Re: [PATCH] mm-slab: allocate kmem_cache with __GFP_REPEAT

From: Konstantin Khlebnikov
Date: Wed Jul 20 2011 - 10:32:44 EST


Mel Gorman wrote:
On Wed, Jul 20, 2011 at 08:54:10AM -0500, Christoph Lameter wrote:
On Wed, 20 Jul 2011, Pekka Enberg wrote:

On Wed, 20 Jul 2011, Konstantin Khlebnikov wrote:
The changelog isn't that convincing, really. This is kmem_cache_create()
so I'm surprised we'd ever get NULL here in practice. Does this fix some
problem you're seeing? If this is really an issue, I'd blame the page
allocator as GFP_KERNEL should just work.

nf_conntrack creates separate slab-cache for each net-namespace,
this patch of course not eliminates the chance of failure, but makes it more
acceptable.

I'm still surprised you are seeing failures. mm/slab.c hasn't changed
significantly in a long time. Why hasn't anyone reported this before? I'd
still be inclined to shift the blame to the page allocator... Mel, Christoph?

There was a lot of recent fiddling with the reclaim logic. Maybe some of
those changes caused the problem?


It's more likely that creating new slabs while under memory pressure
significant enough to fail an order-4 allocation is a situation that is
rarely tested.

What kernel version did this failure occur on? What was the system doing
at the time of failure? Can the page allocation failure message be
posted?


I catch this on our rhel6-openvz kernel, and yes it very patchy,
but I don't see any reasons why this cannot be reproduced on mainline kernel.

there was abount ten containers with random stuff, node already do intensive swapout but still alive,
in this situation starting new containers sometimes (1 per 1000) fails due to kmem_cache_create failures in nf_conntrack,
there no other messages except:
Unable to create nf_conn slab cache
and some
nf_conntrack: falling back to vmalloc.
(it try allocates huge hash table and do it via vmalloc if kmalloc fails)
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/