Re: upcoming kerneloops.org item: get_page_from_freelist
From: David Rientjes
Date: Thu Jun 25 2009 - 18:26:00 EST
On Thu, 25 Jun 2009, Linus Torvalds wrote:
> So the reason I tend to like the kind of "incrementally try harder"
> approaches is two-fold:
>
> - it works well for balancing different choices against each other (like
> on the freeing path, trying to see which kind of memory is most easily
> freed by trying them all first in a "don't try very hard" mode)
>
If we passed alloc_flags into direct reclaim, we'd be able to know that
priority in increasing severity: ALLOC_HARDER, ALLOC_HIGH,
ALLOC_NO_WATERMARKS. We already know if reclaim should be targeted only
to the set of allowable nodes with the gfp_mask, so ALLOC_CPUSET is
already implicitly handled.
The problem with such an approach is that if we were to, as suggested, set
ALLOC_HARDER for __GFP_WAIT when looping is that we'd soon deplete memory
reserves for potentially long-lived allocations and each zone's min
watermark would simply become min / 2.
Setting ALLOC_HARDER when __GFP_WAIT loops would imply that we'd also
defer the oom killer until we've tried with the lower watermark and that
would result in less memory being available for the TIF_MEMDIE thread when
subsequent allocations still fail.
> - it's great for forcing _everybody_ to do part of the work (ie when some
> new thread comes in and tries to allocate, the new thread starts off
> with the lower priority, and as such won't steal a page that an older
> allocator just freed)
>
So you're seeing the space between the zone's min watermark, min, and min
/ 4 (ALLOC_HARDER) as the pool of memory available to older allocators
that have already done reclaim? It wouldn't necessarily prevent rt tasks
from stealing the memory freed by the older allocators, but would stop all
others.
> And I think that second case is true even for the oom killer case, and
> even for __GFP_FS.
>
The oom killer is serialized only on the zonelist and right before it's
called, we try an allocation with the high watermark to catch parallel oom
killings. The oom killer's heuristic for killing preference is one that
should satisfy all concurrent page allocations on that zonelist, so we
probably don't need to worry about it.
> So if people worry about oom, I would suggest that we should not think so
> hard about the GFP_NOFAIL cases (which are relatively small and rare),
If __GFP_NOFAIL is used in the reclaim path, it will loop forever for
~__GFP_NOMEMALLOC allocations (and the same goes for oom killed tasks).
> or
> about things like the above "try harder" when repeating model, but instead
> think about what actually happens during oom: the most common allocations
> will remain to the page allocations for user faults and/or page cache. In
> fact, they get *more* common as you near OOM situation, because you get
> into the whole swap/filemap thrashing situation where you have to re-read
> the same pages over and over again.
>
> So don't worry about NOFS. Instead, look at what GFP_USER and GFP_HIGHUSER
> do. They set the __GFP_HARDWALL bit, and they _always_ check the end
> result and fail gracefully and quickly when the allocation fails.
>
Yeah, and the oom killer prefers to kill tasks that share a set of
allowable __GFP_HARDWALL nodes with current.
> End result? Realistically, I suspect the _best_ thing we can do is to just
> couple that bit with "we're out of memory", and just do something like
>
> if (!did_some_progress && (gfp_flags & __GFP_HARDWALL))
> goto nopage;
>
> rather than anything else. And I suspect that if we do this, we can then
> afford to retry very aggressively for the allocation cases that aren't
> GFP_USER - and that may well be needed in order to make progress.
>
Agreed.
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/