Re: [PATCH] mm: stop kswapd's infinite loop at high orderallocation

From: Mel Gorman
Date: Wed Dec 31 2008 - 07:06:18 EST


On Wed, Dec 31, 2008 at 10:59:58AM +0200, wassim dagash wrote:
> Hi ,
> Thank you all for reviewing.
> Why don't we implement a solution where the order is defined per zone?

kswapd is scanning trying to balance the whole system, not just one of
the zones. You don't know which of the zones would be the best to
rebalance and glancing through your path, the zone that would be
balanced is the last zone in the zonelist. i.e. kswapd will try to
rebalance the least-preferred zone to be used by the calling process.

> I implemented such a solution for my kernel (2.6.18) and tested it, it
> worked fine for me. Attached a patch with a solution for 2.6.28
> (compile tested only).
>

One big one;

zone_watermark_ok() now always calculates order based on
zone->kswapd_max_order. This means that a process that wants an order-0 page
may check watermarks at order-10 because some other process overwrote the
zone value. That is unfair.

One less obvious one;

You add a field that is written often beside a field that is read-mostly
in struct zone. This results in poorer cache behaviour. It's easily
fixed though.

> On Wed, Dec 31, 2008 at 6:54 AM, KOSAKI Motohiro
> <kosaki.motohiro@xxxxxxxxxxxxxx> wrote:
> > Hi
> >
> > thank you for reviewing.
> >
> >>> ==
> >>> From: KOSAKI Motohiro <kosaki.motohiro@xxxxxxxxxxxxxx>
> >>> Subject: [PATCH] mm: kswapd stop infinite loop at high order allocation
> >>>
> >>> Wassim Dagash reported following kswapd infinite loop problem.
> >>>
> >>> kswapd runs in some infinite loop trying to swap until order 10 of zone
> >>> highmem is OK, While zone higmem (as I understand) has nothing to do
> >>> with contiguous memory (cause there is no 1-1 mapping) which means
> >>> kswapd will continue to try to balance order 10 of zone highmem
> >>> forever (or until someone release a very large chunk of highmem).
> >>>
> >>> He proposed remove contenious checking on highmem at all.
> >>> However hugepage on highmem need contenious highmem page.
> >>>
> >>
> >> I'm lacking the original problem report, but contiguous order-10 pages are
> >> indeed required for hugepages in highmem and reclaiming for them should not
> >> be totally disabled at any point. While no 1-1 mapping exists for the kernel,
> >> contiguity is still required.
> >
> > correct.
> > but that's ok.
> >
> > my patch only change corner case bahavior and only disable high-order
> > when priority==0. typical hugepage reclaim don't need and don't reach
> > priority==0.
> >
> > and sorry. I agree with my "2nd loop" word of the patch comment is a
> > bit misleading.
> >
> >
> >> kswapd gets a sc.order when it is known there is a process trying to get
> >> high-order pages so it can reclaim at that order in an attempt to prevent
> >> future direct reclaim at a high-order. Your patch does not appear to depend on
> >> GFP_KERNEL at all so I found the comment misleading. Furthermore, asking it to
> >> loop again at order-0 means it may scan and reclaim more memory unnecessarily
> >> seeing as all_zones_ok was calculated based on a high-order value, not order-0.
> >
> > Yup. my patch doesn't depend on GFP_KERNEL.
> >
> > but, Why order-0 means it may scan more memory unnecessary?
> > all_zones_ok() is calculated by zone_watermark_ok() and zone_watermark_ok()
> > depend on order argument. and my patch set order variable to 0 too.
> >
> >
> >> While constantly looping trying to balance for high-orders is indeed bad,
> >> I'm unconvinced this is the correct change. As we have already gone through
> >> a priorities and scanned everything at the high-order, would it not make
> >> more sense to do just give up with something like the following?
> >>
> >> /*
> >> * If zones are still not balanced, loop again and continue attempting
> >> * to rebalance the system. For high-order allocations, fragmentation
> >> * can prevent the zones being rebalanced no matter how hard kswapd
> >> * works, particularly on systems with little or no swap. For costly
> >> * orders, just give up and assume interested processes will either
> >> * direct reclaim or wake up kswapd as necessary.
> >> */
> >> if (!all_zones_ok && sc.order <= PAGE_ALLOC_COSTLY_ORDER) {
> >> cond_resched();
> >>
> >> try_to_freeze();
> >>
> >> goto loop_again;
> >> }
> >>
> >> I used PAGE_ALLOC_COSTLY_ORDER instead of sc.order == 0 because we are
> >> expected to support allocations up to that order in a fairly reliable fashion.
> >
> > my comment is bellow.
> >
> >
> >> =============
> >> From: Mel Gorman <mel@xxxxxxxxx>
> >> Subject: [PATCH] mm: stop kswapd's infinite loop at high order allocation
> >>
> >> kswapd runs in some infinite loop trying to swap until order 10 of zone
> >> highmem is OK.... kswapd will continue to try to balance order 10 of zone
> >> highmem forever (or until someone release a very large chunk of highmem).
> >>
> >> For costly high-order allocations, the system may never be balanced due to
> >> fragmentation but kswapd should not infinitely loop as a result. The
> >> following patch lets kswapd stop reclaiming in the event it cannot
> >> balance zones and the order is high-order.
> >>
> >> Reported-by: wassim dagash <wassim.dagash@xxxxxxxxx>
> >> Signed-off-by: Mel Gorman <mel@xxxxxxxxx>
> >>
> >> diff --git a/mm/vmscan.c b/mm/vmscan.c
> >> index 62e7f62..03ed9a0 100644
> >> --- a/mm/vmscan.c
> >> +++ b/mm/vmscan.c
> >> @@ -1867,7 +1867,16 @@ out:
> >>
> >> zone->prev_priority = temp_priority[i];
> >> }
> >> - if (!all_zones_ok) {
> >> +
> >> + /*
> >> + * If zones are still not balanced, loop again and continue attempting
> >> + * to rebalance the system. For high-order allocations, fragmentation
> >> + * can prevent the zones being rebalanced no matter how hard kswapd
> >> + * works, particularly on systems with little or no swap. For costly
> >> + * orders, just give up and assume interested processes will either
> >> + * direct reclaim or wake up kswapd as necessary.
> >> + */
> >> + if (!all_zones_ok && sc.order <= PAGE_ALLOC_COSTLY_ORDER) {
> >> cond_resched();
> >>
> >> try_to_freeze();
> >
> > this patch seems no good.
> > kswapd come this point every SWAP_CLUSTER_MAX reclaimed because to avoid
> > unnecessary priority variable decreasing.
> > then "nr_reclaimed >= SWAP_CLUSTER_MAX" indicate kswapd need reclaim more.
> >
> > kswapd purpose is "reclaim until pages_high", not reclaim
> > SWAP_CLUSTER_MAX pages.
> >
> > if your patch applied and kswapd start to reclaim for hugepage, kswapd
> > exit balance_pgdat() function after to reclaim only 32 pages
> > (SWAP_CLUSTER_MAX).
> >
> > In the other hand, "nr_reclaimed < SWAP_CLUSTER_MAX" mean kswapd can't
> > reclaim enough
> > page although priority == 0.
> > in this case, retry is worthless.
> >
> > sorting out again.
> > "goto loop_again" reaching happend by two case.
> >
> > 1. kswapd reclaimed SWAP_CLUSTER_MAX pages.
> > at that time, kswapd reset priority variable to prevent
> > unnecessary priority decreasing.
> > I don't hope this behavior change.
> > 2. kswapd scanned until priority==0.
> > this case is debatable. my patch reset any order to 0. but
> > following code is also considerable to me. (sorry for tab corrupted,
> > current my mail environment is very poor)
> >
> >
> > code-A:
> > if (!all_zones_ok) {
> > if ((nr_reclaimed >= SWAP_CLUSTER_MAX) ||
> > (sc.order <= PAGE_ALLOC_COSTLY_ORDER)) {
> > cond_resched();
> > try_to_freeze();
> > goto loop_again;
> > }
> > }
> >
> > or
> >
> > code-B:
> > if (!all_zones_ok) {
> > cond_resched();
> > try_to_freeze();
> >
> > if (nr_reclaimed >= SWAP_CLUSTER_MAX)
> > goto loop_again;
> >
> > if (sc.order <= PAGE_ALLOC_COSTLY_ORDER)) {
> > order = sc.order = 0;
> > goto loop_again;
> > }
> > }
> >
> >
> > However, I still like my original proposal because ..
> > - code-A forget to order-1 (for stack) allocation also can cause
> > infinite loop.
> > - code-B doesn't simpler than my original proposal.
> >
> > What do you think it?
> >
>
>
>
> --
> too much is never enough!!!!!



--
Mel Gorman
Part-time Phd Student Linux Technology Center
University of Limerick IBM Dublin Software Lab
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/