Re: [PATCH] mm: stop kswapd's infinite loop at high order allocation
From: KOSAKI Motohiro
Date: Tue Dec 30 2008 - 23:54:31 EST
Hi
thank you for reviewing.
>> ==
>> From: KOSAKI Motohiro <kosaki.motohiro@xxxxxxxxxxxxxx>
>> Subject: [PATCH] mm: kswapd stop infinite loop at high order allocation
>>
>> Wassim Dagash reported following kswapd infinite loop problem.
>>
>> kswapd runs in some infinite loop trying to swap until order 10 of zone
>> highmem is OK, While zone higmem (as I understand) has nothing to do
>> with contiguous memory (cause there is no 1-1 mapping) which means
>> kswapd will continue to try to balance order 10 of zone highmem
>> forever (or until someone release a very large chunk of highmem).
>>
>> He proposed remove contenious checking on highmem at all.
>> However hugepage on highmem need contenious highmem page.
>>
>
> I'm lacking the original problem report, but contiguous order-10 pages are
> indeed required for hugepages in highmem and reclaiming for them should not
> be totally disabled at any point. While no 1-1 mapping exists for the kernel,
> contiguity is still required.
correct.
but that's ok.
my patch only change corner case bahavior and only disable high-order
when priority==0. typical hugepage reclaim don't need and don't reach
priority==0.
and sorry. I agree with my "2nd loop" word of the patch comment is a
bit misleading.
> kswapd gets a sc.order when it is known there is a process trying to get
> high-order pages so it can reclaim at that order in an attempt to prevent
> future direct reclaim at a high-order. Your patch does not appear to depend on
> GFP_KERNEL at all so I found the comment misleading. Furthermore, asking it to
> loop again at order-0 means it may scan and reclaim more memory unnecessarily
> seeing as all_zones_ok was calculated based on a high-order value, not order-0.
Yup. my patch doesn't depend on GFP_KERNEL.
but, Why order-0 means it may scan more memory unnecessary?
all_zones_ok() is calculated by zone_watermark_ok() and zone_watermark_ok()
depend on order argument. and my patch set order variable to 0 too.
> While constantly looping trying to balance for high-orders is indeed bad,
> I'm unconvinced this is the correct change. As we have already gone through
> a priorities and scanned everything at the high-order, would it not make
> more sense to do just give up with something like the following?
>
> /*
> * If zones are still not balanced, loop again and continue attempting
> * to rebalance the system. For high-order allocations, fragmentation
> * can prevent the zones being rebalanced no matter how hard kswapd
> * works, particularly on systems with little or no swap. For costly
> * orders, just give up and assume interested processes will either
> * direct reclaim or wake up kswapd as necessary.
> */
> if (!all_zones_ok && sc.order <= PAGE_ALLOC_COSTLY_ORDER) {
> cond_resched();
>
> try_to_freeze();
>
> goto loop_again;
> }
>
> I used PAGE_ALLOC_COSTLY_ORDER instead of sc.order == 0 because we are
> expected to support allocations up to that order in a fairly reliable fashion.
my comment is bellow.
> =============
> From: Mel Gorman <mel@xxxxxxxxx>
> Subject: [PATCH] mm: stop kswapd's infinite loop at high order allocation
>
> kswapd runs in some infinite loop trying to swap until order 10 of zone
> highmem is OK.... kswapd will continue to try to balance order 10 of zone
> highmem forever (or until someone release a very large chunk of highmem).
>
> For costly high-order allocations, the system may never be balanced due to
> fragmentation but kswapd should not infinitely loop as a result. The
> following patch lets kswapd stop reclaiming in the event it cannot
> balance zones and the order is high-order.
>
> Reported-by: wassim dagash <wassim.dagash@xxxxxxxxx>
> Signed-off-by: Mel Gorman <mel@xxxxxxxxx>
>
> diff --git a/mm/vmscan.c b/mm/vmscan.c
> index 62e7f62..03ed9a0 100644
> --- a/mm/vmscan.c
> +++ b/mm/vmscan.c
> @@ -1867,7 +1867,16 @@ out:
>
> zone->prev_priority = temp_priority[i];
> }
> - if (!all_zones_ok) {
> +
> + /*
> + * If zones are still not balanced, loop again and continue attempting
> + * to rebalance the system. For high-order allocations, fragmentation
> + * can prevent the zones being rebalanced no matter how hard kswapd
> + * works, particularly on systems with little or no swap. For costly
> + * orders, just give up and assume interested processes will either
> + * direct reclaim or wake up kswapd as necessary.
> + */
> + if (!all_zones_ok && sc.order <= PAGE_ALLOC_COSTLY_ORDER) {
> cond_resched();
>
> try_to_freeze();
this patch seems no good.
kswapd come this point every SWAP_CLUSTER_MAX reclaimed because to avoid
unnecessary priority variable decreasing.
then "nr_reclaimed >= SWAP_CLUSTER_MAX" indicate kswapd need reclaim more.
kswapd purpose is "reclaim until pages_high", not reclaim
SWAP_CLUSTER_MAX pages.
if your patch applied and kswapd start to reclaim for hugepage, kswapd
exit balance_pgdat() function after to reclaim only 32 pages
(SWAP_CLUSTER_MAX).
In the other hand, "nr_reclaimed < SWAP_CLUSTER_MAX" mean kswapd can't
reclaim enough
page although priority == 0.
in this case, retry is worthless.
sorting out again.
"goto loop_again" reaching happend by two case.
1. kswapd reclaimed SWAP_CLUSTER_MAX pages.
at that time, kswapd reset priority variable to prevent
unnecessary priority decreasing.
I don't hope this behavior change.
2. kswapd scanned until priority==0.
this case is debatable. my patch reset any order to 0. but
following code is also considerable to me. (sorry for tab corrupted,
current my mail environment is very poor)
code-A:
if (!all_zones_ok) {
if ((nr_reclaimed >= SWAP_CLUSTER_MAX) ||
(sc.order <= PAGE_ALLOC_COSTLY_ORDER)) {
cond_resched();
try_to_freeze();
goto loop_again;
}
}
or
code-B:
if (!all_zones_ok) {
cond_resched();
try_to_freeze();
if (nr_reclaimed >= SWAP_CLUSTER_MAX)
goto loop_again;
if (sc.order <= PAGE_ALLOC_COSTLY_ORDER)) {
order = sc.order = 0;
goto loop_again;
}
}
However, I still like my original proposal because ..
- code-A forget to order-1 (for stack) allocation also can cause
infinite loop.
- code-B doesn't simpler than my original proposal.
What do you think it?
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/