Re: [patch 0/8] mm: memcg naturalization -rc2

From: Greg Thelen
Date: Wed Jun 01 2011 - 20:36:14 EST


On Wed, Jun 1, 2011 at 4:52 PM, Hiroyuki Kamezawa
<kamezawa.hiroyuki@xxxxxxxxx> wrote:
> 2011/6/1 Johannes Weiner <hannes@xxxxxxxxxxx>:
>> Hi,
>>
>> this is the second version of the memcg naturalization series.  The
>> notable changes since the first submission are:
>>
>>    o the hierarchy walk is now intermittent and will abort and
>>      remember the last scanned child after sc->nr_to_reclaim pages
>>      have been reclaimed during the walk in one zone (Rik)
>>
>>    o the global lru lists are never scanned when memcg is enabled
>>      after #2 'memcg-aware global reclaim', which makes this patch
>>      self-sufficient and complete without requiring the per-memcg lru
>>      lists to be exclusive (Michal)
>>
>>    o renamed sc->memcg and sc->current_memcg to sc->target_mem_cgroup
>>      and sc->mem_cgroup and fixed their documentation, I hope this is
>>      better understandable now (Rik)
>>
>>    o the reclaim statistic counters have been renamed.  there is no
>>      more distinction between 'pgfree' and 'pgsteal', it is now
>>      'pgreclaim' in both cases; 'kswapd' has been replaced by
>>      'background'
>>
>>    o fixed a nasty crash in the hierarchical soft limit check that
>>      happened during global reclaim in memcgs that are hierarchical
>>      but have no hierarchical parents themselves
>>
>>    o properly implemented the memcg-aware unevictable page rescue
>>      scanner, there were several blatant bugs in there
>>
>>    o documentation on new public interfaces
>>
>> Thanks for your input on the first version.
>>
>> I ran microbenchmarks (sparse file catting, essentially) to stress
>> reclaim and LRU operations.  There is no measurable overhead for
>> !CONFIG_MEMCG, memcg disabled during boot, memcg enabled but no
>> configured groups, and hard limit reclaim.
>>
>> I also ran single-threaded kernbenchs in four unlimited memcgs in
>> parallel, contained in a hard-limited hierarchical parent that put
>> constant pressure on the workload.  There is no measurable difference
>> in runtime, the pgpgin/pgpgout counters, and fairness among memcgs in
>> this test compared to an unpatched kernel.  Needs more evaluation,
>> especially with a higher number of memcgs.
>>
>> The soft limit changes are also proven to work in so far that it is
>> possible to prioritize between children in a hierarchy under pressure
>> and that runtime differences corresponded directly to the soft limit
>> settings in the previously described kernbench setup with staggered
>> soft limits on the groups, but this needs quantification.
>>
>> Based on v2.6.39.
>>
>
> Hmm, I welcome and will review this patches but.....some points I want to say.
>
> 1. No more conflict with Ying's work ?
>    Could you explain what she has and what you don't in this v2 ?
>    If Ying's one has something good to be merged to your set, please
> include it.
>
> 2. it's required to see performance score in commit log.
>
> 3. I think dirty_ratio as 1st big patch to be merged. (But...hmm..Greg ?
>    My patches for asynchronous reclaim is not very important. I can rework it.

I am testing the next version (v8) of the memcg dirty ratio patches. I expect
to have it posted for review later this week.

> 4. This work can be splitted into some small works.
>     a) fix for current code and clean ups
>     a') statistics
>     b) soft limit rework
>     c) change global reclaim
>
>  I like (a)->(b)->(c) order. and while (b) you can merge your work
> with Ying's one.
>  And for a') , I'd like to add a new file memory.reclaim_stat as I've
> already shown.
>  and allow resetting.
>
>  Hmm, how about splitting patch 2/8 into small patches and see what happens in
>  3.2 or 3.3 ? While that, we can make softlimit works better.
>  (and once we do 2/8, our direction will be fixed to the direction to
> remove global LRU.)
>
> 5. please write documentation to explain what new LRU do.
>
> BTW, after this work, lists of ROOT cgroup comes again. I may need to check
> codes which see memcg is ROOT or not. Because we removed many atomic
> ops in memcg, I wonder ROOT cgroup can be accounted again..
>
> Thanks,
> -Kame
>
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/