Re: [patch 12/13] mm: memcontrol: rewrite charge API
From: Johannes Weiner
Date: Mon Jul 14 2014 - 13:13:47 EST
On Mon, Jul 14, 2014 at 05:04:46PM +0200, Michal Hocko wrote:
> Hi,
> I've finally manage to untagle myself from internal stuff...
>
> On Wed 18-06-14 16:40:44, Johannes Weiner wrote:
> > The memcg charge API charges pages before they are rmapped - i.e. have
> > an actual "type" - and so every callsite needs its own set of charge
> > and uncharge functions to know what type is being operated on. Worse,
> > uncharge has to happen from a context that is still type-specific,
> > rather than at the end of the page's lifetime with exclusive access,
> > and so requires a lot of synchronization.
> >
> > Rewrite the charge API to provide a generic set of try_charge(),
> > commit_charge() and cancel_charge() transaction operations, much like
> > what's currently done for swap-in:
> >
> > mem_cgroup_try_charge() attempts to reserve a charge, reclaiming
> > pages from the memcg if necessary.
> >
> > mem_cgroup_commit_charge() commits the page to the charge once it
> > has a valid page->mapping and PageAnon() reliably tells the type.
> >
> > mem_cgroup_cancel_charge() aborts the transaction.
> >
> > This reduces the charge API and enables subsequent patches to
> > drastically simplify uncharging.
> >
> > As pages need to be committed after rmap is established but before
> > they are added to the LRU, page_add_new_anon_rmap() must stop doing
> > LRU additions again. Revive lru_cache_add_active_or_unevictable().
>
> I think it would make more sense to do
> lru_cache_add_active_or_unevictable in a separate patch for easier
> review. Too late, though...
>
> Few comments bellow
> > Signed-off-by: Johannes Weiner <hannes@xxxxxxxxxxx>
>
> The patch looks correct but the code is quite tricky so I hope I didn't
> miss anything.
>
> Acked-by: Michal Hocko <mhocko@xxxxxxx>
Thanks!
> > @@ -54,28 +54,11 @@ struct mem_cgroup_reclaim_cookie {
> > };
> >
> > #ifdef CONFIG_MEMCG
> > -/*
> > - * All "charge" functions with gfp_mask should use GFP_KERNEL or
> > - * (gfp_mask & GFP_RECLAIM_MASK). In current implementatin, memcg doesn't
> > - * alloc memory but reclaims memory from all available zones. So, "where I want
> > - * memory from" bits of gfp_mask has no meaning. So any bits of that field is
> > - * available but adding a rule is better. charge functions' gfp_mask should
> > - * be set to GFP_KERNEL or gfp_mask & GFP_RECLAIM_MASK for avoiding ambiguous
> > - * codes.
> > - * (Of course, if memcg does memory allocation in future, GFP_KERNEL is sane.)
> > - */
>
> I think we should slightly modify the comment but the primary idea
> should stay there. What about the following?
> /*
> * Although memcg charge functions do not allocate any memory they are
> * still getting GFP mask to control the reclaim process (therefore
> * gfp_mask & GFP_RECLAIM_MASK is expected).
> * GFP_KERNEL should be used for the general charge path without any
> * constraints for the reclaim
> * __GFP_WAIT should be cleared for atomic contexts
> * __GFP_NORETRY should be set for charges which might fail rather than
> * spend too much time reclaiming
> * __GFP_NOFAIL should be set for charges which cannot fail.
> */
What *is* the primary idea here?
Taking any kind of gfp mask and interpreting the bits that pertain to
you is done in a lot of places already, and there really is no need to
duplicate the documentation and risk it getting stale and misleading.
> > @@ -948,6 +951,7 @@ static int do_huge_pmd_wp_page_fallback(struct mm_struct *mm,
> > struct page *page,
> > unsigned long haddr)
> > {
> > + struct mem_cgroup *memcg;
> > spinlock_t *ptl;
> > pgtable_t pgtable;
> > pmd_t _pmd;
> > @@ -968,20 +972,21 @@ static int do_huge_pmd_wp_page_fallback(struct mm_struct *mm,
> > __GFP_OTHER_NODE,
> > vma, address, page_to_nid(page));
> > if (unlikely(!pages[i] ||
> > - mem_cgroup_charge_anon(pages[i], mm,
> > - GFP_KERNEL))) {
> > + mem_cgroup_try_charge(pages[i], mm, GFP_KERNEL,
> > + &memcg))) {
> > if (pages[i])
> > put_page(pages[i]);
> > - mem_cgroup_uncharge_start();
> > while (--i >= 0) {
> > - mem_cgroup_uncharge_page(pages[i]);
> > + memcg = (void *)page_private(pages[i]);
>
> Hmm, OK the memcg couldn't go away even if mm owner has left it because
> the charge is already there and the page is not on LRU so the
> mem_cgroup_css_free will wait until we uncharge it or put to LRU.
Yep, res_counter charges have always pinned the memcg. We already
used this exact protocol and relied on the same lifetime rules for
swapin charging.
> > +/**
> > + * mem_cgroup_commit_charge - commit a page charge
> > + * @page: page to charge
> > + * @memcg: memcg to charge the page to
> > + * @lrucare: page might be on LRU already
> > + *
> > + * Finalize a charge transaction started by mem_cgroup_try_charge(),
> > + * after page->mapping has been set up. This must happen atomically
> > + * as part of the page instantiation, i.e. under the page table lock
> > + * for anonymous pages, under the page lock for page and swap cache.
> > + *
> > + * In addition, the page must not be on the LRU during the commit, to
> > + * prevent racing with task migration. If it might be, use @lrucare.
> > + *
> > + * Use mem_cgroup_cancel_charge() to cancel the transaction instead.
> > + */
> > +void mem_cgroup_commit_charge(struct page *page, struct mem_cgroup *memcg,
> > + bool lrucare)
>
> I think we should be explicit that this is only required for LRU pages.
> kmem doesn't have to finalize the transaction.
The function itself only applies to user/LRU pages. kmem has its own
separate API for charge/commit/cancel/uncharge.
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/