Re: [patch] mm: memcontrol: lockless page counters
From: Johannes Weiner
Date: Tue Sep 23 2014 - 13:05:40 EST
On Tue, Sep 23, 2014 at 07:21:50PM +0400, Vladimir Davydov wrote:
> On Tue, Sep 23, 2014 at 09:28:01AM -0400, Johannes Weiner wrote:
> > On Tue, Sep 23, 2014 at 03:06:34PM +0400, Vladimir Davydov wrote:
> > > On Mon, Sep 22, 2014 at 02:57:36PM -0400, Johannes Weiner wrote:
> > > > On Mon, Sep 22, 2014 at 06:41:58PM +0400, Vladimir Davydov wrote:
> > > > > On Fri, Sep 19, 2014 at 09:22:08AM -0400, Johannes Weiner wrote:
> > > > > > diff --git a/include/linux/memcontrol.h b/include/linux/memcontrol.h
> > > > > > index 19df5d857411..bf8fb1a05597 100644
> > > > > > --- a/include/linux/memcontrol.h
> > > > > > +++ b/include/linux/memcontrol.h
> > > > > > @@ -54,6 +54,38 @@ struct mem_cgroup_reclaim_cookie {
> > > > > > };
> > > > > >
> > > > > > #ifdef CONFIG_MEMCG
> > > > > > +
> > > > > > +struct page_counter {
> > > > >
> > > > > I'd place it in a separate file, say
> > > > >
> > > > > include/linux/page_counter.h
> > > > > mm/page_counter.c
> > > > >
> > > > > just to keep mm/memcontrol.c clean.
> > > >
> > > > The page counters are the very core of the memory controller and, as I
> > > > said to Michal, I want to integrate the hugetlb controller into memcg
> > > > as well, at which point there won't be any outside users anymore. So
> > > > I think this is the right place for it.
> > >
> > > Hmm, there might be memcg users out there that don't want to pay for
> > > hugetlb accounting. Or is the overhead supposed to be negligible?
> >
> > Yes. But if it gets in the way, it creates pressure to optimize it.
>
> There always will be an overhead no matter how we optimize it.
>
> I think we should only merge them if it could really help simplify the
> code, for instance if they were dependant on each other. Anyway, I'm not
> an expert in the hugetlb cgroup, so I can't judge whether it's good or
> not. I believe you should raise this topic separately and see if others
> agree with you.
It's not a dependency, it's that there is a lot of redundancy in the
code because they do pretty much the same thing, and ongoing breakage
by stringing such a foreign object along. Those two things have been
a problem with memcg from the beginning and created a lot of grief.
But I agree that it's a separate topic. The only question for now is
whether the page counter should be in its own file. They are pretty
much half of what memory does (account & limit, track & reclaim), so
they are not misplaced in memcontrol.c, and there is one measly user
outside of memcg proper, which is not hurt by having to compile memcg
into the kernel.
> > That's the same reason why I've been trying to integrate memcg into
> > the rest of the VM for over two years now - aside from resulting in
> > much more unified code, it forces us to compete, and it increases our
> > testing exposure by several orders of magnitude.
> >
> > The only reason we are discussing lockless page counters right now is
> > because we got rid of "memcg specialness" and exposed res_counters to
> > the rest of the world; and boy did that instantly raise the bar on us.
> >
> > > Anyway, I still don't think it's a good idea to keep all the definitions
> > > in the same file. memcontrol.c is already huge. Adding more code to it
> > > is not desirable, especially if it can naturally live in a separate
> > > file. And since the page_counter is independent of the memcg core and
> > > *looks* generic, I believe we should keep it separately.
> >
> > It's less code than what I just deleted, and half of it seems
> > redundant when integrated into memcg. This code would benefit a lot
> > from being part of memcg, and memcg could reduce its public API.
>
> I think I understand. You are afraid that other users of the
> page_counter will show up one day, and you won't be able to modify its
> API freely. That's reasonable. But we can solve it while still keeping
> page_counter separately. For example, put the header to mm/ and state
> clearly that it's for memcontrol.c and nobody is allowed to use it w/o a
> special permission.
>
> My point is that it's easier to maintain the code divided in logical
> parts. And page_counter seems to be such a logical part.
It's not about freely changing it, it's that there is no user outside
of memory controlling proper, and there is none in sight.
What makes memcontrol.c hard to deal with is that different things are
interspersed. The page_counter is in its own little section in there,
and the compiler can optimize and inline the important fastpaths.
> Coming to think about placing page_counter.h to mm/, another question
> springs into my mind. Why do you force kmem.tcp code to use the
> page_counter instead of the res_counter? Nobody seems to use it and it
> should pass away sooner or later. May be it's worth leaving kmem.tcp
> using res_counter? I think we could isolate kmem.tcp under a separate
> config option depending on the RES_COUNTER, and mark them both as
> deprecated somehow.
What we usually do is keep changing deprecated or unused code with
minimal/compile-only testing. Eventually somebody will notice that
something major broke while doing this but that noone has complained
in months or even years, at which point we remove it.
I'm curious why you reach the conclusion that we should string *more*
code along for unused interfaces, rather than less?
> > > > > > +int page_counter_charge(struct page_counter *counter, unsigned long nr_pages,
> > > > > > + struct page_counter **fail);
> > > > > > +int page_counter_uncharge(struct page_counter *counter, unsigned long nr_pages);
> > > > > > +int page_counter_limit(struct page_counter *counter, unsigned long limit);
> > > > >
> > > > > Hmm, why not page_counter_set_limit?
> > > >
> > > > Limit is used as a verb here, "to limit". Getters and setters are
> > > > usually wrappers around unusual/complex data structure access,
> > >
> > > Not necessarily. Look at percpu_counter_read e.g. It's a one-line
> > > getter, which we could easily live w/o, but still it's there.
> >
> > It abstracts an unusual and error-prone access to a counter value,
> > i.e. reading an unsigned quantity out of a signed variable.
>
> Wait, percpu_counter_read retval has type s64 and it returns
> percpu_counter->count, which is also s64. So it's a 100% equivalent of
> plain reading of percpu_counter->count.
My bad, I read page_counter instead of percpu_counter. Nonetheless, I
did add page_counter_read().
> > > > but this function does a lot more, so I'm not fond of _set_limit().
> > >
> > > Nevertheless, everything it does can be perfectly described in one
> > > sentence "it tries to set the new value of the limit", so it does
> > > function as a setter. And if there's a setter, there must be a getter
> > > IMO.
> >
> > That's oversimplifying things. Setting a limit requires enforcing a
> > whole bunch of rules and synchronization, whereas reading a limit is
> > accessing an unsigned long.
>
> Let's look at percpu_counter once again (excuse me for sticking to it,
> but it seems to be a good example): percpu_counter_set requires
> enforcing a whole bunch of rules and synchronization, whereas reading
> percpu_counter value is accessing an s64. Nevertheless, they're
> *semantically* a getter and a setter. The same is fair for the
> page_counter IMO.
>
> I don't want to enforce you to introduce the getter, it's not that
> important to me. Just reasoning.
You can always find such comparisons, but the only thing that counts
is the engineering merit, which I don't see for the page limit.
percpu_counter_read() is more obvious, because it's an API that is
expected to be widely used, and the "counter" is actually a complex
data structure. That accessor might choose to do postprocessing like
underflow or percpu-variance correction at some point, and then it can
change callers all over the tree in a single place.
None of this really applies to the page counter limit, however.
> > > > > > break;
> > > > > > case RES_FAILCNT:
> > > > > > - res_counter_reset_failcnt(&h_cg->hugepage[idx]);
> > > > > > + counter->limited = 0;
> > > > >
> > > > > page_counter_reset_failcnt?
> > > >
> > > > That would be more obscure than counter->failcnt = 0, I think.
> > >
> > > There's one thing that bothers me about this patch. Before, all the
> > > functions operating on res_counter were mutually smp-safe, now they
> > > aren't. E.g. if the failcnt reset races with the falcnt increment from
> > > page_counter_try_charge, the reset might be skipped. You only use the
> > > atomic type for the counter, but my guess is that failcnt and watermark
> > > should be atomic too, at least if we're not going to get rid of them
> > > soon. Otherwise, it should be clearly stated that failcnt and watermark
> > > are racy.
> >
> > It's fair enough that the raciness should be documented, but both
> > counters are such roundabout metrics to begin with that it really
> > doesn't matter. What's the difference between a failcnt of 590 and
> > 600 in practical terms? And what does it matter if the counter
> > watermark is off by a few pages when there are per-cpu caches on top
> > of the counters, and the majority of workloads peg the watermark to
> > the limit during startup anyway?
>
> Suppose failcnt=42000. The user resets it and gets 42001 instead of 0.
> That's a huge difference.
I guess that's true, but I still have a really hard time caring. Who
resets this in the middle of ongoing execution? Who resets this at
all instead of just comparing before/after snapshots, like with all
other mm stats? And is anybody even using these pointless metrics?
I'm inclined to just put stop_machine() into the reset functions...
> > > > > > +int page_counter_limit(struct page_counter *counter, unsigned long limit)
> > > > > > +{
> > > > > > + for (;;) {
> > > > > > + unsigned long count;
> > > > > > + unsigned long old;
> > > > > > +
> > > > > > + count = atomic_long_read(&counter->count);
> > > > > > +
> > > > > > + old = xchg(&counter->limit, limit);
> > > > > > +
> > > > > > + if (atomic_long_read(&counter->count) != count) {
> > > > > > + counter->limit = old;
> > > > >
> > > > > I wonder what can happen if two threads execute this function
> > > > > concurrently... or may be it's not supposed to be smp-safe?
> > > >
> > > > memcg already holds the set_limit_mutex here. I updated the tcp and
> > > > hugetlb controllers accordingly to take limit locks as well.
> > >
> > > I would prefer page_counter to handle it internally, because we won't
> > > need the set_limit_mutex once memsw is converted to plain swap
> > > accounting.
> > >
> > > Besides, memcg_update_kmem_limit doesn't take it. Any chance to
> > > achieve that w/o spinlocks, using only atomic variables?
> >
> > We still need it to serialize concurrent access to the memory limit,
> > and I updated the patch to have kmem take it as well. It's such a
> > cold path that using a lockless scheme and worrying about coherency
> > between updates is not worth it, I think.
>
> OK, it's not that important actually. Please state explicitly in the
> comment to the function definition that it needs external
> synchronization then.
Yeah, I documented everything now.
How about the following update? Don't be thrown by the
page_counter_cancel(), I went back to it until we find something more
suitable. But as long as it's documented and has only 1.5 callsites,
it shouldn't matter all that much TBH.
Thanks for your invaluable feedback so far, and sorry if the original
patch was hard to review. I'll try to break it up, to me it's usually
easier to verify new functions by looking at the callers in the same
patch, but I can probably remove the res_counter in a follow-up patch.
---