Re: [PATCH v3 05/10] lib/refcount: Improve performance of generic REFCOUNT_FULL code

From: Will Deacon
Date: Wed Oct 09 2019 - 12:44:41 EST


Hi Peter,

Thanks for having another look, even though you don't like it.

On Wed, Oct 09, 2019 at 11:25:08AM +0200, Peter Zijlstra wrote:
> On Mon, Oct 07, 2019 at 04:46:58PM +0100, Will Deacon wrote:
> > Rewrite the generic REFCOUNT_FULL implementation so that the saturation
> > point is moved to INT_MIN / 2. This allows us to defer the sanity checks
> > until after the atomic operation, which removes many uses of cmpxchg()
> > in favour of atomic_fetch_{add,sub}().
>
> It also radicaly changes behaviour, and afaict is subtly broken, see
> below.

Cheers. Replies below.

> > Some crude perf results obtained from lkdtm show substantially less
> > overhead, despite the checking:
> >
> > $ perf stat -r 3 -B -- echo {ATOMIC,REFCOUNT}_TIMING >/sys/kernel/debug/provoke-crash/DIRECT
> >
> > # arm64
> > ATOMIC_TIMING: 46.50451 +- 0.00134 seconds time elapsed ( +- 0.00% )
> > REFCOUNT_TIMING (REFCOUNT_FULL, mainline): 77.57522 +- 0.00982 seconds time elapsed ( +- 0.01% )
> > REFCOUNT_TIMING (REFCOUNT_FULL, this series): 48.7181 +- 0.0256 seconds time elapsed ( +- 0.05% )
> >
> > # x86
> > ATOMIC_TIMING: 31.6225 +- 0.0776 seconds time elapsed ( +- 0.25% )
> > REFCOUNT_TIMING (!REFCOUNT_FULL, mainline/x86 asm): 31.6689 +- 0.0901 seconds time elapsed ( +- 0.28% )
> > REFCOUNT_TIMING (REFCOUNT_FULL, mainline): 53.203 +- 0.138 seconds time elapsed ( +- 0.26% )
> > REFCOUNT_TIMING (REFCOUNT_FULL, this series): 31.7408 +- 0.0486 seconds time elapsed ( +- 0.15% )
>
> I would _really_ like words on how this is racy and how it probably
> doesn't matter.

Sure thing, but I'll stick that into include/linux/refcount.h so that
it can sit alongside the other comments in there.

> > Cc: Ingo Molnar <mingo@xxxxxxxxxx>
> > Cc: Elena Reshetova <elena.reshetova@xxxxxxxxx>
> > Cc: Peter Zijlstra <peterz@xxxxxxxxxxxxx>
> > Cc: Ard Biesheuvel <ard.biesheuvel@xxxxxxxxxx>
> > Tested-by: Hanjun Guo <guohanjun@xxxxxxxxxx>
> > Tested-by: Jan Glauber <jglauber@xxxxxxxxxxx>
> > Reviewed-by: Kees Cook <keescook@xxxxxxxxxxxx>
> > Signed-off-by: Will Deacon <will@xxxxxxxxxx>
> > ---
> > include/linux/refcount.h | 87 ++++++++++++++++------------------------
> > 1 file changed, 34 insertions(+), 53 deletions(-)
> >
> > diff --git a/include/linux/refcount.h b/include/linux/refcount.h
> > index e719b5b1220e..7f9aa6511142 100644
> > +++ b/include/linux/refcount.h
> > @@ -47,8 +47,8 @@ static inline unsigned int refcount_read(const refcount_t *r)
> > #ifdef CONFIG_REFCOUNT_FULL
> > #include <linux/bug.h>
> >
> > +#define REFCOUNT_MAX INT_MAX
> > +#define REFCOUNT_SATURATED (INT_MIN / 2)
> >
> > /*
> > * Variant of atomic_t specialized for reference counts.
> > @@ -109,25 +109,19 @@ static inline unsigned int refcount_read(const refcount_t *r)
> > */
> > static inline __must_check bool refcount_add_not_zero(int i, refcount_t *r)
> > {
> > + int old = refcount_read(r);
> >
> > do {
> > + if (!old)
> > + break;
> > + } while (!atomic_try_cmpxchg_relaxed(&r->refs, &old, old + i));
> >
> > + if (unlikely(old < 0 || old + i < 0)) {
>
> So this is obviously racy against itself and other operations.
> Particularly refcount_read(), as the sole API member that actually
> exposes the value, is affected.
>
> Yes, it shouldn't happen and we'll have trouble if we ever hit this, but
> are all refcount_read() users sane enough to not cause further trouble?

Most refcount_read() users seem to be checking against a value of 0, 1
or 2, so yes, if you overflowed all the way around then you could cause
them to go wrong but I think that's just the same old discussion we've
been having about the impractical race. Given the racy nature of
using the result of refcount_read(), I don't see the saturated value
causing issues here (it will just look like a large refcount value).

There are more general issues around whether or not saturation semantics
can lead to bugs other that memory leaks, but they're not unique to this
patch series.

> > + refcount_set(r, REFCOUNT_SATURATED);
> > + WARN_ONCE(1, "refcount_t: saturated; leaking memory.\n");
> > + }
> >
> > + return old;
> > }
> >
> > /**
> > @@ -148,7 +142,13 @@ static inline __must_check bool refcount_add_not_zero(int i, refcount_t *r)
> > */
> > static inline void refcount_add(int i, refcount_t *r)
> > {
> > + int old = atomic_fetch_add_relaxed(i, &r->refs);
> > +
> > + WARN_ONCE(!old, "refcount_t: addition on 0; use-after-free.\n");
>
> This is a change in behaviour vs the old one; the previous verion would
> not change the value this will.

Right. The new algorithm very much performs the bad operation and *then*
checks to see if it was bad, at which point it saturates. That's what
allows us to remove the cmpxchg() loops, but it does so at the cost of
opening up the race where you repeatedly perform operations on the bad
state before anybody has a chance to saturate.

> Is it important, I don't know, but it's not documented.

I'll try to document this too.

> > + if (unlikely(old <= 0 || old + i <= 0)) {
> > + refcount_set(r, REFCOUNT_SATURATED);
> > + WARN_ONCE(old, "refcount_t: saturated; leaking memory.\n");
> > + }
> > }
> >
> > /**
>
> > @@ -224,26 +208,19 @@ static inline void refcount_inc(refcount_t *r)
> > */
> > static inline __must_check bool refcount_sub_and_test(int i, refcount_t *r)
> > {
> > + int old = atomic_fetch_sub_release(i, &r->refs);
> >
> > + if (old == i) {
> > smp_acquire__after_ctrl_dep();
> > return true;
> > }
> >
> > + if (unlikely(old - i < 0)) {
> > + refcount_set(r, REFCOUNT_SATURATED);
> > + WARN_ONCE(1, "refcount_t: underflow; use-after-free.\n");
> > + }
>
> I'm failing to see how this preserves REFCOUNT_SATURATED for
> non-underflow. AFAICT this should have:
>
> if (unlikely(old == REFCOUNT_SATURATED || old - i < 0))

Well spotted! I think we just want:

if (unlikely(old < 0 || old - i < 0))

here, which is reassuringly similar to the logic in refcount_add() and
refcount_add_not_zero().

> > + return false;
> > }
> >
> > /**
> > @@ -276,9 +253,13 @@ static inline __must_check bool refcount_dec_and_test(refcount_t *r)
> > */
> > static inline void refcount_dec(refcount_t *r)
> > {
> > + int old = atomic_fetch_sub_release(1, &r->refs);
> >
> > + if (unlikely(old <= 1)) {
>
> Idem.

Hmm, I don't get what you mean with the one, since we're looking at the
old value. REFCOUNT_SATURATED is negative, so it will do the right thing.

> > + refcount_set(r, REFCOUNT_SATURATED);
> > + WARN_ONCE(1, "refcount_t: decrement hit 0; leaking memory.\n");
> > + }
> > +}
>
> Also, things like refcount_dec_not_one() might need fixing to preserve
> REFCOUNT_SATURATED, because they're not expecting that value to actually
> change, but you do!

refcount_dec_not_one() already checks for REFCOUNT_SATURATED and, in the
case of a racing thread setting the saturated value, the cmpxchg() will
fail if the saturated value is written after the check or the saturated
value will overwrite the value written by the cmpxchg(). Is there another
race that you're thinking of?

Cheers,

Will