Re: [PATCH] sched: Fix data-race in wakeup

From: Peter Zijlstra
Date: Tue Nov 17 2020 - 04:30:09 EST


On Tue, Nov 17, 2020 at 09:15:46AM +0000, Will Deacon wrote:
> On Tue, Nov 17, 2020 at 09:30:16AM +0100, Peter Zijlstra wrote:

> > Subject: sched: Fix data-race in wakeup
> > From: Peter Zijlstra <peterz@xxxxxxxxxxxxx>
> > Date: Tue Nov 17 09:08:41 CET 2020
> >
> > Mel reported that on some ARM64 platforms loadavg goes bananas and
> > tracked it down to the following data race:
> >
> > CPU0 CPU1
> >
> > schedule()
> > prev->sched_contributes_to_load = X;
> > deactivate_task(prev);
> >
> > try_to_wake_up()
> > if (p->on_rq &&) // false
> > if (smp_load_acquire(&p->on_cpu) && // true
> > ttwu_queue_wakelist())
> > p->sched_remote_wakeup = Y;
> >
> > smp_store_release(prev->on_cpu, 0);
>
> (nit: I suggested this race over at [1] ;)

Ah, I'll ammend and get you a Debugged-by line or something ;-)

> > where both p->sched_contributes_to_load and p->sched_remote_wakeup are
> > in the same word, and thus the stores X and Y race (and can clobber
> > one another's data).
> >
> > Whereas prior to commit c6e7bd7afaeb ("sched/core: Optimize ttwu()
> > spinning on p->on_cpu") the p->on_cpu handoff serialized access to
> > p->sched_remote_wakeup (just as it still does with
> > p->sched_contributes_to_load) that commit broke that by calling
> > ttwu_queue_wakelist() with p->on_cpu != 0.
> >
> > However, due to
> >
> > p->XXX ttwu()
> > schedule() if (p->on_rq && ...) // false
> > smp_mb__after_spinlock() if (smp_load_acquire(&p->on_cpu) &&
> > deactivate_task() ttwu_queue_wakelist())
> > p->on_rq = 0; p->sched_remote_wakeup = X;
> >
> > We can be sure any 'current' store is complete and 'current' is
> > guaranteed asleep. Therefore we can move p->sched_remote_wakeup into
> > the current flags word.
> >
> > Note: while the observed failure was loadavg accounting gone wrong due
> > to ttwu() cobbering p->sched_contributes_to_load, the reverse problem
> > is also possible where schedule() clobbers p->sched_remote_wakeup,
> > this could result in enqueue_entity() wrecking ->vruntime and causing
> > scheduling artifacts.
> >
> > Fixes: c6e7bd7afaeb ("sched/core: Optimize ttwu() spinning on p->on_cpu")
> > Reported-by: Mel Gorman <mgorman@xxxxxxxxxxxxxxxxxxx>
> > Signed-off-by: Peter Zijlstra (Intel) <peterz@xxxxxxxxxxxxx>
> > ---
> > include/linux/sched.h | 13 ++++++++++++-
> > 1 file changed, 12 insertions(+), 1 deletion(-)
> >
> > --- a/include/linux/sched.h
> > +++ b/include/linux/sched.h
> > @@ -775,7 +775,6 @@ struct task_struct {
> > unsigned sched_reset_on_fork:1;
> > unsigned sched_contributes_to_load:1;
> > unsigned sched_migrated:1;
> > - unsigned sched_remote_wakeup:1;
> > #ifdef CONFIG_PSI
> > unsigned sched_psi_wake_requeue:1;
> > #endif
> > @@ -785,6 +784,18 @@ struct task_struct {
> >
> > /* Unserialized, strictly 'current' */
> >
> > + /*
> > + * p->in_iowait = 1; ttwu()
> > + * schedule() if (p->on_rq && ..) // false
> > + * smp_mb__after_spinlock(); if (smp_load_acquire(&p->on_cpu) && //true
> > + * deactivate_task() ttwu_queue_wakelist())
> > + * p->on_rq = 0; p->sched_remote_wakeup = X;
> > + *
> > + * Guarantees all stores of 'current' are visible before
> > + * ->sched_remote_wakeup gets used.
>
> I'm still not sure this is particularly clear -- don't we want to highlight
> that the store of p->on_rq is unordered wrt the update to
> p->sched_contributes_to_load() in deactivate_task()?

I can explicitly call that out I suppose.

> I dislike bitfields with a passion, but the fix looks good:

I don't particularly hate them, they're just a flag field with names on
(in this case).

> Acked-by: Will Deacon <will@xxxxxxxxxx>

Thanks!

> Now the million dollar question is why KCSAN hasn't run into this. Hrmph.

kernel/sched/Makefile:KCSAN_SANITIZE := n

might have something to do with that, I suppose.