Re: [RFC PATCH v7 08/23] sched: Add core wide task selection and scheduling.
From: Peter Zijlstra
Date: Fri Aug 28 2020 - 16:52:59 EST
On Fri, Aug 28, 2020 at 03:51:09PM -0400, Julien Desfossez wrote:
> + smt_weight = cpumask_weight(smt_mask);
> + for_each_cpu_wrap_or(i, smt_mask, cpumask_of(cpu), cpu) {
> + struct rq *rq_i = cpu_rq(i);
> + struct task_struct *p;
> +
> + /*
> + * During hotplug online a sibling can be added in
> + * the smt_mask * while we are here. If so, we would
> + * need to restart selection by resetting all over.
> + */
> + if (unlikely(smt_weight != cpumask_weight(smt_mask)))
> + goto retry_select;
cpumask_weigt() is fairly expensive, esp. for something that should
'never' happen.
What exactly is the race here?
We'll update the cpu_smt_mask() fairly early in secondary bringup, but
where does it become a problem?
The moment the new thread starts scheduling it'll block on the common
rq->lock and then it'll cycle task_seq and do a new pick.
So where do things go side-ways?
Can we please split out this hotplug 'fix' into a separate patch with a
coherent changelog.