Re: [PATCH v2 1/7] sched/fair: Generalize asym_packing logic for SMT local sched group
From: Ricardo Neri
Date: Wed Dec 28 2022 - 22:51:32 EST
On Thu, Dec 22, 2022 at 04:55:58PM +0000, Valentin Schneider wrote:
> On 22/11/22 12:35, Ricardo Neri wrote:
> > @@ -8926,25 +8924,16 @@ static bool asym_smt_can_pull_tasks(int dst_cpu, struct sd_lb_stats *sds,
> > return sched_asym_prefer(dst_cpu, sg->asym_prefer_cpu);
> > }
> >
> > - /* @dst_cpu has SMT siblings. */
> > -
> > - if (sg_is_smt) {
> > - int local_busy_cpus = sds->local->group_weight -
> > - sds->local_stat.idle_cpus;
> > - int busy_cpus_delta = sg_busy_cpus - local_busy_cpus;
> > -
> > - if (busy_cpus_delta == 1)
> > - return sched_asym_prefer(dst_cpu, sg->asym_prefer_cpu);
> > -
> > - return false;
> > - }
> > -
> > /*
> > - * @sg does not have SMT siblings. Ensure that @sds::local does not end
> > - * up with more than one busy SMT sibling and only pull tasks if there
> > - * are not busy CPUs (i.e., no CPU has running tasks).
> > + * @dst_cpu has SMT siblings. Do asym_packing load balancing only if
> > + * all its siblings are idle (moving tasks between physical cores in
> > + * which some SMT siblings are busy results in the same throughput).
> > + *
> > + * If the difference in the number of busy CPUs is two or more, let
> > + * find_busiest_group() take care of it. We only care if @sg has
> > + * exactly one busy CPU. This covers SMT and non-SMT sched groups.
> > */
> > - if (!sds->local_stat.sum_nr_running)
> > + if (sg_busy_cpus == 1 && !sds->local_stat.sum_nr_running)
> > return sched_asym_prefer(dst_cpu, sg->asym_prefer_cpu);
> >
>
> Some of this is new to me - I had missed the original series introducing
> this. However ISTM that this is conflating two concepts: SMT occupancy
> balancing, and asym packing.
>
> Take the !local_is_smt :: sg_busy_cpus >= 2 :: return true; path. It does
> not involve asym packing priorities at all. This can end up in an
> ASYM_PACKING load balance,
Yes, this the desired result: an idle low-priority CPU should help a high-
priority core with more than one busy SMT sibling. But yes, it does not
relate to priorities and can be implemented differently.
> which per calculate_imbalance() tries to move
> *all* tasks to the higher priority CPU - in the case of SMT balancing,
> we don't want to totally empty the core, just its siblings.
But it will not empty the core, only one of its SMT siblings. A single
sibling will be selected in find_busiest_queue(). The other siblings will
be unaffected.
>
> Is there an ITMT/big.LITTLE (or however x86 calls it) case that invalidates
> the above?
Please see below.
>
> Say, what's not sufficient with the below? AFAICT the only thing that isn't
> covered is the sg_busy_cpus >= 2 thing, but IMO that's SMT balancing, not
> asym packing - if the current calculate_imbalance() doesn't do it, it
> should be fixed to do it.
Agreed.
>Looking at the
>
> local->group_type == group_has_spare
>
> case, it looks like it should DTRT.
I had tried (and failed) to have find_busiest_group() handle the
!local_is_smt :: sg_busy_cpus >= 2 case. Now I think I made it work.
When the busiest group is not overloaded, find_busiest_group() and
local->group = group_has_spare during an idle load balance events the
number of *idle* CPUs. However, this does not work if the local and busiest
groups have different weights. In SMT2, for instance, if busiest has 2 busy
CPUs (i.e., 0 idle CPUs) and the destination CPU is idle, the difference in
the number of idle CPUs is 1. find_busiest_group() will incorrectly goto
out_balanced.
This issue very visible in Intel hybrid processors because the big cores
have SMT but the small cores do not. It can, however, be reproduced in non-
hybrid processors by offlining the SMT siblings of some cores.
The problem can be fixed by instead balancing the number of *busy* CPUs,
which is what in general we want, IMO. (When two groups have the same
weight, it is equivalent to balancing the number of idle CPUs).
This patch worked for me:
@@ -9787,14 +9787,18 @@ static inline void calculate_imbalance(struct lb_env *env, struct sd_lb_stats *s
lsub_positive(&nr_diff, local->sum_nr_running);
env->imbalance = nr_diff;
} else {
+ unsigned int busiest_busy_cpus, local_busy_cpus;
+
+ busiest_busy_cpus = busiest->group_weight - busiest->idle_cpus;
+ local_busy_cpus = local->group_weight - local->idle_cpus;
/*
* If there is no overload, we just want to even the number of
- * idle cpus.
+ * busy cpus.
*/
env->migration_type = migrate_task;
- env->imbalance = max_t(long, 0,
- (local->idle_cpus - busiest->idle_cpus));
+ env->imbalance = max_t(long, 0 ,
+ (busiest_busy_cpus - local_busy_cpus));
}
#ifdef CONFIG_NUMA
@@ -9981,18 +9985,24 @@ static struct sched_group *find_busiest_group(struct lb_env *env)
*/
goto out_balanced;
- if (busiest->group_weight > 1 &&
- local->idle_cpus <= (busiest->idle_cpus + 1))
- /*
- * If the busiest group is not overloaded
- * and there is no imbalance between this and busiest
- * group wrt idle CPUs, it is balanced. The imbalance
- * becomes significant if the diff is greater than 1
- * otherwise we might end up to just move the imbalance
- * on another group. Of course this applies only if
- * there is more than 1 CPU per group.
- */
- goto out_balanced;
+ if (busiest->group_weight > 1) {
+ unsigned int local_busy_cpus, busiest_busy_cpus;
+
+ local_busy_cpus = local->group_weight - local->idle_cpus;
+ busiest_busy_cpus = busiest->group_weight - busiest->idle_cpus;
+
+ if (busiest_busy_cpus <= local_busy_cpus + 1)
+ /*
+ * If the busiest group is not overloaded
+ * and there is no imbalance between this and busiest
+ * group wrt busy CPUs, it is balanced. The imbalance
+ * becomes significant if the diff is greater than 1
+ * otherwise we might end up to just move the imbalance
+ * on another group. Of course this applies only if
+ * there is more than 1 CPU per group.
+ */
+ goto out_balanced;
+ }
if (busiest->sum_h_nr_running == 1)
/*
With this I can remove the sg_busy_cpus >=2 thing from asym_smt_can_pull_tasks().
>
> ---
> diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c
> index 224107278471f..15eb2d3cff186 100644
> --- a/kernel/sched/fair.c
> +++ b/kernel/sched/fair.c
> @@ -9176,12 +9176,15 @@ static inline bool
> sched_asym(struct lb_env *env, struct sd_lb_stats *sds, struct sg_lb_stats *sgs,
> struct sched_group *group)
> {
> - /* Only do SMT checks if either local or candidate have SMT siblings */
> - if ((sds->local->flags & SD_SHARE_CPUCAPACITY) ||
> - (group->flags & SD_SHARE_CPUCAPACITY))
> - return asym_smt_can_pull_tasks(env->dst_cpu, sds, sgs, group);
> + /*
> + * For SMT, env->idle != CPU_NOT_IDLE isn't sufficient, we need to make
> + * sure the whole core is idle.
> + */
> + if (((sds->local->flags & SD_SHARE_CPUCAPACITY) ||
> + (group->flags & SD_SHARE_CPUCAPACITY)) &&
> + !test_idle_cores(env->dst_cpu))
But test_idle_cores() tests for *any* idle core in the highest domain with the
SD_SHARE_PKG_RESOURCES flag. Here we are only interested in the SMT siblings
of env->dst_cpu. If is_core_idle(env->dst_cpu) is used, then I agree with the
change.
Thanks and BR,
Ricardo