Re: [PATCH v6 07/14] sched/topology: Introduce sched_energy_present static key
From: Patrick Bellasi
Date: Thu Aug 30 2018 - 05:23:39 EST
On 29-Aug 18:20, Quentin Perret wrote:
> On Wednesday 29 Aug 2018 at 17:50:58 (+0100), Patrick Bellasi wrote:
> > > +/*
> > > + * The complexity of the Energy Model is defined as: nr_pd * (nr_cpus + nr_cs)
> > > + * with: 'nr_pd' the number of performance domains; 'nr_cpus' the number of
> > > + * CPUs; and 'nr_cs' the sum of the capacity states numbers of all performance
> > > + * domains.
> > > + *
> > > + * It is generally not a good idea to use such a model in the wake-up path on
> > > + * very complex platforms because of the associated scheduling overheads. The
> > > + * arbitrary constraint below prevents that. It makes EAS usable up to 16 CPUs
> > > + * with per-CPU DVFS and less than 8 capacity states each, for example.
> >
> > According to the formula above, that should give a "complexity value" of:
> >
> > 16 * (16 + 9) = 384
> >
> > while, 2K complexity seems more like a 40xCPUs system with 8 OPPs.
> >
> > Maybe we should update either the example or the constant below ?
>
> Hmm I guess the example isn't really clear. 'nr_cs' is the _sum_ of the
> number of OPPs of all perf. domains. So, in the example above, if you
> have 16 CPUs with per-CPU DVFS, and each DVFS island has 8 OPPs, then
> nr_cs = 16 * 8 = 128.
>
> So if you apply the formula you get C = 16 * (16 + 128) = 2304, which is
> more than EM_MAX_COMPLEXITY, so EAS cannot start.
>
> If the DVFS island had 7 OPPs instead of 8 (for example) you would get
> nr_cs = 112, C = 2048, and so EAS could start.
Right, I see now.
> I can try to re-work that comment to explain things a bit better ...
Yes, dunno if it's just me but perhaps a bit of rephrasing could help.
Alternatively, why not having this comment and check after patches
11 and 12 ?
> > > + */
> > > +#define EM_MAX_COMPLEXITY 2048
> > > +
> > > static void build_perf_domains(const struct cpumask *cpu_map)
> > > {
> > > + int i, nr_pd = 0, nr_cs = 0, nr_cpus = cpumask_weight(cpu_map);
> > > struct perf_domain *pd = NULL, *tmp;
> > > int cpu = cpumask_first(cpu_map);
> > > struct root_domain *rd = cpu_rq(cpu)->rd;
> > > - int i;
> > > +
> > > + /* EAS is enabled for asymmetric CPU capacity topologies. */
> > > + if (!per_cpu(sd_asym_cpucapacity, cpu)) {
> > > + if (sched_debug()) {
> > > + pr_info("rd %*pbl: CPUs do not have asymmetric capacities\n",
> > > + cpumask_pr_args(cpu_map));
> > > + }
> > > + goto free;
> > > + }
> > >
> > > for_each_cpu(i, cpu_map) {
> > > /* Skip already covered CPUs. */
> > > @@ -288,6 +318,21 @@ static void build_perf_domains(const struct cpumask *cpu_map)
> > > goto free;
> > > tmp->next = pd;
> > > pd = tmp;
> > > +
> > > + /*
> > > + * Count performance domains and capacity states for the
> > > + * complexity check.
> > > + */
> > > + nr_pd++;
> >
> > A special case where EAS is not going to be used is for systems where
> > nr_pd matches the number of online CPUs, isn't it ?
>
> Well, it depends. Say you have only 4 CPUs with 3 OPPs each. Even with
> per-CPU DVFS the complexity is low enough to start EAS. I don't really
> see a good reason for not doing so no ?
Right... I was totally confused by the idea that we don't
run EAS if we just have 1 CPU per PD... my bad!
Although on those systems, since we don't have idle costs, should not
be a spreading strategy always the best from an energy efficiency
standpoint ?
> > If that's the case, then, by caching this nr_pd you can probably check
> > this condition in the sched_energy_start() and bail out even faster by
> > avoiding to scan all the doms_new's pd ?
> >
> >
> > > + nr_cs += em_pd_nr_cap_states(pd->obj);
> > > + }
> > > +
> > > + /* Bail out if the Energy Model complexity is too high. */
> > > + if (nr_pd * (nr_cs + nr_cpus) > EM_MAX_COMPLEXITY) {
> > > + if (sched_debug())
> > > + pr_info("rd %*pbl: EM complexity is too high\n ",
> > > + cpumask_pr_args(cpu_map));
> > > + goto free;
> > > }
> > >
> > > perf_domain_debug(cpu_map, pd);
> > > @@ -307,6 +352,35 @@ static void build_perf_domains(const struct cpumask *cpu_map)
> > > if (tmp)
> > > call_rcu(&tmp->rcu, destroy_perf_domain_rcu);
> > > }
> > > +
> > > +static void sched_energy_start(int ndoms_new, cpumask_var_t doms_new[])
> > > +{
> > > + /*
> > > + * The conditions for EAS to start are checked during the creation of
> > > + * root domains. If one of them meets all conditions, it will have a
> > > + * non-null list of performance domains.
> > > + */
> > > + while (ndoms_new) {
> > > + if (cpu_rq(cpumask_first(doms_new[ndoms_new - 1]))->rd->pd)
> > > + goto enable;
> > > + ndoms_new--;
> > > + }
> > > +
> > > + if (static_branch_unlikely(&sched_energy_present)) {
> > ^^^^^^^^
> > Is this defined unlikely to reduce overheads on systems which never
> > satisfy all the conditions above while still rebuild SDs from time to
> > time ?
>
> Something like that. I just thought that the case where EAS needs to be
> disabled after being enabled isn't very common. I mean, the most typical
> use-case is, EAS is enabled at boot and stays enabled forever, or EAS
> never gets enabled.
Right, if we have EAS compiled in... we are likely to have it enabled.
> Enabling/disabling EAS because of hotplug (for example) can definitely
> happen, but that shouldn't be the case very often in practice, I think.
Would say yes on sane platform, i.e. where hotplug is not being used
for power/thermal management... but hopefully EAS should improve on
that side ;)
> So we can optimize things out a bit I suppose.
Right thanks!
--
#include <best/regards.h>
Patrick Bellasi