Re: [PATCH v2 1/3] sched/uclamp: Set max_spare_cap_cpu even if max_spare_cap is 0
From: Qais Yousef
Date: Thu Feb 23 2023 - 10:12:26 EST
On 02/21/23 13:20, Dietmar Eggemann wrote:
> I analyzed one of the last traces I got with my example:
>
> During the initial wakeup the system is in CPU OU. So feec() returns
> `target = -1` and `sis()->sic()` (a) does the initial placement for all
> the 6 tasks.
>
> CPU (a) (b) (c) (d) (e)
> ^ ^ ^ ^ ^
> 0 t1-------->| |
> 1 t0 t3 |t5 |t1 |t4| | | | | ...
> 2 t2--------------------->|
> 3
> 4 t4------------>| |
> 5 t5---->| |
>
> (b) feec() for t5:
>
> eenv_pd_busy_time()
>
> pd_busy_time = 1998 = 994 (CPU1) + 1004 (CPU2)
>
> compute_energy()
>
> busy_time = min(pd_capacity, pd_busy_time + task_busy_time)
>
> = min(2048, 1998 + 921)
>
> = 2048
>
> We go into saturation here and EAS assumes it can place t5 for the
> cost of 2048 - 1998 = 50 util where in reality t5 has a util of
> ~921. And that's why t5 ends up on CPU1 too.
So to rephrase as it was hard to follow your line thoughts.
The problem you're highlighting is that with uclamp_max we can end up with
a fully utilized pd. But the energy cost of this always-runing pd is capped to
a constant value. But you think this should be modeled better to reflect it'll
run for longer period of time, hence cost more energy.
There are multiple compound issue that makes this difficult to be reflected
accurately and makes think this best-effort is not really as bad as you think:
1. The capacity of little cores has become small, 158 on pixel 6. This makes
the definition of 'busy' task from the little's point of view rather
relaxed/small. But there's a strong desire to enable uclamp_max to act as
a soft affinity to prevent these tasks from interfering with more important
work on bigger cores and potentially waste power.
2. If the task is truly long running 1024 tasks (even on big core), then
knowing its actual utilization and runtime would be impossible and we'd have
to cap it to something. It's basically inifinity.
3. Because of uclamp_max capping, the util_avg could be wrong - depends on
actually how big the task is compared the uclamp_max. In worst case scenario
if no idle time is seen util_avg can easily grow into 1024 - although if the
task is allowed to run uncapped it actually might be 300 or something in
reality.
4. Because of max aggregation, the uclamp_max tasks are a frequency spike
hazard. Again if the task is a true 1024 (and capping those tasks is one of
the major use cases for uclamp_max, if not The Major use case) then as soon
as uncapped task wakes up on the CPU, there'd be a frequency spike
regardless of how small this task is.
If at wake-up it thought a big core is okay, and ends up running for the new
few 100s ms with randodm rt/kworkers waking up on that core - the big core
will run enough times at most energy enefficient point of the system. Which
is a bigger problem as the power cost will be very high and noticeable. And
uclamp_max isn't being used because of this already today.
And because this is an always running task for the foreseeable future from
the scheduler PoV, and, the lack of a downmigration mechanism to move these
tasks away (I have patches for that - but trying to send fixes one by one),
these frequency spikes poses a bigger hazard of wasting power than making
a one off miscalculation at wakeup time. The decision at wake up is sampling
at that exact instant of time. But generally uclamp_max is desired for those
long running tasks whose potential energy cost over a long period of time
(many ticks worth of realtime) is the worrisome. Immediately after that
sampling time the condition could change even today and render our decision
completely wrong. And there's no fixing for that. It's the nature of the
beast.
5 One of the desired properties of uclamp_max in practice is to act as a soft
affinity. In Android background tasks are restricted by cpuset. But this
could potentially lead to lost opportunities of better energy placement under
normal lightly loaded circumstances where there are a lot of small tasks
running but nothing particularly busy. EAS shines under those scenarios. But
fails for long running tasks - and here where uclamp_max is expected to help.
>
> (c) & (d) similar to (b)
>
> (e) From here on no wakeups anymore. Only tick preemption on CPU1 every
> 4ms (250Hz) between the 5 tasks and t2 finishing on CPU2 eventually.
>
> >>> Which is the desired behavior for uclamp_max?
> >>
> >> Not sure. Essentially, EAS can't do its job properly if there is no idle
> >
> > This "not sure" statement is making me worried. Are you not sure about how
> > uclamp_max should work in force fitting big tasks into small cores? Or just
> > about handling some of the corner cases? I understood the former, not the
> > latter.
>
> I'm not sure if we can call the issue that EAS doesn't work in
> saturation anymore a corner case. In case uclamp_max has an effect, it
> can quite easily create these saturation scenarios in which EAS is still
> enabled but it can't do its job anymore.
> Making sure that we probe each PD is not getting rid of this more
> fundamental issue.
I disagree with there being a fundamental issue in regards to energy
calculation. I see a potential nice-to-have improvement for estimating energy
calculation of long running tasks.
What I see as a fundamental error that this series is fixing is that the hint
to keep tasks on little cores doesn't work. Being able to consider a better
energy efficient CPU is less of a problem and not sure if it's really affecting
anyone in practice. Today's mainline will not consider any busy task as
a candidate for little core even of uclamp_max says it's okay. The task has to
be small enough (< 158 on pixel 6) for EAS to consider it, but those tasks are
not the ones that need uclamp_max hint.
We could try to remove the min(). But I worry this is premature now before
first introducing my fixes to teach the load balancer to do down migration. And
I think we need to do more to address the frequency spike problem - which I do
have proposals ready to address this problem too.
As it stands we can't use uclamp_max in products, and there are a series of
fixes required to achieve that. And what you see as a fundamental issue is not
one of the blocking issues I am aware of. It seems a nice enhancement to make
the wakeup energy estimation even better. But I do think it requires some
crystal ball and it'll remain a best effort even after that.
I think the current outcome is the desired behavior to be honest. Maybe we can
do better, but that's a separate problem/question and not a fundamental
blocking to enabling uclamp_max to do what it says on the tin.
I can add this as another limitation in the uclamp doc.
The only thing that probably should do now is add a limit to how much cramming
we can do before we say it's too much for this pd even with uclamp_max. Maybe
we can put a limit based on load_avg as I think it'll continue to grow unlike
util_avg which will settle on 1024.
Thanks!
--
Qais Yousef