Re: [PATCH v7] sched: Consolidate cpufreq updates

From: Qais Yousef
Date: Thu Aug 01 2024 - 08:23:07 EST


On 07/29/24 17:01, Metin Kaya wrote:
> On 28/07/2024 7:45 pm, Qais Yousef wrote:
> > Improve the interaction with cpufreq governors by making the
> > cpufreq_update_util() calls more intentional.
>
> [snip]
>
> > We also ensure to ignore cpufreq udpates for sugov workers at context
>
> Nit: s/udpates/updates/
>
> > switch if it was prev task.
>
> [snip]
>
> > +static __always_inline void
> > +__update_cpufreq_ctx_switch(struct rq *rq, struct task_struct *prev)
> > +{
> > +#ifdef CONFIG_CPU_FREQ
> > + if (prev && prev->dl.flags & SCHED_FLAG_SUGOV) {
> > + /* Sugov just did an update, don't be too aggressive */
> > + return;
> > + }
> > +
> > + /*
> > + * RT and DL should always send a freq update. But we can do some
> > + * simple checks to avoid it when we know it's not necessary.
> > + *
> > + * iowait_boost will always trigger a freq update too.
> > + *
> > + * Fair tasks will only trigger an update if the root cfs_rq has
> > + * decayed.
> > + *
> > + * Everything else should do nothing.
> > + */
> > + switch (current->policy) {
> > + case SCHED_NORMAL:
> > + case SCHED_BATCH:
> > + case SCHED_IDLE:
> > + if (unlikely(current->in_iowait)) {
> > + cpufreq_update_util(rq, SCHED_CPUFREQ_IOWAIT | SCHED_CPUFREQ_FORCE_UPDATE);
> > + return;
> > + }
> > +
> > +#ifdef CONFIG_SMP
> > + /*
> > + * Send an update if we switched from RT or DL as they tend to
> > + * boost the CPU and we are likely able to reduce the freq now.
> > + */
> > + rq->cfs.decayed |= prev && (rt_policy(prev->policy) || dl_policy(prev->policy));
> > +
> > + if (unlikely(rq->cfs.decayed)) {
> > + rq->cfs.decayed = false;
> > + cpufreq_update_util(rq, 0);
> > + return;
> > + }
> > +#else
> > + cpufreq_update_util(rq, 0);
> > +#endif
> > + return; /* ! */
> > + case SCHED_FIFO:
> > + case SCHED_RR:
> > + if (prev && rt_policy(prev->policy)) {
> > +#ifdef CONFIG_UCLAMP_TASK
> > + unsigned long curr_uclamp_min = uclamp_eff_value(current, UCLAMP_MIN);
> > + unsigned long prev_uclamp_min = uclamp_eff_value(prev, UCLAMP_MIN);
> > +
> > + if (curr_uclamp_min == prev_uclamp_min)
> > +#endif
> > + return;
> > + }
> > +#ifdef CONFIG_SMP
> > + /* Stopper task masquerades as RT */
> > + if (unlikely(current->sched_class == &stop_sched_class))
> > + return;
> > +#endif
> > + cpufreq_update_util(rq, SCHED_CPUFREQ_FORCE_UPDATE);
> > + return; /* ! */
> > + case SCHED_DEADLINE:
> > + /*
> > + * This is handled at enqueue to avoid breaking DL bandwidth
> > + * rules when multiple DL tasks are running on the same CPU.
> > + * Deferring till context switch here could mean the bandwidth
> > + * calculations would be broken to ensure all the DL tasks meet
> > + * their deadlines.
> > + */
> > + return; /* ! */
> > + default:
> > + return; /* ! */
> > + }
>
> Nit: would it be more conventional to replace marked `return` statements
> above with `break`s?

Thanks Metin. I think return and break are both fine here.

>
> > +#endif
> > +}
> > +
> > +/*
> > + * Call when currently running task had an attribute change that requires
> > + * an immediate cpufreq update.
> > + */
> > +void update_cpufreq_current(struct rq *rq)
> > +{
> > + __update_cpufreq_ctx_switch(rq, NULL);
> > +}
> > +
>
> [snip]
>