Re: [PATCH v2 3/3] timers: Exclude isolated cpus from timer migation
From: Gabriele Monaco
Date: Tue Apr 15 2025 - 11:49:50 EST
On Tue, 2025-04-15 at 11:30 -0400, Waiman Long wrote:
>
> On 4/15/25 6:25 AM, Gabriele Monaco wrote:
> > The timer migration mechanism allows active CPUs to pull timers
> > from
> > idle ones to improve the overall idle time. This is however
> > undesired
> > when CPU intensive workloads run on isolated cores, as the
> > algorithm
> > would move the timers from housekeeping to isolated cores,
> > negatively
> > affecting the isolation.
> >
> > This effect was noticed on a 128 cores machine running oslat on the
> > isolated cores (1-31,33-63,65-95,97-127). The tool monopolises
> > CPUs,
> > and the CPU with lowest count in a timer migration hierarchy (here
> > 1
> > and 65) appears as always active and continuously pulls global
> > timers,
> > from the housekeeping CPUs. This ends up moving driver work (e.g.
> > delayed work) to isolated CPUs and causes latency spikes:
> >
> > before the change:
> >
> > # oslat -c 1-31,33-63,65-95,97-127 -D 62s
> > ...
> > Maximum: 1203 10 3 4 ... 5 (us)
> >
> > after the change:
> >
> > # oslat -c 1-31,33-63,65-95,97-127 -D 62s
> > ...
> > Maximum: 10 4 3 4 3 ... 5 (us)
> >
> > Exclude isolated cores from the timer migration algorithm, extend
> > the
> > concept of unavailable cores, currently used for offline ones, to
> > isolated ones:
> > * A core is unavailable if isolated or offline;
> > * A core is available if isolated and offline;
> I think you mean "A core is available if NOT isolated and NOT
> offline".
> Right?
Yes, of course.. My bad. Thanks for spotting.
> >
> > A core is considered unavailable as idle if:
> > * is in the isolcpus list
> > * is in the nohz_full list
> > * is in an isolated cpuset
> >
> > Due to how the timer migration algorithm works, any CPU part of the
> > hierarchy can have their global timers pulled by remote CPUs and
> > have to
> > pull remote timers, only skipping pulling remote timers would break
> > the
> > logic.
> > For this reason, we prevents isolated CPUs from pulling remote
> > global
> > timers, but also the other way around: any global timer started on
> > an
> > isolated CPU will run there. This does not break the concept of
> > isolation (global timers don't come from outside the CPU) and, if
> > considered inappropriate, can usually be mitigated with other
> > isolation
> > techniques (e.g. IRQ pinning).
> >
> > Signed-off-by: Gabriele Monaco <gmonaco@xxxxxxxxxx>
> > ---
> > include/linux/timer.h | 6 ++++++
> > kernel/cgroup/cpuset.c | 14 ++++++++------
> > kernel/time/tick-internal.h | 1 +
> > kernel/time/timer.c | 10 ++++++++++
> > kernel/time/timer_migration.c | 24 +++++++++++++++++++++---
> > 5 files changed, 46 insertions(+), 9 deletions(-)
> >
> > diff --git a/include/linux/timer.h b/include/linux/timer.h
> > index 10596d7c3a346..4722e075d9843 100644
> > --- a/include/linux/timer.h
> > +++ b/include/linux/timer.h
> > @@ -190,4 +190,10 @@ int timers_dead_cpu(unsigned int cpu);
> > #define timers_dead_cpu NULL
> > #endif
> >
> > +#if defined(CONFIG_SMP) && defined(CONFIG_NO_HZ_COMMON)
> > +extern void tmigr_isolated_exclude_cpumask(cpumask_var_t
> > exclude_cpumask);
> > +#else
> > +static inline void tmigr_isolated_exclude_cpumask(cpumask_var_t
> > exclude_cpumask) { }
> > +#endif
> > +
> > #endif
> > diff --git a/kernel/cgroup/cpuset.c b/kernel/cgroup/cpuset.c
> > index 306b604300914..866b4b8188118 100644
> > --- a/kernel/cgroup/cpuset.c
> > +++ b/kernel/cgroup/cpuset.c
> > @@ -1323,7 +1323,7 @@ static bool partition_xcpus_del(int old_prs,
> > struct cpuset *parent,
> > return isolcpus_updated;
> > }
> >
> > -static void update_unbound_workqueue_cpumask(bool
> > isolcpus_updated)
> > +static void update_exclusion_cpumasks(bool isolcpus_updated)
> > {
> > int ret;
> >
> > @@ -1334,6 +1334,8 @@ static void
> > update_unbound_workqueue_cpumask(bool isolcpus_updated)
> >
> > ret = workqueue_unbound_exclude_cpumask(isolated_cpus);
> > WARN_ON_ONCE(ret < 0);
> > +
> > + tmigr_isolated_exclude_cpumask(isolated_cpus);
> > }
> >
> > /**
> > @@ -1454,7 +1456,7 @@ static int remote_partition_enable(struct
> > cpuset *cs, int new_prs,
> > list_add(&cs->remote_sibling, &remote_children);
> > cpumask_copy(cs->effective_xcpus, tmp->new_cpus);
> > spin_unlock_irq(&callback_lock);
> > - update_unbound_workqueue_cpumask(isolcpus_updated);
> > + update_exclusion_cpumasks(isolcpus_updated);
> > cpuset_force_rebuild();
> > cs->prs_err = 0;
> >
> > @@ -1495,7 +1497,7 @@ static void remote_partition_disable(struct
> > cpuset *cs, struct tmpmasks *tmp)
> > compute_effective_exclusive_cpumask(cs, NULL, NULL);
> > reset_partition_data(cs);
> > spin_unlock_irq(&callback_lock);
> > - update_unbound_workqueue_cpumask(isolcpus_updated);
> > + update_exclusion_cpumasks(isolcpus_updated);
> > cpuset_force_rebuild();
> >
> > /*
> > @@ -1563,7 +1565,7 @@ static void remote_cpus_update(struct cpuset
> > *cs, struct cpumask *xcpus,
> > if (xcpus)
> > cpumask_copy(cs->exclusive_cpus, xcpus);
> > spin_unlock_irq(&callback_lock);
> > - update_unbound_workqueue_cpumask(isolcpus_updated);
> > + update_exclusion_cpumasks(isolcpus_updated);
> > if (adding || deleting)
> > cpuset_force_rebuild();
> >
> > @@ -1906,7 +1908,7 @@ static int
> > update_parent_effective_cpumask(struct cpuset *cs, int cmd,
> > WARN_ON_ONCE(parent->nr_subparts < 0);
> > }
> > spin_unlock_irq(&callback_lock);
> > - update_unbound_workqueue_cpumask(isolcpus_updated);
> > + update_exclusion_cpumasks(isolcpus_updated);
> >
> > if ((old_prs != new_prs) && (cmd == partcmd_update))
> > update_partition_exclusive_flag(cs, new_prs);
> > @@ -2931,7 +2933,7 @@ static int update_prstate(struct cpuset *cs,
> > int new_prs)
> > else if (isolcpus_updated)
> > isolated_cpus_update(old_prs, new_prs, cs-
> > >effective_xcpus);
> > spin_unlock_irq(&callback_lock);
> > - update_unbound_workqueue_cpumask(isolcpus_updated);
> > + update_exclusion_cpumasks(isolcpus_updated);
> >
> > /* Force update if switching back to member & update
> > effective_xcpus */
> > update_cpumasks_hier(cs, &tmpmask, !new_prs);
> > diff --git a/kernel/time/tick-internal.h b/kernel/time/tick-
> > internal.h
> > index faac36de35b9e..75580f7c69c64 100644
> > --- a/kernel/time/tick-internal.h
> > +++ b/kernel/time/tick-internal.h
> > @@ -167,6 +167,7 @@ extern void
> > fetch_next_timer_interrupt_remote(unsigned long basej, u64 basem,
> > extern void timer_lock_remote_bases(unsigned int cpu);
> > extern void timer_unlock_remote_bases(unsigned int cpu);
> > extern bool timer_base_is_idle(void);
> > +extern bool timer_base_remote_is_idle(unsigned int cpu);
> > extern void timer_expire_remote(unsigned int cpu);
> > # endif
> > #else /* CONFIG_NO_HZ_COMMON */
> > diff --git a/kernel/time/timer.c b/kernel/time/timer.c
> > index 4d915c0a263c3..f04960091eba9 100644
> > --- a/kernel/time/timer.c
> > +++ b/kernel/time/timer.c
> > @@ -2162,6 +2162,16 @@ bool timer_base_is_idle(void)
> > return __this_cpu_read(timer_bases[BASE_LOCAL].is_idle);
> > }
> >
> > +/**
> > + * timer_base_remote_is_idle() - Return whether timer base is set
> > idle for cpu
> > + *
> > + * Returns value of local timer base is_idle value for remote cpu.
> > + */
> > +bool timer_base_remote_is_idle(unsigned int cpu)
> > +{
> > + return per_cpu(timer_bases[BASE_LOCAL].is_idle, cpu);
> > +}
> > +
> > static void __run_timer_base(struct timer_base *base);
> >
> > /**
> > diff --git a/kernel/time/timer_migration.c
> > b/kernel/time/timer_migration.c
> > index 1fae38fbac8c2..6fe6ca798e98d 100644
> > --- a/kernel/time/timer_migration.c
> > +++ b/kernel/time/timer_migration.c
> > @@ -10,6 +10,7 @@
> > #include <linux/spinlock.h>
> > #include <linux/timerqueue.h>
> > #include <trace/events/ipi.h>
> > +#include <linux/sched/isolation.h>
> >
> > #include "timer_migration.h"
> > #include "tick-internal.h"
> > @@ -1445,7 +1446,7 @@ static long tmigr_trigger_active(void
> > *unused)
> >
> > static int tmigr_cpu_unavailable(unsigned int cpu)
> > {
> > - struct tmigr_cpu *tmc = this_cpu_ptr(&tmigr_cpu);
> > + struct tmigr_cpu *tmc = per_cpu_ptr(&tmigr_cpu, cpu);
> > int migrator;
> > u64 firstexp;
> >
> > @@ -1472,15 +1473,18 @@ static int tmigr_cpu_unavailable(unsigned
> > int cpu)
> >
> > static int tmigr_cpu_available(unsigned int cpu)
> > {
> > - struct tmigr_cpu *tmc = this_cpu_ptr(&tmigr_cpu);
> > + struct tmigr_cpu *tmc = per_cpu_ptr(&tmigr_cpu, cpu);
> >
> > /* Check whether CPU data was successfully initialized */
> > if (WARN_ON_ONCE(!tmc->tmgroup))
> > return -EINVAL;
> >
> > + /* Isolated CPUs don't participate in timer migration */
> > + if (cpu_is_isolated(cpu))
> > + return 0;
>
> There are two main sets of isolated CPUs used by cpu_is_isolated() -
> boot-time isolated CPUs via "isolcpus" and "nohz_full" boot command
> time
> options and runtime isolated CPUs via cpuset isolated partitions. The
> check for runtime isolated CPUs is redundant here as those CPUs won't
> be
> passed to tmigr_cpu_available().
Since tmigr_cpu_available is shared between isolated and offline CPUs,
I added this check also to make sure bringing an isolated CPU back
online won't make it available for tmigr.
> So this call is effectively removing
> the boot time isolated CPUs away from the available cpumask
> especially
> during the boot up process. Maybe you can add some comment about this
> behavioral change.
>
Do you mean I should make clear that the check in tmigr_cpu_available
is especially meaningful at boot time (i.e. when CPUs are first brought
online)?
Yeah, I probably should, good point. I had that kind of comment in v1
while allocating the mask and removed it while changing a few things.
I'm going to make that comment more verbose to clarify when exactly
it's needed.
>
> > raw_spin_lock_irq(&tmc->lock);
> > trace_tmigr_cpu_available(tmc);
> > - tmc->idle = timer_base_is_idle();
> > + tmc->idle = timer_base_remote_is_idle(cpu);
> > if (!tmc->idle)
> > __tmigr_cpu_activate(tmc);
> > tmc->available = true;
> > @@ -1489,6 +1493,20 @@ static int tmigr_cpu_available(unsigned int
> > cpu)
> > return 0;
> > }
> >
> > +void tmigr_isolated_exclude_cpumask(cpumask_var_t exclude_cpumask)
> > +{
> > + int cpu;
> > +
> > + lockdep_assert_cpus_held();
> > +
> > + for_each_cpu_and(cpu, exclude_cpumask,
> > tmigr_available_cpumask)
> > + tmigr_cpu_unavailable(cpu);
> > +
> > + for_each_cpu_andnot(cpu, cpu_online_mask, exclude_cpumask)
> > + if (!cpumask_test_cpu(cpu,
> > tmigr_available_cpumask))
> > + tmigr_cpu_available(cpu);
> > +}
> > +
> > static void tmigr_init_group(struct tmigr_group *group, unsigned
> > int lvl,
> > int node)
> > {
>
> So far, I haven't seen any major issue with this patch series.
>
Thanks for the review!
Cheers,
Gabriele