Re: [RFC PATCH 3/3] softirq: Avoid waking up ksoftirqd from flush_smp_call_function_queue()

From: Peter Zijlstra
Date: Wed Jul 10 2024 - 11:06:22 EST


On Wed, Jul 10, 2024 at 09:02:10AM +0000, K Prateek Nayak wrote:

> diff --git a/kernel/sched/smp.h b/kernel/sched/smp.h
> index 21ac44428bb0..3731e79fe19b 100644
> --- a/kernel/sched/smp.h
> +++ b/kernel/sched/smp.h
> @@ -9,7 +9,9 @@ extern void sched_ttwu_pending(void *arg);
> extern bool call_function_single_prep_ipi(int cpu);
>
> #ifdef CONFIG_SMP
> +extern bool do_softirq_pending(void);
> extern void flush_smp_call_function_queue(void);
> #else
> +static inline bool do_softirq_pending(void) { return false; }
> static inline void flush_smp_call_function_queue(void) { }
> #endif
> diff --git a/kernel/smp.c b/kernel/smp.c
> index f085ebcdf9e7..2eab5e1d5cef 100644
> --- a/kernel/smp.c
> +++ b/kernel/smp.c
> @@ -559,6 +559,36 @@ static void __flush_smp_call_function_queue(bool warn_cpu_offline)
> }
> }
>
> +/* Indicate an impending call to do_softirq_post_smp_call_flush() */
> +static DEFINE_PER_CPU_ALIGNED(bool, will_do_softirq_post_flush);
> +
> +static __always_inline void __set_will_do_softirq_post_flush(void)
> +{
> + this_cpu_write(will_do_softirq_post_flush, true);
> +}
> +
> +static __always_inline void __clr_will_do_softirq_post_flush(void)
> +{
> + this_cpu_write(will_do_softirq_post_flush, false);
> +}
> +
> +/**
> + * do_softirq_pending - Check if do_softirq_post_smp_call_flush() will
> + * be called after the invocation of
> + * __flush_smp_call_function_queue()
> + *
> + * When flush_smp_call_function_queue() executes in the context of idle,
> + * migration thread, a softirq raised from the smp-call-function ends up
> + * waking ksoftirqd despite an impending softirq processing via
> + * do_softirq_post_smp_call_flush().
> + *
> + * Indicate an impending do_softirq() to should_wake_ksoftirqd() despite
> + * not being in an interrupt context.
> + */
> +__always_inline bool do_softirq_pending(void)
> +{
> + return this_cpu_read(will_do_softirq_post_flush);
> +}
>
> /**
> * flush_smp_call_function_queue - Flush pending smp-call-function callbacks
> @@ -583,7 +613,9 @@ void flush_smp_call_function_queue(void)
> local_irq_save(flags);
> /* Get the already pending soft interrupts for RT enabled kernels */
> was_pending = local_softirq_pending();
> + __set_will_do_softirq_post_flush();
> __flush_smp_call_function_queue(true);
> + __clr_will_do_softirq_post_flush();
> if (local_softirq_pending())
> do_softirq_post_smp_call_flush(was_pending);
>
> diff --git a/kernel/softirq.c b/kernel/softirq.c
> index 02582017759a..b39eeed03042 100644
> --- a/kernel/softirq.c
> +++ b/kernel/softirq.c
> @@ -34,6 +34,8 @@
> #define CREATE_TRACE_POINTS
> #include <trace/events/irq.h>
>
> +#include "sched/smp.h"
> +
> /*
> - No shared variables, all the data are CPU local.
> - If a softirq needs serialization, let it serialize itself
> @@ -413,7 +415,13 @@ static inline void ksoftirqd_run_end(void)
>
> static inline bool should_wake_ksoftirqd(void)
> {
> - return true;
> + /*
> + * Avoid waking up ksoftirqd when a softirq is raised from a
> + * call-function executed by flush_smp_call_function_queue()
> + * in idle, migration thread's context since it'll soon call
> + * do_softirq_post_smp_call_flush().
> + */
> + return !do_softirq_pending();
> }

On first reading I wonder why you've not re-used and hooked into the
PREEMPT_RT variant of should_wake_ksoftirqd(). That already has a per
CPU variable to do exactly this.