Re: [patch V9 02/39] rcu: Abstract out rcu_irq_enter_check_tick() from rcu_nmi_enter()
From: Paul E. McKenney
Date: Thu May 21 2020 - 17:03:42 EST
On Thu, May 21, 2020 at 10:05:15PM +0200, Thomas Gleixner wrote:
> From: Paul E. McKenney <paulmck@xxxxxxxxxx>
>
> There will likely be exception handlers that can sleep, which rules
> out the usual approach of invoking rcu_nmi_enter() on entry and also
> rcu_nmi_exit() on all exit paths. However, the alternative approach of
> just not calling anything can prevent RCU from coaxing quiescent states
> from nohz_full CPUs that are looping in the kernel: RCU must instead
> IPI them explicitly. It would be better to enable the scheduler tick
> on such CPUs to interact with RCU in a lighter-weight manner, and this
> enabling is one of the things that rcu_nmi_enter() currently does.
>
> What is needed is something that helps RCU coax quiescent states while
> not preventing subsequent sleeps. This commit therefore splits out the
> nohz_full scheduler-tick enabling from the rest of the rcu_nmi_enter()
> logic into a new function named rcu_irq_enter_check_tick().
>
> [ tglx: Renamed the function and made it a nop when context tracking is off ]
The new name works for me! A couple of nits called out below.
Thanx, Paul
> Suggested-by: Andy Lutomirski <luto@xxxxxxxxxx>
> Signed-off-by: Paul E. McKenney <paulmck@xxxxxxxxxx>
> Signed-off-by: Thomas Gleixner <tglx@xxxxxxxxxxxxx>
> ---
> V9: New patch
> ---
> include/linux/hardirq.h | 9 +++++
> kernel/rcu/tree.c | 82 ++++++++++++++++++++++++++++++++++++------------
> 2 files changed, 71 insertions(+), 20 deletions(-)
>
> --- a/include/linux/hardirq.h
> +++ b/include/linux/hardirq.h
> @@ -2,6 +2,7 @@
> #ifndef LINUX_HARDIRQ_H
> #define LINUX_HARDIRQ_H
>
> +#include <linux/context_tracking_state.h>
> #include <linux/preempt.h>
> #include <linux/lockdep.h>
> #include <linux/ftrace_irq.h>
> @@ -27,6 +28,14 @@ extern void rcu_nmi_enter(void);
> extern void rcu_nmi_exit(void);
> #endif
>
> +void __rcu_irq_enter_check_tick(void);
> +
> +static __always_inline void rcu_irq_enter_check_tick(void)
> +{
> + if (context_tracking_enabled())
> + __rcu_irq_enter_check_tick();
I suggest moving the WARN_ON_ONCE(in_nmi()) check here to avoid calling
in_nmi() twice. Because of the READ_ONCE(), the compiler cannot (had
better not!) eliminate the double call.
> +}
> +
> /*
> * It is safe to do non-atomic ops on ->hardirq_context,
> * because NMI handlers may not preempt and the ops are
> --- a/kernel/rcu/tree.c
> +++ b/kernel/rcu/tree.c
> @@ -848,6 +848,67 @@ void noinstr rcu_user_exit(void)
> {
> rcu_eqs_exit(1);
> }
> +
> +/**
> + * __rcu_irq_enter_check_tick - Enable scheduler tick on CPU if RCU needs it.
> + *
> + * The scheduler tick is not normally enabled when CPUs enter the kernel
> + * from nohz_full userspace execution. After all, nohz_full userspace
> + * execution is an RCU quiescent state and the time executing in the kernel
> + * is quite short. Except of course when it isn't. And it is not hard to
> + * cause a large system to spend tens of seconds or even minutes looping
> + * in the kernel, which can cause a number of problems, include RCU CPU
> + * stall warnings.
> + *
> + * Therefore, if a nohz_full CPU fails to report a quiescent state
> + * in a timely manner, the RCU grace-period kthread sets that CPU's
> + * ->rcu_urgent_qs flag with the expectation that the next interrupt or
> + * exception will invoke this function, which will turn on the scheduler
> + * tick, which will enable RCU to detect that CPU's quiescent states,
> + * for example, due to cond_resched() calls in CONFIG_PREEMPT=n kernels.
> + * The tick will be disabled once a quiescent state is reported for
> + * this CPU.
> + *
> + * Of course, in carefully tuned systems, there might never be an
> + * interrupt or exception. In that case, the RCU grace-period kthread
> + * will eventually cause one to happen. However, in less carefully
> + * controlled environments, this function allows RCU to get what it
> + * needs without creating otherwise useless interruptions.
> + */
> +void __rcu_irq_enter_check_tick(void)
> +{
> + struct rcu_data *rdp = this_cpu_ptr(&rcu_data);
> +
> + // Enabling the tick is unsafe in NMI handlers.
There is an extra space before the "//", probably the one that used to
be after the ";" below. ;-)
> + if (WARN_ON_ONCE(in_nmi()))
> + return;
> +
> + RCU_LOCKDEP_WARN(rcu_dynticks_curr_cpu_in_eqs(),
> + "Illegal rcu_irq_enter_check_tick() from extended quiescent state");
The instrumentation_begin() has disappeared, presumably because
instrumentation is already enabled in the non-RCU code that directly calls
rcu_irq_enter_check_tick(). (I do see the calls in rcu_nmi_enter() below.)
> +
> + if (!tick_nohz_full_cpu(rdp->cpu) ||
> + !READ_ONCE(rdp->rcu_urgent_qs) ||
> + READ_ONCE(rdp->rcu_forced_tick)) {
> + // RCU doesn't need nohz_full help from this CPU, or it is
> + // already getting that help.
> + return;
> + }
> +
> + // We get here only when not in an extended quiescent state and
> + // from interrupts (as opposed to NMIs). Therefore, (1) RCU is
> + // already watching and (2) The fact that we are in an interrupt
> + // handler and that the rcu_node lock is an irq-disabled lock
> + // prevents self-deadlock. So we can safely recheck under the lock.
> + // Note that the nohz_full state currently cannot change.
> + raw_spin_lock_rcu_node(rdp->mynode);
> + if (rdp->rcu_urgent_qs && !rdp->rcu_forced_tick) {
> + // A nohz_full CPU is in the kernel and RCU needs a
> + // quiescent state. Turn on the tick!
> + WRITE_ONCE(rdp->rcu_forced_tick, true);
> + tick_dep_set_cpu(rdp->cpu, TICK_DEP_BIT_RCU);
> + }
> + raw_spin_unlock_rcu_node(rdp->mynode);
> +}
> #endif /* CONFIG_NO_HZ_FULL */
>
> /**
> @@ -894,26 +955,7 @@ noinstr void rcu_nmi_enter(void)
> incby = 1;
> } else if (!in_nmi()) {
This can just be "else" given the in_nmi() check in
__rcu_irq_enter_check_tick(), right? Ah, that check got a
WARN_ON_ONCE(), so never mind!
I guess that will discourage NMI-handler calls to
rcu_irq_enter_check_tick(). ;-)
It does mean a double call to in_nmi(), though, so should that
WARN_ON_ONCE(in_nmi()) check go into the rcu_irq_enter_check_tick()
wrapper? Or do modern compilers figure this one out? Given the
READ_ONCE() in preempt_count(), I have to say that I hope not.
So see my comment above on rcu_irq_enter_check_tick().
> instrumentation_begin();
> - if (tick_nohz_full_cpu(rdp->cpu) &&
> - rdp->dynticks_nmi_nesting == DYNTICK_IRQ_NONIDLE &&
> - READ_ONCE(rdp->rcu_urgent_qs) &&
> - !READ_ONCE(rdp->rcu_forced_tick)) {
> - // We get here only if we had already exited the
> - // extended quiescent state and this was an
> - // interrupt (not an NMI). Therefore, (1) RCU is
> - // already watching and (2) The fact that we are in
> - // an interrupt handler and that the rcu_node lock
> - // is an irq-disabled lock prevents self-deadlock.
> - // So we can safely recheck under the lock.
> - raw_spin_lock_rcu_node(rdp->mynode);
> - if (rdp->rcu_urgent_qs && !rdp->rcu_forced_tick) {
> - // A nohz_full CPU is in the kernel and RCU
> - // needs a quiescent state. Turn on the tick!
> - WRITE_ONCE(rdp->rcu_forced_tick, true);
> - tick_dep_set_cpu(rdp->cpu, TICK_DEP_BIT_RCU);
> - }
> - raw_spin_unlock_rcu_node(rdp->mynode);
> - }
> + rcu_irq_enter_check_tick();
> instrumentation_end();
> }
> instrumentation_begin();
>