Re: [PATCH RFC tip/core/rcu 0/4] Accelerate RCU Tasks Trace updates

From: Paul E. McKenney
Date: Thu Sep 17 2020 - 17:06:57 EST


Hello!

This series accelerates RCU Tasks Trace updates, reducing the average
grace-period latencies from about a second to about 20 milliseconds on my
x86 laptop. These are benchmark numbers, based on a previously posted
patch to rcuscale.c [1] running on my x86 laptop. Additional patches
provide a compiler-warning cleanup and fix several race conditions that
were exposed by the faster grace periods. The patches in this series
are as follows:

1. Prevent complaints of unused show_rcu_tasks_classic_gp_kthread().
This is not related to the problem at hand, but it is
a pre-existing patch that provides a simple cleanup.
The grace-period latency thus remains at 980 milliseconds.

2. Mark variables static, noted during this effort but otherwise
unconnected. This change has no effect, so that the average
grace-period latency remains at 980 milliseconds.

3. Use more aggressive polling for RCU Tasks Trace. This polling
starts at five-millisecond intervals instead of the prior
100-millisecond intervals. As before, the polling interval
increases in duration as the grace period ages, and again as
before is capped at one second. This change reduces the
average grace-period latency to about 620 milliseconds.

4. Selectively enable more RCU Tasks Trace IPIs. This retains
the old behavior of suppressing IPIs for grace periods that are
younger than 500 milliseconds for CONFIG_TASKS_TRACE_RCU_READ_MB=y
kernels, including CONFIG_PREEMPT_RT=y kernels, but allows IPIs
immediately on other kernels. It is quite possible that a more
sophisticated decision procedure will be required, and changes
to RCU's dyntick-idle code might also be needed. This change
(along with the earlier ones) reduces the average grace-period
latency to about 120 milliseconds.

5. Shorten per-grace-period sleep for RCU Tasks Trace. The
current code sleeps for 100 milliseconds after the end of
each grace period, which by itself prevents a back-to-back
string of grace-period waits from completing faster than
ten per second. This patch also retains this old behavior
for CONFIG_TASKS_TRACE_RCU_READ_MB=y (and again thus also
for CONFIG_PREEMPT_RT=y kernels). For other kernels, this
post-grace-period sleep is reduced to five milliseconds.
This change (along with the earlier ones) reduced the average
grace-period latency to about 18 milliseconds, for an overall
factor-of-50 reduction in latency.

6. Fix a deadlock-inducing race between rcu_read_unlock_trace()
and trc_read_check_handler(). The race window was only a few
instructions wide, but please see the commit log for the full
sad story. The grace-period speedup made this race 50 times
more probable, and thus reduced the rcutorture runtime required
to reproduce it from about five months to about four days.

7. Fix a low-probability race between rcu_read_unlock_trace()
and the RCU Tasks Trace CPU stall reporting loop in
rcu_tasks_trace_postgp(). This race could result in leaking
task_struct structures.

8. Fix a low-probability race between the RCU Tasks Trace CPU
stall reporting loop in rcu_tasks_trace_postgp() and task exit.
This race could result in use-after-free errors.

Alexei Starovoitov benchmarked an earlier patch [2], producing results
that are roughly consistent with the above reduction in latency [3].

Changes since last week's RFC version:

o Added patch #1, which cleans up a compiler warning.

o Renumbered patches 1-4 to 2-5.

o Add Ccs to patches 3, 4, and 5.

o Add patches 6-8 to fix race conditions exposed by 50x faster
grace periods. These are either rare on the one hand or both
rare and occurring only during an RCU Tasks Trace CPU stall
warning -and- rare on the other. Still, they need to be fixed.

o This series maintains the sub-20-millisecond update-side
grace-period delays of the RFC series.

o Fixing the first of the race conditions required that a
compiler barrier be added to rcu_read_lock_trace() and that
another compiler barrier along with a WRITE_ONCE() be added to
rcu_read_unlock_trace(). This fix therefore adds a fraction of
a nanosecond to read-side overhead. On my laptop, the increase
is from about 2.6 nanoseconds to about 3 nanoseconds. This
small increase should not cause noticeable problems.

Thanx, Paul

[1] https://lore.kernel.org/bpf/20200909193900.GK29330@paulmck-ThinkPad-P72/
[2] https://lore.kernel.org/bpf/20200910052727.GA4351@paulmck-ThinkPad-P72/
[3] https://lore.kernel.org/bpf/619554b2-4746-635e-22f3-7f0f09d97760@xxxxxx/

------------------------------------------------------------------------

include/linux/rcupdate_trace.h | 4 +
kernel/rcu/tasks.h | 92 +++++++++++++++++++++++++++++++----------
2 files changed, 75 insertions(+), 21 deletions(-)