Re: [xfs] 6df693ed7b: aim7.jobs-per-min -15.7% regression
From: Dave Chinner
Date: Sat Aug 14 2021 - 19:28:53 EST
On Mon, Aug 09, 2021 at 02:42:48PM +0800, kernel test robot wrote:
>
>
> Greeting,
>
> FYI, we noticed a -15.7% regression of aim7.jobs-per-min due to commit:
>
>
> commit: 6df693ed7ba9ec03cafc38d5064de376a11243e2 ("xfs: per-cpu deferred inode inactivation queues")
> https://git.kernel.org/cgit/linux/kernel/git/djwong/xfs-linux.git xfs-5.15-merge
>
>
> in testcase: aim7
> on test machine: 88 threads 2 sockets Intel(R) Xeon(R) Gold 6238M CPU @ 2.10GHz with 128G memory
> with following parameters:
>
> disk: 4BRD_12G
> md: RAID1
> fs: xfs
> test: disk_wrt
> load: 3000
> cpufreq_governor: performance
> ucode: 0x5003006
>
> test-description: AIM7 is a traditional UNIX system level benchmark suite which is used to test and measure the performance of multiuser system.
> test-url: https://sourceforge.net/projects/aimbench/files/aim-suite7/
.....
> commit:
> 52cba078c8 ("xfs: detach dquots from inode if we don't need to inactivate it")
> 6df693ed7b ("xfs: per-cpu deferred inode inactivation queues")
>
> 52cba078c8b4b003 6df693ed7ba9ec03cafc38d5064
> ---------------- ---------------------------
> %stddev %change %stddev
> \ | \
> 539418 -15.7% 454630 aim7.jobs-per-min
> 33.57 +18.5% 39.79 aim7.time.elapsed_time
> 33.57 +18.5% 39.79 aim7.time.elapsed_time.max
> 2056 ± 7% +779.6% 18087 ± 2% aim7.time.involuntary_context_switches
> 673.92 ± 4% +29.2% 870.54 ± 2% aim7.time.system_time
> 912022 -34.2% 599694 aim7.time.voluntary_context_switches
OK, performance went down, system time went up massively. I'm
betting the improvement made something else fast enough to trigger
spinning lock breakdown somewhere in the kernel...
> 0.01 ±223% +0.9 0.93 ± 4% perf-profile.children.cycles-pp.ret_from_fork
> 0.01 ±223% +0.9 0.93 ± 4% perf-profile.children.cycles-pp.kthread
> 67.40 ± 5% +8.7 76.14 ± 2% perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
> 67.02 ± 5% +8.8 75.81 ± 2% perf-profile.children.cycles-pp.do_syscall_64
> 2.02 ± 3% +15.4 17.42 ± 2% perf-profile.children.cycles-pp.creat64
> 1.99 ± 4% +15.4 17.40 ± 2% perf-profile.children.cycles-pp.do_sys_open
> 1.99 ± 3% +15.4 17.40 ± 2% perf-profile.children.cycles-pp.do_sys_openat2
> 1.96 ± 4% +15.4 17.38 ± 2% perf-profile.children.cycles-pp.path_openat
> 1.96 ± 3% +15.4 17.38 ± 2% perf-profile.children.cycles-pp.do_filp_open
> 2.23 ± 6% +15.6 17.82 ± 2% perf-profile.children.cycles-pp.unlink
> 2.18 ± 6% +15.6 17.78 ± 2% perf-profile.children.cycles-pp.do_unlinkat
> 0.66 ± 19% +30.7 31.40 ± 2% perf-profile.children.cycles-pp.osq_lock
> 2.34 ± 8% +31.0 33.33 ± 2% perf-profile.children.cycles-pp.rwsem_down_write_slowpath
Well, looky there. Lots of new rwsem lock contention.
....
> 1.54 ± 3% +0.2 1.74 perf-profile.self.cycles-pp.rwsem_spin_on_owner
> 0.66 ± 19% +30.6 31.21 ± 2% perf-profile.self.cycles-pp.osq_lock
Yup, we now have catastrophic spin-on-onwer breakdown of a rwsem.
IOWs, what this commit has done is put pressure on a rwsem in a
different way, and on this specific machine configuration on this
specific workload, it results in the rwsem breaking down into
catastrophic spin-on-owner contention. This looks like a rwsem bug,
not a bug in the XFS code.
Given that this is showing up in the open and unlink paths, this
is likely the parent directory inode lock being contended due to
concurrent modifications in the same directory.
That correlates with the change that the deferred inactivation
brings to unlink worklaods - the unlink() syscall does about a third
of the work it used to, so results in locking the directory inode
*much* more frequently with only very short pauses in userspace to
make the next unlink() call.
Because of the way the worker threads are CPU bound and all the XFS
objects involved in repeated directory ops will stay CPU affine, the
unlink() syscall is likely to run hot and not block until the queue
limits are hit and it is forced to throttle and let the worker run
to drain the queue.
Now, rwsems are *supposed to be sleeping locks*. In which case, we
should switch away on contention and let the CPU be used for some
other useful work until we are granted the lock. But, no, spinning
on exclusive locks makes some other benchmark go faster, so now on
this benchmark burn *30% of 88 CPUs* spinning on rwsems across this
benchmark.
So this regression is caused by an rwsem bug. XFS is, as usual, just
the messenger for iproblems arising from the misbehaviour of rwsems.
Try turning off rwsem spin-on-owner behaviour and see what
difference that makes to performance...
Cheers,
Dave.
--
Dave Chinner
dchinner@xxxxxxxxxx