On Wed, Jun 07, 2023 at 12:06:03AM +0300, Kirill Tkhai wrote:
On 06.06.2023 01:32, Dave Chinner wrote:
On Mon, Jun 05, 2023 at 10:02:46PM +0300, Kirill Tkhai wrote:
This patch set introduces a new scheme of shrinker unregistration. It allows to split
the unregistration in two parts: fast and slow. This allows to hide slow part from
a user, so user-visible unregistration becomes fast.
This fixes the -88.8% regression of stress-ng.ramfs.ops_per_sec noticed
by kernel test robot:
https://lore.kernel.org/lkml/202305230837.db2c233f-yujie.liu@xxxxxxxxx/
---
Kirill Tkhai (2):
mm: Split unregister_shrinker() in fast and slow part
fs: Use delayed shrinker unregistration
Did you test any filesystem other than ramfs?
Filesystems more complex than ramfs have internal shrinkers, and so
they will still be running the slow synchronize_srcu() - potentially
multiple times! - in every unmount. Both XFS and ext4 have 3
internal shrinker instances per mount, so they will still call
synchronize_srcu() at least 3 times per unmount after this change.
What about any other subsystem that runs a shrinker - do they have
context depedent shrinker instances that get frequently created and
destroyed? They'll need the same treatment.
Of course, all of shrinkers should be fixed. This patch set just aims to describe
the idea more wider, because I'm not sure most people read replys to kernel robot reports.
This is my suggestion of way to go. Probably, Qi is right person to ask whether
we're going to extend this and to maintain f95bdb700bc6 in tree.
There is not much time. Unfortunately, kernel test robot reported this significantly late.
And that's why it should be reverted rather than trying to rush to
try to fix it.
I'm kind of tired of finding out about mm reclaim regressions only
when I see patches making naive and/or broken changes to subsystem
shrinker implementations without any real clue about what they are
doing. If people/subsystems who maintain shrinker implementations
were cc'd on the changes to the shrinker implementation, this would
have all been resolved before merging occurred....
Lockless shrinker lists need a heap of supporting changes to be done
first so that they aren't reliant on synchronise_srcu() *at all*. If
the code was properly designed in the first place (i.e. dynamic
shrinker structures freed via call_rcu()), we wouldn't be in rushing
to fix weird regressions right now.
Can we please revert this and start again with a properly throught
out and reveiwed design?
-Dave.