Re: [PATCH slab/for-next-fixes] mm/slab: allow sheaf refill if blocking is not allowed

From: Harry Yoo

Date: Tue Mar 03 2026 - 22:05:53 EST


On Mon, Mar 02, 2026 at 10:55:37AM +0100, Vlastimil Babka (SUSE) wrote:
> Ming Lei reported [1] a regression in the ublk null target benchmark due
> to sheaves. The profile shows that the alloc_from_pcs() fastpath fails
> and allocations fall back to ___slab_alloc(). It also shows the
> allocations happen through mempool_alloc().
>
> The strategy of mempool_alloc() is to call the underlying allocator
> (here slab) without __GFP_DIRECT_RECLAIM first. This does not play well
> with __pcs_replace_empty_main() checking for gfpflags_allow_blocking()
> to decide if it should refill an empty sheaf or fallback to the
> slowpath, so we end up falling back.
>
> We could change the mempool strategy but there might be other paths
> doing the same ting. So instead allow sheaf refill when blocking is not
> allowed, changing the condition to gfpflags_allow_spinning(). The
> original condition was unnecessarily restrictive.
>
> Note this doesn't fully resolve the regression [1] as another component
> of that are memoryless nodes, which is to be addressed separately.
>
> Reported-by: Ming Lei <ming.lei@xxxxxxxxxx>
> Fixes: e47c897a2949 ("slab: add sheaves to most caches")
> Link: https://lore.kernel.org/all/aZ0SbIqaIkwoW2mB@fedora/
> Signed-off-by: Vlastimil Babka (SUSE) <vbabka@xxxxxxxxxx>
> ---
> mm/slub.c | 21 +++++++++------------
> 1 file changed, 9 insertions(+), 12 deletions(-)
>
> diff --git a/mm/slub.c b/mm/slub.c
> index b1e9f16ba435..17b200695e9b 100644
> --- a/mm/slub.c
> +++ b/mm/slub.c
> @@ -4632,11 +4631,8 @@ __pcs_replace_empty_main(struct kmem_cache *s, struct slub_percpu_sheaves *pcs,
> if (!full)
> return NULL;
>
> - /*
> - * we can reach here only when gfpflags_allow_blocking
> - * so this must not be an irq
> - */
> - local_lock(&s->cpu_sheaves->lock);
> + if (!local_trylock(&s->cpu_sheaves->lock))
> + goto barn_put;

My AI buddy says (don't worry, I filtered it):
| When local_trylock() fails above, the function jumps to barn_put and returns
| pcs without holding the lock. This appears to violate the function's contract
| documented in the comment at the beginning of __pcs_replace_empty_main():
|
| "If not successful, returns NULL and the local lock unlocked."
|
| The caller in alloc_from_pcs() checks for NULL to detect failure:
|
| if (unlikely(pcs->main->size == 0)) {
| pcs = __pcs_replace_empty_main(s, pcs, gfp);
| if (unlikely(!pcs))
| return NULL;
| }
|
| If the trylock fails and pcs (non-NULL) is returned, the caller proceeds
| without realizing the lock was never re-acquired. This leads to accessing
| pcs->main without the lock and later trying to unlock a lock that isn't held.

And the analysis sounds correct to me.

perhaps it should be:

if (!local_trylock(&s->cpu_sheaves->lock)) {
pcs = NULL;
goto barn_put;
}

> pcs = this_cpu_ptr(s->cpu_sheaves);
>
> /*
> @@ -4667,6 +4663,7 @@ __pcs_replace_empty_main(struct kmem_cache *s, struct slub_percpu_sheaves *pcs,
> return pcs;
> }
>
> +barn_put:
> barn_put_full_sheaf(barn, full);
> stat(s, BARN_PUT);

--
Cheers,
Harry / Hyeonggon