Re: [PATCH v5 8/9] slub: Rename all *unfreeze_partials* functions to *put_partials*

From: Hyeonggon Yoo
Date: Sun Dec 03 2023 - 04:28:00 EST


On Thu, Nov 2, 2023 at 12:25 PM <chengming.zhou@xxxxxxxxx> wrote:
>
> From: Chengming Zhou <zhouchengming@xxxxxxxxxxxxx>
>
> Since all partial slabs on the CPU partial list are not frozen anymore,
> we don't unfreeze when moving cpu partial slabs to node partial list,
> it's better to rename these functions.
>
> Signed-off-by: Chengming Zhou <zhouchengming@xxxxxxxxxxxxx>
> Reviewed-by: Vlastimil Babka <vbabka@xxxxxxx>
> Tested-by: Hyeonggon Yoo <42.hyeyoo@xxxxxxxxx>
> ---
> mm/slub.c | 34 +++++++++++++++++-----------------
> 1 file changed, 17 insertions(+), 17 deletions(-)
>
> diff --git a/mm/slub.c b/mm/slub.c
> index d137468fe4b9..c20bdf5dab0f 100644
> --- a/mm/slub.c
> +++ b/mm/slub.c
> @@ -2546,7 +2546,7 @@ static void deactivate_slab(struct kmem_cache *s, struct slab *slab,
> }
>
> #ifdef CONFIG_SLUB_CPU_PARTIAL
> -static void __unfreeze_partials(struct kmem_cache *s, struct slab *partial_slab)
> +static void __put_partials(struct kmem_cache *s, struct slab *partial_slab)
> {
> struct kmem_cache_node *n = NULL, *n2 = NULL;
> struct slab *slab, *slab_to_discard = NULL;
> @@ -2588,9 +2588,9 @@ static void __unfreeze_partials(struct kmem_cache *s, struct slab *partial_slab)
> }
>
> /*
> - * Unfreeze all the cpu partial slabs.
> + * Put all the cpu partial slabs to the node partial list.
> */
> -static void unfreeze_partials(struct kmem_cache *s)
> +static void put_partials(struct kmem_cache *s)
> {
> struct slab *partial_slab;
> unsigned long flags;
> @@ -2601,11 +2601,11 @@ static void unfreeze_partials(struct kmem_cache *s)
> local_unlock_irqrestore(&s->cpu_slab->lock, flags);
>
> if (partial_slab)
> - __unfreeze_partials(s, partial_slab);
> + __put_partials(s, partial_slab);
> }
>
> -static void unfreeze_partials_cpu(struct kmem_cache *s,
> - struct kmem_cache_cpu *c)
> +static void put_partials_cpu(struct kmem_cache *s,
> + struct kmem_cache_cpu *c)
> {
> struct slab *partial_slab;
>
> @@ -2613,7 +2613,7 @@ static void unfreeze_partials_cpu(struct kmem_cache *s,
> c->partial = NULL;
>
> if (partial_slab)
> - __unfreeze_partials(s, partial_slab);
> + __put_partials(s, partial_slab);
> }
>
> /*
> @@ -2626,7 +2626,7 @@ static void unfreeze_partials_cpu(struct kmem_cache *s,
> static void put_cpu_partial(struct kmem_cache *s, struct slab *slab, int drain)
> {
> struct slab *oldslab;
> - struct slab *slab_to_unfreeze = NULL;
> + struct slab *slab_to_put = NULL;
> unsigned long flags;
> int slabs = 0;
>
> @@ -2641,7 +2641,7 @@ static void put_cpu_partial(struct kmem_cache *s, struct slab *slab, int drain)
> * per node partial list. Postpone the actual unfreezing
> * outside of the critical section.
> */
> - slab_to_unfreeze = oldslab;
> + slab_to_put = oldslab;
> oldslab = NULL;
> } else {
> slabs = oldslab->slabs;
> @@ -2657,17 +2657,17 @@ static void put_cpu_partial(struct kmem_cache *s, struct slab *slab, int drain)
>
> local_unlock_irqrestore(&s->cpu_slab->lock, flags);
>
> - if (slab_to_unfreeze) {
> - __unfreeze_partials(s, slab_to_unfreeze);
> + if (slab_to_put) {
> + __put_partials(s, slab_to_put);
> stat(s, CPU_PARTIAL_DRAIN);
> }
> }
>
> #else /* CONFIG_SLUB_CPU_PARTIAL */
>
> -static inline void unfreeze_partials(struct kmem_cache *s) { }
> -static inline void unfreeze_partials_cpu(struct kmem_cache *s,
> - struct kmem_cache_cpu *c) { }
> +static inline void put_partials(struct kmem_cache *s) { }
> +static inline void put_partials_cpu(struct kmem_cache *s,
> + struct kmem_cache_cpu *c) { }
>
> #endif /* CONFIG_SLUB_CPU_PARTIAL */
>
> @@ -2709,7 +2709,7 @@ static inline void __flush_cpu_slab(struct kmem_cache *s, int cpu)
> stat(s, CPUSLAB_FLUSH);
> }
>
> - unfreeze_partials_cpu(s, c);
> + put_partials_cpu(s, c);
> }
>
> struct slub_flush_work {
> @@ -2737,7 +2737,7 @@ static void flush_cpu_slab(struct work_struct *w)
> if (c->slab)
> flush_slab(s, c);
>
> - unfreeze_partials(s);
> + put_partials(s);
> }
>
> static bool has_cpu_slab(int cpu, struct kmem_cache *s)
> @@ -3168,7 +3168,7 @@ static void *___slab_alloc(struct kmem_cache *s, gfp_t gfpflags, int node,
> if (unlikely(!node_match(slab, node) ||
> !pfmemalloc_match(slab, gfpflags))) {
> slab->next = NULL;
> - __unfreeze_partials(s, slab);
> + __put_partials(s, slab);
> continue;
> }
>
> --

Looks good to me,
Reviewed-by: Hyeonggon Yoo <42.hyeyoo@xxxxxxxxx>

Thanks!

> 2.20.1
>