Re: [PATCH] mm, slab: eagerly delete inactive offlined SLABs
From: Michal Hocko
Date: Fri Mar 23 2018 - 09:45:24 EST
[Cc Vladimir]
On Wed 21-03-18 15:43:01, Shakeel Butt wrote:
> With kmem cgroup support, high memcgs churn can leave behind a lot of
> empty kmem_caches. Usually such kmem_caches will be destroyed when the
> corresponding memcg gets released but the memcg release can be
> arbitrarily delayed. These empty kmem_caches wastes cache_reaper's time.
> So, the reaper should destroy such empty offlined kmem_caches.
>
> Signed-off-by: Shakeel Butt <shakeelb@xxxxxxxxxx>
> ---
> mm/slab.c | 18 ++++++++++++++++--
> mm/slab.h | 15 +++++++++++++++
> mm/slab_common.c | 2 +-
> 3 files changed, 32 insertions(+), 3 deletions(-)
>
> diff --git a/mm/slab.c b/mm/slab.c
> index 66f2db98f026..9c174a799ffb 100644
> --- a/mm/slab.c
> +++ b/mm/slab.c
> @@ -4004,6 +4004,16 @@ static void drain_array(struct kmem_cache *cachep, struct kmem_cache_node *n,
> slabs_destroy(cachep, &list);
> }
>
> +static bool is_slab_active(struct kmem_cache *cachep)
> +{
> + int node;
> + struct kmem_cache_node *n;
> +
> + for_each_kmem_cache_node(cachep, node, n)
> + if (READ_ONCE(n->total_slabs) - n->free_slabs)
> + return true;
> + return false;
> +}
> /**
> * cache_reap - Reclaim memory from caches.
> * @w: work descriptor
> @@ -4018,7 +4028,7 @@ static void drain_array(struct kmem_cache *cachep, struct kmem_cache_node *n,
> */
> static void cache_reap(struct work_struct *w)
> {
> - struct kmem_cache *searchp;
> + struct kmem_cache *searchp, *tmp;
> struct kmem_cache_node *n;
> int node = numa_mem_id();
> struct delayed_work *work = to_delayed_work(w);
> @@ -4027,7 +4037,7 @@ static void cache_reap(struct work_struct *w)
> /* Give up. Setup the next iteration. */
> goto out;
>
> - list_for_each_entry(searchp, &slab_caches, list) {
> + list_for_each_entry_safe(searchp, tmp, &slab_caches, list) {
> check_irq_on();
>
> /*
> @@ -4061,6 +4071,10 @@ static void cache_reap(struct work_struct *w)
> 5 * searchp->num - 1) / (5 * searchp->num));
> STATS_ADD_REAPED(searchp, freed);
> }
> +
> + /* Eagerly delete inactive kmem_cache of an offlined memcg. */
> + if (!is_memcg_online(searchp) && !is_slab_active(searchp))
> + shutdown_cache(searchp);
> next:
> cond_resched();
> }
> diff --git a/mm/slab.h b/mm/slab.h
> index e8981e811c45..e911b10efae7 100644
> --- a/mm/slab.h
> +++ b/mm/slab.h
> @@ -166,6 +166,7 @@ static inline slab_flags_t kmem_cache_flags(unsigned int object_size,
> SLAB_TEMPORARY | \
> SLAB_ACCOUNT)
>
> +int shutdown_cache(struct kmem_cache *s);
> int __kmem_cache_shutdown(struct kmem_cache *);
> void __kmem_cache_release(struct kmem_cache *);
> int __kmem_cache_shrink(struct kmem_cache *);
> @@ -290,6 +291,15 @@ static __always_inline void memcg_uncharge_slab(struct page *page, int order,
> memcg_kmem_uncharge(page, order);
> }
>
> +static __always_inline bool is_memcg_online(struct kmem_cache *s)
> +{
> + if (!memcg_kmem_enabled())
> + return true;
> + if (is_root_cache(s))
> + return true;
> + return mem_cgroup_online(s->memcg_params.memcg);
> +}
> +
> extern void slab_init_memcg_params(struct kmem_cache *);
> extern void memcg_link_cache(struct kmem_cache *s);
> extern void slab_deactivate_memcg_cache_rcu_sched(struct kmem_cache *s,
> @@ -342,6 +352,11 @@ static inline void memcg_uncharge_slab(struct page *page, int order,
> {
> }
>
> +static inline bool is_memcg_online(struct kmem_cache *s)
> +{
> + return true;
> +}
> +
> static inline void slab_init_memcg_params(struct kmem_cache *s)
> {
> }
> diff --git a/mm/slab_common.c b/mm/slab_common.c
> index 61ab2ca8bea7..d197e878636b 100644
> --- a/mm/slab_common.c
> +++ b/mm/slab_common.c
> @@ -573,7 +573,7 @@ static void slab_caches_to_rcu_destroy_workfn(struct work_struct *work)
> }
> }
>
> -static int shutdown_cache(struct kmem_cache *s)
> +int shutdown_cache(struct kmem_cache *s)
> {
> /* free asan quarantined objects */
> kasan_cache_shutdown(s);
> --
> 2.17.0.rc0.231.g781580f067-goog
--
Michal Hocko
SUSE Labs