Re: 答复: [PATCH v3] mm/compaction:let proactive compaction order configurable
From: David Rientjes
Date: Sun Apr 25 2021 - 21:52:11 EST
On Mon, 26 Apr 2021, Chu,Kaiping wrote:
> Hi Rientjes
> I already answered your question in 4.19.
> " We turn off the transparent huge page in our machines, so we don't care about the order 9.
> There are many user space applications, different application maybe allocate different order of memory, we can't know the "known order of interest" in advance. Our purpose is to keep the overall fragment index as low as possible, not care about the specific order.
Ok, so you don't care about a specific order but you are adding a
vm.compaction_order sysctl?
I think what you're trying to do is invoke full compaction (cc.order = -1)
at some point in time that will (1) keep node-wide fragmentation low over
the long run and (2) be relatively lightweight at the time it is done.
I can certainly understand (1) on your configuration that is mostly
consumed by 1GB gigantic pages, you are likely dealing with significant
memory pressure that causes fragmentation to increase over time and
eventually become unrecoverable for the most part.
And for (2), yes, using vm.compact_memory will become very heavyweight if
it's done too late.
So since proactive compaction uses cc.order = 1, same as
vm.compact_memory, it should be possible to monitor extfrag_index under
debugfs and manually trigger compaction when necessary without
intervention of the kernel.
I think we can both agree that we wouldn't want to add obscure and
undocumented sysctls that that can easily be replaced by a userspace
implementation.
> Although current proactive compaction mechanism only check the fragment index of specific order, but it can do memory compaction for all order(.order = -1 in proactive_compact_node), so it's still useful for us.
> We set the compaction_order according to the average fragment index of all our machines, it's an experience value, it's a compromise of keep memory fragment index low and not trigger background compaction too much, this value can be changed in future.
> We did periodically memory compaction by command "echo 1 > /proc/sys/vm/compact_memory " previously, but it's not good enough, it's will compact all memory forcibly, it may lead to lots of memory move in short time, and affect the performance of application."
>
>
> BR,
> Chu Kaiping
>
> -----邮件原件-----
> 发件人: David Rientjes <rientjes@xxxxxxxxxx>
> 发送时间: 2021年4月26日 9:15
> 收件人: Chu,Kaiping <chukaiping@xxxxxxxxx>
> 抄送: mcgrof@xxxxxxxxxx; keescook@xxxxxxxxxxxx; yzaikin@xxxxxxxxxx; akpm@xxxxxxxxxxxxxxxxxxxx; vbabka@xxxxxxx; nigupta@xxxxxxxxxx; bhe@xxxxxxxxxx; khalid.aziz@xxxxxxxxxx; iamjoonsoo.kim@xxxxxxx; mateusznosek0@xxxxxxxxx; sh_def@xxxxxxx; linux-kernel@xxxxxxxxxxxxxxx; linux-fsdevel@xxxxxxxxxxxxxxx; linux-mm@xxxxxxxxx
> 主题: Re: [PATCH v3] mm/compaction:let proactive compaction order configurable
>
> On Sun, 25 Apr 2021, chukaiping wrote:
>
> > Currently the proactive compaction order is fixed to
> > COMPACTION_HPAGE_ORDER(9), it's OK in most machines with lots of
> > normal 4KB memory, but it's too high for the machines with small
> > normal memory, for example the machines with most memory configured as
> > 1GB hugetlbfs huge pages. In these machines the max order of free
> > pages is often below 9, and it's always below 9 even with hard
> > compaction. This will lead to proactive compaction be triggered very
> > frequently. In these machines we only care about order of 3 or 4.
> > This patch export the oder to proc and let it configurable by user,
> > and the default value is still COMPACTION_HPAGE_ORDER.
> >
>
> As asked in the review of the v1 of the patch, why is this not a userspace policy decision? If you are interested in order-3 or order-4 fragmentation, for whatever reason, you could periodically check /proc/buddyinfo and manually invoke compaction on the system.
>
> In other words, why does this need to live in the kernel?
>
> > Signed-off-by: chukaiping <chukaiping@xxxxxxxxx>
> > Reported-by: kernel test robot <lkp@xxxxxxxxx>
> > ---
> >
> > Changes in v3:
> > - change the min value of compaction_order to 1 because the fragmentation
> > index of order 0 is always 0
> > - move the definition of max_buddy_zone into #ifdef
> > CONFIG_COMPACTION
> >
> > Changes in v2:
> > - fix the compile error in ia64 and powerpc, move the initialization
> > of sysctl_compaction_order to kcompactd_init because
> > COMPACTION_HPAGE_ORDER is a variable in these architectures
> > - change the hard coded max order number from 10 to MAX_ORDER - 1
> >
> > include/linux/compaction.h | 1 +
> > kernel/sysctl.c | 10 ++++++++++
> > mm/compaction.c | 9 ++++++---
> > 3 files changed, 17 insertions(+), 3 deletions(-)
> >
> > diff --git a/include/linux/compaction.h b/include/linux/compaction.h
> > index ed4070e..151ccd1 100644
> > --- a/include/linux/compaction.h
> > +++ b/include/linux/compaction.h
> > @@ -83,6 +83,7 @@ static inline unsigned long compact_gap(unsigned int
> > order) #ifdef CONFIG_COMPACTION extern int sysctl_compact_memory;
> > extern unsigned int sysctl_compaction_proactiveness;
> > +extern unsigned int sysctl_compaction_order;
> > extern int sysctl_compaction_handler(struct ctl_table *table, int write,
> > void *buffer, size_t *length, loff_t *ppos); extern int
> > sysctl_extfrag_threshold; diff --git a/kernel/sysctl.c
> > b/kernel/sysctl.c index 62fbd09..e50f7d2 100644
> > --- a/kernel/sysctl.c
> > +++ b/kernel/sysctl.c
> > @@ -196,6 +196,7 @@ enum sysctl_writes_mode { #endif /*
> > CONFIG_SCHED_DEBUG */
> >
> > #ifdef CONFIG_COMPACTION
> > +static int max_buddy_zone = MAX_ORDER - 1;
> > static int min_extfrag_threshold;
> > static int max_extfrag_threshold = 1000; #endif @@ -2871,6 +2872,15
> > @@ int proc_do_static_key(struct ctl_table *table, int write,
> > .extra2 = &one_hundred,
> > },
> > {
> > + .procname = "compaction_order",
> > + .data = &sysctl_compaction_order,
> > + .maxlen = sizeof(sysctl_compaction_order),
> > + .mode = 0644,
> > + .proc_handler = proc_dointvec_minmax,
> > + .extra1 = SYSCTL_ONE,
> > + .extra2 = &max_buddy_zone,
> > + },
> > + {
> > .procname = "extfrag_threshold",
> > .data = &sysctl_extfrag_threshold,
> > .maxlen = sizeof(int),
> > diff --git a/mm/compaction.c b/mm/compaction.c index e04f447..70c0acd
> > 100644
> > --- a/mm/compaction.c
> > +++ b/mm/compaction.c
> > @@ -1925,16 +1925,16 @@ static bool kswapd_is_running(pg_data_t
> > *pgdat)
> >
> > /*
> > * A zone's fragmentation score is the external fragmentation wrt to
> > the
> > - * COMPACTION_HPAGE_ORDER. It returns a value in the range [0, 100].
> > + * sysctl_compaction_order. It returns a value in the range [0, 100].
> > */
> > static unsigned int fragmentation_score_zone(struct zone *zone) {
> > - return extfrag_for_order(zone, COMPACTION_HPAGE_ORDER);
> > + return extfrag_for_order(zone, sysctl_compaction_order);
> > }
> >
> > /*
> > * A weighted zone's fragmentation score is the external
> > fragmentation
> > - * wrt to the COMPACTION_HPAGE_ORDER scaled by the zone's size. It
> > + * wrt to the sysctl_compaction_order scaled by the zone's size. It
> > * returns a value in the range [0, 100].
> > *
> > * The scaling factor ensures that proactive compaction focuses on
> > larger @@ -2666,6 +2666,7 @@ static void compact_nodes(void)
> > * background. It takes values in the range [0, 100].
> > */
> > unsigned int __read_mostly sysctl_compaction_proactiveness = 20;
> > +unsigned int __read_mostly sysctl_compaction_order;
> >
> > /*
> > * This is the entry point for compacting all nodes via @@ -2958,6
> > +2959,8 @@ static int __init kcompactd_init(void)
> > int nid;
> > int ret;
> >
> > + sysctl_compaction_order = COMPACTION_HPAGE_ORDER;
> > +
> > ret = cpuhp_setup_state_nocalls(CPUHP_AP_ONLINE_DYN,
> > "mm/compaction:online",
> > kcompactd_cpu_online, NULL);
> > --
> > 1.7.1
> >
> >
>