Re: [RFC PATCH v2 7/7] sched/fair: alternative way of accounting throttle time

From: Aaron Lu
Date: Wed Apr 09 2025 - 10:25:01 EST


On Wed, Apr 09, 2025 at 08:07:46PM +0800, Aaron Lu wrote:
> Implement an alternative way of accounting cfs_rq throttle time which:
> - starts accounting when a throttled cfs_rq has no tasks enqueued and its
> throttled list is not empty;
> - stops accounting when this cfs_rq gets unthrottled or a task gets
> enqueued.
>
> This way, the accounted throttle time is when the cfs_rq has absolutely
> no tasks enqueued and has tasks throttled.
>
> Signed-off-by: Aaron Lu <ziqianlu@xxxxxxxxxxxxx>
> ---
> kernel/sched/fair.c | 112 ++++++++++++++++++++++++++++++++-----------
> kernel/sched/sched.h | 4 ++
> 2 files changed, 89 insertions(+), 27 deletions(-)
>
> diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c
> index 20471a3aa35e6..70f7de82d1d9d 100644
> --- a/kernel/sched/fair.c
> +++ b/kernel/sched/fair.c
> @@ -5300,6 +5300,7 @@ place_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags)
>
> static void check_enqueue_throttle(struct cfs_rq *cfs_rq);
> static inline int cfs_rq_throttled(struct cfs_rq *cfs_rq);
> +static void account_cfs_rq_throttle_self(struct cfs_rq *cfs_rq);
>
> static void
> requeue_delayed_entity(struct sched_entity *se);
> @@ -5362,10 +5363,14 @@ enqueue_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags)
> if (throttled_hierarchy(cfs_rq)) {
> struct rq *rq = rq_of(cfs_rq);
>
> - if (cfs_rq_throttled(cfs_rq) && !cfs_rq->throttled_clock)
> - cfs_rq->throttled_clock = rq_clock(rq);
> - if (!cfs_rq->throttled_clock_self)
> - cfs_rq->throttled_clock_self = rq_clock(rq);
> + if (cfs_rq->throttled_clock) {
> + cfs_rq->throttled_time +=
> + rq_clock(rq) - cfs_rq->throttled_clock;
> + cfs_rq->throttled_clock = 0;
> + }

This probably needs more explanation.

We can also take cfs_b->lock and directly accounts the time into
cfs_b->throttled_time, but considering enqueue can be frequent so to
avoid possible lock contention, I chose to account this time to the cpu
local cfs_rq and on unthrottle, add the local accounted time to
cfs_b->throttled_time.

This has a side effect though: when reading cpu.stat and cpu.stat.local
for a task group with quota setting, the throttled_usec in cpu.stat can
be slightly smaller than throttled_usec in cpu.stat.local since some
throttled time is not accounted to cfs_b yet...

> +
> + if (cfs_rq->throttled_clock_self)
> + account_cfs_rq_throttle_self(cfs_rq);
> }
> #endif
> }