Re: [PATCH v5 1/1] sched/fair: Fix low cpu usage with high throttling by removing expiration of cpu-local slices

From: bsegall
Date: Fri Jul 12 2019 - 14:01:40 EST


Dave Chiluk <chiluk+linux@xxxxxxxxxx> writes:

> So I spent some more time testing this new patch as is *(interrupts disabled). I know I probably should have fixed the patch, but it's hard to get time on big test hardware sometimes, and I was already well along my way with testing.
>
> In regards to the quota usage overage I was seeing earlier: I have a
> theory as to what might be happening here, and I'm pretty sure it's
> related to the IRQs being disabled during the rq->lock walk. I think
> that the main fast thread was able to use an excess amount of quota
> because the timer interrupt meant to stop it wasn't being handled
> timely due to the interrupts being disabled. On my 8 core machine this
> resulted in a what looked like simply improved usage of the quota, but
> when I ran the test on an 80 core machine I saw a massive overage of
> cpu usage when running fibtest. Specifically when running fibtest for
> 5 seconds with 50ms quota/100ms period expecting ~2500ms of quota
> usage; I got 3731 ms of cpu usage which was an unexpected overage of
> 1231ms. Is that a reasonable theory?

Tht doesn't seem likely - taking 1ms would be way longer than I'd expect
to begin with, and runtime_remaining can go negative for that sort of
reason anyways assuming the irq time is even counted towards the task.
Also I don't that the enable-irqs version will help for the scheduler
tick at least without rt patchsets.

That is still also too much for what I was thinking of though. I'll have
to look into this more.

>
> I'll try to get some time again tomorrow to test with IRQs disabled before the walk. Ben if you have a chance to fix and resend the patch that'd help.
>
> I'm really starting to think that simply removing the quota expiration
> may be the best solution here. Mathmatically it works out, it makes
> the code simpler, it doesn't have any of the lock walk issues, it
> doesn't add extra latency or overhead due to the slack timer,

It works out _for the job that is supposed to be throttled_. If the job
then gets a burst of actually-expensive work on many threads it can then
use NCPUs extra ms, adding latency to any other job on the system. Given
that it's still only 1ms on each runqueue, maybe this isn't the end of
the world, but the fail case does exist.

(We have to do exactly the same locking stuff on distribute, both more
rarely on the period timer, and on the currently existing slack timer)

> and that behavior is exactly what the kernel was doing for 5 years with few complaints about overage afaik.
>
> Either way, I'm very glad that we are getting to the end of this one, and all solutions appear to solve the core of the problem. I thank you all the work you guys have put into this.
>
> On Thu, Jul 11, 2019 at 12:46 PM <bsegall@xxxxxxxxxx> wrote:
>
> Peter Zijlstra <peterz@xxxxxxxxxxxxx> writes:
>
> > FWIW, good to see progress, still waiting for you guys to agree :-)
> >
> > On Mon, Jul 01, 2019 at 01:15:44PM -0700, bsegall@xxxxxxxxxx wrote:
> >
> >> - Taking up-to-every rq->lock is bad and expensive and 5ms may be too
> >> short a delay for this. I haven't tried microbenchmarks on the cost of
> >> this vs min_cfs_rq_runtime = 0 vs baseline.
> >
> > Yes, that's tricky, SGI/HPE have definite ideas about that.
> >
> >> @@ -4781,12 +4790,41 @@ static __always_inline void return_cfs_rq_runtime(struct cfs_rq *cfs_rq)
> >> */
> >> static void do_sched_cfs_slack_timer(struct cfs_bandwidth *cfs_b)
> >> {
> >> - u64 runtime = 0, slice = sched_cfs_bandwidth_slice();
> >> + u64 runtime = 0;
> >> unsigned long flags;
> >> u64 expires;
> >> + struct cfs_rq *cfs_rq, *temp;
> >> + LIST_HEAD(temp_head);
> >> +
> >> + local_irq_save(flags);
> >> +
> >> + raw_spin_lock(&cfs_b->lock);
> >> + cfs_b->slack_started = false;
> >> + list_splice_init(&cfs_b->slack_cfs_rq, &temp_head);
> >> + raw_spin_unlock(&cfs_b->lock);
> >> +
> >> +
> >> + /* Gather all left over runtime from all rqs */
> >> + list_for_each_entry_safe(cfs_rq, temp, &temp_head, slack_list) {
> >> + struct rq *rq = rq_of(cfs_rq);
> >> + struct rq_flags rf;
> >> +
> >> + rq_lock(rq, &rf);
> >> +
> >> + raw_spin_lock(&cfs_b->lock);
> >> + list_del_init(&cfs_rq->slack_list);
> >> + if (!cfs_rq->nr_running && cfs_rq->runtime_remaining > 0 &&
> >> + cfs_rq->runtime_expires == cfs_b->runtime_expires) {
> >> + cfs_b->runtime += cfs_rq->runtime_remaining;
> >> + cfs_rq->runtime_remaining = 0;
> >> + }
> >> + raw_spin_unlock(&cfs_b->lock);
> >> +
> >> + rq_unlock(rq, &rf);
> >> + }
> >
> > But worse still, you take possibly every rq->lock without ever
> > re-enabling IRQs.
> >
>
> Yeah, I'm not sure why I did that, it isn't correctness.