Re: [PATCH v7 6/7] sched: replace capacity_factor by usage
From: Peter Zijlstra
Date: Thu Oct 09 2014 - 10:58:43 EST
On Tue, Oct 07, 2014 at 02:13:36PM +0200, Vincent Guittot wrote:
> @@ -6214,17 +6178,21 @@ static inline void update_sd_lb_stats(struct lb_env *env, struct sd_lb_stats *sd
>
> /*
> * In case the child domain prefers tasks go to siblings
> - * first, lower the sg capacity factor to one so that we'll try
> + * first, lower the sg capacity to one so that we'll try
> * and move all the excess tasks away. We lower the capacity
> * of a group only if the local group has the capacity to fit
> - * these excess tasks, i.e. nr_running < group_capacity_factor. The
> + * these excess tasks, i.e. group_capacity > 0. The
> * extra check prevents the case where you always pull from the
> * heaviest group when it is already under-utilized (possible
> * with a large weight task outweighs the tasks on the system).
> */
> if (prefer_sibling && sds->local &&
> - sds->local_stat.group_has_free_capacity)
> - sgs->group_capacity_factor = min(sgs->group_capacity_factor, 1U);
> + group_has_capacity(env, &sds->local_stat)) {
> + if (sgs->sum_nr_running > 1)
> + sgs->group_no_capacity = 1;
> + sgs->group_capacity = min(sgs->group_capacity,
> + SCHED_CAPACITY_SCALE);
> + }
>
> if (update_sd_pick_busiest(env, sds, sg, sgs)) {
> sds->busiest = sg;
So this is your PREFER_SIBLING implementation, why is this a good one?
That is, the current PREFER_SIBLING works because we account against
nr_running, and setting it to 1 makes 2 tasks too much and we end up
moving stuff away.
But if I understand things right, we're now measuring tasks in
'utilization' against group_capacity, so setting group_capacity to
CAPACITY_SCALE, means we can end up with many tasks on the one cpu
before we move over to another group, right?
So I think that for 'idle' systems we want to do the
nr_running/work-conserving thing -- get as many cpus running
'something' and avoid queueing like the plague.
Then when there's some queueing, we want to go do the utilization thing,
basically minimize queueing by leveling utilization.
Once all cpus are fully utilized, we switch to fair/load based balancing
and try and get equal load on cpus.
Does that make sense?
If so, how about adding a group_type and splitting group_other into say
group_idle and group_util:
enum group_type {
group_idle = 0,
group_util,
group_imbalanced,
group_overloaded,
}
we change group_classify() into something like:
if (sgs->group_usage > sgs->group_capacity)
return group_overloaded;
if (sg_imbalanced(group))
return group_imbalanced;
if (sgs->nr_running < sgs->weight)
return group_idle;
return group_util;
And then have update_sd_pick_busiest() something like:
if (sgs->group_type > busiest->group_type)
return true;
if (sgs->group_type < busiest->group_type)
return false;
switch (sgs->group_type) {
case group_idle:
if (sgs->nr_running < busiest->nr_running)
return false;
break;
case group_util:
if (sgs->group_usage < busiest->group_usage)
return false;
break;
default:
if (sgs->avg_load < busiest->avg_load)
return false;
break;
}
....
And then some calculate_imbalance() magic to complete it..
If we have that, we can play tricks with the exact busiest condition in
update_sd_pick_busiest() to implement PREFER_SIBLING or so.
Makes sense?
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/