[PATCH 09/11] sched: rt-group: dynamic period ticks

From: Peter Zijlstra
Date: Sun Jan 06 2008 - 11:25:53 EST


Disable the period updates for inactive groups.

Signed-off-by: Peter Zijlstra <a.p.zijlstra@xxxxxxxxx>
---
kernel/sched.c | 158 ------------------------------------------------------
kernel/sched_rt.c | 54 ++++++++++++++++++
2 files changed, 53 insertions(+), 159 deletions(-)

Index: linux-2.6/kernel/sched.c
===================================================================
--- linux-2.6.orig/kernel/sched.c
+++ linux-2.6/kernel/sched.c
@@ -5277,158 +5277,6 @@ static inline void sched_init_granularit
sysctl_sched_batch_wakeup_granularity *= factor;
}

-static enum hrtimer_restart sched_rt_period_timer(struct hrtimer *timer)
-{
- struct rt_rq *rt_rq =
- container_of(timer, struct rt_rq, rt_period_timer);
- struct rq *rq = rq_of_rt_rq(rt_rq);
- ktime_t now = ktime_get();
-
- WARN_ON(smp_processor_id() != cpu_of(rq));
- WARN_ON(!in_irq());
-
- spin_lock(&rq->lock);
- update_sched_rt_period(rt_rq);
- spin_unlock(&rq->lock);
-
- hrtimer_forward(timer, now, sched_rt_period(rt_rq));
- return HRTIMER_RESTART;
-}
-
-static void sched_rt_period_start(struct rt_rq *rt_rq)
-{
- ktime_t period = sched_rt_period(rt_rq);
-
- WARN_ON(smp_processor_id() != cpu_of(rq_of_rt_rq(rt_rq)));
-
- for (;;) {
- ktime_t now = ktime_get();
- hrtimer_forward(&rt_rq->rt_period_timer, now, period);
- hrtimer_start(&rt_rq->rt_period_timer,
- rt_rq->rt_period_timer.expires,
- HRTIMER_MODE_ABS);
- if (hrtimer_active(&rt_rq->rt_period_timer))
- break;
- }
-}
-
-#if defined CONFIG_SMP || defined CONFIG_FAIR_GROUP_SCHED
-static void sched_rt_period_stop(struct rt_rq *rt_rq)
-{
- hrtimer_cancel(&rt_rq->rt_period_timer);
-}
-#endif
-
-static void sched_rt_period_start_cpu(int cpu)
-{
- struct rq *rq = cpu_rq(cpu);
- struct rt_rq *rt_rq;
-
- for_each_leaf_rt_rq(rt_rq, rq)
- sched_rt_period_start(rt_rq);
-}
-
-#ifdef CONFIG_SMP
-static void sched_rt_period_stop_cpu(int cpu)
-{
- struct rq *rq = cpu_rq(cpu);
- struct rt_rq *rt_rq;
-
- for_each_leaf_rt_rq(rt_rq, rq)
- sched_rt_period_stop(rt_rq);
-}
-
-static int sched_rt_period_hotplug(struct notifier_block *nfb,
- unsigned long action, void *hcpu)
-{
- int cpu = (unsigned long)hcpu;
-
- switch (action) {
- case CPU_UP_PREPARE:
- case CPU_UP_PREPARE_FROZEN:
- case CPU_DOWN_FAILED:
- case CPU_DOWN_FAILED_FROZEN:
- sched_rt_period_start_cpu(cpu);
- return NOTIFY_OK;
-
- case CPU_DOWN_PREPARE:
- case CPU_DOWN_PREPARE_FROZEN:
- case CPU_UP_CANCELED:
- case CPU_UP_CANCELED_FROZEN:
- sched_rt_period_stop_cpu(cpu);
- return NOTIFY_OK;
-
- case CPU_ONLINE:
- case CPU_ONLINE_FROZEN:
- case CPU_DEAD:
- case CPU_DEAD_FROZEN:
- return NOTIFY_OK;
-
- default:
- return NOTIFY_DONE;
- }
-
- return NOTIFY_OK;
-}
-
-static void __init __sched_rt_period_init(void *arg)
-{
- int cpu = smp_processor_id();
- sched_rt_period_start_cpu(cpu);
-}
-
-static void __init sched_rt_period_init(void)
-{
- on_each_cpu(__sched_rt_period_init, NULL, 0, 1);
- hotcpu_notifier(sched_rt_period_hotplug, 0);
-}
-
-#ifdef CONFIG_FAIR_GROUP_SCHED
-static void __sched_rt_period_init_tg(void *arg)
-{
- struct task_group *tg = arg;
- int cpu = smp_processor_id();
-
- sched_rt_period_start(tg->rt_rq[cpu]);
-}
-
-static void sched_rt_period_init_tg(struct task_group *tg)
-{
- on_each_cpu(__sched_rt_period_init_tg, tg, 0, 1);
-}
-
-static void __sched_rt_period_destroy_tg(void *arg)
-{
- struct task_group *tg = arg;
- int cpu = smp_processor_id();
-
- sched_rt_period_stop(tg->rt_rq[cpu]);
-}
-
-static void sched_rt_period_destroy_tg(struct task_group *tg)
-{
- on_each_cpu(__sched_rt_period_destroy_tg, tg, 0, 1);
-}
-#endif /* CONFIG_FAIR_GROUP_SCHED */
-#else /* CONFIG_SMP */
-static void __init sched_rt_period_init(void)
-{
- sched_rt_period_start_cpu(0);
-}
-
-#ifdef CONFIG_FAIR_GROUP_SCHED
-static void sched_rt_period_init_tg(struct task_group *tg)
-{
- sched_rt_period_start(tg->rt_rq[0]);
-}
-
-static void sched_rt_period_destroy_tg(struct task_group *tg)
-{
- sched_rt_period_stop(tg->rt_rq[0]);
-}
-#endif /* CONFIG_FAIR_GROUP_SCHED */
-#endif /* CONFIG_SMP */
-
#ifdef CONFIG_SMP
/*
* This is how migration works:
@@ -7210,7 +7058,6 @@ void __init sched_init_smp(void)
if (set_cpus_allowed(current, non_isolated_cpus) < 0)
BUG();
sched_init_granularity();
- sched_rt_period_init();

#ifdef CONFIG_FAIR_GROUP_SCHED
if (nr_cpu_ids == 1)
@@ -7231,7 +7078,6 @@ void __init sched_init_smp(void)
void __init sched_init_smp(void)
{
sched_init_granularity();
- sched_rt_period_init();
}
#endif /* CONFIG_SMP */

@@ -7788,8 +7634,6 @@ struct task_group *sched_create_group(vo
list_add_rcu(&tg->list, &task_groups);
unlock_task_group_list();

- sched_rt_period_init_tg(tg);
-
return tg;

err:
@@ -7811,8 +7655,6 @@ void sched_destroy_group(struct task_gro
struct rt_rq *rt_rq = NULL;
int i;

- sched_rt_period_destroy_tg(tg);
-
lock_task_group_list();
for_each_possible_cpu(i) {
cfs_rq = tg->cfs_rq[i];
Index: linux-2.6/kernel/sched_rt.c
===================================================================
--- linux-2.6.orig/kernel/sched_rt.c
+++ linux-2.6/kernel/sched_rt.c
@@ -221,8 +221,10 @@ static int sched_rt_ratio_exceeded(struc

if (rt_rq->rt_time > ratio) {
rt_rq->rt_throttled = 1;
- if (rt_rq_throttled(rt_rq))
+ if (rt_rq_throttled(rt_rq)) {
+ WARN_ON(!hrtimer_active(&rt_rq->rt_period_timer));
sched_rt_ratio_dequeue(rt_rq);
+ }
}

out:
@@ -242,6 +244,52 @@ static void update_sched_rt_period(struc
}
}

+static enum hrtimer_restart sched_rt_period_timer(struct hrtimer *timer)
+{
+ struct rt_rq *rt_rq =
+ container_of(timer, struct rt_rq, rt_period_timer);
+ struct rq *rq = rq_of_rt_rq(rt_rq);
+ ktime_t now = ktime_get();
+
+ WARN_ON(smp_processor_id() != cpu_of(rq));
+ WARN_ON(!in_irq());
+
+ spin_lock(&rq->lock);
+ update_sched_rt_period(rt_rq);
+ spin_unlock(&rq->lock);
+
+ hrtimer_forward(timer, now, sched_rt_period(rt_rq));
+
+ return HRTIMER_RESTART;
+}
+
+static void sched_rt_period_start(struct rt_rq *rt_rq)
+{
+ ktime_t period;
+
+ WARN_ON(smp_processor_id() != cpu_of(rq_of_rt_rq(rt_rq)));
+
+ if (hrtimer_active(&rt_rq->rt_period_timer))
+ return;
+
+ period = sched_rt_period(rt_rq);
+
+ for (;;) {
+ ktime_t now = ktime_get();
+ hrtimer_forward(&rt_rq->rt_period_timer, now, period);
+ hrtimer_start(&rt_rq->rt_period_timer,
+ rt_rq->rt_period_timer.expires,
+ HRTIMER_MODE_ABS);
+ if (hrtimer_active(&rt_rq->rt_period_timer))
+ break;
+ }
+}
+
+static void sched_rt_period_stop(struct rt_rq *rt_rq)
+{
+ hrtimer_cancel(&rt_rq->rt_period_timer);
+}
+
/*
* Update the current task's runtime statistics. Skip current tasks that
* are not in our scheduling class.
@@ -274,6 +322,8 @@ static inline
void inc_rt_tasks(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq)
{
WARN_ON(!rt_prio(rt_se_prio(rt_se)));
+ if (!rt_rq->rt_nr_running && !group_rt_rq(rt_se))
+ sched_rt_period_start(rt_rq);
rt_rq->rt_nr_running++;
#if defined CONFIG_SMP || defined CONFIG_FAIR_GROUP_SCHED
if (rt_se_prio(rt_se) < rt_rq->highest_prio)
@@ -299,6 +349,8 @@ void dec_rt_tasks(struct sched_rt_entity
WARN_ON(!rt_prio(rt_se_prio(rt_se)));
WARN_ON(!rt_rq->rt_nr_running);
rt_rq->rt_nr_running--;
+ if (!rt_rq->rt_nr_running && !group_rt_rq(rt_se))
+ sched_rt_period_stop(rt_rq);
#if defined CONFIG_SMP || defined CONFIG_FAIR_GROUP_SCHED
if (rt_rq->rt_nr_running) {
struct rt_prio_array *array;

--

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/