[PATCH][1/2] sched: sched.c task_t cleanup
From: Con Kolivas
Date: Fri Mar 17 2006 - 08:01:54 EST
Replace all task_struct instances in sched.c with task_t
Signed-off-by: Con Kolivas <kernel@xxxxxxxxxxx>
---
kernel/sched.c | 63 ++++++++++++++++++++++++++++-----------------------------
1 files changed, 31 insertions(+), 32 deletions(-)
Index: linux-2.6.16-rc6-mm1/kernel/sched.c
===================================================================
--- linux-2.6.16-rc6-mm1.orig/kernel/sched.c 2006-03-17 23:32:09.000000000 +1100
+++ linux-2.6.16-rc6-mm1/kernel/sched.c 2006-03-18 00:02:37.000000000 +1100
@@ -607,7 +607,7 @@ static inline void sched_info_switch(tas
/*
* Adding/removing a task to/from a priority array:
*/
-static void dequeue_task(struct task_struct *p, prio_array_t *array)
+static void dequeue_task(task_t *p, prio_array_t *array)
{
array->nr_active--;
list_del(&p->run_list);
@@ -615,7 +615,7 @@ static void dequeue_task(struct task_str
__clear_bit(p->prio, array->bitmap);
}
-static void enqueue_task(struct task_struct *p, prio_array_t *array)
+static void enqueue_task(task_t *p, prio_array_t *array)
{
sched_info_queued(p);
list_add_tail(&p->run_list, array->queue + p->prio);
@@ -628,12 +628,12 @@ static void enqueue_task(struct task_str
* Put task to the end of the run list without the overhead of dequeue
* followed by enqueue.
*/
-static void requeue_task(struct task_struct *p, prio_array_t *array)
+static void requeue_task(task_t *p, prio_array_t *array)
{
list_move_tail(&p->run_list, array->queue + p->prio);
}
-static inline void enqueue_task_head(struct task_struct *p, prio_array_t *array)
+static inline void enqueue_task_head(task_t *p, prio_array_t *array)
{
list_add(&p->run_list, array->queue + p->prio);
__set_bit(p->prio, array->bitmap);
@@ -865,7 +865,7 @@ static void activate_task(task_t *p, run
/*
* deactivate_task - remove a task from the runqueue.
*/
-static void deactivate_task(struct task_struct *p, runqueue_t *rq)
+static void deactivate_task(task_t *p, runqueue_t *rq)
{
dec_nr_running(p, rq);
dequeue_task(p, p->array);
@@ -1058,7 +1058,7 @@ static inline unsigned long cpu_avg_load
* domain.
*/
static struct sched_group *
-find_idlest_group(struct sched_domain *sd, struct task_struct *p, int this_cpu)
+find_idlest_group(struct sched_domain *sd, task_t *p, int this_cpu)
{
struct sched_group *idlest = NULL, *this = NULL, *group = sd->groups;
unsigned long min_load = ULONG_MAX, this_load = 0;
@@ -1112,7 +1112,7 @@ nextgroup:
* find_idlest_queue - find the idlest runqueue among the cpus in group.
*/
static int
-find_idlest_cpu(struct sched_group *group, struct task_struct *p, int this_cpu)
+find_idlest_cpu(struct sched_group *group, task_t *p, int this_cpu)
{
cpumask_t tmp;
unsigned long load, min_load = ULONG_MAX;
@@ -1147,7 +1147,7 @@ find_idlest_cpu(struct sched_group *grou
*/
static int sched_balance_self(int cpu, int flag)
{
- struct task_struct *t = current;
+ task_t *t = current;
struct sched_domain *tmp, *sd = NULL;
for_each_domain(cpu, tmp)
@@ -1672,7 +1672,7 @@ asmlinkage void schedule_tail(task_t *pr
* thread's register state.
*/
static inline
-task_t * context_switch(runqueue_t *rq, task_t *prev, task_t *next)
+task_t *context_switch(runqueue_t *rq, task_t *prev, task_t *next)
{
struct mm_struct *mm = next->mm;
struct mm_struct *oldmm = prev->active_mm;
@@ -1831,7 +1831,7 @@ static void sched_migrate_task(task_t *p
/* force the process onto the specified CPU */
if (migrate_task(p, dest_cpu, &req)) {
/* Need to wait for migration thread (might exit: take ref). */
- struct task_struct *mt = rq->migration_thread;
+ task_t *mt = rq->migration_thread;
get_task_struct(mt);
task_rq_unlock(rq, &flags);
wake_up_process(mt);
@@ -2615,7 +2615,7 @@ unsigned long long current_sched_time(co
* @hardirq_offset: the offset to subtract from hardirq_count()
* @cputime: the cpu time spent in user space since the last update
*/
-void account_user_time(struct task_struct *p, cputime_t cputime)
+void account_user_time(task_t *p, cputime_t cputime)
{
struct cpu_usage_stat *cpustat = &kstat_this_cpu.cpustat;
cputime64_t tmp;
@@ -2636,8 +2636,7 @@ void account_user_time(struct task_struc
* @hardirq_offset: the offset to subtract from hardirq_count()
* @cputime: the cpu time spent in kernel space since the last update
*/
-void account_system_time(struct task_struct *p, int hardirq_offset,
- cputime_t cputime)
+void account_system_time(task_t *p, int hardirq_offset, cputime_t cputime)
{
struct cpu_usage_stat *cpustat = &kstat_this_cpu.cpustat;
runqueue_t *rq = this_rq();
@@ -2666,7 +2665,7 @@ void account_system_time(struct task_str
* @p: the process from which the cpu time has been stolen
* @steal: the cpu time spent in involuntary wait
*/
-void account_steal_time(struct task_struct *p, cputime_t steal)
+void account_steal_time(task_t *p, cputime_t steal)
{
struct cpu_usage_stat *cpustat = &kstat_this_cpu.cpustat;
cputime64_t tmp = cputime_to_cputime64(steal);
@@ -3190,7 +3189,7 @@ asmlinkage void __sched preempt_schedule
{
struct thread_info *ti = current_thread_info();
#ifdef CONFIG_PREEMPT_BKL
- struct task_struct *task = current;
+ task_t *task = current;
int saved_lock_depth;
#endif
/*
@@ -3235,7 +3234,7 @@ asmlinkage void __sched preempt_schedule
{
struct thread_info *ti = current_thread_info();
#ifdef CONFIG_PREEMPT_BKL
- struct task_struct *task = current;
+ task_t *task = current;
int saved_lock_depth;
#endif
/* Catch callers which need to be fixed*/
@@ -3645,7 +3644,7 @@ int can_nice(const task_t *p, const int
capable(CAP_SYS_NICE));
}
-struct task_struct *kgdb_get_idle(int this_cpu)
+task_t *kgdb_get_idle(int this_cpu)
{
return cpu_rq(this_cpu)->idle;
}
@@ -3744,7 +3743,7 @@ static inline task_t *find_process_by_pi
}
/* Actually do priority change: must hold rq lock. */
-static void __setscheduler(struct task_struct *p, int policy, int prio)
+static void __setscheduler(task_t *p, int policy, int prio)
{
BUG_ON(p->array);
p->policy = policy;
@@ -3769,7 +3768,7 @@ static void __setscheduler(struct task_s
* @policy: new policy.
* @param: structure containing the new RT priority.
*/
-int sched_setscheduler(struct task_struct *p, int policy,
+int sched_setscheduler(task_t *p, int policy,
struct sched_param *param)
{
int retval;
@@ -3864,7 +3863,7 @@ do_sched_setscheduler(pid_t pid, int pol
{
int retval;
struct sched_param lparam;
- struct task_struct *p;
+ task_t *p;
if (!param || pid < 0)
return -EINVAL;
@@ -4364,22 +4363,22 @@ out_unlock:
return retval;
}
-static inline struct task_struct *eldest_child(struct task_struct *p)
+static inline task_t *eldest_child(task_t *p)
{
if (list_empty(&p->children)) return NULL;
- return list_entry(p->children.next,struct task_struct,sibling);
+ return list_entry(p->children.next,task_t,sibling);
}
-static inline struct task_struct *older_sibling(struct task_struct *p)
+static inline task_t *older_sibling(task_t *p)
{
if (p->sibling.prev==&p->parent->children) return NULL;
- return list_entry(p->sibling.prev,struct task_struct,sibling);
+ return list_entry(p->sibling.prev,task_t,sibling);
}
-static inline struct task_struct *younger_sibling(struct task_struct *p)
+static inline task_t *younger_sibling(task_t *p)
{
if (p->sibling.next==&p->parent->children) return NULL;
- return list_entry(p->sibling.next,struct task_struct,sibling);
+ return list_entry(p->sibling.next,task_t,sibling);
}
static void show_task(task_t *p)
@@ -4576,7 +4575,7 @@ EXPORT_SYMBOL_GPL(set_cpus_allowed);
* So we race with normal scheduler movements, but that's OK, as long
* as the task is no longer on this CPU.
*/
-static void __migrate_task(struct task_struct *p, int src_cpu, int dest_cpu)
+static void __migrate_task(task_t *p, int src_cpu, int dest_cpu)
{
runqueue_t *rq_dest, *rq_src;
@@ -4679,7 +4678,7 @@ wait_to_die:
#ifdef CONFIG_HOTPLUG_CPU
/* Figure out where task on dead CPU should go, use force if neccessary. */
-static void move_task_off_dead_cpu(int dead_cpu, struct task_struct *tsk)
+static void move_task_off_dead_cpu(int dead_cpu, task_t *tsk)
{
int dest_cpu;
cpumask_t mask;
@@ -4734,7 +4733,7 @@ static void migrate_nr_uninterruptible(r
/* Run through task list and migrate tasks from the dead cpu. */
static void migrate_live_tasks(int src_cpu)
{
- struct task_struct *tsk, *t;
+ task_t *tsk, *t;
write_lock_irq(&tasklist_lock);
@@ -4757,7 +4756,7 @@ void sched_idle_next(void)
{
int cpu = smp_processor_id();
runqueue_t *rq = this_rq();
- struct task_struct *p = rq->idle;
+ task_t *p = rq->idle;
unsigned long flags;
/* cpu has to be offline */
@@ -4953,7 +4952,7 @@ static int migration_call(struct notifie
void *hcpu)
{
int cpu = (long)hcpu;
- struct task_struct *p;
+ task_t *p;
struct runqueue *rq;
unsigned long flags;
@@ -6412,7 +6411,7 @@ EXPORT_SYMBOL(__might_sleep);
#ifdef CONFIG_MAGIC_SYSRQ
void normalize_rt_tasks(void)
{
- struct task_struct *p;
+ task_t *p;
prio_array_t *array;
unsigned long flags;
runqueue_t *rq;
-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/