Re: [PATCH v10 4/7] sched: Add move_queued_task_locked helper

From: Qais Yousef
Date: Tue Jun 04 2024 - 09:16:52 EST


On 05/06/24 21:54, John Stultz wrote:
> From: Connor O'Brien <connoro@xxxxxxxxxx>
>
> Switch logic that deactivates, sets the task cpu,
> and reactivates a task on a different rq to use a
> helper that will be later extended to push entire
> blocked task chains.
>
> This patch was broken out from a larger chain migration
> patch originally by Connor O'Brien.
>
> Cc: Joel Fernandes <joelaf@xxxxxxxxxx>
> Cc: Qais Yousef <qyousef@xxxxxxxxxxx>
> Cc: Ingo Molnar <mingo@xxxxxxxxxx>
> Cc: Peter Zijlstra <peterz@xxxxxxxxxxxxx>
> Cc: Juri Lelli <juri.lelli@xxxxxxxxxx>
> Cc: Vincent Guittot <vincent.guittot@xxxxxxxxxx>
> Cc: Dietmar Eggemann <dietmar.eggemann@xxxxxxx>
> Cc: Valentin Schneider <vschneid@xxxxxxxxxx>
> Cc: Steven Rostedt <rostedt@xxxxxxxxxxx>
> Cc: Ben Segall <bsegall@xxxxxxxxxx>
> Cc: Zimuzo Ezeozue <zezeozue@xxxxxxxxxx>
> Cc: Youssef Esmat <youssefesmat@xxxxxxxxxx>
> Cc: Mel Gorman <mgorman@xxxxxxx>
> Cc: Daniel Bristot de Oliveira <bristot@xxxxxxxxxx>
> Cc: Will Deacon <will@xxxxxxxxxx>
> Cc: Waiman Long <longman@xxxxxxxxxx>
> Cc: Boqun Feng <boqun.feng@xxxxxxxxx>
> Cc: "Paul E. McKenney" <paulmck@xxxxxxxxxx>
> Cc: Metin Kaya <Metin.Kaya@xxxxxxx>
> Cc: Xuewen Yan <xuewen.yan94@xxxxxxxxx>
> Cc: K Prateek Nayak <kprateek.nayak@xxxxxxx>
> Cc: Thomas Gleixner <tglx@xxxxxxxxxxxxx>
> Cc: kernel-team@xxxxxxxxxxx
> Tested-by: K Prateek Nayak <kprateek.nayak@xxxxxxx>
> Tested-by: Metin Kaya <metin.kaya@xxxxxxx>
> Reviewed-by: Metin Kaya <metin.kaya@xxxxxxx>
> Reviewed-by: Valentin Schneider <vschneid@xxxxxxxxxx>
> Signed-off-by: Connor O'Brien <connoro@xxxxxxxxxx>
> [jstultz: split out from larger chain migration patch]
> Signed-off-by: John Stultz <jstultz@xxxxxxxxxx>
> ---
> v8:
> * Renamed from push_task_chain to do_push_task so it makes
> more sense without proxy-execution
> v10:
> * Changed name to move_queued_task_locked as suggested by Valentin
> ---
> kernel/sched/core.c | 4 +---
> kernel/sched/deadline.c | 8 ++------
> kernel/sched/rt.c | 8 ++------
> kernel/sched/sched.h | 11 +++++++++++
> 4 files changed, 16 insertions(+), 15 deletions(-)
>
> diff --git a/kernel/sched/core.c b/kernel/sched/core.c
> index 7019a40457a6..48f0d4b381d5 100644
> --- a/kernel/sched/core.c
> +++ b/kernel/sched/core.c
> @@ -2712,9 +2712,7 @@ int push_cpu_stop(void *arg)
>
> // XXX validate p is still the highest prio task
> if (task_rq(p) == rq) {
> - deactivate_task(rq, p, 0);
> - set_task_cpu(p, lowest_rq->cpu);
> - activate_task(lowest_rq, p, 0);
> + move_queued_task_locked(rq, lowest_rq, p);
> resched_curr(lowest_rq);
> }
>
> diff --git a/kernel/sched/deadline.c b/kernel/sched/deadline.c
> index a04a436af8cc..eaedc69c5e30 100644
> --- a/kernel/sched/deadline.c
> +++ b/kernel/sched/deadline.c
> @@ -2443,9 +2443,7 @@ static int push_dl_task(struct rq *rq)
> goto retry;
> }
>
> - deactivate_task(rq, next_task, 0);
> - set_task_cpu(next_task, later_rq->cpu);
> - activate_task(later_rq, next_task, 0);
> + move_queued_task_locked(rq, later_rq, next_task);
> ret = 1;
>
> resched_curr(later_rq);
> @@ -2531,9 +2529,7 @@ static void pull_dl_task(struct rq *this_rq)
> if (is_migration_disabled(p)) {
> push_task = get_push_task(src_rq);
> } else {
> - deactivate_task(src_rq, p, 0);
> - set_task_cpu(p, this_cpu);
> - activate_task(this_rq, p, 0);
> + move_queued_task_locked(src_rq, this_rq, p);
> dmin = p->dl.deadline;
> resched = true;
> }
> diff --git a/kernel/sched/rt.c b/kernel/sched/rt.c
> index 3261b067b67e..975cb49a64dc 100644
> --- a/kernel/sched/rt.c
> +++ b/kernel/sched/rt.c
> @@ -2106,9 +2106,7 @@ static int push_rt_task(struct rq *rq, bool pull)
> goto retry;
> }
>
> - deactivate_task(rq, next_task, 0);
> - set_task_cpu(next_task, lowest_rq->cpu);
> - activate_task(lowest_rq, next_task, 0);
> + move_queued_task_locked(rq, lowest_rq, next_task);
> resched_curr(lowest_rq);
> ret = 1;
>
> @@ -2379,9 +2377,7 @@ static void pull_rt_task(struct rq *this_rq)
> if (is_migration_disabled(p)) {
> push_task = get_push_task(src_rq);
> } else {
> - deactivate_task(src_rq, p, 0);
> - set_task_cpu(p, this_cpu);
> - activate_task(this_rq, p, 0);
> + move_queued_task_locked(src_rq, this_rq, p);
> resched = true;
> }
> /*
> diff --git a/kernel/sched/sched.h b/kernel/sched/sched.h
> index ae50f212775e..2d41ebe200c7 100644
> --- a/kernel/sched/sched.h
> +++ b/kernel/sched/sched.h
> @@ -3480,5 +3480,16 @@ static inline void init_sched_mm_cid(struct task_struct *t) { }
>
> extern u64 avg_vruntime(struct cfs_rq *cfs_rq);
> extern int entity_eligible(struct cfs_rq *cfs_rq, struct sched_entity *se);
> +#ifdef CONFIG_SMP
> +static inline
> +void move_queued_task_locked(struct rq *rq, struct rq *dst_rq, struct task_struct *task)
> +{
> + lockdep_assert_rq_held(rq);
> + lockdep_assert_rq_held(dst_rq);
> + deactivate_task(rq, task, 0);
> + set_task_cpu(task, dst_rq->cpu);
> + activate_task(dst_rq, task, 0);
> +}
> +#endif

I see this pattern in __migrate_swap_task() and try_steal_cookie(), should they
be converted to?

Beside this

Reviewed-by: Qais Yousef <qyousef@xxxxxxxxxxx>

>
> #endif /* _KERNEL_SCHED_SCHED_H */
> --
> 2.45.0.rc1.225.g2a3ae87e7f-goog
>