Skip to content

Commit be39617

Browse files
Valentin SchneiderPeter Zijlstra
authored andcommitted
sched: Fix proxy/current (push,pull)ability
Proxy execution forms atomic pairs of tasks: The waiting donor task (scheduling context) and a proxy (execution context). The donor task, along with the rest of the blocked chain, follows the proxy wrt CPU placement. They can be the same task, in which case push/pull doesn't need any modification. When they are different, however, FIFO1 & FIFO42: ,-> RT42 | | blocked-on | v blocked_donor | mutex | | owner | v `-- RT1 RT1 RT42 CPU0 CPU1 ^ ^ | | overloaded !overloaded rq prio = 42 rq prio = 0 RT1 is eligible to be pushed to CPU1, but should that happen it will "carry" RT42 along. Clearly here neither RT1 nor RT42 must be seen as push/pullable. Unfortunately, only the donor task is usually dequeued from the rq, and the proxy'ed execution context (rq->curr) remains on the rq. This can cause RT1 to be selected for migration from logic like the rt pushable_list. Thus, adda a dequeue/enqueue cycle on the proxy task before __schedule returns, which allows the sched class logic to avoid adding the now current task to the pushable_list. Furthermore, tasks becoming blocked on a mutex don't need an explicit dequeue/enqueue cycle to be made (push/pull)able: they have to be running to block on a mutex, thus they will eventually hit put_prev_task(). Signed-off-by: Valentin Schneider <[email protected]> Signed-off-by: Connor O'Brien <[email protected]> Signed-off-by: John Stultz <[email protected]> Signed-off-by: Peter Zijlstra (Intel) <[email protected]> Tested-by: K Prateek Nayak <[email protected]> Link: https://lkml.kernel.org/r/[email protected]
1 parent be41bde commit be39617

File tree

3 files changed

+37
-0
lines changed

3 files changed

+37
-0
lines changed

kernel/sched/core.c

Lines changed: 25 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -6654,6 +6654,23 @@ find_proxy_task(struct rq *rq, struct task_struct *donor, struct rq_flags *rf)
66546654
}
66556655
#endif /* SCHED_PROXY_EXEC */
66566656

6657+
static inline void proxy_tag_curr(struct rq *rq, struct task_struct *owner)
6658+
{
6659+
if (!sched_proxy_exec())
6660+
return;
6661+
/*
6662+
* pick_next_task() calls set_next_task() on the chosen task
6663+
* at some point, which ensures it is not push/pullable.
6664+
* However, the chosen/donor task *and* the mutex owner form an
6665+
* atomic pair wrt push/pull.
6666+
*
6667+
* Make sure owner we run is not pushable. Unfortunately we can
6668+
* only deal with that by means of a dequeue/enqueue cycle. :-/
6669+
*/
6670+
dequeue_task(rq, owner, DEQUEUE_NOCLOCK | DEQUEUE_SAVE);
6671+
enqueue_task(rq, owner, ENQUEUE_NOCLOCK | ENQUEUE_RESTORE);
6672+
}
6673+
66576674
/*
66586675
* __schedule() is the main scheduler function.
66596676
*
@@ -6798,6 +6815,10 @@ static void __sched notrace __schedule(int sched_mode)
67986815
* changes to task_struct made by pick_next_task().
67996816
*/
68006817
RCU_INIT_POINTER(rq->curr, next);
6818+
6819+
if (!task_current_donor(rq, next))
6820+
proxy_tag_curr(rq, next);
6821+
68016822
/*
68026823
* The membarrier system call requires each architecture
68036824
* to have a full memory barrier after updating
@@ -6832,6 +6853,10 @@ static void __sched notrace __schedule(int sched_mode)
68326853
/* Also unlocks the rq: */
68336854
rq = context_switch(rq, prev, next, &rf);
68346855
} else {
6856+
/* In case next was already curr but just got blocked_donor */
6857+
if (!task_current_donor(rq, next))
6858+
proxy_tag_curr(rq, next);
6859+
68356860
rq_unpin_lock(rq, &rf);
68366861
__balance_callbacks(rq);
68376862
raw_spin_rq_unlock_irq(rq);

kernel/sched/deadline.c

Lines changed: 7 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -2121,6 +2121,9 @@ static void enqueue_task_dl(struct rq *rq, struct task_struct *p, int flags)
21212121
if (dl_server(&p->dl))
21222122
return;
21232123

2124+
if (task_is_blocked(p))
2125+
return;
2126+
21242127
if (!task_current(rq, p) && !p->dl.dl_throttled && p->nr_cpus_allowed > 1)
21252128
enqueue_pushable_dl_task(rq, p);
21262129
}
@@ -2415,6 +2418,10 @@ static void put_prev_task_dl(struct rq *rq, struct task_struct *p, struct task_s
24152418
update_curr_dl(rq);
24162419

24172420
update_dl_rq_load_avg(rq_clock_pelt(rq), rq, 1);
2421+
2422+
if (task_is_blocked(p))
2423+
return;
2424+
24182425
if (on_dl_rq(&p->dl) && p->nr_cpus_allowed > 1)
24192426
enqueue_pushable_dl_task(rq, p);
24202427
}

kernel/sched/rt.c

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1440,6 +1440,9 @@ enqueue_task_rt(struct rq *rq, struct task_struct *p, int flags)
14401440

14411441
enqueue_rt_entity(rt_se, flags);
14421442

1443+
if (task_is_blocked(p))
1444+
return;
1445+
14431446
if (!task_current(rq, p) && p->nr_cpus_allowed > 1)
14441447
enqueue_pushable_task(rq, p);
14451448
}
@@ -1716,6 +1719,8 @@ static void put_prev_task_rt(struct rq *rq, struct task_struct *p, struct task_s
17161719

17171720
update_rt_rq_load_avg(rq_clock_pelt(rq), rq, 1);
17181721

1722+
if (task_is_blocked(p))
1723+
return;
17191724
/*
17201725
* The previous task needs to be made eligible for pushing
17211726
* if it is still active

0 commit comments

Comments
 (0)