Skip to content

Commit be41bde

Browse files
johnstultz-workPeter Zijlstra
authored andcommitted
sched: Add an initial sketch of the find_proxy_task() function
Add a find_proxy_task() function which doesn't do much. When we select a blocked task to run, we will just deactivate it and pick again. The exception being if it has become unblocked after find_proxy_task() was called. This allows us to validate keeping blocked tasks on the runqueue and later deactivating them is working ok, stressing the failure cases for when a proxy isn't found. Greatly simplified from patch by: Peter Zijlstra (Intel) <[email protected]> Juri Lelli <[email protected]> Valentin Schneider <[email protected]> Connor O'Brien <[email protected]> [jstultz: Split out from larger proxy patch and simplified for review and testing.] Signed-off-by: John Stultz <[email protected]> Signed-off-by: Peter Zijlstra (Intel) <[email protected]> Tested-by: K Prateek Nayak <[email protected]> Link: https://lkml.kernel.org/r/[email protected]
1 parent aa4f74d commit be41bde

File tree

2 files changed

+121
-6
lines changed

2 files changed

+121
-6
lines changed

kernel/sched/core.c

Lines changed: 112 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -6522,11 +6522,13 @@ pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
65226522
/*
65236523
* Helper function for __schedule()
65246524
*
6525-
* If a task does not have signals pending, deactivate it
6526-
* Otherwise marks the task's __state as RUNNING
6525+
* Tries to deactivate the task, unless the should_block arg
6526+
* is false or if a signal is pending. In the case a signal
6527+
* is pending, marks the task's __state as RUNNING (and clear
6528+
* blocked_on).
65276529
*/
65286530
static bool try_to_block_task(struct rq *rq, struct task_struct *p,
6529-
unsigned long *task_state_p)
6531+
unsigned long *task_state_p, bool should_block)
65306532
{
65316533
unsigned long task_state = *task_state_p;
65326534
int flags = DEQUEUE_NOCLOCK;
@@ -6537,6 +6539,16 @@ static bool try_to_block_task(struct rq *rq, struct task_struct *p,
65376539
return false;
65386540
}
65396541

6542+
/*
6543+
* We check should_block after signal_pending because we
6544+
* will want to wake the task in that case. But if
6545+
* should_block is false, its likely due to the task being
6546+
* blocked on a mutex, and we want to keep it on the runqueue
6547+
* to be selectable for proxy-execution.
6548+
*/
6549+
if (!should_block)
6550+
return false;
6551+
65406552
p->sched_contributes_to_load =
65416553
(task_state & TASK_UNINTERRUPTIBLE) &&
65426554
!(task_state & TASK_NOLOAD) &&
@@ -6560,6 +6572,88 @@ static bool try_to_block_task(struct rq *rq, struct task_struct *p,
65606572
return true;
65616573
}
65626574

6575+
#ifdef CONFIG_SCHED_PROXY_EXEC
6576+
static inline void proxy_resched_idle(struct rq *rq)
6577+
{
6578+
put_prev_set_next_task(rq, rq->donor, rq->idle);
6579+
rq_set_donor(rq, rq->idle);
6580+
set_tsk_need_resched(rq->idle);
6581+
}
6582+
6583+
static bool __proxy_deactivate(struct rq *rq, struct task_struct *donor)
6584+
{
6585+
unsigned long state = READ_ONCE(donor->__state);
6586+
6587+
/* Don't deactivate if the state has been changed to TASK_RUNNING */
6588+
if (state == TASK_RUNNING)
6589+
return false;
6590+
/*
6591+
* Because we got donor from pick_next_task(), it is *crucial*
6592+
* that we call proxy_resched_idle() before we deactivate it.
6593+
* As once we deactivate donor, donor->on_rq is set to zero,
6594+
* which allows ttwu() to immediately try to wake the task on
6595+
* another rq. So we cannot use *any* references to donor
6596+
* after that point. So things like cfs_rq->curr or rq->donor
6597+
* need to be changed from next *before* we deactivate.
6598+
*/
6599+
proxy_resched_idle(rq);
6600+
return try_to_block_task(rq, donor, &state, true);
6601+
}
6602+
6603+
static struct task_struct *proxy_deactivate(struct rq *rq, struct task_struct *donor)
6604+
{
6605+
if (!__proxy_deactivate(rq, donor)) {
6606+
/*
6607+
* XXX: For now, if deactivation failed, set donor
6608+
* as unblocked, as we aren't doing proxy-migrations
6609+
* yet (more logic will be needed then).
6610+
*/
6611+
donor->blocked_on = NULL;
6612+
}
6613+
return NULL;
6614+
}
6615+
6616+
/*
6617+
* Initial simple sketch that just deactivates the blocked task
6618+
* chosen by pick_next_task() so we can then pick something that
6619+
* isn't blocked.
6620+
*/
6621+
static struct task_struct *
6622+
find_proxy_task(struct rq *rq, struct task_struct *donor, struct rq_flags *rf)
6623+
{
6624+
struct mutex *mutex;
6625+
6626+
mutex = donor->blocked_on;
6627+
/* Something changed in the chain, so pick again */
6628+
if (!mutex)
6629+
return NULL;
6630+
/*
6631+
* By taking mutex->wait_lock we hold off concurrent mutex_unlock()
6632+
* and ensure @owner sticks around.
6633+
*/
6634+
guard(raw_spinlock)(&mutex->wait_lock);
6635+
6636+
/* Check again that donor is blocked with blocked_lock held */
6637+
if (!task_is_blocked(donor) || mutex != __get_task_blocked_on(donor)) {
6638+
/*
6639+
* Something changed in the blocked_on chain and
6640+
* we don't know if only at this level. So, let's
6641+
* just bail out completely and let __schedule()
6642+
* figure things out (pick_again loop).
6643+
*/
6644+
return NULL; /* do pick_next_task() again */
6645+
}
6646+
return proxy_deactivate(rq, donor);
6647+
}
6648+
#else /* SCHED_PROXY_EXEC */
6649+
static struct task_struct *
6650+
find_proxy_task(struct rq *rq, struct task_struct *donor, struct rq_flags *rf)
6651+
{
6652+
WARN_ONCE(1, "This should never be called in the !SCHED_PROXY_EXEC case\n");
6653+
return donor;
6654+
}
6655+
#endif /* SCHED_PROXY_EXEC */
6656+
65636657
/*
65646658
* __schedule() is the main scheduler function.
65656659
*
@@ -6672,12 +6766,25 @@ static void __sched notrace __schedule(int sched_mode)
66726766
goto picked;
66736767
}
66746768
} else if (!preempt && prev_state) {
6675-
try_to_block_task(rq, prev, &prev_state);
6769+
/*
6770+
* We pass task_is_blocked() as the should_block arg
6771+
* in order to keep mutex-blocked tasks on the runqueue
6772+
* for slection with proxy-exec (without proxy-exec
6773+
* task_is_blocked() will always be false).
6774+
*/
6775+
try_to_block_task(rq, prev, &prev_state,
6776+
!task_is_blocked(prev));
66766777
switch_count = &prev->nvcsw;
66776778
}
66786779

6679-
next = pick_next_task(rq, prev, &rf);
6780+
pick_again:
6781+
next = pick_next_task(rq, rq->donor, &rf);
66806782
rq_set_donor(rq, next);
6783+
if (unlikely(task_is_blocked(next))) {
6784+
next = find_proxy_task(rq, next, &rf);
6785+
if (!next)
6786+
goto pick_again;
6787+
}
66816788
picked:
66826789
clear_tsk_need_resched(prev);
66836790
clear_preempt_need_resched();

kernel/sched/sched.h

Lines changed: 9 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2262,6 +2262,14 @@ static inline int task_current_donor(struct rq *rq, struct task_struct *p)
22622262
return rq->donor == p;
22632263
}
22642264

2265+
static inline bool task_is_blocked(struct task_struct *p)
2266+
{
2267+
if (!sched_proxy_exec())
2268+
return false;
2269+
2270+
return !!p->blocked_on;
2271+
}
2272+
22652273
static inline int task_on_cpu(struct rq *rq, struct task_struct *p)
22662274
{
22672275
return p->on_cpu;
@@ -2459,7 +2467,7 @@ static inline void put_prev_set_next_task(struct rq *rq,
24592467
struct task_struct *prev,
24602468
struct task_struct *next)
24612469
{
2462-
WARN_ON_ONCE(rq->curr != prev);
2470+
WARN_ON_ONCE(rq->donor != prev);
24632471

24642472
__put_prev_set_next_dl_server(rq, prev, next);
24652473

0 commit comments

Comments
 (0)