Skip to content

Commit 3077805

Browse files
committed
Merge tag 'sched-urgent-2020-07-25' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip into master
Pull scheduler fixes from Ingo Molnar: "Fix a race introduced by the recent loadavg race fix, plus add a debug check for a hard to debug case of bogus wakeup function flags" * tag 'sched-urgent-2020-07-25' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip: sched: Warn if garbage is passed to default_wake_function() sched: Fix race against ptrace_freeze_trace()
2 parents 17baa44 + 062d3f9 commit 3077805

File tree

1 file changed

+15
-10
lines changed

1 file changed

+15
-10
lines changed

kernel/sched/core.c

Lines changed: 15 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -4119,9 +4119,6 @@ static void __sched notrace __schedule(bool preempt)
41194119
local_irq_disable();
41204120
rcu_note_context_switch(preempt);
41214121

4122-
/* See deactivate_task() below. */
4123-
prev_state = prev->state;
4124-
41254122
/*
41264123
* Make sure that signal_pending_state()->signal_pending() below
41274124
* can't be reordered with __set_current_state(TASK_INTERRUPTIBLE)
@@ -4145,11 +4142,16 @@ static void __sched notrace __schedule(bool preempt)
41454142
update_rq_clock(rq);
41464143

41474144
switch_count = &prev->nivcsw;
4145+
41484146
/*
4149-
* We must re-load prev->state in case ttwu_remote() changed it
4150-
* before we acquired rq->lock.
4147+
* We must load prev->state once (task_struct::state is volatile), such
4148+
* that:
4149+
*
4150+
* - we form a control dependency vs deactivate_task() below.
4151+
* - ptrace_{,un}freeze_traced() can change ->state underneath us.
41514152
*/
4152-
if (!preempt && prev_state && prev_state == prev->state) {
4153+
prev_state = prev->state;
4154+
if (!preempt && prev_state) {
41534155
if (signal_pending_state(prev_state, prev)) {
41544156
prev->state = TASK_RUNNING;
41554157
} else {
@@ -4163,10 +4165,12 @@ static void __sched notrace __schedule(bool preempt)
41634165

41644166
/*
41654167
* __schedule() ttwu()
4166-
* prev_state = prev->state; if (READ_ONCE(p->on_rq) && ...)
4167-
* LOCK rq->lock goto out;
4168-
* smp_mb__after_spinlock(); smp_acquire__after_ctrl_dep();
4169-
* p->on_rq = 0; p->state = TASK_WAKING;
4168+
* prev_state = prev->state; if (p->on_rq && ...)
4169+
* if (prev_state) goto out;
4170+
* p->on_rq = 0; smp_acquire__after_ctrl_dep();
4171+
* p->state = TASK_WAKING
4172+
*
4173+
* Where __schedule() and ttwu() have matching control dependencies.
41704174
*
41714175
* After this, schedule() must not care about p->state any more.
41724176
*/
@@ -4481,6 +4485,7 @@ asmlinkage __visible void __sched preempt_schedule_irq(void)
44814485
int default_wake_function(wait_queue_entry_t *curr, unsigned mode, int wake_flags,
44824486
void *key)
44834487
{
4488+
WARN_ON_ONCE(IS_ENABLED(CONFIG_SCHED_DEBUG) && wake_flags & ~WF_SYNC);
44844489
return try_to_wake_up(curr->private, mode, wake_flags);
44854490
}
44864491
EXPORT_SYMBOL(default_wake_function);

0 commit comments

Comments
 (0)