@@ -416,7 +416,7 @@ struct sched_ext_ops {
416
416
417
417
/**
418
418
* @update_idle: Update the idle state of a CPU
419
- * @cpu: CPU to udpate the idle state for
419
+ * @cpu: CPU to update the idle state for
420
420
* @idle: whether entering or exiting the idle state
421
421
*
422
422
* This operation is called when @rq's CPU goes or leaves the idle
@@ -1214,7 +1214,7 @@ static bool scx_kf_allowed_if_unlocked(void)
1214
1214
1215
1215
/**
1216
1216
* nldsq_next_task - Iterate to the next task in a non-local DSQ
1217
- * @dsq: user dsq being interated
1217
+ * @dsq: user dsq being iterated
1218
1218
* @cur: current position, %NULL to start iteration
1219
1219
* @rev: walk backwards
1220
1220
*
@@ -2078,7 +2078,7 @@ static void set_task_runnable(struct rq *rq, struct task_struct *p)
2078
2078
2079
2079
/*
2080
2080
* list_add_tail() must be used. scx_ops_bypass() depends on tasks being
2081
- * appened to the runnable_list.
2081
+ * appended to the runnable_list.
2082
2082
*/
2083
2083
list_add_tail (& p -> scx .runnable_node , & rq -> scx .runnable_list );
2084
2084
}
@@ -2480,7 +2480,7 @@ static struct rq *move_task_between_dsqs(struct task_struct *p, u64 enq_flags,
2480
2480
/*
2481
2481
* A poorly behaving BPF scheduler can live-lock the system by e.g. incessantly
2482
2482
* banging on the same DSQ on a large NUMA system to the point where switching
2483
- * to the bypass mode can take a long time. Inject artifical delays while the
2483
+ * to the bypass mode can take a long time. Inject artificial delays while the
2484
2484
* bypass mode is switching to guarantee timely completion.
2485
2485
*/
2486
2486
static void scx_ops_breather (struct rq * rq )
@@ -3144,7 +3144,7 @@ static struct task_struct *pick_task_scx(struct rq *rq)
3144
3144
*
3145
3145
* Unless overridden by ops.core_sched_before(), @p->scx.core_sched_at is used
3146
3146
* to implement the default task ordering. The older the timestamp, the higher
3147
- * prority the task - the global FIFO ordering matching the default scheduling
3147
+ * priority the task - the global FIFO ordering matching the default scheduling
3148
3148
* behavior.
3149
3149
*
3150
3150
* When ops.core_sched_before() is enabled, @p->scx.core_sched_at is used to
@@ -4590,7 +4590,7 @@ static int scx_cgroup_init(void)
4590
4590
cgroup_warned_missing_idle = false;
4591
4591
4592
4592
/*
4593
- * scx_tg_on/offline() are excluded thorugh scx_cgroup_rwsem. If we walk
4593
+ * scx_tg_on/offline() are excluded through scx_cgroup_rwsem. If we walk
4594
4594
* cgroups and init, all online cgroups are initialized.
4595
4595
*/
4596
4596
rcu_read_lock ();
0 commit comments