1313bool z_priq_rb_lessthan (struct rbnode * a , struct rbnode * b );
1414
1515/* Dumb Scheduling */
16- #if defined(CONFIG_SCHED_DUMB )
17- #define _priq_run_init z_priq_dumb_init
18- #define _priq_run_add z_priq_dumb_add
19- #define _priq_run_remove z_priq_dumb_remove
20- #define _priq_run_yield z_priq_dumb_yield
16+ #if defined(CONFIG_SCHED_SIMPLE )
17+ #define _priq_run_init z_priq_simple_init
18+ #define _priq_run_add z_priq_simple_add
19+ #define _priq_run_remove z_priq_simple_remove
20+ #define _priq_run_yield z_priq_simple_yield
2121# if defined(CONFIG_SCHED_CPU_MASK )
22- # define _priq_run_best z_priq_dumb_mask_best
22+ # define _priq_run_best z_priq_simple_mask_best
2323# else
24- # define _priq_run_best z_priq_dumb_best
24+ # define _priq_run_best z_priq_simple_best
2525# endif /* CONFIG_SCHED_CPU_MASK */
2626/* Scalable Scheduling */
2727#elif defined(CONFIG_SCHED_SCALABLE )
@@ -45,10 +45,10 @@ bool z_priq_rb_lessthan(struct rbnode *a, struct rbnode *b);
4545#define _priq_wait_remove z_priq_rb_remove
4646#define _priq_wait_best z_priq_rb_best
4747/* Dumb Wait Queue */
48- #elif defined(CONFIG_WAITQ_DUMB )
49- #define _priq_wait_add z_priq_dumb_add
50- #define _priq_wait_remove z_priq_dumb_remove
51- #define _priq_wait_best z_priq_dumb_best
48+ #elif defined(CONFIG_WAITQ_SIMPLE )
49+ #define _priq_wait_add z_priq_simple_add
50+ #define _priq_wait_remove z_priq_simple_remove
51+ #define _priq_wait_best z_priq_simple_best
5252#endif
5353
5454#if defined(CONFIG_64BIT )
@@ -59,7 +59,7 @@ bool z_priq_rb_lessthan(struct rbnode *a, struct rbnode *b);
5959#define TRAILING_ZEROS u32_count_trailing_zeros
6060#endif /* CONFIG_64BIT */
6161
62- static ALWAYS_INLINE void z_priq_dumb_init (sys_dlist_t * pq )
62+ static ALWAYS_INLINE void z_priq_simple_init (sys_dlist_t * pq )
6363{
6464 sys_dlist_init (pq );
6565}
@@ -105,7 +105,7 @@ static ALWAYS_INLINE int32_t z_sched_prio_cmp(struct k_thread *thread_1, struct
105105 return 0 ;
106106}
107107
108- static ALWAYS_INLINE void z_priq_dumb_add (sys_dlist_t * pq , struct k_thread * thread )
108+ static ALWAYS_INLINE void z_priq_simple_add (sys_dlist_t * pq , struct k_thread * thread )
109109{
110110 struct k_thread * t ;
111111
@@ -119,14 +119,14 @@ static ALWAYS_INLINE void z_priq_dumb_add(sys_dlist_t *pq, struct k_thread *thre
119119 sys_dlist_append (pq , & thread -> base .qnode_dlist );
120120}
121121
122- static ALWAYS_INLINE void z_priq_dumb_remove (sys_dlist_t * pq , struct k_thread * thread )
122+ static ALWAYS_INLINE void z_priq_simple_remove (sys_dlist_t * pq , struct k_thread * thread )
123123{
124124 ARG_UNUSED (pq );
125125
126126 sys_dlist_remove (& thread -> base .qnode_dlist );
127127}
128128
129- static ALWAYS_INLINE void z_priq_dumb_yield (sys_dlist_t * pq )
129+ static ALWAYS_INLINE void z_priq_simple_yield (sys_dlist_t * pq )
130130{
131131#ifndef CONFIG_SMP
132132 sys_dnode_t * n ;
@@ -157,7 +157,7 @@ static ALWAYS_INLINE void z_priq_dumb_yield(sys_dlist_t *pq)
157157#endif
158158}
159159
160- static ALWAYS_INLINE struct k_thread * z_priq_dumb_best (sys_dlist_t * pq )
160+ static ALWAYS_INLINE struct k_thread * z_priq_simple_best (sys_dlist_t * pq )
161161{
162162 struct k_thread * thread = NULL ;
163163 sys_dnode_t * n = sys_dlist_peek_head (pq );
@@ -169,7 +169,7 @@ static ALWAYS_INLINE struct k_thread *z_priq_dumb_best(sys_dlist_t *pq)
169169}
170170
171171#ifdef CONFIG_SCHED_CPU_MASK
172- static ALWAYS_INLINE struct k_thread * z_priq_dumb_mask_best (sys_dlist_t * pq )
172+ static ALWAYS_INLINE struct k_thread * z_priq_simple_mask_best (sys_dlist_t * pq )
173173{
174174 /* With masks enabled we need to be prepared to walk the list
175175 * looking for one we can run
0 commit comments