File tree Expand file tree Collapse file tree 8 files changed +13
-13
lines changed Expand file tree Collapse file tree 8 files changed +13
-13
lines changed Original file line number Diff line number Diff line change @@ -158,7 +158,7 @@ void rcu_scheduler_starting(void);
158
158
static inline void rcu_end_inkernel_boot (void ) { }
159
159
static inline bool rcu_inkernel_boot_has_ended (void ) { return true; }
160
160
static inline bool rcu_is_watching (void ) { return true; }
161
- static inline void rcu_momentary_dyntick_idle (void ) { }
161
+ static inline void rcu_momentary_eqs (void ) { }
162
162
static inline void kfree_rcu_scheduler_running (void ) { }
163
163
static inline bool rcu_gp_might_be_stalled (void ) { return false; }
164
164
Original file line number Diff line number Diff line change @@ -37,7 +37,7 @@ void synchronize_rcu_expedited(void);
37
37
void kvfree_call_rcu (struct rcu_head * head , void * ptr );
38
38
39
39
void rcu_barrier (void );
40
- void rcu_momentary_dyntick_idle (void );
40
+ void rcu_momentary_eqs (void );
41
41
void kfree_rcu_scheduler_running (void );
42
42
bool rcu_gp_might_be_stalled (void );
43
43
Original file line number Diff line number Diff line change @@ -2680,7 +2680,7 @@ static unsigned long rcu_torture_fwd_prog_cbfree(struct rcu_fwd *rfp)
2680
2680
rcu_torture_fwd_prog_cond_resched (freed );
2681
2681
if (tick_nohz_full_enabled ()) {
2682
2682
local_irq_save (flags );
2683
- rcu_momentary_dyntick_idle ();
2683
+ rcu_momentary_eqs ();
2684
2684
local_irq_restore (flags );
2685
2685
}
2686
2686
}
@@ -2830,7 +2830,7 @@ static void rcu_torture_fwd_prog_cr(struct rcu_fwd *rfp)
2830
2830
rcu_torture_fwd_prog_cond_resched (n_launders + n_max_cbs );
2831
2831
if (tick_nohz_full_enabled ()) {
2832
2832
local_irq_save (flags );
2833
- rcu_momentary_dyntick_idle ();
2833
+ rcu_momentary_eqs ();
2834
2834
local_irq_restore (flags );
2835
2835
}
2836
2836
}
Original file line number Diff line number Diff line change @@ -367,7 +367,7 @@ bool rcu_watching_zero_in_eqs(int cpu, int *vp)
367
367
*
368
368
* The caller must have disabled interrupts and must not be idle.
369
369
*/
370
- notrace void rcu_momentary_dyntick_idle (void )
370
+ notrace void rcu_momentary_eqs (void )
371
371
{
372
372
int seq ;
373
373
@@ -377,7 +377,7 @@ notrace void rcu_momentary_dyntick_idle(void)
377
377
WARN_ON_ONCE (!(seq & CT_RCU_WATCHING ));
378
378
rcu_preempt_deferred_qs (current );
379
379
}
380
- EXPORT_SYMBOL_GPL (rcu_momentary_dyntick_idle );
380
+ EXPORT_SYMBOL_GPL (rcu_momentary_eqs );
381
381
382
382
/**
383
383
* rcu_is_cpu_rrupt_from_idle - see if 'interrupted' from idle
Original file line number Diff line number Diff line change @@ -917,7 +917,7 @@ static void nocb_cb_wait(struct rcu_data *rdp)
917
917
WARN_ON_ONCE (!rcu_rdp_is_offloaded (rdp ));
918
918
919
919
local_irq_save (flags );
920
- rcu_momentary_dyntick_idle ();
920
+ rcu_momentary_eqs ();
921
921
local_irq_restore (flags );
922
922
/*
923
923
* Disable BH to provide the expected environment. Also, when
Original file line number Diff line number Diff line change @@ -869,7 +869,7 @@ static void rcu_qs(void)
869
869
870
870
/*
871
871
* Register an urgently needed quiescent state. If there is an
872
- * emergency, invoke rcu_momentary_dyntick_idle () to do a heavy-weight
872
+ * emergency, invoke rcu_momentary_eqs () to do a heavy-weight
873
873
* dyntick-idle quiescent state visible to other CPUs, which will in
874
874
* some cases serve for expedited as well as normal grace periods.
875
875
* Either way, register a lightweight quiescent state.
@@ -889,7 +889,7 @@ void rcu_all_qs(void)
889
889
this_cpu_write (rcu_data .rcu_urgent_qs , false);
890
890
if (unlikely (raw_cpu_read (rcu_data .rcu_need_heavy_qs ))) {
891
891
local_irq_save (flags );
892
- rcu_momentary_dyntick_idle ();
892
+ rcu_momentary_eqs ();
893
893
local_irq_restore (flags );
894
894
}
895
895
rcu_qs ();
@@ -909,7 +909,7 @@ void rcu_note_context_switch(bool preempt)
909
909
goto out ;
910
910
this_cpu_write (rcu_data .rcu_urgent_qs , false);
911
911
if (unlikely (raw_cpu_read (rcu_data .rcu_need_heavy_qs )))
912
- rcu_momentary_dyntick_idle ();
912
+ rcu_momentary_eqs ();
913
913
out :
914
914
rcu_tasks_qs (current , preempt );
915
915
trace_rcu_utilization (TPS ("End context switch" ));
Original file line number Diff line number Diff line change @@ -251,7 +251,7 @@ static int multi_cpu_stop(void *data)
251
251
*/
252
252
touch_nmi_watchdog ();
253
253
}
254
- rcu_momentary_dyntick_idle ();
254
+ rcu_momentary_eqs ();
255
255
} while (curstate != MULTI_STOP_EXIT );
256
256
257
257
local_irq_restore (flags );
Original file line number Diff line number Diff line change @@ -1535,7 +1535,7 @@ static int run_osnoise(void)
1535
1535
* This will eventually cause unwarranted noise as PREEMPT_RCU
1536
1536
* will force preemption as the means of ending the current
1537
1537
* grace period. We avoid this problem by calling
1538
- * rcu_momentary_dyntick_idle (), which performs a zero duration
1538
+ * rcu_momentary_eqs (), which performs a zero duration
1539
1539
* EQS allowing PREEMPT_RCU to end the current grace period.
1540
1540
* This call shouldn't be wrapped inside an RCU critical
1541
1541
* section.
@@ -1547,7 +1547,7 @@ static int run_osnoise(void)
1547
1547
if (!disable_irq )
1548
1548
local_irq_disable ();
1549
1549
1550
- rcu_momentary_dyntick_idle ();
1550
+ rcu_momentary_eqs ();
1551
1551
1552
1552
if (!disable_irq )
1553
1553
local_irq_enable ();
You can’t perform that action at this time.
0 commit comments