@@ -5344,8 +5344,6 @@ int apply_workqueue_attrs(struct workqueue_struct *wq,
5344
5344
* wq_update_pod - update pod affinity of a wq for CPU hot[un]plug
5345
5345
* @wq: the target workqueue
5346
5346
* @cpu: the CPU to update pool association for
5347
- * @hotplug_cpu: the CPU coming up or going down
5348
- * @online: whether @cpu is coming up or going down
5349
5347
*
5350
5348
* This function is to be called from %CPU_DOWN_PREPARE, %CPU_ONLINE and
5351
5349
* %CPU_DOWN_FAILED. @cpu is being hot[un]plugged, update pod affinity of
@@ -5362,8 +5360,7 @@ int apply_workqueue_attrs(struct workqueue_struct *wq,
5362
5360
* CPU_DOWN. If a workqueue user wants strict affinity, it's the user's
5363
5361
* responsibility to flush the work item from CPU_DOWN_PREPARE.
5364
5362
*/
5365
- static void wq_update_pod (struct workqueue_struct * wq , int cpu ,
5366
- int hotplug_cpu , bool online )
5363
+ static void wq_update_pod (struct workqueue_struct * wq , int cpu )
5367
5364
{
5368
5365
struct pool_workqueue * old_pwq = NULL , * pwq ;
5369
5366
struct workqueue_attrs * target_attrs ;
@@ -6584,7 +6581,7 @@ int workqueue_online_cpu(unsigned int cpu)
6584
6581
int tcpu ;
6585
6582
6586
6583
for_each_cpu (tcpu , pt -> pod_cpus [pt -> cpu_pod [cpu ]])
6587
- wq_update_pod (wq , tcpu , cpu , true );
6584
+ wq_update_pod (wq , tcpu );
6588
6585
6589
6586
mutex_lock (& wq -> mutex );
6590
6587
wq_update_node_max_active (wq , -1 );
@@ -6619,7 +6616,7 @@ int workqueue_offline_cpu(unsigned int cpu)
6619
6616
int tcpu ;
6620
6617
6621
6618
for_each_cpu (tcpu , pt -> pod_cpus [pt -> cpu_pod [cpu ]])
6622
- wq_update_pod (wq , tcpu , cpu , false );
6619
+ wq_update_pod (wq , tcpu );
6623
6620
6624
6621
mutex_lock (& wq -> mutex );
6625
6622
wq_update_node_max_active (wq , cpu );
@@ -6908,7 +6905,7 @@ static int wq_affn_dfl_set(const char *val, const struct kernel_param *kp)
6908
6905
6909
6906
list_for_each_entry (wq , & workqueues , list ) {
6910
6907
for_each_online_cpu (cpu ) {
6911
- wq_update_pod (wq , cpu , cpu , true );
6908
+ wq_update_pod (wq , cpu );
6912
6909
}
6913
6910
}
6914
6911
@@ -7927,7 +7924,7 @@ void __init workqueue_init_topology(void)
7927
7924
*/
7928
7925
list_for_each_entry (wq , & workqueues , list ) {
7929
7926
for_each_online_cpu (cpu )
7930
- wq_update_pod (wq , cpu , cpu , true );
7927
+ wq_update_pod (wq , cpu );
7931
7928
if (wq -> flags & WQ_UNBOUND ) {
7932
7929
mutex_lock (& wq -> mutex );
7933
7930
wq_update_node_max_active (wq , -1 );
0 commit comments