@@ -176,7 +176,6 @@ static void irq_process_pending_llist(struct idxd_irq_entry *irq_entry)
176
176
{
177
177
struct idxd_desc * desc , * t ;
178
178
struct llist_node * head ;
179
- unsigned long flags ;
180
179
181
180
head = llist_del_all (& irq_entry -> pending_llist );
182
181
if (!head )
@@ -197,27 +196,26 @@ static void irq_process_pending_llist(struct idxd_irq_entry *irq_entry)
197
196
198
197
complete_desc (desc , IDXD_COMPLETE_NORMAL );
199
198
} else {
200
- spin_lock_irqsave (& irq_entry -> list_lock , flags );
199
+ spin_lock (& irq_entry -> list_lock );
201
200
list_add_tail (& desc -> list ,
202
201
& irq_entry -> work_list );
203
- spin_unlock_irqrestore (& irq_entry -> list_lock , flags );
202
+ spin_unlock (& irq_entry -> list_lock );
204
203
}
205
204
}
206
205
}
207
206
208
207
static void irq_process_work_list (struct idxd_irq_entry * irq_entry )
209
208
{
210
- unsigned long flags ;
211
209
LIST_HEAD (flist );
212
210
struct idxd_desc * desc , * n ;
213
211
214
212
/*
215
213
* This lock protects list corruption from access of list outside of the irq handler
216
214
* thread.
217
215
*/
218
- spin_lock_irqsave (& irq_entry -> list_lock , flags );
216
+ spin_lock (& irq_entry -> list_lock );
219
217
if (list_empty (& irq_entry -> work_list )) {
220
- spin_unlock_irqrestore (& irq_entry -> list_lock , flags );
218
+ spin_unlock (& irq_entry -> list_lock );
221
219
return ;
222
220
}
223
221
@@ -228,7 +226,7 @@ static void irq_process_work_list(struct idxd_irq_entry *irq_entry)
228
226
}
229
227
}
230
228
231
- spin_unlock_irqrestore (& irq_entry -> list_lock , flags );
229
+ spin_unlock (& irq_entry -> list_lock );
232
230
233
231
list_for_each_entry (desc , & flist , list ) {
234
232
/*
0 commit comments