@@ -2447,26 +2447,6 @@ static bool io_cqring_event_overflow(struct io_ring_ctx *ctx, u64 user_data,
2447
2447
return true;
2448
2448
}
2449
2449
2450
- static inline bool __io_fill_cqe (struct io_ring_ctx * ctx , u64 user_data ,
2451
- s32 res , u32 cflags )
2452
- {
2453
- struct io_uring_cqe * cqe ;
2454
-
2455
- /*
2456
- * If we can't get a cq entry, userspace overflowed the
2457
- * submission (by quite a lot). Increment the overflow count in
2458
- * the ring.
2459
- */
2460
- cqe = io_get_cqe (ctx );
2461
- if (likely (cqe )) {
2462
- WRITE_ONCE (cqe -> user_data , user_data );
2463
- WRITE_ONCE (cqe -> res , res );
2464
- WRITE_ONCE (cqe -> flags , cflags );
2465
- return true;
2466
- }
2467
- return io_cqring_event_overflow (ctx , user_data , res , cflags , 0 , 0 );
2468
- }
2469
-
2470
2450
static inline bool __io_fill_cqe_req (struct io_ring_ctx * ctx ,
2471
2451
struct io_kiocb * req )
2472
2452
{
@@ -2523,9 +2503,24 @@ static inline bool __io_fill_cqe_req(struct io_ring_ctx *ctx,
2523
2503
static noinline bool io_fill_cqe_aux (struct io_ring_ctx * ctx , u64 user_data ,
2524
2504
s32 res , u32 cflags )
2525
2505
{
2506
+ struct io_uring_cqe * cqe ;
2507
+
2526
2508
ctx -> cq_extra ++ ;
2527
2509
trace_io_uring_complete (ctx , NULL , user_data , res , cflags , 0 , 0 );
2528
- return __io_fill_cqe (ctx , user_data , res , cflags );
2510
+
2511
+ /*
2512
+ * If we can't get a cq entry, userspace overflowed the
2513
+ * submission (by quite a lot). Increment the overflow count in
2514
+ * the ring.
2515
+ */
2516
+ cqe = io_get_cqe (ctx );
2517
+ if (likely (cqe )) {
2518
+ WRITE_ONCE (cqe -> user_data , user_data );
2519
+ WRITE_ONCE (cqe -> res , res );
2520
+ WRITE_ONCE (cqe -> flags , cflags );
2521
+ return true;
2522
+ }
2523
+ return io_cqring_event_overflow (ctx , user_data , res , cflags , 0 , 0 );
2529
2524
}
2530
2525
2531
2526
static void __io_req_complete_put (struct io_kiocb * req )
0 commit comments