io_uring: optimise locking for local tw with submit_wait
authorPavel Begunkov <asml.silence@gmail.com>
Thu, 6 Oct 2022 20:42:33 +0000 (21:42 +0100)
committerJens Axboe <axboe@kernel.dk>
Wed, 12 Oct 2022 22:30:56 +0000 (16:30 -0600)
Running local task_work requires taking uring_lock, for submit + wait we
can try to run them right after submit while we still hold the lock and
save one lock/unlokc pair. The optimisation was implemented in the first
local tw patches but got dropped for simplicity.

Suggested-by: Dylan Yudaken <dylany@fb.com>
Signed-off-by: Pavel Begunkov <asml.silence@gmail.com>
Link: https://lore.kernel.org/r/281fc79d98b5d91fe4778c5137a17a2ab4693e5c.1665088876.git.asml.silence@gmail.com
Signed-off-by: Jens Axboe <axboe@kernel.dk>
io_uring/io_uring.c
io_uring/io_uring.h

index 12870cd7cb07cd0bba0c713fa42478abb1f3c960..de08d9902b30b59efaea8784172ff5a3dd3140ca 100644 (file)
@@ -3227,8 +3227,16 @@ SYSCALL_DEFINE6(io_uring_enter, unsigned int, fd, u32, to_submit,
                        mutex_unlock(&ctx->uring_lock);
                        goto out;
                }
-               if ((flags & IORING_ENTER_GETEVENTS) && ctx->syscall_iopoll)
-                       goto iopoll_locked;
+               if (flags & IORING_ENTER_GETEVENTS) {
+                       if (ctx->syscall_iopoll)
+                               goto iopoll_locked;
+                       /*
+                        * Ignore errors, we'll soon call io_cqring_wait() and
+                        * it should handle ownership problems if any.
+                        */
+                       if (ctx->flags & IORING_SETUP_DEFER_TASKRUN)
+                               (void)io_run_local_work_locked(ctx);
+               }
                mutex_unlock(&ctx->uring_lock);
        }
 
index 47d4cad1e9c4684a922902b71d7b0ad120f50b34..ef77d2aa3172ca470a7364af92abf57232ddbc8b 100644 (file)
@@ -275,6 +275,13 @@ static inline int io_run_task_work_ctx(struct io_ring_ctx *ctx)
        return ret;
 }
 
+static inline int io_run_local_work_locked(struct io_ring_ctx *ctx)
+{
+       if (llist_empty(&ctx->work_llist))
+               return 0;
+       return __io_run_local_work(ctx, true);
+}
+
 static inline void io_tw_lock(struct io_ring_ctx *ctx, bool *locked)
 {
        if (!*locked) {