Commit 25399321 authored by Pavel Begunkov's avatar Pavel Begunkov Committed by Jens Axboe

io_uring: introduce locking helpers for CQE posting

spin_lock(&ctx->completion_lock);
/* post CQEs */
io_commit_cqring(ctx);
spin_unlock(&ctx->completion_lock);
io_cqring_ev_posted(ctx);

We have many places repeating this sequence, and the three function
unlock section is not perfect from the maintainance perspective and also
makes it harder to add new locking/sync trick.

Introduce two helpers. io_cq_lock(), which is simple and only grabs
->completion_lock, and io_cq_unlock_post() encapsulating the three call
section.
Signed-off-by: default avatarPavel Begunkov <asml.silence@gmail.com>
Link: https://lore.kernel.org/r/fe0c682bf7f7b55d9be55b0d034be9c1949277dc.1655684496.git.asml.silence@gmail.comSigned-off-by: default avatarJens Axboe <axboe@kernel.dk>
parent 305bef98
...@@ -527,7 +527,7 @@ void __io_commit_cqring_flush(struct io_ring_ctx *ctx) ...@@ -527,7 +527,7 @@ void __io_commit_cqring_flush(struct io_ring_ctx *ctx)
io_eventfd_signal(ctx); io_eventfd_signal(ctx);
} }
void io_cqring_ev_posted(struct io_ring_ctx *ctx) static inline void io_cqring_ev_posted(struct io_ring_ctx *ctx)
{ {
if (unlikely(ctx->off_timeout_used || ctx->drain_active || if (unlikely(ctx->off_timeout_used || ctx->drain_active ||
ctx->has_evfd)) ctx->has_evfd))
...@@ -536,6 +536,19 @@ void io_cqring_ev_posted(struct io_ring_ctx *ctx) ...@@ -536,6 +536,19 @@ void io_cqring_ev_posted(struct io_ring_ctx *ctx)
io_cqring_wake(ctx); io_cqring_wake(ctx);
} }
static inline void __io_cq_unlock_post(struct io_ring_ctx *ctx)
__releases(ctx->completion_lock)
{
io_commit_cqring(ctx);
spin_unlock(&ctx->completion_lock);
io_cqring_ev_posted(ctx);
}
void io_cq_unlock_post(struct io_ring_ctx *ctx)
{
__io_cq_unlock_post(ctx);
}
/* Returns true if there are no backlogged entries after the flush */ /* Returns true if there are no backlogged entries after the flush */
static bool __io_cqring_overflow_flush(struct io_ring_ctx *ctx, bool force) static bool __io_cqring_overflow_flush(struct io_ring_ctx *ctx, bool force)
{ {
...@@ -548,7 +561,7 @@ static bool __io_cqring_overflow_flush(struct io_ring_ctx *ctx, bool force) ...@@ -548,7 +561,7 @@ static bool __io_cqring_overflow_flush(struct io_ring_ctx *ctx, bool force)
if (ctx->flags & IORING_SETUP_CQE32) if (ctx->flags & IORING_SETUP_CQE32)
cqe_size <<= 1; cqe_size <<= 1;
spin_lock(&ctx->completion_lock); io_cq_lock(ctx);
while (!list_empty(&ctx->cq_overflow_list)) { while (!list_empty(&ctx->cq_overflow_list)) {
struct io_uring_cqe *cqe = io_get_cqe(ctx); struct io_uring_cqe *cqe = io_get_cqe(ctx);
struct io_overflow_cqe *ocqe; struct io_overflow_cqe *ocqe;
...@@ -572,9 +585,7 @@ static bool __io_cqring_overflow_flush(struct io_ring_ctx *ctx, bool force) ...@@ -572,9 +585,7 @@ static bool __io_cqring_overflow_flush(struct io_ring_ctx *ctx, bool force)
atomic_andnot(IORING_SQ_CQ_OVERFLOW, &ctx->rings->sq_flags); atomic_andnot(IORING_SQ_CQ_OVERFLOW, &ctx->rings->sq_flags);
} }
io_commit_cqring(ctx); io_cq_unlock_post(ctx);
spin_unlock(&ctx->completion_lock);
io_cqring_ev_posted(ctx);
return all_flushed; return all_flushed;
} }
...@@ -760,11 +771,9 @@ bool io_post_aux_cqe(struct io_ring_ctx *ctx, ...@@ -760,11 +771,9 @@ bool io_post_aux_cqe(struct io_ring_ctx *ctx,
{ {
bool filled; bool filled;
spin_lock(&ctx->completion_lock); io_cq_lock(ctx);
filled = io_fill_cqe_aux(ctx, user_data, res, cflags); filled = io_fill_cqe_aux(ctx, user_data, res, cflags);
io_commit_cqring(ctx); io_cq_unlock_post(ctx);
spin_unlock(&ctx->completion_lock);
io_cqring_ev_posted(ctx);
return filled; return filled;
} }
...@@ -810,11 +819,9 @@ void io_req_complete_post(struct io_kiocb *req) ...@@ -810,11 +819,9 @@ void io_req_complete_post(struct io_kiocb *req)
{ {
struct io_ring_ctx *ctx = req->ctx; struct io_ring_ctx *ctx = req->ctx;
spin_lock(&ctx->completion_lock); io_cq_lock(ctx);
__io_req_complete_post(req); __io_req_complete_post(req);
io_commit_cqring(ctx); io_cq_unlock_post(ctx);
spin_unlock(&ctx->completion_lock);
io_cqring_ev_posted(ctx);
} }
inline void __io_req_complete(struct io_kiocb *req, unsigned issue_flags) inline void __io_req_complete(struct io_kiocb *req, unsigned issue_flags)
...@@ -946,11 +953,9 @@ static void __io_req_find_next_prep(struct io_kiocb *req) ...@@ -946,11 +953,9 @@ static void __io_req_find_next_prep(struct io_kiocb *req)
{ {
struct io_ring_ctx *ctx = req->ctx; struct io_ring_ctx *ctx = req->ctx;
spin_lock(&ctx->completion_lock); io_cq_lock(ctx);
io_disarm_next(req); io_disarm_next(req);
io_commit_cqring(ctx); io_cq_unlock_post(ctx);
spin_unlock(&ctx->completion_lock);
io_cqring_ev_posted(ctx);
} }
static inline struct io_kiocb *io_req_find_next(struct io_kiocb *req) static inline struct io_kiocb *io_req_find_next(struct io_kiocb *req)
...@@ -984,13 +989,6 @@ static void ctx_flush_and_put(struct io_ring_ctx *ctx, bool *locked) ...@@ -984,13 +989,6 @@ static void ctx_flush_and_put(struct io_ring_ctx *ctx, bool *locked)
percpu_ref_put(&ctx->refs); percpu_ref_put(&ctx->refs);
} }
static inline void ctx_commit_and_unlock(struct io_ring_ctx *ctx)
{
io_commit_cqring(ctx);
spin_unlock(&ctx->completion_lock);
io_cqring_ev_posted(ctx);
}
static void handle_prev_tw_list(struct io_wq_work_node *node, static void handle_prev_tw_list(struct io_wq_work_node *node,
struct io_ring_ctx **ctx, bool *uring_locked) struct io_ring_ctx **ctx, bool *uring_locked)
{ {
...@@ -1006,7 +1004,7 @@ static void handle_prev_tw_list(struct io_wq_work_node *node, ...@@ -1006,7 +1004,7 @@ static void handle_prev_tw_list(struct io_wq_work_node *node,
if (req->ctx != *ctx) { if (req->ctx != *ctx) {
if (unlikely(!*uring_locked && *ctx)) if (unlikely(!*uring_locked && *ctx))
ctx_commit_and_unlock(*ctx); io_cq_unlock_post(*ctx);
ctx_flush_and_put(*ctx, uring_locked); ctx_flush_and_put(*ctx, uring_locked);
*ctx = req->ctx; *ctx = req->ctx;
...@@ -1014,7 +1012,7 @@ static void handle_prev_tw_list(struct io_wq_work_node *node, ...@@ -1014,7 +1012,7 @@ static void handle_prev_tw_list(struct io_wq_work_node *node,
*uring_locked = mutex_trylock(&(*ctx)->uring_lock); *uring_locked = mutex_trylock(&(*ctx)->uring_lock);
percpu_ref_get(&(*ctx)->refs); percpu_ref_get(&(*ctx)->refs);
if (unlikely(!*uring_locked)) if (unlikely(!*uring_locked))
spin_lock(&(*ctx)->completion_lock); io_cq_lock(*ctx);
} }
if (likely(*uring_locked)) { if (likely(*uring_locked)) {
req->io_task_work.func(req, uring_locked); req->io_task_work.func(req, uring_locked);
...@@ -1026,7 +1024,7 @@ static void handle_prev_tw_list(struct io_wq_work_node *node, ...@@ -1026,7 +1024,7 @@ static void handle_prev_tw_list(struct io_wq_work_node *node,
} while (node); } while (node);
if (unlikely(!*uring_locked)) if (unlikely(!*uring_locked))
ctx_commit_and_unlock(*ctx); io_cq_unlock_post(*ctx);
} }
static void handle_tw_list(struct io_wq_work_node *node, static void handle_tw_list(struct io_wq_work_node *node,
...@@ -1261,10 +1259,7 @@ static void __io_submit_flush_completions(struct io_ring_ctx *ctx) ...@@ -1261,10 +1259,7 @@ static void __io_submit_flush_completions(struct io_ring_ctx *ctx)
if (!(req->flags & REQ_F_CQE_SKIP)) if (!(req->flags & REQ_F_CQE_SKIP))
__io_fill_cqe_req(ctx, req); __io_fill_cqe_req(ctx, req);
} }
__io_cq_unlock_post(ctx);
io_commit_cqring(ctx);
spin_unlock(&ctx->completion_lock);
io_cqring_ev_posted(ctx);
io_free_batch_list(ctx, state->compl_reqs.first); io_free_batch_list(ctx, state->compl_reqs.first);
INIT_WQ_LIST(&state->compl_reqs); INIT_WQ_LIST(&state->compl_reqs);
......
...@@ -24,7 +24,6 @@ void __io_req_complete(struct io_kiocb *req, unsigned issue_flags); ...@@ -24,7 +24,6 @@ void __io_req_complete(struct io_kiocb *req, unsigned issue_flags);
void io_req_complete_post(struct io_kiocb *req); void io_req_complete_post(struct io_kiocb *req);
void __io_req_complete_post(struct io_kiocb *req); void __io_req_complete_post(struct io_kiocb *req);
bool io_post_aux_cqe(struct io_ring_ctx *ctx, u64 user_data, s32 res, u32 cflags); bool io_post_aux_cqe(struct io_ring_ctx *ctx, u64 user_data, s32 res, u32 cflags);
void io_cqring_ev_posted(struct io_ring_ctx *ctx);
void __io_commit_cqring_flush(struct io_ring_ctx *ctx); void __io_commit_cqring_flush(struct io_ring_ctx *ctx);
struct page **io_pin_pages(unsigned long ubuf, unsigned long len, int *npages); struct page **io_pin_pages(unsigned long ubuf, unsigned long len, int *npages);
...@@ -66,6 +65,14 @@ bool io_match_task_safe(struct io_kiocb *head, struct task_struct *task, ...@@ -66,6 +65,14 @@ bool io_match_task_safe(struct io_kiocb *head, struct task_struct *task,
#define io_for_each_link(pos, head) \ #define io_for_each_link(pos, head) \
for (pos = (head); pos; pos = pos->link) for (pos = (head); pos; pos = pos->link)
static inline void io_cq_lock(struct io_ring_ctx *ctx)
__acquires(ctx->completion_lock)
{
spin_lock(&ctx->completion_lock);
}
void io_cq_unlock_post(struct io_ring_ctx *ctx);
static inline struct io_uring_cqe *io_get_cqe(struct io_ring_ctx *ctx) static inline struct io_uring_cqe *io_get_cqe(struct io_ring_ctx *ctx)
{ {
if (likely(ctx->cqe_cached < ctx->cqe_sentinel)) { if (likely(ctx->cqe_cached < ctx->cqe_sentinel)) {
......
...@@ -617,7 +617,7 @@ __cold bool io_kill_timeouts(struct io_ring_ctx *ctx, struct task_struct *tsk, ...@@ -617,7 +617,7 @@ __cold bool io_kill_timeouts(struct io_ring_ctx *ctx, struct task_struct *tsk,
struct io_timeout *timeout, *tmp; struct io_timeout *timeout, *tmp;
int canceled = 0; int canceled = 0;
spin_lock(&ctx->completion_lock); io_cq_lock(ctx);
spin_lock_irq(&ctx->timeout_lock); spin_lock_irq(&ctx->timeout_lock);
list_for_each_entry_safe(timeout, tmp, &ctx->timeout_list, list) { list_for_each_entry_safe(timeout, tmp, &ctx->timeout_list, list) {
struct io_kiocb *req = cmd_to_io_kiocb(timeout); struct io_kiocb *req = cmd_to_io_kiocb(timeout);
...@@ -627,8 +627,6 @@ __cold bool io_kill_timeouts(struct io_ring_ctx *ctx, struct task_struct *tsk, ...@@ -627,8 +627,6 @@ __cold bool io_kill_timeouts(struct io_ring_ctx *ctx, struct task_struct *tsk,
canceled++; canceled++;
} }
spin_unlock_irq(&ctx->timeout_lock); spin_unlock_irq(&ctx->timeout_lock);
io_commit_cqring(ctx); io_cq_unlock_post(ctx);
spin_unlock(&ctx->completion_lock);
io_cqring_ev_posted(ctx);
return canceled != 0; return canceled != 0;
} }
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment