Commit f39c8a5b authored by Pavel Begunkov's avatar Pavel Begunkov Committed by Jens Axboe

io_uring: inline io_iopoll_getevents()

io_iopoll_getevents() is of no use to us anymore, io_iopoll_check()
handles all the cases.
Signed-off-by: default avatarPavel Begunkov <asml.silence@gmail.com>
Link: https://lore.kernel.org/r/7e50b8917390f38bee4f822c6f4a6a98a27be037.1618278933.git.asml.silence@gmail.comSigned-off-by: default avatarJens Axboe <axboe@kernel.dk>
parent e9979b36
...@@ -2329,27 +2329,6 @@ static int io_do_iopoll(struct io_ring_ctx *ctx, unsigned int *nr_events, ...@@ -2329,27 +2329,6 @@ static int io_do_iopoll(struct io_ring_ctx *ctx, unsigned int *nr_events,
return ret; return ret;
} }
/*
* Poll for a minimum of 'min' events. Note that if min == 0 we consider that a
* non-spinning poll check - we'll still enter the driver poll loop, but only
* as a non-spinning completion check.
*/
static int io_iopoll_getevents(struct io_ring_ctx *ctx, unsigned int *nr_events,
long min)
{
while (!list_empty(&ctx->iopoll_list) && !need_resched()) {
int ret;
ret = io_do_iopoll(ctx, nr_events, min);
if (ret < 0)
return ret;
if (*nr_events >= min)
return 0;
}
return 1;
}
/* /*
* We can't just wait for polled events to come to us, we have to actively * We can't just wait for polled events to come to us, we have to actively
* find and complete them. * find and complete them.
...@@ -2393,7 +2372,6 @@ static int io_iopoll_check(struct io_ring_ctx *ctx, long min) ...@@ -2393,7 +2372,6 @@ static int io_iopoll_check(struct io_ring_ctx *ctx, long min)
* that got punted to a workqueue. * that got punted to a workqueue.
*/ */
mutex_lock(&ctx->uring_lock); mutex_lock(&ctx->uring_lock);
do {
/* /*
* Don't enter poll loop if we already have events pending. * Don't enter poll loop if we already have events pending.
* If we do, we can potentially be spinning for commands that * If we do, we can potentially be spinning for commands that
...@@ -2402,8 +2380,8 @@ static int io_iopoll_check(struct io_ring_ctx *ctx, long min) ...@@ -2402,8 +2380,8 @@ static int io_iopoll_check(struct io_ring_ctx *ctx, long min)
if (test_bit(0, &ctx->cq_check_overflow)) if (test_bit(0, &ctx->cq_check_overflow))
__io_cqring_overflow_flush(ctx, false); __io_cqring_overflow_flush(ctx, false);
if (io_cqring_events(ctx)) if (io_cqring_events(ctx))
break; goto out;
do {
/* /*
* If a submit got punted to a workqueue, we can have the * If a submit got punted to a workqueue, we can have the
* application entering polling for a command before it gets * application entering polling for a command before it gets
...@@ -2422,13 +2400,9 @@ static int io_iopoll_check(struct io_ring_ctx *ctx, long min) ...@@ -2422,13 +2400,9 @@ static int io_iopoll_check(struct io_ring_ctx *ctx, long min)
if (list_empty(&ctx->iopoll_list)) if (list_empty(&ctx->iopoll_list))
break; break;
} }
ret = io_do_iopoll(ctx, &nr_events, min);
ret = io_iopoll_getevents(ctx, &nr_events, min); } while (!ret && nr_events < min && !need_resched());
if (ret <= 0) out:
break;
ret = 0;
} while (min && !nr_events && !need_resched());
mutex_unlock(&ctx->uring_lock); mutex_unlock(&ctx->uring_lock);
return ret; return ret;
} }
...@@ -2539,7 +2513,7 @@ static void io_complete_rw_iopoll(struct kiocb *kiocb, long res, long res2) ...@@ -2539,7 +2513,7 @@ static void io_complete_rw_iopoll(struct kiocb *kiocb, long res, long res2)
/* /*
* After the iocb has been issued, it's safe to be found on the poll list. * After the iocb has been issued, it's safe to be found on the poll list.
* Adding the kiocb to the list AFTER submission ensures that we don't * Adding the kiocb to the list AFTER submission ensures that we don't
* find it from a io_iopoll_getevents() thread before the issuer is done * find it from a io_do_iopoll() thread before the issuer is done
* accessing the kiocb cookie. * accessing the kiocb cookie.
*/ */
static void io_iopoll_req_issued(struct io_kiocb *req, bool in_async) static void io_iopoll_req_issued(struct io_kiocb *req, bool in_async)
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment