Commit 8237e045 authored by Jens Axboe's avatar Jens Axboe

io_uring: wrap multi-req freeing in struct req_batch

This cleans up the code a bit, and it allows us to build on top of the
multi-req freeing.
Signed-off-by: default avatarJens Axboe <axboe@kernel.dk>
parent 2b85edfc
...@@ -1132,14 +1132,19 @@ static struct io_kiocb *io_get_req(struct io_ring_ctx *ctx, ...@@ -1132,14 +1132,19 @@ static struct io_kiocb *io_get_req(struct io_ring_ctx *ctx,
return NULL; return NULL;
} }
static void io_free_req_many(struct io_ring_ctx *ctx, void **reqs, int *nr) struct req_batch {
void *reqs[IO_IOPOLL_BATCH];
int to_free;
};
static void io_free_req_many(struct io_ring_ctx *ctx, struct req_batch *rb)
{ {
if (*nr) { if (!rb->to_free)
kmem_cache_free_bulk(req_cachep, *nr, reqs); return;
percpu_ref_put_many(&ctx->refs, *nr); kmem_cache_free_bulk(req_cachep, rb->to_free, rb->reqs);
percpu_ref_put_many(&ctx->file_data->refs, *nr); percpu_ref_put_many(&ctx->refs, rb->to_free);
*nr = 0; percpu_ref_put_many(&ctx->file_data->refs, rb->to_free);
} rb->to_free = 0;
} }
static void __io_req_do_free(struct io_kiocb *req) static void __io_req_do_free(struct io_kiocb *req)
...@@ -1371,7 +1376,7 @@ static inline unsigned int io_sqring_entries(struct io_ring_ctx *ctx) ...@@ -1371,7 +1376,7 @@ static inline unsigned int io_sqring_entries(struct io_ring_ctx *ctx)
return smp_load_acquire(&rings->sq.tail) - ctx->cached_sq_head; return smp_load_acquire(&rings->sq.tail) - ctx->cached_sq_head;
} }
static inline bool io_req_multi_free(struct io_kiocb *req) static inline bool io_req_multi_free(struct req_batch *rb, struct io_kiocb *req)
{ {
/* /*
* If we're not using fixed files, we have to pair the completion part * If we're not using fixed files, we have to pair the completion part
...@@ -1379,8 +1384,12 @@ static inline bool io_req_multi_free(struct io_kiocb *req) ...@@ -1379,8 +1384,12 @@ static inline bool io_req_multi_free(struct io_kiocb *req)
* free for fixed file and non-linked commands. * free for fixed file and non-linked commands.
*/ */
if (((req->flags & (REQ_F_FIXED_FILE|REQ_F_LINK)) == REQ_F_FIXED_FILE) if (((req->flags & (REQ_F_FIXED_FILE|REQ_F_LINK)) == REQ_F_FIXED_FILE)
&& !io_is_fallback_req(req) && !req->io) && !io_is_fallback_req(req) && !req->io) {
rb->reqs[rb->to_free++] = req;
if (unlikely(rb->to_free == ARRAY_SIZE(rb->reqs)))
io_free_req_many(req->ctx, rb);
return true; return true;
}
return false; return false;
} }
...@@ -1391,11 +1400,10 @@ static inline bool io_req_multi_free(struct io_kiocb *req) ...@@ -1391,11 +1400,10 @@ static inline bool io_req_multi_free(struct io_kiocb *req)
static void io_iopoll_complete(struct io_ring_ctx *ctx, unsigned int *nr_events, static void io_iopoll_complete(struct io_ring_ctx *ctx, unsigned int *nr_events,
struct list_head *done) struct list_head *done)
{ {
void *reqs[IO_IOPOLL_BATCH]; struct req_batch rb;
struct io_kiocb *req; struct io_kiocb *req;
int to_free;
to_free = 0; rb.to_free = 0;
while (!list_empty(done)) { while (!list_empty(done)) {
req = list_first_entry(done, struct io_kiocb, list); req = list_first_entry(done, struct io_kiocb, list);
list_del(&req->list); list_del(&req->list);
...@@ -1403,19 +1411,13 @@ static void io_iopoll_complete(struct io_ring_ctx *ctx, unsigned int *nr_events, ...@@ -1403,19 +1411,13 @@ static void io_iopoll_complete(struct io_ring_ctx *ctx, unsigned int *nr_events,
io_cqring_fill_event(req, req->result); io_cqring_fill_event(req, req->result);
(*nr_events)++; (*nr_events)++;
if (refcount_dec_and_test(&req->refs)) { if (refcount_dec_and_test(&req->refs) &&
if (io_req_multi_free(req)) { !io_req_multi_free(&rb, req))
reqs[to_free++] = req;
if (to_free == ARRAY_SIZE(reqs))
io_free_req_many(ctx, reqs, &to_free);
} else {
io_free_req(req); io_free_req(req);
} }
}
}
io_commit_cqring(ctx); io_commit_cqring(ctx);
io_free_req_many(ctx, reqs, &to_free); io_free_req_many(ctx, &rb);
} }
static int io_do_iopoll(struct io_ring_ctx *ctx, unsigned int *nr_events, static int io_do_iopoll(struct io_ring_ctx *ctx, unsigned int *nr_events,
...@@ -3221,30 +3223,25 @@ static void io_poll_complete_work(struct io_wq_work **workptr) ...@@ -3221,30 +3223,25 @@ static void io_poll_complete_work(struct io_wq_work **workptr)
static void __io_poll_flush(struct io_ring_ctx *ctx, struct llist_node *nodes) static void __io_poll_flush(struct io_ring_ctx *ctx, struct llist_node *nodes)
{ {
void *reqs[IO_IOPOLL_BATCH];
struct io_kiocb *req, *tmp; struct io_kiocb *req, *tmp;
int to_free = 0; struct req_batch rb;
rb.to_free = 0;
spin_lock_irq(&ctx->completion_lock); spin_lock_irq(&ctx->completion_lock);
llist_for_each_entry_safe(req, tmp, nodes, llist_node) { llist_for_each_entry_safe(req, tmp, nodes, llist_node) {
hash_del(&req->hash_node); hash_del(&req->hash_node);
io_poll_complete(req, req->result, 0); io_poll_complete(req, req->result, 0);
if (refcount_dec_and_test(&req->refs)) { if (refcount_dec_and_test(&req->refs) &&
if (io_req_multi_free(req)) { !io_req_multi_free(&rb, req)) {
reqs[to_free++] = req;
if (to_free == ARRAY_SIZE(reqs))
io_free_req_many(ctx, reqs, &to_free);
} else {
req->flags |= REQ_F_COMP_LOCKED; req->flags |= REQ_F_COMP_LOCKED;
io_free_req(req); io_free_req(req);
} }
} }
}
spin_unlock_irq(&ctx->completion_lock); spin_unlock_irq(&ctx->completion_lock);
io_cqring_ev_posted(ctx); io_cqring_ev_posted(ctx);
io_free_req_many(ctx, reqs, &to_free); io_free_req_many(ctx, &rb);
} }
static void io_poll_flush(struct io_wq_work **workptr) static void io_poll_flush(struct io_wq_work **workptr)
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment