Commit 87dd1d63 authored by Christoph Hellwig's avatar Christoph Hellwig Committed by Jens Axboe

block: move blk_mq_sched_assign_ioc to blk-ioc.c

Move blk_mq_sched_assign_ioc so that many interfaces from the file can
be marked static.  Rename the function to ioc_find_get_icq as well and
return the icq to simplify the interface.
Signed-off-by: default avatarChristoph Hellwig <hch@lst.de>
Link: https://lore.kernel.org/r/20211126115817.2087431-8-hch@lst.deSigned-off-by: default avatarJens Axboe <axboe@kernel.dk>
parent 33047425
...@@ -6666,7 +6666,7 @@ static struct bfq_queue *bfq_get_bfqq_handle_split(struct bfq_data *bfqd, ...@@ -6666,7 +6666,7 @@ static struct bfq_queue *bfq_get_bfqq_handle_split(struct bfq_data *bfqd,
*/ */
static void bfq_prepare_request(struct request *rq) static void bfq_prepare_request(struct request *rq)
{ {
blk_mq_sched_assign_ioc(rq); rq->elv.icq = ioc_find_get_icq(rq->q);
/* /*
* Regardless of whether we have an icq attached, we have to * Regardless of whether we have an icq attached, we have to
......
...@@ -24,7 +24,7 @@ static struct kmem_cache *iocontext_cachep; ...@@ -24,7 +24,7 @@ static struct kmem_cache *iocontext_cachep;
* *
* Increment reference count to @ioc. * Increment reference count to @ioc.
*/ */
void get_io_context(struct io_context *ioc) static void get_io_context(struct io_context *ioc)
{ {
BUG_ON(atomic_long_read(&ioc->refcount) <= 0); BUG_ON(atomic_long_read(&ioc->refcount) <= 0);
atomic_long_inc(&ioc->refcount); atomic_long_inc(&ioc->refcount);
...@@ -248,7 +248,8 @@ void ioc_clear_queue(struct request_queue *q) ...@@ -248,7 +248,8 @@ void ioc_clear_queue(struct request_queue *q)
__ioc_clear_queue(&icq_list); __ioc_clear_queue(&icq_list);
} }
int create_task_io_context(struct task_struct *task, gfp_t gfp_flags, int node) static int create_task_io_context(struct task_struct *task, gfp_t gfp_flags,
int node)
{ {
struct io_context *ioc; struct io_context *ioc;
int ret; int ret;
...@@ -397,8 +398,8 @@ EXPORT_SYMBOL(ioc_lookup_icq); ...@@ -397,8 +398,8 @@ EXPORT_SYMBOL(ioc_lookup_icq);
* The caller is responsible for ensuring @ioc won't go away and @q is * The caller is responsible for ensuring @ioc won't go away and @q is
* alive and will stay alive until this function returns. * alive and will stay alive until this function returns.
*/ */
struct io_cq *ioc_create_icq(struct io_context *ioc, struct request_queue *q, static struct io_cq *ioc_create_icq(struct io_context *ioc,
gfp_t gfp_mask) struct request_queue *q, gfp_t gfp_mask)
{ {
struct elevator_type *et = q->elevator->type; struct elevator_type *et = q->elevator->type;
struct io_cq *icq; struct io_cq *icq;
...@@ -441,6 +442,36 @@ struct io_cq *ioc_create_icq(struct io_context *ioc, struct request_queue *q, ...@@ -441,6 +442,36 @@ struct io_cq *ioc_create_icq(struct io_context *ioc, struct request_queue *q,
return icq; return icq;
} }
struct io_cq *ioc_find_get_icq(struct request_queue *q)
{
struct io_context *ioc;
struct io_cq *icq;
/* create task io_context, if we don't have one already */
if (unlikely(!current->io_context))
create_task_io_context(current, GFP_ATOMIC, q->node);
/*
* May not have an IO context if it's a passthrough request
*/
ioc = current->io_context;
if (!ioc)
return NULL;
spin_lock_irq(&q->queue_lock);
icq = ioc_lookup_icq(ioc, q);
spin_unlock_irq(&q->queue_lock);
if (!icq) {
icq = ioc_create_icq(ioc, q, GFP_ATOMIC);
if (!icq)
return NULL;
}
get_io_context(icq->ioc);
return icq;
}
EXPORT_SYMBOL_GPL(ioc_find_get_icq);
static int __init blk_ioc_init(void) static int __init blk_ioc_init(void)
{ {
iocontext_cachep = kmem_cache_create("blkdev_ioc", iocontext_cachep = kmem_cache_create("blkdev_ioc",
......
...@@ -18,37 +18,6 @@ ...@@ -18,37 +18,6 @@
#include "blk-mq-tag.h" #include "blk-mq-tag.h"
#include "blk-wbt.h" #include "blk-wbt.h"
void blk_mq_sched_assign_ioc(struct request *rq)
{
struct request_queue *q = rq->q;
struct io_context *ioc;
struct io_cq *icq;
/* create task io_context, if we don't have one already */
if (unlikely(!current->io_context))
create_task_io_context(current, GFP_ATOMIC, q->node);
/*
* May not have an IO context if it's a passthrough request
*/
ioc = current->io_context;
if (!ioc)
return;
spin_lock_irq(&q->queue_lock);
icq = ioc_lookup_icq(ioc, q);
spin_unlock_irq(&q->queue_lock);
if (!icq) {
icq = ioc_create_icq(ioc, q, GFP_ATOMIC);
if (!icq)
return;
}
get_io_context(icq->ioc);
rq->elv.icq = icq;
}
EXPORT_SYMBOL_GPL(blk_mq_sched_assign_ioc);
/* /*
* Mark a hardware queue as needing a restart. For shared queues, maintain * Mark a hardware queue as needing a restart. For shared queues, maintain
* a count of how many hardware queues are marked for restart. * a count of how many hardware queues are marked for restart.
......
...@@ -8,8 +8,6 @@ ...@@ -8,8 +8,6 @@
#define MAX_SCHED_RQ (16 * BLKDEV_DEFAULT_RQ) #define MAX_SCHED_RQ (16 * BLKDEV_DEFAULT_RQ)
void blk_mq_sched_assign_ioc(struct request *rq);
bool blk_mq_sched_try_merge(struct request_queue *q, struct bio *bio, bool blk_mq_sched_try_merge(struct request_queue *q, struct bio *bio,
unsigned int nr_segs, struct request **merged_request); unsigned int nr_segs, struct request **merged_request);
bool blk_mq_sched_bio_merge(struct request_queue *q, struct bio *bio, bool blk_mq_sched_bio_merge(struct request_queue *q, struct bio *bio,
......
...@@ -363,14 +363,10 @@ static inline unsigned int bio_aligned_discard_max_sectors( ...@@ -363,14 +363,10 @@ static inline unsigned int bio_aligned_discard_max_sectors(
/* /*
* Internal io_context interface * Internal io_context interface
*/ */
void get_io_context(struct io_context *ioc); struct io_cq *ioc_find_get_icq(struct request_queue *q);
struct io_cq *ioc_lookup_icq(struct io_context *ioc, struct request_queue *q); struct io_cq *ioc_lookup_icq(struct io_context *ioc, struct request_queue *q);
struct io_cq *ioc_create_icq(struct io_context *ioc, struct request_queue *q,
gfp_t gfp_mask);
void ioc_clear_queue(struct request_queue *q); void ioc_clear_queue(struct request_queue *q);
int create_task_io_context(struct task_struct *task, gfp_t gfp_mask, int node);
#ifdef CONFIG_BLK_DEV_THROTTLING_LOW #ifdef CONFIG_BLK_DEV_THROTTLING_LOW
extern ssize_t blk_throtl_sample_time_show(struct request_queue *q, char *page); extern ssize_t blk_throtl_sample_time_show(struct request_queue *q, char *page);
extern ssize_t blk_throtl_sample_time_store(struct request_queue *q, extern ssize_t blk_throtl_sample_time_store(struct request_queue *q,
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment