提交 44e8c2bf 编写于 作者: C Christoph Hellwig 提交者: Jens Axboe

blk-mq: refactor blk_mq_sched_assign_ioc

blk_mq_sched_assign_ioc now only handles the assigned of the ioc if
the schedule needs it (bfq only at the moment).  The caller to the
per-request initializer is moved out so that it can be merged with
a similar call for the kyber I/O scheduler.
Signed-off-by: NChristoph Hellwig <hch@lst.de>
Signed-off-by: NJens Axboe <axboe@kernel.dk>
上级 9f210738
...@@ -31,12 +31,10 @@ void blk_mq_sched_free_hctx_data(struct request_queue *q, ...@@ -31,12 +31,10 @@ void blk_mq_sched_free_hctx_data(struct request_queue *q,
} }
EXPORT_SYMBOL_GPL(blk_mq_sched_free_hctx_data); EXPORT_SYMBOL_GPL(blk_mq_sched_free_hctx_data);
static void __blk_mq_sched_assign_ioc(struct request_queue *q, void blk_mq_sched_assign_ioc(struct request *rq, struct bio *bio)
struct request *rq,
struct bio *bio,
struct io_context *ioc)
{ {
struct elevator_queue *e = q->elevator; struct request_queue *q = rq->q;
struct io_context *ioc = rq_ioc(bio);
struct io_cq *icq; struct io_cq *icq;
spin_lock_irq(q->queue_lock); spin_lock_irq(q->queue_lock);
...@@ -48,26 +46,8 @@ static void __blk_mq_sched_assign_ioc(struct request_queue *q, ...@@ -48,26 +46,8 @@ static void __blk_mq_sched_assign_ioc(struct request_queue *q,
if (!icq) if (!icq)
return; return;
} }
rq->elv.icq = icq;
if (e && e->type->ops.mq.get_rq_priv &&
e->type->ops.mq.get_rq_priv(q, rq, bio)) {
rq->elv.icq = NULL;
return;
}
rq->rq_flags |= RQF_ELVPRIV;
get_io_context(icq->ioc); get_io_context(icq->ioc);
} rq->elv.icq = icq;
void blk_mq_sched_assign_ioc(struct request_queue *q, struct request *rq,
struct bio *bio)
{
struct io_context *ioc;
ioc = rq_ioc(bio);
if (ioc)
__blk_mq_sched_assign_ioc(q, rq, bio, ioc);
} }
void blk_mq_sched_dispatch_requests(struct blk_mq_hw_ctx *hctx) void blk_mq_sched_dispatch_requests(struct blk_mq_hw_ctx *hctx)
......
...@@ -7,8 +7,7 @@ ...@@ -7,8 +7,7 @@
void blk_mq_sched_free_hctx_data(struct request_queue *q, void blk_mq_sched_free_hctx_data(struct request_queue *q,
void (*exit)(struct blk_mq_hw_ctx *)); void (*exit)(struct blk_mq_hw_ctx *));
void blk_mq_sched_assign_ioc(struct request_queue *q, struct request *rq, void blk_mq_sched_assign_ioc(struct request *rq, struct bio *bio);
struct bio *bio);
void blk_mq_sched_request_inserted(struct request *rq); void blk_mq_sched_request_inserted(struct request *rq);
bool blk_mq_sched_try_merge(struct request_queue *q, struct bio *bio, bool blk_mq_sched_try_merge(struct request_queue *q, struct bio *bio,
......
...@@ -315,8 +315,18 @@ static struct request *blk_mq_get_request(struct request_queue *q, ...@@ -315,8 +315,18 @@ static struct request *blk_mq_get_request(struct request_queue *q,
if (!op_is_flush(op)) { if (!op_is_flush(op)) {
rq->elv.icq = NULL; rq->elv.icq = NULL;
if (e && e->type->icq_cache) if (e && e->type->ops.mq.get_rq_priv) {
blk_mq_sched_assign_ioc(q, rq, bio); if (e->type->icq_cache && rq_ioc(bio))
blk_mq_sched_assign_ioc(rq, bio);
if (e->type->ops.mq.get_rq_priv(q, rq, bio)) {
if (rq->elv.icq)
put_io_context(rq->elv.icq->ioc);
rq->elv.icq = NULL;
} else {
rq->rq_flags |= RQF_ELVPRIV;
}
}
} }
data->hctx->queued++; data->hctx->queued++;
return rq; return rq;
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册