blk-mq: move the blk_mq_hctx_stopped check in __blk_mq_delay_run_hw_queue
authorChristoph Hellwig <hch@lst.de>
Thu, 13 Apr 2023 06:06:49 +0000 (08:06 +0200)
committerJens Axboe <axboe@kernel.dk>
Thu, 13 Apr 2023 12:57:18 +0000 (06:57 -0600)
For the in-context dispatch, blk_mq_hctx_stopped is alredy checked in
blk_mq_sched_dispatch_requests under blk_mq_run_dispatch_ops() protection.
For the async dispatch case having a check before scheduling the work
still makes sense to avoid needless workqueue scheduling, so just keep it
for that case.

Signed-off-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Damien Le Moal <dlemoal@kernel.org>
Link: https://lore.kernel.org/r/20230413060651.694656-4-hch@lst.de
Signed-off-by: Jens Axboe <axboe@kernel.dk>
block/blk-mq.c

index ad13d2a..2213117 100644 (file)
@@ -2223,9 +2223,6 @@ select_cpu:
 static void __blk_mq_delay_run_hw_queue(struct blk_mq_hw_ctx *hctx, bool async,
                                        unsigned long msecs)
 {
-       if (unlikely(blk_mq_hctx_stopped(hctx)))
-               return;
-
        if (!async && !(hctx->flags & BLK_MQ_F_BLOCKING)) {
                if (cpumask_test_cpu(raw_smp_processor_id(), hctx->cpumask)) {
                        __blk_mq_run_hw_queue(hctx);
@@ -2233,6 +2230,8 @@ static void __blk_mq_delay_run_hw_queue(struct blk_mq_hw_ctx *hctx, bool async,
                }
        }
 
+       if (unlikely(blk_mq_hctx_stopped(hctx)))
+               return;
        kblockd_mod_delayed_work_on(blk_mq_hctx_next_cpu(hctx), &hctx->run_work,
                                    msecs_to_jiffies(msecs));
 }