block: let blkcg_gq grab request queue's refcnt
authorMing Lei <ming.lei@redhat.com>
Fri, 18 Mar 2022 13:01:43 +0000 (21:01 +0800)
committerJens Axboe <axboe@kernel.dk>
Fri, 18 Mar 2022 15:57:56 +0000 (09:57 -0600)
In the whole lifetime of blkcg_gq instance, ->q will be referred, such
as, ->pd_free_fn() is called in blkg_free, and throtl_pd_free() still
may touch the request queue via &tg->service_queue.pending_timer which
is handled by throtl_pending_timer_fn(), so it is reasonable to grab
request queue's refcnt by blkcg_gq instance.

Previously blkcg_exit_queue() is called from blk_release_queue, and it
is hard to avoid the use-after-free. But recently commit 1059699f87eb ("block:
move blkcg initialization/destroy into disk allocation/release handler")
is merged to for-5.18/block, it becomes simple to fix the issue by simply
grabbing request queue's refcnt.

Reported-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Ming Lei <ming.lei@redhat.com>
Link: https://lore.kernel.org/r/20220318130144.1066064-3-ming.lei@redhat.com
Signed-off-by: Jens Axboe <axboe@kernel.dk>
block/blk-cgroup.c

index fa063c6c0338edc3df20295f84cbcae2c1b46572..d53b0d69dd7363df56ea0d30a908f852a8e1e15d 100644 (file)
@@ -82,6 +82,8 @@ static void blkg_free(struct blkcg_gq *blkg)
                if (blkg->pd[i])
                        blkcg_policy[i]->pd_free_fn(blkg->pd[i]);
 
+       if (blkg->q)
+               blk_put_queue(blkg->q);
        free_percpu(blkg->iostat_cpu);
        percpu_ref_exit(&blkg->refcnt);
        kfree(blkg);
@@ -167,6 +169,9 @@ static struct blkcg_gq *blkg_alloc(struct blkcg *blkcg, struct request_queue *q,
        if (!blkg->iostat_cpu)
                goto err_free;
 
+       if (!blk_get_queue(q))
+               goto err_free;
+
        blkg->q = q;
        INIT_LIST_HEAD(&blkg->q_node);
        spin_lock_init(&blkg->async_bio_lock);