1 // SPDX-License-Identifier: GPL-2.0
3 * Copyright (c) 2019 Christoph Hellwig.
7 static inline unsigned int bio_max_vecs(unsigned int count)
9 return bio_max_segs(howmany(count, PAGE_SIZE));
13 xfs_flush_bdev_async_endio(
16 complete(bio->bi_private);
20 * Submit a request for an async cache flush to run. If the request queue does
21 * not require flush operations, just skip it altogether. If the caller needs
22 * to wait for the flush completion at a later point in time, they must supply a
23 * valid completion. This will be signalled when the flush completes. The
24 * caller never sees the bio that is issued here.
29 struct block_device *bdev,
30 struct completion *done)
32 struct request_queue *q = bdev->bd_disk->queue;
34 if (!test_bit(QUEUE_FLAG_WC, &q->queue_flags)) {
39 bio_init(bio, bdev, NULL, 0, REQ_OP_WRITE | REQ_PREFLUSH | REQ_SYNC);
40 bio->bi_private = done;
41 bio->bi_end_io = xfs_flush_bdev_async_endio;
47 struct block_device *bdev,
54 unsigned int is_vmalloc = is_vmalloc_addr(data);
55 unsigned int left = count;
59 if (is_vmalloc && op == REQ_OP_WRITE)
60 flush_kernel_vmap_range(data, count);
62 bio = bio_alloc(bdev, bio_max_vecs(left), op | REQ_META | REQ_SYNC,
64 bio->bi_iter.bi_sector = sector;
67 struct page *page = kmem_to_page(data);
68 unsigned int off = offset_in_page(data);
69 unsigned int len = min_t(unsigned, left, PAGE_SIZE - off);
71 while (bio_add_page(bio, page, len, off) != len) {
72 struct bio *prev = bio;
74 bio = bio_alloc(prev->bi_bdev, bio_max_vecs(left),
75 prev->bi_opf, GFP_KERNEL);
76 bio->bi_iter.bi_sector = bio_end_sector(prev);
86 error = submit_bio_wait(bio);
89 if (is_vmalloc && op == REQ_OP_READ)
90 invalidate_kernel_vmap_range(data, count);