* just before they are sent down the IO stack.
*/
struct async_submit_bio {
- struct inode *inode;
+ struct btrfs_inode *inode;
struct bio *bio;
enum btrfs_wq_submit_cmd submit_cmd;
int mirror_num;
ret = btree_submit_bio_start(async->bio);
break;
case WQ_SUBMIT_DATA:
- ret = btrfs_submit_bio_start(async->inode, async->bio);
+ ret = btrfs_submit_bio_start(&async->inode->vfs_inode, async->bio);
break;
case WQ_SUBMIT_DATA_DIO:
- ret = btrfs_submit_bio_start_direct_io(async->inode, async->bio,
- async->dio_file_offset);
+ ret = btrfs_submit_bio_start_direct_io(&async->inode->vfs_inode,
+ async->bio, async->dio_file_offset);
break;
}
if (ret)
{
struct async_submit_bio *async =
container_of(work, struct async_submit_bio, work);
- struct inode *inode = async->inode;
+ struct btrfs_inode *inode = async->inode;
struct btrfs_bio *bbio = btrfs_bio(async->bio);
/* If an error occurred we just want to clean up the bio and move on */
* This changes nothing when cgroups aren't in use.
*/
async->bio->bi_opf |= REQ_CGROUP_PUNT;
- btrfs_submit_bio(btrfs_sb(inode->i_sb), async->bio, async->mirror_num);
+ btrfs_submit_bio(inode->root->fs_info, async->bio, async->mirror_num);
}
static void run_one_async_free(struct btrfs_work *work)
if (!async)
return false;
- async->inode = inode;
+ async->inode = BTRFS_I(inode);
async->bio = bio;
async->mirror_num = mirror_num;
async->submit_cmd = cmd;