btrfs: fix race between quota disable and quota assign ioctls
[platform/kernel/linux-rpi.git] / fs / btrfs / relocation.c
index 914d403..becf339 100644 (file)
@@ -1147,7 +1147,8 @@ int replace_file_extents(struct btrfs_trans_handle *trans,
                                       num_bytes, parent);
                ref.real_root = root->root_key.objectid;
                btrfs_init_data_ref(&ref, btrfs_header_owner(leaf),
-                                   key.objectid, key.offset);
+                                   key.objectid, key.offset,
+                                   root->root_key.objectid, false);
                ret = btrfs_inc_extent_ref(trans, &ref);
                if (ret) {
                        btrfs_abort_transaction(trans, ret);
@@ -1158,7 +1159,8 @@ int replace_file_extents(struct btrfs_trans_handle *trans,
                                       num_bytes, parent);
                ref.real_root = root->root_key.objectid;
                btrfs_init_data_ref(&ref, btrfs_header_owner(leaf),
-                                   key.objectid, key.offset);
+                                   key.objectid, key.offset,
+                                   root->root_key.objectid, false);
                ret = btrfs_free_extent(trans, &ref);
                if (ret) {
                        btrfs_abort_transaction(trans, ret);
@@ -1324,7 +1326,9 @@ again:
                btrfs_release_path(path);
 
                path->lowest_level = level;
+               set_bit(BTRFS_ROOT_RESET_LOCKDEP_CLASS, &src->state);
                ret = btrfs_search_slot(trans, src, &key, path, 0, 1);
+               clear_bit(BTRFS_ROOT_RESET_LOCKDEP_CLASS, &src->state);
                path->lowest_level = 0;
                if (ret) {
                        if (ret > 0)
@@ -1368,7 +1372,8 @@ again:
                btrfs_init_generic_ref(&ref, BTRFS_ADD_DELAYED_REF, old_bytenr,
                                       blocksize, path->nodes[level]->start);
                ref.skip_qgroup = true;
-               btrfs_init_tree_ref(&ref, level - 1, src->root_key.objectid);
+               btrfs_init_tree_ref(&ref, level - 1, src->root_key.objectid,
+                                   0, true);
                ret = btrfs_inc_extent_ref(trans, &ref);
                if (ret) {
                        btrfs_abort_transaction(trans, ret);
@@ -1377,7 +1382,8 @@ again:
                btrfs_init_generic_ref(&ref, BTRFS_ADD_DELAYED_REF, new_bytenr,
                                       blocksize, 0);
                ref.skip_qgroup = true;
-               btrfs_init_tree_ref(&ref, level - 1, dest->root_key.objectid);
+               btrfs_init_tree_ref(&ref, level - 1, dest->root_key.objectid, 0,
+                                   true);
                ret = btrfs_inc_extent_ref(trans, &ref);
                if (ret) {
                        btrfs_abort_transaction(trans, ret);
@@ -1386,7 +1392,8 @@ again:
 
                btrfs_init_generic_ref(&ref, BTRFS_DROP_DELAYED_REF, new_bytenr,
                                       blocksize, path->nodes[level]->start);
-               btrfs_init_tree_ref(&ref, level - 1, src->root_key.objectid);
+               btrfs_init_tree_ref(&ref, level - 1, src->root_key.objectid,
+                                   0, true);
                ref.skip_qgroup = true;
                ret = btrfs_free_extent(trans, &ref);
                if (ret) {
@@ -1396,7 +1403,8 @@ again:
 
                btrfs_init_generic_ref(&ref, BTRFS_DROP_DELAYED_REF, old_bytenr,
                                       blocksize, 0);
-               btrfs_init_tree_ref(&ref, level - 1, dest->root_key.objectid);
+               btrfs_init_tree_ref(&ref, level - 1, dest->root_key.objectid,
+                                   0, true);
                ref.skip_qgroup = true;
                ret = btrfs_free_extent(trans, &ref);
                if (ret) {
@@ -2475,7 +2483,8 @@ static int do_relocation(struct btrfs_trans_handle *trans,
                                               upper->eb->start);
                        ref.real_root = root->root_key.objectid;
                        btrfs_init_tree_ref(&ref, node->level,
-                                           btrfs_header_owner(upper->eb));
+                                           btrfs_header_owner(upper->eb),
+                                           root->root_key.objectid, false);
                        ret = btrfs_inc_extent_ref(trans, &ref);
                        if (!ret)
                                ret = btrfs_drop_subtree(trans, root, eb,
@@ -2691,8 +2700,12 @@ static int relocate_tree_block(struct btrfs_trans_handle *trans,
                        list_add_tail(&node->list, &rc->backref_cache.changed);
                } else {
                        path->lowest_level = node->level;
+                       if (root == root->fs_info->chunk_root)
+                               btrfs_reserve_chunk_metadata(trans, false);
                        ret = btrfs_search_slot(trans, root, key, path, 0, 1);
                        btrfs_release_path(path);
+                       if (root == root->fs_info->chunk_root)
+                               btrfs_trans_release_chunk_metadata(trans);
                        if (ret > 0)
                                ret = 0;
                }
@@ -2852,31 +2865,6 @@ static noinline_for_stack int prealloc_file_extent_cluster(
        if (ret)
                return ret;
 
-       /*
-        * On a zoned filesystem, we cannot preallocate the file region.
-        * Instead, we dirty and fiemap_write the region.
-        */
-       if (btrfs_is_zoned(inode->root->fs_info)) {
-               struct btrfs_root *root = inode->root;
-               struct btrfs_trans_handle *trans;
-
-               end = cluster->end - offset + 1;
-               trans = btrfs_start_transaction(root, 1);
-               if (IS_ERR(trans))
-                       return PTR_ERR(trans);
-
-               inode->vfs_inode.i_ctime = current_time(&inode->vfs_inode);
-               i_size_write(&inode->vfs_inode, end);
-               ret = btrfs_update_inode(trans, root, inode);
-               if (ret) {
-                       btrfs_abort_transaction(trans, ret);
-                       btrfs_end_transaction(trans);
-                       return ret;
-               }
-
-               return btrfs_end_transaction(trans);
-       }
-
        btrfs_inode_lock(&inode->vfs_inode, 0);
        for (nr = 0; nr < cluster->nr; nr++) {
                start = cluster->boundary[nr] - offset;
@@ -3084,7 +3072,6 @@ release_page:
 static int relocate_file_extent_cluster(struct inode *inode,
                                        struct file_extent_cluster *cluster)
 {
-       struct btrfs_fs_info *fs_info = btrfs_sb(inode->i_sb);
        u64 offset = BTRFS_I(inode)->index_cnt;
        unsigned long index;
        unsigned long last_index;
@@ -3114,8 +3101,6 @@ static int relocate_file_extent_cluster(struct inode *inode,
        for (index = (cluster->start - offset) >> PAGE_SHIFT;
             index <= last_index && !ret; index++)
                ret = relocate_one_page(inode, ra, cluster, &cluster_nr, index);
-       if (btrfs_is_zoned(fs_info) && !ret)
-               ret = btrfs_wait_ordered_range(inode, 0, (u64)-1);
        if (ret == 0)
                WARN_ON(cluster_nr != cluster->nr);
 out:
@@ -3593,7 +3578,12 @@ int prepare_to_relocate(struct reloc_control *rc)
                 */
                return PTR_ERR(trans);
        }
-       return btrfs_commit_transaction(trans);
+
+       ret = btrfs_commit_transaction(trans);
+       if (ret)
+               unset_reloc_control(rc);
+
+       return ret;
 }
 
 static noinline_for_stack int relocate_block_group(struct reloc_control *rc)
@@ -3770,12 +3760,8 @@ static int __insert_orphan_inode(struct btrfs_trans_handle *trans,
        struct btrfs_path *path;
        struct btrfs_inode_item *item;
        struct extent_buffer *leaf;
-       u64 flags = BTRFS_INODE_NOCOMPRESS | BTRFS_INODE_PREALLOC;
        int ret;
 
-       if (btrfs_is_zoned(trans->fs_info))
-               flags &= ~BTRFS_INODE_PREALLOC;
-
        path = btrfs_alloc_path();
        if (!path)
                return -ENOMEM;
@@ -3790,7 +3776,8 @@ static int __insert_orphan_inode(struct btrfs_trans_handle *trans,
        btrfs_set_inode_generation(leaf, item, 1);
        btrfs_set_inode_size(leaf, item, 0);
        btrfs_set_inode_mode(leaf, item, S_IFREG | 0600);
-       btrfs_set_inode_flags(leaf, item, flags);
+       btrfs_set_inode_flags(leaf, item, BTRFS_INODE_NOCOMPRESS |
+                                         BTRFS_INODE_PREALLOC);
        btrfs_mark_buffer_dirty(leaf);
 out:
        btrfs_free_path(path);
@@ -3885,25 +3872,14 @@ out:
  *   0             success
  *   -EINPROGRESS  operation is already in progress, that's probably a bug
  *   -ECANCELED    cancellation request was set before the operation started
- *   -EAGAIN       can not start because there are ongoing send operations
  */
 static int reloc_chunk_start(struct btrfs_fs_info *fs_info)
 {
-       spin_lock(&fs_info->send_reloc_lock);
-       if (fs_info->send_in_progress) {
-               btrfs_warn_rl(fs_info,
-"cannot run relocation while send operations are in progress (%d in progress)",
-                             fs_info->send_in_progress);
-               spin_unlock(&fs_info->send_reloc_lock);
-               return -EAGAIN;
-       }
        if (test_and_set_bit(BTRFS_FS_RELOC_RUNNING, &fs_info->flags)) {
                /* This should not happen */
-               spin_unlock(&fs_info->send_reloc_lock);
                btrfs_err(fs_info, "reloc already running, cannot start");
                return -EINPROGRESS;
        }
-       spin_unlock(&fs_info->send_reloc_lock);
 
        if (atomic_read(&fs_info->reloc_cancel_req) > 0) {
                btrfs_info(fs_info, "chunk relocation canceled on start");
@@ -3925,9 +3901,7 @@ static void reloc_chunk_end(struct btrfs_fs_info *fs_info)
        /* Requested after start, clear bit first so any waiters can continue */
        if (atomic_read(&fs_info->reloc_cancel_req) > 0)
                btrfs_info(fs_info, "chunk relocation canceled during operation");
-       spin_lock(&fs_info->send_reloc_lock);
        clear_and_wake_up_bit(BTRFS_FS_RELOC_RUNNING, &fs_info->flags);
-       spin_unlock(&fs_info->send_reloc_lock);
        atomic_set(&fs_info->reloc_cancel_req, 0);
 }
 
@@ -3998,6 +3972,19 @@ int btrfs_relocate_block_group(struct btrfs_fs_info *fs_info, u64 group_start)
        int rw = 0;
        int err = 0;
 
+       /*
+        * This only gets set if we had a half-deleted snapshot on mount.  We
+        * cannot allow relocation to start while we're still trying to clean up
+        * these pending deletions.
+        */
+       ret = wait_on_bit(&fs_info->flags, BTRFS_FS_UNFINISHED_DROPS, TASK_INTERRUPTIBLE);
+       if (ret)
+               return ret;
+
+       /* We may have been woken up by close_ctree, so bail if we're closing. */
+       if (btrfs_fs_closing(fs_info))
+               return -EINTR;
+
        bg = btrfs_lookup_block_group(fs_info, group_start);
        if (!bg)
                return -ENOENT;
@@ -4386,8 +4373,7 @@ int btrfs_reloc_cow_block(struct btrfs_trans_handle *trans,
        if (!rc)
                return 0;
 
-       BUG_ON(rc->stage == UPDATE_DATA_PTRS &&
-              root->root_key.objectid == BTRFS_DATA_RELOC_TREE_OBJECTID);
+       BUG_ON(rc->stage == UPDATE_DATA_PTRS && btrfs_is_data_reloc_root(root));
 
        level = btrfs_header_level(buf);
        if (btrfs_header_generation(buf) <=