* will return -ENOSPC.
*/
static int may_commit_transaction(struct btrfs_fs_info *fs_info,
- struct btrfs_space_info *space_info,
- u64 bytes_needed)
+ struct btrfs_space_info *space_info)
{
struct reserve_ticket *ticket = NULL;
struct btrfs_block_rsv *delayed_rsv = &fs_info->delayed_block_rsv;
struct btrfs_block_rsv *trans_rsv = &fs_info->trans_block_rsv;
struct btrfs_trans_handle *trans;
u64 reclaim_bytes = 0;
+ u64 bytes_needed = 0;
u64 cur_free_bytes = 0;
- bool do_commit = false;
trans = (struct btrfs_trans_handle *)current->journal_info;
if (trans)
return -EAGAIN;
- /*
- * If we are data and have passed in U64_MAX we just want to
- * unconditionally commit the transaction to match the previous data
- * flushing behavior.
- */
- if ((space_info->flags & BTRFS_BLOCK_GROUP_DATA) &&
- bytes_needed == U64_MAX) {
- do_commit = true;
- goto check_pinned;
- }
-
spin_lock(&space_info->lock);
cur_free_bytes = btrfs_space_info_used(space_info, true);
if (cur_free_bytes < space_info->total_bytes)
if (!bytes_needed)
return 0;
-check_pinned:
trans = btrfs_join_transaction(fs_info->extent_root);
if (IS_ERR(trans))
return PTR_ERR(trans);
* we have block groups that are going to be freed, allowing us to
* possibly do a chunk allocation the next loop through.
*/
- if (do_commit ||
- test_bit(BTRFS_TRANS_HAVE_FREE_BGS, &trans->transaction->flags) ||
+ if (test_bit(BTRFS_TRANS_HAVE_FREE_BGS, &trans->transaction->flags) ||
__percpu_counter_compare(&space_info->total_bytes_pinned,
bytes_needed,
BTRFS_TOTAL_BYTES_PINNED_BATCH) >= 0)
btrfs_wait_on_delayed_iputs(fs_info);
break;
case COMMIT_TRANS:
- ret = may_commit_transaction(fs_info, space_info, num_bytes);
+ ret = may_commit_transaction(fs_info, space_info);
break;
default:
ret = -ENOSPC;