#define BTRFS_BLOCK_GROUP_SYSTEM (1 << 1)
#define BTRFS_BLOCK_GROUP_METADATA (1 << 2)
#define BTRFS_BLOCK_GROUP_RAID0 (1 << 3)
+#define BTRFS_BLOCK_GROUP_RAID1 (1 << 4)
struct btrfs_block_group_item {
u64 delalloc_bytes;
u64 last_alloc;
u64 last_data_alloc;
- int extra_data_alloc_bits;
- int extra_alloc_bits;
+ int avail_data_alloc_bits;
+ int avail_metadata_alloc_bits;
+ int avail_system_alloc_bits;
+ int data_alloc_profile;
+ int metadata_alloc_profile;
+ int system_alloc_profile;
};
/*
GFP_NOFS);
struct btrfs_root *tree_root = kmalloc(sizeof(struct btrfs_root),
GFP_NOFS);
- struct btrfs_fs_info *fs_info = kmalloc(sizeof(*fs_info),
+ struct btrfs_fs_info *fs_info = kzalloc(sizeof(*fs_info),
GFP_NOFS);
struct btrfs_root *chunk_root = kmalloc(sizeof(struct btrfs_root),
GFP_NOFS);
int ret;
int err = -EINVAL;
struct btrfs_super_block *disk_super;
+
if (!extent_root || !tree_root || !fs_info) {
err = -ENOMEM;
goto fail;
spin_lock_init(&fs_info->delalloc_lock);
spin_lock_init(&fs_info->new_trans_lock);
- memset(&fs_info->super_kobj, 0, sizeof(fs_info->super_kobj));
init_completion(&fs_info->kobj_unregister);
sb_set_blocksize(sb, 4096);
- fs_info->running_transaction = NULL;
- fs_info->last_trans_committed = 0;
fs_info->tree_root = tree_root;
fs_info->extent_root = extent_root;
fs_info->chunk_root = chunk_root;
INIT_LIST_HEAD(&fs_info->space_info);
btrfs_mapping_init(&fs_info->mapping_tree);
fs_info->sb = sb;
- fs_info->throttles = 0;
- fs_info->mount_opt = 0;
fs_info->max_extent = (u64)-1;
fs_info->max_inline = 8192 * 1024;
- fs_info->delalloc_bytes = 0;
setup_bdi(fs_info, &fs_info->bdi);
fs_info->btree_inode = new_inode(sb);
fs_info->btree_inode->i_ino = 1;
extent_io_tree_init(&fs_info->extent_ins,
fs_info->btree_inode->i_mapping, GFP_NOFS);
fs_info->do_barriers = 1;
- fs_info->closing = 0;
- fs_info->total_pinned = 0;
- fs_info->last_alloc = 0;
- fs_info->last_data_alloc = 0;
- fs_info->extra_alloc_bits = 0;
- fs_info->extra_data_alloc_bits = 0;
#if LINUX_VERSION_CODE <= KERNEL_VERSION(2,6,18)
INIT_WORK(&fs_info->trans_work, btrfs_transaction_cleaner, fs_info);
btrfs_read_block_groups(extent_root);
fs_info->generation = btrfs_super_generation(disk_super) + 1;
+ if (btrfs_super_num_devices(disk_super) > 0) {
+ fs_info->data_alloc_profile = BTRFS_BLOCK_GROUP_RAID0;
+ fs_info->metadata_alloc_profile = BTRFS_BLOCK_GROUP_RAID1;
+ fs_info->system_alloc_profile = BTRFS_BLOCK_GROUP_RAID0;
+ }
mutex_unlock(&fs_info->fs_mutex);
return tree_root;
goto new_group;
if (start + num > total_fs_bytes)
goto new_group;
+ if (!block_group_bits(cache, data)) {
+ printk("block group bits don't match %Lu %Lu\n", cache->flags, data);
+ }
*start_ret = start;
return 0;
- } out:
+ }
+out:
cache = btrfs_lookup_block_group(root->fs_info, search_start);
if (!cache) {
printk("Unable to find block group for %Lu\n", search_start);
if (cache->key.objectid > total_fs_bytes)
break;
- if (full_search)
- free_check = cache->key.offset;
- else
- free_check = div_factor(cache->key.offset, factor);
+ if (block_group_bits(cache, data)) {
+ if (full_search)
+ free_check = cache->key.offset;
+ else
+ free_check = div_factor(cache->key.offset,
+ factor);
- if (used + cache->pinned < free_check) {
- found_group = cache;
- goto found;
+ if (used + cache->pinned < free_check) {
+ found_group = cache;
+ goto found;
+ }
}
cond_resched();
}
return 0;
}
+static void set_avail_alloc_bits(struct btrfs_fs_info *fs_info, u64 flags)
+{
+ u64 extra_flags = flags & (BTRFS_BLOCK_GROUP_RAID0 |
+ BTRFS_BLOCK_GROUP_RAID1);
+ if (extra_flags) {
+ if (flags & BTRFS_BLOCK_GROUP_DATA)
+ fs_info->avail_data_alloc_bits |= extra_flags;
+ if (flags & BTRFS_BLOCK_GROUP_METADATA)
+ fs_info->avail_metadata_alloc_bits |= extra_flags;
+ if (flags & BTRFS_BLOCK_GROUP_SYSTEM)
+ fs_info->avail_system_alloc_bits |= extra_flags;
+ }
+}
static int do_chunk_alloc(struct btrfs_trans_handle *trans,
struct btrfs_root *extent_root, u64 alloc_bytes,
if (space_info->full)
return 0;
- thresh = div_factor(space_info->total_bytes, 7);
+ thresh = div_factor(space_info->total_bytes, 6);
if ((space_info->bytes_used + space_info->bytes_pinned + alloc_bytes) <
thresh)
return 0;
start, num_bytes);
BUG_ON(ret);
- if (flags & BTRFS_BLOCK_GROUP_RAID0) {
- if (flags & BTRFS_BLOCK_GROUP_DATA) {
- extent_root->fs_info->extra_data_alloc_bits =
- BTRFS_BLOCK_GROUP_RAID0;
- }
- if (flags & BTRFS_BLOCK_GROUP_METADATA) {
- extent_root->fs_info->extra_alloc_bits =
- BTRFS_BLOCK_GROUP_RAID0;
- }
- }
+ set_avail_alloc_bits(extent_root->fs_info, flags);
return 0;
}
if (data & BTRFS_BLOCK_GROUP_METADATA) {
last_ptr = &root->fs_info->last_alloc;
+ empty_cluster = 256 * 1024;
}
if ((data & BTRFS_BLOCK_GROUP_DATA) && btrfs_test_opt(root, SSD)) {
u64 root_used;
u64 search_start = 0;
u64 new_hint;
+ u64 alloc_profile;
u32 sizes[2];
struct btrfs_fs_info *info = root->fs_info;
struct btrfs_root *extent_root = info->extent_root;
struct btrfs_extent_ref *ref;
struct btrfs_path *path;
struct btrfs_key keys[2];
- int extra_chunk_alloc_bits = 0;
if (data) {
- data = BTRFS_BLOCK_GROUP_DATA | info->extra_data_alloc_bits;
+ alloc_profile = info->avail_data_alloc_bits &
+ info->data_alloc_profile;
+ data = BTRFS_BLOCK_GROUP_DATA | alloc_profile;
} else if (root == root->fs_info->chunk_root) {
- data = BTRFS_BLOCK_GROUP_SYSTEM;
+ alloc_profile = info->avail_system_alloc_bits &
+ info->system_alloc_profile;
+ data = BTRFS_BLOCK_GROUP_SYSTEM | alloc_profile;
} else {
- data = BTRFS_BLOCK_GROUP_METADATA | info->extra_alloc_bits;
+ alloc_profile = info->avail_metadata_alloc_bits &
+ info->metadata_alloc_profile;
+ data = BTRFS_BLOCK_GROUP_METADATA | alloc_profile;
}
- if (btrfs_super_num_devices(&info->super_copy) > 1 &&
- !(data & BTRFS_BLOCK_GROUP_SYSTEM))
- extra_chunk_alloc_bits = BTRFS_BLOCK_GROUP_RAID0;
if (root->ref_cows) {
if (!(data & BTRFS_BLOCK_GROUP_METADATA)) {
ret = do_chunk_alloc(trans, root->fs_info->extent_root,
2 * 1024 * 1024,
BTRFS_BLOCK_GROUP_METADATA |
- info->extra_alloc_bits |
- extra_chunk_alloc_bits);
+ (info->metadata_alloc_profile &
+ info->avail_metadata_alloc_bits));
BUG_ON(ret);
}
ret = do_chunk_alloc(trans, root->fs_info->extent_root,
- num_bytes + 2 * 1024 * 1024, data |
- extra_chunk_alloc_bits);
+ num_bytes + 2 * 1024 * 1024, data);
BUG_ON(ret);
}
if (!next || !btrfs_buffer_uptodate(next)) {
free_extent_buffer(next);
reada_walk_down(root, cur, path->slots[*level]);
+
+ mutex_unlock(&root->fs_info->fs_mutex);
next = read_tree_block(root, bytenr, blocksize);
+ mutex_lock(&root->fs_info->fs_mutex);
- /* we used to drop the lock above, keep the
- * code to double check so that we won't forget
- * when we drop the lock again in the future
- */
+ /* we've dropped the lock, double check */
ret = lookup_extent_ref(trans, root, bytenr,
blocksize, &refs);
BUG_ON(ret);
} else if (cache->flags & BTRFS_BLOCK_GROUP_METADATA) {
bit = BLOCK_GROUP_METADATA;
}
- if (cache->flags & BTRFS_BLOCK_GROUP_RAID0) {
- if (cache->flags & BTRFS_BLOCK_GROUP_DATA) {
- info->extra_data_alloc_bits =
- BTRFS_BLOCK_GROUP_RAID0;
- }
- if (cache->flags & BTRFS_BLOCK_GROUP_METADATA) {
- info->extra_alloc_bits =
- BTRFS_BLOCK_GROUP_RAID0;
- }
- }
+ set_avail_alloc_bits(info, cache->flags);
ret = update_space_info(info, cache->flags, found_key.offset,
btrfs_block_group_used(&cache->item),
u64 physical;
u64 length = 0;
u64 map_length;
+ int total_devs;
struct bio_vec *bvec;
int i;
int ret;
}
map_tree = &root->fs_info->mapping_tree;
map_length = length;
- ret = btrfs_map_block(map_tree, logical, &physical, &map_length, &dev);
+ ret = btrfs_map_block(map_tree, READ, 0, logical, &physical,
+ &map_length, &dev, &total_devs);
if (map_length < length + size) {
return 1;
}
u64 physical;
};
+struct multi_bio {
+ atomic_t stripes;
+ bio_end_io_t *end_io;
+ void *private;
+ int error;
+};
+
struct map_lookup {
u64 type;
int io_align;
if (list_empty(dev_list))
return -ENOSPC;
- if (type & BTRFS_BLOCK_GROUP_RAID0)
+ if (type & (BTRFS_BLOCK_GROUP_RAID0))
num_stripes = btrfs_super_num_devices(&info->super_copy);
- if (type & BTRFS_BLOCK_GROUP_DATA)
- stripe_len = 64 * 1024;
- if (type & (BTRFS_BLOCK_GROUP_METADATA | BTRFS_BLOCK_GROUP_SYSTEM))
- stripe_len = 32 * 1024;
+ if (type & (BTRFS_BLOCK_GROUP_RAID1)) {
+ num_stripes = min_t(u64, 2,
+ btrfs_super_num_devices(&info->super_copy));
+ }
again:
INIT_LIST_HEAD(&private_devs);
cur = dev_list->next;
stripes = &chunk->stripe;
- *num_bytes = calc_size * num_stripes;
+ if (type & BTRFS_BLOCK_GROUP_RAID1)
+ *num_bytes = calc_size;
+ else
+ *num_bytes = calc_size * num_stripes;
+
index = 0;
while(index < num_stripes) {
BUG_ON(list_empty(&private_devs));
key.objectid,
calc_size, &dev_offset);
BUG_ON(ret);
-printk("alloc chunk size %Lu from dev %Lu\n", calc_size, device->devid);
+printk("alloc chunk start %Lu size %Lu from dev %Lu type %Lu\n", key.objectid, calc_size, device->devid, type);
device->bytes_used += calc_size;
ret = btrfs_update_device(trans, device);
BUG_ON(ret);
}
}
-int btrfs_map_block(struct btrfs_mapping_tree *map_tree,
- u64 logical, u64 *phys, u64 *length,
- struct btrfs_device **dev)
+int btrfs_map_block(struct btrfs_mapping_tree *map_tree, int rw,
+ int dev_nr, u64 logical, u64 *phys, u64 *length,
+ struct btrfs_device **dev, int *total_devs)
{
struct extent_map *em;
struct map_lookup *map;
/* stripe_offset is the offset of this block in its stripe*/
stripe_offset = offset - stripe_offset;
- /*
- * after this do_div call, stripe_nr is the number of stripes
- * on this device we have to walk to find the data, and
- * stripe_index is the number of our device in the stripe array
- */
- stripe_index = do_div(stripe_nr, map->num_stripes);
-
+ if (map->type & BTRFS_BLOCK_GROUP_RAID1) {
+ stripe_index = dev_nr;
+ if (rw & (1 << BIO_RW))
+ *total_devs = map->num_stripes;
+ else {
+ int i;
+ u64 least = (u64)-1;
+ struct btrfs_device *cur;
+
+ for (i = 0; i < map->num_stripes; i++) {
+ cur = map->stripes[i].dev;
+ spin_lock(&cur->io_lock);
+ if (cur->total_ios < least) {
+ least = cur->total_ios;
+ stripe_index = i;
+ }
+ spin_unlock(&cur->io_lock);
+ }
+ *total_devs = 1;
+ }
+ } else {
+ /*
+ * after this do_div call, stripe_nr is the number of stripes
+ * on this device we have to walk to find the data, and
+ * stripe_index is the number of our device in the stripe array
+ */
+ stripe_index = do_div(stripe_nr, map->num_stripes);
+ }
BUG_ON(stripe_index >= map->num_stripes);
-
*phys = map->stripes[stripe_index].physical + stripe_offset +
stripe_nr * map->stripe_len;
- if (map->type & BTRFS_BLOCK_GROUP_RAID0) {
+ if (map->type & (BTRFS_BLOCK_GROUP_RAID0 | BTRFS_BLOCK_GROUP_RAID1)) {
/* we limit the length of each bio to what fits in a stripe */
*length = min_t(u64, em->len - offset,
map->stripe_len - stripe_offset);
return 0;
}
+#if LINUX_VERSION_CODE > KERNEL_VERSION(2,6,23)
+static void end_bio_multi_stripe(struct bio *bio, int err)
+#else
+static int end_bio_multi_stripe(struct bio *bio,
+ unsigned int bytes_done, int err)
+#endif
+{
+ struct multi_bio *multi = bio->bi_private;
+
+#if LINUX_VERSION_CODE <= KERNEL_VERSION(2,6,23)
+ if (bio->bi_size)
+ return 1;
+#endif
+ if (err)
+ multi->error = err;
+
+ if (atomic_dec_and_test(&multi->stripes)) {
+ bio->bi_private = multi->private;
+ bio->bi_end_io = multi->end_io;
+
+ if (!err && multi->error)
+ err = multi->error;
+ kfree(multi);
+
+ bio_endio(bio, err);
+ } else {
+ bio_put(bio);
+ }
+#if LINUX_VERSION_CODE <= KERNEL_VERSION(2,6,23)
+ return 0;
+#endif
+}
+
int btrfs_map_bio(struct btrfs_root *root, int rw, struct bio *bio)
{
struct btrfs_mapping_tree *map_tree;
struct btrfs_device *dev;
+ struct bio *first_bio = bio;
u64 logical = bio->bi_sector << 9;
u64 physical;
u64 length = 0;
u64 map_length;
struct bio_vec *bvec;
+ struct multi_bio *multi = NULL;
int i;
int ret;
+ int dev_nr = 0;
+ int total_devs = 1;
bio_for_each_segment(bvec, bio, i) {
length += bvec->bv_len;
}
+
map_tree = &root->fs_info->mapping_tree;
map_length = length;
- ret = btrfs_map_block(map_tree, logical, &physical, &map_length, &dev);
- if (map_length < length) {
- printk("mapping failed logical %Lu bio len %Lu physical %Lu "
- "len %Lu\n", logical, length, physical, map_length);
- BUG();
+ while(dev_nr < total_devs) {
+ ret = btrfs_map_block(map_tree, rw, dev_nr, logical,
+ &physical, &map_length, &dev,
+ &total_devs);
+ if (map_length < length) {
+ printk("mapping failed logical %Lu bio len %Lu physical %Lu "
+ "len %Lu\n", logical, length, physical, map_length);
+ BUG();
+ }
+ BUG_ON(map_length < length);
+ if (total_devs > 1) {
+ if (!multi) {
+ multi = kmalloc(sizeof(*multi), GFP_NOFS);
+ atomic_set(&multi->stripes, 1);
+ multi->end_io = bio->bi_end_io;
+ multi->private = first_bio->bi_private;
+ multi->error = 0;
+ } else {
+ atomic_inc(&multi->stripes);
+ }
+ if (dev_nr < total_devs - 1) {
+ bio = bio_clone(first_bio, GFP_NOFS);
+ BUG_ON(!bio);
+ } else {
+ bio = first_bio;
+ }
+ bio->bi_private = multi;
+ bio->bi_end_io = end_bio_multi_stripe;
+ }
+ bio->bi_sector = physical >> 9;
+ bio->bi_bdev = dev->bdev;
+ spin_lock(&dev->io_lock);
+ dev->total_ios++;
+ spin_unlock(&dev->io_lock);
+ submit_bio(rw, bio);
+ dev_nr++;
}
- BUG_ON(map_length < length);
- bio->bi_sector = physical >> 9;
- bio->bi_bdev = dev->bdev;
- submit_bio(rw, bio);
return 0;
}
return -ENOMEM;
list_add(&device->dev_list,
&root->fs_info->fs_devices->devices);
+ device->total_ios = 0;
+ spin_lock_init(&device->io_lock);
}
fill_device_from_item(leaf, dev_item, device);
#ifndef __BTRFS_VOLUMES_
#define __BTRFS_VOLUMES_
+
struct btrfs_device {
struct list_head dev_list;
struct btrfs_root *dev_root;
+ spinlock_t io_lock;
struct block_device *bdev;
+ u64 total_ios;
+
char *name;
/* the internal btrfs device id */
int btrfs_alloc_dev_extent(struct btrfs_trans_handle *trans,
struct btrfs_device *device,
u64 owner, u64 num_bytes, u64 *start);
-int btrfs_map_block(struct btrfs_mapping_tree *map_tree,
+int btrfs_map_block(struct btrfs_mapping_tree *map_tree, int rw, int stripe_nr,
u64 logical, u64 *phys, u64 *length,
- struct btrfs_device **dev);
+ struct btrfs_device **dev, int *total_stripes);
int btrfs_read_sys_array(struct btrfs_root *root);
int btrfs_read_chunk_tree(struct btrfs_root *root);
int btrfs_alloc_chunk(struct btrfs_trans_handle *trans,
void btrfs_mapping_tree_free(struct btrfs_mapping_tree *tree);
int btrfs_map_bio(struct btrfs_root *root, int rw, struct bio *bio);
int btrfs_read_super_device(struct btrfs_root *root, struct extent_buffer *buf);
-int btrfs_map_block(struct btrfs_mapping_tree *map_tree,
- u64 logical, u64 *phys, u64 *length,
- struct btrfs_device **dev);
int btrfs_open_devices(struct btrfs_fs_devices *fs_devices,
int flags, void *holder);
int btrfs_scan_one_device(const char *path, int flags, void *holder,