Commit fb38f9b8 authored by Linus Torvalds's avatar Linus Torvalds

Merge branch 'for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/mason/linux-btrfs

* 'for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/mason/linux-btrfs:
  Btrfs: drop spin lock when memory alloc fails
  Btrfs: check if the to-be-added device is writable
  Btrfs: try cluster but don't advance in search list
  Btrfs: try to allocate from cluster even at LOOP_NO_EMPTY_SIZE
parents 8bd1c881 1cf4ffdb
...@@ -5107,11 +5107,11 @@ static noinline int find_free_extent(struct btrfs_trans_handle *trans, ...@@ -5107,11 +5107,11 @@ static noinline int find_free_extent(struct btrfs_trans_handle *trans,
struct btrfs_root *root = orig_root->fs_info->extent_root; struct btrfs_root *root = orig_root->fs_info->extent_root;
struct btrfs_free_cluster *last_ptr = NULL; struct btrfs_free_cluster *last_ptr = NULL;
struct btrfs_block_group_cache *block_group = NULL; struct btrfs_block_group_cache *block_group = NULL;
struct btrfs_block_group_cache *used_block_group;
int empty_cluster = 2 * 1024 * 1024; int empty_cluster = 2 * 1024 * 1024;
int allowed_chunk_alloc = 0; int allowed_chunk_alloc = 0;
int done_chunk_alloc = 0; int done_chunk_alloc = 0;
struct btrfs_space_info *space_info; struct btrfs_space_info *space_info;
int last_ptr_loop = 0;
int loop = 0; int loop = 0;
int index = 0; int index = 0;
int alloc_type = (data & BTRFS_BLOCK_GROUP_DATA) ? int alloc_type = (data & BTRFS_BLOCK_GROUP_DATA) ?
...@@ -5173,6 +5173,7 @@ static noinline int find_free_extent(struct btrfs_trans_handle *trans, ...@@ -5173,6 +5173,7 @@ static noinline int find_free_extent(struct btrfs_trans_handle *trans,
ideal_cache: ideal_cache:
block_group = btrfs_lookup_block_group(root->fs_info, block_group = btrfs_lookup_block_group(root->fs_info,
search_start); search_start);
used_block_group = block_group;
/* /*
* we don't want to use the block group if it doesn't match our * we don't want to use the block group if it doesn't match our
* allocation bits, or if its not cached. * allocation bits, or if its not cached.
...@@ -5210,6 +5211,7 @@ static noinline int find_free_extent(struct btrfs_trans_handle *trans, ...@@ -5210,6 +5211,7 @@ static noinline int find_free_extent(struct btrfs_trans_handle *trans,
u64 offset; u64 offset;
int cached; int cached;
used_block_group = block_group;
btrfs_get_block_group(block_group); btrfs_get_block_group(block_group);
search_start = block_group->key.objectid; search_start = block_group->key.objectid;
...@@ -5286,71 +5288,62 @@ static noinline int find_free_extent(struct btrfs_trans_handle *trans, ...@@ -5286,71 +5288,62 @@ static noinline int find_free_extent(struct btrfs_trans_handle *trans,
spin_unlock(&block_group->free_space_ctl->tree_lock); spin_unlock(&block_group->free_space_ctl->tree_lock);
/* /*
* Ok we want to try and use the cluster allocator, so lets look * Ok we want to try and use the cluster allocator, so
* there, unless we are on LOOP_NO_EMPTY_SIZE, since we will * lets look there
* have tried the cluster allocator plenty of times at this */
* point and not have found anything, so we are likely way too if (last_ptr) {
* fragmented for the clustering stuff to find anything, so lets
* just skip it and let the allocator find whatever block it can
* find
*/
if (last_ptr && loop < LOOP_NO_EMPTY_SIZE) {
/* /*
* the refill lock keeps out other * the refill lock keeps out other
* people trying to start a new cluster * people trying to start a new cluster
*/ */
spin_lock(&last_ptr->refill_lock); spin_lock(&last_ptr->refill_lock);
if (!last_ptr->block_group || used_block_group = last_ptr->block_group;
last_ptr->block_group->ro || if (used_block_group != block_group &&
!block_group_bits(last_ptr->block_group, data)) (!used_block_group ||
used_block_group->ro ||
!block_group_bits(used_block_group, data))) {
used_block_group = block_group;
goto refill_cluster; goto refill_cluster;
}
offset = btrfs_alloc_from_cluster(block_group, last_ptr, if (used_block_group != block_group)
num_bytes, search_start); btrfs_get_block_group(used_block_group);
offset = btrfs_alloc_from_cluster(used_block_group,
last_ptr, num_bytes, used_block_group->key.objectid);
if (offset) { if (offset) {
/* we have a block, we're done */ /* we have a block, we're done */
spin_unlock(&last_ptr->refill_lock); spin_unlock(&last_ptr->refill_lock);
goto checks; goto checks;
} }
spin_lock(&last_ptr->lock); WARN_ON(last_ptr->block_group != used_block_group);
/* if (used_block_group != block_group) {
* whoops, this cluster doesn't actually point to btrfs_put_block_group(used_block_group);
* this block group. Get a ref on the block used_block_group = block_group;
* group is does point to and try again
*/
if (!last_ptr_loop && last_ptr->block_group &&
last_ptr->block_group != block_group &&
index <=
get_block_group_index(last_ptr->block_group)) {
btrfs_put_block_group(block_group);
block_group = last_ptr->block_group;
btrfs_get_block_group(block_group);
spin_unlock(&last_ptr->lock);
spin_unlock(&last_ptr->refill_lock);
last_ptr_loop = 1;
search_start = block_group->key.objectid;
/*
* we know this block group is properly
* in the list because
* btrfs_remove_block_group, drops the
* cluster before it removes the block
* group from the list
*/
goto have_block_group;
} }
spin_unlock(&last_ptr->lock);
refill_cluster: refill_cluster:
BUG_ON(used_block_group != block_group);
/* If we are on LOOP_NO_EMPTY_SIZE, we can't
* set up a new clusters, so lets just skip it
* and let the allocator find whatever block
* it can find. If we reach this point, we
* will have tried the cluster allocator
* plenty of times and not have found
* anything, so we are likely way too
* fragmented for the clustering stuff to find
* anything. */
if (loop >= LOOP_NO_EMPTY_SIZE) {
spin_unlock(&last_ptr->refill_lock);
goto unclustered_alloc;
}
/* /*
* this cluster didn't work out, free it and * this cluster didn't work out, free it and
* start over * start over
*/ */
btrfs_return_cluster_to_free_space(NULL, last_ptr); btrfs_return_cluster_to_free_space(NULL, last_ptr);
last_ptr_loop = 0;
/* allocate a cluster in this block group */ /* allocate a cluster in this block group */
ret = btrfs_find_space_cluster(trans, root, ret = btrfs_find_space_cluster(trans, root,
block_group, last_ptr, block_group, last_ptr,
...@@ -5390,6 +5383,7 @@ static noinline int find_free_extent(struct btrfs_trans_handle *trans, ...@@ -5390,6 +5383,7 @@ static noinline int find_free_extent(struct btrfs_trans_handle *trans,
goto loop; goto loop;
} }
unclustered_alloc:
offset = btrfs_find_space_for_alloc(block_group, search_start, offset = btrfs_find_space_for_alloc(block_group, search_start,
num_bytes, empty_size); num_bytes, empty_size);
/* /*
...@@ -5416,14 +5410,14 @@ static noinline int find_free_extent(struct btrfs_trans_handle *trans, ...@@ -5416,14 +5410,14 @@ static noinline int find_free_extent(struct btrfs_trans_handle *trans,
search_start = stripe_align(root, offset); search_start = stripe_align(root, offset);
/* move on to the next group */ /* move on to the next group */
if (search_start + num_bytes >= search_end) { if (search_start + num_bytes >= search_end) {
btrfs_add_free_space(block_group, offset, num_bytes); btrfs_add_free_space(used_block_group, offset, num_bytes);
goto loop; goto loop;
} }
/* move on to the next group */ /* move on to the next group */
if (search_start + num_bytes > if (search_start + num_bytes >
block_group->key.objectid + block_group->key.offset) { used_block_group->key.objectid + used_block_group->key.offset) {
btrfs_add_free_space(block_group, offset, num_bytes); btrfs_add_free_space(used_block_group, offset, num_bytes);
goto loop; goto loop;
} }
...@@ -5431,14 +5425,14 @@ static noinline int find_free_extent(struct btrfs_trans_handle *trans, ...@@ -5431,14 +5425,14 @@ static noinline int find_free_extent(struct btrfs_trans_handle *trans,
ins->offset = num_bytes; ins->offset = num_bytes;
if (offset < search_start) if (offset < search_start)
btrfs_add_free_space(block_group, offset, btrfs_add_free_space(used_block_group, offset,
search_start - offset); search_start - offset);
BUG_ON(offset > search_start); BUG_ON(offset > search_start);
ret = btrfs_update_reserved_bytes(block_group, num_bytes, ret = btrfs_update_reserved_bytes(used_block_group, num_bytes,
alloc_type); alloc_type);
if (ret == -EAGAIN) { if (ret == -EAGAIN) {
btrfs_add_free_space(block_group, offset, num_bytes); btrfs_add_free_space(used_block_group, offset, num_bytes);
goto loop; goto loop;
} }
...@@ -5447,15 +5441,19 @@ static noinline int find_free_extent(struct btrfs_trans_handle *trans, ...@@ -5447,15 +5441,19 @@ static noinline int find_free_extent(struct btrfs_trans_handle *trans,
ins->offset = num_bytes; ins->offset = num_bytes;
if (offset < search_start) if (offset < search_start)
btrfs_add_free_space(block_group, offset, btrfs_add_free_space(used_block_group, offset,
search_start - offset); search_start - offset);
BUG_ON(offset > search_start); BUG_ON(offset > search_start);
if (used_block_group != block_group)
btrfs_put_block_group(used_block_group);
btrfs_put_block_group(block_group); btrfs_put_block_group(block_group);
break; break;
loop: loop:
failed_cluster_refill = false; failed_cluster_refill = false;
failed_alloc = false; failed_alloc = false;
BUG_ON(index != get_block_group_index(block_group)); BUG_ON(index != get_block_group_index(block_group));
if (used_block_group != block_group)
btrfs_put_block_group(used_block_group);
btrfs_put_block_group(block_group); btrfs_put_block_group(block_group);
} }
up_read(&space_info->groups_sem); up_read(&space_info->groups_sem);
......
...@@ -935,8 +935,10 @@ int convert_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, ...@@ -935,8 +935,10 @@ int convert_extent_bit(struct extent_io_tree *tree, u64 start, u64 end,
node = tree_search(tree, start); node = tree_search(tree, start);
if (!node) { if (!node) {
prealloc = alloc_extent_state_atomic(prealloc); prealloc = alloc_extent_state_atomic(prealloc);
if (!prealloc) if (!prealloc) {
return -ENOMEM; err = -ENOMEM;
goto out;
}
err = insert_state(tree, prealloc, start, end, &bits); err = insert_state(tree, prealloc, start, end, &bits);
prealloc = NULL; prealloc = NULL;
BUG_ON(err == -EEXIST); BUG_ON(err == -EEXIST);
...@@ -992,8 +994,10 @@ int convert_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, ...@@ -992,8 +994,10 @@ int convert_extent_bit(struct extent_io_tree *tree, u64 start, u64 end,
*/ */
if (state->start < start) { if (state->start < start) {
prealloc = alloc_extent_state_atomic(prealloc); prealloc = alloc_extent_state_atomic(prealloc);
if (!prealloc) if (!prealloc) {
return -ENOMEM; err = -ENOMEM;
goto out;
}
err = split_state(tree, state, prealloc, start); err = split_state(tree, state, prealloc, start);
BUG_ON(err == -EEXIST); BUG_ON(err == -EEXIST);
prealloc = NULL; prealloc = NULL;
...@@ -1024,8 +1028,10 @@ int convert_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, ...@@ -1024,8 +1028,10 @@ int convert_extent_bit(struct extent_io_tree *tree, u64 start, u64 end,
this_end = last_start - 1; this_end = last_start - 1;
prealloc = alloc_extent_state_atomic(prealloc); prealloc = alloc_extent_state_atomic(prealloc);
if (!prealloc) if (!prealloc) {
return -ENOMEM; err = -ENOMEM;
goto out;
}
/* /*
* Avoid to free 'prealloc' if it can be merged with * Avoid to free 'prealloc' if it can be merged with
...@@ -1051,8 +1057,10 @@ int convert_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, ...@@ -1051,8 +1057,10 @@ int convert_extent_bit(struct extent_io_tree *tree, u64 start, u64 end,
*/ */
if (state->start <= end && state->end > end) { if (state->start <= end && state->end > end) {
prealloc = alloc_extent_state_atomic(prealloc); prealloc = alloc_extent_state_atomic(prealloc);
if (!prealloc) if (!prealloc) {
return -ENOMEM; err = -ENOMEM;
goto out;
}
err = split_state(tree, state, prealloc, end + 1); err = split_state(tree, state, prealloc, end + 1);
BUG_ON(err == -EEXIST); BUG_ON(err == -EEXIST);
......
...@@ -1611,7 +1611,7 @@ int btrfs_init_new_device(struct btrfs_root *root, char *device_path) ...@@ -1611,7 +1611,7 @@ int btrfs_init_new_device(struct btrfs_root *root, char *device_path)
if ((sb->s_flags & MS_RDONLY) && !root->fs_info->fs_devices->seeding) if ((sb->s_flags & MS_RDONLY) && !root->fs_info->fs_devices->seeding)
return -EINVAL; return -EINVAL;
bdev = blkdev_get_by_path(device_path, FMODE_EXCL, bdev = blkdev_get_by_path(device_path, FMODE_WRITE | FMODE_EXCL,
root->fs_info->bdev_holder); root->fs_info->bdev_holder);
if (IS_ERR(bdev)) if (IS_ERR(bdev))
return PTR_ERR(bdev); return PTR_ERR(bdev);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment