summaryrefslogtreecommitdiff
path: root/fs/btrfs/extent-tree.c
diff options
context:
space:
mode:
authorChris Mason <chris.mason@oracle.com>2008-06-25 16:01:30 -0400
committerChris Mason <chris.mason@oracle.com>2008-09-25 11:04:03 -0400
commit333db94cdde9e6dfdedab9290d04d812f83e0922 (patch)
treef4925bc36a8f4177d5a4f3a303ae47f3ea592b60 /fs/btrfs/extent-tree.c
parent5cd57b2cbbb06a350df2698314e4e6a80805fc2f (diff)
downloadlwn-333db94cdde9e6dfdedab9290d04d812f83e0922.tar.gz
lwn-333db94cdde9e6dfdedab9290d04d812f83e0922.zip
Btrfs: Fix snapshot deletion to release the alloc_mutex much more often.
This lowers the impact of snapshot deletion on the rest of the FS. Signed-off-by: Chris Mason <chris.mason@oracle.com>
Diffstat (limited to 'fs/btrfs/extent-tree.c')
-rw-r--r--fs/btrfs/extent-tree.c28
1 files changed, 20 insertions, 8 deletions
diff --git a/fs/btrfs/extent-tree.c b/fs/btrfs/extent-tree.c
index 544fc3f2fe6c..6274f30031db 100644
--- a/fs/btrfs/extent-tree.c
+++ b/fs/btrfs/extent-tree.c
@@ -1223,8 +1223,8 @@ printk("space info full %Lu\n", flags);
ret = btrfs_make_block_group(trans, extent_root, 0, flags,
BTRFS_FIRST_CHUNK_TREE_OBJECTID, start, num_bytes);
BUG_ON(ret);
- mutex_unlock(&extent_root->fs_info->chunk_mutex);
out:
+ mutex_unlock(&extent_root->fs_info->chunk_mutex);
return 0;
}
@@ -2181,18 +2181,30 @@ static void noinline reada_walk_down(struct btrfs_root *root,
continue;
}
}
- mutex_unlock(&root->fs_info->alloc_mutex);
ret = readahead_tree_block(root, bytenr, blocksize,
btrfs_node_ptr_generation(node, i));
last = bytenr + blocksize;
cond_resched();
- mutex_lock(&root->fs_info->alloc_mutex);
if (ret)
break;
}
}
/*
+ * we want to avoid as much random IO as we can with the alloc mutex
+ * held, so drop the lock and do the lookup, then do it again with the
+ * lock held.
+ */
+int drop_snap_lookup_refcount(struct btrfs_root *root, u64 start, u64 len,
+ u32 *refs)
+{
+ mutex_unlock(&root->fs_info->alloc_mutex);
+ lookup_extent_ref(NULL, root, start, len, refs);
+ mutex_lock(&root->fs_info->alloc_mutex);
+ return lookup_extent_ref(NULL, root, start, len, refs);
+}
+
+/*
* helper function for drop_snapshot, this walks down the tree dropping ref
* counts as it goes.
*/
@@ -2215,8 +2227,7 @@ static int noinline walk_down_tree(struct btrfs_trans_handle *trans,
WARN_ON(*level < 0);
WARN_ON(*level >= BTRFS_MAX_LEVEL);
- ret = lookup_extent_ref(trans, root,
- path->nodes[*level]->start,
+ ret = drop_snap_lookup_refcount(root, path->nodes[*level]->start,
path->nodes[*level]->len, &refs);
BUG_ON(ret);
if (refs > 1)
@@ -2245,7 +2256,7 @@ static int noinline walk_down_tree(struct btrfs_trans_handle *trans,
ptr_gen = btrfs_node_ptr_generation(cur, path->slots[*level]);
blocksize = btrfs_level_size(root, *level - 1);
- ret = lookup_extent_ref(trans, root, bytenr, blocksize, &refs);
+ ret = drop_snap_lookup_refcount(root, bytenr, blocksize, &refs);
BUG_ON(ret);
if (refs != 1) {
parent = path->nodes[*level];
@@ -2261,15 +2272,16 @@ static int noinline walk_down_tree(struct btrfs_trans_handle *trans,
next = btrfs_find_tree_block(root, bytenr, blocksize);
if (!next || !btrfs_buffer_uptodate(next, ptr_gen)) {
free_extent_buffer(next);
+ mutex_unlock(&root->fs_info->alloc_mutex);
+
reada_walk_down(root, cur, path->slots[*level]);
- mutex_unlock(&root->fs_info->alloc_mutex);
next = read_tree_block(root, bytenr, blocksize,
ptr_gen);
mutex_lock(&root->fs_info->alloc_mutex);
/* we've dropped the lock, double check */
- ret = lookup_extent_ref(trans, root, bytenr,
+ ret = drop_snap_lookup_refcount(root, bytenr,
blocksize, &refs);
BUG_ON(ret);
if (refs != 1) {