diff options
author | Josef Bacik <josef@toxicpanda.com> | 2020-07-21 10:48:46 -0400 |
---|---|---|
committer | David Sterba <dsterba@suse.com> | 2020-07-27 12:55:47 +0200 |
commit | 162e0a16b7d93629c8325ea32d99b38d17bc03eb (patch) | |
tree | 5e637f16ec455e25ae80f18b2d7dd22f323ee5ae /fs | |
parent | 349e120ecebeb984376c8edb89bf0bfb85bc16f7 (diff) | |
download | lwn-162e0a16b7d93629c8325ea32d99b38d17bc03eb.tar.gz lwn-162e0a16b7d93629c8325ea32d99b38d17bc03eb.zip |
btrfs: if we're restriping, use the target restripe profile
Previously we depended on some weird behavior in our chunk allocator to
force the allocation of new stripes, so by the time we got to doing the
reduce we would usually already have a chunk with the proper target.
However that behavior causes other problems and needs to be removed.
First however we need to remove this check to only restripe if we
already have those available profiles, because if we're allocating our
first chunk it obviously will not be available. Simply use the target
as specified, and if that fails it'll be because we're out of space.
Tested-by: Holger Hoffstätte <holger@applied-asynchrony.com>
Signed-off-by: Josef Bacik <josef@toxicpanda.com>
Signed-off-by: David Sterba <dsterba@suse.com>
Diffstat (limited to 'fs')
-rw-r--r-- | fs/btrfs/block-group.c | 7 |
1 files changed, 2 insertions, 5 deletions
diff --git a/fs/btrfs/block-group.c b/fs/btrfs/block-group.c index 652b35d5a773..613920c17ac1 100644 --- a/fs/btrfs/block-group.c +++ b/fs/btrfs/block-group.c @@ -65,11 +65,8 @@ static u64 btrfs_reduce_alloc_profile(struct btrfs_fs_info *fs_info, u64 flags) spin_lock(&fs_info->balance_lock); target = get_restripe_target(fs_info, flags); if (target) { - /* Pick target profile only if it's already available */ - if ((flags & target) & BTRFS_EXTENDED_PROFILE_MASK) { - spin_unlock(&fs_info->balance_lock); - return extended_to_chunk(target); - } + spin_unlock(&fs_info->balance_lock); + return extended_to_chunk(target); } spin_unlock(&fs_info->balance_lock); |