From: Josef Bacik <josef@toxicpanda.com>
To: Naohiro Aota <naohiro.aota@wdc.com>,
linux-btrfs@vger.kernel.org, David Sterba <dsterba@suse.com>
Cc: Chris Mason <clm@fb.com>, Nikolay Borisov <nborisov@suse.com>,
Damien Le Moal <damien.lemoal@wdc.com>,
Johannes Thumshirn <Johannes.Thumshirn@wdc.com>,
Hannes Reinecke <hare@suse.com>,
Anand Jain <anand.jain@oracle.com>,
linux-fsdevel@vger.kernel.org
Subject: Re: [PATCH 04/20] btrfs: introduce alloc_chunk_ctl
Date: Thu, 6 Feb 2020 11:38:14 -0500 [thread overview]
Message-ID: <bcccefe4-8cff-d50d-ddd1-784e3d194607@toxicpanda.com> (raw)
In-Reply-To: <20200206104214.400857-5-naohiro.aota@wdc.com>
On 2/6/20 5:41 AM, Naohiro Aota wrote:
> Introduce "struct alloc_chunk_ctl" to wrap needed parameters for the chunk
> allocation. This will be used to split __btrfs_alloc_chunk() into smaller
> functions.
>
> This commit folds a number of local variables in __btrfs_alloc_chunk() into
> one "struct alloc_chunk_ctl ctl". There is no functional change.
>
> Signed-off-by: Naohiro Aota <naohiro.aota@wdc.com>
> ---
> fs/btrfs/volumes.c | 143 +++++++++++++++++++++++++--------------------
> 1 file changed, 81 insertions(+), 62 deletions(-)
>
> diff --git a/fs/btrfs/volumes.c b/fs/btrfs/volumes.c
> index 9bb673df777a..cfde302bf297 100644
> --- a/fs/btrfs/volumes.c
> +++ b/fs/btrfs/volumes.c
> @@ -4818,6 +4818,29 @@ static void check_raid1c34_incompat_flag(struct btrfs_fs_info *info, u64 type)
> btrfs_set_fs_incompat(info, RAID1C34);
> }
>
> +/*
> + * Structure used internally for __btrfs_alloc_chunk() function.
> + * Wraps needed parameters.
> + */
> +struct alloc_chunk_ctl {
> + u64 start;
> + u64 type;
> + int num_stripes; /* total number of stripes to allocate */
> + int sub_stripes; /* sub_stripes info for map */
> + int dev_stripes; /* stripes per dev */
> + int devs_max; /* max devs to use */
> + int devs_min; /* min devs needed */
> + int devs_increment; /* ndevs has to be a multiple of this */
> + int ncopies; /* how many copies to data has */
> + int nparity; /* number of stripes worth of bytes to
> + store parity information */
> + u64 max_stripe_size;
> + u64 max_chunk_size;
> + u64 stripe_size;
> + u64 chunk_size;
> + int ndevs;
> +};
> +
> static int __btrfs_alloc_chunk(struct btrfs_trans_handle *trans,
> u64 start, u64 type)
> {
> @@ -4828,23 +4851,11 @@ static int __btrfs_alloc_chunk(struct btrfs_trans_handle *trans,
> struct extent_map_tree *em_tree;
> struct extent_map *em;
> struct btrfs_device_info *devices_info = NULL;
> + struct alloc_chunk_ctl ctl;
> u64 total_avail;
> - int num_stripes; /* total number of stripes to allocate */
> int data_stripes; /* number of stripes that count for
> block group size */
> - int sub_stripes; /* sub_stripes info for map */
> - int dev_stripes; /* stripes per dev */
> - int devs_max; /* max devs to use */
> - int devs_min; /* min devs needed */
> - int devs_increment; /* ndevs has to be a multiple of this */
> - int ncopies; /* how many copies to data has */
> - int nparity; /* number of stripes worth of bytes to
> - store parity information */
> int ret;
> - u64 max_stripe_size;
> - u64 max_chunk_size;
> - u64 stripe_size;
> - u64 chunk_size;
> int ndevs;
> int i;
> int j;
> @@ -4858,32 +4869,36 @@ static int __btrfs_alloc_chunk(struct btrfs_trans_handle *trans,
> return -ENOSPC;
> }
>
> + ctl.start = start;
> + ctl.type = type;
> +
> index = btrfs_bg_flags_to_raid_index(type);
>
> - sub_stripes = btrfs_raid_array[index].sub_stripes;
> - dev_stripes = btrfs_raid_array[index].dev_stripes;
> - devs_max = btrfs_raid_array[index].devs_max;
> - if (!devs_max)
> - devs_max = BTRFS_MAX_DEVS(info);
> - devs_min = btrfs_raid_array[index].devs_min;
> - devs_increment = btrfs_raid_array[index].devs_increment;
> - ncopies = btrfs_raid_array[index].ncopies;
> - nparity = btrfs_raid_array[index].nparity;
> + ctl.sub_stripes = btrfs_raid_array[index].sub_stripes;
> + ctl.dev_stripes = btrfs_raid_array[index].dev_stripes;
> + ctl.devs_max = btrfs_raid_array[index].devs_max;
> + if (!ctl.devs_max)
> + ctl.devs_max = BTRFS_MAX_DEVS(info);
> + ctl.devs_min = btrfs_raid_array[index].devs_min;
> + ctl.devs_increment = btrfs_raid_array[index].devs_increment;
> + ctl.ncopies = btrfs_raid_array[index].ncopies;
> + ctl.nparity = btrfs_raid_array[index].nparity;
>
> if (type & BTRFS_BLOCK_GROUP_DATA) {
> - max_stripe_size = SZ_1G;
> - max_chunk_size = BTRFS_MAX_DATA_CHUNK_SIZE;
> + ctl.max_stripe_size = SZ_1G;
> + ctl.max_chunk_size = BTRFS_MAX_DATA_CHUNK_SIZE;
> } else if (type & BTRFS_BLOCK_GROUP_METADATA) {
> /* for larger filesystems, use larger metadata chunks */
> if (fs_devices->total_rw_bytes > 50ULL * SZ_1G)
> - max_stripe_size = SZ_1G;
> + ctl.max_stripe_size = SZ_1G;
> else
> - max_stripe_size = SZ_256M;
> - max_chunk_size = max_stripe_size;
> + ctl.max_stripe_size = SZ_256M;
> + ctl.max_chunk_size = ctl.max_stripe_size;
> } else if (type & BTRFS_BLOCK_GROUP_SYSTEM) {
> - max_stripe_size = SZ_32M;
> - max_chunk_size = 2 * max_stripe_size;
> - devs_max = min_t(int, devs_max, BTRFS_MAX_DEVS_SYS_CHUNK);
> + ctl.max_stripe_size = SZ_32M;
> + ctl.max_chunk_size = 2 * ctl.max_stripe_size;
> + ctl.devs_max = min_t(int, ctl.devs_max,
> + BTRFS_MAX_DEVS_SYS_CHUNK);
> } else {
> btrfs_err(info, "invalid chunk type 0x%llx requested",
> type);
> @@ -4891,8 +4906,8 @@ static int __btrfs_alloc_chunk(struct btrfs_trans_handle *trans,
> }
>
> /* We don't want a chunk larger than 10% of writable space */
> - max_chunk_size = min(div_factor(fs_devices->total_rw_bytes, 1),
> - max_chunk_size);
> + ctl.max_chunk_size = min(div_factor(fs_devices->total_rw_bytes, 1),
> + ctl.max_chunk_size);
>
> devices_info = kcalloc(fs_devices->rw_devices, sizeof(*devices_info),
> GFP_NOFS);
> @@ -4929,20 +4944,20 @@ static int __btrfs_alloc_chunk(struct btrfs_trans_handle *trans,
> continue;
>
> ret = find_free_dev_extent(device,
> - max_stripe_size * dev_stripes,
> + ctl.max_stripe_size * ctl.dev_stripes,
> &dev_offset, &max_avail);
If you are going to adjust the indentation of arguments, you need to adjust them
all.
> if (ret && ret != -ENOSPC)
> goto error;
>
> if (ret == 0)
> - max_avail = max_stripe_size * dev_stripes;
> + max_avail = ctl.max_stripe_size * ctl.dev_stripes;
>
> - if (max_avail < BTRFS_STRIPE_LEN * dev_stripes) {
> + if (max_avail < BTRFS_STRIPE_LEN * ctl.dev_stripes) {
> if (btrfs_test_opt(info, ENOSPC_DEBUG))
> btrfs_debug(info,
> "%s: devid %llu has no free space, have=%llu want=%u",
> __func__, device->devid, max_avail,
> - BTRFS_STRIPE_LEN * dev_stripes);
> + BTRFS_STRIPE_LEN * ctl.dev_stripes);
Same here.
> continue;
> }
>
> @@ -4957,30 +4972,31 @@ static int __btrfs_alloc_chunk(struct btrfs_trans_handle *trans,
> devices_info[ndevs].dev = device;
> ++ndevs;
> }
> + ctl.ndevs = ndevs;
>
> /*
> * now sort the devices by hole size / available space
> */
> - sort(devices_info, ndevs, sizeof(struct btrfs_device_info),
> + sort(devices_info, ctl.ndevs, sizeof(struct btrfs_device_info),
> btrfs_cmp_device_info, NULL);
>
> /*
> * Round down to number of usable stripes, devs_increment can be any
> * number so we can't use round_down()
> */
> - ndevs -= ndevs % devs_increment;
> + ctl.ndevs -= ctl.ndevs % ctl.devs_increment;
>
> - if (ndevs < devs_min) {
> + if (ctl.ndevs < ctl.devs_min) {
> ret = -ENOSPC;
> if (btrfs_test_opt(info, ENOSPC_DEBUG)) {
> btrfs_debug(info,
> "%s: not enough devices with free space: have=%d minimum required=%d",
> - __func__, ndevs, devs_min);
> + __func__, ctl.ndevs, ctl.devs_min);
> }
> goto error;
> }
>
> - ndevs = min(ndevs, devs_max);
> + ctl.ndevs = min(ctl.ndevs, ctl.devs_max);
>
> /*
> * The primary goal is to maximize the number of stripes, so use as
> @@ -4989,14 +5005,15 @@ static int __btrfs_alloc_chunk(struct btrfs_trans_handle *trans,
> * The DUP profile stores more than one stripe per device, the
> * max_avail is the total size so we have to adjust.
> */
> - stripe_size = div_u64(devices_info[ndevs - 1].max_avail, dev_stripes);
> - num_stripes = ndevs * dev_stripes;
> + ctl.stripe_size = div_u64(devices_info[ctl.ndevs - 1].max_avail,
> + ctl.dev_stripes);
> + ctl.num_stripes = ctl.ndevs * ctl.dev_stripes;
>
> /*
> * this will have to be fixed for RAID1 and RAID10 over
> * more drives
> */
> - data_stripes = (num_stripes - nparity) / ncopies;
> + data_stripes = (ctl.num_stripes - ctl.nparity) / ctl.ncopies;
>
> /*
> * Use the number of data stripes to figure out how big this chunk
> @@ -5004,44 +5021,44 @@ static int __btrfs_alloc_chunk(struct btrfs_trans_handle *trans,
> * and compare that answer with the max chunk size. If it's higher,
> * we try to reduce stripe_size.
> */
> - if (stripe_size * data_stripes > max_chunk_size) {
> + if (ctl.stripe_size * data_stripes > ctl.max_chunk_size) {
> /*
> * Reduce stripe_size, round it up to a 16MB boundary again and
> * then use it, unless it ends up being even bigger than the
> * previous value we had already.
> */
> - stripe_size = min(round_up(div_u64(max_chunk_size,
> + ctl.stripe_size = min(round_up(div_u64(ctl.max_chunk_size,
> data_stripes), SZ_16M),
> - stripe_size);
> + ctl.stripe_size);
And here. Thanks,
Josef
next prev parent reply other threads:[~2020-02-06 16:38 UTC|newest]
Thread overview: 52+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-02-06 10:41 [PATCH 00/20] btrfs: refactor and generalize chunk/dev_extent/extent allocation Naohiro Aota
2020-02-06 10:41 ` [PATCH 01/20] btrfs: change type of full_search to bool Naohiro Aota
2020-02-06 11:26 ` Johannes Thumshirn
2020-02-06 16:03 ` Josef Bacik
2020-02-06 10:41 ` [PATCH 02/20] btrfs: introduce chunk allocation policy Naohiro Aota
2020-02-06 11:30 ` Johannes Thumshirn
2020-02-07 6:11 ` Naohiro Aota
2020-02-06 16:06 ` Josef Bacik
2020-02-06 16:07 ` Josef Bacik
2020-02-06 10:41 ` [PATCH 03/20] btrfs: refactor find_free_dev_extent_start() Naohiro Aota
2020-02-06 12:02 ` Johannes Thumshirn
2020-02-07 6:22 ` Naohiro Aota
2020-02-06 16:34 ` Josef Bacik
2020-02-06 10:41 ` [PATCH 04/20] btrfs: introduce alloc_chunk_ctl Naohiro Aota
2020-02-06 12:07 ` Johannes Thumshirn
2020-02-06 16:38 ` Josef Bacik [this message]
2020-02-07 7:08 ` Naohiro Aota
2020-02-06 10:41 ` [PATCH 05/20] btrfs: factor out set_parameters() Naohiro Aota
2020-02-06 13:51 ` Johannes Thumshirn
2020-02-06 16:40 ` Josef Bacik
2020-02-07 7:59 ` Naohiro Aota
2020-02-06 10:42 ` [PATCH 06/20] btrfs: factor out gather_device_info() Naohiro Aota
2020-02-06 15:43 ` Johannes Thumshirn
2020-02-07 9:54 ` Naohiro Aota
2020-02-06 16:44 ` Josef Bacik
2020-02-06 10:42 ` [PATCH 07/20] btrfs: factor out decide_stripe_size() Naohiro Aota
2020-02-06 15:59 ` Johannes Thumshirn
2020-02-06 16:47 ` Josef Bacik
2020-02-06 10:42 ` [PATCH 08/20] btrfs: factor out create_chunk() Naohiro Aota
2020-02-06 16:49 ` Josef Bacik
2020-02-07 9:17 ` Naohiro Aota
2020-02-06 10:42 ` [PATCH 09/20] btrfs: parameterize dev_extent_min Naohiro Aota
2020-02-06 16:52 ` Josef Bacik
2020-02-07 9:00 ` Naohiro Aota
2020-02-06 10:42 ` [PATCH 10/20] btrfs: introduce extent allocation policy Naohiro Aota
2020-02-06 10:42 ` [PATCH 11/20] btrfs: move hint_byte into find_free_extent_ctl Naohiro Aota
2020-02-06 10:42 ` [PATCH 12/20] btrfs: introduce clustered_alloc_info Naohiro Aota
2020-02-06 12:44 ` Su Yue
2020-02-07 9:25 ` Naohiro Aota
2020-02-06 17:01 ` Josef Bacik
2020-02-07 9:53 ` Naohiro Aota
2020-02-06 10:42 ` [PATCH 13/20] btrfs: factor out do_allocation() Naohiro Aota
2020-02-06 10:42 ` [PATCH 14/20] btrfs: drop unnecessary arguments from clustered allocation functions Naohiro Aota
2020-02-06 10:42 ` [PATCH 15/20] btrfs: factor out release_block_group() Naohiro Aota
2020-02-06 10:42 ` [PATCH 16/20] btrfs: factor out found_extent() Naohiro Aota
2020-02-06 10:42 ` [PATCH 17/20] btrfs: drop unnecessary arguments from find_free_extent_update_loop() Naohiro Aota
2020-02-06 10:42 ` [PATCH 18/20] btrfs: factor out chunk_allocation_failed() Naohiro Aota
2020-02-06 10:42 ` [PATCH 19/20] btrfs: skip LOOP_NO_EMPTY_SIZE if not clustered allocation Naohiro Aota
2020-02-06 10:42 ` [PATCH 20/20] btrfs: factor out prepare_allocation() Naohiro Aota
2020-02-06 11:43 ` [PATCH 00/20] btrfs: refactor and generalize chunk/dev_extent/extent allocation Martin Steigerwald
2020-02-07 6:06 ` Naohiro Aota
2020-02-07 8:02 ` Martin Steigerwald
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=bcccefe4-8cff-d50d-ddd1-784e3d194607@toxicpanda.com \
--to=josef@toxicpanda.com \
--cc=Johannes.Thumshirn@wdc.com \
--cc=anand.jain@oracle.com \
--cc=clm@fb.com \
--cc=damien.lemoal@wdc.com \
--cc=dsterba@suse.com \
--cc=hare@suse.com \
--cc=linux-btrfs@vger.kernel.org \
--cc=linux-fsdevel@vger.kernel.org \
--cc=naohiro.aota@wdc.com \
--cc=nborisov@suse.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox