From: Anand Jain <anand.jain@oracle.com>
To: waxhead@dirtcellar.net, linux-btrfs@vger.kernel.org
Cc: dsterba@suse.com, wqu@suse.com, hrx@bupt.moe
Subject: Re: [PATCH v2 0/3] raid1 balancing methods
Date: Mon, 21 Oct 2024 23:44:42 +0800 [thread overview]
Message-ID: <b38111e8-6b62-40a7-96b4-3f512c55c82d@oracle.com> (raw)
In-Reply-To: <09a3eabc-5e03-3ec9-d867-f86d4b40e2da@dirtcellar.net>
On 21/10/24 22:32, waxhead wrote:
> Anand Jain wrote:
>> v2:
>> 1. Move new features to CONFIG_BTRFS_EXPERIMENTAL instead of
>> CONFIG_BTRFS_DEBUG.
>> 2. Correct the typo from %est_wait to %best_wait.
>> 3. Initialize %best_wait to U64_MAX and remove the check for 0.
>> 4. Implement rotation with a minimum contiguous read threshold before
>> switching to the next stripe. Configure this, using:
>>
>> echo rotation:[min_contiguous_read] > /sys/fs/btrfs/<uuid>/
>> read_policy
>>
>> The default value is the sector size, and the min_contiguous_read
>> value must be a multiple of the sector size.
>>
>> 5. Tested FIO random read/write and defrag compression workloads with
>> min_contiguous_read set to sector size, 192k, and 256k.
>>
>> RAID1 balancing method rotation is better for multi-process workloads
>> such as fio and also single-process workload such as defragmentation.
>
> With this functionality added, would it not also make sense to add a
> RAID0/10 profile that limits the stripe width, so a stripe does not
> spawn more than n disk (for example n=4).
>
> On systems with for example 24 disks in RAID10, a read may activate
12 > disks at the same time which could easily saturate the bus.
>
> Therefore if a storage profile that limits the number of devices a
> stripe occupy existed, it seems like there might be posibillities for
> RAID0/10 as well.
>
> Note that as of writing this I believe that RAID0/10/5/6 make the stripe
> as wide as the number of storage devices available for the filesystem.
> If I am wrong about this please ignore my jabbering and move on.
That's correct. I previously attempted to come up with a fix using
the device grouping method. If there's a convincing and more generic
way to specify how the devices should be grouped, we could consider
that.
Thanks, Anand
next prev parent reply other threads:[~2024-10-21 15:45 UTC|newest]
Thread overview: 15+ messages / expand[flat|nested] mbox.gz Atom feed top
2024-10-11 2:49 [PATCH v2 0/3] raid1 balancing methods Anand Jain
2024-10-11 2:49 ` [PATCH v2 1/3] btrfs: introduce RAID1 round-robin read balancing Anand Jain
2024-10-11 2:49 ` [PATCH v2 2/3] btrfs: use the path with the lowest latency for RAID1 reads Anand Jain
2024-10-11 2:49 ` [PATCH v2 3/3] btrfs: add RAID1 preferred read device Anand Jain
2024-10-11 3:35 ` [PATCH v2 0/3] raid1 balancing methods Anand Jain
2024-10-11 4:59 ` Qu Wenruo
2024-10-11 6:04 ` Anand Jain
2024-10-21 14:05 ` David Sterba
2024-10-21 15:36 ` Anand Jain
2024-10-21 18:42 ` David Sterba
2024-10-22 0:31 ` Anand Jain
2024-10-21 14:32 ` waxhead
2024-10-21 15:44 ` Anand Jain [this message]
2024-10-22 7:07 ` Johannes Thumshirn
2024-10-24 4:39 ` Qu Wenruo
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=b38111e8-6b62-40a7-96b4-3f512c55c82d@oracle.com \
--to=anand.jain@oracle.com \
--cc=dsterba@suse.com \
--cc=hrx@bupt.moe \
--cc=linux-btrfs@vger.kernel.org \
--cc=waxhead@dirtcellar.net \
--cc=wqu@suse.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).