From: Bill Davidsen <davidsen@tmr.com>
To: Beolach <beolach@gmail.com>
Cc: Doug Ledford <dledford@redhat.com>,
Leslie Rhorer <lrhorer@satx.rr.com>,
linux-raid@vger.kernel.org
Subject: Re: Successful RAID 6 setup
Date: Mon, 09 Nov 2009 12:37:57 -0500 [thread overview]
Message-ID: <4AF85375.9@tmr.com> (raw)
In-Reply-To: <aebf5d970911072242l3876282fm93dc938dd3df1990@mail.gmail.com>
Beolach wrote:
> On Sat, Nov 7, 2009 at 11:35, Doug Ledford <dledford@redhat.com> wrote:
>
>> On 11/04/2009 01:40 PM, Leslie Rhorer wrote:
>>
>>> I would recommend a larger chunk size. I'm using 256K, and even
>>> 512K or 1024K probably would not be excessive.
>>>
>> OK, I've got some data that I'm not quite ready to send out yet, but it
>> maps out the relationship between max_sectors_kb (largest request size a
>> disk can process, which varies based upon scsi host adapter in question,
>> but for SATA adapters is capped at and defaults to 512KB max per
>> request) and chunk size for a raid0 array across 4 disks or 5 disks (I
>> could run other array sizes too, and that's part of what I'm waiting on
>> before sending the data out). The point here being that a raid0 array
>> will show up more of the md/lower layer block device interactions where
>> as raid5/6 would muddy the waters with other stuff. The results of the
>> tests I ran were pretty conclusive that the sweet spot for chunk size is
>> when chunk size is == max_sectors_kb, and since SATA is the predominant
>> thing today and it defaults to 512K, that gives a 512K chunk as the
>> sweet spot. Given that the chunk size is generally about optimizing
>> block device operations at the command/queue level, it should transfer
>> directly to raid5/6 as well.
>>
>>
>
> This only really applies for large sequential io loads, right? I seem
> to recall
> smaller chunk sizes being more effective for smaller random io loads.
>
Not true now (if it ever was). The operative limit here is seek time,
not transfer time. Back in the day of old and slow drives, hanging off
old and slow connections, the time to transfer the data was somewhat of
an issue. Current SATA drives and controllers have higher transfer
rates, and until SSD make seek times smaller bigger is better within reason.
Related question: that said, why is a six drive raid6 slower than a four
drive? On a small write all the data chunks have to be read, but that
can be done in parallel, so the limit should stay at the seek time of
the slowest drive. In practice it behaves as if the data chunks were
being read one at a time. Is that real, or just fallout from not a long
enough test to smooth out the data?
--
Bill Davidsen <davidsen@tmr.com>
"We can't solve today's problems by using the same thinking we
used in creating them." - Einstein
next prev parent reply other threads:[~2009-11-09 17:37 UTC|newest]
Thread overview: 9+ messages / expand[flat|nested] mbox.gz Atom feed top
2009-11-04 18:03 Successful RAID 6 setup Andrew Dunn
2009-11-04 18:40 ` Leslie Rhorer
2009-11-07 18:35 ` Doug Ledford
2009-11-08 6:42 ` Beolach
2009-11-08 16:15 ` Doug Ledford
2009-11-09 17:37 ` Bill Davidsen [this message]
2009-11-06 21:22 ` Thomas Fjellstrom
2009-11-08 7:07 ` Beolach
2009-11-08 13:31 ` Andrew Dunn
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=4AF85375.9@tmr.com \
--to=davidsen@tmr.com \
--cc=beolach@gmail.com \
--cc=dledford@redhat.com \
--cc=linux-raid@vger.kernel.org \
--cc=lrhorer@satx.rr.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).