From: Stan Hoeppner <stan@hardwarefreak.com>
To: daobang wang <wangdb1981@gmail.com>
Cc: "Mathias Burén" <mathias.buren@gmail.com>,
linux-raid <linux-raid@vger.kernel.org>
Subject: Re: RAID5 created by 8 disks works with xfs
Date: Sat, 31 Mar 2012 23:52:47 -0500 [thread overview]
Message-ID: <4F77DF1F.6080103@hardwarefreak.com> (raw)
In-Reply-To: <CACwgYDOSrEYUps2VLpFSZ-irH7Mn_-BWrOYYDgXS=ULrmVuEPw@mail.gmail.com>
Flaky mouse button cause a premature send. Sorry for the duplicate.
On 3/31/2012 8:16 PM, daobang wang wrote:
> Thanks to Mathias and stan, Here is the detail of the configuration.
>
> 1. RAID5 with 8 2TB ST32000644NS disks , i can extend to 16 disks.
> the RAID5 created with Chunk Size of 64K and left-symmetric
>
> 2. Volume Group on RAID5 with full capacity
>
> 3. Logical Volume on the Volume Group with full capacity
LVM will create unnecessary overhead for this workload. Do not use it.
Directly format the md device with XFS and proper alignment. Again,
mkfs.xfs will do this automatically.
> 4. XFS filesystem created on the Logical Volume with option "-f -i
> size=512", and mount option is "-t xfs -o
> defaults,usrquota,grpquota,noatime,nodiratime,nobarrier,delaylog,logbsize=262144",
What kernel version are you using? logbsize=262144 has been the default
for quite some time now.
NEVER disable barriers unless you have a quality BBWC RAID controller
and a good working UPS. Are these 8 disks connected to a BBWC RAID
card? Have you verified the write back cache is working properly?
> 5. The real application is 200 D1(2Mb/s) video streams write 500MB
> files on the XFS.
This is a 50 MB/s raw stream rate with 200 writers to 200 target files.
It is very likely that neither 8 nor 16 disks in RAID5 will be able to
sync this rate due to excessive head seeking, as I previously mentioned.
Having LVM layered between XFS and mdraid will make this situation even
worse.
> This is the pressure testing, just verify the reliability of the
> system, we will not use it in real envrionment, 100 video streams
> writen is our goal.
So now you have an aggregate 25 MB/s random write workload with 100
writers. This is still going to likely be too much for an 8 or 16 disk
RAID5 array for the same reason as 200 threads--too many disk seeks.
> is there any clue for optimize the application?
This is the Linux-raid list. We can help you optimize your RAID, and
since many of us use XFS, we can help you there as well. I've never
seen your application so it is impossible to suggest how to optimize it.
As far as optimizing your 16 disks with mdraid and XFS, I've already
given you pointers on how you need to configure your storage for optimal
performance with this workload. However, experience tells me you simply
don't have enough spindles to sync all these parallel writers if you
also need redundancy.
The fact that you're disabling barriers likely in lieu of BBWC seems to
indicate you're not terribly worried about losing all your data in the
event of a drive failure. If this is the case, 16 x 7.2k spindles with
6 parallel writers per spindle at .25 MB/s each might work. Simply
create an md linear array with 16 disks, no LVM, format it with
$ mkfs.xfs -d agcount=48 /dev/md0
and configure your application to write each video stream file to a
different directory. Writing to multiple directories is what drives XFS
parallelism. This will allow for 96 parallel write streams, 6 to each
drive concurrently, for 6*.25 MB/s = 1.5MB/s per drive, 24 MB/s
aggregate. This might keep seeks low enough to achieve the parallel
throughput you need.
I'm guessing since you didn't comment on my previous recommendation that
you don't yet understand the benefits of linear concatenation, and using
XFS allocation groups to drive parallelism, instead of blindly assuming
striping will do the job--it often cannot. XFS on linear array can
decrease the seek rate on each drive substantially vs striped RAID.
Keeping the seek rate low increases throughput of parallel writers.
--
Stan
next prev parent reply other threads:[~2012-04-01 4:52 UTC|newest]
Thread overview: 37+ messages / expand[flat|nested] mbox.gz Atom feed top
2012-03-31 1:22 RAID5 created by 8 disks works with xfs daobang wang
2012-03-31 7:59 ` Mathias Burén
2012-03-31 20:09 ` Stan Hoeppner
2012-04-01 1:16 ` daobang wang
2012-04-01 2:05 ` daobang wang
2012-04-01 5:13 ` Stan Hoeppner
2012-04-01 3:51 ` Stan Hoeppner
2012-04-01 5:12 ` daobang wang
2012-04-01 5:40 ` Stan Hoeppner
2012-04-01 5:59 ` daobang wang
2012-04-01 6:20 ` daobang wang
2012-04-01 7:08 ` Marcus Sorensen
2012-04-02 3:47 ` Stan Hoeppner
2012-04-05 0:48 ` daobang wang
[not found] ` <CACwgYDOtCoVF-p+KKqPYxHhA4vWF78Ueecx9hcVWLoyxFWzV9Q@mail.gmail.com>
2012-04-05 21:01 ` Stan Hoeppner
2012-04-06 0:25 ` daobang wang
2012-04-06 2:33 ` daobang wang
2012-04-06 6:00 ` Jack Wang
2012-04-06 6:45 ` daobang wang
2012-04-06 6:49 ` daobang wang
2012-04-06 8:18 ` Stan Hoeppner
2012-04-06 8:45 ` daobang wang
2012-04-06 11:12 ` Stan Hoeppner
2012-04-18 2:23 ` daobang wang
2012-04-02 3:12 ` Stan Hoeppner
2012-04-01 10:33 ` David Brown
2012-04-01 12:28 ` John Robinson
2012-04-02 6:59 ` David Brown
[not found] ` <CA+res+QkLi7sxZrD-XOcbR47CeJ5gADf7P6pa1w1oMf8CKSB4g@mail.gmail.com>
2012-04-02 8:01 ` John Robinson
2012-04-02 10:01 ` Jack Wang
2012-04-02 10:28 ` John Robinson
2012-04-02 20:41 ` Stan Hoeppner
2012-04-02 5:43 ` Stan Hoeppner
2012-04-02 7:04 ` David Brown
2012-04-02 20:21 ` Stan Hoeppner
2012-04-01 4:52 ` Stan Hoeppner [this message]
2012-04-01 8:06 ` John Robinson
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=4F77DF1F.6080103@hardwarefreak.com \
--to=stan@hardwarefreak.com \
--cc=linux-raid@vger.kernel.org \
--cc=mathias.buren@gmail.com \
--cc=wangdb1981@gmail.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).