* New RAID-5 800GB array, wich fs ? wich stripe block size ?
@ 2004-07-19 9:19 Aurélien Gouny
2004-07-19 20:36 ` Bernhard Dobbels
0 siblings, 1 reply; 5+ messages in thread
From: Aurélien Gouny @ 2004-07-19 9:19 UTC (permalink / raw)
To: linux-raid
Hi all,
I have a Promise SX6000 card and i juste have bought 6x160 GB HDs that
make a 800 GB RAID-5 array.
I was wondering wich filesytem I should use (ext3/reiserfs/xfs/jfs/...),
if I should make one or more partitions and wich stripe block size to
use knowing that:
- it will be mainly a file-server with mp3s, divx, pictures (1 to 4MB
each) and some files (openoffice, ...) all shared with Samba.
- it will have a small postgresql database
If someone need more information in order to determine the stripe block
size or else, just ask ;)
Thanks a lot, (sorry for bad english)
Aurélien
-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: New RAID-5 800GB array, wich fs ? wich stripe block size ?
2004-07-19 9:19 New RAID-5 800GB array, wich fs ? wich stripe block size ? Aurélien Gouny
@ 2004-07-19 20:36 ` Bernhard Dobbels
2004-07-19 21:05 ` Mike Hardy
0 siblings, 1 reply; 5+ messages in thread
From: Bernhard Dobbels @ 2004-07-19 20:36 UTC (permalink / raw)
To: Aurélien Gouny; +Cc: linux-raid
800 GB in one filesystem is possible, but I think it'll make life easier
if you have multiple partitions which you can mount wherever you want.
Technically speaking, I prefer the LVM2 solution on which you can make
and resize partitions on the fly.
I must admit I lost my raid5 with LVM2. The cause will probably have
been my fault, although I don't know what I did wrong, so I suggest if
you have critical data, a backup on another medium is still advised.
Good luck.
Bernhard
Aurélien Gouny wrote:
> Hi all,
>
> I have a Promise SX6000 card and i juste have bought 6x160 GB HDs that
> make a 800 GB RAID-5 array.
> I was wondering wich filesytem I should use (ext3/reiserfs/xfs/jfs/...),
> if I should make one or more partitions and wich stripe block size to
> use knowing that:
>
> - it will be mainly a file-server with mp3s, divx, pictures (1 to 4MB
> each) and some files (openoffice, ...) all shared with Samba.
> - it will have a small postgresql database
>
> If someone need more information in order to determine the stripe block
> size or else, just ask ;)
>
> Thanks a lot, (sorry for bad english)
> Aurélien
> -
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at http://vger.kernel.org/majordomo-info.html
>
-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: New RAID-5 800GB array, wich fs ? wich stripe block size ?
2004-07-19 20:36 ` Bernhard Dobbels
@ 2004-07-19 21:05 ` Mike Hardy
2004-07-20 2:39 ` TJ Harrell
2004-07-20 8:29 ` Gordon Henderson
0 siblings, 2 replies; 5+ messages in thread
From: Mike Hardy @ 2004-07-19 21:05 UTC (permalink / raw)
To: Bernhard Dobbels; +Cc: Aurélien Gouny, linux-raid
That was a tepid review at best! No offense :-)
FWIW - I've got a 1.7TB RAID5 array that handles exactly the files you
describe (we're probably doing the same thing - networked media server,
right? With some other stuff on the side). In addition, it also handles
rsync backups for all the other machines using the --link-dest argument
and a rotations so it has an unbelievable amount of symlinks in the
filesystem.
I also have a 1.0TB RAID5 array in a separate machine that I use as a
backup for the main server, where all the media stuff gets backed up,
but I don't backup the backups since that would be silly, or overkill at
least.
I just chose plain vanilla ext3 ('mke2fs -j -m1 /dev/md2') on top of
plain ol' raid 5. The only option that's different there is the "m1"
since with an enormous filesystem, reserving 5% of it for root use is a
bit silly.
With regard to performance, the first thing you'll notice is that unless
you have gigabit to everywhere, you're limited by network I/O. I am
anyway. I could saturate a 100Mbit network connection with the read
speed, and after that, who cares?
Different filesystems will clearly be better for different situations,
but if you're just looking to serve files over the network you're really
not going to need to work hard to get it set up "good enough".
I will say though, I've had hardware failures and machine failures take
the array out before - remember that MTBF is divided by the number of
parts and arrays usually have lots of parts. Don't forget to backup
early and often...
-Mike
Bernhard Dobbels wrote:
> 800 GB in one filesystem is possible, but I think it'll make life easier
> if you have multiple partitions which you can mount wherever you want.
>
> Technically speaking, I prefer the LVM2 solution on which you can make
> and resize partitions on the fly.
>
> I must admit I lost my raid5 with LVM2. The cause will probably have
> been my fault, although I don't know what I did wrong, so I suggest if
> you have critical data, a backup on another medium is still advised.
>
> Good luck.
>
> Bernhard
> Aurélien Gouny wrote:
>
>> Hi all,
>>
>> I have a Promise SX6000 card and i juste have bought 6x160 GB HDs that
>> make a 800 GB RAID-5 array.
>> I was wondering wich filesytem I should use
>> (ext3/reiserfs/xfs/jfs/...), if I should make one or more partitions
>> and wich stripe block size to use knowing that:
>>
>> - it will be mainly a file-server with mp3s, divx, pictures (1 to 4MB
>> each) and some files (openoffice, ...) all shared with Samba.
>> - it will have a small postgresql database
>>
>> If someone need more information in order to determine the stripe
>> block size or else, just ask ;)
>>
>> Thanks a lot, (sorry for bad english)
>> Aurélien
>> -
>> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
>> the body of a message to majordomo@vger.kernel.org
>> More majordomo info at http://vger.kernel.org/majordomo-info.html
>>
> -
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at http://vger.kernel.org/majordomo-info.html
-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: New RAID-5 800GB array, wich fs ? wich stripe block size ?
2004-07-19 21:05 ` Mike Hardy
@ 2004-07-20 2:39 ` TJ Harrell
2004-07-20 8:29 ` Gordon Henderson
1 sibling, 0 replies; 5+ messages in thread
From: TJ Harrell @ 2004-07-20 2:39 UTC (permalink / raw)
To: Aurélien Gouny, linux-raid
> I just chose plain vanilla ext3 ('mke2fs -j -m1 /dev/md2') on top of
> plain ol' raid 5. The only option that's different there is the "m1"
> since with an enormous filesystem, reserving 5% of it for root use is a
> bit silly.
I have a similar set up. A 600GB RAID-5 serving mp3's, and divx. I use ext3.
The journalling is a lifesaver if you ever have a crash for some reason.
Fsck
on a file-system that large is amazingly slow. I also mount it in
data=writeback
mode to improve performance. The -m option is important, otherwise you
waste a great deal of space. Additionally, the stripe and chunck sizes
should
be matched to the array when you issue the mke2fs. That's something that
was difficult for me to understand, so someone might want to explain that
one.
> With regard to performance, the first thing you'll notice is that unless
> you have gigabit to everywhere, you're limited by network I/O. I am
> anyway. I could saturate a 100Mbit network connection with the read
> speed, and after that, who cares?
I do have gigabit from my fileserver :) The disk bandwidth is actually not
that
great, though. I have yet to figure out why.
> I will say though, I've had hardware failures and machine failures take
> the array out before - remember that MTBF is divided by the number of
> parts and arrays usually have lots of parts. Don't forget to backup
> early and often...
Yes.. RAID is no replacement for backing up!!! My RAID array has proven
extremely durable and reliable through several hardware failures, however,
two drives dying at once, a simple power surge from a faulty PSU or
lightning,
or many other things could kill it. Even mistyping a command during a
recovery
could destroy the array.
> Bernhard Dobbels wrote:
> > Technically speaking, I prefer the LVM2 solution on which you can make
> > and resize partitions on the fly.
> >
> > I must admit I lost my raid5 with LVM2. The cause will probably have
> > been my fault, although I don't know what I did wrong, so I suggest if
> > you have critical data, a backup on another medium is still advised.
I mount my array on /home, and have all data files in the home directory of
my
user account, myself. I can create additional home directories for different
shares.
This is convient to me, although other people may feel otherwise. I do not
use LVM2
for two reasons. It is extra for me to learn to use, and it is an extra
possible point
of failure. The RAID code alone has been found to have bugs and weak code in
the
past. Using LVM2 on top of RAID, which also may have weak code or bugs, is
just
an additional possible fail point. In my opinion, it should not be used
unless it is
significantly useful. In my case, it isn't. I get by without needing it.
I wish you success,
TJ Harrell
^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: New RAID-5 800GB array, wich fs ? wich stripe block size ?
2004-07-19 21:05 ` Mike Hardy
2004-07-20 2:39 ` TJ Harrell
@ 2004-07-20 8:29 ` Gordon Henderson
1 sibling, 0 replies; 5+ messages in thread
From: Gordon Henderson @ 2004-07-20 8:29 UTC (permalink / raw)
To: linux-raid
On Mon, 19 Jul 2004, Mike Hardy wrote:
> I just chose plain vanilla ext3 ('mke2fs -j -m1 /dev/md2') on top of
> plain ol' raid 5. The only option that's different there is the "m1"
> since with an enormous filesystem, reserving 5% of it for root use is a
> bit silly.
I'm using ext3 on some biggish systems too. Never thought about the -m
option, but I always thought the 5% was for efficiency to stop
fragmentation?
> With regard to performance, the first thing you'll notice is that unless
> you have gigabit to everywhere, you're limited by network I/O. I am
> anyway. I could saturate a 100Mbit network connection with the read
> speed, and after that, who cares?
One thing you might want to look into with ext2/3 is the stride size.
Theres an option in mkfs.ext2/3 to let you set this, and words in the S/W
RAID HowTo describing its use with a RAID5 setup.
However, as you say, it probably doesn't matter if you only have 100Mb
networking! DAThe only time I've seen this make a difference though is
when backing up to a local DLT tape drive.
Basically, you set the stride size to the chunk-size / 4 (if you are using
the default 4K block size) So with a chunk-size of 64, the stride size
will be 16; mkfs -t ext3 -R stride=16 /dev/mdX
See
http://www.tldp.org/HOWTO/Software-RAID-HOWTO-5.html#ss5.10
and read down to the bottom of the page.
However, I feel that this sort of tuning will really only matter if your
data is of a nature to take full advantage of it - eg. all big files,
access sequentially rather than lots of little files access randomly, or
whatever! I suspect the only way to see what will work for you is to suck
it and see!
> Different filesystems will clearly be better for different situations,
> but if you're just looking to serve files over the network you're really
> not going to need to work hard to get it set up "good enough".
Indeed. One other thing that you might want to invest in is a managed
Ethernet switch. That way to can see if it really is the filesystem thats
a bottleneck (assuming it's a server and not just a filestore for local
applications). I recently checked a network I look after, after the users
were wondering if the network was running flat-out (it's only 100Mb) and
was able to show that actually, 10Mb switches would be almost good enough
for the most-part. MRTG is your friend here! I was able to get more or
less full bandwidth out of it via NFS (running Bonnie) and over 5 times
that locally (again with Bonnie) so I was happy with the servers
performance and happy to blame the applications for being slow to write
their data ;-)
> I will say though, I've had hardware failures and machine failures take
> the array out before - remember that MTBF is divided by the number of
> parts and arrays usually have lots of parts. Don't forget to backup
> early and often...
Absolutely.
Gordon
^ permalink raw reply [flat|nested] 5+ messages in thread
end of thread, other threads:[~2004-07-20 8:29 UTC | newest]
Thread overview: 5+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2004-07-19 9:19 New RAID-5 800GB array, wich fs ? wich stripe block size ? Aurélien Gouny
2004-07-19 20:36 ` Bernhard Dobbels
2004-07-19 21:05 ` Mike Hardy
2004-07-20 2:39 ` TJ Harrell
2004-07-20 8:29 ` Gordon Henderson
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).