* RAID6 stable enough for production?
@ 2015-10-14 20:19 Sjoerd
2015-10-14 20:23 ` Donald Pearson
2015-10-15 1:55 ` Duncan
0 siblings, 2 replies; 11+ messages in thread
From: Sjoerd @ 2015-10-14 20:19 UTC (permalink / raw)
To: linux-btrfs
Hi all,
Is RAID6 still considered unstable so I shouldn't use it in production?
The latest I could find about a test scenario is more than a year ago
(http://marc.merlins.org/perso/btrfs/post_2014-03-23_Btrfs-Raid5-Status.html)
I want to build a new NAS (6 disks of 4TB) on RAID6 and prefer to use btrfs
over zfs, but the latter is proven stable and I am unsure about btrfs...
Main usage for me would be to able to replace 1 or 2 failing (or going to
fail) drives and be able to extend it in the future with more disks. The data
on it shouldn't get corrupted unless the building were it's in is destroyed ;)
So should I go for btrfs?
NB: I am running happily a RAID5 btrfs with 4x2TB disks in it, but you'll just
know the value of the filesystem when something goes wrong. Yes I know RAID5/6
is not a backup ;)
Cheers,
Sjoerd
^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: RAID6 stable enough for production?
2015-10-14 20:19 RAID6 stable enough for production? Sjoerd
@ 2015-10-14 20:23 ` Donald Pearson
2015-10-14 20:34 ` Lionel Bouton
2015-10-15 1:55 ` Duncan
1 sibling, 1 reply; 11+ messages in thread
From: Donald Pearson @ 2015-10-14 20:23 UTC (permalink / raw)
To: Sjoerd; +Cc: Btrfs BTRFS
I would not use Raid56 in production. I've tried using it a few
different ways but have run in to trouble with stability and
performance. Raid10 has been working excellently for me.
On Wed, Oct 14, 2015 at 3:19 PM, Sjoerd <sjoerd@sjomar.eu> wrote:
> Hi all,
>
> Is RAID6 still considered unstable so I shouldn't use it in production?
> The latest I could find about a test scenario is more than a year ago
> (http://marc.merlins.org/perso/btrfs/post_2014-03-23_Btrfs-Raid5-Status.html)
>
> I want to build a new NAS (6 disks of 4TB) on RAID6 and prefer to use btrfs
> over zfs, but the latter is proven stable and I am unsure about btrfs...
> Main usage for me would be to able to replace 1 or 2 failing (or going to
> fail) drives and be able to extend it in the future with more disks. The data
> on it shouldn't get corrupted unless the building were it's in is destroyed ;)
>
> So should I go for btrfs?
>
> NB: I am running happily a RAID5 btrfs with 4x2TB disks in it, but you'll just
> know the value of the filesystem when something goes wrong. Yes I know RAID5/6
> is not a backup ;)
>
> Cheers,
> Sjoerd
>
> --
> To unsubscribe from this list: send the line "unsubscribe linux-btrfs" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at http://vger.kernel.org/majordomo-info.html
^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: RAID6 stable enough for production?
2015-10-14 20:23 ` Donald Pearson
@ 2015-10-14 20:34 ` Lionel Bouton
2015-10-14 20:53 ` Donald Pearson
0 siblings, 1 reply; 11+ messages in thread
From: Lionel Bouton @ 2015-10-14 20:34 UTC (permalink / raw)
To: Donald Pearson, Sjoerd; +Cc: Btrfs BTRFS
Le 14/10/2015 22:23, Donald Pearson a écrit :
> I would not use Raid56 in production. I've tried using it a few
> different ways but have run in to trouble with stability and
> performance. Raid10 has been working excellently for me.
Hi, could you elaborate on the stability and performance problems you
had? Which kernels were you using at the time you were testing?
I'm interested because I have some RAID10 installations of 7 disks which
don't need much write performance (large backup servers with few clients
and few updates but very large datasets) that I plan to migrate to RAID6
when they approach their storage capacity (at least theoretically with 7
disks this will give better read performance and better protection
against disk failures). 3.19 brought full RAID5/6 support and from what
I remember there were some initial quirks but I'm unaware of any big
RAID5/6 problem in 4.1+ kernels.
Best regards,
Lionel
^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: RAID6 stable enough for production?
2015-10-14 20:34 ` Lionel Bouton
@ 2015-10-14 20:53 ` Donald Pearson
2015-10-14 21:15 ` Rich Freeman
2015-10-14 21:16 ` Lionel Bouton
0 siblings, 2 replies; 11+ messages in thread
From: Donald Pearson @ 2015-10-14 20:53 UTC (permalink / raw)
To: Lionel Bouton; +Cc: Sjoerd, Btrfs BTRFS
I've used it from 3.8 something to current, it does not handle drive
failure well at all, which is the point of parity raid. I had a 10disk
Raid6 array on 4.1.1 and a drive failure put the filesystem in an
irrecoverable state. Scrub speeds are also an order of magnitude or
more slower in my own experience. The issue isn't filesystem
read/write performance, it's maintenance and operation.
That 10 drive system was rebuilt as raid10 and I haven't had problems
since, and it's handled hdd problems reasonably well.
I finally moved away from raid56 yesterday because of the time it took
to scrub. This was a 4x3tb array raid6 that i only used for backups.
I attempted to just rebalance in to raid10 but partway through the
balance the filesystem ran in to problems, forcing the filesystem into
readonly. I tried some things to overcome that but ultimately just
wiped it out and recreated as raid10. I suspect one of the drives may
be having problems so I'm running tests on it now.
Personally I would still recommend zfs on illumos in production,
because it's nearly unshakeable and the creative things you can do to
deal with problems are pretty remarkable. The unfortunate reality is
though that over time your system will probably grow and expand and
zfs is very locked in to the original configuration. Adding vdevs is
a poor solution IMO.
On Wed, Oct 14, 2015 at 3:34 PM, Lionel Bouton
<lionel-subscription@bouton.name> wrote:
> Le 14/10/2015 22:23, Donald Pearson a écrit :
>> I would not use Raid56 in production. I've tried using it a few
>> different ways but have run in to trouble with stability and
>> performance. Raid10 has been working excellently for me.
>
> Hi, could you elaborate on the stability and performance problems you
> had? Which kernels were you using at the time you were testing?
>
> I'm interested because I have some RAID10 installations of 7 disks which
> don't need much write performance (large backup servers with few clients
> and few updates but very large datasets) that I plan to migrate to RAID6
> when they approach their storage capacity (at least theoretically with 7
> disks this will give better read performance and better protection
> against disk failures). 3.19 brought full RAID5/6 support and from what
> I remember there were some initial quirks but I'm unaware of any big
> RAID5/6 problem in 4.1+ kernels.
>
> Best regards,
>
> Lionel
^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: RAID6 stable enough for production?
2015-10-14 20:53 ` Donald Pearson
@ 2015-10-14 21:15 ` Rich Freeman
2015-10-14 21:19 ` Donald Pearson
2015-10-15 1:47 ` Chris Murphy
2015-10-14 21:16 ` Lionel Bouton
1 sibling, 2 replies; 11+ messages in thread
From: Rich Freeman @ 2015-10-14 21:15 UTC (permalink / raw)
To: Donald Pearson; +Cc: Lionel Bouton, Sjoerd, Btrfs BTRFS
On Wed, Oct 14, 2015 at 4:53 PM, Donald Pearson
<donaldwhpearson@gmail.com> wrote:
>
> Personally I would still recommend zfs on illumos in production,
> because it's nearly unshakeable and the creative things you can do to
> deal with problems are pretty remarkable. The unfortunate reality is
> though that over time your system will probably grow and expand and
> zfs is very locked in to the original configuration. Adding vdevs is
> a poor solution IMO.
>
This is the main thing that has kept me away from zfs - you can't
modify a vdev, like you can with an md array or btrfs. I don't think
zfs makes use of all your space if you have mixed disk sizes in a
raid-z either - it works like mdadm. I'm not sure whether btrfs will
be any better in that regard (if I have 2x3TB and 3x1TB drives in a
RAID5 I should get 6TB of usable space, not 4TB, without messing with
partitioning).
So, I am running raid1 btrfs in the hope that I'll be able to move to
something more efficient in the future.
However, I would not personally be using raid5/6 for anything but pure
experimentation on btrfs anytime soon. I don't even trust the 4.1
kernel series for btrfs at all just yet, and you're not going to be
running older than that for raid5/6.
--
Rich
^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: RAID6 stable enough for production?
2015-10-14 20:53 ` Donald Pearson
2015-10-14 21:15 ` Rich Freeman
@ 2015-10-14 21:16 ` Lionel Bouton
1 sibling, 0 replies; 11+ messages in thread
From: Lionel Bouton @ 2015-10-14 21:16 UTC (permalink / raw)
To: Donald Pearson; +Cc: Sjoerd, Btrfs BTRFS
Le 14/10/2015 22:53, Donald Pearson a écrit :
> I've used it from 3.8 something to current, it does not handle drive
> failure well at all, which is the point of parity raid. I had a 10disk
> Raid6 array on 4.1.1 and a drive failure put the filesystem in an
> irrecoverable state. Scrub speeds are also an order of magnitude or
> more slower in my own experience. The issue isn't filesystem
> read/write performance, it's maintenance and operation.
Thanks, I'll proceed with caution...
When 3.19 got out I tried various tests with loopback devices in RAID6
(dd if=/dev/random in the middle of one loopback device guaranteed to
have file data while using the filesystem for example) and didn't manage
to break it but it was arguably simple situations (either missing device
or corrupted data on device, not something behaving really erratically
like failing hardware).
Lionel
^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: RAID6 stable enough for production?
2015-10-14 21:15 ` Rich Freeman
@ 2015-10-14 21:19 ` Donald Pearson
2015-10-15 1:47 ` Chris Murphy
1 sibling, 0 replies; 11+ messages in thread
From: Donald Pearson @ 2015-10-14 21:19 UTC (permalink / raw)
To: Rich Freeman; +Cc: Lionel Bouton, Sjoerd, Btrfs BTRFS
btrfs does handle mixed device sizes really well actually. And you're
right, zfs is limited to the smallest drive x vdev width. The rest
goes unused. You can do things like pre-slice the drives with sparse
files and create zfs on those files, but then you'll load up those
larger drives with a lot more iop requests and you may dramatically
slow things down. It can expand a vdev by replacing drives with
larger ones. The other drawback is the single-drive performance of a
single vdev regardless of the number of drives in it. I love zfs for
a lot of reasons, and dislike it for a lot too. I ultimately decided
to use btrfs on my personal equipment because it promises to be more
organic and my commodity hardware definitely likes to play the organic
role. :)
On Wed, Oct 14, 2015 at 4:15 PM, Rich Freeman
<r-btrfs@thefreemanclan.net> wrote:
> On Wed, Oct 14, 2015 at 4:53 PM, Donald Pearson
> <donaldwhpearson@gmail.com> wrote:
>>
>> Personally I would still recommend zfs on illumos in production,
>> because it's nearly unshakeable and the creative things you can do to
>> deal with problems are pretty remarkable. The unfortunate reality is
>> though that over time your system will probably grow and expand and
>> zfs is very locked in to the original configuration. Adding vdevs is
>> a poor solution IMO.
>>
>
> This is the main thing that has kept me away from zfs - you can't
> modify a vdev, like you can with an md array or btrfs. I don't think
> zfs makes use of all your space if you have mixed disk sizes in a
> raid-z either - it works like mdadm. I'm not sure whether btrfs will
> be any better in that regard (if I have 2x3TB and 3x1TB drives in a
> RAID5 I should get 6TB of usable space, not 4TB, without messing with
> partitioning).
>
> So, I am running raid1 btrfs in the hope that I'll be able to move to
> something more efficient in the future.
>
> However, I would not personally be using raid5/6 for anything but pure
> experimentation on btrfs anytime soon. I don't even trust the 4.1
> kernel series for btrfs at all just yet, and you're not going to be
> running older than that for raid5/6.
>
> --
> Rich
^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: RAID6 stable enough for production?
2015-10-14 21:15 ` Rich Freeman
2015-10-14 21:19 ` Donald Pearson
@ 2015-10-15 1:47 ` Chris Murphy
2015-10-15 16:40 ` Rich Freeman
1 sibling, 1 reply; 11+ messages in thread
From: Chris Murphy @ 2015-10-15 1:47 UTC (permalink / raw)
To: Rich Freeman, Btrfs BTRFS
On Wed, Oct 14, 2015 at 3:15 PM, Rich Freeman
<r-btrfs@thefreemanclan.net> wrote:
> This is the main thing that has kept me away from zfs - you can't
> modify a vdev, like you can with an md array or btrfs.
A possible work around is ZoL (ZFS on Linux) used as a GlusterFS brick.
For that matter, now that GlusterFS has checksums and snapshots, if
your workflow permits use a glusterfs only workflow (using SMB or NFS
for Windows or OS X, and the libvirt glusterfs backend for images) you
could build a conventional md/lvm RAID+XFS brick and then glusterfs on
that. And with distributed-replicated you can bail on raid6 and just
go raid 5. Heck if you have enough bricks you could do raid0 and just
let the bricks implode if need be.
--
Chris Murphy
^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: RAID6 stable enough for production?
2015-10-14 20:19 RAID6 stable enough for production? Sjoerd
2015-10-14 20:23 ` Donald Pearson
@ 2015-10-15 1:55 ` Duncan
1 sibling, 0 replies; 11+ messages in thread
From: Duncan @ 2015-10-15 1:55 UTC (permalink / raw)
To: linux-btrfs
Sjoerd posted on Wed, 14 Oct 2015 22:19:50 +0200 as excerpted:
> Is RAID6 still considered unstable so I shouldn't use it in production?
> The latest I could find about a test scenario is more than a year ago
> (http://marc.merlins.org/perso/btrfs/post_2014-03-23_Btrfs-Raid5-
Status.html)
>
> I want to build a new NAS (6 disks of 4TB) on RAID6 and prefer to use
> btrfs over zfs, but the latter is proven stable and I am unsure about
> btrfs...
My general recommendation for new btrfs features is to let them stabilize
for a year -- five kernel series -- after they are nominally complete,
before you consider them roughly as stable as btrfs itself. Given that
btrfs itself is definitely stabilizing, but not yet fully stable and
mature (a status that it's likely to maintain for... probably another
year at least, IMO, I'll skip the supporting detail arguments here), that
would be the "stable" target for new features, as well.
Since btrfs raid56 mode was nominally complete in 3.19, that would place
"stable as btrfs in general" at 4.4, and raid56 mode does indeed seem to
be healthy and maturing, with no new show-stopping bugs since 4.1, such
that I'm reasonably confident in that 4.4 prediction.
Throwing in another variable, that of LTS-stable kernels, which are very
likely to be the ones of interest to folks looking at production usage,
the current two latest LTS series are 3.18, which was pre-raid56-
completion, and 4.1, which was just after the last known-to-date raid56
show-stopping bug.
Given that situation and the above 4.4 prediction, the absolute earliest
LTS-series "production" recommendation I could comfortably make would be
4.1 series, after it has time to integrate any 4.4 series back-patches,
so say around kernels 4.5 or possibly 4.6, but deploying (after site
testing of course) the 4.1 LTS series as stable at that point.
An arguably more conservative position would be to declare the 4.1 LTS
series still too early as it didn't originate after the year
stabilization period, and wait for the next LTS series /after/ 4.4
(possibly including 4.4 if it is picked up as such). Once that series,
whatever it may be, is declared LTS, start site-testing it, and deploy it
after you're comfortable with the results of those tests, presumably at
least a couple kernel cycles later, so if for example 4.4 itself is
declared LTS, again, possibly around 4.6 or so, but deploying the LTS
series not the just released kernel.
That of course is using my general btrfs feature stability rules as
developed after some time as a regular on this list, not after any raid56
specific experience as Donald Pearson and others in-thread are reporting,
since my own use-case is btrfs raid1 mode (deployed as multiple small
pair-device btrfs raid1 filesystems on partitioned SSDs, without use of
the subvolume/snapshot or btrfs send/receive features).
Tho it can be noted that my recommendations and theirs dovetail at this
point, that btrfs raid56 isn't ready /yet/ for production usage. I'm
just predicting based on general btrfs experience that given general
experience, it should be "as ready as btrfs itself is" come 4.4 or the
next LTS kernel series thereafter, whereas I don't read them as making
such predictions at this point... which of course they couldn't, since a
feature-specific experience-based recommendation obviously rests on
current experience, and everyone appears to agree that it simply isn't
there at this point. I'm simply predicting that it well /could/ be, by
LTS-after-4.4 time, even if it isn't, now.
--
Duncan - List replies preferred. No HTML msgs.
"Every nonfree program has a lord, a master --
and if you use the program, he is your master." Richard Stallman
^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: RAID6 stable enough for production?
2015-10-15 1:47 ` Chris Murphy
@ 2015-10-15 16:40 ` Rich Freeman
2015-10-15 19:04 ` Chris Murphy
0 siblings, 1 reply; 11+ messages in thread
From: Rich Freeman @ 2015-10-15 16:40 UTC (permalink / raw)
To: Chris Murphy; +Cc: Btrfs BTRFS
On Wed, Oct 14, 2015 at 9:47 PM, Chris Murphy <lists@colorremedies.com> wrote:
>
> For that matter, now that GlusterFS has checksums and snapshots...
Interesting - I haven't kept up with that. Does it actually do
end-to-end checksums? That is, compute the checksum at the time of
storage, store the checksum in the metadata somehow, and ensure the
checksum matches when data is retrieved?
I forget whether it was glusterfs or ceph I was looking at, but some
of those distributed filesystems will only checksum data while in
transit, but not while it is at rest. So, if a server claims it has a
copy of the file, then it is assumed to be a good copy and you never
realize that even though you have 5 copies of that file distributed
around the server you ended up using differs from the other 4.
I'm also not sure if it supports an n+1/2 model like raid5/6, or if it
is just a 2*n model like raid1. If I want to store 5TB of data with
redundancy, I'd prefer to not need 10TB worth of drives to do it,
regardless of how many systems they're spread across.
--
Rich
^ permalink raw reply [flat|nested] 11+ messages in thread
* Re: RAID6 stable enough for production?
2015-10-15 16:40 ` Rich Freeman
@ 2015-10-15 19:04 ` Chris Murphy
0 siblings, 0 replies; 11+ messages in thread
From: Chris Murphy @ 2015-10-15 19:04 UTC (permalink / raw)
To: Rich Freeman; +Cc: Chris Murphy, Btrfs BTRFS
On Thu, Oct 15, 2015 at 10:40 AM, Rich Freeman
<r-btrfs@thefreemanclan.net> wrote:
> On Wed, Oct 14, 2015 at 9:47 PM, Chris Murphy <lists@colorremedies.com> wrote:
>>
>> For that matter, now that GlusterFS has checksums and snapshots...
>
> Interesting - I haven't kept up with that. Does it actually do
> end-to-end checksums? That is, compute the checksum at the time of
> storage, store the checksum in the metadata somehow, and ensure the
> checksum matches when data is retrieved?
http://www.gluster.org/community/documentation/index.php/Features/BitRot
It could be argued that since checksums are computed after writing, that
SDC has already happened by the time the checksum is computed and written
(I also don't know if the metadata itself is checksummed so the checksum
could be wrong maybe and we don't know that?).
But yes it's stored checksum. It can be enabled to check on read/open but
by default it only does verification with scrubs. And the checksums are per
file, and are SHA256 based. So it's different than Btrfs, and it's passive.
But it's there.
--
Chris Murphy
^ permalink raw reply [flat|nested] 11+ messages in thread
end of thread, other threads:[~2015-10-15 19:04 UTC | newest]
Thread overview: 11+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2015-10-14 20:19 RAID6 stable enough for production? Sjoerd
2015-10-14 20:23 ` Donald Pearson
2015-10-14 20:34 ` Lionel Bouton
2015-10-14 20:53 ` Donald Pearson
2015-10-14 21:15 ` Rich Freeman
2015-10-14 21:19 ` Donald Pearson
2015-10-15 1:47 ` Chris Murphy
2015-10-15 16:40 ` Rich Freeman
2015-10-15 19:04 ` Chris Murphy
2015-10-14 21:16 ` Lionel Bouton
2015-10-15 1:55 ` Duncan
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).