From: "John Stoffel" <john@stoffel.org>
To: Ian Pilcher <arequipeno@gmail.com>
Cc: John Stoffel <john@stoffel.org>, linux-raid@vger.kernel.org
Subject: Re: RAID types & chunks sizes for new NAS drives
Date: Tue, 23 Jun 2020 20:34:18 -0400 [thread overview]
Message-ID: <24306.40842.482838.682896@quad.stoffel.home> (raw)
In-Reply-To: <40bf8a08-61a6-2b50-b9c6-240e384de80d@gmail.com>
>>>>> "Ian" == Ian Pilcher <arequipeno@gmail.com> writes:
Ian> On 6/23/20 4:30 PM, John Stoffel wrote:
>> Well, as you add LVM volumes to a VG, I don't honestly know offhand if
>> the areas are pre-allocated, or not, I think they are pre-allocated,
>> but if you add/remove/resize LVs, you can start to get fragmentation,
>> which will hurt performance.
Ian> LVs are pre-allocated, and they definitely can become fragmented.
Ian> That's orthogonal to whether the VG is on a single RAID device or a
Ian> set of smaller adjacent RAID devices.
>> No, you still do not want the partitioned setup, becuase if you lose a
>> disk, you want to rebuild it entirely, all at once. Personally, 5 x
>> 8Tb disks setup in RAID10 with a hot spare sounds just fine to me.
>> You can survive a two disk failure if it doesn't hit both halves of
>> the mirror. But the hot spare should help protect you.
Ian> It depends on what sort of failure you're trying to protect against. If
Ian> you lose the entire disk (because of an electronic/mechanical failure,
Ian> for example) your doing either an 8TB rebuild/resync or (for example)
Ian> 16x 512GB rebuild/resyncs, which is effectively the same thing.
Ian> OTOH, if you have a patch of sectors go bad in the partitioned case,
Ian> the RAID layer is only going to automatically rebuild/resync one of the
Ian> partition-based RAID devices. To my thinking, this will reduce the
Ian> chance of a double-failure.
Once a disk starts throwing errors like this, it's toast. Get rid of
it now.
Ian> I think it's important to state that this NAS is pretty actively
Ian> monitored/managed. So if such a failure were to occur, I would
Ian> absolutely be taking steps to retire the drive with the failed sectors.
Ian> But that's something that I'd rather do manually, rather than kicking
Ian> off (for example) and 8TB rebuild with a hot-spare.
Sure, if you think that's going to happen when you're on vacation and
out of town and the disk starts flaking out... :-)
>> One thing I really like to do is mix vendors in my array, just so I
>> dont' get caught by a bad batch. And the RAID10 performance advantage
>> over RAID6 is big. You'd only get 8Tb (only! :-) more space, but much
>> worse interactive response.
Ian> Mixing vendors (or at least channels) is one of those things that I
Ian> know that I should do, but I always get impatient.
Ian> But do I need the better performance. Choices, choices ... :-)
>> Physics sucks, don't it? :-)
Ian> LOL! Indeed it does!
>> What I do is have a pair of mirrored SSDs setup to cache my RAID1
>> arrays, to give me more performance. Not really sure if it's helping
>> or hurting really. dm-cache isn't really great at reporting stats,
>> and I never bothered to test it hard.
Ian> I've played with both bcache and dm-cache, although it's been a few
Ian> years. Neither one really did much for me, but that's probably because
Ian> I was using write-through caching, as I didn't trust "newfangled" SSDs
Ian> at the time.
Sure, I understand that. It makes a difference for me when doing
kernel builds... not that I regularly upgrade.
>> My main box is an old AMD Phenom(tm) II X4 945 Processor, which is now
>> something like 10 years old. It's fast enough for what I do. I'm
>> more concerned with data loss than I am performance.
Ian> Same here. I mainly want to feel comfortable that I haven't crippled my
Ian> performance by doing something stupid, but as long as the NAS can stream
Ian> a movie to media room it's good enough.
Ian> My NAS has an Atom D2550, so it's almost certainly slower than your
Ian> Phenom.
Yeah, so that's another strike (possibly) against RAID6, since it will
be more CPU overhead, esp if you're running VMs at the same time on
there.
prev parent reply other threads:[~2020-06-24 0:34 UTC|newest]
Thread overview: 16+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-06-21 16:23 RAID types & chunks sizes for new NAS drives Ian Pilcher
2020-06-23 1:45 ` John Stoffel
2020-06-23 2:31 ` o1bigtenor
2020-06-23 17:01 ` John Stoffel
2020-06-24 22:13 ` o1bigtenor
2020-06-23 12:26 ` Nix
2020-06-23 18:50 ` John Stoffel
2020-06-23 15:36 ` antlists
2020-06-23 18:55 ` John Stoffel
2020-06-24 12:32 ` Phil Turmel
2020-06-24 14:49 ` John Stoffel
2020-06-24 18:41 ` Wols Lists
2020-06-23 20:27 ` Ian Pilcher
2020-06-23 21:30 ` John Stoffel
2020-06-23 23:16 ` Ian Pilcher
2020-06-24 0:34 ` John Stoffel [this message]
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=24306.40842.482838.682896@quad.stoffel.home \
--to=john@stoffel.org \
--cc=arequipeno@gmail.com \
--cc=linux-raid@vger.kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox