* Home desktop/server RAID upgrade @ 2014-05-30 19:04 Mark Knecht 2014-05-30 19:29 ` L.M.J 0 siblings, 1 reply; 13+ messages in thread From: Mark Knecht @ 2014-05-30 19:04 UTC (permalink / raw) To: Linux-RAID Hi, My main home machine is a Gentoo machine built around a 5-disk RAID6 using this WD 500GB RAID Edition drive: http://www.amazon.com/gp/product/B001EMZPD0 I get about 1.4TB of storage after removing some space for boot and swap. I have only about 100GB left which will be used up in the next few months so I need to make some changes. The current RAID uses mdadm and the motherboard SATA2 ports. Even using RAID6 I never run out of compute cycles and have lots of memory so I'm fine with overall performance of the machine. I'm just short of storage. The predominate use of this machine is trading in the stock market, but nothing fast. It stores a lot of stock market data but it's not doing anything like high-speed trading. Just a lot of numerical analysis using R and then running a couple of Windows VMs for the trading part. I have no issues if the machine was to go down for a day or two as I have backups and other machines that can get me through a rough spot until this box runs again. After looking at a few options (like just adding NAS externally, RAID10, hardware RAID, etc.) I'm thinking I'll go the simple route, use RAID1 with two or possibly three SATA3 3TB drives hooked to the existing SATA2 ports. The two drives I'm considering and want to ask here about are WD Red and WD Se versions: WD Red 3TB: http://www.amazon.com/dp/B008JJLW4M WD Se 3TB: http://www.amazon.com/dp/B00CYSYZZC I expect that either of these in RAID1 will be considerably faster than my existing RAID6 and at 3TB will basically double my storage which is more than I will likely need on this machine over the next couple of years. If I need more I'll probably do some sort of NAS and expect I'll use my current 500GB drives as a first pass on that. Anyway, I've got a few questions: 1) Are the WD Red drives appropriate for this sort of desktop/server application or does the power saving stuff cause problems like the older WD Green drives did? I.e. - Are the Red drives really just for NAS applications? 2) Assuming the Red drives are OK then for roughly the same cost I could get 3 WD Red or 2 WD Se drives. In a RAID1 would a 3-drive Red RAID1 possibly be faster than the 2-drive Se RAID1 and at the same time give me more safety? (Eliminating issues like whether all the drives come from the same manufacturing lot and all fail at the same time, etc.) This isn't overly important. I could always buy a 3rd Se drive later. 3) If down the road I decided to buy a hardware SATA3 RAID controller can I just plug these drives in and expect the new RAID1 to work on a hardware controller or does that require a complete rebuild? For now that's it. Please let me know if I should supply more info. Thanks in advance, Mark ^ permalink raw reply [flat|nested] 13+ messages in thread
* Re: Home desktop/server RAID upgrade 2014-05-30 19:04 Home desktop/server RAID upgrade Mark Knecht @ 2014-05-30 19:29 ` L.M.J 2014-05-30 20:14 ` Mark Knecht 0 siblings, 1 reply; 13+ messages in thread From: L.M.J @ 2014-05-30 19:29 UTC (permalink / raw) To: Linux-RAID Le Fri, 30 May 2014 12:04:07 -0700, Mark Knecht <markknecht@gmail.com> a écrit : > In a RAID1 would a 3-drive Red > RAID1 possibly be faster than the 2-drive Se RAID1 and at the same > time give me more safety? Just a question inside the question : how do you manager a RAID1 with 3 drives ? Maybe you're talking about RAID5 then ? -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html ^ permalink raw reply [flat|nested] 13+ messages in thread
* Re: Home desktop/server RAID upgrade 2014-05-30 19:29 ` L.M.J @ 2014-05-30 20:14 ` Mark Knecht 2014-05-30 20:36 ` Mark Knecht 2014-05-31 10:52 ` David Brown 0 siblings, 2 replies; 13+ messages in thread From: Mark Knecht @ 2014-05-30 20:14 UTC (permalink / raw) To: L.M.J; +Cc: Linux-RAID On Fri, May 30, 2014 at 12:29 PM, L.M.J <linuxmasterjedi@free.fr> wrote: > Le Fri, 30 May 2014 12:04:07 -0700, > Mark Knecht <markknecht@gmail.com> a écrit : > >> In a RAID1 would a 3-drive Red >> RAID1 possibly be faster than the 2-drive Se RAID1 and at the same >> time give me more safety? > > Just a question inside the question : how do you manager a RAID1 with 3 drives ? Maybe you're talking about > RAID5 then ? OK, I'm no RAID expert but RAID1 is just drives in parallel right. 2 drives, 3 drives, 4 drives, all holding exactly the same data. In the case of a 3-drive RAID1 - if there is such a beast - I could safely lose 2 drives. You ask a reasonable question though as maybe the way this is actually done is 2 drives + a hot spare in the box that gets sync'ed if and only if one drive fails. Not sure and maybe I'm totally wrong about that. A 3-drive RAID5 would be 2 drives in series - in this case making 6TB - and then the 3rd drive being the redundancy. In the case of a 3-drive RAID5 I could safely lose 1 drive. In my case I don't need more than 3TB, so an option would be a 3-drive RAID5 made out of 2TB drives which would give me 4TB but I don't need the space as much as I want the redundancy and I think RAID5 is slower than RAID1. Additionally some more mdadm RAID knowledgeable people on other lists say Linux mdadm RAID1 would be faster as it will get data from more than one drive at a time. (Or possibly get data from which ever drive returns it the fastest. Not sure.) I believe one good option if I wanted 4 physical drives would be RAID10 but that's getting more complicated again which I didn't really want to do. So maybe it is just 2 drives and the 3 drive version isn't even a possibility? Could be. -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html ^ permalink raw reply [flat|nested] 13+ messages in thread
* Re: Home desktop/server RAID upgrade 2014-05-30 20:14 ` Mark Knecht @ 2014-05-30 20:36 ` Mark Knecht 2014-05-30 20:58 ` Roberto Spadim 2014-05-31 10:52 ` David Brown 1 sibling, 1 reply; 13+ messages in thread From: Mark Knecht @ 2014-05-30 20:36 UTC (permalink / raw) To: L.M.J; +Cc: Linux-RAID On Fri, May 30, 2014 at 1:14 PM, Mark Knecht <markknecht@gmail.com> wrote: > On Fri, May 30, 2014 at 12:29 PM, L.M.J <linuxmasterjedi@free.fr> wrote: >> Le Fri, 30 May 2014 12:04:07 -0700, >> Mark Knecht <markknecht@gmail.com> a écrit : >> >>> In a RAID1 would a 3-drive Red >>> RAID1 possibly be faster than the 2-drive Se RAID1 and at the same >>> time give me more safety? >> >> Just a question inside the question : how do you manager a RAID1 with 3 drives ? Maybe you're talking about >> RAID5 then ? > > OK, I'm no RAID expert but RAID1 is just drives in parallel right. 2 > drives, 3 drives, 4 drives, all holding exactly the same data. In the > case of a 3-drive RAID1 - if there is such a beast - I could safely > lose 2 drives. You ask a reasonable question though as maybe the way > this is actually done is 2 drives + a hot spare in the box that gets > sync'ed if and only if one drive fails. Not sure and maybe I'm totally > wrong about that. > > A 3-drive RAID5 would be 2 drives in series - in this case making 6TB > - and then the 3rd drive being the redundancy. In the case of a > 3-drive RAID5 I could safely lose 1 drive. > > In my case I don't need more than 3TB, so an option would be a 3-drive > RAID5 made out of 2TB drives which would give me 4TB but I don't need > the space as much as I want the redundancy and I think RAID5 is slower > than RAID1. Additionally some more mdadm RAID knowledgeable people on > other lists say Linux mdadm RAID1 would be faster as it will get data > from more than one drive at a time. (Or possibly get data from which > ever drive returns it the fastest. Not sure.) > > I believe one good option if I wanted 4 physical drives would be > RAID10 but that's getting more complicated again which I didn't really > want to do. > > So maybe it is just 2 drives and the 3 drive version isn't even a > possibility? Could be. Using the instructions here: http://sempike.blogspot.com/2012/06/linux-software-raid-mdadm-with-virtual.html I just built a 3 device RAID1 using loop devices and it seems to have worked. Below md50 did not exist, I created it as a 3 device RAID1 and then mdadm shows it's there. I have no idea if it's a good thing to do but mdadm doesn't stop me. I would need to test other real things like putting a file system on it, mounting it, etc., to be more confident but this much seems to work fine. Great question and good experience for me doing this. Thanks! Cheers, Mark c2RAID6 loopraid # cat /proc/mdstat Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid4] md3 : active raid6 sdb3[9] sdf3[5] sde3[6] sdd3[7] sdc3[8] 1452264480 blocks super 1.2 level 6, 16k chunk, algorithm 2 [5/5] [UUUUU] unused devices: <none> c2RAID6 loopraid # mdadm --create /dev/md50 --level=1 --raid-devices=3 /dev/loop[012] mdadm: Note: this array has metadata at the start and may not be suitable as a boot device. If you plan to store '/boot' on this device please ensure that your boot-loader understands md/v1.x metadata, or use --metadata=0.90 Continue creating array? y mdadm: Defaulting to version 1.2 metadata mdadm: array /dev/md50 started. c2RAID6 loopraid # cat /proc/mdstat Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid4] md50 : active raid1 loop2[2] loop1[1] loop0[0] 20416 blocks super 1.2 [3/3] [UUU] md3 : active raid6 sdb3[9] sdf3[5] sde3[6] sdd3[7] sdc3[8] 1452264480 blocks super 1.2 level 6, 16k chunk, algorithm 2 [5/5] [UUUUU] unused devices: <none> c2RAID6 loopraid # -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html ^ permalink raw reply [flat|nested] 13+ messages in thread
* Re: Home desktop/server RAID upgrade 2014-05-30 20:36 ` Mark Knecht @ 2014-05-30 20:58 ` Roberto Spadim 0 siblings, 0 replies; 13+ messages in thread From: Roberto Spadim @ 2014-05-30 20:58 UTC (permalink / raw) To: Mark Knecht; +Cc: L.M.J, Linux-RAID no problem using 3 hdd with raid1 about speed diference: write speed is the same (slowest disk give the write speed of raid array) read speed is the same (each disk have a mb/s rate) but ... number of hdd heads is diferent, in other words, if you read, part 1%-20% 40%-60% 90%-100% with 3 threads, you will end faster with 3 disks than with 2 disks raid1 with many disks give you a better parallel read speed (more disks = more heads = more read threads), but write speed is "the same" as a single disk 2014-05-30 17:36 GMT-03:00 Mark Knecht <markknecht@gmail.com>: > On Fri, May 30, 2014 at 1:14 PM, Mark Knecht <markknecht@gmail.com> wrote: >> On Fri, May 30, 2014 at 12:29 PM, L.M.J <linuxmasterjedi@free.fr> wrote: >>> Le Fri, 30 May 2014 12:04:07 -0700, >>> Mark Knecht <markknecht@gmail.com> a écrit : >>> >>>> In a RAID1 would a 3-drive Red >>>> RAID1 possibly be faster than the 2-drive Se RAID1 and at the same >>>> time give me more safety? >>> >>> Just a question inside the question : how do you manager a RAID1 with 3 drives ? Maybe you're talking about >>> RAID5 then ? >> >> OK, I'm no RAID expert but RAID1 is just drives in parallel right. 2 >> drives, 3 drives, 4 drives, all holding exactly the same data. In the >> case of a 3-drive RAID1 - if there is such a beast - I could safely >> lose 2 drives. You ask a reasonable question though as maybe the way >> this is actually done is 2 drives + a hot spare in the box that gets >> sync'ed if and only if one drive fails. Not sure and maybe I'm totally >> wrong about that. >> >> A 3-drive RAID5 would be 2 drives in series - in this case making 6TB >> - and then the 3rd drive being the redundancy. In the case of a >> 3-drive RAID5 I could safely lose 1 drive. >> >> In my case I don't need more than 3TB, so an option would be a 3-drive >> RAID5 made out of 2TB drives which would give me 4TB but I don't need >> the space as much as I want the redundancy and I think RAID5 is slower >> than RAID1. Additionally some more mdadm RAID knowledgeable people on >> other lists say Linux mdadm RAID1 would be faster as it will get data >> from more than one drive at a time. (Or possibly get data from which >> ever drive returns it the fastest. Not sure.) >> >> I believe one good option if I wanted 4 physical drives would be >> RAID10 but that's getting more complicated again which I didn't really >> want to do. >> >> So maybe it is just 2 drives and the 3 drive version isn't even a >> possibility? Could be. > > Using the instructions here: > > http://sempike.blogspot.com/2012/06/linux-software-raid-mdadm-with-virtual.html > > I just built a 3 device RAID1 using loop devices and it seems to have > worked. Below md50 did not exist, I created it as a 3 device RAID1 and > then mdadm shows it's there. I have no idea if it's a good thing to do > but mdadm doesn't stop me. I would need to test other real things like > putting a file system on it, mounting it, etc., to be more confident > but this much seems to work fine. > > Great question and good experience for me doing this. Thanks! > > Cheers, > Mark > > > c2RAID6 loopraid # cat /proc/mdstat > Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid4] > md3 : active raid6 sdb3[9] sdf3[5] sde3[6] sdd3[7] sdc3[8] > 1452264480 blocks super 1.2 level 6, 16k chunk, algorithm 2 [5/5] [UUUUU] > > unused devices: <none> > c2RAID6 loopraid # mdadm --create /dev/md50 --level=1 --raid-devices=3 > /dev/loop[012] > mdadm: Note: this array has metadata at the start and > may not be suitable as a boot device. If you plan to > store '/boot' on this device please ensure that > your boot-loader understands md/v1.x metadata, or use > --metadata=0.90 > Continue creating array? y > mdadm: Defaulting to version 1.2 metadata > mdadm: array /dev/md50 started. > c2RAID6 loopraid # cat /proc/mdstat > Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid4] > md50 : active raid1 loop2[2] loop1[1] loop0[0] > 20416 blocks super 1.2 [3/3] [UUU] > > md3 : active raid6 sdb3[9] sdf3[5] sde3[6] sdd3[7] sdc3[8] > 1452264480 blocks super 1.2 level 6, 16k chunk, algorithm 2 [5/5] [UUUUU] > > unused devices: <none> > c2RAID6 loopraid # > -- > To unsubscribe from this list: send the line "unsubscribe linux-raid" in > the body of a message to majordomo@vger.kernel.org > More majordomo info at http://vger.kernel.org/majordomo-info.html -- Roberto Spadim -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html ^ permalink raw reply [flat|nested] 13+ messages in thread
* Re: Home desktop/server RAID upgrade 2014-05-30 20:14 ` Mark Knecht 2014-05-30 20:36 ` Mark Knecht @ 2014-05-31 10:52 ` David Brown [not found] ` <8mtskybo2j1i4l2bqu51l7ll.1401554092920@email.android.com> 1 sibling, 1 reply; 13+ messages in thread From: David Brown @ 2014-05-31 10:52 UTC (permalink / raw) To: Mark Knecht, L.M.J; +Cc: Linux-RAID On 30/05/14 22:14, Mark Knecht wrote: > On Fri, May 30, 2014 at 12:29 PM, L.M.J <linuxmasterjedi@free.fr> > wrote: >> Le Fri, 30 May 2014 12:04:07 -0700, Mark Knecht >> <markknecht@gmail.com> a écrit : >> >>> In a RAID1 would a 3-drive Red RAID1 possibly be faster than the >>> 2-drive Se RAID1 and at the same time give me more safety? >> >> Just a question inside the question : how do you manager a RAID1 >> with 3 drives ? Maybe you're talking about RAID5 then ? > > OK, I'm no RAID expert but RAID1 is just drives in parallel right. 2 > drives, 3 drives, 4 drives, all holding exactly the same data. In > the case of a 3-drive RAID1 - if there is such a beast - I could > safely lose 2 drives. You ask a reasonable question though as maybe > the way this is actually done is 2 drives + a hot spare in the box > that gets sync'ed if and only if one drive fails. Not sure and maybe > I'm totally wrong about that. > > A 3-drive RAID5 would be 2 drives in series - in this case making > 6TB - and then the 3rd drive being the redundancy. In the case of a > 3-drive RAID5 I could safely lose 1 drive. > > In my case I don't need more than 3TB, so an option would be a > 3-drive RAID5 made out of 2TB drives which would give me 4TB but I > don't need the space as much as I want the redundancy and I think > RAID5 is slower than RAID1. Additionally some more mdadm RAID > knowledgeable people on other lists say Linux mdadm RAID1 would be > faster as it will get data from more than one drive at a time. (Or > possibly get data from which ever drive returns it the fastest. Not > sure.) > > I believe one good option if I wanted 4 physical drives would be > RAID10 but that's getting more complicated again which I didn't > really want to do. > > So maybe it is just 2 drives and the 3 drive version isn't even a > possibility? Could be. With 3 drives, you have several possibilities. Raid5 makes "stripes" across the three drives, with 2 parts holding data and one part holding parity to provide redundancy. Raid1 is commonly called "mirroring", because you get the same data on each disk. md raid has no problem making a 3-way mirror, so that each disk is identical. This gives you excellent redundancy, and you can make three different reads in parallel - but writes have to go to each disk, which can be a little slower than using 2 disks. It's not often that people need that level of redundancy. Another option with md raid is the raid10 setups. For many uses, the fastest arrangement is raid10,f2. This means there is two copies of all your data (f3 would be three copies), with a "far" layout. <http://en.wikipedia.org/wiki/Linux_MD_RAID_10#LINUX-MD-RAID-10> With this arrangement, reads are striped across all three disks, which is fast for large reads. Small reads can be handled in parallel. Most reads while be handled from the outer half of the disk, which is faster and needs less head movement - so reading is on average faster than a raid0 on the same disks. Small writes are fast, but large writes require quite a bit of head movement to get everything written twice to different parts of the disks. The "best" option always depends on your needs - how you want to access your files. A layout geared to fast striped reads of large files will be poorer for parallel small writes, and vice versa. raid10,f2 is often the best choice for a desktop or small system - but it is not very flexible if you later want to add new disks or replace the disks with bigger ones. md raid is flexible enough that it will even let you make a 3 disk raid6 array if you want - but a 3-way raid1 mirror will give you the same disk space and much better performance. -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html ^ permalink raw reply [flat|nested] 13+ messages in thread
[parent not found: <8mtskybo2j1i4l2bqu51l7ll.1401554092920@email.android.com>]
* Re: Home desktop/server RAID upgrade [not found] ` <8mtskybo2j1i4l2bqu51l7ll.1401554092920@email.android.com> @ 2014-06-01 14:25 ` Mark Knecht 2014-06-01 15:06 ` David Brown 0 siblings, 1 reply; 13+ messages in thread From: Mark Knecht @ 2014-06-01 14:25 UTC (permalink / raw) To: Craig Curtin; +Cc: David Brown, L.M.J, Linux-RAID Hi Craig, Responding to both you and David Brown. Thanks for your ideas. - Mark On Sat, May 31, 2014 at 9:40 AM, Craig Curtin <craigc@prosis.com.au> wrote: > It sounds like the op has additional data ports on his MOBO - wouldn't he be > better off looking at a couple of SSDs in raid 1 for his OS, swap etc and > his VMs and then leave the rest for data as raid5 - By moving the things > from the existing drives he gets back space and only purchases a couple of > good sized fast SSDs now > It's a possibility. I can get 240GB SSDs in the $120 range so that's $240 for RAID1. If I take the five existing 500GB drives and reconfigure for RAID5 that's 2TB. Overall it's not bad going from 1.4TB to about 2.2TB but being it's not all one big disk I'll likely never use it all as efficiently. Still, it's an option. I do in fact have extra ports: c2RAID6 ~ # lspci | grep SATA 00:1f.2 IDE interface: Intel Corporation 82801JI (ICH10 Family) 4 port SATA IDE Controller #1 00:1f.5 IDE interface: Intel Corporation 82801JI (ICH10 Family) 2 port SATA IDE Controller #2 03:00.0 SATA controller: Marvell Technology Group Ltd. 88SE9123 PCIe SATA 6.0 Gb/s controller (rev 11) 06:00.0 SATA controller: JMicron Technology Corp. JMB363 SATA/IDE Controller (rev 03) 06:00.1 IDE interface: JMicron Technology Corp. JMB363 SATA/IDE Controller (rev 03) c2RAID6 ~ # Currently my 5-drive RAID6 uses 5 of the Intel ports. The 6th port goes to the CD/DVD drive. Some time ago I bought the SATA3 Marvell card and a smaller (120GB) SSD. I put Gentoo on it and played around a bit but I've never really used it day-to-day. Part of my 2-drive RAID1 thinking was that I could build the new RAID1 on the SATA3 controller not even touch the existing RAID6. If it works reliably on that controller I'd be done and have 3TB. I think David's RAID10 3-drive solution could possibly work if I buy 3 of the lower cost new WD drives. I'll need to think about that. Not sure. Thanks, Mark On Sat, May 31, 2014 at 9:40 AM, Craig Curtin <craigc@prosis.com.au> wrote: > It sounds like the op has additional data ports on his MOBO - wouldn't he be > better off looking at a couple of SSDs in raid 1 for his OS, swap etc and > his VMs and then leave the rest for data as raid5 - By moving the things > from the existing drives he gets back space and only purchases a couple of > good sized fast SSDs now > > > Sent from my Samsung tablet > > . > > > -------- Original message -------- > From: David Brown > Date:31/05/2014 21:01 (GMT+10:00) > To: Mark Knecht ,"L.M.J" > Cc: Linux-RAID > Subject: Re: Home desktop/server RAID upgrade > > On 30/05/14 22:14, Mark Knecht wrote: >> On Fri, May 30, 2014 at 12:29 PM, L.M.J <linuxmasterjedi@free.fr> >> wrote: >>> Le Fri, 30 May 2014 12:04:07 -0700, Mark Knecht >>> <markknecht@gmail.com> a écrit : >>> >>>> In a RAID1 would a 3-drive Red RAID1 possibly be faster than the >>>> 2-drive Se RAID1 and at the same time give me more safety? >>> >>> Just a question inside the question : how do you manager a RAID1 >>> with 3 drives ? Maybe you're talking about RAID5 then ? >> >> OK, I'm no RAID expert but RAID1 is just drives in parallel right. 2 >> drives, 3 drives, 4 drives, all holding exactly the same data. In >> the case of a 3-drive RAID1 - if there is such a beast - I could >> safely lose 2 drives. You ask a reasonable question though as maybe >> the way this is actually done is 2 drives + a hot spare in the box >> that gets sync'ed if and only if one drive fails. Not sure and maybe >> I'm totally wrong about that. >> >> A 3-drive RAID5 would be 2 drives in series - in this case making >> 6TB - and then the 3rd drive being the redundancy. In the case of a >> 3-drive RAID5 I could safely lose 1 drive. >> >> In my case I don't need more than 3TB, so an option would be a >> 3-drive RAID5 made out of 2TB drives which would give me 4TB but I >> don't need the space as much as I want the redundancy and I think >> RAID5 is slower than RAID1. Additionally some more mdadm RAID >> knowledgeable people on other lists say Linux mdadm RAID1 would be >> faster as it will get data from more than one drive at a time. (Or >> possibly get data from which ever drive returns it the fastest. Not >> sure.) >> >> I believe one good option if I wanted 4 physical drives would be >> RAID10 but that's getting more complicated again which I didn't >> really want to do. >> >> So maybe it is just 2 drives and the 3 drive version isn't even a >> possibility? Could be. > > With 3 drives, you have several possibilities. > > Raid5 makes "stripes" across the three drives, with 2 parts holding data > and one part holding parity to provide redundancy. > > Raid1 is commonly called "mirroring", because you get the same data on > each disk. md raid has no problem making a 3-way mirror, so that each > disk is identical. This gives you excellent redundancy, and you can > make three different reads in parallel - but writes have to go to each > disk, which can be a little slower than using 2 disks. It's not often > that people need that level of redundancy. > > Another option with md raid is the raid10 setups. For many uses, the > fastest arrangement is raid10,f2. This means there is two copies of all > your data (f3 would be three copies), with a "far" layout. > > <http://en.wikipedia.org/wiki/Linux_MD_RAID_10#LINUX-MD-RAID-10> > > With this arrangement, reads are striped across all three disks, which > is fast for large reads. Small reads can be handled in parallel. Most > reads while be handled from the outer half of the disk, which is faster > and needs less head movement - so reading is on average faster than a > raid0 on the same disks. Small writes are fast, but large writes > require quite a bit of head movement to get everything written twice to > different parts of the disks. > > The "best" option always depends on your needs - how you want to access > your files. A layout geared to fast striped reads of large files will > be poorer for parallel small writes, and vice versa. raid10,f2 is often > the best choice for a desktop or small system - but it is not very > flexible if you later want to add new disks or replace the disks with > bigger ones. > > md raid is flexible enough that it will even let you make a 3 disk raid6 > array if you want - but a 3-way raid1 mirror will give you the same disk > space and much better performance. > > > -- > To unsubscribe from this list: send the line "unsubscribe linux-raid" in > the body of a message to majordomo@vger.kernel.org > More majordomo info at http://vger.kernel.org/majordomo-info.html > > Disclaimer > > CONFIDENTIAL > > This message contains confidential information and is intended only for the > intended recipients. If you are not the named addressee you should not > disseminate, distribute or copy this e-mail. Please notify the sender > immediately by e-mail if you have received this e-mail by mistake and delete > this e-mail from your system. > > Disclaimer > CONFIDENTIAL > This message contains confidential information and is intended only for the > intended recipients. If you are not the named addressee you should not > disseminate, distribute or copy this e-mail. Please notify the sender > immediately by e-mail if you have received this e-mail by mistake and delete > this e-mail from your system. -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html ^ permalink raw reply [flat|nested] 13+ messages in thread
* Re: Home desktop/server RAID upgrade 2014-06-01 14:25 ` Mark Knecht @ 2014-06-01 15:06 ` David Brown 2014-06-01 15:59 ` Mark Knecht 0 siblings, 1 reply; 13+ messages in thread From: David Brown @ 2014-06-01 15:06 UTC (permalink / raw) To: Mark Knecht, Craig Curtin; +Cc: L.M.J, Linux-RAID Hi Mark, What would be really useful here is a description of what you actually /want/. What do you want to do with these drives? What sort of files are they - big or small? Do you need fast access for large files? Do you need fast access for many files in parallel? How important is the data? How important is uptime? What sort of backups do you have? What will the future be like - are you making one big system to last for the foreseeable future, or do you need something that can easily be expanded? Are you looking for "fun, interesting and modern" or "boring but well-tested" solutions? Then you need to make a list of the hardware you have, or the budget for new hardware. Without know at least roughly what you are looking for, it's easy to end up with expensive SSDs because they are "cool", even though you might get more speed for your money with a couple of slow rust disks and a bit more ram in your system. It may be that there is no need for any sort of raid at all - perhaps one big main disk is fine, and the rest of the money spent on a backup disk (possibly external) with rsync'd copies of your data. This would mean longer downtime if your main disk failed - but it also gives some protection against user error. And perhaps btrfs with raid1 would be the best choice. A raid10,f2 is often the best choice for desktops or workstations with 2 or 3 hard disks, but it is not necessarily /the/ best choice. mvh., David On 01/06/14 16:25, Mark Knecht wrote: > Hi Craig, > Responding to both you and David Brown. Thanks for your ideas. > > - Mark > > On Sat, May 31, 2014 at 9:40 AM, Craig Curtin <craigc@prosis.com.au> wrote: >> It sounds like the op has additional data ports on his MOBO - wouldn't he be >> better off looking at a couple of SSDs in raid 1 for his OS, swap etc and >> his VMs and then leave the rest for data as raid5 - By moving the things >> from the existing drives he gets back space and only purchases a couple of >> good sized fast SSDs now >> > > It's a possibility. I can get 240GB SSDs in the $120 range so that's > $240 for RAID1. If I take the five existing 500GB drives and > reconfigure for RAID5 that's 2TB. Overall it's not bad going from > 1.4TB to about 2.2TB but being it's not all one big disk I'll likely > never use it all as efficiently. Still, it's an option. > > I do in fact have extra ports: > > c2RAID6 ~ # lspci | grep SATA > 00:1f.2 IDE interface: Intel Corporation 82801JI (ICH10 Family) 4 port > SATA IDE Controller #1 > 00:1f.5 IDE interface: Intel Corporation 82801JI (ICH10 Family) 2 port > SATA IDE Controller #2 > 03:00.0 SATA controller: Marvell Technology Group Ltd. 88SE9123 PCIe > SATA 6.0 Gb/s controller (rev 11) > 06:00.0 SATA controller: JMicron Technology Corp. JMB363 SATA/IDE > Controller (rev 03) > 06:00.1 IDE interface: JMicron Technology Corp. JMB363 SATA/IDE > Controller (rev 03) > c2RAID6 ~ # > > Currently my 5-drive RAID6 uses 5 of the Intel ports. The 6th port > goes to the CD/DVD drive. Some time ago I bought the SATA3 Marvell > card and a smaller (120GB) SSD. I put Gentoo on it and played around a > bit but I've never really used it day-to-day. Part of my 2-drive RAID1 > thinking was that I could build the new RAID1 on the SATA3 controller > not even touch the existing RAID6. If it works reliably on that > controller I'd be done and have 3TB. > > I think David's RAID10 3-drive solution could possibly work if I buy 3 > of the lower cost new WD drives. I'll need to think about that. Not > sure. > > Thanks, > Mark > > > On Sat, May 31, 2014 at 9:40 AM, Craig Curtin <craigc@prosis.com.au> wrote: >> It sounds like the op has additional data ports on his MOBO - wouldn't he be >> better off looking at a couple of SSDs in raid 1 for his OS, swap etc and >> his VMs and then leave the rest for data as raid5 - By moving the things >> from the existing drives he gets back space and only purchases a couple of >> good sized fast SSDs now >> >> >> Sent from my Samsung tablet >> >> . >> >> >> -------- Original message -------- >> From: David Brown >> Date:31/05/2014 21:01 (GMT+10:00) >> To: Mark Knecht ,"L.M.J" >> Cc: Linux-RAID >> Subject: Re: Home desktop/server RAID upgrade >> >> On 30/05/14 22:14, Mark Knecht wrote: >>> On Fri, May 30, 2014 at 12:29 PM, L.M.J <linuxmasterjedi@free.fr> >>> wrote: >>>> Le Fri, 30 May 2014 12:04:07 -0700, Mark Knecht >>>> <markknecht@gmail.com> a écrit : >>>> >>>>> In a RAID1 would a 3-drive Red RAID1 possibly be faster than the >>>>> 2-drive Se RAID1 and at the same time give me more safety? >>>> >>>> Just a question inside the question : how do you manager a RAID1 >>>> with 3 drives ? Maybe you're talking about RAID5 then ? >>> >>> OK, I'm no RAID expert but RAID1 is just drives in parallel right. 2 >>> drives, 3 drives, 4 drives, all holding exactly the same data. In >>> the case of a 3-drive RAID1 - if there is such a beast - I could >>> safely lose 2 drives. You ask a reasonable question though as maybe >>> the way this is actually done is 2 drives + a hot spare in the box >>> that gets sync'ed if and only if one drive fails. Not sure and maybe >>> I'm totally wrong about that. >>> >>> A 3-drive RAID5 would be 2 drives in series - in this case making >>> 6TB - and then the 3rd drive being the redundancy. In the case of a >>> 3-drive RAID5 I could safely lose 1 drive. >>> >>> In my case I don't need more than 3TB, so an option would be a >>> 3-drive RAID5 made out of 2TB drives which would give me 4TB but I >>> don't need the space as much as I want the redundancy and I think >>> RAID5 is slower than RAID1. Additionally some more mdadm RAID >>> knowledgeable people on other lists say Linux mdadm RAID1 would be >>> faster as it will get data from more than one drive at a time. (Or >>> possibly get data from which ever drive returns it the fastest. Not >>> sure.) >>> >>> I believe one good option if I wanted 4 physical drives would be >>> RAID10 but that's getting more complicated again which I didn't >>> really want to do. >>> >>> So maybe it is just 2 drives and the 3 drive version isn't even a >>> possibility? Could be. >> >> With 3 drives, you have several possibilities. >> >> Raid5 makes "stripes" across the three drives, with 2 parts holding data >> and one part holding parity to provide redundancy. >> >> Raid1 is commonly called "mirroring", because you get the same data on >> each disk. md raid has no problem making a 3-way mirror, so that each >> disk is identical. This gives you excellent redundancy, and you can >> make three different reads in parallel - but writes have to go to each >> disk, which can be a little slower than using 2 disks. It's not often >> that people need that level of redundancy. >> >> Another option with md raid is the raid10 setups. For many uses, the >> fastest arrangement is raid10,f2. This means there is two copies of all >> your data (f3 would be three copies), with a "far" layout. >> >> <http://en.wikipedia.org/wiki/Linux_MD_RAID_10#LINUX-MD-RAID-10> >> >> With this arrangement, reads are striped across all three disks, which >> is fast for large reads. Small reads can be handled in parallel. Most >> reads while be handled from the outer half of the disk, which is faster >> and needs less head movement - so reading is on average faster than a >> raid0 on the same disks. Small writes are fast, but large writes >> require quite a bit of head movement to get everything written twice to >> different parts of the disks. >> >> The "best" option always depends on your needs - how you want to access >> your files. A layout geared to fast striped reads of large files will >> be poorer for parallel small writes, and vice versa. raid10,f2 is often >> the best choice for a desktop or small system - but it is not very >> flexible if you later want to add new disks or replace the disks with >> bigger ones. >> >> md raid is flexible enough that it will even let you make a 3 disk raid6 >> array if you want - but a 3-way raid1 mirror will give you the same disk >> space and much better performance. >> -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html ^ permalink raw reply [flat|nested] 13+ messages in thread
* Re: Home desktop/server RAID upgrade 2014-06-01 15:06 ` David Brown @ 2014-06-01 15:59 ` Mark Knecht 2014-06-02 23:04 ` David Brown 2014-06-04 12:29 ` Brad Campbell 0 siblings, 2 replies; 13+ messages in thread From: Mark Knecht @ 2014-06-01 15:59 UTC (permalink / raw) To: David Brown; +Cc: Craig Curtin, L.M.J, Linux-RAID David, You are correct and I'm sorry I didn't do that. I started this question on a Gentoo list where I put a lot more information about the machine/ When I came here I should have included more. The machine is used 7 days a week. I'm self employed writing software analyzing the stock & futures markets. Most of it is written in R in Linux, some of it in proprietary languages in Windows. Some of it is quite computational but mostly it's just looking at a _lot_ of locally stored financial data. Almost all financial data is currently stored on the machine in Linux in ext4. Over the past year this data has been growing at around 30GB/month. With 100GB left on my current RAID6 I don't have much time before I'm full. When I'm actually trading in the market I have a few Virtualbox VMs running Windows 7. They aren't overly large in terms of disk space. (Currently about 150GB total.) The VMs are each stored in massive single files which I suspect basically represent a hard drive to Virtualbox. I have no idea what size any IO might be coming from the VM. The financial data in the previous paragraph is available to these Windows VMs as a network mount from the Windows perspective. Read & write speeds of this data in Windows is not overly high. These VMs are the area where my current RAID6 (5 drive, 16k chunk size) seems to have been a bad decision. The machine is powered off every night. Loading these VMs takes at least 10-15 minutes each morning where I see disk activity lights just grinding away the whole time. If I had a single _performance_ goal in upgrading the disks it would be to improve this significantly. Craig's SSD RAID1 suggestion would certainly help here but at 240GB there wouldn't be a lot of room left. That may be OK though. The last area is video storage. Write speed is unimportant, read speeds are quite low. Over time I hope to migrate it off to a NAS box but for now this is where it's stored. This is currently using about 1/2 the storage my RAID6 provides. Most important to me is data safety. I currently do weekly rotational backups to a couple of USB drives. I have no real-time issues at all if the machine goes down. I have 2 other machines I can do day-to-day work on while I fix this machine. What I am most concerned about is not losing anything more than a couple of previous days work. If I took a week to rebuild the machine after a failure it's pretty much a non-issue to me. Thanks, Mark On Sun, Jun 1, 2014 at 8:06 AM, David Brown <david.brown@hesbynett.no> wrote: > Hi Mark, > > What would be really useful here is a description of what you actually > /want/. What do you want to do with these drives? What sort of files are > they - big or small? Do you need fast access for large files? Do you need > fast access for many files in parallel? How important is the data? How > important is uptime? What sort of backups do you have? What will the > future be like - are you making one big system to last for the foreseeable > future, or do you need something that can easily be expanded? Are you > looking for "fun, interesting and modern" or "boring but well-tested" > solutions? > > Then you need to make a list of the hardware you have, or the budget for new > hardware. > > Without know at least roughly what you are looking for, it's easy to end up > with expensive SSDs because they are "cool", even though you might get more > speed for your money with a couple of slow rust disks and a bit more ram in > your system. It may be that there is no need for any sort of raid at all - > perhaps one big main disk is fine, and the rest of the money spent on a > backup disk (possibly external) with rsync'd copies of your data. This > would mean longer downtime if your main disk failed - but it also gives some > protection against user error. > > And perhaps btrfs with raid1 would be the best choice. > > A raid10,f2 is often the best choice for desktops or workstations with 2 or > 3 hard disks, but it is not necessarily /the/ best choice. > > mvh., > > David > > > > On 01/06/14 16:25, Mark Knecht wrote: >> >> Hi Craig, >> Responding to both you and David Brown. Thanks for your ideas. >> >> - Mark >> >> On Sat, May 31, 2014 at 9:40 AM, Craig Curtin <craigc@prosis.com.au> >> wrote: >>> >>> It sounds like the op has additional data ports on his MOBO - wouldn't he >>> be >>> better off looking at a couple of SSDs in raid 1 for his OS, swap etc and >>> his VMs and then leave the rest for data as raid5 - By moving the things >>> from the existing drives he gets back space and only purchases a couple >>> of >>> good sized fast SSDs now >>> >> >> It's a possibility. I can get 240GB SSDs in the $120 range so that's >> $240 for RAID1. If I take the five existing 500GB drives and >> reconfigure for RAID5 that's 2TB. Overall it's not bad going from >> 1.4TB to about 2.2TB but being it's not all one big disk I'll likely >> never use it all as efficiently. Still, it's an option. >> >> I do in fact have extra ports: >> >> c2RAID6 ~ # lspci | grep SATA >> 00:1f.2 IDE interface: Intel Corporation 82801JI (ICH10 Family) 4 port >> SATA IDE Controller #1 >> 00:1f.5 IDE interface: Intel Corporation 82801JI (ICH10 Family) 2 port >> SATA IDE Controller #2 >> 03:00.0 SATA controller: Marvell Technology Group Ltd. 88SE9123 PCIe >> SATA 6.0 Gb/s controller (rev 11) >> 06:00.0 SATA controller: JMicron Technology Corp. JMB363 SATA/IDE >> Controller (rev 03) >> 06:00.1 IDE interface: JMicron Technology Corp. JMB363 SATA/IDE >> Controller (rev 03) >> c2RAID6 ~ # >> >> Currently my 5-drive RAID6 uses 5 of the Intel ports. The 6th port >> goes to the CD/DVD drive. Some time ago I bought the SATA3 Marvell >> card and a smaller (120GB) SSD. I put Gentoo on it and played around a >> bit but I've never really used it day-to-day. Part of my 2-drive RAID1 >> thinking was that I could build the new RAID1 on the SATA3 controller >> not even touch the existing RAID6. If it works reliably on that >> controller I'd be done and have 3TB. >> >> I think David's RAID10 3-drive solution could possibly work if I buy 3 >> of the lower cost new WD drives. I'll need to think about that. Not >> sure. >> >> Thanks, >> Mark >> >> >> On Sat, May 31, 2014 at 9:40 AM, Craig Curtin <craigc@prosis.com.au> >> wrote: >>> >>> It sounds like the op has additional data ports on his MOBO - wouldn't he >>> be >>> better off looking at a couple of SSDs in raid 1 for his OS, swap etc and >>> his VMs and then leave the rest for data as raid5 - By moving the things >>> from the existing drives he gets back space and only purchases a couple >>> of >>> good sized fast SSDs now >>> >>> >>> Sent from my Samsung tablet >>> >>> . >>> >>> >>> -------- Original message -------- >>> From: David Brown >>> Date:31/05/2014 21:01 (GMT+10:00) >>> To: Mark Knecht ,"L.M.J" >>> Cc: Linux-RAID >>> Subject: Re: Home desktop/server RAID upgrade >>> >>> On 30/05/14 22:14, Mark Knecht wrote: >>>> >>>> On Fri, May 30, 2014 at 12:29 PM, L.M.J <linuxmasterjedi@free.fr> >>>> wrote: >>>>> >>>>> Le Fri, 30 May 2014 12:04:07 -0700, Mark Knecht >>>>> <markknecht@gmail.com> a écrit : >>>>> >>>>>> In a RAID1 would a 3-drive Red RAID1 possibly be faster than the >>>>>> 2-drive Se RAID1 and at the same time give me more safety? >>>>> >>>>> >>>>> Just a question inside the question : how do you manager a RAID1 >>>>> with 3 drives ? Maybe you're talking about RAID5 then ? >>>> >>>> >>>> OK, I'm no RAID expert but RAID1 is just drives in parallel right. 2 >>>> drives, 3 drives, 4 drives, all holding exactly the same data. In >>>> the case of a 3-drive RAID1 - if there is such a beast - I could >>>> safely lose 2 drives. You ask a reasonable question though as maybe >>>> the way this is actually done is 2 drives + a hot spare in the box >>>> that gets sync'ed if and only if one drive fails. Not sure and maybe >>>> I'm totally wrong about that. >>>> >>>> A 3-drive RAID5 would be 2 drives in series - in this case making >>>> 6TB - and then the 3rd drive being the redundancy. In the case of a >>>> 3-drive RAID5 I could safely lose 1 drive. >>>> >>>> In my case I don't need more than 3TB, so an option would be a >>>> 3-drive RAID5 made out of 2TB drives which would give me 4TB but I >>>> don't need the space as much as I want the redundancy and I think >>>> RAID5 is slower than RAID1. Additionally some more mdadm RAID >>>> knowledgeable people on other lists say Linux mdadm RAID1 would be >>>> faster as it will get data from more than one drive at a time. (Or >>>> possibly get data from which ever drive returns it the fastest. Not >>>> sure.) >>>> >>>> I believe one good option if I wanted 4 physical drives would be >>>> RAID10 but that's getting more complicated again which I didn't >>>> really want to do. >>>> >>>> So maybe it is just 2 drives and the 3 drive version isn't even a >>>> possibility? Could be. >>> >>> >>> With 3 drives, you have several possibilities. >>> >>> Raid5 makes "stripes" across the three drives, with 2 parts holding data >>> and one part holding parity to provide redundancy. >>> >>> Raid1 is commonly called "mirroring", because you get the same data on >>> each disk. md raid has no problem making a 3-way mirror, so that each >>> disk is identical. This gives you excellent redundancy, and you can >>> make three different reads in parallel - but writes have to go to each >>> disk, which can be a little slower than using 2 disks. It's not often >>> that people need that level of redundancy. >>> >>> Another option with md raid is the raid10 setups. For many uses, the >>> fastest arrangement is raid10,f2. This means there is two copies of all >>> your data (f3 would be three copies), with a "far" layout. >>> >>> <http://en.wikipedia.org/wiki/Linux_MD_RAID_10#LINUX-MD-RAID-10> >>> >>> With this arrangement, reads are striped across all three disks, which >>> is fast for large reads. Small reads can be handled in parallel. Most >>> reads while be handled from the outer half of the disk, which is faster >>> and needs less head movement - so reading is on average faster than a >>> raid0 on the same disks. Small writes are fast, but large writes >>> require quite a bit of head movement to get everything written twice to >>> different parts of the disks. >>> >>> The "best" option always depends on your needs - how you want to access >>> your files. A layout geared to fast striped reads of large files will >>> be poorer for parallel small writes, and vice versa. raid10,f2 is often >>> the best choice for a desktop or small system - but it is not very >>> flexible if you later want to add new disks or replace the disks with >>> bigger ones. >>> >>> md raid is flexible enough that it will even let you make a 3 disk raid6 >>> array if you want - but a 3-way raid1 mirror will give you the same disk >>> space and much better performance. >>> > -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html ^ permalink raw reply [flat|nested] 13+ messages in thread
* Re: Home desktop/server RAID upgrade 2014-06-01 15:59 ` Mark Knecht @ 2014-06-02 23:04 ` David Brown [not found] ` <E78FE8BDBAD07C43A60163E7D1716EEC01839CFA3D@PROSIS-W2K8-1.prosis.local> 2014-06-04 12:29 ` Brad Campbell 1 sibling, 1 reply; 13+ messages in thread From: David Brown @ 2014-06-02 23:04 UTC (permalink / raw) To: Mark Knecht; +Cc: Craig Curtin, L.M.J, Linux-RAID Hi Mark, I would say forget the SSD's - they are not ideal for VM files, and I don't think they would be worth the cost. Raid 10 (any arrangement) is likely to give the best speed for such files, and would do a lot better than raid 6. Raid 10,f2 is probably a good choice - but you might want to test things out a bit if that is possible. I don't know how much ram you've got in the machine, but if you can afford more, it will always help (especially if you make sure the VM's use the host's cache rather than direct writes). mvh., David On 01/06/14 17:59, Mark Knecht wrote: > David, > You are correct and I'm sorry I didn't do that. I started this > question on a Gentoo list where I put a lot more information about the > machine/ When I came here I should have included more. > > The machine is used 7 days a week. I'm self employed writing > software analyzing the stock & futures markets. Most of it is written > in R in Linux, some of it in proprietary languages in Windows. Some of > it is quite computational but mostly it's just looking at a _lot_ of > locally stored financial data. Almost all financial data is currently > stored on the machine in Linux in ext4. Over the past year this data > has been growing at around 30GB/month. With 100GB left on my current > RAID6 I don't have much time before I'm full. > > When I'm actually trading in the market I have a few Virtualbox VMs > running Windows 7. They aren't overly large in terms of disk space. > (Currently about 150GB total.) The VMs are each stored in massive > single files which I suspect basically represent a hard drive to > Virtualbox. I have no idea what size any IO might be coming from the > VM. The financial data in the previous paragraph is available to these > Windows VMs as a network mount from the Windows perspective. Read & > write speeds of this data in Windows is not overly high. > > These VMs are the area where my current RAID6 (5 drive, 16k chunk > size) seems to have been a bad decision. The machine is powered off > every night. Loading these VMs takes at least 10-15 minutes each > morning where I see disk activity lights just grinding away the whole > time. If I had a single _performance_ goal in upgrading the disks it > would be to improve this significantly. Craig's SSD RAID1 suggestion > would certainly help here but at 240GB there wouldn't be a lot of room > left. That may be OK though. > > The last area is video storage. Write speed is unimportant, read > speeds are quite low. Over time I hope to migrate it off to a NAS box > but for now this is where it's stored. This is currently using about > 1/2 the storage my RAID6 provides. > > Most important to me is data safety. I currently do weekly > rotational backups to a couple of USB drives. I have no real-time > issues at all if the machine goes down. I have 2 other machines I can > do day-to-day work on while I fix this machine. What I am most > concerned about is not losing anything more than a couple of previous > days work. If I took a week to rebuild the machine after a failure > it's pretty much a non-issue to me. > > Thanks, > Mark > > On Sun, Jun 1, 2014 at 8:06 AM, David Brown <david.brown@hesbynett.no> wrote: >> Hi Mark, >> >> What would be really useful here is a description of what you actually >> /want/. What do you want to do with these drives? What sort of files are >> they - big or small? Do you need fast access for large files? Do you need >> fast access for many files in parallel? How important is the data? How >> important is uptime? What sort of backups do you have? What will the >> future be like - are you making one big system to last for the foreseeable >> future, or do you need something that can easily be expanded? Are you >> looking for "fun, interesting and modern" or "boring but well-tested" >> solutions? >> >> Then you need to make a list of the hardware you have, or the budget for new >> hardware. >> >> Without know at least roughly what you are looking for, it's easy to end up >> with expensive SSDs because they are "cool", even though you might get more >> speed for your money with a couple of slow rust disks and a bit more ram in >> your system. It may be that there is no need for any sort of raid at all - >> perhaps one big main disk is fine, and the rest of the money spent on a >> backup disk (possibly external) with rsync'd copies of your data. This >> would mean longer downtime if your main disk failed - but it also gives some >> protection against user error. >> >> And perhaps btrfs with raid1 would be the best choice. >> >> A raid10,f2 is often the best choice for desktops or workstations with 2 or >> 3 hard disks, but it is not necessarily /the/ best choice. >> >> mvh., >> >> David >> >> >> >> On 01/06/14 16:25, Mark Knecht wrote: >>> >>> Hi Craig, >>> Responding to both you and David Brown. Thanks for your ideas. >>> >>> - Mark >>> >>> On Sat, May 31, 2014 at 9:40 AM, Craig Curtin <craigc@prosis.com.au> >>> wrote: >>>> >>>> It sounds like the op has additional data ports on his MOBO - wouldn't he >>>> be >>>> better off looking at a couple of SSDs in raid 1 for his OS, swap etc and >>>> his VMs and then leave the rest for data as raid5 - By moving the things >>>> from the existing drives he gets back space and only purchases a couple >>>> of >>>> good sized fast SSDs now >>>> >>> >>> It's a possibility. I can get 240GB SSDs in the $120 range so that's >>> $240 for RAID1. If I take the five existing 500GB drives and >>> reconfigure for RAID5 that's 2TB. Overall it's not bad going from >>> 1.4TB to about 2.2TB but being it's not all one big disk I'll likely >>> never use it all as efficiently. Still, it's an option. >>> >>> I do in fact have extra ports: >>> >>> c2RAID6 ~ # lspci | grep SATA >>> 00:1f.2 IDE interface: Intel Corporation 82801JI (ICH10 Family) 4 port >>> SATA IDE Controller #1 >>> 00:1f.5 IDE interface: Intel Corporation 82801JI (ICH10 Family) 2 port >>> SATA IDE Controller #2 >>> 03:00.0 SATA controller: Marvell Technology Group Ltd. 88SE9123 PCIe >>> SATA 6.0 Gb/s controller (rev 11) >>> 06:00.0 SATA controller: JMicron Technology Corp. JMB363 SATA/IDE >>> Controller (rev 03) >>> 06:00.1 IDE interface: JMicron Technology Corp. JMB363 SATA/IDE >>> Controller (rev 03) >>> c2RAID6 ~ # >>> >>> Currently my 5-drive RAID6 uses 5 of the Intel ports. The 6th port >>> goes to the CD/DVD drive. Some time ago I bought the SATA3 Marvell >>> card and a smaller (120GB) SSD. I put Gentoo on it and played around a >>> bit but I've never really used it day-to-day. Part of my 2-drive RAID1 >>> thinking was that I could build the new RAID1 on the SATA3 controller >>> not even touch the existing RAID6. If it works reliably on that >>> controller I'd be done and have 3TB. >>> >>> I think David's RAID10 3-drive solution could possibly work if I buy 3 >>> of the lower cost new WD drives. I'll need to think about that. Not >>> sure. >>> >>> Thanks, >>> Mark >>> >>> >>> On Sat, May 31, 2014 at 9:40 AM, Craig Curtin <craigc@prosis.com.au> >>> wrote: >>>> >>>> It sounds like the op has additional data ports on his MOBO - wouldn't he >>>> be >>>> better off looking at a couple of SSDs in raid 1 for his OS, swap etc and >>>> his VMs and then leave the rest for data as raid5 - By moving the things >>>> from the existing drives he gets back space and only purchases a couple >>>> of >>>> good sized fast SSDs now >>>> >>>> >>>> Sent from my Samsung tablet >>>> >>>> . >>>> >>>> >>>> -------- Original message -------- >>>> From: David Brown >>>> Date:31/05/2014 21:01 (GMT+10:00) >>>> To: Mark Knecht ,"L.M.J" >>>> Cc: Linux-RAID >>>> Subject: Re: Home desktop/server RAID upgrade >>>> >>>> On 30/05/14 22:14, Mark Knecht wrote: >>>>> >>>>> On Fri, May 30, 2014 at 12:29 PM, L.M.J <linuxmasterjedi@free.fr> >>>>> wrote: >>>>>> >>>>>> Le Fri, 30 May 2014 12:04:07 -0700, Mark Knecht >>>>>> <markknecht@gmail.com> a écrit : >>>>>> >>>>>>> In a RAID1 would a 3-drive Red RAID1 possibly be faster than the >>>>>>> 2-drive Se RAID1 and at the same time give me more safety? >>>>>> >>>>>> >>>>>> Just a question inside the question : how do you manager a RAID1 >>>>>> with 3 drives ? Maybe you're talking about RAID5 then ? >>>>> >>>>> >>>>> OK, I'm no RAID expert but RAID1 is just drives in parallel right. 2 >>>>> drives, 3 drives, 4 drives, all holding exactly the same data. In >>>>> the case of a 3-drive RAID1 - if there is such a beast - I could >>>>> safely lose 2 drives. You ask a reasonable question though as maybe >>>>> the way this is actually done is 2 drives + a hot spare in the box >>>>> that gets sync'ed if and only if one drive fails. Not sure and maybe >>>>> I'm totally wrong about that. >>>>> >>>>> A 3-drive RAID5 would be 2 drives in series - in this case making >>>>> 6TB - and then the 3rd drive being the redundancy. In the case of a >>>>> 3-drive RAID5 I could safely lose 1 drive. >>>>> >>>>> In my case I don't need more than 3TB, so an option would be a >>>>> 3-drive RAID5 made out of 2TB drives which would give me 4TB but I >>>>> don't need the space as much as I want the redundancy and I think >>>>> RAID5 is slower than RAID1. Additionally some more mdadm RAID >>>>> knowledgeable people on other lists say Linux mdadm RAID1 would be >>>>> faster as it will get data from more than one drive at a time. (Or >>>>> possibly get data from which ever drive returns it the fastest. Not >>>>> sure.) >>>>> >>>>> I believe one good option if I wanted 4 physical drives would be >>>>> RAID10 but that's getting more complicated again which I didn't >>>>> really want to do. >>>>> >>>>> So maybe it is just 2 drives and the 3 drive version isn't even a >>>>> possibility? Could be. >>>> >>>> >>>> With 3 drives, you have several possibilities. >>>> >>>> Raid5 makes "stripes" across the three drives, with 2 parts holding data >>>> and one part holding parity to provide redundancy. >>>> >>>> Raid1 is commonly called "mirroring", because you get the same data on >>>> each disk. md raid has no problem making a 3-way mirror, so that each >>>> disk is identical. This gives you excellent redundancy, and you can >>>> make three different reads in parallel - but writes have to go to each >>>> disk, which can be a little slower than using 2 disks. It's not often >>>> that people need that level of redundancy. >>>> >>>> Another option with md raid is the raid10 setups. For many uses, the >>>> fastest arrangement is raid10,f2. This means there is two copies of all >>>> your data (f3 would be three copies), with a "far" layout. >>>> >>>> <http://en.wikipedia.org/wiki/Linux_MD_RAID_10#LINUX-MD-RAID-10> >>>> >>>> With this arrangement, reads are striped across all three disks, which >>>> is fast for large reads. Small reads can be handled in parallel. Most >>>> reads while be handled from the outer half of the disk, which is faster >>>> and needs less head movement - so reading is on average faster than a >>>> raid0 on the same disks. Small writes are fast, but large writes >>>> require quite a bit of head movement to get everything written twice to >>>> different parts of the disks. >>>> >>>> The "best" option always depends on your needs - how you want to access >>>> your files. A layout geared to fast striped reads of large files will >>>> be poorer for parallel small writes, and vice versa. raid10,f2 is often >>>> the best choice for a desktop or small system - but it is not very >>>> flexible if you later want to add new disks or replace the disks with >>>> bigger ones. >>>> >>>> md raid is flexible enough that it will even let you make a 3 disk raid6 >>>> array if you want - but a 3-way raid1 mirror will give you the same disk >>>> space and much better performance. >>>> >> -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html ^ permalink raw reply [flat|nested] 13+ messages in thread
[parent not found: <E78FE8BDBAD07C43A60163E7D1716EEC01839CFA3D@PROSIS-W2K8-1.prosis.local>]
* Re: Home desktop/server RAID upgrade [not found] ` <E78FE8BDBAD07C43A60163E7D1716EEC01839CFA3D@PROSIS-W2K8-1.prosis.local> @ 2014-06-03 7:58 ` David Brown 2014-06-03 14:59 ` Roberto Spadim 0 siblings, 1 reply; 13+ messages in thread From: David Brown @ 2014-06-03 7:58 UTC (permalink / raw) To: Craig Curtin, Mark Knecht; +Cc: L.M.J, Linux-RAID Hi, It's not that SSD's are /bad/ for VM images - it is simply that they are not so much better than HD's that they are worth the money. VM files are big - it costs a lot to get that much SSD space, so you have to be sure that the faster IOPs and faster random access is actually worth that cost. SSD's are not significantly faster in bandwidth than HD's - for the price of one high throughput SSD, you can buy four HD's in raid10,f2 with a higher throughput. The OP needs HD's for the space. So the question is whether he should spend additional money on two good quality SSDs - or should he spend it on an extra HD, more ram, and perhaps a little UPS? (I'm assuming he has a limited budget.) I don't think the IOPs rate of SSDs will make such a difference over the layers of indirection - Windows on the VM's, the VM's disk caching system, the VM image file format, the caches on the host ram, the raid layers, etc. These all conspire to add latency and reduce the peak IOPs - within the VM, you are never going to see anything like the SSD's theoretical IOPs rate. You will get a little higher IOPs than with HD's at the back end, but not much more. The VM's will see high IOP's if and only if the data is in ram cache somewhere, regardless of the disk type - so more ram will always help. Of course, there are other reasons you might prefer SSD's - size, space, power, noise, reliability, etc. mvh., David On 03/06/14 01:13, Craig Curtin wrote: > Dave, > > What part of a VM is not ideally suited to running from SSDs. The > right SSDs support a high level of IOPS (much higher sustained that > any SATA based RAID array is going to get to) and has he has > predefined (preallocated/thick) disks already defined for the VMs > they are ideal candidates to move onto SSDs. > > As a real world example - I have 4 HP N40L microservers running in a > VMware Cluster at home - they all source their VMs from another N40L > that has a HP P410 RAID controller in it and dual gigabit Ethernet > ports. > > The box running as the disk store is running Centos 6.3. > > It has two RAID sets defined on the P410 - a pair of Samsung EVO > 240GB SSDs in RAID 1 and 4 x WD (Enterprise Series) 500GB SATA drives > in RAID0+1 > > I can categorically state that the throughput from the SSD VMs is > approx. 4 times more than I can sustain to the SATA drives - the SATA > drives come out at around 1/2 the throughput of a single Gigabit card > whilst the SSDs flood both channels of the card. You can also see the > point where the cache on the controller is flooded when writing to > the SATA drives as everything slows down - whereas with the SSDs this > never happens. This is doing disk intensive operations like live > state migrating VMs etc. > > Craig > > -----Original Message----- From: linux-raid-owner@vger.kernel.org > [mailto:linux-raid-owner@vger.kernel.org] On Behalf Of David Brown > Sent: Tuesday, 3 June 2014 9:05 AM To: Mark Knecht Cc: Craig Curtin; > L.M.J; Linux-RAID Subject: Re: Home desktop/server RAID upgrade > > Hi Mark, > > I would say forget the SSD's - they are not ideal for VM files, and I > don't think they would be worth the cost. Raid 10 (any arrangement) > is likely to give the best speed for such files, and would do a lot > better than raid 6. Raid 10,f2 is probably a good choice - but you > might want to test things out a bit if that is possible. > > I don't know how much ram you've got in the machine, but if you can > afford more, it will always help (especially if you make sure the > VM's use the host's cache rather than direct writes). > > mvh., > > David > > > On 01/06/14 17:59, Mark Knecht wrote: >> David, You are correct and I'm sorry I didn't do that. I started >> this question on a Gentoo list where I put a lot more information >> about the machine/ When I came here I should have included more. >> >> The machine is used 7 days a week. I'm self employed writing >> software analyzing the stock & futures markets. Most of it is >> written in R in Linux, some of it in proprietary languages in >> Windows. Some of it is quite computational but mostly it's just >> looking at a _lot_ of locally stored financial data. Almost all >> financial data is currently stored on the machine in Linux in ext4. >> Over the past year this data has been growing at around 30GB/month. >> With 100GB left on my current RAID6 I don't have much time before >> I'm full. >> >> When I'm actually trading in the market I have a few Virtualbox VMs >> running Windows 7. They aren't overly large in terms of disk >> space. (Currently about 150GB total.) The VMs are each stored in >> massive single files which I suspect basically represent a hard >> drive to Virtualbox. I have no idea what size any IO might be >> coming from the VM. The financial data in the previous paragraph is >> available to these Windows VMs as a network mount from the Windows >> perspective. Read & write speeds of this data in Windows is not >> overly high. >> >> These VMs are the area where my current RAID6 (5 drive, 16k chunk >> size) seems to have been a bad decision. The machine is powered >> off every night. Loading these VMs takes at least 10-15 minutes >> each morning where I see disk activity lights just grinding away >> the whole time. If I had a single _performance_ goal in upgrading >> the disks it would be to improve this significantly. Craig's SSD >> RAID1 suggestion would certainly help here but at 240GB there >> wouldn't be a lot of room left. That may be OK though. >> >> The last area is video storage. Write speed is unimportant, read >> speeds are quite low. Over time I hope to migrate it off to a NAS >> box but for now this is where it's stored. This is currently using >> about 1/2 the storage my RAID6 provides. >> >> Most important to me is data safety. I currently do weekly >> rotational backups to a couple of USB drives. I have no real-time >> issues at all if the machine goes down. I have 2 other machines I >> can do day-to-day work on while I fix this machine. What I am most >> concerned about is not losing anything more than a couple of >> previous days work. If I took a week to rebuild the machine after >> a failure it's pretty much a non-issue to me. >> >> Thanks, Mark >> >> On Sun, Jun 1, 2014 at 8:06 AM, David Brown >> <david.brown@hesbynett.no> wrote: >>> Hi Mark, >>> >>> What would be really useful here is a description of what you >>> actually /want/. What do you want to do with these drives? >>> What sort of files are they - big or small? Do you need fast >>> access for large files? Do you need fast access for many files >>> in parallel? How important is the data? How important is uptime? >>> What sort of backups do you have? What will the future be like - >>> are you making one big system to last for the foreseeable future, >>> or do you need something that can easily be expanded? Are you >>> looking for "fun, interesting and modern" or "boring but >>> well-tested" solutions? >>> >>> Then you need to make a list of the hardware you have, or the >>> budget for new hardware. >>> >>> Without know at least roughly what you are looking for, it's easy >>> to end up with expensive SSDs because they are "cool", even >>> though you might get more speed for your money with a couple of >>> slow rust disks and a bit more ram in your system. It may be >>> that there is no need for any sort of raid at all - perhaps one >>> big main disk is fine, and the rest of the money spent on a >>> backup disk (possibly external) with rsync'd copies of your data. >>> This would mean longer downtime if your main disk failed - but it >>> also gives some protection against user error. >>> >>> And perhaps btrfs with raid1 would be the best choice. >>> >>> A raid10,f2 is often the best choice for desktops or >>> workstations with 2 or 3 hard disks, but it is not necessarily >>> /the/ best choice. >>> >>> mvh., >>> >>> David >>> >>> >>> >>> On 01/06/14 16:25, Mark Knecht wrote: >>>> >>>> Hi Craig, Responding to both you and David Brown. Thanks for >>>> your ideas. >>>> >>>> - Mark >>>> >>>> On Sat, May 31, 2014 at 9:40 AM, Craig Curtin >>>> <craigc@prosis.com.au> wrote: >>>>> >>>>> It sounds like the op has additional data ports on his MOBO >>>>> - wouldn't he be better off looking at a couple of SSDs in >>>>> raid 1 for his OS, swap etc and his VMs and then leave the >>>>> rest for data as raid5 - By moving the things from the >>>>> existing drives he gets back space and only purchases a >>>>> couple of good sized fast SSDs now >>>>> >>>> >>>> It's a possibility. I can get 240GB SSDs in the $120 range so >>>> that's $240 for RAID1. If I take the five existing 500GB drives >>>> and reconfigure for RAID5 that's 2TB. Overall it's not bad >>>> going from 1.4TB to about 2.2TB but being it's not all one big >>>> disk I'll likely never use it all as efficiently. Still, it's >>>> an option. >>>> >>>> I do in fact have extra ports: >>>> >>>> c2RAID6 ~ # lspci | grep SATA 00:1f.2 IDE interface: Intel >>>> Corporation 82801JI (ICH10 Family) 4 port SATA IDE Controller >>>> #1 00:1f.5 IDE interface: Intel Corporation 82801JI (ICH10 >>>> Family) 2 port SATA IDE Controller #2 03:00.0 SATA controller: >>>> Marvell Technology Group Ltd. 88SE9123 PCIe SATA 6.0 Gb/s >>>> controller (rev 11) 06:00.0 SATA controller: JMicron Technology >>>> Corp. JMB363 SATA/IDE Controller (rev 03) 06:00.1 IDE >>>> interface: JMicron Technology Corp. JMB363 SATA/IDE Controller >>>> (rev 03) c2RAID6 ~ # >>>> >>>> Currently my 5-drive RAID6 uses 5 of the Intel ports. The 6th >>>> port goes to the CD/DVD drive. Some time ago I bought the SATA3 >>>> Marvell card and a smaller (120GB) SSD. I put Gentoo on it and >>>> played around a bit but I've never really used it day-to-day. >>>> Part of my 2-drive RAID1 thinking was that I could build the >>>> new RAID1 on the SATA3 controller not even touch the existing >>>> RAID6. If it works reliably on that controller I'd be done and >>>> have 3TB. >>>> >>>> I think David's RAID10 3-drive solution could possibly work if >>>> I buy 3 of the lower cost new WD drives. I'll need to think >>>> about that. Not sure. >>>> >>>> Thanks, Mark >>>> >>>> >>>> On Sat, May 31, 2014 at 9:40 AM, Craig Curtin >>>> <craigc@prosis.com.au> wrote: >>>>> >>>>> It sounds like the op has additional data ports on his MOBO >>>>> - wouldn't he be better off looking at a couple of SSDs in >>>>> raid 1 for his OS, swap etc and his VMs and then leave the >>>>> rest for data as raid5 - By moving the things from the >>>>> existing drives he gets back space and only purchases a >>>>> couple of good sized fast SSDs now >>>>> >>>>> >>>>> Sent from my Samsung tablet >>>>> >>>>> . >>>>> >>>>> >>>>> -------- Original message -------- From: David Brown >>>>> Date:31/05/2014 21:01 (GMT+10:00) To: Mark Knecht ,"L.M.J" >>>>> Cc: Linux-RAID Subject: Re: Home desktop/server RAID upgrade >>>>> >>>>> On 30/05/14 22:14, Mark Knecht wrote: >>>>>> >>>>>> On Fri, May 30, 2014 at 12:29 PM, L.M.J >>>>>> <linuxmasterjedi@free.fr> wrote: >>>>>>> >>>>>>> Le Fri, 30 May 2014 12:04:07 -0700, Mark Knecht >>>>>>> <markknecht@gmail.com> a écrit : >>>>>>> >>>>>>>> In a RAID1 would a 3-drive Red RAID1 possibly be faster >>>>>>>> than the 2-drive Se RAID1 and at the same time give me >>>>>>>> more safety? >>>>>>> >>>>>>> >>>>>>> Just a question inside the question : how do you manager >>>>>>> a RAID1 with 3 drives ? Maybe you're talking about RAID5 >>>>>>> then ? >>>>>> >>>>>> >>>>>> OK, I'm no RAID expert but RAID1 is just drives in parallel >>>>>> right. 2 drives, 3 drives, 4 drives, all holding exactly >>>>>> the same data. In the case of a 3-drive RAID1 - if there is >>>>>> such a beast - I could safely lose 2 drives. You ask a >>>>>> reasonable question though as maybe the way this is >>>>>> actually done is 2 drives + a hot spare in the box that >>>>>> gets sync'ed if and only if one drive fails. Not sure and >>>>>> maybe I'm totally wrong about that. >>>>>> >>>>>> A 3-drive RAID5 would be 2 drives in series - in this case >>>>>> making 6TB - and then the 3rd drive being the redundancy. >>>>>> In the case of a 3-drive RAID5 I could safely lose 1 >>>>>> drive. >>>>>> >>>>>> In my case I don't need more than 3TB, so an option would >>>>>> be a 3-drive RAID5 made out of 2TB drives which would give >>>>>> me 4TB but I don't need the space as much as I want the >>>>>> redundancy and I think RAID5 is slower than RAID1. >>>>>> Additionally some more mdadm RAID knowledgeable people on >>>>>> other lists say Linux mdadm RAID1 would be faster as it >>>>>> will get data from more than one drive at a time. (Or >>>>>> possibly get data from which ever drive returns it the >>>>>> fastest. Not sure.) >>>>>> >>>>>> I believe one good option if I wanted 4 physical drives >>>>>> would be RAID10 but that's getting more complicated again >>>>>> which I didn't really want to do. >>>>>> >>>>>> So maybe it is just 2 drives and the 3 drive version isn't >>>>>> even a possibility? Could be. >>>>> >>>>> >>>>> With 3 drives, you have several possibilities. >>>>> >>>>> Raid5 makes "stripes" across the three drives, with 2 parts >>>>> holding data and one part holding parity to provide >>>>> redundancy. >>>>> >>>>> Raid1 is commonly called "mirroring", because you get the >>>>> same data on each disk. md raid has no problem making a >>>>> 3-way mirror, so that each disk is identical. This gives you >>>>> excellent redundancy, and you can make three different reads >>>>> in parallel - but writes have to go to each disk, which can >>>>> be a little slower than using 2 disks. It's not often that >>>>> people need that level of redundancy. >>>>> >>>>> Another option with md raid is the raid10 setups. For many >>>>> uses, the fastest arrangement is raid10,f2. This means there >>>>> is two copies of all your data (f3 would be three copies), >>>>> with a "far" layout. >>>>> >>>>> <http://en.wikipedia.org/wiki/Linux_MD_RAID_10#LINUX-MD-RAID-10> >>>>> >>>>> >>>>> With this arrangement, reads are striped across all three disks, >>>>> which is fast for large reads. Small reads can be handled >>>>> in parallel. Most reads while be handled from the outer half >>>>> of the disk, which is faster and needs less head movement - >>>>> so reading is on average faster than a raid0 on the same >>>>> disks. Small writes are fast, but large writes require quite >>>>> a bit of head movement to get everything written twice to >>>>> different parts of the disks. >>>>> >>>>> The "best" option always depends on your needs - how you want >>>>> to access your files. A layout geared to fast striped reads >>>>> of large files will be poorer for parallel small writes, and >>>>> vice versa. raid10,f2 is often the best choice for a desktop >>>>> or small system - but it is not very flexible if you later >>>>> want to add new disks or replace the disks with bigger ones. >>>>> >>>>> md raid is flexible enough that it will even let you make a 3 >>>>> disk raid6 array if you want - but a 3-way raid1 mirror will >>>>> give you the same disk space and much better performance. >>>>> >>> > -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html ^ permalink raw reply [flat|nested] 13+ messages in thread
* Re: Home desktop/server RAID upgrade 2014-06-03 7:58 ` David Brown @ 2014-06-03 14:59 ` Roberto Spadim 0 siblings, 0 replies; 13+ messages in thread From: Roberto Spadim @ 2014-06-03 14:59 UTC (permalink / raw) To: David Brown; +Cc: Craig Curtin, Mark Knecht, L.M.J, Linux-RAID i think i asked it before with brown, a raid card with memory is a nice tool here but maybe... you could use a ssd and a bcache, flashcache, or md cache solution to increase performace of hdds (ok, ram/raid card are faster than ssd, but just another idea) 2014-06-03 4:58 GMT-03:00 David Brown <david.brown@hesbynett.no>: > Hi, > > It's not that SSD's are /bad/ for VM images - it is simply that they are > not so much better than HD's that they are worth the money. VM files > are big - it costs a lot to get that much SSD space, so you have to be > sure that the faster IOPs and faster random access is actually worth > that cost. SSD's are not significantly faster in bandwidth than HD's - > for the price of one high throughput SSD, you can buy four HD's in > raid10,f2 with a higher throughput. > > The OP needs HD's for the space. So the question is whether he should > spend additional money on two good quality SSDs - or should he spend it > on an extra HD, more ram, and perhaps a little UPS? (I'm assuming he > has a limited budget.) > > I don't think the IOPs rate of SSDs will make such a difference over the > layers of indirection - Windows on the VM's, the VM's disk caching > system, the VM image file format, the caches on the host ram, the raid > layers, etc. These all conspire to add latency and reduce the peak IOPs > - within the VM, you are never going to see anything like the SSD's > theoretical IOPs rate. You will get a little higher IOPs than with HD's > at the back end, but not much more. The VM's will see high IOP's if and > only if the data is in ram cache somewhere, regardless of the disk type > - so more ram will always help. > > Of course, there are other reasons you might prefer SSD's - size, space, > power, noise, reliability, etc. > > mvh., > > David > > > > On 03/06/14 01:13, Craig Curtin wrote: >> Dave, >> >> What part of a VM is not ideally suited to running from SSDs. The >> right SSDs support a high level of IOPS (much higher sustained that >> any SATA based RAID array is going to get to) and has he has >> predefined (preallocated/thick) disks already defined for the VMs >> they are ideal candidates to move onto SSDs. >> >> As a real world example - I have 4 HP N40L microservers running in a >> VMware Cluster at home - they all source their VMs from another N40L >> that has a HP P410 RAID controller in it and dual gigabit Ethernet >> ports. >> >> The box running as the disk store is running Centos 6.3. >> >> It has two RAID sets defined on the P410 - a pair of Samsung EVO >> 240GB SSDs in RAID 1 and 4 x WD (Enterprise Series) 500GB SATA drives >> in RAID0+1 >> >> I can categorically state that the throughput from the SSD VMs is >> approx. 4 times more than I can sustain to the SATA drives - the SATA >> drives come out at around 1/2 the throughput of a single Gigabit card >> whilst the SSDs flood both channels of the card. You can also see the >> point where the cache on the controller is flooded when writing to >> the SATA drives as everything slows down - whereas with the SSDs this >> never happens. This is doing disk intensive operations like live >> state migrating VMs etc. >> >> Craig >> >> -----Original Message----- From: linux-raid-owner@vger.kernel.org >> [mailto:linux-raid-owner@vger.kernel.org] On Behalf Of David Brown >> Sent: Tuesday, 3 June 2014 9:05 AM To: Mark Knecht Cc: Craig Curtin; >> L.M.J; Linux-RAID Subject: Re: Home desktop/server RAID upgrade >> >> Hi Mark, >> >> I would say forget the SSD's - they are not ideal for VM files, and I >> don't think they would be worth the cost. Raid 10 (any arrangement) >> is likely to give the best speed for such files, and would do a lot >> better than raid 6. Raid 10,f2 is probably a good choice - but you >> might want to test things out a bit if that is possible. >> >> I don't know how much ram you've got in the machine, but if you can >> afford more, it will always help (especially if you make sure the >> VM's use the host's cache rather than direct writes). >> >> mvh., >> >> David >> >> >> On 01/06/14 17:59, Mark Knecht wrote: >>> David, You are correct and I'm sorry I didn't do that. I started >>> this question on a Gentoo list where I put a lot more information >>> about the machine/ When I came here I should have included more. >>> >>> The machine is used 7 days a week. I'm self employed writing >>> software analyzing the stock & futures markets. Most of it is >>> written in R in Linux, some of it in proprietary languages in >>> Windows. Some of it is quite computational but mostly it's just >>> looking at a _lot_ of locally stored financial data. Almost all >>> financial data is currently stored on the machine in Linux in ext4. >>> Over the past year this data has been growing at around 30GB/month. >>> With 100GB left on my current RAID6 I don't have much time before >>> I'm full. >>> >>> When I'm actually trading in the market I have a few Virtualbox VMs >>> running Windows 7. They aren't overly large in terms of disk >>> space. (Currently about 150GB total.) The VMs are each stored in >>> massive single files which I suspect basically represent a hard >>> drive to Virtualbox. I have no idea what size any IO might be >>> coming from the VM. The financial data in the previous paragraph is >>> available to these Windows VMs as a network mount from the Windows >>> perspective. Read & write speeds of this data in Windows is not >>> overly high. >>> >>> These VMs are the area where my current RAID6 (5 drive, 16k chunk >>> size) seems to have been a bad decision. The machine is powered >>> off every night. Loading these VMs takes at least 10-15 minutes >>> each morning where I see disk activity lights just grinding away >>> the whole time. If I had a single _performance_ goal in upgrading >>> the disks it would be to improve this significantly. Craig's SSD >>> RAID1 suggestion would certainly help here but at 240GB there >>> wouldn't be a lot of room left. That may be OK though. >>> >>> The last area is video storage. Write speed is unimportant, read >>> speeds are quite low. Over time I hope to migrate it off to a NAS >>> box but for now this is where it's stored. This is currently using >>> about 1/2 the storage my RAID6 provides. >>> >>> Most important to me is data safety. I currently do weekly >>> rotational backups to a couple of USB drives. I have no real-time >>> issues at all if the machine goes down. I have 2 other machines I >>> can do day-to-day work on while I fix this machine. What I am most >>> concerned about is not losing anything more than a couple of >>> previous days work. If I took a week to rebuild the machine after >>> a failure it's pretty much a non-issue to me. >>> >>> Thanks, Mark >>> >>> On Sun, Jun 1, 2014 at 8:06 AM, David Brown >>> <david.brown@hesbynett.no> wrote: >>>> Hi Mark, >>>> >>>> What would be really useful here is a description of what you >>>> actually /want/. What do you want to do with these drives? >>>> What sort of files are they - big or small? Do you need fast >>>> access for large files? Do you need fast access for many files >>>> in parallel? How important is the data? How important is uptime? >>>> What sort of backups do you have? What will the future be like - >>>> are you making one big system to last for the foreseeable future, >>>> or do you need something that can easily be expanded? Are you >>>> looking for "fun, interesting and modern" or "boring but >>>> well-tested" solutions? >>>> >>>> Then you need to make a list of the hardware you have, or the >>>> budget for new hardware. >>>> >>>> Without know at least roughly what you are looking for, it's easy >>>> to end up with expensive SSDs because they are "cool", even >>>> though you might get more speed for your money with a couple of >>>> slow rust disks and a bit more ram in your system. It may be >>>> that there is no need for any sort of raid at all - perhaps one >>>> big main disk is fine, and the rest of the money spent on a >>>> backup disk (possibly external) with rsync'd copies of your data. >>>> This would mean longer downtime if your main disk failed - but it >>>> also gives some protection against user error. >>>> >>>> And perhaps btrfs with raid1 would be the best choice. >>>> >>>> A raid10,f2 is often the best choice for desktops or >>>> workstations with 2 or 3 hard disks, but it is not necessarily >>>> /the/ best choice. >>>> >>>> mvh., >>>> >>>> David >>>> >>>> >>>> >>>> On 01/06/14 16:25, Mark Knecht wrote: >>>>> >>>>> Hi Craig, Responding to both you and David Brown. Thanks for >>>>> your ideas. >>>>> >>>>> - Mark >>>>> >>>>> On Sat, May 31, 2014 at 9:40 AM, Craig Curtin >>>>> <craigc@prosis.com.au> wrote: >>>>>> >>>>>> It sounds like the op has additional data ports on his MOBO >>>>>> - wouldn't he be better off looking at a couple of SSDs in >>>>>> raid 1 for his OS, swap etc and his VMs and then leave the >>>>>> rest for data as raid5 - By moving the things from the >>>>>> existing drives he gets back space and only purchases a >>>>>> couple of good sized fast SSDs now >>>>>> >>>>> >>>>> It's a possibility. I can get 240GB SSDs in the $120 range so >>>>> that's $240 for RAID1. If I take the five existing 500GB drives >>>>> and reconfigure for RAID5 that's 2TB. Overall it's not bad >>>>> going from 1.4TB to about 2.2TB but being it's not all one big >>>>> disk I'll likely never use it all as efficiently. Still, it's >>>>> an option. >>>>> >>>>> I do in fact have extra ports: >>>>> >>>>> c2RAID6 ~ # lspci | grep SATA 00:1f.2 IDE interface: Intel >>>>> Corporation 82801JI (ICH10 Family) 4 port SATA IDE Controller >>>>> #1 00:1f.5 IDE interface: Intel Corporation 82801JI (ICH10 >>>>> Family) 2 port SATA IDE Controller #2 03:00.0 SATA controller: >>>>> Marvell Technology Group Ltd. 88SE9123 PCIe SATA 6.0 Gb/s >>>>> controller (rev 11) 06:00.0 SATA controller: JMicron Technology >>>>> Corp. JMB363 SATA/IDE Controller (rev 03) 06:00.1 IDE >>>>> interface: JMicron Technology Corp. JMB363 SATA/IDE Controller >>>>> (rev 03) c2RAID6 ~ # >>>>> >>>>> Currently my 5-drive RAID6 uses 5 of the Intel ports. The 6th >>>>> port goes to the CD/DVD drive. Some time ago I bought the SATA3 >>>>> Marvell card and a smaller (120GB) SSD. I put Gentoo on it and >>>>> played around a bit but I've never really used it day-to-day. >>>>> Part of my 2-drive RAID1 thinking was that I could build the >>>>> new RAID1 on the SATA3 controller not even touch the existing >>>>> RAID6. If it works reliably on that controller I'd be done and >>>>> have 3TB. >>>>> >>>>> I think David's RAID10 3-drive solution could possibly work if >>>>> I buy 3 of the lower cost new WD drives. I'll need to think >>>>> about that. Not sure. >>>>> >>>>> Thanks, Mark >>>>> >>>>> >>>>> On Sat, May 31, 2014 at 9:40 AM, Craig Curtin >>>>> <craigc@prosis.com.au> wrote: >>>>>> >>>>>> It sounds like the op has additional data ports on his MOBO >>>>>> - wouldn't he be better off looking at a couple of SSDs in >>>>>> raid 1 for his OS, swap etc and his VMs and then leave the >>>>>> rest for data as raid5 - By moving the things from the >>>>>> existing drives he gets back space and only purchases a >>>>>> couple of good sized fast SSDs now >>>>>> >>>>>> >>>>>> Sent from my Samsung tablet >>>>>> >>>>>> . >>>>>> >>>>>> >>>>>> -------- Original message -------- From: David Brown >>>>>> Date:31/05/2014 21:01 (GMT+10:00) To: Mark Knecht ,"L.M.J" >>>>>> Cc: Linux-RAID Subject: Re: Home desktop/server RAID upgrade >>>>>> >>>>>> On 30/05/14 22:14, Mark Knecht wrote: >>>>>>> >>>>>>> On Fri, May 30, 2014 at 12:29 PM, L.M.J >>>>>>> <linuxmasterjedi@free.fr> wrote: >>>>>>>> >>>>>>>> Le Fri, 30 May 2014 12:04:07 -0700, Mark Knecht >>>>>>>> <markknecht@gmail.com> a écrit : >>>>>>>> >>>>>>>>> In a RAID1 would a 3-drive Red RAID1 possibly be faster >>>>>>>>> than the 2-drive Se RAID1 and at the same time give me >>>>>>>>> more safety? >>>>>>>> >>>>>>>> >>>>>>>> Just a question inside the question : how do you manager >>>>>>>> a RAID1 with 3 drives ? Maybe you're talking about RAID5 >>>>>>>> then ? >>>>>>> >>>>>>> >>>>>>> OK, I'm no RAID expert but RAID1 is just drives in parallel >>>>>>> right. 2 drives, 3 drives, 4 drives, all holding exactly >>>>>>> the same data. In the case of a 3-drive RAID1 - if there is >>>>>>> such a beast - I could safely lose 2 drives. You ask a >>>>>>> reasonable question though as maybe the way this is >>>>>>> actually done is 2 drives + a hot spare in the box that >>>>>>> gets sync'ed if and only if one drive fails. Not sure and >>>>>>> maybe I'm totally wrong about that. >>>>>>> >>>>>>> A 3-drive RAID5 would be 2 drives in series - in this case >>>>>>> making 6TB - and then the 3rd drive being the redundancy. >>>>>>> In the case of a 3-drive RAID5 I could safely lose 1 >>>>>>> drive. >>>>>>> >>>>>>> In my case I don't need more than 3TB, so an option would >>>>>>> be a 3-drive RAID5 made out of 2TB drives which would give >>>>>>> me 4TB but I don't need the space as much as I want the >>>>>>> redundancy and I think RAID5 is slower than RAID1. >>>>>>> Additionally some more mdadm RAID knowledgeable people on >>>>>>> other lists say Linux mdadm RAID1 would be faster as it >>>>>>> will get data from more than one drive at a time. (Or >>>>>>> possibly get data from which ever drive returns it the >>>>>>> fastest. Not sure.) >>>>>>> >>>>>>> I believe one good option if I wanted 4 physical drives >>>>>>> would be RAID10 but that's getting more complicated again >>>>>>> which I didn't really want to do. >>>>>>> >>>>>>> So maybe it is just 2 drives and the 3 drive version isn't >>>>>>> even a possibility? Could be. >>>>>> >>>>>> >>>>>> With 3 drives, you have several possibilities. >>>>>> >>>>>> Raid5 makes "stripes" across the three drives, with 2 parts >>>>>> holding data and one part holding parity to provide >>>>>> redundancy. >>>>>> >>>>>> Raid1 is commonly called "mirroring", because you get the >>>>>> same data on each disk. md raid has no problem making a >>>>>> 3-way mirror, so that each disk is identical. This gives you >>>>>> excellent redundancy, and you can make three different reads >>>>>> in parallel - but writes have to go to each disk, which can >>>>>> be a little slower than using 2 disks. It's not often that >>>>>> people need that level of redundancy. >>>>>> >>>>>> Another option with md raid is the raid10 setups. For many >>>>>> uses, the fastest arrangement is raid10,f2. This means there >>>>>> is two copies of all your data (f3 would be three copies), >>>>>> with a "far" layout. >>>>>> >>>>>> <http://en.wikipedia.org/wiki/Linux_MD_RAID_10#LINUX-MD-RAID-10> >>>>>> >>>>>> >>>>>> > With this arrangement, reads are striped across all three disks, >>>>>> which is fast for large reads. Small reads can be handled >>>>>> in parallel. Most reads while be handled from the outer half >>>>>> of the disk, which is faster and needs less head movement - >>>>>> so reading is on average faster than a raid0 on the same >>>>>> disks. Small writes are fast, but large writes require quite >>>>>> a bit of head movement to get everything written twice to >>>>>> different parts of the disks. >>>>>> >>>>>> The "best" option always depends on your needs - how you want >>>>>> to access your files. A layout geared to fast striped reads >>>>>> of large files will be poorer for parallel small writes, and >>>>>> vice versa. raid10,f2 is often the best choice for a desktop >>>>>> or small system - but it is not very flexible if you later >>>>>> want to add new disks or replace the disks with bigger ones. >>>>>> >>>>>> md raid is flexible enough that it will even let you make a 3 >>>>>> disk raid6 array if you want - but a 3-way raid1 mirror will >>>>>> give you the same disk space and much better performance. >>>>>> >>>> >> > -- > To unsubscribe from this list: send the line "unsubscribe linux-raid" in > the body of a message to majordomo@vger.kernel.org > More majordomo info at http://vger.kernel.org/majordomo-info.html -- Roberto Spadim -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html ^ permalink raw reply [flat|nested] 13+ messages in thread
* Re: Home desktop/server RAID upgrade 2014-06-01 15:59 ` Mark Knecht 2014-06-02 23:04 ` David Brown @ 2014-06-04 12:29 ` Brad Campbell 1 sibling, 0 replies; 13+ messages in thread From: Brad Campbell @ 2014-06-04 12:29 UTC (permalink / raw) To: Mark Knecht, David Brown; +Cc: Craig Curtin, L.M.J, Linux-RAID On 01/06/14 23:59, Mark Knecht wrote: > > These VMs are the area where my current RAID6 (5 drive, 16k chunk > size) seems to have been a bad decision. The machine is powered off > every night. Loading these VMs takes at least 10-15 minutes each > morning where I see disk activity lights just grinding away the whole > time. If I had a single _performance_ goal in upgrading the disks it > would be to improve this significantly. Craig's SSD RAID1 suggestion > would certainly help here but at 240GB there wouldn't be a lot of room > left. That may be OK though. > Just a data point. I run between 4 & 10 VM's here 24/7 (always 4, sometimes 10). I started them out on a 3 drive RAID-5 on 7200RPM disks, then migrated them across to a RAID-10 of 15k SAS drives, and from there to a 6x240GB RAID-10 of SSD's. I would _never_ go back to rotating media now. Some of my batch tasks in a Windows 7 VM took up to an hour on the disks. I'm down to 20-30 minutes max now. Granted my workload is pretty atypical. Much less power, less noise and less heat. Lots to love. 240GB drives are a bit of a price sweet spot, and still were when I bought these 6 over a year ago. I just went with 3 from Intel and 3 from Samsung just to spread the risk a little bit. 3 are on the MB AHCI controller and 3 on an LSI SAS. Regards, Brad -- Dolphins are so intelligent that within a few weeks they can train Americans to stand at the edge of the pool and throw them fish. ^ permalink raw reply [flat|nested] 13+ messages in thread
end of thread, other threads:[~2014-06-04 12:29 UTC | newest]
Thread overview: 13+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2014-05-30 19:04 Home desktop/server RAID upgrade Mark Knecht
2014-05-30 19:29 ` L.M.J
2014-05-30 20:14 ` Mark Knecht
2014-05-30 20:36 ` Mark Knecht
2014-05-30 20:58 ` Roberto Spadim
2014-05-31 10:52 ` David Brown
[not found] ` <8mtskybo2j1i4l2bqu51l7ll.1401554092920@email.android.com>
2014-06-01 14:25 ` Mark Knecht
2014-06-01 15:06 ` David Brown
2014-06-01 15:59 ` Mark Knecht
2014-06-02 23:04 ` David Brown
[not found] ` <E78FE8BDBAD07C43A60163E7D1716EEC01839CFA3D@PROSIS-W2K8-1.prosis.local>
2014-06-03 7:58 ` David Brown
2014-06-03 14:59 ` Roberto Spadim
2014-06-04 12:29 ` Brad Campbell
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox; as well as URLs for NNTP newsgroup(s).