* [linux-lvm] Q: LVM over RAID, or plain disks? A:"Yes" = best of both worlds? [not found] <AANLkTimTYse7vnqZydc1Z5z60vNuQw3juj7J_W7+bQEm@mail.gmail.com> @ 2010-11-28 15:31 ` hansbkk 2010-11-29 16:27 ` Lars Ellenberg 2010-11-29 18:57 ` Nataraj [not found] ` <2D.98.07087.690A2FC4@cdptpa-omtalb.mail.rr.com> 1 sibling, 2 replies; 18+ messages in thread From: hansbkk @ 2010-11-28 15:31 UTC (permalink / raw) To: linux-lvm �- - - - - - My abject apologies to all for improper addressing in my previous messages (thanks to all those who set me straight :) Hope you're all still willing to consider my request for feedback. Start with a bit of context: - SAN/NAS (call it FILER-A) hosting say a dozen TB and servicing a few dozen client machines and servers, mostly virtual hosts. Another, larger (FILER-B - still just tens of TB) host's drives are used for storing backup sets, via not only Amanda, but also filesystems comprising gazillions of hard-linked archive sets created by (eg) rdiff-backup, rsnapshot and BackupPC. We're on a very limited budget, therefore no tape storage for backups. - I plan to run LVM over RAID (likely RAID1 or RAID10) for IMO an ideal combination of fault tolerance, performance and flexibility. - I am not at this point overly concerned about performance issues - reliability/redundancy and ease of recovery are my main priorities. Problem: For off-site data rotation, the hard-linked filesystems on FILER-B require full filesystem cloning with block-level tools rather than file-level copying or sync'ing. My current plan is to swap out disks mirrored via RAID, marking them as "failed" and then rebuilding using the (re-initialized) incoming rotation set. HOWEVER - the use of LVM (and possibly RAID10) adds complexity to the filesystems, which makes disaster recovery from the detached disk sets much more difficult than regular partitions on physical disks. Theoretical solution: Use RAID1 on the "top layer" to mirror the data stored in an LVM (set of) disk(s) on the one hand (call it TopRAID1) to ***regular partitions*** on actual physical disks on the other (call this the TopRAID2 side). (ASCII art best viewed with a monospaced font) "TopRAID1" side �______________________________________ | � � � � � � � �LVM VG � � � � � � � �| | �_____ � _____________ � __________ �| | | LV1 | | � � LV2 � � | | � �LV3 � | | | | � � | | � � � � � � | | � � � � �| | | | � � | | � � � � � � | | � � � � �| | | | � � | | � � � � � � | | � � � � �| | | | � � | | � � � � � � | | � � � � �| | | | � � | | � � � � � � | | � � � � �| | | |_____| |_____________| |__________| | |____v___________v______________v______| � � v � � � � � v � � � � � � �v � � v � � � � � v � � � � � � �v � RAID1 � � � RAID1 � � � � �RAID1 � � v � � � � � v � � � � � � �v � __v__ � ______v______ � _____v____ �| HD1 | | � � HD2 � � | | � �HD3 � | �| � � | | � � � � � � | | � � � � �| �| � � | | � � � � � � | | � � � � �| �| � � | | � � � � � � | | � � � � �| �| � � | | � � � � � � | | � � � � �| �| � � | | � � � � � � | | � � � � �| �|_____| |_____________| |__________| "TopRAID2" side The mirroring at the top level would be set up between the individual LVs on the TopRAID1 side and regular filesystem partitions (no RAID or LVM) on the TopRAID2 side. In the event of a massive host failure, the filesystems on the TopRAID2 side could be easily mounted for data recovery and/or service resumption on another machine, and the TopRAID1 disk set rebuilt from scratch and then re-mirrored from the TopRAID2 disks. One design goal would be to not allow any LV to get so large that it won't fit on a single physical disk on the TopRAID2 side. If this is not possible, then the corresponding TopRAID2 side would need to comprise a multiple disk set, perhaps striped by RAID0 - not as straightforward to recover as single disks, but at least without the LVM layer. Remember, the main purpose of this arrangement is so that the disks in the TopRAID2 set can be rotated out for offsite storage. Ideally this would be done by using an extra identical set (TopRAID2a and TopRAID2b) to minimize the time windows when the live data is running on TopRAID1 only. Note that on the TopRAID1 side the LVM layers could be running on top of another set of RAID disks (call it the BottomRAID), again either RAID1 or perhaps RAID10 mirroring at the lowest layer. This disk set could be allowed to grow in both size and complexity, with an expectation that in the event of massive failure I won't even attempt to rebuild/recover it, just tear it down and set it up again from scratch, then mirror the data back from TopRAID2. At this point this is all idle speculation on my part, and although I think the technology makes it possible, I don't know whether it is a practical scheme. An enhancement of this idea would be to implement the "TopRAID" with a full-server mirror using drdb and heartbeat, perhaps eliminating the need for intra-server disk mirroring. In this case the TopRAID1 server would have the flexibile disk space allocation of LVM, while the TopRAID2 server's disks would all be just regular partitions (no LVM), again, easily swapped out for offsite rotation. Any feedback on these ideas would be most appreciated. ^ permalink raw reply [flat|nested] 18+ messages in thread
* Re: [linux-lvm] Q: LVM over RAID, or plain disks? A:"Yes" = best of both worlds? 2010-11-28 15:31 ` [linux-lvm] Q: LVM over RAID, or plain disks? A:"Yes" = best of both worlds? hansbkk @ 2010-11-29 16:27 ` Lars Ellenberg 2010-11-29 17:00 ` hansbkk 2010-11-29 18:57 ` Nataraj 1 sibling, 1 reply; 18+ messages in thread From: Lars Ellenberg @ 2010-11-29 16:27 UTC (permalink / raw) To: linux-lvm On Sun, Nov 28, 2010 at 10:31:51PM +0700, hansbkk@gmail.com wrote: > �- - - - - - My abject apologies to all for improper addressing in my > previous messages (thanks to all those who set me straight :) > > Hope you're all still willing to consider my request for feedback. > Start with a bit of context: > > - SAN/NAS (call it FILER-A) hosting say a dozen TB and servicing a few > dozen client machines and servers, mostly virtual hosts. Another, > larger (FILER-B - still just tens of TB) host's drives are used for > storing backup sets, via not only Amanda, but also filesystems > comprising gazillions of hard-linked archive sets created by (eg) > rdiff-backup, rsnapshot and BackupPC. We're on a very limited budget, > therefore no tape storage for backups. > > - I plan to run LVM over RAID (likely RAID1 or RAID10) for IMO an > ideal combination of fault tolerance, performance and flexibility. > > - I am not at this point overly concerned about performance issues - > reliability/redundancy and ease of recovery are my main priorities. > > > Problem: > > For off-site data rotation, the hard-linked filesystems on FILER-B > require full filesystem cloning with block-level tools rather than > file-level copying or sync'ing. My current plan is to swap out disks > mirrored via RAID, marking them as "failed" and then rebuilding using > the (re-initialized) incoming rotation set. Did you consider DRBD? Use DRBD, protocol A, potentially with drbd-proxy in between to mitigate impact on primary site performance due to high latency connection to the desaster recovery site. At your option have it replicate continuously, or cron-job triggered sync it up every "rotation interval", then disconnect again. Depending on what you want, you can have one DRBD on top of each LV, or have one DRBD by the PV of a VG. Depending on what you do to the devices on the DR site, you will likely be able to always do an incremental resync only (ship only blocks that changed). If you want to discuss that further, feel free to followup here, on drbd-user, on freenode #drbd, or contact LINBIT directly. -- : Lars Ellenberg : LINBIT | Your Way to High Availability : DRBD/HA support and consulting http://www.linbit.com DRBD� and LINBIT� are registered trademarks of LINBIT, Austria. ^ permalink raw reply [flat|nested] 18+ messages in thread
* Re: [linux-lvm] Q: LVM over RAID, or plain disks? A:"Yes" = best of both worlds? 2010-11-29 16:27 ` Lars Ellenberg @ 2010-11-29 17:00 ` hansbkk 0 siblings, 0 replies; 18+ messages in thread From: hansbkk @ 2010-11-29 17:00 UTC (permalink / raw) To: LVM general discussion and development On Mon, Nov 29, 2010 at 11:27 PM, Lars Ellenberg <lars.ellenberg@linbit.com> wrote: > Did you consider DRBD? > I mentioned that as a "down the road" possibility in my last paragraph - I just spelled it wrong :) > An enhancement of this idea would be to implement the "TopRAID" with a full-server mirror using drdb and heartbeat. . . No budget at the moment for a extra filer, nor for the hefty WAN needed. ^ permalink raw reply [flat|nested] 18+ messages in thread
* Re: [linux-lvm] Q: LVM over RAID, or plain disks? A:"Yes" = best of both worlds? 2010-11-28 15:31 ` [linux-lvm] Q: LVM over RAID, or plain disks? A:"Yes" = best of both worlds? hansbkk 2010-11-29 16:27 ` Lars Ellenberg @ 2010-11-29 18:57 ` Nataraj 2010-11-30 5:20 ` hansbkk 1 sibling, 1 reply; 18+ messages in thread From: Nataraj @ 2010-11-29 18:57 UTC (permalink / raw) To: LVM general discussion and development hansbkk@gmail.com wrote: > - - - - - - My abject apologies to all for improper addressing in my > previous messages (thanks to all those who set me straight :) > > Hope you're all still willing to consider my request for feedback. > Start with a bit of context: > > - SAN/NAS (call it FILER-A) hosting say a dozen TB and servicing a few > dozen client machines and servers, mostly virtual hosts. Another, > larger (FILER-B - still just tens of TB) host's drives are used for > storing backup sets, via not only Amanda, but also filesystems > comprising gazillions of hard-linked archive sets created by (eg) > rdiff-backup, rsnapshot and BackupPC. We're on a very limited budget, > therefore no tape storage for backups. > > - I plan to run LVM over RAID (likely RAID1 or RAID10) for IMO an > ideal combination of fault tolerance, performance and flexibility. > > - I am not at this point overly concerned about performance issues - > reliability/redundancy and ease of recovery are my main priorities. > > > Problem: > > For off-site data rotation, the hard-linked filesystems on FILER-B > require full filesystem cloning with block-level tools rather than > file-level copying or sync'ing. My current plan is to swap out disks > mirrored via RAID, marking them as "failed" and then rebuilding using > the (re-initialized) incoming rotation set. > > HOWEVER - the use of LVM (and possibly RAID10) adds complexity to the > filesystems, which makes disaster recovery from the detached disk sets > much more difficult than regular partitions on physical disks. > > > Theoretical solution: > > Use RAID1 on the "top layer" to mirror the data stored in an LVM (set > of) disk(s) on the one hand (call it TopRAID1) to ***regular > partitions*** on actual physical disks on the other (call this the > TopRAID2 side). > > Your proposed solution is a bit confusing to understand, however raid1 works for doing backups in the manner that you describe . I use it myself and I have, over time read about others doing so as well. Be sure to create your volumes with --bitmap=internal, that way when you swap in a drive, it won't need to replicate the entire drive, only the part that is changed. If your not going to manage the drives yourself, you will need an operations staff that has a pretty good understanding of how raid works and/or possibly write a robust set of scripts to manage the drives, ensure that the correct volume is mounted, etc. Also, I don't personally feel that disks are a suitable media for long term data archival, so if that is really your purpose, as opposed to a quick way to recover from a disk failure, then you might consider doing periodic tape or optical media backups as well. Nataraj ^ permalink raw reply [flat|nested] 18+ messages in thread
* Re: [linux-lvm] Q: LVM over RAID, or plain disks? A:"Yes" = best of both worlds? 2010-11-29 18:57 ` Nataraj @ 2010-11-30 5:20 ` hansbkk 2010-11-30 7:14 ` Nataraj 0 siblings, 1 reply; 18+ messages in thread From: hansbkk @ 2010-11-30 5:20 UTC (permalink / raw) To: Nataraj; +Cc: linux-raid, linux-lvm On Tue, Nov 30, 2010 at 1:57 AM, Nataraj <incoming-redhat@rjl.com> wrote: > Your proposed solution is a bit confusing to understand, Thanks for taking the time to give me feedback! And sorry I couldn't make it clearer, but if the diagram is pasted into a monospace screen it becomes more so - duplicated below. The key is that I want to RAID1 from a given **LV** within the TopRAID1 set to regular partitions on physical disks on the TopRAID2 side. > however raid1 works for doing backups in the manner that you describe . > I use it myself and I > have, over time read about others doing so as well. Be sure to create your > volumes with --bitmap=internal, that way when you swap in a drive, it won't > need to replicate the entire drive, only the part that is changed. TopRAID1's LVM is likely to be running over a RAID6 set , so I'm not depending on the TopRAID mirroring for reliability, just using it for the above volume cloning. When TopRAID1 is running with the TopRAID2 side marked faulty or missing, will there be a performance hit or other negative consequences? If so, would it be possible/better for the host In normal operations to mount the underlying LV directly rather than the degraded top-level RAID1? Obviously I'm asking the whole list here, not just Nataraj :) > If your not going to manage the drives yourself, you will need an operations > staff that has a pretty good understanding of how raid works and/or possibly > write a robust set of scripts to manage the drives, ensure that the correct > volume is mounted, etc. Just me, and I'm sure my understanding will grow over time :) We're not keeping archive sets of the drives themselves, rather managing the archives created by BackupPC/rsnapshot/rdiff-backup and Amanda, keeping them "live" in the disk sets discussed here, each in its own LV(s), which data is constantly being refreshed onto the rotating drives through the RAID re-mirroring process. > Also, I don't personally feel that disks are a > suitable media for long term data archival, so if that is really your > purpose, as opposed to a quick way to recover from a disk failure, then you > might consider doing periodic tape or optical media backups as well. Remember we're talking about tens of terabytes here! I do also make occasional DVD runs of key data sets + 20% PAR2s, but even then, I'm not counting on these lasting more than 3 years, maybe 5 if I use well-recommended media (checking the media ID). If we could afford Verbatim "archive quality" I'd probably invest in a tape unit. I just keep in mind Mr Keynes' "In the long run, we're all dead." Thanks to all for your help! ---------------------------------------------- Repeat of conceptual diagram, best viewed with a monospaced font "TopRAID1" side - LV's within a VG running on a RAID6 set ______________________________________ | LVM VG | | _____ _____________ __________ | | | LV1 | | LV2 | | LV3 | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | |_____| |_____________| |__________| | |____v___________v______________v______| v v v v v v RAID1 RAID1 RAID1 v v v __v__ ______v______ _____v____ | HD1 | | HD2 | | HD3 | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | |_____| |_____________| |__________| "TopRAID2" side - regular partitions on physical disks ^ permalink raw reply [flat|nested] 18+ messages in thread
* Re: [linux-lvm] Q: LVM over RAID, or plain disks? A:"Yes" = best of both worlds? 2010-11-30 5:20 ` hansbkk @ 2010-11-30 7:14 ` Nataraj 2010-11-30 7:34 ` hansbkk 0 siblings, 1 reply; 18+ messages in thread From: Nataraj @ 2010-11-30 7:14 UTC (permalink / raw) To: hansbkk; +Cc: linux-lvm hansbkk@gmail.com wrote: > On Tue, Nov 30, 2010 at 1:57 AM, Nataraj <incoming-redhat@rjl.com> wrote: > >> Your proposed solution is a bit confusing to understand, >> > > Thanks for taking the time to give me feedback! And sorry I couldn't > make it clearer, but if the diagram is pasted into a monospace screen > it becomes more so - duplicated below. > > The key is that I want to RAID1 from a given **LV** within the > TopRAID1 set to regular partitions on physical disks on the TopRAID2 > side. > > >> however raid1 works for doing backups in the manner that you describe . >> I use it myself and I >> have, over time read about others doing so as well. Be sure to create your >> volumes with --bitmap=internal, that way when you swap in a drive, it won't >> need to replicate the entire drive, only the part that is changed. >> > > TopRAID1's LVM is likely to be running over a RAID6 set , so I'm not > depending on the TopRAID mirroring for reliability, just using it for > the above volume cloning. > Your raid 1 backups won't mirror any snapshots of your LV's unless you specifically setup mirroring of the snapshots after they exist. > When TopRAID1 is running with the TopRAID2 side marked faulty or > missing, will there be a performance hit or other negative > consequences? > My guess would be no, but the only way to really tell is to run tests. It would depend somewhat on how your drives and controllers are configured, but writes might be faster with only a single drive online, whereas reads may be slower, since you don't have the ability to read from both drives. But since your LV is already raid 6, it's hard to tell what the behavior of raid 1 on top of raid 6 will be. I think you would just have to try it. The syncing of data when you readd a drive into the raid 1 set tends to be very fast (with the internal bitmap). > If so, would it be possible/better for the host In normal operations > to mount the underlying LV directly rather than the degraded top-level > RAID1? > > No, you want to have mdadm assemble the raid volume, even if in degraded mode with only one drive and then access the LV on top of the md device. Even if you were able to mount the LV and bypass raid, that would be pointless because you would not update the bitmap and superblock and the integrety of the raid set would be lost. Nataraj ^ permalink raw reply [flat|nested] 18+ messages in thread
* Re: [linux-lvm] Q: LVM over RAID, or plain disks? A:"Yes" = best of both worlds? 2010-11-30 7:14 ` Nataraj @ 2010-11-30 7:34 ` hansbkk 2010-11-30 13:13 ` Phil Turmel 2010-11-30 15:41 ` Andrew Gideon 0 siblings, 2 replies; 18+ messages in thread From: hansbkk @ 2010-11-30 7:34 UTC (permalink / raw) To: Nataraj; +Cc: linux-raid, linux-lvm On Tue, Nov 30, 2010 at 2:14 PM, Nataraj <incoming-centos@rjl.com> wrote: >> TopRAID1's LVM is likely to be running over a RAID6 set , so I'm not >> depending on the TopRAID mirroring for reliability, just using it for >> the above volume cloning. > > Your raid 1 backups won't mirror any snapshots of your LV's unless you > specifically setup mirroring of the snapshots after they exist. Ah, getting clearer to me, I was thinking I'd be mirroring the LV itself, but you're right, taking a snapshot and mirroring that is a much better idea. So here's a summary of steps, please confirm: - create a snapshot of a given volume - create a new RAID1 mdN between that and a physical partition (blank?) - let that get sync'd up - break the RAID (fail the partition?), remove the drive - delete the snapshot The below is less clear to me, especially if the above is correct: >> If so, would it be possible/better for the host In normal operations >> to mount the underlying LV directly rather than the degraded top-level >> RAID1? > > No, you want to have mdadm assemble the raid volume, even if in degraded > mode with only one drive and then access the LV on top of the md device. > �Even if you were able to mount the LV and bypass raid, that would be > pointless because you would not update the bitmap and superblock and the > integrety of the raid set would be lost. During normal operations - when I'm not in the process of taking a RAID-backup of my LV snapshot - it seems to me that the "Top-RAID" mdN doesn't even exist right? It's set up to mirror between a snapshot and a regular partition, both of which don't even exist during these normal operations. Therefore during normal operations the host *is* mounting the LV directly, not via the "Top-RAID"mdN. I wasn't talking about accessing the "Bottom-RAID" which creates the underlying PV - this is transparent to LVM anyway, right? Thanks again for your help (or in advance to anyone else who would like to chime in) ^ permalink raw reply [flat|nested] 18+ messages in thread
* Re: [linux-lvm] Q: LVM over RAID, or plain disks? A:"Yes" = best of both worlds? 2010-11-30 7:34 ` hansbkk @ 2010-11-30 13:13 ` Phil Turmel 2010-11-30 15:39 ` hansbkk 2010-11-30 15:41 ` Andrew Gideon 1 sibling, 1 reply; 18+ messages in thread From: Phil Turmel @ 2010-11-30 13:13 UTC (permalink / raw) To: hansbkk; +Cc: linux-raid, Nataraj, linux-lvm Hi Hans, (?) On 11/30/2010 02:34 AM, hansbkk@gmail.com wrote: > On Tue, Nov 30, 2010 at 2:14 PM, Nataraj <incoming-centos@rjl.com> wrote: >>> TopRAID1's LVM is likely to be running over a RAID6 set , so I'm not >>> depending on the TopRAID mirroring for reliability, just using it for >>> the above volume cloning. >> >> Your raid 1 backups won't mirror any snapshots of your LV's unless you >> specifically setup mirroring of the snapshots after they exist. > > Ah, getting clearer to me, I was thinking I'd be mirroring the LV > itself, but you're right, taking a snapshot and mirroring that is a > much better idea. I think you are making this overly complex, insisting on a RAID1 operation to backup from on filer to the other. Consider having each disk on filer #2 configured as a single LVM PV/VG, so it can stand alone in a rotation. The try the alternate below. > So here's a summary of steps, please confirm: > - create a snapshot of a given volume Here's where you are over-complicating things: > - create a new RAID1 mdN between that and a physical partition (blank?) > - let that get sync'd up > - break the RAID (fail the partition?), remove the drive As an alternate, with simpler recovery semantics: Create matching LV on non-RAID PV/VG on filer #2 dd + netcat + dd or other technique to dup the snapshot on filer #1 to filer #2 > - delete the snapshot Now, you have a single disk in your backup set that can be mounted on either filer, and either copied back into service, or in an emergency, used directly (live) in filer #1. This approach also gives you the *option* to implement the backup transfer with file system conversions, compression, free space removal, or any other administrative adjustments you need. A RAID mirror can only duplicate the raw block device. HTH, Phil ^ permalink raw reply [flat|nested] 18+ messages in thread
* Re: [linux-lvm] Q: LVM over RAID, or plain disks? A:"Yes" = best of both worlds? 2010-11-30 13:13 ` Phil Turmel @ 2010-11-30 15:39 ` hansbkk 2010-11-30 16:56 ` Phil Turmel 0 siblings, 1 reply; 18+ messages in thread From: hansbkk @ 2010-11-30 15:39 UTC (permalink / raw) To: Phil Turmel; +Cc: linux-raid, Nataraj, linux-lvm On Tue, Nov 30, 2010 at 8:13 PM, Phil Turmel <philip@turmel.org> wrote: > I think you are making this overly complex, insisting on a RAID1 operation to backup from on filer to the other. Consider having each disk on filer #2 configured as a single LVM PV/VG, so it can stand alone in a rotation. The try the alternate below. <snip> > As an alternate, with simpler recovery semantics: > Create matching LV on non-RAID PV/VG on filer #2 > dd + netcat + dd or other technique to dup the snapshot on filer #1 to filer #2 > >> - delete the snapshot Thanks Phil, and yes I do tend to complicate things during a learning/planning stage, then when I get to implementation reality tends to force simplification :) Perhaps I miscommunicated, but your suggestion misses a couple of factors: - this process is (so far) only targeting the backup data hosted on Filer-B - the backup-mirroring I'm considering is all within that one server, not across a network connection - I can't have "each disk configured as a single VG", since the whole point of my using LVM is to get as much flexibility as possible for dozens of hosts to share my dozen-terabyte disk space, when I only have a half-dozen disks. - my goal is for the snapshot-copy to end up in a regular partition on a physical disk, without any RAID/LVM layers standing between the data and easy recovery However your main point is perfectly valid - dd basically does the same block-level data transfer as mdm RAID1 - which is a must for the gazillions-of-hardlinks backup filesystems (as opposed to file-level tools like rsync). So adapting your suggestion to fit (my perception of) my needs: - create an LV snapshot - mount a plain partition on a physical hard disk (preferably on a separate controller?) - dd the data from the LV snapshot over to the partition - delete the snapshot > A RAID mirror can only duplicate the raw block device. Isn't that all that dd is doing as well? My perception is that software like mdm, designed as it is to maximize data reliability would handle the cloning more reliably than dd - isn't there more error-checking going on during a RAID1 re-mirroring? Your main (and quite valid) point is that a user-space tool designed to do the job is probably more appropriate than putting RAID1 to use in a way beyond what was originally intended. So I guess my question becomes: What is the best tool to block-level clone an LV snapshot to a regular disk partition? - "best" = as close to 100% reliably as possible, speed isn't nearly as important Would a COT cloning package (something like Acronis TrueImage) have data reliability advantages (like mdm's) over dd? ^ permalink raw reply [flat|nested] 18+ messages in thread
* Re: [linux-lvm] Q: LVM over RAID, or plain disks? A:"Yes" = best of both worlds? 2010-11-30 15:39 ` hansbkk @ 2010-11-30 16:56 ` Phil Turmel 2010-12-01 4:45 ` hansbkk 0 siblings, 1 reply; 18+ messages in thread From: Phil Turmel @ 2010-11-30 16:56 UTC (permalink / raw) To: hansbkk; +Cc: linux-raid, Nataraj, linux-lvm On 11/30/2010 10:39 AM, hansbkk@gmail.com wrote: > On Tue, Nov 30, 2010 at 8:13 PM, Phil Turmel <philip@turmel.org> wrote: >> I think you are making this overly complex, insisting on a RAID1 operation to backup from on filer to the other. Consider having each disk on filer #2 configured as a single LVM PV/VG, so it can stand alone in a rotation. The try the alternate below. > > <snip> > >> As an alternate, with simpler recovery semantics: >> Create matching LV on non-RAID PV/VG on filer #2 >> dd + netcat + dd or other technique to dup the snapshot on filer #1 to filer #2 >> >>> - delete the snapshot > > Thanks Phil, and yes I do tend to complicate things during a > learning/planning stage, then when I get to implementation reality > tends to force simplification :) > > Perhaps I miscommunicated, but your suggestion misses a couple of factors: > - this process is (so far) only targeting the backup data hosted on Filer-B > - the backup-mirroring I'm considering is all within that one > server, not across a network connection OK. > - I can't have "each disk configured as a single VG", since the > whole point of my using LVM is to get as much flexibility as possible > for dozens of hosts to share my dozen-terabyte disk space, when I only > have a half-dozen disks. I meant only the disks that will be rotated offline/offsite should be set up as "solo" Volume Groups. > - my goal is for the snapshot-copy to end up in a regular partition > on a physical disk, without any RAID/LVM layers standing between the > data and easy recovery This wasn't clear, and is a key factor. You can wipe the superblock from an mdraid mirror member, and if the SB was at the end of the partition, be left with usable partition (with a little waste space at the end). But this doesn't help you, because the partition you want to clone is an LV inside LVM. lvconvert can add a mirror to a partition on the fly, and split it off at an appropriate time, but only if the mirror is also inside LVM. lvconvert can do some neat stuff, but its all inside LVM. > However your main point is perfectly valid - dd basically does the > same block-level data transfer as mdm RAID1 - which is a must for the > gazillions-of-hardlinks backup filesystems (as opposed to file-level > tools like rsync). (Actually, rsync and tar are both hardlink-aware, at least the versions I use.) > So adapting your suggestion to fit (my perception of) my needs: > > - create an LV snapshot > - mount a plain partition on a physical hard disk (preferably on a > separate controller?) > - dd the data from the LV snapshot over to the partition > - delete the snapshot Yep, this is basically what I recommended. >> A RAID mirror can only duplicate the raw block device. > > Isn't that all that dd is doing as well? The raid operation carries a lot of metadata to track what is in sync vs. what's not, because it is expected to work on live filesystems. If you are duping a snapshot, the snapshot is static, and dd will give you the same result, without the metadata overhead. > My perception is that software like mdm, designed as it is to maximize > data reliability would handle the cloning more reliably than dd - > isn't there more error-checking going on during a RAID1 re-mirroring? > Your main (and quite valid) point is that a user-space tool designed > to do the job is probably more appropriate than putting RAID1 to use > in a way beyond what was originally intended. Well, both mdraid and dd will choke on a write error on the target. > So I guess my question becomes: > > What is the best tool to block-level clone an LV snapshot to a regular > disk partition? > > - "best" = as close to 100% reliably as possible, speed isn't nearly > as important I would use dd. > Would a COT cloning package (something like Acronis TrueImage) have > data reliability advantages (like mdm's) over dd? Not really. But they may offer various forms of compression/sparsification/error detection if you wish to store the final backups as files. Of course, if you do that, you might as well use tar+gzip+md5sum. I'm guessing that you'll be scripting all of this, in which case I'd recommend sticking to some combination of lvcreate -s, lvconvert, dd, and possibly tar+gzip. You want your dismountable disks to be accessible stand-alone, but I don't see why that would preclude setting them up so each is a unique LVM VG. HTH, Phil ^ permalink raw reply [flat|nested] 18+ messages in thread
* Re: [linux-lvm] Q: LVM over RAID, or plain disks? A:"Yes" = best of both worlds? 2010-11-30 16:56 ` Phil Turmel @ 2010-12-01 4:45 ` hansbkk 2010-12-01 12:50 ` Phil Turmel 0 siblings, 1 reply; 18+ messages in thread From: hansbkk @ 2010-12-01 4:45 UTC (permalink / raw) To: Phil Turmel; +Cc: linux-raid, linux-lvm On Tue, Nov 30, 2010 at 11:56 PM, Phil Turmel <philip@turmel.org> wrote: > (Actually, rsync and tar are both hardlink-aware, at least the versions I use.) My backup filesystems contain so many hardlinks (millions, constantly growing) that file-level tools choke - this really must be done at the block device level - see my previous post for more detail. It's also now clear to me that rsync is the tool to use for this for all the other LVs without such problematic filesystems, as I know the tool and trust its error-checking routines. >> So adapting your suggestion to fit (my perception of) my needs: >> >> � - create an LV snapshot >> � - mount a plain partition on a physical hard disk (preferably on a >> separate controller?) >> � - dd the data from the LV snapshot over to the partition >> � - delete the snapshot > > Yep, this is basically what I recommended. >> So I guess my question becomes: >> >> What is the best tool to block-level clone an LV snapshot to a regular >> disk partition? >> >> � - "best" = as close to 100% reliably as possible, speed isn't nearly >> as important > > I would use dd. OK, that's clear, thanks. >> Would a COT cloning package (something like Acronis TrueImage) have >> data reliability advantages (like mdm's) over dd? > > Not really. �But they may offer various forms of compression/sparsification/error detection if you wish to store the final backups as files. �Of course, if you do that, you might as well use tar+gzip+md5sum. No, I'm talking about partition-to-partition cloning operations, which some of these do support. The error detection is critical, and why I was looking at mdraid in the first place. > You want your dismountable disks to be accessible stand-alone, but I don't see why that would preclude setting them up so each is a unique LVM VG. It doesn't preclude it, but it's a layer of complexity during the data recovery process I'm trying to avoid. The ultimate goal is a plain partition on a plain disk that can be directly mounted on a SATA2 host via a normal recovery/LiveCD by a user that's never heard of RAID or LVM. To summarize your feedback: - mdraid's sync error-checking routines don't add value over dd to ensure accurate cloning of a static partition; its metadata is just useless overhead in this case. - dd is reliable enough One last question (and I do realize it's now OT for here, so I won't be hurt if it's ignored :) Does dd already do some sort of "verify after copy"? I will likely investigate the available COTS partition cloning tools as well. Thanks for all your help, not least in helping me to clarify my own goals ^ permalink raw reply [flat|nested] 18+ messages in thread
* Re: [linux-lvm] Q: LVM over RAID, or plain disks? A:"Yes" = best of both worlds? 2010-12-01 4:45 ` hansbkk @ 2010-12-01 12:50 ` Phil Turmel 2010-12-01 19:47 ` hansbkk 0 siblings, 1 reply; 18+ messages in thread From: Phil Turmel @ 2010-12-01 12:50 UTC (permalink / raw) To: hansbkk; +Cc: linux-raid, linux-lvm On 11/30/2010 11:45 PM, hansbkk@gmail.com wrote: > On Tue, Nov 30, 2010 at 11:56 PM, Phil Turmel <philip@turmel.org> wrote: > >> (Actually, rsync and tar are both hardlink-aware, at least the versions I use.) > > My backup filesystems contain so many hardlinks (millions, constantly > growing) that file-level tools choke - this really must be done at the > block device level - see my previous post for more detail. Ah -- I did miss that detail. > It's also now clear to me that rsync is the tool to use for this for > all the other LVs without such problematic filesystems, as I know the > tool and trust its error-checking routines. Indeed. I push my own critical systems offsite with rsync+ssh. [snip /] >> I would use dd. > > OK, that's clear, thanks. > > >> You want your dismountable disks to be accessible stand-alone, but I don't see why that would preclude setting them up so each is a unique LVM VG. > > It doesn't preclude it, but it's a layer of complexity during the data > recovery process I'm trying to avoid. > > The ultimate goal is a plain partition on a plain disk that can be > directly mounted on a SATA2 host via a normal recovery/LiveCD by a > user that's never heard of RAID or LVM. Ah -- not *you*. And you wouldn't be mixing customers on a disk, I presume? > To summarize your feedback: > > - mdraid's sync error-checking routines don't add value over dd to > ensure accurate cloning of a static partition; its metadata is just > useless overhead in this case. Right. > - dd is reliable enough I guess if your filer lacks ECC ram, you could have a bit flip between reading and writing that would be missed. It's really an end-to-end hardware integrity issue at this level. But an undetected read error between the platter and RAM will be propagated by every block-level software tool out there, including software raid. btrfs can do data-checksumming, but that's at the FS level. > One last question (and I do realize it's now OT for here, so I won't > be hurt if it's ignored :) > > Does dd already do some sort of "verify after copy"? I will likely > investigate the available COTS partition cloning tools as well. Not natively, but it fairly easy to pipe a dd reader through md5sum to a dd writer, then follow up with a dd read + md5sum of the copied partition (taking care to read precisely the same number of sectors). The various flavors of ddrescue might have something like this.. didn't check. > Thanks for all your help, not least in helping me to clarify my own goals You're welcome. Phil ^ permalink raw reply [flat|nested] 18+ messages in thread
* Re: [linux-lvm] Q: LVM over RAID, or plain disks? A:"Yes" = best of both worlds? 2010-12-01 12:50 ` Phil Turmel @ 2010-12-01 19:47 ` hansbkk 0 siblings, 0 replies; 18+ messages in thread From: hansbkk @ 2010-12-01 19:47 UTC (permalink / raw) To: Phil Turmel; +Cc: linux-raid, linux-lvm On Wed, Dec 1, 2010 at 7:50 PM, Phil Turmel <philip@turmel.org> wrote: >> Does dd already do some sort of "verify after copy"? I will likely >> investigate the available COTS partition cloning tools as well. > > Not natively, but it fairly easy to pipe a dd reader through md5sum to a dd writer, then follow up with a dd read + md5sum of the copied partition (taking care to read precisely the same number of sectors). > > The various flavors of ddrescue might have something like this.. �didn't check. Sorry it's a bit OT, but for the sake of future googlers thought I'd point to this tool I found: http://dc3dd.sourceforge.net/ http://www.forensicswiki.org/wiki/Dc3dd ^ permalink raw reply [flat|nested] 18+ messages in thread
* Re: [linux-lvm] Q: LVM over RAID, or plain disks? A:"Yes" = best of both worlds? 2010-11-30 7:34 ` hansbkk 2010-11-30 13:13 ` Phil Turmel @ 2010-11-30 15:41 ` Andrew Gideon 2010-11-30 15:53 ` hansbkk 1 sibling, 1 reply; 18+ messages in thread From: Andrew Gideon @ 2010-11-30 15:41 UTC (permalink / raw) To: linux-lvm On Tue, 30 Nov 2010 14:34:05 +0700, hansbkk wrote: > So here's a summary of steps, please confirm: > - create a snapshot of a given volume - create a new RAID1 mdN between > that and a physical partition (blank?) - let that get sync'd up - > break the RAID (fail the partition?), remove the drive - delete the > snapshot I'm having some difficulty following this thread, so perhaps I've missed some other aspect to this. But the above makes it seem like you simply wish to transport snapshots. Why not use something as straightforward as dd for this rather than RAID1? RAID1 tools (mdadm, drbd, etc.) have the benefit of keeping volumes in sync over time. But the above process describes a synchronization over a single instant (the instant of the snapshot being taken). So why bother with the extra work/capability/ complexity? I'm also puzzled by something you wrote in your very first message on this thread: "the hard-linked filesystems on FILER-B require full filesystem cloning with block-level tools rather than file-level copying or sync'ing." Why? rsync -H handles cloning of hard links (though at a performance cost). - Andrew ^ permalink raw reply [flat|nested] 18+ messages in thread
* Re: [linux-lvm] Q: LVM over RAID, or plain disks? A:"Yes" = best of both worlds? 2010-11-30 15:41 ` Andrew Gideon @ 2010-11-30 15:53 ` hansbkk 2010-11-30 15:54 ` hansbkk 0 siblings, 1 reply; 18+ messages in thread From: hansbkk @ 2010-11-30 15:53 UTC (permalink / raw) To: LVM general discussion and development On Tue, Nov 30, 2010 at 10:41 PM, Andrew Gideon <ag2827189@tagmall.com> wrote: > I'm having some difficulty following this thread, so perhaps I've missed > some other aspect to this. �But the above makes it seem like you simply > wish to transport snapshots. �Why not use something as straightforward as > dd for this rather than RAID1? �RAID1 tools (mdadm, drbd, etc.) have the > benefit of keeping volumes in sync over time. �But the above process > describes a synchronization over a single instant (the instant of the > snapshot being taken). �So why bother with the extra work/capability/ > complexity? Our posts must have crossed in the mail - please see my message immediately prior to this one. > I'm also puzzled by something you wrote in your very first message on > this thread: "the hard-linked filesystems on FILER-B require full > filesystem cloning with block-level tools rather than file-level copying > or sync'ing." �Why? �rsync -H handles cloning of hard links (though at a > performance cost). The performance cost becomes unworkable when you are talking about a filesystem with literally millions of hardlinks, as any reasonably well-used rsnapshot/rdiff-backup/BackupPC installation will create. Perhaps my current problem will be overcome by adding 64GB of RAM to my host, but as the number of hardlinks grows. . . It is precisely this problem I'm trying to address. http://www.google.co.th/search?hl=en&q=backuppc+pool+rsync+hard-link ^ permalink raw reply [flat|nested] 18+ messages in thread
* [linux-lvm] Q: LVM over RAID, or plain disks? A:"Yes" = best of both worlds? 2010-11-30 15:53 ` hansbkk @ 2010-11-30 15:54 ` hansbkk 0 siblings, 0 replies; 18+ messages in thread From: hansbkk @ 2010-11-30 15:54 UTC (permalink / raw) To: linux-lvm On Tue, Nov 30, 2010 at 10:41 PM, Andrew Gideon <ag2827189@tagmall.com> wrote: > I'm having some difficulty following this thread, so perhaps I've missed > some other aspect to this. �But the above makes it seem like you simply > wish to transport snapshots. �Why not use something as straightforward as > dd for this rather than RAID1? �RAID1 tools (mdadm, drbd, etc.) have the > benefit of keeping volumes in sync over time. �But the above process > describes a synchronization over a single instant (the instant of the > snapshot being taken). �So why bother with the extra work/capability/ > complexity? Our posts must have crossed in the mail - please see my message immediately prior to this one. > I'm also puzzled by something you wrote in your very first message on > this thread: "the hard-linked filesystems on FILER-B require full > filesystem cloning with block-level tools rather than file-level copying > or sync'ing." �Why? �rsync -H handles cloning of hard links (though at a > performance cost). The performance cost becomes unworkable when you are talking about a filesystem with literally millions of hardlinks, as any reasonably well-used rsnapshot/rdiff-backup/BackupPC installation will create. Perhaps my current problem will be overcome by adding 64GB of RAM to my host, but as the number of hardlinks grows. . . It is precisely this problem I'm trying to address. http://www.google.co.th/search?hl=en&q=backuppc+pool+rsync+hard-link ^ permalink raw reply [flat|nested] 18+ messages in thread
[parent not found: <2D.98.07087.690A2FC4@cdptpa-omtalb.mail.rr.com>]
* Re: [linux-lvm] LVM over RAID, or plain disks? A:"Yes" = best of both worlds? [not found] ` <2D.98.07087.690A2FC4@cdptpa-omtalb.mail.rr.com> @ 2010-11-29 11:01 ` hansbkk 2010-11-29 15:29 ` Keld Jørn Simonsen 0 siblings, 1 reply; 18+ messages in thread From: hansbkk @ 2010-11-29 11:01 UTC (permalink / raw) To: Leslie Rhorer; +Cc: linux-raid, linux-lvm On Mon, Nov 29, 2010 at 1:34 AM, Leslie Rhorer <lrhorer@satx.rr.com> wrote: >> - I am not at this point overly concerned about performance issues - >> reliability/redundancy and ease of recovery are my main priorities. > > � � � �In that case, I'm not sure about your desire for the additional > complexity. �Someone else suggested RAID6, which from an operational > standpoint is much simpler. �The main reason, it seems to me, for the more > complex topologies would be enhancing perfromance. I can see how RAID6 is simpler than RAID10, but compared to RAID1? >> The mirroring at the top level would be set up between the individual >> LVs on the TopRAID1 side and regular filesystem partitions (no RAID or >> LVM) on the TopRAID2 side. In the event of a massive host failure, the >> filesystems on the TopRAID2 side could be easily mounted for data >> recovery and/or service resumption on another machine, and the >> TopRAID1 disk set rebuilt from scratch and then re-mirrored from the >> TopRAID2 disks. > > � � � �This will certainly work. �Let me ask you this: �What advantage do > you seek from using LVM on side 1? �Growth? �Re-sizing? It's a filer, servicing an ever-changing mix of dozens of clients (from a storage POV, both clients and servers), possibly using iSCSI, definitely samba and nfs. Creating new shares and/or iSCSI targets or re-configuring existing ones can't interfere with the operations of the other hosts. I can't see doing this without LVM myself. I'm likely to implement this with Openfiler, and in fact I believe it's install/configuration routines force the use of LVM for exactly this reason. >> One design goal would be to not allow any LV to get so large that it >> won't fit on a single physical disk on the TopRAID2 side. If this is >> not possible, then the corresponding TopRAID2 side would need to >> comprise a multiple disk set, perhaps striped by RAID0 - not as >> straightforward to recover as single disks, but at least without the >> LVM layer. > > � � � �To my mind, a solution which assumes the data payload will not > exceed a certain physical size is poorly considered unless the server hosts > only a very specific application whose data extents are specifically limited > by the application itself. �The beauty, as it were, of LVM is that it can > easily accommodate changing space needs within a pool of available storage > space. �If there are strict limits on the payload size, then I'm not certain > how LVM would offer advantages, especially if you are disallowing striping > the volumes. Sorry I wasn't more clear. I'm using LVM on the TopRAID1 side for exactly this reason. The TopRAID2 side is simply to enable cloning a given LV to a straightforward partition for easy disaster recovery. If and when a given LV needs to be resized, I would break the RAID1 to the TopRAID2 beforehand and manually prepare an appropriate new "static" target on the TopRAID1 side and then re-mirror. > � � � �What you are proposing can certainly be done, but it sounds awfully > frail, to me. �I think a better solution would be to be a RAID1 mirror of > every volume - each volume being a RAID or LVM of some number of disks, with > no disk being host to more than one volume on either side. �Then, rather > than failing an element, you can fail the entire volume set, stop the volume > and take all the disks offline at once. �Not only that, but if I were you, I > would make it a 3 volume mirror, keeping at least 2 volumes online at one > time. �Given your need to rotate the entire volume set, I don't believe > RAID10 will be ideal for what you want, but if not a relatively simple > matrix of RAID0 and RAID1 should work. �I'm not all that familiar with > RAID10, though, so one of the more expert members of the list may want to > chime in. > � � � �It sounds practical to me in terms of setting it up, but not in > terms of being reliable, given the way you intend to use it. Yes, I see that I shouldn't rely on the TopRAID2 side for reliability, and now plan to make the TopRAID1 side reliable in and of itself, and of course am also taking frequent backups. This now allows me to consider the TopRAID2 side as "expendable" only for the purpose of cloning a given LV to a straightforward partition, and I can test and experiment as I like. It would certainly be more conventional to just use a partition cloning tool like Acronis or BESR to make a copy of an LVM snapshot into a file; I was hoping I could use mdm instead to allow me to take normal partitions offsite ready to mount, rather than filesystems needing a restore procedure. ^ permalink raw reply [flat|nested] 18+ messages in thread
* Re: [linux-lvm] LVM over RAID, or plain disks? A:"Yes" = best of both worlds? 2010-11-29 11:01 ` [linux-lvm] " hansbkk @ 2010-11-29 15:29 ` Keld Jørn Simonsen 0 siblings, 0 replies; 18+ messages in thread From: Keld Jørn Simonsen @ 2010-11-29 15:29 UTC (permalink / raw) To: hansbkk; +Cc: Leslie Rhorer, linux-raid, linux-lvm On Mon, Nov 29, 2010 at 06:01:11PM +0700, hansbkk@gmail.com wrote: > On Mon, Nov 29, 2010 at 1:34 AM, Leslie Rhorer <lrhorer@satx.rr.com> wrote: > >> - I am not at this point overly concerned about performance issues - > >> reliability/redundancy and ease of recovery are my main priorities. > > > > � � � �In that case, I'm not sure about your desire for the additional > > complexity. �Someone else suggested RAID6, which from an operational > > standpoint is much simpler. �The main reason, it seems to me, for the more > > complex topologies would be enhancing perfromance. > > I can see how RAID6 is simpler than RAID10, but compared to RAID1? Hmm, does not compute by me. RAID1 and RAID10 are the same in complexity, RAID10 is just a modern RAID1, and should actually have been called RAID1. best regards keld ^ permalink raw reply [flat|nested] 18+ messages in thread
end of thread, other threads:[~2010-12-01 19:47 UTC | newest] Thread overview: 18+ messages (download: mbox.gz follow: Atom feed -- links below jump to the message on this page -- [not found] <AANLkTimTYse7vnqZydc1Z5z60vNuQw3juj7J_W7+bQEm@mail.gmail.com> 2010-11-28 15:31 ` [linux-lvm] Q: LVM over RAID, or plain disks? A:"Yes" = best of both worlds? hansbkk 2010-11-29 16:27 ` Lars Ellenberg 2010-11-29 17:00 ` hansbkk 2010-11-29 18:57 ` Nataraj 2010-11-30 5:20 ` hansbkk 2010-11-30 7:14 ` Nataraj 2010-11-30 7:34 ` hansbkk 2010-11-30 13:13 ` Phil Turmel 2010-11-30 15:39 ` hansbkk 2010-11-30 16:56 ` Phil Turmel 2010-12-01 4:45 ` hansbkk 2010-12-01 12:50 ` Phil Turmel 2010-12-01 19:47 ` hansbkk 2010-11-30 15:41 ` Andrew Gideon 2010-11-30 15:53 ` hansbkk 2010-11-30 15:54 ` hansbkk [not found] ` <2D.98.07087.690A2FC4@cdptpa-omtalb.mail.rr.com> 2010-11-29 11:01 ` [linux-lvm] " hansbkk 2010-11-29 15:29 ` Keld Jørn Simonsen
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox; as well as URLs for NNTP newsgroup(s).