* please help - raid 1 degraded
@ 2015-02-11 18:04 sunruh
2015-02-11 22:12 ` Adam Goryachev
0 siblings, 1 reply; 8+ messages in thread
From: sunruh @ 2015-02-11 18:04 UTC (permalink / raw)
To: linux-raid
centos 6.6
2x 240gig ssd in raid1
this is a live running production machine and the raid1 is for /u of
users home dirs.
1 ssd went totally offline and i replaced it after noticing the firmware
levels are not the same. the new ssd has the same level firmware.
/dev/sdb is the good ssd
/dev/sdc is the new blank ssd
when working it was /u1 from /dev/md127p1 and /u2 from /dev/md127p2
p1 is 80gig and p2 is 160gig for the full 240gig size of the ssd
> ls -al /dev/md*
brw-rw---- 1 root disk 9, 127 Feb 11 11:09 /dev/md127
brw-rw---- 1 root disk 259, 0 Feb 10 20:23 /dev/md127p1
brw-rw---- 1 root disk 259, 1 Feb 10 20:23 /dev/md127p2
/dev/md:
total 8
drwxr-xr-x 2 root root 140 Feb 10 20:24 .
drwxr-xr-x 20 root root 3980 Feb 10 20:24 ..
lrwxrwxrwx 1 root root 8 Feb 11 11:09 240ssd_0 -> ../md127
lrwxrwxrwx 1 root root 10 Feb 10 20:23 240ssd_0p1 -> ../md127p1
lrwxrwxrwx 1 root root 10 Feb 10 20:23 240ssd_0p2 -> ../md127p2
-rw-r--r-- 1 root root 5 Feb 10 20:24 autorebuild.pid
-rw------- 1 root root 63 Feb 10 20:23 md-device-map
> ps -eaf | grep mdadm
root 2188 1 0 Feb10 ? 00:00:00 mdadm --monitor --scan -f --pid-file=/var/run/mdadm/mdadm.pid
how do i rebuild /dev/sdc into the mirror of /dev/sdb?
and thanks much for the help!
steve
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: please help - raid 1 degraded
2015-02-11 18:04 please help - raid 1 degraded sunruh
@ 2015-02-11 22:12 ` Adam Goryachev
2015-02-12 0:09 ` sunruh
0 siblings, 1 reply; 8+ messages in thread
From: Adam Goryachev @ 2015-02-11 22:12 UTC (permalink / raw)
To: sunruh, linux-raid
On 12/02/15 05:04, sunruh@prismnet.com wrote:
> centos 6.6
> 2x 240gig ssd in raid1
> this is a live running production machine and the raid1 is for /u of
> users home dirs.
>
> 1 ssd went totally offline and i replaced it after noticing the firmware
> levels are not the same. the new ssd has the same level firmware.
>
> /dev/sdb is the good ssd
> /dev/sdc is the new blank ssd
>
> when working it was /u1 from /dev/md127p1 and /u2 from /dev/md127p2
> p1 is 80gig and p2 is 160gig for the full 240gig size of the ssd
>
>> ls -al /dev/md*
> brw-rw---- 1 root disk 9, 127 Feb 11 11:09 /dev/md127
> brw-rw---- 1 root disk 259, 0 Feb 10 20:23 /dev/md127p1
> brw-rw---- 1 root disk 259, 1 Feb 10 20:23 /dev/md127p2
>
> /dev/md:
> total 8
> drwxr-xr-x 2 root root 140 Feb 10 20:24 .
> drwxr-xr-x 20 root root 3980 Feb 10 20:24 ..
> lrwxrwxrwx 1 root root 8 Feb 11 11:09 240ssd_0 -> ../md127
> lrwxrwxrwx 1 root root 10 Feb 10 20:23 240ssd_0p1 -> ../md127p1
> lrwxrwxrwx 1 root root 10 Feb 10 20:23 240ssd_0p2 -> ../md127p2
> -rw-r--r-- 1 root root 5 Feb 10 20:24 autorebuild.pid
> -rw------- 1 root root 63 Feb 10 20:23 md-device-map
>
>> ps -eaf | grep mdadm
> root 2188 1 0 Feb10 ? 00:00:00 mdadm --monitor --scan -f --pid-file=/var/run/mdadm/mdadm.pid
>
> how do i rebuild /dev/sdc into the mirror of /dev/sdb?
>
Please send the output of fdisk -lu /dev/sd[bc] and cat /proc/mdstat
(preferably both when it was working and current).
In general, when replacing a failed RAID1 disk, and assuming you
configured it the way I think you did:
1) fdisk -lu /dev/sdb
Find out the exact partition sizes
2) fdisk /dev/sdc
Create the new partitions exactly the same as /dev/sdb
3) mdadm --manage /dev/md127 --add /dev/sdb1
Add the partition to the array
4) cat /proc/mdstat
Watch the rebuild progress, once it is complete, relax.
PS, steps 1 and 2 may not be needed if you are using the full block
device instead of a partition. Also, change the command in step 3 to
"mdadm --manage /dev/md127 --add /dev/sdb"
PPS, if this is a bootable disk, you will probably also need to do
something with your boot manager to get that installed onto the new disk
as well.
Hope this helps, otherwise, please provide more information.
Regards,
Adam
--
Adam Goryachev Website Managers www.websitemanagers.com.au
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: please help - raid 1 degraded
2015-02-11 22:12 ` Adam Goryachev
@ 2015-02-12 0:09 ` sunruh
2015-02-12 0:36 ` Adam Goryachev
2015-02-12 10:11 ` Roaming
0 siblings, 2 replies; 8+ messages in thread
From: sunruh @ 2015-02-12 0:09 UTC (permalink / raw)
To: Adam Goryachev; +Cc: sunruh, linux-raid
On Thu, Feb 12, 2015 at 09:12:50AM +1100, Adam Goryachev wrote:
> On 12/02/15 05:04, sunruh@prismnet.com wrote:
> > centos 6.6
> > 2x 240gig ssd in raid1
> > this is a live running production machine and the raid1 is for /u of
> > users home dirs.
> >
> > 1 ssd went totally offline and i replaced it after noticing the firmware
> > levels are not the same. the new ssd has the same level firmware.
> >
> > /dev/sdb is the good ssd
> > /dev/sdc is the new blank ssd
> >
> > when working it was /u1 from /dev/md127p1 and /u2 from /dev/md127p2
> > p1 is 80gig and p2 is 160gig for the full 240gig size of the ssd
> >
> >> ls -al /dev/md*
> > brw-rw---- 1 root disk 9, 127 Feb 11 11:09 /dev/md127
> > brw-rw---- 1 root disk 259, 0 Feb 10 20:23 /dev/md127p1
> > brw-rw---- 1 root disk 259, 1 Feb 10 20:23 /dev/md127p2
> >
> > /dev/md:
> > total 8
> > drwxr-xr-x 2 root root 140 Feb 10 20:24 .
> > drwxr-xr-x 20 root root 3980 Feb 10 20:24 ..
> > lrwxrwxrwx 1 root root 8 Feb 11 11:09 240ssd_0 -> ../md127
> > lrwxrwxrwx 1 root root 10 Feb 10 20:23 240ssd_0p1 -> ../md127p1
> > lrwxrwxrwx 1 root root 10 Feb 10 20:23 240ssd_0p2 -> ../md127p2
> > -rw-r--r-- 1 root root 5 Feb 10 20:24 autorebuild.pid
> > -rw------- 1 root root 63 Feb 10 20:23 md-device-map
> >
> >> ps -eaf | grep mdadm
> > root 2188 1 0 Feb10 ? 00:00:00 mdadm --monitor --scan -f --pid-file=/var/run/mdadm/mdadm.pid
> >
> > how do i rebuild /dev/sdc into the mirror of /dev/sdb?
> >
>
> Please send the output of fdisk -lu /dev/sd[bc] and cat /proc/mdstat
> (preferably both when it was working and current).
>
> In general, when replacing a failed RAID1 disk, and assuming you
> configured it the way I think you did:
> 1) fdisk -lu /dev/sdb
> Find out the exact partition sizes
> 2) fdisk /dev/sdc
> Create the new partitions exactly the same as /dev/sdb
> 3) mdadm --manage /dev/md127 --add /dev/sdb1
> Add the partition to the array
> 4) cat /proc/mdstat
> Watch the rebuild progress, once it is complete, relax.
>
> PS, steps 1 and 2 may not be needed if you are using the full block
> device instead of a partition. Also, change the command in step 3 to
> "mdadm --manage /dev/md127 --add /dev/sdb"
>
> PPS, if this is a bootable disk, you will probably also need to do
> something with your boot manager to get that installed onto the new disk
> as well.
>
> Hope this helps, otherwise, please provide more information.
>
>
> Regards,
> Adam
>
> --
> Adam Goryachev Website Managers www.websitemanagers.com.au
Adam (and anybody else that can help),
after issue i do not have before. and no they are not bootable.
[root@shell ~]# fdisk -lu /dev/sd[bc]
Disk /dev/sdb: 240.1 GB, 240057409536 bytes
255 heads, 63 sectors/track, 29185 cylinders, total 468862128 sectors
Units = sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x0001a740
Disk /dev/sdc: 240.1 GB, 240057409536 bytes
255 heads, 63 sectors/track, 29185 cylinders, total 468862128 sectors
Units = sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x00000000
[root@shell ~]# cat /proc/mdstat
Personalities : [raid1]
md127 : active raid1 sdb[2]
234299840 blocks super 1.2 [2/1] [U_]
unused devices: <none>
[root@shell ~]# fdisk -lu /dev/sdb
Disk /dev/sdb: 240.1 GB, 240057409536 bytes
255 heads, 63 sectors/track, 29185 cylinders, total 468862128 sectors
Units = sectors of 1 * 512 = 512 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x0001a740
i dont seem to be seeing the partition sizes or im stupid.
couldnt i just dd if=/dev/sdb of=/dev/sdc bs=1G count=240 and then do the
mdadm?
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: please help - raid 1 degraded
2015-02-12 0:09 ` sunruh
@ 2015-02-12 0:36 ` Adam Goryachev
2015-02-12 1:02 ` sunruh
2015-02-12 10:11 ` Roaming
1 sibling, 1 reply; 8+ messages in thread
From: Adam Goryachev @ 2015-02-12 0:36 UTC (permalink / raw)
To: sunruh; +Cc: linux-raid
On 12/02/15 11:09, sunruh@prismnet.com wrote:
> On Thu, Feb 12, 2015 at 09:12:50AM +1100, Adam Goryachev wrote:
>> On 12/02/15 05:04, sunruh@prismnet.com wrote:
>>> centos 6.6
>>> 2x 240gig ssd in raid1
>>> this is a live running production machine and the raid1 is for /u of
>>> users home dirs.
>>>
>>> 1 ssd went totally offline and i replaced it after noticing the firmware
>>> levels are not the same. the new ssd has the same level firmware.
>>>
>>> /dev/sdb is the good ssd
>>> /dev/sdc is the new blank ssd
>>>
>>> when working it was /u1 from /dev/md127p1 and /u2 from /dev/md127p2
>>> p1 is 80gig and p2 is 160gig for the full 240gig size of the ssd
>>>
>>>> ls -al /dev/md*
>>> brw-rw---- 1 root disk 9, 127 Feb 11 11:09 /dev/md127
>>> brw-rw---- 1 root disk 259, 0 Feb 10 20:23 /dev/md127p1
>>> brw-rw---- 1 root disk 259, 1 Feb 10 20:23 /dev/md127p2
>>>
>>> /dev/md:
>>> total 8
>>> drwxr-xr-x 2 root root 140 Feb 10 20:24 .
>>> drwxr-xr-x 20 root root 3980 Feb 10 20:24 ..
>>> lrwxrwxrwx 1 root root 8 Feb 11 11:09 240ssd_0 -> ../md127
>>> lrwxrwxrwx 1 root root 10 Feb 10 20:23 240ssd_0p1 -> ../md127p1
>>> lrwxrwxrwx 1 root root 10 Feb 10 20:23 240ssd_0p2 -> ../md127p2
>>> -rw-r--r-- 1 root root 5 Feb 10 20:24 autorebuild.pid
>>> -rw------- 1 root root 63 Feb 10 20:23 md-device-map
>>>
>>>> ps -eaf | grep mdadm
>>> root 2188 1 0 Feb10 ? 00:00:00 mdadm --monitor --scan -f --pid-file=/var/run/mdadm/mdadm.pid
>>>
>>> how do i rebuild /dev/sdc into the mirror of /dev/sdb?
>>>
>> Please send the output of fdisk -lu /dev/sd[bc] and cat /proc/mdstat
>> (preferably both when it was working and current).
>>
>> In general, when replacing a failed RAID1 disk, and assuming you
>> configured it the way I think you did:
>> 1) fdisk -lu /dev/sdb
>> Find out the exact partition sizes
>> 2) fdisk /dev/sdc
>> Create the new partitions exactly the same as /dev/sdb
>> 3) mdadm --manage /dev/md127 --add /dev/sdb1
>> Add the partition to the array
>> 4) cat /proc/mdstat
>> Watch the rebuild progress, once it is complete, relax.
>>
>> PS, steps 1 and 2 may not be needed if you are using the full block
>> device instead of a partition. Also, change the command in step 3 to
>> "mdadm --manage /dev/md127 --add /dev/sdb"
>>
>> PPS, if this is a bootable disk, you will probably also need to do
>> something with your boot manager to get that installed onto the new disk
>> as well.
>>
>> Hope this helps, otherwise, please provide more information.
>>
>>
>> Regards,
>> Adam
>>
>> --
>> Adam Goryachev Website Managers www.websitemanagers.com.au
> Adam (and anybody else that can help),
> after issue i do not have before. and no they are not bootable.
>
> [root@shell ~]# fdisk -lu /dev/sd[bc]
>
> Disk /dev/sdb: 240.1 GB, 240057409536 bytes
> 255 heads, 63 sectors/track, 29185 cylinders, total 468862128 sectors
> Units = sectors of 1 * 512 = 512 bytes
> Sector size (logical/physical): 512 bytes / 512 bytes
> I/O size (minimum/optimal): 512 bytes / 512 bytes
> Disk identifier: 0x0001a740
>
>
> Disk /dev/sdc: 240.1 GB, 240057409536 bytes
> 255 heads, 63 sectors/track, 29185 cylinders, total 468862128 sectors
> Units = sectors of 1 * 512 = 512 bytes
> Sector size (logical/physical): 512 bytes / 512 bytes
> I/O size (minimum/optimal): 512 bytes / 512 bytes
> Disk identifier: 0x00000000
>
> [root@shell ~]# cat /proc/mdstat
> Personalities : [raid1]
> md127 : active raid1 sdb[2]
> 234299840 blocks super 1.2 [2/1] [U_]
>
> unused devices: <none>
> i dont seem to be seeing the partition sizes or im stupid.
> couldnt i just dd if=/dev/sdb of=/dev/sdc bs=1G count=240 and then do the
> mdadm?
OK, so you aren't using partitioned disks, so it is as simple as what I
said above (with one minor correction):
"mdadm --manage /dev/md127 --add /dev/sdc"
/dev/sdc is the new blank ssd, so that is the one to add, the above
command with /dev/sdb wouldn't have done anything at all .... So just
run that command, and then do "watch cat /proc/mdstat" until the good
stuff is completed.
Regards,
Adam
--
Adam Goryachev Website Managers www.websitemanagers.com.au
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: please help - raid 1 degraded
2015-02-12 0:36 ` Adam Goryachev
@ 2015-02-12 1:02 ` sunruh
2015-02-12 1:10 ` Adam Goryachev
0 siblings, 1 reply; 8+ messages in thread
From: sunruh @ 2015-02-12 1:02 UTC (permalink / raw)
To: Adam Goryachev; +Cc: sunruh, linux-raid
On Thu, Feb 12, 2015 at 11:36:23AM +1100, Adam Goryachev wrote:
> On 12/02/15 11:09, sunruh@prismnet.com wrote:
> > On Thu, Feb 12, 2015 at 09:12:50AM +1100, Adam Goryachev wrote:
> >> On 12/02/15 05:04, sunruh@prismnet.com wrote:
> >>> centos 6.6
> >>> 2x 240gig ssd in raid1
> >>> this is a live running production machine and the raid1 is for /u of
> >>> users home dirs.
> >>>
> >>> 1 ssd went totally offline and i replaced it after noticing the firmware
> >>> levels are not the same. the new ssd has the same level firmware.
> >>>
> >>> /dev/sdb is the good ssd
> >>> /dev/sdc is the new blank ssd
> >>>
> >>> when working it was /u1 from /dev/md127p1 and /u2 from /dev/md127p2
> >>> p1 is 80gig and p2 is 160gig for the full 240gig size of the ssd
> >>>
> >>>> ls -al /dev/md*
> >>> brw-rw---- 1 root disk 9, 127 Feb 11 11:09 /dev/md127
> >>> brw-rw---- 1 root disk 259, 0 Feb 10 20:23 /dev/md127p1
> >>> brw-rw---- 1 root disk 259, 1 Feb 10 20:23 /dev/md127p2
> >>>
> >>> /dev/md:
> >>> total 8
> >>> drwxr-xr-x 2 root root 140 Feb 10 20:24 .
> >>> drwxr-xr-x 20 root root 3980 Feb 10 20:24 ..
> >>> lrwxrwxrwx 1 root root 8 Feb 11 11:09 240ssd_0 -> ../md127
> >>> lrwxrwxrwx 1 root root 10 Feb 10 20:23 240ssd_0p1 -> ../md127p1
> >>> lrwxrwxrwx 1 root root 10 Feb 10 20:23 240ssd_0p2 -> ../md127p2
> >>> -rw-r--r-- 1 root root 5 Feb 10 20:24 autorebuild.pid
> >>> -rw------- 1 root root 63 Feb 10 20:23 md-device-map
> >>>
> >>>> ps -eaf | grep mdadm
> >>> root 2188 1 0 Feb10 ? 00:00:00 mdadm --monitor --scan -f --pid-file=/var/run/mdadm/mdadm.pid
> >>>
> >>> how do i rebuild /dev/sdc into the mirror of /dev/sdb?
> >>>
> >> Please send the output of fdisk -lu /dev/sd[bc] and cat /proc/mdstat
> >> (preferably both when it was working and current).
> >>
> >> In general, when replacing a failed RAID1 disk, and assuming you
> >> configured it the way I think you did:
> >> 1) fdisk -lu /dev/sdb
> >> Find out the exact partition sizes
> >> 2) fdisk /dev/sdc
> >> Create the new partitions exactly the same as /dev/sdb
> >> 3) mdadm --manage /dev/md127 --add /dev/sdb1
> >> Add the partition to the array
> >> 4) cat /proc/mdstat
> >> Watch the rebuild progress, once it is complete, relax.
> >>
> >> PS, steps 1 and 2 may not be needed if you are using the full block
> >> device instead of a partition. Also, change the command in step 3 to
> >> "mdadm --manage /dev/md127 --add /dev/sdb"
> >>
> >> PPS, if this is a bootable disk, you will probably also need to do
> >> something with your boot manager to get that installed onto the new disk
> >> as well.
> >>
> >> Hope this helps, otherwise, please provide more information.
> >>
> >>
> >> Regards,
> >> Adam
> >>
> >> --
> >> Adam Goryachev Website Managers www.websitemanagers.com.au
> > Adam (and anybody else that can help),
> > after issue i do not have before. and no they are not bootable.
> >
> > [root@shell ~]# fdisk -lu /dev/sd[bc]
> >
> > Disk /dev/sdb: 240.1 GB, 240057409536 bytes
> > 255 heads, 63 sectors/track, 29185 cylinders, total 468862128 sectors
> > Units = sectors of 1 * 512 = 512 bytes
> > Sector size (logical/physical): 512 bytes / 512 bytes
> > I/O size (minimum/optimal): 512 bytes / 512 bytes
> > Disk identifier: 0x0001a740
> >
> >
> > Disk /dev/sdc: 240.1 GB, 240057409536 bytes
> > 255 heads, 63 sectors/track, 29185 cylinders, total 468862128 sectors
> > Units = sectors of 1 * 512 = 512 bytes
> > Sector size (logical/physical): 512 bytes / 512 bytes
> > I/O size (minimum/optimal): 512 bytes / 512 bytes
> > Disk identifier: 0x00000000
> >
> > [root@shell ~]# cat /proc/mdstat
> > Personalities : [raid1]
> > md127 : active raid1 sdb[2]
> > 234299840 blocks super 1.2 [2/1] [U_]
> >
> > unused devices: <none>
>
> > i dont seem to be seeing the partition sizes or im stupid.
> > couldnt i just dd if=/dev/sdb of=/dev/sdc bs=1G count=240 and then do the
> > mdadm?
> OK, so you aren't using partitioned disks, so it is as simple as what I
> said above (with one minor correction):
>
> "mdadm --manage /dev/md127 --add /dev/sdc"
>
>
> /dev/sdc is the new blank ssd, so that is the one to add, the above
> command with /dev/sdb wouldn't have done anything at all .... So just
> run that command, and then do "watch cat /proc/mdstat" until the good
> stuff is completed.
>
> Regards,
> Adam
>
> --
> Adam Goryachev Website Managers www.websitemanagers.com.au
awesome sauce!
it is recovering and at a fast pace too. says it will be done in 16mins.
ok, so now the really important questions:
once done, what files/stats do i need to save off for the next time it
craters?
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: please help - raid 1 degraded
2015-02-12 1:02 ` sunruh
@ 2015-02-12 1:10 ` Adam Goryachev
2015-02-12 3:12 ` Eyal Lebedinsky
0 siblings, 1 reply; 8+ messages in thread
From: Adam Goryachev @ 2015-02-12 1:10 UTC (permalink / raw)
To: sunruh; +Cc: linux-raid
On 12/02/15 12:02, sunruh@prismnet.com wrote:
> ok, so now the really important questions: once done, what files/stats
> do i need to save off for the next time it craters?
I think the usual information requested is the following:
fdisk -lu /dev/sd?
mdadm --manage --query /dev/sd?
mdadm --manage --detail /dev/md*
mdadm --manage --examine /dev/sd?
cat /proc/mdstat
ls -l /dev/disk/by-id/
If you can keep a copy of all those things, then you will be much
further ahead than many people. Of course, RAID1 is just so much
easier/simpler than RAID5/RAID6, so usually you won't need any of that.
RAID1 is simple mirror, so if you have two disks, one with data, one
without, then you just need to decide which disk has the data, and start
with that.
It is even possible to start two MD arrays, one from each disk, and then
compare the contents to decide which one you want to keep.
Or, you can simply mount the device directly (skipping any MD data at
the beginning if needed).
Like I said, RAID1 is by far the simplest type of RAID if you want
redundancy and can fit your dataset onto a single device.
Glad you had a successful recovery :)
Regards,
Adam
--
Adam Goryachev Website Managers www.websitemanagers.com.au
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: please help - raid 1 degraded
2015-02-12 1:10 ` Adam Goryachev
@ 2015-02-12 3:12 ` Eyal Lebedinsky
0 siblings, 0 replies; 8+ messages in thread
From: Eyal Lebedinsky @ 2015-02-12 3:12 UTC (permalink / raw)
Cc: linux-raid
On 12/02/15 12:10, Adam Goryachev wrote:
> On 12/02/15 12:02, sunruh@prismnet.com wrote:
>> ok, so now the really important questions: once done, what files/stats do i need to save off for the next time it craters?
>
> I think the usual information requested is the following:
> fdisk -lu /dev/sd?
> mdadm --manage --query /dev/sd?
> mdadm --manage --detail /dev/md*
> mdadm --manage --examine /dev/sd?
Maybe
mdadm --misc --query /dev/md*
mdadm --misc --detail /dev/md*
mdadm --misc --examine /dev/sd*
> cat /proc/mdstat
> ls -l /dev/disk/by-id/
>
> If you can keep a copy of all those things, then you will be much further ahead than many people. Of course, RAID1 is just so much easier/simpler than RAID5/RAID6, so usually you won't need any of that. RAID1 is simple mirror, so if you have two disks, one with data, one without, then you just need to decide which disk has the data, and start with that.
> It is even possible to start two MD arrays, one from each disk, and then compare the contents to decide which one you want to keep.
> Or, you can simply mount the device directly (skipping any MD data at the beginning if needed).
>
> Like I said, RAID1 is by far the simplest type of RAID if you want redundancy and can fit your dataset onto a single device.
>
> Glad you had a successful recovery :)
>
> Regards,
> Adam
>
--
Eyal Lebedinsky (eyal@eyal.emu.id.au)
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: please help - raid 1 degraded
2015-02-12 0:09 ` sunruh
2015-02-12 0:36 ` Adam Goryachev
@ 2015-02-12 10:11 ` Roaming
1 sibling, 0 replies; 8+ messages in thread
From: Roaming @ 2015-02-12 10:11 UTC (permalink / raw)
To: sunruh; +Cc: linux-raid
On 12/02/2015 00:09, sunruh@prismnet.com wrote:
> i dont seem to be seeing the partition sizes or im stupid.
> couldnt i just dd if=/dev/sdb of=/dev/sdc bs=1G count=240 and then do the
> mdadm?
NOT a good idea. I don't know what it would do in your case, where you
are using the entire disk, but if you're using a partition table you
would suddenly end up with a bunch of duplicate GUIDs. Bearing in mind
the "GU" stands for "globally unique", your management tools are likely
to get confused ... not a good idea especially wrt raid.
Cheers,
Wol
^ permalink raw reply [flat|nested] 8+ messages in thread
end of thread, other threads:[~2015-02-12 10:11 UTC | newest]
Thread overview: 8+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2015-02-11 18:04 please help - raid 1 degraded sunruh
2015-02-11 22:12 ` Adam Goryachev
2015-02-12 0:09 ` sunruh
2015-02-12 0:36 ` Adam Goryachev
2015-02-12 1:02 ` sunruh
2015-02-12 1:10 ` Adam Goryachev
2015-02-12 3:12 ` Eyal Lebedinsky
2015-02-12 10:11 ` Roaming
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).