linux-raid.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
* Can't add disk to failed raid array
@ 2006-07-16  0:56 Paul Waldo
  2006-07-16 10:19 ` Neil Brown
                   ` (2 more replies)
  0 siblings, 3 replies; 22+ messages in thread
From: Paul Waldo @ 2006-07-16  0:56 UTC (permalink / raw)
  To: linux-raid

Hi all,

I have a RAID6 array where a disk went bad.  I removed the old disk, put in an 
identical one, and repartitioned the new disk.  I am now trying to add the 
new partition to the array, but I get this error: 

[root@paul ~]# mdadm --add /dev/md1  /dev/hdd2
mdadm: add new device failed for /dev/hdd2 as 2: Invalid argument

When I perform that command, /var/log/messages says this:
Jul 15 20:48:39 paul kernel: md: hdd2 has invalid sb, not importing!
Jul 15 20:48:39 paul kernel: md: md_import_device returned -22


Below is the relevant data.  What might I be doing wrong?  Thanks in advance!

Paul

[root@paul ~]# cat /proc/mdstat
Personalities : [raid6] [raid1]
md0 : active raid1 hdd1[6](S) hda1[0] hdc1[1] hde1[2](S) hdg1[3](S) sda1[4](S) 
sdb1[5](S)
      979840 blocks [2/2] [UU]

md1 : active raid6 sdb2[6] sda2[5] hdg2[4] hde2[3] hdc2[1] hda2[0]
      776541440 blocks level 6, 256k chunk, algorithm 2 [7/6] [UU_UUUU]

unused devices: <none>
[root@paul ~]# fdisk -l /dev/hdd

Disk /dev/hdd: 160.0 GB, 160029999616 bytes
255 heads, 63 sectors/track, 19455 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

   Device Boot      Start         End      Blocks   Id  System
/dev/hdd1               1         122      979933+  fd  Linux raid autodetect
/dev/hdd2             123       19455   155292322+  fd  Linux raid autodetect

[root@paul log]# mdadm --detail /dev/md1
/dev/md1:
        Version : 00.90.03
  Creation Time : Fri Jun 23 22:35:27 2006
     Raid Level : raid6
     Array Size : 776541440 (740.57 GiB 795.18 GB)
    Device Size : 155308288 (148.11 GiB 159.04 GB)
   Raid Devices : 7
  Total Devices : 6
Preferred Minor : 1
    Persistence : Superblock is persistent

    Update Time : Sat Jul 15 20:53:29 2006
          State : clean, degraded
 Active Devices : 6
Working Devices : 6
 Failed Devices : 0
  Spare Devices : 0

     Chunk Size : 256K

           UUID : 2e316d9e:20cac82a:2555918e:bb9acc07
         Events : 0.1396384

    Number   Major   Minor   RaidDevice State
       0       3        2        0      active sync   /dev/hda2
       1      22        2        1      active sync   /dev/hdc2
   3157553       0        0        5      removed
       3      33        2        3      active sync   /dev/hde2
       4      34        2        4      active sync   /dev/hdg2
       5       8        2        5      active sync   /dev/sda2
       6       8       18        6      active sync   /dev/sdb2

[root@paul log]# uname -rv
2.6.17-1.2139_FC5 #1 Fri Jun 23 12:40:16 EDT 2006

^ permalink raw reply	[flat|nested] 22+ messages in thread

* Re: Can't add disk to failed raid array
  2006-07-16  0:56 Can't add disk to failed raid array Paul Waldo
@ 2006-07-16 10:19 ` Neil Brown
  2006-07-16 12:24   ` Paul Waldo
       [not found] ` <200607160913.32005.pwaldo@waldoware.com>
  2006-07-17 18:36 ` Paul Waldo
  2 siblings, 1 reply; 22+ messages in thread
From: Neil Brown @ 2006-07-16 10:19 UTC (permalink / raw)
  To: Paul Waldo; +Cc: linux-raid

On Saturday July 15, pwaldo@waldoware.com wrote:
> Hi all,
> 
> I have a RAID6 array where a disk went bad.  I removed the old disk, put in an 
> identical one, and repartitioned the new disk.  I am now trying to add the 
> new partition to the array, but I get this error: 
> 
> [root@paul ~]# mdadm --add /dev/md1  /dev/hdd2
> mdadm: add new device failed for /dev/hdd2 as 2: Invalid argument
> 
> When I perform that command, /var/log/messages says this:
> Jul 15 20:48:39 paul kernel: md: hdd2 has invalid sb, not importing!
> Jul 15 20:48:39 paul kernel: md: md_import_device returned -22


Rings a bell, but I cannot quite place it..

What version of mdadm are you running?  If not 2.5.2, try that.

...
>     Number   Major   Minor   RaidDevice State
>        0       3        2        0      active sync   /dev/hda2
>        1      22        2        1      active sync   /dev/hdc2
>    3157553       0        0        5      removed
     ^^^^^^^
That looks very odd.  If 2.5.2 does that I'll have to look into why.

NeilBrown


>        3      33        2        3      active sync   /dev/hde2
>        4      34        2        4      active sync   /dev/hdg2
>        5       8        2        5      active sync   /dev/sda2
>        6       8       18        6      active sync   /dev/sdb2
> 
> [root@paul log]# uname -rv
> 2.6.17-1.2139_FC5 #1 Fri Jun 23 12:40:16 EDT 2006
> -
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 22+ messages in thread

* Re: Can't add disk to failed raid array
  2006-07-16 10:19 ` Neil Brown
@ 2006-07-16 12:24   ` Paul Waldo
  2006-07-18  5:36     ` Neil Brown
  0 siblings, 1 reply; 22+ messages in thread
From: Paul Waldo @ 2006-07-16 12:24 UTC (permalink / raw)
  To: Neil Brown; +Cc: linux-raid

Thanks for the reply, Neil.  Here is my version:
[root@paul log]# mdadm --version
mdadm - v2.3.1 - 6 February 2006

This is a somewhat production system, running Fedora Core 5.  Official 
packages containing mdadm at version 2.5.2 aren't available (to my 
knowledge), and I am very hesitant to experiment with non-official 
software :-(

Would "mdadm --assemble" be of use to me here?  From the message "md: hdd2 has 
invalid sb, not importing!", it seems I need to get an "sb" (assumed to be 
super block) on that partition.  Would --assemble work and not destroy the 
existing good array?

As another data point, when I replaced the bad disk, I was able to 
add /dev/hdd1 to /dev/md0 with no problem.  md0 is a RAID 1 array and hdd1 
was a spare, so there may be no relation at all...

Thanks for your help!

On Sunday 16 July 2006 6:19 am, Neil Brown wrote:
> On Saturday July 15, pwaldo@waldoware.com wrote:
> > Hi all,
> >
> > I have a RAID6 array where a disk went bad.  I removed the old disk, put
> > in an identical one, and repartitioned the new disk.  I am now trying to
> > add the new partition to the array, but I get this error:
> >
> > [root@paul ~]# mdadm --add /dev/md1  /dev/hdd2
> > mdadm: add new device failed for /dev/hdd2 as 2: Invalid argument
> >
> > When I perform that command, /var/log/messages says this:
> > Jul 15 20:48:39 paul kernel: md: hdd2 has invalid sb, not importing!
> > Jul 15 20:48:39 paul kernel: md: md_import_device returned -22
>
> Rings a bell, but I cannot quite place it..
>
> What version of mdadm are you running?  If not 2.5.2, try that.
>
> ...
>
> >     Number   Major   Minor   RaidDevice State
> >        0       3        2        0      active sync   /dev/hda2
> >        1      22        2        1      active sync   /dev/hdc2
> >    3157553       0        0        5      removed
>
>      ^^^^^^^
> That looks very odd.  If 2.5.2 does that I'll have to look into why.
>
> NeilBrown
>
> >        3      33        2        3      active sync   /dev/hde2
> >        4      34        2        4      active sync   /dev/hdg2
> >        5       8        2        5      active sync   /dev/sda2
> >        6       8       18        6      active sync   /dev/sdb2
> >
> > [root@paul log]# uname -rv
> > 2.6.17-1.2139_FC5 #1 Fri Jun 23 12:40:16 EDT 2006
> > -
> > To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> > the body of a message to majordomo@vger.kernel.org
> > More majordomo info at  http://vger.kernel.org/majordomo-info.html
>
> -
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 22+ messages in thread

* Re: Can't add disk to failed raid array
       [not found]   ` <Pine.LNX.4.62.0607161524170.7520@uplift.swm.pp.se>
@ 2006-07-16 14:39     ` Paul Waldo
  0 siblings, 0 replies; 22+ messages in thread
From: Paul Waldo @ 2006-07-16 14:39 UTC (permalink / raw)
  To: linux-raid; +Cc: Mikael Abrahamsson

Still no joy after zeroing the superblock,  Mikael.  :-(

[root@paul ~]# mdadm --examine /dev/hdd2
/dev/hdd2:
          Magic : a92b4efc
        Version : 00.90.00
           UUID : 2e316d9e:20cac82a:2555918e:bb9acc07
  Creation Time : Fri Jun 23 22:35:27 2006
     Raid Level : raid6
    Device Size : 155308288 (148.11 GiB 159.04 GB)
     Array Size : 776541440 (740.57 GiB 795.18 GB)
   Raid Devices : 7
  Total Devices : 6
Preferred Minor : 1

    Update Time : Sun Jul 16 09:10:41 2006
          State : clean
 Active Devices : 6
Working Devices : 6
 Failed Devices : 1
  Spare Devices : 0
       Checksum : 6bdf5797 - correct
         Events : 0.1401144


      Number   Major   Minor   RaidDevice State
this     2      22       66       -1      sync   /dev/hdd2

   0     0       3        2        0      active sync   /dev/hda2
   1     1      22        2        1      active sync   /dev/hdc2
   2     2      22       66       -1      sync   /dev/hdd2
   3     3      33        2        3      active sync   /dev/hde2
   4     4      34        2        4      active sync   /dev/hdg2
   5     5       8        2        5      active sync   /dev/sda2
   6     6       8       18        6      active sync   /dev/sdb2
[root@paul ~]# cat /proc/mdstat
Personalities : [raid6] [raid1]
md0 : active raid1 hda1[0] hdc1[1] hdd1[2](S) hde1[3](S) hdg1[4](S) sda1[5](S) 
sdb1[6](S)
      979840 blocks [2/2] [UU]

md1 : active raid6 sdb2[6] sda2[5] hdg2[4] hde2[3] hdc2[1] hda2[0]
      776541440 blocks level 6, 256k chunk, algorithm 2 [7/6] [UU_UUUU]

unused devices: <none>
[root@paul ~]# mdadm --examine /dev/hdd2|grep -i super
[root@paul ~]# mdadm --misc --zero-superblock /dev/hdd2
[root@paul ~]# cat /proc/mdstat
Personalities : [raid6] [raid1]
md0 : active raid1 hda1[0] hdc1[1] hdd1[2](S) hde1[3](S) hdg1[4](S) sda1[5](S) 
sdb1[6](S)
      979840 blocks [2/2] [UU]

md1 : active raid6 sdb2[6] sda2[5] hdg2[4] hde2[3] hdc2[1] hda2[0]
      776541440 blocks level 6, 256k chunk, algorithm 2 [7/6] [UU_UUUU]

unused devices: <none>
[root@paul ~]# mdadm --add /dev/md1  /dev/hdd2
mdadm: add new device failed for /dev/hdd2 as 2: Invalid argument
[root@paul ~]# tail -5 /var/log/messages
Jul 16 09:10:42 paul kernel: md: md_import_device returned -22
Jul 16 09:10:42 paul kernel: md: hdd2 has invalid sb, not importing!
Jul 16 09:10:42 paul kernel: md: md_import_device returned -22
Jul 16 10:33:33 paul kernel: md: hdd2 has invalid sb, not importing!
Jul 16 10:33:33 paul kernel: md: md_import_device returned -22


On Sunday 16 July 2006 9:26 am, you wrote:
> On Sun, 16 Jul 2006, Paul Waldo wrote:
>
> The superblock is at the end so you probably didnt clear it. Use "mdadm
> --examine /dev/hdd2" and see what it says. If it does say it has a
> superblock, use "--misc --zero-superblock /dev/hdd2"

^ permalink raw reply	[flat|nested] 22+ messages in thread

* Re: Can't add disk to failed raid array
  2006-07-16  0:56 Can't add disk to failed raid array Paul Waldo
  2006-07-16 10:19 ` Neil Brown
       [not found] ` <200607160913.32005.pwaldo@waldoware.com>
@ 2006-07-17 18:36 ` Paul Waldo
  2 siblings, 0 replies; 22+ messages in thread
From: Paul Waldo @ 2006-07-17 18:36 UTC (permalink / raw)
  To: linux-raid

All has been quiet on this topic for a while--any more takers?  Please 
help if you can!  Thanks in advance.  Here is the current state of affairs:

[root@paul ~]# mdadm --add /dev/md1 /dev/hdd2
mdadm: add new device failed for /dev/hdd2 as 2: Invalid argument
[root@paul ~]# mdadm --detail /dev/md1
/dev/md1:
         Version : 00.90.03
   Creation Time : Fri Jun 23 22:35:27 2006
      Raid Level : raid6
      Array Size : 776541440 (740.57 GiB 795.18 GB)
     Device Size : 155308288 (148.11 GiB 159.04 GB)
    Raid Devices : 7
   Total Devices : 6
Preferred Minor : 1
     Persistence : Superblock is persistent

     Update Time : Mon Jul 17 14:02:13 2006
           State : clean, degraded
  Active Devices : 6
Working Devices : 6
  Failed Devices : 0
   Spare Devices : 0

      Chunk Size : 256K

            UUID : 2e316d9e:20cac82a:2555918e:bb9acc07
          Events : 0.1416894

     Number   Major   Minor   RaidDevice State
        0       3        2        0      active sync   /dev/hda2
        1      22        2        1      active sync   /dev/hdc2
        0       0        0    1883272037      removed
        3      33        2        3      active sync   /dev/hde2
        4      34        2        4      active sync   /dev/hdg2
        5       8        2        5      active sync   /dev/sda2
        6       8       18        6      active sync   /dev/sdb2
[root@paul ~]# fdisk -l /dev/hdd

Disk /dev/hdd: 160.0 GB, 160029999616 bytes
255 heads, 63 sectors/track, 19455 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

    Device Boot      Start         End      Blocks   Id  System
/dev/hdd1               1         122      979933+  fd  Linux raid 
autodetect
/dev/hdd2             123       19455   155292322+  fd  Linux raid 
autodetect

[root@paul ~]# tail -5 /var/log/messages
Jul 17 14:02:08 paul kernel: md: md_import_device returned -22
Jul 17 14:35:47 paul kernel: md: hdd2 has invalid sb, not importing!
Jul 17 14:35:47 paul kernel: md: md_import_device returned -22
Jul 17 14:35:47 paul kernel: md: hdd2 has invalid sb, not importing!
Jul 17 14:35:47 paul kernel: md: md_import_device returned -22


Paul Waldo wrote:
> Hi all,
> 
> I have a RAID6 array where a disk went bad.  I removed the old disk, put in an 
> identical one, and repartitioned the new disk.  I am now trying to add the 
> new partition to the array, but I get this error: 
> 
> [root@paul ~]# mdadm --add /dev/md1  /dev/hdd2
> mdadm: add new device failed for /dev/hdd2 as 2: Invalid argument
> 
> When I perform that command, /var/log/messages says this:
> Jul 15 20:48:39 paul kernel: md: hdd2 has invalid sb, not importing!
> Jul 15 20:48:39 paul kernel: md: md_import_device returned -22
> 
> 
> Below is the relevant data.  What might I be doing wrong?  Thanks in advance!
> 
> Paul
> 
> [root@paul ~]# cat /proc/mdstat
> Personalities : [raid6] [raid1]
> md0 : active raid1 hdd1[6](S) hda1[0] hdc1[1] hde1[2](S) hdg1[3](S) sda1[4](S) 
> sdb1[5](S)
>       979840 blocks [2/2] [UU]
> 
> md1 : active raid6 sdb2[6] sda2[5] hdg2[4] hde2[3] hdc2[1] hda2[0]
>       776541440 blocks level 6, 256k chunk, algorithm 2 [7/6] [UU_UUUU]
> 
> unused devices: <none>
> [root@paul ~]# fdisk -l /dev/hdd
> 
> Disk /dev/hdd: 160.0 GB, 160029999616 bytes
> 255 heads, 63 sectors/track, 19455 cylinders
> Units = cylinders of 16065 * 512 = 8225280 bytes
> 
>    Device Boot      Start         End      Blocks   Id  System
> /dev/hdd1               1         122      979933+  fd  Linux raid autodetect
> /dev/hdd2             123       19455   155292322+  fd  Linux raid autodetect
> 
> [root@paul log]# mdadm --detail /dev/md1
> /dev/md1:
>         Version : 00.90.03
>   Creation Time : Fri Jun 23 22:35:27 2006
>      Raid Level : raid6
>      Array Size : 776541440 (740.57 GiB 795.18 GB)
>     Device Size : 155308288 (148.11 GiB 159.04 GB)
>    Raid Devices : 7
>   Total Devices : 6
> Preferred Minor : 1
>     Persistence : Superblock is persistent
> 
>     Update Time : Sat Jul 15 20:53:29 2006
>           State : clean, degraded
>  Active Devices : 6
> Working Devices : 6
>  Failed Devices : 0
>   Spare Devices : 0
> 
>      Chunk Size : 256K
> 
>            UUID : 2e316d9e:20cac82a:2555918e:bb9acc07
>          Events : 0.1396384
> 
>     Number   Major   Minor   RaidDevice State
>        0       3        2        0      active sync   /dev/hda2
>        1      22        2        1      active sync   /dev/hdc2
>    3157553       0        0        5      removed
>        3      33        2        3      active sync   /dev/hde2
>        4      34        2        4      active sync   /dev/hdg2
>        5       8        2        5      active sync   /dev/sda2
>        6       8       18        6      active sync   /dev/sdb2
> 
> [root@paul log]# uname -rv
> 2.6.17-1.2139_FC5 #1 Fri Jun 23 12:40:16 EDT 2006


^ permalink raw reply	[flat|nested] 22+ messages in thread

* Re: Can't add disk to failed raid array
  2006-07-16 12:24   ` Paul Waldo
@ 2006-07-18  5:36     ` Neil Brown
  2006-07-23 11:04       ` In Trouble--Please Help! (was Re: Can't add disk to failed raid array) Paul Waldo
  0 siblings, 1 reply; 22+ messages in thread
From: Neil Brown @ 2006-07-18  5:36 UTC (permalink / raw)
  To: Paul Waldo; +Cc: linux-raid

On Sunday July 16, pwaldo@waldoware.com wrote:
> Thanks for the reply, Neil.  Here is my version:
> [root@paul log]# mdadm --version
> mdadm - v2.3.1 - 6 February 2006

Positively ancient :-)  Nothing obvious in the change log since then.

Can you show me the output of 
  mdadm -E /dev/hdd2
  mdadm -E /dev/hda2

immediately after the failed attempt to add hdd2.

Thanks,
NeilBrown

^ permalink raw reply	[flat|nested] 22+ messages in thread

* In Trouble--Please Help!  (was Re: Can't add disk to failed raid array)
  2006-07-18  5:36     ` Neil Brown
@ 2006-07-23 11:04       ` Paul Waldo
  2006-07-23 11:25         ` Neil Brown
  0 siblings, 1 reply; 22+ messages in thread
From: Paul Waldo @ 2006-07-23 11:04 UTC (permalink / raw)
  To: Neil Brown; +Cc: linux-raid

Please, please! I am dead in the water!

To recap, I have a RAID6 array on a Fedora Core 5 system, using /dev/hd[acdeg]
2 and /dev/sd[ab]2.  /dev/hdd went bad so I replaced the drive and tried to 
add it back to the array.  Here is what happens:

#mdadm --assemble /dev/md1 /dev/hd[acdeg]2 /dev/sd[ab]2
mdadm: no RAID superblock on /dev/hdd2
mdadm: /dev/hdd2 has no superblock - assembly aborted

I also tried assembling without the new drive:
#mdadm --assemble /dev/md1 /dev/hd[aceg]2 /dev/sd[ab]2
mdadm: failed to RUN_ARRAY /dev/md1: Input/output error

Forgive me if the messages are not exactly correct, as I am booted into the FC 
rescue disk and I am transcribing what I see to another computer.

How can I get mdadm to use the new drive?  Am I completely fscked?  Thanks in 
advance for any help!!!!

Paul


On Tuesday 18 July 2006 01:36, Neil Brown wrote:
> On Sunday July 16, pwaldo@waldoware.com wrote:
> > Thanks for the reply, Neil.  Here is my version:
> > [root@paul log]# mdadm --version
> > mdadm - v2.3.1 - 6 February 2006
>
> Positively ancient :-)  Nothing obvious in the change log since then.
>
> Can you show me the output of
>   mdadm -E /dev/hdd2
>   mdadm -E /dev/hda2
>
> immediately after the failed attempt to add hdd2.
>
> Thanks,
> NeilBrown
> -
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 22+ messages in thread

* Re: In Trouble--Please Help!  (was Re: Can't add disk to failed raid array)
  2006-07-23 11:04       ` In Trouble--Please Help! (was Re: Can't add disk to failed raid array) Paul Waldo
@ 2006-07-23 11:25         ` Neil Brown
  2006-07-23 11:50           ` Paul Waldo
  2006-07-23 11:53           ` Paul Waldo
  0 siblings, 2 replies; 22+ messages in thread
From: Neil Brown @ 2006-07-23 11:25 UTC (permalink / raw)
  To: Paul Waldo; +Cc: linux-raid

On Sunday July 23, pwaldo@waldoware.com wrote:
> Please, please! I am dead in the water!
> 
> To recap, I have a RAID6 array on a Fedora Core 5 system, using /dev/hd[acdeg]
> 2 and /dev/sd[ab]2.  /dev/hdd went bad so I replaced the drive and tried to 
> add it back to the array.  Here is what happens:
> 
> #mdadm --assemble /dev/md1 /dev/hd[acdeg]2 /dev/sd[ab]2
> mdadm: no RAID superblock on /dev/hdd2
> mdadm: /dev/hdd2 has no superblock - assembly aborted
> 
> I also tried assembling without the new drive:
> #mdadm --assemble /dev/md1 /dev/hd[aceg]2 /dev/sd[ab]2
> mdadm: failed to RUN_ARRAY /dev/md1: Input/output error
> 
> Forgive me if the messages are not exactly correct, as I am booted into the FC 
> rescue disk and I am transcribing what I see to another computer.
> 
> How can I get mdadm to use the new drive?  Am I completely fscked?  Thanks in 
> advance for any help!!!!

Sorry for not following through on this earlier.
I think I know what the problem is.

 From the "-E" output you have me :
     Device Size : 155308288 (148.11 GiB 159.04 GB)

 From the fdisk output in the original email:

[root@paul ~]# fdisk -l /dev/hdd

Disk /dev/hdd: 160.0 GB, 160029999616 bytes
255 heads, 63 sectors/track, 19455 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

   Device Boot      Start         End      Blocks   Id  System
/dev/hdd1               1         122      979933+  fd  Linux raid autodetect
/dev/hdd2             123       19455   155292322+  fd  Linux raid autodetect


Notice that hdd2 is 155292322, but the device needs to be 
155308288.  It isn't quite big enough.
mdadm should pick this up, but obviously doesn't, and the error message
isn't at all helpful.
I will fix that in the next release.
For now, repartition your drive so that hdd2 is slightly larger.

NeilBrown

^ permalink raw reply	[flat|nested] 22+ messages in thread

* Re: In Trouble--Please Help!  (was Re: Can't add disk to failed raid array)
  2006-07-23 11:25         ` Neil Brown
@ 2006-07-23 11:50           ` Paul Waldo
  2006-07-23 11:53           ` Paul Waldo
  1 sibling, 0 replies; 22+ messages in thread
From: Paul Waldo @ 2006-07-23 11:50 UTC (permalink / raw)
  To: Neil Brown; +Cc: linux-raid

On Sunday 23 July 2006 07:25, Neil Brown wrote:
> On Sunday July 23, pwaldo@waldoware.com wrote:
> > Please, please! I am dead in the water!
> >
> > To recap, I have a RAID6 array on a Fedora Core 5 system, using
> > /dev/hd[acdeg] 2 and /dev/sd[ab]2.  /dev/hdd went bad so I replaced the
> > drive and tried to add it back to the array.  Here is what happens:
> >
> > #mdadm --assemble /dev/md1 /dev/hd[acdeg]2 /dev/sd[ab]2
> > mdadm: no RAID superblock on /dev/hdd2
> > mdadm: /dev/hdd2 has no superblock - assembly aborted
> >
> > I also tried assembling without the new drive:
> > #mdadm --assemble /dev/md1 /dev/hd[aceg]2 /dev/sd[ab]2
> > mdadm: failed to RUN_ARRAY /dev/md1: Input/output error
> >
> > Forgive me if the messages are not exactly correct, as I am booted into
> > the FC rescue disk and I am transcribing what I see to another computer.
> >
> > How can I get mdadm to use the new drive?  Am I completely fscked? 
> > Thanks in advance for any help!!!!
>
> Sorry for not following through on this earlier.
> I think I know what the problem is.
>
>  From the "-E" output you have me :
>      Device Size : 155308288 (148.11 GiB 159.04 GB)
>
>  From the fdisk output in the original email:
>
> [root@paul ~]# fdisk -l /dev/hdd
>
> Disk /dev/hdd: 160.0 GB, 160029999616 bytes
> 255 heads, 63 sectors/track, 19455 cylinders
> Units = cylinders of 16065 * 512 = 8225280 bytes
>
>    Device Boot      Start         End      Blocks   Id  System
> /dev/hdd1               1         122      979933+  fd  Linux raid
> autodetect /dev/hdd2             123       19455   155292322+  fd  Linux
> raid autodetect
>
>
> Notice that hdd2 is 155292322, but the device needs to be
> 155308288.  It isn't quite big enough.
> mdadm should pick this up, but obviously doesn't, and the error message
> isn't at all helpful.
> I will fix that in the next release.
> For now, repartition your drive so that hdd2 is slightly larger.
>
> NeilBrown

Hi Neil,

Thanks for the quick reply.  I tried your suggestion.  I now see 
that /dev/hda2 (good partition) is 155308387 blocks.  I 
repartitioned /dev/hdd such that /dev/hdd2 is also 155308387, but I still get 
the message about no superblock on hdd2 :-(.  I used --force.

Paul

^ permalink raw reply	[flat|nested] 22+ messages in thread

* Re: In Trouble--Please Help!  (was Re: Can't add disk to failed raid array)
  2006-07-23 11:25         ` Neil Brown
  2006-07-23 11:50           ` Paul Waldo
@ 2006-07-23 11:53           ` Paul Waldo
  2006-07-23 11:59             ` Neil Brown
  1 sibling, 1 reply; 22+ messages in thread
From: Paul Waldo @ 2006-07-23 11:53 UTC (permalink / raw)
  To: Neil Brown; +Cc: linux-raid

On Sunday 23 July 2006 07:25, Neil Brown wrote:
> On Sunday July 23, pwaldo@waldoware.com wrote:
> > Please, please! I am dead in the water!
> >
> > To recap, I have a RAID6 array on a Fedora Core 5 system, using
> > /dev/hd[acdeg] 2 and /dev/sd[ab]2.  /dev/hdd went bad so I replaced the
> > drive and tried to add it back to the array.  Here is what happens:
> >
> > #mdadm --assemble /dev/md1 /dev/hd[acdeg]2 /dev/sd[ab]2
> > mdadm: no RAID superblock on /dev/hdd2
> > mdadm: /dev/hdd2 has no superblock - assembly aborted
> >
> > I also tried assembling without the new drive:
> > #mdadm --assemble /dev/md1 /dev/hd[aceg]2 /dev/sd[ab]2
> > mdadm: failed to RUN_ARRAY /dev/md1: Input/output error
> >
> > Forgive me if the messages are not exactly correct, as I am booted into
> > the FC rescue disk and I am transcribing what I see to another computer.
> >
> > How can I get mdadm to use the new drive?  Am I completely fscked? 
> > Thanks in advance for any help!!!!
>
> Sorry for not following through on this earlier.
> I think I know what the problem is.
>
>  From the "-E" output you have me :
>      Device Size : 155308288 (148.11 GiB 159.04 GB)
>
>  From the fdisk output in the original email:
>
> [root@paul ~]# fdisk -l /dev/hdd
>
> Disk /dev/hdd: 160.0 GB, 160029999616 bytes
> 255 heads, 63 sectors/track, 19455 cylinders
> Units = cylinders of 16065 * 512 = 8225280 bytes
>
>    Device Boot      Start         End      Blocks   Id  System
> /dev/hdd1               1         122      979933+  fd  Linux raid
> autodetect /dev/hdd2             123       19455   155292322+  fd  Linux
> raid autodetect
>
>
> Notice that hdd2 is 155292322, but the device needs to be
> 155308288.  It isn't quite big enough.
> mdadm should pick this up, but obviously doesn't, and the error message
> isn't at all helpful.
> I will fix that in the next release.
> For now, repartition your drive so that hdd2 is slightly larger.
>
> NeilBrown


At this point, I'd just be happy to be able to get the degraded array back up 
and running.  Is there any way to do that?  Thanks!

Paul

^ permalink raw reply	[flat|nested] 22+ messages in thread

* Re: In Trouble--Please Help!  (was Re: Can't add disk to failed raid array)
  2006-07-23 11:53           ` Paul Waldo
@ 2006-07-23 11:59             ` Neil Brown
  2006-07-23 12:32               ` Paul Waldo
  0 siblings, 1 reply; 22+ messages in thread
From: Neil Brown @ 2006-07-23 11:59 UTC (permalink / raw)
  To: Paul Waldo; +Cc: linux-raid

On Sunday July 23, pwaldo@waldoware.com wrote:
> 
> 
> At this point, I'd just be happy to be able to get the degraded array back up 
> and running.  Is there any way to do that?  Thanks!

 mdadm --assemble --force /dev/md1 /dev/hd[aceg]2 /dev/sd[ab]2

should get you the degraded array.  If not, what kernel log messages
do you get?



^ permalink raw reply	[flat|nested] 22+ messages in thread

* Re: In Trouble--Please Help!  (was Re: Can't add disk to failed raid array)
  2006-07-23 11:59             ` Neil Brown
@ 2006-07-23 12:32               ` Paul Waldo
  2006-07-24 18:26                 ` Dan Williams
  2006-07-24 23:24                 ` Neil Brown
  0 siblings, 2 replies; 22+ messages in thread
From: Paul Waldo @ 2006-07-23 12:32 UTC (permalink / raw)
  To: Neil Brown; +Cc: linux-raid

[-- Attachment #1: Type: text/plain, Size: 480 bytes --]

Here is the dmesg output.  No log files are created with the FC5 rescue disk.  
Thanks!



On Sunday 23 July 2006 07:59, Neil Brown wrote:
> On Sunday July 23, pwaldo@waldoware.com wrote:
> > At this point, I'd just be happy to be able to get the degraded array
> > back up and running.  Is there any way to do that?  Thanks!
>
>  mdadm --assemble --force /dev/md1 /dev/hd[aceg]2 /dev/sd[ab]2
>
> should get you the degraded array.  If not, what kernel log messages
> do you get?

[-- Attachment #2: dmesg.log --]
[-- Type: text/x-log, Size: 7704 bytes --]

[APCG] enabled at IRQ 20
ACPI: PCI Interrupt 0000:00:02.1[B] -> Link [APCG] -> GSI 20 (level, high) -> IRQ 19
PCI: Setting latency timer of device 0000:00:02.1 to 64
ohci_hcd 0000:00:02.1: OHCI Host Controller
ohci_hcd 0000:00:02.1: new USB bus registered, assigned bus number 3
ohci_hcd 0000:00:02.1: irq 19, io mem 0xe2082000
usb usb3: configuration #1 chosen from 1 choice
hub 3-0:1.0: USB hub found
hub 3-0:1.0: 3 ports detected
usb 1-3: configuration #1 chosen from 1 choice
Initializing USB Mass Storage driver...
scsi0 : SCSI emulation for USB Mass Storage devices
usb-storage: device found at 2
usb-storage: waiting for device to settle before scanning
usbcore: registered new driver usb-storage
USB Mass Storage support registered.
  Vendor: HL-DT-ST  Model: DVDRRW GSA-2166D  Rev: 1.01
  Type:   CD-ROM                             ANSI SCSI revision: 00
sr0: scsi3-mmc drive: 48x/48x writer dvd-ram cd/rw xa/form2 cdda tray
Uniform CD-ROM driver Revision: 3.20
sr 0:0:0:0: Attached scsi CD-ROM sr0
usb-storage: device scan complete
libata version 1.20 loaded.
sata_sil 0000:01:0d.0: version 0.9
ACPI: PCI Interrupt Link [APC3] enabled at IRQ 18
ACPI: PCI Interrupt 0000:01:0d.0[A] -> Link [APC3] -> GSI 18 (level, high) -> IRQ 20
ata1: SATA max UDMA/100 cmd 0xE090C080 ctl 0xE090C08A bmdma 0xE090C000 irq 20
ata2: SATA max UDMA/100 cmd 0xE090C0C0 ctl 0xE090C0CA bmdma 0xE090C008 irq 20
ata1: SATA link up 1.5 Gbps (SStatus 113)
ata1: dev 0 cfg 49:2f00 82:346b 83:7f01 84:4003 85:3c69 86:3c01 87:4003 88:20ff
ata1: dev 0 ATA-7, max UDMA7, 312581808 sectors: LBA48
ata1: dev 0 configured for UDMA/100
scsi1 : sata_sil
ata2: SATA link up 1.5 Gbps (SStatus 113)
ata2: dev 0 cfg 49:2f00 82:346b 83:7f01 84:4003 85:3c69 86:3c01 87:4003 88:20ff
ata2: dev 0 ATA-7, max UDMA7, 312581808 sectors: LBA48
ata2: dev 0 configured for UDMA/100
scsi2 : sata_sil
  Vendor: ATA       Model: SAMSUNG SP1614C   Rev: SW10
  Type:   Direct-Access                      ANSI SCSI revision: 05
SCSI device sda: 312581808 512-byte hdwr sectors (160042 MB)
sda: Write Protect is off
sda: Mode Sense: 00 3a 00 00
SCSI device sda: drive cache: write back
SCSI device sda: 312581808 512-byte hdwr sectors (160042 MB)
sda: Write Protect is off
sda: Mode Sense: 00 3a 00 00
SCSI device sda: drive cache: write back
 sda: sda1 sda2
sd 1:0:0:0: Attached scsi disk sda
  Vendor: ATA       Model: SAMSUNG SP1614C   Rev: SW10
  Type:   Direct-Access                      ANSI SCSI revision: 05
SCSI device sdb: 312581808 512-byte hdwr sectors (160042 MB)
sdb: Write Protect is off
sdb: Mode Sense: 00 3a 00 00
SCSI device sdb: drive cache: write back
SCSI device sdb: 312581808 512-byte hdwr sectors (160042 MB)
sdb: Write Protect is off
sdb: Mode Sense: 00 3a 00 00
SCSI device sdb: drive cache: write back
 sdb: sdb1 sdb2
sd 2:0:0:0: Attached scsi disk sdb
forcedeth.c: Reverse Engineered nForce ethernet driver. Version 0.49.
ACPI: PCI Interrupt Link [APCH] enabled at IRQ 22
ACPI: PCI Interrupt 0000:00:04.0[A] -> Link [APCH] -> GSI 22 (level, high) -> IRQ 17
PCI: Setting latency timer of device 0000:00:04.0 to 64
eth0: forcedeth.c: subsystem: 010de:05b2 bound to 0000:00:04.0
ISO 9660 Extensions: Microsoft Joliet Level 3
Unable to load NLS charset utf8
Unable to load NLS charset utf8
ISO 9660 Extensions: RRIP_1991A
Unable to identify CD-ROM format.
VFS: Can't find an ext2 filesystem on dev loop0.
security:  3 users, 6 roles, 1161 types, 135 bools, 1 sens, 256 cats
security:  55 classes, 38679 rules
SELinux:  Completing initialization.
SELinux:  Setting up existing superblocks.
SELinux: initialized (dev loop0, type squashfs), not configured for labeling
SELinux: initialized (dev usbfs, type usbfs), uses genfs_contexts
SELinux: initialized (dev ramfs, type ramfs), uses genfs_contexts
SELinux: initialized (dev ramfs, type ramfs), uses genfs_contexts
SELinux: initialized (dev tmpfs, type tmpfs), uses transition SIDs
SELinux: initialized (dev debugfs, type debugfs), uses genfs_contexts
SELinux: initialized (dev selinuxfs, type selinuxfs), uses genfs_contexts
SELinux: initialized (dev mqueue, type mqueue), uses transition SIDs
SELinux: initialized (dev hugetlbfs, type hugetlbfs), uses genfs_contexts
SELinux: initialized (dev devpts, type devpts), uses transition SIDs
SELinux: initialized (dev eventpollfs, type eventpollfs), uses genfs_contexts
SELinux: initialized (dev inotifyfs, type inotifyfs), uses genfs_contexts
SELinux: initialized (dev tmpfs, type tmpfs), uses transition SIDs
SELinux: initialized (dev futexfs, type futexfs), uses genfs_contexts
SELinux: initialized (dev pipefs, type pipefs), uses task SIDs
SELinux: initialized (dev sockfs, type sockfs), uses task SIDs
SELinux: initialized (dev proc, type proc), uses genfs_contexts
SELinux: initialized (dev bdev, type bdev), uses genfs_contexts
SELinux: initialized (dev rootfs, type rootfs), uses genfs_contexts
SELinux: initialized (dev sysfs, type sysfs), uses genfs_contexts
audit(1153657490.719:2): avc:  denied  { transition } for  pid=532 comm="loader" name="bash" dev=loop0 ino=1485 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:anaconda_t:s0 tclass=process
md: raid0 personality registered for level 0
md: raid1 personality registered for level 1
raid5: automatically using best checksumming function: pIII_sse
   pIII_sse  :  4977.000 MB/sec
raid5: using function: pIII_sse (4977.000 MB/sec)
md: raid5 personality registered for level 5
md: raid4 personality registered for level 4
raid6: int32x1    757 MB/s
raid6: int32x2    835 MB/s
raid6: int32x4    779 MB/s
raid6: int32x8    598 MB/s
raid6: mmxx1     1687 MB/s
raid6: mmxx2     2966 MB/s
raid6: sse1x1    1550 MB/s
raid6: sse1x2    2705 MB/s
raid6: using algorithm sse1x2 (2705 MB/s)
md: raid6 personality registered for level 6
JFS: nTxBlock = 4030, nTxLock = 32241
SGI XFS with ACLs, security attributes, large block numbers, no debug enabled
SGI XFS Quota Management subsystem
device-mapper: 4.5.0-ioctl (2005-10-04) initialised: dm-devel@redhat.com
program anaconda is using a deprecated SCSI ioctl, please convert it to SG_IO
program anaconda is using a deprecated SCSI ioctl, please convert it to SG_IO
usb 1-6: new high speed USB device using ehci_hcd and address 3
usb 1-6: configuration #1 chosen from 1 choice
scsi3 : SCSI emulation for USB Mass Storage devices
usb-storage: device found at 3
usb-storage: waiting for device to settle before scanning
  Vendor: SanDisk   Model: Cruzer Mini       Rev: 0.1 
  Type:   Direct-Access                      ANSI SCSI revision: 02
SCSI device sdc: 250879 512-byte hdwr sectors (128 MB)
sdc: Write Protect is off
sdc: Mode Sense: 03 00 00 00
sdc: assuming drive cache: write through
SCSI device sdc: 250879 512-byte hdwr sectors (128 MB)
sdc: Write Protect is off
sdc: Mode Sense: 03 00 00 00
sdc: assuming drive cache: write through
 sdc: sdc1
sd 3:0:0:0: Attached scsi removable disk sdc
usb-storage: device scan complete
md: md1 stopped.
md: bind<hdc2>
md: bind<hde2>
md: bind<hdg2>
md: bind<sda2>
md: bind<sdb2>
md: bind<hda2>
md: md1: raid array is not clean -- starting background reconstruction
raid6: device hda2 operational as raid disk 0
raid6: device sdb2 operational as raid disk 6
raid6: device sda2 operational as raid disk 5
raid6: device hdg2 operational as raid disk 4
raid6: device hde2 operational as raid disk 3
raid6: device hdc2 operational as raid disk 1
raid6: cannot start dirty degraded array for md1
RAID6 conf printout:
 --- rd:7 wd:6 fd:1
 disk 0, o:1, dev:hda2
 disk 1, o:1, dev:hdc2
 disk 3, o:1, dev:hde2
 disk 4, o:1, dev:hdg2
 disk 5, o:1, dev:sda2
 disk 6, o:1, dev:sdb2
raid6: failed to run raid set md1
md: pers->run() failed ...
ext3: No journal on filesystem on fd0

^ permalink raw reply	[flat|nested] 22+ messages in thread

* Re: In Trouble--Please Help! (was Re: Can't add disk to failed raid array)
  2006-07-23 12:32               ` Paul Waldo
@ 2006-07-24 18:26                 ` Dan Williams
  2006-07-24 18:35                   ` Paul Waldo
  2006-07-25 14:27                   ` Paul Waldo
  2006-07-24 23:24                 ` Neil Brown
  1 sibling, 2 replies; 22+ messages in thread
From: Dan Williams @ 2006-07-24 18:26 UTC (permalink / raw)
  To: Paul Waldo; +Cc: Neil Brown, linux-raid

On 7/23/06, Paul Waldo <pwaldo@waldoware.com> wrote:
> Here is the dmesg output.  No log files are created with the FC5 rescue disk.
> Thanks!
I ran into this as well, I believe at this point you want to set:

md-mod.start_dirty_degraded=1

as part of your boot options.  Understand you may see some filesystem
corruption as noted in the documentation.

See:
http://www.linux-m32r.org/lxr/http/source/Documentation/md.txt?v=2.6.17#L54

Regards,

Dan

^ permalink raw reply	[flat|nested] 22+ messages in thread

* Re: In Trouble--Please Help! (was Re: Can't add disk to failed raid array)
  2006-07-24 18:26                 ` Dan Williams
@ 2006-07-24 18:35                   ` Paul Waldo
  2006-07-24 19:06                     ` Dan Williams
  2006-07-25 14:27                   ` Paul Waldo
  1 sibling, 1 reply; 22+ messages in thread
From: Paul Waldo @ 2006-07-24 18:35 UTC (permalink / raw)
  To: Dan Williams; +Cc: Neil Brown, linux-raid

Dan Williams wrote:
> On 7/23/06, Paul Waldo <pwaldo@waldoware.com> wrote:
>> Here is the dmesg output.  No log files are created with the FC5 
>> rescue disk.
>> Thanks!
> I ran into this as well, I believe at this point you want to set:
> 
> md-mod.start_dirty_degraded=1
> 
> as part of your boot options.  Understand you may see some filesystem
> corruption as noted in the documentation.
> 
> See:
> http://www.linux-m32r.org/lxr/http/source/Documentation/md.txt?v=2.6.17#L54
> 
> Regards,
> 
> Dan

I'll certainly give that a try later on, as I need physical access to 
the box.

The corruption part is worrisome...  When you did this, did you 
experience corruption?  I'm running RAID6 with 7 disks; presumably even 
with two disks out of whack, I should be in good shape...???

Paul

^ permalink raw reply	[flat|nested] 22+ messages in thread

* Re: In Trouble--Please Help! (was Re: Can't add disk to failed raid array)
  2006-07-24 18:35                   ` Paul Waldo
@ 2006-07-24 19:06                     ` Dan Williams
  0 siblings, 0 replies; 22+ messages in thread
From: Dan Williams @ 2006-07-24 19:06 UTC (permalink / raw)
  To: Paul Waldo; +Cc: Neil Brown, linux-raid

> I'll certainly give that a try later on, as I need physical access to
> the box.
>
> The corruption part is worrisome...  When you did this, did you
> experience corruption?  I'm running RAID6 with 7 disks; presumably even
> with two disks out of whack, I should be in good shape...???
>
I was running a 5 disk RAID-5 and did not detect any corruption.  Neil
correct me if I am wrong, but I believe that since your failure
occured without power loss that the chances for data corruption in
this case are small.

Dan

^ permalink raw reply	[flat|nested] 22+ messages in thread

* Re: In Trouble--Please Help!  (was Re: Can't add disk to failed raid array)
  2006-07-23 12:32               ` Paul Waldo
  2006-07-24 18:26                 ` Dan Williams
@ 2006-07-24 23:24                 ` Neil Brown
  2006-08-02 19:15                   ` Converting Ext3 to Ext3 under RAID 1 Dan Graham
  1 sibling, 1 reply; 22+ messages in thread
From: Neil Brown @ 2006-07-24 23:24 UTC (permalink / raw)
  To: Paul Waldo; +Cc: linux-raid

On Sunday July 23, pwaldo@waldoware.com wrote:
> Here is the dmesg output.  No log files are created with the FC5 rescue disk.  
> Thanks!
> 
> 
> 
> On Sunday 23 July 2006 07:59, Neil Brown wrote:
> > On Sunday July 23, pwaldo@waldoware.com wrote:
> > > At this point, I'd just be happy to be able to get the degraded array
> > > back up and running.  Is there any way to do that?  Thanks!
> >
> >  mdadm --assemble --force /dev/md1 /dev/hd[aceg]2 /dev/sd[ab]2
> >
....
> raid6: cannot start dirty degraded array for md1

Just checked, and this is fixed in the latest mdadm.  If you use 2.5.2
this --assemble --force will work.

Alternately booting with
  md-mod.start_dirty_degraded=1

or running
  echo 1 >  /sys/module/md_mod/parameters/start_dirty_degraded 

before the --assemble

should work.

NeilBrown

^ permalink raw reply	[flat|nested] 22+ messages in thread

* Re: In Trouble--Please Help! (was Re: Can't add disk to failed raid array)
  2006-07-24 18:26                 ` Dan Williams
  2006-07-24 18:35                   ` Paul Waldo
@ 2006-07-25 14:27                   ` Paul Waldo
  1 sibling, 0 replies; 22+ messages in thread
From: Paul Waldo @ 2006-07-25 14:27 UTC (permalink / raw)
  To: Dan Williams; +Cc: Neil Brown, linux-raid

Dan Williams wrote:
> On 7/23/06, Paul Waldo <pwaldo@waldoware.com> wrote:
>> Here is the dmesg output.  No log files are created with the FC5 
>> rescue disk.
>> Thanks!
> I ran into this as well, I believe at this point you want to set:
> 
> md-mod.start_dirty_degraded=1
> 
> as part of your boot options.  Understand you may see some filesystem
> corruption as noted in the documentation.
> 
> See:
> http://www.linux-m32r.org/lxr/http/source/Documentation/md.txt?v=2.6.17#L54
> 
> Regards,
> 
> Dan

Woo Hoo!  I am back in the running!
The md-mod.start_dirty_degraded enabled me to get the array running and 
I can now boot the machine.

Per Neil's comments, I adjusted the partitioning on the new drive 
(/dev/hdd2) to exactly match that of the other partitions in the array. 
  Success!  The array is now rebuilding with the new disk!!

Thanks for everyone's help on this--I'd be dead without it.  In return, 
maybe I can impart a lesson learned.  My big problem was that the new 
replacement disk, even though the same model as the original it was 
replacing, did not have the same geometry as the original.  It was short 
by 2 cylinders, which prevented it from being added to the array.  Next 
time I create an array of "identical" disks, I am going to keep a few 
cylinders on each one unused for just this type of problem.

Again, thanks for all the help!

Paul

^ permalink raw reply	[flat|nested] 22+ messages in thread

* Converting Ext3 to Ext3 under RAID 1
  2006-07-24 23:24                 ` Neil Brown
@ 2006-08-02 19:15                   ` Dan Graham
  2006-08-02 19:28                     ` dean gaudet
                                       ` (2 more replies)
  0 siblings, 3 replies; 22+ messages in thread
From: Dan Graham @ 2006-08-02 19:15 UTC (permalink / raw)
  To: linux-raid

Hello;
   I have an existing, active ext3 filesystem which I would like to convert to
a RAID 1 ext3 filesystem with minimal down time.  After casting about the web 
and experimenting some on a test system, I believe that I can accomplish this 
in the following manner.

   - Dismount the filesystem.
   - Shrink the filesystem to leave room for the RAID superblock at the end
     while leaving the partition size untouched (shrinking by 16 blocks seems to
     work )
   - Create a degraded array with only the partition carrying the shrunk ext3
     system
   - start the array and mount the array.
   - hot add the mirroring partitions.

The questions I have for those who know Linux-Raid better than I.

    Is this scheme even half-way sane?
    Is 16 blocks a large enough area?


Thanks in advance for any and all feed-back.

-- 
Daniel Graham
graham@molbio.uoregon.edu

541-346-5079 (voice)
541-346-4854 (FAX)

Institute of Molecular Biology
1229 University of Oregon
Eugene, OR 97403-1229


^ permalink raw reply	[flat|nested] 22+ messages in thread

* Re: Converting Ext3 to Ext3 under RAID 1
  2006-08-02 19:15                   ` Converting Ext3 to Ext3 under RAID 1 Dan Graham
@ 2006-08-02 19:28                     ` dean gaudet
  2006-08-02 19:31                     ` Paul Clements
  2006-08-02 22:38                     ` Robert Heinzmann
  2 siblings, 0 replies; 22+ messages in thread
From: dean gaudet @ 2006-08-02 19:28 UTC (permalink / raw)
  To: Dan Graham; +Cc: linux-raid

On Wed, 2 Aug 2006, Dan Graham wrote:

> Hello;
>   I have an existing, active ext3 filesystem which I would like to convert to
> a RAID 1 ext3 filesystem with minimal down time.  After casting about the web
> and experimenting some on a test system, I believe that I can accomplish this
> in the following manner.
> 
>   - Dismount the filesystem.
>   - Shrink the filesystem to leave room for the RAID superblock at the end
>     while leaving the partition size untouched (shrinking by 16 blocks seems
> to
>     work )
>   - Create a degraded array with only the partition carrying the shrunk ext3
>     system
>   - start the array and mount the array.
>   - hot add the mirroring partitions.
> 
> The questions I have for those who know Linux-Raid better than I.
> 
>    Is this scheme even half-way sane?

yes

>    Is 16 blocks a large enough area?

i always err on the side of caution and take a few meg off then resize it 
back up to full size after creating the degraded raid1.  (hmm maybe mdadm 
has some way to tell you how large the resulting partitions would be... 
i've never looked.)

you pretty much have to do this all using a recovery or live CD...

don't forget to rebuild your initrds... all of them including older 
kernels... otherwise there could be one of them still mounting the 
filesystem without using the md device name (and destroying integrity).

don't forget to set the second disk boot partition active and install grub 
so that you can boot from it when the first fails... (after you've 
mirrored the boot or root partition).

-dean

^ permalink raw reply	[flat|nested] 22+ messages in thread

* Re: Converting Ext3 to Ext3 under RAID 1
  2006-08-02 19:15                   ` Converting Ext3 to Ext3 under RAID 1 Dan Graham
  2006-08-02 19:28                     ` dean gaudet
@ 2006-08-02 19:31                     ` Paul Clements
  2006-08-03  7:50                       ` Michael Tokarev
  2006-08-02 22:38                     ` Robert Heinzmann
  2 siblings, 1 reply; 22+ messages in thread
From: Paul Clements @ 2006-08-02 19:31 UTC (permalink / raw)
  To: Dan Graham; +Cc: linux-raid

Dan Graham wrote:

>    Is this scheme even half-way sane?

Yes. It sounds correct.

>    Is 16 blocks a large enough area?

Maybe. The superblock will be between 64KB and 128KB from the end of the 
partition. This depends on the size of the partition:

SB_LOC = PART_SIZE - 64K - (PART_SIZE & (64K-1))

So, by 16 blocks, I assume you mean 16 filesystem blocks (which are 
generally 4KB for ext3). So as long as your partition ends exactly on a 
64KB boundary, you should be OK.

Personally, I would err on the safe side and just shorten the filesystem 
by 128KB. It's not like you're going to miss the extra 64KB.

--
Paul

^ permalink raw reply	[flat|nested] 22+ messages in thread

* Re: Converting Ext3 to Ext3 under RAID 1
  2006-08-02 19:15                   ` Converting Ext3 to Ext3 under RAID 1 Dan Graham
  2006-08-02 19:28                     ` dean gaudet
  2006-08-02 19:31                     ` Paul Clements
@ 2006-08-02 22:38                     ` Robert Heinzmann
  2 siblings, 0 replies; 22+ messages in thread
From: Robert Heinzmann @ 2006-08-02 22:38 UTC (permalink / raw)
  To: Dan Graham; +Cc: linux-raid

Hi Dan,

see thread http://www.spinics.net/lists/raid/msg07742.html.

Regards,
Robert

Dan Graham schrieb:
> Hello;
>   I have an existing, active ext3 filesystem which I would like to 
> convert to
> a RAID 1 ext3 filesystem with minimal down time.  After casting about 
> the web and experimenting some on a test system, I believe that I can 
> accomplish this in the following manner.
>
>   - Dismount the filesystem.
>   - Shrink the filesystem to leave room for the RAID superblock at the 
> end
>     while leaving the partition size untouched (shrinking by 16 blocks 
> seems to
>     work )
>   - Create a degraded array with only the partition carrying the 
> shrunk ext3
>     system
>   - start the array and mount the array.
>   - hot add the mirroring partitions.
>
> The questions I have for those who know Linux-Raid better than I.
>
>    Is this scheme even half-way sane?
>    Is 16 blocks a large enough area?
>
>
> Thanks in advance for any and all feed-back.
>

^ permalink raw reply	[flat|nested] 22+ messages in thread

* Re: Converting Ext3 to Ext3 under RAID 1
  2006-08-02 19:31                     ` Paul Clements
@ 2006-08-03  7:50                       ` Michael Tokarev
  0 siblings, 0 replies; 22+ messages in thread
From: Michael Tokarev @ 2006-08-03  7:50 UTC (permalink / raw)
  To: Paul Clements; +Cc: Dan Graham, linux-raid

Paul Clements wrote:
>>    Is 16 blocks a large enough area?
> 
> Maybe. The superblock will be between 64KB and 128KB from the end of the
> partition. This depends on the size of the partition:
> 
> SB_LOC = PART_SIZE - 64K - (PART_SIZE & (64K-1))
> 
> So, by 16 blocks, I assume you mean 16 filesystem blocks (which are
> generally 4KB for ext3). So as long as your partition ends exactly on a
> 64KB boundary, you should be OK.
> 
> Personally, I would err on the safe side and just shorten the filesystem
> by 128KB. It's not like you're going to miss the extra 64KB.

Or, better yet, shrink it by 1Mb or even 10Mb, whatever, convert
to raid, and - the point - resize it to the max size of the raid
device (ie, don't give "size" argument to resize2fs).  This way,
you will be both safe and will use 100% of the available size.

/mjt

^ permalink raw reply	[flat|nested] 22+ messages in thread

end of thread, other threads:[~2006-08-03  7:50 UTC | newest]

Thread overview: 22+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2006-07-16  0:56 Can't add disk to failed raid array Paul Waldo
2006-07-16 10:19 ` Neil Brown
2006-07-16 12:24   ` Paul Waldo
2006-07-18  5:36     ` Neil Brown
2006-07-23 11:04       ` In Trouble--Please Help! (was Re: Can't add disk to failed raid array) Paul Waldo
2006-07-23 11:25         ` Neil Brown
2006-07-23 11:50           ` Paul Waldo
2006-07-23 11:53           ` Paul Waldo
2006-07-23 11:59             ` Neil Brown
2006-07-23 12:32               ` Paul Waldo
2006-07-24 18:26                 ` Dan Williams
2006-07-24 18:35                   ` Paul Waldo
2006-07-24 19:06                     ` Dan Williams
2006-07-25 14:27                   ` Paul Waldo
2006-07-24 23:24                 ` Neil Brown
2006-08-02 19:15                   ` Converting Ext3 to Ext3 under RAID 1 Dan Graham
2006-08-02 19:28                     ` dean gaudet
2006-08-02 19:31                     ` Paul Clements
2006-08-03  7:50                       ` Michael Tokarev
2006-08-02 22:38                     ` Robert Heinzmann
     [not found] ` <200607160913.32005.pwaldo@waldoware.com>
     [not found]   ` <Pine.LNX.4.62.0607161524170.7520@uplift.swm.pp.se>
2006-07-16 14:39     ` Can't add disk to failed raid array Paul Waldo
2006-07-17 18:36 ` Paul Waldo

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).