linux-raid.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
* RAID5 in sync does not populate slots sequentially, shows array as (somewhat) faulty
@ 2008-03-03 10:42 Peter Rabbitson
  2008-03-03 11:44 ` Tor Arne Vestbø
  2008-03-03 18:34 ` Bill Davidsen
  0 siblings, 2 replies; 8+ messages in thread
From: Peter Rabbitson @ 2008-03-03 10:42 UTC (permalink / raw)
  To: linux-raid

Hello,

Noticing the problems Tor Vestbø is having, I remembered that I have an array 
in a similar state, which I never figured out. The array has been working 
flawlessly for 3 months, the monthly 'check' runs come back with everything 
being clean. However this is how the array looks through mdadm's eyes:

root@Thesaurus:~# mdadm -D /dev/md5
/dev/md5:
         Version : 01.01.03
   Creation Time : Tue Jan 22 03:52:42 2008
      Raid Level : raid5
      Array Size : 865081344 (825.01 GiB 885.84 GB)
   Used Dev Size : 576720896 (275.00 GiB 295.28 GB)
    Raid Devices : 4
   Total Devices : 4
Preferred Minor : 5
     Persistence : Superblock is persistent

     Update Time : Mon Mar  3 11:34:36 2008
           State : clean
  Active Devices : 4
Working Devices : 4
  Failed Devices : 0
   Spare Devices : 0

          Layout : left-symmetric
      Chunk Size : 2048K

            Name : Thesaurus:Crypta  (local to host Thesaurus)
            UUID : 1decb2d1:ebf16128:a240938a:669b0999
          Events : 9090

     Number   Major   Minor   RaidDevice State
        4       8        3        0      active sync   /dev/sda3	<-- why at 
#4, not #0?
        1       8       19        1      active sync   /dev/sdb3
        2       8       35        2      active sync   /dev/sdc3
        3       8       51        3      active sync   /dev/sdd3
root@Thesaurus:~#

root@Thesaurus:~# mdadm -E /dev/sda3
/dev/sda3:
           Magic : a92b4efc
         Version : 1.1
     Feature Map : 0x0
      Array UUID : 1decb2d1:ebf16128:a240938a:669b0999
            Name : Thesaurus:Crypta  (local to host Thesaurus)
   Creation Time : Tue Jan 22 03:52:42 2008
      Raid Level : raid5
    Raid Devices : 4

  Avail Dev Size : 576733236 (275.01 GiB 295.29 GB)
      Array Size : 1730162688 (825.01 GiB 885.84 GB)
   Used Dev Size : 576720896 (275.00 GiB 295.28 GB)
     Data Offset : 264 sectors		<-- not sure why this is different from the rest
    Super Offset : 0 sectors
           State : active
     Device UUID : 72d9ceb4:89a20f49:50d51384:9b164f72

     Update Time : Mon Mar  3 11:34:26 2008
        Checksum : 4bd3ae95 - correct
          Events : 9090

          Layout : left-symmetric
      Chunk Size : 2048K

     Array Slot : 4 (failed, 1, 2, 3, 0)
    Array State : Uuuu 1 failed			<-- why is that?
root@Thesaurus:~#

root@Thesaurus:~# mdadm -E /dev/sdb3
/dev/sdb3:
           Magic : a92b4efc
         Version : 1.1
     Feature Map : 0x0
      Array UUID : 1decb2d1:ebf16128:a240938a:669b0999
            Name : Thesaurus:Crypta  (local to host Thesaurus)
   Creation Time : Tue Jan 22 03:52:42 2008
      Raid Level : raid5
    Raid Devices : 4

  Avail Dev Size : 576733364 (275.01 GiB 295.29 GB)
      Array Size : 1730162688 (825.01 GiB 885.84 GB)
   Used Dev Size : 576720896 (275.00 GiB 295.28 GB)
     Data Offset : 136 sectors
    Super Offset : 0 sectors
           State : active
     Device UUID : 0ba427b2:6cc26a34:cc9838d9:1dc77772

     Update Time : Mon Mar  3 11:39:01 2008
        Checksum : 89d50e1f - correct
          Events : 9090

          Layout : left-symmetric
      Chunk Size : 2048K

     Array Slot : 1 (failed, 1, 2, 3, 0)
    Array State : uUuu 1 failed			<-- and that?
root@Thesaurus:~#

root@Thesaurus:~# mdadm -E /dev/sdc3
/dev/sdc3:
           Magic : a92b4efc
         Version : 1.1
     Feature Map : 0x0
      Array UUID : 1decb2d1:ebf16128:a240938a:669b0999
            Name : Thesaurus:Crypta  (local to host Thesaurus)
   Creation Time : Tue Jan 22 03:52:42 2008
      Raid Level : raid5
    Raid Devices : 4

  Avail Dev Size : 576733364 (275.01 GiB 295.29 GB)
      Array Size : 1730162688 (825.01 GiB 885.84 GB)
   Used Dev Size : 576720896 (275.00 GiB 295.28 GB)
     Data Offset : 136 sectors
    Super Offset : 0 sectors
           State : active
     Device UUID : ada93838:5e3fef07:fd478ba7:8fcf75e6

     Update Time : Mon Mar  3 11:39:44 2008
        Checksum : 25bb4882 - correct
          Events : 9090

          Layout : left-symmetric
      Chunk Size : 2048K

     Array Slot : 2 (failed, 1, 2, 3, 0)
    Array State : uuUu 1 failed			<-- same here
root@Thesaurus:~#

root@Thesaurus:~# mdadm -E /dev/sdd3
/dev/sdd3:
           Magic : a92b4efc
         Version : 1.1
     Feature Map : 0x0
      Array UUID : 1decb2d1:ebf16128:a240938a:669b0999
            Name : Thesaurus:Crypta  (local to host Thesaurus)
   Creation Time : Tue Jan 22 03:52:42 2008
      Raid Level : raid5
    Raid Devices : 4

  Avail Dev Size : 576733364 (275.01 GiB 295.29 GB)
      Array Size : 1730162688 (825.01 GiB 885.84 GB)
   Used Dev Size : 576720896 (275.00 GiB 295.28 GB)
     Data Offset : 136 sectors
    Super Offset : 0 sectors
           State : active
     Device UUID : be5f202f:c9097996:98d66036:9b6453ee

     Update Time : Mon Mar  3 11:40:11 2008
        Checksum : 41dfecc1 - correct
          Events : 9090

          Layout : left-symmetric
      Chunk Size : 2048K

     Array Slot : 3 (failed, 1, 2, 3, 0)
    Array State : uuuU 1 failed
root@Thesaurus:~# 				<-- and here



--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 8+ messages in thread

* Re: RAID5 in sync does not populate slots sequentially, shows array as (somewhat) faulty
  2008-03-03 10:42 RAID5 in sync does not populate slots sequentially, shows array as (somewhat) faulty Peter Rabbitson
@ 2008-03-03 11:44 ` Tor Arne Vestbø
  2008-03-03 18:34 ` Bill Davidsen
  1 sibling, 0 replies; 8+ messages in thread
From: Tor Arne Vestbø @ 2008-03-03 11:44 UTC (permalink / raw)
  To: linux-raid

Peter Rabbitson wrote:
> Hello,
> 
> Noticing the problems Tor Vestbø is having, I remembered that I have an 
> array in a similar state, which I never figured out. 

[snip]

>     Array Slot : 4 (failed, 1, 2, 3, 0)
>    Array State : Uuuu 1 failed            <-- why is that?
> root@Thesaurus:~#

I've seen this in my array too. Was expecting Uuu_, or Uuuu after 
rebuild, but it also said "1 failed".

Tor Arne
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 8+ messages in thread

* Re: RAID5 in sync does not populate slots sequentially, shows array as (somewhat) faulty
  2008-03-03 10:42 RAID5 in sync does not populate slots sequentially, shows array as (somewhat) faulty Peter Rabbitson
  2008-03-03 11:44 ` Tor Arne Vestbø
@ 2008-03-03 18:34 ` Bill Davidsen
  2008-03-04 10:25   ` Peter Rabbitson
  1 sibling, 1 reply; 8+ messages in thread
From: Bill Davidsen @ 2008-03-03 18:34 UTC (permalink / raw)
  To: Peter Rabbitson; +Cc: linux-raid

Peter Rabbitson wrote:
> Hello,
>
> Noticing the problems Tor Vestbø is having, I remembered that I have 
> an array in a similar state, which I never figured out. The array has 
> been working flawlessly for 3 months, the monthly 'check' runs come 
> back with everything being clean. However this is how the array looks 
> through mdadm's eyes:

I'm in agreement that something is odd about the disk numbers here, and 
I'm suspicious because I have never seen this with 0.90 superblocks. 
That doesn't mean it couldn't happen and I never noticed, it's certainly 
odd that four drives wouldn't be numbered 0..3, in raid5 they are all 
equally out of sync.

-- 
Bill Davidsen <davidsen@tmr.com>
  "Woe unto the statesman who makes war without a reason that will still
  be valid when the war is over..." Otto von Bismark 



--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 8+ messages in thread

* Re: RAID5 in sync does not populate slots sequentially, shows array as (somewhat) faulty
  2008-03-03 18:34 ` Bill Davidsen
@ 2008-03-04 10:25   ` Peter Rabbitson
  2008-03-04 10:42     ` Tor Arne Vestbø
  0 siblings, 1 reply; 8+ messages in thread
From: Peter Rabbitson @ 2008-03-04 10:25 UTC (permalink / raw)
  To: Bill Davidsen; +Cc: linux-raid

Bill Davidsen wrote:
> Peter Rabbitson wrote:
>> Hello,
>>
>> Noticing the problems Tor Vestbø is having, I remembered that I have 
>> an array in a similar state, which I never figured out. The array has 
>> been working flawlessly for 3 months, the monthly 'check' runs come 
>> back with everything being clean. However this is how the array looks 
>> through mdadm's eyes:
> 
> I'm in agreement that something is odd about the disk numbers here, and 
> I'm suspicious because I have never seen this with 0.90 superblocks. 
> That doesn't mean it couldn't happen and I never noticed, it's certainly 
> odd that four drives wouldn't be numbered 0..3, in raid5 they are all 
> equally out of sync.
> 

After Tor Arne reported his success I figured I will simply fail/remove sda3, 
scrape it clean, and will add it back. I zeroed superblocks beforehand and 
also wrote zeros (dd  if=/dev/zero) to the drives start and end just to make 
sure everythign is off. After resync I am back at square one - the offset of 
sda3 is different than everything else and the array has one failed drive. If 
someone can shed some light I made snapshots of the superblocks[1] alongside 
with the current output of mdadm at http://rabbit.us/pool/md5_problem.tar.bz2.

[1] dd if=/dev/sdX3 of=sdX_sb count=<Data Offset> bs=512

Here is my system config:

root@Thesaurus:/arx/space/pool# fdisk -l /dev/sd[abcd]

Disk /dev/sda: 400.0 GB, 400088457216 bytes
255 heads, 63 sectors/track, 48641 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Disk identifier: 0x00000000

    Device Boot      Start         End      Blocks   Id  System
/dev/sda1               1           7       56196   fd  Linux raid autodetect
/dev/sda2               8         507     4016250   fd  Linux raid autodetect
/dev/sda3             508       36407   288366750   83  Linux
/dev/sda4           36408       48641    98269605   83  Linux

Disk /dev/sdb: 320.0 GB, 320072933376 bytes
255 heads, 63 sectors/track, 38913 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Disk identifier: 0x00000000

    Device Boot      Start         End      Blocks   Id  System
/dev/sdb1               1           7       56196   fd  Linux raid autodetect
/dev/sdb2               8         507     4016250   fd  Linux raid autodetect
/dev/sdb3             508       36407   288366750   83  Linux
/dev/sdb4           36408       38913    20129445   83  Linux

Disk /dev/sdc: 300.0 GB, 300090728448 bytes
255 heads, 63 sectors/track, 36483 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Disk identifier: 0x00000000

    Device Boot      Start         End      Blocks   Id  System
/dev/sdc1               1           7       56196   fd  Linux raid autodetect
/dev/sdc2               8         507     4016250   fd  Linux raid autodetect
/dev/sdc3             508       36407   288366750   83  Linux
/dev/sdc4           36408       36483      610470   83  Linux

Disk /dev/sdd: 300.0 GB, 300090728448 bytes
255 heads, 63 sectors/track, 36483 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Disk identifier: 0x00000000

    Device Boot      Start         End      Blocks   Id  System
/dev/sdd1               1           7       56196   fd  Linux raid autodetect
/dev/sdd2               8         507     4016250   fd  Linux raid autodetect
/dev/sdd3             508       36407   288366750   83  Linux
/dev/sdd4           36408       36483      610470   83  Linux
root@Thesaurus:/arx/space/pool#

root@Thesaurus:~# cat /proc/mdstat
Personalities : [raid0] [raid1] [raid10] [raid6] [raid5] [raid4]
md5 : active raid5 sda3[4] sdd3[3] sdc3[2] sdb3[1]
       865081344 blocks super 1.1 level 5, 2048k chunk, algorithm 2 [4/4] [UUUU]

md1 : active raid1 sdd1[3] sdc1[2] sdb1[1] sda1[0]
       56128 blocks [4/4] [UUUU]

md10 : active raid10 sdd2[3] sdc2[2] sdb2[1] sda2[0]
       5353472 blocks 1024K chunks 3 far-copies [4/4] [UUUU]

unused devices: <none>
root@Thesaurus:~#



--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 8+ messages in thread

* Re: RAID5 in sync does not populate slots sequentially, shows array as (somewhat) faulty
  2008-03-04 10:25   ` Peter Rabbitson
@ 2008-03-04 10:42     ` Tor Arne Vestbø
  2008-03-04 10:52       ` Peter Rabbitson
  2008-03-06 14:51       ` Rui Santos
  0 siblings, 2 replies; 8+ messages in thread
From: Tor Arne Vestbø @ 2008-03-04 10:42 UTC (permalink / raw)
  To: linux-raid

Peter Rabbitson wrote:
> After Tor Arne reported his success I figured I will simply fail/remove 
> sda3, scrape it clean, and will add it back. I zeroed superblocks 
> beforehand and also wrote zeros (dd  if=/dev/zero) to the drives start 
> and end just to make sure everythign is off. After resync I am back at 
> square one - the offset of sda3 is different than everything else and 
> the array has one failed drive. If someone can shed some light I made 
> snapshots of the superblocks[1] alongside with the current output of 
> mdadm at http://rabbit.us/pool/md5_problem.tar.bz2.

Not sure if this is at all related to your problem, but one of the 
things I tried was to shred all the old drives in the system that were 
not going to be part of the array.

/dev/sda system (250GB) <-- shred
/dev/sdb home (250GB) <-- shred

/dev/sdc raid (750GB)
/dev/sdd raid (750GB)
/dev/sde raid (750GB)
/dev/sdf raid (750GB)

The reason I did this was because /dev/sda and /dev/sdb used to be part 
of a RAID1 array, but were now used as system disk and home disk 
respectively. I was afraid that mdadm would pick up on some of the 
lingering RAID superblocks on those disks when reporting, so I shredded 
them both using 'shred -n 1' and reinstalled.

Don't know if that affected anything at all for me, since the actual 
problem was that I didn't wait for a full resync, but now you know :)

Tor Arne









> 
> [1] dd if=/dev/sdX3 of=sdX_sb count=<Data Offset> bs=512
> 
> Here is my system config:
> 
> root@Thesaurus:/arx/space/pool# fdisk -l /dev/sd[abcd]
> 
> Disk /dev/sda: 400.0 GB, 400088457216 bytes
> 255 heads, 63 sectors/track, 48641 cylinders
> Units = cylinders of 16065 * 512 = 8225280 bytes
> Disk identifier: 0x00000000
> 
>    Device Boot      Start         End      Blocks   Id  System
> /dev/sda1               1           7       56196   fd  Linux raid 
> autodetect
> /dev/sda2               8         507     4016250   fd  Linux raid 
> autodetect
> /dev/sda3             508       36407   288366750   83  Linux
> /dev/sda4           36408       48641    98269605   83  Linux
> 
> Disk /dev/sdb: 320.0 GB, 320072933376 bytes
> 255 heads, 63 sectors/track, 38913 cylinders
> Units = cylinders of 16065 * 512 = 8225280 bytes
> Disk identifier: 0x00000000
> 
>    Device Boot      Start         End      Blocks   Id  System
> /dev/sdb1               1           7       56196   fd  Linux raid 
> autodetect
> /dev/sdb2               8         507     4016250   fd  Linux raid 
> autodetect
> /dev/sdb3             508       36407   288366750   83  Linux
> /dev/sdb4           36408       38913    20129445   83  Linux
> 
> Disk /dev/sdc: 300.0 GB, 300090728448 bytes
> 255 heads, 63 sectors/track, 36483 cylinders
> Units = cylinders of 16065 * 512 = 8225280 bytes
> Disk identifier: 0x00000000
> 
>    Device Boot      Start         End      Blocks   Id  System
> /dev/sdc1               1           7       56196   fd  Linux raid 
> autodetect
> /dev/sdc2               8         507     4016250   fd  Linux raid 
> autodetect
> /dev/sdc3             508       36407   288366750   83  Linux
> /dev/sdc4           36408       36483      610470   83  Linux
> 
> Disk /dev/sdd: 300.0 GB, 300090728448 bytes
> 255 heads, 63 sectors/track, 36483 cylinders
> Units = cylinders of 16065 * 512 = 8225280 bytes
> Disk identifier: 0x00000000
> 
>    Device Boot      Start         End      Blocks   Id  System
> /dev/sdd1               1           7       56196   fd  Linux raid 
> autodetect
> /dev/sdd2               8         507     4016250   fd  Linux raid 
> autodetect
> /dev/sdd3             508       36407   288366750   83  Linux
> /dev/sdd4           36408       36483      610470   83  Linux
> root@Thesaurus:/arx/space/pool#
> 
> root@Thesaurus:~# cat /proc/mdstat
> Personalities : [raid0] [raid1] [raid10] [raid6] [raid5] [raid4]
> md5 : active raid5 sda3[4] sdd3[3] sdc3[2] sdb3[1]
>       865081344 blocks super 1.1 level 5, 2048k chunk, algorithm 2 [4/4] 
> [UUUU]
> 
> md1 : active raid1 sdd1[3] sdc1[2] sdb1[1] sda1[0]
>       56128 blocks [4/4] [UUUU]
> 
> md10 : active raid10 sdd2[3] sdc2[2] sdb2[1] sda2[0]
>       5353472 blocks 1024K chunks 3 far-copies [4/4] [UUUU]
> 
> unused devices: <none>
> root@Thesaurus:~#
> 
> 
> 
> -- 
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html


^ permalink raw reply	[flat|nested] 8+ messages in thread

* Re: RAID5 in sync does not populate slots sequentially, shows array as (somewhat) faulty
  2008-03-04 10:42     ` Tor Arne Vestbø
@ 2008-03-04 10:52       ` Peter Rabbitson
  2008-03-04 10:58         ` Tor Arne Vestbø
  2008-03-06 14:51       ` Rui Santos
  1 sibling, 1 reply; 8+ messages in thread
From: Peter Rabbitson @ 2008-03-04 10:52 UTC (permalink / raw)
  To: Tor Arne Vestbø; +Cc: linux-raid

Tor Arne Vestbø wrote:
> The reason I did this was because /dev/sda and /dev/sdb used to be part 
> of a RAID1 array, but were now used as system disk and home disk 
> respectively. I was afraid that mdadm would pick up on some of the 
> lingering RAID superblocks on those disks when reporting, so I shredded 
> them both using 'shred -n 1' and reinstalled.
> 

This is irrelevant to 1.x superblocks, and largely insignificant for 0.9 
superblocks (baring some really bizzare cases). In either case mdadm 
--zero-superblock /dev/XX (possibly executed multiple times) would save you a 
lot of disk churning :)
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 8+ messages in thread

* Re: RAID5 in sync does not populate slots sequentially, shows array as (somewhat) faulty
  2008-03-04 10:52       ` Peter Rabbitson
@ 2008-03-04 10:58         ` Tor Arne Vestbø
  0 siblings, 0 replies; 8+ messages in thread
From: Tor Arne Vestbø @ 2008-03-04 10:58 UTC (permalink / raw)
  To: linux-raid

Peter Rabbitson wrote:
>> Tor Arne Vestbø wrote:
>> [...] so I shredded them both using 'shred -n 1' and reinstalled.
> 
> This is irrelevant to 1.x superblocks, and largely insignificant for 0.9 
> superblocks (baring some really bizzare cases). In either case mdadm 
> --zero-superblock /dev/XX (possibly executed multiple times) would save 
> you a lot of disk churning :)

The only way to learn is the hard way ;)

Tor Arne

--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 8+ messages in thread

* Re: RAID5 in sync does not populate slots sequentially, shows array as (somewhat) faulty
  2008-03-04 10:42     ` Tor Arne Vestbø
  2008-03-04 10:52       ` Peter Rabbitson
@ 2008-03-06 14:51       ` Rui Santos
  1 sibling, 0 replies; 8+ messages in thread
From: Rui Santos @ 2008-03-06 14:51 UTC (permalink / raw)
  To: linux-raid; +Cc: Tor Arne Vestbø

Tor Arne Vestbø wrote:
> Not sure if this is at all related to your problem, but one of the 
> things I tried was to shred all the old drives in the system that were 
> not going to be part of the array.
>
> /dev/sda system (250GB) <-- shred
> /dev/sdb home (250GB) <-- shred
>
> /dev/sdc raid (750GB)
> /dev/sdd raid (750GB)
> /dev/sde raid (750GB)
> /dev/sdf raid (750GB)
>
> The reason I did this was because /dev/sda and /dev/sdb used to be 
> part of a RAID1 array, but were now used as system disk and home disk 
> respectively. I was afraid that mdadm would pick up on some of the 
> lingering RAID superblocks on those disks when reporting, so I 
> shredded them both using 'shred -n 1' and reinstalled.
>
> Don't know if that affected anything at all for me, since the actual 
> problem was that I didn't wait for a full resync, but now you know :)
>
> Tor Arne
Hi all,

I have a identical problem. But it didn't went way with the zeroing 
superblock / shred procedure. I also have no extra discs.
Here is my config:

~# cat /proc/mdstat:
md0 : active raid1 sda1[0] sdc1[2] sdb1[1]
          136508 blocks super 1.0 [3/3] [UUU]
          bitmap: 0/9 pages [0KB], 8KB chunk

md1 : active raid5 sdb2[0] sda2[3] sdc2[1]
          1060096 blocks super 1.0 level 5, 128k chunk, algorithm 2 
[3/3] [UUU]

md2 : active raid5 sda3[0] sdc3[3] sdb3[1]
          780083968 blocks super 1.0 level 5, 128k chunk, algorithm 2 
[3/3] [UUU]


~# mdadm -D /dev/md{0,1,2}
/dev/md0:
        Version : 01.00.03
  Creation Time : Wed Feb 27 15:38:43 2008
     Raid Level : raid1
     Array Size : 136508 (133.33 MiB 139.78 MB)
  Used Dev Size : 136508 (133.33 MiB 139.78 MB)
   Raid Devices : 3
  Total Devices : 3
Preferred Minor : 0
    Persistence : Superblock is persistent

  Intent Bitmap : Internal

    Update Time : Thu Mar  6 14:38:08 2008
          State : active
 Active Devices : 3
Working Devices : 3
 Failed Devices : 0
  Spare Devices : 0

           Name : 0
           UUID : c5e9420d:67e022ae:eaf9fc3e:4949a042
         Events : 20

    Number   Major   Minor   RaidDevice State
       0       8        1        0      active sync   /dev/sda1
       1       8       17        1      active sync   /dev/sdb1
       2       8       33        2      active sync   /dev/sdc1
/dev/md1:
        Version : 01.00.03
  Creation Time : Thu Mar  6 13:14:36 2008
     Raid Level : raid5
     Array Size : 1060096 (1035.42 MiB 1085.54 MB)
  Used Dev Size : 530048 (517.71 MiB 542.77 MB)
   Raid Devices : 3
  Total Devices : 3
Preferred Minor : 1
    Persistence : Superblock is persistent

    Update Time : Thu Mar  6 14:38:08 2008
          State : clean
 Active Devices : 3
Working Devices : 3
 Failed Devices : 0
  Spare Devices : 0

         Layout : left-symmetric
     Chunk Size : 128K

           Name : 1
           UUID : 3e76c555:d423cd7d:b1454b79:f34e6322
         Events : 4

    Number   Major   Minor   RaidDevice State
       0       8       18        0      active sync   /dev/sdb2
       1       8       34        1      active sync   /dev/sdc2
       3       8        2        2      active sync   /dev/sda2
/dev/md2:
        Version : 01.00.03
  Creation Time : Wed Feb 27 15:38:46 2008
     Raid Level : raid5
     Array Size : 780083968 (743.95 GiB 798.81 GB)
  Used Dev Size : 780083968 (371.97 GiB 399.40 GB)
   Raid Devices : 3
  Total Devices : 3
Preferred Minor : 2
    Persistence : Superblock is persistent

    Update Time : Thu Mar  6 14:38:08 2008
          State : clean
 Active Devices : 3
Working Devices : 3
 Failed Devices : 0
  Spare Devices : 0

         Layout : left-symmetric
     Chunk Size : 128K

           Name : 2
           UUID : a55c4f4e:8cba34b7:b5f70bb0:97fd1366
         Events : 5070

    Number   Major   Minor   RaidDevice State
       0       8        3        0      active sync   /dev/sda3
       1       8       19        1      active sync   /dev/sdb3
       3       8       35        2      active sync   /dev/sdc3

unused devices: <none>



~# mdadm -E /dev/sd{a1,b1,c1,a2,b2,c2,a3,b3,c3}
/dev/sda1:
          Magic : a92b4efc
        Version : 01
    Feature Map : 0x1
     Array UUID : c5e9420d:67e022ae:eaf9fc3e:4949a042
           Name : 0
  Creation Time : Wed Feb 27 15:38:43 2008
     Raid Level : raid1
   Raid Devices : 3

  Used Dev Size : 273016 (133.33 MiB 139.78 MB)
     Array Size : 273016 (133.33 MiB 139.78 MB)
   Super Offset : 273024 sectors
          State : clean
    Device UUID : 678a316f:e7a2a641:14c1c3f0:b55d6aae

Internal Bitmap : 2 sectors from superblock
    Update Time : Thu Mar  6 14:38:11 2008
       Checksum : c5d6af8b - correct
         Events : 20


    Array Slot : 0 (0, 1, 2)
   Array State : Uuu
/dev/sdb1:
          Magic : a92b4efc
        Version : 01
    Feature Map : 0x1
     Array UUID : c5e9420d:67e022ae:eaf9fc3e:4949a042
           Name : 0
  Creation Time : Wed Feb 27 15:38:43 2008
     Raid Level : raid1
   Raid Devices : 3

  Used Dev Size : 273016 (133.33 MiB 139.78 MB)
     Array Size : 273016 (133.33 MiB 139.78 MB)
   Super Offset : 273024 sectors
          State : clean
    Device UUID : 17525ff0:fe48f81d:8f28e04c:34901f21

Internal Bitmap : 2 sectors from superblock
    Update Time : Thu Mar  6 14:38:11 2008
       Checksum : f227b74c - correct
         Events : 20


    Array Slot : 1 (0, 1, 2)
   Array State : uUu
/dev/sdc1:
          Magic : a92b4efc
        Version : 01
    Feature Map : 0x1
     Array UUID : c5e9420d:67e022ae:eaf9fc3e:4949a042
           Name : 0
  Creation Time : Wed Feb 27 15:38:43 2008
     Raid Level : raid1
   Raid Devices : 3

  Used Dev Size : 273016 (133.33 MiB 139.78 MB)
     Array Size : 273016 (133.33 MiB 139.78 MB)
   Super Offset : 273024 sectors
          State : clean
    Device UUID : c3e00260:dfa90f02:3c39380b:b090375e

Internal Bitmap : 2 sectors from superblock
    Update Time : Thu Mar  6 14:38:11 2008
       Checksum : 4152b803 - correct
         Events : 20


    Array Slot : 2 (0, 1, 2)
   Array State : uuU
/dev/sda2:
          Magic : a92b4efc
        Version : 01
    Feature Map : 0x0
     Array UUID : 3e76c555:d423cd7d:b1454b79:f34e6322
           Name : 1
  Creation Time : Thu Mar  6 13:14:36 2008
     Raid Level : raid5
   Raid Devices : 3

  Used Dev Size : 1060264 (517.79 MiB 542.86 MB)
     Array Size : 2120192 (1035.42 MiB 1085.54 MB)
      Used Size : 1060096 (517.71 MiB 542.77 MB)
   Super Offset : 1060272 sectors
          State : clean
    Device UUID : 2bd8b81f:6e38a263:9c1a48f5:81c2cbc8

    Update Time : Thu Mar  6 14:38:11 2008
       Checksum : e9afe694 - correct
         Events : 4

         Layout : left-symmetric
     Chunk Size : 128K

    Array Slot : 3 (0, 1, failed, 2)
   Array State : uuU 1 failed
/dev/sdb2:
          Magic : a92b4efc
        Version : 01
    Feature Map : 0x0
     Array UUID : 3e76c555:d423cd7d:b1454b79:f34e6322
           Name : 1
  Creation Time : Thu Mar  6 13:14:36 2008
     Raid Level : raid5
   Raid Devices : 3

  Used Dev Size : 1060264 (517.79 MiB 542.86 MB)
     Array Size : 2120192 (1035.42 MiB 1085.54 MB)
      Used Size : 1060096 (517.71 MiB 542.77 MB)
   Super Offset : 1060272 sectors
          State : clean
    Device UUID : 18c77f52:4bbbf090:31a3724c:b7cafa3c

    Update Time : Thu Mar  6 14:38:11 2008
       Checksum : 151ee926 - correct
         Events : 4

         Layout : left-symmetric
     Chunk Size : 128K

    Array Slot : 0 (0, 1, failed, 2)
   Array State : Uuu 1 failed
/dev/sdc2:
          Magic : a92b4efc
        Version : 01
    Feature Map : 0x0
     Array UUID : 3e76c555:d423cd7d:b1454b79:f34e6322
           Name : 1
  Creation Time : Thu Mar  6 13:14:36 2008
     Raid Level : raid5
   Raid Devices : 3

  Used Dev Size : 1060264 (517.79 MiB 542.86 MB)
     Array Size : 2120192 (1035.42 MiB 1085.54 MB)
      Used Size : 1060096 (517.71 MiB 542.77 MB)
   Super Offset : 1060272 sectors
          State : clean
    Device UUID : 19fd5caf:ff6a3e82:95c84b1d:8ec60429

    Update Time : Thu Mar  6 14:38:11 2008
       Checksum : 202cf017 - correct
         Events : 4

         Layout : left-symmetric
     Chunk Size : 128K

    Array Slot : 1 (0, 1, failed, 2)
   Array State : uUu 1 failed
/dev/sda3:
          Magic : a92b4efc
        Version : 01
    Feature Map : 0x0
     Array UUID : a55c4f4e:8cba34b7:b5f70bb0:97fd1366
           Name : 2
  Creation Time : Wed Feb 27 15:38:46 2008
     Raid Level : raid5
   Raid Devices : 3

  Used Dev Size : 780083992 (371.97 GiB 399.40 GB)
     Array Size : 1560167936 (743.95 GiB 798.81 GB)
      Used Size : 780083968 (371.97 GiB 399.40 GB)
   Super Offset : 780084248 sectors
          State : active
    Device UUID : 590ac9c2:e4ae82b3:1248d87a:d655dd7c

    Update Time : Thu Mar  6 14:39:49 2008
       Checksum : 4de7b99b - correct
         Events : 5071

         Layout : left-symmetric
     Chunk Size : 128K

    Array Slot : 0 (0, 1, failed, 2)
   Array State : Uuu 1 failed
/dev/sdb3:
          Magic : a92b4efc
        Version : 01
    Feature Map : 0x0
     Array UUID : a55c4f4e:8cba34b7:b5f70bb0:97fd1366
           Name : 2
  Creation Time : Wed Feb 27 15:38:46 2008
     Raid Level : raid5
   Raid Devices : 3

  Used Dev Size : 780083992 (371.97 GiB 399.40 GB)
     Array Size : 1560167936 (743.95 GiB 798.81 GB)
      Used Size : 780083968 (371.97 GiB 399.40 GB)
   Super Offset : 780084248 sectors
          State : active
    Device UUID : 9589b278:38932876:414d8879:b9c70fe7

    Update Time : Thu Mar  6 14:39:49 2008
       Checksum : 2f59943e - correct
         Events : 5071

         Layout : left-symmetric
     Chunk Size : 128K

    Array Slot : 1 (0, 1, failed, 2)
   Array State : uUu 1 failed
/dev/sdc3:
          Magic : a92b4efc
        Version : 01
    Feature Map : 0x0
     Array UUID : a55c4f4e:8cba34b7:b5f70bb0:97fd1366
           Name : 2
  Creation Time : Wed Feb 27 15:38:46 2008
     Raid Level : raid5
   Raid Devices : 3

  Used Dev Size : 780083992 (371.97 GiB 399.40 GB)
     Array Size : 1560167936 (743.95 GiB 798.81 GB)
      Used Size : 780083968 (371.97 GiB 399.40 GB)
   Super Offset : 780084248 sectors
          State : active
    Device UUID : a8a110fa:75d91ef2:5e0376a7:dad76b1a

    Update Time : Thu Mar  6 14:39:49 2008
       Checksum : 8df7b8ce - correct
         Events : 5071

         Layout : left-symmetric
     Chunk Size : 128K

    Array Slot : 3 (0, 1, failed, 2)
   Array State : uuU 1 failed


As you can see, all of RAID5 devices state that it is failed. On the 
other hand, the RAID device reports the array as clean. You can also see 
the discrepancies in the array slots composition.
Anyway, this only happens when I use 1.0 or 1.1 superblock ( didn't try 
1.2 ). If I use the 0.9 superblocks all problems are solved. I'm able to 
test with one of the arrays, if some test is needed or advised.
I'm using a 2.6.24 x86_64 kernel ( SuSE ) and mdadm v2.6.2 ( also tried 
v2.6.4 with no success ). Also as stated, on this setup, exist only 
three disk with three partitions ( 0xFD ) each to build three raid arrays.

Thanks for your help,
Rui Santos

--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 8+ messages in thread

end of thread, other threads:[~2008-03-06 14:51 UTC | newest]

Thread overview: 8+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2008-03-03 10:42 RAID5 in sync does not populate slots sequentially, shows array as (somewhat) faulty Peter Rabbitson
2008-03-03 11:44 ` Tor Arne Vestbø
2008-03-03 18:34 ` Bill Davidsen
2008-03-04 10:25   ` Peter Rabbitson
2008-03-04 10:42     ` Tor Arne Vestbø
2008-03-04 10:52       ` Peter Rabbitson
2008-03-04 10:58         ` Tor Arne Vestbø
2008-03-06 14:51       ` Rui Santos

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).