linux-raid.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
* strange behavior of a raid5 array after system crash
@ 2005-03-04 19:42 hpg
  2005-03-04 20:10 ` hostRAID, AIC-7901, disable? Stefan Eckert
  2005-03-04 20:57 ` strange behavior of a raid5 array after system crash Mike Tran
  0 siblings, 2 replies; 3+ messages in thread
From: hpg @ 2005-03-04 19:42 UTC (permalink / raw)
  To: linux-raid

Hello everyone,

I need your help with a strange behavior of a raid5 array.

My Linux fileserver was frozen for unknown reason. No mouse movement, 
no console, no disk activity nothing.
So I had to hit the reset button.

At boot time 5 raid5 arrays have been active without any faults.
Two other raid5 arrays resynchronized successfully.
Only one had some trouble to recover.

Because I am using LVM2 on top of all my raid5 arrays and have the root 
filesystem in that volume group which is using the raid5 array in
question.
I had to boot from a Fedora Core 3 Rescue CDROM.

# uname -a 
Linux localhost.localdomain 2.6.9-1.667 #1 Tue Nov 2 14:41:31 EST 2004
i686 unknown

On boot time I get the following:

[...]
md: autorun ...
md: considering hdi7 ...
md:  adding hdi7 ...
md:  adding hdk9 ...
md:  adding hdg5 ...
md:  adding hde10 ...
md:  adding hda11 ...
md: created md4
md: bind<hda11>
md: bind<hde10>
md: bind<hdg5>
md: bind<hdk9>
md: bind<hdi7>
md: running: <hdi7><hdk9><hdg5><hde10><hda11>
md: kicking non-fresh hde10 from array!
md: unbind<hde10>
md: export_rdev(hde10)
md: md4: raid array is not clean -- starting background reconstruction
raid5: device hdi7 operational as raid disk 4
raid5: device hdk9 operational as raid disk 3
raid5: device hdg5 operational as raid disk 2
raid5: device hda11 operational as raid disk 0
raid5: cannot start dirty degraded array for md4
RAID5 conf printout:
 --- rd:5 wd:4 fd:1
 disk 0, o:1, dev:hda11
 disk 2, o:1, dev:hdg5
 disk 3, o:1, dev:hdk9
 disk 4, o:1, dev:hdi7
raid5: failed to run raid set md4
md: pers->run() failed ...
md :do_md_run() returned -22
md: md4 stopped.
md: unbind<hdi7>
md: export_rdev(hdi7)
md: unbind<hdk9>
md: export_rdev(hdk9)
md: unbind<hdg5>
md: export_rdev(hdg5)
md: unbind<hda11>
md: export_rdev(hda11)
md: ... autorun DONE.
[...]

So I tried to reassemble the array:

# mdadm --assemble /dev/md4 /dev/hda11 /dev/hde10 /dev/hdg5 /dev/hdk9
/dev/hdi7
mdadm: /dev/md4 assembled from 4 drives - need all 5 to start it (use
--run to insist)

# dmesg
[...]
md: md4 stopped.
md: bind<hde10>
md: bind<hdg5>
md: bind<hdk9>
md: bind<hdi7>
md: bind<hda11>

# cat /proc/mdstat
Personalities : [raid0] [raid1] [raid5] [raid6]
md1 : active raid5 hdi1[4] hdk1[3] hdg1[2] hde7[1] hda3[0]
      81919744 blocks level 5, 64k chunk, algorithm 2 [5/5] [UUUUU]

md2 : active raid5 hdi2[4] hdk2[3] hdg2[2] hde8[1] hda5[0]
      81919744 blocks level 5, 64k chunk, algorithm 2 [5/5] [UUUUU]

md3 : active raid5 hdi3[4] hdk3[3] hdg3[2] hde9[1] hda6[0]
      81919744 blocks level 5, 64k chunk, algorithm 2 [5/5] [UUUUU]

md4 : inactive hda11[0] hdi7[4] hdk9[3] hdg5[2] hde10[1]
      65246272 blocks
md5 : active raid5 hdl5[3] hdi5[2] hdk5[1] hda7[0]
      61439808 blocks level 5, 64k chunk, algorithm 0 [4/4] [UUUU]

md6 : active raid5 hdl6[3] hdi6[2] hdk6[1] hda8[0]
      61439808 blocks level 5, 64k chunk, algorithm 0 [4/4] [UUUU]

md7 : active raid5 hdl7[2] hdk7[1] hda9[0]
      40965504 blocks level 5, 64k chunk, algorithm 0 [3/3] [UUU]

md8 : active raid5 hdl8[2] hdk8[1] hda10[0]
      40965504 blocks level 5, 64k chunk, algorithm 0 [3/3] [UUU]

unused devices: <none>


# mdadm --stop /dev/md4
# mdadm --assemble --run /dev/md4 /dev/hda11 /dev/hde10 /dev/hdg5
/dev/hdk9 /dev/hdi7
mdadm: /dev/md4 has been started with 4 drives (out of 5).

# cat /proc/mdstat
[...]
md4 : active raid5 hda11[0] hdi7[4] hdk9[3] hdg5[2]
      49126144 blocks level 5, 64k chunk, algorithm 2 [5/4] [U_UUU]
[...]

# dmesg
[...]
md: bind<hde10>
md: bind<hdg5>
md: bind<hdk9>
md: bind<hdi7>
md: bind<hda11>
md: kicking non-fresh hde10 from array!
md: unbind<hde10>
md: export_rdev(hde10)
raid5: device hda11 operational as raid disk 0
raid5: device hdi7 operational as raid disk 4
raid5: device hdk9 operational as raid disk 3
raid5: device hdg5 operational as raid disk 2
raid5: allocated 5248kB for md4
raid5: raid level 5 set md4 active with 4 out of 5 devices, algorithm 2
RAID5 conf printout:
 --- rd:5 wd:4 fd:1
 disk 0, o:1, dev:hda11
 disk 2, o:1, dev:hdg5
 disk 3, o:1, dev:hdk9
 disk 4, o:1, dev:hdi7


So far everything looks ok for me.
But now things become funny:

# dd if=/dev/md4 of=/dev/null
0+0 records in
0+0 records out

# mdadm --stop /dev/md4
mdadm: fail to stop array /dev/md4: Device or resource busy

# dmesg
[...]
md: md4 still in use.

# dd if=/dev/hda11 of=/dev/null count=1000
1000+0 records in
1000+0 records out
# dd if=/dev/hde10 of=/dev/null count=1000
1000+0 records in
1000+0 records out
# dd if=/dev/hdg5 of=/dev/null count=1000
1000+0 records in
1000+0 records out
# dd if=/dev/hdi7 of=/dev/null count=1000
1000+0 records in
1000+0 records out
# dd if=/dev/hdk9 of=/dev/null count=1000
1000+0 records in
1000+0 records out
# dd if=/dev/md1 of=/dev/null count=1000
1000+0 records in
1000+0 records out
# dd if=/dev/md2 of=/dev/null count=1000
1000+0 records in
1000+0 records out
# dd if=/dev/md3 of=/dev/null count=1000
1000+0 records in
1000+0 records out
# dd if=/dev/md5 of=/dev/null count=1000
1000+0 records in
1000+0 records out
# dd if=/dev/md6 of=/dev/null count=1000
1000+0 records in
1000+0 records out
# dd if=/dev/md7 of=/dev/null count=1000
1000+0 records in
1000+0 records out
# dd if=/dev/md8 of=/dev/null count=1000
1000+0 records in
1000+0 records out


Now some still missing details:

# mdadm --detail /dev/md4
/dev/md4:
        Version : 00.90.01
  Creation Time : Sat Jul 24 12:38:25 2004
     Raid Level : raid5
    Device Size : 12281536 (11.71 GiB 12.58 GB)
   Raid Devices : 5
  Total Devices : 4
Preferred Minor : 4
    Persistence : Superblock is persistent

    Update Time : Mon Feb 28 21:10:13 2005
          State : clean, degraded
 Active Devices : 4
Working Devices : 4
 Failed Devices : 0
  Spare Devices : 0

         Layout : left-symmetric
     Chunk Size : 64K

    Number   Major   Minor   RaidDevice State
       0       3       11        0      active sync   /dev/hda11
       1       0        0       -1      removed
       2      34        5        2      active sync   /dev/hdg5
       3      57        9        3      active sync   /dev/hdk9
       4      56        7        4      active sync   /dev/hdi7
           UUID : 1da63142:e1bcc45b:e0287a1a:f9c7c3a8
         Events : 0.26324

# mdadm --examine /dev/hda11 /dev/hde10 /dev/hdg5 /dev/hdi7 /dev/hdk9
/dev/hda11:
          Magic : a92b4efc
        Version : 00.90.00
           UUID : 1da63142:e1bcc45b:e0287a1a:f9c7c3a8
  Creation Time : Sat Jul 24 12:38:25 2004
     Raid Level : raid5
    Device Size : 12281536 (11.71 GiB 12.58 GB)
   Raid Devices : 5
  Total Devices : 5
Preferred Minor : 4

    Update Time : Mon Feb 28 21:10:13 2005
          State : clean
 Active Devices : 5
Working Devices : 5
 Failed Devices : 0
  Spare Devices : 0
       Checksum : 661328a - correct
         Events : 0.26324

         Layout : left-symmetric
     Chunk Size : 64K

      Number   Major   Minor   RaidDevice State
this     0       3       11        0      active sync   /dev/hda11
   0     0       3       11        0      active sync   /dev/hda11
   1     1      33       10        1      active sync   /dev/hde10
   2     2      34        5        2      active sync   /dev/hdg5
   3     3      57        9        3      active sync   /dev/hdk9
   4     4      56        7        4      active sync   /dev/hdi7
/dev/hde10:
          Magic : a92b4efc
        Version : 00.90.00
           UUID : 1da63142:e1bcc45b:e0287a1a:f9c7c3a8
  Creation Time : Sat Jul 24 12:38:25 2004
     Raid Level : raid5
    Device Size : 12281536 (11.71 GiB 12.58 GB)
   Raid Devices : 5
  Total Devices : 5
Preferred Minor : 4

    Update Time : Mon Feb 28 21:10:13 2005
          State : dirty
 Active Devices : 5
Working Devices : 5
 Failed Devices : 0
  Spare Devices : 0
       Checksum : 66132a6 - correct
         Events : 0.26322

         Layout : left-symmetric
     Chunk Size : 64K

      Number   Major   Minor   RaidDevice State
this     1      33       10        1      active sync   /dev/hde10
   0     0       3       11        0      active sync   /dev/hda11
   1     1      33       10        1      active sync   /dev/hde10
   2     2      34        5        2      active sync   /dev/hdg5
   3     3      57        9        3      active sync   /dev/hdk9
   4     4      56        7        4      active sync   /dev/hdi7
/dev/hdg5:
          Magic : a92b4efc
        Version : 00.90.00
           UUID : 1da63142:e1bcc45b:e0287a1a:f9c7c3a8
  Creation Time : Sat Jul 24 12:38:25 2004
     Raid Level : raid5
    Device Size : 12281536 (11.71 GiB 12.58 GB)
   Raid Devices : 5
  Total Devices : 5
Preferred Minor : 4

    Update Time : Mon Feb 28 21:10:13 2005
          State : dirty
 Active Devices : 5
Working Devices : 5
 Failed Devices : 0
  Spare Devices : 0
       Checksum : 66132a6 - correct
         Events : 0.26324

         Layout : left-symmetric
     Chunk Size : 64K

      Number   Major   Minor   RaidDevice State
this     2      34        5        2      active sync   /dev/hdg5
   0     0       3       11        0      active sync   /dev/hda11
   1     1      33       10        1      active sync   /dev/hde10
   2     2      34        5        2      active sync   /dev/hdg5
   3     3      57        9        3      active sync   /dev/hdk9
   4     4      56        7        4      active sync   /dev/hdi7
/dev/hdi7:
          Magic : a92b4efc
        Version : 00.90.00
           UUID : 1da63142:e1bcc45b:e0287a1a:f9c7c3a8
  Creation Time : Sat Jul 24 12:38:25 2004
     Raid Level : raid5
    Device Size : 12281536 (11.71 GiB 12.58 GB)
   Raid Devices : 5
  Total Devices : 5
Preferred Minor : 4

    Update Time : Mon Feb 28 21:10:13 2005
          State : dirty
 Active Devices : 5
Working Devices : 5
 Failed Devices : 0
  Spare Devices : 0
       Checksum : 66132c2 - correct
         Events : 0.26324

         Layout : left-symmetric
     Chunk Size : 64K

      Number   Major   Minor   RaidDevice State
this     4      56        7        4      active sync   /dev/hdi7
   0     0       3       11        0      active sync   /dev/hda11
   1     1      33       10        1      active sync   /dev/hde10
   2     2      34        5        2      active sync   /dev/hdg5
   3     3      57        9        3      active sync   /dev/hdk9
   4     4      56        7        4      active sync   /dev/hdi7
/dev/hdk9:
          Magic : a92b4efc
        Version : 00.90.00
           UUID : 1da63142:e1bcc45b:e0287a1a:f9c7c3a8
  Creation Time : Sat Jul 24 12:38:25 2004
     Raid Level : raid5
    Device Size : 12281536 (11.71 GiB 12.58 GB)
   Raid Devices : 5
  Total Devices : 5
Preferred Minor : 4

    Update Time : Mon Feb 28 21:10:13 2005
          State : dirty
 Active Devices : 5
Working Devices : 5
 Failed Devices : 0
  Spare Devices : 0
       Checksum : 66132c3 - correct
         Events : 0.26324

         Layout : left-symmetric
     Chunk Size : 64K

      Number   Major   Minor   RaidDevice State
this     3      57        9        3      active sync   /dev/hdk9
   0     0       3       11        0      active sync   /dev/hda11
   1     1      33       10        1      active sync   /dev/hde10
   2     2      34        5        2      active sync   /dev/hdg5
   3     3      57        9        3      active sync   /dev/hdk9
   4     4      56        7        4      active sync   /dev/hdi7


I really would appreciate some help.

Regards,
Peter

-- 
Hans Peter Gundelwein
Email: hpg@Gundelwein.de


^ permalink raw reply	[flat|nested] 3+ messages in thread

* Re: hostRAID, AIC-7901, disable?
  2005-03-04 19:42 strange behavior of a raid5 array after system crash hpg
@ 2005-03-04 20:10 ` Stefan Eckert
  2005-03-04 20:57 ` strange behavior of a raid5 array after system crash Mike Tran
  1 sibling, 0 replies; 3+ messages in thread
From: Stefan Eckert @ 2005-03-04 20:10 UTC (permalink / raw)
  To: linux-raid

> 
>  We have recently aquired a new Dell Precision 670 workstation, on 
>  which
>  we intended to install linux, (Suse 9.1). 
>  In the install, is reports it can not find any hard drives.
>  We think we have tracked this problem down to the AIC-7901
>  hostRAID 
>  SCSI controller. From what I we can gather, linux no longer has
>  drivers for hostRAID hardware. We don't care about having RAID, >     
>  just
>  a working system, so we would be content to simply disable the 
>  hostRAID. the problem is, the setup tool does not have any option
>  to disable hostRAID.
>  

Hello Murray,

I have got the same machine (Dell Precision 670 workstation) here, 
and after trying severel ways of installing SuSE Linux 9.2 on it, 
I'm almost becoming desperate and I begin to hate SuSE Linux.
I just wonder if you finally succeeded to install SuSE 9.1 on your 
machine? If so, I'm asking you, how you did it. I really hope you 
can give me some advice.

Thanks in advance,
   Stefan

^ permalink raw reply	[flat|nested] 3+ messages in thread

* Re: strange behavior of a raid5 array after system crash
  2005-03-04 19:42 strange behavior of a raid5 array after system crash hpg
  2005-03-04 20:10 ` hostRAID, AIC-7901, disable? Stefan Eckert
@ 2005-03-04 20:57 ` Mike Tran
  1 sibling, 0 replies; 3+ messages in thread
From: Mike Tran @ 2005-03-04 20:57 UTC (permalink / raw)
  To: linux-raid

Hello Hans,

I would try to re-add the out-of-sync disk (hde10) back to the degraded
raid5 array (md4).  If hde10 got kicked out again, it's time to replace
it with another disk.

--
Regards,
Mike T.

On Fri, 2005-03-04 at 13:42, hpg@gundelwein.de wrote:
> Hello everyone,
> 
> I need your help with a strange behavior of a raid5 array.
> 
> My Linux fileserver was frozen for unknown reason. No mouse movement, 
> no console, no disk activity nothing.
> So I had to hit the reset button.
> 
> At boot time 5 raid5 arrays have been active without any faults.
> Two other raid5 arrays resynchronized successfully.
> Only one had some trouble to recover.
> 
> Because I am using LVM2 on top of all my raid5 arrays and have the root 
> filesystem in that volume group which is using the raid5 array in
> question.
> I had to boot from a Fedora Core 3 Rescue CDROM.
> 
> # uname -a 
> Linux localhost.localdomain 2.6.9-1.667 #1 Tue Nov 2 14:41:31 EST 2004
> i686 unknown
> 
> On boot time I get the following:
> 
> [...]
> md: autorun ...
> md: considering hdi7 ...
> md:  adding hdi7 ...
> md:  adding hdk9 ...
> md:  adding hdg5 ...
> md:  adding hde10 ...
> md:  adding hda11 ...
> md: created md4
> md: bind<hda11>
> md: bind<hde10>
> md: bind<hdg5>
> md: bind<hdk9>
> md: bind<hdi7>
> md: running: <hdi7><hdk9><hdg5><hde10><hda11>
> md: kicking non-fresh hde10 from array!
> md: unbind<hde10>
> md: export_rdev(hde10)
> md: md4: raid array is not clean -- starting background reconstruction
> raid5: device hdi7 operational as raid disk 4
> raid5: device hdk9 operational as raid disk 3
> raid5: device hdg5 operational as raid disk 2
> raid5: device hda11 operational as raid disk 0
> raid5: cannot start dirty degraded array for md4
> RAID5 conf printout:
>  --- rd:5 wd:4 fd:1
>  disk 0, o:1, dev:hda11
>  disk 2, o:1, dev:hdg5
>  disk 3, o:1, dev:hdk9
>  disk 4, o:1, dev:hdi7
> raid5: failed to run raid set md4
> md: pers->run() failed ...
> md :do_md_run() returned -22
> md: md4 stopped.
> md: unbind<hdi7>
> md: export_rdev(hdi7)
> md: unbind<hdk9>
> md: export_rdev(hdk9)
> md: unbind<hdg5>
> md: export_rdev(hdg5)
> md: unbind<hda11>
> md: export_rdev(hda11)
> md: ... autorun DONE.
> [...]
> 
> So I tried to reassemble the array:
> 
> # mdadm --assemble /dev/md4 /dev/hda11 /dev/hde10 /dev/hdg5 /dev/hdk9
> /dev/hdi7
> mdadm: /dev/md4 assembled from 4 drives - need all 5 to start it (use
> --run to insist)
> 
> # dmesg
> [...]
> md: md4 stopped.
> md: bind<hde10>
> md: bind<hdg5>
> md: bind<hdk9>
> md: bind<hdi7>
> md: bind<hda11>
> 
> # cat /proc/mdstat
> Personalities : [raid0] [raid1] [raid5] [raid6]
> md1 : active raid5 hdi1[4] hdk1[3] hdg1[2] hde7[1] hda3[0]
>       81919744 blocks level 5, 64k chunk, algorithm 2 [5/5] [UUUUU]
> 
> md2 : active raid5 hdi2[4] hdk2[3] hdg2[2] hde8[1] hda5[0]
>       81919744 blocks level 5, 64k chunk, algorithm 2 [5/5] [UUUUU]
> 
> md3 : active raid5 hdi3[4] hdk3[3] hdg3[2] hde9[1] hda6[0]
>       81919744 blocks level 5, 64k chunk, algorithm 2 [5/5] [UUUUU]
> 
> md4 : inactive hda11[0] hdi7[4] hdk9[3] hdg5[2] hde10[1]
>       65246272 blocks
> md5 : active raid5 hdl5[3] hdi5[2] hdk5[1] hda7[0]
>       61439808 blocks level 5, 64k chunk, algorithm 0 [4/4] [UUUU]
> 
> md6 : active raid5 hdl6[3] hdi6[2] hdk6[1] hda8[0]
>       61439808 blocks level 5, 64k chunk, algorithm 0 [4/4] [UUUU]
> 
> md7 : active raid5 hdl7[2] hdk7[1] hda9[0]
>       40965504 blocks level 5, 64k chunk, algorithm 0 [3/3] [UUU]
> 
> md8 : active raid5 hdl8[2] hdk8[1] hda10[0]
>       40965504 blocks level 5, 64k chunk, algorithm 0 [3/3] [UUU]
> 
> unused devices: <none>
> 
> 
> # mdadm --stop /dev/md4
> # mdadm --assemble --run /dev/md4 /dev/hda11 /dev/hde10 /dev/hdg5
> /dev/hdk9 /dev/hdi7
> mdadm: /dev/md4 has been started with 4 drives (out of 5).
> 
> # cat /proc/mdstat
> [...]
> md4 : active raid5 hda11[0] hdi7[4] hdk9[3] hdg5[2]
>       49126144 blocks level 5, 64k chunk, algorithm 2 [5/4] [U_UUU]
> [...]
> 
> # dmesg
> [...]
> md: bind<hde10>
> md: bind<hdg5>
> md: bind<hdk9>
> md: bind<hdi7>
> md: bind<hda11>
> md: kicking non-fresh hde10 from array!
> md: unbind<hde10>
> md: export_rdev(hde10)
> raid5: device hda11 operational as raid disk 0
> raid5: device hdi7 operational as raid disk 4
> raid5: device hdk9 operational as raid disk 3
> raid5: device hdg5 operational as raid disk 2
> raid5: allocated 5248kB for md4
> raid5: raid level 5 set md4 active with 4 out of 5 devices, algorithm 2
> RAID5 conf printout:
>  --- rd:5 wd:4 fd:1
>  disk 0, o:1, dev:hda11
>  disk 2, o:1, dev:hdg5
>  disk 3, o:1, dev:hdk9
>  disk 4, o:1, dev:hdi7
> 
> 
> So far everything looks ok for me.
> But now things become funny:
> 
> # dd if=/dev/md4 of=/dev/null
> 0+0 records in
> 0+0 records out
> 
> # mdadm --stop /dev/md4
> mdadm: fail to stop array /dev/md4: Device or resource busy
> 
> # dmesg
> [...]
> md: md4 still in use.
> 
> # dd if=/dev/hda11 of=/dev/null count=1000
> 1000+0 records in
> 1000+0 records out
> # dd if=/dev/hde10 of=/dev/null count=1000
> 1000+0 records in
> 1000+0 records out
> # dd if=/dev/hdg5 of=/dev/null count=1000
> 1000+0 records in
> 1000+0 records out
> # dd if=/dev/hdi7 of=/dev/null count=1000
> 1000+0 records in
> 1000+0 records out
> # dd if=/dev/hdk9 of=/dev/null count=1000
> 1000+0 records in
> 1000+0 records out
> # dd if=/dev/md1 of=/dev/null count=1000
> 1000+0 records in
> 1000+0 records out
> # dd if=/dev/md2 of=/dev/null count=1000
> 1000+0 records in
> 1000+0 records out
> # dd if=/dev/md3 of=/dev/null count=1000
> 1000+0 records in
> 1000+0 records out
> # dd if=/dev/md5 of=/dev/null count=1000
> 1000+0 records in
> 1000+0 records out
> # dd if=/dev/md6 of=/dev/null count=1000
> 1000+0 records in
> 1000+0 records out
> # dd if=/dev/md7 of=/dev/null count=1000
> 1000+0 records in
> 1000+0 records out
> # dd if=/dev/md8 of=/dev/null count=1000
> 1000+0 records in
> 1000+0 records out
> 
> 
> Now some still missing details:
> 
> # mdadm --detail /dev/md4
> /dev/md4:
>         Version : 00.90.01
>   Creation Time : Sat Jul 24 12:38:25 2004
>      Raid Level : raid5
>     Device Size : 12281536 (11.71 GiB 12.58 GB)
>    Raid Devices : 5
>   Total Devices : 4
> Preferred Minor : 4
>     Persistence : Superblock is persistent
> 
>     Update Time : Mon Feb 28 21:10:13 2005
>           State : clean, degraded
>  Active Devices : 4
> Working Devices : 4
>  Failed Devices : 0
>   Spare Devices : 0
> 
>          Layout : left-symmetric
>      Chunk Size : 64K
> 
>     Number   Major   Minor   RaidDevice State
>        0       3       11        0      active sync   /dev/hda11
>        1       0        0       -1      removed
>        2      34        5        2      active sync   /dev/hdg5
>        3      57        9        3      active sync   /dev/hdk9
>        4      56        7        4      active sync   /dev/hdi7
>            UUID : 1da63142:e1bcc45b:e0287a1a:f9c7c3a8
>          Events : 0.26324
> 
> # mdadm --examine /dev/hda11 /dev/hde10 /dev/hdg5 /dev/hdi7 /dev/hdk9
> /dev/hda11:
>           Magic : a92b4efc
>         Version : 00.90.00
>            UUID : 1da63142:e1bcc45b:e0287a1a:f9c7c3a8
>   Creation Time : Sat Jul 24 12:38:25 2004
>      Raid Level : raid5
>     Device Size : 12281536 (11.71 GiB 12.58 GB)
>    Raid Devices : 5
>   Total Devices : 5
> Preferred Minor : 4
> 
>     Update Time : Mon Feb 28 21:10:13 2005
>           State : clean
>  Active Devices : 5
> Working Devices : 5
>  Failed Devices : 0
>   Spare Devices : 0
>        Checksum : 661328a - correct
>          Events : 0.26324
> 
>          Layout : left-symmetric
>      Chunk Size : 64K
> 
>       Number   Major   Minor   RaidDevice State
> this     0       3       11        0      active sync   /dev/hda11
>    0     0       3       11        0      active sync   /dev/hda11
>    1     1      33       10        1      active sync   /dev/hde10
>    2     2      34        5        2      active sync   /dev/hdg5
>    3     3      57        9        3      active sync   /dev/hdk9
>    4     4      56        7        4      active sync   /dev/hdi7
> /dev/hde10:
>           Magic : a92b4efc
>         Version : 00.90.00
>            UUID : 1da63142:e1bcc45b:e0287a1a:f9c7c3a8
>   Creation Time : Sat Jul 24 12:38:25 2004
>      Raid Level : raid5
>     Device Size : 12281536 (11.71 GiB 12.58 GB)
>    Raid Devices : 5
>   Total Devices : 5
> Preferred Minor : 4
> 
>     Update Time : Mon Feb 28 21:10:13 2005
>           State : dirty
>  Active Devices : 5
> Working Devices : 5
>  Failed Devices : 0
>   Spare Devices : 0
>        Checksum : 66132a6 - correct
>          Events : 0.26322
> 
>          Layout : left-symmetric
>      Chunk Size : 64K
> 
>       Number   Major   Minor   RaidDevice State
> this     1      33       10        1      active sync   /dev/hde10
>    0     0       3       11        0      active sync   /dev/hda11
>    1     1      33       10        1      active sync   /dev/hde10
>    2     2      34        5        2      active sync   /dev/hdg5
>    3     3      57        9        3      active sync   /dev/hdk9
>    4     4      56        7        4      active sync   /dev/hdi7
> /dev/hdg5:
>           Magic : a92b4efc
>         Version : 00.90.00
>            UUID : 1da63142:e1bcc45b:e0287a1a:f9c7c3a8
>   Creation Time : Sat Jul 24 12:38:25 2004
>      Raid Level : raid5
>     Device Size : 12281536 (11.71 GiB 12.58 GB)
>    Raid Devices : 5
>   Total Devices : 5
> Preferred Minor : 4
> 
>     Update Time : Mon Feb 28 21:10:13 2005
>           State : dirty
>  Active Devices : 5
> Working Devices : 5
>  Failed Devices : 0
>   Spare Devices : 0
>        Checksum : 66132a6 - correct
>          Events : 0.26324
> 
>          Layout : left-symmetric
>      Chunk Size : 64K
> 
>       Number   Major   Minor   RaidDevice State
> this     2      34        5        2      active sync   /dev/hdg5
>    0     0       3       11        0      active sync   /dev/hda11
>    1     1      33       10        1      active sync   /dev/hde10
>    2     2      34        5        2      active sync   /dev/hdg5
>    3     3      57        9        3      active sync   /dev/hdk9
>    4     4      56        7        4      active sync   /dev/hdi7
> /dev/hdi7:
>           Magic : a92b4efc
>         Version : 00.90.00
>            UUID : 1da63142:e1bcc45b:e0287a1a:f9c7c3a8
>   Creation Time : Sat Jul 24 12:38:25 2004
>      Raid Level : raid5
>     Device Size : 12281536 (11.71 GiB 12.58 GB)
>    Raid Devices : 5
>   Total Devices : 5
> Preferred Minor : 4
> 
>     Update Time : Mon Feb 28 21:10:13 2005
>           State : dirty
>  Active Devices : 5
> Working Devices : 5
>  Failed Devices : 0
>   Spare Devices : 0
>        Checksum : 66132c2 - correct
>          Events : 0.26324
> 
>          Layout : left-symmetric
>      Chunk Size : 64K
> 
>       Number   Major   Minor   RaidDevice State
> this     4      56        7        4      active sync   /dev/hdi7
>    0     0       3       11        0      active sync   /dev/hda11
>    1     1      33       10        1      active sync   /dev/hde10
>    2     2      34        5        2      active sync   /dev/hdg5
>    3     3      57        9        3      active sync   /dev/hdk9
>    4     4      56        7        4      active sync   /dev/hdi7
> /dev/hdk9:
>           Magic : a92b4efc
>         Version : 00.90.00
>            UUID : 1da63142:e1bcc45b:e0287a1a:f9c7c3a8
>   Creation Time : Sat Jul 24 12:38:25 2004
>      Raid Level : raid5
>     Device Size : 12281536 (11.71 GiB 12.58 GB)
>    Raid Devices : 5
>   Total Devices : 5
> Preferred Minor : 4
> 
>     Update Time : Mon Feb 28 21:10:13 2005
>           State : dirty
>  Active Devices : 5
> Working Devices : 5
>  Failed Devices : 0
>   Spare Devices : 0
>        Checksum : 66132c3 - correct
>          Events : 0.26324
> 
>          Layout : left-symmetric
>      Chunk Size : 64K
> 
>       Number   Major   Minor   RaidDevice State
> this     3      57        9        3      active sync   /dev/hdk9
>    0     0       3       11        0      active sync   /dev/hda11
>    1     1      33       10        1      active sync   /dev/hde10
>    2     2      34        5        2      active sync   /dev/hdg5
>    3     3      57        9        3      active sync   /dev/hdk9
>    4     4      56        7        4      active sync   /dev/hdi7
> 
> 
> I really would appreciate some help.
> 
> Regards,
> Peter


^ permalink raw reply	[flat|nested] 3+ messages in thread

end of thread, other threads:[~2005-03-04 20:57 UTC | newest]

Thread overview: 3+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2005-03-04 19:42 strange behavior of a raid5 array after system crash hpg
2005-03-04 20:10 ` hostRAID, AIC-7901, disable? Stefan Eckert
2005-03-04 20:57 ` strange behavior of a raid5 array after system crash Mike Tran

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).