linux-raid.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
* "cannot start dirty degraded array"
@ 2009-06-10 20:03 Kyler Laird
  2009-06-14 23:32 ` Carlos Carvalho
  2009-06-15 15:54 ` Bill Davidsen
  0 siblings, 2 replies; 5+ messages in thread
From: Kyler Laird @ 2009-06-10 20:03 UTC (permalink / raw)
  To: linux-raid

I'm in a bind.  I have three RAID6s on a Sun X4540.  A bunch of disks
threw error all of a sudden.  Two arrays came back (degraded) on reboot
but the third is having problems.  

	root@00144ff2a334:/tmp# mdadm -E /dev/sdag                                                                                       
	/dev/sdag:                                                                                                                       
		  Magic : a92b4efc                                                                                                       
		Version : 00.90.00                                                                                                       
		   UUID : fe5c8175:b45887a9:0fe23972:c126e925 (local to host 00144ff2a334)                                               
	  Creation Time : Mon Apr 27 22:14:16 2009
	     Raid Level : raid6
	  Used Dev Size : 976762496 (931.51 GiB 1000.20 GB)
	     Array Size : 12697912448 (12109.67 GiB 13002.66 GB)
	   Raid Devices : 15
	  Total Devices : 16
	Preferred Minor : 2

	    Update Time : Sun Jun  7 07:28:53 2009
		  State : clean
	 Active Devices : 15
	Working Devices : 16
	 Failed Devices : 0
	  Spare Devices : 1
	       Checksum : fb439997 - correct
		 Events : 0.16

	     Chunk Size : 16K

	      Number   Major   Minor   RaidDevice State

Here's the dmesg output when I try to assemble using "mdadm --assemble
/dev/md2 /dev/sda[g-v]".


	[ 3469.703048] md: bind<sdag>
	[ 3469.703305] md: bind<sdai>
	[ 3469.703554] md: bind<sdaj>
	[ 3469.703806] md: bind<sdak>
	[ 3469.704043] md: bind<sdal>
	[ 3469.704294] md: bind<sdam>
	[ 3469.704544] md: bind<sdan>
	[ 3469.775701] md: bind<sdao>
	[ 3469.775946] md: bind<sdap>
	[ 3469.776198] md: bind<sdaq>
	[ 3469.776453] md: bind<sdar>
	[ 3469.776695] md: bind<sdas>
	[ 3469.776953] md: bind<sdat>
	[ 3469.777204] md: bind<sdau>
	[ 3469.777442] md: bind<sdav>
	[ 3469.777698] md: bind<sdah>
	[ 3469.777762] md: kicking non-fresh sdag from array!
	[ 3469.777766] md: unbind<sdag>
	[ 3469.801894] md: export_rdev(sdag)
	[ 3469.801898] md: md2: raid array is not clean -- starting background reconstruction
	[ 3469.825589] raid5: device sdah operational as raid disk 1
	[ 3469.825591] raid5: device sdau operational as raid disk 14
	[ 3469.825593] raid5: device sdat operational as raid disk 13
	[ 3469.825594] raid5: device sdas operational as raid disk 12
	[ 3469.825595] raid5: device sdar operational as raid disk 11
	[ 3469.825596] raid5: device sdaq operational as raid disk 10
	[ 3469.825597] raid5: device sdap operational as raid disk 9
	[ 3469.825598] raid5: device sdao operational as raid disk 8
	[ 3469.825599] raid5: device sdan operational as raid disk 7
	[ 3469.825600] raid5: device sdam operational as raid disk 6
	[ 3469.825601] raid5: device sdal operational as raid disk 5
	[ 3469.825602] raid5: device sdak operational as raid disk 4
	[ 3469.825603] raid5: device sdaj operational as raid disk 3
	[ 3469.825604] raid5: device sdai operational as raid disk 2
	[ 3469.825606] raid5: cannot start dirty degraded array for md2
	[ 3469.825674] RAID5 conf printout:
	[ 3469.825675]  --- rd:15 wd:14
	[ 3469.825676]  disk 1, o:1, dev:sdah
	[ 3469.825677]  disk 2, o:1, dev:sdai
	[ 3469.825678]  disk 3, o:1, dev:sdaj
	[ 3469.825679]  disk 4, o:1, dev:sdak
	[ 3469.825679]  disk 5, o:1, dev:sdal
	[ 3469.825680]  disk 6, o:1, dev:sdam
	[ 3469.825681]  disk 7, o:1, dev:sdan
	[ 3469.825682]  disk 8, o:1, dev:sdao
	[ 3469.825683]  disk 9, o:1, dev:sdap
	[ 3469.825684]  disk 10, o:1, dev:sdaq
	[ 3469.825685]  disk 11, o:1, dev:sdar
	[ 3469.825685]  disk 12, o:1, dev:sdas
	[ 3469.825686]  disk 13, o:1, dev:sdat
	[ 3469.825687]  disk 14, o:1, dev:sdau
	[ 3469.825689] raid5: failed to run raid set md2
	[ 3469.825751] md: pers->run() failed ...

I ran a check of the disks' SMART data and they seem to be alright.
We're getting ready to ship the entire unit to a data recovery center
but I'd really like to know if there's something simple we can do first.
I've been frantically reading about this issue but none of the simple
solutions seem to work for me.  I welcome suggestions.

Thanks!

--kyler

^ permalink raw reply	[flat|nested] 5+ messages in thread

* Re: "cannot start dirty degraded array"
  2009-06-10 20:03 "cannot start dirty degraded array" Kyler Laird
@ 2009-06-14 23:32 ` Carlos Carvalho
  2009-06-14 23:56   ` Kyler Laird
  2009-06-15 15:54 ` Bill Davidsen
  1 sibling, 1 reply; 5+ messages in thread
From: Carlos Carvalho @ 2009-06-14 23:32 UTC (permalink / raw)
  To: linux-raid

Kyler Laird (kyler-keyword-linuxraid.6e1399@lairds.com) wrote on 10 June 2009 16:03:
 >I'm in a bind.  I have three RAID6s on a Sun X4540.  A bunch of disks
 >threw error all of a sudden.  Two arrays came back (degraded) on reboot
 >but the third is having problems.  
...
 >Here's the dmesg output when I try to assemble using "mdadm --assemble
 >/dev/md2 /dev/sda[g-v]".
 >
 >
 >	[ 3469.703048] md: bind<sdag>
 >	[ 3469.703305] md: bind<sdai>
 >	[ 3469.703554] md: bind<sdaj>
 >	[ 3469.703806] md: bind<sdak>
 >	[ 3469.704043] md: bind<sdal>
 >	[ 3469.704294] md: bind<sdam>
 >	[ 3469.704544] md: bind<sdan>
 >	[ 3469.775701] md: bind<sdao>
 >	[ 3469.775946] md: bind<sdap>
 >	[ 3469.776198] md: bind<sdaq>
 >	[ 3469.776453] md: bind<sdar>
 >	[ 3469.776695] md: bind<sdas>
 >	[ 3469.776953] md: bind<sdat>
 >	[ 3469.777204] md: bind<sdau>
 >	[ 3469.777442] md: bind<sdav>
 >	[ 3469.777698] md: bind<sdah>
 >	[ 3469.777762] md: kicking non-fresh sdag from array!
 >	[ 3469.777766] md: unbind<sdag>
 >	[ 3469.801894] md: export_rdev(sdag)
 >	[ 3469.801898] md: md2: raid array is not clean -- starting background reconstruction
 >	[ 3469.825589] raid5: device sdah operational as raid disk 1
 >	[ 3469.825591] raid5: device sdau operational as raid disk 14
 >	[ 3469.825593] raid5: device sdat operational as raid disk 13
 >	[ 3469.825594] raid5: device sdas operational as raid disk 12
 >	[ 3469.825595] raid5: device sdar operational as raid disk 11
 >	[ 3469.825596] raid5: device sdaq operational as raid disk 10
 >	[ 3469.825597] raid5: device sdap operational as raid disk 9
 >	[ 3469.825598] raid5: device sdao operational as raid disk 8
 >	[ 3469.825599] raid5: device sdan operational as raid disk 7
 >	[ 3469.825600] raid5: device sdam operational as raid disk 6
 >	[ 3469.825601] raid5: device sdal operational as raid disk 5
 >	[ 3469.825602] raid5: device sdak operational as raid disk 4
 >	[ 3469.825603] raid5: device sdaj operational as raid disk 3
 >	[ 3469.825604] raid5: device sdai operational as raid disk 2
 >	[ 3469.825606] raid5: cannot start dirty degraded array for md2

You can try to use mdadm -A -f /dev/md2 <list of devices> to force the
array to assemble. Should work if all disks stopped simultaneously.

^ permalink raw reply	[flat|nested] 5+ messages in thread

* Re: "cannot start dirty degraded array"
  2009-06-14 23:32 ` Carlos Carvalho
@ 2009-06-14 23:56   ` Kyler Laird
  0 siblings, 0 replies; 5+ messages in thread
From: Kyler Laird @ 2009-06-14 23:56 UTC (permalink / raw)
  To: linux-raid

On Sun, Jun 14, 2009 at 08:32:22PM -0300, Carlos Carvalho wrote:

> You can try to use mdadm -A -f /dev/md2 <list of devices> to force the
> array to assemble. Should work if all disks stopped simultaneously.

I appreciate your response, Carlos.  I did try that before sending the
machine for recovery.  We're now working with a service that seems good
to me.  Here's their initial report.
	Md0; is made up of the first 16 physical drives, and the first 8
	drives are out of sync with the second eight.  Event codes are
	incorrect.  It appears that someone tried to start the raid (as
	in force) with only eight drives.  This raid will not reassemble
	without fixing the superblock hex structure and getting it back
	into alignment.

	Md1; is made up of the next 16 physical drives.  The first 8
	drives think the second set of 8 are faulty, but the event codes
	are OK.

	Md2; is made up of the last set of 16 physical drives.  The
	first two drives in this array think that everything is OK, but
	all the other drives show all manner of faults and drive removals.
$36K for standard recovery.  We're still working on it.

The drives all appear to be alright.  I suspect that there was a
kernel/controller problem.

Thank you.

--kyler

^ permalink raw reply	[flat|nested] 5+ messages in thread

* Re: "cannot start dirty degraded array"
  2009-06-10 20:03 "cannot start dirty degraded array" Kyler Laird
  2009-06-14 23:32 ` Carlos Carvalho
@ 2009-06-15 15:54 ` Bill Davidsen
  2009-06-15 15:57   ` Kyler Laird
  1 sibling, 1 reply; 5+ messages in thread
From: Bill Davidsen @ 2009-06-15 15:54 UTC (permalink / raw)
  To: Kyler Laird; +Cc: linux-raid

Kyler Laird wrote:
> I'm in a bind.  I have three RAID6s on a Sun X4540.  A bunch of disks
> threw error all of a sudden.  Two arrays came back (degraded) on reboot
> but the third is having problems.  
>   

Just a thought, when multiple units have errors at the same time, I 
suspect a power issue. And if these are real SCSI drives, it's possible 
for a drive to fail in such a way that it glitches the SCSI bus and 
causes the controller to think that multiple drives doing concurrent 
seeks have failed. I saw this often enough to have a script to force the 
controller to mark drives good and then test them one at a time when I 
was running ISP servers.

-- 
Bill Davidsen <davidsen@tmr.com>
  Obscure bug of 2004: BASH BUFFER OVERFLOW - if bash is being run by a
normal user and is setuid root, with the "vi" line edit mode selected,
and the character set is "big5," an off-by-one error occurs during
wildcard (glob) expansion.


^ permalink raw reply	[flat|nested] 5+ messages in thread

* Re: "cannot start dirty degraded array"
  2009-06-15 15:54 ` Bill Davidsen
@ 2009-06-15 15:57   ` Kyler Laird
  0 siblings, 0 replies; 5+ messages in thread
From: Kyler Laird @ 2009-06-15 15:57 UTC (permalink / raw)
  To: linux-raid

On Mon, Jun 15, 2009 at 11:54:22AM -0400, Bill Davidsen wrote:

> Just a thought, when multiple units have errors at the same time, I  
> suspect a power issue. And if these are real SCSI drives, it's possible  
> for a drive to fail in such a way that it glitches the SCSI bus and  
> causes the controller to think that multiple drives doing concurrent  
> seeks have failed. I saw this often enough to have a script to force the  
> controller to mark drives good and then test them one at a time when I  
> was running ISP servers.

The X4540 has 48 SATA drives.  We've had a couple X4500s (similar) in
similar service for awhile with no problems.  I hadn't considered power
being a potential problem but I'll keep it in mind.

Thank you.

--kyler

^ permalink raw reply	[flat|nested] 5+ messages in thread

end of thread, other threads:[~2009-06-15 15:57 UTC | newest]

Thread overview: 5+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2009-06-10 20:03 "cannot start dirty degraded array" Kyler Laird
2009-06-14 23:32 ` Carlos Carvalho
2009-06-14 23:56   ` Kyler Laird
2009-06-15 15:54 ` Bill Davidsen
2009-06-15 15:57   ` Kyler Laird

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).