linux-raid.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
* mdadm degraded RAID5 failure
       [not found] <6cc8e9ed0810221350o2b8b3aedm3d1c229fe7e66163@mail.gmail.com>
@ 2008-10-22 20:52 ` Steve Evans
  2008-10-24 18:47   ` Steve Evans
  2008-10-25  6:30   ` Neil Brown
  0 siblings, 2 replies; 7+ messages in thread
From: Steve Evans @ 2008-10-22 20:52 UTC (permalink / raw)
  To: linux-raid

Hi all..

I had one of the disks in my 3 disk RAID5 die on me this week. When
attempting to replace the disk via a hot swap (USB), the RAID didn't
like it. It decided to mark one of my remaining 2 disks as faulty.

Can someone *please* help me get the raid back!?

More details -

Drives are /dev/sdb1, /dev/sdc1 & /dev/sdd1

sdc1 was the one that died earlier this week
sdb1 appears to be the one that was marked as faulty

mdadm detail before sdc1 was plugged in -

root@imp[~]:11 # mdadm --detail /dev/md1
/dev/md1:
Version : 00.90.01
Creation Time : Fri Nov 17 21:28:44 2006
Raid Level : raid5
Array Size : 586067072 (558.92 GiB 600.13 GB)
Device Size : 293033536 (279.46 GiB 300.07 GB)
Raid Devices : 3
Total Devices : 2
Preferred Minor : 1
Persistence : Superblock is persistent

Update Time : Sat Oct 18 20:06:34 2008
State : clean, degraded
Active Devices : 2
Working Devices : 2
Failed Devices : 0
Spare Devices : 0

Layout : left-symmetric
Chunk Size : 64K

UUID : bed40ee2:98523fdd:e4d010fb:894c0966
Events : 0.1474312

Number Major Minor RaidDevice State
0 8 17 0 active sync /dev/sdb1
1 0 0 - removed
2 8 49 2 active sync /dev/sdd1


then after plugging in the replacement sdc1 -

root@imp[~]:13 # mdadm --add /dev/md1 /dev/sdc1
mdadm: hot added /dev/sdc1
root@imp[~]:14 #
root@imp[~]:14 #
root@imp[~]:14 # mdadm --detail /dev/md1
/dev/md1:
Version : 00.90.01
Creation Time : Fri Nov 17 21:28:44 2006
Raid Level : raid5
Array Size : 586067072 (558.92 GiB 600.13 GB)
Device Size : 293033536 (279.46 GiB 300.07 GB)
Raid Devices : 3
Total Devices : 3
Preferred Minor : 1
Persistence : Superblock is persistent

Update Time : Sat Oct 18 22:13:13 2008
State : clean, degraded
Active Devices : 1
Working Devices : 2
Failed Devices : 1
Spare Devices : 1

Layout : left-symmetric
Chunk Size : 64K

UUID : bed40ee2:98523fdd:e4d010fb:894c0966
Events : 0.1480366

Number Major Minor RaidDevice State
0 0 0 - removed
1 0 0 - removed
2 8 49 2 active sync /dev/sdd1

3 8 33 0 spare rebuilding /dev/sdc1
4 8 17 - faulty /dev/sdb1

Shortly after this, subsequent mdadm --details stopped responding.. So
I rebooted in the hope I could reset and problems with the hot add..

Now, I'm unable to assemble the raid with the 2 working drives -

mdadm --assemble /dev/md1 /dev/sdb1 /dev/sdd1

doesn't work -

mdadm: /dev/md1 assembled from 1 drive and 1 spare - not enough to
start the array.

mdadm --assemble --force /dev/md1 /dev/sdb1 /dev/sdd1

doesn't' work either

This -

mdadm --assemble --force --run /dev/md1 /dev/sdb1 /dev/sdd1

Did work partially -

/dev/md1:
Version : 00.90.01
Creation Time : Fri Nov 17 21:28:44 2006
Raid Level : raid5
Device Size : 293033536 (279.46 GiB 300.07 GB)
Raid Devices : 3
Total Devices : 2
Preferred Minor : 1
Persistence : Superblock is persistent

Update Time : Sat Oct 18 22:14:48 2008
State : active, degraded
Active Devices : 1
Working Devices : 2
Failed Devices : 0
Spare Devices : 1

Layout : left-symmetric
Chunk Size : 64K

UUID : bed40ee2:98523fdd:e4d010fb:894c0966
Events : 0.1521614

Number Major Minor RaidDevice State
0 0 0 - removed
1 0 0 - removed
2 8 49 2 active sync /dev/sdd1

3 8 17 - spare /dev/sdb1

Here's the output from mdadm -E on each of the 2 drives -

/dev/sdb1:
Magic : a92b4efc
Version : 00.90.00
UUID : bed40ee2:98523fdd:e4d010fb:894c0966
Creation Time : Fri Nov 17 21:28:44 2006
Raid Level : raid5
Raid Devices : 3
Total Devices : 3
Preferred Minor : 1

Update Time : Sat Oct 18 22:14:48 2008
State : clean
Active Devices : 1
Working Devices : 2
Failed Devices : 2
Spare Devices : 1
Checksum : e6dbf75 - correct
Events : 0.1521614

Layout : left-symmetric
Chunk Size : 64K

Number Major Minor RaidDevice State
this 3 8 33 3 spare /dev/sdc1

0 0 0 0 0 removed
1 1 0 0 1 faulty removed
2 2 8 49 2 active sync /dev/sdd1
3 3 8 33 3 spare /dev/sdc1
/dev/sdd1:
Magic : a92b4efc
Version : 00.90.00
UUID : bed40ee2:98523fdd:e4d010fb:894c0966
Creation Time : Fri Nov 17 21:28:44 2006
Raid Level : raid5
Raid Devices : 3
Total Devices : 3
Preferred Minor : 1

Update Time : Sat Oct 18 22:14:48 2008
State : clean
Active Devices : 1
Working Devices : 2
Failed Devices : 2
Spare Devices : 1
Checksum : e6dbf86 - correct
Events : 0.1521614

Layout : left-symmetric
Chunk Size : 64K

Number Major Minor RaidDevice State
this 2 8 49 2 active sync /dev/sdd1

0 0 0 0 0 removed
1 1 0 0 1 faulty removed
2 2 8 49 2 active sync /dev/sdd1
3 3 8 33 0 spare /dev/sdc1

root@imp[~]:28 # mdadm --version
mdadm - v1.9.0 - 04 February 2005
root@imp[~]:29 # uname -a
Linux imp 2.6.8-3-686 #1 Tue Dec 5 21:26:38 UTC 2006 i686 GNU/Linux


Is all the data lost, or can I recover from this?

Thanks so much!
Steve..

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: mdadm degraded RAID5 failure
  2008-10-22 20:52 ` mdadm degraded RAID5 failure Steve Evans
@ 2008-10-24 18:47   ` Steve Evans
  2008-10-25  6:30   ` Neil Brown
  1 sibling, 0 replies; 7+ messages in thread
From: Steve Evans @ 2008-10-24 18:47 UTC (permalink / raw)
  To: linux-raid

So it would appear the drive number has been remapped somehow on
/dev/sdb1 to be 3 instead of 0? Is there a way to reset/set these to
specific values?

3 8 17 - spare /dev/sdb1

Thanks
Steve..

On Wed, Oct 22, 2008 at 2:52 PM, Steve Evans <jeeping@gmail.com> wrote:
>
> Hi all..
>
> I had one of the disks in my 3 disk RAID5 die on me this week. When
> attempting to replace the disk via a hot swap (USB), the RAID didn't
> like it. It decided to mark one of my remaining 2 disks as faulty.
>
> Can someone *please* help me get the raid back!?
>
> More details -
>
> Drives are /dev/sdb1, /dev/sdc1 & /dev/sdd1
>
> sdc1 was the one that died earlier this week
> sdb1 appears to be the one that was marked as faulty
>
> mdadm detail before sdc1 was plugged in -
>
> root@imp[~]:11 # mdadm --detail /dev/md1
> /dev/md1:
> Version : 00.90.01
> Creation Time : Fri Nov 17 21:28:44 2006
> Raid Level : raid5
> Array Size : 586067072 (558.92 GiB 600.13 GB)
> Device Size : 293033536 (279.46 GiB 300.07 GB)
> Raid Devices : 3
> Total Devices : 2
> Preferred Minor : 1
> Persistence : Superblock is persistent
>
> Update Time : Sat Oct 18 20:06:34 2008
> State : clean, degraded
> Active Devices : 2
> Working Devices : 2
> Failed Devices : 0
> Spare Devices : 0
>
> Layout : left-symmetric
> Chunk Size : 64K
>
> UUID : bed40ee2:98523fdd:e4d010fb:894c0966
> Events : 0.1474312
>
> Number Major Minor RaidDevice State
> 0 8 17 0 active sync /dev/sdb1
> 1 0 0 - removed
> 2 8 49 2 active sync /dev/sdd1
>
>
> then after plugging in the replacement sdc1 -
>
> root@imp[~]:13 # mdadm --add /dev/md1 /dev/sdc1
> mdadm: hot added /dev/sdc1
> root@imp[~]:14 #
> root@imp[~]:14 #
> root@imp[~]:14 # mdadm --detail /dev/md1
> /dev/md1:
> Version : 00.90.01
> Creation Time : Fri Nov 17 21:28:44 2006
> Raid Level : raid5
> Array Size : 586067072 (558.92 GiB 600.13 GB)
> Device Size : 293033536 (279.46 GiB 300.07 GB)
> Raid Devices : 3
> Total Devices : 3
> Preferred Minor : 1
> Persistence : Superblock is persistent
>
> Update Time : Sat Oct 18 22:13:13 2008
> State : clean, degraded
> Active Devices : 1
> Working Devices : 2
> Failed Devices : 1
> Spare Devices : 1
>
> Layout : left-symmetric
> Chunk Size : 64K
>
> UUID : bed40ee2:98523fdd:e4d010fb:894c0966
> Events : 0.1480366
>
> Number Major Minor RaidDevice State
> 0 0 0 - removed
> 1 0 0 - removed
> 2 8 49 2 active sync /dev/sdd1
>
> 3 8 33 0 spare rebuilding /dev/sdc1
> 4 8 17 - faulty /dev/sdb1
>
> Shortly after this, subsequent mdadm --details stopped responding.. So
> I rebooted in the hope I could reset and problems with the hot add..
>
> Now, I'm unable to assemble the raid with the 2 working drives -
>
> mdadm --assemble /dev/md1 /dev/sdb1 /dev/sdd1
>
> doesn't work -
>
> mdadm: /dev/md1 assembled from 1 drive and 1 spare - not enough to
> start the array.
>
> mdadm --assemble --force /dev/md1 /dev/sdb1 /dev/sdd1
>
> doesn't' work either
>
> This -
>
> mdadm --assemble --force --run /dev/md1 /dev/sdb1 /dev/sdd1
>
> Did work partially -
>
> /dev/md1:
> Version : 00.90.01
> Creation Time : Fri Nov 17 21:28:44 2006
> Raid Level : raid5
> Device Size : 293033536 (279.46 GiB 300.07 GB)
> Raid Devices : 3
> Total Devices : 2
> Preferred Minor : 1
> Persistence : Superblock is persistent
>
> Update Time : Sat Oct 18 22:14:48 2008
> State : active, degraded
> Active Devices : 1
> Working Devices : 2
> Failed Devices : 0
> Spare Devices : 1
>
> Layout : left-symmetric
> Chunk Size : 64K
>
> UUID : bed40ee2:98523fdd:e4d010fb:894c0966
> Events : 0.1521614
>
> Number Major Minor RaidDevice State
> 0 0 0 - removed
> 1 0 0 - removed
> 2 8 49 2 active sync /dev/sdd1
>
> 3 8 17 - spare /dev/sdb1
>
> Here's the output from mdadm -E on each of the 2 drives -
>
> /dev/sdb1:
> Magic : a92b4efc
> Version : 00.90.00
> UUID : bed40ee2:98523fdd:e4d010fb:894c0966
> Creation Time : Fri Nov 17 21:28:44 2006
> Raid Level : raid5
> Raid Devices : 3
> Total Devices : 3
> Preferred Minor : 1
>
> Update Time : Sat Oct 18 22:14:48 2008
> State : clean
> Active Devices : 1
> Working Devices : 2
> Failed Devices : 2
> Spare Devices : 1
> Checksum : e6dbf75 - correct
> Events : 0.1521614
>
> Layout : left-symmetric
> Chunk Size : 64K
>
> Number Major Minor RaidDevice State
> this 3 8 33 3 spare /dev/sdc1
>
> 0 0 0 0 0 removed
> 1 1 0 0 1 faulty removed
> 2 2 8 49 2 active sync /dev/sdd1
> 3 3 8 33 3 spare /dev/sdc1
> /dev/sdd1:
> Magic : a92b4efc
> Version : 00.90.00
> UUID : bed40ee2:98523fdd:e4d010fb:894c0966
> Creation Time : Fri Nov 17 21:28:44 2006
> Raid Level : raid5
> Raid Devices : 3
> Total Devices : 3
> Preferred Minor : 1
>
> Update Time : Sat Oct 18 22:14:48 2008
> State : clean
> Active Devices : 1
> Working Devices : 2
> Failed Devices : 2
> Spare Devices : 1
> Checksum : e6dbf86 - correct
> Events : 0.1521614
>
> Layout : left-symmetric
> Chunk Size : 64K
>
> Number Major Minor RaidDevice State
> this 2 8 49 2 active sync /dev/sdd1
>
> 0 0 0 0 0 removed
> 1 1 0 0 1 faulty removed
> 2 2 8 49 2 active sync /dev/sdd1
> 3 3 8 33 0 spare /dev/sdc1
>
> root@imp[~]:28 # mdadm --version
> mdadm - v1.9.0 - 04 February 2005
> root@imp[~]:29 # uname -a
> Linux imp 2.6.8-3-686 #1 Tue Dec 5 21:26:38 UTC 2006 i686 GNU/Linux
>
>
> Is all the data lost, or can I recover from this?
>
> Thanks so much!
> Steve..

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: mdadm degraded RAID5 failure
  2008-10-22 20:52 ` mdadm degraded RAID5 failure Steve Evans
  2008-10-24 18:47   ` Steve Evans
@ 2008-10-25  6:30   ` Neil Brown
  2008-10-25 10:44     ` David Greaves
  2008-10-29 22:16     ` Steve Evans
  1 sibling, 2 replies; 7+ messages in thread
From: Neil Brown @ 2008-10-25  6:30 UTC (permalink / raw)
  To: Steve Evans; +Cc: linux-raid

On Wednesday October 22, jeeping@gmail.com wrote:
> Hi all..

Hi.
You need to get a mail client that doesn't destroy the formatting of
the text that you paste in.  But while it is an inconvenience, we
should be able to persevere...

> 
> I had one of the disks in my 3 disk RAID5 die on me this week. When
> attempting to replace the disk via a hot swap (USB), the RAID didn't
> like it. It decided to mark one of my remaining 2 disks as faulty.

It would be interesting to see the kernel logs at this time.  Maybe
the USB bus glitched while you were plugging the device in.


> 
> Can someone *please* help me get the raid back!?

Probably.

> 
> More details -
> 
> Drives are /dev/sdb1, /dev/sdc1 & /dev/sdd1

... or were.  USB device names can change every time you plug them in.

> 
> sdc1 was the one that died earlier this week
> sdb1 appears to be the one that was marked as faulty
> 
> mdadm detail before sdc1 was plugged in -
> 
> root@imp[~]:11 # mdadm --detail /dev/md1
> /dev/md1:
...
> 
> Number Major Minor RaidDevice State
> 0 8 17 0 active sync /dev/sdb1
> 1 0 0 - removed
> 2 8 49 2 active sync /dev/sdd1

So the array thinks the 2nd of 3 is missing.  That is consistent with
your description.

> 
> 
> then after plugging in the replacement sdc1 -
> 
> root@imp[~]:13 # mdadm --add /dev/md1 /dev/sdc1
> mdadm: hot added /dev/sdc1
> root@imp[~]:14 #
> root@imp[~]:14 #
> root@imp[~]:14 # mdadm --detail /dev/md1
> /dev/md1:
...
> 
> Number Major Minor RaidDevice State
> 0 0 0 - removed
> 1 0 0 - removed
> 2 8 49 2 active sync /dev/sdd1
> 
> 3 8 33 0 spare rebuilding /dev/sdc1
> 4 8 17 - faulty /dev/sdb1

Yes, sdb must have got an error and failed while sdc was rebuilding.
Sad.  That suggests that it didn't fail at the moment of USB
insertion, but a little later.  Not conclusively though.

> 
> Shortly after this, subsequent mdadm --details stopped responding.. So
> I rebooted in the hope I could reset and problems with the hot add..
> 
> Now, I'm unable to assemble the raid with the 2 working drives -
> 
> mdadm --assemble /dev/md1 /dev/sdb1 /dev/sdd1
> 
> doesn't work -
> 
> mdadm: /dev/md1 assembled from 1 drive and 1 spare - not enough to
> start the array.

You have rebooted so device names may have changed.
If it thought you had named a good drive and a spare, it probably saw
the device that was originally sdb (and possibly still is)
and the device that was originally sdc (and now might be sdd).

> 
> mdadm --assemble --force /dev/md1 /dev/sdb1 /dev/sdd1
> 
> doesn't' work either

What error messages?  Always best to be explicit.
Adding "-v" to the --assemble line would help too.

> 
> This -
> 
> mdadm --assemble --force --run /dev/md1 /dev/sdb1 /dev/sdd1
> 
> Did work partially -
> 
Hmm.. That really shouldn't have worked.  The kernel should have
rejected the array...

> 
> Here's the output from mdadm -E on each of the 2 drives -

Uhm... There should be 3 drives?
The 'good' one, the 'new' one, and the one that seemed to fail
immediately after you plugged in the 'new' one.

> 
> /dev/sdb1:
..
> Number Major Minor RaidDevice State
> this 3 8 33 3 spare /dev/sdc1
> 
> 0 0 0 0 0 removed
> 1 1 0 0 1 faulty removed
> 2 2 8 49 2 active sync /dev/sdd1
> 3 3 8 33 3 spare /dev/sdc1

sdb looks like the new one.

> /dev/sdd1:
...
> 
> Number Major Minor RaidDevice State
> this 2 8 49 2 active sync /dev/sdd1
> 
> 0 0 0 0 0 removed
> 1 1 0 0 1 faulty removed
> 2 2 8 49 2 active sync /dev/sdd1
> 3 3 8 33 0 spare /dev/sdc1

sdd looks like the good one.

Where is the "one that seemed to fail" which was once called sdb ??
> 
> Is all the data lost, or can I recover from this?

Try

  mdadm --examine --brief --verbose /dev/sd*

That will list anything that looks like an array.
e.g. (on my devel machine)

# mdadm --examine --brief --verbose /dev/sd*
ARRAY /dev/md0 level=raid5 num-devices=3 UUID=cfd6a841:c24600be:c4297cb4:f8ef633e
   devices=/dev/sdb,/dev/sdc,/dev/sdd
ARRAY /dev/md0 level=raid5 num-devices=2 UUID=cb711aad:db89ffc8:faa4816a:59e602da
   devices=/dev/sda11,/dev/sda12

Take careful note of the "devices=" part.  That lists sets of devices
(maybe only one set in your case) which are all part of an array.
So I have two array, one across /dev/sdb, /dev/sdc, /dev/sdd and
one across /dev/sda11 and /dev/sda12.

Then

  mdadm --assemble --force --verbose /dev/md1 /dev/sd....

where you list all the devices in the device= section for the array
you want to try to start.

Report the output of that command and whether it was successful.

NeilBrown

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: mdadm degraded RAID5 failure
  2008-10-25  6:30   ` Neil Brown
@ 2008-10-25 10:44     ` David Greaves
  2008-10-29 22:16     ` Steve Evans
  1 sibling, 0 replies; 7+ messages in thread
From: David Greaves @ 2008-10-25 10:44 UTC (permalink / raw)
  To: Neil Brown; +Cc: Steve Evans, linux-raid

Neil Brown wrote:
>> mdadm --assemble --force --run /dev/md1 /dev/sdb1 /dev/sdd1
>>
>> Did work partially -
>>
> Hmm.. That really shouldn't have worked.  The kernel should have
> rejected the array...

Did you notice: 2.6.8-3-686
mdadm - v1.9.0 - 04 February 2005

I don't really recall the history of md and mdadm but I'd suggest trying to
repair in a modern recovery/live-CD environment.

David


-- 
"Don't worry, you'll be fine; I saw it work in a cartoon once..."

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: mdadm degraded RAID5 failure
  2008-10-25  6:30   ` Neil Brown
  2008-10-25 10:44     ` David Greaves
@ 2008-10-29 22:16     ` Steve Evans
  2008-11-04 21:35       ` Steve Evans
  1 sibling, 1 reply; 7+ messages in thread
From: Steve Evans @ 2008-10-29 22:16 UTC (permalink / raw)
  To: Neil Brown; +Cc: linux-raid

On Sat, Oct 25, 2008 at 12:30 AM, Neil Brown <neilb@suse.de> wrote:
> On Wednesday October 22, jeeping@gmail.com wrote:
>> Hi all..
>
> Hi.
> You need to get a mail client that doesn't destroy the formatting of
> the text that you paste in.  But while it is an inconvenience, we
> should be able to persevere...
>

Sorry, I attempted a plain text email through gmail.. I probably
messed it up :(  Hopefully this one is better..

>>
>> I had one of the disks in my 3 disk RAID5 die on me this week. When
>> attempting to replace the disk via a hot swap (USB), the RAID didn't
>> like it. It decided to mark one of my remaining 2 disks as faulty.
>
> It would be interesting to see the kernel logs at this time.  Maybe
> the USB bus glitched while you were plugging the device in.
>

Here are some of what I thought were the more relevent entries in the
logs, let me know if you'd like all of them and I can email them
directly to you as attachments -

Oct 18 20:40:27 sjev kernel: usb 4-3.2: USB disconnect, address 4
Oct 18 20:40:27 sjev kernel: usb 4-3.2: new high speed USB device
using address 12
Oct 18 20:40:27 sjev kernel: scsi8 : SCSI emulation for USB Mass Storage devices
Oct 18 20:40:28 sjev kernel:   Vendor: ST330063  Model: 1A
   Rev: 0000
Oct 18 20:40:28 sjev kernel:   Type:   Direct-Access
   ANSI SCSI revision: 02
Oct 18 20:40:28 sjev kernel: SCSI device sdc: 586072368 512-byte hdwr
sectors (300069 MB)
Oct 18 20:40:28 sjev kernel: sdc: assuming drive cache: write through
Oct 18 20:40:28 sjev kernel:  /dev/scsi/host8/bus0/target0/lun0: p1
Oct 18 20:40:28 sjev kernel: Attached scsi disk sdc at scsi8, channel
0, id 0, lun 0
Oct 18 20:40:28 sjev kernel: Attached scsi generic sg1 at scsi8,
channel 0, id 0, lun 0,  type 0
Oct 18 20:40:28 sjev kernel: USB Mass Storage device found at 12
Oct 18 20:40:28 sjev usb.agent[8548]:      usb-storage: already loaded
Oct 18 20:40:29 sjev scsi.agent[8571]:      sd_mod: loaded sucessfully
(for disk)
Oct 18 20:40:29 sjev kernel: scsi1 (0:0): rejecting I/O to dead device
Oct 18 20:40:29 sjev kernel: md: write_disk_sb failed for device sdb1
Oct 18 20:40:29 sjev kernel: md: errors occurred during superblock
update, repeating
Oct 18 20:40:29 sjev kernel: scsi1 (0:0): rejecting I/O to dead device
Oct 18 20:40:29 sjev kernel: md: write_disk_sb failed for device sdb1
Oct 18 20:40:29 sjev kernel: md: errors occurred during superblock
update, repeating
Oct 18 20:40:29 sjev kernel: scsi1 (0:0): rejecting I/O to dead device
Oct 18 20:40:29 sjev kernel: md: write_disk_sb failed for device sdb1
Oct 18 20:40:29 sjev kernel: md: errors occurred during superblock
update, repeating
Oct 18 20:40:29 sjev kernel: scsi1 (0:0): rejecting I/O to dead device

etc..

Oct 18 20:40:34 sjev kernel: md: errors occurred during superblock
update, repeating
Oct 18 20:40:34 sjev kernel: scsi1 (0:0): rejecting I/O to dead device
Oct 18 20:40:34 sjev kernel: md: write_disk_sb failed for device sdb1
Oct 18 20:40:34 sjev kernel: md: errors occurred during superblock
update, repeating
Oct 18 20:40:34 sjev kernel: scsi1 (0:0): rejecting I/O to dead device
Oct 18 20:40:34 sjev kernel: md: write_disk_sb failed for device sdb1
Oct 18 20:40:34 sjev kernel: md: excessive errors occurred during
superblock update, exiting
Oct 18 20:40:34 sjev kernel: scsi1 (0:0): rejecting I/O to dead device
Oct 18 20:40:34 sjev kernel: raid5: Disk failure on sdb1, disabling
device. Operation continuing on 0 devices
Oct 18 20:40:34 sjev kernel: RAID5 conf printout:
Oct 18 20:40:34 sjev kernel:  --- rd:3 wd:0 fd:2
Oct 18 20:40:34 sjev kernel:  disk 0, o:0, dev:sdb1
Oct 18 20:40:34 sjev kernel:  disk 2, o:1, dev:sdd1
Oct 18 20:40:34 sjev kernel: RAID5 conf printout:
Oct 18 20:40:34 sjev kernel:  --- rd:3 wd:0 fd:2
Oct 18 20:40:34 sjev kernel:  disk 2, o:1, dev:sdd1
Oct 18 20:40:34 sjev kernel: Buffer I/O error on device md1, logical block 3601
Oct 18 20:40:34 sjev kernel: lost page write due to I/O error on md1
Oct 18 20:40:34 sjev kernel: Aborting journal on device md1.
Oct 18 20:40:35 sjev kernel: ext3_abort called.
Oct 18 20:40:35 sjev kernel: EXT3-fs abort (device md1):
ext3_journal_start: Detected aborted journal
Oct 18 20:40:35 sjev kernel: Remounting filesystem read-only
Oct 18 20:40:38 sjev kernel: Buffer I/O error on device md1, logical
block 103252006
Oct 18 20:40:38 sjev kernel: lost page write due to I/O error on md1
Oct 18 20:40:38 sjev kernel: Buffer I/O error on device md1, logical
block 103252007
Oct 18 20:40:38 sjev kernel: lost page write due to I/O error on md1
Oct 18 20:40:38 sjev kernel: Buffer I/O error on device md1, logical
block 103252008
Oct 18 20:40:38 sjev kernel: lost page write due to I/O error on md1
Oct 18 20:40:38 sjev kernel: Buffer I/O error on device md1, logical
block 103252009
Oct 18 20:40:38 sjev kernel: lost page write due to I/O error on md1
Oct 18 20:40:38 sjev kernel: Buffer I/O error on device md1, logical
block 103252010
Oct 18 20:40:38 sjev kernel: lost page write due to I/O error on md1
Oct 18 20:40:38 sjev kernel: Buffer I/O error on device md1, logical
block 103252011
Oct 18 20:40:38 sjev kernel: lost page write due to I/O error on md1
Oct 18 20:40:38 sjev kernel: Buffer I/O error on device md1, logical
block 103252012
Oct 18 20:40:38 sjev kernel: lost page write due to I/O error on md1
Oct 18 20:40:38 sjev kernel: Buffer I/O error on device md1, logical
block 103252013
Oct 18 20:40:38 sjev kernel: lost page write due to I/O error on md1
Oct 18 20:40:38 sjev kernel: Buffer I/O error on device md1, logical
block 103252014
Oct 18 20:40:38 sjev kernel: lost page write due to I/O error on md1
Oct 18 20:40:52 sjev kernel: printk: 35 messages suppressed.


later ..

Oct 18 22:12:39 sjev kernel: usb 4-3.3: new high speed USB device
using address 13
Oct 18 22:12:40 sjev usb.agent[21323]:      usb-storage: already loaded
Oct 18 22:12:40 sjev kernel: scsi9 : SCSI emulation for USB Mass Storage devices
Oct 18 22:12:40 sjev kernel:   Vendor: MAXTOR S  Model: TM3320620A
   Rev: 0000
Oct 18 22:12:40 sjev kernel:   Type:   Direct-Access
   ANSI SCSI revision: 02
Oct 18 22:12:40 sjev kernel: SCSI device sde: 625142448 512-byte hdwr
sectors (320073 MB)
Oct 18 22:12:40 sjev kernel: sde: assuming drive cache: write through
Oct 18 22:12:40 sjev kernel:  /dev/scsi/host9/bus0/target0/lun0: p1
Oct 18 22:12:40 sjev kernel: Attached scsi disk sde at scsi9, channel
0, id 0, lun 0
Oct 18 22:12:40 sjev kernel: Attached scsi generic sg2 at scsi9,
channel 0, id 0, lun 0,  type 0
Oct 18 22:12:40 sjev kernel: USB Mass Storage device found at 13
Oct 18 22:12:41 sjev scsi.agent[21357]:      sd_mod: loaded
sucessfully (for disk)
Oct 18 22:13:00 sjev kernel: md: trying to hot-add unknown-block(8,33)
to md1 ...
Oct 18 22:13:00 sjev kernel: md: bind<sdc1>
Oct 18 22:13:00 sjev kernel: RAID5 conf printout:
Oct 18 22:13:00 sjev kernel:  --- rd:3 wd:0 fd:2
Oct 18 22:13:00 sjev kernel:  disk 0, o:1, dev:sdc1
Oct 18 22:13:00 sjev kernel:  disk 2, o:1, dev:sdd1
Oct 18 22:13:00 sjev kernel: md: syncing RAID array md1
Oct 18 22:13:00 sjev kernel: md: minimum _guaranteed_ reconstruction
speed: 1000 KB/sec/disc.
Oct 18 22:13:00 sjev kernel: md: using maximum available idle IO
bandwith (but not more than 200000 KB/sec) for reconstruction.
Oct 18 22:13:00 sjev kernel: md: using 128k window, over a total of
293033536 blocks.
Oct 18 22:13:00 sjev kernel: md: md1: sync done.
Oct 18 22:13:00 sjev kernel: md: syncing RAID array md1
Oct 18 22:13:00 sjev kernel: md: minimum _guaranteed_ reconstruction
speed: 1000 KB/sec/disc.
Oct 18 22:13:00 sjev kernel: md: using maximum available idle IO
bandwith (but not more than 200000 KB/sec) for reconstruction.
Oct 18 22:13:00 sjev kernel: md: using 128k window, over a total of
293033536 blocks.
Oct 18 22:13:00 sjev kernel: md: md1: sync done.
Oct 18 22:13:01 sjev kernel: md: syncing RAID array md1

repeats until..

Oct 18 22:14:48 sjev kernel: md: syncing RAID array md1
Oct 18 22:14:48 sjev kernel: md: minimum _guaranteed_ reconstruction
speed: 1000 KB/sec/disc.
Oct 18 22:14:48 sjev kernel: md: using maximum available idle IO
bandwith (but not more than 200000 KB/sec) for reconstruction.
Oct 18 22:14:48 sjev kernel: md: using 128k window, over a total of
293033536 blocks.
Oct 18 22:14:48 sjev kernel: md: md1: sync done.
Oct 18 22:14:48 sjev kernel: Unable to handle kernel NULL pointer
dereference at virtual address 000000a4
Oct 18 22:14:48 sjev kernel:  printing eip:
Oct 18 22:14:48 sjev kernel: c0124d89
Oct 18 22:14:48 sjev kernel: *pde = 00000000
Oct 18 22:14:48 sjev kernel: Oops: 0000 [#1]
Oct 18 22:14:48 sjev kernel: PREEMPT
Oct 18 22:14:48 sjev kernel: Modules linked in: ipv6 smbfs
snd_intel8x0m snd_intel8x0 snd_ac97_codec snd_pcm snd_timer
snd_page_alloc gameport snd_mpu401_uart snd_rawmidi snd_seq_device snd
capability commoncap raid5 xor sr_mod tsdev mousedev joydev evdev
pcspkr pci_hotplug intel_agp agpgart ide_scsi ide_generic sg font
vesafb cfbcopyarea cfbimgblt cfbfillrect appletalk af_packet hw_random
i810_audio soundcore ac97_codec b44 mii yenta_socket rtc piix unix ds
pcmcia_core usb_storage ext3 mbcache raid1 md jbd ehci_hcd ohci_hcd
uhci_hcd usbcore reiserfs psmouse ide_disk ide_cd ide_core cdrom
sd_mod scsi_mod
Oct 18 22:14:48 sjev kernel: CPU:    0
Oct 18 22:14:48 sjev kernel: EIP:    0060:[sig_ignored+73/112]    Not tainted
Oct 18 22:14:48 sjev kernel: EFLAGS: 00010006   (2.6.8-3-686)
Oct 18 22:14:48 sjev kernel: EIP is at sig_ignored+0x49/0x70
Oct 18 22:14:48 sjev kernel: eax: 000000b4   ebx: 00000000   ecx:
00000008   edx: 00000000
Oct 18 22:14:48 sjev kernel: esi: 00000009   edi: 00000009   ebp:
00000000   esp: cedf3ec0
Oct 18 22:14:48 sjev kernel: ds: 007b   es: 007b   ss: 0068
Oct 18 22:14:48 sjev kernel: Process md1_raid5 (pid: 685,
threadinfo=cedf2000 task=cedef3e0)
Oct 18 22:14:48 sjev kernel: Stack: cf10e1b0 00000001 c01259f3
cf10e1b0 00000009 c86194a0 cf99771c 00000202
Oct 18 22:14:48 sjev kernel:        cedf2000 cf997680 cf222c00
c0126565 00000009 00000001 cf10e1b0 c86194a0
Oct 18 22:14:48 sjev kernel:        cedf3f30 cf997680 d093eb7d
00000009 00000001 cf10e1b0 d093ebcd c86194a0
Oct 18 22:14:48 sjev kernel: Call Trace:
Oct 18 22:14:48 sjev kernel:  [specific_send_sig_info+83/224]
specific_send_sig_info+0x53/0xe0
Oct 18 22:14:48 sjev kernel:  [send_sig_info+69/128] send_sig_info+0x45/0x80
Oct 18 22:14:48 sjev kernel:  [__crc_sb_min_blocksize+815035/1015327]
md_interrupt_thread+0x4d/0x60 [md]
Oct 18 22:14:48 sjev kernel:  [__crc_sb_min_blocksize+815115/1015327]
md_unregister_thread+0x3d/0x60 [md]
Oct 18 22:14:48 sjev kernel:  [recalc_task_prio+168/416]
recalc_task_prio+0xa8/0x1a0
Oct 18 22:14:48 sjev kernel:  [__crc_sb_min_blocksize+821862/1015327]
md_check_recovery+0x288/0x300 [md]
Oct 18 22:14:48 sjev kernel:  [__crc_fb_pan_display+1312520/2923165]
raid5d+0x19/0x150 [raid5]
Oct 18 22:14:48 sjev kernel:  [__crc_sb_min_blocksize+814642/1015327]
md_thread+0x164/0x1d0 [md]
Oct 18 22:14:48 sjev kernel:  [autoremove_wake_function+0/96]
autoremove_wake_function+0x0/0x60
Oct 18 22:14:48 sjev kernel:  [ret_from_fork+6/20] ret_from_fork+0x6/0x14
Oct 18 22:14:48 sjev kernel:  [autoremove_wake_function+0/96]
autoremove_wake_function+0x0/0x60
Oct 18 22:14:48 sjev kernel:  [__crc_sb_min_blocksize+814286/1015327]
md_thread+0x0/0x1d0 [md]
Oct 18 22:14:48 sjev kernel:  [kernel_thread_helper+5/24]
kernel_thread_helper+0x5/0x18
Oct 18 22:14:48 sjev kernel: Code: 8b 40 f0 83 f8 01 74 18 85 c0 74 04
89 d3 eb c1 83 fe 1f 7f
Oct 18 22:14:48 sjev kernel:  <6>note: md1_raid5[685] exited with
preempt_count 2


>
>>
>> Can someone *please* help me get the raid back!?
>
> Probably.
>

I like the optimism! Thanks!

>>
>> More details -
>>
>> Drives are /dev/sdb1, /dev/sdc1 & /dev/sdd1
>
> ... or were.  USB device names can change every time you plug them in.
>
>>
>> sdc1 was the one that died earlier this week
>> sdb1 appears to be the one that was marked as faulty
>>
>> mdadm detail before sdc1 was plugged in -
>>
>> root@imp[~]:11 # mdadm --detail /dev/md1
>> /dev/md1:
> ...
>>
>> Number Major Minor RaidDevice State
>> 0 8 17 0 active sync /dev/sdb1
>> 1 0 0 - removed
>> 2 8 49 2 active sync /dev/sdd1
>
> So the array thinks the 2nd of 3 is missing.  That is consistent with
> your description.
>
>>
>>
>> then after plugging in the replacement sdc1 -
>>
>> root@imp[~]:13 # mdadm --add /dev/md1 /dev/sdc1
>> mdadm: hot added /dev/sdc1
>> root@imp[~]:14 #
>> root@imp[~]:14 #
>> root@imp[~]:14 # mdadm --detail /dev/md1
>> /dev/md1:
> ...
>>
>> Number Major Minor RaidDevice State
>> 0 0 0 - removed
>> 1 0 0 - removed
>> 2 8 49 2 active sync /dev/sdd1
>>
>> 3 8 33 0 spare rebuilding /dev/sdc1
>> 4 8 17 - faulty /dev/sdb1
>
> Yes, sdb must have got an error and failed while sdc was rebuilding.
> Sad.  That suggests that it didn't fail at the moment of USB
> insertion, but a little later.  Not conclusively though.
>
>>
>> Shortly after this, subsequent mdadm --details stopped responding.. So
>> I rebooted in the hope I could reset and problems with the hot add..
>>
>> Now, I'm unable to assemble the raid with the 2 working drives -
>>
>> mdadm --assemble /dev/md1 /dev/sdb1 /dev/sdd1
>>
>> doesn't work -
>>
>> mdadm: /dev/md1 assembled from 1 drive and 1 spare - not enough to
>> start the array.
>
> You have rebooted so device names may have changed.
> If it thought you had named a good drive and a spare, it probably saw
> the device that was originally sdb (and possibly still is)
> and the device that was originally sdc (and now might be sdd).
>
>>
>> mdadm --assemble --force /dev/md1 /dev/sdb1 /dev/sdd1
>>
>> doesn't' work either
>
> What error messages?  Always best to be explicit.
> Adding "-v" to the --assemble line would help too.
>
>>
>> This -
>>
>> mdadm --assemble --force --run /dev/md1 /dev/sdb1 /dev/sdd1
>>
>> Did work partially -
>>
> Hmm.. That really shouldn't have worked.  The kernel should have
> rejected the array...
>
>>
>> Here's the output from mdadm -E on each of the 2 drives -
>
> Uhm... There should be 3 drives?
> The 'good' one, the 'new' one, and the one that seemed to fail
> immediately after you plugged in the 'new' one.
>

Sorry, here are all 3 -

root@imp[~]:3 # mdadm -E /dev/sd[bcd]1
/dev/sdb1:
          Magic : a92b4efc
        Version : 00.90.00
           UUID : bed40ee2:98523fdd:e4d010fb:894c0966
  Creation Time : Fri Nov 17 21:28:44 2006
     Raid Level : raid5
   Raid Devices : 3
  Total Devices : 3
Preferred Minor : 1

    Update Time : Sat Oct 18 22:14:48 2008
          State : clean
 Active Devices : 1
Working Devices : 2
 Failed Devices : 2
  Spare Devices : 1
       Checksum : e6dbf86 - correct
         Events : 0.1521614

         Layout : left-symmetric
     Chunk Size : 64K

      Number   Major   Minor   RaidDevice State
this     2       8       49        2      active sync   /dev/sdd1

   0     0       0        0        0      removed
   1     1       0        0        1      faulty removed
   2     2       8       49        2      active sync   /dev/sdd1
   3     3       8       33        0      spare   /dev/sdc1
/dev/sdc1:
          Magic : a92b4efc
        Version : 00.90.00
           UUID : bed40ee2:98523fdd:e4d010fb:894c0966
  Creation Time : Fri Nov 17 21:28:44 2006
     Raid Level : raid5
   Raid Devices : 3
  Total Devices : 3
Preferred Minor : 1

    Update Time : Fri Oct 17 22:30:49 2008
          State : clean
 Active Devices : 2
Working Devices : 3
 Failed Devices : 1
  Spare Devices : 1
       Checksum : e6ae9ea - correct
         Events : 0.1471469

         Layout : left-symmetric
     Chunk Size : 64K

      Number   Major   Minor   RaidDevice State
this     3       8       33        3      spare   /dev/sdc1

   0     0       8       17        0      active sync   /dev/sdb1
   1     1       0        0        1      faulty removed
   2     2       8       49        2      active sync   /dev/sdd1
   3     3       8       33        3      spare   /dev/sdc1
/dev/sdd1:
          Magic : a92b4efc
        Version : 00.90.00
           UUID : bed40ee2:98523fdd:e4d010fb:894c0966
  Creation Time : Fri Nov 17 21:28:44 2006
     Raid Level : raid5
   Raid Devices : 3
  Total Devices : 3
Preferred Minor : 1

    Update Time : Sat Oct 18 22:14:48 2008
          State : clean
 Active Devices : 1
Working Devices : 2
 Failed Devices : 2
  Spare Devices : 1
       Checksum : e6dbf75 - correct
         Events : 0.1521614

         Layout : left-symmetric
     Chunk Size : 64K

      Number   Major   Minor   RaidDevice State
this     3       8       33        3      spare   /dev/sdc1

   0     0       0        0        0      removed
   1     1       0        0        1      faulty removed
   2     2       8       49        2      active sync   /dev/sdd1
   3     3       8       33        3      spare   /dev/sdc1

fdisk details too -

root@imp[~]:7 # fdisk -l /dev/sd[bcd]

Disk /dev/sdb: 300.0 GB, 300069052416 bytes
255 heads, 63 sectors/track, 36481 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

   Device Boot      Start         End      Blocks   Id  System
/dev/sdb1               1       36481   293033601   fd  Linux raid autodetect

Disk /dev/sdc: 320.0 GB, 320072933376 bytes
255 heads, 63 sectors/track, 38913 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

   Device Boot      Start         End      Blocks   Id  System
/dev/sdc1               1       36481   293033601   fd  Linux raid autodetect

Disk /dev/sdd: 300.0 GB, 300069052416 bytes
255 heads, 63 sectors/track, 36481 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes

   Device Boot      Start         End      Blocks   Id  System
/dev/sdd1               1       36481   293033601   fd  Linux raid autodetect


>>
>> /dev/sdb1:
> ..
>> Number Major Minor RaidDevice State
>> this 3 8 33 3 spare /dev/sdc1
>>
>> 0 0 0 0 0 removed
>> 1 1 0 0 1 faulty removed
>> 2 2 8 49 2 active sync /dev/sdd1
>> 3 3 8 33 3 spare /dev/sdc1
>
> sdb looks like the new one.
>
>> /dev/sdd1:
> ...
>>
>> Number Major Minor RaidDevice State
>> this 2 8 49 2 active sync /dev/sdd1
>>
>> 0 0 0 0 0 removed
>> 1 1 0 0 1 faulty removed
>> 2 2 8 49 2 active sync /dev/sdd1
>> 3 3 8 33 0 spare /dev/sdc1
>
> sdd looks like the good one.
>
> Where is the "one that seemed to fail" which was once called sdb ??
>>
>> Is all the data lost, or can I recover from this?
>
> Try
>
>  mdadm --examine --brief --verbose /dev/sd*
>

ARRAY /dev/md1 level=raid5 num-devices=3
UUID=bed40ee2:98523fdd:e4d010fb:894c0966
   devices=/dev/sdb1,/dev/sdc1,/dev/sdd1
ARRAY /dev/md4 level=raid1 num-devices=2
UUID=6fded12b:6ecdca8a:18400b9a:df6a2ffc
   devices=/dev/sda5
ARRAY /dev/md0 level=raid1 num-devices=2
UUID=c94d0631:20f0db42:9c6ab972:19acc617
   devices=/dev/sda1

>
> Then
>
>  mdadm --assemble --force --verbose /dev/md1 /dev/sd....
>
> where you list all the devices in the device= section for the array
> you want to try to start.
>
> Report the output of that command and whether it was successful.

root@imp[~]:9 # mdadm --assemble --force --verbose /dev/md1 /dev/sdb1
/dev/sdc1 /dev/sdd1
mdadm: looking for devices for /dev/md1
mdadm: /dev/sdb1 is identified as a member of /dev/md1, slot 2.
mdadm: /dev/sdc1 is identified as a member of /dev/md1, slot 3.
mdadm: /dev/sdd1 is identified as a member of /dev/md1, slot 3.
mdadm: no uptodate device for slot 0 of /dev/md1
mdadm: no uptodate device for slot 1 of /dev/md1
mdadm: added /dev/sdd1 to /dev/md1 as 3
mdadm: added /dev/sdb1 to /dev/md1 as 2
mdadm: /dev/md1 assembled from 1 drive and 1 spare - not enough to
start the array.
root@imp[~]:10 #

Oct 29 14:52:41 sjev kernel: md: md1 stopped.
Oct 29 14:52:41 sjev kernel: md: unbind<sdb1>
Oct 29 14:52:41 sjev kernel: md: export_rdev(sdb1)
Oct 29 14:52:41 sjev kernel: md: unbind<sdd1>
Oct 29 14:52:41 sjev kernel: md: export_rdev(sdd1)
Oct 29 14:52:41 sjev kernel: md: bind<sdd1>
Oct 29 14:52:41 sjev kernel: md: bind<sdb1>
Oct 29 14:58:07 sjev smartd[2302]: Device: /dev/hdc, SMART Usage
Attribute: 190 Unknown_Attribute changed from 49 to 48
Oct 29 14:58:07 sjev smartd[2302]: Device: /dev/hdc, SMART Usage
Attribute: 194 Temperature_Celsius changed from 51 to 52

I've held off upgrading mdadm to the latest version until I know it's
the best option (vs recovering the raid 1st before upgrading), so you
agree?

>
> NeilBrown
>

Thanks for your patience and help!
Regards,
Steve..

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: mdadm degraded RAID5 failure
  2008-10-29 22:16     ` Steve Evans
@ 2008-11-04 21:35       ` Steve Evans
  2008-11-06  5:41         ` Neil Brown
  0 siblings, 1 reply; 7+ messages in thread
From: Steve Evans @ 2008-11-04 21:35 UTC (permalink / raw)
  To: Neil Brown; +Cc: linux-raid

Hi Neil and others,

Just a couple of questions, I know you're busy -

Do you recommend that I attempt to upgrade mdadm to a more recent
version before any other recovery attempts? If so, which version?

I noted my replacement drive (sdc1) got a smart error (during the
rebuild?), would you recommend replacing it or removing it altogether
until I get the other 2 drives back online (if I even can)?

Is there a way to correct the drive names -

> /dev/sdb1:
> this     2       8       49        2      active sync   /dev/sdd1


> /dev/sdc1:
> this     3       8       33        3      spare   /dev/sdc1


> /dev/sdd1:
> this     3       8       33        3      spare   /dev/sdc1

I'm inclined to believe (but am not sure at all) that -

sdb1 should be sdd1
sdc1 is correct
sdd1 should be sdb1

Thanks!
Steve..

On Wed, Oct 29, 2008 at 3:16 PM, Steve Evans <jeeping@gmail.com> wrote:
> On Sat, Oct 25, 2008 at 12:30 AM, Neil Brown <neilb@suse.de> wrote:
>> On Wednesday October 22, jeeping@gmail.com wrote:
>>> Hi all..
>>
>> Hi.
>> You need to get a mail client that doesn't destroy the formatting of
>> the text that you paste in.  But while it is an inconvenience, we
>> should be able to persevere...
>>
>
> Sorry, I attempted a plain text email through gmail.. I probably
> messed it up :(  Hopefully this one is better..
>
>>>
>>> I had one of the disks in my 3 disk RAID5 die on me this week. When
>>> attempting to replace the disk via a hot swap (USB), the RAID didn't
>>> like it. It decided to mark one of my remaining 2 disks as faulty.
>>
>> It would be interesting to see the kernel logs at this time.  Maybe
>> the USB bus glitched while you were plugging the device in.
>>
>
> Here are some of what I thought were the more relevent entries in the
> logs, let me know if you'd like all of them and I can email them
> directly to you as attachments -
>
> Oct 18 20:40:27 sjev kernel: usb 4-3.2: USB disconnect, address 4
> Oct 18 20:40:27 sjev kernel: usb 4-3.2: new high speed USB device
> using address 12
> Oct 18 20:40:27 sjev kernel: scsi8 : SCSI emulation for USB Mass Storage devices
> Oct 18 20:40:28 sjev kernel:   Vendor: ST330063  Model: 1A
>   Rev: 0000
> Oct 18 20:40:28 sjev kernel:   Type:   Direct-Access
>   ANSI SCSI revision: 02
> Oct 18 20:40:28 sjev kernel: SCSI device sdc: 586072368 512-byte hdwr
> sectors (300069 MB)
> Oct 18 20:40:28 sjev kernel: sdc: assuming drive cache: write through
> Oct 18 20:40:28 sjev kernel:  /dev/scsi/host8/bus0/target0/lun0: p1
> Oct 18 20:40:28 sjev kernel: Attached scsi disk sdc at scsi8, channel
> 0, id 0, lun 0
> Oct 18 20:40:28 sjev kernel: Attached scsi generic sg1 at scsi8,
> channel 0, id 0, lun 0,  type 0
> Oct 18 20:40:28 sjev kernel: USB Mass Storage device found at 12
> Oct 18 20:40:28 sjev usb.agent[8548]:      usb-storage: already loaded
> Oct 18 20:40:29 sjev scsi.agent[8571]:      sd_mod: loaded sucessfully
> (for disk)
> Oct 18 20:40:29 sjev kernel: scsi1 (0:0): rejecting I/O to dead device
> Oct 18 20:40:29 sjev kernel: md: write_disk_sb failed for device sdb1
> Oct 18 20:40:29 sjev kernel: md: errors occurred during superblock
> update, repeating
> Oct 18 20:40:29 sjev kernel: scsi1 (0:0): rejecting I/O to dead device
> Oct 18 20:40:29 sjev kernel: md: write_disk_sb failed for device sdb1
> Oct 18 20:40:29 sjev kernel: md: errors occurred during superblock
> update, repeating
> Oct 18 20:40:29 sjev kernel: scsi1 (0:0): rejecting I/O to dead device
> Oct 18 20:40:29 sjev kernel: md: write_disk_sb failed for device sdb1
> Oct 18 20:40:29 sjev kernel: md: errors occurred during superblock
> update, repeating
> Oct 18 20:40:29 sjev kernel: scsi1 (0:0): rejecting I/O to dead device
>
> etc..
>
> Oct 18 20:40:34 sjev kernel: md: errors occurred during superblock
> update, repeating
> Oct 18 20:40:34 sjev kernel: scsi1 (0:0): rejecting I/O to dead device
> Oct 18 20:40:34 sjev kernel: md: write_disk_sb failed for device sdb1
> Oct 18 20:40:34 sjev kernel: md: errors occurred during superblock
> update, repeating
> Oct 18 20:40:34 sjev kernel: scsi1 (0:0): rejecting I/O to dead device
> Oct 18 20:40:34 sjev kernel: md: write_disk_sb failed for device sdb1
> Oct 18 20:40:34 sjev kernel: md: excessive errors occurred during
> superblock update, exiting
> Oct 18 20:40:34 sjev kernel: scsi1 (0:0): rejecting I/O to dead device
> Oct 18 20:40:34 sjev kernel: raid5: Disk failure on sdb1, disabling
> device. Operation continuing on 0 devices
> Oct 18 20:40:34 sjev kernel: RAID5 conf printout:
> Oct 18 20:40:34 sjev kernel:  --- rd:3 wd:0 fd:2
> Oct 18 20:40:34 sjev kernel:  disk 0, o:0, dev:sdb1
> Oct 18 20:40:34 sjev kernel:  disk 2, o:1, dev:sdd1
> Oct 18 20:40:34 sjev kernel: RAID5 conf printout:
> Oct 18 20:40:34 sjev kernel:  --- rd:3 wd:0 fd:2
> Oct 18 20:40:34 sjev kernel:  disk 2, o:1, dev:sdd1
> Oct 18 20:40:34 sjev kernel: Buffer I/O error on device md1, logical block 3601
> Oct 18 20:40:34 sjev kernel: lost page write due to I/O error on md1
> Oct 18 20:40:34 sjev kernel: Aborting journal on device md1.
> Oct 18 20:40:35 sjev kernel: ext3_abort called.
> Oct 18 20:40:35 sjev kernel: EXT3-fs abort (device md1):
> ext3_journal_start: Detected aborted journal
> Oct 18 20:40:35 sjev kernel: Remounting filesystem read-only
> Oct 18 20:40:38 sjev kernel: Buffer I/O error on device md1, logical
> block 103252006
> Oct 18 20:40:38 sjev kernel: lost page write due to I/O error on md1
> Oct 18 20:40:38 sjev kernel: Buffer I/O error on device md1, logical
> block 103252007
> Oct 18 20:40:38 sjev kernel: lost page write due to I/O error on md1
> Oct 18 20:40:38 sjev kernel: Buffer I/O error on device md1, logical
> block 103252008
> Oct 18 20:40:38 sjev kernel: lost page write due to I/O error on md1
> Oct 18 20:40:38 sjev kernel: Buffer I/O error on device md1, logical
> block 103252009
> Oct 18 20:40:38 sjev kernel: lost page write due to I/O error on md1
> Oct 18 20:40:38 sjev kernel: Buffer I/O error on device md1, logical
> block 103252010
> Oct 18 20:40:38 sjev kernel: lost page write due to I/O error on md1
> Oct 18 20:40:38 sjev kernel: Buffer I/O error on device md1, logical
> block 103252011
> Oct 18 20:40:38 sjev kernel: lost page write due to I/O error on md1
> Oct 18 20:40:38 sjev kernel: Buffer I/O error on device md1, logical
> block 103252012
> Oct 18 20:40:38 sjev kernel: lost page write due to I/O error on md1
> Oct 18 20:40:38 sjev kernel: Buffer I/O error on device md1, logical
> block 103252013
> Oct 18 20:40:38 sjev kernel: lost page write due to I/O error on md1
> Oct 18 20:40:38 sjev kernel: Buffer I/O error on device md1, logical
> block 103252014
> Oct 18 20:40:38 sjev kernel: lost page write due to I/O error on md1
> Oct 18 20:40:52 sjev kernel: printk: 35 messages suppressed.
>
>
> later ..
>
> Oct 18 22:12:39 sjev kernel: usb 4-3.3: new high speed USB device
> using address 13
> Oct 18 22:12:40 sjev usb.agent[21323]:      usb-storage: already loaded
> Oct 18 22:12:40 sjev kernel: scsi9 : SCSI emulation for USB Mass Storage devices
> Oct 18 22:12:40 sjev kernel:   Vendor: MAXTOR S  Model: TM3320620A
>   Rev: 0000
> Oct 18 22:12:40 sjev kernel:   Type:   Direct-Access
>   ANSI SCSI revision: 02
> Oct 18 22:12:40 sjev kernel: SCSI device sde: 625142448 512-byte hdwr
> sectors (320073 MB)
> Oct 18 22:12:40 sjev kernel: sde: assuming drive cache: write through
> Oct 18 22:12:40 sjev kernel:  /dev/scsi/host9/bus0/target0/lun0: p1
> Oct 18 22:12:40 sjev kernel: Attached scsi disk sde at scsi9, channel
> 0, id 0, lun 0
> Oct 18 22:12:40 sjev kernel: Attached scsi generic sg2 at scsi9,
> channel 0, id 0, lun 0,  type 0
> Oct 18 22:12:40 sjev kernel: USB Mass Storage device found at 13
> Oct 18 22:12:41 sjev scsi.agent[21357]:      sd_mod: loaded
> sucessfully (for disk)
> Oct 18 22:13:00 sjev kernel: md: trying to hot-add unknown-block(8,33)
> to md1 ...
> Oct 18 22:13:00 sjev kernel: md: bind<sdc1>
> Oct 18 22:13:00 sjev kernel: RAID5 conf printout:
> Oct 18 22:13:00 sjev kernel:  --- rd:3 wd:0 fd:2
> Oct 18 22:13:00 sjev kernel:  disk 0, o:1, dev:sdc1
> Oct 18 22:13:00 sjev kernel:  disk 2, o:1, dev:sdd1
> Oct 18 22:13:00 sjev kernel: md: syncing RAID array md1
> Oct 18 22:13:00 sjev kernel: md: minimum _guaranteed_ reconstruction
> speed: 1000 KB/sec/disc.
> Oct 18 22:13:00 sjev kernel: md: using maximum available idle IO
> bandwith (but not more than 200000 KB/sec) for reconstruction.
> Oct 18 22:13:00 sjev kernel: md: using 128k window, over a total of
> 293033536 blocks.
> Oct 18 22:13:00 sjev kernel: md: md1: sync done.
> Oct 18 22:13:00 sjev kernel: md: syncing RAID array md1
> Oct 18 22:13:00 sjev kernel: md: minimum _guaranteed_ reconstruction
> speed: 1000 KB/sec/disc.
> Oct 18 22:13:00 sjev kernel: md: using maximum available idle IO
> bandwith (but not more than 200000 KB/sec) for reconstruction.
> Oct 18 22:13:00 sjev kernel: md: using 128k window, over a total of
> 293033536 blocks.
> Oct 18 22:13:00 sjev kernel: md: md1: sync done.
> Oct 18 22:13:01 sjev kernel: md: syncing RAID array md1
>
> repeats until..
>
> Oct 18 22:14:48 sjev kernel: md: syncing RAID array md1
> Oct 18 22:14:48 sjev kernel: md: minimum _guaranteed_ reconstruction
> speed: 1000 KB/sec/disc.
> Oct 18 22:14:48 sjev kernel: md: using maximum available idle IO
> bandwith (but not more than 200000 KB/sec) for reconstruction.
> Oct 18 22:14:48 sjev kernel: md: using 128k window, over a total of
> 293033536 blocks.
> Oct 18 22:14:48 sjev kernel: md: md1: sync done.
> Oct 18 22:14:48 sjev kernel: Unable to handle kernel NULL pointer
> dereference at virtual address 000000a4
> Oct 18 22:14:48 sjev kernel:  printing eip:
> Oct 18 22:14:48 sjev kernel: c0124d89
> Oct 18 22:14:48 sjev kernel: *pde = 00000000
> Oct 18 22:14:48 sjev kernel: Oops: 0000 [#1]
> Oct 18 22:14:48 sjev kernel: PREEMPT
> Oct 18 22:14:48 sjev kernel: Modules linked in: ipv6 smbfs
> snd_intel8x0m snd_intel8x0 snd_ac97_codec snd_pcm snd_timer
> snd_page_alloc gameport snd_mpu401_uart snd_rawmidi snd_seq_device snd
> capability commoncap raid5 xor sr_mod tsdev mousedev joydev evdev
> pcspkr pci_hotplug intel_agp agpgart ide_scsi ide_generic sg font
> vesafb cfbcopyarea cfbimgblt cfbfillrect appletalk af_packet hw_random
> i810_audio soundcore ac97_codec b44 mii yenta_socket rtc piix unix ds
> pcmcia_core usb_storage ext3 mbcache raid1 md jbd ehci_hcd ohci_hcd
> uhci_hcd usbcore reiserfs psmouse ide_disk ide_cd ide_core cdrom
> sd_mod scsi_mod
> Oct 18 22:14:48 sjev kernel: CPU:    0
> Oct 18 22:14:48 sjev kernel: EIP:    0060:[sig_ignored+73/112]    Not tainted
> Oct 18 22:14:48 sjev kernel: EFLAGS: 00010006   (2.6.8-3-686)
> Oct 18 22:14:48 sjev kernel: EIP is at sig_ignored+0x49/0x70
> Oct 18 22:14:48 sjev kernel: eax: 000000b4   ebx: 00000000   ecx:
> 00000008   edx: 00000000
> Oct 18 22:14:48 sjev kernel: esi: 00000009   edi: 00000009   ebp:
> 00000000   esp: cedf3ec0
> Oct 18 22:14:48 sjev kernel: ds: 007b   es: 007b   ss: 0068
> Oct 18 22:14:48 sjev kernel: Process md1_raid5 (pid: 685,
> threadinfo=cedf2000 task=cedef3e0)
> Oct 18 22:14:48 sjev kernel: Stack: cf10e1b0 00000001 c01259f3
> cf10e1b0 00000009 c86194a0 cf99771c 00000202
> Oct 18 22:14:48 sjev kernel:        cedf2000 cf997680 cf222c00
> c0126565 00000009 00000001 cf10e1b0 c86194a0
> Oct 18 22:14:48 sjev kernel:        cedf3f30 cf997680 d093eb7d
> 00000009 00000001 cf10e1b0 d093ebcd c86194a0
> Oct 18 22:14:48 sjev kernel: Call Trace:
> Oct 18 22:14:48 sjev kernel:  [specific_send_sig_info+83/224]
> specific_send_sig_info+0x53/0xe0
> Oct 18 22:14:48 sjev kernel:  [send_sig_info+69/128] send_sig_info+0x45/0x80
> Oct 18 22:14:48 sjev kernel:  [__crc_sb_min_blocksize+815035/1015327]
> md_interrupt_thread+0x4d/0x60 [md]
> Oct 18 22:14:48 sjev kernel:  [__crc_sb_min_blocksize+815115/1015327]
> md_unregister_thread+0x3d/0x60 [md]
> Oct 18 22:14:48 sjev kernel:  [recalc_task_prio+168/416]
> recalc_task_prio+0xa8/0x1a0
> Oct 18 22:14:48 sjev kernel:  [__crc_sb_min_blocksize+821862/1015327]
> md_check_recovery+0x288/0x300 [md]
> Oct 18 22:14:48 sjev kernel:  [__crc_fb_pan_display+1312520/2923165]
> raid5d+0x19/0x150 [raid5]
> Oct 18 22:14:48 sjev kernel:  [__crc_sb_min_blocksize+814642/1015327]
> md_thread+0x164/0x1d0 [md]
> Oct 18 22:14:48 sjev kernel:  [autoremove_wake_function+0/96]
> autoremove_wake_function+0x0/0x60
> Oct 18 22:14:48 sjev kernel:  [ret_from_fork+6/20] ret_from_fork+0x6/0x14
> Oct 18 22:14:48 sjev kernel:  [autoremove_wake_function+0/96]
> autoremove_wake_function+0x0/0x60
> Oct 18 22:14:48 sjev kernel:  [__crc_sb_min_blocksize+814286/1015327]
> md_thread+0x0/0x1d0 [md]
> Oct 18 22:14:48 sjev kernel:  [kernel_thread_helper+5/24]
> kernel_thread_helper+0x5/0x18
> Oct 18 22:14:48 sjev kernel: Code: 8b 40 f0 83 f8 01 74 18 85 c0 74 04
> 89 d3 eb c1 83 fe 1f 7f
> Oct 18 22:14:48 sjev kernel:  <6>note: md1_raid5[685] exited with
> preempt_count 2
>
>
>>
>>>
>>> Can someone *please* help me get the raid back!?
>>
>> Probably.
>>
>
> I like the optimism! Thanks!
>
>>>
>>> More details -
>>>
>>> Drives are /dev/sdb1, /dev/sdc1 & /dev/sdd1
>>
>> ... or were.  USB device names can change every time you plug them in.
>>
>>>
>>> sdc1 was the one that died earlier this week
>>> sdb1 appears to be the one that was marked as faulty
>>>
>>> mdadm detail before sdc1 was plugged in -
>>>
>>> root@imp[~]:11 # mdadm --detail /dev/md1
>>> /dev/md1:
>> ...
>>>
>>> Number Major Minor RaidDevice State
>>> 0 8 17 0 active sync /dev/sdb1
>>> 1 0 0 - removed
>>> 2 8 49 2 active sync /dev/sdd1
>>
>> So the array thinks the 2nd of 3 is missing.  That is consistent with
>> your description.
>>
>>>
>>>
>>> then after plugging in the replacement sdc1 -
>>>
>>> root@imp[~]:13 # mdadm --add /dev/md1 /dev/sdc1
>>> mdadm: hot added /dev/sdc1
>>> root@imp[~]:14 #
>>> root@imp[~]:14 #
>>> root@imp[~]:14 # mdadm --detail /dev/md1
>>> /dev/md1:
>> ...
>>>
>>> Number Major Minor RaidDevice State
>>> 0 0 0 - removed
>>> 1 0 0 - removed
>>> 2 8 49 2 active sync /dev/sdd1
>>>
>>> 3 8 33 0 spare rebuilding /dev/sdc1
>>> 4 8 17 - faulty /dev/sdb1
>>
>> Yes, sdb must have got an error and failed while sdc was rebuilding.
>> Sad.  That suggests that it didn't fail at the moment of USB
>> insertion, but a little later.  Not conclusively though.
>>
>>>
>>> Shortly after this, subsequent mdadm --details stopped responding.. So
>>> I rebooted in the hope I could reset and problems with the hot add..
>>>
>>> Now, I'm unable to assemble the raid with the 2 working drives -
>>>
>>> mdadm --assemble /dev/md1 /dev/sdb1 /dev/sdd1
>>>
>>> doesn't work -
>>>
>>> mdadm: /dev/md1 assembled from 1 drive and 1 spare - not enough to
>>> start the array.
>>
>> You have rebooted so device names may have changed.
>> If it thought you had named a good drive and a spare, it probably saw
>> the device that was originally sdb (and possibly still is)
>> and the device that was originally sdc (and now might be sdd).
>>
>>>
>>> mdadm --assemble --force /dev/md1 /dev/sdb1 /dev/sdd1
>>>
>>> doesn't' work either
>>
>> What error messages?  Always best to be explicit.
>> Adding "-v" to the --assemble line would help too.
>>
>>>
>>> This -
>>>
>>> mdadm --assemble --force --run /dev/md1 /dev/sdb1 /dev/sdd1
>>>
>>> Did work partially -
>>>
>> Hmm.. That really shouldn't have worked.  The kernel should have
>> rejected the array...
>>
>>>
>>> Here's the output from mdadm -E on each of the 2 drives -
>>
>> Uhm... There should be 3 drives?
>> The 'good' one, the 'new' one, and the one that seemed to fail
>> immediately after you plugged in the 'new' one.
>>
>
> Sorry, here are all 3 -
>
> root@imp[~]:3 # mdadm -E /dev/sd[bcd]1
> /dev/sdb1:
>          Magic : a92b4efc
>        Version : 00.90.00
>           UUID : bed40ee2:98523fdd:e4d010fb:894c0966
>  Creation Time : Fri Nov 17 21:28:44 2006
>     Raid Level : raid5
>   Raid Devices : 3
>  Total Devices : 3
> Preferred Minor : 1
>
>    Update Time : Sat Oct 18 22:14:48 2008
>          State : clean
>  Active Devices : 1
> Working Devices : 2
>  Failed Devices : 2
>  Spare Devices : 1
>       Checksum : e6dbf86 - correct
>         Events : 0.1521614
>
>         Layout : left-symmetric
>     Chunk Size : 64K
>
>      Number   Major   Minor   RaidDevice State
> this     2       8       49        2      active sync   /dev/sdd1
>
>   0     0       0        0        0      removed
>   1     1       0        0        1      faulty removed
>   2     2       8       49        2      active sync   /dev/sdd1
>   3     3       8       33        0      spare   /dev/sdc1
> /dev/sdc1:
>          Magic : a92b4efc
>        Version : 00.90.00
>           UUID : bed40ee2:98523fdd:e4d010fb:894c0966
>  Creation Time : Fri Nov 17 21:28:44 2006
>     Raid Level : raid5
>   Raid Devices : 3
>  Total Devices : 3
> Preferred Minor : 1
>
>    Update Time : Fri Oct 17 22:30:49 2008
>          State : clean
>  Active Devices : 2
> Working Devices : 3
>  Failed Devices : 1
>  Spare Devices : 1
>       Checksum : e6ae9ea - correct
>         Events : 0.1471469
>
>         Layout : left-symmetric
>     Chunk Size : 64K
>
>      Number   Major   Minor   RaidDevice State
> this     3       8       33        3      spare   /dev/sdc1
>
>   0     0       8       17        0      active sync   /dev/sdb1
>   1     1       0        0        1      faulty removed
>   2     2       8       49        2      active sync   /dev/sdd1
>   3     3       8       33        3      spare   /dev/sdc1
> /dev/sdd1:
>          Magic : a92b4efc
>        Version : 00.90.00
>           UUID : bed40ee2:98523fdd:e4d010fb:894c0966
>  Creation Time : Fri Nov 17 21:28:44 2006
>     Raid Level : raid5
>   Raid Devices : 3
>  Total Devices : 3
> Preferred Minor : 1
>
>    Update Time : Sat Oct 18 22:14:48 2008
>          State : clean
>  Active Devices : 1
> Working Devices : 2
>  Failed Devices : 2
>  Spare Devices : 1
>       Checksum : e6dbf75 - correct
>         Events : 0.1521614
>
>         Layout : left-symmetric
>     Chunk Size : 64K
>
>      Number   Major   Minor   RaidDevice State
> this     3       8       33        3      spare   /dev/sdc1
>
>   0     0       0        0        0      removed
>   1     1       0        0        1      faulty removed
>   2     2       8       49        2      active sync   /dev/sdd1
>   3     3       8       33        3      spare   /dev/sdc1
>
> fdisk details too -
>
> root@imp[~]:7 # fdisk -l /dev/sd[bcd]
>
> Disk /dev/sdb: 300.0 GB, 300069052416 bytes
> 255 heads, 63 sectors/track, 36481 cylinders
> Units = cylinders of 16065 * 512 = 8225280 bytes
>
>   Device Boot      Start         End      Blocks   Id  System
> /dev/sdb1               1       36481   293033601   fd  Linux raid autodetect
>
> Disk /dev/sdc: 320.0 GB, 320072933376 bytes
> 255 heads, 63 sectors/track, 38913 cylinders
> Units = cylinders of 16065 * 512 = 8225280 bytes
>
>   Device Boot      Start         End      Blocks   Id  System
> /dev/sdc1               1       36481   293033601   fd  Linux raid autodetect
>
> Disk /dev/sdd: 300.0 GB, 300069052416 bytes
> 255 heads, 63 sectors/track, 36481 cylinders
> Units = cylinders of 16065 * 512 = 8225280 bytes
>
>   Device Boot      Start         End      Blocks   Id  System
> /dev/sdd1               1       36481   293033601   fd  Linux raid autodetect
>
>
>>>
>>> /dev/sdb1:
>> ..
>>> Number Major Minor RaidDevice State
>>> this 3 8 33 3 spare /dev/sdc1
>>>
>>> 0 0 0 0 0 removed
>>> 1 1 0 0 1 faulty removed
>>> 2 2 8 49 2 active sync /dev/sdd1
>>> 3 3 8 33 3 spare /dev/sdc1
>>
>> sdb looks like the new one.
>>
>>> /dev/sdd1:
>> ...
>>>
>>> Number Major Minor RaidDevice State
>>> this 2 8 49 2 active sync /dev/sdd1
>>>
>>> 0 0 0 0 0 removed
>>> 1 1 0 0 1 faulty removed
>>> 2 2 8 49 2 active sync /dev/sdd1
>>> 3 3 8 33 0 spare /dev/sdc1
>>
>> sdd looks like the good one.
>>
>> Where is the "one that seemed to fail" which was once called sdb ??
>>>
>>> Is all the data lost, or can I recover from this?
>>
>> Try
>>
>>  mdadm --examine --brief --verbose /dev/sd*
>>
>
> ARRAY /dev/md1 level=raid5 num-devices=3
> UUID=bed40ee2:98523fdd:e4d010fb:894c0966
>   devices=/dev/sdb1,/dev/sdc1,/dev/sdd1
> ARRAY /dev/md4 level=raid1 num-devices=2
> UUID=6fded12b:6ecdca8a:18400b9a:df6a2ffc
>   devices=/dev/sda5
> ARRAY /dev/md0 level=raid1 num-devices=2
> UUID=c94d0631:20f0db42:9c6ab972:19acc617
>   devices=/dev/sda1
>
>>
>> Then
>>
>>  mdadm --assemble --force --verbose /dev/md1 /dev/sd....
>>
>> where you list all the devices in the device= section for the array
>> you want to try to start.
>>
>> Report the output of that command and whether it was successful.
>
> root@imp[~]:9 # mdadm --assemble --force --verbose /dev/md1 /dev/sdb1
> /dev/sdc1 /dev/sdd1
> mdadm: looking for devices for /dev/md1
> mdadm: /dev/sdb1 is identified as a member of /dev/md1, slot 2.
> mdadm: /dev/sdc1 is identified as a member of /dev/md1, slot 3.
> mdadm: /dev/sdd1 is identified as a member of /dev/md1, slot 3.
> mdadm: no uptodate device for slot 0 of /dev/md1
> mdadm: no uptodate device for slot 1 of /dev/md1
> mdadm: added /dev/sdd1 to /dev/md1 as 3
> mdadm: added /dev/sdb1 to /dev/md1 as 2
> mdadm: /dev/md1 assembled from 1 drive and 1 spare - not enough to
> start the array.
> root@imp[~]:10 #
>
> Oct 29 14:52:41 sjev kernel: md: md1 stopped.
> Oct 29 14:52:41 sjev kernel: md: unbind<sdb1>
> Oct 29 14:52:41 sjev kernel: md: export_rdev(sdb1)
> Oct 29 14:52:41 sjev kernel: md: unbind<sdd1>
> Oct 29 14:52:41 sjev kernel: md: export_rdev(sdd1)
> Oct 29 14:52:41 sjev kernel: md: bind<sdd1>
> Oct 29 14:52:41 sjev kernel: md: bind<sdb1>
> Oct 29 14:58:07 sjev smartd[2302]: Device: /dev/hdc, SMART Usage
> Attribute: 190 Unknown_Attribute changed from 49 to 48
> Oct 29 14:58:07 sjev smartd[2302]: Device: /dev/hdc, SMART Usage
> Attribute: 194 Temperature_Celsius changed from 51 to 52
>
> I've held off upgrading mdadm to the latest version until I know it's
> the best option (vs recovering the raid 1st before upgrading), so you
> agree?
>
>>
>> NeilBrown
>>
>
> Thanks for your patience and help!
> Regards,
> Steve..
>

^ permalink raw reply	[flat|nested] 7+ messages in thread

* Re: mdadm degraded RAID5 failure
  2008-11-04 21:35       ` Steve Evans
@ 2008-11-06  5:41         ` Neil Brown
  0 siblings, 0 replies; 7+ messages in thread
From: Neil Brown @ 2008-11-06  5:41 UTC (permalink / raw)
  To: Steve Evans; +Cc: linux-raid

On Tuesday November 4, jeeping@gmail.com wrote:
> Hi Neil and others,
> 
> Just a couple of questions, I know you're busy -
> 
> Do you recommend that I attempt to upgrade mdadm to a more recent
> version before any other recovery attempts? If so, which version?

Yes.  2.6.7.1 (the latest).

> 
> I noted my replacement drive (sdc1) got a smart error (during the
> rebuild?), would you recommend replacing it or removing it altogether
> until I get the other 2 drives back online (if I even can)?

There seem to be different opinion on how much weight to put on SMART
errors.   So I make no recommendations based on them.

> 
> Is there a way to correct the drive names -

When you assemble the array again, it will update the device names to
what they are at the time.

As you have 2 devices that think they are 'spare', you won't be able
to assemble a working array using "--assemble".

What you will need to do is recreate the array over just two devices
and make sure you get them in the right order.

The one that claims to be device '2' (sdb1 below) certainly is device
2 (i.e. the last device: they are numbered 0,1,2).  The others I can
not be so sure of.

So I would recreate the array with e.g.

  mdadm -C /dev/md0 -l5 -n2 /dev/sdc1 missing /dev/sdb1

And check it with e.g.
  fsck -n -f /dev/md0

If fsck is happy: good.  If not, try again with a different
arrangement:

   mdadm -C /dev/md0 -l5 -n2 missing /dev/sdc1 /dev/sdb1

etc.  I don't know which of c1 and d1 is more likely to have good
data.   Keep going until you get a good 'fsck'.

Make very sure to use the "-n" option to fsck to ensure it doesn't try
to 'fix' the mess it finds.

Also, before doing the above, run "mdadm --examine /dev/sdb1" and keep
a record of that.  Check the 'chunksize'.  If it isn't 64, you will
need to explicitly give then number to "mdadm -C".  Also check the
layout and possibly set that explicitly when doing "mdadm -C".

good luck.

NeilBrown


> 
> > /dev/sdb1:
> > this     2       8       49        2      active sync   /dev/sdd1
> 
> 
> > /dev/sdc1:
> > this     3       8       33        3      spare   /dev/sdc1
> 
> 
> > /dev/sdd1:
> > this     3       8       33        3      spare   /dev/sdc1
> 
> I'm inclined to believe (but am not sure at all) that -
> 
> sdb1 should be sdd1
> sdc1 is correct
> sdd1 should be sdb1
> 
> Thanks!
> Steve..
> 

^ permalink raw reply	[flat|nested] 7+ messages in thread

end of thread, other threads:[~2008-11-06  5:41 UTC | newest]

Thread overview: 7+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
     [not found] <6cc8e9ed0810221350o2b8b3aedm3d1c229fe7e66163@mail.gmail.com>
2008-10-22 20:52 ` mdadm degraded RAID5 failure Steve Evans
2008-10-24 18:47   ` Steve Evans
2008-10-25  6:30   ` Neil Brown
2008-10-25 10:44     ` David Greaves
2008-10-29 22:16     ` Steve Evans
2008-11-04 21:35       ` Steve Evans
2008-11-06  5:41         ` Neil Brown

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).