linux-raid.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: David Greaves <david@dgreaves.com>
To: Mark Davies <mark@curly.ii.net>, Neil Brown <neilb@suse.de>
Cc: linux-raid@vger.kernel.org
Subject: Re: raid5 recovery dramas.
Date: Thu, 26 Jun 2008 14:38:38 +0100	[thread overview]
Message-ID: <48639BDE.70001@dgreaves.com> (raw)
In-Reply-To: <4863024A.4030909@curly.ii.net>

Mark Davies wrote:
> No takers?  Is there a different list anyone can suggest I repost this
> to, and any extra information I could include?

You are in the right place - but this may be a nasty problem.
I'd wait for Neil to comment (cc'ed to attract his attention to this one)

You've grown an array from 4-5 and had a 2 disk failure part way through - ouch!!

However, you've recovered the 2 failed disks using ddrescue but of course the
superblock event counts are wrong.

It may be that a simple --assemble --force would work. I've not had enough
experience of failed grow operations.

the /dev/sde1 problem *may* be caused by lvm - try stopping that. However doing
this from an uptodate rescue CD sounds sensible.

You *don't* want to mess with --create and --permute. That's almost guaranteed
to kill the array in this case (due to the reshape).

David

> 
> I found a link to a mdadm create/permutation script
> 
> http://linux-raid.osdl.org/index.php/Permute_array.pl
> 
> Would that appear to be useful in my situation?
> 
> My problematic array was created with mdadm version:
> 
> mdadm --version
> mdadm - v2.5.6 - 9 November 2006
> 
> If I was to boot with a LiveCD (to get around this:
> 
> mdadm: failed to add /dev/sde1 to /dev/md1: Device or resource busy
> 
> error, would the version of mdadm have to be the same, or just more recent?
> 
> Oh, and I'm willing to send a sixpack of beer or whatever in thanks.  :)
> 
> 
> 
> Regards,
> 
> 
> Mark.
> 
> 
> 
> Mark Davies wrote:
>> Hi all,
>>
>> Hoping to find some information to help me recover my software raid5
>> array.
>>
>> Some background information first (excuse the hostname)
>>
>> uname -a
>> Linux Fuckyfucky3 2.6.18-4-686 #1 SMP Wed May 9 23:03:12 UTC 2007 i686
>> GNU/Linux
>>
>>
>> It's a debian box that initially had 4 disks in a software raid5 array.
>>
>> The problem started when I attempted to add another disk and grow the
>> array.  I'd already done this from 3-4 disks using the instruction on
>> this page:  "http://scotgate.org/?p=107".
>>
>> However this time I unmounted the volume, but didn't do a fsck before
>> starting.  I also discovered that for some reason mdadm wasn't
>> monitoring the array.
>>
>> Bad mistakes obviously - and I hope I've learnt from them.
>>
>> Short version is that two of the disks had errors on them, and so
>> mdadm disabled those disks about 50MB into the reshape.  Both failed
>> SMART tests subsequently.
>>
>> I bought two new disks, and used dd-recue to make copies of them,
>> which seemed to work well.
>>
>> Now however I can't restart the array.
>>
>> I can see all 5 superblocks:
>>
>> :~# mdadm --examine /dev/sd?1
>> /dev/sda1:
>>           Magic : a92b4efc
>>         Version : 01
>>     Feature Map : 0x4
>>      Array UUID : 43eff327:8d1aa506:c0df2849:005c003f
>>            Name : 'Fuckyfucky3':1
>>   Creation Time : Sun Dec 23 01:28:08 2007
>>      Raid Level : raid5
>>    Raid Devices : 5
>>
>>     Device Size : 976767856 (465.76 GiB 500.11 GB)
>>      Array Size : 3907069952 (1863.04 GiB 2000.42 GB)
>>       Used Size : 976767488 (465.76 GiB 500.10 GB)
>>    Super Offset : 976767984 sectors
>>           State : clean
>>     Device UUID : 5b38c5a2:798c6793:91ad6d1e:9cfee153
>>
>>   Reshape pos'n : 143872 (140.52 MiB 147.32 MB)
>>   Delta Devices : 1 (4->5)
>>
>>     Update Time : Fri May 16 23:55:29 2008
>>        Checksum : 5354498d - correct
>>          Events : 1420762
>>
>>          Layout : left-symmetric
>>      Chunk Size : 128K
>>
>>     Array Slot : 3 (failed, 1, failed, 2, failed, 0)
>>    Array State : uuU__ 3 failed
>> /dev/sdb1:
>>           Magic : a92b4efc
>>         Version : 01
>>     Feature Map : 0x4
>>      Array UUID : 43eff327:8d1aa506:c0df2849:005c003f
>>            Name : 'Fuckyfucky3':1
>>   Creation Time : Sun Dec 23 01:28:08 2007
>>      Raid Level : raid5
>>    Raid Devices : 5
>>
>>     Device Size : 976767856 (465.76 GiB 500.11 GB)
>>      Array Size : 3907069952 (1863.04 GiB 2000.42 GB)
>>       Used Size : 976767488 (465.76 GiB 500.10 GB)
>>    Super Offset : 976767984 sectors
>>           State : clean
>>     Device UUID : 673ba6d4:6c46fd55:745c9c93:3fa8bf21
>>
>>   Reshape pos'n : 143872 (140.52 MiB 147.32 MB)
>>   Delta Devices : 1 (4->5)
>>
>>     Update Time : Fri May 16 23:55:29 2008
>>        Checksum : 8ad75f10 - correct
>>          Events : 1420762
>>
>>          Layout : left-symmetric
>>      Chunk Size : 128K
>>
>>     Array Slot : 1 (failed, 1, failed, 2, failed, 0)
>>    Array State : uUu__ 3 failed
>> /dev/sdc1:
>>           Magic : a92b4efc
>>         Version : 01
>>     Feature Map : 0x4
>>      Array UUID : 43eff327:8d1aa506:c0df2849:005c003f
>>            Name : 'Fuckyfucky3':1
>>   Creation Time : Sun Dec 23 01:28:08 2007
>>      Raid Level : raid5
>>    Raid Devices : 5
>>
>>     Device Size : 976767856 (465.76 GiB 500.11 GB)
>>      Array Size : 3907069952 (1863.04 GiB 2000.42 GB)
>>       Used Size : 976767488 (465.76 GiB 500.10 GB)
>>    Super Offset : 976767984 sectors
>>           State : clean
>>     Device UUID : 99b87c50:a919bd63:599a135f:9af385ba
>>
>>   Reshape pos'n : 143872 (140.52 MiB 147.32 MB)
>>   Delta Devices : 1 (4->5)
>>
>>     Update Time : Fri May 16 23:55:29 2008
>>        Checksum : 78ab38c3 - correct
>>          Events : 1420762
>>
>>          Layout : left-symmetric
>>      Chunk Size : 128K
>>
>>     Array Slot : 5 (failed, 1, failed, 2, failed, 0)
>>    Array State : Uuu__ 3 failed
>> /dev/sdd1:
>>           Magic : a92b4efc
>>         Version : 01
>>     Feature Map : 0x4
>>      Array UUID : 43eff327:8d1aa506:c0df2849:005c003f
>>            Name : 'Fuckyfucky3':1
>>   Creation Time : Sun Dec 23 01:28:08 2007
>>      Raid Level : raid5
>>    Raid Devices : 5
>>
>>     Device Size : 976767856 (465.76 GiB 500.11 GB)
>>      Array Size : 3907069952 (1863.04 GiB 2000.42 GB)
>>       Used Size : 976767488 (465.76 GiB 500.10 GB)
>>    Super Offset : 976767984 sectors
>>           State : clean
>>     Device UUID : 89201477:8e950d20:9193016d:f5c9deb0
>>
>>   Reshape pos'n : 143872 (140.52 MiB 147.32 MB)
>>   Delta Devices : 1 (4->5)
>>
>>     Update Time : Fri May 16 23:55:29 2008
>>        Checksum : 5fc43e52 - correct
>>          Events : 0
>>
>>          Layout : left-symmetric
>>      Chunk Size : 128K
>>
>>     Array Slot : 6 (failed, 1, failed, 2, failed, 0)
>>    Array State : uuu__ 3 failed
>> /dev/sde1:
>>           Magic : a92b4efc
>>         Version : 01
>>     Feature Map : 0x4
>>      Array UUID : 43eff327:8d1aa506:c0df2849:005c003f
>>            Name : 'Fuckyfucky3':1
>>   Creation Time : Sun Dec 23 01:28:08 2007
>>      Raid Level : raid5
>>    Raid Devices : 5
>>
>>     Device Size : 976767856 (465.76 GiB 500.11 GB)
>>      Array Size : 3907069952 (1863.04 GiB 2000.42 GB)
>>       Used Size : 976767488 (465.76 GiB 500.10 GB)
>>    Super Offset : 976767984 sectors
>>           State : clean
>>     Device UUID : 89b53542:d1d820bc:f2ece884:4785869a
>>
>>   Reshape pos'n : 143872 (140.52 MiB 147.32 MB)
>>   Delta Devices : 1 (4->5)
>>
>>     Update Time : Fri May 16 23:55:29 2008
>>        Checksum : c89dd220 - correct
>>          Events : 1418968
>>
>>          Layout : left-symmetric
>>      Chunk Size : 128K
>>
>>     Array Slot : 6 (failed, 1, failed, 2, failed, 0)
>>    Array State : uuu__ 3 failed
>>
>>
>>
>>
>> When I try to start the array, I get:
>>
>> ~# mdadm --assemble --verbose /dev/md1 /dev/sda1 /dev/sdb1 /dev/sdc1
>> /dev/sdd1 /dev/sde1
>> mdadm: looking for devices for /dev/md1
>> mdadm: /dev/sda1 is identified as a member of /dev/md1, slot 2.
>> mdadm: /dev/sdb1 is identified as a member of /dev/md1, slot 1.
>> mdadm: /dev/sdc1 is identified as a member of /dev/md1, slot 0.
>> mdadm: /dev/sdd1 is identified as a member of /dev/md1, slot -1.
>> mdadm: /dev/sde1 is identified as a member of /dev/md1, slot -1.
>> mdadm: added /dev/sdb1 to /dev/md1 as 1
>> mdadm: added /dev/sda1 to /dev/md1 as 2
>> mdadm: no uptodate device for slot 3 of /dev/md1
>> mdadm: no uptodate device for slot 4 of /dev/md1
>> mdadm: added /dev/sdd1 to /dev/md1 as -1
>> mdadm: failed to add /dev/sde1 to /dev/md1: Device or resource busy
>> mdadm: added /dev/sdc1 to /dev/md1 as 0
>> mdadm: /dev/md1 assembled from 3 drives and -1 spares - not enough to
>> start the array.
>>
>>
>>
>>
>> Any help would be much appreciated.   If I can provide any more
>> information, just ask.
>>
>> As to why /dev/sde1 is busy, I don't know.  lsof shows no files open.
>>
>>
>> Regards,
>>
>>
>> Mark.
>> -- 
>> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
>> the body of a message to majordomo@vger.kernel.org
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> -- 
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html


  reply	other threads:[~2008-06-26 13:38 UTC|newest]

Thread overview: 8+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2008-06-24  6:05 raid5 recovery dramas Mark Davies
2008-06-26  2:43 ` Mark Davies
2008-06-26 13:38   ` David Greaves [this message]
2008-06-26 14:25     ` Mark Davies
2008-06-27 10:28 ` Neil Brown
2008-06-27 11:14   ` Mark Davies
2008-06-27 20:44     ` Neil Brown
2008-06-30  7:03       ` Mark Davies

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=48639BDE.70001@dgreaves.com \
    --to=david@dgreaves.com \
    --cc=linux-raid@vger.kernel.org \
    --cc=mark@curly.ii.net \
    --cc=neilb@suse.de \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).