linux-raid.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: Mark Davies <mark@curly.ii.net>
To: linux-raid@vger.kernel.org
Subject: Re: raid5 recovery dramas.
Date: Thu, 26 Jun 2008 10:43:22 +0800	[thread overview]
Message-ID: <4863024A.4030909@curly.ii.net> (raw)
In-Reply-To: <48608E9B.6080904@curly.ii.net>

No takers?  Is there a different list anyone can suggest I repost this 
to, and any extra information I could include?

I found a link to a mdadm create/permutation script

http://linux-raid.osdl.org/index.php/Permute_array.pl

Would that appear to be useful in my situation?

My problematic array was created with mdadm version:

mdadm --version
mdadm - v2.5.6 - 9 November 2006

If I was to boot with a LiveCD (to get around this:

mdadm: failed to add /dev/sde1 to /dev/md1: Device or resource busy

error, would the version of mdadm have to be the same, or just more recent?

Oh, and I'm willing to send a sixpack of beer or whatever in thanks.  :)



Regards,


Mark.



Mark Davies wrote:
> Hi all,
> 
> Hoping to find some information to help me recover my software raid5 array.
> 
> Some background information first (excuse the hostname)
> 
> uname -a
> Linux Fuckyfucky3 2.6.18-4-686 #1 SMP Wed May 9 23:03:12 UTC 2007 i686 
> GNU/Linux
> 
> 
> It's a debian box that initially had 4 disks in a software raid5 array.
> 
> The problem started when I attempted to add another disk and grow the 
> array.  I'd already done this from 3-4 disks using the instruction on 
> this page:  "http://scotgate.org/?p=107".
> 
> However this time I unmounted the volume, but didn't do a fsck before 
> starting.  I also discovered that for some reason mdadm wasn't 
> monitoring the array.
> 
> Bad mistakes obviously - and I hope I've learnt from them.
> 
> Short version is that two of the disks had errors on them, and so mdadm 
> disabled those disks about 50MB into the reshape.  Both failed SMART 
> tests subsequently.
> 
> I bought two new disks, and used dd-recue to make copies of them, which 
> seemed to work well.
> 
> Now however I can't restart the array.
> 
> I can see all 5 superblocks:
> 
> :~# mdadm --examine /dev/sd?1
> /dev/sda1:
>           Magic : a92b4efc
>         Version : 01
>     Feature Map : 0x4
>      Array UUID : 43eff327:8d1aa506:c0df2849:005c003f
>            Name : 'Fuckyfucky3':1
>   Creation Time : Sun Dec 23 01:28:08 2007
>      Raid Level : raid5
>    Raid Devices : 5
> 
>     Device Size : 976767856 (465.76 GiB 500.11 GB)
>      Array Size : 3907069952 (1863.04 GiB 2000.42 GB)
>       Used Size : 976767488 (465.76 GiB 500.10 GB)
>    Super Offset : 976767984 sectors
>           State : clean
>     Device UUID : 5b38c5a2:798c6793:91ad6d1e:9cfee153
> 
>   Reshape pos'n : 143872 (140.52 MiB 147.32 MB)
>   Delta Devices : 1 (4->5)
> 
>     Update Time : Fri May 16 23:55:29 2008
>        Checksum : 5354498d - correct
>          Events : 1420762
> 
>          Layout : left-symmetric
>      Chunk Size : 128K
> 
>     Array Slot : 3 (failed, 1, failed, 2, failed, 0)
>    Array State : uuU__ 3 failed
> /dev/sdb1:
>           Magic : a92b4efc
>         Version : 01
>     Feature Map : 0x4
>      Array UUID : 43eff327:8d1aa506:c0df2849:005c003f
>            Name : 'Fuckyfucky3':1
>   Creation Time : Sun Dec 23 01:28:08 2007
>      Raid Level : raid5
>    Raid Devices : 5
> 
>     Device Size : 976767856 (465.76 GiB 500.11 GB)
>      Array Size : 3907069952 (1863.04 GiB 2000.42 GB)
>       Used Size : 976767488 (465.76 GiB 500.10 GB)
>    Super Offset : 976767984 sectors
>           State : clean
>     Device UUID : 673ba6d4:6c46fd55:745c9c93:3fa8bf21
> 
>   Reshape pos'n : 143872 (140.52 MiB 147.32 MB)
>   Delta Devices : 1 (4->5)
> 
>     Update Time : Fri May 16 23:55:29 2008
>        Checksum : 8ad75f10 - correct
>          Events : 1420762
> 
>          Layout : left-symmetric
>      Chunk Size : 128K
> 
>     Array Slot : 1 (failed, 1, failed, 2, failed, 0)
>    Array State : uUu__ 3 failed
> /dev/sdc1:
>           Magic : a92b4efc
>         Version : 01
>     Feature Map : 0x4
>      Array UUID : 43eff327:8d1aa506:c0df2849:005c003f
>            Name : 'Fuckyfucky3':1
>   Creation Time : Sun Dec 23 01:28:08 2007
>      Raid Level : raid5
>    Raid Devices : 5
> 
>     Device Size : 976767856 (465.76 GiB 500.11 GB)
>      Array Size : 3907069952 (1863.04 GiB 2000.42 GB)
>       Used Size : 976767488 (465.76 GiB 500.10 GB)
>    Super Offset : 976767984 sectors
>           State : clean
>     Device UUID : 99b87c50:a919bd63:599a135f:9af385ba
> 
>   Reshape pos'n : 143872 (140.52 MiB 147.32 MB)
>   Delta Devices : 1 (4->5)
> 
>     Update Time : Fri May 16 23:55:29 2008
>        Checksum : 78ab38c3 - correct
>          Events : 1420762
> 
>          Layout : left-symmetric
>      Chunk Size : 128K
> 
>     Array Slot : 5 (failed, 1, failed, 2, failed, 0)
>    Array State : Uuu__ 3 failed
> /dev/sdd1:
>           Magic : a92b4efc
>         Version : 01
>     Feature Map : 0x4
>      Array UUID : 43eff327:8d1aa506:c0df2849:005c003f
>            Name : 'Fuckyfucky3':1
>   Creation Time : Sun Dec 23 01:28:08 2007
>      Raid Level : raid5
>    Raid Devices : 5
> 
>     Device Size : 976767856 (465.76 GiB 500.11 GB)
>      Array Size : 3907069952 (1863.04 GiB 2000.42 GB)
>       Used Size : 976767488 (465.76 GiB 500.10 GB)
>    Super Offset : 976767984 sectors
>           State : clean
>     Device UUID : 89201477:8e950d20:9193016d:f5c9deb0
> 
>   Reshape pos'n : 143872 (140.52 MiB 147.32 MB)
>   Delta Devices : 1 (4->5)
> 
>     Update Time : Fri May 16 23:55:29 2008
>        Checksum : 5fc43e52 - correct
>          Events : 0
> 
>          Layout : left-symmetric
>      Chunk Size : 128K
> 
>     Array Slot : 6 (failed, 1, failed, 2, failed, 0)
>    Array State : uuu__ 3 failed
> /dev/sde1:
>           Magic : a92b4efc
>         Version : 01
>     Feature Map : 0x4
>      Array UUID : 43eff327:8d1aa506:c0df2849:005c003f
>            Name : 'Fuckyfucky3':1
>   Creation Time : Sun Dec 23 01:28:08 2007
>      Raid Level : raid5
>    Raid Devices : 5
> 
>     Device Size : 976767856 (465.76 GiB 500.11 GB)
>      Array Size : 3907069952 (1863.04 GiB 2000.42 GB)
>       Used Size : 976767488 (465.76 GiB 500.10 GB)
>    Super Offset : 976767984 sectors
>           State : clean
>     Device UUID : 89b53542:d1d820bc:f2ece884:4785869a
> 
>   Reshape pos'n : 143872 (140.52 MiB 147.32 MB)
>   Delta Devices : 1 (4->5)
> 
>     Update Time : Fri May 16 23:55:29 2008
>        Checksum : c89dd220 - correct
>          Events : 1418968
> 
>          Layout : left-symmetric
>      Chunk Size : 128K
> 
>     Array Slot : 6 (failed, 1, failed, 2, failed, 0)
>    Array State : uuu__ 3 failed
> 
> 
> 
> 
> When I try to start the array, I get:
> 
> ~# mdadm --assemble --verbose /dev/md1 /dev/sda1 /dev/sdb1 /dev/sdc1 
> /dev/sdd1 /dev/sde1
> mdadm: looking for devices for /dev/md1
> mdadm: /dev/sda1 is identified as a member of /dev/md1, slot 2.
> mdadm: /dev/sdb1 is identified as a member of /dev/md1, slot 1.
> mdadm: /dev/sdc1 is identified as a member of /dev/md1, slot 0.
> mdadm: /dev/sdd1 is identified as a member of /dev/md1, slot -1.
> mdadm: /dev/sde1 is identified as a member of /dev/md1, slot -1.
> mdadm: added /dev/sdb1 to /dev/md1 as 1
> mdadm: added /dev/sda1 to /dev/md1 as 2
> mdadm: no uptodate device for slot 3 of /dev/md1
> mdadm: no uptodate device for slot 4 of /dev/md1
> mdadm: added /dev/sdd1 to /dev/md1 as -1
> mdadm: failed to add /dev/sde1 to /dev/md1: Device or resource busy
> mdadm: added /dev/sdc1 to /dev/md1 as 0
> mdadm: /dev/md1 assembled from 3 drives and -1 spares - not enough to 
> start the array.
> 
> 
> 
> 
> Any help would be much appreciated.   If I can provide any more 
> information, just ask.
> 
> As to why /dev/sde1 is busy, I don't know.  lsof shows no files open.
> 
> 
> Regards,
> 
> 
> Mark.
> -- 
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

  reply	other threads:[~2008-06-26  2:43 UTC|newest]

Thread overview: 8+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2008-06-24  6:05 raid5 recovery dramas Mark Davies
2008-06-26  2:43 ` Mark Davies [this message]
2008-06-26 13:38   ` David Greaves
2008-06-26 14:25     ` Mark Davies
2008-06-27 10:28 ` Neil Brown
2008-06-27 11:14   ` Mark Davies
2008-06-27 20:44     ` Neil Brown
2008-06-30  7:03       ` Mark Davies

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=4863024A.4030909@curly.ii.net \
    --to=mark@curly.ii.net \
    --cc=linux-raid@vger.kernel.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).