linux-raid.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: Craig Haskins <reptile55@gmail.com>
To: linux-raid@vger.kernel.org
Subject: Re: Raid5 Reshape failed can't assemble array
Date: Sat, 23 Jan 2010 19:38:35 +0800	[thread overview]
Message-ID: <4B5ADFBB.9040009@gmail.com> (raw)
In-Reply-To: <4B5A2A48.3070208@gmail.com>

I figured out my problem, just had to stop the array first.  Strangely 
though assemble reported not enough drives to start array but when I 
checked /dev/md0 it shows it as degraded and recovering.  Anyway looks 
like my file system is fine, that is a relief.

dev/md0:
         Version : 0.91
   Creation Time : Sun Mar 15 05:22:48 2009
      Raid Level : raid5
      Array Size : 3907039744 (3726.04 GiB 4000.81 GB)
   Used Dev Size : 976759936 (931.51 GiB 1000.20 GB)
    Raid Devices : 6
   Total Devices : 5
Preferred Minor : 0
     Persistence : Superblock is persistent

     Update Time : Sat Jan 23 19:33:23 2010
           State : clean, degraded, recovering
  Active Devices : 5
Working Devices : 5
  Failed Devices : 0
   Spare Devices : 0

          Layout : left-symmetric
      Chunk Size : 64K

  Reshape Status : 86% complete
   Delta Devices : 1, (5->6)

            UUID : be282ff9:764d9beb:74ac4a35:dfcae213
          Events : 0.764194

     Number   Major   Minor   RaidDevice State
        0       8       81        0      active sync   /dev/sdf1
        1       8       65        1      active sync   /dev/sde1
        2       8      113        2      active sync   /dev/sdh1
        3       8       97        3      active sync   /dev/sdg1
        4       8       49        4      active sync   /dev/sdd1
        5       0        0        5      removed



Craig Haskins wrote:
> Hi,
> 
> I wonder if anyone can help me, one of my drives was kicked during 
> reshape (growing from 5 to 6 devices) and after a reboot my array 
> refused to start.  It's very similar to the case here 
> http://marc.info/?t=125218236000001&r=1&w=2 .  I have followed the steps 
> presented in that thread but so far my array refuses to assemble.
> 
> I'm running Ubutunu 9.10 which had mdadm v2.7, that segfaulted when 
> trying to assemble so I got v3.1.1 but still the assemble isn't working:
> 
>  >root@aura:/home/craigh# mdadm -Af --verbose /dev/md0
> mdadm: looking for devices for /dev/md0
> mdadm: /dev/block/252:0 is not one of 
> /dev/sdf1,/dev/sde1,/dev/sdh1,/dev/sdg1,/dev/sdd1,dev/sdc1
> mdadm: cannot open device /dev/sdh1: Device or resource busy
> mdadm: /dev/sdh1 has wrong uuid.
> mdadm: /dev/sdh is not one of 
> /dev/sdf1,/dev/sde1,/dev/sdh1,/dev/sdg1,/dev/sdd1,dev/sdc1
> mdadm: cannot open device /dev/sdg1: Device or resource busy
> mdadm: /dev/sdg1 has wrong uuid.
> mdadm: /dev/sdg is not one of 
> /dev/sdf1,/dev/sde1,/dev/sdh1,/dev/sdg1,/dev/sdd1,dev/sdc1
> mdadm: cannot open device /dev/sdf1: Device or resource busy
> mdadm: /dev/sdf1 has wrong uuid.
> mdadm: /dev/sdf is not one of 
> /dev/sdf1,/dev/sde1,/dev/sdh1,/dev/sdg1,/dev/sdd1,dev/sdc1
> mdadm: cannot open device /dev/sde1: Device or resource busy
> mdadm: /dev/sde1 has wrong uuid.
> mdadm: /dev/sde is not one of 
> /dev/sdf1,/dev/sde1,/dev/sdh1,/dev/sdg1,/dev/sdd1,dev/sdc1
> mdadm: cannot open device /dev/sdd1: Device or resource busy
> mdadm: /dev/sdd1 has wrong uuid.
> mdadm: /dev/sdd is not one of 
> /dev/sdf1,/dev/sde1,/dev/sdh1,/dev/sdg1,/dev/sdd1,dev/sdc1
> mdadm: /dev/sdc1 is not one of 
> /dev/sdf1,/dev/sde1,/dev/sdh1,/dev/sdg1,/dev/sdd1,dev/sdc1
> mdadm: /dev/sdc is not one of 
> /dev/sdf1,/dev/sde1,/dev/sdh1,/dev/sdg1,/dev/sdd1,dev/sdc1
> mdadm: /dev/sdb5 is not one of 
> /dev/sdf1,/dev/sde1,/dev/sdh1,/dev/sdg1,/dev/sdd1,dev/sdc1
> mdadm: /dev/sdb2 is not one of 
> /dev/sdf1,/dev/sde1,/dev/sdh1,/dev/sdg1,/dev/sdd1,dev/sdc1
> mdadm: /dev/sdb1 is not one of 
> /dev/sdf1,/dev/sde1,/dev/sdh1,/dev/sdg1,/dev/sdd1,dev/sdc1
> mdadm: /dev/sdb is not one of 
> /dev/sdf1,/dev/sde1,/dev/sdh1,/dev/sdg1,/dev/sdd1,dev/sdc1
> mdadm: /dev/sda5 is not one of 
> /dev/sdf1,/dev/sde1,/dev/sdh1,/dev/sdg1,/dev/sdd1,dev/sdc1
> mdadm: /dev/sda2 is not one of 
> /dev/sdf1,/dev/sde1,/dev/sdh1,/dev/sdg1,/dev/sdd1,dev/sdc1
> mdadm: /dev/sda1 is not one of 
> /dev/sdf1,/dev/sde1,/dev/sdh1,/dev/sdg1,/dev/sdd1,dev/sdc1
> mdadm: /dev/sda is not one of 
> /dev/sdf1,/dev/sde1,/dev/sdh1,/dev/sdg1,/dev/sdd1,dev/sdc1
> 
> Not sure why it says uuid is wrong as it is set correctly in the conf
> 
> # mdadm.conf
> #
> # Please refer to mdadm.conf(5) for information about this file.
> #
> 
> # by default, scan all partitions (/proc/partitions) for MD superblocks.
> # alternatively, specify devices to scan, using wildcards if desired.
> DEVICE partitions
> 
> # auto-create devices with Debian standard permissions
> CREATE owner=root group=disk mode=0660 auto=yes
> 
> # automatically tag new arrays as belonging to the local system
> HOMEHOST <system>
> 
> # instruct the monitoring daemon where to send mail alerts
> MAILADDR root
> 
> # definitions of existing MD arrays
> ARRAY /dev/md0 level=raid5 num-devices=5
>    UUID=be282ff9:764d9beb:74ac4a35:dfcae213
>    devices=/dev/sdf1,/dev/sde1,/dev/sdh1,/dev/sdg1,/dev/sdd1,dev/sdc1
> 
> # This file was auto-generated on Thu, 12 Nov 2009 01:12:06 +0800
> # by mkconf $Id$
> 
> and here are details of my array
>  >> mdadm -E /dev/sd[cdefgh]1
> 
> /dev/sdc1:
>          Magic : a92b4efc
>        Version : 0.91.00
>           UUID : be282ff9:764d9beb:74ac4a35:dfcae213
>  Creation Time : Sun Mar 15 05:22:48 2009
>     Raid Level : raid5
>  Used Dev Size : 976759936 (931.51 GiB 1000.20 GB)
>     Array Size : 4883799680 (4657.55 GiB 5001.01 GB)
>   Raid Devices : 6
>  Total Devices : 6
> Preferred Minor : 0
> 
>  Reshape pos'n : 4156779840 (3964.21 GiB 4256.54 GB)
>  Delta Devices : 1 (5->6)
> 
>    Update Time : Mon Jan 18 00:00:51 2010
>          State : clean
> Active Devices : 5
> Working Devices : 6
> Failed Devices : 1
>  Spare Devices : 1
>       Checksum : f0ffe91a - correct
>         Events : 764121
> 
>         Layout : left-symmetric
>     Chunk Size : 64K
> 
>      Number   Major   Minor   RaidDevice State
> this     6       8       33        6      spare   /dev/sdc1
> 
>   0     0       8       81        0      active sync   /dev/sdf1
>   1     1       8       65        1      active sync   /dev/sde1
>   2     2       8      113        2      active sync   /dev/sdh1
>   3     3       8       97        3      active sync   /dev/sdg1
>   4     4       8       49        4      active sync   /dev/sdd1
>   5     5       0        0        5      faulty removed
>   6     6       8       33        6      spare   /dev/sdc1
> /dev/sdd1:
>          Magic : a92b4efc
>        Version : 0.91.00
>           UUID : be282ff9:764d9beb:74ac4a35:dfcae213
>  Creation Time : Sun Mar 15 05:22:48 2009
>     Raid Level : raid5
>  Used Dev Size : 976759936 (931.51 GiB 1000.20 GB)
>     Array Size : 4883799680 (4657.55 GiB 5001.01 GB)
>   Raid Devices : 6
>  Total Devices : 6
> Preferred Minor : 0
> 
>  Reshape pos'n : 4157438400 (3964.84 GiB 4257.22 GB)
>  Delta Devices : 1 (5->6)
> 
>    Update Time : Mon Jan 18 00:01:10 2010
>          State : active
> Active Devices : 5
> Working Devices : 5
> Failed Devices : 1
>  Spare Devices : 0
>       Checksum : f1085969 - correct
>         Events : 764126
> 
>         Layout : left-symmetric
>     Chunk Size : 64K
> 
>      Number   Major   Minor   RaidDevice State
> this     4       8       49        4      active sync   /dev/sdd1
> 
>   0     0       8       81        0      active sync   /dev/sdf1
>   1     1       8       65        1      active sync   /dev/sde1
>   2     2       8      113        2      active sync   /dev/sdh1
>   3     3       8       97        3      active sync   /dev/sdg1
>   4     4       8       49        4      active sync   /dev/sdd1
>   5     5       0        0        5      faulty removed
> /dev/sde1:
>          Magic : a92b4efc
>        Version : 0.91.00
>           UUID : be282ff9:764d9beb:74ac4a35:dfcae213
>  Creation Time : Sun Mar 15 05:22:48 2009
>     Raid Level : raid5
>  Used Dev Size : 976759936 (931.51 GiB 1000.20 GB)
>     Array Size : 4883799680 (4657.55 GiB 5001.01 GB)
>   Raid Devices : 6
>  Total Devices : 6
> Preferred Minor : 0
> 
>  Reshape pos'n : 4157438400 (3964.84 GiB 4257.22 GB)
>  Delta Devices : 1 (5->6)
> 
>    Update Time : Mon Jan 18 00:01:10 2010
>          State : active
> Active Devices : 5
> Working Devices : 5
> Failed Devices : 1
>  Spare Devices : 0
>       Checksum : f1085973 - correct
>         Events : 764126
> 
>         Layout : left-symmetric
>     Chunk Size : 64K
> 
>      Number   Major   Minor   RaidDevice State
> this     1       8       65        1      active sync   /dev/sde1
> 
>   0     0       8       81        0      active sync   /dev/sdf1
>   1     1       8       65        1      active sync   /dev/sde1
>   2     2       8      113        2      active sync   /dev/sdh1
>   3     3       8       97        3      active sync   /dev/sdg1
>   4     4       8       49        4      active sync   /dev/sdd1
>   5     5       0        0        5      faulty removed
> 
> Any help or ideas appreciated.
> 
> Thanks
> 
> Craig
> 


      reply	other threads:[~2010-01-23 11:38 UTC|newest]

Thread overview: 2+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2010-01-22 22:44 Raid5 Reshape failed can't assemble array Craig Haskins
2010-01-23 11:38 ` Craig Haskins [this message]

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=4B5ADFBB.9040009@gmail.com \
    --to=reptile55@gmail.com \
    --cc=linux-raid@vger.kernel.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).