From: NeilBrown <neilb@suse.de>
To: Jeff Johnson <jeff.johnson@aeoncomputing.com>
Cc: linux-raid@vger.kernel.org
Subject: Re: 20 drive raid-10, CentOS5.5, after reboot assemble fails - all drives "non-fresh"
Date: Mon, 8 Aug 2011 12:56:08 +1000 [thread overview]
Message-ID: <20110808125608.7ceeaa02@notabene.brown> (raw)
In-Reply-To: <4E3F4BD0.5060807@aeoncomputing.com>
On Sun, 07 Aug 2011 19:37:04 -0700 Jeff Johnson
<jeff.johnson@aeoncomputing.com> wrote:
> Greetings,
>
> I have a 20 drive raid-10 that has been running well for over one year.
> After the most recently system boot the raid will not assemble.
> /var/log/messages shows that all of the drives are "non-fresh".
> Examining the drives show that the raid partitions are present, the
> superblocks have valid data and the Event ticker for the drives are
> equal for the data drives. The spare drives have a different Event
> ticker value.
>
> I am reluctant to try and use the --force switch with assemble until I
> understand the problem better. There is very important data on this
> volume and it is not backed up to my knowledge. I do not know how the
> machine was brought down prior to this system boot.
>
> With all drives being "non-fresh" I can't start a partial array and then
> re-add the remaining drives. I've unraveled some pretty messed up md
> configs and recovered the underlying filesystem but this one has me at a
> loss.
>
> Any advice is greatly appreciated!
>
> --Jeff
>
> Below is the config file and output from mdadm examine commands:
>
> /* Config file */
> ARRAY /dev/md3 level=raid10 num-devices=20
> UUID=e17a29e8:ec6bce5c:f13d343c:cfba4dc4
> spares=4
> devices=/dev/sdz1,/dev/sdy1,/dev/sdx1,/dev/sdw1,/dev/sdv1,/dev/sdu1,/dev/sdt1,/dev/sd
> s1,/dev/sdr1,/dev/sdq1,/dev/sdp1,/dev/sdo1,/dev/sdn1,/dev/sdm1,/dev/sdl1,/dev/sdk1,/dev/sdj1,/dev/s
> di1,/dev/sdh1,/dev/sdg1,/dev/sdf1,/dev/sde1,/dev/sdd1,/dev/sdc1
You really don't want that 'devices=" clause in there. Device names can
change...
>
> /* mdadm -E /dev/sd[cdefghijklmnopqrstuvwxyz]1 | grep Event */
> Events : 90
> Events : 90
> Events : 90
> Events : 90
> Events : 90
> Events : 90
> Events : 90
> Events : 90
> Events : 90
> Events : 90
> Events : 90
> Events : 90
> Events : 90
> Events : 90
> Events : 90
> Events : 90
> Events : 90
> Events : 90
> Events : 90
> Events : 90
> Events : 92
> Events : 92
> Events : 92
> Events : 92
So the spares are '92' and the others are '90'. That is weird...
However you should be able to assemble the array by simply listing all the
non-spare devices:
mdadm -A /dev/md3 /dev/sd[c-v]1
NeilBrown
>
> /* mdadm -E /dev/sdc1 */
> /dev/sdc1:
> Magic : a92b4efc
> Version : 0.90.00
> UUID : e17a29e8:ec6bce5c:f13d343c:cfba4dc4
> Creation Time : Fri Sep 24 12:06:37 2010
> Raid Level : raid10
> Used Dev Size : 99924096 (95.30 GiB 102.32 GB)
> Array Size : 999240960 (952.95 GiB 1023.22 GB)
> Raid Devices : 20
> Total Devices : 24
> Preferred Minor : 3
>
> Update Time : Sat Aug 6 05:54:37 2011
> State : clean
> Active Devices : 20
> Working Devices : 24
> Failed Devices : 0
> Spare Devices : 4
> Checksum : d8d97049 - correct
> Events : 90
>
> Layout : near=2
> Chunk Size : 128K
>
> Number Major Minor RaidDevice State
> this 0 8 33 0 active sync /dev/sdc1
>
> 0 0 8 33 0 active sync /dev/sdc1
> 1 1 8 49 1 active sync /dev/sdd1
> 2 2 8 65 2 active sync /dev/sde1
> 3 3 8 81 3 active sync /dev/sdf1
> 4 4 8 97 4 active sync /dev/sdg1
> 5 5 8 113 5 active sync /dev/sdh1
> 6 6 8 129 6 active sync /dev/sdi1
> 7 7 8 145 7 active sync /dev/sdj1
> 8 8 8 161 8 active sync /dev/sdk1
> 9 9 8 177 9 active sync /dev/sdl1
> 10 10 8 193 10 active sync /dev/sdm1
> 11 11 8 209 11 active sync /dev/sdn1
> 12 12 8 225 12 active sync /dev/sdo1
> 13 13 8 241 13 active sync /dev/sdp1
> 14 14 65 1 14 active sync /dev/sdq1
> 15 15 65 17 15 active sync /dev/sdr1
> 16 16 65 33 16 active sync /dev/sds1
> 17 17 65 49 17 active sync /dev/sdt1
> 18 18 65 65 18 active sync /dev/sdu1
> 19 19 65 81 19 active sync /dev/sdv1
> 20 20 65 145 20 spare /dev/sdz1
> 21 21 65 129 21 spare /dev/sdy1
> 22 22 65 113 22 spare /dev/sdx1
> 23 23 65 97 23 spare /dev/sdw1
>
next prev parent reply other threads:[~2011-08-08 2:56 UTC|newest]
Thread overview: 8+ messages / expand[flat|nested] mbox.gz Atom feed top
2011-08-08 2:37 20 drive raid-10, CentOS5.5, after reboot assemble fails - all drives "non-fresh" Jeff Johnson
2011-08-08 2:56 ` NeilBrown [this message]
2011-08-08 4:32 ` Jeff Johnson
2011-08-08 4:40 ` Joe Landman
2011-08-08 4:54 ` Jeff Johnson
2011-08-08 5:04 ` Joe Landman
2011-08-08 5:55 ` Jeff Johnson
2011-08-08 16:17 ` Joe Landman
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20110808125608.7ceeaa02@notabene.brown \
--to=neilb@suse.de \
--cc=jeff.johnson@aeoncomputing.com \
--cc=linux-raid@vger.kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox