linux-raid.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: "\"Jörg Habenicht\"" <j.habenicht@gmx.de>
To: linux-raid@vger.kernel.org
Subject: Followup: how to restore a raid5 with 1 disk destroyed and 1 kicked out?
Date: Tue, 07 Apr 2009 16:30:38 +0200	[thread overview]
Message-ID: <20090407143038.140430@gmx.net> (raw)
In-Reply-To: <20090407140225.140400@gmx.net>

http://www.mail-archive.com/linux-raid@vger.kernel.org/msg07640.html suggests sending the event count:

The event count of the devices is
/dev/hda1: 0.3088065
/dev/sdb1: 0.3088063 * (out of sync)
/dev/sdc1: 0.3088065
/dev/sdd1: 0.3088062 ** (dropped dead)
/dev/sde1: 0.3088065
/dev/sdf1: 0.3088065


cu
Jörg

-------- Original-Nachricht --------
> Datum: Tue, 07 Apr 2009 16:02:25 +0200
> Von: "Jörg Habenicht" <j.habenicht@gmx.de>
> An: linux-raid@vger.kernel.org
> Betreff: how to restore a raid5 with 1 disk destroyed and 1 kicked out?

> Hello list, hello Neil,
> 
> I hope you may help me with this one:
> 
> During a RAID5 synchronisation with one faulty disk my server crashed and
> left my RAID in an unsynced state. I'd like to get the content from the
> array back to freshen my last backup (4 months ago) and then build the array
> anew.
> 
> The array consists of 6 disks, right now 1 is dead (hw failure) and one is
> "out of sync". I assume the latter is just marked out of sync without
> being so.(*)
> 
> 
> In http://www.mail-archive.com/linux-raid@vger.kernel.org/msg08909.html
> and http://www.mail-archive.com/linux-raid@vger.kernel.org/msg06332.html you
> suggested to recreate the array with --assume-clean.
> In http://www.mail-archive.com/linux-raid@vger.kernel.org/msg08162.html
> you advised not to use --assume-clean for RAID5, it may very well break.
> Is it ok to use --assume-clean on a degraded array (5 out of 6, RAID5)?
> 
> 
> Would it be better to recreate the array without --assume-clean? (E.g.
> like "mdadm --create /dev/md0 disk1 disk2 disk3 missing disk5 disk6")
> 
> Just out of interest: What is the difference between the two commands?
> 
> I think /deb/sdb1 is marked "out of sync" without being so. Is
> "--assume-clean" used exactly for this case?
> 
> 
> Thank you in advance for any advice or help
> cu
> Jörg
> 
> 
> 
> (*)
> /dev/sdd1 dropped dead
> /dev/sdb1 is marked 'out of sync', but I think the content on the disk is
> in sync with the array
> 
> 
>  Now the dirty details:
> 
> ~ # mdadm -S /dev/md0
> mdadm: stopped /dev/md0
> 
> ~ # mdadm --assemble --force /dev/md0  --run --verbose  /dev/sdb1
> /dev/hda1 /dev/sdc1  /dev/sde1  /dev/sdf1
> mdadm: looking for devices for /dev/md0
> mdadm: /dev/sdb1 is identified as a member of /dev/md0, slot 6.
> mdadm: /dev/hda1 is identified as a member of /dev/md0, slot 1.
> mdadm: /dev/sdc1 is identified as a member of /dev/md0, slot 2.
> mdadm: /dev/sde1 is identified as a member of /dev/md0, slot 3.
> mdadm: /dev/sdf1 is identified as a member of /dev/md0, slot 5.
> mdadm: no uptodate device for slot 0 of /dev/md0
> mdadm: added /dev/sdc1 to /dev/md0 as 2
> mdadm: added /dev/sde1 to /dev/md0 as 3
> mdadm: no uptodate device for slot 4 of /dev/md0
> mdadm: added /dev/sdf1 to /dev/md0 as 5
> mdadm: added /dev/sdb1 to /dev/md0 as 6
> mdadm: added /dev/hda1 to /dev/md0 as 1
> mdadm: failed to RUN_ARRAY /dev/md0: Input/output error
> mdadm: Not enough devices to start the array.
> 
> ~ # cat /proc/mdstat
> Personalities : [raid6] [raid5] [raid4] [raid1]
> md0 : inactive hda1[1] sdf1[5] sde1[3] sdc1[2]
>       976791680 blocks
> 
> ~ # mdadm -D /dev/md0
> /dev/md0:
>         Version : 00.90.03
>   Creation Time : Mon Apr  3 12:35:48 2006
>      Raid Level : raid5
>   Used Dev Size : 244195904 (232.88 GiB 250.06 GB)
>    Raid Devices : 6
>   Total Devices : 4
> Preferred Minor : 0
>     Persistence : Superblock is persistent
> 
>     Update Time : Tue Apr  7 12:02:58 2009
>           State : active, degraded, Not Started
>  Active Devices : 4
> Working Devices : 4
>  Failed Devices : 0
>   Spare Devices : 0
> 
>          Layout : left-symmetric
>      Chunk Size : 64K
> 
>            UUID : b72d31b8:f6bbac3d:c1c586ef:bb458af6
>          Events : 0.3088065
> 
>     Number   Major   Minor   RaidDevice State
>        0       0        0        0      removed
>        1       3        1        1      active sync   /dev/hda1
>        2       8       33        2      active sync   /dev/sdc1
>        3       8       65        3      active sync   /dev/sde1
>        4       0        0        4      removed
>        5       8       81        5      active sync   /dev/sdf1
> 
> ~ # mdadm -IR /dev/sdb1
> mdadm: /dev/sdb1 attached to /dev/md0, not enough to start (4).
> 
> ~ # cat /proc/mdstat
> Personalities : [raid6] [raid5] [raid4] [raid1]
> md0 : inactive sdb1[0] hda1[1] sdf1[5] sde1[3] sdc1[2]
>       1220987584 blocks
> 
> ~ # mdadm -D /dev/md0
> (.....)
>     Number   Major   Minor   RaidDevice State
>        0       8       17        0      spare rebuilding   /dev/sdb1
>        1       3        1        1      active sync   /dev/hda1
>        2       8       33        2      active sync   /dev/sdc1
>        3       8       65        3      active sync   /dev/sde1
>        4       0        0        4      removed
>        5       8       81        5      active sync   /dev/sdf1
> 
> 
> 
> I already tried
> - "mdadm --assemble /dev/md0 /dev/sdb1 /dev/hda1 /dev/sdc1 /dev/sde1 
> /dev/sdf1"
> - "mdadm --assemble --force /dev/md0 /dev/sdb1 /dev/hda1 /dev/sdc1 
> /dev/sde1  /dev/sdf1"
> 
> 
> 
> -- 
> cu,
> Joerg
> 
> --
> THE full automatic planets host
> :-)                            http://planets.unix-ag.uni-hannover.de
> 
> 
> 
> Neu: GMX FreeDSL Komplettanschluss mit DSL 6.000 Flatrate +
> Telefonanschluss für nur 17,95 Euro/mtl.!*
> http://dslspecial.gmx.de/freedsl-surfflat/?ac=OM.AD.PD003K11308T4569a
> --
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

-- 
cu,
Joerg

--
THE full automatic planets host
:-)                            http://planets.unix-ag.uni-hannover.de



Psssst! Schon vom neuen GMX MultiMessenger gehört? Der kann`s mit allen: http://www.gmx.net/de/go/multimessenger01
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

      reply	other threads:[~2009-04-07 14:30 UTC|newest]

Thread overview: 2+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2009-04-07 14:02 how to restore a raid5 with 1 disk destroyed and 1 kicked out? "Jörg Habenicht"
2009-04-07 14:30 ` "Jörg Habenicht" [this message]

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20090407143038.140430@gmx.net \
    --to=j.habenicht@gmx.de \
    --cc=linux-raid@vger.kernel.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).