From: jim@bwns.be
To: linux-raid@vger.kernel.org
Subject: RAID 10 problems, two disks marked as spare
Date: Fri, 28 Feb 2014 20:40:59 +0100 [thread overview]
Message-ID: <4ae13e60b43f187d9409032854fb119e@bwns.be> (raw)
Hi all,
I'm having a bit of trouble with my NAS and I was hoping some of you
lads would be able to help my out.
First of all my setup:
The NAS itself is an Iomega StorCenter ix4-200d. It has 4x 1TiB drives,
configured to use RAID10. I have already replaced one drive once. The
NAS itself doesn't come with shell access but it's fairly easy to 'root'
it anyway (which I did).
Several days ago the NAS send me an email that a certain drive was
degraded:
> The Iomega StorCenter device is degraded and data protection is at
> risk. A drive may have either failed or been removed from your Iomega
> StorCenter device. Visit the Dashboard on the management interface for
> details. To prevent possible data loss, this issue should be repaired
> as soon as possible.
I decided to first to try and reboot the device, maybe it was a simple
error. After reboot I received the following:
> Data protection is being reconstructed on your Iomega StorCenter device
So I was happy until several hours later I received the following
messages (all at the same time)
> The Iomega StorCenter device has completed data protection
> reconstruction.
> The Iomega StorCenter device has failed and some data loss may have
> occurred. Multiple drives may have either failed or been removed from
> your storage system. Visit the Dashboard on the management interface
> for details.
> Drive number 4 encountered a recoverable error.
No data was accessible anymore. After that I opened a shell to the
device and try to trouble shout it. But I didn’t manage to get it
working. The only solution I currently see it to try and rebuild the
RAID array but as I have hardly any experience with the mdadm tool I
decided to ask the opinions of the people here.
Here is some information regarding the setup:
root@BauwensNAS:/# mdadm -D /dev/md1
/dev/md1:
Version : 01.00
Creation Time : Mon Jan 24 20:57:43 2011
Raid Level : raid10
Used Dev Size : 974722176 (929.57 GiB 998.12 GB)
Raid Devices : 4
Total Devices : 4
Preferred Minor : 1
Persistence : Superblock is persistent
Update Time : Wed Feb 26 02:44:57 2014
State : active, degraded, Not Started
Active Devices : 2
Working Devices : 4
Failed Devices : 0
Spare Devices : 2
Layout : near=2, far=1
Chunk Size : 64K
Name : bwns:1
UUID : 43a1e240:956c3131:df9f6e66:bd9a071e
Events : 133470
Number Major Minor RaidDevice State
0 8 2 0 active sync /dev/sda2
4 8 18 1 active sync /dev/sdb2
2 0 0 2 removed
3 0 0 3 removed
2 8 34 - spare /dev/sdc2
3 8 50 - spare /dev/sdd2
As you can see, the two last drives are marked as spare. My multiple
attempts to get the drive running with all the disks have only been
failures (but I assume that's also due to me not having experience with
the tools).
Also, the disks themself appear to be fine (also because the md0 device
that hosts /boot works properly)
Some more info:
root@BauwensNAS:/# mdadm --examine /dev/sd[abcd]2
/dev/sda2:
Magic : a92b4efc
Version : 1.0
Feature Map : 0x0
Array UUID : 43a1e240:956c3131:df9f6e66:bd9a071e
Name : bwns:1
Creation Time : Mon Jan 24 20:57:43 2011
Raid Level : raid10
Raid Devices : 4
Avail Dev Size : 1949444384 (929.57 GiB 998.12 GB)
Array Size : 3898888704 (1859.14 GiB 1996.23 GB)
Used Dev Size : 1949444352 (929.57 GiB 998.12 GB)
Super Offset : 1949444640 sectors
State : clean
Device UUID : b05f4c40:819ddbef:76872d9f:abacf3c9
Update Time : Wed Feb 26 02:44:57 2014
Checksum : a94e1ae6 - correct
Events : 133470
Layout : near=2, far=1
Chunk Size : 64K
Array Slot : 0 (0, failed, empty, empty, 1)
Array State : Uu__ 1 failed
/dev/sdb2:
Magic : a92b4efc
Version : 1.0
Feature Map : 0x0
Array UUID : 43a1e240:956c3131:df9f6e66:bd9a071e
Name : bwns:1
Creation Time : Mon Jan 24 20:57:43 2011
Raid Level : raid10
Raid Devices : 4
Avail Dev Size : 1949444384 (929.57 GiB 998.12 GB)
Array Size : 3898888704 (1859.14 GiB 1996.23 GB)
Used Dev Size : 1949444352 (929.57 GiB 998.12 GB)
Super Offset : 1949444640 sectors
State : clean
Device UUID : 6c34331a:0fda7f73:a1f76d41:a826ac1f
Update Time : Wed Feb 26 02:44:57 2014
Checksum : fed2165a - correct
Events : 133470
Layout : near=2, far=1
Chunk Size : 64K
Array Slot : 4 (0, failed, empty, empty, 1)
Array State : uU__ 1 failed
/dev/sdc2:
Magic : a92b4efc
Version : 1.0
Feature Map : 0x0
Array UUID : 43a1e240:956c3131:df9f6e66:bd9a071e
Name : bwns:1
Creation Time : Mon Jan 24 20:57:43 2011
Raid Level : raid10
Raid Devices : 4
Avail Dev Size : 1949444384 (929.57 GiB 998.12 GB)
Array Size : 3898888704 (1859.14 GiB 1996.23 GB)
Used Dev Size : 1949444352 (929.57 GiB 998.12 GB)
Super Offset : 1949444640 sectors
State : clean
Device UUID : 773dbfec:07467e62:de7be59b:5c680df5
Update Time : Wed Feb 26 02:44:57 2014
Checksum : f035517e - correct
Events : 133470
Layout : near=2, far=1
Chunk Size : 64K
Array Slot : 2 (0, failed, empty, empty, 1)
Array State : uu__ 1 failed
/dev/sdd2:
Magic : a92b4efc
Version : 1.0
Feature Map : 0x0
Array UUID : 43a1e240:956c3131:df9f6e66:bd9a071e
Name : bwns:1
Creation Time : Mon Jan 24 20:57:43 2011
Raid Level : raid10
Raid Devices : 4
Avail Dev Size : 1949444384 (929.57 GiB 998.12 GB)
Array Size : 3898888704 (1859.14 GiB 1996.23 GB)
Used Dev Size : 1949444352 (929.57 GiB 998.12 GB)
Super Offset : 1949444640 sectors
State : clean
Device UUID : dbd27546:6b623b53:8f887960:b7cbf424
Update Time : Wed Feb 26 02:44:57 2014
Checksum : 2f247322 - correct
Events : 133470
Layout : near=2, far=1
Chunk Size : 64K
Array Slot : 3 (0, failed, empty, empty, 1)
Array State : uu__ 1 failed
Taking a look at the event count they seem to be synchronized, so I'm
not really sure what's going on here.
root@BauwensNAS:/# cat /proc/mdstat
Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5]
[raid4]
md1 : inactive sda2[0] sdd2[3](S) sdc2[2](S) sdb2[4]
3898888704 blocks super 1.0
md0 : active raid1 sda1[0] sdd1[3] sdc1[2] sdb1[1]
2040128 blocks [4/4] [UUUU]
unused devices: <none>
Anyone have an idea how I could resolve this problem (hoping that I
don't have any data loss...)? Any help is greatly appreciated. I sure
regret rebooting the device without taking some extra backups.
TIA!
Jim
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
next reply other threads:[~2014-02-28 19:40 UTC|newest]
Thread overview: 2+ messages / expand[flat|nested] mbox.gz Atom feed top
2014-02-28 19:40 jim [this message]
2014-03-06 10:11 ` RAID 10 problems, two disks marked as spare Jim Bauwens
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=4ae13e60b43f187d9409032854fb119e@bwns.be \
--to=jim@bwns.be \
--cc=linux-raid@vger.kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).