linux-raid.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: jim@bwns.be
To: linux-raid@vger.kernel.org
Subject: RAID 10 problems, two disks marked as spare
Date: Fri, 28 Feb 2014 20:40:59 +0100	[thread overview]
Message-ID: <4ae13e60b43f187d9409032854fb119e@bwns.be> (raw)

Hi all,

I'm having a bit of trouble with my NAS and I was hoping some of you 
lads would be able to help my out.

First of all my setup:

The NAS itself is an Iomega StorCenter ix4-200d. It has 4x 1TiB drives, 
configured to use RAID10. I have already replaced one drive once. The 
NAS itself doesn't come with shell access but it's fairly easy to 'root' 
it anyway (which I did).

Several days ago the NAS send me an email that a certain drive was 
degraded:

> The Iomega StorCenter device is degraded and data protection is at 
> risk. A drive may have either failed or been removed from your Iomega 
> StorCenter device. Visit the Dashboard on the management interface for 
> details. To prevent possible data loss, this issue should be repaired 
> as soon as possible.

I decided to first to try and reboot the device, maybe it was a simple 
error. After reboot I received the following:

> Data protection is being reconstructed on your Iomega StorCenter device

So I was happy until several hours later I received the following 
messages (all at the same time)

> The Iomega StorCenter device has completed data protection 
> reconstruction.

> The Iomega StorCenter device has failed and some data loss may have 
> occurred. Multiple drives may have either failed or been removed from 
> your storage system. Visit the Dashboard on the management interface 
> for details.

> Drive number 4 encountered a recoverable error.

No data was accessible anymore. After that I opened a shell to the 
device and try to trouble shout it. But I didn’t manage to get it 
working. The only solution I currently see it to try and rebuild the 
RAID array but as I have hardly any experience with the mdadm tool I 
decided to ask the opinions of the people here.

Here is some information regarding the setup:

  root@BauwensNAS:/# mdadm -D /dev/md1
/dev/md1:
         Version : 01.00
   Creation Time : Mon Jan 24 20:57:43 2011
      Raid Level : raid10
   Used Dev Size : 974722176 (929.57 GiB 998.12 GB)
    Raid Devices : 4
   Total Devices : 4
Preferred Minor : 1
     Persistence : Superblock is persistent

     Update Time : Wed Feb 26 02:44:57 2014
           State : active, degraded, Not Started
  Active Devices : 2
Working Devices : 4
  Failed Devices : 0
   Spare Devices : 2

          Layout : near=2, far=1
      Chunk Size : 64K

            Name : bwns:1
            UUID : 43a1e240:956c3131:df9f6e66:bd9a071e
          Events : 133470

     Number   Major   Minor   RaidDevice State
        0       8        2        0      active sync   /dev/sda2
        4       8       18        1      active sync   /dev/sdb2
        2       0        0        2      removed
        3       0        0        3      removed

        2       8       34        -      spare   /dev/sdc2
        3       8       50        -      spare   /dev/sdd2

As you can see, the two last drives are marked as spare. My multiple 
attempts to get the drive running with all the disks have only been 
failures (but I assume that's also due to me not having experience with 
the tools).

Also, the disks themself appear to be fine (also because the md0 device 
that hosts /boot works properly)

Some more info:

root@BauwensNAS:/# mdadm --examine /dev/sd[abcd]2
/dev/sda2:
           Magic : a92b4efc
         Version : 1.0
     Feature Map : 0x0
      Array UUID : 43a1e240:956c3131:df9f6e66:bd9a071e
            Name : bwns:1
   Creation Time : Mon Jan 24 20:57:43 2011
      Raid Level : raid10
    Raid Devices : 4

  Avail Dev Size : 1949444384 (929.57 GiB 998.12 GB)
      Array Size : 3898888704 (1859.14 GiB 1996.23 GB)
   Used Dev Size : 1949444352 (929.57 GiB 998.12 GB)
    Super Offset : 1949444640 sectors
           State : clean
     Device UUID : b05f4c40:819ddbef:76872d9f:abacf3c9

     Update Time : Wed Feb 26 02:44:57 2014
        Checksum : a94e1ae6 - correct
          Events : 133470

          Layout : near=2, far=1
      Chunk Size : 64K

     Array Slot : 0 (0, failed, empty, empty, 1)
    Array State : Uu__ 1 failed
/dev/sdb2:
           Magic : a92b4efc
         Version : 1.0
     Feature Map : 0x0
      Array UUID : 43a1e240:956c3131:df9f6e66:bd9a071e
            Name : bwns:1
   Creation Time : Mon Jan 24 20:57:43 2011
      Raid Level : raid10
    Raid Devices : 4

  Avail Dev Size : 1949444384 (929.57 GiB 998.12 GB)
      Array Size : 3898888704 (1859.14 GiB 1996.23 GB)
   Used Dev Size : 1949444352 (929.57 GiB 998.12 GB)
    Super Offset : 1949444640 sectors
           State : clean
     Device UUID : 6c34331a:0fda7f73:a1f76d41:a826ac1f

     Update Time : Wed Feb 26 02:44:57 2014
        Checksum : fed2165a - correct
          Events : 133470

          Layout : near=2, far=1
      Chunk Size : 64K

     Array Slot : 4 (0, failed, empty, empty, 1)
    Array State : uU__ 1 failed
/dev/sdc2:
           Magic : a92b4efc
         Version : 1.0
     Feature Map : 0x0
      Array UUID : 43a1e240:956c3131:df9f6e66:bd9a071e
            Name : bwns:1
   Creation Time : Mon Jan 24 20:57:43 2011
      Raid Level : raid10
    Raid Devices : 4

  Avail Dev Size : 1949444384 (929.57 GiB 998.12 GB)
      Array Size : 3898888704 (1859.14 GiB 1996.23 GB)
   Used Dev Size : 1949444352 (929.57 GiB 998.12 GB)
    Super Offset : 1949444640 sectors
           State : clean
     Device UUID : 773dbfec:07467e62:de7be59b:5c680df5

     Update Time : Wed Feb 26 02:44:57 2014
        Checksum : f035517e - correct
          Events : 133470

          Layout : near=2, far=1
      Chunk Size : 64K

     Array Slot : 2 (0, failed, empty, empty, 1)
    Array State : uu__ 1 failed
/dev/sdd2:
           Magic : a92b4efc
         Version : 1.0
     Feature Map : 0x0
      Array UUID : 43a1e240:956c3131:df9f6e66:bd9a071e
            Name : bwns:1
   Creation Time : Mon Jan 24 20:57:43 2011
      Raid Level : raid10
    Raid Devices : 4

  Avail Dev Size : 1949444384 (929.57 GiB 998.12 GB)
      Array Size : 3898888704 (1859.14 GiB 1996.23 GB)
   Used Dev Size : 1949444352 (929.57 GiB 998.12 GB)
    Super Offset : 1949444640 sectors
           State : clean
     Device UUID : dbd27546:6b623b53:8f887960:b7cbf424

     Update Time : Wed Feb 26 02:44:57 2014
        Checksum : 2f247322 - correct
          Events : 133470

          Layout : near=2, far=1
      Chunk Size : 64K

     Array Slot : 3 (0, failed, empty, empty, 1)
    Array State : uu__ 1 failed


Taking a look at the event count they seem to be synchronized, so I'm 
not really sure what's going on here.

root@BauwensNAS:/# cat /proc/mdstat
Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] 
[raid4]
md1 : inactive sda2[0] sdd2[3](S) sdc2[2](S) sdb2[4]
       3898888704 blocks super 1.0

md0 : active raid1 sda1[0] sdd1[3] sdc1[2] sdb1[1]
       2040128 blocks [4/4] [UUUU]

unused devices: <none>

Anyone have an idea how I could resolve this problem (hoping that I 
don't have any data loss...)? Any help is greatly appreciated. I sure 
regret rebooting the device without taking some extra backups.

TIA!
Jim
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

             reply	other threads:[~2014-02-28 19:40 UTC|newest]

Thread overview: 2+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2014-02-28 19:40 jim [this message]
2014-03-06 10:11 ` RAID 10 problems, two disks marked as spare Jim Bauwens

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=4ae13e60b43f187d9409032854fb119e@bwns.be \
    --to=jim@bwns.be \
    --cc=linux-raid@vger.kernel.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).