linux-raid.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
* RAID 10 problems, two disks marked as spare
@ 2014-02-28 19:40 jim
  2014-03-06 10:11 ` Jim Bauwens
  0 siblings, 1 reply; 2+ messages in thread
From: jim @ 2014-02-28 19:40 UTC (permalink / raw)
  To: linux-raid

Hi all,

I'm having a bit of trouble with my NAS and I was hoping some of you 
lads would be able to help my out.

First of all my setup:

The NAS itself is an Iomega StorCenter ix4-200d. It has 4x 1TiB drives, 
configured to use RAID10. I have already replaced one drive once. The 
NAS itself doesn't come with shell access but it's fairly easy to 'root' 
it anyway (which I did).

Several days ago the NAS send me an email that a certain drive was 
degraded:

> The Iomega StorCenter device is degraded and data protection is at 
> risk. A drive may have either failed or been removed from your Iomega 
> StorCenter device. Visit the Dashboard on the management interface for 
> details. To prevent possible data loss, this issue should be repaired 
> as soon as possible.

I decided to first to try and reboot the device, maybe it was a simple 
error. After reboot I received the following:

> Data protection is being reconstructed on your Iomega StorCenter device

So I was happy until several hours later I received the following 
messages (all at the same time)

> The Iomega StorCenter device has completed data protection 
> reconstruction.

> The Iomega StorCenter device has failed and some data loss may have 
> occurred. Multiple drives may have either failed or been removed from 
> your storage system. Visit the Dashboard on the management interface 
> for details.

> Drive number 4 encountered a recoverable error.

No data was accessible anymore. After that I opened a shell to the 
device and try to trouble shout it. But I didn’t manage to get it 
working. The only solution I currently see it to try and rebuild the 
RAID array but as I have hardly any experience with the mdadm tool I 
decided to ask the opinions of the people here.

Here is some information regarding the setup:

  root@BauwensNAS:/# mdadm -D /dev/md1
/dev/md1:
         Version : 01.00
   Creation Time : Mon Jan 24 20:57:43 2011
      Raid Level : raid10
   Used Dev Size : 974722176 (929.57 GiB 998.12 GB)
    Raid Devices : 4
   Total Devices : 4
Preferred Minor : 1
     Persistence : Superblock is persistent

     Update Time : Wed Feb 26 02:44:57 2014
           State : active, degraded, Not Started
  Active Devices : 2
Working Devices : 4
  Failed Devices : 0
   Spare Devices : 2

          Layout : near=2, far=1
      Chunk Size : 64K

            Name : bwns:1
            UUID : 43a1e240:956c3131:df9f6e66:bd9a071e
          Events : 133470

     Number   Major   Minor   RaidDevice State
        0       8        2        0      active sync   /dev/sda2
        4       8       18        1      active sync   /dev/sdb2
        2       0        0        2      removed
        3       0        0        3      removed

        2       8       34        -      spare   /dev/sdc2
        3       8       50        -      spare   /dev/sdd2

As you can see, the two last drives are marked as spare. My multiple 
attempts to get the drive running with all the disks have only been 
failures (but I assume that's also due to me not having experience with 
the tools).

Also, the disks themself appear to be fine (also because the md0 device 
that hosts /boot works properly)

Some more info:

root@BauwensNAS:/# mdadm --examine /dev/sd[abcd]2
/dev/sda2:
           Magic : a92b4efc
         Version : 1.0
     Feature Map : 0x0
      Array UUID : 43a1e240:956c3131:df9f6e66:bd9a071e
            Name : bwns:1
   Creation Time : Mon Jan 24 20:57:43 2011
      Raid Level : raid10
    Raid Devices : 4

  Avail Dev Size : 1949444384 (929.57 GiB 998.12 GB)
      Array Size : 3898888704 (1859.14 GiB 1996.23 GB)
   Used Dev Size : 1949444352 (929.57 GiB 998.12 GB)
    Super Offset : 1949444640 sectors
           State : clean
     Device UUID : b05f4c40:819ddbef:76872d9f:abacf3c9

     Update Time : Wed Feb 26 02:44:57 2014
        Checksum : a94e1ae6 - correct
          Events : 133470

          Layout : near=2, far=1
      Chunk Size : 64K

     Array Slot : 0 (0, failed, empty, empty, 1)
    Array State : Uu__ 1 failed
/dev/sdb2:
           Magic : a92b4efc
         Version : 1.0
     Feature Map : 0x0
      Array UUID : 43a1e240:956c3131:df9f6e66:bd9a071e
            Name : bwns:1
   Creation Time : Mon Jan 24 20:57:43 2011
      Raid Level : raid10
    Raid Devices : 4

  Avail Dev Size : 1949444384 (929.57 GiB 998.12 GB)
      Array Size : 3898888704 (1859.14 GiB 1996.23 GB)
   Used Dev Size : 1949444352 (929.57 GiB 998.12 GB)
    Super Offset : 1949444640 sectors
           State : clean
     Device UUID : 6c34331a:0fda7f73:a1f76d41:a826ac1f

     Update Time : Wed Feb 26 02:44:57 2014
        Checksum : fed2165a - correct
          Events : 133470

          Layout : near=2, far=1
      Chunk Size : 64K

     Array Slot : 4 (0, failed, empty, empty, 1)
    Array State : uU__ 1 failed
/dev/sdc2:
           Magic : a92b4efc
         Version : 1.0
     Feature Map : 0x0
      Array UUID : 43a1e240:956c3131:df9f6e66:bd9a071e
            Name : bwns:1
   Creation Time : Mon Jan 24 20:57:43 2011
      Raid Level : raid10
    Raid Devices : 4

  Avail Dev Size : 1949444384 (929.57 GiB 998.12 GB)
      Array Size : 3898888704 (1859.14 GiB 1996.23 GB)
   Used Dev Size : 1949444352 (929.57 GiB 998.12 GB)
    Super Offset : 1949444640 sectors
           State : clean
     Device UUID : 773dbfec:07467e62:de7be59b:5c680df5

     Update Time : Wed Feb 26 02:44:57 2014
        Checksum : f035517e - correct
          Events : 133470

          Layout : near=2, far=1
      Chunk Size : 64K

     Array Slot : 2 (0, failed, empty, empty, 1)
    Array State : uu__ 1 failed
/dev/sdd2:
           Magic : a92b4efc
         Version : 1.0
     Feature Map : 0x0
      Array UUID : 43a1e240:956c3131:df9f6e66:bd9a071e
            Name : bwns:1
   Creation Time : Mon Jan 24 20:57:43 2011
      Raid Level : raid10
    Raid Devices : 4

  Avail Dev Size : 1949444384 (929.57 GiB 998.12 GB)
      Array Size : 3898888704 (1859.14 GiB 1996.23 GB)
   Used Dev Size : 1949444352 (929.57 GiB 998.12 GB)
    Super Offset : 1949444640 sectors
           State : clean
     Device UUID : dbd27546:6b623b53:8f887960:b7cbf424

     Update Time : Wed Feb 26 02:44:57 2014
        Checksum : 2f247322 - correct
          Events : 133470

          Layout : near=2, far=1
      Chunk Size : 64K

     Array Slot : 3 (0, failed, empty, empty, 1)
    Array State : uu__ 1 failed


Taking a look at the event count they seem to be synchronized, so I'm 
not really sure what's going on here.

root@BauwensNAS:/# cat /proc/mdstat
Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] 
[raid4]
md1 : inactive sda2[0] sdd2[3](S) sdc2[2](S) sdb2[4]
       3898888704 blocks super 1.0

md0 : active raid1 sda1[0] sdd1[3] sdc1[2] sdb1[1]
       2040128 blocks [4/4] [UUUU]

unused devices: <none>

Anyone have an idea how I could resolve this problem (hoping that I 
don't have any data loss...)? Any help is greatly appreciated. I sure 
regret rebooting the device without taking some extra backups.

TIA!
Jim
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 2+ messages in thread

* Re: RAID 10 problems, two disks marked as spare
  2014-02-28 19:40 RAID 10 problems, two disks marked as spare jim
@ 2014-03-06 10:11 ` Jim Bauwens
  0 siblings, 0 replies; 2+ messages in thread
From: Jim Bauwens @ 2014-03-06 10:11 UTC (permalink / raw)
  To: linux-raid

Hi all,

I decided to take the plunge and recreate the array with "mdadm --create --assume-clean --level=10 --raid-devices=4 --size=974722176 /dev/md1 /dev/sda2 /dev/sdb2 /dev/sdc2 /dev/sdd2”. The array is working correctly again, however for now I mounted it read only and I’m going to take some backups to another raid system just in case something terribly goes wrong (I have to reboot the device so that the Iomega NAS system can detect the drives again).

Greetings and thanks,
Jim



On 28 Feb 2014, at 20:40, jim@bwns.be wrote:

> Hi all,
> 
> I'm having a bit of trouble with my NAS and I was hoping some of you lads would be able to help my out.
> 
> First of all my setup:
> 
> The NAS itself is an Iomega StorCenter ix4-200d. It has 4x 1TiB drives, configured to use RAID10. I have already replaced one drive once. The NAS itself doesn't come with shell access but it's fairly easy to 'root' it anyway (which I did).
> 
> Several days ago the NAS send me an email that a certain drive was degraded:
> 
>> The Iomega StorCenter device is degraded and data protection is at risk. A drive may have either failed or been removed from your Iomega StorCenter device. Visit the Dashboard on the management interface for details. To prevent possible data loss, this issue should be repaired as soon as possible.
> 
> I decided to first to try and reboot the device, maybe it was a simple error. After reboot I received the following:
> 
>> Data protection is being reconstructed on your Iomega StorCenter device
> 
> So I was happy until several hours later I received the following messages (all at the same time)
> 
>> The Iomega StorCenter device has completed data protection reconstruction.
> 
>> The Iomega StorCenter device has failed and some data loss may have occurred. Multiple drives may have either failed or been removed from your storage system. Visit the Dashboard on the management interface for details.
> 
>> Drive number 4 encountered a recoverable error.
> 
> No data was accessible anymore. After that I opened a shell to the device and try to trouble shout it. But I didn’t manage to get it working. The only solution I currently see it to try and rebuild the RAID array but as I have hardly any experience with the mdadm tool I decided to ask the opinions of the people here.
> 
> Here is some information regarding the setup:
> 
> root@BauwensNAS:/# mdadm -D /dev/md1
> /dev/md1:
>        Version : 01.00
>  Creation Time : Mon Jan 24 20:57:43 2011
>     Raid Level : raid10
>  Used Dev Size : 974722176 (929.57 GiB 998.12 GB)
>   Raid Devices : 4
>  Total Devices : 4
> Preferred Minor : 1
>    Persistence : Superblock is persistent
> 
>    Update Time : Wed Feb 26 02:44:57 2014
>          State : active, degraded, Not Started
> Active Devices : 2
> Working Devices : 4
> Failed Devices : 0
>  Spare Devices : 2
> 
>         Layout : near=2, far=1
>     Chunk Size : 64K
> 
>           Name : bwns:1
>           UUID : 43a1e240:956c3131:df9f6e66:bd9a071e
>         Events : 133470
> 
>    Number   Major   Minor   RaidDevice State
>       0       8        2        0      active sync   /dev/sda2
>       4       8       18        1      active sync   /dev/sdb2
>       2       0        0        2      removed
>       3       0        0        3      removed
> 
>       2       8       34        -      spare   /dev/sdc2
>       3       8       50        -      spare   /dev/sdd2
> 
> As you can see, the two last drives are marked as spare. My multiple attempts to get the drive running with all the disks have only been failures (but I assume that's also due to me not having experience with the tools).
> 
> Also, the disks themself appear to be fine (also because the md0 device that hosts /boot works properly)
> 
> Some more info:
> 
> root@BauwensNAS:/# mdadm --examine /dev/sd[abcd]2
> /dev/sda2:
>          Magic : a92b4efc
>        Version : 1.0
>    Feature Map : 0x0
>     Array UUID : 43a1e240:956c3131:df9f6e66:bd9a071e
>           Name : bwns:1
>  Creation Time : Mon Jan 24 20:57:43 2011
>     Raid Level : raid10
>   Raid Devices : 4
> 
> Avail Dev Size : 1949444384 (929.57 GiB 998.12 GB)
>     Array Size : 3898888704 (1859.14 GiB 1996.23 GB)
>  Used Dev Size : 1949444352 (929.57 GiB 998.12 GB)
>   Super Offset : 1949444640 sectors
>          State : clean
>    Device UUID : b05f4c40:819ddbef:76872d9f:abacf3c9
> 
>    Update Time : Wed Feb 26 02:44:57 2014
>       Checksum : a94e1ae6 - correct
>         Events : 133470
> 
>         Layout : near=2, far=1
>     Chunk Size : 64K
> 
>    Array Slot : 0 (0, failed, empty, empty, 1)
>   Array State : Uu__ 1 failed
> /dev/sdb2:
>          Magic : a92b4efc
>        Version : 1.0
>    Feature Map : 0x0
>     Array UUID : 43a1e240:956c3131:df9f6e66:bd9a071e
>           Name : bwns:1
>  Creation Time : Mon Jan 24 20:57:43 2011
>     Raid Level : raid10
>   Raid Devices : 4
> 
> Avail Dev Size : 1949444384 (929.57 GiB 998.12 GB)
>     Array Size : 3898888704 (1859.14 GiB 1996.23 GB)
>  Used Dev Size : 1949444352 (929.57 GiB 998.12 GB)
>   Super Offset : 1949444640 sectors
>          State : clean
>    Device UUID : 6c34331a:0fda7f73:a1f76d41:a826ac1f
> 
>    Update Time : Wed Feb 26 02:44:57 2014
>       Checksum : fed2165a - correct
>         Events : 133470
> 
>         Layout : near=2, far=1
>     Chunk Size : 64K
> 
>    Array Slot : 4 (0, failed, empty, empty, 1)
>   Array State : uU__ 1 failed
> /dev/sdc2:
>          Magic : a92b4efc
>        Version : 1.0
>    Feature Map : 0x0
>     Array UUID : 43a1e240:956c3131:df9f6e66:bd9a071e
>           Name : bwns:1
>  Creation Time : Mon Jan 24 20:57:43 2011
>     Raid Level : raid10
>   Raid Devices : 4
> 
> Avail Dev Size : 1949444384 (929.57 GiB 998.12 GB)
>     Array Size : 3898888704 (1859.14 GiB 1996.23 GB)
>  Used Dev Size : 1949444352 (929.57 GiB 998.12 GB)
>   Super Offset : 1949444640 sectors
>          State : clean
>    Device UUID : 773dbfec:07467e62:de7be59b:5c680df5
> 
>    Update Time : Wed Feb 26 02:44:57 2014
>       Checksum : f035517e - correct
>         Events : 133470
> 
>         Layout : near=2, far=1
>     Chunk Size : 64K
> 
>    Array Slot : 2 (0, failed, empty, empty, 1)
>   Array State : uu__ 1 failed
> /dev/sdd2:
>          Magic : a92b4efc
>        Version : 1.0
>    Feature Map : 0x0
>     Array UUID : 43a1e240:956c3131:df9f6e66:bd9a071e
>           Name : bwns:1
>  Creation Time : Mon Jan 24 20:57:43 2011
>     Raid Level : raid10
>   Raid Devices : 4
> 
> Avail Dev Size : 1949444384 (929.57 GiB 998.12 GB)
>     Array Size : 3898888704 (1859.14 GiB 1996.23 GB)
>  Used Dev Size : 1949444352 (929.57 GiB 998.12 GB)
>   Super Offset : 1949444640 sectors
>          State : clean
>    Device UUID : dbd27546:6b623b53:8f887960:b7cbf424
> 
>    Update Time : Wed Feb 26 02:44:57 2014
>       Checksum : 2f247322 - correct
>         Events : 133470
> 
>         Layout : near=2, far=1
>     Chunk Size : 64K
> 
>    Array Slot : 3 (0, failed, empty, empty, 1)
>   Array State : uu__ 1 failed
> 
> 
> Taking a look at the event count they seem to be synchronized, so I'm not really sure what's going on here.
> 
> root@BauwensNAS:/# cat /proc/mdstat
> Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid4]
> md1 : inactive sda2[0] sdd2[3](S) sdc2[2](S) sdb2[4]
>      3898888704 blocks super 1.0
> 
> md0 : active raid1 sda1[0] sdd1[3] sdc1[2] sdb1[1]
>      2040128 blocks [4/4] [UUUU]
> 
> unused devices: <none>
> 
> Anyone have an idea how I could resolve this problem (hoping that I don't have any data loss...)? Any help is greatly appreciated. I sure regret rebooting the device without taking some extra backups.
> 
> TIA!
> Jim
> --
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

^ permalink raw reply	[flat|nested] 2+ messages in thread

end of thread, other threads:[~2014-03-06 10:11 UTC | newest]

Thread overview: 2+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2014-02-28 19:40 RAID 10 problems, two disks marked as spare jim
2014-03-06 10:11 ` Jim Bauwens

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).