linux-raid.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
* flagging a disk faulty during raid-5 reshape/grow operation
@ 2008-11-03 15:48 max power
  2008-11-03 16:47 ` David Greaves
  0 siblings, 1 reply; 2+ messages in thread
From: max power @ 2008-11-03 15:48 UTC (permalink / raw)
  To: linux-raid

Hi list,

This morning, I tried to resize our raid-5 array from 3 to 4 disks.
Immediately after issuing the 'mdadm --grow -n 4' command, I found out
that the new disk is actually 'faulty'.  There are no unrecoverable
errors yet, but my smartctl shows a lot of hardware ECC and seek
errors going on. The read errors are causing the whole array to
perform very poorly, and the resize is said to take another 2-3 days
(according to mdstat).

Is it safe to mark the new drive as faulty before the reshape
operation finishes, or is there a risk that this will corrupt the
raid?

I'm using kernel version 2.6.24-21-generic #1 SMP Mon Aug 25 17:32:09
UTC 2008 i686 GNU/Linux (Ubuntu Intrepid), and mdadm version mdadm -
v2.6.7 - 6th June 2008.

output from mdadm  --detail:

/dev/md2:
        Version : 00.91
  Creation Time : Mon Nov 19 14:04:02 2007
     Raid Level : raid5
     Array Size : 535173120 (510.38 GiB 548.02 GB)
  Used Dev Size : 267586560 (255.19 GiB 274.01 GB)
   Raid Devices : 4
  Total Devices : 4
Preferred Minor : 2
    Persistence : Superblock is persistent

    Update Time : Mon Nov  3 16:34:08 2008
          State : clean, recovering
 Active Devices : 4
Working Devices : 4
 Failed Devices : 0
  Spare Devices : 0

         Layout : left-symmetric
     Chunk Size : 64K

 Reshape Status : 1% complete
  Delta Devices : 1, (3->4)

           UUID : 36b972e8:26bae0a4:de6b0f36:
485dee50
         Events : 0.6496

    Number   Major   Minor   RaidDevice State
       0       8        7        0      active sync   /dev/sda7
       1       8       39        1      active sync   /dev/sdc7
       2       8       55        2      active sync   /dev/sdd7
       3       8       23        3      active sync   /dev/sdb7



Thanks in advance,
Jorik

^ permalink raw reply	[flat|nested] 2+ messages in thread

* Re: flagging a disk faulty during raid-5 reshape/grow operation
  2008-11-03 15:48 flagging a disk faulty during raid-5 reshape/grow operation max power
@ 2008-11-03 16:47 ` David Greaves
  0 siblings, 0 replies; 2+ messages in thread
From: David Greaves @ 2008-11-03 16:47 UTC (permalink / raw)
  To: max power; +Cc: linux-raid

Hi avoozl ;)

From a message from Alex on 6/2/08:

If a drive failes during reshape the reshape will just continue.
The blocks which were on the failed drive are calculated from the the other
disks and writes to the failed disk are simply omitted.
The result is a raid5 with a failed drive.
You should get a new drive asap to restore the redundancy.
Also it's kinda important that you don't run 2.6.23 because it has a nasty bug
which would be triggered in this scenario.
The reshape probably increased in speed after the system was no longer actively
used and io bandwidth freed up.


So mark it as faulty and add a new RMA'd disk ASAP.

David

max power wrote:
> Hi list,
> 
> This morning, I tried to resize our raid-5 array from 3 to 4 disks.
> Immediately after issuing the 'mdadm --grow -n 4' command, I found out
> that the new disk is actually 'faulty'.  There are no unrecoverable
> errors yet, but my smartctl shows a lot of hardware ECC and seek
> errors going on. The read errors are causing the whole array to
> perform very poorly, and the resize is said to take another 2-3 days
> (according to mdstat).
> 
> Is it safe to mark the new drive as faulty before the reshape
> operation finishes, or is there a risk that this will corrupt the
> raid?
> 
> I'm using kernel version 2.6.24-21-generic #1 SMP Mon Aug 25 17:32:09
> UTC 2008 i686 GNU/Linux (Ubuntu Intrepid), and mdadm version mdadm -
> v2.6.7 - 6th June 2008.
> 
> output from mdadm  --detail:
> 
> /dev/md2:
>         Version : 00.91
>   Creation Time : Mon Nov 19 14:04:02 2007
>      Raid Level : raid5
>      Array Size : 535173120 (510.38 GiB 548.02 GB)
>   Used Dev Size : 267586560 (255.19 GiB 274.01 GB)
>    Raid Devices : 4
>   Total Devices : 4
> Preferred Minor : 2
>     Persistence : Superblock is persistent
> 
>     Update Time : Mon Nov  3 16:34:08 2008
>           State : clean, recovering
>  Active Devices : 4
> Working Devices : 4
>  Failed Devices : 0
>   Spare Devices : 0
> 
>          Layout : left-symmetric
>      Chunk Size : 64K
> 
>  Reshape Status : 1% complete
>   Delta Devices : 1, (3->4)
> 
>            UUID : 36b972e8:26bae0a4:de6b0f36:
> 485dee50
>          Events : 0.6496
> 
>     Number   Major   Minor   RaidDevice State
>        0       8        7        0      active sync   /dev/sda7
>        1       8       39        1      active sync   /dev/sdc7
>        2       8       55        2      active sync   /dev/sdd7
>        3       8       23        3      active sync   /dev/sdb7
> 
> 
> 
> Thanks in advance,
> Jorik
> --
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html


-- 
"Don't worry, you'll be fine; I saw it work in a cartoon once..."

^ permalink raw reply	[flat|nested] 2+ messages in thread

end of thread, other threads:[~2008-11-03 16:47 UTC | newest]

Thread overview: 2+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2008-11-03 15:48 flagging a disk faulty during raid-5 reshape/grow operation max power
2008-11-03 16:47 ` David Greaves

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).