linux-raid.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
* Busted disks caused healthy ones to fail
@ 2004-12-14  6:42 comsatcat
  2004-12-14  6:55 ` Guy
  0 siblings, 1 reply; 16+ messages in thread
From: comsatcat @ 2004-12-14  6:42 UTC (permalink / raw)
  To: linux-raid

An odd thing happened this weekend.  We were doing some heavy I/O when
one of our servers had two drives in two seperate raid1 mirrors pop.
This was not odd as these drives are old and the batch they are from
have been failing on other boxen as well.  What is odd is that our brand
new disks which the OS resides on (2 drives in raid 1) half busted.

There are 4 md devices

md/0  
md/1
md/2
md/3

md3, md2, and md1 all lost the 2nd drive in the array (sdh3, sdh6, and
sdh5).  md0 however was fine with sdh1 being fine.  Why would losing
disks cause a seemingly healthy disk to go astray?

P.S. I have pull out tons of syslogs showing the two bad disks failing
if that would help.


Thanks,
Ben


^ permalink raw reply	[flat|nested] 16+ messages in thread

* RE: Busted disks caused healthy ones to fail
  2004-12-14  6:42 Busted disks caused healthy ones to fail comsatcat
@ 2004-12-14  6:55 ` Guy
  2004-12-14  8:28   ` comsatcat
  0 siblings, 1 reply; 16+ messages in thread
From: Guy @ 2004-12-14  6:55 UTC (permalink / raw)
  To: comsatcat, linux-raid

Did the disks that failed have anything in common?

SCSI:
If you have disks on 1 SCSI bus, a single failed disk can affect other
disks.  By removing the bad disk you correct the problems with the others.

IDE:  (or what ever they call it today)
2 disks on 1 bus, 1 drive failure will cause the other to fail most of the
time.

Power supply:
If you have external disks, they will have another power supply.  If you
have problems with this power supply, they all could be affected.  Even a
common power cable can cause multi drive failures.

Temperature:
Disks getting too hot can cause failures.

Kids:
Someone turned the disk cabinet off?

I am sure this list is not complete.  But it may help.

Guy

-----Original Message-----
From: linux-raid-owner@vger.kernel.org
[mailto:linux-raid-owner@vger.kernel.org] On Behalf Of comsatcat
Sent: Tuesday, December 14, 2004 1:42 AM
To: linux-raid@vger.kernel.org
Subject: Busted disks caused healthy ones to fail

An odd thing happened this weekend.  We were doing some heavy I/O when
one of our servers had two drives in two seperate raid1 mirrors pop.
This was not odd as these drives are old and the batch they are from
have been failing on other boxen as well.  What is odd is that our brand
new disks which the OS resides on (2 drives in raid 1) half busted.

There are 4 md devices

md/0  
md/1
md/2
md/3

md3, md2, and md1 all lost the 2nd drive in the array (sdh3, sdh6, and
sdh5).  md0 however was fine with sdh1 being fine.  Why would losing
disks cause a seemingly healthy disk to go astray?

P.S. I have pull out tons of syslogs showing the two bad disks failing
if that would help.


Thanks,
Ben

-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html


^ permalink raw reply	[flat|nested] 16+ messages in thread

* RE: Busted disks caused healthy ones to fail
  2004-12-14  6:55 ` Guy
@ 2004-12-14  8:28   ` comsatcat
  2004-12-14 14:11     ` Michael Stumpf
  2004-12-14 15:22     ` Guy
  0 siblings, 2 replies; 16+ messages in thread
From: comsatcat @ 2004-12-14  8:28 UTC (permalink / raw)
  To: Guy; +Cc: linux-raid

The two disks that were actually dead were both on a different bus.  The
OS disk that died was on scsi0.

Is there a way around this behavior (ie: kernel params that can be
adjusted such as timeout values and queuing)?  It never really recovered
correctly after the disks died, a manual reboot as required.
Applications which were using the failed devices would hang forever (I'm
assuming they were waiting for queued commands to complete).

IDE: not in use
Power: 14 internal drives, no external
Temp: fust fine
Kids: Upstairs taking tech calls.


Thanks,
Ben


On Tue, 2004-12-14 at 01:55 -0500, Guy wrote:
> Did the disks that failed have anything in common?
> 
> SCSI:
> If you have disks on 1 SCSI bus, a single failed disk can affect other
> disks.  By removing the bad disk you correct the problems with the others.
> 
> IDE:  (or what ever they call it today)
> 2 disks on 1 bus, 1 drive failure will cause the other to fail most of the
> time.
> 
> Power supply:
> If you have external disks, they will have another power supply.  If you
> have problems with this power supply, they all could be affected.  Even a
> common power cable can cause multi drive failures.
> 
> Temperature:
> Disks getting too hot can cause failures.
> 
> Kids:
> Someone turned the disk cabinet off?
> 
> I am sure this list is not complete.  But it may help.
> 
> Guy
> 
> -----Original Message-----
> From: linux-raid-owner@vger.kernel.org
> [mailto:linux-raid-owner@vger.kernel.org] On Behalf Of comsatcat
> Sent: Tuesday, December 14, 2004 1:42 AM
> To: linux-raid@vger.kernel.org
> Subject: Busted disks caused healthy ones to fail
> 
> An odd thing happened this weekend.  We were doing some heavy I/O when
> one of our servers had two drives in two seperate raid1 mirrors pop.
> This was not odd as these drives are old and the batch they are from
> have been failing on other boxen as well.  What is odd is that our brand
> new disks which the OS resides on (2 drives in raid 1) half busted.
> 
> There are 4 md devices
> 
> md/0  
> md/1
> md/2
> md/3
> 
> md3, md2, and md1 all lost the 2nd drive in the array (sdh3, sdh6, and
> sdh5).  md0 however was fine with sdh1 being fine.  Why would losing
> disks cause a seemingly healthy disk to go astray?
> 
> P.S. I have pull out tons of syslogs showing the two bad disks failing
> if that would help.
> 
> 
> Thanks,
> Ben
> 
> -
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> 
> -
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html


^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: Busted disks caused healthy ones to fail
  2004-12-14  8:28   ` comsatcat
@ 2004-12-14 14:11     ` Michael Stumpf
  2004-12-14 22:34       ` comsatcat
  2004-12-14 15:22     ` Guy
  1 sibling, 1 reply; 16+ messages in thread
From: Michael Stumpf @ 2004-12-14 14:11 UTC (permalink / raw)
  To: comsatcat, linux-raid

14 Internal drives on a single power supply plus the mb/cpu/etc?  Oy; 
I've got 15 + a p2-400 spinning between 2 550w power supplies, and I'm 
worried it is getting overloaded.  I might be paranoid, but I had some 
flakiness that was pretty much impossible to debug, so I took broad 
steps and overestimated.  Figured that maybe a heavily loaded supply 
could hiccup under an unusual condition if too many were attached to 
one..  and, while anecdotal, my once-a-month drive hiccup (require 
re-add to array, nothing else) problem did go away when I added a power 
supply.

comsatcat wrote:

>The two disks that were actually dead were both on a different bus.  The
>OS disk that died was on scsi0.
>
>Is there a way around this behavior (ie: kernel params that can be
>adjusted such as timeout values and queuing)?  It never really recovered
>correctly after the disks died, a manual reboot as required.
>Applications which were using the failed devices would hang forever (I'm
>assuming they were waiting for queued commands to complete).
>
>IDE: not in use
>Power: 14 internal drives, no external
>Temp: fust fine
>Kids: Upstairs taking tech calls.
>
>
>Thanks,
>Ben
>
>
>On Tue, 2004-12-14 at 01:55 -0500, Guy wrote:
>  
>
>>Did the disks that failed have anything in common?
>>
>>SCSI:
>>If you have disks on 1 SCSI bus, a single failed disk can affect other
>>disks.  By removing the bad disk you correct the problems with the others.
>>
>>IDE:  (or what ever they call it today)
>>2 disks on 1 bus, 1 drive failure will cause the other to fail most of the
>>time.
>>
>>Power supply:
>>If you have external disks, they will have another power supply.  If you
>>have problems with this power supply, they all could be affected.  Even a
>>common power cable can cause multi drive failures.
>>
>>Temperature:
>>Disks getting too hot can cause failures.
>>
>>Kids:
>>Someone turned the disk cabinet off?
>>
>>I am sure this list is not complete.  But it may help.
>>
>>Guy
>>
>>-----Original Message-----
>>From: linux-raid-owner@vger.kernel.org
>>[mailto:linux-raid-owner@vger.kernel.org] On Behalf Of comsatcat
>>Sent: Tuesday, December 14, 2004 1:42 AM
>>To: linux-raid@vger.kernel.org
>>Subject: Busted disks caused healthy ones to fail
>>
>>An odd thing happened this weekend.  We were doing some heavy I/O when
>>one of our servers had two drives in two seperate raid1 mirrors pop.
>>This was not odd as these drives are old and the batch they are from
>>have been failing on other boxen as well.  What is odd is that our brand
>>new disks which the OS resides on (2 drives in raid 1) half busted.
>>
>>There are 4 md devices
>>
>>md/0  
>>md/1
>>md/2
>>md/3
>>
>>md3, md2, and md1 all lost the 2nd drive in the array (sdh3, sdh6, and
>>sdh5).  md0 however was fine with sdh1 being fine.  Why would losing
>>disks cause a seemingly healthy disk to go astray?
>>
>>P.S. I have pull out tons of syslogs showing the two bad disks failing
>>if that would help.
>>
>>
>>Thanks,
>>Ben
>>
>>-
>>To unsubscribe from this list: send the line "unsubscribe linux-raid" in
>>the body of a message to majordomo@vger.kernel.org
>>More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>
>>-
>>To unsubscribe from this list: send the line "unsubscribe linux-raid" in
>>the body of a message to majordomo@vger.kernel.org
>>More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>    
>>
>
>-
>To unsubscribe from this list: send the line "unsubscribe linux-raid" in
>the body of a message to majordomo@vger.kernel.org
>More majordomo info at  http://vger.kernel.org/majordomo-info.html
>
>
>  
>


--------------------------------------------
My mailbox is spam-free with ChoiceMail, the leader in personal and corporate anti-spam solutions. Download your free copy of ChoiceMail from www.choicemailfree.com


^ permalink raw reply	[flat|nested] 16+ messages in thread

* RE: Busted disks caused healthy ones to fail
  2004-12-14  8:28   ` comsatcat
  2004-12-14 14:11     ` Michael Stumpf
@ 2004-12-14 15:22     ` Guy
  2004-12-14 20:13       ` Brad Campbell
  1 sibling, 1 reply; 16+ messages in thread
From: Guy @ 2004-12-14 15:22 UTC (permalink / raw)
  To: comsatcat; +Cc: linux-raid

14 drives in 1 case?  That's a big box!

Did you ask your kids for help?  :)

Guy

-----Original Message-----
From: linux-raid-owner@vger.kernel.org
[mailto:linux-raid-owner@vger.kernel.org] On Behalf Of comsatcat
Sent: Tuesday, December 14, 2004 3:29 AM
To: Guy
Cc: linux-raid@vger.kernel.org
Subject: RE: Busted disks caused healthy ones to fail

The two disks that were actually dead were both on a different bus.  The
OS disk that died was on scsi0.

Is there a way around this behavior (ie: kernel params that can be
adjusted such as timeout values and queuing)?  It never really recovered
correctly after the disks died, a manual reboot as required.
Applications which were using the failed devices would hang forever (I'm
assuming they were waiting for queued commands to complete).

IDE: not in use
Power: 14 internal drives, no external
Temp: fust fine
Kids: Upstairs taking tech calls.


Thanks,
Ben


On Tue, 2004-12-14 at 01:55 -0500, Guy wrote:
> Did the disks that failed have anything in common?
> 
> SCSI:
> If you have disks on 1 SCSI bus, a single failed disk can affect other
> disks.  By removing the bad disk you correct the problems with the others.
> 
> IDE:  (or what ever they call it today)
> 2 disks on 1 bus, 1 drive failure will cause the other to fail most of the
> time.
> 
> Power supply:
> If you have external disks, they will have another power supply.  If you
> have problems with this power supply, they all could be affected.  Even a
> common power cable can cause multi drive failures.
> 
> Temperature:
> Disks getting too hot can cause failures.
> 
> Kids:
> Someone turned the disk cabinet off?
> 
> I am sure this list is not complete.  But it may help.
> 
> Guy
> 
> -----Original Message-----
> From: linux-raid-owner@vger.kernel.org
> [mailto:linux-raid-owner@vger.kernel.org] On Behalf Of comsatcat
> Sent: Tuesday, December 14, 2004 1:42 AM
> To: linux-raid@vger.kernel.org
> Subject: Busted disks caused healthy ones to fail
> 
> An odd thing happened this weekend.  We were doing some heavy I/O when
> one of our servers had two drives in two seperate raid1 mirrors pop.
> This was not odd as these drives are old and the batch they are from
> have been failing on other boxen as well.  What is odd is that our brand
> new disks which the OS resides on (2 drives in raid 1) half busted.
> 
> There are 4 md devices
> 
> md/0  
> md/1
> md/2
> md/3
> 
> md3, md2, and md1 all lost the 2nd drive in the array (sdh3, sdh6, and
> sdh5).  md0 however was fine with sdh1 being fine.  Why would losing
> disks cause a seemingly healthy disk to go astray?
> 
> P.S. I have pull out tons of syslogs showing the two bad disks failing
> if that would help.
> 
> 
> Thanks,
> Ben
> 
> -
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> 
> -
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html


^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: Busted disks caused healthy ones to fail
  2004-12-14 15:22     ` Guy
@ 2004-12-14 20:13       ` Brad Campbell
  2004-12-14 21:47         ` Guy
  2004-12-14 21:49         ` Jim Paris
  0 siblings, 2 replies; 16+ messages in thread
From: Brad Campbell @ 2004-12-14 20:13 UTC (permalink / raw)
  To: Guy; +Cc: comsatcat, linux-raid

Guy wrote:
> 14 drives in 1 case?  That's a big box!
> 

It's not that hard.
I have 4 drives loaded in the rear bays and 2 x 5 Way SATA Hotswap bays in the 6 front 5.25 inch 
bays. 14 Drives. And yes, they are on a single 420w PSU along with the motherboard, Athlon XP 2600+. 
and 5 80mm fans. Not much else though.

I'm working my way towards a 15 drive box now. Just waiting for the new enclosures to arrive.

-- 
Brad
                    /"\
Save the Forests   \ /     ASCII RIBBON CAMPAIGN
Burn a Greenie.     X      AGAINST HTML MAIL
                    / \

^ permalink raw reply	[flat|nested] 16+ messages in thread

* RE: Busted disks caused healthy ones to fail
  2004-12-14 20:13       ` Brad Campbell
@ 2004-12-14 21:47         ` Guy
  2004-12-14 23:54           ` Alvin Oga
  2004-12-14 21:49         ` Jim Paris
  1 sibling, 1 reply; 16+ messages in thread
From: Guy @ 2004-12-14 21:47 UTC (permalink / raw)
  To: 'Brad Campbell'; +Cc: comsatcat, linux-raid

My disk drives are rated to use 19.1 watts while active.
My disk drives also get very hot without forced air movement.
14 such disks use 267.4 watts.  That would give you 152.6 watts for
everything else.  I bet your disks use less power than mine.  I agree with
Michael, you may be exceeding the power rating of your power supply.

Also, I would not want to push a power supply to the max rating.  It should
be sized 50 watts or more beyond what you need.

Guy

-----Original Message-----
From: linux-raid-owner@vger.kernel.org
[mailto:linux-raid-owner@vger.kernel.org] On Behalf Of Brad Campbell
Sent: Tuesday, December 14, 2004 3:13 PM
To: Guy
Cc: comsatcat@earthlink.net; linux-raid@vger.kernel.org
Subject: Re: Busted disks caused healthy ones to fail

Guy wrote:
> 14 drives in 1 case?  That's a big box!
> 

It's not that hard.
I have 4 drives loaded in the rear bays and 2 x 5 Way SATA Hotswap bays in
the 6 front 5.25 inch 
bays. 14 Drives. And yes, they are on a single 420w PSU along with the
motherboard, Athlon XP 2600+. 
and 5 80mm fans. Not much else though.

I'm working my way towards a 15 drive box now. Just waiting for the new
enclosures to arrive.

-- 
Brad
                    /"\
Save the Forests   \ /     ASCII RIBBON CAMPAIGN
Burn a Greenie.     X      AGAINST HTML MAIL
                    / \
-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html


^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: Busted disks caused healthy ones to fail
  2004-12-14 20:13       ` Brad Campbell
  2004-12-14 21:47         ` Guy
@ 2004-12-14 21:49         ` Jim Paris
  2004-12-14 22:13           ` Guy
  2004-12-15  4:46           ` Brad Campbell
  1 sibling, 2 replies; 16+ messages in thread
From: Jim Paris @ 2004-12-14 21:49 UTC (permalink / raw)
  To: Brad Campbell; +Cc: Guy, comsatcat, linux-raid

> It's not that hard.
> I have 4 drives loaded in the rear bays and 2 x 5 Way SATA Hotswap bays in 
> the 6 front 5.25 inch bays. 14 Drives. And yes, they are on a single 420w 
> PSU along with the motherboard, Athlon XP 2600+. and 5 80mm fans. Not much 
> else though.

!!!!!!  Holy crap!

Let's pick a random typical hard drive, a Seagate 120GB SATA:
http://www.mittoni.com.au/catalog/product_info.php/products_id/1690
It lists maximum current draw as 2.8 A on the +12V line.
Multiply that by 14 drives and we get __39.2 amps__.

Now let's pick a random 420W power supply:
http://www.newegg.com/app/viewproductdesc.asp?submit=Go&description=N82E16817103445
Note how it's +12V output is rated for only __15 amps__.

Your numbers might differ a bit.  But it is NO surprise that your
drives are failing.  The surprising part is that they and your power
supply have worked this long.

-jim

^ permalink raw reply	[flat|nested] 16+ messages in thread

* RE: Busted disks caused healthy ones to fail
  2004-12-14 21:49         ` Jim Paris
@ 2004-12-14 22:13           ` Guy
  2004-12-15  4:46           ` Brad Campbell
  1 sibling, 0 replies; 16+ messages in thread
From: Guy @ 2004-12-14 22:13 UTC (permalink / raw)
  To: 'Jim Paris', 'Brad Campbell'; +Cc: comsatcat, linux-raid

If all of my disks were to start at the same time (assuming I had 14), they
would require 483 watts while starting up!

My disks use 2.5 amps on startup.  On the 12V line.  But use .98 amps when
idle.  And .90 amps on the 5V line.  I have my system to delay the starting
of the disks based on SCSI ID, so only one starts at a time.  If your disks
start at the same time, then OUCH!  How can it work?

You should find the specs on your disks and do some math just to be sure.

If you want to compare, specs for my disks:
http://www.seagate.com/support/disc/specs/scsi/st118202lc.html

Guy

-----Original Message-----
From: linux-raid-owner@vger.kernel.org
[mailto:linux-raid-owner@vger.kernel.org] On Behalf Of Jim Paris
Sent: Tuesday, December 14, 2004 4:50 PM
To: Brad Campbell
Cc: Guy; comsatcat@earthlink.net; linux-raid@vger.kernel.org
Subject: Re: Busted disks caused healthy ones to fail

> It's not that hard.
> I have 4 drives loaded in the rear bays and 2 x 5 Way SATA Hotswap bays in

> the 6 front 5.25 inch bays. 14 Drives. And yes, they are on a single 420w 
> PSU along with the motherboard, Athlon XP 2600+. and 5 80mm fans. Not much

> else though.

!!!!!!  Holy crap!

Let's pick a random typical hard drive, a Seagate 120GB SATA:
http://www.mittoni.com.au/catalog/product_info.php/products_id/1690
It lists maximum current draw as 2.8 A on the +12V line.
Multiply that by 14 drives and we get __39.2 amps__.

Now let's pick a random 420W power supply:
http://www.newegg.com/app/viewproductdesc.asp?submit=Go&description=N82E1681
7103445
Note how it's +12V output is rated for only __15 amps__.

Your numbers might differ a bit.  But it is NO surprise that your
drives are failing.  The surprising part is that they and your power
supply have worked this long.

-jim
-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html


^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: Busted disks caused healthy ones to fail
  2004-12-14 14:11     ` Michael Stumpf
@ 2004-12-14 22:34       ` comsatcat
  0 siblings, 0 replies; 16+ messages in thread
From: comsatcat @ 2004-12-14 22:34 UTC (permalink / raw)
  To: mjstumpf; +Cc: linux-raid

Not 1, 3 power supplys.

On Tue, 2004-12-14 at 08:11 -0600, Michael Stumpf wrote:
> 14 Internal drives on a single power supply plus the mb/cpu/etc?  Oy; 
> I've got 15 + a p2-400 spinning between 2 550w power supplies, and I'm 
> worried it is getting overloaded.  I might be paranoid, but I had some 
> flakiness that was pretty much impossible to debug, so I took broad 
> steps and overestimated.  Figured that maybe a heavily loaded supply 
> could hiccup under an unusual condition if too many were attached to 
> one..  and, while anecdotal, my once-a-month drive hiccup (require 
> re-add to array, nothing else) problem did go away when I added a power 
> supply.
> 
> comsatcat wrote:
> 
> >The two disks that were actually dead were both on a different bus.  The
> >OS disk that died was on scsi0.
> >
> >Is there a way around this behavior (ie: kernel params that can be
> >adjusted such as timeout values and queuing)?  It never really recovered
> >correctly after the disks died, a manual reboot as required.
> >Applications which were using the failed devices would hang forever (I'm
> >assuming they were waiting for queued commands to complete).
> >
> >IDE: not in use
> >Power: 14 internal drives, no external
> >Temp: fust fine
> >Kids: Upstairs taking tech calls.
> >
> >
> >Thanks,
> >Ben
> >
> >
> >On Tue, 2004-12-14 at 01:55 -0500, Guy wrote:
> >  
> >
> >>Did the disks that failed have anything in common?
> >>
> >>SCSI:
> >>If you have disks on 1 SCSI bus, a single failed disk can affect other
> >>disks.  By removing the bad disk you correct the problems with the others.
> >>
> >>IDE:  (or what ever they call it today)
> >>2 disks on 1 bus, 1 drive failure will cause the other to fail most of the
> >>time.
> >>
> >>Power supply:
> >>If you have external disks, they will have another power supply.  If you
> >>have problems with this power supply, they all could be affected.  Even a
> >>common power cable can cause multi drive failures.
> >>
> >>Temperature:
> >>Disks getting too hot can cause failures.
> >>
> >>Kids:
> >>Someone turned the disk cabinet off?
> >>
> >>I am sure this list is not complete.  But it may help.
> >>
> >>Guy
> >>
> >>-----Original Message-----
> >>From: linux-raid-owner@vger.kernel.org
> >>[mailto:linux-raid-owner@vger.kernel.org] On Behalf Of comsatcat
> >>Sent: Tuesday, December 14, 2004 1:42 AM
> >>To: linux-raid@vger.kernel.org
> >>Subject: Busted disks caused healthy ones to fail
> >>
> >>An odd thing happened this weekend.  We were doing some heavy I/O when
> >>one of our servers had two drives in two seperate raid1 mirrors pop.
> >>This was not odd as these drives are old and the batch they are from
> >>have been failing on other boxen as well.  What is odd is that our brand
> >>new disks which the OS resides on (2 drives in raid 1) half busted.
> >>
> >>There are 4 md devices
> >>
> >>md/0  
> >>md/1
> >>md/2
> >>md/3
> >>
> >>md3, md2, and md1 all lost the 2nd drive in the array (sdh3, sdh6, and
> >>sdh5).  md0 however was fine with sdh1 being fine.  Why would losing
> >>disks cause a seemingly healthy disk to go astray?
> >>
> >>P.S. I have pull out tons of syslogs showing the two bad disks failing
> >>if that would help.
> >>
> >>
> >>Thanks,
> >>Ben
> >>
> >>-
> >>To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> >>the body of a message to majordomo@vger.kernel.org
> >>More majordomo info at  http://vger.kernel.org/majordomo-info.html
> >>
> >>-
> >>To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> >>the body of a message to majordomo@vger.kernel.org
> >>More majordomo info at  http://vger.kernel.org/majordomo-info.html
> >>    
> >>
> >
> >-
> >To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> >the body of a message to majordomo@vger.kernel.org
> >More majordomo info at  http://vger.kernel.org/majordomo-info.html
> >
> >
> >  
> >
> 
> 
> --------------------------------------------
> My mailbox is spam-free with ChoiceMail, the leader in personal and corporate anti-spam solutions. Download your free copy of ChoiceMail from www.choicemailfree.com
> 
> -
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html


^ permalink raw reply	[flat|nested] 16+ messages in thread

* RE: Busted disks caused healthy ones to fail
  2004-12-14 21:47         ` Guy
@ 2004-12-14 23:54           ` Alvin Oga
  2004-12-15  1:03             ` Guy
  0 siblings, 1 reply; 16+ messages in thread
From: Alvin Oga @ 2004-12-14 23:54 UTC (permalink / raw)
  To: Guy; +Cc: 'Brad Campbell', comsatcat, linux-raid



On Tue, 14 Dec 2004, Guy wrote:

> My disk drives are rated to use 19.1 watts while active.
> My disk drives also get very hot without forced air movement.
> 14 such disks use 267.4 watts.  That would give you 152.6 watts for
> everything else.  I bet your disks use less power than mine.  I agree with
> Michael, you may be exceeding the power rating of your power supply.

an overworked power supply will fail "faster"

for "wattage" ... using a sharp needle, "pop" ... that is NOT how "wattage
works"

	- for a given voltage, you should be operating at 1/2 of its rated
	amperage

	- at 12v ... what are the disks rated at for power up ( spinup )
	vs ambeint normal operation
		- typically 1A to power up and 0.5A for normal spinning
		operation

	for a powersupply rated at 12V for 10A .. you can run 5 disks

	if you run more disks ... your power supply will die faster
	and/or your data gets corrupted during power ups when the
	system and the drives think you're in "normal operation"
	and enables write  while in fact, its is still in its bootup

> Also, I would not want to push a power supply to the max rating.  It should
> be sized 50 watts or more beyond what you need.

make that 2x the total wattage needed by the system .. NOT 50W 
	- add up the amps needed on the 3.3V, 5V, 12V 
	- use the MAX needed for operation, which is the power up sequence
	to get a "random number" of operating current ..

	Total watts == 12 * 12v(maxcurrent) + 5v * 5v(maxcurrent)
			3.3V * 3.3v(maxcurrent)

	- normal operation usually does NOT need as much current

- switched power supplies can output higher current than it is rated
  at, but it will not be able to sustain that "extra current load"
  for more than a few seconds before its over-current circuitry kicks
  in to shut itself down

	- you will be bouncing up and down in current on each power
	line till the system is all booted

	- put a digital storage scope on the 12V and 5V and 3.3V line
	and an ampmeter on each power line and watch it go bonkers

c ya
alvin


^ permalink raw reply	[flat|nested] 16+ messages in thread

* RE: Busted disks caused healthy ones to fail
  2004-12-14 23:54           ` Alvin Oga
@ 2004-12-15  1:03             ` Guy
  2004-12-15  1:23               ` Alvin Oga
  0 siblings, 1 reply; 16+ messages in thread
From: Guy @ 2004-12-15  1:03 UTC (permalink / raw)
  To: 'Alvin Oga'; +Cc: 'Brad Campbell', comsatcat, linux-raid

Just an FYI...
My disks take 2.5 amps on the 12V line (30 watts) to start.
Someone else said his require 2.9 amps (34.8 watts).
The above does not include the 5V line.

2 X startup wattage or amperage seems excessive.
Power supply delivers 50% during startup, and 25% while the system is in
use.  Seems wrong to me.

Guy

-----Original Message-----
From: Alvin Oga [mailto:aoga@ns.Linux-Consulting.com] 
Sent: Tuesday, December 14, 2004 6:55 PM
To: Guy
Cc: 'Brad Campbell'; comsatcat@earthlink.net; linux-raid@vger.kernel.org
Subject: RE: Busted disks caused healthy ones to fail



On Tue, 14 Dec 2004, Guy wrote:

> My disk drives are rated to use 19.1 watts while active.
> My disk drives also get very hot without forced air movement.
> 14 such disks use 267.4 watts.  That would give you 152.6 watts for
> everything else.  I bet your disks use less power than mine.  I agree with
> Michael, you may be exceeding the power rating of your power supply.

an overworked power supply will fail "faster"

for "wattage" ... using a sharp needle, "pop" ... that is NOT how "wattage
works"

	- for a given voltage, you should be operating at 1/2 of its rated
	amperage

	- at 12v ... what are the disks rated at for power up ( spinup )
	vs ambeint normal operation
		- typically 1A to power up and 0.5A for normal spinning
		operation

	for a powersupply rated at 12V for 10A .. you can run 5 disks

	if you run more disks ... your power supply will die faster
	and/or your data gets corrupted during power ups when the
	system and the drives think you're in "normal operation"
	and enables write  while in fact, its is still in its bootup

> Also, I would not want to push a power supply to the max rating.  It
should
> be sized 50 watts or more beyond what you need.

make that 2x the total wattage needed by the system .. NOT 50W 
	- add up the amps needed on the 3.3V, 5V, 12V 
	- use the MAX needed for operation, which is the power up sequence
	to get a "random number" of operating current ..

	Total watts == 12 * 12v(maxcurrent) + 5v * 5v(maxcurrent)
			3.3V * 3.3v(maxcurrent)

	- normal operation usually does NOT need as much current

- switched power supplies can output higher current than it is rated
  at, but it will not be able to sustain that "extra current load"
  for more than a few seconds before its over-current circuitry kicks
  in to shut itself down

	- you will be bouncing up and down in current on each power
	line till the system is all booted

	- put a digital storage scope on the 12V and 5V and 3.3V line
	and an ampmeter on each power line and watch it go bonkers

c ya
alvin


^ permalink raw reply	[flat|nested] 16+ messages in thread

* RE: Busted disks caused healthy ones to fail
  2004-12-15  1:03             ` Guy
@ 2004-12-15  1:23               ` Alvin Oga
  0 siblings, 0 replies; 16+ messages in thread
From: Alvin Oga @ 2004-12-15  1:23 UTC (permalink / raw)
  To: Guy; +Cc: 'Alvin Oga', 'Brad Campbell', comsatcat,
	linux-raid



On Tue, 14 Dec 2004, Guy wrote:

> Just an FYI...
> My disks take 2.5 amps on the 12V line (30 watts) to start.
> Someone else said his require 2.9 amps (34.8 watts).

the max rated startup current is NOT necessarily the real
current
	- use a storage digital scope to get "REAL" numbers

> The above does not include the 5V line.
> 
> 2 X startup wattage or amperage seems excessive.

comes from the last 50 years of "rule of thumb" 
when things was not as reliable  

another rule of thumb ...

	- things will die 2x as fast if the average sustained operating
	temp goes up by 10C

	- it works out just about right .. when you take the
	1,000,000 hour MTBF and factor in all the temps and
	24x7 operation and you find out that the little disk
	dies 1 month after the warranty expired :-)

c ya
alvin


^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: Busted disks caused healthy ones to fail
  2004-12-14 21:49         ` Jim Paris
  2004-12-14 22:13           ` Guy
@ 2004-12-15  4:46           ` Brad Campbell
  2004-12-15  5:04             ` Guy
  1 sibling, 1 reply; 16+ messages in thread
From: Brad Campbell @ 2004-12-15  4:46 UTC (permalink / raw)
  To: Jim Paris; +Cc: Guy, comsatcat, linux-raid

Jim Paris wrote:
>>It's not that hard.
>>I have 4 drives loaded in the rear bays and 2 x 5 Way SATA Hotswap bays in 
>>the 6 front 5.25 inch bays. 14 Drives. And yes, they are on a single 420w 
>>PSU along with the motherboard, Athlon XP 2600+. and 5 80mm fans. Not much 
>>else though.
> 
> 
> !!!!!!  Holy crap!
> 
> Let's pick a random typical hard drive, a Seagate 120GB SATA:
> http://www.mittoni.com.au/catalog/product_info.php/products_id/1690
> It lists maximum current draw as 2.8 A on the +12V line.
> Multiply that by 14 drives and we get __39.2 amps__.

Now, lets actually pick my hard drives shall we?

Max current draw on the 12v line is 1.56A at spinup, dropping to 600mA at seek and 556mA at Idle.

So _worst_ case is 21.84A for about 2 seconds (which does actually exceed the PSU ratings by nearly 
3 amps). This machine only gets power cycled about once every three months and I did actually 
monitor the 12V rail with a CRO to check specs and ripple and they never budged.

Worst case running load is 8.4A which leaves ~10A on my 12V rail for my motherboard. Ample.


> Now let's pick a random 420W power supply:
> http://www.newegg.com/app/viewproductdesc.asp?submit=Go&description=N82E16817103445
> Note how it's +12V output is rated for only __15 amps__.

Now lets pick my power supply http://www.wasp.net.au/~brad/p1000256.jpg

So yes, on spinup I'm exceeding my 12V rail by 3 Amps for about 1.5 Seconds (Which this supply has 
amply proven capable of handling). Outside that I don't see an issue.

> Your numbers might differ a bit.  But it is NO surprise that your
> drives are failing.  The surprising part is that they and your power
> supply have worked this long.

I never said anything about failing disks! In fact, if you check back you will see me commenting I 
have a bucket load of Maxtor Maxline-II drives in there that have been flawless to date. (In fact I 
have just ordered 25 more, 15 for me and 10 for a mate. That should increase the sample a little) 
They all sit at below 40 Degrees C and the PSU remains quite cool. (I'm an electronic technician by 
trade and have several thermocouples I use to verify measurements).

Here is the reason the drives stay nice and cool. http://www.wasp.net.au/~brad/p1000250.jpg

-- 
Brad
                    /"\
Save the Forests   \ /     ASCII RIBBON CAMPAIGN
Burn a Greenie.     X      AGAINST HTML MAIL
                    / \

^ permalink raw reply	[flat|nested] 16+ messages in thread

* RE: Busted disks caused healthy ones to fail
  2004-12-15  4:46           ` Brad Campbell
@ 2004-12-15  5:04             ` Guy
  2004-12-15  5:22               ` Brad Campbell
  0 siblings, 1 reply; 16+ messages in thread
From: Guy @ 2004-12-15  5:04 UTC (permalink / raw)
  To: 'Brad Campbell', 'Jim Paris'; +Cc: comsatcat, linux-raid

Maxtor drives....and no problems....You must be crazy!  :)

Well then, uh, it could...., humm....  That only leaves crazy!  :)

I give up!  You defended yourself well.  I have no idea.

Guy

-----Original Message-----
From: linux-raid-owner@vger.kernel.org
[mailto:linux-raid-owner@vger.kernel.org] On Behalf Of Brad Campbell
Sent: Tuesday, December 14, 2004 11:46 PM
To: Jim Paris
Cc: Guy; comsatcat@earthlink.net; linux-raid@vger.kernel.org
Subject: Re: Busted disks caused healthy ones to fail

Jim Paris wrote:
>>It's not that hard.
>>I have 4 drives loaded in the rear bays and 2 x 5 Way SATA Hotswap bays in

>>the 6 front 5.25 inch bays. 14 Drives. And yes, they are on a single 420w 
>>PSU along with the motherboard, Athlon XP 2600+. and 5 80mm fans. Not much

>>else though.
> 
> 
> !!!!!!  Holy crap!
> 
> Let's pick a random typical hard drive, a Seagate 120GB SATA:
> http://www.mittoni.com.au/catalog/product_info.php/products_id/1690
> It lists maximum current draw as 2.8 A on the +12V line.
> Multiply that by 14 drives and we get __39.2 amps__.

Now, lets actually pick my hard drives shall we?

Max current draw on the 12v line is 1.56A at spinup, dropping to 600mA at
seek and 556mA at Idle.

So _worst_ case is 21.84A for about 2 seconds (which does actually exceed
the PSU ratings by nearly 
3 amps). This machine only gets power cycled about once every three months
and I did actually 
monitor the 12V rail with a CRO to check specs and ripple and they never
budged.

Worst case running load is 8.4A which leaves ~10A on my 12V rail for my
motherboard. Ample.


> Now let's pick a random 420W power supply:
>
http://www.newegg.com/app/viewproductdesc.asp?submit=Go&description=N82E1681
7103445
> Note how it's +12V output is rated for only __15 amps__.

Now lets pick my power supply http://www.wasp.net.au/~brad/p1000256.jpg

So yes, on spinup I'm exceeding my 12V rail by 3 Amps for about 1.5 Seconds
(Which this supply has 
amply proven capable of handling). Outside that I don't see an issue.

> Your numbers might differ a bit.  But it is NO surprise that your
> drives are failing.  The surprising part is that they and your power
> supply have worked this long.

I never said anything about failing disks! In fact, if you check back you
will see me commenting I 
have a bucket load of Maxtor Maxline-II drives in there that have been
flawless to date. (In fact I 
have just ordered 25 more, 15 for me and 10 for a mate. That should increase
the sample a little) 
They all sit at below 40 Degrees C and the PSU remains quite cool. (I'm an
electronic technician by 
trade and have several thermocouples I use to verify measurements).

Here is the reason the drives stay nice and cool.
http://www.wasp.net.au/~brad/p1000250.jpg

-- 
Brad
                    /"\
Save the Forests   \ /     ASCII RIBBON CAMPAIGN
Burn a Greenie.     X      AGAINST HTML MAIL
                    / \
-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html


^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: Busted disks caused healthy ones to fail
  2004-12-15  5:04             ` Guy
@ 2004-12-15  5:22               ` Brad Campbell
  0 siblings, 0 replies; 16+ messages in thread
From: Brad Campbell @ 2004-12-15  5:22 UTC (permalink / raw)
  To: Guy; +Cc: 'Jim Paris', comsatcat, linux-raid

Guy wrote:
> Maxtor drives....and no problems....You must be crazy!  :)
> 
> Well then, uh, it could...., humm....  That only leaves crazy!  :)
> 
> I give up!  You defended yourself well.  I have no idea.
> 

Note that it was not me who had the failing disk in the first place. I was just responding to the 
comment that 14 drives in a case sounded ludicrous. If the setup is reasonably well thought out and 
well cooled I see no issue. Yes, I do exceed my PSU continuous rating every time I spin up, but as I 
said I timed that overload at about 1.5 seconds and the PSU does not make any form of complaint.

I hope I'm more than just lucky with the Maxtor drives, but honestly I have had just as bad a run 
with every brand except Quantum (and who owns them now?). I figure by keeping the drives as cool as 
practical, in a stable environment with minimal temperature fluctuation and power cycling (this 
machine is a 24x7 server) then I'm probably fairly likely to get a better than average lifetime.

Sure with 29 Identical Maxtor drives I expect failures. I have a cold spare on standby just in case 
and the new 15 drive box will run Raid-6. In addition, this is a home entertainment system, it's not 
mission critical. Just a bit of fun on the weekends.

It's also a good test of the md and libata drivers. All up between the 29 drives I now have 7 
Promise SATA150TX4 controllers. Looking forward to hotswap :p)

-- 
Brad
                    /"\
Save the Forests   \ /     ASCII RIBBON CAMPAIGN
Burn a Greenie.     X      AGAINST HTML MAIL
                    / \

^ permalink raw reply	[flat|nested] 16+ messages in thread

end of thread, other threads:[~2004-12-15  5:22 UTC | newest]

Thread overview: 16+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2004-12-14  6:42 Busted disks caused healthy ones to fail comsatcat
2004-12-14  6:55 ` Guy
2004-12-14  8:28   ` comsatcat
2004-12-14 14:11     ` Michael Stumpf
2004-12-14 22:34       ` comsatcat
2004-12-14 15:22     ` Guy
2004-12-14 20:13       ` Brad Campbell
2004-12-14 21:47         ` Guy
2004-12-14 23:54           ` Alvin Oga
2004-12-15  1:03             ` Guy
2004-12-15  1:23               ` Alvin Oga
2004-12-14 21:49         ` Jim Paris
2004-12-14 22:13           ` Guy
2004-12-15  4:46           ` Brad Campbell
2004-12-15  5:04             ` Guy
2004-12-15  5:22               ` Brad Campbell

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).