From: maarten van den Berg <maarten@vbvb.nl>
To: 'LinuxRaid' <linux-raid@vger.kernel.org>
Subject: Re: Hard drive Reliability?
Date: Thu, 20 May 2004 01:40:26 +0200 [thread overview]
Message-ID: <200405200140.26636.maarten@vbvb.nl> (raw)
In-Reply-To: <200405192218.i4JMI9B16930@www.watkins-home.com>
On Thursday 20 May 2004 00:18, Guy wrote:
> I think they fudge the MTBF! They say 1,000,000 hours MTBF.
> That's over 114 years!
> Drives don't last anywhere near that long.
Hear hear! (although the MTBF does mean entirely something else, its use in
marketing, and the ensuing misplaced consumer-confidence, is atrocious.)
> Someone I know has 4 IBM disks. 3 of 4 have failed.
> 1 or 2 were replaced and the replacement drive(s) have failed.
Yeah, it's weird. For me Western Digital drives have failed consistently on
me, but I've had very good experience with Maxtor (read: Quantum) and Hitachi
(read: IBM). As opposed to other posts in this thread...
> All still under warranty! He gave up on IBM since the MTBF seems to be
> less than 1 year. This was about 2-3 years ago. He mirrors thing most of
> the time.
One thing I've come to believe over the years is that heat is a very important
factor that's killing drives. So I now take great care in ensuring good heat
dissipation from the drives. This entails amongst others that you should
never 'sandwich' drives in their 3,5" slots (I can't believe case
manufacturers still have not woken up to this need!). Instead I often arrange
for them to go in 5,25" slots so they have plenty of air around. If I need to
put them in 3,5" slots I always leave 1 unit space around a drive.
In the servers I deploy I take bigger measures, like a bigass 120mm fan just
in front of the drives (accomplished either by dremel or by case design)
> In the past I have almost never had a disk failure. Almost all on my
> drives became too small to use before they failed. The drives ranged from
> 10Meg to 3Gig. Drives larger than 3Gig seem to fail before you out grow
> them.
Hm, no. I did observe some real bad brands / series, even all the way back to
30 MB(M!) RLL drives and a particularly bad batch of 48 MB scsi-1 seagate
ones. But I'll admit that those were the exception to the rule way back then.
> I would love to see real live stats. Claimed MTBF and actual MTBF.
MTBF is measured in a purely statistical way, not taking any _real_ wear and
tear or aging into account. They run 10000 drives for a month and
extrapolate the MTBF from there. The figure is close to meaningless. For
starters, it does not guarantee _anything_. If you have 5 out of 5 failures
within the first six months, that still fits fine inside the statistical
model, unless a lot of others have that same rate of failure. Secondly, one
just cannot extrapolate the life expectancy of a drive in this way and get
useable figures. I can take a statistical test with 20.000 babies during a
year, and perhaps extrapolate from there that the MTBF for humans is 210
years. And boy, we all know that statistic paints a dead wrong picture...!
You need to disregard any and all MTBF values. They serve no purpose for us
end-users. They only serve a purpose for vendors (expected rate of return),
manufacturers and, probably, insurance companies...
> I just checked Seagate and Maxtor. They don't give a MTBF anymore.
> When did that happen!
Well, It was more or less useless anyway. I can tell you just offhand that you
can lengthen the life expectancy of a drive maybe four-fold if you make sure
it stays below 35 degrees celsius its entire life, instead of ~45 degrees.
Don't hold me up on that, but you know what I mean, and it is true. :-)
> Just Service life and Warranty.
> Anyway the best indicator of expected life, the warranty. If the
> manufacture thinks the drive will only last 1 or 3 years (depending on size
> or model), who am I to argue?
Times have indeed changed. 5 or 10 years ago, I would not have hesitated to
put all my data (of which I had little or no backups) on a single 120MB or 2
GB disk. Nowadays, I hardly ever put valueable data on single disks. Either
it has good backups or it goes onto raid 1 or 5 arrays. I've seen it happen
too many times at customers... I do take my precautions now.
(I've been there myself too, and got the T-shirt...)
Not that that guarantees anything... The lightning might strike my 8-disk
fileserver and take out everything. The lightning may hit my house as well
and take any and all but some very very old backups along with it.
But still, chances are much lower and that is what counts, innit ?
If / when a real disaster happens, I'll still live through it. But I just
*need* it to have a much better reason than the ubiquitous drive-failure,
user-error of virus, because *that* I will not forgive myself...
Maarten
--
Yes of course I'm sure it's the red cable. I guarante[^%!/+)F#0c|'NO CARRIER
next prev parent reply other threads:[~2004-05-19 23:40 UTC|newest]
Thread overview: 31+ messages / expand[flat|nested] mbox.gz Atom feed top
2004-05-19 19:58 Hard drive Reliability? John Lange
2004-05-19 20:49 ` Måns Rullgård
2004-05-19 21:28 ` Sevatio
2004-05-19 22:02 ` John Lange
2004-05-19 22:42 ` jim
2004-05-19 22:26 ` Guy
2004-05-19 23:53 ` maarten van den Berg
2004-05-20 0:49 ` TJ Harrell
2004-05-20 1:13 ` berk walker
2004-05-20 6:39 ` Måns Rullgård
2004-05-19 23:44 ` maarten van den Berg
2004-05-19 22:04 ` berk walker
2004-05-19 22:18 ` Guy
2004-05-19 23:40 ` maarten van den Berg [this message]
2004-05-20 0:34 ` Guy
2004-05-20 1:46 ` John Lange
2004-05-20 22:27 ` Russ Price
2004-05-20 15:51 ` Sevatio
2004-05-20 4:39 ` Mark Hahn
2004-05-20 7:15 ` John Lange
2004-05-20 12:15 ` Mark Hahn
2004-05-20 13:26 ` jim
2004-05-20 13:31 ` John Lange
2004-05-20 14:20 ` Mark Hahn
2004-05-20 13:32 ` Tim Grant
2004-05-20 14:38 ` Robin Bowes
2004-05-22 5:15 ` Brad Campbell
2004-05-24 13:25 ` Frank van Maarseveen
2004-05-24 18:35 ` Brad Campbell
2004-05-24 22:38 ` maarten van den Berg
-- strict thread matches above, loose matches on Subject: below --
2004-05-20 13:54 Cress, Andrew R
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=200405200140.26636.maarten@vbvb.nl \
--to=maarten@vbvb.nl \
--cc=linux-raid@vger.kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).