From: Stephan von Krawczynski <skraw@ithnet.com>
To: John Bradford <john@grabjohn.com>
Cc: linux-kernel@vger.kernel.org
Subject: Re: Are linux-fs's drive-fault-tolerant by concept?
Date: Sat, 19 Apr 2003 22:38:58 +0200 [thread overview]
Message-ID: <20030419223858.6781064d.skraw@ithnet.com> (raw)
In-Reply-To: <200304192004.h3JK4Sgc000152@81-2-122-30.bradfords.org.uk>
On Sat, 19 Apr 2003 21:04:28 +0100 (BST)
John Bradford <john@grabjohn.com> wrote:
> > > > I wonder whether it would be a good idea to give the linux-fs
> > > > (namely my preferred reiser and ext2 :-) some fault-tolerance.
> > >
> > > Fault tollerance should be done at a lower level than the filesystem.
> >
> > I know it _should_ to live in a nice and easy world. Unfortunately
> > real life is different. The simple question is: you have tons of
> > low-level drivers for all kinds of storage media, but you have
> > comparably few filesystems. To me this sound like the preferred
> > place for this type of behaviour can be fs, because all drivers
> > inherit the feature if it lives in fs.
>
> The sort of feature you are describing would really belong in a
> separate layer, somewhat analogous to the MD driver, but for defect
> management. You could create a virtual block device that has 90% of
> the capacity of the real block device, then allocte spare blocks from
> the real device if and when blocks failed.
Well, of all work-arounds for the problem this one is probably the worst: it
wastes space on good drives and runs out of space for sure on bad ones.
What is so bad about the simple way: the one who wants to write (e.g. fs) and
knows _where_ to write simply uses another newly allocated block and dumps the
old one on a blacklist. The blacklist only for being able to count them (or get
the sector-numbers) in case you are interested. If you weren't you might as
well mark them allocated and that's it (which I would presume a _bad_ idea). If
there are no free blocks left, well, then the medium is full. And that is just
about the only cause for a write error then (if the medium is writeable at
all).
Don't make the thing bigger than it really is...
>
> > > The filesystem doesn't know or care what device it is stored on, and
> > > therefore shouldn't try to predict likely failiures.
> >
> > It should not predict failures, it should possibly only say: "ok, driver
> > told me the block I just wanted to write has an error, so lets try a
> > different one and mark this block in my personal bad-block list as
> > unusable. This does not sound all-too complex. There is a free-block-list
> > anyway...
>
> Modern disk devices typically already do this kind of defect management.
_should_ do, or do you know for sure?
Take real life: it is near midnight and your ide hd has just filled its last
available bad-block-mapping. The next write comes in and your fs goes boom. You
notice next morning and work is waiting for you. Can you mount it again? Is
your data being corrupted? Well, you need some luck ...
On the other hand, you could have been informed by mail that your ide-drives'
fs has just mapped another bad block and the total number of bad blocks just
reached 4 percent of the available space on the drive.
Why do you trust hd manufacturing sites in Malaysia or Hungary or whereever,
when you can make _sure_ yourself?
I don't get it, sorry.
> > > [1] Although it is uncommon to use a tape as a block device, you never
> > > know. It's certainly possible, (not necessarily with Linux).
> >
> > From the fs point of view it makes no difference living on disk or tape or
> > a tesa-strip.
>
> It does if you are trying to avoid likely failiures, which is what I
> was originally thinking about.
Well, this is for sure nothing for a fs, you are right. But lets fix the _easy_
things first ...
--
Regards,
Stephan
next prev parent reply other threads:[~2003-04-19 20:27 UTC|newest]
Thread overview: 74+ messages / expand[flat|nested] mbox.gz Atom feed top
2003-04-19 16:04 Are linux-fs's drive-fault-tolerant by concept? Stephan von Krawczynski
2003-04-19 15:29 ` Alan Cox
2003-04-19 17:00 ` Stephan von Krawczynski
2003-04-19 22:04 ` Alan Cox
2003-04-20 16:24 ` Stephan von Krawczynski
2003-04-20 13:59 ` John Bradford
2003-04-20 16:55 ` Stephan von Krawczynski
2003-04-20 17:12 ` John Bradford
2003-04-20 17:21 ` Stephan von Krawczynski
2003-04-20 18:48 ` Alan Cox
2003-04-20 20:00 ` John Bradford
2003-04-21 1:51 ` jw schultz
2003-04-19 21:13 ` Jos Hulzink
2003-04-20 16:07 ` Stephan von Krawczynski
2003-04-20 16:40 ` John Bradford
2003-04-20 17:01 ` Stephan von Krawczynski
2003-04-20 17:20 ` John Bradford
2003-04-21 9:32 ` Stephan von Krawczynski
2003-04-21 9:55 ` John Bradford
2003-04-21 11:24 ` Stephan von Krawczynski
2003-04-21 11:50 ` Alan Cox
2003-04-21 12:14 ` John Bradford
2003-04-19 16:22 ` John Bradford
2003-04-19 16:36 ` Russell King
2003-04-19 16:45 ` John Bradford
2003-04-19 16:52 ` Stephan von Krawczynski
2003-04-19 20:04 ` John Bradford
2003-04-19 20:33 ` Andreas Dilger
2003-04-21 9:25 ` Denis Vlasenko
2003-04-21 9:42 ` John Bradford
2003-04-21 10:25 ` Stephan von Krawczynski
2003-04-21 10:50 ` John Bradford
2003-04-19 20:38 ` Stephan von Krawczynski [this message]
2003-04-20 14:21 ` John Bradford
2003-04-21 9:09 ` Denis Vlasenko
2003-04-21 9:35 ` John Bradford
2003-04-21 11:03 ` Stephan von Krawczynski
2003-04-21 12:04 ` John Bradford
2003-04-21 11:22 ` Denis Vlasenko
2003-04-21 11:46 ` Stephan von Krawczynski
2003-04-21 12:13 ` John Bradford
2003-04-19 20:05 ` John Bradford
2003-04-19 23:13 ` Arnaldo Carvalho de Melo
2003-04-19 17:54 ` Felipe Alfaro Solana
2003-04-25 0:07 ` Stewart Smith
2003-04-25 0:52 ` Richard B. Johnson
2003-04-25 7:13 ` John Bradford
[not found] ` <20030419161011$0136@gated-at.bofh.it>
2003-04-19 17:18 ` Florian Weimer
2003-04-19 18:07 ` Stephan von Krawczynski
2003-04-19 18:41 ` Dr. David Alan Gilbert
2003-04-19 20:56 ` Helge Hafting
2003-04-19 21:15 ` Valdis.Kletnieks
2003-04-20 10:51 ` Helge Hafting
2003-04-20 19:04 ` Valdis.Kletnieks
2003-04-19 21:57 ` Alan Cox
2003-04-20 10:09 ` Geert Uytterhoeven
2003-04-21 8:37 ` Denis Vlasenko
2003-05-05 12:38 ` Pavel Machek
2003-04-19 22:02 ` Alan Cox
2003-04-20 8:41 ` Arjan van de Ven
2003-04-25 0:11 ` Stewart Smith
-- strict thread matches above, loose matches on Subject: below --
2003-04-20 15:06 Chuck Ebbert
2003-04-20 15:19 ` John Bradford
2003-04-20 17:03 Chuck Ebbert
2003-04-20 17:25 ` John Bradford
2003-04-20 17:28 Chuck Ebbert
2003-04-21 9:36 ` Stephan von Krawczynski
2003-04-20 17:28 Chuck Ebbert
2003-04-20 17:44 Chuck Ebbert
2003-04-20 17:44 Chuck Ebbert
[not found] <mail.linux.kernel/20030420185512.763df745.skraw@ithnet.com>
[not found] ` <03Apr21.020150edt.41463@gpu.utcc.utoronto.ca>
2003-04-21 11:19 ` Stephan von Krawczynski
2003-04-21 11:52 ` Alan Cox
2003-04-21 14:14 ` Valdis.Kletnieks
2003-05-06 7:03 ` Mike Fedyk
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20030419223858.6781064d.skraw@ithnet.com \
--to=skraw@ithnet.com \
--cc=john@grabjohn.com \
--cc=linux-kernel@vger.kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox