From: Andreas Dilger <adilger@sun.com>
To: Valdis.Kletnieks@vt.edu
Cc: Andrew Morton <akpm@linux-foundation.org>,
Abhishek Rai <abhishekrai@google.com>,
linux-kernel@vger.kernel.org, rohitseth@google.com,
linux-ext4@vger.kernel.org
Subject: Re: [CALL FOR TESTING] Make Ext3 fsck way faster [2.6.24-rc6 -mm patch]
Date: Thu, 17 Jan 2008 04:36:05 -0700 [thread overview]
Message-ID: <20080117113605.GT3351@webber.adilger.int> (raw)
In-Reply-To: <25815.1200457538@turing-police.cc.vt.edu>
On Jan 15, 2008 23:25 -0500, Valdis.Kletnieks@vt.edu wrote:
> I've got multiple boxes across the hall that have 50T of disk on them, in one
> case as one large filesystem, and the users want *more* *bigger* still (damned
> researchers - you put a 15 teraflop supercomputer in the room, and then they
> want someplace to *put* all the numbers that come spewing out of there.. ;)
>
> There comes a point where that downtime gets too long to be politically
> expedient. 6->2 may not be a biggie, because you can likely get a 6 hour
> window. 24->8 suddenly looks a lot different.
>
> (Having said that, I'll admit the one 52T filesystem is an SGI Itanium box
> running Suse and using XFS rather than ext3).
>
> Has anybody done a back-of-envelope of what this would do for fsck times for
> a "max realistically achievable ext3 filesystem" (i.e. 100T-200T or ext3
> design limit, whichever is smaller)?
This is exactly the kind of environment that Lustre is designed for.
Not only do you get parallel IO performance, but you can also do parallel
e2fsck on the individual filesystems when you need it. Not that we aren't
also working on improving e2fsck performance, but 100 * 4TB e2fsck in
parallel is much better than 1 * 400TB e2fsck (probably not possible on
a system today due to RAM constraints though I haven't really done any
calculations either way). I know customers were having RAM problems with
3 * 2TB e2fsck in parallel on a 2GB node.
Most customers these days use 2-4 4-8TB filesystems per server
(inexpensive SMP node with 2-4GB RAM instead of a single monstrous SGI box).
We have many Lustre filesystems in the 100-200TB range today, some over 1PB
already and much larger ones being planned.
> (And one of the research crew had a not-totally-on-crack proposal to get a
> petabyte of spinning oxide. Figuring out how to back that up would probably
> have landed firmly in my lap. Ouch. ;)
Charge them for 2PB of storage, and use rsync :-).
Cheers, Andreas
--
Andreas Dilger
Sr. Staff Engineer, Lustre Group
Sun Microsystems of Canada, Inc.
next prev parent reply other threads:[~2008-01-17 11:36 UTC|newest]
Thread overview: 17+ messages / expand[flat|nested] mbox.gz Atom feed top
[not found] <200801140839.01986.abhishekrai@google.com>
2008-01-15 0:34 ` [CALL FOR TESTING] Make Ext3 fsck way faster [2.6.24-rc6 -mm patch] Andrew Morton
2008-01-15 11:04 ` Andrew Morton
2008-01-15 13:15 ` Christoph Hellwig
2008-01-15 13:16 ` Christoph Hellwig
2008-01-15 15:28 ` Theodore Tso
2008-01-17 12:47 ` Abhishek Rai
2008-01-20 4:10 ` Daniel Phillips
2008-01-21 2:51 ` Theodore Tso
2008-01-24 19:04 ` Daniel Phillips
2008-01-15 15:09 ` Ric Wheeler
2008-01-16 5:08 ` Valdis.Kletnieks
2008-01-16 4:25 ` Valdis.Kletnieks
2008-01-17 11:36 ` Andreas Dilger [this message]
2008-01-20 3:55 ` Daniel Phillips
2008-01-23 9:12 Abhishek Rai
2008-01-24 7:49 ` Andrew Morton
2008-01-24 13:14 ` Abhishek Rai
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20080117113605.GT3351@webber.adilger.int \
--to=adilger@sun.com \
--cc=Valdis.Kletnieks@vt.edu \
--cc=abhishekrai@google.com \
--cc=akpm@linux-foundation.org \
--cc=linux-ext4@vger.kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=rohitseth@google.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox