public inbox for linux-xfs@vger.kernel.org
 help / color / mirror / Atom feed
From: Emmanuel Florac <eflorac@intellique.com>
To: Christian Herzog <horeizo@phys.ethz.ch>
Cc: xfs@oss.sgi.com
Subject: Re: XFS errors on large Infiniband fileserver setup
Date: Fri, 24 Sep 2010 15:19:18 +0200	[thread overview]
Message-ID: <20100924151918.598148c6@harpe.intellique.com> (raw)
In-Reply-To: <2276da2491527ca8044fa1daec496b48@phys.ethz.ch>

Le Fri, 24 Sep 2010 07:41:53 +0200
Christian Herzog <horeizo@phys.ethz.ch> écrivait:

>  We start off
> with 52T and can easily add additional disk units to the Infiniband
> switch.

I don't know if using iscsi-over-infiniband is optimal. The problem is
that you plan to expand your existing XFS filesystem by leaps and
bounds up to some extremely large size, through a simple raid-0 like
aggregation, that's really fragile.

It's a configuration that's supposed to work well with something like
ZFS, though in real life setups (no I won't tell who sent back recently
all of a 2 PB cluster to Sun but that happened in 2010 :) large raid-z
iscsi clusters aren't so great :)

For similar setups, I used PVFS2 by aggregating 40 TB nodes. PVFS2 is
known to scale up to petabytes, and (contrary to XFS over RAID-0) is
extremely tolerant to node failure (though it is not redundant); if a
node crashes, the cluster IO may freeze (though write activity can
usually go on) but restart instantly when the failed node is revived.

However PVFS isn't made for general purpose file sharing (though it
works with both samba and nfs), but really flies when used with
applications properly set up (MPIO). It's tailored for scientific work
and heavy computation clusters.

In contrast with Lustre, PVFS2 is very easy to set up, and very easy
to extend if you planned it from the start (Lustre is a fantastic PITA
to set up and administer, and don't even talk about NFS sharing).

So I would set up a storage cluster this way : each storage node is a
PVFS server, the PVFS data resides on an XFS filesystem (officially
recommended by PVFS developers anyway).
You can expand the PVFS filesystem either by enlarging the XFS on the
storage nodes, or by adding new independant storage nodes.

Each storage node can be a PVFS client too, and use its computing power
to crunch data. 

As I said, the main problem is to know how you plan to make space
available to clients systems. You can use NFS/CIFS for ease of use,
desktop access, etc but performance will be low. However native PVFS
performance can be huge over infiniband (in the several GB/s range).
And the more storage nodes you're adding, the more performance you get.

-- 
------------------------------------------------------------------------
Emmanuel Florac     |   Direction technique
                    |   Intellique
                    |	<eflorac@intellique.com>
                    |   +33 1 78 94 84 02
------------------------------------------------------------------------

_______________________________________________
xfs mailing list
xfs@oss.sgi.com
http://oss.sgi.com/mailman/listinfo/xfs

  parent reply	other threads:[~2010-09-24 13:18 UTC|newest]

Thread overview: 12+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2010-09-23  7:22 XFS errors on large Infiniband fileserver setup Christian Herzog
2010-09-23 10:01 ` Christian Herzog
2010-09-23 10:38   ` Christoph Hellwig
2010-09-23 11:52     ` Christian Herzog
2010-09-23 13:54       ` Christoph Hellwig
2010-09-23 14:10         ` Christian Herzog
2010-09-23 23:53 ` Dave Chinner
2010-09-24  5:41   ` Christian Herzog
2010-09-24  8:56     ` Stan Hoeppner
2010-09-24 13:19     ` Emmanuel Florac [this message]
2010-09-24 13:19   ` Emmanuel Florac
2010-09-27  5:49 ` Christian Herzog

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20100924151918.598148c6@harpe.intellique.com \
    --to=eflorac@intellique.com \
    --cc=horeizo@phys.ethz.ch \
    --cc=xfs@oss.sgi.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox