linux-raid.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: Stan Hoeppner <stan@hardwarefreak.com>
To: John Robinson <john.robinson@anonymous.org.uk>
Cc: Alexander Schreiber <als@thangorodrim.de>,
	Linux-RAID <linux-raid@vger.kernel.org>
Subject: Re: raid over ethernet
Date: Sat, 29 Jan 2011 17:04:31 -0600	[thread overview]
Message-ID: <4D449CFF.9010701@hardwarefreak.com> (raw)
In-Reply-To: <4D448CAF.6020409@anonymous.org.uk>

John Robinson put forth on 1/29/2011 3:54 PM:

> Now that is interesting, to me at least. More as a thought experiment for now, I
> was wondering how one would go about setting up a small cluster of commodity
> servers (maybe 8 machines) running Xen (or perhaps now KVM) VMs, such that if
> one (or potentially two) of the machines died, the VMs could be picked up by the
> other machines in the cluster, and only using locally-attached SATA/SAS discs in
> each machine.

Doing N-way active replication with DRBD increases network utilization
substantially.  With two DRBD active nodes you will have a maximum of _2_
simultaneous data streams, one in each direction.  With 8 active nodes you will
have a maximum of _56_ simultaneous data streams.  Your scenario requires all
nodes be active.

This may work for a hobby cluster or something with very low volume of data
being written to disk.  This solution most likely won't scale for a cluster with
any amount of real traffic.  GbE peaks at 100 MB/s.  Therefore each node will
have only about 12 MB/s of bidirectional bandwidth for each other cluster member
if my math is correct.  A single SATA disk run about 80-120 MB/s, so your
network DRBD disk bandwidth is about 1/7th to 1/10th that of a single local
disk.  In a 2 node cluster it's closer to 1:1.  For you scenario to actually be
feasible, you'd need at least bonded quad GbE interfaces if not single 10 GbE
interfaces to get all the bandwidth you'd need.

You'd be _MUCH_ better off using 2 active DRBD mirrored NFS servers with GFS2
filesystems and having the aforementioned 8 nodes do their data sharing via NFS.
 In this setup each node only writes once (to NFS) dramatically reducing network
bandwidth required per node, with only 16 maximum data streams instead of 56.
If you need more bandwidth or IOPS than a single disk NFS server can produce,
simply RAID 4-10 disks on each NFS server via RAID 10, then mirror the two RAIDs
with DRBD.

You may need 2-4 GbE interfaces between the two NFS servers just for DRBD
traffic, but the cost of that is much less than having the same number of
interfaces in each of 8 cluster nodes.  This will also give you much better
performance after a node or two fails and you have to boot their VM guests on
other hosts.  Having fast central RAID storage will allow those guests to boot
much more quickly and without causing degraded performance on the other nodes
due to lack of disk bandwidth in your suggested model.

-- 
Stan

  reply	other threads:[~2011-01-29 23:04 UTC|newest]

Thread overview: 22+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2011-01-29  1:58 raid over ethernet Roberto Spadim
2011-01-29  5:41 ` Jérôme Poulin
2011-01-29  6:42   ` Roberto Spadim
2011-01-29 13:29     ` Alexander Schreiber
2011-01-29  6:42 ` Mikael Abrahamsson
2011-01-29  6:44   ` Roberto Spadim
2011-01-29  6:48     ` Roberto Spadim
     [not found]       ` <AANLkTikdahgMoJjGr2otTS70LSM77GNpW_vAkZf15Kph@mail.gmail.com>
2011-01-29 11:47         ` Roberto Spadim
2011-01-29 13:34     ` Alexander Schreiber
     [not found]       ` <AANLkTi=6ridRPnHpfdOC=f2_ESndSARmQRkvT_shYO3s@mail.gmail.com>
2011-01-29 14:25         ` Denis
2011-01-29 21:08         ` Alexander Schreiber
2011-01-29 21:54           ` John Robinson
2011-01-29 23:04             ` Stan Hoeppner [this message]
2011-01-29 23:06             ` Miles Fidelman
2011-01-30  1:43             ` Alexander Schreiber
2011-01-31  8:42           ` Denis
2011-01-31 13:03             ` Alexander Schreiber
2011-01-31 14:45               ` Roberto Spadim
2011-01-31 16:15                 ` Alexander Schreiber
2011-01-31 17:37                   ` Roberto Spadim
2011-01-29 15:30     ` Spelic
2011-01-29 18:34   ` David Brown

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=4D449CFF.9010701@hardwarefreak.com \
    --to=stan@hardwarefreak.com \
    --cc=als@thangorodrim.de \
    --cc=john.robinson@anonymous.org.uk \
    --cc=linux-raid@vger.kernel.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).