From: Joe Landman <landman@scalableinformatics.com>
To: xfs@oss.sgi.com
Subject: Re: Performance problem - reads slower than writes
Date: Sat, 04 Feb 2012 15:44:25 -0500 [thread overview]
Message-ID: <4F2D98A9.4090709@scalableinformatics.com> (raw)
In-Reply-To: <20120204200417.GA3362@nsrc.org>
On 02/04/2012 03:04 PM, Brian Candler wrote:
> On Sat, Feb 04, 2012 at 06:49:23AM -0600, Stan Hoeppner wrote:
[...]
> Sure it can. A gluster volume consists of "bricks". Each brick is served by
> a glusterd process listening on a different TCP port. Those bricks can be on
> the same server or on different servers.
I seem to remember that the Gluster folks abandoned this model (using
their code versus MD raid) on single servers due to performance issues.
We did play with this a few times, and the performance wasn't that
good. Basically limited by single disk seek/write speed.
>
>> Even if what you describe can be done with Gluster, the performance will
>> likely be significantly less than a properly setup mdraid or hardware
>> raid. Again, if it can be done, I'd test it head-to-head against RAID.
>
> I'd expect similar throughput but higher latency. Given that I'm using low
My recollection is that this wasn't the case. Performance was
suboptimal in all cases we tried.
> RPM drives which already have high latency, I'm hoping the additional
> latency will be insignificant. Anyway, I'll know more once I've done the
> measurements.
I did this with the 3.0.x and the 2.x series of Gluster. Usually atop
xfs of some flavor.
>
>> I've never been a fan of parity RAID, let alone double parity RAID.
>
> I'm with you on that one.
RAID's entire purpose in life is to give an administrator time to run in
and change a disk. RAID isn't a backup, or even a guarantee of data
retention. Many do treat it this way though, to their (and their
data's) peril.
> The attractions of gluster are:
> - being able to scale a volume across many nodes, transparently to
> the clients
This does work, though rebalance is as much a function of the seek and
bandwidth of the slowest link as other things. So if you have 20
drives, and you do a rebalance to add 5 more, its gonna be slow for a
while.
> - being able to take a whole node out of service, while clients
> automatically flip over to the other
>
I hate to put it like this, but this is true for various definitions of
the word "automatically". You need to make sure that your definitions
line up with the reality of "automatic".
If a brick goes away, and you have a file on this brick you want to
overwrite, it doesn't (unless you have a mirror) flip over to another
unit "automatically" or otherwise.
RAID in this case can protect you from some of these issues (single disk
failure issues, being replaced by RAID issues), but unless you are
building mirror pairs of bricks on separate units, this magical
"automatic" isn't quite so.
Moreover, last I checked, Gluster made no guarantees as to the ordering
of the layout for mirrors. So if you have more than one brick per node,
and build mirror pairs with the "replicate" option, you have to check
the actual hashing to make sure it did what you expect. Or build up the
mirror pairs more carefully.
At this point, it sounds like there is a gluster side of this discussion
that I'd recommend you take to the gluster list. There is an xfs
portion as well which is fine here.
Disclosure: we build/sell/support gluster (and other) based systems
atop xfs based RAID units (both hardware and software RAID;
1,10,6,60,...) so we have inherent biases. Your mileage may vary. See
your doctor if your re-balance exceeds 4 hours.
> Regards,
>
> Brian.
Joe
--
Joseph Landman, Ph.D
Founder and CEO
Scalable Informatics Inc.
email: landman@scalableinformatics.com
web : http://scalableinformatics.com
http://scalableinformatics.com/sicluster
phone: +1 734 786 8423 x121
fax : +1 866 888 3112
cell : +1 734 612 4615
_______________________________________________
xfs mailing list
xfs@oss.sgi.com
http://oss.sgi.com/mailman/listinfo/xfs
next prev parent reply other threads:[~2012-02-04 20:44 UTC|newest]
Thread overview: 30+ messages / expand[flat|nested] mbox.gz Atom feed top
2012-01-30 22:00 Performance problem - reads slower than writes Brian Candler
2012-01-31 2:05 ` Dave Chinner
2012-01-31 10:31 ` Brian Candler
2012-01-31 14:16 ` Brian Candler
2012-01-31 20:25 ` Dave Chinner
2012-02-01 7:29 ` Stan Hoeppner
2012-02-03 18:47 ` Brian Candler
2012-02-03 19:03 ` Christoph Hellwig
2012-02-03 21:01 ` Brian Candler
2012-02-03 21:17 ` Brian Candler
2012-02-05 22:50 ` Dave Chinner
2012-02-05 22:43 ` Dave Chinner
2012-01-31 14:52 ` Christoph Hellwig
2012-01-31 21:52 ` Brian Candler
2012-02-01 0:50 ` Raghavendra D Prabhu
2012-02-01 3:59 ` Dave Chinner
2012-02-03 11:54 ` Brian Candler
2012-02-03 19:42 ` Stan Hoeppner
2012-02-03 22:10 ` Brian Candler
2012-02-04 9:59 ` Stan Hoeppner
2012-02-04 11:24 ` Brian Candler
2012-02-04 12:49 ` Stan Hoeppner
2012-02-04 20:04 ` Brian Candler
2012-02-04 20:44 ` Joe Landman [this message]
2012-02-06 10:40 ` Brian Candler
2012-02-07 17:30 ` Brian Candler
2012-02-05 5:16 ` Stan Hoeppner
2012-02-05 9:05 ` Brian Candler
2012-01-31 20:06 ` Dave Chinner
2012-01-31 21:35 ` Brian Candler
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=4F2D98A9.4090709@scalableinformatics.com \
--to=landman@scalableinformatics.com \
--cc=xfs@oss.sgi.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox