From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from cuda.sgi.com (cuda2.sgi.com [192.48.176.25]) by oss.sgi.com (8.14.3/8.14.3/SuSE Linux 0.8) with ESMTP id q14K4LPh237382 for ; Sat, 4 Feb 2012 14:04:21 -0600 Received: from smtp.pobox.com (b-pb-sasl-quonix.pobox.com [208.72.237.35]) by cuda.sgi.com with ESMTP id ozhagcqAOKWORhzo for ; Sat, 04 Feb 2012 12:04:20 -0800 (PST) Date: Sat, 4 Feb 2012 20:04:17 +0000 From: Brian Candler Subject: Re: Performance problem - reads slower than writes Message-ID: <20120204200417.GA3362@nsrc.org> References: <20120130220019.GA45782@nsrc.org> <20120131020508.GF9090@dastard> <20120131103126.GA46170@nsrc.org> <20120131145205.GA6607@infradead.org> <20120203115434.GA649@nsrc.org> <4F2C38BE.2010002@hardwarefreak.com> <20120203221015.GA2675@nsrc.org> <4F2D016C.9020406@hardwarefreak.com> <20120204112436.GA3167@nsrc.org> <4F2D2953.2020906@hardwarefreak.com> MIME-Version: 1.0 Content-Disposition: inline In-Reply-To: <4F2D2953.2020906@hardwarefreak.com> List-Id: XFS Filesystem from SGI List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Content-Type: text/plain; charset="us-ascii" Content-Transfer-Encoding: 7bit Sender: xfs-bounces@oss.sgi.com Errors-To: xfs-bounces@oss.sgi.com To: Stan Hoeppner Cc: Christoph Hellwig , xfs@oss.sgi.com On Sat, Feb 04, 2012 at 06:49:23AM -0600, Stan Hoeppner wrote: > Apparently you've read of a different GlusterFS. The one I know of is > for aggregating multiple storage hosts into a cloud storage resource. > It is not designed to replace striping or concatenation of disks within > a single host. Sure it can. A gluster volume consists of "bricks". Each brick is served by a glusterd process listening on a different TCP port. Those bricks can be on the same server or on different servers. > Even if what you describe can be done with Gluster, the performance will > likely be significantly less than a properly setup mdraid or hardware > raid. Again, if it can be done, I'd test it head-to-head against RAID. I'd expect similar throughput but higher latency. Given that I'm using low RPM drives which already have high latency, I'm hoping the additional latency will be insignificant. Anyway, I'll know more once I've done the measurements. > I've never been a fan of parity RAID, let alone double parity RAID. I'm with you on that one. The attractions of gluster are: - being able to scale a volume across many nodes, transparently to the clients - being able to take a whole node out of service, while clients automatically flip over to the other Regards, Brian. _______________________________________________ xfs mailing list xfs@oss.sgi.com http://oss.sgi.com/mailman/listinfo/xfs