From: Martin Steigerwald <Martin@lichtvoll.de>
To: xfs@oss.sgi.com
Cc: Stefan Ring <stefanrin@gmail.com>, Christoph Hellwig <hch@infradead.org>
Subject: Re: XFS: Abysmal write performance because of excessive seeking (allocation groups to blame?)
Date: Sat, 7 Apr 2012 20:57:38 +0200 [thread overview]
Message-ID: <201204072057.38286.Martin@lichtvoll.de> (raw)
In-Reply-To: <CAAxjCExBcaB6J-u7ivZKWnKiF7oP10JRxzKzQNRuppHkTE2Tzw@mail.gmail.com>
Am Freitag, 6. April 2012 schrieb Stefan Ring:
> > thanks for the detailed report.
>
> Thanks for the detailed and kind answer.
>
> > Can you try a few mount options for me both all together and if you
> > have some time also individually.
> >
> > -o inode64
> >
> > This allows inodes to be close to data even for >1TB
> > filesystems. It's something we hope to make the default soon.
>
> The filesystem is not that large. It’s only 400GB. I turned it on
> anyway. No difference.
>
> > -o filestreams
> >
> > This keeps data written in a single directory group together.
> > Not sure your directories are large enough to really benefit
> > from it, but it's worth a try.
> > -o allocsize=4k
> >
> > This disables the agressive file preallocation we do in XFS,
> > which sounds like it's not useful for your workload.
>
> inode64+filestreams: no difference
> inode64+allocsize: no difference
> inode64+filestreams+allocsize: no difference :(
>
> > For metadata intensive workloads like yours you would be much better
> > using a non-striping raid, e.g. concatentation and mirroring instead
> > of raid 5 or raid 6. I know this has a cost in terms of "wasted"
> > space, but for IOPs bound workload the difference is dramatic.
>
> Hmm, I’m sure you’re right, but I’m out of luck here. If I had 24
> drives, I could think about a different organization. But with only 6
> bays, I cannot give up all that space.
>
> Although *in theory*, it *should* be possible to run fast for
> write-only workloads. The stripe size is 64 KB (4x16), and it’s not
> like data is written all over the place. So it should very well be
> possible to write the data out in some reasonably sized and aligned
> chunks. The filesystem partition itself is nicely aligned.
And is XFS aligned to the RAID 6?
What does xfs_info display on it?
--
Martin 'Helios' Steigerwald - http://www.Lichtvoll.de
GPG: 03B0 0D6C 0040 0710 4AFA B82F 991B EAAC A599 84C7
_______________________________________________
xfs mailing list
xfs@oss.sgi.com
http://oss.sgi.com/mailman/listinfo/xfs
next prev parent reply other threads:[~2012-04-07 18:57 UTC|newest]
Thread overview: 64+ messages / expand[flat|nested] mbox.gz Atom feed top
2012-04-05 18:10 XFS: Abysmal write performance because of excessive seeking (allocation groups to blame?) Stefan Ring
2012-04-05 19:56 ` Peter Grandi
2012-04-05 22:41 ` Peter Grandi
2012-04-06 14:36 ` Peter Grandi
2012-04-06 15:37 ` Stefan Ring
2012-04-07 13:33 ` Peter Grandi
2012-04-05 21:37 ` Christoph Hellwig
2012-04-06 1:09 ` Peter Grandi
2012-04-06 8:25 ` Stefan Ring
2012-04-07 18:57 ` Martin Steigerwald [this message]
2012-04-10 14:02 ` Stefan Ring
2012-04-10 14:32 ` Joe Landman
2012-04-10 15:56 ` Stefan Ring
2012-04-10 18:13 ` Martin Steigerwald
2012-04-10 20:44 ` Stan Hoeppner
2012-04-10 21:00 ` Stefan Ring
2012-04-05 22:32 ` Roger Willcocks
2012-04-06 7:11 ` Stefan Ring
2012-04-06 8:24 ` Stefan Ring
2012-04-05 23:07 ` Peter Grandi
2012-04-06 0:13 ` Peter Grandi
2012-04-06 7:27 ` Stefan Ring
2012-04-06 23:28 ` Stan Hoeppner
2012-04-07 7:27 ` Stefan Ring
2012-04-07 8:53 ` Emmanuel Florac
2012-04-07 14:57 ` Stan Hoeppner
2012-04-09 11:02 ` Stefan Ring
2012-04-09 12:48 ` Emmanuel Florac
2012-04-09 12:53 ` Stefan Ring
2012-04-09 13:03 ` Emmanuel Florac
2012-04-09 23:38 ` Stan Hoeppner
2012-04-10 6:11 ` Stefan Ring
2012-04-10 20:29 ` Stan Hoeppner
2012-04-10 20:43 ` Stefan Ring
2012-04-10 21:29 ` Stan Hoeppner
2012-04-09 0:19 ` Dave Chinner
2012-04-09 11:39 ` Emmanuel Florac
2012-04-09 21:47 ` Dave Chinner
2012-04-07 8:49 ` Emmanuel Florac
2012-04-08 20:33 ` Stan Hoeppner
2012-04-08 21:45 ` Emmanuel Florac
2012-04-09 5:27 ` Stan Hoeppner
2012-04-09 12:45 ` Emmanuel Florac
2012-04-13 19:36 ` Stefan Ring
2012-04-14 7:32 ` Stan Hoeppner
2012-04-14 11:30 ` Stefan Ring
2012-04-09 14:21 ` Geoffrey Wehrman
2012-04-10 19:30 ` Stan Hoeppner
2012-04-11 22:19 ` Geoffrey Wehrman
2012-04-07 16:50 ` Peter Grandi
2012-04-07 17:10 ` Joe Landman
2012-04-08 21:42 ` Stan Hoeppner
2012-04-09 5:13 ` Stan Hoeppner
2012-04-09 11:52 ` Stefan Ring
2012-04-10 7:34 ` Stan Hoeppner
2012-04-10 13:59 ` Stefan Ring
2012-04-09 9:23 ` Stefan Ring
2012-04-09 23:06 ` Stan Hoeppner
2012-04-06 0:53 ` Peter Grandi
2012-04-06 7:32 ` Stefan Ring
2012-04-06 5:53 ` Stefan Ring
2012-04-06 15:35 ` Peter Grandi
2012-04-10 14:05 ` Stefan Ring
2012-04-07 19:11 ` Peter Grandi
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=201204072057.38286.Martin@lichtvoll.de \
--to=martin@lichtvoll.de \
--cc=hch@infradead.org \
--cc=stefanrin@gmail.com \
--cc=xfs@oss.sgi.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox