public inbox for linux-xfs@vger.kernel.org
 help / color / mirror / Atom feed
From: Stan Hoeppner <stan@hardwarefreak.com>
To: Carsten Aulbert <Carsten.Aulbert@aei.mpg.de>,
	Brian Foster <bfoster@redhat.com>
Cc: xfs@oss.sgi.com
Subject: Re: extremely slow file creation/deletion after xfs ran full
Date: Tue, 13 Jan 2015 14:43:19 -0600	[thread overview]
Message-ID: <54B58367.1070008@hardwarefreak.com> (raw)
In-Reply-To: <54B57C59.9070207@aei.mpg.de>

On 01/13/2015 02:13 PM, Carsten Aulbert wrote:
> Hi Stan
> 
> On 01/13/2015 09:06 PM, Stan Hoeppner wrote:
>> This workload seems more suited to a database than a filesystem. Though
>> surely you've already considered such, and chose not to go that route.
>>
> 
> Yepp, but as we do not fully control the server software and need to
> work further on the binary blobs arriving, a database is also not that
> well suited for it, but yes, we looked into it (and run mysql, marida,
> cassandra, mongo, postgresql, ...)
> 
>> With high fragmentation you get lots of seeking.  What model disks are
>> these?  What is your RAID10 geometry?  Are your partitions properly
>> aligned to that geometry, and to the drives (512n/512e)?
> 
> Disks are 2TB Hitachi SATA drives (Ultrastar, HUA722020ALA330). As these
> are some yrs old, they are native 512byte ones. They are connected via
> an Areca 1261ML controller with a Supermicro backplane.
> 
> RAID striping is not ideal (128kByte per member disk) and thus our xfs
> layout is not ideal as well. Things we plan to change with the next
> attempt ;)

With your file sizes that seems a recipe for hotspots.  What do the
controller metrics tell you about IOs per drive, bytes per drive?  Are
they balanced?

> After the arrival of "advanced format" HDD and SSDs we usually try to
> align everything to full 1 MByte or larger, just to be sure any
> combination of 512b, 4kb, ... will eventually align :)

It's not that simple with striping.  Partitions need to start and end on
stripe boundaries, not simply multiples of 4KB or 1MB as with single
disks. If you use non power of 2 drive counts in a stripe, aligning at
multiples of 1MB will screw ya, e.g. 12 drives * 64KB is a 768KB stripe.

Stan



_______________________________________________
xfs mailing list
xfs@oss.sgi.com
http://oss.sgi.com/mailman/listinfo/xfs

  reply	other threads:[~2015-01-13 20:42 UTC|newest]

Thread overview: 14+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2015-01-12  8:36 extremely slow file creation/deletion after xfs ran full Carsten Aulbert
2015-01-12 12:44 ` Brian Foster
2015-01-12 13:30 ` Carsten Aulbert
2015-01-12 15:52   ` Brian Foster
2015-01-12 16:09     ` Carsten Aulbert
2015-01-12 16:37       ` Brian Foster
2015-01-12 17:33         ` Carsten Aulbert
2015-01-13 20:06           ` Stan Hoeppner
2015-01-13 20:13             ` Carsten Aulbert
2015-01-13 20:43               ` Stan Hoeppner [this message]
2015-01-14  6:07                 ` Carsten Aulbert
2015-01-13 20:33           ` Dave Chinner
2015-01-14  6:12             ` Carsten Aulbert
2015-01-16 15:35               ` Carlos Maiolino

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=54B58367.1070008@hardwarefreak.com \
    --to=stan@hardwarefreak.com \
    --cc=Carsten.Aulbert@aei.mpg.de \
    --cc=bfoster@redhat.com \
    --cc=xfs@oss.sgi.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox