public inbox for linux-btrfs@vger.kernel.org
 help / color / mirror / Atom feed
From: Chris Mason <chris.mason@oracle.com>
To: Steven Pratt <slpratt@austin.ibm.com>
Cc: linux-btrfs@vger.kernel.org
Subject: Re: Btrfs experimental branch updates
Date: Tue, 17 Mar 2009 21:18:02 -0400	[thread overview]
Message-ID: <1237339082.31273.49.camel@think.oraclecorp.com> (raw)
In-Reply-To: <49C00EB8.6050200@austin.ibm.com>

On Tue, 2009-03-17 at 15:57 -0500, Steven Pratt wrote:
> Chris Mason wrote:
> > On Sun, 2009-03-15 at 09:38 -0500, Steven Pratt wrote:
> >   
> >>> Thanks for running this, but the main performance fixes for your test
> >>> are still in testing locally.  One thing that makes a huge difference on
> >>> the random write run is to mount -o ssd.
> >>>
> >>>   
> >>>       
> >> Tried a run with -o ssd on the raid system. It made some minor 
> >> improvements in random write performance. Helps more on odirect, but 
> >> mainly at the 16thread count. Single and 128 threads it doesn't make 
> >> much difference.
> >>
> >> Results syncing now to history boxacle
> >> http://btrfs.boxacle.net/repository/raid/history/History.html
> >>     
> >
> > Well, still completely different from my test rig ;)  For the random
> > write run, yours runs at 580 trans/sec for btrfs and mine is going along
> > at 8000 trans/sec.
> >   
> That is odd.  However, I think I have found 1 factor.  In rerunning with 
> blktrace and sysrq an interesting thing happened.  The results got a lot 
> faster.  What I did was just run the 128 thread odirect random write 
> test.  Instead of 2.8MB/sec, I got 17MB/sec. Still far below the 100+ of 
> ext4 and JFS, but one heck of a difference.  Here is what I think is 
> going on.  We make use of a flag in FFSB to reuse the existing fileset 
> if the fileset meets the setup criteria exactly.  For the test I am 
> running that is 1024 100MB files.  Since all of the random write test 
> are doing overwrites within the file, the file sizes do not change and 
> therefore the fileset is valid for reuse.  

Oh! In that case you're stuck waiting to cache the extents already used
in a block group.  At least I hope that's what sysrq-w will show us.
The first mods to a block group after a mount are slow while we read in
the free extents.

-chris



  reply	other threads:[~2009-03-18  1:18 UTC|newest]

Thread overview: 9+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2009-03-13 15:56 Btrfs experimental branch updates Chris Mason
2009-03-13 22:52 ` Steven Pratt
2009-03-14  1:33   ` Chris Mason
2009-03-15 14:38     ` Steven Pratt
2009-03-17  1:24       ` Chris Mason
2009-03-17 20:57         ` Steven Pratt
2009-03-18  1:18           ` Chris Mason [this message]
2009-03-15 19:13 ` Grigory Makarevich
2009-03-16 13:31   ` Chris Mason

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=1237339082.31273.49.camel@think.oraclecorp.com \
    --to=chris.mason@oracle.com \
    --cc=linux-btrfs@vger.kernel.org \
    --cc=slpratt@austin.ibm.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox