From: "Alex,Shi" <alex.shi@intel.com>
To: Jeff Moyer <jmoyer@redhat.com>
Cc: Jan Kara <jack@suse.cz>, Corrado Zoccolo <czoccolo@gmail.com>,
"Li, Shaohua" <shaohua.li@intel.com>,
Vivek Goyal <vgoyal@redhat.com>, "tytso@mit.edu" <tytso@mit.edu>,
"jaxboe@fusionio.com" <jaxboe@fusionio.com>,
"linux-kernel@vger.kernel.org" <linux-kernel@vger.kernel.org>,
"Chen, Tim C" <tim.c.chen@intel.com>
Subject: Re: [performance bug] kernel building regression on 64 LCPUs machine
Date: Tue, 22 Mar 2011 15:38:19 +0800 [thread overview]
Message-ID: <1300779499.30136.353.camel@debian> (raw)
In-Reply-To: <x491v2mivv6.fsf@segfault.boston.devel.redhat.com>
On Sat, 2011-03-05 at 02:27 +0800, Jeff Moyer wrote:
> Jeff Moyer <jmoyer@redhat.com> writes:
>
> > Jan Kara <jack@suse.cz> writes:
> >
> >> I'm not so happy with ext4 results. The difference between ext3 and ext4
> >> might be that amount of data written by kjournald in ext3 is considerably
> >> larger if it ends up pushing out data (because of data=ordered mode) as
> >> well. With ext4, all data are written by filemap_fdatawrite() from fsync
> >> because of delayed allocation. And thus maybe for ext4 WRITE_SYNC_PLUG
> >> is hurting us with your fast storage and small amount of written data? With
> >> WRITE_SYNC, data would be already on it's way to storage before we get to
> >> wait for them...
> >
> >> Or it could be that we really send more data in WRITE mode rather than in
> >> WRITE_SYNC mode with the patch on ext4 (that should be verifiable with
> >> blktrace). But I wonder how that could happen...
> >
> > It looks like this is the case, the I/O isn't coming down as
> > synchronous. I'm seeing a lot of writes, very few write sync's, which
> > means that the write stream will be preempted by the incoming reads.
> >
> > Time to audit that fsync path and make sure it's marked properly, I
> > guess.
>
> OK, I spoke too soon. Here's the blktrace summary information (I re-ran
> the tests using 3 samples, the blktrace is from the last run of the
> three in each case):
>
> Vanilla
> -------
> fs_mark: 307.288 files/sec
> fio: 286509 KB/s
>
> Total (sde):
> Reads Queued: 341,558, 84,994MiB Writes Queued: 1,561K, 6,244MiB
> Read Dispatches: 341,493, 84,994MiB Write Dispatches: 648,046, 6,244MiB
> Reads Requeued: 0 Writes Requeued: 27
> Reads Completed: 341,491, 84,994MiB Writes Completed: 648,021, 6,244MiB
> Read Merges: 65, 2,780KiB Write Merges: 913,076, 3,652MiB
> IO unplugs: 578,102 Timer unplugs: 0
>
> Throughput (R/W): 282,797KiB/s / 20,776KiB/s
> Events (sde): 16,724,303 entries
>
> Patched
> -------
> fs_mark: 278.587 files/sec
> fio: 298007 KB/s
>
> Total (sde):
> Reads Queued: 345,407, 86,834MiB Writes Queued: 1,566K, 6,264MiB
> Read Dispatches: 345,391, 86,834MiB Write Dispatches: 327,404, 6,264MiB
> Reads Requeued: 0 Writes Requeued: 33
> Reads Completed: 345,391, 86,834MiB Writes Completed: 327,371, 6,264MiB
> Read Merges: 16, 1,576KiB Write Merges: 1,238K, 4,954MiB
> IO unplugs: 580,308 Timer unplugs: 0
>
> Throughput (R/W): 288,771KiB/s / 20,832KiB/s
> Events (sde): 14,030,610 entries
>
> So, it appears we flush out writes much more aggressively without the
> patch in place. I'm not sure why the write bandwidth looks to be higher
> in the patched case... odd.
>
Jan:
Do you have new idea on this?
next prev parent reply other threads:[~2011-03-22 7:36 UTC|newest]
Thread overview: 38+ messages / expand[flat|nested] mbox.gz Atom feed top
2011-01-19 1:55 [performance bug] kernel building regression on 64 LCPUs machine Alex,Shi
2011-01-19 2:03 ` Shaohua Li
2011-01-19 12:56 ` Jan Kara
2011-01-20 7:52 ` Alex,Shi
2011-01-20 15:16 ` Vivek Goyal
2011-01-21 7:17 ` Shaohua Li
2011-01-26 8:15 ` Shaohua Li
2011-02-12 9:21 ` Alex,Shi
2011-02-12 18:25 ` Corrado Zoccolo
2011-02-14 2:25 ` Alex,Shi
2011-02-15 1:10 ` Shaohua Li
2011-02-21 16:49 ` Jan Kara
2011-02-23 8:24 ` Alex,Shi
2011-02-24 12:13 ` Jan Kara
2011-02-25 0:44 ` Alex Shi
2011-02-26 14:45 ` Corrado Zoccolo
2011-03-01 19:56 ` Jeff Moyer
2011-03-02 9:42 ` Jan Kara
2011-03-02 16:13 ` Jeff Moyer
2011-03-02 21:17 ` Jan Kara
2011-03-02 21:20 ` Jeff Moyer
2011-03-03 1:14 ` Jeff Moyer
2011-03-04 15:32 ` Jan Kara
2011-03-04 15:40 ` Jeff Moyer
2011-03-04 15:50 ` Jeff Moyer
2011-03-04 18:27 ` Jeff Moyer
2011-03-22 7:38 ` Alex,Shi [this message]
2011-03-22 16:14 ` Jan Kara
2011-03-22 17:46 ` Jeff Moyer
2011-03-24 6:45 ` Alex,Shi
2011-03-28 19:48 ` Jan Kara
2011-01-19 14:32 ` Ted Ts'o
2011-01-20 2:12 ` Shaohua Li
2011-01-21 7:23 ` Corrado Zoccolo
2011-01-21 7:47 ` Alex,Shi
2011-01-21 7:52 ` Alex,Shi
2011-01-21 8:13 ` Corrado Zoccolo
2011-01-21 8:20 ` Shaohua Li
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=1300779499.30136.353.camel@debian \
--to=alex.shi@intel.com \
--cc=czoccolo@gmail.com \
--cc=jack@suse.cz \
--cc=jaxboe@fusionio.com \
--cc=jmoyer@redhat.com \
--cc=linux-kernel@vger.kernel.org \
--cc=shaohua.li@intel.com \
--cc=tim.c.chen@intel.com \
--cc=tytso@mit.edu \
--cc=vgoyal@redhat.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox