From: Badari Pulavarty <pbadari@us.ibm.com>
To: Bryan Henderson <hbryan@us.ibm.com>
Cc: Andreas Dilger <adilger@clusterfs.com>,
linux-fsdevel@vger.kernel.org, Sonny Rao <sonny@burdell.org>
Subject: Re: ext3 writepages ?
Date: 10 Feb 2005 10:32:05 -0800 [thread overview]
Message-ID: <1108060325.20053.1145.camel@dyn318077bld.beaverton.ibm.com> (raw)
In-Reply-To: <OF61009C01.96F70B2E-ON88256FA4.006218C6-88256FA4.0062EA88@us.ibm.com>
On Thu, 2005-02-10 at 10:00, Bryan Henderson wrote:
> >Don't you think, filesystems submitting biggest chunks of IO
> >possible is better than submitting 1k-4k chunks and hoping that
> >IO schedulers do the perfect job ?
>
> No, I don't see why it would better. In fact intuitively, I think the I/O
> scheduler, being closer to the device, should do a better job of deciding
> in what packages I/O should go to the device. After all, there exist
> block devices that don't process big chunks faster than small ones. But
>
> So this starts to look like something where you withhold data from the I/O
> scheduler in order to prevent it from scheduling the I/O wrongly because
> you (the pager/filesystem driver) know better. That shouldn't be the
> architecture.
>
> So I'd like still like to see a theory that explains why submitting the
> I/O a little at a time (i.e. including the bio_submit() in the loop that
> assembles the I/O) causes the device to be idle more.
>
> >We all learnt thro 2.4 RAW code about the overhead of doing 512bytes
> >IO and making the elevator merge all the peices together.
>
> That was CPU time, right? In the present case, the numbers say it takes
> the same amount of CPU time to assemble the I/O above the I/O scheduler as
> inside it.
One clear distinction between submitting smaller chunks vs larger
ones is - number of call backs we get and the processing we need to
do.
I don't think we have enough numbers here to get to bottom of this.
CPU utilization remains same in both cases, doesn't mean that - the
test took exactly same amount of time. I don't even think that we
are doing a fixed number of IOs. Its possible that by doing larger
IOs we save CPU and use that CPU to push more data ?
Thanks,
Badari
next prev parent reply other threads:[~2005-02-10 18:31 UTC|newest]
Thread overview: 21+ messages / expand[flat|nested] mbox.gz Atom feed top
2005-02-02 15:32 ext3 writepages ? Badari Pulavarty
2005-02-02 20:19 ` Sonny Rao
2005-02-03 15:51 ` Badari Pulavarty
2005-02-03 17:00 ` Sonny Rao
2005-02-03 16:56 ` Badari Pulavarty
2005-02-03 17:24 ` Sonny Rao
2005-02-03 20:50 ` Sonny Rao
2005-02-08 1:33 ` Andreas Dilger
2005-02-08 5:38 ` Sonny Rao
2005-02-09 21:11 ` Sonny Rao
2005-02-09 22:29 ` Badari Pulavarty
2005-02-10 2:05 ` Bryan Henderson
2005-02-10 2:45 ` Sonny Rao
2005-02-10 17:51 ` Bryan Henderson
2005-02-10 19:02 ` Sonny Rao
2005-02-10 16:02 ` Badari Pulavarty
2005-02-10 18:00 ` Bryan Henderson
2005-02-10 18:32 ` Badari Pulavarty [this message]
2005-02-10 20:30 ` Bryan Henderson
2005-02-10 20:25 ` Sonny Rao
2005-02-11 0:20 ` Bryan Henderson
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=1108060325.20053.1145.camel@dyn318077bld.beaverton.ibm.com \
--to=pbadari@us.ibm.com \
--cc=adilger@clusterfs.com \
--cc=hbryan@us.ibm.com \
--cc=linux-fsdevel@vger.kernel.org \
--cc=sonny@burdell.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).