linux-kernel.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: Jeff Moyer <jmoyer@redhat.com>
To: Vivek Goyal <vgoyal@redhat.com>
Cc: Jens Axboe <axboe@kernel.dk>, linux-kernel@vger.kernel.org
Subject: Re: [PATCH 2/2] cfq: allow dispatching of both sync and async I/O together
Date: Tue, 22 Jun 2010 08:45:54 -0400	[thread overview]
Message-ID: <x49sk4ffd6l.fsf@segfault.boston.devel.redhat.com> (raw)
In-Reply-To: <20100622040700.GA12502@redhat.com> (Vivek Goyal's message of "Tue, 22 Jun 2010 00:07:00 -0400")

Vivek Goyal <vgoyal@redhat.com> writes:

> On Mon, Jun 21, 2010 at 07:22:08PM -0400, Vivek Goyal wrote:
>> On Mon, Jun 21, 2010 at 09:59:48PM +0200, Jens Axboe wrote:
>> > On 21/06/10 21.49, Jeff Moyer wrote:
>> > > Hi,
>> > > 
>> > > In testing a workload that has a single fsync-ing process and another
>> > > process that does a sequential buffered read, I was unable to tune CFQ
>> > > to reach the throughput of deadline.  This patch, along with the previous
>> > > one, brought CFQ in line with deadline when setting slice_idle to 0.
>> > > 
>> > > I'm not sure what the original reason for not allowing sync and async
>> > > I/O to be dispatched together was.  If there is a workload I should be
>> > > testing that shows the inherent problems of this, please point me at it
>> > > and I will resume testing.  Until and unless that workload is identified,
>> > > please consider applying this patch.
>> > 
>> > The problematic case is/was a normal SATA drive with a buffered
>> > writer and an occasional reader. I'll have to double check my
>> > mail tomorrow, but iirc the issue was that the occasional reader
>> > would suffer great latencies since service times for that single
>> > IO would be delayed at the drive side. It could perhaps just be
>> > a bug in how we handle the slice idling on the read side when the
>> > IO gets delayed initially.
>> > 
>> > So if my memory is correct, google for the fsync madness and
>> > interactiveness thread that we had some months ago and which
>> > caused a lot of tweaking. The commit adding this is
>> > 5ad531db6e0f3c3c985666e83d3c1c4d53acccf9 and was added back
>> > in July last year. So it was around that time that the mails went
>> > around.
>> 
>> Hi Jens,
>> 
>> I suspect we might have introduced this patch because mike galbraith
>> had issues which application interactiveness (reading data back from swap)
>> in the prence of heavy writeout on SATA disk.
>> 
>> After this patch we did two enhancements.
>> 
>> - You introduced the logic of building write queue depth gradually.
>> - Corrado introduced the logic of idling on the random reader service
>>   tree.
>> 
>> In the past random reader were not protected from WRITES as there was no
>> idling on random readers. But with corrado's changes of idling on
>> sync-noidle service tree, I think this problem might have been solved to
>> a great extent.
>> 
>> Getting rid of this exclusivity of either SYNC/ASYNC requests in request
>> queue might help us with throughput on storage arrys without loosing
>> protection for random reader on SATA. 
>> 
>> I will do some testing with and without patch and see if above is true
>> or not.
>
> Some primilinary testing results with and without patch. I started a
> buffered writer and started firefox and monitored how much time firefox
> took.
>
> dd if=/dev/zero of=zerofile bs=4K count=1024M
>
> 2.6.35-rc3 vanilla
> ==================
> real    0m22.546s
> user    0m0.566s
> sys     0m0.107s
>
>
> real    0m21.410s
> user    0m0.527s
> sys     0m0.095s
>
>
> real    0m27.594s
> user    0m1.256s
> sys     0m0.483s
>
> 2.6.35-rc3 + jeff's patches
> ===========================
> real    0m20.372s
> user    0m0.635s
> sys     0m0.128s
>
> real    0m22.281s
> user    0m0.509s
> sys     0m0.093s
>
> real    0m23.211s
> user    0m0.674s
> sys     0m0.140s
>
> So looks like firefox launching times have not changed much in the presence
> of heavy buffered writting going on root disk. I will do more testing tomorrow.

Was the buffered writer actually hitting disk?  How much memory is on
your system?

Cheers,
Jeff

  reply	other threads:[~2010-06-22 12:46 UTC|newest]

Thread overview: 19+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2010-06-21 19:49 [PATCH 0/2] cfq: fixes to bring cfq in line with deadline performance for mid- to high-end storage Jeff Moyer
2010-06-21 19:49 ` [PATCH 1/2] cfq: always return false from should_idle if slice_idle is set to zero Jeff Moyer
2010-06-21 20:00   ` Jens Axboe
2010-06-28 18:41     ` Jeff Moyer
2010-06-28 18:50       ` Jens Axboe
2010-06-28 18:54         ` Jeff Moyer
2010-06-21 23:05   ` Vivek Goyal
2010-06-21 19:49 ` [PATCH 2/2] cfq: allow dispatching of both sync and async I/O together Jeff Moyer
2010-06-21 19:59   ` Jens Axboe
2010-06-21 20:05     ` Jeff Moyer
2010-06-21 23:22     ` Vivek Goyal
2010-06-22  4:07       ` Vivek Goyal
2010-06-22 12:45         ` Jeff Moyer [this message]
2010-06-22 13:18           ` Vivek Goyal
2010-06-22 13:21             ` Jens Axboe
2010-06-22 14:24               ` Vivek Goyal
2010-06-22 14:27             ` Jeff Moyer
2010-06-28 18:40         ` Jeff Moyer
2010-06-28 18:48           ` Jens Axboe

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=x49sk4ffd6l.fsf@segfault.boston.devel.redhat.com \
    --to=jmoyer@redhat.com \
    --cc=axboe@kernel.dk \
    --cc=linux-kernel@vger.kernel.org \
    --cc=vgoyal@redhat.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).