public inbox for linux-kernel@vger.kernel.org
 help / color / mirror / Atom feed
From: Fengguang Wu <fengguang.wu@gmail.com>
To: Helge Hafting <helgehaf@aitel.hist.no>
Cc: Helge Hafting <helge.hafting@aitel.hist.no>,
	Linux Kernel Mailing List <linux-kernel@vger.kernel.org>
Subject: Re: New readahead - ups and downs new test
Date: Mon, 3 Jul 2006 23:39:30 +0800	[thread overview]
Message-ID: <351941126.25373@ustc.edu.cn> (raw)
Message-ID: <20060703153930.GC5874@mail.ustc.edu.cn> (raw)
In-Reply-To: <20060703135027.GA4440@aitel.hist.no>

On Mon, Jul 03, 2006 at 03:50:27PM +0200, Helge Hafting wrote:
> On Mon, Jul 03, 2006 at 07:55:16AM +0800, Fengguang Wu wrote:
> > Hi Helge,
> > 
> > On Tue, Jun 27, 2006 at 03:07:16PM +0200, Helge Hafting wrote:
> > > I made my own little io-intensive test, that shows a case where
> > > performance drops.
> > > 
> > > I boot the machine, and starts "debsums", a debian utility that
> > > checksums every file managed by debian package management.
> > > As soon as the machine starts swapping, I also start
> > > start a process that applies an mm-patch to the kernel tree, and
> > > times this.
> > > 
> > > This patching took 1m28s with cold cache, without debsums running.
> > > With the 2.6.15 kernel (old readahead), and debsums running, this
> > > took 2m20s to complete, and 360kB in swap at the worst.
> > > 
> > > With the new readahead in 2.6.17-mm3 I get 6m22s for patching,
> > > and 22MB in swap at the most.  Runs with mm1 and mm2 were
> > > similiar, 5-6 minutes patching and 22MB swap.
> > > 
> > > My patching clearly takes more times this way.  I don't know
> > > if debsums improved though, it could be as simple as a fairness
> > > issue.  Memory pressure definitely went up.
> > 
> > There are a lot changes between 2.6.15 and 2.6.17-mmX. Would you use
> > the single 2.6.17-mm5 kernel for benchmarking? It's easy:
> > 
> >         - select old readahead:
> >                 echo 1 > /proc/sys/vm/readahead_ratio
> > 
> >         - select new readahead:
> >                 echo 50 > /proc/sys/vm/readahead_ratio
> > 
> >
> I just tried this with 2.5.17-mm5.  I did in on a faster
> machine (opteron cpu, but still 512MB) so don't compare with
> my previous test which ran on a pentium-IV.
> Single cpu in both cases.
> 
> Test procdure:
> 1. Reboot, log in through xdm
> 2. run vmstat 10 for swap monitoring
> 3. time debsums -s
> 4. As soon as the machine touches swap, launch
>    time bzcat 2.6.15-mm5.bz2 | patch -p1
> 
> In either case, testing starts with 320MB free memory after boot,
> which debsums caching eats in about a minute and swapping starts.
> Then I start the patching, which finished before debsums.
> 
> Old readahed:
> Max swap was 700kB, but it dropped back to 244kB after 10s
> and stayed there.  
> Patch timing:
> real    0m37.662s
> user    0m5.002s
> sys     0m2.023s
> debsums timing:
> real    5m50.333s
> user    0m21.127s
> sys     0m14.506s
> 
> New readahead:
> Max swap: 244kB.  (On another try it jumped to 816kB and then fell back
> to 244kB).
> patch timing:
> real    0m40.951s
> user    0m5.043s
> sys     0m2.061s
> debsums timing:
> real    5m46.555s
> user    0m21.195s
> sys     0m13.918s
> 
> Timing and memory load seems to be almost identical this time,
> perhaps this is a load where the type of readahead doesn't
> matter.  

Thanks. You are right, the readahead logic won't affect the swap cache.
Nor will the readahead size, I guess. But to be sure, you can do one
more test on it with the following command, using the same 2.5.17-mm5:

        blockdev --setra /dev/hda1 256

Please replace /dev/hda1 with the root device on your system, thanks.

Wu

  reply	other threads:[~2006-07-03 15:38 UTC|newest]

Thread overview: 8+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2006-06-27 13:07 New readahead - ups and downs Helge Hafting
     [not found] ` <20060627160624.GB6014@mail.ustc.edu.cn>
2006-06-27 16:06   ` Fengguang Wu
     [not found] ` <20060702235516.GA6034@mail.ustc.edu.cn>
2006-07-02 23:55   ` Fengguang Wu
2006-07-03 13:50   ` New readahead - ups and downs new test Helge Hafting
     [not found]     ` <20060703153930.GC5874@mail.ustc.edu.cn>
2006-07-03 15:39       ` Fengguang Wu [this message]
2006-07-03 20:36       ` Helge Hafting
2006-07-03 21:42       ` New readahead - ups and downs new test. Vm oddities Helge Hafting
     [not found]         ` <20060704012621.GA7236@mail.ustc.edu.cn>
2006-07-04  1:26           ` Fengguang Wu

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=351941126.25373@ustc.edu.cn \
    --to=fengguang.wu@gmail.com \
    --cc=helge.hafting@aitel.hist.no \
    --cc=helgehaf@aitel.hist.no \
    --cc=linux-kernel@vger.kernel.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox