* reaim - 2.6.3-mm1 IO performance down.
@ 2004-02-25 0:20 cliff white
2004-02-25 1:03 ` Andrew Morton
0 siblings, 1 reply; 8+ messages in thread
From: cliff white @ 2004-02-25 0:20 UTC (permalink / raw)
To: linux-kernel; +Cc: akpm
Running the reaim 'new_fserver' workload, we now see a performance drop
on 2.6.3-mm1, ext3 filesystem
Kernel JPM Max Users Percent change
inux-2.6.3 10347.87 172 0.0
2.6.3-rc3-mm1 9826.35 164 -5.07
2.6.3-mm1 8938.17 140 (run 1) -13.65
2.6.3-mm1 9100.39 136 (run 2)-12.08
I have done some comparions graphs here, with readprofile data:
http://developer.osdl.org/cliffw/reaim/compares/r_comp/2.6.3_vs_mm1_r1/index.html
http://developer.osdl.org/cliffw/reaim/compares/r_comp/2.6.3_vs_mm1_r2/index.html
The interesting graph is the bottom one on the page, comparing client
run-time to number of children. The -mm1 kernel has a spike in run time, becoming
very noticealbe around 60-80 lusers.
kernel Users JPM Run Time
linux-2.6.3 60 10327.87 34.16 seconds
2.6.3-mm1 60 8279.75 42.61 seconds
Linux-2.6.3 80 10275.23 45.78 seconds
2.6.3-mm1 80 7841.31 59.99 seconds
For the same test on the same machine, results from 2.6.2-rc1-mm2 and 2.6.2-rc3-mm1
were within 1.0% of the linux-2.6.2 runs. So this is new.
More data and tests if requested - are there some patch sets we should try reverting?
cliffw
--
The church is near, but the road is icy.
The bar is far, but i will walk carefully. - Russian proverb
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: reaim - 2.6.3-mm1 IO performance down.
2004-02-25 0:20 cliff white
@ 2004-02-25 1:03 ` Andrew Morton
2004-02-25 17:40 ` Cliff White
2004-02-25 23:36 ` Cliff White
0 siblings, 2 replies; 8+ messages in thread
From: Andrew Morton @ 2004-02-25 1:03 UTC (permalink / raw)
To: cliff white; +Cc: linux-kernel
cliff white <cliffw@osdl.org> wrote:
>
> For the same test on the same machine, results from 2.6.2-rc1-mm2 and 2.6.2-rc3-mm1
> were within 1.0% of the linux-2.6.2 runs. So this is new.
>
> More data and tests if requested - are there some patch sets we should try reverting?
Thanks. You could try reverting adaptive-lazy-readahead.patch. If it is
not that I'd be suspecting CPU scheduler changes. Do you have uniprocessor
test results?
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: reaim - 2.6.3-mm1 IO performance down.
@ 2004-02-25 2:00 Con Kolivas
2004-02-25 3:31 ` Nick Piggin
0 siblings, 1 reply; 8+ messages in thread
From: Con Kolivas @ 2004-02-25 2:00 UTC (permalink / raw)
To: linux-kernel; +Cc: Cliff White, Andrew Morton
>Running the reaim 'new_fserver' workload, we now see a performance drop on
>2.6.3-mm1, ext3 filesystem
I observed a serious slowdown on non numa, non smt machines with kernbench and
the scheduler changes (posted results a week ago here:
http://marc.theaimsgroup.com/?l=linux-kernel&m=107719112225482&w=2 )
A summary of those results is half job load (-j4 on 8x):
2.6.3: Elapsed Time 231.274
2.6.3-mm1: Elapsed Time 273.688
The drop in reaim performance is possibly related.
Con
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: reaim - 2.6.3-mm1 IO performance down.
2004-02-25 2:00 reaim - 2.6.3-mm1 IO performance down Con Kolivas
@ 2004-02-25 3:31 ` Nick Piggin
0 siblings, 0 replies; 8+ messages in thread
From: Nick Piggin @ 2004-02-25 3:31 UTC (permalink / raw)
To: Con Kolivas; +Cc: linux-kernel, Cliff White, Andrew Morton
Con Kolivas wrote:
>
>>Running the reaim 'new_fserver' workload, we now see a performance drop on
>>2.6.3-mm1, ext3 filesystem
>>
>
>I observed a serious slowdown on non numa, non smt machines with kernbench and
>the scheduler changes (posted results a week ago here:
>http://marc.theaimsgroup.com/?l=linux-kernel&m=107719112225482&w=2 )
>
>A summary of those results is half job load (-j4 on 8x):
>2.6.3: Elapsed Time 231.274
>2.6.3-mm1: Elapsed Time 273.688
>
>The drop in reaim performance is possibly related.
>
>
I have been meaning to look at that. The STP wasn't working
for me for a couple of days, but Cliff fixed that so I'll get
on it.
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: reaim - 2.6.3-mm1 IO performance down.
2004-02-25 1:03 ` Andrew Morton
@ 2004-02-25 17:40 ` Cliff White
2004-02-25 19:16 ` Andrew Morton
2004-02-25 23:36 ` Cliff White
1 sibling, 1 reply; 8+ messages in thread
From: Cliff White @ 2004-02-25 17:40 UTC (permalink / raw)
To: Andrew Morton; +Cc: linux-kernel
> cliff white <cliffw@osdl.org> wrote:
> >
> > For the same test on the same machine, results from 2.6.2-rc1-mm2 and 2.6.2
> -rc3-mm1
> > were within 1.0% of the linux-2.6.2 runs. So this is new.
> >
> > More data and tests if requested - are there some patch sets we should try
> reverting?
>
> Thanks. You could try reverting adaptive-lazy-readahead.patch. If it is
> not that I'd be suspecting CPU scheduler changes. Do you have uniprocessor
> test results?
I have them for 2.6.3-mm3, am re-running 2.6.3-mm1 right now.
Gross results are within 1%, but looking at the detail, i do see badness,
example:
Kernel Users Run time
2.6.3 20 32.11
2.6.3-mm3 20 35.47
2.6.3 40 63.64
2.6.3-mm2 40 66.33
Again, this shows up best on the bottom graph on the page.
Graphn of 2.6.3 vs 2.6.3-mm3 :
http://developer.osdl.org/cliffw/reaim/compares/r_comp/2.6.3_vs_mm1_1cpu/index.html
cliffw
>
>
> -
> To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at http://vger.kernel.org/majordomo-info.html
> Please read the FAQ at http://www.tux.org/lkml/
>
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: reaim - 2.6.3-mm1 IO performance down.
2004-02-25 17:40 ` Cliff White
@ 2004-02-25 19:16 ` Andrew Morton
2004-02-25 20:52 ` Cliff White
0 siblings, 1 reply; 8+ messages in thread
From: Andrew Morton @ 2004-02-25 19:16 UTC (permalink / raw)
To: Cliff White; +Cc: linux-kernel
Cliff White <cliffw@osdl.org> wrote:
>
> > Thanks. You could try reverting adaptive-lazy-readahead.patch. If it is
> > not that I'd be suspecting CPU scheduler changes. Do you have uniprocessor
> > test results?
>
> I have them for 2.6.3-mm3, am re-running 2.6.3-mm1 right now.
> Gross results are within 1%, but looking at the detail, i do see badness,
OK, I'll do the bsearch. Again. Could you please tell me what would be an
appropriate reaim command line with which to reproduce this?
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: reaim - 2.6.3-mm1 IO performance down.
2004-02-25 19:16 ` Andrew Morton
@ 2004-02-25 20:52 ` Cliff White
0 siblings, 0 replies; 8+ messages in thread
From: Cliff White @ 2004-02-25 20:52 UTC (permalink / raw)
To: Andrew Morton; +Cc: linux-kernel
> Cliff White <cliffw@osdl.org> wrote:
> >
> > > Thanks. You could try reverting adaptive-lazy-readahead.patch. If it i
> s
> > > not that I'd be suspecting CPU scheduler changes. Do you have uniproces
> sor
> > > test results?
> >
> > I have them for 2.6.3-mm3, am re-running 2.6.3-mm1 right now.
> > Gross results are within 1%, but looking at the detail, i do see badness,
>
> OK, I'll do the bsearch. Again. Could you please tell me what would be an
> appropriate reaim command line with which to reproduce this?
Sure.
We use the 'workfile.new_fserver' for this load. Comes in the kit.
This test does IO, so you'll need to list some directories and a filepool
size
in a text file. Ours looks like this:
stp.config
-------------------
FILESIZE 80k
POOLSIZE 1024k
DISKDIR /mnt/disk1
DISKDIR /mnt/disk2
DISKDIR /mnt/disk3
DISKDIR /mnt/disk4
--------------------
If you want the shortest possible run, shrink POOLSIZE and
go with the defaults:
reaim -fworkfile.new_fserver -l./stp.config
That will run 1->5 users, should be < 1 egg timer in duration
on modern iron. I dunno if that'll be a useful data point.
To replicate at a specific number of users, pick a number
( example: 20 )
, and use this:
reaim -s20 -e20 -fworkfile.new_fserver -l./stp.config
That will do one pass at 20 users. Should be quick, maybe 2-4
egg timers, and will give you one of the data points I reported.
If you have more time, this runs range 20->80 with a 10 user increment:
reaim -s20 -e80 -i10 -t -fworkfile.new_fserver -l./stp.config
( the -t turns off the AIM7 adaptive increment )
The STP runs are very long. Many egg-timers.
We invoke them like this, where NCPU is the number of CPUS we have:
The 'quick convergence run' ( repeats three times )
reaim -s$NCPU -q -t -i$NCPU -fworkfile.new_fserver -r3 -b -l./stp.config
This is the 'maximum' longest run
reaim -s$NCPU -s -t -i$NCPU -fworkfile.new_fserver -r3 -b -l./stp.config
If you want those, quicker to request via STP/PLM -
cliffw
>
> -
> To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at http://vger.kernel.org/majordomo-info.html
> Please read the FAQ at http://www.tux.org/lkml/
>
^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: reaim - 2.6.3-mm1 IO performance down.
2004-02-25 1:03 ` Andrew Morton
2004-02-25 17:40 ` Cliff White
@ 2004-02-25 23:36 ` Cliff White
1 sibling, 0 replies; 8+ messages in thread
From: Cliff White @ 2004-02-25 23:36 UTC (permalink / raw)
To: Andrew Morton; +Cc: cliff white, linux-kernel, cliffw
> cliff white <cliffw@osdl.org> wrote:
> >
> > For the same test on the same machine, results from 2.6.2-rc1-mm2 and 2.6.2
> -rc3-mm1
> > were within 1.0% of the linux-2.6.2 runs. So this is new.
> >
> > More data and tests if requested - are there some patch sets we should try
> reverting?
>
> Thanks. You could try reverting adaptive-lazy-readahead.patch. If it is
> not that I'd be suspecting CPU scheduler changes. Do you have uniprocessor
> test results?
adaptive-lazy-readahead reverted, not really much change here:
Kernel Users JPM Run Time
2.6.3 60 10327.87 34.16 seconds
2.6.3-mm1 60 8279.75 42.61 seconds
2.6.3-mm1-noalr 60 7731.76 45.63 seconds
2.6.3 80 10275.23 45.78 seconds
2.6.3-mm1 80 7841.31 59.99 seconds
2.6.3-mm1-noalr 80 8565.19 54.92 seconds
Full details, reaim tarball:
http://developer.osdl.org/cliffw
cliffw
>
>
> -
> To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at http://vger.kernel.org/majordomo-info.html
> Please read the FAQ at http://www.tux.org/lkml/
>
^ permalink raw reply [flat|nested] 8+ messages in thread
end of thread, other threads:[~2004-02-25 23:42 UTC | newest]
Thread overview: 8+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2004-02-25 2:00 reaim - 2.6.3-mm1 IO performance down Con Kolivas
2004-02-25 3:31 ` Nick Piggin
-- strict thread matches above, loose matches on Subject: below --
2004-02-25 0:20 cliff white
2004-02-25 1:03 ` Andrew Morton
2004-02-25 17:40 ` Cliff White
2004-02-25 19:16 ` Andrew Morton
2004-02-25 20:52 ` Cliff White
2004-02-25 23:36 ` Cliff White
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox