From: Mel Gorman <mel@csn.ul.ie>
To: Christoph Lameter <clameter@sgi.com>
Cc: Pekka Enberg <penberg@cs.helsinki.fi>,
Matt Mackall <mpm@selenic.com>,
linux-mm@kvack.org
Subject: Re: [patch 0/8] slub: Fallback to order 0 and variable order slab support
Date: Fri, 7 Mar 2008 12:17:48 +0000 [thread overview]
Message-ID: <20080307121748.GF26229@csn.ul.ie> (raw)
In-Reply-To: <Pine.LNX.4.64.0803061409150.15083@schroedinger.engr.sgi.com>
On (06/03/08 14:18), Christoph Lameter didst pronounce:
> On Thu, 6 Mar 2008, Mel Gorman wrote:
>
> > For huge page allocation success rates, the high order never helper the
> > situation but it was nowhere near as severe as it was for the slub-defrag
> > patches (ironically enough). Only one machine showed significantly worse
>
> Well the slub-defrag tree is not really in shape for testing at this
> point and I was working on it the last week. So not sure what tree was
> picked up and thus not sure what to deduce from it. It may be too
> aggressive in defragmentation attempts.
>
That sounds fair, I didn't make any attempt to figure out what was going
on. But minimally, what I tested didn't blow up so that in itself is a
plus. We'll pick it up again later.
> > results. The rest were comparable for this set of tests at least but I would
> > still be wary of the long-lived behaviour of high-order slab allocations
> > slowly fragmenting memory due to pageblock fallbacks. Will think of how to
> > prove that in some way but just re-running the tests multiple times
> > without reboot may be enough.
>
> Well maybe we could tune the page allocator a bit? There is the order 0
> issue. We could also make all slab allocations use the same slab order in
> order to reduce fragmentation problems.
>
I don't think it would reduce them unless everyone was always using the
same order. Once slub is using a higher order than everywhere else, it
is possible it will use an alternative pageblock type just for the high
order.
The only tuning of the page allocator I can think of is to teach
rmqueue_bulk() to use the fewer high-order allocations to batch refill
the pcp queues. It's not very straight-forward though as when I tried
this a bit over a year ago, it cause fragmentation problems of its own.
I'll see about trying again.
> > Setting the order to 3 had vaguely similar results. The two outlier
> > machines had even worse negatives than order-4. With those machines
> > omitted the results were
>
> Wonder what made them go worse.
>
No idea.
> > Same story, hackbench-pipes and dbench suffer badly on some machines.
> > It's a similar story for order-1. With machine omitted it's
> >
> > Kernbench Elapsed time -0.14 to 0.24
> > Kernbench Total CPU -0.13 to 0.11
> > Hackbench pipes-1 -11.90 to 5.39
> > Hackbench pipes-4 -7.01 to 2.06
> > Hackbench pipes-8 -5.49 to 1.66
> > Hackbench pipes-16 -6.08 to 2.72
> > Hackbench sockets-1 0.28 to 6.99
> > Hackbench sockets-4 0.63 to 5.50
> > Hackbench sockets-8 -10.95 to 7.70
> > Hackbench sockets-16 0.64 to 12.16
> > TBench clients-1 -3.94 to 1.05
> > TBench clients-2 -11.96 to 3.25
> > TBench clients-4 -12.48 to -1.12
> > TBench clients-8 -11.82 to -8.56
> > DBench clients-1-ext2 -12.20 to 2.27
> > DBench clients-2-ext2 -4.23 to 0.57
> > DBench clients-4-ext2 -2.31 to 3.96
> > DBench clients-8-ext2 -3.65 to 6.09
>
> Well in that case there is something going on very strange performance
> wise. The results should be equal to upstream since the same orders
> are used.
Really, order-1 is used by default by SLUB upstream? I missed that and
it doesn't appear to be the case on 2.6.25-rc2-mm1 at least according to
slabinfo. If it was the difference between order-0 and order-1, it may be
explained by the pcp allocator being bypassed.
> The only change in the hotpaths is another lookup which cannot
> really account for the variances we see here. An 12% improvement because
> logic was added to the hotpath?
Presuming you are referring to hackbench sockets-16, it could be because
the same objects were being reused again and the cache-hotness offset
the additional logic? Dunno, it's all handwaving. Unfortunately I don't
have what is needed in place to gather profiles automatically. It's on
the ever larger todo list :(
> There should be a significant regression
> tbench (2%-4%) because the 4k slab cache must cause trouble.
>
> > Based on this set of tests, it's clear that raising the order can be a big
> > win but setting it as default is less clear-cut.
>
> There is something wrong here and we need to figure out what it is. The
> order-1 test should fairly accurately reproduce upstream performance
> characteristics.
>
--
Mel Gorman
Part-time Phd Student Linux Technology Center
University of Limerick IBM Dublin Software Lab
--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org. For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>
next prev parent reply other threads:[~2008-03-07 12:17 UTC|newest]
Thread overview: 26+ messages / expand[flat|nested] mbox.gz Atom feed top
[not found] <20080229044803.482012397@sgi.com>
[not found] ` <20080229044820.044485187@sgi.com>
2008-02-29 8:13 ` [patch 7/8] slub: Make the order configurable for each slab cache Pekka Enberg
2008-02-29 19:37 ` Christoph Lameter
2008-03-01 9:47 ` Pekka Enberg
2008-03-03 17:49 ` Christoph Lameter
2008-03-03 22:56 ` Pekka Enberg
2008-03-03 23:36 ` Christoph Lameter
[not found] ` <20080229044820.298792748@sgi.com>
2008-02-29 8:13 ` [patch 8/8] slub: Simplify any_slab_object checks Pekka Enberg
[not found] ` <20080229044819.800974712@sgi.com>
2008-02-29 8:19 ` [patch 6/8] slub: Adjust order boundaries and minimum objects per slab Pekka Enberg
2008-02-29 19:41 ` Christoph Lameter
2008-03-01 9:58 ` Pekka J Enberg
2008-03-03 17:52 ` Christoph Lameter
2008-03-03 21:34 ` Matt Mackall
2008-03-03 22:36 ` Christoph Lameter
[not found] ` <20080229044818.999367120@sgi.com>
2008-02-29 8:59 ` [patch 3/8] slub: Update statistics handling for variable order slabs Pekka Enberg
2008-02-29 19:43 ` Christoph Lameter
2008-03-01 10:29 ` Pekka Enberg
2008-03-04 12:20 ` [patch 0/8] slub: Fallback to order 0 and variable order slab support Mel Gorman
2008-03-04 18:53 ` Christoph Lameter
2008-03-05 18:28 ` Mel Gorman
2008-03-05 18:52 ` Christoph Lameter
2008-03-06 22:04 ` Mel Gorman
2008-03-06 22:18 ` Christoph Lameter
2008-03-07 12:17 ` Mel Gorman [this message]
2008-03-07 19:50 ` Christoph Lameter
2008-03-04 19:01 ` Matt Mackall
2008-03-05 0:04 ` Christoph Lameter
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20080307121748.GF26229@csn.ul.ie \
--to=mel@csn.ul.ie \
--cc=clameter@sgi.com \
--cc=linux-mm@kvack.org \
--cc=mpm@selenic.com \
--cc=penberg@cs.helsinki.fi \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).