linux-mm.kvack.org archive mirror
 help / color / mirror / Atom feed
From: Mel Gorman <mel@csn.ul.ie>
To: Christoph Lameter <clameter@sgi.com>
Cc: Pekka Enberg <penberg@cs.helsinki.fi>,
	Matt Mackall <mpm@selenic.com>,
	linux-mm@kvack.org
Subject: Re: [patch 0/8] slub: Fallback to order 0 and variable order slab support
Date: Fri, 7 Mar 2008 12:17:48 +0000	[thread overview]
Message-ID: <20080307121748.GF26229@csn.ul.ie> (raw)
In-Reply-To: <Pine.LNX.4.64.0803061409150.15083@schroedinger.engr.sgi.com>

On (06/03/08 14:18), Christoph Lameter didst pronounce:
> On Thu, 6 Mar 2008, Mel Gorman wrote:
> 
> > For huge page allocation success rates, the high order never helper the
> > situation but it was nowhere near as severe as it was for the slub-defrag
> > patches (ironically enough). Only one machine showed significantly worse
> 
> Well the slub-defrag tree is not really in shape for testing at this 
> point and I was working on it the last week. So not sure what tree was 
> picked up and thus not sure what to deduce from it. It may be too 
> aggressive in defragmentation attempts.
> 

That sounds fair, I didn't make any attempt to figure out what was going
on. But minimally, what I tested didn't blow up so that in itself is a
plus. We'll pick it up again later.

> > results. The rest were comparable for this set of tests at least but I would
> > still be wary of the long-lived behaviour of high-order slab allocations
> > slowly fragmenting memory due to pageblock fallbacks. Will think of how to
> > prove that in some way but just re-running the tests multiple times
> > without reboot may be enough.
> 
> Well maybe we could tune the page allocator a bit? There is the order 0 
> issue. We could also make all slab allocations use the same slab order in 
> order to reduce fragmentation problems.
>  

I don't think it would reduce them unless everyone was always using the
same order. Once slub is using a higher order than everywhere else, it
is possible it will use an alternative pageblock type just for the high
order.

The only tuning of the page allocator I can think of is to teach
rmqueue_bulk() to use the fewer high-order allocations to batch refill
the pcp queues. It's not very straight-forward though as when I tried
this a bit over a year ago, it cause fragmentation problems of its own.
I'll see about trying again.

> > Setting the order to 3 had vaguely similar results. The two outlier
> > machines had even worse negatives than order-4. With those machines
> > omitted the results were
> 
> Wonder what made them go worse.
> 

No idea.

> > Same story, hackbench-pipes and dbench suffer badly on some machines.
> > It's a similar story for order-1. With machine omitted it's
> > 
> > Kernbench Elapsed    time      -0.14      to 0.24
> > Kernbench Total  CPU           -0.13      to 0.11
> > Hackbench pipes-1              -11.90     to 5.39
> > Hackbench pipes-4              -7.01      to 2.06
> > Hackbench pipes-8              -5.49      to 1.66
> > Hackbench pipes-16             -6.08      to 2.72
> > Hackbench sockets-1            0.28       to 6.99
> > Hackbench sockets-4            0.63       to 5.50
> > Hackbench sockets-8            -10.95     to 7.70
> > Hackbench sockets-16           0.64       to 12.16
> > TBench    clients-1            -3.94      to 1.05
> > TBench    clients-2            -11.96     to 3.25
> > TBench    clients-4            -12.48     to -1.12
> > TBench    clients-8            -11.82     to -8.56
> > DBench    clients-1-ext2       -12.20     to 2.27
> > DBench    clients-2-ext2       -4.23      to 0.57
> > DBench    clients-4-ext2       -2.31      to 3.96
> > DBench    clients-8-ext2       -3.65      to 6.09
> 
> Well in that case there is something going on very strange performance
> wise. The results should be equal to upstream since the same orders 
> are used.

Really, order-1 is used by default by SLUB upstream? I missed that and
it doesn't appear to be the case on 2.6.25-rc2-mm1 at least according to
slabinfo. If it was the difference between order-0 and order-1, it may be
explained by the pcp allocator being bypassed.

> The only change in the hotpaths is another lookup which cannot 
> really account for the variances we see here. An 12% improvement because 
> logic was added to the hotpath?

Presuming you are referring to hackbench sockets-16, it could be because
the same objects were being reused again and the cache-hotness offset
the additional logic? Dunno, it's all handwaving. Unfortunately I don't
have what is needed in place to gather profiles automatically. It's on
the ever larger todo list :(

> There should be a significant regression 
> tbench (2%-4%) because the 4k slab cache must cause trouble.
> 
> > Based on this set of tests, it's clear that raising the order can be a big
> > win but setting it as default is less clear-cut.
> 
> There is something wrong here and we need to figure out what it is. The 
> order-1 test should fairly accurately reproduce upstream performance 
> characteristics.
> 

-- 
Mel Gorman
Part-time Phd Student                          Linux Technology Center
University of Limerick                         IBM Dublin Software Lab

--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org.  For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>

  reply	other threads:[~2008-03-07 12:17 UTC|newest]

Thread overview: 26+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
     [not found] <20080229044803.482012397@sgi.com>
     [not found] ` <20080229044820.044485187@sgi.com>
2008-02-29  8:13   ` [patch 7/8] slub: Make the order configurable for each slab cache Pekka Enberg
2008-02-29 19:37     ` Christoph Lameter
2008-03-01  9:47       ` Pekka Enberg
2008-03-03 17:49         ` Christoph Lameter
2008-03-03 22:56           ` Pekka Enberg
2008-03-03 23:36             ` Christoph Lameter
     [not found] ` <20080229044820.298792748@sgi.com>
2008-02-29  8:13   ` [patch 8/8] slub: Simplify any_slab_object checks Pekka Enberg
     [not found] ` <20080229044819.800974712@sgi.com>
2008-02-29  8:19   ` [patch 6/8] slub: Adjust order boundaries and minimum objects per slab Pekka Enberg
2008-02-29 19:41     ` Christoph Lameter
2008-03-01  9:58       ` Pekka J Enberg
2008-03-03 17:52         ` Christoph Lameter
2008-03-03 21:34           ` Matt Mackall
2008-03-03 22:36             ` Christoph Lameter
     [not found] ` <20080229044818.999367120@sgi.com>
2008-02-29  8:59   ` [patch 3/8] slub: Update statistics handling for variable order slabs Pekka Enberg
2008-02-29 19:43     ` Christoph Lameter
2008-03-01 10:29   ` Pekka Enberg
2008-03-04 12:20 ` [patch 0/8] slub: Fallback to order 0 and variable order slab support Mel Gorman
2008-03-04 18:53   ` Christoph Lameter
2008-03-05 18:28     ` Mel Gorman
2008-03-05 18:52       ` Christoph Lameter
2008-03-06 22:04         ` Mel Gorman
2008-03-06 22:18           ` Christoph Lameter
2008-03-07 12:17             ` Mel Gorman [this message]
2008-03-07 19:50               ` Christoph Lameter
2008-03-04 19:01   ` Matt Mackall
2008-03-05  0:04     ` Christoph Lameter

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20080307121748.GF26229@csn.ul.ie \
    --to=mel@csn.ul.ie \
    --cc=clameter@sgi.com \
    --cc=linux-mm@kvack.org \
    --cc=mpm@selenic.com \
    --cc=penberg@cs.helsinki.fi \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).