linux-mm.kvack.org archive mirror
 help / color / mirror / Atom feed
From: Christoph Lameter <cl@linux.com>
To: Pekka Enberg <penberg@cs.helsinki.fi>
Cc: David Rientjes <rientjes@google.com>,
	Hugh Dickins <hughd@google.com>,
	Eric Dumazet <eric.dumazet@gmail.com>,
	"H. Peter Anvin" <hpa@zytor.com>,
	Mathieu Desnoyers <mathieu.desnoyers@efficios.com>,
	linux-mm@kvack.org
Subject: [slubllv333num@/21] slub: Per object NUMA support
Date: Fri, 15 Apr 2011 15:12:52 -0500	[thread overview]
Message-ID: <20110415201258.442821107@linux.com> (raw)
In-Reply-To: 20110415201246.096634892@linux.com

[-- Attachment #1: rr_slabs --]
[-- Type: text/plain, Size: 1897 bytes --]

Currently slub applies NUMA policies per allocated slab page. Change
that to apply memory policies for each individual object allocated.

F.e. before this patch MPOL_INTERLEAVE would return objects from the
same slab page until a new slab page was allocated. Now an object
from a different page is taken for each allocation.

This increases the overhead of the fastpath under NUMA.

Signed-off-by: Christoph Lameter <cl@linux.com>

---
 mm/slub.c |   16 ++++++++++++++++
 1 file changed, 16 insertions(+)

Index: linux-2.6/mm/slub.c
===================================================================
--- linux-2.6.orig/mm/slub.c	2011-04-15 12:54:42.000000000 -0500
+++ linux-2.6/mm/slub.c	2011-04-15 13:11:25.000000000 -0500
@@ -1887,6 +1887,21 @@ debug:
 	goto unlock_out;
 }
 
+static __always_inline int alternate_slab_node(struct kmem_cache *s,
+						gfp_t flags, int node)
+{
+#ifdef CONFIG_NUMA
+	if (unlikely(node == NUMA_NO_NODE &&
+			!(flags & __GFP_THISNODE) &&
+			!in_interrupt())) {
+		if ((s->flags & SLAB_MEM_SPREAD) && cpuset_do_slab_mem_spread())
+			node = cpuset_slab_spread_node();
+		else if (current->mempolicy)
+		node = slab_node(current->mempolicy);
+	}
+#endif
+	return node;
+}
 /*
  * Inlined fastpath so that allocation functions (kmalloc, kmem_cache_alloc)
  * have the fastpath folded into their functions. So no function call
@@ -1911,6 +1926,7 @@ static __always_inline void *slab_alloc(
 	if (slab_pre_alloc_hook(s, gfpflags))
 		return NULL;
 
+	node = alternate_slab_node(s, gfpflags, node);
 #ifndef CONFIG_CMPXCHG_LOCAL
 	local_irq_save(flags);
 #else

--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org.  For more info on Linux MM,
see: http://www.linux-mm.org/ .
Fight unfair telecom internet charges in Canada: sign http://stopthemeter.ca/
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>

  parent reply	other threads:[~2011-04-15 20:13 UTC|newest]

Thread overview: 22+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2011-04-15 20:12 [slubllv333num@/21] SLUB: Lockless freelists for objects V3 Christoph Lameter
2011-04-15 20:12 ` [slubllv333num@/21] slub: Use NUMA_NO_NODE in get_partial Christoph Lameter
2011-04-15 20:12 ` [slubllv333num@/21] slub: get_map() function to establish map of free objects in a slab Christoph Lameter
2011-04-15 20:12 ` [slubllv333num@/21] slub: Eliminate repeated use of c->page through a new page variable Christoph Lameter
2011-04-15 20:12 ` [slubllv333num@/21] slub: Move node determination out of hotpath Christoph Lameter
2011-04-15 20:12 ` [slubllv333num@/21] slub: Move debug handlign in __slab_free Christoph Lameter
2011-04-15 20:12 ` Christoph Lameter [this message]
2011-04-15 20:12 ` [slubllv333num@/21] slub: Do not use frozen page flag but a bit in the page counters Christoph Lameter
2011-04-15 20:12 ` [slubllv333num@/21] slub: Move page->frozen handling near where the page->freelist handling occurs Christoph Lameter
2011-04-15 20:12 ` [slubllv333num@/21] x86: Add support for cmpxchg_double Christoph Lameter
2011-04-15 20:12 ` [slubllv333num@/21] mm: Rearrange struct page Christoph Lameter
2011-04-15 20:12 ` [slubllv333num@/21] slub: Add cmpxchg_double_slab() Christoph Lameter
2011-04-15 20:12 ` [slubllv333num@/21] slub: explicit list_lock taking Christoph Lameter
2011-04-15 20:12 ` [slubllv333num@/21] slub: Pass kmem_cache struct to lock and freeze slab Christoph Lameter
2011-04-15 20:13 ` [slubllv333num@/21] slub: Rework allocator fastpaths Christoph Lameter
2011-04-15 20:13 ` [slubllv333num@/21] slub: Invert locking and avoid slab lock Christoph Lameter
2011-04-15 20:13 ` [slubllv333num@/21] slub: Disable interrupts in free_debug processing Christoph Lameter
2011-04-15 20:13 ` [slubllv333num@/21] slub: Avoid disabling interrupts in free slowpath Christoph Lameter
2011-04-15 20:13 ` [slubllv333num@/21] slub: Get rid of the another_slab label Christoph Lameter
2011-04-15 20:13 ` [slubllv333num@/21] slub: fast release on full slab Christoph Lameter
2011-04-15 20:13 ` [slubllv333num@/21] slub: Not necessary to check for empty slab on load_freelist Christoph Lameter
2011-04-15 20:13 ` [slubllv333num@/21] slub: update statistics for cmpxchg handling Christoph Lameter

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20110415201258.442821107@linux.com \
    --to=cl@linux.com \
    --cc=eric.dumazet@gmail.com \
    --cc=hpa@zytor.com \
    --cc=hughd@google.com \
    --cc=linux-mm@kvack.org \
    --cc=mathieu.desnoyers@efficios.com \
    --cc=penberg@cs.helsinki.fi \
    --cc=rientjes@google.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).