From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from psmtp.com (na3sys010amx116.postini.com [74.125.245.116]) by kanga.kvack.org (Postfix) with SMTP id E29EF6B0068 for ; Mon, 30 Jul 2012 09:11:22 -0400 (EDT) Message-ID: <501687EA.5050401@parallels.com> Date: Mon, 30 Jul 2012 17:11:06 +0400 From: Glauber Costa MIME-Version: 1.0 Subject: Re: [PATCH 06/10] sl[au]b: Allocate objects from memcg cache References: <1343227101-14217-1-git-send-email-glommer@parallels.com> <1343227101-14217-7-git-send-email-glommer@parallels.com> <20120730125851.GB27293@shutemov.name> In-Reply-To: <20120730125851.GB27293@shutemov.name> Content-Type: text/plain; charset="ISO-8859-1" Content-Transfer-Encoding: 7bit Sender: owner-linux-mm@kvack.org List-ID: To: "Kirill A. Shutemov" Cc: linux-kernel@vger.kernel.org, linux-mm@kvack.org, Andrew Morton , Christoph Lameter , David Rientjes , Pekka Enberg , Greg Thelen , Johannes Weiner , Michal Hocko , Frederic Weisbecker , devel@openvz.org, cgroups@vger.kernel.org, Pekka Enberg , Kamezawa Hiroyuki , Suleiman Souhlal On 07/30/2012 04:58 PM, Kirill A. Shutemov wrote: > On Wed, Jul 25, 2012 at 06:38:17PM +0400, Glauber Costa wrote: >> We are able to match a cache allocation to a particular memcg. If the >> task doesn't change groups during the allocation itself - a rare event, >> this will give us a good picture about who is the first group to touch a >> cache page. >> >> This patch uses the now available infrastructure by calling >> memcg_kmem_get_cache() before all the cache allocations. >> >> Signed-off-by: Glauber Costa >> CC: Christoph Lameter >> CC: Pekka Enberg >> CC: Michal Hocko >> CC: Kamezawa Hiroyuki >> CC: Johannes Weiner >> CC: Suleiman Souhlal >> --- >> include/linux/slub_def.h | 18 +++++++++++++----- >> mm/memcontrol.c | 2 ++ >> mm/slab.c | 4 ++++ >> mm/slub.c | 1 + >> 4 files changed, 20 insertions(+), 5 deletions(-) >> >> diff --git a/include/linux/slub_def.h b/include/linux/slub_def.h >> index 8bb8ad2..148000a 100644 >> --- a/include/linux/slub_def.h >> +++ b/include/linux/slub_def.h >> @@ -13,6 +13,8 @@ >> #include >> >> #include >> +#include >> +#include >> >> enum stat_item { >> ALLOC_FASTPATH, /* Allocation from cpu slab */ >> @@ -209,14 +211,14 @@ static __always_inline int kmalloc_index(size_t size) >> * This ought to end up with a global pointer to the right cache >> * in kmalloc_caches. >> */ >> -static __always_inline struct kmem_cache *kmalloc_slab(size_t size) >> +static __always_inline struct kmem_cache *kmalloc_slab(gfp_t flags, size_t size) >> { >> int index = kmalloc_index(size); >> >> if (index == 0) >> return NULL; >> >> - return kmalloc_caches[index]; >> + return memcg_kmem_get_cache(kmalloc_caches[index], flags); >> } >> >> void *kmem_cache_alloc(struct kmem_cache *, gfp_t); >> @@ -225,7 +227,13 @@ void *__kmalloc(size_t size, gfp_t flags); >> static __always_inline void * >> kmalloc_order(size_t size, gfp_t flags, unsigned int order) >> { >> - void *ret = (void *) __get_free_pages(flags | __GFP_COMP, order); >> + void *ret; >> + >> + flags = __GFP_COMP; >> +#ifdef CONFIG_MEMCG_KMEM >> + flags |= __GFP_KMEMCG; >> +#endif > > Em.. I don't see where __GFP_KMEMCG is defined. > It should be 0 for !CONFIG_MEMCG_KMEM. > It is not, sorry. As I said, this is dependent on another patch series. My main goal while sending this was to get the slab part - that will eventually come ontop of that - discussed. Because they are both quite complex, I believe they benefit from being discussed separately. You can find the latest version of that here: https://lkml.org/lkml/2012/6/25/251 -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: email@kvack.org