From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mail-pa0-f46.google.com (mail-pa0-f46.google.com [209.85.220.46]) by kanga.kvack.org (Postfix) with ESMTP id 20C036B026A for ; Wed, 30 Mar 2016 03:11:01 -0400 (EDT) Received: by mail-pa0-f46.google.com with SMTP id tt10so33224957pab.3 for ; Wed, 30 Mar 2016 00:11:01 -0700 (PDT) Received: from lgeamrelo12.lge.com (LGEAMRELO12.lge.com. [156.147.23.52]) by mx.google.com with ESMTP id xt3si4361777pab.145.2016.03.30.00.10.40 for ; Wed, 30 Mar 2016 00:10:41 -0700 (PDT) From: Minchan Kim Subject: [PATCH v3 11/16] zsmalloc: separate free_zspage from putback_zspage Date: Wed, 30 Mar 2016 16:12:10 +0900 Message-Id: <1459321935-3655-12-git-send-email-minchan@kernel.org> In-Reply-To: <1459321935-3655-1-git-send-email-minchan@kernel.org> References: <1459321935-3655-1-git-send-email-minchan@kernel.org> Sender: owner-linux-mm@kvack.org List-ID: To: Andrew Morton Cc: linux-kernel@vger.kernel.org, linux-mm@kvack.org, jlayton@poochiereds.net, bfields@fieldses.org, Vlastimil Babka , Joonsoo Kim , koct9i@gmail.com, aquini@redhat.com, virtualization@lists.linux-foundation.org, Mel Gorman , Hugh Dickins , Sergey Senozhatsky , Rik van Riel , rknize@motorola.com, Gioh Kim , Sangseok Lee , Chan Gyun Jeong , Al Viro , YiPing Xu , Minchan Kim Currently, putback_zspage does free zspage under class->lock if fullness become ZS_EMPTY but it makes trouble to implement locking scheme for new zspage migration. So, this patch is to separate free_zspage from putback_zspage and free zspage out of class->lock which is preparation for zspage migration. Signed-off-by: Minchan Kim --- mm/zsmalloc.c | 46 +++++++++++++++++++++++----------------------- 1 file changed, 23 insertions(+), 23 deletions(-) diff --git a/mm/zsmalloc.c b/mm/zsmalloc.c index 14bcc741eead..b11dcd718502 100644 --- a/mm/zsmalloc.c +++ b/mm/zsmalloc.c @@ -949,7 +949,8 @@ static void reset_page(struct page *page) page->freelist = NULL; } -static void free_zspage(struct page *first_page) +static void free_zspage(struct zs_pool *pool, struct size_class *class, + struct page *first_page) { struct page *nextp, *tmp, *head_extra; @@ -972,6 +973,11 @@ static void free_zspage(struct page *first_page) } reset_page(head_extra); __free_page(head_extra); + + zs_stat_dec(class, OBJ_ALLOCATED, get_maxobj_per_zspage( + class->size, class->pages_per_zspage)); + atomic_long_sub(class->pages_per_zspage, + &pool->pages_allocated); } /* Initialize a newly allocated zspage */ @@ -1559,13 +1565,8 @@ void zs_free(struct zs_pool *pool, unsigned long handle) spin_lock(&class->lock); obj_free(class, obj); fullness = fix_fullness_group(class, first_page); - if (fullness == ZS_EMPTY) { - zs_stat_dec(class, OBJ_ALLOCATED, get_maxobj_per_zspage( - class->size, class->pages_per_zspage)); - atomic_long_sub(class->pages_per_zspage, - &pool->pages_allocated); - free_zspage(first_page); - } + if (fullness == ZS_EMPTY) + free_zspage(pool, class, first_page); spin_unlock(&class->lock); unpin_tag(handle); @@ -1752,7 +1753,7 @@ static struct page *isolate_target_page(struct size_class *class) * @class: destination class * @first_page: target page * - * Return @fist_page's fullness_group + * Return @first_page's updated fullness_group */ static enum fullness_group putback_zspage(struct zs_pool *pool, struct size_class *class, @@ -1764,15 +1765,6 @@ static enum fullness_group putback_zspage(struct zs_pool *pool, insert_zspage(class, fullness, first_page); set_zspage_mapping(first_page, class->index, fullness); - if (fullness == ZS_EMPTY) { - zs_stat_dec(class, OBJ_ALLOCATED, get_maxobj_per_zspage( - class->size, class->pages_per_zspage)); - atomic_long_sub(class->pages_per_zspage, - &pool->pages_allocated); - - free_zspage(first_page); - } - return fullness; } @@ -1835,23 +1827,31 @@ static void __zs_compact(struct zs_pool *pool, struct size_class *class) if (!migrate_zspage(pool, class, &cc)) break; - putback_zspage(pool, class, dst_page); + VM_BUG_ON_PAGE(putback_zspage(pool, class, + dst_page) == ZS_EMPTY, dst_page); } /* Stop if we couldn't find slot */ if (dst_page == NULL) break; - putback_zspage(pool, class, dst_page); - if (putback_zspage(pool, class, src_page) == ZS_EMPTY) + VM_BUG_ON_PAGE(putback_zspage(pool, class, + dst_page) == ZS_EMPTY, dst_page); + if (putback_zspage(pool, class, src_page) == ZS_EMPTY) { pool->stats.pages_compacted += class->pages_per_zspage; - spin_unlock(&class->lock); + spin_unlock(&class->lock); + free_zspage(pool, class, src_page); + } else { + spin_unlock(&class->lock); + } + cond_resched(); spin_lock(&class->lock); } if (src_page) - putback_zspage(pool, class, src_page); + VM_BUG_ON_PAGE(putback_zspage(pool, class, + src_page) == ZS_EMPTY, src_page); spin_unlock(&class->lock); } -- 1.9.1 -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: email@kvack.org