From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mail-wg0-f44.google.com (mail-wg0-f44.google.com [74.125.82.44]) by kanga.kvack.org (Postfix) with ESMTP id EBF456B0036 for ; Fri, 18 Jul 2014 04:15:45 -0400 (EDT) Received: by mail-wg0-f44.google.com with SMTP id m15so3103540wgh.3 for ; Fri, 18 Jul 2014 01:15:45 -0700 (PDT) Received: from szxga02-in.huawei.com (szxga02-in.huawei.com. [119.145.14.65]) by mx.google.com with ESMTPS id da1si1959833wib.71.2014.07.18.01.15.43 for (version=TLSv1 cipher=RC4-SHA bits=128/128); Fri, 18 Jul 2014 01:15:44 -0700 (PDT) From: Wang Nan Subject: [PATCH 1/5] memory-hotplug: x86_64: suitable memory should go to ZONE_MOVABLE Date: Fri, 18 Jul 2014 15:55:59 +0800 Message-ID: <1405670163-53747-2-git-send-email-wangnan0@huawei.com> In-Reply-To: <1405670163-53747-1-git-send-email-wangnan0@huawei.com> References: <1405670163-53747-1-git-send-email-wangnan0@huawei.com> MIME-Version: 1.0 Content-Type: text/plain Sender: owner-linux-mm@kvack.org List-ID: To: Ingo Molnar , Yinghai Lu , Mel Gorman , Andrew Morton Cc: Pei Feiyue , linux-mm@kvack.org, x86@kernel.org, linux-ia64@vger.kernel.org, linuxppc-dev@lists.ozlabs.org, linux-sh@vger.kernel.org, linux-kernel@vger.kernel.org This patch add new memory to ZONE_MOVABLE if movable zone is setup and lower than newly added memory for x86_64. Signed-off-by: Wang Nan --- arch/x86/mm/init_64.c | 10 ++++++++-- 1 file changed, 8 insertions(+), 2 deletions(-) diff --git a/arch/x86/mm/init_64.c b/arch/x86/mm/init_64.c index df1a992..825915e 100644 --- a/arch/x86/mm/init_64.c +++ b/arch/x86/mm/init_64.c @@ -685,17 +685,23 @@ static void update_end_of_memory_vars(u64 start, u64 size) } /* - * Memory is added always to NORMAL zone. This means you will never get - * additional DMA/DMA32 memory. + * Memory is added always to NORMAL or MOVABLE zone. This means you + * will never get additional DMA/DMA32 memory. */ int arch_add_memory(int nid, u64 start, u64 size) { struct pglist_data *pgdat = NODE_DATA(nid); struct zone *zone = pgdat->node_zones + ZONE_NORMAL; + struct zone *movable_zone = pgdat->node_zones + ZONE_MOVABLE; unsigned long start_pfn = start >> PAGE_SHIFT; unsigned long nr_pages = size >> PAGE_SHIFT; int ret; + if (!zone_is_empty(movable_zone)) + if (zone_spans_pfn(movable_zone, start_pfn) || + (zone_end_pfn(movable_zone) <= start_pfn)) + zone = movable_zone; + init_memory_mapping(start, start + size); ret = __add_pages(nid, zone, start_pfn, nr_pages); -- 1.8.4 -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: email@kvack.org