From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mail-wr0-f197.google.com (mail-wr0-f197.google.com [209.85.128.197]) by kanga.kvack.org (Postfix) with ESMTP id 0CED56B0317 for ; Tue, 13 Jun 2017 05:00:55 -0400 (EDT) Received: by mail-wr0-f197.google.com with SMTP id n18so28281546wra.11 for ; Tue, 13 Jun 2017 02:00:55 -0700 (PDT) Received: from mail-wr0-f193.google.com (mail-wr0-f193.google.com. [209.85.128.193]) by mx.google.com with ESMTPS id x4si3319884wmb.93.2017.06.13.02.00.53 for (version=TLS1_2 cipher=ECDHE-RSA-AES128-GCM-SHA256 bits=128/128); Tue, 13 Jun 2017 02:00:53 -0700 (PDT) Received: by mail-wr0-f193.google.com with SMTP id g76so27856594wrd.2 for ; Tue, 13 Jun 2017 02:00:53 -0700 (PDT) From: Michal Hocko Subject: [RFC PATCH 4/4] mm, hugetlb, soft_offline: use new_page_nodemask for soft offline migration Date: Tue, 13 Jun 2017 11:00:39 +0200 Message-Id: <20170613090039.14393-5-mhocko@kernel.org> In-Reply-To: <20170613090039.14393-1-mhocko@kernel.org> References: <20170613090039.14393-1-mhocko@kernel.org> Sender: owner-linux-mm@kvack.org List-ID: To: linux-mm@kvack.org Cc: Naoya Horiguchi , Mike Kravetz , Mel Gorman , Vlastimil Babka , Andrew Morton , LKML , Michal Hocko From: Michal Hocko new_page is yet another duplication of the migration callback which has to handle hugetlb migration specially. We can safely use the generic new_page_nodemask for the same purpose. Please note that gigantic hugetlb pages do not need any special handling because alloc_huge_page_nodemask will make sure to check pages in all per node pools. The reason this was done previously was that alloc_huge_page_node treated NO_NUMA_NODE and a specific node differently and so alloc_huge_page_node(nid) would check on this specific node. Noticed-by: Vlastimil Babka Signed-off-by: Michal Hocko --- mm/memory-failure.c | 10 +--------- 1 file changed, 1 insertion(+), 9 deletions(-) diff --git a/mm/memory-failure.c b/mm/memory-failure.c index 3615bffbd269..7040f60ecb71 100644 --- a/mm/memory-failure.c +++ b/mm/memory-failure.c @@ -1487,16 +1487,8 @@ EXPORT_SYMBOL(unpoison_memory); static struct page *new_page(struct page *p, unsigned long private, int **x) { int nid = page_to_nid(p); - if (PageHuge(p)) { - struct hstate *hstate = page_hstate(compound_head(p)); - if (hstate_is_gigantic(hstate)) - return alloc_huge_page_node(hstate, NUMA_NO_NODE); - - return alloc_huge_page_node(hstate, nid); - } else { - return __alloc_pages_node(nid, GFP_HIGHUSER_MOVABLE, 0); - } + return new_page_nodemask(p, nid, &node_states[N_MEMORY]); } /* -- 2.11.0 -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: email@kvack.org