From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1752091AbaEZTJ0 (ORCPT ); Mon, 26 May 2014 15:09:26 -0400 Received: from mx1.redhat.com ([209.132.183.28]:63788 "EHLO mx1.redhat.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1751507AbaEZTJZ (ORCPT ); Mon, 26 May 2014 15:09:25 -0400 Date: Mon, 26 May 2014 15:53:44 -0300 From: Marcelo Tosatti To: linux-mm@kvack.org, linux-kernel@vger.kernel.org Cc: Lai Jiangshan , Mel Gorman , Tejun Heo , Christoph Lameter , David Rientjes , Andrew Morton , Li Zefan Subject: [PATCH] page_alloc: skip cpuset enforcement for lower zone allocations (v2) Message-ID: <20140526185344.GA19976@amt.cnet> References: <20140523193706.GA22854@amt.cnet> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20140523193706.GA22854@amt.cnet> User-Agent: Mutt/1.5.21 (2010-09-15) Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Zone specific allocations, such as GFP_DMA32, should not be restricted to cpusets allowed node list: the zones which such allocations demand might be contained in particular nodes outside the cpuset node list. The alternative would be to not perform such allocations from applications which are cpuset restricted, which is unrealistic. Fixes KVM's alloc_page(gfp_mask=GFP_DMA32) with cpuset as explained. Signed-off-by: Marcelo Tosatti v2: fix slowpath as well (David Rientjes) diff --git a/kernel/cpuset.c b/kernel/cpuset.c index 3d54c41..b70a336 100644 --- a/kernel/cpuset.c +++ b/kernel/cpuset.c @@ -2392,6 +2392,10 @@ int __cpuset_node_allowed_softwall(int node, gfp_t gfp_mask) if (in_interrupt() || (gfp_mask & __GFP_THISNODE)) return 1; +#ifdef CONFIG_NUMA + if (gfp_zone(gfp_mask) < policy_zone) + return 1; +#endif might_sleep_if(!(gfp_mask & __GFP_HARDWALL)); if (node_isset(node, current->mems_allowed)) return 1; diff --git a/mm/page_alloc.c b/mm/page_alloc.c index 5dba293..dfea3dc 100644 --- a/mm/page_alloc.c +++ b/mm/page_alloc.c @@ -2698,6 +2698,7 @@ __alloc_pages_nodemask(gfp_t gfp_mask, unsigned int order, unsigned int cpuset_mems_cookie; int alloc_flags = ALLOC_WMARK_LOW|ALLOC_CPUSET|ALLOC_FAIR; struct mem_cgroup *memcg = NULL; + nodemask_t *cpuset_mems_allowed = &cpuset_current_mems_allowed; gfp_mask &= gfp_allowed_mask; @@ -2726,9 +2727,14 @@ __alloc_pages_nodemask(gfp_t gfp_mask, unsigned int order, retry_cpuset: cpuset_mems_cookie = read_mems_allowed_begin(); +#ifdef CONFIG_NUMA + if (gfp_zone(gfp_mask) < policy_zone) + cpuset_mems_allowed = NULL; +#endif + /* The preferred zone is used for statistics later */ first_zones_zonelist(zonelist, high_zoneidx, - nodemask ? : &cpuset_current_mems_allowed, + nodemask ? : cpuset_mems_allowed, &preferred_zone); if (!preferred_zone) goto out;