From mboxrd@z Thu Jan 1 00:00:00 1970 Received: from smtp.kernel.org (aws-us-west-2-korg-mail-1.web.codeaurora.org [10.30.226.201]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 568D92BD0B for ; Sun, 1 Mar 2026 01:22:31 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=10.30.226.201 ARC-Seal:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1772328151; cv=none; b=CusQnhcJzeIFlLR02wC+NXOdhZwj0OfIAvxJeBJDY5Bk/b7bveffzD91XXC7PCKi3SqO4TnGC0gS4qllWqqPfnUfnBp5s6dbv1odwA3kwtkG09xYiraiqKVX1x1i97FxOxPPpHW0rCTDqVZvQVTPs4oRgUzMCZ/u5F5pCYloMu8= ARC-Message-Signature:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1772328151; c=relaxed/simple; bh=mjakhQFwOfu3c6K0n9rHkBJ6KvJ7nFTHNV6JA2AQPiM=; h=From:To:Cc:Subject:Date:Message-ID:MIME-Version; b=MurwQWngOlsKDxANY2WauG2t921ljsG/w7CizFCWfI/moGOYsVBZb0yJbPxTJgdJdwfh16wIgIsH938shLzZHm+YYsD5Z6DaMMdNeHSeyOVLuDGJNuf4+mmRPoCWqWyfmXkIb8X/rxCbZdG1VcNtl1q9FZCc9VSpcAd4GSrf874= ARC-Authentication-Results:i=1; smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=kernel.org header.i=@kernel.org header.b=n4aaAXBX; arc=none smtp.client-ip=10.30.226.201 Authentication-Results: smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=kernel.org header.i=@kernel.org header.b="n4aaAXBX" Received: by smtp.kernel.org (Postfix) with ESMTPSA id 9ADDAC19424; Sun, 1 Mar 2026 01:22:29 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=k20201202; t=1772328151; bh=mjakhQFwOfu3c6K0n9rHkBJ6KvJ7nFTHNV6JA2AQPiM=; h=From:To:Cc:Subject:Date:From; b=n4aaAXBXqBchi18REblnAOMYR03FMx896G3ChFEJMhQD8eTf0rzQcPxHNdsdnIau6 2qKpbb7HcX9jeLQi3SUmt2c7YVD4rym59gKEhVljeYwdV+pL4rQ6B3o/qBPiqCKMJ9 P30zZb9gYaisRj2xlkuCAhqhXjCm1Zz8/Xxsqnopf1XWGJb3Sm0jUejN38zERQ3Ob4 YGz6yJtMwV2HmtQebIM9CEsOHA/I2CgImiCXTiFOhBbczjIvu/T/gV4voviMayAJiP Oh5/yieCwQRJVcX35BQwm3QUphkpb/5lUzWMP0hH/F6NFKz33GztX8qsiRTg/EQLjD 690ydPRC8cUiA== From: Sasha Levin To: stable@vger.kernel.org, vbabka@kernel.org Cc: Vlastimil Babka , Michal Hocko , Johannes Weiner , Pedro Falcato , Zi Yan , Brendan Jackman , "David Hildenbrand (Red Hat)" , David Rientjes , Joshua Hahn , Liam Howlett , Lorenzo Stoakes , Mike Rapoport , Suren Baghdasaryan , Andrew Morton , linux-mm@kvack.org Subject: FAILED: Patch "mm, page_alloc, thp: prevent reclaim for __GFP_THISNODE THP allocations" failed to apply to 6.12-stable tree Date: Sat, 28 Feb 2026 20:22:28 -0500 Message-ID: <20260301012228.1678891-1-sashal@kernel.org> X-Mailer: git-send-email 2.51.0 Precedence: bulk X-Mailing-List: stable@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 X-Patchwork-Hint: ignore X-stable: review Content-Transfer-Encoding: 8bit The patch below does not apply to the 6.12-stable tree. If someone wants it applied there, or to any other stable or longterm tree, then please email the backport, including the original git commit id to . Thanks, Sasha ------------------ original commit in Linus's tree ------------------ >From 9c9828d3ead69416d731b1238802af31760c823e Mon Sep 17 00:00:00 2001 From: Vlastimil Babka Date: Fri, 19 Dec 2025 17:31:57 +0100 Subject: [PATCH] mm, page_alloc, thp: prevent reclaim for __GFP_THISNODE THP allocations Since commit cc638f329ef6 ("mm, thp: tweak reclaim/compaction effort of local-only and all-node allocations"), THP page fault allocations have settled on the following scheme (from the commit log): 1. local node only THP allocation with no reclaim, just compaction. 2. for madvised VMA's or when synchronous compaction is enabled always - THP allocation from any node with effort determined by global defrag setting and VMA madvise 3. fallback to base pages on any node Recent customer reports however revealed we have a gap in step 1 above. What we have seen is excessive reclaim due to THP page faults on a NUMA node that's close to its high watermark, while other nodes have plenty of free memory. The problem with step 1 is that it promises no reclaim after the compaction attempt, however reclaim is only avoided for certain compaction outcomes (deferred, or skipped due to insufficient free base pages), and not e.g. when compaction is actually performed but fails (we did see compact_fail vmstat counter increasing). THP page faults can therefore exhibit a zone_reclaim_mode-like behavior, which is not the intention. Thus add a check for __GFP_THISNODE that corresponds to this exact situation and prevents continuing with reclaim/compaction once the initial compaction attempt isn't successful in allocating the page. Note that commit cc638f329ef6 has not introduced this over-reclaim possibility; it appears to exist in some form since commit 2f0799a0ffc0 ("mm, thp: restore node-local hugepage allocations"). Followup commits b39d0ee2632d ("mm, page_alloc: avoid expensive reclaim when compaction may not succeed") and cc638f329ef6 have moved in the right direction, but left the abovementioned gap. Link: https://lkml.kernel.org/r/20251219-costly-noretry-thisnode-fix-v1-1-e1085a4a0c34@suse.cz Fixes: 2f0799a0ffc0 ("mm, thp: restore node-local hugepage allocations") Signed-off-by: Vlastimil Babka Acked-by: Michal Hocko Acked-by: Johannes Weiner Acked-by: Pedro Falcato Acked-by: Zi Yan Cc: Brendan Jackman Cc: "David Hildenbrand (Red Hat)" Cc: David Rientjes Cc: Joshua Hahn Cc: Liam Howlett Cc: Lorenzo Stoakes Cc: Mike Rapoport Cc: Suren Baghdasaryan Cc: Signed-off-by: Andrew Morton --- mm/page_alloc.c | 14 ++++++++++++++ 1 file changed, 14 insertions(+) diff --git a/mm/page_alloc.c b/mm/page_alloc.c index e1cc0c9ed9479..3333524e879c4 100644 --- a/mm/page_alloc.c +++ b/mm/page_alloc.c @@ -4818,6 +4818,20 @@ __alloc_pages_slowpath(gfp_t gfp_mask, unsigned int order, compact_result == COMPACT_DEFERRED) goto nopage; + /* + * THP page faults may attempt local node only first, + * but are then allowed to only compact, not reclaim, + * see alloc_pages_mpol(). + * + * Compaction can fail for other reasons than those + * checked above and we don't want such THP allocations + * to put reclaim pressure on a single node in a + * situation where other nodes might have plenty of + * available memory. + */ + if (gfp_mask & __GFP_THISNODE) + goto nopage; + /* * Looks like reclaim/compaction is worth trying, but * sync compaction could be very expensive, so keep -- 2.51.0