From mboxrd@z Thu Jan 1 00:00:00 1970 Received: from smtp-out2.suse.de (smtp-out2.suse.de [195.135.223.131]) (using TLSv1.2 with cipher ECDHE-RSA-AES128-GCM-SHA256 (128/128 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id E0A9C363C4C for ; Sun, 26 Apr 2026 12:58:01 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=195.135.223.131 ARC-Seal:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1777208283; cv=none; b=mrL0aio25hbUEdqVogDfXFDbJes8mfFRcYawdlVOkTTvTCW9hhK9jv+vPCinc7a0AsnU50rlViOggdrqkzY6HwR+eZuX6MCHPk11OjRh6TyZvJpadRzfyHaKLGQlLpK60W1SfQqNvkTCpbn4IzHRp6kIcAwEtWNyJuNXmCxVf/M= ARC-Message-Signature:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1777208283; c=relaxed/simple; bh=u+G4G8+46MJGdJqtWB7PvKTkUbu7xjBk6O6YB079jbs=; h=From:To:Cc:Subject:Date:Message-ID:In-Reply-To:References: MIME-Version; b=XhSn3bkVcuBT3UZV12jxa7M8W0M2QP6Y/cZ96lXoI0O2WFdMn1slB+HMNlSmx00TgYZOsDsoeX5aH7UJ7nOj3cDV16zmCdBR7ZIyOaZdZq/0R9NHYzklnB4A/a8qFJBLVNAgaLMdedHveiXVcvUO39XtoZU/dhQ+yVCY4yRRFc4= ARC-Authentication-Results:i=1; smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=suse.de; spf=pass smtp.mailfrom=suse.de; arc=none smtp.client-ip=195.135.223.131 Authentication-Results: smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=suse.de Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=suse.de Received: from imap1.dmz-prg2.suse.org (imap1.dmz-prg2.suse.org [IPv6:2a07:de40:b281:104:10:150:64:97]) (using TLSv1.3 with cipher TLS_AES_256_GCM_SHA384 (256/256 bits) key-exchange X25519 server-signature RSA-PSS (4096 bits) server-digest SHA256) (No client certificate requested) by smtp-out2.suse.de (Postfix) with ESMTPS id 751B35BD0A; Sun, 26 Apr 2026 12:57:42 +0000 (UTC) Authentication-Results: smtp-out2.suse.de; none Received: from imap1.dmz-prg2.suse.org (localhost [127.0.0.1]) (using TLSv1.3 with cipher TLS_AES_256_GCM_SHA384 (256/256 bits) key-exchange X25519 server-signature RSA-PSS (4096 bits) server-digest SHA256) (No client certificate requested) by imap1.dmz-prg2.suse.org (Postfix) with ESMTPS id 14739593B1; Sun, 26 Apr 2026 12:57:42 +0000 (UTC) Received: from dovecot-director2.suse.de ([2a07:de40:b281:106:10:150:64:167]) by imap1.dmz-prg2.suse.org with ESMTPSA id kGzYAsYL7mmtRAAAD6G6ig (envelope-from ); Sun, 26 Apr 2026 12:57:42 +0000 From: Oscar Salvador To: Andrew Morton Cc: David Hildenbrand , Michal Hocko , Muchun Song , Vlastimil Babka , Lorenzo Stoakes , linux-kernel@vger.kernel.org, linux-mm@kvack.org, Oscar Salvador Subject: [RFC PATCH v2 6/7] mm: Make /proc/pid/numa_maps use the new generic pagewalk API Date: Sun, 26 Apr 2026 14:57:18 +0200 Message-ID: <20260426125719.24698-7-osalvador@suse.de> X-Mailer: git-send-email 2.49.0 In-Reply-To: <20260426125719.24698-1-osalvador@suse.de> References: <20260426125719.24698-1-osalvador@suse.de> Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 Content-Transfer-Encoding: 8bit X-Rspamd-Pre-Result: action=no action; module=replies; Message is reply to one we originated X-Rspamd-Queue-Id: 751B35BD0A X-Rspamd-Pre-Result: action=no action; module=replies; Message is reply to one we originated X-Rspamd-Action: no action X-Spam-Score: -4.00 X-Spam-Level: X-Spam-Flag: NO X-Spamd-Result: default: False [-4.00 / 50.00]; REPLY(-4.00)[] X-Rspamd-Server: rspamd1.dmz-prg2.suse.org Have /proc/pid/numa_maps make use of the new generic API, and remove the code which was using the old one Signed-off-by: Oscar Salvador --- fs/proc/task_mmu.c | 159 +++++++++------------------------------------ 1 file changed, 32 insertions(+), 127 deletions(-) diff --git a/fs/proc/task_mmu.c b/fs/proc/task_mmu.c index 382c6b02d0e1..5c8a4b5250a1 100644 --- a/fs/proc/task_mmu.c +++ b/fs/proc/task_mmu.c @@ -3061,131 +3061,6 @@ static void gather_stats(struct page *page, struct numa_maps *md, int pte_dirty, md->node[folio_nid(folio)] += nr_pages; } -static struct page *can_gather_numa_stats(pte_t pte, struct vm_area_struct *vma, - unsigned long addr) -{ - struct page *page; - int nid; - - if (!pte_present(pte)) - return NULL; - - page = vm_normal_page(vma, addr, pte); - if (!page || is_zone_device_page(page)) - return NULL; - - if (PageReserved(page)) - return NULL; - - nid = page_to_nid(page); - if (!node_isset(nid, node_states[N_MEMORY])) - return NULL; - - return page; -} - -#ifdef CONFIG_TRANSPARENT_HUGEPAGE -static struct page *can_gather_numa_stats_pmd(pmd_t pmd, - struct vm_area_struct *vma, - unsigned long addr) -{ - struct page *page; - int nid; - - if (!pmd_present(pmd)) - return NULL; - - page = vm_normal_page_pmd(vma, addr, pmd); - if (!page) - return NULL; - - if (PageReserved(page)) - return NULL; - - nid = page_to_nid(page); - if (!node_isset(nid, node_states[N_MEMORY])) - return NULL; - - return page; -} -#endif - -static int gather_pte_stats(pmd_t *pmd, unsigned long addr, - unsigned long end, struct mm_walk *walk) -{ - struct numa_maps *md = walk->private; - struct vm_area_struct *vma = walk->vma; - spinlock_t *ptl; - pte_t *orig_pte; - pte_t *pte; - -#ifdef CONFIG_TRANSPARENT_HUGEPAGE - ptl = pmd_trans_huge_lock(pmd, vma); - if (ptl) { - struct page *page; - - page = can_gather_numa_stats_pmd(*pmd, vma, addr); - if (page) - gather_stats(page, md, pmd_dirty(*pmd), - HPAGE_PMD_SIZE/PAGE_SIZE); - spin_unlock(ptl); - return 0; - } -#endif - orig_pte = pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); - if (!pte) { - walk->action = ACTION_AGAIN; - return 0; - } - do { - pte_t ptent = ptep_get(pte); - struct page *page = can_gather_numa_stats(ptent, vma, addr); - if (!page) - continue; - gather_stats(page, md, pte_dirty(ptent), 1); - - } while (pte++, addr += PAGE_SIZE, addr != end); - pte_unmap_unlock(orig_pte, ptl); - cond_resched(); - return 0; -} -#ifdef CONFIG_HUGETLB_PAGE -static int gather_hugetlb_stats(pte_t *pte, unsigned long hmask, - unsigned long addr, unsigned long end, struct mm_walk *walk) -{ - pte_t huge_pte; - struct numa_maps *md; - struct page *page; - spinlock_t *ptl; - - ptl = huge_pte_lock(hstate_vma(walk->vma), walk->mm, pte); - huge_pte = huge_ptep_get(walk->mm, addr, pte); - if (!pte_present(huge_pte)) - goto out; - - page = pte_page(huge_pte); - - md = walk->private; - gather_stats(page, md, pte_dirty(huge_pte), 1); -out: - spin_unlock(ptl); - return 0; -} - -#else -static int gather_hugetlb_stats(pte_t *pte, unsigned long hmask, - unsigned long addr, unsigned long end, struct mm_walk *walk) -{ - return 0; -} -#endif - -static const struct mm_walk_ops show_numa_ops = { - .hugetlb_entry = gather_hugetlb_stats, - .pmd_entry = gather_pte_stats, - .walk_lock = PGWALK_RDLOCK, -}; - /* * Display pages allocated per node and memory policy via /proc. */ @@ -3197,9 +3072,15 @@ static int show_numa_map(struct seq_file *m, void *v) struct numa_maps *md = &numa_priv->md; struct file *file = vma->vm_file; struct mm_struct *mm = vma->vm_mm; + struct pt_range_walk ptw = { + .mm = mm + }; + enum pt_range_walk_type type; + pt_type_flags_t flags; char buffer[64]; struct mempolicy *pol; pgoff_t ilx; + int nr_pages; int nid; if (!mm) @@ -3230,8 +3111,32 @@ static int show_numa_map(struct seq_file *m, void *v) if (is_vm_hugetlb_page(vma)) seq_puts(m, " huge"); - /* mmap_lock is held by m_start */ - walk_page_vma(vma, &show_numa_ops, md); + flags = PT_TYPE_FOLIO; + type = pt_range_walk_start(&ptw, vma, vma->vm_start, vma->vm_end, flags); + while (type != PTW_DONE) { + + if (!ptw.folio || !ptw.page || PageReserved(ptw.page)) + goto not_found; + + nid = page_to_nid(ptw.page); + if (!node_isset(nid, node_states[N_MEMORY])) + goto not_found; + + if (is_vm_hugetlb_page(vma)) + /* + * As opposed to THP, HugeTLB counts the entire huge + * page as one unit size. + */ + nr_pages = ptw.nr_entries; + else + nr_pages = ptw.size / PAGE_SIZE; + + gather_stats(ptw.page, md, ptw.dirty, nr_pages); +not_found: + type = pt_range_walk_next(&ptw, vma, vma->vm_start, vma->vm_end, flags); + + } + pt_range_walk_done(&ptw); if (!md->pages) goto out; -- 2.35.3