From: Dan Williams <dan.j.williams@intel.com>
To: Alistair Popple <apopple@nvidia.com>, <linux-mm@kvack.org>
Cc: Alistair Popple <apopple@nvidia.com>,
<gerald.schaefer@linux.ibm.com>, <dan.j.williams@intel.com>,
<jgg@ziepe.ca>, <willy@infradead.org>, <david@redhat.com>,
<linux-kernel@vger.kernel.org>, <nvdimm@lists.linux.dev>,
<linux-fsdevel@vger.kernel.org>, <linux-ext4@vger.kernel.org>,
<linux-xfs@vger.kernel.org>, <jhubbard@nvidia.com>, <hch@lst.de>,
<zhang.lyra@gmail.com>, <debug@rivosinc.com>, <bjorn@kernel.org>,
<balbirs@nvidia.com>, <lorenzo.stoakes@oracle.com>,
<linux-arm-kernel@lists.infradead.org>,
<loongarch@lists.linux.dev>, <linuxppc-dev@lists.ozlabs.org>,
<linux-riscv@lists.infradead.org>, <linux-cxl@vger.kernel.org>,
<dri-devel@lists.freedesktop.org>, <John@groves.net>
Subject: Re: [PATCH 03/12] mm/pagewalk: Skip dax pages in pagewalk
Date: Wed, 4 Jun 2025 18:59:09 -0700 [thread overview]
Message-ID: <6840f9ed3785a_249110084@dwillia2-xfh.jf.intel.com.notmuch> (raw)
In-Reply-To: <1799c6772825e1401e7ccad81a10646118201953.1748500293.git-series.apopple@nvidia.com>
Alistair Popple wrote:
> Previously dax pages were skipped by the pagewalk code as pud_special() or
> vm_normal_page{_pmd}() would be false for DAX pages. Now that dax pages are
> refcounted normally that is no longer the case, so add explicit checks to
> skip them.
>
> Signed-off-by: Alistair Popple <apopple@nvidia.com>
> ---
> include/linux/memremap.h | 11 +++++++++++
> mm/pagewalk.c | 12 ++++++++++--
> 2 files changed, 21 insertions(+), 2 deletions(-)
>
> diff --git a/include/linux/memremap.h b/include/linux/memremap.h
> index 4aa1519..54e8b57 100644
> --- a/include/linux/memremap.h
> +++ b/include/linux/memremap.h
> @@ -198,6 +198,17 @@ static inline bool folio_is_fsdax(const struct folio *folio)
> return is_fsdax_page(&folio->page);
> }
>
> +static inline bool is_devdax_page(const struct page *page)
> +{
> + return is_zone_device_page(page) &&
> + page_pgmap(page)->type == MEMORY_DEVICE_GENERIC;
> +}
> +
> +static inline bool folio_is_devdax(const struct folio *folio)
> +{
> + return is_devdax_page(&folio->page);
> +}
> +
> #ifdef CONFIG_ZONE_DEVICE
> void zone_device_page_init(struct page *page);
> void *memremap_pages(struct dev_pagemap *pgmap, int nid);
> diff --git a/mm/pagewalk.c b/mm/pagewalk.c
> index e478777..0dfb9c2 100644
> --- a/mm/pagewalk.c
> +++ b/mm/pagewalk.c
> @@ -884,6 +884,12 @@ struct folio *folio_walk_start(struct folio_walk *fw,
> * support PUD mappings in VM_PFNMAP|VM_MIXEDMAP VMAs.
> */
> page = pud_page(pud);
> +
> + if (is_devdax_page(page)) {
> + spin_unlock(ptl);
> + goto not_found;
> + }
> +
> goto found;
> }
>
> @@ -911,7 +917,8 @@ struct folio *folio_walk_start(struct folio_walk *fw,
> goto pte_table;
> } else if (pmd_present(pmd)) {
> page = vm_normal_page_pmd(vma, addr, pmd);
> - if (page) {
> + if (page && !is_devdax_page(page) &&
> + !is_fsdax_page(page)) {
It just looks awkward to say "yup, normal page, but not *that*
'normal'".
What about something like the below? Either way you can add:
Reviewed-by: Dan Williams <dan.j.williams@intel.com>
diff --git a/include/linux/mm.h b/include/linux/mm.h
index 12d96659e8b4..4e549669166b 100644
--- a/include/linux/mm.h
+++ b/include/linux/mm.h
@@ -2471,6 +2471,27 @@ struct folio *vm_normal_folio_pmd(struct vm_area_struct *vma,
struct page *vm_normal_page_pmd(struct vm_area_struct *vma, unsigned long addr,
pmd_t pmd);
+/* return normal pages backed by the page allocator */
+static inline struct page *vm_normal_gfp_pmd(struct vm_area_struct *vma,
+ unsigned long addr, pmd_t pmd)
+{
+ struct page *page = vm_normal_page_pmd(vma, addr, pmd);
+
+ if (!is_devdax_page(page) && !is_fsdax_page(page))
+ return page;
+ return NULL;
+}
+
+static inline struct page *vm_normal_gfp_pte(struct vm_area_struct *vma,
+ unsigned long addr, pte_t pte)
+{
+ struct page *page = vm_normal_page(vma, addr, pte);
+
+ if (!is_devdax_page(page) && !is_fsdax_page(page))
+ return page;
+ return NULL;
+}
+
void zap_vma_ptes(struct vm_area_struct *vma, unsigned long address,
unsigned long size);
void zap_page_range_single(struct vm_area_struct *vma, unsigned long address,
diff --git a/mm/pagewalk.c b/mm/pagewalk.c
index cca170fe5be5..54bfece05323 100644
--- a/mm/pagewalk.c
+++ b/mm/pagewalk.c
@@ -914,9 +914,8 @@ struct folio *folio_walk_start(struct folio_walk *fw,
spin_unlock(ptl);
goto pte_table;
} else if (pmd_present(pmd)) {
- page = vm_normal_page_pmd(vma, addr, pmd);
- if (page && !is_devdax_page(page) &&
- !is_fsdax_page(page)) {
+ page = vm_normal_gfp_pmd(vma, addr, pmd);
+ if (page) {
goto found;
} else if ((flags & FW_ZEROPAGE) &&
is_huge_zero_pmd(pmd)) {
@@ -949,9 +948,8 @@ struct folio *folio_walk_start(struct folio_walk *fw,
fw->pte = pte;
if (pte_present(pte)) {
- page = vm_normal_page(vma, addr, pte);
- if (page && !is_devdax_page(page) &&
- !is_fsdax_page(page))
+ page = vm_normal_gfp_pte(vma, addr, pte);
+ if (page)
goto found;
if ((flags & FW_ZEROPAGE) &&
is_zero_pfn(pte_pfn(pte))) {
_______________________________________________
linux-riscv mailing list
linux-riscv@lists.infradead.org
http://lists.infradead.org/mailman/listinfo/linux-riscv
next prev parent reply other threads:[~2025-06-05 2:01 UTC|newest]
Thread overview: 59+ messages / expand[flat|nested] mbox.gz Atom feed top
2025-05-29 6:32 [PATCH 00/12] mm: Remove pXX_devmap page table bit and pfn_t type Alistair Popple
2025-05-29 6:32 ` [PATCH 01/12] mm: Remove PFN_MAP, PFN_SG_CHAIN and PFN_SG_LAST Alistair Popple
2025-05-29 11:46 ` Jonathan Cameron
2025-06-04 3:22 ` Alistair Popple
2025-05-30 9:33 ` David Hildenbrand
2025-06-02 4:54 ` Christoph Hellwig
2025-06-04 3:23 ` Alistair Popple
2025-06-03 13:34 ` Jason Gunthorpe
2025-06-04 21:05 ` Dan Williams
2025-05-29 6:32 ` [PATCH 02/12] mm: Convert pXd_devmap checks to vma_is_dax Alistair Popple
2025-05-30 9:37 ` David Hildenbrand
2025-06-12 6:55 ` Alistair Popple
2025-06-03 13:35 ` Jason Gunthorpe
2025-06-05 1:37 ` Dan Williams
2025-05-29 6:32 ` [PATCH 03/12] mm/pagewalk: Skip dax pages in pagewalk Alistair Popple
2025-05-30 9:42 ` David Hildenbrand
2025-06-03 13:36 ` Jason Gunthorpe
2025-06-05 1:59 ` Dan Williams [this message]
2025-06-05 7:46 ` Christoph Hellwig
2025-06-05 7:49 ` David Hildenbrand
2025-06-05 16:21 ` Dan Williams
2025-06-12 7:02 ` Alistair Popple
2025-06-12 8:47 ` Alistair Popple
2025-06-12 14:15 ` Lorenzo Stoakes
2025-06-12 22:50 ` Alistair Popple
2025-05-29 6:32 ` [PATCH 04/12] mm: Convert vmf_insert_mixed() from using pte_devmap to pte_special Alistair Popple
2025-06-03 13:37 ` Jason Gunthorpe
2025-06-05 2:02 ` Dan Williams
2025-05-29 6:32 ` [PATCH 05/12] mm: Remove remaining uses of PFN_DEV Alistair Popple
2025-06-03 13:38 ` Jason Gunthorpe
2025-06-05 2:02 ` Dan Williams
2025-05-29 6:32 ` [PATCH 06/12] mm/gup: Remove pXX_devmap usage from get_user_pages() Alistair Popple
2025-06-03 13:47 ` Jason Gunthorpe
2025-06-05 2:04 ` Dan Williams
2025-05-29 6:32 ` [PATCH 07/12] mm: Remove redundant pXd_devmap calls Alistair Popple
2025-05-29 11:54 ` Jonathan Cameron
2025-06-02 9:33 ` David Hildenbrand
2025-06-02 12:20 ` David Hildenbrand
2025-06-03 13:48 ` Jason Gunthorpe
2025-06-05 2:35 ` Dan Williams
2025-06-05 12:09 ` Jason Gunthorpe
2025-06-05 12:21 ` David Hildenbrand
2025-06-05 16:30 ` Dan Williams
2025-06-05 17:04 ` David Hildenbrand
2025-06-05 16:22 ` Dan Williams
2025-05-29 6:32 ` [PATCH 08/12] mm/khugepaged: Remove redundant pmd_devmap() check Alistair Popple
2025-06-02 11:45 ` David Hildenbrand
2025-06-03 13:48 ` Jason Gunthorpe
2025-05-29 6:32 ` [PATCH 09/12] powerpc: Remove checks for devmap pages and PMDs/PUDs Alistair Popple
2025-06-03 13:49 ` Jason Gunthorpe
2025-05-29 6:32 ` [PATCH 10/12] mm: Remove devmap related functions and page table bits Alistair Popple
2025-06-03 13:50 ` Jason Gunthorpe
2025-05-29 6:32 ` [PATCH 11/12] mm: Remove callers of pfn_t functionality Alistair Popple
2025-06-02 4:44 ` Michael Kelley
2025-06-03 13:50 ` Jason Gunthorpe
2025-05-29 6:32 ` [PATCH 12/12] mm/memremap: Remove unused devmap_managed_key Alistair Popple
2025-06-03 13:51 ` Jason Gunthorpe
2025-06-02 10:31 ` [PATCH 00/12] mm: Remove pXX_devmap page table bit and pfn_t type David Hildenbrand
2025-06-05 1:39 ` Dan Williams
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=6840f9ed3785a_249110084@dwillia2-xfh.jf.intel.com.notmuch \
--to=dan.j.williams@intel.com \
--cc=John@groves.net \
--cc=apopple@nvidia.com \
--cc=balbirs@nvidia.com \
--cc=bjorn@kernel.org \
--cc=david@redhat.com \
--cc=debug@rivosinc.com \
--cc=dri-devel@lists.freedesktop.org \
--cc=gerald.schaefer@linux.ibm.com \
--cc=hch@lst.de \
--cc=jgg@ziepe.ca \
--cc=jhubbard@nvidia.com \
--cc=linux-arm-kernel@lists.infradead.org \
--cc=linux-cxl@vger.kernel.org \
--cc=linux-ext4@vger.kernel.org \
--cc=linux-fsdevel@vger.kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=linux-riscv@lists.infradead.org \
--cc=linux-xfs@vger.kernel.org \
--cc=linuxppc-dev@lists.ozlabs.org \
--cc=loongarch@lists.linux.dev \
--cc=lorenzo.stoakes@oracle.com \
--cc=nvdimm@lists.linux.dev \
--cc=willy@infradead.org \
--cc=zhang.lyra@gmail.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).