* [PATCH v2 0/2] mm: hugetlb: fix mremap tlb flush @ 2023-08-01 2:31 Kefeng Wang 2023-08-01 2:31 ` [PATCH v2 1/2] mm: hugetlb: use flush_hugetlb_tlb_range() in move_hugetlb_page_tables() Kefeng Wang 2023-08-01 2:31 ` [PATCH v2 2/2] arm64: hugetlb: enable __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE Kefeng Wang 0 siblings, 2 replies; 12+ messages in thread From: Kefeng Wang @ 2023-08-01 2:31 UTC (permalink / raw) To: Andrew Morton, Catalin Marinas, Will Deacon, Mike Kravetz, Muchun Song, Mina Almasry, kirill, joel, william.kucharski, kaleshsingh, linux-mm Cc: linux-arm-kernel, linux-kernel, 21cnbao, Kefeng Wang The first patch uses correct flush tlb functions when move page tables, and second patch is a small tlb flush optimization for hugepage on arm64. v2: - drop uncorrected move_normal_pmd/pud() changes - add flush_hugetlb_tlb_range() on arm64 instead of changing generic flush_tlb_range() - collect RB/ACK Kefeng Wang (2): mm: hugetlb: use flush_hugetlb_tlb_range() in move_hugetlb_page_tables() arm64: hugetlb: enable __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE arch/arm64/include/asm/hugetlb.h | 12 ++++++++++++ mm/hugetlb.c | 4 ++-- 2 files changed, 14 insertions(+), 2 deletions(-) -- 2.41.0 ^ permalink raw reply [flat|nested] 12+ messages in thread
* [PATCH v2 1/2] mm: hugetlb: use flush_hugetlb_tlb_range() in move_hugetlb_page_tables() 2023-08-01 2:31 [PATCH v2 0/2] mm: hugetlb: fix mremap tlb flush Kefeng Wang @ 2023-08-01 2:31 ` Kefeng Wang 2023-08-01 2:31 ` [PATCH v2 2/2] arm64: hugetlb: enable __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE Kefeng Wang 1 sibling, 0 replies; 12+ messages in thread From: Kefeng Wang @ 2023-08-01 2:31 UTC (permalink / raw) To: Andrew Morton, Catalin Marinas, Will Deacon, Mike Kravetz, Muchun Song, Mina Almasry, kirill, joel, william.kucharski, kaleshsingh, linux-mm Cc: linux-arm-kernel, linux-kernel, 21cnbao, Kefeng Wang, Muchun Song Archs may need to do special things when flushing hugepage tlb, so use the more applicable flush_hugetlb_tlb_range() instead of flush_tlb_range(). Reviewed-by: Mike Kravetz <mike.kravetz@oracle.com> Acked-by: Muchun Song <songmuchun@bytedance.com> Fixes: 550a7d60bd5e ("mm, hugepages: add mremap() support for hugepage backed vma") Signed-off-by: Kefeng Wang <wangkefeng.wang@huawei.com> --- mm/hugetlb.c | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/mm/hugetlb.c b/mm/hugetlb.c index 64a3239b6407..ac876bfba340 100644 --- a/mm/hugetlb.c +++ b/mm/hugetlb.c @@ -5281,9 +5281,9 @@ int move_hugetlb_page_tables(struct vm_area_struct *vma, } if (shared_pmd) - flush_tlb_range(vma, range.start, range.end); + flush_hugetlb_tlb_range(vma, range.start, range.end); else - flush_tlb_range(vma, old_end - len, old_end); + flush_hugetlb_tlb_range(vma, old_end - len, old_end); mmu_notifier_invalidate_range_end(&range); i_mmap_unlock_write(mapping); hugetlb_vma_unlock_write(vma); -- 2.41.0 ^ permalink raw reply related [flat|nested] 12+ messages in thread
* [PATCH v2 2/2] arm64: hugetlb: enable __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE 2023-08-01 2:31 [PATCH v2 0/2] mm: hugetlb: fix mremap tlb flush Kefeng Wang 2023-08-01 2:31 ` [PATCH v2 1/2] mm: hugetlb: use flush_hugetlb_tlb_range() in move_hugetlb_page_tables() Kefeng Wang @ 2023-08-01 2:31 ` Kefeng Wang 2023-08-01 2:31 ` Muchun Song ` (2 more replies) 1 sibling, 3 replies; 12+ messages in thread From: Kefeng Wang @ 2023-08-01 2:31 UTC (permalink / raw) To: Andrew Morton, Catalin Marinas, Will Deacon, Mike Kravetz, Muchun Song, Mina Almasry, kirill, joel, william.kucharski, kaleshsingh, linux-mm Cc: linux-arm-kernel, linux-kernel, 21cnbao, Kefeng Wang It is better to use huge page size instead of PAGE_SIZE for stride when flush hugepage, which reduces the loop in __flush_tlb_range(). Let's support arch's flush_hugetlb_tlb_range(), which is used in hugetlb_unshare_all_pmds(), move_hugetlb_page_tables() and hugetlb_change_protection() for now. Note, for hugepages based on contiguous bit, it has to be invalidated individually since the contiguous PTE bit is just a hint, the hardware may or may not take it into account. Signed-off-by: Kefeng Wang <wangkefeng.wang@huawei.com> --- arch/arm64/include/asm/hugetlb.h | 12 ++++++++++++ 1 file changed, 12 insertions(+) diff --git a/arch/arm64/include/asm/hugetlb.h b/arch/arm64/include/asm/hugetlb.h index 6a4a1ab8eb23..e5c2e3dd9cf0 100644 --- a/arch/arm64/include/asm/hugetlb.h +++ b/arch/arm64/include/asm/hugetlb.h @@ -60,4 +60,16 @@ extern void huge_ptep_modify_prot_commit(struct vm_area_struct *vma, #include <asm-generic/hugetlb.h> +#define __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE +static inline void flush_hugetlb_tlb_range(struct vm_area_struct *vma, + unsigned long start, + unsigned long end) +{ + unsigned long stride = huge_page_size(hstate_vma(vma)); + + if (stride != PMD_SIZE && stride != PUD_SIZE) + stride = PAGE_SIZE; + __flush_tlb_range(vma, start, end, stride, false, 0); +} + #endif /* __ASM_HUGETLB_H */ -- 2.41.0 ^ permalink raw reply related [flat|nested] 12+ messages in thread
* Re: [PATCH v2 2/2] arm64: hugetlb: enable __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE 2023-08-01 2:31 ` [PATCH v2 2/2] arm64: hugetlb: enable __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE Kefeng Wang @ 2023-08-01 2:31 ` Muchun Song 2023-08-01 11:09 ` Catalin Marinas 2023-08-01 13:56 ` [PATCH v3] " Kefeng Wang 2 siblings, 0 replies; 12+ messages in thread From: Muchun Song @ 2023-08-01 2:31 UTC (permalink / raw) To: Kefeng Wang Cc: Andrew Morton, Catalin Marinas, Will Deacon, Mike Kravetz, Mina Almasry, kirill, joel, William Kucharski, kaleshsingh, Linux-MM, linux-arm-kernel, LKML, 21cnbao > On Aug 1, 2023, at 10:31, Kefeng Wang <wangkefeng.wang@huawei.com> wrote: > > It is better to use huge page size instead of PAGE_SIZE > for stride when flush hugepage, which reduces the loop > in __flush_tlb_range(). > > Let's support arch's flush_hugetlb_tlb_range(), which is > used in hugetlb_unshare_all_pmds(), move_hugetlb_page_tables() > and hugetlb_change_protection() for now. > > Note, for hugepages based on contiguous bit, it has to be > invalidated individually since the contiguous PTE bit is > just a hint, the hardware may or may not take it into account. > > Signed-off-by: Kefeng Wang <wangkefeng.wang@huawei.com> Reviewed-by: Muchun Song <songmuchun@bytedance.com> ^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: [PATCH v2 2/2] arm64: hugetlb: enable __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE 2023-08-01 2:31 ` [PATCH v2 2/2] arm64: hugetlb: enable __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE Kefeng Wang 2023-08-01 2:31 ` Muchun Song @ 2023-08-01 11:09 ` Catalin Marinas 2023-08-01 11:22 ` Kefeng Wang 2023-08-01 13:56 ` [PATCH v3] " Kefeng Wang 2 siblings, 1 reply; 12+ messages in thread From: Catalin Marinas @ 2023-08-01 11:09 UTC (permalink / raw) To: Kefeng Wang Cc: Andrew Morton, Will Deacon, Mike Kravetz, Muchun Song, Mina Almasry, kirill, joel, william.kucharski, kaleshsingh, linux-mm, linux-arm-kernel, linux-kernel, 21cnbao On Tue, Aug 01, 2023 at 10:31:45AM +0800, Kefeng Wang wrote: > +#define __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE > +static inline void flush_hugetlb_tlb_range(struct vm_area_struct *vma, > + unsigned long start, > + unsigned long end) > +{ > + unsigned long stride = huge_page_size(hstate_vma(vma)); > + > + if (stride != PMD_SIZE && stride != PUD_SIZE) > + stride = PAGE_SIZE; > + __flush_tlb_range(vma, start, end, stride, false, 0); We could use some hints here for the tlb_level (2 for pmd, 1 for pud). Regarding the last_level argument to __flush_tlb_range(), I think it needs to stay false since this function is also called on the hugetlb_unshare_pmds() path where the pud is cleared and needs invalidating. That said, maybe you can rewrite it as a switch statement and call flush_pmd_tlb_range() or flush_pud_tlb_range() (just make sure these are defined when CONFIG_HUGETLBFS is enabled). -- Catalin ^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: [PATCH v2 2/2] arm64: hugetlb: enable __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE 2023-08-01 11:09 ` Catalin Marinas @ 2023-08-01 11:22 ` Kefeng Wang 2023-08-01 13:36 ` Kefeng Wang 0 siblings, 1 reply; 12+ messages in thread From: Kefeng Wang @ 2023-08-01 11:22 UTC (permalink / raw) To: Catalin Marinas Cc: Andrew Morton, Will Deacon, Mike Kravetz, Muchun Song, Mina Almasry, kirill, joel, william.kucharski, kaleshsingh, linux-mm, linux-arm-kernel, linux-kernel, 21cnbao On 2023/8/1 19:09, Catalin Marinas wrote: > On Tue, Aug 01, 2023 at 10:31:45AM +0800, Kefeng Wang wrote: >> +#define __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE >> +static inline void flush_hugetlb_tlb_range(struct vm_area_struct *vma, >> + unsigned long start, >> + unsigned long end) >> +{ >> + unsigned long stride = huge_page_size(hstate_vma(vma)); >> + >> + if (stride != PMD_SIZE && stride != PUD_SIZE) >> + stride = PAGE_SIZE; >> + __flush_tlb_range(vma, start, end, stride, false, 0); > > We could use some hints here for the tlb_level (2 for pmd, 1 for pud). > Regarding the last_level argument to __flush_tlb_range(), I think it > needs to stay false since this function is also called on the > hugetlb_unshare_pmds() path where the pud is cleared and needs > invalidating. > > That said, maybe you can rewrite it as a switch statement and call > flush_pmd_tlb_range() or flush_pud_tlb_range() (just make sure these are > defined when CONFIG_HUGETLBFS is enabled). > How about this way, not involved with thp ? diff --git a/arch/arm64/include/asm/hugetlb.h b/arch/arm64/include/asm/hugetlb.h index e5c2e3dd9cf0..a7ce59d3388e 100644 --- a/arch/arm64/include/asm/hugetlb.h +++ b/arch/arm64/include/asm/hugetlb.h @@ -66,10 +66,22 @@ static inline void flush_hugetlb_tlb_range(struct vm_area_struct *vma, unsigned long end) { unsigned long stride = huge_page_size(hstate_vma(vma)); + int tlb_level = 0; - if (stride != PMD_SIZE && stride != PUD_SIZE) + switch (stride) { +#ifndef __PAGETABLE_PMD_FOLDED + case PUD_SIZE: + tlb_level = 1; + break; +#endif + case PMD_SIZE: + tlb_level = 2; + break; + default: stride = PAGE_SIZE; - __flush_tlb_range(vma, start, end, stride, false, 0); + } + + __flush_tlb_range(vma, start, end, stride, false, tlb_level); } ^ permalink raw reply related [flat|nested] 12+ messages in thread
* Re: [PATCH v2 2/2] arm64: hugetlb: enable __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE 2023-08-01 11:22 ` Kefeng Wang @ 2023-08-01 13:36 ` Kefeng Wang 0 siblings, 0 replies; 12+ messages in thread From: Kefeng Wang @ 2023-08-01 13:36 UTC (permalink / raw) To: Catalin Marinas Cc: Andrew Morton, Will Deacon, Mike Kravetz, Muchun Song, Mina Almasry, kirill, joel, william.kucharski, kaleshsingh, linux-mm, linux-arm-kernel, linux-kernel, 21cnbao On 2023/8/1 19:22, Kefeng Wang wrote: > > > On 2023/8/1 19:09, Catalin Marinas wrote: >> On Tue, Aug 01, 2023 at 10:31:45AM +0800, Kefeng Wang wrote: >>> +#define __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE >>> +static inline void flush_hugetlb_tlb_range(struct vm_area_struct *vma, >>> + unsigned long start, >>> + unsigned long end) >>> +{ >>> + unsigned long stride = huge_page_size(hstate_vma(vma)); >>> + >>> + if (stride != PMD_SIZE && stride != PUD_SIZE) >>> + stride = PAGE_SIZE; >>> + __flush_tlb_range(vma, start, end, stride, false, 0); >> >> We could use some hints here for the tlb_level (2 for pmd, 1 for pud). >> Regarding the last_level argument to __flush_tlb_range(), I think it >> needs to stay false since this function is also called on the >> hugetlb_unshare_pmds() path where the pud is cleared and needs >> invalidating. >> > That said, maybe you can rewrite it as a switch statement and call >> flush_pmd_tlb_range() or flush_pud_tlb_range() (just make sure these are >> defined when CONFIG_HUGETLBFS is enabled). I try the way your mentioned, it won't change much, will send v3, thanks. ^ permalink raw reply [flat|nested] 12+ messages in thread
* [PATCH v3] arm64: hugetlb: enable __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE 2023-08-01 2:31 ` [PATCH v2 2/2] arm64: hugetlb: enable __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE Kefeng Wang 2023-08-01 2:31 ` Muchun Song 2023-08-01 11:09 ` Catalin Marinas @ 2023-08-01 13:56 ` Kefeng Wang 2023-08-01 15:28 ` Catalin Marinas 2023-08-02 1:27 ` [PATCH v4] " Kefeng Wang 2 siblings, 2 replies; 12+ messages in thread From: Kefeng Wang @ 2023-08-01 13:56 UTC (permalink / raw) To: Andrew Morton, Catalin Marinas, Will Deacon, Mike Kravetz, Muchun Song, Mina Almasry, kirill, joel, william.kucharski, kaleshsingh, linux-mm Cc: linux-arm-kernel, linux-kernel, 21cnbao, Kefeng Wang It is better to use huge page size instead of PAGE_SIZE for stride when flush hugepage, which reduces the loop in __flush_tlb_range(). Let's support arch's flush_hugetlb_tlb_range(), which is used in hugetlb_unshare_all_pmds(), move_hugetlb_page_tables() and hugetlb_change_protection() for now. Note, for hugepages based on contiguous bit, it has to be invalidated individually since the contiguous PTE bit is just a hint, the hardware may or may not take it into account. Signed-off-by: Kefeng Wang <wangkefeng.wang@huawei.com> --- v3: add tlb_level hint by using flush_pud/pmd_tlb_range, suggested by Catalin Marinas arch/arm64/include/asm/hugetlb.h | 21 +++++++++++++++++++++ arch/arm64/include/asm/pgtable.h | 4 ++-- 2 files changed, 23 insertions(+), 2 deletions(-) diff --git a/arch/arm64/include/asm/hugetlb.h b/arch/arm64/include/asm/hugetlb.h index 6a4a1ab8eb23..0acb1e8b41e9 100644 --- a/arch/arm64/include/asm/hugetlb.h +++ b/arch/arm64/include/asm/hugetlb.h @@ -60,4 +60,25 @@ extern void huge_ptep_modify_prot_commit(struct vm_area_struct *vma, #include <asm-generic/hugetlb.h> +#define __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE +static inline void flush_hugetlb_tlb_range(struct vm_area_struct *vma, + unsigned long start, + unsigned long end) +{ + unsigned long stride = huge_page_size(hstate_vma(vma)); + + switch (stride) { +#ifndef __PAGETABLE_PMD_FOLDED + case PUD_SIZE: + flush_pud_tlb_range(vma, start, end); + break; +#endif + case PMD_SIZE: + flush_pmd_tlb_range(vma, start, end); + break; + default: + __flush_tlb_range(vma, start, end, PAGE_SIZE, false, 0); + } +} + #endif /* __ASM_HUGETLB_H */ diff --git a/arch/arm64/include/asm/pgtable.h b/arch/arm64/include/asm/pgtable.h index 0bd18de9fd97..def402afcbe9 100644 --- a/arch/arm64/include/asm/pgtable.h +++ b/arch/arm64/include/asm/pgtable.h @@ -35,7 +35,7 @@ #include <linux/sched.h> #include <linux/page_table_check.h> -#ifdef CONFIG_TRANSPARENT_HUGEPAGE +#if defined(CONFIG_TRANSPARENT_HUGEPAGE) || defined(CONFIG_HUGETLB_PAGE) #define __HAVE_ARCH_FLUSH_PMD_TLB_RANGE /* Set stride and tlb_level in flush_*_tlb_range */ @@ -43,7 +43,7 @@ __flush_tlb_range(vma, addr, end, PMD_SIZE, false, 2) #define flush_pud_tlb_range(vma, addr, end) \ __flush_tlb_range(vma, addr, end, PUD_SIZE, false, 1) -#endif /* CONFIG_TRANSPARENT_HUGEPAGE */ +#endif /* CONFIG_TRANSPARENT_HUGEPAGE || CONFIG_HUGETLB_PAGE */ static inline bool arch_thp_swp_supported(void) { -- 2.41.0 ^ permalink raw reply related [flat|nested] 12+ messages in thread
* Re: [PATCH v3] arm64: hugetlb: enable __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE 2023-08-01 13:56 ` [PATCH v3] " Kefeng Wang @ 2023-08-01 15:28 ` Catalin Marinas 2023-08-02 1:27 ` [PATCH v4] " Kefeng Wang 1 sibling, 0 replies; 12+ messages in thread From: Catalin Marinas @ 2023-08-01 15:28 UTC (permalink / raw) To: Kefeng Wang Cc: Andrew Morton, Will Deacon, Mike Kravetz, Muchun Song, Mina Almasry, kirill, joel, william.kucharski, kaleshsingh, linux-mm, linux-arm-kernel, linux-kernel, 21cnbao On Tue, Aug 01, 2023 at 09:56:16PM +0800, Kefeng Wang wrote: > +#define __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE > +static inline void flush_hugetlb_tlb_range(struct vm_area_struct *vma, > + unsigned long start, > + unsigned long end) > +{ > + unsigned long stride = huge_page_size(hstate_vma(vma)); > + > + switch (stride) { > +#ifndef __PAGETABLE_PMD_FOLDED > + case PUD_SIZE: > + flush_pud_tlb_range(vma, start, end); > + break; > +#endif > + case PMD_SIZE: > + flush_pmd_tlb_range(vma, start, end); > + break; > + default: > + __flush_tlb_range(vma, start, end, PAGE_SIZE, false, 0); > + } > +} I think we should be consistent and either use __flush_tlb_range() everywhere or flush_p*d_tlb_range() together with flush_tlb_range(). Maybe using __flush_tlb_range() for the pmd/pud is not too bad, smaller patch. That said, I'd avoid the #ifndef and just go for an if/else statement: if (stride == PMD_SIZE) __flush_tlb_range(vma, start, end, stride, false, 2); else if (stride == PUD_SIZE) __flush_tlb_range(vma, start, end, stride, false, 1); else __flush_tlb_range(vma, start, end, PAGE_SIZE, 0); With the pmd folded, the P*D_SIZE is the same and the compiler should eliminate the second branch. -- Catalin ^ permalink raw reply [flat|nested] 12+ messages in thread
* [PATCH v4] arm64: hugetlb: enable __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE 2023-08-01 13:56 ` [PATCH v3] " Kefeng Wang 2023-08-01 15:28 ` Catalin Marinas @ 2023-08-02 1:27 ` Kefeng Wang 2023-08-02 1:40 ` Muchun Song 2023-08-02 10:41 ` Catalin Marinas 1 sibling, 2 replies; 12+ messages in thread From: Kefeng Wang @ 2023-08-02 1:27 UTC (permalink / raw) To: Andrew Morton, Catalin Marinas, Will Deacon, Mike Kravetz, Muchun Song, Mina Almasry, kirill, joel, william.kucharski, kaleshsingh, linux-mm Cc: linux-arm-kernel, linux-kernel, 21cnbao, Kefeng Wang It is better to use huge page size instead of PAGE_SIZE for stride when flush hugepage, which reduces the loop in __flush_tlb_range(). Let's support arch's flush_hugetlb_tlb_range(), which is used in hugetlb_unshare_all_pmds(), move_hugetlb_page_tables() and hugetlb_change_protection() for now. Note, for hugepages based on contiguous bit, it has to be invalidated individually since the contiguous PTE bit is just a hint, the hardware may or may not take it into account. Signed-off-by: Kefeng Wang <wangkefeng.wang@huawei.com> --- v4: directly pass tlb_level to __flush_tlb_range() with PMD/PUD size, suggested by Catalin v3: add tlb_level hint by using flush_pud/pmd_tlb_range, suggested by Catalin arch/arm64/include/asm/hugetlb.h | 15 +++++++++++++++ 1 file changed, 15 insertions(+) diff --git a/arch/arm64/include/asm/hugetlb.h b/arch/arm64/include/asm/hugetlb.h index 6a4a1ab8eb23..a91d6219aa78 100644 --- a/arch/arm64/include/asm/hugetlb.h +++ b/arch/arm64/include/asm/hugetlb.h @@ -60,4 +60,19 @@ extern void huge_ptep_modify_prot_commit(struct vm_area_struct *vma, #include <asm-generic/hugetlb.h> +#define __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE +static inline void flush_hugetlb_tlb_range(struct vm_area_struct *vma, + unsigned long start, + unsigned long end) +{ + unsigned long stride = huge_page_size(hstate_vma(vma)); + + if (stride == PMD_SIZE) + __flush_tlb_range(vma, start, end, stride, false, 2); + else if (stride == PUD_SIZE) + __flush_tlb_range(vma, start, end, stride, false, 1); + else + __flush_tlb_range(vma, start, end, PAGE_SIZE, false, 0); +} + #endif /* __ASM_HUGETLB_H */ -- 2.41.0 ^ permalink raw reply related [flat|nested] 12+ messages in thread
* Re: [PATCH v4] arm64: hugetlb: enable __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE 2023-08-02 1:27 ` [PATCH v4] " Kefeng Wang @ 2023-08-02 1:40 ` Muchun Song 2023-08-02 10:41 ` Catalin Marinas 1 sibling, 0 replies; 12+ messages in thread From: Muchun Song @ 2023-08-02 1:40 UTC (permalink / raw) To: Kefeng Wang Cc: Andrew Morton, Catalin Marinas, Will Deacon, Mike Kravetz, Mina Almasry, kirill, joel, William Kucharski, kaleshsingh, Linux-MM, linux-arm-kernel, LKML, 21cnbao > On Aug 2, 2023, at 09:27, Kefeng Wang <wangkefeng.wang@huawei.com> wrote: > > It is better to use huge page size instead of PAGE_SIZE > for stride when flush hugepage, which reduces the loop > in __flush_tlb_range(). > > Let's support arch's flush_hugetlb_tlb_range(), which is > used in hugetlb_unshare_all_pmds(), move_hugetlb_page_tables() > and hugetlb_change_protection() for now. > > Note, for hugepages based on contiguous bit, it has to be > invalidated individually since the contiguous PTE bit is > just a hint, the hardware may or may not take it into account. > > Signed-off-by: Kefeng Wang <wangkefeng.wang@huawei.com> Reviewed-by: Muchun Song <songmuchun@bytedance.com> ^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: [PATCH v4] arm64: hugetlb: enable __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE 2023-08-02 1:27 ` [PATCH v4] " Kefeng Wang 2023-08-02 1:40 ` Muchun Song @ 2023-08-02 10:41 ` Catalin Marinas 1 sibling, 0 replies; 12+ messages in thread From: Catalin Marinas @ 2023-08-02 10:41 UTC (permalink / raw) To: Kefeng Wang Cc: Andrew Morton, Will Deacon, Mike Kravetz, Muchun Song, Mina Almasry, kirill, joel, william.kucharski, kaleshsingh, linux-mm, linux-arm-kernel, linux-kernel, 21cnbao On Wed, Aug 02, 2023 at 09:27:31AM +0800, Kefeng Wang wrote: > It is better to use huge page size instead of PAGE_SIZE > for stride when flush hugepage, which reduces the loop > in __flush_tlb_range(). > > Let's support arch's flush_hugetlb_tlb_range(), which is > used in hugetlb_unshare_all_pmds(), move_hugetlb_page_tables() > and hugetlb_change_protection() for now. > > Note, for hugepages based on contiguous bit, it has to be > invalidated individually since the contiguous PTE bit is > just a hint, the hardware may or may not take it into account. > > Signed-off-by: Kefeng Wang <wangkefeng.wang@huawei.com> Reviewed-by: Catalin Marinas <catalin.marinas@arm.com> ^ permalink raw reply [flat|nested] 12+ messages in thread
end of thread, other threads:[~2023-08-02 10:41 UTC | newest] Thread overview: 12+ messages (download: mbox.gz follow: Atom feed -- links below jump to the message on this page -- 2023-08-01 2:31 [PATCH v2 0/2] mm: hugetlb: fix mremap tlb flush Kefeng Wang 2023-08-01 2:31 ` [PATCH v2 1/2] mm: hugetlb: use flush_hugetlb_tlb_range() in move_hugetlb_page_tables() Kefeng Wang 2023-08-01 2:31 ` [PATCH v2 2/2] arm64: hugetlb: enable __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE Kefeng Wang 2023-08-01 2:31 ` Muchun Song 2023-08-01 11:09 ` Catalin Marinas 2023-08-01 11:22 ` Kefeng Wang 2023-08-01 13:36 ` Kefeng Wang 2023-08-01 13:56 ` [PATCH v3] " Kefeng Wang 2023-08-01 15:28 ` Catalin Marinas 2023-08-02 1:27 ` [PATCH v4] " Kefeng Wang 2023-08-02 1:40 ` Muchun Song 2023-08-02 10:41 ` Catalin Marinas
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox; as well as URLs for NNTP newsgroup(s).