* [PATCH v2 0/2] mm: hugetlb: fix mremap tlb flush
@ 2023-08-01 2:31 Kefeng Wang
2023-08-01 2:31 ` [PATCH v2 1/2] mm: hugetlb: use flush_hugetlb_tlb_range() in move_hugetlb_page_tables() Kefeng Wang
2023-08-01 2:31 ` [PATCH v2 2/2] arm64: hugetlb: enable __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE Kefeng Wang
0 siblings, 2 replies; 12+ messages in thread
From: Kefeng Wang @ 2023-08-01 2:31 UTC (permalink / raw)
To: Andrew Morton, Catalin Marinas, Will Deacon, Mike Kravetz,
Muchun Song, Mina Almasry, kirill, joel, william.kucharski,
kaleshsingh, linux-mm
Cc: linux-arm-kernel, linux-kernel, 21cnbao, Kefeng Wang
The first patch uses correct flush tlb functions when move page tables,
and second patch is a small tlb flush optimization for hugepage on arm64.
v2:
- drop uncorrected move_normal_pmd/pud() changes
- add flush_hugetlb_tlb_range() on arm64 instead of changing generic
flush_tlb_range()
- collect RB/ACK
Kefeng Wang (2):
mm: hugetlb: use flush_hugetlb_tlb_range() in
move_hugetlb_page_tables()
arm64: hugetlb: enable __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE
arch/arm64/include/asm/hugetlb.h | 12 ++++++++++++
mm/hugetlb.c | 4 ++--
2 files changed, 14 insertions(+), 2 deletions(-)
--
2.41.0
^ permalink raw reply [flat|nested] 12+ messages in thread
* [PATCH v2 1/2] mm: hugetlb: use flush_hugetlb_tlb_range() in move_hugetlb_page_tables()
2023-08-01 2:31 [PATCH v2 0/2] mm: hugetlb: fix mremap tlb flush Kefeng Wang
@ 2023-08-01 2:31 ` Kefeng Wang
2023-08-01 2:31 ` [PATCH v2 2/2] arm64: hugetlb: enable __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE Kefeng Wang
1 sibling, 0 replies; 12+ messages in thread
From: Kefeng Wang @ 2023-08-01 2:31 UTC (permalink / raw)
To: Andrew Morton, Catalin Marinas, Will Deacon, Mike Kravetz,
Muchun Song, Mina Almasry, kirill, joel, william.kucharski,
kaleshsingh, linux-mm
Cc: linux-arm-kernel, linux-kernel, 21cnbao, Kefeng Wang, Muchun Song
Archs may need to do special things when flushing hugepage tlb,
so use the more applicable flush_hugetlb_tlb_range() instead of
flush_tlb_range().
Reviewed-by: Mike Kravetz <mike.kravetz@oracle.com>
Acked-by: Muchun Song <songmuchun@bytedance.com>
Fixes: 550a7d60bd5e ("mm, hugepages: add mremap() support for hugepage backed vma")
Signed-off-by: Kefeng Wang <wangkefeng.wang@huawei.com>
---
mm/hugetlb.c | 4 ++--
1 file changed, 2 insertions(+), 2 deletions(-)
diff --git a/mm/hugetlb.c b/mm/hugetlb.c
index 64a3239b6407..ac876bfba340 100644
--- a/mm/hugetlb.c
+++ b/mm/hugetlb.c
@@ -5281,9 +5281,9 @@ int move_hugetlb_page_tables(struct vm_area_struct *vma,
}
if (shared_pmd)
- flush_tlb_range(vma, range.start, range.end);
+ flush_hugetlb_tlb_range(vma, range.start, range.end);
else
- flush_tlb_range(vma, old_end - len, old_end);
+ flush_hugetlb_tlb_range(vma, old_end - len, old_end);
mmu_notifier_invalidate_range_end(&range);
i_mmap_unlock_write(mapping);
hugetlb_vma_unlock_write(vma);
--
2.41.0
^ permalink raw reply related [flat|nested] 12+ messages in thread
* [PATCH v2 2/2] arm64: hugetlb: enable __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE
2023-08-01 2:31 [PATCH v2 0/2] mm: hugetlb: fix mremap tlb flush Kefeng Wang
2023-08-01 2:31 ` [PATCH v2 1/2] mm: hugetlb: use flush_hugetlb_tlb_range() in move_hugetlb_page_tables() Kefeng Wang
@ 2023-08-01 2:31 ` Kefeng Wang
2023-08-01 2:31 ` Muchun Song
` (2 more replies)
1 sibling, 3 replies; 12+ messages in thread
From: Kefeng Wang @ 2023-08-01 2:31 UTC (permalink / raw)
To: Andrew Morton, Catalin Marinas, Will Deacon, Mike Kravetz,
Muchun Song, Mina Almasry, kirill, joel, william.kucharski,
kaleshsingh, linux-mm
Cc: linux-arm-kernel, linux-kernel, 21cnbao, Kefeng Wang
It is better to use huge page size instead of PAGE_SIZE
for stride when flush hugepage, which reduces the loop
in __flush_tlb_range().
Let's support arch's flush_hugetlb_tlb_range(), which is
used in hugetlb_unshare_all_pmds(), move_hugetlb_page_tables()
and hugetlb_change_protection() for now.
Note, for hugepages based on contiguous bit, it has to be
invalidated individually since the contiguous PTE bit is
just a hint, the hardware may or may not take it into account.
Signed-off-by: Kefeng Wang <wangkefeng.wang@huawei.com>
---
arch/arm64/include/asm/hugetlb.h | 12 ++++++++++++
1 file changed, 12 insertions(+)
diff --git a/arch/arm64/include/asm/hugetlb.h b/arch/arm64/include/asm/hugetlb.h
index 6a4a1ab8eb23..e5c2e3dd9cf0 100644
--- a/arch/arm64/include/asm/hugetlb.h
+++ b/arch/arm64/include/asm/hugetlb.h
@@ -60,4 +60,16 @@ extern void huge_ptep_modify_prot_commit(struct vm_area_struct *vma,
#include <asm-generic/hugetlb.h>
+#define __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE
+static inline void flush_hugetlb_tlb_range(struct vm_area_struct *vma,
+ unsigned long start,
+ unsigned long end)
+{
+ unsigned long stride = huge_page_size(hstate_vma(vma));
+
+ if (stride != PMD_SIZE && stride != PUD_SIZE)
+ stride = PAGE_SIZE;
+ __flush_tlb_range(vma, start, end, stride, false, 0);
+}
+
#endif /* __ASM_HUGETLB_H */
--
2.41.0
^ permalink raw reply related [flat|nested] 12+ messages in thread
* Re: [PATCH v2 2/2] arm64: hugetlb: enable __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE
2023-08-01 2:31 ` [PATCH v2 2/2] arm64: hugetlb: enable __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE Kefeng Wang
@ 2023-08-01 2:31 ` Muchun Song
2023-08-01 11:09 ` Catalin Marinas
2023-08-01 13:56 ` [PATCH v3] " Kefeng Wang
2 siblings, 0 replies; 12+ messages in thread
From: Muchun Song @ 2023-08-01 2:31 UTC (permalink / raw)
To: Kefeng Wang
Cc: Andrew Morton, Catalin Marinas, Will Deacon, Mike Kravetz,
Mina Almasry, kirill, joel, William Kucharski, kaleshsingh,
Linux-MM, linux-arm-kernel, LKML, 21cnbao
> On Aug 1, 2023, at 10:31, Kefeng Wang <wangkefeng.wang@huawei.com> wrote:
>
> It is better to use huge page size instead of PAGE_SIZE
> for stride when flush hugepage, which reduces the loop
> in __flush_tlb_range().
>
> Let's support arch's flush_hugetlb_tlb_range(), which is
> used in hugetlb_unshare_all_pmds(), move_hugetlb_page_tables()
> and hugetlb_change_protection() for now.
>
> Note, for hugepages based on contiguous bit, it has to be
> invalidated individually since the contiguous PTE bit is
> just a hint, the hardware may or may not take it into account.
>
> Signed-off-by: Kefeng Wang <wangkefeng.wang@huawei.com>
Reviewed-by: Muchun Song <songmuchun@bytedance.com>
^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: [PATCH v2 2/2] arm64: hugetlb: enable __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE
2023-08-01 2:31 ` [PATCH v2 2/2] arm64: hugetlb: enable __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE Kefeng Wang
2023-08-01 2:31 ` Muchun Song
@ 2023-08-01 11:09 ` Catalin Marinas
2023-08-01 11:22 ` Kefeng Wang
2023-08-01 13:56 ` [PATCH v3] " Kefeng Wang
2 siblings, 1 reply; 12+ messages in thread
From: Catalin Marinas @ 2023-08-01 11:09 UTC (permalink / raw)
To: Kefeng Wang
Cc: Andrew Morton, Will Deacon, Mike Kravetz, Muchun Song,
Mina Almasry, kirill, joel, william.kucharski, kaleshsingh,
linux-mm, linux-arm-kernel, linux-kernel, 21cnbao
On Tue, Aug 01, 2023 at 10:31:45AM +0800, Kefeng Wang wrote:
> +#define __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE
> +static inline void flush_hugetlb_tlb_range(struct vm_area_struct *vma,
> + unsigned long start,
> + unsigned long end)
> +{
> + unsigned long stride = huge_page_size(hstate_vma(vma));
> +
> + if (stride != PMD_SIZE && stride != PUD_SIZE)
> + stride = PAGE_SIZE;
> + __flush_tlb_range(vma, start, end, stride, false, 0);
We could use some hints here for the tlb_level (2 for pmd, 1 for pud).
Regarding the last_level argument to __flush_tlb_range(), I think it
needs to stay false since this function is also called on the
hugetlb_unshare_pmds() path where the pud is cleared and needs
invalidating.
That said, maybe you can rewrite it as a switch statement and call
flush_pmd_tlb_range() or flush_pud_tlb_range() (just make sure these are
defined when CONFIG_HUGETLBFS is enabled).
--
Catalin
^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: [PATCH v2 2/2] arm64: hugetlb: enable __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE
2023-08-01 11:09 ` Catalin Marinas
@ 2023-08-01 11:22 ` Kefeng Wang
2023-08-01 13:36 ` Kefeng Wang
0 siblings, 1 reply; 12+ messages in thread
From: Kefeng Wang @ 2023-08-01 11:22 UTC (permalink / raw)
To: Catalin Marinas
Cc: Andrew Morton, Will Deacon, Mike Kravetz, Muchun Song,
Mina Almasry, kirill, joel, william.kucharski, kaleshsingh,
linux-mm, linux-arm-kernel, linux-kernel, 21cnbao
On 2023/8/1 19:09, Catalin Marinas wrote:
> On Tue, Aug 01, 2023 at 10:31:45AM +0800, Kefeng Wang wrote:
>> +#define __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE
>> +static inline void flush_hugetlb_tlb_range(struct vm_area_struct *vma,
>> + unsigned long start,
>> + unsigned long end)
>> +{
>> + unsigned long stride = huge_page_size(hstate_vma(vma));
>> +
>> + if (stride != PMD_SIZE && stride != PUD_SIZE)
>> + stride = PAGE_SIZE;
>> + __flush_tlb_range(vma, start, end, stride, false, 0);
>
> We could use some hints here for the tlb_level (2 for pmd, 1 for pud).
> Regarding the last_level argument to __flush_tlb_range(), I think it
> needs to stay false since this function is also called on the
> hugetlb_unshare_pmds() path where the pud is cleared and needs
> invalidating.
> > That said, maybe you can rewrite it as a switch statement and call
> flush_pmd_tlb_range() or flush_pud_tlb_range() (just make sure these are
> defined when CONFIG_HUGETLBFS is enabled).
>
How about this way, not involved with thp ?
diff --git a/arch/arm64/include/asm/hugetlb.h
b/arch/arm64/include/asm/hugetlb.h
index e5c2e3dd9cf0..a7ce59d3388e 100644
--- a/arch/arm64/include/asm/hugetlb.h
+++ b/arch/arm64/include/asm/hugetlb.h
@@ -66,10 +66,22 @@ static inline void flush_hugetlb_tlb_range(struct
vm_area_struct *vma,
unsigned long end)
{
unsigned long stride = huge_page_size(hstate_vma(vma));
+ int tlb_level = 0;
- if (stride != PMD_SIZE && stride != PUD_SIZE)
+ switch (stride) {
+#ifndef __PAGETABLE_PMD_FOLDED
+ case PUD_SIZE:
+ tlb_level = 1;
+ break;
+#endif
+ case PMD_SIZE:
+ tlb_level = 2;
+ break;
+ default:
stride = PAGE_SIZE;
- __flush_tlb_range(vma, start, end, stride, false, 0);
+ }
+
+ __flush_tlb_range(vma, start, end, stride, false, tlb_level);
}
^ permalink raw reply related [flat|nested] 12+ messages in thread
* Re: [PATCH v2 2/2] arm64: hugetlb: enable __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE
2023-08-01 11:22 ` Kefeng Wang
@ 2023-08-01 13:36 ` Kefeng Wang
0 siblings, 0 replies; 12+ messages in thread
From: Kefeng Wang @ 2023-08-01 13:36 UTC (permalink / raw)
To: Catalin Marinas
Cc: Andrew Morton, Will Deacon, Mike Kravetz, Muchun Song,
Mina Almasry, kirill, joel, william.kucharski, kaleshsingh,
linux-mm, linux-arm-kernel, linux-kernel, 21cnbao
On 2023/8/1 19:22, Kefeng Wang wrote:
>
>
> On 2023/8/1 19:09, Catalin Marinas wrote:
>> On Tue, Aug 01, 2023 at 10:31:45AM +0800, Kefeng Wang wrote:
>>> +#define __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE
>>> +static inline void flush_hugetlb_tlb_range(struct vm_area_struct *vma,
>>> + unsigned long start,
>>> + unsigned long end)
>>> +{
>>> + unsigned long stride = huge_page_size(hstate_vma(vma));
>>> +
>>> + if (stride != PMD_SIZE && stride != PUD_SIZE)
>>> + stride = PAGE_SIZE;
>>> + __flush_tlb_range(vma, start, end, stride, false, 0);
>>
>> We could use some hints here for the tlb_level (2 for pmd, 1 for pud).
>> Regarding the last_level argument to __flush_tlb_range(), I think it
>> needs to stay false since this function is also called on the
>> hugetlb_unshare_pmds() path where the pud is cleared and needs
>> invalidating.
>> > That said, maybe you can rewrite it as a switch statement and call
>> flush_pmd_tlb_range() or flush_pud_tlb_range() (just make sure these are
>> defined when CONFIG_HUGETLBFS is enabled).
I try the way your mentioned, it won't change much, will send v3, thanks.
^ permalink raw reply [flat|nested] 12+ messages in thread
* [PATCH v3] arm64: hugetlb: enable __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE
2023-08-01 2:31 ` [PATCH v2 2/2] arm64: hugetlb: enable __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE Kefeng Wang
2023-08-01 2:31 ` Muchun Song
2023-08-01 11:09 ` Catalin Marinas
@ 2023-08-01 13:56 ` Kefeng Wang
2023-08-01 15:28 ` Catalin Marinas
2023-08-02 1:27 ` [PATCH v4] " Kefeng Wang
2 siblings, 2 replies; 12+ messages in thread
From: Kefeng Wang @ 2023-08-01 13:56 UTC (permalink / raw)
To: Andrew Morton, Catalin Marinas, Will Deacon, Mike Kravetz,
Muchun Song, Mina Almasry, kirill, joel, william.kucharski,
kaleshsingh, linux-mm
Cc: linux-arm-kernel, linux-kernel, 21cnbao, Kefeng Wang
It is better to use huge page size instead of PAGE_SIZE
for stride when flush hugepage, which reduces the loop
in __flush_tlb_range().
Let's support arch's flush_hugetlb_tlb_range(), which is
used in hugetlb_unshare_all_pmds(), move_hugetlb_page_tables()
and hugetlb_change_protection() for now.
Note, for hugepages based on contiguous bit, it has to be
invalidated individually since the contiguous PTE bit is
just a hint, the hardware may or may not take it into account.
Signed-off-by: Kefeng Wang <wangkefeng.wang@huawei.com>
---
v3: add tlb_level hint by using flush_pud/pmd_tlb_range,
suggested by Catalin Marinas
arch/arm64/include/asm/hugetlb.h | 21 +++++++++++++++++++++
arch/arm64/include/asm/pgtable.h | 4 ++--
2 files changed, 23 insertions(+), 2 deletions(-)
diff --git a/arch/arm64/include/asm/hugetlb.h b/arch/arm64/include/asm/hugetlb.h
index 6a4a1ab8eb23..0acb1e8b41e9 100644
--- a/arch/arm64/include/asm/hugetlb.h
+++ b/arch/arm64/include/asm/hugetlb.h
@@ -60,4 +60,25 @@ extern void huge_ptep_modify_prot_commit(struct vm_area_struct *vma,
#include <asm-generic/hugetlb.h>
+#define __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE
+static inline void flush_hugetlb_tlb_range(struct vm_area_struct *vma,
+ unsigned long start,
+ unsigned long end)
+{
+ unsigned long stride = huge_page_size(hstate_vma(vma));
+
+ switch (stride) {
+#ifndef __PAGETABLE_PMD_FOLDED
+ case PUD_SIZE:
+ flush_pud_tlb_range(vma, start, end);
+ break;
+#endif
+ case PMD_SIZE:
+ flush_pmd_tlb_range(vma, start, end);
+ break;
+ default:
+ __flush_tlb_range(vma, start, end, PAGE_SIZE, false, 0);
+ }
+}
+
#endif /* __ASM_HUGETLB_H */
diff --git a/arch/arm64/include/asm/pgtable.h b/arch/arm64/include/asm/pgtable.h
index 0bd18de9fd97..def402afcbe9 100644
--- a/arch/arm64/include/asm/pgtable.h
+++ b/arch/arm64/include/asm/pgtable.h
@@ -35,7 +35,7 @@
#include <linux/sched.h>
#include <linux/page_table_check.h>
-#ifdef CONFIG_TRANSPARENT_HUGEPAGE
+#if defined(CONFIG_TRANSPARENT_HUGEPAGE) || defined(CONFIG_HUGETLB_PAGE)
#define __HAVE_ARCH_FLUSH_PMD_TLB_RANGE
/* Set stride and tlb_level in flush_*_tlb_range */
@@ -43,7 +43,7 @@
__flush_tlb_range(vma, addr, end, PMD_SIZE, false, 2)
#define flush_pud_tlb_range(vma, addr, end) \
__flush_tlb_range(vma, addr, end, PUD_SIZE, false, 1)
-#endif /* CONFIG_TRANSPARENT_HUGEPAGE */
+#endif /* CONFIG_TRANSPARENT_HUGEPAGE || CONFIG_HUGETLB_PAGE */
static inline bool arch_thp_swp_supported(void)
{
--
2.41.0
^ permalink raw reply related [flat|nested] 12+ messages in thread
* Re: [PATCH v3] arm64: hugetlb: enable __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE
2023-08-01 13:56 ` [PATCH v3] " Kefeng Wang
@ 2023-08-01 15:28 ` Catalin Marinas
2023-08-02 1:27 ` [PATCH v4] " Kefeng Wang
1 sibling, 0 replies; 12+ messages in thread
From: Catalin Marinas @ 2023-08-01 15:28 UTC (permalink / raw)
To: Kefeng Wang
Cc: Andrew Morton, Will Deacon, Mike Kravetz, Muchun Song,
Mina Almasry, kirill, joel, william.kucharski, kaleshsingh,
linux-mm, linux-arm-kernel, linux-kernel, 21cnbao
On Tue, Aug 01, 2023 at 09:56:16PM +0800, Kefeng Wang wrote:
> +#define __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE
> +static inline void flush_hugetlb_tlb_range(struct vm_area_struct *vma,
> + unsigned long start,
> + unsigned long end)
> +{
> + unsigned long stride = huge_page_size(hstate_vma(vma));
> +
> + switch (stride) {
> +#ifndef __PAGETABLE_PMD_FOLDED
> + case PUD_SIZE:
> + flush_pud_tlb_range(vma, start, end);
> + break;
> +#endif
> + case PMD_SIZE:
> + flush_pmd_tlb_range(vma, start, end);
> + break;
> + default:
> + __flush_tlb_range(vma, start, end, PAGE_SIZE, false, 0);
> + }
> +}
I think we should be consistent and either use __flush_tlb_range()
everywhere or flush_p*d_tlb_range() together with flush_tlb_range().
Maybe using __flush_tlb_range() for the pmd/pud is not too bad, smaller
patch.
That said, I'd avoid the #ifndef and just go for an if/else statement:
if (stride == PMD_SIZE)
__flush_tlb_range(vma, start, end, stride, false, 2);
else if (stride == PUD_SIZE)
__flush_tlb_range(vma, start, end, stride, false, 1);
else
__flush_tlb_range(vma, start, end, PAGE_SIZE, 0);
With the pmd folded, the P*D_SIZE is the same and the compiler should
eliminate the second branch.
--
Catalin
^ permalink raw reply [flat|nested] 12+ messages in thread
* [PATCH v4] arm64: hugetlb: enable __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE
2023-08-01 13:56 ` [PATCH v3] " Kefeng Wang
2023-08-01 15:28 ` Catalin Marinas
@ 2023-08-02 1:27 ` Kefeng Wang
2023-08-02 1:40 ` Muchun Song
2023-08-02 10:41 ` Catalin Marinas
1 sibling, 2 replies; 12+ messages in thread
From: Kefeng Wang @ 2023-08-02 1:27 UTC (permalink / raw)
To: Andrew Morton, Catalin Marinas, Will Deacon, Mike Kravetz,
Muchun Song, Mina Almasry, kirill, joel, william.kucharski,
kaleshsingh, linux-mm
Cc: linux-arm-kernel, linux-kernel, 21cnbao, Kefeng Wang
It is better to use huge page size instead of PAGE_SIZE
for stride when flush hugepage, which reduces the loop
in __flush_tlb_range().
Let's support arch's flush_hugetlb_tlb_range(), which is
used in hugetlb_unshare_all_pmds(), move_hugetlb_page_tables()
and hugetlb_change_protection() for now.
Note, for hugepages based on contiguous bit, it has to be
invalidated individually since the contiguous PTE bit is
just a hint, the hardware may or may not take it into account.
Signed-off-by: Kefeng Wang <wangkefeng.wang@huawei.com>
---
v4: directly pass tlb_level to __flush_tlb_range() with PMD/PUD size,
suggested by Catalin
v3: add tlb_level hint by using flush_pud/pmd_tlb_range,
suggested by Catalin
arch/arm64/include/asm/hugetlb.h | 15 +++++++++++++++
1 file changed, 15 insertions(+)
diff --git a/arch/arm64/include/asm/hugetlb.h b/arch/arm64/include/asm/hugetlb.h
index 6a4a1ab8eb23..a91d6219aa78 100644
--- a/arch/arm64/include/asm/hugetlb.h
+++ b/arch/arm64/include/asm/hugetlb.h
@@ -60,4 +60,19 @@ extern void huge_ptep_modify_prot_commit(struct vm_area_struct *vma,
#include <asm-generic/hugetlb.h>
+#define __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE
+static inline void flush_hugetlb_tlb_range(struct vm_area_struct *vma,
+ unsigned long start,
+ unsigned long end)
+{
+ unsigned long stride = huge_page_size(hstate_vma(vma));
+
+ if (stride == PMD_SIZE)
+ __flush_tlb_range(vma, start, end, stride, false, 2);
+ else if (stride == PUD_SIZE)
+ __flush_tlb_range(vma, start, end, stride, false, 1);
+ else
+ __flush_tlb_range(vma, start, end, PAGE_SIZE, false, 0);
+}
+
#endif /* __ASM_HUGETLB_H */
--
2.41.0
^ permalink raw reply related [flat|nested] 12+ messages in thread
* Re: [PATCH v4] arm64: hugetlb: enable __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE
2023-08-02 1:27 ` [PATCH v4] " Kefeng Wang
@ 2023-08-02 1:40 ` Muchun Song
2023-08-02 10:41 ` Catalin Marinas
1 sibling, 0 replies; 12+ messages in thread
From: Muchun Song @ 2023-08-02 1:40 UTC (permalink / raw)
To: Kefeng Wang
Cc: Andrew Morton, Catalin Marinas, Will Deacon, Mike Kravetz,
Mina Almasry, kirill, joel, William Kucharski, kaleshsingh,
Linux-MM, linux-arm-kernel, LKML, 21cnbao
> On Aug 2, 2023, at 09:27, Kefeng Wang <wangkefeng.wang@huawei.com> wrote:
>
> It is better to use huge page size instead of PAGE_SIZE
> for stride when flush hugepage, which reduces the loop
> in __flush_tlb_range().
>
> Let's support arch's flush_hugetlb_tlb_range(), which is
> used in hugetlb_unshare_all_pmds(), move_hugetlb_page_tables()
> and hugetlb_change_protection() for now.
>
> Note, for hugepages based on contiguous bit, it has to be
> invalidated individually since the contiguous PTE bit is
> just a hint, the hardware may or may not take it into account.
>
> Signed-off-by: Kefeng Wang <wangkefeng.wang@huawei.com>
Reviewed-by: Muchun Song <songmuchun@bytedance.com>
^ permalink raw reply [flat|nested] 12+ messages in thread
* Re: [PATCH v4] arm64: hugetlb: enable __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE
2023-08-02 1:27 ` [PATCH v4] " Kefeng Wang
2023-08-02 1:40 ` Muchun Song
@ 2023-08-02 10:41 ` Catalin Marinas
1 sibling, 0 replies; 12+ messages in thread
From: Catalin Marinas @ 2023-08-02 10:41 UTC (permalink / raw)
To: Kefeng Wang
Cc: Andrew Morton, Will Deacon, Mike Kravetz, Muchun Song,
Mina Almasry, kirill, joel, william.kucharski, kaleshsingh,
linux-mm, linux-arm-kernel, linux-kernel, 21cnbao
On Wed, Aug 02, 2023 at 09:27:31AM +0800, Kefeng Wang wrote:
> It is better to use huge page size instead of PAGE_SIZE
> for stride when flush hugepage, which reduces the loop
> in __flush_tlb_range().
>
> Let's support arch's flush_hugetlb_tlb_range(), which is
> used in hugetlb_unshare_all_pmds(), move_hugetlb_page_tables()
> and hugetlb_change_protection() for now.
>
> Note, for hugepages based on contiguous bit, it has to be
> invalidated individually since the contiguous PTE bit is
> just a hint, the hardware may or may not take it into account.
>
> Signed-off-by: Kefeng Wang <wangkefeng.wang@huawei.com>
Reviewed-by: Catalin Marinas <catalin.marinas@arm.com>
^ permalink raw reply [flat|nested] 12+ messages in thread
end of thread, other threads:[~2023-08-02 10:41 UTC | newest]
Thread overview: 12+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2023-08-01 2:31 [PATCH v2 0/2] mm: hugetlb: fix mremap tlb flush Kefeng Wang
2023-08-01 2:31 ` [PATCH v2 1/2] mm: hugetlb: use flush_hugetlb_tlb_range() in move_hugetlb_page_tables() Kefeng Wang
2023-08-01 2:31 ` [PATCH v2 2/2] arm64: hugetlb: enable __HAVE_ARCH_FLUSH_HUGETLB_TLB_RANGE Kefeng Wang
2023-08-01 2:31 ` Muchun Song
2023-08-01 11:09 ` Catalin Marinas
2023-08-01 11:22 ` Kefeng Wang
2023-08-01 13:36 ` Kefeng Wang
2023-08-01 13:56 ` [PATCH v3] " Kefeng Wang
2023-08-01 15:28 ` Catalin Marinas
2023-08-02 1:27 ` [PATCH v4] " Kefeng Wang
2023-08-02 1:40 ` Muchun Song
2023-08-02 10:41 ` Catalin Marinas
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).