* [PATCH v6 0/1] Risc-V Svinval support @ 2024-06-09 11:21 Mayuresh Chitale 2024-06-09 11:21 ` [PATCH v6 1/1] riscv: mm: Add support for Svinval extension Mayuresh Chitale 0 siblings, 1 reply; 4+ messages in thread From: Mayuresh Chitale @ 2024-06-09 11:21 UTC (permalink / raw) To: linux-riscv, linux-kernel Cc: Mayuresh Chitale, Paul Walmsley, Palmer Dabbelt, Albert Ou, Alexandre Ghiti, Samuel Holland, Andrew Jones This patch adds support for the Svinval extension as defined in the Risc V Privileged specification. Changes in v6: - Rebase on latest torvalds/master Changes in v5: - Reduce tlb flush threshold to 64 - Improve implementation of local_flush_tlb* functions Changes in v4: - Rebase and refactor as per latest changes on torvalds/master - Drop patch 1 in the series Changes in v3: - Fix incorrect vma used for sinval instructions - Use unified static key mechanism for svinval - Rebased on torvalds/master Changes in v2: - Rebased on 5.18-rc3 - update riscv_fill_hwcap to probe Svinval extension Mayuresh Chitale (1): riscv: mm: Add support for Svinval extension arch/riscv/mm/tlbflush.c | 58 ++++++++++++++++++++++++++++++++++++++++ 1 file changed, 58 insertions(+) -- 2.34.1 ^ permalink raw reply [flat|nested] 4+ messages in thread
* [PATCH v6 1/1] riscv: mm: Add support for Svinval extension 2024-06-09 11:21 [PATCH v6 0/1] Risc-V Svinval support Mayuresh Chitale @ 2024-06-09 11:21 ` Mayuresh Chitale 2024-06-21 9:15 ` Andrew Jones 0 siblings, 1 reply; 4+ messages in thread From: Mayuresh Chitale @ 2024-06-09 11:21 UTC (permalink / raw) To: linux-riscv, linux-kernel Cc: Mayuresh Chitale, Paul Walmsley, Palmer Dabbelt, Albert Ou, Alexandre Ghiti, Samuel Holland, Andrew Jones The Svinval extension splits SFENCE.VMA instruction into finer-grained invalidation and ordering operations and is mandatory for RVA23S64 profile. When Svinval is enabled the local_flush_tlb_range_threshold_asid function should use the following sequence to optimize the tlb flushes instead of a simple sfence.vma: sfence.w.inval svinval.vma . . svinval.vma sfence.inval.ir The maximum number of consecutive svinval.vma instructions that can be executed in local_flush_tlb_range_threshold_asid function is limited to 64. This is required to avoid soft lockups and the approach is similar to that used in arm64. Signed-off-by: Mayuresh Chitale <mchitale@ventanamicro.com> --- arch/riscv/mm/tlbflush.c | 58 ++++++++++++++++++++++++++++++++++++++++ 1 file changed, 58 insertions(+) diff --git a/arch/riscv/mm/tlbflush.c b/arch/riscv/mm/tlbflush.c index 9b6e86ce3867..49d7978ac8d3 100644 --- a/arch/riscv/mm/tlbflush.c +++ b/arch/riscv/mm/tlbflush.c @@ -6,6 +6,54 @@ #include <linux/hugetlb.h> #include <asm/sbi.h> #include <asm/mmu_context.h> +#include <asm/cpufeature.h> + +#define has_svinval() riscv_has_extension_unlikely(RISCV_ISA_EXT_SVINVAL) + +static inline void local_sfence_inval_ir(void) +{ + /* + * SFENCE.INVAL.IR + * 0001100 00001 00000 000 00000 1110011 + */ + __asm__ __volatile__ (".word 0x18100073" ::: "memory"); +} + +static inline void local_sfence_w_inval(void) +{ + /* + * SFENCE.W.INVAL + * 0001100 00000 00000 000 00000 1110011 + */ + __asm__ __volatile__ (".word 0x18000073" ::: "memory"); +} + +static inline void local_sinval_vma_asid(unsigned long vma, unsigned long asid) +{ + if (asid != FLUSH_TLB_NO_ASID) { + /* + * rs1 = a0 (VMA) + * rs2 = a1 (asid) + * SINVAL.VMA a0, a1 + * 0001011 01011 01010 000 00000 1110011 + */ + __asm__ __volatile__ ("add a0, %0, zero\n" + "add a1, %1, zero\n" + ".word 0x16B50073\n" + :: "r" (vma), "r" (asid) + : "a0", "a1", "memory"); + } else { + /* + * rs1 = a0 (VMA) + * rs2 = 0 + * SINVAL.VMA a0 + * 0001011 00000 01010 000 00000 1110011 + */ + __asm__ __volatile__ ("add a0, %0, zero\n" + ".word 0x16050073\n" + :: "r" (vma) : "a0", "memory"); + } +} /* * Flush entire TLB if number of entries to be flushed is greater @@ -26,6 +74,16 @@ static void local_flush_tlb_range_threshold_asid(unsigned long start, return; } + if (has_svinval()) { + local_sfence_w_inval(); + for (i = 0; i < nr_ptes_in_range; ++i) { + local_sinval_vma_asid(start, asid); + start += stride; + } + local_sfence_inval_ir(); + return; + } + for (i = 0; i < nr_ptes_in_range; ++i) { local_flush_tlb_page_asid(start, asid); start += stride; -- 2.34.1 ^ permalink raw reply related [flat|nested] 4+ messages in thread
* Re: [PATCH v6 1/1] riscv: mm: Add support for Svinval extension 2024-06-09 11:21 ` [PATCH v6 1/1] riscv: mm: Add support for Svinval extension Mayuresh Chitale @ 2024-06-21 9:15 ` Andrew Jones 2024-06-21 15:23 ` Alexandre Ghiti 0 siblings, 1 reply; 4+ messages in thread From: Andrew Jones @ 2024-06-21 9:15 UTC (permalink / raw) To: Mayuresh Chitale Cc: linux-riscv, linux-kernel, Paul Walmsley, Palmer Dabbelt, Albert Ou, Alexandre Ghiti, Samuel Holland On Sun, Jun 09, 2024 at 04:51:03PM GMT, Mayuresh Chitale wrote: > The Svinval extension splits SFENCE.VMA instruction into finer-grained > invalidation and ordering operations and is mandatory for RVA23S64 profile. > When Svinval is enabled the local_flush_tlb_range_threshold_asid function > should use the following sequence to optimize the tlb flushes instead of > a simple sfence.vma: > > sfence.w.inval > svinval.vma > . > . > svinval.vma > sfence.inval.ir > > The maximum number of consecutive svinval.vma instructions that > can be executed in local_flush_tlb_range_threshold_asid function > is limited to 64. This is required to avoid soft lockups and the > approach is similar to that used in arm64. > > Signed-off-by: Mayuresh Chitale <mchitale@ventanamicro.com> > --- > arch/riscv/mm/tlbflush.c | 58 ++++++++++++++++++++++++++++++++++++++++ > 1 file changed, 58 insertions(+) > > diff --git a/arch/riscv/mm/tlbflush.c b/arch/riscv/mm/tlbflush.c > index 9b6e86ce3867..49d7978ac8d3 100644 > --- a/arch/riscv/mm/tlbflush.c > +++ b/arch/riscv/mm/tlbflush.c > @@ -6,6 +6,54 @@ > #include <linux/hugetlb.h> > #include <asm/sbi.h> > #include <asm/mmu_context.h> > +#include <asm/cpufeature.h> > + > +#define has_svinval() riscv_has_extension_unlikely(RISCV_ISA_EXT_SVINVAL) > + > +static inline void local_sfence_inval_ir(void) > +{ > + /* > + * SFENCE.INVAL.IR > + * 0001100 00001 00000 000 00000 1110011 > + */ > + __asm__ __volatile__ (".word 0x18100073" ::: "memory"); > +} > + > +static inline void local_sfence_w_inval(void) > +{ > + /* > + * SFENCE.W.INVAL > + * 0001100 00000 00000 000 00000 1110011 > + */ > + __asm__ __volatile__ (".word 0x18000073" ::: "memory"); > +} > + > +static inline void local_sinval_vma_asid(unsigned long vma, unsigned long asid) Please name this to void local_sinval_vma(unsigned long vaddr, unsigned long asid) to match the spec's naming. But, if we want the arguments in the name, then it should be something like void local_sinval_vma_addr_asid(unsigned long vaddr, unsigned long asid) but I think that's unnecessary. > +{ > + if (asid != FLUSH_TLB_NO_ASID) { > + /* > + * rs1 = a0 (VMA) > + * rs2 = a1 (asid) > + * SINVAL.VMA a0, a1 > + * 0001011 01011 01010 000 00000 1110011 > + */ > + __asm__ __volatile__ ("add a0, %0, zero\n" > + "add a1, %1, zero\n" > + ".word 0x16B50073\n" > + :: "r" (vma), "r" (asid) > + : "a0", "a1", "memory"); > + } else { > + /* > + * rs1 = a0 (VMA) > + * rs2 = 0 > + * SINVAL.VMA a0 > + * 0001011 00000 01010 000 00000 1110011 > + */ > + __asm__ __volatile__ ("add a0, %0, zero\n" > + ".word 0x16050073\n" > + :: "r" (vma) : "a0", "memory"); > + } Please create an SINVAL_VMA instruction in insn-def.h to allow the compiler to choose which registers will be used for asid and vaddr. And, since SINVAL_VMA will be in insn-def.h, then we might as well add SFENCE_INVAL_IR and SFENCE_W_INVAL there too for consistency, even though they don't have operands. We should still create the three static inline wrapper functions here though. > +} > > /* > * Flush entire TLB if number of entries to be flushed is greater > @@ -26,6 +74,16 @@ static void local_flush_tlb_range_threshold_asid(unsigned long start, > return; > } > > + if (has_svinval()) { > + local_sfence_w_inval(); > + for (i = 0; i < nr_ptes_in_range; ++i) { Where do we limit this to 64 as the commit message states it does? > + local_sinval_vma_asid(start, asid); > + start += stride; > + } > + local_sfence_inval_ir(); > + return; > + } > + > for (i = 0; i < nr_ptes_in_range; ++i) { > local_flush_tlb_page_asid(start, asid); > start += stride; > -- > 2.34.1 > Thanks, drew ^ permalink raw reply [flat|nested] 4+ messages in thread
* Re: [PATCH v6 1/1] riscv: mm: Add support for Svinval extension 2024-06-21 9:15 ` Andrew Jones @ 2024-06-21 15:23 ` Alexandre Ghiti 0 siblings, 0 replies; 4+ messages in thread From: Alexandre Ghiti @ 2024-06-21 15:23 UTC (permalink / raw) To: Andrew Jones, Mayuresh Chitale Cc: linux-riscv, linux-kernel, Paul Walmsley, Palmer Dabbelt, Albert Ou, Alexandre Ghiti, Samuel Holland Hi Mayuresh, Andrew, On 21/06/2024 11:15, Andrew Jones wrote: > On Sun, Jun 09, 2024 at 04:51:03PM GMT, Mayuresh Chitale wrote: >> The Svinval extension splits SFENCE.VMA instruction into finer-grained >> invalidation and ordering operations and is mandatory for RVA23S64 profile. >> When Svinval is enabled the local_flush_tlb_range_threshold_asid function >> should use the following sequence to optimize the tlb flushes instead of >> a simple sfence.vma: >> >> sfence.w.inval >> svinval.vma >> . >> . >> svinval.vma >> sfence.inval.ir >> >> The maximum number of consecutive svinval.vma instructions that >> can be executed in local_flush_tlb_range_threshold_asid function >> is limited to 64. This is required to avoid soft lockups and the >> approach is similar to that used in arm64. >> >> Signed-off-by: Mayuresh Chitale <mchitale@ventanamicro.com> >> --- >> arch/riscv/mm/tlbflush.c | 58 ++++++++++++++++++++++++++++++++++++++++ >> 1 file changed, 58 insertions(+) >> >> diff --git a/arch/riscv/mm/tlbflush.c b/arch/riscv/mm/tlbflush.c >> index 9b6e86ce3867..49d7978ac8d3 100644 >> --- a/arch/riscv/mm/tlbflush.c >> +++ b/arch/riscv/mm/tlbflush.c >> @@ -6,6 +6,54 @@ >> #include <linux/hugetlb.h> >> #include <asm/sbi.h> >> #include <asm/mmu_context.h> >> +#include <asm/cpufeature.h> >> + >> +#define has_svinval() riscv_has_extension_unlikely(RISCV_ISA_EXT_SVINVAL) >> + >> +static inline void local_sfence_inval_ir(void) >> +{ >> + /* >> + * SFENCE.INVAL.IR >> + * 0001100 00001 00000 000 00000 1110011 >> + */ >> + __asm__ __volatile__ (".word 0x18100073" ::: "memory"); >> +} >> + >> +static inline void local_sfence_w_inval(void) >> +{ >> + /* >> + * SFENCE.W.INVAL >> + * 0001100 00000 00000 000 00000 1110011 >> + */ >> + __asm__ __volatile__ (".word 0x18000073" ::: "memory"); >> +} >> + >> +static inline void local_sinval_vma_asid(unsigned long vma, unsigned long asid) > Please name this to > > void local_sinval_vma(unsigned long vaddr, unsigned long asid) > > to match the spec's naming. But, if we want the arguments in the name, > then it should be something like > > void local_sinval_vma_addr_asid(unsigned long vaddr, unsigned long asid) > > but I think that's unnecessary. > >> +{ >> + if (asid != FLUSH_TLB_NO_ASID) { >> + /* >> + * rs1 = a0 (VMA) >> + * rs2 = a1 (asid) >> + * SINVAL.VMA a0, a1 >> + * 0001011 01011 01010 000 00000 1110011 >> + */ >> + __asm__ __volatile__ ("add a0, %0, zero\n" >> + "add a1, %1, zero\n" >> + ".word 0x16B50073\n" >> + :: "r" (vma), "r" (asid) >> + : "a0", "a1", "memory"); >> + } else { >> + /* >> + * rs1 = a0 (VMA) >> + * rs2 = 0 >> + * SINVAL.VMA a0 >> + * 0001011 00000 01010 000 00000 1110011 >> + */ >> + __asm__ __volatile__ ("add a0, %0, zero\n" >> + ".word 0x16050073\n" >> + :: "r" (vma) : "a0", "memory"); >> + } > Please create an SINVAL_VMA instruction in insn-def.h to allow the > compiler to choose which registers will be used for asid and vaddr. And, > since SINVAL_VMA will be in insn-def.h, then we might as well add > SFENCE_INVAL_IR and SFENCE_W_INVAL there too for consistency, even though > they don't have operands. We should still create the three static inline > wrapper functions here though. > >> +} >> >> /* >> * Flush entire TLB if number of entries to be flushed is greater >> @@ -26,6 +74,16 @@ static void local_flush_tlb_range_threshold_asid(unsigned long start, >> return; >> } >> >> + if (has_svinval()) { >> + local_sfence_w_inval(); >> + for (i = 0; i < nr_ptes_in_range; ++i) { > Where do we limit this to 64 as the commit message states it does? If the number of ptes to flush is greater than tlb_flush_all_threshold (= 64), we don't get there so this is indeed limited to 64 entries max. I think this limit should be different when using svinval, but we can do that later and the goal is to allow the vendors to come with their own threshold anyway. Thanks for reviving this Mayuresh, I'll add my RB in the next version when you fix Andrew's comments. Alex > >> + local_sinval_vma_asid(start, asid); >> + start += stride; >> + } >> + local_sfence_inval_ir(); >> + return; >> + } >> + >> for (i = 0; i < nr_ptes_in_range; ++i) { >> local_flush_tlb_page_asid(start, asid); >> start += stride; >> -- >> 2.34.1 >> > Thanks, > drew > > _______________________________________________ > linux-riscv mailing list > linux-riscv@lists.infradead.org > http://lists.infradead.org/mailman/listinfo/linux-riscv ^ permalink raw reply [flat|nested] 4+ messages in thread
end of thread, other threads:[~2024-06-21 15:23 UTC | newest] Thread overview: 4+ messages (download: mbox.gz follow: Atom feed -- links below jump to the message on this page -- 2024-06-09 11:21 [PATCH v6 0/1] Risc-V Svinval support Mayuresh Chitale 2024-06-09 11:21 ` [PATCH v6 1/1] riscv: mm: Add support for Svinval extension Mayuresh Chitale 2024-06-21 9:15 ` Andrew Jones 2024-06-21 15:23 ` Alexandre Ghiti
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox