From: Frederic Barrat <fbarrat@linux.vnet.ibm.com>
To: mpe@ellerman.id.au, aneesh.kumar@linux.vnet.ibm.com,
bsingharora@gmail.com, linuxppc-dev@lists.ozlabs.org
Cc: clombard@linux.vnet.ibm.com, alistair@popple.id.au,
vaibhav@linux.vnet.ibm.com
Subject: [PATCH v3 1/3] powerpc/mm: Add marker for contexts requiring global TLB invalidations
Date: Wed, 2 Aug 2017 22:29:28 +0200 [thread overview]
Message-ID: <20170802202930.5616-2-fbarrat@linux.vnet.ibm.com> (raw)
In-Reply-To: <20170802202930.5616-1-fbarrat@linux.vnet.ibm.com>
Introduce a new 'flags' attribute per context and define its first bit
to be a marker requiring all TLBIs for that context to be broadcasted
globally. Once that marker is set on a context, it cannot be removed.
Such a marker is useful for memory contexts used by devices behind the
NPU and CAPP/PSL. The NPU and the PSL keep their own translation cache
so they need to see all the TLBIs for those contexts.
Rename mm_is_thread_local() to mm_is_invalidation_local() to better
describe what it's doing.
Signed-off-by: Frederic Barrat <fbarrat@linux.vnet.ibm.com>
---
arch/powerpc/include/asm/book3s/64/mmu.h | 18 ++++++++++++++++++
arch/powerpc/include/asm/tlb.h | 27 +++++++++++++++++++++++----
arch/powerpc/mm/mmu_context_book3s64.c | 1 +
arch/powerpc/mm/tlb-radix.c | 8 ++++----
arch/powerpc/mm/tlb_hash64.c | 3 ++-
5 files changed, 48 insertions(+), 9 deletions(-)
diff --git a/arch/powerpc/include/asm/book3s/64/mmu.h b/arch/powerpc/include/asm/book3s/64/mmu.h
index 5b4023c616f7..03d4515ecfa6 100644
--- a/arch/powerpc/include/asm/book3s/64/mmu.h
+++ b/arch/powerpc/include/asm/book3s/64/mmu.h
@@ -79,8 +79,12 @@ struct spinlock;
/* Maximum possible number of NPUs in a system. */
#define NV_MAX_NPUS 8
+/* Bits definition for the context flags */
+#define MM_GLOBAL_TLBIE 0 /* TLBI must be global */
+
typedef struct {
mm_context_id_t id;
+ unsigned long flags;
u16 user_psize; /* page size index */
/* NPU NMMU context */
@@ -165,5 +169,19 @@ extern void radix_init_pseries(void);
static inline void radix_init_pseries(void) { };
#endif
+/*
+ * Mark the memory context as requiring global TLBIs, when used by
+ * GPUs or CAPI accelerators managing their own TLB or ERAT.
+ */
+static inline void mm_context_set_global_tlbi(mm_context_t *ctx)
+{
+ set_bit(MM_GLOBAL_TLBIE, &ctx->flags);
+}
+
+static inline bool mm_context_get_global_tlbi(mm_context_t *ctx)
+{
+ return test_bit(MM_GLOBAL_TLBIE, &ctx->flags);
+}
+
#endif /* __ASSEMBLY__ */
#endif /* _ASM_POWERPC_BOOK3S_64_MMU_H_ */
diff --git a/arch/powerpc/include/asm/tlb.h b/arch/powerpc/include/asm/tlb.h
index 609557569f65..f06dcac82097 100644
--- a/arch/powerpc/include/asm/tlb.h
+++ b/arch/powerpc/include/asm/tlb.h
@@ -69,10 +69,29 @@ static inline int mm_is_core_local(struct mm_struct *mm)
topology_sibling_cpumask(smp_processor_id()));
}
-static inline int mm_is_thread_local(struct mm_struct *mm)
+static inline int mm_is_invalidation_local(struct mm_struct *mm)
{
- return cpumask_equal(mm_cpumask(mm),
- cpumask_of(smp_processor_id()));
+ int rc;
+
+ rc = cpumask_equal(mm_cpumask(mm),
+ cpumask_of(smp_processor_id()));
+#ifdef CONFIG_PPC_BOOK3S_64
+ if (rc) {
+ /*
+ * Check if context requires global TLBI.
+ *
+ * We need to make sure the PTE update is happening
+ * before reading the context global flag. Otherwise,
+ * reading the flag may be re-ordered and happen
+ * first, and we could end up in a situation where the
+ * old PTE was seen by the NPU/PSL/device, but the
+ * TLBI is local.
+ */
+ mb();
+ rc = !mm_context_get_global_tlbi(&mm->context);
+ }
+#endif
+ return rc;
}
#else
@@ -81,7 +100,7 @@ static inline int mm_is_core_local(struct mm_struct *mm)
return 1;
}
-static inline int mm_is_thread_local(struct mm_struct *mm)
+static inline int mm_is_invalidation_local(struct mm_struct *mm)
{
return 1;
}
diff --git a/arch/powerpc/mm/mmu_context_book3s64.c b/arch/powerpc/mm/mmu_context_book3s64.c
index a75f63833284..4dfe57f9c3b0 100644
--- a/arch/powerpc/mm/mmu_context_book3s64.c
+++ b/arch/powerpc/mm/mmu_context_book3s64.c
@@ -165,6 +165,7 @@ int init_new_context(struct task_struct *tsk, struct mm_struct *mm)
return index;
mm->context.id = index;
+ mm->context.flags = 0;
#ifdef CONFIG_PPC_ICSWX
mm->context.cop_lockp = kmalloc(sizeof(spinlock_t), GFP_KERNEL);
if (!mm->context.cop_lockp) {
diff --git a/arch/powerpc/mm/tlb-radix.c b/arch/powerpc/mm/tlb-radix.c
index 16ae1bbe13f0..620ed7dced44 100644
--- a/arch/powerpc/mm/tlb-radix.c
+++ b/arch/powerpc/mm/tlb-radix.c
@@ -207,7 +207,7 @@ void radix__flush_tlb_mm(struct mm_struct *mm)
if (unlikely(pid == MMU_NO_CONTEXT))
goto no_context;
- if (!mm_is_thread_local(mm))
+ if (!mm_is_invalidation_local(mm))
_tlbie_pid(pid, RIC_FLUSH_ALL);
else
_tlbiel_pid(pid, RIC_FLUSH_ALL);
@@ -233,7 +233,7 @@ void radix__flush_tlb_pwc(struct mmu_gather *tlb, unsigned long addr)
if (unlikely(pid == MMU_NO_CONTEXT))
goto no_context;
- if (!mm_is_thread_local(mm))
+ if (!mm_is_invalidation_local(mm))
_tlbie_pid(pid, RIC_FLUSH_PWC);
else
tlbiel_pwc(pid);
@@ -252,7 +252,7 @@ void radix__flush_tlb_page_psize(struct mm_struct *mm, unsigned long vmaddr,
pid = mm ? mm->context.id : 0;
if (unlikely(pid == MMU_NO_CONTEXT))
goto bail;
- if (!mm_is_thread_local(mm))
+ if (!mm_is_invalidation_local(mm))
_tlbie_va(vmaddr, pid, ap, RIC_FLUSH_TLB);
else
_tlbiel_va(vmaddr, pid, ap, RIC_FLUSH_TLB);
@@ -335,7 +335,7 @@ void radix__flush_tlb_range_psize(struct mm_struct *mm, unsigned long start,
{
unsigned long pid;
unsigned long addr;
- int local = mm_is_thread_local(mm);
+ int local = mm_is_invalidation_local(mm);
unsigned long ap = mmu_get_ap(psize);
unsigned long page_size = 1UL << mmu_psize_defs[psize].shift;
diff --git a/arch/powerpc/mm/tlb_hash64.c b/arch/powerpc/mm/tlb_hash64.c
index b5b0fb97b9c0..618865cdc793 100644
--- a/arch/powerpc/mm/tlb_hash64.c
+++ b/arch/powerpc/mm/tlb_hash64.c
@@ -96,7 +96,8 @@ void hpte_need_flush(struct mm_struct *mm, unsigned long addr,
* flush now and return.
*/
if (!batch->active) {
- flush_hash_page(vpn, rpte, psize, ssize, mm_is_thread_local(mm));
+ flush_hash_page(vpn, rpte, psize, ssize,
+ mm_is_invalidation_local(mm));
put_cpu_var(ppc64_tlb_batch);
return;
}
--
2.11.0
next prev parent reply other threads:[~2017-08-02 20:29 UTC|newest]
Thread overview: 9+ messages / expand[flat|nested] mbox.gz Atom feed top
2017-08-02 20:29 [PATCH v3 0/3] powerpc/mm: Mark memory contexts requiring global TLBIs Frederic Barrat
2017-08-02 20:29 ` Frederic Barrat [this message]
2017-08-03 7:16 ` [PATCH v3 1/3] powerpc/mm: Add marker for contexts requiring global TLB invalidations Balbir Singh
2017-08-22 3:11 ` Alistair Popple
2017-08-02 20:29 ` [PATCH v3 2/3] cxl: Mark context requiring global TLBIs Frederic Barrat
2017-08-03 7:22 ` Balbir Singh
2017-08-22 3:09 ` Alistair Popple
2017-08-02 20:29 ` [PATCH v3 3/3] cxl: Add memory barrier to guarantee TLBI scope Frederic Barrat
2017-08-30 13:29 ` [PATCH v3 0/3] powerpc/mm: Mark memory contexts requiring global TLBIs Frederic Barrat
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20170802202930.5616-2-fbarrat@linux.vnet.ibm.com \
--to=fbarrat@linux.vnet.ibm.com \
--cc=alistair@popple.id.au \
--cc=aneesh.kumar@linux.vnet.ibm.com \
--cc=bsingharora@gmail.com \
--cc=clombard@linux.vnet.ibm.com \
--cc=linuxppc-dev@lists.ozlabs.org \
--cc=mpe@ellerman.id.au \
--cc=vaibhav@linux.vnet.ibm.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).