* [PATCH v3 00/20] accel/tcg: Introduce tlb_fill_align hook
@ 2024-10-09 0:04 Richard Henderson
2024-10-09 0:04 ` [PATCH v3 01/20] accel/tcg: Assert noreturn from write-only page for atomics Richard Henderson
` (19 more replies)
0 siblings, 20 replies; 33+ messages in thread
From: Richard Henderson @ 2024-10-09 0:04 UTC (permalink / raw)
To: qemu-devel; +Cc: deller, peter.maydell, alex.bennee, linux-parisc, qemu-arm
This new hook will allow targets to recognize an alignment
fault with the correct priority with respect to other faults
that can be raised by paging.
This should fix several hppa fault priority issues, most
importantly that access permissions come before alignment.
This should fix the documented error in the Arm alignment
fault due to memory type.
Changes for v3:
- Change the tlb_fill_hook to return data into provided
CPUTLBEntryFull, instead of calling tlb_set_page*.
- I have dropped some r-b, correspondingly.
Patches needing review:
05-accel-tcg-Add-TCGCPUOps.tlb_fill_align.patch
06-accel-tcg-Use-the-alignment-test-in-tlb_fill_alig.patch
11-target-hppa-Implement-TCGCPUOps.tlb_fill_align.patch
19-target-arm-Implement-TCGCPUOps.tlb_fill_align.patch
r~
Richard Henderson (20):
accel/tcg: Assert noreturn from write-only page for atomics
include/exec/memop: Move get_alignment_bits from tcg.h
include/exec/memop: Rename get_alignment_bits
include/exec/memop: Introduce memop_atomicity_bits
accel/tcg: Add TCGCPUOps.tlb_fill_align
accel/tcg: Use the alignment test in tlb_fill_align
target/hppa: Add MemOp argument to hppa_get_physical_address
target/hppa: Perform access rights before protection id check
target/hppa: Fix priority of T, D, and B page faults
target/hppa: Handle alignment faults in hppa_get_physical_address
target/hppa: Implement TCGCPUOps.tlb_fill_align
target/arm: Pass MemOp to get_phys_addr
target/arm: Pass MemOp to get_phys_addr_with_space_nogpc
target/arm: Pass MemOp to get_phys_addr_gpc
target/arm: Pass MemOp to get_phys_addr_nogpc
target/arm: Pass MemOp through get_phys_addr_twostage
target/arm: Pass MemOp to get_phys_addr_lpae
target/arm: Move device detection earlier in get_phys_addr_lpae
target/arm: Implement TCGCPUOps.tlb_fill_align
target/arm: Fix alignment fault priority in get_phys_addr_lpae
include/exec/memop.h | 47 ++++++++++
include/hw/core/cpu.h | 4 +-
include/hw/core/tcg-cpu-ops.h | 26 ++++++
include/qemu/typedefs.h | 1 +
include/tcg/tcg.h | 23 -----
target/arm/internals.h | 12 +--
target/hppa/cpu.h | 8 +-
accel/tcg/cputlb.c | 160 +++++++++++++++++----------------
accel/tcg/user-exec.c | 4 +-
target/arm/cpu.c | 2 +-
target/arm/helper.c | 9 +-
target/arm/ptw.c | 141 ++++++++++++++++-------------
target/arm/tcg/cpu-v7m.c | 2 +-
target/arm/tcg/m_helper.c | 8 +-
target/arm/tcg/tlb_helper.c | 49 ++++------
target/arm/tcg/translate-a64.c | 4 +-
target/hppa/cpu.c | 2 +-
target/hppa/int_helper.c | 2 +-
target/hppa/mem_helper.c | 55 +++++++-----
target/hppa/op_helper.c | 2 +-
target/xtensa/translate.c | 2 +-
tcg/tcg-op-ldst.c | 6 +-
tcg/tcg.c | 2 +-
tcg/arm/tcg-target.c.inc | 4 +-
tcg/sparc64/tcg-target.c.inc | 2 +-
25 files changed, 326 insertions(+), 251 deletions(-)
--
2.43.0
^ permalink raw reply [flat|nested] 33+ messages in thread
* [PATCH v3 01/20] accel/tcg: Assert noreturn from write-only page for atomics
2024-10-09 0:04 [PATCH v3 00/20] accel/tcg: Introduce tlb_fill_align hook Richard Henderson
@ 2024-10-09 0:04 ` Richard Henderson
2024-10-09 0:04 ` [PATCH v3 02/20] include/exec/memop: Move get_alignment_bits from tcg.h Richard Henderson
` (18 subsequent siblings)
19 siblings, 0 replies; 33+ messages in thread
From: Richard Henderson @ 2024-10-09 0:04 UTC (permalink / raw)
To: qemu-devel
Cc: deller, peter.maydell, alex.bennee, linux-parisc, qemu-arm,
Helge Deller
There should be no "just in case"; the page is already
in the tlb, and known to be not readable.
Reviewed-by: Helge Deller <deller@gmx.de>
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
---
accel/tcg/cputlb.c | 5 ++---
1 file changed, 2 insertions(+), 3 deletions(-)
diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c
index 117b516739..fd6459b695 100644
--- a/accel/tcg/cputlb.c
+++ b/accel/tcg/cputlb.c
@@ -1852,10 +1852,9 @@ static void *atomic_mmu_lookup(CPUState *cpu, vaddr addr, MemOpIdx oi,
/*
* Since we don't support reads and writes to different
* addresses, and we do have the proper page loaded for
- * write, this shouldn't ever return. But just in case,
- * handle via stop-the-world.
+ * write, this shouldn't ever return.
*/
- goto stop_the_world;
+ g_assert_not_reached();
}
/* Collect tlb flags for read. */
tlb_addr |= tlbe->addr_read;
--
2.43.0
^ permalink raw reply related [flat|nested] 33+ messages in thread
* [PATCH v3 02/20] include/exec/memop: Move get_alignment_bits from tcg.h
2024-10-09 0:04 [PATCH v3 00/20] accel/tcg: Introduce tlb_fill_align hook Richard Henderson
2024-10-09 0:04 ` [PATCH v3 01/20] accel/tcg: Assert noreturn from write-only page for atomics Richard Henderson
@ 2024-10-09 0:04 ` Richard Henderson
2024-10-09 0:04 ` [PATCH v3 03/20] include/exec/memop: Rename get_alignment_bits Richard Henderson
` (17 subsequent siblings)
19 siblings, 0 replies; 33+ messages in thread
From: Richard Henderson @ 2024-10-09 0:04 UTC (permalink / raw)
To: qemu-devel
Cc: deller, peter.maydell, alex.bennee, linux-parisc, qemu-arm,
Helge Deller, Philippe Mathieu-Daudé
This function is specific to MemOp, not TCG in general.
Reviewed-by: Helge Deller <deller@gmx.de>
Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org>
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
---
include/exec/memop.h | 23 +++++++++++++++++++++++
include/tcg/tcg.h | 23 -----------------------
2 files changed, 23 insertions(+), 23 deletions(-)
diff --git a/include/exec/memop.h b/include/exec/memop.h
index f881fe7af4..97720a8ee7 100644
--- a/include/exec/memop.h
+++ b/include/exec/memop.h
@@ -170,4 +170,27 @@ static inline bool memop_big_endian(MemOp op)
return (op & MO_BSWAP) == MO_BE;
}
+/**
+ * get_alignment_bits
+ * @memop: MemOp value
+ *
+ * Extract the alignment size from the memop.
+ */
+static inline unsigned get_alignment_bits(MemOp memop)
+{
+ unsigned a = memop & MO_AMASK;
+
+ if (a == MO_UNALN) {
+ /* No alignment required. */
+ a = 0;
+ } else if (a == MO_ALIGN) {
+ /* A natural alignment requirement. */
+ a = memop & MO_SIZE;
+ } else {
+ /* A specific alignment requirement. */
+ a = a >> MO_ASHIFT;
+ }
+ return a;
+}
+
#endif
diff --git a/include/tcg/tcg.h b/include/tcg/tcg.h
index 21d5884741..824fb3560d 100644
--- a/include/tcg/tcg.h
+++ b/include/tcg/tcg.h
@@ -281,29 +281,6 @@ static inline int tcg_type_size(TCGType t)
return 4 << i;
}
-/**
- * get_alignment_bits
- * @memop: MemOp value
- *
- * Extract the alignment size from the memop.
- */
-static inline unsigned get_alignment_bits(MemOp memop)
-{
- unsigned a = memop & MO_AMASK;
-
- if (a == MO_UNALN) {
- /* No alignment required. */
- a = 0;
- } else if (a == MO_ALIGN) {
- /* A natural alignment requirement. */
- a = memop & MO_SIZE;
- } else {
- /* A specific alignment requirement. */
- a = a >> MO_ASHIFT;
- }
- return a;
-}
-
typedef tcg_target_ulong TCGArg;
/* Define type and accessor macros for TCG variables.
--
2.43.0
^ permalink raw reply related [flat|nested] 33+ messages in thread
* [PATCH v3 03/20] include/exec/memop: Rename get_alignment_bits
2024-10-09 0:04 [PATCH v3 00/20] accel/tcg: Introduce tlb_fill_align hook Richard Henderson
2024-10-09 0:04 ` [PATCH v3 01/20] accel/tcg: Assert noreturn from write-only page for atomics Richard Henderson
2024-10-09 0:04 ` [PATCH v3 02/20] include/exec/memop: Move get_alignment_bits from tcg.h Richard Henderson
@ 2024-10-09 0:04 ` Richard Henderson
2024-10-09 0:04 ` [PATCH v3 04/20] include/exec/memop: Introduce memop_atomicity_bits Richard Henderson
` (16 subsequent siblings)
19 siblings, 0 replies; 33+ messages in thread
From: Richard Henderson @ 2024-10-09 0:04 UTC (permalink / raw)
To: qemu-devel
Cc: deller, peter.maydell, alex.bennee, linux-parisc, qemu-arm,
Helge Deller, Philippe Mathieu-Daudé
Rename to use "memop_" prefix, like other functions
that operate on MemOp.
Reviewed-by: Helge Deller <deller@gmx.de>
Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org>
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
---
include/exec/memop.h | 4 ++--
accel/tcg/cputlb.c | 4 ++--
accel/tcg/user-exec.c | 4 ++--
target/arm/tcg/translate-a64.c | 4 ++--
target/xtensa/translate.c | 2 +-
tcg/tcg-op-ldst.c | 6 +++---
tcg/tcg.c | 2 +-
tcg/arm/tcg-target.c.inc | 4 ++--
tcg/sparc64/tcg-target.c.inc | 2 +-
9 files changed, 16 insertions(+), 16 deletions(-)
diff --git a/include/exec/memop.h b/include/exec/memop.h
index 97720a8ee7..f53bf618c6 100644
--- a/include/exec/memop.h
+++ b/include/exec/memop.h
@@ -171,12 +171,12 @@ static inline bool memop_big_endian(MemOp op)
}
/**
- * get_alignment_bits
+ * memop_alignment_bits:
* @memop: MemOp value
*
* Extract the alignment size from the memop.
*/
-static inline unsigned get_alignment_bits(MemOp memop)
+static inline unsigned memop_alignment_bits(MemOp memop)
{
unsigned a = memop & MO_AMASK;
diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c
index fd6459b695..a975fe5f89 100644
--- a/accel/tcg/cputlb.c
+++ b/accel/tcg/cputlb.c
@@ -1709,7 +1709,7 @@ static bool mmu_lookup(CPUState *cpu, vaddr addr, MemOpIdx oi,
tcg_debug_assert(l->mmu_idx < NB_MMU_MODES);
/* Handle CPU specific unaligned behaviour */
- a_bits = get_alignment_bits(l->memop);
+ a_bits = memop_alignment_bits(l->memop);
if (addr & ((1 << a_bits) - 1)) {
cpu_unaligned_access(cpu, addr, type, l->mmu_idx, ra);
}
@@ -1797,7 +1797,7 @@ static void *atomic_mmu_lookup(CPUState *cpu, vaddr addr, MemOpIdx oi,
{
uintptr_t mmu_idx = get_mmuidx(oi);
MemOp mop = get_memop(oi);
- int a_bits = get_alignment_bits(mop);
+ int a_bits = memop_alignment_bits(mop);
uintptr_t index;
CPUTLBEntry *tlbe;
vaddr tlb_addr;
diff --git a/accel/tcg/user-exec.c b/accel/tcg/user-exec.c
index 7ddc47b0ba..08a6df9987 100644
--- a/accel/tcg/user-exec.c
+++ b/accel/tcg/user-exec.c
@@ -959,7 +959,7 @@ void page_reset_target_data(target_ulong start, target_ulong last) { }
static void *cpu_mmu_lookup(CPUState *cpu, vaddr addr,
MemOp mop, uintptr_t ra, MMUAccessType type)
{
- int a_bits = get_alignment_bits(mop);
+ int a_bits = memop_alignment_bits(mop);
void *ret;
/* Enforce guest required alignment. */
@@ -1241,7 +1241,7 @@ static void *atomic_mmu_lookup(CPUState *cpu, vaddr addr, MemOpIdx oi,
int size, uintptr_t retaddr)
{
MemOp mop = get_memop(oi);
- int a_bits = get_alignment_bits(mop);
+ int a_bits = memop_alignment_bits(mop);
void *ret;
/* Enforce guest required alignment. */
diff --git a/target/arm/tcg/translate-a64.c b/target/arm/tcg/translate-a64.c
index 071b6349fc..ec0b1ee252 100644
--- a/target/arm/tcg/translate-a64.c
+++ b/target/arm/tcg/translate-a64.c
@@ -294,7 +294,7 @@ static TCGv_i64 gen_mte_check1_mmuidx(DisasContext *s, TCGv_i64 addr,
desc = FIELD_DP32(desc, MTEDESC, TBI, s->tbid);
desc = FIELD_DP32(desc, MTEDESC, TCMA, s->tcma);
desc = FIELD_DP32(desc, MTEDESC, WRITE, is_write);
- desc = FIELD_DP32(desc, MTEDESC, ALIGN, get_alignment_bits(memop));
+ desc = FIELD_DP32(desc, MTEDESC, ALIGN, memop_alignment_bits(memop));
desc = FIELD_DP32(desc, MTEDESC, SIZEM1, memop_size(memop) - 1);
ret = tcg_temp_new_i64();
@@ -326,7 +326,7 @@ TCGv_i64 gen_mte_checkN(DisasContext *s, TCGv_i64 addr, bool is_write,
desc = FIELD_DP32(desc, MTEDESC, TBI, s->tbid);
desc = FIELD_DP32(desc, MTEDESC, TCMA, s->tcma);
desc = FIELD_DP32(desc, MTEDESC, WRITE, is_write);
- desc = FIELD_DP32(desc, MTEDESC, ALIGN, get_alignment_bits(single_mop));
+ desc = FIELD_DP32(desc, MTEDESC, ALIGN, memop_alignment_bits(single_mop));
desc = FIELD_DP32(desc, MTEDESC, SIZEM1, total_size - 1);
ret = tcg_temp_new_i64();
diff --git a/target/xtensa/translate.c b/target/xtensa/translate.c
index 75b7bfda4c..f4da4a40f9 100644
--- a/target/xtensa/translate.c
+++ b/target/xtensa/translate.c
@@ -521,7 +521,7 @@ static MemOp gen_load_store_alignment(DisasContext *dc, MemOp mop,
mop |= MO_ALIGN;
}
if (!option_enabled(dc, XTENSA_OPTION_UNALIGNED_EXCEPTION)) {
- tcg_gen_andi_i32(addr, addr, ~0 << get_alignment_bits(mop));
+ tcg_gen_andi_i32(addr, addr, ~0 << memop_alignment_bits(mop));
}
return mop;
}
diff --git a/tcg/tcg-op-ldst.c b/tcg/tcg-op-ldst.c
index 23dc807f11..a318011229 100644
--- a/tcg/tcg-op-ldst.c
+++ b/tcg/tcg-op-ldst.c
@@ -45,7 +45,7 @@ static void check_max_alignment(unsigned a_bits)
static MemOp tcg_canonicalize_memop(MemOp op, bool is64, bool st)
{
- unsigned a_bits = get_alignment_bits(op);
+ unsigned a_bits = memop_alignment_bits(op);
check_max_alignment(a_bits);
@@ -559,7 +559,7 @@ static void tcg_gen_qemu_ld_i128_int(TCGv_i128 val, TCGTemp *addr,
TCGv_i64 ext_addr = NULL;
TCGOpcode opc;
- check_max_alignment(get_alignment_bits(memop));
+ check_max_alignment(memop_alignment_bits(memop));
tcg_gen_req_mo(TCG_MO_LD_LD | TCG_MO_ST_LD);
/* In serial mode, reduce atomicity. */
@@ -676,7 +676,7 @@ static void tcg_gen_qemu_st_i128_int(TCGv_i128 val, TCGTemp *addr,
TCGv_i64 ext_addr = NULL;
TCGOpcode opc;
- check_max_alignment(get_alignment_bits(memop));
+ check_max_alignment(memop_alignment_bits(memop));
tcg_gen_req_mo(TCG_MO_ST_LD | TCG_MO_ST_ST);
/* In serial mode, reduce atomicity. */
diff --git a/tcg/tcg.c b/tcg/tcg.c
index 34e3056380..5decd83cf4 100644
--- a/tcg/tcg.c
+++ b/tcg/tcg.c
@@ -5506,7 +5506,7 @@ static void tcg_reg_alloc_call(TCGContext *s, TCGOp *op)
static TCGAtomAlign atom_and_align_for_opc(TCGContext *s, MemOp opc,
MemOp host_atom, bool allow_two_ops)
{
- MemOp align = get_alignment_bits(opc);
+ MemOp align = memop_alignment_bits(opc);
MemOp size = opc & MO_SIZE;
MemOp half = size ? size - 1 : 0;
MemOp atom = opc & MO_ATOM_MASK;
diff --git a/tcg/arm/tcg-target.c.inc b/tcg/arm/tcg-target.c.inc
index 3de5f50b62..56072d89a2 100644
--- a/tcg/arm/tcg-target.c.inc
+++ b/tcg/arm/tcg-target.c.inc
@@ -1587,7 +1587,7 @@ static void tcg_out_qemu_ld_direct(TCGContext *s, MemOp opc, TCGReg datalo,
tcg_debug_assert((datalo & 1) == 0);
tcg_debug_assert(datahi == datalo + 1);
/* LDRD requires alignment; double-check that. */
- if (get_alignment_bits(opc) >= MO_64) {
+ if (memop_alignment_bits(opc) >= MO_64) {
if (h.index < 0) {
tcg_out_ldrd_8(s, h.cond, datalo, h.base, 0);
break;
@@ -1691,7 +1691,7 @@ static void tcg_out_qemu_st_direct(TCGContext *s, MemOp opc, TCGReg datalo,
tcg_debug_assert((datalo & 1) == 0);
tcg_debug_assert(datahi == datalo + 1);
/* STRD requires alignment; double-check that. */
- if (get_alignment_bits(opc) >= MO_64) {
+ if (memop_alignment_bits(opc) >= MO_64) {
if (h.index < 0) {
tcg_out_strd_8(s, h.cond, datalo, h.base, 0);
} else {
diff --git a/tcg/sparc64/tcg-target.c.inc b/tcg/sparc64/tcg-target.c.inc
index 176c98740b..32f9ec24b5 100644
--- a/tcg/sparc64/tcg-target.c.inc
+++ b/tcg/sparc64/tcg-target.c.inc
@@ -1133,7 +1133,7 @@ static TCGLabelQemuLdst *prepare_host_addr(TCGContext *s, HostAddress *h,
* Otherwise, test for at least natural alignment and defer
* everything else to the helper functions.
*/
- if (s_bits != get_alignment_bits(opc)) {
+ if (s_bits != memop_alignment_bits(opc)) {
tcg_debug_assert(check_fit_tl(a_mask, 13));
tcg_out_arithi(s, TCG_REG_G0, addr_reg, a_mask, ARITH_ANDCC);
--
2.43.0
^ permalink raw reply related [flat|nested] 33+ messages in thread
* [PATCH v3 04/20] include/exec/memop: Introduce memop_atomicity_bits
2024-10-09 0:04 [PATCH v3 00/20] accel/tcg: Introduce tlb_fill_align hook Richard Henderson
` (2 preceding siblings ...)
2024-10-09 0:04 ` [PATCH v3 03/20] include/exec/memop: Rename get_alignment_bits Richard Henderson
@ 2024-10-09 0:04 ` Richard Henderson
2024-10-09 18:55 ` Philippe Mathieu-Daudé
2024-10-09 0:04 ` [PATCH v3 05/20] accel/tcg: Add TCGCPUOps.tlb_fill_align Richard Henderson
` (15 subsequent siblings)
19 siblings, 1 reply; 33+ messages in thread
From: Richard Henderson @ 2024-10-09 0:04 UTC (permalink / raw)
To: qemu-devel
Cc: deller, peter.maydell, alex.bennee, linux-parisc, qemu-arm,
Helge Deller
Split out of mmu_lookup.
Reviewed-by: Helge Deller <deller@gmx.de>
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
---
include/exec/memop.h | 24 ++++++++++++++++++++++++
accel/tcg/cputlb.c | 16 ++--------------
2 files changed, 26 insertions(+), 14 deletions(-)
diff --git a/include/exec/memop.h b/include/exec/memop.h
index f53bf618c6..b699bf7688 100644
--- a/include/exec/memop.h
+++ b/include/exec/memop.h
@@ -193,4 +193,28 @@ static inline unsigned memop_alignment_bits(MemOp memop)
return a;
}
+/*
+ * memop_atomicity_bits:
+ * @memop: MemOp value
+ *
+ * Extract the atomicity size from the memop.
+ */
+static inline unsigned memop_atomicity_bits(MemOp memop)
+{
+ unsigned size = memop & MO_SIZE;
+
+ switch (memop & MO_ATOM_MASK) {
+ case MO_ATOM_NONE:
+ size = MO_8;
+ break;
+ case MO_ATOM_IFALIGN_PAIR:
+ case MO_ATOM_WITHIN16_PAIR:
+ size = size ? size - 1 : 0;
+ break;
+ default:
+ break;
+ }
+ return size;
+}
+
#endif
diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c
index a975fe5f89..35cda1e2b0 100644
--- a/accel/tcg/cputlb.c
+++ b/accel/tcg/cputlb.c
@@ -1767,20 +1767,8 @@ static bool mmu_lookup(CPUState *cpu, vaddr addr, MemOpIdx oi,
* Device memory type require alignment.
*/
if (unlikely(flags & TLB_CHECK_ALIGNED)) {
- MemOp size = l->memop & MO_SIZE;
-
- switch (l->memop & MO_ATOM_MASK) {
- case MO_ATOM_NONE:
- size = MO_8;
- break;
- case MO_ATOM_IFALIGN_PAIR:
- case MO_ATOM_WITHIN16_PAIR:
- size = size ? size - 1 : 0;
- break;
- default:
- break;
- }
- if (addr & ((1 << size) - 1)) {
+ a_bits = memop_atomicity_bits(l->memop);
+ if (addr & ((1 << a_bits) - 1)) {
cpu_unaligned_access(cpu, addr, type, l->mmu_idx, ra);
}
}
--
2.43.0
^ permalink raw reply related [flat|nested] 33+ messages in thread
* [PATCH v3 05/20] accel/tcg: Add TCGCPUOps.tlb_fill_align
2024-10-09 0:04 [PATCH v3 00/20] accel/tcg: Introduce tlb_fill_align hook Richard Henderson
` (3 preceding siblings ...)
2024-10-09 0:04 ` [PATCH v3 04/20] include/exec/memop: Introduce memop_atomicity_bits Richard Henderson
@ 2024-10-09 0:04 ` Richard Henderson
2024-10-09 19:46 ` Helge Deller
2024-10-10 10:31 ` Peter Maydell
2024-10-09 0:04 ` [PATCH v3 06/20] accel/tcg: Use the alignment test in tlb_fill_align Richard Henderson
` (14 subsequent siblings)
19 siblings, 2 replies; 33+ messages in thread
From: Richard Henderson @ 2024-10-09 0:04 UTC (permalink / raw)
To: qemu-devel; +Cc: deller, peter.maydell, alex.bennee, linux-parisc, qemu-arm
Add a new callback to handle softmmu paging. Return the page
details directly, instead of passing them indirectly to
tlb_set_page. Handle alignment simultaneously with paging so
that faults are handled with target-specific priority.
Route all calls the two hooks through a tlb_fill_align
function local to cputlb.c.
As yet no targets implement the new hook.
As yet cputlb.c does not use the new alignment check.
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
---
include/hw/core/cpu.h | 4 +--
include/hw/core/tcg-cpu-ops.h | 26 +++++++++++++++
include/qemu/typedefs.h | 1 +
accel/tcg/cputlb.c | 61 ++++++++++++++++++++++-------------
4 files changed, 67 insertions(+), 25 deletions(-)
diff --git a/include/hw/core/cpu.h b/include/hw/core/cpu.h
index 04e9ad4996..d21a24c82f 100644
--- a/include/hw/core/cpu.h
+++ b/include/hw/core/cpu.h
@@ -205,7 +205,7 @@ struct CPUClass {
* so the layout is not as critical as that of CPUTLBEntry. This is
* also why we don't want to combine the two structs.
*/
-typedef struct CPUTLBEntryFull {
+struct CPUTLBEntryFull {
/*
* @xlat_section contains:
* - in the lower TARGET_PAGE_BITS, a physical section number
@@ -261,7 +261,7 @@ typedef struct CPUTLBEntryFull {
bool guarded;
} arm;
} extra;
-} CPUTLBEntryFull;
+};
/*
* Data elements that are per MMU mode, minus the bits accessed by
diff --git a/include/hw/core/tcg-cpu-ops.h b/include/hw/core/tcg-cpu-ops.h
index 34318cf0e6..c932690621 100644
--- a/include/hw/core/tcg-cpu-ops.h
+++ b/include/hw/core/tcg-cpu-ops.h
@@ -13,6 +13,7 @@
#include "exec/breakpoint.h"
#include "exec/hwaddr.h"
#include "exec/memattrs.h"
+#include "exec/memop.h"
#include "exec/mmu-access-type.h"
#include "exec/vaddr.h"
@@ -131,6 +132,31 @@ struct TCGCPUOps {
* same function signature.
*/
bool (*cpu_exec_halt)(CPUState *cpu);
+ /**
+ * @tlb_fill_align: Handle a softmmu tlb miss
+ * @cpu: cpu context
+ * @out: output page properties
+ * @addr: virtual address
+ * @access_type: read, write or execute
+ * @mmu_idx: mmu context
+ * @memop: memory operation for the access
+ * @size: memory access size, or 0 for whole page
+ * @probe: test only, no fault
+ * @ra: host return address for exception unwind
+ *
+ * If the access is valid, fill in @out and return true.
+ * Otherwise if probe is true, return false.
+ * Otherwise raise and exception and do not return.
+ *
+ * The alignment check for the access is deferred to this hook,
+ * so that the target can determine the priority of any alignment
+ * fault with respect to other potential faults from paging.
+ * Zero may be passed for @memop to skip any alignment check
+ * for non-memory-access operations such as probing.
+ */
+ bool (*tlb_fill_align)(CPUState *cpu, CPUTLBEntryFull *out, vaddr addr,
+ MMUAccessType access_type, int mmu_idx,
+ MemOp memop, int size, bool probe, uintptr_t ra);
/**
* @tlb_fill: Handle a softmmu tlb miss
*
diff --git a/include/qemu/typedefs.h b/include/qemu/typedefs.h
index 9d222dc376..3d84efcac4 100644
--- a/include/qemu/typedefs.h
+++ b/include/qemu/typedefs.h
@@ -40,6 +40,7 @@ typedef struct ConfidentialGuestSupport ConfidentialGuestSupport;
typedef struct CPUArchState CPUArchState;
typedef struct CPUPluginState CPUPluginState;
typedef struct CPUState CPUState;
+typedef struct CPUTLBEntryFull CPUTLBEntryFull;
typedef struct DeviceState DeviceState;
typedef struct DirtyBitmapSnapshot DirtyBitmapSnapshot;
typedef struct DisasContextBase DisasContextBase;
diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c
index 35cda1e2b0..d72f454e9e 100644
--- a/accel/tcg/cputlb.c
+++ b/accel/tcg/cputlb.c
@@ -1221,22 +1221,35 @@ void tlb_set_page(CPUState *cpu, vaddr addr,
}
/*
- * Note: tlb_fill() can trigger a resize of the TLB. This means that all of the
- * caller's prior references to the TLB table (e.g. CPUTLBEntry pointers) must
- * be discarded and looked up again (e.g. via tlb_entry()).
+ * Note: tlb_fill_align() can trigger a resize of the TLB.
+ * This means that all of the caller's prior references to the TLB table
+ * (e.g. CPUTLBEntry pointers) must be discarded and looked up again
+ * (e.g. via tlb_entry()).
*/
-static void tlb_fill(CPUState *cpu, vaddr addr, int size,
- MMUAccessType access_type, int mmu_idx, uintptr_t retaddr)
+static bool tlb_fill_align(CPUState *cpu, vaddr addr, MMUAccessType type,
+ int mmu_idx, MemOp memop, int size,
+ bool probe, uintptr_t ra)
{
- bool ok;
+ const TCGCPUOps *ops = cpu->cc->tcg_ops;
+ CPUTLBEntryFull full;
- /*
- * This is not a probe, so only valid return is success; failure
- * should result in exception + longjmp to the cpu loop.
- */
- ok = cpu->cc->tcg_ops->tlb_fill(cpu, addr, size,
- access_type, mmu_idx, false, retaddr);
- assert(ok);
+ if (ops->tlb_fill_align) {
+ if (ops->tlb_fill_align(cpu, &full, addr, type, mmu_idx,
+ memop, size, probe, ra)) {
+ tlb_set_page_full(cpu, mmu_idx, addr, &full);
+ return true;
+ }
+ } else {
+ /* Legacy behaviour is alignment before paging. */
+ if (addr & ((1u << memop_alignment_bits(memop)) - 1)) {
+ ops->do_unaligned_access(cpu, addr, type, mmu_idx, ra);
+ }
+ if (ops->tlb_fill(cpu, addr, size, type, mmu_idx, probe, ra)) {
+ return true;
+ }
+ }
+ assert(probe);
+ return false;
}
static inline void cpu_unaligned_access(CPUState *cpu, vaddr addr,
@@ -1351,22 +1364,22 @@ static int probe_access_internal(CPUState *cpu, vaddr addr,
if (!tlb_hit_page(tlb_addr, page_addr)) {
if (!victim_tlb_hit(cpu, mmu_idx, index, access_type, page_addr)) {
- if (!cpu->cc->tcg_ops->tlb_fill(cpu, addr, fault_size, access_type,
- mmu_idx, nonfault, retaddr)) {
+ if (!tlb_fill_align(cpu, addr, access_type, mmu_idx,
+ 0, fault_size, nonfault, retaddr)) {
/* Non-faulting page table read failed. */
*phost = NULL;
*pfull = NULL;
return TLB_INVALID_MASK;
}
- /* TLB resize via tlb_fill may have moved the entry. */
+ /* TLB resize via tlb_fill_align may have moved the entry. */
index = tlb_index(cpu, mmu_idx, addr);
entry = tlb_entry(cpu, mmu_idx, addr);
/*
* With PAGE_WRITE_INV, we set TLB_INVALID_MASK immediately,
- * to force the next access through tlb_fill. We've just
- * called tlb_fill, so we know that this entry *is* valid.
+ * to force the next access through tlb_fill_align. We've just
+ * called tlb_fill_align, so we know that this entry *is* valid.
*/
flags &= ~TLB_INVALID_MASK;
}
@@ -1613,7 +1626,7 @@ typedef struct MMULookupLocals {
*
* Resolve the translation for the one page at @data.addr, filling in
* the rest of @data with the results. If the translation fails,
- * tlb_fill will longjmp out. Return true if the softmmu tlb for
+ * tlb_fill_align will longjmp out. Return true if the softmmu tlb for
* @mmu_idx may have resized.
*/
static bool mmu_lookup1(CPUState *cpu, MMULookupPageData *data,
@@ -1631,7 +1644,8 @@ static bool mmu_lookup1(CPUState *cpu, MMULookupPageData *data,
if (!tlb_hit(tlb_addr, addr)) {
if (!victim_tlb_hit(cpu, mmu_idx, index, access_type,
addr & TARGET_PAGE_MASK)) {
- tlb_fill(cpu, addr, data->size, access_type, mmu_idx, ra);
+ tlb_fill_align(cpu, addr, access_type, mmu_idx,
+ 0, data->size, false, ra);
maybe_resized = true;
index = tlb_index(cpu, mmu_idx, addr);
entry = tlb_entry(cpu, mmu_idx, addr);
@@ -1821,8 +1835,8 @@ static void *atomic_mmu_lookup(CPUState *cpu, vaddr addr, MemOpIdx oi,
if (!tlb_hit(tlb_addr, addr)) {
if (!victim_tlb_hit(cpu, mmu_idx, index, MMU_DATA_STORE,
addr & TARGET_PAGE_MASK)) {
- tlb_fill(cpu, addr, size,
- MMU_DATA_STORE, mmu_idx, retaddr);
+ tlb_fill_align(cpu, addr, MMU_DATA_STORE, mmu_idx,
+ 0, size, false, retaddr);
index = tlb_index(cpu, mmu_idx, addr);
tlbe = tlb_entry(cpu, mmu_idx, addr);
}
@@ -1836,7 +1850,8 @@ static void *atomic_mmu_lookup(CPUState *cpu, vaddr addr, MemOpIdx oi,
* but addr_read will only be -1 if PAGE_READ was unset.
*/
if (unlikely(tlbe->addr_read == -1)) {
- tlb_fill(cpu, addr, size, MMU_DATA_LOAD, mmu_idx, retaddr);
+ tlb_fill_align(cpu, addr, MMU_DATA_LOAD, mmu_idx,
+ 0, size, false, retaddr);
/*
* Since we don't support reads and writes to different
* addresses, and we do have the proper page loaded for
--
2.43.0
^ permalink raw reply related [flat|nested] 33+ messages in thread
* [PATCH v3 06/20] accel/tcg: Use the alignment test in tlb_fill_align
2024-10-09 0:04 [PATCH v3 00/20] accel/tcg: Introduce tlb_fill_align hook Richard Henderson
` (4 preceding siblings ...)
2024-10-09 0:04 ` [PATCH v3 05/20] accel/tcg: Add TCGCPUOps.tlb_fill_align Richard Henderson
@ 2024-10-09 0:04 ` Richard Henderson
2024-10-09 21:29 ` Helge Deller
2024-10-10 10:44 ` Peter Maydell
2024-10-09 0:04 ` [PATCH v3 07/20] target/hppa: Add MemOp argument to hppa_get_physical_address Richard Henderson
` (13 subsequent siblings)
19 siblings, 2 replies; 33+ messages in thread
From: Richard Henderson @ 2024-10-09 0:04 UTC (permalink / raw)
To: qemu-devel; +Cc: deller, peter.maydell, alex.bennee, linux-parisc, qemu-arm
When we have a tlb miss, defer the alignment check to
the new tlb_fill_align hook. Move the existing alignment
check so that we only perform it with a tlb hit.
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
---
accel/tcg/cputlb.c | 88 ++++++++++++++++++++++++----------------------
1 file changed, 45 insertions(+), 43 deletions(-)
diff --git a/accel/tcg/cputlb.c b/accel/tcg/cputlb.c
index d72f454e9e..b76a4eac4e 100644
--- a/accel/tcg/cputlb.c
+++ b/accel/tcg/cputlb.c
@@ -1620,6 +1620,7 @@ typedef struct MMULookupLocals {
* mmu_lookup1: translate one page
* @cpu: generic cpu state
* @data: lookup parameters
+ * @memop: memory operation for the access, or 0
* @mmu_idx: virtual address context
* @access_type: load/store/code
* @ra: return address into tcg generated code, or 0
@@ -1629,7 +1630,7 @@ typedef struct MMULookupLocals {
* tlb_fill_align will longjmp out. Return true if the softmmu tlb for
* @mmu_idx may have resized.
*/
-static bool mmu_lookup1(CPUState *cpu, MMULookupPageData *data,
+static bool mmu_lookup1(CPUState *cpu, MMULookupPageData *data, MemOp memop,
int mmu_idx, MMUAccessType access_type, uintptr_t ra)
{
vaddr addr = data->addr;
@@ -1645,7 +1646,7 @@ static bool mmu_lookup1(CPUState *cpu, MMULookupPageData *data,
if (!victim_tlb_hit(cpu, mmu_idx, index, access_type,
addr & TARGET_PAGE_MASK)) {
tlb_fill_align(cpu, addr, access_type, mmu_idx,
- 0, data->size, false, ra);
+ memop, data->size, false, ra);
maybe_resized = true;
index = tlb_index(cpu, mmu_idx, addr);
entry = tlb_entry(cpu, mmu_idx, addr);
@@ -1657,6 +1658,25 @@ static bool mmu_lookup1(CPUState *cpu, MMULookupPageData *data,
flags = tlb_addr & (TLB_FLAGS_MASK & ~TLB_FORCE_SLOW);
flags |= full->slow_flags[access_type];
+ if (likely(!maybe_resized)) {
+ /* Alignment has not been checked by tlb_fill_align. */
+ int a_bits = memop_alignment_bits(memop);
+
+ /*
+ * This alignment check differs from the one above, in that this is
+ * based on the atomicity of the operation. The intended use case is
+ * the ARM memory type field of each PTE, where access to pages with
+ * Device memory type require alignment.
+ */
+ if (unlikely(flags & TLB_CHECK_ALIGNED)) {
+ int at_bits = memop_atomicity_bits(memop);
+ a_bits = MAX(a_bits, at_bits);
+ }
+ if (unlikely(addr & ((1 << a_bits) - 1))) {
+ cpu_unaligned_access(cpu, addr, access_type, mmu_idx, ra);
+ }
+ }
+
data->full = full;
data->flags = flags;
/* Compute haddr speculatively; depending on flags it might be invalid. */
@@ -1713,7 +1733,6 @@ static void mmu_watch_or_dirty(CPUState *cpu, MMULookupPageData *data,
static bool mmu_lookup(CPUState *cpu, vaddr addr, MemOpIdx oi,
uintptr_t ra, MMUAccessType type, MMULookupLocals *l)
{
- unsigned a_bits;
bool crosspage;
int flags;
@@ -1722,12 +1741,6 @@ static bool mmu_lookup(CPUState *cpu, vaddr addr, MemOpIdx oi,
tcg_debug_assert(l->mmu_idx < NB_MMU_MODES);
- /* Handle CPU specific unaligned behaviour */
- a_bits = memop_alignment_bits(l->memop);
- if (addr & ((1 << a_bits) - 1)) {
- cpu_unaligned_access(cpu, addr, type, l->mmu_idx, ra);
- }
-
l->page[0].addr = addr;
l->page[0].size = memop_size(l->memop);
l->page[1].addr = (addr + l->page[0].size - 1) & TARGET_PAGE_MASK;
@@ -1735,7 +1748,7 @@ static bool mmu_lookup(CPUState *cpu, vaddr addr, MemOpIdx oi,
crosspage = (addr ^ l->page[1].addr) & TARGET_PAGE_MASK;
if (likely(!crosspage)) {
- mmu_lookup1(cpu, &l->page[0], l->mmu_idx, type, ra);
+ mmu_lookup1(cpu, &l->page[0], l->memop, l->mmu_idx, type, ra);
flags = l->page[0].flags;
if (unlikely(flags & (TLB_WATCHPOINT | TLB_NOTDIRTY))) {
@@ -1754,8 +1767,8 @@ static bool mmu_lookup(CPUState *cpu, vaddr addr, MemOpIdx oi,
* Lookup both pages, recognizing exceptions from either. If the
* second lookup potentially resized, refresh first CPUTLBEntryFull.
*/
- mmu_lookup1(cpu, &l->page[0], l->mmu_idx, type, ra);
- if (mmu_lookup1(cpu, &l->page[1], l->mmu_idx, type, ra)) {
+ mmu_lookup1(cpu, &l->page[0], l->memop, l->mmu_idx, type, ra);
+ if (mmu_lookup1(cpu, &l->page[1], 0, l->mmu_idx, type, ra)) {
uintptr_t index = tlb_index(cpu, l->mmu_idx, addr);
l->page[0].full = &cpu->neg.tlb.d[l->mmu_idx].fulltlb[index];
}
@@ -1774,19 +1787,6 @@ static bool mmu_lookup(CPUState *cpu, vaddr addr, MemOpIdx oi,
tcg_debug_assert((flags & TLB_BSWAP) == 0);
}
- /*
- * This alignment check differs from the one above, in that this is
- * based on the atomicity of the operation. The intended use case is
- * the ARM memory type field of each PTE, where access to pages with
- * Device memory type require alignment.
- */
- if (unlikely(flags & TLB_CHECK_ALIGNED)) {
- a_bits = memop_atomicity_bits(l->memop);
- if (addr & ((1 << a_bits) - 1)) {
- cpu_unaligned_access(cpu, addr, type, l->mmu_idx, ra);
- }
- }
-
return crosspage;
}
@@ -1799,34 +1799,18 @@ static void *atomic_mmu_lookup(CPUState *cpu, vaddr addr, MemOpIdx oi,
{
uintptr_t mmu_idx = get_mmuidx(oi);
MemOp mop = get_memop(oi);
- int a_bits = memop_alignment_bits(mop);
uintptr_t index;
CPUTLBEntry *tlbe;
vaddr tlb_addr;
void *hostaddr;
CPUTLBEntryFull *full;
+ bool did_tlb_fill = false;
tcg_debug_assert(mmu_idx < NB_MMU_MODES);
/* Adjust the given return address. */
retaddr -= GETPC_ADJ;
- /* Enforce guest required alignment. */
- if (unlikely(a_bits > 0 && (addr & ((1 << a_bits) - 1)))) {
- /* ??? Maybe indicate atomic op to cpu_unaligned_access */
- cpu_unaligned_access(cpu, addr, MMU_DATA_STORE,
- mmu_idx, retaddr);
- }
-
- /* Enforce qemu required alignment. */
- if (unlikely(addr & (size - 1))) {
- /* We get here if guest alignment was not requested,
- or was not enforced by cpu_unaligned_access above.
- We might widen the access and emulate, but for now
- mark an exception and exit the cpu loop. */
- goto stop_the_world;
- }
-
index = tlb_index(cpu, mmu_idx, addr);
tlbe = tlb_entry(cpu, mmu_idx, addr);
@@ -1836,7 +1820,8 @@ static void *atomic_mmu_lookup(CPUState *cpu, vaddr addr, MemOpIdx oi,
if (!victim_tlb_hit(cpu, mmu_idx, index, MMU_DATA_STORE,
addr & TARGET_PAGE_MASK)) {
tlb_fill_align(cpu, addr, MMU_DATA_STORE, mmu_idx,
- 0, size, false, retaddr);
+ mop, size, false, retaddr);
+ did_tlb_fill = true;
index = tlb_index(cpu, mmu_idx, addr);
tlbe = tlb_entry(cpu, mmu_idx, addr);
}
@@ -1859,6 +1844,23 @@ static void *atomic_mmu_lookup(CPUState *cpu, vaddr addr, MemOpIdx oi,
*/
g_assert_not_reached();
}
+
+ /* Enforce guest required alignment, if not handled by tlb_fill_align. */
+ if (!did_tlb_fill && (addr & ((1 << memop_alignment_bits(mop)) - 1))) {
+ cpu_unaligned_access(cpu, addr, MMU_DATA_STORE, mmu_idx, retaddr);
+ }
+
+ /* Enforce qemu required alignment. */
+ if (unlikely(addr & (size - 1))) {
+ /*
+ * We get here if guest alignment was not requested, or was not
+ * enforced by cpu_unaligned_access or tlb_fill_align above.
+ * We might widen the access and emulate, but for now
+ * mark an exception and exit the cpu loop.
+ */
+ goto stop_the_world;
+ }
+
/* Collect tlb flags for read. */
tlb_addr |= tlbe->addr_read;
--
2.43.0
^ permalink raw reply related [flat|nested] 33+ messages in thread
* [PATCH v3 07/20] target/hppa: Add MemOp argument to hppa_get_physical_address
2024-10-09 0:04 [PATCH v3 00/20] accel/tcg: Introduce tlb_fill_align hook Richard Henderson
` (5 preceding siblings ...)
2024-10-09 0:04 ` [PATCH v3 06/20] accel/tcg: Use the alignment test in tlb_fill_align Richard Henderson
@ 2024-10-09 0:04 ` Richard Henderson
2024-10-09 0:04 ` [PATCH v3 08/20] target/hppa: Perform access rights before protection id check Richard Henderson
` (12 subsequent siblings)
19 siblings, 0 replies; 33+ messages in thread
From: Richard Henderson @ 2024-10-09 0:04 UTC (permalink / raw)
To: qemu-devel
Cc: deller, peter.maydell, alex.bennee, linux-parisc, qemu-arm,
Helge Deller, Philippe Mathieu-Daudé
Just add the argument, unused at this point.
Zero is the safe do-nothing value for all callers.
Reviewed-by: Helge Deller <deller@gmx.de>
Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org>
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
---
target/hppa/cpu.h | 2 +-
target/hppa/int_helper.c | 2 +-
target/hppa/mem_helper.c | 9 +++++----
target/hppa/op_helper.c | 2 +-
4 files changed, 8 insertions(+), 7 deletions(-)
diff --git a/target/hppa/cpu.h b/target/hppa/cpu.h
index f4e051f176..526855f982 100644
--- a/target/hppa/cpu.h
+++ b/target/hppa/cpu.h
@@ -369,7 +369,7 @@ bool hppa_cpu_tlb_fill(CPUState *cs, vaddr address, int size,
void hppa_cpu_do_interrupt(CPUState *cpu);
bool hppa_cpu_exec_interrupt(CPUState *cpu, int int_req);
int hppa_get_physical_address(CPUHPPAState *env, vaddr addr, int mmu_idx,
- int type, hwaddr *pphys, int *pprot);
+ int type, MemOp mop, hwaddr *pphys, int *pprot);
void hppa_cpu_do_transaction_failed(CPUState *cs, hwaddr physaddr,
vaddr addr, unsigned size,
MMUAccessType access_type,
diff --git a/target/hppa/int_helper.c b/target/hppa/int_helper.c
index 391f32f27d..58695def82 100644
--- a/target/hppa/int_helper.c
+++ b/target/hppa/int_helper.c
@@ -167,7 +167,7 @@ void hppa_cpu_do_interrupt(CPUState *cs)
vaddr = hppa_form_gva_psw(old_psw, env->iasq_f, vaddr);
t = hppa_get_physical_address(env, vaddr, MMU_KERNEL_IDX,
- 0, &paddr, &prot);
+ 0, 0, &paddr, &prot);
if (t >= 0) {
/* We can't re-load the instruction. */
env->cr[CR_IIR] = 0;
diff --git a/target/hppa/mem_helper.c b/target/hppa/mem_helper.c
index b984f730aa..a386c80fa4 100644
--- a/target/hppa/mem_helper.c
+++ b/target/hppa/mem_helper.c
@@ -197,7 +197,7 @@ static int match_prot_id64(CPUHPPAState *env, uint32_t access_id)
}
int hppa_get_physical_address(CPUHPPAState *env, vaddr addr, int mmu_idx,
- int type, hwaddr *pphys, int *pprot)
+ int type, MemOp mop, hwaddr *pphys, int *pprot)
{
hwaddr phys;
int prot, r_prot, w_prot, x_prot, priv;
@@ -340,7 +340,7 @@ hwaddr hppa_cpu_get_phys_page_debug(CPUState *cs, vaddr addr)
mmu_idx = (cpu->env.psw & PSW_D ? MMU_KERNEL_IDX :
cpu->env.psw & PSW_W ? MMU_ABS_W_IDX : MMU_ABS_IDX);
- excp = hppa_get_physical_address(&cpu->env, addr, mmu_idx, 0,
+ excp = hppa_get_physical_address(&cpu->env, addr, mmu_idx, 0, 0,
&phys, &prot);
/* Since we're translating for debugging, the only error that is a
@@ -438,7 +438,8 @@ bool hppa_cpu_tlb_fill(CPUState *cs, vaddr addr, int size,
break;
}
- excp = hppa_get_physical_address(env, addr, mmu_idx, a_prot, &phys, &prot);
+ excp = hppa_get_physical_address(env, addr, mmu_idx, a_prot, 0,
+ &phys, &prot);
if (unlikely(excp >= 0)) {
if (probe) {
return false;
@@ -678,7 +679,7 @@ target_ulong HELPER(lpa)(CPUHPPAState *env, target_ulong addr)
hwaddr phys;
int prot, excp;
- excp = hppa_get_physical_address(env, addr, MMU_KERNEL_IDX, 0,
+ excp = hppa_get_physical_address(env, addr, MMU_KERNEL_IDX, 0, 0,
&phys, &prot);
if (excp >= 0) {
if (excp == EXCP_DTLB_MISS) {
diff --git a/target/hppa/op_helper.c b/target/hppa/op_helper.c
index 7f79196fff..744325969f 100644
--- a/target/hppa/op_helper.c
+++ b/target/hppa/op_helper.c
@@ -334,7 +334,7 @@ target_ulong HELPER(probe)(CPUHPPAState *env, target_ulong addr,
}
mmu_idx = PRIV_P_TO_MMU_IDX(level, env->psw & PSW_P);
- excp = hppa_get_physical_address(env, addr, mmu_idx, 0, &phys, &prot);
+ excp = hppa_get_physical_address(env, addr, mmu_idx, 0, 0, &phys, &prot);
if (excp >= 0) {
cpu_restore_state(env_cpu(env), GETPC());
hppa_set_ior_and_isr(env, addr, MMU_IDX_MMU_DISABLED(mmu_idx));
--
2.43.0
^ permalink raw reply related [flat|nested] 33+ messages in thread
* [PATCH v3 08/20] target/hppa: Perform access rights before protection id check
2024-10-09 0:04 [PATCH v3 00/20] accel/tcg: Introduce tlb_fill_align hook Richard Henderson
` (6 preceding siblings ...)
2024-10-09 0:04 ` [PATCH v3 07/20] target/hppa: Add MemOp argument to hppa_get_physical_address Richard Henderson
@ 2024-10-09 0:04 ` Richard Henderson
2024-10-09 0:04 ` [PATCH v3 09/20] target/hppa: Fix priority of T, D, and B page faults Richard Henderson
` (11 subsequent siblings)
19 siblings, 0 replies; 33+ messages in thread
From: Richard Henderson @ 2024-10-09 0:04 UTC (permalink / raw)
To: qemu-devel
Cc: deller, peter.maydell, alex.bennee, linux-parisc, qemu-arm,
Helge Deller
In Chapter 5, Interruptions, the group 3 exceptions lists
"Data memory access rights trap" in priority order ahead of
"Data memory protection ID trap".
Swap these checks in hppa_get_physical_address.
Reviewed-by: Helge Deller <deller@gmx.de>
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
---
target/hppa/mem_helper.c | 12 ++++++------
1 file changed, 6 insertions(+), 6 deletions(-)
diff --git a/target/hppa/mem_helper.c b/target/hppa/mem_helper.c
index a386c80fa4..f027c494e2 100644
--- a/target/hppa/mem_helper.c
+++ b/target/hppa/mem_helper.c
@@ -267,6 +267,12 @@ int hppa_get_physical_address(CPUHPPAState *env, vaddr addr, int mmu_idx,
goto egress;
}
+ if (unlikely(!(prot & type))) {
+ /* Not allowed -- Inst/Data Memory Access Rights Fault. */
+ ret = (type & PAGE_EXEC) ? EXCP_IMP : EXCP_DMAR;
+ goto egress;
+ }
+
/* access_id == 0 means public page and no check is performed */
if (ent->access_id && MMU_IDX_TO_P(mmu_idx)) {
int access_prot = (hppa_is_pa20(env)
@@ -281,12 +287,6 @@ int hppa_get_physical_address(CPUHPPAState *env, vaddr addr, int mmu_idx,
prot &= access_prot;
}
- if (unlikely(!(prot & type))) {
- /* Not allowed -- Inst/Data Memory Access Rights Fault. */
- ret = (type & PAGE_EXEC) ? EXCP_IMP : EXCP_DMAR;
- goto egress;
- }
-
/*
* In priority order, check for conditions which raise faults.
* Remove PROT bits that cover the condition we want to check,
--
2.43.0
^ permalink raw reply related [flat|nested] 33+ messages in thread
* [PATCH v3 09/20] target/hppa: Fix priority of T, D, and B page faults
2024-10-09 0:04 [PATCH v3 00/20] accel/tcg: Introduce tlb_fill_align hook Richard Henderson
` (7 preceding siblings ...)
2024-10-09 0:04 ` [PATCH v3 08/20] target/hppa: Perform access rights before protection id check Richard Henderson
@ 2024-10-09 0:04 ` Richard Henderson
2024-10-14 18:02 ` Michael Tokarev
2024-10-09 0:04 ` [PATCH v3 10/20] target/hppa: Handle alignment faults in hppa_get_physical_address Richard Henderson
` (10 subsequent siblings)
19 siblings, 1 reply; 33+ messages in thread
From: Richard Henderson @ 2024-10-09 0:04 UTC (permalink / raw)
To: qemu-devel
Cc: deller, peter.maydell, alex.bennee, linux-parisc, qemu-arm,
Helge Deller, Philippe Mathieu-Daudé
Drop the 'else' so that ret is overridden with the
highest priority fault.
Fixes: d8bc1381250 ("target/hppa: Implement PSW_X")
Reviewed-by: Helge Deller <deller@gmx.de>
Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org>
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
---
target/hppa/mem_helper.c | 8 +++++---
1 file changed, 5 insertions(+), 3 deletions(-)
diff --git a/target/hppa/mem_helper.c b/target/hppa/mem_helper.c
index f027c494e2..f71cedd7a9 100644
--- a/target/hppa/mem_helper.c
+++ b/target/hppa/mem_helper.c
@@ -288,7 +288,7 @@ int hppa_get_physical_address(CPUHPPAState *env, vaddr addr, int mmu_idx,
}
/*
- * In priority order, check for conditions which raise faults.
+ * In reverse priority order, check for conditions which raise faults.
* Remove PROT bits that cover the condition we want to check,
* so that the resulting PROT will force a re-check of the
* architectural TLB entry for the next access.
@@ -299,13 +299,15 @@ int hppa_get_physical_address(CPUHPPAState *env, vaddr addr, int mmu_idx,
/* The T bit is set -- Page Reference Fault. */
ret = EXCP_PAGE_REF;
}
- } else if (!ent->d) {
+ }
+ if (unlikely(!ent->d)) {
prot &= PAGE_READ | PAGE_EXEC;
if (type & PAGE_WRITE) {
/* The D bit is not set -- TLB Dirty Bit Fault. */
ret = EXCP_TLB_DIRTY;
}
- } else if (unlikely(ent->b)) {
+ }
+ if (unlikely(ent->b)) {
prot &= PAGE_READ | PAGE_EXEC;
if (type & PAGE_WRITE) {
/*
--
2.43.0
^ permalink raw reply related [flat|nested] 33+ messages in thread
* [PATCH v3 10/20] target/hppa: Handle alignment faults in hppa_get_physical_address
2024-10-09 0:04 [PATCH v3 00/20] accel/tcg: Introduce tlb_fill_align hook Richard Henderson
` (8 preceding siblings ...)
2024-10-09 0:04 ` [PATCH v3 09/20] target/hppa: Fix priority of T, D, and B page faults Richard Henderson
@ 2024-10-09 0:04 ` Richard Henderson
2024-10-09 0:04 ` [PATCH v3 11/20] target/hppa: Implement TCGCPUOps.tlb_fill_align Richard Henderson
` (9 subsequent siblings)
19 siblings, 0 replies; 33+ messages in thread
From: Richard Henderson @ 2024-10-09 0:04 UTC (permalink / raw)
To: qemu-devel
Cc: deller, peter.maydell, alex.bennee, linux-parisc, qemu-arm,
Helge Deller
In Chapter 5, Interruptions, the group 3 exceptions lists
"Unaligned data reference trap" has higher priority than
"Data memory break trap".
Reviewed-by: Helge Deller <deller@gmx.de>
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
---
target/hppa/mem_helper.c | 7 ++++++-
1 file changed, 6 insertions(+), 1 deletion(-)
diff --git a/target/hppa/mem_helper.c b/target/hppa/mem_helper.c
index f71cedd7a9..d38054da8a 100644
--- a/target/hppa/mem_helper.c
+++ b/target/hppa/mem_helper.c
@@ -221,7 +221,7 @@ int hppa_get_physical_address(CPUHPPAState *env, vaddr addr, int mmu_idx,
g_assert_not_reached();
}
prot = PAGE_READ | PAGE_WRITE | PAGE_EXEC;
- goto egress;
+ goto egress_align;
}
/* Find a valid tlb entry that matches the virtual address. */
@@ -323,6 +323,11 @@ int hppa_get_physical_address(CPUHPPAState *env, vaddr addr, int mmu_idx,
}
}
+ egress_align:
+ if (addr & ((1u << memop_alignment_bits(mop)) - 1)) {
+ ret = EXCP_UNALIGN;
+ }
+
egress:
*pphys = phys;
*pprot = prot;
--
2.43.0
^ permalink raw reply related [flat|nested] 33+ messages in thread
* [PATCH v3 11/20] target/hppa: Implement TCGCPUOps.tlb_fill_align
2024-10-09 0:04 [PATCH v3 00/20] accel/tcg: Introduce tlb_fill_align hook Richard Henderson
` (9 preceding siblings ...)
2024-10-09 0:04 ` [PATCH v3 10/20] target/hppa: Handle alignment faults in hppa_get_physical_address Richard Henderson
@ 2024-10-09 0:04 ` Richard Henderson
2024-10-09 21:33 ` Helge Deller
2024-10-09 0:04 ` [PATCH v3 12/20] target/arm: Pass MemOp to get_phys_addr Richard Henderson
` (8 subsequent siblings)
19 siblings, 1 reply; 33+ messages in thread
From: Richard Henderson @ 2024-10-09 0:04 UTC (permalink / raw)
To: qemu-devel; +Cc: deller, peter.maydell, alex.bennee, linux-parisc, qemu-arm
Convert hppa_cpu_tlb_fill to hppa_cpu_tlb_fill_align so that we
can recognize alignment exceptions in the correct priority order.
Resolves: https://bugzilla.kernel.org/show_bug.cgi?id=219339
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
---
target/hppa/cpu.h | 6 +++---
target/hppa/cpu.c | 2 +-
target/hppa/mem_helper.c | 21 ++++++++++++---------
3 files changed, 16 insertions(+), 13 deletions(-)
diff --git a/target/hppa/cpu.h b/target/hppa/cpu.h
index 526855f982..e45ba50a59 100644
--- a/target/hppa/cpu.h
+++ b/target/hppa/cpu.h
@@ -363,9 +363,9 @@ void hppa_cpu_dump_state(CPUState *cs, FILE *f, int);
void hppa_ptlbe(CPUHPPAState *env);
hwaddr hppa_cpu_get_phys_page_debug(CPUState *cs, vaddr addr);
void hppa_set_ior_and_isr(CPUHPPAState *env, vaddr addr, bool mmu_disabled);
-bool hppa_cpu_tlb_fill(CPUState *cs, vaddr address, int size,
- MMUAccessType access_type, int mmu_idx,
- bool probe, uintptr_t retaddr);
+bool hppa_cpu_tlb_fill_align(CPUState *cs, CPUTLBEntryFull *out, vaddr addr,
+ MMUAccessType access_type, int mmu_idx,
+ MemOp memop, int size, bool probe, uintptr_t ra);
void hppa_cpu_do_interrupt(CPUState *cpu);
bool hppa_cpu_exec_interrupt(CPUState *cpu, int int_req);
int hppa_get_physical_address(CPUHPPAState *env, vaddr addr, int mmu_idx,
diff --git a/target/hppa/cpu.c b/target/hppa/cpu.c
index 7cf2e2f266..c38439c180 100644
--- a/target/hppa/cpu.c
+++ b/target/hppa/cpu.c
@@ -226,7 +226,7 @@ static const TCGCPUOps hppa_tcg_ops = {
.restore_state_to_opc = hppa_restore_state_to_opc,
#ifndef CONFIG_USER_ONLY
- .tlb_fill = hppa_cpu_tlb_fill,
+ .tlb_fill_align = hppa_cpu_tlb_fill_align,
.cpu_exec_interrupt = hppa_cpu_exec_interrupt,
.cpu_exec_halt = hppa_cpu_has_work,
.do_interrupt = hppa_cpu_do_interrupt,
diff --git a/target/hppa/mem_helper.c b/target/hppa/mem_helper.c
index d38054da8a..b8c3e55170 100644
--- a/target/hppa/mem_helper.c
+++ b/target/hppa/mem_helper.c
@@ -424,12 +424,11 @@ void hppa_cpu_do_transaction_failed(CPUState *cs, hwaddr physaddr,
}
}
-bool hppa_cpu_tlb_fill(CPUState *cs, vaddr addr, int size,
- MMUAccessType type, int mmu_idx,
- bool probe, uintptr_t retaddr)
+bool hppa_cpu_tlb_fill_align(CPUState *cs, CPUTLBEntryFull *out, vaddr addr,
+ MMUAccessType type, int mmu_idx,
+ MemOp memop, int size, bool probe, uintptr_t ra)
{
- HPPACPU *cpu = HPPA_CPU(cs);
- CPUHPPAState *env = &cpu->env;
+ CPUHPPAState *env = cpu_env(cs);
int prot, excp, a_prot;
hwaddr phys;
@@ -445,7 +444,7 @@ bool hppa_cpu_tlb_fill(CPUState *cs, vaddr addr, int size,
break;
}
- excp = hppa_get_physical_address(env, addr, mmu_idx, a_prot, 0,
+ excp = hppa_get_physical_address(env, addr, mmu_idx, a_prot, memop,
&phys, &prot);
if (unlikely(excp >= 0)) {
if (probe) {
@@ -454,7 +453,7 @@ bool hppa_cpu_tlb_fill(CPUState *cs, vaddr addr, int size,
trace_hppa_tlb_fill_excp(env, addr, size, type, mmu_idx);
/* Failure. Raise the indicated exception. */
- raise_exception_with_ior(env, excp, retaddr, addr,
+ raise_exception_with_ior(env, excp, ra, addr,
MMU_IDX_MMU_DISABLED(mmu_idx));
}
@@ -468,8 +467,12 @@ bool hppa_cpu_tlb_fill(CPUState *cs, vaddr addr, int size,
* the large page protection mask. We do not require this,
* because we record the large page here in the hppa tlb.
*/
- tlb_set_page(cs, addr & TARGET_PAGE_MASK, phys & TARGET_PAGE_MASK,
- prot, mmu_idx, TARGET_PAGE_SIZE);
+ memset(out, 0, sizeof(*out));
+ out->phys_addr = phys;
+ out->prot = prot;
+ out->attrs = MEMTXATTRS_UNSPECIFIED;
+ out->lg_page_size = TARGET_PAGE_BITS;
+
return true;
}
--
2.43.0
^ permalink raw reply related [flat|nested] 33+ messages in thread
* [PATCH v3 12/20] target/arm: Pass MemOp to get_phys_addr
2024-10-09 0:04 [PATCH v3 00/20] accel/tcg: Introduce tlb_fill_align hook Richard Henderson
` (10 preceding siblings ...)
2024-10-09 0:04 ` [PATCH v3 11/20] target/hppa: Implement TCGCPUOps.tlb_fill_align Richard Henderson
@ 2024-10-09 0:04 ` Richard Henderson
2024-10-09 0:04 ` [PATCH v3 13/20] target/arm: Pass MemOp to get_phys_addr_with_space_nogpc Richard Henderson
` (7 subsequent siblings)
19 siblings, 0 replies; 33+ messages in thread
From: Richard Henderson @ 2024-10-09 0:04 UTC (permalink / raw)
To: qemu-devel
Cc: deller, peter.maydell, alex.bennee, linux-parisc, qemu-arm,
Helge Deller, Philippe Mathieu-Daudé
Zero is the safe do-nothing value for callers to use.
Reviewed-by: Helge Deller <deller@gmx.de>
Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org>
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
---
target/arm/internals.h | 3 ++-
target/arm/ptw.c | 2 +-
target/arm/tcg/m_helper.c | 8 ++++----
target/arm/tcg/tlb_helper.c | 2 +-
4 files changed, 8 insertions(+), 7 deletions(-)
diff --git a/target/arm/internals.h b/target/arm/internals.h
index 1e5da81ce9..2b16579fa5 100644
--- a/target/arm/internals.h
+++ b/target/arm/internals.h
@@ -1432,6 +1432,7 @@ typedef struct GetPhysAddrResult {
* @env: CPUARMState
* @address: virtual address to get physical address for
* @access_type: 0 for read, 1 for write, 2 for execute
+ * @memop: memory operation feeding this access, or 0 for none
* @mmu_idx: MMU index indicating required translation regime
* @result: set on translation success.
* @fi: set to fault info if the translation fails
@@ -1450,7 +1451,7 @@ typedef struct GetPhysAddrResult {
* value.
*/
bool get_phys_addr(CPUARMState *env, vaddr address,
- MMUAccessType access_type, ARMMMUIdx mmu_idx,
+ MMUAccessType access_type, MemOp memop, ARMMMUIdx mmu_idx,
GetPhysAddrResult *result, ARMMMUFaultInfo *fi)
__attribute__((nonnull));
diff --git a/target/arm/ptw.c b/target/arm/ptw.c
index 659855133c..373095a339 100644
--- a/target/arm/ptw.c
+++ b/target/arm/ptw.c
@@ -3572,7 +3572,7 @@ bool get_phys_addr_with_space_nogpc(CPUARMState *env, vaddr address,
}
bool get_phys_addr(CPUARMState *env, vaddr address,
- MMUAccessType access_type, ARMMMUIdx mmu_idx,
+ MMUAccessType access_type, MemOp memop, ARMMMUIdx mmu_idx,
GetPhysAddrResult *result, ARMMMUFaultInfo *fi)
{
S1Translate ptw = {
diff --git a/target/arm/tcg/m_helper.c b/target/arm/tcg/m_helper.c
index 23d7f73035..f7354f3c6e 100644
--- a/target/arm/tcg/m_helper.c
+++ b/target/arm/tcg/m_helper.c
@@ -222,7 +222,7 @@ static bool v7m_stack_write(ARMCPU *cpu, uint32_t addr, uint32_t value,
int exc;
bool exc_secure;
- if (get_phys_addr(env, addr, MMU_DATA_STORE, mmu_idx, &res, &fi)) {
+ if (get_phys_addr(env, addr, MMU_DATA_STORE, 0, mmu_idx, &res, &fi)) {
/* MPU/SAU lookup failed */
if (fi.type == ARMFault_QEMU_SFault) {
if (mode == STACK_LAZYFP) {
@@ -311,7 +311,7 @@ static bool v7m_stack_read(ARMCPU *cpu, uint32_t *dest, uint32_t addr,
bool exc_secure;
uint32_t value;
- if (get_phys_addr(env, addr, MMU_DATA_LOAD, mmu_idx, &res, &fi)) {
+ if (get_phys_addr(env, addr, MMU_DATA_LOAD, 0, mmu_idx, &res, &fi)) {
/* MPU/SAU lookup failed */
if (fi.type == ARMFault_QEMU_SFault) {
qemu_log_mask(CPU_LOG_INT,
@@ -2009,7 +2009,7 @@ static bool v7m_read_half_insn(ARMCPU *cpu, ARMMMUIdx mmu_idx, bool secure,
"...really SecureFault with SFSR.INVEP\n");
return false;
}
- if (get_phys_addr(env, addr, MMU_INST_FETCH, mmu_idx, &res, &fi)) {
+ if (get_phys_addr(env, addr, MMU_INST_FETCH, 0, mmu_idx, &res, &fi)) {
/* the MPU lookup failed */
env->v7m.cfsr[env->v7m.secure] |= R_V7M_CFSR_IACCVIOL_MASK;
armv7m_nvic_set_pending(env->nvic, ARMV7M_EXCP_MEM, env->v7m.secure);
@@ -2045,7 +2045,7 @@ static bool v7m_read_sg_stack_word(ARMCPU *cpu, ARMMMUIdx mmu_idx,
ARMMMUFaultInfo fi = {};
uint32_t value;
- if (get_phys_addr(env, addr, MMU_DATA_LOAD, mmu_idx, &res, &fi)) {
+ if (get_phys_addr(env, addr, MMU_DATA_LOAD, 0, mmu_idx, &res, &fi)) {
/* MPU/SAU lookup failed */
if (fi.type == ARMFault_QEMU_SFault) {
qemu_log_mask(CPU_LOG_INT,
diff --git a/target/arm/tcg/tlb_helper.c b/target/arm/tcg/tlb_helper.c
index 885bf4ec14..1d8b7bcaa2 100644
--- a/target/arm/tcg/tlb_helper.c
+++ b/target/arm/tcg/tlb_helper.c
@@ -344,7 +344,7 @@ bool arm_cpu_tlb_fill(CPUState *cs, vaddr address, int size,
* return false. Otherwise populate fsr with ARM DFSR/IFSR fault
* register format, and signal the fault.
*/
- ret = get_phys_addr(&cpu->env, address, access_type,
+ ret = get_phys_addr(&cpu->env, address, access_type, 0,
core_to_arm_mmu_idx(&cpu->env, mmu_idx),
&res, fi);
if (likely(!ret)) {
--
2.43.0
^ permalink raw reply related [flat|nested] 33+ messages in thread
* [PATCH v3 13/20] target/arm: Pass MemOp to get_phys_addr_with_space_nogpc
2024-10-09 0:04 [PATCH v3 00/20] accel/tcg: Introduce tlb_fill_align hook Richard Henderson
` (11 preceding siblings ...)
2024-10-09 0:04 ` [PATCH v3 12/20] target/arm: Pass MemOp to get_phys_addr Richard Henderson
@ 2024-10-09 0:04 ` Richard Henderson
2024-10-09 0:04 ` [PATCH v3 14/20] target/arm: Pass MemOp to get_phys_addr_gpc Richard Henderson
` (6 subsequent siblings)
19 siblings, 0 replies; 33+ messages in thread
From: Richard Henderson @ 2024-10-09 0:04 UTC (permalink / raw)
To: qemu-devel
Cc: deller, peter.maydell, alex.bennee, linux-parisc, qemu-arm,
Helge Deller, Philippe Mathieu-Daudé
Zero is the safe do-nothing value for callers to use.
Reviewed-by: Helge Deller <deller@gmx.de>
Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org>
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
---
target/arm/internals.h | 3 ++-
target/arm/helper.c | 9 +++++----
target/arm/ptw.c | 2 +-
3 files changed, 8 insertions(+), 6 deletions(-)
diff --git a/target/arm/internals.h b/target/arm/internals.h
index 2b16579fa5..a6088d551c 100644
--- a/target/arm/internals.h
+++ b/target/arm/internals.h
@@ -1461,6 +1461,7 @@ bool get_phys_addr(CPUARMState *env, vaddr address,
* @env: CPUARMState
* @address: virtual address to get physical address for
* @access_type: 0 for read, 1 for write, 2 for execute
+ * @memop: memory operation feeding this access, or 0 for none
* @mmu_idx: MMU index indicating required translation regime
* @space: security space for the access
* @result: set on translation success.
@@ -1470,7 +1471,7 @@ bool get_phys_addr(CPUARMState *env, vaddr address,
* a Granule Protection Check on the resulting address.
*/
bool get_phys_addr_with_space_nogpc(CPUARMState *env, vaddr address,
- MMUAccessType access_type,
+ MMUAccessType access_type, MemOp memop,
ARMMMUIdx mmu_idx, ARMSecuritySpace space,
GetPhysAddrResult *result,
ARMMMUFaultInfo *fi)
diff --git a/target/arm/helper.c b/target/arm/helper.c
index 3f77b40734..0a731a38e8 100644
--- a/target/arm/helper.c
+++ b/target/arm/helper.c
@@ -3599,11 +3599,12 @@ static uint64_t do_ats_write(CPUARMState *env, uint64_t value,
GetPhysAddrResult res = {};
/*
- * I_MXTJT: Granule protection checks are not performed on the final address
- * of a successful translation.
+ * I_MXTJT: Granule protection checks are not performed on the final
+ * address of a successful translation. This is a translation not a
+ * memory reference, so "memop = none = 0".
*/
- ret = get_phys_addr_with_space_nogpc(env, value, access_type, mmu_idx, ss,
- &res, &fi);
+ ret = get_phys_addr_with_space_nogpc(env, value, access_type, 0,
+ mmu_idx, ss, &res, &fi);
/*
* ATS operations only do S1 or S1+S2 translations, so we never
diff --git a/target/arm/ptw.c b/target/arm/ptw.c
index 373095a339..9af86da597 100644
--- a/target/arm/ptw.c
+++ b/target/arm/ptw.c
@@ -3559,7 +3559,7 @@ static bool get_phys_addr_gpc(CPUARMState *env, S1Translate *ptw,
}
bool get_phys_addr_with_space_nogpc(CPUARMState *env, vaddr address,
- MMUAccessType access_type,
+ MMUAccessType access_type, MemOp memop,
ARMMMUIdx mmu_idx, ARMSecuritySpace space,
GetPhysAddrResult *result,
ARMMMUFaultInfo *fi)
--
2.43.0
^ permalink raw reply related [flat|nested] 33+ messages in thread
* [PATCH v3 14/20] target/arm: Pass MemOp to get_phys_addr_gpc
2024-10-09 0:04 [PATCH v3 00/20] accel/tcg: Introduce tlb_fill_align hook Richard Henderson
` (12 preceding siblings ...)
2024-10-09 0:04 ` [PATCH v3 13/20] target/arm: Pass MemOp to get_phys_addr_with_space_nogpc Richard Henderson
@ 2024-10-09 0:04 ` Richard Henderson
2024-10-09 0:04 ` [PATCH v3 15/20] target/arm: Pass MemOp to get_phys_addr_nogpc Richard Henderson
` (5 subsequent siblings)
19 siblings, 0 replies; 33+ messages in thread
From: Richard Henderson @ 2024-10-09 0:04 UTC (permalink / raw)
To: qemu-devel
Cc: deller, peter.maydell, alex.bennee, linux-parisc, qemu-arm,
Helge Deller, Philippe Mathieu-Daudé
Zero is the safe do-nothing value for callers to use.
Pass the value through from get_phys_addr.
Reviewed-by: Helge Deller <deller@gmx.de>
Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org>
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
---
target/arm/ptw.c | 11 ++++++-----
1 file changed, 6 insertions(+), 5 deletions(-)
diff --git a/target/arm/ptw.c b/target/arm/ptw.c
index 9af86da597..e92537d8f2 100644
--- a/target/arm/ptw.c
+++ b/target/arm/ptw.c
@@ -81,7 +81,7 @@ static bool get_phys_addr_nogpc(CPUARMState *env, S1Translate *ptw,
static bool get_phys_addr_gpc(CPUARMState *env, S1Translate *ptw,
vaddr address,
- MMUAccessType access_type,
+ MMUAccessType access_type, MemOp memop,
GetPhysAddrResult *result,
ARMMMUFaultInfo *fi);
@@ -579,7 +579,7 @@ static bool S1_ptw_translate(CPUARMState *env, S1Translate *ptw,
};
GetPhysAddrResult s2 = { };
- if (get_phys_addr_gpc(env, &s2ptw, addr, MMU_DATA_LOAD, &s2, fi)) {
+ if (get_phys_addr_gpc(env, &s2ptw, addr, MMU_DATA_LOAD, 0, &s2, fi)) {
goto fail;
}
@@ -3543,7 +3543,7 @@ static bool get_phys_addr_nogpc(CPUARMState *env, S1Translate *ptw,
static bool get_phys_addr_gpc(CPUARMState *env, S1Translate *ptw,
vaddr address,
- MMUAccessType access_type,
+ MMUAccessType access_type, MemOp memop,
GetPhysAddrResult *result,
ARMMMUFaultInfo *fi)
{
@@ -3641,7 +3641,8 @@ bool get_phys_addr(CPUARMState *env, vaddr address,
}
ptw.in_space = ss;
- return get_phys_addr_gpc(env, &ptw, address, access_type, result, fi);
+ return get_phys_addr_gpc(env, &ptw, address, access_type,
+ memop, result, fi);
}
hwaddr arm_cpu_get_phys_page_attrs_debug(CPUState *cs, vaddr addr,
@@ -3660,7 +3661,7 @@ hwaddr arm_cpu_get_phys_page_attrs_debug(CPUState *cs, vaddr addr,
ARMMMUFaultInfo fi = {};
bool ret;
- ret = get_phys_addr_gpc(env, &ptw, addr, MMU_DATA_LOAD, &res, &fi);
+ ret = get_phys_addr_gpc(env, &ptw, addr, MMU_DATA_LOAD, 0, &res, &fi);
*attrs = res.f.attrs;
if (ret) {
--
2.43.0
^ permalink raw reply related [flat|nested] 33+ messages in thread
* [PATCH v3 15/20] target/arm: Pass MemOp to get_phys_addr_nogpc
2024-10-09 0:04 [PATCH v3 00/20] accel/tcg: Introduce tlb_fill_align hook Richard Henderson
` (13 preceding siblings ...)
2024-10-09 0:04 ` [PATCH v3 14/20] target/arm: Pass MemOp to get_phys_addr_gpc Richard Henderson
@ 2024-10-09 0:04 ` Richard Henderson
2024-10-09 0:04 ` [PATCH v3 16/20] target/arm: Pass MemOp through get_phys_addr_twostage Richard Henderson
` (4 subsequent siblings)
19 siblings, 0 replies; 33+ messages in thread
From: Richard Henderson @ 2024-10-09 0:04 UTC (permalink / raw)
To: qemu-devel
Cc: deller, peter.maydell, alex.bennee, linux-parisc, qemu-arm,
Helge Deller, Philippe Mathieu-Daudé
Zero is the safe do-nothing value for callers to use.
Pass the value through from get_phys_addr_gpc and
get_phys_addr_with_space_nogpc.
Reviewed-by: Helge Deller <deller@gmx.de>
Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org>
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
---
target/arm/ptw.c | 14 ++++++++------
1 file changed, 8 insertions(+), 6 deletions(-)
diff --git a/target/arm/ptw.c b/target/arm/ptw.c
index e92537d8f2..0445c3ccf3 100644
--- a/target/arm/ptw.c
+++ b/target/arm/ptw.c
@@ -75,7 +75,7 @@ typedef struct S1Translate {
static bool get_phys_addr_nogpc(CPUARMState *env, S1Translate *ptw,
vaddr address,
- MMUAccessType access_type,
+ MMUAccessType access_type, MemOp memop,
GetPhysAddrResult *result,
ARMMMUFaultInfo *fi);
@@ -3313,7 +3313,7 @@ static bool get_phys_addr_twostage(CPUARMState *env, S1Translate *ptw,
ARMSecuritySpace ipa_space;
uint64_t hcr;
- ret = get_phys_addr_nogpc(env, ptw, address, access_type, result, fi);
+ ret = get_phys_addr_nogpc(env, ptw, address, access_type, 0, result, fi);
/* If S1 fails, return early. */
if (ret) {
@@ -3339,7 +3339,7 @@ static bool get_phys_addr_twostage(CPUARMState *env, S1Translate *ptw,
cacheattrs1 = result->cacheattrs;
memset(result, 0, sizeof(*result));
- ret = get_phys_addr_nogpc(env, ptw, ipa, access_type, result, fi);
+ ret = get_phys_addr_nogpc(env, ptw, ipa, access_type, 0, result, fi);
fi->s2addr = ipa;
/* Combine the S1 and S2 perms. */
@@ -3406,7 +3406,7 @@ static bool get_phys_addr_twostage(CPUARMState *env, S1Translate *ptw,
static bool get_phys_addr_nogpc(CPUARMState *env, S1Translate *ptw,
vaddr address,
- MMUAccessType access_type,
+ MMUAccessType access_type, MemOp memop,
GetPhysAddrResult *result,
ARMMMUFaultInfo *fi)
{
@@ -3547,7 +3547,8 @@ static bool get_phys_addr_gpc(CPUARMState *env, S1Translate *ptw,
GetPhysAddrResult *result,
ARMMMUFaultInfo *fi)
{
- if (get_phys_addr_nogpc(env, ptw, address, access_type, result, fi)) {
+ if (get_phys_addr_nogpc(env, ptw, address, access_type,
+ memop, result, fi)) {
return true;
}
if (!granule_protection_check(env, result->f.phys_addr,
@@ -3568,7 +3569,8 @@ bool get_phys_addr_with_space_nogpc(CPUARMState *env, vaddr address,
.in_mmu_idx = mmu_idx,
.in_space = space,
};
- return get_phys_addr_nogpc(env, &ptw, address, access_type, result, fi);
+ return get_phys_addr_nogpc(env, &ptw, address, access_type,
+ memop, result, fi);
}
bool get_phys_addr(CPUARMState *env, vaddr address,
--
2.43.0
^ permalink raw reply related [flat|nested] 33+ messages in thread
* [PATCH v3 16/20] target/arm: Pass MemOp through get_phys_addr_twostage
2024-10-09 0:04 [PATCH v3 00/20] accel/tcg: Introduce tlb_fill_align hook Richard Henderson
` (14 preceding siblings ...)
2024-10-09 0:04 ` [PATCH v3 15/20] target/arm: Pass MemOp to get_phys_addr_nogpc Richard Henderson
@ 2024-10-09 0:04 ` Richard Henderson
2024-10-09 0:04 ` [PATCH v3 17/20] target/arm: Pass MemOp to get_phys_addr_lpae Richard Henderson
` (3 subsequent siblings)
19 siblings, 0 replies; 33+ messages in thread
From: Richard Henderson @ 2024-10-09 0:04 UTC (permalink / raw)
To: qemu-devel
Cc: deller, peter.maydell, alex.bennee, linux-parisc, qemu-arm,
Helge Deller, Philippe Mathieu-Daudé
Pass memop through get_phys_addr_twostage with its
recursion with get_phys_addr_nogpc.
Reviewed-by: Helge Deller <deller@gmx.de>
Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org>
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
---
target/arm/ptw.c | 10 ++++++----
1 file changed, 6 insertions(+), 4 deletions(-)
diff --git a/target/arm/ptw.c b/target/arm/ptw.c
index 0445c3ccf3..f1fca086a4 100644
--- a/target/arm/ptw.c
+++ b/target/arm/ptw.c
@@ -3301,7 +3301,7 @@ static bool get_phys_addr_disabled(CPUARMState *env,
static bool get_phys_addr_twostage(CPUARMState *env, S1Translate *ptw,
vaddr address,
- MMUAccessType access_type,
+ MMUAccessType access_type, MemOp memop,
GetPhysAddrResult *result,
ARMMMUFaultInfo *fi)
{
@@ -3313,7 +3313,8 @@ static bool get_phys_addr_twostage(CPUARMState *env, S1Translate *ptw,
ARMSecuritySpace ipa_space;
uint64_t hcr;
- ret = get_phys_addr_nogpc(env, ptw, address, access_type, 0, result, fi);
+ ret = get_phys_addr_nogpc(env, ptw, address, access_type,
+ memop, result, fi);
/* If S1 fails, return early. */
if (ret) {
@@ -3339,7 +3340,8 @@ static bool get_phys_addr_twostage(CPUARMState *env, S1Translate *ptw,
cacheattrs1 = result->cacheattrs;
memset(result, 0, sizeof(*result));
- ret = get_phys_addr_nogpc(env, ptw, ipa, access_type, 0, result, fi);
+ ret = get_phys_addr_nogpc(env, ptw, ipa, access_type,
+ memop, result, fi);
fi->s2addr = ipa;
/* Combine the S1 and S2 perms. */
@@ -3469,7 +3471,7 @@ static bool get_phys_addr_nogpc(CPUARMState *env, S1Translate *ptw,
if (arm_feature(env, ARM_FEATURE_EL2) &&
!regime_translation_disabled(env, ARMMMUIdx_Stage2, ptw->in_space)) {
return get_phys_addr_twostage(env, ptw, address, access_type,
- result, fi);
+ memop, result, fi);
}
/* fall through */
--
2.43.0
^ permalink raw reply related [flat|nested] 33+ messages in thread
* [PATCH v3 17/20] target/arm: Pass MemOp to get_phys_addr_lpae
2024-10-09 0:04 [PATCH v3 00/20] accel/tcg: Introduce tlb_fill_align hook Richard Henderson
` (15 preceding siblings ...)
2024-10-09 0:04 ` [PATCH v3 16/20] target/arm: Pass MemOp through get_phys_addr_twostage Richard Henderson
@ 2024-10-09 0:04 ` Richard Henderson
2024-10-09 0:04 ` [PATCH v3 18/20] target/arm: Move device detection earlier in get_phys_addr_lpae Richard Henderson
` (2 subsequent siblings)
19 siblings, 0 replies; 33+ messages in thread
From: Richard Henderson @ 2024-10-09 0:04 UTC (permalink / raw)
To: qemu-devel
Cc: deller, peter.maydell, alex.bennee, linux-parisc, qemu-arm,
Helge Deller, Philippe Mathieu-Daudé
Pass the value through from get_phys_addr_nogpc.
Reviewed-by: Helge Deller <deller@gmx.de>
Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org>
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
---
target/arm/ptw.c | 6 ++++--
1 file changed, 4 insertions(+), 2 deletions(-)
diff --git a/target/arm/ptw.c b/target/arm/ptw.c
index f1fca086a4..238b2c92a9 100644
--- a/target/arm/ptw.c
+++ b/target/arm/ptw.c
@@ -1684,12 +1684,13 @@ static bool nv_nv1_enabled(CPUARMState *env, S1Translate *ptw)
* @ptw: Current and next stage parameters for the walk.
* @address: virtual address to get physical address for
* @access_type: MMU_DATA_LOAD, MMU_DATA_STORE or MMU_INST_FETCH
+ * @memop: memory operation feeding this access, or 0 for none
* @result: set on translation success,
* @fi: set to fault info if the translation fails
*/
static bool get_phys_addr_lpae(CPUARMState *env, S1Translate *ptw,
uint64_t address,
- MMUAccessType access_type,
+ MMUAccessType access_type, MemOp memop,
GetPhysAddrResult *result, ARMMMUFaultInfo *fi)
{
ARMCPU *cpu = env_archcpu(env);
@@ -3534,7 +3535,8 @@ static bool get_phys_addr_nogpc(CPUARMState *env, S1Translate *ptw,
}
if (regime_using_lpae_format(env, mmu_idx)) {
- return get_phys_addr_lpae(env, ptw, address, access_type, result, fi);
+ return get_phys_addr_lpae(env, ptw, address, access_type,
+ memop, result, fi);
} else if (arm_feature(env, ARM_FEATURE_V7) ||
regime_sctlr(env, mmu_idx) & SCTLR_XP) {
return get_phys_addr_v6(env, ptw, address, access_type, result, fi);
--
2.43.0
^ permalink raw reply related [flat|nested] 33+ messages in thread
* [PATCH v3 18/20] target/arm: Move device detection earlier in get_phys_addr_lpae
2024-10-09 0:04 [PATCH v3 00/20] accel/tcg: Introduce tlb_fill_align hook Richard Henderson
` (16 preceding siblings ...)
2024-10-09 0:04 ` [PATCH v3 17/20] target/arm: Pass MemOp to get_phys_addr_lpae Richard Henderson
@ 2024-10-09 0:04 ` Richard Henderson
2024-10-09 0:04 ` [PATCH v3 19/20] target/arm: Implement TCGCPUOps.tlb_fill_align Richard Henderson
2024-10-09 0:04 ` [PATCH v3 20/20] target/arm: Fix alignment fault priority in get_phys_addr_lpae Richard Henderson
19 siblings, 0 replies; 33+ messages in thread
From: Richard Henderson @ 2024-10-09 0:04 UTC (permalink / raw)
To: qemu-devel; +Cc: deller, peter.maydell, alex.bennee, linux-parisc, qemu-arm
Determine cache attributes, and thence Device vs Normal memory,
earlier in the function. We have an existing regime_is_stage2
if block into which this can be slotted.
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
---
target/arm/ptw.c | 49 ++++++++++++++++++++++++------------------------
1 file changed, 25 insertions(+), 24 deletions(-)
diff --git a/target/arm/ptw.c b/target/arm/ptw.c
index 238b2c92a9..0a1a820362 100644
--- a/target/arm/ptw.c
+++ b/target/arm/ptw.c
@@ -2029,8 +2029,20 @@ static bool get_phys_addr_lpae(CPUARMState *env, S1Translate *ptw,
xn = extract64(attrs, 53, 2);
result->f.prot = get_S2prot(env, ap, xn, ptw->in_s1_is_el0);
}
+
+ result->cacheattrs.is_s2_format = true;
+ result->cacheattrs.attrs = extract32(attrs, 2, 4);
+ /*
+ * Security state does not really affect HCR_EL2.FWB;
+ * we only need to filter FWB for aa32 or other FEAT.
+ */
+ device = S2_attrs_are_device(arm_hcr_el2_eff(env),
+ result->cacheattrs.attrs);
} else {
int nse, ns = extract32(attrs, 5, 1);
+ uint8_t attrindx;
+ uint64_t mair;
+
switch (out_space) {
case ARMSS_Root:
/*
@@ -2102,6 +2114,19 @@ static bool get_phys_addr_lpae(CPUARMState *env, S1Translate *ptw,
*/
result->f.prot = get_S1prot(env, mmu_idx, aarch64, ap, xn, pxn,
result->f.attrs.space, out_space);
+
+ /* Index into MAIR registers for cache attributes */
+ attrindx = extract32(attrs, 2, 3);
+ mair = env->cp15.mair_el[regime_el(env, mmu_idx)];
+ assert(attrindx <= 7);
+ result->cacheattrs.is_s2_format = false;
+ result->cacheattrs.attrs = extract64(mair, attrindx * 8, 8);
+
+ /* When in aarch64 mode, and BTI is enabled, remember GP in the TLB. */
+ if (aarch64 && cpu_isar_feature(aa64_bti, cpu)) {
+ result->f.extra.arm.guarded = extract64(attrs, 50, 1); /* GP */
+ }
+ device = S1_attrs_are_device(result->cacheattrs.attrs);
}
if (!(result->f.prot & (1 << access_type))) {
@@ -2131,30 +2156,6 @@ static bool get_phys_addr_lpae(CPUARMState *env, S1Translate *ptw,
result->f.attrs.space = out_space;
result->f.attrs.secure = arm_space_is_secure(out_space);
- if (regime_is_stage2(mmu_idx)) {
- result->cacheattrs.is_s2_format = true;
- result->cacheattrs.attrs = extract32(attrs, 2, 4);
- /*
- * Security state does not really affect HCR_EL2.FWB;
- * we only need to filter FWB for aa32 or other FEAT.
- */
- device = S2_attrs_are_device(arm_hcr_el2_eff(env),
- result->cacheattrs.attrs);
- } else {
- /* Index into MAIR registers for cache attributes */
- uint8_t attrindx = extract32(attrs, 2, 3);
- uint64_t mair = env->cp15.mair_el[regime_el(env, mmu_idx)];
- assert(attrindx <= 7);
- result->cacheattrs.is_s2_format = false;
- result->cacheattrs.attrs = extract64(mair, attrindx * 8, 8);
-
- /* When in aarch64 mode, and BTI is enabled, remember GP in the TLB. */
- if (aarch64 && cpu_isar_feature(aa64_bti, cpu)) {
- result->f.extra.arm.guarded = extract64(attrs, 50, 1); /* GP */
- }
- device = S1_attrs_are_device(result->cacheattrs.attrs);
- }
-
/*
* Enable alignment checks on Device memory.
*
--
2.43.0
^ permalink raw reply related [flat|nested] 33+ messages in thread
* [PATCH v3 19/20] target/arm: Implement TCGCPUOps.tlb_fill_align
2024-10-09 0:04 [PATCH v3 00/20] accel/tcg: Introduce tlb_fill_align hook Richard Henderson
` (17 preceding siblings ...)
2024-10-09 0:04 ` [PATCH v3 18/20] target/arm: Move device detection earlier in get_phys_addr_lpae Richard Henderson
@ 2024-10-09 0:04 ` Richard Henderson
2024-10-10 11:54 ` Peter Maydell
2024-10-09 0:04 ` [PATCH v3 20/20] target/arm: Fix alignment fault priority in get_phys_addr_lpae Richard Henderson
19 siblings, 1 reply; 33+ messages in thread
From: Richard Henderson @ 2024-10-09 0:04 UTC (permalink / raw)
To: qemu-devel; +Cc: deller, peter.maydell, alex.bennee, linux-parisc, qemu-arm
Fill in the tlb_fill_align hook. Handle alignment not due to
memory type, since that's no longer handled by generic code.
Pass memop to get_phys_addr.
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
---
target/arm/internals.h | 6 ++---
target/arm/cpu.c | 2 +-
target/arm/tcg/cpu-v7m.c | 2 +-
target/arm/tcg/tlb_helper.c | 49 ++++++++++++++-----------------------
4 files changed, 24 insertions(+), 35 deletions(-)
diff --git a/target/arm/internals.h b/target/arm/internals.h
index a6088d551c..299a96a81a 100644
--- a/target/arm/internals.h
+++ b/target/arm/internals.h
@@ -816,9 +816,9 @@ void arm_cpu_record_sigsegv(CPUState *cpu, vaddr addr,
void arm_cpu_record_sigbus(CPUState *cpu, vaddr addr,
MMUAccessType access_type, uintptr_t ra);
#else
-bool arm_cpu_tlb_fill(CPUState *cs, vaddr address, int size,
- MMUAccessType access_type, int mmu_idx,
- bool probe, uintptr_t retaddr);
+bool arm_cpu_tlb_fill_align(CPUState *cs, CPUTLBEntryFull *out, vaddr addr,
+ MMUAccessType access_type, int mmu_idx,
+ MemOp memop, int size, bool probe, uintptr_t ra);
#endif
static inline int arm_to_core_mmu_idx(ARMMMUIdx mmu_idx)
diff --git a/target/arm/cpu.c b/target/arm/cpu.c
index 19191c2391..1320fd8c8f 100644
--- a/target/arm/cpu.c
+++ b/target/arm/cpu.c
@@ -2663,7 +2663,7 @@ static const TCGCPUOps arm_tcg_ops = {
.record_sigsegv = arm_cpu_record_sigsegv,
.record_sigbus = arm_cpu_record_sigbus,
#else
- .tlb_fill = arm_cpu_tlb_fill,
+ .tlb_fill_align = arm_cpu_tlb_fill_align,
.cpu_exec_interrupt = arm_cpu_exec_interrupt,
.cpu_exec_halt = arm_cpu_exec_halt,
.do_interrupt = arm_cpu_do_interrupt,
diff --git a/target/arm/tcg/cpu-v7m.c b/target/arm/tcg/cpu-v7m.c
index 5496f14dc1..58e54578d6 100644
--- a/target/arm/tcg/cpu-v7m.c
+++ b/target/arm/tcg/cpu-v7m.c
@@ -242,7 +242,7 @@ static const TCGCPUOps arm_v7m_tcg_ops = {
.record_sigsegv = arm_cpu_record_sigsegv,
.record_sigbus = arm_cpu_record_sigbus,
#else
- .tlb_fill = arm_cpu_tlb_fill,
+ .tlb_fill_align = arm_cpu_tlb_fill_align,
.cpu_exec_interrupt = arm_v7m_cpu_exec_interrupt,
.cpu_exec_halt = arm_cpu_exec_halt,
.do_interrupt = arm_v7m_cpu_do_interrupt,
diff --git a/target/arm/tcg/tlb_helper.c b/target/arm/tcg/tlb_helper.c
index 1d8b7bcaa2..8841f039bc 100644
--- a/target/arm/tcg/tlb_helper.c
+++ b/target/arm/tcg/tlb_helper.c
@@ -318,14 +318,13 @@ void arm_cpu_do_transaction_failed(CPUState *cs, hwaddr physaddr,
arm_deliver_fault(cpu, addr, access_type, mmu_idx, &fi);
}
-bool arm_cpu_tlb_fill(CPUState *cs, vaddr address, int size,
- MMUAccessType access_type, int mmu_idx,
- bool probe, uintptr_t retaddr)
+bool arm_cpu_tlb_fill_align(CPUState *cs, CPUTLBEntryFull *out, vaddr address,
+ MMUAccessType access_type, int mmu_idx,
+ MemOp memop, int size, bool probe, uintptr_t ra)
{
ARMCPU *cpu = ARM_CPU(cs);
GetPhysAddrResult res = {};
ARMMMUFaultInfo local_fi, *fi;
- int ret;
/*
* Allow S1_ptw_translate to see any fault generated here.
@@ -339,37 +338,27 @@ bool arm_cpu_tlb_fill(CPUState *cs, vaddr address, int size,
}
/*
- * Walk the page table and (if the mapping exists) add the page
- * to the TLB. On success, return true. Otherwise, if probing,
- * return false. Otherwise populate fsr with ARM DFSR/IFSR fault
- * register format, and signal the fault.
+ * Per R_XCHFJ, alignment fault not due to memory type has
+ * highest precedence. Otherwise, walk the page table and
+ * and collect the page description.
*/
- ret = get_phys_addr(&cpu->env, address, access_type, 0,
- core_to_arm_mmu_idx(&cpu->env, mmu_idx),
- &res, fi);
- if (likely(!ret)) {
- /*
- * Map a single [sub]page. Regions smaller than our declared
- * target page size are handled specially, so for those we
- * pass in the exact addresses.
- */
- if (res.f.lg_page_size >= TARGET_PAGE_BITS) {
- res.f.phys_addr &= TARGET_PAGE_MASK;
- address &= TARGET_PAGE_MASK;
- }
-
+ if (address & ((1 << memop_alignment_bits(memop)) - 1)) {
+ fi->type = ARMFault_Alignment;
+ } else if (!get_phys_addr(&cpu->env, address, access_type, memop,
+ core_to_arm_mmu_idx(&cpu->env, mmu_idx),
+ &res, fi)) {
res.f.extra.arm.pte_attrs = res.cacheattrs.attrs;
res.f.extra.arm.shareability = res.cacheattrs.shareability;
-
- tlb_set_page_full(cs, mmu_idx, address, &res.f);
+ *out = res.f;
return true;
- } else if (probe) {
- return false;
- } else {
- /* now we have a real cpu fault */
- cpu_restore_state(cs, retaddr);
- arm_deliver_fault(cpu, address, access_type, mmu_idx, fi);
}
+ if (probe) {
+ return false;
+ }
+
+ /* Now we have a real cpu fault. */
+ cpu_restore_state(cs, ra);
+ arm_deliver_fault(cpu, address, access_type, mmu_idx, fi);
}
#else
void arm_cpu_record_sigsegv(CPUState *cs, vaddr addr,
--
2.43.0
^ permalink raw reply related [flat|nested] 33+ messages in thread
* [PATCH v3 20/20] target/arm: Fix alignment fault priority in get_phys_addr_lpae
2024-10-09 0:04 [PATCH v3 00/20] accel/tcg: Introduce tlb_fill_align hook Richard Henderson
` (18 preceding siblings ...)
2024-10-09 0:04 ` [PATCH v3 19/20] target/arm: Implement TCGCPUOps.tlb_fill_align Richard Henderson
@ 2024-10-09 0:04 ` Richard Henderson
19 siblings, 0 replies; 33+ messages in thread
From: Richard Henderson @ 2024-10-09 0:04 UTC (permalink / raw)
To: qemu-devel; +Cc: deller, peter.maydell, alex.bennee, linux-parisc, qemu-arm
Now that we have the MemOp for the access, we can order
the alignment fault caused by memory type before the
permission fault for the page.
For subsequent page hits, permission and stage 2 checks
are known to pass, and so the TLB_CHECK_ALIGNED fault
raised in generic code is not mis-ordered.
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
---
target/arm/ptw.c | 51 ++++++++++++++++++++++++++++--------------------
1 file changed, 30 insertions(+), 21 deletions(-)
diff --git a/target/arm/ptw.c b/target/arm/ptw.c
index 0a1a820362..dd40268397 100644
--- a/target/arm/ptw.c
+++ b/target/arm/ptw.c
@@ -2129,6 +2129,36 @@ static bool get_phys_addr_lpae(CPUARMState *env, S1Translate *ptw,
device = S1_attrs_are_device(result->cacheattrs.attrs);
}
+ /*
+ * Enable alignment checks on Device memory.
+ *
+ * Per R_XCHFJ, the correct ordering for alignment, permission,
+ * and stage 2 faults is:
+ * - Alignment fault caused by the memory type
+ * - Permission fault
+ * - A stage 2 fault on the memory access
+ * Perform the alignment check now, so that we recognize it in
+ * the correct order. Set TLB_CHECK_ALIGNED so that any subsequent
+ * softmmu tlb hit will also check the alignment; clear along the
+ * non-device path so that tlb_fill_flags is consistent in the
+ * event of restart_atomic_update.
+ *
+ * In v7, for a CPU without the Virtualization Extensions this
+ * access is UNPREDICTABLE; we choose to make it take the alignment
+ * fault as is required for a v7VE CPU. (QEMU doesn't emulate any
+ * CPUs with ARM_FEATURE_LPAE but not ARM_FEATURE_V7VE anyway.)
+ */
+ if (device) {
+ unsigned a_bits = memop_atomicity_bits(memop);
+ if (address & ((1 << a_bits) - 1)) {
+ fi->type = ARMFault_Alignment;
+ goto do_fault;
+ }
+ result->f.tlb_fill_flags = TLB_CHECK_ALIGNED;
+ } else {
+ result->f.tlb_fill_flags = 0;
+ }
+
if (!(result->f.prot & (1 << access_type))) {
fi->type = ARMFault_Permission;
goto do_fault;
@@ -2156,27 +2186,6 @@ static bool get_phys_addr_lpae(CPUARMState *env, S1Translate *ptw,
result->f.attrs.space = out_space;
result->f.attrs.secure = arm_space_is_secure(out_space);
- /*
- * Enable alignment checks on Device memory.
- *
- * Per R_XCHFJ, this check is mis-ordered. The correct ordering
- * for alignment, permission, and stage 2 faults should be:
- * - Alignment fault caused by the memory type
- * - Permission fault
- * - A stage 2 fault on the memory access
- * but due to the way the TCG softmmu TLB operates, we will have
- * implicitly done the permission check and the stage2 lookup in
- * finding the TLB entry, so the alignment check cannot be done sooner.
- *
- * In v7, for a CPU without the Virtualization Extensions this
- * access is UNPREDICTABLE; we choose to make it take the alignment
- * fault as is required for a v7VE CPU. (QEMU doesn't emulate any
- * CPUs with ARM_FEATURE_LPAE but not ARM_FEATURE_V7VE anyway.)
- */
- if (device) {
- result->f.tlb_fill_flags |= TLB_CHECK_ALIGNED;
- }
-
/*
* For FEAT_LPA2 and effective DS, the SH field in the attributes
* was re-purposed for output address bits. The SH attribute in
--
2.43.0
^ permalink raw reply related [flat|nested] 33+ messages in thread
* Re: [PATCH v3 04/20] include/exec/memop: Introduce memop_atomicity_bits
2024-10-09 0:04 ` [PATCH v3 04/20] include/exec/memop: Introduce memop_atomicity_bits Richard Henderson
@ 2024-10-09 18:55 ` Philippe Mathieu-Daudé
0 siblings, 0 replies; 33+ messages in thread
From: Philippe Mathieu-Daudé @ 2024-10-09 18:55 UTC (permalink / raw)
To: Richard Henderson, qemu-devel
Cc: deller, peter.maydell, alex.bennee, linux-parisc, qemu-arm,
Helge Deller
On 8/10/24 21:04, Richard Henderson wrote:
> Split out of mmu_lookup.
>
> Reviewed-by: Helge Deller <deller@gmx.de>
> Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
> Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
> ---
> include/exec/memop.h | 24 ++++++++++++++++++++++++
> accel/tcg/cputlb.c | 16 ++--------------
> 2 files changed, 26 insertions(+), 14 deletions(-)
Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org>
^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: [PATCH v3 05/20] accel/tcg: Add TCGCPUOps.tlb_fill_align
2024-10-09 0:04 ` [PATCH v3 05/20] accel/tcg: Add TCGCPUOps.tlb_fill_align Richard Henderson
@ 2024-10-09 19:46 ` Helge Deller
2024-10-10 10:31 ` Peter Maydell
1 sibling, 0 replies; 33+ messages in thread
From: Helge Deller @ 2024-10-09 19:46 UTC (permalink / raw)
To: Richard Henderson, qemu-devel
Cc: deller, peter.maydell, alex.bennee, linux-parisc, qemu-arm
On 10/9/24 02:04, Richard Henderson wrote:
> Add a new callback to handle softmmu paging. Return the page
> details directly, instead of passing them indirectly to
> tlb_set_page. Handle alignment simultaneously with paging so
> that faults are handled with target-specific priority.
>
> Route all calls the two hooks through a tlb_fill_align
Route all calls (of?) the two hooks... ?
> function local to cputlb.c.
>
> As yet no targets implement the new hook.
> As yet cputlb.c does not use the new alignment check.
>
> Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
> ---
> include/hw/core/cpu.h | 4 +--
> include/hw/core/tcg-cpu-ops.h | 26 +++++++++++++++
> include/qemu/typedefs.h | 1 +
> accel/tcg/cputlb.c | 61 ++++++++++++++++++++++-------------
> 4 files changed, 67 insertions(+), 25 deletions(-)
>
> diff --git a/include/hw/core/cpu.h b/include/hw/core/cpu.h
> index 04e9ad4996..d21a24c82f 100644
> --- a/include/hw/core/cpu.h
> +++ b/include/hw/core/cpu.h
> @@ -205,7 +205,7 @@ struct CPUClass {
> * so the layout is not as critical as that of CPUTLBEntry. This is
> * also why we don't want to combine the two structs.
> */
> -typedef struct CPUTLBEntryFull {
> +struct CPUTLBEntryFull {
> /*
> * @xlat_section contains:
> * - in the lower TARGET_PAGE_BITS, a physical section number
> @@ -261,7 +261,7 @@ typedef struct CPUTLBEntryFull {
> bool guarded;
> } arm;
> } extra;
> -} CPUTLBEntryFull;
> +};
>
> /*
> * Data elements that are per MMU mode, minus the bits accessed by
> diff --git a/include/hw/core/tcg-cpu-ops.h b/include/hw/core/tcg-cpu-ops.h
> index 34318cf0e6..c932690621 100644
> --- a/include/hw/core/tcg-cpu-ops.h
> +++ b/include/hw/core/tcg-cpu-ops.h
> @@ -13,6 +13,7 @@
> #include "exec/breakpoint.h"
> #include "exec/hwaddr.h"
> #include "exec/memattrs.h"
> +#include "exec/memop.h"
> #include "exec/mmu-access-type.h"
> #include "exec/vaddr.h"
>
> @@ -131,6 +132,31 @@ struct TCGCPUOps {
> * same function signature.
> */
> bool (*cpu_exec_halt)(CPUState *cpu);
> + /**
> + * @tlb_fill_align: Handle a softmmu tlb miss
> + * @cpu: cpu context
> + * @out: output page properties
> + * @addr: virtual address
> + * @access_type: read, write or execute
> + * @mmu_idx: mmu context
> + * @memop: memory operation for the access
> + * @size: memory access size, or 0 for whole page
> + * @probe: test only, no fault
> + * @ra: host return address for exception unwind
> + *
> + * If the access is valid, fill in @out and return true.
> + * Otherwise if probe is true, return false.
> + * Otherwise raise and exception and do not return.
Otherwise raise an(?) exception...
Reviewed-by: Helge Deller <deller@gmx.de>
^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: [PATCH v3 06/20] accel/tcg: Use the alignment test in tlb_fill_align
2024-10-09 0:04 ` [PATCH v3 06/20] accel/tcg: Use the alignment test in tlb_fill_align Richard Henderson
@ 2024-10-09 21:29 ` Helge Deller
2024-10-10 10:44 ` Peter Maydell
1 sibling, 0 replies; 33+ messages in thread
From: Helge Deller @ 2024-10-09 21:29 UTC (permalink / raw)
To: Richard Henderson, qemu-devel
Cc: deller, peter.maydell, alex.bennee, linux-parisc, qemu-arm
On 10/9/24 02:04, Richard Henderson wrote:
> When we have a tlb miss, defer the alignment check to
> the new tlb_fill_align hook. Move the existing alignment
> check so that we only perform it with a tlb hit.
>
> Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
Reviewed-by: Helge Deller <deller@gmx.de>
^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: [PATCH v3 11/20] target/hppa: Implement TCGCPUOps.tlb_fill_align
2024-10-09 0:04 ` [PATCH v3 11/20] target/hppa: Implement TCGCPUOps.tlb_fill_align Richard Henderson
@ 2024-10-09 21:33 ` Helge Deller
0 siblings, 0 replies; 33+ messages in thread
From: Helge Deller @ 2024-10-09 21:33 UTC (permalink / raw)
To: Richard Henderson, qemu-devel
Cc: deller, peter.maydell, alex.bennee, linux-parisc, qemu-arm
On 10/9/24 02:04, Richard Henderson wrote:
> Convert hppa_cpu_tlb_fill to hppa_cpu_tlb_fill_align so that we
> can recognize alignment exceptions in the correct priority order.
>
> Resolves: https://bugzilla.kernel.org/show_bug.cgi?id=219339
> Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
Reviewed-by: Helge Deller <deller@gmx.de>
Tested-by: Helge Deller <deller@gmx.de>
> ---
> target/hppa/cpu.h | 6 +++---
> target/hppa/cpu.c | 2 +-
> target/hppa/mem_helper.c | 21 ++++++++++++---------
> 3 files changed, 16 insertions(+), 13 deletions(-)
^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: [PATCH v3 05/20] accel/tcg: Add TCGCPUOps.tlb_fill_align
2024-10-09 0:04 ` [PATCH v3 05/20] accel/tcg: Add TCGCPUOps.tlb_fill_align Richard Henderson
2024-10-09 19:46 ` Helge Deller
@ 2024-10-10 10:31 ` Peter Maydell
1 sibling, 0 replies; 33+ messages in thread
From: Peter Maydell @ 2024-10-10 10:31 UTC (permalink / raw)
To: Richard Henderson; +Cc: qemu-devel, deller, alex.bennee, linux-parisc, qemu-arm
On Wed, 9 Oct 2024 at 01:05, Richard Henderson
<richard.henderson@linaro.org> wrote:
>
> Add a new callback to handle softmmu paging. Return the page
> details directly, instead of passing them indirectly to
> tlb_set_page. Handle alignment simultaneously with paging so
> that faults are handled with target-specific priority.
>
> Route all calls the two hooks through a tlb_fill_align
> function local to cputlb.c.
>
> As yet no targets implement the new hook.
> As yet cputlb.c does not use the new alignment check.
>
> Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
> ---
Other than the typos Helge spotted
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
thanks
-- PMM
^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: [PATCH v3 06/20] accel/tcg: Use the alignment test in tlb_fill_align
2024-10-09 0:04 ` [PATCH v3 06/20] accel/tcg: Use the alignment test in tlb_fill_align Richard Henderson
2024-10-09 21:29 ` Helge Deller
@ 2024-10-10 10:44 ` Peter Maydell
2024-10-11 15:22 ` Richard Henderson
1 sibling, 1 reply; 33+ messages in thread
From: Peter Maydell @ 2024-10-10 10:44 UTC (permalink / raw)
To: Richard Henderson; +Cc: qemu-devel, deller, alex.bennee, linux-parisc, qemu-arm
On Wed, 9 Oct 2024 at 01:05, Richard Henderson
<richard.henderson@linaro.org> wrote:
>
> When we have a tlb miss, defer the alignment check to
> the new tlb_fill_align hook. Move the existing alignment
> check so that we only perform it with a tlb hit.
>
> Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
> @@ -1754,8 +1767,8 @@ static bool mmu_lookup(CPUState *cpu, vaddr addr, MemOpIdx oi,
> * Lookup both pages, recognizing exceptions from either. If the
> * second lookup potentially resized, refresh first CPUTLBEntryFull.
> */
> - mmu_lookup1(cpu, &l->page[0], l->mmu_idx, type, ra);
> - if (mmu_lookup1(cpu, &l->page[1], l->mmu_idx, type, ra)) {
> + mmu_lookup1(cpu, &l->page[0], l->memop, l->mmu_idx, type, ra);
> + if (mmu_lookup1(cpu, &l->page[1], 0, l->mmu_idx, type, ra)) {
Is 0 the right thing here ? Given that alignment-requirements can
differ per-page, what happens for the case of an unaligned access
that crosses a page boundary where the first page is "memory that
doesn't care about alignment" and the second page is "memory that
does enforce alignment" ?
For Arm this is moot, because an access that crosses a page
boundary into something with different memory attributes is
CONSTRAINED UNPREDICTABLE (per Arm ARM rev K.a B2.15.3), but
maybe other architectures are less flexible.
Also, the comment does say "recognizing exceptions from either",
and we aren't going to do that for alignment issues in page 2,
so if we're OK with ignoring this we should update the comment.
thanks
-- PMM
^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: [PATCH v3 19/20] target/arm: Implement TCGCPUOps.tlb_fill_align
2024-10-09 0:04 ` [PATCH v3 19/20] target/arm: Implement TCGCPUOps.tlb_fill_align Richard Henderson
@ 2024-10-10 11:54 ` Peter Maydell
0 siblings, 0 replies; 33+ messages in thread
From: Peter Maydell @ 2024-10-10 11:54 UTC (permalink / raw)
To: Richard Henderson; +Cc: qemu-devel, deller, alex.bennee, linux-parisc, qemu-arm
On Wed, 9 Oct 2024 at 01:05, Richard Henderson
<richard.henderson@linaro.org> wrote:
>
> Fill in the tlb_fill_align hook. Handle alignment not due to
> memory type, since that's no longer handled by generic code.
> Pass memop to get_phys_addr.
>
> Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
> ---
> target/arm/internals.h | 6 ++---
> target/arm/cpu.c | 2 +-
> target/arm/tcg/cpu-v7m.c | 2 +-
> target/arm/tcg/tlb_helper.c | 49 ++++++++++++++-----------------------
> 4 files changed, 24 insertions(+), 35 deletions(-)
>
Reviewed-by: Peter Maydell <peter.maydell@linaro.org>
thanks
-- PMM
^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: [PATCH v3 06/20] accel/tcg: Use the alignment test in tlb_fill_align
2024-10-10 10:44 ` Peter Maydell
@ 2024-10-11 15:22 ` Richard Henderson
2024-10-11 15:36 ` Peter Maydell
0 siblings, 1 reply; 33+ messages in thread
From: Richard Henderson @ 2024-10-11 15:22 UTC (permalink / raw)
To: Peter Maydell; +Cc: qemu-devel, deller, alex.bennee, linux-parisc, qemu-arm
On 10/10/24 03:44, Peter Maydell wrote:
> On Wed, 9 Oct 2024 at 01:05, Richard Henderson
> <richard.henderson@linaro.org> wrote:
>>
>> When we have a tlb miss, defer the alignment check to
>> the new tlb_fill_align hook. Move the existing alignment
>> check so that we only perform it with a tlb hit.
>>
>> Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
>
>
>> @@ -1754,8 +1767,8 @@ static bool mmu_lookup(CPUState *cpu, vaddr addr, MemOpIdx oi,
>> * Lookup both pages, recognizing exceptions from either. If the
>> * second lookup potentially resized, refresh first CPUTLBEntryFull.
>> */
>> - mmu_lookup1(cpu, &l->page[0], l->mmu_idx, type, ra);
>> - if (mmu_lookup1(cpu, &l->page[1], l->mmu_idx, type, ra)) {
>> + mmu_lookup1(cpu, &l->page[0], l->memop, l->mmu_idx, type, ra);
>> + if (mmu_lookup1(cpu, &l->page[1], 0, l->mmu_idx, type, ra)) {
>
> Is 0 the right thing here ? Given that alignment-requirements can
> differ per-page, what happens for the case of an unaligned access
> that crosses a page boundary where the first page is "memory that
> doesn't care about alignment" and the second page is "memory that
> does enforce alignment" ?
I can't think of anything else that makes sense. The access to the second page is
"aligned" in the sense that it begins at offset 0.
Anyway, alignment as a page property is unique to Arm and...
> For Arm this is moot, because an access that crosses a page
> boundary into something with different memory attributes is
> CONSTRAINED UNPREDICTABLE (per Arm ARM rev K.a B2.15.3), but
> maybe other architectures are less flexible.
... as you say.
> Also, the comment does say "recognizing exceptions from either",
> and we aren't going to do that for alignment issues in page 2,
> so if we're OK with ignoring this we should update the comment.
The second page can still raise page faults. How would you re-word this?
r~
^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: [PATCH v3 06/20] accel/tcg: Use the alignment test in tlb_fill_align
2024-10-11 15:22 ` Richard Henderson
@ 2024-10-11 15:36 ` Peter Maydell
0 siblings, 0 replies; 33+ messages in thread
From: Peter Maydell @ 2024-10-11 15:36 UTC (permalink / raw)
To: Richard Henderson; +Cc: qemu-devel, deller, alex.bennee, linux-parisc, qemu-arm
On Fri, 11 Oct 2024 at 16:22, Richard Henderson
<richard.henderson@linaro.org> wrote:
>
> On 10/10/24 03:44, Peter Maydell wrote:
> > On Wed, 9 Oct 2024 at 01:05, Richard Henderson
> > <richard.henderson@linaro.org> wrote:
> >>
> >> When we have a tlb miss, defer the alignment check to
> >> the new tlb_fill_align hook. Move the existing alignment
> >> check so that we only perform it with a tlb hit.
> >>
> >> Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
> >
> >
> >> @@ -1754,8 +1767,8 @@ static bool mmu_lookup(CPUState *cpu, vaddr addr, MemOpIdx oi,
> >> * Lookup both pages, recognizing exceptions from either. If the
> >> * second lookup potentially resized, refresh first CPUTLBEntryFull.
> >> */
> >> - mmu_lookup1(cpu, &l->page[0], l->mmu_idx, type, ra);
> >> - if (mmu_lookup1(cpu, &l->page[1], l->mmu_idx, type, ra)) {
> >> + mmu_lookup1(cpu, &l->page[0], l->memop, l->mmu_idx, type, ra);
> >> + if (mmu_lookup1(cpu, &l->page[1], 0, l->mmu_idx, type, ra)) {
> >
> > Is 0 the right thing here ? Given that alignment-requirements can
> > differ per-page, what happens for the case of an unaligned access
> > that crosses a page boundary where the first page is "memory that
> > doesn't care about alignment" and the second page is "memory that
> > does enforce alignment" ?
>
> I can't think of anything else that makes sense. The access to the second page is
> "aligned" in the sense that it begins at offset 0.
Ah, yes, you don't get to see how unaligned the access
is in the second mmu lookup.
> Anyway, alignment as a page property is unique to Arm and...
>
>
> > For Arm this is moot, because an access that crosses a page
> > boundary into something with different memory attributes is
> > CONSTRAINED UNPREDICTABLE (per Arm ARM rev K.a B2.15.3), but
> > maybe other architectures are less flexible.
>
> ... as you say.
>
> > Also, the comment does say "recognizing exceptions from either",
> > and we aren't going to do that for alignment issues in page 2,
> > so if we're OK with ignoring this we should update the comment.
>
> The second page can still raise page faults. How would you re-word this?
Mmm, so there's also the case of an unaligned access across the
page boundary where page 1 is present and page 2 is not -- do
we prioritize the alignment fault or the page fault? This
patch will always prioritize the alignment fault. For Arm
that's fine, as we architecturally should do the checks in
that order.
-- PMM
^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: [PATCH v3 09/20] target/hppa: Fix priority of T, D, and B page faults
2024-10-09 0:04 ` [PATCH v3 09/20] target/hppa: Fix priority of T, D, and B page faults Richard Henderson
@ 2024-10-14 18:02 ` Michael Tokarev
2024-10-14 19:31 ` Richard Henderson
0 siblings, 1 reply; 33+ messages in thread
From: Michael Tokarev @ 2024-10-14 18:02 UTC (permalink / raw)
To: Richard Henderson, qemu-devel
Cc: deller, peter.maydell, alex.bennee, linux-parisc, qemu-arm,
Helge Deller, Philippe Mathieu-Daudé
On 09.10.2024 03:04, Richard Henderson wrote:
> Drop the 'else' so that ret is overridden with the
> highest priority fault.
>
> Fixes: d8bc1381250 ("target/hppa: Implement PSW_X")
> Reviewed-by: Helge Deller <deller@gmx.de>
> Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org>
> Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
Is this a qemu-stable material? For now I assume yes, please
let me know if it is not.
Thanks,
/mjt
^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: [PATCH v3 09/20] target/hppa: Fix priority of T, D, and B page faults
2024-10-14 18:02 ` Michael Tokarev
@ 2024-10-14 19:31 ` Richard Henderson
2024-10-14 20:17 ` Michael Tokarev
0 siblings, 1 reply; 33+ messages in thread
From: Richard Henderson @ 2024-10-14 19:31 UTC (permalink / raw)
To: Michael Tokarev, qemu-devel
Cc: deller, peter.maydell, alex.bennee, linux-parisc, qemu-arm,
Helge Deller, Philippe Mathieu-Daudé
On 10/14/24 11:02, Michael Tokarev wrote:
> On 09.10.2024 03:04, Richard Henderson wrote:
>> Drop the 'else' so that ret is overridden with the
>> highest priority fault.
>>
>> Fixes: d8bc1381250 ("target/hppa: Implement PSW_X")
>> Reviewed-by: Helge Deller <deller@gmx.de>
>> Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org>
>> Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
>
> Is this a qemu-stable material? For now I assume yes, please
> let me know if it is not.
The kernel fault thing is pretty nasty.
Fixing that probably requires all of patches 2-11.
I don't think the arm fix is serious enough to backport.
r~
^ permalink raw reply [flat|nested] 33+ messages in thread
* Re: [PATCH v3 09/20] target/hppa: Fix priority of T, D, and B page faults
2024-10-14 19:31 ` Richard Henderson
@ 2024-10-14 20:17 ` Michael Tokarev
0 siblings, 0 replies; 33+ messages in thread
From: Michael Tokarev @ 2024-10-14 20:17 UTC (permalink / raw)
To: Richard Henderson, qemu-devel
Cc: deller, peter.maydell, alex.bennee, linux-parisc, qemu-arm,
Helge Deller, Philippe Mathieu-Daudé
On 14.10.2024 22:31, Richard Henderson wrote:
> On 10/14/24 11:02, Michael Tokarev wrote:
>> On 09.10.2024 03:04, Richard Henderson wrote:
>>> Drop the 'else' so that ret is overridden with the
>>> highest priority fault.
>>>
>>> Fixes: d8bc1381250 ("target/hppa: Implement PSW_X")
>>> Reviewed-by: Helge Deller <deller@gmx.de>
>>> Reviewed-by: Philippe Mathieu-Daudé <philmd@linaro.org>
>>> Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
>>
>> Is this a qemu-stable material? For now I assume yes, please
>> let me know if it is not.
>
> The kernel fault thing is pretty nasty.
> Fixing that probably requires all of patches 2-11.
>
> I don't think the arm fix is serious enough to backport.
Count me confused.
So, am I right you suggest picking up changes 2-11 for 9.1?
It looks maybe a bit too much, no?
Thanks,
/mjt
^ permalink raw reply [flat|nested] 33+ messages in thread
end of thread, other threads:[~2024-10-14 20:18 UTC | newest]
Thread overview: 33+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2024-10-09 0:04 [PATCH v3 00/20] accel/tcg: Introduce tlb_fill_align hook Richard Henderson
2024-10-09 0:04 ` [PATCH v3 01/20] accel/tcg: Assert noreturn from write-only page for atomics Richard Henderson
2024-10-09 0:04 ` [PATCH v3 02/20] include/exec/memop: Move get_alignment_bits from tcg.h Richard Henderson
2024-10-09 0:04 ` [PATCH v3 03/20] include/exec/memop: Rename get_alignment_bits Richard Henderson
2024-10-09 0:04 ` [PATCH v3 04/20] include/exec/memop: Introduce memop_atomicity_bits Richard Henderson
2024-10-09 18:55 ` Philippe Mathieu-Daudé
2024-10-09 0:04 ` [PATCH v3 05/20] accel/tcg: Add TCGCPUOps.tlb_fill_align Richard Henderson
2024-10-09 19:46 ` Helge Deller
2024-10-10 10:31 ` Peter Maydell
2024-10-09 0:04 ` [PATCH v3 06/20] accel/tcg: Use the alignment test in tlb_fill_align Richard Henderson
2024-10-09 21:29 ` Helge Deller
2024-10-10 10:44 ` Peter Maydell
2024-10-11 15:22 ` Richard Henderson
2024-10-11 15:36 ` Peter Maydell
2024-10-09 0:04 ` [PATCH v3 07/20] target/hppa: Add MemOp argument to hppa_get_physical_address Richard Henderson
2024-10-09 0:04 ` [PATCH v3 08/20] target/hppa: Perform access rights before protection id check Richard Henderson
2024-10-09 0:04 ` [PATCH v3 09/20] target/hppa: Fix priority of T, D, and B page faults Richard Henderson
2024-10-14 18:02 ` Michael Tokarev
2024-10-14 19:31 ` Richard Henderson
2024-10-14 20:17 ` Michael Tokarev
2024-10-09 0:04 ` [PATCH v3 10/20] target/hppa: Handle alignment faults in hppa_get_physical_address Richard Henderson
2024-10-09 0:04 ` [PATCH v3 11/20] target/hppa: Implement TCGCPUOps.tlb_fill_align Richard Henderson
2024-10-09 21:33 ` Helge Deller
2024-10-09 0:04 ` [PATCH v3 12/20] target/arm: Pass MemOp to get_phys_addr Richard Henderson
2024-10-09 0:04 ` [PATCH v3 13/20] target/arm: Pass MemOp to get_phys_addr_with_space_nogpc Richard Henderson
2024-10-09 0:04 ` [PATCH v3 14/20] target/arm: Pass MemOp to get_phys_addr_gpc Richard Henderson
2024-10-09 0:04 ` [PATCH v3 15/20] target/arm: Pass MemOp to get_phys_addr_nogpc Richard Henderson
2024-10-09 0:04 ` [PATCH v3 16/20] target/arm: Pass MemOp through get_phys_addr_twostage Richard Henderson
2024-10-09 0:04 ` [PATCH v3 17/20] target/arm: Pass MemOp to get_phys_addr_lpae Richard Henderson
2024-10-09 0:04 ` [PATCH v3 18/20] target/arm: Move device detection earlier in get_phys_addr_lpae Richard Henderson
2024-10-09 0:04 ` [PATCH v3 19/20] target/arm: Implement TCGCPUOps.tlb_fill_align Richard Henderson
2024-10-10 11:54 ` Peter Maydell
2024-10-09 0:04 ` [PATCH v3 20/20] target/arm: Fix alignment fault priority in get_phys_addr_lpae Richard Henderson
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).