* [PATCH] hw/riscv/riscv-iommu: Get target page info using runtime helpers
@ 2025-03-04 23:19 Philippe Mathieu-Daudé
2025-03-05 2:17 ` Pierrick Bouvier
` (2 more replies)
0 siblings, 3 replies; 4+ messages in thread
From: Philippe Mathieu-Daudé @ 2025-03-04 23:19 UTC (permalink / raw)
To: qemu-devel
Cc: Liu Zhiwei, Bin Meng, qemu-riscv, Daniel Henrique Barboza,
Alistair Francis, Palmer Dabbelt, Pierrick Bouvier, Weiwei Li,
Philippe Mathieu-Daudé
Prefer runtime helpers to get target page size / mask / bits
rather than compile time definitions. This will help to build
these files once for all RISC-V binaries.
Signed-off-by: Philippe Mathieu-Daudé <philmd@linaro.org>
---
hw/riscv/riscv-iommu-pci.c | 4 +++-
hw/riscv/riscv-iommu.c | 29 +++++++++++++++--------------
2 files changed, 18 insertions(+), 15 deletions(-)
diff --git a/hw/riscv/riscv-iommu-pci.c b/hw/riscv/riscv-iommu-pci.c
index 12451869e41..e2b893c5898 100644
--- a/hw/riscv/riscv-iommu-pci.c
+++ b/hw/riscv/riscv-iommu-pci.c
@@ -23,6 +23,7 @@
#include "hw/qdev-properties.h"
#include "hw/riscv/riscv_hart.h"
#include "migration/vmstate.h"
+#include "exec/target_page.h"
#include "qapi/error.h"
#include "qemu/error-report.h"
#include "qemu/host-utils.h"
@@ -102,7 +103,8 @@ static void riscv_iommu_pci_realize(PCIDevice *dev, Error **errp)
qdev_realize(DEVICE(iommu), NULL, errp);
memory_region_init(&s->bar0, OBJECT(s), "riscv-iommu-bar0",
- QEMU_ALIGN_UP(memory_region_size(&iommu->regs_mr), TARGET_PAGE_SIZE));
+ QEMU_ALIGN_UP(memory_region_size(&iommu->regs_mr),
+ qemu_target_page_size()));
memory_region_add_subregion(&s->bar0, 0, &iommu->regs_mr);
pcie_endpoint_cap_init(dev, 0);
diff --git a/hw/riscv/riscv-iommu.c b/hw/riscv/riscv-iommu.c
index e7568ca227a..8bbb33b8b53 100644
--- a/hw/riscv/riscv-iommu.c
+++ b/hw/riscv/riscv-iommu.c
@@ -23,6 +23,7 @@
#include "hw/qdev-properties.h"
#include "hw/riscv/riscv_hart.h"
#include "migration/vmstate.h"
+#include "exec/target_page.h"
#include "qapi/error.h"
#include "qemu/timer.h"
@@ -300,14 +301,14 @@ static int riscv_iommu_spa_fetch(RISCVIOMMUState *s, RISCVIOMMUContext *ctx,
riscv_iommu_msi_check(s, ctx, iotlb->iova)) {
iotlb->target_as = &s->trap_as;
iotlb->translated_addr = iotlb->iova;
- iotlb->addr_mask = ~TARGET_PAGE_MASK;
+ iotlb->addr_mask = ~qemu_target_page_mask();
return 0;
}
/* Exit early for pass-through mode. */
if (!(en_s || en_g)) {
iotlb->translated_addr = iotlb->iova;
- iotlb->addr_mask = ~TARGET_PAGE_MASK;
+ iotlb->addr_mask = ~qemu_target_page_mask();
/* Allow R/W in pass-through mode */
iotlb->perm = IOMMU_RW;
return 0;
@@ -390,7 +391,7 @@ static int riscv_iommu_spa_fetch(RISCVIOMMUState *s, RISCVIOMMUContext *ctx,
do {
const unsigned widened = (pass && !sc[pass].step) ? 2 : 0;
const unsigned va_bits = widened + sc[pass].ptidxbits;
- const unsigned va_skip = TARGET_PAGE_BITS + sc[pass].ptidxbits *
+ const unsigned va_skip = qemu_target_page_bits() + sc[pass].ptidxbits *
(sc[pass].levels - 1 - sc[pass].step);
const unsigned idx = (addr >> va_skip) & ((1 << va_bits) - 1);
const dma_addr_t pte_addr = base + idx * sc[pass].ptesize;
@@ -448,7 +449,7 @@ static int riscv_iommu_spa_fetch(RISCVIOMMUState *s, RISCVIOMMUContext *ctx,
break; /* Reserved leaf PTE flags: PTE_W */
} else if ((pte & (PTE_R | PTE_W | PTE_X)) == (PTE_W | PTE_X)) {
break; /* Reserved leaf PTE flags: PTE_W + PTE_X */
- } else if (ppn & ((1ULL << (va_skip - TARGET_PAGE_BITS)) - 1)) {
+ } else if (ppn & ((1ULL << (va_skip - qemu_target_page_bits())) - 1)) {
break; /* Misaligned PPN */
} else if ((iotlb->perm & IOMMU_RO) && !(pte & PTE_R)) {
break; /* Read access check failed */
@@ -480,7 +481,7 @@ static int riscv_iommu_spa_fetch(RISCVIOMMUState *s, RISCVIOMMUContext *ctx,
riscv_iommu_msi_check(s, ctx, base)) {
/* Trap MSI writes and return GPA address. */
iotlb->target_as = &s->trap_as;
- iotlb->addr_mask = ~TARGET_PAGE_MASK;
+ iotlb->addr_mask = ~qemu_target_page_mask();
return 0;
}
@@ -950,7 +951,7 @@ static int riscv_iommu_ctx_fetch(RISCVIOMMUState *s, RISCVIOMMUContext *ctx)
* device index: [23:16][15:7][6:0]
*/
const int split = depth * 9 + 6 + dc_fmt;
- addr |= ((ctx->devid >> split) << 3) & ~TARGET_PAGE_MASK;
+ addr |= ((ctx->devid >> split) << 3) & ~qemu_target_page_mask();
if (dma_memory_read(s->target_as, addr, &de, sizeof(de),
MEMTXATTRS_UNSPECIFIED) != MEMTX_OK) {
return RISCV_IOMMU_FQ_CAUSE_DDT_LOAD_FAULT;
@@ -968,7 +969,7 @@ static int riscv_iommu_ctx_fetch(RISCVIOMMUState *s, RISCVIOMMUContext *ctx)
}
/* index into device context entry page */
- addr |= (ctx->devid * dc_len) & ~TARGET_PAGE_MASK;
+ addr |= (ctx->devid * dc_len) & ~qemu_target_page_mask();
memset(&dc, 0, sizeof(dc));
if (dma_memory_read(s->target_as, addr, &dc, dc_len,
@@ -1037,7 +1038,7 @@ static int riscv_iommu_ctx_fetch(RISCVIOMMUState *s, RISCVIOMMUContext *ctx)
* level. See IOMMU Specification, 2.2. Process-Directory-Table.
*/
const int split = depth * 9 + 8;
- addr |= ((ctx->process_id >> split) << 3) & ~TARGET_PAGE_MASK;
+ addr |= ((ctx->process_id >> split) << 3) & ~qemu_target_page_mask();
if (dma_memory_read(s->target_as, addr, &de, sizeof(de),
MEMTXATTRS_UNSPECIFIED) != MEMTX_OK) {
return RISCV_IOMMU_FQ_CAUSE_PDT_LOAD_FAULT;
@@ -1050,7 +1051,7 @@ static int riscv_iommu_ctx_fetch(RISCVIOMMUState *s, RISCVIOMMUContext *ctx)
}
/* Leaf entry in PDT */
- addr |= (ctx->process_id << 4) & ~TARGET_PAGE_MASK;
+ addr |= (ctx->process_id << 4) & ~qemu_target_page_mask();
if (dma_memory_read(s->target_as, addr, &dc.ta, sizeof(uint64_t) * 2,
MEMTXATTRS_UNSPECIFIED) != MEMTX_OK) {
return RISCV_IOMMU_FQ_CAUSE_PDT_LOAD_FAULT;
@@ -1440,7 +1441,7 @@ static int riscv_iommu_translate(RISCVIOMMUState *s, RISCVIOMMUContext *ctx,
perm = iot ? iot->perm : IOMMU_NONE;
if (perm != IOMMU_NONE) {
iotlb->translated_addr = PPN_PHYS(iot->phys);
- iotlb->addr_mask = ~TARGET_PAGE_MASK;
+ iotlb->addr_mask = ~qemu_target_page_mask();
iotlb->perm = perm;
fault = 0;
goto done;
@@ -1481,7 +1482,7 @@ done:
RISCV_IOMMU_PREQ_HDR_PID, ctx->process_id);
}
pr.hdr = set_field(pr.hdr, RISCV_IOMMU_PREQ_HDR_DID, ctx->devid);
- pr.payload = (iotlb->iova & TARGET_PAGE_MASK) |
+ pr.payload = (iotlb->iova & qemu_target_page_mask()) |
RISCV_IOMMU_PREQ_PAYLOAD_M;
riscv_iommu_pri(s, &pr);
return fault;
@@ -1683,7 +1684,7 @@ static void riscv_iommu_process_cq_tail(RISCVIOMMUState *s)
RISCV_IOMMU_CMD_IOTINVAL_GSCID);
uint32_t pscid = get_field(cmd.dword0,
RISCV_IOMMU_CMD_IOTINVAL_PSCID);
- hwaddr iova = (cmd.dword1 << 2) & TARGET_PAGE_MASK;
+ hwaddr iova = (cmd.dword1 << 2) & qemu_target_page_mask();
if (pscv) {
/* illegal command arguments IOTINVAL.GVMA & PSCV == 1 */
@@ -1715,7 +1716,7 @@ static void riscv_iommu_process_cq_tail(RISCVIOMMUState *s)
RISCV_IOMMU_CMD_IOTINVAL_GSCID);
uint32_t pscid = get_field(cmd.dword0,
RISCV_IOMMU_CMD_IOTINVAL_PSCID);
- hwaddr iova = (cmd.dword1 << 2) & TARGET_PAGE_MASK;
+ hwaddr iova = (cmd.dword1 << 2) & qemu_target_page_mask();
RISCVIOMMUTransTag transtag;
if (gv) {
@@ -1928,7 +1929,7 @@ static void riscv_iommu_process_dbg(RISCVIOMMUState *s)
iova = RISCV_IOMMU_TR_RESPONSE_FAULT | (((uint64_t) fault) << 10);
} else {
iova = iotlb.translated_addr & ~iotlb.addr_mask;
- iova >>= TARGET_PAGE_BITS;
+ iova >>= qemu_target_page_bits();
iova &= RISCV_IOMMU_TR_RESPONSE_PPN;
/* We do not support superpages (> 4kbs) for now */
--
2.47.1
^ permalink raw reply related [flat|nested] 4+ messages in thread
* Re: [PATCH] hw/riscv/riscv-iommu: Get target page info using runtime helpers
2025-03-04 23:19 [PATCH] hw/riscv/riscv-iommu: Get target page info using runtime helpers Philippe Mathieu-Daudé
@ 2025-03-05 2:17 ` Pierrick Bouvier
2025-03-06 3:49 ` Alistair Francis
2025-03-06 4:02 ` Alistair Francis
2 siblings, 0 replies; 4+ messages in thread
From: Pierrick Bouvier @ 2025-03-05 2:17 UTC (permalink / raw)
To: Philippe Mathieu-Daudé, qemu-devel
Cc: Liu Zhiwei, Bin Meng, qemu-riscv, Daniel Henrique Barboza,
Alistair Francis, Palmer Dabbelt, Weiwei Li
On 3/4/25 15:19, Philippe Mathieu-Daudé wrote:
> Prefer runtime helpers to get target page size / mask / bits
> rather than compile time definitions. This will help to build
> these files once for all RISC-V binaries.
>
> Signed-off-by: Philippe Mathieu-Daudé <philmd@linaro.org>
> ---
> hw/riscv/riscv-iommu-pci.c | 4 +++-
> hw/riscv/riscv-iommu.c | 29 +++++++++++++++--------------
> 2 files changed, 18 insertions(+), 15 deletions(-)
>
> diff --git a/hw/riscv/riscv-iommu-pci.c b/hw/riscv/riscv-iommu-pci.c
> index 12451869e41..e2b893c5898 100644
> --- a/hw/riscv/riscv-iommu-pci.c
> +++ b/hw/riscv/riscv-iommu-pci.c
> @@ -23,6 +23,7 @@
> #include "hw/qdev-properties.h"
> #include "hw/riscv/riscv_hart.h"
> #include "migration/vmstate.h"
> +#include "exec/target_page.h"
> #include "qapi/error.h"
> #include "qemu/error-report.h"
> #include "qemu/host-utils.h"
> @@ -102,7 +103,8 @@ static void riscv_iommu_pci_realize(PCIDevice *dev, Error **errp)
> qdev_realize(DEVICE(iommu), NULL, errp);
>
> memory_region_init(&s->bar0, OBJECT(s), "riscv-iommu-bar0",
> - QEMU_ALIGN_UP(memory_region_size(&iommu->regs_mr), TARGET_PAGE_SIZE));
> + QEMU_ALIGN_UP(memory_region_size(&iommu->regs_mr),
> + qemu_target_page_size()));
> memory_region_add_subregion(&s->bar0, 0, &iommu->regs_mr);
>
> pcie_endpoint_cap_init(dev, 0);
> diff --git a/hw/riscv/riscv-iommu.c b/hw/riscv/riscv-iommu.c
> index e7568ca227a..8bbb33b8b53 100644
> --- a/hw/riscv/riscv-iommu.c
> +++ b/hw/riscv/riscv-iommu.c
> @@ -23,6 +23,7 @@
> #include "hw/qdev-properties.h"
> #include "hw/riscv/riscv_hart.h"
> #include "migration/vmstate.h"
> +#include "exec/target_page.h"
> #include "qapi/error.h"
> #include "qemu/timer.h"
>
> @@ -300,14 +301,14 @@ static int riscv_iommu_spa_fetch(RISCVIOMMUState *s, RISCVIOMMUContext *ctx,
> riscv_iommu_msi_check(s, ctx, iotlb->iova)) {
> iotlb->target_as = &s->trap_as;
> iotlb->translated_addr = iotlb->iova;
> - iotlb->addr_mask = ~TARGET_PAGE_MASK;
> + iotlb->addr_mask = ~qemu_target_page_mask();
> return 0;
> }
>
> /* Exit early for pass-through mode. */
> if (!(en_s || en_g)) {
> iotlb->translated_addr = iotlb->iova;
> - iotlb->addr_mask = ~TARGET_PAGE_MASK;
> + iotlb->addr_mask = ~qemu_target_page_mask();
> /* Allow R/W in pass-through mode */
> iotlb->perm = IOMMU_RW;
> return 0;
> @@ -390,7 +391,7 @@ static int riscv_iommu_spa_fetch(RISCVIOMMUState *s, RISCVIOMMUContext *ctx,
> do {
> const unsigned widened = (pass && !sc[pass].step) ? 2 : 0;
> const unsigned va_bits = widened + sc[pass].ptidxbits;
> - const unsigned va_skip = TARGET_PAGE_BITS + sc[pass].ptidxbits *
> + const unsigned va_skip = qemu_target_page_bits() + sc[pass].ptidxbits *
> (sc[pass].levels - 1 - sc[pass].step);
> const unsigned idx = (addr >> va_skip) & ((1 << va_bits) - 1);
> const dma_addr_t pte_addr = base + idx * sc[pass].ptesize;
> @@ -448,7 +449,7 @@ static int riscv_iommu_spa_fetch(RISCVIOMMUState *s, RISCVIOMMUContext *ctx,
> break; /* Reserved leaf PTE flags: PTE_W */
> } else if ((pte & (PTE_R | PTE_W | PTE_X)) == (PTE_W | PTE_X)) {
> break; /* Reserved leaf PTE flags: PTE_W + PTE_X */
> - } else if (ppn & ((1ULL << (va_skip - TARGET_PAGE_BITS)) - 1)) {
> + } else if (ppn & ((1ULL << (va_skip - qemu_target_page_bits())) - 1)) {
> break; /* Misaligned PPN */
> } else if ((iotlb->perm & IOMMU_RO) && !(pte & PTE_R)) {
> break; /* Read access check failed */
> @@ -480,7 +481,7 @@ static int riscv_iommu_spa_fetch(RISCVIOMMUState *s, RISCVIOMMUContext *ctx,
> riscv_iommu_msi_check(s, ctx, base)) {
> /* Trap MSI writes and return GPA address. */
> iotlb->target_as = &s->trap_as;
> - iotlb->addr_mask = ~TARGET_PAGE_MASK;
> + iotlb->addr_mask = ~qemu_target_page_mask();
> return 0;
> }
>
> @@ -950,7 +951,7 @@ static int riscv_iommu_ctx_fetch(RISCVIOMMUState *s, RISCVIOMMUContext *ctx)
> * device index: [23:16][15:7][6:0]
> */
> const int split = depth * 9 + 6 + dc_fmt;
> - addr |= ((ctx->devid >> split) << 3) & ~TARGET_PAGE_MASK;
> + addr |= ((ctx->devid >> split) << 3) & ~qemu_target_page_mask();
> if (dma_memory_read(s->target_as, addr, &de, sizeof(de),
> MEMTXATTRS_UNSPECIFIED) != MEMTX_OK) {
> return RISCV_IOMMU_FQ_CAUSE_DDT_LOAD_FAULT;
> @@ -968,7 +969,7 @@ static int riscv_iommu_ctx_fetch(RISCVIOMMUState *s, RISCVIOMMUContext *ctx)
> }
>
> /* index into device context entry page */
> - addr |= (ctx->devid * dc_len) & ~TARGET_PAGE_MASK;
> + addr |= (ctx->devid * dc_len) & ~qemu_target_page_mask();
>
> memset(&dc, 0, sizeof(dc));
> if (dma_memory_read(s->target_as, addr, &dc, dc_len,
> @@ -1037,7 +1038,7 @@ static int riscv_iommu_ctx_fetch(RISCVIOMMUState *s, RISCVIOMMUContext *ctx)
> * level. See IOMMU Specification, 2.2. Process-Directory-Table.
> */
> const int split = depth * 9 + 8;
> - addr |= ((ctx->process_id >> split) << 3) & ~TARGET_PAGE_MASK;
> + addr |= ((ctx->process_id >> split) << 3) & ~qemu_target_page_mask();
> if (dma_memory_read(s->target_as, addr, &de, sizeof(de),
> MEMTXATTRS_UNSPECIFIED) != MEMTX_OK) {
> return RISCV_IOMMU_FQ_CAUSE_PDT_LOAD_FAULT;
> @@ -1050,7 +1051,7 @@ static int riscv_iommu_ctx_fetch(RISCVIOMMUState *s, RISCVIOMMUContext *ctx)
> }
>
> /* Leaf entry in PDT */
> - addr |= (ctx->process_id << 4) & ~TARGET_PAGE_MASK;
> + addr |= (ctx->process_id << 4) & ~qemu_target_page_mask();
> if (dma_memory_read(s->target_as, addr, &dc.ta, sizeof(uint64_t) * 2,
> MEMTXATTRS_UNSPECIFIED) != MEMTX_OK) {
> return RISCV_IOMMU_FQ_CAUSE_PDT_LOAD_FAULT;
> @@ -1440,7 +1441,7 @@ static int riscv_iommu_translate(RISCVIOMMUState *s, RISCVIOMMUContext *ctx,
> perm = iot ? iot->perm : IOMMU_NONE;
> if (perm != IOMMU_NONE) {
> iotlb->translated_addr = PPN_PHYS(iot->phys);
> - iotlb->addr_mask = ~TARGET_PAGE_MASK;
> + iotlb->addr_mask = ~qemu_target_page_mask();
> iotlb->perm = perm;
> fault = 0;
> goto done;
> @@ -1481,7 +1482,7 @@ done:
> RISCV_IOMMU_PREQ_HDR_PID, ctx->process_id);
> }
> pr.hdr = set_field(pr.hdr, RISCV_IOMMU_PREQ_HDR_DID, ctx->devid);
> - pr.payload = (iotlb->iova & TARGET_PAGE_MASK) |
> + pr.payload = (iotlb->iova & qemu_target_page_mask()) |
> RISCV_IOMMU_PREQ_PAYLOAD_M;
> riscv_iommu_pri(s, &pr);
> return fault;
> @@ -1683,7 +1684,7 @@ static void riscv_iommu_process_cq_tail(RISCVIOMMUState *s)
> RISCV_IOMMU_CMD_IOTINVAL_GSCID);
> uint32_t pscid = get_field(cmd.dword0,
> RISCV_IOMMU_CMD_IOTINVAL_PSCID);
> - hwaddr iova = (cmd.dword1 << 2) & TARGET_PAGE_MASK;
> + hwaddr iova = (cmd.dword1 << 2) & qemu_target_page_mask();
>
> if (pscv) {
> /* illegal command arguments IOTINVAL.GVMA & PSCV == 1 */
> @@ -1715,7 +1716,7 @@ static void riscv_iommu_process_cq_tail(RISCVIOMMUState *s)
> RISCV_IOMMU_CMD_IOTINVAL_GSCID);
> uint32_t pscid = get_field(cmd.dword0,
> RISCV_IOMMU_CMD_IOTINVAL_PSCID);
> - hwaddr iova = (cmd.dword1 << 2) & TARGET_PAGE_MASK;
> + hwaddr iova = (cmd.dword1 << 2) & qemu_target_page_mask();
> RISCVIOMMUTransTag transtag;
>
> if (gv) {
> @@ -1928,7 +1929,7 @@ static void riscv_iommu_process_dbg(RISCVIOMMUState *s)
> iova = RISCV_IOMMU_TR_RESPONSE_FAULT | (((uint64_t) fault) << 10);
> } else {
> iova = iotlb.translated_addr & ~iotlb.addr_mask;
> - iova >>= TARGET_PAGE_BITS;
> + iova >>= qemu_target_page_bits();
> iova &= RISCV_IOMMU_TR_RESPONSE_PPN;
>
> /* We do not support superpages (> 4kbs) for now */
Great and simple cleanup.
Reviewed-by: Pierrick Bouvier <pierrick.bouvier@linaro.org>
^ permalink raw reply [flat|nested] 4+ messages in thread
* Re: [PATCH] hw/riscv/riscv-iommu: Get target page info using runtime helpers
2025-03-04 23:19 [PATCH] hw/riscv/riscv-iommu: Get target page info using runtime helpers Philippe Mathieu-Daudé
2025-03-05 2:17 ` Pierrick Bouvier
@ 2025-03-06 3:49 ` Alistair Francis
2025-03-06 4:02 ` Alistair Francis
2 siblings, 0 replies; 4+ messages in thread
From: Alistair Francis @ 2025-03-06 3:49 UTC (permalink / raw)
To: Philippe Mathieu-Daudé
Cc: qemu-devel, Liu Zhiwei, Bin Meng, qemu-riscv,
Daniel Henrique Barboza, Alistair Francis, Palmer Dabbelt,
Pierrick Bouvier, Weiwei Li
On Wed, Mar 5, 2025 at 9:20 AM Philippe Mathieu-Daudé <philmd@linaro.org> wrote:
>
> Prefer runtime helpers to get target page size / mask / bits
> rather than compile time definitions. This will help to build
> these files once for all RISC-V binaries.
>
> Signed-off-by: Philippe Mathieu-Daudé <philmd@linaro.org>
Reviewed-by: Alistair Francis <alistair.francis@wdc.com>
Alistair
> ---
> hw/riscv/riscv-iommu-pci.c | 4 +++-
> hw/riscv/riscv-iommu.c | 29 +++++++++++++++--------------
> 2 files changed, 18 insertions(+), 15 deletions(-)
>
> diff --git a/hw/riscv/riscv-iommu-pci.c b/hw/riscv/riscv-iommu-pci.c
> index 12451869e41..e2b893c5898 100644
> --- a/hw/riscv/riscv-iommu-pci.c
> +++ b/hw/riscv/riscv-iommu-pci.c
> @@ -23,6 +23,7 @@
> #include "hw/qdev-properties.h"
> #include "hw/riscv/riscv_hart.h"
> #include "migration/vmstate.h"
> +#include "exec/target_page.h"
> #include "qapi/error.h"
> #include "qemu/error-report.h"
> #include "qemu/host-utils.h"
> @@ -102,7 +103,8 @@ static void riscv_iommu_pci_realize(PCIDevice *dev, Error **errp)
> qdev_realize(DEVICE(iommu), NULL, errp);
>
> memory_region_init(&s->bar0, OBJECT(s), "riscv-iommu-bar0",
> - QEMU_ALIGN_UP(memory_region_size(&iommu->regs_mr), TARGET_PAGE_SIZE));
> + QEMU_ALIGN_UP(memory_region_size(&iommu->regs_mr),
> + qemu_target_page_size()));
> memory_region_add_subregion(&s->bar0, 0, &iommu->regs_mr);
>
> pcie_endpoint_cap_init(dev, 0);
> diff --git a/hw/riscv/riscv-iommu.c b/hw/riscv/riscv-iommu.c
> index e7568ca227a..8bbb33b8b53 100644
> --- a/hw/riscv/riscv-iommu.c
> +++ b/hw/riscv/riscv-iommu.c
> @@ -23,6 +23,7 @@
> #include "hw/qdev-properties.h"
> #include "hw/riscv/riscv_hart.h"
> #include "migration/vmstate.h"
> +#include "exec/target_page.h"
> #include "qapi/error.h"
> #include "qemu/timer.h"
>
> @@ -300,14 +301,14 @@ static int riscv_iommu_spa_fetch(RISCVIOMMUState *s, RISCVIOMMUContext *ctx,
> riscv_iommu_msi_check(s, ctx, iotlb->iova)) {
> iotlb->target_as = &s->trap_as;
> iotlb->translated_addr = iotlb->iova;
> - iotlb->addr_mask = ~TARGET_PAGE_MASK;
> + iotlb->addr_mask = ~qemu_target_page_mask();
> return 0;
> }
>
> /* Exit early for pass-through mode. */
> if (!(en_s || en_g)) {
> iotlb->translated_addr = iotlb->iova;
> - iotlb->addr_mask = ~TARGET_PAGE_MASK;
> + iotlb->addr_mask = ~qemu_target_page_mask();
> /* Allow R/W in pass-through mode */
> iotlb->perm = IOMMU_RW;
> return 0;
> @@ -390,7 +391,7 @@ static int riscv_iommu_spa_fetch(RISCVIOMMUState *s, RISCVIOMMUContext *ctx,
> do {
> const unsigned widened = (pass && !sc[pass].step) ? 2 : 0;
> const unsigned va_bits = widened + sc[pass].ptidxbits;
> - const unsigned va_skip = TARGET_PAGE_BITS + sc[pass].ptidxbits *
> + const unsigned va_skip = qemu_target_page_bits() + sc[pass].ptidxbits *
> (sc[pass].levels - 1 - sc[pass].step);
> const unsigned idx = (addr >> va_skip) & ((1 << va_bits) - 1);
> const dma_addr_t pte_addr = base + idx * sc[pass].ptesize;
> @@ -448,7 +449,7 @@ static int riscv_iommu_spa_fetch(RISCVIOMMUState *s, RISCVIOMMUContext *ctx,
> break; /* Reserved leaf PTE flags: PTE_W */
> } else if ((pte & (PTE_R | PTE_W | PTE_X)) == (PTE_W | PTE_X)) {
> break; /* Reserved leaf PTE flags: PTE_W + PTE_X */
> - } else if (ppn & ((1ULL << (va_skip - TARGET_PAGE_BITS)) - 1)) {
> + } else if (ppn & ((1ULL << (va_skip - qemu_target_page_bits())) - 1)) {
> break; /* Misaligned PPN */
> } else if ((iotlb->perm & IOMMU_RO) && !(pte & PTE_R)) {
> break; /* Read access check failed */
> @@ -480,7 +481,7 @@ static int riscv_iommu_spa_fetch(RISCVIOMMUState *s, RISCVIOMMUContext *ctx,
> riscv_iommu_msi_check(s, ctx, base)) {
> /* Trap MSI writes and return GPA address. */
> iotlb->target_as = &s->trap_as;
> - iotlb->addr_mask = ~TARGET_PAGE_MASK;
> + iotlb->addr_mask = ~qemu_target_page_mask();
> return 0;
> }
>
> @@ -950,7 +951,7 @@ static int riscv_iommu_ctx_fetch(RISCVIOMMUState *s, RISCVIOMMUContext *ctx)
> * device index: [23:16][15:7][6:0]
> */
> const int split = depth * 9 + 6 + dc_fmt;
> - addr |= ((ctx->devid >> split) << 3) & ~TARGET_PAGE_MASK;
> + addr |= ((ctx->devid >> split) << 3) & ~qemu_target_page_mask();
> if (dma_memory_read(s->target_as, addr, &de, sizeof(de),
> MEMTXATTRS_UNSPECIFIED) != MEMTX_OK) {
> return RISCV_IOMMU_FQ_CAUSE_DDT_LOAD_FAULT;
> @@ -968,7 +969,7 @@ static int riscv_iommu_ctx_fetch(RISCVIOMMUState *s, RISCVIOMMUContext *ctx)
> }
>
> /* index into device context entry page */
> - addr |= (ctx->devid * dc_len) & ~TARGET_PAGE_MASK;
> + addr |= (ctx->devid * dc_len) & ~qemu_target_page_mask();
>
> memset(&dc, 0, sizeof(dc));
> if (dma_memory_read(s->target_as, addr, &dc, dc_len,
> @@ -1037,7 +1038,7 @@ static int riscv_iommu_ctx_fetch(RISCVIOMMUState *s, RISCVIOMMUContext *ctx)
> * level. See IOMMU Specification, 2.2. Process-Directory-Table.
> */
> const int split = depth * 9 + 8;
> - addr |= ((ctx->process_id >> split) << 3) & ~TARGET_PAGE_MASK;
> + addr |= ((ctx->process_id >> split) << 3) & ~qemu_target_page_mask();
> if (dma_memory_read(s->target_as, addr, &de, sizeof(de),
> MEMTXATTRS_UNSPECIFIED) != MEMTX_OK) {
> return RISCV_IOMMU_FQ_CAUSE_PDT_LOAD_FAULT;
> @@ -1050,7 +1051,7 @@ static int riscv_iommu_ctx_fetch(RISCVIOMMUState *s, RISCVIOMMUContext *ctx)
> }
>
> /* Leaf entry in PDT */
> - addr |= (ctx->process_id << 4) & ~TARGET_PAGE_MASK;
> + addr |= (ctx->process_id << 4) & ~qemu_target_page_mask();
> if (dma_memory_read(s->target_as, addr, &dc.ta, sizeof(uint64_t) * 2,
> MEMTXATTRS_UNSPECIFIED) != MEMTX_OK) {
> return RISCV_IOMMU_FQ_CAUSE_PDT_LOAD_FAULT;
> @@ -1440,7 +1441,7 @@ static int riscv_iommu_translate(RISCVIOMMUState *s, RISCVIOMMUContext *ctx,
> perm = iot ? iot->perm : IOMMU_NONE;
> if (perm != IOMMU_NONE) {
> iotlb->translated_addr = PPN_PHYS(iot->phys);
> - iotlb->addr_mask = ~TARGET_PAGE_MASK;
> + iotlb->addr_mask = ~qemu_target_page_mask();
> iotlb->perm = perm;
> fault = 0;
> goto done;
> @@ -1481,7 +1482,7 @@ done:
> RISCV_IOMMU_PREQ_HDR_PID, ctx->process_id);
> }
> pr.hdr = set_field(pr.hdr, RISCV_IOMMU_PREQ_HDR_DID, ctx->devid);
> - pr.payload = (iotlb->iova & TARGET_PAGE_MASK) |
> + pr.payload = (iotlb->iova & qemu_target_page_mask()) |
> RISCV_IOMMU_PREQ_PAYLOAD_M;
> riscv_iommu_pri(s, &pr);
> return fault;
> @@ -1683,7 +1684,7 @@ static void riscv_iommu_process_cq_tail(RISCVIOMMUState *s)
> RISCV_IOMMU_CMD_IOTINVAL_GSCID);
> uint32_t pscid = get_field(cmd.dword0,
> RISCV_IOMMU_CMD_IOTINVAL_PSCID);
> - hwaddr iova = (cmd.dword1 << 2) & TARGET_PAGE_MASK;
> + hwaddr iova = (cmd.dword1 << 2) & qemu_target_page_mask();
>
> if (pscv) {
> /* illegal command arguments IOTINVAL.GVMA & PSCV == 1 */
> @@ -1715,7 +1716,7 @@ static void riscv_iommu_process_cq_tail(RISCVIOMMUState *s)
> RISCV_IOMMU_CMD_IOTINVAL_GSCID);
> uint32_t pscid = get_field(cmd.dword0,
> RISCV_IOMMU_CMD_IOTINVAL_PSCID);
> - hwaddr iova = (cmd.dword1 << 2) & TARGET_PAGE_MASK;
> + hwaddr iova = (cmd.dword1 << 2) & qemu_target_page_mask();
> RISCVIOMMUTransTag transtag;
>
> if (gv) {
> @@ -1928,7 +1929,7 @@ static void riscv_iommu_process_dbg(RISCVIOMMUState *s)
> iova = RISCV_IOMMU_TR_RESPONSE_FAULT | (((uint64_t) fault) << 10);
> } else {
> iova = iotlb.translated_addr & ~iotlb.addr_mask;
> - iova >>= TARGET_PAGE_BITS;
> + iova >>= qemu_target_page_bits();
> iova &= RISCV_IOMMU_TR_RESPONSE_PPN;
>
> /* We do not support superpages (> 4kbs) for now */
> --
> 2.47.1
>
>
^ permalink raw reply [flat|nested] 4+ messages in thread
* Re: [PATCH] hw/riscv/riscv-iommu: Get target page info using runtime helpers
2025-03-04 23:19 [PATCH] hw/riscv/riscv-iommu: Get target page info using runtime helpers Philippe Mathieu-Daudé
2025-03-05 2:17 ` Pierrick Bouvier
2025-03-06 3:49 ` Alistair Francis
@ 2025-03-06 4:02 ` Alistair Francis
2 siblings, 0 replies; 4+ messages in thread
From: Alistair Francis @ 2025-03-06 4:02 UTC (permalink / raw)
To: Philippe Mathieu-Daudé
Cc: qemu-devel, Liu Zhiwei, Bin Meng, qemu-riscv,
Daniel Henrique Barboza, Alistair Francis, Palmer Dabbelt,
Pierrick Bouvier, Weiwei Li
On Wed, Mar 5, 2025 at 9:20 AM Philippe Mathieu-Daudé <philmd@linaro.org> wrote:
>
> Prefer runtime helpers to get target page size / mask / bits
> rather than compile time definitions. This will help to build
> these files once for all RISC-V binaries.
>
> Signed-off-by: Philippe Mathieu-Daudé <philmd@linaro.org>
Thanks!
Do you mind rebasing on
https://github.com/alistair23/qemu/tree/riscv-to-apply.next ?
Alistair
> ---
> hw/riscv/riscv-iommu-pci.c | 4 +++-
> hw/riscv/riscv-iommu.c | 29 +++++++++++++++--------------
> 2 files changed, 18 insertions(+), 15 deletions(-)
>
> diff --git a/hw/riscv/riscv-iommu-pci.c b/hw/riscv/riscv-iommu-pci.c
> index 12451869e41..e2b893c5898 100644
> --- a/hw/riscv/riscv-iommu-pci.c
> +++ b/hw/riscv/riscv-iommu-pci.c
> @@ -23,6 +23,7 @@
> #include "hw/qdev-properties.h"
> #include "hw/riscv/riscv_hart.h"
> #include "migration/vmstate.h"
> +#include "exec/target_page.h"
> #include "qapi/error.h"
> #include "qemu/error-report.h"
> #include "qemu/host-utils.h"
> @@ -102,7 +103,8 @@ static void riscv_iommu_pci_realize(PCIDevice *dev, Error **errp)
> qdev_realize(DEVICE(iommu), NULL, errp);
>
> memory_region_init(&s->bar0, OBJECT(s), "riscv-iommu-bar0",
> - QEMU_ALIGN_UP(memory_region_size(&iommu->regs_mr), TARGET_PAGE_SIZE));
> + QEMU_ALIGN_UP(memory_region_size(&iommu->regs_mr),
> + qemu_target_page_size()));
> memory_region_add_subregion(&s->bar0, 0, &iommu->regs_mr);
>
> pcie_endpoint_cap_init(dev, 0);
> diff --git a/hw/riscv/riscv-iommu.c b/hw/riscv/riscv-iommu.c
> index e7568ca227a..8bbb33b8b53 100644
> --- a/hw/riscv/riscv-iommu.c
> +++ b/hw/riscv/riscv-iommu.c
> @@ -23,6 +23,7 @@
> #include "hw/qdev-properties.h"
> #include "hw/riscv/riscv_hart.h"
> #include "migration/vmstate.h"
> +#include "exec/target_page.h"
> #include "qapi/error.h"
> #include "qemu/timer.h"
>
> @@ -300,14 +301,14 @@ static int riscv_iommu_spa_fetch(RISCVIOMMUState *s, RISCVIOMMUContext *ctx,
> riscv_iommu_msi_check(s, ctx, iotlb->iova)) {
> iotlb->target_as = &s->trap_as;
> iotlb->translated_addr = iotlb->iova;
> - iotlb->addr_mask = ~TARGET_PAGE_MASK;
> + iotlb->addr_mask = ~qemu_target_page_mask();
> return 0;
> }
>
> /* Exit early for pass-through mode. */
> if (!(en_s || en_g)) {
> iotlb->translated_addr = iotlb->iova;
> - iotlb->addr_mask = ~TARGET_PAGE_MASK;
> + iotlb->addr_mask = ~qemu_target_page_mask();
> /* Allow R/W in pass-through mode */
> iotlb->perm = IOMMU_RW;
> return 0;
> @@ -390,7 +391,7 @@ static int riscv_iommu_spa_fetch(RISCVIOMMUState *s, RISCVIOMMUContext *ctx,
> do {
> const unsigned widened = (pass && !sc[pass].step) ? 2 : 0;
> const unsigned va_bits = widened + sc[pass].ptidxbits;
> - const unsigned va_skip = TARGET_PAGE_BITS + sc[pass].ptidxbits *
> + const unsigned va_skip = qemu_target_page_bits() + sc[pass].ptidxbits *
> (sc[pass].levels - 1 - sc[pass].step);
> const unsigned idx = (addr >> va_skip) & ((1 << va_bits) - 1);
> const dma_addr_t pte_addr = base + idx * sc[pass].ptesize;
> @@ -448,7 +449,7 @@ static int riscv_iommu_spa_fetch(RISCVIOMMUState *s, RISCVIOMMUContext *ctx,
> break; /* Reserved leaf PTE flags: PTE_W */
> } else if ((pte & (PTE_R | PTE_W | PTE_X)) == (PTE_W | PTE_X)) {
> break; /* Reserved leaf PTE flags: PTE_W + PTE_X */
> - } else if (ppn & ((1ULL << (va_skip - TARGET_PAGE_BITS)) - 1)) {
> + } else if (ppn & ((1ULL << (va_skip - qemu_target_page_bits())) - 1)) {
> break; /* Misaligned PPN */
> } else if ((iotlb->perm & IOMMU_RO) && !(pte & PTE_R)) {
> break; /* Read access check failed */
> @@ -480,7 +481,7 @@ static int riscv_iommu_spa_fetch(RISCVIOMMUState *s, RISCVIOMMUContext *ctx,
> riscv_iommu_msi_check(s, ctx, base)) {
> /* Trap MSI writes and return GPA address. */
> iotlb->target_as = &s->trap_as;
> - iotlb->addr_mask = ~TARGET_PAGE_MASK;
> + iotlb->addr_mask = ~qemu_target_page_mask();
> return 0;
> }
>
> @@ -950,7 +951,7 @@ static int riscv_iommu_ctx_fetch(RISCVIOMMUState *s, RISCVIOMMUContext *ctx)
> * device index: [23:16][15:7][6:0]
> */
> const int split = depth * 9 + 6 + dc_fmt;
> - addr |= ((ctx->devid >> split) << 3) & ~TARGET_PAGE_MASK;
> + addr |= ((ctx->devid >> split) << 3) & ~qemu_target_page_mask();
> if (dma_memory_read(s->target_as, addr, &de, sizeof(de),
> MEMTXATTRS_UNSPECIFIED) != MEMTX_OK) {
> return RISCV_IOMMU_FQ_CAUSE_DDT_LOAD_FAULT;
> @@ -968,7 +969,7 @@ static int riscv_iommu_ctx_fetch(RISCVIOMMUState *s, RISCVIOMMUContext *ctx)
> }
>
> /* index into device context entry page */
> - addr |= (ctx->devid * dc_len) & ~TARGET_PAGE_MASK;
> + addr |= (ctx->devid * dc_len) & ~qemu_target_page_mask();
>
> memset(&dc, 0, sizeof(dc));
> if (dma_memory_read(s->target_as, addr, &dc, dc_len,
> @@ -1037,7 +1038,7 @@ static int riscv_iommu_ctx_fetch(RISCVIOMMUState *s, RISCVIOMMUContext *ctx)
> * level. See IOMMU Specification, 2.2. Process-Directory-Table.
> */
> const int split = depth * 9 + 8;
> - addr |= ((ctx->process_id >> split) << 3) & ~TARGET_PAGE_MASK;
> + addr |= ((ctx->process_id >> split) << 3) & ~qemu_target_page_mask();
> if (dma_memory_read(s->target_as, addr, &de, sizeof(de),
> MEMTXATTRS_UNSPECIFIED) != MEMTX_OK) {
> return RISCV_IOMMU_FQ_CAUSE_PDT_LOAD_FAULT;
> @@ -1050,7 +1051,7 @@ static int riscv_iommu_ctx_fetch(RISCVIOMMUState *s, RISCVIOMMUContext *ctx)
> }
>
> /* Leaf entry in PDT */
> - addr |= (ctx->process_id << 4) & ~TARGET_PAGE_MASK;
> + addr |= (ctx->process_id << 4) & ~qemu_target_page_mask();
> if (dma_memory_read(s->target_as, addr, &dc.ta, sizeof(uint64_t) * 2,
> MEMTXATTRS_UNSPECIFIED) != MEMTX_OK) {
> return RISCV_IOMMU_FQ_CAUSE_PDT_LOAD_FAULT;
> @@ -1440,7 +1441,7 @@ static int riscv_iommu_translate(RISCVIOMMUState *s, RISCVIOMMUContext *ctx,
> perm = iot ? iot->perm : IOMMU_NONE;
> if (perm != IOMMU_NONE) {
> iotlb->translated_addr = PPN_PHYS(iot->phys);
> - iotlb->addr_mask = ~TARGET_PAGE_MASK;
> + iotlb->addr_mask = ~qemu_target_page_mask();
> iotlb->perm = perm;
> fault = 0;
> goto done;
> @@ -1481,7 +1482,7 @@ done:
> RISCV_IOMMU_PREQ_HDR_PID, ctx->process_id);
> }
> pr.hdr = set_field(pr.hdr, RISCV_IOMMU_PREQ_HDR_DID, ctx->devid);
> - pr.payload = (iotlb->iova & TARGET_PAGE_MASK) |
> + pr.payload = (iotlb->iova & qemu_target_page_mask()) |
> RISCV_IOMMU_PREQ_PAYLOAD_M;
> riscv_iommu_pri(s, &pr);
> return fault;
> @@ -1683,7 +1684,7 @@ static void riscv_iommu_process_cq_tail(RISCVIOMMUState *s)
> RISCV_IOMMU_CMD_IOTINVAL_GSCID);
> uint32_t pscid = get_field(cmd.dword0,
> RISCV_IOMMU_CMD_IOTINVAL_PSCID);
> - hwaddr iova = (cmd.dword1 << 2) & TARGET_PAGE_MASK;
> + hwaddr iova = (cmd.dword1 << 2) & qemu_target_page_mask();
>
> if (pscv) {
> /* illegal command arguments IOTINVAL.GVMA & PSCV == 1 */
> @@ -1715,7 +1716,7 @@ static void riscv_iommu_process_cq_tail(RISCVIOMMUState *s)
> RISCV_IOMMU_CMD_IOTINVAL_GSCID);
> uint32_t pscid = get_field(cmd.dword0,
> RISCV_IOMMU_CMD_IOTINVAL_PSCID);
> - hwaddr iova = (cmd.dword1 << 2) & TARGET_PAGE_MASK;
> + hwaddr iova = (cmd.dword1 << 2) & qemu_target_page_mask();
> RISCVIOMMUTransTag transtag;
>
> if (gv) {
> @@ -1928,7 +1929,7 @@ static void riscv_iommu_process_dbg(RISCVIOMMUState *s)
> iova = RISCV_IOMMU_TR_RESPONSE_FAULT | (((uint64_t) fault) << 10);
> } else {
> iova = iotlb.translated_addr & ~iotlb.addr_mask;
> - iova >>= TARGET_PAGE_BITS;
> + iova >>= qemu_target_page_bits();
> iova &= RISCV_IOMMU_TR_RESPONSE_PPN;
>
> /* We do not support superpages (> 4kbs) for now */
> --
> 2.47.1
>
>
^ permalink raw reply [flat|nested] 4+ messages in thread
end of thread, other threads:[~2025-03-06 4:03 UTC | newest]
Thread overview: 4+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2025-03-04 23:19 [PATCH] hw/riscv/riscv-iommu: Get target page info using runtime helpers Philippe Mathieu-Daudé
2025-03-05 2:17 ` Pierrick Bouvier
2025-03-06 3:49 ` Alistair Francis
2025-03-06 4:02 ` Alistair Francis
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).