From: Richard Henderson <richard.henderson@linaro.org>
To: qemu-devel@nongnu.org
Cc: Ilya Leoshkevich <iii@linux.ibm.com>,
Pierrick Bouvier <pierrick.bouvier@linaro.org>,
Helge Deller <deller@gmx.de>
Subject: [PATCH 26/60] linux-user: Remove qemu_host_page_{size, mask} from mmap.c
Date: Fri, 1 Mar 2024 13:05:45 -1000 [thread overview]
Message-ID: <20240301230619.661008-27-richard.henderson@linaro.org> (raw)
In-Reply-To: <20240301230619.661008-1-richard.henderson@linaro.org>
Use qemu_real_host_page_size instead.
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
Reviewed-by: Ilya Leoshkevich <iii@linux.ibm.com>
Reviewed-by: Pierrick Bouvier <pierrick.bouvier@linaro.org>
Acked-by: Helge Deller <deller@gmx.de>
Message-Id: <20240102015808.132373-9-richard.henderson@linaro.org>
---
linux-user/mmap.c | 66 +++++++++++++++++++++++------------------------
1 file changed, 33 insertions(+), 33 deletions(-)
diff --git a/linux-user/mmap.c b/linux-user/mmap.c
index 96c9433e27..4d3c8717b9 100644
--- a/linux-user/mmap.c
+++ b/linux-user/mmap.c
@@ -165,6 +165,7 @@ static int target_to_host_prot(int prot)
/* NOTE: all the constants are the HOST ones, but addresses are target. */
int target_mprotect(abi_ulong start, abi_ulong len, int target_prot)
{
+ int host_page_size = qemu_real_host_page_size();
abi_ulong starts[3];
abi_ulong lens[3];
int prots[3];
@@ -189,13 +190,13 @@ int target_mprotect(abi_ulong start, abi_ulong len, int target_prot)
}
last = start + len - 1;
- host_start = start & qemu_host_page_mask;
+ host_start = start & -host_page_size;
host_last = HOST_PAGE_ALIGN(last) - 1;
nranges = 0;
mmap_lock();
- if (host_last - host_start < qemu_host_page_size) {
+ if (host_last - host_start < host_page_size) {
/* Single host page contains all guest pages: sum the prot. */
prot1 = target_prot;
for (abi_ulong a = host_start; a < start; a += TARGET_PAGE_SIZE) {
@@ -205,7 +206,7 @@ int target_mprotect(abi_ulong start, abi_ulong len, int target_prot)
prot1 |= page_get_flags(a + 1);
}
starts[nranges] = host_start;
- lens[nranges] = qemu_host_page_size;
+ lens[nranges] = host_page_size;
prots[nranges] = prot1;
nranges++;
} else {
@@ -218,10 +219,10 @@ int target_mprotect(abi_ulong start, abi_ulong len, int target_prot)
/* If the resulting sum differs, create a new range. */
if (prot1 != target_prot) {
starts[nranges] = host_start;
- lens[nranges] = qemu_host_page_size;
+ lens[nranges] = host_page_size;
prots[nranges] = prot1;
nranges++;
- host_start += qemu_host_page_size;
+ host_start += host_page_size;
}
}
@@ -233,9 +234,9 @@ int target_mprotect(abi_ulong start, abi_ulong len, int target_prot)
}
/* If the resulting sum differs, create a new range. */
if (prot1 != target_prot) {
- host_last -= qemu_host_page_size;
+ host_last -= host_page_size;
starts[nranges] = host_last + 1;
- lens[nranges] = qemu_host_page_size;
+ lens[nranges] = host_page_size;
prots[nranges] = prot1;
nranges++;
}
@@ -270,6 +271,7 @@ int target_mprotect(abi_ulong start, abi_ulong len, int target_prot)
static bool mmap_frag(abi_ulong real_start, abi_ulong start, abi_ulong last,
int prot, int flags, int fd, off_t offset)
{
+ int host_page_size = qemu_real_host_page_size();
abi_ulong real_last;
void *host_start;
int prot_old, prot_new;
@@ -286,7 +288,7 @@ static bool mmap_frag(abi_ulong real_start, abi_ulong start, abi_ulong last,
return false;
}
- real_last = real_start + qemu_host_page_size - 1;
+ real_last = real_start + host_page_size - 1;
host_start = g2h_untagged(real_start);
/* Get the protection of the target pages outside the mapping. */
@@ -304,12 +306,12 @@ static bool mmap_frag(abi_ulong real_start, abi_ulong start, abi_ulong last,
* outside of the fragment we need to map. Allocate a new host
* page to cover, discarding whatever else may have been present.
*/
- void *p = mmap(host_start, qemu_host_page_size,
+ void *p = mmap(host_start, host_page_size,
target_to_host_prot(prot),
flags | MAP_ANONYMOUS, -1, 0);
if (p != host_start) {
if (p != MAP_FAILED) {
- munmap(p, qemu_host_page_size);
+ munmap(p, host_page_size);
errno = EEXIST;
}
return false;
@@ -324,7 +326,7 @@ static bool mmap_frag(abi_ulong real_start, abi_ulong start, abi_ulong last,
/* Adjust protection to be able to write. */
if (!(host_prot_old & PROT_WRITE)) {
host_prot_old |= PROT_WRITE;
- mprotect(host_start, qemu_host_page_size, host_prot_old);
+ mprotect(host_start, host_page_size, host_prot_old);
}
/* Read or zero the new guest pages. */
@@ -338,7 +340,7 @@ static bool mmap_frag(abi_ulong real_start, abi_ulong start, abi_ulong last,
/* Put final protection */
if (host_prot_new != host_prot_old) {
- mprotect(host_start, qemu_host_page_size, host_prot_new);
+ mprotect(host_start, host_page_size, host_prot_new);
}
return true;
}
@@ -373,17 +375,18 @@ static abi_ulong mmap_find_vma_reserved(abi_ulong start, abi_ulong size,
*/
abi_ulong mmap_find_vma(abi_ulong start, abi_ulong size, abi_ulong align)
{
+ int host_page_size = qemu_real_host_page_size();
void *ptr, *prev;
abi_ulong addr;
int wrapped, repeat;
- align = MAX(align, qemu_host_page_size);
+ align = MAX(align, host_page_size);
/* If 'start' == 0, then a default start address is used. */
if (start == 0) {
start = mmap_next_start;
} else {
- start &= qemu_host_page_mask;
+ start &= -host_page_size;
}
start = ROUND_UP(start, align);
@@ -492,6 +495,7 @@ abi_ulong mmap_find_vma(abi_ulong start, abi_ulong size, abi_ulong align)
abi_long target_mmap(abi_ulong start, abi_ulong len, int target_prot,
int flags, int fd, off_t offset)
{
+ int host_page_size = qemu_real_host_page_size();
abi_ulong ret, last, real_start, real_last, retaddr, host_len;
abi_ulong passthrough_start = -1, passthrough_last = 0;
int page_flags;
@@ -537,8 +541,8 @@ abi_long target_mmap(abi_ulong start, abi_ulong len, int target_prot,
}
}
- real_start = start & qemu_host_page_mask;
- host_offset = offset & qemu_host_page_mask;
+ real_start = start & -host_page_size;
+ host_offset = offset & -host_page_size;
/*
* If the user is asking for the kernel to find a location, do that
@@ -567,8 +571,7 @@ abi_long target_mmap(abi_ulong start, abi_ulong len, int target_prot,
* may need to truncate file maps at EOF and add extra anonymous pages
* up to the targets page boundary.
*/
- if ((qemu_real_host_page_size() < qemu_host_page_size) &&
- !(flags & MAP_ANONYMOUS)) {
+ if (host_page_size < TARGET_PAGE_SIZE && !(flags & MAP_ANONYMOUS)) {
struct stat sb;
if (fstat(fd, &sb) == -1) {
@@ -595,11 +598,7 @@ abi_long target_mmap(abi_ulong start, abi_ulong len, int target_prot,
host_len = HOST_PAGE_ALIGN(host_len);
host_prot = target_to_host_prot(target_prot);
- /*
- * Note: we prefer to control the mapping address. It is
- * especially important if qemu_host_page_size >
- * qemu_real_host_page_size.
- */
+ /* Note: we prefer to control the mapping address. */
p = mmap(g2h_untagged(start), host_len, host_prot,
flags | MAP_FIXED | MAP_ANONYMOUS, -1, 0);
if (p == MAP_FAILED) {
@@ -665,7 +664,7 @@ abi_long target_mmap(abi_ulong start, abi_ulong len, int target_prot,
* aligned, so we read it
*/
if (!(flags & MAP_ANONYMOUS) &&
- (offset & ~qemu_host_page_mask) != (start & ~qemu_host_page_mask)) {
+ (offset & (host_page_size - 1)) != (start & (host_page_size - 1))) {
/*
* msync() won't work here, so we return an error if write is
* possible while it is a shared mapping
@@ -694,7 +693,7 @@ abi_long target_mmap(abi_ulong start, abi_ulong len, int target_prot,
/* handle the start of the mapping */
if (start > real_start) {
- if (real_last == real_start + qemu_host_page_size - 1) {
+ if (real_last == real_start + host_page_size - 1) {
/* one single host page */
if (!mmap_frag(real_start, start, last,
target_prot, flags, fd, offset)) {
@@ -703,21 +702,21 @@ abi_long target_mmap(abi_ulong start, abi_ulong len, int target_prot,
goto the_end1;
}
if (!mmap_frag(real_start, start,
- real_start + qemu_host_page_size - 1,
+ real_start + host_page_size - 1,
target_prot, flags, fd, offset)) {
goto fail;
}
- real_start += qemu_host_page_size;
+ real_start += host_page_size;
}
/* handle the end of the mapping */
if (last < real_last) {
- abi_ulong real_page = real_last - qemu_host_page_size + 1;
+ abi_ulong real_page = real_last - host_page_size + 1;
if (!mmap_frag(real_page, real_page, last,
target_prot, flags, fd,
offset + real_page - start)) {
goto fail;
}
- real_last -= qemu_host_page_size;
+ real_last -= host_page_size;
}
/* map the middle (easier) */
@@ -784,6 +783,7 @@ fail:
static int mmap_reserve_or_unmap(abi_ulong start, abi_ulong len)
{
+ int host_page_size = qemu_real_host_page_size();
abi_ulong real_start;
abi_ulong real_last;
abi_ulong real_len;
@@ -793,7 +793,7 @@ static int mmap_reserve_or_unmap(abi_ulong start, abi_ulong len)
int prot;
last = start + len - 1;
- real_start = start & qemu_host_page_mask;
+ real_start = start & -host_page_size;
real_last = HOST_PAGE_ALIGN(last) - 1;
/*
@@ -802,7 +802,7 @@ static int mmap_reserve_or_unmap(abi_ulong start, abi_ulong len)
* The single page special case is required for the last page,
* lest real_start overflow to zero.
*/
- if (real_last - real_start < qemu_host_page_size) {
+ if (real_last - real_start < host_page_size) {
prot = 0;
for (a = real_start; a < start; a += TARGET_PAGE_SIZE) {
prot |= page_get_flags(a);
@@ -818,14 +818,14 @@ static int mmap_reserve_or_unmap(abi_ulong start, abi_ulong len)
prot |= page_get_flags(a);
}
if (prot != 0) {
- real_start += qemu_host_page_size;
+ real_start += host_page_size;
}
for (prot = 0, a = last; a < real_last; a += TARGET_PAGE_SIZE) {
prot |= page_get_flags(a + 1);
}
if (prot != 0) {
- real_last -= qemu_host_page_size;
+ real_last -= host_page_size;
}
if (real_last < real_start) {
--
2.34.1
next prev parent reply other threads:[~2024-03-01 23:13 UTC|newest]
Thread overview: 64+ messages / expand[flat|nested] mbox.gz Atom feed top
2024-03-01 23:05 [PATCH 00/60] linux-user and tcg patch queue Richard Henderson
2024-03-01 23:05 ` [PATCH 01/60] linux-user/elfload: Disable core dump if getrlimit fails Richard Henderson
2024-03-01 23:05 ` [PATCH 02/60] linux-user/elfload: Merge init_note_info and fill_note_info Richard Henderson
2024-03-01 23:05 ` [PATCH 03/60] linux-user/elfload: Tidy fill_note_info and struct elf_note_info Richard Henderson
2024-03-01 23:05 ` [PATCH 04/60] linux-user/elfload: Stack allocate struct mm_struct Richard Henderson
2024-03-01 23:05 ` [PATCH 05/60] linux-user/elfload: Latch errno before cleanup in elf_core_dump Richard Henderson
2024-03-01 23:05 ` [PATCH 06/60] linux-user/elfload: Open core file after vma_init Richard Henderson
2024-03-01 23:05 ` [PATCH 07/60] linux-user/elfload: Truncate core file on open Richard Henderson
2024-03-01 23:05 ` [PATCH 08/60] linux-user/elfload: Lock cpu list and mmap during elf_core_dump Richard Henderson
2024-03-01 23:05 ` [PATCH 09/60] linux-user/elfload: Size corefile before opening Richard Henderson
2024-03-01 23:05 ` [PATCH 10/60] linux-user/elfload: Write corefile elf header in one block Richard Henderson
2024-03-01 23:05 ` [PATCH 11/60] linux-user/elfload: Write process memory to core file in larger chunks Richard Henderson
2024-03-01 23:05 ` [PATCH 12/60] linux-user/elfload: Simplify vma_dump_size Richard Henderson
2024-03-01 23:05 ` [PATCH 13/60] linux-user/elfload: Rely on walk_memory_regions for vmas Richard Henderson
2024-03-01 23:05 ` [PATCH 14/60] linux-user/elfload: Unprotect regions before core dump Richard Henderson
2024-03-01 23:05 ` [PATCH 15/60] tcg/aarch64: Apple does not align __int128_t in even registers Richard Henderson
2024-03-01 23:05 ` [PATCH 16/60] accel/tcg: Set can_do_io at at start of lookup_tb_ptr helper Richard Henderson
2024-03-01 23:05 ` [PATCH 17/60] tcg: Avoid double lock if page tables happen to be in mmio memory Richard Henderson
2024-03-01 23:05 ` [PATCH 18/60] accel/tcg: Remove qemu_host_page_size from page_protect/page_unprotect Richard Henderson
2024-03-01 23:05 ` [PATCH 19/60] linux-user: Adjust SVr4 NULL page mapping Richard Henderson
2024-03-01 23:05 ` [PATCH 20/60] linux-user: Remove qemu_host_page_{size, mask} in probe_guest_base Richard Henderson
2024-03-01 23:05 ` [PATCH 21/60] linux-user: Remove qemu_host_page_size from create_elf_tables Richard Henderson
2024-03-01 23:05 ` [PATCH 22/60] linux-user/hppa: Simplify init_guest_commpage Richard Henderson
2024-03-01 23:05 ` [PATCH 23/60] linux-user/nios2: Remove qemu_host_page_size from init_guest_commpage Richard Henderson
2024-03-01 23:05 ` [PATCH 24/60] linux-user/arm: " Richard Henderson
2024-03-01 23:05 ` [PATCH 25/60] linux-user: Remove qemu_host_page_size from elf_core_dump Richard Henderson
2024-03-01 23:05 ` Richard Henderson [this message]
2024-03-01 23:05 ` [PATCH 27/60] linux-user: Remove REAL_HOST_PAGE_ALIGN from mmap.c Richard Henderson
2024-03-01 23:05 ` [PATCH 28/60] linux-user: Remove HOST_PAGE_ALIGN " Richard Henderson
2024-03-01 23:05 ` [PATCH 29/60] migration: Remove qemu_host_page_size Richard Henderson
2024-03-01 23:05 ` [PATCH 30/60] hw/tpm: Remove HOST_PAGE_ALIGN from tpm_ppi_init Richard Henderson
2024-03-01 23:05 ` [PATCH 31/60] softmmu/physmem: Remove qemu_host_page_size Richard Henderson
2024-03-01 23:05 ` [PATCH 32/60] softmmu/physmem: Remove HOST_PAGE_ALIGN Richard Henderson
2024-03-01 23:05 ` [PATCH 33/60] linux-user: Remove qemu_host_page_size from main Richard Henderson
2024-03-01 23:05 ` [PATCH 34/60] linux-user: Split out target_mmap__locked Richard Henderson
2024-03-01 23:05 ` [PATCH 35/60] linux-user: Move some mmap checks outside the lock Richard Henderson
2024-03-01 23:05 ` [PATCH 36/60] linux-user: Fix sub-host-page mmap Richard Henderson
2024-03-01 23:05 ` [PATCH 37/60] linux-user: Split out mmap_end Richard Henderson
2024-03-01 23:05 ` [PATCH 38/60] linux-user: Do early mmap placement only for reserved_va Richard Henderson
2024-03-01 23:05 ` [PATCH 39/60] linux-user: Split out do_munmap Richard Henderson
2024-03-01 23:05 ` [PATCH 40/60] linux-user: Use do_munmap for target_mmap failure Richard Henderson
2024-03-01 23:06 ` [PATCH 41/60] linux-user: Split out mmap_h_eq_g Richard Henderson
2024-03-01 23:06 ` [PATCH 42/60] linux-user: Split out mmap_h_lt_g Richard Henderson
2024-03-01 23:06 ` [PATCH 43/60] linux-user: Split out mmap_h_gt_g Richard Henderson
2024-03-01 23:06 ` [PATCH 44/60] tests/tcg: Remove run-test-mmap-* Richard Henderson
2024-03-01 23:06 ` [PATCH 45/60] tests/tcg: Extend file in linux-madvise.c Richard Henderson
2024-03-01 23:06 ` [PATCH 46/60] *-user: Deprecate and disable -p pagesize Richard Henderson
2024-03-01 23:06 ` [PATCH 47/60] cpu: Remove page_size_init Richard Henderson
2024-03-01 23:06 ` [PATCH 48/60] accel/tcg: Disconnect TargetPageDataNode from page size Richard Henderson
2024-03-01 23:06 ` [PATCH 49/60] linux-user: Allow TARGET_PAGE_BITS_VARY Richard Henderson
2024-03-01 23:06 ` [PATCH 50/60] target/arm: Enable TARGET_PAGE_BITS_VARY for AArch64 user-only Richard Henderson
2024-03-01 23:06 ` [PATCH 51/60] linux-user: Bound mmap_min_addr by host page size Richard Henderson
2024-03-01 23:06 ` [PATCH 52/60] target/ppc: Enable TARGET_PAGE_BITS_VARY for user-only Richard Henderson
2024-03-01 23:06 ` [PATCH 53/60] target/alpha: " Richard Henderson
2024-03-01 23:06 ` [PATCH 54/60] linux-user: Remove pgb_dynamic alignment assertion Richard Henderson
2024-03-01 23:06 ` [PATCH 55/60] tcg/optimize: fix uninitialized variable Richard Henderson
2024-03-01 23:06 ` [PATCH 56/60] linux-user/x86_64: Handle the vsyscall page in open_self_maps_{2, 4} Richard Henderson
2024-04-11 22:09 ` Michael Tokarev
2024-03-01 23:06 ` [PATCH 57/60] linux-user/loongarch64: Remove TARGET_FORCE_SHMLBA Richard Henderson
2024-03-01 23:06 ` [PATCH 58/60] linux-user: Add strace for shmat Richard Henderson
2024-03-01 23:06 ` [PATCH 59/60] linux-user: Rewrite target_shmat Richard Henderson
2024-03-01 23:06 ` [PATCH 60/60] tests/tcg: Check that shmat() does not break /proc/self/maps Richard Henderson
2024-03-01 23:13 ` [PULL 00/60] linux-user and tcg patch queue Richard Henderson
2024-03-05 11:17 ` [PATCH " Peter Maydell
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20240301230619.661008-27-richard.henderson@linaro.org \
--to=richard.henderson@linaro.org \
--cc=deller@gmx.de \
--cc=iii@linux.ibm.com \
--cc=pierrick.bouvier@linaro.org \
--cc=qemu-devel@nongnu.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).