All of lore.kernel.org
 help / color / mirror / Atom feed
From: Richard Henderson <richard.henderson@linaro.org>
To: qemu-devel@nongnu.org
Cc: Ilya Leoshkevich <iii@linux.ibm.com>,
	Pierrick Bouvier <pierrick.bouvier@linaro.org>,
	Helge Deller <deller@gmx.de>
Subject: [PATCH 26/60] linux-user: Remove qemu_host_page_{size, mask} from mmap.c
Date: Fri,  1 Mar 2024 13:05:45 -1000	[thread overview]
Message-ID: <20240301230619.661008-27-richard.henderson@linaro.org> (raw)
In-Reply-To: <20240301230619.661008-1-richard.henderson@linaro.org>

Use qemu_real_host_page_size instead.

Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
Reviewed-by: Ilya Leoshkevich <iii@linux.ibm.com>
Reviewed-by: Pierrick Bouvier <pierrick.bouvier@linaro.org>
Acked-by: Helge Deller <deller@gmx.de>
Message-Id: <20240102015808.132373-9-richard.henderson@linaro.org>
---
 linux-user/mmap.c | 66 +++++++++++++++++++++++------------------------
 1 file changed, 33 insertions(+), 33 deletions(-)

diff --git a/linux-user/mmap.c b/linux-user/mmap.c
index 96c9433e27..4d3c8717b9 100644
--- a/linux-user/mmap.c
+++ b/linux-user/mmap.c
@@ -165,6 +165,7 @@ static int target_to_host_prot(int prot)
 /* NOTE: all the constants are the HOST ones, but addresses are target. */
 int target_mprotect(abi_ulong start, abi_ulong len, int target_prot)
 {
+    int host_page_size = qemu_real_host_page_size();
     abi_ulong starts[3];
     abi_ulong lens[3];
     int prots[3];
@@ -189,13 +190,13 @@ int target_mprotect(abi_ulong start, abi_ulong len, int target_prot)
     }
 
     last = start + len - 1;
-    host_start = start & qemu_host_page_mask;
+    host_start = start & -host_page_size;
     host_last = HOST_PAGE_ALIGN(last) - 1;
     nranges = 0;
 
     mmap_lock();
 
-    if (host_last - host_start < qemu_host_page_size) {
+    if (host_last - host_start < host_page_size) {
         /* Single host page contains all guest pages: sum the prot. */
         prot1 = target_prot;
         for (abi_ulong a = host_start; a < start; a += TARGET_PAGE_SIZE) {
@@ -205,7 +206,7 @@ int target_mprotect(abi_ulong start, abi_ulong len, int target_prot)
             prot1 |= page_get_flags(a + 1);
         }
         starts[nranges] = host_start;
-        lens[nranges] = qemu_host_page_size;
+        lens[nranges] = host_page_size;
         prots[nranges] = prot1;
         nranges++;
     } else {
@@ -218,10 +219,10 @@ int target_mprotect(abi_ulong start, abi_ulong len, int target_prot)
             /* If the resulting sum differs, create a new range. */
             if (prot1 != target_prot) {
                 starts[nranges] = host_start;
-                lens[nranges] = qemu_host_page_size;
+                lens[nranges] = host_page_size;
                 prots[nranges] = prot1;
                 nranges++;
-                host_start += qemu_host_page_size;
+                host_start += host_page_size;
             }
         }
 
@@ -233,9 +234,9 @@ int target_mprotect(abi_ulong start, abi_ulong len, int target_prot)
             }
             /* If the resulting sum differs, create a new range. */
             if (prot1 != target_prot) {
-                host_last -= qemu_host_page_size;
+                host_last -= host_page_size;
                 starts[nranges] = host_last + 1;
-                lens[nranges] = qemu_host_page_size;
+                lens[nranges] = host_page_size;
                 prots[nranges] = prot1;
                 nranges++;
             }
@@ -270,6 +271,7 @@ int target_mprotect(abi_ulong start, abi_ulong len, int target_prot)
 static bool mmap_frag(abi_ulong real_start, abi_ulong start, abi_ulong last,
                       int prot, int flags, int fd, off_t offset)
 {
+    int host_page_size = qemu_real_host_page_size();
     abi_ulong real_last;
     void *host_start;
     int prot_old, prot_new;
@@ -286,7 +288,7 @@ static bool mmap_frag(abi_ulong real_start, abi_ulong start, abi_ulong last,
         return false;
     }
 
-    real_last = real_start + qemu_host_page_size - 1;
+    real_last = real_start + host_page_size - 1;
     host_start = g2h_untagged(real_start);
 
     /* Get the protection of the target pages outside the mapping. */
@@ -304,12 +306,12 @@ static bool mmap_frag(abi_ulong real_start, abi_ulong start, abi_ulong last,
          * outside of the fragment we need to map.  Allocate a new host
          * page to cover, discarding whatever else may have been present.
          */
-        void *p = mmap(host_start, qemu_host_page_size,
+        void *p = mmap(host_start, host_page_size,
                        target_to_host_prot(prot),
                        flags | MAP_ANONYMOUS, -1, 0);
         if (p != host_start) {
             if (p != MAP_FAILED) {
-                munmap(p, qemu_host_page_size);
+                munmap(p, host_page_size);
                 errno = EEXIST;
             }
             return false;
@@ -324,7 +326,7 @@ static bool mmap_frag(abi_ulong real_start, abi_ulong start, abi_ulong last,
     /* Adjust protection to be able to write. */
     if (!(host_prot_old & PROT_WRITE)) {
         host_prot_old |= PROT_WRITE;
-        mprotect(host_start, qemu_host_page_size, host_prot_old);
+        mprotect(host_start, host_page_size, host_prot_old);
     }
 
     /* Read or zero the new guest pages. */
@@ -338,7 +340,7 @@ static bool mmap_frag(abi_ulong real_start, abi_ulong start, abi_ulong last,
 
     /* Put final protection */
     if (host_prot_new != host_prot_old) {
-        mprotect(host_start, qemu_host_page_size, host_prot_new);
+        mprotect(host_start, host_page_size, host_prot_new);
     }
     return true;
 }
@@ -373,17 +375,18 @@ static abi_ulong mmap_find_vma_reserved(abi_ulong start, abi_ulong size,
  */
 abi_ulong mmap_find_vma(abi_ulong start, abi_ulong size, abi_ulong align)
 {
+    int host_page_size = qemu_real_host_page_size();
     void *ptr, *prev;
     abi_ulong addr;
     int wrapped, repeat;
 
-    align = MAX(align, qemu_host_page_size);
+    align = MAX(align, host_page_size);
 
     /* If 'start' == 0, then a default start address is used. */
     if (start == 0) {
         start = mmap_next_start;
     } else {
-        start &= qemu_host_page_mask;
+        start &= -host_page_size;
     }
     start = ROUND_UP(start, align);
 
@@ -492,6 +495,7 @@ abi_ulong mmap_find_vma(abi_ulong start, abi_ulong size, abi_ulong align)
 abi_long target_mmap(abi_ulong start, abi_ulong len, int target_prot,
                      int flags, int fd, off_t offset)
 {
+    int host_page_size = qemu_real_host_page_size();
     abi_ulong ret, last, real_start, real_last, retaddr, host_len;
     abi_ulong passthrough_start = -1, passthrough_last = 0;
     int page_flags;
@@ -537,8 +541,8 @@ abi_long target_mmap(abi_ulong start, abi_ulong len, int target_prot,
         }
     }
 
-    real_start = start & qemu_host_page_mask;
-    host_offset = offset & qemu_host_page_mask;
+    real_start = start & -host_page_size;
+    host_offset = offset & -host_page_size;
 
     /*
      * If the user is asking for the kernel to find a location, do that
@@ -567,8 +571,7 @@ abi_long target_mmap(abi_ulong start, abi_ulong len, int target_prot,
      * may need to truncate file maps at EOF and add extra anonymous pages
      * up to the targets page boundary.
      */
-    if ((qemu_real_host_page_size() < qemu_host_page_size) &&
-        !(flags & MAP_ANONYMOUS)) {
+    if (host_page_size < TARGET_PAGE_SIZE && !(flags & MAP_ANONYMOUS)) {
         struct stat sb;
 
         if (fstat(fd, &sb) == -1) {
@@ -595,11 +598,7 @@ abi_long target_mmap(abi_ulong start, abi_ulong len, int target_prot,
         host_len = HOST_PAGE_ALIGN(host_len);
         host_prot = target_to_host_prot(target_prot);
 
-        /*
-         * Note: we prefer to control the mapping address. It is
-         * especially important if qemu_host_page_size >
-         * qemu_real_host_page_size.
-         */
+        /* Note: we prefer to control the mapping address. */
         p = mmap(g2h_untagged(start), host_len, host_prot,
                  flags | MAP_FIXED | MAP_ANONYMOUS, -1, 0);
         if (p == MAP_FAILED) {
@@ -665,7 +664,7 @@ abi_long target_mmap(abi_ulong start, abi_ulong len, int target_prot,
          * aligned, so we read it
          */
         if (!(flags & MAP_ANONYMOUS) &&
-            (offset & ~qemu_host_page_mask) != (start & ~qemu_host_page_mask)) {
+            (offset & (host_page_size - 1)) != (start & (host_page_size - 1))) {
             /*
              * msync() won't work here, so we return an error if write is
              * possible while it is a shared mapping
@@ -694,7 +693,7 @@ abi_long target_mmap(abi_ulong start, abi_ulong len, int target_prot,
 
         /* handle the start of the mapping */
         if (start > real_start) {
-            if (real_last == real_start + qemu_host_page_size - 1) {
+            if (real_last == real_start + host_page_size - 1) {
                 /* one single host page */
                 if (!mmap_frag(real_start, start, last,
                                target_prot, flags, fd, offset)) {
@@ -703,21 +702,21 @@ abi_long target_mmap(abi_ulong start, abi_ulong len, int target_prot,
                 goto the_end1;
             }
             if (!mmap_frag(real_start, start,
-                           real_start + qemu_host_page_size - 1,
+                           real_start + host_page_size - 1,
                            target_prot, flags, fd, offset)) {
                 goto fail;
             }
-            real_start += qemu_host_page_size;
+            real_start += host_page_size;
         }
         /* handle the end of the mapping */
         if (last < real_last) {
-            abi_ulong real_page = real_last - qemu_host_page_size + 1;
+            abi_ulong real_page = real_last - host_page_size + 1;
             if (!mmap_frag(real_page, real_page, last,
                            target_prot, flags, fd,
                            offset + real_page - start)) {
                 goto fail;
             }
-            real_last -= qemu_host_page_size;
+            real_last -= host_page_size;
         }
 
         /* map the middle (easier) */
@@ -784,6 +783,7 @@ fail:
 
 static int mmap_reserve_or_unmap(abi_ulong start, abi_ulong len)
 {
+    int host_page_size = qemu_real_host_page_size();
     abi_ulong real_start;
     abi_ulong real_last;
     abi_ulong real_len;
@@ -793,7 +793,7 @@ static int mmap_reserve_or_unmap(abi_ulong start, abi_ulong len)
     int prot;
 
     last = start + len - 1;
-    real_start = start & qemu_host_page_mask;
+    real_start = start & -host_page_size;
     real_last = HOST_PAGE_ALIGN(last) - 1;
 
     /*
@@ -802,7 +802,7 @@ static int mmap_reserve_or_unmap(abi_ulong start, abi_ulong len)
      * The single page special case is required for the last page,
      * lest real_start overflow to zero.
      */
-    if (real_last - real_start < qemu_host_page_size) {
+    if (real_last - real_start < host_page_size) {
         prot = 0;
         for (a = real_start; a < start; a += TARGET_PAGE_SIZE) {
             prot |= page_get_flags(a);
@@ -818,14 +818,14 @@ static int mmap_reserve_or_unmap(abi_ulong start, abi_ulong len)
             prot |= page_get_flags(a);
         }
         if (prot != 0) {
-            real_start += qemu_host_page_size;
+            real_start += host_page_size;
         }
 
         for (prot = 0, a = last; a < real_last; a += TARGET_PAGE_SIZE) {
             prot |= page_get_flags(a + 1);
         }
         if (prot != 0) {
-            real_last -= qemu_host_page_size;
+            real_last -= host_page_size;
         }
 
         if (real_last < real_start) {
-- 
2.34.1



  parent reply	other threads:[~2024-03-01 23:13 UTC|newest]

Thread overview: 64+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2024-03-01 23:05 [PATCH 00/60] linux-user and tcg patch queue Richard Henderson
2024-03-01 23:05 ` [PATCH 01/60] linux-user/elfload: Disable core dump if getrlimit fails Richard Henderson
2024-03-01 23:05 ` [PATCH 02/60] linux-user/elfload: Merge init_note_info and fill_note_info Richard Henderson
2024-03-01 23:05 ` [PATCH 03/60] linux-user/elfload: Tidy fill_note_info and struct elf_note_info Richard Henderson
2024-03-01 23:05 ` [PATCH 04/60] linux-user/elfload: Stack allocate struct mm_struct Richard Henderson
2024-03-01 23:05 ` [PATCH 05/60] linux-user/elfload: Latch errno before cleanup in elf_core_dump Richard Henderson
2024-03-01 23:05 ` [PATCH 06/60] linux-user/elfload: Open core file after vma_init Richard Henderson
2024-03-01 23:05 ` [PATCH 07/60] linux-user/elfload: Truncate core file on open Richard Henderson
2024-03-01 23:05 ` [PATCH 08/60] linux-user/elfload: Lock cpu list and mmap during elf_core_dump Richard Henderson
2024-03-01 23:05 ` [PATCH 09/60] linux-user/elfload: Size corefile before opening Richard Henderson
2024-03-01 23:05 ` [PATCH 10/60] linux-user/elfload: Write corefile elf header in one block Richard Henderson
2024-03-01 23:05 ` [PATCH 11/60] linux-user/elfload: Write process memory to core file in larger chunks Richard Henderson
2024-03-01 23:05 ` [PATCH 12/60] linux-user/elfload: Simplify vma_dump_size Richard Henderson
2024-03-01 23:05 ` [PATCH 13/60] linux-user/elfload: Rely on walk_memory_regions for vmas Richard Henderson
2024-03-01 23:05 ` [PATCH 14/60] linux-user/elfload: Unprotect regions before core dump Richard Henderson
2024-03-01 23:05 ` [PATCH 15/60] tcg/aarch64: Apple does not align __int128_t in even registers Richard Henderson
2024-03-01 23:05 ` [PATCH 16/60] accel/tcg: Set can_do_io at at start of lookup_tb_ptr helper Richard Henderson
2024-03-01 23:05 ` [PATCH 17/60] tcg: Avoid double lock if page tables happen to be in mmio memory Richard Henderson
2024-03-01 23:05 ` [PATCH 18/60] accel/tcg: Remove qemu_host_page_size from page_protect/page_unprotect Richard Henderson
2024-03-01 23:05 ` [PATCH 19/60] linux-user: Adjust SVr4 NULL page mapping Richard Henderson
2024-03-01 23:05 ` [PATCH 20/60] linux-user: Remove qemu_host_page_{size, mask} in probe_guest_base Richard Henderson
2024-03-01 23:05 ` [PATCH 21/60] linux-user: Remove qemu_host_page_size from create_elf_tables Richard Henderson
2024-03-01 23:05 ` [PATCH 22/60] linux-user/hppa: Simplify init_guest_commpage Richard Henderson
2024-03-01 23:05 ` [PATCH 23/60] linux-user/nios2: Remove qemu_host_page_size from init_guest_commpage Richard Henderson
2024-03-01 23:05 ` [PATCH 24/60] linux-user/arm: " Richard Henderson
2024-03-01 23:05 ` [PATCH 25/60] linux-user: Remove qemu_host_page_size from elf_core_dump Richard Henderson
2024-03-01 23:05 ` Richard Henderson [this message]
2024-03-01 23:05 ` [PATCH 27/60] linux-user: Remove REAL_HOST_PAGE_ALIGN from mmap.c Richard Henderson
2024-03-01 23:05 ` [PATCH 28/60] linux-user: Remove HOST_PAGE_ALIGN " Richard Henderson
2024-03-01 23:05 ` [PATCH 29/60] migration: Remove qemu_host_page_size Richard Henderson
2024-03-01 23:05 ` [PATCH 30/60] hw/tpm: Remove HOST_PAGE_ALIGN from tpm_ppi_init Richard Henderson
2024-03-01 23:05 ` [PATCH 31/60] softmmu/physmem: Remove qemu_host_page_size Richard Henderson
2024-03-01 23:05 ` [PATCH 32/60] softmmu/physmem: Remove HOST_PAGE_ALIGN Richard Henderson
2024-03-01 23:05 ` [PATCH 33/60] linux-user: Remove qemu_host_page_size from main Richard Henderson
2024-03-01 23:05 ` [PATCH 34/60] linux-user: Split out target_mmap__locked Richard Henderson
2024-03-01 23:05 ` [PATCH 35/60] linux-user: Move some mmap checks outside the lock Richard Henderson
2024-03-01 23:05 ` [PATCH 36/60] linux-user: Fix sub-host-page mmap Richard Henderson
2024-03-01 23:05 ` [PATCH 37/60] linux-user: Split out mmap_end Richard Henderson
2024-03-01 23:05 ` [PATCH 38/60] linux-user: Do early mmap placement only for reserved_va Richard Henderson
2024-03-01 23:05 ` [PATCH 39/60] linux-user: Split out do_munmap Richard Henderson
2024-03-01 23:05 ` [PATCH 40/60] linux-user: Use do_munmap for target_mmap failure Richard Henderson
2024-03-01 23:06 ` [PATCH 41/60] linux-user: Split out mmap_h_eq_g Richard Henderson
2024-03-01 23:06 ` [PATCH 42/60] linux-user: Split out mmap_h_lt_g Richard Henderson
2024-03-01 23:06 ` [PATCH 43/60] linux-user: Split out mmap_h_gt_g Richard Henderson
2024-03-01 23:06 ` [PATCH 44/60] tests/tcg: Remove run-test-mmap-* Richard Henderson
2024-03-01 23:06 ` [PATCH 45/60] tests/tcg: Extend file in linux-madvise.c Richard Henderson
2024-03-01 23:06 ` [PATCH 46/60] *-user: Deprecate and disable -p pagesize Richard Henderson
2024-03-01 23:06 ` [PATCH 47/60] cpu: Remove page_size_init Richard Henderson
2024-03-01 23:06 ` [PATCH 48/60] accel/tcg: Disconnect TargetPageDataNode from page size Richard Henderson
2024-03-01 23:06 ` [PATCH 49/60] linux-user: Allow TARGET_PAGE_BITS_VARY Richard Henderson
2024-03-01 23:06 ` [PATCH 50/60] target/arm: Enable TARGET_PAGE_BITS_VARY for AArch64 user-only Richard Henderson
2024-03-01 23:06 ` [PATCH 51/60] linux-user: Bound mmap_min_addr by host page size Richard Henderson
2024-03-01 23:06 ` [PATCH 52/60] target/ppc: Enable TARGET_PAGE_BITS_VARY for user-only Richard Henderson
2024-03-01 23:06 ` [PATCH 53/60] target/alpha: " Richard Henderson
2024-03-01 23:06 ` [PATCH 54/60] linux-user: Remove pgb_dynamic alignment assertion Richard Henderson
2024-03-01 23:06 ` [PATCH 55/60] tcg/optimize: fix uninitialized variable Richard Henderson
2024-03-01 23:06 ` [PATCH 56/60] linux-user/x86_64: Handle the vsyscall page in open_self_maps_{2, 4} Richard Henderson
2024-04-11 22:09   ` Michael Tokarev
2024-03-01 23:06 ` [PATCH 57/60] linux-user/loongarch64: Remove TARGET_FORCE_SHMLBA Richard Henderson
2024-03-01 23:06 ` [PATCH 58/60] linux-user: Add strace for shmat Richard Henderson
2024-03-01 23:06 ` [PATCH 59/60] linux-user: Rewrite target_shmat Richard Henderson
2024-03-01 23:06 ` [PATCH 60/60] tests/tcg: Check that shmat() does not break /proc/self/maps Richard Henderson
2024-03-01 23:13 ` [PULL 00/60] linux-user and tcg patch queue Richard Henderson
2024-03-05 11:17 ` [PATCH " Peter Maydell

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20240301230619.661008-27-richard.henderson@linaro.org \
    --to=richard.henderson@linaro.org \
    --cc=deller@gmx.de \
    --cc=iii@linux.ibm.com \
    --cc=pierrick.bouvier@linaro.org \
    --cc=qemu-devel@nongnu.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.