From: Janosch Frank <frankja@linux.ibm.com>
To: kvm@vger.kernel.org
Cc: borntraeger@de.ibm.com, david@redhat.com,
linux-s390@vger.kernel.org, imbrenda@linux.ibm.com
Subject: [PATCH 06/14] s390/mm: Provide vmaddr to pmd notification
Date: Wed, 13 Jan 2021 09:41:05 +0000 [thread overview]
Message-ID: <20210113094113.133668-7-frankja@linux.ibm.com> (raw)
In-Reply-To: <20210113094113.133668-1-frankja@linux.ibm.com>
It will be needed for shadow tables.
Signed-off-by: Janosch Frank <frankja@linux.ibm.com>
---
arch/s390/mm/gmap.c | 52 +++++++++++++++++++++++----------------------
1 file changed, 27 insertions(+), 25 deletions(-)
diff --git a/arch/s390/mm/gmap.c b/arch/s390/mm/gmap.c
index d8c9b295294b..b7199c55f98a 100644
--- a/arch/s390/mm/gmap.c
+++ b/arch/s390/mm/gmap.c
@@ -530,10 +530,10 @@ void gmap_unlink(struct mm_struct *mm, unsigned long *table,
}
static void gmap_pmdp_xchg(struct gmap *gmap, pmd_t *old, pmd_t new,
- unsigned long gaddr);
+ unsigned long gaddr, unsigned long vmaddr);
static void gmap_pmd_split(struct gmap *gmap, unsigned long gaddr,
- pmd_t *pmdp, struct page *page);
+ unsigned long vmaddr, pmd_t *pmdp, struct page *page);
/**
* gmap_link - set up shadow page tables to connect a host to a guest address
@@ -632,7 +632,8 @@ int __gmap_link(struct gmap *gmap, unsigned long gaddr, unsigned long vmaddr)
} else if (*table & _SEGMENT_ENTRY_PROTECT &&
!(pmd_val(*pmd) & _SEGMENT_ENTRY_PROTECT)) {
if (page) {
- gmap_pmd_split(gmap, gaddr, (pmd_t *)table, page);
+ gmap_pmd_split(gmap, gaddr, vmaddr,
+ (pmd_t *)table, page);
page = NULL;
} else {
spin_unlock(ptl);
@@ -952,19 +953,15 @@ static void gmap_pte_op_end(spinlock_t *ptl)
* Returns a pointer to the pmd for a guest address, or NULL
*/
static inline pmd_t *gmap_pmd_op_walk(struct gmap *gmap, unsigned long gaddr,
- spinlock_t **ptl)
+ unsigned long vmaddr, spinlock_t **ptl)
{
pmd_t *pmdp, *hpmdp;
- unsigned long vmaddr;
BUG_ON(gmap_is_shadow(gmap));
*ptl = NULL;
if (gmap->mm->context.allow_gmap_hpage_1m) {
- vmaddr = __gmap_translate(gmap, gaddr);
- if (IS_ERR_VALUE(vmaddr))
- return NULL;
hpmdp = pmd_alloc_map(gmap->mm, vmaddr);
if (!hpmdp)
return NULL;
@@ -1047,7 +1044,7 @@ static inline void gmap_pmd_split_free(struct gmap *gmap, pmd_t *pmdp)
* aren't tracked anywhere else.
*/
static void gmap_pmd_split(struct gmap *gmap, unsigned long gaddr,
- pmd_t *pmdp, struct page *page)
+ unsigned long vmaddr, pmd_t *pmdp, struct page *page)
{
unsigned long *ptable = (unsigned long *) page_to_phys(page);
pmd_t new;
@@ -1069,7 +1066,7 @@ static void gmap_pmd_split(struct gmap *gmap, unsigned long gaddr,
spin_lock(&gmap->split_list_lock);
list_add(&page->lru, &gmap->split_list);
spin_unlock(&gmap->split_list_lock);
- gmap_pmdp_xchg(gmap, pmdp, new, gaddr);
+ gmap_pmdp_xchg(gmap, pmdp, new, gaddr, vmaddr);
}
/*
@@ -1087,7 +1084,8 @@ static void gmap_pmd_split(struct gmap *gmap, unsigned long gaddr,
* guest_table_lock held.
*/
static int gmap_protect_pmd(struct gmap *gmap, unsigned long gaddr,
- pmd_t *pmdp, int prot, unsigned long bits)
+ unsigned long vmaddr, pmd_t *pmdp, int prot,
+ unsigned long bits)
{
int pmd_i = pmd_val(*pmdp) & _SEGMENT_ENTRY_INVALID;
int pmd_p = pmd_val(*pmdp) & _SEGMENT_ENTRY_PROTECT;
@@ -1099,13 +1097,13 @@ static int gmap_protect_pmd(struct gmap *gmap, unsigned long gaddr,
if (prot == PROT_NONE && !pmd_i) {
pmd_val(new) |= _SEGMENT_ENTRY_INVALID;
- gmap_pmdp_xchg(gmap, pmdp, new, gaddr);
+ gmap_pmdp_xchg(gmap, pmdp, new, gaddr, vmaddr);
}
if (prot == PROT_READ && !pmd_p) {
pmd_val(new) &= ~_SEGMENT_ENTRY_INVALID;
pmd_val(new) |= _SEGMENT_ENTRY_PROTECT;
- gmap_pmdp_xchg(gmap, pmdp, new, gaddr);
+ gmap_pmdp_xchg(gmap, pmdp, new, gaddr, vmaddr);
}
if (bits & GMAP_NOTIFY_MPROT)
@@ -1168,10 +1166,14 @@ static int gmap_protect_range(struct gmap *gmap, unsigned long gaddr,
int rc;
BUG_ON(gmap_is_shadow(gmap));
+
while (len) {
rc = -EAGAIN;
-
- pmdp = gmap_pmd_op_walk(gmap, gaddr, &ptl_pmd);
+ vmaddr = __gmap_translate(gmap, gaddr);
+ if (IS_ERR_VALUE(vmaddr))
+ return vmaddr;
+ vmaddr |= gaddr & ~PMD_MASK;
+ pmdp = gmap_pmd_op_walk(gmap, gaddr, vmaddr, &ptl_pmd);
if (pmdp && !(pmd_val(*pmdp) & _SEGMENT_ENTRY_INVALID)) {
if (!pmd_large(*pmdp)) {
ptep = gmap_pte_from_pmd(gmap, pmdp, gaddr,
@@ -1196,7 +1198,7 @@ static int gmap_protect_range(struct gmap *gmap, unsigned long gaddr,
return -ENOMEM;
continue;
} else {
- gmap_pmd_split(gmap, gaddr,
+ gmap_pmd_split(gmap, gaddr, vmaddr,
pmdp, page);
page = NULL;
gmap_pmd_op_end(ptl_pmd);
@@ -1214,9 +1216,6 @@ static int gmap_protect_range(struct gmap *gmap, unsigned long gaddr,
return rc;
/* -EAGAIN, fixup of userspace mm and gmap */
- vmaddr = __gmap_translate(gmap, gaddr);
- if (IS_ERR_VALUE(vmaddr))
- return vmaddr;
rc = gmap_fixup(gmap, gaddr, vmaddr, prot);
if (rc)
return rc;
@@ -2441,6 +2440,7 @@ static inline void pmdp_notify_split(struct gmap *gmap, pmd_t *pmdp,
static void pmdp_notify_gmap(struct gmap *gmap, pmd_t *pmdp,
unsigned long gaddr, unsigned long vmaddr)
{
+ BUG_ON((gaddr & ~HPAGE_MASK) || (vmaddr & ~HPAGE_MASK));
if (gmap_pmd_is_split(pmdp))
return pmdp_notify_split(gmap, pmdp, gaddr, vmaddr);
@@ -2461,10 +2461,11 @@ static void pmdp_notify_gmap(struct gmap *gmap, pmd_t *pmdp,
* held.
*/
static void gmap_pmdp_xchg(struct gmap *gmap, pmd_t *pmdp, pmd_t new,
- unsigned long gaddr)
+ unsigned long gaddr, unsigned long vmaddr)
{
gaddr &= HPAGE_MASK;
- pmdp_notify_gmap(gmap, pmdp, gaddr, 0);
+ vmaddr &= HPAGE_MASK;
+ pmdp_notify_gmap(gmap, pmdp, gaddr, vmaddr);
if (pmd_large(new))
pmd_val(new) &= ~GMAP_SEGMENT_NOTIFY_BITS;
if (MACHINE_HAS_TLB_GUEST)
@@ -2612,7 +2613,8 @@ EXPORT_SYMBOL_GPL(gmap_pmdp_idte_global);
* held.
*/
static bool gmap_test_and_clear_dirty_pmd(struct gmap *gmap, pmd_t *pmdp,
- unsigned long gaddr)
+ unsigned long gaddr,
+ unsigned long vmaddr)
{
if (pmd_val(*pmdp) & _SEGMENT_ENTRY_INVALID)
return false;
@@ -2624,7 +2626,7 @@ static bool gmap_test_and_clear_dirty_pmd(struct gmap *gmap, pmd_t *pmdp,
/* Clear UC indication and reset protection */
pmd_val(*pmdp) &= ~_SEGMENT_ENTRY_GMAP_UC;
- gmap_protect_pmd(gmap, gaddr, pmdp, PROT_READ, 0);
+ gmap_protect_pmd(gmap, gaddr, vmaddr, pmdp, PROT_READ, 0);
return true;
}
@@ -2647,12 +2649,12 @@ void gmap_sync_dirty_log_pmd(struct gmap *gmap, unsigned long bitmap[4],
spinlock_t *ptl_pmd = NULL;
spinlock_t *ptl_pte = NULL;
- pmdp = gmap_pmd_op_walk(gmap, gaddr, &ptl_pmd);
+ pmdp = gmap_pmd_op_walk(gmap, gaddr, vmaddr, &ptl_pmd);
if (!pmdp)
return;
if (pmd_large(*pmdp)) {
- if (gmap_test_and_clear_dirty_pmd(gmap, pmdp, gaddr))
+ if (gmap_test_and_clear_dirty_pmd(gmap, pmdp, gaddr, vmaddr))
bitmap_fill(bitmap, _PAGE_ENTRIES);
} else {
for (i = 0; i < _PAGE_ENTRIES; i++, vmaddr += PAGE_SIZE) {
--
2.27.0
next prev parent reply other threads:[~2021-01-13 9:42 UTC|newest]
Thread overview: 15+ messages / expand[flat|nested] mbox.gz Atom feed top
2021-01-13 9:40 [PATCH 00/14] KVM: s390: Add huge page VSIE support Janosch Frank
2021-01-13 9:41 ` [PATCH 01/14] s390/mm: Code cleanups Janosch Frank
2021-01-13 9:41 ` [PATCH 02/14] s390/mm: Improve locking for huge page backings Janosch Frank
2021-01-13 9:41 ` [PATCH 03/14] s390/mm: Take locking out of gmap_protect_pte Janosch Frank
2021-01-13 9:41 ` [PATCH 04/14] s390/mm: split huge pages in GMAP when protecting Janosch Frank
2021-01-13 9:41 ` [PATCH 05/14] s390/mm: Split huge pages when migrating Janosch Frank
2021-01-13 9:41 ` Janosch Frank [this message]
2021-01-13 9:41 ` [PATCH 07/14] s390/mm: factor out idte global flush into gmap_idte_global Janosch Frank
2021-01-13 9:41 ` [PATCH 08/14] s390/mm: Make gmap_read_table EDAT1 compatible Janosch Frank
2021-01-13 9:41 ` [PATCH 09/14] s390/mm: Make gmap_protect_rmap " Janosch Frank
2021-01-13 9:41 ` [PATCH 10/14] s390/mm: Add simple ptep shadow function Janosch Frank
2021-01-13 9:41 ` [PATCH 11/14] s390/mm: Add gmap shadowing for large pmds Janosch Frank
2021-01-13 9:41 ` [PATCH 12/14] s390/mm: Add gmap lock classes Janosch Frank
2021-01-13 9:41 ` [PATCH 13/14] s390/mm: Pull pmd invalid check in gmap_pmd_op_walk Janosch Frank
2021-01-13 9:41 ` [PATCH 14/14] KVM: s390: Allow the VSIE to be used with huge pages Janosch Frank
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20210113094113.133668-7-frankja@linux.ibm.com \
--to=frankja@linux.ibm.com \
--cc=borntraeger@de.ibm.com \
--cc=david@redhat.com \
--cc=imbrenda@linux.ibm.com \
--cc=kvm@vger.kernel.org \
--cc=linux-s390@vger.kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).