All of lore.kernel.org
 help / color / mirror / Atom feed
From: Alex Williamson <alex.williamson@redhat.com>
To: qemu-devel@nongnu.org
Cc: Kirti Wankhede <kwankhede@nvidia.com>, Yan Zhao <yan.y.zhao@intel.com>
Subject: [PULL v2 14/32] vfio: Dirty page tracking when vIOMMU is enabled
Date: Wed, 28 Oct 2020 10:41:36 -0600	[thread overview]
Message-ID: <160390329612.12234.6184919772225463675.stgit@gimli.home> (raw)
In-Reply-To: <160390309510.12234.8858324597971641979.stgit@gimli.home>

From: Kirti Wankhede <kwankhede@nvidia.com>

When vIOMMU is enabled, register MAP notifier from log_sync when all
devices in container are in stop and copy phase of migration. Call replay
and get dirty pages from notifier callback.

Suggested-by: Alex Williamson <alex.williamson@redhat.com>
Signed-off-by: Kirti Wankhede <kwankhede@nvidia.com>
Reviewed-by: Yan Zhao <yan.y.zhao@intel.com>
Signed-off-by: Alex Williamson <alex.williamson@redhat.com>
---
 hw/vfio/common.c     |   88 +++++++++++++++++++++++++++++++++++++++++++++++---
 hw/vfio/trace-events |    1 +
 2 files changed, 83 insertions(+), 6 deletions(-)

diff --git a/hw/vfio/common.c b/hw/vfio/common.c
index 2b1455e78089..32d536e32507 100644
--- a/hw/vfio/common.c
+++ b/hw/vfio/common.c
@@ -442,8 +442,8 @@ static bool vfio_listener_skipped_section(MemoryRegionSection *section)
 }
 
 /* Called with rcu_read_lock held.  */
-static bool vfio_get_vaddr(IOMMUTLBEntry *iotlb, void **vaddr,
-                           bool *read_only)
+static bool vfio_get_xlat_addr(IOMMUTLBEntry *iotlb, void **vaddr,
+                               ram_addr_t *ram_addr, bool *read_only)
 {
     MemoryRegion *mr;
     hwaddr xlat;
@@ -474,8 +474,17 @@ static bool vfio_get_vaddr(IOMMUTLBEntry *iotlb, void **vaddr,
         return false;
     }
 
-    *vaddr = memory_region_get_ram_ptr(mr) + xlat;
-    *read_only = !writable || mr->readonly;
+    if (vaddr) {
+        *vaddr = memory_region_get_ram_ptr(mr) + xlat;
+    }
+
+    if (ram_addr) {
+        *ram_addr = memory_region_get_ram_addr(mr) + xlat;
+    }
+
+    if (read_only) {
+        *read_only = !writable || mr->readonly;
+    }
 
     return true;
 }
@@ -485,7 +494,6 @@ static void vfio_iommu_map_notify(IOMMUNotifier *n, IOMMUTLBEntry *iotlb)
     VFIOGuestIOMMU *giommu = container_of(n, VFIOGuestIOMMU, n);
     VFIOContainer *container = giommu->container;
     hwaddr iova = iotlb->iova + giommu->iommu_offset;
-    bool read_only;
     void *vaddr;
     int ret;
 
@@ -501,7 +509,9 @@ static void vfio_iommu_map_notify(IOMMUNotifier *n, IOMMUTLBEntry *iotlb)
     rcu_read_lock();
 
     if ((iotlb->perm & IOMMU_RW) != IOMMU_NONE) {
-        if (!vfio_get_vaddr(iotlb, &vaddr, &read_only)) {
+        bool read_only;
+
+        if (!vfio_get_xlat_addr(iotlb, &vaddr, NULL, &read_only)) {
             goto out;
         }
         /*
@@ -899,11 +909,77 @@ err_out:
     return ret;
 }
 
+typedef struct {
+    IOMMUNotifier n;
+    VFIOGuestIOMMU *giommu;
+} vfio_giommu_dirty_notifier;
+
+static void vfio_iommu_map_dirty_notify(IOMMUNotifier *n, IOMMUTLBEntry *iotlb)
+{
+    vfio_giommu_dirty_notifier *gdn = container_of(n,
+                                                vfio_giommu_dirty_notifier, n);
+    VFIOGuestIOMMU *giommu = gdn->giommu;
+    VFIOContainer *container = giommu->container;
+    hwaddr iova = iotlb->iova + giommu->iommu_offset;
+    ram_addr_t translated_addr;
+
+    trace_vfio_iommu_map_dirty_notify(iova, iova + iotlb->addr_mask);
+
+    if (iotlb->target_as != &address_space_memory) {
+        error_report("Wrong target AS \"%s\", only system memory is allowed",
+                     iotlb->target_as->name ? iotlb->target_as->name : "none");
+        return;
+    }
+
+    rcu_read_lock();
+    if (vfio_get_xlat_addr(iotlb, NULL, &translated_addr, NULL)) {
+        int ret;
+
+        ret = vfio_get_dirty_bitmap(container, iova, iotlb->addr_mask + 1,
+                                    translated_addr);
+        if (ret) {
+            error_report("vfio_iommu_map_dirty_notify(%p, 0x%"HWADDR_PRIx", "
+                         "0x%"HWADDR_PRIx") = %d (%m)",
+                         container, iova,
+                         iotlb->addr_mask + 1, ret);
+        }
+    }
+    rcu_read_unlock();
+}
+
 static int vfio_sync_dirty_bitmap(VFIOContainer *container,
                                   MemoryRegionSection *section)
 {
     ram_addr_t ram_addr;
 
+    if (memory_region_is_iommu(section->mr)) {
+        VFIOGuestIOMMU *giommu;
+
+        QLIST_FOREACH(giommu, &container->giommu_list, giommu_next) {
+            if (MEMORY_REGION(giommu->iommu) == section->mr &&
+                giommu->n.start == section->offset_within_region) {
+                Int128 llend;
+                vfio_giommu_dirty_notifier gdn = { .giommu = giommu };
+                int idx = memory_region_iommu_attrs_to_index(giommu->iommu,
+                                                       MEMTXATTRS_UNSPECIFIED);
+
+                llend = int128_add(int128_make64(section->offset_within_region),
+                                   section->size);
+                llend = int128_sub(llend, int128_one());
+
+                iommu_notifier_init(&gdn.n,
+                                    vfio_iommu_map_dirty_notify,
+                                    IOMMU_NOTIFIER_MAP,
+                                    section->offset_within_region,
+                                    int128_get64(llend),
+                                    idx);
+                memory_region_iommu_replay(giommu->iommu, &gdn.n);
+                break;
+            }
+        }
+        return 0;
+    }
+
     ram_addr = memory_region_get_ram_addr(section->mr) +
                section->offset_within_region;
 
diff --git a/hw/vfio/trace-events b/hw/vfio/trace-events
index dd991bd8f265..c0e75f24b76d 100644
--- a/hw/vfio/trace-events
+++ b/hw/vfio/trace-events
@@ -164,3 +164,4 @@ vfio_load_state(const char *name, uint64_t data) " (%s) data 0x%"PRIx64
 vfio_load_state_device_data(const char *name, uint64_t data_offset, uint64_t data_size) " (%s) Offset 0x%"PRIx64" size 0x%"PRIx64
 vfio_load_cleanup(const char *name) " (%s)"
 vfio_get_dirty_bitmap(int fd, uint64_t iova, uint64_t size, uint64_t bitmap_size, uint64_t start) "container fd=%d, iova=0x%"PRIx64" size= 0x%"PRIx64" bitmap_size=0x%"PRIx64" start=0x%"PRIx64
+vfio_iommu_map_dirty_notify(uint64_t iova_start, uint64_t iova_end) "iommu dirty @ 0x%"PRIx64" - 0x%"PRIx64



  parent reply	other threads:[~2020-10-28 16:51 UTC|newest]

Thread overview: 35+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2020-10-28 16:39 [PULL v2 00/32] VFIO updates 2020-10-28 (for QEMU 5.2 soft-freeze) Alex Williamson
2020-10-28 16:39 ` [PULL v2 01/32] vfio: Add function to unmap VFIO region Alex Williamson
2020-10-28 16:39 ` [PULL v2 02/32] vfio: Add vfio_get_object callback to VFIODeviceOps Alex Williamson
2020-10-28 16:39 ` [PULL v2 03/32] vfio: Add save and load functions for VFIO PCI devices Alex Williamson
2020-10-28 16:40 ` [PULL v2 04/32] vfio: Add migration region initialization and finalize function Alex Williamson
2020-10-28 16:40 ` [PULL v2 05/32] vfio: Add VM state change handler to know state of VM Alex Williamson
2020-10-28 16:40 ` [PULL v2 06/32] vfio: Add migration state change notifier Alex Williamson
2020-10-28 16:40 ` [PULL v2 07/32] vfio: Register SaveVMHandlers for VFIO device Alex Williamson
2020-10-28 16:40 ` [PULL v2 08/32] vfio: Add save state functions to SaveVMHandlers Alex Williamson
2020-10-28 16:40 ` [PULL v2 09/32] vfio: Add load " Alex Williamson
2020-10-28 16:40 ` [PULL v2 10/32] memory: Set DIRTY_MEMORY_MIGRATION when IOMMU is enabled Alex Williamson
2020-10-28 16:41 ` [PULL v2 11/32] vfio: Get migration capability flags for container Alex Williamson
2020-10-28 16:41 ` [PULL v2 12/32] vfio: Add function to start and stop dirty pages tracking Alex Williamson
2020-10-28 16:41 ` [PULL v2 13/32] vfio: Add vfio_listener_log_sync to mark dirty pages Alex Williamson
2020-10-28 16:41 ` Alex Williamson [this message]
2020-10-28 16:41 ` [PULL v2 15/32] vfio: Add ioctl to get dirty pages bitmap during dma unmap Alex Williamson
2020-10-28 16:41 ` [PULL v2 16/32] vfio: Make vfio-pci device migration capable Alex Williamson
2020-10-28 16:42 ` [PULL v2 17/32] qapi: Add VFIO devices migration stats in Migration stats Alex Williamson
2020-10-28 16:42 ` [PULL v2 18/32] update-linux-headers: Add vfio_zdev.h Alex Williamson
2020-10-28 16:42 ` [PULL v2 19/32] linux-headers: update against 5.10-rc1 Alex Williamson
2020-10-28 16:42 ` [PULL v2 20/32] s390x/pci: Move header files to include/hw/s390x Alex Williamson
2020-10-28 16:42 ` [PULL v2 21/32] vfio: Create shared routine for scanning info capabilities Alex Williamson
2020-10-28 16:42 ` [PULL v2 22/32] vfio: Find DMA available capability Alex Williamson
2020-10-28 16:42 ` [PULL v2 23/32] s390x/pci: Add routine to get the vfio dma available count Alex Williamson
2020-10-28 16:42 ` [PULL v2 24/32] s390x/pci: Honor DMA limits set by vfio Alex Williamson
2020-10-28 16:43 ` [PULL v2 25/32] s390x/pci: create a header dedicated to PCI CLP Alex Williamson
2020-10-28 16:43 ` [PULL v2 26/32] s390x/pci: use a PCI Group structure Alex Williamson
2020-10-28 16:43 ` [PULL v2 27/32] s390x/pci: clean up s390 PCI groups Alex Williamson
2020-10-28 16:43 ` [PULL v2 28/32] s390x/pci: use a PCI Function structure Alex Williamson
2020-10-28 16:43 ` [PULL v2 29/32] vfio: Add routine for finding VFIO_DEVICE_GET_INFO capabilities Alex Williamson
2020-10-28 16:43 ` [PULL v2 30/32] s390x/pci: get zPCI function info from host Alex Williamson
2020-10-28 16:44 ` [PULL v2 31/32] hw/vfio: Use lock guard macros Alex Williamson
2020-10-28 16:44 ` [PULL v2 32/32] vfio: fix incorrect print type Alex Williamson
2020-10-31 14:54 ` [PULL v2 00/32] VFIO updates 2020-10-28 (for QEMU 5.2 soft-freeze) Peter Maydell
2020-11-01 20:46   ` Alex Williamson

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=160390329612.12234.6184919772225463675.stgit@gimli.home \
    --to=alex.williamson@redhat.com \
    --cc=kwankhede@nvidia.com \
    --cc=qemu-devel@nongnu.org \
    --cc=yan.y.zhao@intel.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.