linux-kernel.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
* [RFC] vfio iommu type1: improve memory pinning process for raw PFN mapping
@ 2018-02-24  5:44 jason
  2018-02-26 19:19 ` Alex Williamson
  0 siblings, 1 reply; 5+ messages in thread
From: jason @ 2018-02-24  5:44 UTC (permalink / raw)
  To: jason, alex.williamson, pbonzini, kvm, linux-kernel, linux-mm; +Cc: gnehzuil

When using vfio to pass through a PCIe device (e.g. a GPU card) that
has a huge BAR (e.g. 16GB), a lot of cycles are wasted on memory
pinning because PFNs of PCI BAR are not backed by struct page, and
the corresponding VMA has flags VM_IO|VM_PFNMAP.

With this change, memory pinning process will firstly try to figure
out whether the corresponding region is a raw PFN mapping, and if so
it can skip unnecessary user memory pinning process.

Even though it commes with a little overhead, finding vma and testing
flags, on each call, it can significantly improve VM's boot up time
when passing through devices via VFIO.
---
 drivers/vfio/vfio_iommu_type1.c | 22 ++++++++++++++++++++++
 1 file changed, 22 insertions(+)

diff --git a/drivers/vfio/vfio_iommu_type1.c b/drivers/vfio/vfio_iommu_type1.c
index e30e29ae4819..1a471ece3f9c 100644
--- a/drivers/vfio/vfio_iommu_type1.c
+++ b/drivers/vfio/vfio_iommu_type1.c
@@ -374,6 +374,24 @@ static int vaddr_get_pfn(struct mm_struct *mm, unsigned long vaddr,
        return ret;
 }

+static int try_io_pfnmap(struct mm_struct *mm, unsigned long vaddr, long npage,
+                        unsigned long *pfn)
+{
+       struct vm_area_struct *vma;
+       int pinned = 0;
+
+       down_read(&mm->mmap_sem);
+       vma = find_vma_intersection(mm, vaddr, vaddr + 1);
+       if (vma && vma->vm_flags & (VM_IO | VM_PFNMAP)) {
+               *pfn = ((vaddr - vma->vm_start) >> PAGE_SHIFT) + vma->vm_pgoff;
+               if (is_invalid_reserved_pfn(*pfn))
+                       pinned = min(npage, (long)vma_pages(vma));
+       }
+       up_read(&mm->mmap_sem);
+
+       return pinned;
+}
+
 /*
  * Attempt to pin pages.  We really don't want to track all the pfns and
  * the iommu can only map chunks of consecutive pfns anyway, so get the
@@ -392,6 +410,10 @@ static long vfio_pin_pages_remote(struct vfio_dma *dma, unsigned long vaddr,
        if (!current->mm)
                return -ENODEV;

+       ret = try_io_pfnmap(current->mm, vaddr, npage, pfn_base);
+       if (ret)
+               return ret;
+
        ret = vaddr_get_pfn(current->mm, vaddr, dma->prot, pfn_base);
        if (ret)
                return ret;
--
2.13.6

^ permalink raw reply related	[flat|nested] 5+ messages in thread

end of thread, other threads:[~2018-03-12 22:06 UTC | newest]

Thread overview: 5+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2018-02-24  5:44 [RFC] vfio iommu type1: improve memory pinning process for raw PFN mapping jason
2018-02-26 19:19 ` Alex Williamson
2018-02-27  7:44   ` Jason Cai (Xiang Feng)
2018-03-03 12:10   ` [RFC v2] " Jason Cai (Xiang Feng)
2018-03-12 22:06     ` Alex Williamson

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).