linux-kernel.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: Alex Williamson <alex.williamson@redhat.com>
To: alex.williamson@redhat.com
Cc: Jason Gunthorpe <jgg@nvidia.com>,
	linux-kernel@vger.kernel.org, kvm@vger.kernel.org,
	jgg@nvidia.com, peterx@redhat.com
Subject: [PATCH 4/7] vfio,vfio-pci: Add vma to pfn callback
Date: Thu, 05 Aug 2021 11:07:47 -0600	[thread overview]
Message-ID: <162818326742.1511194.1366505678218237973.stgit@omen> (raw)
In-Reply-To: <162818167535.1511194.6614962507750594786.stgit@omen>

Add a new vfio_device_ops callback to allow the vfio device driver to
translate a vma mapping of a vfio device fd to a pfn.  Implementation
limited to vfio-pci here for the purpose of supporting the reverse of
unmap_mapping_range(), but expected to be implemented for all vfio
device drivers supporting DMA mapping of device memory mmaps.

Suggested-by: Jason Gunthorpe <jgg@nvidia.com>
Signed-off-by: Alex Williamson <alex.williamson@redhat.com>
---
 drivers/vfio/pci/vfio_pci.c |    9 ++++++---
 drivers/vfio/vfio.c         |   18 ++++++++++++++++--
 include/linux/vfio.h        |    6 ++++++
 3 files changed, 28 insertions(+), 5 deletions(-)

diff --git a/drivers/vfio/pci/vfio_pci.c b/drivers/vfio/pci/vfio_pci.c
index c526edbf1173..7a9f67cfc0a2 100644
--- a/drivers/vfio/pci/vfio_pci.c
+++ b/drivers/vfio/pci/vfio_pci.c
@@ -1440,10 +1440,12 @@ void vfio_pci_memory_unlock_and_restore(struct vfio_pci_device *vdev, u16 cmd)
 	up_write(&vdev->memory_lock);
 }
 
-static int vfio_pci_bar_vma_to_pfn(struct vm_area_struct *vma,
+static int vfio_pci_bar_vma_to_pfn(struct vfio_device *core_vdev,
+				   struct vm_area_struct *vma,
 				   unsigned long *pfn)
 {
-	struct vfio_pci_device *vdev = vma->vm_private_data;
+	struct vfio_pci_device *vdev =
+			container_of(core_vdev, struct vfio_pci_device, vdev);
 	struct pci_dev *pdev = vdev->pdev;
 	int index;
 	u64 pgoff;
@@ -1469,7 +1471,7 @@ static vm_fault_t vfio_pci_mmap_fault(struct vm_fault *vmf)
 	unsigned long vaddr, pfn;
 	vm_fault_t ret = VM_FAULT_SIGBUS;
 
-	if (vfio_pci_bar_vma_to_pfn(vma, &pfn))
+	if (vfio_pci_bar_vma_to_pfn(&vdev->vdev, vma, &pfn))
 		return ret;
 
 	down_read(&vdev->memory_lock);
@@ -1742,6 +1744,7 @@ static const struct vfio_device_ops vfio_pci_ops = {
 	.mmap		= vfio_pci_mmap,
 	.request	= vfio_pci_request,
 	.match		= vfio_pci_match,
+	.vma_to_pfn	= vfio_pci_bar_vma_to_pfn,
 };
 
 static int vfio_pci_reflck_attach(struct vfio_pci_device *vdev);
diff --git a/drivers/vfio/vfio.c b/drivers/vfio/vfio.c
index 1e4fc69fee7d..42ca93be152a 100644
--- a/drivers/vfio/vfio.c
+++ b/drivers/vfio/vfio.c
@@ -875,6 +875,22 @@ struct vfio_device *vfio_device_get_from_dev(struct device *dev)
 }
 EXPORT_SYMBOL_GPL(vfio_device_get_from_dev);
 
+static const struct file_operations vfio_device_fops;
+
+int vfio_device_vma_to_pfn(struct vfio_device *device,
+			   struct vm_area_struct *vma, unsigned long *pfn)
+{
+	if (WARN_ON(!vma->vm_file || vma->vm_file->f_op != &vfio_device_fops ||
+		    vma->vm_file->private_data != device))
+		return -EINVAL;
+
+	if (unlikely(!device->ops->vma_to_pfn))
+		return -EPERM;
+
+	return device->ops->vma_to_pfn(device, vma, pfn);
+}
+EXPORT_SYMBOL_GPL(vfio_device_vma_to_pfn);
+
 static struct vfio_device *vfio_device_get_from_name(struct vfio_group *group,
 						     char *buf)
 {
@@ -1407,8 +1423,6 @@ static int vfio_group_add_container_user(struct vfio_group *group)
 	return 0;
 }
 
-static const struct file_operations vfio_device_fops;
-
 static int vfio_group_get_device_fd(struct vfio_group *group, char *buf)
 {
 	struct vfio_device *device;
diff --git a/include/linux/vfio.h b/include/linux/vfio.h
index 712813703e5a..5f07ebe0f85d 100644
--- a/include/linux/vfio.h
+++ b/include/linux/vfio.h
@@ -41,6 +41,7 @@ struct vfio_device {
  * @match: Optional device name match callback (return: 0 for no-match, >0 for
  *         match, -errno for abort (ex. match with insufficient or incorrect
  *         additional args)
+ * @vma_to_pfn: Optional pfn from vma lookup against vma mapping device fd
  */
 struct vfio_device_ops {
 	char	*name;
@@ -55,6 +56,8 @@ struct vfio_device_ops {
 	int	(*mmap)(struct vfio_device *vdev, struct vm_area_struct *vma);
 	void	(*request)(struct vfio_device *vdev, unsigned int count);
 	int	(*match)(struct vfio_device *vdev, char *buf);
+	int	(*vma_to_pfn)(struct vfio_device *vdev,
+			      struct vm_area_struct *vma, unsigned long *pfn);
 };
 
 extern struct iommu_group *vfio_iommu_group_get(struct device *dev);
@@ -68,6 +71,9 @@ extern struct vfio_device *vfio_device_get_from_dev(struct device *dev);
 extern void vfio_device_put(struct vfio_device *device);
 extern void vfio_device_unmap_mapping_range(struct vfio_device *device,
 					    loff_t start, loff_t len);
+extern int vfio_device_vma_to_pfn(struct vfio_device *device,
+				  struct vm_area_struct *vma,
+				  unsigned long *pfn);
 
 /* events for the backend driver notify callback */
 enum vfio_iommu_notify_type {



  parent reply	other threads:[~2021-08-05 17:08 UTC|newest]

Thread overview: 38+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2021-08-05 17:06 [PATCH 0/7] vfio: device fd address space and vfio-pci mmap invalidation cleanup Alex Williamson
2021-08-05 17:07 ` [PATCH 1/7] vfio: Create vfio_fs_type with inode per device Alex Williamson
2021-08-10  8:43   ` Christoph Hellwig
2021-08-10 14:52     ` Alex Williamson
2021-08-10 14:57       ` Christoph Hellwig
2021-08-10 18:49         ` Peter Xu
2021-08-10 21:16           ` Alex Williamson
2021-08-10 22:18             ` Peter Xu
2021-08-05 17:07 ` [PATCH 2/7] vfio: Export unmap_mapping_range() wrapper Alex Williamson
2021-08-10  8:45   ` Christoph Hellwig
2021-08-10 18:56   ` Peter Xu
2021-08-05 17:07 ` [PATCH 3/7] vfio/pci: Use vfio_device_unmap_mapping_range() Alex Williamson
2021-08-06  1:04   ` Jason Gunthorpe
2021-08-06 20:17     ` Alex Williamson
2021-08-10  8:51       ` Christoph Hellwig
2021-08-10 11:57         ` Jason Gunthorpe
2021-08-10 12:55           ` Christoph Hellwig
2021-08-10 21:50             ` Alex Williamson
2021-08-11 11:57               ` Jason Gunthorpe
2021-08-10  8:53   ` Christoph Hellwig
2021-08-10 19:02     ` Peter Xu
2021-08-10 20:51       ` Alex Williamson
2021-08-10 18:48   ` Peter Xu
2021-08-10 19:59     ` Alex Williamson
2021-08-10 20:20       ` Peter Xu
2021-08-05 17:07 ` Alex Williamson [this message]
2021-08-06  1:01   ` [PATCH 4/7] vfio,vfio-pci: Add vma to pfn callback Jason Gunthorpe
2021-08-10  9:00     ` Christoph Hellwig
2021-08-10  9:00   ` Christoph Hellwig
2021-08-05 17:08 ` [PATCH 5/7] mm/interval_tree.c: Export vma interval tree iterators Alex Williamson
2021-08-05 17:08 ` [PATCH 6/7] vfio: Add vfio_device_io_remap_mapping_range() Alex Williamson
2021-08-10  9:04   ` Christoph Hellwig
2021-08-05 17:08 ` [PATCH 7/7] vfio/pci: Remove map-on-fault behavior Alex Williamson
2021-08-10  9:11   ` Christoph Hellwig
2021-08-10 15:04     ` Alex Williamson
2021-08-10 20:54   ` Peter Xu
2021-08-10 21:45     ` Alex Williamson
2021-08-10 22:27       ` Peter Xu

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=162818326742.1511194.1366505678218237973.stgit@omen \
    --to=alex.williamson@redhat.com \
    --cc=jgg@nvidia.com \
    --cc=kvm@vger.kernel.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=peterx@redhat.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).