From: Eli Cohen <elic@nvidia.com>
To: Jason Wang <jasowang@redhat.com>
Cc: <mst@redhat.com>, <eperezma@redhat.com>, <kvm@vger.kernel.org>,
<virtualization@lists.linux-foundation.org>,
<netdev@vger.kernel.org>, <linux-kernel@vger.kernel.org>,
<lulu@redhat.com>, <eli@mellanox.com>, <lingshan.zhu@intel.com>,
<rob.miller@broadcom.com>, <stefanha@redhat.com>,
<sgarzare@redhat.com>
Subject: Re: [PATCH 07/21] vdpa: multiple address spaces support
Date: Tue, 29 Dec 2020 09:28:32 +0200 [thread overview]
Message-ID: <20201229072832.GA195479@mtl-vdi-166.wap.labs.mlnx> (raw)
In-Reply-To: <20201216064818.48239-8-jasowang@redhat.com>
On Wed, Dec 16, 2020 at 02:48:04PM +0800, Jason Wang wrote:
> This patches introduces the multiple address spaces support for vDPA
> device. This idea is to identify a specific address space via an
> dedicated identifier - ASID.
>
> During vDPA device allocation, vDPA device driver needs to report the
> number of address spaces supported by the device then the DMA mapping
> ops of the vDPA device needs to be extended to support ASID.
>
> This helps to isolate the environments for the virtqueue that will not
> be assigned directly. E.g in the case of virtio-net, the control
> virtqueue will not be assigned directly to guest.
>
> As a start, simply claim 1 virtqueue groups and 1 address spaces for
> all vDPA devices. And vhost-vDPA will simply reject the device with
> more than 1 virtqueue groups or address spaces.
>
> Signed-off-by: Jason Wang <jasowang@redhat.com>
> ---
> drivers/vdpa/ifcvf/ifcvf_main.c | 2 +-
> drivers/vdpa/mlx5/net/mlx5_vnet.c | 5 +++--
> drivers/vdpa/vdpa.c | 4 +++-
> drivers/vdpa/vdpa_sim/vdpa_sim.c | 10 ++++++----
> drivers/vhost/vdpa.c | 14 +++++++++-----
> include/linux/vdpa.h | 23 ++++++++++++++++-------
> 6 files changed, 38 insertions(+), 20 deletions(-)
>
> diff --git a/drivers/vdpa/ifcvf/ifcvf_main.c b/drivers/vdpa/ifcvf/ifcvf_main.c
> index c629f4fcc738..8a43f562b169 100644
> --- a/drivers/vdpa/ifcvf/ifcvf_main.c
> +++ b/drivers/vdpa/ifcvf/ifcvf_main.c
> @@ -445,7 +445,7 @@ static int ifcvf_probe(struct pci_dev *pdev, const struct pci_device_id *id)
>
> adapter = vdpa_alloc_device(struct ifcvf_adapter, vdpa,
> dev, &ifc_vdpa_ops,
> - IFCVF_MAX_QUEUE_PAIRS * 2, 1);
> + IFCVF_MAX_QUEUE_PAIRS * 2, 1, 1);
>
> if (adapter == NULL) {
> IFCVF_ERR(pdev, "Failed to allocate vDPA structure");
> diff --git a/drivers/vdpa/mlx5/net/mlx5_vnet.c b/drivers/vdpa/mlx5/net/mlx5_vnet.c
> index 719b52fcc547..7aaf0a4ee80d 100644
> --- a/drivers/vdpa/mlx5/net/mlx5_vnet.c
> +++ b/drivers/vdpa/mlx5/net/mlx5_vnet.c
> @@ -1804,7 +1804,8 @@ static u32 mlx5_vdpa_get_generation(struct vdpa_device *vdev)
> return mvdev->generation;
> }
>
> -static int mlx5_vdpa_set_map(struct vdpa_device *vdev, struct vhost_iotlb *iotlb)
> +static int mlx5_vdpa_set_map(struct vdpa_device *vdev, unsigned int asid,
> + struct vhost_iotlb *iotlb)
> {
> struct mlx5_vdpa_dev *mvdev = to_mvdev(vdev);
> struct mlx5_vdpa_net *ndev = to_mlx5_vdpa_ndev(mvdev);
> @@ -1947,7 +1948,7 @@ void *mlx5_vdpa_add_dev(struct mlx5_core_dev *mdev)
> max_vqs = min_t(u32, max_vqs, MLX5_MAX_SUPPORTED_VQS);
>
> ndev = vdpa_alloc_device(struct mlx5_vdpa_net, mvdev.vdev, mdev->device, &mlx5_vdpa_ops,
> - 2 * mlx5_vdpa_max_qps(max_vqs), 1);
> + 2 * mlx5_vdpa_max_qps(max_vqs), 1, 1);
> if (IS_ERR(ndev))
> return ndev;
>
> diff --git a/drivers/vdpa/vdpa.c b/drivers/vdpa/vdpa.c
> index 46399746ec7c..05195fa7865d 100644
> --- a/drivers/vdpa/vdpa.c
> +++ b/drivers/vdpa/vdpa.c
> @@ -63,6 +63,7 @@ static void vdpa_release_dev(struct device *d)
> * @config: the bus operations that is supported by this device
> * @nvqs: number of virtqueues supported by this device
> * @ngroups: number of groups supported by this device
> + * @nas: number of address spaces supported by this device
> * @size: size of the parent structure that contains private data
> *
> * Driver should use vdpa_alloc_device() wrapper macro instead of
> @@ -74,7 +75,7 @@ static void vdpa_release_dev(struct device *d)
> struct vdpa_device *__vdpa_alloc_device(struct device *parent,
> const struct vdpa_config_ops *config,
> int nvqs, unsigned int ngroups,
> - size_t size)
> + unsigned int nas, size_t size)
> {
> struct vdpa_device *vdev;
> int err = -EINVAL;
> @@ -102,6 +103,7 @@ struct vdpa_device *__vdpa_alloc_device(struct device *parent,
> vdev->features_valid = false;
> vdev->nvqs = nvqs;
> vdev->ngroups = ngroups;
> + vdev->nas = nas;
>
> err = dev_set_name(&vdev->dev, "vdpa%u", vdev->index);
> if (err)
> diff --git a/drivers/vdpa/vdpa_sim/vdpa_sim.c b/drivers/vdpa/vdpa_sim/vdpa_sim.c
> index 5d554b3cd152..140de45ffff2 100644
> --- a/drivers/vdpa/vdpa_sim/vdpa_sim.c
> +++ b/drivers/vdpa/vdpa_sim/vdpa_sim.c
> @@ -359,7 +359,7 @@ static struct vdpasim *vdpasim_create(void)
> ops = &vdpasim_net_config_ops;
>
> vdpasim = vdpa_alloc_device(struct vdpasim, vdpa, NULL, ops,
> - VDPASIM_VQ_NUM, 1);
> + VDPASIM_VQ_NUM, 1, 1);
> if (!vdpasim)
> goto err_alloc;
>
> @@ -606,7 +606,7 @@ static struct vdpa_iova_range vdpasim_get_iova_range(struct vdpa_device *vdpa)
> return range;
> }
>
> -static int vdpasim_set_map(struct vdpa_device *vdpa,
> +static int vdpasim_set_map(struct vdpa_device *vdpa, unsigned int asid,
> struct vhost_iotlb *iotlb)
> {
> struct vdpasim *vdpasim = vdpa_to_sim(vdpa);
> @@ -633,7 +633,8 @@ static int vdpasim_set_map(struct vdpa_device *vdpa,
> return ret;
> }
>
> -static int vdpasim_dma_map(struct vdpa_device *vdpa, u64 iova, u64 size,
> +static int vdpasim_dma_map(struct vdpa_device *vdpa, unsigned int asid,
> + u64 iova, u64 size,
> u64 pa, u32 perm)
> {
> struct vdpasim *vdpasim = vdpa_to_sim(vdpa);
> @@ -647,7 +648,8 @@ static int vdpasim_dma_map(struct vdpa_device *vdpa, u64 iova, u64 size,
> return ret;
> }
>
> -static int vdpasim_dma_unmap(struct vdpa_device *vdpa, u64 iova, u64 size)
> +static int vdpasim_dma_unmap(struct vdpa_device *vdpa, unsigned int asid,
> + u64 iova, u64 size)
> {
> struct vdpasim *vdpasim = vdpa_to_sim(vdpa);
>
> diff --git a/drivers/vhost/vdpa.c b/drivers/vhost/vdpa.c
> index 9bcc03d4e68b..03a9b3311c6c 100644
> --- a/drivers/vhost/vdpa.c
> +++ b/drivers/vhost/vdpa.c
> @@ -570,10 +570,10 @@ static int vhost_vdpa_map(struct vhost_vdpa *v, struct vhost_iotlb *iotlb,
> return r;
>
> if (ops->dma_map) {
> - r = ops->dma_map(vdpa, iova, size, pa, perm);
> + r = ops->dma_map(vdpa, 0, iova, size, pa, perm);
> } else if (ops->set_map) {
> if (!v->in_batch)
> - r = ops->set_map(vdpa, iotlb);
> + r = ops->set_map(vdpa, 0, iotlb);
> } else {
> r = iommu_map(v->domain, iova, pa, size,
> perm_to_iommu_flags(perm));
> @@ -597,10 +597,10 @@ static void vhost_vdpa_unmap(struct vhost_vdpa *v,
> vhost_vdpa_iotlb_unmap(v, iotlb, iova, iova + size - 1);
>
> if (ops->dma_map) {
> - ops->dma_unmap(vdpa, iova, size);
> + ops->dma_unmap(vdpa, 0, iova, size);
> } else if (ops->set_map) {
> if (!v->in_batch)
> - ops->set_map(vdpa, iotlb);
> + ops->set_map(vdpa, 0, iotlb);
> } else {
> iommu_unmap(v->domain, iova, size);
> }
> @@ -764,7 +764,7 @@ static int vhost_vdpa_process_iotlb_msg(struct vhost_dev *dev,
> break;
> case VHOST_IOTLB_BATCH_END:
> if (v->in_batch && ops->set_map)
> - ops->set_map(vdpa, iotlb);
> + ops->set_map(vdpa, 0, iotlb);
> v->in_batch = false;
> break;
> default:
> @@ -1032,6 +1032,10 @@ static int vhost_vdpa_probe(struct vdpa_device *vdpa)
> int minor;
> int r;
>
> + /* Only support 1 address space and 1 groups */
> + if (vdpa->ngroups != 1 || vdpa->nas != 1)
> + return -ENOTSUPP;
> +
> /* Currently, we only accept the network devices. */
> if (ops->get_device_id(vdpa) != VIRTIO_ID_NET)
> return -ENOTSUPP;
> diff --git a/include/linux/vdpa.h b/include/linux/vdpa.h
> index bfc6790b263e..0a9a754f8180 100644
> --- a/include/linux/vdpa.h
> +++ b/include/linux/vdpa.h
> @@ -43,6 +43,8 @@ struct vdpa_vq_state {
> * @index: device index
> * @features_valid: were features initialized? for legacy guests
> * @nvqs: the number of virtqueues
> + * @ngroups: the number of virtqueue groups
> + * @nas: the number of address spaces
I am not sure these can be categorised as part of the state of the VQ.
It's more of a property so maybe we can have a callback to get the
properties of the VQ?
Moreover, I think they should be handled in the hardware drivers to
return 1 for both ngroups and nas.
> */
> struct vdpa_device {
> struct device dev;
> @@ -52,6 +54,7 @@ struct vdpa_device {
> bool features_valid;
> int nvqs;
> unsigned int ngroups;
> + unsigned int nas;
> };
>
> /**
> @@ -175,6 +178,7 @@ struct vdpa_iova_range {
> * Needed for device that using device
> * specific DMA translation (on-chip IOMMU)
> * @vdev: vdpa device
> + * @asid: address space identifier
> * @iotlb: vhost memory mapping to be
> * used by the vDPA
> * Returns integer: success (0) or error (< 0)
> @@ -183,6 +187,7 @@ struct vdpa_iova_range {
> * specific DMA translation (on-chip IOMMU)
> * and preferring incremental map.
> * @vdev: vdpa device
> + * @asid: address space identifier
> * @iova: iova to be mapped
> * @size: size of the area
> * @pa: physical address for the map
> @@ -194,6 +199,7 @@ struct vdpa_iova_range {
> * specific DMA translation (on-chip IOMMU)
> * and preferring incremental unmap.
> * @vdev: vdpa device
> + * @asid: address space identifier
> * @iova: iova to be unmapped
> * @size: size of the area
> * Returns integer: success (0) or error (< 0)
> @@ -240,10 +246,12 @@ struct vdpa_config_ops {
> struct vdpa_iova_range (*get_iova_range)(struct vdpa_device *vdev);
>
> /* DMA ops */
> - int (*set_map)(struct vdpa_device *vdev, struct vhost_iotlb *iotlb);
> - int (*dma_map)(struct vdpa_device *vdev, u64 iova, u64 size,
> - u64 pa, u32 perm);
> - int (*dma_unmap)(struct vdpa_device *vdev, u64 iova, u64 size);
> + int (*set_map)(struct vdpa_device *vdev, unsigned int asid,
> + struct vhost_iotlb *iotlb);
> + int (*dma_map)(struct vdpa_device *vdev, unsigned int asid,
> + u64 iova, u64 size, u64 pa, u32 perm);
> + int (*dma_unmap)(struct vdpa_device *vdev, unsigned int asid,
> + u64 iova, u64 size);
>
> /* Free device resources */
> void (*free)(struct vdpa_device *vdev);
> @@ -252,11 +260,12 @@ struct vdpa_config_ops {
> struct vdpa_device *__vdpa_alloc_device(struct device *parent,
> const struct vdpa_config_ops *config,
> int nvqs, unsigned int ngroups,
> - size_t size);
> + unsigned int nas, size_t size);
>
> -#define vdpa_alloc_device(dev_struct, member, parent, config, nvqs, ngroups) \
> +#define vdpa_alloc_device(dev_struct, member, parent, config, nvqs, \
> + ngroups, nas) \
> container_of(__vdpa_alloc_device( \
> - parent, config, nvqs, ngroups, \
> + parent, config, nvqs, ngroups, nas, \
> sizeof(dev_struct) + \
> BUILD_BUG_ON_ZERO(offsetof( \
> dev_struct, member))), \
> --
> 2.25.1
>
next prev parent reply other threads:[~2020-12-29 7:30 UTC|newest]
Thread overview: 92+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-12-16 6:47 [PATCH 00/21] Control VQ support in vDPA Jason Wang
2020-12-16 6:47 ` [PATCH 01/21] vhost: move the backend feature bits to vhost_types.h Jason Wang
2020-12-16 6:47 ` [PATCH 02/21] virtio-vdpa: don't set callback if virtio doesn't need it Jason Wang
2020-12-16 6:48 ` [PATCH 03/21] vhost-vdpa: passing iotlb to IOMMU mapping helpers Jason Wang
2020-12-16 6:48 ` [PATCH 04/21] vhost-vdpa: switch to use vhost-vdpa specific IOTLB Jason Wang
2020-12-16 6:48 ` [PATCH 05/21] vdpa: add the missing comment for nvqs in struct vdpa_device Jason Wang
2020-12-16 6:48 ` [PATCH 06/21] vdpa: introduce virtqueue groups Jason Wang
2021-01-04 10:04 ` Stefan Hajnoczi
2021-01-05 4:13 ` Jason Wang
2020-12-16 6:48 ` [PATCH 07/21] vdpa: multiple address spaces support Jason Wang
2020-12-29 7:28 ` Eli Cohen [this message]
2020-12-30 4:00 ` Jason Wang
2020-12-30 4:04 ` Jason Wang
2020-12-30 9:44 ` Eli Cohen
2020-12-16 6:48 ` [PATCH 08/21] vdpa: introduce config operations for associating ASID to a virtqueue group Jason Wang
2020-12-16 6:48 ` [PATCH 09/21] vhost_iotlb: split out IOTLB initialization Jason Wang
2020-12-16 6:48 ` [PATCH 10/21] vhost: support ASID in IOTLB API Jason Wang
2020-12-29 10:20 ` Eli Cohen
2020-12-30 4:27 ` Jason Wang
2020-12-16 6:48 ` [PATCH 11/21] vhost-vdpa: introduce asid based IOTLB Jason Wang
2020-12-29 11:41 ` Eli Cohen
2020-12-30 6:23 ` Jason Wang
2020-12-29 11:53 ` Eli Cohen
2020-12-30 6:34 ` Jason Wang
2020-12-29 12:05 ` Eli Cohen
2020-12-30 6:33 ` Jason Wang
2020-12-16 6:48 ` [PATCH 12/21] vhost-vdpa: introduce uAPI to get the number of virtqueue groups Jason Wang
2020-12-29 12:24 ` Eli Cohen
2020-12-30 6:49 ` Jason Wang
2020-12-30 10:05 ` Eli Cohen
2020-12-31 2:36 ` Jason Wang
2020-12-16 6:48 ` [PATCH 13/21] vhost-vdpa: introduce uAPI to get the number of address spaces Jason Wang
2020-12-16 6:48 ` [PATCH 14/21] vhost-vdpa: uAPI to get virtqueue group id Jason Wang
2020-12-16 6:48 ` [PATCH 15/21] vhost-vdpa: introduce uAPI to set group ASID Jason Wang
2020-12-16 6:48 ` [PATCH 16/21] vhost-vdpa: support ASID based IOTLB API Jason Wang
2020-12-16 6:48 ` [PATCH 17/21] vdpa_sim: split vdpasim_virtqueue's iov field in out_iov and in_iov Jason Wang
2020-12-16 6:48 ` [PATCH 18/21] vdpa_sim: advertise VIRTIO_NET_F_MTU Jason Wang
2020-12-16 6:48 ` [PATCH 19/21] vdpa_sim: factor out buffer completion logic Jason Wang
2020-12-16 6:48 ` [PATCH 20/21] vdpa_sim: filter destination mac address Jason Wang
2020-12-16 6:48 ` [PATCH 21/21] vdpasim: control virtqueue support Jason Wang
2020-12-17 20:19 ` kernel test robot
2021-01-11 12:26 ` Eli Cohen
2021-01-12 3:11 ` Jason Wang
2021-01-22 19:43 ` Eugenio Perez Martin
2021-01-25 3:16 ` Jason Wang
2020-12-16 9:47 ` [PATCH 00/21] Control VQ support in vDPA Michael S. Tsirkin
2020-12-17 3:30 ` Jason Wang
2020-12-17 7:58 ` Michael S. Tsirkin
2020-12-17 9:02 ` Jason Wang
2020-12-17 22:28 ` Michael S. Tsirkin
2020-12-18 2:56 ` Jason Wang
2020-12-17 7:26 ` Eli Cohen
2022-02-24 21:22 ` [RFC PATCH v2 00/19] " Gautam Dawar
2022-02-24 21:22 ` [RFC PATCH v2 01/19] vhost: move the backend feature bits to vhost_types.h Gautam Dawar
2022-02-24 21:22 ` [RFC PATCH v2 02/19] virtio-vdpa: don't set callback if virtio doesn't need it Gautam Dawar
2022-02-24 21:22 ` [RFC PATCH v2 03/19] vhost-vdpa: passing iotlb to IOMMU mapping helpers Gautam Dawar
2022-02-24 21:22 ` [RFC PATCH v2 04/19] vhost-vdpa: switch to use vhost-vdpa specific IOTLB Gautam Dawar
2022-02-24 21:22 ` [RFC PATCH v2 05/19] vdpa: introduce virtqueue groups Gautam Dawar
2022-02-28 8:07 ` Jason Wang
2022-02-28 10:57 ` Gautam Dawar
2022-02-24 21:22 ` [RFC PATCH v2 06/19] vdpa: multiple address spaces support Gautam Dawar
2022-03-03 19:39 ` Eugenio Perez Martin
2022-03-04 6:30 ` Gautam Dawar
2022-03-04 17:45 ` Eugenio Perez Martin
2022-02-24 21:22 ` [RFC PATCH v2 07/19] vdpa: introduce config operations for associating ASID to a virtqueue group Gautam Dawar
2022-03-04 9:54 ` Eugenio Perez Martin
2022-03-04 17:48 ` Gautam Dawar
2022-02-24 21:22 ` [RFC PATCH v2 08/19] vhost_iotlb: split out IOTLB initialization Gautam Dawar
2022-02-24 21:22 ` [RFC PATCH v2 09/19] vhost: support ASID in IOTLB API Gautam Dawar
2022-03-04 10:25 ` Eugenio Perez Martin
2022-03-04 17:52 ` Gautam Dawar
2022-02-24 21:22 ` [RFC PATCH v2 10/19] vhost-vdpa: introduce asid based IOTLB Gautam Dawar
2022-03-04 17:56 ` Eugenio Perez Martin
2022-03-07 10:07 ` Gautam Dawar
2022-02-24 21:22 ` [RFC PATCH v2 11/19] vhost-vdpa: introduce uAPI to get the number of virtqueue groups Gautam Dawar
2022-02-24 21:22 ` [RFC PATCH v2 12/19] vhost-vdpa: introduce uAPI to get the number of address spaces Gautam Dawar
2022-02-24 21:22 ` [RFC PATCH v2 13/19] vhost-vdpa: uAPI to get virtqueue group id Gautam Dawar
2022-02-24 21:22 ` [RFC PATCH v2 14/19] vhost-vdpa: introduce uAPI to set group ASID Gautam Dawar
2022-02-24 21:22 ` [RFC PATCH v2 15/19] vhost-vdpa: support ASID based IOTLB API Gautam Dawar
2022-03-04 18:04 ` Eugenio Perez Martin
2022-03-07 10:23 ` Gautam Dawar
2022-02-24 21:22 ` [RFC PATCH v2 16/19] vdpa_sim: advertise VIRTIO_NET_F_MTU Gautam Dawar
2022-03-10 17:55 ` Eugenio Perez Martin
2022-02-24 21:22 ` [RFC PATCH v2 17/19] vdpa_sim: factor out buffer completion logic Gautam Dawar
2022-02-24 21:22 ` [RFC PATCH v2 18/19] vdpa_sim: filter destination mac address Gautam Dawar
2022-03-10 18:22 ` Eugenio Perez Martin
2022-02-24 21:22 ` [RFC PATCH v2 19/19] vdpasim: control virtqueue support Gautam Dawar
2022-03-10 18:20 ` Eugenio Perez Martin
2022-03-18 7:35 ` Eugenio Perez Martin
2022-03-22 8:46 ` Gautam Dawar
2022-02-28 8:17 ` [RFC PATCH v2 00/19] Control VQ support in vDPA Jason Wang
2022-02-28 10:56 ` Gautam Dawar
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20201229072832.GA195479@mtl-vdi-166.wap.labs.mlnx \
--to=elic@nvidia.com \
--cc=eli@mellanox.com \
--cc=eperezma@redhat.com \
--cc=jasowang@redhat.com \
--cc=kvm@vger.kernel.org \
--cc=lingshan.zhu@intel.com \
--cc=linux-kernel@vger.kernel.org \
--cc=lulu@redhat.com \
--cc=mst@redhat.com \
--cc=netdev@vger.kernel.org \
--cc=rob.miller@broadcom.com \
--cc=sgarzare@redhat.com \
--cc=stefanha@redhat.com \
--cc=virtualization@lists.linux-foundation.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).