From: Mike Christie <michael.christie@oracle.com> To: martin.petersen@oracle.com, linux-scsi@vger.kernel.org, target-devel@vger.kernel.org, mst@redhat.com, jasowang@redhat.com, pbonzini@redhat.com, stefanha@redhat.com, virtualization@lists.linux-foundation.org Subject: [PATCH 05/17] vhost: move vq iovec allocation to dev init time Date: Thu, 22 Oct 2020 00:34:51 +0000 [thread overview] Message-ID: <1603326903-27052-6-git-send-email-michael.christie@oracle.com> (raw) In-Reply-To: <1603326903-27052-1-git-send-email-michael.christie@oracle.com> The next patches allow us to create vqs on demand after vhost_dev_init and vhost_dev_set_owner have been called. For vhost-scsi we don't know the number of vqs we really want until the vring/vq setup operations have started up. For other devices we know the number of vqs at vhost_dev_init time, so for those devs we init the vq and allocate the needed iovecs. For vhost-scsi we will do it later when userspace has instructed us to create a vq. Signed-off-by: Mike Christie <michael.christie@oracle.com> --- drivers/vhost/vhost.c | 71 +++++++++++++++++++++++++++------------------------ 1 file changed, 38 insertions(+), 33 deletions(-) diff --git a/drivers/vhost/vhost.c b/drivers/vhost/vhost.c index b35229e..a4a4450 100644 --- a/drivers/vhost/vhost.c +++ b/drivers/vhost/vhost.c @@ -383,29 +383,27 @@ static void vhost_vq_free_iovecs(struct vhost_virtqueue *vq) vq->heads = NULL; } -/* Helper to allocate iovec buffers for all vqs. */ -static long vhost_dev_alloc_iovecs(struct vhost_dev *dev) +static int vhost_vq_alloc_iovecs(struct vhost_dev *dev, + struct vhost_virtqueue *vq) { - struct vhost_virtqueue *vq; - int i; + vq->indirect = kmalloc_array(UIO_MAXIOV, sizeof(*vq->indirect), + GFP_KERNEL); + if (!vq->indirect) + return -ENOMEM; + + if (!dev->iov_limit) + return 0; + + vq->log = kmalloc_array(dev->iov_limit, sizeof(*vq->log), GFP_KERNEL); + vq->heads = kmalloc_array(dev->iov_limit, sizeof(*vq->heads), + GFP_KERNEL); + if (!vq->log || !vq->heads) + goto err_nomem; - for (i = 0; i < dev->nvqs; ++i) { - vq = dev->vqs[i]; - vq->indirect = kmalloc_array(UIO_MAXIOV, - sizeof(*vq->indirect), - GFP_KERNEL); - vq->log = kmalloc_array(dev->iov_limit, sizeof(*vq->log), - GFP_KERNEL); - vq->heads = kmalloc_array(dev->iov_limit, sizeof(*vq->heads), - GFP_KERNEL); - if (!vq->indirect || !vq->log || !vq->heads) - goto err_nomem; - } return 0; err_nomem: - for (; i >= 0; --i) - vhost_vq_free_iovecs(dev->vqs[i]); + vhost_vq_free_iovecs(vq); return -ENOMEM; } @@ -458,6 +456,21 @@ static size_t vhost_get_desc_size(struct vhost_virtqueue *vq, return sizeof(*vq->desc) * num; } +static int vhost_vq_init(struct vhost_dev *dev, struct vhost_virtqueue *vq) +{ + vq->log = NULL; + vq->indirect = NULL; + vq->heads = NULL; + vq->dev = dev; + mutex_init(&vq->mutex); + vhost_vq_reset(dev, vq); + + if (vq->handle_kick) + vhost_poll_init(&vq->poll, vq->handle_kick, EPOLLIN, dev); + + return vhost_vq_alloc_iovecs(dev, vq); +} + int vhost_dev_init(struct vhost_dev *dev, struct vhost_virtqueue **vqs, int nvqs, int iov_limit, int weight, int byte_weight, @@ -465,7 +478,6 @@ int vhost_dev_init(struct vhost_dev *dev, int (*msg_handler)(struct vhost_dev *dev, struct vhost_iotlb_msg *msg)) { - struct vhost_virtqueue *vq; int i; dev->vqs = vqs; @@ -489,19 +501,16 @@ int vhost_dev_init(struct vhost_dev *dev, for (i = 0; i < dev->nvqs; ++i) { - vq = dev->vqs[i]; - vq->log = NULL; - vq->indirect = NULL; - vq->heads = NULL; - vq->dev = dev; - mutex_init(&vq->mutex); - vhost_vq_reset(dev, vq); - if (vq->handle_kick) - vhost_poll_init(&vq->poll, vq->handle_kick, - EPOLLIN, dev); + if (vhost_vq_init(dev, dev->vqs[i])) + goto err_vq_init; } return 0; + +err_vq_init: + for (--i; i >= 0; --i) + vhost_vq_free_iovecs(dev->vqs[i]); + return -ENOMEM; } EXPORT_SYMBOL_GPL(vhost_dev_init); @@ -606,10 +615,6 @@ long vhost_dev_set_owner(struct vhost_dev *dev) goto err_cgroup; } - err = vhost_dev_alloc_iovecs(dev); - if (err) - goto err_cgroup; - return 0; err_cgroup: if (dev->worker) { -- 1.8.3.1
WARNING: multiple messages have this Message-ID (diff)
From: Mike Christie <michael.christie@oracle.com> To: martin.petersen@oracle.com, linux-scsi@vger.kernel.org, target-devel@vger.kernel.org, mst@redhat.com, jasowang@redhat.com, pbonzini@redhat.com, stefanha@redhat.com, virtualization@lists.linux-foundation.org Subject: [PATCH 05/17] vhost: move vq iovec allocation to dev init time Date: Wed, 21 Oct 2020 19:34:51 -0500 [thread overview] Message-ID: <1603326903-27052-6-git-send-email-michael.christie@oracle.com> (raw) In-Reply-To: <1603326903-27052-1-git-send-email-michael.christie@oracle.com> The next patches allow us to create vqs on demand after vhost_dev_init and vhost_dev_set_owner have been called. For vhost-scsi we don't know the number of vqs we really want until the vring/vq setup operations have started up. For other devices we know the number of vqs at vhost_dev_init time, so for those devs we init the vq and allocate the needed iovecs. For vhost-scsi we will do it later when userspace has instructed us to create a vq. Signed-off-by: Mike Christie <michael.christie@oracle.com> --- drivers/vhost/vhost.c | 71 +++++++++++++++++++++++++++------------------------ 1 file changed, 38 insertions(+), 33 deletions(-) diff --git a/drivers/vhost/vhost.c b/drivers/vhost/vhost.c index b35229e..a4a4450 100644 --- a/drivers/vhost/vhost.c +++ b/drivers/vhost/vhost.c @@ -383,29 +383,27 @@ static void vhost_vq_free_iovecs(struct vhost_virtqueue *vq) vq->heads = NULL; } -/* Helper to allocate iovec buffers for all vqs. */ -static long vhost_dev_alloc_iovecs(struct vhost_dev *dev) +static int vhost_vq_alloc_iovecs(struct vhost_dev *dev, + struct vhost_virtqueue *vq) { - struct vhost_virtqueue *vq; - int i; + vq->indirect = kmalloc_array(UIO_MAXIOV, sizeof(*vq->indirect), + GFP_KERNEL); + if (!vq->indirect) + return -ENOMEM; + + if (!dev->iov_limit) + return 0; + + vq->log = kmalloc_array(dev->iov_limit, sizeof(*vq->log), GFP_KERNEL); + vq->heads = kmalloc_array(dev->iov_limit, sizeof(*vq->heads), + GFP_KERNEL); + if (!vq->log || !vq->heads) + goto err_nomem; - for (i = 0; i < dev->nvqs; ++i) { - vq = dev->vqs[i]; - vq->indirect = kmalloc_array(UIO_MAXIOV, - sizeof(*vq->indirect), - GFP_KERNEL); - vq->log = kmalloc_array(dev->iov_limit, sizeof(*vq->log), - GFP_KERNEL); - vq->heads = kmalloc_array(dev->iov_limit, sizeof(*vq->heads), - GFP_KERNEL); - if (!vq->indirect || !vq->log || !vq->heads) - goto err_nomem; - } return 0; err_nomem: - for (; i >= 0; --i) - vhost_vq_free_iovecs(dev->vqs[i]); + vhost_vq_free_iovecs(vq); return -ENOMEM; } @@ -458,6 +456,21 @@ static size_t vhost_get_desc_size(struct vhost_virtqueue *vq, return sizeof(*vq->desc) * num; } +static int vhost_vq_init(struct vhost_dev *dev, struct vhost_virtqueue *vq) +{ + vq->log = NULL; + vq->indirect = NULL; + vq->heads = NULL; + vq->dev = dev; + mutex_init(&vq->mutex); + vhost_vq_reset(dev, vq); + + if (vq->handle_kick) + vhost_poll_init(&vq->poll, vq->handle_kick, EPOLLIN, dev); + + return vhost_vq_alloc_iovecs(dev, vq); +} + int vhost_dev_init(struct vhost_dev *dev, struct vhost_virtqueue **vqs, int nvqs, int iov_limit, int weight, int byte_weight, @@ -465,7 +478,6 @@ int vhost_dev_init(struct vhost_dev *dev, int (*msg_handler)(struct vhost_dev *dev, struct vhost_iotlb_msg *msg)) { - struct vhost_virtqueue *vq; int i; dev->vqs = vqs; @@ -489,19 +501,16 @@ int vhost_dev_init(struct vhost_dev *dev, for (i = 0; i < dev->nvqs; ++i) { - vq = dev->vqs[i]; - vq->log = NULL; - vq->indirect = NULL; - vq->heads = NULL; - vq->dev = dev; - mutex_init(&vq->mutex); - vhost_vq_reset(dev, vq); - if (vq->handle_kick) - vhost_poll_init(&vq->poll, vq->handle_kick, - EPOLLIN, dev); + if (vhost_vq_init(dev, dev->vqs[i])) + goto err_vq_init; } return 0; + +err_vq_init: + for (--i; i >= 0; --i) + vhost_vq_free_iovecs(dev->vqs[i]); + return -ENOMEM; } EXPORT_SYMBOL_GPL(vhost_dev_init); @@ -606,10 +615,6 @@ long vhost_dev_set_owner(struct vhost_dev *dev) goto err_cgroup; } - err = vhost_dev_alloc_iovecs(dev); - if (err) - goto err_cgroup; - return 0; err_cgroup: if (dev->worker) { -- 1.8.3.1
next prev parent reply other threads:[~2020-10-22 0:34 UTC|newest] Thread overview: 102+ messages / expand[flat|nested] mbox.gz Atom feed top 2020-10-22 0:34 [PATCH 00/17 V3] vhost: fix scsi cmd handling and cgroup support Mike Christie 2020-10-22 0:34 ` Mike Christie 2020-10-22 0:34 ` [PATCH 01/17] vhost scsi: add lun parser helper Mike Christie 2020-10-22 0:34 ` Mike Christie 2020-10-26 3:33 ` Jason Wang 2020-10-26 3:33 ` Jason Wang 2020-10-26 3:33 ` Jason Wang 2020-10-22 0:34 ` [PATCH 02/17] vhost: remove work arg from vhost_work_flush Mike Christie 2020-10-22 0:34 ` Mike Christie 2020-10-22 0:51 ` Chaitanya Kulkarni 2020-10-22 0:51 ` Chaitanya Kulkarni 2020-10-22 0:34 ` [PATCH 03/17] vhost net: use goto error handling in open Mike Christie 2020-10-22 0:34 ` Mike Christie 2020-10-22 0:45 ` Chaitanya Kulkarni 2020-10-22 0:45 ` Chaitanya Kulkarni 2020-10-26 3:34 ` Jason Wang 2020-10-26 3:34 ` Jason Wang 2020-10-26 3:34 ` Jason Wang 2020-10-22 0:34 ` [PATCH 04/17] vhost: prep vhost_dev_init users to handle failures Mike Christie 2020-10-22 0:34 ` Mike Christie 2020-10-22 5:22 ` kernel test robot 2020-10-22 5:22 ` kernel test robot 2020-10-22 5:22 ` kernel test robot 2020-10-22 5:22 ` kernel test robot 2020-10-23 16:15 ` Mike Christie 2020-10-23 16:15 ` Mike Christie 2020-11-02 5:57 ` Jason Wang 2020-11-02 5:57 ` Jason Wang 2020-11-02 5:57 ` Jason Wang 2020-11-03 10:04 ` Dan Carpenter 2020-11-03 10:04 ` Dan Carpenter 2020-11-03 10:04 ` Dan Carpenter 2020-11-03 10:04 ` Dan Carpenter 2020-11-03 10:04 ` Dan Carpenter 2020-10-22 0:34 ` Mike Christie [this message] 2020-10-22 0:34 ` [PATCH 05/17] vhost: move vq iovec allocation to dev init time Mike Christie 2020-10-22 0:34 ` [PATCH 06/17] vhost: support delayed vq creation Mike Christie 2020-10-22 0:34 ` Mike Christie 2020-10-22 0:34 ` [PATCH 07/17] vhost scsi: support delayed IO " Mike Christie 2020-10-22 0:34 ` Mike Christie 2020-10-26 3:51 ` Jason Wang 2020-10-26 3:51 ` Jason Wang 2020-10-26 3:51 ` Jason Wang 2020-10-27 5:47 ` Mike Christie 2020-10-27 5:47 ` Mike Christie 2020-10-28 1:55 ` Jason Wang 2020-10-28 1:55 ` Jason Wang 2020-10-28 1:55 ` Jason Wang 2020-10-30 8:47 ` Michael S. Tsirkin 2020-10-30 8:47 ` Michael S. Tsirkin 2020-10-30 8:47 ` Michael S. Tsirkin 2020-10-30 16:30 ` Mike Christie 2020-10-30 16:30 ` Mike Christie 2020-10-30 17:26 ` Mike Christie 2020-10-30 17:26 ` Mike Christie 2020-11-01 22:06 ` Mike Christie 2020-11-01 22:06 ` Mike Christie 2020-11-02 6:36 ` Jason Wang 2020-11-02 6:36 ` Jason Wang 2020-11-02 6:36 ` Jason Wang 2020-11-02 6:49 ` Jason Wang 2020-11-02 6:49 ` Jason Wang 2020-11-02 6:49 ` Jason Wang 2020-11-02 16:19 ` Mike Christie 2020-11-02 16:19 ` Mike Christie 2020-10-22 0:34 ` [PATCH 08/17] vhost scsi: alloc cmds per vq instead of session Mike Christie 2020-10-22 0:34 ` Mike Christie 2020-10-22 0:34 ` [PATCH 09/17] vhost scsi: fix cmd completion race Mike Christie 2020-10-22 0:34 ` Mike Christie 2020-10-27 13:07 ` Maurizio Lombardi 2020-10-27 13:07 ` Maurizio Lombardi 2020-10-30 8:51 ` Michael S. Tsirkin 2020-10-30 8:51 ` Michael S. Tsirkin 2020-10-30 8:51 ` Michael S. Tsirkin 2020-10-30 16:04 ` Paolo Bonzini 2020-10-30 16:04 ` Paolo Bonzini 2020-10-30 16:04 ` Paolo Bonzini 2020-10-22 0:34 ` [PATCH 10/17] vhost scsi: Add support for LUN resets Mike Christie 2020-10-22 0:34 ` Mike Christie 2020-10-22 0:34 ` [PATCH 11/17] vhost scsi: remove extra flushes Mike Christie 2020-10-22 0:34 ` Mike Christie 2020-10-22 0:34 ` [PATCH 12/17] vhost poll: fix coding style Mike Christie 2020-10-22 0:34 ` Mike Christie 2020-10-22 0:39 ` Chaitanya Kulkarni 2020-10-22 0:39 ` Chaitanya Kulkarni 2020-10-22 0:34 ` [PATCH 13/17] vhost: support multiple worker threads Mike Christie 2020-10-22 0:34 ` Mike Christie 2020-10-22 0:35 ` [PATCH 14/17] vhost: poll support support multiple workers Mike Christie 2020-10-22 0:35 ` Mike Christie 2020-10-22 0:35 ` [PATCH 15/17] host: support delayed vq creation Mike Christie 2020-10-22 0:35 ` Mike Christie 2020-10-22 0:50 ` Mike Christie 2020-10-22 0:50 ` Mike Christie 2020-10-22 0:35 ` [PATCH 16/17] vhost scsi: multiple worker support Mike Christie 2020-10-22 0:35 ` Mike Christie 2020-10-22 0:35 ` [PATCH 17/17] vhost scsi: drop submission workqueue Mike Christie 2020-10-22 0:35 ` Mike Christie 2020-10-29 21:47 ` [PATCH 00/17 V3] vhost: fix scsi cmd handling and cgroup support Michael S. Tsirkin 2020-10-29 21:47 ` Michael S. Tsirkin 2020-10-29 21:47 ` Michael S. Tsirkin 2020-10-29 22:19 ` Mike Christie 2020-10-29 22:19 ` Mike Christie
Reply instructions: You may reply publicly to this message via plain-text email using any one of the following methods: * Save the following mbox file, import it into your mail client, and reply-to-all from there: mbox Avoid top-posting and favor interleaved quoting: https://en.wikipedia.org/wiki/Posting_style#Interleaved_style * Reply using the --to, --cc, and --in-reply-to switches of git-send-email(1): git send-email \ --in-reply-to=1603326903-27052-6-git-send-email-michael.christie@oracle.com \ --to=michael.christie@oracle.com \ --cc=jasowang@redhat.com \ --cc=linux-scsi@vger.kernel.org \ --cc=martin.petersen@oracle.com \ --cc=mst@redhat.com \ --cc=pbonzini@redhat.com \ --cc=stefanha@redhat.com \ --cc=target-devel@vger.kernel.org \ --cc=virtualization@lists.linux-foundation.org \ /path/to/YOUR_REPLY https://kernel.org/pub/software/scm/git/docs/git-send-email.html * If your mail client supports setting the In-Reply-To header via mailto: links, try the mailto: linkBe sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes, see mirroring instructions on how to clone and mirror all data and code used by this external index.