From: "Michael S. Tsirkin" <mst@redhat.com>
To: qemu-devel@nongnu.org
Cc: Peter Maydell <peter.maydell@linaro.org>,
Jason Wang <jasowang@redhat.com>
Subject: [PULL v2 36/44] vhost-net: control virtqueue support
Date: Wed, 20 Oct 2021 06:20:53 -0400 [thread overview]
Message-ID: <20211020101844.988480-37-mst@redhat.com> (raw)
In-Reply-To: <20211020101844.988480-1-mst@redhat.com>
From: Jason Wang <jasowang@redhat.com>
We assume there's no cvq in the past, this is not true when we need
control virtqueue support for vhost-user backends. So this patch
implements the control virtqueue support for vhost-net. As datapath,
the control virtqueue is also required to be coupled with the
NetClientState. The vhost_net_start/stop() are tweaked to accept the
number of datapath queue pairs plus the the number of control
virtqueue for us to start and stop the vhost device.
Signed-off-by: Jason Wang <jasowang@redhat.com>
Message-Id: <20211020045600.16082-7-jasowang@redhat.com>
Reviewed-by: Michael S. Tsirkin <mst@redhat.com>
Signed-off-by: Michael S. Tsirkin <mst@redhat.com>
---
include/net/vhost_net.h | 6 ++++--
hw/net/vhost_net-stub.c | 4 ++--
hw/net/vhost_net.c | 43 ++++++++++++++++++++++++++++++-----------
hw/net/virtio-net.c | 4 ++--
4 files changed, 40 insertions(+), 17 deletions(-)
diff --git a/include/net/vhost_net.h b/include/net/vhost_net.h
index fba40cf695..387e913e4e 100644
--- a/include/net/vhost_net.h
+++ b/include/net/vhost_net.h
@@ -21,8 +21,10 @@ typedef struct VhostNetOptions {
uint64_t vhost_net_get_max_queues(VHostNetState *net);
struct vhost_net *vhost_net_init(VhostNetOptions *options);
-int vhost_net_start(VirtIODevice *dev, NetClientState *ncs, int total_queues);
-void vhost_net_stop(VirtIODevice *dev, NetClientState *ncs, int total_queues);
+int vhost_net_start(VirtIODevice *dev, NetClientState *ncs,
+ int data_queue_pairs, int cvq);
+void vhost_net_stop(VirtIODevice *dev, NetClientState *ncs,
+ int data_queue_pairs, int cvq);
void vhost_net_cleanup(VHostNetState *net);
diff --git a/hw/net/vhost_net-stub.c b/hw/net/vhost_net-stub.c
index a7f4252630..89d71cfb8e 100644
--- a/hw/net/vhost_net-stub.c
+++ b/hw/net/vhost_net-stub.c
@@ -33,13 +33,13 @@ struct vhost_net *vhost_net_init(VhostNetOptions *options)
int vhost_net_start(VirtIODevice *dev,
NetClientState *ncs,
- int total_queues)
+ int data_queue_pairs, int cvq)
{
return -ENOSYS;
}
void vhost_net_stop(VirtIODevice *dev,
NetClientState *ncs,
- int total_queues)
+ int data_queue_pairs, int cvq)
{
}
diff --git a/hw/net/vhost_net.c b/hw/net/vhost_net.c
index 386ec2eaa2..e1e9d1ec89 100644
--- a/hw/net/vhost_net.c
+++ b/hw/net/vhost_net.c
@@ -315,11 +315,14 @@ static void vhost_net_stop_one(struct vhost_net *net,
}
int vhost_net_start(VirtIODevice *dev, NetClientState *ncs,
- int total_queues)
+ int data_queue_pairs, int cvq)
{
BusState *qbus = BUS(qdev_get_parent_bus(DEVICE(dev)));
VirtioBusState *vbus = VIRTIO_BUS(qbus);
VirtioBusClass *k = VIRTIO_BUS_GET_CLASS(vbus);
+ int total_notifiers = data_queue_pairs * 2 + cvq;
+ VirtIONet *n = VIRTIO_NET(dev);
+ int nvhosts = data_queue_pairs + cvq;
struct vhost_net *net;
int r, e, i;
NetClientState *peer;
@@ -329,9 +332,14 @@ int vhost_net_start(VirtIODevice *dev, NetClientState *ncs,
return -ENOSYS;
}
- for (i = 0; i < total_queues; i++) {
+ for (i = 0; i < nvhosts; i++) {
+
+ if (i < data_queue_pairs) {
+ peer = qemu_get_peer(ncs, i);
+ } else { /* Control Virtqueue */
+ peer = qemu_get_peer(ncs, n->max_queues);
+ }
- peer = qemu_get_peer(ncs, i);
net = get_vhost_net(peer);
vhost_net_set_vq_index(net, i * 2);
@@ -344,14 +352,18 @@ int vhost_net_start(VirtIODevice *dev, NetClientState *ncs,
}
}
- r = k->set_guest_notifiers(qbus->parent, total_queues * 2, true);
+ r = k->set_guest_notifiers(qbus->parent, total_notifiers, true);
if (r < 0) {
error_report("Error binding guest notifier: %d", -r);
goto err;
}
- for (i = 0; i < total_queues; i++) {
- peer = qemu_get_peer(ncs, i);
+ for (i = 0; i < nvhosts; i++) {
+ if (i < data_queue_pairs) {
+ peer = qemu_get_peer(ncs, i);
+ } else {
+ peer = qemu_get_peer(ncs, n->max_queues);
+ }
r = vhost_net_start_one(get_vhost_net(peer), dev);
if (r < 0) {
@@ -375,7 +387,7 @@ err_start:
peer = qemu_get_peer(ncs , i);
vhost_net_stop_one(get_vhost_net(peer), dev);
}
- e = k->set_guest_notifiers(qbus->parent, total_queues * 2, false);
+ e = k->set_guest_notifiers(qbus->parent, total_notifiers, false);
if (e < 0) {
fprintf(stderr, "vhost guest notifier cleanup failed: %d\n", e);
fflush(stderr);
@@ -385,18 +397,27 @@ err:
}
void vhost_net_stop(VirtIODevice *dev, NetClientState *ncs,
- int total_queues)
+ int data_queue_pairs, int cvq)
{
BusState *qbus = BUS(qdev_get_parent_bus(DEVICE(dev)));
VirtioBusState *vbus = VIRTIO_BUS(qbus);
VirtioBusClass *k = VIRTIO_BUS_GET_CLASS(vbus);
+ VirtIONet *n = VIRTIO_NET(dev);
+ NetClientState *peer;
+ int total_notifiers = data_queue_pairs * 2 + cvq;
+ int nvhosts = data_queue_pairs + cvq;
int i, r;
- for (i = 0; i < total_queues; i++) {
- vhost_net_stop_one(get_vhost_net(ncs[i].peer), dev);
+ for (i = 0; i < nvhosts; i++) {
+ if (i < data_queue_pairs) {
+ peer = qemu_get_peer(ncs, i);
+ } else {
+ peer = qemu_get_peer(ncs, n->max_queues);
+ }
+ vhost_net_stop_one(get_vhost_net(peer), dev);
}
- r = k->set_guest_notifiers(qbus->parent, total_queues * 2, false);
+ r = k->set_guest_notifiers(qbus->parent, total_notifiers, false);
if (r < 0) {
fprintf(stderr, "vhost guest notifier cleanup failed: %d\n", r);
fflush(stderr);
diff --git a/hw/net/virtio-net.c b/hw/net/virtio-net.c
index 3dd2896ff9..5ee6729662 100644
--- a/hw/net/virtio-net.c
+++ b/hw/net/virtio-net.c
@@ -285,14 +285,14 @@ static void virtio_net_vhost_status(VirtIONet *n, uint8_t status)
}
n->vhost_started = 1;
- r = vhost_net_start(vdev, n->nic->ncs, queues);
+ r = vhost_net_start(vdev, n->nic->ncs, queues, 0);
if (r < 0) {
error_report("unable to start vhost net: %d: "
"falling back on userspace virtio", -r);
n->vhost_started = 0;
}
} else {
- vhost_net_stop(vdev, n->nic->ncs, queues);
+ vhost_net_stop(vdev, n->nic->ncs, queues, 0);
n->vhost_started = 0;
}
}
--
MST
next prev parent reply other threads:[~2021-10-20 11:17 UTC|newest]
Thread overview: 48+ messages / expand[flat|nested] mbox.gz Atom feed top
2021-10-20 10:19 [PULL v2 00/44] pc,pci,virtio: features, fixes, tests Michael S. Tsirkin
2021-10-20 10:19 ` [PULL v2 01/44] tests: acpi: dump table with failed checksum Michael S. Tsirkin
2021-10-20 10:19 ` [PULL v2 02/44] tests: qtest: add qtest_has_accel() to check if tested binary supports accelerator Michael S. Tsirkin
2021-10-22 13:04 ` Jason Andryuk
2021-10-26 13:41 ` Igor Mammedov
2021-10-20 10:19 ` [PULL v2 03/44] tests: acpi: whitelist expected tables for acpi/q35/xapic testcase Michael S. Tsirkin
2021-10-20 10:19 ` [PULL v2 04/44] tests: acpi: q35: test for x2APIC entries in SRAT Michael S. Tsirkin
2021-10-20 10:19 ` [PULL v2 05/44] tests: acpi: update expected tables blobs Michael S. Tsirkin
2021-10-20 10:19 ` [PULL v2 06/44] tests: acpi: whitelist new expected table tests/data/acpi/q35/DMAR.dmar Michael S. Tsirkin
2021-10-20 10:19 ` [PULL v2 07/44] tests: acpi: add testcase for intel_iommu (DMAR table) Michael S. Tsirkin
2021-10-20 10:19 ` [PULL v2 08/44] tests: acpi: add expected blob for DMAR table Michael S. Tsirkin
2021-10-20 10:19 ` [PULL v2 09/44] tests: acpi: whitelist expected blobs for new acpi/q35/ivrs testcase Michael S. Tsirkin
2021-10-20 10:19 ` [PULL v2 10/44] tests: acpi: add testcase for amd-iommu (IVRS table) Michael S. Tsirkin
2021-10-20 10:19 ` [PULL v2 11/44] tests: acpi: update expected blobs Michael S. Tsirkin
2021-10-20 10:19 ` [PULL v2 12/44] tests: acpi: arm/virt: drop redundant test_acpi_one() in test_acpi_virt_tcg() Michael S. Tsirkin
2021-10-20 10:19 ` [PULL v2 13/44] tests: arm-cpu-features: use qtest_has_kvm() API Michael S. Tsirkin
2021-10-20 10:19 ` [PULL v2 14/44] tests: migration-test: use qtest_has_accel() API Michael S. Tsirkin
2021-10-20 10:19 ` [PULL v2 15/44] tests: bios-tables-test: use qtest_has_accel() API to register TCG only tests Michael S. Tsirkin
2021-10-20 10:19 ` [PULL v2 16/44] vdpa: Skip protected ram IOMMU mappings Michael S. Tsirkin
2021-10-20 10:19 ` [PULL v2 17/44] vdpa: Add vhost_vdpa_section_end Michael S. Tsirkin
2021-10-20 10:20 ` [PULL v2 18/44] vdpa: Check for iova range at mappings changes Michael S. Tsirkin
2021-10-20 10:20 ` [PULL v2 19/44] virtio-iommu: Remove the non transitional name Michael S. Tsirkin
2021-10-20 10:20 ` [PULL v2 20/44] virtio-iommu: Drop base_name and change generic_name Michael S. Tsirkin
2021-10-20 10:20 ` [PULL v2 21/44] libvhost-user: fix VHOST_USER_REM_MEM_REG skipping mmap_addr Michael S. Tsirkin
2021-10-20 10:20 ` [PULL v2 22/44] vhost-user-rng: Add vhost-user-rng implementation Michael S. Tsirkin
2021-10-20 10:20 ` [PULL v2 23/44] vhost-user-rng-pci: Add vhost-user-rng-pci implementation Michael S. Tsirkin
2021-10-20 10:20 ` [PULL v2 24/44] docs: Add documentation for vhost based RNG implementation Michael S. Tsirkin
2021-10-20 10:20 ` [PULL v2 25/44] vhost-user: fix duplicated notifier MR init Michael S. Tsirkin
2021-10-20 10:20 ` [PULL v2 26/44] failover: fix a regression introduced by JSON'ification of -device Michael S. Tsirkin
2021-10-20 10:20 ` [PULL v2 27/44] vhost-user-blk-test: pass vhost-user socket fds to QSD Michael S. Tsirkin
2021-10-20 10:20 ` [PULL v2 28/44] qdev/qbus: remove failover specific code Michael S. Tsirkin
2021-10-20 10:20 ` [PULL v2 29/44] rebuild-expected-aml.sh: allow partial target list Michael S. Tsirkin
2021-10-20 10:20 ` [PULL v2 30/44] bios-tables-test: don't disassemble empty files Michael S. Tsirkin
2021-10-20 10:20 ` [PULL v2 31/44] vhost-vdpa: open device fd in net_init_vhost_vdpa() Michael S. Tsirkin
2021-10-20 10:20 ` [PULL v2 32/44] vhost-vdpa: classify one time request Michael S. Tsirkin
2021-10-20 10:20 ` [PULL v2 33/44] vhost-vdpa: prepare for the multiqueue support Michael S. Tsirkin
2021-10-20 10:20 ` [PULL v2 34/44] vhost-vdpa: let net_vhost_vdpa_init() returns NetClientState * Michael S. Tsirkin
2021-10-20 10:20 ` [PULL v2 35/44] net: introduce control client Michael S. Tsirkin
2021-10-20 10:20 ` Michael S. Tsirkin [this message]
2021-10-20 10:20 ` [PULL v2 37/44] virtio-net: use "queue_pairs" instead of "queues" when possible Michael S. Tsirkin
2021-10-20 10:20 ` [PULL v2 38/44] vhost: record the last virtqueue index for the virtio device Michael S. Tsirkin
2021-10-20 10:21 ` [PULL v2 39/44] virtio-net: vhost control virtqueue support Michael S. Tsirkin
2021-10-20 10:21 ` [PULL v2 40/44] vhost-vdpa: multiqueue support Michael S. Tsirkin
2021-10-20 10:21 ` [PULL v2 41/44] pci: fix PCI resource reserve capability on BE Michael S. Tsirkin
2021-10-20 10:21 ` [PULL v2 42/44] tests/acpi/bios-tables-test: add and allow changes to a new q35 DSDT table blob Michael S. Tsirkin
2021-10-20 10:21 ` [PULL v2 43/44] tests/acpi/pcihp: add unit tests for hotplug on multifunction bridges for q35 Michael S. Tsirkin
2021-10-20 10:21 ` [PULL v2 44/44] tests/acpi/bios-tables-test: update DSDT blob for multifunction bridge test Michael S. Tsirkin
2021-10-20 17:56 ` [PULL v2 00/44] pc,pci,virtio: features, fixes, tests Richard Henderson
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20211020101844.988480-37-mst@redhat.com \
--to=mst@redhat.com \
--cc=jasowang@redhat.com \
--cc=peter.maydell@linaro.org \
--cc=qemu-devel@nongnu.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).