From: Jason Wang <jasowang@redhat.com>
To: Cindy Lu <lulu@redhat.com>,
mst@redhat.com, armbru@redhat.com, eblake@redhat.com,
cohuck@redhat.com
Cc: mhabets@solarflare.com, qemu-devel@nongnu.org,
rob.miller@broadcom.com, saugatm@xilinx.com, hanand@xilinx.com,
hch@infradead.org, eperezma@redhat.com, jgg@mellanox.com,
shahafs@mellanox.com, kevin.tian@intel.com, parav@mellanox.com,
vmireyno@marvell.com, cunming.liang@intel.com, gdawar@xilinx.com,
jiri@mellanox.com, xiao.w.wang@intel.com, stefanha@redhat.com,
zhihong.wang@intel.com, Tiwei Bie <tiwei.bie@intel.com>,
aadam@redhat.com, rdunlap@infradead.org,
maxime.coquelin@redhat.com, lingshan.zhu@intel.com
Subject: Re: [RFC v2 4/9] vhost-vdpa: introduce vhost-vdpa net client
Date: Sat, 9 May 2020 10:40:17 +0800 [thread overview]
Message-ID: <570f92fa-dbf4-e9d0-e00d-18f02fafd0ce@redhat.com> (raw)
In-Reply-To: <20200508163218.22592-5-lulu@redhat.com>
On 2020/5/9 上午12:32, Cindy Lu wrote:
> From: Tiwei Bie <tiwei.bie@intel.com>
If you think you've done a huge refactor on the code, you can change the
author but need to keep the sob of Tiwei.
>
> This patch set introduces a new net client type: vhost-vdpa.
> vhost-vdpa net client will set up a vDPA device which is specified
> by a "vhostdev" parameter.
>
> Co-authored-by: Lingshan Zhu <lingshan.zhu@intel.com>
> Signed-off-by: Cindy Lu <lulu@redhat.com>
> ---
> configure | 21 ++++
> include/net/vhost-vdpa.h | 19 ++++
> include/net/vhost_net.h | 1 +
Patch 5 which is the infrastructure of vhost-vpda should come first.
Please re-order the patch in next version.
> net/Makefile.objs | 2 +-
> net/clients.h | 2 +
> net/net.c | 3 +
> net/vhost-vdpa.c | 227 +++++++++++++++++++++++++++++++++++++++
> qapi/net.json | 22 +++-
> qemu-options.hx | 19 ++++
> 9 files changed, 313 insertions(+), 3 deletions(-)
> create mode 100644 include/net/vhost-vdpa.h
> create mode 100644 net/vhost-vdpa.c
>
> diff --git a/configure b/configure
> index 6099be1d84..bdd732e3bb 100755
> --- a/configure
> +++ b/configure
> @@ -1505,6 +1505,10 @@ for opt do
> ;;
> --enable-vhost-user) vhost_user="yes"
> ;;
> + --disable-vhost-vdpa) vhost_vdpa="no"
> + ;;
> + --enable-vhost-vdpa) vhost_vdpa="yes"
> + ;;
> --disable-vhost-kernel) vhost_kernel="no"
> ;;
> --enable-vhost-kernel) vhost_kernel="yes"
> @@ -1780,6 +1784,7 @@ disabled with --disable-FEATURE, default is enabled if available:
> vhost-crypto vhost-user-crypto backend support
> vhost-kernel vhost kernel backend support
> vhost-user vhost-user backend support
> + vhost-vdpa vhost-vdpa backend support
Maybe "vhost-vdpa kernel backend support" is better.
> spice spice
> rbd rados block device (rbd)
> libiscsi iscsi support
> @@ -2241,6 +2246,10 @@ test "$vhost_user" = "" && vhost_user=yes
> if test "$vhost_user" = "yes" && test "$mingw32" = "yes"; then
> error_exit "vhost-user isn't available on win32"
> fi
> +test "$vhost_vdpa" = "" && vhost_vdpa=yes
> +if test "$vhost_vdpa" = "yes" && test "$mingw32" = "yes"; then
> + error_exit "vhost-vdpa isn't available on win32"
> +fi
Let's add a check for Linux like vhost kernel below.
> test "$vhost_kernel" = "" && vhost_kernel=$linux
> if test "$vhost_kernel" = "yes" && test "$linux" != "yes"; then
> error_exit "vhost-kernel is only available on Linux"
> @@ -2269,6 +2278,11 @@ test "$vhost_user_fs" = "" && vhost_user_fs=$vhost_user
> if test "$vhost_user_fs" = "yes" && test "$vhost_user" = "no"; then
> error_exit "--enable-vhost-user-fs requires --enable-vhost-user"
> fi
> +#vhost-vdpa backends
> +test "$vhost_net_vdpa" = "" && vhost_net_vdpa=$vhost_vdpa
> +if test "$vhost_net_vdpa" = "yes" && test "$vhost_vdpa" = "no"; then
> + error_exit "--enable-vhost-net-vdpa requires --enable-vhost-vdpa"
> +fi
>
> # OR the vhost-kernel and vhost-user values for simplicity
> if test "$vhost_net" = ""; then
> @@ -6543,6 +6557,7 @@ echo "vhost-scsi support $vhost_scsi"
> echo "vhost-vsock support $vhost_vsock"
> echo "vhost-user support $vhost_user"
> echo "vhost-user-fs support $vhost_user_fs"
> +echo "vhost-vdpa support $vhost_vdpa"
> echo "Trace backends $trace_backends"
> if have_backend "simple"; then
> echo "Trace output file $trace_file-<pid>"
> @@ -7031,6 +7046,9 @@ fi
> if test "$vhost_net_user" = "yes" ; then
> echo "CONFIG_VHOST_NET_USER=y" >> $config_host_mak
> fi
> +if test "$vhost_net_vdpa" = "yes" ; then
> + echo "CONFIG_VHOST_NET_VDPA=y" >> $config_host_mak
> +fi
> if test "$vhost_crypto" = "yes" ; then
> echo "CONFIG_VHOST_CRYPTO=y" >> $config_host_mak
> fi
> @@ -7043,6 +7061,9 @@ fi
> if test "$vhost_user" = "yes" ; then
> echo "CONFIG_VHOST_USER=y" >> $config_host_mak
> fi
> +if test "$vhost_vdpa" = "yes" ; then
> + echo "CONFIG_VHOST_VDPA=y" >> $config_host_mak
> +fi
> if test "$vhost_user_fs" = "yes" ; then
> echo "CONFIG_VHOST_USER_FS=y" >> $config_host_mak
> fi
> diff --git a/include/net/vhost-vdpa.h b/include/net/vhost-vdpa.h
> new file mode 100644
> index 0000000000..6ce0d04f72
> --- /dev/null
> +++ b/include/net/vhost-vdpa.h
> @@ -0,0 +1,19 @@
> +/*
> + * vhost-vdpa.h
> + *
> + * Copyright(c) 2017-2018 Intel Corporation.
> + * Copyright(c) 2020 Red Hat, Inc.
> + *
> + * This work is licensed under the terms of the GNU GPL, version 2 or later.
> + * See the COPYING file in the top-level directory.
> + *
> + */
> +
> +#ifndef VHOST_VDPA_H
> +#define VHOST_VDPA_H
> +
> +struct vhost_net;
> +struct vhost_net *vhost_vdpa_get_vhost_net(NetClientState *nc);
> +uint64_t vhost_vdpa_get_acked_features(NetClientState *nc);
> +
> +#endif /* VHOST_VDPA_H */
> diff --git a/include/net/vhost_net.h b/include/net/vhost_net.h
> index 6548a5a105..b47844bf29 100644
> --- a/include/net/vhost_net.h
> +++ b/include/net/vhost_net.h
> @@ -40,4 +40,5 @@ uint64_t vhost_net_get_acked_features(VHostNetState *net);
>
> int vhost_net_set_mtu(struct vhost_net *net, uint16_t mtu);
> int vhost_set_state(NetClientState *nc, uint8_t state);
> +int vhost_net_get_device_id(struct vhost_net *net, uint32_t *device_id);
Let's move this function to vhost-vdpa generic header instead of net.
> #endif
> diff --git a/net/Makefile.objs b/net/Makefile.objs
> index c5d076d19c..5ab45545db 100644
> --- a/net/Makefile.objs
> +++ b/net/Makefile.objs
> @@ -26,7 +26,7 @@ tap-obj-$(CONFIG_SOLARIS) = tap-solaris.o
> tap-obj-y ?= tap-stub.o
> common-obj-$(CONFIG_POSIX) += tap.o $(tap-obj-y)
> common-obj-$(CONFIG_WIN32) += tap-win32.o
> -
> +common-obj-$(CONFIG_VHOST_NET_VDPA) += vhost-vdpa.o
> vde.o-libs = $(VDE_LIBS)
>
> common-obj-$(CONFIG_CAN_BUS) += can/
> diff --git a/net/clients.h b/net/clients.h
> index a6ef267e19..92f9b59aed 100644
> --- a/net/clients.h
> +++ b/net/clients.h
> @@ -61,4 +61,6 @@ int net_init_netmap(const Netdev *netdev, const char *name,
> int net_init_vhost_user(const Netdev *netdev, const char *name,
> NetClientState *peer, Error **errp);
>
> +int net_init_vhost_vdpa(const Netdev *netdev, const char *name,
> + NetClientState *peer, Error **errp);
> #endif /* QEMU_NET_CLIENTS_H */
> diff --git a/net/net.c b/net/net.c
> index b3192deaf1..9eff1ae982 100644
> --- a/net/net.c
> +++ b/net/net.c
> @@ -965,6 +965,9 @@ static int (* const net_client_init_fun[NET_CLIENT_DRIVER__MAX])(
> #ifdef CONFIG_VHOST_NET_USER
> [NET_CLIENT_DRIVER_VHOST_USER] = net_init_vhost_user,
> #endif
> +#ifdef CONFIG_VHOST_NET_VDPA
> + [NET_CLIENT_DRIVER_VHOST_VDPA] = net_init_vhost_vdpa,
> +#endif
> #ifdef CONFIG_L2TPV3
> [NET_CLIENT_DRIVER_L2TPV3] = net_init_l2tpv3,
> #endif
> diff --git a/net/vhost-vdpa.c b/net/vhost-vdpa.c
> new file mode 100644
> index 0000000000..c29678fdf1
> --- /dev/null
> +++ b/net/vhost-vdpa.c
> @@ -0,0 +1,227 @@
> +/*
> + * vhost-vdpa.c
> + *
> + * Copyright(c) 2017-2018 Intel Corporation.
> + * Copyright(c) 2020 Red Hat, Inc.
> + *
> + * This work is licensed under the terms of the GNU GPL, version 2 or later.
> + * See the COPYING file in the top-level directory.
> + *
> + */
> +
> +#include "qemu/osdep.h"
> +#include "clients.h"
> +#include "net/vhost_net.h"
> +#include "net/vhost-vdpa.h"
> +#include "hw/virtio/vhost-vdpa.h"
> +#include "qemu/config-file.h"
> +#include "qemu/error-report.h"
> +#include "qemu/option.h"
> +#include "qapi/error.h"
> +#include <linux/vfio.h>
No need any more.
> +#include <sys/ioctl.h>
> +#include <err.h>
> +#include <linux/virtio_net.h>
That's not the way we include standard linux headers, qemu maintain a
copy of standard linux headers.
Need use #include "standard-headers/linux/xxx.h"
> +#include "monitor/monitor.h"
> +#include "qemu/sockets.h"
Do we really need this?
> +#include "hw/virtio/vhost.h"
> +
> +/* Todo:need to add the multiqueue support here */
> +typedef struct VhostVDPAState {
> + NetClientState nc;
> + struct vhost_vdpa vhost_vdpa;
> + VHostNetState *vhost_net;
> + uint64_t acked_features;
> + bool started;
> +} VhostVDPAState;
> +
> +VHostNetState *vhost_vdpa_get_vhost_net(NetClientState *nc)
> +{
> + VhostVDPAState *s = DO_UPCAST(VhostVDPAState, nc, nc);
> + assert(nc->info->type == NET_CLIENT_DRIVER_VHOST_VDPA);
> + return s->vhost_net;
> +}
> +
> +uint64_t vhost_vdpa_get_acked_features(NetClientState *nc)
> +{
> + VhostVDPAState *s = DO_UPCAST(VhostVDPAState, nc, nc);
> + assert(nc->info->type == NET_CLIENT_DRIVER_VHOST_VDPA);
> + return s->acked_features;
> +}
> +
> +static void vhost_vdpa_del(NetClientState *ncs)
> +{
> + VhostVDPAState *s;
> +
> + assert(ncs->info->type == NET_CLIENT_DRIVER_VHOST_VDPA);
> +
> + s = DO_UPCAST(VhostVDPAState, nc, ncs);
> +
> + if (s->vhost_net) {
> + /* save acked features */
> + uint64_t features = vhost_net_get_acked_features(s->vhost_net);
> + if (features) {
> + s->acked_features = features;
> + }
> + vhost_net_cleanup(s->vhost_net);
> + }
> +}
> +
> +static int vhost_vdpa_add(NetClientState *ncs, void *be)
> +{
> + VhostNetOptions options;
> + struct vhost_net *net = NULL;
> + VhostVDPAState *s;
> +
> + options.backend_type = VHOST_BACKEND_TYPE_VDPA;
> +
> + assert(ncs->info->type == NET_CLIENT_DRIVER_VHOST_VDPA);
> +
> + s = DO_UPCAST(VhostVDPAState, nc, ncs);
> +
> + options.net_backend = ncs;
> + options.opaque = be;
> + options.busyloop_timeout = 0;
> + net = vhost_net_init(&options);
> + if (!net) {
> + error_report("failed to init vhost_net for queue");
> + goto err;
> + }
> +
> + if (s->vhost_net) {
> + vhost_net_cleanup(s->vhost_net);
> + g_free(s->vhost_net);
> + }
> + s->vhost_net = net;
> +
> + return 0;
> +
> +err:
> + if (net) {
> + vhost_net_cleanup(net);
> + }
> + vhost_vdpa_del(ncs);
> + return -1;
> +}
> +static void vhost_vdpa_cleanup(NetClientState *nc)
> +{
> + VhostVDPAState *s = DO_UPCAST(VhostVDPAState, nc, nc);
> +
> + if (s->vhost_net) {
> + vhost_net_cleanup(s->vhost_net);
> + g_free(s->vhost_net);
> + s->vhost_net = NULL;
> + }
> +
> + qemu_purge_queued_packets(nc);
> +}
> +
> +static bool vhost_vdpa_has_vnet_hdr(NetClientState *nc)
> +{
> + assert(nc->info->type == NET_CLIENT_DRIVER_VHOST_VDPA);
> +
> + return true;
> +}
> +
> +static bool vhost_vdpa_has_ufo(NetClientState *nc)
> +{
> + assert(nc->info->type == NET_CLIENT_DRIVER_VHOST_VDPA);
> + VhostVDPAState *s = DO_UPCAST(VhostVDPAState, nc, nc);
> + uint64_t features = 0;
> +
> + features |= (1ULL << VIRTIO_NET_F_HOST_UFO);
> + features = vhost_net_get_features(s->vhost_net, features);
> + return !!(features & (1ULL << VIRTIO_NET_F_HOST_UFO));
> +
> +}
> +
> +static int vhost_vdpa_check_device_id(NetClientState *nc)
> +{
> + uint32_t device_id;
> + int ret;
> + assert(nc->info->type == NET_CLIENT_DRIVER_VHOST_VDPA);
> + VhostVDPAState *s = DO_UPCAST(VhostVDPAState, nc, nc);
> + /* Get the device id from hw*/
> + ret = vhost_net_get_device_id(s->vhost_net, &device_id);
> + if (device_id != VIRTIO_ID_NET) {
> + return -ENOTSUP;
> + }
> + return ret;
> +}
> +
> +static NetClientInfo net_vhost_vdpa_info = {
> + .type = NET_CLIENT_DRIVER_VHOST_VDPA,
> + .size = sizeof(VhostVDPAState),
> + .cleanup = vhost_vdpa_cleanup,
> + .has_vnet_hdr = vhost_vdpa_has_vnet_hdr,
> + .has_ufo = vhost_vdpa_has_ufo,
> +};
> +
> +static int net_vhost_vdpa_init(NetClientState *peer, const char *device,
> + const char *name, const char *vhostdev,
> + bool has_fd, char *fd)
> +{
> + NetClientState *nc = NULL;
> + VhostVDPAState *s;
> + int vdpa_device_fd = -1;
> + Error *err = NULL;
> +
> + assert(name);
> +
> + nc = qemu_new_net_client(&net_vhost_vdpa_info, peer, device, name);
> + snprintf(nc->info_str, sizeof(nc->info_str), "vhost-vdpa");
> + nc->queue_index = 0;
> +
> + s = DO_UPCAST(VhostVDPAState, nc, nc);
> +
> + if (has_fd) {
> + vdpa_device_fd = monitor_fd_param(cur_mon, fd, &err);
> + } else{
> + vdpa_device_fd = open(vhostdev, O_RDWR);
> + }
> +
> + if (vdpa_device_fd == -1) {
> + return -errno;
> + }
> + s->vhost_vdpa.device_fd = vdpa_device_fd;
> + vhost_vdpa_add(nc, (void *)&s->vhost_vdpa);
> + assert(s->vhost_net);
> + /* check the device id for vdpa */
> + return vhost_vdpa_check_device_id(nc);
We probably need to the check earlier. If we do things like this, we
will probably leak vhost_device_fd.
> +}
> +
> +static int net_vhost_check_net(void *opaque, QemuOpts *opts, Error **errp)
> +{
> + const char *name = opaque;
> + const char *driver, *netdev;
> +
> + driver = qemu_opt_get(opts, "driver");
> + netdev = qemu_opt_get(opts, "netdev");
> + if (!driver || !netdev) {
> + return 0;
> + }
> +
> + if (strcmp(netdev, name) == 0 &&
> + !g_str_has_prefix(driver, "virtio-net-")) {
> + error_setg(errp, "vhost-vdpa requires frontend driver virtio-net-*");
> + return -1;
> + }
> +
> + return 0;
> +}
> +
> +int net_init_vhost_vdpa(const Netdev *netdev, const char *name,
> + NetClientState *peer, Error **errp)
> +{
> + const NetdevVhostVDPAOptions *opts;
> +
> + assert(netdev->type == NET_CLIENT_DRIVER_VHOST_VDPA);
> + opts = &netdev->u.vhost_vdpa;
> + /* verify net frontend */
> + if (qemu_opts_foreach(qemu_find_opts("device"), net_vhost_check_net,
> + (char *)name, errp)) {
> + return -1;
> + }
> + return net_vhost_vdpa_init(peer, "vhost_vdpa", name, opts->vhostdev,
> + opts->has_fd, opts->fd);
> +}
> diff --git a/qapi/net.json b/qapi/net.json
> index 335295be50..0f4fa6fffc 100644
> --- a/qapi/net.json
> +++ b/qapi/net.json
> @@ -441,6 +441,23 @@
> '*queues': 'int' } }
>
> ##
> +# @NetdevVhostVDPAOptions:
> +#
> +# Vhost-vdpa network backend
> +#
> +# @vhostdev: name of a vdpa dev path in sysfs
> +#
> +# @queues: number of queues to be created for multiqueue vhost-vdpa
> +# (default: 1) (Since 5.1)
> +#
> +# Since: 5.1
> +##
> +{ 'struct': 'NetdevVhostVDPAOptions',
> + 'data': {
> + '*vhostdev': 'str',
> + '*fd': 'str',
> + '*queues': 'int' } }
> +##
> # @NetClientDriver:
> #
> # Available netdev drivers.
> @@ -451,7 +468,7 @@
> ##
> { 'enum': 'NetClientDriver',
> 'data': [ 'none', 'nic', 'user', 'tap', 'l2tpv3', 'socket', 'vde',
> - 'bridge', 'hubport', 'netmap', 'vhost-user' ] }
> + 'bridge', 'hubport', 'netmap', 'vhost-user', 'vhost-vdpa' ] }
>
> ##
> # @Netdev:
> @@ -479,7 +496,8 @@
> 'bridge': 'NetdevBridgeOptions',
> 'hubport': 'NetdevHubPortOptions',
> 'netmap': 'NetdevNetmapOptions',
> - 'vhost-user': 'NetdevVhostUserOptions' } }
> + 'vhost-user': 'NetdevVhostUserOptions',
> + 'vhost-vdpa': 'NetdevVhostVDPAOptions' } }
>
> ##
> # @NetLegacy:
> diff --git a/qemu-options.hx b/qemu-options.hx
> index 65c9473b73..08256d9d4e 100644
> --- a/qemu-options.hx
> +++ b/qemu-options.hx
> @@ -2291,6 +2291,10 @@ DEF("netdev", HAS_ARG, QEMU_OPTION_netdev,
> #ifdef CONFIG_POSIX
> "-netdev vhost-user,id=str,chardev=dev[,vhostforce=on|off]\n"
> " configure a vhost-user network, backed by a chardev 'dev'\n"
> +#endif
> +#ifdef CONFIG_POSIX
> + "-netdev vhost-vdpa,id=str,vhostdev=/path/to/dev\n"
> + " configure a vhost-vdpa network, backed by a vhostdev 'dev'\n"
> #endif
> "-netdev hubport,id=str,hubid=n[,netdev=nd]\n"
> " configure a hub port on the hub with ID 'n'\n", QEMU_ARCH_ALL)
> @@ -2310,6 +2314,9 @@ DEF("nic", HAS_ARG, QEMU_OPTION_nic,
> #endif
> #ifdef CONFIG_POSIX
> "vhost-user|"
> +#endif
> +#ifdef CONFIG_POSIX
> + "vhost-vdpa|"
> #endif
> "socket][,option][,...][mac=macaddr]\n"
> " initialize an on-board / default host NIC (using MAC address\n"
> @@ -2749,6 +2756,18 @@ qemu -m 512 -object memory-backend-file,id=mem,size=512M,mem-path=/hugetlbfs,sha
> -device virtio-net-pci,netdev=net0
> @end example
>
> +@item -netdev vhost-vdpa,vhostdev=/path/to/dev
> +Establish a vhost-vdpa netdev, backed by a vhostdev. The chardev should
> +be a unix domain socket backed one.
This seems wrong, we don't use unix domain socket.
Thanks
> The vhost-vdpa uses a specifically defined
> +protocol to pass vhost ioctl replacement messages to an application on the other
> +end of the socket.
> +Example:
> +@example
> +qemu -m 512 -object memory-backend-file,id=mem,size=512M,mem-path=/hugetlbfs,share=on \
> + -numa node,memdev=mem \
> + -netdev type=vhost-vdpa,id=net0,vhostdev=/path/to/dev \
> + -device virtio-net-pci,netdev=net0
> +@end example
> @item -netdev hubport,id=@var{id},hubid=@var{hubid}[,netdev=@var{nd}]
>
> Create a hub port on the emulated hub with ID @var{hubid}.
next prev parent reply other threads:[~2020-05-09 2:41 UTC|newest]
Thread overview: 33+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-05-08 16:32 [RFC v2 0/9] vDPA support in qemu Cindy Lu
2020-05-08 16:32 ` [RFC v2 1/9] net: introduce qemu_get_peer Cindy Lu
2020-05-08 16:32 ` [RFC v2 2/9] net: use the function qemu_get_peer Cindy Lu
2020-05-09 2:19 ` Jason Wang
2020-05-09 6:49 ` Cindy Lu
2020-05-08 16:32 ` [RFC v2 3/9] virtio_net: introduce vhost_set_state Cindy Lu
2020-05-09 2:25 ` Jason Wang
2020-05-09 7:08 ` Cindy Lu
2020-05-08 16:32 ` [RFC v2 4/9] vhost-vdpa: introduce vhost-vdpa net client Cindy Lu
2020-05-08 16:41 ` Eric Blake
2020-05-09 7:17 ` Cindy Lu
2020-05-09 2:40 ` Jason Wang [this message]
2020-05-09 7:31 ` Cindy Lu
2020-05-08 16:32 ` [RFC v2 5/9] vhost-vdpa: implement vhost-vdpa backend Cindy Lu
2020-05-09 3:00 ` Jason Wang
2020-05-09 3:07 ` Jason Wang
2020-05-09 8:14 ` Cindy Lu
2020-05-21 12:40 ` Stefan Hajnoczi
2020-05-25 15:33 ` Cindy Lu
2020-05-25 16:15 ` Stefan Hajnoczi
2020-05-08 16:32 ` [RFC v2 6/9] virtio-bus: introduce queue_enabled method Cindy Lu
2020-05-09 3:01 ` Jason Wang
2020-05-09 6:50 ` Cindy Lu
2020-05-08 16:32 ` [RFC v2 7/9] virito-pci: implement " Cindy Lu
2020-05-09 3:02 ` Jason Wang
2020-05-09 12:01 ` Philippe Mathieu-Daudé
2020-05-11 5:08 ` Cindy Lu
2020-05-08 16:32 ` [RFC v2 8/9] vhost_net: set vq ready during start if necessary Cindy Lu
2020-05-09 3:03 ` Jason Wang
2020-05-09 6:51 ` Cindy Lu
2020-05-08 16:32 ` [RFC v2 9/9] vhost: introduce vhost_set_vring_ready method Cindy Lu
2020-05-09 3:05 ` Jason Wang
2020-05-09 3:10 ` [RFC v2 0/9] vDPA support in qemu Jason Wang
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=570f92fa-dbf4-e9d0-e00d-18f02fafd0ce@redhat.com \
--to=jasowang@redhat.com \
--cc=aadam@redhat.com \
--cc=armbru@redhat.com \
--cc=cohuck@redhat.com \
--cc=cunming.liang@intel.com \
--cc=eblake@redhat.com \
--cc=eperezma@redhat.com \
--cc=gdawar@xilinx.com \
--cc=hanand@xilinx.com \
--cc=hch@infradead.org \
--cc=jgg@mellanox.com \
--cc=jiri@mellanox.com \
--cc=kevin.tian@intel.com \
--cc=lingshan.zhu@intel.com \
--cc=lulu@redhat.com \
--cc=maxime.coquelin@redhat.com \
--cc=mhabets@solarflare.com \
--cc=mst@redhat.com \
--cc=parav@mellanox.com \
--cc=qemu-devel@nongnu.org \
--cc=rdunlap@infradead.org \
--cc=rob.miller@broadcom.com \
--cc=saugatm@xilinx.com \
--cc=shahafs@mellanox.com \
--cc=stefanha@redhat.com \
--cc=tiwei.bie@intel.com \
--cc=vmireyno@marvell.com \
--cc=xiao.w.wang@intel.com \
--cc=zhihong.wang@intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).