From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-8.1 required=3.0 tests=DKIM_INVALID,DKIM_SIGNED, HEADER_FROM_DIFFERENT_DOMAINS,INCLUDES_PATCH,MAILING_LIST_MULTI,SIGNED_OFF_BY, SPF_HELO_NONE,SPF_PASS,USER_AGENT_SANE_1 autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 20979C38A2A for ; Sat, 9 May 2020 03:02:50 +0000 (UTC) Received: from lists.gnu.org (lists.gnu.org [209.51.188.17]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPS id D144C21655 for ; Sat, 9 May 2020 03:02:49 +0000 (UTC) Authentication-Results: mail.kernel.org; dkim=fail reason="signature verification failed" (1024-bit key) header.d=redhat.com header.i=@redhat.com header.b="JqQ/05JN" DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org D144C21655 Authentication-Results: mail.kernel.org; dmarc=fail (p=none dis=none) header.from=redhat.com Authentication-Results: mail.kernel.org; spf=pass smtp.mailfrom=qemu-devel-bounces+qemu-devel=archiver.kernel.org@nongnu.org Received: from localhost ([::1]:49730 helo=lists1p.gnu.org) by lists.gnu.org with esmtp (Exim 4.90_1) (envelope-from ) id 1jXFlJ-00040k-3C for qemu-devel@archiver.kernel.org; Fri, 08 May 2020 23:02:49 -0400 Received: from eggs.gnu.org ([2001:470:142:3::10]:39960) by lists.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1jXFjH-0001ZP-0h for qemu-devel@nongnu.org; Fri, 08 May 2020 23:00:43 -0400 Received: from us-smtp-1.mimecast.com ([207.211.31.81]:35515 helo=us-smtp-delivery-1.mimecast.com) by eggs.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_CBC_SHA1:256) (Exim 4.90_1) (envelope-from ) id 1jXFjD-0001Eh-Vt for qemu-devel@nongnu.org; Fri, 08 May 2020 23:00:42 -0400 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1588993238; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=+NxIzCdwBZYDTajFz8MOs4MhTjPLKkzURoGFeEBZhgA=; b=JqQ/05JNcJf1owBeYgTC8y/99FuQRNMka4NSaLH87iRKLPOBtMOpOLUWn6aqAF0zYJOhJC dof17QR5ux466LnDhJ8bEKcz4+obMyrL8A6tVreyjtjZ8nNRNRcIk/DzI6/pvbMVbllCBh vRRDA3zC4vl3zFB35P+IcCrULhRi1YQ= Received: from mimecast-mx01.redhat.com (mimecast-mx01.redhat.com [209.132.183.4]) (Using TLS) by relay.mimecast.com with ESMTP id us-mta-157-AbVBhUsdN32VUkKxj9XN_Q-1; Fri, 08 May 2020 23:00:34 -0400 X-MC-Unique: AbVBhUsdN32VUkKxj9XN_Q-1 Received: from smtp.corp.redhat.com (int-mx04.intmail.prod.int.phx2.redhat.com [10.5.11.14]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mimecast-mx01.redhat.com (Postfix) with ESMTPS id B1CE5835B40; Sat, 9 May 2020 03:00:31 +0000 (UTC) Received: from [10.72.13.128] (ovpn-13-128.pek2.redhat.com [10.72.13.128]) by smtp.corp.redhat.com (Postfix) with ESMTP id 706495D9CD; Sat, 9 May 2020 03:00:13 +0000 (UTC) Subject: Re: [RFC v2 5/9] vhost-vdpa: implement vhost-vdpa backend To: Cindy Lu , mst@redhat.com, armbru@redhat.com, eblake@redhat.com, cohuck@redhat.com References: <20200508163218.22592-1-lulu@redhat.com> <20200508163218.22592-6-lulu@redhat.com> From: Jason Wang Message-ID: Date: Sat, 9 May 2020 11:00:11 +0800 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:68.0) Gecko/20100101 Thunderbird/68.7.0 MIME-Version: 1.0 In-Reply-To: <20200508163218.22592-6-lulu@redhat.com> Content-Language: en-US X-Scanned-By: MIMEDefang 2.79 on 10.5.11.14 X-Mimecast-Spam-Score: 0 X-Mimecast-Originator: redhat.com Content-Type: text/plain; charset=utf-8; format=flowed Content-Transfer-Encoding: 8bit Received-SPF: pass client-ip=207.211.31.81; envelope-from=jasowang@redhat.com; helo=us-smtp-delivery-1.mimecast.com X-detected-operating-system: by eggs.gnu.org: First seen = 2020/05/08 22:40:43 X-ACL-Warn: Detected OS = Linux 2.2.x-3.x [generic] [fuzzy] X-Spam_score_int: -20 X-Spam_score: -2.1 X-Spam_bar: -- X-Spam_report: (-2.1 / 5.0 requ) BAYES_00=-1.9, DKIMWL_WL_HIGH=0.001, DKIM_SIGNED=0.1, DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, DKIM_VALID_EF=-0.1, RCVD_IN_DNSWL_NONE=-0.0001, RCVD_IN_MSPIKE_H3=0.001, RCVD_IN_MSPIKE_WL=0.001, SPF_PASS=-0.001, URIBL_BLOCKED=0.001 autolearn=_AUTOLEARN X-Spam_action: no action X-BeenThere: qemu-devel@nongnu.org X-Mailman-Version: 2.1.23 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: mhabets@solarflare.com, qemu-devel@nongnu.org, rob.miller@broadcom.com, saugatm@xilinx.com, hanand@xilinx.com, hch@infradead.org, eperezma@redhat.com, jgg@mellanox.com, shahafs@mellanox.com, kevin.tian@intel.com, parav@mellanox.com, vmireyno@marvell.com, cunming.liang@intel.com, gdawar@xilinx.com, jiri@mellanox.com, xiao.w.wang@intel.com, stefanha@redhat.com, zhihong.wang@intel.com, Tiwei Bie , aadam@redhat.com, rdunlap@infradead.org, maxime.coquelin@redhat.com, lingshan.zhu@intel.com Errors-To: qemu-devel-bounces+qemu-devel=archiver.kernel.org@nongnu.org Sender: "Qemu-devel" On 2020/5/9 上午12:32, Cindy Lu wrote: > From: Tiwei Bie > > Currently we have 2 types of vhost backends in QEMU: vhost kernel and > vhost-user. The above patch provides a generic device for vDPA purpose, > this vDPA device exposes to user space a non-vendor-specific configuration > interface for setting up a vhost HW accelerator, this patch set introduces > a third vhost backend called vhost-vdpa based on the vDPA interface. > > Vhost-vdpa usage: > > qemu-system-x86_64 -cpu host -enable-kvm \ > ...... > -netdev type=vhost-vdpa,vhostdev=/dev/vhost-vdpa-id,id=vhost-vdpa0 \ > -device virtio-net-pci,netdev=vhost-vdpa0,page-per-vq=on \ > > Co-Authored-By: Lingshan zhu > Signed-off-by: Cindy Lu Signed-off-by: Jason Wang > --- > hw/net/vhost_net.c | 39 ++- > hw/virtio/Makefile.objs | 1 + > hw/virtio/vhost-backend.c | 5 + > hw/virtio/vhost-vdpa.c | 447 ++++++++++++++++++++++++++++++ > hw/virtio/vhost.c | 14 + > include/hw/virtio/vhost-backend.h | 8 +- > include/hw/virtio/vhost-vdpa.h | 25 ++ > include/hw/virtio/vhost.h | 1 + > 8 files changed, 538 insertions(+), 2 deletions(-) > create mode 100644 hw/virtio/vhost-vdpa.c > create mode 100644 include/hw/virtio/vhost-vdpa.h > > diff --git a/hw/net/vhost_net.c b/hw/net/vhost_net.c > index 63b2a85d6e..1af39abaf3 100644 > --- a/hw/net/vhost_net.c > +++ b/hw/net/vhost_net.c > @@ -17,8 +17,10 @@ > #include "net/net.h" > #include "net/tap.h" > #include "net/vhost-user.h" > +#include "net/vhost-vdpa.h" > > #include "standard-headers/linux/vhost_types.h" > +#include "linux-headers/linux/vhost.h" > #include "hw/virtio/virtio-net.h" > #include "net/vhost_net.h" > #include "qemu/error-report.h" > @@ -85,6 +87,29 @@ static const int user_feature_bits[] = { > VHOST_INVALID_FEATURE_BIT > }; > > +static const int vdpa_feature_bits[] = { > + VIRTIO_F_NOTIFY_ON_EMPTY, > + VIRTIO_RING_F_INDIRECT_DESC, > + VIRTIO_RING_F_EVENT_IDX, > + VIRTIO_F_ANY_LAYOUT, > + VIRTIO_F_VERSION_1, > + VIRTIO_NET_F_CSUM, > + VIRTIO_NET_F_GUEST_CSUM, > + VIRTIO_NET_F_GSO, > + VIRTIO_NET_F_GUEST_TSO4, > + VIRTIO_NET_F_GUEST_TSO6, > + VIRTIO_NET_F_GUEST_ECN, > + VIRTIO_NET_F_GUEST_UFO, > + VIRTIO_NET_F_HOST_TSO4, > + VIRTIO_NET_F_HOST_TSO6, > + VIRTIO_NET_F_HOST_ECN, > + VIRTIO_NET_F_HOST_UFO, > + VIRTIO_NET_F_MRG_RXBUF, > + VIRTIO_NET_F_MTU, > + VIRTIO_F_IOMMU_PLATFORM, > + VIRTIO_NET_F_GUEST_ANNOUNCE, > + VHOST_INVALID_FEATURE_BIT > +}; The framework should be ready for packed ring, let's add that. > static const int *vhost_net_get_feature_bits(struct vhost_net *net) > { > const int *feature_bits = 0; > @@ -96,6 +121,9 @@ static const int *vhost_net_get_feature_bits(struct vhost_net *net) > case NET_CLIENT_DRIVER_VHOST_USER: > feature_bits = user_feature_bits; > break; > + case NET_CLIENT_DRIVER_VHOST_VDPA: > + feature_bits = vdpa_feature_bits; > + break; > default: > error_report("Feature bits not defined for this type: %d", > net->nc->info->type); > @@ -110,7 +138,10 @@ uint64_t vhost_net_get_features(struct vhost_net *net, uint64_t features) > return vhost_get_features(&net->dev, vhost_net_get_feature_bits(net), > features); > } > - > +int vhost_net_get_device_id(struct vhost_net *net, uint32_t * device_id) > +{ > + return vhost_dev_get_device_id(&net->dev, device_id); > +} > void vhost_net_ack_features(struct vhost_net *net, uint64_t features) > { > net->dev.acked_features = net->dev.backend_features; > @@ -433,6 +464,12 @@ VHostNetState *get_vhost_net(NetClientState *nc) > vhost_net = vhost_user_get_vhost_net(nc); > assert(vhost_net); > break; > +#endif > +#ifdef CONFIG_VHOST_NET_VDPA > + case NET_CLIENT_DRIVER_VHOST_VDPA: > + vhost_net = vhost_vdpa_get_vhost_net(nc); > + assert(vhost_net); > + break; > #endif > default: > break; > diff --git a/hw/virtio/Makefile.objs b/hw/virtio/Makefile.objs > index e2f70fbb89..e7c5d4a862 100644 > --- a/hw/virtio/Makefile.objs > +++ b/hw/virtio/Makefile.objs > @@ -5,6 +5,7 @@ obj-y += virtio.o > obj-$(call lor,$(CONFIG_VHOST_USER),$(CONFIG_VHOST_KERNEL)) += vhost.o vhost-backend.o > common-obj-$(call lnot,$(call lor,$(CONFIG_VHOST_USER),$(CONFIG_VHOST_KERNEL))) += vhost-stub.o > obj-$(CONFIG_VHOST_USER) += vhost-user.o > +obj-$(CONFIG_VHOST_VDPA) += vhost-vdpa.o > > common-obj-$(CONFIG_VIRTIO_RNG) += virtio-rng.o > common-obj-$(CONFIG_VIRTIO_PCI) += virtio-pci.o > diff --git a/hw/virtio/vhost-backend.c b/hw/virtio/vhost-backend.c > index 48905383f8..069ddb423d 100644 > --- a/hw/virtio/vhost-backend.c > +++ b/hw/virtio/vhost-backend.c > @@ -285,6 +285,11 @@ int vhost_set_backend_type(struct vhost_dev *dev, VhostBackendType backend_type) > case VHOST_BACKEND_TYPE_USER: > dev->vhost_ops = &user_ops; > break; > +#endif > +#ifdef CONFIG_VHOST_VDPA > + case VHOST_BACKEND_TYPE_VDPA: > + dev->vhost_ops = &vdpa_ops; > + break; > #endif > default: > error_report("Unknown vhost backend type"); > diff --git a/hw/virtio/vhost-vdpa.c b/hw/virtio/vhost-vdpa.c > new file mode 100644 > index 0000000000..bac8a8aa2a > --- /dev/null > +++ b/hw/virtio/vhost-vdpa.c > @@ -0,0 +1,447 @@ > +/* > + * vhost-vdpa > + * > + * Copyright(c) 2017-2018 Intel Corporation. > + * Copyright(c) 2020 Red Hat, Inc. > + * > + * This work is licensed under the terms of the GNU GPL, version 2 or later. > + * See the COPYING file in the top-level directory. > + * > + */ > + > +#include "qemu/osdep.h" > +#include > +#include > +#include > +#include > +#include "hw/virtio/vhost.h" > +#include "hw/virtio/vhost-backend.h" > +#include "hw/virtio/virtio-net.h" > +#include "hw/virtio/vhost-vdpa.h" > +#include "qemu/main-loop.h" > +#include > +#include "sysemu/kvm.h" > + > + > +static bool vhost_vdpa_listener_skipped_section(MemoryRegionSection *section) > +{ > + return (!memory_region_is_ram(section->mr) && > + !memory_region_is_iommu(section->mr)) || > + /* > + * Sizing an enabled 64-bit BAR can cause spurious mappings to > + * addresses in the upper part of the 64-bit address space. These > + * are never accessed by the CPU and beyond the address width of > + * some IOMMU hardware. TODO: VDPA should tell us the IOMMU width. > + */ > + section->offset_within_address_space & (1ULL << 63); > +} > + > +static int vhost_vdpa_dma_map(struct vhost_vdpa *v, hwaddr iova, hwaddr size, > + void *vaddr, bool readonly) > +{ > + struct vhost_msg_v2 msg; > + int fd = v->device_fd; > + int ret = 0; > + > + msg.type = v->msg_type; > + msg.iotlb.iova = iova; > + msg.iotlb.size = size; > + msg.iotlb.uaddr = (uint64_t)vaddr; > + msg.iotlb.perm = readonly ? VHOST_ACCESS_RO : VHOST_ACCESS_RW; > + msg.iotlb.type = VHOST_IOTLB_UPDATE; > + > + if (write(fd, &msg, sizeof(msg)) != sizeof(msg)) { > + error_report("failed to write, fd=%d, errno=%d (%s)", > + fd, errno, strerror(errno)); > + return -EIO ; > + } > + > + return ret; > +} > + > +static int vhost_vdpa_dma_unmap(struct vhost_vdpa *v, hwaddr iova, > + hwaddr size) > +{ > + struct vhost_msg_v2 msg; > + int fd = v->device_fd; > + int ret = 0; > + > + msg.type = v->msg_type; > + msg.iotlb.iova = iova; > + msg.iotlb.size = size; > + msg.iotlb.type = VHOST_IOTLB_INVALIDATE; > + > + if (write(fd, &msg, sizeof(msg)) != sizeof(msg)) { > + error_report("failed to write, fd=%d, errno=%d (%s)", > + fd, errno, strerror(errno)); > + return -EIO ; > + } > + > + return ret; > +} > + > +static void vhost_vdpa_listener_region_add(MemoryListener *listener, > + MemoryRegionSection *section) > +{ > + struct vhost_vdpa *v = container_of(listener, struct vhost_vdpa, listener); > + hwaddr iova; > + Int128 llend, llsize; > + void *vaddr; > + int ret; > + > + if (vhost_vdpa_listener_skipped_section(section)) { > + return; > + } > + > + if (unlikely((section->offset_within_address_space & ~TARGET_PAGE_MASK) != > + (section->offset_within_region & ~TARGET_PAGE_MASK))) { > + error_report("%s received unaligned region", __func__); > + return; > + } > + > + iova = TARGET_PAGE_ALIGN(section->offset_within_address_space); > + llend = int128_make64(section->offset_within_address_space); > + llend = int128_add(llend, section->size); > + llend = int128_and(llend, int128_exts64(TARGET_PAGE_MASK)); > + > + if (int128_ge(int128_make64(iova), llend)) { > + return; > + } > + > + memory_region_ref(section->mr); > + > + /* Here we assume that memory_region_is_ram(section->mr)==true */ > + > + vaddr = memory_region_get_ram_ptr(section->mr) + > + section->offset_within_region + > + (iova - section->offset_within_address_space); > + > + llsize = int128_sub(llend, int128_make64(iova)); > + > + ret = vhost_vdpa_dma_map(v, iova, int128_get64(llsize), > + vaddr, section->readonly); > + if (ret) { > + error_report("vhost vdpa map fail!"); > + if (memory_region_is_ram_device(section->mr)) { > + /* Allow unexpected mappings not to be fatal for RAM devices */ > + error_report("map ram fail!"); > + return ; > + } > + goto fail; > + } > + > + return; > + > +fail: > + if (memory_region_is_ram_device(section->mr)) { > + error_report("failed to vdpa_dma_map. pci p2p may not work"); > + return; > + > + } > + /* > + * On the initfn path, store the first error in the container so we > + * can gracefully fail. Runtime, there's not much we can do other > + * than throw a hardware error. > + */ > + error_report("vhost-vdpa: DMA mapping failed, unable to continue"); > + return; > + > +} > + > +static void vhost_vdpa_listener_region_del(MemoryListener *listener, > + MemoryRegionSection *section) > +{ > + struct vhost_vdpa *v = container_of(listener, struct vhost_vdpa, listener); > + hwaddr iova; > + Int128 llend, llsize; > + int ret; > + bool try_unmap = true; > + > + if (vhost_vdpa_listener_skipped_section(section)) { > + return; > + } > + > + if (unlikely((section->offset_within_address_space & ~TARGET_PAGE_MASK) != > + (section->offset_within_region & ~TARGET_PAGE_MASK))) { > + error_report("%s received unaligned region", __func__); > + return; > + } > + > + iova = TARGET_PAGE_ALIGN(section->offset_within_address_space); > + llend = int128_make64(section->offset_within_address_space); > + llend = int128_add(llend, section->size); > + llend = int128_and(llend, int128_exts64(TARGET_PAGE_MASK)); > + > + if (int128_ge(int128_make64(iova), llend)) { > + return; > + } > + > + llsize = int128_sub(llend, int128_make64(iova)); > + > + if (try_unmap) { > + ret = vhost_vdpa_dma_unmap(v, iova, int128_get64(llsize)); > + if (ret) { > + error_report("vhost_vdpa dma unmap error!"); > + } > + } > + > + memory_region_unref(section->mr); > +} > + > +static const MemoryListener vhost_vdpa_memory_listener = { > + .region_add = vhost_vdpa_listener_region_add, > + .region_del = vhost_vdpa_listener_region_del, > +}; > + Need comment to explain why vhost-vdpa requires a specific memory listener other than just reusing the one that is provided by vhost. > + This newline is unnecessary. > +static int vhost_vdpa_call(struct vhost_dev *dev, unsigned long int request, > + void *arg) > +{ > + struct vhost_vdpa *v = dev->opaque; > + int fd = v->device_fd; > + > + assert(dev->vhost_ops->backend_type == VHOST_BACKEND_TYPE_VDPA); > + > + return ioctl(fd, request, arg); > +} > + > +static int vhost_vdpa_get_backend_features(struct vhost_dev *dev, > + uint64_t *features) > +{ > + return vhost_vdpa_call(dev, VHOST_GET_BACKEND_FEATURES, features); > +} > + > +static int vhost_vdpa_init(struct vhost_dev *dev, void *opaque) > +{ > + struct vhost_vdpa *v; > + uint64_t backend_features = 0; > + int ret = 0; > + > + assert(dev->vhost_ops->backend_type == VHOST_BACKEND_TYPE_VDPA); > + > + v = opaque; > + dev->opaque = opaque ; > + > + v->listener = vhost_vdpa_memory_listener; > + memory_listener_register(&v->listener, &address_space_memory); > + > + ret = vhost_vdpa_get_backend_features(dev, &backend_features); > + /* > + * In order to compatible with older version kernel, > + * If the kernel not support this ioctl, > + * set the msg_type for v2 by defeault > + */ > + if (ret) { > + v->msg_type = VHOST_IOTLB_MSG_V2; > + return 0; > + } > + if (backend_features & (1ULL << VHOST_BACKEND_F_IOTLB_MSG_V2)) { > + v->msg_type = VHOST_IOTLB_MSG_V2; > + } else { > + v->msg_type = VHOST_IOTLB_MSG; > + } But in vhost_vdpa_dma_unmap/vhost_vdpa_dma_map, v2 is still being used. Consider vdpa implies msg_v2, and it doesn't implement get_backend_features. We use v2 unconditionally and leave this check in the future. > + return 0; > +} > + > +static int vhost_vdpa_cleanup(struct vhost_dev *dev) > +{ > + struct vhost_vdpa *v; > + assert(dev->vhost_ops->backend_type == VHOST_BACKEND_TYPE_VDPA); > + > + v = dev->opaque; > + memory_listener_unregister(&v->listener); > + > + dev->opaque = NULL; > + return 0; > +} > + > +static int vhost_vdpa_memslots_limit(struct vhost_dev *dev) > +{ > + /*Using IOTLB API here,NOTE:please set the ulimit before using*/ The "note" seems unrelated to the memslot here. > + return INT_MAX; > +} > + > +static int vhost_vdpa_set_log_base(struct vhost_dev *dev, uint64_t base, > + struct vhost_log *log) > +{ > + return -1; > +} > + > +static int vhost_vdpa_set_mem_table(struct vhost_dev *dev, > + struct vhost_memory *mem) > +{ > + > + if (mem->padding) { > + return -1; > + } > + > + return 0; > +} > + > +static int vhost_vdpa_set_vring_addr(struct vhost_dev *dev, > + struct vhost_vring_addr *addr) > +{ > + return vhost_vdpa_call(dev, VHOST_SET_VRING_ADDR, addr); > +} > + > +static int vhost_vdpa_set_vring_num(struct vhost_dev *dev, > + struct vhost_vring_state *ring) > +{ > + return vhost_vdpa_call(dev, VHOST_SET_VRING_NUM, ring); > +} > + > +static int vhost_vdpa_set_vring_base(struct vhost_dev *dev, > + struct vhost_vring_state *ring) > +{ > + return vhost_vdpa_call(dev, VHOST_GET_VRING_BASE, ring); > +} > + > +static int vhost_vdpa_get_vring_base(struct vhost_dev *dev, > + struct vhost_vring_state *ring) > +{ > + > + return vhost_vdpa_call(dev, VHOST_GET_VRING_BASE, ring); > +} > + > +static int vhost_vdpa_set_vring_kick(struct vhost_dev *dev, > + struct vhost_vring_file *file) > +{ > + return vhost_vdpa_call(dev, VHOST_SET_VRING_KICK, file); > +} > + > +static int vhost_vdpa_set_vring_call(struct vhost_dev *dev, > + struct vhost_vring_file *file) > +{ > + return vhost_vdpa_call(dev, VHOST_SET_VRING_CALL, file); > +} > + > +static int vhost_vdpa_set_features(struct vhost_dev *dev, > + uint64_t features) > +{ > + return vhost_vdpa_call(dev, VHOST_SET_FEATURES, &features); > +} > + > +static int vhost_vdpa_get_features(struct vhost_dev *dev, > + uint64_t *features) > +{ > + return vhost_vdpa_call(dev, VHOST_GET_FEATURES, features); > +} > + > +static int vhost_vdpa_get_device_id(struct vhost_dev *dev, > + uint32_t *device_id) > +{ > + return vhost_vdpa_call(dev, VHOST_VDPA_GET_DEVICE_ID, device_id); > +} > + > +static int vhost_vdpa_set_owner(struct vhost_dev *dev) > +{ > + return vhost_vdpa_call(dev, VHOST_SET_OWNER, NULL); > +} > + > +static int vhost_vdpa_reset_device(struct vhost_dev *dev) > +{ > + return vhost_vdpa_call(dev, VHOST_RESET_OWNER, NULL); > +} I think we may want to reuse the helpers of kenrel backends instead of duplicating functions here. > + > +static int vhost_vdpa_get_vq_index(struct vhost_dev *dev, int idx) > +{ > + assert(idx >= dev->vq_index && idx < dev->vq_index + dev->nvqs); > + > + return idx - dev->vq_index; > +} > + > +static int vhost_vdpa_set_vring_ready(struct vhost_dev *dev) > +{ > + int i; > + for (i = 0; i < dev->nvqs; ++i) { > + struct vhost_vring_state state = { > + .index = dev->vq_index + i, > + .num = 1, > + }; > + vhost_vdpa_call(dev, VHOST_VDPA_SET_VRING_ENABLE, &state); > + } > + return 0; > +} > +static int vhost_vdpa_set_config(struct vhost_dev *dev, const uint8_t *data, One unnecessary space between int and "vhost_vdpa_set_config" > + uint32_t offset, uint32_t size, > + uint32_t flags) > +{ > + > + struct vhost_vdpa_config *config; > + int ret; > + if ((size > VHOST_VDPA_MAX_CONFIG_SIZE) || (data == NULL)) { > + return -1; > + } > + config = g_new0(struct vhost_vdpa_config, 1); It's better to just use a temporary variable instead of allocate it dynamically? > + if (config == NULL) { > + return -EINVAL; > + } > + config->off = 0; > + config->len = size; > + memcpy(config->buf, data, size); > + ret = vhost_vdpa_call(dev, VHOST_VDPA_SET_CONFIG, config); > + g_free(config); > + return ret; > +} > + > + static int vhost_vdpa_get_config(struct vhost_dev *dev, uint8_t *config, > + uint32_t config_len) > +{ > + struct vhost_vdpa_config *v_config = NULL; > + int ret; > + > + if ((config_len > VHOST_VDPA_MAX_CONFIG_SIZE) || (config == NULL)) { > + return -1; > + } > + v_config = g_new0(struct vhost_vdpa_config, 1); Let's use a variable on the stack. > + > + if (v_config == NULL) { > + return -EINVAL; > + } > + ret = vhost_vdpa_call(dev, VHOST_VDPA_GET_CONFIG, v_config); One extra space after '='. > + if ((v_config->len > config_len - v_config->off) || (v_config->len == 0)) { > + g_free(v_config); > + return -EINVAL; > + } Indentation is wired and why need such check? > + > + memcpy(config, v_config->buf + v_config->off, config_len); > + g_free(v_config); > + return ret; > + } Newline is needed here. > +static int vhost_vdpa_set_state(struct vhost_dev *dev, uint8_t state) > +{ > + return vhost_vdpa_call(dev, VHOST_VDPA_SET_STATUS, &state); > +} > + > + This new line is not needed. > +const VhostOps vdpa_ops = { > + .backend_type = VHOST_BACKEND_TYPE_VDPA, > + .vhost_backend_init = vhost_vdpa_init, > + .vhost_backend_cleanup = vhost_vdpa_cleanup, > + .vhost_backend_memslots_limit = vhost_vdpa_memslots_limit, > + .vhost_set_log_base = vhost_vdpa_set_log_base, > + .vhost_set_mem_table = vhost_vdpa_set_mem_table, > + .vhost_set_vring_addr = vhost_vdpa_set_vring_addr, > + .vhost_set_vring_endian = NULL, > + .vhost_set_vring_num = vhost_vdpa_set_vring_num, > + .vhost_set_vring_base = vhost_vdpa_set_vring_base, > + .vhost_get_vring_base = vhost_vdpa_get_vring_base, > + .vhost_set_vring_kick = vhost_vdpa_set_vring_kick, > + .vhost_set_vring_call = vhost_vdpa_set_vring_call, > + .vhost_set_features = vhost_vdpa_set_features, > + .vhost_get_features = vhost_vdpa_get_features, > + .vhost_set_owner = vhost_vdpa_set_owner, > + .vhost_reset_device = vhost_vdpa_reset_device, > + .vhost_get_vq_index = vhost_vdpa_get_vq_index, > + .vhost_set_vring_ready = vhost_vdpa_set_vring_ready, > + .vhost_get_config = vhost_vdpa_get_config, > + .vhost_set_config = vhost_vdpa_set_config, > + .vhost_requires_shm_log = NULL, > + .vhost_migration_done = NULL, > + .vhost_backend_can_merge = NULL, > + .vhost_net_set_mtu = NULL, > + .vhost_set_iotlb_callback = NULL, > + .vhost_send_device_iotlb_msg = NULL, > + .vhost_set_state = vhost_vdpa_set_state, > + .vhost_get_device_id = vhost_vdpa_get_device_id, > +}; As said previously, need to consider to reuse vhost kernel helpers. > diff --git a/hw/virtio/vhost.c b/hw/virtio/vhost.c > index 4da0d5a6c5..6d2aa461d6 100644 > --- a/hw/virtio/vhost.c > +++ b/hw/virtio/vhost.c > @@ -746,6 +746,12 @@ static int vhost_virtqueue_set_addr(struct vhost_dev *dev, > .log_guest_addr = vq->used_phys, > .flags = enable_log ? (1 << VHOST_VRING_F_LOG) : 0, > }; > + /*vDPA need to use the phys address here to set to hardware*/ > + if (dev->vhost_ops->backend_type == VHOST_BACKEND_TYPE_VDPA) { > + addr.desc_user_addr = (uint64_t)(unsigned long)vq->desc_phys; > + addr.avail_user_addr = (uint64_t)(unsigned long)vq->avail_phys; > + addr.used_user_addr = (uint64_t)(unsigned long)vq->used_phys; > + } > int r = dev->vhost_ops->vhost_set_vring_addr(dev, &addr); > if (r < 0) { > VHOST_OPS_DEBUG("vhost_set_vring_addr failed"); > @@ -1493,6 +1499,14 @@ int vhost_dev_set_config(struct vhost_dev *hdev, const uint8_t *data, > return -1; > } > > +int vhost_dev_get_device_id(struct vhost_dev *hdev, uint32_t *device_id) > +{ > + assert(hdev->vhost_ops); > + if (hdev->vhost_ops->vhost_get_device_id) { If you introduce a new vhost_ops, please use a separate patch. > + return hdev->vhost_ops->vhost_get_device_id(hdev, device_id); > + } > + return -1; > +} > void vhost_dev_set_config_notifier(struct vhost_dev *hdev, > const VhostDevConfigOps *ops) > { > diff --git a/include/hw/virtio/vhost-backend.h b/include/hw/virtio/vhost-backend.h > index f823055167..8ac898a987 100644 > --- a/include/hw/virtio/vhost-backend.h > +++ b/include/hw/virtio/vhost-backend.h > @@ -17,7 +17,8 @@ typedef enum VhostBackendType { > VHOST_BACKEND_TYPE_NONE = 0, > VHOST_BACKEND_TYPE_KERNEL = 1, > VHOST_BACKEND_TYPE_USER = 2, > - VHOST_BACKEND_TYPE_MAX = 3, > + VHOST_BACKEND_TYPE_VDPA = 3, > + VHOST_BACKEND_TYPE_MAX = 4, > } VhostBackendType; > > typedef enum VhostSetConfigType { > @@ -77,6 +78,7 @@ typedef int (*vhost_reset_device_op)(struct vhost_dev *dev); > typedef int (*vhost_get_vq_index_op)(struct vhost_dev *dev, int idx); > typedef int (*vhost_set_vring_enable_op)(struct vhost_dev *dev, > int enable); > +typedef int (*vhost_set_vring_ready_op)(struct vhost_dev *dev); A new patch for this patch. Thanks > typedef bool (*vhost_requires_shm_log_op)(struct vhost_dev *dev); > typedef int (*vhost_migration_done_op)(struct vhost_dev *dev, > char *mac_addr); > @@ -113,6 +115,7 @@ typedef int (*vhost_set_inflight_fd_op)(struct vhost_dev *dev, > struct vhost_inflight *inflight); > > typedef int (*vhost_set_state_op)(struct vhost_dev *dev, uint8_t state); > +typedef int (*vhost_get_device_id_op)(struct vhost_dev *dev, uint32_t *dev_id); > typedef struct VhostOps { > VhostBackendType backend_type; > vhost_backend_init vhost_backend_init; > @@ -139,6 +142,7 @@ typedef struct VhostOps { > vhost_reset_device_op vhost_reset_device; > vhost_get_vq_index_op vhost_get_vq_index; > vhost_set_vring_enable_op vhost_set_vring_enable; > + vhost_set_vring_ready_op vhost_set_vring_ready; > vhost_requires_shm_log_op vhost_requires_shm_log; > vhost_migration_done_op vhost_migration_done; > vhost_backend_can_merge_op vhost_backend_can_merge; > @@ -154,9 +158,11 @@ typedef struct VhostOps { > vhost_get_inflight_fd_op vhost_get_inflight_fd; > vhost_set_inflight_fd_op vhost_set_inflight_fd; > vhost_set_state_op vhost_set_state; > + vhost_get_device_id_op vhost_get_device_id; > } VhostOps; > > extern const VhostOps user_ops; > +extern const VhostOps vdpa_ops; > > int vhost_set_backend_type(struct vhost_dev *dev, > VhostBackendType backend_type); > diff --git a/include/hw/virtio/vhost-vdpa.h b/include/hw/virtio/vhost-vdpa.h > new file mode 100644 > index 0000000000..9d64fbfea9 > --- /dev/null > +++ b/include/hw/virtio/vhost-vdpa.h > @@ -0,0 +1,25 @@ > +/* > + * vhost-vdpa.h > + * > + * Copyright(c) 2017-2018 Intel Corporation. > + * Copyright(c) 2020 Red Hat, Inc. > + * > + * This work is licensed under the terms of the GNU GPL, version 2 or later. > + * See the COPYING file in the top-level directory. > + * > + */ > + > +#ifndef HW_VIRTIO_VHOST_VDPA_H > +#define HW_VIRTIO_VHOST_VDPA_H > + > +#include "hw/virtio/virtio.h" > + > +typedef struct vhost_vdpa { > + int device_fd; > + uint32_t msg_type; > + MemoryListener listener; > +} VhostVDPA; > + > +extern AddressSpace address_space_memory; > + > +#endif > diff --git a/include/hw/virtio/vhost.h b/include/hw/virtio/vhost.h > index 085450c6f8..020956d5a9 100644 > --- a/include/hw/virtio/vhost.h > +++ b/include/hw/virtio/vhost.h > @@ -124,6 +124,7 @@ int vhost_dev_get_config(struct vhost_dev *dev, uint8_t *config, > uint32_t config_len); > int vhost_dev_set_config(struct vhost_dev *dev, const uint8_t *data, > uint32_t offset, uint32_t size, uint32_t flags); > +int vhost_dev_get_device_id(struct vhost_dev *hdev, uint32_t *device_id); > /* notifier callback in case vhost device config space changed > */ > void vhost_dev_set_config_notifier(struct vhost_dev *dev,