From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-15.1 required=3.0 tests=BAYES_00,DKIM_INVALID, DKIM_SIGNED,HEADER_FROM_DIFFERENT_DOMAINS,INCLUDES_CR_TRAILER,INCLUDES_PATCH, MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS,USER_AGENT_SANE_1 autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id AFF28C48BE6 for ; Wed, 16 Jun 2021 12:37:36 +0000 (UTC) Received: from lists.gnu.org (lists.gnu.org [209.51.188.17]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPS id 1DD6D61351 for ; Wed, 16 Jun 2021 12:37:36 +0000 (UTC) DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org 1DD6D61351 Authentication-Results: mail.kernel.org; dmarc=fail (p=none dis=none) header.from=redhat.com Authentication-Results: mail.kernel.org; spf=pass smtp.mailfrom=qemu-devel-bounces+qemu-devel=archiver.kernel.org@nongnu.org Received: from localhost ([::1]:36174 helo=lists1p.gnu.org) by lists.gnu.org with esmtp (Exim 4.90_1) (envelope-from ) id 1ltUnX-0002gP-Bj for qemu-devel@archiver.kernel.org; Wed, 16 Jun 2021 08:37:35 -0400 Received: from eggs.gnu.org ([2001:470:142:3::10]:33986) by lists.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1ltUmX-0001Zf-Mn for qemu-devel@nongnu.org; Wed, 16 Jun 2021 08:36:33 -0400 Received: from us-smtp-delivery-124.mimecast.com ([170.10.133.124]:36048) by eggs.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1ltUmQ-00053M-On for qemu-devel@nongnu.org; Wed, 16 Jun 2021 08:36:33 -0400 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1623846984; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: in-reply-to:in-reply-to:references:references; bh=8y36fb2x5cbcB43v1CCw0BWOzIq7vl25hvrDmYZ2x4E=; b=aaU1llt3wcIdG6cRqcTmo+g+xIw0O50s5gOYw+Nw8bHkkXX1de1ij2etug+nPWphlEB27i P66owwtbpfaprWbJ63jKsHREClurRgDIdmEs9jXzAu2c5XC6WIqpEYs3SQYxfOsdtwB44M 3Z17yZK0OlSaq4QyqPHPz7BL8Rm9XnU= Received: from mimecast-mx01.redhat.com (mimecast-mx01.redhat.com [209.132.183.4]) (Using TLS) by relay.mimecast.com with ESMTP id us-mta-96-A7Jnqw4GP8O-DavvhA54Jg-1; Wed, 16 Jun 2021 08:36:22 -0400 X-MC-Unique: A7Jnqw4GP8O-DavvhA54Jg-1 Received: from smtp.corp.redhat.com (int-mx01.intmail.prod.int.phx2.redhat.com [10.5.11.11]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mimecast-mx01.redhat.com (Postfix) with ESMTPS id 1FB9FC740C; Wed, 16 Jun 2021 12:36:22 +0000 (UTC) Received: from work-vm (ovpn-115-42.ams2.redhat.com [10.36.115.42]) by smtp.corp.redhat.com (Postfix) with ESMTPS id 993E96062F; Wed, 16 Jun 2021 12:36:13 +0000 (UTC) Date: Wed, 16 Jun 2021 13:36:10 +0100 From: "Dr. David Alan Gilbert" To: Stefan Hajnoczi Subject: Re: [PATCH v3 26/26] virtiofsd: Ask qemu to drop CAP_FSETID if client asked for it Message-ID: References: <20210428110100.27757-1-dgilbert@redhat.com> <20210428110100.27757-27-dgilbert@redhat.com> <20210506160223.GA277745@redhat.com> <20210510152324.GB150402@horse> MIME-Version: 1.0 In-Reply-To: User-Agent: Mutt/2.0.7 (2021-05-04) X-Scanned-By: MIMEDefang 2.79 on 10.5.11.11 Authentication-Results: relay.mimecast.com; auth=pass smtp.auth=CUSA124A263 smtp.mailfrom=dgilbert@redhat.com X-Mimecast-Spam-Score: 0 X-Mimecast-Originator: redhat.com Content-Type: text/plain; charset=us-ascii Content-Disposition: inline Received-SPF: pass client-ip=170.10.133.124; envelope-from=dgilbert@redhat.com; helo=us-smtp-delivery-124.mimecast.com X-Spam_score_int: -29 X-Spam_score: -3.0 X-Spam_bar: --- X-Spam_report: (-3.0 / 5.0 requ) BAYES_00=-1.9, DKIMWL_WL_HIGH=-0.199, DKIM_SIGNED=0.1, DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, DKIM_VALID_EF=-0.1, RCVD_IN_DNSWL_LOW=-0.7, RCVD_IN_MSPIKE_H4=0.001, RCVD_IN_MSPIKE_WL=0.001, SPF_HELO_NONE=0.001, SPF_PASS=-0.001 autolearn=ham autolearn_force=no X-Spam_action: no action X-BeenThere: qemu-devel@nongnu.org X-Mailman-Version: 2.1.23 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: virtio-fs@redhat.com, qemu-devel@nongnu.org, Vivek Goyal , groug@kaod.org Errors-To: qemu-devel-bounces+qemu-devel=archiver.kernel.org@nongnu.org Sender: "Qemu-devel" * Stefan Hajnoczi (stefanha@redhat.com) wrote: > On Thu, Jun 10, 2021 at 04:29:42PM +0100, Dr. David Alan Gilbert wrote: > > * Dr. David Alan Gilbert (dgilbert@redhat.com) wrote: > > > * Stefan Hajnoczi (stefanha@redhat.com) wrote: > > > > > > > > > > Instead I was thinking about VHOST_USER_DMA_READ/WRITE messages > > > > containing the address (a device IOVA, it could just be a guest physical > > > > memory address in most cases) and the length. The WRITE message would > > > > also contain the data that the vhost-user device wishes to write. The > > > > READ message reply would contain the data that the device read from > > > > QEMU. > > > > > > > > QEMU would implement this using QEMU's address_space_read/write() APIs. > > > > > > > > So basically just a new vhost-user protocol message to do a memcpy(), > > > > but with guest addresses and vIOMMU support :). > > > > > > This doesn't actually feel that hard - ignoring vIOMMU for a minute > > > which I know very little about - I'd have to think where the data > > > actually flows, probably the slave fd. > > > > > > > The vhost-user device will need to do bounce buffering so using these > > > > new messages is slower than zero-copy I/O to shared guest RAM. > > > > > > I guess the theory is it's only in the weird corner cases anyway. > > The feature is also useful if DMA isolation is desirable (i.e. > security/reliability are more important than performance). Once this new > vhost-user protocol feature is available it will be possible to run > vhost-user devices without shared memory or with limited shared memory > (e.g. just the vring). I don't see it ever being efficient, so that case is going to be pretty limited. > > The direction I'm going is something like the following; > > the idea is that the master will have to handle the requests on a > > separate thread, to avoid any problems with side effects from the memory > > accesses; the slave will then have to parkt he requests somewhere and > > handle them later. > > > > > > From 07aacff77c50c8a2b588b2513f2dfcfb8f5aa9df Mon Sep 17 00:00:00 2001 > > From: "Dr. David Alan Gilbert" > > Date: Thu, 10 Jun 2021 15:34:04 +0100 > > Subject: [PATCH] WIP: vhost-user: DMA type interface > > > > A DMA type interface where the slave can ask for a stream of bytes > > to be read/written to the guests memory by the master. > > The interface is asynchronous, since a request may have side effects > > inside the guest. > > > > Signed-off-by: Dr. David Alan Gilbert > > --- > > docs/interop/vhost-user.rst | 33 +++++++++++++++++++++++ > > hw/virtio/vhost-user.c | 4 +++ > > subprojects/libvhost-user/libvhost-user.h | 24 +++++++++++++++++ > > 3 files changed, 61 insertions(+) > > Use of the word "RAM" in this patch is a little unclear since we need > these new messages precisely when it's not ordinary guest RAM :-). Maybe > referring to the address space is more general. Yeh, I'll try and spot those. > > diff --git a/docs/interop/vhost-user.rst b/docs/interop/vhost-user.rst > > index 9ebd05e2bf..b9b5322147 100644 > > --- a/docs/interop/vhost-user.rst > > +++ b/docs/interop/vhost-user.rst > > @@ -1347,6 +1347,15 @@ Master message types > > query the backend for its device status as defined in the Virtio > > specification. > > > > +``VHOST_USER_MEM_DATA`` > > + :id: 41 > > + :equivalent ioctl: N/A > > + :slave payload: N/A > > + :master payload: ``struct VhostUserMemReply`` > > + > > + This message is an asynchronous response to a ``VHOST_USER_SLAVE_MEM_ACCESS`` > > + message. Where the request was for the master to read data, this > > + message will be followed by the data that was read. > > Please explain why this message is asynchronous. Implementors will need > to understand the gotchas around deadlocks, etc. I've added: Making this a separate asynchronous response message (rather than just a reply to the ``VHOST_USER_SLAVE_MEM_ACCESS``) makes it much easier for the master to deal with any side effects the access may have, and in particular avoid deadlocks they might cause if an access triggers another vhost_user message. > > > > Slave message types > > ------------------- > > @@ -1469,6 +1478,30 @@ Slave message types > > The ``VHOST_USER_FS_FLAG_MAP_W`` flag must be set in the ``flags`` field to > > write to the file from RAM. > > > > +``VHOST_USER_SLAVE_MEM_ACCESS`` > > + :id: 9 > > + :equivalent ioctl: N/A > > + :slave payload: ``struct VhostUserMemAccess`` > > + :master payload: N/A > > + > > + Requests that the master perform a range of memory accesses on behalf > > + of the slave that the slave can't perform itself. > > + > > + The ``VHOST_USER_MEM_FLAG_TO_MASTER`` flag must be set in the ``flags`` > > + field for the slave to write data into the RAM of the master. In this > > + case the data to write follows the ``VhostUserMemAccess`` on the fd. > > + The ``VHOST_USER_MEM_FLAG_FROM_MASTER`` flag must be set in the ``flags`` > > + field for the slave to read data from the RAM of the master. > > + > > + When the master has completed the access it replies on the main fd with > > + a ``VHOST_USER_MEM_DATA`` message. > > + > > + The master is allowed to complete part of the request and reply stating > > + the amount completed, leaving it to the slave to resend further components. > > + This may happen to limit memory allocations in the master or to simplify > > + the implementation. > > + > > + > > .. _reply_ack: > > > > VHOST_USER_PROTOCOL_F_REPLY_ACK > > diff --git a/hw/virtio/vhost-user.c b/hw/virtio/vhost-user.c > > index 39a0e55cca..a3fefc4c1d 100644 > > --- a/hw/virtio/vhost-user.c > > +++ b/hw/virtio/vhost-user.c > > @@ -126,6 +126,9 @@ typedef enum VhostUserRequest { > > VHOST_USER_GET_MAX_MEM_SLOTS = 36, > > VHOST_USER_ADD_MEM_REG = 37, > > VHOST_USER_REM_MEM_REG = 38, > > + VHOST_USER_SET_STATUS = 39, > > + VHOST_USER_GET_STATUS = 40, > > + VHOST_USER_MEM_DATA = 41, > > VHOST_USER_MAX > > } VhostUserRequest; > > > > @@ -139,6 +142,7 @@ typedef enum VhostUserSlaveRequest { > > VHOST_USER_SLAVE_FS_MAP = 6, > > VHOST_USER_SLAVE_FS_UNMAP = 7, > > VHOST_USER_SLAVE_FS_IO = 8, > > + VHOST_USER_SLAVE_MEM_ACCESS = 9, > > VHOST_USER_SLAVE_MAX > > } VhostUserSlaveRequest; > > > > diff --git a/subprojects/libvhost-user/libvhost-user.h b/subprojects/libvhost-user/libvhost-user.h > > index eee611a2f6..b5444f4f6f 100644 > > --- a/subprojects/libvhost-user/libvhost-user.h > > +++ b/subprojects/libvhost-user/libvhost-user.h > > @@ -109,6 +109,9 @@ typedef enum VhostUserRequest { > > VHOST_USER_GET_MAX_MEM_SLOTS = 36, > > VHOST_USER_ADD_MEM_REG = 37, > > VHOST_USER_REM_MEM_REG = 38, > > + VHOST_USER_SET_STATUS = 39, > > + VHOST_USER_GET_STATUS = 40, > > + VHOST_USER_MEM_DATA = 41, > > VHOST_USER_MAX > > } VhostUserRequest; > > > > @@ -122,6 +125,7 @@ typedef enum VhostUserSlaveRequest { > > VHOST_USER_SLAVE_FS_MAP = 6, > > VHOST_USER_SLAVE_FS_UNMAP = 7, > > VHOST_USER_SLAVE_FS_IO = 8, > > + VHOST_USER_SLAVE_MEM_ACCESS = 9, > > VHOST_USER_SLAVE_MAX > > } VhostUserSlaveRequest; > > > > @@ -220,6 +224,24 @@ typedef struct VhostUserInflight { > > uint16_t queue_size; > > } VhostUserInflight; > > > > +/* For the flags field of VhostUserMemAccess and VhostUserMemReply */ > > +#define VHOST_USER_MEM_FLAG_TO_MASTER (1u << 0) > > +#define VHOST_USER_MEM_FLAG_FROM_MASTER (1u << 1) > > +typedef struct VhostUserMemAccess { > > + uint32_t id; /* Included in the reply */ > > + uint32_t flags; > > Is VHOST_USER_MEM_FLAG_TO_MASTER | VHOST_USER_MEM_FLAG_FROM_MASTER > valid? No; I've changed the docs to state: One (and only one) of the ``VHOST_USER_MEM_FLAG_TO_MASTER`` and ``VHOST_USER_MEM_FLAG_FROM_MASTER`` flags must be set in the ``flags`` field. > > + uint64_t addr; /* In the bus address of the device */ > > Please check the spec for preferred terminology. "bus address" isn't > used in the spec, so there's probably another term for it. I'm not seeing anything useful in the virtio spec; it mostly talks about guest physical addresses; it does say 'bus addresses' in the definition of 'VIRTIO_F_ACCESS_PLATFORM' . > > + uint64_t len; /* In bytes */ > > +} VhostUserMemAccess; > > + > > +typedef struct VhostUserMemReply { > > + uint32_t id; /* From the request */ > > + uint32_t flags; > > Are any flags defined? Currently they're a copy of the TO/FROM _MASTER flags that were in the request; which are useful to the device to make it easy to know whether there's data following on the stream. > > + uint32_t err; /* 0 on success */ > > + uint32_t align; > > Is this a reserved padding field? "align" is confusing because it could > refer to some kind of memory alignment value. "reserved" or "padding" is > clearer. Changed to 'padding' Dave > > + uint64_t len; > > +} VhostUserMemReply; > > + > > #if defined(_WIN32) && (defined(__x86_64__) || defined(__i386__)) > > # define VU_PACKED __attribute__((gcc_struct, packed)) > > #else > > @@ -248,6 +270,8 @@ typedef struct VhostUserMsg { > > VhostUserVringArea area; > > VhostUserInflight inflight; > > VhostUserFSSlaveMsgMax fs_max; > > + VhostUserMemAccess memaccess; > > + VhostUserMemReply memreply; > > } payload; > > > > int fds[VHOST_MEMORY_BASELINE_NREGIONS]; > > -- > > 2.31.1 > > > > -- > > Dr. David Alan Gilbert / dgilbert@redhat.com / Manchester, UK > > -- Dr. David Alan Gilbert / dgilbert@redhat.com / Manchester, UK From mboxrd@z Thu Jan 1 00:00:00 1970 Date: Wed, 16 Jun 2021 13:36:10 +0100 From: "Dr. David Alan Gilbert" Message-ID: References: <20210428110100.27757-1-dgilbert@redhat.com> <20210428110100.27757-27-dgilbert@redhat.com> <20210506160223.GA277745@redhat.com> <20210510152324.GB150402@horse> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: Subject: Re: [Virtio-fs] [PATCH v3 26/26] virtiofsd: Ask qemu to drop CAP_FSETID if client asked for it List-Id: Development discussions about virtio-fs List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , To: Stefan Hajnoczi Cc: virtio-fs@redhat.com, qemu-devel@nongnu.org, Vivek Goyal * Stefan Hajnoczi (stefanha@redhat.com) wrote: > On Thu, Jun 10, 2021 at 04:29:42PM +0100, Dr. David Alan Gilbert wrote: > > * Dr. David Alan Gilbert (dgilbert@redhat.com) wrote: > > > * Stefan Hajnoczi (stefanha@redhat.com) wrote: > > > > > > > > > > Instead I was thinking about VHOST_USER_DMA_READ/WRITE messages > > > > containing the address (a device IOVA, it could just be a guest physical > > > > memory address in most cases) and the length. The WRITE message would > > > > also contain the data that the vhost-user device wishes to write. The > > > > READ message reply would contain the data that the device read from > > > > QEMU. > > > > > > > > QEMU would implement this using QEMU's address_space_read/write() APIs. > > > > > > > > So basically just a new vhost-user protocol message to do a memcpy(), > > > > but with guest addresses and vIOMMU support :). > > > > > > This doesn't actually feel that hard - ignoring vIOMMU for a minute > > > which I know very little about - I'd have to think where the data > > > actually flows, probably the slave fd. > > > > > > > The vhost-user device will need to do bounce buffering so using these > > > > new messages is slower than zero-copy I/O to shared guest RAM. > > > > > > I guess the theory is it's only in the weird corner cases anyway. > > The feature is also useful if DMA isolation is desirable (i.e. > security/reliability are more important than performance). Once this new > vhost-user protocol feature is available it will be possible to run > vhost-user devices without shared memory or with limited shared memory > (e.g. just the vring). I don't see it ever being efficient, so that case is going to be pretty limited. > > The direction I'm going is something like the following; > > the idea is that the master will have to handle the requests on a > > separate thread, to avoid any problems with side effects from the memory > > accesses; the slave will then have to parkt he requests somewhere and > > handle them later. > > > > > > From 07aacff77c50c8a2b588b2513f2dfcfb8f5aa9df Mon Sep 17 00:00:00 2001 > > From: "Dr. David Alan Gilbert" > > Date: Thu, 10 Jun 2021 15:34:04 +0100 > > Subject: [PATCH] WIP: vhost-user: DMA type interface > > > > A DMA type interface where the slave can ask for a stream of bytes > > to be read/written to the guests memory by the master. > > The interface is asynchronous, since a request may have side effects > > inside the guest. > > > > Signed-off-by: Dr. David Alan Gilbert > > --- > > docs/interop/vhost-user.rst | 33 +++++++++++++++++++++++ > > hw/virtio/vhost-user.c | 4 +++ > > subprojects/libvhost-user/libvhost-user.h | 24 +++++++++++++++++ > > 3 files changed, 61 insertions(+) > > Use of the word "RAM" in this patch is a little unclear since we need > these new messages precisely when it's not ordinary guest RAM :-). Maybe > referring to the address space is more general. Yeh, I'll try and spot those. > > diff --git a/docs/interop/vhost-user.rst b/docs/interop/vhost-user.rst > > index 9ebd05e2bf..b9b5322147 100644 > > --- a/docs/interop/vhost-user.rst > > +++ b/docs/interop/vhost-user.rst > > @@ -1347,6 +1347,15 @@ Master message types > > query the backend for its device status as defined in the Virtio > > specification. > > > > +``VHOST_USER_MEM_DATA`` > > + :id: 41 > > + :equivalent ioctl: N/A > > + :slave payload: N/A > > + :master payload: ``struct VhostUserMemReply`` > > + > > + This message is an asynchronous response to a ``VHOST_USER_SLAVE_MEM_ACCESS`` > > + message. Where the request was for the master to read data, this > > + message will be followed by the data that was read. > > Please explain why this message is asynchronous. Implementors will need > to understand the gotchas around deadlocks, etc. I've added: Making this a separate asynchronous response message (rather than just a reply to the ``VHOST_USER_SLAVE_MEM_ACCESS``) makes it much easier for the master to deal with any side effects the access may have, and in particular avoid deadlocks they might cause if an access triggers another vhost_user message. > > > > Slave message types > > ------------------- > > @@ -1469,6 +1478,30 @@ Slave message types > > The ``VHOST_USER_FS_FLAG_MAP_W`` flag must be set in the ``flags`` field to > > write to the file from RAM. > > > > +``VHOST_USER_SLAVE_MEM_ACCESS`` > > + :id: 9 > > + :equivalent ioctl: N/A > > + :slave payload: ``struct VhostUserMemAccess`` > > + :master payload: N/A > > + > > + Requests that the master perform a range of memory accesses on behalf > > + of the slave that the slave can't perform itself. > > + > > + The ``VHOST_USER_MEM_FLAG_TO_MASTER`` flag must be set in the ``flags`` > > + field for the slave to write data into the RAM of the master. In this > > + case the data to write follows the ``VhostUserMemAccess`` on the fd. > > + The ``VHOST_USER_MEM_FLAG_FROM_MASTER`` flag must be set in the ``flags`` > > + field for the slave to read data from the RAM of the master. > > + > > + When the master has completed the access it replies on the main fd with > > + a ``VHOST_USER_MEM_DATA`` message. > > + > > + The master is allowed to complete part of the request and reply stating > > + the amount completed, leaving it to the slave to resend further components. > > + This may happen to limit memory allocations in the master or to simplify > > + the implementation. > > + > > + > > .. _reply_ack: > > > > VHOST_USER_PROTOCOL_F_REPLY_ACK > > diff --git a/hw/virtio/vhost-user.c b/hw/virtio/vhost-user.c > > index 39a0e55cca..a3fefc4c1d 100644 > > --- a/hw/virtio/vhost-user.c > > +++ b/hw/virtio/vhost-user.c > > @@ -126,6 +126,9 @@ typedef enum VhostUserRequest { > > VHOST_USER_GET_MAX_MEM_SLOTS = 36, > > VHOST_USER_ADD_MEM_REG = 37, > > VHOST_USER_REM_MEM_REG = 38, > > + VHOST_USER_SET_STATUS = 39, > > + VHOST_USER_GET_STATUS = 40, > > + VHOST_USER_MEM_DATA = 41, > > VHOST_USER_MAX > > } VhostUserRequest; > > > > @@ -139,6 +142,7 @@ typedef enum VhostUserSlaveRequest { > > VHOST_USER_SLAVE_FS_MAP = 6, > > VHOST_USER_SLAVE_FS_UNMAP = 7, > > VHOST_USER_SLAVE_FS_IO = 8, > > + VHOST_USER_SLAVE_MEM_ACCESS = 9, > > VHOST_USER_SLAVE_MAX > > } VhostUserSlaveRequest; > > > > diff --git a/subprojects/libvhost-user/libvhost-user.h b/subprojects/libvhost-user/libvhost-user.h > > index eee611a2f6..b5444f4f6f 100644 > > --- a/subprojects/libvhost-user/libvhost-user.h > > +++ b/subprojects/libvhost-user/libvhost-user.h > > @@ -109,6 +109,9 @@ typedef enum VhostUserRequest { > > VHOST_USER_GET_MAX_MEM_SLOTS = 36, > > VHOST_USER_ADD_MEM_REG = 37, > > VHOST_USER_REM_MEM_REG = 38, > > + VHOST_USER_SET_STATUS = 39, > > + VHOST_USER_GET_STATUS = 40, > > + VHOST_USER_MEM_DATA = 41, > > VHOST_USER_MAX > > } VhostUserRequest; > > > > @@ -122,6 +125,7 @@ typedef enum VhostUserSlaveRequest { > > VHOST_USER_SLAVE_FS_MAP = 6, > > VHOST_USER_SLAVE_FS_UNMAP = 7, > > VHOST_USER_SLAVE_FS_IO = 8, > > + VHOST_USER_SLAVE_MEM_ACCESS = 9, > > VHOST_USER_SLAVE_MAX > > } VhostUserSlaveRequest; > > > > @@ -220,6 +224,24 @@ typedef struct VhostUserInflight { > > uint16_t queue_size; > > } VhostUserInflight; > > > > +/* For the flags field of VhostUserMemAccess and VhostUserMemReply */ > > +#define VHOST_USER_MEM_FLAG_TO_MASTER (1u << 0) > > +#define VHOST_USER_MEM_FLAG_FROM_MASTER (1u << 1) > > +typedef struct VhostUserMemAccess { > > + uint32_t id; /* Included in the reply */ > > + uint32_t flags; > > Is VHOST_USER_MEM_FLAG_TO_MASTER | VHOST_USER_MEM_FLAG_FROM_MASTER > valid? No; I've changed the docs to state: One (and only one) of the ``VHOST_USER_MEM_FLAG_TO_MASTER`` and ``VHOST_USER_MEM_FLAG_FROM_MASTER`` flags must be set in the ``flags`` field. > > + uint64_t addr; /* In the bus address of the device */ > > Please check the spec for preferred terminology. "bus address" isn't > used in the spec, so there's probably another term for it. I'm not seeing anything useful in the virtio spec; it mostly talks about guest physical addresses; it does say 'bus addresses' in the definition of 'VIRTIO_F_ACCESS_PLATFORM' . > > + uint64_t len; /* In bytes */ > > +} VhostUserMemAccess; > > + > > +typedef struct VhostUserMemReply { > > + uint32_t id; /* From the request */ > > + uint32_t flags; > > Are any flags defined? Currently they're a copy of the TO/FROM _MASTER flags that were in the request; which are useful to the device to make it easy to know whether there's data following on the stream. > > + uint32_t err; /* 0 on success */ > > + uint32_t align; > > Is this a reserved padding field? "align" is confusing because it could > refer to some kind of memory alignment value. "reserved" or "padding" is > clearer. Changed to 'padding' Dave > > + uint64_t len; > > +} VhostUserMemReply; > > + > > #if defined(_WIN32) && (defined(__x86_64__) || defined(__i386__)) > > # define VU_PACKED __attribute__((gcc_struct, packed)) > > #else > > @@ -248,6 +270,8 @@ typedef struct VhostUserMsg { > > VhostUserVringArea area; > > VhostUserInflight inflight; > > VhostUserFSSlaveMsgMax fs_max; > > + VhostUserMemAccess memaccess; > > + VhostUserMemReply memreply; > > } payload; > > > > int fds[VHOST_MEMORY_BASELINE_NREGIONS]; > > -- > > 2.31.1 > > > > -- > > Dr. David Alan Gilbert / dgilbert@redhat.com / Manchester, UK > > -- Dr. David Alan Gilbert / dgilbert@redhat.com / Manchester, UK