From: "Michael S. Tsirkin" <mst@redhat.com>
To: Raphael Norwitz <raphael.norwitz@nutanix.com>
Cc: Peter Turschmid <peter.turschm@nutanix.com>,
raphael.s.norwitz@gmail.com, qemu-devel@nongnu.org
Subject: Re: [PATCH] Refactor vhost_user_set_mem_table functions
Date: Sun, 29 Mar 2020 09:31:58 -0400 [thread overview]
Message-ID: <20200329093125-mutt-send-email-mst@kernel.org> (raw)
In-Reply-To: <1585132506-13316-1-git-send-email-raphael.norwitz@nutanix.com>
On Wed, Mar 25, 2020 at 06:35:06AM -0400, Raphael Norwitz wrote:
> vhost_user_set_mem_table() and vhost_user_set_mem_table_postcopy() have
> gotten convoluted, and have some identical code.
>
> This change moves the logic populating the VhostUserMemory struct and
> fds array from vhost_user_set_mem_table() and
> vhost_user_set_mem_table_postcopy() to a new function,
> vhost_user_fill_set_mem_table_msg().
>
> No functionality is impacted.
>
> Signed-off-by: Raphael Norwitz <raphael.norwitz@nutanix.com>
> Signed-off-by: Peter Turschmid <peter.turschm@nutanix.com>
Thanks!
I'd like to queue it for merge after the release. If possible
please ping me after the release to help make sure it didn't get
dropped.
> ---
> hw/virtio/vhost-user.c | 143 +++++++++++++++++++++++--------------------------
> 1 file changed, 67 insertions(+), 76 deletions(-)
>
> diff --git a/hw/virtio/vhost-user.c b/hw/virtio/vhost-user.c
> index 08e7e63..ec21e8f 100644
> --- a/hw/virtio/vhost-user.c
> +++ b/hw/virtio/vhost-user.c
> @@ -407,18 +407,79 @@ static int vhost_user_set_log_base(struct vhost_dev *dev, uint64_t base,
> return 0;
> }
>
> +static int vhost_user_fill_set_mem_table_msg(struct vhost_user *u,
> + struct vhost_dev *dev,
> + VhostUserMsg *msg,
> + int *fds, size_t *fd_num,
> + bool track_ramblocks)
> +{
> + int i, fd;
> + ram_addr_t offset;
> + MemoryRegion *mr;
> + struct vhost_memory_region *reg;
> +
> + msg->hdr.request = VHOST_USER_SET_MEM_TABLE;
> +
> + for (i = 0; i < dev->mem->nregions; ++i) {
> + reg = dev->mem->regions + i;
> +
> + assert((uintptr_t)reg->userspace_addr == reg->userspace_addr);
> + mr = memory_region_from_host((void *)(uintptr_t)reg->userspace_addr,
> + &offset);
> + fd = memory_region_get_fd(mr);
> + if (fd > 0) {
> + if (track_ramblocks) {
> + assert(*fd_num < VHOST_MEMORY_MAX_NREGIONS);
> + trace_vhost_user_set_mem_table_withfd(*fd_num, mr->name,
> + reg->memory_size,
> + reg->guest_phys_addr,
> + reg->userspace_addr,
> + offset);
> + u->region_rb_offset[i] = offset;
> + u->region_rb[i] = mr->ram_block;
> + } else if (*fd_num == VHOST_MEMORY_MAX_NREGIONS) {
> + error_report("Failed preparing vhost-user memory table msg");
> + return -1;
> + }
> + msg->payload.memory.regions[*fd_num].userspace_addr =
> + reg->userspace_addr;
> + msg->payload.memory.regions[*fd_num].memory_size =
> + reg->memory_size;
> + msg->payload.memory.regions[*fd_num].guest_phys_addr =
> + reg->guest_phys_addr;
> + msg->payload.memory.regions[*fd_num].mmap_offset = offset;
> + fds[(*fd_num)++] = fd;
> + } else if (track_ramblocks) {
> + u->region_rb_offset[i] = 0;
> + u->region_rb[i] = NULL;
> + }
> + }
> +
> + msg->payload.memory.nregions = *fd_num;
> +
> + if (!*fd_num) {
> + error_report("Failed initializing vhost-user memory map, "
> + "consider using -object memory-backend-file share=on");
> + return -1;
> + }
> +
> + msg->hdr.size = sizeof(msg->payload.memory.nregions);
> + msg->hdr.size += sizeof(msg->payload.memory.padding);
> + msg->hdr.size += *fd_num * sizeof(VhostUserMemoryRegion);
> +
> + return 1;
> +}
> +
> static int vhost_user_set_mem_table_postcopy(struct vhost_dev *dev,
> struct vhost_memory *mem)
> {
> struct vhost_user *u = dev->opaque;
> int fds[VHOST_MEMORY_MAX_NREGIONS];
> - int i, fd;
> size_t fd_num = 0;
> VhostUserMsg msg_reply;
> int region_i, msg_i;
>
> VhostUserMsg msg = {
> - .hdr.request = VHOST_USER_SET_MEM_TABLE,
> .hdr.flags = VHOST_USER_VERSION,
> };
>
> @@ -433,48 +494,11 @@ static int vhost_user_set_mem_table_postcopy(struct vhost_dev *dev,
> u->region_rb_len = dev->mem->nregions;
> }
>
> - for (i = 0; i < dev->mem->nregions; ++i) {
> - struct vhost_memory_region *reg = dev->mem->regions + i;
> - ram_addr_t offset;
> - MemoryRegion *mr;
> -
> - assert((uintptr_t)reg->userspace_addr == reg->userspace_addr);
> - mr = memory_region_from_host((void *)(uintptr_t)reg->userspace_addr,
> - &offset);
> - fd = memory_region_get_fd(mr);
> - if (fd > 0) {
> - assert(fd_num < VHOST_MEMORY_MAX_NREGIONS);
> - trace_vhost_user_set_mem_table_withfd(fd_num, mr->name,
> - reg->memory_size,
> - reg->guest_phys_addr,
> - reg->userspace_addr, offset);
> - u->region_rb_offset[i] = offset;
> - u->region_rb[i] = mr->ram_block;
> - msg.payload.memory.regions[fd_num].userspace_addr =
> - reg->userspace_addr;
> - msg.payload.memory.regions[fd_num].memory_size = reg->memory_size;
> - msg.payload.memory.regions[fd_num].guest_phys_addr =
> - reg->guest_phys_addr;
> - msg.payload.memory.regions[fd_num].mmap_offset = offset;
> - fds[fd_num++] = fd;
> - } else {
> - u->region_rb_offset[i] = 0;
> - u->region_rb[i] = NULL;
> - }
> - }
> -
> - msg.payload.memory.nregions = fd_num;
> -
> - if (!fd_num) {
> - error_report("Failed initializing vhost-user memory map, "
> - "consider using -object memory-backend-file share=on");
> + if (vhost_user_fill_set_mem_table_msg(u, dev, &msg, fds, &fd_num,
> + true) < 0) {
> return -1;
> }
>
> - msg.hdr.size = sizeof(msg.payload.memory.nregions);
> - msg.hdr.size += sizeof(msg.payload.memory.padding);
> - msg.hdr.size += fd_num * sizeof(VhostUserMemoryRegion);
> -
> if (vhost_user_write(dev, &msg, fds, fd_num) < 0) {
> return -1;
> }
> @@ -545,7 +569,6 @@ static int vhost_user_set_mem_table(struct vhost_dev *dev,
> {
> struct vhost_user *u = dev->opaque;
> int fds[VHOST_MEMORY_MAX_NREGIONS];
> - int i, fd;
> size_t fd_num = 0;
> bool do_postcopy = u->postcopy_listen && u->postcopy_fd.handler;
> bool reply_supported = virtio_has_feature(dev->protocol_features,
> @@ -559,7 +582,6 @@ static int vhost_user_set_mem_table(struct vhost_dev *dev,
> }
>
> VhostUserMsg msg = {
> - .hdr.request = VHOST_USER_SET_MEM_TABLE,
> .hdr.flags = VHOST_USER_VERSION,
> };
>
> @@ -567,42 +589,11 @@ static int vhost_user_set_mem_table(struct vhost_dev *dev,
> msg.hdr.flags |= VHOST_USER_NEED_REPLY_MASK;
> }
>
> - for (i = 0; i < dev->mem->nregions; ++i) {
> - struct vhost_memory_region *reg = dev->mem->regions + i;
> - ram_addr_t offset;
> - MemoryRegion *mr;
> -
> - assert((uintptr_t)reg->userspace_addr == reg->userspace_addr);
> - mr = memory_region_from_host((void *)(uintptr_t)reg->userspace_addr,
> - &offset);
> - fd = memory_region_get_fd(mr);
> - if (fd > 0) {
> - if (fd_num == VHOST_MEMORY_MAX_NREGIONS) {
> - error_report("Failed preparing vhost-user memory table msg");
> - return -1;
> - }
> - msg.payload.memory.regions[fd_num].userspace_addr =
> - reg->userspace_addr;
> - msg.payload.memory.regions[fd_num].memory_size = reg->memory_size;
> - msg.payload.memory.regions[fd_num].guest_phys_addr =
> - reg->guest_phys_addr;
> - msg.payload.memory.regions[fd_num].mmap_offset = offset;
> - fds[fd_num++] = fd;
> - }
> - }
> -
> - msg.payload.memory.nregions = fd_num;
> -
> - if (!fd_num) {
> - error_report("Failed initializing vhost-user memory map, "
> - "consider using -object memory-backend-file share=on");
> + if (vhost_user_fill_set_mem_table_msg(u, dev, &msg, fds, &fd_num,
> + false) < 0) {
> return -1;
> }
>
> - msg.hdr.size = sizeof(msg.payload.memory.nregions);
> - msg.hdr.size += sizeof(msg.payload.memory.padding);
> - msg.hdr.size += fd_num * sizeof(VhostUserMemoryRegion);
> -
> if (vhost_user_write(dev, &msg, fds, fd_num) < 0) {
> return -1;
> }
> --
> 1.8.3.1
prev parent reply other threads:[~2020-03-29 13:33 UTC|newest]
Thread overview: 3+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-03-25 10:35 [PATCH] Refactor vhost_user_set_mem_table functions Raphael Norwitz
2020-03-26 11:26 ` Marc-André Lureau
2020-03-29 13:31 ` Michael S. Tsirkin [this message]
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20200329093125-mutt-send-email-mst@kernel.org \
--to=mst@redhat.com \
--cc=peter.turschm@nutanix.com \
--cc=qemu-devel@nongnu.org \
--cc=raphael.norwitz@nutanix.com \
--cc=raphael.s.norwitz@gmail.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).