From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-2.3 required=3.0 tests=HEADER_FROM_DIFFERENT_DOMAINS, MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS,USER_AGENT_SANE_1 autolearn=no autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 4A81CC433FF for ; Thu, 1 Aug 2019 13:36:27 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [209.132.180.67]) by mail.kernel.org (Postfix) with ESMTP id 21C5920679 for ; Thu, 1 Aug 2019 13:36:27 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726583AbfHANgX (ORCPT ); Thu, 1 Aug 2019 09:36:23 -0400 Received: from mail-wr1-f68.google.com ([209.85.221.68]:39131 "EHLO mail-wr1-f68.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1728578AbfHANgX (ORCPT ); Thu, 1 Aug 2019 09:36:23 -0400 Received: by mail-wr1-f68.google.com with SMTP id x4so20476236wrt.6 for ; Thu, 01 Aug 2019 06:36:20 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:date:from:to:cc:subject:message-id:references :mime-version:content-disposition:in-reply-to:user-agent; bh=NJ2gMngpf0kg912re0Xn0ToSKHMrEUKjnCdafaP8Gg0=; b=QOsHBvrMv5ek3nWe2FblZ3wqdZ4yB5/At9wQsdUMvsONN5u5MMXAYYN0aoyeuMgK7P N/pEtxBfdFGbvvE2Y6mFonGy3AC7dfrOaaRHLdZ+DH8QWWFHcsLKluIlOZOsxG4QZVk7 5FQoFU2cAcVz67ZA38O25y6ZhNrjESa3dvrw4Ve26iNtsTywjuFNrt1Zut8xkv2FbsgG IMfZB9vFVJcSUsw7asjkY2QOiiCmGezsfZxYzAtC2H/7KEQOh0TjlkYSazWU83cS7bqz yCdluW0FhR5+0j4/V7wSLCZ8wroyn5M2+GBcII2+w7rTR0fQ7XhvEgVM34KfUcsXqxuL exOw== X-Gm-Message-State: APjAAAXXzxP2wLtIzR6WpbD9QA2Af2XCn7OSkqoD23PRCTMscm+gMG9R 0720lNOrjxbuLoB5tLc6UW7Gpg== X-Google-Smtp-Source: APXvYqzkDCVti1BqaiCg8+9sPv45Y+e5KZ0R5OR8H+9HynT6a2FYU3NisphROmhm6qQjJuqB7DyA/A== X-Received: by 2002:adf:f104:: with SMTP id r4mr50858021wro.140.1564666580137; Thu, 01 Aug 2019 06:36:20 -0700 (PDT) Received: from steredhat (host122-201-dynamic.13-79-r.retail.telecomitalia.it. [79.13.201.122]) by smtp.gmail.com with ESMTPSA id p14sm59528390wrx.17.2019.08.01.06.36.18 (version=TLS1_3 cipher=AEAD-AES256-GCM-SHA384 bits=256/256); Thu, 01 Aug 2019 06:36:19 -0700 (PDT) Date: Thu, 1 Aug 2019 15:36:16 +0200 From: Stefano Garzarella To: "Michael S. Tsirkin" Cc: netdev@vger.kernel.org, linux-kernel@vger.kernel.org, Stefan Hajnoczi , "David S. Miller" , virtualization@lists.linux-foundation.org, Jason Wang , kvm@vger.kernel.org Subject: Re: [PATCH v4 1/5] vsock/virtio: limit the memory used per-socket Message-ID: <20190801133616.sik5drn6ecesukbb@steredhat> References: <20190729095956-mutt-send-email-mst@kernel.org> <20190729153656.zk4q4rob5oi6iq7l@steredhat> <20190729114302-mutt-send-email-mst@kernel.org> <20190729161903.yhaj5rfcvleexkhc@steredhat> <20190729165056.r32uzj6om3o6vfvp@steredhat> <20190729143622-mutt-send-email-mst@kernel.org> <20190730093539.dcksure3vrykir3g@steredhat> <20190730163807-mutt-send-email-mst@kernel.org> <20190801104754.lb3ju5xjfmnxioii@steredhat> <20190801091106-mutt-send-email-mst@kernel.org> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20190801091106-mutt-send-email-mst@kernel.org> User-Agent: NeoMutt/20180716 Sender: netdev-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: netdev@vger.kernel.org On Thu, Aug 01, 2019 at 09:21:15AM -0400, Michael S. Tsirkin wrote: > On Thu, Aug 01, 2019 at 12:47:54PM +0200, Stefano Garzarella wrote: > > On Tue, Jul 30, 2019 at 04:42:25PM -0400, Michael S. Tsirkin wrote: > > > On Tue, Jul 30, 2019 at 11:35:39AM +0200, Stefano Garzarella wrote: > > > > (...) > > > > > > > > > > The problem here is the compatibility. Before this series virtio-vsock > > > > and vhost-vsock modules had the RX buffer size hard-coded > > > > (VIRTIO_VSOCK_DEFAULT_RX_BUF_SIZE = 4K). So, if we send a buffer smaller > > > > of 4K, there might be issues. > > > > > > Shouldn't be if they are following the spec. If not let's fix > > > the broken parts. > > > > > > > > > > > Maybe it is the time to add add 'features' to virtio-vsock device. > > > > > > > > Thanks, > > > > Stefano > > > > > > Why would a remote care about buffer sizes? > > > > > > Let's first see what the issues are. If they exist > > > we can either fix the bugs, or code the bug as a feature in spec. > > > > > > > The vhost_transport '.stream_enqueue' callback > > [virtio_transport_stream_enqueue()] calls the virtio_transport_send_pkt_info(), > > passing the user message. This function allocates a new packet, copying > > the user message, but (before this series) it limits the packet size to > > the VIRTIO_VSOCK_DEFAULT_RX_BUF_SIZE (4K): > > > > static int virtio_transport_send_pkt_info(struct vsock_sock *vsk, > > struct virtio_vsock_pkt_info *info) > > { > > ... > > /* we can send less than pkt_len bytes */ > > if (pkt_len > VIRTIO_VSOCK_DEFAULT_RX_BUF_SIZE) > > pkt_len = VIRTIO_VSOCK_DEFAULT_RX_BUF_SIZE; > > > > /* virtio_transport_get_credit might return less than pkt_len credit */ > > pkt_len = virtio_transport_get_credit(vvs, pkt_len); > > > > /* Do not send zero length OP_RW pkt */ > > if (pkt_len == 0 && info->op == VIRTIO_VSOCK_OP_RW) > > return pkt_len; > > ... > > } > > > > then it queues the packet for the TX worker calling .send_pkt() > > [vhost_transport_send_pkt() in the vhost_transport case] > > > > The main function executed by the TX worker is > > vhost_transport_do_send_pkt() that picks up a buffer from the virtqueue > > and it tries to copy the packet (up to 4K) on it. If the buffer > > allocated from the guest will be smaller then 4K, I think here it will > > be discarded with an error: > > I'm adding more lines to explain better. > > static void > > vhost_transport_do_send_pkt(struct vhost_vsock *vsock, > > struct vhost_virtqueue *vq) > > { ... head = vhost_get_vq_desc(vq, vq->iov, ARRAY_SIZE(vq->iov), &out, &in, NULL, NULL); ... len = iov_length(&vq->iov[out], in); iov_iter_init(&iov_iter, READ, &vq->iov[out], in, len); nbytes = copy_to_iter(&pkt->hdr, sizeof(pkt->hdr), &iov_iter); if (nbytes != sizeof(pkt->hdr)) { virtio_transport_free_pkt(pkt); vq_err(vq, "Faulted on copying pkt hdr\n"); break; } > > ... > > nbytes = copy_to_iter(pkt->buf, pkt->len, &iov_iter); > > isn't pck len the actual length though? > It is the length of the packet that we are copying in the guest RX buffers pointed by the iov_iter. The guest allocates an iovec with 2 buffers, one for the header and one for the payload (4KB). > > if (nbytes != pkt->len) { > > virtio_transport_free_pkt(pkt); > > vq_err(vq, "Faulted on copying pkt buf\n"); > > break; > > } > > ... > > } > > > > > > This series changes this behavior since now we will split the packet in > > vhost_transport_do_send_pkt() depending on the buffer found in the > > virtqueue. > > > > We didn't change the buffer size in this series, so we still backward > > compatible, but if we will use buffers smaller than 4K, we should > > encounter the error described above. > > > > How do you suggest we proceed if we want to change the buffer size? > > Maybe adding a feature to "support any buffer size"? > > > > Thanks, > > Stefano > > --