From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-5.3 required=3.0 tests=HEADER_FROM_DIFFERENT_DOMAINS, MAILING_LIST_MULTI,SIGNED_OFF_BY,SPF_HELO_NONE,SPF_PASS,USER_AGENT_SANE_1 autolearn=unavailable autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id A6A06C76193 for ; Mon, 29 Jul 2019 16:19:11 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [209.132.180.67]) by mail.kernel.org (Postfix) with ESMTP id 876E22067D for ; Mon, 29 Jul 2019 16:19:11 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1728667AbfG2QTK (ORCPT ); Mon, 29 Jul 2019 12:19:10 -0400 Received: from mail-wr1-f68.google.com ([209.85.221.68]:42690 "EHLO mail-wr1-f68.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1727646AbfG2QTJ (ORCPT ); Mon, 29 Jul 2019 12:19:09 -0400 Received: by mail-wr1-f68.google.com with SMTP id x1so12605109wrr.9 for ; Mon, 29 Jul 2019 09:19:07 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:date:from:to:cc:subject:message-id:references :mime-version:content-disposition:in-reply-to:user-agent; bh=ZKDqUs0ayhzO8fC3NvwV+tCFuuA/FPgD3ePdCx1GJqM=; b=Kj/30/bKk7ieRflxh/wAnNuTHFcsyVHcuZDfJWaJT+c70UnUHHdSqR3mCrhregdmmb +f1qVak1A9kubqcGowCtjw82tczxjIJAm3QL21M2wvkTu0AOs4XQUpE2pEQn56fdBZOY nzySs0KPbOVo2TCJ0dKFsrh8fcpZ3R3QPNRvO6jdVvx0lRDOeE9X/viwMNuHpf/Gk1bi w/4IgVXtnoSK62kdmbKQVsiJgLBIr3qSKdRbO4Yk8D9Ge7NOtpYxspVJ7xaVKTLgbq8D wFGE4WTGOe6P9jRQUUn6U8Wyq7PIUKjBNQOAHBhImxl1ZSJ7qxTCweOyI5yauzR36zk5 9LIg== X-Gm-Message-State: APjAAAXubjQRer5RrvWc8GzGXMQvVnmWKFJFo+VVCBe83Lcv5CUCIRM3 tArBx4nW2bb+pECHe2Xx2v/JWw== X-Google-Smtp-Source: APXvYqx0sc12Pv/jNOPaKvwFRyWt3K0PAbYM2mQ0eNUoR5RWLAC9VV53Iw6ZEaEpC4k4BBzOXgmD7Q== X-Received: by 2002:a5d:4b8b:: with SMTP id b11mr40501951wrt.294.1564417146999; Mon, 29 Jul 2019 09:19:06 -0700 (PDT) Received: from steredhat (host122-201-dynamic.13-79-r.retail.telecomitalia.it. [79.13.201.122]) by smtp.gmail.com with ESMTPSA id c11sm104255286wrq.45.2019.07.29.09.19.05 (version=TLS1_3 cipher=AEAD-AES256-GCM-SHA384 bits=256/256); Mon, 29 Jul 2019 09:19:06 -0700 (PDT) Date: Mon, 29 Jul 2019 18:19:03 +0200 From: Stefano Garzarella To: "Michael S. Tsirkin" Cc: netdev@vger.kernel.org, linux-kernel@vger.kernel.org, Stefan Hajnoczi , "David S. Miller" , virtualization@lists.linux-foundation.org, Jason Wang , kvm@vger.kernel.org Subject: Re: [PATCH v4 1/5] vsock/virtio: limit the memory used per-socket Message-ID: <20190729161903.yhaj5rfcvleexkhc@steredhat> References: <20190717113030.163499-1-sgarzare@redhat.com> <20190717113030.163499-2-sgarzare@redhat.com> <20190729095956-mutt-send-email-mst@kernel.org> <20190729153656.zk4q4rob5oi6iq7l@steredhat> <20190729114302-mutt-send-email-mst@kernel.org> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20190729114302-mutt-send-email-mst@kernel.org> User-Agent: NeoMutt/20180716 Sender: netdev-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: netdev@vger.kernel.org On Mon, Jul 29, 2019 at 11:49:02AM -0400, Michael S. Tsirkin wrote: > On Mon, Jul 29, 2019 at 05:36:56PM +0200, Stefano Garzarella wrote: > > On Mon, Jul 29, 2019 at 10:04:29AM -0400, Michael S. Tsirkin wrote: > > > On Wed, Jul 17, 2019 at 01:30:26PM +0200, Stefano Garzarella wrote: > > > > Since virtio-vsock was introduced, the buffers filled by the host > > > > and pushed to the guest using the vring, are directly queued in > > > > a per-socket list. These buffers are preallocated by the guest > > > > with a fixed size (4 KB). > > > > > > > > The maximum amount of memory used by each socket should be > > > > controlled by the credit mechanism. > > > > The default credit available per-socket is 256 KB, but if we use > > > > only 1 byte per packet, the guest can queue up to 262144 of 4 KB > > > > buffers, using up to 1 GB of memory per-socket. In addition, the > > > > guest will continue to fill the vring with new 4 KB free buffers > > > > to avoid starvation of other sockets. > > > > > > > > This patch mitigates this issue copying the payload of small > > > > packets (< 128 bytes) into the buffer of last packet queued, in > > > > order to avoid wasting memory. > > > > > > > > Reviewed-by: Stefan Hajnoczi > > > > Signed-off-by: Stefano Garzarella > > > > > > This is good enough for net-next, but for net I think we > > > should figure out how to address the issue completely. > > > Can we make the accounting precise? What happens to > > > performance if we do? > > > > > > > In order to do more precise accounting maybe we can use the buffer size, > > instead of payload size when we update the credit available. > > In this way, the credit available for each socket will reflect the memory > > actually used. > > > > I should check better, because I'm not sure what happen if the peer sees > > 1KB of space available, then it sends 1KB of payload (using a 4KB > > buffer). > > > > The other option is to copy each packet in a new buffer like I did in > > the v2 [2], but this forces us to make a copy for each packet that does > > not fill the entire buffer, perhaps too expensive. > > > > [2] https://patchwork.kernel.org/patch/10938741/ > > > > > > Thanks, > > Stefano > > Interesting. You are right, and at some level the protocol forces copies. > > We could try to detect that the actual memory is getting close to > admin limits and force copies on queued packets after the fact. > Is that practical? Yes, I think it is doable! We can decrease the credit available with the buffer size queued, and when the buffer size of packet to queue is bigger than the credit available, we can copy it. > > And yes we can extend the credit accounting to include buffer size. > That's a protocol change but maybe it makes sense. Since we send to the other peer the credit available, maybe this change can be backwards compatible (I'll check better this). Thanks, Stefano