From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-7.5 required=3.0 tests=BAYES_00,DKIMWL_WL_HIGH, DKIM_SIGNED,DKIM_VALID,DKIM_VALID_AU,HEADER_FROM_DIFFERENT_DOMAINS, MAILING_LIST_MULTI,NICE_REPLY_A,SPF_HELO_NONE,SPF_PASS,USER_AGENT_SANE_1 autolearn=no autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id D4200C43381 for ; Thu, 14 Jan 2021 03:41:39 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id B64EB23770 for ; Thu, 14 Jan 2021 03:41:39 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1727861AbhANDlH (ORCPT ); Wed, 13 Jan 2021 22:41:07 -0500 Received: from us-smtp-delivery-124.mimecast.com ([63.128.21.124]:45356 "EHLO us-smtp-delivery-124.mimecast.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1727768AbhANDlD (ORCPT ); Wed, 13 Jan 2021 22:41:03 -0500 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1610595576; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=qtF1idSqCy5ua8Az2XtZ/1G+SswhJSq5w0eZMMNKzU8=; b=jWecqsOadZPzEyOtTJA7pjDYatlvo1kIZCJUd771oQ6mMbGkpg421V0SfPNe1k31zCtXwR ipRfUVXpubPQNpfGpuHCQDSNRI1loJXTInkQlxh2ozV6+rr4N4AZVR9mc8Su9Dwl3wTe50 hrhcacw/VX1/E4uCW+5+8czppvO+foM= Received: from mimecast-mx01.redhat.com (mimecast-mx01.redhat.com [209.132.183.4]) (Using TLS) by relay.mimecast.com with ESMTP id us-mta-542--aJiRCZ_NrCYrBKuyUt_sg-1; Wed, 13 Jan 2021 22:39:34 -0500 X-MC-Unique: -aJiRCZ_NrCYrBKuyUt_sg-1 Received: from smtp.corp.redhat.com (int-mx02.intmail.prod.int.phx2.redhat.com [10.5.11.12]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mimecast-mx01.redhat.com (Postfix) with ESMTPS id 638DB15720; Thu, 14 Jan 2021 03:39:30 +0000 (UTC) Received: from [10.72.12.100] (ovpn-12-100.pek2.redhat.com [10.72.12.100]) by smtp.corp.redhat.com (Postfix) with ESMTP id 1C48760C0F; Thu, 14 Jan 2021 03:39:04 +0000 (UTC) Subject: Re: [RFC PATCH 0/7] Support for virtio-net hash reporting To: Willem de Bruijn Cc: Yuri Benditovich , "David S. Miller" , Jakub Kicinski , "Michael S . Tsirkin" , Alexei Starovoitov , Daniel Borkmann , Andrii Nakryiko , Martin KaFai Lau , Song Liu , Yonghong Song , John Fastabend , KP Singh , rdunlap@infradead.org, "Gustavo A . R . Silva" , Herbert Xu , Steffen Klassert , Pablo Neira Ayuso , decui@microsoft.com, cai@lca.pw, Jakub Sitnicki , Marco Elver , Paolo Abeni , Network Development , linux-kernel , kvm@vger.kernel.org, virtualization@lists.linux-foundation.org, bpf , Yan Vugenfirer References: <20210112194143.1494-1-yuri.benditovich@daynix.com> <78bbc518-4b73-4629-68fb-2713250f8967@redhat.com> From: Jason Wang Message-ID: <8ea218a8-a068-1ed9-929d-67ad30111c3c@redhat.com> Date: Thu, 14 Jan 2021 11:38:48 +0800 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:68.0) Gecko/20100101 Thunderbird/68.10.0 MIME-Version: 1.0 In-Reply-To: Content-Type: text/plain; charset=utf-8; format=flowed Content-Transfer-Encoding: 8bit Content-Language: en-US X-Scanned-By: MIMEDefang 2.79 on 10.5.11.12 Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 2021/1/13 下午10:33, Willem de Bruijn wrote: > On Tue, Jan 12, 2021 at 11:11 PM Jason Wang wrote: >> >> On 2021/1/13 上午7:47, Willem de Bruijn wrote: >>> On Tue, Jan 12, 2021 at 3:29 PM Yuri Benditovich >>> wrote: >>>> On Tue, Jan 12, 2021 at 9:49 PM Yuri Benditovich >>>> wrote: >>>>> On Tue, Jan 12, 2021 at 9:41 PM Yuri Benditovich >>>>> wrote: >>>>>> Existing TUN module is able to use provided "steering eBPF" to >>>>>> calculate per-packet hash and derive the destination queue to >>>>>> place the packet to. The eBPF uses mapped configuration data >>>>>> containing a key for hash calculation and indirection table >>>>>> with array of queues' indices. >>>>>> >>>>>> This series of patches adds support for virtio-net hash reporting >>>>>> feature as defined in virtio specification. It extends the TUN module >>>>>> and the "steering eBPF" as follows: >>>>>> >>>>>> Extended steering eBPF calculates the hash value and hash type, keeps >>>>>> hash value in the skb->hash and returns index of destination virtqueue >>>>>> and the type of the hash. TUN module keeps returned hash type in >>>>>> (currently unused) field of the skb. >>>>>> skb->__unused renamed to 'hash_report_type'. >>>>>> >>>>>> When TUN module is called later to allocate and fill the virtio-net >>>>>> header and push it to destination virtqueue it populates the hash >>>>>> and the hash type into virtio-net header. >>>>>> >>>>>> VHOST driver is made aware of respective virtio-net feature that >>>>>> extends the virtio-net header to report the hash value and hash report >>>>>> type. >>>>> Comment from Willem de Bruijn: >>>>> >>>>> Skbuff fields are in short supply. I don't think we need to add one >>>>> just for this narrow path entirely internal to the tun device. >>>>> >>>> We understand that and try to minimize the impact by using an already >>>> existing unused field of skb. >>> Not anymore. It was repurposed as a flags field very recently. >>> >>> This use case is also very narrow in scope. And a very short path from >>> data producer to consumer. So I don't think it needs to claim scarce >>> bits in the skb. >>> >>> tun_ebpf_select_queue stores the field, tun_put_user reads it and >>> converts it to the virtio_net_hdr in the descriptor. >>> >>> tun_ebpf_select_queue is called from .ndo_select_queue. Storing the >>> field in skb->cb is fragile, as in theory some code could overwrite >>> that between field between ndo_select_queue and >>> ndo_start_xmit/tun_net_xmit, from which point it is fully under tun >>> control again. But in practice, I don't believe anything does. >>> >>> Alternatively an existing skb field that is used only on disjoint >>> datapaths, such as ingress-only, could be viable. >> >> A question here. We had metadata support in XDP for cooperation between >> eBPF programs. Do we have something similar in the skb? >> >> E.g in the RSS, if we want to pass some metadata information between >> eBPF program and the logic that generates the vnet header (either hard >> logic in the kernel or another eBPF program). Is there any way that can >> avoid the possible conflicts of qdiscs? > Not that I am aware of. The closest thing is cb[]. > > It'll have to aliase a field like that, that is known unused for the given path. Right, we need to make sure cb is not used by other ones. I'm not sure how hard to achieve that consider Qemu installs the eBPF program but it doesn't deal with networking configurations. > > One other approach that has been used within linear call stacks is out > of band. Like percpu variables softnet_data.xmit.more and > mirred_rec_level. But that is perhaps a bit overwrought for this use > case. Yes, and if we go that way then eBPF turns out to be a burden since we need to invent helpers to access those auxiliary data structure. It would be better then to hard-coded the RSS in the kernel. > >>>>> Instead, you could just run the flow_dissector in tun_put_user if the >>>>> feature is negotiated. Indeed, the flow dissector seems more apt to me >>>>> than BPF here. Note that the flow dissector internally can be >>>>> overridden by a BPF program if the admin so chooses. >>>>> >>>> When this set of patches is related to hash delivery in the virtio-net >>>> packet in general, >>>> it was prepared in context of RSS feature implementation as defined in >>>> virtio spec [1] >>>> In case of RSS it is not enough to run the flow_dissector in tun_put_user: >>>> in tun_ebpf_select_queue the TUN calls eBPF to calculate the hash, >>>> hash type and queue index >>>> according to the (mapped) parameters (key, hash types, indirection >>>> table) received from the guest. >>> TUNSETSTEERINGEBPF was added to support more diverse queue selection >>> than the default in case of multiqueue tun. Not sure what the exact >>> use cases are. >>> >>> But RSS is exactly the purpose of the flow dissector. It is used for >>> that purpose in the software variant RPS. The flow dissector >>> implements a superset of the RSS spec, and certainly computes a >>> four-tuple for TCP/IPv6. In the case of RPS, it is skipped if the NIC >>> has already computed a 4-tuple hash. >>> >>> What it does not give is a type indication, such as >>> VIRTIO_NET_HASH_TYPE_TCPv6. I don't understand how this would be used. >>> In datapaths where the NIC has already computed the four-tuple hash >>> and stored it in skb->hash --the common case for servers--, That type >>> field is the only reason to have to compute again. >> >> The problem is there's no guarantee that the packet comes from the NIC, >> it could be a simple VM2VM or host2VM packet. >> >> And even if the packet is coming from the NIC that calculates the hash >> there's no guarantee that it's the has that guest want (guest may use >> different RSS keys). > Ah yes, of course. > > I would still revisit the need to store a detailed hash_type along with > the hash, as as far I can tell that conveys no actionable information > to the guest. Yes, need to figure out its usage. According to [1], it only mention that storing has type is a charge of driver. Maybe Yuri can answer this. Thanks [1] https://docs.microsoft.com/en-us/windows-hardware/drivers/network/indicating-rss-receive-data >