From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-5.6 required=3.0 tests=BAYES_00,DKIMWL_WL_HIGH, DKIM_SIGNED,DKIM_VALID,DKIM_VALID_AU,HEADER_FROM_DIFFERENT_DOMAINS, MAILING_LIST_MULTI,NICE_REPLY_A,SPF_HELO_NONE,SPF_PASS,USER_AGENT_SANE_1 autolearn=no autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 06B1CC433DF for ; Thu, 16 Jul 2020 04:20:30 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id D3BAF206C1 for ; Thu, 16 Jul 2020 04:20:29 +0000 (UTC) Authentication-Results: mail.kernel.org; dkim=pass (1024-bit key) header.d=redhat.com header.i=@redhat.com header.b="Ucpn2NKD" Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1725965AbgGPEU3 (ORCPT ); Thu, 16 Jul 2020 00:20:29 -0400 Received: from us-smtp-2.mimecast.com ([207.211.31.81]:35938 "EHLO us-smtp-delivery-1.mimecast.com" rhost-flags-OK-OK-OK-FAIL) by vger.kernel.org with ESMTP id S1725270AbgGPEU2 (ORCPT ); Thu, 16 Jul 2020 00:20:28 -0400 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1594873226; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=NMgT9K8k4h6U8jZfojucjF8UBsRWGKHtNgtyATuyZFk=; b=Ucpn2NKD7zzd/idUNilvfLLXE/oFYzpLr+/wlaM7jHKOxcETG8L48PdvaX3NPaCSnNudmi RFeA/VmJIg/FjDc90RKSbeGyPEgo83sR3YyhyGq+6/Jv7SWudt2wI5WPXgR95sfOFjzLSc ZfvytDZDfuLldIbMeXWfaLxPsomr0pM= Received: from mimecast-mx01.redhat.com (mimecast-mx01.redhat.com [209.132.183.4]) (Using TLS) by relay.mimecast.com with ESMTP id us-mta-263-gQeUVn6SPeqxFZab6ULayQ-1; Thu, 16 Jul 2020 00:20:25 -0400 X-MC-Unique: gQeUVn6SPeqxFZab6ULayQ-1 Received: from smtp.corp.redhat.com (int-mx04.intmail.prod.int.phx2.redhat.com [10.5.11.14]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mimecast-mx01.redhat.com (Postfix) with ESMTPS id C6DF3102C7F3; Thu, 16 Jul 2020 04:20:23 +0000 (UTC) Received: from [10.72.12.131] (ovpn-12-131.pek2.redhat.com [10.72.12.131]) by smtp.corp.redhat.com (Postfix) with ESMTP id 8D9175D9DC; Thu, 16 Jul 2020 04:20:10 +0000 (UTC) Subject: Re: [PATCH 0/7] *** IRQ offloading for vDPA *** To: "Zhu, Lingshan" , "Michael S. Tsirkin" , alex williamson , pbonzini@redhat.com, sean.j.christopherson@intel.com, wanpengli@tencent.com Cc: "virtualization@lists.linux-foundation.org" , "netdev@vger.kernel.org" , "kvm@vger.kernel.org" , dan daly References: <1594565524-3394-1-git-send-email-lingshan.zhu@intel.com> <70244d80-08a4-da91-3226-7bfd2019467e@redhat.com> <97032c51-3265-c94a-9ce1-f42fcc6d3075@intel.com> <77318609-85ef-f169-2a1e-500473976d84@redhat.com> <29ab6da8-ed8e-6b91-d658-f3d240543b29@intel.com> <1e91d9dd-d787-beff-2c14-9c76ffc3b285@redhat.com> From: Jason Wang Message-ID: <67c4c41d-9e95-2270-4acb-6f04668c34fa@redhat.com> Date: Thu, 16 Jul 2020 12:20:09 +0800 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:68.0) Gecko/20100101 Thunderbird/68.10.0 MIME-Version: 1.0 In-Reply-To: Content-Type: text/plain; charset=utf-8; format=flowed Content-Transfer-Encoding: 8bit Content-Language: en-US X-Scanned-By: MIMEDefang 2.79 on 10.5.11.14 Sender: netdev-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: netdev@vger.kernel.org On 2020/7/16 下午12:13, Zhu, Lingshan wrote: > > > On 7/16/2020 12:02 PM, Jason Wang wrote: >> >> On 2020/7/16 上午11:59, Zhu, Lingshan wrote: >>> >>> On 7/16/2020 10:59 AM, Jason Wang wrote: >>>> >>>> On 2020/7/16 上午9:39, Zhu, Lingshan wrote: >>>>> >>>>> >>>>> On 7/15/2020 9:43 PM, Jason Wang wrote: >>>>>> >>>>>> On 2020/7/12 下午10:52, Zhu Lingshan wrote: >>>>>>> Hi All, >>>>>>> >>>>>>> This series intends to implement IRQ offloading for >>>>>>> vhost_vdpa. >>>>>>> >>>>>>> By the feat of irq forwarding facilities like posted >>>>>>> interrupt on X86, irq bypass can  help deliver >>>>>>> interrupts to vCPU directly. >>>>>>> >>>>>>> vDPA devices have dedicated hardware backends like VFIO >>>>>>> pass-throughed devices. So it would be possible to setup >>>>>>> irq offloading(irq bypass) for vDPA devices and gain >>>>>>> performance improvements. >>>>>>> >>>>>>> In my testing, with this feature, we can save 0.1ms >>>>>>> in a ping between two VFs on average. >>>>>> >>>>>> >>>>>> Hi Lingshan: >>>>>> >>>>>> During the virtio-networking meeting, Michael spots two possible >>>>>> issues: >>>>>> >>>>>> 1) do we need an new uAPI to stop the irq offloading? >>>>>> 2) can interrupt lost during the eventfd ctx? >>>>>> >>>>>> For 1) I think we probably not, we can allocate an independent >>>>>> eventfd which does not map to MSIX. So the consumer can't match >>>>>> the producer and we fallback to eventfd based irq. >>>>> Hi Jason, >>>>> >>>>> I wonder why we need to stop irq offloading, but if we need to do >>>>> so, maybe a new uAPI would be more intuitive to me, >>>>> but why and who(user? qemu?) shall initialize this process, based >>>>> on what kinda of basis to make the decision? >>>> >>>> >>>> The reason is we may want to fallback to software datapath for some >>>> reason (e.g software assisted live migration). In this case we need >>>> intercept device write to used ring so we can not offloading >>>> virtqueue interrupt in this case. >>> so add a VHOST_VDPA_STOP_IRQ_OFFLOADING? Then do we need a >>> VHOST_VDPA_START_IRQ_OFFLOADING, then let userspace fully control >>> this? Or any better approaches? >> >> >> Probably not, it's as simple as allocating another eventfd (but not >> irqfd), and pass it to vhost-vdpa. Then the offloading is disabled >> since it doesn't have a consumer. > OK, sounds like QEMU work, no need to take care in this series, right? That's my understanding. Thanks > > Thanks, > BR > Zhu Lingshan >> >> Thanks >> >> From mboxrd@z Thu Jan 1 00:00:00 1970 From: Jason Wang Subject: Re: [PATCH 0/7] *** IRQ offloading for vDPA *** Date: Thu, 16 Jul 2020 12:20:09 +0800 Message-ID: <67c4c41d-9e95-2270-4acb-6f04668c34fa@redhat.com> References: <1594565524-3394-1-git-send-email-lingshan.zhu@intel.com> <70244d80-08a4-da91-3226-7bfd2019467e@redhat.com> <97032c51-3265-c94a-9ce1-f42fcc6d3075@intel.com> <77318609-85ef-f169-2a1e-500473976d84@redhat.com> <29ab6da8-ed8e-6b91-d658-f3d240543b29@intel.com> <1e91d9dd-d787-beff-2c14-9c76ffc3b285@redhat.com> Mime-Version: 1.0 Content-Type: text/plain; charset=utf-8; format=flowed Content-Transfer-Encoding: 8bit Return-path: In-Reply-To: Content-Language: en-US Sender: netdev-owner@vger.kernel.org To: "Zhu, Lingshan" , "Michael S. Tsirkin" , alex williamson , pbonzini@redhat.com, sean.j.christopherson@intel.com, wanpengli@tencent.com Cc: "virtualization@lists.linux-foundation.org" , "netdev@vger.kernel.org" , "kvm@vger.kernel.org" , dan daly List-Id: virtualization@lists.linuxfoundation.org On 2020/7/16 下午12:13, Zhu, Lingshan wrote: > > > On 7/16/2020 12:02 PM, Jason Wang wrote: >> >> On 2020/7/16 上午11:59, Zhu, Lingshan wrote: >>> >>> On 7/16/2020 10:59 AM, Jason Wang wrote: >>>> >>>> On 2020/7/16 上午9:39, Zhu, Lingshan wrote: >>>>> >>>>> >>>>> On 7/15/2020 9:43 PM, Jason Wang wrote: >>>>>> >>>>>> On 2020/7/12 下午10:52, Zhu Lingshan wrote: >>>>>>> Hi All, >>>>>>> >>>>>>> This series intends to implement IRQ offloading for >>>>>>> vhost_vdpa. >>>>>>> >>>>>>> By the feat of irq forwarding facilities like posted >>>>>>> interrupt on X86, irq bypass can  help deliver >>>>>>> interrupts to vCPU directly. >>>>>>> >>>>>>> vDPA devices have dedicated hardware backends like VFIO >>>>>>> pass-throughed devices. So it would be possible to setup >>>>>>> irq offloading(irq bypass) for vDPA devices and gain >>>>>>> performance improvements. >>>>>>> >>>>>>> In my testing, with this feature, we can save 0.1ms >>>>>>> in a ping between two VFs on average. >>>>>> >>>>>> >>>>>> Hi Lingshan: >>>>>> >>>>>> During the virtio-networking meeting, Michael spots two possible >>>>>> issues: >>>>>> >>>>>> 1) do we need an new uAPI to stop the irq offloading? >>>>>> 2) can interrupt lost during the eventfd ctx? >>>>>> >>>>>> For 1) I think we probably not, we can allocate an independent >>>>>> eventfd which does not map to MSIX. So the consumer can't match >>>>>> the producer and we fallback to eventfd based irq. >>>>> Hi Jason, >>>>> >>>>> I wonder why we need to stop irq offloading, but if we need to do >>>>> so, maybe a new uAPI would be more intuitive to me, >>>>> but why and who(user? qemu?) shall initialize this process, based >>>>> on what kinda of basis to make the decision? >>>> >>>> >>>> The reason is we may want to fallback to software datapath for some >>>> reason (e.g software assisted live migration). In this case we need >>>> intercept device write to used ring so we can not offloading >>>> virtqueue interrupt in this case. >>> so add a VHOST_VDPA_STOP_IRQ_OFFLOADING? Then do we need a >>> VHOST_VDPA_START_IRQ_OFFLOADING, then let userspace fully control >>> this? Or any better approaches? >> >> >> Probably not, it's as simple as allocating another eventfd (but not >> irqfd), and pass it to vhost-vdpa. Then the offloading is disabled >> since it doesn't have a consumer. > OK, sounds like QEMU work, no need to take care in this series, right? That's my understanding. Thanks > > Thanks, > BR > Zhu Lingshan >> >> Thanks >> >>