From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-7.3 required=3.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,HEADER_FROM_DIFFERENT_DOMAINS,MAILING_LIST_MULTI, NICE_REPLY_A,SPF_HELO_NONE,SPF_PASS,USER_AGENT_SANE_1 autolearn=no autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 53D54C433DB for ; Fri, 12 Mar 2021 18:25:58 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id 23A4164F72 for ; Fri, 12 Mar 2021 18:25:58 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S232975AbhCLSZZ (ORCPT ); Fri, 12 Mar 2021 13:25:25 -0500 Received: from ale.deltatee.com ([204.191.154.188]:53516 "EHLO ale.deltatee.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S233559AbhCLSZH (ORCPT ); Fri, 12 Mar 2021 13:25:07 -0500 DKIM-Signature: v=1; a=rsa-sha256; q=dns/txt; c=relaxed/relaxed; d=deltatee.com; s=20200525; h=Subject:In-Reply-To:MIME-Version:Date: Message-ID:From:References:Cc:To:content-disposition; bh=0Bko3Ie4Uvc54xcT0jJ3kOx/RyID/leZ9OA8KNpjYtM=; b=CvA51+dv8WoK6+gl9j6losyRNn z37elIJkHQw426VAvjyMmd5/0Bd+jcoHqQyCw3hNZXYLCffs7knNFKg37WykjcodyfX3G0aUY9C/N 87MYNvRj2Eqb6cgu0n88BCBDou0hrsXHvB4Nmj2WQOYVly/WtQI9tInfq2+WIY2k0r5NA4XjiQnC8 jSbqLYAUD3AAr9GYj35jdwDqzulDyyaRmX0uMROhg7AzCCnxVMOvjv7zXs8EERUsmx2i5k9gx9zxT EjrLzWUJo5R3pYhi7CTAcr61789tRhIu7upZlkGmerp5l9Bp0T4JHFW5BD5in+k90UHt3tBzvYj37 TaJqDq4A==; Received: from s01060023bee90a7d.cg.shawcable.net ([24.64.145.4] helo=[192.168.0.10]) by ale.deltatee.com with esmtpsa (TLS1.3:ECDHE_RSA_AES_128_GCM_SHA256:128) (Exim 4.92) (envelope-from ) id 1lKmT0-0000WH-0K; Fri, 12 Mar 2021 11:24:55 -0700 To: Robin Murphy , linux-kernel@vger.kernel.org, linux-nvme@lists.infradead.org, linux-block@vger.kernel.org, linux-pci@vger.kernel.org, linux-mm@kvack.org, iommu@lists.linux-foundation.org Cc: Minturn Dave B , John Hubbard , Dave Hansen , Ira Weiny , Matthew Wilcox , =?UTF-8?Q?Christian_K=c3=b6nig?= , Jason Gunthorpe , Jason Ekstrand , Daniel Vetter , Dan Williams , Stephen Bates , Jakowski Andrzej , Christoph Hellwig , Xiong Jianxin References: <20210311233142.7900-1-logang@deltatee.com> <6b9be188-1ec7-527c-ae47-3f5b4e153758@arm.com> <90a2825c-da2f-c031-a70f-08c5efb3db56@arm.com> From: Logan Gunthorpe Message-ID: Date: Fri, 12 Mar 2021 11:24:49 -0700 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:78.0) Gecko/20100101 Thunderbird/78.8.0 MIME-Version: 1.0 In-Reply-To: <90a2825c-da2f-c031-a70f-08c5efb3db56@arm.com> Content-Type: text/plain; charset=utf-8 Content-Language: en-CA Content-Transfer-Encoding: 8bit X-SA-Exim-Connect-IP: 24.64.145.4 X-SA-Exim-Rcpt-To: jianxin.xiong@intel.com, hch@lst.de, andrzej.jakowski@intel.com, sbates@raithlin.com, dan.j.williams@intel.com, daniel.vetter@ffwll.ch, jason@jlekstrand.net, jgg@ziepe.ca, christian.koenig@amd.com, willy@infradead.org, iweiny@intel.com, dave.hansen@linux.intel.com, jhubbard@nvidia.com, dave.b.minturn@intel.com, iommu@lists.linux-foundation.org, linux-mm@kvack.org, linux-pci@vger.kernel.org, linux-block@vger.kernel.org, linux-nvme@lists.infradead.org, linux-kernel@vger.kernel.org, robin.murphy@arm.com X-SA-Exim-Mail-From: logang@deltatee.com Subject: Re: [RFC PATCH v2 00/11] Add support to dma_map_sg for P2PDMA X-SA-Exim-Version: 4.2.1 (built Wed, 08 May 2019 21:11:16 +0000) X-SA-Exim-Scanned: Yes (on ale.deltatee.com) Precedence: bulk List-ID: X-Mailing-List: linux-pci@vger.kernel.org On 2021-03-12 10:46 a.m., Robin Murphy wrote: > On 2021-03-12 16:18, Logan Gunthorpe wrote: >> >> >> On 2021-03-12 8:51 a.m., Robin Murphy wrote: >>> On 2021-03-11 23:31, Logan Gunthorpe wrote: >>>> Hi, >>>> >>>> This is a rework of the first half of my RFC for doing P2PDMA in >>>> userspace >>>> with O_DIRECT[1]. >>>> >>>> The largest issue with that series was the gross way of flagging P2PDMA >>>> SGL segments. This RFC proposes a different approach, (suggested by >>>> Dan Williams[2]) which uses the third bit in the page_link field of the >>>> SGL. >>>> >>>> This approach is a lot less hacky but comes at the cost of adding a >>>> CONFIG_64BIT dependency to CONFIG_PCI_P2PDMA and using up the last >>>> scarce bit in the page_link. For our purposes, a 64BIT restriction is >>>> acceptable but it's not clear if this is ok for all usecases hoping >>>> to make use of P2PDMA. >>>> >>>> Matthew Wilcox has already suggested (off-list) that this is the wrong >>>> approach, preferring a new dma mapping operation and an SGL >>>> replacement. I >>>> don't disagree that something along those lines would be a better long >>>> term solution, but it involves overcoming a lot of challenges to get >>>> there. Creating a new mapping operation still means adding support to >>>> more >>>> than 25 dma_map_ops implementations (many of which are on obscure >>>> architectures) or creating a redundant path to fallback with >>>> dma_map_sg() >>>> for every driver that uses the new operation. This RFC is an approach >>>> that doesn't require overcoming these blocks. >>> >>> I don't really follow that argument - you're only adding support to two >>> implementations with the awkward flag, so why would using a dedicated >>> operation instead be any different? Whatever callers need to do if >>> dma_pci_p2pdma_supported() says no, they could equally do if >>> dma_map_p2p_sg() (or whatever) returns -ENXIO, no? >> >> The thing is if the dma_map_sg doesn't support P2PDMA then P2PDMA >> transactions cannot be done, but regular transactions can still go >> through as they always did. >> >> But replacing dma_map_sg() with dma_map_new() affects all operations, >> P2PDMA or otherwise. If dma_map_new() isn't supported it can't simply >> not support P2PDMA; it has to maintain a fallback path to dma_map_sg(). > > But AFAICS the equivalent fallback path still has to exist either way. > My impression so far is that callers would end up looking something like > this: > >     if (dma_pci_p2pdma_supported()) { >         if (dma_map_sg(...) < 0) >             //do non-p2p fallback due to p2p failure >     } else { >         //do non-p2p fallback due to lack of support >     } > > at which point, simply: > >     if (dma_map_sg_p2p(...) < 0) >         //do non-p2p fallback either way > > seems entirely reasonable. What am I missing? No, that's not how it works. The dma_pci_p2pdma_supported() flag gates whether P2PDMA pages will be used at a much higher level. Currently with NVMe, if P2PDMA is supported, it sets the QUEUE_FLAG_PCI_P2PDMA on the block queue. This is then tested with blk_queue_pci_p2pdma() before any P2PDMA transaction with the block device is started. In the following patches that could add userspace support, blk_queue_pci_p2pdma() is used to add a flag to GUP which allows P2PDMA pages into the driver via O_DIRECT. There is no fallback path on the dma_map_sg() operation if p2pdma is not supported. dma_map_sg() is always used. The code simply prevents pages from getting there in the first place. A new DMA operation would have to be: if (dma_has_new_operation()) { // create input for dma_map_new_op() // map with dma_map_new_op() // parse output of dma_map_new_op() } else { // create SGL dma_map_sg() // convert SGL to hardware map } And this if statement has nothing to do with p2pdma support or not. > > Let's not pretend that overloading an existing API means we can start > feeding P2P pages into any old subsystem/driver without further changes > - there already *are* at least some that retry ad infinitum if DMA > mapping fails (the USB layer springs to mind...) and thus wouldn't > handle the PCI_P2PDMA_MAP_NOT_SUPPORTED case acceptably. Yes, nobody is pretending that at all. We are being very careful with P2PDMA pages and we don't want them to get into any driver that isn't explicitly written to expect them. With the code in the current kernel this is all very explicit and done ahead of time (with QUEUE_FLAG_PCI_P2PDMA and similar). Once the pages get into userspace, GUP will reject them unless the driver getting the pages specifically sets a flag indicating support for them. >> Given that the inputs and outputs for dma_map_new() will be completely >> different data structures this will be quite a lot of similar paths >> required in the driver. (ie mapping a bvec to the input struct and the >> output struct to hardware requirements) If a bug crops up in the old >> dma_map_sg(), developers might not notice it for some time seeing it >> won't be used on the most popular architectures. > > Huh? I'm specifically suggesting a new interface that takes the *same* > data structure (at least to begin with), but just gives us more > flexibility in terms of introducing p2p-aware behaviour somewhat more > safely. Yes, we already know that we ultimately want something better > than scatterlists for representing things like this and dma-buf imports, > but that hardly has to happen overnight. Ok, well that's not what Matthew had suggested off list. He specifically was suggesting new data structures. And yes, that isn't going to happen overnight. If we have a dma_map_sg() and a dma_map_sg_p2p() that are identical except dma_map_sg_p2p() supports p2pdma memory and can return -1 then that doesn't sound so bad to me. So dma_map_sg() would simply be call dma_map_sg_p2p() and add the BUG() on the return code. Though I really don't see the benefit of the extra annotations. I don't think it really adds any value. The tricky thing in the API is the SGL which needs to flag segments for P2PDMA and the new function doesn't solve that. Logan