From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-8.2 required=3.0 tests=DKIM_SIGNED,DKIM_VALID, HEADER_FROM_DIFFERENT_DOMAINS,INCLUDES_PATCH,MAILING_LIST_MULTI,SIGNED_OFF_BY, SPF_HELO_NONE,SPF_PASS,URIBL_BLOCKED,USER_AGENT_SANE_1 autolearn=unavailable autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id BCEF7C433DF for ; Fri, 19 Jun 2020 14:15:10 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id 9481820776 for ; Fri, 19 Jun 2020 14:15:10 +0000 (UTC) Authentication-Results: mail.kernel.org; dkim=pass (2048-bit key) header.d=kernel-dk.20150623.gappssmtp.com header.i=@kernel-dk.20150623.gappssmtp.com header.b="d1Z7NBN/" Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1733081AbgFSOPJ (ORCPT ); Fri, 19 Jun 2020 10:15:09 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:48654 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1733063AbgFSOPF (ORCPT ); Fri, 19 Jun 2020 10:15:05 -0400 Received: from mail-pf1-x444.google.com (mail-pf1-x444.google.com [IPv6:2607:f8b0:4864:20::444]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 7748AC0613EF for ; Fri, 19 Jun 2020 07:15:04 -0700 (PDT) Received: by mail-pf1-x444.google.com with SMTP id b201so4505414pfb.0 for ; Fri, 19 Jun 2020 07:15:04 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=kernel-dk.20150623.gappssmtp.com; s=20150623; h=subject:to:cc:references:from:message-id:date:user-agent :mime-version:in-reply-to:content-language:content-transfer-encoding; bh=vKgaeJnyg9cLq81+E0+J+QD8Om9Rb9c3m9dc4xG9lHc=; b=d1Z7NBN/J4rcYQozroSRSec7V4c8J8uQC7Sagh9+3gHV4/EUGZjouMvG8jmc1IERbT KeSibIf9YZQpn3XPSo3FsVPSQaInICZH8MzKT3/o9y7ADIsXYfXi7S8q2iG0/OqCUS2x xzhP8+m1g2l8uPg2MY1CJ9tnWZE+b71fSw0yzq4gsDq3oGThXsuqRq/+X6JVmBwUxVR8 XIncsGKi0j3r8UBMdHnElslFfoK/euFbUagqpNTJ0/epDTJxZ9ttSTTUJetKEtxHDmkT AL4Biv192ULTPLTQYQxQs4X6brZMECUR7nzq5Uze2Fmi2xrsmvgCbtgWbxgIFSm9VJpL CCCA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:subject:to:cc:references:from:message-id:date :user-agent:mime-version:in-reply-to:content-language :content-transfer-encoding; bh=vKgaeJnyg9cLq81+E0+J+QD8Om9Rb9c3m9dc4xG9lHc=; b=t0QJ+KatLCBhgCnx1Yw9sc3AFz1AENyddoVabn0ExhQMkJHIrWvLYAG2RvR6qRkwSH xTAiKCUV6zMrrPNfFrCht5U2bttEGhhHuQXQRdr0o84BKBPcCv5DTWRWdGTZz8TiIlsR 7d/fGhGS45Fd0HCP2vtH+U85NnVBObaDbeSzlXdYPkE2rGNc+Rym0xRztaKm2ghtzmTc PRttqLH5qZznTVMsZ67IevrTTWchxNdzl8B37mYL/628ElWIAbolZVEbJvkKg5vrtTwY sDCNrAHkIZdg0Nt5WgadEgFoArss30uFu9JK5qRj+J3OVyF1AbVC4N1GbiXz9ps92RuH Jnqg== X-Gm-Message-State: AOAM533zDbHjaFC0Lm5GQm5IOmxE2YBJn0Lg5Vce3+CxmiHLby1j4aZR cMXLKrFB20EG3L3YPqpnwNK9Ww== X-Google-Smtp-Source: ABdhPJxJ5rUdZQlZ85uNdWYu0aORrSm4Hs/gqNUu5fZPN0AfOVDyqgnm2fFHhgtfgfuyb46iF3hR1Q== X-Received: by 2002:a63:7987:: with SMTP id u129mr3109208pgc.353.1592576103577; Fri, 19 Jun 2020 07:15:03 -0700 (PDT) Received: from [192.168.1.188] ([66.219.217.173]) by smtp.gmail.com with ESMTPSA id f29sm2504625pga.59.2020.06.19.07.15.01 (version=TLS1_3 cipher=TLS_AES_128_GCM_SHA256 bits=128/128); Fri, 19 Jun 2020 07:15:03 -0700 (PDT) Subject: Re: [PATCH 3/3] io_uring: add support for zone-append To: "javier.gonz@samsung.com" , Damien Le Moal Cc: Kanchan Joshi , "viro@zeniv.linux.org.uk" , "bcrl@kvack.org" , "linux-fsdevel@vger.kernel.org" , "linux-kernel@vger.kernel.org" , "linux-aio@kvack.org" , "io-uring@vger.kernel.org" , "linux-block@vger.kernel.org" , "selvakuma.s1@samsung.com" , "nj.shetty@samsung.com" References: <1592414619-5646-1-git-send-email-joshi.k@samsung.com> <1592414619-5646-4-git-send-email-joshi.k@samsung.com> <20200618083529.ciifu4chr4vrv2j5@mpHalley.local> <20200618091113.eu2xdp6zmdooy5d2@mpHalley.local> <20200619094149.uaorbger326s6yzz@mpHalley.local> From: Jens Axboe Message-ID: <31f1c27e-4a3d-a411-3d3b-f88a2d92ce7b@kernel.dk> Date: Fri, 19 Jun 2020 08:15:01 -0600 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:68.0) Gecko/20100101 Thunderbird/68.8.0 MIME-Version: 1.0 In-Reply-To: <20200619094149.uaorbger326s6yzz@mpHalley.local> Content-Type: text/plain; charset=utf-8 Content-Language: en-US Content-Transfer-Encoding: 7bit Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 6/19/20 3:41 AM, javier.gonz@samsung.com wrote: > Jens, > > Would you have time to answer a question below in this thread? > > On 18.06.2020 11:11, javier.gonz@samsung.com wrote: >> On 18.06.2020 08:47, Damien Le Moal wrote: >>> On 2020/06/18 17:35, javier.gonz@samsung.com wrote: >>>> On 18.06.2020 07:39, Damien Le Moal wrote: >>>>> On 2020/06/18 2:27, Kanchan Joshi wrote: >>>>>> From: Selvakumar S >>>>>> >>>>>> Introduce three new opcodes for zone-append - >>>>>> >>>>>> IORING_OP_ZONE_APPEND : non-vectord, similiar to IORING_OP_WRITE >>>>>> IORING_OP_ZONE_APPENDV : vectored, similar to IORING_OP_WRITEV >>>>>> IORING_OP_ZONE_APPEND_FIXED : append using fixed-buffers >>>>>> >>>>>> Repurpose cqe->flags to return zone-relative offset. >>>>>> >>>>>> Signed-off-by: SelvaKumar S >>>>>> Signed-off-by: Kanchan Joshi >>>>>> Signed-off-by: Nitesh Shetty >>>>>> Signed-off-by: Javier Gonzalez >>>>>> --- >>>>>> fs/io_uring.c | 72 +++++++++++++++++++++++++++++++++++++++++-- >>>>>> include/uapi/linux/io_uring.h | 8 ++++- >>>>>> 2 files changed, 77 insertions(+), 3 deletions(-) >>>>>> >>>>>> diff --git a/fs/io_uring.c b/fs/io_uring.c >>>>>> index 155f3d8..c14c873 100644 >>>>>> --- a/fs/io_uring.c >>>>>> +++ b/fs/io_uring.c >>>>>> @@ -649,6 +649,10 @@ struct io_kiocb { >>>>>> unsigned long fsize; >>>>>> u64 user_data; >>>>>> u32 result; >>>>>> +#ifdef CONFIG_BLK_DEV_ZONED >>>>>> + /* zone-relative offset for append, in bytes */ >>>>>> + u32 append_offset; >>>>> >>>>> this can overflow. u64 is needed. >>>> >>>> We chose to do it this way to start with because struct io_uring_cqe >>>> only has space for u32 when we reuse the flags. >>>> >>>> We can of course create a new cqe structure, but that will come with >>>> larger changes to io_uring for supporting append. >>>> >>>> Do you believe this is a better approach? >>> >>> The problem is that zone size are 32 bits in the kernel, as a number >>> of sectors. So any device that has a zone size smaller or equal to >>> 2^31 512B sectors can be accepted. Using a zone relative offset in >>> bytes for returning zone append result is OK-ish, but to match the >>> kernel supported range of possible zone size, you need 31+9 bits... >>> 32 does not cut it. >> >> Agree. Our initial assumption was that u32 would cover current zone size >> requirements, but if this is a no-go, we will take the longer path. > > Converting to u64 will require a new version of io_uring_cqe, where we > extend at least 32 bits. I believe this will need a whole new allocation > and probably ioctl(). > > Is this an acceptable change for you? We will of course add support for > liburing when we agree on the right way to do this. If you need 64-bit of return value, then it's not going to work. Even with the existing patches, reusing cqe->flags isn't going to fly, as it would conflict with eg doing zone append writes with automatic buffer selection. We're not changing the io_uring_cqe. It's important to keep it lean, and any other request type is generally fine with 64-bit tag + 32-bit result (and 32-bit flags on the side) for completions. Only viable alternative I see would be to provide an area to store this information, and pass in a pointer to this at submission time through the sqe. One issue I do see with that is if we only have this information available at completion time, then we'd need some async punt to copy it to user space... Generally not ideal. A hackier approach would be to play some tricks with cqe->res and cqe->flags, setting aside a flag to denote an extension of cqe->res. That would mean excluding zone append (etc) from using buffer selection, which probably isn't a huge deal. It'd be more problematic for any other future flags. But if you just need 40 bits, then it could certainly work. Rigth now, if cqe->flags & 1 is set, then (cqe->flags >> 16) is the buffer ID. You could define IORING_CQE_F_ZONE_FOO to be bit 1, so that: uint64_t val = cqe->res; // assuming non-error here if (cqe->flags & IORING_CQE_F_ZONE_FOO) val |= (cqe->flags >> 16) << 32ULL; and hence use the upper 16 bits of cqe->flags for the upper bits of your (then) 48-bit total value. -- Jens Axboe