From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-15.3 required=3.0 tests=BAYES_00, HEADER_FROM_DIFFERENT_DOMAINS,INCLUDES_CR_TRAILER,INCLUDES_PATCH, MAILING_LIST_MULTI,NICE_REPLY_A,SPF_HELO_NONE,SPF_PASS,UNPARSEABLE_RELAY, USER_AGENT_SANE_1 autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 8649DC433DB for ; Wed, 17 Mar 2021 03:49:47 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id 4F1BD64F37 for ; Wed, 17 Mar 2021 03:49:46 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S229519AbhCQDtO (ORCPT ); Tue, 16 Mar 2021 23:49:14 -0400 Received: from out30-43.freemail.mail.aliyun.com ([115.124.30.43]:34141 "EHLO out30-43.freemail.mail.aliyun.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229688AbhCQDtD (ORCPT ); Tue, 16 Mar 2021 23:49:03 -0400 X-Alimail-AntiSpam: AC=PASS;BC=-1|-1;BR=01201311R591e4;CH=green;DM=||false|;DS=||;FP=0|-1|-1|-1|0|-1|-1|-1;HT=e01e04400;MF=jefflexu@linux.alibaba.com;NM=1;PH=DS;RN=6;SR=0;TI=SMTPD_---0USDZqg4_1615952940; Received: from admindeMacBook-Pro-2.local(mailfrom:jefflexu@linux.alibaba.com fp:SMTPD_---0USDZqg4_1615952940) by smtp.aliyun-inc.com(127.0.0.1); Wed, 17 Mar 2021 11:49:00 +0800 Subject: Re: [RFC PATCH 08/11] block: use per-task poll context to implement bio based io poll From: JeffleXu To: Ming Lei Cc: Jens Axboe , linux-block@vger.kernel.org, Christoph Hellwig , Mike Snitzer , dm-devel@redhat.com References: <20210316031523.864506-1-ming.lei@redhat.com> <20210316031523.864506-9-ming.lei@redhat.com> Message-ID: Date: Wed, 17 Mar 2021 11:49:00 +0800 User-Agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.15; rv:78.0) Gecko/20100101 Thunderbird/78.7.0 MIME-Version: 1.0 In-Reply-To: Content-Type: text/plain; charset=utf-8 Content-Language: en-US Content-Transfer-Encoding: 7bit Precedence: bulk List-ID: X-Mailing-List: linux-block@vger.kernel.org On 3/16/21 7:00 PM, JeffleXu wrote: > > > On 3/16/21 3:17 PM, Ming Lei wrote: >> On Tue, Mar 16, 2021 at 02:46:08PM +0800, JeffleXu wrote: >>> It is a giant progress to gather all split bios that need to be polled >>> in a per-task queue. Still some comments below. >>> >>> >>> On 3/16/21 11:15 AM, Ming Lei wrote: >>>> Currently bio based IO poll needs to poll all hw queue blindly, this way >>>> is very inefficient, and the big reason is that we can't pass bio >>>> submission result to io poll task. >>>> >>>> In IO submission context, store associated underlying bios into the >>>> submission queue and save 'cookie' poll data in bio->bi_iter.bi_private_data, >>>> and return current->pid to caller of submit_bio() for any DM or bio based >>>> driver's IO, which is submitted from FS. >>>> >>>> In IO poll context, the passed cookie tells us the PID of submission >>>> context, and we can find the bio from that submission context. Moving >>>> bio from submission queue to poll queue of the poll context, and keep >>>> polling until these bios are ended. Remove bio from poll queue if the >>>> bio is ended. Add BIO_DONE and BIO_END_BY_POLL for such purpose. >>>> >>>> Usually submission shares context with io poll. The per-task poll context >>>> is just like stack variable, and it is cheap to move data between the two >>>> per-task queues. >>>> >>>> Signed-off-by: Ming Lei >>>> --- >>>> block/bio.c | 5 ++ >>>> block/blk-core.c | 74 +++++++++++++++++- >>>> block/blk-mq.c | 156 +++++++++++++++++++++++++++++++++++++- >>>> include/linux/blk_types.h | 3 + >>>> 4 files changed, 235 insertions(+), 3 deletions(-) >>>> >>>> diff --git a/block/bio.c b/block/bio.c >>>> index a1c4d2900c7a..bcf5eca0e8e3 100644 >>>> --- a/block/bio.c >>>> +++ b/block/bio.c >>>> @@ -1402,6 +1402,11 @@ static inline bool bio_remaining_done(struct bio *bio) >>>> **/ >>>> void bio_endio(struct bio *bio) >>>> { >>>> + /* BIO_END_BY_POLL has to be set before calling submit_bio */ >>>> + if (bio_flagged(bio, BIO_END_BY_POLL)) { >>>> + bio_set_flag(bio, BIO_DONE); >>>> + return; >>>> + } >>>> again: >>>> if (!bio_remaining_done(bio)) >>>> return; >>>> diff --git a/block/blk-core.c b/block/blk-core.c >>>> index a082bbc856fb..970b23fa2e6e 100644 >>>> --- a/block/blk-core.c >>>> +++ b/block/blk-core.c >>>> @@ -854,6 +854,40 @@ static inline void blk_bio_poll_preprocess(struct request_queue *q, >>>> bio->bi_opf |= REQ_TAG; >>>> } >>>> >>>> +static bool blk_bio_poll_prep_submit(struct io_context *ioc, struct bio *bio) >>>> +{ >>>> + struct blk_bio_poll_data data = { >>>> + .bio = bio, >>>> + }; >>>> + struct blk_bio_poll_ctx *pc = ioc->data; >>>> + unsigned int queued; >>>> + >>>> + /* lock is required if there is more than one writer */ >>>> + if (unlikely(atomic_read(&ioc->nr_tasks) > 1)) { >>>> + spin_lock(&pc->lock); >>>> + queued = kfifo_put(&pc->sq, data); >>>> + spin_unlock(&pc->lock); >>>> + } else { >>>> + queued = kfifo_put(&pc->sq, data); >>>> + } >>>> + >>>> + /* >>>> + * Now the bio is added per-task fifo, mark it as END_BY_POLL, >>>> + * so we can save cookie into this bio after submit_bio(). >>>> + */ >>>> + if (queued) >>>> + bio_set_flag(bio, BIO_END_BY_POLL); >>>> + else >>>> + bio->bi_opf &= ~(REQ_HIPRI | REQ_TAG); >>>> + >>>> + return queued; >>>> +} >>> >>> The size of kfifo is limited, and it seems that once the sq of kfifio is >>> full, REQ_HIPRI flag is cleared and the corresponding bio is actually >>> enqueued into the default hw queue, which is IRQ driven. >> >> Yeah, this patch starts with 64 queue depth, and we can increase it to >> 128, which should cover most of cases. >> >>> >>> >>>> + >>>> +static void blk_bio_poll_post_submit(struct bio *bio, blk_qc_t cookie) >>>> +{ >>>> + bio->bi_iter.bi_private_data = cookie; >>>> +} >>>> + >>>> static noinline_for_stack bool submit_bio_checks(struct bio *bio) >>>> { >>>> struct block_device *bdev = bio->bi_bdev; >>>> @@ -1008,7 +1042,7 @@ static blk_qc_t __submit_bio(struct bio *bio) >>>> * bio_list_on_stack[1] contains bios that were submitted before the current >>>> * ->submit_bio_bio, but that haven't been processed yet. >>>> */ >>>> -static blk_qc_t __submit_bio_noacct(struct bio *bio) >>>> +static blk_qc_t __submit_bio_noacct_int(struct bio *bio, struct io_context *ioc) >>>> { >>>> struct bio_list bio_list_on_stack[2]; >>>> blk_qc_t ret = BLK_QC_T_NONE; >>>> @@ -1031,7 +1065,16 @@ static blk_qc_t __submit_bio_noacct(struct bio *bio) >>>> bio_list_on_stack[1] = bio_list_on_stack[0]; >>>> bio_list_init(&bio_list_on_stack[0]); >>>> >>>> - ret = __submit_bio(bio); >>>> + if (ioc && queue_is_mq(q) && >>>> + (bio->bi_opf & (REQ_HIPRI | REQ_TAG))) { >>>> + bool queued = blk_bio_poll_prep_submit(ioc, bio); >>>> + >>>> + ret = __submit_bio(bio); >>>> + if (queued) >>>> + blk_bio_poll_post_submit(bio, ret); >>>> + } else { >>>> + ret = __submit_bio(bio); >>>> + } >>>> >>>> /* >>>> * Sort new bios into those for a lower level and those for the >>>> @@ -1057,6 +1100,33 @@ static blk_qc_t __submit_bio_noacct(struct bio *bio) >>>> return ret; >>>> } >>>> >>>> +static inline blk_qc_t __submit_bio_noacct_poll(struct bio *bio, >>>> + struct io_context *ioc) >>>> +{ >>>> + struct blk_bio_poll_ctx *pc = ioc->data; >>>> + int entries = kfifo_len(&pc->sq); >>>> + >>>> + __submit_bio_noacct_int(bio, ioc); >>>> + >>>> + /* bio submissions queued to per-task poll context */ >>>> + if (kfifo_len(&pc->sq) > entries) >>>> + return current->pid; >>>> + >>>> + /* swapper's pid is 0, but it can't submit poll IO for us */ >>>> + return 0; >>>> +} >>>> + >>>> +static inline blk_qc_t __submit_bio_noacct(struct bio *bio) >>>> +{ >>>> + struct io_context *ioc = current->io_context; >>>> + >>>> + if (ioc && ioc->data && (bio->bi_opf & REQ_HIPRI)) >>>> + return __submit_bio_noacct_poll(bio, ioc); >>>> + >>>> + return __submit_bio_noacct_int(bio, NULL); >>>> +} >>>> + >>>> + >>>> static blk_qc_t __submit_bio_noacct_mq(struct bio *bio) >>>> { >>>> struct bio_list bio_list[2] = { }; >>>> diff --git a/block/blk-mq.c b/block/blk-mq.c >>>> index 03f59915fe2c..4e6f1467d303 100644 >>>> --- a/block/blk-mq.c >>>> +++ b/block/blk-mq.c >>>> @@ -3865,14 +3865,168 @@ static inline int blk_mq_poll_hctx(struct request_queue *q, >>>> return ret; >>>> } >>>> >>>> +static blk_qc_t bio_get_poll_cookie(struct bio *bio) >>>> +{ >>>> + return bio->bi_iter.bi_private_data; >>>> +} >>>> + >>>> +static int blk_mq_poll_io(struct bio *bio) >>>> +{ >>>> + struct request_queue *q = bio->bi_bdev->bd_disk->queue; >>>> + blk_qc_t cookie = bio_get_poll_cookie(bio); >>>> + int ret = 0; >>>> + >>>> + if (!bio_flagged(bio, BIO_DONE) && blk_qc_t_valid(cookie)) { >>>> + struct blk_mq_hw_ctx *hctx = >>>> + q->queue_hw_ctx[blk_qc_t_to_queue_num(cookie)]; >>>> + >>>> + ret += blk_mq_poll_hctx(q, hctx); >>>> + } >>>> + return ret; >>>> +} >>>> + >>>> +static int blk_bio_poll_and_end_io(struct request_queue *q, >>>> + struct blk_bio_poll_ctx *poll_ctx) >>>> +{ >>>> + struct blk_bio_poll_data *poll_data = &poll_ctx->pq[0]; >>>> + int ret = 0; >>>> + int i; >>>> + >>>> + for (i = 0; i < BLK_BIO_POLL_PQ_SZ; i++) { >>>> + struct bio *bio = poll_data[i].bio; >>>> + >>>> + if (!bio) >>>> + continue; >>>> + >>>> + ret += blk_mq_poll_io(bio); >>>> + if (bio_flagged(bio, BIO_DONE)) { >>>> + poll_data[i].bio = NULL; >>>> + >>>> + /* clear BIO_END_BY_POLL and end me really */ >>>> + bio_clear_flag(bio, BIO_END_BY_POLL); >>>> + bio_endio(bio); >>>> + } >>>> + } >>>> + return ret; >>>> +} >>> >>> When there are multiple threads polling, saying thread A and thread B, >>> then there's one bio which should be polled by thread A (the pid is >>> passed to thread A), while it's actually completed by thread B. In this >>> case, when the bio is completed by thread B, the bio is not really >>> completed and one extra blk_poll() still needs to be called. >> >> When this happens, the dm bio can't be completed, and the associated >> kiocb can't be completed too, io_uring or other poll code context will >> keep calling blk_poll() by passing thread A's pid until this dm bio is >> done, since the dm bio is submitted from thread A. >> > > This will affect the multi-thread polling performance. I tested > dm-stripe, in which every bio will be split and enqueued into all > underlying devices, and thus amplify the interference between multiple > threads. > > Test Result: > IOPS: 332k (IRQ) -> 363k (iopoll), aka ~10% performance gain Sorry this performance drop is not related to this bio refcount issue here. Still it's due to the limited kfifo size. I did another through test on another machine (aarch64 with more nvme disks). - Without mentioned specifically, the configuration is 'iodepth=128, kfifo queue depth =128'. - The number before '->' indicates the IOPS in IRQ mode, i.e., 'hipri=0', while the number after '->' indicates the IOPS in polling mode, i.e., 'hipri=1'. ``` 3-threads dm-linear-3 targets (4k randread IOPS, unit K) 5.12-rc1: 667 leiming: 674 -> 849 ours 8353c1a: 623 -> 811 3-threads dm-stripe-3 targets (12k randread IOPS, unit K) 5.12-rc1: 321 leiming: 313 -> 349 leiming : 313 -> 409 (iodepth=32, kfifo queue depth =128) leiming : 314 -> 409 (iodepth=128, kfifo queue depth =512) ours 8353c1a: 310 -> 406 1-thread dm-linear-3 targets (4k randread IOPS, unit K) 5.12-rc1: 224 leiming: 218 -> 288 ours 8353c1a: 210 -> 280 1-threads dm-stripe-3 targets (12k randread IOPS, unit K) 5.12-rc1: 109 leiming: 107 -> 120 leiming : 107 -> 145 (iodepth=32, kfifo queue depth =128) leiming : 108 -> 145 (iodepth=128, kfifo queue depth =512) ours 8353c1a: 107 -> 146 ``` Some hints: 1. When configured as 'iodepth=128, kfifo queue depth =128', dm-stripe doesn't perform well in polling mode. It's because it's more likely that the original bio will be split into split bios in dm-stripe, and thus kfifo will be more likely used up in this case. So the size of kfifo need to be tuned according to iodepth and the IO load. Thus exporting the size of kfifo as a sysfs entry may be need in the following patch. 2. It indicates a performance drop of my patch in IRQ mode, compared to the original 5.12-rc1. I doubt maybe it's due to extra code mixed in blk-core, such as __submit_bio_noacct()... > > > Test Environment: > > nvme.poll_queues = 3 > > BLK_BIO_POLL_SQ_SZ = 128 > > dmsetup create testdev --table "0 629145600 striped 3 8 /dev/nvme0n1 0 > /dev/nvme1n1 0 /dev/nvme4n1 0" > > > ``` > $cat fio.conf > [global] > name=iouring-sqpoll-iopoll-1 > ioengine=io_uring > iodepth=128 > numjobs=1 > thread > rw=randread > direct=1 > hipri=1 > runtime=10 > time_based > group_reporting > randrepeat=0 > filename=/dev/mapper/testdev > bs=12k > > [job-1] > cpus_allowed=14 > > [job-2] > cpus_allowed=16 > > [job-3] > cpus_allowed=84 > ``` > -- Thanks, Jeffle From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-15.3 required=3.0 tests=BAYES_00, HEADER_FROM_DIFFERENT_DOMAINS,INCLUDES_CR_TRAILER,INCLUDES_PATCH, MAILING_LIST_MULTI,NICE_REPLY_A,SPF_HELO_NONE,SPF_PASS,UNPARSEABLE_RELAY, USER_AGENT_SANE_1 autolearn=unavailable autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id CEB8AC43381 for ; Thu, 18 Mar 2021 09:14:57 +0000 (UTC) Received: from us-smtp-delivery-124.mimecast.com (us-smtp-delivery-124.mimecast.com [216.205.24.124]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-SHA384 (256/256 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPS id 3895D64F33 for ; Thu, 18 Mar 2021 09:14:57 +0000 (UTC) DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org 3895D64F33 Authentication-Results: mail.kernel.org; dmarc=fail (p=none dis=none) header.from=linux.alibaba.com Authentication-Results: mail.kernel.org; spf=tempfail smtp.mailfrom=dm-devel-bounces@redhat.com Received: from mimecast-mx01.redhat.com (mimecast-mx01.redhat.com [209.132.183.4]) (Using TLS) by relay.mimecast.com with ESMTP id us-mta-349-8itDVSHfOSqckU4xfdW8ow-1; Thu, 18 Mar 2021 05:14:52 -0400 X-MC-Unique: 8itDVSHfOSqckU4xfdW8ow-1 Received: from smtp.corp.redhat.com (int-mx05.intmail.prod.int.phx2.redhat.com [10.5.11.15]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mimecast-mx01.redhat.com (Postfix) with ESMTPS id D306C81621; Thu, 18 Mar 2021 09:14:47 +0000 (UTC) Received: from colo-mx.corp.redhat.com (colo-mx02.intmail.prod.int.phx2.redhat.com [10.5.11.21]) by smtp.corp.redhat.com (Postfix) with ESMTPS id B1D6263634; Thu, 18 Mar 2021 09:14:47 +0000 (UTC) Received: from lists01.pubmisc.prod.ext.phx2.redhat.com (lists01.pubmisc.prod.ext.phx2.redhat.com [10.5.19.33]) by colo-mx.corp.redhat.com (Postfix) with ESMTP id 5DBAF4BB7C; Thu, 18 Mar 2021 09:14:47 +0000 (UTC) Received: from smtp.corp.redhat.com (int-mx03.intmail.prod.int.rdu2.redhat.com [10.11.54.3]) by lists01.pubmisc.prod.ext.phx2.redhat.com (8.13.8/8.13.8) with ESMTP id 12H3nAJG014372 for ; Tue, 16 Mar 2021 23:49:10 -0400 Received: by smtp.corp.redhat.com (Postfix) id 50A361003348; Wed, 17 Mar 2021 03:49:10 +0000 (UTC) Received: from mimecast-mx02.redhat.com (mimecast01.extmail.prod.ext.rdu2.redhat.com [10.11.55.17]) by smtp.corp.redhat.com (Postfix) with ESMTPS id 4C4E110FFE70 for ; Wed, 17 Mar 2021 03:49:08 +0000 (UTC) Received: from us-smtp-1.mimecast.com (us-smtp-delivery-1.mimecast.com [205.139.110.120]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-SHA384 (256/256 bits)) (No client certificate requested) by mimecast-mx02.redhat.com (Postfix) with ESMTPS id 11EE285A5AA for ; Wed, 17 Mar 2021 03:49:08 +0000 (UTC) Received: from out30-43.freemail.mail.aliyun.com (out30-43.freemail.mail.aliyun.com [115.124.30.43]) (Using TLS) by relay.mimecast.com with ESMTP id us-mta-546-uDBmwH8POtaefwUNGR_TJw-1; Tue, 16 Mar 2021 23:49:05 -0400 X-MC-Unique: uDBmwH8POtaefwUNGR_TJw-1 X-Alimail-AntiSpam: AC=PASS; BC=-1|-1; BR=01201311R591e4; CH=green; DM=||false|; DS=||; FP=0|-1|-1|-1|0|-1|-1|-1; HT=e01e04400; MF=jefflexu@linux.alibaba.com; NM=1; PH=DS; RN=6; SR=0; TI=SMTPD_---0USDZqg4_1615952940 Received: from admindeMacBook-Pro-2.local(mailfrom:jefflexu@linux.alibaba.com fp:SMTPD_---0USDZqg4_1615952940) by smtp.aliyun-inc.com(127.0.0.1); Wed, 17 Mar 2021 11:49:00 +0800 From: JeffleXu To: Ming Lei References: <20210316031523.864506-1-ming.lei@redhat.com> <20210316031523.864506-9-ming.lei@redhat.com> Message-ID: Date: Wed, 17 Mar 2021 11:49:00 +0800 User-Agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.15; rv:78.0) Gecko/20100101 Thunderbird/78.7.0 MIME-Version: 1.0 In-Reply-To: X-Mimecast-Impersonation-Protect: Policy=CLT - Impersonation Protection Definition; Similar Internal Domain=false; Similar Monitored External Domain=false; Custom External Domain=false; Mimecast External Domain=false; Newly Observed Domain=false; Internal User Name=false; Custom Display Name List=false; Reply-to Address Mismatch=false; Targeted Threat Dictionary=false; Mimecast Threat Dictionary=false; Custom Threat Dictionary=false X-Scanned-By: MIMEDefang 2.78 on 10.11.54.3 X-loop: dm-devel@redhat.com X-Mailman-Approved-At: Thu, 18 Mar 2021 05:14:45 -0400 Cc: Jens Axboe , linux-block@vger.kernel.org, dm-devel@redhat.com, Christoph Hellwig , Mike Snitzer Subject: Re: [dm-devel] [RFC PATCH 08/11] block: use per-task poll context to implement bio based io poll X-BeenThere: dm-devel@redhat.com X-Mailman-Version: 2.1.12 Precedence: junk List-Id: device-mapper development List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Sender: dm-devel-bounces@redhat.com Errors-To: dm-devel-bounces@redhat.com X-Scanned-By: MIMEDefang 2.79 on 10.5.11.15 Authentication-Results: relay.mimecast.com; auth=pass smtp.auth=CUSA124A263 smtp.mailfrom=dm-devel-bounces@redhat.com X-Mimecast-Spam-Score: 0 X-Mimecast-Originator: redhat.com Content-Language: en-US Content-Type: text/plain; charset="us-ascii" Content-Transfer-Encoding: 7bit On 3/16/21 7:00 PM, JeffleXu wrote: > > > On 3/16/21 3:17 PM, Ming Lei wrote: >> On Tue, Mar 16, 2021 at 02:46:08PM +0800, JeffleXu wrote: >>> It is a giant progress to gather all split bios that need to be polled >>> in a per-task queue. Still some comments below. >>> >>> >>> On 3/16/21 11:15 AM, Ming Lei wrote: >>>> Currently bio based IO poll needs to poll all hw queue blindly, this way >>>> is very inefficient, and the big reason is that we can't pass bio >>>> submission result to io poll task. >>>> >>>> In IO submission context, store associated underlying bios into the >>>> submission queue and save 'cookie' poll data in bio->bi_iter.bi_private_data, >>>> and return current->pid to caller of submit_bio() for any DM or bio based >>>> driver's IO, which is submitted from FS. >>>> >>>> In IO poll context, the passed cookie tells us the PID of submission >>>> context, and we can find the bio from that submission context. Moving >>>> bio from submission queue to poll queue of the poll context, and keep >>>> polling until these bios are ended. Remove bio from poll queue if the >>>> bio is ended. Add BIO_DONE and BIO_END_BY_POLL for such purpose. >>>> >>>> Usually submission shares context with io poll. The per-task poll context >>>> is just like stack variable, and it is cheap to move data between the two >>>> per-task queues. >>>> >>>> Signed-off-by: Ming Lei >>>> --- >>>> block/bio.c | 5 ++ >>>> block/blk-core.c | 74 +++++++++++++++++- >>>> block/blk-mq.c | 156 +++++++++++++++++++++++++++++++++++++- >>>> include/linux/blk_types.h | 3 + >>>> 4 files changed, 235 insertions(+), 3 deletions(-) >>>> >>>> diff --git a/block/bio.c b/block/bio.c >>>> index a1c4d2900c7a..bcf5eca0e8e3 100644 >>>> --- a/block/bio.c >>>> +++ b/block/bio.c >>>> @@ -1402,6 +1402,11 @@ static inline bool bio_remaining_done(struct bio *bio) >>>> **/ >>>> void bio_endio(struct bio *bio) >>>> { >>>> + /* BIO_END_BY_POLL has to be set before calling submit_bio */ >>>> + if (bio_flagged(bio, BIO_END_BY_POLL)) { >>>> + bio_set_flag(bio, BIO_DONE); >>>> + return; >>>> + } >>>> again: >>>> if (!bio_remaining_done(bio)) >>>> return; >>>> diff --git a/block/blk-core.c b/block/blk-core.c >>>> index a082bbc856fb..970b23fa2e6e 100644 >>>> --- a/block/blk-core.c >>>> +++ b/block/blk-core.c >>>> @@ -854,6 +854,40 @@ static inline void blk_bio_poll_preprocess(struct request_queue *q, >>>> bio->bi_opf |= REQ_TAG; >>>> } >>>> >>>> +static bool blk_bio_poll_prep_submit(struct io_context *ioc, struct bio *bio) >>>> +{ >>>> + struct blk_bio_poll_data data = { >>>> + .bio = bio, >>>> + }; >>>> + struct blk_bio_poll_ctx *pc = ioc->data; >>>> + unsigned int queued; >>>> + >>>> + /* lock is required if there is more than one writer */ >>>> + if (unlikely(atomic_read(&ioc->nr_tasks) > 1)) { >>>> + spin_lock(&pc->lock); >>>> + queued = kfifo_put(&pc->sq, data); >>>> + spin_unlock(&pc->lock); >>>> + } else { >>>> + queued = kfifo_put(&pc->sq, data); >>>> + } >>>> + >>>> + /* >>>> + * Now the bio is added per-task fifo, mark it as END_BY_POLL, >>>> + * so we can save cookie into this bio after submit_bio(). >>>> + */ >>>> + if (queued) >>>> + bio_set_flag(bio, BIO_END_BY_POLL); >>>> + else >>>> + bio->bi_opf &= ~(REQ_HIPRI | REQ_TAG); >>>> + >>>> + return queued; >>>> +} >>> >>> The size of kfifo is limited, and it seems that once the sq of kfifio is >>> full, REQ_HIPRI flag is cleared and the corresponding bio is actually >>> enqueued into the default hw queue, which is IRQ driven. >> >> Yeah, this patch starts with 64 queue depth, and we can increase it to >> 128, which should cover most of cases. >> >>> >>> >>>> + >>>> +static void blk_bio_poll_post_submit(struct bio *bio, blk_qc_t cookie) >>>> +{ >>>> + bio->bi_iter.bi_private_data = cookie; >>>> +} >>>> + >>>> static noinline_for_stack bool submit_bio_checks(struct bio *bio) >>>> { >>>> struct block_device *bdev = bio->bi_bdev; >>>> @@ -1008,7 +1042,7 @@ static blk_qc_t __submit_bio(struct bio *bio) >>>> * bio_list_on_stack[1] contains bios that were submitted before the current >>>> * ->submit_bio_bio, but that haven't been processed yet. >>>> */ >>>> -static blk_qc_t __submit_bio_noacct(struct bio *bio) >>>> +static blk_qc_t __submit_bio_noacct_int(struct bio *bio, struct io_context *ioc) >>>> { >>>> struct bio_list bio_list_on_stack[2]; >>>> blk_qc_t ret = BLK_QC_T_NONE; >>>> @@ -1031,7 +1065,16 @@ static blk_qc_t __submit_bio_noacct(struct bio *bio) >>>> bio_list_on_stack[1] = bio_list_on_stack[0]; >>>> bio_list_init(&bio_list_on_stack[0]); >>>> >>>> - ret = __submit_bio(bio); >>>> + if (ioc && queue_is_mq(q) && >>>> + (bio->bi_opf & (REQ_HIPRI | REQ_TAG))) { >>>> + bool queued = blk_bio_poll_prep_submit(ioc, bio); >>>> + >>>> + ret = __submit_bio(bio); >>>> + if (queued) >>>> + blk_bio_poll_post_submit(bio, ret); >>>> + } else { >>>> + ret = __submit_bio(bio); >>>> + } >>>> >>>> /* >>>> * Sort new bios into those for a lower level and those for the >>>> @@ -1057,6 +1100,33 @@ static blk_qc_t __submit_bio_noacct(struct bio *bio) >>>> return ret; >>>> } >>>> >>>> +static inline blk_qc_t __submit_bio_noacct_poll(struct bio *bio, >>>> + struct io_context *ioc) >>>> +{ >>>> + struct blk_bio_poll_ctx *pc = ioc->data; >>>> + int entries = kfifo_len(&pc->sq); >>>> + >>>> + __submit_bio_noacct_int(bio, ioc); >>>> + >>>> + /* bio submissions queued to per-task poll context */ >>>> + if (kfifo_len(&pc->sq) > entries) >>>> + return current->pid; >>>> + >>>> + /* swapper's pid is 0, but it can't submit poll IO for us */ >>>> + return 0; >>>> +} >>>> + >>>> +static inline blk_qc_t __submit_bio_noacct(struct bio *bio) >>>> +{ >>>> + struct io_context *ioc = current->io_context; >>>> + >>>> + if (ioc && ioc->data && (bio->bi_opf & REQ_HIPRI)) >>>> + return __submit_bio_noacct_poll(bio, ioc); >>>> + >>>> + return __submit_bio_noacct_int(bio, NULL); >>>> +} >>>> + >>>> + >>>> static blk_qc_t __submit_bio_noacct_mq(struct bio *bio) >>>> { >>>> struct bio_list bio_list[2] = { }; >>>> diff --git a/block/blk-mq.c b/block/blk-mq.c >>>> index 03f59915fe2c..4e6f1467d303 100644 >>>> --- a/block/blk-mq.c >>>> +++ b/block/blk-mq.c >>>> @@ -3865,14 +3865,168 @@ static inline int blk_mq_poll_hctx(struct request_queue *q, >>>> return ret; >>>> } >>>> >>>> +static blk_qc_t bio_get_poll_cookie(struct bio *bio) >>>> +{ >>>> + return bio->bi_iter.bi_private_data; >>>> +} >>>> + >>>> +static int blk_mq_poll_io(struct bio *bio) >>>> +{ >>>> + struct request_queue *q = bio->bi_bdev->bd_disk->queue; >>>> + blk_qc_t cookie = bio_get_poll_cookie(bio); >>>> + int ret = 0; >>>> + >>>> + if (!bio_flagged(bio, BIO_DONE) && blk_qc_t_valid(cookie)) { >>>> + struct blk_mq_hw_ctx *hctx = >>>> + q->queue_hw_ctx[blk_qc_t_to_queue_num(cookie)]; >>>> + >>>> + ret += blk_mq_poll_hctx(q, hctx); >>>> + } >>>> + return ret; >>>> +} >>>> + >>>> +static int blk_bio_poll_and_end_io(struct request_queue *q, >>>> + struct blk_bio_poll_ctx *poll_ctx) >>>> +{ >>>> + struct blk_bio_poll_data *poll_data = &poll_ctx->pq[0]; >>>> + int ret = 0; >>>> + int i; >>>> + >>>> + for (i = 0; i < BLK_BIO_POLL_PQ_SZ; i++) { >>>> + struct bio *bio = poll_data[i].bio; >>>> + >>>> + if (!bio) >>>> + continue; >>>> + >>>> + ret += blk_mq_poll_io(bio); >>>> + if (bio_flagged(bio, BIO_DONE)) { >>>> + poll_data[i].bio = NULL; >>>> + >>>> + /* clear BIO_END_BY_POLL and end me really */ >>>> + bio_clear_flag(bio, BIO_END_BY_POLL); >>>> + bio_endio(bio); >>>> + } >>>> + } >>>> + return ret; >>>> +} >>> >>> When there are multiple threads polling, saying thread A and thread B, >>> then there's one bio which should be polled by thread A (the pid is >>> passed to thread A), while it's actually completed by thread B. In this >>> case, when the bio is completed by thread B, the bio is not really >>> completed and one extra blk_poll() still needs to be called. >> >> When this happens, the dm bio can't be completed, and the associated >> kiocb can't be completed too, io_uring or other poll code context will >> keep calling blk_poll() by passing thread A's pid until this dm bio is >> done, since the dm bio is submitted from thread A. >> > > This will affect the multi-thread polling performance. I tested > dm-stripe, in which every bio will be split and enqueued into all > underlying devices, and thus amplify the interference between multiple > threads. > > Test Result: > IOPS: 332k (IRQ) -> 363k (iopoll), aka ~10% performance gain Sorry this performance drop is not related to this bio refcount issue here. Still it's due to the limited kfifo size. I did another through test on another machine (aarch64 with more nvme disks). - Without mentioned specifically, the configuration is 'iodepth=128, kfifo queue depth =128'. - The number before '->' indicates the IOPS in IRQ mode, i.e., 'hipri=0', while the number after '->' indicates the IOPS in polling mode, i.e., 'hipri=1'. ``` 3-threads dm-linear-3 targets (4k randread IOPS, unit K) 5.12-rc1: 667 leiming: 674 -> 849 ours 8353c1a: 623 -> 811 3-threads dm-stripe-3 targets (12k randread IOPS, unit K) 5.12-rc1: 321 leiming: 313 -> 349 leiming : 313 -> 409 (iodepth=32, kfifo queue depth =128) leiming : 314 -> 409 (iodepth=128, kfifo queue depth =512) ours 8353c1a: 310 -> 406 1-thread dm-linear-3 targets (4k randread IOPS, unit K) 5.12-rc1: 224 leiming: 218 -> 288 ours 8353c1a: 210 -> 280 1-threads dm-stripe-3 targets (12k randread IOPS, unit K) 5.12-rc1: 109 leiming: 107 -> 120 leiming : 107 -> 145 (iodepth=32, kfifo queue depth =128) leiming : 108 -> 145 (iodepth=128, kfifo queue depth =512) ours 8353c1a: 107 -> 146 ``` Some hints: 1. When configured as 'iodepth=128, kfifo queue depth =128', dm-stripe doesn't perform well in polling mode. It's because it's more likely that the original bio will be split into split bios in dm-stripe, and thus kfifo will be more likely used up in this case. So the size of kfifo need to be tuned according to iodepth and the IO load. Thus exporting the size of kfifo as a sysfs entry may be need in the following patch. 2. It indicates a performance drop of my patch in IRQ mode, compared to the original 5.12-rc1. I doubt maybe it's due to extra code mixed in blk-core, such as __submit_bio_noacct()... > > > Test Environment: > > nvme.poll_queues = 3 > > BLK_BIO_POLL_SQ_SZ = 128 > > dmsetup create testdev --table "0 629145600 striped 3 8 /dev/nvme0n1 0 > /dev/nvme1n1 0 /dev/nvme4n1 0" > > > ``` > $cat fio.conf > [global] > name=iouring-sqpoll-iopoll-1 > ioengine=io_uring > iodepth=128 > numjobs=1 > thread > rw=randread > direct=1 > hipri=1 > runtime=10 > time_based > group_reporting > randrepeat=0 > filename=/dev/mapper/testdev > bs=12k > > [job-1] > cpus_allowed=14 > > [job-2] > cpus_allowed=16 > > [job-3] > cpus_allowed=84 > ``` > -- Thanks, Jeffle -- dm-devel mailing list dm-devel@redhat.com https://listman.redhat.com/mailman/listinfo/dm-devel