From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-16.4 required=3.0 tests=BAYES_00,DKIMWL_WL_HIGH, DKIM_SIGNED,DKIM_VALID,DKIM_VALID_AU,HEADER_FROM_DIFFERENT_DOMAINS, INCLUDES_CR_TRAILER,INCLUDES_PATCH,MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 038E0C4338F for ; Fri, 6 Aug 2021 02:34:57 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id DB236611C2 for ; Fri, 6 Aug 2021 02:34:56 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S242395AbhHFCfK (ORCPT ); Thu, 5 Aug 2021 22:35:10 -0400 Received: from us-smtp-delivery-124.mimecast.com ([216.205.24.124]:54661 "EHLO us-smtp-delivery-124.mimecast.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S242433AbhHFCfH (ORCPT ); Thu, 5 Aug 2021 22:35:07 -0400 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1628217292; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=pBKFu8rMlBA6IQdDgtRI7aE84cYglZpHOGd90HePPTo=; b=KYxSRMCNgVqNu0faiKyFYvYG8ZvHjvsPAM24P5Rd+nsJvT+MVPtloUoGIRDSuRPNXzwbKd styz2DBl4JnF/+115RCy7p1aam0nxEJTuvXcNUUghH/jDoqZbjuvj6p+jFtJXf8obCTrzZ RpIzl/MrJz49zTKZ811O7Qc1GtAzmDU= Received: from mimecast-mx01.redhat.com (mimecast-mx01.redhat.com [209.132.183.4]) (Using TLS) by relay.mimecast.com with ESMTP id us-mta-33-fRw-QmrvPb6NO5aNdEmK_g-1; Thu, 05 Aug 2021 22:34:51 -0400 X-MC-Unique: fRw-QmrvPb6NO5aNdEmK_g-1 Received: from smtp.corp.redhat.com (int-mx07.intmail.prod.int.phx2.redhat.com [10.5.11.22]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mimecast-mx01.redhat.com (Postfix) with ESMTPS id BC45D192CC40; Fri, 6 Aug 2021 02:34:49 +0000 (UTC) Received: from localhost (ovpn-12-45.pek2.redhat.com [10.72.12.45]) by smtp.corp.redhat.com (Postfix) with ESMTP id C6B121036D11; Fri, 6 Aug 2021 02:34:45 +0000 (UTC) From: Ming Lei To: Jens Axboe Cc: linux-kernel@vger.kernel.org, linux-block@vger.kernel.org, Christoph Hellwig , Dan Schatzberg , Ming Lei Subject: [PATCH V3 2/7] loop: clean up blkcg association Date: Fri, 6 Aug 2021 10:34:18 +0800 Message-Id: <20210806023423.131060-3-ming.lei@redhat.com> In-Reply-To: <20210806023423.131060-1-ming.lei@redhat.com> References: <20210806023423.131060-1-ming.lei@redhat.com> MIME-Version: 1.0 Content-Transfer-Encoding: 8bit X-Scanned-By: MIMEDefang 2.84 on 10.5.11.22 Precedence: bulk List-ID: X-Mailing-List: linux-block@vger.kernel.org Each loop_worker is responsible for running requests originated from same blkcg, so: 1) associate with kthread in the entry of loop_process_work(), and disassociate in the end of this function, then we can avoid to do both for each request. 2) remove ->blkcg_css and ->memcg_css from 'loop_cmd' since both are per loop_worker. Also kill #ifdef in the related functions. Acked-by: Dan Schatzberg Signed-off-by: Ming Lei --- drivers/block/loop.c | 81 +++++++++++++++++++++++--------------------- drivers/block/loop.h | 2 -- 2 files changed, 43 insertions(+), 40 deletions(-) diff --git a/drivers/block/loop.c b/drivers/block/loop.c index fa1c298a8cfb..e2b305d0c48d 100644 --- a/drivers/block/loop.c +++ b/drivers/block/loop.c @@ -989,23 +989,46 @@ static inline int queue_on_root_worker(struct cgroup_subsys_state *css) { return !css || css == blkcg_root_css; } +static struct cgroup_subsys_state *loop_rq_blkcg_css(struct loop_cmd *cmd) +{ + struct request *rq = blk_mq_rq_from_pdu(cmd); + /* always use the first bio's css */ + struct blkcg *blkcg = bio_blkcg(rq->bio); + + if (blkcg) + return &blkcg->css; + return NULL; +} #else static inline int queue_on_root_worker(struct cgroup_subsys_state *css) { return !css; } +static struct cgroup_subsys_state *loop_rq_blkcg_css(struct loop_cmd *cmd) +{ + return NULL; +} #endif +static struct cgroup_subsys_state *loop_rq_get_memcg_css( + struct cgroup_subsys_state *blkcg_css) +{ + if (blkcg_css) + return memcg_get_e_css(blkcg_css->cgroup); + return NULL; +} + static void loop_queue_work(struct loop_device *lo, struct loop_cmd *cmd) { struct rb_node **node = &(lo->worker_tree.rb_node), *parent = NULL; struct loop_worker *cur_worker, *worker = NULL; struct work_struct *work; struct list_head *cmd_list; + struct cgroup_subsys_state *blkcg_css = loop_rq_blkcg_css(cmd); spin_lock_irq(&lo->lo_work_lock); - if (queue_on_root_worker(cmd->blkcg_css)) + if (queue_on_root_worker(blkcg_css)) goto queue_work; node = &lo->worker_tree.rb_node; @@ -1013,10 +1036,10 @@ static void loop_queue_work(struct loop_device *lo, struct loop_cmd *cmd) while (*node) { parent = *node; cur_worker = container_of(*node, struct loop_worker, rb_node); - if (cur_worker->blkcg_css == cmd->blkcg_css) { + if (cur_worker->blkcg_css == blkcg_css) { worker = cur_worker; break; - } else if ((long)cur_worker->blkcg_css < (long)cmd->blkcg_css) { + } else if ((long)cur_worker->blkcg_css < (long)blkcg_css) { node = &(*node)->rb_left; } else { node = &(*node)->rb_right; @@ -1030,15 +1053,10 @@ static void loop_queue_work(struct loop_device *lo, struct loop_cmd *cmd) * In the event we cannot allocate a worker, just queue on the * rootcg worker and issue the I/O as the rootcg */ - if (!worker) { - cmd->blkcg_css = NULL; - if (cmd->memcg_css) - css_put(cmd->memcg_css); - cmd->memcg_css = NULL; + if (!worker) goto queue_work; - } - worker->blkcg_css = cmd->blkcg_css; + worker->blkcg_css = blkcg_css; css_get(worker->blkcg_css); INIT_WORK(&worker->work, loop_workfn); INIT_LIST_HEAD(&worker->cmd_list); @@ -2162,19 +2180,6 @@ static blk_status_t loop_queue_rq(struct blk_mq_hw_ctx *hctx, break; } - /* always use the first bio's css */ - cmd->blkcg_css = NULL; - cmd->memcg_css = NULL; -#ifdef CONFIG_BLK_CGROUP - if (rq->bio && rq->bio->bi_blkg) { - cmd->blkcg_css = &bio_blkcg(rq->bio)->css; -#ifdef CONFIG_MEMCG - cmd->memcg_css = - cgroup_get_e_css(cmd->blkcg_css->cgroup, - &memory_cgrp_subsys); -#endif - } -#endif loop_queue_work(lo, cmd); return BLK_STS_OK; @@ -2186,28 +2191,14 @@ static void loop_handle_cmd(struct loop_cmd *cmd) const bool write = op_is_write(req_op(rq)); struct loop_device *lo = rq->q->queuedata; int ret = 0; - struct mem_cgroup *old_memcg = NULL; if (write && (lo->lo_flags & LO_FLAGS_READ_ONLY)) { ret = -EIO; goto failed; } - if (cmd->blkcg_css) - kthread_associate_blkcg(cmd->blkcg_css); - if (cmd->memcg_css) - old_memcg = set_active_memcg( - mem_cgroup_from_css(cmd->memcg_css)); - ret = do_req_filebacked(lo, rq); - if (cmd->blkcg_css) - kthread_associate_blkcg(NULL); - - if (cmd->memcg_css) { - set_active_memcg(old_memcg); - css_put(cmd->memcg_css); - } failed: /* complete non-aio request */ if (!cmd->use_aio || ret) { @@ -2263,7 +2254,21 @@ static void loop_workfn(struct work_struct *work) { struct loop_worker *worker = container_of(work, struct loop_worker, work); - loop_process_work(worker, &worker->cmd_list, worker->lo); + struct mem_cgroup *old_memcg = NULL; + struct cgroup_subsys_state *memcg_css = NULL; + + kthread_associate_blkcg(worker->blkcg_css); + memcg_css = loop_rq_get_memcg_css(worker->blkcg_css); + if (memcg_css) { + old_memcg = set_active_memcg( + mem_cgroup_from_css(memcg_css)); + loop_process_work(worker, &worker->cmd_list, worker->lo); + set_active_memcg(old_memcg); + css_put(memcg_css); + } else { + loop_process_work(worker, &worker->cmd_list, worker->lo); + } + kthread_associate_blkcg(NULL); } static void loop_rootcg_workfn(struct work_struct *work) diff --git a/drivers/block/loop.h b/drivers/block/loop.h index 1988899db63a..a52a3fd89457 100644 --- a/drivers/block/loop.h +++ b/drivers/block/loop.h @@ -77,8 +77,6 @@ struct loop_cmd { long ret; struct kiocb iocb; struct bio_vec *bvec; - struct cgroup_subsys_state *blkcg_css; - struct cgroup_subsys_state *memcg_css; }; /* Support for loadable transfer modules */ -- 2.31.1