From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-15.8 required=3.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,FREEMAIL_FORGED_FROMDOMAIN,FREEMAIL_FROM, HEADER_FROM_DIFFERENT_DOMAINS,INCLUDES_CR_TRAILER,INCLUDES_PATCH, MAILING_LIST_MULTI,USER_AGENT_GIT autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 062FEC001D5 for ; Thu, 1 Apr 2021 17:45:54 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id 48ADF613CC for ; Thu, 1 Apr 2021 17:45:50 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S236250AbhDARoM (ORCPT ); Thu, 1 Apr 2021 13:44:12 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:57240 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S234569AbhDARiF (ORCPT ); Thu, 1 Apr 2021 13:38:05 -0400 Received: from mail-wm1-x331.google.com (mail-wm1-x331.google.com [IPv6:2a00:1450:4864:20::331]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 86FD9C0045F2 for ; Thu, 1 Apr 2021 07:48:17 -0700 (PDT) Received: by mail-wm1-x331.google.com with SMTP id j20-20020a05600c1914b029010f31e15a7fso2995479wmq.1 for ; Thu, 01 Apr 2021 07:48:17 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20161025; h=from:to:subject:date:message-id:in-reply-to:references:mime-version :content-transfer-encoding; bh=SElIx8WYfnaJOAOkAGWeSVpXCAGq8pQPhktAFTf4pvQ=; b=bhZn7on+oMNmQSXAV2jJKFxZdtK+Ev9oNI5n49Hors0pC/U5q5DgTx4WhqWjrAamQt +B8aEJTNjBSaaoBqI2ZaORBsoVhJAZQzvTAsU84fh2+8aPi4F8Fmrujrc81Kce6F7ZiP VpwE5556/Nm0BzXtPRJC92ItZUz5hB/FL7rOkdA72IOaSyFh9eBXUwZwBmdSB/1txc/H SG0gT3C8nsy3Z9qEdZZ6CY6Gpch8kj9pKAt7zt1iL1WJGnVWNCF+4dhYD1gjzfNXypus AJcBo0f6dyiG7coPYt3SQ8f0ixpogDX7gR7ujmgBiPoucllplDZbLP+pjFqsTO17JPPG IJJg== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:from:to:subject:date:message-id:in-reply-to :references:mime-version:content-transfer-encoding; bh=SElIx8WYfnaJOAOkAGWeSVpXCAGq8pQPhktAFTf4pvQ=; b=nmSuLlQ1Ot7U1P54cpT44dEv6ZBiAXFM0Qpd/D8gWkl/cewOlsNStcRL7dyQg5tyC9 UvuaYf3yogHrJlUOIRR3XW33/rLPCC4Dmi21jQ5fjJnH5ibfM4JA8TGk02c7agf/kQXx vrLiAh19uKbAdNe3tIt1ye+O2dR8ROjMCyDy90UdXMk3n+Vkvy7cHygy6W5w8c9ieeKo Xq/kT8vEK6NdaaOy1Y41X/Ad0KdGNuSlMDtbhGiqddrACxR3UdTP4AOONYfd9LQyqkQ6 nJ0i0hC/lLZLtpwKIfDETxaIeqhT/4WO/ZtMoVoMPO9KR9/OcqZjDP/9jejVzIvQVLQL IFeg== X-Gm-Message-State: AOAM533Oq8oG17YTWxqGJeXFyXZvgEI3PUbP40gcQbtOEE3OjaVv2S3E JWAg6LvCNGOQdINLFrSqPYA= X-Google-Smtp-Source: ABdhPJx4e7MiP8OGI6m5oRf6jsY06gq91UgB7JsKPV2LXBwRNNlxQRUur3d2cAZNSXIt5nQjqc2s+g== X-Received: by 2002:a05:600c:2197:: with SMTP id e23mr8275371wme.39.1617288496213; Thu, 01 Apr 2021 07:48:16 -0700 (PDT) Received: from localhost.localdomain ([148.252.132.152]) by smtp.gmail.com with ESMTPSA id x13sm8183948wmp.39.2021.04.01.07.48.15 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Thu, 01 Apr 2021 07:48:15 -0700 (PDT) From: Pavel Begunkov To: Jens Axboe , io-uring@vger.kernel.org Subject: [PATCH v4 01/26] io_uring: name rsrc bits consistently Date: Thu, 1 Apr 2021 15:43:40 +0100 Message-Id: <962f5acdf810f3a62831e65da3932cde24f6d9df.1617287883.git.asml.silence@gmail.com> X-Mailer: git-send-email 2.24.0 In-Reply-To: References: MIME-Version: 1.0 Content-Transfer-Encoding: 8bit Precedence: bulk List-ID: X-Mailing-List: io-uring@vger.kernel.org Keep resource related structs' and functions' naming consistent, in particular use "io_rsrc" prefix for everything. Signed-off-by: Pavel Begunkov --- fs/io_uring.c | 150 ++++++++++++++++++++++++-------------------------- 1 file changed, 71 insertions(+), 79 deletions(-) diff --git a/fs/io_uring.c b/fs/io_uring.c index 8ad6ab708083..5dfd33753471 100644 --- a/fs/io_uring.c +++ b/fs/io_uring.c @@ -219,22 +219,22 @@ struct fixed_rsrc_table { struct file **files; }; -struct fixed_rsrc_ref_node { +struct io_rsrc_node { struct percpu_ref refs; struct list_head node; struct list_head rsrc_list; - struct fixed_rsrc_data *rsrc_data; + struct io_rsrc_data *rsrc_data; void (*rsrc_put)(struct io_ring_ctx *ctx, struct io_rsrc_put *prsrc); struct llist_node llist; bool done; }; -struct fixed_rsrc_data { +struct io_rsrc_data { struct fixed_rsrc_table *table; struct io_ring_ctx *ctx; - struct fixed_rsrc_ref_node *node; + struct io_rsrc_node *node; struct percpu_ref refs; struct completion done; bool quiesce; @@ -393,7 +393,7 @@ struct io_ring_ctx { * readers must ensure that ->refs is alive as long as the file* is * used. Only updated through io_uring_register(2). */ - struct fixed_rsrc_data *file_data; + struct io_rsrc_data *file_data; unsigned nr_user_files; /* if used, fixed mapped user buffers */ @@ -447,7 +447,7 @@ struct io_ring_ctx { struct llist_head rsrc_put_llist; struct list_head rsrc_ref_list; spinlock_t rsrc_ref_lock; - struct fixed_rsrc_ref_node *rsrc_backup_node; + struct io_rsrc_node *rsrc_backup_node; struct io_restriction restrictions; @@ -1024,9 +1024,8 @@ static void io_uring_try_cancel_requests(struct io_ring_ctx *ctx, struct task_struct *task, struct files_struct *files); static void io_uring_cancel_sqpoll(struct io_ring_ctx *ctx); -static void destroy_fixed_rsrc_ref_node(struct fixed_rsrc_ref_node *ref_node); -static struct fixed_rsrc_ref_node *alloc_fixed_rsrc_ref_node( - struct io_ring_ctx *ctx); +static void io_rsrc_node_destroy(struct io_rsrc_node *ref_node); +static struct io_rsrc_node *io_rsrc_node_alloc(struct io_ring_ctx *ctx); static void io_ring_file_put(struct io_ring_ctx *ctx, struct io_rsrc_put *prsrc); static void io_cqring_fill_event(struct io_kiocb *req, long res); @@ -1072,7 +1071,7 @@ EXPORT_SYMBOL(io_uring_get_socket); #define io_for_each_link(pos, head) \ for (pos = (head); pos; pos = pos->link) -static inline void io_set_resource_node(struct io_kiocb *req) +static inline void io_req_set_rsrc_node(struct io_kiocb *req) { struct io_ring_ctx *ctx = req->ctx; @@ -6272,7 +6271,7 @@ static void io_wq_submit_work(struct io_wq_work *work) #endif #define FFS_MASK ~(FFS_ASYNC_READ|FFS_ASYNC_WRITE|FFS_ISREG) -static inline struct file **io_fixed_file_slot(struct fixed_rsrc_data *file_data, +static inline struct file **io_fixed_file_slot(struct io_rsrc_data *file_data, unsigned i) { struct fixed_rsrc_table *table; @@ -6306,7 +6305,7 @@ static struct file *io_file_get(struct io_submit_state *state, file_ptr &= ~FFS_MASK; /* mask in overlapping REQ_F and FFS bits */ req->flags |= (file_ptr << REQ_F_ASYNC_READ_BIT); - io_set_resource_node(req); + io_req_set_rsrc_node(req); } else { trace_io_uring_file_get(ctx, fd); file = __io_file_get(state, fd); @@ -7060,9 +7059,8 @@ static void __io_sqe_files_unregister(struct io_ring_ctx *ctx) static void io_rsrc_data_ref_zero(struct percpu_ref *ref) { - struct fixed_rsrc_data *data; + struct io_rsrc_data *data = container_of(ref, struct io_rsrc_data, refs); - data = container_of(ref, struct fixed_rsrc_data, refs); complete(&data->done); } @@ -7076,20 +7074,20 @@ static inline void io_rsrc_ref_unlock(struct io_ring_ctx *ctx) spin_unlock_bh(&ctx->rsrc_ref_lock); } -static void io_sqe_rsrc_set_node(struct io_ring_ctx *ctx, - struct fixed_rsrc_data *rsrc_data, - struct fixed_rsrc_ref_node *ref_node) +static void io_rsrc_node_set(struct io_ring_ctx *ctx, + struct io_rsrc_data *rsrc_data, + struct io_rsrc_node *rsrc_node) { io_rsrc_ref_lock(ctx); - rsrc_data->node = ref_node; - list_add_tail(&ref_node->node, &ctx->rsrc_ref_list); + rsrc_data->node = rsrc_node; + list_add_tail(&rsrc_node->node, &ctx->rsrc_ref_list); io_rsrc_ref_unlock(ctx); percpu_ref_get(&rsrc_data->refs); } -static void io_sqe_rsrc_kill_node(struct io_ring_ctx *ctx, struct fixed_rsrc_data *data) +static void io_rsrc_node_kill(struct io_ring_ctx *ctx, struct io_rsrc_data *data) { - struct fixed_rsrc_ref_node *ref_node = NULL; + struct io_rsrc_node *ref_node = NULL; io_rsrc_ref_lock(ctx); ref_node = data->node; @@ -7099,21 +7097,21 @@ static void io_sqe_rsrc_kill_node(struct io_ring_ctx *ctx, struct fixed_rsrc_dat percpu_ref_kill(&ref_node->refs); } -static int io_rsrc_refnode_prealloc(struct io_ring_ctx *ctx) +static int io_rsrc_node_prealloc(struct io_ring_ctx *ctx) { if (ctx->rsrc_backup_node) return 0; - ctx->rsrc_backup_node = alloc_fixed_rsrc_ref_node(ctx); + ctx->rsrc_backup_node = io_rsrc_node_alloc(ctx); return ctx->rsrc_backup_node ? 0 : -ENOMEM; } -static struct fixed_rsrc_ref_node * -io_rsrc_refnode_get(struct io_ring_ctx *ctx, - struct fixed_rsrc_data *rsrc_data, - void (*rsrc_put)(struct io_ring_ctx *ctx, - struct io_rsrc_put *prsrc)) +static struct io_rsrc_node * +io_rsrc_node_get(struct io_ring_ctx *ctx, + struct io_rsrc_data *rsrc_data, + void (*rsrc_put)(struct io_ring_ctx *ctx, + struct io_rsrc_put *prsrc)) { - struct fixed_rsrc_ref_node *node = ctx->rsrc_backup_node; + struct io_rsrc_node *node = ctx->rsrc_backup_node; WARN_ON_ONCE(!node); @@ -7123,12 +7121,12 @@ io_rsrc_refnode_get(struct io_ring_ctx *ctx, return node; } -static int io_rsrc_ref_quiesce(struct fixed_rsrc_data *data, +static int io_rsrc_ref_quiesce(struct io_rsrc_data *data, struct io_ring_ctx *ctx, void (*rsrc_put)(struct io_ring_ctx *ctx, struct io_rsrc_put *prsrc)) { - struct fixed_rsrc_ref_node *node; + struct io_rsrc_node *node; int ret; if (data->quiesce) @@ -7136,10 +7134,10 @@ static int io_rsrc_ref_quiesce(struct fixed_rsrc_data *data, data->quiesce = true; do { - ret = io_rsrc_refnode_prealloc(ctx); + ret = io_rsrc_node_prealloc(ctx); if (ret) break; - io_sqe_rsrc_kill_node(ctx, data); + io_rsrc_node_kill(ctx, data); percpu_ref_kill(&data->refs); flush_delayed_work(&ctx->rsrc_put_work); @@ -7148,8 +7146,8 @@ static int io_rsrc_ref_quiesce(struct fixed_rsrc_data *data, break; percpu_ref_resurrect(&data->refs); - node = io_rsrc_refnode_get(ctx, data, rsrc_put); - io_sqe_rsrc_set_node(ctx, data, node); + node = io_rsrc_node_get(ctx, data, rsrc_put); + io_rsrc_node_set(ctx, data, node); reinit_completion(&data->done); mutex_unlock(&ctx->uring_lock); @@ -7161,9 +7159,9 @@ static int io_rsrc_ref_quiesce(struct fixed_rsrc_data *data, return ret; } -static struct fixed_rsrc_data *alloc_fixed_rsrc_data(struct io_ring_ctx *ctx) +static struct io_rsrc_data *io_rsrc_data_alloc(struct io_ring_ctx *ctx) { - struct fixed_rsrc_data *data; + struct io_rsrc_data *data; data = kzalloc(sizeof(*data), GFP_KERNEL); if (!data) @@ -7179,7 +7177,7 @@ static struct fixed_rsrc_data *alloc_fixed_rsrc_data(struct io_ring_ctx *ctx) return data; } -static void free_fixed_rsrc_data(struct fixed_rsrc_data *data) +static void io_rsrc_data_free(struct io_rsrc_data *data) { percpu_ref_exit(&data->refs); kfree(data->table); @@ -7188,7 +7186,7 @@ static void free_fixed_rsrc_data(struct fixed_rsrc_data *data) static int io_sqe_files_unregister(struct io_ring_ctx *ctx) { - struct fixed_rsrc_data *data = ctx->file_data; + struct io_rsrc_data *data = ctx->file_data; unsigned nr_tables, i; int ret; @@ -7207,7 +7205,7 @@ static int io_sqe_files_unregister(struct io_ring_ctx *ctx) nr_tables = DIV_ROUND_UP(ctx->nr_user_files, IORING_MAX_FILES_TABLE); for (i = 0; i < nr_tables; i++) kfree(data->table[i].files); - free_fixed_rsrc_data(data); + io_rsrc_data_free(data); ctx->file_data = NULL; ctx->nr_user_files = 0; return 0; @@ -7436,7 +7434,7 @@ static int io_sqe_files_scm(struct io_ring_ctx *ctx) } #endif -static int io_sqe_alloc_file_tables(struct fixed_rsrc_data *file_data, +static int io_sqe_alloc_file_tables(struct io_rsrc_data *file_data, unsigned nr_tables, unsigned nr_files) { int i; @@ -7526,9 +7524,9 @@ static void io_ring_file_put(struct io_ring_ctx *ctx, struct io_rsrc_put *prsrc) #endif } -static void __io_rsrc_put_work(struct fixed_rsrc_ref_node *ref_node) +static void __io_rsrc_put_work(struct io_rsrc_node *ref_node) { - struct fixed_rsrc_data *rsrc_data = ref_node->rsrc_data; + struct io_rsrc_data *rsrc_data = ref_node->rsrc_data; struct io_ring_ctx *ctx = rsrc_data->ctx; struct io_rsrc_put *prsrc, *tmp; @@ -7552,10 +7550,10 @@ static void io_rsrc_put_work(struct work_struct *work) node = llist_del_all(&ctx->rsrc_put_llist); while (node) { - struct fixed_rsrc_ref_node *ref_node; + struct io_rsrc_node *ref_node; struct llist_node *next = node->next; - ref_node = llist_entry(node, struct fixed_rsrc_ref_node, llist); + ref_node = llist_entry(node, struct io_rsrc_node, llist); __io_rsrc_put_work(ref_node); node = next; } @@ -7563,27 +7561,23 @@ static void io_rsrc_put_work(struct work_struct *work) static void io_rsrc_node_ref_zero(struct percpu_ref *ref) { - struct fixed_rsrc_ref_node *ref_node; - struct fixed_rsrc_data *data; - struct io_ring_ctx *ctx; + struct io_rsrc_node *node = container_of(ref, struct io_rsrc_node, refs); + struct io_rsrc_data *data = node->rsrc_data; + struct io_ring_ctx *ctx = data->ctx; bool first_add = false; int delay = HZ; - ref_node = container_of(ref, struct fixed_rsrc_ref_node, refs); - data = ref_node->rsrc_data; - ctx = data->ctx; - io_rsrc_ref_lock(ctx); - ref_node->done = true; + node->done = true; while (!list_empty(&ctx->rsrc_ref_list)) { - ref_node = list_first_entry(&ctx->rsrc_ref_list, - struct fixed_rsrc_ref_node, node); + node = list_first_entry(&ctx->rsrc_ref_list, + struct io_rsrc_node, node); /* recycle ref nodes in order */ - if (!ref_node->done) + if (!node->done) break; - list_del(&ref_node->node); - first_add |= llist_add(&ref_node->llist, &ctx->rsrc_put_llist); + list_del(&node->node); + first_add |= llist_add(&node->llist, &ctx->rsrc_put_llist); } io_rsrc_ref_unlock(ctx); @@ -7596,10 +7590,9 @@ static void io_rsrc_node_ref_zero(struct percpu_ref *ref) queue_delayed_work(system_wq, &ctx->rsrc_put_work, delay); } -static struct fixed_rsrc_ref_node *alloc_fixed_rsrc_ref_node( - struct io_ring_ctx *ctx) +static struct io_rsrc_node *io_rsrc_node_alloc(struct io_ring_ctx *ctx) { - struct fixed_rsrc_ref_node *ref_node; + struct io_rsrc_node *ref_node; ref_node = kzalloc(sizeof(*ref_node), GFP_KERNEL); if (!ref_node) @@ -7617,19 +7610,18 @@ static struct fixed_rsrc_ref_node *alloc_fixed_rsrc_ref_node( } static void init_fixed_file_ref_node(struct io_ring_ctx *ctx, - struct fixed_rsrc_ref_node *ref_node) + struct io_rsrc_node *ref_node) { ref_node->rsrc_data = ctx->file_data; ref_node->rsrc_put = io_ring_file_put; } -static void destroy_fixed_rsrc_ref_node(struct fixed_rsrc_ref_node *ref_node) +static void io_rsrc_node_destroy(struct io_rsrc_node *ref_node) { percpu_ref_exit(&ref_node->refs); kfree(ref_node); } - static int io_sqe_files_register(struct io_ring_ctx *ctx, void __user *arg, unsigned nr_args) { @@ -7637,8 +7629,8 @@ static int io_sqe_files_register(struct io_ring_ctx *ctx, void __user *arg, unsigned nr_tables, i; struct file *file; int fd, ret = -ENOMEM; - struct fixed_rsrc_ref_node *ref_node; - struct fixed_rsrc_data *file_data; + struct io_rsrc_node *ref_node; + struct io_rsrc_data *file_data; if (ctx->file_data) return -EBUSY; @@ -7647,7 +7639,7 @@ static int io_sqe_files_register(struct io_ring_ctx *ctx, void __user *arg, if (nr_args > IORING_MAX_FIXED_FILES) return -EMFILE; - file_data = alloc_fixed_rsrc_data(ctx); + file_data = io_rsrc_data_alloc(ctx); if (!file_data) return -ENOMEM; ctx->file_data = file_data; @@ -7704,14 +7696,14 @@ static int io_sqe_files_register(struct io_ring_ctx *ctx, void __user *arg, return ret; } - ref_node = alloc_fixed_rsrc_ref_node(ctx); + ref_node = io_rsrc_node_alloc(ctx); if (!ref_node) { io_sqe_files_unregister(ctx); return -ENOMEM; } init_fixed_file_ref_node(ctx, ref_node); - io_sqe_rsrc_set_node(ctx, file_data, ref_node); + io_rsrc_node_set(ctx, file_data, ref_node); return ret; out_fput: for (i = 0; i < ctx->nr_user_files; i++) { @@ -7723,7 +7715,7 @@ static int io_sqe_files_register(struct io_ring_ctx *ctx, void __user *arg, kfree(file_data->table[i].files); ctx->nr_user_files = 0; out_free: - free_fixed_rsrc_data(ctx->file_data); + io_rsrc_data_free(ctx->file_data); ctx->file_data = NULL; return ret; } @@ -7771,10 +7763,10 @@ static int io_sqe_file_register(struct io_ring_ctx *ctx, struct file *file, #endif } -static int io_queue_rsrc_removal(struct fixed_rsrc_data *data, void *rsrc) +static int io_queue_rsrc_removal(struct io_rsrc_data *data, void *rsrc) { struct io_rsrc_put *prsrc; - struct fixed_rsrc_ref_node *ref_node = data->node; + struct io_rsrc_node *ref_node = data->node; prsrc = kzalloc(sizeof(*prsrc), GFP_KERNEL); if (!prsrc) @@ -7786,7 +7778,7 @@ static int io_queue_rsrc_removal(struct fixed_rsrc_data *data, void *rsrc) return 0; } -static inline int io_queue_file_removal(struct fixed_rsrc_data *data, +static inline int io_queue_file_removal(struct io_rsrc_data *data, struct file *file) { return io_queue_rsrc_removal(data, (void *)file); @@ -7796,8 +7788,8 @@ static int __io_sqe_files_update(struct io_ring_ctx *ctx, struct io_uring_rsrc_update *up, unsigned nr_args) { - struct fixed_rsrc_data *data = ctx->file_data; - struct fixed_rsrc_ref_node *ref_node; + struct io_rsrc_data *data = ctx->file_data; + struct io_rsrc_node *ref_node; struct file *file, **file_slot; __s32 __user *fds; int fd, i, err; @@ -7808,7 +7800,7 @@ static int __io_sqe_files_update(struct io_ring_ctx *ctx, return -EOVERFLOW; if (done > ctx->nr_user_files) return -EINVAL; - err = io_rsrc_refnode_prealloc(ctx); + err = io_rsrc_node_prealloc(ctx); if (err) return err; @@ -7864,8 +7856,8 @@ static int __io_sqe_files_update(struct io_ring_ctx *ctx, if (needs_switch) { percpu_ref_kill(&data->node->refs); - ref_node = io_rsrc_refnode_get(ctx, data, io_ring_file_put); - io_sqe_rsrc_set_node(ctx, data, ref_node); + ref_node = io_rsrc_node_get(ctx, data, io_ring_file_put); + io_rsrc_node_set(ctx, data, ref_node); } return done ? done : err; } @@ -8541,7 +8533,7 @@ static void io_ring_ctx_free(struct io_ring_ctx *ctx) io_destroy_buffers(ctx); if (ctx->rsrc_backup_node) - destroy_fixed_rsrc_ref_node(ctx->rsrc_backup_node); + io_rsrc_node_destroy(ctx->rsrc_backup_node); #if defined(CONFIG_UNIX) if (ctx->ring_sock) { -- 2.24.0