From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-15.4 required=3.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,HEADER_FROM_DIFFERENT_DOMAINS,INCLUDES_CR_TRAILER, INCLUDES_PATCH,MAILING_LIST_MULTI,MISSING_SUBJECT,SPF_HELO_NONE,SPF_PASS, URIBL_BLOCKED,USER_AGENT_SANE_1 autolearn=unavailable autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id B326CC64E7A for ; Tue, 1 Dec 2020 11:55:22 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id 4FCA42084C for ; Tue, 1 Dec 2020 11:55:22 +0000 (UTC) Authentication-Results: mail.kernel.org; dkim=pass (1024-bit key) header.d=sargun.me header.i=@sargun.me header.b="E/h4fMA+" Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1728294AbgLALzW (ORCPT ); Tue, 1 Dec 2020 06:55:22 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:50106 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1728263AbgLALzV (ORCPT ); Tue, 1 Dec 2020 06:55:21 -0500 Received: from mail-io1-xd34.google.com (mail-io1-xd34.google.com [IPv6:2607:f8b0:4864:20::d34]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 7FFD1C0613D4 for ; Tue, 1 Dec 2020 03:54:41 -0800 (PST) Received: by mail-io1-xd34.google.com with SMTP id q137so1276248iod.9 for ; Tue, 01 Dec 2020 03:54:41 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=sargun.me; s=google; h=date:from:to:cc:message-id:mime-version:content-disposition :user-agent; bh=7+fTf4fqXiwm8yOFyjgfaZN3jbaTriUpbBa9OV5cRKc=; b=E/h4fMA+2bro8w6bO14kcveBcnU55PQJbTzecEnm29p0Ex+wBliWfrj0Hc3bZI+IBp hWHyS4nsSKl7n5JcWZOl6TmHzmERtakJgvYaXW0mQvmr7Q0e00UesY6zi+cYZDr1atDd 00yIKo0T9bLclPhXeOBmUGJnK/fXIGY+CbIOo= X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:date:from:to:cc:message-id:mime-version :content-disposition:user-agent; bh=7+fTf4fqXiwm8yOFyjgfaZN3jbaTriUpbBa9OV5cRKc=; b=t0w+1drVMQ+5tGith73ahhac/GwNji0NOJSZxR8+YNnV25fmQ2qn3s6LYwdmaDtHvV ZJQP5KeRlYVtGfajwBIvCUV+OnCQbrRE2O9dgBkrvnDnKmfZPRUM94L2oIEhfDDyav81 6q+gsQjno1AIQqATqWV3+/G6wonONwQFGAr/XmTAV976yCSxnB/kf8TWCwsbps8GF29i cQyTOxeBSJm/UCGLtY2jwNcEXHphF2EOFKDPQ+cnH/OUXA/GlGEojqERTQjdS/eEe3dY U24V/KJ4yW+K4dnzy0KxU4WcC8WfT0j5DYTMricZ82lAgqCupDuKh1z7q5/4BeLvUKPl htgw== X-Gm-Message-State: AOAM533+GvWcuVoMxJOCt1/TpQSW22PSuuIqclT71MW7QMwGMaMH10j8 cRhYAu01CuQaY6IkmQvyhDhJgQ== X-Google-Smtp-Source: ABdhPJzS9gW1wrBxiMgJVaNqp/THljI13jwd+rspiXSW3D7BFCzGwUS8JE6N7meYpzNcJPc9HcUK+w== X-Received: by 2002:a02:8622:: with SMTP id e31mr2191644jai.88.1606823680697; Tue, 01 Dec 2020 03:54:40 -0800 (PST) Received: from ircssh-2.c.rugged-nimbus-611.internal (80.60.198.104.bc.googleusercontent.com. [104.198.60.80]) by smtp.gmail.com with ESMTPSA id q140sm679761iod.43.2020.12.01.03.54.39 (version=TLS1_2 cipher=ECDHE-ECDSA-CHACHA20-POLY1305 bits=256/256); Tue, 01 Dec 2020 03:54:40 -0800 (PST) Date: Tue, 1 Dec 2020 11:54:38 +0000 From: Sargun Dhillon To: Vivek Goyal Cc: linux-unionfs@vger.kernel.org, miklos@szeredi.hu, Alexander Viro , Amir Goldstein , Giuseppe Scrivano , Daniel J Walsh , linux-fsdevel@vger.kernel.org, David Howells , Jeff Layton Message-ID: <20201201115437.GB24837@ircssh-2.c.rugged-nimbus-611.internal> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline User-Agent: Mutt/1.9.4 (2018-02-28) Precedence: bulk List-ID: X-Mailing-List: linux-unionfs@vger.kernel.org Bcc: Subject: Re: [PATCH v2 4/4] overlay: Add rudimentary checking of writeback errseq on volatile remount Reply-To: In-Reply-To: <20201130193342.GD14328@redhat.com> On Mon, Nov 30, 2020 at 02:33:42PM -0500, Vivek Goyal wrote: > On Fri, Nov 27, 2020 at 01:20:58AM -0800, Sargun Dhillon wrote: > > Volatile remounts validate the following at the moment: > > * Has the module been reloaded / the system rebooted > > * Has the workdir been remounted > > > > This adds a new check for errors detected via the superblock's > > errseq_t. At mount time, the errseq_t is snapshotted to disk, > > and upon remount it's re-verified. This allows for kernel-level > > detection of errors without forcing userspace to perform a > > sync and allows for the hidden detection of writeback errors. > > > > Signed-off-by: Sargun Dhillon > > Cc: linux-fsdevel@vger.kernel.org > > Cc: linux-unionfs@vger.kernel.org > > Cc: Miklos Szeredi > > Cc: Amir Goldstein > > Cc: Vivek Goyal > > --- > > fs/overlayfs/overlayfs.h | 1 + > > fs/overlayfs/readdir.c | 6 ++++++ > > fs/overlayfs/super.c | 1 + > > 3 files changed, 8 insertions(+) > > > > diff --git a/fs/overlayfs/overlayfs.h b/fs/overlayfs/overlayfs.h > > index de694ee99d7c..e8a711953b64 100644 > > --- a/fs/overlayfs/overlayfs.h > > +++ b/fs/overlayfs/overlayfs.h > > @@ -85,6 +85,7 @@ struct ovl_volatile_info { > > */ > > uuid_t ovl_boot_id; /* Must stay first member */ > > u64 s_instance_id; > > + errseq_t errseq; /* Implemented as a u32 */ > > } __packed; > > > > /* > > diff --git a/fs/overlayfs/readdir.c b/fs/overlayfs/readdir.c > > index 7b66fbb20261..5795b28bb4cf 100644 > > --- a/fs/overlayfs/readdir.c > > +++ b/fs/overlayfs/readdir.c > > @@ -1117,6 +1117,12 @@ static int ovl_verify_volatile_info(struct ovl_fs *ofs, > > return -EINVAL; > > } > > > > + err = errseq_check(&volatiledir->d_sb->s_wb_err, info.errseq); > > Might be a stupid question. Will ask anyway. > > But what protects against wrapping of counter. IOW, Say we stored info.errseq > value as A. It is possible that bunch of errors occurred and at remount > time ->s_wb_err is back to A and we pass the check. (Despite the fact lots > of errors have occurred since we sampled). > > Thanks > Vivek > +Jeff Layton Nothing. The current errseq API works like this today where if you have 2^20 (1048576) errors, and syncfs (or other calls that mark the errseq as seen), and the error that occured 1048575 times ago was the same error as you just last had, and the error on the upperdir has already been marked as seen, the error will be swallowed up silently. This exists throughout all of VFS. I think we're potentially making this more likely by checkpointing to disk. The one aspect which is a little different about the usecase in the patch is that it relies on this mechanism to determine if an error has occured after the entire FS was constructed, so it's somewhat more consequential than the current issue in VFS which will just bubble up errors in a few files. On my system syncfs takes about 2 milliseconds, so you have a chance to experience this every ~30 minutes if the syscalls align in the right way. If we expanded the errseq_t to u64, we would potentially get a collision every 4503599627370496 calls, or assuming the 2 millisecond invariant holds, every 285 years. Now, we probably don't want to make errseq_t into a u64 because of performance reasons (not all systems have native u64 cmpxchg), and the extra memory it'd take up. If we really want to avoid this case, I can think of one "simple" solution, which is something like laying out errseq_t as something like a errseq_t_src that's 64-bits, and all readers just look at the lower 32-bits. The longer errseq_t would exist on super_blocks, but files would still get the shorter one. To potentially avoid the performance penalty of atomic longs, we could also do something like this: typedef struct { atomic_t overflow; u32 errseq; } errseq_t_big; And in errseq_set, do: /* Wraps */ if (new < old) atomic_inc(&eseq->overflow); *shrug* I don't think that the above scenario is likely though.