From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-8.2 required=3.0 tests=HEADER_FROM_DIFFERENT_DOMAINS, INCLUDES_PATCH,MAILING_LIST_MULTI,SIGNED_OFF_BY,SPF_PASS,URIBL_BLOCKED, USER_AGENT_MUTT autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 5841BC65BA7 for ; Fri, 5 Oct 2018 05:28:17 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [209.132.180.67]) by mail.kernel.org (Postfix) with ESMTP id 26CC5208E7 for ; Fri, 5 Oct 2018 05:28:17 +0000 (UTC) DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org 26CC5208E7 Authentication-Results: mail.kernel.org; dmarc=none (p=none dis=none) header.from=fromorbit.com Authentication-Results: mail.kernel.org; spf=none smtp.mailfrom=linux-btrfs-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1727545AbeJEMZT (ORCPT ); Fri, 5 Oct 2018 08:25:19 -0400 Received: from ipmail07.adl2.internode.on.net ([150.101.137.131]:46499 "EHLO ipmail07.adl2.internode.on.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726939AbeJEMZT (ORCPT ); Fri, 5 Oct 2018 08:25:19 -0400 Received: from ppp59-167-129-252.static.internode.on.net (HELO dastard) ([59.167.129.252]) by ipmail07.adl2.internode.on.net with ESMTP; 05 Oct 2018 14:58:10 +0930 Received: from dave by dastard with local (Exim 4.80) (envelope-from ) id 1g8Ien-0003BU-DI; Fri, 05 Oct 2018 15:28:09 +1000 Date: Fri, 5 Oct 2018 15:28:09 +1000 From: Dave Chinner To: "Darrick J. Wong" Cc: linux-xfs@vger.kernel.org, linux-fsdevel@vger.kernel.org, linux-btrfs@vger.kernel.org, ocfs2-devel@oss.oracle.com, sandeen@redhat.com Subject: Re: [PATCH 02/15] xfs: refactor clonerange preparation into a separate helper Message-ID: <20181005052809.GB12041@dastard> References: <153870027422.29072.7433543674436957232.stgit@magnolia> <153870028762.29072.5369530877410002226.stgit@magnolia> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <153870028762.29072.5369530877410002226.stgit@magnolia> User-Agent: Mutt/1.5.21 (2010-09-15) Sender: linux-btrfs-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-btrfs@vger.kernel.org On Thu, Oct 04, 2018 at 05:44:47PM -0700, Darrick J. Wong wrote: > From: Darrick J. Wong > > Refactor all the reflink preparation steps into a separate helper that > we'll use to land all the upcoming fixes for insufficient input checks. If I've read the patch right, this also changes the location of the page cache truncation, right? i.e. it now happens in the xfs_reflink_remap_prep() function instead of after the remap? I think the commit message needs to mention that because it's a fix to incorrect behaviour.... I've added: -- This rework also moves the invalidation of the destination range to the prep function so that it is done before the range is remapped. This ensures that nobody can access the data in range being remapped until the remap is complete. -- Sound OK? Otherwise this looks fine. Reviewed-by: Dave Chinner -Dave. > > Signed-off-by: Darrick J. Wong > --- > fs/xfs/xfs_reflink.c | 96 +++++++++++++++++++++++++++++++++++++------------- > 1 file changed, 71 insertions(+), 25 deletions(-) > > > diff --git a/fs/xfs/xfs_reflink.c b/fs/xfs/xfs_reflink.c > index 38f405415b88..80ca9b6793cd 100644 > --- a/fs/xfs/xfs_reflink.c > +++ b/fs/xfs/xfs_reflink.c > @@ -1195,11 +1195,33 @@ xfs_iolock_two_inodes_and_break_layout( > return 0; > } > > +/* Unlock both inodes after they've been prepped for a range clone. */ > +STATIC void > +xfs_reflink_remap_unlock( > + struct file *file_in, > + struct file *file_out) > +{ > + struct inode *inode_in = file_inode(file_in); > + struct xfs_inode *src = XFS_I(inode_in); > + struct inode *inode_out = file_inode(file_out); > + struct xfs_inode *dest = XFS_I(inode_out); > + bool same_inode = (inode_in == inode_out); > + > + xfs_iunlock(dest, XFS_MMAPLOCK_EXCL); > + if (!same_inode) > + xfs_iunlock(src, XFS_MMAPLOCK_SHARED); > + inode_unlock(inode_out); > + if (!same_inode) > + inode_unlock_shared(inode_in); > +} > + > /* > - * Link a range of blocks from one file to another. > + * Prepare two files for range cloning. Upon a successful return both inodes > + * will have the iolock and mmaplock held, the page cache of the out file > + * will be truncated, and any leases on the out file will have been broken. > */ > -int > -xfs_reflink_remap_range( > +STATIC int > +xfs_reflink_remap_prep( > struct file *file_in, > loff_t pos_in, > struct file *file_out, > @@ -1211,19 +1233,9 @@ xfs_reflink_remap_range( > struct xfs_inode *src = XFS_I(inode_in); > struct inode *inode_out = file_inode(file_out); > struct xfs_inode *dest = XFS_I(inode_out); > - struct xfs_mount *mp = src->i_mount; > bool same_inode = (inode_in == inode_out); > - xfs_fileoff_t sfsbno, dfsbno; > - xfs_filblks_t fsblen; > - xfs_extlen_t cowextsize; > ssize_t ret; > > - if (!xfs_sb_version_hasreflink(&mp->m_sb)) > - return -EOPNOTSUPP; > - > - if (XFS_FORCED_SHUTDOWN(mp)) > - return -EIO; > - > /* Lock both files against IO */ > ret = xfs_iolock_two_inodes_and_break_layout(inode_in, inode_out); > if (ret) > @@ -1254,8 +1266,6 @@ xfs_reflink_remap_range( > if (ret) > goto out_unlock; > > - trace_xfs_reflink_remap_range(src, pos_in, len, dest, pos_out); > - > /* > * Clear out post-eof preallocations because we don't have page cache > * backing the delayed allocations and they'll never get freed on > @@ -1272,6 +1282,51 @@ xfs_reflink_remap_range( > if (ret) > goto out_unlock; > > + /* Zap any page cache for the destination file's range. */ > + truncate_inode_pages_range(&inode_out->i_data, pos_out, > + PAGE_ALIGN(pos_out + len) - 1); > + return 0; > +out_unlock: > + xfs_reflink_remap_unlock(file_in, file_out); > + return ret; > +} > + > +/* > + * Link a range of blocks from one file to another. > + */ > +int > +xfs_reflink_remap_range( > + struct file *file_in, > + loff_t pos_in, > + struct file *file_out, > + loff_t pos_out, > + u64 len, > + bool is_dedupe) > +{ > + struct inode *inode_in = file_inode(file_in); > + struct xfs_inode *src = XFS_I(inode_in); > + struct inode *inode_out = file_inode(file_out); > + struct xfs_inode *dest = XFS_I(inode_out); > + struct xfs_mount *mp = src->i_mount; > + xfs_fileoff_t sfsbno, dfsbno; > + xfs_filblks_t fsblen; > + xfs_extlen_t cowextsize; > + ssize_t ret; > + > + if (!xfs_sb_version_hasreflink(&mp->m_sb)) > + return -EOPNOTSUPP; > + > + if (XFS_FORCED_SHUTDOWN(mp)) > + return -EIO; > + > + /* Prepare and then clone file data. */ > + ret = xfs_reflink_remap_prep(file_in, pos_in, file_out, pos_out, > + len, is_dedupe); > + if (ret) > + return ret; > + > + trace_xfs_reflink_remap_range(src, pos_in, len, dest, pos_out); > + > dfsbno = XFS_B_TO_FSBT(mp, pos_out); > sfsbno = XFS_B_TO_FSBT(mp, pos_in); > fsblen = XFS_B_TO_FSB(mp, len); > @@ -1280,10 +1335,6 @@ xfs_reflink_remap_range( > if (ret) > goto out_unlock; > > - /* Zap any page cache for the destination file's range. */ > - truncate_inode_pages_range(&inode_out->i_data, pos_out, > - PAGE_ALIGN(pos_out + len) - 1); > - > /* > * Carry the cowextsize hint from src to dest if we're sharing the > * entire source file to the entire destination file, the source file > @@ -1300,12 +1351,7 @@ xfs_reflink_remap_range( > is_dedupe); > > out_unlock: > - xfs_iunlock(dest, XFS_MMAPLOCK_EXCL); > - if (!same_inode) > - xfs_iunlock(src, XFS_MMAPLOCK_SHARED); > - inode_unlock(inode_out); > - if (!same_inode) > - inode_unlock_shared(inode_in); > + xfs_reflink_remap_unlock(file_in, file_out); > if (ret) > trace_xfs_reflink_remap_range_error(dest, ret, _RET_IP_); > return ret; > > -- Dave Chinner david@fromorbit.com From mboxrd@z Thu Jan 1 00:00:00 1970 From: Dave Chinner Date: Fri, 5 Oct 2018 15:28:09 +1000 Subject: [Ocfs2-devel] [PATCH 02/15] xfs: refactor clonerange preparation into a separate helper In-Reply-To: <153870028762.29072.5369530877410002226.stgit@magnolia> References: <153870027422.29072.7433543674436957232.stgit@magnolia> <153870028762.29072.5369530877410002226.stgit@magnolia> Message-ID: <20181005052809.GB12041@dastard> List-Id: MIME-Version: 1.0 Content-Type: text/plain; charset="us-ascii" Content-Transfer-Encoding: 7bit To: "Darrick J. Wong" Cc: linux-xfs@vger.kernel.org, linux-fsdevel@vger.kernel.org, linux-btrfs@vger.kernel.org, ocfs2-devel@oss.oracle.com, sandeen@redhat.com On Thu, Oct 04, 2018 at 05:44:47PM -0700, Darrick J. Wong wrote: > From: Darrick J. Wong > > Refactor all the reflink preparation steps into a separate helper that > we'll use to land all the upcoming fixes for insufficient input checks. If I've read the patch right, this also changes the location of the page cache truncation, right? i.e. it now happens in the xfs_reflink_remap_prep() function instead of after the remap? I think the commit message needs to mention that because it's a fix to incorrect behaviour.... I've added: -- This rework also moves the invalidation of the destination range to the prep function so that it is done before the range is remapped. This ensures that nobody can access the data in range being remapped until the remap is complete. -- Sound OK? Otherwise this looks fine. Reviewed-by: Dave Chinner -Dave. > > Signed-off-by: Darrick J. Wong > --- > fs/xfs/xfs_reflink.c | 96 +++++++++++++++++++++++++++++++++++++------------- > 1 file changed, 71 insertions(+), 25 deletions(-) > > > diff --git a/fs/xfs/xfs_reflink.c b/fs/xfs/xfs_reflink.c > index 38f405415b88..80ca9b6793cd 100644 > --- a/fs/xfs/xfs_reflink.c > +++ b/fs/xfs/xfs_reflink.c > @@ -1195,11 +1195,33 @@ xfs_iolock_two_inodes_and_break_layout( > return 0; > } > > +/* Unlock both inodes after they've been prepped for a range clone. */ > +STATIC void > +xfs_reflink_remap_unlock( > + struct file *file_in, > + struct file *file_out) > +{ > + struct inode *inode_in = file_inode(file_in); > + struct xfs_inode *src = XFS_I(inode_in); > + struct inode *inode_out = file_inode(file_out); > + struct xfs_inode *dest = XFS_I(inode_out); > + bool same_inode = (inode_in == inode_out); > + > + xfs_iunlock(dest, XFS_MMAPLOCK_EXCL); > + if (!same_inode) > + xfs_iunlock(src, XFS_MMAPLOCK_SHARED); > + inode_unlock(inode_out); > + if (!same_inode) > + inode_unlock_shared(inode_in); > +} > + > /* > - * Link a range of blocks from one file to another. > + * Prepare two files for range cloning. Upon a successful return both inodes > + * will have the iolock and mmaplock held, the page cache of the out file > + * will be truncated, and any leases on the out file will have been broken. > */ > -int > -xfs_reflink_remap_range( > +STATIC int > +xfs_reflink_remap_prep( > struct file *file_in, > loff_t pos_in, > struct file *file_out, > @@ -1211,19 +1233,9 @@ xfs_reflink_remap_range( > struct xfs_inode *src = XFS_I(inode_in); > struct inode *inode_out = file_inode(file_out); > struct xfs_inode *dest = XFS_I(inode_out); > - struct xfs_mount *mp = src->i_mount; > bool same_inode = (inode_in == inode_out); > - xfs_fileoff_t sfsbno, dfsbno; > - xfs_filblks_t fsblen; > - xfs_extlen_t cowextsize; > ssize_t ret; > > - if (!xfs_sb_version_hasreflink(&mp->m_sb)) > - return -EOPNOTSUPP; > - > - if (XFS_FORCED_SHUTDOWN(mp)) > - return -EIO; > - > /* Lock both files against IO */ > ret = xfs_iolock_two_inodes_and_break_layout(inode_in, inode_out); > if (ret) > @@ -1254,8 +1266,6 @@ xfs_reflink_remap_range( > if (ret) > goto out_unlock; > > - trace_xfs_reflink_remap_range(src, pos_in, len, dest, pos_out); > - > /* > * Clear out post-eof preallocations because we don't have page cache > * backing the delayed allocations and they'll never get freed on > @@ -1272,6 +1282,51 @@ xfs_reflink_remap_range( > if (ret) > goto out_unlock; > > + /* Zap any page cache for the destination file's range. */ > + truncate_inode_pages_range(&inode_out->i_data, pos_out, > + PAGE_ALIGN(pos_out + len) - 1); > + return 0; > +out_unlock: > + xfs_reflink_remap_unlock(file_in, file_out); > + return ret; > +} > + > +/* > + * Link a range of blocks from one file to another. > + */ > +int > +xfs_reflink_remap_range( > + struct file *file_in, > + loff_t pos_in, > + struct file *file_out, > + loff_t pos_out, > + u64 len, > + bool is_dedupe) > +{ > + struct inode *inode_in = file_inode(file_in); > + struct xfs_inode *src = XFS_I(inode_in); > + struct inode *inode_out = file_inode(file_out); > + struct xfs_inode *dest = XFS_I(inode_out); > + struct xfs_mount *mp = src->i_mount; > + xfs_fileoff_t sfsbno, dfsbno; > + xfs_filblks_t fsblen; > + xfs_extlen_t cowextsize; > + ssize_t ret; > + > + if (!xfs_sb_version_hasreflink(&mp->m_sb)) > + return -EOPNOTSUPP; > + > + if (XFS_FORCED_SHUTDOWN(mp)) > + return -EIO; > + > + /* Prepare and then clone file data. */ > + ret = xfs_reflink_remap_prep(file_in, pos_in, file_out, pos_out, > + len, is_dedupe); > + if (ret) > + return ret; > + > + trace_xfs_reflink_remap_range(src, pos_in, len, dest, pos_out); > + > dfsbno = XFS_B_TO_FSBT(mp, pos_out); > sfsbno = XFS_B_TO_FSBT(mp, pos_in); > fsblen = XFS_B_TO_FSB(mp, len); > @@ -1280,10 +1335,6 @@ xfs_reflink_remap_range( > if (ret) > goto out_unlock; > > - /* Zap any page cache for the destination file's range. */ > - truncate_inode_pages_range(&inode_out->i_data, pos_out, > - PAGE_ALIGN(pos_out + len) - 1); > - > /* > * Carry the cowextsize hint from src to dest if we're sharing the > * entire source file to the entire destination file, the source file > @@ -1300,12 +1351,7 @@ xfs_reflink_remap_range( > is_dedupe); > > out_unlock: > - xfs_iunlock(dest, XFS_MMAPLOCK_EXCL); > - if (!same_inode) > - xfs_iunlock(src, XFS_MMAPLOCK_SHARED); > - inode_unlock(inode_out); > - if (!same_inode) > - inode_unlock_shared(inode_in); > + xfs_reflink_remap_unlock(file_in, file_out); > if (ret) > trace_xfs_reflink_remap_range_error(dest, ret, _RET_IP_); > return ret; > > -- Dave Chinner david at fromorbit.com