From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: from mx1.redhat.com ([209.132.183.28]:52116 "EHLO mx1.redhat.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S2390384AbfFJOCm (ORCPT ); Mon, 10 Jun 2019 10:02:42 -0400 Date: Mon, 10 Jun 2019 10:02:39 -0400 From: Brian Foster Subject: Re: [PATCH 05/10] xfs: move bulkstat ichunk helpers to iwalk code Message-ID: <20190610140236.GE6473@bfoster> References: <155968496814.1657646.13743491598480818627.stgit@magnolia> <155968499971.1657646.988325107267126890.stgit@magnolia> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <155968499971.1657646.988325107267126890.stgit@magnolia> Sender: linux-xfs-owner@vger.kernel.org List-ID: List-Id: xfs To: "Darrick J. Wong" Cc: linux-xfs@vger.kernel.org On Tue, Jun 04, 2019 at 02:49:59PM -0700, Darrick J. Wong wrote: > From: Darrick J. Wong > > Now that we've reworked the bulkstat code to use iwalk, we can move the > old bulkstat ichunk helpers to xfs_iwalk.c. No functional changes here. > > Signed-off-by: Darrick J. Wong > --- Reviewed-by: Brian Foster > fs/xfs/xfs_itable.c | 93 -------------------------------------------------- > fs/xfs/xfs_itable.h | 8 ---- > fs/xfs/xfs_iwalk.c | 95 ++++++++++++++++++++++++++++++++++++++++++++++++++- > 3 files changed, 93 insertions(+), 103 deletions(-) > > > diff --git a/fs/xfs/xfs_itable.c b/fs/xfs/xfs_itable.c > index 87c597ea1df7..06abe5c9c0ee 100644 > --- a/fs/xfs/xfs_itable.c > +++ b/fs/xfs/xfs_itable.c > @@ -186,99 +186,6 @@ xfs_bulkstat_one( > return error; > } > > -/* > - * Loop over all clusters in a chunk for a given incore inode allocation btree > - * record. Do a readahead if there are any allocated inodes in that cluster. > - */ > -void > -xfs_bulkstat_ichunk_ra( > - struct xfs_mount *mp, > - xfs_agnumber_t agno, > - struct xfs_inobt_rec_incore *irec) > -{ > - struct xfs_ino_geometry *igeo = M_IGEO(mp); > - xfs_agblock_t agbno; > - struct blk_plug plug; > - int i; /* inode chunk index */ > - > - agbno = XFS_AGINO_TO_AGBNO(mp, irec->ir_startino); > - > - blk_start_plug(&plug); > - for (i = 0; > - i < XFS_INODES_PER_CHUNK; > - i += igeo->inodes_per_cluster, > - agbno += igeo->blocks_per_cluster) { > - if (xfs_inobt_maskn(i, igeo->inodes_per_cluster) & > - ~irec->ir_free) { > - xfs_btree_reada_bufs(mp, agno, agbno, > - igeo->blocks_per_cluster, > - &xfs_inode_buf_ops); > - } > - } > - blk_finish_plug(&plug); > -} > - > -/* > - * Lookup the inode chunk that the given inode lives in and then get the record > - * if we found the chunk. If the inode was not the last in the chunk and there > - * are some left allocated, update the data for the pointed-to record as well as > - * return the count of grabbed inodes. > - */ > -int > -xfs_bulkstat_grab_ichunk( > - struct xfs_btree_cur *cur, /* btree cursor */ > - xfs_agino_t agino, /* starting inode of chunk */ > - int *icount,/* return # of inodes grabbed */ > - struct xfs_inobt_rec_incore *irec) /* btree record */ > -{ > - int idx; /* index into inode chunk */ > - int stat; > - int error = 0; > - > - /* Lookup the inode chunk that this inode lives in */ > - error = xfs_inobt_lookup(cur, agino, XFS_LOOKUP_LE, &stat); > - if (error) > - return error; > - if (!stat) { > - *icount = 0; > - return error; > - } > - > - /* Get the record, should always work */ > - error = xfs_inobt_get_rec(cur, irec, &stat); > - if (error) > - return error; > - XFS_WANT_CORRUPTED_RETURN(cur->bc_mp, stat == 1); > - > - /* Check if the record contains the inode in request */ > - if (irec->ir_startino + XFS_INODES_PER_CHUNK <= agino) { > - *icount = 0; > - return 0; > - } > - > - idx = agino - irec->ir_startino + 1; > - if (idx < XFS_INODES_PER_CHUNK && > - (xfs_inobt_maskn(idx, XFS_INODES_PER_CHUNK - idx) & ~irec->ir_free)) { > - int i; > - > - /* We got a right chunk with some left inodes allocated at it. > - * Grab the chunk record. Mark all the uninteresting inodes > - * free -- because they're before our start point. > - */ > - for (i = 0; i < idx; i++) { > - if (XFS_INOBT_MASK(i) & ~irec->ir_free) > - irec->ir_freecount++; > - } > - > - irec->ir_free |= xfs_inobt_maskn(0, idx); > - *icount = irec->ir_count - irec->ir_freecount; > - } > - > - return 0; > -} > - > -#define XFS_BULKSTAT_UBLEFT(ubleft) ((ubleft) >= statstruct_size) > - > static int > xfs_bulkstat_iwalk( > struct xfs_mount *mp, > diff --git a/fs/xfs/xfs_itable.h b/fs/xfs/xfs_itable.h > index 366d391eb11f..a2562fe8d282 100644 > --- a/fs/xfs/xfs_itable.h > +++ b/fs/xfs/xfs_itable.h > @@ -67,12 +67,4 @@ xfs_inumbers( > void __user *buffer, /* buffer with inode info */ > inumbers_fmt_pf formatter); > > -/* Temporarily needed while we refactor functions. */ > -struct xfs_btree_cur; > -struct xfs_inobt_rec_incore; > -void xfs_bulkstat_ichunk_ra(struct xfs_mount *mp, xfs_agnumber_t agno, > - struct xfs_inobt_rec_incore *irec); > -int xfs_bulkstat_grab_ichunk(struct xfs_btree_cur *cur, xfs_agino_t agino, > - int *icount, struct xfs_inobt_rec_incore *irec); > - > #endif /* __XFS_ITABLE_H__ */ > diff --git a/fs/xfs/xfs_iwalk.c b/fs/xfs/xfs_iwalk.c > index 3e6c06e69c75..bef0c4907781 100644 > --- a/fs/xfs/xfs_iwalk.c > +++ b/fs/xfs/xfs_iwalk.c > @@ -66,6 +66,97 @@ struct xfs_iwalk_ag { > void *data; > }; > > +/* > + * Loop over all clusters in a chunk for a given incore inode allocation btree > + * record. Do a readahead if there are any allocated inodes in that cluster. > + */ > +STATIC void > +xfs_iwalk_ichunk_ra( > + struct xfs_mount *mp, > + xfs_agnumber_t agno, > + struct xfs_inobt_rec_incore *irec) > +{ > + struct xfs_ino_geometry *igeo = M_IGEO(mp); > + xfs_agblock_t agbno; > + struct blk_plug plug; > + int i; /* inode chunk index */ > + > + agbno = XFS_AGINO_TO_AGBNO(mp, irec->ir_startino); > + > + blk_start_plug(&plug); > + for (i = 0; > + i < XFS_INODES_PER_CHUNK; > + i += igeo->inodes_per_cluster, > + agbno += igeo->blocks_per_cluster) { > + if (xfs_inobt_maskn(i, igeo->inodes_per_cluster) & > + ~irec->ir_free) { > + xfs_btree_reada_bufs(mp, agno, agbno, > + igeo->blocks_per_cluster, > + &xfs_inode_buf_ops); > + } > + } > + blk_finish_plug(&plug); > +} > + > +/* > + * Lookup the inode chunk that the given inode lives in and then get the record > + * if we found the chunk. If the inode was not the last in the chunk and there > + * are some left allocated, update the data for the pointed-to record as well as > + * return the count of grabbed inodes. > + */ > +STATIC int > +xfs_iwalk_grab_ichunk( > + struct xfs_btree_cur *cur, /* btree cursor */ > + xfs_agino_t agino, /* starting inode of chunk */ > + int *icount,/* return # of inodes grabbed */ > + struct xfs_inobt_rec_incore *irec) /* btree record */ > +{ > + int idx; /* index into inode chunk */ > + int stat; > + int error = 0; > + > + /* Lookup the inode chunk that this inode lives in */ > + error = xfs_inobt_lookup(cur, agino, XFS_LOOKUP_LE, &stat); > + if (error) > + return error; > + if (!stat) { > + *icount = 0; > + return error; > + } > + > + /* Get the record, should always work */ > + error = xfs_inobt_get_rec(cur, irec, &stat); > + if (error) > + return error; > + XFS_WANT_CORRUPTED_RETURN(cur->bc_mp, stat == 1); > + > + /* Check if the record contains the inode in request */ > + if (irec->ir_startino + XFS_INODES_PER_CHUNK <= agino) { > + *icount = 0; > + return 0; > + } > + > + idx = agino - irec->ir_startino + 1; > + if (idx < XFS_INODES_PER_CHUNK && > + (xfs_inobt_maskn(idx, XFS_INODES_PER_CHUNK - idx) & ~irec->ir_free)) { > + int i; > + > + /* We got a right chunk with some left inodes allocated at it. > + * Grab the chunk record. Mark all the uninteresting inodes > + * free -- because they're before our start point. > + */ > + for (i = 0; i < idx; i++) { > + if (XFS_INOBT_MASK(i) & ~irec->ir_free) > + irec->ir_freecount++; > + } > + > + irec->ir_free |= xfs_inobt_maskn(0, idx); > + *icount = irec->ir_count - irec->ir_freecount; > + } > + > + return 0; > +} > + > /* Allocate memory for a walk. */ > STATIC int > xfs_iwalk_alloc( > @@ -190,7 +281,7 @@ xfs_iwalk_ag_start( > * We require a lookup cache of at least two elements so that we don't > * have to deal with tearing down the cursor to walk the records. > */ > - error = xfs_bulkstat_grab_ichunk(*curpp, agino - 1, &icount, > + error = xfs_iwalk_grab_ichunk(*curpp, agino - 1, &icount, > &iwag->recs[iwag->nr_recs]); > if (error) > return error; > @@ -295,7 +386,7 @@ xfs_iwalk_ag( > * Start readahead for this inode chunk in anticipation of > * walking the inodes. > */ > - xfs_bulkstat_ichunk_ra(mp, agno, irec); > + xfs_iwalk_ichunk_ra(mp, agno, irec); > > /* > * If there's space in the buffer for more records, increment >