All of lore.kernel.org
 help / color / mirror / Atom feed
* [PATCH] xfs: don't include bnobt blocks when reserving free block pool
@ 2022-03-14 18:08 Darrick J. Wong
  2022-03-16 11:28 ` Brian Foster
  0 siblings, 1 reply; 14+ messages in thread
From: Darrick J. Wong @ 2022-03-14 18:08 UTC (permalink / raw)
  To: Brian Foster; +Cc: xfs

From: Darrick J. Wong <djwong@kernel.org>

xfs_reserve_blocks controls the size of the user-visible free space
reserve pool.  Given the difference between the current and requested
pool sizes, it will try to reserve free space from fdblocks.  However,
the amount requested from fdblocks is also constrained by the amount of
space that we think xfs_mod_fdblocks will give us.  We'll keep trying to
reserve space so long as xfs_mod_fdblocks returns ENOSPC.

In commit fd43cf600cf6, we decided that xfs_mod_fdblocks should not hand
out the "free space" used by the free space btrees, because some portion
of the free space btrees hold in reserve space for future btree
expansion.  Unfortunately, xfs_reserve_blocks' estimation of the number
of blocks that it could request from xfs_mod_fdblocks was not updated to
include m_allocbt_blks, so if space is extremely low, the caller hangs.

Fix this by including m_allocbt_blks in the estimation, and modify the
loop so that it will not retry infinitely.

Found by running xfs/306 (which formats a single-AG 20MB filesystem)
with an fstests configuration that specifies a 1k blocksize and a
specially crafted log size that will consume 7/8 of the space (17920
blocks, specifically) in that AG.

Cc: Brian Foster <bfoster@redhat.com>
Fixes: fd43cf600cf6 ("xfs: set aside allocation btree blocks from block reservation")
Signed-off-by: Darrick J. Wong <djwong@kernel.org>
---
 fs/xfs/xfs_fsops.c |   16 ++++++++++++----
 1 file changed, 12 insertions(+), 4 deletions(-)

diff --git a/fs/xfs/xfs_fsops.c b/fs/xfs/xfs_fsops.c
index 33e26690a8c4..78b6982ea5b0 100644
--- a/fs/xfs/xfs_fsops.c
+++ b/fs/xfs/xfs_fsops.c
@@ -379,6 +379,7 @@ xfs_reserve_blocks(
 	int64_t			fdblks_delta = 0;
 	uint64_t		request;
 	int64_t			free;
+	unsigned int		tries;
 	int			error = 0;
 
 	/* If inval is null, report current values and return */
@@ -432,9 +433,16 @@ xfs_reserve_blocks(
 	 * perform a partial reservation if the request exceeds free space.
 	 */
 	error = -ENOSPC;
-	do {
-		free = percpu_counter_sum(&mp->m_fdblocks) -
-						mp->m_alloc_set_aside;
+	for (tries = 0; tries < 30 && error == -ENOSPC; tries++) {
+		/*
+		 * The reservation pool cannot take space that xfs_mod_fdblocks
+		 * will not give us.  This includes the per-AG set-aside space
+		 * and free space btree blocks that are not available for
+		 * allocation due to per-AG metadata reservations.
+		 */
+		free = percpu_counter_sum(&mp->m_fdblocks);
+		free -= mp->m_alloc_set_aside;
+		free -= atomic64_read(&mp->m_allocbt_blks);
 		if (free <= 0)
 			break;
 
@@ -459,7 +467,7 @@ xfs_reserve_blocks(
 		spin_unlock(&mp->m_sb_lock);
 		error = xfs_mod_fdblocks(mp, -fdblks_delta, 0);
 		spin_lock(&mp->m_sb_lock);
-	} while (error == -ENOSPC);
+	}
 
 	/*
 	 * Update the reserve counters if blocks have been successfully

^ permalink raw reply related	[flat|nested] 14+ messages in thread

* Re: [PATCH] xfs: don't include bnobt blocks when reserving free block pool
  2022-03-14 18:08 [PATCH] xfs: don't include bnobt blocks when reserving free block pool Darrick J. Wong
@ 2022-03-16 11:28 ` Brian Foster
  2022-03-16 16:32   ` Darrick J. Wong
  0 siblings, 1 reply; 14+ messages in thread
From: Brian Foster @ 2022-03-16 11:28 UTC (permalink / raw)
  To: Darrick J. Wong; +Cc: xfs

On Mon, Mar 14, 2022 at 11:08:47AM -0700, Darrick J. Wong wrote:
> From: Darrick J. Wong <djwong@kernel.org>
> 
> xfs_reserve_blocks controls the size of the user-visible free space
> reserve pool.  Given the difference between the current and requested
> pool sizes, it will try to reserve free space from fdblocks.  However,
> the amount requested from fdblocks is also constrained by the amount of
> space that we think xfs_mod_fdblocks will give us.  We'll keep trying to
> reserve space so long as xfs_mod_fdblocks returns ENOSPC.
> 
> In commit fd43cf600cf6, we decided that xfs_mod_fdblocks should not hand
> out the "free space" used by the free space btrees, because some portion
> of the free space btrees hold in reserve space for future btree
> expansion.  Unfortunately, xfs_reserve_blocks' estimation of the number
> of blocks that it could request from xfs_mod_fdblocks was not updated to
> include m_allocbt_blks, so if space is extremely low, the caller hangs.
> 
> Fix this by including m_allocbt_blks in the estimation, and modify the
> loop so that it will not retry infinitely.
> 
> Found by running xfs/306 (which formats a single-AG 20MB filesystem)
> with an fstests configuration that specifies a 1k blocksize and a
> specially crafted log size that will consume 7/8 of the space (17920
> blocks, specifically) in that AG.
> 
> Cc: Brian Foster <bfoster@redhat.com>
> Fixes: fd43cf600cf6 ("xfs: set aside allocation btree blocks from block reservation")
> Signed-off-by: Darrick J. Wong <djwong@kernel.org>
> ---
>  fs/xfs/xfs_fsops.c |   16 ++++++++++++----
>  1 file changed, 12 insertions(+), 4 deletions(-)
> 
> diff --git a/fs/xfs/xfs_fsops.c b/fs/xfs/xfs_fsops.c
> index 33e26690a8c4..78b6982ea5b0 100644
> --- a/fs/xfs/xfs_fsops.c
> +++ b/fs/xfs/xfs_fsops.c
> @@ -379,6 +379,7 @@ xfs_reserve_blocks(
>  	int64_t			fdblks_delta = 0;
>  	uint64_t		request;
>  	int64_t			free;
> +	unsigned int		tries;
>  	int			error = 0;
>  
>  	/* If inval is null, report current values and return */
> @@ -432,9 +433,16 @@ xfs_reserve_blocks(
>  	 * perform a partial reservation if the request exceeds free space.
>  	 */
>  	error = -ENOSPC;
> -	do {
> -		free = percpu_counter_sum(&mp->m_fdblocks) -
> -						mp->m_alloc_set_aside;
> +	for (tries = 0; tries < 30 && error == -ENOSPC; tries++) {

Any reason for the magic number of retries as opposed to perhaps just
not retrying at all? This seems a little odd when you think about it
given that the request is already intended to take available space into
account and modify the request from userspace. OTOH, another
consideration could be to retry some (really large?) number of times and
then bail out if we happen to iterate without an observable change in
free space (i.e., something is wrong), however I suppose that could be
racy as well. *shrug*

> +		/*
> +		 * The reservation pool cannot take space that xfs_mod_fdblocks
> +		 * will not give us.  This includes the per-AG set-aside space
> +		 * and free space btree blocks that are not available for
> +		 * allocation due to per-AG metadata reservations.
> +		 */
> +		free = percpu_counter_sum(&mp->m_fdblocks);
> +		free -= mp->m_alloc_set_aside;
> +		free -= atomic64_read(&mp->m_allocbt_blks);

Seems reasonable. Do we want to consider ->m_allocbt_blks in other
places where ->m_alloc_set_aside is used (i.e. xfs_fs_statfs(), etc.)?
Not sure how much it matters for space reporting purposes, but if so, it
might also be worth reconsidering the usefulness of a static field and
initialization helper (i.e. xfs_alloc_set_aside()) if the majority of
uses involve a dynamic calculation (due to ->m_allocbt_blks).

Brian

>  		if (free <= 0)
>  			break;
>  
> @@ -459,7 +467,7 @@ xfs_reserve_blocks(
>  		spin_unlock(&mp->m_sb_lock);
>  		error = xfs_mod_fdblocks(mp, -fdblks_delta, 0);
>  		spin_lock(&mp->m_sb_lock);
> -	} while (error == -ENOSPC);
> +	}
>  
>  	/*
>  	 * Update the reserve counters if blocks have been successfully
> 


^ permalink raw reply	[flat|nested] 14+ messages in thread

* Re: [PATCH] xfs: don't include bnobt blocks when reserving free block pool
  2022-03-16 11:28 ` Brian Foster
@ 2022-03-16 16:32   ` Darrick J. Wong
  2022-03-16 17:29     ` Brian Foster
  0 siblings, 1 reply; 14+ messages in thread
From: Darrick J. Wong @ 2022-03-16 16:32 UTC (permalink / raw)
  To: Brian Foster; +Cc: xfs

On Wed, Mar 16, 2022 at 07:28:18AM -0400, Brian Foster wrote:
> On Mon, Mar 14, 2022 at 11:08:47AM -0700, Darrick J. Wong wrote:
> > From: Darrick J. Wong <djwong@kernel.org>
> > 
> > xfs_reserve_blocks controls the size of the user-visible free space
> > reserve pool.  Given the difference between the current and requested
> > pool sizes, it will try to reserve free space from fdblocks.  However,
> > the amount requested from fdblocks is also constrained by the amount of
> > space that we think xfs_mod_fdblocks will give us.  We'll keep trying to
> > reserve space so long as xfs_mod_fdblocks returns ENOSPC.
> > 
> > In commit fd43cf600cf6, we decided that xfs_mod_fdblocks should not hand
> > out the "free space" used by the free space btrees, because some portion
> > of the free space btrees hold in reserve space for future btree
> > expansion.  Unfortunately, xfs_reserve_blocks' estimation of the number
> > of blocks that it could request from xfs_mod_fdblocks was not updated to
> > include m_allocbt_blks, so if space is extremely low, the caller hangs.
> > 
> > Fix this by including m_allocbt_blks in the estimation, and modify the
> > loop so that it will not retry infinitely.
> > 
> > Found by running xfs/306 (which formats a single-AG 20MB filesystem)
> > with an fstests configuration that specifies a 1k blocksize and a
> > specially crafted log size that will consume 7/8 of the space (17920
> > blocks, specifically) in that AG.
> > 
> > Cc: Brian Foster <bfoster@redhat.com>
> > Fixes: fd43cf600cf6 ("xfs: set aside allocation btree blocks from block reservation")
> > Signed-off-by: Darrick J. Wong <djwong@kernel.org>
> > ---
> >  fs/xfs/xfs_fsops.c |   16 ++++++++++++----
> >  1 file changed, 12 insertions(+), 4 deletions(-)
> > 
> > diff --git a/fs/xfs/xfs_fsops.c b/fs/xfs/xfs_fsops.c
> > index 33e26690a8c4..78b6982ea5b0 100644
> > --- a/fs/xfs/xfs_fsops.c
> > +++ b/fs/xfs/xfs_fsops.c
> > @@ -379,6 +379,7 @@ xfs_reserve_blocks(
> >  	int64_t			fdblks_delta = 0;
> >  	uint64_t		request;
> >  	int64_t			free;
> > +	unsigned int		tries;
> >  	int			error = 0;
> >  
> >  	/* If inval is null, report current values and return */
> > @@ -432,9 +433,16 @@ xfs_reserve_blocks(
> >  	 * perform a partial reservation if the request exceeds free space.
> >  	 */
> >  	error = -ENOSPC;
> > -	do {
> > -		free = percpu_counter_sum(&mp->m_fdblocks) -
> > -						mp->m_alloc_set_aside;
> > +	for (tries = 0; tries < 30 && error == -ENOSPC; tries++) {
> 
> Any reason for the magic number of retries as opposed to perhaps just
> not retrying at all?

I /think/ the origins of the loop was commit dbcabad19aa9 ("[XFS] Fix
block reservation mechanism."), where I guess Dave decided that we
should loop forever trying to satisfy a request from userspace to
increase the reserve pool.  OFC you and I have been patching this
function to fix all its horrible warts over the years, so maybe you're
right that this should only try once...

(For the mount time default reservation, we should only iterate the loop
once (provided the accounting is correct ;) since nobody else is
touching the free space counters.)

> This seems a little odd when you think about it
> given that the request is already intended to take available space into
> account and modify the request from userspace. OTOH, another
> consideration could be to retry some (really large?) number of times and
> then bail out if we happen to iterate without an observable change in
> free space (i.e., something is wrong), however I suppose that could be
> racy as well. *shrug*

...but if you're the sysadmin desperately trying to increase the size of
the reserve pool when the fs is running near ENOSPC, you're going to be
racing with fdblocks bouncing up and down.  The @free samples that we
take here in the loop body are indeed racy since we can't tell the
difference between @free being unchanged from the last iteration because
someone freed a block and someone else immediately consumed it, or a
totally idle system.

Either way, it's better than hanging the whole system. :)

What if I augment the loop control with a comment capturing some of this:

	/*
	 * The loop body estimates how many blocks it can request from
	 * fdblocks to stash in the reserve pool.  This is a classic
	 * TOCTOU race since fdblocks updates are not always coordinated
	 * via m_sb_lock.  We also cannot tell if @free remaining
	 * unchanged between iterations is due to an idle system or
	 * freed blocks being consumed immediately, so we'll try a
	 * finite number of times to satisfy the request.
	 */
	for (tries = 0; tries < 30...) {

> 
> > +		/*
> > +		 * The reservation pool cannot take space that xfs_mod_fdblocks
> > +		 * will not give us.  This includes the per-AG set-aside space
> > +		 * and free space btree blocks that are not available for
> > +		 * allocation due to per-AG metadata reservations.
> > +		 */
> > +		free = percpu_counter_sum(&mp->m_fdblocks);
> > +		free -= mp->m_alloc_set_aside;
> > +		free -= atomic64_read(&mp->m_allocbt_blks);
> 
> Seems reasonable. Do we want to consider ->m_allocbt_blks in other
> places where ->m_alloc_set_aside is used (i.e. xfs_fs_statfs(), etc.)?
> Not sure how much it matters for space reporting purposes, but if so, it
> might also be worth reconsidering the usefulness of a static field and
> initialization helper (i.e. xfs_alloc_set_aside()) if the majority of
> uses involve a dynamic calculation (due to ->m_allocbt_blks).

When I was writing this patch, I very nearly decided to make those three
lines above their own helper.  I didn't see any other spots that looked
like obvious candidates for such a calculation outside of statfs.

Subtracting m_allocbt_blks from statfs' avail field is a behavior
change, since we always used to consider bnobt blocks as available.  We
don't have an exact count of how many blocks are needed to hide the
per-ag reserved extents, so in the end we have to decide whether we want
to hear complaints about over- or under-estimation of available blocks.

So I think the statfs stuff is a separate patch. :)

--D

> 
> Brian
> 
> >  		if (free <= 0)
> >  			break;
> >  
> > @@ -459,7 +467,7 @@ xfs_reserve_blocks(
> >  		spin_unlock(&mp->m_sb_lock);
> >  		error = xfs_mod_fdblocks(mp, -fdblks_delta, 0);
> >  		spin_lock(&mp->m_sb_lock);
> > -	} while (error == -ENOSPC);
> > +	}
> >  
> >  	/*
> >  	 * Update the reserve counters if blocks have been successfully
> > 
> 

^ permalink raw reply	[flat|nested] 14+ messages in thread

* Re: [PATCH] xfs: don't include bnobt blocks when reserving free block pool
  2022-03-16 16:32   ` Darrick J. Wong
@ 2022-03-16 17:29     ` Brian Foster
  2022-03-16 18:17       ` Darrick J. Wong
  0 siblings, 1 reply; 14+ messages in thread
From: Brian Foster @ 2022-03-16 17:29 UTC (permalink / raw)
  To: Darrick J. Wong; +Cc: xfs

On Wed, Mar 16, 2022 at 09:32:16AM -0700, Darrick J. Wong wrote:
> On Wed, Mar 16, 2022 at 07:28:18AM -0400, Brian Foster wrote:
> > On Mon, Mar 14, 2022 at 11:08:47AM -0700, Darrick J. Wong wrote:
> > > From: Darrick J. Wong <djwong@kernel.org>
> > > 
> > > xfs_reserve_blocks controls the size of the user-visible free space
> > > reserve pool.  Given the difference between the current and requested
> > > pool sizes, it will try to reserve free space from fdblocks.  However,
> > > the amount requested from fdblocks is also constrained by the amount of
> > > space that we think xfs_mod_fdblocks will give us.  We'll keep trying to
> > > reserve space so long as xfs_mod_fdblocks returns ENOSPC.
> > > 
> > > In commit fd43cf600cf6, we decided that xfs_mod_fdblocks should not hand
> > > out the "free space" used by the free space btrees, because some portion
> > > of the free space btrees hold in reserve space for future btree
> > > expansion.  Unfortunately, xfs_reserve_blocks' estimation of the number
> > > of blocks that it could request from xfs_mod_fdblocks was not updated to
> > > include m_allocbt_blks, so if space is extremely low, the caller hangs.
> > > 
> > > Fix this by including m_allocbt_blks in the estimation, and modify the
> > > loop so that it will not retry infinitely.
> > > 
> > > Found by running xfs/306 (which formats a single-AG 20MB filesystem)
> > > with an fstests configuration that specifies a 1k blocksize and a
> > > specially crafted log size that will consume 7/8 of the space (17920
> > > blocks, specifically) in that AG.
> > > 
> > > Cc: Brian Foster <bfoster@redhat.com>
> > > Fixes: fd43cf600cf6 ("xfs: set aside allocation btree blocks from block reservation")
> > > Signed-off-by: Darrick J. Wong <djwong@kernel.org>
> > > ---
> > >  fs/xfs/xfs_fsops.c |   16 ++++++++++++----
> > >  1 file changed, 12 insertions(+), 4 deletions(-)
> > > 
> > > diff --git a/fs/xfs/xfs_fsops.c b/fs/xfs/xfs_fsops.c
> > > index 33e26690a8c4..78b6982ea5b0 100644
> > > --- a/fs/xfs/xfs_fsops.c
> > > +++ b/fs/xfs/xfs_fsops.c
> > > @@ -379,6 +379,7 @@ xfs_reserve_blocks(
> > >  	int64_t			fdblks_delta = 0;
> > >  	uint64_t		request;
> > >  	int64_t			free;
> > > +	unsigned int		tries;
> > >  	int			error = 0;
> > >  
> > >  	/* If inval is null, report current values and return */
> > > @@ -432,9 +433,16 @@ xfs_reserve_blocks(
> > >  	 * perform a partial reservation if the request exceeds free space.
> > >  	 */
> > >  	error = -ENOSPC;
> > > -	do {
> > > -		free = percpu_counter_sum(&mp->m_fdblocks) -
> > > -						mp->m_alloc_set_aside;
> > > +	for (tries = 0; tries < 30 && error == -ENOSPC; tries++) {
> > 
> > Any reason for the magic number of retries as opposed to perhaps just
> > not retrying at all?
> 
> I /think/ the origins of the loop was commit dbcabad19aa9 ("[XFS] Fix
> block reservation mechanism."), where I guess Dave decided that we
> should loop forever trying to satisfy a request from userspace to
> increase the reserve pool.  OFC you and I have been patching this
> function to fix all its horrible warts over the years, so maybe you're
> right that this should only try once...
> 
> (For the mount time default reservation, we should only iterate the loop
> once (provided the accounting is correct ;) since nobody else is
> touching the free space counters.)
> 
> > This seems a little odd when you think about it
> > given that the request is already intended to take available space into
> > account and modify the request from userspace. OTOH, another
> > consideration could be to retry some (really large?) number of times and
> > then bail out if we happen to iterate without an observable change in
> > free space (i.e., something is wrong), however I suppose that could be
> > racy as well. *shrug*
> 
> ...but if you're the sysadmin desperately trying to increase the size of
> the reserve pool when the fs is running near ENOSPC, you're going to be
> racing with fdblocks bouncing up and down.  The @free samples that we
> take here in the loop body are indeed racy since we can't tell the
> difference between @free being unchanged from the last iteration because
> someone freed a block and someone else immediately consumed it, or a
> totally idle system.
> 
> Either way, it's better than hanging the whole system. :)
> 

Yeah.. I'm not bothered much by whether we retry once, 42 times or
forever. I think what this boils down to for me is whether it's worth
the risk of a behavior change of an -ENOSPC return causing something
unexpected for some random user or use case. Could we just do this in
two separate patches? Patch 1 fixes the calculation and targets stable,
patch 2 does whatever to the retry loop that potentially changes retry
semantics (and doesn't really need backporting)..?

> What if I augment the loop control with a comment capturing some of this:
> 
> 	/*
> 	 * The loop body estimates how many blocks it can request from
> 	 * fdblocks to stash in the reserve pool.  This is a classic
> 	 * TOCTOU race since fdblocks updates are not always coordinated
> 	 * via m_sb_lock.  We also cannot tell if @free remaining
> 	 * unchanged between iterations is due to an idle system or
> 	 * freed blocks being consumed immediately, so we'll try a
> 	 * finite number of times to satisfy the request.
> 	 */
> 	for (tries = 0; tries < 30...) {
> 
> > 
> > > +		/*
> > > +		 * The reservation pool cannot take space that xfs_mod_fdblocks
> > > +		 * will not give us.  This includes the per-AG set-aside space
> > > +		 * and free space btree blocks that are not available for
> > > +		 * allocation due to per-AG metadata reservations.
> > > +		 */
> > > +		free = percpu_counter_sum(&mp->m_fdblocks);
> > > +		free -= mp->m_alloc_set_aside;
> > > +		free -= atomic64_read(&mp->m_allocbt_blks);
> > 
> > Seems reasonable. Do we want to consider ->m_allocbt_blks in other
> > places where ->m_alloc_set_aside is used (i.e. xfs_fs_statfs(), etc.)?
> > Not sure how much it matters for space reporting purposes, but if so, it
> > might also be worth reconsidering the usefulness of a static field and
> > initialization helper (i.e. xfs_alloc_set_aside()) if the majority of
> > uses involve a dynamic calculation (due to ->m_allocbt_blks).
> 
> When I was writing this patch, I very nearly decided to make those three
> lines above their own helper.  I didn't see any other spots that looked
> like obvious candidates for such a calculation outside of statfs.
> 

Indeed..

> Subtracting m_allocbt_blks from statfs' avail field is a behavior
> change, since we always used to consider bnobt blocks as available.  We
> don't have an exact count of how many blocks are needed to hide the
> per-ag reserved extents, so in the end we have to decide whether we want
> to hear complaints about over- or under-estimation of available blocks.
> 
> So I think the statfs stuff is a separate patch. :)
> 

Similar deal as above.. I'm more interested in a potential cleanup of
the code that helps prevent this sort of buglet for the next user of
->m_alloc_set_aside that will (expectedly) have no idea about this
subtle quirk than I am about what's presented in the free space
counters. ISTM that we ought to just ditch ->m_alloc_set_aside, replace
the existing xfs_alloc_set_aside() with an XFS_ALLOC_FS_RESERVED() macro
or something that just does the (agcount << 3) thing, and then define a
new xfs_alloc_set_aside() that combines the macro calculation with
->m_allocbt_blks. Then the whole "set aside" concept is calculated and
documented in one place. Hm?

Brian

> --D
> 
> > 
> > Brian
> > 
> > >  		if (free <= 0)
> > >  			break;
> > >  
> > > @@ -459,7 +467,7 @@ xfs_reserve_blocks(
> > >  		spin_unlock(&mp->m_sb_lock);
> > >  		error = xfs_mod_fdblocks(mp, -fdblks_delta, 0);
> > >  		spin_lock(&mp->m_sb_lock);
> > > -	} while (error == -ENOSPC);
> > > +	}
> > >  
> > >  	/*
> > >  	 * Update the reserve counters if blocks have been successfully
> > > 
> > 
> 


^ permalink raw reply	[flat|nested] 14+ messages in thread

* Re: [PATCH] xfs: don't include bnobt blocks when reserving free block pool
  2022-03-16 17:29     ` Brian Foster
@ 2022-03-16 18:17       ` Darrick J. Wong
  2022-03-16 18:48         ` Brian Foster
  2022-03-17  2:05         ` Dave Chinner
  0 siblings, 2 replies; 14+ messages in thread
From: Darrick J. Wong @ 2022-03-16 18:17 UTC (permalink / raw)
  To: Brian Foster; +Cc: xfs

On Wed, Mar 16, 2022 at 01:29:01PM -0400, Brian Foster wrote:
> On Wed, Mar 16, 2022 at 09:32:16AM -0700, Darrick J. Wong wrote:
> > On Wed, Mar 16, 2022 at 07:28:18AM -0400, Brian Foster wrote:
> > > On Mon, Mar 14, 2022 at 11:08:47AM -0700, Darrick J. Wong wrote:
> > > > From: Darrick J. Wong <djwong@kernel.org>
> > > > 
> > > > xfs_reserve_blocks controls the size of the user-visible free space
> > > > reserve pool.  Given the difference between the current and requested
> > > > pool sizes, it will try to reserve free space from fdblocks.  However,
> > > > the amount requested from fdblocks is also constrained by the amount of
> > > > space that we think xfs_mod_fdblocks will give us.  We'll keep trying to
> > > > reserve space so long as xfs_mod_fdblocks returns ENOSPC.
> > > > 
> > > > In commit fd43cf600cf6, we decided that xfs_mod_fdblocks should not hand
> > > > out the "free space" used by the free space btrees, because some portion
> > > > of the free space btrees hold in reserve space for future btree
> > > > expansion.  Unfortunately, xfs_reserve_blocks' estimation of the number
> > > > of blocks that it could request from xfs_mod_fdblocks was not updated to
> > > > include m_allocbt_blks, so if space is extremely low, the caller hangs.
> > > > 
> > > > Fix this by including m_allocbt_blks in the estimation, and modify the
> > > > loop so that it will not retry infinitely.
> > > > 
> > > > Found by running xfs/306 (which formats a single-AG 20MB filesystem)
> > > > with an fstests configuration that specifies a 1k blocksize and a
> > > > specially crafted log size that will consume 7/8 of the space (17920
> > > > blocks, specifically) in that AG.
> > > > 
> > > > Cc: Brian Foster <bfoster@redhat.com>
> > > > Fixes: fd43cf600cf6 ("xfs: set aside allocation btree blocks from block reservation")
> > > > Signed-off-by: Darrick J. Wong <djwong@kernel.org>
> > > > ---
> > > >  fs/xfs/xfs_fsops.c |   16 ++++++++++++----
> > > >  1 file changed, 12 insertions(+), 4 deletions(-)
> > > > 
> > > > diff --git a/fs/xfs/xfs_fsops.c b/fs/xfs/xfs_fsops.c
> > > > index 33e26690a8c4..78b6982ea5b0 100644
> > > > --- a/fs/xfs/xfs_fsops.c
> > > > +++ b/fs/xfs/xfs_fsops.c
> > > > @@ -379,6 +379,7 @@ xfs_reserve_blocks(
> > > >  	int64_t			fdblks_delta = 0;
> > > >  	uint64_t		request;
> > > >  	int64_t			free;
> > > > +	unsigned int		tries;
> > > >  	int			error = 0;
> > > >  
> > > >  	/* If inval is null, report current values and return */
> > > > @@ -432,9 +433,16 @@ xfs_reserve_blocks(
> > > >  	 * perform a partial reservation if the request exceeds free space.
> > > >  	 */
> > > >  	error = -ENOSPC;
> > > > -	do {
> > > > -		free = percpu_counter_sum(&mp->m_fdblocks) -
> > > > -						mp->m_alloc_set_aside;
> > > > +	for (tries = 0; tries < 30 && error == -ENOSPC; tries++) {
> > > 
> > > Any reason for the magic number of retries as opposed to perhaps just
> > > not retrying at all?
> > 
> > I /think/ the origins of the loop was commit dbcabad19aa9 ("[XFS] Fix
> > block reservation mechanism."), where I guess Dave decided that we
> > should loop forever trying to satisfy a request from userspace to
> > increase the reserve pool.  OFC you and I have been patching this
> > function to fix all its horrible warts over the years, so maybe you're
> > right that this should only try once...
> > 
> > (For the mount time default reservation, we should only iterate the loop
> > once (provided the accounting is correct ;) since nobody else is
> > touching the free space counters.)
> > 
> > > This seems a little odd when you think about it
> > > given that the request is already intended to take available space into
> > > account and modify the request from userspace. OTOH, another
> > > consideration could be to retry some (really large?) number of times and
> > > then bail out if we happen to iterate without an observable change in
> > > free space (i.e., something is wrong), however I suppose that could be
> > > racy as well. *shrug*
> > 
> > ...but if you're the sysadmin desperately trying to increase the size of
> > the reserve pool when the fs is running near ENOSPC, you're going to be
> > racing with fdblocks bouncing up and down.  The @free samples that we
> > take here in the loop body are indeed racy since we can't tell the
> > difference between @free being unchanged from the last iteration because
> > someone freed a block and someone else immediately consumed it, or a
> > totally idle system.
> > 
> > Either way, it's better than hanging the whole system. :)
> > 
> 
> Yeah.. I'm not bothered much by whether we retry once, 42 times or
> forever. I think what this boils down to for me is whether it's worth
> the risk of a behavior change of an -ENOSPC return causing something
> unexpected for some random user or use case. Could we just do this in
> two separate patches? Patch 1 fixes the calculation and targets stable,
> patch 2 does whatever to the retry loop that potentially changes retry
> semantics (and doesn't really need backporting)..?

Splitting the two patches sounds good to me.

> > What if I augment the loop control with a comment capturing some of this:
> > 
> > 	/*
> > 	 * The loop body estimates how many blocks it can request from
> > 	 * fdblocks to stash in the reserve pool.  This is a classic
> > 	 * TOCTOU race since fdblocks updates are not always coordinated
> > 	 * via m_sb_lock.  We also cannot tell if @free remaining
> > 	 * unchanged between iterations is due to an idle system or
> > 	 * freed blocks being consumed immediately, so we'll try a
> > 	 * finite number of times to satisfy the request.
> > 	 */
> > 	for (tries = 0; tries < 30...) {
> > 
> > > 
> > > > +		/*
> > > > +		 * The reservation pool cannot take space that xfs_mod_fdblocks
> > > > +		 * will not give us.  This includes the per-AG set-aside space
> > > > +		 * and free space btree blocks that are not available for
> > > > +		 * allocation due to per-AG metadata reservations.
> > > > +		 */
> > > > +		free = percpu_counter_sum(&mp->m_fdblocks);
> > > > +		free -= mp->m_alloc_set_aside;
> > > > +		free -= atomic64_read(&mp->m_allocbt_blks);
> > > 
> > > Seems reasonable. Do we want to consider ->m_allocbt_blks in other
> > > places where ->m_alloc_set_aside is used (i.e. xfs_fs_statfs(), etc.)?
> > > Not sure how much it matters for space reporting purposes, but if so, it
> > > might also be worth reconsidering the usefulness of a static field and
> > > initialization helper (i.e. xfs_alloc_set_aside()) if the majority of
> > > uses involve a dynamic calculation (due to ->m_allocbt_blks).
> > 
> > When I was writing this patch, I very nearly decided to make those three
> > lines above their own helper.  I didn't see any other spots that looked
> > like obvious candidates for such a calculation outside of statfs.
> > 
> 
> Indeed..
> 
> > Subtracting m_allocbt_blks from statfs' avail field is a behavior
> > change, since we always used to consider bnobt blocks as available.  We
> > don't have an exact count of how many blocks are needed to hide the
> > per-ag reserved extents, so in the end we have to decide whether we want
> > to hear complaints about over- or under-estimation of available blocks.
> > 
> > So I think the statfs stuff is a separate patch. :)
> > 
> 
> Similar deal as above.. I'm more interested in a potential cleanup of
> the code that helps prevent this sort of buglet for the next user of
> ->m_alloc_set_aside that will (expectedly) have no idea about this
> subtle quirk than I am about what's presented in the free space
> counters. ISTM that we ought to just ditch ->m_alloc_set_aside, replace
> the existing xfs_alloc_set_aside() with an XFS_ALLOC_FS_RESERVED() macro
> or something that just does the (agcount << 3) thing, and then define a

I'm not sure that the current xfs_alloc_set_aside code is correct.
Right now it comes with this comment:

"We need to reserve 4 fsbs _per AG_ for the freelist and 4 more to
handle a potential split of the file's bmap btree."

I think the first part ("4 fsbs _per AG_ for the freelist") is wrong.
AFAICT, that part refers to the number of blocks we need to keep free in
case we have to replenish a completely empty AGFL.  The hardcoded value
of 4 seems wrong, since xfs_alloc_min_freelist() is what _fix_freelist
uses to decide how big the AGFL needs to be, and it returns 6 on a
filesystem that has rmapbt enabled.  So I think XFS_ALLOC_AGFL_RESERVE
is wrong here and should be replaced with the function call.

I also think the second part ("and 4 more to handle a split of the
file's bmap btree") is wrong.  If we're really supposed to save enough
blocks to handle a bmbt split, then I think this ought to be
(mp->m_bm_maxlevels[0] - 1), not 4, right?  According to xfs_db, bmap
btrees can be 9 levels tall:

xfs_db> btheight bmapbt -n 4294967296 -b 512
bmapbt: worst case per 512-byte block: 13 records (leaf) / 13 keyptrs (node)
level 0: 4294967296 records, 330382100 blocks
level 1: 330382100 records, 25414008 blocks
level 2: 25414008 records, 1954924 blocks
level 3: 1954924 records, 150379 blocks
level 4: 150379 records, 11568 blocks
level 5: 11568 records, 890 blocks
level 6: 890 records, 69 blocks
level 7: 69 records, 6 blocks
level 8: 6 records, 1 block
9 levels, 357913945 blocks total

The root level is in the inode, so we need 8 blocks to handle a full
split on a 512b block filesystem.

Granted, it's at least correct for 4k+ block filesystems:

xfs_db> btheight bmapbt -n 4294967296
bmapbt: worst case per 4096-byte block: 125 records (leaf) / 125 keyptrs (node)
level 0: 4294967296 records, 34359739 blocks
level 1: 34359739 records, 274878 blocks
level 2: 274878 records, 2200 blocks
level 3: 2200 records, 18 blocks
level 4: 18 records, 1 block
5 levels, 34636836 blocks total

So in the end, I think that calculation should become:

unsigned int
xfs_alloc_set_aside(
	struct xfs_mount	*mp)
{
	unsigned int		min-agfl = xfs_alloc_min_freelist(mp, NULL);

	return mp->m_sb.sb_agcount * (min_agfl + mp->m_bm_maxlevels[0] - 1);
}

> new xfs_alloc_set_aside() that combines the macro calculation with
> ->m_allocbt_blks. Then the whole "set aside" concept is calculated and
> documented in one place. Hm?

I think I'd rather call the new function xfs_fdblocks_avail() over
reusing an existing name, because I fear that zapping an old function
and replacing it with a new function with the same name will cause
confusion for anyone backporting patches or reading code after an
absence.

Also the only reason we have a mount variable and a function (instead of
a macro) is that Dave asked me to change the codebase away from the
XFS_ALLOC_AG_MAX_USABLE/XFS_ALLOC_SET_ASIDE macros as part of merging
reflink.

--D

> Brian
> 
> > --D
> > 
> > > 
> > > Brian
> > > 
> > > >  		if (free <= 0)
> > > >  			break;
> > > >  
> > > > @@ -459,7 +467,7 @@ xfs_reserve_blocks(
> > > >  		spin_unlock(&mp->m_sb_lock);
> > > >  		error = xfs_mod_fdblocks(mp, -fdblks_delta, 0);
> > > >  		spin_lock(&mp->m_sb_lock);
> > > > -	} while (error == -ENOSPC);
> > > > +	}
> > > >  
> > > >  	/*
> > > >  	 * Update the reserve counters if blocks have been successfully
> > > > 
> > > 
> > 
> 

^ permalink raw reply	[flat|nested] 14+ messages in thread

* Re: [PATCH] xfs: don't include bnobt blocks when reserving free block pool
  2022-03-16 18:17       ` Darrick J. Wong
@ 2022-03-16 18:48         ` Brian Foster
  2022-03-16 19:17           ` Brian Foster
  2022-03-17  2:05         ` Dave Chinner
  1 sibling, 1 reply; 14+ messages in thread
From: Brian Foster @ 2022-03-16 18:48 UTC (permalink / raw)
  To: Darrick J. Wong; +Cc: xfs

On Wed, Mar 16, 2022 at 11:17:26AM -0700, Darrick J. Wong wrote:
> On Wed, Mar 16, 2022 at 01:29:01PM -0400, Brian Foster wrote:
> > On Wed, Mar 16, 2022 at 09:32:16AM -0700, Darrick J. Wong wrote:
> > > On Wed, Mar 16, 2022 at 07:28:18AM -0400, Brian Foster wrote:
> > > > On Mon, Mar 14, 2022 at 11:08:47AM -0700, Darrick J. Wong wrote:
> > > > > From: Darrick J. Wong <djwong@kernel.org>
> > > > > 
> > > > > xfs_reserve_blocks controls the size of the user-visible free space
> > > > > reserve pool.  Given the difference between the current and requested
> > > > > pool sizes, it will try to reserve free space from fdblocks.  However,
> > > > > the amount requested from fdblocks is also constrained by the amount of
> > > > > space that we think xfs_mod_fdblocks will give us.  We'll keep trying to
> > > > > reserve space so long as xfs_mod_fdblocks returns ENOSPC.
> > > > > 
> > > > > In commit fd43cf600cf6, we decided that xfs_mod_fdblocks should not hand
> > > > > out the "free space" used by the free space btrees, because some portion
> > > > > of the free space btrees hold in reserve space for future btree
> > > > > expansion.  Unfortunately, xfs_reserve_blocks' estimation of the number
> > > > > of blocks that it could request from xfs_mod_fdblocks was not updated to
> > > > > include m_allocbt_blks, so if space is extremely low, the caller hangs.
> > > > > 
> > > > > Fix this by including m_allocbt_blks in the estimation, and modify the
> > > > > loop so that it will not retry infinitely.
> > > > > 
> > > > > Found by running xfs/306 (which formats a single-AG 20MB filesystem)
> > > > > with an fstests configuration that specifies a 1k blocksize and a
> > > > > specially crafted log size that will consume 7/8 of the space (17920
> > > > > blocks, specifically) in that AG.
> > > > > 
> > > > > Cc: Brian Foster <bfoster@redhat.com>
> > > > > Fixes: fd43cf600cf6 ("xfs: set aside allocation btree blocks from block reservation")
> > > > > Signed-off-by: Darrick J. Wong <djwong@kernel.org>
> > > > > ---
> > > > >  fs/xfs/xfs_fsops.c |   16 ++++++++++++----
> > > > >  1 file changed, 12 insertions(+), 4 deletions(-)
> > > > > 
> > > > > diff --git a/fs/xfs/xfs_fsops.c b/fs/xfs/xfs_fsops.c
> > > > > index 33e26690a8c4..78b6982ea5b0 100644
> > > > > --- a/fs/xfs/xfs_fsops.c
> > > > > +++ b/fs/xfs/xfs_fsops.c
> > > > > @@ -379,6 +379,7 @@ xfs_reserve_blocks(
> > > > >  	int64_t			fdblks_delta = 0;
> > > > >  	uint64_t		request;
> > > > >  	int64_t			free;
> > > > > +	unsigned int		tries;
> > > > >  	int			error = 0;
> > > > >  
> > > > >  	/* If inval is null, report current values and return */
> > > > > @@ -432,9 +433,16 @@ xfs_reserve_blocks(
> > > > >  	 * perform a partial reservation if the request exceeds free space.
> > > > >  	 */
> > > > >  	error = -ENOSPC;
> > > > > -	do {
> > > > > -		free = percpu_counter_sum(&mp->m_fdblocks) -
> > > > > -						mp->m_alloc_set_aside;
> > > > > +	for (tries = 0; tries < 30 && error == -ENOSPC; tries++) {
> > > > 
> > > > Any reason for the magic number of retries as opposed to perhaps just
> > > > not retrying at all?
> > > 
> > > I /think/ the origins of the loop was commit dbcabad19aa9 ("[XFS] Fix
> > > block reservation mechanism."), where I guess Dave decided that we
> > > should loop forever trying to satisfy a request from userspace to
> > > increase the reserve pool.  OFC you and I have been patching this
> > > function to fix all its horrible warts over the years, so maybe you're
> > > right that this should only try once...
> > > 
> > > (For the mount time default reservation, we should only iterate the loop
> > > once (provided the accounting is correct ;) since nobody else is
> > > touching the free space counters.)
> > > 
> > > > This seems a little odd when you think about it
> > > > given that the request is already intended to take available space into
> > > > account and modify the request from userspace. OTOH, another
> > > > consideration could be to retry some (really large?) number of times and
> > > > then bail out if we happen to iterate without an observable change in
> > > > free space (i.e., something is wrong), however I suppose that could be
> > > > racy as well. *shrug*
> > > 
> > > ...but if you're the sysadmin desperately trying to increase the size of
> > > the reserve pool when the fs is running near ENOSPC, you're going to be
> > > racing with fdblocks bouncing up and down.  The @free samples that we
> > > take here in the loop body are indeed racy since we can't tell the
> > > difference between @free being unchanged from the last iteration because
> > > someone freed a block and someone else immediately consumed it, or a
> > > totally idle system.
> > > 
> > > Either way, it's better than hanging the whole system. :)
> > > 
> > 
> > Yeah.. I'm not bothered much by whether we retry once, 42 times or
> > forever. I think what this boils down to for me is whether it's worth
> > the risk of a behavior change of an -ENOSPC return causing something
> > unexpected for some random user or use case. Could we just do this in
> > two separate patches? Patch 1 fixes the calculation and targets stable,
> > patch 2 does whatever to the retry loop that potentially changes retry
> > semantics (and doesn't really need backporting)..?
> 
> Splitting the two patches sounds good to me.
> 
> > > What if I augment the loop control with a comment capturing some of this:
> > > 
> > > 	/*
> > > 	 * The loop body estimates how many blocks it can request from
> > > 	 * fdblocks to stash in the reserve pool.  This is a classic
> > > 	 * TOCTOU race since fdblocks updates are not always coordinated
> > > 	 * via m_sb_lock.  We also cannot tell if @free remaining
> > > 	 * unchanged between iterations is due to an idle system or
> > > 	 * freed blocks being consumed immediately, so we'll try a
> > > 	 * finite number of times to satisfy the request.
> > > 	 */
> > > 	for (tries = 0; tries < 30...) {
> > > 
> > > > 
> > > > > +		/*
> > > > > +		 * The reservation pool cannot take space that xfs_mod_fdblocks
> > > > > +		 * will not give us.  This includes the per-AG set-aside space
> > > > > +		 * and free space btree blocks that are not available for
> > > > > +		 * allocation due to per-AG metadata reservations.
> > > > > +		 */
> > > > > +		free = percpu_counter_sum(&mp->m_fdblocks);
> > > > > +		free -= mp->m_alloc_set_aside;
> > > > > +		free -= atomic64_read(&mp->m_allocbt_blks);
> > > > 
> > > > Seems reasonable. Do we want to consider ->m_allocbt_blks in other
> > > > places where ->m_alloc_set_aside is used (i.e. xfs_fs_statfs(), etc.)?
> > > > Not sure how much it matters for space reporting purposes, but if so, it
> > > > might also be worth reconsidering the usefulness of a static field and
> > > > initialization helper (i.e. xfs_alloc_set_aside()) if the majority of
> > > > uses involve a dynamic calculation (due to ->m_allocbt_blks).
> > > 
> > > When I was writing this patch, I very nearly decided to make those three
> > > lines above their own helper.  I didn't see any other spots that looked
> > > like obvious candidates for such a calculation outside of statfs.
> > > 
> > 
> > Indeed..
> > 
> > > Subtracting m_allocbt_blks from statfs' avail field is a behavior
> > > change, since we always used to consider bnobt blocks as available.  We
> > > don't have an exact count of how many blocks are needed to hide the
> > > per-ag reserved extents, so in the end we have to decide whether we want
> > > to hear complaints about over- or under-estimation of available blocks.
> > > 
> > > So I think the statfs stuff is a separate patch. :)
> > > 
> > 
> > Similar deal as above.. I'm more interested in a potential cleanup of
> > the code that helps prevent this sort of buglet for the next user of
> > ->m_alloc_set_aside that will (expectedly) have no idea about this
> > subtle quirk than I am about what's presented in the free space
> > counters. ISTM that we ought to just ditch ->m_alloc_set_aside, replace
> > the existing xfs_alloc_set_aside() with an XFS_ALLOC_FS_RESERVED() macro
> > or something that just does the (agcount << 3) thing, and then define a
> 
> I'm not sure that the current xfs_alloc_set_aside code is correct.
> Right now it comes with this comment:
> 
> "We need to reserve 4 fsbs _per AG_ for the freelist and 4 more to
> handle a potential split of the file's bmap btree."
> 
> I think the first part ("4 fsbs _per AG_ for the freelist") is wrong.
> AFAICT, that part refers to the number of blocks we need to keep free in
> case we have to replenish a completely empty AGFL.  The hardcoded value
> of 4 seems wrong, since xfs_alloc_min_freelist() is what _fix_freelist
> uses to decide how big the AGFL needs to be, and it returns 6 on a
> filesystem that has rmapbt enabled.  So I think XFS_ALLOC_AGFL_RESERVE
> is wrong here and should be replaced with the function call.
> 
> I also think the second part ("and 4 more to handle a split of the
> file's bmap btree") is wrong.  If we're really supposed to save enough
> blocks to handle a bmbt split, then I think this ought to be
> (mp->m_bm_maxlevels[0] - 1), not 4, right?  According to xfs_db, bmap
> btrees can be 9 levels tall:
> 
> xfs_db> btheight bmapbt -n 4294967296 -b 512
> bmapbt: worst case per 512-byte block: 13 records (leaf) / 13 keyptrs (node)
> level 0: 4294967296 records, 330382100 blocks
> level 1: 330382100 records, 25414008 blocks
> level 2: 25414008 records, 1954924 blocks
> level 3: 1954924 records, 150379 blocks
> level 4: 150379 records, 11568 blocks
> level 5: 11568 records, 890 blocks
> level 6: 890 records, 69 blocks
> level 7: 69 records, 6 blocks
> level 8: 6 records, 1 block
> 9 levels, 357913945 blocks total
> 
> The root level is in the inode, so we need 8 blocks to handle a full
> split on a 512b block filesystem.
> 
> Granted, it's at least correct for 4k+ block filesystems:
> 
> xfs_db> btheight bmapbt -n 4294967296
> bmapbt: worst case per 4096-byte block: 125 records (leaf) / 125 keyptrs (node)
> level 0: 4294967296 records, 34359739 blocks
> level 1: 34359739 records, 274878 blocks
> level 2: 274878 records, 2200 blocks
> level 3: 2200 records, 18 blocks
> level 4: 18 records, 1 block
> 5 levels, 34636836 blocks total
> 
> So in the end, I think that calculation should become:
> 
> unsigned int
> xfs_alloc_set_aside(
> 	struct xfs_mount	*mp)
> {
> 	unsigned int		min-agfl = xfs_alloc_min_freelist(mp, NULL);
> 
> 	return mp->m_sb.sb_agcount * (min_agfl + mp->m_bm_maxlevels[0] - 1);
> }
> 

I'm not familiar enough with the original context, but that seems
reasonable to me at a glance. If we do want to change this, I'd again
just suggest a separate patch since presumably there isn't any known bug
report or anything associated with the current code.

> > new xfs_alloc_set_aside() that combines the macro calculation with
> > ->m_allocbt_blks. Then the whole "set aside" concept is calculated and
> > documented in one place. Hm?
> 
> I think I'd rather call the new function xfs_fdblocks_avail() over
> reusing an existing name, because I fear that zapping an old function
> and replacing it with a new function with the same name will cause
> confusion for anyone backporting patches or reading code after an
> absence.
> 

The logical change is basically just to include the allocbt block count
in the "set aside" calculation, not necessarily redefine what "set
aside" means, so I don't really see that as a problem. The function is
only used in a couple places, but we've managed to make it confusing by
conflating the naming of the function, xfs_mount field and local
variables in certain contexts.

BTW, do we ever really expect the "set aside" value to be of any
significance relative to the perag reservation values for any moderately
recent kernel? IIRC (most) perag res is not presented as free space, so
I'm curious if it would even be noticeable if we just fixed and used
xfs_alloc_set_aside() consistently across the board..

Brian

> Also the only reason we have a mount variable and a function (instead of
> a macro) is that Dave asked me to change the codebase away from the
> XFS_ALLOC_AG_MAX_USABLE/XFS_ALLOC_SET_ASIDE macros as part of merging
> reflink.
> 
> --D
> 
> > Brian
> > 
> > > --D
> > > 
> > > > 
> > > > Brian
> > > > 
> > > > >  		if (free <= 0)
> > > > >  			break;
> > > > >  
> > > > > @@ -459,7 +467,7 @@ xfs_reserve_blocks(
> > > > >  		spin_unlock(&mp->m_sb_lock);
> > > > >  		error = xfs_mod_fdblocks(mp, -fdblks_delta, 0);
> > > > >  		spin_lock(&mp->m_sb_lock);
> > > > > -	} while (error == -ENOSPC);
> > > > > +	}
> > > > >  
> > > > >  	/*
> > > > >  	 * Update the reserve counters if blocks have been successfully
> > > > > 
> > > > 
> > > 
> > 
> 


^ permalink raw reply	[flat|nested] 14+ messages in thread

* Re: [PATCH] xfs: don't include bnobt blocks when reserving free block pool
  2022-03-16 18:48         ` Brian Foster
@ 2022-03-16 19:17           ` Brian Foster
  2022-03-16 21:15             ` Darrick J. Wong
  0 siblings, 1 reply; 14+ messages in thread
From: Brian Foster @ 2022-03-16 19:17 UTC (permalink / raw)
  To: Darrick J. Wong; +Cc: xfs

On Wed, Mar 16, 2022 at 02:48:43PM -0400, Brian Foster wrote:
> On Wed, Mar 16, 2022 at 11:17:26AM -0700, Darrick J. Wong wrote:
> > On Wed, Mar 16, 2022 at 01:29:01PM -0400, Brian Foster wrote:
> > > On Wed, Mar 16, 2022 at 09:32:16AM -0700, Darrick J. Wong wrote:
> > > > On Wed, Mar 16, 2022 at 07:28:18AM -0400, Brian Foster wrote:
> > > > > On Mon, Mar 14, 2022 at 11:08:47AM -0700, Darrick J. Wong wrote:
> > > > > > From: Darrick J. Wong <djwong@kernel.org>
> > > > > > 
> > > > > > xfs_reserve_blocks controls the size of the user-visible free space
> > > > > > reserve pool.  Given the difference between the current and requested
> > > > > > pool sizes, it will try to reserve free space from fdblocks.  However,
> > > > > > the amount requested from fdblocks is also constrained by the amount of
> > > > > > space that we think xfs_mod_fdblocks will give us.  We'll keep trying to
> > > > > > reserve space so long as xfs_mod_fdblocks returns ENOSPC.
> > > > > > 
> > > > > > In commit fd43cf600cf6, we decided that xfs_mod_fdblocks should not hand
> > > > > > out the "free space" used by the free space btrees, because some portion
> > > > > > of the free space btrees hold in reserve space for future btree
> > > > > > expansion.  Unfortunately, xfs_reserve_blocks' estimation of the number
> > > > > > of blocks that it could request from xfs_mod_fdblocks was not updated to
> > > > > > include m_allocbt_blks, so if space is extremely low, the caller hangs.
> > > > > > 
> > > > > > Fix this by including m_allocbt_blks in the estimation, and modify the
> > > > > > loop so that it will not retry infinitely.
> > > > > > 
> > > > > > Found by running xfs/306 (which formats a single-AG 20MB filesystem)
> > > > > > with an fstests configuration that specifies a 1k blocksize and a
> > > > > > specially crafted log size that will consume 7/8 of the space (17920
> > > > > > blocks, specifically) in that AG.
> > > > > > 
> > > > > > Cc: Brian Foster <bfoster@redhat.com>
> > > > > > Fixes: fd43cf600cf6 ("xfs: set aside allocation btree blocks from block reservation")
> > > > > > Signed-off-by: Darrick J. Wong <djwong@kernel.org>
> > > > > > ---
> > > > > >  fs/xfs/xfs_fsops.c |   16 ++++++++++++----
> > > > > >  1 file changed, 12 insertions(+), 4 deletions(-)
> > > > > > 
> > > > > > diff --git a/fs/xfs/xfs_fsops.c b/fs/xfs/xfs_fsops.c
> > > > > > index 33e26690a8c4..78b6982ea5b0 100644
> > > > > > --- a/fs/xfs/xfs_fsops.c
> > > > > > +++ b/fs/xfs/xfs_fsops.c
> > > > > > @@ -379,6 +379,7 @@ xfs_reserve_blocks(
> > > > > >  	int64_t			fdblks_delta = 0;
> > > > > >  	uint64_t		request;
> > > > > >  	int64_t			free;
> > > > > > +	unsigned int		tries;
> > > > > >  	int			error = 0;
> > > > > >  
> > > > > >  	/* If inval is null, report current values and return */
> > > > > > @@ -432,9 +433,16 @@ xfs_reserve_blocks(
> > > > > >  	 * perform a partial reservation if the request exceeds free space.
> > > > > >  	 */
> > > > > >  	error = -ENOSPC;
> > > > > > -	do {
> > > > > > -		free = percpu_counter_sum(&mp->m_fdblocks) -
> > > > > > -						mp->m_alloc_set_aside;
> > > > > > +	for (tries = 0; tries < 30 && error == -ENOSPC; tries++) {
> > > > > 
> > > > > Any reason for the magic number of retries as opposed to perhaps just
> > > > > not retrying at all?
> > > > 
> > > > I /think/ the origins of the loop was commit dbcabad19aa9 ("[XFS] Fix
> > > > block reservation mechanism."), where I guess Dave decided that we
> > > > should loop forever trying to satisfy a request from userspace to
> > > > increase the reserve pool.  OFC you and I have been patching this
> > > > function to fix all its horrible warts over the years, so maybe you're
> > > > right that this should only try once...
> > > > 
> > > > (For the mount time default reservation, we should only iterate the loop
> > > > once (provided the accounting is correct ;) since nobody else is
> > > > touching the free space counters.)
> > > > 
> > > > > This seems a little odd when you think about it
> > > > > given that the request is already intended to take available space into
> > > > > account and modify the request from userspace. OTOH, another
> > > > > consideration could be to retry some (really large?) number of times and
> > > > > then bail out if we happen to iterate without an observable change in
> > > > > free space (i.e., something is wrong), however I suppose that could be
> > > > > racy as well. *shrug*
> > > > 
> > > > ...but if you're the sysadmin desperately trying to increase the size of
> > > > the reserve pool when the fs is running near ENOSPC, you're going to be
> > > > racing with fdblocks bouncing up and down.  The @free samples that we
> > > > take here in the loop body are indeed racy since we can't tell the
> > > > difference between @free being unchanged from the last iteration because
> > > > someone freed a block and someone else immediately consumed it, or a
> > > > totally idle system.
> > > > 
> > > > Either way, it's better than hanging the whole system. :)
> > > > 
> > > 
> > > Yeah.. I'm not bothered much by whether we retry once, 42 times or
> > > forever. I think what this boils down to for me is whether it's worth
> > > the risk of a behavior change of an -ENOSPC return causing something
> > > unexpected for some random user or use case. Could we just do this in
> > > two separate patches? Patch 1 fixes the calculation and targets stable,
> > > patch 2 does whatever to the retry loop that potentially changes retry
> > > semantics (and doesn't really need backporting)..?
> > 
> > Splitting the two patches sounds good to me.
> > 
> > > > What if I augment the loop control with a comment capturing some of this:
> > > > 
> > > > 	/*
> > > > 	 * The loop body estimates how many blocks it can request from
> > > > 	 * fdblocks to stash in the reserve pool.  This is a classic
> > > > 	 * TOCTOU race since fdblocks updates are not always coordinated
> > > > 	 * via m_sb_lock.  We also cannot tell if @free remaining
> > > > 	 * unchanged between iterations is due to an idle system or
> > > > 	 * freed blocks being consumed immediately, so we'll try a
> > > > 	 * finite number of times to satisfy the request.
> > > > 	 */
> > > > 	for (tries = 0; tries < 30...) {
> > > > 
> > > > > 
> > > > > > +		/*
> > > > > > +		 * The reservation pool cannot take space that xfs_mod_fdblocks
> > > > > > +		 * will not give us.  This includes the per-AG set-aside space
> > > > > > +		 * and free space btree blocks that are not available for
> > > > > > +		 * allocation due to per-AG metadata reservations.
> > > > > > +		 */
> > > > > > +		free = percpu_counter_sum(&mp->m_fdblocks);
> > > > > > +		free -= mp->m_alloc_set_aside;
> > > > > > +		free -= atomic64_read(&mp->m_allocbt_blks);
> > > > > 
> > > > > Seems reasonable. Do we want to consider ->m_allocbt_blks in other
> > > > > places where ->m_alloc_set_aside is used (i.e. xfs_fs_statfs(), etc.)?
> > > > > Not sure how much it matters for space reporting purposes, but if so, it
> > > > > might also be worth reconsidering the usefulness of a static field and
> > > > > initialization helper (i.e. xfs_alloc_set_aside()) if the majority of
> > > > > uses involve a dynamic calculation (due to ->m_allocbt_blks).
> > > > 
> > > > When I was writing this patch, I very nearly decided to make those three
> > > > lines above their own helper.  I didn't see any other spots that looked
> > > > like obvious candidates for such a calculation outside of statfs.
> > > > 
> > > 
> > > Indeed..
> > > 
> > > > Subtracting m_allocbt_blks from statfs' avail field is a behavior
> > > > change, since we always used to consider bnobt blocks as available.  We
> > > > don't have an exact count of how many blocks are needed to hide the
> > > > per-ag reserved extents, so in the end we have to decide whether we want
> > > > to hear complaints about over- or under-estimation of available blocks.
> > > > 
> > > > So I think the statfs stuff is a separate patch. :)
> > > > 
> > > 
> > > Similar deal as above.. I'm more interested in a potential cleanup of
> > > the code that helps prevent this sort of buglet for the next user of
> > > ->m_alloc_set_aside that will (expectedly) have no idea about this
> > > subtle quirk than I am about what's presented in the free space
> > > counters. ISTM that we ought to just ditch ->m_alloc_set_aside, replace
> > > the existing xfs_alloc_set_aside() with an XFS_ALLOC_FS_RESERVED() macro
> > > or something that just does the (agcount << 3) thing, and then define a
> > 
> > I'm not sure that the current xfs_alloc_set_aside code is correct.
> > Right now it comes with this comment:
> > 
> > "We need to reserve 4 fsbs _per AG_ for the freelist and 4 more to
> > handle a potential split of the file's bmap btree."
> > 
> > I think the first part ("4 fsbs _per AG_ for the freelist") is wrong.
> > AFAICT, that part refers to the number of blocks we need to keep free in
> > case we have to replenish a completely empty AGFL.  The hardcoded value
> > of 4 seems wrong, since xfs_alloc_min_freelist() is what _fix_freelist
> > uses to decide how big the AGFL needs to be, and it returns 6 on a
> > filesystem that has rmapbt enabled.  So I think XFS_ALLOC_AGFL_RESERVE
> > is wrong here and should be replaced with the function call.
> > 
> > I also think the second part ("and 4 more to handle a split of the
> > file's bmap btree") is wrong.  If we're really supposed to save enough
> > blocks to handle a bmbt split, then I think this ought to be
> > (mp->m_bm_maxlevels[0] - 1), not 4, right?  According to xfs_db, bmap
> > btrees can be 9 levels tall:
> > 
> > xfs_db> btheight bmapbt -n 4294967296 -b 512
> > bmapbt: worst case per 512-byte block: 13 records (leaf) / 13 keyptrs (node)
> > level 0: 4294967296 records, 330382100 blocks
> > level 1: 330382100 records, 25414008 blocks
> > level 2: 25414008 records, 1954924 blocks
> > level 3: 1954924 records, 150379 blocks
> > level 4: 150379 records, 11568 blocks
> > level 5: 11568 records, 890 blocks
> > level 6: 890 records, 69 blocks
> > level 7: 69 records, 6 blocks
> > level 8: 6 records, 1 block
> > 9 levels, 357913945 blocks total
> > 
> > The root level is in the inode, so we need 8 blocks to handle a full
> > split on a 512b block filesystem.
> > 
> > Granted, it's at least correct for 4k+ block filesystems:
> > 
> > xfs_db> btheight bmapbt -n 4294967296
> > bmapbt: worst case per 4096-byte block: 125 records (leaf) / 125 keyptrs (node)
> > level 0: 4294967296 records, 34359739 blocks
> > level 1: 34359739 records, 274878 blocks
> > level 2: 274878 records, 2200 blocks
> > level 3: 2200 records, 18 blocks
> > level 4: 18 records, 1 block
> > 5 levels, 34636836 blocks total
> > 
> > So in the end, I think that calculation should become:
> > 
> > unsigned int
> > xfs_alloc_set_aside(
> > 	struct xfs_mount	*mp)
> > {
> > 	unsigned int		min-agfl = xfs_alloc_min_freelist(mp, NULL);
> > 
> > 	return mp->m_sb.sb_agcount * (min_agfl + mp->m_bm_maxlevels[0] - 1);
> > }
> > 
> 
> I'm not familiar enough with the original context, but that seems
> reasonable to me at a glance. If we do want to change this, I'd again
> just suggest a separate patch since presumably there isn't any known bug
> report or anything associated with the current code.
> 
> > > new xfs_alloc_set_aside() that combines the macro calculation with
> > > ->m_allocbt_blks. Then the whole "set aside" concept is calculated and
> > > documented in one place. Hm?
> > 
> > I think I'd rather call the new function xfs_fdblocks_avail() over
> > reusing an existing name, because I fear that zapping an old function
> > and replacing it with a new function with the same name will cause
> > confusion for anyone backporting patches or reading code after an
> > absence.
> > 
> 
> The logical change is basically just to include the allocbt block count
> in the "set aside" calculation, not necessarily redefine what "set
> aside" means, so I don't really see that as a problem. The function is
> only used in a couple places, but we've managed to make it confusing by
> conflating the naming of the function, xfs_mount field and local
> variables in certain contexts.
> 
> BTW, do we ever really expect the "set aside" value to be of any
> significance relative to the perag reservation values for any moderately
> recent kernel? IIRC (most) perag res is not presented as free space, so
> I'm curious if it would even be noticeable if we just fixed and used
> xfs_alloc_set_aside() consistently across the board..
> 

I suppose it's reasonable to want to include allocbt blocks as free
space in general, though, since technically they are eventually usable
blocks. So naming and whatnot aside, at the end the day I'm good with
anything that refactors this code one way or another such that it's
harder to repeat the same mistake.

Brian

> Brian
> 
> > Also the only reason we have a mount variable and a function (instead of
> > a macro) is that Dave asked me to change the codebase away from the
> > XFS_ALLOC_AG_MAX_USABLE/XFS_ALLOC_SET_ASIDE macros as part of merging
> > reflink.
> > 
> > --D
> > 
> > > Brian
> > > 
> > > > --D
> > > > 
> > > > > 
> > > > > Brian
> > > > > 
> > > > > >  		if (free <= 0)
> > > > > >  			break;
> > > > > >  
> > > > > > @@ -459,7 +467,7 @@ xfs_reserve_blocks(
> > > > > >  		spin_unlock(&mp->m_sb_lock);
> > > > > >  		error = xfs_mod_fdblocks(mp, -fdblks_delta, 0);
> > > > > >  		spin_lock(&mp->m_sb_lock);
> > > > > > -	} while (error == -ENOSPC);
> > > > > > +	}
> > > > > >  
> > > > > >  	/*
> > > > > >  	 * Update the reserve counters if blocks have been successfully
> > > > > > 
> > > > > 
> > > > 
> > > 
> > 


^ permalink raw reply	[flat|nested] 14+ messages in thread

* Re: [PATCH] xfs: don't include bnobt blocks when reserving free block pool
  2022-03-16 19:17           ` Brian Foster
@ 2022-03-16 21:15             ` Darrick J. Wong
  2022-03-17  2:19               ` Dave Chinner
  2022-03-17 12:53               ` Brian Foster
  0 siblings, 2 replies; 14+ messages in thread
From: Darrick J. Wong @ 2022-03-16 21:15 UTC (permalink / raw)
  To: Brian Foster; +Cc: xfs

On Wed, Mar 16, 2022 at 03:17:27PM -0400, Brian Foster wrote:
> On Wed, Mar 16, 2022 at 02:48:43PM -0400, Brian Foster wrote:
> > On Wed, Mar 16, 2022 at 11:17:26AM -0700, Darrick J. Wong wrote:
> > > On Wed, Mar 16, 2022 at 01:29:01PM -0400, Brian Foster wrote:
> > > > On Wed, Mar 16, 2022 at 09:32:16AM -0700, Darrick J. Wong wrote:
> > > > > On Wed, Mar 16, 2022 at 07:28:18AM -0400, Brian Foster wrote:
> > > > > > On Mon, Mar 14, 2022 at 11:08:47AM -0700, Darrick J. Wong wrote:
> > > > > > > From: Darrick J. Wong <djwong@kernel.org>
> > > > > > > 
> > > > > > > xfs_reserve_blocks controls the size of the user-visible free space
> > > > > > > reserve pool.  Given the difference between the current and requested
> > > > > > > pool sizes, it will try to reserve free space from fdblocks.  However,
> > > > > > > the amount requested from fdblocks is also constrained by the amount of
> > > > > > > space that we think xfs_mod_fdblocks will give us.  We'll keep trying to
> > > > > > > reserve space so long as xfs_mod_fdblocks returns ENOSPC.
> > > > > > > 
> > > > > > > In commit fd43cf600cf6, we decided that xfs_mod_fdblocks should not hand
> > > > > > > out the "free space" used by the free space btrees, because some portion
> > > > > > > of the free space btrees hold in reserve space for future btree
> > > > > > > expansion.  Unfortunately, xfs_reserve_blocks' estimation of the number
> > > > > > > of blocks that it could request from xfs_mod_fdblocks was not updated to
> > > > > > > include m_allocbt_blks, so if space is extremely low, the caller hangs.
> > > > > > > 
> > > > > > > Fix this by including m_allocbt_blks in the estimation, and modify the
> > > > > > > loop so that it will not retry infinitely.
> > > > > > > 
> > > > > > > Found by running xfs/306 (which formats a single-AG 20MB filesystem)
> > > > > > > with an fstests configuration that specifies a 1k blocksize and a
> > > > > > > specially crafted log size that will consume 7/8 of the space (17920
> > > > > > > blocks, specifically) in that AG.
> > > > > > > 
> > > > > > > Cc: Brian Foster <bfoster@redhat.com>
> > > > > > > Fixes: fd43cf600cf6 ("xfs: set aside allocation btree blocks from block reservation")
> > > > > > > Signed-off-by: Darrick J. Wong <djwong@kernel.org>
> > > > > > > ---
> > > > > > >  fs/xfs/xfs_fsops.c |   16 ++++++++++++----
> > > > > > >  1 file changed, 12 insertions(+), 4 deletions(-)
> > > > > > > 
> > > > > > > diff --git a/fs/xfs/xfs_fsops.c b/fs/xfs/xfs_fsops.c
> > > > > > > index 33e26690a8c4..78b6982ea5b0 100644
> > > > > > > --- a/fs/xfs/xfs_fsops.c
> > > > > > > +++ b/fs/xfs/xfs_fsops.c
> > > > > > > @@ -379,6 +379,7 @@ xfs_reserve_blocks(
> > > > > > >  	int64_t			fdblks_delta = 0;
> > > > > > >  	uint64_t		request;
> > > > > > >  	int64_t			free;
> > > > > > > +	unsigned int		tries;
> > > > > > >  	int			error = 0;
> > > > > > >  
> > > > > > >  	/* If inval is null, report current values and return */
> > > > > > > @@ -432,9 +433,16 @@ xfs_reserve_blocks(
> > > > > > >  	 * perform a partial reservation if the request exceeds free space.
> > > > > > >  	 */
> > > > > > >  	error = -ENOSPC;
> > > > > > > -	do {
> > > > > > > -		free = percpu_counter_sum(&mp->m_fdblocks) -
> > > > > > > -						mp->m_alloc_set_aside;
> > > > > > > +	for (tries = 0; tries < 30 && error == -ENOSPC; tries++) {
> > > > > > 
> > > > > > Any reason for the magic number of retries as opposed to perhaps just
> > > > > > not retrying at all?
> > > > > 
> > > > > I /think/ the origins of the loop was commit dbcabad19aa9 ("[XFS] Fix
> > > > > block reservation mechanism."), where I guess Dave decided that we
> > > > > should loop forever trying to satisfy a request from userspace to
> > > > > increase the reserve pool.  OFC you and I have been patching this
> > > > > function to fix all its horrible warts over the years, so maybe you're
> > > > > right that this should only try once...
> > > > > 
> > > > > (For the mount time default reservation, we should only iterate the loop
> > > > > once (provided the accounting is correct ;) since nobody else is
> > > > > touching the free space counters.)
> > > > > 
> > > > > > This seems a little odd when you think about it
> > > > > > given that the request is already intended to take available space into
> > > > > > account and modify the request from userspace. OTOH, another
> > > > > > consideration could be to retry some (really large?) number of times and
> > > > > > then bail out if we happen to iterate without an observable change in
> > > > > > free space (i.e., something is wrong), however I suppose that could be
> > > > > > racy as well. *shrug*
> > > > > 
> > > > > ...but if you're the sysadmin desperately trying to increase the size of
> > > > > the reserve pool when the fs is running near ENOSPC, you're going to be
> > > > > racing with fdblocks bouncing up and down.  The @free samples that we
> > > > > take here in the loop body are indeed racy since we can't tell the
> > > > > difference between @free being unchanged from the last iteration because
> > > > > someone freed a block and someone else immediately consumed it, or a
> > > > > totally idle system.
> > > > > 
> > > > > Either way, it's better than hanging the whole system. :)
> > > > > 
> > > > 
> > > > Yeah.. I'm not bothered much by whether we retry once, 42 times or
> > > > forever. I think what this boils down to for me is whether it's worth
> > > > the risk of a behavior change of an -ENOSPC return causing something
> > > > unexpected for some random user or use case. Could we just do this in
> > > > two separate patches? Patch 1 fixes the calculation and targets stable,
> > > > patch 2 does whatever to the retry loop that potentially changes retry
> > > > semantics (and doesn't really need backporting)..?
> > > 
> > > Splitting the two patches sounds good to me.
> > > 
> > > > > What if I augment the loop control with a comment capturing some of this:
> > > > > 
> > > > > 	/*
> > > > > 	 * The loop body estimates how many blocks it can request from
> > > > > 	 * fdblocks to stash in the reserve pool.  This is a classic
> > > > > 	 * TOCTOU race since fdblocks updates are not always coordinated
> > > > > 	 * via m_sb_lock.  We also cannot tell if @free remaining
> > > > > 	 * unchanged between iterations is due to an idle system or
> > > > > 	 * freed blocks being consumed immediately, so we'll try a
> > > > > 	 * finite number of times to satisfy the request.
> > > > > 	 */
> > > > > 	for (tries = 0; tries < 30...) {
> > > > > 
> > > > > > 
> > > > > > > +		/*
> > > > > > > +		 * The reservation pool cannot take space that xfs_mod_fdblocks
> > > > > > > +		 * will not give us.  This includes the per-AG set-aside space
> > > > > > > +		 * and free space btree blocks that are not available for
> > > > > > > +		 * allocation due to per-AG metadata reservations.
> > > > > > > +		 */
> > > > > > > +		free = percpu_counter_sum(&mp->m_fdblocks);
> > > > > > > +		free -= mp->m_alloc_set_aside;
> > > > > > > +		free -= atomic64_read(&mp->m_allocbt_blks);
> > > > > > 
> > > > > > Seems reasonable. Do we want to consider ->m_allocbt_blks in other
> > > > > > places where ->m_alloc_set_aside is used (i.e. xfs_fs_statfs(), etc.)?
> > > > > > Not sure how much it matters for space reporting purposes, but if so, it
> > > > > > might also be worth reconsidering the usefulness of a static field and
> > > > > > initialization helper (i.e. xfs_alloc_set_aside()) if the majority of
> > > > > > uses involve a dynamic calculation (due to ->m_allocbt_blks).
> > > > > 
> > > > > When I was writing this patch, I very nearly decided to make those three
> > > > > lines above their own helper.  I didn't see any other spots that looked
> > > > > like obvious candidates for such a calculation outside of statfs.
> > > > > 
> > > > 
> > > > Indeed..
> > > > 
> > > > > Subtracting m_allocbt_blks from statfs' avail field is a behavior
> > > > > change, since we always used to consider bnobt blocks as available.  We
> > > > > don't have an exact count of how many blocks are needed to hide the
> > > > > per-ag reserved extents, so in the end we have to decide whether we want
> > > > > to hear complaints about over- or under-estimation of available blocks.
> > > > > 
> > > > > So I think the statfs stuff is a separate patch. :)
> > > > > 
> > > > 
> > > > Similar deal as above.. I'm more interested in a potential cleanup of
> > > > the code that helps prevent this sort of buglet for the next user of
> > > > ->m_alloc_set_aside that will (expectedly) have no idea about this
> > > > subtle quirk than I am about what's presented in the free space
> > > > counters. ISTM that we ought to just ditch ->m_alloc_set_aside, replace
> > > > the existing xfs_alloc_set_aside() with an XFS_ALLOC_FS_RESERVED() macro
> > > > or something that just does the (agcount << 3) thing, and then define a
> > > 
> > > I'm not sure that the current xfs_alloc_set_aside code is correct.
> > > Right now it comes with this comment:
> > > 
> > > "We need to reserve 4 fsbs _per AG_ for the freelist and 4 more to
> > > handle a potential split of the file's bmap btree."
> > > 
> > > I think the first part ("4 fsbs _per AG_ for the freelist") is wrong.
> > > AFAICT, that part refers to the number of blocks we need to keep free in
> > > case we have to replenish a completely empty AGFL.  The hardcoded value
> > > of 4 seems wrong, since xfs_alloc_min_freelist() is what _fix_freelist
> > > uses to decide how big the AGFL needs to be, and it returns 6 on a
> > > filesystem that has rmapbt enabled.  So I think XFS_ALLOC_AGFL_RESERVE
> > > is wrong here and should be replaced with the function call.
> > > 
> > > I also think the second part ("and 4 more to handle a split of the
> > > file's bmap btree") is wrong.  If we're really supposed to save enough
> > > blocks to handle a bmbt split, then I think this ought to be
> > > (mp->m_bm_maxlevels[0] - 1), not 4, right?  According to xfs_db, bmap
> > > btrees can be 9 levels tall:
> > > 
> > > xfs_db> btheight bmapbt -n 4294967296 -b 512
> > > bmapbt: worst case per 512-byte block: 13 records (leaf) / 13 keyptrs (node)
> > > level 0: 4294967296 records, 330382100 blocks
> > > level 1: 330382100 records, 25414008 blocks
> > > level 2: 25414008 records, 1954924 blocks
> > > level 3: 1954924 records, 150379 blocks
> > > level 4: 150379 records, 11568 blocks
> > > level 5: 11568 records, 890 blocks
> > > level 6: 890 records, 69 blocks
> > > level 7: 69 records, 6 blocks
> > > level 8: 6 records, 1 block
> > > 9 levels, 357913945 blocks total
> > > 
> > > The root level is in the inode, so we need 8 blocks to handle a full
> > > split on a 512b block filesystem.
> > > 
> > > Granted, it's at least correct for 4k+ block filesystems:
> > > 
> > > xfs_db> btheight bmapbt -n 4294967296
> > > bmapbt: worst case per 4096-byte block: 125 records (leaf) / 125 keyptrs (node)
> > > level 0: 4294967296 records, 34359739 blocks
> > > level 1: 34359739 records, 274878 blocks
> > > level 2: 274878 records, 2200 blocks
> > > level 3: 2200 records, 18 blocks
> > > level 4: 18 records, 1 block
> > > 5 levels, 34636836 blocks total
> > > 
> > > So in the end, I think that calculation should become:
> > > 
> > > unsigned int
> > > xfs_alloc_set_aside(
> > > 	struct xfs_mount	*mp)
> > > {
> > > 	unsigned int		min-agfl = xfs_alloc_min_freelist(mp, NULL);
> > > 
> > > 	return mp->m_sb.sb_agcount * (min_agfl + mp->m_bm_maxlevels[0] - 1);
> > > }
> > > 
> > 
> > I'm not familiar enough with the original context, but that seems
> > reasonable to me at a glance. If we do want to change this, I'd again
> > just suggest a separate patch since presumably there isn't any known bug
> > report or anything associated with the current code.

Oh definitely, each of these little fixes nets their own patch.

> > > > new xfs_alloc_set_aside() that combines the macro calculation with
> > > > ->m_allocbt_blks. Then the whole "set aside" concept is calculated and
> > > > documented in one place. Hm?
> > > 
> > > I think I'd rather call the new function xfs_fdblocks_avail() over
> > > reusing an existing name, because I fear that zapping an old function
> > > and replacing it with a new function with the same name will cause
> > > confusion for anyone backporting patches or reading code after an
> > > absence.
> > > 
> > 
> > The logical change is basically just to include the allocbt block count
> > in the "set aside" calculation, not necessarily redefine what "set
> > aside" means, so I don't really see that as a problem. The function is
> > only used in a couple places, but we've managed to make it confusing by
> > conflating the naming of the function, xfs_mount field and local
> > variables in certain contexts.

Hmm.  The way I see things, the setaside is the *minimum* amount of free
space that we need to withhold from the user to guarantee that we can
always make progress in a file mapping update.  m_allocbt_blks is the
amount of space used by free space btrees (currently, the entire btree)
that we withhold from the user to guarantee that btree expansions from
the per-AG reservations can actually be filled.

The reserve pool looks to me like a user-controlled chunk of space that
we also withhold to avoid ENOSPC failure from within writeback; and the
per-AG reservations of course are even more space withheld to guarantee
btree expansion.

IOWs, there are four (I hope!) different free space pools that we use to
guarantee forward progress on this or that operation, and the only thing
they have in common is they don't get reported to userspace as available
blocks.  I don't know that I want to go mixing them too much.

Though, I agree that all of this is really confusing and needs a central
function that makes it obvious how we compute available blocks given all
the free space pools that we have.

> > BTW, do we ever really expect the "set aside" value to be of any
> > significance relative to the perag reservation values for any moderately
> > recent kernel?

Nope.  Any non-tiny filesystem with at least finobt=1 is going to have
way more per-AG reservation than either the setaside or the reserve
pool.

> > IIRC (most) perag res is not presented as free space, so
> > I'm curious if it would even be noticeable if we just fixed and used
> > xfs_alloc_set_aside() consistently across the board..

It isn't.  On a 500G test image with reflink, rmap, and finobt turned
on, the per-AG reservations consume ~630kblocks for each of four AGs of
about ~32.7mblocks apiece.  That's about 2%.  The reservation pool is
8192 blocks and the set-aside is 4*8=32 blocks, both of which are
rounding error by comparison.

At maximal free space fragmentation, the free space btrees would consume
(worst case) about 65kblocks per AG.  That might be noticeable.

> I suppose it's reasonable to want to include allocbt blocks as free
> space in general, though, since technically they are eventually usable
> blocks. So naming and whatnot aside, at the end the day I'm good with
> anything that refactors this code one way or another such that it's
> harder to repeat the same mistake.

Yeah, I at least like the idea of having a function that estimates how
much free space we can try to reserve without using reserve pool blocks.
I also am warming to the idea of telling users they can't have /any/ of
those reserved blocks... particularly since I haven't even gotten to the
"I fill my fs to 99% full and only then buy one more gigabyte of space,
wash, rinse, repeat" complaints.

--D

> 
> Brian
> 
> > Brian
> > 
> > > Also the only reason we have a mount variable and a function (instead of
> > > a macro) is that Dave asked me to change the codebase away from the
> > > XFS_ALLOC_AG_MAX_USABLE/XFS_ALLOC_SET_ASIDE macros as part of merging
> > > reflink.
> > > 
> > > --D
> > > 
> > > > Brian
> > > > 
> > > > > --D
> > > > > 
> > > > > > 
> > > > > > Brian
> > > > > > 
> > > > > > >  		if (free <= 0)
> > > > > > >  			break;
> > > > > > >  
> > > > > > > @@ -459,7 +467,7 @@ xfs_reserve_blocks(
> > > > > > >  		spin_unlock(&mp->m_sb_lock);
> > > > > > >  		error = xfs_mod_fdblocks(mp, -fdblks_delta, 0);
> > > > > > >  		spin_lock(&mp->m_sb_lock);
> > > > > > > -	} while (error == -ENOSPC);
> > > > > > > +	}
> > > > > > >  
> > > > > > >  	/*
> > > > > > >  	 * Update the reserve counters if blocks have been successfully
> > > > > > > 
> > > > > > 
> > > > > 
> > > > 
> > > 
> 

^ permalink raw reply	[flat|nested] 14+ messages in thread

* Re: [PATCH] xfs: don't include bnobt blocks when reserving free block pool
  2022-03-16 18:17       ` Darrick J. Wong
  2022-03-16 18:48         ` Brian Foster
@ 2022-03-17  2:05         ` Dave Chinner
  2022-03-17 12:56           ` Brian Foster
  2022-03-17 17:05           ` Darrick J. Wong
  1 sibling, 2 replies; 14+ messages in thread
From: Dave Chinner @ 2022-03-17  2:05 UTC (permalink / raw)
  To: Darrick J. Wong; +Cc: Brian Foster, xfs

On Wed, Mar 16, 2022 at 11:17:26AM -0700, Darrick J. Wong wrote:
> On Wed, Mar 16, 2022 at 01:29:01PM -0400, Brian Foster wrote:
> > On Wed, Mar 16, 2022 at 09:32:16AM -0700, Darrick J. Wong wrote:
> > > On Wed, Mar 16, 2022 at 07:28:18AM -0400, Brian Foster wrote:
> > > > On Mon, Mar 14, 2022 at 11:08:47AM -0700, Darrick J. Wong wrote:
> > Similar deal as above.. I'm more interested in a potential cleanup of
> > the code that helps prevent this sort of buglet for the next user of
> > ->m_alloc_set_aside that will (expectedly) have no idea about this
> > subtle quirk than I am about what's presented in the free space
> > counters. ISTM that we ought to just ditch ->m_alloc_set_aside, replace
> > the existing xfs_alloc_set_aside() with an XFS_ALLOC_FS_RESERVED() macro
> > or something that just does the (agcount << 3) thing, and then define a
> 
> I'm not sure that the current xfs_alloc_set_aside code is correct.
> Right now it comes with this comment:
> 
> "We need to reserve 4 fsbs _per AG_ for the freelist and 4 more to
> handle a potential split of the file's bmap btree."
>
> I think the first part ("4 fsbs _per AG_ for the freelist") is wrong.
> AFAICT, that part refers to the number of blocks we need to keep free in
> case we have to replenish a completely empty AGFL.  The hardcoded value
> of 4 seems wrong, since xfs_alloc_min_freelist() is what _fix_freelist
> uses to decide how big the AGFL needs to be, and it returns 6 on a
> filesystem that has rmapbt enabled.  So I think XFS_ALLOC_AGFL_RESERVE
> is wrong here and should be replaced with the function call.

Back when I wrote that code (circa 2007, IIRC), that was actually
correct according to the reservations that were made when freeing
an extent at ENOSPC.

We needed 4 blocks for the AGFL fixup to always succeed  - 2 blocks
for each BNO and CNT btrees, and, IIRC, the extent free reservation
was just 4 blocks at that time. Hence the 4+4 value.

However, you are right that rmap also adds another per-ag btree that
is allocated from the agfl and that set_aside() should be taking
that into accout. That said, I think that xfs_alloc_min_freelist()
might even be wrong by just adding 2 blocks to the AGFL for the
rmapbt.

That is, at ENOSPC the rmapbt can be a *big* btree. It's not like
the BNO and CNT btrees which are completely empty at that point in
time; the RMAP tree could be one level below max height, and freeing
a single block could split a rmap rec and trigger a full height RMAP
split.

So the minimum free list length in that case is 2 + 2 + MAX_RMAP_HEIGHT.

> I also think the second part ("and 4 more to handle a split of the
> file's bmap btree") is wrong.  If we're really supposed to save enough
> blocks to handle a bmbt split, then I think this ought to be
> (mp->m_bm_maxlevels[0] - 1), not 4, right?  According to xfs_db, bmap
> btrees can be 9 levels tall:

Yes, we've changed the BMBT reservations in the years since that
code was written to handle max height reservations correctly, too.
So, like the RMAP btree reservation, we probably should be reserving
MAX_BMAP_HEIGHT in the set-aside calculation.

refcount btree space is handled by the ag_resv code and blocks
aren't allocated from the AGFL, so I don't think we need to take
taht into account for xfs_alloc_set_aside().

> So in the end, I think that calculation should become:
> 
> unsigned int
> xfs_alloc_set_aside(
> 	struct xfs_mount	*mp)
> {
> 	unsigned int		min-agfl = xfs_alloc_min_freelist(mp, NULL);
> 
> 	return mp->m_sb.sb_agcount * (min_agfl + mp->m_bm_maxlevels[0] - 1);
> }

*nod*, but with the proviso that xfs_alloc_min_freelist() doesn't
appear to be correct, either....

Also, that's a fixed value for the physical geometry of the
filesystem, so it should be calculated once at mount time and stored
in the xfs_mount (and only updated if needed at growfs time)...

> > new xfs_alloc_set_aside() that combines the macro calculation with
> > ->m_allocbt_blks. Then the whole "set aside" concept is calculated and
> > documented in one place. Hm?
> 
> I think I'd rather call the new function xfs_fdblocks_avail() over
> reusing an existing name, because I fear that zapping an old function
> and replacing it with a new function with the same name will cause
> confusion for anyone backporting patches or reading code after an
> absence.
> 
> Also the only reason we have a mount variable and a function (instead of
> a macro) is that Dave asked me to change the codebase away from the
> XFS_ALLOC_AG_MAX_USABLE/XFS_ALLOC_SET_ASIDE macros as part of merging
> reflink.

Yeah, macros wrapping a variable or repeated constant calculation
are bad, and it's something we've been cleaning up for a long
time...

Cheers,

Dave.
-- 
Dave Chinner
david@fromorbit.com

^ permalink raw reply	[flat|nested] 14+ messages in thread

* Re: [PATCH] xfs: don't include bnobt blocks when reserving free block pool
  2022-03-16 21:15             ` Darrick J. Wong
@ 2022-03-17  2:19               ` Dave Chinner
  2022-03-17 12:53               ` Brian Foster
  1 sibling, 0 replies; 14+ messages in thread
From: Dave Chinner @ 2022-03-17  2:19 UTC (permalink / raw)
  To: Darrick J. Wong; +Cc: Brian Foster, xfs

On Wed, Mar 16, 2022 at 02:15:26PM -0700, Darrick J. Wong wrote:
> Yeah, I at least like the idea of having a function that estimates how
> much free space we can try to reserve without using reserve pool blocks.
> I also am warming to the idea of telling users they can't have /any/ of
> those reserved blocks... particularly since I haven't even gotten to the
> "I fill my fs to 99% full and only then buy one more gigabyte of space,
> wash, rinse, repeat" complaints.

Go look at what I considered for the "thinp aware" filesystems a few
years ago:

https://lore.kernel.org/linux-xfs/20171026083322.20428-1-david@fromorbit.com/

The available space the filesystem presented to the user is
completely divorced from the physical device block count and
internal filesystem metadata space usage. Users simply didn't see
any of the space we reserved for metadata as space they could
allocate. i.e. it split the superblock sb_dblocks value away from
the number of data blocks we present to the user that they can
allocate as user data.

(https://lore.kernel.org/linux-xfs/20171026083322.20428-10-david@fromorbit.com/)

So, yeah, moving towards a model that essentially hides all the
reserved space from users and only presenting them with the amount
of space that is available without dipping into any reserves would
make things an awful lot simpler from an architectural POV...

Cheers,

Dave.
-- 
Dave Chinner
david@fromorbit.com

^ permalink raw reply	[flat|nested] 14+ messages in thread

* Re: [PATCH] xfs: don't include bnobt blocks when reserving free block pool
  2022-03-16 21:15             ` Darrick J. Wong
  2022-03-17  2:19               ` Dave Chinner
@ 2022-03-17 12:53               ` Brian Foster
  1 sibling, 0 replies; 14+ messages in thread
From: Brian Foster @ 2022-03-17 12:53 UTC (permalink / raw)
  To: Darrick J. Wong; +Cc: xfs

On Wed, Mar 16, 2022 at 02:15:26PM -0700, Darrick J. Wong wrote:
> On Wed, Mar 16, 2022 at 03:17:27PM -0400, Brian Foster wrote:
> > On Wed, Mar 16, 2022 at 02:48:43PM -0400, Brian Foster wrote:
> > > On Wed, Mar 16, 2022 at 11:17:26AM -0700, Darrick J. Wong wrote:
> > > > On Wed, Mar 16, 2022 at 01:29:01PM -0400, Brian Foster wrote:
> > > > > On Wed, Mar 16, 2022 at 09:32:16AM -0700, Darrick J. Wong wrote:
> > > > > > On Wed, Mar 16, 2022 at 07:28:18AM -0400, Brian Foster wrote:
> > > > > > > On Mon, Mar 14, 2022 at 11:08:47AM -0700, Darrick J. Wong wrote:
> > > > > > > > From: Darrick J. Wong <djwong@kernel.org>
> > > > > > > > 
> > > > > > > > xfs_reserve_blocks controls the size of the user-visible free space
> > > > > > > > reserve pool.  Given the difference between the current and requested
> > > > > > > > pool sizes, it will try to reserve free space from fdblocks.  However,
> > > > > > > > the amount requested from fdblocks is also constrained by the amount of
> > > > > > > > space that we think xfs_mod_fdblocks will give us.  We'll keep trying to
> > > > > > > > reserve space so long as xfs_mod_fdblocks returns ENOSPC.
> > > > > > > > 
> > > > > > > > In commit fd43cf600cf6, we decided that xfs_mod_fdblocks should not hand
> > > > > > > > out the "free space" used by the free space btrees, because some portion
> > > > > > > > of the free space btrees hold in reserve space for future btree
> > > > > > > > expansion.  Unfortunately, xfs_reserve_blocks' estimation of the number
> > > > > > > > of blocks that it could request from xfs_mod_fdblocks was not updated to
> > > > > > > > include m_allocbt_blks, so if space is extremely low, the caller hangs.
> > > > > > > > 
> > > > > > > > Fix this by including m_allocbt_blks in the estimation, and modify the
> > > > > > > > loop so that it will not retry infinitely.
> > > > > > > > 
> > > > > > > > Found by running xfs/306 (which formats a single-AG 20MB filesystem)
> > > > > > > > with an fstests configuration that specifies a 1k blocksize and a
> > > > > > > > specially crafted log size that will consume 7/8 of the space (17920
> > > > > > > > blocks, specifically) in that AG.
> > > > > > > > 
> > > > > > > > Cc: Brian Foster <bfoster@redhat.com>
> > > > > > > > Fixes: fd43cf600cf6 ("xfs: set aside allocation btree blocks from block reservation")
> > > > > > > > Signed-off-by: Darrick J. Wong <djwong@kernel.org>
> > > > > > > > ---
> > > > > > > >  fs/xfs/xfs_fsops.c |   16 ++++++++++++----
> > > > > > > >  1 file changed, 12 insertions(+), 4 deletions(-)
> > > > > > > > 
> > > > > > > > diff --git a/fs/xfs/xfs_fsops.c b/fs/xfs/xfs_fsops.c
> > > > > > > > index 33e26690a8c4..78b6982ea5b0 100644
> > > > > > > > --- a/fs/xfs/xfs_fsops.c
> > > > > > > > +++ b/fs/xfs/xfs_fsops.c
> > > > > > > > @@ -379,6 +379,7 @@ xfs_reserve_blocks(
> > > > > > > >  	int64_t			fdblks_delta = 0;
> > > > > > > >  	uint64_t		request;
> > > > > > > >  	int64_t			free;
> > > > > > > > +	unsigned int		tries;
> > > > > > > >  	int			error = 0;
> > > > > > > >  
> > > > > > > >  	/* If inval is null, report current values and return */
> > > > > > > > @@ -432,9 +433,16 @@ xfs_reserve_blocks(
> > > > > > > >  	 * perform a partial reservation if the request exceeds free space.
> > > > > > > >  	 */
> > > > > > > >  	error = -ENOSPC;
> > > > > > > > -	do {
> > > > > > > > -		free = percpu_counter_sum(&mp->m_fdblocks) -
> > > > > > > > -						mp->m_alloc_set_aside;
> > > > > > > > +	for (tries = 0; tries < 30 && error == -ENOSPC; tries++) {
> > > > > > > 
> > > > > > > Any reason for the magic number of retries as opposed to perhaps just
> > > > > > > not retrying at all?
> > > > > > 
> > > > > > I /think/ the origins of the loop was commit dbcabad19aa9 ("[XFS] Fix
> > > > > > block reservation mechanism."), where I guess Dave decided that we
> > > > > > should loop forever trying to satisfy a request from userspace to
> > > > > > increase the reserve pool.  OFC you and I have been patching this
> > > > > > function to fix all its horrible warts over the years, so maybe you're
> > > > > > right that this should only try once...
> > > > > > 
> > > > > > (For the mount time default reservation, we should only iterate the loop
> > > > > > once (provided the accounting is correct ;) since nobody else is
> > > > > > touching the free space counters.)
> > > > > > 
> > > > > > > This seems a little odd when you think about it
> > > > > > > given that the request is already intended to take available space into
> > > > > > > account and modify the request from userspace. OTOH, another
> > > > > > > consideration could be to retry some (really large?) number of times and
> > > > > > > then bail out if we happen to iterate without an observable change in
> > > > > > > free space (i.e., something is wrong), however I suppose that could be
> > > > > > > racy as well. *shrug*
> > > > > > 
> > > > > > ...but if you're the sysadmin desperately trying to increase the size of
> > > > > > the reserve pool when the fs is running near ENOSPC, you're going to be
> > > > > > racing with fdblocks bouncing up and down.  The @free samples that we
> > > > > > take here in the loop body are indeed racy since we can't tell the
> > > > > > difference between @free being unchanged from the last iteration because
> > > > > > someone freed a block and someone else immediately consumed it, or a
> > > > > > totally idle system.
> > > > > > 
> > > > > > Either way, it's better than hanging the whole system. :)
> > > > > > 
> > > > > 
> > > > > Yeah.. I'm not bothered much by whether we retry once, 42 times or
> > > > > forever. I think what this boils down to for me is whether it's worth
> > > > > the risk of a behavior change of an -ENOSPC return causing something
> > > > > unexpected for some random user or use case. Could we just do this in
> > > > > two separate patches? Patch 1 fixes the calculation and targets stable,
> > > > > patch 2 does whatever to the retry loop that potentially changes retry
> > > > > semantics (and doesn't really need backporting)..?
> > > > 
> > > > Splitting the two patches sounds good to me.
> > > > 
> > > > > > What if I augment the loop control with a comment capturing some of this:
> > > > > > 
> > > > > > 	/*
> > > > > > 	 * The loop body estimates how many blocks it can request from
> > > > > > 	 * fdblocks to stash in the reserve pool.  This is a classic
> > > > > > 	 * TOCTOU race since fdblocks updates are not always coordinated
> > > > > > 	 * via m_sb_lock.  We also cannot tell if @free remaining
> > > > > > 	 * unchanged between iterations is due to an idle system or
> > > > > > 	 * freed blocks being consumed immediately, so we'll try a
> > > > > > 	 * finite number of times to satisfy the request.
> > > > > > 	 */
> > > > > > 	for (tries = 0; tries < 30...) {
> > > > > > 
> > > > > > > 
> > > > > > > > +		/*
> > > > > > > > +		 * The reservation pool cannot take space that xfs_mod_fdblocks
> > > > > > > > +		 * will not give us.  This includes the per-AG set-aside space
> > > > > > > > +		 * and free space btree blocks that are not available for
> > > > > > > > +		 * allocation due to per-AG metadata reservations.
> > > > > > > > +		 */
> > > > > > > > +		free = percpu_counter_sum(&mp->m_fdblocks);
> > > > > > > > +		free -= mp->m_alloc_set_aside;
> > > > > > > > +		free -= atomic64_read(&mp->m_allocbt_blks);
> > > > > > > 
> > > > > > > Seems reasonable. Do we want to consider ->m_allocbt_blks in other
> > > > > > > places where ->m_alloc_set_aside is used (i.e. xfs_fs_statfs(), etc.)?
> > > > > > > Not sure how much it matters for space reporting purposes, but if so, it
> > > > > > > might also be worth reconsidering the usefulness of a static field and
> > > > > > > initialization helper (i.e. xfs_alloc_set_aside()) if the majority of
> > > > > > > uses involve a dynamic calculation (due to ->m_allocbt_blks).
> > > > > > 
> > > > > > When I was writing this patch, I very nearly decided to make those three
> > > > > > lines above their own helper.  I didn't see any other spots that looked
> > > > > > like obvious candidates for such a calculation outside of statfs.
> > > > > > 
> > > > > 
> > > > > Indeed..
> > > > > 
> > > > > > Subtracting m_allocbt_blks from statfs' avail field is a behavior
> > > > > > change, since we always used to consider bnobt blocks as available.  We
> > > > > > don't have an exact count of how many blocks are needed to hide the
> > > > > > per-ag reserved extents, so in the end we have to decide whether we want
> > > > > > to hear complaints about over- or under-estimation of available blocks.
> > > > > > 
> > > > > > So I think the statfs stuff is a separate patch. :)
> > > > > > 
> > > > > 
> > > > > Similar deal as above.. I'm more interested in a potential cleanup of
> > > > > the code that helps prevent this sort of buglet for the next user of
> > > > > ->m_alloc_set_aside that will (expectedly) have no idea about this
> > > > > subtle quirk than I am about what's presented in the free space
> > > > > counters. ISTM that we ought to just ditch ->m_alloc_set_aside, replace
> > > > > the existing xfs_alloc_set_aside() with an XFS_ALLOC_FS_RESERVED() macro
> > > > > or something that just does the (agcount << 3) thing, and then define a
> > > > 
> > > > I'm not sure that the current xfs_alloc_set_aside code is correct.
> > > > Right now it comes with this comment:
> > > > 
> > > > "We need to reserve 4 fsbs _per AG_ for the freelist and 4 more to
> > > > handle a potential split of the file's bmap btree."
> > > > 
> > > > I think the first part ("4 fsbs _per AG_ for the freelist") is wrong.
> > > > AFAICT, that part refers to the number of blocks we need to keep free in
> > > > case we have to replenish a completely empty AGFL.  The hardcoded value
> > > > of 4 seems wrong, since xfs_alloc_min_freelist() is what _fix_freelist
> > > > uses to decide how big the AGFL needs to be, and it returns 6 on a
> > > > filesystem that has rmapbt enabled.  So I think XFS_ALLOC_AGFL_RESERVE
> > > > is wrong here and should be replaced with the function call.
> > > > 
> > > > I also think the second part ("and 4 more to handle a split of the
> > > > file's bmap btree") is wrong.  If we're really supposed to save enough
> > > > blocks to handle a bmbt split, then I think this ought to be
> > > > (mp->m_bm_maxlevels[0] - 1), not 4, right?  According to xfs_db, bmap
> > > > btrees can be 9 levels tall:
> > > > 
> > > > xfs_db> btheight bmapbt -n 4294967296 -b 512
> > > > bmapbt: worst case per 512-byte block: 13 records (leaf) / 13 keyptrs (node)
> > > > level 0: 4294967296 records, 330382100 blocks
> > > > level 1: 330382100 records, 25414008 blocks
> > > > level 2: 25414008 records, 1954924 blocks
> > > > level 3: 1954924 records, 150379 blocks
> > > > level 4: 150379 records, 11568 blocks
> > > > level 5: 11568 records, 890 blocks
> > > > level 6: 890 records, 69 blocks
> > > > level 7: 69 records, 6 blocks
> > > > level 8: 6 records, 1 block
> > > > 9 levels, 357913945 blocks total
> > > > 
> > > > The root level is in the inode, so we need 8 blocks to handle a full
> > > > split on a 512b block filesystem.
> > > > 
> > > > Granted, it's at least correct for 4k+ block filesystems:
> > > > 
> > > > xfs_db> btheight bmapbt -n 4294967296
> > > > bmapbt: worst case per 4096-byte block: 125 records (leaf) / 125 keyptrs (node)
> > > > level 0: 4294967296 records, 34359739 blocks
> > > > level 1: 34359739 records, 274878 blocks
> > > > level 2: 274878 records, 2200 blocks
> > > > level 3: 2200 records, 18 blocks
> > > > level 4: 18 records, 1 block
> > > > 5 levels, 34636836 blocks total
> > > > 
> > > > So in the end, I think that calculation should become:
> > > > 
> > > > unsigned int
> > > > xfs_alloc_set_aside(
> > > > 	struct xfs_mount	*mp)
> > > > {
> > > > 	unsigned int		min-agfl = xfs_alloc_min_freelist(mp, NULL);
> > > > 
> > > > 	return mp->m_sb.sb_agcount * (min_agfl + mp->m_bm_maxlevels[0] - 1);
> > > > }
> > > > 
> > > 
> > > I'm not familiar enough with the original context, but that seems
> > > reasonable to me at a glance. If we do want to change this, I'd again
> > > just suggest a separate patch since presumably there isn't any known bug
> > > report or anything associated with the current code.
> 
> Oh definitely, each of these little fixes nets their own patch.
> 
> > > > > new xfs_alloc_set_aside() that combines the macro calculation with
> > > > > ->m_allocbt_blks. Then the whole "set aside" concept is calculated and
> > > > > documented in one place. Hm?
> > > > 
> > > > I think I'd rather call the new function xfs_fdblocks_avail() over
> > > > reusing an existing name, because I fear that zapping an old function
> > > > and replacing it with a new function with the same name will cause
> > > > confusion for anyone backporting patches or reading code after an
> > > > absence.
> > > > 
> > > 
> > > The logical change is basically just to include the allocbt block count
> > > in the "set aside" calculation, not necessarily redefine what "set
> > > aside" means, so I don't really see that as a problem. The function is
> > > only used in a couple places, but we've managed to make it confusing by
> > > conflating the naming of the function, xfs_mount field and local
> > > variables in certain contexts.
> 
> Hmm.  The way I see things, the setaside is the *minimum* amount of free
> space that we need to withhold from the user to guarantee that we can
> always make progress in a file mapping update.  m_allocbt_blks is the
> amount of space used by free space btrees (currently, the entire btree)
> that we withhold from the user to guarantee that btree expansions from
> the per-AG reservations can actually be filled.
> 
> The reserve pool looks to me like a user-controlled chunk of space that
> we also withhold to avoid ENOSPC failure from within writeback; and the
> per-AG reservations of course are even more space withheld to guarantee
> btree expansion.
> 
> IOWs, there are four (I hope!) different free space pools that we use to
> guarantee forward progress on this or that operation, and the only thing
> they have in common is they don't get reported to userspace as available
> blocks.  I don't know that I want to go mixing them too much.
> 

Yeah, I'm.. not sure where the idea of mixing them comes from.

To try and clarify, I was just prioritizing reducing confusion of the
set aside bits over whether they happen to be reported as available
space or not. I.e., if we wanted to consolidate to a single helper and
use that everywhere with the side effect that set aside blocks are not
shown as free, then that sounds reasonable to me (re: to Dave's follow
up point as well). OTOH, if you had some reason to not do the latter,
then we should at least try to eliminate the need to duplicate the set
aside naming/calculation where it does happen to be used, because either
way the current code is sort of written to nearly ensure it won't be
used correctly.

Brian

> Though, I agree that all of this is really confusing and needs a central
> function that makes it obvious how we compute available blocks given all
> the free space pools that we have.
> 
> > > BTW, do we ever really expect the "set aside" value to be of any
> > > significance relative to the perag reservation values for any moderately
> > > recent kernel?
> 
> Nope.  Any non-tiny filesystem with at least finobt=1 is going to have
> way more per-AG reservation than either the setaside or the reserve
> pool.
> 
> > > IIRC (most) perag res is not presented as free space, so
> > > I'm curious if it would even be noticeable if we just fixed and used
> > > xfs_alloc_set_aside() consistently across the board..
> 
> It isn't.  On a 500G test image with reflink, rmap, and finobt turned
> on, the per-AG reservations consume ~630kblocks for each of four AGs of
> about ~32.7mblocks apiece.  That's about 2%.  The reservation pool is
> 8192 blocks and the set-aside is 4*8=32 blocks, both of which are
> rounding error by comparison.
> 
> At maximal free space fragmentation, the free space btrees would consume
> (worst case) about 65kblocks per AG.  That might be noticeable.
> 
> > I suppose it's reasonable to want to include allocbt blocks as free
> > space in general, though, since technically they are eventually usable
> > blocks. So naming and whatnot aside, at the end the day I'm good with
> > anything that refactors this code one way or another such that it's
> > harder to repeat the same mistake.
> 
> Yeah, I at least like the idea of having a function that estimates how
> much free space we can try to reserve without using reserve pool blocks.
> I also am warming to the idea of telling users they can't have /any/ of
> those reserved blocks... particularly since I haven't even gotten to the
> "I fill my fs to 99% full and only then buy one more gigabyte of space,
> wash, rinse, repeat" complaints.
> 
> --D
> 
> > 
> > Brian
> > 
> > > Brian
> > > 
> > > > Also the only reason we have a mount variable and a function (instead of
> > > > a macro) is that Dave asked me to change the codebase away from the
> > > > XFS_ALLOC_AG_MAX_USABLE/XFS_ALLOC_SET_ASIDE macros as part of merging
> > > > reflink.
> > > > 
> > > > --D
> > > > 
> > > > > Brian
> > > > > 
> > > > > > --D
> > > > > > 
> > > > > > > 
> > > > > > > Brian
> > > > > > > 
> > > > > > > >  		if (free <= 0)
> > > > > > > >  			break;
> > > > > > > >  
> > > > > > > > @@ -459,7 +467,7 @@ xfs_reserve_blocks(
> > > > > > > >  		spin_unlock(&mp->m_sb_lock);
> > > > > > > >  		error = xfs_mod_fdblocks(mp, -fdblks_delta, 0);
> > > > > > > >  		spin_lock(&mp->m_sb_lock);
> > > > > > > > -	} while (error == -ENOSPC);
> > > > > > > > +	}
> > > > > > > >  
> > > > > > > >  	/*
> > > > > > > >  	 * Update the reserve counters if blocks have been successfully
> > > > > > > > 
> > > > > > > 
> > > > > > 
> > > > > 
> > > > 
> > 
> 


^ permalink raw reply	[flat|nested] 14+ messages in thread

* Re: [PATCH] xfs: don't include bnobt blocks when reserving free block pool
  2022-03-17  2:05         ` Dave Chinner
@ 2022-03-17 12:56           ` Brian Foster
  2022-03-17 15:46             ` Darrick J. Wong
  2022-03-17 17:05           ` Darrick J. Wong
  1 sibling, 1 reply; 14+ messages in thread
From: Brian Foster @ 2022-03-17 12:56 UTC (permalink / raw)
  To: Dave Chinner; +Cc: Darrick J. Wong, xfs

On Thu, Mar 17, 2022 at 01:05:26PM +1100, Dave Chinner wrote:
> On Wed, Mar 16, 2022 at 11:17:26AM -0700, Darrick J. Wong wrote:
> > On Wed, Mar 16, 2022 at 01:29:01PM -0400, Brian Foster wrote:
> > > On Wed, Mar 16, 2022 at 09:32:16AM -0700, Darrick J. Wong wrote:
> > > > On Wed, Mar 16, 2022 at 07:28:18AM -0400, Brian Foster wrote:
> > > > > On Mon, Mar 14, 2022 at 11:08:47AM -0700, Darrick J. Wong wrote:
> > > Similar deal as above.. I'm more interested in a potential cleanup of
> > > the code that helps prevent this sort of buglet for the next user of
> > > ->m_alloc_set_aside that will (expectedly) have no idea about this
> > > subtle quirk than I am about what's presented in the free space
> > > counters. ISTM that we ought to just ditch ->m_alloc_set_aside, replace
> > > the existing xfs_alloc_set_aside() with an XFS_ALLOC_FS_RESERVED() macro
> > > or something that just does the (agcount << 3) thing, and then define a
> > 
> > I'm not sure that the current xfs_alloc_set_aside code is correct.
> > Right now it comes with this comment:
> > 
> > "We need to reserve 4 fsbs _per AG_ for the freelist and 4 more to
> > handle a potential split of the file's bmap btree."
> >
> > I think the first part ("4 fsbs _per AG_ for the freelist") is wrong.
> > AFAICT, that part refers to the number of blocks we need to keep free in
> > case we have to replenish a completely empty AGFL.  The hardcoded value
> > of 4 seems wrong, since xfs_alloc_min_freelist() is what _fix_freelist
> > uses to decide how big the AGFL needs to be, and it returns 6 on a
> > filesystem that has rmapbt enabled.  So I think XFS_ALLOC_AGFL_RESERVE
> > is wrong here and should be replaced with the function call.
> 
> Back when I wrote that code (circa 2007, IIRC), that was actually
> correct according to the reservations that were made when freeing
> an extent at ENOSPC.
> 
> We needed 4 blocks for the AGFL fixup to always succeed  - 2 blocks
> for each BNO and CNT btrees, and, IIRC, the extent free reservation
> was just 4 blocks at that time. Hence the 4+4 value.
> 
> However, you are right that rmap also adds another per-ag btree that
> is allocated from the agfl and that set_aside() should be taking
> that into accout. That said, I think that xfs_alloc_min_freelist()
> might even be wrong by just adding 2 blocks to the AGFL for the
> rmapbt.
> 
> That is, at ENOSPC the rmapbt can be a *big* btree. It's not like
> the BNO and CNT btrees which are completely empty at that point in
> time; the RMAP tree could be one level below max height, and freeing
> a single block could split a rmap rec and trigger a full height RMAP
> split.
> 
> So the minimum free list length in that case is 2 + 2 + MAX_RMAP_HEIGHT.
> 
> > I also think the second part ("and 4 more to handle a split of the
> > file's bmap btree") is wrong.  If we're really supposed to save enough
> > blocks to handle a bmbt split, then I think this ought to be
> > (mp->m_bm_maxlevels[0] - 1), not 4, right?  According to xfs_db, bmap
> > btrees can be 9 levels tall:
> 
> Yes, we've changed the BMBT reservations in the years since that
> code was written to handle max height reservations correctly, too.
> So, like the RMAP btree reservation, we probably should be reserving
> MAX_BMAP_HEIGHT in the set-aside calculation.
> 
> refcount btree space is handled by the ag_resv code and blocks
> aren't allocated from the AGFL, so I don't think we need to take
> taht into account for xfs_alloc_set_aside().
> 
> > So in the end, I think that calculation should become:
> > 
> > unsigned int
> > xfs_alloc_set_aside(
> > 	struct xfs_mount	*mp)
> > {
> > 	unsigned int		min-agfl = xfs_alloc_min_freelist(mp, NULL);
> > 
> > 	return mp->m_sb.sb_agcount * (min_agfl + mp->m_bm_maxlevels[0] - 1);
> > }
> 
> *nod*, but with the proviso that xfs_alloc_min_freelist() doesn't
> appear to be correct, either....
> 
> Also, that's a fixed value for the physical geometry of the
> filesystem, so it should be calculated once at mount time and stored
> in the xfs_mount (and only updated if needed at growfs time)...
> 

To my earlier point... please just don't call this fixed mount value
"set_aside" if that's not what it actually is. Rename the field and
helper to something self-descriptive based on whatever fixed components
it's made up of (you could even qualify it as a subcomponent of
set_aside with something like ".._agfl_bmap_set_aside" or whatever) then
reserve the _set_aside() name for the helper that calculates and
documents what the actual/final/dynamic "set aside" value is.

Brian

> > > new xfs_alloc_set_aside() that combines the macro calculation with
> > > ->m_allocbt_blks. Then the whole "set aside" concept is calculated and
> > > documented in one place. Hm?
> > 
> > I think I'd rather call the new function xfs_fdblocks_avail() over
> > reusing an existing name, because I fear that zapping an old function
> > and replacing it with a new function with the same name will cause
> > confusion for anyone backporting patches or reading code after an
> > absence.
> > 
> > Also the only reason we have a mount variable and a function (instead of
> > a macro) is that Dave asked me to change the codebase away from the
> > XFS_ALLOC_AG_MAX_USABLE/XFS_ALLOC_SET_ASIDE macros as part of merging
> > reflink.
> 
> Yeah, macros wrapping a variable or repeated constant calculation
> are bad, and it's something we've been cleaning up for a long
> time...
> 
> Cheers,
> 
> Dave.
> -- 
> Dave Chinner
> david@fromorbit.com
> 


^ permalink raw reply	[flat|nested] 14+ messages in thread

* Re: [PATCH] xfs: don't include bnobt blocks when reserving free block pool
  2022-03-17 12:56           ` Brian Foster
@ 2022-03-17 15:46             ` Darrick J. Wong
  0 siblings, 0 replies; 14+ messages in thread
From: Darrick J. Wong @ 2022-03-17 15:46 UTC (permalink / raw)
  To: Brian Foster; +Cc: Dave Chinner, xfs

On Thu, Mar 17, 2022 at 08:56:55AM -0400, Brian Foster wrote:
> On Thu, Mar 17, 2022 at 01:05:26PM +1100, Dave Chinner wrote:
> > On Wed, Mar 16, 2022 at 11:17:26AM -0700, Darrick J. Wong wrote:
> > > On Wed, Mar 16, 2022 at 01:29:01PM -0400, Brian Foster wrote:
> > > > On Wed, Mar 16, 2022 at 09:32:16AM -0700, Darrick J. Wong wrote:
> > > > > On Wed, Mar 16, 2022 at 07:28:18AM -0400, Brian Foster wrote:
> > > > > > On Mon, Mar 14, 2022 at 11:08:47AM -0700, Darrick J. Wong wrote:
> > > > Similar deal as above.. I'm more interested in a potential cleanup of
> > > > the code that helps prevent this sort of buglet for the next user of
> > > > ->m_alloc_set_aside that will (expectedly) have no idea about this
> > > > subtle quirk than I am about what's presented in the free space
> > > > counters. ISTM that we ought to just ditch ->m_alloc_set_aside, replace
> > > > the existing xfs_alloc_set_aside() with an XFS_ALLOC_FS_RESERVED() macro
> > > > or something that just does the (agcount << 3) thing, and then define a
> > > 
> > > I'm not sure that the current xfs_alloc_set_aside code is correct.
> > > Right now it comes with this comment:
> > > 
> > > "We need to reserve 4 fsbs _per AG_ for the freelist and 4 more to
> > > handle a potential split of the file's bmap btree."
> > >
> > > I think the first part ("4 fsbs _per AG_ for the freelist") is wrong.
> > > AFAICT, that part refers to the number of blocks we need to keep free in
> > > case we have to replenish a completely empty AGFL.  The hardcoded value
> > > of 4 seems wrong, since xfs_alloc_min_freelist() is what _fix_freelist
> > > uses to decide how big the AGFL needs to be, and it returns 6 on a
> > > filesystem that has rmapbt enabled.  So I think XFS_ALLOC_AGFL_RESERVE
> > > is wrong here and should be replaced with the function call.
> > 
> > Back when I wrote that code (circa 2007, IIRC), that was actually
> > correct according to the reservations that were made when freeing
> > an extent at ENOSPC.
> > 
> > We needed 4 blocks for the AGFL fixup to always succeed  - 2 blocks
> > for each BNO and CNT btrees, and, IIRC, the extent free reservation
> > was just 4 blocks at that time. Hence the 4+4 value.
> > 
> > However, you are right that rmap also adds another per-ag btree that
> > is allocated from the agfl and that set_aside() should be taking
> > that into accout. That said, I think that xfs_alloc_min_freelist()
> > might even be wrong by just adding 2 blocks to the AGFL for the
> > rmapbt.
> > 
> > That is, at ENOSPC the rmapbt can be a *big* btree. It's not like
> > the BNO and CNT btrees which are completely empty at that point in
> > time; the RMAP tree could be one level below max height, and freeing
> > a single block could split a rmap rec and trigger a full height RMAP
> > split.
> > 
> > So the minimum free list length in that case is 2 + 2 + MAX_RMAP_HEIGHT.
> > 
> > > I also think the second part ("and 4 more to handle a split of the
> > > file's bmap btree") is wrong.  If we're really supposed to save enough
> > > blocks to handle a bmbt split, then I think this ought to be
> > > (mp->m_bm_maxlevels[0] - 1), not 4, right?  According to xfs_db, bmap
> > > btrees can be 9 levels tall:
> > 
> > Yes, we've changed the BMBT reservations in the years since that
> > code was written to handle max height reservations correctly, too.
> > So, like the RMAP btree reservation, we probably should be reserving
> > MAX_BMAP_HEIGHT in the set-aside calculation.
> > 
> > refcount btree space is handled by the ag_resv code and blocks
> > aren't allocated from the AGFL, so I don't think we need to take
> > taht into account for xfs_alloc_set_aside().
> > 
> > > So in the end, I think that calculation should become:
> > > 
> > > unsigned int
> > > xfs_alloc_set_aside(
> > > 	struct xfs_mount	*mp)
> > > {
> > > 	unsigned int		min-agfl = xfs_alloc_min_freelist(mp, NULL);
> > > 
> > > 	return mp->m_sb.sb_agcount * (min_agfl + mp->m_bm_maxlevels[0] - 1);
> > > }
> > 
> > *nod*, but with the proviso that xfs_alloc_min_freelist() doesn't
> > appear to be correct, either....
> > 
> > Also, that's a fixed value for the physical geometry of the
> > filesystem, so it should be calculated once at mount time and stored
> > in the xfs_mount (and only updated if needed at growfs time)...
> > 
> 
> To my earlier point... please just don't call this fixed mount value
> "set_aside" if that's not what it actually is. Rename the field and
> helper to something self-descriptive based on whatever fixed components
> it's made up of (you could even qualify it as a subcomponent of
> set_aside with something like ".._agfl_bmap_set_aside" or whatever) then
> reserve the _set_aside() name for the helper that calculates and
> documents what the actual/final/dynamic "set aside" value is.

Ahh, ok, that's what you were getting at.  Every time I look at
"alloc_set_aside" I have to figure out what that /really/ does, and I
think "agfl_bmap_setaside" (or even "bmbt_split_setaside") is a better
hint to what this actually does.

--D

> Brian
> 
> > > > new xfs_alloc_set_aside() that combines the macro calculation with
> > > > ->m_allocbt_blks. Then the whole "set aside" concept is calculated and
> > > > documented in one place. Hm?
> > > 
> > > I think I'd rather call the new function xfs_fdblocks_avail() over
> > > reusing an existing name, because I fear that zapping an old function
> > > and replacing it with a new function with the same name will cause
> > > confusion for anyone backporting patches or reading code after an
> > > absence.
> > > 
> > > Also the only reason we have a mount variable and a function (instead of
> > > a macro) is that Dave asked me to change the codebase away from the
> > > XFS_ALLOC_AG_MAX_USABLE/XFS_ALLOC_SET_ASIDE macros as part of merging
> > > reflink.
> > 
> > Yeah, macros wrapping a variable or repeated constant calculation
> > are bad, and it's something we've been cleaning up for a long
> > time...
> > 
> > Cheers,
> > 
> > Dave.
> > -- 
> > Dave Chinner
> > david@fromorbit.com
> > 
> 

^ permalink raw reply	[flat|nested] 14+ messages in thread

* Re: [PATCH] xfs: don't include bnobt blocks when reserving free block pool
  2022-03-17  2:05         ` Dave Chinner
  2022-03-17 12:56           ` Brian Foster
@ 2022-03-17 17:05           ` Darrick J. Wong
  1 sibling, 0 replies; 14+ messages in thread
From: Darrick J. Wong @ 2022-03-17 17:05 UTC (permalink / raw)
  To: Dave Chinner; +Cc: Brian Foster, xfs

On Thu, Mar 17, 2022 at 01:05:26PM +1100, Dave Chinner wrote:
> On Wed, Mar 16, 2022 at 11:17:26AM -0700, Darrick J. Wong wrote:
> > On Wed, Mar 16, 2022 at 01:29:01PM -0400, Brian Foster wrote:
> > > On Wed, Mar 16, 2022 at 09:32:16AM -0700, Darrick J. Wong wrote:
> > > > On Wed, Mar 16, 2022 at 07:28:18AM -0400, Brian Foster wrote:
> > > > > On Mon, Mar 14, 2022 at 11:08:47AM -0700, Darrick J. Wong wrote:
> > > Similar deal as above.. I'm more interested in a potential cleanup of
> > > the code that helps prevent this sort of buglet for the next user of
> > > ->m_alloc_set_aside that will (expectedly) have no idea about this
> > > subtle quirk than I am about what's presented in the free space
> > > counters. ISTM that we ought to just ditch ->m_alloc_set_aside, replace
> > > the existing xfs_alloc_set_aside() with an XFS_ALLOC_FS_RESERVED() macro
> > > or something that just does the (agcount << 3) thing, and then define a
> > 
> > I'm not sure that the current xfs_alloc_set_aside code is correct.
> > Right now it comes with this comment:
> > 
> > "We need to reserve 4 fsbs _per AG_ for the freelist and 4 more to
> > handle a potential split of the file's bmap btree."
> >
> > I think the first part ("4 fsbs _per AG_ for the freelist") is wrong.
> > AFAICT, that part refers to the number of blocks we need to keep free in
> > case we have to replenish a completely empty AGFL.  The hardcoded value
> > of 4 seems wrong, since xfs_alloc_min_freelist() is what _fix_freelist
> > uses to decide how big the AGFL needs to be, and it returns 6 on a
> > filesystem that has rmapbt enabled.  So I think XFS_ALLOC_AGFL_RESERVE
> > is wrong here and should be replaced with the function call.
> 
> Back when I wrote that code (circa 2007, IIRC), that was actually
> correct according to the reservations that were made when freeing
> an extent at ENOSPC.
> 
> We needed 4 blocks for the AGFL fixup to always succeed  - 2 blocks
> for each BNO and CNT btrees, and, IIRC, the extent free reservation
> was just 4 blocks at that time. Hence the 4+4 value.
> 
> However, you are right that rmap also adds another per-ag btree that
> is allocated from the agfl and that set_aside() should be taking
> that into accout. That said, I think that xfs_alloc_min_freelist()
> might even be wrong by just adding 2 blocks to the AGFL for the
> rmapbt.
> 
> That is, at ENOSPC the rmapbt can be a *big* btree. It's not like
> the BNO and CNT btrees which are completely empty at that point in
> time; the RMAP tree could be one level below max height, and freeing
> a single block could split a rmap rec and trigger a full height RMAP
> split.
> 
> So the minimum free list length in that case is 2 + 2 + MAX_RMAP_HEIGHT.

The rmap btree can become a big btree, but the per-ag rmapbt reservation
ensures that there's enough free space to refill the AGFL to handle the
rmap btree expanding to its maximum allowable size.  XFS_AG_RESV_RMAPBT
is subtracted from fdblocks, so I don't think alloc_set_aside ought to
withhold even more blocks from xfs_mod_fdblocks.

IOWS, I was wrong earlier -- we only need to withhold enough space from
fdblocks to handle splits of the bnobt and cntbt at or near ENOSPC.  The
value 4 is actually correct, but needs much better explanation.
Especially for benefit of the original author. ;)

> > I also think the second part ("and 4 more to handle a split of the
> > file's bmap btree") is wrong.  If we're really supposed to save enough
> > blocks to handle a bmbt split, then I think this ought to be
> > (mp->m_bm_maxlevels[0] - 1), not 4, right?  According to xfs_db, bmap
> > btrees can be 9 levels tall:
> 
> Yes, we've changed the BMBT reservations in the years since that
> code was written to handle max height reservations correctly, too.
> So, like the RMAP btree reservation, we probably should be reserving
> MAX_BMAP_HEIGHT in the set-aside calculation.

Right.

> refcount btree space is handled by the ag_resv code and blocks
> aren't allocated from the AGFL, so I don't think we need to take
> taht into account for xfs_alloc_set_aside().

Right.

> > So in the end, I think that calculation should become:
> > 
> > unsigned int
> > xfs_alloc_set_aside(
> > 	struct xfs_mount	*mp)
> > {
> > 	unsigned int		min-agfl = xfs_alloc_min_freelist(mp, NULL);
> > 
> > 	return mp->m_sb.sb_agcount * (min_agfl + mp->m_bm_maxlevels[0] - 1);
> > }
> 
> *nod*, but with the proviso that xfs_alloc_min_freelist() doesn't
> appear to be correct, either....
> 
> Also, that's a fixed value for the physical geometry of the
> filesystem, so it should be calculated once at mount time and stored
> in the xfs_mount (and only updated if needed at growfs time)...

There are three callers of xfs_alloc_min_freelist(, NULL) now.  One of
them is the the function that does the root inode calculation, which we
only use in mkfs and repair.

The other two are xfs_alloc_set_aside and xfs_alloc_ag_max_usable, and
we already cache the return value of those two functions, so I don't see
why we need to cache xfs_alloc_min_freelist separately?

(Or even touch it at all, really...)

--D

> > > new xfs_alloc_set_aside() that combines the macro calculation with
> > > ->m_allocbt_blks. Then the whole "set aside" concept is calculated and
> > > documented in one place. Hm?
> > 
> > I think I'd rather call the new function xfs_fdblocks_avail() over
> > reusing an existing name, because I fear that zapping an old function
> > and replacing it with a new function with the same name will cause
> > confusion for anyone backporting patches or reading code after an
> > absence.
> > 
> > Also the only reason we have a mount variable and a function (instead of
> > a macro) is that Dave asked me to change the codebase away from the
> > XFS_ALLOC_AG_MAX_USABLE/XFS_ALLOC_SET_ASIDE macros as part of merging
> > reflink.
> 
> Yeah, macros wrapping a variable or repeated constant calculation
> are bad, and it's something we've been cleaning up for a long
> time...
> 
> Cheers,
> 
> Dave.
> -- 
> Dave Chinner
> david@fromorbit.com

^ permalink raw reply	[flat|nested] 14+ messages in thread

end of thread, other threads:[~2022-03-17 17:05 UTC | newest]

Thread overview: 14+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2022-03-14 18:08 [PATCH] xfs: don't include bnobt blocks when reserving free block pool Darrick J. Wong
2022-03-16 11:28 ` Brian Foster
2022-03-16 16:32   ` Darrick J. Wong
2022-03-16 17:29     ` Brian Foster
2022-03-16 18:17       ` Darrick J. Wong
2022-03-16 18:48         ` Brian Foster
2022-03-16 19:17           ` Brian Foster
2022-03-16 21:15             ` Darrick J. Wong
2022-03-17  2:19               ` Dave Chinner
2022-03-17 12:53               ` Brian Foster
2022-03-17  2:05         ` Dave Chinner
2022-03-17 12:56           ` Brian Foster
2022-03-17 15:46             ` Darrick J. Wong
2022-03-17 17:05           ` Darrick J. Wong

This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.