linux-fsdevel.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: Josef Bacik <josef@toxicpanda.com>
To: Naohiro Aota <naohiro.aota@wdc.com>,
	linux-btrfs@vger.kernel.org, dsterba@suse.com
Cc: hare@suse.com, linux-fsdevel@vger.kernel.org
Subject: Re: [PATCH v9 12/41] btrfs: implement zoned chunk allocator
Date: Mon, 2 Nov 2020 15:09:58 -0500	[thread overview]
Message-ID: <54c78c27-37ad-7b42-7d0a-2b1b46c69dac@toxicpanda.com> (raw)
In-Reply-To: <5b9798f6e4c317e6a2c433ef88ffeabe00b93bb3.1604065695.git.naohiro.aota@wdc.com>

On 10/30/20 9:51 AM, Naohiro Aota wrote:
> This commit implements a zoned chunk/dev_extent allocator. The zoned
> allocator aligns the device extents to zone boundaries, so that a zone
> reset affects only the device extent and does not change the state of
> blocks in the neighbor device extents.
> 
> Also, it checks that a region allocation is not overlapping any of the
> super block zones, and ensures the region is empty.
> 
> Signed-off-by: Naohiro Aota <naohiro.aota@wdc.com>
> ---
>   fs/btrfs/volumes.c | 131 +++++++++++++++++++++++++++++++++++++++++++++
>   fs/btrfs/volumes.h |   1 +
>   fs/btrfs/zoned.c   | 126 +++++++++++++++++++++++++++++++++++++++++++
>   fs/btrfs/zoned.h   |  30 +++++++++++
>   4 files changed, 288 insertions(+)
> 
> diff --git a/fs/btrfs/volumes.c b/fs/btrfs/volumes.c
> index db884b96a5ea..78c62ef02e6f 100644
> --- a/fs/btrfs/volumes.c
> +++ b/fs/btrfs/volumes.c
> @@ -1416,6 +1416,14 @@ static bool contains_pending_extent(struct btrfs_device *device, u64 *start,
>   	return false;
>   }
>   
> +static inline u64 dev_extent_search_start_zoned(struct btrfs_device *device,
> +						u64 start)
> +{
> +	start = max_t(u64, start,
> +		      max_t(u64, device->zone_info->zone_size, SZ_1M));
> +	return btrfs_zone_align(device, start);
> +}
> +
>   static u64 dev_extent_search_start(struct btrfs_device *device, u64 start)
>   {
>   	switch (device->fs_devices->chunk_alloc_policy) {
> @@ -1426,11 +1434,57 @@ static u64 dev_extent_search_start(struct btrfs_device *device, u64 start)
>   		 * make sure to start at an offset of at least 1MB.
>   		 */
>   		return max_t(u64, start, SZ_1M);
> +	case BTRFS_CHUNK_ALLOC_ZONED:
> +		return dev_extent_search_start_zoned(device, start);
>   	default:
>   		BUG();
>   	}
>   }
>   
> +static bool dev_extent_hole_check_zoned(struct btrfs_device *device,
> +					u64 *hole_start, u64 *hole_size,
> +					u64 num_bytes)
> +{
> +	u64 zone_size = device->zone_info->zone_size;
> +	u64 pos;
> +	int ret;
> +	int changed = 0;
> +
> +	ASSERT(IS_ALIGNED(*hole_start, zone_size));
> +
> +	while (*hole_size > 0) {
> +		pos = btrfs_find_allocatable_zones(device, *hole_start,
> +						   *hole_start + *hole_size,
> +						   num_bytes);
> +		if (pos != *hole_start) {
> +			*hole_size = *hole_start + *hole_size - pos;
> +			*hole_start = pos;
> +			changed = 1;
> +			if (*hole_size < num_bytes)
> +				break;
> +		}
> +
> +		ret = btrfs_ensure_empty_zones(device, pos, num_bytes);
> +
> +		/* range is ensured to be empty */
> +		if (!ret)
> +			return changed;
> +
> +		/* given hole range was invalid (outside of device) */
> +		if (ret == -ERANGE) {
> +			*hole_start += *hole_size;
> +			*hole_size = 0;
> +			return 1;
> +		}
> +
> +		*hole_start += zone_size;
> +		*hole_size -= zone_size;
> +		changed = 1;
> +	}
> +
> +	return changed;
> +}
> +
>   /**
>    * dev_extent_hole_check - check if specified hole is suitable for allocation
>    * @device:	the device which we have the hole
> @@ -1463,6 +1517,10 @@ static bool dev_extent_hole_check(struct btrfs_device *device, u64 *hole_start,
>   	case BTRFS_CHUNK_ALLOC_REGULAR:
>   		/* No extra check */
>   		break;
> +	case BTRFS_CHUNK_ALLOC_ZONED:
> +		changed |= dev_extent_hole_check_zoned(device, hole_start,
> +						       hole_size, num_bytes);
I'm confused here, we check to make sure the pending stuff doesn't overlap with 
non-empty zones.  However we don't ever actually mark zones as non-empty except 
on mount.  I realize that if we allocate this zone then it appears pending and 
thus we won't allocate with this zone again while the fs is mounted, but it took 
me a while to realize this.  Is there a reason to not mark a zone as non-empty 
when we allocate from it?


> +		break;
>   	default:
>   		BUG();
>   	}
> @@ -1517,6 +1575,9 @@ static int find_free_dev_extent_start(struct btrfs_device *device,
>   
>   	search_start = dev_extent_search_start(device, search_start);
>   
> +	WARN_ON(device->zone_info &&
> +		!IS_ALIGNED(num_bytes, device->zone_info->zone_size));
> +
>   	path = btrfs_alloc_path();
>   	if (!path)
>   		return -ENOMEM;
> @@ -4907,6 +4968,37 @@ static void init_alloc_chunk_ctl_policy_regular(
>   	ctl->dev_extent_min = BTRFS_STRIPE_LEN * ctl->dev_stripes;
>   }
>   
> +static void
> +init_alloc_chunk_ctl_policy_zoned(struct btrfs_fs_devices *fs_devices,
> +				  struct alloc_chunk_ctl *ctl)
> +{
> +	u64 zone_size = fs_devices->fs_info->zone_size;
> +	u64 limit;
> +	int min_num_stripes = ctl->devs_min * ctl->dev_stripes;
> +	int min_data_stripes = (min_num_stripes - ctl->nparity) / ctl->ncopies;
> +	u64 min_chunk_size = min_data_stripes * zone_size;
> +	u64 type = ctl->type;
> +
> +	ctl->max_stripe_size = zone_size;
> +	if (type & BTRFS_BLOCK_GROUP_DATA) {
> +		ctl->max_chunk_size = round_down(BTRFS_MAX_DATA_CHUNK_SIZE,
> +						 zone_size);
> +	} else if (type & BTRFS_BLOCK_GROUP_METADATA) {
> +		ctl->max_chunk_size = ctl->max_stripe_size;
> +	} else if (type & BTRFS_BLOCK_GROUP_SYSTEM) {
> +		ctl->max_chunk_size = 2 * ctl->max_stripe_size;
> +		ctl->devs_max = min_t(int, ctl->devs_max,
> +				      BTRFS_MAX_DEVS_SYS_CHUNK);
> +	}
> +
> +	/* We don't want a chunk larger than 10% of writable space */
> +	limit = max(round_down(div_factor(fs_devices->total_rw_bytes, 1),
> +			       zone_size),
> +		    min_chunk_size);
> +	ctl->max_chunk_size = min(limit, ctl->max_chunk_size);
> +	ctl->dev_extent_min = zone_size * ctl->dev_stripes;
> +}
> +
>   static void init_alloc_chunk_ctl(struct btrfs_fs_devices *fs_devices,
>   				 struct alloc_chunk_ctl *ctl)
>   {
> @@ -4927,6 +5019,9 @@ static void init_alloc_chunk_ctl(struct btrfs_fs_devices *fs_devices,
>   	case BTRFS_CHUNK_ALLOC_REGULAR:
>   		init_alloc_chunk_ctl_policy_regular(fs_devices, ctl);
>   		break;
> +	case BTRFS_CHUNK_ALLOC_ZONED:
> +		init_alloc_chunk_ctl_policy_zoned(fs_devices, ctl);
> +		break;
>   	default:
>   		BUG();
>   	}
> @@ -5053,6 +5148,40 @@ static int decide_stripe_size_regular(struct alloc_chunk_ctl *ctl,
>   	return 0;
>   }
>   
> +static int decide_stripe_size_zoned(struct alloc_chunk_ctl *ctl,
> +				    struct btrfs_device_info *devices_info)
> +{
> +	u64 zone_size = devices_info[0].dev->zone_info->zone_size;
> +	/* number of stripes that count for block group size */
> +	int data_stripes;
> +
> +	/*
> +	 * It should hold because:
> +	 *    dev_extent_min == dev_extent_want == zone_size * dev_stripes
> +	 */
> +	ASSERT(devices_info[ctl->ndevs - 1].max_avail == ctl->dev_extent_min);
> +
> +	ctl->stripe_size = zone_size;
> +	ctl->num_stripes = ctl->ndevs * ctl->dev_stripes;
> +	data_stripes = (ctl->num_stripes - ctl->nparity) / ctl->ncopies;
> +
> +	/*
> +	 * stripe_size is fixed in ZONED. Reduce ndevs instead.
> +	 */
> +	if (ctl->stripe_size * data_stripes > ctl->max_chunk_size) {
> +		ctl->ndevs = div_u64(div_u64(ctl->max_chunk_size * ctl->ncopies,
> +					     ctl->stripe_size) + ctl->nparity,
> +				     ctl->dev_stripes);
> +		ctl->num_stripes = ctl->ndevs * ctl->dev_stripes;
> +		data_stripes = (ctl->num_stripes - ctl->nparity) / ctl->ncopies;
> +		ASSERT(ctl->stripe_size * data_stripes <= ctl->max_chunk_size);
> +	}
> +
> +	ctl->chunk_size = ctl->stripe_size * data_stripes;
> +
> +	return 0;
> +}
> +
>   static int decide_stripe_size(struct btrfs_fs_devices *fs_devices,
>   			      struct alloc_chunk_ctl *ctl,
>   			      struct btrfs_device_info *devices_info)
> @@ -5080,6 +5209,8 @@ static int decide_stripe_size(struct btrfs_fs_devices *fs_devices,
>   	switch (fs_devices->chunk_alloc_policy) {
>   	case BTRFS_CHUNK_ALLOC_REGULAR:
>   		return decide_stripe_size_regular(ctl, devices_info);
> +	case BTRFS_CHUNK_ALLOC_ZONED:
> +		return decide_stripe_size_zoned(ctl, devices_info);
>   	default:
>   		BUG();
>   	}
> diff --git a/fs/btrfs/volumes.h b/fs/btrfs/volumes.h
> index 9c07b97a2260..0249aca668fb 100644
> --- a/fs/btrfs/volumes.h
> +++ b/fs/btrfs/volumes.h
> @@ -213,6 +213,7 @@ BTRFS_DEVICE_GETSET_FUNCS(bytes_used);
>   
>   enum btrfs_chunk_allocation_policy {
>   	BTRFS_CHUNK_ALLOC_REGULAR,
> +	BTRFS_CHUNK_ALLOC_ZONED,
>   };
>   
>   struct btrfs_fs_devices {
> diff --git a/fs/btrfs/zoned.c b/fs/btrfs/zoned.c
> index d5487cba203b..4411d786597a 100644
> --- a/fs/btrfs/zoned.c
> +++ b/fs/btrfs/zoned.c
> @@ -1,11 +1,13 @@
>   // SPDX-License-Identifier: GPL-2.0
>   
> +#include <linux/bitops.h>
>   #include <linux/slab.h>
>   #include <linux/blkdev.h>
>   #include "ctree.h"
>   #include "volumes.h"
>   #include "zoned.h"
>   #include "rcu-string.h"
> +#include "disk-io.h"
>   
>   /* Maximum number of zones to report per blkdev_report_zones() call */
>   #define BTRFS_REPORT_NR_ZONES   4096
> @@ -328,6 +330,7 @@ int btrfs_check_zoned_mode(struct btrfs_fs_info *fs_info)
>   
>   	fs_info->zone_size = zone_size;
>   	fs_info->max_zone_append_size = max_zone_append_size;
> +	fs_info->fs_devices->chunk_alloc_policy = BTRFS_CHUNK_ALLOC_ZONED;
>   
>   	btrfs_info(fs_info, "ZONED mode enabled, zone size %llu B",
>   		   fs_info->zone_size);
> @@ -607,3 +610,126 @@ int btrfs_reset_sb_log_zones(struct block_device *bdev, int mirror)
>   				sb_zone << zone_sectors_shift, zone_sectors * 2,
>   				GFP_NOFS);
>   }
> +
> +/*
> + * btrfs_check_allocatable_zones - find allocatable zones within give region
> + * @device:	the device to allocate a region
> + * @hole_start: the position of the hole to allocate the region
> + * @num_bytes:	the size of wanted region
> + * @hole_size:	the size of hole
> + *
> + * Allocatable region should not contain any superblock locations.
> + */
> +u64 btrfs_find_allocatable_zones(struct btrfs_device *device, u64 hole_start,
> +				 u64 hole_end, u64 num_bytes)
> +{
> +	struct btrfs_zoned_device_info *zinfo = device->zone_info;
> +	u8 shift = zinfo->zone_size_shift;
> +	u64 nzones = num_bytes >> shift;
> +	u64 pos = hole_start;
> +	u64 begin, end;
> +	u64 sb_pos;
> +	bool have_sb;
> +	int i;
> +
> +	ASSERT(IS_ALIGNED(hole_start, zinfo->zone_size));
> +	ASSERT(IS_ALIGNED(num_bytes, zinfo->zone_size));
> +
> +	while (pos < hole_end) {
> +		begin = pos >> shift;
> +		end = begin + nzones;
> +
> +		if (end > zinfo->nr_zones)
> +			return hole_end;
> +
> +		/* check if zones in the region are all empty */
> +		if (btrfs_dev_is_sequential(device, pos) &&
> +		    find_next_zero_bit(zinfo->empty_zones, end, begin) != end) {
> +			pos += zinfo->zone_size;
> +			continue;
> +		}
> +
> +		have_sb = false;
> +		for (i = 0; i < BTRFS_SUPER_MIRROR_MAX; i++) {
> +			sb_pos = sb_zone_number(zinfo->zone_size, i);
> +			if (!(end < sb_pos || sb_pos + 1 < begin)) {
> +				have_sb = true;
> +				pos = (sb_pos + 2) << shift;
> +				break;
> +			}
> +		}
> +		if (!have_sb)
> +			break;
> +	}
> +
> +	return pos;
> +}
> +
> +int btrfs_reset_device_zone(struct btrfs_device *device, u64 physical,
> +			    u64 length, u64 *bytes)
> +{
> +	int ret;
> +
> +	*bytes = 0;
> +	ret = blkdev_zone_mgmt(device->bdev, REQ_OP_ZONE_RESET,
> +			       physical >> SECTOR_SHIFT, length >> SECTOR_SHIFT,
> +			       GFP_NOFS);
> +	if (ret)
> +		return ret;
> +
> +	*bytes = length;
> +	while (length) {
> +		btrfs_dev_set_zone_empty(device, physical);
> +		physical += device->zone_info->zone_size;
> +		length -= device->zone_info->zone_size;
> +	}
> +
> +	return 0;
> +}
> +
> +int btrfs_ensure_empty_zones(struct btrfs_device *device, u64 start, u64 size)
> +{
> +	struct btrfs_zoned_device_info *zinfo = device->zone_info;
> +	u8 shift = zinfo->zone_size_shift;
> +	unsigned long begin = start >> shift;
> +	unsigned long end = (start + size) >> shift;
> +	u64 pos;
> +	int ret;
> +
> +	ASSERT(IS_ALIGNED(start, zinfo->zone_size));
> +	ASSERT(IS_ALIGNED(size, zinfo->zone_size));
> +
> +	if (end > zinfo->nr_zones)
> +		return -ERANGE;
> +
> +	/* all the zones are conventional */
> +	if (find_next_bit(zinfo->seq_zones, begin, end) == end)
> +		return 0;
> +

This check is duplicated below.

> +	/* all the zones are sequential and empty */
> +	if (find_next_zero_bit(zinfo->seq_zones, begin, end) == end &&
> +	    find_next_zero_bit(zinfo->empty_zones, begin, end) == end)
> +		return 0;
> +
> +	for (pos = start; pos < start + size; pos += zinfo->zone_size) {
> +		u64 reset_bytes;
> +
> +		if (!btrfs_dev_is_sequential(device, pos) ||
> +		    btrfs_dev_is_empty_zone(device, pos))
> +			continue;
> +
> +		/* free regions should be empty */
> +		btrfs_warn_in_rcu(
> +			device->fs_info,
> +			"resetting device %s zone %llu for allocation",
> +			rcu_str_deref(device->name), pos >> shift);
> +		WARN_ON_ONCE(1);
> +
> +		ret = btrfs_reset_device_zone(device, pos, zinfo->zone_size,
> +					      &reset_bytes);
> +		if (ret)
> +			return ret;

This seems bad, as we could just have corruption right?  So we're resetting the 
zone which could lose us data right?  Shouldn't we just bail here?  Thanks,

Josef

  reply	other threads:[~2020-11-02 20:10 UTC|newest]

Thread overview: 119+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2020-10-30 13:51 [PATCH v9 00/41] btrfs: zoned block device support Naohiro Aota
2020-10-30 13:51 ` [PATCH v9 01/41] block: add bio_add_zone_append_page Naohiro Aota
2020-10-30 13:51   ` [PATCH v9 02/41] iomap: support REQ_OP_ZONE_APPEND Naohiro Aota
2020-11-02  5:34     ` Naohiro Aota
2020-11-02 16:55     ` Darrick J. Wong
2020-11-02 17:39       ` Johannes Thumshirn
2020-10-30 13:51   ` [PATCH v9 03/41] btrfs: introduce ZONED feature flag Naohiro Aota
2020-10-30 13:51   ` [PATCH v9 04/41] btrfs: Get zone information of zoned block devices Naohiro Aota
2020-11-02 16:53     ` Josef Bacik
2020-11-02 16:58       ` Johannes Thumshirn
2020-11-02 21:07       ` Naohiro Aota
2020-11-03 12:02     ` David Sterba
2020-10-30 13:51   ` [PATCH v9 05/41] btrfs: Check and enable ZONED mode Naohiro Aota
2020-11-03 12:13     ` David Sterba
2020-11-06  9:36       ` Naohiro Aota
2020-10-30 13:51   ` [PATCH v9 06/41] btrfs: introduce max_zone_append_size Naohiro Aota
2020-11-02 16:57     ` Josef Bacik
2020-11-03 12:16     ` David Sterba
2020-10-30 13:51   ` [PATCH v9 07/41] btrfs: disallow space_cache in ZONED mode Naohiro Aota
2020-11-02 17:02     ` Josef Bacik
2020-11-02 17:37       ` Johannes Thumshirn
2020-11-03 12:48     ` David Sterba
2020-11-10 10:14       ` Naohiro Aota
2020-10-30 13:51   ` [PATCH v9 08/41] btrfs: disallow NODATACOW " Naohiro Aota
2020-11-02 17:05     ` Josef Bacik
2020-11-03 12:57     ` David Sterba
2020-10-30 13:51   ` [PATCH v9 09/41] btrfs: disable fallocate " Naohiro Aota
2020-11-03 13:00     ` David Sterba
2020-10-30 13:51   ` [PATCH v9 10/41] btrfs: disallow mixed-bg " Naohiro Aota
2020-11-03 13:01     ` David Sterba
2020-10-30 13:51   ` [PATCH v9 11/41] btrfs: implement log-structured superblock for " Naohiro Aota
2020-11-02 18:22     ` Josef Bacik
2020-11-02 18:53       ` Johannes Thumshirn
2020-11-02 19:01         ` Josef Bacik
2020-11-02 19:31           ` Johannes Thumshirn
2020-11-03  8:21       ` Naohiro Aota
2020-11-02 18:54     ` Josef Bacik
2020-11-03  3:31       ` Naohiro Aota
2020-11-03 13:15     ` David Sterba
2020-11-03 14:10     ` David Sterba
2020-11-06 10:37       ` Naohiro Aota
2020-10-30 13:51   ` [PATCH v9 12/41] btrfs: implement zoned chunk allocator Naohiro Aota
2020-11-02 20:09     ` Josef Bacik [this message]
2020-11-02 22:21       ` Naohiro Aota
2020-11-03 13:23     ` David Sterba
2020-10-30 13:51   ` [PATCH v9 13/41] btrfs: verify device extent is aligned to zone Naohiro Aota
2020-11-02 20:14     ` Josef Bacik
2020-10-30 13:51   ` [PATCH v9 14/41] btrfs: load zone's alloction offset Naohiro Aota
2020-11-02 20:25     ` Josef Bacik
2020-11-02 20:29       ` Josef Bacik
2020-11-02 22:43         ` Naohiro Aota
2020-11-02 22:40       ` Naohiro Aota
2020-11-03 13:28     ` David Sterba
2020-10-30 13:51   ` [PATCH v9 15/41] btrfs: emulate write pointer for conventional zones Naohiro Aota
2020-11-02 20:37     ` Josef Bacik
2020-11-03  1:25       ` Naohiro Aota
2020-11-03 13:32     ` David Sterba
2020-10-30 13:51   ` [PATCH v9 16/41] btrfs: track unusable bytes for zones Naohiro Aota
2020-11-03 14:25     ` Josef Bacik
2020-10-30 13:51   ` [PATCH v9 17/41] btrfs: do sequential extent allocation in ZONED mode Naohiro Aota
2020-11-03 14:28     ` Josef Bacik
2020-10-30 13:51   ` [PATCH v9 18/41] btrfs: reset zones of unused block groups Naohiro Aota
2020-11-03 14:34     ` Josef Bacik
2020-11-10 10:40       ` Naohiro Aota
2020-10-30 13:51   ` [PATCH v9 19/41] btrfs: redirty released extent buffers in ZONED mode Naohiro Aota
2020-11-03 14:41     ` Josef Bacik
2020-11-06  9:11       ` Johannes Thumshirn
2020-11-06 15:01         ` Josef Bacik
2020-10-30 13:51   ` [PATCH v9 20/41] btrfs: extract page adding function Naohiro Aota
2020-11-03 14:45     ` Josef Bacik
2020-10-30 13:51   ` [PATCH v9 21/41] btrfs: use bio_add_zone_append_page for zoned btrfs Naohiro Aota
2020-11-03 14:55     ` Josef Bacik
2020-11-10 10:42       ` Naohiro Aota
2020-10-30 13:51   ` [PATCH v9 22/41] btrfs: handle REQ_OP_ZONE_APPEND as writing Naohiro Aota
2020-11-03 14:57     ` Josef Bacik
2020-10-30 13:51   ` [PATCH v9 23/41] btrfs: split ordered extent when bio is sent Naohiro Aota
2020-11-03 15:29     ` Josef Bacik
2020-10-30 13:51   ` [PATCH v9 24/41] btrfs: extend btrfs_rmap_block for specifying a device Naohiro Aota
2020-11-03 15:32     ` Josef Bacik
2020-11-06 10:52       ` Johannes Thumshirn
2020-10-30 13:51   ` [PATCH v9 25/41] btrfs: use ZONE_APPEND write for ZONED btrfs Naohiro Aota
2020-11-03 15:55     ` Josef Bacik
2020-10-30 13:51   ` [PATCH v9 26/41] btrfs: enable zone append writing for direct IO Naohiro Aota
2020-11-03 15:56     ` Josef Bacik
2020-10-30 13:51   ` [PATCH v9 27/41] btrfs: introduce dedicated data write path for ZONED mode Naohiro Aota
2020-11-03 15:57     ` Josef Bacik
2020-10-30 13:51   ` [PATCH v9 28/41] btrfs: serialize meta IOs on " Naohiro Aota
2020-11-03 16:04     ` Josef Bacik
2020-10-30 13:51   ` [PATCH v9 29/41] btrfs: wait existing extents before truncating Naohiro Aota
2020-10-30 13:51   ` [PATCH v9 30/41] btrfs: avoid async metadata checksum on ZONED mode Naohiro Aota
2020-11-03 16:05     ` Josef Bacik
2020-10-30 13:51   ` [PATCH v9 31/41] btrfs: mark block groups to copy for device-replace Naohiro Aota
2020-11-03 17:09     ` Josef Bacik
2020-10-30 13:51   ` [PATCH v9 32/41] btrfs: implement cloning for ZONED device-replace Naohiro Aota
2020-11-03 17:15     ` Josef Bacik
2020-10-30 13:51   ` [PATCH v9 33/41] btrfs: implement copying " Naohiro Aota
2020-11-03 17:19     ` Josef Bacik
2020-11-10 11:09       ` Naohiro Aota
2020-10-30 13:51   ` [PATCH v9 34/41] btrfs: support dev-replace in ZONED mode Naohiro Aota
2020-11-03 20:34     ` Josef Bacik
2020-10-30 13:51   ` [PATCH v9 35/41] btrfs: enable relocation " Naohiro Aota
2020-11-03 20:39     ` Josef Bacik
2020-10-30 13:51   ` [PATCH v9 36/41] btrfs: relocate block group to repair IO failure in ZONED Naohiro Aota
2020-10-30 13:51   ` [PATCH v9 37/41] btrfs: split alloc_log_tree() Naohiro Aota
2020-11-03 20:42     ` Josef Bacik
2020-11-03 22:10       ` Amy Parker
2020-11-10 11:12         ` Naohiro Aota
2020-10-30 13:51   ` [PATCH v9 38/41] btrfs: extend zoned allocator to use dedicated tree-log block group Naohiro Aota
2020-11-03 20:47     ` Josef Bacik
2020-11-10  6:37       ` Naohiro Aota
2020-10-30 13:51   ` [PATCH v9 39/41] btrfs: serialize log transaction on ZONED mode Naohiro Aota
2020-11-03 20:49     ` Josef Bacik
2020-10-30 13:51   ` [PATCH v9 40/41] btrfs: reorder log node allocation Naohiro Aota
2020-11-03 20:49     ` Josef Bacik
2020-10-30 13:51   ` [PATCH v9 41/41] btrfs: enable to mount ZONED incompat flag Naohiro Aota
2020-10-31  3:40   ` [PATCH v9 01/41] block: add bio_add_zone_append_page Jens Axboe
2020-11-02  5:15     ` Naohiro Aota
2020-11-02  8:24     ` Johannes Thumshirn
2020-11-03 11:54 ` [PATCH v9 00/41] btrfs: zoned block device support David Sterba

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=54c78c27-37ad-7b42-7d0a-2b1b46c69dac@toxicpanda.com \
    --to=josef@toxicpanda.com \
    --cc=dsterba@suse.com \
    --cc=hare@suse.com \
    --cc=linux-btrfs@vger.kernel.org \
    --cc=linux-fsdevel@vger.kernel.org \
    --cc=naohiro.aota@wdc.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).