All of lore.kernel.org
 help / color / mirror / Atom feed
* [PATCH 00/10] Freespace tree repair support v2
@ 2018-10-01 14:46 Nikolay Borisov
  2018-10-01 14:46 ` [PATCH 01/10] btrfs-progs: Add support for freespace tree in btrfs_read_fs_root Nikolay Borisov
                   ` (10 more replies)
  0 siblings, 11 replies; 27+ messages in thread
From: Nikolay Borisov @ 2018-10-01 14:46 UTC (permalink / raw)
  To: linux-btrfs; +Cc: Nikolay Borisov

Hello, 

Here is the v2 of the freespace tree repair support patches. Version 1 can be 
found at [0]. For background on the series please refer to the initial cover 
letter. This time round a number of changes have been incorporated based on 
feedback from Omar. Namely: 

1. Added 2 new patches, patches 3 and 4,  which refactor the userspace
impelementation of various bit manipulation functions. This was required to
implement the bitmap conversion code from kernel space. 

2. Extended patch 5 to include the bitmap conversion code, now userspace 
handles the case of fragmented fst. 

3. Re-ordered patches 6 and 7 so that we first hook into the extent 
alloc/dealloc sequence (patch 6) and finally enable support for fst in patch 7.

4. Updated the FST test (patch 9) to work with the new bitmap conversion code 
since it changed the layout of the expected freespace tree. Now the test 
passes again. Also removed an unused function. 

5. Added patch 10 which fixes a similar bug to e444c7bfa65f ("btrfs-progs:
check: Fix wrong error message in case of corrupted extent")

Following the updates, I rerun all misc/fsck-tests and no failures were 
observed. 

More feedback and suggestions are always welcomed.

[0] https://lore.kernel.org/linux-btrfs/1529060762-4372-1-git-send-email-nborisov@suse.com/


Nikolay Borisov (10):
  btrfs-progs: Add support for freespace tree in btrfs_read_fs_root
  btrfs-progs: Add extent buffer bitmap manipulation infrastructure
  btrfs-progs: Replace homegrown bitops related functions with kernel
    counterparts
  btrfs-progs: Implement find_*_bit_le operations
  btrfs-progs: Pull free space tree related code from kernel
  btrfs-progs: Hook FST code in extent (de)alloc
  btrfs-progs: Add freespace tree as compat_ro supported feature
  btrfs-progs: check: Add support for freespace tree fixing
  btrfs-progs: tests: Test for FST corruption detection/repair
  btrfs-progs: check: Fix wrong error message in case of corrupted
    bitmap

 check/main.c                                      |   47 +-
 ctree.c                                           |   77 ++
 ctree.h                                           |   19 +-
 disk-io.c                                         |    3 +
 extent-tree.c                                     |   11 +
 extent_io.c                                       |   56 +
 extent_io.h                                       |    4 +
 free-space-tree.c                                 | 1255 ++++++++++++++++++++-
 free-space-tree.h                                 |   13 +-
 kerncompat.h                                      |    6 +
 kernel-lib/bitops.h                               |  210 ++--
 tests/fsck-tests/036-freespacetree-repair/test.sh |   76 ++
 12 files changed, 1662 insertions(+), 115 deletions(-)
 create mode 100755 tests/fsck-tests/036-freespacetree-repair/test.sh

-- 
2.7.4


^ permalink raw reply	[flat|nested] 27+ messages in thread

* [PATCH 01/10] btrfs-progs: Add support for freespace tree in btrfs_read_fs_root
  2018-10-01 14:46 [PATCH 00/10] Freespace tree repair support v2 Nikolay Borisov
@ 2018-10-01 14:46 ` Nikolay Borisov
  2018-10-02 19:20   ` Omar Sandoval
  2018-10-04  1:05   ` Su Yue
  2018-10-01 14:46 ` [PATCH 02/10] btrfs-progs: Add extent buffer bitmap manipulation infrastructure Nikolay Borisov
                   ` (9 subsequent siblings)
  10 siblings, 2 replies; 27+ messages in thread
From: Nikolay Borisov @ 2018-10-01 14:46 UTC (permalink / raw)
  To: linux-btrfs; +Cc: Nikolay Borisov

For completeness sake add code to btrfs_read_fs_root so that it can
handle the freespace tree.

Signed-off-by: Nikolay Borisov <nborisov@suse.com>
---
 disk-io.c | 3 +++
 1 file changed, 3 insertions(+)

diff --git a/disk-io.c b/disk-io.c
index 2e6d56a36af9..14f0fd5c2f0c 100644
--- a/disk-io.c
+++ b/disk-io.c
@@ -668,6 +668,9 @@ struct btrfs_root *btrfs_read_fs_root(struct btrfs_fs_info *fs_info,
 	if (location->objectid == BTRFS_QUOTA_TREE_OBJECTID)
 		return fs_info->quota_enabled ? fs_info->quota_root :
 				ERR_PTR(-ENOENT);
+	if (location->objectid == BTRFS_FREE_SPACE_TREE_OBJECTID)
+        return fs_info->free_space_root ? fs_info->free_space_root :
+                                                  ERR_PTR(-ENOENT);
 
 	BUG_ON(location->objectid == BTRFS_TREE_RELOC_OBJECTID ||
 	       location->offset != (u64)-1);
-- 
2.7.4


^ permalink raw reply related	[flat|nested] 27+ messages in thread

* [PATCH 02/10] btrfs-progs: Add extent buffer bitmap manipulation infrastructure
  2018-10-01 14:46 [PATCH 00/10] Freespace tree repair support v2 Nikolay Borisov
  2018-10-01 14:46 ` [PATCH 01/10] btrfs-progs: Add support for freespace tree in btrfs_read_fs_root Nikolay Borisov
@ 2018-10-01 14:46 ` Nikolay Borisov
  2018-10-02 19:24   ` Omar Sandoval
  2018-10-04  1:31   ` Su Yue
  2018-10-01 14:46 ` [PATCH 03/10] btrfs-progs: Replace homegrown bitops related functions with kernel counterparts Nikolay Borisov
                   ` (8 subsequent siblings)
  10 siblings, 2 replies; 27+ messages in thread
From: Nikolay Borisov @ 2018-10-01 14:46 UTC (permalink / raw)
  To: linux-btrfs; +Cc: Nikolay Borisov

Those functions are in preparation for adding the freespace tree
repair code since it needs to be able to deal with bitmap based fsts.
This patch adds extent_buffer_bitmap_set and extent_buffer_bitmap_clear
functions. Since in userspace we don't have to deal with page mappings
their implementation is vastly simplified by simply setting each bit in
the passed range.

Signed-off-by: Nikolay Borisov <nborisov@suse.com>
---
 extent_io.c | 56 ++++++++++++++++++++++++++++++++++++++++++++++++++++++++
 extent_io.h |  4 ++++
 2 files changed, 60 insertions(+)

diff --git a/extent_io.c b/extent_io.c
index 198492699438..de47c2c59ae9 100644
--- a/extent_io.c
+++ b/extent_io.c
@@ -204,6 +204,62 @@ static int clear_state_bit(struct extent_io_tree *tree,
 	return ret;
 }
 
+/**
+ * extent_buffer_bitmap_set - set an area of a bitmap
+ * @eb: the extent buffer
+ * @start: offset of the bitmap item in the extent buffer
+ * @pos: bit number of the first bit
+ * @len: number of bits to set
+ */
+void extent_buffer_bitmap_set(struct extent_buffer *eb, unsigned long start,
+                              unsigned long pos, unsigned long len)
+{
+	u8 *p = (u8 *)eb->data + start + BIT_BYTE(pos);
+	const unsigned int size = pos + len;
+	int bits_to_set = BITS_PER_BYTE - (pos % BITS_PER_BYTE);
+	u8 mask_to_set = BITMAP_FIRST_BYTE_MASK(pos);
+
+	while (len >= bits_to_set) {
+		*p |= mask_to_set;
+		len -= bits_to_set;
+		bits_to_set = BITS_PER_BYTE;
+		mask_to_set = ~0;
+		p++;
+	}
+	if (len) {
+		mask_to_set &= BITMAP_LAST_BYTE_MASK(size);
+		*p |= mask_to_set;
+	}
+}
+
+
+/**
+ * extent_buffer_bitmap_clear - clear an area of a bitmap
+ * @eb: the extent buffer
+ * @start: offset of the bitmap item in the extent buffer
+ * @pos: bit number of the first bit
+ * @len: number of bits to clear
+ */
+void extent_buffer_bitmap_clear(struct extent_buffer *eb, unsigned long start,
+                                unsigned long pos, unsigned long len)
+{
+	u8 *p = (u8 *)eb->data + start + BIT_BYTE(pos);
+	const unsigned int size = pos + len;
+	int bits_to_clear = BITS_PER_BYTE - (pos % BITS_PER_BYTE);
+	u8 mask_to_clear = BITMAP_FIRST_BYTE_MASK(pos);
+
+	while (len >= bits_to_clear) {
+		*p &= ~mask_to_clear;
+		len -= bits_to_clear;
+		bits_to_clear = BITS_PER_BYTE;
+		mask_to_clear = ~0;
+		p++;
+	}
+	if (len) {
+		mask_to_clear &= BITMAP_LAST_BYTE_MASK(size);
+		*p &= ~mask_to_clear;
+	}
+}
 /*
  * clear some bits on a range in the tree.
  */
diff --git a/extent_io.h b/extent_io.h
index d407d93d617e..b67c6fc40e89 100644
--- a/extent_io.h
+++ b/extent_io.h
@@ -175,4 +175,8 @@ int read_data_from_disk(struct btrfs_fs_info *info, void *buf, u64 offset,
 			u64 bytes, int mirror);
 int write_data_to_disk(struct btrfs_fs_info *info, void *buf, u64 offset,
 		       u64 bytes, int mirror);
+void extent_buffer_bitmap_clear(struct extent_buffer *eb, unsigned long start,
+                                unsigned long pos, unsigned long len);
+void extent_buffer_bitmap_set(struct extent_buffer *eb, unsigned long start,
+                              unsigned long pos, unsigned long len);
 #endif
-- 
2.7.4


^ permalink raw reply related	[flat|nested] 27+ messages in thread

* [PATCH 03/10] btrfs-progs: Replace homegrown bitops related functions with kernel counterparts
  2018-10-01 14:46 [PATCH 00/10] Freespace tree repair support v2 Nikolay Borisov
  2018-10-01 14:46 ` [PATCH 01/10] btrfs-progs: Add support for freespace tree in btrfs_read_fs_root Nikolay Borisov
  2018-10-01 14:46 ` [PATCH 02/10] btrfs-progs: Add extent buffer bitmap manipulation infrastructure Nikolay Borisov
@ 2018-10-01 14:46 ` Nikolay Borisov
  2018-10-02 23:32   ` Omar Sandoval
  2018-10-01 14:46 ` [PATCH 04/10] btrfs-progs: Implement find_*_bit_le operations Nikolay Borisov
                   ` (7 subsequent siblings)
  10 siblings, 1 reply; 27+ messages in thread
From: Nikolay Borisov @ 2018-10-01 14:46 UTC (permalink / raw)
  To: linux-btrfs; +Cc: Nikolay Borisov

Replace existing find_*_bit functions with kernel equivalent. This
reduces duplication, simplifies the code (we really have one worker
function _find_next_bit) and is quite likely faster. No functional
changes.

Signed-off-by: Nikolay Borisov <nborisov@suse.com>
---
 kernel-lib/bitops.h | 142 +++++++++++++++++-----------------------------------
 1 file changed, 46 insertions(+), 96 deletions(-)

diff --git a/kernel-lib/bitops.h b/kernel-lib/bitops.h
index 5b35f9fc5213..78256adf55be 100644
--- a/kernel-lib/bitops.h
+++ b/kernel-lib/bitops.h
@@ -2,6 +2,7 @@
 #define _PERF_LINUX_BITOPS_H_
 
 #include <linux/kernel.h>
+#include "internal.h"
 
 #ifndef DIV_ROUND_UP
 #define DIV_ROUND_UP(n, d) (((n) + (d) - 1) / (d))
@@ -109,116 +110,65 @@ static __always_inline unsigned long __ffs(unsigned long word)
 
 #define ffz(x) __ffs(~(x))
 
+#define BITMAP_FIRST_WORD_MASK(start) (~0UL << ((start) & (BITS_PER_LONG - 1))) 
+#define BITMAP_LAST_WORD_MASK(nbits) (~0UL >> (-(nbits) & (BITS_PER_LONG - 1)))
+
 /*
- * Find the first set bit in a memory region.
+ * This is a common helper function for find_next_bit, find_next_zero_bit, and
+ * find_next_and_bit. The differences are:
+ *  - The "invert" argument, which is XORed with each fetched word before
+ *    searching it for one bits.
+ *  - The optional "addr2", which is anded with "addr1" if present.
  */
-static inline unsigned long
-find_first_bit(const unsigned long *addr, unsigned long size)
+static inline unsigned long _find_next_bit(const unsigned long *addr1,
+		const unsigned long *addr2, unsigned long nbits,
+		unsigned long start, unsigned long invert)
 {
-	const unsigned long *p = addr;
-	unsigned long result = 0;
 	unsigned long tmp;
 
-	while (size & ~(BITS_PER_LONG-1)) {
-		if ((tmp = *(p++)))
-			goto found;
-		result += BITS_PER_LONG;
-		size -= BITS_PER_LONG;
+	if (start >= nbits)
+		return nbits;
+
+	tmp = addr1[start / BITS_PER_LONG];
+	if (addr2)
+		tmp &= addr2[start / BITS_PER_LONG];
+	tmp ^= invert;
+
+	/* Handle 1st word. */
+	tmp &= BITMAP_FIRST_WORD_MASK(start);
+	start = round_down(start, BITS_PER_LONG);
+
+	while (!tmp) {
+		start += BITS_PER_LONG;
+		if (start >= nbits)
+			return nbits;
+
+		tmp = addr1[start / BITS_PER_LONG];
+		if (addr2)
+			tmp &= addr2[start / BITS_PER_LONG];
+		tmp ^= invert;
 	}
-	if (!size)
-		return result;
-
-	tmp = (*p) & (~0UL >> (BITS_PER_LONG - size));
-	if (tmp == 0UL)		/* Are any bits set? */
-		return result + size;	/* Nope. */
-found:
-	return result + __ffs(tmp);
+
+	return min(start + __ffs(tmp), nbits);
 }
 
 /*
  * Find the next set bit in a memory region.
  */
-static inline unsigned long
-find_next_bit(const unsigned long *addr, unsigned long size,
-	      unsigned long offset)
+static inline unsigned long find_next_bit(const unsigned long *addr,
+					  unsigned long size,
+					  unsigned long offset)
 {
-	const unsigned long *p = addr + BITOP_WORD(offset);
-	unsigned long result = offset & ~(BITS_PER_LONG-1);
-	unsigned long tmp;
-
-	if (offset >= size)
-		return size;
-	size -= result;
-	offset %= BITS_PER_LONG;
-	if (offset) {
-		tmp = *(p++);
-		tmp &= (~0UL << offset);
-		if (size < BITS_PER_LONG)
-			goto found_first;
-		if (tmp)
-			goto found_middle;
-		size -= BITS_PER_LONG;
-		result += BITS_PER_LONG;
-	}
-	while (size & ~(BITS_PER_LONG-1)) {
-		if ((tmp = *(p++)))
-			goto found_middle;
-		result += BITS_PER_LONG;
-		size -= BITS_PER_LONG;
-	}
-	if (!size)
-		return result;
-	tmp = *p;
-
-found_first:
-	tmp &= (~0UL >> (BITS_PER_LONG - size));
-	if (tmp == 0UL)		/* Are any bits set? */
-		return result + size;	/* Nope. */
-found_middle:
-	return result + __ffs(tmp);
+	return _find_next_bit(addr, NULL, size, offset, 0UL);
 }
 
-/*
- * This implementation of find_{first,next}_zero_bit was stolen from
- * Linus' asm-alpha/bitops.h.
- */
-static inline unsigned long
-find_next_zero_bit(const unsigned long *addr, unsigned long size,
-		   unsigned long offset)
+static inline unsigned long find_next_zero_bit(const unsigned long *addr,
+					       unsigned long size,
+					       unsigned long offset)
 {
-	const unsigned long *p = addr + BITOP_WORD(offset);
-	unsigned long result = offset & ~(BITS_PER_LONG-1);
-	unsigned long tmp;
-
-	if (offset >= size)
-		return size;
-	size -= result;
-	offset %= BITS_PER_LONG;
-	if (offset) {
-		tmp = *(p++);
-		tmp |= ~0UL >> (BITS_PER_LONG - offset);
-		if (size < BITS_PER_LONG)
-			goto found_first;
-		if (~tmp)
-			goto found_middle;
-		size -= BITS_PER_LONG;
-		result += BITS_PER_LONG;
-	}
-	while (size & ~(BITS_PER_LONG-1)) {
-		if (~(tmp = *(p++)))
-			goto found_middle;
-		result += BITS_PER_LONG;
-		size -= BITS_PER_LONG;
-	}
-	if (!size)
-		return result;
-	tmp = *p;
-
-found_first:
-	tmp |= ~0UL << size;
-	if (tmp == ~0UL)	/* Are any bits zero? */
-		return result + size;	/* Nope. */
-found_middle:
-	return result + ffz(tmp);
+	return _find_next_bit(addr, NULL, size, offset, ~0UL);
 }
+
+#define find_first_bit(addr, size) find_next_bit((addr), (size), 0)
+
 #endif
-- 
2.7.4


^ permalink raw reply related	[flat|nested] 27+ messages in thread

* [PATCH 04/10] btrfs-progs: Implement find_*_bit_le operations
  2018-10-01 14:46 [PATCH 00/10] Freespace tree repair support v2 Nikolay Borisov
                   ` (2 preceding siblings ...)
  2018-10-01 14:46 ` [PATCH 03/10] btrfs-progs: Replace homegrown bitops related functions with kernel counterparts Nikolay Borisov
@ 2018-10-01 14:46 ` Nikolay Borisov
  2018-10-04 18:08   ` Omar Sandoval
  2018-10-01 14:46 ` [PATCH 05/10] btrfs-progs: Pull free space tree related code from kernel Nikolay Borisov
                   ` (6 subsequent siblings)
  10 siblings, 1 reply; 27+ messages in thread
From: Nikolay Borisov @ 2018-10-01 14:46 UTC (permalink / raw)
  To: linux-btrfs; +Cc: Nikolay Borisov

This commit introduces explicit little endian bit operations. The only
difference with the existing bitops implementation is that bswap(32|64)
is called when the _le versions are invoked on a big-endian machine.
This is in preparation for adding free space tree conversion support.

Signed-off-by: Nikolay Borisov <nborisov@suse.com>
---
 kernel-lib/bitops.h | 82 +++++++++++++++++++++++++++++++++++++++++++++++++++++
 1 file changed, 82 insertions(+)

diff --git a/kernel-lib/bitops.h b/kernel-lib/bitops.h
index 78256adf55be..5030bfa2815e 100644
--- a/kernel-lib/bitops.h
+++ b/kernel-lib/bitops.h
@@ -2,6 +2,7 @@
 #define _PERF_LINUX_BITOPS_H_
 
 #include <linux/kernel.h>
+#include <endian.h>
 #include "internal.h"
 
 #ifndef DIV_ROUND_UP
@@ -170,5 +171,86 @@ static inline unsigned long find_next_zero_bit(const unsigned long *addr,
 }
 
 #define find_first_bit(addr, size) find_next_bit((addr), (size), 0)
+#define find_first_zero_bit(addr, size) find_next_zero_bit((addr), (size), 0)
+
+#if __BYTE_ORDER == __BIG_ENDIAN
+
+static inline unsigned long ext2_swab(const unsigned long y)
+{
+#if BITS_PER_LONG == 64
+	return (unsigned long) bswap64((u64) y);
+#elif BITS_PER_LONG == 32
+	return (unsigned long) bswap32((u32) y);
+#else
+#error BITS_PER_LONG not defined
+#endif
+}
+
+static inline unsigned long _find_next_bit_le(const unsigned long *addr1,
+		const unsigned long *addr2, unsigned long nbits,
+		unsigned long start, unsigned long invert)
+{
+	unsigned long tmp;
+
+	if (start >= nbits)
+		return nbits;
+
+	tmp = addr1[start / BITS_PER_LONG];
+	if (addr2)
+		tmp &= addr2[start / BITS_PER_LONG];
+	tmp ^= invert;
+
+	/* Handle 1st word. */
+	tmp &= ext2_swab(BITMAP_FIRST_WORD_MASK(start));
+	start = round_down(start, BITS_PER_LONG);
+
+	while (!tmp) {
+		start += BITS_PER_LONG;
+		if (start >= nbits)
+			return nbits;
+
+		tmp = addr1[start / BITS_PER_LONG];
+		if (addr2)
+			tmp &= addr2[start / BITS_PER_LONG];
+		tmp ^= invert;
+	}
+
+	return min(start + __ffs(ext2_swab(tmp)), nbits);
+}
+
+unsigned long find_next_zero_bit_le(const void *addr, unsigned
+		long size, unsigned long offset)
+{
+	return _find_next_bit_le(addr, NULL, size, offset, ~0UL);
+}
+
+
+unsigned long find_next_bit_le(const void *addr, unsigned
+		long size, unsigned long offset)
+{
+	return _find_next_bit_le(addr, NULL, size, offset, 0UL);
+}
+
+#else
+
+static inline unsigned long find_next_zero_bit_le(const void *addr,
+                unsigned long size, unsigned long offset)
+{
+        return find_next_zero_bit(addr, size, offset);
+}
+
+static inline unsigned long find_next_bit_le(const void *addr,
+                unsigned long size, unsigned long offset)
+{
+        return find_next_bit(addr, size, offset);
+}
+
+static inline unsigned long find_first_zero_bit_le(const void *addr,
+                unsigned long size)
+{
+        return find_first_zero_bit(addr, size);
+}
+
+#endif
 
 #endif
-- 
2.7.4


^ permalink raw reply related	[flat|nested] 27+ messages in thread

* [PATCH 05/10] btrfs-progs: Pull free space tree related code from kernel
  2018-10-01 14:46 [PATCH 00/10] Freespace tree repair support v2 Nikolay Borisov
                   ` (3 preceding siblings ...)
  2018-10-01 14:46 ` [PATCH 04/10] btrfs-progs: Implement find_*_bit_le operations Nikolay Borisov
@ 2018-10-01 14:46 ` Nikolay Borisov
  2018-10-04 18:26   ` Omar Sandoval
  2018-10-01 14:46 ` [PATCH 06/10] btrfs-progs: Hook FST code in extent (de)alloc Nikolay Borisov
                   ` (5 subsequent siblings)
  10 siblings, 1 reply; 27+ messages in thread
From: Nikolay Borisov @ 2018-10-01 14:46 UTC (permalink / raw)
  To: linux-btrfs; +Cc: Nikolay Borisov

To help implement free space tree checker in user space some kernel
function are necessary, namely iterating/deleting/adding freespace
items, some internal search functions. Functions to populate a block
group based on the extent tree. The code is largely copy/paste from
the kernel with locking eliminated (i.e free_space_lock). It supports
reading/writing of both bitmap and extent based FST trees.

Signed-off-by: Nikolay Borisov <nborisov@suse.com>
---
 ctree.c           |   77 ++++
 ctree.h           |   15 +
 free-space-tree.c | 1253 ++++++++++++++++++++++++++++++++++++++++++++++++++++-
 free-space-tree.h |   13 +-
 kerncompat.h      |    6 +
 5 files changed, 1357 insertions(+), 7 deletions(-)

diff --git a/ctree.c b/ctree.c
index d8a6883aa85f..aa1568620205 100644
--- a/ctree.c
+++ b/ctree.c
@@ -1226,6 +1226,83 @@ int btrfs_search_slot(struct btrfs_trans_handle *trans, struct btrfs_root
 }
 
 /*
+ * helper to use instead of search slot if no exact match is needed but
+ * instead the next or previous item should be returned.
+ * When find_higher is true, the next higher item is returned, the next lower
+ * otherwise.
+ * When return_any and find_higher are both true, and no higher item is found,
+ * return the next lower instead.
+ * When return_any is true and find_higher is false, and no lower item is found,
+ * return the next higher instead.
+ * It returns 0 if any item is found, 1 if none is found (tree empty), and
+ * < 0 on error
+ */
+int btrfs_search_slot_for_read(struct btrfs_root *root,
+                               const struct btrfs_key *key,
+                               struct btrfs_path *p, int find_higher,
+                               int return_any)
+{
+        int ret;
+        struct extent_buffer *leaf;
+
+again:
+        ret = btrfs_search_slot(NULL, root, key, p, 0, 0);
+        if (ret <= 0)
+                return ret;
+        /*
+         * a return value of 1 means the path is at the position where the
+         * item should be inserted. Normally this is the next bigger item,
+         * but in case the previous item is the last in a leaf, path points
+         * to the first free slot in the previous leaf, i.e. at an invalid
+         * item.
+         */
+        leaf = p->nodes[0];
+
+        if (find_higher) {
+                if (p->slots[0] >= btrfs_header_nritems(leaf)) {
+                        ret = btrfs_next_leaf(root, p);
+                        if (ret <= 0)
+                                return ret;
+                        if (!return_any)
+                                return 1;
+                        /*
+                         * no higher item found, return the next
+                         * lower instead
+                         */
+                        return_any = 0;
+                        find_higher = 0;
+                        btrfs_release_path(p);
+                        goto again;
+                }
+        } else {
+                if (p->slots[0] == 0) {
+                        ret = btrfs_prev_leaf(root, p);
+                        if (ret < 0)
+                                return ret;
+                        if (!ret) {
+                                leaf = p->nodes[0];
+                                if (p->slots[0] == btrfs_header_nritems(leaf))
+                                        p->slots[0]--;
+                                return 0;
+                        }
+                        if (!return_any)
+                                return 1;
+                        /*
+                         * no lower item found, return the next
+                         * higher instead
+                         */
+                        return_any = 0;
+                        find_higher = 1;
+                        btrfs_release_path(p);
+                        goto again;
+                } else {
+                        --p->slots[0];
+                }
+        }
+        return 0;
+}
+
+/*
  * adjust the pointers going up the tree, starting at level
  * making sure the right key of each node is points to 'key'.
  * This is used after shifting pointers to the left, so it stops
diff --git a/ctree.h b/ctree.h
index 49f0f5181512..a6d6c3decd87 100644
--- a/ctree.h
+++ b/ctree.h
@@ -1071,6 +1071,17 @@ struct btrfs_block_group_cache {
 	u64 flags;
 	int cached;
 	int ro;
+	/*
+         * If the free space extent count exceeds this number, convert the block
+         * group to bitmaps.
+         */
+        u32 bitmap_high_thresh;
+        /*
+         * If the free space extent count drops below this number, convert the
+         * block group back to extents.
+         */
+        u32 bitmap_low_thresh;
+
 };
 
 struct btrfs_device;
@@ -2596,6 +2607,10 @@ int btrfs_split_item(struct btrfs_trans_handle *trans,
 int btrfs_search_slot(struct btrfs_trans_handle *trans, struct btrfs_root
 		      *root, struct btrfs_key *key, struct btrfs_path *p, int
 		      ins_len, int cow);
+int btrfs_search_slot_for_read(struct btrfs_root *root,
+                               const struct btrfs_key *key,
+                               struct btrfs_path *p, int find_higher,
+                               int return_any);
 int btrfs_find_item(struct btrfs_root *fs_root, struct btrfs_path *found_path,
 		u64 iobjectid, u64 ioff, u8 key_type,
 		struct btrfs_key *found_key);
diff --git a/free-space-tree.c b/free-space-tree.c
index b439b6b43146..3b7e8a3fe4f5 100644
--- a/free-space-tree.c
+++ b/free-space-tree.c
@@ -21,6 +21,37 @@
 #include "free-space-cache.h"
 #include "free-space-tree.h"
 #include "transaction.h"
+#include "bitops.h"
+#include "internal.h"
+
+void set_free_space_tree_thresholds(struct btrfs_block_group_cache *cache,
+				    u64 sectorsize)
+{
+	u32 bitmap_range;
+	size_t bitmap_size;
+	u64 num_bitmaps, total_bitmap_size;
+
+	/*
+	 * We convert to bitmaps when the disk space required for using extents
+	 * exceeds that required for using bitmaps.
+	 */
+	bitmap_range = sectorsize * BTRFS_FREE_SPACE_BITMAP_BITS;
+	num_bitmaps = div_u64(cache->key.offset + bitmap_range - 1,
+			      bitmap_range);
+	bitmap_size = sizeof(struct btrfs_item) + BTRFS_FREE_SPACE_BITMAP_SIZE;
+	total_bitmap_size = num_bitmaps * bitmap_size;
+	cache->bitmap_high_thresh = div_u64(total_bitmap_size,
+					    sizeof(struct btrfs_item));
+
+	/*
+	 * We allow for a small buffer between the high threshold and low
+	 * threshold to avoid thrashing back and forth between the two formats.
+	 */
+	if (cache->bitmap_high_thresh > 100)
+		cache->bitmap_low_thresh = cache->bitmap_high_thresh - 100;
+	else
+		cache->bitmap_low_thresh = 0;
+}
 
 static struct btrfs_free_space_info *
 search_free_space_info(struct btrfs_trans_handle *trans,
@@ -47,8 +78,7 @@ search_free_space_info(struct btrfs_trans_handle *trans,
 }
 
 static int free_space_test_bit(struct btrfs_block_group_cache *block_group,
-			       struct btrfs_path *path, u64 offset,
-			       u64 sectorsize)
+			       struct btrfs_path *path, u64 offset)
 {
 	struct extent_buffer *leaf;
 	struct btrfs_key key;
@@ -64,10 +94,1085 @@ static int free_space_test_bit(struct btrfs_block_group_cache *block_group,
 	ASSERT(offset >= found_start && offset < found_end);
 
 	ptr = btrfs_item_ptr_offset(leaf, path->slots[0]);
-	i = (offset - found_start) / sectorsize;
+	i = (offset - found_start) / leaf->fs_info->sectorsize;
 	return !!extent_buffer_test_bit(leaf, ptr, i);
 }
 
+/*
+ * btrfs_search_slot() but we're looking for the greatest key less than the
+ * passed key.
+ */
+static int btrfs_search_prev_slot(struct btrfs_trans_handle *trans,
+                                  struct btrfs_root *root,
+                                  struct btrfs_key *key, struct btrfs_path *p,
+                                  int ins_len, int cow)
+{
+	int ret;
+
+	ret = btrfs_search_slot(trans, root, key, p, ins_len, cow);
+	if (ret < 0)
+		return ret;
+
+	if (ret == 0) {
+		ASSERT(0);
+		return -EIO;
+	}
+
+	if (p->slots[0] == 0) {
+		ASSERT(0);
+		return -EIO;
+	}
+	p->slots[0]--;
+
+	return 0;
+}
+
+static int add_new_free_space_info(struct btrfs_trans_handle *trans,
+                                   struct btrfs_block_group_cache *block_group,
+                                   struct btrfs_path *path)
+{
+	struct btrfs_root *root = trans->fs_info->free_space_root;
+	struct btrfs_free_space_info *info;
+	struct btrfs_key key;
+	struct extent_buffer *leaf;
+	int ret;
+
+	key.objectid = block_group->key.objectid;
+	key.type = BTRFS_FREE_SPACE_INFO_KEY;
+	key.offset = block_group->key.offset;
+
+	ret = btrfs_insert_empty_item(trans, root, path, &key, sizeof(*info));
+	if (ret)
+		goto out;
+
+	leaf = path->nodes[0];
+	info = btrfs_item_ptr(leaf, path->slots[0],
+	                      struct btrfs_free_space_info);
+	btrfs_set_free_space_extent_count(leaf, info, 0);
+	btrfs_set_free_space_flags(leaf, info, 0);
+	btrfs_mark_buffer_dirty(leaf);
+
+	ret = 0;
+out:
+	btrfs_release_path(path);
+	return ret;
+}
+
+static inline u32 free_space_bitmap_size(u64 size, u32 sectorsize)
+{
+	return DIV_ROUND_UP((u32)div_u64(size, sectorsize), BITS_PER_BYTE);
+}
+
+static unsigned long *alloc_bitmap(u32 bitmap_size)
+{
+	unsigned long *ret;
+	unsigned int nofs_flag;
+	u32 bitmap_rounded_size = round_up(bitmap_size, sizeof(unsigned long));
+
+	/*
+	 * GFP_NOFS doesn't work with kvmalloc(), but we really can't recurse
+	 * into the filesystem as the free space bitmap can be modified in the
+	 * critical section of a transaction commit.
+	 *
+	 * TODO: push the memalloc_nofs_{save,restore}() to the caller where we
+	 * know that recursion is unsafe.
+	 */
+	nofs_flag = memalloc_nofs_save();
+	ret = kvzalloc(bitmap_rounded_size, GFP_KERNEL);
+	memalloc_nofs_restore(nofs_flag);
+	return ret;
+}
+
+static void le_bitmap_set(unsigned long *map, unsigned int start, int len)
+{
+	u8 *p = ((u8 *)map) + BIT_BYTE(start);
+	const unsigned int size = start + len;
+	int bits_to_set = BITS_PER_BYTE - (start % BITS_PER_BYTE);
+	u8 mask_to_set = BITMAP_FIRST_BYTE_MASK(start);
+
+	while (len - bits_to_set >= 0) {
+		*p |= mask_to_set;
+		len -= bits_to_set;
+		bits_to_set = BITS_PER_BYTE;
+		mask_to_set = ~0;
+		p++;
+	}
+	if (len) {
+		mask_to_set &= BITMAP_LAST_BYTE_MASK(size);
+		*p |= mask_to_set;
+	}
+}
+
+int convert_free_space_to_bitmaps(struct btrfs_trans_handle *trans,
+				  struct btrfs_block_group_cache *block_group,
+				  struct btrfs_path *path)
+{
+	struct btrfs_fs_info *fs_info = trans->fs_info;
+	struct btrfs_root *root = fs_info->free_space_root;
+	struct btrfs_free_space_info *info;
+	struct btrfs_key key, found_key;
+	struct extent_buffer *leaf;
+	unsigned long *bitmap;
+	char *bitmap_cursor;
+	u64 start, end;
+	u64 bitmap_range, i;
+	u32 bitmap_size, flags, expected_extent_count;
+	u32 extent_count = 0;
+	int done = 0, nr;
+	int ret;
+
+	bitmap_size = free_space_bitmap_size(block_group->key.offset,
+					     fs_info->sectorsize);
+	bitmap = alloc_bitmap(bitmap_size);
+	if (!bitmap) {
+		ret = -ENOMEM;
+		goto out;
+	}
+
+	start = block_group->key.objectid;
+	end = block_group->key.objectid + block_group->key.offset;
+
+	key.objectid = end - 1;
+	key.type = (u8)-1;
+	key.offset = (u64)-1;
+
+	while (!done) {
+		ret = btrfs_search_prev_slot(trans, root, &key, path, -1, 1);
+		if (ret)
+			goto out;
+
+		leaf = path->nodes[0];
+		nr = 0;
+		path->slots[0]++;
+		while (path->slots[0] > 0) {
+			btrfs_item_key_to_cpu(leaf, &found_key, path->slots[0] - 1);
+
+			if (found_key.type == BTRFS_FREE_SPACE_INFO_KEY) {
+				ASSERT(found_key.objectid == block_group->key.objectid);
+				ASSERT(found_key.offset == block_group->key.offset);
+				done = 1;
+				break;
+			} else if (found_key.type == BTRFS_FREE_SPACE_EXTENT_KEY) {
+				u64 first, last;
+
+				ASSERT(found_key.objectid >= start);
+				ASSERT(found_key.objectid < end);
+				ASSERT(found_key.objectid + found_key.offset <= end);
+
+				first = div_u64(found_key.objectid - start,
+						fs_info->sectorsize);
+				last = div_u64(found_key.objectid + found_key.offset - start,
+					       fs_info->sectorsize);
+				le_bitmap_set(bitmap, first, last - first);
+
+				extent_count++;
+				nr++;
+				path->slots[0]--;
+			} else {
+				ASSERT(0);
+			}
+		}
+
+		ret = btrfs_del_items(trans, root, path, path->slots[0], nr);
+		if (ret)
+			goto out;
+		btrfs_release_path(path);
+	}
+
+	info = search_free_space_info(trans, fs_info, block_group, path, 1);
+	if (IS_ERR(info)) {
+		ret = PTR_ERR(info);
+		goto out;
+	}
+	leaf = path->nodes[0];
+	flags = btrfs_free_space_flags(leaf, info);
+	flags |= BTRFS_FREE_SPACE_USING_BITMAPS;
+	btrfs_set_free_space_flags(leaf, info, flags);
+	expected_extent_count = btrfs_free_space_extent_count(leaf, info);
+	btrfs_mark_buffer_dirty(leaf);
+	btrfs_release_path(path);
+
+	if (extent_count != expected_extent_count) {
+		fprintf(stderr,
+			"incorrect extent count for %llu; counted %u, expected %u",
+			block_group->key.objectid, extent_count,
+			expected_extent_count);
+		ASSERT(0);
+		ret = -EIO;
+		goto out;
+	}
+
+	bitmap_cursor = (char *)bitmap;
+	bitmap_range = fs_info->sectorsize * BTRFS_FREE_SPACE_BITMAP_BITS;
+	i = start;
+	while (i < end) {
+		unsigned long ptr;
+		u64 extent_size;
+		u32 data_size;
+
+		extent_size = min(end - i, bitmap_range);
+		data_size = free_space_bitmap_size(extent_size,
+						   fs_info->sectorsize);
+
+		key.objectid = i;
+		key.type = BTRFS_FREE_SPACE_BITMAP_KEY;
+		key.offset = extent_size;
+
+		ret = btrfs_insert_empty_item(trans, root, path, &key,
+					      data_size);
+		if (ret)
+			goto out;
+
+		leaf = path->nodes[0];
+		ptr = btrfs_item_ptr_offset(leaf, path->slots[0]);
+		write_extent_buffer(leaf, bitmap_cursor, ptr,
+				    data_size);
+		btrfs_mark_buffer_dirty(leaf);
+		btrfs_release_path(path);
+
+		i += extent_size;
+		bitmap_cursor += data_size;
+	}
+
+	ret = 0;
+out:
+	kvfree(bitmap);
+	if (ret)
+		btrfs_abort_transaction(trans, ret);
+	return ret;
+}
+
+int convert_free_space_to_extents(struct btrfs_trans_handle *trans,
+				  struct btrfs_block_group_cache *block_group,
+				  struct btrfs_path *path)
+{
+	struct btrfs_fs_info *fs_info = trans->fs_info;
+	struct btrfs_root *root = fs_info->free_space_root;
+	struct btrfs_free_space_info *info;
+	struct btrfs_key key, found_key;
+	struct extent_buffer *leaf;
+	unsigned long *bitmap;
+	u64 start, end;
+	u32 bitmap_size, flags, expected_extent_count;
+	unsigned long nrbits, start_bit, end_bit;
+	u32 extent_count = 0;
+	int done = 0, nr;
+	int ret;
+
+	bitmap_size = free_space_bitmap_size(block_group->key.offset,
+					     fs_info->sectorsize);
+	bitmap = alloc_bitmap(bitmap_size);
+	if (!bitmap) {
+		ret = -ENOMEM;
+		goto out;
+	}
+
+	start = block_group->key.objectid;
+	end = block_group->key.objectid + block_group->key.offset;
+
+	key.objectid = end - 1;
+	key.type = (u8)-1;
+	key.offset = (u64)-1;
+
+	while (!done) {
+		ret = btrfs_search_prev_slot(trans, root, &key, path, -1, 1);
+		if (ret)
+			goto out;
+
+		leaf = path->nodes[0];
+		nr = 0;
+		path->slots[0]++;
+		while (path->slots[0] > 0) {
+			btrfs_item_key_to_cpu(leaf, &found_key, path->slots[0] - 1);
+
+			if (found_key.type == BTRFS_FREE_SPACE_INFO_KEY) {
+				ASSERT(found_key.objectid == block_group->key.objectid);
+				ASSERT(found_key.offset == block_group->key.offset);
+				done = 1;
+				break;
+			} else if (found_key.type == BTRFS_FREE_SPACE_BITMAP_KEY) {
+				unsigned long ptr;
+				char *bitmap_cursor;
+				u32 bitmap_pos, data_size;
+
+				ASSERT(found_key.objectid >= start);
+				ASSERT(found_key.objectid < end);
+				ASSERT(found_key.objectid + found_key.offset <= end);
+
+				bitmap_pos = div_u64(found_key.objectid - start,
+						     fs_info->sectorsize *
+						     BITS_PER_BYTE);
+				bitmap_cursor = ((char *)bitmap) + bitmap_pos;
+				data_size = free_space_bitmap_size(found_key.offset,
+								   fs_info->sectorsize);
+
+				ptr = btrfs_item_ptr_offset(leaf, path->slots[0] - 1);
+				read_extent_buffer(leaf, bitmap_cursor, ptr,
+						   data_size);
+
+				nr++;
+				path->slots[0]--;
+			} else {
+				ASSERT(0);
+			}
+		}
+
+		ret = btrfs_del_items(trans, root, path, path->slots[0], nr);
+		if (ret)
+			goto out;
+		btrfs_release_path(path);
+	}
+
+	info = search_free_space_info(trans, fs_info, block_group, path, 1);
+	if (IS_ERR(info)) {
+		ret = PTR_ERR(info);
+		goto out;
+	}
+	leaf = path->nodes[0];
+	flags = btrfs_free_space_flags(leaf, info);
+	flags &= ~BTRFS_FREE_SPACE_USING_BITMAPS;
+	btrfs_set_free_space_flags(leaf, info, flags);
+	expected_extent_count = btrfs_free_space_extent_count(leaf, info);
+	btrfs_mark_buffer_dirty(leaf);
+	btrfs_release_path(path);
+
+	nrbits = div_u64(block_group->key.offset, fs_info->sectorsize);
+	start_bit = find_next_bit_le(bitmap, nrbits, 0);
+
+	while (start_bit < nrbits) {
+		end_bit = find_next_zero_bit_le(bitmap, nrbits, start_bit);
+		ASSERT(start_bit < end_bit);
+
+		key.objectid = start + start_bit * fs_info->sectorsize;
+		key.type = BTRFS_FREE_SPACE_EXTENT_KEY;
+		key.offset = (end_bit - start_bit) * fs_info->sectorsize;
+
+		ret = btrfs_insert_empty_item(trans, root, path, &key, 0);
+		if (ret)
+			goto out;
+		btrfs_release_path(path);
+
+		extent_count++;
+
+		start_bit = find_next_bit_le(bitmap, nrbits, end_bit);
+	}
+
+	if (extent_count != expected_extent_count) {
+		fprintf(stderr,
+			"incorrect extent count for %llu; counted %u, expected %u",
+			block_group->key.objectid, extent_count,
+			expected_extent_count);
+		ASSERT(0);
+		ret = -EIO;
+		goto out;
+	}
+
+	ret = 0;
+out:
+	kvfree(bitmap);
+	if (ret)
+		btrfs_abort_transaction(trans, ret);
+	return ret;
+}
+
+static int update_free_space_extent_count(struct btrfs_trans_handle *trans,
+                                          struct btrfs_block_group_cache *block_group,
+                                          struct btrfs_path *path,
+                                          int new_extents)
+{
+	struct btrfs_free_space_info *info;
+	u32 flags;
+	u32 extent_count;
+	int ret = 0;
+
+	if (new_extents == 0)
+		return 0;
+
+	info = search_free_space_info(trans, trans->fs_info, block_group, path,
+				1);
+	if (IS_ERR(info)) {
+		ret = PTR_ERR(info);
+		goto out;
+	}
+	flags = btrfs_free_space_flags(path->nodes[0], info);
+	extent_count = btrfs_free_space_extent_count(path->nodes[0], info);
+
+	extent_count += new_extents;
+	btrfs_set_free_space_extent_count(path->nodes[0], info, extent_count);
+	btrfs_mark_buffer_dirty(path->nodes[0]);
+	btrfs_release_path(path);
+
+	if (!(flags & BTRFS_FREE_SPACE_USING_BITMAPS) &&
+	    extent_count > block_group->bitmap_high_thresh) {
+		ret = convert_free_space_to_bitmaps(trans, block_group, path);
+	} else if ((flags & BTRFS_FREE_SPACE_USING_BITMAPS) &&
+		   extent_count < block_group->bitmap_low_thresh) {
+		ret = convert_free_space_to_extents(trans, block_group, path);
+	}
+
+
+out:
+	return ret;
+}
+
+
+static void free_space_set_bits(struct btrfs_block_group_cache *block_group,
+                                struct btrfs_path *path, u64 *start, u64 *size,
+                                int bit)
+{
+        struct extent_buffer *leaf = path->nodes[0];
+        struct btrfs_fs_info *fs_info = leaf->fs_info;
+        struct btrfs_key key;
+        u64 end = *start + *size;
+        u64 found_start, found_end;
+        unsigned long ptr, first, last;
+
+        btrfs_item_key_to_cpu(leaf, &key, path->slots[0]);
+        ASSERT(key.type == BTRFS_FREE_SPACE_BITMAP_KEY);
+
+        found_start = key.objectid;
+        found_end = key.objectid + key.offset;
+        ASSERT(*start >= found_start && *start < found_end);
+        ASSERT(end > found_start);
+
+        if (end > found_end)
+                end = found_end;
+
+        ptr = btrfs_item_ptr_offset(leaf, path->slots[0]);
+        first = (*start - found_start) / fs_info->sectorsize;
+        last = (end - found_start) / fs_info->sectorsize;
+        if (bit)
+                extent_buffer_bitmap_set(leaf, ptr, first, last - first);
+        else
+                extent_buffer_bitmap_clear(leaf, ptr, first, last - first);
+        btrfs_mark_buffer_dirty(leaf);
+
+        *size -= end - *start;
+        *start = end;
+}
+
+/*
+ * We can't use btrfs_next_item() in modify_free_space_bitmap() because
+ * btrfs_next_leaf() doesn't get the path for writing. We can forgo the fancy
+ * tree walking in btrfs_next_leaf() anyways because we know exactly what we're
+ * looking for.
+ */
+static int free_space_next_bitmap(struct btrfs_trans_handle *trans,
+                                  struct btrfs_root *root, struct btrfs_path *p)
+{
+	struct btrfs_key key;
+
+	if (p->slots[0] + 1 < btrfs_header_nritems(p->nodes[0])) {
+		p->slots[0]++;
+		return 0;
+	}
+
+	btrfs_item_key_to_cpu(p->nodes[0], &key, p->slots[0]);
+	btrfs_release_path(p);
+
+	key.objectid += key.offset;
+	key.type = (u8)-1;
+	key.offset = (u64)-1;
+
+	return btrfs_search_prev_slot(trans, root, &key, p, 0, 1);
+}
+
+/*
+ * If remove is 1, then we are removing free space, thus clearing bits in the
+ * bitmap. If remove is 0, then we are adding free space, thus setting bits in
+ * the bitmap.
+ */
+static int modify_free_space_bitmap(struct btrfs_trans_handle *trans,
+                                    struct btrfs_block_group_cache *block_group,
+                                    struct btrfs_path *path,
+                                    u64 start, u64 size, int remove)
+{
+        struct btrfs_root *root = trans->fs_info->free_space_root;
+        struct btrfs_key key;
+        u64 end = start + size;
+        u64 cur_start, cur_size;
+        int prev_bit, next_bit;
+        int new_extents;
+        int ret;
+
+        /*
+         * Read the bit for the block immediately before the extent of space if
+         * that block is within the block group.
+         */
+        if (start > block_group->key.objectid) {
+                u64 prev_block = start - trans->fs_info->sectorsize;
+
+                key.objectid = prev_block;
+                key.type = (u8)-1;
+                key.offset = (u64)-1;
+
+                ret = btrfs_search_prev_slot(trans, root, &key, path, 0, 1);
+                if (ret)
+                        goto out;
+
+                prev_bit = free_space_test_bit(block_group, path, prev_block);
+
+                /* The previous block may have been in the previous bitmap. */
+                btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]);
+                if (start >= key.objectid + key.offset) {
+                        ret = free_space_next_bitmap(trans, root, path);
+                        if (ret)
+                                goto out;
+                }
+        } else {
+                key.objectid = start;
+                key.type = (u8)-1;
+                key.offset = (u64)-1;
+
+                ret = btrfs_search_prev_slot(trans, root, &key, path, 0, 1);
+                if (ret)
+                        goto out;
+
+                prev_bit = -1;
+        }
+
+        /*
+         * Iterate over all of the bitmaps overlapped by the extent of space,
+         * clearing/setting bits as required.
+         */
+        cur_start = start;
+        cur_size = size;
+        while (1) {
+                free_space_set_bits(block_group, path, &cur_start, &cur_size,
+                                    !remove);
+                if (cur_size == 0)
+                        break;
+                ret = free_space_next_bitmap(trans, root, path);
+                if (ret)
+                        goto out;
+        }
+
+	/*
+         * Read the bit for the block immediately after the extent of space if
+         * that block is within the block group.
+         */
+        if (end < block_group->key.objectid + block_group->key.offset) {
+                /* The next block may be in the next bitmap. */
+                btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]);
+                if (end >= key.objectid + key.offset) {
+                        ret = free_space_next_bitmap(trans, root, path);
+                        if (ret)
+                                goto out;
+                }
+
+                next_bit = free_space_test_bit(block_group, path, end);
+        } else {
+                next_bit = -1;
+        }
+
+        if (remove) {
+                new_extents = -1;
+                if (prev_bit == 1) {
+                        /* Leftover on the left. */
+                        new_extents++;
+                }
+                if (next_bit == 1) {
+                        /* Leftover on the right. */
+                        new_extents++;
+                }
+        } else {
+                new_extents = 1;
+                if (prev_bit == 1) {
+                        /* Merging with neighbor on the left. */
+                        new_extents--;
+                }
+                if (next_bit == 1) {
+                        /* Merging with neighbor on the right. */
+                        new_extents--;
+                }
+        }
+
+        btrfs_release_path(path);
+        ret = update_free_space_extent_count(trans, block_group, path,
+                                             new_extents);
+
+out:
+        return ret;
+}
+
+static int remove_free_space_extent(struct btrfs_trans_handle *trans,
+				    struct btrfs_block_group_cache *block_group,
+				    struct btrfs_path *path,
+				    u64 start, u64 size)
+{
+	struct btrfs_root *root = trans->fs_info->free_space_root;
+	struct btrfs_key key;
+	u64 found_start, found_end;
+	u64 end = start + size;
+	int new_extents = -1;
+	int ret;
+
+	key.objectid = start;
+	key.type = (u8)-1;
+	key.offset = (u64)-1;
+
+	ret = btrfs_search_prev_slot(trans, root, &key, path, -1, 1);
+	if (ret)
+		goto out;
+
+	btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]);
+
+	ASSERT(key.type == BTRFS_FREE_SPACE_EXTENT_KEY);
+
+	found_start = key.objectid;
+	found_end = key.objectid + key.offset;
+	ASSERT(start >= found_start && end <= found_end);
+
+	/*
+	 * Okay, now that we've found the free space extent which contains the
+	 * free space that we are removing, there are four cases:
+	 *
+	 * 1. We're using the whole extent: delete the key we found and
+	 * decrement the free space extent count.
+	 * 2. We are using part of the extent starting at the beginning: delete
+	 * the key we found and insert a new key representing the leftover at
+	 * the end. There is no net change in the number of extents.
+	 * 3. We are using part of the extent ending at the end: delete the key
+	 * we found and insert a new key representing the leftover at the
+	 * beginning. There is no net change in the number of extents.
+	 * 4. We are using part of the extent in the middle: delete the key we
+	 * found and insert two new keys representing the leftovers on each
+	 * side. Where we used to have one extent, we now have two, so increment
+	 * the extent count. We may need to convert the block group to bitmaps
+	 * as a result.
+	 */
+
+	/* Delete the existing key (cases 1-4). */
+	ret = btrfs_del_item(trans, root, path);
+	if (ret)
+		goto out;
+
+	/* Add a key for leftovers at the beginning (cases 3 and 4). */
+	if (start > found_start) {
+		key.objectid = found_start;
+		key.type = BTRFS_FREE_SPACE_EXTENT_KEY;
+		key.offset = start - found_start;
+
+		btrfs_release_path(path);
+		ret = btrfs_insert_empty_item(trans, root, path, &key, 0);
+		if (ret)
+			goto out;
+		new_extents++;
+	}
+
+	/* Add a key for leftovers at the end (cases 2 and 4). */
+	if (end < found_end) {
+		key.objectid = end;
+		key.type = BTRFS_FREE_SPACE_EXTENT_KEY;
+		key.offset = found_end - end;
+
+		btrfs_release_path(path);
+		ret = btrfs_insert_empty_item(trans, root, path, &key, 0);
+		if (ret)
+			goto out;
+		new_extents++;
+	}
+
+	btrfs_release_path(path);
+	ret = update_free_space_extent_count(trans, block_group, path,
+					     new_extents);
+
+out:
+	return ret;
+}
+
+int __remove_from_free_space_tree(struct btrfs_trans_handle *trans,
+                                  struct btrfs_block_group_cache *block_group,
+                                  struct btrfs_path *path, u64 start, u64 size)
+{
+	struct btrfs_free_space_info *info;
+	u32 flags;
+
+	info = search_free_space_info(NULL, trans->fs_info, block_group, path,
+	                              0);
+	if (IS_ERR(info))
+		return PTR_ERR(info);
+	flags = btrfs_free_space_flags(path->nodes[0], info);
+	btrfs_release_path(path);
+
+	if (flags & BTRFS_FREE_SPACE_USING_BITMAPS) {
+		return modify_free_space_bitmap(trans, block_group, path,
+	                                        start, size, 1);
+	} else {
+		return remove_free_space_extent(trans, block_group, path,
+	                                        start, size);
+	}
+}
+
+int remove_from_free_space_tree(struct btrfs_trans_handle *trans,
+				u64 start, u64 size)
+{
+	struct btrfs_block_group_cache *block_group;
+	struct btrfs_path *path;
+	int ret;
+
+	if (!btrfs_fs_compat_ro(trans->fs_info, FREE_SPACE_TREE))
+		return 0;
+
+	path = btrfs_alloc_path();
+	if (!path) {
+		ret = -ENOMEM;
+		goto out;
+	}
+
+	block_group = btrfs_lookup_block_group(trans->fs_info, start);
+	if (!block_group) {
+		ASSERT(0);
+		ret = -ENOENT;
+		goto out;
+	}
+
+	ret = __remove_from_free_space_tree(trans, block_group, path, start,
+					    size);
+out:
+	btrfs_free_path(path);
+	if (ret)
+		btrfs_abort_transaction(trans, ret);
+	return ret;
+}
+
+static int add_free_space_extent(struct btrfs_trans_handle *trans,
+                                 struct btrfs_block_group_cache *block_group,
+                                 struct btrfs_path *path,
+                                 u64 start, u64 size)
+{
+        struct btrfs_root *root = trans->fs_info->free_space_root;
+        struct btrfs_key key, new_key;
+        u64 found_start, found_end;
+        u64 end = start + size;
+        int new_extents = 1;
+        int ret;
+
+        /*
+         * We are adding a new extent of free space, but we need to merge
+         * extents. There are four cases here:
+         *
+         * 1. The new extent does not have any immediate neighbors to merge
+         * with: add the new key and increment the free space extent count. We
+         * may need to convert the block group to bitmaps as a result.
+         * 2. The new extent has an immediate neighbor before it: remove the
+         * previous key and insert a new key combining both of them. There is no
+         * net change in the number of extents.
+         * 3. The new extent has an immediate neighbor after it: remove the next
+         * key and insert a new key combining both of them. There is no net
+         * change in the number of extents.
+         * 4. The new extent has immediate neighbors on both sides: remove both
+         * of the keys and insert a new key combining all of them. Where we used
+         * to have two extents, we now have one, so decrement the extent count.
+         */
+
+        new_key.objectid = start;
+        new_key.type = BTRFS_FREE_SPACE_EXTENT_KEY;
+        new_key.offset = size;
+
+        /* Search for a neighbor on the left. */
+        if (start == block_group->key.objectid)
+                goto right;
+        key.objectid = start - 1;
+        key.type = (u8)-1;
+        key.offset = (u64)-1;
+
+        ret = btrfs_search_prev_slot(trans, root, &key, path, -1, 1);
+        if (ret)
+                goto out;
+
+        btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]);
+
+        if (key.type != BTRFS_FREE_SPACE_EXTENT_KEY) {
+                ASSERT(key.type == BTRFS_FREE_SPACE_INFO_KEY);
+                btrfs_release_path(path);
+                goto right;
+        }
+
+        found_start = key.objectid;
+        found_end = key.objectid + key.offset;
+        ASSERT(found_start >= block_group->key.objectid &&
+               found_end > block_group->key.objectid);
+        ASSERT(found_start < start && found_end <= start);
+
+        /*
+         * Delete the neighbor on the left and absorb it into the new key (cases
+         * 2 and 4).
+         */
+        if (found_end == start) {
+                ret = btrfs_del_item(trans, root, path);
+                if (ret)
+                        goto out;
+                new_key.objectid = found_start;
+                new_key.offset += key.offset;
+                new_extents--;
+        }
+        btrfs_release_path(path);
+right:
+        /* Search for a neighbor on the right. */
+        if (end == block_group->key.objectid + block_group->key.offset)
+                goto insert;
+        key.objectid = end;
+        key.type = (u8)-1;
+        key.offset = (u64)-1;
+
+        ret = btrfs_search_prev_slot(trans, root, &key, path, -1, 1);
+        if (ret)
+                goto out;
+
+        btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]);
+
+        if (key.type != BTRFS_FREE_SPACE_EXTENT_KEY) {
+                ASSERT(key.type == BTRFS_FREE_SPACE_INFO_KEY);
+                btrfs_release_path(path);
+                goto insert;
+        }
+
+        found_start = key.objectid;
+        found_end = key.objectid + key.offset;
+        ASSERT(found_start >= block_group->key.objectid &&
+               found_end > block_group->key.objectid);
+        ASSERT((found_start < start && found_end <= start) ||
+               (found_start >= end && found_end > end));
+
+        /*
+         * Delete the neighbor on the right and absorb it into the new key
+         * (cases 3 and 4).
+         */
+        if (found_start == end) {
+                ret = btrfs_del_item(trans, root, path);
+                if (ret)
+                        goto out;
+                new_key.offset += key.offset;
+                new_extents--;
+        }
+        btrfs_release_path(path);
+
+insert:
+        /* Insert the new key (cases 1-4). */
+        ret = btrfs_insert_empty_item(trans, root, path, &new_key, 0);
+        if (ret)
+                goto out;
+
+        btrfs_release_path(path);
+        ret = update_free_space_extent_count(trans, block_group, path,
+                                             new_extents);
+
+out:
+        return ret;
+}
+
+int __add_to_free_space_tree(struct btrfs_trans_handle *trans,
+                             struct btrfs_block_group_cache *block_group,
+                             struct btrfs_path *path, u64 start, u64 size)
+{
+	struct btrfs_fs_info *fs_info = trans->fs_info;
+	struct btrfs_free_space_info *info;
+	u32 flags;
+
+	info = search_free_space_info(NULL, fs_info, block_group, path, 0);
+	if (IS_ERR(info))
+	        return PTR_ERR(info);
+	flags = btrfs_free_space_flags(path->nodes[0], info);
+	btrfs_release_path(path);
+
+	if (flags & BTRFS_FREE_SPACE_USING_BITMAPS) {
+	        return modify_free_space_bitmap(trans, block_group, path,
+	                                        start, size, 0);
+	} else {
+	        return add_free_space_extent(trans, block_group, path, start,
+	                                     size);
+	}
+}
+
+
+int add_to_free_space_tree(struct btrfs_trans_handle *trans,
+			   u64 start, u64 size)
+{
+	struct btrfs_block_group_cache *block_group;
+	struct btrfs_path *path;
+	int ret;
+
+	if (!btrfs_fs_compat_ro(trans->fs_info, FREE_SPACE_TREE))
+		return 0;
+
+	path = btrfs_alloc_path();
+	if (!path) {
+		ret = -ENOMEM;
+		goto out;
+	}
+
+	block_group = btrfs_lookup_block_group(trans->fs_info, start);
+	if (!block_group) {
+		ASSERT(0);
+		ret = -ENOENT;
+		goto out;
+	}
+
+	ret = __add_to_free_space_tree(trans, block_group, path, start, size);
+out:
+	btrfs_free_path(path);
+	if (ret)
+		btrfs_abort_transaction(trans, ret);
+	return ret;
+}
+
+int populate_free_space_tree(struct btrfs_trans_handle *trans,
+			     struct btrfs_block_group_cache *block_group)
+{
+        struct btrfs_root *extent_root = trans->fs_info->extent_root;
+        struct btrfs_path *path, *path2;
+        struct btrfs_key key;
+        u64 start, end;
+        int ret;
+
+        path = btrfs_alloc_path();
+        if (!path)
+                return -ENOMEM;
+        path->reada = READA_FORWARD;
+
+        path2 = btrfs_alloc_path();
+        if (!path2) {
+                btrfs_free_path(path);
+                return -ENOMEM;
+        }
+
+        ret = add_new_free_space_info(trans, block_group, path2);
+        if (ret)
+                goto out;
+
+        /*
+         * Iterate through all of the extent and metadata items in this block
+         * group, adding the free space between them and the free space at the
+         * end. Note that EXTENT_ITEM and METADATA_ITEM are less than
+         * BLOCK_GROUP_ITEM, so an extent may precede the block group that it's
+         * contained in.
+         */
+        key.objectid = block_group->key.objectid;
+        key.type = BTRFS_EXTENT_ITEM_KEY;
+        key.offset = 0;
+
+        ret = btrfs_search_slot_for_read(extent_root, &key, path, 1, 0);
+        if (ret < 0)
+                goto out;
+        ASSERT(ret == 0);
+
+        start = block_group->key.objectid;
+        end = block_group->key.objectid + block_group->key.offset;
+        while (1) {
+                btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]);
+
+                if (key.type == BTRFS_EXTENT_ITEM_KEY ||
+                    key.type == BTRFS_METADATA_ITEM_KEY) {
+                        if (key.objectid >= end)
+                                break;
+
+                        if (start < key.objectid) {
+                                ret = __add_to_free_space_tree(trans,
+                                                               block_group,
+                                                               path2, start,
+                                                               key.objectid -
+                                                               start);
+                                if (ret)
+                                        goto out;
+                        }
+                        start = key.objectid;
+                        if (key.type == BTRFS_METADATA_ITEM_KEY)
+                                start += trans->fs_info->nodesize;
+                        else
+                                start += key.offset;
+                } else if (key.type == BTRFS_BLOCK_GROUP_ITEM_KEY) {
+                        if (key.objectid != block_group->key.objectid)
+                                break;
+                }
+
+                ret = btrfs_next_item(extent_root, path);
+                if (ret < 0)
+                        goto out;
+                if (ret)
+                        break;
+        }
+        if (start < end) {
+                ret = __add_to_free_space_tree(trans, block_group, path2,
+                                               start, end - start);
+                if (ret)
+                        goto out;
+        }
+
+        ret = 0;
+out:
+        btrfs_free_path(path2);
+        btrfs_free_path(path);
+        return ret;
+}
+
+int remove_block_group_free_space(struct btrfs_trans_handle *trans,
+				  struct btrfs_block_group_cache *block_group)
+{
+	struct btrfs_root *root = trans->fs_info->free_space_root;
+	struct btrfs_path *path;
+	struct btrfs_key key, found_key;
+	struct extent_buffer *leaf;
+	u64 start, end;
+	int done = 0, nr;
+	int ret;
+
+	path = btrfs_alloc_path();
+	if (!path) {
+		ret = -ENOMEM;
+		goto out;
+	}
+
+	start = block_group->key.objectid;
+	end = block_group->key.objectid + block_group->key.offset;
+
+	key.objectid = end - 1;
+	key.type = (u8)-1;
+	key.offset = (u64)-1;
+
+	while (!done) {
+		ret = btrfs_search_prev_slot(trans, root, &key, path, -1, 1);
+		if (ret)
+			goto out;
+
+		leaf = path->nodes[0];
+		nr = 0;
+		path->slots[0]++;
+		while (path->slots[0] > 0) {
+			btrfs_item_key_to_cpu(leaf, &found_key, path->slots[0] - 1);
+
+			if (found_key.type == BTRFS_FREE_SPACE_INFO_KEY) {
+				ASSERT(found_key.objectid == block_group->key.objectid);
+				ASSERT(found_key.offset == block_group->key.offset);
+				done = 1;
+				nr++;
+				path->slots[0]--;
+				break;
+			} else if (found_key.type == BTRFS_FREE_SPACE_EXTENT_KEY ||
+				   found_key.type == BTRFS_FREE_SPACE_BITMAP_KEY) {
+				ASSERT(found_key.objectid >= start);
+				ASSERT(found_key.objectid < end);
+				ASSERT(found_key.objectid + found_key.offset <= end);
+				nr++;
+				path->slots[0]--;
+			} else {
+				ASSERT(0);
+			}
+		}
+
+		ret = btrfs_del_items(trans, root, path, path->slots[0], nr);
+		if (ret)
+			goto out;
+		btrfs_release_path(path);
+	}
+
+	ret = 0;
+out:
+	btrfs_free_path(path);
+	if (ret)
+		btrfs_abort_transaction(trans, ret);
+	return ret;
+}
 static int clear_free_space_tree(struct btrfs_trans_handle *trans,
 				 struct btrfs_root *root)
 {
@@ -204,8 +1309,8 @@ static int load_free_space_bitmaps(struct btrfs_fs_info *fs_info,
 
 		offset = key.objectid;
 		while (offset < key.objectid + key.offset) {
-			bit = free_space_test_bit(block_group, path, offset,
-						  fs_info->sectorsize);
+			bit = free_space_test_bit(block_group, path, offset);
+
 			if (prev_bit == 0 && bit == 1) {
 				extent_start = offset;
 			} else if (prev_bit == 1 && bit == 0) {
@@ -320,6 +1425,142 @@ static int load_free_space_extents(struct btrfs_fs_info *fs_info,
 	return ret;
 }
 
+struct btrfs_root *btrfs_create_tree(struct btrfs_trans_handle *trans,
+                                     struct btrfs_fs_info *fs_info,
+                                     u64 objectid)
+{
+	struct extent_buffer *leaf;
+	struct btrfs_root *tree_root = fs_info->tree_root;
+	struct btrfs_root *root;
+	struct btrfs_key key;
+	int ret = 0;
+
+	root = kzalloc(sizeof(*root), GFP_KERNEL);
+	if (!root)
+		return ERR_PTR(-ENOMEM);
+
+	btrfs_setup_root(root, fs_info, objectid);
+	root->root_key.objectid = objectid;
+	root->root_key.type = BTRFS_ROOT_ITEM_KEY;
+	root->root_key.offset = 0;
+
+	leaf = btrfs_alloc_free_block(trans, root, fs_info->nodesize, objectid, NULL, 0, 0, 0);
+	if (IS_ERR(leaf)) {
+		ret = PTR_ERR(leaf);
+		leaf = NULL;
+		goto fail;
+	}
+
+	memset_extent_buffer(leaf, 0, 0, sizeof(struct btrfs_header));
+	btrfs_set_header_bytenr(leaf, leaf->start);
+	btrfs_set_header_generation(leaf, trans->transid);
+	btrfs_set_header_backref_rev(leaf, BTRFS_MIXED_BACKREF_REV);
+	btrfs_set_header_owner(leaf, objectid);
+	root->node = leaf;
+	write_extent_buffer(leaf, fs_info->fsid, btrfs_header_fsid(), BTRFS_FSID_SIZE);
+	write_extent_buffer(leaf, fs_info->chunk_tree_uuid,
+			    btrfs_header_chunk_tree_uuid(leaf),
+			    BTRFS_UUID_SIZE);
+	btrfs_mark_buffer_dirty(leaf);
+
+	extent_buffer_get(root->node);
+	root->commit_root = root->node;
+	root->track_dirty = 1;
+
+	root->root_item.flags = 0;
+	root->root_item.byte_limit = 0;
+	btrfs_set_root_bytenr(&root->root_item, leaf->start);
+	btrfs_set_root_generation(&root->root_item, trans->transid);
+	btrfs_set_root_level(&root->root_item, 0);
+	btrfs_set_root_refs(&root->root_item, 1);
+	btrfs_set_root_used(&root->root_item, leaf->len);
+	btrfs_set_root_last_snapshot(&root->root_item, 0);
+	btrfs_set_root_dirid(&root->root_item, 0);
+	memset(root->root_item.uuid, 0, BTRFS_UUID_SIZE);
+	root->root_item.drop_level = 0;
+
+	key.objectid = objectid;
+	key.type = BTRFS_ROOT_ITEM_KEY;
+	key.offset = 0;
+	ret = btrfs_insert_root(trans, tree_root, &key, &root->root_item);
+	if (ret)
+		goto fail;
+
+	return root;
+
+fail:
+	if (leaf)
+		free_extent_buffer(leaf);
+
+	kfree(root);
+	return ERR_PTR(ret);
+}
+
+#define btrfs_set_fs_compat_ro(__fs_info, opt) \
+        __btrfs_set_fs_compat_ro((__fs_info), BTRFS_FEATURE_COMPAT_RO_##opt)
+
+static inline void __btrfs_set_fs_compat_ro(struct btrfs_fs_info *fs_info,
+                                            u64 flag)
+{
+	struct btrfs_super_block *disk_super;
+	u64 features;
+
+	disk_super = fs_info->super_copy;
+	features = btrfs_super_compat_ro_flags(disk_super);
+	if (!(features & flag)) {
+		features = btrfs_super_compat_ro_flags(disk_super);
+		if (!(features & flag)) {
+			features |= flag;
+			btrfs_set_super_compat_ro_flags(disk_super, features);
+		}
+	}
+}
+
+int btrfs_create_free_space_tree(struct btrfs_fs_info *fs_info)
+{
+	struct btrfs_trans_handle *trans;
+	struct btrfs_root *tree_root = fs_info->tree_root;
+	struct btrfs_root *free_space_root;
+	struct btrfs_block_group_cache *block_group;
+	u64 start = BTRFS_SUPER_INFO_OFFSET + BTRFS_SUPER_INFO_SIZE;
+	int ret;
+
+	trans = btrfs_start_transaction(tree_root, 0);
+	if (IS_ERR(trans))
+		return PTR_ERR(trans);
+
+	free_space_root = btrfs_create_tree(trans, fs_info,
+					    BTRFS_FREE_SPACE_TREE_OBJECTID);
+	if (IS_ERR(free_space_root)) {
+		ret = PTR_ERR(free_space_root);
+		goto abort;
+	}
+	fs_info->free_space_root = free_space_root;
+
+	do {
+		block_group = btrfs_lookup_first_block_group(fs_info, start);
+		if (!block_group)
+			break;
+		start = block_group->key.objectid + block_group->key.offset;
+		ret = populate_free_space_tree(trans, block_group);
+		if (ret)
+			goto abort;
+	} while (block_group);
+
+	btrfs_set_fs_compat_ro(fs_info, FREE_SPACE_TREE);
+	btrfs_set_fs_compat_ro(fs_info, FREE_SPACE_TREE_VALID);
+
+	ret = btrfs_commit_transaction(trans, tree_root);
+	if (ret)
+		return ret;
+
+	return 0;
+
+abort:
+	btrfs_abort_transaction(trans, ret);
+	return ret;
+}
+
 int load_free_space_tree(struct btrfs_fs_info *fs_info,
 			 struct btrfs_block_group_cache *block_group)
 {
@@ -332,7 +1573,7 @@ int load_free_space_tree(struct btrfs_fs_info *fs_info,
 	path = btrfs_alloc_path();
 	if (!path)
 		return -ENOMEM;
-	path->reada = 1;
+	path->reada = READA_BACK;
 
 	info = search_free_space_info(NULL, fs_info, block_group, path, 0);
 	if (IS_ERR(info)) {
diff --git a/free-space-tree.h b/free-space-tree.h
index 4845f13e6808..9530c2882358 100644
--- a/free-space-tree.h
+++ b/free-space-tree.h
@@ -19,8 +19,19 @@
 #ifndef __BTRFS_FREE_SPACE_TREE_H__
 #define __BTRFS_FREE_SPACE_TREE_H__
 
+#define BTRFS_FREE_SPACE_BITMAP_SIZE 256
+#define BTRFS_FREE_SPACE_BITMAP_BITS (BTRFS_FREE_SPACE_BITMAP_SIZE * BITS_PER_BYTE)
+
 int btrfs_clear_free_space_tree(struct btrfs_fs_info *fs_info);
 int load_free_space_tree(struct btrfs_fs_info *fs_info,
 			 struct btrfs_block_group_cache *block_group);
-
+int populate_free_space_tree(struct btrfs_trans_handle *trans,
+			     struct btrfs_block_group_cache *block_group);
+int remove_block_group_free_space(struct btrfs_trans_handle *trans,
+				  struct btrfs_block_group_cache *block_group);
+int add_to_free_space_tree(struct btrfs_trans_handle *trans, u64 start,
+			   u64 size);
+int remove_from_free_space_tree(struct btrfs_trans_handle *trans, u64 start,
+				u64 size);
+int btrfs_create_free_space_tree(struct btrfs_fs_info *info);
 #endif
diff --git a/kerncompat.h b/kerncompat.h
index 1a2bc18c3ac2..a223a7f009bd 100644
--- a/kerncompat.h
+++ b/kerncompat.h
@@ -263,6 +263,8 @@ static inline int IS_ERR_OR_NULL(const void *ptr)
 	return !ptr || IS_ERR(ptr);
 }
 
+#define div_u64(x, y) ((x) / (y))
+
 /**
  * swap - swap values of @a and @b
  * @a: first value
@@ -297,6 +299,10 @@ static inline int IS_ERR_OR_NULL(const void *ptr)
 #define kfree(x) free(x)
 #define vmalloc(x) malloc(x)
 #define vfree(x) free(x)
+#define kvzalloc(x, y) kzalloc(x,y)
+#define kvfree(x) free(x)
+#define memalloc_nofs_save() (0)
+#define memalloc_nofs_restore(x)
 
 #ifndef BTRFS_DISABLE_BACKTRACE
 static inline void assert_trace(const char *assertion, const char *filename,
-- 
2.7.4


^ permalink raw reply related	[flat|nested] 27+ messages in thread

* [PATCH 06/10] btrfs-progs: Hook FST code in extent (de)alloc
  2018-10-01 14:46 [PATCH 00/10] Freespace tree repair support v2 Nikolay Borisov
                   ` (4 preceding siblings ...)
  2018-10-01 14:46 ` [PATCH 05/10] btrfs-progs: Pull free space tree related code from kernel Nikolay Borisov
@ 2018-10-01 14:46 ` Nikolay Borisov
  2018-10-01 14:46 ` [PATCH 07/10] btrfs-progs: Add freespace tree as compat_ro supported feature Nikolay Borisov
                   ` (4 subsequent siblings)
  10 siblings, 0 replies; 27+ messages in thread
From: Nikolay Borisov @ 2018-10-01 14:46 UTC (permalink / raw)
  To: linux-btrfs; +Cc: Nikolay Borisov

For now this doesn't change the functionality since FST code is not
yet enabled via the compat bits. But this will be needed when it's
enabled so that the FST is correctly modified during repair operations
that allocate/deallocate  extents.

Signed-off-by: Nikolay Borisov <nborisov@suse.com>
---
 extent-tree.c | 11 +++++++++++
 1 file changed, 11 insertions(+)

diff --git a/extent-tree.c b/extent-tree.c
index b9a30644720b..7adda557f8e6 100644
--- a/extent-tree.c
+++ b/extent-tree.c
@@ -29,6 +29,7 @@
 #include "crc32c.h"
 #include "volumes.h"
 #include "free-space-cache.h"
+#include "free-space-tree.h"
 #include "utils.h"
 
 #define PENDING_EXTENT_INSERT 0
@@ -2276,6 +2277,11 @@ static int __free_extent(struct btrfs_trans_handle *trans,
 			BUG_ON(ret);
 		}
 
+		ret = add_to_free_space_tree(trans, bytenr, num_bytes);
+		if (ret) {
+			goto fail;
+		}
+
 		update_block_group(trans->fs_info, bytenr, num_bytes, 0,
 				   mark_free);
 	}
@@ -2595,6 +2601,11 @@ static int alloc_reserved_tree_block(struct btrfs_trans_handle *trans,
 	btrfs_mark_buffer_dirty(leaf);
 	btrfs_free_path(path);
 
+	ret = remove_from_free_space_tree(trans, ins.objectid,
+					  fs_info->nodesize);
+	if (ret)
+		return ret;
+
 	ret = update_block_group(fs_info, ins.objectid, fs_info->nodesize, 1,
 				 0);
 
-- 
2.7.4


^ permalink raw reply related	[flat|nested] 27+ messages in thread

* [PATCH 07/10] btrfs-progs: Add freespace tree as compat_ro supported feature
  2018-10-01 14:46 [PATCH 00/10] Freespace tree repair support v2 Nikolay Borisov
                   ` (5 preceding siblings ...)
  2018-10-01 14:46 ` [PATCH 06/10] btrfs-progs: Hook FST code in extent (de)alloc Nikolay Borisov
@ 2018-10-01 14:46 ` Nikolay Borisov
  2018-10-04 18:30   ` Omar Sandoval
  2018-10-01 14:46 ` [PATCH 08/10] btrfs-progs: check: Add support for freespace tree fixing Nikolay Borisov
                   ` (3 subsequent siblings)
  10 siblings, 1 reply; 27+ messages in thread
From: Nikolay Borisov @ 2018-10-01 14:46 UTC (permalink / raw)
  To: linux-btrfs; +Cc: Nikolay Borisov

The RO_FREE_SPACE_TREE(_VALID) flags are required in order to be able
to open an FST filesystem in repair mode. Add them to
BTRFS_FEATURE_COMPAT_RO_SUPP.

Signed-off-by: Nikolay Borisov <nborisov@suse.com>
---
 ctree.h | 4 +++-
 1 file changed, 3 insertions(+), 1 deletion(-)

diff --git a/ctree.h b/ctree.h
index a6d6c3decd87..3c396e7d293d 100644
--- a/ctree.h
+++ b/ctree.h
@@ -497,7 +497,9 @@ struct btrfs_super_block {
  * added here until read-write support for the free space tree is implemented in
  * btrfs-progs.
  */
-#define BTRFS_FEATURE_COMPAT_RO_SUPP		0ULL
+#define BTRFS_FEATURE_COMPAT_RO_SUPP			\
+	(BTRFS_FEATURE_COMPAT_RO_FREE_SPACE_TREE |	\
+	 BTRFS_FEATURE_COMPAT_RO_FREE_SPACE_TREE_VALID)
 
 #define BTRFS_FEATURE_INCOMPAT_SUPP			\
 	(BTRFS_FEATURE_INCOMPAT_MIXED_BACKREF |		\
-- 
2.7.4


^ permalink raw reply related	[flat|nested] 27+ messages in thread

* [PATCH 08/10] btrfs-progs: check: Add support for freespace tree fixing
  2018-10-01 14:46 [PATCH 00/10] Freespace tree repair support v2 Nikolay Borisov
                   ` (6 preceding siblings ...)
  2018-10-01 14:46 ` [PATCH 07/10] btrfs-progs: Add freespace tree as compat_ro supported feature Nikolay Borisov
@ 2018-10-01 14:46 ` Nikolay Borisov
  2018-10-04 19:16   ` Omar Sandoval
  2018-10-01 14:46 ` [PATCH 09/10] btrfs-progs: tests: Test for FST corruption detection/repair Nikolay Borisov
                   ` (2 subsequent siblings)
  10 siblings, 1 reply; 27+ messages in thread
From: Nikolay Borisov @ 2018-10-01 14:46 UTC (permalink / raw)
  To: linux-btrfs; +Cc: Nikolay Borisov

Now that all the prerequisite code for proper support of free space
tree repair is in, it's time to wire it in. This is achieved by first
hooking the freespace tree to the __free_extent/alloc_reserved_tree_block
functions. And then introducing a wrapper function to contains the
existing check_space_cache and the newly introduced repair code.
Finally, it's important to note that FST repair code first clears the
existing FST in case of any problem found and rebuilds it from scratch.

Signed-off-by: Nikolay Borisov <nborisov@suse.com>
---
 check/main.c | 47 ++++++++++++++++++++++++++++++-----------------
 1 file changed, 30 insertions(+), 17 deletions(-)

diff --git a/check/main.c b/check/main.c
index b361cd7e26a0..4daf85aad82c 100644
--- a/check/main.c
+++ b/check/main.c
@@ -5392,14 +5392,6 @@ static int check_space_cache(struct btrfs_root *root)
 	int ret;
 	int error = 0;
 
-	if (btrfs_super_cache_generation(root->fs_info->super_copy) != -1ULL &&
-	    btrfs_super_generation(root->fs_info->super_copy) !=
-	    btrfs_super_cache_generation(root->fs_info->super_copy)) {
-		printf("cache and super generation don't match, space cache "
-		       "will be invalidated\n");
-		return 0;
-	}
-
 	while (1) {
 		ctx.item_count++;
 		cache = btrfs_lookup_first_block_group(root->fs_info, start);
@@ -9417,7 +9409,6 @@ static int do_clear_free_space_cache(struct btrfs_fs_info *fs_info,
 			ret = 1;
 			goto close_out;
 		}
-		printf("Clearing free space cache\n");
 		ret = clear_free_space_cache(fs_info);
 		if (ret) {
 			error("failed to clear free space cache");
@@ -9444,6 +9435,35 @@ static int do_clear_free_space_cache(struct btrfs_fs_info *fs_info,
 	return ret;
 }
 
+static int validate_free_space_cache(struct btrfs_root *root)
+{
+
+	int ret;
+
+	if (btrfs_super_cache_generation(root->fs_info->super_copy) != -1ULL &&
+	    btrfs_super_generation(root->fs_info->super_copy) !=
+	    btrfs_super_cache_generation(root->fs_info->super_copy)) {
+		printf("cache and super generation don't match, space cache "
+		       "will be invalidated\n");
+		return 0;
+	}
+
+	ret = check_space_cache(root);
+	if (ret && btrfs_fs_compat_ro(global_info, FREE_SPACE_TREE)
+	    && repair) {
+		ret = do_clear_free_space_cache(global_info, 2);
+		if (ret)
+			goto out;
+
+		ret = btrfs_create_free_space_tree(global_info);
+		if (ret)
+			error("couldn't repair freespace tree");
+	}
+
+out:
+	return ret ? -EINVAL : 0;
+}
+
 const char * const cmd_check_usage[] = {
 	"btrfs check [options] <device>",
 	"Check structural integrity of a filesystem (unmounted).",
@@ -9877,16 +9897,9 @@ int cmd_check(int argc, char **argv)
 		task_start(ctx.info, &ctx.start_time, &ctx.item_count);
 	}
 
-	ret = check_space_cache(root);
+	ret = validate_free_space_cache(root);
 	task_stop(ctx.info);
 	err |= !!ret;
-	if (ret) {
-		if (is_free_space_tree)
-			error("errors found in free space tree");
-		else
-			error("errors found in free space cache");
-		goto out;
-	}
 
 	/*
 	 * We used to have to have these hole extents in between our real
-- 
2.7.4


^ permalink raw reply related	[flat|nested] 27+ messages in thread

* [PATCH 09/10] btrfs-progs: tests: Test for FST corruption detection/repair
  2018-10-01 14:46 [PATCH 00/10] Freespace tree repair support v2 Nikolay Borisov
                   ` (7 preceding siblings ...)
  2018-10-01 14:46 ` [PATCH 08/10] btrfs-progs: check: Add support for freespace tree fixing Nikolay Borisov
@ 2018-10-01 14:46 ` Nikolay Borisov
  2018-10-01 14:46 ` [PATCH 10/10] btrfs-progs: check: Fix wrong error message in case of corrupted bitmap Nikolay Borisov
  2018-10-23 15:00 ` [PATCH 00/10] Freespace tree repair support v2 David Sterba
  10 siblings, 0 replies; 27+ messages in thread
From: Nikolay Borisov @ 2018-10-01 14:46 UTC (permalink / raw)
  To: linux-btrfs; +Cc: Nikolay Borisov

Simple test case which preps a filesystem, then corrupts the FST and
finally repairs it. Tests both extent based and bitmap based FSTs.

Signed-off-by: Nikolay Borisov <nborisov@suse.com>
---
 tests/fsck-tests/036-freespacetree-repair/test.sh | 76 +++++++++++++++++++++++
 1 file changed, 76 insertions(+)
 create mode 100755 tests/fsck-tests/036-freespacetree-repair/test.sh

diff --git a/tests/fsck-tests/036-freespacetree-repair/test.sh b/tests/fsck-tests/036-freespacetree-repair/test.sh
new file mode 100755
index 000000000000..8c91ce9dedbc
--- /dev/null
+++ b/tests/fsck-tests/036-freespacetree-repair/test.sh
@@ -0,0 +1,76 @@
+#!/bin/bash
+# Corrupt a filesystem that is using freespace tree and then ensure that
+# btrfs check is able to repair it. This tests correct detection/repair of
+# both a FREE_SPACE_EXTENT based FST and a FREE_SPACE_BITMAP based FST.
+
+source "$TEST_TOP/common"
+
+# wrapper for btrfs-corrupt-item
+# $1: Type of item we want to corrupt - extent or bitmap
+corrupt_fst_item()
+{
+	local type
+	local objectid
+	local offset
+	type="$1"
+
+	if [[ $type == "bitmap" ]]; then
+		type=200
+		objectid=$("$TOP/btrfs" inspect-internal dump-tree -t 10 "$TEST_DEV" | \
+			grep -o "[[:digit:]]* FREE_SPACE_BITMAP [[:digit:]]*" | \
+			cut -d' ' -f1 | tail -2 | head -1)
+		offset=$("$TOP/btrfs" inspect-internal dump-tree -t 10 "$TEST_DEV" | \
+			grep -o "[[:digit:]]* FREE_SPACE_BITMAP [[:digit:]]*" | \
+			cut -d' ' -f3 |tail -2 | head -1)
+		echo "Corrupting $objectid,FREE_SPACE_BITMAP,$offset" >> "$RESULTS"
+	elif [[ $type == "extent" ]]; then
+		type=199
+		objectid=$("$TOP/btrfs" inspect-internal dump-tree -t 10 "$TEST_DEV" | \
+			grep -o "[[:digit:]]* FREE_SPACE_EXTENT [[:digit:]]*" | \
+			cut -d' ' -f1 | tail -2 | head -1)
+		offset=$("$TOP/btrfs" inspect-internal dump-tree -t 10 "$TEST_DEV" | \
+			grep -o "[[:digit:]]* FREE_SPACE_EXTENT [[:digit:]]*" | \
+			cut -d' ' -f3 | tail -2 | head -1)
+		echo "Corrupting $objectid,FREE_SPACE_EXTENT,$offset" >> "$RESULTS"
+	else
+		_fail "Unknown item type for corruption"
+	fi
+
+	run_check "$TOP/btrfs-corrupt-block" -r 10 -K "$objectid,$type,$offset" \
+		-f offset "$TEST_DEV"
+}
+
+check_prereq btrfs
+check_prereq mkfs.btrfs
+check_global_prereq grep
+check_global_prereq tail
+check_global_prereq head
+check_global_prereq cut
+
+setup_root_helper
+prepare_test_dev 256M
+
+run_check "$TOP/mkfs.btrfs" -n 4k -f "$TEST_DEV"
+run_check_mount_test_dev -oclear_cache,space_cache=v2
+
+#create files which will populate the FST
+for i in {1..3000}; do
+	fallocate -l 4k "$TEST_MNT/file.$i"
+done
+
+run_check_umount_test_dev
+
+#now corrupt one of the bitmap items
+corrupt_fst_item "bitmap"
+check_image "$TEST_DEV"
+
+# change the freespace such that we now have at least one free_space_extent
+# object
+run_check_mount_test_dev
+rm -rf "$TEST_MNT/file.*"
+fallocate -l 50m "$TEST_MNT/file"
+run_check_umount_test_dev
+
+#now corrupt an extent
+corrupt_fst_item "extent"
+check_image "$TEST_DEV"
-- 
2.7.4


^ permalink raw reply related	[flat|nested] 27+ messages in thread

* [PATCH 10/10] btrfs-progs: check: Fix wrong error message in case of corrupted bitmap
  2018-10-01 14:46 [PATCH 00/10] Freespace tree repair support v2 Nikolay Borisov
                   ` (8 preceding siblings ...)
  2018-10-01 14:46 ` [PATCH 09/10] btrfs-progs: tests: Test for FST corruption detection/repair Nikolay Borisov
@ 2018-10-01 14:46 ` Nikolay Borisov
  2018-10-04 19:18   ` Omar Sandoval
  2018-10-23 15:00 ` [PATCH 00/10] Freespace tree repair support v2 David Sterba
  10 siblings, 1 reply; 27+ messages in thread
From: Nikolay Borisov @ 2018-10-01 14:46 UTC (permalink / raw)
  To: linux-btrfs; +Cc: Nikolay Borisov

Similarly to the fix in e444c7bfa65f ("btrfs-progs: check: Fix wrong
error message in case of corrupted extent") this commits addresses the
same problem but for corrupted bitmap objects.

Signed-off-by: Nikolay Borisov <nborisov@suse.com>
---
 free-space-tree.c | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)

diff --git a/free-space-tree.c b/free-space-tree.c
index 3b7e8a3fe4f5..690da44a739d 100644
--- a/free-space-tree.c
+++ b/free-space-tree.c
@@ -1302,7 +1302,7 @@ static int load_free_space_bitmaps(struct btrfs_fs_info *fs_info,
 		if (key.objectid + key.offset > end) {
 			fprintf(stderr,
 	"free space bitmap ends at %llu, beyond end of block group %llu-%llu\n",
-				key.objectid, start, end);
+				key.objectid + key.offset, start, end);
 			(*errors)++;
 			break;
 		}
-- 
2.7.4


^ permalink raw reply related	[flat|nested] 27+ messages in thread

* Re: [PATCH 01/10] btrfs-progs: Add support for freespace tree in btrfs_read_fs_root
  2018-10-01 14:46 ` [PATCH 01/10] btrfs-progs: Add support for freespace tree in btrfs_read_fs_root Nikolay Borisov
@ 2018-10-02 19:20   ` Omar Sandoval
  2018-10-04  1:05   ` Su Yue
  1 sibling, 0 replies; 27+ messages in thread
From: Omar Sandoval @ 2018-10-02 19:20 UTC (permalink / raw)
  To: Nikolay Borisov; +Cc: linux-btrfs

On Mon, Oct 01, 2018 at 05:46:12PM +0300, Nikolay Borisov wrote:
> For completeness sake add code to btrfs_read_fs_root so that it can
> handle the freespace tree.

Reviewed-by: Omar Sandoval <osandov@fb.com>

> Signed-off-by: Nikolay Borisov <nborisov@suse.com>
> ---
>  disk-io.c | 3 +++
>  1 file changed, 3 insertions(+)
> 
> diff --git a/disk-io.c b/disk-io.c
> index 2e6d56a36af9..14f0fd5c2f0c 100644
> --- a/disk-io.c
> +++ b/disk-io.c
> @@ -668,6 +668,9 @@ struct btrfs_root *btrfs_read_fs_root(struct btrfs_fs_info *fs_info,
>  	if (location->objectid == BTRFS_QUOTA_TREE_OBJECTID)
>  		return fs_info->quota_enabled ? fs_info->quota_root :
>  				ERR_PTR(-ENOENT);
> +	if (location->objectid == BTRFS_FREE_SPACE_TREE_OBJECTID)
> +        return fs_info->free_space_root ? fs_info->free_space_root :
> +                                                  ERR_PTR(-ENOENT);
>  
>  	BUG_ON(location->objectid == BTRFS_TREE_RELOC_OBJECTID ||
>  	       location->offset != (u64)-1);
> -- 
> 2.7.4
> 

^ permalink raw reply	[flat|nested] 27+ messages in thread

* Re: [PATCH 02/10] btrfs-progs: Add extent buffer bitmap manipulation infrastructure
  2018-10-01 14:46 ` [PATCH 02/10] btrfs-progs: Add extent buffer bitmap manipulation infrastructure Nikolay Borisov
@ 2018-10-02 19:24   ` Omar Sandoval
  2018-10-04  1:31   ` Su Yue
  1 sibling, 0 replies; 27+ messages in thread
From: Omar Sandoval @ 2018-10-02 19:24 UTC (permalink / raw)
  To: Nikolay Borisov; +Cc: linux-btrfs

On Mon, Oct 01, 2018 at 05:46:13PM +0300, Nikolay Borisov wrote:
> Those functions are in preparation for adding the freespace tree
> repair code since it needs to be able to deal with bitmap based fsts.
> This patch adds extent_buffer_bitmap_set and extent_buffer_bitmap_clear
> functions. Since in userspace we don't have to deal with page mappings
> their implementation is vastly simplified by simply setting each bit in
> the passed range.

Reviewed-by: Omar Sandoval <osandov@fb.com>

> Signed-off-by: Nikolay Borisov <nborisov@suse.com>
> ---
>  extent_io.c | 56 ++++++++++++++++++++++++++++++++++++++++++++++++++++++++
>  extent_io.h |  4 ++++
>  2 files changed, 60 insertions(+)
> 
> diff --git a/extent_io.c b/extent_io.c
> index 198492699438..de47c2c59ae9 100644
> --- a/extent_io.c
> +++ b/extent_io.c
> @@ -204,6 +204,62 @@ static int clear_state_bit(struct extent_io_tree *tree,
>  	return ret;
>  }
>  
> +/**
> + * extent_buffer_bitmap_set - set an area of a bitmap
> + * @eb: the extent buffer
> + * @start: offset of the bitmap item in the extent buffer
> + * @pos: bit number of the first bit
> + * @len: number of bits to set
> + */
> +void extent_buffer_bitmap_set(struct extent_buffer *eb, unsigned long start,
> +                              unsigned long pos, unsigned long len)
> +{
> +	u8 *p = (u8 *)eb->data + start + BIT_BYTE(pos);
> +	const unsigned int size = pos + len;
> +	int bits_to_set = BITS_PER_BYTE - (pos % BITS_PER_BYTE);
> +	u8 mask_to_set = BITMAP_FIRST_BYTE_MASK(pos);
> +
> +	while (len >= bits_to_set) {
> +		*p |= mask_to_set;
> +		len -= bits_to_set;
> +		bits_to_set = BITS_PER_BYTE;
> +		mask_to_set = ~0;
> +		p++;
> +	}
> +	if (len) {
> +		mask_to_set &= BITMAP_LAST_BYTE_MASK(size);
> +		*p |= mask_to_set;
> +	}
> +}
> +
> +
> +/**
> + * extent_buffer_bitmap_clear - clear an area of a bitmap
> + * @eb: the extent buffer
> + * @start: offset of the bitmap item in the extent buffer
> + * @pos: bit number of the first bit
> + * @len: number of bits to clear
> + */
> +void extent_buffer_bitmap_clear(struct extent_buffer *eb, unsigned long start,
> +                                unsigned long pos, unsigned long len)
> +{
> +	u8 *p = (u8 *)eb->data + start + BIT_BYTE(pos);
> +	const unsigned int size = pos + len;
> +	int bits_to_clear = BITS_PER_BYTE - (pos % BITS_PER_BYTE);
> +	u8 mask_to_clear = BITMAP_FIRST_BYTE_MASK(pos);
> +
> +	while (len >= bits_to_clear) {
> +		*p &= ~mask_to_clear;
> +		len -= bits_to_clear;
> +		bits_to_clear = BITS_PER_BYTE;
> +		mask_to_clear = ~0;
> +		p++;
> +	}
> +	if (len) {
> +		mask_to_clear &= BITMAP_LAST_BYTE_MASK(size);
> +		*p &= ~mask_to_clear;
> +	}
> +}
>  /*
>   * clear some bits on a range in the tree.
>   */
> diff --git a/extent_io.h b/extent_io.h
> index d407d93d617e..b67c6fc40e89 100644
> --- a/extent_io.h
> +++ b/extent_io.h
> @@ -175,4 +175,8 @@ int read_data_from_disk(struct btrfs_fs_info *info, void *buf, u64 offset,
>  			u64 bytes, int mirror);
>  int write_data_to_disk(struct btrfs_fs_info *info, void *buf, u64 offset,
>  		       u64 bytes, int mirror);
> +void extent_buffer_bitmap_clear(struct extent_buffer *eb, unsigned long start,
> +                                unsigned long pos, unsigned long len);
> +void extent_buffer_bitmap_set(struct extent_buffer *eb, unsigned long start,
> +                              unsigned long pos, unsigned long len);
>  #endif
> -- 
> 2.7.4
> 

^ permalink raw reply	[flat|nested] 27+ messages in thread

* Re: [PATCH 03/10] btrfs-progs: Replace homegrown bitops related functions with kernel counterparts
  2018-10-01 14:46 ` [PATCH 03/10] btrfs-progs: Replace homegrown bitops related functions with kernel counterparts Nikolay Borisov
@ 2018-10-02 23:32   ` Omar Sandoval
  0 siblings, 0 replies; 27+ messages in thread
From: Omar Sandoval @ 2018-10-02 23:32 UTC (permalink / raw)
  To: Nikolay Borisov; +Cc: linux-btrfs

On Mon, Oct 01, 2018 at 05:46:14PM +0300, Nikolay Borisov wrote:
> Replace existing find_*_bit functions with kernel equivalent. This
> reduces duplication, simplifies the code (we really have one worker
> function _find_next_bit) and is quite likely faster. No functional
> changes.

Reviewed-by: Omar Sandoval <osandov@fb.com>

> Signed-off-by: Nikolay Borisov <nborisov@suse.com>
> ---
>  kernel-lib/bitops.h | 142 +++++++++++++++++-----------------------------------
>  1 file changed, 46 insertions(+), 96 deletions(-)
> 
> diff --git a/kernel-lib/bitops.h b/kernel-lib/bitops.h
> index 5b35f9fc5213..78256adf55be 100644
> --- a/kernel-lib/bitops.h
> +++ b/kernel-lib/bitops.h
> @@ -2,6 +2,7 @@
>  #define _PERF_LINUX_BITOPS_H_
>  
>  #include <linux/kernel.h>
> +#include "internal.h"
>  
>  #ifndef DIV_ROUND_UP
>  #define DIV_ROUND_UP(n, d) (((n) + (d) - 1) / (d))
> @@ -109,116 +110,65 @@ static __always_inline unsigned long __ffs(unsigned long word)
>  
>  #define ffz(x) __ffs(~(x))
>  
> +#define BITMAP_FIRST_WORD_MASK(start) (~0UL << ((start) & (BITS_PER_LONG - 1))) 
> +#define BITMAP_LAST_WORD_MASK(nbits) (~0UL >> (-(nbits) & (BITS_PER_LONG - 1)))
> +
>  /*
> - * Find the first set bit in a memory region.
> + * This is a common helper function for find_next_bit, find_next_zero_bit, and
> + * find_next_and_bit. The differences are:
> + *  - The "invert" argument, which is XORed with each fetched word before
> + *    searching it for one bits.
> + *  - The optional "addr2", which is anded with "addr1" if present.
>   */
> -static inline unsigned long
> -find_first_bit(const unsigned long *addr, unsigned long size)
> +static inline unsigned long _find_next_bit(const unsigned long *addr1,
> +		const unsigned long *addr2, unsigned long nbits,
> +		unsigned long start, unsigned long invert)
>  {
> -	const unsigned long *p = addr;
> -	unsigned long result = 0;
>  	unsigned long tmp;
>  
> -	while (size & ~(BITS_PER_LONG-1)) {
> -		if ((tmp = *(p++)))
> -			goto found;
> -		result += BITS_PER_LONG;
> -		size -= BITS_PER_LONG;
> +	if (start >= nbits)
> +		return nbits;
> +
> +	tmp = addr1[start / BITS_PER_LONG];
> +	if (addr2)
> +		tmp &= addr2[start / BITS_PER_LONG];
> +	tmp ^= invert;
> +
> +	/* Handle 1st word. */
> +	tmp &= BITMAP_FIRST_WORD_MASK(start);
> +	start = round_down(start, BITS_PER_LONG);
> +
> +	while (!tmp) {
> +		start += BITS_PER_LONG;
> +		if (start >= nbits)
> +			return nbits;
> +
> +		tmp = addr1[start / BITS_PER_LONG];
> +		if (addr2)
> +			tmp &= addr2[start / BITS_PER_LONG];
> +		tmp ^= invert;
>  	}
> -	if (!size)
> -		return result;
> -
> -	tmp = (*p) & (~0UL >> (BITS_PER_LONG - size));
> -	if (tmp == 0UL)		/* Are any bits set? */
> -		return result + size;	/* Nope. */
> -found:
> -	return result + __ffs(tmp);
> +
> +	return min(start + __ffs(tmp), nbits);
>  }
>  
>  /*
>   * Find the next set bit in a memory region.
>   */
> -static inline unsigned long
> -find_next_bit(const unsigned long *addr, unsigned long size,
> -	      unsigned long offset)
> +static inline unsigned long find_next_bit(const unsigned long *addr,
> +					  unsigned long size,
> +					  unsigned long offset)
>  {
> -	const unsigned long *p = addr + BITOP_WORD(offset);
> -	unsigned long result = offset & ~(BITS_PER_LONG-1);
> -	unsigned long tmp;
> -
> -	if (offset >= size)
> -		return size;
> -	size -= result;
> -	offset %= BITS_PER_LONG;
> -	if (offset) {
> -		tmp = *(p++);
> -		tmp &= (~0UL << offset);
> -		if (size < BITS_PER_LONG)
> -			goto found_first;
> -		if (tmp)
> -			goto found_middle;
> -		size -= BITS_PER_LONG;
> -		result += BITS_PER_LONG;
> -	}
> -	while (size & ~(BITS_PER_LONG-1)) {
> -		if ((tmp = *(p++)))
> -			goto found_middle;
> -		result += BITS_PER_LONG;
> -		size -= BITS_PER_LONG;
> -	}
> -	if (!size)
> -		return result;
> -	tmp = *p;
> -
> -found_first:
> -	tmp &= (~0UL >> (BITS_PER_LONG - size));
> -	if (tmp == 0UL)		/* Are any bits set? */
> -		return result + size;	/* Nope. */
> -found_middle:
> -	return result + __ffs(tmp);
> +	return _find_next_bit(addr, NULL, size, offset, 0UL);
>  }
>  
> -/*
> - * This implementation of find_{first,next}_zero_bit was stolen from
> - * Linus' asm-alpha/bitops.h.
> - */
> -static inline unsigned long
> -find_next_zero_bit(const unsigned long *addr, unsigned long size,
> -		   unsigned long offset)
> +static inline unsigned long find_next_zero_bit(const unsigned long *addr,
> +					       unsigned long size,
> +					       unsigned long offset)
>  {
> -	const unsigned long *p = addr + BITOP_WORD(offset);
> -	unsigned long result = offset & ~(BITS_PER_LONG-1);
> -	unsigned long tmp;
> -
> -	if (offset >= size)
> -		return size;
> -	size -= result;
> -	offset %= BITS_PER_LONG;
> -	if (offset) {
> -		tmp = *(p++);
> -		tmp |= ~0UL >> (BITS_PER_LONG - offset);
> -		if (size < BITS_PER_LONG)
> -			goto found_first;
> -		if (~tmp)
> -			goto found_middle;
> -		size -= BITS_PER_LONG;
> -		result += BITS_PER_LONG;
> -	}
> -	while (size & ~(BITS_PER_LONG-1)) {
> -		if (~(tmp = *(p++)))
> -			goto found_middle;
> -		result += BITS_PER_LONG;
> -		size -= BITS_PER_LONG;
> -	}
> -	if (!size)
> -		return result;
> -	tmp = *p;
> -
> -found_first:
> -	tmp |= ~0UL << size;
> -	if (tmp == ~0UL)	/* Are any bits zero? */
> -		return result + size;	/* Nope. */
> -found_middle:
> -	return result + ffz(tmp);
> +	return _find_next_bit(addr, NULL, size, offset, ~0UL);
>  }
> +
> +#define find_first_bit(addr, size) find_next_bit((addr), (size), 0)
> +
>  #endif
> -- 
> 2.7.4
> 

^ permalink raw reply	[flat|nested] 27+ messages in thread

* Re: [PATCH 01/10] btrfs-progs: Add support for freespace tree in btrfs_read_fs_root
  2018-10-01 14:46 ` [PATCH 01/10] btrfs-progs: Add support for freespace tree in btrfs_read_fs_root Nikolay Borisov
  2018-10-02 19:20   ` Omar Sandoval
@ 2018-10-04  1:05   ` Su Yue
  1 sibling, 0 replies; 27+ messages in thread
From: Su Yue @ 2018-10-04  1:05 UTC (permalink / raw)
  To: Nikolay Borisov, linux-btrfs



On 10/1/18 10:46 PM, Nikolay Borisov wrote:
> For completeness sake add code to btrfs_read_fs_root so that it can
> handle the freespace tree.
> 
> Signed-off-by: Nikolay Borisov <nborisov@suse.com>

Reviewed-by: Su Yue <suy.fnst@cn.fujitsu.com>
> ---
>   disk-io.c | 3 +++
>   1 file changed, 3 insertions(+)
> 
> diff --git a/disk-io.c b/disk-io.c
> index 2e6d56a36af9..14f0fd5c2f0c 100644
> --- a/disk-io.c
> +++ b/disk-io.c
> @@ -668,6 +668,9 @@ struct btrfs_root *btrfs_read_fs_root(struct btrfs_fs_info *fs_info,
>   	if (location->objectid == BTRFS_QUOTA_TREE_OBJECTID)
>   		return fs_info->quota_enabled ? fs_info->quota_root :
>   				ERR_PTR(-ENOENT);
> +	if (location->objectid == BTRFS_FREE_SPACE_TREE_OBJECTID)
> +        return fs_info->free_space_root ? fs_info->free_space_root :
> +                                                  ERR_PTR(-ENOENT);
>   
>   	BUG_ON(location->objectid == BTRFS_TREE_RELOC_OBJECTID ||
>   	       location->offset != (u64)-1);
> 



^ permalink raw reply	[flat|nested] 27+ messages in thread

* Re: [PATCH 02/10] btrfs-progs: Add extent buffer bitmap manipulation infrastructure
  2018-10-01 14:46 ` [PATCH 02/10] btrfs-progs: Add extent buffer bitmap manipulation infrastructure Nikolay Borisov
  2018-10-02 19:24   ` Omar Sandoval
@ 2018-10-04  1:31   ` Su Yue
  1 sibling, 0 replies; 27+ messages in thread
From: Su Yue @ 2018-10-04  1:31 UTC (permalink / raw)
  To: Nikolay Borisov, linux-btrfs



On 10/1/18 10:46 PM, Nikolay Borisov wrote:
> Those functions are in preparation for adding the freespace tree
> repair code since it needs to be able to deal with bitmap based fsts.
> This patch adds extent_buffer_bitmap_set and extent_buffer_bitmap_clear
> functions. Since in userspace we don't have to deal with page mappings
> their implementation is vastly simplified by simply setting each bit in
> the passed range.
> 
> Signed-off-by: Nikolay Borisov <nborisov@suse.com>

Reviewed-by: Su Yue <suy.fnst@cn.fujitsu.com>
> ---
>   extent_io.c | 56 ++++++++++++++++++++++++++++++++++++++++++++++++++++++++
>   extent_io.h |  4 ++++
>   2 files changed, 60 insertions(+)
> 
> diff --git a/extent_io.c b/extent_io.c
> index 198492699438..de47c2c59ae9 100644
> --- a/extent_io.c
> +++ b/extent_io.c
> @@ -204,6 +204,62 @@ static int clear_state_bit(struct extent_io_tree *tree,
>   	return ret;
>   }
>   
> +/**
> + * extent_buffer_bitmap_set - set an area of a bitmap
> + * @eb: the extent buffer
> + * @start: offset of the bitmap item in the extent buffer
> + * @pos: bit number of the first bit
> + * @len: number of bits to set
> + */
> +void extent_buffer_bitmap_set(struct extent_buffer *eb, unsigned long start,
> +                              unsigned long pos, unsigned long len)
> +{
> +	u8 *p = (u8 *)eb->data + start + BIT_BYTE(pos);
> +	const unsigned int size = pos + len;
> +	int bits_to_set = BITS_PER_BYTE - (pos % BITS_PER_BYTE);
> +	u8 mask_to_set = BITMAP_FIRST_BYTE_MASK(pos);
> +
> +	while (len >= bits_to_set) {
> +		*p |= mask_to_set;
> +		len -= bits_to_set;
> +		bits_to_set = BITS_PER_BYTE;
> +		mask_to_set = ~0;
> +		p++;
> +	}
> +	if (len) {
> +		mask_to_set &= BITMAP_LAST_BYTE_MASK(size);
> +		*p |= mask_to_set;
> +	}
> +}
> +
> +
> +/**
> + * extent_buffer_bitmap_clear - clear an area of a bitmap
> + * @eb: the extent buffer
> + * @start: offset of the bitmap item in the extent buffer
> + * @pos: bit number of the first bit
> + * @len: number of bits to clear
> + */
> +void extent_buffer_bitmap_clear(struct extent_buffer *eb, unsigned long start,
> +                                unsigned long pos, unsigned long len)
> +{
> +	u8 *p = (u8 *)eb->data + start + BIT_BYTE(pos);
> +	const unsigned int size = pos + len;
> +	int bits_to_clear = BITS_PER_BYTE - (pos % BITS_PER_BYTE);
> +	u8 mask_to_clear = BITMAP_FIRST_BYTE_MASK(pos);
> +
> +	while (len >= bits_to_clear) {
> +		*p &= ~mask_to_clear;
> +		len -= bits_to_clear;
> +		bits_to_clear = BITS_PER_BYTE;
> +		mask_to_clear = ~0;
> +		p++;
> +	}
> +	if (len) {
> +		mask_to_clear &= BITMAP_LAST_BYTE_MASK(size);
> +		*p &= ~mask_to_clear;
> +	}
> +}
>   /*
>    * clear some bits on a range in the tree.
>    */
> diff --git a/extent_io.h b/extent_io.h
> index d407d93d617e..b67c6fc40e89 100644
> --- a/extent_io.h
> +++ b/extent_io.h
> @@ -175,4 +175,8 @@ int read_data_from_disk(struct btrfs_fs_info *info, void *buf, u64 offset,
>   			u64 bytes, int mirror);
>   int write_data_to_disk(struct btrfs_fs_info *info, void *buf, u64 offset,
>   		       u64 bytes, int mirror);
> +void extent_buffer_bitmap_clear(struct extent_buffer *eb, unsigned long start,
> +                                unsigned long pos, unsigned long len);
> +void extent_buffer_bitmap_set(struct extent_buffer *eb, unsigned long start,
> +                              unsigned long pos, unsigned long len);
>   #endif
> 



^ permalink raw reply	[flat|nested] 27+ messages in thread

* Re: [PATCH 04/10] btrfs-progs: Implement find_*_bit_le operations
  2018-10-01 14:46 ` [PATCH 04/10] btrfs-progs: Implement find_*_bit_le operations Nikolay Borisov
@ 2018-10-04 18:08   ` Omar Sandoval
  2018-10-04 18:09     ` Nikolay Borisov
  0 siblings, 1 reply; 27+ messages in thread
From: Omar Sandoval @ 2018-10-04 18:08 UTC (permalink / raw)
  To: Nikolay Borisov; +Cc: linux-btrfs

On Mon, Oct 01, 2018 at 05:46:15PM +0300, Nikolay Borisov wrote:
> This commit introduces explicit little endian bit operations. The only
> difference with the existing bitops implementation is that bswap(32|64)
> is called when the _le versions are invoked on a big-endian machine.
> This is in preparation for adding free space tree conversion support.

I had to check, but it looks like these are also pulled from the kernel
source, so

Reviewed-by: Omar Sandoval <osandov@fb.com>

> Signed-off-by: Nikolay Borisov <nborisov@suse.com>
> ---
>  kernel-lib/bitops.h | 82 +++++++++++++++++++++++++++++++++++++++++++++++++++++
>  1 file changed, 82 insertions(+)

^ permalink raw reply	[flat|nested] 27+ messages in thread

* Re: [PATCH 04/10] btrfs-progs: Implement find_*_bit_le operations
  2018-10-04 18:08   ` Omar Sandoval
@ 2018-10-04 18:09     ` Nikolay Borisov
  0 siblings, 0 replies; 27+ messages in thread
From: Nikolay Borisov @ 2018-10-04 18:09 UTC (permalink / raw)
  To: Omar Sandoval; +Cc: linux-btrfs



On  4.10.2018 21:08, Omar Sandoval wrote:
> On Mon, Oct 01, 2018 at 05:46:15PM +0300, Nikolay Borisov wrote:
>> This commit introduces explicit little endian bit operations. The only
>> difference with the existing bitops implementation is that bswap(32|64)
>> is called when the _le versions are invoked on a big-endian machine.
>> This is in preparation for adding free space tree conversion support.
> 
> I had to check, but it looks like these are also pulled from the kernel
> source, so


Indeed and I did spend time to convince myself how the code works and
that it works :)

> 
> Reviewed-by: Omar Sandoval <osandov@fb.com>
> 
>> Signed-off-by: Nikolay Borisov <nborisov@suse.com>
>> ---
>>  kernel-lib/bitops.h | 82 +++++++++++++++++++++++++++++++++++++++++++++++++++++
>>  1 file changed, 82 insertions(+)
> 

^ permalink raw reply	[flat|nested] 27+ messages in thread

* Re: [PATCH 05/10] btrfs-progs: Pull free space tree related code from kernel
  2018-10-01 14:46 ` [PATCH 05/10] btrfs-progs: Pull free space tree related code from kernel Nikolay Borisov
@ 2018-10-04 18:26   ` Omar Sandoval
  2018-10-04 18:34     ` Nikolay Borisov
  2018-10-23 14:05     ` David Sterba
  0 siblings, 2 replies; 27+ messages in thread
From: Omar Sandoval @ 2018-10-04 18:26 UTC (permalink / raw)
  To: Nikolay Borisov; +Cc: linux-btrfs

On Mon, Oct 01, 2018 at 05:46:16PM +0300, Nikolay Borisov wrote:
> To help implement free space tree checker in user space some kernel
> function are necessary, namely iterating/deleting/adding freespace
> items, some internal search functions. Functions to populate a block
> group based on the extent tree. The code is largely copy/paste from
> the kernel with locking eliminated (i.e free_space_lock). It supports
> reading/writing of both bitmap and extent based FST trees.

For some reason, a lot of this added code uses spaces instead of tabs,
so I had to fix that in order to compare it to the kernel code (some of
the functions were reordered, too).

The only functional difference I noticed was that this is missing the
code to insert the block group header in the free space tree:

	if (block_group->needs_free_space) {
		ret = __add_block_group_free_space(trans, block_group, path);
		if (ret)
			return ret;
	}

Was that intentionally omitted? Without it, the free space tree is
pretty broken :(

> Signed-off-by: Nikolay Borisov <nborisov@suse.com>
> ---
>  ctree.c           |   77 ++++
>  ctree.h           |   15 +
>  free-space-tree.c | 1253 ++++++++++++++++++++++++++++++++++++++++++++++++++++-
>  free-space-tree.h |   13 +-
>  kerncompat.h      |    6 +
>  5 files changed, 1357 insertions(+), 7 deletions(-)
> 
> diff --git a/ctree.c b/ctree.c
> index d8a6883aa85f..aa1568620205 100644
> --- a/ctree.c
> +++ b/ctree.c
> @@ -1226,6 +1226,83 @@ int btrfs_search_slot(struct btrfs_trans_handle *trans, struct btrfs_root
>  }
>  
>  /*
> + * helper to use instead of search slot if no exact match is needed but
> + * instead the next or previous item should be returned.
> + * When find_higher is true, the next higher item is returned, the next lower
> + * otherwise.
> + * When return_any and find_higher are both true, and no higher item is found,
> + * return the next lower instead.
> + * When return_any is true and find_higher is false, and no lower item is found,
> + * return the next higher instead.
> + * It returns 0 if any item is found, 1 if none is found (tree empty), and
> + * < 0 on error
> + */
> +int btrfs_search_slot_for_read(struct btrfs_root *root,
> +                               const struct btrfs_key *key,
> +                               struct btrfs_path *p, int find_higher,
> +                               int return_any)
> +{
> +        int ret;
> +        struct extent_buffer *leaf;
> +
> +again:
> +        ret = btrfs_search_slot(NULL, root, key, p, 0, 0);
> +        if (ret <= 0)
> +                return ret;
> +        /*
> +         * a return value of 1 means the path is at the position where the
> +         * item should be inserted. Normally this is the next bigger item,
> +         * but in case the previous item is the last in a leaf, path points
> +         * to the first free slot in the previous leaf, i.e. at an invalid
> +         * item.
> +         */
> +        leaf = p->nodes[0];
> +
> +        if (find_higher) {
> +                if (p->slots[0] >= btrfs_header_nritems(leaf)) {
> +                        ret = btrfs_next_leaf(root, p);
> +                        if (ret <= 0)
> +                                return ret;
> +                        if (!return_any)
> +                                return 1;
> +                        /*
> +                         * no higher item found, return the next
> +                         * lower instead
> +                         */
> +                        return_any = 0;
> +                        find_higher = 0;
> +                        btrfs_release_path(p);
> +                        goto again;
> +                }
> +        } else {
> +                if (p->slots[0] == 0) {
> +                        ret = btrfs_prev_leaf(root, p);
> +                        if (ret < 0)
> +                                return ret;
> +                        if (!ret) {
> +                                leaf = p->nodes[0];
> +                                if (p->slots[0] == btrfs_header_nritems(leaf))
> +                                        p->slots[0]--;
> +                                return 0;
> +                        }
> +                        if (!return_any)
> +                                return 1;
> +                        /*
> +                         * no lower item found, return the next
> +                         * higher instead
> +                         */
> +                        return_any = 0;
> +                        find_higher = 1;
> +                        btrfs_release_path(p);
> +                        goto again;
> +                } else {
> +                        --p->slots[0];
> +                }
> +        }
> +        return 0;
> +}
> +
> +/*
>   * adjust the pointers going up the tree, starting at level
>   * making sure the right key of each node is points to 'key'.
>   * This is used after shifting pointers to the left, so it stops
> diff --git a/ctree.h b/ctree.h
> index 49f0f5181512..a6d6c3decd87 100644
> --- a/ctree.h
> +++ b/ctree.h
> @@ -1071,6 +1071,17 @@ struct btrfs_block_group_cache {
>  	u64 flags;
>  	int cached;
>  	int ro;
> +	/*
> +         * If the free space extent count exceeds this number, convert the block
> +         * group to bitmaps.
> +         */
> +        u32 bitmap_high_thresh;
> +        /*
> +         * If the free space extent count drops below this number, convert the
> +         * block group back to extents.
> +         */
> +        u32 bitmap_low_thresh;
> +
>  };
>  
>  struct btrfs_device;
> @@ -2596,6 +2607,10 @@ int btrfs_split_item(struct btrfs_trans_handle *trans,
>  int btrfs_search_slot(struct btrfs_trans_handle *trans, struct btrfs_root
>  		      *root, struct btrfs_key *key, struct btrfs_path *p, int
>  		      ins_len, int cow);
> +int btrfs_search_slot_for_read(struct btrfs_root *root,
> +                               const struct btrfs_key *key,
> +                               struct btrfs_path *p, int find_higher,
> +                               int return_any);
>  int btrfs_find_item(struct btrfs_root *fs_root, struct btrfs_path *found_path,
>  		u64 iobjectid, u64 ioff, u8 key_type,
>  		struct btrfs_key *found_key);
> diff --git a/free-space-tree.c b/free-space-tree.c
> index b439b6b43146..3b7e8a3fe4f5 100644
> --- a/free-space-tree.c
> +++ b/free-space-tree.c
> @@ -21,6 +21,37 @@
>  #include "free-space-cache.h"
>  #include "free-space-tree.h"
>  #include "transaction.h"
> +#include "bitops.h"
> +#include "internal.h"
> +
> +void set_free_space_tree_thresholds(struct btrfs_block_group_cache *cache,
> +				    u64 sectorsize)
> +{
> +	u32 bitmap_range;
> +	size_t bitmap_size;
> +	u64 num_bitmaps, total_bitmap_size;
> +
> +	/*
> +	 * We convert to bitmaps when the disk space required for using extents
> +	 * exceeds that required for using bitmaps.
> +	 */
> +	bitmap_range = sectorsize * BTRFS_FREE_SPACE_BITMAP_BITS;
> +	num_bitmaps = div_u64(cache->key.offset + bitmap_range - 1,
> +			      bitmap_range);
> +	bitmap_size = sizeof(struct btrfs_item) + BTRFS_FREE_SPACE_BITMAP_SIZE;
> +	total_bitmap_size = num_bitmaps * bitmap_size;
> +	cache->bitmap_high_thresh = div_u64(total_bitmap_size,
> +					    sizeof(struct btrfs_item));
> +
> +	/*
> +	 * We allow for a small buffer between the high threshold and low
> +	 * threshold to avoid thrashing back and forth between the two formats.
> +	 */
> +	if (cache->bitmap_high_thresh > 100)
> +		cache->bitmap_low_thresh = cache->bitmap_high_thresh - 100;
> +	else
> +		cache->bitmap_low_thresh = 0;
> +}
>  
>  static struct btrfs_free_space_info *
>  search_free_space_info(struct btrfs_trans_handle *trans,
> @@ -47,8 +78,7 @@ search_free_space_info(struct btrfs_trans_handle *trans,
>  }
>  
>  static int free_space_test_bit(struct btrfs_block_group_cache *block_group,
> -			       struct btrfs_path *path, u64 offset,
> -			       u64 sectorsize)
> +			       struct btrfs_path *path, u64 offset)
>  {
>  	struct extent_buffer *leaf;
>  	struct btrfs_key key;
> @@ -64,10 +94,1085 @@ static int free_space_test_bit(struct btrfs_block_group_cache *block_group,
>  	ASSERT(offset >= found_start && offset < found_end);
>  
>  	ptr = btrfs_item_ptr_offset(leaf, path->slots[0]);
> -	i = (offset - found_start) / sectorsize;
> +	i = (offset - found_start) / leaf->fs_info->sectorsize;
>  	return !!extent_buffer_test_bit(leaf, ptr, i);
>  }
>  
> +/*
> + * btrfs_search_slot() but we're looking for the greatest key less than the
> + * passed key.
> + */
> +static int btrfs_search_prev_slot(struct btrfs_trans_handle *trans,
> +                                  struct btrfs_root *root,
> +                                  struct btrfs_key *key, struct btrfs_path *p,
> +                                  int ins_len, int cow)
> +{
> +	int ret;
> +
> +	ret = btrfs_search_slot(trans, root, key, p, ins_len, cow);
> +	if (ret < 0)
> +		return ret;
> +
> +	if (ret == 0) {
> +		ASSERT(0);
> +		return -EIO;
> +	}
> +
> +	if (p->slots[0] == 0) {
> +		ASSERT(0);
> +		return -EIO;
> +	}
> +	p->slots[0]--;
> +
> +	return 0;
> +}
> +
> +static int add_new_free_space_info(struct btrfs_trans_handle *trans,
> +                                   struct btrfs_block_group_cache *block_group,
> +                                   struct btrfs_path *path)
> +{
> +	struct btrfs_root *root = trans->fs_info->free_space_root;
> +	struct btrfs_free_space_info *info;
> +	struct btrfs_key key;
> +	struct extent_buffer *leaf;
> +	int ret;
> +
> +	key.objectid = block_group->key.objectid;
> +	key.type = BTRFS_FREE_SPACE_INFO_KEY;
> +	key.offset = block_group->key.offset;
> +
> +	ret = btrfs_insert_empty_item(trans, root, path, &key, sizeof(*info));
> +	if (ret)
> +		goto out;
> +
> +	leaf = path->nodes[0];
> +	info = btrfs_item_ptr(leaf, path->slots[0],
> +	                      struct btrfs_free_space_info);
> +	btrfs_set_free_space_extent_count(leaf, info, 0);
> +	btrfs_set_free_space_flags(leaf, info, 0);
> +	btrfs_mark_buffer_dirty(leaf);
> +
> +	ret = 0;
> +out:
> +	btrfs_release_path(path);
> +	return ret;
> +}
> +
> +static inline u32 free_space_bitmap_size(u64 size, u32 sectorsize)
> +{
> +	return DIV_ROUND_UP((u32)div_u64(size, sectorsize), BITS_PER_BYTE);
> +}
> +
> +static unsigned long *alloc_bitmap(u32 bitmap_size)
> +{
> +	unsigned long *ret;
> +	unsigned int nofs_flag;
> +	u32 bitmap_rounded_size = round_up(bitmap_size, sizeof(unsigned long));
> +
> +	/*
> +	 * GFP_NOFS doesn't work with kvmalloc(), but we really can't recurse
> +	 * into the filesystem as the free space bitmap can be modified in the
> +	 * critical section of a transaction commit.
> +	 *
> +	 * TODO: push the memalloc_nofs_{save,restore}() to the caller where we
> +	 * know that recursion is unsafe.
> +	 */
> +	nofs_flag = memalloc_nofs_save();
> +	ret = kvzalloc(bitmap_rounded_size, GFP_KERNEL);
> +	memalloc_nofs_restore(nofs_flag);
> +	return ret;
> +}
> +
> +static void le_bitmap_set(unsigned long *map, unsigned int start, int len)
> +{
> +	u8 *p = ((u8 *)map) + BIT_BYTE(start);
> +	const unsigned int size = start + len;
> +	int bits_to_set = BITS_PER_BYTE - (start % BITS_PER_BYTE);
> +	u8 mask_to_set = BITMAP_FIRST_BYTE_MASK(start);
> +
> +	while (len - bits_to_set >= 0) {
> +		*p |= mask_to_set;
> +		len -= bits_to_set;
> +		bits_to_set = BITS_PER_BYTE;
> +		mask_to_set = ~0;
> +		p++;
> +	}
> +	if (len) {
> +		mask_to_set &= BITMAP_LAST_BYTE_MASK(size);
> +		*p |= mask_to_set;
> +	}
> +}
> +
> +int convert_free_space_to_bitmaps(struct btrfs_trans_handle *trans,
> +				  struct btrfs_block_group_cache *block_group,
> +				  struct btrfs_path *path)
> +{
> +	struct btrfs_fs_info *fs_info = trans->fs_info;
> +	struct btrfs_root *root = fs_info->free_space_root;
> +	struct btrfs_free_space_info *info;
> +	struct btrfs_key key, found_key;
> +	struct extent_buffer *leaf;
> +	unsigned long *bitmap;
> +	char *bitmap_cursor;
> +	u64 start, end;
> +	u64 bitmap_range, i;
> +	u32 bitmap_size, flags, expected_extent_count;
> +	u32 extent_count = 0;
> +	int done = 0, nr;
> +	int ret;
> +
> +	bitmap_size = free_space_bitmap_size(block_group->key.offset,
> +					     fs_info->sectorsize);
> +	bitmap = alloc_bitmap(bitmap_size);
> +	if (!bitmap) {
> +		ret = -ENOMEM;
> +		goto out;
> +	}
> +
> +	start = block_group->key.objectid;
> +	end = block_group->key.objectid + block_group->key.offset;
> +
> +	key.objectid = end - 1;
> +	key.type = (u8)-1;
> +	key.offset = (u64)-1;
> +
> +	while (!done) {
> +		ret = btrfs_search_prev_slot(trans, root, &key, path, -1, 1);
> +		if (ret)
> +			goto out;
> +
> +		leaf = path->nodes[0];
> +		nr = 0;
> +		path->slots[0]++;
> +		while (path->slots[0] > 0) {
> +			btrfs_item_key_to_cpu(leaf, &found_key, path->slots[0] - 1);
> +
> +			if (found_key.type == BTRFS_FREE_SPACE_INFO_KEY) {
> +				ASSERT(found_key.objectid == block_group->key.objectid);
> +				ASSERT(found_key.offset == block_group->key.offset);
> +				done = 1;
> +				break;
> +			} else if (found_key.type == BTRFS_FREE_SPACE_EXTENT_KEY) {
> +				u64 first, last;
> +
> +				ASSERT(found_key.objectid >= start);
> +				ASSERT(found_key.objectid < end);
> +				ASSERT(found_key.objectid + found_key.offset <= end);
> +
> +				first = div_u64(found_key.objectid - start,
> +						fs_info->sectorsize);
> +				last = div_u64(found_key.objectid + found_key.offset - start,
> +					       fs_info->sectorsize);
> +				le_bitmap_set(bitmap, first, last - first);
> +
> +				extent_count++;
> +				nr++;
> +				path->slots[0]--;
> +			} else {
> +				ASSERT(0);
> +			}
> +		}
> +
> +		ret = btrfs_del_items(trans, root, path, path->slots[0], nr);
> +		if (ret)
> +			goto out;
> +		btrfs_release_path(path);
> +	}
> +
> +	info = search_free_space_info(trans, fs_info, block_group, path, 1);
> +	if (IS_ERR(info)) {
> +		ret = PTR_ERR(info);
> +		goto out;
> +	}
> +	leaf = path->nodes[0];
> +	flags = btrfs_free_space_flags(leaf, info);
> +	flags |= BTRFS_FREE_SPACE_USING_BITMAPS;
> +	btrfs_set_free_space_flags(leaf, info, flags);
> +	expected_extent_count = btrfs_free_space_extent_count(leaf, info);
> +	btrfs_mark_buffer_dirty(leaf);
> +	btrfs_release_path(path);
> +
> +	if (extent_count != expected_extent_count) {
> +		fprintf(stderr,
> +			"incorrect extent count for %llu; counted %u, expected %u",
> +			block_group->key.objectid, extent_count,
> +			expected_extent_count);
> +		ASSERT(0);
> +		ret = -EIO;
> +		goto out;
> +	}
> +
> +	bitmap_cursor = (char *)bitmap;
> +	bitmap_range = fs_info->sectorsize * BTRFS_FREE_SPACE_BITMAP_BITS;
> +	i = start;
> +	while (i < end) {
> +		unsigned long ptr;
> +		u64 extent_size;
> +		u32 data_size;
> +
> +		extent_size = min(end - i, bitmap_range);
> +		data_size = free_space_bitmap_size(extent_size,
> +						   fs_info->sectorsize);
> +
> +		key.objectid = i;
> +		key.type = BTRFS_FREE_SPACE_BITMAP_KEY;
> +		key.offset = extent_size;
> +
> +		ret = btrfs_insert_empty_item(trans, root, path, &key,
> +					      data_size);
> +		if (ret)
> +			goto out;
> +
> +		leaf = path->nodes[0];
> +		ptr = btrfs_item_ptr_offset(leaf, path->slots[0]);
> +		write_extent_buffer(leaf, bitmap_cursor, ptr,
> +				    data_size);
> +		btrfs_mark_buffer_dirty(leaf);
> +		btrfs_release_path(path);
> +
> +		i += extent_size;
> +		bitmap_cursor += data_size;
> +	}
> +
> +	ret = 0;
> +out:
> +	kvfree(bitmap);
> +	if (ret)
> +		btrfs_abort_transaction(trans, ret);
> +	return ret;
> +}
> +
> +int convert_free_space_to_extents(struct btrfs_trans_handle *trans,
> +				  struct btrfs_block_group_cache *block_group,
> +				  struct btrfs_path *path)
> +{
> +	struct btrfs_fs_info *fs_info = trans->fs_info;
> +	struct btrfs_root *root = fs_info->free_space_root;
> +	struct btrfs_free_space_info *info;
> +	struct btrfs_key key, found_key;
> +	struct extent_buffer *leaf;
> +	unsigned long *bitmap;
> +	u64 start, end;
> +	u32 bitmap_size, flags, expected_extent_count;
> +	unsigned long nrbits, start_bit, end_bit;
> +	u32 extent_count = 0;
> +	int done = 0, nr;
> +	int ret;
> +
> +	bitmap_size = free_space_bitmap_size(block_group->key.offset,
> +					     fs_info->sectorsize);
> +	bitmap = alloc_bitmap(bitmap_size);
> +	if (!bitmap) {
> +		ret = -ENOMEM;
> +		goto out;
> +	}
> +
> +	start = block_group->key.objectid;
> +	end = block_group->key.objectid + block_group->key.offset;
> +
> +	key.objectid = end - 1;
> +	key.type = (u8)-1;
> +	key.offset = (u64)-1;
> +
> +	while (!done) {
> +		ret = btrfs_search_prev_slot(trans, root, &key, path, -1, 1);
> +		if (ret)
> +			goto out;
> +
> +		leaf = path->nodes[0];
> +		nr = 0;
> +		path->slots[0]++;
> +		while (path->slots[0] > 0) {
> +			btrfs_item_key_to_cpu(leaf, &found_key, path->slots[0] - 1);
> +
> +			if (found_key.type == BTRFS_FREE_SPACE_INFO_KEY) {
> +				ASSERT(found_key.objectid == block_group->key.objectid);
> +				ASSERT(found_key.offset == block_group->key.offset);
> +				done = 1;
> +				break;
> +			} else if (found_key.type == BTRFS_FREE_SPACE_BITMAP_KEY) {
> +				unsigned long ptr;
> +				char *bitmap_cursor;
> +				u32 bitmap_pos, data_size;
> +
> +				ASSERT(found_key.objectid >= start);
> +				ASSERT(found_key.objectid < end);
> +				ASSERT(found_key.objectid + found_key.offset <= end);
> +
> +				bitmap_pos = div_u64(found_key.objectid - start,
> +						     fs_info->sectorsize *
> +						     BITS_PER_BYTE);
> +				bitmap_cursor = ((char *)bitmap) + bitmap_pos;
> +				data_size = free_space_bitmap_size(found_key.offset,
> +								   fs_info->sectorsize);
> +
> +				ptr = btrfs_item_ptr_offset(leaf, path->slots[0] - 1);
> +				read_extent_buffer(leaf, bitmap_cursor, ptr,
> +						   data_size);
> +
> +				nr++;
> +				path->slots[0]--;
> +			} else {
> +				ASSERT(0);
> +			}
> +		}
> +
> +		ret = btrfs_del_items(trans, root, path, path->slots[0], nr);
> +		if (ret)
> +			goto out;
> +		btrfs_release_path(path);
> +	}
> +
> +	info = search_free_space_info(trans, fs_info, block_group, path, 1);
> +	if (IS_ERR(info)) {
> +		ret = PTR_ERR(info);
> +		goto out;
> +	}
> +	leaf = path->nodes[0];
> +	flags = btrfs_free_space_flags(leaf, info);
> +	flags &= ~BTRFS_FREE_SPACE_USING_BITMAPS;
> +	btrfs_set_free_space_flags(leaf, info, flags);
> +	expected_extent_count = btrfs_free_space_extent_count(leaf, info);
> +	btrfs_mark_buffer_dirty(leaf);
> +	btrfs_release_path(path);
> +
> +	nrbits = div_u64(block_group->key.offset, fs_info->sectorsize);
> +	start_bit = find_next_bit_le(bitmap, nrbits, 0);
> +
> +	while (start_bit < nrbits) {
> +		end_bit = find_next_zero_bit_le(bitmap, nrbits, start_bit);
> +		ASSERT(start_bit < end_bit);
> +
> +		key.objectid = start + start_bit * fs_info->sectorsize;
> +		key.type = BTRFS_FREE_SPACE_EXTENT_KEY;
> +		key.offset = (end_bit - start_bit) * fs_info->sectorsize;
> +
> +		ret = btrfs_insert_empty_item(trans, root, path, &key, 0);
> +		if (ret)
> +			goto out;
> +		btrfs_release_path(path);
> +
> +		extent_count++;
> +
> +		start_bit = find_next_bit_le(bitmap, nrbits, end_bit);
> +	}
> +
> +	if (extent_count != expected_extent_count) {
> +		fprintf(stderr,
> +			"incorrect extent count for %llu; counted %u, expected %u",
> +			block_group->key.objectid, extent_count,
> +			expected_extent_count);
> +		ASSERT(0);
> +		ret = -EIO;
> +		goto out;
> +	}
> +
> +	ret = 0;
> +out:
> +	kvfree(bitmap);
> +	if (ret)
> +		btrfs_abort_transaction(trans, ret);
> +	return ret;
> +}
> +
> +static int update_free_space_extent_count(struct btrfs_trans_handle *trans,
> +                                          struct btrfs_block_group_cache *block_group,
> +                                          struct btrfs_path *path,
> +                                          int new_extents)
> +{
> +	struct btrfs_free_space_info *info;
> +	u32 flags;
> +	u32 extent_count;
> +	int ret = 0;
> +
> +	if (new_extents == 0)
> +		return 0;
> +
> +	info = search_free_space_info(trans, trans->fs_info, block_group, path,
> +				1);
> +	if (IS_ERR(info)) {
> +		ret = PTR_ERR(info);
> +		goto out;
> +	}
> +	flags = btrfs_free_space_flags(path->nodes[0], info);
> +	extent_count = btrfs_free_space_extent_count(path->nodes[0], info);
> +
> +	extent_count += new_extents;
> +	btrfs_set_free_space_extent_count(path->nodes[0], info, extent_count);
> +	btrfs_mark_buffer_dirty(path->nodes[0]);
> +	btrfs_release_path(path);
> +
> +	if (!(flags & BTRFS_FREE_SPACE_USING_BITMAPS) &&
> +	    extent_count > block_group->bitmap_high_thresh) {
> +		ret = convert_free_space_to_bitmaps(trans, block_group, path);
> +	} else if ((flags & BTRFS_FREE_SPACE_USING_BITMAPS) &&
> +		   extent_count < block_group->bitmap_low_thresh) {
> +		ret = convert_free_space_to_extents(trans, block_group, path);
> +	}
> +
> +
> +out:
> +	return ret;
> +}
> +
> +
> +static void free_space_set_bits(struct btrfs_block_group_cache *block_group,
> +                                struct btrfs_path *path, u64 *start, u64 *size,
> +                                int bit)
> +{
> +        struct extent_buffer *leaf = path->nodes[0];
> +        struct btrfs_fs_info *fs_info = leaf->fs_info;
> +        struct btrfs_key key;
> +        u64 end = *start + *size;
> +        u64 found_start, found_end;
> +        unsigned long ptr, first, last;
> +
> +        btrfs_item_key_to_cpu(leaf, &key, path->slots[0]);
> +        ASSERT(key.type == BTRFS_FREE_SPACE_BITMAP_KEY);
> +
> +        found_start = key.objectid;
> +        found_end = key.objectid + key.offset;
> +        ASSERT(*start >= found_start && *start < found_end);
> +        ASSERT(end > found_start);
> +
> +        if (end > found_end)
> +                end = found_end;
> +
> +        ptr = btrfs_item_ptr_offset(leaf, path->slots[0]);
> +        first = (*start - found_start) / fs_info->sectorsize;
> +        last = (end - found_start) / fs_info->sectorsize;
> +        if (bit)
> +                extent_buffer_bitmap_set(leaf, ptr, first, last - first);
> +        else
> +                extent_buffer_bitmap_clear(leaf, ptr, first, last - first);
> +        btrfs_mark_buffer_dirty(leaf);
> +
> +        *size -= end - *start;
> +        *start = end;
> +}
> +
> +/*
> + * We can't use btrfs_next_item() in modify_free_space_bitmap() because
> + * btrfs_next_leaf() doesn't get the path for writing. We can forgo the fancy
> + * tree walking in btrfs_next_leaf() anyways because we know exactly what we're
> + * looking for.
> + */
> +static int free_space_next_bitmap(struct btrfs_trans_handle *trans,
> +                                  struct btrfs_root *root, struct btrfs_path *p)
> +{
> +	struct btrfs_key key;
> +
> +	if (p->slots[0] + 1 < btrfs_header_nritems(p->nodes[0])) {
> +		p->slots[0]++;
> +		return 0;
> +	}
> +
> +	btrfs_item_key_to_cpu(p->nodes[0], &key, p->slots[0]);
> +	btrfs_release_path(p);
> +
> +	key.objectid += key.offset;
> +	key.type = (u8)-1;
> +	key.offset = (u64)-1;
> +
> +	return btrfs_search_prev_slot(trans, root, &key, p, 0, 1);
> +}
> +
> +/*
> + * If remove is 1, then we are removing free space, thus clearing bits in the
> + * bitmap. If remove is 0, then we are adding free space, thus setting bits in
> + * the bitmap.
> + */
> +static int modify_free_space_bitmap(struct btrfs_trans_handle *trans,
> +                                    struct btrfs_block_group_cache *block_group,
> +                                    struct btrfs_path *path,
> +                                    u64 start, u64 size, int remove)
> +{
> +        struct btrfs_root *root = trans->fs_info->free_space_root;
> +        struct btrfs_key key;
> +        u64 end = start + size;
> +        u64 cur_start, cur_size;
> +        int prev_bit, next_bit;
> +        int new_extents;
> +        int ret;
> +
> +        /*
> +         * Read the bit for the block immediately before the extent of space if
> +         * that block is within the block group.
> +         */
> +        if (start > block_group->key.objectid) {
> +                u64 prev_block = start - trans->fs_info->sectorsize;
> +
> +                key.objectid = prev_block;
> +                key.type = (u8)-1;
> +                key.offset = (u64)-1;
> +
> +                ret = btrfs_search_prev_slot(trans, root, &key, path, 0, 1);
> +                if (ret)
> +                        goto out;
> +
> +                prev_bit = free_space_test_bit(block_group, path, prev_block);
> +
> +                /* The previous block may have been in the previous bitmap. */
> +                btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]);
> +                if (start >= key.objectid + key.offset) {
> +                        ret = free_space_next_bitmap(trans, root, path);
> +                        if (ret)
> +                                goto out;
> +                }
> +        } else {
> +                key.objectid = start;
> +                key.type = (u8)-1;
> +                key.offset = (u64)-1;
> +
> +                ret = btrfs_search_prev_slot(trans, root, &key, path, 0, 1);
> +                if (ret)
> +                        goto out;
> +
> +                prev_bit = -1;
> +        }
> +
> +        /*
> +         * Iterate over all of the bitmaps overlapped by the extent of space,
> +         * clearing/setting bits as required.
> +         */
> +        cur_start = start;
> +        cur_size = size;
> +        while (1) {
> +                free_space_set_bits(block_group, path, &cur_start, &cur_size,
> +                                    !remove);
> +                if (cur_size == 0)
> +                        break;
> +                ret = free_space_next_bitmap(trans, root, path);
> +                if (ret)
> +                        goto out;
> +        }
> +
> +	/*
> +         * Read the bit for the block immediately after the extent of space if
> +         * that block is within the block group.
> +         */
> +        if (end < block_group->key.objectid + block_group->key.offset) {
> +                /* The next block may be in the next bitmap. */
> +                btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]);
> +                if (end >= key.objectid + key.offset) {
> +                        ret = free_space_next_bitmap(trans, root, path);
> +                        if (ret)
> +                                goto out;
> +                }
> +
> +                next_bit = free_space_test_bit(block_group, path, end);
> +        } else {
> +                next_bit = -1;
> +        }
> +
> +        if (remove) {
> +                new_extents = -1;
> +                if (prev_bit == 1) {
> +                        /* Leftover on the left. */
> +                        new_extents++;
> +                }
> +                if (next_bit == 1) {
> +                        /* Leftover on the right. */
> +                        new_extents++;
> +                }
> +        } else {
> +                new_extents = 1;
> +                if (prev_bit == 1) {
> +                        /* Merging with neighbor on the left. */
> +                        new_extents--;
> +                }
> +                if (next_bit == 1) {
> +                        /* Merging with neighbor on the right. */
> +                        new_extents--;
> +                }
> +        }
> +
> +        btrfs_release_path(path);
> +        ret = update_free_space_extent_count(trans, block_group, path,
> +                                             new_extents);
> +
> +out:
> +        return ret;
> +}
> +
> +static int remove_free_space_extent(struct btrfs_trans_handle *trans,
> +				    struct btrfs_block_group_cache *block_group,
> +				    struct btrfs_path *path,
> +				    u64 start, u64 size)
> +{
> +	struct btrfs_root *root = trans->fs_info->free_space_root;
> +	struct btrfs_key key;
> +	u64 found_start, found_end;
> +	u64 end = start + size;
> +	int new_extents = -1;
> +	int ret;
> +
> +	key.objectid = start;
> +	key.type = (u8)-1;
> +	key.offset = (u64)-1;
> +
> +	ret = btrfs_search_prev_slot(trans, root, &key, path, -1, 1);
> +	if (ret)
> +		goto out;
> +
> +	btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]);
> +
> +	ASSERT(key.type == BTRFS_FREE_SPACE_EXTENT_KEY);
> +
> +	found_start = key.objectid;
> +	found_end = key.objectid + key.offset;
> +	ASSERT(start >= found_start && end <= found_end);
> +
> +	/*
> +	 * Okay, now that we've found the free space extent which contains the
> +	 * free space that we are removing, there are four cases:
> +	 *
> +	 * 1. We're using the whole extent: delete the key we found and
> +	 * decrement the free space extent count.
> +	 * 2. We are using part of the extent starting at the beginning: delete
> +	 * the key we found and insert a new key representing the leftover at
> +	 * the end. There is no net change in the number of extents.
> +	 * 3. We are using part of the extent ending at the end: delete the key
> +	 * we found and insert a new key representing the leftover at the
> +	 * beginning. There is no net change in the number of extents.
> +	 * 4. We are using part of the extent in the middle: delete the key we
> +	 * found and insert two new keys representing the leftovers on each
> +	 * side. Where we used to have one extent, we now have two, so increment
> +	 * the extent count. We may need to convert the block group to bitmaps
> +	 * as a result.
> +	 */
> +
> +	/* Delete the existing key (cases 1-4). */
> +	ret = btrfs_del_item(trans, root, path);
> +	if (ret)
> +		goto out;
> +
> +	/* Add a key for leftovers at the beginning (cases 3 and 4). */
> +	if (start > found_start) {
> +		key.objectid = found_start;
> +		key.type = BTRFS_FREE_SPACE_EXTENT_KEY;
> +		key.offset = start - found_start;
> +
> +		btrfs_release_path(path);
> +		ret = btrfs_insert_empty_item(trans, root, path, &key, 0);
> +		if (ret)
> +			goto out;
> +		new_extents++;
> +	}
> +
> +	/* Add a key for leftovers at the end (cases 2 and 4). */
> +	if (end < found_end) {
> +		key.objectid = end;
> +		key.type = BTRFS_FREE_SPACE_EXTENT_KEY;
> +		key.offset = found_end - end;
> +
> +		btrfs_release_path(path);
> +		ret = btrfs_insert_empty_item(trans, root, path, &key, 0);
> +		if (ret)
> +			goto out;
> +		new_extents++;
> +	}
> +
> +	btrfs_release_path(path);
> +	ret = update_free_space_extent_count(trans, block_group, path,
> +					     new_extents);
> +
> +out:
> +	return ret;
> +}
> +
> +int __remove_from_free_space_tree(struct btrfs_trans_handle *trans,
> +                                  struct btrfs_block_group_cache *block_group,
> +                                  struct btrfs_path *path, u64 start, u64 size)
> +{
> +	struct btrfs_free_space_info *info;
> +	u32 flags;
> +
> +	info = search_free_space_info(NULL, trans->fs_info, block_group, path,
> +	                              0);
> +	if (IS_ERR(info))
> +		return PTR_ERR(info);
> +	flags = btrfs_free_space_flags(path->nodes[0], info);
> +	btrfs_release_path(path);
> +
> +	if (flags & BTRFS_FREE_SPACE_USING_BITMAPS) {
> +		return modify_free_space_bitmap(trans, block_group, path,
> +	                                        start, size, 1);
> +	} else {
> +		return remove_free_space_extent(trans, block_group, path,
> +	                                        start, size);
> +	}
> +}
> +
> +int remove_from_free_space_tree(struct btrfs_trans_handle *trans,
> +				u64 start, u64 size)
> +{
> +	struct btrfs_block_group_cache *block_group;
> +	struct btrfs_path *path;
> +	int ret;
> +
> +	if (!btrfs_fs_compat_ro(trans->fs_info, FREE_SPACE_TREE))
> +		return 0;
> +
> +	path = btrfs_alloc_path();
> +	if (!path) {
> +		ret = -ENOMEM;
> +		goto out;
> +	}
> +
> +	block_group = btrfs_lookup_block_group(trans->fs_info, start);
> +	if (!block_group) {
> +		ASSERT(0);
> +		ret = -ENOENT;
> +		goto out;
> +	}
> +
> +	ret = __remove_from_free_space_tree(trans, block_group, path, start,
> +					    size);
> +out:
> +	btrfs_free_path(path);
> +	if (ret)
> +		btrfs_abort_transaction(trans, ret);
> +	return ret;
> +}
> +
> +static int add_free_space_extent(struct btrfs_trans_handle *trans,
> +                                 struct btrfs_block_group_cache *block_group,
> +                                 struct btrfs_path *path,
> +                                 u64 start, u64 size)
> +{
> +        struct btrfs_root *root = trans->fs_info->free_space_root;
> +        struct btrfs_key key, new_key;
> +        u64 found_start, found_end;
> +        u64 end = start + size;
> +        int new_extents = 1;
> +        int ret;
> +
> +        /*
> +         * We are adding a new extent of free space, but we need to merge
> +         * extents. There are four cases here:
> +         *
> +         * 1. The new extent does not have any immediate neighbors to merge
> +         * with: add the new key and increment the free space extent count. We
> +         * may need to convert the block group to bitmaps as a result.
> +         * 2. The new extent has an immediate neighbor before it: remove the
> +         * previous key and insert a new key combining both of them. There is no
> +         * net change in the number of extents.
> +         * 3. The new extent has an immediate neighbor after it: remove the next
> +         * key and insert a new key combining both of them. There is no net
> +         * change in the number of extents.
> +         * 4. The new extent has immediate neighbors on both sides: remove both
> +         * of the keys and insert a new key combining all of them. Where we used
> +         * to have two extents, we now have one, so decrement the extent count.
> +         */
> +
> +        new_key.objectid = start;
> +        new_key.type = BTRFS_FREE_SPACE_EXTENT_KEY;
> +        new_key.offset = size;
> +
> +        /* Search for a neighbor on the left. */
> +        if (start == block_group->key.objectid)
> +                goto right;
> +        key.objectid = start - 1;
> +        key.type = (u8)-1;
> +        key.offset = (u64)-1;
> +
> +        ret = btrfs_search_prev_slot(trans, root, &key, path, -1, 1);
> +        if (ret)
> +                goto out;
> +
> +        btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]);
> +
> +        if (key.type != BTRFS_FREE_SPACE_EXTENT_KEY) {
> +                ASSERT(key.type == BTRFS_FREE_SPACE_INFO_KEY);
> +                btrfs_release_path(path);
> +                goto right;
> +        }
> +
> +        found_start = key.objectid;
> +        found_end = key.objectid + key.offset;
> +        ASSERT(found_start >= block_group->key.objectid &&
> +               found_end > block_group->key.objectid);
> +        ASSERT(found_start < start && found_end <= start);
> +
> +        /*
> +         * Delete the neighbor on the left and absorb it into the new key (cases
> +         * 2 and 4).
> +         */
> +        if (found_end == start) {
> +                ret = btrfs_del_item(trans, root, path);
> +                if (ret)
> +                        goto out;
> +                new_key.objectid = found_start;
> +                new_key.offset += key.offset;
> +                new_extents--;
> +        }
> +        btrfs_release_path(path);
> +right:
> +        /* Search for a neighbor on the right. */
> +        if (end == block_group->key.objectid + block_group->key.offset)
> +                goto insert;
> +        key.objectid = end;
> +        key.type = (u8)-1;
> +        key.offset = (u64)-1;
> +
> +        ret = btrfs_search_prev_slot(trans, root, &key, path, -1, 1);
> +        if (ret)
> +                goto out;
> +
> +        btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]);
> +
> +        if (key.type != BTRFS_FREE_SPACE_EXTENT_KEY) {
> +                ASSERT(key.type == BTRFS_FREE_SPACE_INFO_KEY);
> +                btrfs_release_path(path);
> +                goto insert;
> +        }
> +
> +        found_start = key.objectid;
> +        found_end = key.objectid + key.offset;
> +        ASSERT(found_start >= block_group->key.objectid &&
> +               found_end > block_group->key.objectid);
> +        ASSERT((found_start < start && found_end <= start) ||
> +               (found_start >= end && found_end > end));
> +
> +        /*
> +         * Delete the neighbor on the right and absorb it into the new key
> +         * (cases 3 and 4).
> +         */
> +        if (found_start == end) {
> +                ret = btrfs_del_item(trans, root, path);
> +                if (ret)
> +                        goto out;
> +                new_key.offset += key.offset;
> +                new_extents--;
> +        }
> +        btrfs_release_path(path);
> +
> +insert:
> +        /* Insert the new key (cases 1-4). */
> +        ret = btrfs_insert_empty_item(trans, root, path, &new_key, 0);
> +        if (ret)
> +                goto out;
> +
> +        btrfs_release_path(path);
> +        ret = update_free_space_extent_count(trans, block_group, path,
> +                                             new_extents);
> +
> +out:
> +        return ret;
> +}
> +
> +int __add_to_free_space_tree(struct btrfs_trans_handle *trans,
> +                             struct btrfs_block_group_cache *block_group,
> +                             struct btrfs_path *path, u64 start, u64 size)
> +{
> +	struct btrfs_fs_info *fs_info = trans->fs_info;
> +	struct btrfs_free_space_info *info;
> +	u32 flags;
> +
> +	info = search_free_space_info(NULL, fs_info, block_group, path, 0);
> +	if (IS_ERR(info))
> +	        return PTR_ERR(info);
> +	flags = btrfs_free_space_flags(path->nodes[0], info);
> +	btrfs_release_path(path);
> +
> +	if (flags & BTRFS_FREE_SPACE_USING_BITMAPS) {
> +	        return modify_free_space_bitmap(trans, block_group, path,
> +	                                        start, size, 0);
> +	} else {
> +	        return add_free_space_extent(trans, block_group, path, start,
> +	                                     size);
> +	}
> +}
> +
> +
> +int add_to_free_space_tree(struct btrfs_trans_handle *trans,
> +			   u64 start, u64 size)
> +{
> +	struct btrfs_block_group_cache *block_group;
> +	struct btrfs_path *path;
> +	int ret;
> +
> +	if (!btrfs_fs_compat_ro(trans->fs_info, FREE_SPACE_TREE))
> +		return 0;
> +
> +	path = btrfs_alloc_path();
> +	if (!path) {
> +		ret = -ENOMEM;
> +		goto out;
> +	}
> +
> +	block_group = btrfs_lookup_block_group(trans->fs_info, start);
> +	if (!block_group) {
> +		ASSERT(0);
> +		ret = -ENOENT;
> +		goto out;
> +	}
> +
> +	ret = __add_to_free_space_tree(trans, block_group, path, start, size);
> +out:
> +	btrfs_free_path(path);
> +	if (ret)
> +		btrfs_abort_transaction(trans, ret);
> +	return ret;
> +}
> +
> +int populate_free_space_tree(struct btrfs_trans_handle *trans,
> +			     struct btrfs_block_group_cache *block_group)
> +{
> +        struct btrfs_root *extent_root = trans->fs_info->extent_root;
> +        struct btrfs_path *path, *path2;
> +        struct btrfs_key key;
> +        u64 start, end;
> +        int ret;
> +
> +        path = btrfs_alloc_path();
> +        if (!path)
> +                return -ENOMEM;
> +        path->reada = READA_FORWARD;
> +
> +        path2 = btrfs_alloc_path();
> +        if (!path2) {
> +                btrfs_free_path(path);
> +                return -ENOMEM;
> +        }
> +
> +        ret = add_new_free_space_info(trans, block_group, path2);
> +        if (ret)
> +                goto out;
> +
> +        /*
> +         * Iterate through all of the extent and metadata items in this block
> +         * group, adding the free space between them and the free space at the
> +         * end. Note that EXTENT_ITEM and METADATA_ITEM are less than
> +         * BLOCK_GROUP_ITEM, so an extent may precede the block group that it's
> +         * contained in.
> +         */
> +        key.objectid = block_group->key.objectid;
> +        key.type = BTRFS_EXTENT_ITEM_KEY;
> +        key.offset = 0;
> +
> +        ret = btrfs_search_slot_for_read(extent_root, &key, path, 1, 0);
> +        if (ret < 0)
> +                goto out;
> +        ASSERT(ret == 0);
> +
> +        start = block_group->key.objectid;
> +        end = block_group->key.objectid + block_group->key.offset;
> +        while (1) {
> +                btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]);
> +
> +                if (key.type == BTRFS_EXTENT_ITEM_KEY ||
> +                    key.type == BTRFS_METADATA_ITEM_KEY) {
> +                        if (key.objectid >= end)
> +                                break;
> +
> +                        if (start < key.objectid) {
> +                                ret = __add_to_free_space_tree(trans,
> +                                                               block_group,
> +                                                               path2, start,
> +                                                               key.objectid -
> +                                                               start);
> +                                if (ret)
> +                                        goto out;
> +                        }
> +                        start = key.objectid;
> +                        if (key.type == BTRFS_METADATA_ITEM_KEY)
> +                                start += trans->fs_info->nodesize;
> +                        else
> +                                start += key.offset;
> +                } else if (key.type == BTRFS_BLOCK_GROUP_ITEM_KEY) {
> +                        if (key.objectid != block_group->key.objectid)
> +                                break;
> +                }
> +
> +                ret = btrfs_next_item(extent_root, path);
> +                if (ret < 0)
> +                        goto out;
> +                if (ret)
> +                        break;
> +        }
> +        if (start < end) {
> +                ret = __add_to_free_space_tree(trans, block_group, path2,
> +                                               start, end - start);
> +                if (ret)
> +                        goto out;
> +        }
> +
> +        ret = 0;
> +out:
> +        btrfs_free_path(path2);
> +        btrfs_free_path(path);
> +        return ret;
> +}
> +
> +int remove_block_group_free_space(struct btrfs_trans_handle *trans,
> +				  struct btrfs_block_group_cache *block_group)
> +{
> +	struct btrfs_root *root = trans->fs_info->free_space_root;
> +	struct btrfs_path *path;
> +	struct btrfs_key key, found_key;
> +	struct extent_buffer *leaf;
> +	u64 start, end;
> +	int done = 0, nr;
> +	int ret;
> +
> +	path = btrfs_alloc_path();
> +	if (!path) {
> +		ret = -ENOMEM;
> +		goto out;
> +	}
> +
> +	start = block_group->key.objectid;
> +	end = block_group->key.objectid + block_group->key.offset;
> +
> +	key.objectid = end - 1;
> +	key.type = (u8)-1;
> +	key.offset = (u64)-1;
> +
> +	while (!done) {
> +		ret = btrfs_search_prev_slot(trans, root, &key, path, -1, 1);
> +		if (ret)
> +			goto out;
> +
> +		leaf = path->nodes[0];
> +		nr = 0;
> +		path->slots[0]++;
> +		while (path->slots[0] > 0) {
> +			btrfs_item_key_to_cpu(leaf, &found_key, path->slots[0] - 1);
> +
> +			if (found_key.type == BTRFS_FREE_SPACE_INFO_KEY) {
> +				ASSERT(found_key.objectid == block_group->key.objectid);
> +				ASSERT(found_key.offset == block_group->key.offset);
> +				done = 1;
> +				nr++;
> +				path->slots[0]--;
> +				break;
> +			} else if (found_key.type == BTRFS_FREE_SPACE_EXTENT_KEY ||
> +				   found_key.type == BTRFS_FREE_SPACE_BITMAP_KEY) {
> +				ASSERT(found_key.objectid >= start);
> +				ASSERT(found_key.objectid < end);
> +				ASSERT(found_key.objectid + found_key.offset <= end);
> +				nr++;
> +				path->slots[0]--;
> +			} else {
> +				ASSERT(0);
> +			}
> +		}
> +
> +		ret = btrfs_del_items(trans, root, path, path->slots[0], nr);
> +		if (ret)
> +			goto out;
> +		btrfs_release_path(path);
> +	}
> +
> +	ret = 0;
> +out:
> +	btrfs_free_path(path);
> +	if (ret)
> +		btrfs_abort_transaction(trans, ret);
> +	return ret;
> +}
>  static int clear_free_space_tree(struct btrfs_trans_handle *trans,
>  				 struct btrfs_root *root)
>  {
> @@ -204,8 +1309,8 @@ static int load_free_space_bitmaps(struct btrfs_fs_info *fs_info,
>  
>  		offset = key.objectid;
>  		while (offset < key.objectid + key.offset) {
> -			bit = free_space_test_bit(block_group, path, offset,
> -						  fs_info->sectorsize);
> +			bit = free_space_test_bit(block_group, path, offset);
> +
>  			if (prev_bit == 0 && bit == 1) {
>  				extent_start = offset;
>  			} else if (prev_bit == 1 && bit == 0) {
> @@ -320,6 +1425,142 @@ static int load_free_space_extents(struct btrfs_fs_info *fs_info,
>  	return ret;
>  }
>  
> +struct btrfs_root *btrfs_create_tree(struct btrfs_trans_handle *trans,
> +                                     struct btrfs_fs_info *fs_info,
> +                                     u64 objectid)
> +{
> +	struct extent_buffer *leaf;
> +	struct btrfs_root *tree_root = fs_info->tree_root;
> +	struct btrfs_root *root;
> +	struct btrfs_key key;
> +	int ret = 0;
> +
> +	root = kzalloc(sizeof(*root), GFP_KERNEL);
> +	if (!root)
> +		return ERR_PTR(-ENOMEM);
> +
> +	btrfs_setup_root(root, fs_info, objectid);
> +	root->root_key.objectid = objectid;
> +	root->root_key.type = BTRFS_ROOT_ITEM_KEY;
> +	root->root_key.offset = 0;
> +
> +	leaf = btrfs_alloc_free_block(trans, root, fs_info->nodesize, objectid, NULL, 0, 0, 0);
> +	if (IS_ERR(leaf)) {
> +		ret = PTR_ERR(leaf);
> +		leaf = NULL;
> +		goto fail;
> +	}
> +
> +	memset_extent_buffer(leaf, 0, 0, sizeof(struct btrfs_header));
> +	btrfs_set_header_bytenr(leaf, leaf->start);
> +	btrfs_set_header_generation(leaf, trans->transid);
> +	btrfs_set_header_backref_rev(leaf, BTRFS_MIXED_BACKREF_REV);
> +	btrfs_set_header_owner(leaf, objectid);
> +	root->node = leaf;
> +	write_extent_buffer(leaf, fs_info->fsid, btrfs_header_fsid(), BTRFS_FSID_SIZE);
> +	write_extent_buffer(leaf, fs_info->chunk_tree_uuid,
> +			    btrfs_header_chunk_tree_uuid(leaf),
> +			    BTRFS_UUID_SIZE);
> +	btrfs_mark_buffer_dirty(leaf);
> +
> +	extent_buffer_get(root->node);
> +	root->commit_root = root->node;
> +	root->track_dirty = 1;
> +
> +	root->root_item.flags = 0;
> +	root->root_item.byte_limit = 0;
> +	btrfs_set_root_bytenr(&root->root_item, leaf->start);
> +	btrfs_set_root_generation(&root->root_item, trans->transid);
> +	btrfs_set_root_level(&root->root_item, 0);
> +	btrfs_set_root_refs(&root->root_item, 1);
> +	btrfs_set_root_used(&root->root_item, leaf->len);
> +	btrfs_set_root_last_snapshot(&root->root_item, 0);
> +	btrfs_set_root_dirid(&root->root_item, 0);
> +	memset(root->root_item.uuid, 0, BTRFS_UUID_SIZE);
> +	root->root_item.drop_level = 0;
> +
> +	key.objectid = objectid;
> +	key.type = BTRFS_ROOT_ITEM_KEY;
> +	key.offset = 0;
> +	ret = btrfs_insert_root(trans, tree_root, &key, &root->root_item);
> +	if (ret)
> +		goto fail;
> +
> +	return root;
> +
> +fail:
> +	if (leaf)
> +		free_extent_buffer(leaf);
> +
> +	kfree(root);
> +	return ERR_PTR(ret);
> +}
> +
> +#define btrfs_set_fs_compat_ro(__fs_info, opt) \
> +        __btrfs_set_fs_compat_ro((__fs_info), BTRFS_FEATURE_COMPAT_RO_##opt)
> +
> +static inline void __btrfs_set_fs_compat_ro(struct btrfs_fs_info *fs_info,
> +                                            u64 flag)
> +{
> +	struct btrfs_super_block *disk_super;
> +	u64 features;
> +
> +	disk_super = fs_info->super_copy;
> +	features = btrfs_super_compat_ro_flags(disk_super);
> +	if (!(features & flag)) {
> +		features = btrfs_super_compat_ro_flags(disk_super);
> +		if (!(features & flag)) {
> +			features |= flag;
> +			btrfs_set_super_compat_ro_flags(disk_super, features);
> +		}
> +	}
> +}
> +
> +int btrfs_create_free_space_tree(struct btrfs_fs_info *fs_info)
> +{
> +	struct btrfs_trans_handle *trans;
> +	struct btrfs_root *tree_root = fs_info->tree_root;
> +	struct btrfs_root *free_space_root;
> +	struct btrfs_block_group_cache *block_group;
> +	u64 start = BTRFS_SUPER_INFO_OFFSET + BTRFS_SUPER_INFO_SIZE;
> +	int ret;
> +
> +	trans = btrfs_start_transaction(tree_root, 0);
> +	if (IS_ERR(trans))
> +		return PTR_ERR(trans);
> +
> +	free_space_root = btrfs_create_tree(trans, fs_info,
> +					    BTRFS_FREE_SPACE_TREE_OBJECTID);
> +	if (IS_ERR(free_space_root)) {
> +		ret = PTR_ERR(free_space_root);
> +		goto abort;
> +	}
> +	fs_info->free_space_root = free_space_root;
> +
> +	do {
> +		block_group = btrfs_lookup_first_block_group(fs_info, start);
> +		if (!block_group)
> +			break;
> +		start = block_group->key.objectid + block_group->key.offset;
> +		ret = populate_free_space_tree(trans, block_group);
> +		if (ret)
> +			goto abort;
> +	} while (block_group);
> +
> +	btrfs_set_fs_compat_ro(fs_info, FREE_SPACE_TREE);
> +	btrfs_set_fs_compat_ro(fs_info, FREE_SPACE_TREE_VALID);
> +
> +	ret = btrfs_commit_transaction(trans, tree_root);
> +	if (ret)
> +		return ret;
> +
> +	return 0;
> +
> +abort:
> +	btrfs_abort_transaction(trans, ret);
> +	return ret;
> +}
> +
>  int load_free_space_tree(struct btrfs_fs_info *fs_info,
>  			 struct btrfs_block_group_cache *block_group)
>  {
> @@ -332,7 +1573,7 @@ int load_free_space_tree(struct btrfs_fs_info *fs_info,
>  	path = btrfs_alloc_path();
>  	if (!path)
>  		return -ENOMEM;
> -	path->reada = 1;
> +	path->reada = READA_BACK;
>  
>  	info = search_free_space_info(NULL, fs_info, block_group, path, 0);
>  	if (IS_ERR(info)) {
> diff --git a/free-space-tree.h b/free-space-tree.h
> index 4845f13e6808..9530c2882358 100644
> --- a/free-space-tree.h
> +++ b/free-space-tree.h
> @@ -19,8 +19,19 @@
>  #ifndef __BTRFS_FREE_SPACE_TREE_H__
>  #define __BTRFS_FREE_SPACE_TREE_H__
>  
> +#define BTRFS_FREE_SPACE_BITMAP_SIZE 256
> +#define BTRFS_FREE_SPACE_BITMAP_BITS (BTRFS_FREE_SPACE_BITMAP_SIZE * BITS_PER_BYTE)
> +
>  int btrfs_clear_free_space_tree(struct btrfs_fs_info *fs_info);
>  int load_free_space_tree(struct btrfs_fs_info *fs_info,
>  			 struct btrfs_block_group_cache *block_group);
> -
> +int populate_free_space_tree(struct btrfs_trans_handle *trans,
> +			     struct btrfs_block_group_cache *block_group);
> +int remove_block_group_free_space(struct btrfs_trans_handle *trans,
> +				  struct btrfs_block_group_cache *block_group);
> +int add_to_free_space_tree(struct btrfs_trans_handle *trans, u64 start,
> +			   u64 size);
> +int remove_from_free_space_tree(struct btrfs_trans_handle *trans, u64 start,
> +				u64 size);
> +int btrfs_create_free_space_tree(struct btrfs_fs_info *info);
>  #endif
> diff --git a/kerncompat.h b/kerncompat.h
> index 1a2bc18c3ac2..a223a7f009bd 100644
> --- a/kerncompat.h
> +++ b/kerncompat.h
> @@ -263,6 +263,8 @@ static inline int IS_ERR_OR_NULL(const void *ptr)
>  	return !ptr || IS_ERR(ptr);
>  }
>  
> +#define div_u64(x, y) ((x) / (y))
> +
>  /**
>   * swap - swap values of @a and @b
>   * @a: first value
> @@ -297,6 +299,10 @@ static inline int IS_ERR_OR_NULL(const void *ptr)
>  #define kfree(x) free(x)
>  #define vmalloc(x) malloc(x)
>  #define vfree(x) free(x)
> +#define kvzalloc(x, y) kzalloc(x,y)
> +#define kvfree(x) free(x)
> +#define memalloc_nofs_save() (0)
> +#define memalloc_nofs_restore(x)
>  
>  #ifndef BTRFS_DISABLE_BACKTRACE
>  static inline void assert_trace(const char *assertion, const char *filename,
> -- 
> 2.7.4
> 

^ permalink raw reply	[flat|nested] 27+ messages in thread

* Re: [PATCH 07/10] btrfs-progs: Add freespace tree as compat_ro supported feature
  2018-10-01 14:46 ` [PATCH 07/10] btrfs-progs: Add freespace tree as compat_ro supported feature Nikolay Borisov
@ 2018-10-04 18:30   ` Omar Sandoval
  2018-10-04 18:36     ` Nikolay Borisov
  0 siblings, 1 reply; 27+ messages in thread
From: Omar Sandoval @ 2018-10-04 18:30 UTC (permalink / raw)
  To: Nikolay Borisov; +Cc: linux-btrfs

On Mon, Oct 01, 2018 at 05:46:18PM +0300, Nikolay Borisov wrote:
> The RO_FREE_SPACE_TREE(_VALID) flags are required in order to be able
> to open an FST filesystem in repair mode. Add them to
> BTRFS_FEATURE_COMPAT_RO_SUPP.
> 
> Signed-off-by: Nikolay Borisov <nborisov@suse.com>
> ---
>  ctree.h | 4 +++-
>  1 file changed, 3 insertions(+), 1 deletion(-)
> 
> diff --git a/ctree.h b/ctree.h
> index a6d6c3decd87..3c396e7d293d 100644
> --- a/ctree.h
> +++ b/ctree.h
> @@ -497,7 +497,9 @@ struct btrfs_super_block {
>   * added here until read-write support for the free space tree is implemented in
>   * btrfs-progs.
>   */

This comment should go away.

> -#define BTRFS_FEATURE_COMPAT_RO_SUPP		0ULL
> +#define BTRFS_FEATURE_COMPAT_RO_SUPP			\
> +	(BTRFS_FEATURE_COMPAT_RO_FREE_SPACE_TREE |	\
> +	 BTRFS_FEATURE_COMPAT_RO_FREE_SPACE_TREE_VALID)
>  
>  #define BTRFS_FEATURE_INCOMPAT_SUPP			\
>  	(BTRFS_FEATURE_INCOMPAT_MIXED_BACKREF |		\

To repeat my question from before, did you test whether we can properly
change the filesystem with, e.g., btrfstune or btrfs fi label? Given
that some critical code was missing in the free space tree code, I'd be
surprised if it worked correctly.

^ permalink raw reply	[flat|nested] 27+ messages in thread

* Re: [PATCH 05/10] btrfs-progs: Pull free space tree related code from kernel
  2018-10-04 18:26   ` Omar Sandoval
@ 2018-10-04 18:34     ` Nikolay Borisov
  2018-10-04 19:01       ` Omar Sandoval
  2018-10-23 14:05     ` David Sterba
  1 sibling, 1 reply; 27+ messages in thread
From: Nikolay Borisov @ 2018-10-04 18:34 UTC (permalink / raw)
  To: Omar Sandoval; +Cc: linux-btrfs



On  4.10.2018 21:26, Omar Sandoval wrote:
> On Mon, Oct 01, 2018 at 05:46:16PM +0300, Nikolay Borisov wrote:
>> To help implement free space tree checker in user space some kernel
>> function are necessary, namely iterating/deleting/adding freespace
>> items, some internal search functions. Functions to populate a block
>> group based on the extent tree. The code is largely copy/paste from
>> the kernel with locking eliminated (i.e free_space_lock). It supports
>> reading/writing of both bitmap and extent based FST trees.
> 
> For some reason, a lot of this added code uses spaces instead of tabs,
> so I had to fix that in order to compare it to the kernel code (some of
> the functions were reordered, too).
> 
> The only functional difference I noticed was that this is missing the
> code to insert the block group header in the free space tree:
> 
> 	if (block_group->needs_free_space) {
> 		ret = __add_block_group_free_space(trans, block_group, path);
> 		if (ret)
> 			return ret;
> 	}
> 
> Was that intentionally omitted? Without it, the free space tree is
> pretty broken :(

Yes, it was intentional. If you remember I even emailed you about this
particular piece of code and you said you needed to have it this way in
case delayed refs were run before the space tree was initialized.

In user space AFAIK we don't have that problem since
add_new_free_space_info/__add_to_free_space_tree is called via

btrfs_Create_free_space_tree
   populate_free_space_tree
     add_new_free_space_info
     __add_to_free_space_tree



> 
>> Signed-off-by: Nikolay Borisov <nborisov@suse.com>
>> ---
>>  ctree.c           |   77 ++++
>>  ctree.h           |   15 +
>>  free-space-tree.c | 1253 ++++++++++++++++++++++++++++++++++++++++++++++++++++-
>>  free-space-tree.h |   13 +-
>>  kerncompat.h      |    6 +
>>  5 files changed, 1357 insertions(+), 7 deletions(-)
>>
>> diff --git a/ctree.c b/ctree.c
>> index d8a6883aa85f..aa1568620205 100644
>> --- a/ctree.c
>> +++ b/ctree.c
>> @@ -1226,6 +1226,83 @@ int btrfs_search_slot(struct btrfs_trans_handle *trans, struct btrfs_root
>>  }
>>  
>>  /*
>> + * helper to use instead of search slot if no exact match is needed but
>> + * instead the next or previous item should be returned.
>> + * When find_higher is true, the next higher item is returned, the next lower
>> + * otherwise.
>> + * When return_any and find_higher are both true, and no higher item is found,
>> + * return the next lower instead.
>> + * When return_any is true and find_higher is false, and no lower item is found,
>> + * return the next higher instead.
>> + * It returns 0 if any item is found, 1 if none is found (tree empty), and
>> + * < 0 on error
>> + */
>> +int btrfs_search_slot_for_read(struct btrfs_root *root,
>> +                               const struct btrfs_key *key,
>> +                               struct btrfs_path *p, int find_higher,
>> +                               int return_any)
>> +{
>> +        int ret;
>> +        struct extent_buffer *leaf;
>> +
>> +again:
>> +        ret = btrfs_search_slot(NULL, root, key, p, 0, 0);
>> +        if (ret <= 0)
>> +                return ret;
>> +        /*
>> +         * a return value of 1 means the path is at the position where the
>> +         * item should be inserted. Normally this is the next bigger item,
>> +         * but in case the previous item is the last in a leaf, path points
>> +         * to the first free slot in the previous leaf, i.e. at an invalid
>> +         * item.
>> +         */
>> +        leaf = p->nodes[0];
>> +
>> +        if (find_higher) {
>> +                if (p->slots[0] >= btrfs_header_nritems(leaf)) {
>> +                        ret = btrfs_next_leaf(root, p);
>> +                        if (ret <= 0)
>> +                                return ret;
>> +                        if (!return_any)
>> +                                return 1;
>> +                        /*
>> +                         * no higher item found, return the next
>> +                         * lower instead
>> +                         */
>> +                        return_any = 0;
>> +                        find_higher = 0;
>> +                        btrfs_release_path(p);
>> +                        goto again;
>> +                }
>> +        } else {
>> +                if (p->slots[0] == 0) {
>> +                        ret = btrfs_prev_leaf(root, p);
>> +                        if (ret < 0)
>> +                                return ret;
>> +                        if (!ret) {
>> +                                leaf = p->nodes[0];
>> +                                if (p->slots[0] == btrfs_header_nritems(leaf))
>> +                                        p->slots[0]--;
>> +                                return 0;
>> +                        }
>> +                        if (!return_any)
>> +                                return 1;
>> +                        /*
>> +                         * no lower item found, return the next
>> +                         * higher instead
>> +                         */
>> +                        return_any = 0;
>> +                        find_higher = 1;
>> +                        btrfs_release_path(p);
>> +                        goto again;
>> +                } else {
>> +                        --p->slots[0];
>> +                }
>> +        }
>> +        return 0;
>> +}
>> +
>> +/*
>>   * adjust the pointers going up the tree, starting at level
>>   * making sure the right key of each node is points to 'key'.
>>   * This is used after shifting pointers to the left, so it stops
>> diff --git a/ctree.h b/ctree.h
>> index 49f0f5181512..a6d6c3decd87 100644
>> --- a/ctree.h
>> +++ b/ctree.h
>> @@ -1071,6 +1071,17 @@ struct btrfs_block_group_cache {
>>  	u64 flags;
>>  	int cached;
>>  	int ro;
>> +	/*
>> +         * If the free space extent count exceeds this number, convert the block
>> +         * group to bitmaps.
>> +         */
>> +        u32 bitmap_high_thresh;
>> +        /*
>> +         * If the free space extent count drops below this number, convert the
>> +         * block group back to extents.
>> +         */
>> +        u32 bitmap_low_thresh;
>> +
>>  };
>>  
>>  struct btrfs_device;
>> @@ -2596,6 +2607,10 @@ int btrfs_split_item(struct btrfs_trans_handle *trans,
>>  int btrfs_search_slot(struct btrfs_trans_handle *trans, struct btrfs_root
>>  		      *root, struct btrfs_key *key, struct btrfs_path *p, int
>>  		      ins_len, int cow);
>> +int btrfs_search_slot_for_read(struct btrfs_root *root,
>> +                               const struct btrfs_key *key,
>> +                               struct btrfs_path *p, int find_higher,
>> +                               int return_any);
>>  int btrfs_find_item(struct btrfs_root *fs_root, struct btrfs_path *found_path,
>>  		u64 iobjectid, u64 ioff, u8 key_type,
>>  		struct btrfs_key *found_key);
>> diff --git a/free-space-tree.c b/free-space-tree.c
>> index b439b6b43146..3b7e8a3fe4f5 100644
>> --- a/free-space-tree.c
>> +++ b/free-space-tree.c
>> @@ -21,6 +21,37 @@
>>  #include "free-space-cache.h"
>>  #include "free-space-tree.h"
>>  #include "transaction.h"
>> +#include "bitops.h"
>> +#include "internal.h"
>> +
>> +void set_free_space_tree_thresholds(struct btrfs_block_group_cache *cache,
>> +				    u64 sectorsize)
>> +{
>> +	u32 bitmap_range;
>> +	size_t bitmap_size;
>> +	u64 num_bitmaps, total_bitmap_size;
>> +
>> +	/*
>> +	 * We convert to bitmaps when the disk space required for using extents
>> +	 * exceeds that required for using bitmaps.
>> +	 */
>> +	bitmap_range = sectorsize * BTRFS_FREE_SPACE_BITMAP_BITS;
>> +	num_bitmaps = div_u64(cache->key.offset + bitmap_range - 1,
>> +			      bitmap_range);
>> +	bitmap_size = sizeof(struct btrfs_item) + BTRFS_FREE_SPACE_BITMAP_SIZE;
>> +	total_bitmap_size = num_bitmaps * bitmap_size;
>> +	cache->bitmap_high_thresh = div_u64(total_bitmap_size,
>> +					    sizeof(struct btrfs_item));
>> +
>> +	/*
>> +	 * We allow for a small buffer between the high threshold and low
>> +	 * threshold to avoid thrashing back and forth between the two formats.
>> +	 */
>> +	if (cache->bitmap_high_thresh > 100)
>> +		cache->bitmap_low_thresh = cache->bitmap_high_thresh - 100;
>> +	else
>> +		cache->bitmap_low_thresh = 0;
>> +}
>>  
>>  static struct btrfs_free_space_info *
>>  search_free_space_info(struct btrfs_trans_handle *trans,
>> @@ -47,8 +78,7 @@ search_free_space_info(struct btrfs_trans_handle *trans,
>>  }
>>  
>>  static int free_space_test_bit(struct btrfs_block_group_cache *block_group,
>> -			       struct btrfs_path *path, u64 offset,
>> -			       u64 sectorsize)
>> +			       struct btrfs_path *path, u64 offset)
>>  {
>>  	struct extent_buffer *leaf;
>>  	struct btrfs_key key;
>> @@ -64,10 +94,1085 @@ static int free_space_test_bit(struct btrfs_block_group_cache *block_group,
>>  	ASSERT(offset >= found_start && offset < found_end);
>>  
>>  	ptr = btrfs_item_ptr_offset(leaf, path->slots[0]);
>> -	i = (offset - found_start) / sectorsize;
>> +	i = (offset - found_start) / leaf->fs_info->sectorsize;
>>  	return !!extent_buffer_test_bit(leaf, ptr, i);
>>  }
>>  
>> +/*
>> + * btrfs_search_slot() but we're looking for the greatest key less than the
>> + * passed key.
>> + */
>> +static int btrfs_search_prev_slot(struct btrfs_trans_handle *trans,
>> +                                  struct btrfs_root *root,
>> +                                  struct btrfs_key *key, struct btrfs_path *p,
>> +                                  int ins_len, int cow)
>> +{
>> +	int ret;
>> +
>> +	ret = btrfs_search_slot(trans, root, key, p, ins_len, cow);
>> +	if (ret < 0)
>> +		return ret;
>> +
>> +	if (ret == 0) {
>> +		ASSERT(0);
>> +		return -EIO;
>> +	}
>> +
>> +	if (p->slots[0] == 0) {
>> +		ASSERT(0);
>> +		return -EIO;
>> +	}
>> +	p->slots[0]--;
>> +
>> +	return 0;
>> +}
>> +
>> +static int add_new_free_space_info(struct btrfs_trans_handle *trans,
>> +                                   struct btrfs_block_group_cache *block_group,
>> +                                   struct btrfs_path *path)
>> +{
>> +	struct btrfs_root *root = trans->fs_info->free_space_root;
>> +	struct btrfs_free_space_info *info;
>> +	struct btrfs_key key;
>> +	struct extent_buffer *leaf;
>> +	int ret;
>> +
>> +	key.objectid = block_group->key.objectid;
>> +	key.type = BTRFS_FREE_SPACE_INFO_KEY;
>> +	key.offset = block_group->key.offset;
>> +
>> +	ret = btrfs_insert_empty_item(trans, root, path, &key, sizeof(*info));
>> +	if (ret)
>> +		goto out;
>> +
>> +	leaf = path->nodes[0];
>> +	info = btrfs_item_ptr(leaf, path->slots[0],
>> +	                      struct btrfs_free_space_info);
>> +	btrfs_set_free_space_extent_count(leaf, info, 0);
>> +	btrfs_set_free_space_flags(leaf, info, 0);
>> +	btrfs_mark_buffer_dirty(leaf);
>> +
>> +	ret = 0;
>> +out:
>> +	btrfs_release_path(path);
>> +	return ret;
>> +}
>> +
>> +static inline u32 free_space_bitmap_size(u64 size, u32 sectorsize)
>> +{
>> +	return DIV_ROUND_UP((u32)div_u64(size, sectorsize), BITS_PER_BYTE);
>> +}
>> +
>> +static unsigned long *alloc_bitmap(u32 bitmap_size)
>> +{
>> +	unsigned long *ret;
>> +	unsigned int nofs_flag;
>> +	u32 bitmap_rounded_size = round_up(bitmap_size, sizeof(unsigned long));
>> +
>> +	/*
>> +	 * GFP_NOFS doesn't work with kvmalloc(), but we really can't recurse
>> +	 * into the filesystem as the free space bitmap can be modified in the
>> +	 * critical section of a transaction commit.
>> +	 *
>> +	 * TODO: push the memalloc_nofs_{save,restore}() to the caller where we
>> +	 * know that recursion is unsafe.
>> +	 */
>> +	nofs_flag = memalloc_nofs_save();
>> +	ret = kvzalloc(bitmap_rounded_size, GFP_KERNEL);
>> +	memalloc_nofs_restore(nofs_flag);
>> +	return ret;
>> +}
>> +
>> +static void le_bitmap_set(unsigned long *map, unsigned int start, int len)
>> +{
>> +	u8 *p = ((u8 *)map) + BIT_BYTE(start);
>> +	const unsigned int size = start + len;
>> +	int bits_to_set = BITS_PER_BYTE - (start % BITS_PER_BYTE);
>> +	u8 mask_to_set = BITMAP_FIRST_BYTE_MASK(start);
>> +
>> +	while (len - bits_to_set >= 0) {
>> +		*p |= mask_to_set;
>> +		len -= bits_to_set;
>> +		bits_to_set = BITS_PER_BYTE;
>> +		mask_to_set = ~0;
>> +		p++;
>> +	}
>> +	if (len) {
>> +		mask_to_set &= BITMAP_LAST_BYTE_MASK(size);
>> +		*p |= mask_to_set;
>> +	}
>> +}
>> +
>> +int convert_free_space_to_bitmaps(struct btrfs_trans_handle *trans,
>> +				  struct btrfs_block_group_cache *block_group,
>> +				  struct btrfs_path *path)
>> +{
>> +	struct btrfs_fs_info *fs_info = trans->fs_info;
>> +	struct btrfs_root *root = fs_info->free_space_root;
>> +	struct btrfs_free_space_info *info;
>> +	struct btrfs_key key, found_key;
>> +	struct extent_buffer *leaf;
>> +	unsigned long *bitmap;
>> +	char *bitmap_cursor;
>> +	u64 start, end;
>> +	u64 bitmap_range, i;
>> +	u32 bitmap_size, flags, expected_extent_count;
>> +	u32 extent_count = 0;
>> +	int done = 0, nr;
>> +	int ret;
>> +
>> +	bitmap_size = free_space_bitmap_size(block_group->key.offset,
>> +					     fs_info->sectorsize);
>> +	bitmap = alloc_bitmap(bitmap_size);
>> +	if (!bitmap) {
>> +		ret = -ENOMEM;
>> +		goto out;
>> +	}
>> +
>> +	start = block_group->key.objectid;
>> +	end = block_group->key.objectid + block_group->key.offset;
>> +
>> +	key.objectid = end - 1;
>> +	key.type = (u8)-1;
>> +	key.offset = (u64)-1;
>> +
>> +	while (!done) {
>> +		ret = btrfs_search_prev_slot(trans, root, &key, path, -1, 1);
>> +		if (ret)
>> +			goto out;
>> +
>> +		leaf = path->nodes[0];
>> +		nr = 0;
>> +		path->slots[0]++;
>> +		while (path->slots[0] > 0) {
>> +			btrfs_item_key_to_cpu(leaf, &found_key, path->slots[0] - 1);
>> +
>> +			if (found_key.type == BTRFS_FREE_SPACE_INFO_KEY) {
>> +				ASSERT(found_key.objectid == block_group->key.objectid);
>> +				ASSERT(found_key.offset == block_group->key.offset);
>> +				done = 1;
>> +				break;
>> +			} else if (found_key.type == BTRFS_FREE_SPACE_EXTENT_KEY) {
>> +				u64 first, last;
>> +
>> +				ASSERT(found_key.objectid >= start);
>> +				ASSERT(found_key.objectid < end);
>> +				ASSERT(found_key.objectid + found_key.offset <= end);
>> +
>> +				first = div_u64(found_key.objectid - start,
>> +						fs_info->sectorsize);
>> +				last = div_u64(found_key.objectid + found_key.offset - start,
>> +					       fs_info->sectorsize);
>> +				le_bitmap_set(bitmap, first, last - first);
>> +
>> +				extent_count++;
>> +				nr++;
>> +				path->slots[0]--;
>> +			} else {
>> +				ASSERT(0);
>> +			}
>> +		}
>> +
>> +		ret = btrfs_del_items(trans, root, path, path->slots[0], nr);
>> +		if (ret)
>> +			goto out;
>> +		btrfs_release_path(path);
>> +	}
>> +
>> +	info = search_free_space_info(trans, fs_info, block_group, path, 1);
>> +	if (IS_ERR(info)) {
>> +		ret = PTR_ERR(info);
>> +		goto out;
>> +	}
>> +	leaf = path->nodes[0];
>> +	flags = btrfs_free_space_flags(leaf, info);
>> +	flags |= BTRFS_FREE_SPACE_USING_BITMAPS;
>> +	btrfs_set_free_space_flags(leaf, info, flags);
>> +	expected_extent_count = btrfs_free_space_extent_count(leaf, info);
>> +	btrfs_mark_buffer_dirty(leaf);
>> +	btrfs_release_path(path);
>> +
>> +	if (extent_count != expected_extent_count) {
>> +		fprintf(stderr,
>> +			"incorrect extent count for %llu; counted %u, expected %u",
>> +			block_group->key.objectid, extent_count,
>> +			expected_extent_count);
>> +		ASSERT(0);
>> +		ret = -EIO;
>> +		goto out;
>> +	}
>> +
>> +	bitmap_cursor = (char *)bitmap;
>> +	bitmap_range = fs_info->sectorsize * BTRFS_FREE_SPACE_BITMAP_BITS;
>> +	i = start;
>> +	while (i < end) {
>> +		unsigned long ptr;
>> +		u64 extent_size;
>> +		u32 data_size;
>> +
>> +		extent_size = min(end - i, bitmap_range);
>> +		data_size = free_space_bitmap_size(extent_size,
>> +						   fs_info->sectorsize);
>> +
>> +		key.objectid = i;
>> +		key.type = BTRFS_FREE_SPACE_BITMAP_KEY;
>> +		key.offset = extent_size;
>> +
>> +		ret = btrfs_insert_empty_item(trans, root, path, &key,
>> +					      data_size);
>> +		if (ret)
>> +			goto out;
>> +
>> +		leaf = path->nodes[0];
>> +		ptr = btrfs_item_ptr_offset(leaf, path->slots[0]);
>> +		write_extent_buffer(leaf, bitmap_cursor, ptr,
>> +				    data_size);
>> +		btrfs_mark_buffer_dirty(leaf);
>> +		btrfs_release_path(path);
>> +
>> +		i += extent_size;
>> +		bitmap_cursor += data_size;
>> +	}
>> +
>> +	ret = 0;
>> +out:
>> +	kvfree(bitmap);
>> +	if (ret)
>> +		btrfs_abort_transaction(trans, ret);
>> +	return ret;
>> +}
>> +
>> +int convert_free_space_to_extents(struct btrfs_trans_handle *trans,
>> +				  struct btrfs_block_group_cache *block_group,
>> +				  struct btrfs_path *path)
>> +{
>> +	struct btrfs_fs_info *fs_info = trans->fs_info;
>> +	struct btrfs_root *root = fs_info->free_space_root;
>> +	struct btrfs_free_space_info *info;
>> +	struct btrfs_key key, found_key;
>> +	struct extent_buffer *leaf;
>> +	unsigned long *bitmap;
>> +	u64 start, end;
>> +	u32 bitmap_size, flags, expected_extent_count;
>> +	unsigned long nrbits, start_bit, end_bit;
>> +	u32 extent_count = 0;
>> +	int done = 0, nr;
>> +	int ret;
>> +
>> +	bitmap_size = free_space_bitmap_size(block_group->key.offset,
>> +					     fs_info->sectorsize);
>> +	bitmap = alloc_bitmap(bitmap_size);
>> +	if (!bitmap) {
>> +		ret = -ENOMEM;
>> +		goto out;
>> +	}
>> +
>> +	start = block_group->key.objectid;
>> +	end = block_group->key.objectid + block_group->key.offset;
>> +
>> +	key.objectid = end - 1;
>> +	key.type = (u8)-1;
>> +	key.offset = (u64)-1;
>> +
>> +	while (!done) {
>> +		ret = btrfs_search_prev_slot(trans, root, &key, path, -1, 1);
>> +		if (ret)
>> +			goto out;
>> +
>> +		leaf = path->nodes[0];
>> +		nr = 0;
>> +		path->slots[0]++;
>> +		while (path->slots[0] > 0) {
>> +			btrfs_item_key_to_cpu(leaf, &found_key, path->slots[0] - 1);
>> +
>> +			if (found_key.type == BTRFS_FREE_SPACE_INFO_KEY) {
>> +				ASSERT(found_key.objectid == block_group->key.objectid);
>> +				ASSERT(found_key.offset == block_group->key.offset);
>> +				done = 1;
>> +				break;
>> +			} else if (found_key.type == BTRFS_FREE_SPACE_BITMAP_KEY) {
>> +				unsigned long ptr;
>> +				char *bitmap_cursor;
>> +				u32 bitmap_pos, data_size;
>> +
>> +				ASSERT(found_key.objectid >= start);
>> +				ASSERT(found_key.objectid < end);
>> +				ASSERT(found_key.objectid + found_key.offset <= end);
>> +
>> +				bitmap_pos = div_u64(found_key.objectid - start,
>> +						     fs_info->sectorsize *
>> +						     BITS_PER_BYTE);
>> +				bitmap_cursor = ((char *)bitmap) + bitmap_pos;
>> +				data_size = free_space_bitmap_size(found_key.offset,
>> +								   fs_info->sectorsize);
>> +
>> +				ptr = btrfs_item_ptr_offset(leaf, path->slots[0] - 1);
>> +				read_extent_buffer(leaf, bitmap_cursor, ptr,
>> +						   data_size);
>> +
>> +				nr++;
>> +				path->slots[0]--;
>> +			} else {
>> +				ASSERT(0);
>> +			}
>> +		}
>> +
>> +		ret = btrfs_del_items(trans, root, path, path->slots[0], nr);
>> +		if (ret)
>> +			goto out;
>> +		btrfs_release_path(path);
>> +	}
>> +
>> +	info = search_free_space_info(trans, fs_info, block_group, path, 1);
>> +	if (IS_ERR(info)) {
>> +		ret = PTR_ERR(info);
>> +		goto out;
>> +	}
>> +	leaf = path->nodes[0];
>> +	flags = btrfs_free_space_flags(leaf, info);
>> +	flags &= ~BTRFS_FREE_SPACE_USING_BITMAPS;
>> +	btrfs_set_free_space_flags(leaf, info, flags);
>> +	expected_extent_count = btrfs_free_space_extent_count(leaf, info);
>> +	btrfs_mark_buffer_dirty(leaf);
>> +	btrfs_release_path(path);
>> +
>> +	nrbits = div_u64(block_group->key.offset, fs_info->sectorsize);
>> +	start_bit = find_next_bit_le(bitmap, nrbits, 0);
>> +
>> +	while (start_bit < nrbits) {
>> +		end_bit = find_next_zero_bit_le(bitmap, nrbits, start_bit);
>> +		ASSERT(start_bit < end_bit);
>> +
>> +		key.objectid = start + start_bit * fs_info->sectorsize;
>> +		key.type = BTRFS_FREE_SPACE_EXTENT_KEY;
>> +		key.offset = (end_bit - start_bit) * fs_info->sectorsize;
>> +
>> +		ret = btrfs_insert_empty_item(trans, root, path, &key, 0);
>> +		if (ret)
>> +			goto out;
>> +		btrfs_release_path(path);
>> +
>> +		extent_count++;
>> +
>> +		start_bit = find_next_bit_le(bitmap, nrbits, end_bit);
>> +	}
>> +
>> +	if (extent_count != expected_extent_count) {
>> +		fprintf(stderr,
>> +			"incorrect extent count for %llu; counted %u, expected %u",
>> +			block_group->key.objectid, extent_count,
>> +			expected_extent_count);
>> +		ASSERT(0);
>> +		ret = -EIO;
>> +		goto out;
>> +	}
>> +
>> +	ret = 0;
>> +out:
>> +	kvfree(bitmap);
>> +	if (ret)
>> +		btrfs_abort_transaction(trans, ret);
>> +	return ret;
>> +}
>> +
>> +static int update_free_space_extent_count(struct btrfs_trans_handle *trans,
>> +                                          struct btrfs_block_group_cache *block_group,
>> +                                          struct btrfs_path *path,
>> +                                          int new_extents)
>> +{
>> +	struct btrfs_free_space_info *info;
>> +	u32 flags;
>> +	u32 extent_count;
>> +	int ret = 0;
>> +
>> +	if (new_extents == 0)
>> +		return 0;
>> +
>> +	info = search_free_space_info(trans, trans->fs_info, block_group, path,
>> +				1);
>> +	if (IS_ERR(info)) {
>> +		ret = PTR_ERR(info);
>> +		goto out;
>> +	}
>> +	flags = btrfs_free_space_flags(path->nodes[0], info);
>> +	extent_count = btrfs_free_space_extent_count(path->nodes[0], info);
>> +
>> +	extent_count += new_extents;
>> +	btrfs_set_free_space_extent_count(path->nodes[0], info, extent_count);
>> +	btrfs_mark_buffer_dirty(path->nodes[0]);
>> +	btrfs_release_path(path);
>> +
>> +	if (!(flags & BTRFS_FREE_SPACE_USING_BITMAPS) &&
>> +	    extent_count > block_group->bitmap_high_thresh) {
>> +		ret = convert_free_space_to_bitmaps(trans, block_group, path);
>> +	} else if ((flags & BTRFS_FREE_SPACE_USING_BITMAPS) &&
>> +		   extent_count < block_group->bitmap_low_thresh) {
>> +		ret = convert_free_space_to_extents(trans, block_group, path);
>> +	}
>> +
>> +
>> +out:
>> +	return ret;
>> +}
>> +
>> +
>> +static void free_space_set_bits(struct btrfs_block_group_cache *block_group,
>> +                                struct btrfs_path *path, u64 *start, u64 *size,
>> +                                int bit)
>> +{
>> +        struct extent_buffer *leaf = path->nodes[0];
>> +        struct btrfs_fs_info *fs_info = leaf->fs_info;
>> +        struct btrfs_key key;
>> +        u64 end = *start + *size;
>> +        u64 found_start, found_end;
>> +        unsigned long ptr, first, last;
>> +
>> +        btrfs_item_key_to_cpu(leaf, &key, path->slots[0]);
>> +        ASSERT(key.type == BTRFS_FREE_SPACE_BITMAP_KEY);
>> +
>> +        found_start = key.objectid;
>> +        found_end = key.objectid + key.offset;
>> +        ASSERT(*start >= found_start && *start < found_end);
>> +        ASSERT(end > found_start);
>> +
>> +        if (end > found_end)
>> +                end = found_end;
>> +
>> +        ptr = btrfs_item_ptr_offset(leaf, path->slots[0]);
>> +        first = (*start - found_start) / fs_info->sectorsize;
>> +        last = (end - found_start) / fs_info->sectorsize;
>> +        if (bit)
>> +                extent_buffer_bitmap_set(leaf, ptr, first, last - first);
>> +        else
>> +                extent_buffer_bitmap_clear(leaf, ptr, first, last - first);
>> +        btrfs_mark_buffer_dirty(leaf);
>> +
>> +        *size -= end - *start;
>> +        *start = end;
>> +}
>> +
>> +/*
>> + * We can't use btrfs_next_item() in modify_free_space_bitmap() because
>> + * btrfs_next_leaf() doesn't get the path for writing. We can forgo the fancy
>> + * tree walking in btrfs_next_leaf() anyways because we know exactly what we're
>> + * looking for.
>> + */
>> +static int free_space_next_bitmap(struct btrfs_trans_handle *trans,
>> +                                  struct btrfs_root *root, struct btrfs_path *p)
>> +{
>> +	struct btrfs_key key;
>> +
>> +	if (p->slots[0] + 1 < btrfs_header_nritems(p->nodes[0])) {
>> +		p->slots[0]++;
>> +		return 0;
>> +	}
>> +
>> +	btrfs_item_key_to_cpu(p->nodes[0], &key, p->slots[0]);
>> +	btrfs_release_path(p);
>> +
>> +	key.objectid += key.offset;
>> +	key.type = (u8)-1;
>> +	key.offset = (u64)-1;
>> +
>> +	return btrfs_search_prev_slot(trans, root, &key, p, 0, 1);
>> +}
>> +
>> +/*
>> + * If remove is 1, then we are removing free space, thus clearing bits in the
>> + * bitmap. If remove is 0, then we are adding free space, thus setting bits in
>> + * the bitmap.
>> + */
>> +static int modify_free_space_bitmap(struct btrfs_trans_handle *trans,
>> +                                    struct btrfs_block_group_cache *block_group,
>> +                                    struct btrfs_path *path,
>> +                                    u64 start, u64 size, int remove)
>> +{
>> +        struct btrfs_root *root = trans->fs_info->free_space_root;
>> +        struct btrfs_key key;
>> +        u64 end = start + size;
>> +        u64 cur_start, cur_size;
>> +        int prev_bit, next_bit;
>> +        int new_extents;
>> +        int ret;
>> +
>> +        /*
>> +         * Read the bit for the block immediately before the extent of space if
>> +         * that block is within the block group.
>> +         */
>> +        if (start > block_group->key.objectid) {
>> +                u64 prev_block = start - trans->fs_info->sectorsize;
>> +
>> +                key.objectid = prev_block;
>> +                key.type = (u8)-1;
>> +                key.offset = (u64)-1;
>> +
>> +                ret = btrfs_search_prev_slot(trans, root, &key, path, 0, 1);
>> +                if (ret)
>> +                        goto out;
>> +
>> +                prev_bit = free_space_test_bit(block_group, path, prev_block);
>> +
>> +                /* The previous block may have been in the previous bitmap. */
>> +                btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]);
>> +                if (start >= key.objectid + key.offset) {
>> +                        ret = free_space_next_bitmap(trans, root, path);
>> +                        if (ret)
>> +                                goto out;
>> +                }
>> +        } else {
>> +                key.objectid = start;
>> +                key.type = (u8)-1;
>> +                key.offset = (u64)-1;
>> +
>> +                ret = btrfs_search_prev_slot(trans, root, &key, path, 0, 1);
>> +                if (ret)
>> +                        goto out;
>> +
>> +                prev_bit = -1;
>> +        }
>> +
>> +        /*
>> +         * Iterate over all of the bitmaps overlapped by the extent of space,
>> +         * clearing/setting bits as required.
>> +         */
>> +        cur_start = start;
>> +        cur_size = size;
>> +        while (1) {
>> +                free_space_set_bits(block_group, path, &cur_start, &cur_size,
>> +                                    !remove);
>> +                if (cur_size == 0)
>> +                        break;
>> +                ret = free_space_next_bitmap(trans, root, path);
>> +                if (ret)
>> +                        goto out;
>> +        }
>> +
>> +	/*
>> +         * Read the bit for the block immediately after the extent of space if
>> +         * that block is within the block group.
>> +         */
>> +        if (end < block_group->key.objectid + block_group->key.offset) {
>> +                /* The next block may be in the next bitmap. */
>> +                btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]);
>> +                if (end >= key.objectid + key.offset) {
>> +                        ret = free_space_next_bitmap(trans, root, path);
>> +                        if (ret)
>> +                                goto out;
>> +                }
>> +
>> +                next_bit = free_space_test_bit(block_group, path, end);
>> +        } else {
>> +                next_bit = -1;
>> +        }
>> +
>> +        if (remove) {
>> +                new_extents = -1;
>> +                if (prev_bit == 1) {
>> +                        /* Leftover on the left. */
>> +                        new_extents++;
>> +                }
>> +                if (next_bit == 1) {
>> +                        /* Leftover on the right. */
>> +                        new_extents++;
>> +                }
>> +        } else {
>> +                new_extents = 1;
>> +                if (prev_bit == 1) {
>> +                        /* Merging with neighbor on the left. */
>> +                        new_extents--;
>> +                }
>> +                if (next_bit == 1) {
>> +                        /* Merging with neighbor on the right. */
>> +                        new_extents--;
>> +                }
>> +        }
>> +
>> +        btrfs_release_path(path);
>> +        ret = update_free_space_extent_count(trans, block_group, path,
>> +                                             new_extents);
>> +
>> +out:
>> +        return ret;
>> +}
>> +
>> +static int remove_free_space_extent(struct btrfs_trans_handle *trans,
>> +				    struct btrfs_block_group_cache *block_group,
>> +				    struct btrfs_path *path,
>> +				    u64 start, u64 size)
>> +{
>> +	struct btrfs_root *root = trans->fs_info->free_space_root;
>> +	struct btrfs_key key;
>> +	u64 found_start, found_end;
>> +	u64 end = start + size;
>> +	int new_extents = -1;
>> +	int ret;
>> +
>> +	key.objectid = start;
>> +	key.type = (u8)-1;
>> +	key.offset = (u64)-1;
>> +
>> +	ret = btrfs_search_prev_slot(trans, root, &key, path, -1, 1);
>> +	if (ret)
>> +		goto out;
>> +
>> +	btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]);
>> +
>> +	ASSERT(key.type == BTRFS_FREE_SPACE_EXTENT_KEY);
>> +
>> +	found_start = key.objectid;
>> +	found_end = key.objectid + key.offset;
>> +	ASSERT(start >= found_start && end <= found_end);
>> +
>> +	/*
>> +	 * Okay, now that we've found the free space extent which contains the
>> +	 * free space that we are removing, there are four cases:
>> +	 *
>> +	 * 1. We're using the whole extent: delete the key we found and
>> +	 * decrement the free space extent count.
>> +	 * 2. We are using part of the extent starting at the beginning: delete
>> +	 * the key we found and insert a new key representing the leftover at
>> +	 * the end. There is no net change in the number of extents.
>> +	 * 3. We are using part of the extent ending at the end: delete the key
>> +	 * we found and insert a new key representing the leftover at the
>> +	 * beginning. There is no net change in the number of extents.
>> +	 * 4. We are using part of the extent in the middle: delete the key we
>> +	 * found and insert two new keys representing the leftovers on each
>> +	 * side. Where we used to have one extent, we now have two, so increment
>> +	 * the extent count. We may need to convert the block group to bitmaps
>> +	 * as a result.
>> +	 */
>> +
>> +	/* Delete the existing key (cases 1-4). */
>> +	ret = btrfs_del_item(trans, root, path);
>> +	if (ret)
>> +		goto out;
>> +
>> +	/* Add a key for leftovers at the beginning (cases 3 and 4). */
>> +	if (start > found_start) {
>> +		key.objectid = found_start;
>> +		key.type = BTRFS_FREE_SPACE_EXTENT_KEY;
>> +		key.offset = start - found_start;
>> +
>> +		btrfs_release_path(path);
>> +		ret = btrfs_insert_empty_item(trans, root, path, &key, 0);
>> +		if (ret)
>> +			goto out;
>> +		new_extents++;
>> +	}
>> +
>> +	/* Add a key for leftovers at the end (cases 2 and 4). */
>> +	if (end < found_end) {
>> +		key.objectid = end;
>> +		key.type = BTRFS_FREE_SPACE_EXTENT_KEY;
>> +		key.offset = found_end - end;
>> +
>> +		btrfs_release_path(path);
>> +		ret = btrfs_insert_empty_item(trans, root, path, &key, 0);
>> +		if (ret)
>> +			goto out;
>> +		new_extents++;
>> +	}
>> +
>> +	btrfs_release_path(path);
>> +	ret = update_free_space_extent_count(trans, block_group, path,
>> +					     new_extents);
>> +
>> +out:
>> +	return ret;
>> +}
>> +
>> +int __remove_from_free_space_tree(struct btrfs_trans_handle *trans,
>> +                                  struct btrfs_block_group_cache *block_group,
>> +                                  struct btrfs_path *path, u64 start, u64 size)
>> +{
>> +	struct btrfs_free_space_info *info;
>> +	u32 flags;
>> +
>> +	info = search_free_space_info(NULL, trans->fs_info, block_group, path,
>> +	                              0);
>> +	if (IS_ERR(info))
>> +		return PTR_ERR(info);
>> +	flags = btrfs_free_space_flags(path->nodes[0], info);
>> +	btrfs_release_path(path);
>> +
>> +	if (flags & BTRFS_FREE_SPACE_USING_BITMAPS) {
>> +		return modify_free_space_bitmap(trans, block_group, path,
>> +	                                        start, size, 1);
>> +	} else {
>> +		return remove_free_space_extent(trans, block_group, path,
>> +	                                        start, size);
>> +	}
>> +}
>> +
>> +int remove_from_free_space_tree(struct btrfs_trans_handle *trans,
>> +				u64 start, u64 size)
>> +{
>> +	struct btrfs_block_group_cache *block_group;
>> +	struct btrfs_path *path;
>> +	int ret;
>> +
>> +	if (!btrfs_fs_compat_ro(trans->fs_info, FREE_SPACE_TREE))
>> +		return 0;
>> +
>> +	path = btrfs_alloc_path();
>> +	if (!path) {
>> +		ret = -ENOMEM;
>> +		goto out;
>> +	}
>> +
>> +	block_group = btrfs_lookup_block_group(trans->fs_info, start);
>> +	if (!block_group) {
>> +		ASSERT(0);
>> +		ret = -ENOENT;
>> +		goto out;
>> +	}
>> +
>> +	ret = __remove_from_free_space_tree(trans, block_group, path, start,
>> +					    size);
>> +out:
>> +	btrfs_free_path(path);
>> +	if (ret)
>> +		btrfs_abort_transaction(trans, ret);
>> +	return ret;
>> +}
>> +
>> +static int add_free_space_extent(struct btrfs_trans_handle *trans,
>> +                                 struct btrfs_block_group_cache *block_group,
>> +                                 struct btrfs_path *path,
>> +                                 u64 start, u64 size)
>> +{
>> +        struct btrfs_root *root = trans->fs_info->free_space_root;
>> +        struct btrfs_key key, new_key;
>> +        u64 found_start, found_end;
>> +        u64 end = start + size;
>> +        int new_extents = 1;
>> +        int ret;
>> +
>> +        /*
>> +         * We are adding a new extent of free space, but we need to merge
>> +         * extents. There are four cases here:
>> +         *
>> +         * 1. The new extent does not have any immediate neighbors to merge
>> +         * with: add the new key and increment the free space extent count. We
>> +         * may need to convert the block group to bitmaps as a result.
>> +         * 2. The new extent has an immediate neighbor before it: remove the
>> +         * previous key and insert a new key combining both of them. There is no
>> +         * net change in the number of extents.
>> +         * 3. The new extent has an immediate neighbor after it: remove the next
>> +         * key and insert a new key combining both of them. There is no net
>> +         * change in the number of extents.
>> +         * 4. The new extent has immediate neighbors on both sides: remove both
>> +         * of the keys and insert a new key combining all of them. Where we used
>> +         * to have two extents, we now have one, so decrement the extent count.
>> +         */
>> +
>> +        new_key.objectid = start;
>> +        new_key.type = BTRFS_FREE_SPACE_EXTENT_KEY;
>> +        new_key.offset = size;
>> +
>> +        /* Search for a neighbor on the left. */
>> +        if (start == block_group->key.objectid)
>> +                goto right;
>> +        key.objectid = start - 1;
>> +        key.type = (u8)-1;
>> +        key.offset = (u64)-1;
>> +
>> +        ret = btrfs_search_prev_slot(trans, root, &key, path, -1, 1);
>> +        if (ret)
>> +                goto out;
>> +
>> +        btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]);
>> +
>> +        if (key.type != BTRFS_FREE_SPACE_EXTENT_KEY) {
>> +                ASSERT(key.type == BTRFS_FREE_SPACE_INFO_KEY);
>> +                btrfs_release_path(path);
>> +                goto right;
>> +        }
>> +
>> +        found_start = key.objectid;
>> +        found_end = key.objectid + key.offset;
>> +        ASSERT(found_start >= block_group->key.objectid &&
>> +               found_end > block_group->key.objectid);
>> +        ASSERT(found_start < start && found_end <= start);
>> +
>> +        /*
>> +         * Delete the neighbor on the left and absorb it into the new key (cases
>> +         * 2 and 4).
>> +         */
>> +        if (found_end == start) {
>> +                ret = btrfs_del_item(trans, root, path);
>> +                if (ret)
>> +                        goto out;
>> +                new_key.objectid = found_start;
>> +                new_key.offset += key.offset;
>> +                new_extents--;
>> +        }
>> +        btrfs_release_path(path);
>> +right:
>> +        /* Search for a neighbor on the right. */
>> +        if (end == block_group->key.objectid + block_group->key.offset)
>> +                goto insert;
>> +        key.objectid = end;
>> +        key.type = (u8)-1;
>> +        key.offset = (u64)-1;
>> +
>> +        ret = btrfs_search_prev_slot(trans, root, &key, path, -1, 1);
>> +        if (ret)
>> +                goto out;
>> +
>> +        btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]);
>> +
>> +        if (key.type != BTRFS_FREE_SPACE_EXTENT_KEY) {
>> +                ASSERT(key.type == BTRFS_FREE_SPACE_INFO_KEY);
>> +                btrfs_release_path(path);
>> +                goto insert;
>> +        }
>> +
>> +        found_start = key.objectid;
>> +        found_end = key.objectid + key.offset;
>> +        ASSERT(found_start >= block_group->key.objectid &&
>> +               found_end > block_group->key.objectid);
>> +        ASSERT((found_start < start && found_end <= start) ||
>> +               (found_start >= end && found_end > end));
>> +
>> +        /*
>> +         * Delete the neighbor on the right and absorb it into the new key
>> +         * (cases 3 and 4).
>> +         */
>> +        if (found_start == end) {
>> +                ret = btrfs_del_item(trans, root, path);
>> +                if (ret)
>> +                        goto out;
>> +                new_key.offset += key.offset;
>> +                new_extents--;
>> +        }
>> +        btrfs_release_path(path);
>> +
>> +insert:
>> +        /* Insert the new key (cases 1-4). */
>> +        ret = btrfs_insert_empty_item(trans, root, path, &new_key, 0);
>> +        if (ret)
>> +                goto out;
>> +
>> +        btrfs_release_path(path);
>> +        ret = update_free_space_extent_count(trans, block_group, path,
>> +                                             new_extents);
>> +
>> +out:
>> +        return ret;
>> +}
>> +
>> +int __add_to_free_space_tree(struct btrfs_trans_handle *trans,
>> +                             struct btrfs_block_group_cache *block_group,
>> +                             struct btrfs_path *path, u64 start, u64 size)
>> +{
>> +	struct btrfs_fs_info *fs_info = trans->fs_info;
>> +	struct btrfs_free_space_info *info;
>> +	u32 flags;
>> +
>> +	info = search_free_space_info(NULL, fs_info, block_group, path, 0);
>> +	if (IS_ERR(info))
>> +	        return PTR_ERR(info);
>> +	flags = btrfs_free_space_flags(path->nodes[0], info);
>> +	btrfs_release_path(path);
>> +
>> +	if (flags & BTRFS_FREE_SPACE_USING_BITMAPS) {
>> +	        return modify_free_space_bitmap(trans, block_group, path,
>> +	                                        start, size, 0);
>> +	} else {
>> +	        return add_free_space_extent(trans, block_group, path, start,
>> +	                                     size);
>> +	}
>> +}
>> +
>> +
>> +int add_to_free_space_tree(struct btrfs_trans_handle *trans,
>> +			   u64 start, u64 size)
>> +{
>> +	struct btrfs_block_group_cache *block_group;
>> +	struct btrfs_path *path;
>> +	int ret;
>> +
>> +	if (!btrfs_fs_compat_ro(trans->fs_info, FREE_SPACE_TREE))
>> +		return 0;
>> +
>> +	path = btrfs_alloc_path();
>> +	if (!path) {
>> +		ret = -ENOMEM;
>> +		goto out;
>> +	}
>> +
>> +	block_group = btrfs_lookup_block_group(trans->fs_info, start);
>> +	if (!block_group) {
>> +		ASSERT(0);
>> +		ret = -ENOENT;
>> +		goto out;
>> +	}
>> +
>> +	ret = __add_to_free_space_tree(trans, block_group, path, start, size);
>> +out:
>> +	btrfs_free_path(path);
>> +	if (ret)
>> +		btrfs_abort_transaction(trans, ret);
>> +	return ret;
>> +}
>> +
>> +int populate_free_space_tree(struct btrfs_trans_handle *trans,
>> +			     struct btrfs_block_group_cache *block_group)
>> +{
>> +        struct btrfs_root *extent_root = trans->fs_info->extent_root;
>> +        struct btrfs_path *path, *path2;
>> +        struct btrfs_key key;
>> +        u64 start, end;
>> +        int ret;
>> +
>> +        path = btrfs_alloc_path();
>> +        if (!path)
>> +                return -ENOMEM;
>> +        path->reada = READA_FORWARD;
>> +
>> +        path2 = btrfs_alloc_path();
>> +        if (!path2) {
>> +                btrfs_free_path(path);
>> +                return -ENOMEM;
>> +        }
>> +
>> +        ret = add_new_free_space_info(trans, block_group, path2);
>> +        if (ret)
>> +                goto out;
>> +
>> +        /*
>> +         * Iterate through all of the extent and metadata items in this block
>> +         * group, adding the free space between them and the free space at the
>> +         * end. Note that EXTENT_ITEM and METADATA_ITEM are less than
>> +         * BLOCK_GROUP_ITEM, so an extent may precede the block group that it's
>> +         * contained in.
>> +         */
>> +        key.objectid = block_group->key.objectid;
>> +        key.type = BTRFS_EXTENT_ITEM_KEY;
>> +        key.offset = 0;
>> +
>> +        ret = btrfs_search_slot_for_read(extent_root, &key, path, 1, 0);
>> +        if (ret < 0)
>> +                goto out;
>> +        ASSERT(ret == 0);
>> +
>> +        start = block_group->key.objectid;
>> +        end = block_group->key.objectid + block_group->key.offset;
>> +        while (1) {
>> +                btrfs_item_key_to_cpu(path->nodes[0], &key, path->slots[0]);
>> +
>> +                if (key.type == BTRFS_EXTENT_ITEM_KEY ||
>> +                    key.type == BTRFS_METADATA_ITEM_KEY) {
>> +                        if (key.objectid >= end)
>> +                                break;
>> +
>> +                        if (start < key.objectid) {
>> +                                ret = __add_to_free_space_tree(trans,
>> +                                                               block_group,
>> +                                                               path2, start,
>> +                                                               key.objectid -
>> +                                                               start);
>> +                                if (ret)
>> +                                        goto out;
>> +                        }
>> +                        start = key.objectid;
>> +                        if (key.type == BTRFS_METADATA_ITEM_KEY)
>> +                                start += trans->fs_info->nodesize;
>> +                        else
>> +                                start += key.offset;
>> +                } else if (key.type == BTRFS_BLOCK_GROUP_ITEM_KEY) {
>> +                        if (key.objectid != block_group->key.objectid)
>> +                                break;
>> +                }
>> +
>> +                ret = btrfs_next_item(extent_root, path);
>> +                if (ret < 0)
>> +                        goto out;
>> +                if (ret)
>> +                        break;
>> +        }
>> +        if (start < end) {
>> +                ret = __add_to_free_space_tree(trans, block_group, path2,
>> +                                               start, end - start);
>> +                if (ret)
>> +                        goto out;
>> +        }
>> +
>> +        ret = 0;
>> +out:
>> +        btrfs_free_path(path2);
>> +        btrfs_free_path(path);
>> +        return ret;
>> +}
>> +
>> +int remove_block_group_free_space(struct btrfs_trans_handle *trans,
>> +				  struct btrfs_block_group_cache *block_group)
>> +{
>> +	struct btrfs_root *root = trans->fs_info->free_space_root;
>> +	struct btrfs_path *path;
>> +	struct btrfs_key key, found_key;
>> +	struct extent_buffer *leaf;
>> +	u64 start, end;
>> +	int done = 0, nr;
>> +	int ret;
>> +
>> +	path = btrfs_alloc_path();
>> +	if (!path) {
>> +		ret = -ENOMEM;
>> +		goto out;
>> +	}
>> +
>> +	start = block_group->key.objectid;
>> +	end = block_group->key.objectid + block_group->key.offset;
>> +
>> +	key.objectid = end - 1;
>> +	key.type = (u8)-1;
>> +	key.offset = (u64)-1;
>> +
>> +	while (!done) {
>> +		ret = btrfs_search_prev_slot(trans, root, &key, path, -1, 1);
>> +		if (ret)
>> +			goto out;
>> +
>> +		leaf = path->nodes[0];
>> +		nr = 0;
>> +		path->slots[0]++;
>> +		while (path->slots[0] > 0) {
>> +			btrfs_item_key_to_cpu(leaf, &found_key, path->slots[0] - 1);
>> +
>> +			if (found_key.type == BTRFS_FREE_SPACE_INFO_KEY) {
>> +				ASSERT(found_key.objectid == block_group->key.objectid);
>> +				ASSERT(found_key.offset == block_group->key.offset);
>> +				done = 1;
>> +				nr++;
>> +				path->slots[0]--;
>> +				break;
>> +			} else if (found_key.type == BTRFS_FREE_SPACE_EXTENT_KEY ||
>> +				   found_key.type == BTRFS_FREE_SPACE_BITMAP_KEY) {
>> +				ASSERT(found_key.objectid >= start);
>> +				ASSERT(found_key.objectid < end);
>> +				ASSERT(found_key.objectid + found_key.offset <= end);
>> +				nr++;
>> +				path->slots[0]--;
>> +			} else {
>> +				ASSERT(0);
>> +			}
>> +		}
>> +
>> +		ret = btrfs_del_items(trans, root, path, path->slots[0], nr);
>> +		if (ret)
>> +			goto out;
>> +		btrfs_release_path(path);
>> +	}
>> +
>> +	ret = 0;
>> +out:
>> +	btrfs_free_path(path);
>> +	if (ret)
>> +		btrfs_abort_transaction(trans, ret);
>> +	return ret;
>> +}
>>  static int clear_free_space_tree(struct btrfs_trans_handle *trans,
>>  				 struct btrfs_root *root)
>>  {
>> @@ -204,8 +1309,8 @@ static int load_free_space_bitmaps(struct btrfs_fs_info *fs_info,
>>  
>>  		offset = key.objectid;
>>  		while (offset < key.objectid + key.offset) {
>> -			bit = free_space_test_bit(block_group, path, offset,
>> -						  fs_info->sectorsize);
>> +			bit = free_space_test_bit(block_group, path, offset);
>> +
>>  			if (prev_bit == 0 && bit == 1) {
>>  				extent_start = offset;
>>  			} else if (prev_bit == 1 && bit == 0) {
>> @@ -320,6 +1425,142 @@ static int load_free_space_extents(struct btrfs_fs_info *fs_info,
>>  	return ret;
>>  }
>>  
>> +struct btrfs_root *btrfs_create_tree(struct btrfs_trans_handle *trans,
>> +                                     struct btrfs_fs_info *fs_info,
>> +                                     u64 objectid)
>> +{
>> +	struct extent_buffer *leaf;
>> +	struct btrfs_root *tree_root = fs_info->tree_root;
>> +	struct btrfs_root *root;
>> +	struct btrfs_key key;
>> +	int ret = 0;
>> +
>> +	root = kzalloc(sizeof(*root), GFP_KERNEL);
>> +	if (!root)
>> +		return ERR_PTR(-ENOMEM);
>> +
>> +	btrfs_setup_root(root, fs_info, objectid);
>> +	root->root_key.objectid = objectid;
>> +	root->root_key.type = BTRFS_ROOT_ITEM_KEY;
>> +	root->root_key.offset = 0;
>> +
>> +	leaf = btrfs_alloc_free_block(trans, root, fs_info->nodesize, objectid, NULL, 0, 0, 0);
>> +	if (IS_ERR(leaf)) {
>> +		ret = PTR_ERR(leaf);
>> +		leaf = NULL;
>> +		goto fail;
>> +	}
>> +
>> +	memset_extent_buffer(leaf, 0, 0, sizeof(struct btrfs_header));
>> +	btrfs_set_header_bytenr(leaf, leaf->start);
>> +	btrfs_set_header_generation(leaf, trans->transid);
>> +	btrfs_set_header_backref_rev(leaf, BTRFS_MIXED_BACKREF_REV);
>> +	btrfs_set_header_owner(leaf, objectid);
>> +	root->node = leaf;
>> +	write_extent_buffer(leaf, fs_info->fsid, btrfs_header_fsid(), BTRFS_FSID_SIZE);
>> +	write_extent_buffer(leaf, fs_info->chunk_tree_uuid,
>> +			    btrfs_header_chunk_tree_uuid(leaf),
>> +			    BTRFS_UUID_SIZE);
>> +	btrfs_mark_buffer_dirty(leaf);
>> +
>> +	extent_buffer_get(root->node);
>> +	root->commit_root = root->node;
>> +	root->track_dirty = 1;
>> +
>> +	root->root_item.flags = 0;
>> +	root->root_item.byte_limit = 0;
>> +	btrfs_set_root_bytenr(&root->root_item, leaf->start);
>> +	btrfs_set_root_generation(&root->root_item, trans->transid);
>> +	btrfs_set_root_level(&root->root_item, 0);
>> +	btrfs_set_root_refs(&root->root_item, 1);
>> +	btrfs_set_root_used(&root->root_item, leaf->len);
>> +	btrfs_set_root_last_snapshot(&root->root_item, 0);
>> +	btrfs_set_root_dirid(&root->root_item, 0);
>> +	memset(root->root_item.uuid, 0, BTRFS_UUID_SIZE);
>> +	root->root_item.drop_level = 0;
>> +
>> +	key.objectid = objectid;
>> +	key.type = BTRFS_ROOT_ITEM_KEY;
>> +	key.offset = 0;
>> +	ret = btrfs_insert_root(trans, tree_root, &key, &root->root_item);
>> +	if (ret)
>> +		goto fail;
>> +
>> +	return root;
>> +
>> +fail:
>> +	if (leaf)
>> +		free_extent_buffer(leaf);
>> +
>> +	kfree(root);
>> +	return ERR_PTR(ret);
>> +}
>> +
>> +#define btrfs_set_fs_compat_ro(__fs_info, opt) \
>> +        __btrfs_set_fs_compat_ro((__fs_info), BTRFS_FEATURE_COMPAT_RO_##opt)
>> +
>> +static inline void __btrfs_set_fs_compat_ro(struct btrfs_fs_info *fs_info,
>> +                                            u64 flag)
>> +{
>> +	struct btrfs_super_block *disk_super;
>> +	u64 features;
>> +
>> +	disk_super = fs_info->super_copy;
>> +	features = btrfs_super_compat_ro_flags(disk_super);
>> +	if (!(features & flag)) {
>> +		features = btrfs_super_compat_ro_flags(disk_super);
>> +		if (!(features & flag)) {
>> +			features |= flag;
>> +			btrfs_set_super_compat_ro_flags(disk_super, features);
>> +		}
>> +	}
>> +}
>> +
>> +int btrfs_create_free_space_tree(struct btrfs_fs_info *fs_info)
>> +{
>> +	struct btrfs_trans_handle *trans;
>> +	struct btrfs_root *tree_root = fs_info->tree_root;
>> +	struct btrfs_root *free_space_root;
>> +	struct btrfs_block_group_cache *block_group;
>> +	u64 start = BTRFS_SUPER_INFO_OFFSET + BTRFS_SUPER_INFO_SIZE;
>> +	int ret;
>> +
>> +	trans = btrfs_start_transaction(tree_root, 0);
>> +	if (IS_ERR(trans))
>> +		return PTR_ERR(trans);
>> +
>> +	free_space_root = btrfs_create_tree(trans, fs_info,
>> +					    BTRFS_FREE_SPACE_TREE_OBJECTID);
>> +	if (IS_ERR(free_space_root)) {
>> +		ret = PTR_ERR(free_space_root);
>> +		goto abort;
>> +	}
>> +	fs_info->free_space_root = free_space_root;
>> +
>> +	do {
>> +		block_group = btrfs_lookup_first_block_group(fs_info, start);
>> +		if (!block_group)
>> +			break;
>> +		start = block_group->key.objectid + block_group->key.offset;
>> +		ret = populate_free_space_tree(trans, block_group);
>> +		if (ret)
>> +			goto abort;
>> +	} while (block_group);
>> +
>> +	btrfs_set_fs_compat_ro(fs_info, FREE_SPACE_TREE);
>> +	btrfs_set_fs_compat_ro(fs_info, FREE_SPACE_TREE_VALID);
>> +
>> +	ret = btrfs_commit_transaction(trans, tree_root);
>> +	if (ret)
>> +		return ret;
>> +
>> +	return 0;
>> +
>> +abort:
>> +	btrfs_abort_transaction(trans, ret);
>> +	return ret;
>> +}
>> +
>>  int load_free_space_tree(struct btrfs_fs_info *fs_info,
>>  			 struct btrfs_block_group_cache *block_group)
>>  {
>> @@ -332,7 +1573,7 @@ int load_free_space_tree(struct btrfs_fs_info *fs_info,
>>  	path = btrfs_alloc_path();
>>  	if (!path)
>>  		return -ENOMEM;
>> -	path->reada = 1;
>> +	path->reada = READA_BACK;
>>  
>>  	info = search_free_space_info(NULL, fs_info, block_group, path, 0);
>>  	if (IS_ERR(info)) {
>> diff --git a/free-space-tree.h b/free-space-tree.h
>> index 4845f13e6808..9530c2882358 100644
>> --- a/free-space-tree.h
>> +++ b/free-space-tree.h
>> @@ -19,8 +19,19 @@
>>  #ifndef __BTRFS_FREE_SPACE_TREE_H__
>>  #define __BTRFS_FREE_SPACE_TREE_H__
>>  
>> +#define BTRFS_FREE_SPACE_BITMAP_SIZE 256
>> +#define BTRFS_FREE_SPACE_BITMAP_BITS (BTRFS_FREE_SPACE_BITMAP_SIZE * BITS_PER_BYTE)
>> +
>>  int btrfs_clear_free_space_tree(struct btrfs_fs_info *fs_info);
>>  int load_free_space_tree(struct btrfs_fs_info *fs_info,
>>  			 struct btrfs_block_group_cache *block_group);
>> -
>> +int populate_free_space_tree(struct btrfs_trans_handle *trans,
>> +			     struct btrfs_block_group_cache *block_group);
>> +int remove_block_group_free_space(struct btrfs_trans_handle *trans,
>> +				  struct btrfs_block_group_cache *block_group);
>> +int add_to_free_space_tree(struct btrfs_trans_handle *trans, u64 start,
>> +			   u64 size);
>> +int remove_from_free_space_tree(struct btrfs_trans_handle *trans, u64 start,
>> +				u64 size);
>> +int btrfs_create_free_space_tree(struct btrfs_fs_info *info);
>>  #endif
>> diff --git a/kerncompat.h b/kerncompat.h
>> index 1a2bc18c3ac2..a223a7f009bd 100644
>> --- a/kerncompat.h
>> +++ b/kerncompat.h
>> @@ -263,6 +263,8 @@ static inline int IS_ERR_OR_NULL(const void *ptr)
>>  	return !ptr || IS_ERR(ptr);
>>  }
>>  
>> +#define div_u64(x, y) ((x) / (y))
>> +
>>  /**
>>   * swap - swap values of @a and @b
>>   * @a: first value
>> @@ -297,6 +299,10 @@ static inline int IS_ERR_OR_NULL(const void *ptr)
>>  #define kfree(x) free(x)
>>  #define vmalloc(x) malloc(x)
>>  #define vfree(x) free(x)
>> +#define kvzalloc(x, y) kzalloc(x,y)
>> +#define kvfree(x) free(x)
>> +#define memalloc_nofs_save() (0)
>> +#define memalloc_nofs_restore(x)
>>  
>>  #ifndef BTRFS_DISABLE_BACKTRACE
>>  static inline void assert_trace(const char *assertion, const char *filename,
>> -- 
>> 2.7.4
>>
> 

^ permalink raw reply	[flat|nested] 27+ messages in thread

* Re: [PATCH 07/10] btrfs-progs: Add freespace tree as compat_ro supported feature
  2018-10-04 18:30   ` Omar Sandoval
@ 2018-10-04 18:36     ` Nikolay Borisov
  0 siblings, 0 replies; 27+ messages in thread
From: Nikolay Borisov @ 2018-10-04 18:36 UTC (permalink / raw)
  To: Omar Sandoval; +Cc: linux-btrfs



On  4.10.2018 21:30, Omar Sandoval wrote:
> On Mon, Oct 01, 2018 at 05:46:18PM +0300, Nikolay Borisov wrote:
>> The RO_FREE_SPACE_TREE(_VALID) flags are required in order to be able
>> to open an FST filesystem in repair mode. Add them to
>> BTRFS_FEATURE_COMPAT_RO_SUPP.
>>
>> Signed-off-by: Nikolay Borisov <nborisov@suse.com>
>> ---
>>  ctree.h | 4 +++-
>>  1 file changed, 3 insertions(+), 1 deletion(-)
>>
>> diff --git a/ctree.h b/ctree.h
>> index a6d6c3decd87..3c396e7d293d 100644
>> --- a/ctree.h
>> +++ b/ctree.h
>> @@ -497,7 +497,9 @@ struct btrfs_super_block {
>>   * added here until read-write support for the free space tree is implemented in
>>   * btrfs-progs.
>>   */
> 
> This comment should go away.
> 
>> -#define BTRFS_FEATURE_COMPAT_RO_SUPP		0ULL
>> +#define BTRFS_FEATURE_COMPAT_RO_SUPP			\
>> +	(BTRFS_FEATURE_COMPAT_RO_FREE_SPACE_TREE |	\
>> +	 BTRFS_FEATURE_COMPAT_RO_FREE_SPACE_TREE_VALID)
>>  
>>  #define BTRFS_FEATURE_INCOMPAT_SUPP			\
>>  	(BTRFS_FEATURE_INCOMPAT_MIXED_BACKREF |		\
> 
> To repeat my question from before, did you test whether we can properly
> change the filesystem with, e.g., btrfstune or btrfs fi label? Given
> that some critical code was missing in the free space tree code, I'd be
> surprised if it worked correctly.
> 

Yes, it works. I manually tried changing the label of the filesystem
(though now I'm thinking whether I had v2 space cache enabled) and also
run all tests for progs. Will revise this however with explicit v2 fst.

^ permalink raw reply	[flat|nested] 27+ messages in thread

* Re: [PATCH 05/10] btrfs-progs: Pull free space tree related code from kernel
  2018-10-04 18:34     ` Nikolay Borisov
@ 2018-10-04 19:01       ` Omar Sandoval
  0 siblings, 0 replies; 27+ messages in thread
From: Omar Sandoval @ 2018-10-04 19:01 UTC (permalink / raw)
  To: Nikolay Borisov; +Cc: linux-btrfs

On Thu, Oct 04, 2018 at 09:34:01PM +0300, Nikolay Borisov wrote:
> 
> 
> On  4.10.2018 21:26, Omar Sandoval wrote:
> > On Mon, Oct 01, 2018 at 05:46:16PM +0300, Nikolay Borisov wrote:
> >> To help implement free space tree checker in user space some kernel
> >> function are necessary, namely iterating/deleting/adding freespace
> >> items, some internal search functions. Functions to populate a block
> >> group based on the extent tree. The code is largely copy/paste from
> >> the kernel with locking eliminated (i.e free_space_lock). It supports
> >> reading/writing of both bitmap and extent based FST trees.
> > 
> > For some reason, a lot of this added code uses spaces instead of tabs,
> > so I had to fix that in order to compare it to the kernel code (some of
> > the functions were reordered, too).
> > 
> > The only functional difference I noticed was that this is missing the
> > code to insert the block group header in the free space tree:
> > 
> > 	if (block_group->needs_free_space) {
> > 		ret = __add_block_group_free_space(trans, block_group, path);
> > 		if (ret)
> > 			return ret;
> > 	}
> > 
> > Was that intentionally omitted? Without it, the free space tree is
> > pretty broken :(
> 
> Yes, it was intentional. If you remember I even emailed you about this
> particular piece of code and you said you needed to have it this way in
> case delayed refs were run before the space tree was initialized.
> 
> In user space AFAIK we don't have that problem since
> add_new_free_space_info/__add_to_free_space_tree is called via
> 
> btrfs_Create_free_space_tree
>    populate_free_space_tree
>      add_new_free_space_info
>      __add_to_free_space_tree

Ah, will progs never allocate a new block group? If so, then you're
right, this is fine.

^ permalink raw reply	[flat|nested] 27+ messages in thread

* Re: [PATCH 08/10] btrfs-progs: check: Add support for freespace tree fixing
  2018-10-01 14:46 ` [PATCH 08/10] btrfs-progs: check: Add support for freespace tree fixing Nikolay Borisov
@ 2018-10-04 19:16   ` Omar Sandoval
  0 siblings, 0 replies; 27+ messages in thread
From: Omar Sandoval @ 2018-10-04 19:16 UTC (permalink / raw)
  To: Nikolay Borisov; +Cc: linux-btrfs

On Mon, Oct 01, 2018 at 05:46:19PM +0300, Nikolay Borisov wrote:
> Now that all the prerequisite code for proper support of free space
> tree repair is in, it's time to wire it in. This is achieved by first
> hooking the freespace tree to the __free_extent/alloc_reserved_tree_block
> functions. And then introducing a wrapper function to contains the
> existing check_space_cache and the newly introduced repair code.
> Finally, it's important to note that FST repair code first clears the
> existing FST in case of any problem found and rebuilds it from scratch.

Reviewed-by: Omar Sandoval <osandov@fb.com>

A couple of really trivial nitpicks below that you should feel free to
ignore ;)

> Signed-off-by: Nikolay Borisov <nborisov@suse.com>
> ---
>  check/main.c | 47 ++++++++++++++++++++++++++++++-----------------
>  1 file changed, 30 insertions(+), 17 deletions(-)
> 
> diff --git a/check/main.c b/check/main.c
> index b361cd7e26a0..4daf85aad82c 100644
> --- a/check/main.c
> +++ b/check/main.c
> @@ -5392,14 +5392,6 @@ static int check_space_cache(struct btrfs_root *root)
>  	int ret;
>  	int error = 0;
>  
> -	if (btrfs_super_cache_generation(root->fs_info->super_copy) != -1ULL &&
> -	    btrfs_super_generation(root->fs_info->super_copy) !=
> -	    btrfs_super_cache_generation(root->fs_info->super_copy)) {
> -		printf("cache and super generation don't match, space cache "
> -		       "will be invalidated\n");
> -		return 0;
> -	}
> -
>  	while (1) {
>  		ctx.item_count++;
>  		cache = btrfs_lookup_first_block_group(root->fs_info, start);
> @@ -9417,7 +9409,6 @@ static int do_clear_free_space_cache(struct btrfs_fs_info *fs_info,
>  			ret = 1;
>  			goto close_out;
>  		}
> -		printf("Clearing free space cache\n");

Just out of curiosity, why did you delete this message? The one in the
v2 case is still there.

>  		ret = clear_free_space_cache(fs_info);
>  		if (ret) {
>  			error("failed to clear free space cache");
> @@ -9444,6 +9435,35 @@ static int do_clear_free_space_cache(struct btrfs_fs_info *fs_info,
>  	return ret;
>  }
>  
> +static int validate_free_space_cache(struct btrfs_root *root)

At first glance, I wouldn't know what the difference is between
check_space_cache() and validate_free_space_cache(); they sound like the
same thing. Maybe rename this to check_and_repair_space_cache() or just
fold the rebuild into check_space_cache(), to be more in line with the
other check steps in fsck?

^ permalink raw reply	[flat|nested] 27+ messages in thread

* Re: [PATCH 10/10] btrfs-progs: check: Fix wrong error message in case of corrupted bitmap
  2018-10-01 14:46 ` [PATCH 10/10] btrfs-progs: check: Fix wrong error message in case of corrupted bitmap Nikolay Borisov
@ 2018-10-04 19:18   ` Omar Sandoval
  0 siblings, 0 replies; 27+ messages in thread
From: Omar Sandoval @ 2018-10-04 19:18 UTC (permalink / raw)
  To: Nikolay Borisov; +Cc: linux-btrfs

On Mon, Oct 01, 2018 at 05:46:21PM +0300, Nikolay Borisov wrote:
> Similarly to the fix in e444c7bfa65f ("btrfs-progs: check: Fix wrong
> error message in case of corrupted extent") this commits addresses the
> same problem but for corrupted bitmap objects.

Oops.

Reviewed-by: Omar Sandoval <osandov@fb.com>

> Signed-off-by: Nikolay Borisov <nborisov@suse.com>
> ---
>  free-space-tree.c | 2 +-
>  1 file changed, 1 insertion(+), 1 deletion(-)
> 
> diff --git a/free-space-tree.c b/free-space-tree.c
> index 3b7e8a3fe4f5..690da44a739d 100644
> --- a/free-space-tree.c
> +++ b/free-space-tree.c
> @@ -1302,7 +1302,7 @@ static int load_free_space_bitmaps(struct btrfs_fs_info *fs_info,
>  		if (key.objectid + key.offset > end) {
>  			fprintf(stderr,
>  	"free space bitmap ends at %llu, beyond end of block group %llu-%llu\n",
> -				key.objectid, start, end);
> +				key.objectid + key.offset, start, end);
>  			(*errors)++;
>  			break;
>  		}
> -- 
> 2.7.4
> 

^ permalink raw reply	[flat|nested] 27+ messages in thread

* Re: [PATCH 05/10] btrfs-progs: Pull free space tree related code from kernel
  2018-10-04 18:26   ` Omar Sandoval
  2018-10-04 18:34     ` Nikolay Borisov
@ 2018-10-23 14:05     ` David Sterba
  1 sibling, 0 replies; 27+ messages in thread
From: David Sterba @ 2018-10-23 14:05 UTC (permalink / raw)
  To: Omar Sandoval; +Cc: Nikolay Borisov, linux-btrfs

On Thu, Oct 04, 2018 at 11:26:36AM -0700, Omar Sandoval wrote:
> On Mon, Oct 01, 2018 at 05:46:16PM +0300, Nikolay Borisov wrote:
> > To help implement free space tree checker in user space some kernel
> > function are necessary, namely iterating/deleting/adding freespace
> > items, some internal search functions. Functions to populate a block
> > group based on the extent tree. The code is largely copy/paste from
> > the kernel with locking eliminated (i.e free_space_lock). It supports
> > reading/writing of both bitmap and extent based FST trees.
> 
> For some reason, a lot of this added code uses spaces instead of tabs,
> so I had to fix that in order to compare it to the kernel code (some of
> the functions were reordered, too).

That for noticing it, fixing whitespace damage on 300 lines makes me
really annoyed even if vim is a great help here. Patch applied but I'm
not doing that next time.

^ permalink raw reply	[flat|nested] 27+ messages in thread

* Re: [PATCH 00/10] Freespace tree repair support v2
  2018-10-01 14:46 [PATCH 00/10] Freespace tree repair support v2 Nikolay Borisov
                   ` (9 preceding siblings ...)
  2018-10-01 14:46 ` [PATCH 10/10] btrfs-progs: check: Fix wrong error message in case of corrupted bitmap Nikolay Borisov
@ 2018-10-23 15:00 ` David Sterba
  10 siblings, 0 replies; 27+ messages in thread
From: David Sterba @ 2018-10-23 15:00 UTC (permalink / raw)
  To: Nikolay Borisov; +Cc: linux-btrfs

On Mon, Oct 01, 2018 at 05:46:11PM +0300, Nikolay Borisov wrote:
> Here is the v2 of the freespace tree repair support patches. Version 1 can be 
> found at [0]. For background on the series please refer to the initial cover 
> letter. This time round a number of changes have been incorporated based on 
> feedback from Omar. Namely: 
> 
> 1. Added 2 new patches, patches 3 and 4,  which refactor the userspace
> impelementation of various bit manipulation functions. This was required to
> implement the bitmap conversion code from kernel space. 
> 
> 2. Extended patch 5 to include the bitmap conversion code, now userspace 
> handles the case of fragmented fst. 
> 
> 3. Re-ordered patches 6 and 7 so that we first hook into the extent 
> alloc/dealloc sequence (patch 6) and finally enable support for fst in patch 7.
> 
> 4. Updated the FST test (patch 9) to work with the new bitmap conversion code 
> since it changed the layout of the expected freespace tree. Now the test 
> passes again. Also removed an unused function. 
> 
> 5. Added patch 10 which fixes a similar bug to e444c7bfa65f ("btrfs-progs:
> check: Fix wrong error message in case of corrupted extent")
> 
> Following the updates, I rerun all misc/fsck-tests and no failures were 
> observed. 
> 
> More feedback and suggestions are always welcomed.
> 
> [0] https://lore.kernel.org/linux-btrfs/1529060762-4372-1-git-send-email-nborisov@suse.com/
> 
> 
> Nikolay Borisov (10):
>   btrfs-progs: Add support for freespace tree in btrfs_read_fs_root
>   btrfs-progs: Add extent buffer bitmap manipulation infrastructure
>   btrfs-progs: Replace homegrown bitops related functions with kernel
>     counterparts
>   btrfs-progs: Implement find_*_bit_le operations
>   btrfs-progs: Pull free space tree related code from kernel
>   btrfs-progs: Hook FST code in extent (de)alloc
>   btrfs-progs: Add freespace tree as compat_ro supported feature
>   btrfs-progs: check: Add support for freespace tree fixing
>   btrfs-progs: tests: Test for FST corruption detection/repair
>   btrfs-progs: check: Fix wrong error message in case of corrupted
>     bitmap

As there's been sufficient review, I'm adding it to devel and schedule
it for the next major release. There were some fixups needed but not in
the functionality itself. If you have futher updates, plase send them as
separate patches. Thanks.

^ permalink raw reply	[flat|nested] 27+ messages in thread

end of thread, other threads:[~2018-10-23 15:00 UTC | newest]

Thread overview: 27+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2018-10-01 14:46 [PATCH 00/10] Freespace tree repair support v2 Nikolay Borisov
2018-10-01 14:46 ` [PATCH 01/10] btrfs-progs: Add support for freespace tree in btrfs_read_fs_root Nikolay Borisov
2018-10-02 19:20   ` Omar Sandoval
2018-10-04  1:05   ` Su Yue
2018-10-01 14:46 ` [PATCH 02/10] btrfs-progs: Add extent buffer bitmap manipulation infrastructure Nikolay Borisov
2018-10-02 19:24   ` Omar Sandoval
2018-10-04  1:31   ` Su Yue
2018-10-01 14:46 ` [PATCH 03/10] btrfs-progs: Replace homegrown bitops related functions with kernel counterparts Nikolay Borisov
2018-10-02 23:32   ` Omar Sandoval
2018-10-01 14:46 ` [PATCH 04/10] btrfs-progs: Implement find_*_bit_le operations Nikolay Borisov
2018-10-04 18:08   ` Omar Sandoval
2018-10-04 18:09     ` Nikolay Borisov
2018-10-01 14:46 ` [PATCH 05/10] btrfs-progs: Pull free space tree related code from kernel Nikolay Borisov
2018-10-04 18:26   ` Omar Sandoval
2018-10-04 18:34     ` Nikolay Borisov
2018-10-04 19:01       ` Omar Sandoval
2018-10-23 14:05     ` David Sterba
2018-10-01 14:46 ` [PATCH 06/10] btrfs-progs: Hook FST code in extent (de)alloc Nikolay Borisov
2018-10-01 14:46 ` [PATCH 07/10] btrfs-progs: Add freespace tree as compat_ro supported feature Nikolay Borisov
2018-10-04 18:30   ` Omar Sandoval
2018-10-04 18:36     ` Nikolay Borisov
2018-10-01 14:46 ` [PATCH 08/10] btrfs-progs: check: Add support for freespace tree fixing Nikolay Borisov
2018-10-04 19:16   ` Omar Sandoval
2018-10-01 14:46 ` [PATCH 09/10] btrfs-progs: tests: Test for FST corruption detection/repair Nikolay Borisov
2018-10-01 14:46 ` [PATCH 10/10] btrfs-progs: check: Fix wrong error message in case of corrupted bitmap Nikolay Borisov
2018-10-04 19:18   ` Omar Sandoval
2018-10-23 15:00 ` [PATCH 00/10] Freespace tree repair support v2 David Sterba

This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.