linux-fsdevel.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
* [PATCH 0/4] hfsplus: Replace kmap() with kmap_local_page()
@ 2022-08-09 20:31 Fabio M. De Francesco
  2022-08-09 20:31 ` [PATCH 1/4] hfsplus: Unmap the page in the "fail_page" label Fabio M. De Francesco
                   ` (3 more replies)
  0 siblings, 4 replies; 9+ messages in thread
From: Fabio M. De Francesco @ 2022-08-09 20:31 UTC (permalink / raw)
  To: Matthew Wilcox (Oracle),
	Fabio M. De Francesco, Ira Weiny, Jens Axboe, Andrew Morton,
	Bart Van Assche, Kees Cook, Muchun Song, Viacheslav Dubeyko,
	linux-fsdevel, linux-kernel

kmap() is being deprecated in favor of kmap_local_page().

There are two main problems with kmap(): (1) It comes with an overhead as 
mapping space is restricted and protected by a global lock for 
synchronization and (2) it also requires global TLB invalidation when the 
kmap’s pool wraps and it might block when the mapping space is fully 
utilized until a slot becomes available.

With kmap_local_page() the mappings are per thread, CPU local, can take
page faults, and can be called from any context (including interrupts).
It is faster than kmap() in kernels with HIGHMEM enabled. Furthermore,
the tasks can be preempted and, when they are scheduled to run again, the
kernel virtual addresses are restored and still valid.

Since its use in fs/hfsplus is safe everywhere, it should be preferred.

Therefore, replace kmap() with kmap_local_page() in fs/hfsplus. Where
possible, use the suited standard helpers (memzero_page(), memcpy_page())
instead of open coding kmap_local_page() plus memset() or memcpy().

Fix a bug due to a page being not unmapped if the code jumps to the
"fail_page" label (1/4).

Tested in a QEMU/KVM x86_32 VM, 6GB RAM, booting a kernel with
HIGHMEM64GB enabled.

Fabio M. De Francesco (4):
  hfsplus: Unmap the page in the "fail_page" label
  hfsplus: Convert kmap() to kmap_local_page() in bnode.c
  hfsplus: Convert kmap() to kmap_local_page() in bitmap.c
  hfsplus: Convert kmap() to kmap_local_page() in btree.c

 fs/hfsplus/bitmap.c |  20 ++++-----
 fs/hfsplus/bnode.c  | 105 ++++++++++++++++++++------------------------
 fs/hfsplus/btree.c  |  27 ++++++------
 3 files changed, 72 insertions(+), 80 deletions(-)

-- 
2.37.1


^ permalink raw reply	[flat|nested] 9+ messages in thread

* [PATCH 1/4] hfsplus: Unmap the page in the "fail_page" label
  2022-08-09 20:31 [PATCH 0/4] hfsplus: Replace kmap() with kmap_local_page() Fabio M. De Francesco
@ 2022-08-09 20:31 ` Fabio M. De Francesco
  2022-08-09 22:16   ` Viacheslav Dubeyko
  2022-08-09 20:31 ` [PATCH 2/4] hfsplus: Convert kmap() to kmap_local_page() in bnode.c Fabio M. De Francesco
                   ` (2 subsequent siblings)
  3 siblings, 1 reply; 9+ messages in thread
From: Fabio M. De Francesco @ 2022-08-09 20:31 UTC (permalink / raw)
  To: Matthew Wilcox (Oracle),
	Fabio M. De Francesco, Ira Weiny, Jens Axboe, Andrew Morton,
	Bart Van Assche, Kees Cook, Muchun Song, Viacheslav Dubeyko,
	linux-fsdevel, linux-kernel

Several paths within hfs_btree_open() jump to the "fail_page" label
where put_page() is called while the page is still mapped.

Call kunmap() to unmap the page soon before put_page().

Cc: Viacheslav Dubeyko <slava@dubeyko.com>
Reviewed-by: Ira Weiny <ira.weiny@intel.com>
Signed-off-by: Fabio M. De Francesco <fmdefrancesco@gmail.com>
---
 fs/hfsplus/btree.c | 1 +
 1 file changed, 1 insertion(+)

diff --git a/fs/hfsplus/btree.c b/fs/hfsplus/btree.c
index 66774f4cb4fd..3a917a9a4edd 100644
--- a/fs/hfsplus/btree.c
+++ b/fs/hfsplus/btree.c
@@ -245,6 +245,7 @@ struct hfs_btree *hfs_btree_open(struct super_block *sb, u32 id)
 	return tree;
 
  fail_page:
+	kunmap(page);
 	put_page(page);
  free_inode:
 	tree->inode->i_mapping->a_ops = &hfsplus_aops;
-- 
2.37.1


^ permalink raw reply related	[flat|nested] 9+ messages in thread

* [PATCH 2/4] hfsplus: Convert kmap() to kmap_local_page() in bnode.c
  2022-08-09 20:31 [PATCH 0/4] hfsplus: Replace kmap() with kmap_local_page() Fabio M. De Francesco
  2022-08-09 20:31 ` [PATCH 1/4] hfsplus: Unmap the page in the "fail_page" label Fabio M. De Francesco
@ 2022-08-09 20:31 ` Fabio M. De Francesco
  2022-08-09 22:17   ` Viacheslav Dubeyko
  2022-08-09 20:31 ` [PATCH 3/4] hfsplus: Convert kmap() to kmap_local_page() in bitmap.c Fabio M. De Francesco
  2022-08-09 20:31 ` [PATCH 4/4] hfsplus: Convert kmap() to kmap_local_page() in btree.c Fabio M. De Francesco
  3 siblings, 1 reply; 9+ messages in thread
From: Fabio M. De Francesco @ 2022-08-09 20:31 UTC (permalink / raw)
  To: Matthew Wilcox (Oracle),
	Fabio M. De Francesco, Ira Weiny, Jens Axboe, Andrew Morton,
	Bart Van Assche, Kees Cook, Muchun Song, Viacheslav Dubeyko,
	linux-fsdevel, linux-kernel

kmap() is being deprecated in favor of kmap_local_page().

Two main problems with kmap(): (1) It comes with an overhead as mapping
space is restricted and protected by a global lock for synchronization and
(2) it also requires global TLB invalidation when the kmap’s pool wraps
and it might block when the mapping space is fully utilized until a slot
becomes available.

With kmap_local_page() the mappings are per thread, CPU local, can take
page faults, and can be called from any context (including interrupts).
It is faster than kmap() in kernels with HIGHMEM enabled. Furthermore,
the tasks can be preempted and, when they are scheduled to run again, the
kernel virtual addresses are restored and still valid.

Since its use in bnode.c is safe everywhere, it should be preferred.

Therefore, replace kmap() with kmap_local_page() in bnode.c. Where
possible, use the suited standard helpers (memzero_page(), memcpy_page())
instead of open coding kmap_local_page() plus memset() or memcpy().

Tested in a QEMU/KVM x86_32 VM, 6GB RAM, booting a kernel with
HIGHMEM64GB enabled.

Cc: Viacheslav Dubeyko <slava@dubeyko.com>
Suggested-by: Ira Weiny <ira.weiny@intel.com>
Reviewed-by: Ira Weiny <ira.weiny@intel.com>
Signed-off-by: Fabio M. De Francesco <fmdefrancesco@gmail.com>
---
 fs/hfsplus/bnode.c | 105 +++++++++++++++++++++------------------------
 1 file changed, 48 insertions(+), 57 deletions(-)

diff --git a/fs/hfsplus/bnode.c b/fs/hfsplus/bnode.c
index a5ab00e54220..87974d5e6791 100644
--- a/fs/hfsplus/bnode.c
+++ b/fs/hfsplus/bnode.c
@@ -29,14 +29,12 @@ void hfs_bnode_read(struct hfs_bnode *node, void *buf, int off, int len)
 	off &= ~PAGE_MASK;
 
 	l = min_t(int, len, PAGE_SIZE - off);
-	memcpy(buf, kmap(*pagep) + off, l);
-	kunmap(*pagep);
+	memcpy_from_page(buf, *pagep, off, l);
 
 	while ((len -= l) != 0) {
 		buf += l;
 		l = min_t(int, len, PAGE_SIZE);
-		memcpy(buf, kmap(*++pagep), l);
-		kunmap(*pagep);
+		memcpy_from_page(buf, *++pagep, 0, l);
 	}
 }
 
@@ -82,16 +80,14 @@ void hfs_bnode_write(struct hfs_bnode *node, void *buf, int off, int len)
 	off &= ~PAGE_MASK;
 
 	l = min_t(int, len, PAGE_SIZE - off);
-	memcpy(kmap(*pagep) + off, buf, l);
+	memcpy_to_page(*pagep, off, buf, l);
 	set_page_dirty(*pagep);
-	kunmap(*pagep);
 
 	while ((len -= l) != 0) {
 		buf += l;
 		l = min_t(int, len, PAGE_SIZE);
-		memcpy(kmap(*++pagep), buf, l);
+		memcpy_to_page(*++pagep, 0, buf, l);
 		set_page_dirty(*pagep);
-		kunmap(*pagep);
 	}
 }
 
@@ -112,15 +108,13 @@ void hfs_bnode_clear(struct hfs_bnode *node, int off, int len)
 	off &= ~PAGE_MASK;
 
 	l = min_t(int, len, PAGE_SIZE - off);
-	memset(kmap(*pagep) + off, 0, l);
+	memzero_page(*pagep, off, l);
 	set_page_dirty(*pagep);
-	kunmap(*pagep);
 
 	while ((len -= l) != 0) {
 		l = min_t(int, len, PAGE_SIZE);
-		memset(kmap(*++pagep), 0, l);
+		memzero_page(*++pagep, 0, l);
 		set_page_dirty(*pagep);
-		kunmap(*pagep);
 	}
 }
 
@@ -142,24 +136,20 @@ void hfs_bnode_copy(struct hfs_bnode *dst_node, int dst,
 
 	if (src == dst) {
 		l = min_t(int, len, PAGE_SIZE - src);
-		memcpy(kmap(*dst_page) + src, kmap(*src_page) + src, l);
-		kunmap(*src_page);
+		memcpy_page(*dst_page, src, *src_page, src, l);
 		set_page_dirty(*dst_page);
-		kunmap(*dst_page);
 
 		while ((len -= l) != 0) {
 			l = min_t(int, len, PAGE_SIZE);
-			memcpy(kmap(*++dst_page), kmap(*++src_page), l);
-			kunmap(*src_page);
+			memcpy_page(*++dst_page, 0, *++src_page, 0, l);
 			set_page_dirty(*dst_page);
-			kunmap(*dst_page);
 		}
 	} else {
 		void *src_ptr, *dst_ptr;
 
 		do {
-			src_ptr = kmap(*src_page) + src;
-			dst_ptr = kmap(*dst_page) + dst;
+			dst_ptr = kmap_local_page(*dst_page) + dst;
+			src_ptr = kmap_local_page(*src_page) + src;
 			if (PAGE_SIZE - src < PAGE_SIZE - dst) {
 				l = PAGE_SIZE - src;
 				src = 0;
@@ -171,9 +161,9 @@ void hfs_bnode_copy(struct hfs_bnode *dst_node, int dst,
 			}
 			l = min(len, l);
 			memcpy(dst_ptr, src_ptr, l);
-			kunmap(*src_page);
+			kunmap_local(src_ptr);
 			set_page_dirty(*dst_page);
-			kunmap(*dst_page);
+			kunmap_local(dst_ptr);
 			if (!dst)
 				dst_page++;
 			else
@@ -185,6 +175,7 @@ void hfs_bnode_copy(struct hfs_bnode *dst_node, int dst,
 void hfs_bnode_move(struct hfs_bnode *node, int dst, int src, int len)
 {
 	struct page **src_page, **dst_page;
+	void *src_ptr, *dst_ptr;
 	int l;
 
 	hfs_dbg(BNODE_MOD, "movebytes: %u,%u,%u\n", dst, src, len);
@@ -202,27 +193,28 @@ void hfs_bnode_move(struct hfs_bnode *node, int dst, int src, int len)
 
 		if (src == dst) {
 			while (src < len) {
-				memmove(kmap(*dst_page), kmap(*src_page), src);
-				kunmap(*src_page);
+				dst_ptr = kmap_local_page(*dst_page);
+				src_ptr = kmap_local_page(*src_page);
+				memmove(dst_ptr, src_ptr, src);
+				kunmap_local(src_ptr);
 				set_page_dirty(*dst_page);
-				kunmap(*dst_page);
+				kunmap_local(dst_ptr);
 				len -= src;
 				src = PAGE_SIZE;
 				src_page--;
 				dst_page--;
 			}
 			src -= len;
-			memmove(kmap(*dst_page) + src,
-				kmap(*src_page) + src, len);
-			kunmap(*src_page);
+			dst_ptr = kmap_local_page(*dst_page);
+			src_ptr = kmap_local_page(*src_page);
+			memmove(dst_ptr + src, src_ptr + src, len);
+			kunmap_local(src_ptr);
 			set_page_dirty(*dst_page);
-			kunmap(*dst_page);
+			kunmap_local(dst_ptr);
 		} else {
-			void *src_ptr, *dst_ptr;
-
 			do {
-				src_ptr = kmap(*src_page) + src;
-				dst_ptr = kmap(*dst_page) + dst;
+				dst_ptr = kmap_local_page(*dst_page) + dst;
+				src_ptr = kmap_local_page(*src_page) + src;
 				if (src < dst) {
 					l = src;
 					src = PAGE_SIZE;
@@ -234,9 +226,9 @@ void hfs_bnode_move(struct hfs_bnode *node, int dst, int src, int len)
 				}
 				l = min(len, l);
 				memmove(dst_ptr - l, src_ptr - l, l);
-				kunmap(*src_page);
+				kunmap_local(src_ptr);
 				set_page_dirty(*dst_page);
-				kunmap(*dst_page);
+				kunmap_local(dst_ptr);
 				if (dst == PAGE_SIZE)
 					dst_page--;
 				else
@@ -251,26 +243,27 @@ void hfs_bnode_move(struct hfs_bnode *node, int dst, int src, int len)
 
 		if (src == dst) {
 			l = min_t(int, len, PAGE_SIZE - src);
-			memmove(kmap(*dst_page) + src,
-				kmap(*src_page) + src, l);
-			kunmap(*src_page);
+
+			dst_ptr = kmap_local_page(*dst_page) + src;
+			src_ptr = kmap_local_page(*src_page) + src;
+			memmove(dst_ptr, src_ptr, l);
+			kunmap_local(src_ptr);
 			set_page_dirty(*dst_page);
-			kunmap(*dst_page);
+			kunmap_local(dst_ptr);
 
 			while ((len -= l) != 0) {
 				l = min_t(int, len, PAGE_SIZE);
-				memmove(kmap(*++dst_page),
-					kmap(*++src_page), l);
-				kunmap(*src_page);
+				dst_ptr = kmap_local_page(*++dst_page);
+				src_ptr = kmap_local_page(*++src_page);
+				memmove(dst_ptr, src_ptr, l);
+				kunmap_local(src_ptr);
 				set_page_dirty(*dst_page);
-				kunmap(*dst_page);
+				kunmap_local(dst_ptr);
 			}
 		} else {
-			void *src_ptr, *dst_ptr;
-
 			do {
-				src_ptr = kmap(*src_page) + src;
-				dst_ptr = kmap(*dst_page) + dst;
+				dst_ptr = kmap_local_page(*dst_page) + dst;
+				src_ptr = kmap_local_page(*src_page) + src;
 				if (PAGE_SIZE - src <
 						PAGE_SIZE - dst) {
 					l = PAGE_SIZE - src;
@@ -283,9 +276,9 @@ void hfs_bnode_move(struct hfs_bnode *node, int dst, int src, int len)
 				}
 				l = min(len, l);
 				memmove(dst_ptr, src_ptr, l);
-				kunmap(*src_page);
+				kunmap_local(src_ptr);
 				set_page_dirty(*dst_page);
-				kunmap(*dst_page);
+				kunmap_local(dst_ptr);
 				if (!dst)
 					dst_page++;
 				else
@@ -498,14 +491,14 @@ struct hfs_bnode *hfs_bnode_find(struct hfs_btree *tree, u32 num)
 	if (!test_bit(HFS_BNODE_NEW, &node->flags))
 		return node;
 
-	desc = (struct hfs_bnode_desc *)(kmap(node->page[0]) +
-			node->page_offset);
+	desc = (struct hfs_bnode_desc *)(kmap_local_page(node->page[0]) +
+							 node->page_offset);
 	node->prev = be32_to_cpu(desc->prev);
 	node->next = be32_to_cpu(desc->next);
 	node->num_recs = be16_to_cpu(desc->num_recs);
 	node->type = desc->type;
 	node->height = desc->height;
-	kunmap(node->page[0]);
+	kunmap_local(desc);
 
 	switch (node->type) {
 	case HFS_NODE_HEADER:
@@ -589,14 +582,12 @@ struct hfs_bnode *hfs_bnode_create(struct hfs_btree *tree, u32 num)
 	}
 
 	pagep = node->page;
-	memset(kmap(*pagep) + node->page_offset, 0,
-	       min_t(int, PAGE_SIZE, tree->node_size));
+	memzero_page(*pagep, node->page_offset,
+		     min_t(int, PAGE_SIZE, tree->node_size));
 	set_page_dirty(*pagep);
-	kunmap(*pagep);
 	for (i = 1; i < tree->pages_per_bnode; i++) {
-		memset(kmap(*++pagep), 0, PAGE_SIZE);
+		memzero_page(*++pagep, 0, PAGE_SIZE);
 		set_page_dirty(*pagep);
-		kunmap(*pagep);
 	}
 	clear_bit(HFS_BNODE_NEW, &node->flags);
 	wake_up(&node->lock_wq);
-- 
2.37.1


^ permalink raw reply related	[flat|nested] 9+ messages in thread

* [PATCH 3/4] hfsplus: Convert kmap() to kmap_local_page() in bitmap.c
  2022-08-09 20:31 [PATCH 0/4] hfsplus: Replace kmap() with kmap_local_page() Fabio M. De Francesco
  2022-08-09 20:31 ` [PATCH 1/4] hfsplus: Unmap the page in the "fail_page" label Fabio M. De Francesco
  2022-08-09 20:31 ` [PATCH 2/4] hfsplus: Convert kmap() to kmap_local_page() in bnode.c Fabio M. De Francesco
@ 2022-08-09 20:31 ` Fabio M. De Francesco
  2022-08-09 22:18   ` Viacheslav Dubeyko
  2022-08-09 20:31 ` [PATCH 4/4] hfsplus: Convert kmap() to kmap_local_page() in btree.c Fabio M. De Francesco
  3 siblings, 1 reply; 9+ messages in thread
From: Fabio M. De Francesco @ 2022-08-09 20:31 UTC (permalink / raw)
  To: Matthew Wilcox (Oracle),
	Fabio M. De Francesco, Ira Weiny, Jens Axboe, Andrew Morton,
	Bart Van Assche, Kees Cook, Muchun Song, Viacheslav Dubeyko,
	linux-fsdevel, linux-kernel

kmap() is being deprecated in favor of kmap_local_page().

There are two main problems with kmap(): (1) It comes with an overhead as
mapping space is restricted and protected by a global lock for
synchronization and (2) it also requires global TLB invalidation when the
kmap’s pool wraps and it might block when the mapping space is fully
utilized until a slot becomes available.

With kmap_local_page() the mappings are per thread, CPU local, can take
page faults, and can be called from any context (including interrupts).
It is faster than kmap() in kernels with HIGHMEM enabled. Furthermore,
the tasks can be preempted and, when they are scheduled to run again, the
kernel virtual addresses are restored and are still valid.

Since its use in bitmap.c is safe everywhere, it should be preferred.

Therefore, replace kmap() with kmap_local_page() in bitmap.c.

Tested in a QEMU/KVM x86_32 VM, 6GB RAM, booting a kernel with
HIGHMEM64GB enabled.

Cc: Viacheslav Dubeyko <slava@dubeyko.com>
Suggested-by: Ira Weiny <ira.weiny@intel.com>
Reviewed-by: Ira Weiny <ira.weiny@intel.com>
Signed-off-by: Fabio M. De Francesco <fmdefrancesco@gmail.com>
---
 fs/hfsplus/bitmap.c | 20 ++++++++++----------
 1 file changed, 10 insertions(+), 10 deletions(-)

diff --git a/fs/hfsplus/bitmap.c b/fs/hfsplus/bitmap.c
index cebce0cfe340..bd8dcea85588 100644
--- a/fs/hfsplus/bitmap.c
+++ b/fs/hfsplus/bitmap.c
@@ -39,7 +39,7 @@ int hfsplus_block_allocate(struct super_block *sb, u32 size,
 		start = size;
 		goto out;
 	}
-	pptr = kmap(page);
+	pptr = kmap_local_page(page);
 	curr = pptr + (offset & (PAGE_CACHE_BITS - 1)) / 32;
 	i = offset % 32;
 	offset &= ~(PAGE_CACHE_BITS - 1);
@@ -74,7 +74,7 @@ int hfsplus_block_allocate(struct super_block *sb, u32 size,
 			}
 			curr++;
 		}
-		kunmap(page);
+		kunmap_local(pptr);
 		offset += PAGE_CACHE_BITS;
 		if (offset >= size)
 			break;
@@ -84,7 +84,7 @@ int hfsplus_block_allocate(struct super_block *sb, u32 size,
 			start = size;
 			goto out;
 		}
-		curr = pptr = kmap(page);
+		curr = pptr = kmap_local_page(page);
 		if ((size ^ offset) / PAGE_CACHE_BITS)
 			end = pptr + PAGE_CACHE_BITS / 32;
 		else
@@ -127,7 +127,7 @@ int hfsplus_block_allocate(struct super_block *sb, u32 size,
 			len -= 32;
 		}
 		set_page_dirty(page);
-		kunmap(page);
+		kunmap_local(pptr);
 		offset += PAGE_CACHE_BITS;
 		page = read_mapping_page(mapping, offset / PAGE_CACHE_BITS,
 					 NULL);
@@ -135,7 +135,7 @@ int hfsplus_block_allocate(struct super_block *sb, u32 size,
 			start = size;
 			goto out;
 		}
-		pptr = kmap(page);
+		pptr = kmap_local_page(page);
 		curr = pptr;
 		end = pptr + PAGE_CACHE_BITS / 32;
 	}
@@ -151,7 +151,7 @@ int hfsplus_block_allocate(struct super_block *sb, u32 size,
 done:
 	*curr = cpu_to_be32(n);
 	set_page_dirty(page);
-	kunmap(page);
+	kunmap_local(pptr);
 	*max = offset + (curr - pptr) * 32 + i - start;
 	sbi->free_blocks -= *max;
 	hfsplus_mark_mdb_dirty(sb);
@@ -185,7 +185,7 @@ int hfsplus_block_free(struct super_block *sb, u32 offset, u32 count)
 	page = read_mapping_page(mapping, pnr, NULL);
 	if (IS_ERR(page))
 		goto kaboom;
-	pptr = kmap(page);
+	pptr = kmap_local_page(page);
 	curr = pptr + (offset & (PAGE_CACHE_BITS - 1)) / 32;
 	end = pptr + PAGE_CACHE_BITS / 32;
 	len = count;
@@ -215,11 +215,11 @@ int hfsplus_block_free(struct super_block *sb, u32 offset, u32 count)
 		if (!count)
 			break;
 		set_page_dirty(page);
-		kunmap(page);
+		kunmap_local(pptr);
 		page = read_mapping_page(mapping, ++pnr, NULL);
 		if (IS_ERR(page))
 			goto kaboom;
-		pptr = kmap(page);
+		pptr = kmap_local_page(page);
 		curr = pptr;
 		end = pptr + PAGE_CACHE_BITS / 32;
 	}
@@ -231,7 +231,7 @@ int hfsplus_block_free(struct super_block *sb, u32 offset, u32 count)
 	}
 out:
 	set_page_dirty(page);
-	kunmap(page);
+	kunmap_local(pptr);
 	sbi->free_blocks += len;
 	hfsplus_mark_mdb_dirty(sb);
 	mutex_unlock(&sbi->alloc_mutex);
-- 
2.37.1


^ permalink raw reply related	[flat|nested] 9+ messages in thread

* [PATCH 4/4] hfsplus: Convert kmap() to kmap_local_page() in btree.c
  2022-08-09 20:31 [PATCH 0/4] hfsplus: Replace kmap() with kmap_local_page() Fabio M. De Francesco
                   ` (2 preceding siblings ...)
  2022-08-09 20:31 ` [PATCH 3/4] hfsplus: Convert kmap() to kmap_local_page() in bitmap.c Fabio M. De Francesco
@ 2022-08-09 20:31 ` Fabio M. De Francesco
  2022-08-09 22:18   ` Viacheslav Dubeyko
  3 siblings, 1 reply; 9+ messages in thread
From: Fabio M. De Francesco @ 2022-08-09 20:31 UTC (permalink / raw)
  To: Matthew Wilcox (Oracle),
	Fabio M. De Francesco, Ira Weiny, Jens Axboe, Andrew Morton,
	Bart Van Assche, Kees Cook, Muchun Song, Viacheslav Dubeyko,
	linux-fsdevel, linux-kernel

kmap() is being deprecated in favor of kmap_local_page().

There are two main problems with kmap(): (1) It comes with an overhead as
mapping space is restricted and protected by a global lock for
synchronization and (2) it also requires global TLB invalidation when the
kmap’s pool wraps and it might block when the mapping space is fully
utilized until a slot becomes available.

With kmap_local_page() the mappings are per thread, CPU local, can take
page faults, and can be called from any context (including interrupts).
It is faster than kmap() in kernels with HIGHMEM enabled. Furthermore,
the tasks can be preempted and, when they are scheduled to run again, the
kernel virtual addresses are restored and are still valid.

Since its use in btree.c is safe everywhere, it should be preferred.

Therefore, replace kmap() with kmap_local_page() in btree.c.

Tested in a QEMU/KVM x86_32 VM, 6GB RAM, booting a kernel with
HIGHMEM64GB enabled.

Cc: Viacheslav Dubeyko <slava@dubeyko.com>
Suggested-by: Ira Weiny <ira.weiny@intel.com>
Reviewed-by: Ira Weiny <ira.weiny@intel.com>
Signed-off-by: Fabio M. De Francesco <fmdefrancesco@gmail.com>
---
 fs/hfsplus/btree.c | 28 ++++++++++++++--------------
 1 file changed, 14 insertions(+), 14 deletions(-)

diff --git a/fs/hfsplus/btree.c b/fs/hfsplus/btree.c
index 3a917a9a4edd..9e1732a2b92a 100644
--- a/fs/hfsplus/btree.c
+++ b/fs/hfsplus/btree.c
@@ -163,7 +163,7 @@ struct hfs_btree *hfs_btree_open(struct super_block *sb, u32 id)
 		goto free_inode;
 
 	/* Load the header */
-	head = (struct hfs_btree_header_rec *)(kmap(page) +
+	head = (struct hfs_btree_header_rec *)(kmap_local_page(page) +
 		sizeof(struct hfs_bnode_desc));
 	tree->root = be32_to_cpu(head->root);
 	tree->leaf_count = be32_to_cpu(head->leaf_count);
@@ -240,12 +240,12 @@ struct hfs_btree *hfs_btree_open(struct super_block *sb, u32 id)
 		(tree->node_size + PAGE_SIZE - 1) >>
 		PAGE_SHIFT;
 
-	kunmap(page);
+	kunmap_local(head);
 	put_page(page);
 	return tree;
 
  fail_page:
-	kunmap(page);
+	kunmap_local(head);
 	put_page(page);
  free_inode:
 	tree->inode->i_mapping->a_ops = &hfsplus_aops;
@@ -292,7 +292,7 @@ int hfs_btree_write(struct hfs_btree *tree)
 		return -EIO;
 	/* Load the header */
 	page = node->page[0];
-	head = (struct hfs_btree_header_rec *)(kmap(page) +
+	head = (struct hfs_btree_header_rec *)(kmap_local_page(page) +
 		sizeof(struct hfs_bnode_desc));
 
 	head->root = cpu_to_be32(tree->root);
@@ -304,7 +304,7 @@ int hfs_btree_write(struct hfs_btree *tree)
 	head->attributes = cpu_to_be32(tree->attributes);
 	head->depth = cpu_to_be16(tree->depth);
 
-	kunmap(page);
+	kunmap_local(head);
 	set_page_dirty(page);
 	hfs_bnode_put(node);
 	return 0;
@@ -395,7 +395,7 @@ struct hfs_bnode *hfs_bmap_alloc(struct hfs_btree *tree)
 
 	off += node->page_offset;
 	pagep = node->page + (off >> PAGE_SHIFT);
-	data = kmap(*pagep);
+	data = kmap_local_page(*pagep);
 	off &= ~PAGE_MASK;
 	idx = 0;
 
@@ -408,7 +408,7 @@ struct hfs_bnode *hfs_bmap_alloc(struct hfs_btree *tree)
 						idx += i;
 						data[off] |= m;
 						set_page_dirty(*pagep);
-						kunmap(*pagep);
+						kunmap_local(data);
 						tree->free_nodes--;
 						mark_inode_dirty(tree->inode);
 						hfs_bnode_put(node);
@@ -418,14 +418,14 @@ struct hfs_bnode *hfs_bmap_alloc(struct hfs_btree *tree)
 				}
 			}
 			if (++off >= PAGE_SIZE) {
-				kunmap(*pagep);
-				data = kmap(*++pagep);
+				kunmap_local(data);
+				data = kmap_local_page(*++pagep);
 				off = 0;
 			}
 			idx += 8;
 			len--;
 		}
-		kunmap(*pagep);
+		kunmap_local(data);
 		nidx = node->next;
 		if (!nidx) {
 			hfs_dbg(BNODE_MOD, "create new bmap node\n");
@@ -441,7 +441,7 @@ struct hfs_bnode *hfs_bmap_alloc(struct hfs_btree *tree)
 		off = off16;
 		off += node->page_offset;
 		pagep = node->page + (off >> PAGE_SHIFT);
-		data = kmap(*pagep);
+		data = kmap_local_page(*pagep);
 		off &= ~PAGE_MASK;
 	}
 }
@@ -491,7 +491,7 @@ void hfs_bmap_free(struct hfs_bnode *node)
 	}
 	off += node->page_offset + nidx / 8;
 	page = node->page[off >> PAGE_SHIFT];
-	data = kmap(page);
+	data = kmap_local_page(page);
 	off &= ~PAGE_MASK;
 	m = 1 << (~nidx & 7);
 	byte = data[off];
@@ -499,13 +499,13 @@ void hfs_bmap_free(struct hfs_bnode *node)
 		pr_crit("trying to free free bnode "
 				"%u(%d)\n",
 			node->this, node->type);
-		kunmap(page);
+		kunmap_local(data);
 		hfs_bnode_put(node);
 		return;
 	}
 	data[off] = byte & ~m;
 	set_page_dirty(page);
-	kunmap(page);
+	kunmap_local(data);
 	hfs_bnode_put(node);
 	tree->free_nodes++;
 	mark_inode_dirty(tree->inode);
-- 
2.37.1


^ permalink raw reply related	[flat|nested] 9+ messages in thread

* Re: [PATCH 1/4] hfsplus: Unmap the page in the "fail_page" label
  2022-08-09 20:31 ` [PATCH 1/4] hfsplus: Unmap the page in the "fail_page" label Fabio M. De Francesco
@ 2022-08-09 22:16   ` Viacheslav Dubeyko
  0 siblings, 0 replies; 9+ messages in thread
From: Viacheslav Dubeyko @ 2022-08-09 22:16 UTC (permalink / raw)
  To: Fabio M. De Francesco
  Cc: Matthew Wilcox (Oracle),
	Ira Weiny, Jens Axboe, Andrew Morton, Bart Van Assche, Kees Cook,
	Muchun Song, linux-fsdevel, linux-kernel



> On Aug 9, 2022, at 1:31 PM, Fabio M. De Francesco <fmdefrancesco@gmail.com> wrote:
> 
> Several paths within hfs_btree_open() jump to the "fail_page" label
> where put_page() is called while the page is still mapped.
> 
> Call kunmap() to unmap the page soon before put_page().
> 
> Cc: Viacheslav Dubeyko <slava@dubeyko.com>
> Reviewed-by: Ira Weiny <ira.weiny@intel.com>
> Signed-off-by: Fabio M. De Francesco <fmdefrancesco@gmail.com>
> ---
> fs/hfsplus/btree.c | 1 +
> 1 file changed, 1 insertion(+)
> 
> diff --git a/fs/hfsplus/btree.c b/fs/hfsplus/btree.c
> index 66774f4cb4fd..3a917a9a4edd 100644
> --- a/fs/hfsplus/btree.c
> +++ b/fs/hfsplus/btree.c
> @@ -245,6 +245,7 @@ struct hfs_btree *hfs_btree_open(struct super_block *sb, u32 id)
> 	return tree;
> 
>  fail_page:
> +	kunmap(page);
> 	put_page(page);
>  free_inode:
> 	tree->inode->i_mapping->a_ops = &hfsplus_aops;
> -- 
> 2.37.1
> 

Looks good.

Reviewed-by: Viacheslav Dubeyko <slava@dubeyko.com>

Thanks,
Slava.



^ permalink raw reply	[flat|nested] 9+ messages in thread

* Re: [PATCH 2/4] hfsplus: Convert kmap() to kmap_local_page() in bnode.c
  2022-08-09 20:31 ` [PATCH 2/4] hfsplus: Convert kmap() to kmap_local_page() in bnode.c Fabio M. De Francesco
@ 2022-08-09 22:17   ` Viacheslav Dubeyko
  0 siblings, 0 replies; 9+ messages in thread
From: Viacheslav Dubeyko @ 2022-08-09 22:17 UTC (permalink / raw)
  To: Fabio M. De Francesco
  Cc: Matthew Wilcox (Oracle),
	Ira Weiny, Jens Axboe, Andrew Morton, Bart Van Assche, Kees Cook,
	Muchun Song, Linux FS Devel, LKML



> On Aug 9, 2022, at 1:31 PM, Fabio M. De Francesco <fmdefrancesco@gmail.com> wrote:
> 
> kmap() is being deprecated in favor of kmap_local_page().
> 
> Two main problems with kmap(): (1) It comes with an overhead as mapping
> space is restricted and protected by a global lock for synchronization and
> (2) it also requires global TLB invalidation when the kmap’s pool wraps
> and it might block when the mapping space is fully utilized until a slot
> becomes available.
> 
> With kmap_local_page() the mappings are per thread, CPU local, can take
> page faults, and can be called from any context (including interrupts).
> It is faster than kmap() in kernels with HIGHMEM enabled. Furthermore,
> the tasks can be preempted and, when they are scheduled to run again, the
> kernel virtual addresses are restored and still valid.
> 
> Since its use in bnode.c is safe everywhere, it should be preferred.
> 
> Therefore, replace kmap() with kmap_local_page() in bnode.c. Where
> possible, use the suited standard helpers (memzero_page(), memcpy_page())
> instead of open coding kmap_local_page() plus memset() or memcpy().
> 
> Tested in a QEMU/KVM x86_32 VM, 6GB RAM, booting a kernel with
> HIGHMEM64GB enabled.
> 
> Cc: Viacheslav Dubeyko <slava@dubeyko.com>
> Suggested-by: Ira Weiny <ira.weiny@intel.com>
> Reviewed-by: Ira Weiny <ira.weiny@intel.com>
> Signed-off-by: Fabio M. De Francesco <fmdefrancesco@gmail.com>
> ---


Looks good.

Reviewed-by: Viacheslav Dubeyko <slava@dubeyko.com>

Thanks,
Slava.


> fs/hfsplus/bnode.c | 105 +++++++++++++++++++++------------------------
> 1 file changed, 48 insertions(+), 57 deletions(-)
> 
> diff --git a/fs/hfsplus/bnode.c b/fs/hfsplus/bnode.c
> index a5ab00e54220..87974d5e6791 100644
> --- a/fs/hfsplus/bnode.c
> +++ b/fs/hfsplus/bnode.c
> @@ -29,14 +29,12 @@ void hfs_bnode_read(struct hfs_bnode *node, void *buf, int off, int len)
> 	off &= ~PAGE_MASK;
> 
> 	l = min_t(int, len, PAGE_SIZE - off);
> -	memcpy(buf, kmap(*pagep) + off, l);
> -	kunmap(*pagep);
> +	memcpy_from_page(buf, *pagep, off, l);
> 
> 	while ((len -= l) != 0) {
> 		buf += l;
> 		l = min_t(int, len, PAGE_SIZE);
> -		memcpy(buf, kmap(*++pagep), l);
> -		kunmap(*pagep);
> +		memcpy_from_page(buf, *++pagep, 0, l);
> 	}
> }
> 
> @@ -82,16 +80,14 @@ void hfs_bnode_write(struct hfs_bnode *node, void *buf, int off, int len)
> 	off &= ~PAGE_MASK;
> 
> 	l = min_t(int, len, PAGE_SIZE - off);
> -	memcpy(kmap(*pagep) + off, buf, l);
> +	memcpy_to_page(*pagep, off, buf, l);
> 	set_page_dirty(*pagep);
> -	kunmap(*pagep);
> 
> 	while ((len -= l) != 0) {
> 		buf += l;
> 		l = min_t(int, len, PAGE_SIZE);
> -		memcpy(kmap(*++pagep), buf, l);
> +		memcpy_to_page(*++pagep, 0, buf, l);
> 		set_page_dirty(*pagep);
> -		kunmap(*pagep);
> 	}
> }
> 
> @@ -112,15 +108,13 @@ void hfs_bnode_clear(struct hfs_bnode *node, int off, int len)
> 	off &= ~PAGE_MASK;
> 
> 	l = min_t(int, len, PAGE_SIZE - off);
> -	memset(kmap(*pagep) + off, 0, l);
> +	memzero_page(*pagep, off, l);
> 	set_page_dirty(*pagep);
> -	kunmap(*pagep);
> 
> 	while ((len -= l) != 0) {
> 		l = min_t(int, len, PAGE_SIZE);
> -		memset(kmap(*++pagep), 0, l);
> +		memzero_page(*++pagep, 0, l);
> 		set_page_dirty(*pagep);
> -		kunmap(*pagep);
> 	}
> }
> 
> @@ -142,24 +136,20 @@ void hfs_bnode_copy(struct hfs_bnode *dst_node, int dst,
> 
> 	if (src == dst) {
> 		l = min_t(int, len, PAGE_SIZE - src);
> -		memcpy(kmap(*dst_page) + src, kmap(*src_page) + src, l);
> -		kunmap(*src_page);
> +		memcpy_page(*dst_page, src, *src_page, src, l);
> 		set_page_dirty(*dst_page);
> -		kunmap(*dst_page);
> 
> 		while ((len -= l) != 0) {
> 			l = min_t(int, len, PAGE_SIZE);
> -			memcpy(kmap(*++dst_page), kmap(*++src_page), l);
> -			kunmap(*src_page);
> +			memcpy_page(*++dst_page, 0, *++src_page, 0, l);
> 			set_page_dirty(*dst_page);
> -			kunmap(*dst_page);
> 		}
> 	} else {
> 		void *src_ptr, *dst_ptr;
> 
> 		do {
> -			src_ptr = kmap(*src_page) + src;
> -			dst_ptr = kmap(*dst_page) + dst;
> +			dst_ptr = kmap_local_page(*dst_page) + dst;
> +			src_ptr = kmap_local_page(*src_page) + src;
> 			if (PAGE_SIZE - src < PAGE_SIZE - dst) {
> 				l = PAGE_SIZE - src;
> 				src = 0;
> @@ -171,9 +161,9 @@ void hfs_bnode_copy(struct hfs_bnode *dst_node, int dst,
> 			}
> 			l = min(len, l);
> 			memcpy(dst_ptr, src_ptr, l);
> -			kunmap(*src_page);
> +			kunmap_local(src_ptr);
> 			set_page_dirty(*dst_page);
> -			kunmap(*dst_page);
> +			kunmap_local(dst_ptr);
> 			if (!dst)
> 				dst_page++;
> 			else
> @@ -185,6 +175,7 @@ void hfs_bnode_copy(struct hfs_bnode *dst_node, int dst,
> void hfs_bnode_move(struct hfs_bnode *node, int dst, int src, int len)
> {
> 	struct page **src_page, **dst_page;
> +	void *src_ptr, *dst_ptr;
> 	int l;
> 
> 	hfs_dbg(BNODE_MOD, "movebytes: %u,%u,%u\n", dst, src, len);
> @@ -202,27 +193,28 @@ void hfs_bnode_move(struct hfs_bnode *node, int dst, int src, int len)
> 
> 		if (src == dst) {
> 			while (src < len) {
> -				memmove(kmap(*dst_page), kmap(*src_page), src);
> -				kunmap(*src_page);
> +				dst_ptr = kmap_local_page(*dst_page);
> +				src_ptr = kmap_local_page(*src_page);
> +				memmove(dst_ptr, src_ptr, src);
> +				kunmap_local(src_ptr);
> 				set_page_dirty(*dst_page);
> -				kunmap(*dst_page);
> +				kunmap_local(dst_ptr);
> 				len -= src;
> 				src = PAGE_SIZE;
> 				src_page--;
> 				dst_page--;
> 			}
> 			src -= len;
> -			memmove(kmap(*dst_page) + src,
> -				kmap(*src_page) + src, len);
> -			kunmap(*src_page);
> +			dst_ptr = kmap_local_page(*dst_page);
> +			src_ptr = kmap_local_page(*src_page);
> +			memmove(dst_ptr + src, src_ptr + src, len);
> +			kunmap_local(src_ptr);
> 			set_page_dirty(*dst_page);
> -			kunmap(*dst_page);
> +			kunmap_local(dst_ptr);
> 		} else {
> -			void *src_ptr, *dst_ptr;
> -
> 			do {
> -				src_ptr = kmap(*src_page) + src;
> -				dst_ptr = kmap(*dst_page) + dst;
> +				dst_ptr = kmap_local_page(*dst_page) + dst;
> +				src_ptr = kmap_local_page(*src_page) + src;
> 				if (src < dst) {
> 					l = src;
> 					src = PAGE_SIZE;
> @@ -234,9 +226,9 @@ void hfs_bnode_move(struct hfs_bnode *node, int dst, int src, int len)
> 				}
> 				l = min(len, l);
> 				memmove(dst_ptr - l, src_ptr - l, l);
> -				kunmap(*src_page);
> +				kunmap_local(src_ptr);
> 				set_page_dirty(*dst_page);
> -				kunmap(*dst_page);
> +				kunmap_local(dst_ptr);
> 				if (dst == PAGE_SIZE)
> 					dst_page--;
> 				else
> @@ -251,26 +243,27 @@ void hfs_bnode_move(struct hfs_bnode *node, int dst, int src, int len)
> 
> 		if (src == dst) {
> 			l = min_t(int, len, PAGE_SIZE - src);
> -			memmove(kmap(*dst_page) + src,
> -				kmap(*src_page) + src, l);
> -			kunmap(*src_page);
> +
> +			dst_ptr = kmap_local_page(*dst_page) + src;
> +			src_ptr = kmap_local_page(*src_page) + src;
> +			memmove(dst_ptr, src_ptr, l);
> +			kunmap_local(src_ptr);
> 			set_page_dirty(*dst_page);
> -			kunmap(*dst_page);
> +			kunmap_local(dst_ptr);
> 
> 			while ((len -= l) != 0) {
> 				l = min_t(int, len, PAGE_SIZE);
> -				memmove(kmap(*++dst_page),
> -					kmap(*++src_page), l);
> -				kunmap(*src_page);
> +				dst_ptr = kmap_local_page(*++dst_page);
> +				src_ptr = kmap_local_page(*++src_page);
> +				memmove(dst_ptr, src_ptr, l);
> +				kunmap_local(src_ptr);
> 				set_page_dirty(*dst_page);
> -				kunmap(*dst_page);
> +				kunmap_local(dst_ptr);
> 			}
> 		} else {
> -			void *src_ptr, *dst_ptr;
> -
> 			do {
> -				src_ptr = kmap(*src_page) + src;
> -				dst_ptr = kmap(*dst_page) + dst;
> +				dst_ptr = kmap_local_page(*dst_page) + dst;
> +				src_ptr = kmap_local_page(*src_page) + src;
> 				if (PAGE_SIZE - src <
> 						PAGE_SIZE - dst) {
> 					l = PAGE_SIZE - src;
> @@ -283,9 +276,9 @@ void hfs_bnode_move(struct hfs_bnode *node, int dst, int src, int len)
> 				}
> 				l = min(len, l);
> 				memmove(dst_ptr, src_ptr, l);
> -				kunmap(*src_page);
> +				kunmap_local(src_ptr);
> 				set_page_dirty(*dst_page);
> -				kunmap(*dst_page);
> +				kunmap_local(dst_ptr);
> 				if (!dst)
> 					dst_page++;
> 				else
> @@ -498,14 +491,14 @@ struct hfs_bnode *hfs_bnode_find(struct hfs_btree *tree, u32 num)
> 	if (!test_bit(HFS_BNODE_NEW, &node->flags))
> 		return node;
> 
> -	desc = (struct hfs_bnode_desc *)(kmap(node->page[0]) +
> -			node->page_offset);
> +	desc = (struct hfs_bnode_desc *)(kmap_local_page(node->page[0]) +
> +							 node->page_offset);
> 	node->prev = be32_to_cpu(desc->prev);
> 	node->next = be32_to_cpu(desc->next);
> 	node->num_recs = be16_to_cpu(desc->num_recs);
> 	node->type = desc->type;
> 	node->height = desc->height;
> -	kunmap(node->page[0]);
> +	kunmap_local(desc);
> 
> 	switch (node->type) {
> 	case HFS_NODE_HEADER:
> @@ -589,14 +582,12 @@ struct hfs_bnode *hfs_bnode_create(struct hfs_btree *tree, u32 num)
> 	}
> 
> 	pagep = node->page;
> -	memset(kmap(*pagep) + node->page_offset, 0,
> -	       min_t(int, PAGE_SIZE, tree->node_size));
> +	memzero_page(*pagep, node->page_offset,
> +		     min_t(int, PAGE_SIZE, tree->node_size));
> 	set_page_dirty(*pagep);
> -	kunmap(*pagep);
> 	for (i = 1; i < tree->pages_per_bnode; i++) {
> -		memset(kmap(*++pagep), 0, PAGE_SIZE);
> +		memzero_page(*++pagep, 0, PAGE_SIZE);
> 		set_page_dirty(*pagep);
> -		kunmap(*pagep);
> 	}
> 	clear_bit(HFS_BNODE_NEW, &node->flags);
> 	wake_up(&node->lock_wq);
> -- 
> 2.37.1
> 


^ permalink raw reply	[flat|nested] 9+ messages in thread

* Re: [PATCH 3/4] hfsplus: Convert kmap() to kmap_local_page() in bitmap.c
  2022-08-09 20:31 ` [PATCH 3/4] hfsplus: Convert kmap() to kmap_local_page() in bitmap.c Fabio M. De Francesco
@ 2022-08-09 22:18   ` Viacheslav Dubeyko
  0 siblings, 0 replies; 9+ messages in thread
From: Viacheslav Dubeyko @ 2022-08-09 22:18 UTC (permalink / raw)
  To: Fabio M. De Francesco
  Cc: Matthew Wilcox (Oracle),
	Ira Weiny, Jens Axboe, Andrew Morton, Bart Van Assche, Kees Cook,
	Muchun Song, Linux FS Devel, LKML



> On Aug 9, 2022, at 1:31 PM, Fabio M. De Francesco <fmdefrancesco@gmail.com> wrote:
> 
> kmap() is being deprecated in favor of kmap_local_page().
> 
> There are two main problems with kmap(): (1) It comes with an overhead as
> mapping space is restricted and protected by a global lock for
> synchronization and (2) it also requires global TLB invalidation when the
> kmap’s pool wraps and it might block when the mapping space is fully
> utilized until a slot becomes available.
> 
> With kmap_local_page() the mappings are per thread, CPU local, can take
> page faults, and can be called from any context (including interrupts).
> It is faster than kmap() in kernels with HIGHMEM enabled. Furthermore,
> the tasks can be preempted and, when they are scheduled to run again, the
> kernel virtual addresses are restored and are still valid.
> 
> Since its use in bitmap.c is safe everywhere, it should be preferred.
> 
> Therefore, replace kmap() with kmap_local_page() in bitmap.c.
> 
> Tested in a QEMU/KVM x86_32 VM, 6GB RAM, booting a kernel with
> HIGHMEM64GB enabled.
> 
> Cc: Viacheslav Dubeyko <slava@dubeyko.com>
> Suggested-by: Ira Weiny <ira.weiny@intel.com>
> Reviewed-by: Ira Weiny <ira.weiny@intel.com>
> Signed-off-by: Fabio M. De Francesco <fmdefrancesco@gmail.com>
> ---


Looks good.

Reviewed-by: Viacheslav Dubeyko <slava@dubeyko.com>

Thanks,
Slava.


> fs/hfsplus/bitmap.c | 20 ++++++++++----------
> 1 file changed, 10 insertions(+), 10 deletions(-)
> 
> diff --git a/fs/hfsplus/bitmap.c b/fs/hfsplus/bitmap.c
> index cebce0cfe340..bd8dcea85588 100644
> --- a/fs/hfsplus/bitmap.c
> +++ b/fs/hfsplus/bitmap.c
> @@ -39,7 +39,7 @@ int hfsplus_block_allocate(struct super_block *sb, u32 size,
> 		start = size;
> 		goto out;
> 	}
> -	pptr = kmap(page);
> +	pptr = kmap_local_page(page);
> 	curr = pptr + (offset & (PAGE_CACHE_BITS - 1)) / 32;
> 	i = offset % 32;
> 	offset &= ~(PAGE_CACHE_BITS - 1);
> @@ -74,7 +74,7 @@ int hfsplus_block_allocate(struct super_block *sb, u32 size,
> 			}
> 			curr++;
> 		}
> -		kunmap(page);
> +		kunmap_local(pptr);
> 		offset += PAGE_CACHE_BITS;
> 		if (offset >= size)
> 			break;
> @@ -84,7 +84,7 @@ int hfsplus_block_allocate(struct super_block *sb, u32 size,
> 			start = size;
> 			goto out;
> 		}
> -		curr = pptr = kmap(page);
> +		curr = pptr = kmap_local_page(page);
> 		if ((size ^ offset) / PAGE_CACHE_BITS)
> 			end = pptr + PAGE_CACHE_BITS / 32;
> 		else
> @@ -127,7 +127,7 @@ int hfsplus_block_allocate(struct super_block *sb, u32 size,
> 			len -= 32;
> 		}
> 		set_page_dirty(page);
> -		kunmap(page);
> +		kunmap_local(pptr);
> 		offset += PAGE_CACHE_BITS;
> 		page = read_mapping_page(mapping, offset / PAGE_CACHE_BITS,
> 					 NULL);
> @@ -135,7 +135,7 @@ int hfsplus_block_allocate(struct super_block *sb, u32 size,
> 			start = size;
> 			goto out;
> 		}
> -		pptr = kmap(page);
> +		pptr = kmap_local_page(page);
> 		curr = pptr;
> 		end = pptr + PAGE_CACHE_BITS / 32;
> 	}
> @@ -151,7 +151,7 @@ int hfsplus_block_allocate(struct super_block *sb, u32 size,
> done:
> 	*curr = cpu_to_be32(n);
> 	set_page_dirty(page);
> -	kunmap(page);
> +	kunmap_local(pptr);
> 	*max = offset + (curr - pptr) * 32 + i - start;
> 	sbi->free_blocks -= *max;
> 	hfsplus_mark_mdb_dirty(sb);
> @@ -185,7 +185,7 @@ int hfsplus_block_free(struct super_block *sb, u32 offset, u32 count)
> 	page = read_mapping_page(mapping, pnr, NULL);
> 	if (IS_ERR(page))
> 		goto kaboom;
> -	pptr = kmap(page);
> +	pptr = kmap_local_page(page);
> 	curr = pptr + (offset & (PAGE_CACHE_BITS - 1)) / 32;
> 	end = pptr + PAGE_CACHE_BITS / 32;
> 	len = count;
> @@ -215,11 +215,11 @@ int hfsplus_block_free(struct super_block *sb, u32 offset, u32 count)
> 		if (!count)
> 			break;
> 		set_page_dirty(page);
> -		kunmap(page);
> +		kunmap_local(pptr);
> 		page = read_mapping_page(mapping, ++pnr, NULL);
> 		if (IS_ERR(page))
> 			goto kaboom;
> -		pptr = kmap(page);
> +		pptr = kmap_local_page(page);
> 		curr = pptr;
> 		end = pptr + PAGE_CACHE_BITS / 32;
> 	}
> @@ -231,7 +231,7 @@ int hfsplus_block_free(struct super_block *sb, u32 offset, u32 count)
> 	}
> out:
> 	set_page_dirty(page);
> -	kunmap(page);
> +	kunmap_local(pptr);
> 	sbi->free_blocks += len;
> 	hfsplus_mark_mdb_dirty(sb);
> 	mutex_unlock(&sbi->alloc_mutex);
> -- 
> 2.37.1
> 


^ permalink raw reply	[flat|nested] 9+ messages in thread

* Re: [PATCH 4/4] hfsplus: Convert kmap() to kmap_local_page() in btree.c
  2022-08-09 20:31 ` [PATCH 4/4] hfsplus: Convert kmap() to kmap_local_page() in btree.c Fabio M. De Francesco
@ 2022-08-09 22:18   ` Viacheslav Dubeyko
  0 siblings, 0 replies; 9+ messages in thread
From: Viacheslav Dubeyko @ 2022-08-09 22:18 UTC (permalink / raw)
  To: Fabio M. De Francesco
  Cc: Matthew Wilcox (Oracle),
	Ira Weiny, Jens Axboe, Andrew Morton, Bart Van Assche, Kees Cook,
	Muchun Song, Linux FS Devel, LKML



> On Aug 9, 2022, at 1:31 PM, Fabio M. De Francesco <fmdefrancesco@gmail.com> wrote:
> 
> kmap() is being deprecated in favor of kmap_local_page().
> 
> There are two main problems with kmap(): (1) It comes with an overhead as
> mapping space is restricted and protected by a global lock for
> synchronization and (2) it also requires global TLB invalidation when the
> kmap’s pool wraps and it might block when the mapping space is fully
> utilized until a slot becomes available.
> 
> With kmap_local_page() the mappings are per thread, CPU local, can take
> page faults, and can be called from any context (including interrupts).
> It is faster than kmap() in kernels with HIGHMEM enabled. Furthermore,
> the tasks can be preempted and, when they are scheduled to run again, the
> kernel virtual addresses are restored and are still valid.
> 
> Since its use in btree.c is safe everywhere, it should be preferred.
> 
> Therefore, replace kmap() with kmap_local_page() in btree.c.
> 
> Tested in a QEMU/KVM x86_32 VM, 6GB RAM, booting a kernel with
> HIGHMEM64GB enabled.
> 
> Cc: Viacheslav Dubeyko <slava@dubeyko.com>
> Suggested-by: Ira Weiny <ira.weiny@intel.com>
> Reviewed-by: Ira Weiny <ira.weiny@intel.com>
> Signed-off-by: Fabio M. De Francesco <fmdefrancesco@gmail.com>
> ---


Looks good.

Reviewed-by: Viacheslav Dubeyko <slava@dubeyko.com>

Thanks,
Slava.


> fs/hfsplus/btree.c | 28 ++++++++++++++--------------
> 1 file changed, 14 insertions(+), 14 deletions(-)
> 
> diff --git a/fs/hfsplus/btree.c b/fs/hfsplus/btree.c
> index 3a917a9a4edd..9e1732a2b92a 100644
> --- a/fs/hfsplus/btree.c
> +++ b/fs/hfsplus/btree.c
> @@ -163,7 +163,7 @@ struct hfs_btree *hfs_btree_open(struct super_block *sb, u32 id)
> 		goto free_inode;
> 
> 	/* Load the header */
> -	head = (struct hfs_btree_header_rec *)(kmap(page) +
> +	head = (struct hfs_btree_header_rec *)(kmap_local_page(page) +
> 		sizeof(struct hfs_bnode_desc));
> 	tree->root = be32_to_cpu(head->root);
> 	tree->leaf_count = be32_to_cpu(head->leaf_count);
> @@ -240,12 +240,12 @@ struct hfs_btree *hfs_btree_open(struct super_block *sb, u32 id)
> 		(tree->node_size + PAGE_SIZE - 1) >>
> 		PAGE_SHIFT;
> 
> -	kunmap(page);
> +	kunmap_local(head);
> 	put_page(page);
> 	return tree;
> 
>  fail_page:
> -	kunmap(page);
> +	kunmap_local(head);
> 	put_page(page);
>  free_inode:
> 	tree->inode->i_mapping->a_ops = &hfsplus_aops;
> @@ -292,7 +292,7 @@ int hfs_btree_write(struct hfs_btree *tree)
> 		return -EIO;
> 	/* Load the header */
> 	page = node->page[0];
> -	head = (struct hfs_btree_header_rec *)(kmap(page) +
> +	head = (struct hfs_btree_header_rec *)(kmap_local_page(page) +
> 		sizeof(struct hfs_bnode_desc));
> 
> 	head->root = cpu_to_be32(tree->root);
> @@ -304,7 +304,7 @@ int hfs_btree_write(struct hfs_btree *tree)
> 	head->attributes = cpu_to_be32(tree->attributes);
> 	head->depth = cpu_to_be16(tree->depth);
> 
> -	kunmap(page);
> +	kunmap_local(head);
> 	set_page_dirty(page);
> 	hfs_bnode_put(node);
> 	return 0;
> @@ -395,7 +395,7 @@ struct hfs_bnode *hfs_bmap_alloc(struct hfs_btree *tree)
> 
> 	off += node->page_offset;
> 	pagep = node->page + (off >> PAGE_SHIFT);
> -	data = kmap(*pagep);
> +	data = kmap_local_page(*pagep);
> 	off &= ~PAGE_MASK;
> 	idx = 0;
> 
> @@ -408,7 +408,7 @@ struct hfs_bnode *hfs_bmap_alloc(struct hfs_btree *tree)
> 						idx += i;
> 						data[off] |= m;
> 						set_page_dirty(*pagep);
> -						kunmap(*pagep);
> +						kunmap_local(data);
> 						tree->free_nodes--;
> 						mark_inode_dirty(tree->inode);
> 						hfs_bnode_put(node);
> @@ -418,14 +418,14 @@ struct hfs_bnode *hfs_bmap_alloc(struct hfs_btree *tree)
> 				}
> 			}
> 			if (++off >= PAGE_SIZE) {
> -				kunmap(*pagep);
> -				data = kmap(*++pagep);
> +				kunmap_local(data);
> +				data = kmap_local_page(*++pagep);
> 				off = 0;
> 			}
> 			idx += 8;
> 			len--;
> 		}
> -		kunmap(*pagep);
> +		kunmap_local(data);
> 		nidx = node->next;
> 		if (!nidx) {
> 			hfs_dbg(BNODE_MOD, "create new bmap node\n");
> @@ -441,7 +441,7 @@ struct hfs_bnode *hfs_bmap_alloc(struct hfs_btree *tree)
> 		off = off16;
> 		off += node->page_offset;
> 		pagep = node->page + (off >> PAGE_SHIFT);
> -		data = kmap(*pagep);
> +		data = kmap_local_page(*pagep);
> 		off &= ~PAGE_MASK;
> 	}
> }
> @@ -491,7 +491,7 @@ void hfs_bmap_free(struct hfs_bnode *node)
> 	}
> 	off += node->page_offset + nidx / 8;
> 	page = node->page[off >> PAGE_SHIFT];
> -	data = kmap(page);
> +	data = kmap_local_page(page);
> 	off &= ~PAGE_MASK;
> 	m = 1 << (~nidx & 7);
> 	byte = data[off];
> @@ -499,13 +499,13 @@ void hfs_bmap_free(struct hfs_bnode *node)
> 		pr_crit("trying to free free bnode "
> 				"%u(%d)\n",
> 			node->this, node->type);
> -		kunmap(page);
> +		kunmap_local(data);
> 		hfs_bnode_put(node);
> 		return;
> 	}
> 	data[off] = byte & ~m;
> 	set_page_dirty(page);
> -	kunmap(page);
> +	kunmap_local(data);
> 	hfs_bnode_put(node);
> 	tree->free_nodes++;
> 	mark_inode_dirty(tree->inode);
> -- 
> 2.37.1
> 


^ permalink raw reply	[flat|nested] 9+ messages in thread

end of thread, other threads:[~2022-08-09 22:19 UTC | newest]

Thread overview: 9+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2022-08-09 20:31 [PATCH 0/4] hfsplus: Replace kmap() with kmap_local_page() Fabio M. De Francesco
2022-08-09 20:31 ` [PATCH 1/4] hfsplus: Unmap the page in the "fail_page" label Fabio M. De Francesco
2022-08-09 22:16   ` Viacheslav Dubeyko
2022-08-09 20:31 ` [PATCH 2/4] hfsplus: Convert kmap() to kmap_local_page() in bnode.c Fabio M. De Francesco
2022-08-09 22:17   ` Viacheslav Dubeyko
2022-08-09 20:31 ` [PATCH 3/4] hfsplus: Convert kmap() to kmap_local_page() in bitmap.c Fabio M. De Francesco
2022-08-09 22:18   ` Viacheslav Dubeyko
2022-08-09 20:31 ` [PATCH 4/4] hfsplus: Convert kmap() to kmap_local_page() in btree.c Fabio M. De Francesco
2022-08-09 22:18   ` Viacheslav Dubeyko

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).