All of lore.kernel.org
 help / color / mirror / Atom feed
From: "Matthew Wilcox (Oracle)" <willy@infradead.org>
To: linux-kernel@vger.kernel.org
Cc: "Matthew Wilcox (Oracle)" <willy@infradead.org>,
	linux-mm@kvack.org, linux-fsdevel@vger.kernel.org,
	Christoph Hellwig <hch@lst.de>, Jeff Layton <jlayton@kernel.org>,
	"Kirill A . Shutemov" <kirill.shutemov@linux.intel.com>,
	Vlastimil Babka <vbabka@suse.cz>,
	William Kucharski <william.kucharski@oracle.com>,
	David Howells <dhowells@redhat.com>
Subject: [PATCH v13 010/137] mm: Add folio flag manipulation functions
Date: Mon, 12 Jul 2021 04:04:54 +0100	[thread overview]
Message-ID: <20210712030701.4000097-11-willy@infradead.org> (raw)
In-Reply-To: <20210712030701.4000097-1-willy@infradead.org>

These new functions are the folio analogues of the various PageFlags
functions.  If CONFIG_DEBUG_VM_PGFLAGS is enabled, we check the folio
is not a tail page at every invocation.  This will also catch the
PagePoisoned case as a poisoned page has every bit set, which would
include PageTail.

This saves 1684 bytes of text with the distro-derived config that
I'm testing due to removing a double call to compound_head() in
PageSwapCache().

Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Acked-by: Jeff Layton <jlayton@kernel.org>
Acked-by: Kirill A. Shutemov <kirill.shutemov@linux.intel.com>
Acked-by: Vlastimil Babka <vbabka@suse.cz>
Reviewed-by: William Kucharski <william.kucharski@oracle.com>
Reviewed-by: David Howells <dhowells@redhat.com>
---
 include/linux/page-flags.h | 219 ++++++++++++++++++++++++++-----------
 1 file changed, 156 insertions(+), 63 deletions(-)

diff --git a/include/linux/page-flags.h b/include/linux/page-flags.h
index 70ede8345538..fb914468b302 100644
--- a/include/linux/page-flags.h
+++ b/include/linux/page-flags.h
@@ -143,6 +143,8 @@ enum pageflags {
 #endif
 	__NR_PAGEFLAGS,
 
+	PG_readahead = PG_reclaim,
+
 	/* Filesystems */
 	PG_checked = PG_owner_priv_1,
 
@@ -243,6 +245,15 @@ static inline void page_init_poison(struct page *page, size_t size)
 }
 #endif
 
+static unsigned long *folio_flags(struct folio *folio, unsigned n)
+{
+	struct page *page = &folio->page;
+
+	VM_BUG_ON_PGFLAGS(PageTail(page), page);
+	VM_BUG_ON_PGFLAGS(n > 0 && !test_bit(PG_head, &page->flags), page);
+	return &page[n].flags;
+}
+
 /*
  * Page flags policies wrt compound pages
  *
@@ -287,36 +298,64 @@ static inline void page_init_poison(struct page *page, size_t size)
 		VM_BUG_ON_PGFLAGS(!PageHead(page), page);		\
 		PF_POISONED_CHECK(&page[1]); })
 
+/* Which page is the flag stored in */
+#define FOLIO_PF_ANY		0
+#define FOLIO_PF_HEAD		0
+#define FOLIO_PF_ONLY_HEAD	0
+#define FOLIO_PF_NO_TAIL	0
+#define FOLIO_PF_NO_COMPOUND	0
+#define FOLIO_PF_SECOND		1
+
 /*
  * Macros to create function definitions for page flags
  */
 #define TESTPAGEFLAG(uname, lname, policy)				\
+static __always_inline bool folio_##lname(struct folio *folio)		\
+{ return test_bit(PG_##lname, folio_flags(folio, FOLIO_##policy)); }	\
 static __always_inline int Page##uname(struct page *page)		\
-	{ return test_bit(PG_##lname, &policy(page, 0)->flags); }
+{ return test_bit(PG_##lname, &policy(page, 0)->flags); }
 
 #define SETPAGEFLAG(uname, lname, policy)				\
+static __always_inline							\
+void folio_set_##lname##_flag(struct folio *folio)			\
+{ set_bit(PG_##lname, folio_flags(folio, FOLIO_##policy)); }		\
 static __always_inline void SetPage##uname(struct page *page)		\
-	{ set_bit(PG_##lname, &policy(page, 1)->flags); }
+{ set_bit(PG_##lname, &policy(page, 1)->flags); }
 
 #define CLEARPAGEFLAG(uname, lname, policy)				\
+static __always_inline							\
+void folio_clear_##lname##_flag(struct folio *folio)			\
+{ clear_bit(PG_##lname, folio_flags(folio, FOLIO_##policy)); }		\
 static __always_inline void ClearPage##uname(struct page *page)		\
-	{ clear_bit(PG_##lname, &policy(page, 1)->flags); }
+{ clear_bit(PG_##lname, &policy(page, 1)->flags); }
 
 #define __SETPAGEFLAG(uname, lname, policy)				\
+static __always_inline							\
+void __folio_set_##lname##_flag(struct folio *folio)			\
+{ __set_bit(PG_##lname, folio_flags(folio, FOLIO_##policy)); }		\
 static __always_inline void __SetPage##uname(struct page *page)		\
-	{ __set_bit(PG_##lname, &policy(page, 1)->flags); }
+{ __set_bit(PG_##lname, &policy(page, 1)->flags); }
 
 #define __CLEARPAGEFLAG(uname, lname, policy)				\
+static __always_inline							\
+void __folio_clear_##lname##_flag(struct folio *folio)			\
+{ __clear_bit(PG_##lname, folio_flags(folio, FOLIO_##policy)); }	\
 static __always_inline void __ClearPage##uname(struct page *page)	\
-	{ __clear_bit(PG_##lname, &policy(page, 1)->flags); }
+{ __clear_bit(PG_##lname, &policy(page, 1)->flags); }
 
 #define TESTSETFLAG(uname, lname, policy)				\
+static __always_inline							\
+bool folio_test_set_##lname##_flag(struct folio *folio)			\
+{ return test_and_set_bit(PG_##lname, folio_flags(folio, FOLIO_##policy)); } \
 static __always_inline int TestSetPage##uname(struct page *page)	\
-	{ return test_and_set_bit(PG_##lname, &policy(page, 1)->flags); }
+{ return test_and_set_bit(PG_##lname, &policy(page, 1)->flags); }
 
 #define TESTCLEARFLAG(uname, lname, policy)				\
+static __always_inline							\
+bool folio_test_clear_##lname##_flag(struct folio *folio)		\
+{ return test_and_clear_bit(PG_##lname, folio_flags(folio, FOLIO_##policy)); } \
 static __always_inline int TestClearPage##uname(struct page *page)	\
-	{ return test_and_clear_bit(PG_##lname, &policy(page, 1)->flags); }
+{ return test_and_clear_bit(PG_##lname, &policy(page, 1)->flags); }
 
 #define PAGEFLAG(uname, lname, policy)					\
 	TESTPAGEFLAG(uname, lname, policy)				\
@@ -332,29 +371,37 @@ static __always_inline int TestClearPage##uname(struct page *page)	\
 	TESTSETFLAG(uname, lname, policy)				\
 	TESTCLEARFLAG(uname, lname, policy)
 
-#define TESTPAGEFLAG_FALSE(uname)					\
+#define TESTPAGEFLAG_FALSE(uname, lname)				\
+static inline bool folio_##lname(const struct folio *folio) { return 0; } \
 static inline int Page##uname(const struct page *page) { return 0; }
 
-#define SETPAGEFLAG_NOOP(uname)						\
+#define SETPAGEFLAG_NOOP(uname, lname)					\
+static inline void folio_set_##lname##_flag(struct folio *folio) { }	\
 static inline void SetPage##uname(struct page *page) {  }
 
-#define CLEARPAGEFLAG_NOOP(uname)					\
+#define CLEARPAGEFLAG_NOOP(uname, lname)				\
+static inline void folio_clear_##lname##_flag(struct folio *folio) { }	\
 static inline void ClearPage##uname(struct page *page) {  }
 
-#define __CLEARPAGEFLAG_NOOP(uname)					\
+#define __CLEARPAGEFLAG_NOOP(uname, lname)				\
+static inline void __folio_clear_##lname_flags(struct folio *folio) { }	\
 static inline void __ClearPage##uname(struct page *page) {  }
 
-#define TESTSETFLAG_FALSE(uname)					\
+#define TESTSETFLAG_FALSE(uname, lname)					\
+static inline bool folio_test_set_##lname##_flag(struct folio *folio)	\
+{ return 0; }								\
 static inline int TestSetPage##uname(struct page *page) { return 0; }
 
-#define TESTCLEARFLAG_FALSE(uname)					\
+#define TESTCLEARFLAG_FALSE(uname, lname)				\
+static inline bool folio_test_clear_##lname##_flag(struct folio *folio) \
+{ return 0; }								\
 static inline int TestClearPage##uname(struct page *page) { return 0; }
 
-#define PAGEFLAG_FALSE(uname) TESTPAGEFLAG_FALSE(uname)			\
-	SETPAGEFLAG_NOOP(uname) CLEARPAGEFLAG_NOOP(uname)
+#define PAGEFLAG_FALSE(uname, lname) TESTPAGEFLAG_FALSE(uname, lname)	\
+	SETPAGEFLAG_NOOP(uname, lname) CLEARPAGEFLAG_NOOP(uname, lname)
 
-#define TESTSCFLAG_FALSE(uname)						\
-	TESTSETFLAG_FALSE(uname) TESTCLEARFLAG_FALSE(uname)
+#define TESTSCFLAG_FALSE(uname, lname)					\
+	TESTSETFLAG_FALSE(uname, lname) TESTCLEARFLAG_FALSE(uname, lname)
 
 __PAGEFLAG(Locked, locked, PF_NO_TAIL)
 PAGEFLAG(Waiters, waiters, PF_ONLY_HEAD) __CLEARPAGEFLAG(Waiters, waiters, PF_ONLY_HEAD)
@@ -410,8 +457,8 @@ PAGEFLAG(MappedToDisk, mappedtodisk, PF_NO_TAIL)
 /* PG_readahead is only used for reads; PG_reclaim is only for writes */
 PAGEFLAG(Reclaim, reclaim, PF_NO_TAIL)
 	TESTCLEARFLAG(Reclaim, reclaim, PF_NO_TAIL)
-PAGEFLAG(Readahead, reclaim, PF_NO_COMPOUND)
-	TESTCLEARFLAG(Readahead, reclaim, PF_NO_COMPOUND)
+PAGEFLAG(Readahead, readahead, PF_NO_COMPOUND)
+	TESTCLEARFLAG(Readahead, readahead, PF_NO_COMPOUND)
 
 #ifdef CONFIG_HIGHMEM
 /*
@@ -420,22 +467,25 @@ PAGEFLAG(Readahead, reclaim, PF_NO_COMPOUND)
  */
 #define PageHighMem(__p) is_highmem_idx(page_zonenum(__p))
 #else
-PAGEFLAG_FALSE(HighMem)
+PAGEFLAG_FALSE(HighMem, highmem)
 #endif
 
 #ifdef CONFIG_SWAP
-static __always_inline int PageSwapCache(struct page *page)
+static __always_inline bool folio_swapcache(struct folio *folio)
 {
-#ifdef CONFIG_THP_SWAP
-	page = compound_head(page);
-#endif
-	return PageSwapBacked(page) && test_bit(PG_swapcache, &page->flags);
+	return folio_swapbacked(folio) &&
+			test_bit(PG_swapcache, folio_flags(folio, 0));
+}
 
+static __always_inline bool PageSwapCache(struct page *page)
+{
+	return folio_swapcache(page_folio(page));
 }
+
 SETPAGEFLAG(SwapCache, swapcache, PF_NO_TAIL)
 CLEARPAGEFLAG(SwapCache, swapcache, PF_NO_TAIL)
 #else
-PAGEFLAG_FALSE(SwapCache)
+PAGEFLAG_FALSE(SwapCache, swapcache)
 #endif
 
 PAGEFLAG(Unevictable, unevictable, PF_HEAD)
@@ -447,14 +497,14 @@ PAGEFLAG(Mlocked, mlocked, PF_NO_TAIL)
 	__CLEARPAGEFLAG(Mlocked, mlocked, PF_NO_TAIL)
 	TESTSCFLAG(Mlocked, mlocked, PF_NO_TAIL)
 #else
-PAGEFLAG_FALSE(Mlocked) __CLEARPAGEFLAG_NOOP(Mlocked)
-	TESTSCFLAG_FALSE(Mlocked)
+PAGEFLAG_FALSE(Mlocked, mlocked) __CLEARPAGEFLAG_NOOP(Mlocked, mlocked)
+	TESTSCFLAG_FALSE(Mlocked, mlocked)
 #endif
 
 #ifdef CONFIG_ARCH_USES_PG_UNCACHED
 PAGEFLAG(Uncached, uncached, PF_NO_COMPOUND)
 #else
-PAGEFLAG_FALSE(Uncached)
+PAGEFLAG_FALSE(Uncached, uncached)
 #endif
 
 #ifdef CONFIG_MEMORY_FAILURE
@@ -463,7 +513,7 @@ TESTSCFLAG(HWPoison, hwpoison, PF_ANY)
 #define __PG_HWPOISON (1UL << PG_hwpoison)
 extern bool take_page_off_buddy(struct page *page);
 #else
-PAGEFLAG_FALSE(HWPoison)
+PAGEFLAG_FALSE(HWPoison, hwpoison)
 #define __PG_HWPOISON 0
 #endif
 
@@ -477,7 +527,7 @@ PAGEFLAG(Idle, idle, PF_ANY)
 #ifdef CONFIG_KASAN_HW_TAGS
 PAGEFLAG(SkipKASanPoison, skip_kasan_poison, PF_HEAD)
 #else
-PAGEFLAG_FALSE(SkipKASanPoison)
+PAGEFLAG_FALSE(SkipKASanPoison, skip_kasan_poison)
 #endif
 
 /*
@@ -515,10 +565,14 @@ static __always_inline int PageMappingFlags(struct page *page)
 	return ((unsigned long)page->mapping & PAGE_MAPPING_FLAGS) != 0;
 }
 
-static __always_inline int PageAnon(struct page *page)
+static __always_inline bool folio_anon(struct folio *folio)
+{
+	return ((unsigned long)folio->mapping & PAGE_MAPPING_ANON) != 0;
+}
+
+static __always_inline bool PageAnon(struct page *page)
 {
-	page = compound_head(page);
-	return ((unsigned long)page->mapping & PAGE_MAPPING_ANON) != 0;
+	return folio_anon(page_folio(page));
 }
 
 static __always_inline int __PageMovable(struct page *page)
@@ -534,30 +588,32 @@ static __always_inline int __PageMovable(struct page *page)
  * is found in VM_MERGEABLE vmas.  It's a PageAnon page, pointing not to any
  * anon_vma, but to that page's node of the stable tree.
  */
-static __always_inline int PageKsm(struct page *page)
+static __always_inline bool folio_ksm(struct folio *folio)
 {
-	page = compound_head(page);
-	return ((unsigned long)page->mapping & PAGE_MAPPING_FLAGS) ==
+	return ((unsigned long)folio->mapping & PAGE_MAPPING_FLAGS) ==
 				PAGE_MAPPING_KSM;
 }
+
+static __always_inline bool PageKsm(struct page *page)
+{
+	return folio_ksm(page_folio(page));
+}
 #else
-TESTPAGEFLAG_FALSE(Ksm)
+TESTPAGEFLAG_FALSE(Ksm, ksm)
 #endif
 
 u64 stable_page_flags(struct page *page);
 
-static inline int PageUptodate(struct page *page)
+static inline bool folio_uptodate(struct folio *folio)
 {
-	int ret;
-	page = compound_head(page);
-	ret = test_bit(PG_uptodate, &(page)->flags);
+	bool ret = test_bit(PG_uptodate, folio_flags(folio, 0));
 	/*
-	 * Must ensure that the data we read out of the page is loaded
-	 * _after_ we've loaded page->flags to check for PageUptodate.
-	 * We can skip the barrier if the page is not uptodate, because
+	 * Must ensure that the data we read out of the folio is loaded
+	 * _after_ we've loaded folio->flags to check the uptodate bit.
+	 * We can skip the barrier if the folio is not uptodate, because
 	 * we wouldn't be reading anything from it.
 	 *
-	 * See SetPageUptodate() for the other side of the story.
+	 * See folio_mark_uptodate() for the other side of the story.
 	 */
 	if (ret)
 		smp_rmb();
@@ -565,23 +621,36 @@ static inline int PageUptodate(struct page *page)
 	return ret;
 }
 
-static __always_inline void __SetPageUptodate(struct page *page)
+static inline int PageUptodate(struct page *page)
+{
+	return folio_uptodate(page_folio(page));
+}
+
+static __always_inline void __folio_mark_uptodate(struct folio *folio)
 {
-	VM_BUG_ON_PAGE(PageTail(page), page);
 	smp_wmb();
-	__set_bit(PG_uptodate, &page->flags);
+	__set_bit(PG_uptodate, folio_flags(folio, 0));
 }
 
-static __always_inline void SetPageUptodate(struct page *page)
+static __always_inline void folio_mark_uptodate(struct folio *folio)
 {
-	VM_BUG_ON_PAGE(PageTail(page), page);
 	/*
 	 * Memory barrier must be issued before setting the PG_uptodate bit,
-	 * so that all previous stores issued in order to bring the page
-	 * uptodate are actually visible before PageUptodate becomes true.
+	 * so that all previous stores issued in order to bring the folio
+	 * uptodate are actually visible before folio_uptodate becomes true.
 	 */
 	smp_wmb();
-	set_bit(PG_uptodate, &page->flags);
+	set_bit(PG_uptodate, folio_flags(folio, 0));
+}
+
+static __always_inline void __SetPageUptodate(struct page *page)
+{
+	__folio_mark_uptodate((struct folio *)page);
+}
+
+static __always_inline void SetPageUptodate(struct page *page)
+{
+	folio_mark_uptodate((struct folio *)page);
 }
 
 CLEARPAGEFLAG(Uptodate, uptodate, PF_NO_TAIL)
@@ -606,6 +675,17 @@ static inline void set_page_writeback_keepwrite(struct page *page)
 
 __PAGEFLAG(Head, head, PF_ANY) CLEARPAGEFLAG(Head, head, PF_ANY)
 
+/* Whether there are one or multiple pages in a folio */
+static inline bool folio_single(struct folio *folio)
+{
+	return !folio_head(folio);
+}
+
+static inline bool folio_multi(struct folio *folio)
+{
+	return folio_head(folio);
+}
+
 static __always_inline void set_compound_head(struct page *page, struct page *head)
 {
 	WRITE_ONCE(page->compound_head, (unsigned long)head + 1);
@@ -629,12 +709,15 @@ static inline void ClearPageCompound(struct page *page)
 #ifdef CONFIG_HUGETLB_PAGE
 int PageHuge(struct page *page);
 int PageHeadHuge(struct page *page);
+static inline bool folio_hugetlb(struct folio *folio)
+{
+	return PageHeadHuge(&folio->page);
+}
 #else
-TESTPAGEFLAG_FALSE(Huge)
-TESTPAGEFLAG_FALSE(HeadHuge)
+TESTPAGEFLAG_FALSE(Huge, hugetlb)
+TESTPAGEFLAG_FALSE(HeadHuge, headhuge)
 #endif
 
-
 #ifdef CONFIG_TRANSPARENT_HUGEPAGE
 /*
  * PageHuge() only returns true for hugetlbfs pages, but not for
@@ -650,6 +733,11 @@ static inline int PageTransHuge(struct page *page)
 	return PageHead(page);
 }
 
+static inline bool folio_transhuge(struct folio *folio)
+{
+	return folio_head(folio);
+}
+
 /*
  * PageTransCompound returns true for both transparent huge pages
  * and hugetlbfs pages, so it should only be called when it's known
@@ -723,12 +811,12 @@ static inline int PageTransTail(struct page *page)
 PAGEFLAG(DoubleMap, double_map, PF_SECOND)
 	TESTSCFLAG(DoubleMap, double_map, PF_SECOND)
 #else
-TESTPAGEFLAG_FALSE(TransHuge)
-TESTPAGEFLAG_FALSE(TransCompound)
-TESTPAGEFLAG_FALSE(TransCompoundMap)
-TESTPAGEFLAG_FALSE(TransTail)
-PAGEFLAG_FALSE(DoubleMap)
-	TESTSCFLAG_FALSE(DoubleMap)
+TESTPAGEFLAG_FALSE(TransHuge, transhuge)
+TESTPAGEFLAG_FALSE(TransCompound, transcompound)
+TESTPAGEFLAG_FALSE(TransCompoundMap, transcompoundmap)
+TESTPAGEFLAG_FALSE(TransTail, transtail)
+PAGEFLAG_FALSE(DoubleMap, double_map)
+	TESTSCFLAG_FALSE(DoubleMap, double_map)
 #endif
 
 /*
@@ -903,6 +991,11 @@ static inline int page_has_private(struct page *page)
 	return !!(page->flags & PAGE_FLAGS_PRIVATE);
 }
 
+static inline bool folio_has_private(struct folio *folio)
+{
+	return page_has_private(&folio->page);
+}
+
 #undef PF_ANY
 #undef PF_HEAD
 #undef PF_ONLY_HEAD
-- 
2.30.2


  parent reply	other threads:[~2021-07-12  3:14 UTC|newest]

Thread overview: 153+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2021-07-12  3:04 [PATCH v13 000/137] Memory folios Matthew Wilcox (Oracle)
2021-07-12  3:04 ` [PATCH v13 001/137] mm: Convert get_page_unless_zero() to return bool Matthew Wilcox (Oracle)
2021-07-12  3:04 ` [PATCH v13 002/137] mm: Introduce struct folio Matthew Wilcox (Oracle)
2021-07-12  3:04 ` [PATCH v13 003/137] mm: Add folio_pgdat(), folio_zone() and folio_zonenum() Matthew Wilcox (Oracle)
2021-07-12  3:04 ` [PATCH v13 004/137] mm/vmstat: Add functions to account folio statistics Matthew Wilcox (Oracle)
2021-07-12  3:04 ` [PATCH v13 005/137] mm/debug: Add VM_BUG_ON_FOLIO() and VM_WARN_ON_ONCE_FOLIO() Matthew Wilcox (Oracle)
2021-07-12  3:04 ` [PATCH v13 006/137] mm: Add folio reference count functions Matthew Wilcox (Oracle)
2021-07-12  3:04 ` [PATCH v13 007/137] mm: Add folio_put() Matthew Wilcox (Oracle)
2021-07-12  3:04 ` [PATCH v13 008/137] mm: Add folio_get() Matthew Wilcox (Oracle)
2021-07-12  3:04 ` [PATCH v13 009/137] mm: Add folio_try_get_rcu() Matthew Wilcox (Oracle)
2021-07-12  3:04 ` Matthew Wilcox (Oracle) [this message]
2021-07-13  0:24   ` [PATCH v13 010/137] mm: Add folio flag manipulation functions Johannes Weiner
2021-07-13  2:15     ` Matthew Wilcox
2021-07-13  9:15       ` Peter Zijlstra
2021-07-13 15:55         ` Johannes Weiner
2021-07-14  1:55           ` Matthew Wilcox
2021-07-14  1:56           ` Andrew Morton
2021-07-14 14:03             ` Matthew Wilcox
2021-07-14  9:18         ` David Howells
2021-07-12  3:04 ` [PATCH v13 011/137] mm/lru: Add folio LRU functions Matthew Wilcox (Oracle)
2021-07-12  3:04 ` [PATCH v13 012/137] mm: Handle per-folio private data Matthew Wilcox (Oracle)
2021-07-12  3:04 ` [PATCH v13 013/137] mm/filemap: Add folio_index(), folio_file_page() and folio_contains() Matthew Wilcox (Oracle)
2021-07-12  3:04 ` [PATCH v13 014/137] mm/filemap: Add folio_next_index() Matthew Wilcox (Oracle)
2021-07-12  3:04 ` [PATCH v13 015/137] mm/filemap: Add folio_pos() and folio_file_pos() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 016/137] mm/util: Add folio_mapping() and folio_file_mapping() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 017/137] mm/filemap: Add folio_unlock() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 018/137] mm/filemap: Add folio_lock() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 019/137] mm/filemap: Add folio_lock_killable() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 020/137] mm/filemap: Add __folio_lock_async() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 021/137] mm/filemap: Add folio_wait_locked() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 022/137] mm/filemap: Add __folio_lock_or_retry() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 023/137] mm/swap: Add folio_rotate_reclaimable() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 024/137] mm/filemap: Add folio_end_writeback() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 025/137] mm/writeback: Add folio_wait_writeback() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 026/137] mm/writeback: Add folio_wait_stable() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 027/137] mm/filemap: Add folio_wait_bit() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 028/137] mm/filemap: Add folio_wake_bit() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 029/137] mm/filemap: Convert page wait queues to be folios Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 030/137] mm/filemap: Add folio private_2 functions Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 031/137] fs/netfs: Add folio fscache functions Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 032/137] mm: Add folio_mapped() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 033/137] mm: Add folio_nid() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 034/137] mm/memcg: Remove 'page' parameter to mem_cgroup_charge_statistics() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 035/137] mm/memcg: Use the node id in mem_cgroup_update_tree() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 036/137] mm/memcg: Remove soft_limit_tree_node() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 037/137] mm/memcg: Convert memcg_check_events to take a node ID Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 038/137] mm/memcg: Add folio_memcg() and related functions Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 039/137] mm/memcg: Convert commit_charge() to take a folio Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 040/137] mm/memcg: Convert mem_cgroup_charge() " Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 041/137] mm/memcg: Convert uncharge_page() to uncharge_folio() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 042/137] mm/memcg: Convert mem_cgroup_uncharge() to take a folio Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 043/137] mm/memcg: Convert mem_cgroup_migrate() to take folios Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 044/137] mm/memcg: Convert mem_cgroup_track_foreign_dirty_slowpath() to folio Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 045/137] mm/memcg: Add folio_memcg_lock() and folio_memcg_unlock() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 046/137] mm/memcg: Convert mem_cgroup_move_account() to use a folio Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 047/137] mm/memcg: Add folio_lruvec() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 048/137] mm/memcg: Add folio_lruvec_lock() and similar functions Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 049/137] mm/memcg: Add folio_lruvec_relock_irq() and folio_lruvec_relock_irqsave() Matthew Wilcox (Oracle)
2021-07-12  7:18   ` kernel test robot
2021-07-12  3:05 ` [PATCH v13 050/137] mm/workingset: Convert workingset_activation to take a folio Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 051/137] mm: Add folio_pfn() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 052/137] mm: Add folio_raw_mapping() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 053/137] mm: Add flush_dcache_folio() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 054/137] mm: Add kmap_local_folio() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 055/137] mm: Add arch_make_folio_accessible() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 056/137] mm: Add folio_young() and folio_idle() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 057/137] mm/swap: Add folio_activate() Matthew Wilcox (Oracle)
2021-07-12  7:01   ` kernel test robot
2021-07-12  3:05 ` [PATCH v13 058/137] mm/swap: Add folio_mark_accessed() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 059/137] mm/rmap: Add folio_mkclean() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 060/137] mm/migrate: Add folio_migrate_mapping() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 061/137] mm/migrate: Add folio_migrate_flags() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 062/137] mm/migrate: Add folio_migrate_copy() Matthew Wilcox (Oracle)
2021-07-12  7:58   ` kernel test robot
2021-07-12  3:05 ` [PATCH v13 063/137] mm/writeback: Rename __add_wb_stat() to wb_stat_mod() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 064/137] flex_proportions: Allow N events instead of 1 Matthew Wilcox (Oracle)
2021-07-13 14:39   ` Jan Kara
2021-07-12  3:05 ` [PATCH v13 065/137] mm/writeback: Change __wb_writeout_inc() to __wb_writeout_add() Matthew Wilcox (Oracle)
2021-07-13 14:40   ` Jan Kara
2021-07-12  3:05 ` [PATCH v13 066/137] mm/writeback: Add __folio_end_writeback() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 067/137] mm/writeback: Add folio_start_writeback() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 068/137] mm/writeback: Add folio_mark_dirty() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 069/137] mm/writeback: Add __folio_mark_dirty() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 070/137] mm/writeback: Add filemap_dirty_folio() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 071/137] mm/writeback: Add folio_account_cleaned() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 072/137] mm/writeback: Add folio_cancel_dirty() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 073/137] mm/writeback: Add folio_clear_dirty_for_io() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 074/137] mm/writeback: Add folio_account_redirty() Matthew Wilcox (Oracle)
2021-07-12  3:05 ` [PATCH v13 075/137] mm/writeback: Add folio_redirty_for_writepage() Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 076/137] mm/filemap: Add i_blocks_per_folio() Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 077/137] mm/filemap: Add folio_mkwrite_check_truncate() Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 078/137] mm/filemap: Add readahead_folio() Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 079/137] mm/workingset: Convert workingset_refault() to take a folio Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 080/137] mm: Add folio_evictable() Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 081/137] mm/lru: Convert __pagevec_lru_add_fn to take a folio Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 082/137] mm/lru: Add folio_add_lru() Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 083/137] mm/page_alloc: Add folio allocation functions Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 084/137] mm/filemap: Add filemap_alloc_folio Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 085/137] mm/filemap: Add filemap_add_folio() Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 086/137] mm/filemap: Convert mapping_get_entry to return a folio Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 087/137] mm/filemap: Add filemap_get_folio Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 088/137] mm/filemap: Add FGP_STABLE Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 089/137] block: Add bio_add_folio() Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 090/137] block: Add bio_for_each_folio_all() Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 091/137] iomap: Convert to_iomap_page to take a folio Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 092/137] iomap: Convert iomap_page_create " Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 093/137] iomap: Convert iomap_page_release " Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 094/137] iomap: Convert iomap_releasepage to use " Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 095/137] iomap: Convert iomap_invalidatepage " Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 096/137] iomap: Pass the iomap_page into iomap_set_range_uptodate Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 097/137] iomap: Use folio offsets instead of page offsets Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 098/137] iomap: Convert bio completions to use folios Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 099/137] iomap: Convert readahead and readpage to use a folio Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 100/137] iomap: Convert iomap_page_mkwrite " Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 101/137] iomap: Convert iomap_write_begin and iomap_write_end to folios Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 102/137] iomap: Convert iomap_read_inline_data to take a folio Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 103/137] iomap: Convert iomap_write_end_inline " Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 104/137] iomap: Convert iomap_add_to_ioend " Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 105/137] iomap: Convert iomap_do_writepage to use " Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 106/137] iomap: Convert iomap_migrate_page to use folios Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 107/137] mm/filemap: Convert page_cache_delete to take a folio Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 108/137] mm/filemap: Convert unaccount_page_cache_page to filemap_unaccount_folio Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 109/137] mm/filemap: Add filemap_remove_folio and __filemap_remove_folio Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 110/137] mm/filemap: Convert find_get_entry to return a folio Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 111/137] mm/filemap: Convert filemap_get_read_batch to use folios Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 112/137] mm/filemap: Convert find_get_pages_contig to folios Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 113/137] mm/filemap: Convert filemap_read_page to take a folio Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 114/137] mm/filemap: Convert filemap_create_page to folio Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 115/137] mm/filemap: Convert filemap_range_uptodate to folios Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 116/137] mm/filemap: Convert filemap_fault to folio Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 117/137] mm/filemap: Add read_cache_folio and read_mapping_folio Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 118/137] mm/filemap: Convert filemap_get_pages to use folios Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 119/137] mm/filemap: Convert page_cache_delete_batch to folios Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 120/137] mm/filemap: Remove PageHWPoison check from next_uptodate_page() Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 121/137] mm/filemap: Use folios in next_uptodate_page Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 122/137] mm/filemap: Use a folio in filemap_map_pages Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 123/137] fs: Convert vfs_dedupe_file_range_compare to folios Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 124/137] mm/truncate,shmem: Handle truncates that split THPs Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 125/137] mm/filemap: Return only head pages from find_get_entries Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 126/137] mm: Use multi-index entries in the page cache Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 127/137] iomap: Support multi-page folios in invalidatepage Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 128/137] xfs: Support THPs Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 129/137] mm/truncate: Convert invalidate_inode_pages2_range to folios Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 130/137] mm/truncate: Fix invalidate_complete_page2 for THPs Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 131/137] mm/vmscan: Free non-shmem THPs without splitting them Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 132/137] mm: Fix READ_ONLY_THP warning Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 133/137] mm: Support arbitrary THP sizes Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 134/137] mm/filemap: Allow multi-page folios to be added to the page cache Matthew Wilcox (Oracle)
2021-07-12  3:06 ` [PATCH v13 135/137] mm/vmscan: Optimise shrink_page_list for smaller THPs Matthew Wilcox (Oracle)
2021-07-12  3:07 ` [PATCH v13 136/137] mm/readahead: Convert page_cache_async_ra() to take a folio Matthew Wilcox (Oracle)
2021-07-12  3:07 ` [PATCH v13 137/137] mm/readahead: Add multi-page folio readahead Matthew Wilcox (Oracle)
2021-07-12  5:46 ` [PATCH v13 000/137] Memory folios Christoph Hellwig
2021-07-12 11:35   ` Matthew Wilcox

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20210712030701.4000097-11-willy@infradead.org \
    --to=willy@infradead.org \
    --cc=dhowells@redhat.com \
    --cc=hch@lst.de \
    --cc=jlayton@kernel.org \
    --cc=kirill.shutemov@linux.intel.com \
    --cc=linux-fsdevel@vger.kernel.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-mm@kvack.org \
    --cc=vbabka@suse.cz \
    --cc=william.kucharski@oracle.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.