All of lore.kernel.org
 help / color / mirror / Atom feed
From: "Vishal Moola (Oracle)" <vishal.moola@gmail.com>
To: linux-fsdevel@vger.kernel.org
Cc: linux-afs@lists.infradead.org, linux-kernel@vger.kernel.org,
	linux-btrfs@vger.kernel.org, ceph-devel@vger.kernel.org,
	linux-cifs@vger.kernel.org, linux-ext4@vger.kernel.org,
	linux-f2fs-devel@lists.sourceforge.net, cluster-devel@redhat.com,
	linux-nilfs@vger.kernel.org, linux-mm@kvack.org,
	"Vishal Moola (Oracle)" <vishal.moola@gmail.com>,
	Matthew Wilcow <willy@infradead.org>
Subject: [PATCH v4 04/23] page-writeback: Convert write_cache_pages() to use filemap_get_folios_tag()
Date: Wed,  2 Nov 2022 09:10:12 -0700	[thread overview]
Message-ID: <20221102161031.5820-5-vishal.moola@gmail.com> (raw)
In-Reply-To: <20221102161031.5820-1-vishal.moola@gmail.com>

Converted function to use folios throughout. This is in preparation for
the removal of find_get_pages_range_tag(). This change removes 8 calls
to compound_head(), and the function now supports large folios.

Signed-off-by: Vishal Moola (Oracle) <vishal.moola@gmail.com>
Reviewed-by: Matthew Wilcow (Oracle) <willy@infradead.org>
---
 mm/page-writeback.c | 44 +++++++++++++++++++++++---------------------
 1 file changed, 23 insertions(+), 21 deletions(-)

diff --git a/mm/page-writeback.c b/mm/page-writeback.c
index 7e9d8d857ecc..aeec8b196232 100644
--- a/mm/page-writeback.c
+++ b/mm/page-writeback.c
@@ -2286,15 +2286,15 @@ int write_cache_pages(struct address_space *mapping,
 	int ret = 0;
 	int done = 0;
 	int error;
-	struct pagevec pvec;
-	int nr_pages;
+	struct folio_batch fbatch;
+	int nr_folios;
 	pgoff_t index;
 	pgoff_t end;		/* Inclusive */
 	pgoff_t done_index;
 	int range_whole = 0;
 	xa_mark_t tag;
 
-	pagevec_init(&pvec);
+	folio_batch_init(&fbatch);
 	if (wbc->range_cyclic) {
 		index = mapping->writeback_index; /* prev offset */
 		end = -1;
@@ -2314,17 +2314,18 @@ int write_cache_pages(struct address_space *mapping,
 	while (!done && (index <= end)) {
 		int i;
 
-		nr_pages = pagevec_lookup_range_tag(&pvec, mapping, &index, end,
-				tag);
-		if (nr_pages == 0)
+		nr_folios = filemap_get_folios_tag(mapping, &index, end,
+				tag, &fbatch);
+
+		if (nr_folios == 0)
 			break;
 
-		for (i = 0; i < nr_pages; i++) {
-			struct page *page = pvec.pages[i];
+		for (i = 0; i < nr_folios; i++) {
+			struct folio *folio = fbatch.folios[i];
 
-			done_index = page->index;
+			done_index = folio->index;
 
-			lock_page(page);
+			folio_lock(folio);
 
 			/*
 			 * Page truncated or invalidated. We can freely skip it
@@ -2334,30 +2335,30 @@ int write_cache_pages(struct address_space *mapping,
 			 * even if there is now a new, dirty page at the same
 			 * pagecache address.
 			 */
-			if (unlikely(page->mapping != mapping)) {
+			if (unlikely(folio->mapping != mapping)) {
 continue_unlock:
-				unlock_page(page);
+				folio_unlock(folio);
 				continue;
 			}
 
-			if (!PageDirty(page)) {
+			if (!folio_test_dirty(folio)) {
 				/* someone wrote it for us */
 				goto continue_unlock;
 			}
 
-			if (PageWriteback(page)) {
+			if (folio_test_writeback(folio)) {
 				if (wbc->sync_mode != WB_SYNC_NONE)
-					wait_on_page_writeback(page);
+					folio_wait_writeback(folio);
 				else
 					goto continue_unlock;
 			}
 
-			BUG_ON(PageWriteback(page));
-			if (!clear_page_dirty_for_io(page))
+			BUG_ON(folio_test_writeback(folio));
+			if (!folio_clear_dirty_for_io(folio))
 				goto continue_unlock;
 
 			trace_wbc_writepage(wbc, inode_to_bdi(mapping->host));
-			error = (*writepage)(page, wbc, data);
+			error = writepage(&folio->page, wbc, data);
 			if (unlikely(error)) {
 				/*
 				 * Handle errors according to the type of
@@ -2372,11 +2373,12 @@ int write_cache_pages(struct address_space *mapping,
 				 * the first error.
 				 */
 				if (error == AOP_WRITEPAGE_ACTIVATE) {
-					unlock_page(page);
+					folio_unlock(folio);
 					error = 0;
 				} else if (wbc->sync_mode != WB_SYNC_ALL) {
 					ret = error;
-					done_index = page->index + 1;
+					done_index = folio->index +
+						folio_nr_pages(folio);
 					done = 1;
 					break;
 				}
@@ -2396,7 +2398,7 @@ int write_cache_pages(struct address_space *mapping,
 				break;
 			}
 		}
-		pagevec_release(&pvec);
+		folio_batch_release(&fbatch);
 		cond_resched();
 	}
 
-- 
2.38.1


WARNING: multiple messages have this Message-ID (diff)
From: "Vishal Moola (Oracle)" <vishal.moola@gmail.com>
To: linux-fsdevel@vger.kernel.org
Cc: linux-cifs@vger.kernel.org, linux-nilfs@vger.kernel.org,
	"Vishal Moola \(Oracle\)" <vishal.moola@gmail.com>,
	linux-kernel@vger.kernel.org,
	Matthew Wilcow <willy@infradead.org>,
	linux-f2fs-devel@lists.sourceforge.net, cluster-devel@redhat.com,
	linux-mm@kvack.org, ceph-devel@vger.kernel.org,
	linux-ext4@vger.kernel.org, linux-afs@lists.infradead.org,
	linux-btrfs@vger.kernel.org
Subject: [f2fs-dev] [PATCH v4 04/23] page-writeback: Convert write_cache_pages() to use filemap_get_folios_tag()
Date: Wed,  2 Nov 2022 09:10:12 -0700	[thread overview]
Message-ID: <20221102161031.5820-5-vishal.moola@gmail.com> (raw)
In-Reply-To: <20221102161031.5820-1-vishal.moola@gmail.com>

Converted function to use folios throughout. This is in preparation for
the removal of find_get_pages_range_tag(). This change removes 8 calls
to compound_head(), and the function now supports large folios.

Signed-off-by: Vishal Moola (Oracle) <vishal.moola@gmail.com>
Reviewed-by: Matthew Wilcow (Oracle) <willy@infradead.org>
---
 mm/page-writeback.c | 44 +++++++++++++++++++++++---------------------
 1 file changed, 23 insertions(+), 21 deletions(-)

diff --git a/mm/page-writeback.c b/mm/page-writeback.c
index 7e9d8d857ecc..aeec8b196232 100644
--- a/mm/page-writeback.c
+++ b/mm/page-writeback.c
@@ -2286,15 +2286,15 @@ int write_cache_pages(struct address_space *mapping,
 	int ret = 0;
 	int done = 0;
 	int error;
-	struct pagevec pvec;
-	int nr_pages;
+	struct folio_batch fbatch;
+	int nr_folios;
 	pgoff_t index;
 	pgoff_t end;		/* Inclusive */
 	pgoff_t done_index;
 	int range_whole = 0;
 	xa_mark_t tag;
 
-	pagevec_init(&pvec);
+	folio_batch_init(&fbatch);
 	if (wbc->range_cyclic) {
 		index = mapping->writeback_index; /* prev offset */
 		end = -1;
@@ -2314,17 +2314,18 @@ int write_cache_pages(struct address_space *mapping,
 	while (!done && (index <= end)) {
 		int i;
 
-		nr_pages = pagevec_lookup_range_tag(&pvec, mapping, &index, end,
-				tag);
-		if (nr_pages == 0)
+		nr_folios = filemap_get_folios_tag(mapping, &index, end,
+				tag, &fbatch);
+
+		if (nr_folios == 0)
 			break;
 
-		for (i = 0; i < nr_pages; i++) {
-			struct page *page = pvec.pages[i];
+		for (i = 0; i < nr_folios; i++) {
+			struct folio *folio = fbatch.folios[i];
 
-			done_index = page->index;
+			done_index = folio->index;
 
-			lock_page(page);
+			folio_lock(folio);
 
 			/*
 			 * Page truncated or invalidated. We can freely skip it
@@ -2334,30 +2335,30 @@ int write_cache_pages(struct address_space *mapping,
 			 * even if there is now a new, dirty page at the same
 			 * pagecache address.
 			 */
-			if (unlikely(page->mapping != mapping)) {
+			if (unlikely(folio->mapping != mapping)) {
 continue_unlock:
-				unlock_page(page);
+				folio_unlock(folio);
 				continue;
 			}
 
-			if (!PageDirty(page)) {
+			if (!folio_test_dirty(folio)) {
 				/* someone wrote it for us */
 				goto continue_unlock;
 			}
 
-			if (PageWriteback(page)) {
+			if (folio_test_writeback(folio)) {
 				if (wbc->sync_mode != WB_SYNC_NONE)
-					wait_on_page_writeback(page);
+					folio_wait_writeback(folio);
 				else
 					goto continue_unlock;
 			}
 
-			BUG_ON(PageWriteback(page));
-			if (!clear_page_dirty_for_io(page))
+			BUG_ON(folio_test_writeback(folio));
+			if (!folio_clear_dirty_for_io(folio))
 				goto continue_unlock;
 
 			trace_wbc_writepage(wbc, inode_to_bdi(mapping->host));
-			error = (*writepage)(page, wbc, data);
+			error = writepage(&folio->page, wbc, data);
 			if (unlikely(error)) {
 				/*
 				 * Handle errors according to the type of
@@ -2372,11 +2373,12 @@ int write_cache_pages(struct address_space *mapping,
 				 * the first error.
 				 */
 				if (error == AOP_WRITEPAGE_ACTIVATE) {
-					unlock_page(page);
+					folio_unlock(folio);
 					error = 0;
 				} else if (wbc->sync_mode != WB_SYNC_ALL) {
 					ret = error;
-					done_index = page->index + 1;
+					done_index = folio->index +
+						folio_nr_pages(folio);
 					done = 1;
 					break;
 				}
@@ -2396,7 +2398,7 @@ int write_cache_pages(struct address_space *mapping,
 				break;
 			}
 		}
-		pagevec_release(&pvec);
+		folio_batch_release(&fbatch);
 		cond_resched();
 	}
 
-- 
2.38.1



_______________________________________________
Linux-f2fs-devel mailing list
Linux-f2fs-devel@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/linux-f2fs-devel

WARNING: multiple messages have this Message-ID (diff)
From: Vishal Moola (Oracle) <vishal.moola@gmail.com>
To: cluster-devel.redhat.com
Subject: [Cluster-devel] [PATCH v4 04/23] page-writeback: Convert write_cache_pages() to use filemap_get_folios_tag()
Date: Wed,  2 Nov 2022 09:10:12 -0700	[thread overview]
Message-ID: <20221102161031.5820-5-vishal.moola@gmail.com> (raw)
In-Reply-To: <20221102161031.5820-1-vishal.moola@gmail.com>

Converted function to use folios throughout. This is in preparation for
the removal of find_get_pages_range_tag(). This change removes 8 calls
to compound_head(), and the function now supports large folios.

Signed-off-by: Vishal Moola (Oracle) <vishal.moola@gmail.com>
Reviewed-by: Matthew Wilcow (Oracle) <willy@infradead.org>
---
 mm/page-writeback.c | 44 +++++++++++++++++++++++---------------------
 1 file changed, 23 insertions(+), 21 deletions(-)

diff --git a/mm/page-writeback.c b/mm/page-writeback.c
index 7e9d8d857ecc..aeec8b196232 100644
--- a/mm/page-writeback.c
+++ b/mm/page-writeback.c
@@ -2286,15 +2286,15 @@ int write_cache_pages(struct address_space *mapping,
 	int ret = 0;
 	int done = 0;
 	int error;
-	struct pagevec pvec;
-	int nr_pages;
+	struct folio_batch fbatch;
+	int nr_folios;
 	pgoff_t index;
 	pgoff_t end;		/* Inclusive */
 	pgoff_t done_index;
 	int range_whole = 0;
 	xa_mark_t tag;
 
-	pagevec_init(&pvec);
+	folio_batch_init(&fbatch);
 	if (wbc->range_cyclic) {
 		index = mapping->writeback_index; /* prev offset */
 		end = -1;
@@ -2314,17 +2314,18 @@ int write_cache_pages(struct address_space *mapping,
 	while (!done && (index <= end)) {
 		int i;
 
-		nr_pages = pagevec_lookup_range_tag(&pvec, mapping, &index, end,
-				tag);
-		if (nr_pages == 0)
+		nr_folios = filemap_get_folios_tag(mapping, &index, end,
+				tag, &fbatch);
+
+		if (nr_folios == 0)
 			break;
 
-		for (i = 0; i < nr_pages; i++) {
-			struct page *page = pvec.pages[i];
+		for (i = 0; i < nr_folios; i++) {
+			struct folio *folio = fbatch.folios[i];
 
-			done_index = page->index;
+			done_index = folio->index;
 
-			lock_page(page);
+			folio_lock(folio);
 
 			/*
 			 * Page truncated or invalidated. We can freely skip it
@@ -2334,30 +2335,30 @@ int write_cache_pages(struct address_space *mapping,
 			 * even if there is now a new, dirty page at the same
 			 * pagecache address.
 			 */
-			if (unlikely(page->mapping != mapping)) {
+			if (unlikely(folio->mapping != mapping)) {
 continue_unlock:
-				unlock_page(page);
+				folio_unlock(folio);
 				continue;
 			}
 
-			if (!PageDirty(page)) {
+			if (!folio_test_dirty(folio)) {
 				/* someone wrote it for us */
 				goto continue_unlock;
 			}
 
-			if (PageWriteback(page)) {
+			if (folio_test_writeback(folio)) {
 				if (wbc->sync_mode != WB_SYNC_NONE)
-					wait_on_page_writeback(page);
+					folio_wait_writeback(folio);
 				else
 					goto continue_unlock;
 			}
 
-			BUG_ON(PageWriteback(page));
-			if (!clear_page_dirty_for_io(page))
+			BUG_ON(folio_test_writeback(folio));
+			if (!folio_clear_dirty_for_io(folio))
 				goto continue_unlock;
 
 			trace_wbc_writepage(wbc, inode_to_bdi(mapping->host));
-			error = (*writepage)(page, wbc, data);
+			error = writepage(&folio->page, wbc, data);
 			if (unlikely(error)) {
 				/*
 				 * Handle errors according to the type of
@@ -2372,11 +2373,12 @@ int write_cache_pages(struct address_space *mapping,
 				 * the first error.
 				 */
 				if (error == AOP_WRITEPAGE_ACTIVATE) {
-					unlock_page(page);
+					folio_unlock(folio);
 					error = 0;
 				} else if (wbc->sync_mode != WB_SYNC_ALL) {
 					ret = error;
-					done_index = page->index + 1;
+					done_index = folio->index +
+						folio_nr_pages(folio);
 					done = 1;
 					break;
 				}
@@ -2396,7 +2398,7 @@ int write_cache_pages(struct address_space *mapping,
 				break;
 			}
 		}
-		pagevec_release(&pvec);
+		folio_batch_release(&fbatch);
 		cond_resched();
 	}
 
-- 
2.38.1


WARNING: multiple messages have this Message-ID (diff)
From: "Vishal Moola (Oracle)" <vishal.moola@gmail.com>
To: linux-fsdevel@vger.kernel.org
Cc: linux-cifs@vger.kernel.org, linux-nilfs@vger.kernel.org,
	"Vishal Moola (Oracle)" <vishal.moola@gmail.com>,
	linux-kernel@vger.kernel.org,
	Matthew Wilcow <willy@infradead.org>,
	linux-f2fs-devel@lists.sourceforge.net, cluster-devel@redhat.com,
	linux-mm@kvack.org, ceph-devel@vger.kernel.org,
	linux-ext4@vger.kernel.org, linux-afs@lists.infradead.org,
	linux-btrfs@vger.kernel.org
Subject: [PATCH v4 04/23] page-writeback: Convert write_cache_pages() to use filemap_get_folios_tag()
Date: Wed,  2 Nov 2022 09:10:12 -0700	[thread overview]
Message-ID: <20221102161031.5820-5-vishal.moola@gmail.com> (raw)
In-Reply-To: <20221102161031.5820-1-vishal.moola@gmail.com>

Converted function to use folios throughout. This is in preparation for
the removal of find_get_pages_range_tag(). This change removes 8 calls
to compound_head(), and the function now supports large folios.

Signed-off-by: Vishal Moola (Oracle) <vishal.moola@gmail.com>
Reviewed-by: Matthew Wilcow (Oracle) <willy@infradead.org>
---
 mm/page-writeback.c | 44 +++++++++++++++++++++++---------------------
 1 file changed, 23 insertions(+), 21 deletions(-)

diff --git a/mm/page-writeback.c b/mm/page-writeback.c
index 7e9d8d857ecc..aeec8b196232 100644
--- a/mm/page-writeback.c
+++ b/mm/page-writeback.c
@@ -2286,15 +2286,15 @@ int write_cache_pages(struct address_space *mapping,
 	int ret = 0;
 	int done = 0;
 	int error;
-	struct pagevec pvec;
-	int nr_pages;
+	struct folio_batch fbatch;
+	int nr_folios;
 	pgoff_t index;
 	pgoff_t end;		/* Inclusive */
 	pgoff_t done_index;
 	int range_whole = 0;
 	xa_mark_t tag;
 
-	pagevec_init(&pvec);
+	folio_batch_init(&fbatch);
 	if (wbc->range_cyclic) {
 		index = mapping->writeback_index; /* prev offset */
 		end = -1;
@@ -2314,17 +2314,18 @@ int write_cache_pages(struct address_space *mapping,
 	while (!done && (index <= end)) {
 		int i;
 
-		nr_pages = pagevec_lookup_range_tag(&pvec, mapping, &index, end,
-				tag);
-		if (nr_pages == 0)
+		nr_folios = filemap_get_folios_tag(mapping, &index, end,
+				tag, &fbatch);
+
+		if (nr_folios == 0)
 			break;
 
-		for (i = 0; i < nr_pages; i++) {
-			struct page *page = pvec.pages[i];
+		for (i = 0; i < nr_folios; i++) {
+			struct folio *folio = fbatch.folios[i];
 
-			done_index = page->index;
+			done_index = folio->index;
 
-			lock_page(page);
+			folio_lock(folio);
 
 			/*
 			 * Page truncated or invalidated. We can freely skip it
@@ -2334,30 +2335,30 @@ int write_cache_pages(struct address_space *mapping,
 			 * even if there is now a new, dirty page at the same
 			 * pagecache address.
 			 */
-			if (unlikely(page->mapping != mapping)) {
+			if (unlikely(folio->mapping != mapping)) {
 continue_unlock:
-				unlock_page(page);
+				folio_unlock(folio);
 				continue;
 			}
 
-			if (!PageDirty(page)) {
+			if (!folio_test_dirty(folio)) {
 				/* someone wrote it for us */
 				goto continue_unlock;
 			}
 
-			if (PageWriteback(page)) {
+			if (folio_test_writeback(folio)) {
 				if (wbc->sync_mode != WB_SYNC_NONE)
-					wait_on_page_writeback(page);
+					folio_wait_writeback(folio);
 				else
 					goto continue_unlock;
 			}
 
-			BUG_ON(PageWriteback(page));
-			if (!clear_page_dirty_for_io(page))
+			BUG_ON(folio_test_writeback(folio));
+			if (!folio_clear_dirty_for_io(folio))
 				goto continue_unlock;
 
 			trace_wbc_writepage(wbc, inode_to_bdi(mapping->host));
-			error = (*writepage)(page, wbc, data);
+			error = writepage(&folio->page, wbc, data);
 			if (unlikely(error)) {
 				/*
 				 * Handle errors according to the type of
@@ -2372,11 +2373,12 @@ int write_cache_pages(struct address_space *mapping,
 				 * the first error.
 				 */
 				if (error == AOP_WRITEPAGE_ACTIVATE) {
-					unlock_page(page);
+					folio_unlock(folio);
 					error = 0;
 				} else if (wbc->sync_mode != WB_SYNC_ALL) {
 					ret = error;
-					done_index = page->index + 1;
+					done_index = folio->index +
+						folio_nr_pages(folio);
 					done = 1;
 					break;
 				}
@@ -2396,7 +2398,7 @@ int write_cache_pages(struct address_space *mapping,
 				break;
 			}
 		}
-		pagevec_release(&pvec);
+		folio_batch_release(&fbatch);
 		cond_resched();
 	}
 
-- 
2.38.1


  parent reply	other threads:[~2022-11-02 16:11 UTC|newest]

Thread overview: 110+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2022-11-02 16:10 [PATCH v4 00/23] Convert to filemap_get_folios_tag() Vishal Moola (Oracle)
2022-11-02 16:10 ` [Cluster-devel] " Vishal Moola
2022-11-02 16:10 ` [f2fs-dev] " Vishal Moola (Oracle)
2022-11-02 16:10 ` [PATCH v4 01/23] pagemap: Add filemap_grab_folio() Vishal Moola (Oracle)
2022-11-02 16:10   ` [Cluster-devel] " Vishal Moola
2022-11-02 16:10   ` [f2fs-dev] " Vishal Moola (Oracle)
2022-11-02 16:10 ` [PATCH v4 02/23] filemap: Added filemap_get_folios_tag() Vishal Moola (Oracle)
2022-11-02 16:10   ` [Cluster-devel] " Vishal Moola
2022-11-02 16:10   ` [f2fs-dev] " Vishal Moola (Oracle)
2022-11-02 16:10 ` [PATCH v4 03/23] filemap: Convert __filemap_fdatawait_range() to use filemap_get_folios_tag() Vishal Moola (Oracle)
2022-11-02 16:10   ` Vishal Moola (Oracle)
2022-11-02 16:10   ` [Cluster-devel] " Vishal Moola
2022-11-02 16:10   ` [f2fs-dev] " Vishal Moola (Oracle)
2022-11-02 16:10 ` Vishal Moola (Oracle) [this message]
2022-11-02 16:10   ` [PATCH v4 04/23] page-writeback: Convert write_cache_pages() " Vishal Moola (Oracle)
2022-11-02 16:10   ` [Cluster-devel] " Vishal Moola
2022-11-02 16:10   ` [f2fs-dev] " Vishal Moola (Oracle)
2022-11-02 16:10 ` [PATCH v4 05/23] afs: Convert afs_writepages_region() " Vishal Moola (Oracle)
2022-11-02 16:10   ` Vishal Moola (Oracle)
2022-11-02 16:10   ` [Cluster-devel] " Vishal Moola
2022-11-02 16:10   ` [f2fs-dev] " Vishal Moola (Oracle)
2022-11-02 16:10 ` [PATCH v4 06/23] btrfs: Convert btree_write_cache_pages() to use filemap_get_folio_tag() Vishal Moola (Oracle)
2022-11-02 16:10   ` Vishal Moola (Oracle)
2022-11-02 16:10   ` [Cluster-devel] " Vishal Moola
2022-11-02 16:10   ` [f2fs-dev] " Vishal Moola (Oracle)
2022-11-02 16:10 ` [PATCH v4 07/23] btrfs: Convert extent_write_cache_pages() to use filemap_get_folios_tag() Vishal Moola (Oracle)
2022-11-02 16:10   ` Vishal Moola (Oracle)
2022-11-02 16:10   ` [Cluster-devel] " Vishal Moola
2022-11-02 16:10   ` [f2fs-dev] " Vishal Moola (Oracle)
2022-11-02 16:10 ` [PATCH v4 08/23] ceph: Convert ceph_writepages_start() " Vishal Moola (Oracle)
2022-11-02 16:10   ` Vishal Moola (Oracle)
2022-11-02 16:10   ` [Cluster-devel] " Vishal Moola
2022-11-02 16:10   ` [f2fs-dev] " Vishal Moola (Oracle)
2022-11-02 16:10 ` [PATCH v4 09/23] cifs: Convert wdata_alloc_and_fillpages() " Vishal Moola (Oracle)
2022-11-02 16:10   ` Vishal Moola (Oracle)
2022-11-02 16:10   ` [Cluster-devel] " Vishal Moola
2022-11-02 16:10   ` [f2fs-dev] " Vishal Moola (Oracle)
2022-11-02 16:10 ` [PATCH v4 10/23] ext4: Convert mpage_prepare_extent_to_map() " Vishal Moola (Oracle)
2022-11-02 16:10   ` Vishal Moola (Oracle)
2022-11-02 16:10   ` [Cluster-devel] " Vishal Moola
2022-11-02 16:10   ` [f2fs-dev] " Vishal Moola (Oracle)
2022-11-02 16:10 ` [PATCH v4 11/23] f2fs: Convert f2fs_fsync_node_pages() " Vishal Moola (Oracle)
2022-11-02 16:10   ` Vishal Moola (Oracle)
2022-11-02 16:10   ` [Cluster-devel] " Vishal Moola
2022-11-02 16:10   ` [f2fs-dev] " Vishal Moola (Oracle)
2022-11-02 16:10 ` [PATCH v4 12/23] f2fs: Convert f2fs_flush_inline_data() " Vishal Moola (Oracle)
2022-11-02 16:10   ` Vishal Moola (Oracle)
2022-11-02 16:10   ` [Cluster-devel] " Vishal Moola
2022-11-02 16:10   ` [f2fs-dev] " Vishal Moola (Oracle)
2022-11-02 16:10 ` [PATCH v4 13/23] f2fs: Convert f2fs_sync_node_pages() " Vishal Moola (Oracle)
2022-11-02 16:10   ` Vishal Moola (Oracle)
2022-11-02 16:10   ` [Cluster-devel] " Vishal Moola
2022-11-02 16:10   ` [f2fs-dev] " Vishal Moola (Oracle)
2022-11-02 16:10 ` [PATCH v4 14/23] f2fs: Convert f2fs_write_cache_pages() " Vishal Moola (Oracle)
2022-11-02 16:10   ` Vishal Moola (Oracle)
2022-11-02 16:10   ` [Cluster-devel] " Vishal Moola
2022-11-02 16:10   ` [f2fs-dev] " Vishal Moola (Oracle)
2022-11-02 16:10 ` [PATCH v4 15/23] f2fs: Convert last_fsync_dnode() " Vishal Moola (Oracle)
2022-11-02 16:10   ` Vishal Moola (Oracle)
2022-11-02 16:10   ` [Cluster-devel] " Vishal Moola
2022-11-02 16:10   ` [f2fs-dev] " Vishal Moola (Oracle)
2022-12-11  5:54   ` Chao Yu
2022-12-11  5:54     ` [Cluster-devel] " Chao Yu
2022-12-11  5:54     ` Chao Yu
2022-11-02 16:10 ` [PATCH v4 16/23] f2fs: Convert f2fs_sync_meta_pages() " Vishal Moola (Oracle)
2022-11-02 16:10   ` Vishal Moola (Oracle)
2022-11-02 16:10   ` [Cluster-devel] " Vishal Moola
2022-11-02 16:10   ` [f2fs-dev] " Vishal Moola (Oracle)
2022-12-11  6:03   ` Chao Yu
2022-12-11  6:03     ` [Cluster-devel] " Chao Yu
2022-12-11  6:03     ` Chao Yu
2022-11-02 16:10 ` [PATCH v4 17/23] gfs2: Convert gfs2_write_cache_jdata() " Vishal Moola (Oracle)
2022-11-02 16:10   ` Vishal Moola (Oracle)
2022-11-02 16:10   ` [Cluster-devel] " Vishal Moola
2022-11-02 16:10   ` [f2fs-dev] " Vishal Moola (Oracle)
2022-11-02 16:10 ` [PATCH v4 18/23] nilfs2: Convert nilfs_lookup_dirty_data_buffers() " Vishal Moola (Oracle)
2022-11-02 16:10   ` Vishal Moola (Oracle)
2022-11-02 16:10   ` [Cluster-devel] " Vishal Moola
2022-11-02 16:10   ` [f2fs-dev] " Vishal Moola (Oracle)
2022-11-02 16:10 ` [PATCH v4 19/23] nilfs2: Convert nilfs_lookup_dirty_node_buffers() " Vishal Moola (Oracle)
2022-11-02 16:10   ` Vishal Moola (Oracle)
2022-11-02 16:10   ` [Cluster-devel] " Vishal Moola
2022-11-02 16:10   ` [f2fs-dev] " Vishal Moola (Oracle)
2022-11-02 16:10 ` [PATCH v4 20/23] nilfs2: Convert nilfs_btree_lookup_dirty_buffers() " Vishal Moola (Oracle)
2022-11-02 16:10   ` Vishal Moola (Oracle)
2022-11-02 16:10   ` [Cluster-devel] " Vishal Moola
2022-11-02 16:10   ` [f2fs-dev] " Vishal Moola (Oracle)
2022-11-02 16:10 ` [PATCH v4 21/23] nilfs2: Convert nilfs_copy_dirty_pages() " Vishal Moola (Oracle)
2022-11-02 16:10   ` Vishal Moola (Oracle)
2022-11-02 16:10   ` [Cluster-devel] " Vishal Moola
2022-11-02 16:10   ` [f2fs-dev] " Vishal Moola (Oracle)
2022-11-02 16:10 ` [PATCH v4 22/23] nilfs2: Convert nilfs_clear_dirty_pages() " Vishal Moola (Oracle)
2022-11-02 16:10   ` Vishal Moola (Oracle)
2022-11-02 16:10   ` [Cluster-devel] " Vishal Moola
2022-11-02 16:10   ` [f2fs-dev] " Vishal Moola (Oracle)
2022-11-02 16:10 ` [PATCH v4 23/23] filemap: Remove find_get_pages_range_tag() Vishal Moola (Oracle)
2022-11-02 16:10   ` Vishal Moola (Oracle)
2022-11-02 16:10   ` [Cluster-devel] " Vishal Moola
2022-11-02 16:10   ` [f2fs-dev] " Vishal Moola (Oracle)
2022-11-03  7:08 ` [PATCH v4 00/23] Convert to filemap_get_folios_tag() Dave Chinner
2022-11-03  7:08   ` Dave Chinner
2022-11-03  7:08   ` [Cluster-devel] " Dave Chinner
2022-11-03  7:08   ` [f2fs-dev] " Dave Chinner
2022-11-03 16:38   ` Vishal Moola
2022-11-03 16:38     ` Vishal Moola
2022-11-03 16:38     ` [Cluster-devel] " Vishal Moola
2022-11-03 16:38     ` [f2fs-dev] " Vishal Moola
2022-11-03 20:47     ` Dave Chinner
2022-11-03 20:47       ` [Cluster-devel] " Dave Chinner
2022-11-03 20:47       ` [f2fs-dev] " Dave Chinner

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20221102161031.5820-5-vishal.moola@gmail.com \
    --to=vishal.moola@gmail.com \
    --cc=ceph-devel@vger.kernel.org \
    --cc=cluster-devel@redhat.com \
    --cc=linux-afs@lists.infradead.org \
    --cc=linux-btrfs@vger.kernel.org \
    --cc=linux-cifs@vger.kernel.org \
    --cc=linux-ext4@vger.kernel.org \
    --cc=linux-f2fs-devel@lists.sourceforge.net \
    --cc=linux-fsdevel@vger.kernel.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-mm@kvack.org \
    --cc=linux-nilfs@vger.kernel.org \
    --cc=willy@infradead.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.