All of lore.kernel.org
 help / color / mirror / Atom feed
From: Luis Chamberlain <mcgrof@kernel.org>
To: akpm@linux-foundation.org, willy@infradead.org,
	djwong@kernel.org, brauner@kernel.org, david@fromorbit.com,
	chandan.babu@oracle.com
Cc: hare@suse.de, ritesh.list@gmail.com, john.g.garry@oracle.com,
	ziy@nvidia.com, linux-fsdevel@vger.kernel.org,
	linux-xfs@vger.kernel.org, linux-mm@kvack.org,
	linux-block@vger.kernel.org, gost.dev@samsung.com,
	p.raghav@samsung.com, kernel@pankajraghav.com, mcgrof@kernel.org
Subject: [PATCH v5 03/11] filemap: allocate mapping_min_order folios in the page cache
Date: Fri,  3 May 2024 02:53:45 -0700	[thread overview]
Message-ID: <20240503095353.3798063-4-mcgrof@kernel.org> (raw)
In-Reply-To: <20240503095353.3798063-1-mcgrof@kernel.org>

filemap_create_folio() and do_read_cache_folio() were always allocating
folio of order 0. __filemap_get_folio was trying to allocate higher
order folios when fgp_flags had higher order hint set but it will default
to order 0 folio if higher order memory allocation fails.

Supporting mapping_min_order implies that we guarantee each folio in the
page cache has at least an order of mapping_min_order. When adding new
folios to the page cache we must also ensure the index used is aligned to
the mapping_min_order as the page cache requires the index to be aligned
to the order of the folio.

Co-developed-by: Pankaj Raghav <p.raghav@samsung.com>
Signed-off-by: Pankaj Raghav <p.raghav@samsung.com>
Reviewed-by: Hannes Reinecke <hare@suse.de>
Reviewed-by: Darrick J. Wong <djwong@kernel.org>
Signed-off-by: Luis Chamberlain <mcgrof@kernel.org>
---
 mm/filemap.c | 24 +++++++++++++++++-------
 1 file changed, 17 insertions(+), 7 deletions(-)

diff --git a/mm/filemap.c b/mm/filemap.c
index 30de18c4fd28..f0c0cfbbd134 100644
--- a/mm/filemap.c
+++ b/mm/filemap.c
@@ -858,6 +858,8 @@ noinline int __filemap_add_folio(struct address_space *mapping,
 
 	VM_BUG_ON_FOLIO(!folio_test_locked(folio), folio);
 	VM_BUG_ON_FOLIO(folio_test_swapbacked(folio), folio);
+	VM_BUG_ON_FOLIO(folio_order(folio) < mapping_min_folio_order(mapping),
+			folio);
 	mapping_set_update(&xas, mapping);
 
 	if (!huge) {
@@ -1895,8 +1897,10 @@ struct folio *__filemap_get_folio(struct address_space *mapping, pgoff_t index,
 		folio_wait_stable(folio);
 no_page:
 	if (!folio && (fgp_flags & FGP_CREAT)) {
-		unsigned order = FGF_GET_ORDER(fgp_flags);
+		unsigned int min_order = mapping_min_folio_order(mapping);
+		unsigned int order = max(min_order, FGF_GET_ORDER(fgp_flags));
 		int err;
+		index = mapping_align_start_index(mapping, index);
 
 		if ((fgp_flags & FGP_WRITE) && mapping_can_writeback(mapping))
 			gfp |= __GFP_WRITE;
@@ -1936,7 +1940,7 @@ struct folio *__filemap_get_folio(struct address_space *mapping, pgoff_t index,
 				break;
 			folio_put(folio);
 			folio = NULL;
-		} while (order-- > 0);
+		} while (order-- > min_order);
 
 		if (err == -EEXIST)
 			goto repeat;
@@ -2425,13 +2429,16 @@ static int filemap_update_page(struct kiocb *iocb,
 }
 
 static int filemap_create_folio(struct file *file,
-		struct address_space *mapping, pgoff_t index,
+		struct address_space *mapping, loff_t pos,
 		struct folio_batch *fbatch)
 {
 	struct folio *folio;
 	int error;
+	unsigned int min_order = mapping_min_folio_order(mapping);
+	pgoff_t index;
 
-	folio = filemap_alloc_folio(mapping_gfp_mask(mapping), 0);
+	folio = filemap_alloc_folio(mapping_gfp_mask(mapping),
+				    min_order);
 	if (!folio)
 		return -ENOMEM;
 
@@ -2449,6 +2456,8 @@ static int filemap_create_folio(struct file *file,
 	 * well to keep locking rules simple.
 	 */
 	filemap_invalidate_lock_shared(mapping);
+	/* index in PAGE units but aligned to min_order number of pages. */
+	index = (pos >> (PAGE_SHIFT + min_order)) << min_order;
 	error = filemap_add_folio(mapping, folio, index,
 			mapping_gfp_constraint(mapping, GFP_KERNEL));
 	if (error == -EEXIST)
@@ -2509,8 +2518,7 @@ static int filemap_get_pages(struct kiocb *iocb, size_t count,
 	if (!folio_batch_count(fbatch)) {
 		if (iocb->ki_flags & (IOCB_NOWAIT | IOCB_WAITQ))
 			return -EAGAIN;
-		err = filemap_create_folio(filp, mapping,
-				iocb->ki_pos >> PAGE_SHIFT, fbatch);
+		err = filemap_create_folio(filp, mapping, iocb->ki_pos, fbatch);
 		if (err == AOP_TRUNCATED_PAGE)
 			goto retry;
 		return err;
@@ -3708,9 +3716,11 @@ static struct folio *do_read_cache_folio(struct address_space *mapping,
 repeat:
 	folio = filemap_get_folio(mapping, index);
 	if (IS_ERR(folio)) {
-		folio = filemap_alloc_folio(gfp, 0);
+		folio = filemap_alloc_folio(gfp,
+					    mapping_min_folio_order(mapping));
 		if (!folio)
 			return ERR_PTR(-ENOMEM);
+		index = mapping_align_start_index(mapping, index);
 		err = filemap_add_folio(mapping, folio, index, gfp);
 		if (unlikely(err)) {
 			folio_put(folio);
-- 
2.43.0


  parent reply	other threads:[~2024-05-03  9:54 UTC|newest]

Thread overview: 51+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2024-05-03  9:53 [PATCH v5 00/11] enable bs > ps in XFS Luis Chamberlain
2024-05-03  9:53 ` [PATCH v5 01/11] readahead: rework loop in page_cache_ra_unbounded() Luis Chamberlain
2024-05-03  9:53 ` [PATCH v5 02/11] fs: Allow fine-grained control of folio sizes Luis Chamberlain
2024-05-03  9:53 ` Luis Chamberlain [this message]
2024-05-03  9:53 ` [PATCH v5 04/11] readahead: allocate folios with mapping_min_order in readahead Luis Chamberlain
2024-05-03 14:32   ` Hannes Reinecke
2024-05-03  9:53 ` [PATCH v5 05/11] mm: split a folio in minimum folio order chunks Luis Chamberlain
2024-05-03 14:53   ` Zi Yan
2024-05-15 15:32   ` Matthew Wilcox
2024-05-16 14:56     ` Pankaj Raghav (Samsung)
2024-05-03  9:53 ` [PATCH v5 06/11] filemap: cap PTE range to be created to allowed zero fill in folio_map_range() Luis Chamberlain
2024-05-03  9:53 ` [PATCH v5 07/11] iomap: fix iomap_dio_zero() for fs bs > system page size Luis Chamberlain
2024-05-07 14:58   ` [RFC] iomap: use huge zero folio in iomap_dio_zero Pankaj Raghav (Samsung)
2024-05-07 15:11     ` Zi Yan
2024-05-07 16:11     ` Christoph Hellwig
2024-05-08 11:39       ` Pankaj Raghav (Samsung)
2024-05-08 11:43         ` Christoph Hellwig
2024-05-09 12:31           ` Pankaj Raghav (Samsung)
2024-05-09 12:46             ` Christoph Hellwig
2024-05-09 12:55               ` Pankaj Raghav (Samsung)
2024-05-09 12:58                 ` Christoph Hellwig
2024-05-09 14:32                   ` Darrick J. Wong
2024-05-09 15:05                     ` Christoph Hellwig
2024-05-09 15:08                       ` Darrick J. Wong
2024-05-09 15:09                         ` Christoph Hellwig
2024-05-15  0:50     ` Matthew Wilcox
2024-05-15  2:34       ` Keith Busch
2024-05-15  4:04         ` Matthew Wilcox
2024-05-15 15:59           ` Pankaj Raghav (Samsung)
2024-05-15 18:03             ` Matthew Wilcox
2024-05-16 15:02               ` Pankaj Raghav (Samsung)
2024-05-17 12:36                 ` Hannes Reinecke
2024-05-17 12:56                   ` Hannes Reinecke
2024-05-17 13:30                   ` Matthew Wilcox
2024-05-15 11:48       ` Christoph Hellwig
2024-05-07 16:00   ` [PATCH v5 07/11] iomap: fix iomap_dio_zero() for fs bs > system page size Matthew Wilcox
2024-05-07 16:10     ` Christoph Hellwig
2024-05-07 16:11       ` Matthew Wilcox
2024-05-07 16:13         ` Christoph Hellwig
2024-05-08  4:24           ` Matthew Wilcox
2024-05-08 11:22             ` Pankaj Raghav (Samsung)
2024-05-08 11:36             ` Christoph Hellwig
2024-05-08 11:20     ` Pankaj Raghav (Samsung)
2024-05-03  9:53 ` [PATCH v5 08/11] xfs: use kvmalloc for xattr buffers Luis Chamberlain
2024-05-03  9:53 ` [PATCH v5 09/11] xfs: expose block size in stat Luis Chamberlain
2024-05-03  9:53 ` [PATCH v5 10/11] xfs: make the calculation generic in xfs_sb_validate_fsb_count() Luis Chamberlain
2024-05-07  8:40   ` John Garry
2024-05-07 21:13     ` Darrick J. Wong
2024-05-08 11:28       ` Pankaj Raghav (Samsung)
2024-05-03  9:53 ` [PATCH v5 11/11] xfs: enable block size larger than page size support Luis Chamberlain
2024-05-07  0:05   ` Dave Chinner

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20240503095353.3798063-4-mcgrof@kernel.org \
    --to=mcgrof@kernel.org \
    --cc=akpm@linux-foundation.org \
    --cc=brauner@kernel.org \
    --cc=chandan.babu@oracle.com \
    --cc=david@fromorbit.com \
    --cc=djwong@kernel.org \
    --cc=gost.dev@samsung.com \
    --cc=hare@suse.de \
    --cc=john.g.garry@oracle.com \
    --cc=kernel@pankajraghav.com \
    --cc=linux-block@vger.kernel.org \
    --cc=linux-fsdevel@vger.kernel.org \
    --cc=linux-mm@kvack.org \
    --cc=linux-xfs@vger.kernel.org \
    --cc=p.raghav@samsung.com \
    --cc=ritesh.list@gmail.com \
    --cc=willy@infradead.org \
    --cc=ziy@nvidia.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.