From: Matthew Wilcox <willy@infradead.org>
To: linux-fsdevel@vger.kernel.org, linux-mm@kvack.org,
linux-kernel@vger.kernel.org
Cc: "Matthew Wilcox (Oracle)" <willy@infradead.org>
Subject: [PATCH 10/15] mm: Allow find_get_page to be used for large pages
Date: Tue, 24 Sep 2019 17:52:09 -0700 [thread overview]
Message-ID: <20190925005214.27240-11-willy@infradead.org> (raw)
In-Reply-To: <20190925005214.27240-1-willy@infradead.org>
From: "Matthew Wilcox (Oracle)" <willy@infradead.org>
Add FGP_PMD to indicate that we're trying to find-or-create a page that
is at least PMD_ORDER in size. The internal 'conflict' entry usage
is modelled after that in DAX, but the implementations are different
due to DAX using multi-order entries and the page cache using multiple
order-0 entries.
Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org>
---
include/linux/pagemap.h | 13 ++++++
mm/filemap.c | 99 +++++++++++++++++++++++++++++++++++------
2 files changed, 99 insertions(+), 13 deletions(-)
diff --git a/include/linux/pagemap.h b/include/linux/pagemap.h
index d610a49be571..d6d97f9fb762 100644
--- a/include/linux/pagemap.h
+++ b/include/linux/pagemap.h
@@ -248,6 +248,19 @@ pgoff_t page_cache_prev_miss(struct address_space *mapping,
#define FGP_NOFS 0x00000010
#define FGP_NOWAIT 0x00000020
#define FGP_FOR_MMAP 0x00000040
+/*
+ * If you add more flags, increment FGP_ORDER_SHIFT (no further than 25).
+ * Do not insert flags above the FGP order bits.
+ */
+#define FGP_ORDER_SHIFT 7
+#define FGP_PMD ((PMD_SHIFT - PAGE_SHIFT) << FGP_ORDER_SHIFT)
+#define FGP_PUD ((PUD_SHIFT - PAGE_SHIFT) << FGP_ORDER_SHIFT)
+
+#ifdef CONFIG_TRANSPARENT_HUGEPAGE
+#define fgp_order(fgp) ((fgp) >> FGP_ORDER_SHIFT)
+#else
+#define fgp_order(fgp) 0
+#endif
struct page *pagecache_get_page(struct address_space *mapping, pgoff_t offset,
int fgp_flags, gfp_t cache_gfp_mask);
diff --git a/mm/filemap.c b/mm/filemap.c
index afe8f5d95810..8eca91547e40 100644
--- a/mm/filemap.c
+++ b/mm/filemap.c
@@ -1576,7 +1576,71 @@ struct page *find_get_entry(struct address_space *mapping, pgoff_t offset)
return page;
}
-EXPORT_SYMBOL(find_get_entry);
+
+static bool pagecache_is_conflict(struct page *page)
+{
+ return page == XA_RETRY_ENTRY;
+}
+
+/**
+ * __find_get_page - Find and get a page cache entry.
+ * @mapping: The address_space to search.
+ * @offset: The page cache index.
+ * @order: The minimum order of the entry to return.
+ *
+ * Looks up the page cache entries at @mapping between @offset and
+ * @offset + 2^@order. If there is a page cache page, it is returned with
+ * an increased refcount unless it is smaller than @order.
+ *
+ * If the slot holds a shadow entry of a previously evicted page, or a
+ * swap entry from shmem/tmpfs, it is returned.
+ *
+ * Return: the found page, a value indicating a conflicting page or %NULL if
+ * there are no pages in this range.
+ */
+static struct page *__find_get_page(struct address_space *mapping,
+ unsigned long offset, unsigned int order)
+{
+ XA_STATE(xas, &mapping->i_pages, offset);
+ struct page *page;
+
+ rcu_read_lock();
+repeat:
+ xas_reset(&xas);
+ page = xas_find(&xas, offset | ((1UL << order) - 1));
+ if (xas_retry(&xas, page))
+ goto repeat;
+ /*
+ * A shadow entry of a recently evicted page, or a swap entry from
+ * shmem/tmpfs. Skip it; keep looking for pages.
+ */
+ if (xa_is_value(page))
+ goto repeat;
+ if (!page)
+ goto out;
+ if (compound_order(page) < order) {
+ page = XA_RETRY_ENTRY;
+ goto out;
+ }
+
+ if (!page_cache_get_speculative(page))
+ goto repeat;
+
+ /*
+ * Has the page moved or been split?
+ * This is part of the lockless pagecache protocol. See
+ * include/linux/pagemap.h for details.
+ */
+ if (unlikely(page != xas_reload(&xas))) {
+ put_page(page);
+ goto repeat;
+ }
+ page = find_subpage(page, offset);
+out:
+ rcu_read_unlock();
+
+ return page;
+}
/**
* find_lock_entry - locate, pin and lock a page cache entry
@@ -1618,12 +1682,12 @@ EXPORT_SYMBOL(find_lock_entry);
* pagecache_get_page - find and get a page reference
* @mapping: the address_space to search
* @offset: the page index
- * @fgp_flags: PCG flags
+ * @fgp_flags: FGP flags
* @gfp_mask: gfp mask to use for the page cache data page allocation
*
* Looks up the page cache slot at @mapping & @offset.
*
- * PCG flags modify how the page is returned.
+ * FGP flags modify how the page is returned.
*
* @fgp_flags can be:
*
@@ -1636,6 +1700,10 @@ EXPORT_SYMBOL(find_lock_entry);
* - FGP_FOR_MMAP: Similar to FGP_CREAT, only we want to allow the caller to do
* its own locking dance if the page is already in cache, or unlock the page
* before returning if we had to add the page to pagecache.
+ * - FGP_PMD: We're only interested in pages at PMD granularity. If there
+ * is no page here (and FGP_CREATE is set), we'll create one large enough.
+ * If there is a smaller page in the cache that overlaps the PMD page, we
+ * return %NULL and do not attempt to create a page.
*
* If FGP_LOCK or FGP_CREAT are specified then the function may sleep even
* if the GFP flags specified for FGP_CREAT are atomic.
@@ -1649,10 +1717,11 @@ struct page *pagecache_get_page(struct address_space *mapping, pgoff_t offset,
{
struct page *page;
+ BUILD_BUG_ON(((63 << FGP_ORDER_SHIFT) >> FGP_ORDER_SHIFT) != 63);
repeat:
- page = find_get_entry(mapping, offset);
- if (xa_is_value(page))
- page = NULL;
+ page = __find_get_page(mapping, offset, fgp_order(fgp_flags));
+ if (pagecache_is_conflict(page))
+ return NULL;
if (!page)
goto no_page;
@@ -1686,7 +1755,7 @@ struct page *pagecache_get_page(struct address_space *mapping, pgoff_t offset,
if (fgp_flags & FGP_NOFS)
gfp_mask &= ~__GFP_FS;
- page = __page_cache_alloc(gfp_mask);
+ page = __page_cache_alloc_order(gfp_mask, fgp_order(fgp_flags));
if (!page)
return NULL;
@@ -1704,13 +1773,17 @@ struct page *pagecache_get_page(struct address_space *mapping, pgoff_t offset,
if (err == -EEXIST)
goto repeat;
}
+ if (page) {
+ if (fgp_order(fgp_flags))
+ count_vm_event(THP_FILE_ALLOC);
- /*
- * add_to_page_cache_lru locks the page, and for mmap we expect
- * an unlocked page.
- */
- if (page && (fgp_flags & FGP_FOR_MMAP))
- unlock_page(page);
+ /*
+ * add_to_page_cache_lru locks the page, and
+ * for mmap we expect an unlocked page.
+ */
+ if (fgp_flags & FGP_FOR_MMAP)
+ unlock_page(page);
+ }
}
return page;
--
2.23.0
next prev parent reply other threads:[~2019-09-25 0:53 UTC|newest]
Thread overview: 40+ messages / expand[flat|nested] mbox.gz Atom feed top
2019-09-25 0:51 [RFC 00/15] Large pages in the page-cache Matthew Wilcox
2019-09-25 0:52 ` [PATCH 01/15] mm: Use vm_fault error code directly Matthew Wilcox
2019-09-26 13:55 ` Kirill A. Shutemov
2019-09-25 0:52 ` [PATCH 02/15] fs: Introduce i_blocks_per_page Matthew Wilcox
2019-09-25 8:36 ` Dave Chinner
2019-10-04 19:28 ` Matthew Wilcox
2019-10-08 3:53 ` Dave Chinner
2019-09-25 0:52 ` [PATCH 03/15] mm: Add file_offset_of_ helpers Matthew Wilcox
2019-09-26 14:02 ` Kirill A. Shutemov
2019-10-04 19:39 ` Matthew Wilcox
2019-09-25 0:52 ` [PATCH 04/15] iomap: Support large pages Matthew Wilcox
2019-09-25 0:52 ` [PATCH 05/15] xfs: " Matthew Wilcox
2019-09-25 0:52 ` [PATCH 06/15] xfs: Pass a page to xfs_finish_page_writeback Matthew Wilcox
2019-09-25 0:52 ` [PATCH 07/15] mm: Make prep_transhuge_page tail-callable Matthew Wilcox
2019-09-26 14:13 ` Kirill A. Shutemov
2019-09-25 0:52 ` [PATCH 08/15] mm: Add __page_cache_alloc_order Matthew Wilcox
2019-09-26 14:15 ` Kirill A. Shutemov
2019-09-25 0:52 ` [PATCH 09/15] mm: Allow large pages to be added to the page cache Matthew Wilcox
2019-09-26 14:22 ` Kirill A. Shutemov
2019-09-25 0:52 ` Matthew Wilcox [this message]
2019-10-01 10:32 ` [PATCH 10/15] mm: Allow find_get_page to be used for large pages Kirill A. Shutemov
2019-09-25 0:52 ` [PATCH 11/15] mm: Remove hpage_nr_pages Matthew Wilcox
2019-10-01 10:35 ` Kirill A. Shutemov
2019-09-25 0:52 ` [PATCH 12/15] mm: Support removing arbitrary sized pages from mapping Matthew Wilcox
2019-10-01 10:39 ` Kirill A. Shutemov
2019-09-25 0:52 ` [PATCH 13/15] mm: Add a huge page fault handler for files Matthew Wilcox
2019-10-01 10:42 ` Kirill A. Shutemov
2019-09-25 0:52 ` [PATCH 14/15] mm: Align THP mappings for non-DAX Matthew Wilcox
2019-10-01 10:45 ` Kirill A. Shutemov
2019-10-01 11:21 ` William Kucharski
2019-10-01 11:32 ` Kirill A. Shutemov
2019-10-01 12:18 ` William Kucharski
2019-10-01 14:20 ` Kirill A. Shutemov
2019-10-01 16:08 ` William Kucharski
2019-10-02 0:15 ` Kirill A. Shutemov
2019-09-25 0:52 ` [PATCH 15/15] xfs: Use filemap_huge_fault Matthew Wilcox
[not found] ` <20191002130753.7680-1-hdanton@sina.com>
2019-10-04 19:33 ` [PATCH 03/15] mm: Add file_offset_of_ helpers Matthew Wilcox
[not found] ` <20191002133211.15696-1-hdanton@sina.com>
2019-10-04 19:34 ` [PATCH 04/15] iomap: Support large pages Matthew Wilcox
[not found] ` <20191003040846.17604-1-hdanton@sina.com>
2019-10-04 19:35 ` [PATCH 06/15] xfs: Pass a page to xfs_finish_page_writeback Matthew Wilcox
[not found] ` <20191003050859.18140-1-hdanton@sina.com>
2019-10-04 19:36 ` [PATCH 11/15] mm: Remove hpage_nr_pages Matthew Wilcox
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20190925005214.27240-11-willy@infradead.org \
--to=willy@infradead.org \
--cc=linux-fsdevel@vger.kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).