From: Matthew Wilcox <willy@infradead.org>
To: linux-fsdevel@vger.kernel.org
Cc: "Matthew Wilcox (Oracle)" <willy@infradead.org>,
linux-mm@kvack.org, linux-kernel@vger.kernel.org
Subject: [PATCH v4 13/36] iomap: Support arbitrarily many blocks per page
Date: Fri, 15 May 2020 06:16:33 -0700 [thread overview]
Message-ID: <20200515131656.12890-14-willy@infradead.org> (raw)
In-Reply-To: <20200515131656.12890-1-willy@infradead.org>
From: "Matthew Wilcox (Oracle)" <willy@infradead.org>
Size the uptodate array dynamically. Now that this array is protected
by a spinlock, we can use bitmap functions to set the bits in this array
instead of a loop around set_bit().
Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org>
---
fs/iomap/buffered-io.c | 27 +++++++++------------------
1 file changed, 9 insertions(+), 18 deletions(-)
diff --git a/fs/iomap/buffered-io.c b/fs/iomap/buffered-io.c
index 4bc37bf8d057..4a79061073eb 100644
--- a/fs/iomap/buffered-io.c
+++ b/fs/iomap/buffered-io.c
@@ -22,14 +22,14 @@
#include "../internal.h"
/*
- * Structure allocated for each page when block size < PAGE_SIZE to track
+ * Structure allocated for each page when block size < page size to track
* sub-page uptodate status and I/O completions.
*/
struct iomap_page {
atomic_t read_count;
atomic_t write_count;
spinlock_t uptodate_lock;
- DECLARE_BITMAP(uptodate, PAGE_SIZE / 512);
+ unsigned long uptodate[];
};
static inline struct iomap_page *to_iomap_page(struct page *page)
@@ -45,15 +45,14 @@ static struct iomap_page *
iomap_page_create(struct inode *inode, struct page *page)
{
struct iomap_page *iop = to_iomap_page(page);
+ unsigned int nr_blocks = i_blocks_per_page(inode, page);
- if (iop || i_blocks_per_page(inode, page) <= 1)
+ if (iop || nr_blocks <= 1)
return iop;
- iop = kmalloc(sizeof(*iop), GFP_NOFS | __GFP_NOFAIL);
- atomic_set(&iop->read_count, 0);
- atomic_set(&iop->write_count, 0);
+ iop = kzalloc(struct_size(iop, uptodate, BITS_TO_LONGS(nr_blocks)),
+ GFP_NOFS | __GFP_NOFAIL);
spin_lock_init(&iop->uptodate_lock);
- bitmap_zero(iop->uptodate, PAGE_SIZE / SECTOR_SIZE);
/*
* migrate_page_move_mapping() assumes that pages with private data have
@@ -146,20 +145,12 @@ iomap_iop_set_range_uptodate(struct page *page, unsigned off, unsigned len)
struct iomap_page *iop = to_iomap_page(page);
struct inode *inode = page->mapping->host;
unsigned first = off >> inode->i_blkbits;
- unsigned last = (off + len - 1) >> inode->i_blkbits;
- bool uptodate = true;
+ unsigned count = len >> inode->i_blkbits;
unsigned long flags;
- unsigned int i;
spin_lock_irqsave(&iop->uptodate_lock, flags);
- for (i = 0; i < i_blocks_per_page(inode, page); i++) {
- if (i >= first && i <= last)
- set_bit(i, iop->uptodate);
- else if (!test_bit(i, iop->uptodate))
- uptodate = false;
- }
-
- if (uptodate)
+ bitmap_set(iop->uptodate, first, count);
+ if (bitmap_full(iop->uptodate, i_blocks_per_page(inode, page)))
SetPageUptodate(page);
spin_unlock_irqrestore(&iop->uptodate_lock, flags);
}
--
2.26.2
next prev parent reply other threads:[~2020-05-15 13:41 UTC|newest]
Thread overview: 59+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-05-15 13:16 [PATCH v4 00/36] Large pages in the page cache Matthew Wilcox
2020-05-15 13:16 ` [PATCH v4 01/36] mm: Move PageDoubleMap bit Matthew Wilcox
2020-05-15 13:16 ` [PATCH v4 02/36] mm: Simplify PageDoubleMap with PF_SECOND policy Matthew Wilcox
2020-05-15 13:16 ` [PATCH v4 03/36] mm: Allow hpages to be arbitrary order Matthew Wilcox
2020-05-28 14:19 ` Zi Yan
2020-05-15 13:16 ` [PATCH v4 04/36] mm: Introduce thp_size Matthew Wilcox
2020-05-15 13:38 ` David Hildenbrand
2020-05-15 13:16 ` [PATCH v4 05/36] mm: Introduce thp_order Matthew Wilcox
2020-05-15 13:16 ` [PATCH v4 06/36] mm: Introduce offset_in_thp Matthew Wilcox
2020-05-15 13:39 ` David Hildenbrand
2020-05-22 17:15 ` Kirill A. Shutemov
2020-05-29 12:59 ` Matthew Wilcox
2020-05-15 13:16 ` [PATCH v4 07/36] fs: Add a filesystem flag for large pages Matthew Wilcox
2020-05-21 21:55 ` Dave Chinner
2020-05-21 23:29 ` Matthew Wilcox
2020-05-15 13:16 ` [PATCH v4 08/36] fs: Do not update nr_thps for large page mappings Matthew Wilcox
2020-05-15 13:16 ` [PATCH v4 09/36] fs: Introduce i_blocks_per_page Matthew Wilcox
2020-05-15 13:16 ` [PATCH v4 10/36] fs: Make page_mkwrite_check_truncate thp-aware Matthew Wilcox
2020-05-21 22:01 ` Dave Chinner
2020-05-21 23:30 ` Matthew Wilcox
2020-05-15 13:16 ` [PATCH v4 11/36] fs: Support THPs in zero_user_segments Matthew Wilcox
2020-05-25 4:55 ` Kirill A. Shutemov
2020-05-15 13:16 ` [PATCH v4 12/36] bio: Add bio_for_each_thp_segment_all Matthew Wilcox
2020-05-15 13:16 ` Matthew Wilcox [this message]
2020-05-15 13:16 ` [PATCH v4 14/36] iomap: Support large pages in iomap_adjust_read_range Matthew Wilcox
2020-05-21 22:24 ` Dave Chinner
2020-05-21 23:39 ` Matthew Wilcox
2020-05-15 13:16 ` [PATCH v4 15/36] iomap: Support large pages in read paths Matthew Wilcox
2020-05-15 13:16 ` [PATCH v4 16/36] iomap: Support large pages in write paths Matthew Wilcox
2020-05-15 13:16 ` [PATCH v4 17/36] iomap: Inline data shouldn't see large pages Matthew Wilcox
2020-05-15 13:16 ` [PATCH v4 18/36] iomap: Handle tail pages in iomap_page_mkwrite Matthew Wilcox
2020-05-15 13:16 ` [PATCH v4 19/36] xfs: Support large pages Matthew Wilcox
2020-05-15 13:16 ` [PATCH v4 20/36] mm: Make prep_transhuge_page return its argument Matthew Wilcox
2020-05-15 13:16 ` [PATCH v4 21/36] mm: Add __page_cache_alloc_order Matthew Wilcox
2020-05-15 13:16 ` [PATCH v4 22/36] mm: Allow large pages to be added to the page cache Matthew Wilcox
2020-05-15 13:16 ` [PATCH v4 23/36] mm: Allow large pages to be removed from " Matthew Wilcox
2020-05-15 13:16 ` [PATCH v4 24/36] mm: Remove page fault assumption of compound page size Matthew Wilcox
2020-05-25 4:59 ` Kirill A. Shutemov
2020-05-15 13:16 ` [PATCH v4 25/36] mm: Fix total_mapcount assumption of " Matthew Wilcox
2020-05-15 13:16 ` [PATCH v4 26/36] mm: Avoid splitting large pages Matthew Wilcox
2020-05-15 13:16 ` [PATCH v4 27/36] mm: Fix truncation for pages of arbitrary size Matthew Wilcox
2020-05-15 13:16 ` [PATCH v4 28/36] mm: Support storing shadow entries for large pages Matthew Wilcox
2020-05-15 13:16 ` [PATCH v4 29/36] mm: Support retrieving tail pages from the page cache Matthew Wilcox
2020-05-15 13:16 ` [PATCH v4 30/36] mm: Support tail pages in wait_for_stable_page Matthew Wilcox
2020-05-15 13:16 ` [PATCH v4 31/36] mm: Add DEFINE_READAHEAD Matthew Wilcox
2020-05-15 13:16 ` [PATCH v4 32/36] mm: Make page_cache_readahead_unbounded take a readahead_control Matthew Wilcox
2020-05-15 13:16 ` [PATCH v4 33/36] mm: Make __do_page_cache_readahead " Matthew Wilcox
2020-05-15 13:16 ` [PATCH v4 34/36] mm: Allow PageReadahead to be set on head pages Matthew Wilcox
2020-05-15 13:16 ` [PATCH v4 35/36] mm: Add large page readahead Matthew Wilcox
2020-05-15 13:16 ` [PATCH v4 36/36] mm: Align THP mappings for non-DAX Matthew Wilcox
2020-05-26 22:05 ` William Kucharski
2020-05-26 22:20 ` Matthew Wilcox
2020-05-21 22:49 ` [PATCH v4 00/36] Large pages in the page cache Dave Chinner
2020-05-22 0:04 ` Matthew Wilcox
2020-05-22 2:57 ` Dave Chinner
2020-05-22 3:05 ` Matthew Wilcox
2020-05-25 23:07 ` Dave Chinner
2020-05-26 1:21 ` Matthew Wilcox
2020-05-28 11:00 ` William Kucharski
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20200515131656.12890-14-willy@infradead.org \
--to=willy@infradead.org \
--cc=linux-fsdevel@vger.kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).