From: Max Kellermann <max.kellermann@ionos.com>
To: akpm@linux-foundation.org, linux-mm@kvack.org,
linux-kernel@vger.kernel.org
Cc: willy@infradead.org, Max Kellermann <max.kellermann@ionos.com>
Subject: [PATCH v2 07/14] linux/mm.h: move folio_next() to mm/folio_next.h
Date: Thu, 22 Feb 2024 12:50:48 +0100 [thread overview]
Message-ID: <20240222115055.1172877-8-max.kellermann@ionos.com> (raw)
In-Reply-To: <20240222115055.1172877-1-max.kellermann@ionos.com>
Prepare to reduce dependencies on linux/mm.h.
folio_next() is used by linux/bio.h. Moving it to a separate lean
header will allow us to avoid the dependency on linux/mm.h.
Signed-off-by: Max Kellermann <max.kellermann@ionos.com>
---
include/linux/mm.h | 20 +-------------------
include/linux/mm/folio_next.h | 27 +++++++++++++++++++++++++++
2 files changed, 28 insertions(+), 19 deletions(-)
create mode 100644 include/linux/mm/folio_next.h
diff --git a/include/linux/mm.h b/include/linux/mm.h
index 9dea6477b4ac..7d83d6d64f30 100644
--- a/include/linux/mm.h
+++ b/include/linux/mm.h
@@ -2,6 +2,7 @@
#ifndef _LINUX_MM_H
#define _LINUX_MM_H
+#include <linux/mm/folio_next.h>
#include <linux/mm/folio_size.h>
#include <linux/mm/page_address.h>
#include <linux/mm/page_section.h>
@@ -1907,25 +1908,6 @@ static inline void set_page_links(struct page *page, enum zone_type zone,
#endif
}
-/**
- * folio_next - Move to the next physical folio.
- * @folio: The folio we're currently operating on.
- *
- * If you have physically contiguous memory which may span more than
- * one folio (eg a &struct bio_vec), use this function to move from one
- * folio to the next. Do not use it if the memory is only virtually
- * contiguous as the folios are almost certainly not adjacent to each
- * other. This is the folio equivalent to writing ``page++``.
- *
- * Context: We assume that the folios are refcounted and/or locked at a
- * higher level and do not adjust the reference counts.
- * Return: The next struct folio.
- */
-static inline struct folio *folio_next(struct folio *folio)
-{
- return (struct folio *)folio_page(folio, folio_nr_pages(folio));
-}
-
/**
* folio_estimated_sharers - Estimate the number of sharers of a folio.
* @folio: The folio.
diff --git a/include/linux/mm/folio_next.h b/include/linux/mm/folio_next.h
new file mode 100644
index 000000000000..7016e303439c
--- /dev/null
+++ b/include/linux/mm/folio_next.h
@@ -0,0 +1,27 @@
+/* SPDX-License-Identifier: GPL-2.0 */
+#ifndef _LINUX_MM_FOLIO_NEXT_H
+#define _LINUX_MM_FOLIO_NEXT_H
+
+#include <linux/mm/folio_size.h> // for folio_nr_pages()
+#include <linux/mm/page_address.h> // for nth_page(), needed by folio_page()
+
+/**
+ * folio_next - Move to the next physical folio.
+ * @folio: The folio we're currently operating on.
+ *
+ * If you have physically contiguous memory which may span more than
+ * one folio (eg a &struct bio_vec), use this function to move from one
+ * folio to the next. Do not use it if the memory is only virtually
+ * contiguous as the folios are almost certainly not adjacent to each
+ * other. This is the folio equivalent to writing ``page++``.
+ *
+ * Context: We assume that the folios are refcounted and/or locked at a
+ * higher level and do not adjust the reference counts.
+ * Return: The next struct folio.
+ */
+static inline struct folio *folio_next(struct folio *folio)
+{
+ return (struct folio *)folio_page(folio, folio_nr_pages(folio));
+}
+
+#endif /* _LINUX_MM_FOLIO_NEXT_H */
--
2.39.2
next prev parent reply other threads:[~2024-02-22 11:51 UTC|newest]
Thread overview: 15+ messages / expand[flat|nested] mbox.gz Atom feed top
2024-02-22 11:50 [PATCH v2 00/14] Fast kernel headers: split linux/mm.h Max Kellermann
2024-02-22 11:50 ` [PATCH v2 01/14] drivers: add missing includes on linux/mm.h (and others) Max Kellermann
2024-02-22 11:50 ` [PATCH v2 02/14] include/drm/drm_gem.h: add poll_table_struct forward declaration Max Kellermann
2024-02-22 11:50 ` [PATCH v2 03/14] linux/mm.h: move page_kasan_tag() to mm/page_kasan_tag.h Max Kellermann
2024-02-22 11:50 ` [PATCH v2 04/14] linux/mm.h: move section functions to mm/page_section.h Max Kellermann
2024-02-22 11:50 ` [PATCH v2 05/14] linux/mm.h: move page_address() and others to mm/page_address.h Max Kellermann
2024-02-22 11:50 ` [PATCH v2 06/14] linux/mm.h: move folio_size(), ... to mm/folio_size.h Max Kellermann
2024-02-22 11:50 ` Max Kellermann [this message]
2024-02-22 11:50 ` [PATCH v2 08/14] linux/mm.h: move devmap-related declarations to mm/devmap_managed.h Max Kellermann
2024-02-22 11:50 ` [PATCH v2 09/14] linux/mm.h: move usage count functions to mm/folio_usage.h Max Kellermann
2024-02-22 11:50 ` [PATCH v2 10/14] linux/mm.h: move page_zone_id() and more to mm/folio_zone.h Max Kellermann
2024-02-22 11:50 ` [PATCH v2 11/14] linux/mm.h: move pfmemalloc-related functions to pfmemalloc.h Max Kellermann
2024-02-22 11:50 ` [PATCH v2 12/14] linux/mm.h: move is_vmalloc_addr() to mm/vmalloc_addr.h Max Kellermann
2024-02-22 11:50 ` [PATCH v2 13/14] linux/mm.h: move high_memory to mm/high_memory.h Max Kellermann
2024-02-22 11:50 ` [PATCH v2 14/14] include: reduce dependencies on linux/mm.h Max Kellermann
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20240222115055.1172877-8-max.kellermann@ionos.com \
--to=max.kellermann@ionos.com \
--cc=akpm@linux-foundation.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=willy@infradead.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).