All of lore.kernel.org
 help / color / mirror / Atom feed
From: Baoquan He <bhe@redhat.com>
To: linux-mm@kvack.org
Cc: linux-kernel@vger.kernel.org, x86@kernel.org,
	linuxppc-dev@lists.ozlabs.org, akpm@linux-foundation.org,
	rppt@kernel.org, Baoquan He <bhe@redhat.com>
Subject: [PATCH v2 2/6] mm/mm_init.c: remove the useless dma_reserve
Date: Mon, 25 Mar 2024 22:56:42 +0800	[thread overview]
Message-ID: <20240325145646.1044760-3-bhe@redhat.com> (raw)
In-Reply-To: <20240325145646.1044760-1-bhe@redhat.com>

Now nobody calls set_dma_reserve() to set value for dma_reserve, remove
set_dma_reserve(), global variable dma_reserve and the codes using it.

Signed-off-by: Baoquan He <bhe@redhat.com>
---
 include/linux/mm.h |  1 -
 mm/mm_init.c       | 23 -----------------------
 2 files changed, 24 deletions(-)

diff --git a/include/linux/mm.h b/include/linux/mm.h
index 0436b919f1c7..ad19350e1538 100644
--- a/include/linux/mm.h
+++ b/include/linux/mm.h
@@ -3210,7 +3210,6 @@ static inline int early_pfn_to_nid(unsigned long pfn)
 extern int __meminit early_pfn_to_nid(unsigned long pfn);
 #endif
 
-extern void set_dma_reserve(unsigned long new_dma_reserve);
 extern void mem_init(void);
 extern void __init mmap_init(void);
 
diff --git a/mm/mm_init.c b/mm/mm_init.c
index 549e76af8f82..153fb2dc666f 100644
--- a/mm/mm_init.c
+++ b/mm/mm_init.c
@@ -226,7 +226,6 @@ static unsigned long required_movablecore_percent __initdata;
 
 static unsigned long nr_kernel_pages __initdata;
 static unsigned long nr_all_pages __initdata;
-static unsigned long dma_reserve __initdata;
 
 static bool deferred_struct_pages __meminitdata;
 
@@ -1583,12 +1582,6 @@ static void __init free_area_init_core(struct pglist_data *pgdat)
 					zone_names[j], memmap_pages, freesize);
 		}
 
-		/* Account for reserved pages */
-		if (j == 0 && freesize > dma_reserve) {
-			freesize -= dma_reserve;
-			pr_debug("  %s zone: %lu pages reserved\n", zone_names[0], dma_reserve);
-		}
-
 		if (!is_highmem_idx(j))
 			nr_kernel_pages += freesize;
 		/* Charge for highmem memmap if there are enough kernel pages */
@@ -2547,22 +2540,6 @@ void *__init alloc_large_system_hash(const char *tablename,
 	return table;
 }
 
-/**
- * set_dma_reserve - set the specified number of pages reserved in the first zone
- * @new_dma_reserve: The number of pages to mark reserved
- *
- * The per-cpu batchsize and zone watermarks are determined by managed_pages.
- * In the DMA zone, a significant percentage may be consumed by kernel image
- * and other unfreeable allocations which can skew the watermarks badly. This
- * function may optionally be used to account for unfreeable pages in the
- * first zone (e.g., ZONE_DMA). The effect will be lower watermarks and
- * smaller per-cpu batchsize.
- */
-void __init set_dma_reserve(unsigned long new_dma_reserve)
-{
-	dma_reserve = new_dma_reserve;
-}
-
 void __init memblock_free_pages(struct page *page, unsigned long pfn,
 							unsigned int order)
 {
-- 
2.41.0


WARNING: multiple messages have this Message-ID (diff)
From: Baoquan He <bhe@redhat.com>
To: linux-mm@kvack.org
Cc: Baoquan He <bhe@redhat.com>,
	x86@kernel.org, linux-kernel@vger.kernel.org,
	akpm@linux-foundation.org, linuxppc-dev@lists.ozlabs.org,
	rppt@kernel.org
Subject: [PATCH v2 2/6] mm/mm_init.c: remove the useless dma_reserve
Date: Mon, 25 Mar 2024 22:56:42 +0800	[thread overview]
Message-ID: <20240325145646.1044760-3-bhe@redhat.com> (raw)
In-Reply-To: <20240325145646.1044760-1-bhe@redhat.com>

Now nobody calls set_dma_reserve() to set value for dma_reserve, remove
set_dma_reserve(), global variable dma_reserve and the codes using it.

Signed-off-by: Baoquan He <bhe@redhat.com>
---
 include/linux/mm.h |  1 -
 mm/mm_init.c       | 23 -----------------------
 2 files changed, 24 deletions(-)

diff --git a/include/linux/mm.h b/include/linux/mm.h
index 0436b919f1c7..ad19350e1538 100644
--- a/include/linux/mm.h
+++ b/include/linux/mm.h
@@ -3210,7 +3210,6 @@ static inline int early_pfn_to_nid(unsigned long pfn)
 extern int __meminit early_pfn_to_nid(unsigned long pfn);
 #endif
 
-extern void set_dma_reserve(unsigned long new_dma_reserve);
 extern void mem_init(void);
 extern void __init mmap_init(void);
 
diff --git a/mm/mm_init.c b/mm/mm_init.c
index 549e76af8f82..153fb2dc666f 100644
--- a/mm/mm_init.c
+++ b/mm/mm_init.c
@@ -226,7 +226,6 @@ static unsigned long required_movablecore_percent __initdata;
 
 static unsigned long nr_kernel_pages __initdata;
 static unsigned long nr_all_pages __initdata;
-static unsigned long dma_reserve __initdata;
 
 static bool deferred_struct_pages __meminitdata;
 
@@ -1583,12 +1582,6 @@ static void __init free_area_init_core(struct pglist_data *pgdat)
 					zone_names[j], memmap_pages, freesize);
 		}
 
-		/* Account for reserved pages */
-		if (j == 0 && freesize > dma_reserve) {
-			freesize -= dma_reserve;
-			pr_debug("  %s zone: %lu pages reserved\n", zone_names[0], dma_reserve);
-		}
-
 		if (!is_highmem_idx(j))
 			nr_kernel_pages += freesize;
 		/* Charge for highmem memmap if there are enough kernel pages */
@@ -2547,22 +2540,6 @@ void *__init alloc_large_system_hash(const char *tablename,
 	return table;
 }
 
-/**
- * set_dma_reserve - set the specified number of pages reserved in the first zone
- * @new_dma_reserve: The number of pages to mark reserved
- *
- * The per-cpu batchsize and zone watermarks are determined by managed_pages.
- * In the DMA zone, a significant percentage may be consumed by kernel image
- * and other unfreeable allocations which can skew the watermarks badly. This
- * function may optionally be used to account for unfreeable pages in the
- * first zone (e.g., ZONE_DMA). The effect will be lower watermarks and
- * smaller per-cpu batchsize.
- */
-void __init set_dma_reserve(unsigned long new_dma_reserve)
-{
-	dma_reserve = new_dma_reserve;
-}
-
 void __init memblock_free_pages(struct page *page, unsigned long pfn,
 							unsigned int order)
 {
-- 
2.41.0


  parent reply	other threads:[~2024-03-25 14:57 UTC|newest]

Thread overview: 42+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2024-03-25 14:56 [PATCH v2 0/6] mm/mm_init.c: refactor free_area_init_core() Baoquan He
2024-03-25 14:56 ` Baoquan He
2024-03-25 14:56 ` [PATCH v2 1/6] x86: remove unneeded memblock_find_dma_reserve() Baoquan He
2024-03-25 14:56   ` Baoquan He
2024-03-26  6:44   ` Mike Rapoport
2024-03-26  6:44     ` Mike Rapoport
2024-03-25 14:56 ` Baoquan He [this message]
2024-03-25 14:56   ` [PATCH v2 2/6] mm/mm_init.c: remove the useless dma_reserve Baoquan He
2024-03-26  6:44   ` Mike Rapoport
2024-03-26  6:44     ` Mike Rapoport
2024-03-25 14:56 ` [PATCH v2 3/6] mm/mm_init.c: add new function calc_nr_all_pages() Baoquan He
2024-03-25 14:56   ` Baoquan He
2024-03-26  6:57   ` Mike Rapoport
2024-03-26  6:57     ` Mike Rapoport
2024-03-26 13:49     ` Baoquan He
2024-03-26 13:49       ` Baoquan He
2024-03-27 15:40   ` Mike Rapoport
2024-03-27 15:40     ` Mike Rapoport
2024-03-25 14:56 ` [PATCH v2 4/6] mm/mm_init.c: remove meaningless calculation of zone->managed_pages in free_area_init_core() Baoquan He
2024-03-25 14:56   ` Baoquan He
2024-03-27 15:40   ` Mike Rapoport
2024-03-27 15:40     ` Mike Rapoport
2024-03-28  8:32   ` Baoquan He
2024-03-28  8:32     ` Baoquan He
2024-03-28  9:53     ` Mike Rapoport
2024-03-28  9:53       ` Mike Rapoport
2024-03-28 14:46       ` Baoquan He
2024-03-28 14:46         ` Baoquan He
2024-03-28  9:12   ` [PATCH v3 " Baoquan He
2024-03-28  9:12     ` Baoquan He
2024-03-25 14:56 ` [PATCH v2 5/6] mm/mm_init.c: remove unneeded calc_memmap_size() Baoquan He
2024-03-25 14:56   ` Baoquan He
2024-03-27 16:21   ` Mike Rapoport
2024-03-27 16:21     ` Mike Rapoport
2024-03-28  1:24     ` Baoquan He
2024-03-28  1:24       ` Baoquan He
2024-03-25 14:56 ` [PATCH v2 6/6] mm/mm_init.c: remove arch_reserved_kernel_pages() Baoquan He
2024-03-25 14:56   ` Baoquan He
2024-03-26  6:57   ` Mike Rapoport
2024-03-26  6:57     ` Mike Rapoport
2024-03-27 15:41   ` Mike Rapoport
2024-03-27 15:41     ` Mike Rapoport

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20240325145646.1044760-3-bhe@redhat.com \
    --to=bhe@redhat.com \
    --cc=akpm@linux-foundation.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-mm@kvack.org \
    --cc=linuxppc-dev@lists.ozlabs.org \
    --cc=rppt@kernel.org \
    --cc=x86@kernel.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.