linux-mm.kvack.org archive mirror
 help / color / mirror / Atom feed
* [PATCH v2] mm/hugetlb: use helper huge_page_size() to get hugepage size
@ 2021-02-09  2:18 Miaohe Lin
  2021-02-12  9:50 ` David Hildenbrand
  0 siblings, 1 reply; 2+ messages in thread
From: Miaohe Lin @ 2021-02-09  2:18 UTC (permalink / raw)
  To: akpm, mike.kravetz; +Cc: linux-mm, linux-kernel, linmiaohe

We can use helper huge_page_size() to get the hugepage size directly to
simplify the code slightly.

Reviewed-by: Mike Kravetz <mike.kravetz@oracle.com>
Signed-off-by: Miaohe Lin <linmiaohe@huawei.com>
---
v1 -> v2
	change huge_page_size(h) >> 10 to huge_page_size(h) / SZ_1K
	collect Reviewed-by tag
---
 mm/hugetlb.c | 14 ++++++--------
 1 file changed, 6 insertions(+), 8 deletions(-)

diff --git a/mm/hugetlb.c b/mm/hugetlb.c
index 18628f8dbfb0..06719fdf9fd6 100644
--- a/mm/hugetlb.c
+++ b/mm/hugetlb.c
@@ -3199,7 +3199,7 @@ void __init hugetlb_add_hstate(unsigned int order)
 	BUG_ON(order == 0);
 	h = &hstates[hugetlb_max_hstate++];
 	h->order = order;
-	h->mask = ~((1ULL << (order + PAGE_SHIFT)) - 1);
+	h->mask = ~(huge_page_size(h) - 1);
 	for (i = 0; i < MAX_NUMNODES; ++i)
 		INIT_LIST_HEAD(&h->hugepage_freelists[i]);
 	INIT_LIST_HEAD(&h->hugepage_activelist);
@@ -3474,7 +3474,7 @@ void hugetlb_report_meminfo(struct seq_file *m)
 	for_each_hstate(h) {
 		unsigned long count = h->nr_huge_pages;
 
-		total += (PAGE_SIZE << huge_page_order(h)) * count;
+		total += huge_page_size(h) * count;
 
 		if (h == &default_hstate)
 			seq_printf(m,
@@ -3487,10 +3487,10 @@ void hugetlb_report_meminfo(struct seq_file *m)
 				   h->free_huge_pages,
 				   h->resv_huge_pages,
 				   h->surplus_huge_pages,
-				   (PAGE_SIZE << huge_page_order(h)) / 1024);
+				   huge_page_size(h) / SZ_1K);
 	}
 
-	seq_printf(m, "Hugetlb:        %8lu kB\n", total / 1024);
+	seq_printf(m, "Hugetlb:        %8lu kB\n", total / SZ_1K);
 }
 
 int hugetlb_report_node_meminfo(char *buf, int len, int nid)
@@ -3524,7 +3524,7 @@ void hugetlb_show_meminfo(void)
 				h->nr_huge_pages_node[nid],
 				h->free_huge_pages_node[nid],
 				h->surplus_huge_pages_node[nid],
-				1UL << (huge_page_order(h) + PAGE_SHIFT - 10));
+				huge_page_size(h) / SZ_1K);
 }
 
 void hugetlb_report_usage(struct seq_file *m, struct mm_struct *mm)
@@ -3647,9 +3647,7 @@ static int hugetlb_vm_op_split(struct vm_area_struct *vma, unsigned long addr)
 
 static unsigned long hugetlb_vm_op_pagesize(struct vm_area_struct *vma)
 {
-	struct hstate *hstate = hstate_vma(vma);
-
-	return 1UL << huge_page_shift(hstate);
+	return huge_page_size(hstate_vma(vma));
 }
 
 /*
-- 
2.19.1



^ permalink raw reply related	[flat|nested] 2+ messages in thread

* Re: [PATCH v2] mm/hugetlb: use helper huge_page_size() to get hugepage size
  2021-02-09  2:18 [PATCH v2] mm/hugetlb: use helper huge_page_size() to get hugepage size Miaohe Lin
@ 2021-02-12  9:50 ` David Hildenbrand
  0 siblings, 0 replies; 2+ messages in thread
From: David Hildenbrand @ 2021-02-12  9:50 UTC (permalink / raw)
  To: Miaohe Lin, akpm, mike.kravetz; +Cc: linux-mm, linux-kernel

On 09.02.21 03:18, Miaohe Lin wrote:
> We can use helper huge_page_size() to get the hugepage size directly to
> simplify the code slightly.
> 
> Reviewed-by: Mike Kravetz <mike.kravetz@oracle.com>
> Signed-off-by: Miaohe Lin <linmiaohe@huawei.com>
> ---
> v1 -> v2
> 	change huge_page_size(h) >> 10 to huge_page_size(h) / SZ_1K
> 	collect Reviewed-by tag
> ---
>   mm/hugetlb.c | 14 ++++++--------
>   1 file changed, 6 insertions(+), 8 deletions(-)
> 
> diff --git a/mm/hugetlb.c b/mm/hugetlb.c
> index 18628f8dbfb0..06719fdf9fd6 100644
> --- a/mm/hugetlb.c
> +++ b/mm/hugetlb.c
> @@ -3199,7 +3199,7 @@ void __init hugetlb_add_hstate(unsigned int order)
>   	BUG_ON(order == 0);
>   	h = &hstates[hugetlb_max_hstate++];
>   	h->order = order;
> -	h->mask = ~((1ULL << (order + PAGE_SHIFT)) - 1);
> +	h->mask = ~(huge_page_size(h) - 1);
>   	for (i = 0; i < MAX_NUMNODES; ++i)
>   		INIT_LIST_HEAD(&h->hugepage_freelists[i]);
>   	INIT_LIST_HEAD(&h->hugepage_activelist);
> @@ -3474,7 +3474,7 @@ void hugetlb_report_meminfo(struct seq_file *m)
>   	for_each_hstate(h) {
>   		unsigned long count = h->nr_huge_pages;
>   
> -		total += (PAGE_SIZE << huge_page_order(h)) * count;
> +		total += huge_page_size(h) * count;
>   
>   		if (h == &default_hstate)
>   			seq_printf(m,
> @@ -3487,10 +3487,10 @@ void hugetlb_report_meminfo(struct seq_file *m)
>   				   h->free_huge_pages,
>   				   h->resv_huge_pages,
>   				   h->surplus_huge_pages,
> -				   (PAGE_SIZE << huge_page_order(h)) / 1024);
> +				   huge_page_size(h) / SZ_1K);
>   	}
>   
> -	seq_printf(m, "Hugetlb:        %8lu kB\n", total / 1024);
> +	seq_printf(m, "Hugetlb:        %8lu kB\n", total / SZ_1K);
>   }
>   
>   int hugetlb_report_node_meminfo(char *buf, int len, int nid)
> @@ -3524,7 +3524,7 @@ void hugetlb_show_meminfo(void)
>   				h->nr_huge_pages_node[nid],
>   				h->free_huge_pages_node[nid],
>   				h->surplus_huge_pages_node[nid],
> -				1UL << (huge_page_order(h) + PAGE_SHIFT - 10));
> +				huge_page_size(h) / SZ_1K);
>   }
>   
>   void hugetlb_report_usage(struct seq_file *m, struct mm_struct *mm)
> @@ -3647,9 +3647,7 @@ static int hugetlb_vm_op_split(struct vm_area_struct *vma, unsigned long addr)
>   
>   static unsigned long hugetlb_vm_op_pagesize(struct vm_area_struct *vma)
>   {
> -	struct hstate *hstate = hstate_vma(vma);
> -
> -	return 1UL << huge_page_shift(hstate);
> +	return huge_page_size(hstate_vma(vma));
>   }
>   
>   /*
> 

Reviewed-by: David Hildenbrand <david@redhat.com>

Thanks!

-- 
Thanks,

David / dhildenb



^ permalink raw reply	[flat|nested] 2+ messages in thread

end of thread, other threads:[~2021-02-12  9:50 UTC | newest]

Thread overview: 2+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2021-02-09  2:18 [PATCH v2] mm/hugetlb: use helper huge_page_size() to get hugepage size Miaohe Lin
2021-02-12  9:50 ` David Hildenbrand

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).