linux-mm.kvack.org archive mirror
 help / color / mirror / Atom feed
* [PATCH v1 1/2] mm/damon: rename sz_damon_region to damon_region_size
@ 2022-09-26  7:10 Xin Hao
  2022-09-26  7:11 ` [PATCH v1 2/2] mm/damon: use damon_region_size() in appropriate place Xin Hao
  2022-09-26 16:16 ` [PATCH v1 1/2] mm/damon: rename sz_damon_region to damon_region_size SeongJae Park
  0 siblings, 2 replies; 4+ messages in thread
From: Xin Hao @ 2022-09-26  7:10 UTC (permalink / raw)
  To: sj; +Cc: akpm, damon, linux-mm, linux-kernel, xhao

Here, i rename sz_damon_region() to damon_region_size(), and move it to
"include/linux/damon.h", because in many places, we can to use this func.

Signed-off-by: Xin Hao <xhao@linux.alibaba.com>
---
 include/linux/damon.h | 1 +
 mm/damon/core.c       | 6 +++---
 2 files changed, 4 insertions(+), 3 deletions(-)

diff --git a/include/linux/damon.h b/include/linux/damon.h
index ed5470f50bab..21f4bfd0f41f 100644
--- a/include/linux/damon.h
+++ b/include/linux/damon.h
@@ -524,6 +524,7 @@ void damon_add_region(struct damon_region *r, struct damon_target *t);
 void damon_destroy_region(struct damon_region *r, struct damon_target *t);
 int damon_set_regions(struct damon_target *t, struct damon_addr_range *ranges,
 		unsigned int nr_ranges);
+unsigned long damon_region_size(struct damon_region *r);

 struct damos *damon_new_scheme(struct damos_access_pattern *pattern,
 			enum damos_action action, struct damos_quota *quota,
diff --git a/mm/damon/core.c b/mm/damon/core.c
index 4de8c7c52979..74ab45b2c2f1 100644
--- a/mm/damon/core.c
+++ b/mm/damon/core.c
@@ -864,7 +864,7 @@ static void kdamond_apply_schemes(struct damon_ctx *c)
 	}
 }

-static inline unsigned long sz_damon_region(struct damon_region *r)
+unsigned long damon_region_size(struct damon_region *r)
 {
 	return r->ar.end - r->ar.start;
 }
@@ -875,7 +875,7 @@ static inline unsigned long sz_damon_region(struct damon_region *r)
 static void damon_merge_two_regions(struct damon_target *t,
 		struct damon_region *l, struct damon_region *r)
 {
-	unsigned long sz_l = sz_damon_region(l), sz_r = sz_damon_region(r);
+	unsigned long sz_l = damon_region_size(l), sz_r = damon_region_size(r);

 	l->nr_accesses = (l->nr_accesses * sz_l + r->nr_accesses * sz_r) /
 			(sz_l + sz_r);
@@ -904,7 +904,7 @@ static void damon_merge_regions_of(struct damon_target *t, unsigned int thres,

 		if (prev && prev->ar.end == r->ar.start &&
 		    abs(prev->nr_accesses - r->nr_accesses) <= thres &&
-		    sz_damon_region(prev) + sz_damon_region(r) <= sz_limit)
+		    damon_region_size(prev) + damon_region_size(r) <= sz_limit)
 			damon_merge_two_regions(t, prev, r);
 		else
 			prev = r;
--
2.31.0


^ permalink raw reply related	[flat|nested] 4+ messages in thread

* [PATCH v1 2/2] mm/damon: use damon_region_size() in appropriate place
  2022-09-26  7:10 [PATCH v1 1/2] mm/damon: rename sz_damon_region to damon_region_size Xin Hao
@ 2022-09-26  7:11 ` Xin Hao
  2022-09-26 16:16 ` [PATCH v1 1/2] mm/damon: rename sz_damon_region to damon_region_size SeongJae Park
  1 sibling, 0 replies; 4+ messages in thread
From: Xin Hao @ 2022-09-26  7:11 UTC (permalink / raw)
  To: sj; +Cc: akpm, damon, linux-mm, linux-kernel, xhao

In many place, we can use damon_region_size() to instead of "r->ar.end -
r->ar.start".

Signed-off-by: Xin Hao <xhao@linux.alibaba.com>
---
 mm/damon/core.c  | 17 ++++++++---------
 mm/damon/vaddr.c |  4 ++--
 2 files changed, 10 insertions(+), 11 deletions(-)

diff --git a/mm/damon/core.c b/mm/damon/core.c
index 74ab45b2c2f1..0368551c5a8b 100644
--- a/mm/damon/core.c
+++ b/mm/damon/core.c
@@ -490,7 +490,7 @@ static unsigned long damon_region_sz_limit(struct damon_ctx *ctx)

 	damon_for_each_target(t, ctx) {
 		damon_for_each_region(r, t)
-			sz += r->ar.end - r->ar.start;
+			sz += damon_region_size(r);
 	}

 	if (ctx->attrs.min_nr_regions)
@@ -673,7 +673,7 @@ static bool __damos_valid_target(struct damon_region *r, struct damos *s)
 {
 	unsigned long sz;

-	sz = r->ar.end - r->ar.start;
+	sz = damon_region_size(r);
 	return s->pattern.min_sz_region <= sz &&
 		sz <= s->pattern.max_sz_region &&
 		s->pattern.min_nr_accesses <= r->nr_accesses &&
@@ -701,7 +701,7 @@ static void damon_do_apply_schemes(struct damon_ctx *c,

 	damon_for_each_scheme(s, c) {
 		struct damos_quota *quota = &s->quota;
-		unsigned long sz = r->ar.end - r->ar.start;
+		unsigned long sz = damon_region_size(r);
 		struct timespec64 begin, end;
 		unsigned long sz_applied = 0;

@@ -730,14 +730,14 @@ static void damon_do_apply_schemes(struct damon_ctx *c,
 				sz = ALIGN_DOWN(quota->charge_addr_from -
 						r->ar.start, DAMON_MIN_REGION);
 				if (!sz) {
-					if (r->ar.end - r->ar.start <=
-							DAMON_MIN_REGION)
+					if (damon_region_size(r) <=
+					    DAMON_MIN_REGION)
 						continue;
 					sz = DAMON_MIN_REGION;
 				}
 				damon_split_region_at(t, r, sz);
 				r = damon_next_region(r);
-				sz = r->ar.end - r->ar.start;
+				sz = damon_region_size(r);
 			}
 			quota->charge_target_from = NULL;
 			quota->charge_addr_from = 0;
@@ -842,8 +842,7 @@ static void kdamond_apply_schemes(struct damon_ctx *c)
 					continue;
 				score = c->ops.get_scheme_score(
 						c, t, r, s);
-				quota->histogram[score] +=
-					r->ar.end - r->ar.start;
+				quota->histogram[score] += damon_region_size(r);
 				if (score > max_score)
 					max_score = score;
 			}
@@ -962,7 +961,7 @@ static void damon_split_regions_of(struct damon_target *t, int nr_subs)
 	int i;

 	damon_for_each_region_safe(r, next, t) {
-		sz_region = r->ar.end - r->ar.start;
+		sz_region = damon_region_size(r);

 		for (i = 0; i < nr_subs - 1 &&
 				sz_region > 2 * DAMON_MIN_REGION; i++) {
diff --git a/mm/damon/vaddr.c b/mm/damon/vaddr.c
index ea94e0b2c311..bd1ac9db0f49 100644
--- a/mm/damon/vaddr.c
+++ b/mm/damon/vaddr.c
@@ -72,7 +72,7 @@ static int damon_va_evenly_split_region(struct damon_target *t,
 		return -EINVAL;

 	orig_end = r->ar.end;
-	sz_orig = r->ar.end - r->ar.start;
+	sz_orig = damon_region_size(r);
 	sz_piece = ALIGN_DOWN(sz_orig / nr_pieces, DAMON_MIN_REGION);

 	if (!sz_piece)
@@ -618,7 +618,7 @@ static unsigned long damos_madvise(struct damon_target *target,
 {
 	struct mm_struct *mm;
 	unsigned long start = PAGE_ALIGN(r->ar.start);
-	unsigned long len = PAGE_ALIGN(r->ar.end - r->ar.start);
+	unsigned long len = PAGE_ALIGN(damon_region_size(r));
 	unsigned long applied;

 	mm = damon_get_mm(target);
--
2.31.0


^ permalink raw reply related	[flat|nested] 4+ messages in thread

* Re: [PATCH v1 1/2] mm/damon: rename sz_damon_region to damon_region_size
  2022-09-26  7:10 [PATCH v1 1/2] mm/damon: rename sz_damon_region to damon_region_size Xin Hao
  2022-09-26  7:11 ` [PATCH v1 2/2] mm/damon: use damon_region_size() in appropriate place Xin Hao
@ 2022-09-26 16:16 ` SeongJae Park
  2022-09-27  0:04   ` haoxin
  1 sibling, 1 reply; 4+ messages in thread
From: SeongJae Park @ 2022-09-26 16:16 UTC (permalink / raw)
  To: Xin Hao; +Cc: sj, akpm, damon, linux-mm, linux-kernel

On Mon, 26 Sep 2022 15:10:59 +0800 Xin Hao <xhao@linux.alibaba.com> wrote:

> Here, i rename sz_damon_region() to damon_region_size(), and move it to
> "include/linux/damon.h", because in many places, we can to use this func.

Good idea.

> 
> Signed-off-by: Xin Hao <xhao@linux.alibaba.com>
> ---
>  include/linux/damon.h | 1 +
>  mm/damon/core.c       | 6 +++---
>  2 files changed, 4 insertions(+), 3 deletions(-)
> 
> diff --git a/include/linux/damon.h b/include/linux/damon.h
> index ed5470f50bab..21f4bfd0f41f 100644
> --- a/include/linux/damon.h
> +++ b/include/linux/damon.h
> @@ -524,6 +524,7 @@ void damon_add_region(struct damon_region *r, struct damon_target *t);
>  void damon_destroy_region(struct damon_region *r, struct damon_target *t);
>  int damon_set_regions(struct damon_target *t, struct damon_addr_range *ranges,
>  		unsigned int nr_ranges);
> +unsigned long damon_region_size(struct damon_region *r);

This name looks good enough, but I'd like to suggest damon_sz_region().  What
do you think?

Also, I'd like to keep this function 'static inline' and put the definition
just after that of 'damon_first_region()'.


Thanks,
SJ

> 
>  struct damos *damon_new_scheme(struct damos_access_pattern *pattern,
>  			enum damos_action action, struct damos_quota *quota,
> diff --git a/mm/damon/core.c b/mm/damon/core.c
> index 4de8c7c52979..74ab45b2c2f1 100644
> --- a/mm/damon/core.c
> +++ b/mm/damon/core.c
> @@ -864,7 +864,7 @@ static void kdamond_apply_schemes(struct damon_ctx *c)
>  	}
>  }
> 
> -static inline unsigned long sz_damon_region(struct damon_region *r)
> +unsigned long damon_region_size(struct damon_region *r)
>  {
>  	return r->ar.end - r->ar.start;
>  }
> @@ -875,7 +875,7 @@ static inline unsigned long sz_damon_region(struct damon_region *r)
>  static void damon_merge_two_regions(struct damon_target *t,
>  		struct damon_region *l, struct damon_region *r)
>  {
> -	unsigned long sz_l = sz_damon_region(l), sz_r = sz_damon_region(r);
> +	unsigned long sz_l = damon_region_size(l), sz_r = damon_region_size(r);
> 
>  	l->nr_accesses = (l->nr_accesses * sz_l + r->nr_accesses * sz_r) /
>  			(sz_l + sz_r);
> @@ -904,7 +904,7 @@ static void damon_merge_regions_of(struct damon_target *t, unsigned int thres,
> 
>  		if (prev && prev->ar.end == r->ar.start &&
>  		    abs(prev->nr_accesses - r->nr_accesses) <= thres &&
> -		    sz_damon_region(prev) + sz_damon_region(r) <= sz_limit)
> +		    damon_region_size(prev) + damon_region_size(r) <= sz_limit)
>  			damon_merge_two_regions(t, prev, r);
>  		else
>  			prev = r;
> --
> 2.31.0
> 


^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: [PATCH v1 1/2] mm/damon: rename sz_damon_region to damon_region_size
  2022-09-26 16:16 ` [PATCH v1 1/2] mm/damon: rename sz_damon_region to damon_region_size SeongJae Park
@ 2022-09-27  0:04   ` haoxin
  0 siblings, 0 replies; 4+ messages in thread
From: haoxin @ 2022-09-27  0:04 UTC (permalink / raw)
  To: SeongJae Park; +Cc: akpm, damon, linux-mm, linux-kernel


在 2022/9/27 上午12:16, SeongJae Park 写道:
> On Mon, 26 Sep 2022 15:10:59 +0800 Xin Hao <xhao@linux.alibaba.com> wrote:
>
>> Here, i rename sz_damon_region() to damon_region_size(), and move it to
>> "include/linux/damon.h", because in many places, we can to use this func.
> Good idea.
>
>> Signed-off-by: Xin Hao <xhao@linux.alibaba.com>
>> ---
>>   include/linux/damon.h | 1 +
>>   mm/damon/core.c       | 6 +++---
>>   2 files changed, 4 insertions(+), 3 deletions(-)
>>
>> diff --git a/include/linux/damon.h b/include/linux/damon.h
>> index ed5470f50bab..21f4bfd0f41f 100644
>> --- a/include/linux/damon.h
>> +++ b/include/linux/damon.h
>> @@ -524,6 +524,7 @@ void damon_add_region(struct damon_region *r, struct damon_target *t);
>>   void damon_destroy_region(struct damon_region *r, struct damon_target *t);
>>   int damon_set_regions(struct damon_target *t, struct damon_addr_range *ranges,
>>   		unsigned int nr_ranges);
>> +unsigned long damon_region_size(struct damon_region *r);
> This name looks good enough, but I'd like to suggest damon_sz_region().  What
> do you think?
Yes,  the fisrt  i named damon_sz_region, but later i changed it to 
damon_region_size, so i will change it back.
>
> Also, I'd like to keep this function 'static inline' and put the definition
> just after that of 'damon_first_region()'.
Ok, thanks
>
>
> Thanks,
> SJ
>
>>   struct damos *damon_new_scheme(struct damos_access_pattern *pattern,
>>   			enum damos_action action, struct damos_quota *quota,
>> diff --git a/mm/damon/core.c b/mm/damon/core.c
>> index 4de8c7c52979..74ab45b2c2f1 100644
>> --- a/mm/damon/core.c
>> +++ b/mm/damon/core.c
>> @@ -864,7 +864,7 @@ static void kdamond_apply_schemes(struct damon_ctx *c)
>>   	}
>>   }
>>
>> -static inline unsigned long sz_damon_region(struct damon_region *r)
>> +unsigned long damon_region_size(struct damon_region *r)
>>   {
>>   	return r->ar.end - r->ar.start;
>>   }
>> @@ -875,7 +875,7 @@ static inline unsigned long sz_damon_region(struct damon_region *r)
>>   static void damon_merge_two_regions(struct damon_target *t,
>>   		struct damon_region *l, struct damon_region *r)
>>   {
>> -	unsigned long sz_l = sz_damon_region(l), sz_r = sz_damon_region(r);
>> +	unsigned long sz_l = damon_region_size(l), sz_r = damon_region_size(r);
>>
>>   	l->nr_accesses = (l->nr_accesses * sz_l + r->nr_accesses * sz_r) /
>>   			(sz_l + sz_r);
>> @@ -904,7 +904,7 @@ static void damon_merge_regions_of(struct damon_target *t, unsigned int thres,
>>
>>   		if (prev && prev->ar.end == r->ar.start &&
>>   		    abs(prev->nr_accesses - r->nr_accesses) <= thres &&
>> -		    sz_damon_region(prev) + sz_damon_region(r) <= sz_limit)
>> +		    damon_region_size(prev) + damon_region_size(r) <= sz_limit)
>>   			damon_merge_two_regions(t, prev, r);
>>   		else
>>   			prev = r;
>> --
>> 2.31.0
>>


^ permalink raw reply	[flat|nested] 4+ messages in thread

end of thread, other threads:[~2022-09-27  0:04 UTC | newest]

Thread overview: 4+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2022-09-26  7:10 [PATCH v1 1/2] mm/damon: rename sz_damon_region to damon_region_size Xin Hao
2022-09-26  7:11 ` [PATCH v1 2/2] mm/damon: use damon_region_size() in appropriate place Xin Hao
2022-09-26 16:16 ` [PATCH v1 1/2] mm/damon: rename sz_damon_region to damon_region_size SeongJae Park
2022-09-27  0:04   ` haoxin

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).