linux-mm.kvack.org archive mirror
 help / color / mirror / Atom feed
* [PATCH v2] mm/damon: simplify stop mechanism
@ 2021-10-27 13:05 Changbin Du
  2021-10-28  7:08 ` SeongJae Park
  0 siblings, 1 reply; 2+ messages in thread
From: Changbin Du @ 2021-10-27 13:05 UTC (permalink / raw)
  To: Andrew Morton; +Cc: SeongJae Park, linux-mm, linux-kernel, Changbin Du

An kernel thread can exit gracefully with kthread_stop(). So we don't need
a new flag 'kdamond_stop'. And to make sure the task struct is not freed
when accessing it, get reference of it before termination.

Signed-off-by: Changbin Du <changbin.du@gmail.com>
---
 include/linux/damon.h |  1 -
 mm/damon/core.c       | 51 +++++++++++++------------------------------
 2 files changed, 15 insertions(+), 37 deletions(-)

diff --git a/include/linux/damon.h b/include/linux/damon.h
index a14b3cc54cab..50c6eb0dee1f 100644
--- a/include/linux/damon.h
+++ b/include/linux/damon.h
@@ -381,7 +381,6 @@ struct damon_ctx {
 
 /* public: */
 	struct task_struct *kdamond;
-	bool kdamond_stop;
 	struct mutex kdamond_lock;
 
 	struct damon_primitive primitive;
diff --git a/mm/damon/core.c b/mm/damon/core.c
index 46a6afea3030..f37c17b53814 100644
--- a/mm/damon/core.c
+++ b/mm/damon/core.c
@@ -390,17 +390,6 @@ static unsigned long damon_region_sz_limit(struct damon_ctx *ctx)
 	return sz;
 }
 
-static bool damon_kdamond_running(struct damon_ctx *ctx)
-{
-	bool running;
-
-	mutex_lock(&ctx->kdamond_lock);
-	running = ctx->kdamond != NULL;
-	mutex_unlock(&ctx->kdamond_lock);
-
-	return running;
-}
-
 static int kdamond_fn(void *data);
 
 /*
@@ -418,7 +407,6 @@ static int __damon_start(struct damon_ctx *ctx)
 	mutex_lock(&ctx->kdamond_lock);
 	if (!ctx->kdamond) {
 		err = 0;
-		ctx->kdamond_stop = false;
 		ctx->kdamond = kthread_run(kdamond_fn, ctx, "kdamond.%d",
 				nr_running_ctxs);
 		if (IS_ERR(ctx->kdamond)) {
@@ -474,13 +462,15 @@ int damon_start(struct damon_ctx **ctxs, int nr_ctxs)
  */
 static int __damon_stop(struct damon_ctx *ctx)
 {
+	struct task_struct *tsk;
+
 	mutex_lock(&ctx->kdamond_lock);
-	if (ctx->kdamond) {
-		ctx->kdamond_stop = true;
+	tsk = ctx->kdamond;
+	if (tsk) {
+		get_task_struct(tsk);
 		mutex_unlock(&ctx->kdamond_lock);
-		while (damon_kdamond_running(ctx))
-			usleep_range(ctx->sample_interval,
-					ctx->sample_interval * 2);
+		kthread_stop(tsk);
+		put_task_struct(tsk);
 		return 0;
 	}
 	mutex_unlock(&ctx->kdamond_lock);
@@ -925,12 +915,8 @@ static bool kdamond_need_update_primitive(struct damon_ctx *ctx)
 static bool kdamond_need_stop(struct damon_ctx *ctx)
 {
 	struct damon_target *t;
-	bool stop;
 
-	mutex_lock(&ctx->kdamond_lock);
-	stop = ctx->kdamond_stop;
-	mutex_unlock(&ctx->kdamond_lock);
-	if (stop)
+	if (kthread_should_stop())
 		return true;
 
 	if (!ctx->primitive.target_valid)
@@ -1021,13 +1007,6 @@ static int kdamond_wait_activation(struct damon_ctx *ctx)
 	return -EBUSY;
 }
 
-static void set_kdamond_stop(struct damon_ctx *ctx)
-{
-	mutex_lock(&ctx->kdamond_lock);
-	ctx->kdamond_stop = true;
-	mutex_unlock(&ctx->kdamond_lock);
-}
-
 /*
  * The monitoring daemon that runs as a kernel thread
  */
@@ -1038,17 +1017,18 @@ static int kdamond_fn(void *data)
 	struct damon_region *r, *next;
 	unsigned int max_nr_accesses = 0;
 	unsigned long sz_limit = 0;
+	bool done = false;
 
 	pr_debug("kdamond (%d) starts\n", current->pid);
 
 	if (ctx->primitive.init)
 		ctx->primitive.init(ctx);
 	if (ctx->callback.before_start && ctx->callback.before_start(ctx))
-		set_kdamond_stop(ctx);
+		done = true;
 
 	sz_limit = damon_region_sz_limit(ctx);
 
-	while (!kdamond_need_stop(ctx)) {
+	while (!kdamond_need_stop(ctx) && !done) {
 		if (kdamond_wait_activation(ctx))
 			continue;
 
@@ -1056,7 +1036,7 @@ static int kdamond_fn(void *data)
 			ctx->primitive.prepare_access_checks(ctx);
 		if (ctx->callback.after_sampling &&
 				ctx->callback.after_sampling(ctx))
-			set_kdamond_stop(ctx);
+			done = true;
 
 		usleep_range(ctx->sample_interval, ctx->sample_interval + 1);
 
@@ -1069,7 +1049,7 @@ static int kdamond_fn(void *data)
 					sz_limit);
 			if (ctx->callback.after_aggregation &&
 					ctx->callback.after_aggregation(ctx))
-				set_kdamond_stop(ctx);
+				done = true;
 			kdamond_apply_schemes(ctx);
 			kdamond_reset_aggregated(ctx);
 			kdamond_split_regions(ctx);
@@ -1088,9 +1068,8 @@ static int kdamond_fn(void *data)
 			damon_destroy_region(r, t);
 	}
 
-	if (ctx->callback.before_terminate &&
-			ctx->callback.before_terminate(ctx))
-		set_kdamond_stop(ctx);
+	if (ctx->callback.before_terminate)
+		ctx->callback.before_terminate(ctx);
 	if (ctx->primitive.cleanup)
 		ctx->primitive.cleanup(ctx);
 
-- 
2.32.0



^ permalink raw reply related	[flat|nested] 2+ messages in thread

* Re: [PATCH v2] mm/damon: simplify stop mechanism
  2021-10-27 13:05 [PATCH v2] mm/damon: simplify stop mechanism Changbin Du
@ 2021-10-28  7:08 ` SeongJae Park
  0 siblings, 0 replies; 2+ messages in thread
From: SeongJae Park @ 2021-10-28  7:08 UTC (permalink / raw)
  To: Changbin Du; +Cc: Andrew Morton, SeongJae Park, linux-mm, linux-kernel

Thank you for this patch!

On Wed, 27 Oct 2021 21:05:17 +0800 Changbin Du <changbin.du@gmail.com> wrote:

> An kernel thread can exit gracefully with kthread_stop(). So we don't need
> a new flag 'kdamond_stop'. And to make sure the task struct is not freed
> when accessing it, get reference of it before termination.
> 
> Signed-off-by: Changbin Du <changbin.du@gmail.com>

Reviewed-by: SeongJae Park <sj@kernel.org>


Thanks,
SJ

> ---
>  include/linux/damon.h |  1 -
>  mm/damon/core.c       | 51 +++++++++++++------------------------------
>  2 files changed, 15 insertions(+), 37 deletions(-)
> 
> diff --git a/include/linux/damon.h b/include/linux/damon.h
> index a14b3cc54cab..50c6eb0dee1f 100644
> --- a/include/linux/damon.h
> +++ b/include/linux/damon.h
> @@ -381,7 +381,6 @@ struct damon_ctx {
>  
>  /* public: */
>  	struct task_struct *kdamond;
> -	bool kdamond_stop;
>  	struct mutex kdamond_lock;
>  
>  	struct damon_primitive primitive;
> diff --git a/mm/damon/core.c b/mm/damon/core.c
> index 46a6afea3030..f37c17b53814 100644
> --- a/mm/damon/core.c
> +++ b/mm/damon/core.c
> @@ -390,17 +390,6 @@ static unsigned long damon_region_sz_limit(struct damon_ctx *ctx)
>  	return sz;
>  }
>  
> -static bool damon_kdamond_running(struct damon_ctx *ctx)
> -{
> -	bool running;
> -
> -	mutex_lock(&ctx->kdamond_lock);
> -	running = ctx->kdamond != NULL;
> -	mutex_unlock(&ctx->kdamond_lock);
> -
> -	return running;
> -}
> -
>  static int kdamond_fn(void *data);
>  
>  /*
> @@ -418,7 +407,6 @@ static int __damon_start(struct damon_ctx *ctx)
>  	mutex_lock(&ctx->kdamond_lock);
>  	if (!ctx->kdamond) {
>  		err = 0;
> -		ctx->kdamond_stop = false;
>  		ctx->kdamond = kthread_run(kdamond_fn, ctx, "kdamond.%d",
>  				nr_running_ctxs);
>  		if (IS_ERR(ctx->kdamond)) {
> @@ -474,13 +462,15 @@ int damon_start(struct damon_ctx **ctxs, int nr_ctxs)
>   */
>  static int __damon_stop(struct damon_ctx *ctx)
>  {
> +	struct task_struct *tsk;
> +
>  	mutex_lock(&ctx->kdamond_lock);
> -	if (ctx->kdamond) {
> -		ctx->kdamond_stop = true;
> +	tsk = ctx->kdamond;
> +	if (tsk) {
> +		get_task_struct(tsk);
>  		mutex_unlock(&ctx->kdamond_lock);
> -		while (damon_kdamond_running(ctx))
> -			usleep_range(ctx->sample_interval,
> -					ctx->sample_interval * 2);
> +		kthread_stop(tsk);
> +		put_task_struct(tsk);
>  		return 0;
>  	}
>  	mutex_unlock(&ctx->kdamond_lock);
> @@ -925,12 +915,8 @@ static bool kdamond_need_update_primitive(struct damon_ctx *ctx)
>  static bool kdamond_need_stop(struct damon_ctx *ctx)
>  {
>  	struct damon_target *t;
> -	bool stop;
>  
> -	mutex_lock(&ctx->kdamond_lock);
> -	stop = ctx->kdamond_stop;
> -	mutex_unlock(&ctx->kdamond_lock);
> -	if (stop)
> +	if (kthread_should_stop())
>  		return true;
>  
>  	if (!ctx->primitive.target_valid)
> @@ -1021,13 +1007,6 @@ static int kdamond_wait_activation(struct damon_ctx *ctx)
>  	return -EBUSY;
>  }
>  
> -static void set_kdamond_stop(struct damon_ctx *ctx)
> -{
> -	mutex_lock(&ctx->kdamond_lock);
> -	ctx->kdamond_stop = true;
> -	mutex_unlock(&ctx->kdamond_lock);
> -}
> -
>  /*
>   * The monitoring daemon that runs as a kernel thread
>   */
> @@ -1038,17 +1017,18 @@ static int kdamond_fn(void *data)
>  	struct damon_region *r, *next;
>  	unsigned int max_nr_accesses = 0;
>  	unsigned long sz_limit = 0;
> +	bool done = false;
>  
>  	pr_debug("kdamond (%d) starts\n", current->pid);
>  
>  	if (ctx->primitive.init)
>  		ctx->primitive.init(ctx);
>  	if (ctx->callback.before_start && ctx->callback.before_start(ctx))
> -		set_kdamond_stop(ctx);
> +		done = true;
>  
>  	sz_limit = damon_region_sz_limit(ctx);
>  
> -	while (!kdamond_need_stop(ctx)) {
> +	while (!kdamond_need_stop(ctx) && !done) {
>  		if (kdamond_wait_activation(ctx))
>  			continue;
>  
> @@ -1056,7 +1036,7 @@ static int kdamond_fn(void *data)
>  			ctx->primitive.prepare_access_checks(ctx);
>  		if (ctx->callback.after_sampling &&
>  				ctx->callback.after_sampling(ctx))
> -			set_kdamond_stop(ctx);
> +			done = true;
>  
>  		usleep_range(ctx->sample_interval, ctx->sample_interval + 1);
>  
> @@ -1069,7 +1049,7 @@ static int kdamond_fn(void *data)
>  					sz_limit);
>  			if (ctx->callback.after_aggregation &&
>  					ctx->callback.after_aggregation(ctx))
> -				set_kdamond_stop(ctx);
> +				done = true;
>  			kdamond_apply_schemes(ctx);
>  			kdamond_reset_aggregated(ctx);
>  			kdamond_split_regions(ctx);
> @@ -1088,9 +1068,8 @@ static int kdamond_fn(void *data)
>  			damon_destroy_region(r, t);
>  	}
>  
> -	if (ctx->callback.before_terminate &&
> -			ctx->callback.before_terminate(ctx))
> -		set_kdamond_stop(ctx);
> +	if (ctx->callback.before_terminate)
> +		ctx->callback.before_terminate(ctx);
>  	if (ctx->primitive.cleanup)
>  		ctx->primitive.cleanup(ctx);
>  
> -- 
> 2.32.0
> 


^ permalink raw reply	[flat|nested] 2+ messages in thread

end of thread, other threads:[~2021-10-28  7:08 UTC | newest]

Thread overview: 2+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2021-10-27 13:05 [PATCH v2] mm/damon: simplify stop mechanism Changbin Du
2021-10-28  7:08 ` SeongJae Park

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).