linux-kernel.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
* [PATCH v2] locking/rwsem: Disable preemption while trying for rwsem lock
@ 2022-09-08 18:24 Mukesh Ojha
  2022-09-08 18:30 ` Waiman Long
                   ` (2 more replies)
  0 siblings, 3 replies; 9+ messages in thread
From: Mukesh Ojha @ 2022-09-08 18:24 UTC (permalink / raw)
  To: peterz, mingo, will, longman, boqun.feng
  Cc: linux-kernel, Gokul krishna Krishnakumar, Mukesh Ojha

From: Gokul krishna Krishnakumar <quic_gokukris@quicinc.com>

Make the region inside the rwsem_write_trylock non preemptible.

We observe RT task is hogging CPU when trying to acquire rwsem lock
which was acquired by a kworker task but before the rwsem owner was set.

Here is the scenario:
1. CFS task (affined to a particular CPU) takes rwsem lock.

2. CFS task gets preempted by a RT task before setting owner.

3. RT task (FIFO) is trying to acquire the lock, but spinning until
RT throttling happens for the lock as the lock was taken by CFS task.

This patch attempts to fix the above issue by disabling preemption
until owner is set for the lock. While at it also fix the issues
at the places where rwsem_{set,clear}_owner() are called.

This also adds lockdep annotation of preemption disable in
rwsem_{set,clear}_owner() on Peter Z. suggestion.

Signed-off-by: Gokul krishna Krishnakumar <quic_gokukris@quicinc.com>
Signed-off-by: Mukesh Ojha <quic_mojha@quicinc.com>
---
Changes in v2:
 - Remove preempt disable code in rwsem_try_write_lock_unqueued()
 - Addressed suggestion from Peter Z.
 - Modified commit text
 kernel/locking/rwsem.c | 14 ++++++++++++--
 1 file changed, 12 insertions(+), 2 deletions(-)

diff --git a/kernel/locking/rwsem.c b/kernel/locking/rwsem.c
index 65f0262..4487359 100644
--- a/kernel/locking/rwsem.c
+++ b/kernel/locking/rwsem.c
@@ -133,14 +133,19 @@
  * the owner value concurrently without lock. Read from owner, however,
  * may not need READ_ONCE() as long as the pointer value is only used
  * for comparison and isn't being dereferenced.
+ *
+ * Both rwsem_{set,clear}_owner() functions should be in the same
+ * preempt disable section as the atomic op that changes sem->count.
  */
 static inline void rwsem_set_owner(struct rw_semaphore *sem)
 {
+	lockdep_assert_preemption_disabled();
 	atomic_long_set(&sem->owner, (long)current);
 }
 
 static inline void rwsem_clear_owner(struct rw_semaphore *sem)
 {
+	lockdep_assert_preemption_disabled();
 	atomic_long_set(&sem->owner, 0);
 }
 
@@ -251,13 +256,16 @@ static inline bool rwsem_read_trylock(struct rw_semaphore *sem, long *cntp)
 static inline bool rwsem_write_trylock(struct rw_semaphore *sem)
 {
 	long tmp = RWSEM_UNLOCKED_VALUE;
+	bool ret = false;
 
+	preempt_disable();
 	if (atomic_long_try_cmpxchg_acquire(&sem->count, &tmp, RWSEM_WRITER_LOCKED)) {
 		rwsem_set_owner(sem);
-		return true;
+		ret = true;
 	}
 
-	return false;
+	preempt_enable();
+	return ret;
 }
 
 /*
@@ -1352,8 +1360,10 @@ static inline void __up_write(struct rw_semaphore *sem)
 	DEBUG_RWSEMS_WARN_ON((rwsem_owner(sem) != current) &&
 			    !rwsem_test_oflags(sem, RWSEM_NONSPINNABLE), sem);
 
+	preempt_disable();
 	rwsem_clear_owner(sem);
 	tmp = atomic_long_fetch_add_release(-RWSEM_WRITER_LOCKED, &sem->count);
+	preempt_enable();
 	if (unlikely(tmp & RWSEM_FLAG_WAITERS))
 		rwsem_wake(sem);
 }
-- 
2.7.4


^ permalink raw reply related	[flat|nested] 9+ messages in thread

* Re: [PATCH v2] locking/rwsem: Disable preemption while trying for rwsem lock
  2022-09-08 18:24 [PATCH v2] locking/rwsem: Disable preemption while trying for rwsem lock Mukesh Ojha
@ 2022-09-08 18:30 ` Waiman Long
  2022-09-15 14:23 ` [tip: locking/core] " tip-bot2 for Gokul krishna Krishnakumar
  2024-03-05 11:04 ` [PATCH v2] " Aaro Koskinen
  2 siblings, 0 replies; 9+ messages in thread
From: Waiman Long @ 2022-09-08 18:30 UTC (permalink / raw)
  To: Mukesh Ojha, peterz, mingo, will, boqun.feng
  Cc: linux-kernel, Gokul krishna Krishnakumar

On 9/8/22 14:24, Mukesh Ojha wrote:
> From: Gokul krishna Krishnakumar <quic_gokukris@quicinc.com>
>
> Make the region inside the rwsem_write_trylock non preemptible.
>
> We observe RT task is hogging CPU when trying to acquire rwsem lock
> which was acquired by a kworker task but before the rwsem owner was set.
>
> Here is the scenario:
> 1. CFS task (affined to a particular CPU) takes rwsem lock.
>
> 2. CFS task gets preempted by a RT task before setting owner.
>
> 3. RT task (FIFO) is trying to acquire the lock, but spinning until
> RT throttling happens for the lock as the lock was taken by CFS task.
>
> This patch attempts to fix the above issue by disabling preemption
> until owner is set for the lock. While at it also fix the issues
> at the places where rwsem_{set,clear}_owner() are called.
>
> This also adds lockdep annotation of preemption disable in
> rwsem_{set,clear}_owner() on Peter Z. suggestion.
>
> Signed-off-by: Gokul krishna Krishnakumar <quic_gokukris@quicinc.com>
> Signed-off-by: Mukesh Ojha <quic_mojha@quicinc.com>
> ---
> Changes in v2:
>   - Remove preempt disable code in rwsem_try_write_lock_unqueued()
>   - Addressed suggestion from Peter Z.
>   - Modified commit text
>   kernel/locking/rwsem.c | 14 ++++++++++++--
>   1 file changed, 12 insertions(+), 2 deletions(-)
>
> diff --git a/kernel/locking/rwsem.c b/kernel/locking/rwsem.c
> index 65f0262..4487359 100644
> --- a/kernel/locking/rwsem.c
> +++ b/kernel/locking/rwsem.c
> @@ -133,14 +133,19 @@
>    * the owner value concurrently without lock. Read from owner, however,
>    * may not need READ_ONCE() as long as the pointer value is only used
>    * for comparison and isn't being dereferenced.
> + *
> + * Both rwsem_{set,clear}_owner() functions should be in the same
> + * preempt disable section as the atomic op that changes sem->count.
>    */
>   static inline void rwsem_set_owner(struct rw_semaphore *sem)
>   {
> +	lockdep_assert_preemption_disabled();
>   	atomic_long_set(&sem->owner, (long)current);
>   }
>   
>   static inline void rwsem_clear_owner(struct rw_semaphore *sem)
>   {
> +	lockdep_assert_preemption_disabled();
>   	atomic_long_set(&sem->owner, 0);
>   }
>   
> @@ -251,13 +256,16 @@ static inline bool rwsem_read_trylock(struct rw_semaphore *sem, long *cntp)
>   static inline bool rwsem_write_trylock(struct rw_semaphore *sem)
>   {
>   	long tmp = RWSEM_UNLOCKED_VALUE;
> +	bool ret = false;
>   
> +	preempt_disable();
>   	if (atomic_long_try_cmpxchg_acquire(&sem->count, &tmp, RWSEM_WRITER_LOCKED)) {
>   		rwsem_set_owner(sem);
> -		return true;
> +		ret = true;
>   	}
>   
> -	return false;
> +	preempt_enable();
> +	return ret;
>   }
>   
>   /*
> @@ -1352,8 +1360,10 @@ static inline void __up_write(struct rw_semaphore *sem)
>   	DEBUG_RWSEMS_WARN_ON((rwsem_owner(sem) != current) &&
>   			    !rwsem_test_oflags(sem, RWSEM_NONSPINNABLE), sem);
>   
> +	preempt_disable();
>   	rwsem_clear_owner(sem);
>   	tmp = atomic_long_fetch_add_release(-RWSEM_WRITER_LOCKED, &sem->count);
> +	preempt_enable();
>   	if (unlikely(tmp & RWSEM_FLAG_WAITERS))
>   		rwsem_wake(sem);
>   }
Reviewed-by: Waiman Long <longman@redhat.com>

Cheers,
Longman


^ permalink raw reply	[flat|nested] 9+ messages in thread

* [tip: locking/core] locking/rwsem: Disable preemption while trying for rwsem lock
  2022-09-08 18:24 [PATCH v2] locking/rwsem: Disable preemption while trying for rwsem lock Mukesh Ojha
  2022-09-08 18:30 ` Waiman Long
@ 2022-09-15 14:23 ` tip-bot2 for Gokul krishna Krishnakumar
  2024-03-05 11:04 ` [PATCH v2] " Aaro Koskinen
  2 siblings, 0 replies; 9+ messages in thread
From: tip-bot2 for Gokul krishna Krishnakumar @ 2022-09-15 14:23 UTC (permalink / raw)
  To: linux-tip-commits
  Cc: Gokul krishna Krishnakumar, Mukesh Ojha, Peter Zijlstra (Intel),
	Waiman Long, x86, linux-kernel

The following commit has been merged into the locking/core branch of tip:

Commit-ID:     48dfb5d2560d36fb16c7d430c229d1604ea7d185
Gitweb:        https://git.kernel.org/tip/48dfb5d2560d36fb16c7d430c229d1604ea7d185
Author:        Gokul krishna Krishnakumar <quic_gokukris@quicinc.com>
AuthorDate:    Thu, 08 Sep 2022 23:54:27 +05:30
Committer:     Peter Zijlstra <peterz@infradead.org>
CommitterDate: Thu, 15 Sep 2022 16:14:02 +02:00

locking/rwsem: Disable preemption while trying for rwsem lock

Make the region inside the rwsem_write_trylock non preemptible.

We observe RT task is hogging CPU when trying to acquire rwsem lock
which was acquired by a kworker task but before the rwsem owner was set.

Here is the scenario:
1. CFS task (affined to a particular CPU) takes rwsem lock.

2. CFS task gets preempted by a RT task before setting owner.

3. RT task (FIFO) is trying to acquire the lock, but spinning until
RT throttling happens for the lock as the lock was taken by CFS task.

This patch attempts to fix the above issue by disabling preemption
until owner is set for the lock. While at it also fix the issues
at the places where rwsem_{set,clear}_owner() are called.

This also adds lockdep annotation of preemption disable in
rwsem_{set,clear}_owner() on Peter Z. suggestion.

Signed-off-by: Gokul krishna Krishnakumar <quic_gokukris@quicinc.com>
Signed-off-by: Mukesh Ojha <quic_mojha@quicinc.com>
Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Reviewed-by: Waiman Long <longman@redhat.com>
Link: https://lore.kernel.org/r/1662661467-24203-1-git-send-email-quic_mojha@quicinc.com
---
 kernel/locking/rwsem.c | 14 ++++++++++++--
 1 file changed, 12 insertions(+), 2 deletions(-)

diff --git a/kernel/locking/rwsem.c b/kernel/locking/rwsem.c
index 65f0262..4487359 100644
--- a/kernel/locking/rwsem.c
+++ b/kernel/locking/rwsem.c
@@ -133,14 +133,19 @@
  * the owner value concurrently without lock. Read from owner, however,
  * may not need READ_ONCE() as long as the pointer value is only used
  * for comparison and isn't being dereferenced.
+ *
+ * Both rwsem_{set,clear}_owner() functions should be in the same
+ * preempt disable section as the atomic op that changes sem->count.
  */
 static inline void rwsem_set_owner(struct rw_semaphore *sem)
 {
+	lockdep_assert_preemption_disabled();
 	atomic_long_set(&sem->owner, (long)current);
 }
 
 static inline void rwsem_clear_owner(struct rw_semaphore *sem)
 {
+	lockdep_assert_preemption_disabled();
 	atomic_long_set(&sem->owner, 0);
 }
 
@@ -251,13 +256,16 @@ static inline bool rwsem_read_trylock(struct rw_semaphore *sem, long *cntp)
 static inline bool rwsem_write_trylock(struct rw_semaphore *sem)
 {
 	long tmp = RWSEM_UNLOCKED_VALUE;
+	bool ret = false;
 
+	preempt_disable();
 	if (atomic_long_try_cmpxchg_acquire(&sem->count, &tmp, RWSEM_WRITER_LOCKED)) {
 		rwsem_set_owner(sem);
-		return true;
+		ret = true;
 	}
 
-	return false;
+	preempt_enable();
+	return ret;
 }
 
 /*
@@ -1352,8 +1360,10 @@ static inline void __up_write(struct rw_semaphore *sem)
 	DEBUG_RWSEMS_WARN_ON((rwsem_owner(sem) != current) &&
 			    !rwsem_test_oflags(sem, RWSEM_NONSPINNABLE), sem);
 
+	preempt_disable();
 	rwsem_clear_owner(sem);
 	tmp = atomic_long_fetch_add_release(-RWSEM_WRITER_LOCKED, &sem->count);
+	preempt_enable();
 	if (unlikely(tmp & RWSEM_FLAG_WAITERS))
 		rwsem_wake(sem);
 }

^ permalink raw reply related	[flat|nested] 9+ messages in thread

* Re: [PATCH v2] locking/rwsem: Disable preemption while trying for rwsem lock
  2022-09-08 18:24 [PATCH v2] locking/rwsem: Disable preemption while trying for rwsem lock Mukesh Ojha
  2022-09-08 18:30 ` Waiman Long
  2022-09-15 14:23 ` [tip: locking/core] " tip-bot2 for Gokul krishna Krishnakumar
@ 2024-03-05 11:04 ` Aaro Koskinen
  2024-03-05 14:55   ` Waiman Long
  2 siblings, 1 reply; 9+ messages in thread
From: Aaro Koskinen @ 2024-03-05 11:04 UTC (permalink / raw)
  To: Mukesh Ojha, Peter Zijlstra
  Cc: mingo, will, longman, boqun.feng, linux-kernel,
	Gokul krishna Krishnakumar

Hi,

It seems this patch (commit 48dfb5d2560d) is missing from
at least 5.15 stable tree.

Based on quick test, it seems to fix an issue where system locks up
easily when RT throttling is disabled, and also it applies cleanly, so
I think it should be good to have it 5.15?

A.

On Thu, Sep 08, 2022 at 11:54:27PM +0530, Mukesh Ojha wrote:
> From: Gokul krishna Krishnakumar <quic_gokukris@quicinc.com>
> 
> Make the region inside the rwsem_write_trylock non preemptible.
> 
> We observe RT task is hogging CPU when trying to acquire rwsem lock
> which was acquired by a kworker task but before the rwsem owner was set.
> 
> Here is the scenario:
> 1. CFS task (affined to a particular CPU) takes rwsem lock.
> 
> 2. CFS task gets preempted by a RT task before setting owner.
> 
> 3. RT task (FIFO) is trying to acquire the lock, but spinning until
> RT throttling happens for the lock as the lock was taken by CFS task.
> 
> This patch attempts to fix the above issue by disabling preemption
> until owner is set for the lock. While at it also fix the issues
> at the places where rwsem_{set,clear}_owner() are called.
> 
> This also adds lockdep annotation of preemption disable in
> rwsem_{set,clear}_owner() on Peter Z. suggestion.
> 
> Signed-off-by: Gokul krishna Krishnakumar <quic_gokukris@quicinc.com>
> Signed-off-by: Mukesh Ojha <quic_mojha@quicinc.com>
> ---
> Changes in v2:
>  - Remove preempt disable code in rwsem_try_write_lock_unqueued()
>  - Addressed suggestion from Peter Z.
>  - Modified commit text
>  kernel/locking/rwsem.c | 14 ++++++++++++--
>  1 file changed, 12 insertions(+), 2 deletions(-)
> 
> diff --git a/kernel/locking/rwsem.c b/kernel/locking/rwsem.c
> index 65f0262..4487359 100644
> --- a/kernel/locking/rwsem.c
> +++ b/kernel/locking/rwsem.c
> @@ -133,14 +133,19 @@
>   * the owner value concurrently without lock. Read from owner, however,
>   * may not need READ_ONCE() as long as the pointer value is only used
>   * for comparison and isn't being dereferenced.
> + *
> + * Both rwsem_{set,clear}_owner() functions should be in the same
> + * preempt disable section as the atomic op that changes sem->count.
>   */
>  static inline void rwsem_set_owner(struct rw_semaphore *sem)
>  {
> +	lockdep_assert_preemption_disabled();
>  	atomic_long_set(&sem->owner, (long)current);
>  }
>  
>  static inline void rwsem_clear_owner(struct rw_semaphore *sem)
>  {
> +	lockdep_assert_preemption_disabled();
>  	atomic_long_set(&sem->owner, 0);
>  }
>  
> @@ -251,13 +256,16 @@ static inline bool rwsem_read_trylock(struct rw_semaphore *sem, long *cntp)
>  static inline bool rwsem_write_trylock(struct rw_semaphore *sem)
>  {
>  	long tmp = RWSEM_UNLOCKED_VALUE;
> +	bool ret = false;
>  
> +	preempt_disable();
>  	if (atomic_long_try_cmpxchg_acquire(&sem->count, &tmp, RWSEM_WRITER_LOCKED)) {
>  		rwsem_set_owner(sem);
> -		return true;
> +		ret = true;
>  	}
>  
> -	return false;
> +	preempt_enable();
> +	return ret;
>  }
>  
>  /*
> @@ -1352,8 +1360,10 @@ static inline void __up_write(struct rw_semaphore *sem)
>  	DEBUG_RWSEMS_WARN_ON((rwsem_owner(sem) != current) &&
>  			    !rwsem_test_oflags(sem, RWSEM_NONSPINNABLE), sem);
>  
> +	preempt_disable();
>  	rwsem_clear_owner(sem);
>  	tmp = atomic_long_fetch_add_release(-RWSEM_WRITER_LOCKED, &sem->count);
> +	preempt_enable();
>  	if (unlikely(tmp & RWSEM_FLAG_WAITERS))
>  		rwsem_wake(sem);
>  }
> -- 
> 2.7.4
> 
> 

^ permalink raw reply	[flat|nested] 9+ messages in thread

* Re: [PATCH v2] locking/rwsem: Disable preemption while trying for rwsem lock
  2024-03-05 11:04 ` [PATCH v2] " Aaro Koskinen
@ 2024-03-05 14:55   ` Waiman Long
  2024-04-02 17:30     ` Aaro Koskinen
  0 siblings, 1 reply; 9+ messages in thread
From: Waiman Long @ 2024-03-05 14:55 UTC (permalink / raw)
  To: Aaro Koskinen, Mukesh Ojha, Peter Zijlstra, Greg KH
  Cc: mingo, will, boqun.feng, linux-kernel,
	Gokul krishna Krishnakumar, stable

On 3/5/24 06:04, Aaro Koskinen wrote:
> Hi,
>
> It seems this patch (commit 48dfb5d2560d) is missing from
> at least 5.15 stable tree.
>
> Based on quick test, it seems to fix an issue where system locks up
> easily when RT throttling is disabled, and also it applies cleanly, so
> I think it should be good to have it 5.15?

You need to cc stable as the locking maintainers are not responsible for 
merging patches to stable trees.

Cheers,
Longman

>
> A.
>
> On Thu, Sep 08, 2022 at 11:54:27PM +0530, Mukesh Ojha wrote:
>> From: Gokul krishna Krishnakumar <quic_gokukris@quicinc.com>
>>
>> Make the region inside the rwsem_write_trylock non preemptible.
>>
>> We observe RT task is hogging CPU when trying to acquire rwsem lock
>> which was acquired by a kworker task but before the rwsem owner was set.
>>
>> Here is the scenario:
>> 1. CFS task (affined to a particular CPU) takes rwsem lock.
>>
>> 2. CFS task gets preempted by a RT task before setting owner.
>>
>> 3. RT task (FIFO) is trying to acquire the lock, but spinning until
>> RT throttling happens for the lock as the lock was taken by CFS task.
>>
>> This patch attempts to fix the above issue by disabling preemption
>> until owner is set for the lock. While at it also fix the issues
>> at the places where rwsem_{set,clear}_owner() are called.
>>
>> This also adds lockdep annotation of preemption disable in
>> rwsem_{set,clear}_owner() on Peter Z. suggestion.
>>
>> Signed-off-by: Gokul krishna Krishnakumar <quic_gokukris@quicinc.com>
>> Signed-off-by: Mukesh Ojha <quic_mojha@quicinc.com>
>> ---
>> Changes in v2:
>>   - Remove preempt disable code in rwsem_try_write_lock_unqueued()
>>   - Addressed suggestion from Peter Z.
>>   - Modified commit text
>>   kernel/locking/rwsem.c | 14 ++++++++++++--
>>   1 file changed, 12 insertions(+), 2 deletions(-)
>>
>> diff --git a/kernel/locking/rwsem.c b/kernel/locking/rwsem.c
>> index 65f0262..4487359 100644
>> --- a/kernel/locking/rwsem.c
>> +++ b/kernel/locking/rwsem.c
>> @@ -133,14 +133,19 @@
>>    * the owner value concurrently without lock. Read from owner, however,
>>    * may not need READ_ONCE() as long as the pointer value is only used
>>    * for comparison and isn't being dereferenced.
>> + *
>> + * Both rwsem_{set,clear}_owner() functions should be in the same
>> + * preempt disable section as the atomic op that changes sem->count.
>>    */
>>   static inline void rwsem_set_owner(struct rw_semaphore *sem)
>>   {
>> +	lockdep_assert_preemption_disabled();
>>   	atomic_long_set(&sem->owner, (long)current);
>>   }
>>   
>>   static inline void rwsem_clear_owner(struct rw_semaphore *sem)
>>   {
>> +	lockdep_assert_preemption_disabled();
>>   	atomic_long_set(&sem->owner, 0);
>>   }
>>   
>> @@ -251,13 +256,16 @@ static inline bool rwsem_read_trylock(struct rw_semaphore *sem, long *cntp)
>>   static inline bool rwsem_write_trylock(struct rw_semaphore *sem)
>>   {
>>   	long tmp = RWSEM_UNLOCKED_VALUE;
>> +	bool ret = false;
>>   
>> +	preempt_disable();
>>   	if (atomic_long_try_cmpxchg_acquire(&sem->count, &tmp, RWSEM_WRITER_LOCKED)) {
>>   		rwsem_set_owner(sem);
>> -		return true;
>> +		ret = true;
>>   	}
>>   
>> -	return false;
>> +	preempt_enable();
>> +	return ret;
>>   }
>>   
>>   /*
>> @@ -1352,8 +1360,10 @@ static inline void __up_write(struct rw_semaphore *sem)
>>   	DEBUG_RWSEMS_WARN_ON((rwsem_owner(sem) != current) &&
>>   			    !rwsem_test_oflags(sem, RWSEM_NONSPINNABLE), sem);
>>   
>> +	preempt_disable();
>>   	rwsem_clear_owner(sem);
>>   	tmp = atomic_long_fetch_add_release(-RWSEM_WRITER_LOCKED, &sem->count);
>> +	preempt_enable();
>>   	if (unlikely(tmp & RWSEM_FLAG_WAITERS))
>>   		rwsem_wake(sem);
>>   }
>> -- 
>> 2.7.4
>>
>>


^ permalink raw reply	[flat|nested] 9+ messages in thread

* Re: [PATCH v2] locking/rwsem: Disable preemption while trying for rwsem lock
  2024-03-05 14:55   ` Waiman Long
@ 2024-04-02 17:30     ` Aaro Koskinen
  2024-04-02 17:42       ` Greg KH
  0 siblings, 1 reply; 9+ messages in thread
From: Aaro Koskinen @ 2024-04-02 17:30 UTC (permalink / raw)
  To: stable, Greg KH
  Cc: Waiman Long, Mukesh Ojha, Peter Zijlstra, Greg KH, mingo, will,
	boqun.feng, linux-kernel, Gokul krishna Krishnakumar, stable

Stable team,

Please cherry-pick this patch into v5.15 stable:

	locking/rwsem: Disable preemption while trying for rwsem lock

	commit 48dfb5d2560d36fb16c7d430c229d1604ea7d185

It fixes the following bug present in v5.15:

> > > From: Gokul krishna Krishnakumar <quic_gokukris@quicinc.com>
> > > We observe RT task is hogging CPU when trying to acquire rwsem lock
> > > which was acquired by a kworker task but before the rwsem owner was set.
> > > 
> > > Here is the scenario:
> > > 1. CFS task (affined to a particular CPU) takes rwsem lock.
> > > 
> > > 2. CFS task gets preempted by a RT task before setting owner.
> > > 
> > > 3. RT task (FIFO) is trying to acquire the lock, but spinning until
> > > RT throttling happens for the lock as the lock was taken by CFS task.

If the RT throttling is disabled, the RT task will remain looping forever
in the kernel. If the system is UP, it will lock up completely.

The issue can be easily reproduced by running RT task and normal task which
are affined to the same CPU core.

A.

On Tue, Mar 05, 2024 at 09:55:47AM -0500, Waiman Long wrote:
> On 3/5/24 06:04, Aaro Koskinen wrote:
> > Hi,
> > 
> > It seems this patch (commit 48dfb5d2560d) is missing from
> > at least 5.15 stable tree.
> > 
> > Based on quick test, it seems to fix an issue where system locks up
> > easily when RT throttling is disabled, and also it applies cleanly, so
> > I think it should be good to have it 5.15?
> 
> You need to cc stable as the locking maintainers are not responsible for
> merging patches to stable trees.
> 
> Cheers,
> Longman
> 
> > 
> > A.
> > 
> > On Thu, Sep 08, 2022 at 11:54:27PM +0530, Mukesh Ojha wrote:
> > > From: Gokul krishna Krishnakumar <quic_gokukris@quicinc.com>
> > > 
> > > Make the region inside the rwsem_write_trylock non preemptible.
> > > 
> > > We observe RT task is hogging CPU when trying to acquire rwsem lock
> > > which was acquired by a kworker task but before the rwsem owner was set.
> > > 
> > > Here is the scenario:
> > > 1. CFS task (affined to a particular CPU) takes rwsem lock.
> > > 
> > > 2. CFS task gets preempted by a RT task before setting owner.
> > > 
> > > 3. RT task (FIFO) is trying to acquire the lock, but spinning until
> > > RT throttling happens for the lock as the lock was taken by CFS task.
> > > 
> > > This patch attempts to fix the above issue by disabling preemption
> > > until owner is set for the lock. While at it also fix the issues
> > > at the places where rwsem_{set,clear}_owner() are called.
> > > 
> > > This also adds lockdep annotation of preemption disable in
> > > rwsem_{set,clear}_owner() on Peter Z. suggestion.
> > > 
> > > Signed-off-by: Gokul krishna Krishnakumar <quic_gokukris@quicinc.com>
> > > Signed-off-by: Mukesh Ojha <quic_mojha@quicinc.com>
> > > ---
> > > Changes in v2:
> > >   - Remove preempt disable code in rwsem_try_write_lock_unqueued()
> > >   - Addressed suggestion from Peter Z.
> > >   - Modified commit text
> > >   kernel/locking/rwsem.c | 14 ++++++++++++--
> > >   1 file changed, 12 insertions(+), 2 deletions(-)
> > > 
> > > diff --git a/kernel/locking/rwsem.c b/kernel/locking/rwsem.c
> > > index 65f0262..4487359 100644
> > > --- a/kernel/locking/rwsem.c
> > > +++ b/kernel/locking/rwsem.c
> > > @@ -133,14 +133,19 @@
> > >    * the owner value concurrently without lock. Read from owner, however,
> > >    * may not need READ_ONCE() as long as the pointer value is only used
> > >    * for comparison and isn't being dereferenced.
> > > + *
> > > + * Both rwsem_{set,clear}_owner() functions should be in the same
> > > + * preempt disable section as the atomic op that changes sem->count.
> > >    */
> > >   static inline void rwsem_set_owner(struct rw_semaphore *sem)
> > >   {
> > > +	lockdep_assert_preemption_disabled();
> > >   	atomic_long_set(&sem->owner, (long)current);
> > >   }
> > >   static inline void rwsem_clear_owner(struct rw_semaphore *sem)
> > >   {
> > > +	lockdep_assert_preemption_disabled();
> > >   	atomic_long_set(&sem->owner, 0);
> > >   }
> > > @@ -251,13 +256,16 @@ static inline bool rwsem_read_trylock(struct rw_semaphore *sem, long *cntp)
> > >   static inline bool rwsem_write_trylock(struct rw_semaphore *sem)
> > >   {
> > >   	long tmp = RWSEM_UNLOCKED_VALUE;
> > > +	bool ret = false;
> > > +	preempt_disable();
> > >   	if (atomic_long_try_cmpxchg_acquire(&sem->count, &tmp, RWSEM_WRITER_LOCKED)) {
> > >   		rwsem_set_owner(sem);
> > > -		return true;
> > > +		ret = true;
> > >   	}
> > > -	return false;
> > > +	preempt_enable();
> > > +	return ret;
> > >   }
> > >   /*
> > > @@ -1352,8 +1360,10 @@ static inline void __up_write(struct rw_semaphore *sem)
> > >   	DEBUG_RWSEMS_WARN_ON((rwsem_owner(sem) != current) &&
> > >   			    !rwsem_test_oflags(sem, RWSEM_NONSPINNABLE), sem);
> > > +	preempt_disable();
> > >   	rwsem_clear_owner(sem);
> > >   	tmp = atomic_long_fetch_add_release(-RWSEM_WRITER_LOCKED, &sem->count);
> > > +	preempt_enable();
> > >   	if (unlikely(tmp & RWSEM_FLAG_WAITERS))
> > >   		rwsem_wake(sem);
> > >   }
> > > -- 
> > > 2.7.4
> > > 
> > > 
> 
> 

^ permalink raw reply	[flat|nested] 9+ messages in thread

* Re: [PATCH v2] locking/rwsem: Disable preemption while trying for rwsem lock
  2024-04-02 17:30     ` Aaro Koskinen
@ 2024-04-02 17:42       ` Greg KH
  2024-04-02 18:25         ` Aaro Koskinen
  0 siblings, 1 reply; 9+ messages in thread
From: Greg KH @ 2024-04-02 17:42 UTC (permalink / raw)
  To: Aaro Koskinen
  Cc: stable, Waiman Long, Mukesh Ojha, Peter Zijlstra, mingo, will,
	boqun.feng, linux-kernel, Gokul krishna Krishnakumar

On Tue, Apr 02, 2024 at 08:30:32PM +0300, Aaro Koskinen wrote:
> Stable team,
> 
> Please cherry-pick this patch into v5.15 stable:
> 
> 	locking/rwsem: Disable preemption while trying for rwsem lock
> 
> 	commit 48dfb5d2560d36fb16c7d430c229d1604ea7d185
> 
> It fixes the following bug present in v5.15:
> 
> > > > From: Gokul krishna Krishnakumar <quic_gokukris@quicinc.com>
> > > > We observe RT task is hogging CPU when trying to acquire rwsem lock
> > > > which was acquired by a kworker task but before the rwsem owner was set.
> > > > 
> > > > Here is the scenario:
> > > > 1. CFS task (affined to a particular CPU) takes rwsem lock.
> > > > 
> > > > 2. CFS task gets preempted by a RT task before setting owner.
> > > > 
> > > > 3. RT task (FIFO) is trying to acquire the lock, but spinning until
> > > > RT throttling happens for the lock as the lock was taken by CFS task.
> 
> If the RT throttling is disabled, the RT task will remain looping forever
> in the kernel. If the system is UP, it will lock up completely.
> 
> The issue can be easily reproduced by running RT task and normal task which
> are affined to the same CPU core.

Does this only affect the -rt patchset, or is this an issue with a
"clean" kernel release?

thanks,

greg k-h

^ permalink raw reply	[flat|nested] 9+ messages in thread

* Re: [PATCH v2] locking/rwsem: Disable preemption while trying for rwsem lock
  2024-04-02 17:42       ` Greg KH
@ 2024-04-02 18:25         ` Aaro Koskinen
  2024-04-05  6:32           ` Greg KH
  0 siblings, 1 reply; 9+ messages in thread
From: Aaro Koskinen @ 2024-04-02 18:25 UTC (permalink / raw)
  To: Greg KH
  Cc: stable, Waiman Long, Mukesh Ojha, Peter Zijlstra, mingo, will,
	boqun.feng, linux-kernel, Gokul krishna Krishnakumar

Hi,

On Tue, Apr 02, 2024 at 07:42:03PM +0200, Greg KH wrote:
> On Tue, Apr 02, 2024 at 08:30:32PM +0300, Aaro Koskinen wrote:
> > Stable team,
> > 
> > Please cherry-pick this patch into v5.15 stable:
> > 
> > 	locking/rwsem: Disable preemption while trying for rwsem lock
> > 
> > 	commit 48dfb5d2560d36fb16c7d430c229d1604ea7d185
> > 
> > It fixes the following bug present in v5.15:
> > 
> > > > > From: Gokul krishna Krishnakumar <quic_gokukris@quicinc.com>
> > > > > We observe RT task is hogging CPU when trying to acquire rwsem lock
> > > > > which was acquired by a kworker task but before the rwsem owner was set.
> > > > > 
> > > > > Here is the scenario:
> > > > > 1. CFS task (affined to a particular CPU) takes rwsem lock.
> > > > > 
> > > > > 2. CFS task gets preempted by a RT task before setting owner.
> > > > > 
> > > > > 3. RT task (FIFO) is trying to acquire the lock, but spinning until
> > > > > RT throttling happens for the lock as the lock was taken by CFS task.
> > 
> > If the RT throttling is disabled, the RT task will remain looping forever
> > in the kernel. If the system is UP, it will lock up completely.
> > 
> > The issue can be easily reproduced by running RT task and normal task which
> > are affined to the same CPU core.
> 
> Does this only affect the -rt patchset, or is this an issue with a
> "clean" kernel release?

It's with a normal "clean" kernel.

A.

^ permalink raw reply	[flat|nested] 9+ messages in thread

* Re: [PATCH v2] locking/rwsem: Disable preemption while trying for rwsem lock
  2024-04-02 18:25         ` Aaro Koskinen
@ 2024-04-05  6:32           ` Greg KH
  0 siblings, 0 replies; 9+ messages in thread
From: Greg KH @ 2024-04-05  6:32 UTC (permalink / raw)
  To: Aaro Koskinen
  Cc: stable, Waiman Long, Mukesh Ojha, Peter Zijlstra, mingo, will,
	boqun.feng, linux-kernel, Gokul krishna Krishnakumar

On Tue, Apr 02, 2024 at 09:25:11PM +0300, Aaro Koskinen wrote:
> Hi,
> 
> On Tue, Apr 02, 2024 at 07:42:03PM +0200, Greg KH wrote:
> > On Tue, Apr 02, 2024 at 08:30:32PM +0300, Aaro Koskinen wrote:
> > > Stable team,
> > > 
> > > Please cherry-pick this patch into v5.15 stable:
> > > 
> > > 	locking/rwsem: Disable preemption while trying for rwsem lock
> > > 
> > > 	commit 48dfb5d2560d36fb16c7d430c229d1604ea7d185
> > > 
> > > It fixes the following bug present in v5.15:
> > > 
> > > > > > From: Gokul krishna Krishnakumar <quic_gokukris@quicinc.com>
> > > > > > We observe RT task is hogging CPU when trying to acquire rwsem lock
> > > > > > which was acquired by a kworker task but before the rwsem owner was set.
> > > > > > 
> > > > > > Here is the scenario:
> > > > > > 1. CFS task (affined to a particular CPU) takes rwsem lock.
> > > > > > 
> > > > > > 2. CFS task gets preempted by a RT task before setting owner.
> > > > > > 
> > > > > > 3. RT task (FIFO) is trying to acquire the lock, but spinning until
> > > > > > RT throttling happens for the lock as the lock was taken by CFS task.
> > > 
> > > If the RT throttling is disabled, the RT task will remain looping forever
> > > in the kernel. If the system is UP, it will lock up completely.
> > > 
> > > The issue can be easily reproduced by running RT task and normal task which
> > > are affined to the same CPU core.
> > 
> > Does this only affect the -rt patchset, or is this an issue with a
> > "clean" kernel release?
> 
> It's with a normal "clean" kernel.

Fair enough, now queued up, thanks.

greg k-h

^ permalink raw reply	[flat|nested] 9+ messages in thread

end of thread, other threads:[~2024-04-05  6:32 UTC | newest]

Thread overview: 9+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2022-09-08 18:24 [PATCH v2] locking/rwsem: Disable preemption while trying for rwsem lock Mukesh Ojha
2022-09-08 18:30 ` Waiman Long
2022-09-15 14:23 ` [tip: locking/core] " tip-bot2 for Gokul krishna Krishnakumar
2024-03-05 11:04 ` [PATCH v2] " Aaro Koskinen
2024-03-05 14:55   ` Waiman Long
2024-04-02 17:30     ` Aaro Koskinen
2024-04-02 17:42       ` Greg KH
2024-04-02 18:25         ` Aaro Koskinen
2024-04-05  6:32           ` Greg KH

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).