From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-15.8 required=3.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,HEADER_FROM_DIFFERENT_DOMAINS,INCLUDES_CR_TRAILER, INCLUDES_PATCH,MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS,URIBL_BLOCKED autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 053D0C4338F for ; Fri, 30 Jul 2021 14:20:58 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id E04DF60462 for ; Fri, 30 Jul 2021 14:20:57 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S239291AbhG3OU7 (ORCPT ); Fri, 30 Jul 2021 10:20:59 -0400 Received: from Galois.linutronix.de ([193.142.43.55]:55584 "EHLO galois.linutronix.de" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S239170AbhG3ORc (ORCPT ); Fri, 30 Jul 2021 10:17:32 -0400 Message-ID: <20210730135206.301477338@linutronix.de> DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linutronix.de; s=2020; t=1627654646; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding: references:references; bh=8IVT82AeJWY0RxlKDGWc7VOoP+02IvCgg5Oih1Rk2hc=; b=unK2lJDtadGjWrih7vbr4WvG6Q8aBxTFsvuSqmmprHLbKBQi0lPK0x0JiGl3Uk8sPHFXI1 Lfua5Fn8XucWWGabC/69HDRP8W8w8NMEO3eVh7YXGscgRLfN5S5N3isyB2Ra9dgKkWn7ft sidyF04kMcpCJB7wuqcWuId+oCFMxyqkRxRG2IbWzxa+vxstoFPNKxgMQGajilrQlL41xj ZfhvKTOHch4dHOnDoQbIdxmOdu+0qhlGltTVj7MKW0DstydO1shbGvDWElTyEFU2eWgtKD gCeCPRem7TRh0x70wSKY/Ce1ViIPJf0O2upslCC0EtHXpp3/ND3WwizcKBCOBg== DKIM-Signature: v=1; a=ed25519-sha256; c=relaxed/relaxed; d=linutronix.de; s=2020e; t=1627654646; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding: references:references; bh=8IVT82AeJWY0RxlKDGWc7VOoP+02IvCgg5Oih1Rk2hc=; b=42BsIyOPFvCs5cN46S6Ph+dDTJBCwcHI7Mel80hlLuZvwwuUEDJ7Qlbwg155mZEtCvZWPy e26vH2/HmPyG27Cw== Date: Fri, 30 Jul 2021 15:50:27 +0200 From: Thomas Gleixner To: LKML Cc: Peter Zijlstra , Ingo Molnar , Juri Lelli , Steven Rostedt , Daniel Bristot de Oliveira , Will Deacon , Waiman Long , Boqun Feng , Sebastian Andrzej Siewior , Davidlohr Bueso Subject: [patch 20/63] locking/rtmutex: Prepare RT rt_mutex_wake_q for RT locks References: <20210730135007.155909613@linutronix.de> MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-transfer-encoding: 8-bit Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org From: Thomas Gleixner Add a rtlock_task pointer to rt_mutex_wake_q which allows to handle the RT specific wakeup for spin/rwlock waiters. The pointer is just consuming 4/8 bytes on stack so it is provided unconditionaly to avoid #ifdeffery all over the place. No functional change for non-RT enabled kernels. Signed-off-by: Thomas Gleixner --- V2: Make it symmetric (PeterZ) --- include/linux/sched/wake_q.h | 7 ++++++- kernel/locking/rtmutex.c | 12 ++++++++++-- kernel/locking/rtmutex_common.h | 3 +++ kernel/sched/core.c | 4 ++-- 4 files changed, 21 insertions(+), 5 deletions(-) --- --- a/include/linux/sched/wake_q.h +++ b/include/linux/sched/wake_q.h @@ -61,6 +61,11 @@ static inline bool wake_q_empty(struct w extern void wake_q_add(struct wake_q_head *head, struct task_struct *task); extern void wake_q_add_safe(struct wake_q_head *head, struct task_struct *task); -extern void wake_up_q(struct wake_q_head *head); +extern void __wake_up_q(struct wake_q_head *head, unsigned int state); + +static inline void wake_up_q(struct wake_q_head *head) +{ + __wake_up_q(head, TASK_NORMAL); +} #endif /* _LINUX_SCHED_WAKE_Q_H */ --- a/kernel/locking/rtmutex.c +++ b/kernel/locking/rtmutex.c @@ -351,12 +351,20 @@ static __always_inline void rt_mutex_adj static __always_inline void rt_mutex_wake_q_add(struct rt_wake_q_head *wqh, struct rt_mutex_waiter *w) { - wake_q_add(&wqh->head, w->task); + if (IS_ENABLED(CONFIG_PREEMPT_RT) && w->wake_state != TASK_NORMAL) { + wake_q_add(&wqh->rt_head, w->task); + } else { + wake_q_add(&wqh->head, w->task); + } } static __always_inline void rt_mutex_wake_up_q(struct rt_wake_q_head *wqh) { - wake_up_q(&wqh->head); + if (IS_ENABLED(CONFIG_PREEMPT_RT) && !wake_q_empty(&wqh->rt_head)) + __wake_up_q(&wqh->rt_head, TASK_RTLOCK_WAIT); + + if (!wake_q_empty(&wqh->head)) + wake_up_q(&wqh->head); /* Pairs with preempt_disable() in mark_wakeup_next_waiter() */ preempt_enable(); --- a/kernel/locking/rtmutex_common.h +++ b/kernel/locking/rtmutex_common.h @@ -43,14 +43,17 @@ struct rt_mutex_waiter { * rt_wake_q_head - Wrapper around regular wake_q_head to support * "sleeping" spinlocks on RT * @head: The regular wake_q_head for sleeping lock variants + * @rt_head: The wake_q_head for RT lock (spin/rwlock) variants */ struct rt_wake_q_head { struct wake_q_head head; + struct wake_q_head rt_head; }; #define DEFINE_RT_WAKE_Q(name) \ struct rt_wake_q_head name = { \ .head = WAKE_Q_HEAD_INITIALIZER(name.head), \ + .rt_head = WAKE_Q_HEAD_INITIALIZER(name.rt_head),\ } /* --- a/kernel/sched/core.c +++ b/kernel/sched/core.c @@ -916,7 +916,7 @@ void wake_q_add_safe(struct wake_q_head put_task_struct(task); } -void wake_up_q(struct wake_q_head *head) +void __wake_up_q(struct wake_q_head *head, unsigned int state) { struct wake_q_node *node = head->first; @@ -932,7 +932,7 @@ void wake_up_q(struct wake_q_head *head) * wake_up_process() executes a full barrier, which pairs with * the queueing in wake_q_add() so as not to miss wakeups. */ - wake_up_process(task); + wake_up_state(task, state); put_task_struct(task); } }