From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-15.2 required=3.0 tests=BAYES_00, HEADER_FROM_DIFFERENT_DOMAINS,INCLUDES_CR_TRAILER,INCLUDES_PATCH, MAILING_LIST_MULTI,NICE_REPLY_A,SPF_HELO_NONE,SPF_PASS,URIBL_BLOCKED, USER_AGENT_SANE_1 autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 78CB9C64E75 for ; Thu, 26 Nov 2020 09:27:00 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id 2C1DC21741 for ; Thu, 26 Nov 2020 09:27:00 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S2389395AbgKZJ0r (ORCPT ); Thu, 26 Nov 2020 04:26:47 -0500 Received: from mga12.intel.com ([192.55.52.136]:32765 "EHLO mga12.intel.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S2389372AbgKZJ0n (ORCPT ); Thu, 26 Nov 2020 04:26:43 -0500 IronPort-SDR: +QLlXfdZwG8WkdHDIAKnm/tRocwyNJmuBuSTuleRErw1pN2BatVKWNjmmky4FREq8ewzd3G2BZ JNappPQSV5HA== X-IronPort-AV: E=McAfee;i="6000,8403,9816"; a="151518266" X-IronPort-AV: E=Sophos;i="5.78,371,1599548400"; d="scan'208";a="151518266" X-Amp-Result: SKIPPED(no attachment in message) X-Amp-File-Uploaded: False Received: from fmsmga003.fm.intel.com ([10.253.24.29]) by fmsmga106.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 26 Nov 2020 01:26:42 -0800 IronPort-SDR: sKDM7NH8iPhFij26wvPORfs5Xw3WdN5nzYq2bmiL6uOZgQFMjpuF9ZxOKXsMeOhHiR5dCLkTGq K9sul99wqPwQ== X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.78,371,1599548400"; d="scan'208";a="371218354" Received: from cli6-desk1.ccr.corp.intel.com (HELO [10.239.161.125]) ([10.239.161.125]) by FMSMGA003.fm.intel.com with ESMTP; 26 Nov 2020 01:26:32 -0800 Subject: Re: [PATCH -tip 14/32] sched: migration changes for core scheduling To: Balbir Singh Cc: Peter Zijlstra , "Joel Fernandes (Google)" , Nishanth Aravamudan , Julien Desfossez , Tim Chen , Vineeth Pillai , Aaron Lu , Aubrey Li , tglx@linutronix.de, linux-kernel@vger.kernel.org, mingo@kernel.org, torvalds@linux-foundation.org, fweisbec@gmail.com, keescook@chromium.org, kerrnel@google.com, Phil Auld , Valentin Schneider , Mel Gorman , Pawan Gupta , Paolo Bonzini , vineeth@bitbyteword.org, Chen Yu , Christian Brauner , Agata Gruza , Antonio Gomez Iglesias , graf@amazon.com, konrad.wilk@oracle.com, dfaggioli@suse.com, pjt@google.com, rostedt@goodmis.org, derkling@google.com, benbjiang@tencent.com, Alexandre Chartre , James.Bottomley@hansenpartnership.com, OWeisse@umich.edu, Dhaval Giani , Junaid Shahid , jsbarnes@google.com, chris.hyser@oracle.com, Ben Segall , Josh Don , Hao Luo , Tom Lendacky , Aubrey Li , "Paul E. McKenney" , Tim Chen References: <20201117232003.3580179-1-joel@joelfernandes.org> <20201117232003.3580179-15-joel@joelfernandes.org> <20201122235456.GF110669@balbir-desktop> <0b2514ef-6cc3-c1a3-280b-5d9062c80a31@linux.intel.com> <20201124154237.GZ3021@hirez.programming.kicks-ass.net> <20201125225731.GB163610@balbir-desktop> <20201126083250.GI163610@balbir-desktop> From: "Li, Aubrey" Message-ID: Date: Thu, 26 Nov 2020 17:26:31 +0800 User-Agent: Mozilla/5.0 (Windows NT 10.0; WOW64; rv:68.0) Gecko/20100101 Thunderbird/68.9.0 MIME-Version: 1.0 In-Reply-To: <20201126083250.GI163610@balbir-desktop> Content-Type: text/plain; charset=utf-8 Content-Language: en-US Content-Transfer-Encoding: 7bit Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 2020/11/26 16:32, Balbir Singh wrote: > On Thu, Nov 26, 2020 at 11:20:41AM +0800, Li, Aubrey wrote: >> On 2020/11/26 6:57, Balbir Singh wrote: >>> On Wed, Nov 25, 2020 at 11:12:53AM +0800, Li, Aubrey wrote: >>>> On 2020/11/24 23:42, Peter Zijlstra wrote: >>>>> On Mon, Nov 23, 2020 at 12:36:10PM +0800, Li, Aubrey wrote: >>>>>>>> +#ifdef CONFIG_SCHED_CORE >>>>>>>> + /* >>>>>>>> + * Skip this cpu if source task's cookie does not match >>>>>>>> + * with CPU's core cookie. >>>>>>>> + */ >>>>>>>> + if (!sched_core_cookie_match(cpu_rq(cpu), env->p)) >>>>>>>> + continue; >>>>>>>> +#endif >>>>>>>> + >>>>>>> >>>>>>> Any reason this is under an #ifdef? In sched_core_cookie_match() won't >>>>>>> the check for sched_core_enabled() do the right thing even when >>>>>>> CONFIG_SCHED_CORE is not enabed?> >>>>>> Yes, sched_core_enabled works properly when CONFIG_SCHED_CORE is not >>>>>> enabled. But when CONFIG_SCHED_CORE is not enabled, it does not make >>>>>> sense to leave a core scheduler specific function here even at compile >>>>>> time. Also, for the cases in hot path, this saves CPU cycles to avoid >>>>>> a judgment. >>>>> >>>>> No, that's nonsense. If it works, remove the #ifdef. Less (#ifdef) is >>>>> more. >>>>> >>>> >>>> Okay, I pasted the refined patch here. >>>> @Joel, please let me know if you want me to send it in a separated thread. >>>> >>> >>> You still have a bunch of #ifdefs, can't we just do >>> >>> #ifndef CONFIG_SCHED_CORE >>> static inline bool sched_core_enabled(struct rq *rq) >>> { >>> return false; >>> } >>> #endif >>> >>> and frankly I think even that is not needed because there is a jump >>> label __sched_core_enabled that tells us if sched_core is enabled or >>> not. >> >> Hmm..., I need another wrapper for CONFIG_SCHED_CORE specific variables. >> How about this one? >> > > Much better :) > >> Thanks, >> -Aubrey >> >> From 61dac9067e66b5b9ea26c684c8c8235714bab38a Mon Sep 17 00:00:00 2001 >> From: Aubrey Li >> Date: Thu, 26 Nov 2020 03:08:04 +0000 >> Subject: [PATCH] sched: migration changes for core scheduling >> >> - Don't migrate if there is a cookie mismatch >> Load balance tries to move task from busiest CPU to the >> destination CPU. When core scheduling is enabled, if the >> task's cookie does not match with the destination CPU's >> core cookie, this task will be skipped by this CPU. This >> mitigates the forced idle time on the destination CPU. >> >> - Select cookie matched idle CPU >> In the fast path of task wakeup, select the first cookie matched >> idle CPU instead of the first idle CPU. >> >> - Find cookie matched idlest CPU >> In the slow path of task wakeup, find the idlest CPU whose core >> cookie matches with task's cookie >> >> - Don't migrate task if cookie not match >> For the NUMA load balance, don't migrate task to the CPU whose >> core cookie does not match with task's cookie >> >> Tested-by: Julien Desfossez >> Signed-off-by: Aubrey Li >> Signed-off-by: Tim Chen >> Signed-off-by: Vineeth Remanan Pillai >> Signed-off-by: Joel Fernandes (Google) >> --- >> kernel/sched/fair.c | 57 ++++++++++++++++++++++++++++++++++++++++---- >> kernel/sched/sched.h | 43 +++++++++++++++++++++++++++++++++ >> 2 files changed, 95 insertions(+), 5 deletions(-) >> >> diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c >> index de82f88ba98c..70dd013dff1d 100644 >> --- a/kernel/sched/fair.c >> +++ b/kernel/sched/fair.c >> @@ -1921,6 +1921,13 @@ static void task_numa_find_cpu(struct task_numa_env *env, >> if (!cpumask_test_cpu(cpu, env->p->cpus_ptr)) >> continue; >> >> + /* >> + * Skip this cpu if source task's cookie does not match >> + * with CPU's core cookie. >> + */ >> + if (!sched_core_cookie_match(cpu_rq(cpu), env->p)) >> + continue; >> + >> env->dst_cpu = cpu; >> if (task_numa_compare(env, taskimp, groupimp, maymove)) >> break; >> @@ -5867,11 +5874,15 @@ find_idlest_group_cpu(struct sched_group *group, struct task_struct *p, int this >> >> /* Traverse only the allowed CPUs */ >> for_each_cpu_and(i, sched_group_span(group), p->cpus_ptr) { >> + struct rq *rq = cpu_rq(i); >> + >> + if (!sched_core_cookie_match(rq, p)) >> + continue; >> + >> if (sched_idle_cpu(i)) >> return i; >> >> if (available_idle_cpu(i)) { >> - struct rq *rq = cpu_rq(i); >> struct cpuidle_state *idle = idle_get_state(rq); >> if (idle && idle->exit_latency < min_exit_latency) { >> /* >> @@ -6129,8 +6140,19 @@ static int select_idle_cpu(struct task_struct *p, struct sched_domain *sd, int t >> for_each_cpu_wrap(cpu, cpus, target) { >> if (!--nr) >> return -1; >> - if (available_idle_cpu(cpu) || sched_idle_cpu(cpu)) >> - break; >> + >> + if (available_idle_cpu(cpu) || sched_idle_cpu(cpu)) { >> + /* >> + * If Core Scheduling is enabled, select this cpu >> + * only if the process cookie matches core cookie. >> + */ >> + if (sched_core_enabled(cpu_rq(cpu))) { >> + if (__cookie_match(cpu_rq(cpu), p)) >> + break; >> + } else { >> + break; >> + } >> + } > > Isn't this better and equivalent? > > if ((available_idle_cpu(cpu) || sched_idle_cpu(cpu)) && > sched_core_cookie_match(cpu_rq(cpu), p)) > break; > That's my previous implementation in the earlier version. But since here is the hot code path, we want to remove the idle core check in sched_core_cookie_match. >> } >> >> time = cpu_clock(this) - time; >> @@ -7530,8 +7552,9 @@ int can_migrate_task(struct task_struct *p, struct lb_env *env) >> * We do not migrate tasks that are: >> * 1) throttled_lb_pair, or >> * 2) cannot be migrated to this CPU due to cpus_ptr, or >> - * 3) running (obviously), or >> - * 4) are cache-hot on their current CPU. >> + * 3) task's cookie does not match with this CPU's core cookie >> + * 4) running (obviously), or >> + * 5) are cache-hot on their current CPU. >> */ >> if (throttled_lb_pair(task_group(p), env->src_cpu, env->dst_cpu)) >> return 0; >> @@ -7566,6 +7589,13 @@ int can_migrate_task(struct task_struct *p, struct lb_env *env) >> return 0; >> } >> >> + /* >> + * Don't migrate task if the task's cookie does not match >> + * with the destination CPU's core cookie. >> + */ >> + if (!sched_core_cookie_match(cpu_rq(env->dst_cpu), p)) >> + return 0; >> + >> /* Record that we found atleast one task that could run on dst_cpu */ >> env->flags &= ~LBF_ALL_PINNED; >> >> @@ -8792,6 +8822,23 @@ find_idlest_group(struct sched_domain *sd, struct task_struct *p, int this_cpu) >> p->cpus_ptr)) >> continue; >> >> + if (sched_core_enabled(cpu_rq(this_cpu))) { >> + int i = 0; >> + bool cookie_match = false; >> + >> + for_each_cpu(i, sched_group_span(group)) { >> + struct rq *rq = cpu_rq(i); >> + >> + if (sched_core_cookie_match(rq, p)) { >> + cookie_match = true; >> + break; >> + } >> + } >> + /* Skip over this group if no cookie matched */ >> + if (!cookie_match) >> + continue; >> + } >> + > > Again, I think this can be refactored because sched_core_cookie_match checks > for sched_core_enabled() > > int i = 0; > bool cookie_match = false; > for_each_cpu(i, sched_group_span(group)) { > if (sched_core_cookie_match(cpu_rq(i), p)) > break; > } > if (i >= nr_cpu_ids) > continue; There is a loop here when CONFIG_SCHED_CORE=n, which is unwanted I guess. Thanks, -Aubrey