From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-12.8 required=3.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,FREEMAIL_FORGED_FROMDOMAIN,FREEMAIL_FROM, HEADER_FROM_DIFFERENT_DOMAINS,INCLUDES_CR_TRAILER,INCLUDES_PATCH, MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS,URIBL_BLOCKED autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 8DBCDC64E8A for ; Mon, 30 Nov 2020 09:34:26 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id 2CD9A20809 for ; Mon, 30 Nov 2020 09:34:26 +0000 (UTC) Authentication-Results: mail.kernel.org; dkim=pass (2048-bit key) header.d=gmail.com header.i=@gmail.com header.b="Y+2W2pNh" Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1728436AbgK3JeZ (ORCPT ); Mon, 30 Nov 2020 04:34:25 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:57898 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1728117AbgK3JeY (ORCPT ); Mon, 30 Nov 2020 04:34:24 -0500 Received: from mail-pl1-x642.google.com (mail-pl1-x642.google.com [IPv6:2607:f8b0:4864:20::642]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id A1CEEC0613CF for ; Mon, 30 Nov 2020 01:33:38 -0800 (PST) Received: by mail-pl1-x642.google.com with SMTP id v3so1567541plz.13 for ; Mon, 30 Nov 2020 01:33:38 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20161025; h=date:from:to:cc:subject:message-id:references:mime-version :content-disposition:in-reply-to; bh=cbE8Ll3lKAvsCg8dcYoAmRikIYrae4TKYhJEbYG/CxM=; b=Y+2W2pNh/ghungmxtJ8O85k6PrudVUlmno97OoPspBTeX5Fx4BUUMAM4Q9JPjK0ToD SGnb5VrfL0p20tmqLf7hNWVRL7ri+0Dz+W6ua+aBBgXA7G9aAL2HxqOsWFJrOE9Id6Vb kf95PwUzzwoPGC7l+NpXsz1YzL1ED9+montbo4PrDHFX9Qr3P3d6MrSEnKrsISvF6wyk oRg94kstBo8IfCWu4Pq919c9DnOCfYcQDHqCd0VINlv3zYjmskTaKSHeHtCK/SlyQzyj VgCq60UQxVloSWdVWtf1axC7v1P41aRlRni0P/Lik9llOH1AIB1DTvB2/WkfiFl0iHhu pvnw== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:date:from:to:cc:subject:message-id:references :mime-version:content-disposition:in-reply-to; bh=cbE8Ll3lKAvsCg8dcYoAmRikIYrae4TKYhJEbYG/CxM=; b=bBWH/0SQFQ8rJ+DDUmq+WDgaDsiVZ9+N8fMYNpV9mYlZOupNpQjOiUj8uBlzznBeAH aHxnMrTg2q92ETSxzGt+oV8iv6xXmEX3W9EF6xsytayqi4oCeKYRJNrj+e2MXZyLOE51 NH092UzPAD/YwAAaMzoi5GKXhVnD91fgGAklQsZxBQzWm0XdtXGVWpwKWQBd4/nvRFJu VMDxvG/yKW6G3f07qwcPiWq2axERm773+5TME0qSwq1pGNXyMNq95DnPPx45MmgNP+No ggwPgAu8OtyfbfatkORZfNCYDKPNoO2n4fVZTq8e+Ki9fDXHD3hy0z3rdyr+7GN3I/fe lwDg== X-Gm-Message-State: AOAM5314g2Fsn9m2g7LJXKvpgSeLKTFN3R9zU87qUa7f4RDC+5TCcdsY 3Q1tHWHlT+ekCUuMX4KR134= X-Google-Smtp-Source: ABdhPJyb4w7cZyyY75rZ01AdvY7fpLiwhmJZPpBcCe1h6SCowd8C3MjbjMx9VhOgOboAAO401GhDsQ== X-Received: by 2002:a17:90a:4482:: with SMTP id t2mr24703168pjg.44.1606728817860; Mon, 30 Nov 2020 01:33:37 -0800 (PST) Received: from localhost (61-68-227-232.tpgi.com.au. [61.68.227.232]) by smtp.gmail.com with ESMTPSA id r36sm15433246pgb.75.2020.11.30.01.33.35 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Mon, 30 Nov 2020 01:33:36 -0800 (PST) Date: Mon, 30 Nov 2020 20:33:33 +1100 From: Balbir Singh To: "Li, Aubrey" Cc: Peter Zijlstra , "Joel Fernandes (Google)" , Nishanth Aravamudan , Julien Desfossez , Tim Chen , Vineeth Pillai , Aaron Lu , Aubrey Li , tglx@linutronix.de, linux-kernel@vger.kernel.org, mingo@kernel.org, torvalds@linux-foundation.org, fweisbec@gmail.com, keescook@chromium.org, kerrnel@google.com, Phil Auld , Valentin Schneider , Mel Gorman , Pawan Gupta , Paolo Bonzini , vineeth@bitbyteword.org, Chen Yu , Christian Brauner , Agata Gruza , Antonio Gomez Iglesias , graf@amazon.com, konrad.wilk@oracle.com, dfaggioli@suse.com, pjt@google.com, rostedt@goodmis.org, derkling@google.com, benbjiang@tencent.com, Alexandre Chartre , James.Bottomley@hansenpartnership.com, OWeisse@umich.edu, Dhaval Giani , Junaid Shahid , jsbarnes@google.com, chris.hyser@oracle.com, Ben Segall , Josh Don , Hao Luo , Tom Lendacky , Aubrey Li , "Paul E. McKenney" , Tim Chen Subject: Re: [PATCH -tip 14/32] sched: migration changes for core scheduling Message-ID: <20201130093333.GD473773@balbir-desktop> References: <20201117232003.3580179-1-joel@joelfernandes.org> <20201117232003.3580179-15-joel@joelfernandes.org> <20201122235456.GF110669@balbir-desktop> <0b2514ef-6cc3-c1a3-280b-5d9062c80a31@linux.intel.com> <20201124154237.GZ3021@hirez.programming.kicks-ass.net> <20201125225731.GB163610@balbir-desktop> <20201126083250.GI163610@balbir-desktop> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Thu, Nov 26, 2020 at 05:26:31PM +0800, Li, Aubrey wrote: > On 2020/11/26 16:32, Balbir Singh wrote: > > On Thu, Nov 26, 2020 at 11:20:41AM +0800, Li, Aubrey wrote: > >> On 2020/11/26 6:57, Balbir Singh wrote: > >>> On Wed, Nov 25, 2020 at 11:12:53AM +0800, Li, Aubrey wrote: > >>>> On 2020/11/24 23:42, Peter Zijlstra wrote: > >>>>> On Mon, Nov 23, 2020 at 12:36:10PM +0800, Li, Aubrey wrote: > >>>>>>>> +#ifdef CONFIG_SCHED_CORE > >>>>>>>> + /* > >>>>>>>> + * Skip this cpu if source task's cookie does not match > >>>>>>>> + * with CPU's core cookie. > >>>>>>>> + */ > >>>>>>>> + if (!sched_core_cookie_match(cpu_rq(cpu), env->p)) > >>>>>>>> + continue; > >>>>>>>> +#endif > >>>>>>>> + > >>>>>>> > >>>>>>> Any reason this is under an #ifdef? In sched_core_cookie_match() won't > >>>>>>> the check for sched_core_enabled() do the right thing even when > >>>>>>> CONFIG_SCHED_CORE is not enabed?> > >>>>>> Yes, sched_core_enabled works properly when CONFIG_SCHED_CORE is not > >>>>>> enabled. But when CONFIG_SCHED_CORE is not enabled, it does not make > >>>>>> sense to leave a core scheduler specific function here even at compile > >>>>>> time. Also, for the cases in hot path, this saves CPU cycles to avoid > >>>>>> a judgment. > >>>>> > >>>>> No, that's nonsense. If it works, remove the #ifdef. Less (#ifdef) is > >>>>> more. > >>>>> > >>>> > >>>> Okay, I pasted the refined patch here. > >>>> @Joel, please let me know if you want me to send it in a separated thread. > >>>> > >>> > >>> You still have a bunch of #ifdefs, can't we just do > >>> > >>> #ifndef CONFIG_SCHED_CORE > >>> static inline bool sched_core_enabled(struct rq *rq) > >>> { > >>> return false; > >>> } > >>> #endif > >>> > >>> and frankly I think even that is not needed because there is a jump > >>> label __sched_core_enabled that tells us if sched_core is enabled or > >>> not. > >> > >> Hmm..., I need another wrapper for CONFIG_SCHED_CORE specific variables. > >> How about this one? > >> > > > > Much better :) > > > >> Thanks, > >> -Aubrey > >> > >> From 61dac9067e66b5b9ea26c684c8c8235714bab38a Mon Sep 17 00:00:00 2001 > >> From: Aubrey Li > >> Date: Thu, 26 Nov 2020 03:08:04 +0000 > >> Subject: [PATCH] sched: migration changes for core scheduling > >> > >> - Don't migrate if there is a cookie mismatch > >> Load balance tries to move task from busiest CPU to the > >> destination CPU. When core scheduling is enabled, if the > >> task's cookie does not match with the destination CPU's > >> core cookie, this task will be skipped by this CPU. This > >> mitigates the forced idle time on the destination CPU. > >> > >> - Select cookie matched idle CPU > >> In the fast path of task wakeup, select the first cookie matched > >> idle CPU instead of the first idle CPU. > >> > >> - Find cookie matched idlest CPU > >> In the slow path of task wakeup, find the idlest CPU whose core > >> cookie matches with task's cookie > >> > >> - Don't migrate task if cookie not match > >> For the NUMA load balance, don't migrate task to the CPU whose > >> core cookie does not match with task's cookie > >> > >> Tested-by: Julien Desfossez > >> Signed-off-by: Aubrey Li > >> Signed-off-by: Tim Chen > >> Signed-off-by: Vineeth Remanan Pillai > >> Signed-off-by: Joel Fernandes (Google) > >> --- > >> kernel/sched/fair.c | 57 ++++++++++++++++++++++++++++++++++++++++---- > >> kernel/sched/sched.h | 43 +++++++++++++++++++++++++++++++++ > >> 2 files changed, 95 insertions(+), 5 deletions(-) > >> > >> diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c > >> index de82f88ba98c..70dd013dff1d 100644 > >> --- a/kernel/sched/fair.c > >> +++ b/kernel/sched/fair.c > >> @@ -1921,6 +1921,13 @@ static void task_numa_find_cpu(struct task_numa_env *env, > >> if (!cpumask_test_cpu(cpu, env->p->cpus_ptr)) > >> continue; > >> > >> + /* > >> + * Skip this cpu if source task's cookie does not match > >> + * with CPU's core cookie. > >> + */ > >> + if (!sched_core_cookie_match(cpu_rq(cpu), env->p)) > >> + continue; > >> + > >> env->dst_cpu = cpu; > >> if (task_numa_compare(env, taskimp, groupimp, maymove)) > >> break; > >> @@ -5867,11 +5874,15 @@ find_idlest_group_cpu(struct sched_group *group, struct task_struct *p, int this > >> > >> /* Traverse only the allowed CPUs */ > >> for_each_cpu_and(i, sched_group_span(group), p->cpus_ptr) { > >> + struct rq *rq = cpu_rq(i); > >> + > >> + if (!sched_core_cookie_match(rq, p)) > >> + continue; > >> + > >> if (sched_idle_cpu(i)) > >> return i; > >> > >> if (available_idle_cpu(i)) { > >> - struct rq *rq = cpu_rq(i); > >> struct cpuidle_state *idle = idle_get_state(rq); > >> if (idle && idle->exit_latency < min_exit_latency) { > >> /* > >> @@ -6129,8 +6140,19 @@ static int select_idle_cpu(struct task_struct *p, struct sched_domain *sd, int t > >> for_each_cpu_wrap(cpu, cpus, target) { > >> if (!--nr) > >> return -1; > >> - if (available_idle_cpu(cpu) || sched_idle_cpu(cpu)) > >> - break; > >> + > >> + if (available_idle_cpu(cpu) || sched_idle_cpu(cpu)) { > >> + /* > >> + * If Core Scheduling is enabled, select this cpu > >> + * only if the process cookie matches core cookie. > >> + */ > >> + if (sched_core_enabled(cpu_rq(cpu))) { > >> + if (__cookie_match(cpu_rq(cpu), p)) > >> + break; > >> + } else { > >> + break; > >> + } > >> + } > > > > Isn't this better and equivalent? > > > > if ((available_idle_cpu(cpu) || sched_idle_cpu(cpu)) && > > sched_core_cookie_match(cpu_rq(cpu), p)) > > break; > > > > > That's my previous implementation in the earlier version. > But since here is the hot code path, we want to remove the idle > core check in sched_core_cookie_match. I see, so we basically need a jump label, if sched_core_cookie_match can be inlined with a check for sched_core_enabled() upfront, it might solve a lot of the concern, readability of this section of code is not the best. > > >> } > >> > >> time = cpu_clock(this) - time; > >> @@ -7530,8 +7552,9 @@ int can_migrate_task(struct task_struct *p, struct lb_env *env) > >> * We do not migrate tasks that are: > >> * 1) throttled_lb_pair, or > >> * 2) cannot be migrated to this CPU due to cpus_ptr, or > >> - * 3) running (obviously), or > >> - * 4) are cache-hot on their current CPU. > >> + * 3) task's cookie does not match with this CPU's core cookie > >> + * 4) running (obviously), or > >> + * 5) are cache-hot on their current CPU. > >> */ > >> if (throttled_lb_pair(task_group(p), env->src_cpu, env->dst_cpu)) > >> return 0; > >> @@ -7566,6 +7589,13 @@ int can_migrate_task(struct task_struct *p, struct lb_env *env) > >> return 0; > >> } > >> > >> + /* > >> + * Don't migrate task if the task's cookie does not match > >> + * with the destination CPU's core cookie. > >> + */ > >> + if (!sched_core_cookie_match(cpu_rq(env->dst_cpu), p)) > >> + return 0; > >> + > >> /* Record that we found atleast one task that could run on dst_cpu */ > >> env->flags &= ~LBF_ALL_PINNED; > >> > >> @@ -8792,6 +8822,23 @@ find_idlest_group(struct sched_domain *sd, struct task_struct *p, int this_cpu) > >> p->cpus_ptr)) > >> continue; > >> > >> + if (sched_core_enabled(cpu_rq(this_cpu))) { > >> + int i = 0; > >> + bool cookie_match = false; > >> + > >> + for_each_cpu(i, sched_group_span(group)) { > >> + struct rq *rq = cpu_rq(i); > >> + > >> + if (sched_core_cookie_match(rq, p)) { > >> + cookie_match = true; > >> + break; > >> + } > >> + } > >> + /* Skip over this group if no cookie matched */ > >> + if (!cookie_match) > >> + continue; > >> + } > >> + > > > > Again, I think this can be refactored because sched_core_cookie_match checks > > for sched_core_enabled() > > > > int i = 0; > > bool cookie_match = false; > > for_each_cpu(i, sched_group_span(group)) { > > if (sched_core_cookie_match(cpu_rq(i), p)) > > break; > > } > > if (i >= nr_cpu_ids) > > continue; > > There is a loop here when CONFIG_SCHED_CORE=n, which is unwanted I guess. > Yes, potentially, may be abstract the for_each_cpu into a function and then optimize out the case for SCHED_CORE=n, I feel all the extra checks in the various places make the code hard to read. Balbir