linux-kernel.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: Cheng Jian <cj.chengjian@huawei.com>
To: <mingo@kernel.org>, <peterz@infradead.org>,
	<linux-kernel@vger.kernel.org>
Cc: <cj.chengjian@huawei.com>, <chenwandun@huawei.com>,
	<xiexiuqi@huawei.com>, <liwei391@huawei.com>,
	<huawei.libin@huawei.com>, <bobo.shaobowang@huawei.com>,
	<juri.lelli@redhat.com>, <vincent.guittot@linaro.org>
Subject: [PATCH] sched/fair: Optimize select_idle_cpu
Date: Thu, 12 Dec 2019 22:41:02 +0800	[thread overview]
Message-ID: <20191212144102.181510-1-cj.chengjian@huawei.com> (raw)

select_idle_cpu will scan the LLC domain for idle CPUs,
it's always expensive. so commit
    1ad3aaf3fcd2 ("sched/core: Implement new approach to scale select_idle_cpu()")
introduces a way to limit how many CPUs we scan.

But this also lead to the following issue:

Our threads are all bind to the front CPUs of the LLC domain,
and now all the threads runs on the last CPU of them. nr is
always less than the cpumask_weight, for_each_cpu_wrap can't
find the CPU which our threads can run on, so the threads stay
at the last CPU all the time.

Fixes: 1ad3aaf3fcd2 ("sched/core: Implement new approach to scale select_idle_cpu()")
Signed-off-by: Cheng Jian <cj.chengjian@huawei.com>
---
 kernel/sched/fair.c | 7 ++++---
 1 file changed, 4 insertions(+), 3 deletions(-)

diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c
index 08a233e97a01..16a29b570803 100644
--- a/kernel/sched/fair.c
+++ b/kernel/sched/fair.c
@@ -5834,6 +5834,7 @@ static int select_idle_cpu(struct task_struct *p, struct sched_domain *sd, int t
 	s64 delta;
 	int this = smp_processor_id();
 	int cpu, nr = INT_MAX, si_cpu = -1;
+	struct cpumask cpus;
 
 	this_sd = rcu_dereference(*this_cpu_ptr(&sd_llc));
 	if (!this_sd)
@@ -5859,11 +5860,11 @@ static int select_idle_cpu(struct task_struct *p, struct sched_domain *sd, int t
 
 	time = cpu_clock(this);
 
-	for_each_cpu_wrap(cpu, sched_domain_span(sd), target) {
+	cpumask_and(&cpus, sched_domain_span(sd), p->cpus_ptr);
+
+	for_each_cpu_wrap(cpu, &cpus, target) {
 		if (!--nr)
 			return si_cpu;
-		if (!cpumask_test_cpu(cpu, p->cpus_ptr))
-			continue;
 		if (available_idle_cpu(cpu))
 			break;
 		if (si_cpu == -1 && sched_idle_cpu(cpu))
-- 
2.20.1


             reply	other threads:[~2019-12-12 14:43 UTC|newest]

Thread overview: 12+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2019-12-12 14:41 Cheng Jian [this message]
2019-12-12 14:56 ` [PATCH] sched/fair: Optimize select_idle_cpu chengjian (D)
2019-12-12 15:04 ` Peter Zijlstra
2019-12-13  1:51   ` chengjian (D)
2019-12-13  8:10     ` Peter Zijlstra
2019-12-12 15:24 ` Peter Zijlstra
2019-12-12 15:44   ` Valentin Schneider
2019-12-13  9:47     ` chengjian (D)
2019-12-13  9:57   ` chengjian (D)
2019-12-13 11:28     ` Valentin Schneider
2019-12-13 12:09       ` Peter Zijlstra
2019-12-13 12:20         ` Valentin Schneider

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20191212144102.181510-1-cj.chengjian@huawei.com \
    --to=cj.chengjian@huawei.com \
    --cc=bobo.shaobowang@huawei.com \
    --cc=chenwandun@huawei.com \
    --cc=huawei.libin@huawei.com \
    --cc=juri.lelli@redhat.com \
    --cc=linux-kernel@vger.kernel.org \
    --cc=liwei391@huawei.com \
    --cc=mingo@kernel.org \
    --cc=peterz@infradead.org \
    --cc=vincent.guittot@linaro.org \
    --cc=xiexiuqi@huawei.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).