All of lore.kernel.org
 help / color / mirror / Atom feed
From: Ming Lei <ming.lei@redhat.com>
To: Jens Axboe <axboe@kernel.dk>,
	Christoph Hellwig <hch@infradead.org>,
	Thomas Gleixner <tglx@linutronix.de>,
	linux-kernel@vger.kernel.org
Cc: linux-block@vger.kernel.org,
	Laurence Oberman <loberman@redhat.com>,
	Ming Lei <ming.lei@redhat.com>
Subject: [PATCH V3 4/4] genirq/affinity: irq vector spread among online CPUs as far as possible
Date: Thu,  8 Mar 2018 18:53:58 +0800	[thread overview]
Message-ID: <20180308105358.1506-5-ming.lei@redhat.com> (raw)
In-Reply-To: <20180308105358.1506-1-ming.lei@redhat.com>

84676c1f21 ("genirq/affinity: assign vectors to all possible CPUs")
may cause irq vector assigned to all offline CPUs, and this kind of
assignment may cause much less irq vectors mapped to online CPUs, and
performance may get hurt.

For example, in a 8 cores system, 0~3 online, 4~8 offline/not present,
see 'lscpu':

	[ming@box]$lscpu
	Architecture:          x86_64
	CPU op-mode(s):        32-bit, 64-bit
	Byte Order:            Little Endian
	CPU(s):                4
	On-line CPU(s) list:   0-3
	Thread(s) per core:    1
	Core(s) per socket:    2
	Socket(s):             2
	NUMA node(s):          2
	...
	NUMA node0 CPU(s):     0-3
	NUMA node1 CPU(s):
	...

For example, one device has 4 queues:

1) before 84676c1f21 ("genirq/affinity: assign vectors to all possible CPUs")
	irq 39, cpu list 0
	irq 40, cpu list 1
	irq 41, cpu list 2
	irq 42, cpu list 3

2) after 84676c1f21 ("genirq/affinity: assign vectors to all possible CPUs")
	irq 39, cpu list 0-2
	irq 40, cpu list 3-4,6
	irq 41, cpu list 5
	irq 42, cpu list 7

3) after applying this patch against V4.15+:
	irq 39, cpu list 0,4
	irq 40, cpu list 1,6
	irq 41, cpu list 2,5
	irq 42, cpu list 3,7

This patch tries to do irq vector spread among online CPUs as far as
possible by 2 stages spread.

The above assignment 3) isn't the optimal result from NUMA view, but it
returns more irq vectors with online CPU mapped, given in reality one CPU
should be enough to handle one irq vector, so it is better to do this way.

Cc: Thomas Gleixner <tglx@linutronix.de>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Reported-by: Laurence Oberman <loberman@redhat.com>
Signed-off-by: Ming Lei <ming.lei@redhat.com>
---
 kernel/irq/affinity.c | 35 +++++++++++++++++++++++++++++------
 1 file changed, 29 insertions(+), 6 deletions(-)

diff --git a/kernel/irq/affinity.c b/kernel/irq/affinity.c
index 616f040c5d02..253c5bf85d18 100644
--- a/kernel/irq/affinity.c
+++ b/kernel/irq/affinity.c
@@ -106,6 +106,9 @@ static int irq_build_affinity_masks(const struct irq_affinity *affd,
 	nodemask_t nodemsk = NODE_MASK_NONE;
 	int n, nodes, cpus_per_vec, extra_vecs, done = 0;
 
+	if (!cpumask_weight(cpu_mask))
+		return 0;
+
 	nodes = get_nodes_in_cpumask(node_to_cpumask, cpu_mask, &nodemsk);
 
 	/*
@@ -175,9 +178,9 @@ struct cpumask *
 irq_create_affinity_masks(int nvecs, const struct irq_affinity *affd)
 {
 	int affv = nvecs - affd->pre_vectors - affd->post_vectors;
-	int curvec;
+	int curvec, vecs_offline, vecs_online;
 	struct cpumask *masks;
-	cpumask_var_t nmsk, *node_to_cpumask;
+	cpumask_var_t nmsk, cpu_mask, *node_to_cpumask;
 
 	/*
 	 * If there aren't any vectors left after applying the pre/post
@@ -193,9 +196,12 @@ irq_create_affinity_masks(int nvecs, const struct irq_affinity *affd)
 	if (!masks)
 		goto out;
 
+	if (!alloc_cpumask_var(&cpu_mask, GFP_KERNEL))
+		goto out;
+
 	node_to_cpumask = alloc_node_to_cpumask();
 	if (!node_to_cpumask)
-		goto out;
+		goto out_free_cpu_mask;
 
 	/* Fill out vectors at the beginning that don't need affinity */
 	for (curvec = 0; curvec < affd->pre_vectors; curvec++)
@@ -204,15 +210,32 @@ irq_create_affinity_masks(int nvecs, const struct irq_affinity *affd)
 	/* Stabilize the cpumasks */
 	get_online_cpus();
 	build_node_to_cpumask(node_to_cpumask);
-	curvec += irq_build_affinity_masks(affd, curvec, affv,
-					   node_to_cpumask,
-					   cpu_possible_mask, nmsk, masks);
+	/* spread on online CPUs starting from the vector of affd->pre_vectors */
+	vecs_online = irq_build_affinity_masks(affd, curvec, affv,
+					       node_to_cpumask,
+					       cpu_online_mask, nmsk, masks);
+
+	/* spread on offline CPUs starting from the next vector to be handled */
+	if (vecs_online >= affv)
+		curvec = affd->pre_vectors;
+	else
+		curvec = affd->pre_vectors + vecs_online;
+	cpumask_andnot(cpu_mask, cpu_possible_mask, cpu_online_mask);
+	vecs_offline = irq_build_affinity_masks(affd, curvec, affv,
+						node_to_cpumask,
+					        cpu_mask, nmsk, masks);
 	put_online_cpus();
 
 	/* Fill out vectors at the end that don't need affinity */
+	if (vecs_online + vecs_offline >= affv)
+		curvec = affv + affd->pre_vectors;
+	else
+		curvec = affd->pre_vectors + vecs_online + vecs_offline;
 	for (; curvec < nvecs; curvec++)
 		cpumask_copy(masks + curvec, irq_default_affinity);
 	free_node_to_cpumask(node_to_cpumask);
+out_free_cpu_mask:
+	free_cpumask_var(cpu_mask);
 out:
 	free_cpumask_var(nmsk);
 	return masks;
-- 
2.9.5

  parent reply	other threads:[~2018-03-08 10:53 UTC|newest]

Thread overview: 48+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2018-03-08 10:53 [PATCH V3 0/4] genirq/affinity: irq vector spread among online CPUs as far as possible Ming Lei
2018-03-08 10:53 ` [PATCH V3 1/4] genirq/affinity: rename *node_to_possible_cpumask as *node_to_cpumask Ming Lei
2018-04-06 21:52   ` [tip:irq/core] genirq/affinity: Rename " tip-bot for Ming Lei
2018-03-08 10:53 ` [PATCH V3 2/4] genirq/affinity: move actual irq vector spread into one helper Ming Lei
2018-04-06 21:53   ` [tip:irq/core] genirq/affinity: Move actual irq vector spreading into a helper function tip-bot for Ming Lei
2018-03-08 10:53 ` [PATCH V3 3/4] genirq/affinity: support to do irq vectors spread starting from any vector Ming Lei
2018-04-06 21:53   ` [tip:irq/core] genirq/affinity: Allow irq spreading from a given starting point tip-bot for Ming Lei
2018-03-08 10:53 ` Ming Lei [this message]
2018-04-03 13:32   ` [PATCH V3 4/4] genirq/affinity: irq vector spread among online CPUs as far as possible Thomas Gleixner
2018-04-03 16:00     ` Ming Lei
2018-04-04  8:25       ` Thomas Gleixner
2018-04-04 12:45         ` Thomas Gleixner
2018-04-04 15:20           ` Ming Lei
2018-04-05 10:12             ` Thomas Gleixner
2018-04-04 15:08         ` Ming Lei
2018-04-04 19:38           ` Thomas Gleixner
2018-04-06  9:13             ` Ming Lei
2018-04-06  9:46               ` Thomas Gleixner
2018-04-06 21:49                 ` Thomas Gleixner
2018-04-08  3:19                   ` Ming Lei
2018-04-06 21:54   ` [tip:irq/core] genirq/affinity: Spread irq vectors among present " tip-bot for Ming Lei
2018-03-08 13:18 ` [PATCH V3 0/4] genirq/affinity: irq vector spread among online " Artem Bityutskiy
2018-03-08 13:25   ` Artem Bityutskiy
2018-03-08 13:34   ` Ming Lei
2018-03-08 23:20     ` Thomas Gleixner
2018-03-09  1:24       ` Ming Lei
2018-03-09  7:00         ` Artem Bityutskiy
2018-03-09  7:33           ` Ming Lei
2018-03-09 10:08         ` Thomas Gleixner
2018-03-09 12:08           ` Ming Lei
2018-03-09 15:08             ` Thomas Gleixner
2018-03-13  3:11               ` Dou Liyang
2018-03-13  7:38                 ` Artem Bityutskiy
2018-03-13  7:38                   ` Artem Bityutskiy
2018-03-13  8:35                   ` Ming Lei
2018-03-13  8:39                     ` Artem Bityutskiy
2018-03-13  8:39                       ` Artem Bityutskiy
2018-03-13  9:35                       ` Rafael J. Wysocki
2018-03-14  3:29                         ` Dou Liyang
2018-03-14  4:11                           ` Dou Liyang
2018-03-14  9:07                             ` Artem Bityutskiy
2018-03-14  9:47                               ` Dou Liyang
2018-03-13  9:25                 ` Rafael J. Wysocki
2018-03-14  3:30                   ` Dou Liyang
2018-03-30  3:15               ` Ming Lei
2018-04-03 12:55                 ` Thomas Gleixner
2018-03-26  8:39   ` Thorsten Leemhuis
2018-03-28  6:15     ` Artem Bityutskiy

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20180308105358.1506-5-ming.lei@redhat.com \
    --to=ming.lei@redhat.com \
    --cc=axboe@kernel.dk \
    --cc=hch@infradead.org \
    --cc=linux-block@vger.kernel.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=loberman@redhat.com \
    --cc=tglx@linutronix.de \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.