From: Valentin Schneider <valentin.schneider@arm.com>
To: Barry Song <song.bao.hua@hisilicon.com>
Cc: catalin.marinas@arm.com, will@kernel.org, rjw@rjwysocki.net,
lenb@kernel.org, gregkh@linuxfoundation.org,
Jonathan.Cameron@huawei.com, mingo@redhat.com,
peterz@infradead.org, juri.lelli@redhat.com,
vincent.guittot@linaro.org, dietmar.eggemann@arm.com,
rostedt@goodmis.org, bsegall@google.com, mgorman@suse.de,
mark.rutland@arm.com, linux-arm-kernel@lists.infradead.org,
linux-kernel@vger.kernel.org, linux-acpi@vger.kernel.org,
linuxarm@huawei.com, xuwei5@huawei.com, prime.zeng@hisilicon.com
Subject: Re: [RFC PATCH v2 1/2] topology: Represent clusters of CPUs within a die.
Date: Tue, 01 Dec 2020 16:03:46 +0000 [thread overview]
Message-ID: <jhj360pv7h9.mognet@arm.com> (raw)
In-Reply-To: <20201201025944.18260-2-song.bao.hua@hisilicon.com>
On 01/12/20 02:59, Barry Song wrote:
> That means the cost to transfer ownership of a cacheline between CPUs
> within a cluster is lower than between CPUs in different clusters on
> the same die. Hence, it can make sense to tell the scheduler to use
> the cache affinity of the cluster to make better decision on thread
> migration.
>
> This patch simply exposes this information to userspace libraries
> like hwloc by providing cluster_cpus and related sysfs attributes.
> PoC of HWLOC support at [2].
>
> Note this patch only handle the ACPI case.
>
AIUI this requires PPTT to describe your system like so:
{Processor nodes} {Caches}
[Node0] ----------------> [L3]
^
|
[Cluster0] ---------------> []
^
|
[CPU0] ------------> [L1] -> [L2]
which is a bit odd, because there is that middling level without any
private resources. I suppose right now this is the only way to describe
this kind of cache topology via PPTT, but is that widespread?
Now, looking at the Ampere eMAG's PPTT, this has a "similar" shape. The
topology is private L1, L2 shared by pairs of CPUs, shared L3 [1].
If I parse the PPTT thing right this is encoded as:
{Processor nodes} {Caches}
[Cluster0] -------------> ([L3] not present in my PPTT for some reason)
^
|
[ Pair0 ] ------------> [L2]
^ ^
| |
| [CPU1] ------------> [L1]
[CPU0] -----------------> [L1]
So you could spin the same story there were first scanning the pair and
then the cluster could help.
[1]: https://en.wikichip.org/wiki/ampere_computing/emag/8180
> Special consideration is needed for SMT processors, where it is
> necessary to move 2 levels up the hierarchy from the leaf nodes
> (thus skipping the processor core level).
>
> Currently the ID provided is the offset of the Processor
> Hierarchy Nodes Structure within PPTT. Whilst this is unique
> it is not terribly elegant so alternative suggestions welcome.
>
Skimming through the spec, this sounds like something the ID structure
(Type 2) could be used for. However in v1 Jonathan and Sudeep talked about
UID's / DSDT, any news on that?
> Note that arm64 / ACPI does not provide any means of identifying
> a die level in the topology but that may be unrelate to the cluster
> level.
>
> [1] ACPI Specification 6.3 - section 5.2.29.1 processor hierarchy node
> structure (Type 0)
> [2] https://github.com/hisilicon/hwloc/tree/linux-cluster
>
> Signed-off-by: Jonathan Cameron <Jonathan.Cameron@huawei.com>
> Signed-off-by: Barry Song <song.bao.hua@hisilicon.com>
next prev parent reply other threads:[~2020-12-01 16:04 UTC|newest]
Thread overview: 24+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-12-01 2:59 [RFC PATCH v2 0/2] scheduler: expose the topology of clusters and add cluster scheduler Barry Song
2020-12-01 2:59 ` [RFC PATCH v2 1/2] topology: Represent clusters of CPUs within a die Barry Song
2020-12-01 16:03 ` Valentin Schneider [this message]
2020-12-02 9:55 ` Sudeep Holla
2020-12-01 2:59 ` [RFC PATCH v2 2/2] scheduler: add scheduler level for clusters Barry Song
2020-12-01 16:04 ` Valentin Schneider
2020-12-03 9:28 ` Peter Zijlstra
2020-12-03 9:49 ` Mel Gorman
2020-12-03 9:57 ` Song Bao Hua (Barry Song)
2020-12-03 10:07 ` Peter Zijlstra
2020-12-02 8:27 ` Vincent Guittot
2020-12-02 9:20 ` Song Bao Hua (Barry Song)
2020-12-02 10:16 ` Vincent Guittot
2020-12-02 10:45 ` Song Bao Hua (Barry Song)
2020-12-02 10:48 ` Song Bao Hua (Barry Song)
2020-12-02 20:58 ` Song Bao Hua (Barry Song)
2020-12-03 9:03 ` Vincent Guittot
2020-12-03 9:11 ` Song Bao Hua (Barry Song)
2020-12-03 9:39 ` Vincent Guittot
2020-12-03 9:54 ` Vincent Guittot
2020-12-07 9:59 ` Song Bao Hua (Barry Song)
2020-12-07 15:29 ` Vincent Guittot
2020-12-09 11:35 ` Song Bao Hua (Barry Song)
2020-12-01 10:46 ` [RFC PATCH v2 0/2] scheduler: expose the topology of clusters and add cluster scheduler Dietmar Eggemann
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=jhj360pv7h9.mognet@arm.com \
--to=valentin.schneider@arm.com \
--cc=Jonathan.Cameron@huawei.com \
--cc=bsegall@google.com \
--cc=catalin.marinas@arm.com \
--cc=dietmar.eggemann@arm.com \
--cc=gregkh@linuxfoundation.org \
--cc=juri.lelli@redhat.com \
--cc=lenb@kernel.org \
--cc=linux-acpi@vger.kernel.org \
--cc=linux-arm-kernel@lists.infradead.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linuxarm@huawei.com \
--cc=mark.rutland@arm.com \
--cc=mgorman@suse.de \
--cc=mingo@redhat.com \
--cc=peterz@infradead.org \
--cc=prime.zeng@hisilicon.com \
--cc=rjw@rjwysocki.net \
--cc=rostedt@goodmis.org \
--cc=song.bao.hua@hisilicon.com \
--cc=vincent.guittot@linaro.org \
--cc=will@kernel.org \
--cc=xuwei5@huawei.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).