qemu-devel.nongnu.org archive mirror
 help / color / mirror / Atom feed
From: Igor Mammedov <imammedo@redhat.com>
To: Babu Moger <babu.moger@amd.com>
Cc: ehabkost@redhat.com, mst@redhat.com, qemu-devel@nongnu.org,
	pbonzini@redhat.com, rth@twiddle.net
Subject: Re: [PATCH v4 09/16] target/i386: Cleanup and use the EPYC mode topology functions
Date: Tue, 25 Feb 2020 08:49:05 +0100	[thread overview]
Message-ID: <20200225084905.11b9731d@redhat.com> (raw)
In-Reply-To: <85bb2603-115a-1df2-df5d-887faae66bbe@amd.com>

On Mon, 24 Feb 2020 11:29:37 -0600
Babu Moger <babu.moger@amd.com> wrote:

> On 2/24/20 2:52 AM, Igor Mammedov wrote:
> > On Thu, 13 Feb 2020 12:17:25 -0600
> > Babu Moger <babu.moger@amd.com> wrote:
> >   
> >> Use the new functions from topology.h and delete the unused code. Given the
> >> sockets, nodes, cores and threads, the new functions generate apic id for EPYC
> >> mode. Removes all the hardcoded values.
> >>
> >> Signed-off-by: Babu Moger <babu.moger@amd.com>  
> > 
> > modulo MAX() macro, looks fine to me  
>
> Igor, Sorry. What do you mean here?

I meant s/MAX(topo_info->nodes_per_pkg, 1)/topo_info->nodes_per_pkg/

after it's made sure that topo_info->nodes_per_pkg is always valid.


(I believe I've commented on that somewhere. Series isn't split nicely,
so I've ended up applying it all and then reviewing so comments might
look out of the place sometimes, hopefully next revision will be easier
to review)

> >   
> >> ---
> >>  target/i386/cpu.c |  162 +++++++++++------------------------------------------
> >>  1 file changed, 35 insertions(+), 127 deletions(-)
> >>
> >> diff --git a/target/i386/cpu.c b/target/i386/cpu.c
> >> index 5d6edfd09b..19675eb696 100644
> >> --- a/target/i386/cpu.c
> >> +++ b/target/i386/cpu.c
> >> @@ -338,68 +338,15 @@ static void encode_cache_cpuid80000006(CPUCacheInfo *l2,
> >>      }
> >>  }
> >>  
> >> -/*
> >> - * Definitions used for building CPUID Leaf 0x8000001D and 0x8000001E
> >> - * Please refer to the AMD64 Architecture Programmer’s Manual Volume 3.
> >> - * Define the constants to build the cpu topology. Right now, TOPOEXT
> >> - * feature is enabled only on EPYC. So, these constants are based on
> >> - * EPYC supported configurations. We may need to handle the cases if
> >> - * these values change in future.
> >> - */
> >> -/* Maximum core complexes in a node */
> >> -#define MAX_CCX 2
> >> -/* Maximum cores in a core complex */
> >> -#define MAX_CORES_IN_CCX 4
> >> -/* Maximum cores in a node */
> >> -#define MAX_CORES_IN_NODE 8
> >> -/* Maximum nodes in a socket */
> >> -#define MAX_NODES_PER_SOCKET 4
> >> -
> >> -/*
> >> - * Figure out the number of nodes required to build this config.
> >> - * Max cores in a node is 8
> >> - */
> >> -static int nodes_in_socket(int nr_cores)
> >> -{
> >> -    int nodes;
> >> -
> >> -    nodes = DIV_ROUND_UP(nr_cores, MAX_CORES_IN_NODE);
> >> -
> >> -   /* Hardware does not support config with 3 nodes, return 4 in that case */
> >> -    return (nodes == 3) ? 4 : nodes;
> >> -}
> >> -
> >> -/*
> >> - * Decide the number of cores in a core complex with the given nr_cores using
> >> - * following set constants MAX_CCX, MAX_CORES_IN_CCX, MAX_CORES_IN_NODE and
> >> - * MAX_NODES_PER_SOCKET. Maintain symmetry as much as possible
> >> - * L3 cache is shared across all cores in a core complex. So, this will also
> >> - * tell us how many cores are sharing the L3 cache.
> >> - */
> >> -static int cores_in_core_complex(int nr_cores)
> >> -{
> >> -    int nodes;
> >> -
> >> -    /* Check if we can fit all the cores in one core complex */
> >> -    if (nr_cores <= MAX_CORES_IN_CCX) {
> >> -        return nr_cores;
> >> -    }
> >> -    /* Get the number of nodes required to build this config */
> >> -    nodes = nodes_in_socket(nr_cores);
> >> -
> >> -    /*
> >> -     * Divide the cores accros all the core complexes
> >> -     * Return rounded up value
> >> -     */
> >> -    return DIV_ROUND_UP(nr_cores, nodes * MAX_CCX);
> >> -}
> >> -
> >>  /* Encode cache info for CPUID[8000001D] */
> >> -static void encode_cache_cpuid8000001d(CPUCacheInfo *cache, CPUState *cs,
> >> -                                uint32_t *eax, uint32_t *ebx,
> >> -                                uint32_t *ecx, uint32_t *edx)
> >> +static void encode_cache_cpuid8000001d(CPUCacheInfo *cache,
> >> +                                       X86CPUTopoInfo *topo_info,
> >> +                                       uint32_t *eax, uint32_t *ebx,
> >> +                                       uint32_t *ecx, uint32_t *edx)
> >>  {
> >>      uint32_t l3_cores;
> >> +    unsigned nodes = MAX(topo_info->nodes_per_pkg, 1);
> >> +
> >>      assert(cache->size == cache->line_size * cache->associativity *
> >>                            cache->partitions * cache->sets);
> >>  
> >> @@ -408,10 +355,13 @@ static void encode_cache_cpuid8000001d(CPUCacheInfo *cache, CPUState *cs,
> >>  
> >>      /* L3 is shared among multiple cores */
> >>      if (cache->level == 3) {
> >> -        l3_cores = cores_in_core_complex(cs->nr_cores);
> >> -        *eax |= ((l3_cores * cs->nr_threads) - 1) << 14;
> >> +        l3_cores = DIV_ROUND_UP((topo_info->dies_per_pkg *
> >> +                                 topo_info->cores_per_die *
> >> +                                 topo_info->threads_per_core),
> >> +                                 nodes);
> >> +        *eax |= (l3_cores - 1) << 14;
> >>      } else {
> >> -        *eax |= ((cs->nr_threads - 1) << 14);
> >> +        *eax |= ((topo_info->threads_per_core - 1) << 14);
> >>      }
> >>  
> >>      assert(cache->line_size > 0);
> >> @@ -431,55 +381,17 @@ static void encode_cache_cpuid8000001d(CPUCacheInfo *cache, CPUState *cs,
> >>             (cache->complex_indexing ? CACHE_COMPLEX_IDX : 0);
> >>  }
> >>  
> >> -/* Data structure to hold the configuration info for a given core index */
> >> -struct core_topology {
> >> -    /* core complex id of the current core index */
> >> -    int ccx_id;
> >> -    /*
> >> -     * Adjusted core index for this core in the topology
> >> -     * This can be 0,1,2,3 with max 4 cores in a core complex
> >> -     */
> >> -    int core_id;
> >> -    /* Node id for this core index */
> >> -    int node_id;
> >> -    /* Number of nodes in this config */
> >> -    int num_nodes;
> >> -};
> >> -
> >> -/*
> >> - * Build the configuration closely match the EPYC hardware. Using the EPYC
> >> - * hardware configuration values (MAX_CCX, MAX_CORES_IN_CCX, MAX_CORES_IN_NODE)
> >> - * right now. This could change in future.
> >> - * nr_cores : Total number of cores in the config
> >> - * core_id  : Core index of the current CPU
> >> - * topo     : Data structure to hold all the config info for this core index
> >> - */
> >> -static void build_core_topology(int nr_cores, int core_id,
> >> -                                struct core_topology *topo)
> >> -{
> >> -    int nodes, cores_in_ccx;
> >> -
> >> -    /* First get the number of nodes required */
> >> -    nodes = nodes_in_socket(nr_cores);
> >> -
> >> -    cores_in_ccx = cores_in_core_complex(nr_cores);
> >> -
> >> -    topo->node_id = core_id / (cores_in_ccx * MAX_CCX);
> >> -    topo->ccx_id = (core_id % (cores_in_ccx * MAX_CCX)) / cores_in_ccx;
> >> -    topo->core_id = core_id % cores_in_ccx;
> >> -    topo->num_nodes = nodes;
> >> -}
> >> -
> >>  /* Encode cache info for CPUID[8000001E] */
> >> -static void encode_topo_cpuid8000001e(CPUState *cs, X86CPU *cpu,
> >> +static void encode_topo_cpuid8000001e(X86CPUTopoInfo *topo_info, X86CPU *cpu,
> >>                                         uint32_t *eax, uint32_t *ebx,
> >>                                         uint32_t *ecx, uint32_t *edx)
> >>  {
> >> -    struct core_topology topo = {0};
> >> -    unsigned long nodes;
> >> +    X86CPUTopoIDs topo_ids = {0};
> >> +    unsigned long nodes = MAX(topo_info->nodes_per_pkg, 1);
> >>      int shift;
> >>  
> >> -    build_core_topology(cs->nr_cores, cpu->core_id, &topo);
> >> +    x86_topo_ids_from_apicid_epyc(cpu->apic_id, topo_info, &topo_ids);
> >> +
> >>      *eax = cpu->apic_id;
> >>      /*
> >>       * CPUID_Fn8000001E_EBX
> >> @@ -496,12 +408,8 @@ static void encode_topo_cpuid8000001e(CPUState *cs, X86CPU *cpu,
> >>       *             3 Core complex id
> >>       *           1:0 Core id
> >>       */
> >> -    if (cs->nr_threads - 1) {
> >> -        *ebx = ((cs->nr_threads - 1) << 8) | (topo.node_id << 3) |
> >> -                (topo.ccx_id << 2) | topo.core_id;
> >> -    } else {
> >> -        *ebx = (topo.node_id << 4) | (topo.ccx_id << 3) | topo.core_id;
> >> -    }
> >> +    *ebx = ((topo_info->threads_per_core - 1) << 8) | (topo_ids.node_id << 3) |
> >> +            (topo_ids.core_id);
> >>      /*
> >>       * CPUID_Fn8000001E_ECX
> >>       * 31:11 Reserved
> >> @@ -510,9 +418,9 @@ static void encode_topo_cpuid8000001e(CPUState *cs, X86CPU *cpu,
> >>       *         2  Socket id
> >>       *       1:0  Node id
> >>       */
> >> -    if (topo.num_nodes <= 4) {
> >> -        *ecx = ((topo.num_nodes - 1) << 8) | (cpu->socket_id << 2) |
> >> -                topo.node_id;
> >> +
> >> +    if (nodes <= 4) {
> >> +        *ecx = ((nodes - 1) << 8) | (topo_ids.pkg_id << 2) | topo_ids.node_id;
> >>      } else {
> >>          /*
> >>           * Node id fix up. Actual hardware supports up to 4 nodes. But with
> >> @@ -527,10 +435,10 @@ static void encode_topo_cpuid8000001e(CPUState *cs, X86CPU *cpu,
> >>           * number of nodes. find_last_bit returns last set bit(0 based). Left
> >>           * shift(+1) the socket id to represent all the nodes.
> >>           */
> >> -        nodes = topo.num_nodes - 1;
> >> +        nodes -= 1;
> >>          shift = find_last_bit(&nodes, 8);
> >> -        *ecx = ((topo.num_nodes - 1) << 8) | (cpu->socket_id << (shift + 1)) |
> >> -                topo.node_id;
> >> +        *ecx = (nodes << 8) | (topo_ids.pkg_id << (shift + 1)) |
> >> +               topo_ids.node_id;
> >>      }
> >>      *edx = 0;
> >>  }
> >> @@ -5318,6 +5226,7 @@ void cpu_x86_cpuid(CPUX86State *env, uint32_t index, uint32_t count,
> >>      uint32_t signature[3];
> >>      X86CPUTopoInfo topo_info;
> >>  
> >> +    topo_info.nodes_per_pkg = env->nr_nodes;
> >>      topo_info.dies_per_pkg = env->nr_dies;
> >>      topo_info.cores_per_die = cs->nr_cores;
> >>      topo_info.threads_per_core = cs->nr_threads;
> >> @@ -5737,20 +5646,20 @@ void cpu_x86_cpuid(CPUX86State *env, uint32_t index, uint32_t count,
> >>          }
> >>          switch (count) {
> >>          case 0: /* L1 dcache info */
> >> -            encode_cache_cpuid8000001d(env->cache_info_amd.l1d_cache, cs,
> >> -                                       eax, ebx, ecx, edx);
> >> +            encode_cache_cpuid8000001d(env->cache_info_amd.l1d_cache,
> >> +                                       &topo_info, eax, ebx, ecx, edx);
> >>              break;
> >>          case 1: /* L1 icache info */
> >> -            encode_cache_cpuid8000001d(env->cache_info_amd.l1i_cache, cs,
> >> -                                       eax, ebx, ecx, edx);
> >> +            encode_cache_cpuid8000001d(env->cache_info_amd.l1i_cache,
> >> +                                       &topo_info, eax, ebx, ecx, edx);
> >>              break;
> >>          case 2: /* L2 cache info */
> >> -            encode_cache_cpuid8000001d(env->cache_info_amd.l2_cache, cs,
> >> -                                       eax, ebx, ecx, edx);
> >> +            encode_cache_cpuid8000001d(env->cache_info_amd.l2_cache,
> >> +                                       &topo_info, eax, ebx, ecx, edx);
> >>              break;
> >>          case 3: /* L3 cache info */
> >> -            encode_cache_cpuid8000001d(env->cache_info_amd.l3_cache, cs,
> >> -                                       eax, ebx, ecx, edx);
> >> +            encode_cache_cpuid8000001d(env->cache_info_amd.l3_cache,
> >> +                                       &topo_info, eax, ebx, ecx, edx);
> >>              break;
> >>          default: /* end of info */
> >>              *eax = *ebx = *ecx = *edx = 0;
> >> @@ -5759,8 +5668,7 @@ void cpu_x86_cpuid(CPUX86State *env, uint32_t index, uint32_t count,
> >>          break;
> >>      case 0x8000001E:
> >>          assert(cpu->core_id <= 255);
> >> -        encode_topo_cpuid8000001e(cs, cpu,
> >> -                                  eax, ebx, ecx, edx);
> >> +        encode_topo_cpuid8000001e(&topo_info, cpu, eax, ebx, ecx, edx);
> >>          break;
> >>      case 0xC0000000:
> >>          *eax = env->cpuid_xlevel2;
> >>  
> >   
> 



  reply	other threads:[~2020-02-25  7:49 UTC|newest]

Thread overview: 44+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2020-02-13 18:16 [PATCH v4 00/16] APIC ID fixes for AMD EPYC CPU model Babu Moger
2020-02-13 18:16 ` [PATCH v4 01/16] hw/i386: Rename X86CPUTopoInfo structure to X86CPUTopoIDs Babu Moger
2020-02-13 18:16 ` [PATCH v4 02/16] hw/i386: Introduce X86CPUTopoInfo to contain topology info Babu Moger
2020-02-13 18:16 ` [PATCH v4 03/16] hw/i386: Consolidate topology functions Babu Moger
2020-02-13 18:16 ` [PATCH v4 04/16] hw/i386: Introduce init_topo_info to initialize X86CPUTopoInfo Babu Moger
2020-02-21 17:05   ` Igor Mammedov
2020-02-21 17:51     ` Babu Moger
2020-02-24  8:18       ` Igor Mammedov
2020-02-24 16:54         ` Babu Moger
2020-02-13 18:16 ` [PATCH v4 05/16] machine: Add SMP Sockets in CpuTopology Babu Moger
2020-02-24  8:37   ` Igor Mammedov
2020-02-13 18:17 ` [PATCH v4 06/16] hw/i386: Update structures for nodes_per_pkg Babu Moger
2020-02-24  8:34   ` Igor Mammedov
2020-02-24 17:12     ` Babu Moger
2020-02-25  7:42       ` Igor Mammedov
2020-02-13 18:17 ` [PATCH v4 07/16] hw/i386: Rename apicid_from_topo_ids to x86_apicid_from_topo_ids Babu Moger
2020-02-20 12:43   ` Igor Mammedov
2020-02-13 18:17 ` [PATCH v4 08/16] hw/386: Add EPYC mode topology decoding functions Babu Moger
2020-02-24  8:50   ` Igor Mammedov
2020-02-24 17:24     ` Babu Moger
2020-02-13 18:17 ` [PATCH v4 09/16] target/i386: Cleanup and use the EPYC mode topology functions Babu Moger
2020-02-24  8:52   ` Igor Mammedov
2020-02-24 17:29     ` Babu Moger
2020-02-25  7:49       ` Igor Mammedov [this message]
2020-02-25 15:10         ` Babu Moger
2020-03-02 17:09         ` Babu Moger
2020-03-03  8:47           ` Igor Mammedov
2020-02-13 18:17 ` [PATCH v4 10/16] hw/i386: Introduce apicid functions inside X86MachineState Babu Moger
2020-02-24 17:01   ` Igor Mammedov
2020-02-24 17:30     ` Babu Moger
2020-02-13 18:17 ` [PATCH v4 11/16] target/i386: Load apicid model specific handlers from X86CPUDefinition Babu Moger
2020-02-13 18:17 ` [PATCH v4 12/16] hw/i386: Use the apicid handlers from X86MachineState Babu Moger
2020-02-24 17:19   ` Igor Mammedov
2020-02-24 17:58     ` Babu Moger
2020-02-24 22:31       ` Eduardo Habkost
2020-02-24 23:13         ` Babu Moger
2020-02-25 15:32           ` Eduardo Habkost
2020-02-25 15:41             ` Babu Moger
2020-02-25  8:16         ` Igor Mammedov
2020-02-25 15:26           ` Eduardo Habkost
2020-02-13 18:17 ` [PATCH v4 13/16] target/i386: Add EPYC model specific handlers Babu Moger
2020-02-13 18:17 ` [PATCH v4 14/16] hw/i386: Move arch_id decode inside x86_cpus_init Babu Moger
2020-02-13 18:18 ` [PATCH v4 15/16] i386: Fix pkg_id offset for EPYC cpu models Babu Moger
2020-02-13 18:18 ` [PATCH v4 16/16] tests: Update the Unit tests Babu Moger

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20200225084905.11b9731d@redhat.com \
    --to=imammedo@redhat.com \
    --cc=babu.moger@amd.com \
    --cc=ehabkost@redhat.com \
    --cc=mst@redhat.com \
    --cc=pbonzini@redhat.com \
    --cc=qemu-devel@nongnu.org \
    --cc=rth@twiddle.net \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).