Message ID | 1526949675-106737-3-git-send-email-babu.moger@amd.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Babu, If num_sharing_l3_cache() uses MAX_NODES_EPYC, then that function It’s EPYC specific. An alternative would be to use a data member (e.g., max_nodes_per_socket)) that get initialized (via another helper function) to MAX_NODES_EPYC. Basically, ideally the functions that return CPUID information do *not* use EPYC-specific macros, like MAX_NODES_EPYC. Leo. > -----Original Message----- > From: Moger, Babu > Sent: Monday, May 21, 2018 7:41 PM > To: mst@redhat.com; marcel.apfelbaum@gmail.com; pbonzini@redhat.com; > rth@twiddle.net; ehabkost@redhat.com; mtosatti@redhat.com > Cc: qemu-devel@nongnu.org; kvm@vger.kernel.org; Moger, Babu > <Babu.Moger@amd.com>; kash@tripleback.net; geoff@hostfission.com > Subject: [PATCH v10 2/5] i386: Populate AMD Processor Cache Information > for cpuid 0x8000001D > > Add information for cpuid 0x8000001D leaf. Populate cache topology > information for different cache types(Data Cache, Instruction Cache, L2 and > L3) supported by 0x8000001D leaf. Please refer Processor Programming > Reference (PPR) for AMD Family 17h Model for more details. > > Signed-off-by: Babu Moger <babu.moger@amd.com> > --- > target/i386/cpu.c | 103 > ++++++++++++++++++++++++++++++++++++++++++++++++++++++ > target/i386/kvm.c | 29 +++++++++++++-- > 2 files changed, 129 insertions(+), 3 deletions(-) > > diff --git a/target/i386/cpu.c b/target/i386/cpu.c index d9773b6..1dd060a > 100644 > --- a/target/i386/cpu.c > +++ b/target/i386/cpu.c > @@ -336,6 +336,85 @@ static void > encode_cache_cpuid80000006(CPUCacheInfo *l2, > } > } > > +/* Definitions used for building CPUID Leaf 0x8000001D and 0x8000001E > +*/ > +/* Please refer AMD64 Architecture Programmer’s Manual Volume 3 */ > +#define MAX_CCX 2 #define MAX_CORES_IN_CCX 4 #define > MAX_NODES_EPYC 4 > +#define MAX_CORES_IN_NODE 8 > + > +/* Number of logical processors sharing L3 cache */ > +#define NUM_SHARING_CACHE(threads, num_sharing) ((threads > 1) ? \ > + (((num_sharing - 1) * threads) + 1) : \ > + (num_sharing - 1)) > +/* > + * L3 Cache is shared between all the cores in a core complex. > + * Maximum cores that can share L3 is 4. > + */ > +static int num_sharing_l3_cache(int nr_cores) { > + int i, nodes = 1; > + > + /* Check if we can fit all the cores in one CCX */ > + if (nr_cores <= MAX_CORES_IN_CCX) { > + return nr_cores; > + } > + /* > + * Figure out the number of nodes(or dies) required to build > + * this config. Max cores in a node is 8 > + */ > + for (i = nodes; i <= MAX_NODES_EPYC; i++) { > + if (nr_cores <= (i * MAX_CORES_IN_NODE)) { > + nodes = i; > + break; > + } > + /* We support nodes 1, 2, 4 */ > + if (i == 3) { > + continue; > + } > + } > + /* Spread the cores accros all the CCXs and return max cores in a ccx */ > + return (nr_cores / (nodes * MAX_CCX)) + > + ((nr_cores % (nodes * MAX_CCX)) ? 1 : 0); } > + > +/* Encode cache info for CPUID[8000001D] */ static void > +encode_cache_cpuid8000001d(CPUCacheInfo *cache, CPUState *cs, > + uint32_t *eax, uint32_t *ebx, > + uint32_t *ecx, uint32_t *edx) { > + uint32_t num_share_l3; > + assert(cache->size == cache->line_size * cache->associativity * > + cache->partitions * cache->sets); > + > + *eax = CACHE_TYPE(cache->type) | CACHE_LEVEL(cache->level) | > + (cache->self_init ? CACHE_SELF_INIT_LEVEL : 0); > + > + /* L3 is shared among multiple cores */ > + if (cache->level == 3) { > + num_share_l3 = num_sharing_l3_cache(cs->nr_cores); > + *eax |= (NUM_SHARING_CACHE(cs->nr_threads, num_share_l3) << > 14); > + } else { > + *eax |= ((cs->nr_threads - 1) << 14); > + } > + > + assert(cache->line_size > 0); > + assert(cache->partitions > 0); > + assert(cache->associativity > 0); > + /* We don't implement fully-associative caches */ > + assert(cache->associativity < cache->sets); > + *ebx = (cache->line_size - 1) | > + ((cache->partitions - 1) << 12) | > + ((cache->associativity - 1) << 22); > + > + assert(cache->sets > 0); > + *ecx = cache->sets - 1; > + > + *edx = (cache->no_invd_sharing ? CACHE_NO_INVD_SHARING : 0) | > + (cache->inclusive ? CACHE_INCLUSIVE : 0) | > + (cache->complex_indexing ? CACHE_COMPLEX_IDX : 0); } > + > /* > * Definitions of the hardcoded cache entries we expose: > * These are legacy cache values. If there is a need to change any @@ - > 4005,6 +4084,30 @@ void cpu_x86_cpuid(CPUX86State *env, uint32_t index, > uint32_t count, > *edx = 0; > } > break; > + case 0x8000001D: > + *eax = 0; > + switch (count) { > + case 0: /* L1 dcache info */ > + encode_cache_cpuid8000001d(env->cache_info_amd.l1d_cache, cs, > + eax, ebx, ecx, edx); > + break; > + case 1: /* L1 icache info */ > + encode_cache_cpuid8000001d(env->cache_info_amd.l1i_cache, cs, > + eax, ebx, ecx, edx); > + break; > + case 2: /* L2 cache info */ > + encode_cache_cpuid8000001d(env->cache_info_amd.l2_cache, cs, > + eax, ebx, ecx, edx); > + break; > + case 3: /* L3 cache info */ > + encode_cache_cpuid8000001d(env->cache_info_amd.l3_cache, cs, > + eax, ebx, ecx, edx); > + break; > + default: /* end of info */ > + *eax = *ebx = *ecx = *edx = 0; > + break; > + } > + break; > case 0xC0000000: > *eax = env->cpuid_xlevel2; > *ebx = 0; > diff --git a/target/i386/kvm.c b/target/i386/kvm.c index d6666a4..a8bf7eb > 100644 > --- a/target/i386/kvm.c > +++ b/target/i386/kvm.c > @@ -979,9 +979,32 @@ int kvm_arch_init_vcpu(CPUState *cs) > } > c = &cpuid_data.entries[cpuid_i++]; > > - c->function = i; > - c->flags = 0; > - cpu_x86_cpuid(env, i, 0, &c->eax, &c->ebx, &c->ecx, &c->edx); > + switch (i) { > + case 0x8000001d: > + /* Query for all AMD cache information leaves */ > + for (j = 0; ; j++) { > + c->function = i; > + c->flags = KVM_CPUID_FLAG_SIGNIFCANT_INDEX; > + c->index = j; > + cpu_x86_cpuid(env, i, j, &c->eax, &c->ebx, &c->ecx, > + &c->edx); > + > + if (c->eax == 0) { > + break; > + } > + if (cpuid_i == KVM_MAX_CPUID_ENTRIES) { > + fprintf(stderr, "cpuid_data is full, no space for " > + "cpuid(eax:0x%x,ecx:0x%x)\n", i, j); > + abort(); > + } > + c = &cpuid_data.entries[cpuid_i++]; > + } > + break; > + default: > + c->function = i; > + c->flags = 0; > + cpu_x86_cpuid(env, i, 0, &c->eax, &c->ebx, &c->ecx, &c->edx); > + break; > + } > } > > /* Call Centaur's CPUID instructions they are supported. */ > -- > 1.8.3.1
> -----Original Message----- > From: Duran, Leo > Sent: Monday, May 21, 2018 8:32 PM > To: Moger, Babu <Babu.Moger@amd.com>; mst@redhat.com; > marcel.apfelbaum@gmail.com; pbonzini@redhat.com; rth@twiddle.net; > ehabkost@redhat.com; mtosatti@redhat.com > Cc: qemu-devel@nongnu.org; kvm@vger.kernel.org; kash@tripleback.net; > geoff@hostfission.com > Subject: RE: [PATCH v10 2/5] i386: Populate AMD Processor Cache > Information for cpuid 0x8000001D > > Babu, > > If num_sharing_l3_cache() uses MAX_NODES_EPYC, then that function It’s > EPYC specific. > > An alternative would be to use a data member (e.g., > max_nodes_per_socket)) that get initialized (via another helper function) to > MAX_NODES_EPYC. Thanks Leo. Let me see how we can handle this. This requires changes in generic Data structure which I tried to avoid here. I will wait for all the comments for whole series before making this change. Note that right now, this feature is only enabled for EPYC. Yes. I know this could this in future. > Basically, ideally the functions that return CPUID information do *not* use > EPYC-specific macros, like MAX_NODES_EPYC. > > Leo. > > > -----Original Message----- > > From: Moger, Babu > > Sent: Monday, May 21, 2018 7:41 PM > > To: mst@redhat.com; marcel.apfelbaum@gmail.com; > pbonzini@redhat.com; > > rth@twiddle.net; ehabkost@redhat.com; mtosatti@redhat.com > > Cc: qemu-devel@nongnu.org; kvm@vger.kernel.org; Moger, Babu > > <Babu.Moger@amd.com>; kash@tripleback.net; geoff@hostfission.com > > Subject: [PATCH v10 2/5] i386: Populate AMD Processor Cache Information > > for cpuid 0x8000001D > > > > Add information for cpuid 0x8000001D leaf. Populate cache topology > > information for different cache types(Data Cache, Instruction Cache, L2 and > > L3) supported by 0x8000001D leaf. Please refer Processor Programming > > Reference (PPR) for AMD Family 17h Model for more details. > > > > Signed-off-by: Babu Moger <babu.moger@amd.com> > > --- > > target/i386/cpu.c | 103 > > ++++++++++++++++++++++++++++++++++++++++++++++++++++++ > > target/i386/kvm.c | 29 +++++++++++++-- > > 2 files changed, 129 insertions(+), 3 deletions(-) > > > > diff --git a/target/i386/cpu.c b/target/i386/cpu.c index d9773b6..1dd060a > > 100644 > > --- a/target/i386/cpu.c > > +++ b/target/i386/cpu.c > > @@ -336,6 +336,85 @@ static void > > encode_cache_cpuid80000006(CPUCacheInfo *l2, > > } > > } > > > > +/* Definitions used for building CPUID Leaf 0x8000001D and 0x8000001E > > +*/ > > +/* Please refer AMD64 Architecture Programmer’s Manual Volume 3 */ > > +#define MAX_CCX 2 #define MAX_CORES_IN_CCX 4 #define > > MAX_NODES_EPYC 4 > > +#define MAX_CORES_IN_NODE 8 > > + > > +/* Number of logical processors sharing L3 cache */ > > +#define NUM_SHARING_CACHE(threads, num_sharing) ((threads > 1) ? > \ > > + (((num_sharing - 1) * threads) + 1) : \ > > + (num_sharing - 1)) > > +/* > > + * L3 Cache is shared between all the cores in a core complex. > > + * Maximum cores that can share L3 is 4. > > + */ > > +static int num_sharing_l3_cache(int nr_cores) { > > + int i, nodes = 1; > > + > > + /* Check if we can fit all the cores in one CCX */ > > + if (nr_cores <= MAX_CORES_IN_CCX) { > > + return nr_cores; > > + } > > + /* > > + * Figure out the number of nodes(or dies) required to build > > + * this config. Max cores in a node is 8 > > + */ > > + for (i = nodes; i <= MAX_NODES_EPYC; i++) { > > + if (nr_cores <= (i * MAX_CORES_IN_NODE)) { > > + nodes = i; > > + break; > > + } > > + /* We support nodes 1, 2, 4 */ > > + if (i == 3) { > > + continue; > > + } > > + } > > + /* Spread the cores accros all the CCXs and return max cores in a ccx */ > > + return (nr_cores / (nodes * MAX_CCX)) + > > + ((nr_cores % (nodes * MAX_CCX)) ? 1 : 0); } > > + > > +/* Encode cache info for CPUID[8000001D] */ static void > > +encode_cache_cpuid8000001d(CPUCacheInfo *cache, CPUState *cs, > > + uint32_t *eax, uint32_t *ebx, > > + uint32_t *ecx, uint32_t *edx) { > > + uint32_t num_share_l3; > > + assert(cache->size == cache->line_size * cache->associativity * > > + cache->partitions * cache->sets); > > + > > + *eax = CACHE_TYPE(cache->type) | CACHE_LEVEL(cache->level) | > > + (cache->self_init ? CACHE_SELF_INIT_LEVEL : 0); > > + > > + /* L3 is shared among multiple cores */ > > + if (cache->level == 3) { > > + num_share_l3 = num_sharing_l3_cache(cs->nr_cores); > > + *eax |= (NUM_SHARING_CACHE(cs->nr_threads, num_share_l3) << > > 14); > > + } else { > > + *eax |= ((cs->nr_threads - 1) << 14); > > + } > > + > > + assert(cache->line_size > 0); > > + assert(cache->partitions > 0); > > + assert(cache->associativity > 0); > > + /* We don't implement fully-associative caches */ > > + assert(cache->associativity < cache->sets); > > + *ebx = (cache->line_size - 1) | > > + ((cache->partitions - 1) << 12) | > > + ((cache->associativity - 1) << 22); > > + > > + assert(cache->sets > 0); > > + *ecx = cache->sets - 1; > > + > > + *edx = (cache->no_invd_sharing ? CACHE_NO_INVD_SHARING : 0) | > > + (cache->inclusive ? CACHE_INCLUSIVE : 0) | > > + (cache->complex_indexing ? CACHE_COMPLEX_IDX : 0); } > > + > > /* > > * Definitions of the hardcoded cache entries we expose: > > * These are legacy cache values. If there is a need to change any @@ - > > 4005,6 +4084,30 @@ void cpu_x86_cpuid(CPUX86State *env, uint32_t > index, > > uint32_t count, > > *edx = 0; > > } > > break; > > + case 0x8000001D: > > + *eax = 0; > > + switch (count) { > > + case 0: /* L1 dcache info */ > > + encode_cache_cpuid8000001d(env->cache_info_amd.l1d_cache, > cs, > > + eax, ebx, ecx, edx); > > + break; > > + case 1: /* L1 icache info */ > > + encode_cache_cpuid8000001d(env->cache_info_amd.l1i_cache, cs, > > + eax, ebx, ecx, edx); > > + break; > > + case 2: /* L2 cache info */ > > + encode_cache_cpuid8000001d(env->cache_info_amd.l2_cache, cs, > > + eax, ebx, ecx, edx); > > + break; > > + case 3: /* L3 cache info */ > > + encode_cache_cpuid8000001d(env->cache_info_amd.l3_cache, cs, > > + eax, ebx, ecx, edx); > > + break; > > + default: /* end of info */ > > + *eax = *ebx = *ecx = *edx = 0; > > + break; > > + } > > + break; > > case 0xC0000000: > > *eax = env->cpuid_xlevel2; > > *ebx = 0; > > diff --git a/target/i386/kvm.c b/target/i386/kvm.c index d6666a4..a8bf7eb > > 100644 > > --- a/target/i386/kvm.c > > +++ b/target/i386/kvm.c > > @@ -979,9 +979,32 @@ int kvm_arch_init_vcpu(CPUState *cs) > > } > > c = &cpuid_data.entries[cpuid_i++]; > > > > - c->function = i; > > - c->flags = 0; > > - cpu_x86_cpuid(env, i, 0, &c->eax, &c->ebx, &c->ecx, &c->edx); > > + switch (i) { > > + case 0x8000001d: > > + /* Query for all AMD cache information leaves */ > > + for (j = 0; ; j++) { > > + c->function = i; > > + c->flags = KVM_CPUID_FLAG_SIGNIFCANT_INDEX; > > + c->index = j; > > + cpu_x86_cpuid(env, i, j, &c->eax, &c->ebx, &c->ecx, > > + &c->edx); > > + > > + if (c->eax == 0) { > > + break; > > + } > > + if (cpuid_i == KVM_MAX_CPUID_ENTRIES) { > > + fprintf(stderr, "cpuid_data is full, no space for " > > + "cpuid(eax:0x%x,ecx:0x%x)\n", i, j); > > + abort(); > > + } > > + c = &cpuid_data.entries[cpuid_i++]; > > + } > > + break; > > + default: > > + c->function = i; > > + c->flags = 0; > > + cpu_x86_cpuid(env, i, 0, &c->eax, &c->ebx, &c->ecx, &c->edx); > > + break; > > + } > > } > > > > /* Call Centaur's CPUID instructions they are supported. */ > > -- > > 1.8.3.1
On Mon, May 21, 2018 at 08:41:12PM -0400, Babu Moger wrote: > Add information for cpuid 0x8000001D leaf. Populate cache topology information > for different cache types(Data Cache, Instruction Cache, L2 and L3) supported > by 0x8000001D leaf. Please refer Processor Programming Reference (PPR) for AMD > Family 17h Model for more details. > > Signed-off-by: Babu Moger <babu.moger@amd.com> > --- > target/i386/cpu.c | 103 ++++++++++++++++++++++++++++++++++++++++++++++++++++++ > target/i386/kvm.c | 29 +++++++++++++-- > 2 files changed, 129 insertions(+), 3 deletions(-) > > diff --git a/target/i386/cpu.c b/target/i386/cpu.c > index d9773b6..1dd060a 100644 > --- a/target/i386/cpu.c > +++ b/target/i386/cpu.c > @@ -336,6 +336,85 @@ static void encode_cache_cpuid80000006(CPUCacheInfo *l2, > } > } > The number of variables here is large, so maybe we should document what each one mean so it's easier to review: > +/* Definitions used for building CPUID Leaf 0x8000001D and 0x8000001E */ > +/* Please refer AMD64 Architecture Programmer’s Manual Volume 3 */ > +#define MAX_CCX 2 CCX is "core complex", right? A comment would be useful here. > +#define MAX_CORES_IN_CCX 4 > +#define MAX_NODES_EPYC 4 A comment explaining why it's OK to use a EPYC-specific constant here would be useful. > +#define MAX_CORES_IN_NODE 8 > + > +/* Number of logical processors sharing L3 cache */ > +#define NUM_SHARING_CACHE(threads, num_sharing) ((threads > 1) ? \ > + (((num_sharing - 1) * threads) + 1) : \ > + (num_sharing - 1)) This formula is confusing to me. If 4 cores are sharing the cache and threads==1, 4 logical processors share the cache, and we return 3. Sounds OK. But, if 4 cores are sharing the cache and threads==2, the number of logical processors sharing the cache is 8. We should return 7. The formula above returns (((4 - 1) * 2) + 1), which is correct. But isn't it simpler to write this as: #define NUM_SHARING_CACHE(threads, num_sharing) \ (((num_sharing) * (threads)) - 1) (Maybe the "- 1" part could be moved outside the macro for clarity. See below.) > +/* > + * L3 Cache is shared between all the cores in a core complex. > + * Maximum cores that can share L3 is 4. > + */ > +static int num_sharing_l3_cache(int nr_cores) Can we document what exactly this function is going to return? This returns the number of cores sharing l3 cache, not the number of logical processors, correct? > +{ > + int i, nodes = 1; > + > + /* Check if we can fit all the cores in one CCX */ > + if (nr_cores <= MAX_CORES_IN_CCX) { > + return nr_cores; > + } > + /* > + * Figure out the number of nodes(or dies) required to build > + * this config. Max cores in a node is 8 > + */ > + for (i = nodes; i <= MAX_NODES_EPYC; i++) { > + if (nr_cores <= (i * MAX_CORES_IN_NODE)) { > + nodes = i; > + break; > + } > + /* We support nodes 1, 2, 4 */ > + if (i == 3) { > + continue; > + } > + } "continue" as the very last statement of a for loop does nothing, so it looks like this could be written as: for (i = nodes; i <= MAX_NODES_EPYC; i++) { if (nr_cores <= (i * MAX_CORES_IN_NODE)) { nodes = i; break; } } which in turn seems to be the same as: nodes = DIV_ROUND_UP(nr_cores, MAX_CORES_IN_NODE); nodes = MIN(nodes, MAX_NODES_EPYC) But, is this really what we want here? > + /* Spread the cores accros all the CCXs and return max cores in a ccx */ > + return (nr_cores / (nodes * MAX_CCX)) + > + ((nr_cores % (nodes * MAX_CCX)) ? 1 : 0); This also seems to be the same as DIV_ROUND_UP? return DIV_ROUND_UP(nr_cores, nodes * MAX_CCX); I didn't confirm the logic is valid, though, because I don't know what we should expect. What is the expected return value of this function in the following cases? -smp 24,sockets=2,cores=12,threads=1 -smp 64,sockets=2,cores=32,threads=1 > +} > + > +/* Encode cache info for CPUID[8000001D] */ > +static void encode_cache_cpuid8000001d(CPUCacheInfo *cache, CPUState *cs, > + uint32_t *eax, uint32_t *ebx, > + uint32_t *ecx, uint32_t *edx) > +{ > + uint32_t num_share_l3; > + assert(cache->size == cache->line_size * cache->associativity * > + cache->partitions * cache->sets); > + > + *eax = CACHE_TYPE(cache->type) | CACHE_LEVEL(cache->level) | > + (cache->self_init ? CACHE_SELF_INIT_LEVEL : 0); > + > + /* L3 is shared among multiple cores */ > + if (cache->level == 3) { > + num_share_l3 = num_sharing_l3_cache(cs->nr_cores); > + *eax |= (NUM_SHARING_CACHE(cs->nr_threads, num_share_l3) << 14); Considering that the line below has an explicit "- 1", I think the "- 1" part could be moved outside the NUM_SHARING_CACHE macro, and used explicitly here. But then the NUM_SHARING_CACHE would be just a simple multiplication, so this could be simply written as: /* num_sharing_l3_cache() renamed to cores_sharing_l3_cache() */ uint32_t l3_cores = cores_sharing_l3_cache(cs->nr_cores); uint32_t l3_logical_processors = l3_cores * cs->nr_threads; *eax |= (l3_logical_processors - 1) << 14; > + } else { > + *eax |= ((cs->nr_threads - 1) << 14); > + } > + > + assert(cache->line_size > 0); > + assert(cache->partitions > 0); > + assert(cache->associativity > 0); > + /* We don't implement fully-associative caches */ > + assert(cache->associativity < cache->sets); > + *ebx = (cache->line_size - 1) | > + ((cache->partitions - 1) << 12) | > + ((cache->associativity - 1) << 22); > + > + assert(cache->sets > 0); > + *ecx = cache->sets - 1; > + > + *edx = (cache->no_invd_sharing ? CACHE_NO_INVD_SHARING : 0) | > + (cache->inclusive ? CACHE_INCLUSIVE : 0) | > + (cache->complex_indexing ? CACHE_COMPLEX_IDX : 0); > +} > + > /* > * Definitions of the hardcoded cache entries we expose: > * These are legacy cache values. If there is a need to change any > @@ -4005,6 +4084,30 @@ void cpu_x86_cpuid(CPUX86State *env, uint32_t index, uint32_t count, > *edx = 0; > } > break; > + case 0x8000001D: > + *eax = 0; > + switch (count) { > + case 0: /* L1 dcache info */ > + encode_cache_cpuid8000001d(env->cache_info_amd.l1d_cache, cs, > + eax, ebx, ecx, edx); > + break; > + case 1: /* L1 icache info */ > + encode_cache_cpuid8000001d(env->cache_info_amd.l1i_cache, cs, > + eax, ebx, ecx, edx); > + break; > + case 2: /* L2 cache info */ > + encode_cache_cpuid8000001d(env->cache_info_amd.l2_cache, cs, > + eax, ebx, ecx, edx); > + break; > + case 3: /* L3 cache info */ > + encode_cache_cpuid8000001d(env->cache_info_amd.l3_cache, cs, > + eax, ebx, ecx, edx); > + break; > + default: /* end of info */ > + *eax = *ebx = *ecx = *edx = 0; > + break; > + } > + break; > case 0xC0000000: > *eax = env->cpuid_xlevel2; > *ebx = 0; > diff --git a/target/i386/kvm.c b/target/i386/kvm.c > index d6666a4..a8bf7eb 100644 > --- a/target/i386/kvm.c > +++ b/target/i386/kvm.c > @@ -979,9 +979,32 @@ int kvm_arch_init_vcpu(CPUState *cs) > } > c = &cpuid_data.entries[cpuid_i++]; > > - c->function = i; > - c->flags = 0; > - cpu_x86_cpuid(env, i, 0, &c->eax, &c->ebx, &c->ecx, &c->edx); > + switch (i) { > + case 0x8000001d: > + /* Query for all AMD cache information leaves */ > + for (j = 0; ; j++) { > + c->function = i; > + c->flags = KVM_CPUID_FLAG_SIGNIFCANT_INDEX; > + c->index = j; > + cpu_x86_cpuid(env, i, j, &c->eax, &c->ebx, &c->ecx, &c->edx); > + > + if (c->eax == 0) { > + break; > + } > + if (cpuid_i == KVM_MAX_CPUID_ENTRIES) { > + fprintf(stderr, "cpuid_data is full, no space for " > + "cpuid(eax:0x%x,ecx:0x%x)\n", i, j); > + abort(); > + } > + c = &cpuid_data.entries[cpuid_i++]; > + } > + break; > + default: > + c->function = i; > + c->flags = 0; > + cpu_x86_cpuid(env, i, 0, &c->eax, &c->ebx, &c->ecx, &c->edx); > + break; > + } > } > > /* Call Centaur's CPUID instructions they are supported. */ > -- > 1.8.3.1 >
On Tue, May 22, 2018 at 01:32:52PM +0000, Moger, Babu wrote: > > > -----Original Message----- > > From: Duran, Leo > > Sent: Monday, May 21, 2018 8:32 PM > > To: Moger, Babu <Babu.Moger@amd.com>; mst@redhat.com; > > marcel.apfelbaum@gmail.com; pbonzini@redhat.com; rth@twiddle.net; > > ehabkost@redhat.com; mtosatti@redhat.com > > Cc: qemu-devel@nongnu.org; kvm@vger.kernel.org; kash@tripleback.net; > > geoff@hostfission.com > > Subject: RE: [PATCH v10 2/5] i386: Populate AMD Processor Cache > > Information for cpuid 0x8000001D > > > > Babu, > > > > If num_sharing_l3_cache() uses MAX_NODES_EPYC, then that function It’s > > EPYC specific. > > > > An alternative would be to use a data member (e.g., > > max_nodes_per_socket)) that get initialized (via another helper function) to > > MAX_NODES_EPYC. > > Thanks Leo. Let me see how we can handle this. This requires changes in generic > Data structure which I tried to avoid here. I will wait for all the comments for whole > series before making this change. Note that right now, this feature is only enabled > for EPYC. Yes. I know this could this in future. We just need a reasonable default, by now, and it can even be the same value used on EPYC. This default just need to generate reasonable results for other cases that don't match real hardware (like cores=32 or cores=12).
> -----Original Message----- > From: Eduardo Habkost [mailto:ehabkost@redhat.com] > Sent: Tuesday, May 22, 2018 9:04 AM > To: Moger, Babu <Babu.Moger@amd.com> > Cc: Duran, Leo <leo.duran@amd.com>; mst@redhat.com; > marcel.apfelbaum@gmail.com; pbonzini@redhat.com; rth@twiddle.net; > mtosatti@redhat.com; qemu-devel@nongnu.org; kvm@vger.kernel.org; > kash@tripleback.net; geoff@hostfission.com > Subject: Re: [PATCH v10 2/5] i386: Populate AMD Processor Cache > Information for cpuid 0x8000001D > > On Tue, May 22, 2018 at 01:32:52PM +0000, Moger, Babu wrote: > > > > > -----Original Message----- > > > From: Duran, Leo > > > Sent: Monday, May 21, 2018 8:32 PM > > > To: Moger, Babu <Babu.Moger@amd.com>; mst@redhat.com; > > > marcel.apfelbaum@gmail.com; pbonzini@redhat.com; rth@twiddle.net; > > > ehabkost@redhat.com; mtosatti@redhat.com > > > Cc: qemu-devel@nongnu.org; kvm@vger.kernel.org; > kash@tripleback.net; > > > geoff@hostfission.com > > > Subject: RE: [PATCH v10 2/5] i386: Populate AMD Processor Cache > > > Information for cpuid 0x8000001D > > > > > > Babu, > > > > > > If num_sharing_l3_cache() uses MAX_NODES_EPYC, then that function > It’s > > > EPYC specific. > > > > > > An alternative would be to use a data member (e.g., > > > max_nodes_per_socket)) that get initialized (via another helper > function) to > > > MAX_NODES_EPYC. > > > > Thanks Leo. Let me see how we can handle this. This requires changes in > generic > > Data structure which I tried to avoid here. I will wait for all the comments > for whole > > series before making this change. Note that right now, this feature is only > enabled > > for EPYC. Yes. I know this could this in future. > > We just need a reasonable default, by now, and it can even be the > same value used on EPYC. This default just need to generate > reasonable results for other cases that don't match real hardware > (like cores=32 or cores=12). Ok. Will change the name to bit generic for now and keep the defaults as in EPYC. > > -- > Eduardo
Hi Eduardo, Please see my comments below. > -----Original Message----- > From: Eduardo Habkost [mailto:ehabkost@redhat.com] > Sent: Tuesday, May 22, 2018 8:54 AM > To: Moger, Babu <Babu.Moger@amd.com> > Cc: mst@redhat.com; marcel.apfelbaum@gmail.com; pbonzini@redhat.com; > rth@twiddle.net; mtosatti@redhat.com; qemu-devel@nongnu.org; > kvm@vger.kernel.org; kash@tripleback.net; geoff@hostfission.com > Subject: Re: [PATCH v10 2/5] i386: Populate AMD Processor Cache > Information for cpuid 0x8000001D > > On Mon, May 21, 2018 at 08:41:12PM -0400, Babu Moger wrote: > > Add information for cpuid 0x8000001D leaf. Populate cache topology > information > > for different cache types(Data Cache, Instruction Cache, L2 and L3) > supported > > by 0x8000001D leaf. Please refer Processor Programming Reference (PPR) > for AMD > > Family 17h Model for more details. > > > > Signed-off-by: Babu Moger <babu.moger@amd.com> > > --- > > target/i386/cpu.c | 103 > ++++++++++++++++++++++++++++++++++++++++++++++++++++++ > > target/i386/kvm.c | 29 +++++++++++++-- > > 2 files changed, 129 insertions(+), 3 deletions(-) > > > > diff --git a/target/i386/cpu.c b/target/i386/cpu.c > > index d9773b6..1dd060a 100644 > > --- a/target/i386/cpu.c > > +++ b/target/i386/cpu.c > > @@ -336,6 +336,85 @@ static void > encode_cache_cpuid80000006(CPUCacheInfo *l2, > > } > > } > > > > The number of variables here is large, so maybe we should > document what each one mean so it's easier to review: > Sure. Will add more comments. > > > +/* Definitions used for building CPUID Leaf 0x8000001D and 0x8000001E */ > > +/* Please refer AMD64 Architecture Programmer’s Manual Volume 3 */ > > +#define MAX_CCX 2 > > CCX is "core complex", right? A comment would be useful here. Yes. It is core complex. Will add comments. > > > +#define MAX_CORES_IN_CCX 4 > > +#define MAX_NODES_EPYC 4 > > A comment explaining why it's OK to use a EPYC-specific constant > here would be useful. Sure. > > > > +#define MAX_CORES_IN_NODE 8 > > + > > +/* Number of logical processors sharing L3 cache */ > > +#define NUM_SHARING_CACHE(threads, num_sharing) ((threads > 1) ? > \ > > + (((num_sharing - 1) * threads) + 1) : \ > > + (num_sharing - 1)) > > This formula is confusing to me. If 4 cores are sharing the > cache and threads==1, 4 logical processors share the cache, and > we return 3. Sounds OK. > > But, if 4 cores are sharing the cache and threads==2, the number > of logical processors sharing the cache is 8. We should return > 7. The formula above returns (((4 - 1) * 2) + 1), which is > correct. > > But isn't it simpler to write this as: > > #define NUM_SHARING_CACHE(threads, num_sharing) \ > (((num_sharing) * (threads)) - 1) > > > (Maybe the "- 1" part could be moved outside the macro for > clarity. See below.) Yes, If we move -1 outside, then we could simplify it and we don’t need this macro. Will change it. > > > > +/* > > + * L3 Cache is shared between all the cores in a core complex. > > + * Maximum cores that can share L3 is 4. > > + */ > > +static int num_sharing_l3_cache(int nr_cores) > > Can we document what exactly this function is going to return? > This returns the number of cores sharing l3 cache, not the number > of logical processors, correct? Yes. It is the number of cores. Will fix it. > > > > +{ > > + int i, nodes = 1; > > + > > + /* Check if we can fit all the cores in one CCX */ > > + if (nr_cores <= MAX_CORES_IN_CCX) { > > + return nr_cores; > > + } > > + /* > > + * Figure out the number of nodes(or dies) required to build > > + * this config. Max cores in a node is 8 > > + */ > > + for (i = nodes; i <= MAX_NODES_EPYC; i++) { > > + if (nr_cores <= (i * MAX_CORES_IN_NODE)) { > > + nodes = i; > > + break; > > + } > > + /* We support nodes 1, 2, 4 */ > > + if (i == 3) { > > + continue; > > + } > > + } > > "continue" as the very last statement of a for loop does nothing, > so it looks like this could be written as: In real hardware number of nodes 3 is not a valid configuration. I was trying to avoid 3 there. Yes, we can achieve this with DIV_ROUND_UP like below. > > for (i = nodes; i <= MAX_NODES_EPYC; i++) { > if (nr_cores <= (i * MAX_CORES_IN_NODE)) { > nodes = i; > break; > } > } > > which in turn seems to be the same as: > > nodes = DIV_ROUND_UP(nr_cores, MAX_CORES_IN_NODE); > nodes = MIN(nodes, MAX_NODES_EPYC) > > But, is this really what we want here? Number of nodes supported is 1, 2 or 4. Hardware does not support 3. That is what I was trying to achieve there. DIV_ROUND_UP will work with check for 3. If it is 3 then make nodes = 4. Will change it. MIN(nodes, MAX_NODES_EPYC) is not required as I have added a check in patch 4/5 to check topology(function verify_topology). If we go beyond 4 nodes then I am disabling topoext feature. > > > > + /* Spread the cores accros all the CCXs and return max cores in a ccx */ > > + return (nr_cores / (nodes * MAX_CCX)) + > > + ((nr_cores % (nodes * MAX_CCX)) ? 1 : 0); > > This also seems to be the same as DIV_ROUND_UP? > > return DIV_ROUND_UP(nr_cores, nodes * MAX_CCX); > Yes. DIV_ROUND_UP will work. > I didn't confirm the logic is valid, though, because I don't know > what we should expect. What is the expected return value of this > function in the following cases? > > -smp 24,sockets=2,cores=12,threads=1 This should return 3(DIV_ROUND_UP(12, 2 * 2). We can fit in 2 nodes, with 4 core complexes. There will be 3 cores in each core complex. > -smp 64,sockets=2,cores=32,threads=1 This should return 4(DIV_ROUND_UP(32, 4 * 2).. We can fit it in 4 nodes with total 8 core complexes. There will be 4 cores in each core complex. > > > > +} > > + > > +/* Encode cache info for CPUID[8000001D] */ > > +static void encode_cache_cpuid8000001d(CPUCacheInfo *cache, > CPUState *cs, > > + uint32_t *eax, uint32_t *ebx, > > + uint32_t *ecx, uint32_t *edx) > > +{ > > + uint32_t num_share_l3; > > + assert(cache->size == cache->line_size * cache->associativity * > > + cache->partitions * cache->sets); > > + > > + *eax = CACHE_TYPE(cache->type) | CACHE_LEVEL(cache->level) | > > + (cache->self_init ? CACHE_SELF_INIT_LEVEL : 0); > > + > > + /* L3 is shared among multiple cores */ > > + if (cache->level == 3) { > > + num_share_l3 = num_sharing_l3_cache(cs->nr_cores); > > + *eax |= (NUM_SHARING_CACHE(cs->nr_threads, num_share_l3) << > 14); > > Considering that the line below has an explicit "- 1", I think > the "- 1" part could be moved outside the NUM_SHARING_CACHE > macro, and used explicitly here. > > But then the NUM_SHARING_CACHE would be just a simple > multiplication, so this could be simply written as: > > /* num_sharing_l3_cache() renamed to cores_sharing_l3_cache() */ > uint32_t l3_cores = cores_sharing_l3_cache(cs->nr_cores); > uint32_t l3_logical_processors = l3_cores * cs->nr_threads; > *eax |= (l3_logical_processors - 1) << 14; Yes. Will make these changes. > > > + } else { > > + *eax |= ((cs->nr_threads - 1) << 14); > > + } > > + > > + assert(cache->line_size > 0); > > + assert(cache->partitions > 0); > > + assert(cache->associativity > 0); > > + /* We don't implement fully-associative caches */ > > + assert(cache->associativity < cache->sets); > > + *ebx = (cache->line_size - 1) | > > + ((cache->partitions - 1) << 12) | > > + ((cache->associativity - 1) << 22); > > + > > + assert(cache->sets > 0); > > + *ecx = cache->sets - 1; > > + > > + *edx = (cache->no_invd_sharing ? CACHE_NO_INVD_SHARING : 0) | > > + (cache->inclusive ? CACHE_INCLUSIVE : 0) | > > + (cache->complex_indexing ? CACHE_COMPLEX_IDX : 0); > > +} > > + > > /* > > * Definitions of the hardcoded cache entries we expose: > > * These are legacy cache values. If there is a need to change any > > @@ -4005,6 +4084,30 @@ void cpu_x86_cpuid(CPUX86State *env, > uint32_t index, uint32_t count, > > *edx = 0; > > } > > break; > > + case 0x8000001D: > > + *eax = 0; > > + switch (count) { > > + case 0: /* L1 dcache info */ > > + encode_cache_cpuid8000001d(env->cache_info_amd.l1d_cache, > cs, > > + eax, ebx, ecx, edx); > > + break; > > + case 1: /* L1 icache info */ > > + encode_cache_cpuid8000001d(env->cache_info_amd.l1i_cache, cs, > > + eax, ebx, ecx, edx); > > + break; > > + case 2: /* L2 cache info */ > > + encode_cache_cpuid8000001d(env->cache_info_amd.l2_cache, cs, > > + eax, ebx, ecx, edx); > > + break; > > + case 3: /* L3 cache info */ > > + encode_cache_cpuid8000001d(env->cache_info_amd.l3_cache, cs, > > + eax, ebx, ecx, edx); > > + break; > > + default: /* end of info */ > > + *eax = *ebx = *ecx = *edx = 0; > > + break; > > + } > > + break; > > case 0xC0000000: > > *eax = env->cpuid_xlevel2; > > *ebx = 0; > > diff --git a/target/i386/kvm.c b/target/i386/kvm.c > > index d6666a4..a8bf7eb 100644 > > --- a/target/i386/kvm.c > > +++ b/target/i386/kvm.c > > @@ -979,9 +979,32 @@ int kvm_arch_init_vcpu(CPUState *cs) > > } > > c = &cpuid_data.entries[cpuid_i++]; > > > > - c->function = i; > > - c->flags = 0; > > - cpu_x86_cpuid(env, i, 0, &c->eax, &c->ebx, &c->ecx, &c->edx); > > + switch (i) { > > + case 0x8000001d: > > + /* Query for all AMD cache information leaves */ > > + for (j = 0; ; j++) { > > + c->function = i; > > + c->flags = KVM_CPUID_FLAG_SIGNIFCANT_INDEX; > > + c->index = j; > > + cpu_x86_cpuid(env, i, j, &c->eax, &c->ebx, &c->ecx, &c->edx); > > + > > + if (c->eax == 0) { > > + break; > > + } > > + if (cpuid_i == KVM_MAX_CPUID_ENTRIES) { > > + fprintf(stderr, "cpuid_data is full, no space for " > > + "cpuid(eax:0x%x,ecx:0x%x)\n", i, j); > > + abort(); > > + } > > + c = &cpuid_data.entries[cpuid_i++]; > > + } > > + break; > > + default: > > + c->function = i; > > + c->flags = 0; > > + cpu_x86_cpuid(env, i, 0, &c->eax, &c->ebx, &c->ecx, &c->edx); > > + break; > > + } > > } > > > > /* Call Centaur's CPUID instructions they are supported. */ > > -- > > 1.8.3.1 > > > > -- > Eduardo
diff --git a/target/i386/cpu.c b/target/i386/cpu.c index d9773b6..1dd060a 100644 --- a/target/i386/cpu.c +++ b/target/i386/cpu.c @@ -336,6 +336,85 @@ static void encode_cache_cpuid80000006(CPUCacheInfo *l2, } } +/* Definitions used for building CPUID Leaf 0x8000001D and 0x8000001E */ +/* Please refer AMD64 Architecture Programmer’s Manual Volume 3 */ +#define MAX_CCX 2 +#define MAX_CORES_IN_CCX 4 +#define MAX_NODES_EPYC 4 +#define MAX_CORES_IN_NODE 8 + +/* Number of logical processors sharing L3 cache */ +#define NUM_SHARING_CACHE(threads, num_sharing) ((threads > 1) ? \ + (((num_sharing - 1) * threads) + 1) : \ + (num_sharing - 1)) +/* + * L3 Cache is shared between all the cores in a core complex. + * Maximum cores that can share L3 is 4. + */ +static int num_sharing_l3_cache(int nr_cores) +{ + int i, nodes = 1; + + /* Check if we can fit all the cores in one CCX */ + if (nr_cores <= MAX_CORES_IN_CCX) { + return nr_cores; + } + /* + * Figure out the number of nodes(or dies) required to build + * this config. Max cores in a node is 8 + */ + for (i = nodes; i <= MAX_NODES_EPYC; i++) { + if (nr_cores <= (i * MAX_CORES_IN_NODE)) { + nodes = i; + break; + } + /* We support nodes 1, 2, 4 */ + if (i == 3) { + continue; + } + } + /* Spread the cores accros all the CCXs and return max cores in a ccx */ + return (nr_cores / (nodes * MAX_CCX)) + + ((nr_cores % (nodes * MAX_CCX)) ? 1 : 0); +} + +/* Encode cache info for CPUID[8000001D] */ +static void encode_cache_cpuid8000001d(CPUCacheInfo *cache, CPUState *cs, + uint32_t *eax, uint32_t *ebx, + uint32_t *ecx, uint32_t *edx) +{ + uint32_t num_share_l3; + assert(cache->size == cache->line_size * cache->associativity * + cache->partitions * cache->sets); + + *eax = CACHE_TYPE(cache->type) | CACHE_LEVEL(cache->level) | + (cache->self_init ? CACHE_SELF_INIT_LEVEL : 0); + + /* L3 is shared among multiple cores */ + if (cache->level == 3) { + num_share_l3 = num_sharing_l3_cache(cs->nr_cores); + *eax |= (NUM_SHARING_CACHE(cs->nr_threads, num_share_l3) << 14); + } else { + *eax |= ((cs->nr_threads - 1) << 14); + } + + assert(cache->line_size > 0); + assert(cache->partitions > 0); + assert(cache->associativity > 0); + /* We don't implement fully-associative caches */ + assert(cache->associativity < cache->sets); + *ebx = (cache->line_size - 1) | + ((cache->partitions - 1) << 12) | + ((cache->associativity - 1) << 22); + + assert(cache->sets > 0); + *ecx = cache->sets - 1; + + *edx = (cache->no_invd_sharing ? CACHE_NO_INVD_SHARING : 0) | + (cache->inclusive ? CACHE_INCLUSIVE : 0) | + (cache->complex_indexing ? CACHE_COMPLEX_IDX : 0); +} + /* * Definitions of the hardcoded cache entries we expose: * These are legacy cache values. If there is a need to change any @@ -4005,6 +4084,30 @@ void cpu_x86_cpuid(CPUX86State *env, uint32_t index, uint32_t count, *edx = 0; } break; + case 0x8000001D: + *eax = 0; + switch (count) { + case 0: /* L1 dcache info */ + encode_cache_cpuid8000001d(env->cache_info_amd.l1d_cache, cs, + eax, ebx, ecx, edx); + break; + case 1: /* L1 icache info */ + encode_cache_cpuid8000001d(env->cache_info_amd.l1i_cache, cs, + eax, ebx, ecx, edx); + break; + case 2: /* L2 cache info */ + encode_cache_cpuid8000001d(env->cache_info_amd.l2_cache, cs, + eax, ebx, ecx, edx); + break; + case 3: /* L3 cache info */ + encode_cache_cpuid8000001d(env->cache_info_amd.l3_cache, cs, + eax, ebx, ecx, edx); + break; + default: /* end of info */ + *eax = *ebx = *ecx = *edx = 0; + break; + } + break; case 0xC0000000: *eax = env->cpuid_xlevel2; *ebx = 0; diff --git a/target/i386/kvm.c b/target/i386/kvm.c index d6666a4..a8bf7eb 100644 --- a/target/i386/kvm.c +++ b/target/i386/kvm.c @@ -979,9 +979,32 @@ int kvm_arch_init_vcpu(CPUState *cs) } c = &cpuid_data.entries[cpuid_i++]; - c->function = i; - c->flags = 0; - cpu_x86_cpuid(env, i, 0, &c->eax, &c->ebx, &c->ecx, &c->edx); + switch (i) { + case 0x8000001d: + /* Query for all AMD cache information leaves */ + for (j = 0; ; j++) { + c->function = i; + c->flags = KVM_CPUID_FLAG_SIGNIFCANT_INDEX; + c->index = j; + cpu_x86_cpuid(env, i, j, &c->eax, &c->ebx, &c->ecx, &c->edx); + + if (c->eax == 0) { + break; + } + if (cpuid_i == KVM_MAX_CPUID_ENTRIES) { + fprintf(stderr, "cpuid_data is full, no space for " + "cpuid(eax:0x%x,ecx:0x%x)\n", i, j); + abort(); + } + c = &cpuid_data.entries[cpuid_i++]; + } + break; + default: + c->function = i; + c->flags = 0; + cpu_x86_cpuid(env, i, 0, &c->eax, &c->ebx, &c->ecx, &c->edx); + break; + } } /* Call Centaur's CPUID instructions they are supported. */
Add information for cpuid 0x8000001D leaf. Populate cache topology information for different cache types(Data Cache, Instruction Cache, L2 and L3) supported by 0x8000001D leaf. Please refer Processor Programming Reference (PPR) for AMD Family 17h Model for more details. Signed-off-by: Babu Moger <babu.moger@amd.com> --- target/i386/cpu.c | 103 ++++++++++++++++++++++++++++++++++++++++++++++++++++++ target/i386/kvm.c | 29 +++++++++++++-- 2 files changed, 129 insertions(+), 3 deletions(-)