@@ -899,7 +899,7 @@ int LLVMFuzzerTestOneInput(const uint8_t
int rc;
/* Not part of the initializer, for old gcc to cope. */
- ctxt.cpu_policy = &cp;
+ ctxt.cpu_policy = &cpu_policy;
/* Reset all global state variables */
memset(&input, 0, sizeof(input));
@@ -888,7 +888,8 @@ static void zap_fpsel(unsigned int *env,
env[3] &= ~0xffff;
}
- if ( cp.x86_vendor != X86_VENDOR_AMD && cp.x86_vendor != X86_VENDOR_HYGON )
+ if ( cpu_policy.x86_vendor != X86_VENDOR_AMD &&
+ cpu_policy.x86_vendor != X86_VENDOR_HYGON )
return;
if ( is_32bit )
@@ -1022,7 +1023,7 @@ int main(int argc, char **argv)
ctxt.regs = ®s;
ctxt.force_writeback = 0;
- ctxt.cpu_policy = &cp;
+ ctxt.cpu_policy = &cpu_policy;
ctxt.lma = sizeof(void *) == 8;
ctxt.addr_size = 8 * sizeof(void *);
ctxt.sp_size = 8 * sizeof(void *);
@@ -1540,7 +1541,7 @@ int main(int argc, char **argv)
regs.edx = (unsigned long)res;
res[0] = 0x00004444;
res[1] = 0x8888cccc;
- i = cp.extd.nscb; cp.extd.nscb = true; /* for AMD */
+ i = cpu_policy.extd.nscb; cpu_policy.extd.nscb = true; /* for AMD */
rc = x86_emulate(&ctxt, &emulops);
if ( (rc != X86EMUL_OKAY) ||
(regs.eip != (unsigned long)&instr[5]) ||
@@ -1549,7 +1550,7 @@ int main(int argc, char **argv)
goto fail;
printf("okay\n");
- cp.extd.nscb = i;
+ cpu_policy.extd.nscb = i;
emulops.write_segment = NULL;
printf("%-40s", "Testing rdmsrlist...");
@@ -1803,11 +1804,11 @@ int main(int argc, char **argv)
goto fail;
printf("okay\n");
- vendor_native = cp.x86_vendor;
- for ( cp.x86_vendor = X86_VENDOR_AMD; ; )
+ vendor_native = cpu_policy.x86_vendor;
+ for ( cpu_policy.x86_vendor = X86_VENDOR_AMD; ; )
{
- unsigned int v = cp.x86_vendor == X86_VENDOR_INTEL;
- const char *vendor = cp.x86_vendor == X86_VENDOR_INTEL ? "Intel" : "AMD";
+ unsigned int v = cpu_policy.x86_vendor == X86_VENDOR_INTEL;
+ const char *vendor = cpu_policy.x86_vendor == X86_VENDOR_INTEL ? "Intel" : "AMD";
uint64_t *stk = (void *)res + MMAP_SZ - 16;
regs.rcx = 2;
@@ -1843,11 +1844,11 @@ int main(int argc, char **argv)
printf("okay\n");
}
- if ( cp.x86_vendor == X86_VENDOR_INTEL )
+ if ( cpu_policy.x86_vendor == X86_VENDOR_INTEL )
break;
- cp.x86_vendor = X86_VENDOR_INTEL;
+ cpu_policy.x86_vendor = X86_VENDOR_INTEL;
}
- cp.x86_vendor = vendor_native;
+ cpu_policy.x86_vendor = vendor_native;
#endif /* x86-64 */
printf("%-40s", "Testing shld $1,%ecx,(%edx)...");
@@ -25,7 +25,7 @@
#endif
uint32_t mxcsr_mask = 0x0000ffbf;
-struct cpu_policy cp;
+struct cpu_policy cpu_policy;
static char fpu_save_area[0x4000] __attribute__((__aligned__((64))));
static bool use_xsave;
@@ -75,24 +75,24 @@ bool emul_test_init(void)
unsigned long sp;
- x86_cpu_policy_fill_native(&cp);
- x86_cpu_policy_bound_max_leaves(&cp);
+ x86_cpu_policy_fill_native(&cpu_policy);
+ x86_cpu_policy_bound_max_leaves(&cpu_policy);
/*
* The emulator doesn't use these instructions, so can always emulate
* them.
*/
- cp.basic.movbe = true;
- cp.feat.invpcid = true;
- cp.feat.adx = true;
- cp.feat.avx512pf = cp.feat.avx512f;
- cp.feat.rdpid = true;
- cp.feat.lkgs = true;
- cp.feat.wrmsrns = true;
- cp.feat.msrlist = true;
- cp.extd.clzero = true;
+ cpu_policy.basic.movbe = true;
+ cpu_policy.feat.invpcid = true;
+ cpu_policy.feat.adx = true;
+ cpu_policy.feat.avx512pf = cpu_policy.feat.avx512f;
+ cpu_policy.feat.rdpid = true;
+ cpu_policy.feat.lkgs = true;
+ cpu_policy.feat.wrmsrns = true;
+ cpu_policy.feat.msrlist = true;
+ cpu_policy.extd.clzero = true;
- x86_cpu_policy_shrink_max_leaves(&cp);
+ x86_cpu_policy_shrink_max_leaves(&cpu_policy);
if ( cpu_has_xsave )
{
@@ -69,7 +69,7 @@
#define is_canonical_address(x) (((int64_t)(x) >> 47) == ((int64_t)(x) >> 63))
extern uint32_t mxcsr_mask;
-extern struct cpu_policy cp;
+extern struct cpu_policy cpu_policy;
#define MMAP_SZ 16384
bool emul_test_init(void);
@@ -123,7 +123,7 @@ static inline uint64_t xgetbv(uint32_t x
}
/* Intentionally checking OSXSAVE here. */
-#define cpu_has_xsave (cp.basic.raw[1].c & (1u << 27))
+#define cpu_has_xsave (cpu_policy.basic.raw[1].c & (1u << 27))
static inline bool xcr0_mask(uint64_t mask)
{
@@ -133,67 +133,67 @@ static inline bool xcr0_mask(uint64_t ma
unsigned int rdpkru(void);
void wrpkru(unsigned int val);
-#define cache_line_size() (cp.basic.clflush_size * 8)
-#define cpu_has_fpu cp.basic.fpu
-#define cpu_has_mmx cp.basic.mmx
-#define cpu_has_fxsr cp.basic.fxsr
-#define cpu_has_sse cp.basic.sse
-#define cpu_has_sse2 cp.basic.sse2
-#define cpu_has_sse3 cp.basic.sse3
-#define cpu_has_pclmulqdq cp.basic.pclmulqdq
-#define cpu_has_ssse3 cp.basic.ssse3
-#define cpu_has_fma (cp.basic.fma && xcr0_mask(6))
-#define cpu_has_sse4_1 cp.basic.sse4_1
-#define cpu_has_sse4_2 cp.basic.sse4_2
-#define cpu_has_popcnt cp.basic.popcnt
-#define cpu_has_aesni cp.basic.aesni
-#define cpu_has_avx (cp.basic.avx && xcr0_mask(6))
-#define cpu_has_f16c (cp.basic.f16c && xcr0_mask(6))
-
-#define cpu_has_avx2 (cp.feat.avx2 && xcr0_mask(6))
-#define cpu_has_bmi1 cp.feat.bmi1
-#define cpu_has_bmi2 cp.feat.bmi2
-#define cpu_has_avx512f (cp.feat.avx512f && xcr0_mask(0xe6))
-#define cpu_has_avx512dq (cp.feat.avx512dq && xcr0_mask(0xe6))
-#define cpu_has_avx512_ifma (cp.feat.avx512_ifma && xcr0_mask(0xe6))
-#define cpu_has_avx512er (cp.feat.avx512er && xcr0_mask(0xe6))
-#define cpu_has_avx512cd (cp.feat.avx512cd && xcr0_mask(0xe6))
-#define cpu_has_sha cp.feat.sha
-#define cpu_has_avx512bw (cp.feat.avx512bw && xcr0_mask(0xe6))
-#define cpu_has_avx512vl (cp.feat.avx512vl && xcr0_mask(0xe6))
-#define cpu_has_avx512_vbmi (cp.feat.avx512_vbmi && xcr0_mask(0xe6))
-#define cpu_has_avx512_vbmi2 (cp.feat.avx512_vbmi2 && xcr0_mask(0xe6))
-#define cpu_has_gfni cp.feat.gfni
-#define cpu_has_vaes (cp.feat.vaes && xcr0_mask(6))
-#define cpu_has_vpclmulqdq (cp.feat.vpclmulqdq && xcr0_mask(6))
-#define cpu_has_avx512_vnni (cp.feat.avx512_vnni && xcr0_mask(0xe6))
-#define cpu_has_avx512_bitalg (cp.feat.avx512_bitalg && xcr0_mask(0xe6))
-#define cpu_has_avx512_vpopcntdq (cp.feat.avx512_vpopcntdq && xcr0_mask(0xe6))
-#define cpu_has_movdiri cp.feat.movdiri
-#define cpu_has_movdir64b cp.feat.movdir64b
-#define cpu_has_avx512_4vnniw (cp.feat.avx512_4vnniw && xcr0_mask(0xe6))
-#define cpu_has_avx512_4fmaps (cp.feat.avx512_4fmaps && xcr0_mask(0xe6))
-#define cpu_has_avx512_vp2intersect (cp.feat.avx512_vp2intersect && xcr0_mask(0xe6))
-#define cpu_has_serialize cp.feat.serialize
-#define cpu_has_avx512_fp16 (cp.feat.avx512_fp16 && xcr0_mask(0xe6))
-#define cpu_has_sha512 (cp.feat.sha512 && xcr0_mask(6))
-#define cpu_has_sm3 (cp.feat.sm3 && xcr0_mask(6))
-#define cpu_has_sm4 (cp.feat.sm4 && xcr0_mask(6))
-#define cpu_has_avx_vnni (cp.feat.avx_vnni && xcr0_mask(6))
-#define cpu_has_avx512_bf16 (cp.feat.avx512_bf16 && xcr0_mask(0xe6))
-#define cpu_has_cmpccxadd cp.feat.cmpccxadd
-#define cpu_has_avx_ifma (cp.feat.avx_ifma && xcr0_mask(6))
-#define cpu_has_avx_vnni_int8 (cp.feat.avx_vnni_int8 && xcr0_mask(6))
-#define cpu_has_avx_ne_convert (cp.feat.avx_ne_convert && xcr0_mask(6))
-#define cpu_has_avx_vnni_int16 (cp.feat.avx_vnni_int16 && xcr0_mask(6))
-
-#define cpu_has_xgetbv1 (cpu_has_xsave && cp.xstate.xgetbv1)
-
-#define cpu_has_3dnow_ext cp.extd._3dnowext
-#define cpu_has_sse4a cp.extd.sse4a
-#define cpu_has_xop (cp.extd.xop && xcr0_mask(6))
-#define cpu_has_fma4 (cp.extd.fma4 && xcr0_mask(6))
-#define cpu_has_tbm cp.extd.tbm
+#define cache_line_size() (cpu_policy.basic.clflush_size * 8)
+#define cpu_has_fpu cpu_policy.basic.fpu
+#define cpu_has_mmx cpu_policy.basic.mmx
+#define cpu_has_fxsr cpu_policy.basic.fxsr
+#define cpu_has_sse cpu_policy.basic.sse
+#define cpu_has_sse2 cpu_policy.basic.sse2
+#define cpu_has_sse3 cpu_policy.basic.sse3
+#define cpu_has_pclmulqdq cpu_policy.basic.pclmulqdq
+#define cpu_has_ssse3 cpu_policy.basic.ssse3
+#define cpu_has_fma (cpu_policy.basic.fma && xcr0_mask(6))
+#define cpu_has_sse4_1 cpu_policy.basic.sse4_1
+#define cpu_has_sse4_2 cpu_policy.basic.sse4_2
+#define cpu_has_popcnt cpu_policy.basic.popcnt
+#define cpu_has_aesni cpu_policy.basic.aesni
+#define cpu_has_avx (cpu_policy.basic.avx && xcr0_mask(6))
+#define cpu_has_f16c (cpu_policy.basic.f16c && xcr0_mask(6))
+
+#define cpu_has_avx2 (cpu_policy.feat.avx2 && xcr0_mask(6))
+#define cpu_has_bmi1 cpu_policy.feat.bmi1
+#define cpu_has_bmi2 cpu_policy.feat.bmi2
+#define cpu_has_avx512f (cpu_policy.feat.avx512f && xcr0_mask(0xe6))
+#define cpu_has_avx512dq (cpu_policy.feat.avx512dq && xcr0_mask(0xe6))
+#define cpu_has_avx512_ifma (cpu_policy.feat.avx512_ifma && xcr0_mask(0xe6))
+#define cpu_has_avx512er (cpu_policy.feat.avx512er && xcr0_mask(0xe6))
+#define cpu_has_avx512cd (cpu_policy.feat.avx512cd && xcr0_mask(0xe6))
+#define cpu_has_sha cpu_policy.feat.sha
+#define cpu_has_avx512bw (cpu_policy.feat.avx512bw && xcr0_mask(0xe6))
+#define cpu_has_avx512vl (cpu_policy.feat.avx512vl && xcr0_mask(0xe6))
+#define cpu_has_avx512_vbmi (cpu_policy.feat.avx512_vbmi && xcr0_mask(0xe6))
+#define cpu_has_avx512_vbmi2 (cpu_policy.feat.avx512_vbmi2 && xcr0_mask(0xe6))
+#define cpu_has_gfni cpu_policy.feat.gfni
+#define cpu_has_vaes (cpu_policy.feat.vaes && xcr0_mask(6))
+#define cpu_has_vpclmulqdq (cpu_policy.feat.vpclmulqdq && xcr0_mask(6))
+#define cpu_has_avx512_vnni (cpu_policy.feat.avx512_vnni && xcr0_mask(0xe6))
+#define cpu_has_avx512_bitalg (cpu_policy.feat.avx512_bitalg && xcr0_mask(0xe6))
+#define cpu_has_avx512_vpopcntdq (cpu_policy.feat.avx512_vpopcntdq && xcr0_mask(0xe6))
+#define cpu_has_movdiri cpu_policy.feat.movdiri
+#define cpu_has_movdir64b cpu_policy.feat.movdir64b
+#define cpu_has_avx512_4vnniw (cpu_policy.feat.avx512_4vnniw && xcr0_mask(0xe6))
+#define cpu_has_avx512_4fmaps (cpu_policy.feat.avx512_4fmaps && xcr0_mask(0xe6))
+#define cpu_has_avx512_vp2intersect (cpu_policy.feat.avx512_vp2intersect && xcr0_mask(0xe6))
+#define cpu_has_serialize cpu_policy.feat.serialize
+#define cpu_has_avx512_fp16 (cpu_policy.feat.avx512_fp16 && xcr0_mask(0xe6))
+#define cpu_has_sha512 (cpu_policy.feat.sha512 && xcr0_mask(6))
+#define cpu_has_sm3 (cpu_policy.feat.sm3 && xcr0_mask(6))
+#define cpu_has_sm4 (cpu_policy.feat.sm4 && xcr0_mask(6))
+#define cpu_has_avx_vnni (cpu_policy.feat.avx_vnni && xcr0_mask(6))
+#define cpu_has_avx512_bf16 (cpu_policy.feat.avx512_bf16 && xcr0_mask(0xe6))
+#define cpu_has_cmpccxadd cpu_policy.feat.cmpccxadd
+#define cpu_has_avx_ifma (cpu_policy.feat.avx_ifma && xcr0_mask(6))
+#define cpu_has_avx_vnni_int8 (cpu_policy.feat.avx_vnni_int8 && xcr0_mask(6))
+#define cpu_has_avx_ne_convert (cpu_policy.feat.avx_ne_convert && xcr0_mask(6))
+#define cpu_has_avx_vnni_int16 (cpu_policy.feat.avx_vnni_int16 && xcr0_mask(6))
+
+#define cpu_has_xgetbv1 (cpu_has_xsave && cpu_policy.xstate.xgetbv1)
+
+#define cpu_has_3dnow_ext cpu_policy.extd._3dnowext
+#define cpu_has_sse4a cpu_policy.extd.sse4a
+#define cpu_has_xop (cpu_policy.extd.xop && xcr0_mask(6))
+#define cpu_has_fma4 (cpu_policy.extd.fma4 && xcr0_mask(6))
+#define cpu_has_tbm cpu_policy.extd.tbm
int emul_test_cpuid(
uint32_t leaf,
In preparation of introducing a const struct cpu_policy * local in x86_emulate(), rename that global variable to something more suitable: "cp" is our commonly used name for function parameters or local variables of type struct cpu_policy *, and the present name of the global could hence have interfered already. Signed-off-by: Jan Beulich <jbeulich@suse.com>