Message ID | 20250225224559.1226079-1-andrew.cooper3@citrix.com (mailing list archive) |
---|---|
State | New |
Headers | show |
Series | x86/elf: Improve code generation in elf_core_save_regs() | expand |
On 25.02.2025 23:45, Andrew Cooper wrote: > A CALL with 0 displacement is handled specially, and is why this logic > functions even with CET Shadow Stacks active. Nevertheless a rip-relative LEA > is the more normal way of doing this in 64bit code. > > The retrieval of flags modifies the stack pointer so needs to state a > dependency on the stack pointer. Despite it's name, ASM_CALL_CONSTRAINT is > the way to do this. > > read_sreg() forces the answer through a register, causing code generation of > the form: > > mov %gs, %eax > mov %eax, %eax > mov %rax, 0x140(%rsi) > > Encode the reads directly with a memory operand. This results in a 16bit > store instead of an 64bit store, but the backing memory is zeroed. Raises the question whether we shouldn't change read_sreg(). At least the emulator uses of it would also benefit from storing straight to memory. And the remaining uses ought to be optimizable by the compiler, except that I don't expect we'd be able to express the zero-extending nature when the destination is a register. Or wait, maybe I can make up something (whether that's going to be liked is a separate question). Jan
On 26.02.2025 08:44, Jan Beulich wrote: > On 25.02.2025 23:45, Andrew Cooper wrote: >> A CALL with 0 displacement is handled specially, and is why this logic >> functions even with CET Shadow Stacks active. Nevertheless a rip-relative LEA >> is the more normal way of doing this in 64bit code. >> >> The retrieval of flags modifies the stack pointer so needs to state a >> dependency on the stack pointer. Despite it's name, ASM_CALL_CONSTRAINT is >> the way to do this. >> >> read_sreg() forces the answer through a register, causing code generation of >> the form: >> >> mov %gs, %eax >> mov %eax, %eax >> mov %rax, 0x140(%rsi) >> >> Encode the reads directly with a memory operand. This results in a 16bit >> store instead of an 64bit store, but the backing memory is zeroed. > > Raises the question whether we shouldn't change read_sreg(). At least the > emulator uses of it would also benefit from storing straight to memory. And > the remaining uses ought to be optimizable by the compiler, except that I > don't expect we'd be able to express the zero-extending nature when the > destination is a register. Or wait, maybe I can make up something (whether > that's going to be liked is a separate question). Here you go. Jan x86: make read_sreg() "bi-modal" Permit use sites to control whether to store directly to memory; right now both elf_core_save_regs() and the insn emulator's put_fpu() needlessly go through an intermediate GPR. Note that in both cases the apparent loss of zero-extension isn't a problem: The fields written to start out zero-initialized anyway. No change in generated code for the use sites not being touched. Signed-off-by: Jan Beulich <jbeulich@suse.com> --- Whether to make the change to put_fpu() is up for discussion: In my build it increases code size slightly, despite the reduction of number of insns emitted. An alternative (leaving the decision to the compiler) might be to drop the if() and use "=g" as constraint. I was considering to omit the assignment to sel_ on the if() branch, expecting the compiler to then flag uses of the return value (as consuming uninitialized data) when a 2nd argument is passed. However, gcc14 then already flags the "sel_;" at the end of the macro as consuming uninitialized data. --- a/xen/arch/x86/include/asm/regs.h +++ b/xen/arch/x86/include/asm/regs.h @@ -16,10 +16,20 @@ !diff || ((r)->cs != __HYPERVISOR_CS); \ }) -#define read_sreg(name) ({ \ - unsigned int __sel; \ - asm ( "mov %%" STR(name) ",%0" : "=r" (__sel) ); \ - __sel; \ +#define read_sreg(name, dst...) ({ \ + unsigned int sel_; \ + BUILD_BUG_ON(count_args(dst) > 1); \ + if ( count_args(dst) ) \ + { \ + typeof(LASTARG(&sel_, ## dst)) dst_ = \ + LASTARG(&sel_, ## dst); \ + asm ( "mov %%" STR(name) ",%0" : "=m" (*dst_) ); \ + /* The compiler ought to optimize this out. */ \ + sel_ = *dst_; \ + } \ + else \ + asm ( "mov %%" STR(name) ",%0" : "=r" (sel_) ); \ + sel_; \ }) static inline void read_sregs(struct cpu_user_regs *regs) --- a/xen/arch/x86/include/asm/x86_64/elf.h +++ b/xen/arch/x86/include/asm/x86_64/elf.h @@ -55,16 +55,16 @@ static inline void elf_core_save_regs(EL /* orig_rax not filled in for now */ asm ( "call 0f; 0: popq %0" : "=m" (core_regs->rip) ); - core_regs->cs = read_sreg(cs); + read_sreg(cs, &core_regs->cs); asm ( "pushfq; popq %0" : "=m" (core_regs->rflags) ); asm ( "movq %%rsp, %0" : "=m" (core_regs->rsp) ); - core_regs->ss = read_sreg(ss); + read_sreg(ss, &core_regs->ss); rdmsrl(MSR_FS_BASE, core_regs->thread_fs); rdmsrl(MSR_GS_BASE, core_regs->thread_gs); - core_regs->ds = read_sreg(ds); - core_regs->es = read_sreg(es); - core_regs->fs = read_sreg(fs); - core_regs->gs = read_sreg(gs); + read_sreg(ds, &core_regs->ds); + read_sreg(es, &core_regs->es); + read_sreg(fs, &core_regs->fs); + read_sreg(gs, &core_regs->gs); asm ( "mov %%cr0, %0" : "=r" (xen_core_regs->cr0) ); asm ( "mov %%cr2, %0" : "=r" (xen_core_regs->cr2) ); --- a/xen/arch/x86/x86_emulate/x86_emulate.c +++ b/xen/arch/x86/x86_emulate/x86_emulate.c @@ -465,10 +465,10 @@ static void put_fpu( else if ( is_pv_vcpu(current) ) switch ( state->ea.mem.seg ) { - case x86_seg_ds: aux.ds = read_sreg(ds); break; - case x86_seg_es: aux.ds = read_sreg(es); break; - case x86_seg_fs: aux.ds = read_sreg(fs); break; - case x86_seg_gs: aux.ds = read_sreg(gs); break; + case x86_seg_ds: read_sreg(ds, &aux.ds); break; + case x86_seg_es: read_sreg(es, &aux.ds); break; + case x86_seg_fs: read_sreg(fs, &aux.ds); break; + case x86_seg_gs: read_sreg(gs, &aux.ds); break; case x86_seg_ss: aux.ds = ctxt->regs->ss; break; default: ASSERT_UNREACHABLE(); break; }
diff --git a/xen/arch/x86/include/asm/x86_64/elf.h b/xen/arch/x86/include/asm/x86_64/elf.h index 00227e0e120c..f33be46ddec9 100644 --- a/xen/arch/x86/include/asm/x86_64/elf.h +++ b/xen/arch/x86/include/asm/x86_64/elf.h @@ -34,7 +34,7 @@ typedef struct { unsigned long gs; } ELF_Gregset; -static inline void elf_core_save_regs(ELF_Gregset *core_regs, +static inline void elf_core_save_regs(ELF_Gregset *core_regs, crash_xen_core_t *xen_core_regs) { asm ( "movq %%r15, %0" : "=m" (core_regs->r15) ); @@ -54,17 +54,17 @@ static inline void elf_core_save_regs(ELF_Gregset *core_regs, asm ( "movq %%rdi, %0" : "=m" (core_regs->rdi) ); /* orig_rax not filled in for now */ - asm ( "call 0f; 0: popq %0" : "=m" (core_regs->rip) ); - core_regs->cs = read_sreg(cs); - asm ( "pushfq; popq %0" : "=m" (core_regs->rflags) ); + asm ( "lea (%%rip), %0" : "=r" (core_regs->rip) ); + asm ( "mov %%cs, %0" : "=m" (core_regs->cs) ); + asm ( "pushfq; popq %0" : "=m" (core_regs->rflags) ASM_CALL_CONSTRAINT ); asm ( "movq %%rsp, %0" : "=m" (core_regs->rsp) ); - core_regs->ss = read_sreg(ss); + asm ( "mov %%ss, %0" : "=m" (core_regs->ss) ); rdmsrl(MSR_FS_BASE, core_regs->thread_fs); rdmsrl(MSR_GS_BASE, core_regs->thread_gs); - core_regs->ds = read_sreg(ds); - core_regs->es = read_sreg(es); - core_regs->fs = read_sreg(fs); - core_regs->gs = read_sreg(gs); + asm ( "mov %%ds, %0" : "=m" (core_regs->ds) ); + asm ( "mov %%es, %0" : "=m" (core_regs->es) ); + asm ( "mov %%fs, %0" : "=m" (core_regs->fs) ); + asm ( "mov %%gs, %0" : "=m" (core_regs->gs) ); asm ( "mov %%cr0, %0" : "=r" (xen_core_regs->cr0) ); asm ( "mov %%cr2, %0" : "=r" (xen_core_regs->cr2) );
A CALL with 0 displacement is handled specially, and is why this logic functions even with CET Shadow Stacks active. Nevertheless a rip-relative LEA is the more normal way of doing this in 64bit code. The retrieval of flags modifies the stack pointer so needs to state a dependency on the stack pointer. Despite it's name, ASM_CALL_CONSTRAINT is the way to do this. read_sreg() forces the answer through a register, causing code generation of the form: mov %gs, %eax mov %eax, %eax mov %rax, 0x140(%rsi) Encode the reads directly with a memory operand. This results in a 16bit store instead of an 64bit store, but the backing memory is zeroed. While cleaning this up, drop one piece of trailing whitespace. No functional change. Signed-off-by: Andrew Cooper <andrew.cooper3@citrix.com> --- CC: Jan Beulich <JBeulich@suse.com> CC: Roger Pau Monné <roger.pau@citrix.com> This is tidyup from the effort to drop the vm86 regs from cpu_user_regs, but is fairly unrelated to the rest of the work. --- xen/arch/x86/include/asm/x86_64/elf.h | 18 +++++++++--------- 1 file changed, 9 insertions(+), 9 deletions(-)