All of lore.kernel.org
 help / color / mirror / Atom feed
From: Sean Christopherson <seanjc@google.com>
To: Paolo Bonzini <pbonzini@redhat.com>
Cc: linux-kernel@vger.kernel.org, kvm@vger.kernel.org,
	 chang.seok.bae@intel.com, x86@kernel.org
Subject: Re: [PATCH 1/3] KVM: VMX: Macrofy GPR swapping in __vmx_vcpu_run()
Date: Thu, 14 May 2026 12:30:32 -0700	[thread overview]
Message-ID: <agYi2Lc7UbGiw_zL@google.com> (raw)
In-Reply-To: <20260513174948.726812-2-pbonzini@redhat.com>

On Wed, May 13, 2026, Paolo Bonzini wrote:
> From: "Chang S. Bae" <chang.seok.bae@intel.com>
> 
> Convert the repeated register save/restore sequences into macros to
> simplify the VM entry code. 

This is not simpler.  Maybe it ends up being less verbose, but I don't see how
anyone can claim this is simpler.  E.g. for people like me that aren't already
familiar with the .ifc directive, the R32_NUM macro is inscrutable.  I can at
least suss out what e.g. LOAD_REGS and friends are doing, but claiming the code
is "simpler" is rather ridiculous.

> This can also make it easier to extend for additional registers.

For me, this needs to be the focal point of the changelog, and it needs to be
written with --verbose, because the impact of APX on the VM-Entry/VM-Exit code
isn't so obvious that a one-line "this makes future life easier" sufficiently
justifies the macro magic.  And that's coming from someone that generally loves
macro magic :-)
 
> diff --git a/arch/x86/include/asm/kvm_vcpu_regs.h b/arch/x86/include/asm/kvm_vcpu_regs.h
> index 1af2cb59233b..7c12a1fd1ffc 100644
> --- a/arch/x86/include/asm/kvm_vcpu_regs.h
> +++ b/arch/x86/include/asm/kvm_vcpu_regs.h
> @@ -22,4 +22,126 @@
>  #define __VCPU_REGS_R15 15
>  #endif
>  
> +#ifdef __ASSEMBLER__
> +
> +#define REG_NUM_INVALID		100
> +
> +	.macro R32_NUM opd r32
> +	\opd = REG_NUM_INVALID
> +	.ifc \r32,%eax
> +	\opd = __VCPU_REGS_EAX
> +	.endif
> +	.ifc \r32,%ecx
> +	\opd = __VCPU_REGS_ECX
> +	.endif
> +	.ifc \r32,%edx
> +	\opd = __VCPU_REGS_EDX
> +	.endif
> +	.ifc \r32,%ebx
> +	\opd = __VCPU_REGS_EBX
> +	.endif
> +	.ifc \r32,%esp
> +	\opd = __VCPU_REGS_ESP
> +	.endif
> +	.ifc \r32,%ebp
> +	\opd = __VCPU_REGS_EBP
> +	.endif
> +	.ifc \r32,%esi
> +	\opd = __VCPU_REGS_ESI
> +	.endif
> +	.ifc \r32,%edi
> +	\opd = __VCPU_REGS_EDI
> +	.endif
> +#ifdef CONFIG_X86_64
> +	.ifc \r32,%r8d
> +	\opd = 8
> +	.endif
> +	.ifc \r32,%r9d
> +	\opd = 9
> +	.endif
> +	.ifc \r32,%r10d
> +	\opd = 10
> +	.endif
> +	.ifc \r32,%r11d
> +	\opd = 11
> +	.endif
> +	.ifc \r32,%r12d
> +	\opd = 12
> +	.endif
> +	.ifc \r32,%r13d
> +	\opd = 13
> +	.endif
> +	.ifc \r32,%r14d
> +	\opd = 14
> +	.endif
> +	.ifc \r32,%r15d
> +	\opd = 15
> +	.endif
> +#endif
> +	.endm
> +
> +	.macro R64_NUM opd r64
> +	\opd = REG_NUM_INVALID
> +#ifdef CONFIG_X86_64
> +	.ifc \r64,%rax
> +	\opd = __VCPU_REGS_RAX
> +	.endif
> +	.ifc \r64,%rcx
> +	\opd = __VCPU_REGS_RCX
> +	.endif
> +	.ifc \r64,%rdx
> +	\opd = __VCPU_REGS_RDX
> +	.endif
> +	.ifc \r64,%rbx
> +	\opd = __VCPU_REGS_RBX
> +	.endif
> +	.ifc \r64,%rsp
> +	\opd = __VCPU_REGS_RSP
> +	.endif
> +	.ifc \r64,%rbp
> +	\opd = __VCPU_REGS_RBP
> +	.endif
> +	.ifc \r64,%rsi
> +	\opd = __VCPU_REGS_RSI
> +	.endif
> +	.ifc \r64,%rdi
> +	\opd = __VCPU_REGS_RDI
> +	.endif
> +	.ifc \r64,%r8
> +	\opd = 8
> +	.endif
> +	.ifc \r64,%r9
> +	\opd = 9
> +	.endif
> +	.ifc \r64,%r10
> +	\opd = 10
> +	.endif
> +	.ifc \r64,%r11
> +	\opd = 11
> +	.endif
> +	.ifc \r64,%r12
> +	\opd = 12
> +	.endif
> +	.ifc \r64,%r13
> +	\opd = 13
> +	.endif
> +	.ifc \r64,%r14
> +	\opd = 14
> +	.endif
> +	.ifc \r64,%r15
> +	\opd = 15
> +	.endif
> +#endif
> +	.endm
> +
> +.macro REG_NUM reg_num reg
> +#ifdef CONFIG_X86_64
> +	R64_NUM \reg_num \reg
> +#else
> +	R32_NUM \reg_num \reg
> +#endif
> +.endm
> +
> +#endif
> +
>  #endif /* _ASM_X86_KVM_VCPU_REGS_H */
> diff --git a/arch/x86/kvm/svm/vmenter.S b/arch/x86/kvm/svm/vmenter.S
> index f523d9e49839..6a91e1383e8f 100644
> --- a/arch/x86/kvm/svm/vmenter.S
> +++ b/arch/x86/kvm/svm/vmenter.S
> @@ -4,13 +4,10 @@
>  #include <asm/asm-offsets.h>
>  #include <asm/bitsperlong.h>
>  #include <asm/frame.h>
> -#include <asm/kvm_vcpu_regs.h>
>  #include <asm/nospec-branch.h>
>  #include "kvm-asm-offsets.h"
>  #include "vmenter.h"
>  
> -#define WORD_SIZE (BITS_PER_LONG / 8)
> -
>  /* Intentionally omit RAX as it's context switched by hardware */
>  #define VCPU_RCX	(SVM_vcpu_arch_regs + __VCPU_REGS_RCX * WORD_SIZE)
>  #define VCPU_RDX	(SVM_vcpu_arch_regs + __VCPU_REGS_RDX * WORD_SIZE)
> diff --git a/arch/x86/kvm/vmenter.h b/arch/x86/kvm/vmenter.h
> index ba3f71449c62..44574806cb02 100644
> --- a/arch/x86/kvm/vmenter.h
> +++ b/arch/x86/kvm/vmenter.h
> @@ -2,6 +2,8 @@
>  #ifndef __KVM_X86_VMENTER_H
>  #define __KVM_X86_VMENTER_H
>  
> +#include <asm/kvm_vcpu_regs.h>
> +
>  #define KVM_ENTER_VMRESUME			BIT(0)
>  #define KVM_ENTER_SAVE_SPEC_CTRL		BIT(1)
>  #define KVM_ENTER_CLEAR_CPU_BUFFERS_FOR_MMIO	BIT(2)
> @@ -76,5 +78,46 @@
>  	wrmsr
>  .endm
>  
> +#define WORD_SIZE (BITS_PER_LONG / 8)
> +
> +.macro LOAD_REGS src:req, regs_ofs:req, regs:vararg
> +.irp reg, \regs
> +	REG_NUM reg_num \reg
> +	.if reg_num <> REG_NUM_INVALID
> +	mov (\regs_ofs + reg_num * WORD_SIZE)(\src), \reg
> +	.else
> +	.err invalid register \reg
> +	.endif
> +.endr
> +.endm
> +
> +.macro STORE_REGS dst:req, regs_ofs:req, regs:vararg
> +.irp reg, \regs
> +	REG_NUM reg_num \reg
> +	.if reg_num <> REG_NUM_INVALID
> +	mov \reg, (\regs_ofs + reg_num * WORD_SIZE)(\dst)
> +	.else
> +	.err invalid register \reg
> +	.endif
> +.endr
> +.endm
> +
> +.macro POP_REGS dst:req, regs_ofs:req, regs:vararg
> +.irp reg, \regs
> +	REG_NUM reg_num \reg
> +	.if reg_num <> REG_NUM_INVALID
> +	pop (\regs_ofs + reg_num * WORD_SIZE)(\dst)
> +	.else
> +	.err invalid register \reg
> +	.endif
> +.endr
> +.endm
> +
> +.macro CLEAR_REGS regs:vararg
> +.irp reg, \regs
> +	xorl \reg, \reg
> +.endr
> +.endm
> +
>  #endif /* __ASSEMBLER__ */
>  #endif /* __KVM_X86_VMENTER_H */
> diff --git a/arch/x86/kvm/vmx/vmenter.S b/arch/x86/kvm/vmx/vmenter.S
> index 7e4dc17fc0b8..4b7aaa7430fb 100644
> --- a/arch/x86/kvm/vmx/vmenter.S
> +++ b/arch/x86/kvm/vmx/vmenter.S
> @@ -2,35 +2,12 @@
>  #include <linux/linkage.h>
>  #include <asm/asm.h>
>  #include <asm/bitsperlong.h>
> -#include <asm/kvm_vcpu_regs.h>
>  #include <asm/nospec-branch.h>
>  #include <asm/percpu.h>
>  #include <asm/segment.h>
>  #include "kvm-asm-offsets.h"
>  #include "vmenter.h"
>  
> -#define WORD_SIZE (BITS_PER_LONG / 8)
> -
> -#define VCPU_RAX	(VMX_vcpu_arch_regs + __VCPU_REGS_RAX * WORD_SIZE)
> -#define VCPU_RCX	(VMX_vcpu_arch_regs + __VCPU_REGS_RCX * WORD_SIZE)
> -#define VCPU_RDX	(VMX_vcpu_arch_regs + __VCPU_REGS_RDX * WORD_SIZE)
> -#define VCPU_RBX	(VMX_vcpu_arch_regs + __VCPU_REGS_RBX * WORD_SIZE)
> -/* Intentionally omit RSP as it's context switched by hardware */
> -#define VCPU_RBP	(VMX_vcpu_arch_regs + __VCPU_REGS_RBP * WORD_SIZE)
> -#define VCPU_RSI	(VMX_vcpu_arch_regs + __VCPU_REGS_RSI * WORD_SIZE)
> -#define VCPU_RDI	(VMX_vcpu_arch_regs + __VCPU_REGS_RDI * WORD_SIZE)
> -
> -#ifdef CONFIG_X86_64
> -#define VCPU_R8		(VMX_vcpu_arch_regs + __VCPU_REGS_R8  * WORD_SIZE)
> -#define VCPU_R9		(VMX_vcpu_arch_regs + __VCPU_REGS_R9  * WORD_SIZE)
> -#define VCPU_R10	(VMX_vcpu_arch_regs + __VCPU_REGS_R10 * WORD_SIZE)
> -#define VCPU_R11	(VMX_vcpu_arch_regs + __VCPU_REGS_R11 * WORD_SIZE)
> -#define VCPU_R12	(VMX_vcpu_arch_regs + __VCPU_REGS_R12 * WORD_SIZE)
> -#define VCPU_R13	(VMX_vcpu_arch_regs + __VCPU_REGS_R13 * WORD_SIZE)
> -#define VCPU_R14	(VMX_vcpu_arch_regs + __VCPU_REGS_R14 * WORD_SIZE)
> -#define VCPU_R15	(VMX_vcpu_arch_regs + __VCPU_REGS_R15 * WORD_SIZE)
> -#endif
> -
>  .macro VMX_DO_EVENT_IRQOFF call_insn call_target
>  	/*
>  	 * Unconditionally create a stack frame, getting the correct RSP on the
> @@ -114,25 +91,18 @@ SYM_FUNC_START(__vmx_vcpu_run)
>  	 * an LFENCE to stop speculation from skipping the wrmsr.
>  	 */
>  
> -	/* Load guest registers.  Don't clobber flags. */
> -	mov VCPU_RAX(%_ASM_DI), %_ASM_AX
> -	mov VCPU_RCX(%_ASM_DI), %_ASM_CX
> -	mov VCPU_RDX(%_ASM_DI), %_ASM_DX
> -	mov VCPU_RBX(%_ASM_DI), %_ASM_BX
> -	mov VCPU_RBP(%_ASM_DI), %_ASM_BP
> -	mov VCPU_RSI(%_ASM_DI), %_ASM_SI
> +	/*
> +	 * Load guest registers.  Don't clobber flags. Intentionally omit
> +	 * %_ASM_SP as it's context switched by hardware
> +	 */
> +	LOAD_REGS %_ASM_DI, VMX_vcpu_arch_regs, \
> +		  %_ASM_AX, %_ASM_CX, %_ASM_DX, %_ASM_BX, %_ASM_BP, %_ASM_SI
>  #ifdef CONFIG_X86_64
> -	mov VCPU_R8 (%_ASM_DI),  %r8
> -	mov VCPU_R9 (%_ASM_DI),  %r9
> -	mov VCPU_R10(%_ASM_DI), %r10
> -	mov VCPU_R11(%_ASM_DI), %r11
> -	mov VCPU_R12(%_ASM_DI), %r12
> -	mov VCPU_R13(%_ASM_DI), %r13
> -	mov VCPU_R14(%_ASM_DI), %r14
> -	mov VCPU_R15(%_ASM_DI), %r15
> +	LOAD_REGS %_ASM_DI, VMX_vcpu_arch_regs, \
> +		  %r8, %r9, %r10, %r11, %r12, %r13, %r14, %r15
>  #endif
>  	/* Load guest RDI.  This kills the @vmx pointer! */
> -	mov VCPU_RDI(%_ASM_DI), %_ASM_DI
> +	LOAD_REGS %_ASM_DI, VMX_vcpu_arch_regs, %_ASM_DI
>  
>  	/*
>  	 * Note, ALTERNATIVE_2 works in reverse order.  If CLEAR_CPU_BUF_VM is
> @@ -187,23 +157,16 @@ SYM_INNER_LABEL_ALIGN(vmx_vmexit, SYM_L_GLOBAL)
>  	/* Reload @vmx to RDI. */
>  	mov 2*WORD_SIZE(%_ASM_SP), %_ASM_DI
>  
> -	/* Save all guest registers, including RDI from the stack */
> -	mov %_ASM_AX, VCPU_RAX(%_ASM_DI)
> -	mov %_ASM_CX, VCPU_RCX(%_ASM_DI)
> -	mov %_ASM_DX, VCPU_RDX(%_ASM_DI)
> -	mov %_ASM_BX, VCPU_RBX(%_ASM_DI)
> -	mov %_ASM_BP, VCPU_RBP(%_ASM_DI)
> -	mov %_ASM_SI, VCPU_RSI(%_ASM_DI)
> -	pop           VCPU_RDI(%_ASM_DI)
> +	/*
> +	 * Save all guest registers, including RDI from the stack. Intentionally
> +	 * omit %_ASM_SP as it's context switched by hardware
> +	 */
> +	STORE_REGS %_ASM_DI, VMX_vcpu_arch_regs, \
> +		   %_ASM_AX, %_ASM_CX, %_ASM_DX, %_ASM_BX, %_ASM_BP, %_ASM_SI
> +	POP_REGS   %_ASM_DI, VMX_vcpu_arch_regs, %_ASM_DI
>  #ifdef CONFIG_X86_64
> -	mov %r8,  VCPU_R8 (%_ASM_DI)
> -	mov %r9,  VCPU_R9 (%_ASM_DI)
> -	mov %r10, VCPU_R10(%_ASM_DI)
> -	mov %r11, VCPU_R11(%_ASM_DI)
> -	mov %r12, VCPU_R12(%_ASM_DI)
> -	mov %r13, VCPU_R13(%_ASM_DI)
> -	mov %r14, VCPU_R14(%_ASM_DI)
> -	mov %r15, VCPU_R15(%_ASM_DI)
> +	STORE_REGS %_ASM_DI, VMX_vcpu_arch_regs, \
> +		   %r8, %r9, %r10, %r11, %r12, %r13, %r14, %r15
>  #endif
>  
>  	/* Clear return value to indicate VM-Exit (as opposed to VM-Fail). */
> @@ -220,21 +183,9 @@ SYM_INNER_LABEL_ALIGN(vmx_vmexit, SYM_L_GLOBAL)
>  	 * VM-Exit and RBX is explicitly loaded with 0 or 1 to hold the return
>  	 * value.
>  	 */
> -	xor %eax, %eax
> -	xor %ecx, %ecx
> -	xor %edx, %edx
> -	xor %ebp, %ebp
> -	xor %esi, %esi
> -	xor %edi, %edi
> +	CLEAR_REGS %eax, %ecx, %edx, %ebp, %esi, %edi
>  #ifdef CONFIG_X86_64
> -	xor %r8d,  %r8d
> -	xor %r9d,  %r9d
> -	xor %r10d, %r10d
> -	xor %r11d, %r11d
> -	xor %r12d, %r12d
> -	xor %r13d, %r13d
> -	xor %r14d, %r14d
> -	xor %r15d, %r15d
> +	CLEAR_REGS %r8d, %r9d, %r10d, %r11d, %r12d, %r13d, %r14d, %r15d
>  #endif
>  
>  	/*
> -- 
> 2.52.0
> 
> 

  reply	other threads:[~2026-05-14 19:30 UTC|newest]

Thread overview: 6+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2026-05-13 17:49 [PATCH 0/3] KVM: Macrofy GPR swapping in assembly code Paolo Bonzini
2026-05-13 17:49 ` [PATCH 1/3] KVM: VMX: Macrofy GPR swapping in __vmx_vcpu_run() Paolo Bonzini
2026-05-14 19:30   ` Sean Christopherson [this message]
2026-05-15  2:04     ` Chang S. Bae
2026-05-13 17:49 ` [PATCH 2/3] KVM: SVM: Macrofy GPR swapping in __svm_vcpu_run() Paolo Bonzini
2026-05-13 17:49 ` [PATCH 3/3] KVM: SEV: Macrofy GPR swapping in __svm_sev_es_vcpu_run() Paolo Bonzini

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=agYi2Lc7UbGiw_zL@google.com \
    --to=seanjc@google.com \
    --cc=chang.seok.bae@intel.com \
    --cc=kvm@vger.kernel.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=pbonzini@redhat.com \
    --cc=x86@kernel.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.