* [PATCH v2] hvf: use standard CR0 and CR4 register definitions
@ 2020-04-15 5:12 Cameron Esfahani via
2020-04-15 12:06 ` Paolo Bonzini
2020-04-15 13:01 ` Roman Bolshakov
0 siblings, 2 replies; 3+ messages in thread
From: Cameron Esfahani via @ 2020-04-15 5:12 UTC (permalink / raw)
To: qemu-devel; +Cc: pbonzini, r.bolshakov
Signed-off-by: Cameron Esfahani <dirty@apple.com>
---
v2:
+ Fix duplicate line Roman Bolshakov <r.bolshakov@yadro.com> found in review.
---
target/i386/cpu.h | 2 ++
target/i386/hvf/hvf.c | 2 +-
target/i386/hvf/vmx.h | 15 ++++++++-------
target/i386/hvf/x86.c | 6 +++---
target/i386/hvf/x86.h | 34 ----------------------------------
target/i386/hvf/x86_mmu.c | 2 +-
target/i386/hvf/x86_task.c | 3 ++-
7 files changed, 17 insertions(+), 47 deletions(-)
diff --git a/target/i386/cpu.h b/target/i386/cpu.h
index 60d797d594..1286ec6e7a 100644
--- a/target/i386/cpu.h
+++ b/target/i386/cpu.h
@@ -225,6 +225,8 @@ typedef enum X86Seg {
#define CR0_NE_MASK (1U << 5)
#define CR0_WP_MASK (1U << 16)
#define CR0_AM_MASK (1U << 18)
+#define CR0_NW_MASK (1U << 29)
+#define CR0_CD_MASK (1U << 30)
#define CR0_PG_MASK (1U << 31)
#define CR4_VME_MASK (1U << 0)
diff --git a/target/i386/hvf/hvf.c b/target/i386/hvf/hvf.c
index d72543dc31..48f3ef050c 100644
--- a/target/i386/hvf/hvf.c
+++ b/target/i386/hvf/hvf.c
@@ -455,7 +455,7 @@ void hvf_reset_vcpu(CPUState *cpu) {
wvmcs(cpu->hvf_fd, VMCS_GUEST_PDPTE0 + i * 2, pdpte[i]);
}
- macvm_set_cr0(cpu->hvf_fd, 0x60000010);
+ macvm_set_cr0(cpu->hvf_fd, CR0_CD_MASK | CR0_NW_MASK | CR0_ET_MASK);
wvmcs(cpu->hvf_fd, VMCS_CR4_MASK, CR4_VMXE_MASK);
wvmcs(cpu->hvf_fd, VMCS_CR4_SHADOW, 0x0);
diff --git a/target/i386/hvf/vmx.h b/target/i386/hvf/vmx.h
index 03d2c79b9c..8ec2e6414e 100644
--- a/target/i386/hvf/vmx.h
+++ b/target/i386/hvf/vmx.h
@@ -121,9 +121,10 @@ static inline void macvm_set_cr0(hv_vcpuid_t vcpu, uint64_t cr0)
uint64_t pdpte[4] = {0, 0, 0, 0};
uint64_t efer = rvmcs(vcpu, VMCS_GUEST_IA32_EFER);
uint64_t old_cr0 = rvmcs(vcpu, VMCS_GUEST_CR0);
- uint64_t mask = CR0_PG | CR0_CD | CR0_NW | CR0_NE | CR0_ET;
+ uint64_t mask = CR0_PG_MASK | CR0_CD_MASK | CR0_NW_MASK |
+ CR0_NE_MASK | CR0_ET_MASK;
- if ((cr0 & CR0_PG) && (rvmcs(vcpu, VMCS_GUEST_CR4) & CR4_PAE) &&
+ if ((cr0 & CR0_PG_MASK) && (rvmcs(vcpu, VMCS_GUEST_CR4) & CR4_PAE_MASK) &&
!(efer & MSR_EFER_LME)) {
address_space_read(&address_space_memory,
rvmcs(vcpu, VMCS_GUEST_CR3) & ~0x1f,
@@ -138,17 +139,17 @@ static inline void macvm_set_cr0(hv_vcpuid_t vcpu, uint64_t cr0)
wvmcs(vcpu, VMCS_CR0_SHADOW, cr0);
if (efer & MSR_EFER_LME) {
- if (!(old_cr0 & CR0_PG) && (cr0 & CR0_PG)) {
+ if (!(old_cr0 & CR0_PG_MASK) && (cr0 & CR0_PG_MASK)) {
enter_long_mode(vcpu, cr0, efer);
}
- if (/*(old_cr0 & CR0_PG) &&*/ !(cr0 & CR0_PG)) {
+ if (!(cr0 & CR0_PG_MASK)) {
exit_long_mode(vcpu, cr0, efer);
}
}
/* Filter new CR0 after we are finished examining it above. */
- cr0 = (cr0 & ~(mask & ~CR0_PG));
- wvmcs(vcpu, VMCS_GUEST_CR0, cr0 | CR0_NE | CR0_ET);
+ cr0 = (cr0 & ~(mask & ~CR0_PG_MASK));
+ wvmcs(vcpu, VMCS_GUEST_CR0, cr0 | CR0_NE_MASK | CR0_ET_MASK);
hv_vcpu_invalidate_tlb(vcpu);
hv_vcpu_flush(vcpu);
@@ -156,7 +157,7 @@ static inline void macvm_set_cr0(hv_vcpuid_t vcpu, uint64_t cr0)
static inline void macvm_set_cr4(hv_vcpuid_t vcpu, uint64_t cr4)
{
- uint64_t guest_cr4 = cr4 | CR4_VMXE;
+ uint64_t guest_cr4 = cr4 | CR4_VMXE_MASK;
wvmcs(vcpu, VMCS_GUEST_CR4, guest_cr4);
wvmcs(vcpu, VMCS_CR4_SHADOW, cr4);
diff --git a/target/i386/hvf/x86.c b/target/i386/hvf/x86.c
index 3afcedc7fc..668c02de6e 100644
--- a/target/i386/hvf/x86.c
+++ b/target/i386/hvf/x86.c
@@ -119,7 +119,7 @@ bool x86_read_call_gate(struct CPUState *cpu, struct x86_call_gate *idt_desc,
bool x86_is_protected(struct CPUState *cpu)
{
uint64_t cr0 = rvmcs(cpu->hvf_fd, VMCS_GUEST_CR0);
- return cr0 & CR0_PE;
+ return cr0 & CR0_PE_MASK;
}
bool x86_is_real(struct CPUState *cpu)
@@ -150,13 +150,13 @@ bool x86_is_long64_mode(struct CPUState *cpu)
bool x86_is_paging_mode(struct CPUState *cpu)
{
uint64_t cr0 = rvmcs(cpu->hvf_fd, VMCS_GUEST_CR0);
- return cr0 & CR0_PG;
+ return cr0 & CR0_PG_MASK;
}
bool x86_is_pae_enabled(struct CPUState *cpu)
{
uint64_t cr4 = rvmcs(cpu->hvf_fd, VMCS_GUEST_CR4);
- return cr4 & CR4_PAE;
+ return cr4 & CR4_PAE_MASK;
}
target_ulong linear_addr(struct CPUState *cpu, target_ulong addr, X86Seg seg)
diff --git a/target/i386/hvf/x86.h b/target/i386/hvf/x86.h
index c95d5b2116..bc0170b2a8 100644
--- a/target/i386/hvf/x86.h
+++ b/target/i386/hvf/x86.h
@@ -100,40 +100,6 @@ typedef struct x86_reg_flags {
};
} __attribute__ ((__packed__)) x86_reg_flags;
-typedef enum x86_reg_cr0 {
- CR0_PE = (1L << 0),
- CR0_MP = (1L << 1),
- CR0_EM = (1L << 2),
- CR0_TS = (1L << 3),
- CR0_ET = (1L << 4),
- CR0_NE = (1L << 5),
- CR0_WP = (1L << 16),
- CR0_AM = (1L << 18),
- CR0_NW = (1L << 29),
- CR0_CD = (1L << 30),
- CR0_PG = (1L << 31),
-} x86_reg_cr0;
-
-typedef enum x86_reg_cr4 {
- CR4_VME = (1L << 0),
- CR4_PVI = (1L << 1),
- CR4_TSD = (1L << 2),
- CR4_DE = (1L << 3),
- CR4_PSE = (1L << 4),
- CR4_PAE = (1L << 5),
- CR4_MSE = (1L << 6),
- CR4_PGE = (1L << 7),
- CR4_PCE = (1L << 8),
- CR4_OSFXSR = (1L << 9),
- CR4_OSXMMEXCPT = (1L << 10),
- CR4_VMXE = (1L << 13),
- CR4_SMXE = (1L << 14),
- CR4_FSGSBASE = (1L << 16),
- CR4_PCIDE = (1L << 17),
- CR4_OSXSAVE = (1L << 18),
- CR4_SMEP = (1L << 20),
-} x86_reg_cr4;
-
/* 16 bit Task State Segment */
typedef struct x86_tss_segment16 {
uint16_t link;
diff --git a/target/i386/hvf/x86_mmu.c b/target/i386/hvf/x86_mmu.c
index 65d4603dbf..8f38eccffc 100644
--- a/target/i386/hvf/x86_mmu.c
+++ b/target/i386/hvf/x86_mmu.c
@@ -130,7 +130,7 @@ static bool test_pt_entry(struct CPUState *cpu, struct gpt_translation *pt,
uint32_t cr0 = rvmcs(cpu->hvf_fd, VMCS_GUEST_CR0);
/* check protection */
- if (cr0 & CR0_WP) {
+ if (cr0 & CR0_WP_MASK) {
if (pt->write_access && !pte_write_access(pte)) {
return false;
}
diff --git a/target/i386/hvf/x86_task.c b/target/i386/hvf/x86_task.c
index 1daac6cc2b..5e41d09b89 100644
--- a/target/i386/hvf/x86_task.c
+++ b/target/i386/hvf/x86_task.c
@@ -174,7 +174,8 @@ void vmx_handle_task_switch(CPUState *cpu, x68_segment_selector tss_sel, int rea
//ret = task_switch_16(cpu, tss_sel, old_tss_sel, old_tss_base, &next_tss_desc);
VM_PANIC("task_switch_16");
- macvm_set_cr0(cpu->hvf_fd, rvmcs(cpu->hvf_fd, VMCS_GUEST_CR0) | CR0_TS);
+ macvm_set_cr0(cpu->hvf_fd,
+ rvmcs(cpu->hvf_fd, VMCS_GUEST_CR0) | CR0_TS_MASK);
x86_segment_descriptor_to_vmx(cpu, tss_sel, &next_tss_desc, &vmx_seg);
vmx_write_segment_descriptor(cpu, &vmx_seg, R_TR);
--
2.24.0
^ permalink raw reply related [flat|nested] 3+ messages in thread* Re: [PATCH v2] hvf: use standard CR0 and CR4 register definitions
2020-04-15 5:12 [PATCH v2] hvf: use standard CR0 and CR4 register definitions Cameron Esfahani via
@ 2020-04-15 12:06 ` Paolo Bonzini
2020-04-15 13:01 ` Roman Bolshakov
1 sibling, 0 replies; 3+ messages in thread
From: Paolo Bonzini @ 2020-04-15 12:06 UTC (permalink / raw)
To: Cameron Esfahani, qemu-devel; +Cc: r.bolshakov
On 15/04/20 07:12, Cameron Esfahani wrote:
> Signed-off-by: Cameron Esfahani <dirty@apple.com>
> ---
>
> v2:
> + Fix duplicate line Roman Bolshakov <r.bolshakov@yadro.com> found in review.
> ---
> target/i386/cpu.h | 2 ++
> target/i386/hvf/hvf.c | 2 +-
> target/i386/hvf/vmx.h | 15 ++++++++-------
> target/i386/hvf/x86.c | 6 +++---
> target/i386/hvf/x86.h | 34 ----------------------------------
> target/i386/hvf/x86_mmu.c | 2 +-
> target/i386/hvf/x86_task.c | 3 ++-
> 7 files changed, 17 insertions(+), 47 deletions(-)
>
> diff --git a/target/i386/cpu.h b/target/i386/cpu.h
> index 60d797d594..1286ec6e7a 100644
> --- a/target/i386/cpu.h
> +++ b/target/i386/cpu.h
> @@ -225,6 +225,8 @@ typedef enum X86Seg {
> #define CR0_NE_MASK (1U << 5)
> #define CR0_WP_MASK (1U << 16)
> #define CR0_AM_MASK (1U << 18)
> +#define CR0_NW_MASK (1U << 29)
> +#define CR0_CD_MASK (1U << 30)
> #define CR0_PG_MASK (1U << 31)
>
> #define CR4_VME_MASK (1U << 0)
> diff --git a/target/i386/hvf/hvf.c b/target/i386/hvf/hvf.c
> index d72543dc31..48f3ef050c 100644
> --- a/target/i386/hvf/hvf.c
> +++ b/target/i386/hvf/hvf.c
> @@ -455,7 +455,7 @@ void hvf_reset_vcpu(CPUState *cpu) {
> wvmcs(cpu->hvf_fd, VMCS_GUEST_PDPTE0 + i * 2, pdpte[i]);
> }
>
> - macvm_set_cr0(cpu->hvf_fd, 0x60000010);
> + macvm_set_cr0(cpu->hvf_fd, CR0_CD_MASK | CR0_NW_MASK | CR0_ET_MASK);
>
> wvmcs(cpu->hvf_fd, VMCS_CR4_MASK, CR4_VMXE_MASK);
> wvmcs(cpu->hvf_fd, VMCS_CR4_SHADOW, 0x0);
> diff --git a/target/i386/hvf/vmx.h b/target/i386/hvf/vmx.h
> index 03d2c79b9c..8ec2e6414e 100644
> --- a/target/i386/hvf/vmx.h
> +++ b/target/i386/hvf/vmx.h
> @@ -121,9 +121,10 @@ static inline void macvm_set_cr0(hv_vcpuid_t vcpu, uint64_t cr0)
> uint64_t pdpte[4] = {0, 0, 0, 0};
> uint64_t efer = rvmcs(vcpu, VMCS_GUEST_IA32_EFER);
> uint64_t old_cr0 = rvmcs(vcpu, VMCS_GUEST_CR0);
> - uint64_t mask = CR0_PG | CR0_CD | CR0_NW | CR0_NE | CR0_ET;
> + uint64_t mask = CR0_PG_MASK | CR0_CD_MASK | CR0_NW_MASK |
> + CR0_NE_MASK | CR0_ET_MASK;
>
> - if ((cr0 & CR0_PG) && (rvmcs(vcpu, VMCS_GUEST_CR4) & CR4_PAE) &&
> + if ((cr0 & CR0_PG_MASK) && (rvmcs(vcpu, VMCS_GUEST_CR4) & CR4_PAE_MASK) &&
> !(efer & MSR_EFER_LME)) {
> address_space_read(&address_space_memory,
> rvmcs(vcpu, VMCS_GUEST_CR3) & ~0x1f,
> @@ -138,17 +139,17 @@ static inline void macvm_set_cr0(hv_vcpuid_t vcpu, uint64_t cr0)
> wvmcs(vcpu, VMCS_CR0_SHADOW, cr0);
>
> if (efer & MSR_EFER_LME) {
> - if (!(old_cr0 & CR0_PG) && (cr0 & CR0_PG)) {
> + if (!(old_cr0 & CR0_PG_MASK) && (cr0 & CR0_PG_MASK)) {
> enter_long_mode(vcpu, cr0, efer);
> }
> - if (/*(old_cr0 & CR0_PG) &&*/ !(cr0 & CR0_PG)) {
> + if (!(cr0 & CR0_PG_MASK)) {
> exit_long_mode(vcpu, cr0, efer);
> }
> }
>
> /* Filter new CR0 after we are finished examining it above. */
> - cr0 = (cr0 & ~(mask & ~CR0_PG));
> - wvmcs(vcpu, VMCS_GUEST_CR0, cr0 | CR0_NE | CR0_ET);
> + cr0 = (cr0 & ~(mask & ~CR0_PG_MASK));
> + wvmcs(vcpu, VMCS_GUEST_CR0, cr0 | CR0_NE_MASK | CR0_ET_MASK);
>
> hv_vcpu_invalidate_tlb(vcpu);
> hv_vcpu_flush(vcpu);
> @@ -156,7 +157,7 @@ static inline void macvm_set_cr0(hv_vcpuid_t vcpu, uint64_t cr0)
>
> static inline void macvm_set_cr4(hv_vcpuid_t vcpu, uint64_t cr4)
> {
> - uint64_t guest_cr4 = cr4 | CR4_VMXE;
> + uint64_t guest_cr4 = cr4 | CR4_VMXE_MASK;
>
> wvmcs(vcpu, VMCS_GUEST_CR4, guest_cr4);
> wvmcs(vcpu, VMCS_CR4_SHADOW, cr4);
> diff --git a/target/i386/hvf/x86.c b/target/i386/hvf/x86.c
> index 3afcedc7fc..668c02de6e 100644
> --- a/target/i386/hvf/x86.c
> +++ b/target/i386/hvf/x86.c
> @@ -119,7 +119,7 @@ bool x86_read_call_gate(struct CPUState *cpu, struct x86_call_gate *idt_desc,
> bool x86_is_protected(struct CPUState *cpu)
> {
> uint64_t cr0 = rvmcs(cpu->hvf_fd, VMCS_GUEST_CR0);
> - return cr0 & CR0_PE;
> + return cr0 & CR0_PE_MASK;
> }
>
> bool x86_is_real(struct CPUState *cpu)
> @@ -150,13 +150,13 @@ bool x86_is_long64_mode(struct CPUState *cpu)
> bool x86_is_paging_mode(struct CPUState *cpu)
> {
> uint64_t cr0 = rvmcs(cpu->hvf_fd, VMCS_GUEST_CR0);
> - return cr0 & CR0_PG;
> + return cr0 & CR0_PG_MASK;
> }
>
> bool x86_is_pae_enabled(struct CPUState *cpu)
> {
> uint64_t cr4 = rvmcs(cpu->hvf_fd, VMCS_GUEST_CR4);
> - return cr4 & CR4_PAE;
> + return cr4 & CR4_PAE_MASK;
> }
>
> target_ulong linear_addr(struct CPUState *cpu, target_ulong addr, X86Seg seg)
> diff --git a/target/i386/hvf/x86.h b/target/i386/hvf/x86.h
> index c95d5b2116..bc0170b2a8 100644
> --- a/target/i386/hvf/x86.h
> +++ b/target/i386/hvf/x86.h
> @@ -100,40 +100,6 @@ typedef struct x86_reg_flags {
> };
> } __attribute__ ((__packed__)) x86_reg_flags;
>
> -typedef enum x86_reg_cr0 {
> - CR0_PE = (1L << 0),
> - CR0_MP = (1L << 1),
> - CR0_EM = (1L << 2),
> - CR0_TS = (1L << 3),
> - CR0_ET = (1L << 4),
> - CR0_NE = (1L << 5),
> - CR0_WP = (1L << 16),
> - CR0_AM = (1L << 18),
> - CR0_NW = (1L << 29),
> - CR0_CD = (1L << 30),
> - CR0_PG = (1L << 31),
> -} x86_reg_cr0;
> -
> -typedef enum x86_reg_cr4 {
> - CR4_VME = (1L << 0),
> - CR4_PVI = (1L << 1),
> - CR4_TSD = (1L << 2),
> - CR4_DE = (1L << 3),
> - CR4_PSE = (1L << 4),
> - CR4_PAE = (1L << 5),
> - CR4_MSE = (1L << 6),
> - CR4_PGE = (1L << 7),
> - CR4_PCE = (1L << 8),
> - CR4_OSFXSR = (1L << 9),
> - CR4_OSXMMEXCPT = (1L << 10),
> - CR4_VMXE = (1L << 13),
> - CR4_SMXE = (1L << 14),
> - CR4_FSGSBASE = (1L << 16),
> - CR4_PCIDE = (1L << 17),
> - CR4_OSXSAVE = (1L << 18),
> - CR4_SMEP = (1L << 20),
> -} x86_reg_cr4;
> -
> /* 16 bit Task State Segment */
> typedef struct x86_tss_segment16 {
> uint16_t link;
> diff --git a/target/i386/hvf/x86_mmu.c b/target/i386/hvf/x86_mmu.c
> index 65d4603dbf..8f38eccffc 100644
> --- a/target/i386/hvf/x86_mmu.c
> +++ b/target/i386/hvf/x86_mmu.c
> @@ -130,7 +130,7 @@ static bool test_pt_entry(struct CPUState *cpu, struct gpt_translation *pt,
>
> uint32_t cr0 = rvmcs(cpu->hvf_fd, VMCS_GUEST_CR0);
> /* check protection */
> - if (cr0 & CR0_WP) {
> + if (cr0 & CR0_WP_MASK) {
> if (pt->write_access && !pte_write_access(pte)) {
> return false;
> }
> diff --git a/target/i386/hvf/x86_task.c b/target/i386/hvf/x86_task.c
> index 1daac6cc2b..5e41d09b89 100644
> --- a/target/i386/hvf/x86_task.c
> +++ b/target/i386/hvf/x86_task.c
> @@ -174,7 +174,8 @@ void vmx_handle_task_switch(CPUState *cpu, x68_segment_selector tss_sel, int rea
> //ret = task_switch_16(cpu, tss_sel, old_tss_sel, old_tss_base, &next_tss_desc);
> VM_PANIC("task_switch_16");
>
> - macvm_set_cr0(cpu->hvf_fd, rvmcs(cpu->hvf_fd, VMCS_GUEST_CR0) | CR0_TS);
> + macvm_set_cr0(cpu->hvf_fd,
> + rvmcs(cpu->hvf_fd, VMCS_GUEST_CR0) | CR0_TS_MASK);
> x86_segment_descriptor_to_vmx(cpu, tss_sel, &next_tss_desc, &vmx_seg);
> vmx_write_segment_descriptor(cpu, &vmx_seg, R_TR);
>
>
Queued, thanks.
Paolo
^ permalink raw reply [flat|nested] 3+ messages in thread* Re: [PATCH v2] hvf: use standard CR0 and CR4 register definitions
2020-04-15 5:12 [PATCH v2] hvf: use standard CR0 and CR4 register definitions Cameron Esfahani via
2020-04-15 12:06 ` Paolo Bonzini
@ 2020-04-15 13:01 ` Roman Bolshakov
1 sibling, 0 replies; 3+ messages in thread
From: Roman Bolshakov @ 2020-04-15 13:01 UTC (permalink / raw)
To: Cameron Esfahani; +Cc: pbonzini, qemu-devel
On Tue, Apr 14, 2020 at 10:12:12PM -0700, Cameron Esfahani wrote:
> target/i386/cpu.h | 2 ++
> target/i386/hvf/hvf.c | 2 +-
> target/i386/hvf/vmx.h | 15 ++++++++-------
> target/i386/hvf/x86.c | 6 +++---
> target/i386/hvf/x86.h | 34 ----------------------------------
> target/i386/hvf/x86_mmu.c | 2 +-
> target/i386/hvf/x86_task.c | 3 ++-
> 7 files changed, 17 insertions(+), 47 deletions(-)
>
Hi Cameron,
Reviewed-by: Roman Bolshakov <r.bolshakov@yadro.com>
Thanks,
Roman
^ permalink raw reply [flat|nested] 3+ messages in thread
end of thread, other threads:[~2020-04-15 13:02 UTC | newest]
Thread overview: 3+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2020-04-15 5:12 [PATCH v2] hvf: use standard CR0 and CR4 register definitions Cameron Esfahani via
2020-04-15 12:06 ` Paolo Bonzini
2020-04-15 13:01 ` Roman Bolshakov
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).