* [PATCH v5 0/3] KVM: SVM: Flush cache only on CPUs running SEV guest
@ 2025-01-20 12:05 Zheyun Shen
2025-01-20 12:05 ` [PATCH v5 1/3] KVM: x86: Add a wbinvd helper Zheyun Shen
` (2 more replies)
0 siblings, 3 replies; 6+ messages in thread
From: Zheyun Shen @ 2025-01-20 12:05 UTC (permalink / raw)
To: thomas.lendacky, seanjc, pbonzini, tglx, kevinloughlin, mingo, bp
Cc: kvm, linux-kernel, Zheyun Shen
Previous versions pointed out the problem of wbinvd_on_all_cpus() in SEV
and tried to maintain a cpumask to solve it. This version futher removes
unnecessary calls to wbinvd().
Although dirty_mask is not maintained perfectly and may lead to wbinvd on
physical CPUs that are not running a SEV guest, it's still better than
wbinvd_on_all_cpus(). And vcpu migration is designed to be solved in
future work.
---
v4 -> v5:
- Added a commit to remove unnecessary calls to wbinvd().
v3 -> v4:
- Added a wbinvd helper and export it to SEV.
- Changed the struct cpumask in kvm_sev_info into cpumask*, which should
be dynamically allocated and freed.
- Changed the time of recording the CPUs from pre_sev_run() to vcpu_load().
- Removed code of clearing the mask.
v2 -> v3:
- Replaced get_cpu() with parameter cpu in pre_sev_run().
v1 -> v2:
- Added sev_do_wbinvd() to wrap two operations.
- Used cpumask_test_and_clear_cpu() to avoid concurrent problems.
---
Zheyun Shen (3):
KVM: x86: Add a wbinvd helper
KVM: SVM: Remove wbinvd in sev_vm_destroy()
KVM: SVM: Flush cache only on CPUs running SEV guest
arch/x86/kvm/svm/sev.c | 45 +++++++++++++++++++++++++++++++++---------
arch/x86/kvm/svm/svm.c | 2 ++
arch/x86/kvm/svm/svm.h | 5 ++++-
arch/x86/kvm/x86.c | 9 +++++++--
arch/x86/kvm/x86.h | 1 +
5 files changed, 50 insertions(+), 12 deletions(-)
--
2.34.1
^ permalink raw reply [flat|nested] 6+ messages in thread
* [PATCH v5 1/3] KVM: x86: Add a wbinvd helper
2025-01-20 12:05 [PATCH v5 0/3] KVM: SVM: Flush cache only on CPUs running SEV guest Zheyun Shen
@ 2025-01-20 12:05 ` Zheyun Shen
2025-01-20 12:05 ` [PATCH v5 2/3] KVM: SVM: Remove wbinvd in sev_vm_destroy() Zheyun Shen
2025-01-20 12:05 ` [PATCH v5 3/3] KVM: SVM: Flush cache only on CPUs running SEV guest Zheyun Shen
2 siblings, 0 replies; 6+ messages in thread
From: Zheyun Shen @ 2025-01-20 12:05 UTC (permalink / raw)
To: thomas.lendacky, seanjc, pbonzini, tglx, kevinloughlin, mingo, bp
Cc: kvm, linux-kernel, Zheyun Shen
At the moment open-coded calls to on_each_cpu_mask() are used when
emulating wbinvd. A subsequent patch needs the same behavior and the
helper prevents callers from preparing some idential parameters.
Signed-off-by: Zheyun Shen <szy0127@sjtu.edu.cn>
---
arch/x86/kvm/x86.c | 9 +++++++--
arch/x86/kvm/x86.h | 1 +
2 files changed, 8 insertions(+), 2 deletions(-)
diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c
index c79a8cc57..77f656306 100644
--- a/arch/x86/kvm/x86.c
+++ b/arch/x86/kvm/x86.c
@@ -8231,8 +8231,7 @@ static int kvm_emulate_wbinvd_noskip(struct kvm_vcpu *vcpu)
int cpu = get_cpu();
cpumask_set_cpu(cpu, vcpu->arch.wbinvd_dirty_mask);
- on_each_cpu_mask(vcpu->arch.wbinvd_dirty_mask,
- wbinvd_ipi, NULL, 1);
+ wbinvd_on_many_cpus(vcpu->arch.wbinvd_dirty_mask);
put_cpu();
cpumask_clear(vcpu->arch.wbinvd_dirty_mask);
} else
@@ -13971,6 +13970,12 @@ int kvm_sev_es_string_io(struct kvm_vcpu *vcpu, unsigned int size,
}
EXPORT_SYMBOL_GPL(kvm_sev_es_string_io);
+void wbinvd_on_many_cpus(struct cpumask *mask)
+{
+ on_each_cpu_mask(mask, wbinvd_ipi, NULL, 1);
+}
+EXPORT_SYMBOL_GPL(wbinvd_on_many_cpus);
+
EXPORT_TRACEPOINT_SYMBOL_GPL(kvm_entry);
EXPORT_TRACEPOINT_SYMBOL_GPL(kvm_exit);
EXPORT_TRACEPOINT_SYMBOL_GPL(kvm_fast_mmio);
diff --git a/arch/x86/kvm/x86.h b/arch/x86/kvm/x86.h
index ec623d23d..8f715e14b 100644
--- a/arch/x86/kvm/x86.h
+++ b/arch/x86/kvm/x86.h
@@ -611,5 +611,6 @@ int kvm_sev_es_mmio_read(struct kvm_vcpu *vcpu, gpa_t src, unsigned int bytes,
int kvm_sev_es_string_io(struct kvm_vcpu *vcpu, unsigned int size,
unsigned int port, void *data, unsigned int count,
int in);
+void wbinvd_on_many_cpus(struct cpumask *mask);
#endif
--
2.34.1
^ permalink raw reply related [flat|nested] 6+ messages in thread
* [PATCH v5 2/3] KVM: SVM: Remove wbinvd in sev_vm_destroy()
2025-01-20 12:05 [PATCH v5 0/3] KVM: SVM: Flush cache only on CPUs running SEV guest Zheyun Shen
2025-01-20 12:05 ` [PATCH v5 1/3] KVM: x86: Add a wbinvd helper Zheyun Shen
@ 2025-01-20 12:05 ` Zheyun Shen
2025-01-20 12:05 ` [PATCH v5 3/3] KVM: SVM: Flush cache only on CPUs running SEV guest Zheyun Shen
2 siblings, 0 replies; 6+ messages in thread
From: Zheyun Shen @ 2025-01-20 12:05 UTC (permalink / raw)
To: thomas.lendacky, seanjc, pbonzini, tglx, kevinloughlin, mingo, bp
Cc: kvm, linux-kernel, Zheyun Shen
Before sev_vm_destroy() is called, kvm_arch_guest_memory_reclaimed()
has been called for SEV and SEV-ES and kvm_arch_gmem_invalidate()
has been called for SEV-SNP. These functions have already handled
flushing the memory. Therefore, this wbinvd_on_all_cpus() can
simply be dropped.
Suggested-by: Sean Christopherson <seanjc@google.com>
Suggested-by: Kevin Loughlin <kevinloughlin@google.com>
Signed-off-by: Zheyun Shen <szy0127@sjtu.edu.cn>
---
arch/x86/kvm/svm/sev.c | 6 ------
1 file changed, 6 deletions(-)
diff --git a/arch/x86/kvm/svm/sev.c b/arch/x86/kvm/svm/sev.c
index 943bd074a..1ce67de9d 100644
--- a/arch/x86/kvm/svm/sev.c
+++ b/arch/x86/kvm/svm/sev.c
@@ -2899,12 +2899,6 @@ void sev_vm_destroy(struct kvm *kvm)
return;
}
- /*
- * Ensure that all guest tagged cache entries are flushed before
- * releasing the pages back to the system for use. CLFLUSH will
- * not do this, so issue a WBINVD.
- */
- wbinvd_on_all_cpus();
/*
* if userspace was terminated before unregistering the memory regions
--
2.34.1
^ permalink raw reply related [flat|nested] 6+ messages in thread
* [PATCH v5 3/3] KVM: SVM: Flush cache only on CPUs running SEV guest
2025-01-20 12:05 [PATCH v5 0/3] KVM: SVM: Flush cache only on CPUs running SEV guest Zheyun Shen
2025-01-20 12:05 ` [PATCH v5 1/3] KVM: x86: Add a wbinvd helper Zheyun Shen
2025-01-20 12:05 ` [PATCH v5 2/3] KVM: SVM: Remove wbinvd in sev_vm_destroy() Zheyun Shen
@ 2025-01-20 12:05 ` Zheyun Shen
2025-01-21 5:01 ` Nikunj A Dadhania
2 siblings, 1 reply; 6+ messages in thread
From: Zheyun Shen @ 2025-01-20 12:05 UTC (permalink / raw)
To: thomas.lendacky, seanjc, pbonzini, tglx, kevinloughlin, mingo, bp
Cc: kvm, linux-kernel, Zheyun Shen
On AMD CPUs without ensuring cache consistency, each memory page
reclamation in an SEV guest triggers a call to wbinvd_on_all_cpus(),
thereby affecting the performance of other programs on the host.
Typically, an AMD server may have 128 cores or more, while the SEV guest
might only utilize 8 of these cores. Meanwhile, host can use qemu-affinity
to bind these 8 vCPUs to specific physical CPUs.
Therefore, keeping a record of the physical core numbers each time a vCPU
runs can help avoid flushing the cache for all CPUs every time.
Suggested-by: Sean Christopherson <seanjc@google.com>
Signed-off-by: Zheyun Shen <szy0127@sjtu.edu.cn>
---
arch/x86/kvm/svm/sev.c | 39 ++++++++++++++++++++++++++++++++++++---
arch/x86/kvm/svm/svm.c | 2 ++
arch/x86/kvm/svm/svm.h | 5 ++++-
3 files changed, 42 insertions(+), 4 deletions(-)
diff --git a/arch/x86/kvm/svm/sev.c b/arch/x86/kvm/svm/sev.c
index 1ce67de9d..91469edd1 100644
--- a/arch/x86/kvm/svm/sev.c
+++ b/arch/x86/kvm/svm/sev.c
@@ -252,6 +252,36 @@ static void sev_asid_free(struct kvm_sev_info *sev)
sev->misc_cg = NULL;
}
+static struct cpumask *sev_get_wbinvd_dirty_mask(struct kvm *kvm)
+{
+ struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
+
+ return sev->wbinvd_dirty_mask;
+}
+
+void sev_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
+{
+ /*
+ * To optimize cache flushes when memory is reclaimed from an SEV VM,
+ * track physical CPUs that enter the guest for SEV VMs and thus can
+ * have encrypted, dirty data in the cache, and flush caches only for
+ * CPUs that have entered the guest.
+ */
+ cpumask_set_cpu(cpu, sev_get_wbinvd_dirty_mask(vcpu->kvm));
+}
+
+static void sev_do_wbinvd(struct kvm *kvm)
+{
+ struct cpumask *dirty_mask = sev_get_wbinvd_dirty_mask(kvm);
+
+ /*
+ * TODO: Clear CPUs from the bitmap prior to flushing. Doing so
+ * requires serializing multiple calls and having CPUs mark themselves
+ * "dirty" if they are currently running a vCPU for the VM.
+ */
+ wbinvd_on_many_cpus(dirty_mask);
+}
+
static void sev_decommission(unsigned int handle)
{
struct sev_data_decommission decommission;
@@ -448,6 +478,8 @@ static int __sev_guest_init(struct kvm *kvm, struct kvm_sev_cmd *argp,
ret = sev_platform_init(&init_args);
if (ret)
goto e_free;
+ if (!zalloc_cpumask_var(&sev->wbinvd_dirty_mask, GFP_KERNEL_ACCOUNT))
+ goto e_free;
/* This needs to happen after SEV/SNP firmware initialization. */
if (vm_type == KVM_X86_SNP_VM) {
@@ -2778,7 +2810,7 @@ int sev_mem_enc_unregister_region(struct kvm *kvm,
* releasing the pages back to the system for use. CLFLUSH will
* not do this, so issue a WBINVD.
*/
- wbinvd_on_all_cpus();
+ sev_do_wbinvd(kvm);
__unregister_enc_region_locked(kvm, region);
@@ -2926,6 +2958,7 @@ void sev_vm_destroy(struct kvm *kvm)
}
sev_asid_free(sev);
+ free_cpumask_var(sev->wbinvd_dirty_mask);
}
void __init sev_set_cpu_caps(void)
@@ -3129,7 +3162,7 @@ static void sev_flush_encrypted_page(struct kvm_vcpu *vcpu, void *va)
return;
do_wbinvd:
- wbinvd_on_all_cpus();
+ sev_do_wbinvd(vcpu->kvm);
}
void sev_guest_memory_reclaimed(struct kvm *kvm)
@@ -3143,7 +3176,7 @@ void sev_guest_memory_reclaimed(struct kvm *kvm)
if (!sev_guest(kvm) || sev_snp_guest(kvm))
return;
- wbinvd_on_all_cpus();
+ sev_do_wbinvd(kvm);
}
void sev_free_vcpu(struct kvm_vcpu *vcpu)
diff --git a/arch/x86/kvm/svm/svm.c b/arch/x86/kvm/svm/svm.c
index 21dacd312..d2a423c0e 100644
--- a/arch/x86/kvm/svm/svm.c
+++ b/arch/x86/kvm/svm/svm.c
@@ -1565,6 +1565,8 @@ static void svm_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
}
if (kvm_vcpu_apicv_active(vcpu))
avic_vcpu_load(vcpu, cpu);
+ if (sev_guest(vcpu->kvm))
+ sev_vcpu_load(vcpu, cpu);
}
static void svm_vcpu_put(struct kvm_vcpu *vcpu)
diff --git a/arch/x86/kvm/svm/svm.h b/arch/x86/kvm/svm/svm.h
index 43fa6a16e..c8f42cb61 100644
--- a/arch/x86/kvm/svm/svm.h
+++ b/arch/x86/kvm/svm/svm.h
@@ -112,6 +112,8 @@ struct kvm_sev_info {
void *guest_req_buf; /* Bounce buffer for SNP Guest Request input */
void *guest_resp_buf; /* Bounce buffer for SNP Guest Request output */
struct mutex guest_req_mutex; /* Must acquire before using bounce buffers */
+ /* CPUs invoked VMRUN call wbinvd after guest memory is reclaimed */
+ struct cpumask *wbinvd_dirty_mask;
};
struct kvm_svm {
@@ -763,6 +765,7 @@ void sev_snp_init_protected_guest_state(struct kvm_vcpu *vcpu);
int sev_gmem_prepare(struct kvm *kvm, kvm_pfn_t pfn, gfn_t gfn, int max_order);
void sev_gmem_invalidate(kvm_pfn_t start, kvm_pfn_t end);
int sev_private_max_mapping_level(struct kvm *kvm, kvm_pfn_t pfn);
+void sev_vcpu_load(struct kvm_vcpu *vcpu, int cpu);
#else
static inline struct page *snp_safe_alloc_page_node(int node, gfp_t gfp)
{
@@ -793,7 +796,7 @@ static inline int sev_private_max_mapping_level(struct kvm *kvm, kvm_pfn_t pfn)
{
return 0;
}
-
+static inline void sev_vcpu_load(struct kvm_vcpu *vcpu, int cpu) {}
#endif
/* vmenter.S */
--
2.34.1
^ permalink raw reply related [flat|nested] 6+ messages in thread
* Re: [PATCH v5 3/3] KVM: SVM: Flush cache only on CPUs running SEV guest
2025-01-20 12:05 ` [PATCH v5 3/3] KVM: SVM: Flush cache only on CPUs running SEV guest Zheyun Shen
@ 2025-01-21 5:01 ` Nikunj A Dadhania
2025-01-26 10:33 ` Zheyun Shen
0 siblings, 1 reply; 6+ messages in thread
From: Nikunj A Dadhania @ 2025-01-21 5:01 UTC (permalink / raw)
To: Zheyun Shen, thomas.lendacky, seanjc, pbonzini, tglx,
kevinloughlin, mingo, bp
Cc: kvm, linux-kernel, Zheyun Shen
Zheyun Shen <szy0127@sjtu.edu.cn> writes:
> On AMD CPUs without ensuring cache consistency, each memory page
> reclamation in an SEV guest triggers a call to wbinvd_on_all_cpus(),
> thereby affecting the performance of other programs on the host.
>
> Typically, an AMD server may have 128 cores or more, while the SEV guest
> might only utilize 8 of these cores. Meanwhile, host can use qemu-affinity
> to bind these 8 vCPUs to specific physical CPUs.
>
> Therefore, keeping a record of the physical core numbers each time a vCPU
> runs can help avoid flushing the cache for all CPUs every time.
>
> Suggested-by: Sean Christopherson <seanjc@google.com>
> Signed-off-by: Zheyun Shen <szy0127@sjtu.edu.cn>
> ---
> arch/x86/kvm/svm/sev.c | 39 ++++++++++++++++++++++++++++++++++++---
> arch/x86/kvm/svm/svm.c | 2 ++
> arch/x86/kvm/svm/svm.h | 5 ++++-
> 3 files changed, 42 insertions(+), 4 deletions(-)
>
> diff --git a/arch/x86/kvm/svm/sev.c b/arch/x86/kvm/svm/sev.c
> index 1ce67de9d..91469edd1 100644
> --- a/arch/x86/kvm/svm/sev.c
> +++ b/arch/x86/kvm/svm/sev.c
> @@ -252,6 +252,36 @@ static void sev_asid_free(struct kvm_sev_info *sev)
> sev->misc_cg = NULL;
> }
>
> +static struct cpumask *sev_get_wbinvd_dirty_mask(struct kvm *kvm)
> +{
> + struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
There is a helper to get sev_info: to_kvm_sev_info(), if you use that,
sev_get_wbinvd_dirty_mask() helper will not be needed.
> +
> + return sev->wbinvd_dirty_mask;
> +}
> +
> +void sev_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
> +{
> + /*
> + * To optimize cache flushes when memory is reclaimed from an SEV VM,
> + * track physical CPUs that enter the guest for SEV VMs and thus can
> + * have encrypted, dirty data in the cache, and flush caches only for
> + * CPUs that have entered the guest.
> + */
> + cpumask_set_cpu(cpu, sev_get_wbinvd_dirty_mask(vcpu->kvm));
> +}
> +
> +static void sev_do_wbinvd(struct kvm *kvm)
> +{
> + struct cpumask *dirty_mask = sev_get_wbinvd_dirty_mask(kvm);
> +
> + /*
> + * TODO: Clear CPUs from the bitmap prior to flushing. Doing so
> + * requires serializing multiple calls and having CPUs mark themselves
> + * "dirty" if they are currently running a vCPU for the VM.
> + */
> + wbinvd_on_many_cpus(dirty_mask);
> +}
Something like the below
void sev_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
{
/* ... */
cpumask_set_cpu(cpu, to_kvm_sev_info(kvm)->wbinvd_dirty_mask);
}
static void sev_do_wbinvd(struct kvm *kvm)
{
/* ... */
wbinvd_on_many_cpus(to_kvm_sev_info(kvm)->wbinvd_dirty_mask);
}
Regards,
Nikunj
^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: [PATCH v5 3/3] KVM: SVM: Flush cache only on CPUs running SEV guest
2025-01-21 5:01 ` Nikunj A Dadhania
@ 2025-01-26 10:33 ` Zheyun Shen
0 siblings, 0 replies; 6+ messages in thread
From: Zheyun Shen @ 2025-01-26 10:33 UTC (permalink / raw)
To: Nikunj A Dadhania
Cc: thomas.lendacky, seanjc, pbonzini, tglx, kevinloughlin, mingo, bp,
kvm, linux-kernel
> Nikunj A Dadhania <nikunj@amd.com> writes:
>
> Zheyun Shen <szy0127@sjtu.edu.cn> writes:
>
>> On AMD CPUs without ensuring cache consistency, each memory page
>> reclamation in an SEV guest triggers a call to wbinvd_on_all_cpus(),
>> thereby affecting the performance of other programs on the host.
>>
>> Typically, an AMD server may have 128 cores or more, while the SEV guest
>> might only utilize 8 of these cores. Meanwhile, host can use qemu-affinity
>> to bind these 8 vCPUs to specific physical CPUs.
>>
>> Therefore, keeping a record of the physical core numbers each time a vCPU
>> runs can help avoid flushing the cache for all CPUs every time.
>>
>> Suggested-by: Sean Christopherson <seanjc@google.com>
>> Signed-off-by: Zheyun Shen <szy0127@sjtu.edu.cn>
>> ---
>> arch/x86/kvm/svm/sev.c | 39 ++++++++++++++++++++++++++++++++++++---
>> arch/x86/kvm/svm/svm.c | 2 ++
>> arch/x86/kvm/svm/svm.h | 5 ++++-
>> 3 files changed, 42 insertions(+), 4 deletions(-)
>>
>> diff --git a/arch/x86/kvm/svm/sev.c b/arch/x86/kvm/svm/sev.c
>> index 1ce67de9d..91469edd1 100644
>> --- a/arch/x86/kvm/svm/sev.c
>> +++ b/arch/x86/kvm/svm/sev.c
>> @@ -252,6 +252,36 @@ static void sev_asid_free(struct kvm_sev_info *sev)
>> sev->misc_cg = NULL;
>> }
>>
>> +static struct cpumask *sev_get_wbinvd_dirty_mask(struct kvm *kvm)
>> +{
>> + struct kvm_sev_info *sev = &to_kvm_svm(kvm)->sev_info;
>
> There is a helper to get sev_info: to_kvm_sev_info(), if you use that,
> sev_get_wbinvd_dirty_mask() helper will not be needed.
>
>> +
>> + return sev->wbinvd_dirty_mask;
>> +}
>> +
>> +void sev_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
>> +{
>> + /*
>> + * To optimize cache flushes when memory is reclaimed from an SEV VM,
>> + * track physical CPUs that enter the guest for SEV VMs and thus can
>> + * have encrypted, dirty data in the cache, and flush caches only for
>> + * CPUs that have entered the guest.
>> + */
>> + cpumask_set_cpu(cpu, sev_get_wbinvd_dirty_mask(vcpu->kvm));
>> +}
>> +
>> +static void sev_do_wbinvd(struct kvm *kvm)
>> +{
>> + struct cpumask *dirty_mask = sev_get_wbinvd_dirty_mask(kvm);
>> +
>> + /*
>> + * TODO: Clear CPUs from the bitmap prior to flushing. Doing so
>> + * requires serializing multiple calls and having CPUs mark themselves
>> + * "dirty" if they are currently running a vCPU for the VM.
>> + */
>> + wbinvd_on_many_cpus(dirty_mask);
>> +}
>
> Something like the below
>
> void sev_vcpu_load(struct kvm_vcpu *vcpu, int cpu)
> {
> /* ... */
> cpumask_set_cpu(cpu, to_kvm_sev_info(kvm)->wbinvd_dirty_mask);
> }
>
> static void sev_do_wbinvd(struct kvm *kvm)
> {
> /* ... */
> wbinvd_on_many_cpus(to_kvm_sev_info(kvm)->wbinvd_dirty_mask);
> }
>
> Regards,
> Nikunj
>
Got it, thanks.
Regards,
Zheyun Shen
^ permalink raw reply [flat|nested] 6+ messages in thread
end of thread, other threads:[~2025-01-26 10:43 UTC | newest]
Thread overview: 6+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2025-01-20 12:05 [PATCH v5 0/3] KVM: SVM: Flush cache only on CPUs running SEV guest Zheyun Shen
2025-01-20 12:05 ` [PATCH v5 1/3] KVM: x86: Add a wbinvd helper Zheyun Shen
2025-01-20 12:05 ` [PATCH v5 2/3] KVM: SVM: Remove wbinvd in sev_vm_destroy() Zheyun Shen
2025-01-20 12:05 ` [PATCH v5 3/3] KVM: SVM: Flush cache only on CPUs running SEV guest Zheyun Shen
2025-01-21 5:01 ` Nikunj A Dadhania
2025-01-26 10:33 ` Zheyun Shen
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox