* [PATCH v1] kvm: x86: implement PV send_IPI method @ 2025-07-18 6:24 Cindy Lu 2025-07-18 7:52 ` Jason Wang 0 siblings, 1 reply; 8+ messages in thread From: Cindy Lu @ 2025-07-18 6:24 UTC (permalink / raw) To: Paolo Bonzini, Vitaly Kuznetsov, Thomas Gleixner, Ingo Molnar, Borislav Petkov, Dave Hansen, maintainer:X86 ARCHITECTURE (32-BIT AND 64-BIT), H. Peter Anvin, Peter Zijlstra (Intel), Kirill A. Shutemov, Xin Li (Intel), Rik van Riel, Jason Wang, Ahmed S. Darwish, open list:KVM PARAVIRT (KVM/paravirt), open list:X86 ARCHITECTURE (32-BIT AND 64-BIT), lulu From: Jason Wang <jasowang@redhat.com> We used to have PV version of send_IPI_mask and send_IPI_mask_allbutself. This patch implements PV send_IPI method to reduce the number of vmexits. Signed-off-by: Jason Wang <jasowang@redhat.com> Tested-by: Cindy Lu <lulu@redhat.com> --- arch/x86/kernel/kvm.c | 6 ++++++ 1 file changed, 6 insertions(+) diff --git a/arch/x86/kernel/kvm.c b/arch/x86/kernel/kvm.c index 921c1c783bc1..b920cfd10441 100644 --- a/arch/x86/kernel/kvm.c +++ b/arch/x86/kernel/kvm.c @@ -557,6 +557,11 @@ static void __send_ipi_mask(const struct cpumask *mask, int vector) local_irq_restore(flags); } +static void kvm_send_ipi(int cpu, int vector) +{ + __send_ipi_mask(cpumask_of(cpu), vector); +} + static void kvm_send_ipi_mask(const struct cpumask *mask, int vector) { __send_ipi_mask(mask, vector); @@ -628,6 +633,7 @@ late_initcall(setup_efi_kvm_sev_migration); */ static __init void kvm_setup_pv_ipi(void) { + apic_update_callback(send_IPI, kvm_send_ipi); apic_update_callback(send_IPI_mask, kvm_send_ipi_mask); apic_update_callback(send_IPI_mask_allbutself, kvm_send_ipi_mask_allbutself); pr_info("setup PV IPIs\n"); -- 2.45.0 ^ permalink raw reply related [flat|nested] 8+ messages in thread
* Re: [PATCH v1] kvm: x86: implement PV send_IPI method 2025-07-18 6:24 [PATCH v1] kvm: x86: implement PV send_IPI method Cindy Lu @ 2025-07-18 7:52 ` Jason Wang 2025-07-18 11:00 ` Chao Gao 0 siblings, 1 reply; 8+ messages in thread From: Jason Wang @ 2025-07-18 7:52 UTC (permalink / raw) To: Cindy Lu Cc: Paolo Bonzini, Vitaly Kuznetsov, Thomas Gleixner, Ingo Molnar, Borislav Petkov, Dave Hansen, maintainer:X86 ARCHITECTURE (32-BIT AND 64-BIT), H. Peter Anvin, Peter Zijlstra (Intel), Kirill A. Shutemov, Xin Li (Intel), Rik van Riel, Ahmed S. Darwish, open list:KVM PARAVIRT (KVM/paravirt), open list:X86 ARCHITECTURE (32-BIT AND 64-BIT) On Fri, Jul 18, 2025 at 2:25 PM Cindy Lu <lulu@redhat.com> wrote: > > From: Jason Wang <jasowang@redhat.com> > > We used to have PV version of send_IPI_mask and > send_IPI_mask_allbutself. This patch implements PV send_IPI method to > reduce the number of vmexits. > > Signed-off-by: Jason Wang <jasowang@redhat.com> > Tested-by: Cindy Lu <lulu@redhat.com> I think a question here is are we able to see performance improvement in any kind of setup? Thanks ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [PATCH v1] kvm: x86: implement PV send_IPI method 2025-07-18 7:52 ` Jason Wang @ 2025-07-18 11:00 ` Chao Gao 2025-07-18 11:15 ` Jason Wang 0 siblings, 1 reply; 8+ messages in thread From: Chao Gao @ 2025-07-18 11:00 UTC (permalink / raw) To: Jason Wang Cc: Cindy Lu, Paolo Bonzini, Vitaly Kuznetsov, Thomas Gleixner, Ingo Molnar, Borislav Petkov, Dave Hansen, maintainer:X86 ARCHITECTURE (32-BIT AND 64-BIT), H. Peter Anvin, Peter Zijlstra (Intel), Kirill A. Shutemov, Xin Li (Intel), Rik van Riel, Ahmed S. Darwish, open list:KVM PARAVIRT (KVM/paravirt), open list:X86 ARCHITECTURE (32-BIT AND 64-BIT) On Fri, Jul 18, 2025 at 03:52:30PM +0800, Jason Wang wrote: >On Fri, Jul 18, 2025 at 2:25 PM Cindy Lu <lulu@redhat.com> wrote: >> >> From: Jason Wang <jasowang@redhat.com> >> >> We used to have PV version of send_IPI_mask and >> send_IPI_mask_allbutself. This patch implements PV send_IPI method to >> reduce the number of vmexits. It won't reduce the number of VM-exits; in fact, it may increase them on CPUs that support IPI virtualization. With IPI virtualization enabled, *unicast* and physical-addressing IPIs won't cause a VM-exit. Instead, the microcode posts interrupts directly to the target vCPU. The PV version always causes a VM-exit. >> >> Signed-off-by: Jason Wang <jasowang@redhat.com> >> Tested-by: Cindy Lu <lulu@redhat.com> > >I think a question here is are we able to see performance improvement >in any kind of setup? It may result in a negative performance impact. > >Thanks > > ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [PATCH v1] kvm: x86: implement PV send_IPI method 2025-07-18 11:00 ` Chao Gao @ 2025-07-18 11:15 ` Jason Wang 2025-07-18 12:19 ` Chao Gao 0 siblings, 1 reply; 8+ messages in thread From: Jason Wang @ 2025-07-18 11:15 UTC (permalink / raw) To: Chao Gao Cc: Cindy Lu, Paolo Bonzini, Vitaly Kuznetsov, Thomas Gleixner, Ingo Molnar, Borislav Petkov, Dave Hansen, maintainer:X86 ARCHITECTURE (32-BIT AND 64-BIT), H. Peter Anvin, Peter Zijlstra (Intel), Kirill A. Shutemov, Xin Li (Intel), Rik van Riel, Ahmed S. Darwish, open list:KVM PARAVIRT (KVM/paravirt), open list:X86 ARCHITECTURE (32-BIT AND 64-BIT) On Fri, Jul 18, 2025 at 7:01 PM Chao Gao <chao.gao@intel.com> wrote: > > On Fri, Jul 18, 2025 at 03:52:30PM +0800, Jason Wang wrote: > >On Fri, Jul 18, 2025 at 2:25 PM Cindy Lu <lulu@redhat.com> wrote: > >> > >> From: Jason Wang <jasowang@redhat.com> > >> > >> We used to have PV version of send_IPI_mask and > >> send_IPI_mask_allbutself. This patch implements PV send_IPI method to > >> reduce the number of vmexits. > > It won't reduce the number of VM-exits; in fact, it may increase them on CPUs > that support IPI virtualization. Sure, but I wonder if it reduces the vmexits when there's no APICV or L2 VM. I thought it can reduce the 2 vmexits to 1? > > With IPI virtualization enabled, *unicast* and physical-addressing IPIs won't > cause a VM-exit. Right. > Instead, the microcode posts interrupts directly to the target > vCPU. The PV version always causes a VM-exit. Yes, but it applies to all PV IPI I think. > > >> > >> Signed-off-by: Jason Wang <jasowang@redhat.com> > >> Tested-by: Cindy Lu <lulu@redhat.com> > > > >I think a question here is are we able to see performance improvement > >in any kind of setup? > > It may result in a negative performance impact. Userspace can check and enable PV IPI for the case where it suits. For example, HyperV did something like: void __init hv_apic_init(void) { if (ms_hyperv.hints & HV_X64_CLUSTER_IPI_RECOMMENDED) { pr_info("Hyper-V: Using IPI hypercalls\n"); /* * Set the IPI entry points. */ orig_apic = *apic; apic_update_callback(send_IPI, hv_send_ipi); apic_update_callback(send_IPI_mask, hv_send_ipi_mask); apic_update_callback(send_IPI_mask_allbutself, hv_send_ipi_mask_allbutself); apic_update_callback(send_IPI_allbutself, hv_send_ipi_allbutself); apic_update_callback(send_IPI_all, hv_send_ipi_all); apic_update_callback(send_IPI_self, hv_send_ipi_self); } send_IPI_mask is there. Thanks > > > > >Thanks > > > > > ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [PATCH v1] kvm: x86: implement PV send_IPI method 2025-07-18 11:15 ` Jason Wang @ 2025-07-18 12:19 ` Chao Gao 2025-07-18 14:01 ` Sean Christopherson 0 siblings, 1 reply; 8+ messages in thread From: Chao Gao @ 2025-07-18 12:19 UTC (permalink / raw) To: Jason Wang Cc: Cindy Lu, Paolo Bonzini, Vitaly Kuznetsov, Thomas Gleixner, Ingo Molnar, Borislav Petkov, Dave Hansen, maintainer:X86 ARCHITECTURE (32-BIT AND 64-BIT), H. Peter Anvin, Peter Zijlstra (Intel), Kirill A. Shutemov, Xin Li (Intel), Rik van Riel, Ahmed S. Darwish, open list:KVM PARAVIRT (KVM/paravirt), open list:X86 ARCHITECTURE (32-BIT AND 64-BIT) On Fri, Jul 18, 2025 at 07:15:37PM +0800, Jason Wang wrote: >On Fri, Jul 18, 2025 at 7:01 PM Chao Gao <chao.gao@intel.com> wrote: >> >> On Fri, Jul 18, 2025 at 03:52:30PM +0800, Jason Wang wrote: >> >On Fri, Jul 18, 2025 at 2:25 PM Cindy Lu <lulu@redhat.com> wrote: >> >> >> >> From: Jason Wang <jasowang@redhat.com> >> >> >> >> We used to have PV version of send_IPI_mask and >> >> send_IPI_mask_allbutself. This patch implements PV send_IPI method to >> >> reduce the number of vmexits. >> >> It won't reduce the number of VM-exits; in fact, it may increase them on CPUs >> that support IPI virtualization. > >Sure, but I wonder if it reduces the vmexits when there's no APICV or >L2 VM. I thought it can reduce the 2 vmexits to 1? Even without APICv, there is just 1 vmexit due to APIC write (xAPIC mode) or MSR write (x2APIC mode). > >> >> With IPI virtualization enabled, *unicast* and physical-addressing IPIs won't >> cause a VM-exit. > >Right. > >> Instead, the microcode posts interrupts directly to the target >> vCPU. The PV version always causes a VM-exit. > >Yes, but it applies to all PV IPI I think. For multi-cast IPIs, a single hypercall (PV IPI) outperforms multiple ICR writes, even when IPI virtualization is enabled. > >> >> >> >> >> Signed-off-by: Jason Wang <jasowang@redhat.com> >> >> Tested-by: Cindy Lu <lulu@redhat.com> >> > >> >I think a question here is are we able to see performance improvement >> >in any kind of setup? >> >> It may result in a negative performance impact. > >Userspace can check and enable PV IPI for the case where it suits. Yeah, we need to identify the cases. One example may be for TDX guests, using a PV approach (TDVMCALL) can avoid the #VE cost. > >For example, HyperV did something like: > >void __init hv_apic_init(void) >{ > if (ms_hyperv.hints & HV_X64_CLUSTER_IPI_RECOMMENDED) { > pr_info("Hyper-V: Using IPI hypercalls\n"); > /* > * Set the IPI entry points. > */ > orig_apic = *apic; > > apic_update_callback(send_IPI, hv_send_ipi); > apic_update_callback(send_IPI_mask, hv_send_ipi_mask); > apic_update_callback(send_IPI_mask_allbutself, >hv_send_ipi_mask_allbutself); > apic_update_callback(send_IPI_allbutself, >hv_send_ipi_allbutself); > apic_update_callback(send_IPI_all, hv_send_ipi_all); > apic_update_callback(send_IPI_self, hv_send_ipi_self); >} > >send_IPI_mask is there. > >Thanks > >> >> > >> >Thanks >> > >> > >> > ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [PATCH v1] kvm: x86: implement PV send_IPI method 2025-07-18 12:19 ` Chao Gao @ 2025-07-18 14:01 ` Sean Christopherson 2025-07-18 15:15 ` Chao Gao 0 siblings, 1 reply; 8+ messages in thread From: Sean Christopherson @ 2025-07-18 14:01 UTC (permalink / raw) To: Chao Gao Cc: Jason Wang, Cindy Lu, Paolo Bonzini, Vitaly Kuznetsov, Thomas Gleixner, Ingo Molnar, Borislav Petkov, Dave Hansen, maintainer:X86 ARCHITECTURE (32-BIT AND 64-BIT), H. Peter Anvin, Peter Zijlstra (Intel), Kirill A. Shutemov, Xin Li (Intel), Rik van Riel, Ahmed S. Darwish, open list:KVM PARAVIRT (KVM/paravirt), open list:X86 ARCHITECTURE (32-BIT AND 64-BIT) On Fri, Jul 18, 2025, Chao Gao wrote: > On Fri, Jul 18, 2025 at 07:15:37PM +0800, Jason Wang wrote: > >On Fri, Jul 18, 2025 at 7:01 PM Chao Gao <chao.gao@intel.com> wrote: > >> > >> On Fri, Jul 18, 2025 at 03:52:30PM +0800, Jason Wang wrote: > >> >On Fri, Jul 18, 2025 at 2:25 PM Cindy Lu <lulu@redhat.com> wrote: > >> >> > >> >> From: Jason Wang <jasowang@redhat.com> > >> >> > >> >> We used to have PV version of send_IPI_mask and > >> >> send_IPI_mask_allbutself. This patch implements PV send_IPI method to > >> >> reduce the number of vmexits. > >> > >> It won't reduce the number of VM-exits; in fact, it may increase them on CPUs > >> that support IPI virtualization. > > > >Sure, but I wonder if it reduces the vmexits when there's no APICV or > >L2 VM. I thought it can reduce the 2 vmexits to 1? > > Even without APICv, there is just 1 vmexit due to APIC write (xAPIC mode) > or MSR write (x2APIC mode). xAPIC will have two exits: ICR2 and then ICR. If xAPIC vs. x2APIC is stable when kvm_setup_pv_ipi() runs, maybe key off of that? > >> With IPI virtualization enabled, *unicast* and physical-addressing IPIs won't > >> cause a VM-exit. > > > >Right. > > > >> Instead, the microcode posts interrupts directly to the target > >> vCPU. The PV version always causes a VM-exit. > > > >Yes, but it applies to all PV IPI I think. > > For multi-cast IPIs, a single hypercall (PV IPI) outperforms multiple ICR > writes, even when IPI virtualization is enabled. FWIW, I doubt _all_ multi-cast IPIs outperform IPI virtualization. My guess is there's a threshold in the number of targets where the cost of sending multiple virtual IPIs becomes more expensive than the VM-Exit and software processing, and I assume/hope that threshold isn't '2'. > >> >> Signed-off-by: Jason Wang <jasowang@redhat.com> > >> >> Tested-by: Cindy Lu <lulu@redhat.com> > >> > > >> >I think a question here is are we able to see performance improvement > >> >in any kind of setup? > >> > >> It may result in a negative performance impact. > > > >Userspace can check and enable PV IPI for the case where it suits. > > Yeah, we need to identify the cases. One example may be for TDX guests, using > a PV approach (TDVMCALL) can avoid the #VE cost. TDX doesn't need a PV approach. Or rather, TDX already has an "architectural" PV approach. Make a TDVMCALL to request emulation of WRMSR(ICR). Don't plumb more KVM logic into it. ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [PATCH v1] kvm: x86: implement PV send_IPI method 2025-07-18 14:01 ` Sean Christopherson @ 2025-07-18 15:15 ` Chao Gao 2025-07-18 15:22 ` Sean Christopherson 0 siblings, 1 reply; 8+ messages in thread From: Chao Gao @ 2025-07-18 15:15 UTC (permalink / raw) To: Sean Christopherson Cc: Jason Wang, Cindy Lu, Paolo Bonzini, Vitaly Kuznetsov, Thomas Gleixner, Ingo Molnar, Borislav Petkov, Dave Hansen, maintainer:X86 ARCHITECTURE (32-BIT AND 64-BIT), H. Peter Anvin, Peter Zijlstra (Intel), Kirill A. Shutemov, Xin Li (Intel), Rik van Riel, Ahmed S. Darwish, open list:KVM PARAVIRT (KVM/paravirt), open list:X86 ARCHITECTURE (32-BIT AND 64-BIT) >> >> >> From: Jason Wang <jasowang@redhat.com> >> >> >> >> >> >> We used to have PV version of send_IPI_mask and >> >> >> send_IPI_mask_allbutself. This patch implements PV send_IPI method to >> >> >> reduce the number of vmexits. >> >> >> >> It won't reduce the number of VM-exits; in fact, it may increase them on CPUs >> >> that support IPI virtualization. >> > >> >Sure, but I wonder if it reduces the vmexits when there's no APICV or >> >L2 VM. I thought it can reduce the 2 vmexits to 1? >> >> Even without APICv, there is just 1 vmexit due to APIC write (xAPIC mode) >> or MSR write (x2APIC mode). > >xAPIC will have two exits: ICR2 and then ICR. ah, yes. >If xAPIC vs. x2APIC is stable when >kvm_setup_pv_ipi() runs, maybe key off of that? But the guest doesn't know if APICv is enabled or even IPI virtualization is enabled. > >> >> With IPI virtualization enabled, *unicast* and physical-addressing IPIs won't >> >> cause a VM-exit. >> > >> >Right. >> > >> >> Instead, the microcode posts interrupts directly to the target >> >> vCPU. The PV version always causes a VM-exit. >> > >> >Yes, but it applies to all PV IPI I think. >> >> For multi-cast IPIs, a single hypercall (PV IPI) outperforms multiple ICR >> writes, even when IPI virtualization is enabled. > >FWIW, I doubt _all_ multi-cast IPIs outperform IPI virtualization. My guess is >there's a threshold in the number of targets where the cost of sending multiple >virtual IPIs becomes more expensive than the VM-Exit and software processing, >and I assume/hope that threshold isn't '2'. Yes. Determining the threshold is tricky, and it's likely not a constant value across different CPU generations. > >> >> >> Signed-off-by: Jason Wang <jasowang@redhat.com> >> >> >> Tested-by: Cindy Lu <lulu@redhat.com> >> >> > >> >> >I think a question here is are we able to see performance improvement >> >> >in any kind of setup? >> >> >> >> It may result in a negative performance impact. >> > >> >Userspace can check and enable PV IPI for the case where it suits. >> >> Yeah, we need to identify the cases. One example may be for TDX guests, using >> a PV approach (TDVMCALL) can avoid the #VE cost. > >TDX doesn't need a PV approach. Or rather, TDX already has an "architectural" >PV approach. Make a TDVMCALL to request emulation of WRMSR(ICR). Don't plumb >more KVM logic into it. Agree. It should be an optimization for TDX guests, regardless of the underlying hypervisor. ^ permalink raw reply [flat|nested] 8+ messages in thread
* Re: [PATCH v1] kvm: x86: implement PV send_IPI method 2025-07-18 15:15 ` Chao Gao @ 2025-07-18 15:22 ` Sean Christopherson 0 siblings, 0 replies; 8+ messages in thread From: Sean Christopherson @ 2025-07-18 15:22 UTC (permalink / raw) To: Chao Gao Cc: Jason Wang, Cindy Lu, Paolo Bonzini, Vitaly Kuznetsov, Thomas Gleixner, Ingo Molnar, Borislav Petkov, Dave Hansen, maintainer:X86 ARCHITECTURE (32-BIT AND 64-BIT), H. Peter Anvin, Peter Zijlstra (Intel), Kirill A. Shutemov, Xin Li (Intel), Rik van Riel, Ahmed S. Darwish, open list:KVM PARAVIRT (KVM/paravirt), open list:X86 ARCHITECTURE (32-BIT AND 64-BIT) On Fri, Jul 18, 2025, Chao Gao wrote: > >> >> >> From: Jason Wang <jasowang@redhat.com> > >If xAPIC vs. x2APIC is stable when > >kvm_setup_pv_ipi() runs, maybe key off of that? > > But the guest doesn't know if APICv is enabled or even IPI virtualization > is enabled. Oh yeah, duh. Given that KVM emulates x2APIC irrespective of hardware support, and that Linux leans heavily towards x2APIC (thanks MMIO stale data!), my vote is to leave things as they are. ^ permalink raw reply [flat|nested] 8+ messages in thread
end of thread, other threads:[~2025-07-18 15:22 UTC | newest] Thread overview: 8+ messages (download: mbox.gz follow: Atom feed -- links below jump to the message on this page -- 2025-07-18 6:24 [PATCH v1] kvm: x86: implement PV send_IPI method Cindy Lu 2025-07-18 7:52 ` Jason Wang 2025-07-18 11:00 ` Chao Gao 2025-07-18 11:15 ` Jason Wang 2025-07-18 12:19 ` Chao Gao 2025-07-18 14:01 ` Sean Christopherson 2025-07-18 15:15 ` Chao Gao 2025-07-18 15:22 ` Sean Christopherson
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox; as well as URLs for NNTP newsgroup(s).