From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-18.7 required=3.0 tests=BAYES_00,DKIMWL_WL_HIGH, DKIM_SIGNED,DKIM_VALID,DKIM_VALID_AU,HEADER_FROM_DIFFERENT_DOMAINS, INCLUDES_CR_TRAILER,INCLUDES_PATCH,MAILING_LIST_MULTI,NICE_REPLY_A, SPF_HELO_NONE,SPF_PASS,USER_AGENT_SANE_1 autolearn=unavailable autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id D23C8C433EF for ; Wed, 22 Sep 2021 06:41:46 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id ADA8960F13 for ; Wed, 22 Sep 2021 06:41:46 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S232916AbhIVGnP (ORCPT ); Wed, 22 Sep 2021 02:43:15 -0400 Received: from us-smtp-delivery-124.mimecast.com ([216.205.24.124]:46469 "EHLO us-smtp-delivery-124.mimecast.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S232799AbhIVGnO (ORCPT ); Wed, 22 Sep 2021 02:43:14 -0400 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1632292905; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=kLkDWA7zis/qBBg4IjuyViLi2gFI3fX/kzAaOlA6siY=; b=AUkrta5k2ObKRfGty7Cy93vKFJf3RBCacTW6xVBC8r+G9cA/6r1hHuz7M/rNrYyCz4o9Kt Gy8egNtEXA6IKqCrQwwaM9rUvAdAoW7kqAb0deP5IQkTGp++F88NbdYJveLMDTZ1wW/VKp nJmksK+XVZPAr49UHKtodw745z0vWbQ= Received: from mail-ed1-f72.google.com (mail-ed1-f72.google.com [209.85.208.72]) (Using TLS) by relay.mimecast.com with ESMTP id us-mta-321-Fo0t0doCOz6GxqnF07efwQ-1; Wed, 22 Sep 2021 02:41:43 -0400 X-MC-Unique: Fo0t0doCOz6GxqnF07efwQ-1 Received: by mail-ed1-f72.google.com with SMTP id 2-20020a508e02000000b003d871759f5dso1839265edw.10 for ; Tue, 21 Sep 2021 23:41:43 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=x-gm-message-state:subject:to:cc:references:from:message-id:date :user-agent:mime-version:in-reply-to:content-language :content-transfer-encoding; bh=kLkDWA7zis/qBBg4IjuyViLi2gFI3fX/kzAaOlA6siY=; b=Q3qAPk1x1CtX3qHXnD/fZb0JJTBN5zQa+BU2xhfVAhB3FehUoxd0H6Gh6td+yKktb9 fBqEXaGNGd0N/Mq6zaQ0DFYSxfwTIL5iiJJWlm7nMXwZ74jTUTCpfSi/Qgoeb/Rtjr0G 4VPzkkQsRhozPGRn5LR3DxDTTcUPsSqZZ0/TBxeEwTWo7O7iywruLWgcWmFSPzHxKzJk t6ITI3dtWk+8tWX049f0reBDMI/hR/CYdS3C2OBCYbmf0m4B49hPY7cZM7K7gHX9bocy cToDW22W878RBoXYHJTgzi4n6jNkkzL3tcPWLrXJRmkuFtUpublNofn96Pp3vxLihHSh S0ng== X-Gm-Message-State: AOAM531t0jPkYAOZnPOMT/5rH2+h5W5mdySGc0skvOUPVwchD6s4sDH5 vEVmMAhezuxc6/4LJ281w0t+59EzKD4BS5z3TIk/xaR+FfoyOn2sGxzbcVW+qDsTC9nFGKKk+E9 SIvjG8TOh6mpnH0u0SUt6Gbi9D90lhA== X-Received: by 2002:a05:6402:5186:: with SMTP id q6mr40893902edd.64.1632292902398; Tue, 21 Sep 2021 23:41:42 -0700 (PDT) X-Google-Smtp-Source: ABdhPJwP+MY9+n/one3D2apviNbkQobxFZdru25IJ6syxfQ+H3Moyz3xvNvSjJ+iSQi50cWPcONDFw== X-Received: by 2002:a05:6402:5186:: with SMTP id q6mr40893882edd.64.1632292902156; Tue, 21 Sep 2021 23:41:42 -0700 (PDT) Received: from ?IPv6:2001:b07:6468:f312:c8dd:75d4:99ab:290a? ([2001:b07:6468:f312:c8dd:75d4:99ab:290a]) by smtp.gmail.com with ESMTPSA id kt19sm557582ejb.26.2021.09.21.23.41.31 (version=TLS1_3 cipher=TLS_AES_128_GCM_SHA256 bits=128/128); Tue, 21 Sep 2021 23:41:41 -0700 (PDT) Subject: Re: [PATCH v3 12/16] KVM: Move x86's perf guest info callbacks to generic KVM To: Sean Christopherson , Peter Zijlstra , Ingo Molnar , Arnaldo Carvalho de Melo , Will Deacon , Mark Rutland , Marc Zyngier , Guo Ren , Nick Hu , Greentime Hu , Vincent Chen , Paul Walmsley , Palmer Dabbelt , Albert Ou , Boris Ostrovsky , Juergen Gross Cc: Alexander Shishkin , Jiri Olsa , Namhyung Kim , James Morse , Alexandru Elisei , Suzuki K Poulose , Vitaly Kuznetsov , Wanpeng Li , Jim Mattson , Joerg Roedel , Stefano Stabellini , linux-arm-kernel@lists.infradead.org, linux-perf-users@vger.kernel.org, linux-kernel@vger.kernel.org, kvmarm@lists.cs.columbia.edu, linux-csky@vger.kernel.org, linux-riscv@lists.infradead.org, kvm@vger.kernel.org, xen-devel@lists.xenproject.org, Artem Kashkanov , Like Xu , Zhu Lingshan References: <20210922000533.713300-1-seanjc@google.com> <20210922000533.713300-13-seanjc@google.com> From: Paolo Bonzini Message-ID: <50cda5e5-be7e-00f3-886b-21b0fdfd1236@redhat.com> Date: Wed, 22 Sep 2021 08:41:30 +0200 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:78.0) Gecko/20100101 Thunderbird/78.11.0 MIME-Version: 1.0 In-Reply-To: <20210922000533.713300-13-seanjc@google.com> Content-Type: text/plain; charset=utf-8; format=flowed Content-Language: en-US Content-Transfer-Encoding: 7bit Precedence: bulk List-ID: X-Mailing-List: linux-perf-users@vger.kernel.org On 22/09/21 02:05, Sean Christopherson wrote: > Move x86's perf guest callbacks into common KVM, as they are semantically > identical to arm64's callbacks (the only other such KVM callbacks). > arm64 will convert to the common versions in a future patch. > > Implement the necessary arm64 arch hooks now to avoid having to provide > stubs or a temporary #define (from x86) to avoid arm64 compilation errors > when CONFIG_GUEST_PERF_EVENTS=y. > > Signed-off-by: Sean Christopherson > --- > arch/arm64/include/asm/kvm_host.h | 8 +++++ > arch/arm64/kvm/arm.c | 5 +++ > arch/x86/include/asm/kvm_host.h | 3 ++ > arch/x86/kvm/x86.c | 53 +++++++------------------------ > include/linux/kvm_host.h | 10 ++++++ > virt/kvm/kvm_main.c | 44 +++++++++++++++++++++++++ > 6 files changed, 81 insertions(+), 42 deletions(-) > > diff --git a/arch/arm64/include/asm/kvm_host.h b/arch/arm64/include/asm/kvm_host.h > index ed940aec89e0..828b6eaa2c56 100644 > --- a/arch/arm64/include/asm/kvm_host.h > +++ b/arch/arm64/include/asm/kvm_host.h > @@ -673,6 +673,14 @@ int io_mem_abort(struct kvm_vcpu *vcpu, phys_addr_t fault_ipa); > void kvm_perf_init(void); > void kvm_perf_teardown(void); > > +#ifdef CONFIG_GUEST_PERF_EVENTS > +static inline bool kvm_arch_pmi_in_guest(struct kvm_vcpu *vcpu) > +{ > + /* Any callback while a vCPU is loaded is considered to be in guest. */ > + return !!vcpu; > +} > +#endif > + > long kvm_hypercall_pv_features(struct kvm_vcpu *vcpu); > gpa_t kvm_init_stolen_time(struct kvm_vcpu *vcpu); > void kvm_update_stolen_time(struct kvm_vcpu *vcpu); > diff --git a/arch/arm64/kvm/arm.c b/arch/arm64/kvm/arm.c > index e9a2b8f27792..2b542fdc237e 100644 > --- a/arch/arm64/kvm/arm.c > +++ b/arch/arm64/kvm/arm.c > @@ -500,6 +500,11 @@ bool kvm_arch_vcpu_in_kernel(struct kvm_vcpu *vcpu) > return vcpu_mode_priv(vcpu); > } > > +unsigned long kvm_arch_vcpu_get_ip(struct kvm_vcpu *vcpu) > +{ > + return *vcpu_pc(vcpu); > +} > + > /* Just ensure a guest exit from a particular CPU */ > static void exit_vm_noop(void *info) > { > diff --git a/arch/x86/include/asm/kvm_host.h b/arch/x86/include/asm/kvm_host.h > index 2d86a2dfc775..6efe4e03a6d2 100644 > --- a/arch/x86/include/asm/kvm_host.h > +++ b/arch/x86/include/asm/kvm_host.h > @@ -1543,6 +1543,9 @@ static inline int kvm_arch_flush_remote_tlb(struct kvm *kvm) > return -ENOTSUPP; > } > > +#define kvm_arch_pmi_in_guest(vcpu) \ > + ((vcpu) && (vcpu)->arch.handling_intr_from_guest) > + > int kvm_mmu_module_init(void); > void kvm_mmu_module_exit(void); > > diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c > index 412646b973bb..1bea616402e6 100644 > --- a/arch/x86/kvm/x86.c > +++ b/arch/x86/kvm/x86.c > @@ -8264,43 +8264,12 @@ static void kvm_timer_init(void) > kvmclock_cpu_online, kvmclock_cpu_down_prep); > } > > -static inline bool kvm_pmi_in_guest(struct kvm_vcpu *vcpu) > -{ > - return vcpu && vcpu->arch.handling_intr_from_guest; > -} > - > -static unsigned int kvm_guest_state(void) > -{ > - struct kvm_vcpu *vcpu = kvm_get_running_vcpu(); > - unsigned int state; > - > - if (!kvm_pmi_in_guest(vcpu)) > - return 0; > - > - state = PERF_GUEST_ACTIVE; > - if (static_call(kvm_x86_get_cpl)(vcpu)) > - state |= PERF_GUEST_USER; > - > - return state; > -} > - > -static unsigned long kvm_guest_get_ip(void) > -{ > - struct kvm_vcpu *vcpu = kvm_get_running_vcpu(); > - > - /* Retrieving the IP must be guarded by a call to kvm_guest_state(). */ > - if (WARN_ON_ONCE(!kvm_pmi_in_guest(vcpu))) > - return 0; > - > - return kvm_rip_read(vcpu); > -} > - > static unsigned int kvm_handle_intel_pt_intr(void) > { > struct kvm_vcpu *vcpu = kvm_get_running_vcpu(); > > /* '0' on failure so that the !PT case can use a RET0 static call. */ > - if (!kvm_pmi_in_guest(vcpu)) > + if (!kvm_arch_pmi_in_guest(vcpu)) > return 0; > > kvm_make_request(KVM_REQ_PMI, vcpu); > @@ -8309,12 +8278,6 @@ static unsigned int kvm_handle_intel_pt_intr(void) > return 1; > } > > -static struct perf_guest_info_callbacks kvm_guest_cbs = { > - .state = kvm_guest_state, > - .get_ip = kvm_guest_get_ip, > - .handle_intel_pt_intr = NULL, > -}; > - > #ifdef CONFIG_X86_64 > static void pvclock_gtod_update_fn(struct work_struct *work) > { > @@ -11068,9 +11031,11 @@ int kvm_arch_hardware_setup(void *opaque) > memcpy(&kvm_x86_ops, ops->runtime_ops, sizeof(kvm_x86_ops)); > kvm_ops_static_call_update(); > > + /* Temporary ugliness. */ > if (ops->intel_pt_intr_in_guest && ops->intel_pt_intr_in_guest()) > - kvm_guest_cbs.handle_intel_pt_intr = kvm_handle_intel_pt_intr; > - perf_register_guest_info_callbacks(&kvm_guest_cbs); > + kvm_register_perf_callbacks(kvm_handle_intel_pt_intr); > + else > + kvm_register_perf_callbacks(NULL); > > if (!kvm_cpu_cap_has(X86_FEATURE_XSAVES)) > supported_xss = 0; > @@ -11099,8 +11064,7 @@ int kvm_arch_hardware_setup(void *opaque) > > void kvm_arch_hardware_unsetup(void) > { > - perf_unregister_guest_info_callbacks(&kvm_guest_cbs); > - kvm_guest_cbs.handle_intel_pt_intr = NULL; > + kvm_unregister_perf_callbacks(); > > static_call(kvm_x86_hardware_unsetup)(); > } > @@ -11727,6 +11691,11 @@ bool kvm_arch_vcpu_in_kernel(struct kvm_vcpu *vcpu) > return vcpu->arch.preempted_in_kernel; > } > > +unsigned long kvm_arch_vcpu_get_ip(struct kvm_vcpu *vcpu) > +{ > + return kvm_rip_read(vcpu); > +} > + > int kvm_arch_vcpu_should_kick(struct kvm_vcpu *vcpu) > { > return kvm_vcpu_exiting_guest_mode(vcpu) == IN_GUEST_MODE; > diff --git a/include/linux/kvm_host.h b/include/linux/kvm_host.h > index e4d712e9f760..b9255a6439f2 100644 > --- a/include/linux/kvm_host.h > +++ b/include/linux/kvm_host.h > @@ -1163,6 +1163,16 @@ static inline bool kvm_arch_intc_initialized(struct kvm *kvm) > } > #endif > > +#ifdef CONFIG_GUEST_PERF_EVENTS > +unsigned long kvm_arch_vcpu_get_ip(struct kvm_vcpu *vcpu); > + > +void kvm_register_perf_callbacks(unsigned int (*pt_intr_handler)(void)); > +void kvm_unregister_perf_callbacks(void); > +#else > +static inline void kvm_register_perf_callbacks(void *ign) {} > +static inline void kvm_unregister_perf_callbacks(void) {} > +#endif /* CONFIG_GUEST_PERF_EVENTS */ > + > int kvm_arch_init_vm(struct kvm *kvm, unsigned long type); > void kvm_arch_destroy_vm(struct kvm *kvm); > void kvm_arch_sync_events(struct kvm *kvm); > diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c > index 3e67c93ca403..179fb110a00f 100644 > --- a/virt/kvm/kvm_main.c > +++ b/virt/kvm/kvm_main.c > @@ -5460,6 +5460,50 @@ struct kvm_vcpu * __percpu *kvm_get_running_vcpus(void) > return &kvm_running_vcpu; > } > > +#ifdef CONFIG_GUEST_PERF_EVENTS > +static unsigned int kvm_guest_state(void) > +{ > + struct kvm_vcpu *vcpu = kvm_get_running_vcpu(); > + unsigned int state; > + > + if (!kvm_arch_pmi_in_guest(vcpu)) > + return 0; > + > + state = PERF_GUEST_ACTIVE; > + if (!kvm_arch_vcpu_in_kernel(vcpu)) > + state |= PERF_GUEST_USER; > + > + return state; > +} > + > +static unsigned long kvm_guest_get_ip(void) > +{ > + struct kvm_vcpu *vcpu = kvm_get_running_vcpu(); > + > + /* Retrieving the IP must be guarded by a call to kvm_guest_state(). */ > + if (WARN_ON_ONCE(!kvm_arch_pmi_in_guest(vcpu))) > + return 0; > + > + return kvm_arch_vcpu_get_ip(vcpu); > +} > + > +static struct perf_guest_info_callbacks kvm_guest_cbs = { > + .state = kvm_guest_state, > + .get_ip = kvm_guest_get_ip, > + .handle_intel_pt_intr = NULL, > +}; > + > +void kvm_register_perf_callbacks(unsigned int (*pt_intr_handler)(void)) > +{ > + kvm_guest_cbs.handle_intel_pt_intr = pt_intr_handler; > + perf_register_guest_info_callbacks(&kvm_guest_cbs); > +} > +void kvm_unregister_perf_callbacks(void) > +{ > + perf_unregister_guest_info_callbacks(&kvm_guest_cbs); > +} > +#endif > + > struct kvm_cpu_compat_check { > void *opaque; > int *ret; > Reviewed-by: Paolo Bonzini