From mboxrd@z Thu Jan 1 00:00:00 1970 From: Marcelo Tosatti Subject: [patch 3/5] KVM: MMU: add kvm_mmu_shadow_walk helper Date: Tue, 09 Jun 2009 18:30:12 -0300 Message-ID: <20090609213312.838419569@amt.cnet> References: <20090609213009.436123773@amt.cnet> Cc: avi@redhat.com, sheng.yang@intel.com, Marcelo Tosatti To: kvm@vger.kernel.org Return-path: Received: from mx2.redhat.com ([66.187.237.31]:42290 "EHLO mx2.redhat.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1756790AbZFIVe4 (ORCPT ); Tue, 9 Jun 2009 17:34:56 -0400 Content-Disposition: inline; filename=kvm_mmu_walk_shadow Sender: kvm-owner@vger.kernel.org List-ID: Required by EPT misconfiguration handler. Signed-off-by: Marcelo Tosatti Index: kvm/arch/x86/kvm/mmu.c =================================================================== --- kvm.orig/arch/x86/kvm/mmu.c +++ kvm/arch/x86/kvm/mmu.c @@ -3013,6 +3013,26 @@ out: return r; } +void kvm_mmu_shadow_walk(struct kvm_vcpu *vcpu, u64 addr, + struct mmu_shadow_walk *walk) +{ + struct kvm_shadow_walk_iterator iterator; + + spin_lock(&vcpu->kvm->mmu_lock); + for_each_shadow_entry(vcpu, addr, iterator) { + int err; + + err = walk->fn(vcpu, iterator.sptep, iterator.level, walk); + if (err) + break; + + if (!is_shadow_present_pte(*iterator.sptep)) + break; + } + spin_unlock(&vcpu->kvm->mmu_lock); +} +EXPORT_SYMBOL(kvm_mmu_shadow_walk); + #ifdef AUDIT static const char *audit_msg; Index: kvm/arch/x86/kvm/mmu.h =================================================================== --- kvm.orig/arch/x86/kvm/mmu.h +++ kvm/arch/x86/kvm/mmu.h @@ -37,6 +37,14 @@ #define PT32_ROOT_LEVEL 2 #define PT32E_ROOT_LEVEL 3 +struct mmu_shadow_walk { + int (*fn) (struct kvm_vcpu *vcpu, u64 *sptep, int level, + struct mmu_shadow_walk *walk); +}; + +void kvm_mmu_shadow_walk(struct kvm_vcpu *vcpu, u64 addr, + struct mmu_shadow_walk *walk); + static inline void kvm_mmu_free_some_pages(struct kvm_vcpu *vcpu) { if (unlikely(vcpu->kvm->arch.n_free_mmu_pages < KVM_MIN_FREE_MMU_PAGES))