From: Xiao Guangrong <xiaoguangrong@cn.fujitsu.com>
To: Avi Kivity <avi@redhat.com>
Cc: Marcelo Tosatti <mtosatti@redhat.com>,
LKML <linux-kernel@vger.kernel.org>,
KVM list <kvm@vger.kernel.org>
Subject: [PATCH 5/6] KVM: MMU: prefetch ptes when intercepted guest #PF
Date: Tue, 15 Jun 2010 10:47:04 +0800 [thread overview]
Message-ID: <4C16E9A8.10409@cn.fujitsu.com> (raw)
In-Reply-To: <4C16E82E.5010306@cn.fujitsu.com>
Support prefetch ptes when intercept guest #PF, avoid to #PF by later
access
If we meet any failure in the prefetch path, we will exit it and
not try other ptes to avoid become heavy path
Signed-off-by: Xiao Guangrong <xiaoguangrong@cn.fujitsu.com>
---
arch/x86/kvm/mmu.c | 36 +++++++++++++++++++++
arch/x86/kvm/paging_tmpl.h | 76 ++++++++++++++++++++++++++++++++++++++++++++
2 files changed, 112 insertions(+), 0 deletions(-)
diff --git a/arch/x86/kvm/mmu.c b/arch/x86/kvm/mmu.c
index 92ff099..941c86b 100644
--- a/arch/x86/kvm/mmu.c
+++ b/arch/x86/kvm/mmu.c
@@ -89,6 +89,8 @@ module_param(oos_shadow, bool, 0644);
}
#endif
+#define PTE_PREFETCH_NUM 16
+
#define PT_FIRST_AVAIL_BITS_SHIFT 9
#define PT64_SECOND_AVAIL_BITS_SHIFT 52
@@ -2041,6 +2043,39 @@ static void nonpaging_new_cr3(struct kvm_vcpu *vcpu)
{
}
+static void direct_pte_prefetch(struct kvm_vcpu *vcpu, u64 *sptep)
+{
+ struct kvm_mmu_page *sp;
+ int index, i;
+
+ sp = page_header(__pa(sptep));
+ WARN_ON(!sp->role.direct);
+ index = sptep - sp->spt;
+
+ for (i = index + 1; i < min(PT64_ENT_PER_PAGE,
+ index + PTE_PREFETCH_NUM); i++) {
+ gfn_t gfn;
+ pfn_t pfn;
+ u64 *spte = sp->spt + i;
+
+ if (*spte != shadow_trap_nonpresent_pte)
+ continue;
+
+ gfn = sp->gfn + (i << ((sp->role.level - 1) * PT64_LEVEL_BITS));
+
+ pfn = gfn_to_pfn_atomic(vcpu->kvm, gfn);
+ if (is_error_pfn(pfn)) {
+ kvm_release_pfn_clean(pfn);
+ break;
+ }
+ if (pte_prefetch_topup_memory_cache(vcpu))
+ break;
+
+ mmu_set_spte(vcpu, spte, ACC_ALL, ACC_ALL, 0, 0, 1, NULL,
+ sp->role.level, gfn, pfn, true, false);
+ }
+}
+
static int __direct_map(struct kvm_vcpu *vcpu, gpa_t v, int write,
int level, gfn_t gfn, pfn_t pfn)
{
@@ -2055,6 +2090,7 @@ static int __direct_map(struct kvm_vcpu *vcpu, gpa_t v, int write,
0, write, 1, &pt_write,
level, gfn, pfn, false, true);
++vcpu->stat.pf_fixed;
+ direct_pte_prefetch(vcpu, iterator.sptep);
break;
}
diff --git a/arch/x86/kvm/paging_tmpl.h b/arch/x86/kvm/paging_tmpl.h
index eb47148..af4e041 100644
--- a/arch/x86/kvm/paging_tmpl.h
+++ b/arch/x86/kvm/paging_tmpl.h
@@ -291,6 +291,81 @@ static void FNAME(update_pte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp,
gpte_to_gfn(gpte), pfn, true, true);
}
+static void FNAME(pte_prefetch)(struct kvm_vcpu *vcpu, u64 *sptep)
+{
+ struct kvm_mmu_page *sp;
+ pt_element_t *table = NULL;
+ int offset = 0, shift, index, i;
+
+ sp = page_header(__pa(sptep));
+ index = sptep - sp->spt;
+
+ if (PTTYPE == 32) {
+ shift = PAGE_SHIFT - (PT_LEVEL_BITS -
+ PT64_LEVEL_BITS) * sp->role.level;
+ offset = sp->role.quadrant << shift;
+ }
+
+ for (i = index + 1; i < min(PT64_ENT_PER_PAGE,
+ index + PTE_PREFETCH_NUM); i++) {
+ struct page *page;
+ pt_element_t gpte;
+ unsigned pte_access;
+ u64 *spte = sp->spt + i;
+ gfn_t gfn;
+ pfn_t pfn;
+ int dirty;
+
+ if (*spte != shadow_trap_nonpresent_pte)
+ continue;
+
+ pte_access = sp->role.access;
+ if (sp->role.direct) {
+ dirty = 1;
+ gfn = sp->gfn + (i << ((sp->role.level - 1) *
+ PT64_LEVEL_BITS));
+ goto gfn_mapping;
+ }
+
+ if (!table) {
+ page = gfn_to_page_atomic(vcpu->kvm, sp->gfn);
+ if (is_error_page(page)) {
+ kvm_release_page_clean(page);
+ break;
+ }
+ table = kmap_atomic(page, KM_USER0);
+ table = (pt_element_t *)((char *)table + offset);
+ }
+
+ gpte = table[i];
+ if (!(gpte & PT_ACCESSED_MASK))
+ continue;
+
+ if (!is_present_gpte(gpte)) {
+ if (!sp->unsync)
+ *spte = shadow_notrap_nonpresent_pte;
+ continue;
+ }
+ dirty = is_dirty_gpte(gpte);
+ gfn = (gpte & PT64_BASE_ADDR_MASK) >> PAGE_SHIFT;
+ pte_access = pte_access & FNAME(gpte_access)(vcpu, gpte);
+gfn_mapping:
+ pfn = gfn_to_pfn_atomic(vcpu->kvm, gfn);
+ if (is_error_pfn(pfn)) {
+ kvm_release_pfn_clean(pfn);
+ break;
+ }
+
+ if (pte_prefetch_topup_memory_cache(vcpu))
+ break;
+ mmu_set_spte(vcpu, spte, sp->role.access, pte_access, 0, 0,
+ dirty, NULL, sp->role.level, gfn, pfn,
+ true, false);
+ }
+ if (table)
+ kunmap_atomic((char *)table - offset, KM_USER0);
+}
+
/*
* Fetch a shadow pte for a specific level in the paging hierarchy.
*/
@@ -322,6 +397,7 @@ static u64 *FNAME(fetch)(struct kvm_vcpu *vcpu, gva_t addr,
is_dirty_gpte(gw->ptes[gw->level-1]),
ptwrite, level,
gw->gfn, pfn, false, true);
+ FNAME(pte_prefetch)(vcpu, sptep);
break;
}
--
1.6.1.2
next prev parent reply other threads:[~2010-06-15 2:50 UTC|newest]
Thread overview: 23+ messages / expand[flat|nested] mbox.gz Atom feed top
[not found] <4C16E6ED.7020009@cn.fujitsu.com>
2010-06-15 2:46 ` [PATCH 1/6] KVM: MMU: fix gfn got in kvm_mmu_page_get_gfn() Xiao Guangrong
[not found] ` <4C16E75F.6020003@cn.fujitsu.com>
2010-06-15 2:46 ` [PATCH 2/6] KVM: MMU: fix conflict access permissions in direct sp Xiao Guangrong
[not found] ` <4C16E7AD.1060101@cn.fujitsu.com>
2010-06-15 2:46 ` [PATCH 3/6] KVM: MMU: introduce gfn_to_page_atomic() and gfn_to_pfn_atomic() Xiao Guangrong
2010-06-15 11:22 ` Avi Kivity
2010-06-16 7:59 ` Andi Kleen
2010-06-16 8:05 ` Avi Kivity
2010-06-16 8:49 ` Andi Kleen
2010-06-16 9:40 ` Avi Kivity
2010-06-16 10:51 ` huang ying
2010-06-16 11:08 ` Avi Kivity
2010-06-17 6:46 ` Xiao Guangrong
[not found] ` <4C16E7F4.5060801@cn.fujitsu.com>
2010-06-15 2:46 ` [PATCH 4/6] KVM: MMU: introduce mmu_topup_memory_cache_atomic() Xiao Guangrong
[not found] ` <4C16E82E.5010306@cn.fujitsu.com>
2010-06-15 2:47 ` Xiao Guangrong [this message]
2010-06-15 11:41 ` [PATCH 5/6] KVM: MMU: prefetch ptes when intercepted guest #PF Avi Kivity
2010-06-17 7:25 ` Xiao Guangrong
2010-06-17 8:07 ` Avi Kivity
2010-06-17 9:04 ` Xiao Guangrong
2010-06-17 9:20 ` Avi Kivity
2010-06-17 9:29 ` Xiao Guangrong
2010-06-16 3:55 ` Marcelo Tosatti
[not found] ` <4C16E867.8040606@cn.fujitsu.com>
2010-06-15 2:47 ` [PATCH 6/6] KVM: MMU: trace pte prefetch Xiao Guangrong
2010-06-15 12:09 ` Avi Kivity
2010-06-17 7:27 ` Xiao Guangrong
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=4C16E9A8.10409@cn.fujitsu.com \
--to=xiaoguangrong@cn.fujitsu.com \
--cc=avi@redhat.com \
--cc=kvm@vger.kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=mtosatti@redhat.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox