From: Marc Zyngier <marc.zyngier@arm.com>
To: Christoffer Dall <christoffer.dall@linaro.org>
Cc: "kvmarm@lists.cs.columbia.edu" <kvmarm@lists.cs.columbia.edu>,
"kvm@vger.kernel.org" <kvm@vger.kernel.org>
Subject: Re: [PATCH 2/3] arm/arm64: KVM: Implement Stage-2 page aging
Date: Thu, 12 Mar 2015 14:50:32 +0000 [thread overview]
Message-ID: <20150312145032.10c129b6@arm.com> (raw)
In-Reply-To: <20150312114017.GB28863@cbox>
On Thu, 12 Mar 2015 11:40:17 +0000
Christoffer Dall <christoffer.dall@linaro.org> wrote:
> On Wed, Jan 21, 2015 at 06:42:12PM +0000, Marc Zyngier wrote:
> > Until now, KVM/arm didn't care much for page aging (who was swapping
> > anyway?), and simply provided empty hooks to the core KVM code. With
> > server-type systems now being available, things are quite different.
> >
> > This patch implements very simple support for page aging, by
> > clearing the Access flag in the Stage-2 page tables. On access
> > fault, the current fault handling will write the PTE or PMD again,
> > putting the Access flag back on.
> >
> > It should be possible to implement a much faster handling for Access
> > faults, but that's left for a later patch.
> >
> > With this in place, performance in VMs is degraded much more
> > gracefully.
> >
> > Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
> > ---
> > arch/arm/include/asm/kvm_host.h | 13 ++-------
> > arch/arm/kvm/mmu.c | 59
> > ++++++++++++++++++++++++++++++++++++++-
> > arch/arm/kvm/trace.h | 33 ++++++++++++++++++++++
> > arch/arm64/include/asm/kvm_arm.h | 1 +
> > arch/arm64/include/asm/kvm_host.h | 13 ++------- 5 files changed,
> > 96 insertions(+), 23 deletions(-)
> >
> > diff --git a/arch/arm/include/asm/kvm_host.h
> > b/arch/arm/include/asm/kvm_host.h index 04b4ea0..d6b5b85 100644
> > --- a/arch/arm/include/asm/kvm_host.h
> > +++ b/arch/arm/include/asm/kvm_host.h
> > @@ -163,19 +163,10 @@ void kvm_set_spte_hva(struct kvm *kvm,
> > unsigned long hva, pte_t pte);
> > unsigned long kvm_arm_num_regs(struct kvm_vcpu *vcpu);
> > int kvm_arm_copy_reg_indices(struct kvm_vcpu *vcpu, u64 __user
> > *indices); +int kvm_age_hva(struct kvm *kvm, unsigned long start,
> > unsigned long end); +int kvm_test_age_hva(struct kvm *kvm, unsigned
> > long hva);
> > /* We do not have shadow page tables, hence the empty hooks */
> > -static inline int kvm_age_hva(struct kvm *kvm, unsigned long start,
> > - unsigned long end)
> > -{
> > - return 0;
> > -}
> > -
> > -static inline int kvm_test_age_hva(struct kvm *kvm, unsigned long
> > hva) -{
> > - return 0;
> > -}
> > -
> > static inline void kvm_arch_mmu_notifier_invalidate_page(struct
> > kvm *kvm, unsigned long address)
> > {
> > diff --git a/arch/arm/kvm/mmu.c b/arch/arm/kvm/mmu.c
> > index e163a45..ffe89a0 100644
> > --- a/arch/arm/kvm/mmu.c
> > +++ b/arch/arm/kvm/mmu.c
> > @@ -1068,6 +1068,7 @@ static int user_mem_abort(struct kvm_vcpu
> > *vcpu, phys_addr_t fault_ipa,
> > out_unlock:
> > spin_unlock(&kvm->mmu_lock);
> > + kvm_set_pfn_accessed(pfn);
> > kvm_release_pfn_clean(pfn);
> > return ret;
> > }
> > @@ -1102,7 +1103,8 @@ int kvm_handle_guest_abort(struct kvm_vcpu
> > *vcpu, struct kvm_run *run)
> > /* Check the stage-2 fault is trans. fault or write fault
> > */ fault_status = kvm_vcpu_trap_get_fault_type(vcpu);
> > - if (fault_status != FSC_FAULT && fault_status != FSC_PERM)
> > {
> > + if (fault_status != FSC_FAULT && fault_status != FSC_PERM
> > &&
> > + fault_status != FSC_ACCESS) {
> > kvm_err("Unsupported FSC: EC=%#x xFSC=%#lx
> > ESR_EL2=%#lx\n", kvm_vcpu_trap_get_class(vcpu),
> > (unsigned
> > long)kvm_vcpu_trap_get_fault(vcpu), @@ -1237,6 +1239,61 @@ void
> > kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte)
> > handle_hva_to_gpa(kvm, hva, end, &kvm_set_spte_handler,
> > &stage2_pte); }
> > +static int kvm_age_hva_handler(struct kvm *kvm, gpa_t gpa, void
> > *data) +{
> > + pmd_t *pmd;
> > + pte_t *pte;
> > +
> > + pmd = stage2_get_pmd(kvm, NULL, gpa);
> > + if (!pmd || pmd_none(*pmd)) /* Nothing there */
> > + return 0;
> > +
> > + if (kvm_pmd_huge(*pmd)) { /* THP, HugeTLB */
> > + *pmd = pmd_mkold(*pmd);
> > + goto tlbi;
>
> so in this case we'll loop over a huge pmd on a page-by-page basis,
> invalidating the tlb each time, right?
>
> Would it be worth checking of the access flag is already clear
> (!pmd_young()) and in that case exit without doing tlb invalidation?
>
> In fact, shouldn't we only return 1 if the pmd is indeed young and
> then the tlb invalidation will be done once by
> kvm_flush_remote_tlbs() in kvm_mmu_notifier_clear_flush_young() ?
>
> I got a little lost looking at how the core mm code uses the return
> value, but if I read the x86 and powerpc code correctly, they use it
> the way I suggest. Did I get this all wrong?
That's a very good point. I wonder if we could actually optimize
handle_hva_to_gpa to actually respect the mapping boundaries instead of
stupidly iterating over single pages, but that would be a further
improvement.
In the meantime, testing the access flag and doing an early out if
clear seems like the right thing to do.
> > + }
> > +
> > + pte = pte_offset_kernel(pmd, gpa);
> > + if (pte_none(*pte))
> > + return 0;
> > +
> > + *pte = pte_mkold(*pte); /* Just a page... */
>
> same with checking if it's young or not... ?
Yes, same logic.
> > +tlbi:
> > + kvm_tlb_flush_vmid_ipa(kvm, gpa);
> > + return 1;
> > +}
> > +
> > +static int kvm_test_age_hva_handler(struct kvm *kvm, gpa_t gpa,
> > void *data) +{
> > + pmd_t *pmd;
> > + pte_t *pte;
> > +
> > + pmd = stage2_get_pmd(kvm, NULL, gpa);
> > + if (!pmd || pmd_none(*pmd)) /* Nothing there */
> > + return 0;
> > +
> > + if (kvm_pmd_huge(*pmd)) /* THP, HugeTLB */
> > + return pmd_young(*pmd);
> > +
> > + pte = pte_offset_kernel(pmd, gpa);
> > + if (!pte_none(*pte)) /* Just a page... */
> > + return pte_young(*pte);
> > +
> > + return 0;
> > +}
> > +
> > +int kvm_age_hva(struct kvm *kvm, unsigned long start, unsigned
> > long end) +{
> > + trace_kvm_age_hva(start, end);
> > + return handle_hva_to_gpa(kvm, start, end,
> > kvm_age_hva_handler, NULL); +}
> > +
> > +int kvm_test_age_hva(struct kvm *kvm, unsigned long hva)
> > +{
> > + trace_kvm_test_age_hva(hva);
> > + return handle_hva_to_gpa(kvm, hva, hva,
> > kvm_test_age_hva_handler, NULL); +}
> > +
> > void kvm_mmu_free_memory_caches(struct kvm_vcpu *vcpu)
> > {
> > mmu_free_memory_cache(&vcpu->arch.mmu_page_cache);
> > diff --git a/arch/arm/kvm/trace.h b/arch/arm/kvm/trace.h
> > index b6a6e71..364b5382 100644
> > --- a/arch/arm/kvm/trace.h
> > +++ b/arch/arm/kvm/trace.h
> > @@ -203,6 +203,39 @@ TRACE_EVENT(kvm_set_spte_hva,
> > TP_printk("mmu notifier set pte hva: %#08lx", __entry->hva)
> > );
> >
> > +TRACE_EVENT(kvm_age_hva,
> > + TP_PROTO(unsigned long start, unsigned long end),
> > + TP_ARGS(start, end),
> > +
> > + TP_STRUCT__entry(
> > + __field( unsigned long,
> > start )
> > + __field( unsigned long,
> > end )
> > + ),
> > +
> > + TP_fast_assign(
> > + __entry->start = start;
> > + __entry->end = end;
> > + ),
> > +
> > + TP_printk("mmu notifier age hva: %#08lx -- %#08lx",
> > + __entry->start, __entry->end)
> > +);
> > +
> > +TRACE_EVENT(kvm_test_age_hva,
> > + TP_PROTO(unsigned long hva),
> > + TP_ARGS(hva),
> > +
> > + TP_STRUCT__entry(
> > + __field( unsigned long,
> > hva )
> > + ),
> > +
> > + TP_fast_assign(
> > + __entry->hva = hva;
> > + ),
> > +
> > + TP_printk("mmu notifier test age hva: %#08lx",
> > __entry->hva) +);
> > +
> > TRACE_EVENT(kvm_hvc,
> > TP_PROTO(unsigned long vcpu_pc, unsigned long r0, unsigned
> > long imm), TP_ARGS(vcpu_pc, r0, imm),
> > diff --git a/arch/arm64/include/asm/kvm_arm.h
> > b/arch/arm64/include/asm/kvm_arm.h index 8afb863..0d738f2 100644
> > --- a/arch/arm64/include/asm/kvm_arm.h
> > +++ b/arch/arm64/include/asm/kvm_arm.h
> > @@ -212,6 +212,7 @@
> >
> >
> > #define FSC_FAULT (0x04)
> > +#define FSC_ACCESS (0x08)
> > #define FSC_PERM (0x0c)
> >
> > /* Hyp Prefetch Fault Address Register (HPFAR/HDFAR) */
> > diff --git a/arch/arm64/include/asm/kvm_host.h
> > b/arch/arm64/include/asm/kvm_host.h index acd101a..b831710 100644
> > --- a/arch/arm64/include/asm/kvm_host.h
> > +++ b/arch/arm64/include/asm/kvm_host.h
> > @@ -173,19 +173,10 @@ int kvm_unmap_hva(struct kvm *kvm, unsigned
> > long hva); int kvm_unmap_hva_range(struct kvm *kvm,
> > unsigned long start, unsigned long end);
> > void kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t
> > pte); +int kvm_age_hva(struct kvm *kvm, unsigned long start,
> > unsigned long end); +int kvm_test_age_hva(struct kvm *kvm, unsigned
> > long hva);
> > /* We do not have shadow page tables, hence the empty hooks */
> > -static inline int kvm_age_hva(struct kvm *kvm, unsigned long start,
> > - unsigned long end)
> > -{
> > - return 0;
> > -}
> > -
> > -static inline int kvm_test_age_hva(struct kvm *kvm, unsigned long
> > hva) -{
> > - return 0;
> > -}
> > -
> > static inline void kvm_arch_mmu_notifier_invalidate_page(struct
> > kvm *kvm, unsigned long address)
> > {
> > --
> > 2.1.4
> >
> Otherwise, this looks good.
Thanks for having had a look!
M.
--
Jazz is not dead. It just smells funny.
next prev parent reply other threads:[~2015-03-12 14:50 UTC|newest]
Thread overview: 10+ messages / expand[flat|nested] mbox.gz Atom feed top
2015-01-21 18:42 [PATCH 0/3] arm/arm64: KVM: Add support for page aging Marc Zyngier
2015-01-21 18:42 ` [PATCH 1/3] arm/arm64: KVM: Allow handle_hva_to_gpa to return a value Marc Zyngier
2015-03-12 11:40 ` Christoffer Dall
2015-01-21 18:42 ` [PATCH 2/3] arm/arm64: KVM: Implement Stage-2 page aging Marc Zyngier
2015-03-12 11:40 ` Christoffer Dall
2015-03-12 14:50 ` Marc Zyngier [this message]
2015-01-21 18:42 ` [PATCH 3/3] arm/arm64: KVM: Optimize handling of Access Flag faults Marc Zyngier
2015-03-12 11:40 ` Christoffer Dall
2015-03-12 15:04 ` Marc Zyngier
2015-03-12 15:07 ` Marc Zyngier
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20150312145032.10c129b6@arm.com \
--to=marc.zyngier@arm.com \
--cc=christoffer.dall@linaro.org \
--cc=kvm@vger.kernel.org \
--cc=kvmarm@lists.cs.columbia.edu \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox