From: Xiao Guangrong <xiaoguangrong@cn.fujitsu.com>
To: Avi Kivity <avi@redhat.com>
Cc: Marcelo Tosatti <mtosatti@redhat.com>,
LKML <linux-kernel@vger.kernel.org>,
KVM list <kvm@vger.kernel.org>
Subject: Re: [PATCH 5/6] KVM: MMU: prefetch ptes when intercepted guest #PF
Date: Thu, 17 Jun 2010 17:04:50 +0800 [thread overview]
Message-ID: <4C19E532.4070708@cn.fujitsu.com> (raw)
In-Reply-To: <4C19D7B1.6060908@redhat.com>
Avi Kivity wrote:
> On 06/17/2010 10:25 AM, Xiao Guangrong wrote:
>>
>>
>>> Can this in fact work for level != PT_PAGE_TABLE_LEVEL? We might start
>>> at PT_PAGE_DIRECTORY_LEVEL but get 4k pages while iterating.
>>>
>> Ah, i forgot it. We can't assume that the host also support huge page for
>> next gfn, as Marcelo's suggestion, we should "only map with level> 1 if
>> the host page matches the size".
>>
>> Um, the problem is, when we get host page size, we should hold
>> 'mm->mmap_sem',
>> it can't used in atomic context and it's also a slow path, we hope pte
>> prefetch
>> path is fast.
>>
>> How about only allow prefetch for sp.leve = 1 now? i'll improve it in
>> the future,
>> i think it need more time :-)
>>
>
> I don't think prefetch for level > 1 is worthwhile. One fault per 2MB
> is already very good, no need to optimize it further.
>
OK
>>>> +
>>>> + pfn = gfn_to_pfn_atomic(vcpu->kvm, gfn);
>>>> + if (is_error_pfn(pfn)) {
>>>> + kvm_release_pfn_clean(pfn);
>>>> + break;
>>>> + }
>>>> + if (pte_prefetch_topup_memory_cache(vcpu))
>>>> + break;
>>>> +
>>>> + mmu_set_spte(vcpu, spte, ACC_ALL, ACC_ALL, 0, 0, 1, NULL,
>>>> + sp->role.level, gfn, pfn, true, false);
>>>> + }
>>>> +}
>>>>
>>>>
>>> Nice. Direct prefetch should usually succeed.
>>>
>>> Can later augment to call get_users_pages_fast(..., PTE_PREFETCH_NUM,
>>> ...) to reduce gup overhead.
>>>
>> But we can't assume the gfn's hva is consecutive, for example, gfn and
>> gfn+1
>> maybe in the different slots.
>>
>
> Right. We could limit it to one slot then for simplicity.
OK, i'll do it.
>
>>
>>>> +
>>>> + if (!table) {
>>>> + page = gfn_to_page_atomic(vcpu->kvm, sp->gfn);
>>>> + if (is_error_page(page)) {
>>>> + kvm_release_page_clean(page);
>>>> + break;
>>>> + }
>>>> + table = kmap_atomic(page, KM_USER0);
>>>> + table = (pt_element_t *)((char *)table + offset);
>>>> + }
>>>>
>>>>
>>> Why not kvm_read_guest_atomic()? Can do it outside the loop.
>>>
>> Do you mean that read all prefetched sptes at one time?
>>
>
> Yes.
>
>> If prefetch one spte fail, the later sptes that we read is waste, so i
>> choose read next spte only when current spte is prefetched successful.
>>
>> But i not have strong opinion on it since it's fast to read all sptes at
>> one time, at the worst case, only 16 * 8 = 128 bytes we need to read.
>>
>
> In general batching is worthwhile, the cost of the extra bytes is low
> compared to the cost of bringing in the cacheline and error checking.
>
Agreed.
> btw, you could align the prefetch to 16 pte boundary. That would
> improve performance for memory that is scanned backwards.
>
Yeah, good idea.
> So we can change the fault path to always fault 16 ptes, aligned on 16
> pte boundary, with the needed pte called with specualtive=false.
Avi, i not understand it clearly, Could you please explain it? :-(
next prev parent reply other threads:[~2010-06-17 9:08 UTC|newest]
Thread overview: 23+ messages / expand[flat|nested] mbox.gz Atom feed top
[not found] <4C16E6ED.7020009@cn.fujitsu.com>
2010-06-15 2:46 ` [PATCH 1/6] KVM: MMU: fix gfn got in kvm_mmu_page_get_gfn() Xiao Guangrong
[not found] ` <4C16E75F.6020003@cn.fujitsu.com>
2010-06-15 2:46 ` [PATCH 2/6] KVM: MMU: fix conflict access permissions in direct sp Xiao Guangrong
[not found] ` <4C16E7AD.1060101@cn.fujitsu.com>
2010-06-15 2:46 ` [PATCH 3/6] KVM: MMU: introduce gfn_to_page_atomic() and gfn_to_pfn_atomic() Xiao Guangrong
2010-06-15 11:22 ` Avi Kivity
2010-06-16 7:59 ` Andi Kleen
2010-06-16 8:05 ` Avi Kivity
2010-06-16 8:49 ` Andi Kleen
2010-06-16 9:40 ` Avi Kivity
2010-06-16 10:51 ` huang ying
2010-06-16 11:08 ` Avi Kivity
2010-06-17 6:46 ` Xiao Guangrong
[not found] ` <4C16E7F4.5060801@cn.fujitsu.com>
2010-06-15 2:46 ` [PATCH 4/6] KVM: MMU: introduce mmu_topup_memory_cache_atomic() Xiao Guangrong
[not found] ` <4C16E82E.5010306@cn.fujitsu.com>
2010-06-15 2:47 ` [PATCH 5/6] KVM: MMU: prefetch ptes when intercepted guest #PF Xiao Guangrong
2010-06-15 11:41 ` Avi Kivity
2010-06-17 7:25 ` Xiao Guangrong
2010-06-17 8:07 ` Avi Kivity
2010-06-17 9:04 ` Xiao Guangrong [this message]
2010-06-17 9:20 ` Avi Kivity
2010-06-17 9:29 ` Xiao Guangrong
2010-06-16 3:55 ` Marcelo Tosatti
[not found] ` <4C16E867.8040606@cn.fujitsu.com>
2010-06-15 2:47 ` [PATCH 6/6] KVM: MMU: trace pte prefetch Xiao Guangrong
2010-06-15 12:09 ` Avi Kivity
2010-06-17 7:27 ` Xiao Guangrong
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=4C19E532.4070708@cn.fujitsu.com \
--to=xiaoguangrong@cn.fujitsu.com \
--cc=avi@redhat.com \
--cc=kvm@vger.kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=mtosatti@redhat.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox