From: Paolo Bonzini <pbonzini@redhat.com>
To: Sean Christopherson <seanjc@google.com>
Cc: Emanuele Giuseppe Esposito <eesposit@redhat.com>,
David Hildenbrand <david@redhat.com>,
Maxim Levitsky <mlevitsk@redhat.com>,
kvm@vger.kernel.org, Vitaly Kuznetsov <vkuznets@redhat.com>,
Wanpeng Li <wanpengli@tencent.com>,
Jim Mattson <jmattson@google.com>, Joerg Roedel <joro@8bytes.org>,
Thomas Gleixner <tglx@linutronix.de>,
Ingo Molnar <mingo@redhat.com>, Borislav Petkov <bp@alien8.de>,
Dave Hansen <dave.hansen@linux.intel.com>,
x86@kernel.org, "H. Peter Anvin" <hpa@zytor.com>,
linux-kernel@vger.kernel.org, Like Xu <like.xu.linux@gmail.com>
Subject: Re: [RFC PATCH 0/9] kvm: implement atomic memslot updates
Date: Wed, 28 Sep 2022 18:38:00 +0200 [thread overview]
Message-ID: <8534dfe4-bc71-2c14-b268-e610a3111d14@redhat.com> (raw)
In-Reply-To: <YzRvMZDoukMbeaxR@google.com>
On 9/28/22 17:58, Sean Christopherson wrote:
> On Wed, Sep 28, 2022, Paolo Bonzini wrote:
>> On 9/27/22 17:58, Sean Christopherson wrote:
>>> I'm pretty sure this patch will Just Work for QEMU, because QEMU simply resumes
>>> the vCPU if mmio.len==0. It's a bit of a hack, but I don't think it violates KVM's
>>> ABI in any way, and it can even become "official" behavior since KVM x86 doesn't
>>> otherwise exit with mmio.len==0.
>>
>> I think this patch is not a good idea for two reasons:
>>
>> 1) we don't know how userspace behaves if mmio.len is zero. It is of course
>> reasonable to do nothing, but an assertion failure is also a valid behavior
>
> Except that KVM currently does neither. If the fetch happens at CPL>0 and/or in
> L2, KVM injects #UD. That's flat out architecturally invalid. If it's a sticking
> point, the mmio.len==0 hack can be avoided by defining a new exit reason.
I agree that doing this at CPL>0 or in L2 is invalid and makes little
sense (because either way the invalid address cannot be reached without
help from the supervisor or L1's page tables).
>> 2) more important, there is no way to distinguish a failure due to the guest
>> going in the weeds (and then KVM_EXIT_INTERNAL_ERROR is fine) from one due
>> to the KVM_SET_USER_MEMORY_REGION race condition. So this will cause a
>> guest that correctly caused an internal error to loop forever.
>
> Userspace has the GPA and absolutely should be able to detect if the MMIO may have
> been due to its memslot manipulation versus the guest jumping into the weeds.
>
>> While the former could be handled in a "wait and see" manner, the latter in
>> particular is part of the KVM_RUN contract. Of course it is possible for a
>> guest to just loop forever, but in general all of KVM, QEMU and upper
>> userspace layers want a crashed guest to be detected and stopped forever.
>>
>> Yes, QEMU could loop only if memslot updates are in progress, but honestly
>> all the alternatives I have seen to atomic memslot updates are really
>> *awful*. David's patches even invent a new kind of mutex for which I have
>> absolutely no idea what kind of deadlocks one should worry about and why
>> they should not exist; QEMU's locking is already pretty crappy, it's
>> certainly not on my wishlist to make it worse!
>>
>> This is clearly a deficiency in the KVM kernel API, and (thanks to SRCU) the
>> kernel is the only place where you can have a *good* fix. It should have
>> been fixed years ago.
>
> I don't disagree that the memslots API is lacking, but IMO that is somewhat
> orthogonal to fixing KVM x86's "code fetch to MMIO" mess. Such a massive new API
> should be viewed and prioritized as a new feature, not as a bug fix, e.g. I'd
> like to have the luxury of being able to explore ideas beyond "let userspace
> batch memslot updates", and I really don't want to feel pressured to get this
> code reviewed and merge.
I absolutely agree that this is not a bugfix. Most new features for KVM
can be seen as bug fixes if you squint hard enough, but they're still
features.
> E.g. why do a batch update and not provide KVM_SET_ALL_USER_MEMORY_REGIONS to
> do wholesale replacement? That seems like it would be vastly simpler to handle
> on KVM's end. Or maybe there's a solution in the opposite direction, e.g. an
> API that allows 1->N or N->1 conversions but not arbitrary batching.
Wholesale replacement was my first idea when I looked at the issue, I
think at the end of 2020. I never got to a full implementation, but my
impression was that allocating/deallocating dirty bitmaps, rmaps etc.
would make it any easier than arbitrary batch updates.
> And just because QEMU's locking is "already pretty crappy", that's not a good
> reason to drag KVM down into the mud. E.g. taking a lock and conditionally
> releasing it... I get that this is an RFC, but IMO anything that requires such
> shenanigans simply isn't acceptable.
>
> /*
> * Takes kvm->slots_arch_lock, and releases it only if
> * invalid_slot allocation, kvm_prepare_memory_region failed
> * or batch->is_move_delete is true.
> */
> static int kvm_prepare_memslot(struct kvm *kvm,
> struct kvm_internal_memory_region_list *batch)
>
No objection about that. :)
Paolo
next prev parent reply other threads:[~2022-09-28 16:38 UTC|newest]
Thread overview: 58+ messages / expand[flat|nested] mbox.gz Atom feed top
2022-09-09 10:44 [RFC PATCH 0/9] kvm: implement atomic memslot updates Emanuele Giuseppe Esposito
2022-09-09 10:44 ` [RFC PATCH 1/9] kvm_main.c: move slot check in kvm_set_memory_region Emanuele Giuseppe Esposito
2022-09-28 16:41 ` Paolo Bonzini
2022-09-09 10:44 ` [RFC PATCH 2/9] kvm.h: introduce KVM_SET_USER_MEMORY_REGION_LIST ioctl Emanuele Giuseppe Esposito
2022-09-28 16:42 ` Paolo Bonzini
2022-09-09 10:45 ` [RFC PATCH 3/9] kvm_main.c: introduce kvm_internal_memory_region_list Emanuele Giuseppe Esposito
2022-09-28 16:48 ` Paolo Bonzini
2022-09-09 10:45 ` [RFC PATCH 4/9] kvm_main.c: split logic in kvm_set_memslots Emanuele Giuseppe Esposito
2022-09-28 17:04 ` Paolo Bonzini
2022-09-09 10:45 ` [RFC PATCH 5/9] kvm_main.c: split __kvm_set_memory_region logic in kvm_check_mem and kvm_prepare_batch Emanuele Giuseppe Esposito
2022-09-13 2:56 ` Yang, Weijiang
2022-09-18 16:22 ` Emanuele Giuseppe Esposito
2022-09-28 17:11 ` Paolo Bonzini
2022-09-09 10:45 ` [RFC PATCH 6/9] kvm_main.c: simplify change-specific callbacks Emanuele Giuseppe Esposito
2022-09-09 10:45 ` [RFC PATCH 7/9] kvm_main.c: duplicate invalid memslot also in inactive list Emanuele Giuseppe Esposito
2022-09-28 17:18 ` Paolo Bonzini
2022-09-09 10:45 ` [RFC PATCH 8/9] kvm_main.c: find memslots from the inactive memslot list Emanuele Giuseppe Esposito
2022-09-09 10:45 ` [RFC PATCH 9/9] kvm_main.c: handle atomic memslot update Emanuele Giuseppe Esposito
2022-09-13 2:30 ` Yang, Weijiang
2022-09-18 16:18 ` Emanuele Giuseppe Esposito
2022-09-27 7:46 ` David Hildenbrand
2022-09-27 8:35 ` Emanuele Giuseppe Esposito
2022-09-27 9:22 ` David Hildenbrand
2022-09-27 9:32 ` Emanuele Giuseppe Esposito
2022-09-27 14:52 ` David Hildenbrand
2022-09-28 17:29 ` Paolo Bonzini
2022-09-09 14:30 ` [RFC PATCH 0/9] kvm: implement atomic memslot updates Sean Christopherson
2022-09-18 16:13 ` Emanuele Giuseppe Esposito
2022-09-19 7:38 ` Like Xu
2022-09-19 7:53 ` David Hildenbrand
2022-09-19 17:30 ` David Hildenbrand
2022-09-23 13:10 ` Emanuele Giuseppe Esposito
2022-09-23 13:21 ` David Hildenbrand
2022-09-23 13:38 ` Emanuele Giuseppe Esposito
2022-09-26 9:03 ` David Hildenbrand
2022-09-26 21:28 ` Sean Christopherson
2022-09-27 7:38 ` Emanuele Giuseppe Esposito
2022-09-27 15:58 ` Sean Christopherson
2022-09-28 9:11 ` Emanuele Giuseppe Esposito
2022-09-28 11:14 ` Maxim Levitsky
2022-09-28 12:52 ` David Hildenbrand
2022-09-28 15:07 ` Paolo Bonzini
2022-09-28 15:33 ` David Hildenbrand
2022-09-28 15:58 ` Sean Christopherson
2022-09-28 16:38 ` Paolo Bonzini [this message]
2022-09-28 20:41 ` Sean Christopherson
2022-09-29 8:05 ` Emanuele Giuseppe Esposito
2022-09-29 8:24 ` David Hildenbrand
2022-09-29 15:18 ` Sean Christopherson
2022-09-29 15:41 ` Paolo Bonzini
2022-09-29 15:28 ` Paolo Bonzini
2022-09-29 15:40 ` Maxim Levitsky
2022-09-29 16:00 ` David Hildenbrand
2022-09-29 21:39 ` Sean Christopherson
2022-10-13 7:43 ` Emanuele Giuseppe Esposito
2022-10-13 8:44 ` David Hildenbrand
2022-10-13 11:12 ` Emanuele Giuseppe Esposito
2022-10-13 14:45 ` David Hildenbrand
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=8534dfe4-bc71-2c14-b268-e610a3111d14@redhat.com \
--to=pbonzini@redhat.com \
--cc=bp@alien8.de \
--cc=dave.hansen@linux.intel.com \
--cc=david@redhat.com \
--cc=eesposit@redhat.com \
--cc=hpa@zytor.com \
--cc=jmattson@google.com \
--cc=joro@8bytes.org \
--cc=kvm@vger.kernel.org \
--cc=like.xu.linux@gmail.com \
--cc=linux-kernel@vger.kernel.org \
--cc=mingo@redhat.com \
--cc=mlevitsk@redhat.com \
--cc=seanjc@google.com \
--cc=tglx@linutronix.de \
--cc=vkuznets@redhat.com \
--cc=wanpengli@tencent.com \
--cc=x86@kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox