From: "Herongguang (Stephen)" <herongguang.he@huawei.com>
To: Alexey G <x1917x@gmail.com>, Stefano Stabellini <sstabellini@kernel.org>
Cc: hrg <hrgstephen@gmail.com>,
wangxinxin.wang@huawei.com, Alexander Graf <agraf@suse.de>,
qemu-devel@nongnu.org, xen-devel@lists.xensource.com,
Jun Nakajima <jun.nakajima@intel.com>,
Anthony PERARD <anthony.perard@citrix.com>,
xen-devel@lists.xenproject.org, xen-devel@lists.xen.org
Subject: Re: [Qemu-devel] [Xen-devel] [RFC/BUG] xen-mapcache: buggy invalidate map cache?
Date: Wed, 12 Apr 2017 16:28:21 +0800 [thread overview]
Message-ID: <58EDE525.4050905@huawei.com> (raw)
In-Reply-To: <20170412161704.00003875@gmail.com>
On 2017/4/12 14:17, Alexey G wrote:
> On Tue, 11 Apr 2017 15:32:09 -0700 (PDT)
> Stefano Stabellini <sstabellini@kernel.org> wrote:
>
>> On Tue, 11 Apr 2017, hrg wrote:
>>> On Tue, Apr 11, 2017 at 3:50 AM, Stefano Stabellini
>>> <sstabellini@kernel.org> wrote:
>>>> On Mon, 10 Apr 2017, Stefano Stabellini wrote:
>>>>> On Mon, 10 Apr 2017, hrg wrote:
>>>>>> On Sun, Apr 9, 2017 at 11:55 PM, hrg <hrgstephen@gmail.com> wrote:
>>>>>>> On Sun, Apr 9, 2017 at 11:52 PM, hrg <hrgstephen@gmail.com> wrote:
>>>>>>>> Hi,
>>>>>>>>
>>>>>>>> In xen_map_cache_unlocked(), map to guest memory maybe in
>>>>>>>> entry->next instead of first level entry (if map to rom other than
>>>>>>>> guest memory comes first), while in xen_invalidate_map_cache(),
>>>>>>>> when VM ballooned out memory, qemu did not invalidate cache entries
>>>>>>>> in linked list(entry->next), so when VM balloon back in memory,
>>>>>>>> gfns probably mapped to different mfns, thus if guest asks device
>>>>>>>> to DMA to these GPA, qemu may DMA to stale MFNs.
>>>>>>>>
>>>>>>>> So I think in xen_invalidate_map_cache() linked lists should also be
>>>>>>>> checked and invalidated.
>>>>>>>>
>>>>>>>> What’s your opinion? Is this a bug? Is my analyze correct?
>>>>>
>>>>> Yes, you are right. We need to go through the list for each element of
>>>>> the array in xen_invalidate_map_cache. Can you come up with a patch?
>>>>
>>>> I spoke too soon. In the regular case there should be no locked mappings
>>>> when xen_invalidate_map_cache is called (see the DPRINTF warning at the
>>>> beginning of the functions). Without locked mappings, there should never
>>>> be more than one element in each list (see xen_map_cache_unlocked:
>>>> entry->lock == true is a necessary condition to append a new entry to
>>>> the list, otherwise it is just remapped).
>>>>
>>>> Can you confirm that what you are seeing are locked mappings
>>>> when xen_invalidate_map_cache is called? To find out, enable the DPRINTK
>>>> by turning it into a printf or by defininig MAPCACHE_DEBUG.
>>>
>>> In fact, I think the DPRINTF above is incorrect too. In
>>> pci_add_option_rom(), rtl8139 rom is locked mapped in
>>> pci_add_option_rom->memory_region_get_ram_ptr (after
>>> memory_region_init_ram). So actually I think we should remove the
>>> DPRINTF warning as it is normal.
>>
>> Let me explain why the DPRINTF warning is there: emulated dma operations
>> can involve locked mappings. Once a dma operation completes, the related
>> mapping is unlocked and can be safely destroyed. But if we destroy a
>> locked mapping in xen_invalidate_map_cache, while a dma is still
>> ongoing, QEMU will crash. We cannot handle that case.
>>
>> However, the scenario you described is different. It has nothing to do
>> with DMA. It looks like pci_add_option_rom calls
>> memory_region_get_ram_ptr to map the rtl8139 rom. The mapping is a
>> locked mapping and it is never unlocked or destroyed.
>>
>> It looks like "ptr" is not used after pci_add_option_rom returns. Does
>> the append patch fix the problem you are seeing? For the proper fix, I
>> think we probably need some sort of memory_region_unmap wrapper or maybe
>> a call to address_space_unmap.
>
> Hmm, for some reason my message to the Xen-devel list got rejected but was sent
> to Qemu-devel instead, without any notice. Sorry if I'm missing something
> obvious as a list newbie.
>
> Stefano, hrg,
>
> There is an issue with inconsistency between the list of normal MapCacheEntry's
> and their 'reverse' counterparts - MapCacheRev's in locked_entries.
> When bad situation happens, there are multiple (locked) MapCacheEntry
> entries in the bucket's linked list along with a number of MapCacheRev's. And
> when it comes to a reverse lookup, xen-mapcache picks the wrong entry from the
> first list and calculates a wrong pointer from it which may then be caught with
> the "Bad RAM offset" check (or not). Mapcache invalidation might be related to
> this issue as well I think.
>
> I'll try to provide a test code which can reproduce the issue from the
> guest side using an emulated IDE controller, though it's much simpler to achieve
> this result with an AHCI controller using multiple NCQ I/O commands. So far I've
> seen this issue only with Windows 7 (and above) guest on AHCI, but any block I/O
> DMA should be enough I think.
>
Yes, I think there may be other bugs lurking, considering the complexity, though we need to reproduce it if we want to delve into it.
next prev parent reply other threads:[~2017-04-12 8:29 UTC|newest]
Thread overview: 15+ messages / expand[flat|nested] mbox.gz Atom feed top
2017-04-09 15:52 [Qemu-devel] [RFC/BUG] xen-mapcache: buggy invalidate map cache? hrg
2017-04-09 15:55 ` hrg
2017-04-09 16:36 ` hrg
2017-04-09 17:52 ` [Qemu-devel] [Xen-devel] " Alexey G
2017-04-10 19:04 ` [Qemu-devel] " Stefano Stabellini
2017-04-10 19:50 ` Stefano Stabellini
2017-04-11 4:47 ` hrg
2017-04-11 22:32 ` Stefano Stabellini
2017-04-12 6:17 ` [Qemu-devel] [Xen-devel] " Alexey G
2017-04-12 8:28 ` Herongguang (Stephen) [this message]
2017-04-12 23:51 ` Stefano Stabellini
2017-04-12 8:14 ` [Qemu-devel] " Herongguang (Stephen)
2017-04-12 23:51 ` Stefano Stabellini
2017-04-13 5:47 ` Herongguang (Stephen)
2017-04-28 23:51 ` Stefano Stabellini
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=58EDE525.4050905@huawei.com \
--to=herongguang.he@huawei.com \
--cc=agraf@suse.de \
--cc=anthony.perard@citrix.com \
--cc=hrgstephen@gmail.com \
--cc=jun.nakajima@intel.com \
--cc=qemu-devel@nongnu.org \
--cc=sstabellini@kernel.org \
--cc=wangxinxin.wang@huawei.com \
--cc=x1917x@gmail.com \
--cc=xen-devel@lists.xen.org \
--cc=xen-devel@lists.xenproject.org \
--cc=xen-devel@lists.xensource.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).