From: David Hildenbrand <david@redhat.com>
To: "david.dai" <david.dai@montage-tech.com>
Cc: peter.maydell@linaro.org, vsementsov@virtuozzo.com,
eajames@linux.ibm.com, qemu-devel@nongnu.org,
changguo.du@montage-tech.com,
Stefan Hajnoczi <stefanha@redhat.com>,
Igor Mammedov <imammedo@redhat.com>,
kuhn.chenqun@huawei.com
Subject: Re: [PATCH] hw/misc: Add a virtual pci device to dynamically attach memory to QEMU
Date: Wed, 29 Sep 2021 11:30:53 +0200 [thread overview]
Message-ID: <c87c301e-62af-ab5a-2b9c-fa2ef28898f1@redhat.com> (raw)
In-Reply-To: <20210927122848.GB144947@tianmu-host-sw-01>
On 27.09.21 14:28, david.dai wrote:
> On Mon, Sep 27, 2021 at 11:07:43AM +0200, David Hildenbrand (david@redhat.com) wrote:
>>
>> CAUTION: This email originated from outside of the organization. Do not
>> click links or open attachments unless you recognize the sender and know the
>> content is safe.
>>
>>
>> On 27.09.21 10:27, Stefan Hajnoczi wrote:
>>> On Sun, Sep 26, 2021 at 10:16:14AM +0800, David Dai wrote:
>>>> Add a virtual pci to QEMU, the pci device is used to dynamically attach memory
>>>> to VM, so driver in guest can apply host memory in fly without virtualization
>>>> management software's help, such as libvirt/manager. The attached memory is
>>
>> We do have virtio-mem to dynamically attach memory to a VM. It could be
>> extended by a mechanism for the VM to request more/less memory, that's
>> already a planned feature. But yeah, virito-mem memory is exposed as
>> ordinary system RAM, not only via a BAR to mostly be managed by user space
>> completely.
There is a virtio-pmem spec proposal to expose the memory region via a
PCI BAR. We could do something similar for virtio-mem, however, we would
have to wire that new model up differently in QEMU (it's no longer a
"memory device" like a DIMM then).
>>
>
> I wish virtio-mem can solve our problem, but it is a dynamic allocation mechanism
> for system RAM in virtualization. In heterogeneous computing environments, the
> attached memory usually comes from computing device, it should be managed separately.
> we doesn't hope Linux MM controls it.
If that heterogeneous memory would have a dedicated node (which usually
is the case IIRC) , and you let it manage by the Linux kernel
(dax/kmem), you can bind the memory backend of virtio-mem to that
special NUMA node. So all memory managed by that virtio-mem device would
come from that heterogeneous memory.
You could then further use a separate NUMA node for that virtio-mem
device inside the VM. But to the VM it would look like System memory
with different performance characteristics. That would work fore some
use cases I guess, but not sure for which not (I assume you can tell :) ).
We could even write an alternative virtio-mem mode, where device manage
isn't exposed to the buddy but using some different way to user space.
>
>>>> isolated from System RAM, it can be used in heterogeneous memory management for
>>>> virtualization. Multiple VMs dynamically share same computing device memory
>>>> without memory overcommit.
>>
>> This sounds a lot like MemExpand/MemLego ... am I right that this is the
>> original design? I recall that VMs share a memory region and dynamically
>> agree upon which part of the memory region a VM uses. I further recall that
>> there were malloc() hooks that would dynamically allocate such memory in
>> user space from the shared memory region.
>>
>
> Thank you for telling me about Memexpand/MemLego, I have carefully read the paper.
> some ideas from it are same as this patch, such as software model and stack, but
> it may have a security risk that whole shared memory is visible to all VMs.
How will you make sure that not all shared memory can be accessed by the
other VMs? IOW, emulate !shared memory on shared memory?
> -----------------------
> application
> -----------------------
> memory management driver
> -----------------------
> pci driver
> -----------------------
> virtual pci device
> -----------------------
>
>> I can see some use cases for it, although the shared memory design isn't
>> what you typically want in most VM environments.
>>
>
> The original design for this patch is to share a computing device among multipile
> VMs. Each VM runs a computing application(for example, OpenCL application)
> Our computing device can support a few applications in parallel. In addition, it
> supports SVM(shared virtual memory) via IOMMU/ATS/PASID/PRI. Device exposes its
> memory to host vis PCIe bar or CXL.mem, host constructs memory pool to uniformly
> manage device memory, then attach device memory to VM via a virtual PCI device.
How exactly is that memory pool created/managed? Simply dax/kmem and
handling it via the buddy in a special NUMA node.
> but we don't know how much memory should be assigned when creating VM, so we hope
> memory is attached to VM on-demand. driver in guest triggers memory attaching, but
> not outside virtualization management software. so the original requirements are:
> 1> The managed memory comes from device, it should be isolated from system RAM
> 2> The memory can be dynamically attached to VM in fly
> 3> The attached memory supports SVM and DMA operation with IOMMU
>
> Thank you very much.
Thanks for the info. If virtio-mem is not applicable and cannot be
modified for this use case, would it make sense to create a new virtio
device type?
--
Thanks,
David / dhildenb
next prev parent reply other threads:[~2021-09-29 9:35 UTC|newest]
Thread overview: 15+ messages / expand[flat|nested] mbox.gz Atom feed top
2021-09-26 2:16 [PATCH] hw/misc: Add a virtual pci device to dynamically attach memory to QEMU David Dai
2021-09-27 8:27 ` Stefan Hajnoczi
2021-09-27 9:07 ` David Hildenbrand
2021-09-27 12:28 ` david.dai
2021-09-29 9:30 ` David Hildenbrand [this message]
2021-09-30 9:40 ` david.dai
2021-09-30 10:33 ` David Hildenbrand
2021-10-09 9:42 ` david.dai
2021-10-11 7:43 ` David Hildenbrand
2021-10-13 8:13 ` david.dai
2021-10-13 8:33 ` David Hildenbrand
2021-10-15 9:10 ` david.dai
2021-10-15 9:27 ` David Hildenbrand
2021-10-15 9:57 ` david.dai
2021-09-27 12:17 ` david.dai
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=c87c301e-62af-ab5a-2b9c-fa2ef28898f1@redhat.com \
--to=david@redhat.com \
--cc=changguo.du@montage-tech.com \
--cc=david.dai@montage-tech.com \
--cc=eajames@linux.ibm.com \
--cc=imammedo@redhat.com \
--cc=kuhn.chenqun@huawei.com \
--cc=peter.maydell@linaro.org \
--cc=qemu-devel@nongnu.org \
--cc=stefanha@redhat.com \
--cc=vsementsov@virtuozzo.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).