From: Jan Beulich <jbeulich@suse.com>
To: Teddy Astie <teddy.astie@vates.tech>,
Demi Marie Obenour <demiobenour@gmail.com>
Cc: Xen developer discussion <xen-devel@lists.xenproject.org>,
dri-devel@lists.freedesktop.org, linux-mm@kvack.org,
Val Packett <val@invisiblethingslab.com>,
Ariadne Conill <ariadne@ariadne.space>,
Andrew Cooper <andrew.cooper3@citrix.com>,
Juergen Gross <jgross@suse.com>
Subject: Re: Why memory lending is needed for GPU acceleration
Date: Mon, 30 Mar 2026 12:25:25 +0200 [thread overview]
Message-ID: <cf54e793-63d6-417c-b7c5-3301455fcf33@suse.com> (raw)
In-Reply-To: <ce114f02-f45e-4638-84ee-a8fd86ce1c5d@vates.tech>
On 30.03.2026 12:15, Teddy Astie wrote:
> Le 29/03/2026 à 19:32, Demi Marie Obenour a écrit :
May I ask that the two of you please properly separate To: vs Cc:?
Thanks, Jan
>> On 3/24/26 10:17, Demi Marie Obenour wrote:
>>> Here is a proposed design document for supporting mapping GPU VRAM
>>> and/or file-backed memory into other domains. It's not in the form of
>>> a patch because the leading + characters would just make it harder to
>>> read for no particular gain, and because this is still RFC right now.
>>> Once it is ready to merge, I'll send a proper patch. Nevertheless,
>>> you can consider this to be
>>>
>>> Signed-off-by: Demi Marie Obenour <demiobenour@gmail.com>
>>>
>>> This approach is very different from the "frontend-allocates"
>>> approach used elsewhere in Xen. It is very much Linux-centric,
>>> rather than Xen-centric. In fact, MMU notifiers were invented for
>>> KVM, and this approach is exactly the same as the one KVM implements.
>>> However, to the best of my understanding, the design described here is
>>> the only viable one. Linux MM and GPU drivers require it, and changes
>>> to either to relax this requirement will not be accepted upstream.
>>
>> Teddy Astie (CCd) proposed a couple of alternatives on Matrix:
>>
>> 1. Create dma-bufs for guest pages and import them into the host.
>>
>> This is a win not only for Xen, but also for KVM. Right now, shared
>> (CPU) memory buffers must be copied from the guest to the host,
>> which is pointless. So fixing that is a good thing! That said,
>> I'm still concerned about triggering GPU driver code-paths that
>> are not tested on bare metal.
>>
>> 2. Use PASID and 2-stage translation so that the GPU can operate in
>> guest physical memory.
>>
>> This is also a win. AMD XDNA absolutely requires PASID support,
>> and apparently AMD GPUs can also use PASID. So being able to use
>> PASID is certainly helpful.
>>
>> However, I don't think either approach is sufficient for two reasons.
>>
>> First, discrete GPUs have dedicated VRAM, which Xen knows nothing about.
>> Only dom0's GPU drivers can manage VRAM, and they will insist on being
>> able to migrate it between the CPU and the GPU. Furthermore, VRAM
>> can only be allocated using GPU driver ioctls, which will allocate
>> it from dom0-owned memory.
>>
>> Second, Certain Wayland protocols, such as screencapture, require programs
>> to be able to import dmabufs. Both of the above solutions would
>> require that the pages be pinned. I don't think this is an option,
>> as IIUC pin_user_pages() fails on mappings of these dmabufs. It's why
>> direct I/O to dmabufs doesn't work.
>>
>
> I suppose it fails because of the RAM/VRAM constraint you said
> previously. If the location of the memory stays the same (i.e guest
> memory mapping), pin should be almost "no-op".
>
> (though, having dma-buf buffers coming from GPU drivers failing to pin
> is probably not a good thing in term of stability; some stuff like
> cameras probably break as a result; but I'm not a expert on that subject)
>
>> To the best of my knowledge, these problems mean that lending memory
>> is the only way to get robust GPU acceleration for both graphics and
>> compute workloads under Xen. Simpler approaches might work for pure
>> compute workloads, for iGPUs, or for drivers that have Xen-specific
>> changes. None of them, however, support graphics workloads on dGPUs
>> while using the GPU driver the same way bare metal workloads do.
>>
>> Linux's graphics stack is massive, and trying to adapt it to work with
>> Xen isn't going to be sustainable in the long term. Adapting Xen to
>> fit the graphics stack is probably more work up front, but it has the
>> advantage of working with all GPU drivers, including ones that have not
>> been written yet. It also means that the testing done on bare metal is
>> still applicable, and that bugs found when using this driver can either
>> be reproduced on bare metal or can be fixed without driver changes.
>
> One of my main concerns was about whether dma-buf can be used as
> "general purpose" GPU buffers; what I read in driver code suggest it
> should be fine, but it's a bit on the edge.
>
>>
>> Finally, I'm not actually attached to memory lending at all. It's a
>> lot of complexity, and it's not at all similar to how the rest of
>> Xen works. If someone else can come up with a better solution that
>> doesn't require GPU driver changes, I'd be all for it. Unfortunately,
>> I suspect none exists. One can make almost anything work if one is
>> willing to patch the drivers, but I am virtually certain that this
>> will not be long-term sustainable.
>>
>
> There's also the virtio-gpu side to consider. Blob mechanism appears to
> insist that GPU memory to come from the host by allowing buffers that
> aren't bound to virtio-gpu BAR yet (that also complexifies the KVM
> situation).
>
> You can have GPU memory that exists in virtio-gpu, without being
> guest-visible, then the guest can map it on its own BAR.
>
>> If Xen had its own GPU drivers, the situation would be totally
>> different. However, Xen must rely on Linux's GPU drivers, and that
>> means it must play by their rules.
>
>
>
>
> --
> Teddy Astie | Vates XCP-ng Developer
>
> XCP-ng & Xen Orchestra - Vates solutions
>
> web: https://vates.tech
>
>
next prev parent reply other threads:[~2026-03-30 10:25 UTC|newest]
Thread overview: 15+ messages / expand[flat|nested] mbox.gz Atom feed top
2026-03-24 14:17 Mapping non-pinned memory from one Xen domain into another Demi Marie Obenour
2026-03-24 18:00 ` Teddy Astie
2026-03-26 17:18 ` Demi Marie Obenour
2026-03-26 18:26 ` Teddy Astie
2026-03-27 17:18 ` Demi Marie Obenour
2026-03-29 17:32 ` Why memory lending is needed for GPU acceleration Demi Marie Obenour
2026-03-30 10:15 ` Teddy Astie
2026-03-30 10:25 ` Jan Beulich [this message]
2026-03-30 12:24 ` Demi Marie Obenour
2026-03-30 20:07 ` Val Packett
2026-03-31 9:42 ` Teddy Astie
2026-03-31 9:42 ` Teddy Astie
2026-03-31 11:23 ` Val Packett
2026-04-03 21:24 ` Marek Marczykowski-Górecki
2026-03-30 12:13 ` Mapping non-pinned memory from one Xen domain into another Teddy Astie
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=cf54e793-63d6-417c-b7c5-3301455fcf33@suse.com \
--to=jbeulich@suse.com \
--cc=andrew.cooper3@citrix.com \
--cc=ariadne@ariadne.space \
--cc=demiobenour@gmail.com \
--cc=dri-devel@lists.freedesktop.org \
--cc=jgross@suse.com \
--cc=linux-mm@kvack.org \
--cc=teddy.astie@vates.tech \
--cc=val@invisiblethingslab.com \
--cc=xen-devel@lists.xenproject.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.