From: "Dr. David Alan Gilbert" <dgilbert@redhat.com>
To: Jag Raman <jag.raman@oracle.com>
Cc: "eduardo@habkost.net" <eduardo@habkost.net>,
"Elena Ufimtseva" <elena.ufimtseva@oracle.com>,
"John Johnson" <john.g.johnson@oracle.com>,
"Daniel P. Berrangé" <berrange@redhat.com>,
"Beraldo Leal" <bleal@redhat.com>,
"john.levon@nutanix.com" <john.levon@nutanix.com>,
"Michael S. Tsirkin" <mst@redhat.com>,
"armbru@redhat.com" <armbru@redhat.com>,
"quintela@redhat.com" <quintela@redhat.com>,
"Philippe Mathieu-Daudé" <f4bug@amsat.org>,
qemu-devel <qemu-devel@nongnu.org>,
"Marc-André Lureau" <marcandre.lureau@gmail.com>,
"Stefan Hajnoczi" <stefanha@redhat.com>,
"thanos.makatos@nutanix.com" <thanos.makatos@nutanix.com>,
"Paolo Bonzini" <pbonzini@redhat.com>,
"Eric Blake" <eblake@redhat.com>
Subject: Re: [PATCH v5 03/18] pci: isolated address space for PCI bus
Date: Tue, 25 Jan 2022 18:38:32 +0000 [thread overview]
Message-ID: <YfBDqMuO/l/gWzL8@work-vm> (raw)
In-Reply-To: <1CACFB08-1BBC-4ECC-9C0B-6F377018D795@oracle.com>
* Jag Raman (jag.raman@oracle.com) wrote:
>
>
> > On Jan 19, 2022, at 7:12 PM, Michael S. Tsirkin <mst@redhat.com> wrote:
> >
> > On Wed, Jan 19, 2022 at 04:41:52PM -0500, Jagannathan Raman wrote:
> >> Allow PCI buses to be part of isolated CPU address spaces. This has a
> >> niche usage.
> >>
> >> TYPE_REMOTE_MACHINE allows multiple VMs to house their PCI devices in
> >> the same machine/server. This would cause address space collision as
> >> well as be a security vulnerability. Having separate address spaces for
> >> each PCI bus would solve this problem.
> >
> > Fascinating, but I am not sure I understand. any examples?
>
> Hi Michael!
>
> multiprocess QEMU and vfio-user implement a client-server model to allow
> out-of-process emulation of devices. The client QEMU, which makes ioctls
> to the kernel and runs VCPUs, could attach devices running in a server
> QEMU. The server QEMU needs access to parts of the client’s RAM to
> perform DMA.
Do you ever have the opposite problem? i.e. when an emulated PCI device
exposes a chunk of RAM-like space (frame buffer, or maybe a mapped file)
that the client can see. What happens if two emulated devices need to
access each others emulated address space?
Dave
> In the case where multiple clients attach devices that are running on the
> same server, we need to ensure that each devices has isolated memory
> ranges. This ensures that the memory space of one device is not visible
> to other devices in the same server.
>
> >
> > I also wonder whether this special type could be modelled like a special
> > kind of iommu internally.
>
> Could you please provide some more details on the design?
>
> >
> >> Signed-off-by: Elena Ufimtseva <elena.ufimtseva@oracle.com>
> >> Signed-off-by: John G Johnson <john.g.johnson@oracle.com>
> >> Signed-off-by: Jagannathan Raman <jag.raman@oracle.com>
> >> ---
> >> include/hw/pci/pci.h | 2 ++
> >> include/hw/pci/pci_bus.h | 17 +++++++++++++++++
> >> hw/pci/pci.c | 17 +++++++++++++++++
> >> hw/pci/pci_bridge.c | 5 +++++
> >> 4 files changed, 41 insertions(+)
> >>
> >> diff --git a/include/hw/pci/pci.h b/include/hw/pci/pci.h
> >> index 023abc0f79..9bb4472abc 100644
> >> --- a/include/hw/pci/pci.h
> >> +++ b/include/hw/pci/pci.h
> >> @@ -387,6 +387,8 @@ void pci_device_save(PCIDevice *s, QEMUFile *f);
> >> int pci_device_load(PCIDevice *s, QEMUFile *f);
> >> MemoryRegion *pci_address_space(PCIDevice *dev);
> >> MemoryRegion *pci_address_space_io(PCIDevice *dev);
> >> +AddressSpace *pci_isol_as_mem(PCIDevice *dev);
> >> +AddressSpace *pci_isol_as_io(PCIDevice *dev);
> >>
> >> /*
> >> * Should not normally be used by devices. For use by sPAPR target
> >> diff --git a/include/hw/pci/pci_bus.h b/include/hw/pci/pci_bus.h
> >> index 347440d42c..d78258e79e 100644
> >> --- a/include/hw/pci/pci_bus.h
> >> +++ b/include/hw/pci/pci_bus.h
> >> @@ -39,9 +39,26 @@ struct PCIBus {
> >> void *irq_opaque;
> >> PCIDevice *devices[PCI_SLOT_MAX * PCI_FUNC_MAX];
> >> PCIDevice *parent_dev;
> >> +
> >> MemoryRegion *address_space_mem;
> >> MemoryRegion *address_space_io;
> >>
> >> + /**
> >> + * Isolated address spaces - these allow the PCI bus to be part
> >> + * of an isolated address space as opposed to the global
> >> + * address_space_memory & address_space_io.
> >
> > Are you sure address_space_memory & address_space_io are
> > always global? even in the case of an iommu?
>
> On the CPU side of the Root Complex, I believe address_space_memory
> & address_space_io are global.
>
> In the vfio-user case, devices on the same machine (TYPE_REMOTE_MACHINE)
> could be attached to different clients VMs. Each client would have their own address
> space for their CPUs. With isolated address spaces, we ensure that the devices
> see the address space of the CPUs they’re attached to.
>
> Not sure if it’s OK to share weblinks in this mailing list, please let me know if that’s
> not preferred. But I’m referring to the terminology used in the following block diagram:
> https://en.wikipedia.org/wiki/Root_complex#/media/File:Example_PCI_Express_Topology.svg
>
> >
> >> This allows the
> >> + * bus to be attached to CPUs from different machines. The
> >> + * following is not used used commonly.
> >> + *
> >> + * TYPE_REMOTE_MACHINE allows emulating devices from multiple
> >> + * VM clients,
> >
> > what are VM clients?
>
> It’s the client in the client - server model explained above.
>
> Thank you!
> --
> Jag
>
> >
> >> as such it needs the PCI buses in the same machine
> >> + * to be part of different CPU address spaces. The following is
> >> + * useful in that scenario.
> >> + *
> >> + */
> >> + AddressSpace *isol_as_mem;
> >> + AddressSpace *isol_as_io;
> >> +
> >> QLIST_HEAD(, PCIBus) child; /* this will be replaced by qdev later */
> >> QLIST_ENTRY(PCIBus) sibling;/* this will be replaced by qdev later */
> >>
> >> diff --git a/hw/pci/pci.c b/hw/pci/pci.c
> >> index 5d30f9ca60..d5f1c6c421 100644
> >> --- a/hw/pci/pci.c
> >> +++ b/hw/pci/pci.c
> >> @@ -442,6 +442,8 @@ static void pci_root_bus_internal_init(PCIBus *bus, DeviceState *parent,
> >> bus->slot_reserved_mask = 0x0;
> >> bus->address_space_mem = address_space_mem;
> >> bus->address_space_io = address_space_io;
> >> + bus->isol_as_mem = NULL;
> >> + bus->isol_as_io = NULL;
> >> bus->flags |= PCI_BUS_IS_ROOT;
> >>
> >> /* host bridge */
> >> @@ -2676,6 +2678,16 @@ MemoryRegion *pci_address_space_io(PCIDevice *dev)
> >> return pci_get_bus(dev)->address_space_io;
> >> }
> >>
> >> +AddressSpace *pci_isol_as_mem(PCIDevice *dev)
> >> +{
> >> + return pci_get_bus(dev)->isol_as_mem;
> >> +}
> >> +
> >> +AddressSpace *pci_isol_as_io(PCIDevice *dev)
> >> +{
> >> + return pci_get_bus(dev)->isol_as_io;
> >> +}
> >> +
> >> static void pci_device_class_init(ObjectClass *klass, void *data)
> >> {
> >> DeviceClass *k = DEVICE_CLASS(klass);
> >> @@ -2699,6 +2711,7 @@ static void pci_device_class_base_init(ObjectClass *klass, void *data)
> >>
> >> AddressSpace *pci_device_iommu_address_space(PCIDevice *dev)
> >> {
> >> + AddressSpace *iommu_as = NULL;
> >> PCIBus *bus = pci_get_bus(dev);
> >> PCIBus *iommu_bus = bus;
> >> uint8_t devfn = dev->devfn;
> >> @@ -2745,6 +2758,10 @@ AddressSpace *pci_device_iommu_address_space(PCIDevice *dev)
> >> if (!pci_bus_bypass_iommu(bus) && iommu_bus && iommu_bus->iommu_fn) {
> >> return iommu_bus->iommu_fn(bus, iommu_bus->iommu_opaque, devfn);
> >> }
> >> + iommu_as = pci_isol_as_mem(dev);
> >> + if (iommu_as) {
> >> + return iommu_as;
> >> + }
> >> return &address_space_memory;
> >> }
> >>
> >> diff --git a/hw/pci/pci_bridge.c b/hw/pci/pci_bridge.c
> >> index da34c8ebcd..98366768d2 100644
> >> --- a/hw/pci/pci_bridge.c
> >> +++ b/hw/pci/pci_bridge.c
> >> @@ -383,6 +383,11 @@ void pci_bridge_initfn(PCIDevice *dev, const char *typename)
> >> sec_bus->address_space_io = &br->address_space_io;
> >> memory_region_init(&br->address_space_io, OBJECT(br), "pci_bridge_io",
> >> 4 * GiB);
> >> +
> >> + /* This PCI bridge puts the sec_bus in its parent's address space */
> >> + sec_bus->isol_as_mem = pci_isol_as_mem(dev);
> >> + sec_bus->isol_as_io = pci_isol_as_io(dev);
> >> +
> >> br->windows = pci_bridge_region_init(br);
> >> QLIST_INIT(&sec_bus->child);
> >> QLIST_INSERT_HEAD(&parent->child, sec_bus, sibling);
> >> --
> >> 2.20.1
>
--
Dr. David Alan Gilbert / dgilbert@redhat.com / Manchester, UK
next prev parent reply other threads:[~2022-01-25 18:53 UTC|newest]
Thread overview: 99+ messages / expand[flat|nested] mbox.gz Atom feed top
2022-01-19 21:41 [PATCH v5 00/18] vfio-user server in QEMU Jagannathan Raman
2022-01-19 21:41 ` [PATCH v5 01/18] configure, meson: override C compiler for cmake Jagannathan Raman
2022-01-20 13:27 ` Paolo Bonzini
2022-01-20 15:21 ` Jag Raman
2022-02-17 6:10 ` Jag Raman
2022-01-19 21:41 ` [PATCH v5 02/18] tests/avocado: Specify target VM argument to helper routines Jagannathan Raman
2022-01-25 9:40 ` Stefan Hajnoczi
2022-01-19 21:41 ` [PATCH v5 03/18] pci: isolated address space for PCI bus Jagannathan Raman
2022-01-20 0:12 ` Michael S. Tsirkin
2022-01-20 15:20 ` Jag Raman
2022-01-25 18:38 ` Dr. David Alan Gilbert [this message]
2022-01-26 5:27 ` Jag Raman
2022-01-26 9:45 ` Stefan Hajnoczi
2022-01-26 20:07 ` Dr. David Alan Gilbert
2022-01-26 21:13 ` Michael S. Tsirkin
2022-01-27 8:30 ` Stefan Hajnoczi
2022-01-27 12:50 ` Michael S. Tsirkin
2022-01-27 21:22 ` Alex Williamson
2022-01-28 8:19 ` Stefan Hajnoczi
2022-01-28 9:18 ` Stefan Hajnoczi
2022-01-31 16:16 ` Alex Williamson
2022-02-01 9:30 ` Stefan Hajnoczi
2022-02-01 15:24 ` Alex Williamson
2022-02-01 21:24 ` Jag Raman
2022-02-01 22:47 ` Alex Williamson
2022-02-02 1:13 ` Jag Raman
2022-02-02 5:34 ` Alex Williamson
2022-02-02 9:22 ` Stefan Hajnoczi
2022-02-10 0:08 ` Jag Raman
2022-02-10 8:02 ` Michael S. Tsirkin
2022-02-10 22:23 ` Jag Raman
2022-02-10 22:53 ` Michael S. Tsirkin
2022-02-10 23:46 ` Jag Raman
2022-02-10 23:17 ` Alex Williamson
2022-02-10 23:28 ` Michael S. Tsirkin
2022-02-10 23:49 ` Alex Williamson
2022-02-11 0:26 ` Michael S. Tsirkin
2022-02-11 0:54 ` Jag Raman
2022-02-11 0:10 ` Jag Raman
2022-02-02 9:30 ` Peter Maydell
2022-02-02 10:06 ` Michael S. Tsirkin
2022-02-02 15:49 ` Alex Williamson
2022-02-02 16:53 ` Michael S. Tsirkin
2022-02-02 17:12 ` Alex Williamson
2022-02-01 10:42 ` Dr. David Alan Gilbert
2022-01-26 18:13 ` Dr. David Alan Gilbert
2022-01-27 17:43 ` Jag Raman
2022-01-25 9:56 ` Stefan Hajnoczi
2022-01-25 13:49 ` Jag Raman
2022-01-25 14:19 ` Stefan Hajnoczi
2022-01-19 21:41 ` [PATCH v5 04/18] pci: create and free isolated PCI buses Jagannathan Raman
2022-01-25 10:25 ` Stefan Hajnoczi
2022-01-25 14:10 ` Jag Raman
2022-01-19 21:41 ` [PATCH v5 05/18] qdev: unplug blocker for devices Jagannathan Raman
2022-01-25 10:27 ` Stefan Hajnoczi
2022-01-25 14:43 ` Jag Raman
2022-01-26 9:32 ` Stefan Hajnoczi
2022-01-26 15:13 ` Jag Raman
2022-01-19 21:41 ` [PATCH v5 06/18] vfio-user: add HotplugHandler for remote machine Jagannathan Raman
2022-01-25 10:32 ` Stefan Hajnoczi
2022-01-25 18:12 ` Jag Raman
2022-01-26 9:35 ` Stefan Hajnoczi
2022-01-26 15:20 ` Jag Raman
2022-01-26 15:43 ` Stefan Hajnoczi
2022-01-19 21:41 ` [PATCH v5 07/18] vfio-user: set qdev bus callbacks " Jagannathan Raman
2022-01-25 10:44 ` Stefan Hajnoczi
2022-01-25 21:12 ` Jag Raman
2022-01-26 9:37 ` Stefan Hajnoczi
2022-01-26 15:51 ` Jag Raman
2022-01-19 21:41 ` [PATCH v5 08/18] vfio-user: build library Jagannathan Raman
2022-01-19 21:41 ` [PATCH v5 09/18] vfio-user: define vfio-user-server object Jagannathan Raman
2022-01-25 14:40 ` Stefan Hajnoczi
2022-01-19 21:41 ` [PATCH v5 10/18] vfio-user: instantiate vfio-user context Jagannathan Raman
2022-01-25 14:44 ` Stefan Hajnoczi
2022-01-19 21:42 ` [PATCH v5 11/18] vfio-user: find and init PCI device Jagannathan Raman
2022-01-25 14:48 ` Stefan Hajnoczi
2022-01-26 3:14 ` Jag Raman
2022-01-19 21:42 ` [PATCH v5 12/18] vfio-user: run vfio-user context Jagannathan Raman
2022-01-25 15:10 ` Stefan Hajnoczi
2022-01-26 3:26 ` Jag Raman
2022-01-19 21:42 ` [PATCH v5 13/18] vfio-user: handle PCI config space accesses Jagannathan Raman
2022-01-25 15:13 ` Stefan Hajnoczi
2022-01-19 21:42 ` [PATCH v5 14/18] vfio-user: handle DMA mappings Jagannathan Raman
2022-01-19 21:42 ` [PATCH v5 15/18] vfio-user: handle PCI BAR accesses Jagannathan Raman
2022-01-19 21:42 ` [PATCH v5 16/18] vfio-user: handle device interrupts Jagannathan Raman
2022-01-25 15:25 ` Stefan Hajnoczi
2022-01-19 21:42 ` [PATCH v5 17/18] vfio-user: register handlers to facilitate migration Jagannathan Raman
2022-01-25 15:48 ` Stefan Hajnoczi
2022-01-27 17:04 ` Jag Raman
2022-01-28 8:29 ` Stefan Hajnoczi
2022-01-28 14:49 ` Thanos Makatos
2022-02-01 3:49 ` Jag Raman
2022-02-01 9:37 ` Stefan Hajnoczi
2022-01-19 21:42 ` [PATCH v5 18/18] vfio-user: avocado tests for vfio-user Jagannathan Raman
2022-01-26 4:25 ` Philippe Mathieu-Daudé via
2022-01-26 15:12 ` Jag Raman
2022-01-25 16:00 ` [PATCH v5 00/18] vfio-user server in QEMU Stefan Hajnoczi
2022-01-26 5:04 ` Jag Raman
2022-01-26 9:56 ` Stefan Hajnoczi
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=YfBDqMuO/l/gWzL8@work-vm \
--to=dgilbert@redhat.com \
--cc=armbru@redhat.com \
--cc=berrange@redhat.com \
--cc=bleal@redhat.com \
--cc=eblake@redhat.com \
--cc=eduardo@habkost.net \
--cc=elena.ufimtseva@oracle.com \
--cc=f4bug@amsat.org \
--cc=jag.raman@oracle.com \
--cc=john.g.johnson@oracle.com \
--cc=john.levon@nutanix.com \
--cc=marcandre.lureau@gmail.com \
--cc=mst@redhat.com \
--cc=pbonzini@redhat.com \
--cc=qemu-devel@nongnu.org \
--cc=quintela@redhat.com \
--cc=stefanha@redhat.com \
--cc=thanos.makatos@nutanix.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).