qemu-devel.nongnu.org archive mirror
 help / color / mirror / Atom feed
From: David Woodhouse <dwmw2@infradead.org>
To: Peter Maydell <peter.maydell@linaro.org>, qemu-devel@nongnu.org
Cc: "Paolo Bonzini" <pbonzini@redhat.com>,
	"Paul Durrant" <paul@xen.org>,
	"Joao Martins" <joao.m.martins@oracle.com>,
	"Ankur Arora" <ankur.a.arora@oracle.com>,
	"Philippe Mathieu-Daudé" <philmd@linaro.org>,
	"Thomas Huth" <thuth@redhat.com>,
	"Alex Bennée" <alex.bennee@linaro.org>,
	"Juan Quintela" <quintela@redhat.com>,
	"Dr . David Alan Gilbert" <dgilbert@redhat.com>,
	"Claudio Fontana" <cfontana@suse.de>,
	"Julien Grall" <julien@xen.org>,
	"Michael S. Tsirkin" <mst@redhat.com>,
	"Marcel Apfelbaum" <marcel.apfelbaum@gmail.com>,
	armbru@redhat.com, "Stefano Stabellini" <sstabellini@kernel.org>,
	vikram.garhwal@amd.com
Subject: [PATCH v14 52/60] hw/xen: Add basic ring handling to xenstore
Date: Mon, 27 Feb 2023 14:29:00 +0000	[thread overview]
Message-ID: <20230227142908.503517-53-dwmw2@infradead.org> (raw)
In-Reply-To: <20230227142908.503517-1-dwmw2@infradead.org>

From: David Woodhouse <dwmw@amazon.co.uk>

Extract requests, return ENOSYS to all of them. This is enough to allow
older Linux guests to boot, as they need *something* back but it doesn't
matter much what.

A full implementation of a single-tentant internal XenStore copy-on-write
tree with transactions and watches is waiting in the wings to be sent in
a subsequent round of patches along with hooking up the actual PV disk
back end in qemu, but this is enough to get guests booting for now.

Signed-off-by: David Woodhouse <dwmw@amazon.co.uk>
Reviewed-by: Paul Durrant <paul@xen.org>
---
 hw/i386/kvm/xen_xenstore.c | 254 ++++++++++++++++++++++++++++++++++++-
 1 file changed, 251 insertions(+), 3 deletions(-)

diff --git a/hw/i386/kvm/xen_xenstore.c b/hw/i386/kvm/xen_xenstore.c
index e8abddae57..14193ef3f9 100644
--- a/hw/i386/kvm/xen_xenstore.c
+++ b/hw/i386/kvm/xen_xenstore.c
@@ -192,18 +192,266 @@ uint16_t xen_xenstore_get_port(void)
     return s->guest_port;
 }
 
+static bool req_pending(XenXenstoreState *s)
+{
+    struct xsd_sockmsg *req = (struct xsd_sockmsg *)s->req_data;
+
+    return s->req_offset == XENSTORE_HEADER_SIZE + req->len;
+}
+
+static void reset_req(XenXenstoreState *s)
+{
+    memset(s->req_data, 0, sizeof(s->req_data));
+    s->req_offset = 0;
+}
+
+static void reset_rsp(XenXenstoreState *s)
+{
+    s->rsp_pending = false;
+
+    memset(s->rsp_data, 0, sizeof(s->rsp_data));
+    s->rsp_offset = 0;
+}
+
+static void process_req(XenXenstoreState *s)
+{
+    struct xsd_sockmsg *req = (struct xsd_sockmsg *)s->req_data;
+    struct xsd_sockmsg *rsp = (struct xsd_sockmsg *)s->rsp_data;
+    const char enosys[] = "ENOSYS";
+
+    assert(req_pending(s));
+    assert(!s->rsp_pending);
+
+    rsp->type = XS_ERROR;
+    rsp->req_id = req->req_id;
+    rsp->tx_id = req->tx_id;
+    rsp->len = sizeof(enosys);
+    memcpy((void *)&rsp[1], enosys, sizeof(enosys));
+
+    s->rsp_pending = true;
+    reset_req(s);
+}
+
+static unsigned int copy_from_ring(XenXenstoreState *s, uint8_t *ptr,
+                                   unsigned int len)
+{
+    if (!len) {
+        return 0;
+    }
+
+    XENSTORE_RING_IDX prod = qatomic_read(&s->xs->req_prod);
+    XENSTORE_RING_IDX cons = qatomic_read(&s->xs->req_cons);
+    unsigned int copied = 0;
+
+    /* Ensure the ring contents don't cross the req_prod access. */
+    smp_rmb();
+
+    while (len) {
+        unsigned int avail = prod - cons;
+        unsigned int offset = MASK_XENSTORE_IDX(cons);
+        unsigned int copylen = avail;
+
+        if (avail > XENSTORE_RING_SIZE) {
+            error_report("XenStore ring handling error");
+            s->fatal_error = true;
+            break;
+        } else if (avail == 0) {
+            break;
+        }
+
+        if (copylen > len) {
+            copylen = len;
+        }
+        if (copylen > XENSTORE_RING_SIZE - offset) {
+            copylen = XENSTORE_RING_SIZE - offset;
+        }
+
+        memcpy(ptr, &s->xs->req[offset], copylen);
+        copied += copylen;
+
+        ptr += copylen;
+        len -= copylen;
+
+        cons += copylen;
+    }
+
+    /*
+     * Not sure this ever mattered except on Alpha, but this barrier
+     * is to ensure that the update to req_cons is globally visible
+     * only after we have consumed all the data from the ring, and we
+     * don't end up seeing data written to the ring *after* the other
+     * end sees the update and writes more to the ring. Xen's own
+     * xenstored has the same barrier here (although with no comment
+     * at all, obviously, because it's Xen code).
+     */
+    smp_mb();
+
+    qatomic_set(&s->xs->req_cons, cons);
+
+    return copied;
+}
+
+static unsigned int copy_to_ring(XenXenstoreState *s, uint8_t *ptr,
+                                 unsigned int len)
+{
+    if (!len) {
+        return 0;
+    }
+
+    XENSTORE_RING_IDX cons = qatomic_read(&s->xs->rsp_cons);
+    XENSTORE_RING_IDX prod = qatomic_read(&s->xs->rsp_prod);
+    unsigned int copied = 0;
+
+    /*
+     * This matches the barrier in copy_to_ring() (or the guest's
+     * equivalent) betweem writing the data to the ring and updating
+     * rsp_prod. It protects against the pathological case (which
+     * again I think never happened except on Alpha) where our
+     * subsequent writes to the ring could *cross* the read of
+     * rsp_cons and the guest could see the new data when it was
+     * intending to read the old.
+     */
+    smp_mb();
+
+    while (len) {
+        unsigned int avail = cons + XENSTORE_RING_SIZE - prod;
+        unsigned int offset = MASK_XENSTORE_IDX(prod);
+        unsigned int copylen = len;
+
+        if (avail > XENSTORE_RING_SIZE) {
+            error_report("XenStore ring handling error");
+            s->fatal_error = true;
+            break;
+        } else if (avail == 0) {
+            break;
+        }
+
+        if (copylen > avail) {
+            copylen = avail;
+        }
+        if (copylen > XENSTORE_RING_SIZE - offset) {
+            copylen = XENSTORE_RING_SIZE - offset;
+        }
+
+
+        memcpy(&s->xs->rsp[offset], ptr, copylen);
+        copied += copylen;
+
+        ptr += copylen;
+        len -= copylen;
+
+        prod += copylen;
+    }
+
+    /* Ensure the ring contents are seen before rsp_prod update. */
+    smp_wmb();
+
+    qatomic_set(&s->xs->rsp_prod, prod);
+
+    return copied;
+}
+
+static unsigned int get_req(XenXenstoreState *s)
+{
+    unsigned int copied = 0;
+
+    if (s->fatal_error) {
+        return 0;
+    }
+
+    assert(!req_pending(s));
+
+    if (s->req_offset < XENSTORE_HEADER_SIZE) {
+        void *ptr = s->req_data + s->req_offset;
+        unsigned int len = XENSTORE_HEADER_SIZE;
+        unsigned int copylen = copy_from_ring(s, ptr, len);
+
+        copied += copylen;
+        s->req_offset += copylen;
+    }
+
+    if (s->req_offset >= XENSTORE_HEADER_SIZE) {
+        struct xsd_sockmsg *req = (struct xsd_sockmsg *)s->req_data;
+
+        if (req->len > (uint32_t)XENSTORE_PAYLOAD_MAX) {
+            error_report("Illegal XenStore request");
+            s->fatal_error = true;
+            return 0;
+        }
+
+        void *ptr = s->req_data + s->req_offset;
+        unsigned int len = XENSTORE_HEADER_SIZE + req->len - s->req_offset;
+        unsigned int copylen = copy_from_ring(s, ptr, len);
+
+        copied += copylen;
+        s->req_offset += copylen;
+    }
+
+    return copied;
+}
+
+static unsigned int put_rsp(XenXenstoreState *s)
+{
+    if (s->fatal_error) {
+        return 0;
+    }
+
+    assert(s->rsp_pending);
+
+    struct xsd_sockmsg *rsp = (struct xsd_sockmsg *)s->rsp_data;
+    assert(s->rsp_offset < XENSTORE_HEADER_SIZE + rsp->len);
+
+    void *ptr = s->rsp_data + s->rsp_offset;
+    unsigned int len = XENSTORE_HEADER_SIZE + rsp->len - s->rsp_offset;
+    unsigned int copylen = copy_to_ring(s, ptr, len);
+
+    s->rsp_offset += copylen;
+
+    /* Have we produced a complete response? */
+    if (s->rsp_offset == XENSTORE_HEADER_SIZE + rsp->len) {
+        reset_rsp(s);
+    }
+
+    return copylen;
+}
+
 static void xen_xenstore_event(void *opaque)
 {
     XenXenstoreState *s = opaque;
     evtchn_port_t port = xen_be_evtchn_pending(s->eh);
+    unsigned int copied_to, copied_from;
+    bool processed, notify = false;
+
     if (port != s->be_port) {
         return;
     }
-    printf("xenstore event\n");
+
     /* We know this is a no-op. */
     xen_be_evtchn_unmask(s->eh, port);
-    qemu_hexdump(stdout, "", s->xs, sizeof(*s->xs));
-    xen_be_evtchn_notify(s->eh, s->be_port);
+
+    do {
+        copied_to = copied_from = 0;
+        processed = false;
+
+        if (s->rsp_pending) {
+            copied_to = put_rsp(s);
+        }
+
+        if (!req_pending(s)) {
+            copied_from = get_req(s);
+        }
+
+        if (req_pending(s) && !s->rsp_pending) {
+            process_req(s);
+            processed = true;
+        }
+
+        notify |= copied_to || copied_from;
+    } while (copied_to || copied_from || processed);
+
+    if (notify) {
+        xen_be_evtchn_notify(s->eh, s->be_port);
+    }
 }
 
 static void alloc_guest_port(XenXenstoreState *s)
-- 
2.39.0



  parent reply	other threads:[~2023-02-27 14:44 UTC|newest]

Thread overview: 63+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2023-02-27 14:28 [PATCH v14 00/60] Xen HVM support under KVM David Woodhouse
2023-02-27 14:28 ` [PATCH v14 01/60] include: import Xen public headers to hw/xen/interface David Woodhouse
2023-02-27 14:28 ` [PATCH v14 02/60] xen: add CONFIG_XEN_BUS and CONFIG_XEN_EMU options for Xen emulation David Woodhouse
2023-02-27 14:28 ` [PATCH v14 03/60] xen: Add XEN_DISABLED mode and make it default David Woodhouse
2023-02-27 14:28 ` [PATCH v14 04/60] i386/kvm: Add xen-version KVM accelerator property and init KVM Xen support David Woodhouse
2023-02-27 14:28 ` [PATCH v14 05/60] i386/kvm: handle Xen HVM cpuid leaves David Woodhouse
2023-02-27 14:28 ` [PATCH v14 06/60] i386/kvm: Set Xen vCPU ID in KVM David Woodhouse
2023-02-27 14:28 ` [PATCH v14 07/60] xen-platform: exclude vfio-pci from the PCI platform unplug David Woodhouse
2023-02-27 14:28 ` [PATCH v14 08/60] xen-platform: allow its creation with XEN_EMULATE mode David Woodhouse
2023-02-27 14:28 ` [PATCH v14 09/60] i386/xen: handle guest hypercalls David Woodhouse
2023-02-27 14:28 ` [PATCH v14 10/60] i386/xen: implement HYPERVISOR_xen_version David Woodhouse
2023-02-27 14:28 ` [PATCH v14 11/60] i386/xen: implement HYPERVISOR_sched_op, SCHEDOP_shutdown David Woodhouse
2023-02-27 14:28 ` [PATCH v14 12/60] i386/xen: Implement SCHEDOP_poll and SCHEDOP_yield David Woodhouse
2023-02-27 14:28 ` [PATCH v14 13/60] hw/xen: Add xen_overlay device for emulating shared xenheap pages David Woodhouse
2023-02-27 14:28 ` [PATCH v14 14/60] xen: Permit --xen-domid argument when accel is KVM David Woodhouse
2023-02-27 14:28 ` [PATCH v14 15/60] i386/xen: add pc_machine_kvm_type to initialize XEN_EMULATE mode David Woodhouse
2023-02-27 22:37   ` Philippe Mathieu-Daudé
2023-02-28  8:55     ` David Woodhouse
2023-02-27 14:28 ` [PATCH v14 16/60] i386/xen: manage and save/restore Xen guest long_mode setting David Woodhouse
2023-02-27 14:28 ` [PATCH v14 17/60] i386/xen: implement HYPERVISOR_memory_op David Woodhouse
2023-02-27 14:28 ` [PATCH v14 18/60] i386/xen: implement XENMEM_add_to_physmap_batch David Woodhouse
2023-02-27 14:28 ` [PATCH v14 19/60] i386/xen: implement HYPERVISOR_hvm_op David Woodhouse
2023-02-27 14:28 ` [PATCH v14 20/60] i386/xen: implement HYPERVISOR_vcpu_op David Woodhouse
2023-02-27 14:28 ` [PATCH v14 21/60] i386/xen: handle VCPUOP_register_vcpu_info David Woodhouse
2023-02-27 14:28 ` [PATCH v14 22/60] i386/xen: handle VCPUOP_register_vcpu_time_info David Woodhouse
2023-02-27 14:28 ` [PATCH v14 23/60] i386/xen: handle VCPUOP_register_runstate_memory_area David Woodhouse
2023-02-27 14:28 ` [PATCH v14 24/60] i386/xen: implement HYPERVISOR_event_channel_op David Woodhouse
2023-02-27 14:28 ` [PATCH v14 25/60] i386/xen: implement HVMOP_set_evtchn_upcall_vector David Woodhouse
2023-02-27 14:28 ` [PATCH v14 26/60] i386/xen: implement HVMOP_set_param David Woodhouse
2023-02-27 14:28 ` [PATCH v14 27/60] hw/xen: Add xen_evtchn device for event channel emulation David Woodhouse
2023-02-27 14:28 ` [PATCH v14 28/60] i386/xen: Add support for Xen event channel delivery to vCPU David Woodhouse
2023-02-27 14:28 ` [PATCH v14 29/60] hw/xen: Implement EVTCHNOP_status David Woodhouse
2023-02-27 14:28 ` [PATCH v14 30/60] hw/xen: Implement EVTCHNOP_close David Woodhouse
2023-02-27 14:28 ` [PATCH v14 31/60] hw/xen: Implement EVTCHNOP_unmask David Woodhouse
2023-02-27 14:28 ` [PATCH v14 32/60] hw/xen: Implement EVTCHNOP_bind_virq David Woodhouse
2023-02-27 14:28 ` [PATCH v14 33/60] hw/xen: Implement EVTCHNOP_bind_ipi David Woodhouse
2023-02-27 14:28 ` [PATCH v14 34/60] hw/xen: Implement EVTCHNOP_send David Woodhouse
2023-02-27 14:28 ` [PATCH v14 35/60] hw/xen: Implement EVTCHNOP_alloc_unbound David Woodhouse
2023-02-27 14:28 ` [PATCH v14 36/60] hw/xen: Implement EVTCHNOP_bind_interdomain David Woodhouse
2023-02-27 14:28 ` [PATCH v14 37/60] hw/xen: Implement EVTCHNOP_bind_vcpu David Woodhouse
2023-02-27 14:28 ` [PATCH v14 38/60] hw/xen: Implement EVTCHNOP_reset David Woodhouse
2023-02-27 14:28 ` [PATCH v14 39/60] i386/xen: add monitor commands to test event injection David Woodhouse
2023-02-27 14:28 ` [PATCH v14 40/60] hw/xen: Support HVM_PARAM_CALLBACK_TYPE_GSI callback David Woodhouse
2023-02-27 14:28 ` [PATCH v14 41/60] hw/xen: Support HVM_PARAM_CALLBACK_TYPE_PCI_INTX callback David Woodhouse
2023-02-27 14:28 ` [PATCH v14 42/60] kvm/i386: Add xen-gnttab-max-frames property David Woodhouse
2023-02-27 14:28 ` [PATCH v14 43/60] hw/xen: Add xen_gnttab device for grant table emulation David Woodhouse
2023-02-27 14:28 ` [PATCH v14 44/60] hw/xen: Support mapping grant frames David Woodhouse
2023-02-27 14:28 ` [PATCH v14 45/60] i386/xen: Implement HYPERVISOR_grant_table_op and GNTTABOP_[gs]et_verson David Woodhouse
2023-02-27 14:28 ` [PATCH v14 46/60] hw/xen: Implement GNTTABOP_query_size David Woodhouse
2023-02-27 14:28 ` [PATCH v14 47/60] i386/xen: handle PV timer hypercalls David Woodhouse
2023-02-27 14:28 ` [PATCH v14 48/60] i386/xen: Reserve Xen special pages for console, xenstore rings David Woodhouse
2023-02-27 14:28 ` [PATCH v14 49/60] i386/xen: handle HVMOP_get_param David Woodhouse
2023-02-27 14:28 ` [PATCH v14 50/60] hw/xen: Add backend implementation of interdomain event channel support David Woodhouse
2023-02-27 14:28 ` [PATCH v14 51/60] hw/xen: Add xen_xenstore device for xenstore emulation David Woodhouse
2023-02-27 14:29 ` David Woodhouse [this message]
2023-02-27 14:29 ` [PATCH v14 53/60] hw/xen: Automatically add xen-platform PCI device for emulated Xen guests David Woodhouse
2023-02-27 14:29 ` [PATCH v14 54/60] i386/xen: Implement HYPERVISOR_physdev_op David Woodhouse
2023-02-27 14:29 ` [PATCH v14 55/60] hw/xen: Implement emulated PIRQ hypercall support David Woodhouse
2023-02-27 14:29 ` [PATCH v14 56/60] hw/xen: Support GSI mapping to PIRQ David Woodhouse
2023-02-27 14:29 ` [PATCH v14 57/60] hw/xen: Support MSI " David Woodhouse
2023-02-27 14:29 ` [PATCH v14 58/60] kvm/i386: Add xen-evtchn-max-pirq property David Woodhouse
2023-02-27 14:29 ` [PATCH v14 59/60] i386/xen: Document Xen HVM emulation David Woodhouse
2023-02-27 14:29 ` [PATCH v14 60/60] hw/xen: Subsume xen_be_register_common() into xen_be_init() David Woodhouse

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20230227142908.503517-53-dwmw2@infradead.org \
    --to=dwmw2@infradead.org \
    --cc=alex.bennee@linaro.org \
    --cc=ankur.a.arora@oracle.com \
    --cc=armbru@redhat.com \
    --cc=cfontana@suse.de \
    --cc=dgilbert@redhat.com \
    --cc=joao.m.martins@oracle.com \
    --cc=julien@xen.org \
    --cc=marcel.apfelbaum@gmail.com \
    --cc=mst@redhat.com \
    --cc=paul@xen.org \
    --cc=pbonzini@redhat.com \
    --cc=peter.maydell@linaro.org \
    --cc=philmd@linaro.org \
    --cc=qemu-devel@nongnu.org \
    --cc=quintela@redhat.com \
    --cc=sstabellini@kernel.org \
    --cc=thuth@redhat.com \
    --cc=vikram.garhwal@amd.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).