From: Xiao Guangrong <guangrong.xiao@linux.intel.com>
To: pbonzini@redhat.com, imammedo@redhat.com
Cc: Xiao Guangrong <guangrong.xiao@linux.intel.com>,
ehabkost@redhat.com, kvm@vger.kernel.org, mst@redhat.com,
gleb@kernel.org, mtosatti@redhat.com, qemu-devel@nongnu.org,
stefanha@redhat.com, dan.j.williams@intel.com, rth@twiddle.net
Subject: [Qemu-devel] [PATCH v5 32/33] nvdimm acpi: support _FIT method
Date: Wed, 28 Oct 2015 22:26:30 +0000 [thread overview]
Message-ID: <1446071191-62591-33-git-send-email-guangrong.xiao@linux.intel.com> (raw)
In-Reply-To: <1446071191-62591-1-git-send-email-guangrong.xiao@linux.intel.com>
FIT buffer is not completely mapped into guest address space, so a new
function, Read FIT, function index 0xFFFFFFFF, is reserved by QEMU to
read the piece of FIT buffer. The buffer is concatenated before _FIT
return
Refer to docs/specs/acpi-nvdimm.txt for detailed design
Signed-off-by: Xiao Guangrong <guangrong.xiao@linux.intel.com>
---
hw/acpi/nvdimm.c | 168 +++++++++++++++++++++++++++++++++++++++++++++++++++++--
1 file changed, 164 insertions(+), 4 deletions(-)
diff --git a/hw/acpi/nvdimm.c b/hw/acpi/nvdimm.c
index c5a50ea..e1ae6c5 100644
--- a/hw/acpi/nvdimm.c
+++ b/hw/acpi/nvdimm.c
@@ -384,6 +384,18 @@ static void nvdimm_build_nfit(GSList *device_list, GArray *table_offsets,
g_array_free(structures, true);
}
+/*
+ * define UUID for NVDIMM Root Device according to Chapter 3 DSM Interface
+ * for NVDIMM Root Device - Example in DSM Spec Rev1.
+ */
+#define NVDIMM_DSM_ROOT_UUID "2F10E7A4-9E91-11E4-89D3-123B93F75CBA"
+
+/*
+ * Read FIT Function, which is a QEMU internal use only function, more detail
+ * refer to docs/specs/acpi_nvdimm.txt
+ */
+#define NVDIMM_DSM_FUNC_READ_FIT 0xFFFFFFFF
+
/* define NVDIMM DSM return status codes according to DSM Spec Rev1. */
enum {
/* Common return status codes. */
@@ -420,6 +432,11 @@ struct nvdimm_func_in_set_label_data {
} QEMU_PACKED;
typedef struct nvdimm_func_in_set_label_data nvdimm_func_in_set_label_data;
+struct nvdimm_func_in_read_fit {
+ uint32_t offset; /* fit offset */
+} QEMU_PACKED;
+typedef struct nvdimm_func_in_read_fit nvdimm_func_in_read_fit;
+
struct nvdimm_dsm_in {
uint32_t handle;
uint32_t revision;
@@ -429,6 +446,7 @@ struct nvdimm_dsm_in {
uint8_t arg3[0];
nvdimm_func_in_set_label_data func_set_label_data;
nvdimm_func_in_get_label_data func_get_label_data;
+ nvdimm_func_in_read_fit func_read_fit;
};
} QEMU_PACKED;
typedef struct nvdimm_dsm_in nvdimm_dsm_in;
@@ -450,13 +468,71 @@ struct nvdimm_func_out_get_label_data {
} QEMU_PACKED;
typedef struct nvdimm_func_out_get_label_data nvdimm_func_out_get_label_data;
+struct nvdimm_func_out_read_fit {
+ uint32_t status; /* return status code. */
+ uint32_t length; /* the length of fit data we read. */
+ uint8_t fit_data[0]; /* fit data. */
+} QEMU_PACKED;
+typedef struct nvdimm_func_out_read_fit nvdimm_func_out_read_fit;
+
static void nvdimm_dsm_write_status(GArray *out, uint32_t status)
{
status = cpu_to_le32(status);
build_append_int_noprefix(out, status, sizeof(status));
}
-static void nvdimm_dsm_root(nvdimm_dsm_in *in, GArray *out)
+/* Build fit memory which is presented to guest via _FIT method. */
+static void nvdimm_build_fit(AcpiNVDIMMState *state)
+{
+ if (!state->fit) {
+ GSList *device_list = nvdimm_get_plugged_device_list();
+
+ nvdimm_debug("Rebuild FIT...\n");
+ state->fit = nvdimm_build_device_structure(device_list);
+ g_slist_free(device_list);
+ }
+}
+
+/* Read FIT data, defined in docs/specs/acpi_nvdimm.txt. */
+static void nvdimm_dsm_func_read_fit(AcpiNVDIMMState *state,
+ nvdimm_dsm_in *in, GArray *out)
+{
+ nvdimm_func_in_read_fit *read_fit = &in->func_read_fit;
+ nvdimm_func_out_read_fit fit_out;
+ uint32_t read_length = getpagesize() - sizeof(nvdimm_func_out_read_fit);
+ uint32_t status = NVDIMM_DSM_ROOT_DEV_STATUS_INVALID_PARAS;
+
+ nvdimm_build_fit(state);
+
+ le32_to_cpus(&read_fit->offset);
+
+ nvdimm_debug("Read FIT offset %#x.\n", read_fit->offset);
+
+ if (read_fit->offset > state->fit->len) {
+ nvdimm_debug("offset %#x is beyond fit size (%#x).\n",
+ read_fit->offset, state->fit->len);
+ goto exit;
+ }
+
+ read_length = MIN(read_length, state->fit->len - read_fit->offset);
+ nvdimm_debug("read length %#x.\n", read_length);
+
+ fit_out.status = cpu_to_le32(NVDIMM_DSM_STATUS_SUCCESS);
+ fit_out.length = cpu_to_le32(read_length);
+ g_array_append_vals(out, &fit_out, sizeof(fit_out));
+
+ if (read_length) {
+ g_array_append_vals(out, state->fit->data + read_fit->offset,
+ read_length);
+ }
+ return;
+
+exit:
+ nvdimm_dsm_write_status(out, status);
+}
+
+static void nvdimm_dsm_root(AcpiNVDIMMState *state, nvdimm_dsm_in *in,
+ GArray *out)
{
uint32_t status = NVDIMM_DSM_STATUS_NOT_SUPPORTED;
@@ -475,6 +551,10 @@ static void nvdimm_dsm_root(nvdimm_dsm_in *in, GArray *out)
return;
}
+ if (in->function == NVDIMM_DSM_FUNC_READ_FIT /* FIT Read */) {
+ return nvdimm_dsm_func_read_fit(state, in, out);
+ }
+
nvdimm_debug("Return status %#x.\n", status);
nvdimm_dsm_write_status(out, status);
}
@@ -713,7 +793,7 @@ nvdimm_dsm_read(void *opaque, hwaddr addr, unsigned size)
/* Handle 0 is reserved for NVDIMM Root Device. */
if (!in->handle) {
- nvdimm_dsm_root(in, out);
+ nvdimm_dsm_root(state, in, out);
goto exit;
}
@@ -928,8 +1008,88 @@ static void nvdimm_build_acpi_devices(GSList *device_list, Aml *sb_scope)
*/
BUILD_DSM_METHOD(dev, method,
0 /* 0 is reserved for NVDIMM Root Device*/,
- "2F10E7A4-9E91-11E4-89D3-123B93F75CBA"
- /* UUID for NVDIMM Root Devices. */);
+ NVDIMM_DSM_ROOT_UUID /* UUID for NVDIMM Root Devices. */);
+
+ method = aml_method("RFIT", 1);
+ {
+ Aml *ret, *pckg, *ifcond, *ifctx, *dsm_return = aml_local(0);
+
+ aml_append(method, aml_create_dword_field(aml_buffer(4, NULL),
+ aml_int(0), "OFST"));
+
+ /* prepare nvdimm_func_in_read_fit.offset */
+ aml_append(method, aml_store(aml_arg(0), aml_name("OFST")));
+ pckg = aml_package(1);
+ aml_append(pckg, aml_name("OFST"));
+
+ ret = aml_call4("_DSM",
+ aml_touuid(NVDIMM_DSM_ROOT_UUID) /* Root Device UUID */,
+ aml_int(1) /* Revision 1 */,
+ aml_int(NVDIMM_DSM_FUNC_READ_FIT) /* Read FIT
+ Function Index */,
+ pckg);
+ aml_append(method, aml_store(ret, dsm_return));
+
+ aml_append(method, aml_create_dword_field(dsm_return,
+ aml_int(0) /* offset at byte 0 */,
+ "STAU"));
+ /* if something is wrong during _DSM. */
+ ifcond = aml_equal(aml_int(NVDIMM_DSM_STATUS_SUCCESS),
+ aml_name("STAU"));
+ ifctx = aml_if(aml_lnot(ifcond));
+ {
+ aml_append(ifctx, aml_return(aml_buffer(0, NULL)));
+ }
+ aml_append(method, ifctx);
+
+ aml_append(method, aml_create_dword_field(dsm_return,
+ aml_int(4) /* offset at byte 4. */,
+ "BFSZ"));
+ /* if we read the end of fit. */
+ ifctx = aml_if(aml_equal(aml_name("BFSZ"), aml_int(0)));
+ {
+ aml_append(ifctx, aml_return(aml_buffer(0, NULL)));
+ }
+ aml_append(method, ifctx);
+
+ aml_append(method, aml_store(aml_shiftleft(aml_name("BFSZ"),
+ aml_int(3)), aml_local(6)));
+ aml_append(method, aml_create_field(dsm_return,
+ aml_int(8 * BITS_PER_BYTE), /* offset at byte 8.*/
+ aml_local(6), "BUFF"));
+ aml_append(method, aml_return(aml_name("BUFF")));
+ }
+ aml_append(dev, method);
+
+ method = aml_method("_FIT", 0);
+ {
+ Aml *whilectx, *fit = aml_local(0), *offset = aml_local(1);
+
+ aml_append(method, aml_store(aml_buffer(0, NULL), fit));
+ aml_append(method, aml_store(aml_int(0), offset));
+
+ whilectx = aml_while(aml_int(1));
+ {
+ Aml *ifctx, *buf = aml_local(2), *bufsize = aml_local(3);
+
+ aml_append(whilectx, aml_store(aml_call1("RFIT", offset), buf));
+ aml_append(whilectx, aml_store(aml_sizeof(buf), bufsize));
+
+ /* finish fit read if no data is read out. */
+ ifctx = aml_if(aml_equal(bufsize, aml_int(0)));
+ {
+ aml_append(ifctx, aml_return(fit));
+ }
+ aml_append(whilectx, ifctx);
+
+ /* update the offset. */
+ aml_append(whilectx, aml_store(aml_add(offset, bufsize), offset));
+ /* append the data we read out to the fit buffer. */
+ aml_append(whilectx, aml_concatenate(fit, buf, fit));
+ }
+ aml_append(method, whilectx);
+ }
+ aml_append(dev, method);
build_nvdimm_devices(device_list, dev);
--
1.8.3.1
next prev parent reply other threads:[~2015-10-28 14:33 UTC|newest]
Thread overview: 43+ messages / expand[flat|nested] mbox.gz Atom feed top
2015-10-28 22:25 [Qemu-devel] [PATCH v5 00/33] implement vNVDIMM Xiao Guangrong
2015-10-28 22:25 ` [Qemu-devel] [PATCH v5 01/33] acpi: add aml_derefof Xiao Guangrong
2015-10-28 22:26 ` [Qemu-devel] [PATCH v5 02/33] acpi: add aml_sizeof Xiao Guangrong
2015-10-28 22:26 ` [Qemu-devel] [PATCH v5 03/33] acpi: add aml_create_field Xiao Guangrong
2015-10-28 22:26 ` [Qemu-devel] [PATCH v5 04/33] acpi: add aml_concatenate Xiao Guangrong
2015-10-28 22:26 ` [Qemu-devel] [PATCH v5 05/33] acpi: add aml_object_type Xiao Guangrong
2015-10-28 22:26 ` [Qemu-devel] [PATCH v5 06/33] acpi: add aml_method_serialized Xiao Guangrong
2015-10-28 22:26 ` [Qemu-devel] [PATCH v5 07/33] util: introduce qemu_file_get_page_size() Xiao Guangrong
2015-10-28 22:26 ` [Qemu-devel] [PATCH v5 08/33] exec: allow memory to be allocated from any kind of path Xiao Guangrong
2015-10-28 22:26 ` [Qemu-devel] [PATCH v5 09/33] exec: allow file_ram_alloc to work on file Xiao Guangrong
2015-10-28 22:26 ` [Qemu-devel] [PATCH v5 10/33] hostmem-file: clean up memory allocation Xiao Guangrong
2015-10-28 22:26 ` [Qemu-devel] [PATCH v5 11/33] hostmem-file: use whole file size if possible Xiao Guangrong
2015-10-28 22:26 ` [Qemu-devel] [PATCH v5 12/33] pc-dimm: remove DEFAULT_PC_DIMMSIZE Xiao Guangrong
2015-10-28 22:26 ` [Qemu-devel] [PATCH v5 13/33] pc-dimm: make pc_existing_dimms_capacity static and rename it Xiao Guangrong
2015-10-28 22:26 ` [Qemu-devel] [PATCH v5 14/33] pc-dimm: drop the prefix of pc-dimm Xiao Guangrong
2015-10-28 22:26 ` [Qemu-devel] [PATCH v5 15/33] stubs: rename qmp_pc_dimm_device_list.c Xiao Guangrong
2015-10-28 22:26 ` [Qemu-devel] [PATCH v5 16/33] pc-dimm: rename pc-dimm.c and pc-dimm.h Xiao Guangrong
2015-10-28 22:26 ` [Qemu-devel] [PATCH v5 17/33] dimm: abstract dimm device from pc-dimm Xiao Guangrong
2015-10-28 22:26 ` [Qemu-devel] [PATCH v5 18/33] dimm: get mapped memory region from DIMMDeviceClass->get_memory_region Xiao Guangrong
2015-10-28 22:26 ` [Qemu-devel] [PATCH v5 19/33] dimm: keep the state of the whole backend memory Xiao Guangrong
2015-10-28 22:26 ` [Qemu-devel] [PATCH v5 20/33] dimm: introduce realize callback Xiao Guangrong
2015-10-28 22:26 ` [Qemu-devel] [PATCH v5 21/33] nvdimm: implement NVDIMM device abstract Xiao Guangrong
2015-10-28 22:26 ` [Qemu-devel] [PATCH v5 22/33] docs: add NVDIMM ACPI documentation Xiao Guangrong
2015-10-28 22:26 ` [Qemu-devel] [PATCH v5 23/33] nvdimm acpi: init the resource used by NVDIMM ACPI Xiao Guangrong
2015-10-28 22:26 ` [Qemu-devel] [PATCH v5 24/33] nvdimm acpi: build ACPI NFIT table Xiao Guangrong
2015-10-28 22:26 ` [Qemu-devel] [PATCH v5 25/33] nvdimm acpi: build ACPI nvdimm devices Xiao Guangrong
2015-10-28 22:26 ` [Qemu-devel] [PATCH v5 26/33] nvdimm acpi: save arg3 for NVDIMM device _DSM method Xiao Guangrong
2015-10-28 22:26 ` [Qemu-devel] [PATCH v5 27/33] nvdimm acpi: support function 0 Xiao Guangrong
2015-10-28 16:37 ` Stefan Hajnoczi
2015-10-28 22:26 ` [Qemu-devel] [PATCH v5 28/33] nvdimm acpi: support Get Namespace Label Size function Xiao Guangrong
2015-10-28 16:41 ` Stefan Hajnoczi
2015-10-29 2:16 ` Xiao Guangrong
2015-10-29 14:14 ` Stefan Hajnoczi
2015-10-28 22:26 ` [Qemu-devel] [PATCH v5 29/33] nvdimm acpi: support Get Namespace Label Data function Xiao Guangrong
2015-10-28 16:46 ` Stefan Hajnoczi
2015-10-29 2:25 ` Xiao Guangrong
2015-10-29 2:58 ` Xiao Guangrong
2015-10-28 22:26 ` [Qemu-devel] [PATCH v5 30/33] nvdimm acpi: support Set " Xiao Guangrong
2015-10-28 16:48 ` Stefan Hajnoczi
2015-10-29 2:26 ` Xiao Guangrong
2015-10-28 22:26 ` [Qemu-devel] [PATCH v5 31/33] nvdimm: allow using whole backend memory as pmem Xiao Guangrong
2015-10-28 22:26 ` Xiao Guangrong [this message]
2015-10-28 22:26 ` [Qemu-devel] [PATCH v5 33/33] nvdimm: add maintain info Xiao Guangrong
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=1446071191-62591-33-git-send-email-guangrong.xiao@linux.intel.com \
--to=guangrong.xiao@linux.intel.com \
--cc=dan.j.williams@intel.com \
--cc=ehabkost@redhat.com \
--cc=gleb@kernel.org \
--cc=imammedo@redhat.com \
--cc=kvm@vger.kernel.org \
--cc=mst@redhat.com \
--cc=mtosatti@redhat.com \
--cc=pbonzini@redhat.com \
--cc=qemu-devel@nongnu.org \
--cc=rth@twiddle.net \
--cc=stefanha@redhat.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).