|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH 2/2] pc-nvdimm acpi: build ACPI tables for pc-nvdimm devices
CC'ing the Xen tools maintainers and Anthony.
On Tue, 29 Dec 2015, Haozhong Zhang wrote:
> Reuse existing NVDIMM ACPI code to build ACPI tables for pc-nvdimm
> devices. The resulting tables are then copied into Xen guest domain so
> tha they can be later loaded by Xen hvmloader.
>
> Signed-off-by: Haozhong Zhang <haozhong.zhang@xxxxxxxxx>
How much work would it be to generate the nvdimm acpi tables from the
Xen toolstack?
Getting the tables from QEMU doesn't seem like a good idea to me, unless
we start getting the whole set of ACPI tables that way.
> hw/acpi/nvdimm.c | 5 +++-
> hw/i386/pc.c | 6 ++++-
> include/hw/xen/xen.h | 2 ++
> xen-hvm.c | 71
> ++++++++++++++++++++++++++++++++++++++++++++++++++++
> 4 files changed, 82 insertions(+), 2 deletions(-)
>
> diff --git a/hw/acpi/nvdimm.c b/hw/acpi/nvdimm.c
> index df1b176..7c4b931 100644
> --- a/hw/acpi/nvdimm.c
> +++ b/hw/acpi/nvdimm.c
> @@ -29,12 +29,15 @@
> #include "hw/acpi/acpi.h"
> #include "hw/acpi/aml-build.h"
> #include "hw/mem/nvdimm.h"
> +#include "hw/mem/pc-nvdimm.h"
> +#include "hw/xen/xen.h"
>
> static int nvdimm_plugged_device_list(Object *obj, void *opaque)
> {
> GSList **list = opaque;
> + const char *type_name = xen_enabled() ? TYPE_PC_NVDIMM : TYPE_NVDIMM;
>
> - if (object_dynamic_cast(obj, TYPE_NVDIMM)) {
> + if (object_dynamic_cast(obj, type_name)) {
> DeviceState *dev = DEVICE(obj);
>
> if (dev->realized) { /* only realized NVDIMMs matter */
> diff --git a/hw/i386/pc.c b/hw/i386/pc.c
> index 459260b..fadacf5 100644
> --- a/hw/i386/pc.c
> +++ b/hw/i386/pc.c
> @@ -1186,7 +1186,11 @@ void pc_guest_info_machine_done(Notifier *notifier,
> void *data)
> }
> }
>
> - acpi_setup(&guest_info_state->info);
> + if (!xen_enabled()) {
> + acpi_setup(&guest_info_state->info);
> + } else if (xen_hvm_acpi_setup(PC_MACHINE(qdev_get_machine()))) {
> + error_report("Warning: failed to initialize Xen HVM ACPI tables");
> + }
> }
>
> PcGuestInfo *pc_guest_info_init(PCMachineState *pcms)
> diff --git a/include/hw/xen/xen.h b/include/hw/xen/xen.h
> index e90931a..8b705e1 100644
> --- a/include/hw/xen/xen.h
> +++ b/include/hw/xen/xen.h
> @@ -51,4 +51,6 @@ void xen_register_framebuffer(struct MemoryRegion *mr);
> # define HVM_MAX_VCPUS 32
> #endif
>
> +int xen_hvm_acpi_setup(PCMachineState *pcms);
> +
> #endif /* QEMU_HW_XEN_H */
> diff --git a/xen-hvm.c b/xen-hvm.c
> index 6ebf43f..f1f5e77 100644
> --- a/xen-hvm.c
> +++ b/xen-hvm.c
> @@ -26,6 +26,13 @@
> #include <xen/hvm/params.h>
> #include <xen/hvm/e820.h>
>
> +#include "qemu/error-report.h"
> +#include "hw/acpi/acpi.h"
> +#include "hw/acpi/aml-build.h"
> +#include "hw/acpi/bios-linker-loader.h"
> +#include "hw/mem/nvdimm.h"
> +#include "hw/mem/pc-nvdimm.h"
> +
> //#define DEBUG_XEN_HVM
>
> #ifdef DEBUG_XEN_HVM
> @@ -1330,6 +1337,70 @@ int xen_hvm_init(PCMachineState *pcms,
> return 0;
> }
>
> +int xen_hvm_acpi_setup(PCMachineState *pcms)
> +{
> + AcpiBuildTables *hvm_acpi_tables;
> + GArray *tables_blob, *table_offsets;
> +
> + ram_addr_t acpi_tables_addr, acpi_tables_size;
> + void *host;
> +
> + struct xs_handle *xs = NULL;
> + char path[80], value[17];
> +
> + if (!pcms->nvdimm) {
> + return 0;
> + }
> +
> + hvm_acpi_tables = g_malloc0(sizeof(AcpiBuildTables));
> + if (!hvm_acpi_tables) {
> + return -1;
> + }
> + acpi_build_tables_init(hvm_acpi_tables);
> + tables_blob = hvm_acpi_tables->table_data;
> + table_offsets = g_array_new(false, true, sizeof(uint32_t));
> + bios_linker_loader_alloc(hvm_acpi_tables->linker,
> + ACPI_BUILD_TABLE_FILE, 64, false);
> +
> + /* build NFIT tables */
> + nvdimm_build_acpi(table_offsets, tables_blob, hvm_acpi_tables->linker);
> + g_array_free(table_offsets, true);
> +
> + /* copy APCI tables into VM */
> + acpi_tables_size = tables_blob->len;
> + acpi_tables_addr =
> + (pcms->below_4g_mem_size - acpi_tables_size) & XC_PAGE_MASK;
> + host = xc_map_foreign_range(xen_xc, xen_domid,
> + ROUND_UP(acpi_tables_size, XC_PAGE_SIZE),
> + PROT_READ | PROT_WRITE,
> + acpi_tables_addr >> XC_PAGE_SHIFT);
> + memcpy(host, tables_blob->data, acpi_tables_size);
> + munmap(host, ROUND_UP(acpi_tables_size, XC_PAGE_SIZE));
> +
> + /* write address and size of ACPI tables to xenstore */
> + xs = xs_open(0);
> + if (xs == NULL) {
> + error_report("could not contact XenStore\n");
> + return -1;
> + }
> + snprintf(path, sizeof(path),
> + "/local/domain/%d/hvmloader/dm-acpi/address", xen_domid);
> + snprintf(value, sizeof(value), "%"PRIu64, (uint64_t) acpi_tables_addr);
> + if (!xs_write(xs, 0, path, value, strlen(value))) {
> + error_report("failed to write NFIT base address to xenstore\n");
> + return -1;
> + }
> + snprintf(path, sizeof(path),
> + "/local/domain/%d/hvmloader/dm-acpi/length", xen_domid);
> + snprintf(value, sizeof(value), "%"PRIu64, (uint64_t) acpi_tables_size);
> + if (!xs_write(xs, 0, path, value, strlen(value))) {
> + error_report("failed to write NFIT size to xenstore\n");
> + return -1;
> + }
> +
> + return 0;
> +}
> +
> void destroy_hvm_domain(bool reboot)
> {
> XenXC xc_handle;
> --
> 2.4.8
>
>
> _______________________________________________
> Xen-devel mailing list
> Xen-devel@xxxxxxxxxxxxx
> http://lists.xen.org/xen-devel
>
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |