|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH v9 10/15] tools/libxc: x86 HVM save code
Andrew Cooper <andrew.cooper3@xxxxxxxxxx> writes:
> Save the x86 HVM specific parts of the domain. This is considerably simpler
> than an x86 PV domain. Only the HVM_CONTEXT and HVM_PARAMS records are
> needed.
>
> There is no need for any page normalisation.
>
> Signed-off-by: Andrew Cooper <andrew.cooper3@xxxxxxxxxx>
> CC: Ian Campbell <Ian.Campbell@xxxxxxxxxx>
> CC: Ian Jackson <Ian.Jackson@xxxxxxxxxxxxx>
> CC: Wei Liu <wei.liu2@xxxxxxxxxx>
> ---
> tools/libxc/Makefile | 1 +
> tools/libxc/xc_sr_common.h | 1 +
> tools/libxc/xc_sr_save_x86_hvm.c | 220
> ++++++++++++++++++++++++++++++++++++++
> 3 files changed, 222 insertions(+)
> create mode 100644 tools/libxc/xc_sr_save_x86_hvm.c
>
> diff --git a/tools/libxc/Makefile b/tools/libxc/Makefile
> index fc1e668..ba124f4 100644
> --- a/tools/libxc/Makefile
> +++ b/tools/libxc/Makefile
> @@ -58,6 +58,7 @@ GUEST_SRCS-$(CONFIG_X86) += xc_sr_common_x86.c
> GUEST_SRCS-$(CONFIG_X86) += xc_sr_common_x86_pv.c
> GUEST_SRCS-$(CONFIG_X86) += xc_sr_restore_x86_pv.c
> GUEST_SRCS-$(CONFIG_X86) += xc_sr_save_x86_pv.c
> +GUEST_SRCS-$(CONFIG_X86) += xc_sr_save_x86_hvm.c
> GUEST_SRCS-y += xc_sr_restore.c
> GUEST_SRCS-y += xc_sr_save.c
> GUEST_SRCS-y += xc_offline_page.c xc_compression.c
> diff --git a/tools/libxc/xc_sr_common.h b/tools/libxc/xc_sr_common.h
> index 4f33839..ec14d59 100644
> --- a/tools/libxc/xc_sr_common.h
> +++ b/tools/libxc/xc_sr_common.h
> @@ -286,6 +286,7 @@ struct xc_sr_context
> };
>
> extern struct xc_sr_save_ops save_ops_x86_pv;
> +extern struct xc_sr_save_ops save_ops_x86_hvm;
>
> extern struct xc_sr_restore_ops restore_ops_x86_pv;
>
> diff --git a/tools/libxc/xc_sr_save_x86_hvm.c
> b/tools/libxc/xc_sr_save_x86_hvm.c
> new file mode 100644
> index 0000000..0928f19
> --- /dev/null
> +++ b/tools/libxc/xc_sr_save_x86_hvm.c
> @@ -0,0 +1,220 @@
> +#include <assert.h>
> +
> +#include "xc_sr_common_x86.h"
> +
> +#include <xen/hvm/params.h>
> +
> +/*
> + * Query for the HVM context and write an HVM_CONTEXT record into the stream.
> + */
> +static int write_hvm_context(struct xc_sr_context *ctx)
> +{
> + xc_interface *xch = ctx->xch;
> + int rc, hvm_buf_size;
> + struct xc_sr_record hvm_rec =
> + {
> + .type = REC_TYPE_HVM_CONTEXT,
> + };
> +
> + hvm_buf_size = xc_domain_hvm_getcontext(xch, ctx->domid, 0, 0);
> + if ( hvm_buf_size < 0 )
> + {
> + PERROR("Couldn't get HVM context size from Xen");
> + rc = -1;
> + goto out;
> + }
> +
> + hvm_rec.data = malloc(hvm_buf_size);
> + if ( !hvm_rec.data )
> + {
> + PERROR("Couldn't allocate memory");
> + rc = -1;
> + goto out;
> + }
> +
> + hvm_buf_size = xc_domain_hvm_getcontext(xch, ctx->domid,
> + hvm_rec.data, hvm_buf_size);
> + if ( hvm_buf_size < 0 )
> + {
> + PERROR("Couldn't get HVM context from Xen");
> + rc = -1;
> + goto out;
> + }
> +
> + hvm_rec.length = hvm_buf_size;
> + rc = write_record(ctx, &hvm_rec);
> + if ( rc < 0 )
> + {
> + PERROR("error write HVM_CONTEXT record");
> + goto out;
> + }
> +
> + out:
> + free(hvm_rec.data);
> + return rc;
> +}
> +
> +/*
> + * Query for a range of HVM parameters and write an HVM_PARAMS record into
> the
> + * stream.
> + */
> +static int write_hvm_params(struct xc_sr_context *ctx)
> +{
> + static const unsigned int params[] = {
> + HVM_PARAM_STORE_PFN,
> + HVM_PARAM_IOREQ_PFN,
> + HVM_PARAM_BUFIOREQ_PFN,
> + HVM_PARAM_PAGING_RING_PFN,
> + HVM_PARAM_MONITOR_RING_PFN,
> + HVM_PARAM_SHARING_RING_PFN,
> + HVM_PARAM_VM86_TSS,
> + HVM_PARAM_CONSOLE_PFN,
> + HVM_PARAM_ACPI_IOPORTS_LOCATION,
> + HVM_PARAM_VIRIDIAN,
> + HVM_PARAM_IDENT_PT,
> + HVM_PARAM_PAE_ENABLED,
> + HVM_PARAM_VM_GENERATION_ID_ADDR,
> + HVM_PARAM_IOREQ_SERVER_PFN,
> + HVM_PARAM_NR_IOREQ_SERVER_PAGES,
> + };
> +
> + xc_interface *xch = ctx->xch;
> + struct xc_sr_rec_hvm_params_entry entries[ARRAY_SIZE(params)];
> + struct xc_sr_rec_hvm_params hdr = {
> + .count = 0,
> + };
> + struct xc_sr_record rec = {
> + .type = REC_TYPE_HVM_PARAMS,
> + .length = sizeof(hdr),
> + .data = &hdr,
> + };
> + unsigned int i;
> + int rc;
> +
> + for ( i = 0; i < ARRAY_SIZE(params); i++ )
> + {
> + uint32_t index = params[i];
> + uint64_t value;
> +
> + rc = xc_hvm_param_get(xch, ctx->domid, index, &value);
> + if ( rc )
> + {
> + PERROR("Failed to get HVMPARAM at index %u", index);
> + return rc;
> + }
> +
> + if ( value != 0 )
> + {
> + entries[hdr.count].index = index;
> + entries[hdr.count].value = value;
> + hdr.count++;
> + }
> + }
While reviewing my 'soft reset' series Ian C raised a question about the
unsafeness of sequential get/set of HVM params:
http://lists.xen.org/archives/html/xen-devel/2015-01/msg01177.html
In case the concern is valid I think the requirements for migration are
even stricter as we can migrate live. In case it is not, would it be
possible to move your params[] array to some common place so it can be
reused?
> +
> + rc = write_split_record(ctx, &rec, entries, hdr.count *
> sizeof(*entries));
> + if ( rc )
> + PERROR("Failed to write HVM_PARAMS record");
> +
> + return rc;
> +}
> +
> +static xen_pfn_t x86_hvm_pfn_to_gfn(const struct xc_sr_context *ctx,
> + xen_pfn_t pfn)
> +{
> + /* identity map */
> + return pfn;
> +}
> +
> +static int x86_hvm_normalise_page(struct xc_sr_context *ctx,
> + xen_pfn_t type, void **page)
> +{
> + /* no-op */
> + return 0;
> +}
> +
> +static int x86_hvm_setup(struct xc_sr_context *ctx)
> +{
> + xc_interface *xch = ctx->xch;
> +
> + if ( !ctx->save.callbacks->switch_qemu_logdirty )
> + {
> + ERROR("No switch_qemu_logdirty callback provided");
> + errno = EINVAL;
> + return -1;
> + }
> +
> + if ( ctx->save.callbacks->switch_qemu_logdirty(
> + ctx->domid, 1, ctx->save.callbacks->data) )
> + {
> + PERROR("Couldn't enable qemu log-dirty mode");
> + return -1;
> + }
> +
> + ctx->x86_hvm.save.qemu_enabled_logdirty = true;
> +
> + return 0;
> +}
> +
> +static int x86_hvm_start_of_stream(struct xc_sr_context *ctx)
> +{
> + /* no-op */
> + return 0;
> +}
> +
> +static int x86_hvm_end_of_stream(struct xc_sr_context *ctx)
> +{
> + int rc;
> +
> + /* Write the TSC record. */
> + rc = write_tsc_info(ctx);
> + if ( rc )
> + return rc;
> +
> + /* Write the HVM_CONTEXT record. */
> + rc = write_hvm_context(ctx);
> + if ( rc )
> + return rc;
> +
> + /* Write HVM_PARAMS record contains applicable HVM params. */
> + rc = write_hvm_params(ctx);
> + if ( rc )
> + return rc;
> +
> + return rc;
> +}
> +
> +static int x86_hvm_cleanup(struct xc_sr_context *ctx)
> +{
> + xc_interface *xch = ctx->xch;
> +
> + /* If qemu successfully enabled logdirty mode, attempt to disable. */
> + if ( ctx->x86_hvm.save.qemu_enabled_logdirty &&
> + ctx->save.callbacks->switch_qemu_logdirty(
> + ctx->domid, 0, ctx->save.callbacks->data) )
> + {
> + PERROR("Couldn't disable qemu log-dirty mode");
> + return -1;
> + }
> +
> + return 0;
> +}
> +
> +struct xc_sr_save_ops save_ops_x86_hvm =
> +{
> + .pfn_to_gfn = x86_hvm_pfn_to_gfn,
> + .normalise_page = x86_hvm_normalise_page,
> + .setup = x86_hvm_setup,
> + .start_of_stream = x86_hvm_start_of_stream,
> + .end_of_stream = x86_hvm_end_of_stream,
> + .cleanup = x86_hvm_cleanup,
> +};
> +
> +/*
> + * Local variables:
> + * mode: C
> + * c-file-style: "BSD"
> + * c-basic-offset: 4
> + * tab-width: 4
> + * indent-tabs-mode: nil
> + * End:
> + */
--
Vitaly
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |