|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] Ping: [PATCH RFC v2] x86/domctl: Don't pause the whole domain if only getting vcpu state
Any thoughts appreciated.
On Vi, 2017-10-06 at 13:02 +0300, Alexandru Isaila wrote:
> This patch adds the hvm_save_one_cpu_ctxt() function.
> It optimizes by only pausing the vcpu on all HVMSR_PER_VCPU save
> callbacks where only data for one VCPU is required.
>
> Signed-off-by: Alexandru Isaila <aisaila@xxxxxxxxxxxxxxx>
>
> ---
> Changes since V1:
> - Integrated the vcpu check into all the save callbacks
> ---
> tools/tests/vhpet/emul.h | 3 +-
> tools/tests/vhpet/main.c | 2 +-
> xen/arch/x86/cpu/mcheck/vmce.c | 16 ++-
> xen/arch/x86/domctl.c | 2 -
> xen/arch/x86/hvm/hpet.c | 2 +-
> xen/arch/x86/hvm/hvm.c | 280 ++++++++++++++++++++++++++-----
> ----------
> xen/arch/x86/hvm/i8254.c | 2 +-
> xen/arch/x86/hvm/irq.c | 6 +-
> xen/arch/x86/hvm/mtrr.c | 32 ++++-
> xen/arch/x86/hvm/pmtimer.c | 2 +-
> xen/arch/x86/hvm/rtc.c | 2 +-
> xen/arch/x86/hvm/save.c | 71 ++++++++---
> xen/arch/x86/hvm/vioapic.c | 2 +-
> xen/arch/x86/hvm/viridian.c | 17 ++-
> xen/arch/x86/hvm/vlapic.c | 23 +++-
> xen/arch/x86/hvm/vpic.c | 2 +-
> xen/include/asm-x86/hvm/hvm.h | 2 +
> xen/include/asm-x86/hvm/save.h | 5 +-
> 18 files changed, 324 insertions(+), 147 deletions(-)
>
> diff --git a/tools/tests/vhpet/emul.h b/tools/tests/vhpet/emul.h
> index 383acff..99d5bbd 100644
> --- a/tools/tests/vhpet/emul.h
> +++ b/tools/tests/vhpet/emul.h
> @@ -296,7 +296,8 @@ struct hvm_hw_hpet
> };
>
> typedef int (*hvm_save_handler)(struct domain *d,
> - hvm_domain_context_t *h);
> + hvm_domain_context_t *h,
> + unsigned int instance);
> typedef int (*hvm_load_handler)(struct domain *d,
> hvm_domain_context_t *h);
>
> diff --git a/tools/tests/vhpet/main.c b/tools/tests/vhpet/main.c
> index 6fe65ea..3d8e7f5 100644
> --- a/tools/tests/vhpet/main.c
> +++ b/tools/tests/vhpet/main.c
> @@ -177,7 +177,7 @@ void __init hvm_register_savevm(uint16_t
> typecode,
>
> int do_save(uint16_t typecode, struct domain *d,
> hvm_domain_context_t *h)
> {
> - return hvm_sr_handlers[typecode].save(d, h);
> + return hvm_sr_handlers[typecode].save(d, h, d->max_vcpus);
> }
>
> int do_load(uint16_t typecode, struct domain *d,
> hvm_domain_context_t *h)
> diff --git a/xen/arch/x86/cpu/mcheck/vmce.c
> b/xen/arch/x86/cpu/mcheck/vmce.c
> index e07cd2f..a1a12a5 100644
> --- a/xen/arch/x86/cpu/mcheck/vmce.c
> +++ b/xen/arch/x86/cpu/mcheck/vmce.c
> @@ -349,12 +349,24 @@ int vmce_wrmsr(uint32_t msr, uint64_t val)
> return ret;
> }
>
> -static int vmce_save_vcpu_ctxt(struct domain *d,
> hvm_domain_context_t *h)
> +static int vmce_save_vcpu_ctxt(struct domain *d,
> hvm_domain_context_t *h, unsigned int instance)
> {
> struct vcpu *v;
> int err = 0;
>
> - for_each_vcpu ( d, v )
> + if( instance < d->max_vcpus )
> + {
> + struct hvm_vmce_vcpu ctxt;
> +
> + v = d->vcpu[instance];
> + ctxt.caps = v->arch.vmce.mcg_cap;
> + ctxt.mci_ctl2_bank0 = v->arch.vmce.bank[0].mci_ctl2;
> + ctxt.mci_ctl2_bank1 = v->arch.vmce.bank[1].mci_ctl2;
> + ctxt.mcg_ext_ctl = v->arch.vmce.mcg_ext_ctl;
> +
> + err = hvm_save_entry(VMCE_VCPU, v->vcpu_id, h, &ctxt);
> + }
> + else for_each_vcpu ( d, v )
> {
> struct hvm_vmce_vcpu ctxt = {
> .caps = v->arch.vmce.mcg_cap,
> diff --git a/xen/arch/x86/domctl.c b/xen/arch/x86/domctl.c
> index 540ba08..d3c4e14 100644
> --- a/xen/arch/x86/domctl.c
> +++ b/xen/arch/x86/domctl.c
> @@ -624,12 +624,10 @@ long arch_do_domctl(
> !is_hvm_domain(d) )
> break;
>
> - domain_pause(d);
> ret = hvm_save_one(d, domctl->u.hvmcontext_partial.type,
> domctl->u.hvmcontext_partial.instance,
> domctl->u.hvmcontext_partial.buffer,
> &domctl->u.hvmcontext_partial.bufsz);
> - domain_unpause(d);
>
> if ( !ret )
> copyback = true;
> diff --git a/xen/arch/x86/hvm/hpet.c b/xen/arch/x86/hvm/hpet.c
> index 3ea895a..56f4691 100644
> --- a/xen/arch/x86/hvm/hpet.c
> +++ b/xen/arch/x86/hvm/hpet.c
> @@ -509,7 +509,7 @@ static const struct hvm_mmio_ops hpet_mmio_ops =
> {
> };
>
>
> -static int hpet_save(struct domain *d, hvm_domain_context_t *h)
> +static int hpet_save(struct domain *d, hvm_domain_context_t *h,
> unsigned int instance)
> {
> HPETState *hp = domain_vhpet(d);
> struct vcpu *v = pt_global_vcpu_target(d);
> diff --git a/xen/arch/x86/hvm/hvm.c b/xen/arch/x86/hvm/hvm.c
> index 205b4cb..140f2c3 100644
> --- a/xen/arch/x86/hvm/hvm.c
> +++ b/xen/arch/x86/hvm/hvm.c
> @@ -728,13 +728,19 @@ void hvm_domain_destroy(struct domain *d)
> }
> }
>
> -static int hvm_save_tsc_adjust(struct domain *d,
> hvm_domain_context_t *h)
> +static int hvm_save_tsc_adjust(struct domain *d,
> hvm_domain_context_t *h, unsigned int instance)
> {
> struct vcpu *v;
> struct hvm_tsc_adjust ctxt;
> int err = 0;
>
> - for_each_vcpu ( d, v )
> + if( instance < d->max_vcpus )
> + {
> + v = d->vcpu[instance];
> + ctxt.tsc_adjust = v->arch.hvm_vcpu.msr_tsc_adjust;
> + err = hvm_save_entry(TSC_ADJUST, v->vcpu_id, h, &ctxt);
> + }
> + else for_each_vcpu ( d, v )
> {
> ctxt.tsc_adjust = v->arch.hvm_vcpu.msr_tsc_adjust;
> err = hvm_save_entry(TSC_ADJUST, v->vcpu_id, h, &ctxt);
> @@ -768,117 +774,135 @@ static int hvm_load_tsc_adjust(struct domain
> *d, hvm_domain_context_t *h)
> HVM_REGISTER_SAVE_RESTORE(TSC_ADJUST, hvm_save_tsc_adjust,
> hvm_load_tsc_adjust, 1, HVMSR_PER_VCPU);
>
> -static int hvm_save_cpu_ctxt(struct domain *d, hvm_domain_context_t
> *h)
> +void hvm_save_one_cpu_ctxt(struct vcpu *v, struct hvm_hw_cpu *ctxt)
> +{
> + struct segment_register seg;
> +
> + /* Architecture-specific vmcs/vmcb bits */
> + hvm_funcs.save_cpu_ctxt(v, ctxt);
> +
> + ctxt->tsc = hvm_get_guest_tsc_fixed(v, v->domain-
> >arch.hvm_domain.sync_tsc);
> +
> + ctxt->msr_tsc_aux = hvm_msr_tsc_aux(v);
> +
> + hvm_get_segment_register(v, x86_seg_idtr, &seg);
> + ctxt->idtr_limit = seg.limit;
> + ctxt->idtr_base = seg.base;
> +
> + hvm_get_segment_register(v, x86_seg_gdtr, &seg);
> + ctxt->gdtr_limit = seg.limit;
> + ctxt->gdtr_base = seg.base;
> +
> + hvm_get_segment_register(v, x86_seg_cs, &seg);
> + ctxt->cs_sel = seg.sel;
> + ctxt->cs_limit = seg.limit;
> + ctxt->cs_base = seg.base;
> + ctxt->cs_arbytes = seg.attr;
> +
> + hvm_get_segment_register(v, x86_seg_ds, &seg);
> + ctxt->ds_sel = seg.sel;
> + ctxt->ds_limit = seg.limit;
> + ctxt->ds_base = seg.base;
> + ctxt->ds_arbytes = seg.attr;
> +
> + hvm_get_segment_register(v, x86_seg_es, &seg);
> + ctxt->es_sel = seg.sel;
> + ctxt->es_limit = seg.limit;
> + ctxt->es_base = seg.base;
> + ctxt->es_arbytes = seg.attr;
> +
> + hvm_get_segment_register(v, x86_seg_ss, &seg);
> + ctxt->ss_sel = seg.sel;
> + ctxt->ss_limit = seg.limit;
> + ctxt->ss_base = seg.base;
> + ctxt->ss_arbytes = seg.attr;
> +
> + hvm_get_segment_register(v, x86_seg_fs, &seg);
> + ctxt->fs_sel = seg.sel;
> + ctxt->fs_limit = seg.limit;
> + ctxt->fs_base = seg.base;
> + ctxt->fs_arbytes = seg.attr;
> +
> + hvm_get_segment_register(v, x86_seg_gs, &seg);
> + ctxt->gs_sel = seg.sel;
> + ctxt->gs_limit = seg.limit;
> + ctxt->gs_base = seg.base;
> + ctxt->gs_arbytes = seg.attr;
> +
> + hvm_get_segment_register(v, x86_seg_tr, &seg);
> + ctxt->tr_sel = seg.sel;
> + ctxt->tr_limit = seg.limit;
> + ctxt->tr_base = seg.base;
> + ctxt->tr_arbytes = seg.attr;
> +
> + hvm_get_segment_register(v, x86_seg_ldtr, &seg);
> + ctxt->ldtr_sel = seg.sel;
> + ctxt->ldtr_limit = seg.limit;
> + ctxt->ldtr_base = seg.base;
> + ctxt->ldtr_arbytes = seg.attr;
> +
> + if ( v->fpu_initialised )
> + {
> + memcpy(ctxt->fpu_regs, v->arch.fpu_ctxt, sizeof(ctxt-
> >fpu_regs));
> + ctxt->flags = XEN_X86_FPU_INITIALISED;
> + }
> +
> + ctxt->rax = v->arch.user_regs.rax;
> + ctxt->rbx = v->arch.user_regs.rbx;
> + ctxt->rcx = v->arch.user_regs.rcx;
> + ctxt->rdx = v->arch.user_regs.rdx;
> + ctxt->rbp = v->arch.user_regs.rbp;
> + ctxt->rsi = v->arch.user_regs.rsi;
> + ctxt->rdi = v->arch.user_regs.rdi;
> + ctxt->rsp = v->arch.user_regs.rsp;
> + ctxt->rip = v->arch.user_regs.rip;
> + ctxt->rflags = v->arch.user_regs.rflags;
> + ctxt->r8 = v->arch.user_regs.r8;
> + ctxt->r9 = v->arch.user_regs.r9;
> + ctxt->r10 = v->arch.user_regs.r10;
> + ctxt->r11 = v->arch.user_regs.r11;
> + ctxt->r12 = v->arch.user_regs.r12;
> + ctxt->r13 = v->arch.user_regs.r13;
> + ctxt->r14 = v->arch.user_regs.r14;
> + ctxt->r15 = v->arch.user_regs.r15;
> + ctxt->dr0 = v->arch.debugreg[0];
> + ctxt->dr1 = v->arch.debugreg[1];
> + ctxt->dr2 = v->arch.debugreg[2];
> + ctxt->dr3 = v->arch.debugreg[3];
> + ctxt->dr6 = v->arch.debugreg[6];
> + ctxt->dr7 = v->arch.debugreg[7];
> +}
> +
> +static int hvm_save_cpu_ctxt(struct domain *d, hvm_domain_context_t
> *h, unsigned int instance)
> {
> struct vcpu *v;
> struct hvm_hw_cpu ctxt;
> - struct segment_register seg;
>
> - for_each_vcpu ( d, v )
> + if( instance < d->max_vcpus)
> + {
> + v = d->vcpu[instance];
> + if ( v->pause_flags & VPF_down )
> + return 1;
> + memset(&ctxt, 0, sizeof(ctxt));
> +
> + hvm_save_one_cpu_ctxt(v, &ctxt);
> +
> + if ( hvm_save_entry(CPU, v->vcpu_id, h, &ctxt) != 0 )
> + return 1;
> + }
> + else for_each_vcpu ( d, v )
> {
> - /* We don't need to save state for a vcpu that is down; the
> restore
> + /* We don't need to save state for a vcpu that is down; the
> restore
> * code will leave it down if there is nothing saved. */
> if ( v->pause_flags & VPF_down )
> continue;
>
> memset(&ctxt, 0, sizeof(ctxt));
>
> - /* Architecture-specific vmcs/vmcb bits */
> - hvm_funcs.save_cpu_ctxt(v, &ctxt);
> -
> - ctxt.tsc = hvm_get_guest_tsc_fixed(v, d-
> >arch.hvm_domain.sync_tsc);
> -
> - ctxt.msr_tsc_aux = hvm_msr_tsc_aux(v);
> -
> - hvm_get_segment_register(v, x86_seg_idtr, &seg);
> - ctxt.idtr_limit = seg.limit;
> - ctxt.idtr_base = seg.base;
> -
> - hvm_get_segment_register(v, x86_seg_gdtr, &seg);
> - ctxt.gdtr_limit = seg.limit;
> - ctxt.gdtr_base = seg.base;
> -
> - hvm_get_segment_register(v, x86_seg_cs, &seg);
> - ctxt.cs_sel = seg.sel;
> - ctxt.cs_limit = seg.limit;
> - ctxt.cs_base = seg.base;
> - ctxt.cs_arbytes = seg.attr;
> -
> - hvm_get_segment_register(v, x86_seg_ds, &seg);
> - ctxt.ds_sel = seg.sel;
> - ctxt.ds_limit = seg.limit;
> - ctxt.ds_base = seg.base;
> - ctxt.ds_arbytes = seg.attr;
> -
> - hvm_get_segment_register(v, x86_seg_es, &seg);
> - ctxt.es_sel = seg.sel;
> - ctxt.es_limit = seg.limit;
> - ctxt.es_base = seg.base;
> - ctxt.es_arbytes = seg.attr;
> -
> - hvm_get_segment_register(v, x86_seg_ss, &seg);
> - ctxt.ss_sel = seg.sel;
> - ctxt.ss_limit = seg.limit;
> - ctxt.ss_base = seg.base;
> - ctxt.ss_arbytes = seg.attr;
> -
> - hvm_get_segment_register(v, x86_seg_fs, &seg);
> - ctxt.fs_sel = seg.sel;
> - ctxt.fs_limit = seg.limit;
> - ctxt.fs_base = seg.base;
> - ctxt.fs_arbytes = seg.attr;
> -
> - hvm_get_segment_register(v, x86_seg_gs, &seg);
> - ctxt.gs_sel = seg.sel;
> - ctxt.gs_limit = seg.limit;
> - ctxt.gs_base = seg.base;
> - ctxt.gs_arbytes = seg.attr;
> -
> - hvm_get_segment_register(v, x86_seg_tr, &seg);
> - ctxt.tr_sel = seg.sel;
> - ctxt.tr_limit = seg.limit;
> - ctxt.tr_base = seg.base;
> - ctxt.tr_arbytes = seg.attr;
> -
> - hvm_get_segment_register(v, x86_seg_ldtr, &seg);
> - ctxt.ldtr_sel = seg.sel;
> - ctxt.ldtr_limit = seg.limit;
> - ctxt.ldtr_base = seg.base;
> - ctxt.ldtr_arbytes = seg.attr;
> -
> - if ( v->fpu_initialised )
> - {
> - memcpy(ctxt.fpu_regs, v->arch.fpu_ctxt,
> sizeof(ctxt.fpu_regs));
> - ctxt.flags = XEN_X86_FPU_INITIALISED;
> - }
> -
> - ctxt.rax = v->arch.user_regs.rax;
> - ctxt.rbx = v->arch.user_regs.rbx;
> - ctxt.rcx = v->arch.user_regs.rcx;
> - ctxt.rdx = v->arch.user_regs.rdx;
> - ctxt.rbp = v->arch.user_regs.rbp;
> - ctxt.rsi = v->arch.user_regs.rsi;
> - ctxt.rdi = v->arch.user_regs.rdi;
> - ctxt.rsp = v->arch.user_regs.rsp;
> - ctxt.rip = v->arch.user_regs.rip;
> - ctxt.rflags = v->arch.user_regs.rflags;
> - ctxt.r8 = v->arch.user_regs.r8;
> - ctxt.r9 = v->arch.user_regs.r9;
> - ctxt.r10 = v->arch.user_regs.r10;
> - ctxt.r11 = v->arch.user_regs.r11;
> - ctxt.r12 = v->arch.user_regs.r12;
> - ctxt.r13 = v->arch.user_regs.r13;
> - ctxt.r14 = v->arch.user_regs.r14;
> - ctxt.r15 = v->arch.user_regs.r15;
> - ctxt.dr0 = v->arch.debugreg[0];
> - ctxt.dr1 = v->arch.debugreg[1];
> - ctxt.dr2 = v->arch.debugreg[2];
> - ctxt.dr3 = v->arch.debugreg[3];
> - ctxt.dr6 = v->arch.debugreg[6];
> - ctxt.dr7 = v->arch.debugreg[7];
> + hvm_save_one_cpu_ctxt(v, &ctxt);
>
> if ( hvm_save_entry(CPU, v->vcpu_id, h, &ctxt) != 0 )
> - return 1;
> + return 1;
> }
> return 0;
> }
> @@ -1162,7 +1186,8 @@ HVM_REGISTER_SAVE_RESTORE(CPU,
> hvm_save_cpu_ctxt, hvm_load_cpu_ctxt,
> save_area) + \
> xstate_ctxt_size(xcr0))
>
> -static int hvm_save_cpu_xsave_states(struct domain *d,
> hvm_domain_context_t *h)
> +static int hvm_save_cpu_xsave_states(struct domain *d,
> hvm_domain_context_t *h,
> + unsigned int instance)
> {
> struct vcpu *v;
> struct hvm_hw_cpu_xsave *ctxt;
> @@ -1170,7 +1195,27 @@ static int hvm_save_cpu_xsave_states(struct
> domain *d, hvm_domain_context_t *h)
> if ( !cpu_has_xsave )
> return 0; /* do nothing */
>
> - for_each_vcpu ( d, v )
> + if( instance < d->max_vcpus )
> + {
> + unsigned int size;
> +
> + v = d->vcpu[instance];
> + size = HVM_CPU_XSAVE_SIZE(v->arch.xcr0_accum);
> +
> + if ( !xsave_enabled(v) )
> + return 1;
> + if ( _hvm_init_entry(h, CPU_XSAVE_CODE, v->vcpu_id, size) )
> + return 1;
> + ctxt = (struct hvm_hw_cpu_xsave *)&h->data[h->cur];
> + h->cur += size;
> +
> + ctxt->xfeature_mask = xfeature_mask;
> + ctxt->xcr0 = v->arch.xcr0;
> + ctxt->xcr0_accum = v->arch.xcr0_accum;
> + expand_xsave_states(v, &ctxt->save_area,
> + size - offsetof(typeof(*ctxt),
> save_area));
> + }
> + else for_each_vcpu ( d, v )
> {
> unsigned int size = HVM_CPU_XSAVE_SIZE(v->arch.xcr0_accum);
>
> @@ -1324,10 +1369,39 @@ static int hvm_load_cpu_xsave_states(struct
> domain *d, hvm_domain_context_t *h)
> #define HVM_CPU_MSR_SIZE(cnt) offsetof(struct hvm_msr, msr[cnt])
> static unsigned int __read_mostly msr_count_max;
>
> -static int hvm_save_cpu_msrs(struct domain *d, hvm_domain_context_t
> *h)
> +static int hvm_save_cpu_msrs(struct domain *d, hvm_domain_context_t
> *h,
> + unsigned int instance)
> {
> struct vcpu *v;
>
> + if( instance < d->max_vcpus )
> + {
> + struct hvm_msr *ctxt;
> + unsigned int i;
> +
> + v = d->vcpu[instance];
> +
> + if ( _hvm_init_entry(h, CPU_MSR_CODE, v->vcpu_id,
> + HVM_CPU_MSR_SIZE(msr_count_max)) )
> + return 1;
> + ctxt = (struct hvm_msr *)&h->data[h->cur];
> + ctxt->count = 0;
> +
> + if ( hvm_funcs.save_msr )
> + hvm_funcs.save_msr(v, ctxt);
> +
> + ASSERT(ctxt->count <= msr_count_max);
> +
> + for ( i = 0; i < ctxt->count; ++i )
> + ctxt->msr[i]._rsvd = 0;
> +
> + if ( ctxt->count )
> + h->cur += HVM_CPU_MSR_SIZE(ctxt->count);
> + else
> + h->cur -= sizeof(struct hvm_save_descriptor);
> +
> + }
> +
> for_each_vcpu ( d, v )
> {
> struct hvm_msr *ctxt;
> diff --git a/xen/arch/x86/hvm/i8254.c b/xen/arch/x86/hvm/i8254.c
> index 992f08d..143b64d 100644
> --- a/xen/arch/x86/hvm/i8254.c
> +++ b/xen/arch/x86/hvm/i8254.c
> @@ -390,7 +390,7 @@ void pit_stop_channel0_irq(PITState *pit)
> spin_unlock(&pit->lock);
> }
>
> -static int pit_save(struct domain *d, hvm_domain_context_t *h)
> +static int pit_save(struct domain *d, hvm_domain_context_t *h,
> unsigned int instance)
> {
> PITState *pit = domain_vpit(d);
> int rc;
> diff --git a/xen/arch/x86/hvm/irq.c b/xen/arch/x86/hvm/irq.c
> index e425df9..dbbf769 100644
> --- a/xen/arch/x86/hvm/irq.c
> +++ b/xen/arch/x86/hvm/irq.c
> @@ -598,7 +598,7 @@ static int __init dump_irq_info_key_init(void)
> }
> __initcall(dump_irq_info_key_init);
>
> -static int irq_save_pci(struct domain *d, hvm_domain_context_t *h)
> +static int irq_save_pci(struct domain *d, hvm_domain_context_t *h,
> unsigned int instance)
> {
> struct hvm_irq *hvm_irq = hvm_domain_irq(d);
> unsigned int asserted, pdev, pintx;
> @@ -630,7 +630,7 @@ static int irq_save_pci(struct domain *d,
> hvm_domain_context_t *h)
> return rc;
> }
>
> -static int irq_save_isa(struct domain *d, hvm_domain_context_t *h)
> +static int irq_save_isa(struct domain *d, hvm_domain_context_t *h,
> unsigned int instance)
> {
> struct hvm_irq *hvm_irq = hvm_domain_irq(d);
>
> @@ -638,7 +638,7 @@ static int irq_save_isa(struct domain *d,
> hvm_domain_context_t *h)
> return ( hvm_save_entry(ISA_IRQ, 0, h, &hvm_irq->isa_irq) );
> }
>
> -static int irq_save_link(struct domain *d, hvm_domain_context_t *h)
> +static int irq_save_link(struct domain *d, hvm_domain_context_t *h,
> unsigned int instance)
> {
> struct hvm_irq *hvm_irq = hvm_domain_irq(d);
>
> diff --git a/xen/arch/x86/hvm/mtrr.c b/xen/arch/x86/hvm/mtrr.c
> index b721c63..b998d80 100644
> --- a/xen/arch/x86/hvm/mtrr.c
> +++ b/xen/arch/x86/hvm/mtrr.c
> @@ -666,14 +666,42 @@ int hvm_set_mem_pinned_cacheattr(struct domain
> *d, uint64_t gfn_start,
> return 0;
> }
>
> -static int hvm_save_mtrr_msr(struct domain *d, hvm_domain_context_t
> *h)
> +static int hvm_save_mtrr_msr(struct domain *d, hvm_domain_context_t
> *h, unsigned int instance)
> {
> int i;
> struct vcpu *v;
> struct hvm_hw_mtrr hw_mtrr;
> struct mtrr_state *mtrr_state;
> /* save mtrr&pat */
> - for_each_vcpu(d, v)
> + if( instance < d->max_vcpus )
> + {
> + v = d->vcpu[instance];
> + mtrr_state = &v->arch.hvm_vcpu.mtrr;
> +
> + hvm_get_guest_pat(v, &hw_mtrr.msr_pat_cr);
> +
> + hw_mtrr.msr_mtrr_def_type = mtrr_state->def_type
> + | (mtrr_state->enabled << 10);
> + hw_mtrr.msr_mtrr_cap = mtrr_state->mtrr_cap;
> +
> + for ( i = 0; i < MTRR_VCNT; i++ )
> + {
> + /* save physbase */
> + hw_mtrr.msr_mtrr_var[i*2] =
> + ((uint64_t*)mtrr_state->var_ranges)[i*2];
> + /* save physmask */
> + hw_mtrr.msr_mtrr_var[i*2+1] =
> + ((uint64_t*)mtrr_state->var_ranges)[i*2+1];
> + }
> +
> + for ( i = 0; i < NUM_FIXED_MSR; i++ )
> + hw_mtrr.msr_mtrr_fixed[i] =
> + ((uint64_t*)mtrr_state->fixed_ranges)[i];
> +
> + if ( hvm_save_entry(MTRR, v->vcpu_id, h, &hw_mtrr) != 0 )
> + return 1;
> + }
> + else for_each_vcpu(d, v)
> {
> mtrr_state = &v->arch.hvm_vcpu.mtrr;
>
> diff --git a/xen/arch/x86/hvm/pmtimer.c b/xen/arch/x86/hvm/pmtimer.c
> index b70c299..21dcdeb 100644
> --- a/xen/arch/x86/hvm/pmtimer.c
> +++ b/xen/arch/x86/hvm/pmtimer.c
> @@ -249,7 +249,7 @@ static int handle_pmt_io(
> return X86EMUL_OKAY;
> }
>
> -static int acpi_save(struct domain *d, hvm_domain_context_t *h)
> +static int acpi_save(struct domain *d, hvm_domain_context_t *h,
> unsigned int instance)
> {
> struct hvm_hw_acpi *acpi = &d->arch.hvm_domain.acpi;
> PMTState *s = &d->arch.hvm_domain.pl_time->vpmt;
> diff --git a/xen/arch/x86/hvm/rtc.c b/xen/arch/x86/hvm/rtc.c
> index bcfa169..83f339d 100644
> --- a/xen/arch/x86/hvm/rtc.c
> +++ b/xen/arch/x86/hvm/rtc.c
> @@ -737,7 +737,7 @@ void rtc_migrate_timers(struct vcpu *v)
> }
>
> /* Save RTC hardware state */
> -static int rtc_save(struct domain *d, hvm_domain_context_t *h)
> +static int rtc_save(struct domain *d, hvm_domain_context_t *h,
> unsigned int instance)
> {
> RTCState *s = domain_vrtc(d);
> int rc;
> diff --git a/xen/arch/x86/hvm/save.c b/xen/arch/x86/hvm/save.c
> index 8984a23..97b56f7 100644
> --- a/xen/arch/x86/hvm/save.c
> +++ b/xen/arch/x86/hvm/save.c
> @@ -138,6 +138,7 @@ int hvm_save_one(struct domain *d, unsigned int
> typecode, unsigned int instance,
> int rv;
> hvm_domain_context_t ctxt = { };
> const struct hvm_save_descriptor *desc;
> + bool is_single_instance = false;
>
> if ( d->is_dying ||
> typecode > HVM_SAVE_CODE_MAX ||
> @@ -145,41 +146,75 @@ int hvm_save_one(struct domain *d, unsigned int
> typecode, unsigned int instance,
> !hvm_sr_handlers[typecode].save )
> return -EINVAL;
>
> + if( hvm_sr_handlers[typecode].kind == HVMSR_PER_VCPU &&
> + instance < d->max_vcpus )
> + is_single_instance = true;
> +
> ctxt.size = hvm_sr_handlers[typecode].size;
> - if ( hvm_sr_handlers[typecode].kind == HVMSR_PER_VCPU )
> + if ( hvm_sr_handlers[typecode].kind == HVMSR_PER_VCPU &&
> + instance == d->max_vcpus )
> ctxt.size *= d->max_vcpus;
> +
> ctxt.data = xmalloc_bytes(ctxt.size);
> if ( !ctxt.data )
> return -ENOMEM;
>
> - if ( (rv = hvm_sr_handlers[typecode].save(d, &ctxt)) != 0 )
> + if( is_single_instance )
> + vcpu_pause(d->vcpu[instance]);
> + else
> + domain_pause(d);
> +
> + if ( (rv = hvm_sr_handlers[typecode].save(d, &ctxt, instance))
> != 0 )
> printk(XENLOG_G_ERR "HVM%d save: failed to save type
> %"PRIu16" (%d)\n",
> d->domain_id, typecode, rv);
> else if ( rv = -ENOENT, ctxt.cur >= sizeof(*desc) )
> {
> uint32_t off;
>
> - for ( off = 0; off <= (ctxt.cur - sizeof(*desc)); off +=
> desc->length )
> + if( is_single_instance )
> {
> - desc = (void *)(ctxt.data + off);
> + desc = (void *)(ctxt.data);
> /* Move past header */
> - off += sizeof(*desc);
> + off = sizeof(*desc);
> if ( ctxt.cur < desc->length ||
> off > ctxt.cur - desc->length )
> - break;
> - if ( instance == desc->instance )
> + rv = -EFAULT;
> + rv = 0;
> + if ( guest_handle_is_null(handle) )
> + *bufsz = desc->length;
> + else if ( *bufsz < desc->length )
> + rv = -ENOBUFS;
> + else if ( copy_to_guest(handle, ctxt.data + off, desc-
> >length) )
> + rv = -EFAULT;
> + else
> + *bufsz = desc->length;
> + vcpu_unpause(d->vcpu[instance]);
> + }
> + else
> + {
> + for ( off = 0; off <= (ctxt.cur - sizeof(*desc)); off +=
> desc->length )
> {
> - rv = 0;
> - if ( guest_handle_is_null(handle) )
> - *bufsz = desc->length;
> - else if ( *bufsz < desc->length )
> - rv = -ENOBUFS;
> - else if ( copy_to_guest(handle, ctxt.data + off,
> desc->length) )
> - rv = -EFAULT;
> - else
> - *bufsz = desc->length;
> - break;
> + desc = (void *)(ctxt.data + off);
> + /* Move past header */
> + off += sizeof(*desc);
> + if ( ctxt.cur < desc->length ||
> + off > ctxt.cur - desc->length )
> + break;
> + if ( instance == desc->instance )
> + {
> + rv = 0;
> + if ( guest_handle_is_null(handle) )
> + *bufsz = desc->length;
> + else if ( *bufsz < desc->length )
> + rv = -ENOBUFS;
> + else if ( copy_to_guest(handle, ctxt.data + off,
> desc->length) )
> + rv = -EFAULT;
> + else
> + *bufsz = desc->length;
> + break;
> + }
> }
> + domain_unpause(d);
> }
> }
>
> @@ -225,7 +260,7 @@ int hvm_save(struct domain *d,
> hvm_domain_context_t *h)
> {
> printk(XENLOG_G_INFO "HVM%d save: %s\n",
> d->domain_id, hvm_sr_handlers[i].name);
> - if ( handler(d, h) != 0 )
> + if ( handler(d, h, d->max_vcpus) != 0 )
> {
> printk(XENLOG_G_ERR
> "HVM%d save: failed to save type
> %"PRIu16"\n",
> diff --git a/xen/arch/x86/hvm/vioapic.c b/xen/arch/x86/hvm/vioapic.c
> index 97b419f..34d6907 100644
> --- a/xen/arch/x86/hvm/vioapic.c
> +++ b/xen/arch/x86/hvm/vioapic.c
> @@ -569,7 +569,7 @@ int vioapic_get_trigger_mode(const struct domain
> *d, unsigned int gsi)
> return vioapic->redirtbl[pin].fields.trig_mode;
> }
>
> -static int ioapic_save(struct domain *d, hvm_domain_context_t *h)
> +static int ioapic_save(struct domain *d, hvm_domain_context_t *h,
> unsigned int instance)
> {
> struct hvm_vioapic *s;
>
> diff --git a/xen/arch/x86/hvm/viridian.c
> b/xen/arch/x86/hvm/viridian.c
> index f0fa59d..5943bf4 100644
> --- a/xen/arch/x86/hvm/viridian.c
> +++ b/xen/arch/x86/hvm/viridian.c
> @@ -994,7 +994,7 @@ out:
> return HVM_HCALL_completed;
> }
>
> -static int viridian_save_domain_ctxt(struct domain *d,
> hvm_domain_context_t *h)
> +static int viridian_save_domain_ctxt(struct domain *d,
> hvm_domain_context_t *h, unsigned int instance)
> {
> struct hvm_viridian_domain_context ctxt = {
> .time_ref_count = d-
> >arch.hvm_domain.viridian.time_ref_count.val,
> @@ -1030,14 +1030,25 @@ static int viridian_load_domain_ctxt(struct
> domain *d, hvm_domain_context_t *h)
> HVM_REGISTER_SAVE_RESTORE(VIRIDIAN_DOMAIN,
> viridian_save_domain_ctxt,
> viridian_load_domain_ctxt, 1,
> HVMSR_PER_DOM);
>
> -static int viridian_save_vcpu_ctxt(struct domain *d,
> hvm_domain_context_t *h)
> +static int viridian_save_vcpu_ctxt(struct domain *d,
> hvm_domain_context_t *h, unsigned int instance)
> {
> struct vcpu *v;
>
> if ( !is_viridian_domain(d) )
> return 0;
>
> - for_each_vcpu( d, v ) {
> + if( instance < d->max_vcpus )
> + {
> + struct hvm_viridian_vcpu_context ctxt;
> +
> + v = d->vcpu[instance];
> + ctxt.vp_assist_msr = v-
> >arch.hvm_vcpu.viridian.vp_assist.msr.raw;
> + ctxt.vp_assist_vector = v-
> >arch.hvm_vcpu.viridian.vp_assist.vector;
> +
> + if ( hvm_save_entry(VIRIDIAN_VCPU, v->vcpu_id, h, &ctxt) !=
> 0 )
> + return 1;
> + }
> + else for_each_vcpu( d, v ) {
> struct hvm_viridian_vcpu_context ctxt = {
> .vp_assist_msr = v-
> >arch.hvm_vcpu.viridian.vp_assist.msr.raw,
> .vp_assist_vector = v-
> >arch.hvm_vcpu.viridian.vp_assist.vector,
> diff --git a/xen/arch/x86/hvm/vlapic.c b/xen/arch/x86/hvm/vlapic.c
> index 4bfc53e..591631a 100644
> --- a/xen/arch/x86/hvm/vlapic.c
> +++ b/xen/arch/x86/hvm/vlapic.c
> @@ -1387,7 +1387,7 @@ static void lapic_rearm(struct vlapic *s)
> s->timer_last_update = s->pt.last_plt_gtime;
> }
>
> -static int lapic_save_hidden(struct domain *d, hvm_domain_context_t
> *h)
> +static int lapic_save_hidden(struct domain *d, hvm_domain_context_t
> *h, unsigned int instance)
> {
> struct vcpu *v;
> struct vlapic *s;
> @@ -1396,7 +1396,13 @@ static int lapic_save_hidden(struct domain *d,
> hvm_domain_context_t *h)
> if ( !has_vlapic(d) )
> return 0;
>
> - for_each_vcpu ( d, v )
> + if( instance < d->max_vcpus )
> + {
> + v = d->vcpu[instance];
> + s = vcpu_vlapic(v);
> + rc = hvm_save_entry(LAPIC, v->vcpu_id, h, &s->hw);
> + }
> + else for_each_vcpu ( d, v )
> {
> s = vcpu_vlapic(v);
> if ( (rc = hvm_save_entry(LAPIC, v->vcpu_id, h, &s->hw)) !=
> 0 )
> @@ -1406,7 +1412,7 @@ static int lapic_save_hidden(struct domain *d,
> hvm_domain_context_t *h)
> return rc;
> }
>
> -static int lapic_save_regs(struct domain *d, hvm_domain_context_t
> *h)
> +static int lapic_save_regs(struct domain *d, hvm_domain_context_t
> *h, unsigned int instance)
> {
> struct vcpu *v;
> struct vlapic *s;
> @@ -1415,7 +1421,16 @@ static int lapic_save_regs(struct domain *d,
> hvm_domain_context_t *h)
> if ( !has_vlapic(d) )
> return 0;
>
> - for_each_vcpu ( d, v )
> + if( instance < d->max_vcpus )
> + {
> + v = d->vcpu[instance];
> + if ( hvm_funcs.sync_pir_to_irr )
> + hvm_funcs.sync_pir_to_irr(v);
> +
> + s = vcpu_vlapic(v);
> + rc = hvm_save_entry(LAPIC_REGS, v->vcpu_id, h, s->regs);
> + }
> + else for_each_vcpu ( d, v )
> {
> if ( hvm_funcs.sync_pir_to_irr )
> hvm_funcs.sync_pir_to_irr(v);
> diff --git a/xen/arch/x86/hvm/vpic.c b/xen/arch/x86/hvm/vpic.c
> index e160bbd..6b77f3c 100644
> --- a/xen/arch/x86/hvm/vpic.c
> +++ b/xen/arch/x86/hvm/vpic.c
> @@ -371,7 +371,7 @@ static int vpic_intercept_elcr_io(
> return X86EMUL_OKAY;
> }
>
> -static int vpic_save(struct domain *d, hvm_domain_context_t *h)
> +static int vpic_save(struct domain *d, hvm_domain_context_t *h,
> unsigned int instance)
> {
> struct hvm_hw_vpic *s;
> int i;
> diff --git a/xen/include/asm-x86/hvm/hvm.h b/xen/include/asm-
> x86/hvm/hvm.h
> index b687e03..c4b7b3d 100644
> --- a/xen/include/asm-x86/hvm/hvm.h
> +++ b/xen/include/asm-x86/hvm/hvm.h
> @@ -609,6 +609,8 @@ static inline bool altp2m_vcpu_emulate_ve(struct
> vcpu *v)
> return false;
> }
>
> +void hvm_save_one_cpu_ctxt(struct vcpu *v, struct hvm_hw_cpu *ctxt);
> +
> /* Check CR4/EFER values */
> const char *hvm_efer_valid(const struct vcpu *v, uint64_t value,
> signed int cr0_pg);
> diff --git a/xen/include/asm-x86/hvm/save.h b/xen/include/asm-
> x86/hvm/save.h
> index f889e8f..a2c39c4 100644
> --- a/xen/include/asm-x86/hvm/save.h
> +++ b/xen/include/asm-x86/hvm/save.h
> @@ -95,8 +95,9 @@ static inline uint16_t hvm_load_instance(struct
> hvm_domain_context *h)
> * The save handler may save multiple instances of a type into the
> buffer;
> * the load handler will be called once for each instance found when
> * restoring. Both return non-zero on error. */
> -typedef int (*hvm_save_handler) (struct domain *d,
> - hvm_domain_context_t *h);
> +typedef int (*hvm_save_handler) (struct domain *d,
> + hvm_domain_context_t *h,
> + unsigned int instance);
> typedef int (*hvm_load_handler) (struct domain *d,
> hvm_domain_context_t *h);
>
________________________
This email was scanned by Bitdefender
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxxx
https://lists.xenproject.org/mailman/listinfo/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |