[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH v10 10/11] x86/hvm: Remove redundant save functions
This patch removes the redundant save functions and renames the save_one* to save. It then changes the domain param to vcpu in the save funcs. Signed-off-by: Alexandru Isaila <aisaila@xxxxxxxxxxxxxxx> --- Changes since V9: - Add enum return type for save funcs --- xen/arch/x86/cpu/mcheck/vmce.c | 24 +++------ xen/arch/x86/hvm/hpet.c | 8 ++- xen/arch/x86/hvm/hvm.c | 112 ++++++++++++----------------------------- xen/arch/x86/hvm/i8254.c | 9 ++-- xen/arch/x86/hvm/irq.c | 24 ++++++--- xen/arch/x86/hvm/mtrr.c | 22 +++----- xen/arch/x86/hvm/pmtimer.c | 10 ++-- xen/arch/x86/hvm/rtc.c | 9 ++-- xen/arch/x86/hvm/save.c | 29 +++-------- xen/arch/x86/hvm/vioapic.c | 10 ++-- xen/arch/x86/hvm/viridian.c | 36 ++++++------- xen/arch/x86/hvm/vlapic.c | 45 +++++++---------- xen/arch/x86/hvm/vpic.c | 12 +++-- xen/include/asm-x86/hvm/save.h | 8 ++- 14 files changed, 148 insertions(+), 210 deletions(-) diff --git a/xen/arch/x86/cpu/mcheck/vmce.c b/xen/arch/x86/cpu/mcheck/vmce.c index 29898a6..c054ae9 100644 --- a/xen/arch/x86/cpu/mcheck/vmce.c +++ b/xen/arch/x86/cpu/mcheck/vmce.c @@ -349,7 +349,8 @@ int vmce_wrmsr(uint32_t msr, uint64_t val) return ret; } -static int vmce_save_vcpu_ctxt_one(struct vcpu *v, hvm_domain_context_t *h) +static enum save_return_type_t vmce_save_vcpu_ctxt(struct vcpu *v, + hvm_domain_context_t *h) { struct hvm_vmce_vcpu ctxt; @@ -358,24 +359,11 @@ static int vmce_save_vcpu_ctxt_one(struct vcpu *v, hvm_domain_context_t *h) ctxt.mci_ctl2_bank1 = v->arch.vmce.bank[1].mci_ctl2; ctxt.mcg_ext_ctl = v->arch.vmce.mcg_ext_ctl; - return hvm_save_entry(VMCE_VCPU, v->vcpu_id, h, &ctxt); + if ( hvm_save_entry(VMCE_VCPU, v->vcpu_id, h, &ctxt) != 0 ) + return ERR; + return OK; } -static int vmce_save_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h) -{ - struct vcpu *v; - int err = 0; - - for_each_vcpu ( d, v ) - { - err = vmce_save_vcpu_ctxt_one(v, h); - if ( err ) - break; - } - - return err; -} - static int vmce_load_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h) { unsigned int vcpuid = hvm_load_instance(h); @@ -396,7 +384,7 @@ static int vmce_load_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h) } HVM_REGISTER_SAVE_RESTORE(VMCE_VCPU, vmce_save_vcpu_ctxt, - vmce_save_vcpu_ctxt_one, + NULL, vmce_load_vcpu_ctxt, 1, HVMSR_PER_VCPU); /* diff --git a/xen/arch/x86/hvm/hpet.c b/xen/arch/x86/hvm/hpet.c index aff8613..0eb302f 100644 --- a/xen/arch/x86/hvm/hpet.c +++ b/xen/arch/x86/hvm/hpet.c @@ -516,8 +516,10 @@ static const struct hvm_mmio_ops hpet_mmio_ops = { }; -static int hpet_save(struct domain *d, hvm_domain_context_t *h) +static enum save_return_type_t hpet_save(struct vcpu *vcpu, + hvm_domain_context_t *h) { + struct domain *d = vcpu->domain; HPETState *hp = domain_vhpet(d); struct vcpu *v = pt_global_vcpu_target(d); int rc; @@ -575,7 +577,9 @@ static int hpet_save(struct domain *d, hvm_domain_context_t *h) write_unlock(&hp->lock); - return rc; + if ( rc != 0 ) + return ERR; + return OK; } static int hpet_load(struct domain *d, hvm_domain_context_t *h) diff --git a/xen/arch/x86/hvm/hvm.c b/xen/arch/x86/hvm/hvm.c index 7d2a12d..cdf91f4 100644 --- a/xen/arch/x86/hvm/hvm.c +++ b/xen/arch/x86/hvm/hvm.c @@ -740,29 +740,18 @@ void hvm_domain_destroy(struct domain *d) destroy_vpci_mmcfg(d); } -static int hvm_save_tsc_adjust_one(struct vcpu *v, hvm_domain_context_t *h) +static enum save_return_type_t hvm_save_tsc_adjust(struct vcpu *v, + hvm_domain_context_t *h) { struct hvm_tsc_adjust ctxt; ctxt.tsc_adjust = v->arch.hvm_vcpu.msr_tsc_adjust; - return hvm_save_entry(TSC_ADJUST, v->vcpu_id, h, &ctxt); + if ( hvm_save_entry(TSC_ADJUST, v->vcpu_id, h, &ctxt) != 0 ) + return ERR; + return OK; } -static int hvm_save_tsc_adjust(struct domain *d, hvm_domain_context_t *h) -{ - struct vcpu *v; - int err = 0; - - for_each_vcpu ( d, v ) - { - err = hvm_save_tsc_adjust_one(v, h); - if ( err ) - break; - } - return err; -} - static int hvm_load_tsc_adjust(struct domain *d, hvm_domain_context_t *h) { unsigned int vcpuid = hvm_load_instance(h); @@ -784,14 +773,22 @@ static int hvm_load_tsc_adjust(struct domain *d, hvm_domain_context_t *h) } HVM_REGISTER_SAVE_RESTORE(TSC_ADJUST, hvm_save_tsc_adjust, - hvm_save_tsc_adjust_one, + NULL, hvm_load_tsc_adjust, 1, HVMSR_PER_VCPU); -static int hvm_save_cpu_ctxt_one(struct vcpu *v, hvm_domain_context_t *h) +static enum save_return_type_t hvm_save_cpu_ctxt(struct vcpu *v, + hvm_domain_context_t *h) { struct segment_register seg; struct hvm_hw_cpu ctxt; + /* + * We don't need to save state for a vcpu that is down; the restore + * code will leave it down if there is nothing saved. + */ + if ( v->pause_flags & VPF_down ) + return CONTINUE; + memset(&ctxt, 0, sizeof(ctxt)); /* Architecture-specific vmcs/vmcb bits */ @@ -888,27 +885,9 @@ static int hvm_save_cpu_ctxt_one(struct vcpu *v, hvm_domain_context_t *h) ctxt.dr6 = v->arch.debugreg[6]; ctxt.dr7 = v->arch.debugreg[7]; - return hvm_save_entry(CPU, v->vcpu_id, h, &ctxt); -} - -static int hvm_save_cpu_ctxt(struct domain *d, hvm_domain_context_t *h) -{ - struct vcpu *v; - - for_each_vcpu ( d, v ) - { - /* - * We don't need to save state for a vcpu that is down; the restore - * code will leave it down if there is nothing saved. - */ - if ( v->pause_flags & VPF_down ) - continue; - - - if ( hvm_save_cpu_ctxt_one(v, h) != 0 ) - return 1; - } - return 0; + if ( hvm_save_entry(CPU, v->vcpu_id, h, &ctxt) != 0 ) + return ERR; + return OK; } /* Return a string indicating the error, or NULL for valid. */ @@ -1182,7 +1161,7 @@ static int hvm_load_cpu_ctxt(struct domain *d, hvm_domain_context_t *h) return 0; } -HVM_REGISTER_SAVE_RESTORE(CPU, hvm_save_cpu_ctxt, hvm_save_cpu_ctxt_one, +HVM_REGISTER_SAVE_RESTORE(CPU, hvm_save_cpu_ctxt, NULL, hvm_load_cpu_ctxt, 1, HVMSR_PER_VCPU); @@ -1190,16 +1169,18 @@ HVM_REGISTER_SAVE_RESTORE(CPU, hvm_save_cpu_ctxt, hvm_save_cpu_ctxt_one, save_area) + \ xstate_ctxt_size(xcr0)) -static int hvm_save_cpu_xsave_states_one(struct vcpu *v, hvm_domain_context_t *h) +static enum save_return_type_t hvm_save_cpu_xsave_states(struct vcpu *v, + hvm_domain_context_t *h) { struct hvm_hw_cpu_xsave *ctxt; unsigned int size = HVM_CPU_XSAVE_SIZE(v->arch.xcr0_accum); if ( !cpu_has_xsave ) - return 0; /* do nothing */ - + return OK; /* do nothing */ + if ( !xsave_enabled(v) ) + return CONTINUE; if ( _hvm_init_entry(h, CPU_XSAVE_CODE, v->vcpu_id, size) ) - return 1; + return ERR; ctxt = (struct hvm_hw_cpu_xsave *)&h->data[h->cur]; h->cur += size; ctxt->xfeature_mask = xfeature_mask; @@ -1208,25 +1189,9 @@ static int hvm_save_cpu_xsave_states_one(struct vcpu *v, hvm_domain_context_t *h expand_xsave_states(v, &ctxt->save_area, size - offsetof(typeof(*ctxt), save_area)); - return 0; + return OK; } -static int hvm_save_cpu_xsave_states(struct domain *d, hvm_domain_context_t *h) -{ - struct vcpu *v; - - for_each_vcpu ( d, v ) - { - if ( !xsave_enabled(v) ) - continue; - - if ( hvm_save_cpu_xsave_states_one(v, h) != 0 ) - return 1; - } - - return 0; -} - /* * Structure layout conformity checks, documenting correctness of the cast in * the invocation of validate_xstate() below. @@ -1364,7 +1329,8 @@ static const uint32_t msrs_to_send[] = { }; static unsigned int __read_mostly msr_count_max = ARRAY_SIZE(msrs_to_send); -static int hvm_save_cpu_msrs_one(struct vcpu *v, hvm_domain_context_t *h) +static enum save_return_type_t hvm_save_cpu_msrs(struct vcpu *v, + hvm_domain_context_t *h) { unsigned int i; struct hvm_msr *ctxt; @@ -1372,7 +1338,7 @@ static int hvm_save_cpu_msrs_one(struct vcpu *v, hvm_domain_context_t *h) if ( _hvm_init_entry(h, CPU_MSR_CODE, v->vcpu_id, HVM_CPU_MSR_SIZE(msr_count_max)) ) - return 1; + return ERR; ctxt = (struct hvm_msr *)&h->data[h->cur]; ctxt->count = 0; @@ -1421,21 +1387,7 @@ static int hvm_save_cpu_msrs_one(struct vcpu *v, hvm_domain_context_t *h) /* or rewind and remove the descriptor from the stream. */ h->cur -= sizeof(struct hvm_save_descriptor); - return 0; -} - - -static int hvm_save_cpu_msrs(struct domain *d, hvm_domain_context_t *h) -{ - struct vcpu *v; - - for_each_vcpu ( d, v ) - { - if ( hvm_save_cpu_msrs_one(v, h) != 0 ) - return 1; - } - - return 0; + return OK; } static int hvm_load_cpu_msrs(struct domain *d, hvm_domain_context_t *h) @@ -1530,7 +1482,7 @@ static int __init hvm_register_CPU_save_and_restore(void) hvm_register_savevm(CPU_XSAVE_CODE, "CPU_XSAVE", hvm_save_cpu_xsave_states, - hvm_save_cpu_xsave_states_one, + NULL, hvm_load_cpu_xsave_states, HVM_CPU_XSAVE_SIZE(xfeature_mask) + sizeof(struct hvm_save_descriptor), @@ -1543,7 +1495,7 @@ static int __init hvm_register_CPU_save_and_restore(void) hvm_register_savevm(CPU_MSR_CODE, "CPU_MSR", hvm_save_cpu_msrs, - hvm_save_cpu_msrs_one, + NULL, hvm_load_cpu_msrs, HVM_CPU_MSR_SIZE(msr_count_max) + sizeof(struct hvm_save_descriptor), diff --git a/xen/arch/x86/hvm/i8254.c b/xen/arch/x86/hvm/i8254.c index ec77b23..00cabac 100644 --- a/xen/arch/x86/hvm/i8254.c +++ b/xen/arch/x86/hvm/i8254.c @@ -390,8 +390,10 @@ void pit_stop_channel0_irq(PITState *pit) spin_unlock(&pit->lock); } -static int pit_save(struct domain *d, hvm_domain_context_t *h) +static enum save_return_type_t pit_save(struct vcpu *v, + hvm_domain_context_t *h) { + struct domain *d = v->domain; PITState *pit = domain_vpit(d); int rc; @@ -403,8 +405,9 @@ static int pit_save(struct domain *d, hvm_domain_context_t *h) rc = hvm_save_entry(PIT, 0, h, &pit->hw); spin_unlock(&pit->lock); - - return rc; + if ( rc != 0 ) + return ERR; + return OK; } static int pit_load(struct domain *d, hvm_domain_context_t *h) diff --git a/xen/arch/x86/hvm/irq.c b/xen/arch/x86/hvm/irq.c index 770eab7..73f48a9 100644 --- a/xen/arch/x86/hvm/irq.c +++ b/xen/arch/x86/hvm/irq.c @@ -630,8 +630,10 @@ static int __init dump_irq_info_key_init(void) } __initcall(dump_irq_info_key_init); -static int irq_save_pci(struct domain *d, hvm_domain_context_t *h) +static enum save_return_type_t irq_save_pci(struct vcpu *v, + hvm_domain_context_t *h) { + struct domain *d = v->domain; struct hvm_irq *hvm_irq = hvm_domain_irq(d); unsigned int asserted, pdev, pintx; int rc; @@ -659,23 +661,33 @@ static int irq_save_pci(struct domain *d, hvm_domain_context_t *h) spin_unlock(&d->arch.hvm_domain.irq_lock); - return rc; + if ( rc != 0 ) + return ERR; + return OK; } -static int irq_save_isa(struct domain *d, hvm_domain_context_t *h) +static enum save_return_type_t irq_save_isa(struct vcpu *v, + hvm_domain_context_t *h) { + struct domain *d = v->domain; struct hvm_irq *hvm_irq = hvm_domain_irq(d); /* Save ISA IRQ lines */ - return ( hvm_save_entry(ISA_IRQ, 0, h, &hvm_irq->isa_irq) ); + if ( hvm_save_entry(ISA_IRQ, 0, h, &hvm_irq->isa_irq) != 0 ) + return ERR; + return OK; } -static int irq_save_link(struct domain *d, hvm_domain_context_t *h) +static enum save_return_type_t irq_save_link(struct vcpu *v, + hvm_domain_context_t *h) { + struct domain *d = v->domain; struct hvm_irq *hvm_irq = hvm_domain_irq(d); /* Save PCI-ISA link state */ - return ( hvm_save_entry(PCI_LINK, 0, h, &hvm_irq->pci_link) ); + if ( hvm_save_entry(PCI_LINK, 0, h, &hvm_irq->pci_link) != 0 ) + return ERR; + return OK; } static int irq_load_pci(struct domain *d, hvm_domain_context_t *h) diff --git a/xen/arch/x86/hvm/mtrr.c b/xen/arch/x86/hvm/mtrr.c index ff9ff69..b4e1335 100644 --- a/xen/arch/x86/hvm/mtrr.c +++ b/xen/arch/x86/hvm/mtrr.c @@ -718,7 +718,8 @@ int hvm_set_mem_pinned_cacheattr(struct domain *d, uint64_t gfn_start, return 0; } -static int hvm_save_mtrr_msr_one(struct vcpu *v, hvm_domain_context_t *h) +static enum save_return_type_t hvm_save_mtrr_msr(struct vcpu *v, + hvm_domain_context_t *h) { const struct mtrr_state *mtrr_state = &v->arch.hvm_vcpu.mtrr; struct hvm_hw_mtrr hw_mtrr; @@ -753,22 +754,11 @@ static int hvm_save_mtrr_msr_one(struct vcpu *v, hvm_domain_context_t *h) hw_mtrr.msr_mtrr_fixed[i] = ((uint64_t*)mtrr_state->fixed_ranges)[i]; - return hvm_save_entry(MTRR, v->vcpu_id, h, &hw_mtrr); + if ( hvm_save_entry(MTRR, v->vcpu_id, h, &hw_mtrr) != 0 ) + return ERR; + return OK; } -static int hvm_save_mtrr_msr(struct domain *d, hvm_domain_context_t *h) -{ - struct vcpu *v; - - /* save mtrr&pat */ - for_each_vcpu(d, v) - { - if ( hvm_save_mtrr_msr_one(v, h) != 0 ) - return 1; - } - return 0; -} - static int hvm_load_mtrr_msr(struct domain *d, hvm_domain_context_t *h) { int vcpuid, i; @@ -819,7 +809,7 @@ static int hvm_load_mtrr_msr(struct domain *d, hvm_domain_context_t *h) return 0; } -HVM_REGISTER_SAVE_RESTORE(MTRR, hvm_save_mtrr_msr, hvm_save_mtrr_msr_one, +HVM_REGISTER_SAVE_RESTORE(MTRR, hvm_save_mtrr_msr, NULL, hvm_load_mtrr_msr, 1, HVMSR_PER_VCPU); void memory_type_changed(struct domain *d) diff --git a/xen/arch/x86/hvm/pmtimer.c b/xen/arch/x86/hvm/pmtimer.c index 0a5e8ce..b96ab5b 100644 --- a/xen/arch/x86/hvm/pmtimer.c +++ b/xen/arch/x86/hvm/pmtimer.c @@ -249,15 +249,17 @@ static int handle_pmt_io( return X86EMUL_OKAY; } -static int acpi_save(struct domain *d, hvm_domain_context_t *h) +static enum save_return_type_t acpi_save(struct vcpu *v, + hvm_domain_context_t *h) { + struct domain *d = v->domain; struct hvm_hw_acpi *acpi = &d->arch.hvm_domain.acpi; PMTState *s = &d->arch.hvm_domain.pl_time->vpmt; uint32_t x, msb = acpi->tmr_val & TMR_VAL_MSB; int rc; if ( !has_vpm(d) ) - return 0; + return OK; spin_lock(&s->lock); @@ -277,7 +279,9 @@ static int acpi_save(struct domain *d, hvm_domain_context_t *h) spin_unlock(&s->lock); - return rc; + if ( rc != 0 ) + return ERR; + return OK; } static int acpi_load(struct domain *d, hvm_domain_context_t *h) diff --git a/xen/arch/x86/hvm/rtc.c b/xen/arch/x86/hvm/rtc.c index ce7e71b..e4a9720 100644 --- a/xen/arch/x86/hvm/rtc.c +++ b/xen/arch/x86/hvm/rtc.c @@ -737,18 +737,21 @@ void rtc_migrate_timers(struct vcpu *v) } /* Save RTC hardware state */ -static int rtc_save(struct domain *d, hvm_domain_context_t *h) +static enum save_return_type_t rtc_save(struct vcpu *v, hvm_domain_context_t *h) { + struct domain *d = v->domain; RTCState *s = domain_vrtc(d); int rc; if ( !has_vrtc(d) ) - return 0; + return OK; spin_lock(&s->lock); rc = hvm_save_entry(RTC, 0, h, &s->hw); spin_unlock(&s->lock); - return rc; + if ( rc != 0 ) + return ERR; + return OK; } /* Reload the hardware state from a saved domain */ diff --git a/xen/arch/x86/hvm/save.c b/xen/arch/x86/hvm/save.c index 1b28e7f..1230f0e 100644 --- a/xen/arch/x86/hvm/save.c +++ b/xen/arch/x86/hvm/save.c @@ -174,7 +174,7 @@ int hvm_save_one(struct domain *d, unsigned int typecode, unsigned int instance, rv = hvm_sr_handlers[typecode].save_one(d->vcpu[instance], &ctxt); else - rv = hvm_sr_handlers[typecode].save(d, &ctxt); + rv = hvm_sr_handlers[typecode].save(d->vcpu[instance], &ctxt); if ( rv != 0 ) { @@ -207,7 +207,8 @@ int hvm_save_one(struct domain *d, unsigned int typecode, unsigned int instance, { for_each_vcpu ( d, v ) { - if ( (rv = hvm_sr_handlers[typecode].save(d, &ctxt)) != 0 ) + if ( (rv = hvm_sr_handlers[typecode].save(d->vcpu[instance], + &ctxt)) != 0 ) { printk(XENLOG_G_ERR "HVM%d save: failed to save type %"PRIu16" (%d)\n", d->domain_id, typecode, rv); @@ -250,7 +251,6 @@ int hvm_save(struct domain *d, hvm_domain_context_t *h) struct hvm_save_header hdr; struct hvm_save_end end; hvm_save_handler handler; - hvm_save_one_handler save_one_handler; unsigned int i, rc; struct vcpu *v = NULL; @@ -280,14 +280,14 @@ int hvm_save(struct domain *d, hvm_domain_context_t *h) for ( i = 0; i <= HVM_SAVE_CODE_MAX; i++ ) { handler = hvm_sr_handlers[i].save; - save_one_handler = hvm_sr_handlers[i].save_one; - if ( save_one_handler != NULL ) + if ( handler != NULL ) { printk(XENLOG_G_INFO "HVM%d save: %s\n", d->domain_id, hvm_sr_handlers[i].name); + for_each_vcpu ( d, v ) { - rc = save_one_handler(v, h); + rc = handler(v, h); if( rc == CONTINUE ) continue; @@ -300,23 +300,6 @@ int hvm_save(struct domain *d, hvm_domain_context_t *h) } } } - else if ( handler != NULL ) - { - printk(XENLOG_G_INFO "HVM%d save: %s\n", - d->domain_id, hvm_sr_handlers[i].name); - - rc = handler(d, h); - if( rc == CONTINUE ) - continue; - - if( rc != 0 ) - { - printk(XENLOG_G_ERR - "HVM%d save: failed to save type %"PRIu16"\n", - d->domain_id, i); - return -EFAULT; - } - } } /* Save an end-of-file marker */ diff --git a/xen/arch/x86/hvm/vioapic.c b/xen/arch/x86/hvm/vioapic.c index 66f54e4..1f59d64 100644 --- a/xen/arch/x86/hvm/vioapic.c +++ b/xen/arch/x86/hvm/vioapic.c @@ -569,12 +569,14 @@ int vioapic_get_trigger_mode(const struct domain *d, unsigned int gsi) return vioapic->redirtbl[pin].fields.trig_mode; } -static int ioapic_save(struct domain *d, hvm_domain_context_t *h) +static enum save_return_type_t ioapic_save(struct vcpu *v, + hvm_domain_context_t *h) { + struct domain *d = v->domain; struct hvm_vioapic *s; if ( !has_vioapic(d) ) - return 0; + return OK; s = domain_vioapic(d, 0); @@ -582,7 +584,9 @@ static int ioapic_save(struct domain *d, hvm_domain_context_t *h) d->arch.hvm_domain.nr_vioapics != 1 ) return -EOPNOTSUPP; - return hvm_save_entry(IOAPIC, 0, h, &s->domU); + if ( hvm_save_entry(IOAPIC, 0, h, &s->domU) != 0 ) + return ERR; + return OK; } static int ioapic_load(struct domain *d, hvm_domain_context_t *h) diff --git a/xen/arch/x86/hvm/viridian.c b/xen/arch/x86/hvm/viridian.c index 466e015..ddae657 100644 --- a/xen/arch/x86/hvm/viridian.c +++ b/xen/arch/x86/hvm/viridian.c @@ -990,8 +990,10 @@ out: return HVM_HCALL_completed; } -static int viridian_save_domain_ctxt(struct domain *d, hvm_domain_context_t *h) +static enum save_return_type_t viridian_save_domain_ctxt(struct vcpu *v, + hvm_domain_context_t *h) { + struct domain *d = v->domain; struct hvm_viridian_domain_context ctxt = { .time_ref_count = d->arch.hvm_domain.viridian.time_ref_count.val, .hypercall_gpa = d->arch.hvm_domain.viridian.hypercall_gpa.raw, @@ -1000,12 +1002,15 @@ static int viridian_save_domain_ctxt(struct domain *d, hvm_domain_context_t *h) }; if ( !is_viridian_domain(d) ) - return 0; + return OK; - return (hvm_save_entry(VIRIDIAN_DOMAIN, 0, h, &ctxt) != 0); + if ( hvm_save_entry(VIRIDIAN_DOMAIN, 0, h, &ctxt) != 0 ) + return ERR; + return OK; } -static int viridian_load_domain_ctxt(struct domain *d, hvm_domain_context_t *h) +static int viridian_load_domain_ctxt(struct domain *d, + hvm_domain_context_t *h) { struct hvm_viridian_domain_context ctxt; @@ -1026,30 +1031,21 @@ static int viridian_load_domain_ctxt(struct domain *d, hvm_domain_context_t *h) HVM_REGISTER_SAVE_RESTORE(VIRIDIAN_DOMAIN, viridian_save_domain_ctxt, NULL, viridian_load_domain_ctxt, 1, HVMSR_PER_DOM); -static int viridian_save_vcpu_ctxt_one(struct vcpu *v, hvm_domain_context_t *h) +static enum save_return_type_t viridian_save_vcpu_ctxt(struct vcpu *v, + hvm_domain_context_t *h) { struct hvm_viridian_vcpu_context ctxt; if ( !is_viridian_domain(v->domain) ) - return 0; + return OK; memset(&ctxt, 0, sizeof(ctxt)); ctxt.vp_assist_msr = v->arch.hvm_vcpu.viridian.vp_assist.msr.raw; ctxt.vp_assist_pending = v->arch.hvm_vcpu.viridian.vp_assist.pending; - return hvm_save_entry(VIRIDIAN_VCPU, v->vcpu_id, h, &ctxt); -} - -static int viridian_save_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h) -{ - struct vcpu *v; - - for_each_vcpu( d, v ) { - if ( viridian_save_vcpu_ctxt_one(v, h) != 0 ) - return 1; - } - - return 0; + if ( hvm_save_entry(VIRIDIAN_VCPU, v->vcpu_id, h, &ctxt) != 0 ) + return ERR; + return OK; } static int viridian_load_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h) @@ -1083,7 +1079,7 @@ static int viridian_load_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h) } HVM_REGISTER_SAVE_RESTORE(VIRIDIAN_VCPU, viridian_save_vcpu_ctxt, - viridian_save_vcpu_ctxt_one, + NULL, viridian_load_vcpu_ctxt, 1, HVMSR_PER_VCPU); static int __init parse_viridian_version(const char *arg) diff --git a/xen/arch/x86/hvm/vlapic.c b/xen/arch/x86/hvm/vlapic.c index eff6070..227bed4 100644 --- a/xen/arch/x86/hvm/vlapic.c +++ b/xen/arch/x86/hvm/vlapic.c @@ -1435,45 +1435,36 @@ static void lapic_rearm(struct vlapic *s) s->timer_last_update = s->pt.last_plt_gtime; } -static int lapic_save_hidden(struct domain *d, hvm_domain_context_t *h) +static enum save_return_type_t lapic_save_hidden(struct vcpu *v, + hvm_domain_context_t *h) { - struct vcpu *v; + struct domain *d = v->domain; struct vlapic *s; - int rc = 0; if ( !has_vlapic(d) ) - return 0; + return OK; - for_each_vcpu ( d, v ) - { - s = vcpu_vlapic(v); - if ( (rc = hvm_save_entry(LAPIC, v->vcpu_id, h, &s->hw)) != 0 ) - break; - } - - return rc; + s = vcpu_vlapic(v); + if ( hvm_save_entry(LAPIC, v->vcpu_id, h, &s->hw) != 0 ) + return ERR; + return OK; } -static int lapic_save_regs(struct domain *d, hvm_domain_context_t *h) +static enum save_return_type_t lapic_save_regs(struct vcpu *v, + hvm_domain_context_t *h) { - struct vcpu *v; + struct domain *d = v->domain; struct vlapic *s; - int rc = 0; if ( !has_vlapic(d) ) - return 0; - - for_each_vcpu ( d, v ) - { - if ( hvm_funcs.sync_pir_to_irr ) - hvm_funcs.sync_pir_to_irr(v); - - s = vcpu_vlapic(v); - if ( (rc = hvm_save_entry(LAPIC_REGS, v->vcpu_id, h, s->regs)) != 0 ) - break; - } + return OK; + if ( hvm_funcs.sync_pir_to_irr ) + hvm_funcs.sync_pir_to_irr(v); - return rc; + s = vcpu_vlapic(v); + if ( hvm_save_entry(LAPIC_REGS, v->vcpu_id, h, s->regs) != 0 ) + return ERR; + return OK; } /* diff --git a/xen/arch/x86/hvm/vpic.c b/xen/arch/x86/hvm/vpic.c index ca9b4cb..a0140d8 100644 --- a/xen/arch/x86/hvm/vpic.c +++ b/xen/arch/x86/hvm/vpic.c @@ -371,23 +371,25 @@ static int vpic_intercept_elcr_io( return X86EMUL_OKAY; } -static int vpic_save(struct domain *d, hvm_domain_context_t *h) +static enum save_return_type_t vpic_save(struct vcpu *v, + hvm_domain_context_t *h) { + struct domain *d = v->domain; struct hvm_hw_vpic *s; int i; if ( !has_vpic(d) ) - return 0; + return OK; /* Save the state of both PICs */ for ( i = 0; i < 2 ; i++ ) { s = &d->arch.hvm_domain.vpic[i]; - if ( hvm_save_entry(PIC, i, h, s) ) - return 1; + if ( hvm_save_entry(PIC, i, h, s) != 0 ) + return ERR; } - return 0; + return OK; } static int vpic_load(struct domain *d, hvm_domain_context_t *h) diff --git a/xen/include/asm-x86/hvm/save.h b/xen/include/asm-x86/hvm/save.h index 2538628..22e5a92 100644 --- a/xen/include/asm-x86/hvm/save.h +++ b/xen/include/asm-x86/hvm/save.h @@ -91,11 +91,17 @@ static inline uint16_t hvm_load_instance(struct hvm_domain_context *h) return d->instance; } +enum save_return_type_t { + OK, + ERR, + CONTINUE, +}; + /* Handler types for different types of save-file entry. * The save handler may save multiple instances of a type into the buffer; * the load handler will be called once for each instance found when * restoring. Both return non-zero on error. */ -typedef int (*hvm_save_handler) (struct domain *d, +typedef enum save_return_type_t (*hvm_save_handler) (struct vcpu *v, hvm_domain_context_t *h); typedef int (*hvm_save_one_handler)(struct vcpu *v, hvm_domain_context_t *h); -- 2.7.4 _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxxx https://lists.xenproject.org/mailman/listinfo/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |