|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH v10 10/11] x86/hvm: Remove redundant save functions
This patch removes the redundant save functions and renames the
save_one* to save. It then changes the domain param to vcpu in the save
funcs.
Signed-off-by: Alexandru Isaila <aisaila@xxxxxxxxxxxxxxx>
---
Changes since V9:
- Add enum return type for save funcs
---
xen/arch/x86/cpu/mcheck/vmce.c | 24 +++------
xen/arch/x86/hvm/hpet.c | 8 ++-
xen/arch/x86/hvm/hvm.c | 112 ++++++++++++-----------------------------
xen/arch/x86/hvm/i8254.c | 9 ++--
xen/arch/x86/hvm/irq.c | 24 ++++++---
xen/arch/x86/hvm/mtrr.c | 22 +++-----
xen/arch/x86/hvm/pmtimer.c | 10 ++--
xen/arch/x86/hvm/rtc.c | 9 ++--
xen/arch/x86/hvm/save.c | 29 +++--------
xen/arch/x86/hvm/vioapic.c | 10 ++--
xen/arch/x86/hvm/viridian.c | 36 ++++++-------
xen/arch/x86/hvm/vlapic.c | 45 +++++++----------
xen/arch/x86/hvm/vpic.c | 12 +++--
xen/include/asm-x86/hvm/save.h | 8 ++-
14 files changed, 148 insertions(+), 210 deletions(-)
diff --git a/xen/arch/x86/cpu/mcheck/vmce.c b/xen/arch/x86/cpu/mcheck/vmce.c
index 29898a6..c054ae9 100644
--- a/xen/arch/x86/cpu/mcheck/vmce.c
+++ b/xen/arch/x86/cpu/mcheck/vmce.c
@@ -349,7 +349,8 @@ int vmce_wrmsr(uint32_t msr, uint64_t val)
return ret;
}
-static int vmce_save_vcpu_ctxt_one(struct vcpu *v, hvm_domain_context_t *h)
+static enum save_return_type_t vmce_save_vcpu_ctxt(struct vcpu *v,
+ hvm_domain_context_t *h)
{
struct hvm_vmce_vcpu ctxt;
@@ -358,24 +359,11 @@ static int vmce_save_vcpu_ctxt_one(struct vcpu *v,
hvm_domain_context_t *h)
ctxt.mci_ctl2_bank1 = v->arch.vmce.bank[1].mci_ctl2;
ctxt.mcg_ext_ctl = v->arch.vmce.mcg_ext_ctl;
- return hvm_save_entry(VMCE_VCPU, v->vcpu_id, h, &ctxt);
+ if ( hvm_save_entry(VMCE_VCPU, v->vcpu_id, h, &ctxt) != 0 )
+ return ERR;
+ return OK;
}
-static int vmce_save_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h)
-{
- struct vcpu *v;
- int err = 0;
-
- for_each_vcpu ( d, v )
- {
- err = vmce_save_vcpu_ctxt_one(v, h);
- if ( err )
- break;
- }
-
- return err;
-}
-
static int vmce_load_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h)
{
unsigned int vcpuid = hvm_load_instance(h);
@@ -396,7 +384,7 @@ static int vmce_load_vcpu_ctxt(struct domain *d,
hvm_domain_context_t *h)
}
HVM_REGISTER_SAVE_RESTORE(VMCE_VCPU, vmce_save_vcpu_ctxt,
- vmce_save_vcpu_ctxt_one,
+ NULL,
vmce_load_vcpu_ctxt, 1, HVMSR_PER_VCPU);
/*
diff --git a/xen/arch/x86/hvm/hpet.c b/xen/arch/x86/hvm/hpet.c
index aff8613..0eb302f 100644
--- a/xen/arch/x86/hvm/hpet.c
+++ b/xen/arch/x86/hvm/hpet.c
@@ -516,8 +516,10 @@ static const struct hvm_mmio_ops hpet_mmio_ops = {
};
-static int hpet_save(struct domain *d, hvm_domain_context_t *h)
+static enum save_return_type_t hpet_save(struct vcpu *vcpu,
+ hvm_domain_context_t *h)
{
+ struct domain *d = vcpu->domain;
HPETState *hp = domain_vhpet(d);
struct vcpu *v = pt_global_vcpu_target(d);
int rc;
@@ -575,7 +577,9 @@ static int hpet_save(struct domain *d, hvm_domain_context_t
*h)
write_unlock(&hp->lock);
- return rc;
+ if ( rc != 0 )
+ return ERR;
+ return OK;
}
static int hpet_load(struct domain *d, hvm_domain_context_t *h)
diff --git a/xen/arch/x86/hvm/hvm.c b/xen/arch/x86/hvm/hvm.c
index 7d2a12d..cdf91f4 100644
--- a/xen/arch/x86/hvm/hvm.c
+++ b/xen/arch/x86/hvm/hvm.c
@@ -740,29 +740,18 @@ void hvm_domain_destroy(struct domain *d)
destroy_vpci_mmcfg(d);
}
-static int hvm_save_tsc_adjust_one(struct vcpu *v, hvm_domain_context_t *h)
+static enum save_return_type_t hvm_save_tsc_adjust(struct vcpu *v,
+ hvm_domain_context_t *h)
{
struct hvm_tsc_adjust ctxt;
ctxt.tsc_adjust = v->arch.hvm_vcpu.msr_tsc_adjust;
- return hvm_save_entry(TSC_ADJUST, v->vcpu_id, h, &ctxt);
+ if ( hvm_save_entry(TSC_ADJUST, v->vcpu_id, h, &ctxt) != 0 )
+ return ERR;
+ return OK;
}
-static int hvm_save_tsc_adjust(struct domain *d, hvm_domain_context_t *h)
-{
- struct vcpu *v;
- int err = 0;
-
- for_each_vcpu ( d, v )
- {
- err = hvm_save_tsc_adjust_one(v, h);
- if ( err )
- break;
- }
- return err;
-}
-
static int hvm_load_tsc_adjust(struct domain *d, hvm_domain_context_t *h)
{
unsigned int vcpuid = hvm_load_instance(h);
@@ -784,14 +773,22 @@ static int hvm_load_tsc_adjust(struct domain *d,
hvm_domain_context_t *h)
}
HVM_REGISTER_SAVE_RESTORE(TSC_ADJUST, hvm_save_tsc_adjust,
- hvm_save_tsc_adjust_one,
+ NULL,
hvm_load_tsc_adjust, 1, HVMSR_PER_VCPU);
-static int hvm_save_cpu_ctxt_one(struct vcpu *v, hvm_domain_context_t *h)
+static enum save_return_type_t hvm_save_cpu_ctxt(struct vcpu *v,
+ hvm_domain_context_t *h)
{
struct segment_register seg;
struct hvm_hw_cpu ctxt;
+ /*
+ * We don't need to save state for a vcpu that is down; the restore
+ * code will leave it down if there is nothing saved.
+ */
+ if ( v->pause_flags & VPF_down )
+ return CONTINUE;
+
memset(&ctxt, 0, sizeof(ctxt));
/* Architecture-specific vmcs/vmcb bits */
@@ -888,27 +885,9 @@ static int hvm_save_cpu_ctxt_one(struct vcpu *v,
hvm_domain_context_t *h)
ctxt.dr6 = v->arch.debugreg[6];
ctxt.dr7 = v->arch.debugreg[7];
- return hvm_save_entry(CPU, v->vcpu_id, h, &ctxt);
-}
-
-static int hvm_save_cpu_ctxt(struct domain *d, hvm_domain_context_t *h)
-{
- struct vcpu *v;
-
- for_each_vcpu ( d, v )
- {
- /*
- * We don't need to save state for a vcpu that is down; the restore
- * code will leave it down if there is nothing saved.
- */
- if ( v->pause_flags & VPF_down )
- continue;
-
-
- if ( hvm_save_cpu_ctxt_one(v, h) != 0 )
- return 1;
- }
- return 0;
+ if ( hvm_save_entry(CPU, v->vcpu_id, h, &ctxt) != 0 )
+ return ERR;
+ return OK;
}
/* Return a string indicating the error, or NULL for valid. */
@@ -1182,7 +1161,7 @@ static int hvm_load_cpu_ctxt(struct domain *d,
hvm_domain_context_t *h)
return 0;
}
-HVM_REGISTER_SAVE_RESTORE(CPU, hvm_save_cpu_ctxt, hvm_save_cpu_ctxt_one,
+HVM_REGISTER_SAVE_RESTORE(CPU, hvm_save_cpu_ctxt, NULL,
hvm_load_cpu_ctxt,
1, HVMSR_PER_VCPU);
@@ -1190,16 +1169,18 @@ HVM_REGISTER_SAVE_RESTORE(CPU, hvm_save_cpu_ctxt,
hvm_save_cpu_ctxt_one,
save_area) + \
xstate_ctxt_size(xcr0))
-static int hvm_save_cpu_xsave_states_one(struct vcpu *v, hvm_domain_context_t
*h)
+static enum save_return_type_t hvm_save_cpu_xsave_states(struct vcpu *v,
+ hvm_domain_context_t
*h)
{
struct hvm_hw_cpu_xsave *ctxt;
unsigned int size = HVM_CPU_XSAVE_SIZE(v->arch.xcr0_accum);
if ( !cpu_has_xsave )
- return 0; /* do nothing */
-
+ return OK; /* do nothing */
+ if ( !xsave_enabled(v) )
+ return CONTINUE;
if ( _hvm_init_entry(h, CPU_XSAVE_CODE, v->vcpu_id, size) )
- return 1;
+ return ERR;
ctxt = (struct hvm_hw_cpu_xsave *)&h->data[h->cur];
h->cur += size;
ctxt->xfeature_mask = xfeature_mask;
@@ -1208,25 +1189,9 @@ static int hvm_save_cpu_xsave_states_one(struct vcpu *v,
hvm_domain_context_t *h
expand_xsave_states(v, &ctxt->save_area,
size - offsetof(typeof(*ctxt), save_area));
- return 0;
+ return OK;
}
-static int hvm_save_cpu_xsave_states(struct domain *d, hvm_domain_context_t *h)
-{
- struct vcpu *v;
-
- for_each_vcpu ( d, v )
- {
- if ( !xsave_enabled(v) )
- continue;
-
- if ( hvm_save_cpu_xsave_states_one(v, h) != 0 )
- return 1;
- }
-
- return 0;
-}
-
/*
* Structure layout conformity checks, documenting correctness of the cast in
* the invocation of validate_xstate() below.
@@ -1364,7 +1329,8 @@ static const uint32_t msrs_to_send[] = {
};
static unsigned int __read_mostly msr_count_max = ARRAY_SIZE(msrs_to_send);
-static int hvm_save_cpu_msrs_one(struct vcpu *v, hvm_domain_context_t *h)
+static enum save_return_type_t hvm_save_cpu_msrs(struct vcpu *v,
+ hvm_domain_context_t *h)
{
unsigned int i;
struct hvm_msr *ctxt;
@@ -1372,7 +1338,7 @@ static int hvm_save_cpu_msrs_one(struct vcpu *v,
hvm_domain_context_t *h)
if ( _hvm_init_entry(h, CPU_MSR_CODE, v->vcpu_id,
HVM_CPU_MSR_SIZE(msr_count_max)) )
- return 1;
+ return ERR;
ctxt = (struct hvm_msr *)&h->data[h->cur];
ctxt->count = 0;
@@ -1421,21 +1387,7 @@ static int hvm_save_cpu_msrs_one(struct vcpu *v,
hvm_domain_context_t *h)
/* or rewind and remove the descriptor from the stream. */
h->cur -= sizeof(struct hvm_save_descriptor);
- return 0;
-}
-
-
-static int hvm_save_cpu_msrs(struct domain *d, hvm_domain_context_t *h)
-{
- struct vcpu *v;
-
- for_each_vcpu ( d, v )
- {
- if ( hvm_save_cpu_msrs_one(v, h) != 0 )
- return 1;
- }
-
- return 0;
+ return OK;
}
static int hvm_load_cpu_msrs(struct domain *d, hvm_domain_context_t *h)
@@ -1530,7 +1482,7 @@ static int __init hvm_register_CPU_save_and_restore(void)
hvm_register_savevm(CPU_XSAVE_CODE,
"CPU_XSAVE",
hvm_save_cpu_xsave_states,
- hvm_save_cpu_xsave_states_one,
+ NULL,
hvm_load_cpu_xsave_states,
HVM_CPU_XSAVE_SIZE(xfeature_mask) +
sizeof(struct hvm_save_descriptor),
@@ -1543,7 +1495,7 @@ static int __init hvm_register_CPU_save_and_restore(void)
hvm_register_savevm(CPU_MSR_CODE,
"CPU_MSR",
hvm_save_cpu_msrs,
- hvm_save_cpu_msrs_one,
+ NULL,
hvm_load_cpu_msrs,
HVM_CPU_MSR_SIZE(msr_count_max) +
sizeof(struct hvm_save_descriptor),
diff --git a/xen/arch/x86/hvm/i8254.c b/xen/arch/x86/hvm/i8254.c
index ec77b23..00cabac 100644
--- a/xen/arch/x86/hvm/i8254.c
+++ b/xen/arch/x86/hvm/i8254.c
@@ -390,8 +390,10 @@ void pit_stop_channel0_irq(PITState *pit)
spin_unlock(&pit->lock);
}
-static int pit_save(struct domain *d, hvm_domain_context_t *h)
+static enum save_return_type_t pit_save(struct vcpu *v,
+ hvm_domain_context_t *h)
{
+ struct domain *d = v->domain;
PITState *pit = domain_vpit(d);
int rc;
@@ -403,8 +405,9 @@ static int pit_save(struct domain *d, hvm_domain_context_t
*h)
rc = hvm_save_entry(PIT, 0, h, &pit->hw);
spin_unlock(&pit->lock);
-
- return rc;
+ if ( rc != 0 )
+ return ERR;
+ return OK;
}
static int pit_load(struct domain *d, hvm_domain_context_t *h)
diff --git a/xen/arch/x86/hvm/irq.c b/xen/arch/x86/hvm/irq.c
index 770eab7..73f48a9 100644
--- a/xen/arch/x86/hvm/irq.c
+++ b/xen/arch/x86/hvm/irq.c
@@ -630,8 +630,10 @@ static int __init dump_irq_info_key_init(void)
}
__initcall(dump_irq_info_key_init);
-static int irq_save_pci(struct domain *d, hvm_domain_context_t *h)
+static enum save_return_type_t irq_save_pci(struct vcpu *v,
+ hvm_domain_context_t *h)
{
+ struct domain *d = v->domain;
struct hvm_irq *hvm_irq = hvm_domain_irq(d);
unsigned int asserted, pdev, pintx;
int rc;
@@ -659,23 +661,33 @@ static int irq_save_pci(struct domain *d,
hvm_domain_context_t *h)
spin_unlock(&d->arch.hvm_domain.irq_lock);
- return rc;
+ if ( rc != 0 )
+ return ERR;
+ return OK;
}
-static int irq_save_isa(struct domain *d, hvm_domain_context_t *h)
+static enum save_return_type_t irq_save_isa(struct vcpu *v,
+ hvm_domain_context_t *h)
{
+ struct domain *d = v->domain;
struct hvm_irq *hvm_irq = hvm_domain_irq(d);
/* Save ISA IRQ lines */
- return ( hvm_save_entry(ISA_IRQ, 0, h, &hvm_irq->isa_irq) );
+ if ( hvm_save_entry(ISA_IRQ, 0, h, &hvm_irq->isa_irq) != 0 )
+ return ERR;
+ return OK;
}
-static int irq_save_link(struct domain *d, hvm_domain_context_t *h)
+static enum save_return_type_t irq_save_link(struct vcpu *v,
+ hvm_domain_context_t *h)
{
+ struct domain *d = v->domain;
struct hvm_irq *hvm_irq = hvm_domain_irq(d);
/* Save PCI-ISA link state */
- return ( hvm_save_entry(PCI_LINK, 0, h, &hvm_irq->pci_link) );
+ if ( hvm_save_entry(PCI_LINK, 0, h, &hvm_irq->pci_link) != 0 )
+ return ERR;
+ return OK;
}
static int irq_load_pci(struct domain *d, hvm_domain_context_t *h)
diff --git a/xen/arch/x86/hvm/mtrr.c b/xen/arch/x86/hvm/mtrr.c
index ff9ff69..b4e1335 100644
--- a/xen/arch/x86/hvm/mtrr.c
+++ b/xen/arch/x86/hvm/mtrr.c
@@ -718,7 +718,8 @@ int hvm_set_mem_pinned_cacheattr(struct domain *d, uint64_t
gfn_start,
return 0;
}
-static int hvm_save_mtrr_msr_one(struct vcpu *v, hvm_domain_context_t *h)
+static enum save_return_type_t hvm_save_mtrr_msr(struct vcpu *v,
+ hvm_domain_context_t *h)
{
const struct mtrr_state *mtrr_state = &v->arch.hvm_vcpu.mtrr;
struct hvm_hw_mtrr hw_mtrr;
@@ -753,22 +754,11 @@ static int hvm_save_mtrr_msr_one(struct vcpu *v,
hvm_domain_context_t *h)
hw_mtrr.msr_mtrr_fixed[i] =
((uint64_t*)mtrr_state->fixed_ranges)[i];
- return hvm_save_entry(MTRR, v->vcpu_id, h, &hw_mtrr);
+ if ( hvm_save_entry(MTRR, v->vcpu_id, h, &hw_mtrr) != 0 )
+ return ERR;
+ return OK;
}
-static int hvm_save_mtrr_msr(struct domain *d, hvm_domain_context_t *h)
-{
- struct vcpu *v;
-
- /* save mtrr&pat */
- for_each_vcpu(d, v)
- {
- if ( hvm_save_mtrr_msr_one(v, h) != 0 )
- return 1;
- }
- return 0;
-}
-
static int hvm_load_mtrr_msr(struct domain *d, hvm_domain_context_t *h)
{
int vcpuid, i;
@@ -819,7 +809,7 @@ static int hvm_load_mtrr_msr(struct domain *d,
hvm_domain_context_t *h)
return 0;
}
-HVM_REGISTER_SAVE_RESTORE(MTRR, hvm_save_mtrr_msr, hvm_save_mtrr_msr_one,
+HVM_REGISTER_SAVE_RESTORE(MTRR, hvm_save_mtrr_msr, NULL,
hvm_load_mtrr_msr, 1, HVMSR_PER_VCPU);
void memory_type_changed(struct domain *d)
diff --git a/xen/arch/x86/hvm/pmtimer.c b/xen/arch/x86/hvm/pmtimer.c
index 0a5e8ce..b96ab5b 100644
--- a/xen/arch/x86/hvm/pmtimer.c
+++ b/xen/arch/x86/hvm/pmtimer.c
@@ -249,15 +249,17 @@ static int handle_pmt_io(
return X86EMUL_OKAY;
}
-static int acpi_save(struct domain *d, hvm_domain_context_t *h)
+static enum save_return_type_t acpi_save(struct vcpu *v,
+ hvm_domain_context_t *h)
{
+ struct domain *d = v->domain;
struct hvm_hw_acpi *acpi = &d->arch.hvm_domain.acpi;
PMTState *s = &d->arch.hvm_domain.pl_time->vpmt;
uint32_t x, msb = acpi->tmr_val & TMR_VAL_MSB;
int rc;
if ( !has_vpm(d) )
- return 0;
+ return OK;
spin_lock(&s->lock);
@@ -277,7 +279,9 @@ static int acpi_save(struct domain *d, hvm_domain_context_t
*h)
spin_unlock(&s->lock);
- return rc;
+ if ( rc != 0 )
+ return ERR;
+ return OK;
}
static int acpi_load(struct domain *d, hvm_domain_context_t *h)
diff --git a/xen/arch/x86/hvm/rtc.c b/xen/arch/x86/hvm/rtc.c
index ce7e71b..e4a9720 100644
--- a/xen/arch/x86/hvm/rtc.c
+++ b/xen/arch/x86/hvm/rtc.c
@@ -737,18 +737,21 @@ void rtc_migrate_timers(struct vcpu *v)
}
/* Save RTC hardware state */
-static int rtc_save(struct domain *d, hvm_domain_context_t *h)
+static enum save_return_type_t rtc_save(struct vcpu *v, hvm_domain_context_t
*h)
{
+ struct domain *d = v->domain;
RTCState *s = domain_vrtc(d);
int rc;
if ( !has_vrtc(d) )
- return 0;
+ return OK;
spin_lock(&s->lock);
rc = hvm_save_entry(RTC, 0, h, &s->hw);
spin_unlock(&s->lock);
- return rc;
+ if ( rc != 0 )
+ return ERR;
+ return OK;
}
/* Reload the hardware state from a saved domain */
diff --git a/xen/arch/x86/hvm/save.c b/xen/arch/x86/hvm/save.c
index 1b28e7f..1230f0e 100644
--- a/xen/arch/x86/hvm/save.c
+++ b/xen/arch/x86/hvm/save.c
@@ -174,7 +174,7 @@ int hvm_save_one(struct domain *d, unsigned int typecode,
unsigned int instance,
rv = hvm_sr_handlers[typecode].save_one(d->vcpu[instance],
&ctxt);
else
- rv = hvm_sr_handlers[typecode].save(d, &ctxt);
+ rv = hvm_sr_handlers[typecode].save(d->vcpu[instance], &ctxt);
if ( rv != 0 )
{
@@ -207,7 +207,8 @@ int hvm_save_one(struct domain *d, unsigned int typecode,
unsigned int instance,
{
for_each_vcpu ( d, v )
{
- if ( (rv = hvm_sr_handlers[typecode].save(d, &ctxt)) != 0 )
+ if ( (rv = hvm_sr_handlers[typecode].save(d->vcpu[instance],
+ &ctxt)) != 0 )
{
printk(XENLOG_G_ERR "HVM%d save: failed to save type %"PRIu16"
(%d)\n",
d->domain_id, typecode, rv);
@@ -250,7 +251,6 @@ int hvm_save(struct domain *d, hvm_domain_context_t *h)
struct hvm_save_header hdr;
struct hvm_save_end end;
hvm_save_handler handler;
- hvm_save_one_handler save_one_handler;
unsigned int i, rc;
struct vcpu *v = NULL;
@@ -280,14 +280,14 @@ int hvm_save(struct domain *d, hvm_domain_context_t *h)
for ( i = 0; i <= HVM_SAVE_CODE_MAX; i++ )
{
handler = hvm_sr_handlers[i].save;
- save_one_handler = hvm_sr_handlers[i].save_one;
- if ( save_one_handler != NULL )
+ if ( handler != NULL )
{
printk(XENLOG_G_INFO "HVM%d save: %s\n",
d->domain_id, hvm_sr_handlers[i].name);
+
for_each_vcpu ( d, v )
{
- rc = save_one_handler(v, h);
+ rc = handler(v, h);
if( rc == CONTINUE )
continue;
@@ -300,23 +300,6 @@ int hvm_save(struct domain *d, hvm_domain_context_t *h)
}
}
}
- else if ( handler != NULL )
- {
- printk(XENLOG_G_INFO "HVM%d save: %s\n",
- d->domain_id, hvm_sr_handlers[i].name);
-
- rc = handler(d, h);
- if( rc == CONTINUE )
- continue;
-
- if( rc != 0 )
- {
- printk(XENLOG_G_ERR
- "HVM%d save: failed to save type %"PRIu16"\n",
- d->domain_id, i);
- return -EFAULT;
- }
- }
}
/* Save an end-of-file marker */
diff --git a/xen/arch/x86/hvm/vioapic.c b/xen/arch/x86/hvm/vioapic.c
index 66f54e4..1f59d64 100644
--- a/xen/arch/x86/hvm/vioapic.c
+++ b/xen/arch/x86/hvm/vioapic.c
@@ -569,12 +569,14 @@ int vioapic_get_trigger_mode(const struct domain *d,
unsigned int gsi)
return vioapic->redirtbl[pin].fields.trig_mode;
}
-static int ioapic_save(struct domain *d, hvm_domain_context_t *h)
+static enum save_return_type_t ioapic_save(struct vcpu *v,
+ hvm_domain_context_t *h)
{
+ struct domain *d = v->domain;
struct hvm_vioapic *s;
if ( !has_vioapic(d) )
- return 0;
+ return OK;
s = domain_vioapic(d, 0);
@@ -582,7 +584,9 @@ static int ioapic_save(struct domain *d,
hvm_domain_context_t *h)
d->arch.hvm_domain.nr_vioapics != 1 )
return -EOPNOTSUPP;
- return hvm_save_entry(IOAPIC, 0, h, &s->domU);
+ if ( hvm_save_entry(IOAPIC, 0, h, &s->domU) != 0 )
+ return ERR;
+ return OK;
}
static int ioapic_load(struct domain *d, hvm_domain_context_t *h)
diff --git a/xen/arch/x86/hvm/viridian.c b/xen/arch/x86/hvm/viridian.c
index 466e015..ddae657 100644
--- a/xen/arch/x86/hvm/viridian.c
+++ b/xen/arch/x86/hvm/viridian.c
@@ -990,8 +990,10 @@ out:
return HVM_HCALL_completed;
}
-static int viridian_save_domain_ctxt(struct domain *d, hvm_domain_context_t *h)
+static enum save_return_type_t viridian_save_domain_ctxt(struct vcpu *v,
+ hvm_domain_context_t
*h)
{
+ struct domain *d = v->domain;
struct hvm_viridian_domain_context ctxt = {
.time_ref_count = d->arch.hvm_domain.viridian.time_ref_count.val,
.hypercall_gpa = d->arch.hvm_domain.viridian.hypercall_gpa.raw,
@@ -1000,12 +1002,15 @@ static int viridian_save_domain_ctxt(struct domain *d,
hvm_domain_context_t *h)
};
if ( !is_viridian_domain(d) )
- return 0;
+ return OK;
- return (hvm_save_entry(VIRIDIAN_DOMAIN, 0, h, &ctxt) != 0);
+ if ( hvm_save_entry(VIRIDIAN_DOMAIN, 0, h, &ctxt) != 0 )
+ return ERR;
+ return OK;
}
-static int viridian_load_domain_ctxt(struct domain *d, hvm_domain_context_t *h)
+static int viridian_load_domain_ctxt(struct domain *d,
+ hvm_domain_context_t *h)
{
struct hvm_viridian_domain_context ctxt;
@@ -1026,30 +1031,21 @@ static int viridian_load_domain_ctxt(struct domain *d,
hvm_domain_context_t *h)
HVM_REGISTER_SAVE_RESTORE(VIRIDIAN_DOMAIN, viridian_save_domain_ctxt, NULL,
viridian_load_domain_ctxt, 1, HVMSR_PER_DOM);
-static int viridian_save_vcpu_ctxt_one(struct vcpu *v, hvm_domain_context_t *h)
+static enum save_return_type_t viridian_save_vcpu_ctxt(struct vcpu *v,
+ hvm_domain_context_t *h)
{
struct hvm_viridian_vcpu_context ctxt;
if ( !is_viridian_domain(v->domain) )
- return 0;
+ return OK;
memset(&ctxt, 0, sizeof(ctxt));
ctxt.vp_assist_msr = v->arch.hvm_vcpu.viridian.vp_assist.msr.raw;
ctxt.vp_assist_pending = v->arch.hvm_vcpu.viridian.vp_assist.pending;
- return hvm_save_entry(VIRIDIAN_VCPU, v->vcpu_id, h, &ctxt);
-}
-
-static int viridian_save_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h)
-{
- struct vcpu *v;
-
- for_each_vcpu( d, v ) {
- if ( viridian_save_vcpu_ctxt_one(v, h) != 0 )
- return 1;
- }
-
- return 0;
+ if ( hvm_save_entry(VIRIDIAN_VCPU, v->vcpu_id, h, &ctxt) != 0 )
+ return ERR;
+ return OK;
}
static int viridian_load_vcpu_ctxt(struct domain *d, hvm_domain_context_t *h)
@@ -1083,7 +1079,7 @@ static int viridian_load_vcpu_ctxt(struct domain *d,
hvm_domain_context_t *h)
}
HVM_REGISTER_SAVE_RESTORE(VIRIDIAN_VCPU, viridian_save_vcpu_ctxt,
- viridian_save_vcpu_ctxt_one,
+ NULL,
viridian_load_vcpu_ctxt, 1, HVMSR_PER_VCPU);
static int __init parse_viridian_version(const char *arg)
diff --git a/xen/arch/x86/hvm/vlapic.c b/xen/arch/x86/hvm/vlapic.c
index eff6070..227bed4 100644
--- a/xen/arch/x86/hvm/vlapic.c
+++ b/xen/arch/x86/hvm/vlapic.c
@@ -1435,45 +1435,36 @@ static void lapic_rearm(struct vlapic *s)
s->timer_last_update = s->pt.last_plt_gtime;
}
-static int lapic_save_hidden(struct domain *d, hvm_domain_context_t *h)
+static enum save_return_type_t lapic_save_hidden(struct vcpu *v,
+ hvm_domain_context_t *h)
{
- struct vcpu *v;
+ struct domain *d = v->domain;
struct vlapic *s;
- int rc = 0;
if ( !has_vlapic(d) )
- return 0;
+ return OK;
- for_each_vcpu ( d, v )
- {
- s = vcpu_vlapic(v);
- if ( (rc = hvm_save_entry(LAPIC, v->vcpu_id, h, &s->hw)) != 0 )
- break;
- }
-
- return rc;
+ s = vcpu_vlapic(v);
+ if ( hvm_save_entry(LAPIC, v->vcpu_id, h, &s->hw) != 0 )
+ return ERR;
+ return OK;
}
-static int lapic_save_regs(struct domain *d, hvm_domain_context_t *h)
+static enum save_return_type_t lapic_save_regs(struct vcpu *v,
+ hvm_domain_context_t *h)
{
- struct vcpu *v;
+ struct domain *d = v->domain;
struct vlapic *s;
- int rc = 0;
if ( !has_vlapic(d) )
- return 0;
-
- for_each_vcpu ( d, v )
- {
- if ( hvm_funcs.sync_pir_to_irr )
- hvm_funcs.sync_pir_to_irr(v);
-
- s = vcpu_vlapic(v);
- if ( (rc = hvm_save_entry(LAPIC_REGS, v->vcpu_id, h, s->regs)) != 0 )
- break;
- }
+ return OK;
+ if ( hvm_funcs.sync_pir_to_irr )
+ hvm_funcs.sync_pir_to_irr(v);
- return rc;
+ s = vcpu_vlapic(v);
+ if ( hvm_save_entry(LAPIC_REGS, v->vcpu_id, h, s->regs) != 0 )
+ return ERR;
+ return OK;
}
/*
diff --git a/xen/arch/x86/hvm/vpic.c b/xen/arch/x86/hvm/vpic.c
index ca9b4cb..a0140d8 100644
--- a/xen/arch/x86/hvm/vpic.c
+++ b/xen/arch/x86/hvm/vpic.c
@@ -371,23 +371,25 @@ static int vpic_intercept_elcr_io(
return X86EMUL_OKAY;
}
-static int vpic_save(struct domain *d, hvm_domain_context_t *h)
+static enum save_return_type_t vpic_save(struct vcpu *v,
+ hvm_domain_context_t *h)
{
+ struct domain *d = v->domain;
struct hvm_hw_vpic *s;
int i;
if ( !has_vpic(d) )
- return 0;
+ return OK;
/* Save the state of both PICs */
for ( i = 0; i < 2 ; i++ )
{
s = &d->arch.hvm_domain.vpic[i];
- if ( hvm_save_entry(PIC, i, h, s) )
- return 1;
+ if ( hvm_save_entry(PIC, i, h, s) != 0 )
+ return ERR;
}
- return 0;
+ return OK;
}
static int vpic_load(struct domain *d, hvm_domain_context_t *h)
diff --git a/xen/include/asm-x86/hvm/save.h b/xen/include/asm-x86/hvm/save.h
index 2538628..22e5a92 100644
--- a/xen/include/asm-x86/hvm/save.h
+++ b/xen/include/asm-x86/hvm/save.h
@@ -91,11 +91,17 @@ static inline uint16_t hvm_load_instance(struct
hvm_domain_context *h)
return d->instance;
}
+enum save_return_type_t {
+ OK,
+ ERR,
+ CONTINUE,
+};
+
/* Handler types for different types of save-file entry.
* The save handler may save multiple instances of a type into the buffer;
* the load handler will be called once for each instance found when
* restoring. Both return non-zero on error. */
-typedef int (*hvm_save_handler) (struct domain *d,
+typedef enum save_return_type_t (*hvm_save_handler) (struct vcpu *v,
hvm_domain_context_t *h);
typedef int (*hvm_save_one_handler)(struct vcpu *v,
hvm_domain_context_t *h);
--
2.7.4
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxxx
https://lists.xenproject.org/mailman/listinfo/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |