|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH RFC v1 58/74] xen/pvshim: add migration support
From: Roger Pau Monne <roger.pau@xxxxxxxxxx>
Signed-off-by: Roger Pau Monné <roger.pau@xxxxxxxxxx>
---
xen/arch/x86/guest/xen.c | 29 +++++++
xen/arch/x86/mm.c | 3 +-
xen/arch/x86/pv/shim.c | 163 +++++++++++++++++++++++++++++++++++++-
xen/common/domain.c | 11 ++-
xen/common/schedule.c | 3 +-
xen/drivers/char/xen_pv_console.c | 2 +-
xen/include/asm-x86/guest/xen.h | 5 ++
xen/include/asm-x86/pv/shim.h | 5 +-
xen/include/xen/sched.h | 2 +-
9 files changed, 206 insertions(+), 17 deletions(-)
diff --git a/xen/arch/x86/guest/xen.c b/xen/arch/x86/guest/xen.c
index 653a7366ab..54c997b9e0 100644
--- a/xen/arch/x86/guest/xen.c
+++ b/xen/arch/x86/guest/xen.c
@@ -361,6 +361,35 @@ uint32_t hypervisor_cpuid_base(void)
return xen_cpuid_base;
}
+void ap_resume(void *unused)
+{
+ map_vcpuinfo();
+ init_evtchn();
+}
+
+void hypervisor_resume(void)
+{
+ /* Reset shared info page. */
+ map_shared_info();
+
+ /*
+ * Reset vcpu_info. Just clean the mapped bitmap and try to map the vcpu
+ * area again. On failure to map (when it was previously mapped) panic
+ * since it's impossible to safely shut down running guest vCPUs in order
+ * to meet the new XEN_LEGACY_MAX_VCPUS requirement.
+ */
+ memset(vcpu_info_mapped, 0, sizeof(vcpu_info_mapped));
+ if ( map_vcpuinfo() && nr_cpu_ids > XEN_LEGACY_MAX_VCPUS )
+ panic("unable to remap vCPU info and vCPUs > legacy limit");
+
+ /* Setup event channel upcall vector. */
+ init_evtchn();
+ smp_call_function(ap_resume, NULL, 1);
+
+ if ( pv_console )
+ pv_console_init();
+}
+
/*
* Local variables:
* mode: C
diff --git a/xen/arch/x86/mm.c b/xen/arch/x86/mm.c
index 4332d3bb39..e726c62064 100644
--- a/xen/arch/x86/mm.c
+++ b/xen/arch/x86/mm.c
@@ -466,8 +466,7 @@ void share_xen_page_with_guest(
spin_unlock(&d->page_alloc_lock);
}
-int __init unshare_xen_page_with_guest(struct page_info *page,
- struct domain *d)
+int unshare_xen_page_with_guest(struct page_info *page, struct domain *d)
{
if ( page_get_owner(page) != d || !is_xen_heap_page(page) )
return -EINVAL;
diff --git a/xen/arch/x86/pv/shim.c b/xen/arch/x86/pv/shim.c
index 74b24bf950..56ecaea2d2 100644
--- a/xen/arch/x86/pv/shim.c
+++ b/xen/arch/x86/pv/shim.c
@@ -151,10 +151,167 @@ void __init pv_shim_setup_dom(struct domain *d,
l4_pgentry_t *l4start,
}
}
-void pv_shim_shutdown(uint8_t reason)
+static void write_start_info(struct domain *d)
{
- /* XXX: handle suspend */
- xen_hypercall_shutdown(reason);
+ struct cpu_user_regs *regs = guest_cpu_user_regs();
+ start_info_t *si = map_domain_page(_mfn(is_pv_32bit_domain(d) ? regs->edx
+ :
regs->rdx));
+ uint64_t param;
+
+ BUG_ON(!si);
+
+ snprintf(si->magic, sizeof(si->magic), "xen-3.0-x86_%s",
+ is_pv_32bit_domain(d) ? "32p" : "64");
+ si->nr_pages = d->tot_pages;
+ si->shared_info = virt_to_maddr(d->shared_info);
+ si->flags = (xen_processor_pmbits << 8) & SIF_PM_MASK;
+ BUG_ON(xen_hypercall_hvm_get_param(HVM_PARAM_STORE_PFN, &si->store_mfn));
+ BUG_ON(xen_hypercall_hvm_get_param(HVM_PARAM_STORE_EVTCHN, ¶m));
+ si->store_evtchn = param;
+ BUG_ON(xen_hypercall_hvm_get_param(HVM_PARAM_CONSOLE_EVTCHN, ¶m));
+ si->console.domU.evtchn = param;
+ if ( !pv_console )
+ BUG_ON(xen_hypercall_hvm_get_param(HVM_PARAM_CONSOLE_PFN,
+ &si->console.domU.mfn));
+ else
+ si->console.domU.mfn = virt_to_mfn(consoled_get_ring_addr());
+
+ if ( is_pv_32bit_domain(d) )
+ xlat_start_info(si, XLAT_start_info_console_domU);
+
+ unmap_domain_page(si);
+}
+
+int pv_shim_shutdown(uint8_t reason)
+{
+ long rc;
+
+ if ( reason == SHUTDOWN_suspend )
+ {
+ struct domain *d = current->domain;
+ struct vcpu *v;
+ unsigned int i;
+ uint64_t old_store_pfn, old_console_pfn = 0, store_pfn, console_pfn;
+ uint64_t store_evtchn, console_evtchn;
+
+ BUG_ON(current->vcpu_id != 0);
+
+ BUG_ON(xen_hypercall_hvm_get_param(HVM_PARAM_STORE_PFN,
+ &old_store_pfn));
+ if ( !pv_console )
+ BUG_ON(xen_hypercall_hvm_get_param(HVM_PARAM_CONSOLE_PFN,
+ &old_console_pfn));
+
+ /* Pause the other vcpus before starting the migration. */
+ for_each_vcpu(d, v)
+ if ( v != current )
+ vcpu_pause_by_systemcontroller(v);
+
+ rc = xen_hypercall_shutdown(SHUTDOWN_suspend);
+ if ( rc )
+ {
+ for_each_vcpu(d, v)
+ if ( v != current )
+ vcpu_unpause_by_systemcontroller(v);
+
+ return rc;
+ }
+
+ /* Resume the shim itself first. */
+ hypervisor_resume();
+
+ /*
+ * ATM there's nothing Xen can do if the console/store pfn changes,
+ * because Xen won't have a page_info struct for it.
+ */
+ BUG_ON(xen_hypercall_hvm_get_param(HVM_PARAM_STORE_PFN,
+ &store_pfn));
+ BUG_ON(old_store_pfn != store_pfn);
+ if ( !pv_console )
+ {
+ BUG_ON(xen_hypercall_hvm_get_param(HVM_PARAM_CONSOLE_PFN,
+ &console_pfn));
+ BUG_ON(old_console_pfn != console_pfn);
+ }
+
+ /* Update domain id. */
+ d->domain_id = get_dom0_domid();
+
+ /* Clean the iomem range. */
+ BUG_ON(iomem_deny_access(d, 0, ~0UL));
+
+ /* Clean grant frames. */
+ xfree(grant_frames);
+ grant_frames = NULL;
+ nr_grant_list = 0;
+
+ /* Clean event channels. */
+ for ( i = 0; i < EVTCHN_2L_NR_CHANNELS; i++ )
+ {
+ if ( !port_is_valid(d, i) )
+ continue;
+
+ if ( evtchn_handled(d, i) )
+ evtchn_close(d, i, false);
+ else
+ evtchn_free(d, evtchn_from_port(d, i));
+ }
+
+ /* Reserve store/console event channel. */
+ BUG_ON(xen_hypercall_hvm_get_param(HVM_PARAM_STORE_EVTCHN,
+ &store_evtchn));
+ BUG_ON(evtchn_allocate_port(d, store_evtchn));
+ evtchn_reserve(d, store_evtchn);
+ BUG_ON(xen_hypercall_hvm_get_param(HVM_PARAM_CONSOLE_EVTCHN,
+ &console_evtchn));
+ BUG_ON(evtchn_allocate_port(d, console_evtchn));
+ evtchn_reserve(d, console_evtchn);
+
+ /* Clean watchdogs. */
+ watchdog_domain_destroy(d);
+ watchdog_domain_init(d);
+
+ /* Clean the PIRQ EOI page. */
+ if ( d->arch.pirq_eoi_map != NULL )
+ {
+ unmap_domain_page_global(d->arch.pirq_eoi_map);
+ put_page_and_type(mfn_to_page(d->arch.pirq_eoi_map_mfn));
+ d->arch.pirq_eoi_map = NULL;
+ d->arch.pirq_eoi_map_mfn = 0;
+ d->arch.auto_unmask = 0;
+ }
+
+ /*
+ * NB: there's no need to fixup the p2m, since the mfns assigned
+ * to the PV guest have not changed at all. Just re-write the
+ * start_info fields with the appropriate value.
+ */
+ write_start_info(d);
+
+ for_each_vcpu(d, v)
+ {
+ /* Unmap guest vcpu_info pages. */
+ unmap_vcpu_info(v);
+
+ /* Reset the periodic timer to the default value. */
+ v->periodic_period = MILLISECS(10);
+ /* Stop the singleshot timer. */
+ stop_timer(&v->singleshot_timer);
+
+ if ( test_bit(_VPF_down, &v->pause_flags) )
+ BUG_ON(vcpu_reset(v));
+
+ if ( v != current )
+ vcpu_unpause_by_systemcontroller(v);
+ else
+ vcpu_force_reschedule(v);
+ }
+ }
+ else
+ /* Forward to L0. */
+ rc = xen_hypercall_shutdown(reason);
+
+ return rc;
}
long pv_shim_event_channel_op(int cmd, XEN_GUEST_HANDLE_PARAM(void) arg)
diff --git a/xen/common/domain.c b/xen/common/domain.c
index d653a0b0bb..bc2ceb2d36 100644
--- a/xen/common/domain.c
+++ b/xen/common/domain.c
@@ -701,15 +701,12 @@ void __domain_crash_synchronous(void)
}
-void domain_shutdown(struct domain *d, u8 reason)
+int domain_shutdown(struct domain *d, u8 reason)
{
struct vcpu *v;
if ( pv_shim )
- {
- pv_shim_shutdown(reason);
- return;
- }
+ return pv_shim_shutdown(reason);
spin_lock(&d->shutdown_lock);
@@ -723,7 +720,7 @@ void domain_shutdown(struct domain *d, u8 reason)
if ( d->is_shutting_down )
{
spin_unlock(&d->shutdown_lock);
- return;
+ return 0;
}
d->is_shutting_down = 1;
@@ -745,6 +742,8 @@ void domain_shutdown(struct domain *d, u8 reason)
__domain_finalise_shutdown(d);
spin_unlock(&d->shutdown_lock);
+
+ return 0;
}
void domain_resume(struct domain *d)
diff --git a/xen/common/schedule.c b/xen/common/schedule.c
index 88279213e8..b7884263f2 100644
--- a/xen/common/schedule.c
+++ b/xen/common/schedule.c
@@ -1149,11 +1149,10 @@ ret_t do_sched_op(int cmd, XEN_GUEST_HANDLE_PARAM(void)
arg)
if ( copy_from_guest(&sched_shutdown, arg, 1) )
break;
- ret = 0;
TRACE_3D(TRC_SCHED_SHUTDOWN,
current->domain->domain_id, current->vcpu_id,
sched_shutdown.reason);
- domain_shutdown(current->domain, (u8)sched_shutdown.reason);
+ ret = domain_shutdown(current->domain, (u8)sched_shutdown.reason);
break;
}
diff --git a/xen/drivers/char/xen_pv_console.c
b/xen/drivers/char/xen_pv_console.c
index fb5a7893be..d2a28478ac 100644
--- a/xen/drivers/char/xen_pv_console.c
+++ b/xen/drivers/char/xen_pv_console.c
@@ -37,7 +37,7 @@ static DEFINE_SPINLOCK(tx_lock);
bool pv_console;
-void __init pv_console_init(void)
+void pv_console_init(void)
{
struct evtchn_unmask unmask;
long r;
diff --git a/xen/include/asm-x86/guest/xen.h b/xen/include/asm-x86/guest/xen.h
index 94f781c30f..c244569e83 100644
--- a/xen/include/asm-x86/guest/xen.h
+++ b/xen/include/asm-x86/guest/xen.h
@@ -40,6 +40,7 @@ void hypervisor_fixup_e820(struct e820map *e820);
void hypervisor_init_memory(void);
const unsigned long *hypervisor_reserved_pages(unsigned int *size);
uint32_t hypervisor_cpuid_base(void);
+void hypervisor_resume(void);
#else
@@ -91,6 +92,10 @@ static inline uint32_t hypervisor_cpuid_base(void)
ASSERT_UNREACHABLE();
return 0;
};
+static inline void hypervisor_resume(void)
+{
+ ASSERT_UNREACHABLE();
+};
#endif /* CONFIG_XEN_GUEST */
#endif /* __X86_GUEST_XEN_H__ */
diff --git a/xen/include/asm-x86/pv/shim.h b/xen/include/asm-x86/pv/shim.h
index 47bc4267af..0207348a85 100644
--- a/xen/include/asm-x86/pv/shim.h
+++ b/xen/include/asm-x86/pv/shim.h
@@ -35,7 +35,7 @@ void pv_shim_setup_dom(struct domain *d, l4_pgentry_t
*l4start,
unsigned long va_start, unsigned long store_va,
unsigned long console_va, unsigned long vphysmap,
start_info_t *si);
-void pv_shim_shutdown(uint8_t reason);
+int pv_shim_shutdown(uint8_t reason);
long pv_shim_event_channel_op(int cmd, XEN_GUEST_HANDLE_PARAM(void) arg);
void pv_shim_inject_evtchn(unsigned int port);
long pv_shim_grant_table_op(unsigned int cmd, XEN_GUEST_HANDLE_PARAM(void) uop,
@@ -53,9 +53,10 @@ static inline void pv_shim_setup_dom(struct domain *d,
l4_pgentry_t *l4start,
{
ASSERT_UNREACHABLE();
}
-static inline void pv_shim_shutdown(uint8_t reason)
+static inline int pv_shim_shutdown(uint8_t reason)
{
ASSERT_UNREACHABLE();
+ return 0;
}
static inline long pv_shim_event_channel_op(int cmd,
XEN_GUEST_HANDLE_PARAM(void) arg)
diff --git a/xen/include/xen/sched.h b/xen/include/xen/sched.h
index ac65d0c0df..70db377eae 100644
--- a/xen/include/xen/sched.h
+++ b/xen/include/xen/sched.h
@@ -605,7 +605,7 @@ static inline struct domain *rcu_lock_current_domain(void)
struct domain *get_domain_by_id(domid_t dom);
void domain_destroy(struct domain *d);
int domain_kill(struct domain *d);
-void domain_shutdown(struct domain *d, u8 reason);
+int domain_shutdown(struct domain *d, u8 reason);
void domain_resume(struct domain *d);
void domain_pause_for_debugger(void);
--
2.11.0
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxxx
https://lists.xenproject.org/mailman/listinfo/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |