[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH V3] X86/vMCE: handle broken page with regard to migration
Generally, there are 2 cases: 1. broken page occurs before migration 2. broken page occurs during migration We had submitted 2 versions of patches. Their difference is how to handle case 2: V1 patch aborted migration for the case 'broken page occurs during migration'; V2 patch marked broken page to dirty bitmap then got handled next round and no abort; This is V3 patch, adding handle for vMCE occur at last iteration of migration: At the sender For case 1, the broken page will be mapped but not copied to target (otherwise it may trigger more serious error, say, SRAR error). While its pfn_type and pfn number will be transferred to target so that target take appropriate action. For case 2, mce handler marks the broken page to dirty bitmap, so that at copypages stage of migration, its pfn_type and pfn number will be transferred to target and then take appropriate action. At the target When migration target populates pages for guest. As for the case of broken page wrt migration, we prefer to keep the type of the page, for the sake of seamless migration. At target it will set p2m as p2m_ram_broken for broken page. Guest MCE may have good chance to handle its broken page, while if guest access the broken page again it will kill itself as expected. If vMCE occur at the last memory copy iteration of migration, we do one more iteration so that the broken page's pfn_type and pfn number has chance to be transferred to target. Suggested-by: George Dunlap <george.dunlap@xxxxxxxxxxxxx> Signed-off-by: Liu Jinsong <jinsong.liu@xxxxxxxxx> Acked-by: George Dunlap <george.dunlap@xxxxxxxxxxxxx> - for tools part of the patch Acked-by: Ian Jackson <Ian.Jackson@xxxxxxxxxxxxx> - for hypervisor part of the patch Acked-by: Jan Beulich <JBeulich@xxxxxxxx> diff -r 8b93ac0c93f3 -r 446f6b9bfc89 tools/libxc/xc_domain.c --- a/tools/libxc/xc_domain.c Tue Nov 13 11:19:17 2012 +0000 +++ b/tools/libxc/xc_domain.c Sat Nov 17 09:46:05 2012 +0800 @@ -283,6 +283,22 @@ return ret; } +/* set broken page p2m */ +int xc_set_broken_page_p2m(xc_interface *xch, + uint32_t domid, + unsigned long pfn) +{ + int ret; + DECLARE_DOMCTL; + + domctl.cmd = XEN_DOMCTL_set_broken_page_p2m; + domctl.domain = (domid_t)domid; + domctl.u.set_broken_page_p2m.pfn = pfn; + ret = do_domctl(xch, &domctl); + + return ret ? -1 : 0; +} + /* get info from hvm guest for save */ int xc_domain_hvm_getcontext(xc_interface *xch, uint32_t domid, diff -r 8b93ac0c93f3 -r 446f6b9bfc89 tools/libxc/xc_domain_restore.c --- a/tools/libxc/xc_domain_restore.c Tue Nov 13 11:19:17 2012 +0000 +++ b/tools/libxc/xc_domain_restore.c Sat Nov 17 09:46:05 2012 +0800 @@ -1023,9 +1023,15 @@ countpages = count; for (i = oldcount; i < buf->nr_pages; ++i) - if ((buf->pfn_types[i] & XEN_DOMCTL_PFINFO_LTAB_MASK) == XEN_DOMCTL_PFINFO_XTAB - ||(buf->pfn_types[i] & XEN_DOMCTL_PFINFO_LTAB_MASK) == XEN_DOMCTL_PFINFO_XALLOC) + { + unsigned long pagetype; + + pagetype = buf->pfn_types[i] & XEN_DOMCTL_PFINFO_LTAB_MASK; + if ( pagetype == XEN_DOMCTL_PFINFO_XTAB || + pagetype == XEN_DOMCTL_PFINFO_BROKEN || + pagetype == XEN_DOMCTL_PFINFO_XALLOC ) --countpages; + } if (!countpages) return count; @@ -1267,6 +1273,17 @@ /* a bogus/unmapped/allocate-only page: skip it */ continue; + if ( pagetype == XEN_DOMCTL_PFINFO_BROKEN ) + { + if ( xc_set_broken_page_p2m(xch, dom, pfn) ) + { + ERROR("Set p2m for broken page failed, " + "dom=%d, pfn=%lx\n", dom, pfn); + goto err_mapped; + } + continue; + } + if (pfn_err[i]) { ERROR("unexpected PFN mapping failure pfn %lx map_mfn %lx p2m_mfn %lx", diff -r 8b93ac0c93f3 -r 446f6b9bfc89 tools/libxc/xc_domain_save.c --- a/tools/libxc/xc_domain_save.c Tue Nov 13 11:19:17 2012 +0000 +++ b/tools/libxc/xc_domain_save.c Sat Nov 17 09:46:05 2012 +0800 @@ -1118,7 +1118,7 @@ /* Now write out each data page, canonicalising page tables as we go... */ for ( ; ; ) { - unsigned int N, batch, run; + unsigned int N, batch, run, broken_page_num1, broken_page_num2; char reportbuf[80]; snprintf(reportbuf, sizeof(reportbuf), @@ -1270,6 +1270,7 @@ goto out; } + broken_page_num1 = 0; for ( run = j = 0; j < batch; j++ ) { unsigned long gmfn = pfn_batch[j]; @@ -1277,6 +1278,14 @@ if ( !hvm ) gmfn = pfn_to_mfn(gmfn); + if ( pfn_type[j] == XEN_DOMCTL_PFINFO_BROKEN ) + { + pfn_type[j] |= pfn_batch[j]; + ++broken_page_num1; + ++run; + continue; + } + if ( pfn_err[j] ) { if ( pfn_type[j] == XEN_DOMCTL_PFINFO_XTAB ) @@ -1371,8 +1380,12 @@ } } - /* skip pages that aren't present or are alloc-only */ + /* + * skip pages that aren't present, + * or are broken, or are alloc-only + */ if ( pagetype == XEN_DOMCTL_PFINFO_XTAB + || pagetype == XEN_DOMCTL_PFINFO_BROKEN || pagetype == XEN_DOMCTL_PFINFO_XALLOC ) continue; @@ -1484,6 +1497,35 @@ munmap(region_base, batch*PAGE_SIZE); + /* + * if vMCE occur at last iter, do one more iter so that it get + * chance to transfer broken page's pfn_type and pfn number to + * target and then take appropriate action + */ + if ( last_iter ) + { + for ( j = 0; j < batch; j++ ) + { + if ( hvm ) + pfn_type[j] = pfn_batch[j]; + else + pfn_type[j] = pfn_to_mfn(pfn_batch[j]); + } + + if ( xc_get_pfn_type_batch(xch, dom, batch, pfn_type) ) + { + PERROR("get_pfn_type_batch failed"); + goto out; + } + + broken_page_num2 = 0; + for ( j = 0; j < batch; j++ ) + if ( pfn_type[j] == XEN_DOMCTL_PFINFO_BROKEN ) + broken_page_num2++; + + if ( broken_page_num1 < broken_page_num2 ) + last_iter = 0; + } } /* end of this while loop for this iteration */ skip: @@ -1550,23 +1592,22 @@ PERROR("Error when writing to state file (tsc)"); goto out; } + } + } + else + last_iter = 1; + if ( xc_shadow_control(xch, dom, + XEN_DOMCTL_SHADOW_OP_CLEAN, HYPERCALL_BUFFER(to_send), + dinfo->p2m_size, NULL, 0, &shadow_stats) != dinfo->p2m_size ) + { + PERROR("Error flushing shadow PT"); + goto out; + } - } + sent_last_iter = sent_this_iter; - if ( xc_shadow_control(xch, dom, - XEN_DOMCTL_SHADOW_OP_CLEAN, HYPERCALL_BUFFER(to_send), - dinfo->p2m_size, NULL, 0, &shadow_stats) != dinfo->p2m_size ) - { - PERROR("Error flushing shadow PT"); - goto out; - } - - sent_last_iter = sent_this_iter; - - print_stats(xch, dom, sent_this_iter, &time_stats, &shadow_stats, 1); - - } + print_stats(xch, dom, sent_this_iter, &time_stats, &shadow_stats, 1); } /* end of infinite for loop */ DPRINTF("All memory is saved\n"); diff -r 8b93ac0c93f3 -r 446f6b9bfc89 tools/libxc/xenctrl.h --- a/tools/libxc/xenctrl.h Tue Nov 13 11:19:17 2012 +0000 +++ b/tools/libxc/xenctrl.h Sat Nov 17 09:46:05 2012 +0800 @@ -575,6 +575,17 @@ xc_domaininfo_t *info); /** + * This function set p2m for broken page + * &parm xch a handle to an open hypervisor interface + * @parm domid the domain id which broken page belong to + * @parm pfn the pfn number of the broken page + * @return 0 on success, -1 on failure + */ +int xc_set_broken_page_p2m(xc_interface *xch, + uint32_t domid, + unsigned long pfn); + +/** * This function returns information about the context of a hvm domain * @parm xch a handle to an open hypervisor interface * @parm domid the domain to get information from diff -r 8b93ac0c93f3 -r 446f6b9bfc89 xen/arch/x86/cpu/mcheck/mcaction.c --- a/xen/arch/x86/cpu/mcheck/mcaction.c Tue Nov 13 11:19:17 2012 +0000 +++ b/xen/arch/x86/cpu/mcheck/mcaction.c Sat Nov 17 09:46:05 2012 +0800 @@ -1,5 +1,6 @@ #include <xen/types.h> #include <xen/sched.h> +#include <asm/p2m.h> #include "mcaction.h" #include "vmce.h" #include "mce.h" @@ -91,6 +92,24 @@ goto vmce_failed; } + if ( is_hvm_domain(d) && !d->arch.hvm_domain.dirty_vram && + paging_mode_log_dirty(d) ) + { + /* + * vMCE occur during migration + * + * At sender, it marks broken page to dirty bitmap, + * so that at copypages stage of migration, broken + * page's pfn_type and pfn number would be transferred + * to target and then take appropriate action. + * + * At target, it would set p2m as p2m_ram_broken for + * broken page, so that if guest access the broken page + * again, it would kill itself as expected. + */ + paging_mark_dirty(d, mfn); + } + if ( unmmap_broken_page(d, _mfn(mfn), gfn) ) { printk("Unmap broken memory %lx for DOM%d failed\n", diff -r 8b93ac0c93f3 -r 446f6b9bfc89 xen/arch/x86/domctl.c --- a/xen/arch/x86/domctl.c Tue Nov 13 11:19:17 2012 +0000 +++ b/xen/arch/x86/domctl.c Sat Nov 17 09:46:05 2012 +0800 @@ -209,12 +209,18 @@ for ( j = 0; j < k; j++ ) { unsigned long type = 0; + p2m_type_t t; - page = get_page_from_gfn(d, arr[j], NULL, P2M_ALLOC); + page = get_page_from_gfn(d, arr[j], &t, P2M_ALLOC); if ( unlikely(!page) || unlikely(is_xen_heap_page(page)) ) - type = XEN_DOMCTL_PFINFO_XTAB; + { + if ( p2m_is_broken(t) ) + type = XEN_DOMCTL_PFINFO_BROKEN; + else + type = XEN_DOMCTL_PFINFO_XTAB; + } else { switch( page->u.inuse.type_info & PGT_type_mask ) @@ -235,6 +241,9 @@ if ( page->u.inuse.type_info & PGT_pinned ) type |= XEN_DOMCTL_PFINFO_LPINTAB; + + if ( page->count_info & PGC_broken ) + type = XEN_DOMCTL_PFINFO_BROKEN; } if ( page ) @@ -1568,6 +1577,29 @@ } break; + case XEN_DOMCTL_set_broken_page_p2m: + { + struct domain *d; + + d = rcu_lock_domain_by_id(domctl->domain); + if ( d != NULL ) + { + p2m_type_t pt; + unsigned long pfn = domctl->u.set_broken_page_p2m.pfn; + mfn_t mfn = get_gfn_query(d, pfn, &pt); + + if ( unlikely(!mfn_valid(mfn_x(mfn)) || !p2m_is_ram(pt) || + (p2m_change_type(d, pfn, pt, p2m_ram_broken) != pt)) ) + ret = -EINVAL; + + put_gfn(d, pfn); + rcu_unlock_domain(d); + } + else + ret = -ESRCH; + } + break; + default: ret = iommu_do_domctl(domctl, u_domctl); break; diff -r 8b93ac0c93f3 -r 446f6b9bfc89 xen/include/public/domctl.h --- a/xen/include/public/domctl.h Tue Nov 13 11:19:17 2012 +0000 +++ b/xen/include/public/domctl.h Sat Nov 17 09:46:05 2012 +0800 @@ -136,6 +136,7 @@ #define XEN_DOMCTL_PFINFO_LPINTAB (0x1U<<31) #define XEN_DOMCTL_PFINFO_XTAB (0xfU<<28) /* invalid page */ #define XEN_DOMCTL_PFINFO_XALLOC (0xeU<<28) /* allocate-only page */ +#define XEN_DOMCTL_PFINFO_BROKEN (0xdU<<28) /* broken page */ #define XEN_DOMCTL_PFINFO_LTAB_MASK (0xfU<<28) struct xen_domctl_getpageframeinfo { @@ -834,6 +835,12 @@ typedef struct xen_domctl_set_access_required xen_domctl_set_access_required_t; DEFINE_XEN_GUEST_HANDLE(xen_domctl_set_access_required_t); +struct xen_domctl_set_broken_page_p2m { + uint64_aligned_t pfn; +}; +typedef struct xen_domctl_set_broken_page_p2m xen_domctl_set_broken_page_p2m_t; +DEFINE_XEN_GUEST_HANDLE(xen_domctl_set_broken_page_p2m_t); + struct xen_domctl { uint32_t cmd; #define XEN_DOMCTL_createdomain 1 @@ -899,6 +906,7 @@ #define XEN_DOMCTL_set_access_required 64 #define XEN_DOMCTL_audit_p2m 65 #define XEN_DOMCTL_set_virq_handler 66 +#define XEN_DOMCTL_set_broken_page_p2m 67 #define XEN_DOMCTL_gdbsx_guestmemio 1000 #define XEN_DOMCTL_gdbsx_pausevcpu 1001 #define XEN_DOMCTL_gdbsx_unpausevcpu 1002 @@ -954,6 +962,7 @@ struct xen_domctl_audit_p2m audit_p2m; struct xen_domctl_set_virq_handler set_virq_handler; struct xen_domctl_gdbsx_memio gdbsx_guest_memio; + struct xen_domctl_set_broken_page_p2m set_broken_page_p2m; struct xen_domctl_gdbsx_pauseunp_vcpu gdbsx_pauseunp_vcpu; struct xen_domctl_gdbsx_domstatus gdbsx_domstatus; uint8_t pad[128]; _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |