[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-users] nestedhvm.
At 09:56 +0100 on 20 May (1400576182), Ian Campbell wrote: > Adding xen-devel and some relevant maintainers. > > > On 05/19/2014 11:40 AM, Ian Campbell wrote: > > > On Sun, 2014-05-18 at 08:02 -0400, Alvin Starr wrote: > > >> I am trying to run nested hypervisors to do some openstack experiments. > > >> I seem to be able to run xen-on-xen with no problems but if i try to run > > >> kvm-on-xen the system seems to spontaneously reboot. > > > >> I get the same results with xen 4.3 or 4.4. > > >> The dom0 is running fedora-20 > > >> The experiment environment is Centos6 with RDO > > On Mon, 2014-05-19 at 23:53 -0400, Alvin Starr wrote: > > Here is the serial port output. > > boot log along with panic. > > Which contains: > (XEN) mm locking order violation: 260 > 222 > (XEN) Xen BUG at mm-locks.h:118 > (full stack trace is below) > > That lead me to > http://lists.xen.org/archives/html/xen-devel/2013-02/msg01372.html but > not to a patch. Was there one? I've grepped the git logs for hints but > not found it... I don't believe there was, no. I'm not convinced that making shadow code do locked p2m lookups is the right answer, anyway, though I suppose it would stop this particular crash. In the meantime, at least it suggests a workaround, which is to boot the KVM VM with max-mem == memory (or however Openstack expresses that). Tim. > (XEN) ----[ Xen-4.3.2 x86_64 debug=n Not tainted ]---- > (XEN) CPU: 23 > (XEN) RIP: e008:[<ffff82c4c01ec7bb>] p2m_flush_table+0x1db/0x1f0 > (XEN) RFLAGS: 0000000000010286 CONTEXT: hypervisor > (XEN) rax: ffff8308299ed020 rbx: ffff831835cb0540 rcx: 0000000000000000 > (XEN) rdx: ffff8308299e0000 rsi: 000000000000000a rdi: ffff82c4c027d658 > (XEN) rbp: ffff82c4c031b648 rsp: ffff8308299e7998 r8: 0000000000000004 > (XEN) r9: 0000000000000000 r10: ffff82c4c022ce64 r11: 0000000000000003 > (XEN) r12: ffff83202cf99000 r13: 0000000000000000 r14: 0000000000000009 > (XEN) r15: 0000000000000000 cr0: 0000000080050033 cr4: 00000000000406f0 > (XEN) cr3: 0000001834178000 cr2: 0000000000000000 > (XEN) ds: 0000 es: 0000 fs: 0000 gs: 0000 ss: 0000 cs: e008 > (XEN) Xen stack trace from rsp=ffff8308299e7998: > (XEN) 0000000000000008 ffff83202cf99000 0000000000000006 0000000000000000 > (XEN) 0000000000000009 ffff82c4c01f0431 0000000000000000 ffff831835cb0010 > (XEN) 0000000000371600 ffff82c4c01f1dc5 2000000000000000 00000000016e8400 > (XEN) ffff831836e38c58 ffff8308299e7a08 0000000001836e38 ffff831836e38000 > (XEN) 0000000000000000 0000000000000000 0000000000000000 ffff831835cb0010 > (XEN) 00000000000ee200 0000000000000000 0000000000000200 ffff831835cb0010 > (XEN) 0000000000000001 0000000000371600 0000000000000200 ffff82c4c01ecf50 > (XEN) ffff83202cf99000 0000000700000006 0000000001836e37 ffff831835cb0010 > (XEN) ffff83202cf99000 ffff8308299e7af0 0000000000000200 0000000000371600 > (XEN) 00000000016e8400 ffff82c4c01f3c8f ffff8308299e7aec 0000000035cb0010 > (XEN) 0000000000000001 00000000016e8400 0000000000000200 ffff82c400000007 > (XEN) ffff83202cf99000 0000000700000000 ffff83040e4402c4 ffff831835cb0010 > (XEN) 0000000000000009 0000000000f9f600 00000000000ee200 0000000000000200 > (XEN) ffff83202cf99000 ffff82c4c01f6019 00000000000ee200 ffff830800000200 > (XEN) ffff831835cb04f8 ffff8308299e7f18 0000000000000003 ffff8308299e7c68 > (XEN) 0000000000000010 ffff82c4c01bcf83 ffff8308299e7ba0 ffff82c4c01f1222 > (XEN) 6000001800000000 ffffffff810402c4 ffff8308299e7c50 ffff8300aebdd000 > (XEN) ffff8308299e7c50 ffff8300aebdd000 0000000000000000 ffff82c4c01c85dc > (XEN) ffffffff81039e63 0a9b00100000000f 00000000ffffffff 0000000000000000 > (XEN) 00000000ffffffff 0000000000000000 00000000ffffffff ffff831835cb0010 > (XEN) Xen call trace: > (XEN) [<ffff82c4c01ec7bb>] p2m_flush_table+0x1db/0x1f0 > (XEN) [<ffff82c4c01f0431>] p2m_flush_nestedp2m+0x21/0x30 > (XEN) [<ffff82c4c01f1dc5>] p2m_set_entry+0x565/0x650 > (XEN) [<ffff82c4c01ecf50>] set_p2m_entry+0x90/0x130 > (XEN) [<ffff82c4c01f3c8f>] p2m_pod_zero_check_superpage+0x21f/0x460 > (XEN) [<ffff82c4c01f6019>] p2m_pod_demand_populate+0x699/0x890 > (XEN) [<ffff82c4c01bcf83>] hvm_emulate_one+0xc3/0x1f0 > (XEN) [<ffff82c4c01f1222>] p2m_gfn_to_mfn+0x392/0x3c0 > (XEN) [<ffff82c4c01c85dc>] handle_mmio+0x7c/0x1e0 > (XEN) [<ffff82c4c01f10e1>] p2m_gfn_to_mfn+0x251/0x3c0 > (XEN) [<ffff82c4c01eca58>] __get_gfn_type_access+0x68/0x210 > (XEN) [<ffff82c4c01c1843>] hvm_hap_nested_page_fault+0xc3/0x510 > (XEN) [<ffff82c4c011a447>] csched_vcpu_wake+0x367/0x580 > > > >> > > >> Any hints on what the problem may be or a good place to start to look to > > >> diagnose it? > > > You'll need to gather some logs I think. Ideally a serial console log or > > > if not try using "noreboot" on your hypervisor command line to try and > > > see the last messages before it reboots. > > > > > > Ian. > > > > > > > > > > > > _______________________________________________ Xen-users mailing list Xen-users@xxxxxxxxxxxxx http://lists.xen.org/xen-users
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |