[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH v5 00/21] libxl: domain save/restore: run in a separate process
Shriram Rajagopalan writes ("Re: [PATCH v5 00/21] libxl: domain save/restore: run in a separate process"): > Ian, > The code segfaults. Here are the system details and error traces from gdb. Thanks. > My setup: > > dom0 : ubuntu 64bit, 2.6.32-39 (pvops kernel), > running latest xen-4.2-unstable (built from your repo) > tools stack also built from your repo (which I hope has all the > latest patches). > > domU: ubuntu 32bit PV, xenolinux kernel (2.6.32.2 - novel suse version) > with suspend event channel support > > As a sanity check, I tested xl remus with latest tip from xen-unstable > mercurial repo, c/s: 25496:e08cf97e76f0 > > Blackhole replication (to /dev/null) and localhost replication worked as > expected > and the guest recovered properly without any issues. Thanks for the test runes. That didn't work entirely properly for me, even with the xen-unstable baseline. I did this xl -vvvv remus -b -i 100 debian.guest.osstest dummy >remus.log 2>&1 & The result was that the guest's networking broke. The guest shows up in xl list as debian.guest.osstest 7 512 1 ---ss- 5.2 and is still responsive on its pv console. After I killed the remus process, the guest's networking was still broken. At the start, the guest prints this on its console: [ 36.017241] WARNING: g.e. still in use! [ 36.021056] WARNING: g.e. still in use! [ 36.024740] WARNING: g.e. still in use! [ 36.024763] WARNING: g.e. still in use! If I try the rune with "localhost" I would have expected, surely, to see a domain with the incoming migration ? But I don't. I tried killing the `xl remus' process and the guest became wedged. However, when I apply my series, I can indeed produce an assertion failure: xc: detail: All memory is saved xc: error: Could not get domain info (3 = No such process): Internal error libxl: error: libxl.c:388:libxl_domain_resume: xc_domain_resume failed for domain 3077579968: No such process xl: libxl_event.c:1426: libxl__ao_inprogress_gc: Assertion `ao->magic == 0xA0FACE00ul' failed. So I have indeed made matters worse. > Blackhole replication: > ================ > xl error: > ---------- > xc: error: Could not get domain info (3 = No such process): Internal error > libxl: error: libxl.c:388:libxl_domain_resume: xc_domain_resume failed for > domain 4154075147<tel:4154075147>: No such process > libxl: error: libxl_dom.c:1184:libxl__domain_save_device_model: unable to > open qemu save file ?8b: No such file or directory I don't see that at all. NB that PV guests may have a qemu for certain disk backends, or consoles, depending on the configuration. Can you show me your domain config ? Mine is below. > I also ran xl in GDB to get a stack trace and hopefully some useful debug > info. > gdb traces: http://pastebin.com/7zFwFjW4 I get a different crash - see above. > Localhost replication: Partial success, but xl still segfaults > dmesg shows > [ 1399.254849] xl[4716]: segfault at 0 ip 00007f979483a417 sp > 00007fffe06043e0 error 6 in libxenlight.so.2.0.0[7f9794807000+4d000] I see exactly the same thing with `localhost' instead of `dummy'. And I see no incoming domain. I will investigate the crash I see. In the meantime can you try to help me see why it doesn't work me even with the baseline ? Thanks, Ian. # # Configuration file for the Xen instance debian.guest.osstest, created # by xen-tools 4.2 on Thu Apr 5 16:43:43 2012. # # # Kernel + memory size # #kernel = '/boot/vmlinuz-2.6.32.57' #ramdisk = '/boot/initrd.img-2.6.32.57' #bootloader = 'pygrub' bootloader = '/root/strace-pygrub' memory = '512' # # Disk device(s). # root = '/dev/xvda2 ro' disk = [ 'phy:/dev/bedbug/debian.guest.osstest-disk,xvda2,w', 'phy:/dev/bedbug/debian.guest.osstest-swap,xvda1,w', ] # # Physical volumes # # # Hostname # name = 'debian.guest.osstest' # # Networking # #dhcp = 'dhcp' vif = [ 'mac=5a:36:0e:26:00:01' ] # # Behaviour # on_poweroff = 'destroy' on_reboot = 'restart' on_crash='preserve' vcpus = 1 extra='console=hvc0 earlyprintk=xen' _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |