[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH] tools/libxl: Fixes to stream v2 task joining logic
On Thu, Jul 23, 2015 at 12:42:25PM +0100, Andrew Cooper wrote: > On 23/07/15 12:38, Wei Liu wrote: > > On Thu, Jul 23, 2015 at 12:09:38PM +0100, Andrew Cooper wrote: > >> During review of the libxl migration v2 series, I changes the task > >> joining logic, but clearly didn't think the result through > >> properly. This patch fixes several errors. > >> > >> 1) Do not call check_all_finished() in the success cases of > >> libxl__xc_domain_{save,restore}_done(). It serves no specific purpose > >> as the save helper state will report itself as inactive by this point, > >> and avoids triggering a second stream->completion_callback() in the case > >> that write_toolstack_record()/stream_continue() record errors > >> synchronously themselves. > >> > >> 2) Only ever set stream->rc in stream_complete(). The first version of > >> the migration v2 series had separate rc and joined_rc parameters, where > >> this logic worked. However when combining the two, the teardown path > >> fails to trigger if stream_done() records stream->rc itself. A side > >> effect of this is that stream_done() needs to take an rc parameter. > >> > >> 3) Avoid stacking of check_all_finished() via synchronous teardown of > >> tasks. If the _abort() functions call back synchronously, > >> stream->completion_callback() ends up getting called twice, as first and > >> last check_all_finished() frames observe each task being finished. > >> > >> Signed-off-by: Andrew Cooper <andrew.cooper3@xxxxxxxxxx> > >> CC: Ian Campbell <Ian.Campbell@xxxxxxxxxx> > >> CC: Ian Jackson <Ian.Jackson@xxxxxxxxxxxxx> > >> CC: Wei Liu <wei.liu2@xxxxxxxxxx> > >> > >> --- > >> I found this while working to fix the toolstack record issue, but am > >> posting this bugfix ahead of the other work as OSSTest has tripped over > >> the issue. > > This change itself doesn't seem to have anything to do with libxc. In > > OSSTest the error that triggers this knock on effect is the failure of > > xc_map_foreign_bulk. Does that mean this patch only fix half of the > > problem seen in OSSTest? > > Saving to image new xl format (info 0x3/0x0/1797) > xc: info: In experimental xc_domain_save2 > xc: info: Saving domain 2, type x86 HVM > xc: progress: Memory: 67584/1344513 5% > xc: progress: Memory: 135168/1344513 10% > xc: progress: Memory: 201728/1344513 15% > xc: progress: Memory: 269312/1344513 20% > xc: progress: Memory: 336896/1344513 25% > xc: progress: Memory: 403456/1344513 30% > xc: progress: Memory: 471040/1344513 35% > xc: progress: Memory: 538624/1344513 40% > xc: progress: Memory: 605184/1344513 45% > xc: progress: Memory: 672768/1344513 50% > xc: progress: Memory: 740352/1344513 55% > xc: error: xc_map_foreign_bulk: mmap failed (12 = Cannot allocate > memory): Internal error > xc: error: Failed to map guest pages (12 = Cannot allocate memory): > Internal error > xc: error: Save failed (12 = Cannot allocate memory): Internal error > libxl: error: libxl_stream_write.c:267:libxl__xc_domain_save_done: > saving domain: domain responded to suspend request: Cannot allocate memory > xl: libxl_event.c:1866: libxl__ao_inprogress_gc: Assertion > `!ao->complete' failed. > > > I don't know why xc_map_foreign_bulk() failed, but at a guess I would > day dom0 is out of RAM. > I don't think so. There is no OOM error in serial log and dmesg. A similar failure is discovered in QEMU mainline test. That one is using OVMF + QEMU upstream. So this failure is not related to firmware or device model. http://logs.test-lab.xenproject.org/osstest/logs/59819/test-amd64-i386-xl-qemuu-debianhvm-amd64-xsm/info.html It should be a real bug manifests only on 32bit toolstack + guest with large amount of ram (5000 in this case). Looks like we have one more bug to hunt down. Wei. > This patch fixes the libxl assertion. > > ~Andrew _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |