[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] test report for Xen 4.3 RC1
On Mon, May 27, 2013 at 03:49:27AM +0000, Ren, Yongjie wrote: > Hi All, > This is a report based on our testing for Xen 4.3.0 RC1 on Intel platforms. > (Sorry it's a little late. :-) If the status changes, I'll have an update > later.) OK, I've some updates and ideas that can help with narrowing some of these issues down. Thank you for doing this. > > Test environment: > Xen: Xen 4.3 RC1 with qemu-upstream-unstable.git > Dom0: Linux kernel 3.9.3 Could you please test v3.10-rc3. There have been some changes for the VCPU hotplug added in v3.10 that I am not sure whether they are in v3.9? > Hardware: Intel Sandy Bridge, Ivy Bridge, Haswell systems > > Below are the features we tested. > - PV and HVM guest booting (HVM: Ubuntu, Fedora, RHEL, Windows) > - Save/Restore and live migration > - PCI device assignment and SR-IOV > - power management: C-state/P-state, Dom0 S3, HVM S3 > - AVX and XSAVE instruction set > - MCE > - CPU online/offline for Dom0 > - vCPU hot-plug > - Nested Virtualization (Please look at my report in the following link.) > http://lists.xen.org/archives/html/xen-devel/2013-05/msg01145.html > > New bugs (4): (some of which are not regressions) > 1. sometimes failed to online cpu in Dom0 > http://bugzilla-archived.xenproject.org//bugzilla/show_bug.cgi?id=1851 That looks like you are hitting the udev race. Could you verify that these patches: https://lkml.org/lkml/2013/5/13/520 fix the issue (They are destined for v3.11) > 2. dom0 call trace when running sriov hvm guest with igbvf > http://bugzilla-archived.xenproject.org//bugzilla/show_bug.cgi?id=1852 > -- a regression in Linux kernel (Dom0). Hm, the call-trace you refer too: [ 68.404440] Already setup the GSI :37 [ 68.405105] igb 0000:04:00.0: Enabling SR-IOV VFs using the module parameter is deprecated - please use the pci sysfs interface. [ 68.506230] ------------[ cut here ]------------ [ 68.506265] WARNING: at /home/www/builds_xen_unstable/xen-src-27009-20130509/linux-2.6-pvops.git/fs/sysfs/dir.c:536 sysfs_add_one+0xcc/0xf0() [ 68.506279] Hardware name: S2600CP is a deprecated warning. Did you follow the 'pci sysfs' interface way? Looking at da36b64736cf2552e7fb5109c0255d4af804f5e7 ixgbe: Implement PCI SR-IOV sysfs callback operation it says it is using this: commit 1789382a72a537447d65ea4131d8bcc1ad85ce7b Author: Donald Dutile <ddutile@xxxxxxxxxx> Date: Mon Nov 5 15:20:36 2012 -0500 PCI: SRIOV control and status via sysfs Provide files under sysfs to determine the maximum number of VFs an SR-IOV-capable PCIe device supports, and methods to enable and disable the VFs on a per-device basis. Currently, VF enablement by SR-IOV-capable PCIe devices is done via driver-specific module parameters. If not setup in modprobe files, it requires admin to unload & reload PF drivers with number of desired VFs to enable. Additionally, the enablement is system wide: all devices controlled by the same driver have the same number of VFs enabled. Although the latter is probably desired, there are PCI configurations setup by system BIOS that may not enable that to occur. Two files are created for the PF of PCIe devices with SR-IOV support: sriov_totalvfs Contains the maximum number of VFs the device could support as reported by the TotalVFs register in the SR-IOV extended capability. sriov_numvfs Contains the number of VFs currently enabled on this device as reported by the NumVFs register in the SR-IOV extended capability. Writing zero to this file disables all VFs. Writing a positive number to this file enables that number of VFs. These files are readable for all SR-IOV PF devices. Writes to the sriov_numvfs file are effective only if a driver that supports the sriov_configure() method is attached. Signed-off-by: Donald Dutile <ddutile@xxxxxxxxxx> Can you try that please? > 3. Booting multiple guests will lead Dom0 call trace > http://bugzilla-archived.xenproject.org//bugzilla/show_bug.cgi?id=1853 That one worries me. Did you do a git bisect to figure out what is commit is causing this? > 4. After live migration, guest console continuously prints "Clocksource tsc > unstable" > http://bugzilla-archived.xenproject.org//bugzilla/show_bug.cgi?id=1854 This looks like a current bug with QEMU unstable missing a ACPI table? Did you try booting the guest with the old QEMU? device_model_version = 'qemu-xen-traditional' > > Old bugs: (11) > 1. [ACPI] Dom0 can't resume from S3 sleep > http://bugzilla.xen.org/bugzilla/show_bug.cgi?id=1707 That should be fixed in v3.11 (as now we have the fixes) Could you try v3.10 with the Rafael's ACPI tree merged in? (so the patches that he wants to submit for v3.11) > 2. [XL]"xl vcpu-set" causes dom0 crash or panic > http://bugzilla.xen.org/bugzilla/show_bug.cgi?id=1730 That I think is fixed in v3.10. Could you please check v3.10-rc3? > 3. Sometimes Xen panic on ia32pae Sandybridge when restore guest > http://bugzilla.xen.org/bugzilla/show_bug.cgi?id=1747 That looks to be with v2.6.32. Is the issue present with v3.9 or v3.10-rc3? > 4. 'xl vcpu-set' can't decrease the vCPU number of a HVM guest > http://bugzilla.xen.org/bugzilla/show_bug.cgi?id=1822 That I believe was an QEMU bug: http://lists.xen.org/archives/html/xen-devel/2013-05/msg01054.html which should be in QEMU traditional now (05-21 was when it went in the tree) > 5. Dom0 cannot be shutdown before PCI device detachment from guest > http://bugzilla.xen.org/bugzilla/show_bug.cgi?id=1826 Ok, I can reproduce that too. > 6. xl pci-list shows one PCI device (PF or VF) could be assigned to two > different guests > http://bugzilla.xen.org/bugzilla/show_bug.cgi?id=1834 OK, I can reproduce that too: > xl create /vm-pv.cfg Parsing config from /vm-pv.cfg libxl: error: libxl_pci.c:1043:libxl__device_pci_add: PCI device 0:1:0.0 is not assignable Daemon running with PID 3933 15:11:17 # 16 :/mnt/lab/latest/ > xl pci-list 1 Vdev Device 05.0 0000:01:00.0 > xl list Name ID Mem VCPUs State Time(s) Domain-0 0 2047 4 r----- 26.7 latest 1 2043 1 -b---- 5.3 latestadesa 4 1024 3 -b---- 5.1 15:11:20 # 20 :/mnt/lab/latest/ > xl pci-list 4 Vdev Device 00.0 0000:01:00.0 The rest I hadn't had a chance to look at. George, have you seen these issues? > 7. [upstream qemu] Guest free memory with upstream qemu is 14MB lower than > that with qemu-xen-unstable.git > http://bugzilla.xen.org/bugzilla/show_bug.cgi?id=1836 > 8. [upstream qemu]'maxvcpus=NUM' item is not supported in upstream QEMU > http://bugzilla.xen.org/bugzilla/show_bug.cgi?id=1837 > 9. [upstream qemu] Guest console hangs after save/restore or live-migration > when setting 'hpet=0' in guest config file > http://bugzilla.xen.org/bugzilla/show_bug.cgi?id=1838 > 10. [upstream qemu] 'xen_platform_pci=0' setting cannot make the guest use > emulated PCI devices by default > http://bugzilla.xen.org/bugzilla/show_bug.cgi?id=1839 > 11. Live migration fail when migrating the same guest for more than 2 times > http://bugzilla.xen.org/bugzilla/show_bug.cgi?id=1845 > > Best Regards, > Yongjie (Jay) > > > _______________________________________________ > Xen-devel mailing list > Xen-devel@xxxxxxxxxxxxx > http://lists.xen.org/xen-devel > _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |