|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] Xen-unstable-staging: Xen BUG at iommu_map.c:455
On 11/04/15 18:25, Sander Eikelenboom wrote:
> Saturday, April 11, 2015, 6:38:17 PM, you wrote:
>
>> On 11/04/15 17:32, Andrew Cooper wrote:
>>> On 11/04/15 17:21, Sander Eikelenboom wrote:
>>>> Saturday, April 11, 2015, 4:21:56 PM, you wrote:
>>>>
>>>>> On 11/04/15 15:11, Sander Eikelenboom wrote:
>>>>>> Friday, April 10, 2015, 8:55:27 PM, you wrote:
>>>>>>
>>>>>>> On 10/04/15 11:24, Sander Eikelenboom wrote:
>>>>>>>> Hi Andrew,
>>>>>>>>
>>>>>>>> Finally got some time to figure this out .. and i have narrowed it
>>>>>>>> down to:
>>>>>>>> git://xenbits.xen.org/staging/qemu-upstream-unstable.git
>>>>>>>> commit 7665d6ba98e20fb05c420de947c1750fd47e5c07 "Xen: Use the
>>>>>>>> ioreq-server API when available"
>>>>>>>> A straight revert of this commit prevents the issue from happening.
>>>>>>>>
>>>>>>>> The reason i had a hard time figuring this out was:
>>>>>>>> - I wasn't aware of this earlier, since git pulling the main xen tree,
>>>>>>>> doesn't
>>>>>>>> auto update the qemu-* trees.
>>>>>>> This has caught me out so many times. It is very non-obvious behaviour.
>>>>>>>> - So i happen to get this when i cloned a fresh tree to try to figure
>>>>>>>> out the
>>>>>>>> other issue i was seeing.
>>>>>>>> - After that checking out previous versions of the main xen tree
>>>>>>>> didn't resolve
>>>>>>>> this new issue, because the qemu tree doesn't get auto updated and
>>>>>>>> is set
>>>>>>>> "master".
>>>>>>>> - Cloning a xen-stable-4.5.0 made it go away .. because that has a
>>>>>>>> specific
>>>>>>>> git://xenbits.xen.org/staging/qemu-upstream-unstable.git tag which
>>>>>>>> is not
>>>>>>>> master.
>>>>>>>>
>>>>>>>> *sigh*
>>>>>>>>
>>>>>>>> This is tested with xen main tree at last commit
>>>>>>>> 3a28f760508fb35c430edac17a9efde5aff6d1d5
>>>>>>>> (normal xen-unstable, not the staging branch)
>>>>>>>>
>>>>>>>> Ok so i have added some extra debug info (see attached diff) and this
>>>>>>>> is the
>>>>>>>> output when it crashes due to something the commit above triggered,
>>>>>>>> the
>>>>>>>> level is out of bounds and the pfn looks fishy too.
>>>>>>>> Complete serial log from both bad and good (specific commit reverted)
>>>>>>>> are
>>>>>>>> attached.
>>>>>>> Just to confirm, you are positively identifying a qemu changeset as
>>>>>>> causing this crash?
>>>>>>> If so, the qemu change has discovered a pre-existing issue in the
>>>>>>> toolstack pci-passthrough interface. Whatever qemu is or isn't doing,
>>>>>>> it should not be able to cause a crash like this.
>>>>>>> With this in mind, I need to brush up on my AMD-Vi details.
>>>>>>> In the meantime, can you run with the following patch to identify what
>>>>>>> is going on, domctl wise? I assume it is the assign_device which is
>>>>>>> failing, but it will be nice to observe the differences between the
>>>>>>> working and failing case, which might offer a hint.
>>>>>> Hrrm with your patch i end up with a fatal page fault in
>>>>>> iommu_do_pci_domctl:
>>>>>>
>>>>>> (XEN) [2015-04-11 14:03:31.833] ----[ Xen-4.6-unstable x86_64 debug=y
>>>>>> Tainted: C ]----
>>>>>> (XEN) [2015-04-11 14:03:31.857] CPU: 5
>>>>>> (XEN) [2015-04-11 14:03:31.868] RIP: e008:[<ffff82d08014c52c>]
>>>>>> iommu_do_pci_domctl+0x2dc/0x740
>>>>>> (XEN) [2015-04-11 14:03:31.894] RFLAGS: 0000000000010256 CONTEXT:
>>>>>> hypervisor
>>>>>> (XEN) [2015-04-11 14:03:31.915] rax: 0000000000000008 rbx:
>>>>>> 0000000000000800 rcx: ffffffffffebe5ed
>>>>>> (XEN) [2015-04-11 14:03:31.942] rdx: 0000000000000800 rsi:
>>>>>> 0000000000000000 rdi: ffff830256ef7e38
>>>>>> (XEN) [2015-04-11 14:03:31.968] rbp: ffff830256ef7c98 rsp:
>>>>>> ffff830256ef7c08 r8: 00000000deadbeef
>>>>>> (XEN) [2015-04-11 14:03:31.995] r9: 00000000deadbeef r10:
>>>>>> ffff82d08024e500 r11: 0000000000000282
>>>>>> (XEN) [2015-04-11 14:03:32.022] r12: 0000000000000000 r13:
>>>>>> 0000000000000008 r14: 0000000000000000
>>>>>> (XEN) [2015-04-11 14:03:32.049] r15: 0000000000000000 cr0:
>>>>>> 0000000080050033 cr4: 00000000000006f0
>>>>>> (XEN) [2015-04-11 14:03:32.076] cr3: 00000002336a6000 cr2:
>>>>>> 0000000000000000
>>>>>> (XEN) [2015-04-11 14:03:32.096] ds: 0000 es: 0000 fs: 0000 gs:
>>>>>> 0000 ss: e010 cs: e008
>>>>>> (XEN) [2015-04-11 14:03:32.121] Xen stack trace from
>>>>>> rsp=ffff830256ef7c08:
>>>>>> (XEN) [2015-04-11 14:03:32.141] ffff830256ef7c78 ffff82d08012c178
>>>>>> ffff830256ef7c28 ffff830256ef7c28
>>>>>> (XEN) [2015-04-11 14:03:32.168] 0000000000000010 0000000000000000
>>>>>> 0000000000000000 0000000000000000
>>>>>> (XEN) [2015-04-11 14:03:32.195] 00000000000006f0 00007fe300000000
>>>>>> ffff830256eb7790 ffff83025cc6d300
>>>>>> (XEN) [2015-04-11 14:03:32.222] ffff82d080330c60 00007fe396bab004
>>>>>> 0000000000000000 00007fe396bab004
>>>>>> (XEN) [2015-04-11 14:03:32.249] 0000000000000000 0000000000000005
>>>>>> ffff830256ef7ca8 ffff82d08014900b
>>>>>> (XEN) [2015-04-11 14:03:32.276] ffff830256ef7d98 ffff82d080161f2d
>>>>>> 0000000000000010 0000000000000000
>>>>>> (XEN) [2015-04-11 14:03:32.303] 0000000000000000 ffff830256ef7ce8
>>>>>> ffff82d08018b655 ffff830256ef7d48
>>>>>> (XEN) [2015-04-11 14:03:32.330] ffff830256ef7cf8 ffff82d08018b66a
>>>>>> ffff830256ef7d38 ffff82d08012925e
>>>>>> (XEN) [2015-04-11 14:03:32.357] ffff830256efc068 0000000800000001
>>>>>> 800000022e12c167 0000000000000000
>>>>>> (XEN) [2015-04-11 14:03:32.384] 0000000000000002 ffff830256ef7e38
>>>>>> 0000000800000000 800000022e12c167
>>>>>> (XEN) [2015-04-11 14:03:32.411] 0000000000000003 ffff830256ef7db8
>>>>>> 0000000000000000 00007fe396780eb0
>>>>>> (XEN) [2015-04-11 14:03:32.439] 0000000000000202 ffffffffffffffff
>>>>>> 0000000000000000 00007fe396bab004
>>>>>> (XEN) [2015-04-11 14:03:32.466] 0000000000000000 0000000000000005
>>>>>> ffff830256ef7ef8 ffff82d08010497f
>>>>>> (XEN) [2015-04-11 14:03:32.493] 0000000000000001 0000000000100001
>>>>>> 800000022e12c167 ffff88001f7ecc00
>>>>>> (XEN) [2015-04-11 14:03:32.520] 00007fe396780eb0 ffff88001c849508
>>>>>> 0000000e00000007 ffffffff8105594a
>>>>>> (XEN) [2015-04-11 14:03:32.547] 000000000000e033 0000000000000202
>>>>>> ffff88001ece3d40 000000000000e02b
>>>>>> (XEN) [2015-04-11 14:03:32.574] ffff830256ef7e28 ffff82d080194933
>>>>>> 000000000000beef ffffffff81bd6c85
>>>>>> (XEN) [2015-04-11 14:03:32.601] ffff830256ef7f08 ffff82d080193edd
>>>>>> 0000000b0000002d 0000000000000001
>>>>>> (XEN) [2015-04-11 14:03:32.628] 0000000100000800 00007fe3962abbd0
>>>>>> ffff000a81050001 00007fe39656ce6e
>>>>>> (XEN) [2015-04-11 14:03:32.655] 00007ffdf2a654f0 00007fe39656d0c9
>>>>>> 00007fe39656ce6e 00007fe3969a9a55
>>>>>> (XEN) [2015-04-11 14:03:32.682] Xen call trace:
>>>>>> (XEN) [2015-04-11 14:03:32.695] [<ffff82d08014c52c>]
>>>>>> iommu_do_pci_domctl+0x2dc/0x740
>>>>>> (XEN) [2015-04-11 14:03:32.718] [<ffff82d08014900b>]
>>>>>> iommu_do_domctl+0x17/0x1a
>>>>>> (XEN) [2015-04-11 14:03:32.739] [<ffff82d080161f2d>]
>>>>>> arch_do_domctl+0x2469/0x26e1
>>>>>> (XEN) [2015-04-11 14:03:32.762] [<ffff82d08010497f>]
>>>>>> do_domctl+0x1a1f/0x1d60
>>>>>> (XEN) [2015-04-11 14:03:32.783] [<ffff82d080234c6b>]
>>>>>> syscall_enter+0xeb/0x145
>>>>>> (XEN) [2015-04-11 14:03:32.804]
>>>>>> (XEN) [2015-04-11 14:03:32.813] Pagetable walk from 0000000000000000:
>>>>>> (XEN) [2015-04-11 14:03:32.831] L4[0x000] = 0000000234075067
>>>>>> 000000000001f2a8
>>>>>> (XEN) [2015-04-11 14:03:32.852] L3[0x000] = 0000000229ad4067
>>>>>> 0000000000014c49
>>>>>> (XEN) [2015-04-11 14:03:32.873] L2[0x000] = 0000000000000000
>>>>>> ffffffffffffffff
>>>>>> (XEN) [2015-04-11 14:03:32.894]
>>>>>> (XEN) [2015-04-11 14:03:32.903] ****************************************
>>>>>> (XEN) [2015-04-11 14:03:32.922] Panic on CPU 5:
>>>>>> (XEN) [2015-04-11 14:03:32.935] FATAL PAGE FAULT
>>>>>> (XEN) [2015-04-11 14:03:32.948] [error_code=0000]
>>>>>> (XEN) [2015-04-11 14:03:32.961] Faulting linear address: 0000000000000000
>>>>>> (XEN) [2015-04-11 14:03:32.981] ****************************************
>>>>>> (XEN) [2015-04-11 14:03:33.000]
>>>>>> (XEN) [2015-04-11 14:03:33.009] Reboot in five seconds...
>>>>>>
>>>>>> The RIP resolves to the prink added by your patch in:
>>>>>>
>>>>>> case XEN_DOMCTL_test_assign_device:
>>>>>> ret = xsm_test_assign_device(XSM_HOOK,
>>>>>> domctl->u.assign_device.machine_sbdf);
>>>>>> if ( ret )
>>>>>> break;
>>>>>>
>>>>>> seg = domctl->u.assign_device.machine_sbdf >> 16;
>>>>>> bus = (domctl->u.assign_device.machine_sbdf >> 8) & 0xff;
>>>>>> devfn = domctl->u.assign_device.machine_sbdf & 0xff;
>>>>>>
>>>>>> printk("*** %pv->d%d: test_assign_device({%04x:%02x:%02x.%u})\n",
>>>>>> current, d->domain_id,
>>>>>> seg, bus, PCI_SLOT(devfn), PCI_FUNC(devfn));
>>>>>>
>>>>>> if ( device_assigned(seg, bus, devfn) )
>>>>>> {
>>>>>> printk(XENLOG_G_INFO
>>>>>> "%04x:%02x:%02x.%u already assigned, or
>>>>>> non-existent\n",
>>>>>> seg, bus, PCI_SLOT(devfn), PCI_FUNC(devfn));
>>>>>> ret = -EINVAL;
>>>>>> }
>>>>>> break;
>>>>> hmm - 'd' is NULL. This ought to work better.
>>>>> diff --git a/xen/drivers/passthrough/pci.c b/xen/drivers/passthrough/pci.c
>>>>> index 9f3413c..85ff1fc 100644
>>>>> --- a/xen/drivers/passthrough/pci.c
>>>>> +++ b/xen/drivers/passthrough/pci.c
>>>>> @@ -1532,6 +1532,11 @@ int iommu_do_pci_domctl(
>>>>> max_sdevs = domctl->u.get_device_group.max_sdevs;
>>>>> sdevs = domctl->u.get_device_group.sdev_array;
>>>>>
>>>>> + printk("*** %pv->d%d: get_device_group({%04x:%02x:%02x.%u,
>>>>> %u})\n",
>>>>> + current, d->domain_id,
>>>>> + seg, bus, PCI_SLOT(devfn), PCI_FUNC(devfn),
>>>>> + max_sdevs);
>>>>> +
>>>>> ret = iommu_get_device_group(d, seg, bus, devfn, sdevs,
>>>>> max_sdevs);
>>>>> if ( ret < 0 )
>>>>> {
>>>>> @@ -1558,6 +1563,9 @@ int iommu_do_pci_domctl(
>>>>> bus = (domctl->u.assign_device.machine_sbdf >> 8) & 0xff;
>>>>> devfn = domctl->u.assign_device.machine_sbdf & 0xff;
>>>>>
>>>>> + printk("*** %pv: test_assign_device({%04x:%02x:%02x.%u})\n",
>>>>> + current, seg, bus, PCI_SLOT(devfn), PCI_FUNC(devfn));
>>>>> +
>>>>> if ( device_assigned(seg, bus, devfn) )
>>>>> {
>>>>> printk(XENLOG_G_INFO
>>>>> @@ -1582,6 +1590,10 @@ int iommu_do_pci_domctl(
>>>>> bus = (domctl->u.assign_device.machine_sbdf >> 8) & 0xff;
>>>>> devfn = domctl->u.assign_device.machine_sbdf & 0xff;
>>>>>
>>>>> + printk("*** %pv->d%d: assign_device({%04x:%02x:%02x.%u})\n",
>>>>> + current, d->domain_id,
>>>>> + seg, bus, PCI_SLOT(devfn), PCI_FUNC(devfn));
>>>>> +
>>>>> ret = device_assigned(seg, bus, devfn) ?:
>>>>> assign_device(d, seg, bus, devfn);
>>>>> if ( ret == -ERESTART )
>>>>> @@ -1604,6 +1616,10 @@ int iommu_do_pci_domctl(
>>>>> bus = (domctl->u.assign_device.machine_sbdf >> 8) & 0xff;
>>>>> devfn = domctl->u.assign_device.machine_sbdf & 0xff;
>>>>>
>>>>> + printk("*** %pv->d%d: deassign_device({%04x:%02x:%02x.%u})\n",
>>>>> + current, d->domain_id,
>>>>> + seg, bus, PCI_SLOT(devfn), PCI_FUNC(devfn));
>>>>> +
>>>>> spin_lock(&pcidevs_lock);
>>>>> ret = deassign_device(d, seg, bus, devfn);
>>>>> spin_unlock(&pcidevs_lock);
>>>> Hi Andrew,
>>>>
>>>> Attached are the serial logs good (with revert) and bad (without):
>>>>
>>>> Some things that seems strange to me:
>>>> - The numerous calls to get the device 08:00.0 assigned ... for 0a:00.0
>>>> there
>>>> was only one call to both test assign and assign.
>>>> - However these numerous calls are there in both the good and the bad case,
>>>> so perhaps it's strange and wrong .. but not the cause ..
>>>> - I had a hunch it could be due to the 08:00.0 using MSI-X, but when only
>>>> passing through 0a:00.0, i get the same numerous calls but now for the
>>>> 0a:00.0 which uses IntX, so I think that is more related to being the
>>>> *first*
>>>> device to be passed through to a guest.
>>> I have also observed this behaviour, but not had time to investigate.
>>> It doesn't appear problematic in the longrun but it probably a toolstack
>>> issue which wants fixing (if only in the name of efficiency).
>> And just after I sent this email, I have realised why.
>> The first assign device will have to build IO pagetables, which is a
>> long operation and subject to hypercall continuations. The second
>> device will reused the same pagetables, so is quick to complete.
> So .. is the ioreq patch from Paul involved in providing something used in
> building
> the pagetables .. and could it have say some off by one resulting in the
> 0xffffffffffff .. which could lead to the pagetable building going beserk,
> requiring a paging_mode far greater than normally would be required .. which
> get's set .. since that isn't checked properly .. leading to things breaking
> a bit further when it does get checked ?
A -1 is slipping in somewhere and ending up in the gfn field.
The result is that update_paging_mode() attempts to construct
iopagetables to cover a 76bit address space, which is how level ends up
at 8. (Note that a level of 7 is reserved, and a level of anything
greater than 4 is implausible on your system.)
I think the crash is collateral damage following on from
update_paging_mode() not properly sanitising its input, but that there
is still some other issue causing -1 to be passed in the first place.
I am still trying to locate where a -1 might plausibly be coming from.
~Andrew
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |