[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] Xen-unstable: pci-passthrough "irq 16: nobody cared" on HVM guest shutdown on irq of device not passed through.
Thursday, September 25, 2014, 6:14:43 PM, you wrote: >>>> On 25.09.14 at 17:49, <linux@xxxxxxxxxxxxxx> wrote: >> Thursday, September 25, 2014, 5:11:33 PM, you wrote: >> >>>>>> On 25.09.14 at 16:36, <linux@xxxxxxxxxxxxxx> wrote: >>>> - When shutting down the HVM guest when A happens the number of interrupts >>>> in >> >>>> /proc/interrups is still what it was, but when B happens it seems like a >>>> irq >> >>>> storm >>>> and after the irq nobody cared that ends with (always that 200000 so >>>> perhaps a threshold ?): >>>> 16: 200000 0 0 0 0 0 >> xen-pirq-ioapic-level snd_hda_intel >> >>> 100,000 is the traditional threshold, so I would expect the cited >>> instance to be the second one. It didn't really become clear to me >>> - is this observed in Dom0 or in the shutting down guest? >> >> This is from the /proc/interrupts of dom0 after the irq nobody cared message >> appeared in dom0 (so after B happened). Just after host boot and first guest >> boot it was stable around 500. On the next start (and after which B would >> happen on shutting the guest down again) it doubled to about 1000 (perhaps >> when >> the "Unsupported MSI delivery mode 3 for Dom2" occured). > Something odd must then be going on - the threshold _is_ 100,000, > not 200,000. >>> And did you really check that no other device (even if currently not having >>> an interrupt handler bound) is sitting on IRQ 16? >> >> In what way could i check that to be certain ? >> (if it's not bound, lspci and /proc/interrupts will probably be insufficient >> for that ?) > If the BIOS sets these up, lspci might still be of help. Consulting > boot messages of the kernel may also provide some hints. Beyond > that I'm not really sure how to figure out. > Jan lspci gives only one device with IRQ 16, the soundcontroller 00:14.2 Audio device: Advanced Micro Devices [AMD] nee ATI SBx00 Azalia (Intel HDA) (rev 40) Subsystem: Micro-Star International Co., Ltd. Device 7640 Flags: bus master, slow devsel, latency 64, IRQ 16 Memory at fdbf8000 (64-bit, non-prefetchable) [size=16K] Capabilities: [50] Power Management version 2 Kernel driver in use: snd_hda_intel On boot i do get a message "Already setup the GSI :16", however that seems to happen for multiple devices and irq/gsi's. I did have a go at copy and pasting the (hopefully) most relevant messages around IRQ's and MSI's for the different stages. But my untrained eye doesn't spot a difference that i can relate. ##Cold boot of the host system [ 35.556728] xen: registering gsi 16 triggering 0 polarity 1 [ 35.573157] xen: --> pirq=16 -> irq=16 (gsi=16) (XEN) [2014-09-25 13:08:55.771] IOAPIC[0]: Set PCI routing entry (6-16 -> 0x89 -> IRQ 16 Mode:1 Active:1) [ 38.575661] pciback 0000:09:00.0: enabling device (0000 -> 0003) [ 38.593584] xen: registering gsi 32 triggering 0 polarity 1 [ 38.610461] xen: --> pirq=32 -> irq=32 (gsi=32) (XEN) [2014-09-25 13:08:58.809] IOAPIC[1]: Set PCI routing entry (7-8 -> 0xc9 -> IRQ 32 Mode:1 Active:1) [ 42.713230] xen: registering gsi 16 triggering 0 polarity 1 [ 42.713233] Already setup the GSI :16 (XEN) [2014-09-25 13:29:04.111] IRQ: 16 affinity:01 vec:89 type=IO-APIC-level status=00000030 in-flight=0 domain-list=0: 16(---), (XEN) [2014-09-25 13:29:04.370] IRQ: 32 affinity:3f vec:c9 type=IO-APIC-level status=00000002 mapped, unbound (XEN) [2014-09-25 13:29:06.583] IRQ 16 Vec137: (XEN) [2014-09-25 13:29:06.597] Apic 0x00, Pin 16: vec=89 delivery=Fixed dest=L status=0 polarity=1 irr=0 trig=L mask=0 dest_id:1 (XEN) [2014-09-25 13:29:06.978] IRQ 32 Vec201: (XEN) [2014-09-25 13:29:06.991] Apic 0x01, Pin 8: vec=00 delivery=Fixed dest=L status=0 polarity=1 irr=0 trig=L mask=1 dest_id:63 (XEN) [2014-09-25 13:29:19.108] 0000:09:00.0 - dom 0 - MSIs < > (XEN) [2014-09-25 13:29:19.442] 0000:00:14.2 - dom 0 - MSIs < > ##Start of the HVM guest with pci device passed through (dom1). (XEN) [2014-09-25 13:30:32.831] io.c:280: d1: bind: m_gsi=32 g_gsi=36 dev=00.00.5 intx=0 (XEN) [2014-09-25 13:35:10.930] IRQ: 16 affinity:01 vec:89 type=IO-APIC-level status=00000030 in-flight=0 domain-list=0: 16(---), (XEN) [2014-09-25 13:35:11.189] IRQ: 32 affinity:02 vec:c9 type=IO-APIC-level status=00000010 in-flight=0 domain-list=1: 32(-M-), (XEN) [2014-09-25 13:35:12.498] IRQ: 84 affinity:04 vec:aa type=PCI-MSI status=00000030 in-flight=0 domain-list=1: 87(---), (XEN) [2014-09-25 13:35:13.443] IRQ 16 Vec137: (XEN) [2014-09-25 13:35:13.456] Apic 0x00, Pin 16: vec=89 delivery=Fixed dest=L status=0 polarity=1 irr=0 trig=L mask=0 dest_id:1 (XEN) [2014-09-25 13:35:13.837] IRQ 32 Vec201: (XEN) [2014-09-25 13:35:13.851] Apic 0x01, Pin 8: vec=c9 delivery=Fixed dest=L status=0 polarity=1 irr=0 trig=L mask=0 dest_id:2 (XEN) [2014-09-25 13:35:28.164] 0000:09:00.0 - dom 1 - MSIs < 84 > (XEN) [2014-09-25 13:35:28.515] 0000:00:14.2 - dom 0 - MSIs < > (XEN) [2014-09-25 13:35:37.013] MSI 84 vec=aa lowest edge assert log lowest dest=00000004 mask=0/1/? ##Shutdown of the HVM guest with pci device passed through, A happened. (XEN) [2014-09-25 13:38:27.974] IRQ: 16 affinity:01 vec:89 type=IO-APIC-level status=00000030 in-flight=0 domain-list=0: 16(---), (XEN) [2014-09-25 13:38:28.233] IRQ: 32 affinity:02 vec:c9 type=IO-APIC-level status=00000002 mapped, unbound (XEN) [2014-09-25 13:38:30.446] IRQ 16 Vec137: (XEN) [2014-09-25 13:38:30.459] Apic 0x00, Pin 16: vec=89 delivery=Fixed dest=L status=0 polarity=1 irr=0 trig=L mask=0 dest_id:1 (XEN) [2014-09-25 13:38:30.840] IRQ 32 Vec201: (XEN) [2014-09-25 13:38:30.854] Apic 0x01, Pin 8: vec=c9 delivery=Fixed dest=L status=0 polarity=1 irr=0 trig=L mask=1 dest_id:2 (XEN) [2014-09-25 13:38:39.255] 0000:09:00.0 - dom 0 - MSIs < > (XEN) [2014-09-25 13:38:39.590] 0000:00:14.2 - dom 0 - MSIs < > ##Start of the HVM guest with pci device passed through (dom2). (XEN) [2014-09-25 13:39:07.963] io.c:280: d2: bind: m_gsi=32 g_gsi=36 dev=00.00.5 intx=0 (XEN) [2014-09-25 13:39:48.149] d32767v2: Unsupported MSI delivery mode 3 for Dom2 (XEN) [2014-09-25 13:40:44.831] IRQ: 16 affinity:01 vec:89 type=IO-APIC-level status=00000030 in-flight=0 domain-list=0: 16(---), (XEN) [2014-09-25 13:40:45.089] IRQ: 32 affinity:02 vec:c9 type=IO-APIC-level status=00000010 in-flight=0 domain-list=2: 32(-M-), (XEN) [2014-09-25 13:40:46.398] IRQ: 84 affinity:02 vec:b2 type=PCI-MSI status=00000030 in-flight=0 domain-list=2: 87(---), (XEN) [2014-09-25 13:40:47.343] IRQ 16 Vec137: (XEN) [2014-09-25 13:40:47.357] Apic 0x00, Pin 16: vec=89 delivery=Fixed dest=L status=0 polarity=1 irr=0 trig=L mask=0 dest_id:1 (XEN) [2014-09-25 13:40:47.738] IRQ 32 Vec201: (XEN) [2014-09-25 13:40:47.751] Apic 0x01, Pin 8: vec=c9 delivery=Fixed dest=L status=0 polarity=1 irr=0 trig=L mask=0 dest_id:2 (XEN) [2014-09-25 13:40:57.567] 0000:09:00.0 - dom 2 - MSIs < 84 > (XEN) [2014-09-25 13:40:57.901] 0000:00:14.2 - dom 0 - MSIs < > (XEN) [2014-09-25 13:41:01.051] MSI 84 vec=b2 lowest edge assert log lowest dest=00000002 mask=0/1/? ##Shutdown of the HVM guest with pci device passed through, B happened. [ 2265.395971] irq 16: nobody cared (try booting with the "irqpoll" option) <call trace> [ 2266.234031] Disabling IRQ #16 (XEN) [2014-09-25 13:46:54.844] IRQ: 16 affinity:01 vec:89 type=IO-APIC-level status=00000030 in-flight=1 domain-list=0: 16(PMM), (XEN) [2014-09-25 13:46:55.103] IRQ: 32 affinity:02 vec:c9 type=IO-APIC-level status=00000002 mapped, unbound (XEN) [2014-09-25 13:46:57.316] IRQ 16 Vec137: (XEN) [2014-09-25 13:46:57.330] Apic 0x00, Pin 16: vec=89 delivery=Fixed dest=L status=0 polarity=1 irr=1 trig=L mask=0 dest_id:1 (XEN) [2014-09-25 13:46:57.711] IRQ 32 Vec201: (XEN) [2014-09-25 13:46:57.724] Apic 0x01, Pin 8: vec=c9 delivery=Fixed dest=L status=1 polarity=1 irr=0 trig=L mask=1 dest_id:2 (XEN) [2014-09-25 13:47:08.688] 0000:09:00.0 - dom 0 - MSIs < > (XEN) [2014-09-25 13:47:09.022] 0000:00:14.2 - dom 0 - MSIs < > _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |