[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH v6 for-xen-4.5 1/3] dpci: Move from domain centric model to hvm_dirq_dpci model.
When an interrupt for an PCI (or PCIe) passthrough device is to be sent to a guest, we find the appropiate 'hvm_dirq_dpci' structure for the interrupt (PIRQ), set a bit, and schedule an tasklet. Then the 'hvm_dirq_assist' tasklet gets called with the 'struct domain' from which we it iterates over all of the 'hvm_dirq_dpci' which are mapped to the guest. However, it is important to note that when we setup the 'hvm_dirq_dpci' we have a field for the 'struct domain' that we can use instead of having to pass in the 'struct domain'. As such this patch moves the tasklet to the 'struct hvm_dirq_dpci' and sets the 'dom' field to the domain. We have to be careful with this as that means we cannot set 'dom' to NULL until the tasklet has run. As such this patch introduces 'pt_pirq_reset' which resets the 'hvm_dirq_dpci' structure for proper usage during setup. The mechanism to tear it down is more complex as there are two ways it can be executed: a) pci_clean_dpci_irq. This gets called when the guest is being destroyed. We end up calling 'tasklet_kill'. The scenarios in which the 'struct pirq' (and subsequently the 'hvm_pirq_dpci') gets destroyed is when: - guest did not use the pirq at all after setup. - guest did use pirq, but decided to mask and left it in that state. - guest did use pirq, but crashed. In all of those scenarios we end up calling 'tasklet_kill' which will spin on the tasklet if it is running. b) pt_irq_destroy_bind (guest disables the MSI). We double-check that the softirq has run by piggy-backing on the existing 'pirq_cleanup_check' mechanism which calls 'pt_pirq_cleanup_check'. We add the extra call to 'pt_pirq_softirq_active' in 'pt_pirq_cleanup_check'. NOTE: Guests that use event channels unbind first the event channel from PIRQs, so the 'pt_pirq_cleanup_check' won't be called as eventch is set to zero. In that case we either clean it up via the a) mechanism. It is OK to re-use the tasklet when 'pt_irq_create_bind' is called afterwards. There is an extra scenario regardless of eventch being set or not: the guest did 'pt_irq_destroy_bind' while an interrupt was triggered and tasklet was scheduled (but had not been run). It is OK to still run the tasklet as hvm_dirq_assist won't do anything (as the flags are set to zero). Suggested-by: Jan Beulich <jbeulich@xxxxxxxx> Signed-off-by: Konrad Rzeszutek Wilk <konrad.wilk@xxxxxxxxxx> --- xen/drivers/passthrough/io.c | 39 ++++++++++++++++++++++++++++++--------- xen/drivers/passthrough/pci.c | 4 ++-- xen/include/xen/hvm/irq.h | 2 +- 3 files changed, 33 insertions(+), 12 deletions(-) diff --git a/xen/drivers/passthrough/io.c b/xen/drivers/passthrough/io.c index 4cd32b5..59b5c09 100644 --- a/xen/drivers/passthrough/io.c +++ b/xen/drivers/passthrough/io.c @@ -90,6 +90,25 @@ void free_hvm_irq_dpci(struct hvm_irq_dpci *dpci) xfree(dpci); } +/* + * Reset the 'struct hvm_pirq_dpci' values to proper states. + * + * N.B: + * The 'pt_irq_create_bind' can be called right after 'pt_irq_destroy_bind' + * was called. The 'pirq_cleanup_check' which would free the structure + * is only called if the event channel for the PIRQ is active. However + * OS-es that use event channels usually bind the PIRQ to an event channel + * and also unbind it before 'pt_irq_destroy_bind' is called which means + * we end up re-using the 'dpci' structure. This can be easily reproduced + * with unloading and loading the driver for the device. + * + * As such on every 'pt_irq_create_bind' call we MUST reset the values. + */ +static void pt_pirq_reset(struct domain *d, struct hvm_pirq_dpci *dpci) +{ + dpci->dom = d; +} + int pt_irq_create_bind( struct domain *d, xen_domctl_bind_pt_irq_t *pt_irq_bind) { @@ -114,9 +133,6 @@ int pt_irq_create_bind( spin_unlock(&d->event_lock); return -ENOMEM; } - softirq_tasklet_init( - &hvm_irq_dpci->dirq_tasklet, - hvm_dirq_assist, (unsigned long)d); for ( i = 0; i < NR_HVM_IRQS; i++ ) INIT_LIST_HEAD(&hvm_irq_dpci->girq[i]); @@ -130,6 +146,7 @@ int pt_irq_create_bind( return -ENOMEM; } pirq_dpci = pirq_dpci(info); + pt_pirq_reset(d, pirq_dpci); switch ( pt_irq_bind->irq_type ) { @@ -232,7 +249,6 @@ int pt_irq_create_bind( { unsigned int share; - pirq_dpci->dom = d; if ( pt_irq_bind->irq_type == PT_IRQ_TYPE_MSI_TRANSLATE ) { pirq_dpci->flags = HVM_IRQ_DPCI_MAPPED | @@ -258,7 +274,6 @@ int pt_irq_create_bind( { if ( pt_irq_need_timer(pirq_dpci->flags) ) kill_timer(&pirq_dpci->timer); - pirq_dpci->dom = NULL; list_del(&girq->list); list_del(&digl->list); hvm_irq_dpci->link_cnt[link]--; @@ -391,7 +406,6 @@ int pt_irq_destroy_bind( msixtbl_pt_unregister(d, pirq); if ( pt_irq_need_timer(pirq_dpci->flags) ) kill_timer(&pirq_dpci->timer); - pirq_dpci->dom = NULL; pirq_dpci->flags = 0; pirq_cleanup_check(pirq, d); } @@ -415,11 +429,18 @@ void pt_pirq_init(struct domain *d, struct hvm_pirq_dpci *dpci) { INIT_LIST_HEAD(&dpci->digl_list); dpci->gmsi.dest_vcpu_id = -1; + softirq_tasklet_init(&dpci->tasklet, hvm_dirq_assist, (unsigned long)dpci); } bool_t pt_pirq_cleanup_check(struct hvm_pirq_dpci *dpci) { - return !dpci->flags; + if ( !dpci->flags ) + { + tasklet_kill(&dpci->tasklet); + dpci->dom = NULL; + return 1; + } + return 0; } int pt_pirq_iterate(struct domain *d, @@ -459,7 +480,7 @@ int hvm_do_IRQ_dpci(struct domain *d, struct pirq *pirq) return 0; pirq_dpci->masked = 1; - tasklet_schedule(&dpci->dirq_tasklet); + tasklet_schedule(&pirq_dpci->tasklet); return 1; } @@ -564,7 +585,7 @@ static int _hvm_dirq_assist(struct domain *d, struct hvm_pirq_dpci *pirq_dpci, static void hvm_dirq_assist(unsigned long _d) { - struct domain *d = (struct domain *)_d; + struct domain *d = ((struct hvm_pirq_dpci *)_d)->dom; ASSERT(d->arch.hvm_domain.irq.dpci); diff --git a/xen/drivers/passthrough/pci.c b/xen/drivers/passthrough/pci.c index 1eba833..81e8a3a 100644 --- a/xen/drivers/passthrough/pci.c +++ b/xen/drivers/passthrough/pci.c @@ -767,6 +767,8 @@ static int pci_clean_dpci_irq(struct domain *d, xfree(digl); } + tasklet_kill(&pirq_dpci->tasklet); + return 0; } @@ -784,8 +786,6 @@ static void pci_clean_dpci_irqs(struct domain *d) hvm_irq_dpci = domain_get_irq_dpci(d); if ( hvm_irq_dpci != NULL ) { - tasklet_kill(&hvm_irq_dpci->dirq_tasklet); - pt_pirq_iterate(d, pci_clean_dpci_irq, NULL); d->arch.hvm_domain.irq.dpci = NULL; diff --git a/xen/include/xen/hvm/irq.h b/xen/include/xen/hvm/irq.h index c89f4b1..94a550a 100644 --- a/xen/include/xen/hvm/irq.h +++ b/xen/include/xen/hvm/irq.h @@ -88,7 +88,6 @@ struct hvm_irq_dpci { DECLARE_BITMAP(isairq_map, NR_ISAIRQS); /* Record of mapped Links */ uint8_t link_cnt[NR_LINK]; - struct tasklet dirq_tasklet; }; /* Machine IRQ to guest device/intx mapping. */ @@ -100,6 +99,7 @@ struct hvm_pirq_dpci { struct domain *dom; struct hvm_gmsi_info gmsi; struct timer timer; + struct tasklet tasklet; }; void pt_pirq_init(struct domain *, struct hvm_pirq_dpci *); -- 1.9.3 _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |