[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] virq injection probelem in Xen ARM 4.5
On Wed, 22 Apr 2015, ììì wrote: > HI > Please send emails in plain text, it is very hard to read html on my MUA. > > I might find the VIRQ injection in Xen ARM 4.5Âand fix with simple way. > > Check please. > >  > > Problem > > - In Xen ARM 4.5, SPI's pending_irq structure could be accessed > simultaneously from each pCPU. > >  > > Reason > > - When inject SPI VIRQ to domain0, Xen ARM 4.5 set bit HW(31bit) bit in list > register.Âso domain0's EOI commandÂmakes deactivate > IRQ without Xen ARM interference. ButÂSPI's pending_irqÂis not updated > immediately . because Xen ARM 4.5 dosen't use maintenance > interrupt. After a while domain0's vcpu that receive SPI VIRQ is go to hyp > mode and updateÂSPI's pending_irqÂin > enter_hypervisor_head() function.ÂAfter EOI in domain0's, same SPI IRQ can > occur immediately andÂSPI's pending_irq is updated in > vgic_vcpu_inject_irq() function. > > If enter_hypervisor_head() is excecuted in pcpu1 andÂSPI's pending_irq is > updated in pcpu0, SPI's pending_irq is could be > accesedÂsimultaneously. To accessÂSPI's pending_irq Xen have to acquire > v->arch.vgic.lock. But this lock don't guarantee exclusive > access ofÂSPI's pending_irq. > > Example > > * Machine has 2 pcpu > > * Domain0 has 2 vcpu. vcpu0 is running on pcpu0 and vcpu1 is running on pcpu1. > > * All SPI is go to pcpu0 that run Domain0's vcpu0. > > * Assume that second IRQ 100 occur immediately after deactivate first IRQ 100. > > * Assume that first IRQ 100 injectedÂDomain0's vcpu1 and second IRQ 100 > injected Domain0's vcpu0. First IRQ 100 is managed in > pcpu1 and second IRQ 100 is managed in pcpu0. This can only happen if Dom0 writes to the virtual GICD_ITARGETSR register to change the delivery of IRQ 100. If the write is done on vcpu1, such a write would cause an immediate trap in the hypervisor and clearing of the LRs, before the interrupt routing (virtual and physical) is changed. If the write is done on vcpu0, the physical routing of the irq won't be changed yet, see arch_move_irqs. The next interrupt will still be delivererd to pcpu1, that is going to clear the LRs and only then change the physical interrupt routing. Therefore only the third IRQ 100 interrupt will reach pcpu0 directly. > - After deactivate first IRQ 100 in pcpu1, IRQ 100's pending_irq struct is > not updated immediately. But second IRQ 100 occur > andÂIRQ 100's pending_irq pcpu0.Âenter_hypervisor_head() and > vgic_vcpu_inject_irq() could be executedÂsimultaneously andÂIRQ 100's > pending_irq struct could be changedÂsimultaneously. > >  > > Fix > > - To prevent access SPI's pending_irq struct, deactivate irq must be executed > after updatedÂSPI's pending_irq struct. So Xen ARM > 4.5 cannot useÂHW(31bit) bit in list register to prevent domain0 deactivated > SPI.ÂXen ARM 4.5Âhave to manually deactivate SPI > after updatedÂSPI's pending_irq struct. > >  > > Code Changed > > - I Modify below functions. and this functions works well in my SPI Routing > Code in Arndale Board. > > --------------------------------------------------------------------------------------------- > > * > > static void gic_update_one_lr(struct vcpu *v, int i) > > { > >   struct pending_irq *p; > >   int irq; > >   struct gic_lr lr_val; > >  > >   ASSERT(spin_is_locked(&v->arch.vgic.lock)); > >   ASSERT(!local_irq_is_enabled()); > >  > >   gic_hw_ops->read_lr(i, &lr_val); > >   irq = lr_val.virq; > >   p = irq_to_pending(v, irq); > >   if ( lr_val.state & GICH_LR_ACTIVE ) > >   { > >     set_bit(GIC_IRQ_GUEST_ACTIVE, &p->status); > >     if ( test_bit(GIC_IRQ_GUEST_ENABLED, &p->status) && > >       Âtest_and_clear_bit(GIC_IRQ_GUEST_QUEUED, &p->status) ) > >     { > >       if ( p->desc == NULL ) > >       { > >         Âlr_val.state |= GICH_LR_PENDING; > >         Âgic_hw_ops->write_lr(i, &lr_val); > >       } > >       else > >         gdprintk(XENLOG_WARNING, "unable to inject hw irq=%d into > d%dv%d: already active in LR%d\n", > >             Âirq, v->domain->domain_id, v->vcpu_id, i); > >     } > >   } > >   else if ( lr_val.state & GICH_LR_PENDING ) > >   { > >     int q __attribute__ ((unused)) = > test_and_clear_bit(GIC_IRQ_GUEST_QUEUED, &p->status); > > #ifdef GIC_DEBUG > >     if ( q ) > >       gdprintk(XENLOG_DEBUG, "trying to inject irq=%d into d%dv%d, when > it is already pending in LR%d\n", > >           irq, v->domain->domain_id, v->vcpu_id, i); > > #endif > >   } > >   else > >   { > >     gic_hw_ops->clear_lr(i); > >     clear_bit(i, &this_cpu(lr_mask)); > >  > >     clear_bit(GIC_IRQ_GUEST_VISIBLE, &p->status); > >     clear_bit(GIC_IRQ_GUEST_ACTIVE, &p->status); > >     p->lr = GIC_INVALID_LR; > >     if ( test_bit(GIC_IRQ_GUEST_ENABLED, &p->status) && > >       Âtest_bit(GIC_IRQ_GUEST_QUEUED, &p->status) && > >       Â!test_bit(GIC_IRQ_GUEST_MIGRATING, &p->status) ) > >       gic_raise_guest_irq(v, irq, p->priority); > >     else { > >       list_del_init(&p->inflight); > >       if ( test_and_clear_bit(GIC_IRQ_GUEST_MIGRATING, &p->status) ) > >       { > >         struct vcpu *v_target = vgic_get_target_vcpu(v, irq); > >         irq_set_affinity(p->desc, cpumask_of(v_target->processor)); > >       } > >     } > > //* supsupi > >     if ( p->desc != NULL ) > > gic_hw_ops->deactivate_irq(p->desc); > >   } > >  > > } > > --------------------------------------------------------------------------------------------- > > static void gicv2_update_lr(int lr, const struct pending_irq *p, > >               unsigned int state) > > { > >   uint32_t lr_reg; > >  > >   BUG_ON(lr >= gicv2_info.nr_lrs); > >   BUG_ON(lr < 0); > >  > >   lr_reg = (((state & GICH_V2_LR_STATE_MASK) << GICH_V2_LR_STATE_SHIFT) Â| > >        ((GIC_PRI_TO_GUEST(p->priority) & GICH_V2_LR_PRIORITY_MASK) > >                       Â<< GICH_V2_LR_PRIORITY_SHIFT) | > >        ((p->irq & GICH_V2_LR_VIRTUAL_MASK) << > GICH_V2_LR_VIRTUAL_SHIFT)); > >  > > //* supsupi > >   if ( p->desc != NULL ) > > lr_reg |= GICH_V2_LR_MAINTENANCE_IRQ; > >  > >   writel_gich(lr_reg, GICH_LR + lr * 4); > >  > > } > > --------------------------------------------------------------------------------------------- > > Thank you > >  > > [?img=m9Rm%2BBFm%2BBFohAnZFAM9pxMXaAUlMxFoKxUwpoi0Fqk0MrUXFoEqM4ioFrpvtzFXp6UwFSl5WLl51zlqDBFdp6d5MreRhoRq%2Bzk4M6lT7NFdM6i0WzwGW > 40gpBE5Mr0db40%2F74FTWt%3D%3D.gif] > _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |