[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

[Xen-devel] [RFC PATCH] xen/arm: Deliver interrupts to vcpu specified in IROUTER



From: Vijaya Kumar K <Vijaya.Kumar@xxxxxxxxxxxxxxxxxx>

In GICv3 use IROUTER register contents to
deliver irq to specified vcpu's.

In error scenario fallback to vcpu 0.

This patch is similar to Stefano's commit
5b3a817ea33b891caf7d7d788da9ce6deffa82a1 for GICv2

Signed-off-by: Vijaya Kumar K <Vijaya.Kumar@xxxxxxxxxxxxxxxxxx>
---
 xen/arch/arm/vgic-v3.c |  115 ++++++++++++++++++++++++++++++++++++++++++++----
 1 file changed, 106 insertions(+), 9 deletions(-)

diff --git a/xen/arch/arm/vgic-v3.c b/xen/arch/arm/vgic-v3.c
index ecda672..b01a201 100644
--- a/xen/arch/arm/vgic-v3.c
+++ b/xen/arch/arm/vgic-v3.c
@@ -45,6 +45,35 @@
 #define GICV3_GICR_PIDR2  GICV3_GICD_PIDR2
 #define GICV3_GICR_PIDR4  GICV3_GICD_PIDR4
 
+static struct vcpu *vgicv3_irouter_to_vcpu(struct vcpu *v,
+                                           uint64_t irouter_aff)
+{
+    int i;
+    uint64_t cpu_affinity;
+    struct vcpu *v_target = NULL;
+
+    /* If routing mode is set. Fallback to vcpu0 */
+    if ( irouter_aff & GICD_IROUTER_SPI_MODE_ANY )
+        return v->domain->vcpu[0];
+
+    for ( i = 0; i < v->domain->max_vcpus; i++ )
+    {
+        v_target = v->domain->vcpu[i];
+        cpu_affinity = (MPIDR_AFFINITY_LEVEL(v_target->arch.vmpidr, 3) << 32 |
+                        MPIDR_AFFINITY_LEVEL(v_target->arch.vmpidr, 2) << 16 |
+                        MPIDR_AFFINITY_LEVEL(v_target->arch.vmpidr, 1) << 8  |
+                        MPIDR_AFFINITY_LEVEL(v_target->arch.vmpidr, 0));
+
+        if ( irouter_aff == cpu_affinity )
+            return v_target;
+    }
+    gdprintk(XENLOG_WARNING,
+             "d%d: No vcpu match found for irouter 0x%lx\n",
+             v->domain->domain_id, irouter_aff);
+    /* XXX: Fallback to vcpu0? */
+    return v->domain->vcpu[0];
+}
+
 static int __vgic_v3_rdistr_rd_mmio_read(struct vcpu *v, mmio_info_t *info,
                                          uint32_t gicr_reg)
 {
@@ -347,9 +376,9 @@ static int __vgic_v3_distr_common_mmio_write(struct vcpu 
*v, mmio_info_t *info,
         vgic_lock_rank(v, rank, flags);
         tr = rank->ienable;
         rank->ienable |= *r;
-        vgic_unlock_rank(v, rank, flags);
         /* The irq number is extracted from offset. so shift by register size 
*/
         vgic_enable_irqs(v, (*r) & (~tr), (reg - GICD_ISENABLER) >> DABT_WORD);
+        vgic_unlock_rank(v, rank, flags);
         return 1;
     case GICD_ICENABLER ... GICD_ICENABLERN:
         if ( dabt.size != DABT_WORD ) goto bad_width;
@@ -358,9 +387,9 @@ static int __vgic_v3_distr_common_mmio_write(struct vcpu 
*v, mmio_info_t *info,
         vgic_lock_rank(v, rank, flags);
         tr = rank->ienable;
         rank->ienable &= ~*r;
-        vgic_unlock_rank(v, rank, flags);
         /* The irq number is extracted from offset. so shift by register size 
*/
         vgic_disable_irqs(v, (*r) & tr, (reg - GICD_ICENABLER) >> DABT_WORD);
+        vgic_unlock_rank(v, rank, flags);
         return 1;
     case GICD_ISPENDR ... GICD_ISPENDRN:
         if ( dabt.size != DABT_WORD ) goto bad_width;
@@ -749,6 +778,9 @@ static int vgic_v3_distr_mmio_write(struct vcpu *v, 
mmio_info_t *info)
     register_t *r = select_user_reg(regs, dabt.reg);
     struct vgic_irq_rank *rank;
     unsigned long flags;
+    uint64_t new_target, old_target;
+    struct vcpu *old_vcpu, *new_vcpu;
+    int irq;
     int gicd_reg = (int)(info->gpa - v->domain->arch.vgic.dbase);
 
     switch ( gicd_reg )
@@ -804,16 +836,42 @@ static int vgic_v3_distr_mmio_write(struct vcpu *v, 
mmio_info_t *info)
         if ( dabt.size != DABT_DOUBLE_WORD ) goto bad_width;
         rank = vgic_rank_offset(v, 64, gicd_reg - GICD_IROUTER, 
DABT_DOUBLE_WORD);
         if ( rank == NULL) goto write_ignore_64;
-        if ( *r )
+        new_target = *r;
+        vgic_lock_rank(v, rank, flags);
+        old_target = rank->v3.irouter[REG_RANK_INDEX(64,
+                              (gicd_reg - GICD_IROUTER), DABT_DOUBLE_WORD)];
+        if ( *r & GICD_IROUTER_SPI_MODE_ANY )
         {
-            /* TODO: Ignored. We don't support irq delivery for vcpu != 0 */
-            gdprintk(XENLOG_DEBUG,
-                     "SPI delivery to secondary cpus not supported\n");
-            goto write_ignore_64;
+            /* 
+             * IRQ routing mode set. Route any one processor in the entire
+             * system. We chose vcpu 0
+             */
+            rank->v3.irouter[REG_RANK_INDEX(64,
+                     (gicd_reg - GICD_IROUTER), DABT_DOUBLE_WORD)] = *r;
+            vgic_unlock_rank(v, rank, flags);
+            return 1;
         }
-        vgic_lock_rank(v, rank, flags);
+        irq = (gicd_reg - GICD_IROUTER)/8;
+        /* IROUTER now specifies only one cpu affinity for this irq */
+        /* Migrate from any core(vcpu0) to new_target */
+        if ( (old_target & GICD_IROUTER_SPI_MODE_ANY) &&
+             !(new_target & GICD_IROUTER_SPI_MODE_ANY) )
+        {
+            new_vcpu = vgicv3_irouter_to_vcpu(v, new_target);
+            vgic_migrate_irq(v->domain->vcpu[0], new_vcpu, irq);
+        }
+        else
+        {
+            if ( old_target != new_target )
+            {
+                old_vcpu = vgicv3_irouter_to_vcpu(v, old_target);
+                new_vcpu = vgicv3_irouter_to_vcpu(v, new_target);
+                vgic_migrate_irq(old_vcpu, new_vcpu, irq);
+            }
+        }
+ 
         rank->v3.irouter[REG_RANK_INDEX(64,
-                      (gicd_reg - GICD_IROUTER), DABT_DOUBLE_WORD)] = *r;
+                      (gicd_reg - GICD_IROUTER), DABT_DOUBLE_WORD)] = 
new_target;
         vgic_unlock_rank(v, rank, flags);
         return 1;
     case GICD_NSACR ... GICD_NSACRN:
@@ -928,6 +986,44 @@ static const struct mmio_handler_ops 
vgic_distr_mmio_handler = {
     .write_handler = vgic_v3_distr_mmio_write,
 };
 
+static struct vcpu *vgicv3_get_target_vcpu(struct vcpu *v, unsigned int irq)
+{
+    int i;
+    uint64_t irq_affinity, cpu_affinity;
+    struct vcpu *v_target;
+    struct vgic_irq_rank *rank = vgic_rank_irq(v, irq);
+
+    ASSERT(spin_is_locked(&rank->lock));
+
+    irq_affinity = rank->v3.irouter[irq % 32];
+    if ( irq_affinity & GICD_IROUTER_SPI_MODE_ANY )
+    {
+        /* 
+         * IRQ routing mode set. Route any one processor in the entire
+         * system. We chose vcpu 0
+         */
+        return v->domain->vcpu[0];
+    }
+     
+    for ( i = 0; i < v->domain->max_vcpus; i++ )
+    {
+        v_target = v->domain->vcpu[i];
+        cpu_affinity = (MPIDR_AFFINITY_LEVEL(v_target->arch.vmpidr, 3) << 32 |
+                        MPIDR_AFFINITY_LEVEL(v_target->arch.vmpidr, 2) << 16 |
+                        MPIDR_AFFINITY_LEVEL(v_target->arch.vmpidr, 1) << 8  |
+                        MPIDR_AFFINITY_LEVEL(v_target->arch.vmpidr, 0));
+
+        if ( irq_affinity == cpu_affinity )
+            return v_target;
+    }
+    gdprintk(XENLOG_WARNING,
+             "d%d: No vcpu match found for irq %d with irouter 0x%lx\n",
+             v->domain->domain_id, irq, irq_affinity );
+
+    /* XXX:Fallback to vcpu0 ? */
+    return v->domain->vcpu[0];
+}
+
 static int vgicv3_vcpu_init(struct vcpu *v)
 {
     int i;
@@ -968,6 +1064,7 @@ static int vgicv3_domain_init(struct domain *d)
 static const struct vgic_ops v3_ops = {
     .vcpu_init   = vgicv3_vcpu_init,
     .domain_init = vgicv3_domain_init,
+    .get_target_vcpu = vgicv3_get_target_vcpu,
     .emulate_sysreg  = vgicv3_emulate_sysreg,
 };
 
-- 
1.7.9.5


_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.