|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH V3 15/29] x86/vvtd: Process interrupt remapping request
From: Chao Gao <chao.gao@xxxxxxxxx>
When a remapping interrupt request arrives, remapping hardware computes the
interrupt_index per the algorithm described in VTD spec
"Interrupt Remapping Table", interprets the IRTE and generates a remapped
interrupt request.
This patch introduces viommu_handle_irq_request() to emulate the process how
remapping hardware handles a remapping interrupt request.
Signed-off-by: Chao Gao <chao.gao@xxxxxxxxx>
Signed-off-by: Lan Tianyu <tianyu.lan@xxxxxxxxx>
---
v3:
- Encode map_guest_page()'s error into void* to avoid using another parameter
---
xen/drivers/passthrough/vtd/iommu.h | 21 +++
xen/drivers/passthrough/vtd/vvtd.c | 264 +++++++++++++++++++++++++++++++++++-
2 files changed, 284 insertions(+), 1 deletion(-)
diff --git a/xen/drivers/passthrough/vtd/iommu.h
b/xen/drivers/passthrough/vtd/iommu.h
index 703726f..790384f 100644
--- a/xen/drivers/passthrough/vtd/iommu.h
+++ b/xen/drivers/passthrough/vtd/iommu.h
@@ -218,6 +218,21 @@
#define dma_frcd_source_id(c) (c & 0xffff)
#define dma_frcd_page_addr(d) (d & (((u64)-1) << 12)) /* low 64 bit */
+enum VTD_FAULT_TYPE
+{
+ /* Interrupt remapping transition faults */
+ VTD_FR_IR_REQ_RSVD = 0x20, /* One or more IR request reserved
+ * fields set */
+ VTD_FR_IR_INDEX_OVER = 0x21, /* Index value greater than max */
+ VTD_FR_IR_ENTRY_P = 0x22, /* Present (P) not set in IRTE */
+ VTD_FR_IR_ROOT_INVAL = 0x23, /* IR Root table invalid */
+ VTD_FR_IR_IRTE_RSVD = 0x24, /* IRTE Rsvd field non-zero with
+ * Present flag set */
+ VTD_FR_IR_REQ_COMPAT = 0x25, /* Encountered compatible IR
+ * request while disabled */
+ VTD_FR_IR_SID_ERR = 0x26, /* Invalid Source-ID */
+};
+
/*
* 0: Present
* 1-11: Reserved
@@ -358,6 +373,12 @@ struct iremap_entry {
};
/*
+ * When VT-d doesn't enable Extended Interrupt Mode. Hardware only interprets
+ * only 8-bits ([15:8]) of Destination-ID field in the IRTEs.
+ */
+#define IRTE_xAPIC_DEST_MASK 0xff00
+
+/*
* Posted-interrupt descriptor address is 64 bits with 64-byte aligned, only
* the upper 26 bits of lest significiant 32 bits is available.
*/
diff --git a/xen/drivers/passthrough/vtd/vvtd.c
b/xen/drivers/passthrough/vtd/vvtd.c
index a0f63e9..90c00f5 100644
--- a/xen/drivers/passthrough/vtd/vvtd.c
+++ b/xen/drivers/passthrough/vtd/vvtd.c
@@ -23,11 +23,17 @@
#include <xen/types.h>
#include <xen/viommu.h>
#include <xen/xmalloc.h>
+#include <asm/apic.h>
#include <asm/current.h>
+#include <asm/event.h>
#include <asm/hvm/domain.h>
+#include <asm/io_apic.h>
#include <asm/page.h>
+#include <asm/p2m.h>
+#include <asm/viommu.h>
#include "iommu.h"
+#include "vtd.h"
/* Supported capabilities by vvtd */
unsigned int vvtd_caps = VIOMMU_CAP_IRQ_REMAPPING;
@@ -111,6 +117,132 @@ static inline uint64_t vvtd_get_reg_quad(struct vvtd
*vtd, uint32_t reg)
return vtd->regs->data64[reg/sizeof(uint64_t)];
}
+static void* map_guest_page(struct domain *d, uint64_t gfn)
+{
+ struct page_info *p;
+ void *ret;
+
+ p = get_page_from_gfn(d, gfn, NULL, P2M_ALLOC);
+ if ( !p )
+ return ERR_PTR(-EINVAL);
+
+ if ( !get_page_type(p, PGT_writable_page) )
+ {
+ put_page(p);
+ return ERR_PTR(-EINVAL);
+ }
+
+ ret = __map_domain_page_global(p);
+ if ( !ret )
+ {
+ put_page_and_type(p);
+ return ERR_PTR(-ENOMEM);
+ }
+
+ return ret;
+}
+
+static void unmap_guest_page(void *virt)
+{
+ struct page_info *page;
+
+ ASSERT((unsigned long)virt & PAGE_MASK);
+ page = mfn_to_page(domain_page_map_to_mfn(virt));
+
+ unmap_domain_page_global(virt);
+ put_page_and_type(page);
+}
+
+static void vvtd_inj_irq(struct vlapic *target, uint8_t vector,
+ uint8_t trig_mode, uint8_t delivery_mode)
+{
+ vvtd_debug("dest=v%d, delivery_mode=%x vector=%d trig_mode=%d\n",
+ vlapic_vcpu(target)->vcpu_id, delivery_mode, vector, trig_mode);
+
+ ASSERT((delivery_mode == dest_Fixed) ||
+ (delivery_mode == dest_LowestPrio));
+
+ vlapic_set_irq(target, vector, trig_mode);
+}
+
+static int vvtd_delivery(struct domain *d, uint8_t vector,
+ uint32_t dest, uint8_t dest_mode,
+ uint8_t delivery_mode, uint8_t trig_mode)
+{
+ struct vlapic *target;
+ struct vcpu *v;
+
+ switch ( delivery_mode )
+ {
+ case dest_LowestPrio:
+ target = vlapic_lowest_prio(d, NULL, 0, dest, dest_mode);
+ if ( target != NULL )
+ {
+ vvtd_inj_irq(target, vector, trig_mode, delivery_mode);
+ break;
+ }
+ vvtd_debug("null round robin: vector=%02x\n", vector);
+ break;
+
+ case dest_Fixed:
+ for_each_vcpu ( d, v )
+ if ( vlapic_match_dest(vcpu_vlapic(v), NULL, 0, dest, dest_mode) )
+ vvtd_inj_irq(vcpu_vlapic(v), vector, trig_mode, delivery_mode);
+ break;
+
+ case dest_NMI:
+ for_each_vcpu ( d, v )
+ if ( vlapic_match_dest(vcpu_vlapic(v), NULL, 0, dest, dest_mode) &&
+ !test_and_set_bool(v->nmi_pending) )
+ vcpu_kick(v);
+ break;
+
+ default:
+ gdprintk(XENLOG_WARNING, "Unsupported VTD delivery mode %d\n",
+ delivery_mode);
+ return -EINVAL;
+ }
+
+ return 0;
+}
+
+static uint32_t irq_remapping_request_index(
+ const struct arch_irq_remapping_request *irq)
+{
+ if ( irq->type == VIOMMU_REQUEST_IRQ_MSI )
+ {
+ uint32_t index;
+ struct msi_msg_remap_entry msi_msg =
+ {
+ .address_lo = { .val = irq->msg.msi.addr },
+ .data = irq->msg.msi.data,
+ };
+
+ index = (msi_msg.address_lo.index_15 << 15) +
+ msi_msg.address_lo.index_0_14;
+ if ( msi_msg.address_lo.SHV )
+ index += (uint16_t)msi_msg.data;
+
+ return index;
+ }
+ else if ( irq->type == VIOMMU_REQUEST_IRQ_APIC )
+ {
+ struct IO_APIC_route_remap_entry remap_rte = { .val = irq->msg.rte };
+
+ return (remap_rte.index_15 << 15) + remap_rte.index_0_14;
+ }
+ ASSERT_UNREACHABLE();
+
+ return 0;
+}
+
+static inline uint32_t irte_dest(struct vvtd *vvtd, uint32_t dest)
+{
+ /* In xAPIC mode, only 8-bits([15:8]) are valid */
+ return vvtd->status.eim_enabled ? dest :
+ MASK_EXTR(dest, IRTE_xAPIC_DEST_MASK);
+}
+
static void vvtd_handle_gcmd_ire(struct vvtd *vvtd, uint32_t val)
{
vvtd_info("%sable Interrupt Remapping",
@@ -255,6 +387,135 @@ static const struct hvm_mmio_ops vvtd_mmio_ops = {
.write = vvtd_write
};
+static void vvtd_handle_fault(struct vvtd *vvtd,
+ struct arch_irq_remapping_request *irq,
+ struct iremap_entry *irte,
+ unsigned int fault,
+ bool record_fault)
+{
+ if ( !record_fault )
+ return;
+
+ switch ( fault )
+ {
+ case VTD_FR_IR_SID_ERR:
+ case VTD_FR_IR_IRTE_RSVD:
+ case VTD_FR_IR_ENTRY_P:
+ if ( qinval_fault_disable(*irte) )
+ break;
+ /* fall through */
+ case VTD_FR_IR_INDEX_OVER:
+ case VTD_FR_IR_ROOT_INVAL:
+ /* TODO: handle fault (e.g. record and report this fault to VM */
+ break;
+
+ default:
+ gdprintk(XENLOG_INFO, "Can't handle VT-d fault %x\n", fault);
+ }
+ return;
+}
+
+static bool vvtd_irq_request_sanity_check(const struct vvtd *vvtd,
+ struct arch_irq_remapping_request
*irq)
+{
+ if ( irq->type == VIOMMU_REQUEST_IRQ_APIC )
+ {
+ struct IO_APIC_route_remap_entry rte = { .val = irq->msg.rte };
+
+ ASSERT(rte.format);
+ return !!rte.reserved;
+ }
+ else if ( irq->type == VIOMMU_REQUEST_IRQ_MSI )
+ {
+ struct msi_msg_remap_entry msi_msg =
+ { .address_lo = { .val = irq->msg.msi.addr } };
+
+ ASSERT(msi_msg.address_lo.format);
+ return 0;
+ }
+ ASSERT_UNREACHABLE();
+
+ return 0;
+}
+
+/*
+ * 'record_fault' is a flag to indicate whether we need recording a fault
+ * and notifying guest when a fault happens during fetching vIRTE.
+ */
+static int vvtd_get_entry(struct vvtd *vvtd,
+ struct arch_irq_remapping_request *irq,
+ struct iremap_entry *dest,
+ bool record_fault)
+{
+ uint32_t entry = irq_remapping_request_index(irq);
+ struct iremap_entry *irte, *irt_page;
+
+ vvtd_debug("interpret a request with index %x\n", entry);
+
+ if ( vvtd_irq_request_sanity_check(vvtd, irq) )
+ {
+ vvtd_handle_fault(vvtd, irq, NULL, VTD_FR_IR_REQ_RSVD, record_fault);
+ return -EINVAL;
+ }
+
+ if ( entry > vvtd->status.irt_max_entry )
+ {
+ vvtd_handle_fault(vvtd, irq, NULL, VTD_FR_IR_INDEX_OVER, record_fault);
+ return -EACCES;
+ }
+
+ irt_page = map_guest_page(vvtd->domain,
+ vvtd->status.irt + (entry >>
IREMAP_ENTRY_ORDER));
+ if ( IS_ERR(irt_page) )
+ {
+ vvtd_handle_fault(vvtd, irq, NULL, VTD_FR_IR_ROOT_INVAL, record_fault);
+ return PTR_ERR(irt_page);
+ }
+
+ irte = irt_page + (entry % (1 << IREMAP_ENTRY_ORDER));
+ dest->val = irte->val;
+ if ( !qinval_present(*irte) )
+ {
+ vvtd_handle_fault(vvtd, irq, NULL, VTD_FR_IR_ENTRY_P, record_fault);
+ unmap_guest_page(irt_page);
+ return -ENOENT;
+ }
+
+ /* Check reserved bits */
+ if ( (irte->remap.res_1 || irte->remap.res_2 || irte->remap.res_3 ||
+ irte->remap.res_4) )
+ {
+ vvtd_handle_fault(vvtd, irq, NULL, VTD_FR_IR_IRTE_RSVD, record_fault);
+ unmap_guest_page(irt_page);
+ return -EINVAL;
+ }
+
+ /* FIXME: We don't check against the source ID */
+ unmap_guest_page(irt_page);
+
+ return 0;
+}
+
+static int vvtd_handle_irq_request(struct domain *d,
+ struct arch_irq_remapping_request *irq)
+{
+ struct iremap_entry irte;
+ int ret;
+ struct vvtd *vvtd = domain_vvtd(d);
+
+ if ( !vvtd || !vvtd->status.intremap_enabled )
+ return -ENODEV;
+
+ ret = vvtd_get_entry(vvtd, irq, &irte, true);
+ if ( ret )
+ return ret;
+
+ return vvtd_delivery(vvtd->domain, irte.remap.vector,
+ irte_dest(vvtd, irte.remap.dst),
+ irte.remap.dm, irte.remap.dlm,
+ irte.remap.tm);
+}
+
static void vvtd_reset(struct vvtd *vvtd, uint64_t capability)
{
uint64_t cap = cap_set_num_fault_regs(1ULL) |
@@ -324,7 +585,8 @@ static int vvtd_destroy(struct viommu *viommu)
struct viommu_ops vvtd_hvm_vmx_ops = {
.create = vvtd_create,
- .destroy = vvtd_destroy
+ .destroy = vvtd_destroy,
+ .handle_irq_request = vvtd_handle_irq_request
};
static int vvtd_register(void)
--
1.8.3.1
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
https://lists.xen.org/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |