[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH v3 07/18] x86/hvm: unify internal portio and mmio intercepts
The implementation of mmio and portio intercepts is unnecessarily different. This leads to much code duplication. This patch unifies much of the intercept handling, leaving only distinct handlers for stdvga mmio and dpci portio. Subsequent patches will unify those handlers. Signed-off-by: Paul Durrant <paul.durrant@xxxxxxxxxx> Cc: Keir Fraser <keir@xxxxxxx> Cc: Jan Beulich <jbeulich@xxxxxxxx> Cc: Andrew Cooper <andrew.cooper3@xxxxxxxxxx> --- xen/arch/x86/hvm/emulate.c | 11 +- xen/arch/x86/hvm/hpet.c | 4 +- xen/arch/x86/hvm/hvm.c | 7 +- xen/arch/x86/hvm/intercept.c | 502 +++++++++++++---------------- xen/arch/x86/hvm/stdvga.c | 30 +- xen/arch/x86/hvm/vioapic.c | 4 +- xen/arch/x86/hvm/vlapic.c | 5 +- xen/arch/x86/hvm/vmsi.c | 7 +- xen/drivers/passthrough/amd/iommu_guest.c | 30 +- xen/include/asm-x86/hvm/domain.h | 1 + xen/include/asm-x86/hvm/io.h | 119 +++---- 11 files changed, 350 insertions(+), 370 deletions(-) diff --git a/xen/arch/x86/hvm/emulate.c b/xen/arch/x86/hvm/emulate.c index 02796d0..91e57b0 100644 --- a/xen/arch/x86/hvm/emulate.c +++ b/xen/arch/x86/hvm/emulate.c @@ -143,16 +143,7 @@ static int hvmemul_do_io( hvmtrace_io_assist(&p); } - if ( is_mmio ) - { - rc = hvm_mmio_intercept(&p); - if ( rc == X86EMUL_UNHANDLEABLE ) - rc = hvm_buffered_io_intercept(&p); - } - else - { - rc = hvm_portio_intercept(&p); - } + rc = hvm_io_intercept(&p); switch ( rc ) { diff --git a/xen/arch/x86/hvm/hpet.c b/xen/arch/x86/hvm/hpet.c index 9585ca8..8958873 100644 --- a/xen/arch/x86/hvm/hpet.c +++ b/xen/arch/x86/hvm/hpet.c @@ -504,7 +504,7 @@ static int hpet_range(struct vcpu *v, unsigned long addr) (addr < (HPET_BASE_ADDRESS + HPET_MMAP_SIZE)) ); } -const struct hvm_mmio_ops hpet_mmio_ops = { +static const struct hvm_mmio_ops hpet_mmio_ops = { .check = hpet_range, .read = hpet_read, .write = hpet_write @@ -659,6 +659,8 @@ void hpet_init(struct domain *d) h->hpet.comparator64[i] = ~0ULL; h->pt[i].source = PTSRC_isa; } + + register_mmio_handler(d, &hpet_mmio_ops); } void hpet_deinit(struct domain *d) diff --git a/xen/arch/x86/hvm/hvm.c b/xen/arch/x86/hvm/hvm.c index 535d622..c10db78 100644 --- a/xen/arch/x86/hvm/hvm.c +++ b/xen/arch/x86/hvm/hvm.c @@ -1465,11 +1465,12 @@ int hvm_domain_initialise(struct domain *d) goto fail0; d->arch.hvm_domain.params = xzalloc_array(uint64_t, HVM_NR_PARAMS); - d->arch.hvm_domain.io_handler = xmalloc(struct hvm_io_handler); + d->arch.hvm_domain.io_handler = xzalloc_array(struct hvm_io_handler, + NR_IO_HANDLERS); rc = -ENOMEM; if ( !d->arch.hvm_domain.params || !d->arch.hvm_domain.io_handler ) goto fail1; - d->arch.hvm_domain.io_handler->num_slot = 0; + d->arch.hvm_domain.io_handler_count = 0; /* Set the default IO Bitmap. */ if ( is_hardware_domain(d) ) @@ -1506,6 +1507,8 @@ int hvm_domain_initialise(struct domain *d) rtc_init(d); + msixtbl_init(d); + register_portio_handler(d, 0xe9, 1, hvm_print_line); register_portio_handler(d, 0xcf8, 4, hvm_access_cf8); diff --git a/xen/arch/x86/hvm/intercept.c b/xen/arch/x86/hvm/intercept.c index cc44733..4db024e 100644 --- a/xen/arch/x86/hvm/intercept.c +++ b/xen/arch/x86/hvm/intercept.c @@ -32,205 +32,97 @@ #include <xen/event.h> #include <xen/iommu.h> -static const struct hvm_mmio_ops *const -hvm_mmio_handlers[HVM_MMIO_HANDLER_NR] = +static bool_t hvm_mmio_accept(struct hvm_io_handler *handler, + uint64_t addr, + uint64_t size) { - &hpet_mmio_ops, - &vlapic_mmio_ops, - &vioapic_mmio_ops, - &msixtbl_mmio_ops, - &iommu_mmio_ops -}; + BUG_ON(handler->type != IOREQ_TYPE_COPY); + + return handler->u.mmio.ops->check(current, addr); +} -static int hvm_mmio_access(struct vcpu *v, - ioreq_t *p, - hvm_mmio_read_t read, - hvm_mmio_write_t write) +static int hvm_mmio_read(struct hvm_io_handler *handler, + uint64_t addr, + uint64_t size, + uint64_t *data) { - struct hvm_vcpu_io *vio = &v->arch.hvm_vcpu.hvm_io; - unsigned long data; - int rc = X86EMUL_OKAY, i, step = p->df ? -p->size : p->size; + BUG_ON(handler->type != IOREQ_TYPE_COPY); - if ( !p->data_is_ptr ) - { - if ( p->dir == IOREQ_READ ) - { - if ( vio->mmio_retrying ) - { - if ( vio->mmio_large_read_bytes != p->size ) - return X86EMUL_UNHANDLEABLE; - memcpy(&data, vio->mmio_large_read, p->size); - vio->mmio_large_read_bytes = 0; - vio->mmio_retrying = 0; - } - else - rc = read(v, p->addr, p->size, &data); - p->data = data; - } - else /* p->dir == IOREQ_WRITE */ - rc = write(v, p->addr, p->size, p->data); - return rc; - } + return handler->u.mmio.ops->read(current, addr, size, data); +} - if ( p->dir == IOREQ_READ ) - { - for ( i = 0; i < p->count; i++ ) - { - if ( vio->mmio_retrying ) - { - if ( vio->mmio_large_read_bytes != p->size ) - return X86EMUL_UNHANDLEABLE; - memcpy(&data, vio->mmio_large_read, p->size); - vio->mmio_large_read_bytes = 0; - vio->mmio_retrying = 0; - } - else - { - rc = read(v, p->addr + step * i, p->size, &data); - if ( rc != X86EMUL_OKAY ) - break; - } - switch ( hvm_copy_to_guest_phys(p->data + step * i, - &data, p->size) ) - { - case HVMCOPY_okay: - break; - case HVMCOPY_gfn_paged_out: - case HVMCOPY_gfn_shared: - rc = X86EMUL_RETRY; - break; - case HVMCOPY_bad_gfn_to_mfn: - /* Drop the write as real hardware would. */ - continue; - case HVMCOPY_bad_gva_to_gfn: - ASSERT(0); - /* fall through */ - default: - rc = X86EMUL_UNHANDLEABLE; - break; - } - if ( rc != X86EMUL_OKAY) - break; - } +static int hvm_mmio_write(struct hvm_io_handler *handler, + uint64_t addr, + uint64_t size, + uint64_t data) +{ + BUG_ON(handler->type != IOREQ_TYPE_COPY); - if ( rc == X86EMUL_RETRY ) - { - vio->mmio_retry = 1; - vio->mmio_large_read_bytes = p->size; - memcpy(vio->mmio_large_read, &data, p->size); - } - } - else - { - for ( i = 0; i < p->count; i++ ) - { - switch ( hvm_copy_from_guest_phys(&data, p->data + step * i, - p->size) ) - { - case HVMCOPY_okay: - break; - case HVMCOPY_gfn_paged_out: - case HVMCOPY_gfn_shared: - rc = X86EMUL_RETRY; - break; - case HVMCOPY_bad_gfn_to_mfn: - data = ~0; - break; - case HVMCOPY_bad_gva_to_gfn: - ASSERT(0); - /* fall through */ - default: - rc = X86EMUL_UNHANDLEABLE; - break; - } - if ( rc != X86EMUL_OKAY ) - break; - rc = write(v, p->addr + step * i, p->size, data); - if ( rc != X86EMUL_OKAY ) - break; - } + return handler->u.mmio.ops->write(current, addr, size, data); +} - if ( rc == X86EMUL_RETRY ) - vio->mmio_retry = 1; - } +static const struct hvm_io_ops mmio_ops = { + .accept = hvm_mmio_accept, + .read = hvm_mmio_read, + .write = hvm_mmio_write +}; - if ( i != 0 ) - { - p->count = i; - rc = X86EMUL_OKAY; - } +static bool_t hvm_portio_accept(struct hvm_io_handler *handler, + uint64_t addr, + uint64_t size) +{ + BUG_ON(handler->type != IOREQ_TYPE_PIO); - return rc; + return (addr >= handler->u.portio.start) && + ((addr + size) <= handler->u.portio.end); } -bool_t hvm_mmio_internal(paddr_t gpa) +static int hvm_portio_read(struct hvm_io_handler *handler, + uint64_t addr, + uint64_t size, + uint64_t *data) { - struct vcpu *curr = current; - unsigned int i; + uint32_t val = 0; + int rc; - for ( i = 0; i < HVM_MMIO_HANDLER_NR; ++i ) - if ( hvm_mmio_handlers[i]->check(curr, gpa) ) - return 1; + BUG_ON(handler->type != IOREQ_TYPE_PIO); - return 0; + rc = handler->u.portio.action(IOREQ_READ, addr, size, &val); + if ( rc == X86EMUL_OKAY ) + *data = val; + + return rc; } -int hvm_mmio_intercept(ioreq_t *p) +static int hvm_portio_write(struct hvm_io_handler *handler, + uint64_t addr, + uint64_t size, + uint64_t data) { - struct vcpu *v = current; - int i; + uint32_t val = data; - for ( i = 0; i < HVM_MMIO_HANDLER_NR; i++ ) - { - hvm_mmio_check_t check = hvm_mmio_handlers[i]->check; - - if ( check(v, p->addr) ) - { - if ( unlikely(p->count > 1) && - !check(v, unlikely(p->df) - ? p->addr - (p->count - 1L) * p->size - : p->addr + (p->count - 1L) * p->size) ) - p->count = 1; - - return hvm_mmio_access( - v, p, - hvm_mmio_handlers[i]->read, - hvm_mmio_handlers[i]->write); - } - } + BUG_ON(handler->type != IOREQ_TYPE_PIO); - return X86EMUL_UNHANDLEABLE; + return handler->u.portio.action(IOREQ_WRITE, addr, size, &val); } -static int process_portio_intercept(portio_action_t action, ioreq_t *p) +static const struct hvm_io_ops portio_ops = { + .accept = hvm_portio_accept, + .read = hvm_portio_read, + .write = hvm_portio_write +}; + +static int hvm_process_io_intercept(struct hvm_io_handler *handler, + ioreq_t *p) { struct hvm_vcpu_io *vio = ¤t->arch.hvm_vcpu.hvm_io; + const struct hvm_io_ops *ops = + (p->type == IOREQ_TYPE_COPY) ? + &mmio_ops : + &portio_ops; int rc = X86EMUL_OKAY, i, step = p->df ? -p->size : p->size; - uint32_t data; - - if ( !p->data_is_ptr ) - { - if ( p->dir == IOREQ_READ ) - { - if ( vio->mmio_retrying ) - { - if ( vio->mmio_large_read_bytes != p->size ) - return X86EMUL_UNHANDLEABLE; - memcpy(&data, vio->mmio_large_read, p->size); - vio->mmio_large_read_bytes = 0; - vio->mmio_retrying = 0; - } - else - rc = action(IOREQ_READ, p->addr, p->size, &data); - p->data = data; - } - else - { - data = p->data; - rc = action(IOREQ_WRITE, p->addr, p->size, &data); - } - return rc; - } + uint64_t data; + uint64_t addr; if ( p->dir == IOREQ_READ ) { @@ -246,31 +138,40 @@ static int process_portio_intercept(portio_action_t action, ioreq_t *p) } else { - rc = action(IOREQ_READ, p->addr, p->size, &data); + addr = (p->type == IOREQ_TYPE_COPY) ? + p->addr + step * i : + p->addr; + rc = ops->read(handler, addr, p->size, &data); if ( rc != X86EMUL_OKAY ) break; } - switch ( hvm_copy_to_guest_phys(p->data + step * i, - &data, p->size) ) + + if ( p->data_is_ptr ) { - case HVMCOPY_okay: - break; - case HVMCOPY_gfn_paged_out: - case HVMCOPY_gfn_shared: - rc = X86EMUL_RETRY; - break; - case HVMCOPY_bad_gfn_to_mfn: - /* Drop the write as real hardware would. */ - continue; - case HVMCOPY_bad_gva_to_gfn: - ASSERT(0); - /* fall through */ - default: - rc = X86EMUL_UNHANDLEABLE; - break; + switch ( hvm_copy_to_guest_phys(p->data + step * i, + &data, p->size) ) + { + case HVMCOPY_okay: + break; + case HVMCOPY_gfn_paged_out: + case HVMCOPY_gfn_shared: + rc = X86EMUL_RETRY; + break; + case HVMCOPY_bad_gfn_to_mfn: + /* Drop the write as real hardware would. */ + continue; + case HVMCOPY_bad_gva_to_gfn: + ASSERT_UNREACHABLE(); + /* fall through */ + default: + rc = X86EMUL_UNHANDLEABLE; + break; + } + if ( rc != X86EMUL_OKAY ) + break; } - if ( rc != X86EMUL_OKAY) - break; + else + p->data = data; } if ( rc == X86EMUL_RETRY ) @@ -284,29 +185,37 @@ static int process_portio_intercept(portio_action_t action, ioreq_t *p) { for ( i = 0; i < p->count; i++ ) { - data = 0; - switch ( hvm_copy_from_guest_phys(&data, p->data + step * i, - p->size) ) + if ( p->data_is_ptr ) { - case HVMCOPY_okay: - break; - case HVMCOPY_gfn_paged_out: - case HVMCOPY_gfn_shared: - rc = X86EMUL_RETRY; - break; - case HVMCOPY_bad_gfn_to_mfn: - data = ~0; - break; - case HVMCOPY_bad_gva_to_gfn: - ASSERT(0); - /* fall through */ - default: - rc = X86EMUL_UNHANDLEABLE; - break; + switch ( hvm_copy_from_guest_phys(&data, p->data + step * i, + p->size) ) + { + case HVMCOPY_okay: + break; + case HVMCOPY_gfn_paged_out: + case HVMCOPY_gfn_shared: + rc = X86EMUL_RETRY; + break; + case HVMCOPY_bad_gfn_to_mfn: + data = ~0; + break; + case HVMCOPY_bad_gva_to_gfn: + ASSERT_UNREACHABLE(); + /* fall through */ + default: + rc = X86EMUL_UNHANDLEABLE; + break; + } + if ( rc != X86EMUL_OKAY ) + break; } - if ( rc != X86EMUL_OKAY ) - break; - rc = action(IOREQ_WRITE, p->addr, p->size, &data); + else + data = p->data; + + addr = (p->type == IOREQ_TYPE_COPY) ? + p->addr + step * i : + p->addr; + rc = ops->write(handler, addr, p->size, data); if ( rc != X86EMUL_OKAY ) break; } @@ -324,78 +233,133 @@ static int process_portio_intercept(portio_action_t action, ioreq_t *p) return rc; } -/* - * Check if the request is handled inside xen - * return value: 0 --not handled; 1 --handled - */ -int hvm_io_intercept(ioreq_t *p, int type) +static struct hvm_io_handler *hvm_find_io_handler(ioreq_t *p) +{ + struct vcpu *curr = current; + struct domain *curr_d = curr->domain; + const struct hvm_io_ops *ops = + (p->type == IOREQ_TYPE_COPY) ? + &mmio_ops : + &portio_ops; + unsigned int i; + + for ( i = 0; i < curr_d->arch.hvm_domain.io_handler_count; i++ ) + { + struct hvm_io_handler *handler = + &curr_d->arch.hvm_domain.io_handler[i]; + uint64_t start, end, count = p->count, size = p->size; + + if ( handler->type != p->type ) + continue; + + switch ( handler->type ) + { + case IOREQ_TYPE_PIO: + start = p->addr; + end = p->addr + size; + break; + case IOREQ_TYPE_COPY: + if ( p->df ) + { + start = (p->addr - (count - 1) * size); + end = p->addr + size; + } + else + { + start = p->addr; + end = p->addr + count * size; + } + break; + default: + BUG(); + } + + if ( ops->accept(handler, start, end - start) ) + return handler; + } + + return NULL; +} + +int hvm_io_intercept(ioreq_t *p) { - struct vcpu *v = current; - struct hvm_io_handler *handler = v->domain->arch.hvm_domain.io_handler; - int i; - unsigned long addr, size; + struct hvm_io_handler *handler; - if ( type == HVM_PORTIO ) + if ( p->type == IOREQ_TYPE_PIO ) { int rc = dpci_ioport_intercept(p); if ( (rc == X86EMUL_OKAY) || (rc == X86EMUL_RETRY) ) return rc; } - - for ( i = 0; i < handler->num_slot; i++ ) + else if ( p->type == IOREQ_TYPE_COPY ) { - if ( type != handler->hdl_list[i].type ) - continue; - addr = handler->hdl_list[i].addr; - size = handler->hdl_list[i].size; - if ( (p->addr >= addr) && - ((p->addr + p->size) <= (addr + size)) ) - { - if ( type == HVM_PORTIO ) - return process_portio_intercept( - handler->hdl_list[i].action.portio, p); + int rc = stdvga_intercept_mmio(p); + if ( (rc == X86EMUL_OKAY) || (rc == X86EMUL_RETRY) ) + return rc; + } - if ( unlikely(p->count > 1) && - (unlikely(p->df) - ? p->addr - (p->count - 1L) * p->size < addr - : p->addr + p->count * 1L * p->size - 1 >= addr + size) ) - p->count = 1; + handler = hvm_find_io_handler(p); - return handler->hdl_list[i].action.mmio(p); - } - } + if ( handler == NULL ) + return X86EMUL_UNHANDLEABLE; - return X86EMUL_UNHANDLEABLE; + return hvm_process_io_intercept(handler, p); } -void register_io_handler( - struct domain *d, unsigned long addr, unsigned long size, - void *action, int type) +static struct hvm_io_handler *hvm_next_io_handler(struct domain *d) { - struct hvm_io_handler *handler = d->arch.hvm_domain.io_handler; - int num = handler->num_slot; + unsigned int i = d->arch.hvm_domain.io_handler_count++; - BUG_ON(num >= MAX_IO_HANDLER); + if (i == NR_IO_HANDLERS) + domain_crash(d); - handler->hdl_list[num].addr = addr; - handler->hdl_list[num].size = size; - handler->hdl_list[num].action.ptr = action; - handler->hdl_list[num].type = type; - handler->num_slot++; + return &d->arch.hvm_domain.io_handler[i]; } -void relocate_io_handler( - struct domain *d, unsigned long old_addr, unsigned long new_addr, - unsigned long size, int type) +void register_mmio_handler(struct domain *d, const struct hvm_mmio_ops *ops) { - struct hvm_io_handler *handler = d->arch.hvm_domain.io_handler; - int i; - - for ( i = 0; i < handler->num_slot; i++ ) - if ( (handler->hdl_list[i].addr == old_addr) && - (handler->hdl_list[i].size == size) && - (handler->hdl_list[i].type == type) ) - handler->hdl_list[i].addr = new_addr; + struct hvm_io_handler *handler = hvm_next_io_handler(d); + + handler->type = IOREQ_TYPE_COPY; + handler->u.mmio.ops = ops; +} + +void register_portio_handler(struct domain *d, uint32_t addr, + uint32_t size, portio_action_t action) +{ + struct hvm_io_handler *handler = hvm_next_io_handler(d); + + handler->type = IOREQ_TYPE_PIO; + handler->u.portio.start = addr; + handler->u.portio.end = addr + size; + handler->u.portio.action = action; +} + +void relocate_portio_handler(struct domain *d, uint32_t old_addr, + uint32_t new_addr, uint32_t size) +{ + ioreq_t p = { + .type = IOREQ_TYPE_PIO, + .addr = old_addr, + .size = size + }; + struct hvm_io_handler *handler = hvm_find_io_handler(&p); + + if ( handler == NULL ) + return; + + handler->u.portio.start = new_addr; + handler->u.portio.end = new_addr + size; +} + +bool_t hvm_mmio_internal(paddr_t gpa) +{ + ioreq_t p = { + .type = IOREQ_TYPE_COPY, + .addr = gpa + }; + + return (hvm_find_io_handler(&p) != NULL); } /* diff --git a/xen/arch/x86/hvm/stdvga.c b/xen/arch/x86/hvm/stdvga.c index 13d1029..dcd532a 100644 --- a/xen/arch/x86/hvm/stdvga.c +++ b/xen/arch/x86/hvm/stdvga.c @@ -547,12 +547,27 @@ static int mmio_move(struct hvm_hw_stdvga *s, ioreq_t *p) return 1; } -static int stdvga_intercept_mmio(ioreq_t *p) +int stdvga_intercept_mmio(ioreq_t *p) { struct domain *d = current->domain; struct hvm_hw_stdvga *s = &d->arch.hvm_domain.stdvga; + uint64_t start, end, count = p->count, size = p->size; int buf = 0, rc; + if ( p->df ) + { + start = (p->addr - (count - 1) * size); + end = p->addr + size; + } + else + { + start = p->addr; + end = p->addr + count * size; + } + + if ( (start < VGA_MEM_BASE) || (end > (VGA_MEM_BASE + VGA_MEM_SIZE)) ) + return X86EMUL_UNHANDLEABLE; + if ( p->size > 8 ) { gdprintk(XENLOG_WARNING, "invalid mmio size %d\n", (int)p->size); @@ -619,9 +634,6 @@ void stdvga_init(struct domain *d) register_portio_handler(d, 0x3c4, 2, stdvga_intercept_pio); /* Graphics registers. */ register_portio_handler(d, 0x3ce, 2, stdvga_intercept_pio); - /* MMIO. */ - register_buffered_io_handler( - d, VGA_MEM_BASE, VGA_MEM_SIZE, stdvga_intercept_mmio); } } @@ -638,3 +650,13 @@ void stdvga_deinit(struct domain *d) s->vram_page[i] = NULL; } } + +/* + * Local variables: + * mode: C + * c-file-style: "BSD" + * c-basic-offset: 4 + * tab-width: 4 + * indent-tabs-mode: nil + * End: + */ diff --git a/xen/arch/x86/hvm/vioapic.c b/xen/arch/x86/hvm/vioapic.c index cbbef9f..9ad909b 100644 --- a/xen/arch/x86/hvm/vioapic.c +++ b/xen/arch/x86/hvm/vioapic.c @@ -250,7 +250,7 @@ static int vioapic_range(struct vcpu *v, unsigned long addr) (addr < vioapic->base_address + VIOAPIC_MEM_LENGTH))); } -const struct hvm_mmio_ops vioapic_mmio_ops = { +static const struct hvm_mmio_ops vioapic_mmio_ops = { .check = vioapic_range, .read = vioapic_read, .write = vioapic_write @@ -456,6 +456,8 @@ int vioapic_init(struct domain *d) d->arch.hvm_domain.vioapic->domain = d; vioapic_reset(d); + register_mmio_handler(d, &vioapic_mmio_ops); + return 0; } diff --git a/xen/arch/x86/hvm/vlapic.c b/xen/arch/x86/hvm/vlapic.c index d5855cc..f2052cf 100644 --- a/xen/arch/x86/hvm/vlapic.c +++ b/xen/arch/x86/hvm/vlapic.c @@ -996,7 +996,7 @@ static int vlapic_range(struct vcpu *v, unsigned long addr) (offset < PAGE_SIZE); } -const struct hvm_mmio_ops vlapic_mmio_ops = { +static const struct hvm_mmio_ops vlapic_mmio_ops = { .check = vlapic_range, .read = vlapic_read, .write = vlapic_write @@ -1462,6 +1462,9 @@ int vlapic_init(struct vcpu *v) vlapic_init_sipi_action, (unsigned long)v); + if ( v->vcpu_id == 0 ) + register_mmio_handler(v->domain, &vlapic_mmio_ops); + return 0; } diff --git a/xen/arch/x86/hvm/vmsi.c b/xen/arch/x86/hvm/vmsi.c index f89233d..09ea301 100644 --- a/xen/arch/x86/hvm/vmsi.c +++ b/xen/arch/x86/hvm/vmsi.c @@ -344,7 +344,7 @@ static int msixtbl_range(struct vcpu *v, unsigned long addr) return !!desc; } -const struct hvm_mmio_ops msixtbl_mmio_ops = { +static const struct hvm_mmio_ops msixtbl_mmio_ops = { .check = msixtbl_range, .read = msixtbl_read, .write = msixtbl_write @@ -481,6 +481,11 @@ found: spin_unlock_irq(&irq_desc->lock); } +void msixtbl_init(struct domain *d) +{ + register_mmio_handler(d, &msixtbl_mmio_ops); +} + void msixtbl_pt_cleanup(struct domain *d) { struct msixtbl_entry *entry, *temp; diff --git a/xen/drivers/passthrough/amd/iommu_guest.c b/xen/drivers/passthrough/amd/iommu_guest.c index 7b0c102..9c3c488 100644 --- a/xen/drivers/passthrough/amd/iommu_guest.c +++ b/xen/drivers/passthrough/amd/iommu_guest.c @@ -868,6 +868,20 @@ static void guest_iommu_reg_init(struct guest_iommu *iommu) iommu->reg_ext_feature.hi = upper; } +static int guest_iommu_mmio_range(struct vcpu *v, unsigned long addr) +{ + struct guest_iommu *iommu = vcpu_iommu(v); + + return iommu && addr >= iommu->mmio_base && + addr < iommu->mmio_base + IOMMU_MMIO_SIZE; +} + +static const struct hvm_mmio_ops iommu_mmio_ops = { + .check = guest_iommu_mmio_range, + .read = guest_iommu_mmio_read, + .write = guest_iommu_mmio_write +}; + /* Domain specific initialization */ int guest_iommu_init(struct domain* d) { @@ -894,6 +908,8 @@ int guest_iommu_init(struct domain* d) spin_lock_init(&iommu->lock); + register_mmio_handler(d, &iommu_mmio_ops); + return 0; } @@ -910,17 +926,3 @@ void guest_iommu_destroy(struct domain *d) domain_hvm_iommu(d)->arch.g_iommu = NULL; } - -static int guest_iommu_mmio_range(struct vcpu *v, unsigned long addr) -{ - struct guest_iommu *iommu = vcpu_iommu(v); - - return iommu && addr >= iommu->mmio_base && - addr < iommu->mmio_base + IOMMU_MMIO_SIZE; -} - -const struct hvm_mmio_ops iommu_mmio_ops = { - .check = guest_iommu_mmio_range, - .read = guest_iommu_mmio_read, - .write = guest_iommu_mmio_write -}; diff --git a/xen/include/asm-x86/hvm/domain.h b/xen/include/asm-x86/hvm/domain.h index ad68fcf..b612755 100644 --- a/xen/include/asm-x86/hvm/domain.h +++ b/xen/include/asm-x86/hvm/domain.h @@ -94,6 +94,7 @@ struct hvm_domain { struct pl_time pl_time; struct hvm_io_handler *io_handler; + unsigned int io_handler_count; /* Lock protects access to irq, vpic and vioapic. */ spinlock_t irq_lock; diff --git a/xen/include/asm-x86/hvm/io.h b/xen/include/asm-x86/hvm/io.h index f2aaec5..fecd02d 100644 --- a/xen/include/asm-x86/hvm/io.h +++ b/xen/include/asm-x86/hvm/io.h @@ -25,10 +25,7 @@ #include <public/hvm/ioreq.h> #include <public/event_channel.h> -#define MAX_IO_HANDLER 16 - -#define HVM_PORTIO 0 -#define HVM_BUFFERED_IO 2 +#define NR_IO_HANDLERS 32 typedef int (*hvm_mmio_read_t)(struct vcpu *v, unsigned long addr, @@ -40,81 +37,57 @@ typedef int (*hvm_mmio_write_t)(struct vcpu *v, unsigned long val); typedef int (*hvm_mmio_check_t)(struct vcpu *v, unsigned long addr); -typedef int (*portio_action_t)( - int dir, uint32_t port, uint32_t bytes, uint32_t *val); -typedef int (*mmio_action_t)(ioreq_t *); -struct io_handler { - int type; - unsigned long addr; - unsigned long size; - union { - portio_action_t portio; - mmio_action_t mmio; - void *ptr; - } action; -}; - -struct hvm_io_handler { - int num_slot; - struct io_handler hdl_list[MAX_IO_HANDLER]; -}; - struct hvm_mmio_ops { hvm_mmio_check_t check; hvm_mmio_read_t read; hvm_mmio_write_t write; }; -extern const struct hvm_mmio_ops hpet_mmio_ops; -extern const struct hvm_mmio_ops vlapic_mmio_ops; -extern const struct hvm_mmio_ops vioapic_mmio_ops; -extern const struct hvm_mmio_ops msixtbl_mmio_ops; -extern const struct hvm_mmio_ops iommu_mmio_ops; - -#define HVM_MMIO_HANDLER_NR 5 +typedef int (*portio_action_t)( + int dir, uint32_t port, uint32_t bytes, uint32_t *val); -int hvm_io_intercept(ioreq_t *p, int type); -void register_io_handler( - struct domain *d, unsigned long addr, unsigned long size, - void *action, int type); -void relocate_io_handler( - struct domain *d, unsigned long old_addr, unsigned long new_addr, - unsigned long size, int type); +struct hvm_io_handler { + union { + struct { + const struct hvm_mmio_ops *ops; + } mmio; + struct { + uint32_t start, end; + portio_action_t action; + } portio; + } u; + uint8_t type; +}; -static inline int hvm_portio_intercept(ioreq_t *p) -{ - return hvm_io_intercept(p, HVM_PORTIO); -} +typedef int (*hvm_io_read_t)(struct hvm_io_handler *, + uint64_t addr, + uint64_t size, + uint64_t *data); +typedef int (*hvm_io_write_t)(struct hvm_io_handler *, + uint64_t addr, + uint64_t size, + uint64_t data); +typedef bool_t (*hvm_io_accept_t)(struct hvm_io_handler *, + uint64_t addr, + uint64_t size); + +struct hvm_io_ops { + hvm_io_accept_t accept; + hvm_io_read_t read; + hvm_io_write_t write; +}; -static inline int hvm_buffered_io_intercept(ioreq_t *p) -{ - return hvm_io_intercept(p, HVM_BUFFERED_IO); -} +int hvm_io_intercept(ioreq_t *p); +void register_mmio_handler(struct domain *d, + const struct hvm_mmio_ops *ops); +void register_portio_handler(struct domain *d, uint32_t addr, + uint32_t size, portio_action_t action); +void relocate_portio_handler(struct domain *d, uint32_t old_addr, + uint32_t new_addr, uint32_t size); bool_t hvm_mmio_internal(paddr_t gpa); -int hvm_mmio_intercept(ioreq_t *p); -int hvm_buffered_io_send(ioreq_t *p); -static inline void register_portio_handler( - struct domain *d, unsigned long addr, - unsigned long size, portio_action_t action) -{ - register_io_handler(d, addr, size, action, HVM_PORTIO); -} - -static inline void relocate_portio_handler( - struct domain *d, unsigned long old_addr, unsigned long new_addr, - unsigned long size) -{ - relocate_io_handler(d, old_addr, new_addr, size, HVM_PORTIO); -} - -static inline void register_buffered_io_handler( - struct domain *d, unsigned long addr, - unsigned long size, mmio_action_t action) -{ - register_io_handler(d, addr, size, action, HVM_BUFFERED_IO); -} +int hvm_buffered_io_send(ioreq_t *p); void send_timeoffset_req(unsigned long timeoff); void send_invalidate_req(void); @@ -127,6 +100,7 @@ void hvm_io_assist(ioreq_t *p); void hvm_dpci_eoi(struct domain *d, unsigned int guest_irq, const union vioapic_redir_entry *ent); void msix_write_completion(struct vcpu *); +void msixtbl_init(struct domain *d); struct hvm_hw_stdvga { uint8_t sr_index; @@ -141,8 +115,19 @@ struct hvm_hw_stdvga { }; void stdvga_init(struct domain *d); +int stdvga_intercept_mmio(ioreq_t *p); void stdvga_deinit(struct domain *d); extern void hvm_dpci_msi_eoi(struct domain *d, int vector); #endif /* __ASM_X86_HVM_IO_H__ */ + +/* + * Local variables: + * mode: C + * c-file-style: "BSD" + * c-basic-offset: 4 + * tab-width: 4 + * indent-tabs-mode: nil + * End: + */ -- 1.7.10.4 _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |