|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH v6 16/16] x86/hvm: track large memory mapped accesses by buffer offset
>>> On 03.07.15 at 18:25, <paul.durrant@xxxxxxxxxx> wrote:
> @@ -635,13 +605,41 @@ static int hvmemul_phys_mmio_access(
> return rc;
> }
>
> +static struct hvm_mmio_cache *hvmemul_find_mmio_cache(
> + struct hvm_vcpu_io *vio, unsigned long gla, uint8_t dir)
> +{
> + unsigned int i;
> + struct hvm_mmio_cache *cache;
> +
> + for ( i = 0; i < vio->mmio_cache_count; i ++ )
> + {
> + cache = &vio->mmio_cache[i];
> +
> + if ( gla == cache->gla &&
> + dir == cache->dir )
> + return cache;
> + }
> +
> + i = vio->mmio_cache_count++;
> + BUG_ON(i == ARRAY_SIZE(vio->mmio_cache));
> +
> + cache = &vio->mmio_cache[i];
> + memset(cache, 0, sizeof (*cache));
> +
> + cache->gla = gla;
> + cache->dir = dir;
> +
> + return cache;
> +}
There's still a weakness here (but that also applies if you used physical
addresses): Multiple reads to the same address may not return the
same result. I don't think this needs to be addressed here, but adding
a comment clarifying that this case isn't handled correctly would help
future readers easier understand the state of affairs.
As to the BUG_ON() - I think this would better be a domain_crash(),
just in case we overlooked an exotic instruction accessing more than
3 memory locations.
Jan
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |