[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-devel] [PATCH v6 02/16] x86/hvm: remove multiple open coded 'chunking' loops



>>> On 08.07.15 at 18:43, <andrew.cooper3@xxxxxxxxxx> wrote:
> On 08/07/2015 16:57, Paul Durrant wrote:
>>> -----Original Message-----
>>> From: xen-devel-bounces@xxxxxxxxxxxxx [mailto:xen-devel-
>>> bounces@xxxxxxxxxxxxx] On Behalf Of Jan Beulich
>>> Sent: 08 July 2015 16:53
>>> To: Andrew Cooper; Paul Durrant
>>> Cc: xen-devel@xxxxxxxxxxxxxxxxxxxx; Keir (Xen.org)
>>> Subject: Re: [Xen-devel] [PATCH v6 02/16] x86/hvm: remove multiple open
>>> coded 'chunking' loops
>>>
>>>>>> On 03.07.15 at 18:25, <paul.durrant@xxxxxxxxxx> wrote:
>>>> +static int hvmemul_linear_mmio_access(
>>>> +    unsigned long gla, unsigned int size, uint8_t dir, uint8_t *buffer,
>>>> +    uint32_t pfec, struct hvm_emulate_ctxt *hvmemul_ctxt, bool_t
>>> known_gpfn)
>>>> +{
>>>> +    struct hvm_vcpu_io *vio = &current->arch.hvm_vcpu.hvm_io;
>>>> +    unsigned long offset = gla & ~PAGE_MASK;
>>>> +    unsigned int chunk;
>>>> +    paddr_t gpa;
>>>> +    unsigned long one_rep = 1;
>>>> +    int rc;
>>>> +
>>>> +    chunk = min_t(unsigned int, size, PAGE_SIZE - offset);
>>>> +
>>>> +    if ( known_gpfn )
>>>> +        gpa = pfn_to_paddr(vio->mmio_gpfn) | offset;
>>>> +    else
>>>> +    {
>>>> +        rc = hvmemul_linear_to_phys(gla, &gpa, chunk, &one_rep, pfec,
>>>> +                                    hvmemul_ctxt);
>>>> +        if ( rc != X86EMUL_OKAY )
>>>> +            return rc;
>>>> +    }
>>>> +
>>>> +    for ( ;; )
>>>> +    {
>>>> +        rc = hvmemul_phys_mmio_access(gpa, chunk, dir, buffer);
>>>> +        if ( rc != X86EMUL_OKAY )
>>>> +            break;
>>>> +
>>>> +        gla += chunk;
>>>> +        buffer += chunk;
>>>> +        size -= chunk;
>>>> +
>>>> +        if ( size == 0 )
>>>> +            break;
>>>> +
>>>> +        ASSERT((gla & ~PAGE_MASK) == 0);
>>> Does this really matter for the code below?
>>>
>>>> +        chunk = min_t(unsigned int, size, PAGE_SIZE);
>>> Iirc Andrew had asked for this, but I still don't see why: "size" is the
>>> width of an instruction operand, and hence won't even come close
>>> to PAGE_SIZE.
> 
> The original version of the code asserted that size was less than
> PAGE_SIZE around here.  This is not true in the general case, given a
> for loop like this and can in principle be hit if we ever got into a
> position of emulating an xsave instruction to an MMIO region.

That's a hypothetical future extension to XSAVE, isn't it? I.e. only
when AVX-1024 eventually arrives ...

> This specific example is not as far fetched as it seems.  The VM
> instrospection people are looking to emulate more and more instructions,
> while the GVT-g are working on the mmio_write_dm side of things which
> causes real RAM to be treated as MMIO from Xens point of view.

... and when we really think we need to support XSAVEing to MMIO
or MMIO-like memory (which I doubt we will).

> If we had some blanket sanity checks for size at the top of the
> emulation calltree it would be less of an issue, but we don't and I have
> a nagging feeing that assumptions like this are going to bite us in an
> XSA-kind-of-way.

Hence the ASSERT() against PAGE_SIZE that originally was there.

Jan


_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.