[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH v6 02/16] x86/hvm: remove multiple open coded 'chunking' loops
> -----Original Message----- > From: Jan Beulich [mailto:JBeulich@xxxxxxxx] > Sent: 09 July 2015 07:54 > To: Andrew Cooper > Cc: Paul Durrant; xen-devel@xxxxxxxxxxxxxxxxxxxx; Keir (Xen.org) > Subject: Re: [Xen-devel] [PATCH v6 02/16] x86/hvm: remove multiple open > coded 'chunking' loops > > >>> On 08.07.15 at 18:43, <andrew.cooper3@xxxxxxxxxx> wrote: > > On 08/07/2015 16:57, Paul Durrant wrote: > >>> -----Original Message----- > >>> From: xen-devel-bounces@xxxxxxxxxxxxx [mailto:xen-devel- > >>> bounces@xxxxxxxxxxxxx] On Behalf Of Jan Beulich > >>> Sent: 08 July 2015 16:53 > >>> To: Andrew Cooper; Paul Durrant > >>> Cc: xen-devel@xxxxxxxxxxxxxxxxxxxx; Keir (Xen.org) > >>> Subject: Re: [Xen-devel] [PATCH v6 02/16] x86/hvm: remove multiple > open > >>> coded 'chunking' loops > >>> > >>>>>> On 03.07.15 at 18:25, <paul.durrant@xxxxxxxxxx> wrote: > >>>> +static int hvmemul_linear_mmio_access( > >>>> + unsigned long gla, unsigned int size, uint8_t dir, uint8_t *buffer, > >>>> + uint32_t pfec, struct hvm_emulate_ctxt *hvmemul_ctxt, bool_t > >>> known_gpfn) > >>>> +{ > >>>> + struct hvm_vcpu_io *vio = ¤t->arch.hvm_vcpu.hvm_io; > >>>> + unsigned long offset = gla & ~PAGE_MASK; > >>>> + unsigned int chunk; > >>>> + paddr_t gpa; > >>>> + unsigned long one_rep = 1; > >>>> + int rc; > >>>> + > >>>> + chunk = min_t(unsigned int, size, PAGE_SIZE - offset); > >>>> + > >>>> + if ( known_gpfn ) > >>>> + gpa = pfn_to_paddr(vio->mmio_gpfn) | offset; > >>>> + else > >>>> + { > >>>> + rc = hvmemul_linear_to_phys(gla, &gpa, chunk, &one_rep, pfec, > >>>> + hvmemul_ctxt); > >>>> + if ( rc != X86EMUL_OKAY ) > >>>> + return rc; > >>>> + } > >>>> + > >>>> + for ( ;; ) > >>>> + { > >>>> + rc = hvmemul_phys_mmio_access(gpa, chunk, dir, buffer); > >>>> + if ( rc != X86EMUL_OKAY ) > >>>> + break; > >>>> + > >>>> + gla += chunk; > >>>> + buffer += chunk; > >>>> + size -= chunk; > >>>> + > >>>> + if ( size == 0 ) > >>>> + break; > >>>> + > >>>> + ASSERT((gla & ~PAGE_MASK) == 0); > >>> Does this really matter for the code below? > >>> > >>>> + chunk = min_t(unsigned int, size, PAGE_SIZE); > >>> Iirc Andrew had asked for this, but I still don't see why: "size" is the > >>> width of an instruction operand, and hence won't even come close > >>> to PAGE_SIZE. > > > > The original version of the code asserted that size was less than > > PAGE_SIZE around here. This is not true in the general case, given a > > for loop like this and can in principle be hit if we ever got into a > > position of emulating an xsave instruction to an MMIO region. > > That's a hypothetical future extension to XSAVE, isn't it? I.e. only > when AVX-1024 eventually arrives ... > > > This specific example is not as far fetched as it seems. The VM > > instrospection people are looking to emulate more and more instructions, > > while the GVT-g are working on the mmio_write_dm side of things which > > causes real RAM to be treated as MMIO from Xens point of view. > > ... and when we really think we need to support XSAVEing to MMIO > or MMIO-like memory (which I doubt we will). > > > If we had some blanket sanity checks for size at the top of the > > emulation calltree it would be less of an issue, but we don't and I have > > a nagging feeing that assumptions like this are going to bite us in an > > XSA-kind-of-way. > > Hence the ASSERT() against PAGE_SIZE that originally was there. > Is there any harm in leaving the function as-is so that it can cope with size > PAGE_SIZE in future? Paul > Jan _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |