[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH v4 3/5] xen/arm: clean and invalidate all guest caches by VMID after domain build.
On Fri, 2014-02-07 at 14:49 +0000, Stefano Stabellini wrote: > > diff --git a/tools/libxc/xc_dom_core.c b/tools/libxc/xc_dom_core.c > > index 77a4e64..b9d1015 100644 > > --- a/tools/libxc/xc_dom_core.c > > +++ b/tools/libxc/xc_dom_core.c > > @@ -603,6 +603,8 @@ void xc_dom_unmap_one(struct xc_dom_image *dom, > > xen_pfn_t pfn) > > prev->next = phys->next; > > else > > dom->phys_pages = phys->next; > > + > > + xc_domain_cacheflush(dom->xch, dom->guest_domid, phys->first, > > phys->count); > > I am not sure I understand the semantic of xc_dom_unmap_one: from the > name of the function and the parameters I would think that it is > supposed to unmap just one page. In that case we should flush just one > page. However the implementation calls > > munmap(phys->ptr, phys->count << page_shift) > > so it actually can unmap more than a single page, so I think that you > did the right thing by calling xc_domain_cacheflush with phys->first and > phys->count. "one" in the name means "one struct xc_dom_phys", which is a range starting at the given pfn, which is why using phys->{first,count} is correct. > > +void sync_page_to_ram(unsigned long mfn) > > +{ > > + void *p, *v = map_domain_page(mfn); > > + > > + dsb(); /* So the CPU issues all writes to the range */ > > + for ( p = v; p < v + PAGE_SIZE ; p += cacheline_bytes ) > > What about the last few bytes on a page? > Can we assume that PAGE_SIZE is a multiple of cacheline_bytes? I sure hope so! A non power of two cache line size would be pretty crazy! The cacheline is always a 2^N (the register contains log2(cacheline size)) and so is PAGE_SIZE. Ian. _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |