[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH 1/5] arm: shared_info page allocation and mapping
On Thu, 2012-02-23 at 17:13 +0000, Stefano Stabellini wrote: > Allocate the shared_info page at domain creation. > > Implement arch_memory_op, only for XENMEM_add_to_physmap with space == > XENMAPSPACE_shared_info, so that the guest can map the shared_info page. > > Signed-off-by: Stefano Stabellini <stefano.stabellini@xxxxxxxxxxxxx> > --- > xen/arch/arm/domain.c | 8 ++++ > xen/arch/arm/mm.c | 98 > +++++++++++++++++++++++++++++++++++++++++++-- > xen/arch/arm/p2m.c | 15 ++++++- > xen/include/asm-arm/mm.h | 4 ++ > xen/include/asm-arm/p2m.h | 2 + > 5 files changed, 122 insertions(+), 5 deletions(-) > > diff --git a/xen/arch/arm/domain.c b/xen/arch/arm/domain.c > index 0b55934..1e5cca5 100644 > --- a/xen/arch/arm/domain.c > +++ b/xen/arch/arm/domain.c > @@ -235,6 +235,14 @@ int arch_domain_create(struct domain *d, unsigned int > domcr_flags) > if ( (rc = p2m_init(d)) != 0 ) > goto fail; > > + rc = -ENOMEM; > + if ( (d->shared_info = alloc_xenheap_pages(0, MEMF_bits(32))) == NULL ) > + goto fail; > + > + clear_page(d->shared_info); > + share_xen_page_with_guest( > + virt_to_page(d->shared_info), d, XENSHARE_writable); You seem to have some hard tabs here. > + > d->max_vcpus = 8; > > if ( (rc = domain_vgic_init(d)) != 0 ) > diff --git a/xen/arch/arm/mm.c b/xen/arch/arm/mm.c > index a0f39eb..5f4fd6a 100644 > --- a/xen/arch/arm/mm.c > +++ b/xen/arch/arm/mm.c > @@ -25,8 +25,11 @@ > #include <xen/mm.h> > #include <xen/preempt.h> > #include <xen/errno.h> > +#include <xen/guest_access.h> > #include <asm/page.h> > #include <asm/current.h> > +#include <public/memory.h> > +#include <xen/sched.h> > > struct domain *dom_xen, *dom_io; > > @@ -323,17 +326,104 @@ void arch_dump_shared_mem_info(void) > { > } > > -long arch_memory_op(int op, XEN_GUEST_HANDLE(void) arg) > +int donate_page(struct domain *d, struct page_info *page, unsigned int > memflags) > { > + ASSERT(0); > return -ENOSYS; > } > > -int donate_page(struct domain *d, struct page_info *page, unsigned int > memflags) > +void share_xen_page_with_guest(struct page_info *page, > + struct domain *d, int readonly) > { > - ASSERT(0); > - return -ENOSYS; > + if ( page_get_owner(page) == d ) > + return; > + > + spin_lock(&d->page_alloc_lock); > + > + /* The incremented type count pins as writable or read-only. */ > + page->u.inuse.type_info = (readonly ? PGT_none : PGT_writable_page); > + page->u.inuse.type_info |= PGT_validated | 1; > + > + page_set_owner(page, d); > + wmb(); /* install valid domain ptr before updating refcnt. */ > + ASSERT((page->count_info & ~PGC_xen_heap) == 0); > + > + /* Only add to the allocation list if the domain isn't dying. */ > + if ( !d->is_dying ) > + { > + page->count_info |= PGC_allocated | 1; > + if ( unlikely(d->xenheap_pages++ == 0) ) > + get_knownalive_domain(d); > + page_list_add_tail(page, &d->xenpage_list); > + } > + > + spin_unlock(&d->page_alloc_lock); > +} > + > +static int xenmem_add_to_physmap_once( > + struct domain *d, > + const struct xen_add_to_physmap *xatp) > +{ > + unsigned long mfn = 0; > + int rc; > + > + switch ( xatp->space ) > + { > + case XENMAPSPACE_shared_info: > + if ( xatp->idx == 0 ) > + mfn = virt_to_mfn(d->shared_info); > + break; > + default: > + return -ENOSYS; Another w/s snafu. > + } > + > + domain_lock(d); > + > + /* Map at new location. */ > + rc = guest_physmap_add_page(d, xatp->gpfn, mfn); > + > + domain_unlock(d); > + > + return rc; > +} > + > +static int xenmem_add_to_physmap(struct domain *d, > + struct xen_add_to_physmap *xatp) > +{ > + return xenmem_add_to_physmap_once(d, xatp); > } > > +long arch_memory_op(int op, XEN_GUEST_HANDLE(void) arg) > +{ > + int rc; > + > + switch ( op ) > + { > + case XENMEM_add_to_physmap: > + { > + struct xen_add_to_physmap xatp; > + struct domain *d; > + > + if ( copy_from_guest(&xatp, arg, 1) ) > + return -EFAULT; > + > + rc = rcu_lock_target_domain_by_id(xatp.domid, &d); > + if ( rc != 0 ) > + return rc; > + > + rc = xenmem_add_to_physmap(d, &xatp); > + > + rcu_unlock_domain(d); > + > + return rc; > + } > + > + default: > + return -ENOSYS; > + } > + > + return 0; > +} > /* > * Local variables: > * mode: C > diff --git a/xen/arch/arm/p2m.c b/xen/arch/arm/p2m.c > index 14614fd..6ee1b5f 100644 > --- a/xen/arch/arm/p2m.c > +++ b/xen/arch/arm/p2m.c > @@ -118,7 +118,12 @@ static int create_p2m_entries(struct domain *d, > } > /* else: third already valid */ > > - BUG_ON(third[third_table_offset(addr)].p2m.valid); > + if ( third[third_table_offset(addr)].p2m.valid ) > + { > + /* p2m entry already present */ > + free_domheap_page( > + > mfn_to_page(third[third_table_offset(addr)].p2m.base)); > + } Guess what ;) > > /* Allocate a new RAM page and attach */ > if (alloc) > @@ -172,6 +177,14 @@ int map_mmio_regions(struct domain *d, > return create_p2m_entries(d, 0, start_gaddr, end_gaddr, maddr); > } > > +int guest_physmap_add_page(struct domain *d, unsigned long gpfn, > + unsigned long mfn) > +{ > + return create_p2m_entries(d, 0, gpfn << PAGE_SHIFT, > + (gpfn + 1) << PAGE_SHIFT, > + mfn << PAGE_SHIFT); Not sure if this is a hardspace issue or just a strange way of laying it out? > +} > + > int p2m_alloc_table(struct domain *d) > { > struct p2m_domain *p2m = &d->arch.p2m; > diff --git a/xen/include/asm-arm/mm.h b/xen/include/asm-arm/mm.h > index bfc0f76..56ab9415 100644 > --- a/xen/include/asm-arm/mm.h > +++ b/xen/include/asm-arm/mm.h > @@ -78,6 +78,10 @@ struct page_info > #define _PGT_pinned PG_shift(5) > #define PGT_pinned PG_mask(1, 5) > > + /* Has this page been validated for use as its current type? */ > +#define _PGT_validated PG_shift(6) > +#define PGT_validated PG_mask(1, 6) > + > /* Count of uses of this frame as its current type. */ > #define PGT_count_width PG_shift(9) > #define PGT_count_mask ((1UL<<PGT_count_width)-1) > diff --git a/xen/include/asm-arm/p2m.h b/xen/include/asm-arm/p2m.h > index aec52f7..b1d42a8 100644 > --- a/xen/include/asm-arm/p2m.h > +++ b/xen/include/asm-arm/p2m.h > @@ -39,6 +39,8 @@ int p2m_populate_ram(struct domain *d, paddr_t start, > paddr_t end); > * address maddr. */ > int map_mmio_regions(struct domain *d, paddr_t start_gaddr, > paddr_t end_gaddr, paddr_t maddr); > +int guest_physmap_add_page(struct domain *d, unsigned long gpfn, > + unsigned long mfn); > > unsigned long gmfn_to_mfn(struct domain *d, unsigned long gpfn); > _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |