[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH v4 04/14] x86/paging: introduce paging_set_allocation
... and remove hap_set_alloc_for_pvh_dom0. While there also change the last parameter of the {hap/shadow}_set_allocation functions to be a boolean. Signed-off-by: Roger Pau Monné <roger.pau@xxxxxxxxxx> Acked-by: Tim Deegan <tim@xxxxxxx> Acked-by: George Dunlap <george.dunlap@xxxxxxxxxx> Reviewed-by: Jan Beulich <jbeulich@xxxxxxxx> --- Cc: Jan Beulich <jbeulich@xxxxxxxx> Cc: Andrew Cooper <andrew.cooper3@xxxxxxxxxx> Cc: George Dunlap <george.dunlap@xxxxxxxxxxxxx> Cc: Tim Deegan <tim@xxxxxxx> --- Changes since v3: - Rename sh_set_allocation to shadow_set_allocation (public shadow functions use the shadow prefix instead of sh). Changes since v2: - Convert the preempt parameter into a bool. - Fix Dom0 builder comment to reflect that paging.mode should be correct before calling paging_set_allocation. Changes since RFC: - Make paging_set_allocation preemtable. - Move comments. --- xen/arch/x86/domain_build.c | 21 +++++++++++++++------ xen/arch/x86/mm/hap/hap.c | 22 +++++----------------- xen/arch/x86/mm/paging.c | 19 ++++++++++++++++++- xen/arch/x86/mm/shadow/common.c | 31 +++++++++++++------------------ xen/include/asm-x86/hap.h | 4 ++-- xen/include/asm-x86/paging.h | 7 +++++++ xen/include/asm-x86/shadow.h | 11 ++++++++++- 7 files changed, 70 insertions(+), 45 deletions(-) diff --git a/xen/arch/x86/domain_build.c b/xen/arch/x86/domain_build.c index 0a02d65..17f8e91 100644 --- a/xen/arch/x86/domain_build.c +++ b/xen/arch/x86/domain_build.c @@ -35,7 +35,6 @@ #include <asm/setup.h> #include <asm/bzimage.h> /* for bzimage_parse */ #include <asm/io_apic.h> -#include <asm/hap.h> #include <asm/hpet.h> #include <public/version.h> @@ -1383,15 +1382,25 @@ int __init construct_dom0( nr_pages); } - if ( is_pvh_domain(d) ) - hap_set_alloc_for_pvh_dom0(d, dom0_paging_pages(d, nr_pages)); - /* - * We enable paging mode again so guest_physmap_add_page will do the - * right thing for us. + * We enable paging mode again so guest_physmap_add_page and + * paging_set_allocation will do the right thing for us. */ d->arch.paging.mode = save_pvh_pg_mode; + if ( is_pvh_domain(d) ) + { + bool preempted; + + do { + preempted = false; + paging_set_allocation(d, dom0_paging_pages(d, nr_pages), + &preempted); + process_pending_softirqs(); + } while ( preempted ); + } + + /* Write the phys->machine and machine->phys table entries. */ for ( pfn = 0; pfn < count; pfn++ ) { diff --git a/xen/arch/x86/mm/hap/hap.c b/xen/arch/x86/mm/hap/hap.c index b9faba6..e6dc088 100644 --- a/xen/arch/x86/mm/hap/hap.c +++ b/xen/arch/x86/mm/hap/hap.c @@ -334,8 +334,7 @@ hap_get_allocation(struct domain *d) /* Set the pool of pages to the required number of pages. * Returns 0 for success, non-zero for failure. */ -static int -hap_set_allocation(struct domain *d, unsigned int pages, int *preempted) +int hap_set_allocation(struct domain *d, unsigned int pages, bool *preempted) { struct page_info *pg; @@ -381,7 +380,7 @@ hap_set_allocation(struct domain *d, unsigned int pages, int *preempted) /* Check to see if we need to yield and try again */ if ( preempted && general_preempt_check() ) { - *preempted = 1; + *preempted = true; return 0; } } @@ -561,7 +560,7 @@ void hap_final_teardown(struct domain *d) paging_unlock(d); } -void hap_teardown(struct domain *d, int *preempted) +void hap_teardown(struct domain *d, bool *preempted) { struct vcpu *v; mfn_t mfn; @@ -609,7 +608,8 @@ out: int hap_domctl(struct domain *d, xen_domctl_shadow_op_t *sc, XEN_GUEST_HANDLE_PARAM(void) u_domctl) { - int rc, preempted = 0; + int rc; + bool preempted = false; switch ( sc->op ) { @@ -636,18 +636,6 @@ int hap_domctl(struct domain *d, xen_domctl_shadow_op_t *sc, } } -void __init hap_set_alloc_for_pvh_dom0(struct domain *d, - unsigned long hap_pages) -{ - int rc; - - paging_lock(d); - rc = hap_set_allocation(d, hap_pages, NULL); - paging_unlock(d); - - BUG_ON(rc); -} - static const struct paging_mode hap_paging_real_mode; static const struct paging_mode hap_paging_protected_mode; static const struct paging_mode hap_paging_pae_mode; diff --git a/xen/arch/x86/mm/paging.c b/xen/arch/x86/mm/paging.c index cc44682..853a035 100644 --- a/xen/arch/x86/mm/paging.c +++ b/xen/arch/x86/mm/paging.c @@ -809,7 +809,8 @@ long paging_domctl_continuation(XEN_GUEST_HANDLE_PARAM(xen_domctl_t) u_domctl) /* Call when destroying a domain */ int paging_teardown(struct domain *d) { - int rc, preempted = 0; + int rc; + bool preempted = false; if ( hap_enabled(d) ) hap_teardown(d, &preempted); @@ -954,6 +955,22 @@ void paging_write_p2m_entry(struct p2m_domain *p2m, unsigned long gfn, safe_write_pte(p, new); } +int paging_set_allocation(struct domain *d, unsigned int pages, bool *preempted) +{ + int rc; + + ASSERT(paging_mode_enabled(d)); + + paging_lock(d); + if ( hap_enabled(d) ) + rc = hap_set_allocation(d, pages, preempted); + else + rc = shadow_set_allocation(d, pages, preempted); + paging_unlock(d); + + return rc; +} + /* * Local variables: * mode: C diff --git a/xen/arch/x86/mm/shadow/common.c b/xen/arch/x86/mm/shadow/common.c index ddbdb73..9f3bed9 100644 --- a/xen/arch/x86/mm/shadow/common.c +++ b/xen/arch/x86/mm/shadow/common.c @@ -1611,13 +1611,7 @@ shadow_free_p2m_page(struct domain *d, struct page_info *pg) paging_unlock(d); } -/* Set the pool of shadow pages to the required number of pages. - * Input will be rounded up to at least shadow_min_acceptable_pages(), - * plus space for the p2m table. - * Returns 0 for success, non-zero for failure. */ -static int sh_set_allocation(struct domain *d, - unsigned int pages, - int *preempted) +int shadow_set_allocation(struct domain *d, unsigned int pages, bool *preempted) { struct page_info *sp; unsigned int lower_bound; @@ -1683,7 +1677,7 @@ static int sh_set_allocation(struct domain *d, /* Check to see if we need to yield and try again */ if ( preempted && general_preempt_check() ) { - *preempted = 1; + *preempted = true; return 0; } } @@ -3154,10 +3148,10 @@ int shadow_enable(struct domain *d, u32 mode) if ( old_pages == 0 ) { paging_lock(d); - rv = sh_set_allocation(d, 1024, NULL); /* Use at least 4MB */ + rv = shadow_set_allocation(d, 1024, NULL); /* Use at least 4MB */ if ( rv != 0 ) { - sh_set_allocation(d, 0, NULL); + shadow_set_allocation(d, 0, NULL); goto out_locked; } paging_unlock(d); @@ -3239,7 +3233,7 @@ int shadow_enable(struct domain *d, u32 mode) return rv; } -void shadow_teardown(struct domain *d, int *preempted) +void shadow_teardown(struct domain *d, bool *preempted) /* Destroy the shadow pagetables of this domain and free its shadow memory. * Should only be called for dying domains. */ { @@ -3301,7 +3295,7 @@ void shadow_teardown(struct domain *d, int *preempted) if ( d->arch.paging.shadow.total_pages != 0 ) { /* Destroy all the shadows and release memory to domheap */ - sh_set_allocation(d, 0, preempted); + shadow_set_allocation(d, 0, preempted); if ( preempted && *preempted ) goto out; @@ -3366,7 +3360,7 @@ void shadow_final_teardown(struct domain *d) p2m_teardown(p2m_get_hostp2m(d)); /* Free any shadow memory that the p2m teardown released */ paging_lock(d); - sh_set_allocation(d, 0, NULL); + shadow_set_allocation(d, 0, NULL); SHADOW_PRINTK("dom %u final teardown done." " Shadow pages total = %u, free = %u, p2m=%u\n", d->domain_id, @@ -3392,9 +3386,9 @@ static int shadow_one_bit_enable(struct domain *d, u32 mode) if ( d->arch.paging.shadow.total_pages == 0 ) { /* Init the shadow memory allocation if the user hasn't done so */ - if ( sh_set_allocation(d, 1, NULL) != 0 ) + if ( shadow_set_allocation(d, 1, NULL) != 0 ) { - sh_set_allocation(d, 0, NULL); + shadow_set_allocation(d, 0, NULL); return -ENOMEM; } } @@ -3463,7 +3457,7 @@ static int shadow_one_bit_disable(struct domain *d, u32 mode) } /* Pull down the memory allocation */ - if ( sh_set_allocation(d, 0, NULL) != 0 ) + if ( shadow_set_allocation(d, 0, NULL) != 0 ) BUG(); /* In fact, we will have BUG()ed already */ shadow_hash_teardown(d); SHADOW_PRINTK("un-shadowing of domain %u done." @@ -3876,7 +3870,8 @@ int shadow_domctl(struct domain *d, xen_domctl_shadow_op_t *sc, XEN_GUEST_HANDLE_PARAM(void) u_domctl) { - int rc, preempted = 0; + int rc; + bool preempted = false; switch ( sc->op ) { @@ -3907,7 +3902,7 @@ int shadow_domctl(struct domain *d, paging_unlock(d); return -EINVAL; } - rc = sh_set_allocation(d, sc->mb << (20 - PAGE_SHIFT), &preempted); + rc = shadow_set_allocation(d, sc->mb << (20 - PAGE_SHIFT), &preempted); paging_unlock(d); if ( preempted ) /* Not finished. Set up to re-run the call. */ diff --git a/xen/include/asm-x86/hap.h b/xen/include/asm-x86/hap.h index c613836..dedb4b1 100644 --- a/xen/include/asm-x86/hap.h +++ b/xen/include/asm-x86/hap.h @@ -38,7 +38,7 @@ int hap_domctl(struct domain *d, xen_domctl_shadow_op_t *sc, XEN_GUEST_HANDLE_PARAM(void) u_domctl); int hap_enable(struct domain *d, u32 mode); void hap_final_teardown(struct domain *d); -void hap_teardown(struct domain *d, int *preempted); +void hap_teardown(struct domain *d, bool *preempted); void hap_vcpu_init(struct vcpu *v); int hap_track_dirty_vram(struct domain *d, unsigned long begin_pfn, @@ -46,7 +46,7 @@ int hap_track_dirty_vram(struct domain *d, XEN_GUEST_HANDLE_64(uint8) dirty_bitmap); extern const struct paging_mode *hap_paging_get_mode(struct vcpu *); -void hap_set_alloc_for_pvh_dom0(struct domain *d, unsigned long num_pages); +int hap_set_allocation(struct domain *d, unsigned int pages, bool *preempted); #endif /* XEN_HAP_H */ diff --git a/xen/include/asm-x86/paging.h b/xen/include/asm-x86/paging.h index 56eef6b..f83ed8b 100644 --- a/xen/include/asm-x86/paging.h +++ b/xen/include/asm-x86/paging.h @@ -347,6 +347,13 @@ void pagetable_dying(struct domain *d, paddr_t gpa); void paging_dump_domain_info(struct domain *d); void paging_dump_vcpu_info(struct vcpu *v); +/* Set the pool of shadow pages to the required number of pages. + * Input might be rounded up to at minimum amount of pages, plus + * space for the p2m table. + * Returns 0 for success, non-zero for failure. */ +int paging_set_allocation(struct domain *d, unsigned int pages, + bool *preempted); + #endif /* XEN_PAGING_H */ /* diff --git a/xen/include/asm-x86/shadow.h b/xen/include/asm-x86/shadow.h index 6d0aefb..bac952f 100644 --- a/xen/include/asm-x86/shadow.h +++ b/xen/include/asm-x86/shadow.h @@ -73,7 +73,7 @@ int shadow_domctl(struct domain *d, XEN_GUEST_HANDLE_PARAM(void) u_domctl); /* Call when destroying a domain */ -void shadow_teardown(struct domain *d, int *preempted); +void shadow_teardown(struct domain *d, bool *preempted); /* Call once all of the references to the domain have gone away */ void shadow_final_teardown(struct domain *d); @@ -83,6 +83,13 @@ void sh_remove_shadows(struct domain *d, mfn_t gmfn, int fast, int all); /* Discard _all_ mappings from the domain's shadows. */ void shadow_blow_tables_per_domain(struct domain *d); +/* Set the pool of shadow pages to the required number of pages. + * Input will be rounded up to at least shadow_min_acceptable_pages(), + * plus space for the p2m table. + * Returns 0 for success, non-zero for failure. */ +int shadow_set_allocation(struct domain *d, unsigned int pages, + bool *preempted); + #else /* !CONFIG_SHADOW_PAGING */ #define shadow_teardown(d, p) ASSERT(is_pv_domain(d)) @@ -91,6 +98,8 @@ void shadow_blow_tables_per_domain(struct domain *d); ({ ASSERT(is_pv_domain(d)); -EOPNOTSUPP; }) #define shadow_track_dirty_vram(d, begin_pfn, nr, bitmap) \ ({ ASSERT_UNREACHABLE(); -EOPNOTSUPP; }) +#define shadow_set_allocation(d, pages, preempted) \ + ({ ASSERT_UNREACHABLE(); -EOPNOTSUPP; }) static inline void sh_remove_shadows(struct domain *d, mfn_t gmfn, bool_t fast, bool_t all) {} -- 2.9.3 (Apple Git-75) _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx https://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |