|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-changelog] [xen staging] x86: Improve the efficiency of domain_relinquish_resources()
commit 0dfffe01d5681ede6a50c6b57131320d9f4a3361
Author: Andrew Cooper <andrew.cooper3@xxxxxxxxxx>
AuthorDate: Wed Feb 20 13:39:20 2019 +0000
Commit: Andrew Cooper <andrew.cooper3@xxxxxxxxxx>
CommitDate: Fri Feb 22 14:10:42 2019 +0000
x86: Improve the efficiency of domain_relinquish_resources()
pci_release_devices() takes the global PCI lock. Once pci_release_devices()
has completed, it will be called redundantly each time paging_teardown() and
vcpu_destroy_pagetables() continue.
This is liable to be millions of times for a reasonably sized guest, and is
a
serialising bottleneck now that domain_kill() can be run concurrently on
different domains.
Instead of propagating the opencoding of the relinquish state machine, take
the opportunity to clean it up.
Leave a proper set of comments explaining that domain_relinquish_resources()
implements a co-routine. Introduce a documented PROGRESS() macro to avoid
latent bugs such as the RELMEM_xen case, and make the new PROG_* states
private to domain_relinquish_resources().
Signed-off-by: Andrew Cooper <andrew.cooper3@xxxxxxxxxx>
Reviewed-by: Wei Liu <wei.liu2@xxxxxxxxxx>
Acked-by: Jan Beulich <jbeulich@xxxxxxxx>
Release-acked-by: Juergen Gross <jgross@xxxxxxxx>
---
xen/arch/x86/domain.c | 71 +++++++++++++++++++++++++++++---------------
xen/include/asm-x86/domain.h | 10 +------
2 files changed, 48 insertions(+), 33 deletions(-)
diff --git a/xen/arch/x86/domain.c b/xen/arch/x86/domain.c
index 32dc4253ff..7a29435cb3 100644
--- a/xen/arch/x86/domain.c
+++ b/xen/arch/x86/domain.c
@@ -475,8 +475,6 @@ int arch_domain_create(struct domain *d,
int rc;
INIT_LIST_HEAD(&d->arch.pdev_list);
-
- d->arch.relmem = RELMEM_not_started;
INIT_PAGE_LIST_HEAD(&d->arch.relmem_list);
spin_lock_init(&d->arch.e820_lock);
@@ -2020,18 +2018,51 @@ int domain_relinquish_resources(struct domain *d)
BUG_ON(!cpumask_empty(d->dirty_cpumask));
- switch ( d->arch.relmem )
+ /*
+ * This hypercall can take minutes of wallclock time to complete. This
+ * logic implements a co-routine, stashing state in struct domain across
+ * hypercall continuation boundaries.
+ */
+ switch ( d->arch.rel_priv )
{
- case RELMEM_not_started:
+ /*
+ * Record the current progress. Subsequent hypercall continuations
+ * will logically restart work from this point.
+ *
+ * PROGRESS() markers must not be in the middle of loops. The loop
+ * variable isn't preserved across a continuation.
+ *
+ * To avoid redundant work, there should be a marker before each
+ * function which may return -ERESTART.
+ */
+#define PROGRESS(x) \
+ d->arch.rel_priv = PROG_ ## x; /* Fallthrough */ case PROG_ ## x
+
+ enum {
+ PROG_paging = 1,
+ PROG_vcpu_pagetables,
+ PROG_shared,
+ PROG_xen,
+ PROG_l4,
+ PROG_l3,
+ PROG_l2,
+ PROG_done,
+ };
+
+ case 0:
ret = pci_release_devices(d);
if ( ret )
return ret;
+ PROGRESS(paging):
+
/* Tear down paging-assistance stuff. */
ret = paging_teardown(d);
if ( ret )
return ret;
+ PROGRESS(vcpu_pagetables):
+
/* Drop the in-use references to page-table bases. */
for_each_vcpu ( d, v )
{
@@ -2058,10 +2089,7 @@ int domain_relinquish_resources(struct domain *d)
d->arch.auto_unmask = 0;
}
- d->arch.relmem = RELMEM_shared;
- /* fallthrough */
-
- case RELMEM_shared:
+ PROGRESS(shared):
if ( is_hvm_domain(d) )
{
@@ -2072,45 +2100,40 @@ int domain_relinquish_resources(struct domain *d)
return ret;
}
- d->arch.relmem = RELMEM_xen;
-
spin_lock(&d->page_alloc_lock);
page_list_splice(&d->arch.relmem_list, &d->page_list);
INIT_PAGE_LIST_HEAD(&d->arch.relmem_list);
spin_unlock(&d->page_alloc_lock);
- /* Fallthrough. Relinquish every page of memory. */
- case RELMEM_xen:
+ PROGRESS(xen):
+
ret = relinquish_memory(d, &d->xenpage_list, ~0UL);
if ( ret )
return ret;
- d->arch.relmem = RELMEM_l4;
- /* fallthrough */
- case RELMEM_l4:
+ PROGRESS(l4):
+
ret = relinquish_memory(d, &d->page_list, PGT_l4_page_table);
if ( ret )
return ret;
- d->arch.relmem = RELMEM_l3;
- /* fallthrough */
- case RELMEM_l3:
+ PROGRESS(l3):
+
ret = relinquish_memory(d, &d->page_list, PGT_l3_page_table);
if ( ret )
return ret;
- d->arch.relmem = RELMEM_l2;
- /* fallthrough */
- case RELMEM_l2:
+ PROGRESS(l2):
+
ret = relinquish_memory(d, &d->page_list, PGT_l2_page_table);
if ( ret )
return ret;
- d->arch.relmem = RELMEM_done;
- /* fallthrough */
- case RELMEM_done:
+ PROGRESS(done):
break;
+#undef PROGRESS
+
default:
BUG();
}
diff --git a/xen/include/asm-x86/domain.h b/xen/include/asm-x86/domain.h
index 277f99f633..58ade0b48e 100644
--- a/xen/include/asm-x86/domain.h
+++ b/xen/include/asm-x86/domain.h
@@ -314,15 +314,7 @@ struct arch_domain
int page_alloc_unlock_level;
/* Continuable domain_relinquish_resources(). */
- enum {
- RELMEM_not_started,
- RELMEM_shared,
- RELMEM_xen,
- RELMEM_l4,
- RELMEM_l3,
- RELMEM_l2,
- RELMEM_done,
- } relmem;
+ unsigned int rel_priv;
struct page_list_head relmem_list;
const struct arch_csw {
--
generated by git-patchbot for /home/xen/git/xen.git#staging
_______________________________________________
Xen-changelog mailing list
Xen-changelog@xxxxxxxxxxxxxxxxxxxx
https://lists.xenproject.org/xen-changelog
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |