|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH v7 2/3] mm: make pages allocated with MEMF_no_refcount safe to assign
Currently it is unsafe to assign a domheap page allocated with
MEMF_no_refcount to a domain because the domain't 'tot_pages' will not
be incremented, but will be decrement when the page is freed (since
free_domheap_pages() has no way of telling that the increment was skipped).
This patch allocates a new 'count_info' bit for a PGC_extra flag
which is then used to mark pages when alloc_domheap_pages() is called
with MEMF_no_refcount. The MEMF_no_refcount is *not* passed through to
assign_pages() because it still needs to call domain_adjust_tot_pages() to
make sure the domain is appropriately referenced. assign_pages() is
accordingly modified to account pages marked with PGC_extra to an
'extra_pages' counter, which is then subtracted from 'tot_pages' before it
is checked against 'max_pages', thus avoiding over-allocation errors.
NOTE: steal_page() is also modified to decrement extra_pages in the case of
a PGC_extra page being stolen from a domain.
Also, whilst adding the extra_pages counter into struct domain, make
some cosmetic fixes to comments for neighbouring fields.
Signed-off-by: Paul Durrant <pdurrant@xxxxxxxxxx>
---
Cc: Andrew Cooper <andrew.cooper3@xxxxxxxxxx>
Cc: George Dunlap <George.Dunlap@xxxxxxxxxxxxx>
Cc: Ian Jackson <ian.jackson@xxxxxxxxxxxxx>
Cc: Jan Beulich <jbeulich@xxxxxxxx>
Cc: Julien Grall <julien@xxxxxxx>
Cc: Konrad Rzeszutek Wilk <konrad.wilk@xxxxxxxxxx>
Cc: Stefano Stabellini <sstabellini@xxxxxxxxxx>
Cc: Wei Liu <wl@xxxxxxx>
Cc: Volodymyr Babchuk <Volodymyr_Babchuk@xxxxxxxx>
Cc: "Roger Pau Monné" <roger.pau@xxxxxxxxxx>
v7:
- s/PGC_no_refcount/PGC_extra/g
- Re-work allocation to account for 'extra' pages, also making it
safe to assign PGC_extra pages post-allocation
v6:
- Add an extra ASSERT into assign_pages() that PGC_no_refcount is not
set if MEMF_no_refcount is clear
- ASSERT that count_info is 0 in alloc_domheap_pages() and set to
PGC_no_refcount rather than ORing
v5:
- Make sure PGC_no_refcount is set before assign_pages() is called
- Don't bother to clear PGC_no_refcount in free_domheap_pages() and
drop ASSERT in free_heap_pages()
- Don't latch count_info in free_heap_pages()
v4:
- New in v4
---
xen/arch/x86/mm.c | 5 ++++
xen/common/page_alloc.c | 49 +++++++++++++++++++++++++++++-----------
xen/include/asm-arm/mm.h | 5 +++-
xen/include/asm-x86/mm.h | 7 ++++--
xen/include/xen/sched.h | 18 ++++++++-------
5 files changed, 60 insertions(+), 24 deletions(-)
diff --git a/xen/arch/x86/mm.c b/xen/arch/x86/mm.c
index f50c065af3..5b04db8c21 100644
--- a/xen/arch/x86/mm.c
+++ b/xen/arch/x86/mm.c
@@ -4266,6 +4266,11 @@ int steal_page(
page_list_del(page, &d->page_list);
/* Unlink from original owner. */
+ if ( page->count_info & PGC_extra )
+ {
+ ASSERT(d->extra_pages);
+ d->extra_pages--;
+ }
if ( !(memflags & MEMF_no_refcount) && !domain_adjust_tot_pages(d, -1) )
drop_dom_ref = true;
diff --git a/xen/common/page_alloc.c b/xen/common/page_alloc.c
index 919a270587..a2d69f222a 100644
--- a/xen/common/page_alloc.c
+++ b/xen/common/page_alloc.c
@@ -2256,6 +2256,7 @@ int assign_pages(
{
int rc = 0;
unsigned long i;
+ unsigned int extra_pages = 0;
spin_lock(&d->page_alloc_lock);
@@ -2267,13 +2268,19 @@ int assign_pages(
goto out;
}
+ for ( i = 0; i < (1 << order); i++ )
+ if ( pg[i].count_info & PGC_extra )
+ extra_pages++;
+
if ( !(memflags & MEMF_no_refcount) )
{
- if ( unlikely((d->tot_pages + (1 << order)) > d->max_pages) )
+ unsigned int max_pages = d->max_pages - d->extra_pages - extra_pages;
+
+ if ( unlikely((d->tot_pages + (1 << order)) > max_pages) )
{
gprintk(XENLOG_INFO, "Over-allocation for domain %u: "
"%u > %u\n", d->domain_id,
- d->tot_pages + (1 << order), d->max_pages);
+ d->tot_pages + (1 << order), max_pages);
rc = -E2BIG;
goto out;
}
@@ -2282,13 +2289,17 @@ int assign_pages(
get_knownalive_domain(d);
}
+ d->extra_pages += extra_pages;
for ( i = 0; i < (1 << order); i++ )
{
+ unsigned long count_info = pg[i].count_info;
+
ASSERT(page_get_owner(&pg[i]) == NULL);
- ASSERT(!pg[i].count_info);
+ ASSERT(!(count_info & ~PGC_extra));
page_set_owner(&pg[i], d);
smp_wmb(); /* Domain pointer must be visible before updating refcnt. */
- pg[i].count_info = PGC_allocated | 1;
+ count_info &= PGC_extra;
+ pg[i].count_info = count_info | PGC_allocated | 1;
page_list_add_tail(&pg[i], &d->page_list);
}
@@ -2314,11 +2325,6 @@ struct page_info *alloc_domheap_pages(
if ( memflags & MEMF_no_owner )
memflags |= MEMF_no_refcount;
- else if ( (memflags & MEMF_no_refcount) && d )
- {
- ASSERT(!(memflags & MEMF_no_refcount));
- return NULL;
- }
if ( !dma_bitsize )
memflags &= ~MEMF_no_dma;
@@ -2331,11 +2337,23 @@ struct page_info *alloc_domheap_pages(
memflags, d)) == NULL)) )
return NULL;
- if ( d && !(memflags & MEMF_no_owner) &&
- assign_pages(d, pg, order, memflags) )
+ if ( d && !(memflags & MEMF_no_owner) )
{
- free_heap_pages(pg, order, memflags & MEMF_no_scrub);
- return NULL;
+ if ( memflags & MEMF_no_refcount )
+ {
+ unsigned long i;
+
+ for ( i = 0; i < (1ul << order); i++ )
+ {
+ ASSERT(!pg[i].count_info);
+ pg[i].count_info = PGC_extra;
+ }
+ }
+ if ( assign_pages(d, pg, order, memflags & ~MEMF_no_refcount) )
+ {
+ free_heap_pages(pg, order, memflags & MEMF_no_scrub);
+ return NULL;
+ }
}
return pg;
@@ -2383,6 +2401,11 @@ void free_domheap_pages(struct page_info *pg, unsigned
int order)
BUG();
}
arch_free_heap_page(d, &pg[i]);
+ if ( pg[i].count_info & PGC_extra )
+ {
+ ASSERT(d->extra_pages);
+ d->extra_pages--;
+ }
}
drop_dom_ref = !domain_adjust_tot_pages(d, -(1 << order));
diff --git a/xen/include/asm-arm/mm.h b/xen/include/asm-arm/mm.h
index 333efd3a60..7df91280bc 100644
--- a/xen/include/asm-arm/mm.h
+++ b/xen/include/asm-arm/mm.h
@@ -119,9 +119,12 @@ struct page_info
#define PGC_state_offlined PG_mask(2, 9)
#define PGC_state_free PG_mask(3, 9)
#define page_state_is(pg, st) (((pg)->count_info&PGC_state) == PGC_state_##st)
+/* Page is not reference counted */
+#define _PGC_extra PG_shift(10)
+#define PGC_extra PG_mask(1, 10)
/* Count of references to this frame. */
-#define PGC_count_width PG_shift(9)
+#define PGC_count_width PG_shift(10)
#define PGC_count_mask ((1UL<<PGC_count_width)-1)
/*
diff --git a/xen/include/asm-x86/mm.h b/xen/include/asm-x86/mm.h
index 2ca8882ad0..06d64d494d 100644
--- a/xen/include/asm-x86/mm.h
+++ b/xen/include/asm-x86/mm.h
@@ -77,9 +77,12 @@
#define PGC_state_offlined PG_mask(2, 9)
#define PGC_state_free PG_mask(3, 9)
#define page_state_is(pg, st) (((pg)->count_info&PGC_state) == PGC_state_##st)
+/* Page is not reference counted */
+#define _PGC_extra PG_shift(10)
+#define PGC_extra PG_mask(1, 10)
- /* Count of references to this frame. */
-#define PGC_count_width PG_shift(9)
+/* Count of references to this frame. */
+#define PGC_count_width PG_shift(10)
#define PGC_count_mask ((1UL<<PGC_count_width)-1)
/*
diff --git a/xen/include/xen/sched.h b/xen/include/xen/sched.h
index 7c5c437247..763fcd56a4 100644
--- a/xen/include/xen/sched.h
+++ b/xen/include/xen/sched.h
@@ -361,15 +361,17 @@ struct domain
spinlock_t domain_lock;
- spinlock_t page_alloc_lock; /* protects all the following fields */
- struct page_list_head page_list; /* linked list */
+ spinlock_t page_alloc_lock; /* protects all the following fields */
+ struct page_list_head page_list; /* linked list */
struct page_list_head xenpage_list; /* linked list (size xenheap_pages) */
- unsigned int tot_pages; /* number of pages currently possesed */
- unsigned int xenheap_pages; /* # pages allocated from Xen heap */
- unsigned int outstanding_pages; /* pages claimed but not possessed */
- unsigned int max_pages; /* maximum value for tot_pages */
- atomic_t shr_pages; /* number of shared pages */
- atomic_t paged_pages; /* number of paged-out pages */
+ unsigned int tot_pages; /* number of pages currently possesed
*/
+ unsigned int xenheap_pages; /* number of pages from Xen heap */
+ unsigned int outstanding_pages; /* pages claimed but not possessed */
+ unsigned int extra_pages; /* extra pages not limited by
max_pages */
+ unsigned int max_pages; /* maximum value for tot_pages minus */
+ /* extra_pages */
+ atomic_t shr_pages; /* number of shared pages */
+ atomic_t paged_pages; /* number of paged-out pages */
/* Scheduling. */
void *sched_priv; /* scheduler-specific data */
--
2.20.1
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxxx
https://lists.xenproject.org/mailman/listinfo/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |