[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH v3 4/9] mm: Scrub memory from idle loop
>>> On 04.05.17 at 19:09, <boris.ostrovsky@xxxxxxxxxx> wrote: > On 05/04/2017 11:31 AM, Jan Beulich wrote: >>>>> On 14.04.17 at 17:37, <boris.ostrovsky@xxxxxxxxxx> wrote: >>> --- a/xen/common/page_alloc.c >>> +++ b/xen/common/page_alloc.c >>> @@ -1035,16 +1035,82 @@ merge_and_free_buddy(struct page_info *pg, unsigned >>> int node, >>> return pg; >>> } >>> >>> -static void scrub_free_pages(unsigned int node) >>> +static nodemask_t node_scrubbing; >>> + >>> +static unsigned int node_to_scrub(bool get_node) >>> +{ >>> + nodeid_t node = cpu_to_node(smp_processor_id()), local_node; >>> + nodeid_t closest = NUMA_NO_NODE; >>> + u8 dist, shortest = 0xff; >>> + >>> + if ( node == NUMA_NO_NODE ) >>> + node = 0; >>> + >>> + if ( node_need_scrub[node] && >>> + (!get_node || !node_test_and_set(node, node_scrubbing)) ) >>> + return node; >>> + >>> + /* >>> + * See if there are memory-only nodes that need scrubbing and choose >>> + * the closest one. >>> + */ >>> + local_node = node; >>> + while ( 1 ) >>> + { >>> + do { >>> + node = cycle_node(node, node_online_map); >>> + } while ( !cpumask_empty(&node_to_cpumask(node)) && >>> + (node != local_node) ); >>> + >>> + if ( node == local_node ) >>> + break; >>> + >>> + if ( node_need_scrub[node] ) >>> + { >>> + if ( !get_node ) >>> + return node; >> I think the function parameter name is not / no longer suitable. The >> caller wants to get _some_ node in either case. The difference is >> whether it wants to just know whether there's _any_ needing scrub >> work done, or whether it wants _the one_ to actually scrub on. So >> how about "get_any" or "get_any_node" or just "any"? > > Not only to find out whether there is anything to scrub but, if get_node > is true, to actually "get" it, i.e. set the bit in the node_scrubbing > mask. Thus the name. Hmm, okay, in that case at least an explanatory comment should be added. >>> +bool scrub_free_pages(void) >>> { >>> struct page_info *pg; >>> unsigned int zone, order; >>> unsigned long i; >>> + unsigned int cpu = smp_processor_id(); >>> + bool preempt = false; >>> + nodeid_t node; >>> >>> - ASSERT(spin_is_locked(&heap_lock)); >>> + /* >>> + * Don't scrub while dom0 is being constructed since we may >>> + * fail trying to call map_domain_page() from scrub_one_page(). >>> + */ >>> + if ( system_state < SYS_STATE_active ) >>> + return false; >> I assume that's because of the mapcache vcpu override? That's x86 >> specific though, so the restriction here ought to be arch specific. >> Even better would be to find a way to avoid this restriction >> altogether, as on bigger systems only one CPU is actually busy >> while building Dom0, so all others could be happily scrubbing. Could >> that override become a per-CPU one perhaps? > > Is it worth doing though? What you are saying below is exactly why I > simply return here --- there were very few dirty pages. Well, in that case the comment should cover this second reason as well, at the very least. > This may change > if we decide to use idle-loop scrubbing for boot scrubbing as well (as > Andrew suggested earlier) but there is little reason to do it now IMO. Why not? In fact I had meant to ask why your series doesn't include that? >> Otoh there's not much to scrub yet until Dom0 had all its memory >> allocated, and we know which pages truly remain free (wanting >> what is currently the boot time scrubbing done on them). But that >> point in time may still be earlier than when we switch to >> SYS_STATE_active. IOW I think boot scrubbing could be kicked off as soon as Dom0 had the bulk of its memory allocated. >>> @@ -1065,16 +1131,29 @@ static void scrub_free_pages(unsigned int node) >>> pg[i].count_info &= ~PGC_need_scrub; >>> node_need_scrub[node]--; >>> } >>> + if ( softirq_pending(cpu) ) >>> + { >>> + preempt = true; >>> + break; >>> + } >> Isn't this a little too eager, especially if you didn't have to scrub >> the page on this iteration? > > What would be a good place then? Count how actually scrubbed pages and > check for pending interrupts every so many? Yes. > Even if we don't scrub at all walking whole heap can take a while. Correct - you can't skip this check altogether even if no page requires actual scrubbing. >>> @@ -1141,9 +1220,6 @@ static void free_heap_pages( >>> if ( tainted ) >>> reserve_offlined_page(pg); >>> >>> - if ( need_scrub ) >>> - scrub_free_pages(node); >> I'd expect this eliminates the need for the need_scrub variable. > > We still need it to decide whether to set PGC_need_scrub on pages. Oh, of course. Jan _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx https://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |