[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH v2 2/9] mm: Place unscrubbed pages at the end of pagelist
On 04/04/2017 11:29 AM, Jan Beulich wrote: >>>> On 04.04.17 at 17:14, <boris.ostrovsky@xxxxxxxxxx> wrote: >> On 04/04/2017 10:46 AM, Jan Beulich wrote: >>>> @@ -933,6 +952,10 @@ static bool_t can_merge(struct page_info *buddy, >>>> unsigned int node, >>>> (phys_to_nid(page_to_maddr(buddy)) != node) ) >>>> return false; >>>> >>>> + if ( need_scrub != >>>> + !!test_bit(_PGC_need_scrub, &buddy->count_info) ) >>>> + return false; >>> I don't think leaving the tree in a state where larger order chunks >>> don't become available for allocation right away is going to be >>> acceptable. Hence with this issue being dealt with only in patch 7 >>> as it seems, you should state clearly and visibly that (at least) >>> patches 2...7 should only be committed together. >> The dirty pages are available for allocation as result of this patch but >> they might not be merged with higher orders (which is what this check is >> for) > The individual chunks are available for allocation, but not the > combined one (for a suitably high order request). Or am I > missing something? Correct, but this is not changed by any later patch (including patch 7). We only merge with a buddy with the same level of cleanliness (so to speak ;-)) > >>>> @@ -952,9 +977,10 @@ static struct page_info *merge_chunks(struct >>>> page_info *pg, unsigned int node, >>>> { >>>> /* Merge with predecessor block? */ >>>> buddy = pg - mask; >>>> - if ( !can_merge(buddy, node, order) ) >>>> + if ( !can_merge(buddy, node, order, need_scrub) ) >>>> break; >>>> >>>> + pg->count_info &= ~PGC_need_scrub; >>>> pg = buddy; >>>> page_list_del(pg, &heap(node, zone, order)); >>>> } >>>> @@ -962,9 +988,10 @@ static struct page_info *merge_chunks(struct >>>> page_info *pg, unsigned int node, >>>> { >>>> /* Merge with successor block? */ >>>> buddy = pg + mask; >>>> - if ( !can_merge(buddy, node, order) ) >>>> + if ( !can_merge(buddy, node, order, need_scrub) ) >>>> break; >>>> >>>> + buddy->count_info &= ~PGC_need_scrub; >>>> page_list_del(buddy, &heap(node, zone, order)); >>>> } >>> For both of these, how come you can / want to clear the need-scrub >>> flag? Wouldn't it be better for each individual page to retain it, so >>> when encountering a higher-order one you know which pages need >>> scrubbing and which don't? Couldn't that also be used to avoid >>> suppressing their merging here right away? >> I am trying to avoid having to keep dirty bit for each page since a >> buddy is either fully clean or fully dirty. That way we shouldn't need >> to walk the list and clear the bit. (I, in fact, suspect that there may >> be other state bits/fields that we might be able to keep at a buddy only) > But as said - at the expense of not being able to merge early. I > consider this a serious limitation. What do you mean by "early"? At freeing time? But then we will always have to scan the buddy during allocation to see if any pages are dirty. > >>>> --- a/xen/include/asm-x86/mm.h >>>> +++ b/xen/include/asm-x86/mm.h >>>> @@ -233,6 +233,10 @@ struct page_info >>>> #define PGC_count_width PG_shift(9) >>>> #define PGC_count_mask ((1UL<<PGC_count_width)-1) >>>> >>>> +/* Page needs to be scrubbed */ >>>> +#define _PGC_need_scrub PG_shift(10) >>>> +#define PGC_need_scrub PG_mask(1, 10) >>> So why not a new PGC_state_dirty instead of this independent >>> flag? Pages other than PGC_state_free should never make it >>> to the scrubber, so the flag is meaningless for all other >>> PGC_state_*. >> Wouldn't doing this require possibly making two checks --- >> page_state_is(pg, free) || page_state_is(pg, dirty)? > Well, your goal would normally be to first look for pages not > needing scrubbing anyway, so quite likely you'd do two > passes anyway. But of course much depends on whether to > merge early or late. Again, I need to understand what you consider "early" and "late". -boris _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx https://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |