[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-devel] [PATCH v2] xen: fix setup of PVH Dom0 memory map



>>> On 22.05.14 at 19:01, <roger.pau@xxxxxxxxxx> wrote:
> @@ -369,6 +375,100 @@ static __init void pvh_map_all_iomem(struct domain *d)
>          nump = end_pfn - start_pfn;
>          pvh_add_mem_mapping(d, start_pfn, start_pfn, nump);
>      }
> +
> +    /*
> +     * Add the memory removed by the holes at the end of the
> +     * memory map.
> +     */
> +    for ( i = 0, entry = e820.map; i < e820.nr_map && nr_holes > 0;
> +          i++, entry++ )
> +    {
> +        if ( entry->type != E820_RAM )
> +            continue;
> +
> +        end_pfn = PFN_UP(entry->addr + entry->size);
> +        if ( end_pfn <= nr_pages )
> +            continue;
> +
> +        navail = end_pfn - nr_pages;
> +        nmap = navail > nr_holes ? nr_holes : navail;

min()?

> +        nr_holes -= nmap;
> +        start_pfn = PFN_DOWN(entry->addr) < nr_pages ?
> +                        nr_pages : PFN_DOWN(entry->addr);

max_t()?

> +        /*
> +         * Populate this memory region using the pages
> +         * previously removed by the MMIO holes.
> +         */
> +        page_list_for_each ( page, &d->page_list )
> +        {
> +            mfn = page_to_mfn(page);
> +            if ( get_gpfn_from_mfn(mfn) != INVALID_M2P_ENTRY )
> +                continue;

This being inside another loop, this is going to be unnecessarily
expensive on big memory systems without dom0_mem=, as you're
going to repeatedly skip a huge (and increasing) number of pages
here. So apart from needing to add a process_pending_softirqs()
call somewhere, I think you should have a cursor into the list
(which isn't going to change under your feet).

Talking of which, Mukesh, I think the lack of periodic calls to
process_pending_softirqs() is also a latent problem in the function
without Roger's changes. Furthermore now that I look at it again I
wonder whether this whole thing shouldn't be SRAT based (when
there is an SRAT) in order to avoid mapping hotplug memory
regions as MMIO.

> +
> +            rc = guest_physmap_add_page(d, start_pfn, mfn, 0);
> +            if ( rc != 0 )
> +                panic("Unable to add gpfn %#lx mfn %#lx to Dom0 physmap",
> +                      start_pfn, page_to_mfn(page));

Please print rc here too, to help diagnosing eventual problems. Also
please use "mfn" rather than "page_to_mfn(page)".

> +            start_pfn++;
> +            if ( --nmap == 0 )
> +                break;
> +        }
> +        ASSERT(nmap == 0);
> +    }
> +
> +    ASSERT(nr_holes == 0);
> +}
> +
> +static __init void pvh_setup_e820(struct domain *d, unsigned long nr_pages)
> +{
> +    struct e820entry *entry, *entry_guest;
> +    unsigned int i;
> +    unsigned long pages, cur_pages = 0;
> +
> +    /*
> +     * Craft the e820 memory map for Dom0 based on the hardware e820 map.
> +     */
> +    d->arch.e820 = xzalloc_array(struct e820entry, e820.nr_map);
> +    if ( !d->arch.e820 )
> +        panic("Unable to allocate memory for Dom0 e820 map");
> +    entry_guest = d->arch.e820;
> +
> +    /* Clamp e820 memory map to match the memory assigned to Dom0 */
> +    for ( i = 0, entry = e820.map; i < e820.nr_map; i++, entry++ )
> +    {
> +        if ( entry->type != E820_RAM )
> +        {
> +            *entry_guest = *entry;
> +            goto next;
> +        }
> +
> +        if ( nr_pages == cur_pages )
> +        {
> +            /*
> +             * We already have all the assigned memory,
> +             * skip this entry
> +             */
> +            continue;
> +        }
> +
> +        *entry_guest = *entry;
> +        pages = (entry_guest->size + PAGE_SIZE-1) >> PAGE_SHIFT;

PFN_UP()?

> +        if ( (cur_pages + pages) > nr_pages )
> +        {
> +            /* Truncate region */
> +            entry_guest->size = (nr_pages - cur_pages) << PAGE_SHIFT;
> +            cur_pages = nr_pages;
> +        }
> +        else
> +        {
> +            cur_pages += pages;
> +        }
> +next:

Labels indented by at least on space please (avoiding diff's -p option
picking this up as hunk context in future patches).

Jan

_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.