[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-devel] [PATCHv3] x86/xen: avoid m2p lookup when setting early page table entries



On 23/06/16 15:37, David Vrabel wrote:
> On 23/06/16 14:27, Juergen Gross wrote:
>> On 23/06/16 15:06, David Vrabel wrote:
>>> On 23/06/16 13:13, Juergen Gross wrote:
>>>> On 23/06/16 11:51, David Vrabel wrote:
>>>>> When page tables entries are set using xen_set_pte_init() during early
>>>>> boot there is no page fault handler that could handle a fault when
>>>>> performing an M2P lookup.
>>>>>
>>>>> In 64 bit guests (usually dom0) early_ioremap() would fault in
>>>>> xen_set_pte_init() because an M2P lookup faults because the MFN is in
>>>>> MMIO space and not mapped in the M2P.  This lookup is done to see if
>>>>> the PFN in in the range used for the initial page table pages, so that
>>>>> the PTE may be set as read-only.
>>>>>
>>>>> The M2P lookup can be avoided by moving the check (and clear of RW)
>>>>> earlier when the PFN is still available.
>>>>>
>>>>> Signed-off-by: David Vrabel <david.vrabel@xxxxxxxxxx>
>>>>> ---
>>>>> Cc: Kevin Moraga <kmoragas@xxxxxxxxxx>
>>>>>
>>>>> v3:
>>>>> - fold mask_rw_pte()/mask_rw_pteval() into their callers.
>>>>>
>>>>> v2:
>>>>> - Remove __init annotation from xen_make_pte_init() since
>>>>>   PV_CALLEE_SAVE_REGS_THUNK always puts the thunk in .text.
>>>>>
>>>>> - mask_rw_pte() -> mask_rw_pteval() for x86-64.
>>>>> ---
>>>>>  arch/x86/xen/mmu.c | 76 
>>>>> +++++++++++++++++++++++++-----------------------------
>>>>>  1 file changed, 35 insertions(+), 41 deletions(-)
>>>>>
>>>>> diff --git a/arch/x86/xen/mmu.c b/arch/x86/xen/mmu.c
>>>>> index 478a2de..64d8f0b 100644
>>>>> --- a/arch/x86/xen/mmu.c
>>>>> +++ b/arch/x86/xen/mmu.c
>>>>> @@ -1551,41 +1551,6 @@ static void xen_pgd_free(struct mm_struct *mm, 
>>>>> pgd_t *pgd)
>>>>>  #endif
>>>>>  }
>>>>>  
>>>>> -#ifdef CONFIG_X86_32
>>>>> -static pte_t __init mask_rw_pte(pte_t *ptep, pte_t pte)
>>>>> -{
>>>>> - /* If there's an existing pte, then don't allow _PAGE_RW to be set */
>>>>> - if (pte_val_ma(*ptep) & _PAGE_PRESENT)
>>>>> -         pte = __pte_ma(((pte_val_ma(*ptep) & _PAGE_RW) | ~_PAGE_RW) &
>>>>> -                        pte_val_ma(pte));
>>>>> -
>>>>> - return pte;
>>>>> -}
>>>>> -#else /* CONFIG_X86_64 */
>>>>> -static pte_t __init mask_rw_pte(pte_t *ptep, pte_t pte)
>>>>> -{
>>>>> - unsigned long pfn;
>>>>> -
>>>>> - if (xen_feature(XENFEAT_writable_page_tables) ||
>>>>> -     xen_feature(XENFEAT_auto_translated_physmap) ||
>>>>> -     xen_start_info->mfn_list >= __START_KERNEL_map)
>>>>> -         return pte;
>>>>> -
>>>>> - /*
>>>>> -  * Pages belonging to the initial p2m list mapped outside the default
>>>>> -  * address range must be mapped read-only. This region contains the
>>>>> -  * page tables for mapping the p2m list, too, and page tables MUST be
>>>>> -  * mapped read-only.
>>>>> -  */
>>>>> - pfn = pte_pfn(pte);
>>>>> - if (pfn >= xen_start_info->first_p2m_pfn &&
>>>>> -     pfn < xen_start_info->first_p2m_pfn + xen_start_info->nr_p2m_frames)
>>>>> -         pte = __pte_ma(pte_val_ma(pte) & ~_PAGE_RW);
>>>>> -
>>>>> - return pte;
>>>>> -}
>>>>> -#endif /* CONFIG_X86_64 */
>>>>> -
>>>>>  /*
>>>>>   * Init-time set_pte while constructing initial pagetables, which
>>>>>   * doesn't allow RO page table pages to be remapped RW.
>>>>> @@ -1600,13 +1565,41 @@ static pte_t __init mask_rw_pte(pte_t *ptep, 
>>>>> pte_t pte)
>>>>>   * so always write the PTE directly and rely on Xen trapping and
>>>>>   * emulating any updates as necessary.
>>>>>   */
>>>>> -static void __init xen_set_pte_init(pte_t *ptep, pte_t pte)
>>>>> +__visible pte_t xen_make_pte_init(pteval_t pte)
>>>>>  {
>>>>> - if (pte_mfn(pte) != INVALID_P2M_ENTRY)
>>>>> -         pte = mask_rw_pte(ptep, pte);
>>>>> - else
>>>>> -         pte = __pte_ma(0);
>>>>> +#ifdef CONFIG_X86_64
>>>>> + unsigned long pfn;
>>>>> +
>>>>> + /*
>>>>> +  * Pages belonging to the initial p2m list mapped outside the default
>>>>> +  * address range must be mapped read-only. This region contains the
>>>>> +  * page tables for mapping the p2m list, too, and page tables MUST be
>>>>> +  * mapped read-only.
>>>>> +  */
>>>>> + pfn = (pte & PTE_PFN_MASK) >> PAGE_SHIFT;
>>>>> + if (xen_start_info->mfn_list < __START_KERNEL_map &&
>>>>> +     pfn >= xen_start_info->first_p2m_pfn &&
>>>>> +     pfn < xen_start_info->first_p2m_pfn + xen_start_info->nr_p2m_frames)
>>>>> +         pte &= ~_PAGE_RW;
>>>>> +#endif
>>>>> + pte = pte_pfn_to_mfn(pte);
>>>>>  
>>>>> + if ((pte & PTE_PFN_MASK) >> PAGE_SHIFT == INVALID_P2M_ENTRY)
>>>>
>>>> How can this ever be true? I know this is just the open coded
>>>> variant form the original xen_set_pte_init(). Either the if isn't
>>>> needed at all or it should be corrected.
>>>
>>> The frame might be ballooned out.
>>
>> Okay. But how do you get the 64 bit value with all bits set by shifting
>> an unsigned value to the right (I admit that this is possible on a 32
>> bit dom0, but I doubt that was the intention of the test)?
> 
> pte_pfn_to_mfn() already does the right thing so this test is bogus.
> I've removed it.
> 
> Can I add your acked-by or reviewed-by with this change?

Yes, with the test removed: Reviewed-by: Juergen Gross <jgross@xxxxxxxx>


Juergen

_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel

 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.