|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [PATCH v3 4/9] xen/riscv: setup fixmap mapping
On 05.08.2024 18:02, oleksii.kurochko@xxxxxxxxx wrote:
> On Mon, 2024-08-05 at 17:45 +0200, Jan Beulich wrote:
>> On 05.08.2024 17:13, oleksii.kurochko@xxxxxxxxx wrote:
>>> On Mon, 2024-07-29 at 15:35 +0200, Jan Beulich wrote:
>>>>> + }
>>>>> +
>>>>> + BUG_ON(pte_is_valid(*pte));
>>>>> +
>>>>> + tmp = paddr_to_pte(LINK_TO_LOAD((unsigned
>>>>> long)&xen_fixmap),
>>>>> PTE_TABLE);
>>>>
>>>> I'm a little puzzled by the use of LINK_TO_LOAD() (and
>>>> LOAD_TO_LINK()
>>>> a
>>>> little further up) here. Don't you have functioning __pa() and
>>>> __va()?
>>> Can __pa() and __va() be used in this case?
>>>
>>> According to comments for other architectures, these macros are
>>> used
>>> for converting between Xen heap virtual addresses (VA) and machine
>>> addresses (MA). I may have misunderstood what is meant by the Xen
>>> heap
>>> in this context, but I'm not sure if xen_fixmap[] and page tables
>>> are
>>> considered part of the Xen heap.
>>
>> I didn't check Arm, but on x86 virt_to_maddr() (underlying __pa())
>> has
>> special case code to also allow addresses within the Xen image
>> (area).
>
> Yes, it is true for __virt_to_maddr:
> static inline unsigned long __virt_to_maddr(unsigned long va)
> {
> ASSERT(va < DIRECTMAP_VIRT_END);
> if ( va >= DIRECTMAP_VIRT_START )
> va -= DIRECTMAP_VIRT_START;
> else
> {
> BUILD_BUG_ON(XEN_VIRT_END - XEN_VIRT_START != GB(1));
> /* Signed, so ((long)XEN_VIRT_START >> 30) fits in an imm32. */
> ASSERT(((long)va >> (PAGE_ORDER_1G + PAGE_SHIFT)) ==
> ((long)XEN_VIRT_START >> (PAGE_ORDER_1G + PAGE_SHIFT)));
>
> va += xen_phys_start - XEN_VIRT_START;
> }
> return (va & ma_va_bottom_mask) |
> ((va << pfn_pdx_hole_shift) & ma_top_mask);
> }
>
> But in case of __maddr_to_virt ( __va() ) it is using directmap region:
> static inline void *__maddr_to_virt(unsigned long ma)
> {
> ASSERT(pfn_to_pdx(ma >> PAGE_SHIFT) < (DIRECTMAP_SIZE >>
> PAGE_SHIFT));
> return (void *)(DIRECTMAP_VIRT_START +
> ((ma & ma_va_bottom_mask) |
> ((ma & ma_top_mask) >> pfn_pdx_hole_shift)));
> }
>
> But I have to use both __va() and __pa().
> __va() inside cycle to find L1 page table:
>
> for ( i = HYP_PT_ROOT_LEVEL; i-- > 1; )
> {
> BUG_ON(!pte_is_valid(*pte));
>
> pte = (pte_t *)LOAD_TO_LINK(pte_to_paddr(*pte));
> pte = &pte[pt_index(i, FIXMAP_ADDR(0))];
> }
>
> __pa() to set a physical address of L0 page table:
> tmp = paddr_to_pte(LINK_TO_LOAD((unsigned long)&xen_fixmap),
> PTE_TABLE);
> write_pte(pte, tmp);
Hmm, then using at least LINK_TO_LOAD() is going to be unavoidable for the
time being, I guess. Yet midterm I think they should disappear from here.
Jan
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |