|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH v3 03/10] IOMMU/MMU: enhance the call trees of IOMMU unmapping and mapping
>>> On 04.05.16 at 03:45, <kevin.tian@xxxxxxxxx> wrote:
>> From: Xu, Quan
>> Sent: Friday, April 29, 2016 5:25 PM
>> --- a/xen/arch/x86/mm.c
>> +++ b/xen/arch/x86/mm.c
>> @@ -2467,7 +2467,7 @@ static int __get_page_type(struct page_info *page,
>> unsigned
>> long type,
>> int preemptible)
>> {
>> unsigned long nx, x, y = page->u.inuse.type_info;
>> - int rc = 0;
>> + int rc = 0, ret = 0;
>>
>> ASSERT(!(type & ~(PGT_type_mask | PGT_pae_xen_l2)));
>>
>> @@ -2578,11 +2578,11 @@ static int __get_page_type(struct page_info *page,
>> unsigned
>> long type,
>> if ( d && is_pv_domain(d) && unlikely(need_iommu(d)) )
>> {
>> if ( (x & PGT_type_mask) == PGT_writable_page )
>> - iommu_unmap_page(d, mfn_to_gmfn(d, page_to_mfn(page)));
>> + ret = iommu_unmap_page(d, mfn_to_gmfn(d,
>> page_to_mfn(page)));
>> else if ( type == PGT_writable_page )
>> - iommu_map_page(d, mfn_to_gmfn(d, page_to_mfn(page)),
>> - page_to_mfn(page),
>> - IOMMUF_readable|IOMMUF_writable);
>> + ret = iommu_map_page(d, mfn_to_gmfn(d, page_to_mfn(page)),
>> + page_to_mfn(page),
>> + IOMMUF_readable|IOMMUF_writable);
>> }
>> }
>>
>> @@ -2599,6 +2599,9 @@ static int __get_page_type(struct page_info *page,
>> unsigned
>> long type,
>> if ( (x & PGT_partial) && !(nx & PGT_partial) )
>> put_page(page);
>>
>> + if ( !rc )
>> + rc = ret;
>> +
>> return rc;
>> }
>
> I know there were quite some discussions before around above change (sorry I
> didn't remember all of them). Just based on mental picture we should return
> error where it firstly occurs. However above change looks favoring errors in
> later "rc = alloc_page_type" over earlier iommu_map/unmap_page error. Is it
> what we want?
Yes, as that's the primary operation here.
> If there is a reason that we cannot return immediately upon
> iommu_map/unmap,
Since for Dom0 we don't call domain_crash(), we must not bypass
alloc_page_type() here. And even for DomU it would seem at least
fragile if we did - we better don't alter the refcounting behavior.
>> --- a/xen/arch/x86/mm/p2m-ept.c
>> +++ b/xen/arch/x86/mm/p2m-ept.c
>> @@ -821,6 +821,8 @@ out:
>> if ( needs_sync )
>> ept_sync_domain(p2m);
>>
>> + ret = 0;
>> +
>> /* For host p2m, may need to change VT-d page table.*/
>> if ( rc == 0 && p2m_is_hostp2m(p2m) && need_iommu(d) &&
>> need_modify_vtd_table )
>> @@ -831,11 +833,29 @@ out:
>> {
>> if ( iommu_flags )
>> for ( i = 0; i < (1 << order); i++ )
>> - iommu_map_page(d, gfn + i, mfn_x(mfn) + i, iommu_flags);
>> + {
>> + rc = iommu_map_page(d, gfn + i, mfn_x(mfn) + i,
>> iommu_flags);
>> +
>> + if ( !ret && unlikely(rc) )
>
> I think you should move check of ret before iommu_map_page, since we
> should stop map against any error (either from best-effort unmap error side).
Considering ret getting set to zero ahead of the loop plus ...
>> + {
>> + while ( i-- )
>> + iommu_unmap_page(d, gfn + i);
>> +
>> + ret = rc;
>> + break;
... this, it looks to me as if the checking of ret above is simply
unnecessary.
Jan
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |