[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-devel] [PATCH] xen-swiotlb: exchange memory with Xen only when pages are contiguous



> -----Original Message-----
> From: Xen-devel [mailto:xen-devel-bounces@xxxxxxxxxxxxxxxxxxxx] On Behalf
> Of Joe Jin
> Sent: 30 October 2018 14:48
> To: Paul Durrant <Paul.Durrant@xxxxxxxxxx>; Boris Ostrovsky
> <boris.ostrovsky@xxxxxxxxxx>; Konrad Rzeszutek Wilk
> <konrad.wilk@xxxxxxxxxx>
> Cc: John Sobecki <john.sobecki@xxxxxxxxxx>; DONGLI.ZHANG
> <dongli.zhang@xxxxxxxxxx>; linux-kernel@xxxxxxxxxxxxxxx" <linux-
> kernel@xxxxxxxxxxxxxxx>; konrad@xxxxxxxxxx; xen-
> devel@xxxxxxxxxxxxxxxxxxxx; Christoph Helwig <hch@xxxxxx>
> Subject: Re: [Xen-devel] [PATCH] xen-swiotlb: exchange memory with Xen
> only when pages are contiguous
> 
> On 10/30/18 7:21 AM, Paul Durrant wrote:
> >> -----Original Message-----
> >> From: Xen-devel [mailto:xen-devel-bounces@xxxxxxxxxxxxxxxxxxxx] On
> Behalf
> >> Of Joe Jin
> >> Sent: 30 October 2018 14:13
> >> To: Paul Durrant <Paul.Durrant@xxxxxxxxxx>; Boris Ostrovsky
> >> <boris.ostrovsky@xxxxxxxxxx>; Konrad Rzeszutek Wilk
> >> <konrad.wilk@xxxxxxxxxx>
> >> Cc: John Sobecki <john.sobecki@xxxxxxxxxx>; DONGLI.ZHANG
> >> <dongli.zhang@xxxxxxxxxx>; linux-kernel@xxxxxxxxxxxxxxx" <linux-
> >> kernel@xxxxxxxxxxxxxxx>; konrad@xxxxxxxxxx; xen-
> >> devel@xxxxxxxxxxxxxxxxxxxx; Christoph Helwig <hch@xxxxxx>
> >> Subject: Re: [Xen-devel] [PATCH] xen-swiotlb: exchange memory with Xen
> >> only when pages are contiguous
> >>
> >> On 10/30/18 1:59 AM, Paul Durrant wrote:
> >>>> On 10/25/18 11:56 AM, Joe Jin wrote:
> >>>>> I just discussed this patch with Boris in private, his
> opinions(Boris,
> >>>>> please correct me if any misunderstood) are:
> >>>>>
> >>>>> 1. With/without the check, both are incorrect, he thought we need to
> >>>>>    prevented unalloc'd free at here.
> >>>>> 2. On freeing, if upper layer already checked the memory was DMA-
> able,
> >>>>>    the checking at here does not make sense, we can remove all
> checks.
> >>>>> 3. xen_create_contiguous_region() and
> xen_destroy_contiguous_region()
> >>>>>    to come in pairs.
> >>>> I tried to added radix_tree to track allocating/freeing and I found
> >> some
> >>>> memory only allocated but was not freed, I guess it caused by driver
> >> used
> >>>> dma_pool, that means if lots of such requests, the list will consume
> >> lot
> >>>> of memory for it. Will continue to work on it, if anyone have good
> idea
> >>>> for it please let me know, I'd like to try it.
> >>>>
> >>> FWIW, in my Xen PV-IOMMU test patches, I have also tried keeping a
> list
> >> of ranges mapped for DMA and have discovered apparent issues with some
> >> drivers, particularly tg3, that seem to free mappings that have not
> been
> >> allocated (or possibly double-free). I've never fully tracked down the
> >> issue.
> >>
> >> Call trace of first called xen_swiotlb_alloc_coherent(The pages never
> >> backed to Xen):
> >>
> >> [   23.436333]  [<ffffffff814040c9>]
> >> xen_swiotlb_alloc_coherent+0x169/0x510
> >> [   23.436623]  [<ffffffff811eb38d>] ?
> kmem_cache_alloc_trace+0x1ed/0x280
> >> [   23.436900]  [<ffffffff811d72af>] dma_pool_alloc+0x11f/0x260
> >> [   23.437190]  [<ffffffff81537442>] ehci_qh_alloc+0x52/0x120
> >> [   23.437481]  [<ffffffff8153b80f>] ehci_setup+0x2bf/0x8e0
> >> [   23.437760]  [<ffffffff81476d06>] ? __dev_printk+0x46/0xa0
> >> [   23.438042]  [<ffffffff814770b3>] ? _dev_info+0x53/0x60
> >> [   23.438327]  [<ffffffff8153f620>] ehci_pci_setup+0xc0/0x5f0
> >> [   23.438615]  [<ffffffff81519fcd>] usb_add_hcd+0x25d/0xaf0
> >> [   23.438901]  [<ffffffff8152c9a6>] usb_hcd_pci_probe+0x406/0x520
> >> [   23.439177]  [<ffffffff8153f486>] ehci_pci_probe+0x36/0x40
> >> [   23.439469]  [<ffffffff8136e99a>] local_pci_probe+0x4a/0xb0
> >> [   23.439752]  [<ffffffff8136fba5>] ? pci_match_device+0xe5/0x110
> >> [   23.440027]  [<ffffffff8136fce1>] pci_device_probe+0xd1/0x120
> >> [   23.440320]  [<ffffffff8147b13c>] driver_probe_device+0x20c/0x4d0
> >> [   23.440599]  [<ffffffff8147b4eb>] __driver_attach+0x9b/0xa0
> >> [   23.440879]  [<ffffffff8147b450>] ? __device_attach+0x50/0x50
> >>
> >> Above was EHCI used DMA pool to allocate DMA memory.
> >>
> >> During my testing, ~1000 entries was not freed, if more PCI devices
> >> use DMA pool, the tree/list will have more entries, looks it's not a
> >> good idea that use a list to track it.
> >>
> >
> > Yes, it seems pools can hang onto a serious number of allocations so a
> list is probably not wise.
> 
> I agree with you.
> 
> > What I was pointing out, though, is that it appears you can't even track
> mappings (as opposed to allocations) with a list.
> 
> Right.
> 
> 
> > either because drivers apparently try to unmap things they have not
> mapped.
> 
> If this happened, should be fixed by driver :)

Oh, totally agreed. I fear the breakage is quite indemic throughout many 
drivers though :-(

  Paul

> 
> Thanks,
> Joe
> >
> >   Paul
> >
> 
> _______________________________________________
> Xen-devel mailing list
> Xen-devel@xxxxxxxxxxxxxxxxxxxx
> https://lists.xenproject.org/mailman/listinfo/xen-devel
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxxx
https://lists.xenproject.org/mailman/listinfo/xen-devel

 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.