[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [RFC + Queries] Flow of PCI passthrough in ARM
On Thu, 25 Sep 2014, manish jaggi wrote: > On 24 September 2014 19:40, Stefano Stabellini > <stefano.stabellini@xxxxxxxxxxxxx> wrote: > > CC'ing Matt and Dave at ARM for an opinions about device tree, SMMUs and > > stream ids. See below. > > > > On Wed, 24 Sep 2014, manish jaggi wrote: > >> On 22 September 2014 16:15, Stefano Stabellini > >> <stefano.stabellini@xxxxxxxxxxxxx> wrote: > >> > On Thu, 18 Sep 2014, manish jaggi wrote: > >> >> Hi, > >> >> Below is the flow I am working on, Please provide your comments, I > >> >> have a couple of queries as well.. > >> >> > >> >> a) Device tree has smmu nodes and each smmu node has the mmu-master > >> >> property. > >> >> In our Soc DT the mmu-master is a pcie node in device tree. > >> > > >> > Do you mean that both the smmu nodes and the pcie node have the > >> > mmu-master property? The pcie node is the pcie root complex, right? > >> > > >> pci-node is the pcie root complex. pci node is the mmu master in smmu node. > >> > >> smmu1@0x8310,00000000 { > >> ... > >> > >> mmu-masters = <&pcie1 0x100>; > >> }; > >> > >> >> b) Xen parses the device tree and prepares a list which stores the pci > >> >> device tree node pointers. The order in device tree is mapped to > >> >> segment number in subsequent calls. For eg 1st pci node found is > >> >> segment 0, 2nd segment 1 > >> > > >> > What's a segment number? Something from the PCI spec? > >> > If you have several pci nodes on device tree, does that mean that you > >> > have several different pcie root complexes? > >> > > >> yes. > >> segment is the pci rc number. > >> > > >> >> c) During SMMU init the pcie nodes in DT are saved as smmu masters. > >> > > >> > At this point you should also be able to find via DT the stream-id range > >> > supported by each SMMU and program the SMMU with them, assigning > >> > everything to dom0. > >> Currently pcie enumeration is not done in xen, it is done in dom0. > > > > Yes, but we don't really need to walk any PCIe busses in order to > > program the SMMU, right? We only need the requestor id and the stream id > > ranges. We should be able to get them via device tree. > > > Yes, but i have a doubt here > Before booting dom0 for each smmu the mask in SMR can be set to enable > stream ids to dom0. > This can be fixed or read from device tree. > There are 2 points here > a) PCI bus enumeration > b) Programming SMMU for dom0 > For (b) the enumeration is not required provided we set the mask > So are you also saying that (a) should be done in Xen and not in dom0 ? > If yes how would dom0 get to know about PCIe Eps , from its Device tree ? No, I think that doing (a) via PHYSDEVOP_pci_device_add is OK. I am saying that we should consider doing (b) in Xen before booting dom0. > >> >> d) Dom0 Enumerates PCI devices, calls hypercall > >> >> PHYSDEVOP_pci_device_add. > >> >> - In Xen the SMMU iommu_ops add_device is called. I have implemented > >> >> the add_device function. > >> >> - In the add_device function > >> >> the segment number is used to locate the device tree node pointer of > >> >> the pcie node which helps to find out the corresponding smmu. > >> >> - In the same PHYSDEVOP the BAR regions are mapped to Dom0. > >> >> > >> >> Note: The current SMMU driver maps the complete Domain's Address space > >> >> for the device in SMMU hardware. > >> >> > >> >> The above flow works currently for us. > >> > > >> > It would be nice to be able to skip d): in a system where all dma capable > >> > devices are behind smmus, we should be capable of booting dom0 without > >> > the 1:1 mapping hack. If we do that, it would be better to program the > >> > smmus before booting dom0. Otherwise there is a risk that dom0 is going > >> > to start using these devices and doing dma before we manage to secure > >> > the devices via smmus. > >> > > >> In our current case we are programming smmu in > >> PHYSDEVOP_pci_device_add flow so before the domain 0 accesses the > >> device it is mapped, otherwise xen gets a SMMU fault. > > > > Good. > > > > > >> > Of course we can do that if there are no alternatives. But in our case > >> > we should be able to extract the stream-ids from device tree and program > >> > the smmus right away, right? Do we really need to wait for dom0 to call > >> > PHYSDEVOP_pci_device_add? We could just assign everything to dom0 for a > >> > start. > >> > > >> We cannot get streamid from device tree as enumeration is done for the > >> same. > > > > I am not sure what the current state of the device tree spec is, but I > > am pretty sure that the intention is to express stream id and requestor > > id ranges directly in the dts so that the SMMU can be programmed right > > away without walking the PCI bus. > > > > > >> > I would like to know from the x86 guys, if this is really how it is > >> > supposed to work on PVH too. Do we rely on PHYSDEVOP_pci_device_add to > >> > program the IOMMU? > >> > > >> > > >> I was waiting but no one has commented > > > > Me too. Everybody is very busy at the moment with the 4.5 release. > > > > > >> >> Now when I call pci-assignable-add I see that the iommu_ops > >> >> remove_device in smmu driver is not called. If that is not called the > >> >> SMMU would still have the dom0 address space mappings for that device > >> >> > >> >> Can you please suggest the best place (kernel / xl-tools) to put the > >> >> code which would call the remove_device in iommu_opps in the control > >> >> flow from pci-assignable-add. > >> >> > >> >> One way I see is to introduce a DOMCTL_iommu_remove_device in > >> >> pci-assignable-add / pci-detach and DOMCTL_iommu_add_device in > >> >> pci-attach. Is that a valid approach ? > >> > > >> > I am not 100% sure, but I think that before assigning a PCI device to > >> > another guest, you are supposed to bind the device to xen-pciback (see > >> > drivers/xen/xen-pciback, also see > >> > http://wiki.xen.org/wiki/Xen_PCI_Passthrough). The pciback driver is > >> > going hide the device from dom0 and as a consequence > >> > drivers/xen/pci.c:xen_remove_device ends up being called, that issues a > >> > PHYSDEVOP_pci_device_remove hypercall. > >> > >> xen_remove_device is not called at all. in pci-attach > >> iommu_ops->assign_device is called. > >> In Xen the nomenclature is confusing and no comments are there is iommu.h > >> iommu_ops.add_device is when dom0 issues hypercall > >> iommu_ops.assign_dt_device is when xen attaches a device tree device to > >> dom0 > >> iommu_ops.assign_device is when xl pci-attach is called > >> iommu_ops.reassign_device is called when xl pci-detach is called > >> > >> As of now we are able to assign devices to domU and driver in domU is > >> running, we did some hacks like > >> a) in xen pci front driver bus->msi is assigned to its msi_chip > >> > >> ---- pcifront_scan_root() > >> ... > >> b = pci_scan_bus_parented(&pdev->xdev->dev, bus, > >> &pcifront_bus_ops, sd); > >> if (!b) { > >> dev_err(&pdev->xdev->dev, > >> "Error creating PCI Frontend Bus!\n"); > >> err = -ENOMEM; > >> pci_unlock_rescan_remove(); > >> goto err_out; > >> } > >> > >> bus_entry->bus = b; > >> + msi_node = of_find_compatible_node(NULL,NULL, "arm,gic-v3-its"); > >> + if(msi_node) { > >> + b->msi = of_pci_find_msi_chip_by_node(msi_node); > >> + if(!b->msi) { > >> + printk(KERN_ERR"Unable to find bus->msi node \r\n"); > >> + goto err_out; > >> + } > >> + }else { > >> + printk(KERN_ERR"Unable to find arm,gic-v3-its > >> compatible node \r\n"); > >> + goto err_out; > >> + } > > > > It seems to be that of_pci_find_msi_chip_by_node should be called by > > common code somewhere else. Maybe people at linux-arm would know where > > to suggest this initialization should go. > > > This is a workaround to attach a msi-controller to xen pcifront bus. > We are avoiding the xen fronted ops for msi. I think I would need to see a proper patch series to really evaluate this change. > > > >> ---- > >> > >> using this the ITS emulation code in xen is able to trap ITS command > >> writes by driver. > >> But we are facing a problem now, where your help is needed > >> > >> The StreamID is generated by segment: bus : device: number which is > >> fed as DevID in ITS commands. In Dom0 the streamID is correctly > >> generated but in domU the Stream ID for a passthrough device is > >> 0:0:0:0 now when emulating this in Xen it is a problem as xen does not > >> know how to get the physical stream id. > >> > >> (Eg: xl pci-attach 1 0001:00:05.0 > >> DomU has the device but in DomU the id is 0000:00:00.0.) > >> > >> Could you suggest how to go about this. > > > > I don't think that the ITS patches have been posted yet, so it is > > difficult for me to understand the problem and propose a solution. > > In a simpler way, It is more of what the StreamID a driver running in > domU sees. Which is programmed in the ITS commands. > And how to map the domU streamID to actual streamID in Xen when the > ITS command write traps. Wouldn't it be possible to pass the correct StreamID to DomU via device tree? Does it really need to match the PCI BDF? Otherwise if the command trap into Xen, couldn't Xen do the translation? _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |