[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] Re: Question about x86/mm/gup.c's use of disabled interrupts
Jeremy Fitzhardinge wrote: Avi Kivity wrote:Hm, awkward if flush_tlb_others doesn't IPI...How can it avoid flushing the tlb on cpu [01]? It's it's gup_fast()ing a pte, it may as well load it into the tlb.xen_flush_tlb_others uses a hypercall rather than an IPI, so none of the logic which depends on there being an IPI will work. Right, of course, that's what we were talking about. I thought optimizations to avoid IPIs if an mm never visited a cpu. Simplest fix is to make gup_get_pte() a pvop, but that does seem like putting a red flag in front of an inner-loop hotspot, or something...The per-cpu tlb-flush exclusion flag might really be the way to go.I don't see how it will work, without changing Xen to look at the flag?local_irq_disable() is used here to lock out a remote cpu, I don't see why deferring the flush helps.Well, no, not deferring. Making xen_flush_tlb_others() spin waiting for "doing_gup" to clear on the target cpu. Or add an explicit notion of a "pte update barrier" rather than implicitly relying on the tlb IPI (which is extremely convenient when available...). Pick up a percpu flag from all cpus and spin on each? Nasty.You could use the irq enabled flag; it's available and what native spins on (but also means I'll need to add one if I implement this). -- I have a truly marvellous patch that fixes the bug which this signature is too narrow to contain. _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |