[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] Interrupt to CPU routing in HVM domains - again
John Levon wrote: On Fri, Sep 05, 2008 at 01:11:41PM -0400, Steve Ofsthun wrote:I can submit a patch for this, but we feel this is something of a hack.#ifdef IRQ0_SPECIAL_ROUTING /* Force round-robin to pick VCPU 0 */ if ( ((irq == hvm_isa_irq_to_gsi(0)) && pit_channel0_enabled()) || is_hvm_callback_irq(vioapic, irq) ) deliver_bitmask = (uint32_t)1; #endifYes, please - Solaris 10 PV drivers are buggy in that they use the current VCPUs vcpu_info. I just found this bug, and it's getting fixed, but if this makes sense anyway, it'd be good.Yep. OK, I'll throw a patch together. We'd like to provide a more general mechanism for allowing event channel binding to "work" for HVM guests. But to do this, we are trying to address conflicting goals. Either we honor the event channel binding by circumventing the IOAPIC emulation, or we faithfully emulate the IOAPIC and circumvent the event channel binding.Well, this doesn't really make sense anyway as is: the IRQ binding has little to do with where the evtchns are handled (I don't think there's any requirement that they both happen on the same CPU). Yes, there is no requirement, but there is a significant latency penalty for redirecting an event channel interrupt through an IRQ routed to a different VCPU. Think 10s of milliseconds delay minimum due to Xen scheduling on a busy node (the current VCPU will not yield unless it is idle). Add to this the fact that almost any significant I/O load on an HVM Windows guest becomes cpu bound quickly (so your scheduling priority is reduced). Our driver writers would like to see support for multiple callback IRQs. Then particular event channel interrupts could be bound to particular IRQs. This would allow PV device interrupts to be distributed intelligently. It would also allow net and block interrupts to be disentangled for Windows PV drivers.You could do a bunch of that just by distributing them from the single callback IRQ. But I suppose it would be nice to move to a one-IRQ-per-evtchn model. You'd need to keep the ABI of course, so you'd need a feature flag or something. Distributing from a single IRQ works OK for Linux, but doesn't work very well for older versions of Windows. For block handling you want to deliver the real interrupts in SCSI miniport context. The network can deal with the interrupt redirect. But the network easily generates the highest interrupt rates and is sensitive to additional latency. So you end up slowing SCSI down with "extra" network interrupts, and slowing the network down with increased interrupt latency. Delivering net and block interrupts independently would avoid these issues. Delivering interrupts to a bus driver and forwarding these to virtual device drivers directly is only an option on newer versions of Windows. We deal pretty much exclusively with HVM guests, do SMP PV environments selectively bind device interrupts to different VCPUs?For true PV you can bind evtchns at will. regards john _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |