[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH RFC V4 0/5] kvm : Paravirt-spinlock support for KVM guests
On 01/16/2012 09:24 PM, Alexander Graf wrote: > This is true in case you're spinning. If on overcommit spinlocks would > instead of spin just yield(), we wouldn't have any vcpu running that's > just waiting for a late ticket. Yes, but the reality is that most spinlocks are held for a short period of time and there's a low likelihood of being preempted while within a spinlock critical section. Therefore if someone else tries to get the spinlock and there's contention, it's always worth spinning for a little while because the lock will likely become free soon. At least that's the case if the lock has low contention (shallow queue depth and not in slow state). Again, maybe it makes sense to never spin for deep queues or already slowstate locks. > We still have an issue finding the point in time when a vcpu could run again, > which is what this whole series is about. My point above was that instead of > doing a count loop, we could just do the normal spin dance and set the > threshold to when we enable the magic to have another spin lock notify us in > the CPU. That way we > > * don't change the uncontended case I don't follow you. What do you mean by "the normal spin dance"? What do you mean by "have another spinlock notify us in the CPU"? Don't change which uncontended case? Do you mean in the locking path? Or the unlock path? Or both? > * can set the threshold on the host, which knows how contended the system is Hm, I'm not convinced that knowing how contended the system is is all that useful overall. What's important is how contended a particular lock is, and what state the current holder is in. If it's not currently running, then knowing the overall system contention would give you some idea about how long you need to wait for it to be rescheduled, but that's getting pretty indirect. I think the "slowpath if preempted while spinning" idea I mentioned in the other mail is probably worth following up, since that give specific actionable information to the guest from the hypervisor. But lots of caveats. [[ A possible mechanism: * register ranges of [er]ips with the hypervisor * each range is paired with a "resched handler block" * if vcpu is preempted within such a range, make sure it is rescheduled in the resched handler block This is obviously akin to the exception mechanism, but it is partially implemented by the hypervisor. It allows the spinlock code to be unchanged from native, but make use of a resched rather than an explicit counter to determine when to slowpath the lock. And it's a nice general mechanism that could be potentially useful elsewhere. Unfortunately, it doesn't change the unlock path at all; it still needs to explicitly test if a VCPU needs to be kicked on unlock. ]] > And since we control what spin locks look like, we can for example always > keep the pointer to it in a specific register so that we can handle > pv_lock_ops.lock_spinning() inside there and fetch all the information we > need from our pt_regs. You've left a pile of parts of an idea lying around, but I'm not sure what shape you intend it to be. >>> Speaking of which - have you benchmarked performance degradation of pv >>> ticket locks on bare metal? Last time I checked, enabling all the PV ops >>> did incur significant slowdown which is why I went though the work to split >>> the individual pv ops features up to only enable a few for KVM guests. >> The whole point of the pv-ticketlock work is to keep the pvops hooks out of >> the locking fast path, so that the calls are only made on the slow path - >> that is, when spinning too long on a contended lock, and when releasing a >> lock that's in a "slow" state. In the fast path case of no contention, >> there are no pvops, and the executed code path is almost identical to native. > You're still changing a tight loop that does nothing (CPU detects it and > saves power) into something that performs calculations. It still has a "pause" instruction in that loop, so that CPU mechanism will still come into play. "pause" doesn't directly "save power"; it's more about making sure that memory dependence cycles are broken and that two competing threads will make similar progress. Besides I'm not sure adding a dec+test to a loop that's already got a memory read and compare in it is adding much in the way of "calculations". J _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |