[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH v10 11/20] x86/VPMU: Interface for setting PMU mode and flags
>>> On 11.09.14 at 16:12, <boris.ostrovsky@xxxxxxxxxx> wrote: > On 09/11/2014 02:44 AM, Jan Beulich wrote: >>>>> On 10.09.14 at 19:37, <boris.ostrovsky@xxxxxxxxxx> wrote: >>> On 09/10/2014 11:05 AM, Jan Beulich wrote: >>>>>>> On 04.09.14 at 05:41, <boris.ostrovsky@xxxxxxxxxx> wrote: >>>>> + cont_wait: >>>>> + /* >>>>> + * Note that we may fail here if a CPU is hot-(un)plugged while we >>>>> are >>>>> + * waiting. We will then time out. >>>>> + */ >>>>> + while ( atomic_read(&vpmu_sched_counter) != allbutself_num ) >>>>> + { >>>>> + /* Give up after 5 seconds */ >>>>> + if ( NOW() > start + SECONDS(5) ) >>>>> + { >>>>> + printk(XENLOG_WARNING >>>>> + "vpmu_force_context_switch: failed to sync\n"); >>>>> + ret = -EBUSY; >>>>> + break; >>>>> + } >>>>> + cpu_relax(); >>>>> + if ( hypercall_preempt_check() ) >>>>> + return hypercall_create_continuation( >>>>> + __HYPERVISOR_xenpmu_op, "ih", XENPMU_mode_set, arg); >>>>> + } >>>> I wouldn't complain about this not being synchronized with CPU >>>> hotplug if there wasn't this hypercall continuation and relatively >>>> long timeout. Much of the state you latch in static variables will >>>> cause this operation to time out if in between a CPU got brought >>>> down. >>> It seemed to me that if we were to correctly deal with CPU hotplug it >>> would add a bit too much complexity to the code. So I felt that letting >>> the operation timeout would be a better way out. >> The please at least add a code comment making this explicit to >> future readers. > > Is the comment above 'while' keyword not sufficient? Oh, it is of course. Must have not scrolled back enough... >>>> And as already alluded to, all this looks rather fragile anyway, >>>> even if I can't immediately spot any problems with it anymore. >>> The continuation is really a carry-over from earlier patch version when >>> I had double loops over domain and VCPUs to explicitly unload VPMUs. At >>> that time Andrew pointed out that these loops may take really long time >>> and so I added continuations. >>> >>> Now that I changed that after realizing that having each PCPU go through >>> a context switch is sufficient perhaps I don't need it any longer. Is >>> the worst case scenario of being stuck here for 5 seconds (chosen >>> somewhat arbitrary) acceptable without continuation? >> 5 seconds is _way_ too long for doing this without continuation. > > Then I am also adding back your other comment from this thread > > > > +long do_xenpmu_op(int op, XEN_GUEST_HANDLE_PARAM(xen_pmu_params_t) arg) > > > +{ > > > + int ret = -EINVAL; > > > + xen_pmu_params_t pmu_params; > > > + > > > + switch ( op ) > > > + { > > > + case XENPMU_mode_set: > > > + { > > > + static DEFINE_SPINLOCK(xenpmu_mode_lock); > > > + uint32_t current_mode; > > > + > > > + if ( !is_control_domain(current->domain) ) > > > + return -EPERM; > > > + > > > + if ( copy_from_guest(&pmu_params, arg, 1) ) > > > + return -EFAULT; > > > + > > > + if ( pmu_params.val & ~XENPMU_MODE_SELF ) > > > + return -EINVAL; > > > + > > > + /* > > > + * Return error is someone else is in the middle of changing > mode --- > > > + * this is most likely indication of two system administrators > > > + * working against each other > > > + */ > > > + if ( !spin_trylock(&xenpmu_mode_lock) ) > > > + return -EAGAIN; > > > > So what happens if you can't take the lock in a continuation? If > > returning -EAGAIN in that case is not a problem, what do you > > need the continuation for in the first place? > > EAGAIN this case means that the caller was not able to initiate the > operation. Continuation will allow the caller to finish operation in > progress. But that's only what you want, not what the code does. Also now that I look again I don't think the comment really applies to this if(). Jan _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |