[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH v3 for Xen 4.6 1/4] xen: enable per-VCPU parameter settings for RTDS scheduler
On Tue, 2015-07-07 at 09:59 +0100, Jan Beulich wrote: > >>> On 29.06.15 at 04:44, <lichong659@xxxxxxxxx> wrote: > > --- a/xen/common/Makefile > > +++ b/xen/common/Makefile > > @@ -31,7 +31,6 @@ obj-y += rbtree.o > > obj-y += rcupdate.o > > obj-y += sched_credit.o > > obj-y += sched_credit2.o > > -obj-y += sched_sedf.o > > obj-y += sched_arinc653.o > > obj-y += sched_rt.o > > obj-y += schedule.o > > Stray change. Or perhaps the file doesn't build anymore, in which case > you should instead have stated that the patch is dependent upon the > series removing SEDF. > > > @@ -1157,8 +1158,75 @@ rt_dom_cntl( > > list_for_each( iter, &sdom->vcpu ) > > { > > struct rt_vcpu * svc = list_entry(iter, struct rt_vcpu, > > sdom_elem); > > - svc->period = MICROSECS(op->u.rtds.period); /* transfer to > > nanosec */ > > - svc->budget = MICROSECS(op->u.rtds.budget); > > + svc->period = MICROSECS(op->u.d.rtds.period); /* transfer to > > nanosec */ > > + svc->budget = MICROSECS(op->u.d.rtds.budget); > > + } > > + spin_unlock_irqrestore(&prv->lock, flags); > > + break; > > + case XEN_DOMCTL_SCHEDOP_getvcpuinfo: > > + spin_lock_irqsave(&prv->lock, flags); > > + for( index = 0; index < op->u.v.nr_vcpus; index++ ) > > Coding style (more further down). > > > + { > > + if ( copy_from_guest_offset(&local_sched, > > + op->u.v.vcpus, index, 1) ) > > Indentation. > > > + { > > + rc = -EFAULT; > > + break; > > + } > > + if ( local_sched.vcpuid >= d->max_vcpus > > + || d->vcpu[local_sched.vcpuid] == NULL ) > > || belongs at the end of the first line. Indentation. > > > + { > > + rc = -EINVAL; > > + break; > > + } > > + svc = rt_vcpu(d->vcpu[local_sched.vcpuid]); > > + > > + local_sched.vcpuid = svc->vcpu->vcpu_id; > > Why? If at all, this should be an ASSERT(). > > > + local_sched.s.rtds.budget = svc->budget / MICROSECS(1); > > + local_sched.s.rtds.period = svc->period / MICROSECS(1); > > + if( index >= op->u.v.nr_vcpus ) /* not enough guest buffer*/ > > Impossible due to the containing loop's condition. > > > + { > > + rc = -ENOBUFS; > > + break; > > + } > > + if ( copy_to_guest_offset(op->u.v.vcpus, index, > > __copy_to_guest_offset() > > > + case XEN_DOMCTL_SCHEDOP_putvcpuinfo: > > + spin_lock_irqsave(&prv->lock, flags); > > + for( index = 0; index < op->u.v.nr_vcpus; index++ ) > > + { > > + if ( copy_from_guest_offset(&local_sched, > > + op->u.v.vcpus, index, 1) ) > > + { > > + rc = -EFAULT; > > + break; > > + } > > + if ( local_sched.vcpuid >= d->max_vcpus > > + || d->vcpu[local_sched.vcpuid] == NULL ) > > + { > > + rc = -EINVAL; > > + break; > > + } > > + svc = rt_vcpu(d->vcpu[local_sched.vcpuid]); > > + svc->period = MICROSECS(local_sched.s.rtds.period); > > + svc->budget = MICROSECS(local_sched.s.rtds.budget); > > Are all input values valid here? > > > --- a/xen/common/schedule.c > > +++ b/xen/common/schedule.c > > @@ -65,7 +65,6 @@ DEFINE_PER_CPU(struct schedule_data, schedule_data); > > DEFINE_PER_CPU(struct scheduler *, scheduler); > > > > static const struct scheduler *schedulers[] = { > > - &sched_sedf_def, > > &sched_credit_def, > > &sched_credit2_def, > > &sched_arinc653_def, > > See above. > > > @@ -1054,7 +1053,9 @@ long sched_adjust(struct domain *d, struct > > xen_domctl_scheduler_op *op) > > > > if ( (op->sched_id != DOM2OP(d)->sched_id) || > > ((op->cmd != XEN_DOMCTL_SCHEDOP_putinfo) && > > - (op->cmd != XEN_DOMCTL_SCHEDOP_getinfo)) ) > > + (op->cmd != XEN_DOMCTL_SCHEDOP_getinfo) && > > + (op->cmd != XEN_DOMCTL_SCHEDOP_putvcpuinfo) && > > + (op->cmd != XEN_DOMCTL_SCHEDOP_getvcpuinfo)) ) > > return -EINVAL; > > Convert to switch() please. > > > --- a/xen/include/public/domctl.h > > +++ b/xen/include/public/domctl.h > > @@ -330,31 +330,59 @@ DEFINE_XEN_GUEST_HANDLE(xen_domctl_max_vcpus_t); > > #define XEN_SCHEDULER_ARINC653 7 > > #define XEN_SCHEDULER_RTDS 8 > > > > +typedef struct xen_domctl_sched_sedf { > > + uint64_aligned_t period; > > + uint64_aligned_t slice; > > + uint64_aligned_t latency; > > + uint32_t extratime; > > + uint32_t weight; > > +} xen_domctl_sched_sedf_t; > > Indentation. > > > +typedef union xen_domctl_schedparam { > > + xen_domctl_sched_sedf_t sedf; > > + xen_domctl_sched_credit_t credit; > > + xen_domctl_sched_credit2_t credit2; > > + xen_domctl_sched_rtds_t rtds; > > +} xen_domctl_schedparam_t; > > I don't see the need for this extra wrapper type. Nor do I see the > need for the typedef here and above - they're generally only > created if you want to also define a matching guest handle type. > > > +typedef struct xen_domctl_schedparam_vcpu { > > + union { > > + xen_domctl_sched_credit_t credit; > > + xen_domctl_sched_credit2_t credit2; > > + xen_domctl_sched_rtds_t rtds; > > + } s; > > + uint16_t vcpuid; > > Explicit padding please. > > > +} xen_domctl_schedparam_vcpu_t; > > +DEFINE_XEN_GUEST_HANDLE(xen_domctl_schedparam_vcpu_t); > > + > > /* Set or get info? */ > > #define XEN_DOMCTL_SCHEDOP_putinfo 0 > > #define XEN_DOMCTL_SCHEDOP_getinfo 1 > > +#define XEN_DOMCTL_SCHEDOP_putvcpuinfo 2 > > +#define XEN_DOMCTL_SCHEDOP_getvcpuinfo 3 > > struct xen_domctl_scheduler_op { > > uint32_t sched_id; /* XEN_SCHEDULER_* */ > > uint32_t cmd; /* XEN_DOMCTL_SCHEDOP_* */ > > union { > > - struct xen_domctl_sched_sedf { > > - uint64_aligned_t period; > > - uint64_aligned_t slice; > > - uint64_aligned_t latency; > > - uint32_t extratime; > > - uint32_t weight; > > - } sedf; > > - struct xen_domctl_sched_credit { > > - uint16_t weight; > > - uint16_t cap; > > - } credit; > > - struct xen_domctl_sched_credit2 { > > - uint16_t weight; > > - } credit2; > > - struct xen_domctl_sched_rtds { > > - uint32_t period; > > - uint32_t budget; > > - } rtds; > > + xen_domctl_schedparam_t d; > > With this type gone I'm not even sure we need to wrap this in > another union; not doing so would eliminate some of the other > changes in this patch. > So, Jan, just to be sure, do you mean (apart from the explicit padding) something like this (attached, also)? diff --git a/xen/include/public/domctl.h b/xen/include/public/domctl.h index bc45ea5..8210ecb 100644 --- a/xen/include/public/domctl.h +++ b/xen/include/public/domctl.h @@ -330,31 +330,56 @@ DEFINE_XEN_GUEST_HANDLE(xen_domctl_max_vcpus_t); #define XEN_SCHEDULER_ARINC653 7 #define XEN_SCHEDULER_RTDS 8 +struct xen_domctl_sched_sedf { + uint64_aligned_t period; + uint64_aligned_t slice; + uint64_aligned_t latency; + uint32_t extratime; + uint32_t weight; +}; + +struct xen_domctl_sched_credit { + uint16_t weight; + uint16_t cap; +}; + +struct xen_domctl_sched_credit2 { + uint16_t weight; +}; + +struct xen_domctl_sched_rtds { + uint32_t period; + uint32_t budget; +}; + +typedef struct xen_domctl_schedparam_vcpu { + union { + struct xen_domctl_sched_sedf sedf; + struct xen_domctl_sched_credit credit; + struct xen_domctl_sched_credit2 credit2; + struct xen_domctl_sched_rtds rtds; + } s; + uint16_t vcpuid; +} xen_domctl_schedparam_vcpu_t; +DEFINE_XEN_GUEST_HANDLE(xen_domctl_schedparam_vcpu_t); + /* Set or get info? */ #define XEN_DOMCTL_SCHEDOP_putinfo 0 #define XEN_DOMCTL_SCHEDOP_getinfo 1 +#define XEN_DOMCTL_SCHEDOP_putvcpuinfo 2 +#define XEN_DOMCTL_SCHEDOP_getvcpuinfo 3 struct xen_domctl_scheduler_op { uint32_t sched_id; /* XEN_SCHEDULER_* */ uint32_t cmd; /* XEN_DOMCTL_SCHEDOP_* */ union { - struct xen_domctl_sched_sedf { - uint64_aligned_t period; - uint64_aligned_t slice; - uint64_aligned_t latency; - uint32_t extratime; - uint32_t weight; - } sedf; - struct xen_domctl_sched_credit { - uint16_t weight; - uint16_t cap; - } credit; - struct xen_domctl_sched_credit2 { - uint16_t weight; - } credit2; - struct xen_domctl_sched_rtds { - uint32_t period; - uint32_t budget; - } rtds; + struct xen_domctl_sched_sedf sedf; + struct xen_domctl_sched_credit credit; + struct xen_domctl_sched_credit2 credit2; + struct xen_domctl_sched_rtds rtds; + struct { + XEN_GUEST_HANDLE_64(xen_domctl_schedparam_vcpu_t) vcpus; + uint16_t nr_vcpus; + } v; } u; }; typedef struct xen_domctl_scheduler_op xen_domctl_scheduler_op_t; If yes, I'd be fine with it too (George?) Thanks and Regards, Dario -- <<This happens because I choose it to happen!>> (Raistlin Majere) ----------------------------------------------------------------- Dario Faggioli, Ph.D, http://about.me/dario.faggioli Senior Software Engineer, Citrix Systems R&D Ltd., Cambridge (UK) Attachment:
rtds-domctl.patch Attachment:
signature.asc _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |