|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH v1 1/4] xen: enabling XL to set per-VCPU parameters of a domain for RTDS scheduler
Add two hypercalls(XEN_DOMCTL_SCHEDOP_getvcpuinfo/putvcpuinfo) to get/set a
domain's
per-VCPU parameters. Hypercalls are handled in function rt_dom_cntl.
Add an array pointer in struct xen_domctl_sched_rtds(an union in struct
xen_domctl_scheduler_op),
which is used for transferring data between tool and hypervisor.
Signed-off-by: Chong Li <chong.li@xxxxxxxxx>
Signed-off-by: Meng Xu <mengxu@xxxxxxxxxxxxx>
Signed-off-by: Sisu Xi <xisisu@xxxxxxxxx>
---
xen/common/sched_rt.c | 64 +++++++++++++++++++++++++++++++++++++++++++++
xen/common/schedule.c | 4 ++-
xen/include/public/domctl.h | 22 +++++++++++++---
3 files changed, 86 insertions(+), 4 deletions(-)
diff --git a/xen/common/sched_rt.c b/xen/common/sched_rt.c
index 7c39a9e..9add5a4 100644
--- a/xen/common/sched_rt.c
+++ b/xen/common/sched_rt.c
@@ -1085,6 +1085,9 @@ rt_dom_cntl(
struct list_head *iter;
unsigned long flags;
int rc = 0;
+ xen_domctl_sched_rtds_params_t *local_sched;
+ int vcpu_index=0;
+ int i;
switch ( op->cmd )
{
@@ -1110,6 +1113,67 @@ rt_dom_cntl(
}
spin_unlock_irqrestore(&prv->lock, flags);
break;
+ case XEN_DOMCTL_SCHEDOP_getvcpuinfo:
+ op->u.rtds.nr_vcpus = 0;
+ spin_lock_irqsave(&prv->lock, flags);
+ list_for_each( iter, &sdom->vcpu )
+ vcpu_index++;
+ spin_unlock_irqrestore(&prv->lock, flags);
+ op->u.rtds.nr_vcpus = vcpu_index;
+ local_sched = xzalloc_array(xen_domctl_sched_rtds_params_t,
+ vcpu_index);
+ if( local_sched == NULL )
+ {
+ return -ENOMEM;
+ }
+ vcpu_index = 0;
+ spin_lock_irqsave(&prv->lock, flags);
+ list_for_each( iter, &sdom->vcpu )
+ {
+ struct rt_vcpu *svc = list_entry(iter, struct rt_vcpu, sdom_elem);
+
+ local_sched[vcpu_index].budget = svc->budget / MICROSECS(1);
+ local_sched[vcpu_index].period = svc->period / MICROSECS(1);
+ local_sched[vcpu_index].index = vcpu_index;
+ vcpu_index++;
+ }
+ spin_unlock_irqrestore(&prv->lock, flags);
+ copy_to_guest(op->u.rtds.vcpus, local_sched, vcpu_index);
+ xfree(local_sched);
+ rc = 0;
+ break;
+ case XEN_DOMCTL_SCHEDOP_putvcpuinfo:
+ local_sched = xzalloc_array(xen_domctl_sched_rtds_params_t,
+ op->u.rtds.nr_vcpus);
+ if( local_sched == NULL )
+ {
+ return -ENOMEM;
+ }
+ copy_from_guest(local_sched, op->u.rtds.vcpus, op->u.rtds.nr_vcpus);
+
+ for( i = 0; i < op->u.rtds.nr_vcpus; i++ )
+ {
+ vcpu_index = 0;
+ spin_lock_irqsave(&prv->lock, flags);
+ list_for_each( iter, &sdom->vcpu )
+ {
+ struct rt_vcpu *svc = list_entry(iter, struct rt_vcpu,
sdom_elem);
+ if ( local_sched[i].index == vcpu_index )
+ {
+ if ( local_sched[i].period <= 0 || local_sched[i].budget
<= 0 )
+ return -EINVAL;
+
+ svc->period = MICROSECS(local_sched[i].period);
+ svc->budget = MICROSECS(local_sched[i].budget);
+ break;
+ }
+ vcpu_index++;
+ }
+ spin_unlock_irqrestore(&prv->lock, flags);
+ }
+ xfree(local_sched);
+ rc = 0;
+ break;
}
return rc;
diff --git a/xen/common/schedule.c b/xen/common/schedule.c
index f5a2e55..f820946 100644
--- a/xen/common/schedule.c
+++ b/xen/common/schedule.c
@@ -1093,7 +1093,9 @@ long sched_adjust(struct domain *d, struct
xen_domctl_scheduler_op *op)
if ( (op->sched_id != DOM2OP(d)->sched_id) ||
((op->cmd != XEN_DOMCTL_SCHEDOP_putinfo) &&
- (op->cmd != XEN_DOMCTL_SCHEDOP_getinfo)) )
+ (op->cmd != XEN_DOMCTL_SCHEDOP_getinfo) &&
+ (op->cmd != XEN_DOMCTL_SCHEDOP_putvcpuinfo) &&
+ (op->cmd != XEN_DOMCTL_SCHEDOP_getvcpuinfo)) )
return -EINVAL;
/* NB: the pluggable scheduler code needs to take care
diff --git a/xen/include/public/domctl.h b/xen/include/public/domctl.h
index 10b51ef..490a6b6 100644
--- a/xen/include/public/domctl.h
+++ b/xen/include/public/domctl.h
@@ -342,6 +342,16 @@ struct xen_domctl_max_vcpus {
typedef struct xen_domctl_max_vcpus xen_domctl_max_vcpus_t;
DEFINE_XEN_GUEST_HANDLE(xen_domctl_max_vcpus_t);
+struct xen_domctl_sched_rtds_params {
+ /* vcpus' info */
+ uint64_t period;
+ uint64_t budget;
+ uint16_t index;
+ uint16_t padding[3];
+};
+typedef struct xen_domctl_sched_rtds_params xen_domctl_sched_rtds_params_t;
+DEFINE_XEN_GUEST_HANDLE(xen_domctl_sched_rtds_params_t);
+
/* XEN_DOMCTL_scheduler_op */
/* Scheduler types. */
@@ -351,9 +361,12 @@ DEFINE_XEN_GUEST_HANDLE(xen_domctl_max_vcpus_t);
#define XEN_SCHEDULER_ARINC653 7
#define XEN_SCHEDULER_RTDS 8
-/* Set or get info? */
+/* Set or get info */
#define XEN_DOMCTL_SCHEDOP_putinfo 0
#define XEN_DOMCTL_SCHEDOP_getinfo 1
+#define XEN_DOMCTL_SCHEDOP_getvcpuinfo 2
+#define XEN_DOMCTL_SCHEDOP_putvcpuinfo 3
+
struct xen_domctl_scheduler_op {
uint32_t sched_id; /* XEN_SCHEDULER_* */
uint32_t cmd; /* XEN_DOMCTL_SCHEDOP_* */
@@ -373,8 +386,11 @@ struct xen_domctl_scheduler_op {
uint16_t weight;
} credit2;
struct xen_domctl_sched_rtds {
- uint32_t period;
- uint32_t budget;
+ uint64_t period;
+ uint64_t budget;
+ XEN_GUEST_HANDLE_64(xen_domctl_sched_rtds_params_t) vcpus;
+ uint16_t nr_vcpus;
+ uint16_t padding[3];
} rtds;
} u;
};
--
1.9.1
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |