|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH v2 12/16] xenalyze: handle scheduling events
so the trace will show properly decoded info,
rather than just a bunch of hex codes.
Signed-off-by: Dario Faggioli <dario.faggioli@xxxxxxxxxx>
Reviewed-by: Konrad Rzeszutek Wilk <konrad.wilk@xxxxxxxxxx>
---
Cc: George Dunlap <george.dunlap@xxxxxxxxxxxxx>
Cc: Ian Jackson <ian.jackson@xxxxxxxxxxxxx>
Cc: Ian Campbell <ian.campbell@xxxxxxxxxx>
Cc: Wei Liu <wei.liu2@xxxxxxxxxx>
Cc: Olaf Hering <olaf@xxxxxxxxx>
---
Changes from v1:
* SCHED_DOM_{ADD,REM} handling slightly changed, to avoid
confusion with DOM0_DOM_{ADD,REM} (introduced later in
the series);
* '} * r =' turned into '} *r =', as requested
during review.
---
tools/xentrace/xenalyze.c | 156 ++++++++++++++++++++++++++++++++++++---------
1 file changed, 126 insertions(+), 30 deletions(-)
diff --git a/tools/xentrace/xenalyze.c b/tools/xentrace/xenalyze.c
index 6520790..17021f1 100644
--- a/tools/xentrace/xenalyze.c
+++ b/tools/xentrace/xenalyze.c
@@ -1519,27 +1519,6 @@ struct pv_data {
};
/* Sched data */
-
-enum {
- SCHED_DOM_ADD=1,
- SCHED_DOM_REM,
- SCHED_SLEEP,
- SCHED_WAKE,
- SCHED_YIELD,
- SCHED_BLOCK,
- SCHED_SHUTDOWN,
- SCHED_CTL,
- SCHED_ADJDOM,
- SCHED_SWITCH,
- SCHED_S_TIMER_FN,
- SCHED_T_TIMER_FN,
- SCHED_DOM_TIMER_FN,
- SCHED_SWITCH_INFPREV,
- SCHED_SWITCH_INFNEXT,
- SCHED_SHUTDOWN_CODE,
- SCHED_MAX
-};
-
enum {
RUNSTATE_RUNNING=0,
RUNSTATE_RUNNABLE,
@@ -7431,6 +7410,17 @@ no_update:
return;
}
+void dump_sched_switch(struct record_info *ri)
+{
+ struct {
+ unsigned int prev_dom, prev_vcpu, next_dom, next_vcpu;
+ } *r = (typeof(r))ri->d;
+
+ printf(" %s sched_switch prev d%uv%u next d%uv%u\n",
+ ri->dump_header, r->prev_dom, r->prev_vcpu,
+ r->next_dom, r->next_vcpu);
+}
+
void sched_switch_process(struct pcpu_info *p)
{
struct vcpu_data *prev, *next;
@@ -7440,10 +7430,7 @@ void sched_switch_process(struct pcpu_info *p)
} * r = (typeof(r))ri->d;
if(opt.dump_all)
- printf("%s sched_switch prev d%uv%u next d%uv%u\n",
- ri->dump_header,
- r->prev_dom, r->prev_vcpu,
- r->next_dom, r->next_vcpu);
+ dump_sched_switch(ri);
if(r->prev_vcpu > MAX_CPUS)
{
@@ -7559,6 +7546,14 @@ void sched_summary_domain(struct domain_data *d)
}
}
+void dump_sched_vcpu_action(struct record_info *ri, const char *action)
+{
+ struct {
+ unsigned int domid, vcpuid;
+ } *r = (typeof(r))ri->d;
+
+ printf(" %s %s d%uv%u\n", ri->dump_header, action, r->domid, r->vcpuid);
+}
void sched_process(struct pcpu_info *p)
{
@@ -7573,13 +7568,114 @@ void sched_process(struct pcpu_info *p)
default:
process_generic(&p->ri);
}
- } else {
- if(ri->evt.sub == 1)
- sched_runstate_process(p);
- else {
- UPDATE_VOLUME(p, sched_verbose, ri->size);
+ return;
+ }
+
+ if(ri->evt.sub == 1) {
+ /* TRC_SCHED_MIN */
+ sched_runstate_process(p);
+ } else if (ri->evt.sub == 8) {
+ /* TRC_SCHED_VERBOSE */
+ switch(ri->event)
+ {
+ case TRC_SCHED_DOM_ADD:
+ if(opt.dump_all) {
+ struct {
+ unsigned int domid;
+ } *r = (typeof(r))ri->d;
+
+ printf(" %s sched_init_domain d%u\n", ri->dump_header,
r->domid);
+ }
+ break;
+ case TRC_SCHED_DOM_REM:
+ if(opt.dump_all) {
+ struct {
+ unsigned int domid, vcpuid;
+ } *r = (typeof(r))ri->d;
+
+ printf(" %s sched_destroy_domain d%u\n", ri->dump_header,
r->domid);
+ }
+ break;
+ case TRC_SCHED_SLEEP:
+ if(opt.dump_all)
+ dump_sched_vcpu_action(ri, "vcpu_sleep");
+ break;
+ case TRC_SCHED_WAKE:
+ if(opt.dump_all)
+ dump_sched_vcpu_action(ri, "vcpu_wake");
+ break;
+ case TRC_SCHED_YIELD:
+ if(opt.dump_all)
+ dump_sched_vcpu_action(ri, "vcpu_yield");
+ break;
+ case TRC_SCHED_BLOCK:
+ if(opt.dump_all)
+ dump_sched_vcpu_action(ri, "vcpu_block");
+ break;
+ case TRC_SCHED_SHUTDOWN:
+ case TRC_SCHED_SHUTDOWN_CODE:
+ if(opt.dump_all) {
+ struct {
+ unsigned int domid, vcpuid, reason;
+ } *r = (typeof(r))ri->d;
+
+ printf(" %s %s d%uv%u, reason = %u\n", ri->dump_header,
+ ri->event == TRC_SCHED_SHUTDOWN ? "sched_shutdown" :
+ "sched_shutdown_code", r->domid, r->vcpuid, r->reason);
+ }
+ break;
+ case TRC_SCHED_ADJDOM:
+ if(opt.dump_all) {
+ struct {
+ unsigned int domid;
+ } *r = (typeof(r))ri->d;
+
+ printf(" %s sched_adjust d%u\n", ri->dump_header, r->domid);
+ }
+ break;
+ case TRC_SCHED_SWITCH:
+ dump_sched_switch(ri);
+ break;
+ case TRC_SCHED_SWITCH_INFPREV:
+ if(opt.dump_all) {
+ struct {
+ unsigned int domid, runtime;
+ } *r = (typeof(r))ri->d;
+
+ printf(" %s sched_switch prev d%u, run for %u.%uus\n",
+ ri->dump_header, r->domid, r->runtime / 1000,
+ r->runtime % 1000);
+ }
+ break;
+ case TRC_SCHED_SWITCH_INFNEXT:
+ if(opt.dump_all)
+ {
+ struct {
+ unsigned int domid, rsince;
+ int slice;
+ } *r = (typeof(r))ri->d;
+
+ printf(" %s sched_switch next d%u", ri->dump_header, r->domid);
+ if ( r->rsince != 0 )
+ printf(", was runnable for %u.%uus, ", r->rsince / 1000,
+ r->rsince % 1000);
+ if ( r->slice > 0 )
+ printf("next slice %u.%uus\n", r->slice / 1000,
+ r->slice % 1000);
+ printf("\n");
+ }
+ break;
+ case TRC_SCHED_CTL:
+ case TRC_SCHED_S_TIMER_FN:
+ case TRC_SCHED_T_TIMER_FN:
+ case TRC_SCHED_DOM_TIMER_FN:
+ break;
+ default:
process_generic(&p->ri);
}
+ } else {
+ UPDATE_VOLUME(p, sched_verbose, ri->size);
+ process_generic(&p->ri);
}
}
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |