|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-changelog] [xen master] xen: credit2: improve debug dump output.
commit 3af86727b8204eee3ba8b280fb4815d21e0f86f1
Author: Dario Faggioli <dario.faggioli@xxxxxxxxxx>
AuthorDate: Wed Jan 18 02:10:14 2017 +0100
Commit: George Dunlap <george.dunlap@xxxxxxxxxx>
CommitDate: Wed Feb 1 14:31:58 2017 +0000
xen: credit2: improve debug dump output.
Scheduling information debug dump for Credit2 is hard
to read as it contains the same information repeated
multiple time in different ways.
In fact, in Credit2, CPUs are grouped in runqueus. Before
this change, for each CPU, we were printing the while
content of the runqueue, as shown below:
CPU[00] sibling=000003, core=0000ff
run: [32767.0] flags=0 cpu=0 credit=-1073741824 [w=0] load=0 (~0%)
1: [0.0] flags=0 cpu=2 credit=3860932 [w=256] load=262144 (~100%)
2: [0.1] flags=0 cpu=2 credit=3859906 [w=256] load=262144 (~100%)
CPU[01] sibling=000003, core=0000ff
run: [32767.1] flags=0 cpu=1 credit=-1073741824 [w=0] load=0 (~0%)
1: [0.0] flags=0 cpu=2 credit=2859840 [w=256] load=262144 (~100%)
2: [0.3] flags=0 cpu=2 credit=-17466062 [w=256] load=262144 (~100%)
CPU[02] sibling=00000c, core=0000ff
run: [0.0] flags=2 cpu=2 credit=1858628 [w=256] load=262144 (~100%)
1: [0.3] flags=0 cpu=2 credit=-17466062 [w=256] load=262144 (~100%)
2: [0.1] flags=0 cpu=2 credit=-23957055 [w=256] load=262144 (~100%)
CPU[03] sibling=00000c, core=0000ff
run: [32767.3] flags=0 cpu=3 credit=-1073741824 [w=0] load=0 (~0%)
1: [0.1] flags=0 cpu=2 credit=-3957055 [w=256] load=262144 (~100%)
2: [0.0] flags=0 cpu=2 credit=-6216254 [w=256] load=262144 (~100%)
CPU[04] sibling=000030, core=0000ff
run: [32767.4] flags=0 cpu=4 credit=-1073741824 [w=0] load=0 (~0%)
1: [0.1] flags=0 cpu=2 credit=3782667 [w=256] load=262144 (~100%)
2: [0.3] flags=0 cpu=2 credit=-16287483 [w=256] load=262144 (~100%)
As it can be seen, all the CPUs print the whole content
of the runqueue they belong to, at the time of their
sampling, and this is cumbersome and hard to interpret!
In new output format we print, for each CPU, only the vCPU
that is running there (if that's not the idle vCPU, in which
case, nothing is printed), while the runqueues content
is printed only once, in a dedicated section.
An example:
CPUs info:
CPU[02] runq=0, sibling=00000c, core=0000ff
run: [0.3] flags=2 cpu=2 credit=8054391 [w=256] load=262144 (~100%)
CPU[14] runq=1, sibling=00c000, core=00ff00
run: [0.4] flags=2 cpu=14 credit=8771420 [w=256] load=262144 (~100%)
... ... ... ... ... ... ... ... ...
Runqueue info:
runqueue 0:
0: [0.1] flags=0 cpu=2 credit=7869771 [w=256] load=262144 (~100%)
1: [0.0] flags=0 cpu=2 credit=7709649 [w=256] load=262144 (~100%)
runqueue 1:
0: [0.5] flags=0 cpu=14 credit=-1188 [w=256] load=262144 (~100%)
Note that there still is risk of inconsistency between
what is printed in the 'Runqueue info:' and in 'CPUs info:'
sections. That is unavoidable, as the relevant locks are
released and re-acquired, around each single operation.
At least, the inconsistency is less severe than before.
Signed-off-by: Dario Faggioli <dario.faggioli@xxxxxxxxxx>
Reviewed-by: George Dunlap <george.dunlap@xxxxxxxxxx>
---
xen/common/sched_credit2.c | 50 +++++++++++++++++++++++++++-------------------
xen/common/schedule.c | 1 +
2 files changed, 30 insertions(+), 21 deletions(-)
diff --git a/xen/common/sched_credit2.c b/xen/common/sched_credit2.c
index b2f2b17..9b063c0 100644
--- a/xen/common/sched_credit2.c
+++ b/xen/common/sched_credit2.c
@@ -2631,50 +2631,35 @@ static void
csched2_dump_pcpu(const struct scheduler *ops, int cpu)
{
struct csched2_private *prv = CSCHED2_PRIV(ops);
- struct list_head *runq, *iter;
struct csched2_vcpu *svc;
unsigned long flags;
spinlock_t *lock;
- int loop;
#define cpustr keyhandler_scratch
/*
* We need both locks:
+ * - we print current, so we need the runqueue lock for this
+ * cpu (the one of the runqueue this cpu is associated to);
* - csched2_dump_vcpu() wants to access domains' weights,
- * which are protected by the private scheduler lock;
- * - we scan through the runqueue, so we need the proper runqueue
- * lock (the one of the runqueue this cpu is associated to).
+ * which are protected by the private scheduler lock.
*/
read_lock_irqsave(&prv->lock, flags);
lock = per_cpu(schedule_data, cpu).schedule_lock;
spin_lock(lock);
- runq = &RQD(ops, cpu)->runq;
-
cpumask_scnprintf(cpustr, sizeof(cpustr), per_cpu(cpu_sibling_mask, cpu));
- printk(" sibling=%s, ", cpustr);
+ printk(" runq=%d, sibling=%s, ", c2r(ops, cpu), cpustr);
cpumask_scnprintf(cpustr, sizeof(cpustr), per_cpu(cpu_core_mask, cpu));
printk("core=%s\n", cpustr);
- /* current VCPU */
+ /* current VCPU (nothing to say if that's the idle vcpu) */
svc = CSCHED2_VCPU(curr_on_cpu(cpu));
- if ( svc )
+ if ( svc && !is_idle_vcpu(svc->vcpu) )
{
printk("\trun: ");
csched2_dump_vcpu(prv, svc);
}
- loop = 0;
- list_for_each( iter, runq )
- {
- svc = __runq_elem(iter);
- if ( svc )
- {
- printk("\t%3d: ", ++loop);
- csched2_dump_vcpu(prv, svc);
- }
- }
-
spin_unlock(lock);
read_unlock_irqrestore(&prv->lock, flags);
#undef cpustr
@@ -2756,6 +2741,29 @@ csched2_dump(const struct scheduler *ops)
}
}
+ printk("Runqueue info:\n");
+ for_each_cpu(i, &prv->active_queues)
+ {
+ struct csched2_runqueue_data *rqd = prv->rqd + i;
+ struct list_head *iter, *runq = &rqd->runq;
+ int loop = 0;
+
+ /* We need the lock to scan the runqueue. */
+ spin_lock(&rqd->lock);
+ printk("runqueue %d:\n", i);
+ list_for_each( iter, runq )
+ {
+ struct csched2_vcpu *svc = __runq_elem(iter);
+
+ if ( svc )
+ {
+ printk("\t%3d: ", loop++);
+ csched2_dump_vcpu(prv, svc);
+ }
+ }
+ spin_unlock(&rqd->lock);
+ }
+
read_unlock_irqrestore(&prv->lock, flags);
#undef cpustr
}
diff --git a/xen/common/schedule.c b/xen/common/schedule.c
index 43b5b99..ed77990 100644
--- a/xen/common/schedule.c
+++ b/xen/common/schedule.c
@@ -1844,6 +1844,7 @@ void schedule_dump(struct cpupool *c)
cpus = &cpupool_free_cpus;
}
+ printk("CPUs info:\n");
for_each_cpu (i, cpus)
{
printk("CPU[%02d] ", i);
--
generated by git-patchbot for /home/xen/git/xen.git#master
_______________________________________________
Xen-changelog mailing list
Xen-changelog@xxxxxxxxxxxxx
https://lists.xenproject.org/xen-changelog
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |