[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH 2/4] sysctl/libxl: Add interface for returning IO topology data
Make XEN_SYSCTL_topologyinfo more generic so that it can return both CPU and IO topology (support for returning the latter is added in the subsequent patch) To do so move (and rename) previously used cpu_to_core/cpu_to_socket/ cpu_to_node into struct xen_sysctl_cputopo and move it, together with newly added struct xen_sysctl_iotopo, to xen_sysctl_topologyinfo. Add libxl_get_topology() to handle new interface and modify libxl_get_cpu_topology() to be a wrapper around it. Adjust xenpm and python's low-level C libraries for new interface. This change requires bumping XEN_SYSCTL_INTERFACE_VERSION Signed-off-by: Boris Ostrovsky <boris.ostrovsky@xxxxxxxxxx> --- tools/libxl/libxl.c | 79 ++++++++++++++++++++++++--------------- tools/libxl/libxl.h | 4 ++ tools/libxl/libxl_types.idl | 12 ++++++ tools/libxl/libxl_utils.c | 6 +++ tools/misc/xenpm.c | 64 +++++++++++++------------------ tools/python/xen/lowlevel/xc/xc.c | 38 +++++++------------ xen/common/sysctl.c | 43 ++++++++++++--------- xen/include/public/sysctl.h | 36 +++++++++++++----- 8 files changed, 162 insertions(+), 120 deletions(-) diff --git a/tools/libxl/libxl.c b/tools/libxl/libxl.c index de23fec..58cdc3b 100644 --- a/tools/libxl/libxl.c +++ b/tools/libxl/libxl.c @@ -5056,11 +5056,37 @@ int libxl_get_physinfo(libxl_ctx *ctx, libxl_physinfo *physinfo) libxl_cputopology *libxl_get_cpu_topology(libxl_ctx *ctx, int *nb_cpu_out) { GC_INIT(ctx); + libxl_cputopology *cputopology = NULL; + libxl_topology *topo; + int i; + + topo = libxl_get_topology(ctx); + if (topo == NULL) { + *nb_cpu_out = 0; + goto out; + } + + cputopology = libxl__zalloc(NOGC, sizeof(libxl_cputopology) * topo->cpu_num); + for (i = 0; i < topo->cpu_num; i++) { + cputopology[i].core = topo->cpu[i].core; + cputopology[i].socket = topo->cpu[i].socket; + cputopology[i].node = topo->cpu[i].node; + } + *nb_cpu_out = topo->cpu_num; + + libxl_topology_free(topo); + + out: + GC_FREE; + return cputopology; +} + +libxl_topology *libxl_get_topology(libxl_ctx *ctx) +{ + GC_INIT(ctx); xc_topologyinfo_t tinfo; - DECLARE_HYPERCALL_BUFFER(xc_cpu_to_core_t, coremap); - DECLARE_HYPERCALL_BUFFER(xc_cpu_to_socket_t, socketmap); - DECLARE_HYPERCALL_BUFFER(xc_cpu_to_node_t, nodemap); - libxl_cputopology *ret = NULL; + DECLARE_HYPERCALL_BUFFER(xen_sysctl_cputopo_t, cputopo); + libxl_topology *ret = NULL; int i; int max_cpus; @@ -5072,48 +5098,39 @@ libxl_cputopology *libxl_get_cpu_topology(libxl_ctx *ctx, int *nb_cpu_out) goto out; } - coremap = xc_hypercall_buffer_alloc - (ctx->xch, coremap, sizeof(*coremap) * max_cpus); - socketmap = xc_hypercall_buffer_alloc - (ctx->xch, socketmap, sizeof(*socketmap) * max_cpus); - nodemap = xc_hypercall_buffer_alloc - (ctx->xch, nodemap, sizeof(*nodemap) * max_cpus); - if ((coremap == NULL) || (socketmap == NULL) || (nodemap == NULL)) { + cputopo = xc_hypercall_buffer_alloc(ctx->xch, cputopo, + sizeof(*cputopo) * max_cpus); + if (cputopo == NULL) { LIBXL__LOG_ERRNOVAL(ctx, XTL_ERROR, ENOMEM, "Unable to allocate hypercall arguments"); goto fail; } - - set_xen_guest_handle(tinfo.cpu_to_core, coremap); - set_xen_guest_handle(tinfo.cpu_to_socket, socketmap); - set_xen_guest_handle(tinfo.cpu_to_node, nodemap); + set_xen_guest_handle(tinfo.cputopo, cputopo); tinfo.max_cpu_index = max_cpus - 1; + + set_xen_guest_handle(tinfo.iotopo, HYPERCALL_BUFFER_NULL); + if (xc_topologyinfo(ctx->xch, &tinfo) != 0) { LIBXL__LOG_ERRNO(ctx, XTL_ERROR, "Topology info hypercall failed"); goto fail; } - if (tinfo.max_cpu_index < max_cpus - 1) - max_cpus = tinfo.max_cpu_index + 1; + ret = libxl__zalloc(NOGC, sizeof(*ret)); + ret->cpu_num = tinfo.max_cpu_index + 1; + ret->cpu = libxl__zalloc(NOGC, sizeof(*(ret->cpu)) * ret->cpu_num); - ret = libxl__zalloc(NOGC, sizeof(libxl_cputopology) * max_cpus); - - for (i = 0; i < max_cpus; i++) { -#define V(map, i) (map[i] == INVALID_TOPOLOGY_ID) ? \ - LIBXL_CPUTOPOLOGY_INVALID_ENTRY : map[i] - ret[i].core = V(coremap, i); - ret[i].socket = V(socketmap, i); - ret[i].node = V(nodemap, i); + for (i = 0; i < ret->cpu_num; i++) { +#define V(map, i) ( cputopo[i].map == INVALID_TOPOLOGY_ID) ? \ + LIBXL_CPUTOPOLOGY_INVALID_ENTRY : cputopo[i].map + ret->cpu[i].core = V(core, i); + ret->cpu[i].socket = V(socket, i); + ret->cpu[i].node = V(node, i); #undef V } -fail: - xc_hypercall_buffer_free(ctx->xch, coremap); - xc_hypercall_buffer_free(ctx->xch, socketmap); - xc_hypercall_buffer_free(ctx->xch, nodemap); + fail: + xc_hypercall_buffer_free(ctx->xch, cputopo); - if (ret) - *nb_cpu_out = max_cpus; out: GC_FREE; return ret; diff --git a/tools/libxl/libxl.h b/tools/libxl/libxl.h index c3451bd..5ab008d 100644 --- a/tools/libxl/libxl.h +++ b/tools/libxl/libxl.h @@ -1060,6 +1060,10 @@ void libxl_vminfo_list_free(libxl_vminfo *list, int nb_vm); libxl_cputopology *libxl_get_cpu_topology(libxl_ctx *ctx, int *nb_cpu_out); void libxl_cputopology_list_free(libxl_cputopology *, int nb_cpu); +#define LIBXL_TOPOLOGY_INVALID_ENTRY (~(uint32_t)0) +libxl_topology *libxl_get_topology(libxl_ctx *ctx); +void libxl_topology_free(libxl_topology *); + #define LIBXL_NUMAINFO_INVALID_ENTRY (~(uint32_t)0) libxl_numainfo *libxl_get_numainfo(libxl_ctx *ctx, int *nr); void libxl_numainfo_list_free(libxl_numainfo *, int nr); diff --git a/tools/libxl/libxl_types.idl b/tools/libxl/libxl_types.idl index f7fc695..673b273 100644 --- a/tools/libxl/libxl_types.idl +++ b/tools/libxl/libxl_types.idl @@ -642,6 +642,18 @@ libxl_cputopology = Struct("cputopology", [ ("node", uint32), ], dir=DIR_OUT) +libxl_iotopology = Struct("iotopology", [ + ("seg", uint16), + ("bus", uint8), + ("devfn", uint8), + ("node", uint32), + ], dir=DIR_OUT) + +libxl_topology = Struct("topology", [ + ("cpu", Array(libxl_cputopology, "cpu_num")), + ("dev", Array(libxl_iotopology, "dev_num")), + ], dir=DIR_OUT) + libxl_sched_credit_params = Struct("sched_credit_params", [ ("tslice_ms", integer), ("ratelimit_us", integer), diff --git a/tools/libxl/libxl_utils.c b/tools/libxl/libxl_utils.c index 58df4f3..70c21a2 100644 --- a/tools/libxl/libxl_utils.c +++ b/tools/libxl/libxl_utils.c @@ -859,6 +859,12 @@ void libxl_cputopology_list_free(libxl_cputopology *list, int nr) free(list); } +void libxl_topology_free(libxl_topology *tinfo) +{ + libxl_topology_dispose(tinfo); + free(tinfo); +} + void libxl_numainfo_list_free(libxl_numainfo *list, int nr) { int i; diff --git a/tools/misc/xenpm.c b/tools/misc/xenpm.c index e43924c..8fd51d2 100644 --- a/tools/misc/xenpm.c +++ b/tools/misc/xenpm.c @@ -355,16 +355,11 @@ static void signal_int_handler(int signo) int i, j, k; struct timeval tv; int cx_cap = 0, px_cap = 0; - DECLARE_HYPERCALL_BUFFER(uint32_t, cpu_to_core); - DECLARE_HYPERCALL_BUFFER(uint32_t, cpu_to_socket); - DECLARE_HYPERCALL_BUFFER(uint32_t, cpu_to_node); + DECLARE_HYPERCALL_BUFFER(xen_sysctl_cputopo_t, cputopo); xc_topologyinfo_t info = { 0 }; - cpu_to_core = xc_hypercall_buffer_alloc(xc_handle, cpu_to_core, sizeof(*cpu_to_core) * MAX_NR_CPU); - cpu_to_socket = xc_hypercall_buffer_alloc(xc_handle, cpu_to_socket, sizeof(*cpu_to_socket) * MAX_NR_CPU); - cpu_to_node = xc_hypercall_buffer_alloc(xc_handle, cpu_to_node, sizeof(*cpu_to_node) * MAX_NR_CPU); - - if ( cpu_to_core == NULL || cpu_to_socket == NULL || cpu_to_node == NULL ) + cputopo = xc_hypercall_buffer_alloc(xc_handle, cputopo, sizeof(*cputopo) * MAX_NR_CPU); + if ( cputopo == NULL ) { fprintf(stderr, "failed to allocate hypercall buffers\n"); goto out; @@ -448,11 +443,11 @@ static void signal_int_handler(int signo) printf(" Avg freq\t%d\tKHz\n", avgfreq[i]); } - set_xen_guest_handle(info.cpu_to_core, cpu_to_core); - set_xen_guest_handle(info.cpu_to_socket, cpu_to_socket); - set_xen_guest_handle(info.cpu_to_node, cpu_to_node); + set_xen_guest_handle(info.cputopo, cputopo); info.max_cpu_index = MAX_NR_CPU - 1; + set_xen_guest_handle(info.iotopo, HYPERCALL_BUFFER_NULL); + if ( cx_cap && !xc_topologyinfo(xc_handle, &info) ) { uint32_t socket_ids[MAX_NR_CPU]; @@ -465,8 +460,8 @@ static void signal_int_handler(int signo) /* check validity */ for ( i = 0; i <= info.max_cpu_index; i++ ) { - if ( cpu_to_core[i] == INVALID_TOPOLOGY_ID || - cpu_to_socket[i] == INVALID_TOPOLOGY_ID ) + if ( cputopo[i].core == INVALID_TOPOLOGY_ID || + cputopo[i].socket == INVALID_TOPOLOGY_ID ) break; } if ( i > info.max_cpu_index ) @@ -475,20 +470,20 @@ static void signal_int_handler(int signo) for ( i = 0; i <= info.max_cpu_index; i++ ) { for ( j = 0; j < socket_nr; j++ ) - if ( cpu_to_socket[i] == socket_ids[j] ) + if ( cputopo[i].socket == socket_ids[j] ) break; if ( j == socket_nr ) { - socket_ids[j] = cpu_to_socket[i]; + socket_ids[j] = cputopo[i].socket; socket_nr++; } for ( j = 0; j < core_nr; j++ ) - if ( cpu_to_core[i] == core_ids[j] ) + if ( cputopo[i].core == core_ids[j] ) break; if ( j == core_nr ) { - core_ids[j] = cpu_to_core[i]; + core_ids[j] = cputopo[i].core; core_nr++; } } @@ -501,7 +496,7 @@ static void signal_int_handler(int signo) for ( j = 0; j <= info.max_cpu_index; j++ ) { - if ( cpu_to_socket[j] == socket_ids[i] ) + if ( cputopo[j].socket == socket_ids[i] ) break; } printf("\nSocket %d\n", socket_ids[i]); @@ -520,8 +515,8 @@ static void signal_int_handler(int signo) { for ( j = 0; j <= info.max_cpu_index; j++ ) { - if ( cpu_to_socket[j] == socket_ids[i] && - cpu_to_core[j] == core_ids[k] ) + if ( cputopo[j].socket == socket_ids[i] && + cputopo[j].core == core_ids[k] ) break; } printf("\t Core %d CPU %d\n", core_ids[k], j); @@ -556,9 +551,7 @@ static void signal_int_handler(int signo) free(sum); free(avgfreq); out: - xc_hypercall_buffer_free(xc_handle, cpu_to_core); - xc_hypercall_buffer_free(xc_handle, cpu_to_socket); - xc_hypercall_buffer_free(xc_handle, cpu_to_node); + xc_hypercall_buffer_free(xc_handle, cputopo); xc_interface_close(xc_handle); exit(0); } @@ -965,27 +958,22 @@ void scaling_governor_func(int argc, char *argv[]) void cpu_topology_func(int argc, char *argv[]) { - DECLARE_HYPERCALL_BUFFER(uint32_t, cpu_to_core); - DECLARE_HYPERCALL_BUFFER(uint32_t, cpu_to_socket); - DECLARE_HYPERCALL_BUFFER(uint32_t, cpu_to_node); + DECLARE_HYPERCALL_BUFFER(xen_sysctl_cputopo_t, cputopo); xc_topologyinfo_t info = { 0 }; int i, rc = ENOMEM; - cpu_to_core = xc_hypercall_buffer_alloc(xc_handle, cpu_to_core, sizeof(*cpu_to_core) * MAX_NR_CPU); - cpu_to_socket = xc_hypercall_buffer_alloc(xc_handle, cpu_to_socket, sizeof(*cpu_to_socket) * MAX_NR_CPU); - cpu_to_node = xc_hypercall_buffer_alloc(xc_handle, cpu_to_node, sizeof(*cpu_to_node) * MAX_NR_CPU); - - if ( cpu_to_core == NULL || cpu_to_socket == NULL || cpu_to_node == NULL ) + cputopo = xc_hypercall_buffer_alloc(xc_handle, cputopo, sizeof(*cputopo) * MAX_NR_CPU); + if ( cputopo == NULL ) { fprintf(stderr, "failed to allocate hypercall buffers\n"); goto out; } - set_xen_guest_handle(info.cpu_to_core, cpu_to_core); - set_xen_guest_handle(info.cpu_to_socket, cpu_to_socket); - set_xen_guest_handle(info.cpu_to_node, cpu_to_node); + set_xen_guest_handle(info.cputopo, cputopo); info.max_cpu_index = MAX_NR_CPU-1; + set_xen_guest_handle(info.iotopo, HYPERCALL_BUFFER_NULL); + if ( xc_topologyinfo(xc_handle, &info) ) { rc = errno; @@ -1000,16 +988,14 @@ void cpu_topology_func(int argc, char *argv[]) printf("CPU\tcore\tsocket\tnode\n"); for ( i = 0; i <= info.max_cpu_index; i++ ) { - if ( cpu_to_core[i] == INVALID_TOPOLOGY_ID ) + if ( cputopo[i].core == INVALID_TOPOLOGY_ID ) continue; printf("CPU%d\t %d\t %d\t %d\n", - i, cpu_to_core[i], cpu_to_socket[i], cpu_to_node[i]); + i, cputopo[i].core, cputopo[i].socket, cputopo[i].node); } rc = 0; out: - xc_hypercall_buffer_free(xc_handle, cpu_to_core); - xc_hypercall_buffer_free(xc_handle, cpu_to_socket); - xc_hypercall_buffer_free(xc_handle, cpu_to_node); + xc_hypercall_buffer_free(xc_handle, cputopo); if ( rc ) exit(rc); } diff --git a/tools/python/xen/lowlevel/xc/xc.c b/tools/python/xen/lowlevel/xc/xc.c index d95d459..1f9252a 100644 --- a/tools/python/xen/lowlevel/xc/xc.c +++ b/tools/python/xen/lowlevel/xc/xc.c @@ -1226,25 +1226,17 @@ static PyObject *pyxc_topologyinfo(XcObject *self) int i, max_cpu_index; PyObject *ret_obj = NULL; PyObject *cpu_to_core_obj, *cpu_to_socket_obj, *cpu_to_node_obj; - DECLARE_HYPERCALL_BUFFER(xc_cpu_to_core_t, coremap); - DECLARE_HYPERCALL_BUFFER(xc_cpu_to_socket_t, socketmap); - DECLARE_HYPERCALL_BUFFER(xc_cpu_to_node_t, nodemap); - coremap = xc_hypercall_buffer_alloc(self->xc_handle, coremap, sizeof(*coremap) * (MAX_CPU_INDEX+1)); - if ( coremap == NULL ) - goto out; - socketmap = xc_hypercall_buffer_alloc(self->xc_handle, socketmap, sizeof(*socketmap) * (MAX_CPU_INDEX+1)); - if ( socketmap == NULL ) - goto out; - nodemap = xc_hypercall_buffer_alloc(self->xc_handle, nodemap, sizeof(*nodemap) * (MAX_CPU_INDEX+1)); - if ( nodemap == NULL ) - goto out; + DECLARE_HYPERCALL_BUFFER(xen_sysctl_cputopo_t, cputopo); - set_xen_guest_handle(tinfo.cpu_to_core, coremap); - set_xen_guest_handle(tinfo.cpu_to_socket, socketmap); - set_xen_guest_handle(tinfo.cpu_to_node, nodemap); + cputopo = xc_hypercall_buffer_alloc(self->xc_handle, cputopo, sizeof(*cputopo) * (MAX_CPU_INDEX+1)); + if ( cputopo == NULL ) + goto out; + set_xen_guest_handle(tinfo.cputopo, cputopo); tinfo.max_cpu_index = MAX_CPU_INDEX; + set_xen_guest_handle(tinfo.iotopo, HYPERCALL_BUFFER_NULL); + if ( xc_topologyinfo(self->xc_handle, &tinfo) != 0 ) goto out; @@ -1258,35 +1250,35 @@ static PyObject *pyxc_topologyinfo(XcObject *self) cpu_to_node_obj = PyList_New(0); for ( i = 0; i <= max_cpu_index; i++ ) { - if ( coremap[i] == INVALID_TOPOLOGY_ID ) + if ( cputopo[i].core == INVALID_TOPOLOGY_ID ) { PyList_Append(cpu_to_core_obj, Py_None); } else { - PyObject *pyint = PyInt_FromLong(coremap[i]); + PyObject *pyint = PyInt_FromLong(cputopo[i].core); PyList_Append(cpu_to_core_obj, pyint); Py_DECREF(pyint); } - if ( socketmap[i] == INVALID_TOPOLOGY_ID ) + if ( cputopo[i].socket == INVALID_TOPOLOGY_ID ) { PyList_Append(cpu_to_socket_obj, Py_None); } else { - PyObject *pyint = PyInt_FromLong(socketmap[i]); + PyObject *pyint = PyInt_FromLong(cputopo[i].socket); PyList_Append(cpu_to_socket_obj, pyint); Py_DECREF(pyint); } - if ( nodemap[i] == INVALID_TOPOLOGY_ID ) + if ( cputopo[i].node == INVALID_TOPOLOGY_ID ) { PyList_Append(cpu_to_node_obj, Py_None); } else { - PyObject *pyint = PyInt_FromLong(nodemap[i]); + PyObject *pyint = PyInt_FromLong(cputopo[i].node); PyList_Append(cpu_to_node_obj, pyint); Py_DECREF(pyint); } @@ -1304,9 +1296,7 @@ static PyObject *pyxc_topologyinfo(XcObject *self) Py_DECREF(cpu_to_node_obj); out: - xc_hypercall_buffer_free(self->xc_handle, coremap); - xc_hypercall_buffer_free(self->xc_handle, socketmap); - xc_hypercall_buffer_free(self->xc_handle, nodemap); + xc_hypercall_buffer_free(self->xc_handle, cputopo); return ret_obj ? ret_obj : pyxc_error_to_exception(self->xc_handle); #undef MAX_CPU_INDEX } diff --git a/xen/common/sysctl.c b/xen/common/sysctl.c index 0cb6ee1..d4dc8ed 100644 --- a/xen/common/sysctl.c +++ b/xen/common/sysctl.c @@ -327,32 +327,41 @@ long do_sysctl(XEN_GUEST_HANDLE_PARAM(xen_sysctl_t) u_sysctl) last_online_cpu = cpumask_last(&cpu_online_map); max_cpu_index = min_t(uint32_t, ti->max_cpu_index, last_online_cpu); - ti->max_cpu_index = last_online_cpu; + + if ( guest_handle_is_null(ti->cputopo) ) + { + ret = -EINVAL; + break; + } for ( i = 0; i <= max_cpu_index; i++ ) { - if ( !guest_handle_is_null(ti->cpu_to_core) ) - { - uint32_t core = cpu_online(i) ? cpu_to_core(i) : ~0u; - if ( copy_to_guest_offset(ti->cpu_to_core, i, &core, 1) ) - break; - } - if ( !guest_handle_is_null(ti->cpu_to_socket) ) + xen_sysctl_cputopo_t cputopo; + + if ( cpu_online(i) ) { - uint32_t socket = cpu_online(i) ? cpu_to_socket(i) : ~0u; - if ( copy_to_guest_offset(ti->cpu_to_socket, i, &socket, 1) ) - break; + cputopo.core = cpu_to_core(i); + cputopo.socket = cpu_to_socket(i); + cputopo.node = cpu_to_node(i); } - if ( !guest_handle_is_null(ti->cpu_to_node) ) + else + cputopo.core = cputopo.socket = + cputopo.node = INVALID_TOPOLOGY_ID; + + if ( copy_to_guest_offset(ti->cputopo, i, &cputopo, 1) ) { - uint32_t node = cpu_online(i) ? cpu_to_node(i) : ~0u; - if ( copy_to_guest_offset(ti->cpu_to_node, i, &node, 1) ) - break; + ret = -EFAULT; + break; } } - ret = ((i <= max_cpu_index) || copy_to_guest(u_sysctl, op, 1)) - ? -EFAULT : 0; + if ( !ret && (ti->max_cpu_index != last_online_cpu) ) + { + ti->max_cpu_index = last_online_cpu; + if ( __copy_field_to_guest(u_sysctl, op, + u.topologyinfo.max_cpu_index) ) + ret = -EFAULT; + } } break; diff --git a/xen/include/public/sysctl.h b/xen/include/public/sysctl.h index b3713b3..e1d1348 100644 --- a/xen/include/public/sysctl.h +++ b/xen/include/public/sysctl.h @@ -34,7 +34,7 @@ #include "xen.h" #include "domctl.h" -#define XEN_SYSCTL_INTERFACE_VERSION 0x0000000B +#define XEN_SYSCTL_INTERFACE_VERSION 0x0000000C /* * Read console content from Xen buffer ring. @@ -464,26 +464,44 @@ DEFINE_XEN_GUEST_HANDLE(xen_sysctl_lockprof_op_t); /* XEN_SYSCTL_topologyinfo */ #define INVALID_TOPOLOGY_ID (~0U) + +struct xen_sysctl_cputopo { + uint32_t core; + uint32_t socket; + uint32_t node; +}; +typedef struct xen_sysctl_cputopo xen_sysctl_cputopo_t; +DEFINE_XEN_GUEST_HANDLE(xen_sysctl_cputopo_t); + +struct xen_sysctl_iotopo { + uint16_t seg; + uint8_t bus; + uint8_t devfn; + uint32_t node; +}; +typedef struct xen_sysctl_iotopo xen_sysctl_iotopo_t; +DEFINE_XEN_GUEST_HANDLE(xen_sysctl_iotopo_t); + struct xen_sysctl_topologyinfo { /* * IN: maximum addressable entry in the caller-provided arrays. - * OUT: largest cpu identifier in the system. + * OUT: largest cpu identifier or max number of devices in the system. * If OUT is greater than IN then the arrays are truncated! * If OUT is leass than IN then the array tails are not written by sysctl. */ uint32_t max_cpu_index; + uint32_t max_devs; /* * If not NULL, these arrays are filled with core/socket/node identifier - * for each cpu. - * If a cpu has no core/socket/node information (e.g., cpu not present) - * then the sentinel value ~0u is written to each array. - * The number of array elements written by the sysctl is: + * for each cpu and/or node for each PCI device. + * If information for a particular entry is not avalable it is set to + * INVALID_TOPOLOGY_ID. + * The number of array elements for CPU topology written by the sysctl is: * min(@max_cpu_index_IN,@max_cpu_index_OUT)+1 */ - XEN_GUEST_HANDLE_64(uint32) cpu_to_core; - XEN_GUEST_HANDLE_64(uint32) cpu_to_socket; - XEN_GUEST_HANDLE_64(uint32) cpu_to_node; + XEN_GUEST_HANDLE_64(xen_sysctl_cputopo_t) cputopo; + XEN_GUEST_HANDLE_64(xen_sysctl_iotopo_t) iotopo; }; typedef struct xen_sysctl_topologyinfo xen_sysctl_topologyinfo_t; DEFINE_XEN_GUEST_HANDLE(xen_sysctl_topologyinfo_t); -- 1.8.4.2 _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |