[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH v6 01/10] xen: vnuma topology and subop hypercalls
On Tue, Jul 22, 2014 at 11:14 AM, Dario Faggioli <dario.faggioli@xxxxxxxxxx> wrote: > On ven, 2014-07-18 at 01:50 -0400, Elena Ufimtseva wrote: > >> diff --git a/xen/common/domain.c b/xen/common/domain.c >> index cd64aea..895584a 100644 > >> @@ -297,6 +297,144 @@ int vcpuaffinity_params_invalid(const >> xen_domctl_vcpuaffinity_t *vcpuaff) >> guest_handle_is_null(vcpuaff->cpumap_soft.bitmap)); >> } >> >> +/* >> + * Allocates memory for vNUMA, **vnuma should be NULL. >> + * Caller has to make sure that domain has max_pages >> + * and number of vcpus set for domain. >> + * Verifies that single allocation does not exceed >> + * PAGE_SIZE. >> + */ >> +static int vnuma_alloc(struct vnuma_info **vnuma, >> + unsigned int nr_vnodes, >> + unsigned int nr_vcpus, >> + unsigned int dist_size) >> +{ >> + struct vnuma_info *v; >> + >> + if ( vnuma && *vnuma ) >> + return -EINVAL; >> + >> + v = *vnuma; >> > Do you need this? What for? > >> + /* >> + * check if any of xmallocs exeeds PAGE_SIZE. >> + * If yes, consider it as an error for now. >> > Do you mind elaborating a bit more on the 'for now'? Why 'for now'? > What's the plan for the future, etc. ... > >> + */ >> + if ( nr_vnodes > PAGE_SIZE / sizeof(nr_vnodes) || >> + nr_vcpus > PAGE_SIZE / sizeof(nr_vcpus) || >> + nr_vnodes > PAGE_SIZE / sizeof(struct vmemrange) || >> + dist_size > PAGE_SIZE / sizeof(dist_size) ) >> + return -EINVAL; >> + >> + v = xzalloc(struct vnuma_info); >> + if ( !v ) >> + return -ENOMEM; >> + >> + v->vdistance = xmalloc_array(unsigned int, dist_size); >> + v->vmemrange = xmalloc_array(vmemrange_t, nr_vnodes); >> + v->vcpu_to_vnode = xmalloc_array(unsigned int, nr_vcpus); >> + v->vnode_to_pnode = xmalloc_array(unsigned int, nr_vnodes); >> + >> + if ( v->vdistance == NULL || v->vmemrange == NULL || >> + v->vcpu_to_vnode == NULL || v->vnode_to_pnode == NULL ) >> + { >> + vnuma_destroy(v); >> + return -ENOMEM; >> + } >> + >> + *vnuma = v; >> + >> + return 0; >> +} >> + >> +/* >> + * Allocate memory and construct one vNUMA node, >> + * set default parameters, assign all memory and >> + * vcpus to this node, set distance to 10. >> + */ >> +static long vnuma_fallback(const struct domain *d, >> + struct vnuma_info **vnuma) >> +{ >> + > I think I agree with Wei, about this fallback not being necessary. > >> +/* >> + * construct vNUMA topology form u_vnuma struct and return >> + * it in dst. >> + */ >> +long vnuma_init(const struct xen_domctl_vnuma *u_vnuma, >> + const struct domain *d, >> + struct vnuma_info **dst) >> +{ >> + unsigned int dist_size, nr_vnodes = 0; >> + long ret; >> + struct vnuma_info *v = NULL; >> + >> + ret = -EINVAL; >> + > Why not initialize 'ret' while defining it? > >> + /* If vNUMA topology already set, just exit. */ >> + if ( !u_vnuma || *dst ) >> + return ret; >> + >> + nr_vnodes = u_vnuma->nr_vnodes; >> + >> + if ( nr_vnodes == 0 ) >> + return ret; >> + >> + if ( nr_vnodes > (UINT_MAX / nr_vnodes) ) >> + return ret; >> + > Mmmm, do we perhaps want to #define a maximum number of supported vitual > node, put it somewhere in an header, and use it for the check? I mean > something like what we have for the host (in that case, it's called > MAX_NUMNODES). > > I mean, if UINT_MAX is 2^64, would it make sense to allow a 2^32 nodes > guest? True to that, no one needs that many nodes :) Will define a const in v7. Probably, it will make sense to set to the same as vcpu number? > >> + dist_size = nr_vnodes * nr_vnodes; >> + >> + ret = vnuma_alloc(&v, nr_vnodes, d->max_vcpus, dist_size); >> + if ( ret ) >> + return ret; >> + >> + /* On failure, set only one vNUMA node and its success. */ >> + ret = 0; >> + >> + if ( copy_from_guest(v->vdistance, u_vnuma->vdistance, dist_size) ) >> + goto vnuma_onenode; >> + if ( copy_from_guest(v->vmemrange, u_vnuma->vmemrange, nr_vnodes) ) >> + goto vnuma_onenode; >> + if ( copy_from_guest(v->vcpu_to_vnode, u_vnuma->vcpu_to_vnode, >> + d->max_vcpus) ) >> + goto vnuma_onenode; >> + if ( copy_from_guest(v->vnode_to_pnode, u_vnuma->vnode_to_pnode, >> + nr_vnodes) ) >> + goto vnuma_onenode; >> + >> + v->nr_vnodes = nr_vnodes; >> + *dst = v; >> + >> + return ret; >> + >> +vnuma_onenode: >> + vnuma_destroy(v); >> + return vnuma_fallback(d, dst); >> > As said, just report the error and bail in this case. Yes, agree on that. > >> +} >> + >> long do_domctl(XEN_GUEST_HANDLE_PARAM(xen_domctl_t) u_domctl) >> { >> long ret = 0; >> @@ -967,6 +1105,35 @@ long do_domctl(XEN_GUEST_HANDLE_PARAM(xen_domctl_t) >> u_domctl) >> } >> break; >> >> + case XEN_DOMCTL_setvnumainfo: >> + { >> + struct vnuma_info *v = NULL; >> + >> + ret = -EFAULT; >> + if ( guest_handle_is_null(op->u.vnuma.vdistance) || >> + guest_handle_is_null(op->u.vnuma.vmemrange) || >> + guest_handle_is_null(op->u.vnuma.vcpu_to_vnode) || >> + guest_handle_is_null(op->u.vnuma.vnode_to_pnode) ) >> + return ret; >> + >> + ret = -EINVAL; >> + >> + ret = vnuma_init(&op->u.vnuma, d, &v); >> > Rather pointless 'ret=-EINVAL', I would say. :-) > >> + if ( ret < 0 || v == NULL ) >> + break; >> + >> + /* overwrite vnuma for domain */ >> + if ( !d->vnuma ) >> + vnuma_destroy(d->vnuma); >> + >> + domain_lock(d); >> + d->vnuma = v; >> + domain_unlock(d); >> + >> + ret = 0; >> + } >> + break; >> + >> default: >> ret = arch_do_domctl(op, d, u_domctl); >> break; > > Regards, > Dario > > -- > <<This happens because I choose it to happen!>> (Raistlin Majere) > ----------------------------------------------------------------- > Dario Faggioli, Ph.D, http://about.me/dario.faggioli > Senior Software Engineer, Citrix Systems R&D Ltd., Cambridge (UK) > -- Elena _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |