[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH v16 06/10] x86: collect global CMT information
On Thu, Sep 25, 2014 at 06:19:06PM +0800, Chao Peng wrote: > This implementation tries to put all policies into user space, thus some > global CMT information needs to be exposed, such as the total RMID count, > L3 upscaling factor, etc. > > Signed-off-by: Dongxiao Xu <dongxiao.xu@xxxxxxxxx> > Signed-off-by: Chao Peng <chao.p.peng@xxxxxxxxxxxxxxx> > Acked-by: Jan Beulich <jbeulich@xxxxxxxx> > --- > xen/arch/x86/cpu/intel_cacheinfo.c | 49 > ++---------------------------------- > xen/arch/x86/sysctl.c | 43 +++++++++++++++++++++++++++++++ > xen/include/asm-x86/cpufeature.h | 45 +++++++++++++++++++++++++++++++++ > xen/include/public/sysctl.h | 14 +++++++++++ > 4 files changed, 104 insertions(+), 47 deletions(-) > > diff --git a/xen/arch/x86/cpu/intel_cacheinfo.c > b/xen/arch/x86/cpu/intel_cacheinfo.c > index 430f939..48970c0 100644 > --- a/xen/arch/x86/cpu/intel_cacheinfo.c > +++ b/xen/arch/x86/cpu/intel_cacheinfo.c > @@ -81,54 +81,9 @@ static struct _cache_table cache_table[] __cpuinitdata = > { 0x00, 0, 0} > }; > > - > -enum _cache_type > -{ > - CACHE_TYPE_NULL = 0, > - CACHE_TYPE_DATA = 1, > - CACHE_TYPE_INST = 2, > - CACHE_TYPE_UNIFIED = 3 > -}; > - > -union _cpuid4_leaf_eax { > - struct { > - enum _cache_type type:5; > - unsigned int level:3; > - unsigned int is_self_initializing:1; > - unsigned int is_fully_associative:1; > - unsigned int reserved:4; > - unsigned int num_threads_sharing:12; > - unsigned int num_cores_on_die:6; > - } split; > - u32 full; > -}; > - > -union _cpuid4_leaf_ebx { > - struct { > - unsigned int coherency_line_size:12; > - unsigned int physical_line_partition:10; > - unsigned int ways_of_associativity:10; > - } split; > - u32 full; > -}; > - > -union _cpuid4_leaf_ecx { > - struct { > - unsigned int number_of_sets:32; > - } split; > - u32 full; > -}; > - > -struct _cpuid4_info { > - union _cpuid4_leaf_eax eax; > - union _cpuid4_leaf_ebx ebx; > - union _cpuid4_leaf_ecx ecx; > - unsigned long size; > -}; > - > unsigned short num_cache_leaves; > > -static int __cpuinit cpuid4_cache_lookup(int index, struct _cpuid4_info > *this_leaf) > +int cpuid4_cache_lookup(int index, struct cpuid4_info *this_leaf) > { > union _cpuid4_leaf_eax eax; > union _cpuid4_leaf_ebx ebx; > @@ -185,7 +140,7 @@ unsigned int __cpuinit init_intel_cacheinfo(struct > cpuinfo_x86 *c) > * parameters cpuid leaf to find the cache details > */ > for (i = 0; i < num_cache_leaves; i++) { > - struct _cpuid4_info this_leaf; > + struct cpuid4_info this_leaf; > > int retval; > > diff --git a/xen/arch/x86/sysctl.c b/xen/arch/x86/sysctl.c > index 15d4b91..b95408f 100644 > --- a/xen/arch/x86/sysctl.c > +++ b/xen/arch/x86/sysctl.c > @@ -28,6 +28,7 @@ > #include <xen/nodemask.h> > #include <xen/cpu.h> > #include <xsm/xsm.h> > +#include <asm/psr.h> > > #define get_xen_guest_handle(val, hnd) do { val = (hnd).p; } while (0) > > @@ -101,6 +102,48 @@ long arch_do_sysctl( > } > break; > > + case XEN_SYSCTL_psr_cmt_op: > + if ( !psr_cmt_enabled() ) > + return -ENODEV; > + > + if ( sysctl->u.psr_cmt_op.flags != 0 ) > + return -EINVAL; > + > + switch ( sysctl->u.psr_cmt_op.cmd ) > + { > + case XEN_SYSCTL_PSR_CMT_enabled: > + sysctl->u.psr_cmt_op.data = > + (psr_cmt->features & PSR_RESOURCE_TYPE_L3) && > + (psr_cmt->l3.features & PSR_CMT_L3_OCCUPANCY); > + break; > + case XEN_SYSCTL_PSR_CMT_get_total_rmid: > + sysctl->u.psr_cmt_op.data = psr_cmt->rmid_max; > + break; > + case XEN_SYSCTL_PSR_CMT_get_l3_upscaling_factor: > + sysctl->u.psr_cmt_op.data = psr_cmt->l3.upscaling_factor; > + break; > + case XEN_SYSCTL_PSR_CMT_get_l3_cache_size: > + { > + struct cpuid4_info info; > + > + ret = cpuid4_cache_lookup(3, &info); Couldn't you use 'struct cpuinfo_x86' and extend it if you need to? > + if ( ret < 0 ) > + break; > + > + sysctl->u.psr_cmt_op.data = info.size / 1024; /* in KB unit */ With the Haswell EP they have this weird setup where there are 8 cores on one side and 10 cores on another. Also the cache size is different (20MB LLC and 25MB LLC). With that wouldn't you want to enumerate exactly _which_ CPU cache you want instead of the one you running at? Or is my reading of the diagrams wrong and OS never sees the split and gets 45MB? > + } > + break; > + default: > + sysctl->u.psr_cmt_op.data = 0; > + ret = -ENOSYS; > + break; > + } > + > + if ( __copy_to_guest(u_sysctl, sysctl, 1) ) > + ret = -EFAULT; > + > + break; > + > default: > ret = -ENOSYS; > break; > diff --git a/xen/include/asm-x86/cpufeature.h > b/xen/include/asm-x86/cpufeature.h > index 137d75c..d3bd14d 100644 > --- a/xen/include/asm-x86/cpufeature.h > +++ b/xen/include/asm-x86/cpufeature.h > @@ -215,6 +215,51 @@ > #define cpu_has_vmx boot_cpu_has(X86_FEATURE_VMXE) > > #define cpu_has_cpuid_faulting boot_cpu_has(X86_FEATURE_CPUID_FAULTING) > + > +enum _cache_type { > + CACHE_TYPE_NULL = 0, > + CACHE_TYPE_DATA = 1, > + CACHE_TYPE_INST = 2, > + CACHE_TYPE_UNIFIED = 3 > +}; > + > +union _cpuid4_leaf_eax { > + struct { > + enum _cache_type type:5; > + unsigned int level:3; > + unsigned int is_self_initializing:1; > + unsigned int is_fully_associative:1; > + unsigned int reserved:4; > + unsigned int num_threads_sharing:12; > + unsigned int num_cores_on_die:6; > + } split; > + u32 full; > +}; > + > +union _cpuid4_leaf_ebx { > + struct { > + unsigned int coherency_line_size:12; > + unsigned int physical_line_partition:10; > + unsigned int ways_of_associativity:10; > + } split; > + u32 full; > +}; > + > +union _cpuid4_leaf_ecx { > + struct { > + unsigned int number_of_sets:32; > + } split; > + u32 full; > +}; > + > +struct cpuid4_info { > + union _cpuid4_leaf_eax eax; > + union _cpuid4_leaf_ebx ebx; > + union _cpuid4_leaf_ecx ecx; > + unsigned long size; > +}; > + > +int cpuid4_cache_lookup(int index, struct cpuid4_info *this_leaf); > #endif > > #endif /* __ASM_I386_CPUFEATURE_H */ > diff --git a/xen/include/public/sysctl.h b/xen/include/public/sysctl.h > index 3588698..66b6e47 100644 > --- a/xen/include/public/sysctl.h > +++ b/xen/include/public/sysctl.h > @@ -636,6 +636,18 @@ struct xen_sysctl_coverage_op { > typedef struct xen_sysctl_coverage_op xen_sysctl_coverage_op_t; > DEFINE_XEN_GUEST_HANDLE(xen_sysctl_coverage_op_t); > > +#define XEN_SYSCTL_PSR_CMT_get_total_rmid 0 > +#define XEN_SYSCTL_PSR_CMT_get_l3_upscaling_factor 1 > +/* The L3 cache size is returned in KB unit */ > +#define XEN_SYSCTL_PSR_CMT_get_l3_cache_size 2 > +#define XEN_SYSCTL_PSR_CMT_enabled 3 > +struct xen_sysctl_psr_cmt_op { > + uint32_t cmd; > + uint32_t flags; /* padding variable, may be extended for future use > */ > + uint64_t data; > +}; > +typedef struct xen_sysctl_psr_cmt_op xen_sysctl_psr_cmt_op_t; > +DEFINE_XEN_GUEST_HANDLE(xen_sysctl_psr_cmt_op_t); > > struct xen_sysctl { > uint32_t cmd; > @@ -658,6 +670,7 @@ struct xen_sysctl { > #define XEN_SYSCTL_cpupool_op 18 > #define XEN_SYSCTL_scheduler_op 19 > #define XEN_SYSCTL_coverage_op 20 > +#define XEN_SYSCTL_psr_cmt_op 21 > uint32_t interface_version; /* XEN_SYSCTL_INTERFACE_VERSION */ > union { > struct xen_sysctl_readconsole readconsole; > @@ -679,6 +692,7 @@ struct xen_sysctl { > struct xen_sysctl_cpupool_op cpupool_op; > struct xen_sysctl_scheduler_op scheduler_op; > struct xen_sysctl_coverage_op coverage_op; > + struct xen_sysctl_psr_cmt_op psr_cmt_op; > uint8_t pad[128]; > } u; > }; > -- > 1.7.9.5 > > > _______________________________________________ > Xen-devel mailing list > Xen-devel@xxxxxxxxxxxxx > http://lists.xen.org/xen-devel _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |