[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-devel] [PATCH v5 04/17] intel/VPMU: Clean up Intel VPMU code



Am Montag 17 Februar 2014, 12:55:51 schrieb Boris Ostrovsky:
> Remove struct pmumsr and core2_pmu_enable. Replace static MSR structures with
> fields in core2_vpmu_context.
> 
> Call core2_get_pmc_count() once, during initialization.
> 
> Properly clean up when core2_vpmu_alloc_resource() fails and add routines
> to remove MSRs from VMCS.

Only a minor remark below.

> 
> Signed-off-by: Boris Ostrovsky <boris.ostrovsky@xxxxxxxxxx>
> ---
>  xen/arch/x86/hvm/vmx/vmcs.c              |  55 ++++++
>  xen/arch/x86/hvm/vmx/vpmu_core2.c        | 310 
> ++++++++++++++-----------------
>  xen/include/asm-x86/hvm/vmx/vmcs.h       |   2 +
>  xen/include/asm-x86/hvm/vmx/vpmu_core2.h |  19 --
>  4 files changed, 199 insertions(+), 187 deletions(-)
> 
> diff --git a/xen/arch/x86/hvm/vmx/vmcs.c b/xen/arch/x86/hvm/vmx/vmcs.c
> index 44f33cb..5f86b17 100644
> --- a/xen/arch/x86/hvm/vmx/vmcs.c
> +++ b/xen/arch/x86/hvm/vmx/vmcs.c
> @@ -1205,6 +1205,34 @@ int vmx_add_guest_msr(u32 msr)
>      return 0;
>  }
>  
> +void vmx_rm_guest_msr(u32 msr)
> +{
> +    struct vcpu *curr = current;
> +    unsigned int idx, msr_count = curr->arch.hvm_vmx.msr_count;
> +    struct vmx_msr_entry *msr_area = curr->arch.hvm_vmx.msr_area;
> +
> +    if ( msr_area == NULL )
> +        return;
> +
> +    for ( idx = 0; idx < msr_count; idx++ )
> +        if ( msr_area[idx].index == msr )
> +            break;
> +
> +    if ( idx == msr_count )
> +        return;
> +    for ( ; idx < msr_count - 1; idx++ )
> +    {
> +        msr_area[idx].index = msr_area[idx + 1].index;
> +        msr_area[idx].data = msr_area[idx + 1].data;
> +    }
> +    msr_area[msr_count - 1].index = 0;
> +
> +    curr->arch.hvm_vmx.msr_count = --msr_count;
> +    __vmwrite(VM_EXIT_MSR_STORE_COUNT, msr_count);
> +    __vmwrite(VM_ENTRY_MSR_LOAD_COUNT, msr_count);
> +}
> +
>  int vmx_add_host_load_msr(u32 msr)
>  {
>      struct vcpu *curr = current;
> @@ -1235,6 +1263,33 @@ int vmx_add_host_load_msr(u32 msr)
>      return 0;
>  }
>  
> +void vmx_rm_host_load_msr(u32 msr)
> +{
> +    struct vcpu *curr = current;
> +    unsigned int idx,  msr_count = curr->arch.hvm_vmx.host_msr_count;
> +    struct vmx_msr_entry *msr_area = curr->arch.hvm_vmx.host_msr_area;
> +
> +    if ( msr_area == NULL )
> +        return;
> +
> +    for ( idx = 0; idx < msr_count; idx++ )
> +        if ( msr_area[idx].index == msr )
> +            break;
> +
> +    if ( idx == msr_count )
> +        return;
> +
> +    for ( ; idx < msr_count - 1; idx++ )
> +    {
> +        msr_area[idx].index = msr_area[idx + 1].index;
> +        msr_area[idx].data = msr_area[idx + 1].data;
> +    }
> +    msr_area[msr_count - 1].index = 0;
> +
> +    curr->arch.hvm_vmx.host_msr_count = --msr_count;
> +    __vmwrite(VM_EXIT_MSR_LOAD_COUNT, msr_count);
> +}
> +
>  void vmx_set_eoi_exit_bitmap(struct vcpu *v, u8 vector)
>  {
>      if ( !test_and_set_bit(vector, v->arch.hvm_vmx.eoi_exit_bitmap) )
> diff --git a/xen/arch/x86/hvm/vmx/vpmu_core2.c 
> b/xen/arch/x86/hvm/vmx/vpmu_core2.c
> index 1e32ff3..513eca4 100644
> --- a/xen/arch/x86/hvm/vmx/vpmu_core2.c
> +++ b/xen/arch/x86/hvm/vmx/vpmu_core2.c
> @@ -69,6 +69,26 @@
>  static bool_t __read_mostly full_width_write;
>  
>  /*
> + * MSR_CORE_PERF_FIXED_CTR_CTRL contains the configuration of all fixed
> + * counters. 4 bits for every counter.
> + */
> +#define FIXED_CTR_CTRL_BITS 4
> +#define FIXED_CTR_CTRL_MASK ((1 << FIXED_CTR_CTRL_BITS) - 1)
> +
> +#define VPMU_CORE2_MAX_FIXED_PMCS     4
> +struct core2_vpmu_context {
> +    u64 fixed_ctrl;
> +    u64 ds_area;
> +    u64 pebs_enable;
> +    u64 global_ovf_status;
> +    u64 fix_counters[VPMU_CORE2_MAX_FIXED_PMCS];
> +    struct arch_msr_pair arch_msr_pair[1];
> +};
> +
> +/* Number of general-purpose and fixed performance counters */
> +static unsigned int __read_mostly arch_pmc_cnt, fixed_pmc_cnt;
> +
> +/*
>   * QUIRK to workaround an issue on various family 6 cpus.
>   * The issue leads to endless PMC interrupt loops on the processor.
>   * If the interrupt handler is running and a pmc reaches the value 0, this
> @@ -88,11 +108,8 @@ static void check_pmc_quirk(void)
>          is_pmc_quirk = 0;    
>  }
>  
> -static int core2_get_pmc_count(void);
>  static void handle_pmc_quirk(u64 msr_content)
>  {
> -    int num_gen_pmc = core2_get_pmc_count();
> -    int num_fix_pmc  = 3;
>      int i;
>      u64 val;
>  
> @@ -100,7 +117,7 @@ static void handle_pmc_quirk(u64 msr_content)
>          return;
>  
>      val = msr_content;
> -    for ( i = 0; i < num_gen_pmc; i++ )
> +    for ( i = 0; i < arch_pmc_cnt; i++ )
>      {
>          if ( val & 0x1 )
>          {
> @@ -112,7 +129,7 @@ static void handle_pmc_quirk(u64 msr_content)
>          val >>= 1;
>      }
>      val = msr_content >> 32;
> -    for ( i = 0; i < num_fix_pmc; i++ )
> +    for ( i = 0; i < fixed_pmc_cnt; i++ )
>      {
>          if ( val & 0x1 )
>          {
> @@ -125,75 +142,42 @@ static void handle_pmc_quirk(u64 msr_content)
>      }
>  }
>  
> -static const u32 core2_fix_counters_msr[] = {
> -    MSR_CORE_PERF_FIXED_CTR0,
> -    MSR_CORE_PERF_FIXED_CTR1,
> -    MSR_CORE_PERF_FIXED_CTR2
> -};
> -
>  /*
> - * MSR_CORE_PERF_FIXED_CTR_CTRL contains the configuration of all fixed
> - * counters. 4 bits for every counter.
> + * Read the number of general counters via CPUID.EAX[0xa].EAX[8..15]
>   */
> -#define FIXED_CTR_CTRL_BITS 4
> -#define FIXED_CTR_CTRL_MASK ((1 << FIXED_CTR_CTRL_BITS) - 1)
> -
> -/* The index into the core2_ctrls_msr[] of this MSR used in 
> core2_vpmu_dump() */
> -#define MSR_CORE_PERF_FIXED_CTR_CTRL_IDX 0
> -
> -/* Core 2 Non-architectual Performance Control MSRs. */
> -static const u32 core2_ctrls_msr[] = {
> -    MSR_CORE_PERF_FIXED_CTR_CTRL,
> -    MSR_IA32_PEBS_ENABLE,
> -    MSR_IA32_DS_AREA
> -};
> -
> -struct pmumsr {
> -    unsigned int num;
> -    const u32 *msr;
> -};
> -
> -static const struct pmumsr core2_fix_counters = {
> -    VPMU_CORE2_NUM_FIXED,
> -    core2_fix_counters_msr
> -};
> +static int core2_get_arch_pmc_count(void)
> +{
> +    u32 eax;
>  
> -static const struct pmumsr core2_ctrls = {
> -    VPMU_CORE2_NUM_CTRLS,
> -    core2_ctrls_msr
> -};
> -static int arch_pmc_cnt;
> +    eax = cpuid_eax(0xa);
> +    return ( (eax & PMU_GENERAL_NR_MASK) >> PMU_GENERAL_NR_SHIFT );
> +}
>  
>  /*
> - * Read the number of general counters via CPUID.EAX[0xa].EAX[8..15]
> + * Read the number of fixed counters via CPUID.EDX[0xa].EDX[0..4]
>   */
> -static int core2_get_pmc_count(void)
> +static int core2_get_fixed_pmc_count(void)
>  {
> -    u32 eax, ebx, ecx, edx;
> -
> -    if ( arch_pmc_cnt == 0 )
> -    {
> -        cpuid(0xa, &eax, &ebx, &ecx, &edx);
> -        arch_pmc_cnt = (eax & PMU_GENERAL_NR_MASK) >> PMU_GENERAL_NR_SHIFT;
> -    }
> +    u32 eax;
>  
> -    return arch_pmc_cnt;
> +    eax = cpuid_eax(0xa);
> +    return ( (eax & PMU_FIXED_NR_MASK) >> PMU_FIXED_NR_SHIFT );
>  }
>  
>  static u64 core2_calc_intial_glb_ctrl_msr(void)
>  {
> -    int arch_pmc_bits = (1 << core2_get_pmc_count()) - 1;
> -    u64 fix_pmc_bits  = (1 << 3) - 1;
> -    return ((fix_pmc_bits << 32) | arch_pmc_bits);
> +    int arch_pmc_bits = (1 << arch_pmc_cnt) - 1;
> +    u64 fix_pmc_bits  = (1 << fixed_pmc_cnt) - 1;
> +    return ( (fix_pmc_bits << 32) | arch_pmc_bits );
>  }
>  
>  /* edx bits 5-12: Bit width of fixed-function performance counters  */
>  static int core2_get_bitwidth_fix_count(void)
>  {
> -    u32 eax, ebx, ecx, edx;
> +    u32 edx;
>  
> -    cpuid(0xa, &eax, &ebx, &ecx, &edx);
> -    return ((edx & PMU_FIXED_WIDTH_MASK) >> PMU_FIXED_WIDTH_SHIFT);
> +    edx = cpuid_edx(0xa);
> +    return ( (edx & PMU_FIXED_WIDTH_MASK) >> PMU_FIXED_WIDTH_SHIFT );
>  }
>  
>  static int is_core2_vpmu_msr(u32 msr_index, int *type, int *index)
> @@ -201,9 +185,9 @@ static int is_core2_vpmu_msr(u32 msr_index, int *type, 
> int *index)
>      int i;
>      u32 msr_index_pmc;
>  
> -    for ( i = 0; i < core2_fix_counters.num; i++ )
> +    for ( i = 0; i < fixed_pmc_cnt; i++ )
>      {
> -        if ( core2_fix_counters.msr[i] == msr_index )
> +        if ( msr_index == MSR_CORE_PERF_FIXED_CTR0 + i )
>          {
>              *type = MSR_TYPE_COUNTER;
>              *index = i;
> @@ -211,14 +195,12 @@ static int is_core2_vpmu_msr(u32 msr_index, int *type, 
> int *index)
>          }
>      }
>  
> -    for ( i = 0; i < core2_ctrls.num; i++ )
> +    if ( (msr_index == MSR_CORE_PERF_FIXED_CTR_CTRL ) ||
> +        (msr_index == MSR_IA32_DS_AREA) ||
> +        (msr_index == MSR_IA32_PEBS_ENABLE) )
>      {
> -        if ( core2_ctrls.msr[i] == msr_index )
> -        {
> -            *type = MSR_TYPE_CTRL;
> -            *index = i;
> -            return 1;
> -        }
> +        *type = MSR_TYPE_CTRL;
> +        return 1;
>      }
>  
>      if ( (msr_index == MSR_CORE_PERF_GLOBAL_CTRL) ||
> @@ -231,7 +213,7 @@ static int is_core2_vpmu_msr(u32 msr_index, int *type, 
> int *index)
>  
>      msr_index_pmc = msr_index & MSR_PMC_ALIAS_MASK;
>      if ( (msr_index_pmc >= MSR_IA32_PERFCTR0) &&
> -         (msr_index_pmc < (MSR_IA32_PERFCTR0 + core2_get_pmc_count())) )
> +         (msr_index_pmc < (MSR_IA32_PERFCTR0 + arch_pmc_cnt)) )
>      {
>          *type = MSR_TYPE_ARCH_COUNTER;
>          *index = msr_index_pmc - MSR_IA32_PERFCTR0;
> @@ -239,7 +221,7 @@ static int is_core2_vpmu_msr(u32 msr_index, int *type, 
> int *index)
>      }
>  
>      if ( (msr_index >= MSR_P6_EVNTSEL0) &&
> -         (msr_index < (MSR_P6_EVNTSEL0 + core2_get_pmc_count())) )
> +         (msr_index < (MSR_P6_EVNTSEL0 + arch_pmc_cnt)) )
>      {
>          *type = MSR_TYPE_ARCH_CTRL;
>          *index = msr_index - MSR_P6_EVNTSEL0;
> @@ -254,13 +236,13 @@ static void core2_vpmu_set_msr_bitmap(unsigned long 
> *msr_bitmap)
>      int i;
>  
>      /* Allow Read/Write PMU Counters MSR Directly. */
> -    for ( i = 0; i < core2_fix_counters.num; i++ )
> +    for ( i = 0; i < fixed_pmc_cnt; i++ )
>      {
> -        clear_bit(msraddr_to_bitpos(core2_fix_counters.msr[i]), msr_bitmap);
> -        clear_bit(msraddr_to_bitpos(core2_fix_counters.msr[i]),
> +        clear_bit(msraddr_to_bitpos(MSR_CORE_PERF_FIXED_CTR0 + i), 
> msr_bitmap);
> +        clear_bit(msraddr_to_bitpos(MSR_CORE_PERF_FIXED_CTR0 + i),
>                    msr_bitmap + 0x800/BYTES_PER_LONG);
>      }
> -    for ( i = 0; i < core2_get_pmc_count(); i++ )
> +    for ( i = 0; i < arch_pmc_cnt; i++ )
>      {
>          clear_bit(msraddr_to_bitpos(MSR_IA32_PERFCTR0+i), msr_bitmap);
>          clear_bit(msraddr_to_bitpos(MSR_IA32_PERFCTR0+i),
> @@ -275,26 +257,28 @@ static void core2_vpmu_set_msr_bitmap(unsigned long 
> *msr_bitmap)
>      }
>  
>      /* Allow Read PMU Non-global Controls Directly. */
> -    for ( i = 0; i < core2_ctrls.num; i++ )
> -        clear_bit(msraddr_to_bitpos(core2_ctrls.msr[i]), msr_bitmap);
> -    for ( i = 0; i < core2_get_pmc_count(); i++ )
> -        clear_bit(msraddr_to_bitpos(MSR_P6_EVNTSEL0+i), msr_bitmap);
> +    for ( i = 0; i < arch_pmc_cnt; i++ )
> +         clear_bit(msraddr_to_bitpos(MSR_P6_EVNTSEL0 + i), msr_bitmap);
> +
> +    clear_bit(msraddr_to_bitpos(MSR_CORE_PERF_FIXED_CTR_CTRL), msr_bitmap);
> +    clear_bit(msraddr_to_bitpos(MSR_IA32_PEBS_ENABLE), msr_bitmap);
> +    clear_bit(msraddr_to_bitpos(MSR_IA32_DS_AREA), msr_bitmap);
>  }
>  
>  static void core2_vpmu_unset_msr_bitmap(unsigned long *msr_bitmap)
>  {
>      int i;
>  
> -    for ( i = 0; i < core2_fix_counters.num; i++ )
> +    for ( i = 0; i < fixed_pmc_cnt; i++ )
>      {
> -        set_bit(msraddr_to_bitpos(core2_fix_counters.msr[i]), msr_bitmap);
> -        set_bit(msraddr_to_bitpos(core2_fix_counters.msr[i]),
> +        set_bit(msraddr_to_bitpos(MSR_CORE_PERF_FIXED_CTR0 + i), msr_bitmap);
> +        set_bit(msraddr_to_bitpos(MSR_CORE_PERF_FIXED_CTR0 + i),
>                  msr_bitmap + 0x800/BYTES_PER_LONG);
>      }
> -    for ( i = 0; i < core2_get_pmc_count(); i++ )
> +    for ( i = 0; i < arch_pmc_cnt; i++ )
>      {
> -        set_bit(msraddr_to_bitpos(MSR_IA32_PERFCTR0+i), msr_bitmap);
> -        set_bit(msraddr_to_bitpos(MSR_IA32_PERFCTR0+i),
> +        set_bit(msraddr_to_bitpos(MSR_IA32_PERFCTR0 + i), msr_bitmap);
> +        set_bit(msraddr_to_bitpos(MSR_IA32_PERFCTR0 + i),
>                  msr_bitmap + 0x800/BYTES_PER_LONG);
>  
>          if ( full_width_write )
> @@ -305,10 +289,12 @@ static void core2_vpmu_unset_msr_bitmap(unsigned long 
> *msr_bitmap)
>          }
>      }
>  
> -    for ( i = 0; i < core2_ctrls.num; i++ )
> -        set_bit(msraddr_to_bitpos(core2_ctrls.msr[i]), msr_bitmap);
> -    for ( i = 0; i < core2_get_pmc_count(); i++ )
> -        set_bit(msraddr_to_bitpos(MSR_P6_EVNTSEL0+i), msr_bitmap);
> +    for ( i = 0; i < arch_pmc_cnt; i++ )
> +        set_bit(msraddr_to_bitpos(MSR_P6_EVNTSEL0 + i), msr_bitmap);
> +
> +    set_bit(msraddr_to_bitpos(MSR_CORE_PERF_FIXED_CTR_CTRL), msr_bitmap);
> +    set_bit(msraddr_to_bitpos(MSR_IA32_PEBS_ENABLE), msr_bitmap);
> +    set_bit(msraddr_to_bitpos(MSR_IA32_DS_AREA), msr_bitmap);
>  }
>  
>  static inline void __core2_vpmu_save(struct vcpu *v)
> @@ -316,10 +302,10 @@ static inline void __core2_vpmu_save(struct vcpu *v)
>      int i;
>      struct core2_vpmu_context *core2_vpmu_cxt = vcpu_vpmu(v)->context;
>  
> -    for ( i = 0; i < core2_fix_counters.num; i++ )
> -        rdmsrl(core2_fix_counters.msr[i], core2_vpmu_cxt->fix_counters[i]);
> -    for ( i = 0; i < core2_get_pmc_count(); i++ )
> -        rdmsrl(MSR_IA32_PERFCTR0+i, 
> core2_vpmu_cxt->arch_msr_pair[i].counter);
> +    for ( i = 0; i < fixed_pmc_cnt; i++ )
> +        rdmsrl(MSR_CORE_PERF_FIXED_CTR0 + i, 
> core2_vpmu_cxt->fix_counters[i]);
> +    for ( i = 0; i < arch_pmc_cnt; i++ )
> +        rdmsrl(MSR_IA32_PERFCTR0 + i, 
> core2_vpmu_cxt->arch_msr_pair[i].counter);
>  }
>  
>  static int core2_vpmu_save(struct vcpu *v)
> @@ -343,20 +329,22 @@ static inline void __core2_vpmu_load(struct vcpu *v)
>      unsigned int i, pmc_start;
>      struct core2_vpmu_context *core2_vpmu_cxt = vcpu_vpmu(v)->context;
>  
> -    for ( i = 0; i < core2_fix_counters.num; i++ )
> -        wrmsrl(core2_fix_counters.msr[i], core2_vpmu_cxt->fix_counters[i]);
> +    for ( i = 0; i < fixed_pmc_cnt; i++ )
> +        wrmsrl(MSR_CORE_PERF_FIXED_CTR0 + i, 
> core2_vpmu_cxt->fix_counters[i]);
>  
>      if ( full_width_write )
>          pmc_start = MSR_IA32_A_PERFCTR0;
>      else
>          pmc_start = MSR_IA32_PERFCTR0;
> -    for ( i = 0; i < core2_get_pmc_count(); i++ )
> +    for ( i = 0; i < arch_pmc_cnt; i++ )
> +    {
>          wrmsrl(pmc_start + i, core2_vpmu_cxt->arch_msr_pair[i].counter);
> +        wrmsrl(MSR_P6_EVNTSEL0 + i, 
> core2_vpmu_cxt->arch_msr_pair[i].control);
> +    }
>  
> -    for ( i = 0; i < core2_ctrls.num; i++ )
> -        wrmsrl(core2_ctrls.msr[i], core2_vpmu_cxt->ctrls[i]);
> -    for ( i = 0; i < core2_get_pmc_count(); i++ )
> -        wrmsrl(MSR_P6_EVNTSEL0+i, core2_vpmu_cxt->arch_msr_pair[i].control);
> +    wrmsrl(MSR_CORE_PERF_FIXED_CTR_CTRL, core2_vpmu_cxt->fixed_ctrl);
> +    wrmsrl(MSR_IA32_DS_AREA, core2_vpmu_cxt->ds_area);
> +    wrmsrl(MSR_IA32_PEBS_ENABLE, core2_vpmu_cxt->pebs_enable);
>  }
>  
>  static void core2_vpmu_load(struct vcpu *v)
> @@ -375,56 +363,39 @@ static int core2_vpmu_alloc_resource(struct vcpu *v)
>  {
>      struct vpmu_struct *vpmu = vcpu_vpmu(v);
>      struct core2_vpmu_context *core2_vpmu_cxt;
> -    struct core2_pmu_enable *pmu_enable;
>  
>      if ( !acquire_pmu_ownership(PMU_OWNER_HVM) )
>          return 0;
>  
>      wrmsrl(MSR_CORE_PERF_GLOBAL_CTRL, 0);
>      if ( vmx_add_host_load_msr(MSR_CORE_PERF_GLOBAL_CTRL) )
> -        return 0;
> +        goto out_err;
>  
>      if ( vmx_add_guest_msr(MSR_CORE_PERF_GLOBAL_CTRL) )
> -        return 0;
> +        goto out_err;
>      vmx_write_guest_msr(MSR_CORE_PERF_GLOBAL_CTRL,
>                   core2_calc_intial_glb_ctrl_msr());
>  
> -    pmu_enable = xzalloc_bytes(sizeof(struct core2_pmu_enable) +
> -                               core2_get_pmc_count() - 1);
> -    if ( !pmu_enable )
> -        goto out1;
> -
>      core2_vpmu_cxt = xzalloc_bytes(sizeof(struct core2_vpmu_context) +
> -                    (core2_get_pmc_count()-1)*sizeof(struct arch_msr_pair));
> +                    (arch_pmc_cnt-1)*sizeof(struct arch_msr_pair));
>      if ( !core2_vpmu_cxt )
> -        goto out2;
> -    core2_vpmu_cxt->pmu_enable = pmu_enable;
> +        goto out_err;
> +
>      vpmu->context = (void *)core2_vpmu_cxt;
>  
> +    vpmu_set(vpmu, VPMU_CONTEXT_ALLOCATED);
> +
>      return 1;
> - out2:
> -    xfree(pmu_enable);
> - out1:
> -    gdprintk(XENLOG_WARNING, "Insufficient memory for PMU, PMU feature is "
> -             "unavailable on domain %d vcpu %d.\n",
> -             v->vcpu_id, v->domain->domain_id);
> -    return 0;
> -}
>  
> -static void core2_vpmu_save_msr_context(struct vcpu *v, int type,
> -                                       int index, u64 msr_data)
> -{
> -    struct core2_vpmu_context *core2_vpmu_cxt = vcpu_vpmu(v)->context;
> +out_err:
> +    vmx_rm_host_load_msr(MSR_CORE_PERF_GLOBAL_CTRL);
> +    vmx_rm_guest_msr(MSR_CORE_PERF_GLOBAL_CTRL);
> +    release_pmu_ownship(PMU_OWNER_HVM);
>  
> -    switch ( type )
> -    {
> -    case MSR_TYPE_CTRL:
> -        core2_vpmu_cxt->ctrls[index] = msr_data;
> -        break;
> -    case MSR_TYPE_ARCH_CTRL:
> -        core2_vpmu_cxt->arch_msr_pair[index].control = msr_data;
> -        break;
> -    }
> +    printk("Failed to allocate VPMU resources for domain %u vcpu %u\n",
> +           v->vcpu_id, v->domain->domain_id);
> +
> +    return 0;
>  }
>  
>  static int core2_vpmu_msr_common_check(u32 msr_index, int *type, int *index)
> @@ -435,10 +406,8 @@ static int core2_vpmu_msr_common_check(u32 msr_index, 
> int *type, int *index)
>          return 0;
>  
>      if ( unlikely(!vpmu_is_set(vpmu, VPMU_CONTEXT_ALLOCATED)) &&
> -      (vpmu->context != NULL ||
> -       !core2_vpmu_alloc_resource(current)) )
> +         !core2_vpmu_alloc_resource(current) )
>          return 0;
> -    vpmu_set(vpmu, VPMU_CONTEXT_ALLOCATED);
>  
>      /* Do the lazy load staff. */
>      if ( !vpmu_is_set(vpmu, VPMU_CONTEXT_LOADED) )
> @@ -454,7 +423,7 @@ static int core2_vpmu_msr_common_check(u32 msr_index, int 
> *type, int *index)
>  static int core2_vpmu_do_wrmsr(unsigned int msr, uint64_t msr_content)
>  {
>      u64 global_ctrl, non_global_ctrl;
> -    char pmu_enable = 0;
> +    unsigned pmu_enable = 0;
>      int i, tmp;
>      int type = -1, index = -1;
>      struct vcpu *v = current;
> @@ -499,6 +468,7 @@ static int core2_vpmu_do_wrmsr(unsigned int msr, uint64_t 
> msr_content)
>          if ( msr_content & 1 )
>              gdprintk(XENLOG_WARNING, "Guest is trying to enable PEBS, "
>                       "which is not supported.\n");
> +        core2_vpmu_cxt->pebs_enable = msr_content;
>          return 1;
>      case MSR_IA32_DS_AREA:
>          if ( vpmu_is_set(vpmu, VPMU_CPU_HAS_DS) )
> @@ -511,27 +481,25 @@ static int core2_vpmu_do_wrmsr(unsigned int msr, 
> uint64_t msr_content)
>                  hvm_inject_hw_exception(TRAP_gp_fault, 0);
>                  return 1;
>              }
> -            core2_vpmu_cxt->pmu_enable->ds_area_enable = msr_content ? 1 : 0;
> +            core2_vpmu_cxt->ds_area = msr_content;
>              break;
>          }
>          gdprintk(XENLOG_WARNING, "Guest setting of DTS is ignored.\n");
>          return 1;
>      case MSR_CORE_PERF_GLOBAL_CTRL:
>          global_ctrl = msr_content;
> -        for ( i = 0; i < core2_get_pmc_count(); i++ )
> +        for ( i = 0; i < arch_pmc_cnt; i++ )
>          {
>              rdmsrl(MSR_P6_EVNTSEL0+i, non_global_ctrl);
> -            core2_vpmu_cxt->pmu_enable->arch_pmc_enable[i] =
> -                    global_ctrl & (non_global_ctrl >> 22) & 1;
> +            pmu_enable += global_ctrl & (non_global_ctrl >> 22) & 1;
>              global_ctrl >>= 1;
>          }
>  
>          rdmsrl(MSR_CORE_PERF_FIXED_CTR_CTRL, non_global_ctrl);
>          global_ctrl = msr_content >> 32;
> -        for ( i = 0; i < core2_fix_counters.num; i++ )
> +        for ( i = 0; i < fixed_pmc_cnt; i++ )
>          {
> -            core2_vpmu_cxt->pmu_enable->fixed_ctr_enable[i] =
> -                (global_ctrl & 1) & ((non_global_ctrl & 0x3)? 1: 0);
> +            pmu_enable += (global_ctrl & 1) & ((non_global_ctrl & 0x3)? 1 : 
> 0);
>              non_global_ctrl >>= FIXED_CTR_CTRL_BITS;
>              global_ctrl >>= 1;
>          }
> @@ -540,27 +508,27 @@ static int core2_vpmu_do_wrmsr(unsigned int msr, 
> uint64_t msr_content)
>          non_global_ctrl = msr_content;
>          vmx_read_guest_msr(MSR_CORE_PERF_GLOBAL_CTRL, &global_ctrl);
>          global_ctrl >>= 32;
> -        for ( i = 0; i < core2_fix_counters.num; i++ )
> +        for ( i = 0; i < fixed_pmc_cnt; i++ )
>          {
> -            core2_vpmu_cxt->pmu_enable->fixed_ctr_enable[i] =
> -                (global_ctrl & 1) & ((non_global_ctrl & 0x3)? 1: 0);
> +            pmu_enable += (global_ctrl & 1) & ((non_global_ctrl & 0x3)? 1 : 
> 0);
>              non_global_ctrl >>= 4;
>              global_ctrl >>= 1;
>          }
> +        core2_vpmu_cxt->fixed_ctrl = msr_content;
>          break;
>      default:
>          tmp = msr - MSR_P6_EVNTSEL0;
> -        vmx_read_guest_msr(MSR_CORE_PERF_GLOBAL_CTRL, &global_ctrl);
> -        if ( tmp >= 0 && tmp < core2_get_pmc_count() )
> -            core2_vpmu_cxt->pmu_enable->arch_pmc_enable[tmp] =
> -                (global_ctrl >> tmp) & (msr_content >> 22) & 1;
> +        if ( tmp >= 0 && tmp < arch_pmc_cnt )
> +        {
> +            vmx_read_guest_msr(MSR_CORE_PERF_GLOBAL_CTRL, &global_ctrl);
> +            core2_vpmu_cxt->arch_msr_pair[tmp].control = msr_content;
> +            for ( i = 0; i < arch_pmc_cnt && !pmu_enable; i++ )
> +                pmu_enable += (global_ctrl >> i) &
> +                    (core2_vpmu_cxt->arch_msr_pair[i].control >> 22) & 1;
> +        }
>      }
>  
> -    for ( i = 0; i < core2_fix_counters.num; i++ )
> -        pmu_enable |= core2_vpmu_cxt->pmu_enable->fixed_ctr_enable[i];
> -    for ( i = 0; i < core2_get_pmc_count(); i++ )
> -        pmu_enable |= core2_vpmu_cxt->pmu_enable->arch_pmc_enable[i];
> -    pmu_enable |= core2_vpmu_cxt->pmu_enable->ds_area_enable;
> +    pmu_enable += (core2_vpmu_cxt->ds_area != 0);

Maybe it would be better to move this incrementing of pmu_enable to the
"case MSR_IA32_DS_AREA:" above where the ds_area gets handled ? 

Dietmar.

>      if ( pmu_enable )
>          vpmu_set(vpmu, VPMU_RUNNING);
>      else
> @@ -579,7 +547,6 @@ static int core2_vpmu_do_wrmsr(unsigned int msr, uint64_t 
> msr_content)
>          vpmu->hw_lapic_lvtpc = PMU_APIC_VECTOR | APIC_LVT_MASKED;
>      }
>  
> -    core2_vpmu_save_msr_context(v, type, index, msr_content);
>      if ( type != MSR_TYPE_GLOBAL )
>      {
>          u64 mask;
> @@ -595,7 +562,7 @@ static int core2_vpmu_do_wrmsr(unsigned int msr, uint64_t 
> msr_content)
>              if  ( msr == MSR_IA32_DS_AREA )
>                  break;
>              /* 4 bits per counter, currently 3 fixed counters implemented. */
> -            mask = ~((1ull << (VPMU_CORE2_NUM_FIXED * FIXED_CTR_CTRL_BITS)) 
> - 1);
> +            mask = ~((1ull << (fixed_pmc_cnt * FIXED_CTR_CTRL_BITS)) - 1);
>              if (msr_content & mask)
>                  inject_gp = 1;
>              break;
> @@ -680,7 +647,7 @@ static void core2_vpmu_do_cpuid(unsigned int input,
>  static void core2_vpmu_dump(const struct vcpu *v)
>  {
>      const struct vpmu_struct *vpmu = vcpu_vpmu(v);
> -    int i, num;
> +    int i;
>      const struct core2_vpmu_context *core2_vpmu_cxt = NULL;
>      u64 val;
>  
> @@ -698,27 +665,25 @@ static void core2_vpmu_dump(const struct vcpu *v)
>  
>      printk("    vPMU running\n");
>      core2_vpmu_cxt = vpmu->context;
> -    num = core2_get_pmc_count();
> +
>      /* Print the contents of the counter and its configuration msr. */
> -    for ( i = 0; i < num; i++ )
> +    for ( i = 0; i < arch_pmc_cnt; i++ )
>      {
>          const struct arch_msr_pair *msr_pair = core2_vpmu_cxt->arch_msr_pair;
>  
> -        if ( core2_vpmu_cxt->pmu_enable->arch_pmc_enable[i] )
> -            printk("      general_%d: 0x%016lx ctrl: 0x%016lx\n",
> -                   i, msr_pair[i].counter, msr_pair[i].control);
> +        printk("      general_%d: 0x%016lx ctrl: 0x%016lx\n",
> +               i, msr_pair[i].counter, msr_pair[i].control);
>      }
>      /*
>       * The configuration of the fixed counter is 4 bits each in the
>       * MSR_CORE_PERF_FIXED_CTR_CTRL.
>       */
> -    val = core2_vpmu_cxt->ctrls[MSR_CORE_PERF_FIXED_CTR_CTRL_IDX];
> -    for ( i = 0; i < core2_fix_counters.num; i++ )
> +    val = core2_vpmu_cxt->fixed_ctrl;
> +    for ( i = 0; i < fixed_pmc_cnt; i++ )
>      {
> -        if ( core2_vpmu_cxt->pmu_enable->fixed_ctr_enable[i] )
> -            printk("      fixed_%d:   0x%016lx ctrl: %#lx\n",
> -                   i, core2_vpmu_cxt->fix_counters[i],
> -                   val & FIXED_CTR_CTRL_MASK);
> +        printk("      fixed_%d:   0x%016lx ctrl: %#lx\n",
> +               i, core2_vpmu_cxt->fix_counters[i],
> +               val & FIXED_CTR_CTRL_MASK);
>          val >>= FIXED_CTR_CTRL_BITS;
>      }
>  }
> @@ -736,7 +701,7 @@ static int core2_vpmu_do_interrupt(struct cpu_user_regs 
> *regs)
>          if ( is_pmc_quirk )
>              handle_pmc_quirk(msr_content);
>          core2_vpmu_cxt->global_ovf_status |= msr_content;
> -        msr_content = 0xC000000700000000 | ((1 << core2_get_pmc_count()) - 
> 1);
> +        msr_content = 0xC000000700000000 | ((1 << arch_pmc_cnt) - 1);
>          wrmsrl(MSR_CORE_PERF_GLOBAL_OVF_CTRL, msr_content);
>      }
>      else
> @@ -799,18 +764,27 @@ static int core2_vpmu_initialise(struct vcpu *v, 
> unsigned int vpmu_flags)
>          }
>      }
>  func_out:
> +
> +    arch_pmc_cnt = core2_get_arch_pmc_count();
> +    fixed_pmc_cnt = core2_get_fixed_pmc_count();
> +    if ( fixed_pmc_cnt > VPMU_CORE2_MAX_FIXED_PMCS )
> +    {
> +        fixed_pmc_cnt = VPMU_CORE2_MAX_FIXED_PMCS;
> +        printk(XENLOG_G_WARNING "Limiting number of fixed counters to %d\n",
> +               fixed_pmc_cnt);
> +    }
>      check_pmc_quirk();
> +
>      return 0;
>  }
>  
>  static void core2_vpmu_destroy(struct vcpu *v)
>  {
>      struct vpmu_struct *vpmu = vcpu_vpmu(v);
> -    struct core2_vpmu_context *core2_vpmu_cxt = vpmu->context;
>  
>      if ( !vpmu_is_set(vpmu, VPMU_CONTEXT_ALLOCATED) )
>          return;
> -    xfree(core2_vpmu_cxt->pmu_enable);
> +
>      xfree(vpmu->context);
>      if ( cpu_has_vmx_msr_bitmap && is_hvm_domain(v->domain) )
>          core2_vpmu_unset_msr_bitmap(v->arch.hvm_vmx.msr_bitmap);
> diff --git a/xen/include/asm-x86/hvm/vmx/vmcs.h 
> b/xen/include/asm-x86/hvm/vmx/vmcs.h
> index ebaba5c..ed81cfb 100644
> --- a/xen/include/asm-x86/hvm/vmx/vmcs.h
> +++ b/xen/include/asm-x86/hvm/vmx/vmcs.h
> @@ -473,7 +473,9 @@ void vmx_enable_intercept_for_msr(struct vcpu *v, u32 
> msr, int type);
>  int vmx_read_guest_msr(u32 msr, u64 *val);
>  int vmx_write_guest_msr(u32 msr, u64 val);
>  int vmx_add_guest_msr(u32 msr);
> +void vmx_rm_guest_msr(u32 msr);
>  int vmx_add_host_load_msr(u32 msr);
> +void vmx_rm_host_load_msr(u32 msr);
>  void vmx_vmcs_switch(struct vmcs_struct *from, struct vmcs_struct *to);
>  void vmx_set_eoi_exit_bitmap(struct vcpu *v, u8 vector);
>  void vmx_clear_eoi_exit_bitmap(struct vcpu *v, u8 vector);
> diff --git a/xen/include/asm-x86/hvm/vmx/vpmu_core2.h 
> b/xen/include/asm-x86/hvm/vmx/vpmu_core2.h
> index 60b05fd..410372d 100644
> --- a/xen/include/asm-x86/hvm/vmx/vpmu_core2.h
> +++ b/xen/include/asm-x86/hvm/vmx/vpmu_core2.h
> @@ -23,29 +23,10 @@
>  #ifndef __ASM_X86_HVM_VPMU_CORE_H_
>  #define __ASM_X86_HVM_VPMU_CORE_H_
>  
> -/* Currently only 3 fixed counters are supported. */
> -#define VPMU_CORE2_NUM_FIXED 3
> -/* Currently only 3 Non-architectual Performance Control MSRs */
> -#define VPMU_CORE2_NUM_CTRLS 3
> -
>  struct arch_msr_pair {
>      u64 counter;
>      u64 control;
>  };
>  
> -struct core2_pmu_enable {
> -    char ds_area_enable;
> -    char fixed_ctr_enable[VPMU_CORE2_NUM_FIXED];
> -    char arch_pmc_enable[1];
> -};
> -
> -struct core2_vpmu_context {
> -    struct core2_pmu_enable *pmu_enable;
> -    u64 fix_counters[VPMU_CORE2_NUM_FIXED];
> -    u64 ctrls[VPMU_CORE2_NUM_CTRLS];
> -    u64 global_ovf_status;
> -    struct arch_msr_pair arch_msr_pair[1];
> -};
> -
>  #endif /* __ASM_X86_HVM_VPMU_CORE_H_ */
>  
> 

-- 
Company details: http://ts.fujitsu.com/imprint.html

_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.