[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH] x86/AMD: Add support for AMD's OSVW feature in guests
# HG changeset patch # User Boris Ostrovsky <boris.ostrovsky@xxxxxxx> # Date 1326748089 -3600 # Node ID f157d40df95aa3b8becb970968d33f4eca6c7e75 # Parent 5b2676ac13218951698c49fa0350f2ac48220f3d x86/AMD: Add support for AMD's OSVW feature in guests. In some cases guests should not provide workarounds for errata even when the physical processor is affected. For example, because of erratum 400 on family 10h processors a Linux guest will read an MSR (resulting in VMEXIT) before going to idle in order to avoid getting stuck in a non-C0 state. This is not necessary: HLT and IO instructions are intercepted and therefore there is no reason for erratum 400 workaround in the guest. This patch allows us to present a guest with certain errata as fixed, regardless of the state of actual hardware. Signed-off-by: Boris Ostrovsky <boris.ostrovsky@xxxxxxx> diff -r 5b2676ac1321 -r f157d40df95a tools/libxc/xc_cpuid_x86.c --- a/tools/libxc/xc_cpuid_x86.c Mon Jan 09 16:01:44 2012 +0100 +++ b/tools/libxc/xc_cpuid_x86.c Mon Jan 16 22:08:09 2012 +0100 @@ -108,6 +108,7 @@ static void amd_xc_cpuid_policy( bitmaskof(X86_FEATURE_SSE4A) | bitmaskof(X86_FEATURE_MISALIGNSSE) | bitmaskof(X86_FEATURE_3DNOWPREFETCH) | + bitmaskof(X86_FEATURE_OSVW) | bitmaskof(X86_FEATURE_XOP) | bitmaskof(X86_FEATURE_FMA4) | bitmaskof(X86_FEATURE_TBM) | @@ -524,7 +525,6 @@ static void xc_cpuid_pv_policy( clear_bit(X86_FEATURE_RDTSCP, regs[3]); clear_bit(X86_FEATURE_SVM, regs[2]); - clear_bit(X86_FEATURE_OSVW, regs[2]); clear_bit(X86_FEATURE_IBS, regs[2]); clear_bit(X86_FEATURE_SKINIT, regs[2]); clear_bit(X86_FEATURE_WDT, regs[2]); diff -r 5b2676ac1321 -r f157d40df95a xen/arch/x86/cpu/amd.c --- a/xen/arch/x86/cpu/amd.c Mon Jan 09 16:01:44 2012 +0100 +++ b/xen/arch/x86/cpu/amd.c Mon Jan 16 22:08:09 2012 +0100 @@ -32,6 +32,13 @@ static char opt_famrev[14]; string_param("cpuid_mask_cpu", opt_famrev); +/* + * Set osvw_len to higher number when updated Revision Guides + * are published and we know what the new status bits are + */ +static uint64_t osvw_length = 4, osvw_status; +static DEFINE_SPINLOCK(amd_lock); + static inline void wrmsr_amd(unsigned int index, unsigned int lo, unsigned int hi) { @@ -182,6 +189,35 @@ static void __devinit set_cpuidmask(cons } } +static void amd_guest_osvw_init(struct vcpu *vcpu) +{ + if (boot_cpu_data.x86_vendor != X86_VENDOR_AMD) + return; + + /* + * Guests should see errata 400 and 415 as fixed (assuming that + * HLT and IO instructions are intercepted). + */ + vcpu->arch.amd.osvw.length = (osvw_length >= 3) ? (osvw_length) : 3; + vcpu->arch.amd.osvw.status = osvw_status & ~(6ULL); + + /* + * By increasing VCPU's osvw.length to 3 we are telling the guest that + * all osvw.status bits inside that length, including bit 0 (which is + * reserved for erratum 298), are valid. However, if host processor's + * osvw_len is 0 then osvw_status[0] carries no information. We need to + * be conservative here and therefore we tell the guest that erratum 298 + * is present (because we really don't know). + */ + if (osvw_length == 0 && boot_cpu_data.x86 == 0x10) + vcpu->arch.amd.osvw.status |= 1; +} + +void amd_vcpu_initialise(struct vcpu *vcpu) +{ + amd_guest_osvw_init(vcpu); +} + /* * Check for the presence of an AMD erratum. Arguments are defined in amd.h * for each known erratum. Return 1 if erratum is found. @@ -512,6 +548,30 @@ static void __devinit init_amd(struct cp set_cpuidmask(c); check_syscfg_dram_mod_en(); + + /* + * Get OSVW bits. If bits are not the same on different processors then + * choose the worst case (i.e. if erratum is present on one processor and + * not on another assume that the erratum is present everywhere). + */ + if (test_bit(X86_FEATURE_OSVW, &boot_cpu_data.x86_capability)) { + uint64_t len, status; + + if (rdmsr_safe(MSR_AMD_OSVW_ID_LENGTH, len) || + rdmsr_safe(MSR_AMD_OSVW_STATUS, status)) + len = status = 0; + + spin_lock(&amd_lock); + + if (len < osvw_length) + osvw_length = len; + + osvw_status |= status; + osvw_status &= (1ULL << osvw_length) - 1; + + spin_unlock(&amd_lock); + } else + osvw_length = osvw_status = 0; } static struct cpu_dev amd_cpu_dev __cpuinitdata = { diff -r 5b2676ac1321 -r f157d40df95a xen/arch/x86/domain.c --- a/xen/arch/x86/domain.c Mon Jan 09 16:01:44 2012 +0100 +++ b/xen/arch/x86/domain.c Mon Jan 16 22:08:09 2012 +0100 @@ -422,6 +422,10 @@ int vcpu_initialise(struct vcpu *v) if ( (rc = vcpu_init_fpu(v)) != 0 ) return rc; + /* Vendor-specific initialization */ + if (boot_cpu_data.x86_vendor == X86_VENDOR_AMD) + amd_vcpu_initialise(v); + if ( is_hvm_domain(d) ) { rc = hvm_vcpu_initialise(v); diff -r 5b2676ac1321 -r f157d40df95a xen/arch/x86/hvm/svm/svm.c --- a/xen/arch/x86/hvm/svm/svm.c Mon Jan 09 16:01:44 2012 +0100 +++ b/xen/arch/x86/hvm/svm/svm.c Mon Jan 16 22:08:09 2012 +0100 @@ -1044,6 +1044,27 @@ static void svm_init_erratum_383(struct } } +static int svm_handle_osvw(struct vcpu *v, uint32_t msr, uint64_t *val, uint read) +{ + uint eax, ebx, ecx, edx; + + /* Guest OSVW support */ + hvm_cpuid(0x80000001, &eax, &ebx, &ecx, &edx); + if (!test_bit((X86_FEATURE_OSVW & 31), &ecx)) + return -1; + + if (read) { + if (msr == MSR_AMD_OSVW_ID_LENGTH) + *val = v->arch.amd.osvw.length; + else + *val = v->arch.amd.osvw.status; + } + /* Writes are ignored */ + + return 0; +} + + static int svm_cpu_up(void) { uint64_t msr_content; @@ -1385,6 +1406,13 @@ static int svm_msr_read_intercept(unsign vpmu_do_rdmsr(msr, msr_content); break; + case MSR_AMD_OSVW_ID_LENGTH: + case MSR_AMD_OSVW_STATUS: + ret = svm_handle_osvw(v, msr, msr_content, 1); + if (ret < 0) + goto gpf; + break; + default: ret = nsvm_rdmsr(v, msr, msr_content); if ( ret < 0 ) @@ -1509,6 +1537,13 @@ static int svm_msr_write_intercept(unsig */ break; + case MSR_AMD_OSVW_ID_LENGTH: + case MSR_AMD_OSVW_STATUS: + ret = svm_handle_osvw(v, msr, &msr_content, 0); + if (ret < 0) + goto gpf; + break; + default: ret = nsvm_wrmsr(v, msr, msr_content); if ( ret < 0 ) diff -r 5b2676ac1321 -r f157d40df95a xen/arch/x86/traps.c --- a/xen/arch/x86/traps.c Mon Jan 09 16:01:44 2012 +0100 +++ b/xen/arch/x86/traps.c Mon Jan 16 22:08:09 2012 +0100 @@ -71,6 +71,7 @@ #include <asm/apic.h> #include <asm/mc146818rtc.h> #include <asm/hpet.h> +#include <asm/amd.h> #include <public/arch-x86/cpuid.h> #include <xsm/xsm.h> @@ -889,7 +890,6 @@ static void pv_cpuid(struct cpu_user_reg __clear_bit(X86_FEATURE_SVM % 32, &c); if ( !cpu_has_apic ) __clear_bit(X86_FEATURE_EXTAPIC % 32, &c); - __clear_bit(X86_FEATURE_OSVW % 32, &c); __clear_bit(X86_FEATURE_IBS % 32, &c); __clear_bit(X86_FEATURE_SKINIT % 32, &c); __clear_bit(X86_FEATURE_WDT % 32, &c); @@ -2542,6 +2542,15 @@ static int emulate_privileged_op(struct if ( wrmsr_safe(regs->ecx, msr_content) != 0 ) goto fail; break; + case MSR_AMD_OSVW_ID_LENGTH: + case MSR_AMD_OSVW_STATUS: + if ( boot_cpu_data.x86_vendor == X86_VENDOR_AMD ) { + if (!boot_cpu_has(X86_FEATURE_OSVW)) + goto fail; + else + break; /* Writes are ignored */ + } + /* Fall through to default case */ default: if ( wrmsr_hypervisor_regs(regs->ecx, msr_content) ) break; @@ -2573,6 +2582,7 @@ static int emulate_privileged_op(struct break; case 0x32: /* RDMSR */ + switch ( (u32)regs->ecx ) { #ifdef CONFIG_X86_64 @@ -2632,6 +2642,23 @@ static int emulate_privileged_op(struct regs->eax = (uint32_t)msr_content; regs->edx = (uint32_t)(msr_content >> 32); break; + case MSR_AMD_OSVW_ID_LENGTH: + case MSR_AMD_OSVW_STATUS: + if (boot_cpu_data.x86_vendor == X86_VENDOR_AMD) { + if (!boot_cpu_has(X86_FEATURE_OSVW)) + goto fail; + else { + if ((u32)regs->ecx == MSR_AMD_OSVW_ID_LENGTH) + msr_content = v->arch.amd.osvw.length; + else + msr_content = v->arch.amd.osvw.status; + + regs->eax = (uint32_t)msr_content; + regs->edx = (uint32_t)(msr_content >> 32); + } + } else + goto rdmsr_normal; + break; default: if ( rdmsr_hypervisor_regs(regs->ecx, &val) ) { diff -r 5b2676ac1321 -r f157d40df95a xen/include/asm-x86/amd.h --- a/xen/include/asm-x86/amd.h Mon Jan 09 16:01:44 2012 +0100 +++ b/xen/include/asm-x86/amd.h Mon Jan 16 22:08:09 2012 +0100 @@ -140,7 +140,17 @@ AMD_MODEL_RANGE(0x11, 0x0, 0x0, 0xff, 0xf), \ AMD_MODEL_RANGE(0x12, 0x0, 0x0, 0xff, 0xf)) +struct vcpu_amd { + + /* OSVW MSRs */ + struct { + u64 length; + u64 status; + } osvw; +}; + struct cpuinfo_x86; +void amd_vcpu_initialise(struct vcpu *); int cpu_has_amd_erratum(const struct cpuinfo_x86 *, int, ...); #ifdef __x86_64__ diff -r 5b2676ac1321 -r f157d40df95a xen/include/asm-x86/domain.h --- a/xen/include/asm-x86/domain.h Mon Jan 09 16:01:44 2012 +0100 +++ b/xen/include/asm-x86/domain.h Mon Jan 16 22:08:09 2012 +0100 @@ -8,6 +8,7 @@ #include <asm/hvm/domain.h> #include <asm/e820.h> #include <asm/mce.h> +#include <asm/amd.h> #include <public/vcpu.h> #define has_32bit_shinfo(d) ((d)->arch.has_32bit_shinfo) @@ -495,6 +496,11 @@ struct arch_vcpu uint32_t gdbsx_vcpu_event; + /* Vendor-specific data */ + union { + struct vcpu_amd amd; + }; + /* A secondary copy of the vcpu time info. */ XEN_GUEST_HANDLE(vcpu_time_info_t) time_info_guest; _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxx http://lists.xensource.com/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |