|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [PATCH v3 13/22] x86: XFD enabling
Just like for XCR0 we generally want to run with the guest settings of
the involved MSRs while in Xen. Hence saving/restoring of guest values
needs to only happen during context switch.
While adding the feature to libxl's table I've noticed the other XSAVE
sub-features all don't have entries there. These get added at the same
time.
Signed-off-by: Jan Beulich <jbeulich@xxxxxxxx>
---
v3: New.
---
I wasn't sure whether we want to have AMX depend on XFD. Purely from a
spec pov the two are independent features (without XFD one simply
wouldn't have a way to disable it for the purpose of lazy allocation of
state saving space).
--- a/tools/libs/light/libxl_cpuid.c
+++ b/tools/libs/light/libxl_cpuid.c
@@ -237,6 +237,12 @@ int libxl_cpuid_parse_config(libxl_cpuid
{"fsrs", 0x00000007, 1, CPUID_REG_EAX, 11, 1},
{"fsrcs", 0x00000007, 1, CPUID_REG_EAX, 12, 1},
+ {"xsaveopt", 0x0000000d, 1, CPUID_REG_EAX, 0, 1},
+ {"xsavec", 0x0000000d, 1, CPUID_REG_EAX, 1, 1},
+ {"xgetbv1", 0x0000000d, 1, CPUID_REG_EAX, 2, 1},
+ {"xsaves", 0x0000000d, 1, CPUID_REG_EAX, 3, 1},
+ {"xfd", 0x0000000d, 1, CPUID_REG_EAX, 4, 1},
+
{"lahfsahf", 0x80000001, NA, CPUID_REG_ECX, 0, 1},
{"cmplegacy", 0x80000001, NA, CPUID_REG_ECX, 1, 1},
{"svm", 0x80000001, NA, CPUID_REG_ECX, 2, 1},
--- a/tools/misc/xen-cpuid.c
+++ b/tools/misc/xen-cpuid.c
@@ -116,6 +116,7 @@ static const char *const str_Da1[32] =
{
[ 0] = "xsaveopt", [ 1] = "xsavec",
[ 2] = "xgetbv1", [ 3] = "xsaves",
+ [ 4] = "xfd",
};
static const char *const str_7c0[32] =
--- a/xen/arch/x86/cpuid.c
+++ b/xen/arch/x86/cpuid.c
@@ -230,6 +230,7 @@ static void recalculate_xstate(struct cp
p->xstate.comp[i].offset = xstate_offset(i);
p->xstate.comp[i].xss = curr_xstate & XSTATE_XSAVES_ONLY;
p->xstate.comp[i].align = curr_xstate & xstate_align;
+ p->xstate.comp[i].xfd = curr_xstate & xfd_mask;
}
}
--- a/xen/arch/x86/domain.c
+++ b/xen/arch/x86/domain.c
@@ -1837,6 +1837,12 @@ void paravirt_ctxt_switch_to(struct vcpu
if ( cpu_has_msr_tsc_aux )
wrmsr_tsc_aux(v->arch.msrs->tsc_aux);
+
+ if ( v->domain->arch.cpuid->xstate.xfd )
+ {
+ wrmsrl(MSR_XFD, v->arch.msrs->xfd);
+ wrmsrl(MSR_XFD_ERR, v->arch.msrs->xfd_err);
+ }
}
/* Update per-VCPU guest runstate shared memory area (if registered). */
--- a/xen/arch/x86/domctl.c
+++ b/xen/arch/x86/domctl.c
@@ -1101,6 +1101,8 @@ long arch_do_domctl(
static const uint32_t msrs_to_send[] = {
MSR_SPEC_CTRL,
MSR_INTEL_MISC_FEATURES_ENABLES,
+ MSR_XFD,
+ MSR_XFD_ERR,
MSR_TSC_AUX,
MSR_AMD64_DR0_ADDRESS_MASK,
MSR_AMD64_DR1_ADDRESS_MASK,
--- a/xen/arch/x86/hvm/vmx/vmcs.c
+++ b/xen/arch/x86/hvm/vmx/vmcs.c
@@ -1178,6 +1178,12 @@ static int construct_vmcs(struct vcpu *v
if ( (vmexit_ctl & VM_EXIT_CLEAR_BNDCFGS) &&
(vmentry_ctl & VM_ENTRY_LOAD_BNDCFGS) )
vmx_clear_msr_intercept(v, MSR_IA32_BNDCFGS, VMX_MSR_RW);
+
+ if ( d->arch.cpuid->xstate.xfd )
+ {
+ vmx_clear_msr_intercept(v, MSR_XFD, VMX_MSR_RW);
+ vmx_clear_msr_intercept(v, MSR_XFD_ERR, VMX_MSR_RW);
+ }
}
/* I/O access bitmap. */
--- a/xen/arch/x86/hvm/vmx/vmx.c
+++ b/xen/arch/x86/hvm/vmx/vmx.c
@@ -539,6 +539,12 @@ static void vmx_save_guest_msrs(struct v
*/
v->arch.hvm.vmx.shadow_gs = read_gs_shadow();
+ if ( v->domain->arch.cpuid->xstate.xfd )
+ {
+ rdmsrl(MSR_XFD, msrs->xfd);
+ rdmsrl(MSR_XFD_ERR, msrs->xfd_err);
+ }
+
if ( v->arch.hvm.vmx.ipt_active )
{
rdmsrl(MSR_RTIT_OUTPUT_MASK, msrs->rtit.output_mask);
@@ -558,6 +564,12 @@ static void vmx_restore_guest_msrs(struc
if ( cpu_has_msr_tsc_aux )
wrmsr_tsc_aux(msrs->tsc_aux);
+ if ( v->domain->arch.cpuid->xstate.xfd )
+ {
+ wrmsrl(MSR_XFD, msrs->xfd);
+ wrmsrl(MSR_XFD_ERR, msrs->xfd_err);
+ }
+
if ( v->arch.hvm.vmx.ipt_active )
{
wrmsrl(MSR_RTIT_OUTPUT_BASE, page_to_maddr(v->vmtrace.pg));
--- a/xen/arch/x86/msr.c
+++ b/xen/arch/x86/msr.c
@@ -29,6 +29,7 @@
#include <asm/hvm/viridian.h>
#include <asm/msr.h>
#include <asm/setup.h>
+#include <asm/xstate.h>
#include <public/hvm/params.h>
@@ -267,6 +268,18 @@ int guest_rdmsr(struct vcpu *v, uint32_t
*val = msrs->misc_features_enables.raw;
break;
+ case MSR_XFD:
+ case MSR_XFD_ERR:
+ if ( !cp->xstate.xfd )
+ goto gp_fault;
+ if ( v == curr && is_hvm_domain(d) )
+ rdmsrl(msr, *val);
+ else if ( msr == MSR_XFD )
+ *val = msrs->xfd;
+ else
+ *val = msrs->xfd_err;
+ break;
+
case MSR_IA32_MCG_CAP ... MSR_IA32_MCG_CTL: /* 0x179 -> 0x17b */
case MSR_IA32_MCx_CTL2(0) ... MSR_IA32_MCx_CTL2(31): /* 0x280 -> 0x29f */
case MSR_IA32_MCx_CTL(0) ... MSR_IA32_MCx_MISC(31): /* 0x400 -> 0x47f */
@@ -523,6 +536,19 @@ int guest_wrmsr(struct vcpu *v, uint32_t
break;
}
+ case MSR_XFD:
+ case MSR_XFD_ERR:
+ if ( !cp->xstate.xfd ||
+ (val & ~(cpuid_policy_xstates(cp) & xfd_mask)) )
+ goto gp_fault;
+ if ( v == curr )
+ wrmsrl(msr, val);
+ if ( msr == MSR_XFD )
+ msrs->xfd = val;
+ else
+ msrs->xfd_err = val;
+ break;
+
case MSR_IA32_MCG_CAP ... MSR_IA32_MCG_CTL: /* 0x179 -> 0x17b */
case MSR_IA32_MCx_CTL2(0) ... MSR_IA32_MCx_CTL2(31): /* 0x280 -> 0x29f */
case MSR_IA32_MCx_CTL(0) ... MSR_IA32_MCx_MISC(31): /* 0x400 -> 0x47f */
--- a/xen/arch/x86/xstate.c
+++ b/xen/arch/x86/xstate.c
@@ -29,6 +29,10 @@
u64 __read_mostly xfeature_mask;
u64 __read_mostly xstate_align;
+
+/* Mask of XSAVE/XRSTOR features supporting XFD. */
+uint64_t __read_mostly xfd_mask;
+
static unsigned int __read_mostly xstate_features;
uint32_t __read_mostly mxcsr_mask = 0x0000ffbf;
@@ -106,6 +110,11 @@ static int setup_xstate_features(bool bs
__set_bit(leaf, &xstate_align);
else
BUG_ON(!(ecx & XSTATE_ALIGN64) != !test_bit(leaf, &xstate_align));
+
+ if ( bsp && (ecx & XSTATE_XFD) )
+ __set_bit(leaf, &xfd_mask);
+ else
+ BUG_ON(!(ecx & XSTATE_XFD) != !test_bit(leaf, &xfd_mask));
}
return 0;
--- a/xen/include/asm-x86/msr-index.h
+++ b/xen/include/asm-x86/msr-index.h
@@ -69,6 +69,10 @@
#define MSR_MCU_OPT_CTRL 0x00000123
#define MCU_OPT_CTRL_RNGDS_MITG_DIS (_AC(1, ULL) << 0)
+/* Bits within these two MSRs share positions with XCR0. */
+#define MSR_XFD 0x000001c4
+#define MSR_XFD_ERR 0x000001c5
+
#define MSR_RTIT_OUTPUT_BASE 0x00000560
#define MSR_RTIT_OUTPUT_MASK 0x00000561
#define MSR_RTIT_CTL 0x00000570
--- a/xen/include/asm-x86/msr.h
+++ b/xen/include/asm-x86/msr.h
@@ -343,6 +343,12 @@ struct vcpu_msrs
uint64_t raw;
} xss;
+ /* 0x000001c4 - MSR_XFD */
+ uint64_t xfd;
+
+ /* 0x000001c5 - MSR_XFD_ERR */
+ uint64_t xfd_err;
+
/*
* 0xc0000103 - MSR_TSC_AUX
*
--- a/xen/include/asm-x86/xstate.h
+++ b/xen/include/asm-x86/xstate.h
@@ -39,9 +39,11 @@ extern uint32_t mxcsr_mask;
#define XSTATE_COMPACTION_ENABLED (1ULL << 63)
#define XSTATE_ALIGN64 (1U << 1)
+#define XSTATE_XFD (1U << 2)
extern u64 xfeature_mask;
extern u64 xstate_align;
+extern uint64_t xfd_mask;
#define xstate_offset(n) (raw_cpuid_policy.xstate.comp[n].offset)
#define xstate_size(n) (raw_cpuid_policy.xstate.comp[n].size)
--- a/xen/include/public/arch-x86/cpufeatureset.h
+++ b/xen/include/public/arch-x86/cpufeatureset.h
@@ -191,6 +191,7 @@ XEN_CPUFEATURE(XSAVEOPT, 4*32+ 0) /
XEN_CPUFEATURE(XSAVEC, 4*32+ 1) /*A XSAVEC/XRSTORC instructions */
XEN_CPUFEATURE(XGETBV1, 4*32+ 2) /*A XGETBV with %ecx=1 */
XEN_CPUFEATURE(XSAVES, 4*32+ 3) /*S XSAVES/XRSTORS instructions */
+XEN_CPUFEATURE(XFD, 4*32+ 4) /* XFD / XFD_ERR MSRs */
/* Intel-defined CPU features, CPUID level 0x00000007:0.ebx, word 5 */
XEN_CPUFEATURE(FSGSBASE, 5*32+ 0) /*A {RD,WR}{FS,GS}BASE instructions */
--- a/xen/include/xen/lib/x86/cpuid.h
+++ b/xen/include/xen/lib/x86/cpuid.h
@@ -222,7 +222,7 @@ struct cpuid_policy
/* Per-component common state. Valid for i >= 2. */
struct {
uint32_t size, offset;
- bool xss:1, align:1;
+ bool xss:1, align:1, xfd:1;
uint32_t _res_d;
} comp[CPUID_GUEST_NR_XSTATE];
} xstate;
--- a/xen/tools/gen-cpuid.py
+++ b/xen/tools/gen-cpuid.py
@@ -221,7 +221,7 @@ def crunch_numbers(state):
# are instructions built on top of base XSAVE, while others are new
# instruction groups which are specified to require XSAVE for state
# management.
- XSAVE: [XSAVEOPT, XSAVEC, XGETBV1, XSAVES,
+ XSAVE: [XSAVEOPT, XSAVEC, XGETBV1, XSAVES, XFD,
AVX, MPX, PKU, AMX_TILE, LWP],
# AVX is taken to mean hardware support for 256bit registers (which in
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |