|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH 1/6] x86: stop handling MSR_IA32_BNDCFGS save/restore in implementation code
Saving and restoring the value of this MSR is currently handled by
implementation-specific code despite it being architectural. This patch
moves handling of accesses to this MSR from hvm.c into the msr.c, thus
allowing the common MSR save/restore code to handle it.
This patch also changes hvm_get/set_guest_bndcfgs() to check CPUID policy
for the appropriate feature, rather than hardware, and also re-works
the get/set_guest_bndcfgs() hvm_funcs so they are no longer boolean. Uses
of u64 are also converted to uint64_t.
NOTE: Because vmx_get/set_guest_bndcfgs() call vmx_vmcs_enter(), the
struct vcpu pointer passed in cannot be const.
Signed-off-by: Paul Durrant <paul.durrant@xxxxxxxxxx>
---
Cc: Jan Beulich <jbeulich@xxxxxxxx>
Cc: Andrew Cooper <andrew.cooper3@xxxxxxxxxx>
Cc: Wei Liu <wei.liu2@xxxxxxxxxx>
Cc: "Roger Pau Monné" <roger.pau@xxxxxxxxxx>
Cc: Jun Nakajima <jun.nakajima@xxxxxxxxx>
Cc: Kevin Tian <kevin.tian@xxxxxxxxx>
---
xen/arch/x86/hvm/hvm.c | 44 +++++++++++++++++++++-------------
xen/arch/x86/hvm/vmx/vmx.c | 31 +++++-------------------
xen/arch/x86/msr.c | 14 ++++++++++-
xen/arch/x86/pv/emul-priv-op.c | 2 +-
xen/include/asm-x86/hvm/hvm.h | 13 ++++------
xen/include/asm-x86/msr.h | 2 +-
6 files changed, 52 insertions(+), 54 deletions(-)
diff --git a/xen/arch/x86/hvm/hvm.c b/xen/arch/x86/hvm/hvm.c
index 401c4a9312..5fd5478b7d 100644
--- a/xen/arch/x86/hvm/hvm.c
+++ b/xen/arch/x86/hvm/hvm.c
@@ -308,11 +308,16 @@ int hvm_set_guest_pat(struct vcpu *v, u64 guest_pat)
return 1;
}
-bool hvm_set_guest_bndcfgs(struct vcpu *v, u64 val)
+bool hvm_set_guest_bndcfgs(struct vcpu *v, uint64_t val)
{
- if ( !hvm_funcs.set_guest_bndcfgs ||
- !is_canonical_address(val) ||
- (val & IA32_BNDCFGS_RESERVED) )
+ const struct cpuid_policy *cp = v->domain->arch.cpuid;
+
+ if ( !cp->feat.mpx )
+ return false;
+
+ ASSERT(hvm_funcs.set_guest_bndcfgs);
+
+ if ( !is_canonical_address(val) || (val & IA32_BNDCFGS_RESERVED) )
return false;
/*
@@ -342,7 +347,22 @@ bool hvm_set_guest_bndcfgs(struct vcpu *v, u64 val)
/* nothing, best effort only */;
}
- return hvm_funcs.set_guest_bndcfgs(v, val);
+ hvm_funcs.set_guest_bndcfgs(v, val);
+
+ return true;
+}
+
+bool hvm_get_guest_bndcfgs(struct vcpu *v, uint64_t *val)
+{
+ const struct cpuid_policy *cp = v->domain->arch.cpuid;
+
+ if ( !cp->feat.mpx )
+ return false;
+
+ ASSERT(hvm_funcs.get_guest_bndcfgs);
+ *val = hvm_funcs.get_guest_bndcfgs(v);
+
+ return true;
}
/*
@@ -1312,6 +1332,7 @@ static int hvm_load_cpu_xsave_states(struct domain *d,
hvm_domain_context_t *h)
static const uint32_t msrs_to_send[] = {
MSR_SPEC_CTRL,
MSR_INTEL_MISC_FEATURES_ENABLES,
+ MSR_IA32_BNDCFGS,
MSR_AMD64_DR0_ADDRESS_MASK,
MSR_AMD64_DR1_ADDRESS_MASK,
MSR_AMD64_DR2_ADDRESS_MASK,
@@ -1449,6 +1470,7 @@ static int hvm_load_cpu_msrs(struct domain *d,
hvm_domain_context_t *h)
case MSR_SPEC_CTRL:
case MSR_INTEL_MISC_FEATURES_ENABLES:
+ case MSR_IA32_BNDCFGS:
case MSR_AMD64_DR0_ADDRESS_MASK:
case MSR_AMD64_DR1_ADDRESS_MASK ... MSR_AMD64_DR3_ADDRESS_MASK:
rc = guest_wrmsr(v, ctxt->msr[i].index, ctxt->msr[i].val);
@@ -3472,12 +3494,6 @@ int hvm_msr_read_intercept(unsigned int msr, uint64_t
*msr_content)
*msr_content = v->arch.hvm.msr_xss;
break;
- case MSR_IA32_BNDCFGS:
- if ( !d->arch.cpuid->feat.mpx ||
- !hvm_get_guest_bndcfgs(v, msr_content) )
- goto gp_fault;
- break;
-
case MSR_K8_ENABLE_C1E:
case MSR_AMD64_NB_CFG:
/*
@@ -3624,12 +3640,6 @@ int hvm_msr_write_intercept(unsigned int msr, uint64_t
msr_content,
v->arch.hvm.msr_xss = msr_content;
break;
- case MSR_IA32_BNDCFGS:
- if ( !d->arch.cpuid->feat.mpx ||
- !hvm_set_guest_bndcfgs(v, msr_content) )
- goto gp_fault;
- break;
-
case MSR_AMD64_NB_CFG:
/* ignore the write */
break;
diff --git a/xen/arch/x86/hvm/vmx/vmx.c b/xen/arch/x86/hvm/vmx/vmx.c
index 64af8bf943..4bfabe8d0e 100644
--- a/xen/arch/x86/hvm/vmx/vmx.c
+++ b/xen/arch/x86/hvm/vmx/vmx.c
@@ -805,17 +805,6 @@ static unsigned int __init vmx_init_msr(void)
static void vmx_save_msr(struct vcpu *v, struct hvm_msr *ctxt)
{
- vmx_vmcs_enter(v);
-
- if ( cpu_has_mpx && cpu_has_vmx_mpx )
- {
- __vmread(GUEST_BNDCFGS, &ctxt->msr[ctxt->count].val);
- if ( ctxt->msr[ctxt->count].val )
- ctxt->msr[ctxt->count++].index = MSR_IA32_BNDCFGS;
- }
-
- vmx_vmcs_exit(v);
-
if ( cpu_has_xsaves && cpu_has_vmx_xsaves )
{
ctxt->msr[ctxt->count].val = v->arch.hvm.msr_xss;
@@ -835,14 +824,6 @@ static int vmx_load_msr(struct vcpu *v, struct hvm_msr
*ctxt)
{
switch ( ctxt->msr[i].index )
{
- case MSR_IA32_BNDCFGS:
- if ( cpu_has_mpx && cpu_has_vmx_mpx &&
- is_canonical_address(ctxt->msr[i].val) &&
- !(ctxt->msr[i].val & IA32_BNDCFGS_RESERVED) )
- __vmwrite(GUEST_BNDCFGS, ctxt->msr[i].val);
- else if ( ctxt->msr[i].val )
- err = -ENXIO;
- break;
case MSR_IA32_XSS:
if ( cpu_has_xsaves && cpu_has_vmx_xsaves )
v->arch.hvm.msr_xss = ctxt->msr[i].val;
@@ -1204,26 +1185,26 @@ static int vmx_get_guest_pat(struct vcpu *v, u64 *gpat)
return 1;
}
-static bool vmx_set_guest_bndcfgs(struct vcpu *v, u64 val)
+static void vmx_set_guest_bndcfgs(struct vcpu *v, uint64_t val)
{
ASSERT(cpu_has_mpx && cpu_has_vmx_mpx);
vmx_vmcs_enter(v);
__vmwrite(GUEST_BNDCFGS, val);
vmx_vmcs_exit(v);
-
- return true;
}
-static bool vmx_get_guest_bndcfgs(struct vcpu *v, u64 *val)
+static uint64_t vmx_get_guest_bndcfgs(struct vcpu *v)
{
+ uint64_t val;
+
ASSERT(cpu_has_mpx && cpu_has_vmx_mpx);
vmx_vmcs_enter(v);
- __vmread(GUEST_BNDCFGS, val);
+ __vmread(GUEST_BNDCFGS, &val);
vmx_vmcs_exit(v);
- return true;
+ return val;
}
static void vmx_handle_cd(struct vcpu *v, unsigned long value)
diff --git a/xen/arch/x86/msr.c b/xen/arch/x86/msr.c
index 9bb38b6d66..a3406c29a8 100644
--- a/xen/arch/x86/msr.c
+++ b/xen/arch/x86/msr.c
@@ -115,7 +115,7 @@ int init_vcpu_msr_policy(struct vcpu *v)
return 0;
}
-int guest_rdmsr(const struct vcpu *v, uint32_t msr, uint64_t *val)
+int guest_rdmsr(struct vcpu *v, uint32_t msr, uint64_t *val)
{
const struct vcpu *curr = current;
const struct domain *d = v->domain;
@@ -158,6 +158,12 @@ int guest_rdmsr(const struct vcpu *v, uint32_t msr,
uint64_t *val)
ret = guest_rdmsr_x2apic(v, msr, val);
break;
+ case MSR_IA32_BNDCFGS:
+ if ( !is_hvm_domain(d) || !hvm_get_guest_bndcfgs(v, val) )
+ goto gp_fault;
+
+ break;
+
case 0x40000000 ... 0x400001ff:
if ( is_viridian_domain(d) )
{
@@ -319,6 +325,12 @@ int guest_wrmsr(struct vcpu *v, uint32_t msr, uint64_t val)
ret = guest_wrmsr_x2apic(v, msr, val);
break;
+ case MSR_IA32_BNDCFGS:
+ if ( !is_hvm_domain(d) || !hvm_set_guest_bndcfgs(v, val) )
+ goto gp_fault;
+
+ break;
+
case 0x40000000 ... 0x400001ff:
if ( is_viridian_domain(d) )
{
diff --git a/xen/arch/x86/pv/emul-priv-op.c b/xen/arch/x86/pv/emul-priv-op.c
index 942ece2ca0..678dad3792 100644
--- a/xen/arch/x86/pv/emul-priv-op.c
+++ b/xen/arch/x86/pv/emul-priv-op.c
@@ -807,7 +807,7 @@ static inline bool is_cpufreq_controller(const struct
domain *d)
static int read_msr(unsigned int reg, uint64_t *val,
struct x86_emulate_ctxt *ctxt)
{
- const struct vcpu *curr = current;
+ struct vcpu *curr = current;
const struct domain *currd = curr->domain;
bool vpmu_msr = false;
int ret;
diff --git a/xen/include/asm-x86/hvm/hvm.h b/xen/include/asm-x86/hvm/hvm.h
index 0a10b51554..5c8237e087 100644
--- a/xen/include/asm-x86/hvm/hvm.h
+++ b/xen/include/asm-x86/hvm/hvm.h
@@ -149,8 +149,8 @@ struct hvm_function_table {
int (*get_guest_pat)(struct vcpu *v, u64 *);
int (*set_guest_pat)(struct vcpu *v, u64);
- bool (*get_guest_bndcfgs)(struct vcpu *v, u64 *);
- bool (*set_guest_bndcfgs)(struct vcpu *v, u64);
+ uint64_t (*get_guest_bndcfgs)(struct vcpu *v);
+ void (*set_guest_bndcfgs)(struct vcpu *v, uint64_t);
void (*set_tsc_offset)(struct vcpu *v, u64 offset, u64 at_tsc);
@@ -283,8 +283,6 @@ void hvm_get_segment_register(struct vcpu *v, enum
x86_segment seg,
void hvm_set_segment_register(struct vcpu *v, enum x86_segment seg,
struct segment_register *reg);
-bool hvm_set_guest_bndcfgs(struct vcpu *v, u64 val);
-
bool hvm_check_cpuid_faulting(struct vcpu *v);
void hvm_migrate_timers(struct vcpu *v);
void hvm_do_resume(struct vcpu *v);
@@ -446,11 +444,8 @@ static inline unsigned long hvm_get_shadow_gs_base(struct
vcpu *v)
return hvm_funcs.get_shadow_gs_base(v);
}
-static inline bool hvm_get_guest_bndcfgs(struct vcpu *v, u64 *val)
-{
- return hvm_funcs.get_guest_bndcfgs &&
- hvm_funcs.get_guest_bndcfgs(v, val);
-}
+bool hvm_get_guest_bndcfgs(struct vcpu *v, uint64_t *val);
+bool hvm_set_guest_bndcfgs(struct vcpu *v, uint64_t val);
#define has_hvm_params(d) \
((d)->arch.hvm.params != NULL)
diff --git a/xen/include/asm-x86/msr.h b/xen/include/asm-x86/msr.h
index adfa2fa05b..ad8688a61f 100644
--- a/xen/include/asm-x86/msr.h
+++ b/xen/include/asm-x86/msr.h
@@ -320,7 +320,7 @@ int init_vcpu_msr_policy(struct vcpu *v);
* These functions are also used by the migration logic, so need to cope with
* being used outside of v's context.
*/
-int guest_rdmsr(const struct vcpu *v, uint32_t msr, uint64_t *val);
+int guest_rdmsr(struct vcpu *v, uint32_t msr, uint64_t *val);
int guest_wrmsr(struct vcpu *v, uint32_t msr, uint64_t val);
#endif /* !__ASSEMBLY__ */
--
2.20.1
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxxx
https://lists.xenproject.org/mailman/listinfo/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |