[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-devel] [PATCH v2 1/2] x86/mem_event: Deliver gla fault EPT violation information



On 08/07/2014 03:47 PM, Tamas K Lengyel wrote:
On Intel EPT the exit qualification generated by a violation also includes a 
bit (EPT_GLA_FAULT) which describes the following information:
Set if the access causing the EPT violation is to a guest-physical address that 
is the translation of a linear address. Clear if the access causing the EPT 
violation is to a paging-structure entry as part of a page walk or the update 
of an accessed or dirty bit.

For more information see Table 27-7 in the Intel SDM.

This patch extends the mem_event system to deliver this extra information, 
which could be useful for determining the cause of a violation.

v2: Split gla_fault into fault_in_gpt and fault_gla to be more compatible with 
the AMD implementation.

Signed-off-by: Tamas K Lengyel <tamas.lengyel@xxxxxxxxxxxx>
---
  xen/arch/x86/hvm/hvm.c         |  8 ++++++--
  xen/arch/x86/hvm/svm/svm.c     |  2 +-
  xen/arch/x86/hvm/vmx/vmx.c     | 23 ++++++++++++++++++++++-
  xen/arch/x86/mm/p2m.c          |  5 ++++-
  xen/include/asm-x86/hvm/hvm.h  |  5 ++++-
  xen/include/asm-x86/p2m.h      |  3 ++-
  xen/include/public/mem_event.h |  4 +++-
  7 files changed, 42 insertions(+), 8 deletions(-)

diff --git a/xen/arch/x86/hvm/hvm.c b/xen/arch/x86/hvm/hvm.c
index e834406..d7b5e2b 100644
--- a/xen/arch/x86/hvm/hvm.c
+++ b/xen/arch/x86/hvm/hvm.c
@@ -2725,6 +2725,8 @@ void hvm_inject_page_fault(int errcode, unsigned long cr2)
  int hvm_hap_nested_page_fault(paddr_t gpa,
                                bool_t gla_valid,
                                unsigned long gla,
+                              bool_t fault_in_gpt,
+                              bool_t fault_gla,
                                bool_t access_r,
                                bool_t access_w,
                                bool_t access_x)
@@ -2832,8 +2834,10 @@ int hvm_hap_nested_page_fault(paddr_t gpa,
if ( violation )
          {
-            if ( p2m_mem_access_check(gpa, gla_valid, gla, access_r,
-                                        access_w, access_x, &req_ptr) )
+            if ( p2m_mem_access_check(gpa, gla_valid, gla,
+                                        fault_in_gpt, fault_gla,
+                                        access_r, access_w, access_x,
+                                        &req_ptr) )
              {
                  fall_through = 1;
              } else {
diff --git a/xen/arch/x86/hvm/svm/svm.c b/xen/arch/x86/hvm/svm/svm.c
index 76616ac..9e35e7a 100644
--- a/xen/arch/x86/hvm/svm/svm.c
+++ b/xen/arch/x86/hvm/svm/svm.c
@@ -1403,7 +1403,7 @@ static void svm_do_nested_pgfault(struct vcpu *v,
      p2m_access_t p2ma;
      struct p2m_domain *p2m = NULL;
- ret = hvm_hap_nested_page_fault(gpa, 0, ~0ul,
+    ret = hvm_hap_nested_page_fault(gpa, 0, ~0ul, 0, 0,


Why not pass the actual bits that the HW provides?

-boris


                                      1, /* All NPFs count as reads */
                                      npfec & PFEC_write_access,
                                      npfec & PFEC_insn_fetch);
diff --git a/xen/arch/x86/hvm/vmx/vmx.c b/xen/arch/x86/hvm/vmx/vmx.c
index 2caa04a..7ebf289 100644
--- a/xen/arch/x86/hvm/vmx/vmx.c
+++ b/xen/arch/x86/hvm/vmx/vmx.c
@@ -2353,6 +2353,7 @@ static void ept_handle_violation(unsigned long 
qualification, paddr_t gpa)
      p2m_type_t p2mt;
      int ret;
      struct domain *d = current->domain;
+    bool_t fault_in_gpt, fault_gla;
if ( tb_init_done )
      {
@@ -2371,11 +2372,31 @@ static void ept_handle_violation(unsigned long 
qualification, paddr_t gpa)
      }
if ( qualification & EPT_GLA_VALID )
+    {
          __vmread(GUEST_LINEAR_ADDRESS, &gla);
+
+        if ( qualification & EPT_GLA_FAULT )
+        {
+            fault_in_gpt = 0;
+            fault_gla = 1;
+        }
+        else
+        {
+            fault_in_gpt = 1;
+            fault_gla = 0;
+        }
+    }
      else
+    {
          gla = ~0ull;
+        fault_in_gpt = 0;
+        fault_gla = 0;
+    }
      ret = hvm_hap_nested_page_fault(gpa,
-                                    !!(qualification & EPT_GLA_VALID), gla,
+                                    !!(qualification & EPT_GLA_VALID),
+                                    gla,
+                                    fault_in_gpt,
+                                    fault_gla,
                                      !!(qualification & EPT_READ_VIOLATION),
                                      !!(qualification & EPT_WRITE_VIOLATION),
                                      !!(qualification & EPT_EXEC_VIOLATION));
diff --git a/xen/arch/x86/mm/p2m.c b/xen/arch/x86/mm/p2m.c
index bca9f0f..132f0d2 100644
--- a/xen/arch/x86/mm/p2m.c
+++ b/xen/arch/x86/mm/p2m.c
@@ -1323,7 +1323,8 @@ void p2m_mem_paging_resume(struct domain *d)
      }
  }
-bool_t p2m_mem_access_check(paddr_t gpa, bool_t gla_valid, unsigned long gla,
+bool_t p2m_mem_access_check(paddr_t gpa, bool_t gla_valid, unsigned long gla,
+                          bool_t fault_in_gpt, bool_t fault_gla,
                            bool_t access_r, bool_t access_w, bool_t access_x,
                            mem_event_request_t **req_ptr)
  {
@@ -1405,6 +1406,8 @@ bool_t p2m_mem_access_check(paddr_t gpa, bool_t 
gla_valid, unsigned long gla,
          req->offset = gpa & ((1 << PAGE_SHIFT) - 1);
          req->gla_valid = gla_valid;
          req->gla = gla;
+        req->fault_in_gpt = fault_in_gpt;
+        req->fault_gla = fault_gla;
          req->access_r = access_r;
          req->access_w = access_w;
          req->access_x = access_x;
diff --git a/xen/include/asm-x86/hvm/hvm.h b/xen/include/asm-x86/hvm/hvm.h
index 0ebd478..de755b6 100644
--- a/xen/include/asm-x86/hvm/hvm.h
+++ b/xen/include/asm-x86/hvm/hvm.h
@@ -456,7 +456,10 @@ static inline void hvm_invalidate_regs_fields(struct 
cpu_user_regs *regs)
  }
int hvm_hap_nested_page_fault(paddr_t gpa,
-                              bool_t gla_valid, unsigned long gla,
+                              bool_t gla_valid,
+                              unsigned long gla,
+                              bool_t fault_in_gpt,
+                              bool_t fault_gla,
                                bool_t access_r,
                                bool_t access_w,
                                bool_t access_x);
diff --git a/xen/include/asm-x86/p2m.h b/xen/include/asm-x86/p2m.h
index 0ddbadb..59803c5 100644
--- a/xen/include/asm-x86/p2m.h
+++ b/xen/include/asm-x86/p2m.h
@@ -597,7 +597,8 @@ void p2m_mem_paging_resume(struct domain *d);
   * been promoted with no underlying vcpu pause. If the req_ptr has been 
populated,
   * then the caller must put the event in the ring (once having released 
get_gfn*
   * locks -- caller must also xfree the request. */
-bool_t p2m_mem_access_check(paddr_t gpa, bool_t gla_valid, unsigned long gla,
+bool_t p2m_mem_access_check(paddr_t gpa, bool_t gla_valid, unsigned long gla,
+                          bool_t fault_in_gpt, bool_t fault_gla,
                            bool_t access_r, bool_t access_w, bool_t access_x,
                            mem_event_request_t **req_ptr);
  /* Resumes the running of the VCPU, restarting the last instruction */
diff --git a/xen/include/public/mem_event.h b/xen/include/public/mem_event.h
index 3831b41..1ba6863 100644
--- a/xen/include/public/mem_event.h
+++ b/xen/include/public/mem_event.h
@@ -62,7 +62,9 @@ typedef struct mem_event_st {
      uint16_t access_w:1;
      uint16_t access_x:1;
      uint16_t gla_valid:1;
-    uint16_t available:12;
+    uint16_t fault_in_gpt:1;
+    uint16_t fault_gla:1;
+    uint16_t available:10;
uint16_t reason;
  } mem_event_request_t, mem_event_response_t;


_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.