[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-devel] [V3 PATCH 7/9] x86/hvm: pkeys, add pkeys support for guest_walk_tables



On Wed, 2015-12-16 at 02:12 -0700, Jan Beulich wrote:
> > > > On 16.12.15 at 10:03, <huaitong.han@xxxxxxxxx> wrote:
> > On Wed, 2015-12-16 at 01:32 -0700, Jan Beulich wrote:
> > > > > > On 16.12.15 at 09:16, <huaitong.han@xxxxxxxxx> wrote:
> > > > On Tue, 2015-12-15 at 02:02 -0700, Jan Beulich wrote:
> > > > > Well, I wouldn't want you to introduce a brand new function,
> > > > > but
> > > > > instead just factor out the necessary piece from xsave()
> > > > > (making
> > > > > the new one take a struct xsave_struct * instead of a struct
> > > > > vcpu
> > > > > *,
> > > > > and calling it from what is now xsave()).
> > > > So the function looks like this:
> > > > unsigned int get_xsave_pkru(struct vcpu *v)
> > > > {
> > > >     void *offset;
> > > >     struct xsave_struct *xsave_area;
> > > >     uint64_t mask = XSTATE_PKRU;
> > > >     unsigned int index = fls64(mask) - 1;
> > > >     unsigned int pkru = 0;
> > > > 
> > > >     if ( !cpu_has_xsave )
> > > >         return 0;
> > > >     
> > > >     BUG_ON(xsave_cntxt_size < XSTATE_AREA_MIN_SIZE);
> > > >     xsave_area = _xzalloc(xsave_cntxt_size, 64);
> > > >     if ( xsave_area == NULL )
> > > >         return 0;
> > > > 
> > > >     xsave(xsave_area, mask);
> > > >     offset = (void *)xsave_area + (xsave_area_compressed(xsave)
> > > > ? 
> > > >             XSTATE_AREA_MIN_SIZE : xstate_offsets[index] );
> > > >     memcpy(&pkru, offset, sizeof(pkru));
> > > > 
> > > >     xfree(xsave_area);
> > > > 
> > > >     return pkru;
> > > > }
> > > 
> > > Depending on how frequently this might get called, the allocation
> > > overhead may not be tolerable. I.e. you may want to set up e.g.
> > > a per-CPU buffer up front. Or you check whether using RDPKRU
> > > (with temporarily setting CR4.PKE) is cheaper than what you
> > > do right now.
> > RDPKRU does cost less than the function, and if temporarily setting
> > CR4.PKE is accepted, I will use RDPKRU instead of the function.
> 
> The question isn't just the RDPKRU cost, but that of the two CR4
> writes.
Testing result with NOW() function:
Time of the function 10 times execution
(XEN)xsave time is 1376 ns
(XEN)read_pkru time is 28 ns

So, read_pkru function does cost less than get_xsave_pkru, and I will
use read_pkru.


Testing codes:
 static void reboot_machine(unsigned char key, struct cpu_user_regs
*regs)
 {
-    printk("'%c' pressed -> rebooting machine\n", key);
-    machine_restart(0);
+  //  printk("read pkru test\n", key);
+       s_time_t pre, last;
+       unsigned int pkru = 0;
+       unsigned int i = 0;
+       void * offset;
+       struct xsave_struct *save_area = _xzalloc(1024, 64);
+
+       pre = NOW();
+       for (;i< 10;i++){
+       xsave(current, XSTATE_PKRU, save_area);
+       offset = (void *)save_area + XSTATE_AREA_MIN_SIZE;
+       memcpy(&pkru, offset, sizeof(pkru));}
+       last = NOW();
+       printk("xsave time is %lu\n", (last - pre));
+
+       pre = NOW();
+       for (;i< 10;i++){
+       pkru  = read_pkru();
+       }
+       last = NOW();
+       printk("read_pkru time is %lu\n", (last - pre));
+  //  machine_restart(0);
 }

 static inline unsigned int read_pkru(void)
{
    unsigned int pkru;

    set_in_cr4(X86_CR4_PKE);
    asm volatile (".byte 0x0f,0x01,0xee"
        : "=a" (pkru) : "c" (0) : "dx");
    clear_in_cr4(X86_CR4_PKE);

    return pkru;
}

> 
> Jan
> 
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.