vm-assist: slightly reduce source code size Adjust VM_ASSIST() to prepend VMASST_TYPE_. Signed-off-by: Jan Beulich --- v2: Split original patch into two. --- a/xen/arch/x86/domain.c +++ b/xen/arch/x86/domain.c @@ -972,7 +972,7 @@ int arch_set_info_guest( case -ERESTART: break; case 0: - if ( !compat && !VM_ASSIST(d, VMASST_TYPE_m2p_strict) && + if ( !compat && !VM_ASSIST(d, m2p_strict) && !paging_mode_refcounts(d) ) fill_ro_mpt(cr3_gfn); break; @@ -1011,7 +1011,7 @@ int arch_set_info_guest( cr3_page = NULL; break; case 0: - if ( VM_ASSIST(d, VMASST_TYPE_m2p_strict) ) + if ( VM_ASSIST(d, m2p_strict) ) zap_ro_mpt(cr3_gfn); break; } --- a/xen/arch/x86/mm.c +++ b/xen/arch/x86/mm.c @@ -1464,7 +1464,7 @@ static int alloc_l4_table(struct page_in adjust_guest_l4e(pl4e[i], d); } - init_guest_l4_table(pl4e, d, !VM_ASSIST(d, VMASST_TYPE_m2p_strict)); + init_guest_l4_table(pl4e, d, !VM_ASSIST(d, m2p_strict)); unmap_domain_page(pl4e); return rc > 0 ? 0 : rc; @@ -2774,7 +2774,7 @@ int new_guest_cr3(unsigned long mfn) invalidate_shadow_ldt(curr, 0); - if ( !VM_ASSIST(d, VMASST_TYPE_m2p_strict) && !paging_mode_refcounts(d) ) + if ( !VM_ASSIST(d, m2p_strict) && !paging_mode_refcounts(d) ) fill_ro_mpt(mfn); curr->arch.guest_table = pagetable_from_pfn(mfn); update_cr3(curr); @@ -3133,8 +3133,7 @@ long do_mmuext_op( op.arg1.mfn); break; } - if ( VM_ASSIST(d, VMASST_TYPE_m2p_strict) && - !paging_mode_refcounts(d) ) + if ( VM_ASSIST(d, m2p_strict) && !paging_mode_refcounts(d) ) zap_ro_mpt(op.arg1.mfn); } --- a/xen/arch/x86/mm/shadow/multi.c +++ b/xen/arch/x86/mm/shadow/multi.c @@ -1435,7 +1435,7 @@ void sh_install_xen_entries_in_l4(struct shadow_l4e_from_mfn(page_to_mfn(d->arch.perdomain_l3_pg), __PAGE_HYPERVISOR); - if ( !VM_ASSIST(d, VMASST_TYPE_m2p_strict) ) + if ( !VM_ASSIST(d, m2p_strict) ) sl4e[shadow_l4_table_offset(RO_MPT_VIRT_START)] = shadow_l4e_empty(); /* Shadow linear mapping for 4-level shadows. N.B. for 3-level @@ -3983,11 +3983,11 @@ sh_update_cr3(struct vcpu *v, int do_loc shadow_l4e_t *sl4e = v->arch.paging.shadow.guest_vtable; if ( (v->arch.flags & TF_kernel_mode) && - !VM_ASSIST(d, VMASST_TYPE_m2p_strict) ) + !VM_ASSIST(d, m2p_strict) ) sl4e[shadow_l4_table_offset(RO_MPT_VIRT_START)] = idle_pg_table[l4_table_offset(RO_MPT_VIRT_START)]; else if ( !(v->arch.flags & TF_kernel_mode) && - VM_ASSIST(d, VMASST_TYPE_m2p_strict) ) + VM_ASSIST(d, m2p_strict) ) sl4e[shadow_l4_table_offset(RO_MPT_VIRT_START)] = shadow_l4e_empty(); } --- a/xen/arch/x86/traps.c +++ b/xen/arch/x86/traps.c @@ -1441,7 +1441,7 @@ static int fixup_page_fault(unsigned lon !(regs->error_code & (PFEC_reserved_bit | PFEC_insn_fetch)) && (regs->error_code & PFEC_write_access) ) { - if ( VM_ASSIST(d, VMASST_TYPE_writable_pagetables) && + if ( VM_ASSIST(d, writable_pagetables) && /* Do not check if access-protection fault since the page may legitimately be not present in shadow page tables */ (paging_mode_enabled(d) || --- a/xen/common/kernel.c +++ b/xen/common/kernel.c @@ -306,7 +306,7 @@ DO(xen_version)(int cmd, XEN_GUEST_HANDL { case 0: fi.submap = (1U << XENFEAT_memory_op_vnode_supported); - if ( VM_ASSIST(d, VMASST_TYPE_pae_extended_cr3) ) + if ( VM_ASSIST(d, pae_extended_cr3) ) fi.submap |= (1U << XENFEAT_pae_pgdir_above_4gb); if ( paging_mode_translate(d) ) fi.submap |= --- a/xen/include/xen/sched.h +++ b/xen/include/xen/sched.h @@ -833,7 +833,7 @@ void watchdog_domain_destroy(struct doma /* This check is for functionality specific to a control domain */ #define is_control_domain(_d) ((_d)->is_privileged) -#define VM_ASSIST(_d,_t) (test_bit((_t), &(_d)->vm_assist)) +#define VM_ASSIST(d, t) (test_bit(VMASST_TYPE_ ## t, &(d)->vm_assist)) #define is_pv_domain(d) ((d)->guest_type == guest_type_pv) #define is_pv_vcpu(v) (is_pv_domain((v)->domain))