[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH 1/6] x86/mce: adapt mce.{c, h} to Xen hypervisor coding style
Signed-off-by: Haozhong Zhang <haozhong.zhang@xxxxxxxxx> --- xen/arch/x86/cpu/mcheck/mce.c | 536 +++++++++++++++++++++++------------------- xen/arch/x86/cpu/mcheck/mce.h | 21 +- 2 files changed, 311 insertions(+), 246 deletions(-) diff --git a/xen/arch/x86/cpu/mcheck/mce.c b/xen/arch/x86/cpu/mcheck/mce.c index 7affe2591e..580e68d6f2 100644 --- a/xen/arch/x86/cpu/mcheck/mce.c +++ b/xen/arch/x86/cpu/mcheck/mce.c @@ -64,7 +64,7 @@ struct mca_banks *mca_allbanks; int mce_verbosity; static int __init mce_set_verbosity(const char *str) { - if (strcmp("verbose", str) == 0) + if ( strcmp("verbose", str) == 0 ) mce_verbosity = MCE_VERBOSE; else return -EINVAL; @@ -81,7 +81,6 @@ static void unexpected_machine_check(const struct cpu_user_regs *regs) fatal_trap(regs, 1); } - static x86_mce_vector_t _machine_check_vector = unexpected_machine_check; void x86_mce_vector_register(x86_mce_vector_t hdlr) @@ -97,11 +96,13 @@ void do_machine_check(const struct cpu_user_regs *regs) _machine_check_vector(regs); } -/* Init machine check callback handler +/* + * Init machine check callback handler * It is used to collect additional information provided by newer * CPU families/models without the need to duplicate the whole handler. * This avoids having many handlers doing almost nearly the same and each - * with its own tweaks ands bugs. */ + * with its own tweaks ands bugs. + */ static x86_mce_callback_t mc_callback_bank_extended = NULL; void x86_mce_callback_register(x86_mce_callback_t cbfunc) @@ -109,7 +110,8 @@ void x86_mce_callback_register(x86_mce_callback_t cbfunc) mc_callback_bank_extended = cbfunc; } -/* Machine check recoverable judgement callback handler +/* + * Machine check recoverable judgement callback handler * It is used to judge whether an UC error is recoverable by software */ static mce_recoverable_t mc_recoverable_scan = NULL; @@ -124,12 +126,12 @@ struct mca_banks *mcabanks_alloc(void) struct mca_banks *mb; mb = xmalloc(struct mca_banks); - if (!mb) + if ( !mb ) return NULL; mb->bank_map = xzalloc_array(unsigned long, BITS_TO_LONGS(nr_mce_banks)); - if (!mb->bank_map) + if ( !mb->bank_map ) { xfree(mb); return NULL; @@ -142,9 +144,9 @@ struct mca_banks *mcabanks_alloc(void) void mcabanks_free(struct mca_banks *banks) { - if (banks == NULL) + if ( banks == NULL ) return; - if (banks->bank_map) + if ( banks->bank_map ) xfree(banks->bank_map); xfree(banks); } @@ -155,15 +157,16 @@ static void mcabank_clear(int banknum) status = mca_rdmsr(MSR_IA32_MCx_STATUS(banknum)); - if (status & MCi_STATUS_ADDRV) + if ( status & MCi_STATUS_ADDRV ) mca_wrmsr(MSR_IA32_MCx_ADDR(banknum), 0x0ULL); - if (status & MCi_STATUS_MISCV) + if ( status & MCi_STATUS_MISCV ) mca_wrmsr(MSR_IA32_MCx_MISC(banknum), 0x0ULL); mca_wrmsr(MSR_IA32_MCx_STATUS(banknum), 0x0ULL); } -/* Judging whether to Clear Machine Check error bank callback handler +/* + * Judging whether to Clear Machine Check error bank callback handler * According to Intel latest MCA OS Recovery Writer's Guide, * whether the error MCA bank needs to be cleared is decided by the mca_source * and MCi_status bit value. @@ -188,17 +191,15 @@ const struct mca_error_handler *__read_mostly mce_uhandlers; unsigned int __read_mostly mce_dhandler_num; unsigned int __read_mostly mce_uhandler_num; - -static void mca_init_bank(enum mca_source who, - struct mc_info *mi, int bank) +static void mca_init_bank(enum mca_source who, struct mc_info *mi, int bank) { struct mcinfo_bank *mib; - if (!mi) + if ( !mi ) return; mib = x86_mcinfo_reserve(mi, sizeof(*mib), MC_TYPE_BANK); - if (!mib) + if ( !mib ) { mi->flags |= MCINFO_FLAGS_UNCOMPLETE; return; @@ -209,26 +210,27 @@ static void mca_init_bank(enum mca_source who, mib->mc_bank = bank; mib->mc_domid = DOMID_INVALID; - if (mib->mc_status & MCi_STATUS_MISCV) + if ( mib->mc_status & MCi_STATUS_MISCV ) mib->mc_misc = mca_rdmsr(MSR_IA32_MCx_MISC(bank)); - if (mib->mc_status & MCi_STATUS_ADDRV) + if ( mib->mc_status & MCi_STATUS_ADDRV ) mib->mc_addr = mca_rdmsr(MSR_IA32_MCx_ADDR(bank)); - if ((mib->mc_status & MCi_STATUS_MISCV) && - (mib->mc_status & MCi_STATUS_ADDRV) && - (mc_check_addr(mib->mc_status, mib->mc_misc, MC_ADDR_PHYSICAL)) && - (who == MCA_POLLER || who == MCA_CMCI_HANDLER) && - (mfn_valid(_mfn(paddr_to_pfn(mib->mc_addr))))) + if ( (mib->mc_status & MCi_STATUS_MISCV) && + (mib->mc_status & MCi_STATUS_ADDRV) && + (mc_check_addr(mib->mc_status, mib->mc_misc, MC_ADDR_PHYSICAL)) && + (who == MCA_POLLER || who == MCA_CMCI_HANDLER) && + (mfn_valid(_mfn(paddr_to_pfn(mib->mc_addr)))) ) { struct domain *d; d = maddr_get_owner(mib->mc_addr); - if (d) + if ( d ) mib->mc_domid = d->domain_id; } - if (who == MCA_CMCI_HANDLER) { + if ( who == MCA_CMCI_HANDLER ) + { mib->mc_ctrl2 = mca_rdmsr(MSR_IA32_MC0_CTL2 + bank); mib->mc_tsc = rdtsc(); } @@ -252,7 +254,8 @@ static int mca_init_global(uint32_t flags, struct mcinfo_global *mig) &mig->mc_coreid, &mig->mc_core_threadid, &mig->mc_apicid, NULL, NULL, NULL); - if (curr != INVALID_VCPU) { + if ( curr != INVALID_VCPU ) + { mig->mc_domid = curr->domain->domain_id; mig->mc_vcpuid = curr->vcpu_id; } @@ -260,14 +263,17 @@ static int mca_init_global(uint32_t flags, struct mcinfo_global *mig) return 0; } -/* Utility function to perform MCA bank telemetry readout and to push that +/* + * Utility function to perform MCA bank telemetry readout and to push that * telemetry towards an interested dom0 for logging and diagnosis. * The caller - #MC handler or MCA poll function - must arrange that we - * do not migrate cpus. */ + * do not migrate cpus. + */ /* XXFM Could add overflow counting? */ -/* Add out_param clear_bank for Machine Check Handler Caller. +/* + * Add out_param clear_bank for Machine Check Handler Caller. * For Intel latest CPU, whether to clear the error bank status needs to * be judged by the callback function defined above. */ @@ -286,7 +292,8 @@ mcheck_mca_logout(enum mca_source who, struct mca_banks *bankmask, int i; gstatus = mca_rdmsr(MSR_IA32_MCG_STATUS); - switch (who) { + switch ( who ) + { case MCA_MCE_SCAN: mc_flags = MC_FLAG_MCE; which = MC_URGENT; @@ -307,34 +314,42 @@ mcheck_mca_logout(enum mca_source who, struct mca_banks *bankmask, BUG(); } - /* If no mc_recovery_scan callback handler registered, + /* + * If no mc_recovery_scan callback handler registered, * this error is not recoverable */ - recover = (mc_recoverable_scan) ? 1 : 0; + recover = mc_recoverable_scan ? 1 : 0; - for (i = 0; i < nr_mce_banks; i++) { + for ( i = 0; i < nr_mce_banks; i++ ) + { /* Skip bank if corresponding bit in bankmask is clear */ - if (!mcabanks_test(i, bankmask)) + if ( !mcabanks_test(i, bankmask) ) continue; status = mca_rdmsr(MSR_IA32_MCx_STATUS(i)); - if (!(status & MCi_STATUS_VAL)) + if ( !(status & MCi_STATUS_VAL) ) continue; /* this bank has no valid telemetry */ - /* For Intel Latest CPU CMCI/MCE Handler caller, we need to + /* + * For Intel Latest CPU CMCI/MCE Handler caller, we need to * decide whether to clear bank by MCi_STATUS bit value such as * OVER/UC/EN/PCC/S/AR */ if ( mc_need_clearbank_scan ) need_clear = mc_need_clearbank_scan(who, status); - /* If this is the first bank with valid MCA DATA, then + /* + * If this is the first bank with valid MCA DATA, then * try to reserve an entry from the urgent/nonurgent queue * depending on whether we are called from an exception or * a poller; this can fail (for example dom0 may not - * yet have consumed past telemetry). */ - if (errcnt++ == 0) { - if ( (mctc = mctelem_reserve(which)) != NULL ) { + * yet have consumed past telemetry). + */ + if ( errcnt++ == 0 ) + { + mctc = mctelem_reserve(which); + if ( mctc ) + { mci = mctelem_dataptr(mctc); mcinfo_clear(mci); mig = x86_mcinfo_reserve(mci, sizeof(*mig), MC_TYPE_GLOBAL); @@ -342,49 +357,50 @@ mcheck_mca_logout(enum mca_source who, struct mca_banks *bankmask, ASSERT(mig); mca_init_global(mc_flags, mig); /* A hook here to get global extended msrs */ - if (boot_cpu_data.x86_vendor == X86_VENDOR_INTEL) + if ( boot_cpu_data.x86_vendor == X86_VENDOR_INTEL ) intel_get_extended_msrs(mig, mci); } } /* flag for uncorrected errors */ - if (!uc && ((status & MCi_STATUS_UC) != 0)) + if ( !uc && ((status & MCi_STATUS_UC) != 0) ) uc = true; /* flag processor context corrupt */ - if (!pcc && ((status & MCi_STATUS_PCC) != 0)) + if ( !pcc && ((status & MCi_STATUS_PCC) != 0) ) pcc = true; - if (recover && uc) + if ( recover && uc ) /* uc = true, recover = true, we need not panic. */ recover = mc_recoverable_scan(status); mca_init_bank(who, mci, i); - if (mc_callback_bank_extended) + if ( mc_callback_bank_extended ) mc_callback_bank_extended(mci, i, status); /* By default, need_clear = true */ - if (who != MCA_MCE_SCAN && need_clear) + if ( who != MCA_MCE_SCAN && need_clear ) /* Clear bank */ mcabank_clear(i); - else if ( who == MCA_MCE_SCAN && need_clear) + else if ( who == MCA_MCE_SCAN && need_clear ) mcabanks_set(i, clear_bank); wmb(); } - if (mig && errcnt > 0) { - if (pcc) + if ( mig && errcnt > 0 ) + { + if ( pcc ) mig->mc_flags |= MC_FLAG_UNCORRECTABLE; - else if (uc) + else if ( uc ) mig->mc_flags |= MC_FLAG_RECOVERABLE; else mig->mc_flags |= MC_FLAG_CORRECTABLE; } - - if (sp) { + if ( sp ) + { sp->errcnt = errcnt; sp->ripv = (gstatus & MCG_STATUS_RIPV) != 0; sp->eipv = (gstatus & MCG_STATUS_EIPV) != 0; @@ -399,19 +415,20 @@ mcheck_mca_logout(enum mca_source who, struct mca_banks *bankmask, static void mce_spin_lock(spinlock_t *lk) { - while (!spin_trylock(lk)) { - cpu_relax(); - mce_panic_check(); - } + while ( !spin_trylock(lk) ) + { + cpu_relax(); + mce_panic_check(); + } } static void mce_spin_unlock(spinlock_t *lk) { - spin_unlock(lk); + spin_unlock(lk); } static enum mce_result mce_action(const struct cpu_user_regs *regs, - mctelem_cookie_t mctc); + mctelem_cookie_t mctc); /* * Return: @@ -419,11 +436,11 @@ static enum mce_result mce_action(const struct cpu_user_regs *regs, * 0: Continue to next step */ static int mce_urgent_action(const struct cpu_user_regs *regs, - mctelem_cookie_t mctc) + mctelem_cookie_t mctc) { uint64_t gstatus; - if ( mctc == NULL) + if ( mctc == NULL ) return 0; gstatus = mca_rdmsr(MSR_IA32_MCG_STATUS); @@ -460,50 +477,49 @@ void mcheck_cmn_handler(const struct cpu_user_regs *regs) mce_spin_lock(&mce_logout_lock); - if (clear_bank != NULL) { - memset( clear_bank->bank_map, 0x0, - sizeof(long) * BITS_TO_LONGS(clear_bank->num)); - } + if ( clear_bank != NULL ) + memset(clear_bank->bank_map, 0x0, + sizeof(long) * BITS_TO_LONGS(clear_bank->num)); mctc = mcheck_mca_logout(MCA_MCE_SCAN, bankmask, &bs, clear_bank); lmce = bs.lmce; bcast = mce_broadcast && !lmce; - if (bs.errcnt) { + if ( bs.errcnt ) + { /* * Uncorrected errors must be dealt with in softirq context. */ - if (bs.uc || bs.pcc) { + if ( bs.uc || bs.pcc ) + { add_taint(TAINT_MACHINE_CHECK); - if (mctc != NULL) + if ( mctc ) mctelem_defer(mctc, lmce); /* * For PCC=1 and can't be recovered, context is lost, so * reboot now without clearing the banks, and deal with * the telemetry after reboot (the MSRs are sticky) */ - if (bs.pcc || !bs.recoverable) + if ( bs.pcc || !bs.recoverable ) cpumask_set_cpu(smp_processor_id(), &mce_fatal_cpus); - } else { - if (mctc != NULL) - mctelem_commit(mctc); } + else if ( mctc != NULL ) + mctelem_commit(mctc); atomic_set(&found_error, 1); /* The last CPU will be take check/clean-up etc */ atomic_set(&severity_cpu, smp_processor_id()); mce_printk(MCE_CRITICAL, "MCE: clear_bank map %lx on CPU%d\n", - *((unsigned long*)clear_bank), smp_processor_id()); - if (clear_bank != NULL) + *((unsigned long *)clear_bank), smp_processor_id()); + if ( clear_bank != NULL ) mcheck_mca_clearbanks(clear_bank); - } else { - if (mctc != NULL) - mctelem_dismiss(mctc); } + else if ( mctc != NULL ) + mctelem_dismiss(mctc); mce_spin_unlock(&mce_logout_lock); mce_barrier_enter(&mce_trap_bar, bcast); - if ( mctc != NULL && mce_urgent_action(regs, mctc)) + if ( mctc != NULL && mce_urgent_action(regs, mctc) ) cpumask_set_cpu(smp_processor_id(), &mce_fatal_cpus); mce_barrier_exit(&mce_trap_bar, bcast); @@ -511,14 +527,16 @@ void mcheck_cmn_handler(const struct cpu_user_regs *regs) * Wait until everybody has processed the trap. */ mce_barrier_enter(&mce_trap_bar, bcast); - if (lmce || atomic_read(&severity_cpu) == smp_processor_id()) { - /* According to SDM, if no error bank found on any cpus, + if ( lmce || atomic_read(&severity_cpu) == smp_processor_id() ) + { + /* + * According to SDM, if no error bank found on any cpus, * something unexpected happening, we can't do any * recovery job but to reset the system. */ - if (atomic_read(&found_error) == 0) + if ( atomic_read(&found_error) == 0 ) mc_panic("MCE: No CPU found valid MCE, need reset"); - if (!cpumask_empty(&mce_fatal_cpus)) + if ( !cpumask_empty(&mce_fatal_cpus) ) { char *ebufp, ebuf[96] = "MCE: Fatal error happened on CPUs "; ebufp = ebuf + strlen(ebuf); @@ -533,7 +551,8 @@ void mcheck_cmn_handler(const struct cpu_user_regs *regs) /* Clear flags after above fatal check */ mce_barrier_enter(&mce_trap_bar, bcast); gstatus = mca_rdmsr(MSR_IA32_MCG_STATUS); - if ((gstatus & MCG_STATUS_MCIP) != 0) { + if ( (gstatus & MCG_STATUS_MCIP) != 0 ) + { mce_printk(MCE_CRITICAL, "MCE: Clear MCIP@ last step"); mca_wrmsr(MSR_IA32_MCG_STATUS, 0); } @@ -546,8 +565,9 @@ void mcheck_mca_clearbanks(struct mca_banks *bankmask) { int i; - for (i = 0; i < nr_mce_banks; i++) { - if (!mcabanks_test(i, bankmask)) + for ( i = 0; i < nr_mce_banks; i++ ) + { + if ( !mcabanks_test(i, bankmask) ) continue; mcabank_clear(i); } @@ -565,19 +585,16 @@ bool mce_available(const struct cpuinfo_x86 *c) */ unsigned int mce_firstbank(struct cpuinfo_x86 *c) { - if (c->x86 == 6) { - if (c->x86_vendor == X86_VENDOR_INTEL && c->x86_model < 0x1a) - return 1; - } - - return 0; + return c->x86 == 6 && + c->x86_vendor == X86_VENDOR_INTEL && c->x86_model < 0x1a; } int show_mca_info(int inited, struct cpuinfo_x86 *c) { static enum mcheck_type g_type = mcheck_unset; - if (inited != g_type) { + if ( inited != g_type ) + { char prefix[20]; static const char *const type_str[] = { [mcheck_amd_famXX] = "AMD", @@ -589,7 +606,8 @@ int show_mca_info(int inited, struct cpuinfo_x86 *c) g_type != mcheck_unset ? XENLOG_WARNING : XENLOG_INFO, smp_processor_id()); BUG_ON(inited >= ARRAY_SIZE(type_str)); - switch (inited) { + switch ( inited ) + { default: printk("%s%s machine check reporting enabled\n", prefix, type_str[inited]); @@ -618,14 +636,16 @@ static void set_poll_bankmask(struct cpuinfo_x86 *c) mb = per_cpu(poll_bankmask, cpu); BUG_ON(!mb); - if (cmci_support && opt_mce) { + if ( cmci_support && opt_mce ) + { mb->num = per_cpu(no_cmci_banks, cpu)->num; bitmap_copy(mb->bank_map, per_cpu(no_cmci_banks, cpu)->bank_map, nr_mce_banks); } - else { + else + { bitmap_copy(mb->bank_map, mca_allbanks->bank_map, nr_mce_banks); - if (mce_firstbank(c)) + if ( mce_firstbank(c) ) mcabanks_clear(0, mb); } } @@ -637,10 +657,10 @@ int mca_cap_init(void) rdmsrl(MSR_IA32_MCG_CAP, msr_content); - if (msr_content & MCG_CTL_P) /* Control register present ? */ + if ( msr_content & MCG_CTL_P ) /* Control register present ? */ wrmsrl(MSR_IA32_MCG_CTL, 0xffffffffffffffffULL); - if (nr_mce_banks && (msr_content & MCG_CAP_COUNT) != nr_mce_banks) + if ( nr_mce_banks && (msr_content & MCG_CAP_COUNT) != nr_mce_banks ) { dprintk(XENLOG_WARNING, "Different bank number on cpu %x\n", smp_processor_id()); @@ -648,7 +668,7 @@ int mca_cap_init(void) } nr_mce_banks = msr_content & MCG_CAP_COUNT; - if (!nr_mce_banks) + if ( !nr_mce_banks ) { printk(XENLOG_INFO "CPU%u: No MCE banks present. " "Machine check support disabled\n", smp_processor_id()); @@ -656,16 +676,16 @@ int mca_cap_init(void) } /* mcabanks_alloc depends on nr_mce_banks */ - if (!mca_allbanks) + if ( !mca_allbanks ) { int i; mca_allbanks = mcabanks_alloc(); - for ( i = 0; i < nr_mce_banks; i++) + for ( i = 0; i < nr_mce_banks; i++ ) mcabanks_set(i, mca_allbanks); } - return mca_allbanks ? 0:-ENOMEM; + return mca_allbanks ? 0 : -ENOMEM; } static void cpu_bank_free(unsigned int cpu) @@ -731,7 +751,7 @@ void mcheck_init(struct cpuinfo_x86 *c, bool bsp) return; } - if (!mce_available(c)) + if ( !mce_available(c) ) { printk(XENLOG_INFO "CPU%i: No machine check support available\n", smp_processor_id()); @@ -739,20 +759,22 @@ void mcheck_init(struct cpuinfo_x86 *c, bool bsp) } /*Hardware Enable */ - if (mca_cap_init()) + if ( mca_cap_init() ) return; /* Early MCE initialisation for BSP. */ if ( bsp && cpu_bank_alloc(smp_processor_id()) ) BUG(); - switch (c->x86_vendor) { + switch ( c->x86_vendor ) + { case X86_VENDOR_AMD: inited = amd_mcheck_init(c); break; case X86_VENDOR_INTEL: - switch (c->x86) { + switch ( c->x86 ) + { case 6: case 15: inited = intel_mcheck_init(c, bsp); @@ -765,7 +787,7 @@ void mcheck_init(struct cpuinfo_x86 *c, bool bsp) } show_mca_info(inited, c); - if (inited == mcheck_none || inited == mcheck_unset) + if ( inited == mcheck_none || inited == mcheck_unset ) goto out; intpose_init(); @@ -807,15 +829,14 @@ void *x86_mcinfo_reserve(struct mc_info *mi, mic_index = mic_base = x86_mcinfo_first(mi); /* go to first free entry */ - for (i = 0; i < x86_mcinfo_nentries(mi); i++) { + for ( i = 0; i < x86_mcinfo_nentries(mi); i++ ) mic_index = x86_mcinfo_next(mic_index); - } /* check if there is enough size */ end1 = (unsigned long)((uint8_t *)mic_base + sizeof(struct mc_info)); end2 = (unsigned long)((uint8_t *)mic_index + size); - if (end1 < end2) + if ( end1 < end2 ) { mce_printk(MCE_CRITICAL, "mcinfo_add: No space left in mc_info\n"); @@ -854,9 +875,11 @@ static void x86_mcinfo_apei_save( apei_write_mce(&m); } -/* Dump machine check information in a format, +/* + * Dump machine check information in a format, * mcelog can parse. This is used only when - * Dom0 does not take the notification. */ + * Dom0 does not take the notification. + */ void x86_mcinfo_dump(struct mc_info *mi) { struct mcinfo_common *mic = NULL; @@ -865,27 +888,26 @@ void x86_mcinfo_dump(struct mc_info *mi) /* first print the global info */ x86_mcinfo_lookup(mic, mi, MC_TYPE_GLOBAL); - if (mic == NULL) + if ( mic == NULL ) return; mc_global = (struct mcinfo_global *)mic; - if (mc_global->mc_flags & MC_FLAG_MCE) { + if ( mc_global->mc_flags & MC_FLAG_MCE ) printk(XENLOG_WARNING "CPU%d: Machine Check Exception: %16"PRIx64"\n", mc_global->mc_coreid, mc_global->mc_gstatus); - } else if (mc_global->mc_flags & MC_FLAG_CMCI) { + else if ( mc_global->mc_flags & MC_FLAG_CMCI ) printk(XENLOG_WARNING "CMCI occurred on CPU %d.\n", mc_global->mc_coreid); - } else if (mc_global->mc_flags & MC_FLAG_POLLED) { + else if ( mc_global->mc_flags & MC_FLAG_POLLED ) printk(XENLOG_WARNING "POLLED occurred on CPU %d.\n", mc_global->mc_coreid); - } /* then the bank information */ x86_mcinfo_lookup(mic, mi, MC_TYPE_BANK); /* finds the first entry */ do { - if (mic == NULL) + if ( mic == NULL ) return; - if (mic->type != MC_TYPE_BANK) + if ( mic->type != MC_TYPE_BANK ) goto next; mc_bank = (struct mcinfo_bank *)mic; @@ -893,20 +915,20 @@ void x86_mcinfo_dump(struct mc_info *mi) printk(XENLOG_WARNING "Bank %d: %16"PRIx64, mc_bank->mc_bank, mc_bank->mc_status); - if (mc_bank->mc_status & MCi_STATUS_MISCV) + if ( mc_bank->mc_status & MCi_STATUS_MISCV ) printk("[%16"PRIx64"]", mc_bank->mc_misc); - if (mc_bank->mc_status & MCi_STATUS_ADDRV) + if ( mc_bank->mc_status & MCi_STATUS_ADDRV ) printk(" at %16"PRIx64, mc_bank->mc_addr); printk("\n"); - if (is_mc_panic) + if ( is_mc_panic ) x86_mcinfo_apei_save(mc_global, mc_bank); - next: + next: mic = x86_mcinfo_next(mic); /* next entry */ - if ((mic == NULL) || (mic->size == 0)) + if ( (mic == NULL) || (mic->size == 0) ) break; - } while (1); + } while ( 1 ); } static void do_mc_get_cpu_info(void *v) @@ -925,8 +947,9 @@ static void do_mc_get_cpu_info(void *v) /* * Deal with sparse masks, condensed into a contig array. */ - while (cpn >= 0) { - if (cpu_online(cpn)) + while ( cpn >= 0 ) + { + if ( cpu_online(cpn) ) cindex++; cpn--; } @@ -956,14 +979,15 @@ static void do_mc_get_cpu_info(void *v) xcp->mc_msrvalues[0].reg = MSR_IA32_MCG_CAP; rdmsrl(MSR_IA32_MCG_CAP, xcp->mc_msrvalues[0].value); - if (c->cpuid_level >= 1) { + if ( c->cpuid_level >= 1 ) + { cpuid(1, &junk, &ebx, &junk, &junk); xcp->mc_clusterid = (ebx >> 24) & 0xff; - } else + } + else xcp->mc_clusterid = get_apic_id(); } - void x86_mc_get_cpu_info(unsigned cpu, uint32_t *chipid, uint16_t *coreid, uint16_t *threadid, uint32_t *apicid, unsigned *ncores, unsigned *ncores_active, @@ -973,28 +997,31 @@ void x86_mc_get_cpu_info(unsigned cpu, uint32_t *chipid, uint16_t *coreid, *apicid = cpu_physical_id(cpu); c = &cpu_data[cpu]; - if (c->apicid == BAD_APICID) { + if ( c->apicid == BAD_APICID ) + { *chipid = cpu; *coreid = 0; *threadid = 0; - if (ncores != NULL) + if ( ncores != NULL ) *ncores = 1; - if (ncores_active != NULL) + if ( ncores_active != NULL ) *ncores_active = 1; - if (nthreads != NULL) + if ( nthreads != NULL ) *nthreads = 1; - } else { + } + else + { *chipid = c->phys_proc_id; - if (c->x86_max_cores > 1) + if ( c->x86_max_cores > 1 ) *coreid = c->cpu_core_id; else *coreid = 0; *threadid = c->apicid & ((1 << (c->x86_num_siblings - 1)) - 1); - if (ncores != NULL) + if ( ncores != NULL ) *ncores = c->x86_max_cores; - if (ncores_active != NULL) + if ( ncores_active != NULL ) *ncores_active = c->booted_cores; - if (nthreads != NULL) + if ( nthreads != NULL ) *nthreads = c->x86_num_siblings; } } @@ -1002,7 +1029,7 @@ void x86_mc_get_cpu_info(unsigned cpu, uint32_t *chipid, uint16_t *coreid, #define INTPOSE_NENT 50 static struct intpose_ent { - unsigned int cpu_nr; + unsigned int cpu_nr; uint64_t msr; uint64_t val; } intpose_arr[INTPOSE_NENT]; @@ -1012,12 +1039,11 @@ static void intpose_init(void) static int done; int i; - if (done++ > 0) + if ( done++ > 0 ) return; - for (i = 0; i < INTPOSE_NENT; i++) { + for ( i = 0; i < INTPOSE_NENT; i++ ) intpose_arr[i].cpu_nr = -1; - } } @@ -1026,10 +1052,11 @@ struct intpose_ent *intpose_lookup(unsigned int cpu_nr, uint64_t msr, { int i; - for (i = 0; i < INTPOSE_NENT; i++) { - if (intpose_arr[i].cpu_nr == cpu_nr && - intpose_arr[i].msr == msr) { - if (valp != NULL) + for ( i = 0; i < INTPOSE_NENT; i++ ) + { + if ( intpose_arr[i].cpu_nr == cpu_nr && intpose_arr[i].msr == msr ) + { + if ( valp != NULL ) *valp = intpose_arr[i].val; return &intpose_arr[i]; } @@ -1040,16 +1067,19 @@ struct intpose_ent *intpose_lookup(unsigned int cpu_nr, uint64_t msr, static void intpose_add(unsigned int cpu_nr, uint64_t msr, uint64_t val) { - struct intpose_ent *ent; + struct intpose_ent *ent = intpose_lookup(cpu_nr, msr, NULL); int i; - if ((ent = intpose_lookup(cpu_nr, msr, NULL)) != NULL) { + if ( ent ) + { ent->val = val; return; } - for (i = 0, ent = &intpose_arr[0]; i < INTPOSE_NENT; i++, ent++) { - if (ent->cpu_nr == -1) { + for ( i = 0, ent = &intpose_arr[0]; i < INTPOSE_NENT; i++, ent++ ) + { + if ( ent->cpu_nr == -1 ) + { ent->cpu_nr = cpu_nr; ent->msr = msr; ent->val = val; @@ -1083,50 +1113,60 @@ static bool x86_mc_msrinject_verify(struct xen_mc_msrinject *mci) c = &cpu_data[smp_processor_id()]; - for (i = 0; i < mci->mcinj_count; i++) { + for ( i = 0; i < mci->mcinj_count; i++ ) + { uint64_t reg = mci->mcinj_msr[i].reg; const char *reason = NULL; - if (IS_MCA_BANKREG(reg)) { - if (c->x86_vendor == X86_VENDOR_AMD) { - /* On AMD we can set MCi_STATUS_WREN in the + if ( IS_MCA_BANKREG(reg) ) + { + if ( c->x86_vendor == X86_VENDOR_AMD ) + { + /* + * On AMD we can set MCi_STATUS_WREN in the * HWCR MSR to allow non-zero writes to banks * MSRs not to #GP. The injector in dom0 * should set that bit, but we detect when it * is necessary and set it as a courtesy to - * avoid #GP in the hypervisor. */ + * avoid #GP in the hypervisor. + */ mci->mcinj_flags |= _MC_MSRINJ_F_REQ_HWCR_WREN; continue; - } else { - /* No alternative but to interpose, so require - * that the injector specified as such. */ - if (!(mci->mcinj_flags & - MC_MSRINJ_F_INTERPOSE)) { + } + else + { + /* + * No alternative but to interpose, so require + * that the injector specified as such. + */ + if ( !(mci->mcinj_flags & MC_MSRINJ_F_INTERPOSE) ) reason = "must specify interposition"; - } } - } else { - switch (reg) { - /* MSRs acceptable on all x86 cpus */ + } + else + { + switch ( reg ) + { + /* MSRs acceptable on all x86 cpus */ case MSR_IA32_MCG_STATUS: break; case MSR_F10_MC4_MISC1: case MSR_F10_MC4_MISC2: case MSR_F10_MC4_MISC3: - if (c->x86_vendor != X86_VENDOR_AMD) + if ( c->x86_vendor != X86_VENDOR_AMD ) reason = "only supported on AMD"; - else if (c->x86 < 0x10) + else if ( c->x86 < 0x10 ) reason = "only supported on AMD Fam10h+"; break; - /* MSRs that the HV will take care of */ + /* MSRs that the HV will take care of */ case MSR_K8_HWCR: - if (c->x86_vendor == X86_VENDOR_AMD) + if ( c->x86_vendor == X86_VENDOR_AMD ) reason = "HV will operate HWCR"; else - reason ="only supported on AMD"; + reason = "only supported on AMD"; break; default: @@ -1135,7 +1175,8 @@ static bool x86_mc_msrinject_verify(struct xen_mc_msrinject *mci) } } - if (reason != NULL) { + if ( reason != NULL ) + { printk("HV MSR INJECT ERROR: MSR %#Lx %s\n", (unsigned long long)mci->mcinj_msr[i].reg, reason); errs++; @@ -1151,7 +1192,8 @@ static uint64_t x86_mc_hwcr_wren(void) rdmsrl(MSR_K8_HWCR, old); - if (!(old & K8_HWCR_MCi_STATUS_WREN)) { + if ( !(old & K8_HWCR_MCi_STATUS_WREN) ) + { uint64_t new = old | K8_HWCR_MCi_STATUS_WREN; wrmsrl(MSR_K8_HWCR, new); } @@ -1161,7 +1203,7 @@ static uint64_t x86_mc_hwcr_wren(void) static void x86_mc_hwcr_wren_restore(uint64_t hwcr) { - if (!(hwcr & K8_HWCR_MCi_STATUS_WREN)) + if ( !(hwcr & K8_HWCR_MCi_STATUS_WREN) ) wrmsrl(MSR_K8_HWCR, hwcr); } @@ -1173,26 +1215,26 @@ static void x86_mc_msrinject(void *data) int intpose; int i; - if (mci->mcinj_flags & _MC_MSRINJ_F_REQ_HWCR_WREN) + if ( mci->mcinj_flags & _MC_MSRINJ_F_REQ_HWCR_WREN ) hwcr = x86_mc_hwcr_wren(); intpose = (mci->mcinj_flags & MC_MSRINJ_F_INTERPOSE) != 0; - for (i = 0, msr = &mci->mcinj_msr[0]; - i < mci->mcinj_count; i++, msr++) { + for ( i = 0, msr = &mci->mcinj_msr[0]; i < mci->mcinj_count; i++, msr++ ) + { printk("HV MSR INJECT (%s) target %u actual %u MSR %#Lx <-- %#Lx\n", - intpose ? "interpose" : "hardware", + intpose ? "interpose" : "hardware", mci->mcinj_cpunr, smp_processor_id(), (unsigned long long)msr->reg, (unsigned long long)msr->value); - if (intpose) + if ( intpose ) intpose_add(mci->mcinj_cpunr, msr->reg, msr->value); else wrmsrl(msr->reg, msr->value); } - if (mci->mcinj_flags & _MC_MSRINJ_F_REQ_HWCR_WREN) + if ( mci->mcinj_flags & _MC_MSRINJ_F_REQ_HWCR_WREN ) x86_mc_hwcr_wren_restore(hwcr); } @@ -1297,12 +1339,14 @@ long do_mca(XEN_GUEST_HANDLE_PARAM(xen_mc_t) u_xen_mc) if ( op->interface_version != XEN_MCA_INTERFACE_VERSION ) return x86_mcerr("do_mca: interface version mismatch", -EACCES); - switch (op->cmd) { + switch ( op->cmd ) + { case XEN_MC_fetch: mc_fetch.nat = &op->u.mc_fetch; cmdflags = mc_fetch.nat->flags; - switch (cmdflags & (XEN_MC_NONURGENT | XEN_MC_URGENT)) { + switch ( cmdflags & (XEN_MC_NONURGENT | XEN_MC_URGENT) ) + { case XEN_MC_NONURGENT: which = MC_NONURGENT; break; @@ -1317,30 +1361,37 @@ long do_mca(XEN_GUEST_HANDLE_PARAM(xen_mc_t) u_xen_mc) flags = XEN_MC_OK; - if (cmdflags & XEN_MC_ACK) { + if ( cmdflags & XEN_MC_ACK ) + { mctelem_cookie_t cookie = ID2COOKIE(mc_fetch.nat->fetch_id); mctelem_ack(which, cookie); - } else { - if (!is_pv_32bit_vcpu(v) - ? guest_handle_is_null(mc_fetch.nat->data) - : compat_handle_is_null(mc_fetch.cmp->data)) + } + else + { + if ( !is_pv_32bit_vcpu(v) + ? guest_handle_is_null(mc_fetch.nat->data) + : compat_handle_is_null(mc_fetch.cmp->data) ) return x86_mcerr("do_mca fetch: guest buffer " "invalid", -EINVAL); - if ((mctc = mctelem_consume_oldest_begin(which))) { + mctc = mctelem_consume_oldest_begin(which); + if ( mctc ) + { struct mc_info *mcip = mctelem_dataptr(mctc); - if (!is_pv_32bit_vcpu(v) - ? copy_to_guest(mc_fetch.nat->data, mcip, 1) - : copy_to_compat(mc_fetch.cmp->data, - mcip, 1)) { + if ( !is_pv_32bit_vcpu(v) + ? copy_to_guest(mc_fetch.nat->data, mcip, 1) + : copy_to_compat(mc_fetch.cmp->data, mcip, 1) ) + { ret = -EFAULT; flags |= XEN_MC_FETCHFAILED; mc_fetch.nat->fetch_id = 0; - } else { - mc_fetch.nat->fetch_id = COOKIE2ID(mctc); } + else + mc_fetch.nat->fetch_id = COOKIE2ID(mctc); mctelem_consume_oldest_end(mctc); - } else { + } + else + { /* There is no data */ flags |= XEN_MC_NODATA; mc_fetch.nat->fetch_id = 0; @@ -1360,48 +1411,47 @@ long do_mca(XEN_GUEST_HANDLE_PARAM(xen_mc_t) u_xen_mc) mc_physcpuinfo.nat = &op->u.mc_physcpuinfo; nlcpu = num_online_cpus(); - if (!is_pv_32bit_vcpu(v) - ? !guest_handle_is_null(mc_physcpuinfo.nat->info) - : !compat_handle_is_null(mc_physcpuinfo.cmp->info)) { - if (mc_physcpuinfo.nat->ncpus <= 0) + if ( !is_pv_32bit_vcpu(v) + ? !guest_handle_is_null(mc_physcpuinfo.nat->info) + : !compat_handle_is_null(mc_physcpuinfo.cmp->info) ) + { + if ( mc_physcpuinfo.nat->ncpus <= 0 ) return x86_mcerr("do_mca cpuinfo: ncpus <= 0", -EINVAL); nlcpu = min(nlcpu, (int)mc_physcpuinfo.nat->ncpus); log_cpus = xmalloc_array(xen_mc_logical_cpu_t, nlcpu); - if (log_cpus == NULL) + if ( log_cpus == NULL ) return x86_mcerr("do_mca cpuinfo", -ENOMEM); on_each_cpu(do_mc_get_cpu_info, log_cpus, 1); - if (!is_pv_32bit_vcpu(v) - ? copy_to_guest(mc_physcpuinfo.nat->info, - log_cpus, nlcpu) - : copy_to_compat(mc_physcpuinfo.cmp->info, - log_cpus, nlcpu)) + if ( !is_pv_32bit_vcpu(v) + ? copy_to_guest(mc_physcpuinfo.nat->info, log_cpus, nlcpu) + : copy_to_compat(mc_physcpuinfo.cmp->info, log_cpus, nlcpu) ) ret = -EFAULT; xfree(log_cpus); } mc_physcpuinfo.nat->ncpus = nlcpu; - if (copy_to_guest(u_xen_mc, op, 1)) + if ( copy_to_guest(u_xen_mc, op, 1) ) return x86_mcerr("do_mca cpuinfo", -EFAULT); break; case XEN_MC_msrinject: - if (nr_mce_banks == 0) + if ( nr_mce_banks == 0 ) return x86_mcerr("do_mca inject", -ENODEV); mc_msrinject = &op->u.mc_msrinject; target = mc_msrinject->mcinj_cpunr; - if (target >= nr_cpu_ids) + if ( target >= nr_cpu_ids ) return x86_mcerr("do_mca inject: bad target", -EINVAL); - if (!cpu_online(target)) + if ( !cpu_online(target) ) return x86_mcerr("do_mca inject: target offline", -EINVAL); - if (mc_msrinject->mcinj_count == 0) + if ( mc_msrinject->mcinj_count == 0 ) return 0; if ( mc_msrinject->mcinj_flags & MC_MSRINJ_F_GPADDR ) @@ -1450,7 +1500,7 @@ long do_mca(XEN_GUEST_HANDLE_PARAM(xen_mc_t) u_xen_mc) put_domain(d); } - if (!x86_mc_msrinject_verify(mc_msrinject)) + if ( !x86_mc_msrinject_verify(mc_msrinject) ) return x86_mcerr("do_mca inject: illegal MSR", -EINVAL); add_taint(TAINT_ERROR_INJECT); @@ -1461,16 +1511,16 @@ long do_mca(XEN_GUEST_HANDLE_PARAM(xen_mc_t) u_xen_mc) break; case XEN_MC_mceinject: - if (nr_mce_banks == 0) + if ( nr_mce_banks == 0 ) return x86_mcerr("do_mca #MC", -ENODEV); mc_mceinject = &op->u.mc_mceinject; target = mc_mceinject->mceinj_cpunr; - if (target >= nr_cpu_ids) + if ( target >= nr_cpu_ids ) return x86_mcerr("do_mca #MC: bad target", -EINVAL); - if (!cpu_online(target)) + if ( !cpu_online(target) ) return x86_mcerr("do_mca #MC: target offline", -EINVAL); add_taint(TAINT_ERROR_INJECT); @@ -1488,7 +1538,7 @@ long do_mca(XEN_GUEST_HANDLE_PARAM(xen_mc_t) u_xen_mc) cpumask_var_t cmv; bool broadcast = op->u.mc_inject_v2.flags & XEN_MC_INJECT_CPU_BROADCAST; - if (nr_mce_banks == 0) + if ( nr_mce_banks == 0 ) return x86_mcerr("do_mca #MC", -ENODEV); if ( broadcast ) @@ -1510,7 +1560,7 @@ long do_mca(XEN_GUEST_HANDLE_PARAM(xen_mc_t) u_xen_mc) "Not all required CPUs are online\n"); } - switch (op->u.mc_inject_v2.flags & XEN_MC_INJECT_TYPE_MASK) + switch ( op->u.mc_inject_v2.flags & XEN_MC_INJECT_TYPE_MASK ) { case XEN_MC_INJECT_TYPE_MCE: if ( mce_broadcast && @@ -1521,8 +1571,7 @@ long do_mca(XEN_GUEST_HANDLE_PARAM(xen_mc_t) u_xen_mc) case XEN_MC_INJECT_TYPE_CMCI: if ( !cmci_apic_vector ) - ret = x86_mcerr( - "No CMCI supported in platform\n", -EINVAL); + ret = x86_mcerr("No CMCI supported in platform\n", -EINVAL); else { if ( cpumask_test_cpu(smp_processor_id(), cpumap) ) @@ -1557,7 +1606,7 @@ long do_mca(XEN_GUEST_HANDLE_PARAM(xen_mc_t) u_xen_mc) break; } - if (cpumap != &cpu_online_map) + if ( cpumap != &cpu_online_map ) free_cpumask_var(cmv); break; @@ -1608,7 +1657,8 @@ void mc_panic(char *s) panic("HARDWARE ERROR"); } -/* Machine Check owner judge algorithm: +/* + * Machine Check owner judge algorithm: * When error happens, all cpus serially read its msr banks. * The first CPU who fetches the error bank's info will clear * this bank. Later readers can't get any information again. @@ -1625,7 +1675,7 @@ void mc_panic(char *s) /* Maybe called in MCE context, no lock, no printk */ static enum mce_result mce_action(const struct cpu_user_regs *regs, - mctelem_cookie_t mctc) + mctelem_cookie_t mctc) { struct mc_info *local_mi; enum mce_result bank_result = MCER_NOERROR; @@ -1636,15 +1686,16 @@ static enum mce_result mce_action(const struct cpu_user_regs *regs, unsigned int i, handler_num = mce_dhandler_num; /* When in mce context, regs is valid */ - if (regs) + if ( regs ) { handler_num = mce_uhandler_num; handlers = mce_uhandlers; } - local_mi = (struct mc_info*)mctelem_dataptr(mctc); + local_mi = (struct mc_info *)mctelem_dataptr(mctc); x86_mcinfo_lookup(mic, local_mi, MC_TYPE_GLOBAL); - if (mic == NULL) { + if ( mic == NULL ) + { printk(KERN_ERR "MCE: get local buffer entry failed\n "); return MCER_CONTINUE; } @@ -1659,17 +1710,19 @@ static enum mce_result mce_action(const struct cpu_user_regs *regs, for ( ; bank_result != MCER_RESET && mic && mic->size; mic = x86_mcinfo_next(mic) ) { - if (mic->type != MC_TYPE_BANK) { + if ( mic->type != MC_TYPE_BANK ) + { continue; } - binfo.mib = (struct mcinfo_bank*)mic; + binfo.mib = (struct mcinfo_bank *)mic; binfo.bank = binfo.mib->mc_bank; bank_result = MCER_NOERROR; - for ( i = 0; i < handler_num; i++ ) { - if (handlers[i].owned_error(binfo.mib->mc_status)) + for ( i = 0; i < handler_num; i++ ) + { + if ( handlers[i].owned_error(binfo.mib->mc_status) ) { handlers[i].recovery_handler(&binfo, &bank_result, regs); - if (worst_result < bank_result) + if ( worst_result < bank_result ) worst_result = bank_result; break; } @@ -1691,7 +1744,7 @@ static int mce_delayed_action(mctelem_cookie_t mctc) result = mce_action(NULL, mctc); - switch (result) + switch ( result ) { case MCER_RESET: dprintk(XENLOG_ERR, "MCE delayed action failed\n"); @@ -1702,12 +1755,12 @@ static int mce_delayed_action(mctelem_cookie_t mctc) case MCER_RECOVERED: dprintk(XENLOG_INFO, "MCE: Error is successfully recovered\n"); - ret = 1; + ret = 1; break; case MCER_CONTINUE: dprintk(XENLOG_INFO, "MCE: Error can't be recovered, " - "system is tainted\n"); + "system is tainted\n"); x86_mcinfo_dump(mctelem_dataptr(mctc)); ret = 1; break; @@ -1734,7 +1787,8 @@ static void mce_softirq(void) mce_barrier_enter(&mce_inside_bar, bcast); - if (!lmce) { + if ( !lmce ) + { /* * Everybody is here. Now let's see who gets to do the * recovery work. Right now we just see if there's a CPU @@ -1747,28 +1801,31 @@ static void mce_softirq(void) atomic_set(&severity_cpu, cpu); mce_barrier_enter(&mce_severity_bar, bcast); - if (!mctelem_has_deferred(cpu)) + if ( !mctelem_has_deferred(cpu) ) atomic_set(&severity_cpu, cpu); mce_barrier_exit(&mce_severity_bar, bcast); } /* We choose severity_cpu for further processing */ - if (lmce || atomic_read(&severity_cpu) == cpu) { + if ( lmce || atomic_read(&severity_cpu) == cpu ) + { mce_printk(MCE_VERBOSE, "CPU%d handling errors\n", cpu); - /* Step1: Fill DOM0 LOG buffer, vMCE injection buffer and + /* + * Step1: Fill DOM0 LOG buffer, vMCE injection buffer and * vMCE MSRs virtualization buffer */ - if (lmce) + if ( lmce ) mctelem_process_deferred(cpu, mce_delayed_action, true); else for_each_online_cpu(workcpu) mctelem_process_deferred(workcpu, mce_delayed_action, false); /* Step2: Send Log to DOM0 through vIRQ */ - if (dom0_vmce_enabled()) { + if ( dom0_vmce_enabled() ) + { mce_printk(MCE_VERBOSE, "MCE: send MCE# to DOM0 through virq\n"); send_global_virq(VIRQ_MCA); } @@ -1777,7 +1834,8 @@ static void mce_softirq(void) mce_barrier_exit(&mce_inside_bar, bcast); } -/* Machine Check owner judge algorithm: +/* + * Machine Check owner judge algorithm: * When error happens, all cpus serially read its msr banks. * The first CPU who fetches the error bank's info will clear * this bank. Later readers can't get any infor again. @@ -1793,7 +1851,7 @@ static void mce_softirq(void) */ void mce_handler_init(void) { - if (smp_processor_id() != 0) + if ( smp_processor_id() != 0 ) return; /* callback register, do we really need so many callback? */ diff --git a/xen/arch/x86/cpu/mcheck/mce.h b/xen/arch/x86/cpu/mcheck/mce.h index 664161a2af..f2cebccdd1 100644 --- a/xen/arch/x86/cpu/mcheck/mce.h +++ b/xen/arch/x86/cpu/mcheck/mce.h @@ -19,7 +19,8 @@ #define MCE_CRITICAL 2 extern int mce_verbosity; -/* Define the default level of machine check related print. +/* + * Define the default level of machine check related print. * When set mce_verbosity=verbose, all mce debug information * will be printed, otherwise, those information will not be * printed. @@ -62,8 +63,10 @@ void x86_mc_get_cpu_info(unsigned, uint32_t *, uint16_t *, uint16_t *, typedef void (*x86_mce_vector_t)(const struct cpu_user_regs *regs); extern void x86_mce_vector_register(x86_mce_vector_t); -/* Common generic MCE handler that implementations may nominate - * via x86_mce_vector_register. */ +/* + * Common generic MCE handler that implementations may nominate + * via x86_mce_vector_register. + */ extern void mcheck_cmn_handler(const struct cpu_user_regs *regs); /* Register a handler for judging whether mce is recoverable. */ @@ -90,12 +93,14 @@ static inline uint64_t mca_rdmsr(unsigned int msr) } while ( 0 ) -/* Utility function to "logout" all architectural MCA telemetry from the MCA +/* + * Utility function to "logout" all architectural MCA telemetry from the MCA * banks of the current processor. A cookie is returned which may be * uses to reference the data so logged (the cookie can be NULL if * no logout structures were available). The caller can also pass a pointer * to a structure which will be completed with some summary information - * of the MCA data observed in the logout operation. */ + * of the MCA data observed in the logout operation. + */ enum mca_source { MCA_POLLER, @@ -126,7 +131,8 @@ extern void mcheck_mca_clearbanks(struct mca_banks *); extern mctelem_cookie_t mcheck_mca_logout(enum mca_source, struct mca_banks *, struct mca_summary *, struct mca_banks *); -/* Register callbacks to be made during bank telemetry logout. +/* + * Register callbacks to be made during bank telemetry logout. * Those callbacks are only available to those machine check handlers * that call to the common mcheck_cmn_handler or who use the common * telemetry logout function mcheck_mca_logout in error polling. @@ -136,7 +142,8 @@ extern mctelem_cookie_t mcheck_mca_logout(enum mca_source, struct mca_banks *, typedef bool (*mce_need_clearbank_t)(enum mca_source who, u64 status); extern void mce_need_clearbank_register(mce_need_clearbank_t); -/* Register a callback to collect additional information (typically non- +/* + * Register a callback to collect additional information (typically non- * architectural) provided by newer CPU families/models without the need * to duplicate the whole handler resulting in various handlers each with * its own tweaks and bugs. The callback receives an struct mc_info pointer -- 2.11.0 _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx https://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |