On Fri, Jan 08, 2021 at 01:31:56PM +0100, Borislav Petkov wrote: > On Thu, Jan 07, 2021 at 09:08:44AM -0800, Paul E. McKenney wrote: > > Some information is usually better than none. And I bet that failing > > hardware is capable of all sorts of tricks at all sorts of levels. ;-) > > Tell me about it. > > > Updated patch below. Is this what you had in mind? > > Ok, so I've massaged it into the below locally while taking another > detailed look. Made the pr_info pr_emerg and poked at the text more, as > I do. :) > > Lemme know if something else needs to be adjusted, otherwise I'll queue > it.
Looks good to me! I agree that your change to the pr_emerg() string is much better than my original. And good point on your added comment, plus it was fun to see that my original "holdouts" wording has not completely vanished. ;-) Thank you very much!!! Thanx, Paul > Thx. > > --- > Author: Paul E. McKenney <paul...@kernel.org> > Date: Wed Dec 23 17:04:19 2020 -0800 > > x86/mce: Make mce_timed_out() identify holdout CPUs > > The > > "Timeout: Not all CPUs entered broadcast exception handler" > > message will appear from time to time given enough systems, but this > message does not identify which CPUs failed to enter the broadcast > exception handler. This information would be valuable if available, > for example, in order to correlate with other hardware-oriented error > messages. > > Add a cpumask of CPUs which maintains which CPUs have entered this > handler, and print out which ones failed to enter in the event of a > timeout. > > [ bp: Massage. ] > > Reported-by: Jonathan Lemon <b...@fb.com> > Signed-off-by: Paul E. McKenney <paul...@kernel.org> > Signed-off-by: Borislav Petkov <b...@suse.de> > Tested-by: Tony Luck <tony.l...@intel.com> > Link: > https://lkml.kernel.org/r/20210106174102.GA23874@paulmck-ThinkPad-P72 > > diff --git a/arch/x86/kernel/cpu/mce/core.c b/arch/x86/kernel/cpu/mce/core.c > index 13d3f1cbda17..6c81d0998e0a 100644 > --- a/arch/x86/kernel/cpu/mce/core.c > +++ b/arch/x86/kernel/cpu/mce/core.c > @@ -877,6 +877,12 @@ static atomic_t mce_executing; > */ > static atomic_t mce_callin; > > +/* > + * Track which CPUs entered the MCA broadcast synchronization and which not > in > + * order to print holdouts. > + */ > +static cpumask_t mce_missing_cpus = CPU_MASK_ALL; > + > /* > * Check if a timeout waiting for other CPUs happened. > */ > @@ -894,8 +900,12 @@ static int mce_timed_out(u64 *t, const char *msg) > if (!mca_cfg.monarch_timeout) > goto out; > if ((s64)*t < SPINUNIT) { > - if (mca_cfg.tolerant <= 1) > + if (mca_cfg.tolerant <= 1) { > + if (cpumask_and(&mce_missing_cpus, cpu_online_mask, > &mce_missing_cpus)) > + pr_emerg("CPUs not responding to MCE broadcast > (may include false positives): %*pbl\n", > + cpumask_pr_args(&mce_missing_cpus)); > mce_panic(msg, NULL, NULL); > + } > cpu_missing = 1; > return 1; > } > @@ -1006,6 +1016,7 @@ static int mce_start(int *no_way_out) > * is updated before mce_callin. > */ > order = atomic_inc_return(&mce_callin); > + cpumask_clear_cpu(smp_processor_id(), &mce_missing_cpus); > > /* > * Wait for everyone. > @@ -1114,6 +1125,7 @@ static int mce_end(int order) > reset: > atomic_set(&global_nwo, 0); > atomic_set(&mce_callin, 0); > + cpumask_setall(&mce_missing_cpus); > barrier(); > > /* > @@ -2712,6 +2724,7 @@ static void mce_reset(void) > atomic_set(&mce_executing, 0); > atomic_set(&mce_callin, 0); > atomic_set(&global_nwo, 0); > + cpumask_setall(&mce_missing_cpus); > } > > static int fake_panic_get(void *data, u64 *val) > > -- > Regards/Gruss, > Boris. > > https://people.kernel.org/tglx/notes-about-netiquette