On Fri, Sep 25, 2015 at 10:29:01AM +0200, Borislav Petkov wrote: > > > > > > > The last patch of that series had 2 changes. > > > > 1. Allow offline cpu's to participate in the rendezvous. Since in the odd > > chance the offline cpus have any errors collected we can still report them. > > (we changed mce_start/mce_end to use cpu_present_mask instead of just > > online map). > > This is not necessarily wrong - it is just unusual.
Correct!. > > > Without this change today if i were to inject an broadcast MCE > > it ends up hanging, since the offline cpu is also incrementing mce_callin. > > It will always end up more than cpu_online_mask by the number of cpu's > > logically offlined > > Yeah, I'd like to have a bit somewhere which says "don't report MCEs on this > core." But we talked about this already. > > > Consider for e.g. if 2 thread of the core are offline. And the MLC picks up > > What is MLC? Mid Level Cache. This is shared between the 2 threads in that core. As opposed to the Last Level Cache (LLC) which is shared between all the threads in the socket. > > > an error. Other cpus in the socket can't access them. Only way is to let > > those > > CPUs read and report their own banks as they are core scoped. In upcoming > > CPUs > > we have some banks that can be thread scoped as well. > > > > Its understood OS doesn't execute any code on those CPUs. But SMI can still > > run on them, and could collect errors that can be logged. > > Well, that is not our problem, is it? > > I mean, SMM wants to stay undetected. When all of a sudden offlined > cores start reporting MCEs, that's going to raise some brows. You are right.. i was simply trying to state how an offline CPU from the OS perspective could still be collecting errors. Only trying to highlight what happens from a platform level. > > Regardless, there are other reasons why offlined cores might report MCEs > - the fact that logical cores share functional units and data flow goes > through them might trip the reporting on those cores. Yadda yadda... Yep! > > > 2. If the cpu is offline, we copied them to mce_log buffer, and them copy > > those out from the rendezvous master during mce_reign(). > > > > If we were to replace this mce_log_add() with gen_pool_add(), then i would > > have to call mce_gen_pool_add() from the offline CPU. This will end up > > calling > > RCU functions. > > > > We don't want to leave any errors reported by the offline CPU for purpose > > of logging. It is rare, but still interested in capturing those errors if > > they > > were to happen. > > > > Does this help? > > So first of all, we need to hold this down somewhere, maybe in > Documentation/ to explain why we're running on offlined cores. This is > certainly unusual code and people will ask WTF is going on there. Good idea to document these weird cases. I can do it in either the cpu-hotplug.txt, or in the x86/x86_64/machinecheck file as appropriate. Will add that in my next update. > > Then, I really really don't like a static buffer which we will have > to increase with each new bigger machine configuration. This is just > clumsy. > > It'd be probably much better to make that MCE buffer per CPU. We can > say, we're allowed to log 2-3, hell, 5 errors in it and when we're done > with the rendezvous, an online core goes and flushes out the error > records to gen_pool. That makes good sense.. i will make these per-cpu and also look at clearing them during mce_panic to make sure we flush them in fatal cases. per-cpu certainly scales better than a static number.. > > This scales much better than any artificial MCE_LOG_LEN size. > > Oh, and we either overwrite old errors when we fill up the percpu buffer > or we return. that's something we can discuss later. Or we come up with > a bit smarter strategy of selecting which ones to overwrite. > > Just artificially increasing a static buffer is not good design IMO. Agreed.. thanks, i will get another rev rolling. Cheers, Ashok -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/