On Wed, 7 Jan 2015, Linus Torvalds wrote: > On Wed, Jan 7, 2015 at 11:00 AM, Nicolas Pitre <nicolas.pi...@linaro.org> > wrote: > > > > We'll make sure it is scaled properly so not to have orders of magnitude > > discrepancy whether the timer based or the CPU based loop is used for > > the purpose of making people feel good. > > Why? > > You'd basically be lying. And it might actually hide real problems. > If the scaling hides the fact that the timer source cannot do a good > job at microsecond resolution delays, then it's not just lying, it's > lying in ways that hide real issues. So why should that kind of > behavior be encouraged? The actual *real* unscaled resolution of the > timer is valid and real information.
I think you are missing something fundamental in this thread. On ARM, when the timer is used to provide small delays, it is typically the ARM architected timer which by definition must have a constant input clock in the MHz range. This timer clock has *nothing* to do with whatever CPU clock you might be using. On the system I have here, the CPU clock is 2GHz and the timer used for delays is 24MHz. If the CPU clock is scaled down to 180MHz the timer clock remains at 24MHz. The implementation of udelay() in this case is basically doing: void udelay(unsigned long usecs) { unsigned long timer_ticks = usecs * (24000000 / 1000000); unsigned long start = read_timer_count(); while (read_timer_count() - start < timer_ticks); } Some other systems might as well have a completely different timer clock based on what its hardware designers thought was best, or based on what they smoked the night before. There is no calibrating of the timer input clock: it is just set and the timer clock rate for a given hardware implementation is advertised by the firmware. No calibration is necessary. No calibration would even be possible if that's the only time source on the system. Now tell me what this means for bogomips? Nothing. Absolutely nothing. Deriving a bogomips number from a 24MHz timer clock that bares no relationship with the CPU clock is completely useless. We might as well hardcode a constant 1.00 and be done with it. Or hash the machine name and add the RTC time and report that. At least the later would have some entertainment value. What I'm suggesting is to leave the timer based udelay() alone as it doesn't need any loops_per_jiffy or whatnot to operate. Then, for the semi-entertaining value of having *something* displayed alongside "bogomips" in /proc/cpuinfo I'm suggesting to simply calibrate loops_per_jiffy the traditional way in all cases, whether or not a timer based udelay is in use. Why you might have having a problem with that is beyond my understanding. Nicolas -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to majord...@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/