\
> diff --git a/arch/powerpc/kernel/perf_event.c 
> b/arch/powerpc/kernel/perf_event.c
> index 97e0ae4..0a5178f 100644
> --- a/arch/powerpc/kernel/perf_event.c
> +++ b/arch/powerpc/kernel/perf_event.c
> @@ -416,6 +416,15 @@ static void power_pmu_read(struct perf_event *event)
>               prev = local64_read(&event->hw.prev_count);
>               barrier();
>               val = read_pmc(event->hw.idx);
> +             /*
> +              * POWER7 can roll back counter values, if the new value is
> +              * smaller than the previous value it will cause the delta
> +              * and the counter to have bogus values.  If this is the
> +              * case skip updating anything until the counter grows again.
> +              * This can lead to a small lack of precision in the counters.
> +              */
> +             if (val < prev)
> +                     return;
>       } while (local64_cmpxchg(&event->hw.prev_count, prev, val) != prev);

So if the counters can rollover, the above is still wrong... prev could
be ffffffff and val could be 1 ... for example. In this case you really
need to substract and get the sign of the result I'd think...

Cheers,
Ben.

>       /* The counters are only 32 bits wide */
> @@ -439,7 +448,8 @@ static void freeze_limited_counters(struct cpu_hw_events 
> *cpuhw,
>                                   unsigned long pmc5, unsigned long pmc6)
>  {
>       struct perf_event *event;
> -     u64 val, prev, delta;
> +     u64 val, prev;
> +     s32 delta;
>       int i;
>  
>       for (i = 0; i < cpuhw->n_limited; ++i) {
> @@ -449,8 +459,13 @@ static void freeze_limited_counters(struct cpu_hw_events 
> *cpuhw,
>               val = (event->hw.idx == 5) ? pmc5 : pmc6;
>               prev = local64_read(&event->hw.prev_count);
>               event->hw.idx = 0;
> -             delta = (val - prev) & 0xfffffffful;
> -             local64_add(delta, &event->count);
> +             /*
> +              * The PerfMon registers are only 32 bits wide so the
> +              * delta should not overflow.
> +              */
> +             delta = val - prev;
> +             if (delta > 0)
> +                     local64_add(delta, &event->count);
>       }
>  }
>  
> @@ -458,14 +473,16 @@ static void thaw_limited_counters(struct cpu_hw_events 
> *cpuhw,
>                                 unsigned long pmc5, unsigned long pmc6)
>  {
>       struct perf_event *event;
> -     u64 val;
> +     u64 val, prev;
>       int i;
>  
>       for (i = 0; i < cpuhw->n_limited; ++i) {
>               event = cpuhw->limited_counter[i];
>               event->hw.idx = cpuhw->limited_hwidx[i];
>               val = (event->hw.idx == 5) ? pmc5 : pmc6;
> -             local64_set(&event->hw.prev_count, val);
> +             prev = local64_read(&event->hw.prev_count);
> +             if (val > prev)
> +                     local64_set(&event->hw.prev_count, val);
>               perf_event_update_userpage(event);
>       }
>  }
> @@ -1187,7 +1204,8 @@ static void record_and_restart(struct perf_event 
> *event, unsigned long val,
>                              struct pt_regs *regs, int nmi)
>  {
>       u64 period = event->hw.sample_period;
> -     s64 prev, delta, left;
> +     s64 prev, left;
> +     s32 delta;
>       int record = 0;
>  
>       if (event->hw.state & PERF_HES_STOPPED) {
> @@ -1197,7 +1215,9 @@ static void record_and_restart(struct perf_event 
> *event, unsigned long val,
>  
>       /* we don't have to worry about interrupts here */
>       prev = local64_read(&event->hw.prev_count);
> -     delta = (val - prev) & 0xfffffffful;
> +     delta = val - prev;
> +     if (delta < 0)
> +             delta = 0;
>       local64_add(delta, &event->count);
>  
>       /*


_______________________________________________
Linuxppc-dev mailing list
Linuxppc-dev@lists.ozlabs.org
https://lists.ozlabs.org/listinfo/linuxppc-dev

Reply via email to