Hey,

Some test results:

On Thursday 18 Feb 2021 at 16:35:38 (+0000), Ionela Voinescu wrote:
[..]
> > +static void __init cppc_freq_invariance_init(void)
> > +{
[..]
> > +
> > +           ret = cppc_get_perf_ctrs(i, &fb_ctrs);
> > +           if (!ret)
> > +                   per_cpu(cppc_fi->prev_perf_fb_ctrs, i) = fb_ctrs;
> 

After fixing this one:
                        cppc_fi->prev_perf_fb_ctrs = fb_ctrs;

I got the following:

Platform:

 - Juno R2 (CPUs [0-3] are littles, CPUs [4-5] are bigs)
    + PMU counters, used by CPPC through FFH
    + userspace/schedutil


  - Verifying that with userspace governor we see a correct change in
    scale factor:

        root@buildroot:~# dmesg | grep FIE
        [    6.436770] AMU: CPUs[0-3]: AMU counters WON'T be used for FIE.
        [    6.436962] AMU: CPUs[4-5]: AMU counters WON'T be used for FIE.
        [    6.451510] CPPC:CPUs[0-5]: CPPC counters will be used for FIE.

        root@buildroot:~# echo 600000 > policy4/scaling_setspeed
        [  353.939495] CPU4: Invariance(cppc) scale: 512.
        [  353.939497] CPU5: Invariance(cppc) scale: 512.

        root@buildroot:~# echo 1200000 > policy4/scaling_setspeed
        [  372.683511] CPU5: Invariance(cppc) scale: 1024.
        [  372.683518] CPU4: Invariance(cppc) scale: 1024.

        root@buildroot:~# echo 450000 > policy0/scaling_setspeed
        [  641.495513] CPU2: Invariance(cppc) scale: 485.
        [  641.495514] CPU1: Invariance(cppc) scale: 485.
        [  641.495517] CPU0: Invariance(cppc) scale: 485.
        [  641.495542] CPU3: Invariance(cppc) scale: 485.

        root@buildroot:~# echo 950000 > policy0/scaling_setspeed
        [  852.015514] CPU2: Invariance(cppc) scale: 1024.
        [  852.015514] CPU1: Invariance(cppc) scale: 1024.
        [  852.015517] CPU0: Invariance(cppc) scale: 1024.
        [  852.015541] CPU3: Invariance(cppc) scale: 1024.

 - I ran some benchmarks as well (perf, hackbench, dhrystone) on the same
   platform, using the userspace governor at fixed frequency, to evaluate
   the impact of the work we do or don't do on the tick.

   ./perf bench sched pipe
   (10 iterations, higher is better, ops/s, comparisons with
   cpufreq-based FIE)

   cpufreq-based FIE    AMU-based FIE    CPPC-based FIE
   ----------------------------------------------------
   39498.8              40984.7          38893.4
   std: 3.766%          std: 4.461%      std: 0.575%
                        diff: 3.625%     diff: -1.556%

   ./hackbench -l 1000
   (10 iterations, lower is better, seconds, comparison with
   cpufreq-based FIE)

   cpufreq-based FIE    AMU-based FIE    CPPC-based FIE
   ----------------------------------------------------
   6.4207               6.3386           6.7841
   std: 7.298%          std: 2.252%      std: 2.460%
                        diff: -1.295%    diff: 5.356%

   This shows a small regression for the CPPC-based FIE, but within the
   standard deviation.

   I ran some dhrystone benchmarks (./dhrystone -t 2/34/5/6/ -l 5000) as
   well with schedutil governor to understand if an increase in accuracy
   with the AMU/CPPC counters makes a difference. Given the
   characteristics of the platform it's no surprise that the results
   were very similar between the three cases, so I won't bore you with
   the numbers.

Hope it helps,
Ionela.

Reply via email to