Hi Quentin, On 05/06/18 11:57, Quentin Perret wrote:
[...] > What about the diff below (just a quick hack to show the idea) applied > on tip/sched/core ? > > ---8<--- > diff --git a/kernel/sched/cpufreq_schedutil.c > b/kernel/sched/cpufreq_schedutil.c > index a8ba6d1f262a..23a4fb1c2c25 100644 > --- a/kernel/sched/cpufreq_schedutil.c > +++ b/kernel/sched/cpufreq_schedutil.c > @@ -180,9 +180,12 @@ static void sugov_get_util(struct sugov_cpu *sg_cpu) > sg_cpu->util_dl = cpu_util_dl(rq); > } > > +unsigned long scale_rt_capacity(int cpu); > static unsigned long sugov_aggregate_util(struct sugov_cpu *sg_cpu) > { > struct rq *rq = cpu_rq(sg_cpu->cpu); > + int cpu = sg_cpu->cpu; > + unsigned long util, dl_bw; > > if (rq->rt.rt_nr_running) > return sg_cpu->max; > @@ -197,7 +200,14 @@ static unsigned long sugov_aggregate_util(struct > sugov_cpu *sg_cpu) > * util_cfs + util_dl as requested freq. However, cpufreq is not yet > * ready for such an interface. So, we only do the latter for now. > */ > - return min(sg_cpu->max, (sg_cpu->util_dl + sg_cpu->util_cfs)); > + util = arch_scale_cpu_capacity(NULL, cpu) * scale_rt_capacity(cpu); Sorry to be pedantinc, but this (ATM) includes DL avg contribution, so, since we use max below, we will probably have the same problem that we discussed on Vincent's approach (overestimation of DL contribution while we could use running_bw). > + util >>= SCHED_CAPACITY_SHIFT; > + util = arch_scale_cpu_capacity(NULL, cpu) - util; > + util += sg_cpu->util_cfs; > + dl_bw = (rq->dl.this_bw * SCHED_CAPACITY_SCALE) >> BW_SHIFT; Why this_bw instead of running_bw? Thanks, - Juri