Hi Dirk, FYI, we noticed the below changes on commit adacdf3f2b8e65aa441613cf61c4f598e9042690 ("intel_pstate: Remove C0 tracking")
test case: brickland3/aim9/300s-shell_rtns_3 v3.15-rc8 adacdf3f2b8e65aa441613cf6 --------------- ------------------------- 125 ± 5% +119.9% 275 ± 1% TOTAL aim9.shell_rtns_3.ops_per_sec 96.81 ± 3% +51.6% 146.77 ± 5% TOTAL turbostat.Pkg_W 36.74 ± 9% +121.4% 81.34 ±10% TOTAL turbostat.Cor_W 38.36 ± 1% +63.6% 62.76 ± 0% TOTAL turbostat.RAM_W -13794 ±-5% -13.4% -11946 ±-6% TOTAL sched_debug.cfs_rq[1]:/.spread0 -10828 ±-8% -34.7% -7069 ±-20% TOTAL sched_debug.cfs_rq[34]:/.spread0 -14141 ±-8% -19.1% -11441 ±-19% TOTAL sched_debug.cfs_rq[24]:/.spread0 -13819 ±-7% -13.6% -11944 ±-6% TOTAL sched_debug.cfs_rq[7]:/.spread0 6006 ± 8% -71.6% 1703 ±21% TOTAL sched_debug.cpu#33.ttwu_local 6281 ±36% -68.3% 1988 ±49% TOTAL sched_debug.cpu#7.ttwu_count 3177 ±47% +235.5% 10660 ± 6% TOTAL cpuidle.C1-IVT-4S.usage 268 ±45% -68.5% 84 ±31% TOTAL sched_debug.cpu#45.ttwu_local 6658 ±34% -66.1% 2260 ±30% TOTAL sched_debug.cpu#21.ttwu_count 292 ±44% -71.7% 82 ±11% TOTAL sched_debug.cpu#23.ttwu_local 5351 ± 6% -61.8% 2045 ±29% TOTAL sched_debug.cpu#48.ttwu_local 2395 ±29% -62.9% 888 ±14% TOTAL sched_debug.cpu#37.ttwu_count 2269 ±11% +144.0% 5537 ±26% TOTAL sched_debug.cfs_rq[91]:/.blocked_load_avg 2040 ±17% +154.5% 5192 ±14% TOTAL sched_debug.cfs_rq[106]:/.blocked_load_avg 1.24 ± 6% +154.8% 3.15 ± 0% TOTAL turbostat.GHz 2417 ±10% +135.2% 5685 ±25% TOTAL sched_debug.cfs_rq[91]:/.tg_load_contrib 69 ±29% -52.4% 33 ±36% TOTAL sched_debug.cfs_rq[41]:/.avg->runnable_avg_sum 4422 ±29% -57.6% 1875 ± 9% TOTAL sched_debug.cpu#3.ttwu_local 2210 ±14% +140.9% 5324 ±14% TOTAL sched_debug.cfs_rq[106]:/.tg_load_contrib 1445 ±16% +126.6% 3276 ±14% TOTAL sched_debug.cfs_rq[3]:/.blocked_load_avg 1448 ±16% +127.4% 3293 ±14% TOTAL sched_debug.cfs_rq[3]:/.tg_load_contrib 561248 ± 4% +145.3% 1376953 ± 0% TOTAL cpuidle.C6-IVT-4S.usage 4975 ±28% -63.7% 1805 ±13% TOTAL sched_debug.cpu#4.ttwu_local 1348 ±19% +137.8% 3206 ±12% TOTAL sched_debug.cfs_rq[48]:/.blocked_load_avg 1696 ±13% +106.7% 3507 ±15% TOTAL sched_debug.cfs_rq[32]:/.tg_load_contrib 1684 ±13% +106.6% 3478 ±15% TOTAL sched_debug.cfs_rq[32]:/.blocked_load_avg 1619 ±22% +118.1% 3532 ±13% TOTAL sched_debug.cfs_rq[17]:/.blocked_load_avg 1626 ±22% +117.4% 3537 ±13% TOTAL sched_debug.cfs_rq[17]:/.tg_load_contrib 1354 ±19% +137.0% 3209 ±12% TOTAL sched_debug.cfs_rq[48]:/.tg_load_contrib 21314 ± 5% +125.6% 48083 ± 2% TOTAL sched_debug.cfs_rq[85]:/.tg_load_avg 21409 ± 5% +125.1% 48199 ± 1% TOTAL sched_debug.cfs_rq[83]:/.tg_load_avg 21340 ± 5% +125.8% 48193 ± 1% TOTAL sched_debug.cfs_rq[84]:/.tg_load_avg 21291 ± 5% +125.7% 48060 ± 2% TOTAL sched_debug.cfs_rq[86]:/.tg_load_avg 21191 ± 6% +126.2% 47929 ± 1% TOTAL sched_debug.cfs_rq[102]:/.tg_load_avg 21266 ± 5% +126.0% 48058 ± 2% TOTAL sched_debug.cfs_rq[90]:/.tg_load_avg 21289 ± 5% +125.7% 48054 ± 2% TOTAL sched_debug.cfs_rq[89]:/.tg_load_avg 21186 ± 6% +126.2% 47929 ± 1% TOTAL sched_debug.cfs_rq[101]:/.tg_load_avg 21314 ± 6% +125.8% 48131 ± 1% TOTAL sched_debug.cfs_rq[113]:/.tg_load_avg 21266 ± 5% +126.1% 48083 ± 2% TOTAL sched_debug.cfs_rq[91]:/.tg_load_avg 21298 ± 6% +125.3% 47981 ± 1% TOTAL sched_debug.cfs_rq[106]:/.tg_load_avg 21309 ± 6% +125.0% 47953 ± 1% TOTAL sched_debug.cfs_rq[105]:/.tg_load_avg 21178 ± 6% +126.4% 47953 ± 1% TOTAL sched_debug.cfs_rq[100]:/.tg_load_avg 21236 ± 6% +126.5% 48095 ± 1% TOTAL sched_debug.cfs_rq[93]:/.tg_load_avg 21271 ± 5% +126.0% 48077 ± 1% TOTAL sched_debug.cfs_rq[88]:/.tg_load_avg 21286 ± 5% +126.0% 48096 ± 2% TOTAL sched_debug.cfs_rq[87]:/.tg_load_avg 21269 ± 6% +126.1% 48093 ± 1% TOTAL sched_debug.cfs_rq[92]:/.tg_load_avg 21291 ± 6% +125.2% 47956 ± 1% TOTAL sched_debug.cfs_rq[104]:/.tg_load_avg 21303 ± 6% +125.3% 48005 ± 1% TOTAL sched_debug.cfs_rq[107]:/.tg_load_avg 21247 ± 6% +125.7% 47957 ± 1% TOTAL sched_debug.cfs_rq[94]:/.tg_load_avg 21350 ± 6% +125.7% 48185 ± 1% TOTAL sched_debug.cfs_rq[119]:/.tg_load_avg 21357 ± 6% +125.3% 48108 ± 1% TOTAL sched_debug.cfs_rq[114]:/.tg_load_avg 21263 ± 6% +125.6% 47968 ± 1% TOTAL sched_debug.cfs_rq[103]:/.tg_load_avg 21362 ± 6% +125.4% 48154 ± 1% TOTAL sched_debug.cfs_rq[118]:/.tg_load_avg 21470 ± 5% +124.6% 48223 ± 1% TOTAL sched_debug.cfs_rq[82]:/.tg_load_avg 1513 ±24% +122.0% 3358 ±16% TOTAL sched_debug.cfs_rq[47]:/.blocked_load_avg 21170 ± 6% +126.4% 47936 ± 1% TOTAL sched_debug.cfs_rq[98]:/.tg_load_avg 21216 ± 6% +126.0% 47943 ± 1% TOTAL sched_debug.cfs_rq[95]:/.tg_load_avg 21183 ± 6% +126.3% 47936 ± 1% TOTAL sched_debug.cfs_rq[97]:/.tg_load_avg 21351 ± 6% +125.3% 48106 ± 1% TOTAL sched_debug.cfs_rq[115]:/.tg_load_avg 21194 ± 6% +126.2% 47952 ± 1% TOTAL sched_debug.cfs_rq[96]:/.tg_load_avg 21181 ± 6% +126.3% 47929 ± 1% TOTAL sched_debug.cfs_rq[99]:/.tg_load_avg 21366 ± 6% +125.1% 48101 ± 1% TOTAL sched_debug.cfs_rq[116]:/.tg_load_avg 21352 ± 6% +125.5% 48145 ± 1% TOTAL sched_debug.cfs_rq[112]:/.tg_load_avg 21381 ± 6% +125.1% 48131 ± 1% TOTAL sched_debug.cfs_rq[117]:/.tg_load_avg 21507 ± 5% +124.3% 48244 ± 1% TOTAL sched_debug.cfs_rq[81]:/.tg_load_avg 21346 ± 6% +125.5% 48126 ± 1% TOTAL sched_debug.cfs_rq[111]:/.tg_load_avg 22339 ± 4% +124.5% 50156 ± 1% TOTAL sched_debug.cfs_rq[5]:/.tg_load_avg 21569 ± 5% +123.7% 48256 ± 1% TOTAL sched_debug.cfs_rq[80]:/.tg_load_avg 21343 ± 6% +125.0% 48018 ± 1% TOTAL sched_debug.cfs_rq[108]:/.tg_load_avg 1528 ±23% +120.8% 3373 ±16% TOTAL sched_debug.cfs_rq[47]:/.tg_load_contrib 21616 ± 5% +123.2% 48245 ± 2% TOTAL sched_debug.cfs_rq[78]:/.tg_load_avg 21595 ± 4% +124.7% 48525 ± 1% TOTAL sched_debug.cfs_rq[41]:/.tg_load_avg 21622 ± 5% +123.4% 48294 ± 2% TOTAL sched_debug.cfs_rq[77]:/.tg_load_avg 21571 ± 4% +123.7% 48245 ± 2% TOTAL sched_debug.cfs_rq[79]:/.tg_load_avg 22460 ± 4% +124.3% 50377 ± 1% TOTAL sched_debug.cfs_rq[4]:/.tg_load_avg 22257 ± 5% +124.2% 49910 ± 1% TOTAL sched_debug.cfs_rq[8]:/.tg_load_avg 22291 ± 5% +124.0% 49922 ± 1% TOTAL sched_debug.cfs_rq[7]:/.tg_load_avg 22586 ± 4% +123.3% 50430 ± 1% TOTAL sched_debug.cfs_rq[3]:/.tg_load_avg 22236 ± 5% +124.1% 49831 ± 1% TOTAL sched_debug.cfs_rq[9]:/.tg_load_avg 21599 ± 4% +124.4% 48473 ± 1% TOTAL sched_debug.cfs_rq[42]:/.tg_load_avg 22118 ± 6% +123.8% 49501 ± 1% TOTAL sched_debug.cfs_rq[12]:/.tg_load_avg 21591 ± 4% +124.8% 48544 ± 1% TOTAL sched_debug.cfs_rq[40]:/.tg_load_avg 21348 ± 6% +125.3% 48090 ± 1% TOTAL sched_debug.cfs_rq[110]:/.tg_load_avg 21636 ± 4% +123.4% 48331 ± 1% TOTAL sched_debug.cfs_rq[43]:/.tg_load_avg 22170 ± 5% +123.5% 49543 ± 1% TOTAL sched_debug.cfs_rq[11]:/.tg_load_avg 22117 ± 6% +123.8% 49505 ± 1% TOTAL sched_debug.cfs_rq[13]:/.tg_load_avg 21656 ± 5% +122.8% 48260 ± 1% TOTAL sched_debug.cfs_rq[44]:/.tg_load_avg 22206 ± 5% +123.4% 49613 ± 1% TOTAL sched_debug.cfs_rq[10]:/.tg_load_avg 22307 ± 4% +124.3% 50042 ± 1% TOTAL sched_debug.cfs_rq[6]:/.tg_load_avg 1438 ±22% +123.8% 3218 ±13% TOTAL sched_debug.cfs_rq[18]:/.blocked_load_avg 21389 ± 6% +124.7% 48064 ± 1% TOTAL sched_debug.cfs_rq[109]:/.tg_load_avg 1438 ±22% +124.0% 3222 ±13% TOTAL sched_debug.cfs_rq[18]:/.tg_load_contrib 21612 ± 4% +124.6% 48541 ± 1% TOTAL sched_debug.cfs_rq[39]:/.tg_load_avg 103 ±15% -56.4% 45 ±32% TOTAL sched_debug.cfs_rq[30]:/.avg->runnable_avg_sum 22746 ± 4% +122.6% 50626 ± 0% TOTAL sched_debug.cfs_rq[2]:/.tg_load_avg 22689 ± 4% +122.6% 50512 ± 0% TOTAL sched_debug.cfs_rq[1]:/.tg_load_avg 1498 ±12% +118.2% 3271 ±16% TOTAL sched_debug.cfs_rq[33]:/.tg_load_contrib 21710 ± 5% +122.1% 48223 ± 1% TOTAL sched_debug.cfs_rq[45]:/.tg_load_avg 21675 ± 5% +123.9% 48522 ± 1% TOTAL sched_debug.cfs_rq[68]:/.tg_load_avg 22702 ± 4% +123.0% 50638 ± 1% TOTAL sched_debug.cfs_rq[0]:/.tg_load_avg 21791 ± 5% +120.8% 48114 ± 1% TOTAL sched_debug.cfs_rq[47]:/.tg_load_avg 21768 ± 5% +122.1% 48352 ± 1% TOTAL sched_debug.cfs_rq[57]:/.tg_load_avg 21611 ± 4% +123.9% 48393 ± 1% TOTAL sched_debug.cfs_rq[72]:/.tg_load_avg 21668 ± 4% +124.2% 48578 ± 1% TOTAL sched_debug.cfs_rq[38]:/.tg_load_avg 21661 ± 5% +123.5% 48416 ± 1% TOTAL sched_debug.cfs_rq[71]:/.tg_load_avg 21653 ± 5% +123.0% 48291 ± 2% TOTAL sched_debug.cfs_rq[76]:/.tg_load_avg 21748 ± 5% +122.7% 48426 ± 1% TOTAL sched_debug.cfs_rq[62]:/.tg_load_avg 21770 ± 5% +121.2% 48162 ± 1% TOTAL sched_debug.cfs_rq[46]:/.tg_load_avg 21688 ± 5% +122.7% 48295 ± 2% TOTAL sched_debug.cfs_rq[75]:/.tg_load_avg 21651 ± 5% +123.5% 48392 ± 1% TOTAL sched_debug.cfs_rq[74]:/.tg_load_avg 101524 ± 6% +122.3% 225684 ± 4% TOTAL proc-vmstat.pgalloc_dma32 21758 ± 5% +122.5% 48417 ± 1% TOTAL sched_debug.cfs_rq[63]:/.tg_load_avg 21721 ± 5% +123.2% 48488 ± 1% TOTAL sched_debug.cfs_rq[67]:/.tg_load_avg 2057 ±10% +137.4% 4885 ±18% TOTAL sched_debug.cfs_rq[61]:/.blocked_load_avg 21704 ± 5% +123.2% 48439 ± 1% TOTAL sched_debug.cfs_rq[64]:/.tg_load_avg 21695 ± 5% +123.2% 48422 ± 1% TOTAL sched_debug.cfs_rq[70]:/.tg_load_avg 21706 ± 5% +123.1% 48428 ± 1% TOTAL sched_debug.cfs_rq[69]:/.tg_load_avg 21837 ± 5% +121.2% 48294 ± 1% TOTAL sched_debug.cfs_rq[56]:/.tg_load_avg 21761 ± 4% +122.3% 48370 ± 1% TOTAL sched_debug.cfs_rq[61]:/.tg_load_avg 21769 ± 4% +123.2% 48581 ± 1% TOTAL sched_debug.cfs_rq[37]:/.tg_load_avg 21704 ± 5% +123.4% 48479 ± 1% TOTAL sched_debug.cfs_rq[66]:/.tg_load_avg 21643 ± 5% +123.5% 48365 ± 1% TOTAL sched_debug.cfs_rq[73]:/.tg_load_avg 21693 ± 5% +123.5% 48480 ± 1% TOTAL sched_debug.cfs_rq[65]:/.tg_load_avg 1498 ±12% +117.5% 3260 ±16% TOTAL sched_debug.cfs_rq[33]:/.blocked_load_avg 21762 ± 4% +122.0% 48313 ± 1% TOTAL sched_debug.cfs_rq[58]:/.tg_load_avg 21873 ± 5% +120.1% 48132 ± 1% TOTAL sched_debug.cfs_rq[48]:/.tg_load_avg 21778 ± 4% +123.3% 48623 ± 1% TOTAL sched_debug.cfs_rq[36]:/.tg_load_avg 21887 ± 3% +122.3% 48647 ± 1% TOTAL sched_debug.cfs_rq[35]:/.tg_load_avg 22106 ± 6% +123.1% 49330 ± 1% TOTAL sched_debug.cfs_rq[14]:/.tg_load_avg 21743 ± 5% +122.5% 48382 ± 1% TOTAL sched_debug.cfs_rq[60]:/.tg_load_avg 21761 ± 5% +122.2% 48362 ± 1% TOTAL sched_debug.cfs_rq[59]:/.tg_load_avg 21933 ± 3% +121.8% 48653 ± 1% TOTAL sched_debug.cfs_rq[34]:/.tg_load_avg 22003 ± 3% +121.9% 48833 ± 1% TOTAL sched_debug.cfs_rq[29]:/.tg_load_avg 21927 ± 5% +119.2% 48069 ± 1% TOTAL sched_debug.cfs_rq[49]:/.tg_load_avg 21970 ± 4% +123.1% 49009 ± 2% TOTAL sched_debug.cfs_rq[22]:/.tg_load_avg 21979 ± 5% +123.1% 49045 ± 1% TOTAL sched_debug.cfs_rq[21]:/.tg_load_avg 22033 ± 5% +123.1% 49153 ± 1% TOTAL sched_debug.cfs_rq[19]:/.tg_load_avg 21979 ± 5% +118.8% 48084 ± 1% TOTAL sched_debug.cfs_rq[52]:/.tg_load_avg 22059 ± 3% +121.2% 48794 ± 1% TOTAL sched_debug.cfs_rq[28]:/.tg_load_avg 21984 ± 4% +122.9% 48996 ± 2% TOTAL sched_debug.cfs_rq[23]:/.tg_load_avg 21965 ± 4% +119.1% 48128 ± 1% TOTAL sched_debug.cfs_rq[53]:/.tg_load_avg 22037 ± 3% +121.4% 48784 ± 1% TOTAL sched_debug.cfs_rq[30]:/.tg_load_avg 22069 ± 3% +121.1% 48793 ± 1% TOTAL sched_debug.cfs_rq[31]:/.tg_load_avg 21957 ± 4% +119.6% 48208 ± 1% TOTAL sched_debug.cfs_rq[54]:/.tg_load_avg 21954 ± 5% +118.8% 48030 ± 1% TOTAL sched_debug.cfs_rq[50]:/.tg_load_avg 22117 ± 6% +122.8% 49277 ± 1% TOTAL sched_debug.cfs_rq[15]:/.tg_load_avg 21978 ± 5% +118.6% 48046 ± 1% TOTAL sched_debug.cfs_rq[51]:/.tg_load_avg 21963 ± 4% +119.7% 48246 ± 1% TOTAL sched_debug.cfs_rq[55]:/.tg_load_avg 22012 ± 3% +121.3% 48712 ± 1% TOTAL sched_debug.cfs_rq[32]:/.tg_load_avg 22037 ± 3% +121.6% 48838 ± 1% TOTAL sched_debug.cfs_rq[27]:/.tg_load_avg 22029 ± 3% +122.0% 48908 ± 1% TOTAL sched_debug.cfs_rq[26]:/.tg_load_avg 21990 ± 4% +122.3% 48891 ± 1% TOTAL sched_debug.cfs_rq[25]:/.tg_load_avg 22010 ± 4% +122.2% 48914 ± 1% TOTAL sched_debug.cfs_rq[24]:/.tg_load_avg 21985 ± 3% +121.4% 48671 ± 1% TOTAL sched_debug.cfs_rq[33]:/.tg_load_avg 22023 ± 5% +122.8% 49058 ± 2% TOTAL sched_debug.cfs_rq[20]:/.tg_load_avg 22054 ± 5% +123.3% 49240 ± 1% TOTAL sched_debug.cfs_rq[17]:/.tg_load_avg 22112 ± 6% +122.7% 49239 ± 1% TOTAL sched_debug.cfs_rq[16]:/.tg_load_avg 22058 ± 5% +122.8% 49137 ± 1% TOTAL sched_debug.cfs_rq[18]:/.tg_load_avg 2237 ± 9% +124.1% 5013 ±18% TOTAL sched_debug.cfs_rq[61]:/.tg_load_contrib 15018 ± 5% +114.2% 32164 ± 0% TOTAL sched_debug.cpu#61.ttwu_local 15138 ± 5% +113.2% 32273 ± 1% TOTAL sched_debug.cpu#91.ttwu_local 15141 ± 6% +116.6% 32798 ± 4% TOTAL sched_debug.cpu#106.ttwu_local 13386019 ± 4% +113.7% 28610387 ± 2% TOTAL proc-vmstat.pgalloc_normal 13484197 ± 4% +113.8% 28830514 ± 2% TOTAL proc-vmstat.pgfree 76806 ± 5% +111.6% 162514 ± 1% TOTAL sched_debug.cpu#91.nr_switches 76808 ± 5% +111.8% 162667 ± 1% TOTAL sched_debug.cpu#91.sched_count 13084681 ± 4% +113.2% 27900527 ± 2% TOTAL proc-vmstat.numa_local 13084697 ± 4% +113.2% 27900563 ± 2% TOTAL proc-vmstat.numa_hit 76596 ± 6% +112.9% 163099 ± 2% TOTAL sched_debug.cpu#106.nr_switches 30771 ± 6% +111.9% 65201 ± 2% TOTAL sched_debug.cpu#106.sched_goidle 30881 ± 5% +110.9% 65122 ± 2% TOTAL sched_debug.cpu#91.sched_goidle 18389572 ± 5% +111.8% 38945748 ± 1% TOTAL proc-vmstat.pgfault 3250107 ± 5% +111.8% 6885051 ± 1% TOTAL numa-numastat.node1.local_node 3250110 ± 5% +111.8% 6885062 ± 1% TOTAL numa-numastat.node1.numa_hit 3262764 ± 5% +115.3% 7024497 ± 4% TOTAL numa-numastat.node2.numa_hit 3262758 ± 5% +115.3% 7024485 ± 4% TOTAL numa-numastat.node2.local_node 3279215 ± 4% +113.9% 7015147 ± 4% TOTAL numa-numastat.node0.numa_hit 3279211 ± 4% +113.9% 7015136 ± 4% TOTAL numa-numastat.node0.local_node 76121 ± 4% +112.9% 162034 ± 1% TOTAL sched_debug.cpu#61.nr_switches 77527 ± 6% +109.0% 162036 ± 1% TOTAL sched_debug.cpu#61.sched_count 243.30 ±33% -51.6% 117.69 ±29% TOTAL sched_debug.cfs_rq[92]:/.exec_clock 30594 ± 4% +112.1% 64904 ± 2% TOTAL sched_debug.cpu#61.sched_goidle 3281833 ± 4% +109.8% 6886537 ± 1% TOTAL numa-numastat.node3.local_node 3281836 ± 4% +109.8% 6886541 ± 1% TOTAL numa-numastat.node3.numa_hit 78218 ± 6% +109.1% 163583 ± 3% TOTAL sched_debug.cpu#106.sched_count 1727502 ± 6% +107.5% 3583823 ± 4% TOTAL numa-vmstat.node2.numa_local 1742994 ± 5% +103.7% 3550217 ± 1% TOTAL numa-vmstat.node3.numa_local 1794367 ± 5% +101.6% 3617858 ± 1% TOTAL numa-vmstat.node1.numa_local 1810000 ± 5% +102.6% 3666376 ± 3% TOTAL numa-vmstat.node2.numa_hit 1825404 ± 5% +99.0% 3632638 ± 1% TOTAL numa-vmstat.node3.numa_hit 1816414 ± 5% +101.8% 3666109 ± 3% TOTAL numa-vmstat.node0.numa_local 1843627 ± 6% +100.6% 3698703 ± 1% TOTAL numa-vmstat.node1.numa_hit 3135 ±12% -46.7% 1672 ±15% TOTAL sched_debug.cpu#34.ttwu_local 1849929 ± 4% +98.3% 3668167 ± 3% TOTAL numa-vmstat.node0.numa_hit 8992 ±13% -50.9% 4418 ±41% TOTAL sched_debug.cpu#30.sched_count 241.28 ±24% -40.6% 143.43 ±25% TOTAL sched_debug.cfs_rq[11]:/.exec_clock 18020 ±39% -55.2% 8066 ±16% TOTAL sched_debug.cpu#4.ttwu_count 319 ±22% +61.6% 516 ±13% TOTAL cpuidle.C1E-IVT-4S.usage 4156 ±15% -47.6% 2176 ±41% TOTAL sched_debug.cpu#30.sched_goidle 8343 ±15% -47.6% 4375 ±41% TOTAL sched_debug.cpu#30.nr_switches 29165 ± 1% +76.4% 51461 ± 3% TOTAL sched_debug.cpu#106.ttwu_count 28980 ± 2% +73.4% 50247 ± 1% TOTAL sched_debug.cpu#61.ttwu_count 29138 ± 1% +74.5% 50853 ± 1% TOTAL sched_debug.cpu#91.ttwu_count 22537 ± 8% +70.7% 38465 ±19% TOTAL sched_debug.cpu#47.ttwu_count 1641 ± 3% +67.9% 2757 ± 1% TOTAL proc-vmstat.pgactivate 131 ±19% -37.4% 82 ± 5% TOTAL sched_debug.cpu#106.cpu_load[4] 13130 ± 2% +62.4% 21321 ± 8% TOTAL sched_debug.cpu#47.sched_goidle 7089 ±13% +54.9% 10979 ±11% TOTAL sched_debug.cpu#20.sched_goidle 26562 ± 2% +61.5% 42903 ± 7% TOTAL sched_debug.cpu#47.nr_switches 14233 ±13% +54.5% 21991 ±11% TOTAL sched_debug.cpu#20.nr_switches 88 ±17% +54.3% 135 ±25% TOTAL sched_debug.cpu#107.ttwu_local 4777 ±12% +54.7% 7389 ±14% TOTAL sched_debug.cfs_rq[34]:/.min_vruntime 119 ±12% -32.7% 80 ± 8% TOTAL sched_debug.cpu#61.cpu_load[4] 149 ±17% -33.5% 99 ± 9% TOTAL sched_debug.cpu#106.cpu_load[3] 11071 ±17% -26.7% 8120 ±22% TOTAL sched_debug.cpu#34.ttwu_count 3831 ± 6% +42.6% 5463 ± 7% TOTAL numa-meminfo.node2.KernelStack 1712 ±11% -43.0% 975 ±22% TOTAL sched_debug.cpu#1.ttwu_local 239 ± 6% +41.7% 339 ± 7% TOTAL numa-vmstat.node2.nr_kernel_stack 3638 ±24% -32.8% 2443 ±32% TOTAL sched_debug.cpu#1.ttwu_count 135 ± 7% -37.0% 85 ±11% TOTAL sched_debug.cpu#91.cpu_load[4] 5131 ±18% -21.3% 4038 ± 5% TOTAL meminfo.AnonHugePages 227 ±12% -28.6% 162 ±18% TOTAL sched_debug.cpu#91.cpu_load[0] 66199 ± 6% +49.7% 99076 ± 1% TOTAL sched_debug.cpu#106.nr_load_updates 31880 ± 6% +41.2% 45012 ±10% TOTAL sched_debug.cpu#47.sched_count 13581 ± 3% +47.7% 20066 ± 5% TOTAL sched_debug.cpu#32.sched_goidle 29309 ±12% +41.2% 41372 ± 9% TOTAL sched_debug.cpu#32.sched_count 69667 ± 4% +42.5% 99307 ± 1% TOTAL sched_debug.cpu#91.nr_load_updates 160 ±15% -26.9% 117 ±18% TOTAL sched_debug.cfs_rq[61]:/.load 27436 ± 3% +47.3% 40401 ± 5% TOTAL sched_debug.cpu#32.nr_switches 70549 ± 4% +41.8% 100061 ± 1% TOTAL sched_debug.cpu#61.nr_load_updates 13693 ± 5% +48.4% 20325 ± 6% TOTAL sched_debug.cpu#17.sched_goidle 3973 ± 6% +41.4% 5619 ± 6% TOTAL numa-meminfo.node3.KernelStack 27719 ± 5% +47.9% 40984 ± 6% TOTAL sched_debug.cpu#17.nr_switches 248 ± 6% +40.7% 349 ± 6% TOTAL numa-vmstat.node3.nr_kernel_stack 6508 ± 1% +49.1% 9705 ±22% TOTAL sched_debug.cpu#35.sched_goidle 138 ±14% -27.4% 100 ±10% TOTAL sched_debug.cpu#61.cpu_load[3] 13073 ± 2% +48.7% 19438 ±22% TOTAL sched_debug.cpu#35.nr_switches 666 ±14% +50.7% 1004 ±17% TOTAL cpuidle.C3-IVT-4S.usage 80 ±33% -45.9% 43 ±34% TOTAL sched_debug.cfs_rq[39]:/.avg->runnable_avg_sum 19457 ± 7% +31.6% 25610 ± 7% TOTAL sched_debug.cpu#47.nr_load_updates 21711 ±13% +45.4% 31570 ±19% TOTAL sched_debug.cpu#17.ttwu_count 13418 ± 3% +45.3% 19492 ±22% TOTAL sched_debug.cpu#35.sched_count 22622 ± 5% +37.5% 31103 ±15% TOTAL sched_debug.cpu#32.ttwu_count 21 ± 9% -19.6% 17 ±12% TOTAL sched_debug.cpu#99.ttwu_local 191 ±12% -25.1% 143 ±10% TOTAL sched_debug.cpu#91.cpu_load[1] 154 ± 9% -31.6% 105 ±11% TOTAL sched_debug.cpu#91.cpu_load[3] 160 ±15% -27.2% 116 ±14% TOTAL sched_debug.cpu#106.cpu_load[2] 15464 ±15% +43.3% 22163 ±11% TOTAL sched_debug.cpu#20.sched_count 176 ±14% -24.4% 133 ±23% TOTAL sched_debug.cpu#106.cpu_load[1] 169 ±10% -25.2% 126 ±10% TOTAL sched_debug.cpu#91.cpu_load[2] 20 ±10% -15.5% 17 ± 8% TOTAL sched_debug.cpu#74.ttwu_local 87151 ±17% +34.6% 117307 ± 3% TOTAL sched_debug.cfs_rq[106]:/.spread0 23131 ± 9% -20.8% 18314 ±12% TOTAL sched_debug.cpu#33.ttwu_count 9097 ±11% -23.5% 6955 ± 8% TOTAL sched_debug.cfs_rq[0]:/.exec_clock 1485 ± 2% +25.6% 1866 ± 7% TOTAL proc-vmstat.nr_kernel_stack 1788 ± 2% -20.6% 1419 ± 0% TOTAL sched_debug.cfs_rq[106]:/.tg->runnable_avg 1810 ± 2% -20.6% 1436 ± 0% TOTAL sched_debug.cfs_rq[119]:/.tg->runnable_avg 1809 ± 2% -20.7% 1435 ± 0% TOTAL sched_debug.cfs_rq[118]:/.tg->runnable_avg 1784 ± 2% -20.6% 1417 ± 0% TOTAL sched_debug.cfs_rq[105]:/.tg->runnable_avg 1798 ± 2% -20.7% 1426 ± 0% TOTAL sched_debug.cfs_rq[111]:/.tg->runnable_avg 1803 ± 2% -20.5% 1433 ± 0% TOTAL sched_debug.cfs_rq[115]:/.tg->runnable_avg 1801 ± 2% -20.5% 1431 ± 0% TOTAL sched_debug.cfs_rq[114]:/.tg->runnable_avg 1790 ± 2% -20.6% 1421 ± 0% TOTAL sched_debug.cfs_rq[107]:/.tg->runnable_avg 1799 ± 2% -20.6% 1428 ± 0% TOTAL sched_debug.cfs_rq[112]:/.tg->runnable_avg 1792 ± 2% -20.6% 1423 ± 0% TOTAL sched_debug.cfs_rq[108]:/.tg->runnable_avg 1782 ± 2% -20.5% 1415 ± 0% TOTAL sched_debug.cfs_rq[104]:/.tg->runnable_avg 1800 ± 2% -20.6% 1430 ± 0% TOTAL sched_debug.cfs_rq[113]:/.tg->runnable_avg 1805 ± 2% -20.6% 1434 ± 0% TOTAL sched_debug.cfs_rq[116]:/.tg->runnable_avg 1806 ± 2% -20.6% 1435 ± 0% TOTAL sched_debug.cfs_rq[117]:/.tg->runnable_avg 1795 ± 2% -20.7% 1424 ± 0% TOTAL sched_debug.cfs_rq[109]:/.tg->runnable_avg 95310 ± 4% +23.7% 117875 ± 2% TOTAL sched_debug.cfs_rq[91]:/.spread0 1796 ± 2% -20.7% 1425 ± 0% TOTAL sched_debug.cfs_rq[110]:/.tg->runnable_avg 1778 ± 2% -20.5% 1414 ± 0% TOTAL sched_debug.cfs_rq[103]:/.tg->runnable_avg 1771 ± 2% -20.3% 1411 ± 0% TOTAL sched_debug.cfs_rq[100]:/.tg->runnable_avg 1772 ± 2% -20.3% 1413 ± 0% TOTAL sched_debug.cfs_rq[101]:/.tg->runnable_avg 1768 ± 2% -20.3% 1410 ± 0% TOTAL sched_debug.cfs_rq[99]:/.tg->runnable_avg 1774 ± 2% -20.3% 1413 ± 0% TOTAL sched_debug.cfs_rq[102]:/.tg->runnable_avg 97534 ± 4% +21.8% 118768 ± 5% TOTAL sched_debug.cfs_rq[61]:/.spread0 1766 ± 2% -20.2% 1408 ± 0% TOTAL sched_debug.cfs_rq[98]:/.tg->runnable_avg 1762 ± 2% -20.1% 1407 ± 0% TOTAL sched_debug.cfs_rq[97]:/.tg->runnable_avg 1760 ± 2% -20.1% 1405 ± 0% TOTAL sched_debug.cfs_rq[96]:/.tg->runnable_avg 1756 ± 2% -20.0% 1405 ± 0% TOTAL sched_debug.cfs_rq[95]:/.tg->runnable_avg 1747 ± 2% -19.8% 1400 ± 0% TOTAL sched_debug.cfs_rq[92]:/.tg->runnable_avg 1753 ± 2% -19.9% 1404 ± 0% TOTAL sched_debug.cfs_rq[94]:/.tg->runnable_avg 1751 ± 2% -19.9% 1402 ± 0% TOTAL sched_debug.cfs_rq[93]:/.tg->runnable_avg 1743 ± 2% -19.8% 1398 ± 0% TOTAL sched_debug.cfs_rq[91]:/.tg->runnable_avg 23871 ± 2% +24.3% 29667 ± 8% TOTAL meminfo.KernelStack 1739 ± 2% -19.7% 1397 ± 0% TOTAL sched_debug.cfs_rq[90]:/.tg->runnable_avg 1734 ± 2% -19.6% 1395 ± 0% TOTAL sched_debug.cfs_rq[89]:/.tg->runnable_avg 1729 ± 2% -19.4% 1394 ± 0% TOTAL sched_debug.cfs_rq[88]:/.tg->runnable_avg 1725 ± 2% -19.3% 1392 ± 0% TOTAL sched_debug.cfs_rq[87]:/.tg->runnable_avg 1724 ± 2% -19.3% 1390 ± 0% TOTAL sched_debug.cfs_rq[86]:/.tg->runnable_avg 1721 ± 2% -19.3% 1389 ± 0% TOTAL sched_debug.cfs_rq[85]:/.tg->runnable_avg 1718 ± 2% -19.2% 1388 ± 0% TOTAL sched_debug.cfs_rq[84]:/.tg->runnable_avg 102757 ±13% +28.2% 131768 ± 3% TOTAL sched_debug.cfs_rq[106]:/.min_vruntime 1699 ± 2% -19.0% 1376 ± 0% TOTAL sched_debug.cfs_rq[77]:/.tg->runnable_avg 1701 ± 2% -19.0% 1378 ± 0% TOTAL sched_debug.cfs_rq[78]:/.tg->runnable_avg 1692 ± 2% -18.9% 1373 ± 0% TOTAL sched_debug.cfs_rq[75]:/.tg->runnable_avg 1695 ± 2% -19.0% 1373 ± 0% TOTAL sched_debug.cfs_rq[76]:/.tg->runnable_avg 1715 ± 2% -19.1% 1387 ± 0% TOTAL sched_debug.cfs_rq[83]:/.tg->runnable_avg 21038 ± 5% +20.1% 25263 ± 5% TOTAL sched_debug.cpu#17.nr_load_updates 1683 ± 2% -18.8% 1367 ± 0% TOTAL sched_debug.cfs_rq[71]:/.tg->runnable_avg 1709 ± 2% -18.9% 1385 ± 0% TOTAL sched_debug.cfs_rq[82]:/.tg->runnable_avg 1701 ± 2% -18.9% 1379 ± 0% TOTAL sched_debug.cfs_rq[79]:/.tg->runnable_avg 1686 ± 2% -18.8% 1369 ± 0% TOTAL sched_debug.cfs_rq[73]:/.tg->runnable_avg 1681 ± 2% -18.8% 1365 ± 0% TOTAL sched_debug.cfs_rq[70]:/.tg->runnable_avg 1705 ± 2% -18.9% 1382 ± 0% TOTAL sched_debug.cfs_rq[80]:/.tg->runnable_avg 1683 ± 2% -18.7% 1368 ± 0% TOTAL sched_debug.cfs_rq[72]:/.tg->runnable_avg 1672 ± 2% -18.5% 1362 ± 0% TOTAL sched_debug.cfs_rq[67]:/.tg->runnable_avg 1688 ± 2% -18.8% 1371 ± 0% TOTAL sched_debug.cfs_rq[74]:/.tg->runnable_avg 1663 ± 2% -18.5% 1356 ± 0% TOTAL sched_debug.cfs_rq[63]:/.tg->runnable_avg 1679 ± 2% -18.7% 1364 ± 0% TOTAL sched_debug.cfs_rq[69]:/.tg->runnable_avg 1670 ± 2% -18.5% 1362 ± 0% TOTAL sched_debug.cfs_rq[66]:/.tg->runnable_avg 1675 ± 2% -18.6% 1363 ± 0% TOTAL sched_debug.cfs_rq[68]:/.tg->runnable_avg 1665 ± 2% -18.5% 1357 ± 0% TOTAL sched_debug.cfs_rq[64]:/.tg->runnable_avg 1707 ± 2% -18.9% 1384 ± 0% TOTAL sched_debug.cfs_rq[81]:/.tg->runnable_avg 1667 ± 2% -18.5% 1359 ± 0% TOTAL sched_debug.cfs_rq[65]:/.tg->runnable_avg 152 ±14% -18.9% 123 ±12% TOTAL sched_debug.cpu#61.cpu_load[2] 1658 ± 2% -18.3% 1355 ± 0% TOTAL sched_debug.cfs_rq[62]:/.tg->runnable_avg 1652 ± 2% -18.1% 1353 ± 0% TOTAL sched_debug.cfs_rq[61]:/.tg->runnable_avg 1650 ± 2% -18.0% 1352 ± 0% TOTAL sched_debug.cfs_rq[60]:/.tg->runnable_avg 1643 ± 2% -17.9% 1348 ± 0% TOTAL sched_debug.cfs_rq[57]:/.tg->runnable_avg 113140 ± 4% +17.8% 133227 ± 5% TOTAL sched_debug.cfs_rq[61]:/.min_vruntime 1648 ± 2% -18.0% 1351 ± 0% TOTAL sched_debug.cfs_rq[59]:/.tg->runnable_avg 110916 ± 4% +19.3% 132335 ± 2% TOTAL sched_debug.cfs_rq[91]:/.min_vruntime 1625 ± 1% -17.2% 1346 ± 0% TOTAL sched_debug.cfs_rq[55]:/.tg->runnable_avg 1638 ± 2% -17.8% 1347 ± 0% TOTAL sched_debug.cfs_rq[56]:/.tg->runnable_avg 1646 ± 2% -17.9% 1350 ± 0% TOTAL sched_debug.cfs_rq[58]:/.tg->runnable_avg 1615 ± 1% -17.0% 1340 ± 0% TOTAL sched_debug.cfs_rq[51]:/.tg->runnable_avg 1616 ± 1% -17.0% 1341 ± 0% TOTAL sched_debug.cfs_rq[52]:/.tg->runnable_avg 1620 ± 1% -17.1% 1343 ± 0% TOTAL sched_debug.cfs_rq[53]:/.tg->runnable_avg 1611 ± 1% -17.0% 1337 ± 0% TOTAL sched_debug.cfs_rq[50]:/.tg->runnable_avg 1622 ± 1% -17.1% 1345 ± 0% TOTAL sched_debug.cfs_rq[54]:/.tg->runnable_avg 1605 ± 1% -16.8% 1335 ± 0% TOTAL sched_debug.cfs_rq[49]:/.tg->runnable_avg 1601 ± 1% -16.9% 1331 ± 0% TOTAL sched_debug.cfs_rq[48]:/.tg->runnable_avg 14782 ± 5% -18.7% 12017 ± 7% TOTAL sched_debug.cfs_rq[91]:/.avg->runnable_avg_sum 1595 ± 1% -16.8% 1327 ± 0% TOTAL sched_debug.cfs_rq[47]:/.tg->runnable_avg 321 ± 5% -18.9% 260 ± 7% TOTAL sched_debug.cfs_rq[91]:/.tg_runnable_contrib 17810 ±34% +39.1% 24766 ± 4% TOTAL sched_debug.cpu#2.nr_load_updates 1590 ± 1% -16.6% 1326 ± 0% TOTAL sched_debug.cfs_rq[46]:/.tg->runnable_avg 1587 ± 1% -16.5% 1324 ± 0% TOTAL sched_debug.cfs_rq[45]:/.tg->runnable_avg 1581 ± 1% -16.3% 1323 ± 0% TOTAL sched_debug.cfs_rq[44]:/.tg->runnable_avg 20364 ± 6% +21.0% 24646 ± 5% TOTAL sched_debug.cpu#32.nr_load_updates 23451 ±11% +16.0% 27201 ±10% TOTAL sched_debug.cpu#18.nr_load_updates 1576 ± 1% -16.1% 1322 ± 0% TOTAL sched_debug.cfs_rq[43]:/.tg->runnable_avg 4393 ± 3% +17.0% 5138 ± 2% TOTAL slabinfo.signal_cache.num_objs 1573 ± 1% -16.0% 1321 ± 0% TOTAL sched_debug.cfs_rq[42]:/.tg->runnable_avg 1568 ± 1% -15.9% 1319 ± 0% TOTAL sched_debug.cfs_rq[41]:/.tg->runnable_avg 1564 ± 1% -15.7% 1318 ± 0% TOTAL sched_debug.cfs_rq[40]:/.tg->runnable_avg 296 ± 4% -14.6% 253 ± 4% TOTAL sched_debug.cfs_rq[61]:/.tg_runnable_contrib 1560 ± 1% -15.5% 1318 ± 0% TOTAL sched_debug.cfs_rq[39]:/.tg->runnable_avg 1554 ± 1% -15.3% 1317 ± 0% TOTAL sched_debug.cfs_rq[38]:/.tg->runnable_avg 13680 ± 3% -14.7% 11667 ± 4% TOTAL sched_debug.cfs_rq[61]:/.avg->runnable_avg_sum 1544 ± 1% -14.8% 1315 ± 0% TOTAL sched_debug.cfs_rq[37]:/.tg->runnable_avg 3309 ± 5% -17.4% 2734 ± 2% TOTAL sched_debug.cfs_rq[32]:/.exec_clock 1534 ± 1% -14.6% 1310 ± 0% TOTAL sched_debug.cfs_rq[34]:/.tg->runnable_avg 1537 ± 1% -14.6% 1312 ± 0% TOTAL sched_debug.cfs_rq[35]:/.tg->runnable_avg 1540 ± 1% -14.7% 1314 ± 0% TOTAL sched_debug.cfs_rq[36]:/.tg->runnable_avg 553 ±43% +59.0% 879 ± 3% TOTAL numa-vmstat.node0.nr_kernel_stack 1530 ± 1% -14.5% 1308 ± 0% TOTAL sched_debug.cfs_rq[33]:/.tg->runnable_avg 1523 ± 1% -14.3% 1306 ± 0% TOTAL sched_debug.cfs_rq[32]:/.tg->runnable_avg 8851 ±43% +59.3% 14097 ± 3% TOTAL numa-meminfo.node0.KernelStack 1519 ± 1% -14.1% 1306 ± 0% TOTAL sched_debug.cfs_rq[31]:/.tg->runnable_avg 1516 ± 1% -14.0% 1304 ± 0% TOTAL sched_debug.cfs_rq[30]:/.tg->runnable_avg 1510 ± 1% -14.0% 1300 ± 0% TOTAL sched_debug.cfs_rq[28]:/.tg->runnable_avg 1513 ± 1% -14.0% 1302 ± 0% TOTAL sched_debug.cfs_rq[29]:/.tg->runnable_avg 1507 ± 1% -13.9% 1297 ± 0% TOTAL sched_debug.cfs_rq[27]:/.tg->runnable_avg 1504 ± 1% -13.8% 1296 ± 0% TOTAL sched_debug.cfs_rq[26]:/.tg->runnable_avg 1496 ± 1% -13.6% 1293 ± 0% TOTAL sched_debug.cfs_rq[24]:/.tg->runnable_avg 1492 ± 1% -13.5% 1292 ± 0% TOTAL sched_debug.cfs_rq[23]:/.tg->runnable_avg 1499 ± 1% -13.6% 1295 ± 0% TOTAL sched_debug.cfs_rq[25]:/.tg->runnable_avg 795010 ± 3% -10.6% 710653 ± 6% TOTAL sched_debug.cpu#32.avg_idle 1489 ± 1% -13.3% 1291 ± 0% TOTAL sched_debug.cfs_rq[22]:/.tg->runnable_avg 1467 ± 1% -12.9% 1278 ± 0% TOTAL sched_debug.cfs_rq[17]:/.tg->runnable_avg 1485 ± 1% -13.2% 1290 ± 0% TOTAL sched_debug.cfs_rq[21]:/.tg->runnable_avg 1463 ± 1% -12.8% 1276 ± 0% TOTAL sched_debug.cfs_rq[16]:/.tg->runnable_avg 1027 ± 6% +16.2% 1194 ± 4% TOTAL slabinfo.kmalloc-192.active_slabs 1027 ± 6% +16.2% 1194 ± 4% TOTAL slabinfo.kmalloc-192.num_slabs 43031 ± 6% +16.3% 50041 ± 4% TOTAL slabinfo.kmalloc-192.active_objs 43170 ± 6% +16.2% 50161 ± 4% TOTAL slabinfo.kmalloc-192.num_objs 1472 ± 1% -12.9% 1282 ± 0% TOTAL sched_debug.cfs_rq[18]:/.tg->runnable_avg 1479 ± 1% -13.0% 1287 ± 0% TOTAL sched_debug.cfs_rq[20]:/.tg->runnable_avg 1456 ± 1% -12.5% 1273 ± 0% TOTAL sched_debug.cfs_rq[15]:/.tg->runnable_avg 1452 ± 1% -12.3% 1273 ± 0% TOTAL sched_debug.cfs_rq[14]:/.tg->runnable_avg 862 ± 8% -12.9% 750 ± 5% TOTAL slabinfo.RAW.num_objs 862 ± 8% -12.9% 750 ± 5% TOTAL slabinfo.RAW.active_objs 1475 ± 1% -12.9% 1284 ± 0% TOTAL sched_debug.cfs_rq[19]:/.tg->runnable_avg 4393 ± 3% +14.5% 5028 ± 2% TOTAL slabinfo.signal_cache.active_objs 1446 ± 1% -12.1% 1272 ± 0% TOTAL sched_debug.cfs_rq[12]:/.tg->runnable_avg 1448 ± 1% -12.1% 1273 ± 0% TOTAL sched_debug.cfs_rq[13]:/.tg->runnable_avg 1442 ± 1% -11.9% 1271 ± 0% TOTAL sched_debug.cfs_rq[11]:/.tg->runnable_avg 1439 ± 1% -11.7% 1271 ± 0% TOTAL sched_debug.cfs_rq[10]:/.tg->runnable_avg 1437 ± 1% -11.5% 1271 ± 0% TOTAL sched_debug.cfs_rq[9]:/.tg->runnable_avg 1431 ± 1% -11.2% 1270 ± 0% TOTAL sched_debug.cfs_rq[8]:/.tg->runnable_avg 1428 ± 1% -11.1% 1269 ± 0% TOTAL sched_debug.cfs_rq[7]:/.tg->runnable_avg 1423 ± 1% -10.8% 1270 ± 0% TOTAL sched_debug.cfs_rq[6]:/.tg->runnable_avg 1421 ± 1% -10.6% 1270 ± 0% TOTAL sched_debug.cfs_rq[5]:/.tg->runnable_avg 1418 ± 1% -10.5% 1269 ± 0% TOTAL sched_debug.cfs_rq[4]:/.tg->runnable_avg 1417 ± 1% -10.5% 1268 ± 0% TOTAL sched_debug.cfs_rq[3]:/.tg->runnable_avg 5041 ± 4% +12.8% 5687 ± 1% TOTAL slabinfo.task_xstate.active_objs 5041 ± 4% +12.8% 5687 ± 1% TOTAL slabinfo.task_xstate.num_objs 20 ±18% -18.6% 16 ± 2% TOTAL sched_debug.cpu#104.ttwu_local 83828 ± 1% +9.4% 91675 ± 3% TOTAL slabinfo.kmalloc-64.active_objs 1406 ± 1% -10.1% 1264 ± 0% TOTAL sched_debug.cfs_rq[2]:/.tg->runnable_avg 1404 ± 1% -10.1% 1262 ± 0% TOTAL sched_debug.cfs_rq[1]:/.tg->runnable_avg 109592 ± 4% +6.3% 116546 ± 2% TOTAL numa-meminfo.node1.FilePages 27397 ± 4% +6.3% 29136 ± 2% TOTAL numa-vmstat.node1.nr_file_pages 36 ± 2% +8.3% 39 ± 2% TOTAL turbostat.CTMP 1382 ± 1% -9.2% 1255 ± 0% TOTAL sched_debug.cfs_rq[0]:/.tg->runnable_avg 52240 ± 5% +8.9% 56888 ± 4% TOTAL numa-meminfo.node0.Slab 31564 ± 7% +14.9% 36254 ± 5% TOTAL numa-meminfo.node1.Active 1331 ± 0% +8.1% 1439 ± 3% TOTAL slabinfo.kmalloc-64.active_slabs 1331 ± 0% +8.1% 1439 ± 3% TOTAL slabinfo.kmalloc-64.num_slabs 85255 ± 0% +8.1% 92172 ± 3% TOTAL slabinfo.kmalloc-64.num_objs 217201 ± 5% +125.5% 489860 ± 0% TOTAL time.voluntary_context_switches 17206167 ± 5% +118.8% 37639010 ± 1% TOTAL time.minor_page_faults 115930 ± 5% +116.5% 251005 ± 1% TOTAL time.involuntary_context_switches 0.00 ± 9% +121.4% 0.00 ±10% TOTAL energy.energy-cores 0.00 ± 1% +63.6% 0.00 ± 0% TOTAL energy.energy-ram 0.00 ± 3% +51.6% 0.00 ± 5% TOTAL energy.energy-pkg 7352 ± 1% +39.9% 10285 ± 0% TOTAL vmstat.system.cs 89.70 ± 0% -14.0% 77.11 ± 1% TOTAL time.user_time 1.06 ± 0% -12.9% 0.92 ± 0% TOTAL turbostat.%c0 214 ± 0% +5.9% 227 ± 0% TOTAL time.system_time time.user_time 95 ++---------------------------------------------------------------------+ | *.. | | : *.. .*. | 90 *+.*..*..*.*.. : *. *..*..*..*..* | | : | | * | 85 ++ | | | 80 ++ | | O O | | O O O O O O 75 O+ O O O O O O O O | | O O O O O O O O O | | | 70 ++---------------------------------------------------------------------+ time.system_time 235 ++--------------------------------------------------------------------+ | | | O | 230 ++ O O O O O O O O O | O O O O O O O | | O O O O O O O O 225 ++ O | | | 220 ++ | | | | * | 215 ++ + : .*.. | | + : *.*..*. *..*.*..* | *..*..*.*..* : .. | 210 ++---------------*----------------------------------------------------+ time.voluntary_context_switches 500000 ++------------------------------------------------------O--O--O----O | O | 450000 ++ O | | O O | 400000 ++ O O O O O O O O O O O | | O O | 350000 O+ O O O O | | | 300000 ++ | | | 250000 ++ | *..*.*..*..*. .*..*. .*.. | 200000 ++ *..*..*.*. *..*. * | | | 150000 ++-----------------------------------------------------------------+ energy.energy-cores 2.5e-08 ++----------------------------------------------------------------+ | O | | | 2e-08 ++ | | O O O O | O | 1.5e-08 ++ O | | O O O O | 1e-08 O+ O O O O O O O O O O O O O O | | *..*.*..*..*.*..*..*.*..* | | : | 5e-09 ++ : | | : | | : | 0 *+-*-*--*--*------------------------------------------------------+ energy.energy-pkg 4e-08 ++----------------------------------------------------------------+ | O | 3.5e-08 ++ O | O O O O | 3e-08 ++ O | 2.5e-08 ++ O O O O O O O O O O O O O O | O O O O..O.*..*..*.*..*..*.*..* | 2e-08 ++ : | | : | 1.5e-08 ++ : | 1e-08 ++ : | | : | 5e-09 ++ : | | : | 0 *+-*-*--*--*------------------------------------------------------+ aim9.shell_rtns_3.ops_per_sec 300 ++--------------------------------------------------------------------+ 280 ++ O | | O O O O 260 ++ O | 240 ++ O O | | O O O O O O O O O O O | 220 ++ O O | 200 O+ O O O O | 180 ++ | | | 160 ++ | 140 ++ | | .*.*..*..*..*..*.*..* | 120 *+.*..*.*..*..*..*. | 100 ++--------------------------------------------------------------------+ turbostat.%c0 1.12 ++-------------------------------------------------------------------+ 1.1 ++ *..* | | + + | 1.08 ++ .*. .* + | 1.06 *+.*. *..*..*. *..*.*..*..*..* | 1.04 ++ | 1.02 ++ | | | 1 ++ | 0.98 O+ O O O O O | 0.96 ++ O O O O O O O O O O O O O | 0.94 ++ O O | | O | 0.92 ++ O O O O 0.9 ++-------------------------------------------------------------------+ turbostat.Pkg_W 170 ++--------------------------------------------------------------------+ | O | 160 ++ | 150 ++ | | O 140 ++ O O O | | | 130 ++ O | | O | 120 ++ | 110 ++ O O O O | O O O O O O O O O O O O O O O | 100 ++ | *..*..*.*..*..*..*..*.*..*..*..*..*.*..* | 90 ++--------------------------------------------------------------------+ turbostat.Cor_W 100 ++--------------------------------------------------------------O-----+ | | 90 ++ | | | 80 ++ O | O O O | 70 ++ O | | O | 60 ++ | | O | 50 ++ O O O O O O O O O O O O O | O O O O O | 40 ++ | *..*..*.*..*..*..*..*.*..*..*..*..*.*..* | 30 ++--------------------------------------------------------------------+ [*] bisect-good sample [O] bisect-bad sample To reproduce: apt-get install ruby ruby-oj git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git cd lkp-tests bin/setup-local job.yaml # the job file attached in this email bin/run-local job.yaml Disclaimer: Results have been estimated based on internal Intel analysis and are provided for informational purposes only. Any difference in system hardware or software design or configuration may affect actual performance. Thanks, Fengguang
--- testcase: aim9 default_monitors: watch-oom: wait: pre-test uptime: iostat: vmstat: numa-numastat: numa-vmstat: numa-meminfo: proc-vmstat: proc-stat: meminfo: slabinfo: interrupts: lock_stat: latency_stats: softirqs: bdi_dev_mapping: diskstats: energy: cpuidle: cpufreq: turbostat: sched_debug: interval: 10 pmeter: model: Brickland Ivy Bridge-EX nr_cpu: 120 memory: 512G hdd_partitions: swap_partitions: aim9: testtime: 300s test: - shell_rtns_3 branch: linus/master commit: 19583ca584d6f574384e17fe7613dfaeadcdc4a6 repeat_to: 3 enqueue_time: 2014-09-25 21:56:15.069539322 +08:00 testbox: brickland3 kconfig: x86_64-rhel kernel: "/kernel/x86_64-rhel/19583ca584d6f574384e17fe7613dfaeadcdc4a6/vmlinuz-3.16.0" user: lkp queue: wfg result_root: "/result/brickland3/aim9/300s-shell_rtns_3/debian-x86_64.cgz/x86_64-rhel/19583ca584d6f574384e17fe7613dfaeadcdc4a6/0" job_file: "/lkp/scheduled/brickland3/wfg_aim9-300s-shell_rtns_3-x86_64-rhel-19583ca584d6f574384e17fe7613dfaeadcdc4a6-2.yaml" dequeue_time: 2014-09-30 11:42:21.894696145 +08:00 history_time: 348.87 job_state: finished loadavg: 0.84 0.66 0.31 1/967 52904 start_time: '1412048625' end_time: '1412048925' version: "/lkp/lkp/.src-20140929-152043"
_______________________________________________ LKP mailing list l...@linux.intel.com