FYI, we noticed the below changes on

commit fb7332a9fedfd62b1ba6530c86f39f0fa38afd49 ("mmu_gather: move minimal 
range calculations into generic code")


testbox/testcase/testparams: ivb42/will-it-scale/performance-brk1

63648dd20fa0780a  fb7332a9fedfd62b1ba6530c86  
----------------  --------------------------  
         %stddev     %change         %stddev
             \          |                \  
    733067 ±  0%      -6.2%     687433 ±  0%  will-it-scale.per_thread_ops
   3728462 ±  0%      -5.7%    3516927 ±  0%  will-it-scale.per_process_ops
      0.46 ±  0%      -3.3%       0.44 ±  0%  will-it-scale.scalability
    140435 ± 22%     -75.2%      34877 ± 32%  sched_debug.cpu#19.ttwu_count
        85 ± 32%    +144.6%        208 ± 30%  
sched_debug.cfs_rq[8]:/.blocked_load_avg
    892437 ± 34%     -55.5%     397177 ± 16%  sched_debug.cpu#5.nr_switches
    445593 ± 34%     -55.5%     198460 ± 16%  sched_debug.cpu#5.sched_goidle
        92 ± 29%    +136.7%        219 ± 29%  
sched_debug.cfs_rq[8]:/.tg_load_contrib
    374967 ± 14%     -56.3%     163891 ± 35%  sched_debug.cpu#5.ttwu_count
    924309 ± 32%     -51.0%     452960 ±  9%  sched_debug.cpu#5.sched_count
        90 ± 41%     +78.2%        161 ± 18%  
sched_debug.cfs_rq[30]:/.tg_load_contrib
    205529 ± 15%     -38.3%     126863 ± 32%  sched_debug.cpu#40.ttwu_count
      1152 ± 33%     +50.4%       1734 ± 22%  sched_debug.cpu#13.ttwu_local
      3.44 ± 38%     +86.6%       6.42 ± 18%  
perf-profile.cpu-cycles.rwsem_spin_on_owner.rwsem_down_write_failed.call_rwsem_down_write_failed.sys_brk.system_call_fastpath
         6 ± 39%    +100.0%         13 ±  9%  sched_debug.cpu#2.cpu_load[0]
      1.03 ± 11%     +46.6%       1.52 ± 12%  
perf-profile.cpu-cycles.find_vma.do_munmap.sys_brk.system_call_fastpath.brk
      0.76 ± 32%     +65.2%       1.26 ± 14%  
perf-profile.cpu-cycles.up_write.vma_adjust.vma_merge.do_brk.sys_brk
    538886 ± 38%     +70.3%     917481 ± 20%  sched_debug.cpu#26.ttwu_count
      0.76 ± 21%     +69.6%       1.28 ± 14%  
perf-profile.cpu-cycles.find_vma.sys_brk.system_call_fastpath.brk
        16 ±  5%     +25.8%         20 ± 18%  sched_debug.cpu#30.cpu_load[1]
      2224 ± 11%     +21.8%       2709 ± 16%  sched_debug.cpu#41.curr->pid
      3.94 ±  9%     -30.5%       2.74 ± 17%  
perf-profile.cpu-cycles._raw_spin_lock.try_to_wake_up.wake_up_process.__rwsem_do_wake.rwsem_wake
        28 ± 14%     -32.1%         19 ± 14%  sched_debug.cfs_rq[25]:/.load
        16 ±  4%     +25.0%         20 ± 12%  sched_debug.cpu#32.cpu_load[2]
        17 ± 16%     +20.3%         20 ± 10%  
sched_debug.cfs_rq[34]:/.runnable_load_avg
    180505 ± 26%     -43.4%     102128 ± 18%  sched_debug.cpu#44.ttwu_count
      2135 ±  7%     +28.7%       2747 ± 22%  sched_debug.cpu#44.curr->pid
     13.14 ± 10%     +20.4%      15.82 ±  5%  
perf-profile.cpu-cycles.call_rwsem_down_write_failed.sys_brk.system_call_fastpath.brk
     13.05 ± 10%     +20.4%      15.71 ±  5%  
perf-profile.cpu-cycles.rwsem_down_write_failed.call_rwsem_down_write_failed.sys_brk.system_call_fastpath.brk
      2.30 ± 10%     +24.1%       2.86 ±  7%  
perf-profile.cpu-cycles.vma_adjust.vma_merge.do_brk.sys_brk.system_call_fastpath
      1.70 ±  6%     -13.2%       1.47 ± 11%  
perf-profile.cpu-cycles.clockevents_program_event.tick_program_event.__hrtimer_start_range_ns.hrtimer_start_range_ns.tick_nohz_restart
      5512 ±  1%     +27.7%       7040 ± 22%  
sched_debug.cfs_rq[20]:/.exec_clock
        17 ±  5%     -30.9%         11 ± 32%  sched_debug.cpu#40.load
      2.73 ±  9%     +21.0%       3.30 ±  7%  
perf-profile.cpu-cycles.vma_merge.do_brk.sys_brk.system_call_fastpath.brk
    505131 ± 13%     +26.4%     638526 ±  7%  sched_debug.cpu#32.ttwu_count
      1.09 ±  7%     -14.9%       0.93 ± 11%  
perf-profile.cpu-cycles._raw_spin_unlock_irqrestore.rwsem_wake.call_rwsem_wake.sys_brk.system_call_fastpath
        16 ±  2%     +13.8%         18 ±  8%  sched_debug.cpu#32.cpu_load[3]
      1.73 ±  6%     -12.5%       1.52 ± 11%  
perf-profile.cpu-cycles.tick_program_event.__hrtimer_start_range_ns.hrtimer_start_range_ns.tick_nohz_restart.tick_nohz_idle_exit
      1.89 ±  5%     -12.2%       1.66 ±  1%  
perf-profile.cpu-cycles.set_next_entity.pick_next_task_fair.__sched_text_start.schedule_preempt_disabled.cpu_startup_entry
     17.50 ±  5%     -14.4%      14.98 ±  6%  
perf-profile.cpu-cycles.try_to_wake_up.wake_up_process.__rwsem_do_wake.rwsem_wake.call_rwsem_wake
     18.55 ±  5%     -13.8%      16.00 ±  6%  
perf-profile.cpu-cycles.wake_up_process.__rwsem_do_wake.rwsem_wake.call_rwsem_wake.sys_brk
       229 ±  6%     -10.5%        205 ±  0%  
sched_debug.cfs_rq[2]:/.tg_runnable_contrib
     10557 ±  6%     -10.2%       9478 ±  0%  
sched_debug.cfs_rq[2]:/.avg->runnable_avg_sum
     18.66 ±  5%     -13.4%      16.16 ±  5%  
perf-profile.cpu-cycles.__rwsem_do_wake.rwsem_wake.call_rwsem_wake.sys_brk.system_call_fastpath
    745968 ±  4%      +9.1%     813977 ±  5%  sched_debug.cpu#10.avg_idle
      3.53 ±  2%     +13.0%       3.98 ±  5%  
perf-profile.cpu-cycles.unmap_region.do_munmap.sys_brk.system_call_fastpath.brk
    738761 ±  5%     +13.5%     838441 ±  3%  sched_debug.cpu#2.avg_idle
     21.09 ±  5%     -13.4%      18.27 ±  6%  
perf-profile.cpu-cycles.rwsem_wake.call_rwsem_wake.sys_brk.system_call_fastpath.brk
        16 ±  6%     +12.1%         18 ±  2%  sched_debug.cpu#29.cpu_load[2]
     21.21 ±  5%     -13.2%      18.40 ±  6%  
perf-profile.cpu-cycles.call_rwsem_wake.sys_brk.system_call_fastpath.brk
     11.09 ±  3%     +11.3%      12.35 ±  2%  
perf-profile.cpu-cycles.do_munmap.sys_brk.system_call_fastpath.brk
    678615 ±  4%      +7.4%     728663 ±  3%  sched_debug.cpu#25.avg_idle
      2757 ±  5%      -7.4%       2551 ±  2%  sched_debug.cpu#25.curr->pid
       885 ± 11%     +22.0%       1080 ± 11%  slabinfo.RAW.num_objs
       885 ± 11%     +22.0%       1080 ± 11%  slabinfo.RAW.active_objs
      2528 ±  3%      +4.7%       2647 ±  5%  sched_debug.cpu#35.curr->pid
       230 ±  4%     -10.0%        207 ±  1%  
sched_debug.cfs_rq[8]:/.tg_runnable_contrib

testbox/testcase/testparams: lkp-t410/tlbflush/performance-200%-32x-512

63648dd20fa0780a  fb7332a9fedfd62b1ba6530c86  
----------------  --------------------------  
       113 ±  0%      -5.5%        107 ±  0%  tlbflush.mem_acc_cost_ns_time
      8758 ±  0%      +5.4%       9227 ±  0%  tlbflush.mem_acc_time_thread_ms
      3314 ± 11%     -21.2%       2610 ± 15%  slabinfo.anon_vma.num_objs

ivb42: Ivytown Ivy Bridge-EP
Memory: 64G

lkp-t410: Westmere
Memory: 2G




                          tlbflush.mem_acc_time_thread_ms

  9300 ++-O-------------------O--------O-------------------O----------------O
       |                   O        O                            O O     O  |
  9200 O+       O  O    O        O        O O  O  O  O  O     O       O     |
  9100 ++    O                                                              |
       |                                                                    |
  9000 ++            O                                                      |
       |                                                                    |
  8900 ++                                        .*                         |
       |  *                     .*..      *    *.  +                        |
  8800 ++. +        .*..*     *.         + + ..     +  .*..  .*..           |
  8700 *+   +     .*     :   :      *.. +   *        *.    *.    *          |
       |     *..*.        :  :         *                                    |
  8600 ++                 : :                                               |
       |                   *                                                |
  8500 ++-------------------------------------------------------------------+


        [*] bisect-good sample
        [O] bisect-bad  sample

To reproduce:

        apt-get install ruby
        git clone 
git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
        cd lkp-tests
        bin/setup-local job.yaml # the job file attached in this email
        bin/run-local   job.yaml


Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


Thanks,
Fengguang


---
testcase: tlbflush
default_monitors:
  wait: pre-test
  uptime: 
  iostat: 
  vmstat: 
  numa-numastat: 
  numa-vmstat: 
  numa-meminfo: 
  proc-vmstat: 
  proc-stat: 
  meminfo: 
  slabinfo: 
  interrupts: 
  lock_stat: 
  latency_stats: 
  softirqs: 
  bdi_dev_mapping: 
  diskstats: 
  cpuidle: 
  cpufreq: 
  turbostat: 
  sched_debug:
    interval: 10
  pmeter: 
default_watchdogs:
  watch-oom: 
  watchdog: 
cpufreq_governor:
- performance
commit: 97bf6af1f928216fd6c5a66e8a57bfa95a659672
model: Westmere
memory: 2G
hdd_partitions: "/dev/disk/by-id/ata-FUJITSU_MJA2250BH_G2_K95CT9C2G29W-part6"
swap_partitions: 
rootfs_partition: "/dev/disk/by-id/ata-FUJITSU_MJA2250BH_G2_K95CT9C2G29W-part7"
nr_threads: 200%
iterations: 32x
tlbflush:
  entries: 512
testbox: lkp-t410
tbox_group: lkp-t410
kconfig: x86_64-rhel
enqueue_time: 2014-12-15 13:40:54.179622910 +08:00
head_commit: 9718143f8eb121a6c53757855427f41da3a5ad7f
base_commit: 97bf6af1f928216fd6c5a66e8a57bfa95a659672
branch: linux-devel/devel-hourly-2014122315
kernel: 
"/kernel/x86_64-rhel/97bf6af1f928216fd6c5a66e8a57bfa95a659672/vmlinuz-3.19.0-rc1-g97bf6af"
user: lkp
queue: cyclic
rootfs: debian-x86_64.cgz
result_root: 
"/result/lkp-t410/tlbflush/performance-200%-32x-512/debian-x86_64.cgz/x86_64-rhel/97bf6af1f928216fd6c5a66e8a57bfa95a659672/0"
job_file: 
"/lkp/scheduled/lkp-t410/cyclic_tlbflush-performance-200%-32x-512-x86_64-rhel-BASE-97bf6af1f928216fd6c5a66e8a57bfa95a659672-0.yaml"
dequeue_time: 2014-12-23 16:43:56.407165575 +08:00
job_state: finished
loadavg: 0.10 0.13 0.08 1/98 7750
start_time: '1419324268'
end_time: '1419324762'
version: "/lkp/lkp/.src-20141223-112927"
echo performance > /sys/devices/system/cpu/cpu0/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu1/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu2/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu3/cpufreq/scaling_governor
./tlbflush -n 91 -t 8
./tlbflush -n 159 -t 8
./tlbflush -n 205 -t 8
./tlbflush -n 314 -t 8
./tlbflush -n 352 -t 8
./tlbflush -n 430 -t 8
./tlbflush -n 479 -t 8
./tlbflush -n 36 -t 8
./tlbflush -n 76 -t 8
./tlbflush -n 166 -t 8
./tlbflush -n 253 -t 8
./tlbflush -n 271 -t 8
./tlbflush -n 372 -t 8
./tlbflush -n 397 -t 8
./tlbflush -n 511 -t 8
./tlbflush -n 50 -t 8
./tlbflush -n 87 -t 8
./tlbflush -n 162 -t 8
./tlbflush -n 223 -t 8
./tlbflush -n 271 -t 8
./tlbflush -n 378 -t 8
./tlbflush -n 437 -t 8
./tlbflush -n 465 -t 8
./tlbflush -n 27 -t 8
./tlbflush -n 97 -t 8
./tlbflush -n 172 -t 8
./tlbflush -n 225 -t 8
./tlbflush -n 289 -t 8
./tlbflush -n 339 -t 8
./tlbflush -n 436 -t 8
./tlbflush -n 449 -t 8
./tlbflush -n 53 -t 8
_______________________________________________
LKP mailing list
l...@linux.intel.com

Reply via email to