FYI, we noticed the below changes on

commit fc7f0dd381720ea5ee5818645f7d0e9dece41cb0 ("kernel: avoid overflow in 
cmp_range")


testbox/testcase/testparams: lituya/will-it-scale/powersave-mmap2

7ad4b4ae5757b896  fc7f0dd381720ea5ee5818645f  
----------------  --------------------------  
         %stddev     %change         %stddev
             \          |                \  
    252693 ±  0%      -2.2%     247031 ±  0%  will-it-scale.per_thread_ops
      0.18 ±  0%      +1.8%       0.19 ±  0%  will-it-scale.scalability
     43536 ± 24%    +276.2%     163774 ± 33%  sched_debug.cpu#6.ttwu_local
      3.55 ±  2%     +36.2%       4.84 ±  2%  
perf-profile.cpu-cycles.___might_sleep.unmap_page_range.unmap_single_vma.unmap_vmas.unmap_region
      8.49 ± 12%     -29.5%       5.99 ±  5%  
perf-profile.cpu-cycles._raw_spin_lock_irqsave.__percpu_counter_add.do_munmap.vm_munmap.sys_munmap
     12.27 ±  8%     -20.2%       9.80 ±  3%  
perf-profile.cpu-cycles.__percpu_counter_add.do_munmap.vm_munmap.sys_munmap.system_call_fastpath
      7.45 ±  7%     -20.8%       5.90 ±  5%  
perf-profile.cpu-cycles._raw_spin_lock_irqsave.__percpu_counter_add.__vm_enough_memory.selinux_vm_enough_memory.security_vm_enough_memory_mm
     11.11 ±  3%     -12.9%       9.67 ±  3%  
perf-profile.cpu-cycles.__percpu_counter_add.__vm_enough_memory.selinux_vm_enough_memory.security_vm_enough_memory_mm.mmap_region
      2.46 ±  3%     +13.1%       2.78 ±  2%  
perf-profile.cpu-cycles.___might_sleep.unmap_single_vma.unmap_vmas.unmap_region.do_munmap
     11.42 ±  3%     -12.3%      10.01 ±  2%  
perf-profile.cpu-cycles.__vm_enough_memory.selinux_vm_enough_memory.security_vm_enough_memory_mm.mmap_region.do_mmap_pgoff
     12.39 ±  3%     -11.2%      11.00 ±  2%  
perf-profile.cpu-cycles.selinux_vm_enough_memory.security_vm_enough_memory_mm.mmap_region.do_mmap_pgoff.vm_mmap_pgoff
     12.45 ±  3%     -11.1%      11.07 ±  2%  
perf-profile.cpu-cycles.security_vm_enough_memory_mm.mmap_region.do_mmap_pgoff.vm_mmap_pgoff.sys_mmap_pgoff
     14.38 ±  1%      +9.5%      15.75 ±  1%  
perf-profile.cpu-cycles.unmap_page_range.unmap_single_vma.unmap_vmas.unmap_region.do_munmap

testbox/testcase/testparams: lituya/will-it-scale/performance-mmap2

7ad4b4ae5757b896  fc7f0dd381720ea5ee5818645f  
----------------  --------------------------  
    268761 ±  0%      -2.1%     263177 ±  0%  will-it-scale.per_thread_ops
      0.18 ±  0%      +1.8%       0.19 ±  0%  will-it-scale.scalability
      0.01 ± 37%     -99.3%       0.00 ± 12%  sched_debug.rt_rq[10]:/.rt_time
    104123 ± 41%     -63.7%      37788 ± 45%  sched_debug.cpu#5.ttwu_local
    459901 ± 48%     +60.7%     739071 ± 31%  sched_debug.cpu#6.ttwu_count
   1858053 ± 12%     -36.9%    1171826 ± 38%  sched_debug.cpu#10.sched_goidle
   3716823 ± 12%     -36.9%    2344353 ± 38%  sched_debug.cpu#10.nr_switches
   3777468 ± 11%     -36.9%    2383575 ± 36%  sched_debug.cpu#10.sched_count
        36 ± 28%     -40.9%         21 ±  7%  sched_debug.cpu#6.cpu_load[1]
     18042 ± 17%     +54.0%      27789 ± 30%  sched_debug.cfs_rq[4]:/.exec_clock
        56 ± 17%     -48.8%         29 ±  5%  
sched_debug.cfs_rq[6]:/.runnable_load_avg
        36 ± 29%     +43.6%         52 ± 11%  sched_debug.cpu#4.load
    594415 ±  4%     +82.4%    1084432 ± 18%  sched_debug.cpu#2.ttwu_count
        15 ±  0%     +51.1%         22 ± 14%  sched_debug.cpu#4.cpu_load[4]
      2077 ± 11%     -36.7%       1315 ± 15%  sched_debug.cpu#6.curr->pid
        11 ± 28%     +48.6%         17 ± 23%  sched_debug.cpu#7.cpu_load[4]
      0.00 ± 20%     +77.0%       0.00 ± 26%  sched_debug.rt_rq[5]:/.rt_time
        16 ±  5%     +52.1%         24 ±  9%  sched_debug.cpu#4.cpu_load[3]
        17 ± 11%     +50.0%         26 ±  8%  sched_debug.cpu#4.cpu_load[2]
     48035 ±  7%     -22.2%      37362 ± 24%  
sched_debug.cfs_rq[12]:/.exec_clock
        34 ± 12%     -24.5%         25 ± 20%  
sched_debug.cfs_rq[12]:/.runnable_load_avg
        33 ± 11%     -24.2%         25 ± 20%  sched_debug.cpu#12.cpu_load[4]
        19 ± 25%     +50.9%         28 ±  3%  sched_debug.cpu#4.cpu_load[1]
        66 ± 17%     -24.7%         49 ±  5%  sched_debug.cpu#6.load
    421462 ± 16%     +18.8%     500676 ± 13%  
sched_debug.cfs_rq[1]:/.min_vruntime
      3.60 ±  0%     +35.4%       4.87 ±  0%  
perf-profile.cpu-cycles.___might_sleep.unmap_page_range.unmap_single_vma.unmap_vmas.unmap_region
        44 ±  9%     +37.9%         60 ± 17%  sched_debug.cpu#3.load
        37 ±  6%     -17.9%         30 ± 15%  sched_debug.cpu#15.cpu_load[3]
      6.96 ±  4%     -10.4%       6.24 ±  3%  
perf-profile.cpu-cycles._raw_spin_lock_irqsave.__percpu_counter_add.do_munmap.vm_munmap.sys_munmap
        36 ±  6%     +24.1%         44 ±  2%  sched_debug.cpu#2.load
        39 ±  7%     -16.9%         32 ± 12%  sched_debug.cpu#15.cpu_load[2]
   1528695 ±  6%     -19.5%    1230190 ± 16%  sched_debug.cpu#10.ttwu_count
        36 ±  6%     +27.3%         46 ±  9%  sched_debug.cpu#10.load
       447 ±  3%     -13.9%        385 ± 10%  
sched_debug.cfs_rq[15]:/.tg_runnable_contrib
     20528 ±  3%     -13.8%      17701 ± 10%  
sched_debug.cfs_rq[15]:/.avg->runnable_avg_sum
    634808 ±  6%     +50.3%     954347 ± 24%  sched_debug.cpu#2.sched_goidle
   1270648 ±  6%     +50.3%    1909528 ± 24%  sched_debug.cpu#2.nr_switches
   1284042 ±  6%     +51.4%    1944604 ± 23%  sched_debug.cpu#2.sched_count
        55 ± 11%     +28.7%         71 ±  4%  sched_debug.cpu#8.cpu_load[0]
      6.39 ±  0%      -8.7%       5.84 ±  2%  
perf-profile.cpu-cycles._raw_spin_lock_irqsave.__percpu_counter_add.__vm_enough_memory.selinux_vm_enough_memory.security_vm_enough_memory_mm
     48721 ± 11%     +19.1%      58037 ±  5%  sched_debug.cpu#11.nr_load_updates
        53 ±  9%     +16.1%         62 ±  1%  sched_debug.cpu#8.cpu_load[1]
      1909 ±  0%     +22.2%       2333 ±  9%  sched_debug.cpu#3.curr->pid
      0.95 ±  4%      -8.4%       0.87 ±  4%  
perf-profile.cpu-cycles.file_map_prot_check.selinux_mmap_file.security_mmap_file.vm_mmap_pgoff.sys_mmap_pgoff
    567608 ±  8%     +11.0%     629780 ±  4%  
sched_debug.cfs_rq[14]:/.min_vruntime
    804637 ± 15%     +24.4%    1000664 ± 13%  sched_debug.cpu#3.ttwu_count
    684460 ±  5%      -9.6%     618867 ±  3%  sched_debug.cpu#14.avg_idle
      1.02 ±  4%      -7.2%       0.94 ±  4%  
perf-profile.cpu-cycles.selinux_mmap_file.security_mmap_file.vm_mmap_pgoff.sys_mmap_pgoff.sys_mmap
      2605 ±  2%      -5.8%       2454 ±  5%  slabinfo.kmalloc-96.active_objs
      2605 ±  2%      -5.8%       2454 ±  5%  slabinfo.kmalloc-96.num_objs
        50 ±  4%     +11.3%         56 ±  1%  
sched_debug.cfs_rq[8]:/.runnable_load_avg
      1.15 ±  4%      -6.4%       1.08 ±  4%  
perf-profile.cpu-cycles.security_mmap_file.vm_mmap_pgoff.sys_mmap_pgoff.sys_mmap.system_call_fastpath
      1.07 ±  2%      +9.7%       1.17 ±  3%  
perf-profile.cpu-cycles.vma_compute_subtree_gap.__vma_link_rb.vma_link.mmap_region.do_mmap_pgoff

To reproduce:

        apt-get install ruby ruby-oj
        git clone 
git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
        cd lkp-tests
        bin/setup-local job.yaml # the job file attached in this email
        bin/run-local   job.yaml


Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


Thanks,
Huang, Ying

---
testcase: will-it-scale
default-monitors:
  wait: pre-test
  uptime: 
  iostat: 
  vmstat: 
  numa-numastat: 
  numa-vmstat: 
  numa-meminfo: 
  proc-vmstat: 
  proc-stat: 
  meminfo: 
  slabinfo: 
  interrupts: 
  lock_stat: 
  latency_stats: 
  softirqs: 
  bdi_dev_mapping: 
  diskstats: 
  nfsstat: 
  cpuidle: 
  cpufreq-stats: 
  turbostat: 
  pmeter: 
  sched_debug:
    interval: 10
default_watchdogs:
  watch-oom: 
  watchdog: 
cpufreq_governor:
- powersave
commit: ec6f34e5b552fb0a52e6aae1a5afbbb1605cc6cc
model: Grantley Haswell
nr_cpu: 16
memory: 16G
hdd_partitions: 
swap_partitions: 
rootfs_partition: 
perf-profile:
  freq: 800
will-it-scale:
  test:
  - mmap2
testbox: lituya
tbox_group: lituya
kconfig: x86_64-rhel
enqueue_time: 2015-01-18 14:10:07.541442957 +08:00
head_commit: b213d55915f2ee6748ba62f743b5e70564ab31e7
base_commit: ec6f34e5b552fb0a52e6aae1a5afbbb1605cc6cc
branch: linux-devel/devel-hourly-2015011917
kernel: 
"/kernel/x86_64-rhel/ec6f34e5b552fb0a52e6aae1a5afbbb1605cc6cc/vmlinuz-3.19.0-rc5-gec6f34e"
user: lkp
queue: cyclic
rootfs: debian-x86_64-2015-01-13.cgz
result_root: 
"/result/lituya/will-it-scale/powersave-mmap2/debian-x86_64-2015-01-13.cgz/x86_64-rhel/ec6f34e5b552fb0a52e6aae1a5afbbb1605cc6cc/0"
job_file: 
"/lkp/scheduled/lituya/cyclic_will-it-scale-powersave-mmap2-x86_64-rhel-BASE-ec6f34e5b552fb0a52e6aae1a5afbbb1605cc6cc-0.yaml"
dequeue_time: 2015-01-19 18:08:35.232473498 +08:00
job_state: finished
loadavg: 11.32 6.60 2.70 1/178 7099
start_time: '1421662149'
end_time: '1421662453'
version: "/lkp/lkp/.src-20150119-113749"
./runtest.py mmap2 32 1 8 12 16
_______________________________________________
LKP mailing list
l...@linux.intel.com

Reply via email to