[lkp] [mm/vmstat] 6cdb18ad98: -8.5% will-it-scale.per_thread_ops

From: kernel test robot
Date: Tue Jan 05 2016 - 22:21:07 EST


FYI, we noticed the below changes on

https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git master
commit 6cdb18ad98a49f7e9b95d538a0614cde827404b8 ("mm/vmstat: fix overflow in mod_zone_page_state()")


=========================================================================================
compiler/cpufreq_governor/kconfig/rootfs/tbox_group/test/testcase:
gcc-4.9/performance/x86_64-rhel/debian-x86_64-2015-02-07.cgz/ivb42/pread1/will-it-scale

commit:
cc28d6d80f6ab494b10f0e2ec949eacd610f66e3
6cdb18ad98a49f7e9b95d538a0614cde827404b8

cc28d6d80f6ab494 6cdb18ad98a49f7e9b95d538a0
---------------- --------------------------
%stddev %change %stddev
\ | \
2733943 ± 0% -8.5% 2502129 ± 0% will-it-scale.per_thread_ops
3410 ± 0% -2.0% 3343 ± 0% will-it-scale.time.system_time
340.08 ± 0% +19.7% 406.99 ± 0% will-it-scale.time.user_time
69882822 ± 2% -24.3% 52926191 ± 5% cpuidle.C1-IVT.time
340.08 ± 0% +19.7% 406.99 ± 0% time.user_time
491.25 ± 6% -17.7% 404.25 ± 7% numa-vmstat.node0.nr_alloc_batch
2799 ± 20% -36.6% 1776 ± 0% numa-vmstat.node0.nr_mapped
630.00 ±140% +244.4% 2169 ± 1% numa-vmstat.node1.nr_inactive_anon
6440 ± 11% -15.5% 5440 ± 16% numa-vmstat.node1.nr_slab_reclaimable
11204 ± 20% -36.6% 7106 ± 0% numa-meminfo.node0.Mapped
1017 ±173% +450.3% 5598 ± 15% numa-meminfo.node1.AnonHugePages
2521 ±140% +244.1% 8678 ± 1% numa-meminfo.node1.Inactive(anon)
25762 ± 11% -15.5% 21764 ± 16% numa-meminfo.node1.SReclaimable
70103 ± 9% -9.8% 63218 ± 9% numa-meminfo.node1.Slab
2.29 ± 3% +32.8% 3.04 ± 4% perf-profile.cycles-pp.atime_needs_update.touch_atime.shmem_file_read_iter.__vfs_read.vfs_read
1.10 ± 3% -27.4% 0.80 ± 5% perf-profile.cycles-pp.current_fs_time.atime_needs_update.touch_atime.shmem_file_read_iter.__vfs_read
2.33 ± 2% -13.0% 2.02 ± 3% perf-profile.cycles-pp.fput.entry_SYSCALL_64_fastpath
0.89 ± 2% +29.6% 1.15 ± 7% perf-profile.cycles-pp.fsnotify.vfs_read.sys_pread64.entry_SYSCALL_64_fastpath
2.85 ± 2% +45.4% 4.14 ± 5% perf-profile.cycles-pp.touch_atime.shmem_file_read_iter.__vfs_read.vfs_read.sys_pread64
63939 ± 0% +17.9% 75370 ± 15% sched_debug.cfs_rq:/.exec_clock.25
72.50 ± 73% -63.1% 26.75 ± 19% sched_debug.cfs_rq:/.load_avg.1
34.00 ± 62% -61.8% 13.00 ± 12% sched_debug.cfs_rq:/.load_avg.14
18.00 ± 11% -11.1% 16.00 ± 10% sched_debug.cfs_rq:/.load_avg.20
14.75 ± 41% +122.0% 32.75 ± 26% sched_debug.cfs_rq:/.load_avg.25
278.88 ± 11% +18.8% 331.25 ± 7% sched_debug.cfs_rq:/.load_avg.max
51.89 ± 11% +13.6% 58.97 ± 4% sched_debug.cfs_rq:/.load_avg.stddev
7.25 ± 5% +255.2% 25.75 ± 53% sched_debug.cfs_rq:/.runnable_load_avg.25
28.50 ± 1% +55.3% 44.25 ± 46% sched_debug.cfs_rq:/.runnable_load_avg.7
72.50 ± 73% -63.1% 26.75 ± 19% sched_debug.cfs_rq:/.tg_load_avg_contrib.1
34.00 ± 62% -61.8% 13.00 ± 12% sched_debug.cfs_rq:/.tg_load_avg_contrib.14
18.00 ± 11% -11.1% 16.00 ± 10% sched_debug.cfs_rq:/.tg_load_avg_contrib.20
14.75 ± 41% +122.0% 32.75 ± 25% sched_debug.cfs_rq:/.tg_load_avg_contrib.25
279.29 ± 11% +19.1% 332.67 ± 7% sched_debug.cfs_rq:/.tg_load_avg_contrib.max
52.01 ± 11% +13.8% 59.18 ± 4% sched_debug.cfs_rq:/.tg_load_avg_contrib.stddev
359.50 ± 6% +41.5% 508.75 ± 22% sched_debug.cfs_rq:/.util_avg.25
206.25 ± 16% -13.1% 179.25 ± 11% sched_debug.cfs_rq:/.util_avg.40
688.75 ± 1% +18.5% 816.00 ± 1% sched_debug.cfs_rq:/.util_avg.7
953467 ± 1% -17.9% 782518 ± 10% sched_debug.cpu.avg_idle.5
9177 ± 43% +73.9% 15957 ± 29% sched_debug.cpu.nr_switches.13
7365 ± 19% -35.4% 4755 ± 11% sched_debug.cpu.nr_switches.20
12203 ± 28% -62.2% 4608 ± 9% sched_debug.cpu.nr_switches.22
1868 ± 49% -51.1% 913.50 ± 27% sched_debug.cpu.nr_switches.27
2546 ± 56% -70.0% 763.00 ± 18% sched_debug.cpu.nr_switches.28
3003 ± 78% -77.9% 663.00 ± 18% sched_debug.cpu.nr_switches.33
1820 ± 19% +68.0% 3058 ± 33% sched_debug.cpu.nr_switches.8
-4.00 ±-35% -156.2% 2.25 ± 85% sched_debug.cpu.nr_uninterruptible.11
4.00 ±133% -187.5% -3.50 ±-24% sched_debug.cpu.nr_uninterruptible.17
1.75 ± 74% -214.3% -2.00 ±-127% sched_debug.cpu.nr_uninterruptible.25
0.00 ± 2% +Inf% 4.00 ± 39% sched_debug.cpu.nr_uninterruptible.26
2.50 ± 44% -110.0% -0.25 ±-591% sched_debug.cpu.nr_uninterruptible.27
1.33 ±154% -287.5% -2.50 ±-72% sched_debug.cpu.nr_uninterruptible.32
-1.00 ±-244% -250.0% 1.50 ±251% sched_debug.cpu.nr_uninterruptible.45
3.50 ± 82% -135.7% -1.25 ±-66% sched_debug.cpu.nr_uninterruptible.46
-4.50 ±-40% -133.3% 1.50 ±242% sched_debug.cpu.nr_uninterruptible.6
-3.00 ±-78% -433.3% 10.00 ±150% sched_debug.cpu.nr_uninterruptible.7
10124 ± 39% +65.8% 16783 ± 23% sched_debug.cpu.sched_count.13
12833 ± 23% -54.6% 5823 ± 32% sched_debug.cpu.sched_count.22
1934 ± 48% -49.8% 971.00 ± 26% sched_debug.cpu.sched_count.27
3065 ± 76% -76.2% 728.25 ± 16% sched_debug.cpu.sched_count.33
2098 ± 24% +664.1% 16030 ±126% sched_debug.cpu.sched_count.5
4653 ± 33% +83.4% 8536 ± 25% sched_debug.cpu.sched_goidle.15
5061 ± 41% -61.1% 1968 ± 13% sched_debug.cpu.sched_goidle.22
834.75 ± 57% -60.2% 332.00 ± 35% sched_debug.cpu.sched_goidle.27
719.00 ± 71% -63.3% 264.00 ± 19% sched_debug.cpu.sched_goidle.28
943.25 ±115% -76.3% 223.25 ± 21% sched_debug.cpu.sched_goidle.33
2520 ± 26% +112.4% 5353 ± 19% sched_debug.cpu.ttwu_count.13
5324 ± 22% -49.7% 2679 ± 45% sched_debug.cpu.ttwu_count.22
2926 ± 38% +231.1% 9690 ± 37% sched_debug.cpu.ttwu_count.23
277.25 ± 18% +166.7% 739.50 ± 83% sched_debug.cpu.ttwu_count.27
1247 ± 61% -76.6% 292.25 ± 11% sched_debug.cpu.ttwu_count.28
751.75 ± 22% +183.9% 2134 ± 9% sched_debug.cpu.ttwu_count.3
6405 ± 97% -75.9% 1542 ± 48% sched_debug.cpu.ttwu_count.41
5582 ±104% -76.2% 1327 ± 55% sched_debug.cpu.ttwu_count.43
3201 ± 26% -75.1% 796.75 ± 18% sched_debug.cpu.ttwu_local.22


ivb42: Ivytown Ivy Bridge-EP
Memory: 64G

To reproduce:

git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
cd lkp-tests
bin/lkp install job.yaml # job file is attached in this email
bin/lkp run job.yaml


Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


Thanks,
Ying Huang
---
LKP_SERVER: inn
LKP_CGI_PORT: 80
LKP_CIFS_PORT: 139
testcase: will-it-scale
default-monitors:
wait: activate-monitor
kmsg:
uptime:
iostat:
vmstat:
numa-numastat:
numa-vmstat:
numa-meminfo:
proc-vmstat:
proc-stat:
interval: 10
meminfo:
slabinfo:
interrupts:
lock_stat:
latency_stats:
softirqs:
bdi_dev_mapping:
diskstats:
nfsstat:
cpuidle:
cpufreq-stats:
turbostat:
pmeter:
sched_debug:
interval: 60
cpufreq_governor: performance
default-watchdogs:
oom-killer:
watchdog:
commit: 6cdb18ad98a49f7e9b95d538a0614cde827404b8
model: Ivytown Ivy Bridge-EP
nr_cpu: 48
memory: 64G
swap_partitions: LABEL=SWAP
rootfs_partition: LABEL=LKP-ROOTFS
category: benchmark
perf-profile:
freq: 800
will-it-scale:
test: pread1
queue: bisect
testbox: ivb42
tbox_group: ivb42
kconfig: x86_64-rhel
enqueue_time: 2016-01-05 05:00:40.511744641 +08:00
id: 374e605d3cbf102941031de6640b9edf424e5409
user: lkp
compiler: gcc-4.9
head_commit: f4366aad18b531cf15057f70e3cea09fef88c310
base_commit: 168309855a7d1e16db751e9c647119fe2d2dc878
branch: internal-eywa/master
rootfs: debian-x86_64-2015-02-07.cgz
result_root: "/result/will-it-scale/performance-pread1/ivb42/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/6cdb18ad98a49f7e9b95d538a0614cde827404b8/0"
job_file: "/lkp/scheduled/ivb42/bisect_will-it-scale-performance-pread1-debian-x86_64-2015-02-07.cgz-x86_64-rhel-6cdb18ad98a49f7e9b95d538a0614cde827404b8-20160105-44334-2cdw3i-0.yaml"
max_uptime: 1500
initrd: "/osimage/debian/debian-x86_64-2015-02-07.cgz"
bootloader_append:
- root=/dev/ram0
- user=lkp
- job=/lkp/scheduled/ivb42/bisect_will-it-scale-performance-pread1-debian-x86_64-2015-02-07.cgz-x86_64-rhel-6cdb18ad98a49f7e9b95d538a0614cde827404b8-20160105-44334-2cdw3i-0.yaml
- ARCH=x86_64
- kconfig=x86_64-rhel
- branch=internal-eywa/master
- commit=6cdb18ad98a49f7e9b95d538a0614cde827404b8
- BOOT_IMAGE=/pkg/linux/x86_64-rhel/gcc-4.9/6cdb18ad98a49f7e9b95d538a0614cde827404b8/vmlinuz-4.4.0-rc7-00013-g6cdb18a
- max_uptime=1500
- RESULT_ROOT=/result/will-it-scale/performance-pread1/ivb42/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/6cdb18ad98a49f7e9b95d538a0614cde827404b8/0
- LKP_SERVER=inn
- |2-


earlyprintk=ttyS0,115200 systemd.log_level=err
debug apic=debug sysrq_always_enabled rcupdate.rcu_cpu_stall_timeout=100
panic=-1 softlockup_panic=1 nmi_watchdog=panic oops=panic load_ramdisk=2 prompt_ramdisk=0
console=ttyS0,115200 console=tty0 vga=normal

rw
lkp_initrd: "/lkp/lkp/lkp-x86_64.cgz"
modules_initrd: "/pkg/linux/x86_64-rhel/gcc-4.9/6cdb18ad98a49f7e9b95d538a0614cde827404b8/modules.cgz"
bm_initrd: "/osimage/deps/debian-x86_64-2015-02-07.cgz/lkp.cgz,/osimage/deps/debian-x86_64-2015-02-07.cgz/run-ipconfig.cgz,/osimage/deps/debian-x86_64-2015-02-07.cgz/turbostat.cgz,/lkp/benchmarks/turbostat.cgz,/lkp/benchmarks/will-it-scale.cgz"
linux_headers_initrd: "/pkg/linux/x86_64-rhel/gcc-4.9/6cdb18ad98a49f7e9b95d538a0614cde827404b8/linux-headers.cgz"
repeat_to: 2
kernel: "/pkg/linux/x86_64-rhel/gcc-4.9/6cdb18ad98a49f7e9b95d538a0614cde827404b8/vmlinuz-4.4.0-rc7-00013-g6cdb18a"
dequeue_time: 2016-01-05 05:06:46.434314862 +08:00
job_state: finished
loadavg: 38.55 18.46 7.28 1/506 9298
start_time: '1451941650'
end_time: '1451941960'
version: "/lkp/lkp/.src-20160104-165204"

Attachment: reproduce.sh
Description: Bourne shell script