[lkp] [vmpressure] 78d1788350: No primary result change, 274.1% vm-scalability.time.involuntary_context_switches

From: kernel test robot
Date: Sun Jan 31 2016 - 20:43:11 EST


FYI, we noticed the below changes on

https://github.com/0day-ci/linux Vladimir-Davydov/vmpressure-Fix-subtree-pressure-detection/20160128-003153
commit 78d1788350477246d516496d8d7684fa80ef7f18 ("vmpressure: Fix subtree pressure detection")


=========================================================================================
compiler/cpufreq_governor/kconfig/rootfs/runtime/tbox_group/test/testcase:
gcc-4.9/performance/x86_64-rhel/debian-x86_64-2015-02-07.cgz/300s/ivb43/lru-file-mmap-read/vm-scalability

commit:
v4.5-rc1
78d1788350477246d516496d8d7684fa80ef7f18

v4.5-rc1 78d1788350477246d516496d8d
---------------- --------------------------
%stddev %change %stddev
\ | \
195486 ± 1% +274.1% 731248 ± 2% vm-scalability.time.involuntary_context_switches
2939 ± 21% +37.5% 4041 ± 12% numa-vmstat.node1.nr_anon_pages
897378 ± 1% +17.7% 1056357 ± 2% softirqs.RCU
4293 ± 4% +82.5% 7835 ± 2% vmstat.system.cs
120.40 ± 22% +41.6% 170.50 ± 19% cpuidle.C3-IVT.usage
4.723e+08 ± 4% +13.8% 5.375e+08 ± 5% cpuidle.C6-IVT.time
93336 ± 31% +611.8% 664396 ± 88% latency_stats.max.pipe_wait.pipe_read.__vfs_read.vfs_read.SyS_read.entry_SYSCALL_64_fastpath
371333 ±166% -70.1% 110861 ±117% latency_stats.sum.pipe_write.__vfs_write.vfs_write.SyS_write.entry_SYSCALL_64_fastpath
195486 ± 1% +274.1% 731248 ± 2% time.involuntary_context_switches
717.20 ± 14% +20.3% 862.50 ± 6% time.major_page_faults
2724 ± 60% -93.6% 174.25 ±173% numa-meminfo.node0.AnonHugePages
1630 ±122% +222.2% 5254 ± 16% numa-meminfo.node1.AnonHugePages
11760 ± 21% +37.6% 16177 ± 12% numa-meminfo.node1.AnonPages
1.01 ± 4% +26.5% 1.28 ± 10% perf-profile.cycles-pp.__add_to_page_cache_locked.add_to_page_cache_lru.mpage_readpages.xfs_vm_readpages.__do_page_cache_readahead
0.90 ± 56% -72.4% 0.25 ± 95% perf-profile.cycles-pp.__alloc_pages_nodemask.alloc_pages_current.__page_cache_alloc.__do_page_cache_readahead.filemap_fault
0.90 ± 56% -72.5% 0.25 ± 95% perf-profile.cycles-pp.__page_cache_alloc.__do_page_cache_readahead.filemap_fault.xfs_filemap_fault.__do_fault
2.18 ± 2% +15.9% 2.52 ± 7% perf-profile.cycles-pp.__remove_mapping.shrink_page_list.shrink_inactive_list.shrink_zone_memcg.shrink_zone
1.04 ± 4% +16.8% 1.21 ± 7% perf-profile.cycles-pp.__rmqueue.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_current.__page_cache_alloc
1.08 ± 6% +18.1% 1.28 ± 5% perf-profile.cycles-pp._raw_spin_lock_irqsave.__remove_mapping.shrink_page_list.shrink_inactive_list.shrink_zone_memcg
0.90 ± 56% -72.4% 0.25 ± 95% perf-profile.cycles-pp.alloc_pages_current.__page_cache_alloc.__do_page_cache_readahead.filemap_fault.xfs_filemap_fault
0.77 ± 20% +47.0% 1.14 ± 9% perf-profile.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.__remove_mapping.shrink_page_list.shrink_inactive_list
1.38 ± 2% +15.8% 1.60 ± 7% perf-profile.cycles-pp.rmap_walk.try_to_unmap.shrink_page_list.shrink_inactive_list.shrink_zone_memcg
2.17 ± 2% -14.7% 1.85 ± 9% perf-profile.cycles-pp.shrink_inactive_list.shrink_zone_memcg.shrink_zone.kswapd.kthread
2.19 ± 2% -14.5% 1.87 ± 9% perf-profile.cycles-pp.shrink_zone.kswapd.kthread.ret_from_fork
2.18 ± 1% -14.8% 1.86 ± 9% perf-profile.cycles-pp.shrink_zone_memcg.shrink_zone.kswapd.kthread.ret_from_fork
1.46 ± 2% +19.5% 1.75 ± 10% perf-profile.cycles-pp.try_to_unmap.shrink_page_list.shrink_inactive_list.shrink_zone_memcg.shrink_zone
50.20 ±133% +273.0% 187.25 ±109% sched_debug.cfs_rq:/.load.5
56.60 ± 11% +97.4% 111.75 ± 55% sched_debug.cfs_rq:/.load_avg.0
17.60 ± 9% +25.0% 22.00 ± 14% sched_debug.cfs_rq:/.load_avg.19
47.00 ± 33% -54.8% 21.25 ± 13% sched_debug.cfs_rq:/.load_avg.23
21.60 ± 30% +139.6% 51.75 ± 63% sched_debug.cfs_rq:/.load_avg.30
20.80 ± 30% +50.2% 31.25 ± 26% sched_debug.cfs_rq:/.load_avg.36
16.40 ± 2% +18.9% 19.50 ± 11% sched_debug.cfs_rq:/.runnable_load_avg.14
16685 ±278% -374.2% -45753 ±-25% sched_debug.cfs_rq:/.spread0.12
119599 ± 47% -71.8% 33762 ±121% sched_debug.cfs_rq:/.spread0.27
106418 ± 42% -267.6% -178399 ±-221% sched_debug.cfs_rq:/.spread0.28
56797 ± 81% -136.5% -20757 ±-159% sched_debug.cfs_rq:/.spread0.3
63128 ± 63% -157.9% -36549 ±-295% sched_debug.cfs_rq:/.spread0.46
57759 ±100% -676.1% -332737 ±-164% sched_debug.cfs_rq:/.spread0.8
151767 ± 32% -40.2% 90684 ± 14% sched_debug.cfs_rq:/.spread0.max
855.00 ± 1% +7.5% 918.75 ± 7% sched_debug.cfs_rq:/.util_avg.20
16.40 ± 2% +18.9% 19.50 ± 11% sched_debug.cpu.cpu_load[0].14
16.40 ± 2% +18.9% 19.50 ± 11% sched_debug.cpu.cpu_load[1].14
16.40 ± 2% +18.9% 19.50 ± 11% sched_debug.cpu.cpu_load[2].14
16.40 ± 2% +18.9% 19.50 ± 11% sched_debug.cpu.cpu_load[3].14
16.40 ± 2% +18.9% 19.50 ± 11% sched_debug.cpu.cpu_load[4].14
1305 ± 1% +16.6% 1521 ± 13% sched_debug.cpu.curr->pid.23
50.20 ±133% +273.0% 187.25 ±109% sched_debug.cpu.load.5
0.00 ± 0% +Inf% 1.00 ± 0% sched_debug.cpu.nr_running.5
10459 ± 30% +86.1% 19462 ± 13% sched_debug.cpu.nr_switches.12
11042 ±113% +169.5% 29759 ± 31% sched_debug.cpu.nr_switches.14
17516 ± 69% +85.9% 32567 ± 46% sched_debug.cpu.nr_switches.18
8287 ± 69% +271.4% 30775 ± 71% sched_debug.cpu.nr_switches.21
16525 ± 59% +80.6% 29850 ± 41% sched_debug.cpu.nr_switches.23
8306 ± 36% +92.8% 16016 ± 8% sched_debug.cpu.nr_switches.24
9271 ± 64% +149.7% 23153 ± 30% sched_debug.cpu.nr_switches.25
6094 ± 49% +243.8% 20951 ± 31% sched_debug.cpu.nr_switches.26
6569 ± 32% +194.9% 19373 ± 17% sched_debug.cpu.nr_switches.30
10034 ±111% +295.8% 39717 ± 41% sched_debug.cpu.nr_switches.31
7417 ± 70% +205.4% 22657 ± 35% sched_debug.cpu.nr_switches.38
9204 ± 69% +125.4% 20746 ± 26% sched_debug.cpu.nr_switches.42
8742 ± 68% +172.8% 23845 ± 14% sched_debug.cpu.nr_switches.47
11267 ± 75% +311.6% 46375 ± 68% sched_debug.cpu.nr_switches.7
14373 ± 4% +74.8% 25131 ± 3% sched_debug.cpu.nr_switches.avg
68237 ± 15% +30.3% 88895 ± 10% sched_debug.cpu.nr_switches.max
1792 ± 11% +610.9% 12741 ± 4% sched_debug.cpu.nr_switches.min
-1.00 ±-244% -200.0% 1.00 ±212% sched_debug.cpu.nr_uninterruptible.21
-29.60 ±-22% -31.7% -20.21 ±-29% sched_debug.cpu.nr_uninterruptible.min
11570 ± 27% +75.8% 20344 ± 14% sched_debug.cpu.sched_count.12
18187 ± 52% +67.7% 30504 ± 40% sched_debug.cpu.sched_count.23
8874 ± 33% +83.7% 16301 ± 7% sched_debug.cpu.sched_count.24
9847 ± 60% +137.9% 23423 ± 29% sched_debug.cpu.sched_count.25
7299 ± 38% +190.8% 21225 ± 30% sched_debug.cpu.sched_count.26
7156 ± 30% +179.8% 20022 ± 16% sched_debug.cpu.sched_count.30
10606 ±105% +279.7% 40278 ± 41% sched_debug.cpu.sched_count.31
7977 ± 66% +187.5% 22938 ± 35% sched_debug.cpu.sched_count.38
10713 ± 47% +138.5% 25554 ± 54% sched_debug.cpu.sched_count.39
25911 ± 63% +76.9% 45826 ± 39% sched_debug.cpu.sched_count.4
9758 ± 66% +115.6% 21039 ± 26% sched_debug.cpu.sched_count.42
9252 ± 64% +160.6% 24110 ± 14% sched_debug.cpu.sched_count.47
12153 ± 69% +296.0% 48126 ± 65% sched_debug.cpu.sched_count.7
110528 ± 0% +9.1% 120637 ± 0% sched_debug.cpu.sched_count.avg
2364 ± 8% +455.3% 13129 ± 5% sched_debug.cpu.sched_count.min
816.60 ± 32% -40.5% 485.75 ± 16% sched_debug.cpu.sched_goidle.17
383.20 ± 22% +56.1% 598.00 ± 22% sched_debug.cpu.sched_goidle.23
204.20 ± 12% +210.7% 634.50 ± 86% sched_debug.cpu.sched_goidle.25
190.60 ± 26% +54.6% 294.75 ± 12% sched_debug.cpu.sched_goidle.31
4752 ± 36% +107.1% 9844 ± 12% sched_debug.cpu.ttwu_count.12
5428 ±115% +178.2% 15102 ± 31% sched_debug.cpu.ttwu_count.14
8902 ± 73% +80.8% 16095 ± 47% sched_debug.cpu.ttwu_count.18
3698 ± 62% +315.1% 15351 ± 75% sched_debug.cpu.ttwu_count.21
8371 ± 57% +80.6% 15121 ± 41% sched_debug.cpu.ttwu_count.23
4569 ± 33% +85.9% 8496 ± 10% sched_debug.cpu.ttwu_count.24
4884 ± 56% +152.6% 12339 ± 27% sched_debug.cpu.ttwu_count.25
3187 ± 39% +258.7% 11434 ± 23% sched_debug.cpu.ttwu_count.26
3513 ± 31% +183.5% 9962 ± 19% sched_debug.cpu.ttwu_count.30
5462 ± 97% +283.9% 20970 ± 41% sched_debug.cpu.ttwu_count.31
5605 ± 43% +85.8% 10415 ± 34% sched_debug.cpu.ttwu_count.34
4011 ± 66% +190.7% 11661 ± 33% sched_debug.cpu.ttwu_count.38
4305 ± 41% +183.7% 12215 ± 59% sched_debug.cpu.ttwu_count.39
4828 ± 73% +125.9% 10906 ± 25% sched_debug.cpu.ttwu_count.42
4560 ± 74% +168.1% 12227 ± 12% sched_debug.cpu.ttwu_count.47
5663 ± 78% +313.7% 23431 ± 67% sched_debug.cpu.ttwu_count.7
7326 ± 4% +74.8% 12808 ± 3% sched_debug.cpu.ttwu_count.avg
34268 ± 15% +29.6% 44398 ± 9% sched_debug.cpu.ttwu_count.max
792.03 ± 20% +698.7% 6325 ± 5% sched_debug.cpu.ttwu_count.min
3951 ± 45% +119.7% 8681 ± 13% sched_debug.cpu.ttwu_local.12
4585 ±137% +205.0% 13987 ± 33% sched_debug.cpu.ttwu_local.14
8045 ± 80% +88.6% 15173 ± 48% sched_debug.cpu.ttwu_local.18
2941 ± 79% +386.9% 14319 ± 79% sched_debug.cpu.ttwu_local.21
7262 ± 67% +95.1% 14172 ± 44% sched_debug.cpu.ttwu_local.23
3991 ± 40% +100.4% 8000 ± 10% sched_debug.cpu.ttwu_local.24
4237 ± 68% +162.6% 11128 ± 29% sched_debug.cpu.ttwu_local.25
2570 ± 41% +292.2% 10081 ± 30% sched_debug.cpu.ttwu_local.26
2891 ± 40% +215.8% 9132 ± 21% sched_debug.cpu.ttwu_local.30
4665 ±113% +318.2% 19509 ± 43% sched_debug.cpu.ttwu_local.31
4533 ± 50% +115.0% 9748 ± 35% sched_debug.cpu.ttwu_local.34
4131 ± 56% +145.5% 10144 ± 47% sched_debug.cpu.ttwu_local.37
3349 ± 82% +225.3% 10897 ± 34% sched_debug.cpu.ttwu_local.38
3630 ± 48% +219.9% 11611 ± 63% sched_debug.cpu.ttwu_local.39
4138 ± 82% +140.4% 9947 ± 29% sched_debug.cpu.ttwu_local.42
3356 ± 78% +238.4% 11359 ± 14% sched_debug.cpu.ttwu_local.47
4866 ± 91% +360.6% 22415 ± 71% sched_debug.cpu.ttwu_local.7
6406 ± 5% +84.0% 11787 ± 4% sched_debug.cpu.ttwu_local.avg
33023 ± 15% +29.7% 42825 ± 11% sched_debug.cpu.ttwu_local.max
450.27 ± 20% +1199.8% 5852 ± 4% sched_debug.cpu.ttwu_local.min
1.58 ± 4% -55.1% 0.71 ± 99% sched_debug.rt_rq:/.rt_time.12

ivb43: Ivytown Ivy Bridge-EP
Memory: 64G

vm-scalability.time.involuntary_context_switches

800000 ++-----------------------------------------------------------------+
| O O O O O O O |
700000 ++ O O O O O O O O O O O
| |
600000 O+ |
| O O |
500000 ++ O O O O |
| |
400000 ++ |
| |
300000 ++ |
| |
200000 *+.*..*.*..*..*..*..*.*..*..*..*..*.*..*..*..*.*..*..*..*..*.* |
| |
100000 ++-----------------------------------------------------------------+

[*] bisect-good sample
[O] bisect-bad sample

To reproduce:

git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
cd lkp-tests
bin/lkp install job.yaml # job file is attached in this email
bin/lkp run job.yaml


Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


Thanks,
Ying Huang
---
LKP_SERVER: inn
LKP_CGI_PORT: 80
LKP_CIFS_PORT: 139
testcase: vm-scalability
default-monitors:
wait: activate-monitor
kmsg:
uptime:
iostat:
vmstat:
numa-numastat:
numa-vmstat:
numa-meminfo:
proc-vmstat:
proc-stat:
interval: 10
meminfo:
slabinfo:
interrupts:
lock_stat:
latency_stats:
softirqs:
bdi_dev_mapping:
diskstats:
nfsstat:
cpuidle:
cpufreq-stats:
turbostat:
pmeter:
sched_debug:
interval: 60
cpufreq_governor: performance
default-watchdogs:
oom-killer:
watchdog:
commit: 78d1788350477246d516496d8d7684fa80ef7f18
model: Ivytown Ivy Bridge-EP
nr_cpu: 48
memory: 64G
nr_hdd_partitions: 0
hdd_partitions:
rootfs_partition:
swap_partitions:
category: benchmark
perf-profile:
runtime: 300s
size:
vm-scalability:
test: lru-file-mmap-read
queue: bisect
testbox: ivb43
tbox_group: ivb43
kconfig: x86_64-rhel
enqueue_time: 2016-01-31 04:22:27.102563621 +08:00
id: 2542d63be154b791122544fcefff3f1d6d23a064
user: lkp
compiler: gcc-4.9
head_commit: 5edef856d8228f4fc81476cd9ff36693ad0ffccb
base_commit: 92e963f50fc74041b5e9e744c330dca48e04f08d
branch: linux-devel/devel-hourly-2016013009
rootfs: debian-x86_64-2015-02-07.cgz
result_root: "/result/vm-scalability/performance-300s-lru-file-mmap-read/ivb43/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/78d1788350477246d516496d8d7684fa80ef7f18/0"
job_file: "/lkp/scheduled/ivb43/bisect_vm-scalability-performance-300s-lru-file-mmap-read-debian-x86_64-2015-02-07.cgz-x86_64-rhel-78d1788350477246d516496d8d7684fa80ef7f18-20160131-120638-crvdne-0.yaml"
max_uptime: 1500
initrd: "/osimage/debian/debian-x86_64-2015-02-07.cgz"
bootloader_append:
- root=/dev/ram0
- user=lkp
- job=/lkp/scheduled/ivb43/bisect_vm-scalability-performance-300s-lru-file-mmap-read-debian-x86_64-2015-02-07.cgz-x86_64-rhel-78d1788350477246d516496d8d7684fa80ef7f18-20160131-120638-crvdne-0.yaml
- ARCH=x86_64
- kconfig=x86_64-rhel
- branch=linux-devel/devel-hourly-2016013009
- commit=78d1788350477246d516496d8d7684fa80ef7f18
- BOOT_IMAGE=/pkg/linux/x86_64-rhel/gcc-4.9/78d1788350477246d516496d8d7684fa80ef7f18/vmlinuz-4.5.0-rc1-00001-g78d1788
- max_uptime=1500
- RESULT_ROOT=/result/vm-scalability/performance-300s-lru-file-mmap-read/ivb43/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/78d1788350477246d516496d8d7684fa80ef7f18/0
- LKP_SERVER=inn
- |2-


earlyprintk=ttyS0,115200 systemd.log_level=err
debug apic=debug sysrq_always_enabled rcupdate.rcu_cpu_stall_timeout=100
panic=-1 softlockup_panic=1 nmi_watchdog=panic oops=panic load_ramdisk=2 prompt_ramdisk=0
console=ttyS0,115200 console=tty0 vga=normal

rw
lkp_initrd: "/lkp/lkp/lkp-x86_64.cgz"
modules_initrd: "/pkg/linux/x86_64-rhel/gcc-4.9/78d1788350477246d516496d8d7684fa80ef7f18/modules.cgz"
bm_initrd: "/osimage/deps/debian-x86_64-2015-02-07.cgz/lkp.cgz,/osimage/deps/debian-x86_64-2015-02-07.cgz/run-ipconfig.cgz,/osimage/deps/debian-x86_64-2015-02-07.cgz/turbostat.cgz,/lkp/benchmarks/turbostat.cgz,/lkp/benchmarks/vm-scalability.cgz"
linux_headers_initrd: "/pkg/linux/x86_64-rhel/gcc-4.9/78d1788350477246d516496d8d7684fa80ef7f18/linux-headers.cgz"
repeat_to: 2
kernel: "/pkg/linux/x86_64-rhel/gcc-4.9/78d1788350477246d516496d8d7684fa80ef7f18/vmlinuz-4.5.0-rc1-00001-g78d1788"
dequeue_time: 2016-01-31 04:37:47.136664651 +08:00
job_state: finished
loadavg: 23.80 27.32 13.15 2/525 8525
start_time: '1454186311'
end_time: '1454186622'
version: "/lkp/lkp/.src-20160127-223853"

Attachment: reproduce.sh
Description: Bourne shell script