[lkp] [vfs] f3f86e33dc: -5.3% will-it-scale.per_process_ops
From: kernel test robot
Date: Wed Nov 18 2015 - 01:44:58 EST
FYI, we noticed the below changes on
https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git master
commit f3f86e33dc3da437fa4f204588ce7c78ea756982 ("vfs: Fix pathological performance case for __alloc_fd()")
=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/test:
ivb42/will-it-scale/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performance/dup1
commit:
8a28d67457b613258aa0578ccece206d166f2b9f
f3f86e33dc3da437fa4f204588ce7c78ea756982
8a28d67457b61325 f3f86e33dc3da437fa4f204588
---------------- --------------------------
%stddev %change %stddev
\ | \
5994379 ± 0% -5.3% 5678711 ± 0% will-it-scale.per_process_ops
1440545 ± 0% -5.1% 1367766 ± 2% will-it-scale.per_thread_ops
0.57 ± 0% -5.9% 0.54 ± 0% will-it-scale.scalability
4.47 ± 2% -3.1% 4.33 ± 1% turbostat.RAMWatt
59880 ± 5% -13.1% 52055 ± 11% cpuidle.C1-IVT.usage
597.50 ± 4% -19.7% 479.50 ± 16% cpuidle.POLL.usage
15756223 ± 0% +367.7% 73688311 ± 84% latency_stats.avg.nfs_wait_on_request.nfs_updatepage.nfs_write_end.generic_perform_write.__generic_file_write_iter.generic_file_write_iter.nfs_file_write.__vfs_write.vfs_write.SyS_write.entry_SYSCALL_64_fastpath
35858260 ± 0% +113.3% 76474871 ± 77% latency_stats.max.nfs_wait_on_request.nfs_updatepage.nfs_write_end.generic_perform_write.__generic_file_write_iter.generic_file_write_iter.nfs_file_write.__vfs_write.vfs_write.SyS_write.entry_SYSCALL_64_fastpath
1560 ±171% +300.1% 6241 ± 0% numa-numastat.node0.other_node
3101 ± 99% -99.7% 9.50 ± 67% numa-numastat.node1.other_node
2980 ± 3% -13.4% 2582 ± 1% slabinfo.kmalloc-2048.active_objs
3139 ± 3% -12.5% 2746 ± 1% slabinfo.kmalloc-2048.num_objs
5018 ± 14% -50.4% 2487 ± 13% numa-vmstat.node0.nr_active_anon
3121 ± 31% -77.6% 700.00 ±132% numa-vmstat.node0.nr_shmem
3349 ± 20% +76.6% 5916 ± 7% numa-vmstat.node1.nr_active_anon
1210 ± 80% +200.8% 3640 ± 25% numa-vmstat.node1.nr_shmem
70442 ± 5% -12.7% 61484 ± 0% numa-meminfo.node0.Active
20079 ± 14% -50.4% 9954 ± 13% numa-meminfo.node0.Active(anon)
12487 ± 31% -77.6% 2801 ±132% numa-meminfo.node0.Shmem
61970 ± 5% +18.0% 73096 ± 1% numa-meminfo.node1.Active
13402 ± 20% +76.6% 23671 ± 7% numa-meminfo.node1.Active(anon)
4843 ± 80% +200.7% 14564 ± 25% numa-meminfo.node1.Shmem
1999660 ± 2% -9.1% 1817792 ± 6% sched_debug.cfs_rq[0]:/.min_vruntime
814.25 ± 6% -11.9% 717.00 ± 10% sched_debug.cfs_rq[0]:/.util_avg
-220009 ±-25% -83.5% -36294 ±-314% sched_debug.cfs_rq[10]:/.spread0
-220410 ±-25% -82.7% -38205 ±-290% sched_debug.cfs_rq[11]:/.spread0
-868154 ± -5% -20.7% -688065 ±-16% sched_debug.cfs_rq[14]:/.spread0
13.00 ± 0% +105.8% 26.75 ± 61% sched_debug.cfs_rq[15]:/.load
-952278 ±-16% -27.9% -687025 ±-16% sched_debug.cfs_rq[16]:/.spread0
-876660 ± -6% -21.8% -685915 ±-15% sched_debug.cfs_rq[17]:/.spread0
-869841 ± -6% -20.5% -691511 ±-15% sched_debug.cfs_rq[18]:/.spread0
-872906 ± -6% -21.0% -689435 ±-15% sched_debug.cfs_rq[19]:/.spread0
-220042 ±-24% -82.8% -37798 ±-282% sched_debug.cfs_rq[1]:/.spread0
-870736 ± -6% -20.9% -689178 ±-16% sched_debug.cfs_rq[20]:/.spread0
-870782 ± -5% -20.6% -691440 ±-16% sched_debug.cfs_rq[21]:/.spread0
12.50 ± 12% +20.0% 15.00 ± 8% sched_debug.cfs_rq[23]:/.load_avg
-947289 ±-16% -27.8% -684292 ±-16% sched_debug.cfs_rq[23]:/.spread0
12.50 ± 12% +20.0% 15.00 ± 8% sched_debug.cfs_rq[23]:/.tg_load_avg_contrib
424.00 ± 13% +27.4% 540.00 ± 9% sched_debug.cfs_rq[23]:/.util_avg
-180921 ±-30% -100.2% 424.29 ±26645% sched_debug.cfs_rq[25]:/.spread0
-179335 ±-30% -82.3% -31706 ±-346% sched_debug.cfs_rq[26]:/.spread0
-180972 ±-30% -100.1% 163.84 ±68609% sched_debug.cfs_rq[27]:/.spread0
-179636 ±-30% -100.4% 736.15 ±15384% sched_debug.cfs_rq[28]:/.spread0
-180380 ±-30% -101.1% 1963 ±5772% sched_debug.cfs_rq[29]:/.spread0
26.00 ± 3% -18.3% 21.25 ± 22% sched_debug.cfs_rq[2]:/.load
29.50 ± 9% -21.2% 23.25 ± 23% sched_debug.cfs_rq[2]:/.runnable_load_avg
-211354 ±-27% -97.3% -5780 ±-2383% sched_debug.cfs_rq[2]:/.spread0
762.25 ± 6% -17.1% 632.00 ± 11% sched_debug.cfs_rq[2]:/.util_avg
-179346 ±-31% -101.0% 1767 ±6351% sched_debug.cfs_rq[30]:/.spread0
-182129 ±-30% -99.9% -200.51 ±-56625% sched_debug.cfs_rq[31]:/.spread0
-178388 ±-30% -99.9% -162.33 ±-69718% sched_debug.cfs_rq[32]:/.spread0
-178678 ±-30% -100.0% -67.48 ±-166628% sched_debug.cfs_rq[33]:/.spread0
-177514 ±-30% -100.1% 200.37 ±56326% sched_debug.cfs_rq[34]:/.spread0
-178339 ±-29% -101.6% 2870 ±3873% sched_debug.cfs_rq[35]:/.spread0
-795803 ± -8% -34.5% -521305 ±-40% sched_debug.cfs_rq[37]:/.spread0
-783897 ± -6% -22.6% -607100 ±-18% sched_debug.cfs_rq[38]:/.spread0
3.00 ± 0% +250.0% 10.50 ± 40% sched_debug.cfs_rq[39]:/.load_avg
-784040 ± -6% -33.2% -523669 ±-39% sched_debug.cfs_rq[39]:/.spread0
3.00 ± 0% +250.0% 10.50 ± 40% sched_debug.cfs_rq[39]:/.tg_load_avg_contrib
173.75 ± 4% +36.8% 237.75 ± 30% sched_debug.cfs_rq[39]:/.util_avg
-220092 ±-24% -82.4% -38783 ±-288% sched_debug.cfs_rq[3]:/.spread0
-783338 ± -6% -22.8% -604971 ±-18% sched_debug.cfs_rq[41]:/.spread0
-784423 ± -6% -23.1% -603402 ±-17% sched_debug.cfs_rq[42]:/.spread0
-785872 ± -6% -23.0% -605005 ±-18% sched_debug.cfs_rq[43]:/.spread0
-782962 ± -6% -22.9% -603838 ±-19% sched_debug.cfs_rq[44]:/.spread0
-783170 ± -6% -23.2% -601383 ±-18% sched_debug.cfs_rq[45]:/.spread0
-784950 ± -6% -23.2% -602937 ±-18% sched_debug.cfs_rq[46]:/.spread0
32.25 ± 35% -24.8% 24.25 ± 1% sched_debug.cfs_rq[4]:/.load
-217411 ±-24% -83.2% -36433 ±-300% sched_debug.cfs_rq[4]:/.spread0
-219424 ±-25% -83.0% -37233 ±-299% sched_debug.cfs_rq[5]:/.spread0
-219112 ±-25% -82.4% -38536 ±-289% sched_debug.cfs_rq[6]:/.spread0
-218643 ±-24% -82.8% -37629 ±-298% sched_debug.cfs_rq[7]:/.spread0
-220909 ±-24% -85.0% -33175 ±-350% sched_debug.cfs_rq[8]:/.spread0
-220076 ±-25% -85.9% -31115 ±-337% sched_debug.cfs_rq[9]:/.spread0
89160 ± 6% -8.7% 81395 ± 7% sched_debug.cpu#0.nr_load_updates
-2.75 ±-126% -172.7% 2.00 ±136% sched_debug.cpu#12.nr_uninterruptible
16563 ± 21% -39.6% 10009 ± 9% sched_debug.cpu#13.nr_switches
16901 ± 21% -34.5% 11064 ± 9% sched_debug.cpu#13.sched_count
6432 ± 27% -47.2% 3396 ± 38% sched_debug.cpu#13.sched_goidle
7244 ± 14% -45.3% 3961 ± 39% sched_debug.cpu#14.sched_goidle
13.00 ± 0% +105.8% 26.75 ± 61% sched_debug.cpu#15.load
1554 ± 21% +62.9% 2531 ± 30% sched_debug.cpu#16.ttwu_local
6965 ± 21% +48.0% 10308 ± 17% sched_debug.cpu#18.sched_count
28.25 ± 2% -17.7% 23.25 ± 23% sched_debug.cpu#2.cpu_load[4]
26.00 ± 3% -18.3% 21.25 ± 22% sched_debug.cpu#2.load
2703 ± 9% +11.5% 3014 ± 8% sched_debug.cpu#24.curr->pid
420.00 ± 27% -34.1% 276.75 ± 30% sched_debug.cpu#25.sched_goidle
247.00 ± 24% +55.7% 384.50 ± 26% sched_debug.cpu#27.sched_goidle
-2.25 ±-79% -211.1% 2.50 ± 82% sched_debug.cpu#30.nr_uninterruptible
715.75 ± 46% -44.2% 399.50 ± 35% sched_debug.cpu#32.ttwu_count
133.50 ± 22% +99.4% 266.25 ± 29% sched_debug.cpu#33.ttwu_local
1212 ± 47% -46.7% 646.25 ± 25% sched_debug.cpu#35.nr_switches
506.50 ± 46% -51.6% 245.00 ± 30% sched_debug.cpu#35.sched_goidle
32.25 ± 35% -24.8% 24.25 ± 1% sched_debug.cpu#4.load
2973 ± 46% +161.2% 7766 ± 47% sched_debug.cpu#40.nr_switches
3062 ± 46% +156.1% 7843 ± 47% sched_debug.cpu#40.sched_count
1219 ± 55% +155.9% 3121 ± 60% sched_debug.cpu#40.sched_goidle
1429 ± 2% +49.2% 2131 ± 41% sched_debug.cpu#44.curr->pid
1.75 ± 93% -57.1% 0.75 ±145% sched_debug.cpu#45.nr_uninterruptible
433.75 ± 32% +75.1% 759.50 ± 14% sched_debug.cpu#6.ttwu_count
ivb42: Ivytown Ivy Bridge-EP
Memory: 64G
will-it-scale.per_process_ops
6.05e+06 ++---------------------------------------------------------------+
*.*.**.*.*.**.*.*. .*. *. .* .*.**. |
6e+06 ++ **.*.*.** *.* *.*.**.*.* *.* *.*.**.*.*
5.95e+06 ++ |
| |
5.9e+06 ++ |
5.85e+06 ++ |
| |
5.8e+06 ++ |
5.75e+06 ++ |
| O OO O |
5.7e+06 O+ OO O O O O O OO O O |
5.65e+06 ++O O O O O OO O |
| O |
5.6e+06 ++---------------------------------------------------------------+
[*] bisect-good sample
[O] bisect-bad sample
To reproduce:
git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
cd lkp-tests
bin/lkp install job.yaml # job file is attached in this email
bin/lkp run job.yaml
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
Thanks,
Ying Huang
---
LKP_SERVER: inn
LKP_CGI_PORT: 80
LKP_CIFS_PORT: 139
testcase: will-it-scale
default-monitors:
wait: activate-monitor
kmsg:
uptime:
iostat:
vmstat:
numa-numastat:
numa-vmstat:
numa-meminfo:
proc-vmstat:
proc-stat:
interval: 10
meminfo:
slabinfo:
interrupts:
lock_stat:
latency_stats:
softirqs:
bdi_dev_mapping:
diskstats:
nfsstat:
cpuidle:
cpufreq-stats:
turbostat:
pmeter:
sched_debug:
interval: 60
cpufreq_governor: performance
default-watchdogs:
oom-killer:
watchdog:
commit: 8005c49d9aea74d382f474ce11afbbc7d7130bec
model: Ivytown Ivy Bridge-EP
nr_cpu: 48
memory: 64G
swap_partitions: LABEL=SWAP
rootfs_partition: LABEL=LKP-ROOTFS
category: benchmark
perf-profile:
freq: 800
will-it-scale:
test: dup1
queue: cyclic
testbox: ivb42
tbox_group: ivb42
kconfig: x86_64-rhel
enqueue_time: 2015-11-17 08:27:36.309490411 +08:00
id: 7363031303e3969c581a84334a46962a2dffa4c3
user: lkp
compiler: gcc-4.9
head_commit: a25498f782e28fcbd76b93cd9325b9e18c1c829a
base_commit: 8005c49d9aea74d382f474ce11afbbc7d7130bec
branch: linux-devel/devel-hourly-2015111705
kernel: "/pkg/linux/x86_64-rhel/gcc-4.9/8005c49d9aea74d382f474ce11afbbc7d7130bec/vmlinuz-4.4.0-rc1"
rootfs: debian-x86_64-2015-02-07.cgz
result_root: "/result/will-it-scale/performance-dup1/ivb42/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/8005c49d9aea74d382f474ce11afbbc7d7130bec/0"
job_file: "/lkp/scheduled/ivb42/cyclic_will-it-scale-performance-dup1-x86_64-rhel-CYCLIC_BASE-8005c49d9aea74d382f474ce11afbbc7d7130bec-20151117-76241-5xtdl7-0.yaml"
dequeue_time: 2015-11-17 09:51:21.843559599 +08:00
max_uptime: 1500
initrd: "/osimage/debian/debian-x86_64-2015-02-07.cgz"
bootloader_append:
- root=/dev/ram0
- user=lkp
- job=/lkp/scheduled/ivb42/cyclic_will-it-scale-performance-dup1-x86_64-rhel-CYCLIC_BASE-8005c49d9aea74d382f474ce11afbbc7d7130bec-20151117-76241-5xtdl7-0.yaml
- ARCH=x86_64
- kconfig=x86_64-rhel
- branch=linux-devel/devel-hourly-2015111705
- commit=8005c49d9aea74d382f474ce11afbbc7d7130bec
- BOOT_IMAGE=/pkg/linux/x86_64-rhel/gcc-4.9/8005c49d9aea74d382f474ce11afbbc7d7130bec/vmlinuz-4.4.0-rc1
- max_uptime=1500
- RESULT_ROOT=/result/will-it-scale/performance-dup1/ivb42/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/8005c49d9aea74d382f474ce11afbbc7d7130bec/0
- LKP_SERVER=inn
- |2-
earlyprintk=ttyS0,115200 systemd.log_level=err
debug apic=debug sysrq_always_enabled rcupdate.rcu_cpu_stall_timeout=100
panic=-1 softlockup_panic=1 nmi_watchdog=panic oops=panic load_ramdisk=2 prompt_ramdisk=0
console=ttyS0,115200 console=tty0 vga=normal
rw
lkp_initrd: "/lkp/lkp/lkp-x86_64.cgz"
modules_initrd: "/pkg/linux/x86_64-rhel/gcc-4.9/8005c49d9aea74d382f474ce11afbbc7d7130bec/modules.cgz"
bm_initrd: "/osimage/deps/debian-x86_64-2015-02-07.cgz/lkp.cgz,/osimage/deps/debian-x86_64-2015-02-07.cgz/run-ipconfig.cgz,/osimage/deps/debian-x86_64-2015-02-07.cgz/turbostat.cgz,/lkp/benchmarks/turbostat.cgz,/lkp/benchmarks/will-it-scale.cgz"
job_state: finished
loadavg: 41.80 18.86 7.37 1/501 9285
start_time: '1447725123'
end_time: '1447725433'
version: "/lkp/lkp/.src-20151116-235214"
echo performance > /sys/devices/system/cpu/cpu0/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu1/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu10/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu11/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu12/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu13/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu14/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu15/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu16/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu17/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu18/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu19/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu2/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu20/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu21/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu22/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu23/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu24/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu25/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu26/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu27/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu28/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu29/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu3/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu30/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu31/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu32/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu33/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu34/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu35/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu36/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu37/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu38/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu39/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu4/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu40/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu41/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu42/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu43/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu44/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu45/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu46/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu47/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu5/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu6/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu7/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu8/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu9/cpufreq/scaling_governor
./runtest.py dup1 25 both 1 12 24 36 48