[lkp] [x86/entry/64] fa58aafc448: 10.8% aim7.jobs-per-min
From: kernel test robot
Date: Thu Aug 20 2015 - 01:32:16 EST
FYI, we noticed the below changes on
git://git.kernel.org/pub/scm/linux/kernel/git/luto/linux.git x86/entry
commit fa58aafc44805ac425d17c6a8082513b5442ce9d ("x86/entry/64: When returning via SYSRET, POP regs instead of using MOV")
=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/load/test:
lkp-a06/aim7/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/4000/new_raph
commit:
a4be9881623375fd126762af65ef18dc8175c68d
fa58aafc44805ac425d17c6a8082513b5442ce9d
a4be9881623375fd fa58aafc44805ac425d17c6a80
---------------- --------------------------
%stddev %change %stddev
\ | \
184099 ± 0% +10.8% 204000 ± 0% aim7.jobs-per-min
131.89 ± 0% -9.8% 119.00 ± 0% aim7.time.elapsed_time
131.89 ± 0% -9.8% 119.00 ± 0% aim7.time.elapsed_time.max
2215262 ± 0% -92.5% 165275 ± 0% aim7.time.involuntary_context_switches
19.56 ± 1% -65.8% 6.70 ± 5% aim7.time.system_time
435.63 ± 0% -2.8% 423.34 ± 0% aim7.time.user_time
60385 ± 1% -17.3% 49927 ± 0% aim7.time.voluntary_context_switches
131.89 ± 0% -9.8% 119.00 ± 0% time.elapsed_time
131.89 ± 0% -9.8% 119.00 ± 0% time.elapsed_time.max
2215262 ± 0% -92.5% 165275 ± 0% time.involuntary_context_switches
19.56 ± 1% -65.8% 6.70 ± 5% time.system_time
60385 ± 1% -17.3% 49927 ± 0% time.voluntary_context_switches
=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/load/test:
lkp-a06/aim7/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/4000/pipe_cpy
commit:
a4be9881623375fd126762af65ef18dc8175c68d
fa58aafc44805ac425d17c6a8082513b5442ce9d
a4be9881623375fd fa58aafc44805ac425d17c6a80
---------------- --------------------------
%stddev %change %stddev
\ | \
247245 ± 1% +15.6% 285751 ± 1% aim7.jobs-per-min
98.61 ± 1% -13.4% 85.37 ± 1% aim7.time.elapsed_time
98.61 ± 1% -13.4% 85.37 ± 1% aim7.time.elapsed_time.max
2003598 ± 0% -93.3% 133967 ± 2% aim7.time.involuntary_context_switches
266.80 ± 1% -7.1% 247.73 ± 1% aim7.time.system_time
51.41 ± 4% -11.8% 45.32 ± 7% aim7.time.user_time
53934 ± 1% -21.5% 42329 ± 1% aim7.time.voluntary_context_switches
98.61 ± 1% -13.4% 85.37 ± 1% time.elapsed_time
98.61 ± 1% -13.4% 85.37 ± 1% time.elapsed_time.max
2003598 ± 0% -93.3% 133967 ± 2% time.involuntary_context_switches
51.41 ± 4% -11.8% 45.32 ± 7% time.user_time
53934 ± 1% -21.5% 42329 ± 1% time.voluntary_context_switches
=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/nr_threads:
lkp-a06/dbench/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performance/100%
commit:
a4be9881623375fd126762af65ef18dc8175c68d
fa58aafc44805ac425d17c6a8082513b5442ce9d
a4be9881623375fd fa58aafc44805ac425d17c6a80
---------------- --------------------------
%stddev %change %stddev
\ | \
2245160 ± 8% -76.6% 526406 ± 4% dbench.time.involuntary_context_switches
379.50 ± 0% +1.3% 384.50 ± 0% dbench.time.percent_of_cpu_this_job_got
1715 ± 0% +1.7% 1745 ± 0% dbench.time.system_time
2245160 ± 8% -76.6% 526406 ± 4% time.involuntary_context_switches
2.69 ± 11% +81.5% 4.88 ± 37% perf-profile.cpu-cycles.__hrtimer_run_queues.hrtimer_interrupt.local_apic_timer_interrupt.smp_apic_timer_interrupt.apic_timer_interrupt
1.43 ± 2% -10.1% 1.29 ± 2% perf-profile.cpu-cycles.entry_SYSCALL_64_after_swapgs
1.51 ± 8% -26.2% 1.11 ± 10% perf-profile.cpu-cycles.rcu_nocb_kthread.kthread.ret_from_fork
1.20 ± 15% +109.4% 2.51 ± 46% perf-profile.cpu-cycles.scheduler_tick.update_process_times.tick_sched_handle.tick_sched_timer.__hrtimer_run_queues
0.68 ± 16% +110.7% 1.43 ± 47% perf-profile.cpu-cycles.task_tick_fair.scheduler_tick.update_process_times.tick_sched_handle.tick_sched_timer
1.97 ± 11% +96.4% 3.87 ± 40% perf-profile.cpu-cycles.tick_sched_handle.isra.17.tick_sched_timer.__hrtimer_run_queues.hrtimer_interrupt.local_apic_timer_interrupt
2.33 ± 10% +84.8% 4.30 ± 38% perf-profile.cpu-cycles.tick_sched_timer.__hrtimer_run_queues.hrtimer_interrupt.local_apic_timer_interrupt.smp_apic_timer_interrupt
1.90 ± 11% +96.3% 3.72 ± 41% perf-profile.cpu-cycles.update_process_times.tick_sched_handle.tick_sched_timer.__hrtimer_run_queues.hrtimer_interrupt
1.07 ± 2% -9.3% 0.97 ± 3% perf-profile.cpu-cycles.vfs_create.path_openat.do_filp_open.do_sys_open.sys_open
=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/iterations/nr_threads/disk/fs/filesize/test_size/sync_method/nr_directories/nr_files_per_directory:
nhm4/fsmark/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performance/1x/32t/1HDD/ext4/8K/400M/fsyncBeforeClose/16d/256fpd
commit:
a4be9881623375fd126762af65ef18dc8175c68d
fa58aafc44805ac425d17c6a8082513b5442ce9d
a4be9881623375fd fa58aafc44805ac425d17c6a80
---------------- --------------------------
%stddev %change %stddev
\ | \
10935 ± 7% -38.1% 6768 ± 1% fsmark.time.involuntary_context_switches
2989 ± 0% +1.2% 3026 ± 0% fsmark.time.maximum_resident_set_size
10935 ± 7% -38.1% 6768 ± 1% time.involuntary_context_switches
29861 ± 3% -86.7% 3970 ± 1% vmstat.system.cs
13362 ± 3% -97.0% 405.25 ± 1% vmstat.system.in
76414335 ± 1% -55.4% 34106888 ± 4% cpuidle.C1-NHM.time
4836217 ± 0% -92.9% 344308 ± 4% cpuidle.C1-NHM.usage
1310 ± 4% -96.7% 43.00 ± 10% cpuidle.POLL.usage
1.32 ± 2% -43.9% 0.74 ± 0% turbostat.%Busy
39.25 ± 2% -51.6% 19.00 ± 0% turbostat.Avg_MHz
2985 ± 0% -15.9% 2512 ± 0% turbostat.Bzy_MHz
7.68 ± 5% -42.2% 4.44 ± 3% turbostat.CPU%c1
0.00 ± -1% +Inf% 20233 ±125% latency_stats.avg.submit_bio_wait.blkdev_issue_flush.jbd2_cleanup_journal_tail.jbd2_log_do_checkpoint.__jbd2_log_wait_for_space.start_this_handle.jbd2__journal_start.__ext4_journal_start_sb.__ext4_new_inode.ext4_mkdir.vfs_mkdir.SyS_mkdir
4866 ± 28% +42.4% 6930 ±141% latency_stats.max.do_get_write_access.jbd2_journal_get_write_access.__ext4_journal_get_write_access.ext4_reserve_inode_write.ext4_mark_inode_dirty.ext4_dirty_inode.__mark_inode_dirty.generic_update_time.file_update_time.__generic_file_write_iter.ext4_file_write_iter.__vfs_write
8314 ± 73% +365.2% 38680 ± 69% latency_stats.max.do_get_write_access.jbd2_journal_get_write_access.__ext4_journal_get_write_access.ext4_reserve_inode_write.ext4_mark_inode_dirty.ext4_dirty_inode.__mark_inode_dirty.generic_write_end.ext4_da_write_end.generic_perform_write.__generic_file_write_iter.ext4_file_write_iter
0.00 ± -1% +Inf% 24939 ±105% latency_stats.max.submit_bio_wait.blkdev_issue_flush.jbd2_cleanup_journal_tail.jbd2_log_do_checkpoint.__jbd2_log_wait_for_space.start_this_handle.jbd2__journal_start.__ext4_journal_start_sb.__ext4_new_inode.ext4_mkdir.vfs_mkdir.SyS_mkdir
0.00 ± -1% +Inf% 24960 ±105% latency_stats.sum.submit_bio_wait.blkdev_issue_flush.jbd2_cleanup_journal_tail.jbd2_log_do_checkpoint.__jbd2_log_wait_for_space.start_this_handle.jbd2__journal_start.__ext4_journal_start_sb.__ext4_new_inode.ext4_mkdir.vfs_mkdir.SyS_mkdir
5099 ± 5% +8.2% 5517 ± 5% sched_debug.cfs_rq[0]:/.min_vruntime
533.25 ± 3% -9.6% 482.25 ± 2% sched_debug.cfs_rq[0]:/.tg->runnable_avg
537.75 ± 3% -9.9% 484.75 ± 2% sched_debug.cfs_rq[1]:/.tg->runnable_avg
11.50 ± 35% +65.2% 19.00 ± 11% sched_debug.cfs_rq[2]:/.nr_spread_over
538.50 ± 3% -9.7% 486.50 ± 2% sched_debug.cfs_rq[2]:/.tg->runnable_avg
-1924 ±-24% +34.3% -2583 ±-12% sched_debug.cfs_rq[3]:/.spread0
539.75 ± 3% -10.4% 483.75 ± 2% sched_debug.cfs_rq[3]:/.tg->runnable_avg
1006 ± 13% +17.2% 1179 ± 5% sched_debug.cfs_rq[4]:/.exec_clock
2780 ± 16% +20.9% 3361 ± 7% sched_debug.cfs_rq[4]:/.min_vruntime
542.75 ± 3% -10.7% 484.50 ± 2% sched_debug.cfs_rq[4]:/.tg->runnable_avg
2626 ± 5% +41.7% 3723 ± 12% sched_debug.cfs_rq[5]:/.avg->runnable_avg_sum
2463 ± 8% +16.3% 2865 ± 7% sched_debug.cfs_rq[5]:/.min_vruntime
547.00 ± 4% -11.4% 484.50 ± 2% sched_debug.cfs_rq[5]:/.tg->runnable_avg
56.75 ± 4% +41.9% 80.50 ± 13% sched_debug.cfs_rq[5]:/.tg_runnable_contrib
909.00 ± 74% +241.7% 3105 ± 4% sched_debug.cfs_rq[6]:/.blocked_load_avg
549.00 ± 4% -11.5% 486.00 ± 2% sched_debug.cfs_rq[6]:/.tg->runnable_avg
927.25 ± 71% +240.7% 3158 ± 6% sched_debug.cfs_rq[6]:/.tg_load_contrib
4572 ± 22% -49.6% 2303 ± 27% sched_debug.cfs_rq[7]:/.avg->runnable_avg_sum
-1634 ±-23% +55.2% -2535 ±-19% sched_debug.cfs_rq[7]:/.spread0
551.00 ± 4% -11.4% 488.25 ± 3% sched_debug.cfs_rq[7]:/.tg->runnable_avg
98.00 ± 22% -49.7% 49.25 ± 27% sched_debug.cfs_rq[7]:/.tg_runnable_contrib
-9609 ± -7% +10.0% -10571 ± -1% sched_debug.cpu#0.nr_uninterruptible
15.50 ± 79% -91.9% 1.25 ±173% sched_debug.cpu#2.cpu_load[1]
12.75 ± 58% -76.5% 3.00 ±117% sched_debug.cpu#2.cpu_load[2]
11.75 ± 42% -70.2% 3.50 ± 95% sched_debug.cpu#2.cpu_load[3]
11.00 ± 39% -68.2% 3.50 ± 82% sched_debug.cpu#2.cpu_load[4]
851076 ±155% -93.9% 52140 ± 38% sched_debug.cpu#3.nr_switches
1395 ± 4% -8.7% 1274 ± 1% sched_debug.cpu#3.nr_uninterruptible
851137 ±155% -93.9% 52218 ± 38% sched_debug.cpu#3.sched_count
418288 ±157% -94.6% 22436 ± 44% sched_debug.cpu#3.sched_goidle
6.00 ±100% +150.0% 15.00 ± 30% sched_debug.cpu#4.cpu_load[2]
5.25 ± 76% +157.1% 13.50 ± 19% sched_debug.cpu#4.cpu_load[3]
5.25 ± 72% +123.8% 11.75 ± 20% sched_debug.cpu#4.cpu_load[4]
1507 ± 5% +23.3% 1859 ± 5% sched_debug.cpu#5.nr_uninterruptible
811411 ± 8% +10.4% 895772 ± 6% sched_debug.cpu#6.avg_idle
1349 ± 13% +38.2% 1863 ± 3% sched_debug.cpu#6.nr_uninterruptible
=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/iterations/nr_threads/disk/fs/filesize/test_size/sync_method/nr_directories/nr_files_per_directory:
nhm4/fsmark/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performance/1x/32t/1HDD/xfs/8K/400M/fsyncBeforeClose/16d/256fpd
commit:
a4be9881623375fd126762af65ef18dc8175c68d
fa58aafc44805ac425d17c6a8082513b5442ce9d
a4be9881623375fd fa58aafc44805ac425d17c6a80
---------------- --------------------------
%stddev %change %stddev
\ | \
85071 ± 14% -33.4% 56662 ± 0% fsmark.time.involuntary_context_switches
44.50 ± 2% +12.9% 50.25 ± 0% fsmark.time.percent_of_cpu_this_job_got
1173823 ± 2% +25.4% 1472245 ± 6% latency_stats.sum.down.xfs_buf_lock._xfs_buf_find.xfs_buf_get_map.xfs_buf_read_map.xfs_trans_read_buf_map.xfs_read_agi.xfs_ialloc_read_agi.xfs_dialloc.xfs_ialloc.xfs_dir_ialloc.xfs_create
16393 ± 0% +224.5% 53190 ±112% softirqs.TIMER
36.11 ± 2% +148.9% 89.88 ± 94% uptime.boot
178.57 ± 4% +241.2% 609.30 ±111% uptime.idle
178335 ± 0% -80.3% 35149 ± 2% vmstat.system.cs
77433 ± 0% -93.5% 5027 ± 2% vmstat.system.in
28135 ± 3% -12.1% 24722 ± 1% meminfo.Active(anon)
27784 ± 3% -12.3% 24365 ± 1% meminfo.AnonPages
14863 ± 2% -14.8% 12659 ± 2% meminfo.Mapped
6993 ± 3% -11.9% 6160 ± 1% proc-vmstat.nr_active_anon
6906 ± 3% -12.0% 6075 ± 1% proc-vmstat.nr_anon_pages
3703 ± 2% -14.9% 3152 ± 2% proc-vmstat.nr_mapped
85071 ± 14% -33.4% 56662 ± 0% time.involuntary_context_switches
44.50 ± 2% +12.9% 50.25 ± 0% time.percent_of_cpu_this_job_got
5.87 ± 1% +13.5% 6.67 ± 1% time.system_time
10.71 ± 1% -27.2% 7.79 ± 0% turbostat.%Busy
357.25 ± 1% -34.9% 232.50 ± 0% turbostat.Avg_MHz
3333 ± 0% -10.5% 2984 ± 0% turbostat.Bzy_MHz
48.21 ± 5% -23.5% 36.86 ± 4% turbostat.CPU%c1
32.52 ± 5% +22.7% 39.91 ± 5% turbostat.CPU%c3
8.56 ± 11% +80.3% 15.43 ± 5% turbostat.CPU%c6
18315930 ± 4% -46.6% 9777154 ± 8% cpuidle.C1-NHM.time
1153863 ± 2% -94.6% 62163 ± 3% cpuidle.C1-NHM.usage
73216 ± 3% +10.4% 80802 ± 3% cpuidle.C3-NHM.usage
22540985 ± 6% +26.9% 28610584 ± 4% cpuidle.C6-NHM.time
10006 ± 8% +10.7% 11072 ± 3% cpuidle.C6-NHM.usage
43036 ± 99% -98.5% 641.00 ± 24% cpuidle.POLL.time
14491 ±104% -99.6% 51.50 ± 21% cpuidle.POLL.usage
17223 ± 25% -42.3% 9931 ± 35% sched_debug.cfs_rq[0]:/.avg->runnable_avg_sum
2435 ± 2% -10.7% 2174 ± 2% sched_debug.cfs_rq[0]:/.tg->runnable_avg
379.00 ± 25% -42.8% 216.75 ± 34% sched_debug.cfs_rq[0]:/.tg_runnable_contrib
2432 ± 2% -10.7% 2172 ± 2% sched_debug.cfs_rq[1]:/.tg->runnable_avg
12047 ± 12% +26.4% 15233 ± 4% sched_debug.cfs_rq[2]:/.avg->runnable_avg_sum
1122 ± 11% +18.6% 1331 ± 4% sched_debug.cfs_rq[2]:/.min_vruntime
-2608 ± -9% -16.4% -2180 ±-12% sched_debug.cfs_rq[2]:/.spread0
2436 ± 2% -10.9% 2170 ± 2% sched_debug.cfs_rq[2]:/.tg->runnable_avg
262.50 ± 12% +27.0% 333.50 ± 5% sched_debug.cfs_rq[2]:/.tg_runnable_contrib
2435 ± 2% -10.7% 2173 ± 2% sched_debug.cfs_rq[3]:/.tg->runnable_avg
2050 ±120% +731.3% 17041 ± 16% sched_debug.cfs_rq[4]:/.blocked_load_avg
2433 ± 1% -10.3% 2181 ± 2% sched_debug.cfs_rq[4]:/.tg->runnable_avg
2073 ±121% +731.2% 17235 ± 16% sched_debug.cfs_rq[4]:/.tg_load_contrib
1043 ± 19% -35.6% 672.06 ± 20% sched_debug.cfs_rq[5]:/.min_vruntime
2433 ± 1% -10.3% 2184 ± 2% sched_debug.cfs_rq[5]:/.tg->runnable_avg
2433 ± 1% -10.2% 2185 ± 2% sched_debug.cfs_rq[6]:/.tg->runnable_avg
13519 ± 30% -40.0% 8114 ± 35% sched_debug.cfs_rq[7]:/.blocked_load_avg
2429 ± 1% -10.1% 2185 ± 2% sched_debug.cfs_rq[7]:/.tg->runnable_avg
13871 ± 30% -39.9% 8331 ± 35% sched_debug.cfs_rq[7]:/.tg_load_contrib
353549 ± 9% +66.8% 589619 ± 40% sched_debug.cpu#0.avg_idle
21206 ± 3% +253.8% 75034 ±113% sched_debug.cpu#0.clock
21206 ± 3% +253.8% 75034 ±113% sched_debug.cpu#0.clock_task
21207 ± 3% +253.8% 75035 ±113% sched_debug.cpu#1.clock
21207 ± 3% +253.8% 75035 ±113% sched_debug.cpu#1.clock_task
21205 ± 3% +253.9% 75035 ±113% sched_debug.cpu#2.clock
21205 ± 3% +253.9% 75035 ±113% sched_debug.cpu#2.clock_task
5275 ± 21% +95.3% 10300 ± 35% sched_debug.cpu#2.nr_switches
5280 ± 21% +95.4% 10319 ± 35% sched_debug.cpu#2.sched_count
2298 ± 24% +108.5% 4792 ± 37% sched_debug.cpu#2.sched_goidle
2377 ± 31% +96.9% 4680 ± 34% sched_debug.cpu#2.ttwu_count
748.00 ± 47% +284.9% 2879 ± 48% sched_debug.cpu#2.ttwu_local
21208 ± 3% +253.8% 75034 ±113% sched_debug.cpu#3.clock
21208 ± 3% +253.8% 75034 ±113% sched_debug.cpu#3.clock_task
21206 ± 3% +253.8% 75034 ±113% sched_debug.cpu#4.clock
21206 ± 3% +253.8% 75034 ±113% sched_debug.cpu#4.clock_task
73956 ±163% -96.5% 2581 ± 41% sched_debug.cpu#4.nr_switches
73962 ±163% -96.5% 2600 ± 42% sched_debug.cpu#4.sched_count
36498 ±165% -97.4% 950.25 ± 60% sched_debug.cpu#4.sched_goidle
507768 ± 26% +65.5% 840493 ± 20% sched_debug.cpu#5.avg_idle
21207 ± 3% +253.8% 75034 ±113% sched_debug.cpu#5.clock
21207 ± 3% +253.8% 75034 ±113% sched_debug.cpu#5.clock_task
44.75 ± 62% -84.4% 7.00 ± 81% sched_debug.cpu#5.cpu_load[1]
779.25 ± 42% +33.1% 1037 ± 34% sched_debug.cpu#5.nr_load_updates
21207 ± 3% +253.8% 75035 ±113% sched_debug.cpu#6.clock
21207 ± 3% +253.8% 75035 ±113% sched_debug.cpu#6.clock_task
1995 ± 11% +21.6% 2427 ± 17% sched_debug.cpu#6.nr_switches
2001 ± 11% +22.3% 2446 ± 17% sched_debug.cpu#6.sched_count
21206 ± 3% +253.8% 75035 ±113% sched_debug.cpu#7.clock
21206 ± 3% +253.8% 75035 ±113% sched_debug.cpu#7.clock_task
21207 ± 3% +253.8% 75036 ±113% sched_debug.cpu_clk
21049 ± 3% +255.7% 74876 ±113% sched_debug.ktime
21207 ± 3% +253.8% 75036 ±113% sched_debug.sched_clk
=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/nr_threads/iterations/samples:
lituya/ftq/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/powersave/100%/20x/100000ss
commit:
a4be9881623375fd126762af65ef18dc8175c68d
fa58aafc44805ac425d17c6a8082513b5442ce9d
a4be9881623375fd fa58aafc44805ac425d17c6a80
---------------- --------------------------
%stddev %change %stddev
\ | \
7572 ± 0% +6.2% 8040 ± 0% ftq.counts
0.18 ± 2% -82.4% 0.03 ± 8% ftq.stddev
1737203 ± 0% -99.1% 15898 ± 18% ftq.time.involuntary_context_switches
1467 ± 0% +3.4% 1517 ± 0% ftq.time.percent_of_cpu_this_job_got
547.01 ± 0% +3.5% 566.08 ± 0% ftq.time.user_time
16734 ± 0% -13.5% 14475 ± 0% meminfo.Mapped
4181 ± 0% -13.3% 3624 ± 0% proc-vmstat.nr_mapped
1.21 ± 3% -53.7% 0.56 ± 2% turbostat.CPU%c1
5.76 ± 3% +14.6% 6.61 ± 1% turbostat.CPU%c6
97309 ± 0% -96.9% 2991 ± 2% vmstat.system.cs
62011 ± 0% -76.5% 14573 ± 0% vmstat.system.in
1737203 ± 0% -99.1% 15898 ± 18% time.involuntary_context_switches
2.07 ± 6% -47.5% 1.09 ± 2% time.system_time
655.75 ± 36% +55.8% 1021 ± 5% time.voluntary_context_switches
1917711 ± 27% -91.5% 163688 ± 12% cpuidle.C1-HSW.time
144241 ± 3% -99.6% 608.50 ± 13% cpuidle.C1-HSW.usage
13.25 ± 38% -92.5% 1.00 ±100% cpuidle.POLL.time
7.00 ± 30% -92.9% 0.50 ±100% cpuidle.POLL.usage
3330 ± 2% -12.4% 2918 ± 2% sched_debug.cfs_rq[0]:/.tg->runnable_avg
48305 ± 10% +46.6% 70802 ± 3% sched_debug.cfs_rq[0]:/.tg_load_avg
1737 ± 74% +479.4% 10066 ± 59% sched_debug.cfs_rq[10]:/.blocked_load_avg
3330 ± 2% -12.3% 2922 ± 1% sched_debug.cfs_rq[10]:/.tg->runnable_avg
47674 ± 10% +46.5% 69861 ± 3% sched_debug.cfs_rq[10]:/.tg_load_avg
1812 ± 73% +457.2% 10098 ± 59% sched_debug.cfs_rq[10]:/.tg_load_contrib
-4849 ± -2% -17.4% -4006 ±-17% sched_debug.cfs_rq[11]:/.spread0
3330 ± 2% -12.3% 2922 ± 1% sched_debug.cfs_rq[11]:/.tg->runnable_avg
47674 ± 10% +46.5% 69861 ± 3% sched_debug.cfs_rq[11]:/.tg_load_avg
3330 ± 2% -12.0% 2930 ± 1% sched_debug.cfs_rq[12]:/.tg->runnable_avg
47674 ± 10% +46.4% 69806 ± 3% sched_debug.cfs_rq[12]:/.tg_load_avg
3330 ± 2% -12.0% 2930 ± 1% sched_debug.cfs_rq[13]:/.tg->runnable_avg
47674 ± 10% +46.4% 69806 ± 3% sched_debug.cfs_rq[13]:/.tg_load_avg
3338 ± 2% -12.2% 2930 ± 1% sched_debug.cfs_rq[14]:/.tg->runnable_avg
47612 ± 10% +46.6% 69806 ± 3% sched_debug.cfs_rq[14]:/.tg_load_avg
13486 ± 65% -66.1% 4567 ± 44% sched_debug.cfs_rq[15]:/.avg->runnable_avg_sum
3347 ± 2% -12.5% 2930 ± 1% sched_debug.cfs_rq[15]:/.tg->runnable_avg
47536 ± 10% +46.8% 69806 ± 3% sched_debug.cfs_rq[15]:/.tg_load_avg
295.00 ± 67% -66.2% 99.75 ± 45% sched_debug.cfs_rq[15]:/.tg_runnable_contrib
3329 ± 2% -12.4% 2917 ± 2% sched_debug.cfs_rq[1]:/.tg->runnable_avg
48268 ± 10% +46.7% 70802 ± 3% sched_debug.cfs_rq[1]:/.tg_load_avg
611.00 ±164% +895.1% 6080 ± 65% sched_debug.cfs_rq[2]:/.blocked_load_avg
3328 ± 2% -13.0% 2897 ± 1% sched_debug.cfs_rq[2]:/.tg->runnable_avg
48268 ± 10% +45.8% 70372 ± 3% sched_debug.cfs_rq[2]:/.tg_load_avg
611.00 ±164% +961.3% 6484 ± 69% sched_debug.cfs_rq[2]:/.tg_load_contrib
2088 ± 22% -30.6% 1448 ± 2% sched_debug.cfs_rq[3]:/.min_vruntime
3328 ± 2% -13.0% 2897 ± 1% sched_debug.cfs_rq[3]:/.tg->runnable_avg
48268 ± 10% +45.8% 70372 ± 3% sched_debug.cfs_rq[3]:/.tg_load_avg
3330 ± 2% -12.8% 2902 ± 1% sched_debug.cfs_rq[4]:/.tg->runnable_avg
48037 ± 10% +46.3% 70285 ± 3% sched_debug.cfs_rq[4]:/.tg_load_avg
3321 ± 2% -12.5% 2905 ± 1% sched_debug.cfs_rq[5]:/.tg->runnable_avg
48034 ± 10% +46.2% 70241 ± 3% sched_debug.cfs_rq[5]:/.tg_load_avg
5958 ± 58% -79.2% 1239 ± 77% sched_debug.cfs_rq[6]:/.blocked_load_avg
3321 ± 2% -12.4% 2909 ± 1% sched_debug.cfs_rq[6]:/.tg->runnable_avg
48034 ± 10% +46.2% 70222 ± 3% sched_debug.cfs_rq[6]:/.tg_load_avg
6017 ± 57% -79.4% 1239 ± 77% sched_debug.cfs_rq[6]:/.tg_load_contrib
-4384 ±-20% -23.6% -3350 ±-37% sched_debug.cfs_rq[7]:/.spread0
3321 ± 2% -12.4% 2909 ± 1% sched_debug.cfs_rq[7]:/.tg->runnable_avg
47777 ± 10% +47.0% 70222 ± 3% sched_debug.cfs_rq[7]:/.tg_load_avg
6303 ± 42% +54.5% 9736 ± 14% sched_debug.cfs_rq[8]:/.avg->runnable_avg_sum
909.28 ± 36% +91.1% 1738 ± 39% sched_debug.cfs_rq[8]:/.min_vruntime
3330 ± 2% -12.5% 2914 ± 1% sched_debug.cfs_rq[8]:/.tg->runnable_avg
47674 ± 10% +47.2% 70180 ± 3% sched_debug.cfs_rq[8]:/.tg_load_avg
137.75 ± 43% +52.8% 210.50 ± 14% sched_debug.cfs_rq[8]:/.tg_runnable_contrib
99.27 ± 18% +50.2% 149.07 ± 22% sched_debug.cfs_rq[9]:/.exec_clock
3330 ± 2% -12.3% 2922 ± 1% sched_debug.cfs_rq[9]:/.tg->runnable_avg
47674 ± 10% +46.5% 69861 ± 3% sched_debug.cfs_rq[9]:/.tg_load_avg
27.00 ± 43% -55.6% 12.00 ± 5% sched_debug.cpu#0.cpu_load[3]
889905 ± 9% -24.5% 671951 ± 16% sched_debug.cpu#1.avg_idle
10.00 ± 43% -70.0% 3.00 ±102% sched_debug.cpu#1.cpu_load[3]
8.50 ± 52% -79.4% 1.75 ±102% sched_debug.cpu#1.cpu_load[4]
7.75 ± 19% -108.6% -0.67 ±-430% sched_debug.cpu#10.nr_uninterruptible
2398 ± 82% -86.6% 321.25 ± 45% sched_debug.cpu#10.ttwu_count
1835 ± 95% -96.5% 64.00 ± 30% sched_debug.cpu#10.ttwu_local
1368 ± 8% +26.8% 1736 ± 18% sched_debug.cpu#11.nr_switches
1373 ± 8% +26.6% 1738 ± 18% sched_debug.cpu#11.sched_count
509.75 ± 10% +35.9% 693.00 ± 22% sched_debug.cpu#11.sched_goidle
578.00 ± 30% +36.0% 786.25 ± 17% sched_debug.cpu#11.ttwu_count
334.00 ± 36% +115.6% 720.25 ± 16% sched_debug.cpu#13.ttwu_count
588893 ± 42% +64.2% 966897 ± 5% sched_debug.cpu#5.avg_idle
-4.00 ±-68% -118.8% 0.75 ±331% sched_debug.cpu#5.nr_uninterruptible
2.25 ±164% -355.6% -5.75 ±-65% sched_debug.cpu#7.nr_uninterruptible
343.25 ± 42% +86.7% 641.00 ± 43% sched_debug.cpu#9.ttwu_count
=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/nr_threads/iterations/samples:
lituya/fwq/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performance/100%/20x/100000ss
commit:
a4be9881623375fd126762af65ef18dc8175c68d
fa58aafc44805ac425d17c6a8082513b5442ce9d
a4be9881623375fd fa58aafc44805ac425d17c6a80
---------------- --------------------------
%stddev %change %stddev
\ | \
0.11 ± 2% -44.0% 0.06 ± 2% fwq.stddev
3229188 ± 0% -85.9% 455853 ± 11% fwq.time.involuntary_context_switches
13780 ± 1% +5.7% 14566 ± 0% fwq.time.maximum_resident_set_size
176058 ± 20% -31.6% 120345 ± 0% latency_stats.sum.do_wait.SyS_wait4.entry_SYSCALL_64_fastpath
0.30 ± 22% -76.9% 0.07 ± 0% turbostat.CPU%c1
484.63 ± 56% -36.7% 307.00 ± 1% uptime.idle
16399 ± 3% -78.3% 3553 ± 2% vmstat.system.cs
22757 ± 3% -26.2% 16801 ± 0% vmstat.system.in
3907 ± 4% -10.0% 3517 ± 4% slabinfo.anon_vma.active_objs
3907 ± 4% -10.0% 3517 ± 4% slabinfo.anon_vma.num_objs
8215 ± 4% -8.4% 7522 ± 4% slabinfo.kmalloc-512.num_objs
3229188 ± 0% -85.9% 455853 ± 11% time.involuntary_context_switches
1791 ± 3% +10.3% 1976 ± 3% time.minor_page_faults
3.12 ± 1% -72.4% 0.86 ± 6% time.system_time
2392395 ±114% -93.4% 156963 ± 92% cpuidle.C1-HSW.time
48618 ± 12% -99.3% 331.33 ± 14% cpuidle.C1-HSW.usage
2.271e+08 ±130% -82.5% 39690941 ± 3% cpuidle.C6-HSW.time
5212 ± 82% -75.3% 1289 ± 6% cpuidle.C6-HSW.usage
6.25 ± 66% -100.0% 0.00 ± 0% cpuidle.POLL.time
2.75 ± 15% -100.0% 0.00 ± 0% cpuidle.POLL.usage
300.50 ± 49% +108.2% 625.67 ± 33% sched_debug.cfs_rq[0]:/.blocked_load_avg
55.25 ± 4% -7.7% 51.00 ± 4% sched_debug.cfs_rq[0]:/.load
58.75 ± 8% -14.9% 50.00 ± 1% sched_debug.cfs_rq[0]:/.runnable_load_avg
364.25 ± 42% +87.7% 683.67 ± 28% sched_debug.cfs_rq[0]:/.tg_load_contrib
912.75 ± 35% -41.1% 537.33 ± 70% sched_debug.cfs_rq[10]:/.tg_load_contrib
792.33 ± 74% +151.1% 1989 ± 32% sched_debug.cfs_rq[11]:/.blocked_load_avg
659.50 ± 92% +212.5% 2061 ± 30% sched_debug.cfs_rq[11]:/.tg_load_contrib
489.50 ± 59% +158.1% 1263 ± 20% sched_debug.cfs_rq[12]:/.blocked_load_avg
544.75 ± 53% +143.4% 1326 ± 19% sched_debug.cfs_rq[12]:/.tg_load_contrib
98.25 ± 86% +298.3% 391.33 ± 34% sched_debug.cfs_rq[1]:/.blocked_load_avg
157.25 ± 57% +194.9% 463.67 ± 28% sched_debug.cfs_rq[1]:/.tg_load_contrib
324.00 ± 68% +471.2% 1850 ± 22% sched_debug.cfs_rq[3]:/.blocked_load_avg
379.75 ± 60% +402.9% 1909 ± 21% sched_debug.cfs_rq[3]:/.tg_load_contrib
67.75 ± 26% -23.2% 52.00 ± 6% sched_debug.cfs_rq[5]:/.load
1586 ± 85% -64.3% 566.67 ± 46% sched_debug.cfs_rq[6]:/.blocked_load_avg
1679 ± 82% -61.8% 642.00 ± 43% sched_debug.cfs_rq[6]:/.tg_load_contrib
1.25 ± 34% +406.7% 6.33 ± 39% sched_debug.cfs_rq[7]:/.nr_spread_over
59.75 ± 12% -11.3% 53.00 ± 6% sched_debug.cpu#0.cpu_load[0]
55.25 ± 4% -7.7% 51.00 ± 4% sched_debug.cpu#0.load
125050 ± 80% -83.6% 20475 ± 37% sched_debug.cpu#1.nr_switches
1.75 ±240% +776.2% 15.33 ± 13% sched_debug.cpu#1.nr_uninterruptible
125107 ± 80% -83.6% 20557 ± 37% sched_debug.cpu#1.sched_count
54622 ± 93% -78.4% 11825 ± 72% sched_debug.cpu#1.ttwu_count
36441 ± 92% -91.9% 2955 ± 47% sched_debug.cpu#1.ttwu_local
7.75 ± 78% -95.7% 0.33 ±282% sched_debug.cpu#10.nr_uninterruptible
7.25 ± 45% -51.7% 3.50 ± 42% sched_debug.cpu#12.nr_uninterruptible
1584 ± 72% +1888.2% 31493 ±121% sched_debug.cpu#13.ttwu_count
12188 ±141% -91.0% 1100 ± 68% sched_debug.cpu#4.sched_goidle
80.00 ± 15% -27.1% 58.33 ± 7% sched_debug.cpu#6.cpu_load[3]
78.00 ± 16% -27.8% 56.33 ± 4% sched_debug.cpu#6.cpu_load[4]
128000 ±128% -95.1% 6219 ±103% sched_debug.cpu#8.ttwu_count
106189 ±165% -99.7% 357.33 ± 2% sched_debug.cpu#8.ttwu_local
32547 ±143% -89.9% 3291 ± 62% sched_debug.cpu#9.nr_switches
32615 ±143% -89.7% 3352 ± 62% sched_debug.cpu#9.sched_count
26785 ± 79% -85.9% 3781 ± 81% sched_debug.cpu#9.ttwu_count
5.94 ±172% -100.0% 0.00 ± 85% sched_debug.rt_rq[2]:/.rt_time
1.89 ±172% -99.9% 0.00 ± 89% sched_debug.rt_rq[8]:/.rt_time
=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/runtime/nr_threads/cluster/test:
lkp-ne02/netperf/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performance/300s/200%/cs-localhost/SCTP_RR
commit:
a4be9881623375fd126762af65ef18dc8175c68d
fa58aafc44805ac425d17c6a8082513b5442ce9d
a4be9881623375fd fa58aafc44805ac425d17c6a80
---------------- --------------------------
%stddev %change %stddev
\ | \
1233946 ± 1% -93.3% 83018 ± 13% netperf.time.involuntary_context_switches
26623 ±120% -76.8% 6174 ± 63% latency_stats.sum.rpc_wait_bit_killable.__rpc_wait_for_completion_task.nfs4_run_open_task.[nfsv4]._nfs4_open_and_get_state.[nfsv4].nfs4_do_open.[nfsv4].nfs4_atomic_open.[nfsv4].nfs4_file_open.[nfsv4].do_dentry_open.vfs_open.path_openat.do_filp_open.do_sys_open
301360 ± 2% -8.2% 276612 ± 1% softirqs.RCU
1233946 ± 1% -93.3% 83018 ± 13% time.involuntary_context_switches
0.22 ± 12% -27.0% 0.16 ± 13% turbostat.CPU%c1
26675 ± 0% -32.3% 18052 ± 0% vmstat.system.in
9078 ± 5% +12.7% 10235 ± 3% slabinfo.vm_area_struct.active_objs
9128 ± 4% +12.8% 10298 ± 3% slabinfo.vm_area_struct.num_objs
3247591 ± 37% -42.2% 1877235 ± 23% cpuidle.C1-NHM.time
175462 ± 7% -95.2% 8494 ± 6% cpuidle.C1-NHM.usage
863871 ± 11% -100.0% 63.00 ± 8% cpuidle.POLL.time
175219 ± 10% -100.0% 5.50 ± 20% cpuidle.POLL.usage
2731 ±120% +218.4% 8696 ± 3% numa-meminfo.node0.Inactive(anon)
15363 ± 5% +7.2% 16474 ± 4% numa-meminfo.node0.SUnreclaim
6557 ± 50% -91.3% 570.25 ± 49% numa-meminfo.node1.Inactive(anon)
7805 ± 13% -25.5% 5812 ± 0% numa-meminfo.node1.Mapped
18367 ± 2% -7.4% 17002 ± 5% numa-meminfo.node1.SReclaimable
682.50 ±120% +218.5% 2173 ± 3% numa-vmstat.node0.nr_inactive_anon
3840 ± 5% +7.2% 4118 ± 4% numa-vmstat.node0.nr_slab_unreclaimable
1639 ± 50% -91.3% 142.00 ± 49% numa-vmstat.node1.nr_inactive_anon
1950 ± 13% -25.5% 1452 ± 0% numa-vmstat.node1.nr_mapped
4591 ± 2% -7.4% 4250 ± 5% numa-vmstat.node1.nr_slab_reclaimable
1.00 ± 70% +350.0% 4.50 ± 57% sched_debug.cfs_rq[12]:/.nr_spread_over
-1967 ±-3098% +5332.3% -106895 ±-79% sched_debug.cfs_rq[13]:/.spread0
103.00 ± 5% +19.4% 123.00 ± 5% sched_debug.cfs_rq[15]:/.load
95.75 ± 10% +16.7% 111.75 ± 10% sched_debug.cfs_rq[15]:/.runnable_load_avg
-1514 ±-4117% +6467.6% -99452 ±-81% sched_debug.cfs_rq[15]:/.spread0
1116022 ± 6% +11.2% 1240796 ± 4% sched_debug.cfs_rq[2]:/.MIN_vruntime
1116022 ± 6% +11.2% 1240796 ± 4% sched_debug.cfs_rq[2]:/.max_vruntime
1084538 ± 9% +15.0% 1247278 ± 4% sched_debug.cfs_rq[3]:/.MIN_vruntime
1084538 ± 9% +15.0% 1247278 ± 4% sched_debug.cfs_rq[3]:/.max_vruntime
12.25 ± 10% -40.8% 7.25 ± 35% sched_debug.cfs_rq[5]:/.nr_spread_over
-3847 ±-1573% +2484.2% -99431 ±-80% sched_debug.cfs_rq[7]:/.spread0
-2145 ±-139% +451.8% -11836 ±-59% sched_debug.cfs_rq[8]:/.spread0
119.00 ± 7% -23.1% 91.50 ± 12% sched_debug.cpu#0.cpu_load[0]
105.25 ± 3% -15.9% 88.50 ± 6% sched_debug.cpu#0.cpu_load[1]
99.00 ± 4% -13.1% 86.00 ± 5% sched_debug.cpu#0.cpu_load[2]
1480 ± 21% -34.8% 965.50 ± 10% sched_debug.cpu#0.curr->pid
2943 ± 29% +28.1% 3770 ± 1% sched_debug.cpu#0.sched_goidle
1784 ± 77% -75.2% 442.50 ± 7% sched_debug.cpu#13.sched_goidle
88778 ± 53% +58.4% 140611 ± 8% sched_debug.cpu#2.avg_idle
89.00 ± 6% +27.2% 113.25 ± 4% sched_debug.cpu#3.load
2979 ± 51% -67.9% 956.75 ± 19% sched_debug.cpu#7.sched_goidle
1369 ± 29% -26.8% 1002 ± 21% sched_debug.cpu#9.curr->pid
=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/runtime/nr_threads/cluster/test:
lkp-ne02/netperf/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performance/300s/200%/cs-localhost/TCP_SENDFILE
commit:
a4be9881623375fd126762af65ef18dc8175c68d
fa58aafc44805ac425d17c6a8082513b5442ce9d
a4be9881623375fd fa58aafc44805ac425d17c6a80
---------------- --------------------------
%stddev %change %stddev
\ | \
2033310 ± 0% -97.2% 57638 ± 4% netperf.time.involuntary_context_switches
165110 ± 33% +47.2% 243029 ± 20% proc-vmstat.pgalloc_normal
2033310 ± 0% -97.2% 57638 ± 4% time.involuntary_context_switches
3320 ± 7% +12.2% 3725 ± 4% numa-meminfo.node0.KernelStack
2760 ± 8% -15.4% 2335 ± 6% numa-meminfo.node1.KernelStack
78.25 ± 37% -78.3% 17.00 ±139% numa-vmstat.node1.nr_dirtied
76.25 ± 37% -78.4% 16.50 ±138% numa-vmstat.node1.nr_written
0.22 ± 5% -35.6% 0.14 ± 8% turbostat.CPU%c1
0.53 ± 5% +16.0% 0.61 ± 0% turbostat.CPU%c6
49180 ± 0% -48.2% 25479 ± 0% vmstat.system.cs
27651 ± 0% -37.2% 17351 ± 0% vmstat.system.in
4278144 ± 1% -22.0% 3335680 ± 0% latency_stats.hits.sk_wait_data.tcp_recvmsg.inet_recvmsg.sock_recvmsg.SYSC_recvfrom.SyS_recvfrom.entry_SYSCALL_64_fastpath
44720 ± 36% +113.9% 95674 ±108% latency_stats.sum.generic_file_write_iter.nfs_file_write.__vfs_write.vfs_write.SyS_write.entry_SYSCALL_64_fastpath
2.379e+09 ± 0% -1.2% 2.351e+09 ± 0% latency_stats.sum.sk_wait_data.tcp_recvmsg.inet_recvmsg.sock_recvmsg.SYSC_recvfrom.SyS_recvfrom.entry_SYSCALL_64_fastpath
1875754 ± 8% -75.8% 453014 ± 30% cpuidle.C1-NHM.time
149386 ± 1% -95.2% 7207 ± 45% cpuidle.C1-NHM.usage
322.75 ± 51% -65.1% 112.50 ± 16% cpuidle.C1E-NHM.usage
14707 ±165% -99.7% 42.00 ± 41% cpuidle.POLL.time
232.50 ± 82% -98.6% 3.25 ± 45% cpuidle.POLL.usage
106.75 ± 2% -10.5% 95.50 ± 5% sched_debug.cfs_rq[12]:/.load
112.50 ± 7% -15.8% 94.75 ± 5% sched_debug.cfs_rq[12]:/.runnable_load_avg
110.75 ± 3% -12.0% 97.50 ± 15% sched_debug.cfs_rq[13]:/.runnable_load_avg
143.00 ± 98% +628.1% 1041 ± 47% sched_debug.cfs_rq[2]:/.blocked_load_avg
247.75 ± 57% +366.3% 1155 ± 42% sched_debug.cfs_rq[2]:/.tg_load_contrib
42.50 ±158% +665.3% 325.25 ± 68% sched_debug.cfs_rq[3]:/.blocked_load_avg
-75934 ±-49% -129.8% 22591 ± 76% sched_debug.cfs_rq[3]:/.spread0
145.50 ± 46% +199.0% 435.00 ± 51% sched_debug.cfs_rq[3]:/.tg_load_contrib
30.75 ± 31% +74.8% 53.75 ± 25% sched_debug.cfs_rq[4]:/.nr_spread_over
102.75 ± 3% +10.7% 113.75 ± 6% sched_debug.cpu#0.cpu_load[0]
2530 ± 20% +68.6% 4265 ± 23% sched_debug.cpu#0.sched_goidle
350.25 ± 10% +81.1% 634.25 ± 24% sched_debug.cpu#10.sched_goidle
112.00 ± 7% -14.7% 95.50 ± 8% sched_debug.cpu#12.cpu_load[0]
110.25 ± 6% -11.3% 97.75 ± 5% sched_debug.cpu#12.cpu_load[1]
109.50 ± 4% -8.9% 99.75 ± 4% sched_debug.cpu#12.cpu_load[2]
110.75 ± 4% -12.0% 97.50 ± 11% sched_debug.cpu#13.cpu_load[1]
111.25 ± 4% -11.9% 98.00 ± 10% sched_debug.cpu#13.cpu_load[2]
624.00 ± 23% -32.6% 420.50 ± 16% sched_debug.cpu#13.sched_goidle
947672 ± 76% -77.0% 217667 ± 3% sched_debug.cpu#15.nr_switches
947705 ± 76% -77.0% 217685 ± 3% sched_debug.cpu#15.sched_count
592433 ± 65% -66.0% 201467 ± 1% sched_debug.cpu#15.ttwu_local
911.50 ± 8% +9.5% 998.50 ± 11% sched_debug.cpu#3.curr->pid
562814 ± 32% +41.8% 798162 ± 15% sched_debug.cpu#4.avg_idle
277723 ± 12% -17.9% 227889 ± 5% sched_debug.cpu#4.nr_switches
277747 ± 12% -17.9% 227907 ± 5% sched_debug.cpu#4.sched_count
109.00 ± 2% -11.0% 97.00 ± 0% sched_debug.cpu#5.load
=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/nr_threads/iterations/entries:
snb-drag/tlbflush/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performance/200%/32x/512
commit:
a4be9881623375fd126762af65ef18dc8175c68d
fa58aafc44805ac425d17c6a8082513b5442ce9d
a4be9881623375fd fa58aafc44805ac425d17c6a80
---------------- --------------------------
%stddev %change %stddev
\ | \
89406 ± 1% +2.1% 91264 ± 0% tlbflush.mem_acc_time_thread_ms
12692 ± 29% -69.7% 3848 ± 6% tlbflush.time.involuntary_context_switches
45262 ± 14% -20.5% 35996 ± 13% softirqs.SCHED
12692 ± 29% -69.7% 3848 ± 6% time.involuntary_context_switches
5023 ± 14% +24.8% 6271 ± 4% slabinfo.kmalloc-32.active_objs
5023 ± 14% +24.8% 6271 ± 4% slabinfo.kmalloc-32.num_objs
62647 ± 4% -20.7% 49700 ± 4% vmstat.system.cs
26516 ± 6% -24.7% 19964 ± 3% vmstat.system.in
1.486e+08 ± 4% -25.5% 1.108e+08 ± 7% cpuidle.C1-SNB.time
9489652 ± 0% -45.4% 5183155 ± 1% cpuidle.C1-SNB.usage
94983 ± 14% -35.2% 61571 ± 0% cpuidle.POLL.usage
424061 ± 57% -100.0% 0.00 ± -1% latency_stats.avg.rpc_wait_bit_killable.__rpc_wait_for_completion_task.nfs4_run_open_task.[nfsv4]._nfs4_open_and_get_state.[nfsv4].nfs4_do_open.[nfsv4].nfs4_atomic_open.[nfsv4].nfs_atomic_open.path_openat.do_filp_open.do_sys_open.SyS_open.entry_SYSCALL_64_fastpath
424061 ± 57% -100.0% 0.00 ± -1% latency_stats.max.rpc_wait_bit_killable.__rpc_wait_for_completion_task.nfs4_run_open_task.[nfsv4]._nfs4_open_and_get_state.[nfsv4].nfs4_do_open.[nfsv4].nfs4_atomic_open.[nfsv4].nfs_atomic_open.path_openat.do_filp_open.do_sys_open.SyS_open.entry_SYSCALL_64_fastpath
424061 ± 57% -100.0% 0.00 ± -1% latency_stats.sum.rpc_wait_bit_killable.__rpc_wait_for_completion_task.nfs4_run_open_task.[nfsv4]._nfs4_open_and_get_state.[nfsv4].nfs4_do_open.[nfsv4].nfs4_atomic_open.[nfsv4].nfs_atomic_open.path_openat.do_filp_open.do_sys_open.SyS_open.entry_SYSCALL_64_fastpath
4767 ± 56% -94.4% 268.00 ± 30% sched_debug.cfs_rq[0]:/.load
17.25 ± 58% +127.5% 39.25 ± 19% sched_debug.cfs_rq[3]:/.runnable_load_avg
8521655 ± 15% -42.7% 4882199 ± 15% sched_debug.cpu#0.nr_switches
8522183 ± 15% -42.7% 4882721 ± 15% sched_debug.cpu#0.sched_count
4225538 ± 15% -42.7% 2421794 ± 15% sched_debug.cpu#0.sched_goidle
4280766 ± 15% -41.3% 2511288 ± 15% sched_debug.cpu#0.ttwu_count
3693688 ± 17% -48.0% 1919886 ± 18% sched_debug.cpu#0.ttwu_local
10474544 ± 12% -43.9% 5872222 ± 9% sched_debug.cpu#1.nr_switches
10474799 ± 12% -43.9% 5872473 ± 9% sched_debug.cpu#1.sched_count
5198778 ± 12% -43.9% 2917524 ± 9% sched_debug.cpu#1.sched_goidle
5265913 ± 12% -44.2% 2940722 ± 9% sched_debug.cpu#1.ttwu_count
4654824 ± 14% -48.5% 2396748 ± 10% sched_debug.cpu#1.ttwu_local
6.50 ± 50% +138.5% 15.50 ± 25% sched_debug.cpu#3.cpu_load[1]
5.25 ± 28% +114.3% 11.25 ± 25% sched_debug.cpu#3.cpu_load[2]
=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/test:
nhm-white/unixbench/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/shell8
commit:
a4be9881623375fd126762af65ef18dc8175c68d
fa58aafc44805ac425d17c6a8082513b5442ce9d
a4be9881623375fd fa58aafc44805ac425d17c6a80
---------------- --------------------------
%stddev %change %stddev
\ | \
5937622 ± 1% -26.4% 4369228 ± 0% unixbench.time.involuntary_context_switches
51200 ± 3% +15.0% 58880 ± 4% meminfo.DirectMap4k
5937622 ± 1% -26.4% 4369228 ± 0% time.involuntary_context_switches
62839 ± 1% -28.4% 44966 ± 2% vmstat.system.cs
19289 ± 2% -46.2% 10378 ± 2% vmstat.system.in
6.09 ± 0% +8.9% 6.63 ± 2% turbostat.CPU%c3
0.45 ± 10% +50.6% 0.67 ± 20% turbostat.Pkg%pc3
3.17 ± 1% +62.5% 5.15 ± 47% turbostat.Pkg%pc6
45216499 ± 0% -34.6% 29566388 ± 1% cpuidle.C1-NHM.time
1918738 ± 7% -88.4% 222808 ± 2% cpuidle.C1-NHM.usage
220032 ± 10% -90.2% 21647 ± 11% cpuidle.POLL.time
30597 ± 3% -96.6% 1051 ± 2% cpuidle.POLL.usage
1886 ± 45% +73.6% 3275 ± 36% sched_debug.cfs_rq[4]:/.utilization_load_avg
294740 ± 5% +42.0% 418454 ± 8% sched_debug.cpu#1.avg_idle
2624072 ± 63% -60.8% 1029438 ± 2% sched_debug.cpu#1.nr_switches
2624725 ± 63% -60.8% 1030184 ± 2% sched_debug.cpu#1.sched_count
1203729 ± 70% -66.9% 398300 ± 1% sched_debug.cpu#1.ttwu_count
992043 ± 86% -81.8% 180660 ± 3% sched_debug.cpu#1.ttwu_local
15179 ± 13% -43.3% 8606 ± 20% sched_debug.cpu#2.curr->pid
-204.00 ±-22% -47.5% -107.00 ±-43% sched_debug.cpu#2.nr_uninterruptible
184.75 ± 28% -28.1% 132.75 ± 13% sched_debug.cpu#5.nr_uninterruptible
14010 ± 11% -20.8% 11095 ± 16% sched_debug.cpu#7.curr->pid
2209845 ± 57% -56.4% 962613 ± 3% sched_debug.cpu#7.nr_switches
2210474 ± 57% -56.4% 963302 ± 3% sched_debug.cpu#7.sched_count
575333 ± 61% -58.4% 239461 ± 1% sched_debug.cpu#7.sched_goidle
7.45 ±124% -99.9% 0.01 ± 3% sched_debug.rt_rq[5]:/.rt_time
lkp-a06: Atom
Memory: 8G
nhm4: Nehalem
Memory: 4G
lituya: Grantley Haswell
Memory: 16G
lkp-ne02: Nehalem-EP
Memory: 5G
snb-drag: Sandy Bridge
Memory: 6G
nhm-white: Nehalem
Memory: 6G
aim7.time.voluntary_context_switches
58000 ++------------------------------------------------------------------+
| * |
56000 *+ .. + .* .* |
54000 ++*..*.*.*.. .* *.*..* + .* .*.*..*.*. .*. .* + .*.*
| *.* *. + .* *..* *. *. |
52000 ++ *. |
50000 ++ |
| |
48000 ++ |
46000 ++ |
| |
44000 O+ O O O |
42000 ++O O O O O |
| O O O O O O |
40000 ++------------------------------------------------------------------+
aim7.time.involuntary_context_switches
2.2e+06 ++----------------------------------------------------------------+
2e+06 *+*..*.*.*.*.. .*. .*.*.*..*.*.*.*..*.*.*..*.*.*.*..*.*.*.*..*.*
| *.* *. |
1.8e+06 ++ |
1.6e+06 ++ |
1.4e+06 ++ |
1.2e+06 ++ |
| |
1e+06 ++ |
800000 ++ |
600000 ++ |
400000 ++ |
| |
200000 O+O O O O O O O O O O O O O O |
0 ++----------------------------------------------------------------+
[*] bisect-good sample
[O] bisect-bad sample
To reproduce:
git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
cd lkp-tests
bin/lkp install job.yaml # job file is attached in this email
bin/lkp run job.yaml
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
Thanks,
Ying Huang
---
LKP_SERVER: inn
LKP_CGI_PORT: 80
LKP_CIFS_PORT: 139
default-watchdogs:
oom-killer:
watchdog:
commit: 4334e59ef28396425833e152edf40c40e58ccb84
model: Atom
memory: 8G
hdd_partitions: "/dev/disk/by-id/ata-TOSHIBA_MK8061GSYB_22FHP0NGT-part1"
swap_partitions:
rootfs_partition: "/dev/disk/by-id/ata-TOSHIBA_MK8061GSYB_22FHP0NGT-part2"
testcase: aim7
aim7:
load: 4000
test: pipe_cpy
queue: cyclic
testbox: lkp-a06
tbox_group: lkp-a06
kconfig: x86_64-rhel
enqueue_time: 2015-08-13 15:19:11.193423027 +08:00
id: 34aa50e279ab0f513acb7f6abd3a572fe64787a7
user: lkp
compiler: gcc-4.9
head_commit: 4334e59ef28396425833e152edf40c40e58ccb84
base_commit: f7644cbfcdf03528f0f450f3940c4985b2291f49
branch: linux-devel/devel-hourly-2015081302
kernel: "/pkg/linux/x86_64-rhel/gcc-4.9/4334e59ef28396425833e152edf40c40e58ccb84/vmlinuz-4.2.0-rc6-wl-ath-03840-g4334e59"
rootfs: debian-x86_64-2015-02-07.cgz
result_root: "/result/aim7/4000-pipe_cpy/lkp-a06/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/4334e59ef28396425833e152edf40c40e58ccb84/0"
job_file: "/lkp/scheduled/lkp-a06/cyclic_aim7-4000-pipe_cpy-x86_64-rhel-CYCLIC_HEAD-4334e59ef28396425833e152edf40c40e58ccb84-20150813-16309-gx9rp8-0.yaml"
dequeue_time: 2015-08-13 16:29:29.826691265 +08:00
nr_cpu: "$(nproc)"
max_uptime: 3600
initrd: "/osimage/debian/debian-x86_64-2015-02-07.cgz"
bootloader_append:
- root=/dev/ram0
- user=lkp
- job=/lkp/scheduled/lkp-a06/cyclic_aim7-4000-pipe_cpy-x86_64-rhel-CYCLIC_HEAD-4334e59ef28396425833e152edf40c40e58ccb84-20150813-16309-gx9rp8-0.yaml
- ARCH=x86_64
- kconfig=x86_64-rhel
- branch=linux-devel/devel-hourly-2015081302
- commit=4334e59ef28396425833e152edf40c40e58ccb84
- BOOT_IMAGE=/pkg/linux/x86_64-rhel/gcc-4.9/4334e59ef28396425833e152edf40c40e58ccb84/vmlinuz-4.2.0-rc6-wl-ath-03840-g4334e59
- max_uptime=3600
- RESULT_ROOT=/result/aim7/4000-pipe_cpy/lkp-a06/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/4334e59ef28396425833e152edf40c40e58ccb84/0
- LKP_SERVER=inn
- |2-
earlyprintk=ttyS0,115200 systemd.log_level=err
debug apic=debug sysrq_always_enabled rcupdate.rcu_cpu_stall_timeout=100
panic=-1 softlockup_panic=1 nmi_watchdog=panic oops=panic load_ramdisk=2 prompt_ramdisk=0
console=ttyS0,115200 console=tty0 vga=normal
rw
lkp_initrd: "/lkp/lkp/lkp-x86_64.cgz"
modules_initrd: "/pkg/linux/x86_64-rhel/gcc-4.9/4334e59ef28396425833e152edf40c40e58ccb84/modules.cgz"
bm_initrd: "/osimage/deps/debian-x86_64-2015-02-07.cgz/lkp.cgz,/osimage/deps/debian-x86_64-2015-02-07.cgz/run-ipconfig.cgz,/lkp/benchmarks/aim7-x86_64.cgz"
job_state: finished
loadavg: 2048.94 721.78 257.37 1/131 4538
start_time: '1439454675'
end_time: '1439454764'
version: "/lkp/lkp/.src-20150813-001640"
echo performance > /sys/devices/system/cpu/cpu0/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu1/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu10/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu11/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu12/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu13/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu14/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu15/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu2/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu3/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu4/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu5/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu6/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu7/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu8/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu9/cpufreq/scaling_governor
netserver
netperf -t TCP_SENDFILE -c -C -l 300 -H 127.0.0.1
netperf -t TCP_SENDFILE -c -C -l 300 -H 127.0.0.1
netperf -t TCP_SENDFILE -c -C -l 300 -H 127.0.0.1
netperf -t TCP_SENDFILE -c -C -l 300 -H 127.0.0.1
netperf -t TCP_SENDFILE -c -C -l 300 -H 127.0.0.1
netperf -t TCP_SENDFILE -c -C -l 300 -H 127.0.0.1
netperf -t TCP_SENDFILE -c -C -l 300 -H 127.0.0.1
netperf -t TCP_SENDFILE -c -C -l 300 -H 127.0.0.1
netperf -t TCP_SENDFILE -c -C -l 300 -H 127.0.0.1
netperf -t TCP_SENDFILE -c -C -l 300 -H 127.0.0.1
netperf -t TCP_SENDFILE -c -C -l 300 -H 127.0.0.1
netperf -t TCP_SENDFILE -c -C -l 300 -H 127.0.0.1
netperf -t TCP_SENDFILE -c -C -l 300 -H 127.0.0.1
netperf -t TCP_SENDFILE -c -C -l 300 -H 127.0.0.1
netperf -t TCP_SENDFILE -c -C -l 300 -H 127.0.0.1
netperf -t TCP_SENDFILE -c -C -l 300 -H 127.0.0.1
netperf -t TCP_SENDFILE -c -C -l 300 -H 127.0.0.1
netperf -t TCP_SENDFILE -c -C -l 300 -H 127.0.0.1
netperf -t TCP_SENDFILE -c -C -l 300 -H 127.0.0.1
netperf -t TCP_SENDFILE -c -C -l 300 -H 127.0.0.1
netperf -t TCP_SENDFILE -c -C -l 300 -H 127.0.0.1
netperf -t TCP_SENDFILE -c -C -l 300 -H 127.0.0.1
netperf -t TCP_SENDFILE -c -C -l 300 -H 127.0.0.1
netperf -t TCP_SENDFILE -c -C -l 300 -H 127.0.0.1
netperf -t TCP_SENDFILE -c -C -l 300 -H 127.0.0.1
netperf -t TCP_SENDFILE -c -C -l 300 -H 127.0.0.1
netperf -t TCP_SENDFILE -c -C -l 300 -H 127.0.0.1
netperf -t TCP_SENDFILE -c -C -l 300 -H 127.0.0.1
netperf -t TCP_SENDFILE -c -C -l 300 -H 127.0.0.1
netperf -t TCP_SENDFILE -c -C -l 300 -H 127.0.0.1
netperf -t TCP_SENDFILE -c -C -l 300 -H 127.0.0.1
netperf -t TCP_SENDFILE -c -C -l 300 -H 127.0.0.1