[lkp] [nfs] 048883e0b9: No primary result change, -70.4% fsmark.time.involuntary_context_switches
From: kernel test robot
Date: Wed Oct 07 2015 - 23:25:20 EST
FYI, we noticed the below changes on
https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git master
commit 048883e0b934d9a5103d40e209cb14b7f33d2933 ("nfs: fix pg_test page count calculation")
=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/iterations/nr_threads/disk/fs/fs2/filesize/test_size/sync_method/nr_directories/nr_files_per_directory:
lkp-ws02/fsmark/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/1x/32t/1HDD/xfs/nfsv4/16MB/60G/fsyncBeforeClose/16d/256fpd
commit:
a41cbe86df3afbc82311a1640e20858c0cd7e065
048883e0b934d9a5103d40e209cb14b7f33d2933
a41cbe86df3afbc8 048883e0b934d9a5103d40e209
---------------- --------------------------
%stddev %change %stddev
\ | \
261986 ± 0% -70.4% 77543 ± 0% fsmark.time.involuntary_context_switches
272406 ± 0% -17.9% 223687 ± 0% fsmark.time.voluntary_context_switches
5443 ± 0% -38.6% 3342 ± 0% vmstat.system.cs
475248 ± 0% -50.9% 233285 ± 0% softirqs.NET_RX
157367 ± 1% -9.0% 143212 ± 0% softirqs.SCHED
261986 ± 0% -70.4% 77543 ± 0% time.involuntary_context_switches
272406 ± 0% -17.9% 223687 ± 0% time.voluntary_context_switches
248624 ± 1% +294.3% 980340 ± 0% meminfo.Active
223111 ± 2% +328.0% 954877 ± 0% meminfo.Active(file)
65657 ± 0% -13.1% 57050 ± 0% meminfo.SUnreclaim
1.34 ± 0% -5.2% 1.27 ± 0% turbostat.%Busy
5.19 ± 1% -28.5% 3.71 ± 3% turbostat.CPU%c1
12.41 ± 1% -52.3% 5.91 ± 4% turbostat.CPU%c3
14.86 ± 1% -23.3% 11.39 ± 3% turbostat.Pkg%pc3
16.35 ± 1% +41.8% 23.19 ± 1% turbostat.Pkg%pc6
2.675e+08 ± 4% -12.9% 2.329e+08 ± 7% cpuidle.C1-NHM.time
165684 ± 4% +12.4% 186205 ± 2% cpuidle.C1-NHM.usage
1.446e+08 ± 7% -75.9% 34785128 ± 12% cpuidle.C1E-NHM.time
79076 ± 1% -87.7% 9744 ± 3% cpuidle.C1E-NHM.usage
1.618e+09 ± 1% -55.5% 7.193e+08 ± 3% cpuidle.C3-NHM.time
510548 ± 0% -73.4% 135726 ± 1% cpuidle.C3-NHM.usage
1532641 ± 0% -9.8% 1382714 ± 1% cpuidle.C6-NHM.usage
119890 ± 2% +322.8% 506915 ± 5% numa-meminfo.node0.Active
107993 ± 2% +356.4% 492854 ± 5% numa-meminfo.node0.Active(file)
34025 ± 3% -9.9% 30670 ± 1% numa-meminfo.node0.SUnreclaim
128802 ± 4% +267.7% 473544 ± 6% numa-meminfo.node1.Active
115176 ± 4% +301.2% 462134 ± 6% numa-meminfo.node1.Active(file)
1217 ± 4% -20.5% 967.25 ± 22% numa-meminfo.node1.Dirty
9663 ± 24% -29.4% 6824 ± 35% numa-meminfo.node1.Mapped
31637 ± 3% -16.6% 26381 ± 1% numa-meminfo.node1.SUnreclaim
11631847 ± 2% +19.8% 13937484 ± 4% numa-numastat.node0.local_node
3950957 ± 9% +58.3% 6253584 ± 10% numa-numastat.node0.numa_foreign
11631855 ± 2% +19.8% 13937495 ± 4% numa-numastat.node0.numa_hit
4660337 ± 3% -27.1% 3398872 ± 7% numa-numastat.node0.numa_miss
4660345 ± 3% -27.1% 3398883 ± 7% numa-numastat.node0.other_node
13541675 ± 3% -24.6% 10208933 ± 9% numa-numastat.node1.local_node
4660333 ± 3% -27.1% 3398872 ± 7% numa-numastat.node1.numa_foreign
13541683 ± 3% -24.6% 10208939 ± 9% numa-numastat.node1.numa_hit
3950957 ± 9% +58.3% 6253604 ± 10% numa-numastat.node1.numa_miss
3950964 ± 9% +58.3% 6253611 ± 10% numa-numastat.node1.other_node
55777 ± 2% +328.0% 238719 ± 0% proc-vmstat.nr_active_file
16414 ± 0% -13.1% 14262 ± 0% proc-vmstat.nr_slab_unreclaimable
8605572 ± 5% +12.1% 9648367 ± 6% proc-vmstat.numa_foreign
8605590 ± 5% +12.1% 9648366 ± 6% proc-vmstat.numa_miss
8605606 ± 5% +12.1% 9648384 ± 6% proc-vmstat.numa_other
1080 ± 8% -14.5% 924.25 ± 5% proc-vmstat.numa_pages_migrated
68361 ± 5% +620.8% 492764 ± 2% proc-vmstat.pgactivate
1080 ± 8% -14.5% 924.25 ± 5% proc-vmstat.pgmigrate_success
4917430 ± 1% +8.5% 5336634 ± 2% proc-vmstat.pgscan_kswapd_dma32
2245024 ± 0% -70.9% 653056 ± 0% proc-vmstat.slabs_scanned
26997 ± 2% +356.4% 123206 ± 5% numa-vmstat.node0.nr_active_file
8505 ± 3% -9.9% 7667 ± 1% numa-vmstat.node0.nr_slab_unreclaimable
1556010 ± 11% +72.7% 2687972 ± 9% numa-vmstat.node0.numa_foreign
5466532 ± 2% +17.2% 6404739 ± 5% numa-vmstat.node0.numa_hit
5466271 ± 2% +17.2% 6404584 ± 5% numa-vmstat.node0.numa_local
2073926 ± 8% -27.8% 1497829 ± 15% numa-vmstat.node0.numa_miss
2074187 ± 8% -27.8% 1497984 ± 15% numa-vmstat.node0.numa_other
28794 ± 4% +301.2% 115526 ± 6% numa-vmstat.node1.nr_active_file
300.50 ± 4% -27.2% 218.75 ± 17% numa-vmstat.node1.nr_dirty
2415 ± 24% -29.4% 1705 ± 35% numa-vmstat.node1.nr_mapped
7909 ± 3% -16.6% 6595 ± 1% numa-vmstat.node1.nr_slab_unreclaimable
2073932 ± 8% -27.8% 1497832 ± 15% numa-vmstat.node1.numa_foreign
6395016 ± 3% -23.4% 4896548 ± 6% numa-vmstat.node1.numa_hit
6330567 ± 3% -23.7% 4832140 ± 6% numa-vmstat.node1.numa_local
1556010 ± 11% +72.7% 2687974 ± 9% numa-vmstat.node1.numa_miss
1620459 ± 10% +69.9% 2752382 ± 8% numa-vmstat.node1.numa_other
4543 ± 2% -78.3% 986.25 ± 1% slabinfo.RAW.active_objs
129.25 ± 2% -79.3% 26.75 ± 1% slabinfo.RAW.active_slabs
4672 ± 2% -78.9% 986.25 ± 1% slabinfo.RAW.num_objs
129.25 ± 2% -79.3% 26.75 ± 1% slabinfo.RAW.num_slabs
55723 ± 0% -40.0% 33420 ± 0% slabinfo.kmalloc-128.active_objs
1768 ± 0% -39.6% 1067 ± 1% slabinfo.kmalloc-128.active_slabs
56593 ± 0% -39.6% 34163 ± 1% slabinfo.kmalloc-128.num_objs
1768 ± 0% -39.6% 1067 ± 1% slabinfo.kmalloc-128.num_slabs
12134 ± 1% -7.9% 11175 ± 3% slabinfo.kmalloc-192.active_objs
12367 ± 1% -9.0% 11257 ± 2% slabinfo.kmalloc-192.num_objs
14574 ± 1% -32.2% 9875 ± 3% slabinfo.kmalloc-32.active_objs
14574 ± 1% -32.2% 9875 ± 3% slabinfo.kmalloc-32.num_objs
35364 ± 0% -62.7% 13198 ± 2% slabinfo.kmalloc-96.active_objs
854.00 ± 0% -62.5% 320.25 ± 1% slabinfo.kmalloc-96.active_slabs
35880 ± 0% -62.5% 13469 ± 1% slabinfo.kmalloc-96.num_objs
854.00 ± 0% -62.5% 320.25 ± 1% slabinfo.kmalloc-96.num_slabs
8.25 ± 69% +524.2% 51.50 ± 42% sched_debug.cfs_rq[0]:/.load
9290 ± 2% -28.3% 6664 ± 3% sched_debug.cfs_rq[10]:/.exec_clock
267277 ± 2% -9.4% 242184 ± 4% sched_debug.cfs_rq[10]:/.min_vruntime
53.25 ± 6% -24.4% 40.25 ± 2% sched_debug.cfs_rq[10]:/.nr_spread_over
-9975 ±-167% +515.1% -61355 ±-28% sched_debug.cfs_rq[10]:/.spread0
9375 ± 4% -27.8% 6766 ± 4% sched_debug.cfs_rq[11]:/.exec_clock
54.00 ± 6% -28.2% 38.75 ± 9% sched_debug.cfs_rq[11]:/.nr_spread_over
-26030 ±-80% +151.9% -65565 ±-33% sched_debug.cfs_rq[11]:/.spread0
47694 ± 8% +25.3% 59751 ± 6% sched_debug.cfs_rq[12]:/.min_vruntime
18.50 ± 14% +41.9% 26.25 ± 11% sched_debug.cfs_rq[12]:/.nr_spread_over
2556 ± 5% +62.8% 4160 ± 15% sched_debug.cfs_rq[13]:/.exec_clock
43878 ± 8% +51.8% 66592 ± 20% sched_debug.cfs_rq[13]:/.min_vruntime
2813 ± 11% +27.2% 3579 ± 5% sched_debug.cfs_rq[14]:/.exec_clock
7.25 ±173% +331.0% 31.25 ± 73% sched_debug.cfs_rq[14]:/.load
43638 ± 15% +45.7% 63601 ± 18% sched_debug.cfs_rq[14]:/.min_vruntime
5.00 ± 37% +135.0% 11.75 ± 32% sched_debug.cfs_rq[14]:/.util_avg
2665 ± 5% +39.2% 3711 ± 11% sched_debug.cfs_rq[15]:/.exec_clock
40446 ± 14% +34.6% 54455 ± 11% sched_debug.cfs_rq[15]:/.min_vruntime
2510 ± 3% +50.5% 3778 ± 9% sched_debug.cfs_rq[16]:/.exec_clock
14.00 ± 15% +69.6% 23.75 ± 9% sched_debug.cfs_rq[16]:/.nr_spread_over
-218464 ± -6% +10.4% -241264 ± -4% sched_debug.cfs_rq[16]:/.spread0
2750 ± 7% +24.8% 3432 ± 2% sched_debug.cfs_rq[17]:/.exec_clock
3203 ± 5% -28.9% 2278 ± 7% sched_debug.cfs_rq[18]:/.exec_clock
59639 ± 4% -20.3% 47534 ± 6% sched_debug.cfs_rq[18]:/.min_vruntime
-217617 ± -5% +17.6% -256020 ± -5% sched_debug.cfs_rq[18]:/.spread0
2975 ± 5% -22.7% 2301 ± 12% sched_debug.cfs_rq[19]:/.exec_clock
-223305 ± -7% +15.8% -258545 ± -6% sched_debug.cfs_rq[19]:/.spread0
8007 ± 3% +14.1% 9134 ± 4% sched_debug.cfs_rq[1]:/.exec_clock
3094 ± 7% -24.1% 2349 ± 2% sched_debug.cfs_rq[20]:/.exec_clock
54684 ± 15% -25.7% 40653 ± 14% sched_debug.cfs_rq[20]:/.min_vruntime
-222572 ± -7% +18.1% -262902 ± -5% sched_debug.cfs_rq[20]:/.spread0
3057 ± 4% -23.0% 2354 ± 10% sched_debug.cfs_rq[21]:/.exec_clock
2997 ± 4% -23.3% 2298 ± 9% sched_debug.cfs_rq[22]:/.exec_clock
50613 ± 9% -18.3% 41365 ± 16% sched_debug.cfs_rq[22]:/.min_vruntime
19.75 ± 6% -21.5% 15.50 ± 17% sched_debug.cfs_rq[22]:/.nr_spread_over
-226643 ± -4% +15.7% -262191 ± -5% sched_debug.cfs_rq[22]:/.spread0
23.25 ± 47% -72.0% 6.50 ± 65% sched_debug.cfs_rq[22]:/.util_avg
-220864 ± -5% +15.5% -255074 ± -6% sched_debug.cfs_rq[23]:/.spread0
-28540 ±-34% +95.8% -55873 ±-18% sched_debug.cfs_rq[2]:/.spread0
47.25 ± 6% +24.3% 58.75 ± 6% sched_debug.cfs_rq[5]:/.nr_spread_over
9369 ± 1% -23.3% 7187 ± 3% sched_debug.cfs_rq[6]:/.exec_clock
63.25 ± 9% -18.6% 51.50 ± 13% sched_debug.cfs_rq[6]:/.nr_spread_over
9502 ± 2% -26.7% 6965 ± 4% sched_debug.cfs_rq[7]:/.exec_clock
58.25 ± 8% -20.6% 46.25 ± 4% sched_debug.cfs_rq[7]:/.nr_spread_over
-33020 ±-61% +117.1% -71698 ±-19% sched_debug.cfs_rq[7]:/.spread0
9332 ± 3% -26.4% 6864 ± 3% sched_debug.cfs_rq[8]:/.exec_clock
254022 ± 2% -9.0% 231037 ± 1% sched_debug.cfs_rq[8]:/.min_vruntime
-23229 ±-73% +212.1% -72492 ±-19% sched_debug.cfs_rq[8]:/.spread0
9519 ± 3% -26.4% 7007 ± 3% sched_debug.cfs_rq[9]:/.exec_clock
-10930 ±-196% +425.0% -57389 ±-11% sched_debug.cfs_rq[9]:/.spread0
175146 ± 5% -28.7% 124945 ± 5% sched_debug.cpu#0.nr_switches
138.75 ± 21% -32.4% 93.75 ± 28% sched_debug.cpu#0.nr_uninterruptible
65351 ± 7% -20.6% 51877 ± 7% sched_debug.cpu#0.sched_goidle
454109 ± 3% +30.2% 591454 ± 4% sched_debug.cpu#0.ttwu_count
36456 ± 1% -35.0% 23705 ± 5% sched_debug.cpu#0.ttwu_local
159152 ± 5% -26.4% 117194 ± 5% sched_debug.cpu#1.nr_switches
60806 ± 6% -18.8% 49354 ± 5% sched_debug.cpu#1.sched_goidle
31121 ± 5% -35.6% 20032 ± 10% sched_debug.cpu#1.ttwu_local
38223 ± 4% -10.9% 34062 ± 5% sched_debug.cpu#10.nr_load_updates
165227 ± 3% -43.8% 92896 ± 3% sched_debug.cpu#10.nr_switches
257578 ± 4% -33.7% 170890 ± 6% sched_debug.cpu#10.sched_count
59662 ± 4% -33.9% 39423 ± 4% sched_debug.cpu#10.sched_goidle
161422 ± 6% -34.7% 105365 ± 5% sched_debug.cpu#10.ttwu_count
34048 ± 4% -57.5% 14483 ± 11% sched_debug.cpu#10.ttwu_local
39424 ± 7% -16.4% 32956 ± 4% sched_debug.cpu#11.nr_load_updates
172939 ± 10% -47.9% 90094 ± 6% sched_debug.cpu#11.nr_switches
273257 ± 7% -37.4% 171120 ± 6% sched_debug.cpu#11.sched_count
63637 ± 13% -40.3% 38022 ± 8% sched_debug.cpu#11.sched_goidle
155856 ± 5% -32.7% 104830 ± 5% sched_debug.cpu#11.ttwu_count
34928 ± 9% -61.4% 13465 ± 5% sched_debug.cpu#11.ttwu_local
0.50 ±100% +550.0% 3.25 ± 50% sched_debug.cpu#12.cpu_load[0]
0.33 ±141% +2825.0% 9.75 ± 46% sched_debug.cpu#12.cpu_load[1]
0.00 ± 0% +Inf% 10.75 ± 53% sched_debug.cpu#12.cpu_load[2]
34474 ± 6% -20.2% 27520 ± 4% sched_debug.cpu#12.nr_switches
-6.25 ±-44% +292.0% -24.50 ±-19% sched_debug.cpu#12.nr_uninterruptible
60396 ± 5% +29.7% 78305 ± 5% sched_debug.cpu#12.sched_count
10850 ± 3% -16.2% 9096 ± 3% sched_debug.cpu#12.sched_goidle
31456 ± 7% +35.7% 42670 ± 15% sched_debug.cpu#12.ttwu_count
8534 ± 9% -24.4% 6448 ± 6% sched_debug.cpu#12.ttwu_local
34463 ± 9% -24.4% 26054 ± 4% sched_debug.cpu#13.nr_switches
11721 ± 11% -22.0% 9139 ± 4% sched_debug.cpu#13.sched_goidle
28660 ± 21% +32.6% 38008 ± 11% sched_debug.cpu#13.ttwu_count
8099 ± 10% -29.3% 5728 ± 4% sched_debug.cpu#13.ttwu_local
7.25 ±173% +331.0% 31.25 ± 73% sched_debug.cpu#14.load
7767 ± 11% -29.6% 5469 ± 3% sched_debug.cpu#14.ttwu_local
33863 ± 7% -24.1% 25692 ± 5% sched_debug.cpu#15.nr_switches
53077 ± 8% +17.8% 62540 ± 4% sched_debug.cpu#15.sched_count
11485 ± 8% -20.6% 9120 ± 6% sched_debug.cpu#15.sched_goidle
7561 ± 8% -28.0% 5445 ± 4% sched_debug.cpu#15.ttwu_local
0.50 ±781% -1400.0% -6.50 ±-58% sched_debug.cpu#16.nr_uninterruptible
50380 ± 6% +32.3% 66650 ± 9% sched_debug.cpu#16.sched_count
32988 ± 13% +30.3% 42982 ± 10% sched_debug.cpu#16.ttwu_count
7258 ± 3% -24.6% 5474 ± 5% sched_debug.cpu#16.ttwu_local
32669 ± 7% -21.0% 25796 ± 3% sched_debug.cpu#17.nr_switches
51010 ± 7% +27.7% 65151 ± 7% sched_debug.cpu#17.sched_count
10905 ± 7% -15.1% 9263 ± 4% sched_debug.cpu#17.sched_goidle
7554 ± 7% -29.4% 5333 ± 2% sched_debug.cpu#17.ttwu_local
42588 ± 5% -57.5% 18111 ± 5% sched_debug.cpu#18.nr_switches
48946 ± 5% -45.6% 26618 ± 3% sched_debug.cpu#18.sched_count
14079 ± 6% -52.5% 6685 ± 5% sched_debug.cpu#18.sched_goidle
59721 ± 14% -34.8% 38956 ± 15% sched_debug.cpu#18.ttwu_count
9908 ± 4% -60.9% 3871 ± 5% sched_debug.cpu#18.ttwu_local
42395 ± 9% -55.6% 18832 ± 7% sched_debug.cpu#19.nr_switches
48669 ± 8% -41.7% 28376 ± 10% sched_debug.cpu#19.sched_count
14408 ± 11% -51.8% 6938 ± 4% sched_debug.cpu#19.sched_goidle
53917 ± 8% -30.3% 37553 ± 22% sched_debug.cpu#19.ttwu_count
9427 ± 4% -56.6% 4092 ± 11% sched_debug.cpu#19.ttwu_local
147731 ± 6% -23.9% 112487 ± 13% sched_debug.cpu#2.nr_switches
29874 ± 7% -38.2% 18457 ± 4% sched_debug.cpu#2.ttwu_local
40196 ± 1% -53.8% 18575 ± 7% sched_debug.cpu#20.nr_switches
46237 ± 2% -42.8% 26424 ± 7% sched_debug.cpu#20.sched_count
13149 ± 1% -48.0% 6832 ± 9% sched_debug.cpu#20.sched_goidle
58681 ± 17% -32.2% 39806 ± 7% sched_debug.cpu#20.ttwu_count
9561 ± 3% -58.7% 3948 ± 3% sched_debug.cpu#20.ttwu_local
44696 ± 12% -56.2% 19577 ± 4% sched_debug.cpu#21.nr_switches
50784 ± 13% -44.8% 28024 ± 9% sched_debug.cpu#21.sched_count
15527 ± 16% -52.4% 7393 ± 8% sched_debug.cpu#21.sched_goidle
57233 ± 8% -40.5% 34066 ± 11% sched_debug.cpu#21.ttwu_count
9417 ± 2% -56.3% 4111 ± 8% sched_debug.cpu#21.ttwu_local
24.00 ± 74% -99.0% 0.25 ±173% sched_debug.cpu#22.cpu_load[1]
19107 ± 0% -9.1% 17364 ± 2% sched_debug.cpu#22.nr_load_updates
39083 ± 4% -45.3% 21397 ± 18% sched_debug.cpu#22.nr_switches
45608 ± 4% -36.6% 28906 ± 16% sched_debug.cpu#22.sched_count
12789 ± 4% -35.0% 8314 ± 22% sched_debug.cpu#22.sched_goidle
55277 ± 8% -26.6% 40574 ± 16% sched_debug.cpu#22.ttwu_count
9233 ± 5% -57.4% 3931 ± 8% sched_debug.cpu#22.ttwu_local
10.25 ± 85% -95.1% 0.50 ±173% sched_debug.cpu#23.cpu_load[3]
8.50 ± 71% -91.2% 0.75 ±173% sched_debug.cpu#23.cpu_load[4]
40504 ± 5% -50.9% 19893 ± 11% sched_debug.cpu#23.nr_switches
46592 ± 5% -37.8% 28998 ± 12% sched_debug.cpu#23.sched_count
13555 ± 8% -44.8% 7482 ± 12% sched_debug.cpu#23.sched_goidle
9260 ± 2% -57.6% 3929 ± 5% sched_debug.cpu#23.ttwu_local
138922 ± 7% -26.0% 102743 ± 6% sched_debug.cpu#3.nr_switches
50963 ± 9% -16.1% 42759 ± 7% sched_debug.cpu#3.sched_goidle
28726 ± 1% -45.2% 15753 ± 5% sched_debug.cpu#3.ttwu_local
144321 ± 11% -33.2% 96457 ± 4% sched_debug.cpu#4.nr_switches
53378 ± 14% -25.9% 39541 ± 5% sched_debug.cpu#4.sched_goidle
28224 ± 3% -40.6% 16760 ± 8% sched_debug.cpu#4.ttwu_local
150005 ± 9% -25.7% 111426 ± 8% sched_debug.cpu#5.nr_switches
56676 ± 10% -17.4% 46823 ± 10% sched_debug.cpu#5.sched_goidle
27292 ± 8% -35.7% 17541 ± 13% sched_debug.cpu#5.ttwu_local
38010 ± 2% -10.1% 34177 ± 6% sched_debug.cpu#6.nr_load_updates
168056 ± 4% -40.0% 100828 ± 15% sched_debug.cpu#6.nr_switches
262900 ± 3% -31.4% 180287 ± 11% sched_debug.cpu#6.sched_count
61232 ± 6% -29.9% 42898 ± 18% sched_debug.cpu#6.sched_goidle
154457 ± 3% -25.4% 115273 ± 1% sched_debug.cpu#6.ttwu_count
33443 ± 2% -55.3% 14959 ± 9% sched_debug.cpu#6.ttwu_local
170525 ± 5% -42.4% 98162 ± 9% sched_debug.cpu#7.nr_switches
-46.50 ±-23% -69.9% -14.00 ±-67% sched_debug.cpu#7.nr_uninterruptible
265048 ± 4% -30.2% 185134 ± 5% sched_debug.cpu#7.sched_count
62399 ± 7% -33.0% 41796 ± 10% sched_debug.cpu#7.sched_goidle
154392 ± 1% -30.4% 107510 ± 5% sched_debug.cpu#7.ttwu_count
33054 ± 2% -53.2% 15465 ± 15% sched_debug.cpu#7.ttwu_local
37815 ± 2% -11.0% 33641 ± 5% sched_debug.cpu#8.nr_load_updates
163043 ± 5% -43.7% 91771 ± 8% sched_debug.cpu#8.nr_switches
-43.75 ±-27% -81.1% -8.25 ±-136% sched_debug.cpu#8.nr_uninterruptible
258882 ± 3% -34.3% 170056 ± 4% sched_debug.cpu#8.sched_count
58242 ± 6% -33.4% 38773 ± 10% sched_debug.cpu#8.sched_goidle
157201 ± 2% -32.7% 105794 ± 8% sched_debug.cpu#8.ttwu_count
33488 ± 3% -58.9% 13772 ± 9% sched_debug.cpu#8.ttwu_local
37709 ± 2% -11.5% 33383 ± 3% sched_debug.cpu#9.nr_load_updates
172842 ± 10% -45.0% 95016 ± 9% sched_debug.cpu#9.nr_switches
-38.75 ±-45% -76.8% -9.00 ±-82% sched_debug.cpu#9.nr_uninterruptible
270984 ± 6% -35.9% 173630 ± 5% sched_debug.cpu#9.sched_count
62874 ± 12% -36.0% 40237 ± 11% sched_debug.cpu#9.sched_goidle
156190 ± 5% -33.3% 104228 ± 6% sched_debug.cpu#9.ttwu_count
33673 ± 3% -59.4% 13669 ± 5% sched_debug.cpu#9.ttwu_local
lkp-ws02: Westmere-EP
Memory: 16G
fsmark.time.involuntary_context_switches
280000 ++-----------------------------------------------------------------+
260000 *+..*..*...*...*...*..*...*...*...*..*...*...*..*...*...*...* |
| |
240000 ++ |
220000 ++ |
200000 ++ |
180000 ++ |
| |
160000 ++ |
140000 ++ |
120000 ++ |
100000 ++ |
| |
80000 O+ O O O O O O O O O O O O O O O O O O
60000 ++-----------------------------------------------------------------+
[*] bisect-good sample
[O] bisect-bad sample
To reproduce:
git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
cd lkp-tests
bin/lkp install job.yaml # job file is attached in this email
bin/lkp run job.yaml
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
Thanks,
Ying Huang
---
LKP_SERVER: inn
LKP_CGI_PORT: 80
LKP_CIFS_PORT: 139
testcase: fsmark
default-monitors:
wait: activate-monitor
kmsg:
uptime:
iostat:
vmstat:
numa-numastat:
numa-vmstat:
numa-meminfo:
proc-vmstat:
proc-stat:
interval: 10
meminfo:
slabinfo:
interrupts:
lock_stat:
latency_stats:
softirqs:
bdi_dev_mapping:
diskstats:
nfsstat:
cpuidle:
cpufreq-stats:
turbostat:
pmeter:
sched_debug:
interval: 60
cpufreq_governor:
default-watchdogs:
oom-killer:
watchdog:
commit: 9ffecb10283508260936b96022d4ee43a7798b4c
model: Westmere-EP
memory: 16G
nr_hdd_partitions: 11
hdd_partitions: "/dev/disk/by-id/scsi-35000c500*-part1"
swap_partitions:
rootfs_partition: "/dev/disk/by-id/ata-WDC_WD1002FAEX-00Z3A0_WD-WCATR5408564-part3"
category: benchmark
iterations: 1x
nr_threads: 32t
disk: 1HDD
fs: xfs
fs2: nfsv4
fsmark:
filesize: 16MB
test_size: 60G
sync_method: fsyncBeforeClose
nr_directories: 16d
nr_files_per_directory: 256fpd
queue: cyclic
testbox: lkp-ws02
tbox_group: lkp-ws02
kconfig: x86_64-rhel
enqueue_time: 2015-09-29 09:57:55.489991401 +08:00
id: 35ea9a8d063be47c43e385590bfd7cfb058c576b
user: lkp
compiler: gcc-4.9
head_commit: c5cfd4b6bead810e5aeb40a3e1a8793a91554ebc
base_commit: 9ffecb10283508260936b96022d4ee43a7798b4c
branch: linux-devel/devel-hourly-2015092809
kernel: "/pkg/linux/x86_64-rhel/gcc-4.9/9ffecb10283508260936b96022d4ee43a7798b4c/vmlinuz-4.3.0-rc3"
rootfs: debian-x86_64-2015-02-07.cgz
result_root: "/result/fsmark/1x-32t-1HDD-xfs-nfsv4-16MB-60G-fsyncBeforeClose-16d-256fpd/lkp-ws02/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/9ffecb10283508260936b96022d4ee43a7798b4c/0"
job_file: "/lkp/scheduled/lkp-ws02/cyclic_fsmark-1x-32t-1HDD-xfs-nfsv4-16MB-60G-fsyncBeforeClose-16d-256fpd-x86_64-rhel-CYCLIC_BASE-9ffecb10283508260936b96022d4ee43a7798b4c-20150929-57335-5fc4iu-0.yaml"
dequeue_time: 2015-09-29 10:18:12.225271851 +08:00
nr_cpu: "$(nproc)"
max_uptime: 2492.3999999999996
initrd: "/osimage/debian/debian-x86_64-2015-02-07.cgz"
bootloader_append:
- root=/dev/ram0
- user=lkp
- job=/lkp/scheduled/lkp-ws02/cyclic_fsmark-1x-32t-1HDD-xfs-nfsv4-16MB-60G-fsyncBeforeClose-16d-256fpd-x86_64-rhel-CYCLIC_BASE-9ffecb10283508260936b96022d4ee43a7798b4c-20150929-57335-5fc4iu-0.yaml
- ARCH=x86_64
- kconfig=x86_64-rhel
- branch=linux-devel/devel-hourly-2015092809
- commit=9ffecb10283508260936b96022d4ee43a7798b4c
- BOOT_IMAGE=/pkg/linux/x86_64-rhel/gcc-4.9/9ffecb10283508260936b96022d4ee43a7798b4c/vmlinuz-4.3.0-rc3
- max_uptime=2492
- RESULT_ROOT=/result/fsmark/1x-32t-1HDD-xfs-nfsv4-16MB-60G-fsyncBeforeClose-16d-256fpd/lkp-ws02/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/9ffecb10283508260936b96022d4ee43a7798b4c/0
- LKP_SERVER=inn
- |-
ipmi_watchdog.start_now=1
earlyprintk=ttyS0,115200 systemd.log_level=err
debug apic=debug sysrq_always_enabled rcupdate.rcu_cpu_stall_timeout=100
panic=-1 softlockup_panic=1 nmi_watchdog=panic oops=panic load_ramdisk=2 prompt_ramdisk=0
console=ttyS0,115200 console=tty0 vga=normal
rw
lkp_initrd: "/lkp/lkp/lkp-x86_64.cgz"
modules_initrd: "/pkg/linux/x86_64-rhel/gcc-4.9/9ffecb10283508260936b96022d4ee43a7798b4c/modules.cgz"
bm_initrd: "/osimage/deps/debian-x86_64-2015-02-07.cgz/lkp.cgz,/osimage/deps/debian-x86_64-2015-02-07.cgz/run-ipconfig.cgz,/osimage/deps/debian-x86_64-2015-02-07.cgz/turbostat.cgz,/lkp/benchmarks/turbostat.cgz,/osimage/deps/debian-x86_64-2015-02-07.cgz/fs.cgz,/osimage/deps/debian-x86_64-2015-02-07.cgz/fs2.cgz,/lkp/benchmarks/fsmark.cgz"
job_state: finished
loadavg: 33.51 34.65 21.92 1/358 21217
start_time: '1443493409'
end_time: '1443494209'
version: "/lkp/lkp/.src-20150929-091341"
mkfs -t xfs /dev/sdd1
mount -t xfs -o nobarrier,inode64 /dev/sdd1 /fs/sdd1
/etc/init.d/rpcbind start
/etc/init.d/nfs-common start
/etc/init.d/nfs-kernel-server start
mount -t nfs -o vers=4 localhost:/fs/sdd1 /nfs/sdd1
./fs_mark -d /nfs/sdd1/1 -d /nfs/sdd1/2 -d /nfs/sdd1/3 -d /nfs/sdd1/4 -d /nfs/sdd1/5 -d /nfs/sdd1/6 -d /nfs/sdd1/7 -d /nfs/sdd1/8 -d /nfs/sdd1/9 -d /nfs/sdd1/10 -d /nfs/sdd1/11 -d /nfs/sdd1/12 -d /nfs/sdd1/13 -d /nfs/sdd1/14 -d /nfs/sdd1/15 -d /nfs/sdd1/16 -d /nfs/sdd1/17 -d /nfs/sdd1/18 -d /nfs/sdd1/19 -d /nfs/sdd1/20 -d /nfs/sdd1/21 -d /nfs/sdd1/22 -d /nfs/sdd1/23 -d /nfs/sdd1/24 -d /nfs/sdd1/25 -d /nfs/sdd1/26 -d /nfs/sdd1/27 -d /nfs/sdd1/28 -d /nfs/sdd1/29 -d /nfs/sdd1/30 -d /nfs/sdd1/31 -d /nfs/sdd1/32 -D 16 -N 256 -n 120 -L 1 -S 1 -s 16777216