[LKP] [rcu] 00f5159aeb6: +55.3% will-it-scale.time.involuntary_context_switches

From: Huang Ying
Date: Wed Jan 21 2015 - 02:33:22 EST


FYI, we noticed the below changes on

git://git.kernel.org/pub/scm/linux/kernel/git/paulmck/linux-rcu.git stall.2015.01.06a
commit 00f5159aeb68f08ce29597be86be87d6db0c1ba1 ("rcu: Run grace-period kthreads at real-time priority")


testbox/testcase/testparams: lkp-a03/will-it-scale/performance-lock1

e3663b1024d1f946 00f5159aeb68f08ce29597be86
---------------- --------------------------
%stddev %change %stddev
\ | \
11321 Â 5% +55.3% 17582 Â 14% will-it-scale.time.involuntary_context_switches
28999 Â 44% +109.3% 60700 Â 25% sched_debug.cpu#1.ttwu_count
58384 Â 44% +103.6% 118851 Â 24% sched_debug.cpu#1.nr_switches
26144 Â 48% +90.3% 49762 Â 26% sched_debug.cpu#1.sched_goidle
827 Â 34% -57.3% 353 Â 21% cpuidle.C2-ATM.usage
58976 Â 43% +103.0% 119712 Â 24% sched_debug.cpu#1.sched_count
503 Â 29% -56.5% 219 Â 27% cpuidle.C4-ATM.usage
11321 Â 5% +55.3% 17582 Â 14% time.involuntary_context_switches
156 Â 14% -29.6% 110 Â 19% cpuidle.POLL.usage
50583 Â 2% -21.7% 39630 Â 4% cpuidle.C6-ATM.usage
51597156 Â 2% +23.7% 63848564 Â 3% cpuidle.C1E-ATM.time
39349 Â 2% -12.0% 34622 Â 2% softirqs.SCHED
758790 Â 8% -15.3% 642336 Â 12% sched_debug.cpu#2.sched_count
378981 Â 8% -15.6% 319776 Â 12% sched_debug.cpu#2.ttwu_count
374440 Â 8% -15.5% 316513 Â 12% sched_debug.cpu#2.ttwu_local
5307 Â 0% -1.0% 5255 Â 0% vmstat.system.in

testbox/testcase/testparams: wsm/will-it-scale/performance-open2

e3663b1024d1f946 00f5159aeb68f08ce29597be86
---------------- --------------------------
22128 Â 3% +120.2% 48719 Â 8% will-it-scale.time.involuntary_context_switches
5577 Â 45% +1492.2% 88802 Â 33% sched_debug.cpu#6.sched_count
5457 Â 46% +1526.2% 88751 Â 33% sched_debug.cpu#6.nr_switches
0.45 Â 43% -73.3% 0.12 Â 15% turbostat.%c3
10 Â 24% -55.8% 4 Â 40% sched_debug.cpu#9.nr_uninterruptible
147 Â 11% -58.5% 61 Â 32% sched_debug.cpu#11.load
147 Â 11% -58.5% 61 Â 32% sched_debug.cfs_rq[11]:/.load
10821106 Â 39% -59.8% 4347430 Â 20% cpuidle.C3-NHM.time
356 Â 41% +76.6% 629 Â 3% sched_debug.cfs_rq[6]:/.blocked_load_avg
22128 Â 3% +120.2% 48719 Â 8% time.involuntary_context_switches
145 Â 36% -35.9% 93 Â 30% sched_debug.cpu#5.load
469 Â 29% +55.9% 732 Â 4% sched_debug.cfs_rq[6]:/.tg_load_contrib
145 Â 36% -34.0% 95 Â 29% sched_debug.cfs_rq[5]:/.load
99 Â 8% -27.8% 72 Â 24% sched_debug.cpu#9.load
99 Â 8% -27.8% 72 Â 24% sched_debug.cfs_rq[9]:/.load
219871 Â 6% -33.3% 146545 Â 19% sched_debug.cpu#3.ttwu_local
224610 Â 6% -31.3% 154342 Â 18% sched_debug.cpu#3.sched_goidle
451877 Â 6% -31.1% 311378 Â 18% sched_debug.cpu#3.sched_count
451737 Â 6% -31.1% 311279 Â 18% sched_debug.cpu#3.nr_switches
225208 Â 6% -31.3% 154688 Â 18% sched_debug.cpu#3.ttwu_count
4122 Â 7% -23.6% 3148 Â 17% cpuidle.C3-NHM.usage
84 Â 11% -27.4% 61 Â 5% sched_debug.cpu#7.cpu_load[0]
154565 Â 2% -29.6% 108768 Â 9% cpuidle.C6-NHM.usage
124949 Â 25% +41.4% 176670 Â 15% sched_debug.cpu#2.sched_goidle
74177 Â 2% -23.6% 56653 Â 4% sched_debug.cpu#3.nr_load_updates
79 Â 6% -23.1% 60 Â 5% sched_debug.cpu#7.cpu_load[1]
50107 Â 6% -20.9% 39621 Â 7% sched_debug.cfs_rq[3]:/.exec_clock
76 Â 4% -20.6% 60 Â 6% sched_debug.cpu#7.cpu_load[2]
385983 Â 9% -20.4% 307347 Â 12% sched_debug.cfs_rq[3]:/.min_vruntime
74 Â 2% -18.5% 60 Â 6% sched_debug.cpu#7.cpu_load[3]
1.50 Â 6% -10.0% 1.35 Â 9% perf-profile.cpu-cycles.selinux_file_free_security.security_file_free.__fput.____fput.task_work_run
46086 Â 1% -25.7% 34260 Â 5% softirqs.SCHED
72 Â 1% -15.9% 60 Â 7% sched_debug.cpu#7.cpu_load[4]
74 Â 13% -17.7% 61 Â 2% sched_debug.cfs_rq[7]:/.runnable_load_avg
65 Â 2% +16.9% 76 Â 4% sched_debug.cpu#1.cpu_load[3]
1.94 Â 5% -9.9% 1.74 Â 6% perf-profile.cpu-cycles.security_file_free.__fput.____fput.task_work_run.do_notify_resume
58851 Â 1% +20.5% 70942 Â 9% sched_debug.cfs_rq[1]:/.exec_clock
69880 Â 1% -16.9% 58047 Â 11% sched_debug.cpu#7.nr_load_updates
60 Â 5% +9.5% 66 Â 3% sched_debug.cpu#9.cpu_load[4]
12064 Â 7% +12.4% 13561 Â 5% slabinfo.kmalloc-64.num_objs
67 Â 4% +14.9% 77 Â 4% sched_debug.cpu#1.cpu_load[2]
53070 Â 7% +18.3% 62803 Â 5% sched_debug.cfs_rq[9]:/.exec_clock
64 Â 1% +16.3% 74 Â 4% sched_debug.cpu#1.cpu_load[4]
11895 Â 7% +12.9% 13429 Â 5% slabinfo.kmalloc-64.active_objs
67415 Â 1% -16.4% 56345 Â 11% sched_debug.cfs_rq[7]:/.exec_clock
58403 Â 4% +12.0% 65434 Â 6% sched_debug.cpu#9.nr_load_updates
428535 Â 8% +18.3% 507136 Â 6% sched_debug.cfs_rq[9]:/.min_vruntime
505551 Â 1% +16.3% 588207 Â 8% sched_debug.cfs_rq[1]:/.min_vruntime
1818 Â 5% -8.4% 1665 Â 3% slabinfo.proc_inode_cache.active_objs
1818 Â 5% -8.4% 1665 Â 3% slabinfo.proc_inode_cache.num_objs
616 Â 0% +11.4% 686 Â 4% sched_debug.cfs_rq[1]:/.tg_runnable_contrib
2259 Â 14% +23.5% 2790 Â 5% sched_debug.cpu#3.curr->pid
28336 Â 0% +11.2% 31517 Â 4% sched_debug.cfs_rq[1]:/.avg->runnable_avg_sum
576795 Â 1% -14.6% 492717 Â 10% sched_debug.cfs_rq[7]:/.min_vruntime
7176 Â 0% +2.1% 7323 Â 0% vmstat.system.cs
10798 Â 0% -1.1% 10675 Â 0% vmstat.system.in

lkp-a03: Atom
Memory: 8G

wsm: Westmere
Memory: 6G




will-it-scale.time.involuntary_context_switches

70000 ++------------------------------------------------------------------+
| |
60000 ++ O |
O O O O O O O O O O |
50000 ++ O O O O O O O O
| O O O O O O |
40000 ++ O O O O O |
| |
30000 ++ |
| .*. .*.*.. .* |
20000 *+*..* *..*.* *.*.*..*.*.*. *.*.*..*.*.*..*.*.*..* |
| : : |
10000 ++ : : |
| :: |
0 ++--------------*---------------------------------------------------+


[*] bisect-good sample
[O] bisect-bad sample

To reproduce:

apt-get install ruby ruby-oj
git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
cd lkp-tests
bin/setup-local job.yaml # the job file attached in this email
bin/run-local job.yaml


Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


Thanks,
Huang, Ying


---
testcase: will-it-scale
default-monitors:
wait: pre-test
uptime:
iostat:
vmstat:
numa-numastat:
numa-vmstat:
numa-meminfo:
proc-vmstat:
proc-stat:
meminfo:
slabinfo:
interrupts:
lock_stat:
latency_stats:
softirqs:
bdi_dev_mapping:
diskstats:
cpuidle:
cpufreq:
turbostat:
sched_debug:
interval: 10
pmeter:
default_watchdogs:
watch-oom:
watchdog:
cpufreq_governor:
- performance
commit: 634b0bd490b7ebd7a054cea4f7e0d25748bde678
model: Westmere
memory: 6G
nr_hdd_partitions: 1
hdd_partitions:
swap_partitions:
rootfs_partition:
netconsole_port: 6667
perf-profile:
freq: 800
will-it-scale:
test:
- open2
testbox: wsm
tbox_group: wsm
kconfig: x86_64-rhel
enqueue_time: 2015-01-16 17:43:47.976990621 +08:00
head_commit: 634b0bd490b7ebd7a054cea4f7e0d25748bde678
base_commit: eaa27f34e91a14cdceed26ed6c6793ec1d186115
branch: next/master
kernel: "/kernel/x86_64-rhel/634b0bd490b7ebd7a054cea4f7e0d25748bde678/vmlinuz-3.19.0-rc4-next-20150116-g634b0bd"
user: lkp
queue: cyclic
rootfs: debian-x86_64-2015-01-13.cgz
result_root: "/result/wsm/will-it-scale/performance-open2/debian-x86_64-2015-01-13.cgz/x86_64-rhel/634b0bd490b7ebd7a054cea4f7e0d25748bde678/0"
job_file: "/lkp/scheduled/wsm/cyclic_will-it-scale-performance-open2-x86_64-rhel-HEAD-634b0bd490b7ebd7a054cea4f7e0d25748bde678-0.yaml"
dequeue_time: 2015-01-16 23:50:07.247065623 +08:00
nr_cpu: "$(nproc)"
job_state: finished
loadavg: 8.12 4.97 2.06 1/146 6693
start_time: '1421423439'
end_time: '1421423743'
version: "/lkp/lkp/.src-20150116-113525"
./runtest.py open2 32 1 6 9 12
_______________________________________________
LKP mailing list
LKP@xxxxxxxxxxxxxxx