2015-11-18 06:44:18

by kernel test robot

[permalink] [raw]
Subject: [lkp] [vfs] f3f86e33dc: -5.3% will-it-scale.per_process_ops

FYI, we noticed the below changes on

https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git master
commit f3f86e33dc3da437fa4f204588ce7c78ea756982 ("vfs: Fix pathological performance case for __alloc_fd()")


=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/test:
ivb42/will-it-scale/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performance/dup1

commit:
8a28d67457b613258aa0578ccece206d166f2b9f
f3f86e33dc3da437fa4f204588ce7c78ea756982

8a28d67457b61325 f3f86e33dc3da437fa4f204588
---------------- --------------------------
%stddev %change %stddev
\ | \
5994379 ? 0% -5.3% 5678711 ? 0% will-it-scale.per_process_ops
1440545 ? 0% -5.1% 1367766 ? 2% will-it-scale.per_thread_ops
0.57 ? 0% -5.9% 0.54 ? 0% will-it-scale.scalability
4.47 ? 2% -3.1% 4.33 ? 1% turbostat.RAMWatt
59880 ? 5% -13.1% 52055 ? 11% cpuidle.C1-IVT.usage
597.50 ? 4% -19.7% 479.50 ? 16% cpuidle.POLL.usage
15756223 ? 0% +367.7% 73688311 ? 84% latency_stats.avg.nfs_wait_on_request.nfs_updatepage.nfs_write_end.generic_perform_write.__generic_file_write_iter.generic_file_write_iter.nfs_file_write.__vfs_write.vfs_write.SyS_write.entry_SYSCALL_64_fastpath
35858260 ? 0% +113.3% 76474871 ? 77% latency_stats.max.nfs_wait_on_request.nfs_updatepage.nfs_write_end.generic_perform_write.__generic_file_write_iter.generic_file_write_iter.nfs_file_write.__vfs_write.vfs_write.SyS_write.entry_SYSCALL_64_fastpath
1560 ?171% +300.1% 6241 ? 0% numa-numastat.node0.other_node
3101 ? 99% -99.7% 9.50 ? 67% numa-numastat.node1.other_node
2980 ? 3% -13.4% 2582 ? 1% slabinfo.kmalloc-2048.active_objs
3139 ? 3% -12.5% 2746 ? 1% slabinfo.kmalloc-2048.num_objs
5018 ? 14% -50.4% 2487 ? 13% numa-vmstat.node0.nr_active_anon
3121 ? 31% -77.6% 700.00 ?132% numa-vmstat.node0.nr_shmem
3349 ? 20% +76.6% 5916 ? 7% numa-vmstat.node1.nr_active_anon
1210 ? 80% +200.8% 3640 ? 25% numa-vmstat.node1.nr_shmem
70442 ? 5% -12.7% 61484 ? 0% numa-meminfo.node0.Active
20079 ? 14% -50.4% 9954 ? 13% numa-meminfo.node0.Active(anon)
12487 ? 31% -77.6% 2801 ?132% numa-meminfo.node0.Shmem
61970 ? 5% +18.0% 73096 ? 1% numa-meminfo.node1.Active
13402 ? 20% +76.6% 23671 ? 7% numa-meminfo.node1.Active(anon)
4843 ? 80% +200.7% 14564 ? 25% numa-meminfo.node1.Shmem
1999660 ? 2% -9.1% 1817792 ? 6% sched_debug.cfs_rq[0]:/.min_vruntime
814.25 ? 6% -11.9% 717.00 ? 10% sched_debug.cfs_rq[0]:/.util_avg
-220009 ?-25% -83.5% -36294 ?-314% sched_debug.cfs_rq[10]:/.spread0
-220410 ?-25% -82.7% -38205 ?-290% sched_debug.cfs_rq[11]:/.spread0
-868154 ? -5% -20.7% -688065 ?-16% sched_debug.cfs_rq[14]:/.spread0
13.00 ? 0% +105.8% 26.75 ? 61% sched_debug.cfs_rq[15]:/.load
-952278 ?-16% -27.9% -687025 ?-16% sched_debug.cfs_rq[16]:/.spread0
-876660 ? -6% -21.8% -685915 ?-15% sched_debug.cfs_rq[17]:/.spread0
-869841 ? -6% -20.5% -691511 ?-15% sched_debug.cfs_rq[18]:/.spread0
-872906 ? -6% -21.0% -689435 ?-15% sched_debug.cfs_rq[19]:/.spread0
-220042 ?-24% -82.8% -37798 ?-282% sched_debug.cfs_rq[1]:/.spread0
-870736 ? -6% -20.9% -689178 ?-16% sched_debug.cfs_rq[20]:/.spread0
-870782 ? -5% -20.6% -691440 ?-16% sched_debug.cfs_rq[21]:/.spread0
12.50 ? 12% +20.0% 15.00 ? 8% sched_debug.cfs_rq[23]:/.load_avg
-947289 ?-16% -27.8% -684292 ?-16% sched_debug.cfs_rq[23]:/.spread0
12.50 ? 12% +20.0% 15.00 ? 8% sched_debug.cfs_rq[23]:/.tg_load_avg_contrib
424.00 ? 13% +27.4% 540.00 ? 9% sched_debug.cfs_rq[23]:/.util_avg
-180921 ?-30% -100.2% 424.29 ?26645% sched_debug.cfs_rq[25]:/.spread0
-179335 ?-30% -82.3% -31706 ?-346% sched_debug.cfs_rq[26]:/.spread0
-180972 ?-30% -100.1% 163.84 ?68609% sched_debug.cfs_rq[27]:/.spread0
-179636 ?-30% -100.4% 736.15 ?15384% sched_debug.cfs_rq[28]:/.spread0
-180380 ?-30% -101.1% 1963 ?5772% sched_debug.cfs_rq[29]:/.spread0
26.00 ? 3% -18.3% 21.25 ? 22% sched_debug.cfs_rq[2]:/.load
29.50 ? 9% -21.2% 23.25 ? 23% sched_debug.cfs_rq[2]:/.runnable_load_avg
-211354 ?-27% -97.3% -5780 ?-2383% sched_debug.cfs_rq[2]:/.spread0
762.25 ? 6% -17.1% 632.00 ? 11% sched_debug.cfs_rq[2]:/.util_avg
-179346 ?-31% -101.0% 1767 ?6351% sched_debug.cfs_rq[30]:/.spread0
-182129 ?-30% -99.9% -200.51 ?-56625% sched_debug.cfs_rq[31]:/.spread0
-178388 ?-30% -99.9% -162.33 ?-69718% sched_debug.cfs_rq[32]:/.spread0
-178678 ?-30% -100.0% -67.48 ?-166628% sched_debug.cfs_rq[33]:/.spread0
-177514 ?-30% -100.1% 200.37 ?56326% sched_debug.cfs_rq[34]:/.spread0
-178339 ?-29% -101.6% 2870 ?3873% sched_debug.cfs_rq[35]:/.spread0
-795803 ? -8% -34.5% -521305 ?-40% sched_debug.cfs_rq[37]:/.spread0
-783897 ? -6% -22.6% -607100 ?-18% sched_debug.cfs_rq[38]:/.spread0
3.00 ? 0% +250.0% 10.50 ? 40% sched_debug.cfs_rq[39]:/.load_avg
-784040 ? -6% -33.2% -523669 ?-39% sched_debug.cfs_rq[39]:/.spread0
3.00 ? 0% +250.0% 10.50 ? 40% sched_debug.cfs_rq[39]:/.tg_load_avg_contrib
173.75 ? 4% +36.8% 237.75 ? 30% sched_debug.cfs_rq[39]:/.util_avg
-220092 ?-24% -82.4% -38783 ?-288% sched_debug.cfs_rq[3]:/.spread0
-783338 ? -6% -22.8% -604971 ?-18% sched_debug.cfs_rq[41]:/.spread0
-784423 ? -6% -23.1% -603402 ?-17% sched_debug.cfs_rq[42]:/.spread0
-785872 ? -6% -23.0% -605005 ?-18% sched_debug.cfs_rq[43]:/.spread0
-782962 ? -6% -22.9% -603838 ?-19% sched_debug.cfs_rq[44]:/.spread0
-783170 ? -6% -23.2% -601383 ?-18% sched_debug.cfs_rq[45]:/.spread0
-784950 ? -6% -23.2% -602937 ?-18% sched_debug.cfs_rq[46]:/.spread0
32.25 ? 35% -24.8% 24.25 ? 1% sched_debug.cfs_rq[4]:/.load
-217411 ?-24% -83.2% -36433 ?-300% sched_debug.cfs_rq[4]:/.spread0
-219424 ?-25% -83.0% -37233 ?-299% sched_debug.cfs_rq[5]:/.spread0
-219112 ?-25% -82.4% -38536 ?-289% sched_debug.cfs_rq[6]:/.spread0
-218643 ?-24% -82.8% -37629 ?-298% sched_debug.cfs_rq[7]:/.spread0
-220909 ?-24% -85.0% -33175 ?-350% sched_debug.cfs_rq[8]:/.spread0
-220076 ?-25% -85.9% -31115 ?-337% sched_debug.cfs_rq[9]:/.spread0
89160 ? 6% -8.7% 81395 ? 7% sched_debug.cpu#0.nr_load_updates
-2.75 ?-126% -172.7% 2.00 ?136% sched_debug.cpu#12.nr_uninterruptible
16563 ? 21% -39.6% 10009 ? 9% sched_debug.cpu#13.nr_switches
16901 ? 21% -34.5% 11064 ? 9% sched_debug.cpu#13.sched_count
6432 ? 27% -47.2% 3396 ? 38% sched_debug.cpu#13.sched_goidle
7244 ? 14% -45.3% 3961 ? 39% sched_debug.cpu#14.sched_goidle
13.00 ? 0% +105.8% 26.75 ? 61% sched_debug.cpu#15.load
1554 ? 21% +62.9% 2531 ? 30% sched_debug.cpu#16.ttwu_local
6965 ? 21% +48.0% 10308 ? 17% sched_debug.cpu#18.sched_count
28.25 ? 2% -17.7% 23.25 ? 23% sched_debug.cpu#2.cpu_load[4]
26.00 ? 3% -18.3% 21.25 ? 22% sched_debug.cpu#2.load
2703 ? 9% +11.5% 3014 ? 8% sched_debug.cpu#24.curr->pid
420.00 ? 27% -34.1% 276.75 ? 30% sched_debug.cpu#25.sched_goidle
247.00 ? 24% +55.7% 384.50 ? 26% sched_debug.cpu#27.sched_goidle
-2.25 ?-79% -211.1% 2.50 ? 82% sched_debug.cpu#30.nr_uninterruptible
715.75 ? 46% -44.2% 399.50 ? 35% sched_debug.cpu#32.ttwu_count
133.50 ? 22% +99.4% 266.25 ? 29% sched_debug.cpu#33.ttwu_local
1212 ? 47% -46.7% 646.25 ? 25% sched_debug.cpu#35.nr_switches
506.50 ? 46% -51.6% 245.00 ? 30% sched_debug.cpu#35.sched_goidle
32.25 ? 35% -24.8% 24.25 ? 1% sched_debug.cpu#4.load
2973 ? 46% +161.2% 7766 ? 47% sched_debug.cpu#40.nr_switches
3062 ? 46% +156.1% 7843 ? 47% sched_debug.cpu#40.sched_count
1219 ? 55% +155.9% 3121 ? 60% sched_debug.cpu#40.sched_goidle
1429 ? 2% +49.2% 2131 ? 41% sched_debug.cpu#44.curr->pid
1.75 ? 93% -57.1% 0.75 ?145% sched_debug.cpu#45.nr_uninterruptible
433.75 ? 32% +75.1% 759.50 ? 14% sched_debug.cpu#6.ttwu_count


ivb42: Ivytown Ivy Bridge-EP
Memory: 64G




will-it-scale.per_process_ops

6.05e+06 ++---------------------------------------------------------------+
*.*.**.*.*.**.*.*. .*. *. .* .*.**. |
6e+06 ++ **.*.*.** *.* *.*.**.*.* *.* *.*.**.*.*
5.95e+06 ++ |
| |
5.9e+06 ++ |
5.85e+06 ++ |
| |
5.8e+06 ++ |
5.75e+06 ++ |
| O OO O |
5.7e+06 O+ OO O O O O O OO O O |
5.65e+06 ++O O O O O OO O |
| O |
5.6e+06 ++---------------------------------------------------------------+

[*] bisect-good sample
[O] bisect-bad sample

To reproduce:

git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
cd lkp-tests
bin/lkp install job.yaml # job file is attached in this email
bin/lkp run job.yaml


Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


Thanks,
Ying Huang


Attachments:
(No filename) (10.44 kB)
job.yaml (3.08 kB)
reproduce (3.50 kB)
Download all attachments