2022-06-19 15:08:21

by kernel test robot

[permalink] [raw]
Subject: [net] 4890b686f4: netperf.Throughput_Mbps -69.4% regression



Greeting,

FYI, we noticed a -69.4% regression of netperf.Throughput_Mbps due to commit:


commit: 4890b686f4088c90432149bd6de567e621266fa2 ("net: keep sk->sk_forward_alloc as small as possible")
https://git.kernel.org/cgit/linux/kernel/git/next/linux-next.git master

in testcase: netperf
on test machine: 144 threads 4 sockets Intel(R) Xeon(R) Gold 5318H CPU @ 2.50GHz with 128G memory
with following parameters:

ip: ipv4
runtime: 300s
nr_threads: 25%
cluster: cs-localhost
send_size: 10K
test: SCTP_STREAM_MANY
cpufreq_governor: performance
ucode: 0x7002402

test-description: Netperf is a benchmark that can be use to measure various aspect of networking performance.
test-url: http://www.netperf.org/netperf/

In addition to that, the commit also has significant impact on the following tests:

+------------------+----------------------------------------------------------------------------------+
| testcase: change | netperf: netperf.Throughput_Mbps -73.7% regression |
| test machine | 144 threads 4 sockets Intel(R) Xeon(R) Gold 5318H CPU @ 2.50GHz with 128G memory |
| test parameters | cluster=cs-localhost |
| | cpufreq_governor=performance |
| | ip=ipv4 |
| | nr_threads=50% |
| | runtime=300s |
| | send_size=10K |
| | test=SCTP_STREAM_MANY |
| | ucode=0x7002402 |
+------------------+----------------------------------------------------------------------------------+


If you fix the issue, kindly add following tag
Reported-by: kernel test robot <[email protected]>


Details are as below:
-------------------------------------------------------------------------------------------------->


To reproduce:

git clone https://github.com/intel/lkp-tests.git
cd lkp-tests
sudo bin/lkp install job.yaml # job file is attached in this email
bin/lkp split-job --compatible job.yaml # generate the yaml file for lkp run
sudo bin/lkp run generated-yaml-file

# if come across any failure that blocks the test,
# please remove ~/.lkp and /lkp dir to run from a clean state.

=========================================================================================
cluster/compiler/cpufreq_governor/ip/kconfig/nr_threads/rootfs/runtime/send_size/tbox_group/test/testcase/ucode:
cs-localhost/gcc-11/performance/ipv4/x86_64-rhel-8.3/25%/debian-11.1-x86_64-20220510.cgz/300s/10K/lkp-cpl-4sp1/SCTP_STREAM_MANY/netperf/0x7002402

commit:
7c80b038d2 ("net: fix sk_wmem_schedule() and sk_rmem_schedule() errors")
4890b686f4 ("net: keep sk->sk_forward_alloc as small as possible")

7c80b038d23e1f4c 4890b686f4088c90432149bd6de
---------------- ---------------------------
%stddev %change %stddev
\ | \
15855 -69.4% 4854 netperf.Throughput_Mbps
570788 -69.4% 174773 netperf.Throughput_total_Mbps
30543 -59.1% 12480 ? 3% netperf.time.involuntary_context_switches
16661 ? 2% -20.1% 13317 ? 2% netperf.time.minor_page_faults
2076 -58.1% 869.17 ? 2% netperf.time.percent_of_cpu_this_job_got
6118 -57.8% 2583 ? 2% netperf.time.system_time
143.71 ? 6% -72.3% 39.77 netperf.time.user_time
87371 ? 8% +81.1% 158253 netperf.time.voluntary_context_switches
2.09e+09 -69.4% 6.4e+08 netperf.workload
36264 +18.2% 42849 uptime.idle
3.058e+10 +22.0% 3.731e+10 cpuidle..time
6.536e+08 -60.7% 2.566e+08 cpuidle..usage
127819 ? 79% -82.3% 22609 ? 65% numa-meminfo.node3.Inactive
127764 ? 79% -82.4% 22440 ? 66% numa-meminfo.node3.Inactive(anon)
71.38 +14.7 86.09 mpstat.cpu.all.idle%
4.84 -2.1 2.78 ? 2% mpstat.cpu.all.soft%
22.32 -12.1 10.25 ? 2% mpstat.cpu.all.sys%
0.78 ? 3% -0.5 0.25 mpstat.cpu.all.usr%
71.00 +20.9% 85.83 vmstat.cpu.id
4121749 -23.6% 3150424 vmstat.memory.cache
43.00 ? 2% -55.0% 19.33 ? 2% vmstat.procs.r
3921783 -69.4% 1200263 vmstat.system.cs
292198 -1.3% 288478 vmstat.system.in
6.908e+08 ? 3% -70.1% 2.064e+08 ? 9% numa-numastat.node0.local_node
6.907e+08 ? 3% -70.1% 2.065e+08 ? 9% numa-numastat.node0.numa_hit
6.517e+08 ? 2% -65.5% 2.246e+08 ? 10% numa-numastat.node1.local_node
6.519e+08 ? 2% -65.5% 2.247e+08 ? 10% numa-numastat.node1.numa_hit
6.81e+08 ? 3% -72.1% 1.902e+08 ? 8% numa-numastat.node2.local_node
6.811e+08 ? 3% -72.0% 1.905e+08 ? 8% numa-numastat.node2.numa_hit
6.819e+08 ? 3% -69.9% 2.05e+08 ? 6% numa-numastat.node3.local_node
6.823e+08 ? 3% -69.9% 2.051e+08 ? 6% numa-numastat.node3.numa_hit
6.907e+08 ? 3% -70.1% 2.065e+08 ? 9% numa-vmstat.node0.numa_hit
6.908e+08 ? 3% -70.1% 2.064e+08 ? 9% numa-vmstat.node0.numa_local
6.519e+08 ? 2% -65.5% 2.247e+08 ? 10% numa-vmstat.node1.numa_hit
6.517e+08 ? 2% -65.5% 2.246e+08 ? 10% numa-vmstat.node1.numa_local
6.811e+08 ? 3% -72.0% 1.905e+08 ? 8% numa-vmstat.node2.numa_hit
6.81e+08 ? 3% -72.1% 1.902e+08 ? 8% numa-vmstat.node2.numa_local
31948 ? 79% -82.4% 5622 ? 66% numa-vmstat.node3.nr_inactive_anon
31948 ? 79% -82.4% 5622 ? 66% numa-vmstat.node3.nr_zone_inactive_anon
6.823e+08 ? 3% -69.9% 2.051e+08 ? 6% numa-vmstat.node3.numa_hit
6.819e+08 ? 3% -69.9% 2.05e+08 ? 6% numa-vmstat.node3.numa_local
1216292 ? 5% -70.1% 363637 ? 4% meminfo.Active
1215873 ? 5% -70.1% 363097 ? 4% meminfo.Active(anon)
3994598 -24.3% 3022264 meminfo.Cached
1878317 ? 4% -53.7% 870428 meminfo.Committed_AS
431272 ? 8% -28.0% 310461 meminfo.Inactive
428860 ? 9% -28.4% 306912 meminfo.Inactive(anon)
668273 ? 5% -71.3% 192010 ? 4% meminfo.Mapped
5793516 -17.1% 4802753 meminfo.Memused
8320 ? 2% -12.1% 7309 ? 2% meminfo.PageTables
1341918 ? 6% -72.5% 368590 ? 4% meminfo.Shmem
9796929 -25.1% 7337138 meminfo.max_used_kB
1029 -51.6% 498.17 ? 2% turbostat.Avg_MHz
31.32 -16.2 15.17 ? 2% turbostat.Busy%
5.953e+08 -69.1% 1.837e+08 ? 2% turbostat.C1
9.22 -5.2 4.04 ? 4% turbostat.C1%
68.58 +23.5% 84.67 turbostat.CPU%c1
57.67 -16.8% 48.00 ? 3% turbostat.CoreTmp
0.23 ? 2% -31.4% 0.16 turbostat.IPC
18.59 ? 30% -18.6 0.00 turbostat.PKG_%
4524200 ? 6% -81.5% 835280 turbostat.POLL
0.10 ? 3% -0.1 0.02 ? 17% turbostat.POLL%
57.33 -16.6% 47.83 ? 4% turbostat.PkgTmp
561.39 -25.9% 416.03 turbostat.PkgWatt
7.75 +6.8% 8.27 turbostat.RAMWatt
303881 ? 5% -70.2% 90609 ? 4% proc-vmstat.nr_active_anon
998594 -24.4% 755404 proc-vmstat.nr_file_pages
107254 ? 9% -28.4% 76741 proc-vmstat.nr_inactive_anon
167104 ? 5% -71.0% 48414 ? 4% proc-vmstat.nr_mapped
2081 ? 2% -12.1% 1828 ? 2% proc-vmstat.nr_page_table_pages
335422 ? 6% -72.6% 91983 ? 4% proc-vmstat.nr_shmem
32561 -1.7% 31994 proc-vmstat.nr_slab_reclaimable
303881 ? 5% -70.2% 90609 ? 4% proc-vmstat.nr_zone_active_anon
107254 ? 9% -28.4% 76741 proc-vmstat.nr_zone_inactive_anon
267554 ? 15% -25.4% 199518 ? 10% proc-vmstat.numa_hint_faults
82267 ? 20% -27.2% 59929 ? 13% proc-vmstat.numa_hint_faults_local
2.706e+09 -69.5% 8.266e+08 proc-vmstat.numa_hit
2.705e+09 -69.5% 8.262e+08 proc-vmstat.numa_local
489361 ? 9% -19.4% 394229 ? 6% proc-vmstat.numa_pte_updates
510140 ? 18% -75.3% 125829 ? 2% proc-vmstat.pgactivate
7.17e+09 -69.4% 2.197e+09 proc-vmstat.pgalloc_normal
1636476 ? 2% -11.5% 1448944 proc-vmstat.pgfault
7.17e+09 -69.4% 2.197e+09 proc-vmstat.pgfree
30967 ? 41% -88.7% 3488 ? 30% sched_debug.cfs_rq:/.MIN_vruntime.avg
1604384 ? 22% -83.6% 262510 ? 14% sched_debug.cfs_rq:/.MIN_vruntime.max
201146 ? 30% -86.1% 28034 ? 18% sched_debug.cfs_rq:/.MIN_vruntime.stddev
0.29 ? 8% -43.5% 0.16 ? 9% sched_debug.cfs_rq:/.h_nr_running.avg
0.42 ? 3% -15.1% 0.36 ? 3% sched_debug.cfs_rq:/.h_nr_running.stddev
30967 ? 41% -88.7% 3488 ? 30% sched_debug.cfs_rq:/.max_vruntime.avg
1604384 ? 22% -83.6% 262510 ? 14% sched_debug.cfs_rq:/.max_vruntime.max
201146 ? 30% -86.1% 28034 ? 18% sched_debug.cfs_rq:/.max_vruntime.stddev
1526268 ? 11% -80.6% 295647 ? 10% sched_debug.cfs_rq:/.min_vruntime.avg
2336190 ? 12% -79.4% 482042 ? 7% sched_debug.cfs_rq:/.min_vruntime.max
859659 ? 11% -85.8% 121714 ? 18% sched_debug.cfs_rq:/.min_vruntime.min
343902 ? 15% -78.7% 73263 ? 6% sched_debug.cfs_rq:/.min_vruntime.stddev
0.29 ? 8% -43.6% 0.16 ? 9% sched_debug.cfs_rq:/.nr_running.avg
0.42 ? 3% -15.1% 0.36 ? 3% sched_debug.cfs_rq:/.nr_running.stddev
312.39 ? 7% -46.1% 168.29 ? 5% sched_debug.cfs_rq:/.runnable_avg.avg
355.85 ? 2% -27.3% 258.86 ? 3% sched_debug.cfs_rq:/.runnable_avg.stddev
984535 ? 22% -82.1% 176253 ? 31% sched_debug.cfs_rq:/.spread0.max
-491914 -62.6% -184050 sched_debug.cfs_rq:/.spread0.min
343876 ? 15% -78.7% 73261 ? 6% sched_debug.cfs_rq:/.spread0.stddev
312.25 ? 7% -46.1% 168.23 ? 5% sched_debug.cfs_rq:/.util_avg.avg
355.80 ? 2% -27.3% 258.80 ? 3% sched_debug.cfs_rq:/.util_avg.stddev
213.78 ? 8% -62.3% 80.50 ? 11% sched_debug.cfs_rq:/.util_est_enqueued.avg
331.99 ? 3% -37.4% 207.79 ? 4% sched_debug.cfs_rq:/.util_est_enqueued.stddev
630039 +19.0% 749877 ? 3% sched_debug.cpu.avg_idle.avg
4262 +45.9% 6219 ? 3% sched_debug.cpu.avg_idle.min
2345 ? 14% -35.1% 1522 ? 6% sched_debug.cpu.clock_task.stddev
1496 ? 5% -47.9% 779.72 ? 7% sched_debug.cpu.curr->pid.avg
2528 ? 3% -19.7% 2031 ? 3% sched_debug.cpu.curr->pid.stddev
0.00 ? 10% -23.0% 0.00 ? 4% sched_debug.cpu.next_balance.stddev
0.25 ? 6% -48.0% 0.13 ? 8% sched_debug.cpu.nr_running.avg
0.41 ? 2% -19.9% 0.33 ? 3% sched_debug.cpu.nr_running.stddev
3859540 ? 10% -70.7% 1132195 ? 10% sched_debug.cpu.nr_switches.avg
7113795 ? 11% -64.8% 2504199 ? 10% sched_debug.cpu.nr_switches.max
1618475 ? 15% -86.1% 224980 ? 30% sched_debug.cpu.nr_switches.min
1045624 ? 9% -57.2% 447501 ? 9% sched_debug.cpu.nr_switches.stddev
2.446e+10 -66.9% 8.087e+09 perf-stat.i.branch-instructions
2.356e+08 ? 2% -65.6% 80979566 ? 3% perf-stat.i.branch-misses
0.74 ? 8% +5.7 6.47 perf-stat.i.cache-miss-rate%
16166727 ? 10% +236.6% 54420281 ? 2% perf-stat.i.cache-misses
2.609e+09 -67.5% 8.493e+08 ? 2% perf-stat.i.cache-references
3956687 -69.4% 1211999 perf-stat.i.context-switches
1.23 +44.9% 1.79 perf-stat.i.cpi
1.51e+11 -52.3% 7.199e+10 ? 2% perf-stat.i.cpu-cycles
363.75 -38.7% 223.06 perf-stat.i.cpu-migrations
10558 ? 10% -87.4% 1326 perf-stat.i.cycles-between-cache-misses
652403 ?109% -72.7% 178258 ? 8% perf-stat.i.dTLB-load-misses
3.497e+10 -66.9% 1.158e+10 perf-stat.i.dTLB-loads
0.00 ? 3% +0.0 0.00 ? 11% perf-stat.i.dTLB-store-miss-rate%
119514 ? 5% -51.8% 57642 ? 9% perf-stat.i.dTLB-store-misses
2.067e+10 -66.9% 6.833e+09 perf-stat.i.dTLB-stores
74.98 +0.8 75.82 perf-stat.i.iTLB-load-miss-rate%
1.648e+08 ? 3% -67.2% 54063746 ? 4% perf-stat.i.iTLB-load-misses
54768215 -68.8% 17107737 perf-stat.i.iTLB-loads
1.223e+11 -66.9% 4.052e+10 perf-stat.i.instructions
0.81 -30.7% 0.56 perf-stat.i.ipc
1.05 -52.3% 0.50 ? 2% perf-stat.i.metric.GHz
453.64 -46.1% 244.53 perf-stat.i.metric.K/sec
574.31 -66.9% 189.94 perf-stat.i.metric.M/sec
4984 ? 2% -12.4% 4368 perf-stat.i.minor-faults
4306378 ? 13% +145.4% 10568372 ? 2% perf-stat.i.node-load-misses
267971 ? 20% +205.5% 818656 ? 4% perf-stat.i.node-loads
93.32 +4.7 98.02 perf-stat.i.node-store-miss-rate%
1642200 ? 14% +217.5% 5214268 ? 2% perf-stat.i.node-store-misses
198883 ? 16% -39.3% 120629 ? 9% perf-stat.i.node-stores
4985 ? 2% -12.4% 4369 perf-stat.i.page-faults
21.33 -1.7% 20.96 perf-stat.overall.MPKI
0.62 ? 10% +5.8 6.41 perf-stat.overall.cache-miss-rate%
1.23 +44.0% 1.78 perf-stat.overall.cpi
9426 ? 9% -86.0% 1323 perf-stat.overall.cycles-between-cache-misses
0.00 ? 4% +0.0 0.00 ? 9% perf-stat.overall.dTLB-store-miss-rate%
75.04 +0.9 75.95 perf-stat.overall.iTLB-load-miss-rate%
0.81 -30.5% 0.56 perf-stat.overall.ipc
89.11 +8.6 97.73 perf-stat.overall.node-store-miss-rate%
17633 +8.3% 19103 perf-stat.overall.path-length
2.438e+10 -66.9% 8.061e+09 perf-stat.ps.branch-instructions
2.348e+08 ? 2% -65.6% 80709139 ? 3% perf-stat.ps.branch-misses
16118295 ? 10% +236.5% 54240216 ? 2% perf-stat.ps.cache-misses
2.6e+09 -67.4% 8.464e+08 ? 2% perf-stat.ps.cache-references
3943074 -69.4% 1207953 perf-stat.ps.context-switches
1.505e+11 -52.3% 7.175e+10 ? 2% perf-stat.ps.cpu-cycles
362.72 -38.7% 222.48 perf-stat.ps.cpu-migrations
650701 ?109% -72.7% 177916 ? 8% perf-stat.ps.dTLB-load-misses
3.485e+10 -66.9% 1.155e+10 perf-stat.ps.dTLB-loads
119179 ? 5% -51.8% 57464 ? 9% perf-stat.ps.dTLB-store-misses
2.06e+10 -66.9% 6.81e+09 perf-stat.ps.dTLB-stores
1.642e+08 ? 3% -67.2% 53885104 ? 4% perf-stat.ps.iTLB-load-misses
54583301 -68.8% 17050690 perf-stat.ps.iTLB-loads
1.219e+11 -66.9% 4.039e+10 perf-stat.ps.instructions
4965 ? 2% -12.3% 4352 perf-stat.ps.minor-faults
4293134 ? 13% +145.4% 10534117 ? 2% perf-stat.ps.node-load-misses
267097 ? 20% +205.5% 816047 ? 4% perf-stat.ps.node-loads
1637287 ? 13% +217.4% 5197315 ? 2% perf-stat.ps.node-store-misses
198487 ? 15% -39.3% 120430 ? 9% perf-stat.ps.node-stores
4966 ? 2% -12.3% 4353 perf-stat.ps.page-faults
3.686e+13 -66.8% 1.223e+13 perf-stat.total.instructions
9.05 ? 7% -3.4 5.70 ? 6% perf-profile.calltrace.cycles-pp.sctp_packet_pack.sctp_packet_transmit.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm
8.94 ? 7% -3.2 5.69 ? 5% perf-profile.calltrace.cycles-pp.sctp_do_sm.sctp_primitive_SEND.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg
8.61 ? 7% -3.2 5.44 ? 6% perf-profile.calltrace.cycles-pp.memcpy_erms.sctp_packet_pack.sctp_packet_transmit.sctp_outq_flush.sctp_cmd_interpreter
5.69 ? 4% -3.0 2.68 ? 7% perf-profile.calltrace.cycles-pp.mwait_idle_with_hints.intel_idle_irq.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call
5.85 ? 4% -3.0 2.90 ? 8% perf-profile.calltrace.cycles-pp.intel_idle_irq.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
21.32 ? 6% -2.9 18.43 ? 5% perf-profile.calltrace.cycles-pp.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg.____sys_sendmsg.___sys_sendmsg
13.45 ? 7% -2.5 10.92 ? 5% perf-profile.calltrace.cycles-pp.sctp_primitive_SEND.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg.____sys_sendmsg
7.02 ? 7% -2.5 4.51 ? 6% perf-profile.calltrace.cycles-pp.sctp_datamsg_from_user.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg.____sys_sendmsg
6.53 ? 6% -2.5 4.06 ? 6% perf-profile.calltrace.cycles-pp.skb_copy_datagram_iter.sctp_recvmsg.inet_recvmsg.____sys_recvmsg.___sys_recvmsg
6.51 ? 6% -2.5 4.05 ? 6% perf-profile.calltrace.cycles-pp.__skb_datagram_iter.skb_copy_datagram_iter.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
5.68 ? 6% -2.1 3.54 ? 6% perf-profile.calltrace.cycles-pp._copy_to_iter.__skb_datagram_iter.skb_copy_datagram_iter.sctp_recvmsg.inet_recvmsg
5.52 ? 6% -2.1 3.45 ? 6% perf-profile.calltrace.cycles-pp.copyout._copy_to_iter.__skb_datagram_iter.skb_copy_datagram_iter.sctp_recvmsg
5.48 ? 6% -2.1 3.42 ? 6% perf-profile.calltrace.cycles-pp.copy_user_enhanced_fast_string.copyout._copy_to_iter.__skb_datagram_iter.skb_copy_datagram_iter
11.97 ? 7% -2.0 9.98 ? 5% perf-profile.calltrace.cycles-pp.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm.sctp_primitive_SEND.sctp_sendmsg_to_asoc
9.16 ? 7% -1.8 7.37 ? 5% perf-profile.calltrace.cycles-pp.sctp_cmd_interpreter.sctp_do_sm.sctp_primitive_SEND.sctp_sendmsg_to_asoc.sctp_sendmsg
10.42 ? 7% -1.4 9.00 ? 5% perf-profile.calltrace.cycles-pp.sctp_packet_transmit.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm.sctp_primitive_SEND
1.58 ? 6% -1.2 0.35 ? 70% perf-profile.calltrace.cycles-pp.__alloc_pages.kmalloc_large_node.__kmalloc_node_track_caller.kmalloc_reserve.__alloc_skb
3.49 ? 6% -1.2 2.30 ? 7% perf-profile.calltrace.cycles-pp.sctp_user_addto_chunk.sctp_datamsg_from_user.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg
3.15 ? 6% -1.1 2.06 ? 6% perf-profile.calltrace.cycles-pp._copy_from_iter.sctp_user_addto_chunk.sctp_datamsg_from_user.sctp_sendmsg_to_asoc.sctp_sendmsg
2.86 ? 7% -1.1 1.79 ? 7% perf-profile.calltrace.cycles-pp.sctp_make_datafrag_empty.sctp_datamsg_from_user.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg
3.16 ? 7% -1.1 2.11 ? 5% perf-profile.calltrace.cycles-pp.sctp_data_ready.sctp_ulpq_tail_event.sctp_ulpq_tail_data.sctp_cmd_interpreter.sctp_do_sm
3.00 ? 6% -1.0 1.96 ? 6% perf-profile.calltrace.cycles-pp.copyin._copy_from_iter.sctp_user_addto_chunk.sctp_datamsg_from_user.sctp_sendmsg_to_asoc
2.97 ? 6% -1.0 1.93 ? 6% perf-profile.calltrace.cycles-pp.copy_user_enhanced_fast_string.copyin._copy_from_iter.sctp_user_addto_chunk.sctp_datamsg_from_user
3.26 ? 7% -1.0 2.22 ? 5% perf-profile.calltrace.cycles-pp.sctp_ulpq_tail_event.sctp_ulpq_tail_data.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv
2.89 ? 7% -1.0 1.87 ? 5% perf-profile.calltrace.cycles-pp.__wake_up_common_lock.sctp_data_ready.sctp_ulpq_tail_event.sctp_ulpq_tail_data.sctp_cmd_interpreter
2.68 ? 6% -0.9 1.74 ? 5% perf-profile.calltrace.cycles-pp.__wake_up_common.__wake_up_common_lock.sctp_data_ready.sctp_ulpq_tail_event.sctp_ulpq_tail_data
2.40 ? 7% -0.9 1.46 ? 6% perf-profile.calltrace.cycles-pp._sctp_make_chunk.sctp_make_datafrag_empty.sctp_datamsg_from_user.sctp_sendmsg_to_asoc.sctp_sendmsg
2.49 ? 7% -0.9 1.62 ? 5% perf-profile.calltrace.cycles-pp.autoremove_wake_function.__wake_up_common.__wake_up_common_lock.sctp_data_ready.sctp_ulpq_tail_event
2.46 ? 6% -0.9 1.60 ? 5% perf-profile.calltrace.cycles-pp.try_to_wake_up.autoremove_wake_function.__wake_up_common.__wake_up_common_lock.sctp_data_ready
1.86 ? 6% -0.8 1.10 ? 6% perf-profile.calltrace.cycles-pp.schedule_timeout.sctp_skb_recv_datagram.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
1.81 ? 7% -0.8 1.06 ? 5% perf-profile.calltrace.cycles-pp.__alloc_skb._sctp_make_chunk.sctp_make_datafrag_empty.sctp_datamsg_from_user.sctp_sendmsg_to_asoc
1.76 ? 9% -0.7 1.01 ? 6% perf-profile.calltrace.cycles-pp.sctp_ulpevent_free.sctp_recvmsg.inet_recvmsg.____sys_recvmsg.___sys_recvmsg
1.82 ? 6% -0.7 1.08 ? 6% perf-profile.calltrace.cycles-pp.schedule.schedule_timeout.sctp_skb_recv_datagram.sctp_recvmsg.inet_recvmsg
1.50 ? 6% -0.7 0.76 ? 5% perf-profile.calltrace.cycles-pp.skb_release_data.kfree_skb_reason.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
1.79 ? 6% -0.7 1.06 ? 6% perf-profile.calltrace.cycles-pp.__schedule.schedule.schedule_timeout.sctp_skb_recv_datagram.sctp_recvmsg
1.61 ? 6% -0.6 1.01 ? 6% perf-profile.calltrace.cycles-pp.schedule_idle.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
0.86 ? 7% -0.6 0.26 ?100% perf-profile.calltrace.cycles-pp.dequeue_entity.dequeue_task_fair.__schedule.schedule.schedule_timeout
1.56 ? 6% -0.6 0.98 ? 6% perf-profile.calltrace.cycles-pp.__schedule.schedule_idle.do_idle.cpu_startup_entry.start_secondary
1.56 ? 8% -0.6 0.99 ? 6% perf-profile.calltrace.cycles-pp.sctp_outq_flush_data.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm.sctp_primitive_SEND
1.25 ? 7% -0.6 0.68 ? 5% perf-profile.calltrace.cycles-pp.kmalloc_reserve.__alloc_skb._sctp_make_chunk.sctp_make_datafrag_empty.sctp_datamsg_from_user
1.22 ? 7% -0.6 0.66 ? 5% perf-profile.calltrace.cycles-pp.__kmalloc_node_track_caller.kmalloc_reserve.__alloc_skb._sctp_make_chunk.sctp_make_datafrag_empty
1.19 ? 7% -0.5 0.64 ? 5% perf-profile.calltrace.cycles-pp.kmalloc_large_node.__kmalloc_node_track_caller.kmalloc_reserve.__alloc_skb._sctp_make_chunk
0.90 ? 7% -0.5 0.39 ? 70% perf-profile.calltrace.cycles-pp.skb_release_data.consume_skb.sctp_chunk_put.sctp_outq_sack.sctp_cmd_interpreter
0.94 ? 7% -0.5 0.46 ? 45% perf-profile.calltrace.cycles-pp.dequeue_task_fair.__schedule.schedule.schedule_timeout.sctp_skb_recv_datagram
0.74 ? 6% -0.5 0.27 ?100% perf-profile.calltrace.cycles-pp.sctp_endpoint_lookup_assoc.sctp_sendmsg.sock_sendmsg.____sys_sendmsg.___sys_sendmsg
0.98 ? 8% -0.4 0.57 ? 7% perf-profile.calltrace.cycles-pp.sctp_chunk_put.sctp_ulpevent_free.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
1.05 ? 8% -0.4 0.66 ? 7% perf-profile.calltrace.cycles-pp.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv.sctp_rcv
1.10 ? 7% -0.4 0.72 ? 6% perf-profile.calltrace.cycles-pp.sendmsg_copy_msghdr.___sys_sendmsg.__sys_sendmsg.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.46 ? 45% +0.4 0.83 ? 27% perf-profile.calltrace.cycles-pp.menu_select.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary
0.00 +0.8 0.84 ? 24% perf-profile.calltrace.cycles-pp.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state
0.00 +0.8 0.84 ? 24% perf-profile.calltrace.cycles-pp.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter
0.00 +0.9 0.85 ? 6% perf-profile.calltrace.cycles-pp.page_counter_uncharge.drain_stock.refill_stock.__sk_mem_reduce_allocated.skb_release_head_state
0.00 +0.9 0.87 ? 6% perf-profile.calltrace.cycles-pp.drain_stock.refill_stock.__sk_mem_reduce_allocated.skb_release_head_state.kfree_skb_reason
0.00 +1.0 0.98 ? 7% perf-profile.calltrace.cycles-pp.refill_stock.__sk_mem_reduce_allocated.skb_release_head_state.kfree_skb_reason.sctp_recvmsg
0.00 +1.2 1.24 ? 5% perf-profile.calltrace.cycles-pp.page_counter_uncharge.drain_stock.refill_stock.__sk_mem_reduce_allocated.sctp_wfree
0.00 +1.3 1.26 ? 5% perf-profile.calltrace.cycles-pp.drain_stock.refill_stock.__sk_mem_reduce_allocated.sctp_wfree.skb_release_head_state
0.00 +1.3 1.27 ? 19% perf-profile.calltrace.cycles-pp.try_charge_memcg.mem_cgroup_charge_skmem.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_ulpevent_make_rcvmsg
0.00 +1.3 1.29 ? 26% perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call
1.60 ? 10% +1.3 2.92 ? 5% perf-profile.calltrace.cycles-pp.sctp_ulpevent_make_rcvmsg.sctp_ulpq_tail_data.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv
0.00 +1.4 1.35 ? 5% perf-profile.calltrace.cycles-pp.refill_stock.__sk_mem_reduce_allocated.sctp_wfree.skb_release_head_state.consume_skb
0.00 +1.4 1.42 ? 25% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
0.00 +1.6 1.63 ? 5% perf-profile.calltrace.cycles-pp.mem_cgroup_charge_skmem.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_ulpevent_make_rcvmsg.sctp_ulpq_tail_data
0.00 +1.7 1.66 ? 19% perf-profile.calltrace.cycles-pp.page_counter_try_charge.try_charge_memcg.mem_cgroup_charge_skmem.__sk_mem_raise_allocated.__sk_mem_schedule
10.96 ? 7% +1.7 12.67 ? 5% perf-profile.calltrace.cycles-pp.__ip_queue_xmit.sctp_packet_transmit.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm
0.00 +1.7 1.72 ? 5% perf-profile.calltrace.cycles-pp.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_ulpevent_make_rcvmsg.sctp_ulpq_tail_data.sctp_cmd_interpreter
0.00 +1.7 1.73 ? 5% perf-profile.calltrace.cycles-pp.__sk_mem_schedule.sctp_ulpevent_make_rcvmsg.sctp_ulpq_tail_data.sctp_cmd_interpreter.sctp_do_sm
10.84 ? 7% +1.8 12.59 ? 5% perf-profile.calltrace.cycles-pp.ip_finish_output2.__ip_queue_xmit.sctp_packet_transmit.sctp_outq_flush.sctp_cmd_interpreter
10.63 ? 7% +1.8 12.42 ? 5% perf-profile.calltrace.cycles-pp.__local_bh_enable_ip.ip_finish_output2.__ip_queue_xmit.sctp_packet_transmit.sctp_outq_flush
10.58 ? 7% +1.8 12.40 ? 5% perf-profile.calltrace.cycles-pp.do_softirq.__local_bh_enable_ip.ip_finish_output2.__ip_queue_xmit.sctp_packet_transmit
10.53 ? 7% +1.8 12.36 ? 5% perf-profile.calltrace.cycles-pp.__softirqentry_text_start.do_softirq.__local_bh_enable_ip.ip_finish_output2.__ip_queue_xmit
10.36 ? 7% +1.9 12.25 ? 5% perf-profile.calltrace.cycles-pp.net_rx_action.__softirqentry_text_start.do_softirq.__local_bh_enable_ip.ip_finish_output2
10.26 ? 7% +1.9 12.18 ? 5% perf-profile.calltrace.cycles-pp.__napi_poll.net_rx_action.__softirqentry_text_start.do_softirq.__local_bh_enable_ip
10.23 ? 7% +1.9 12.17 ? 5% perf-profile.calltrace.cycles-pp.process_backlog.__napi_poll.net_rx_action.__softirqentry_text_start.do_softirq
10.07 ? 7% +2.0 12.08 ? 5% perf-profile.calltrace.cycles-pp.__netif_receive_skb_one_core.process_backlog.__napi_poll.net_rx_action.__softirqentry_text_start
0.00 +2.0 2.02 ? 6% perf-profile.calltrace.cycles-pp.try_charge_memcg.mem_cgroup_charge_skmem.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_sendmsg_to_asoc
9.84 ? 7% +2.1 11.95 ? 5% perf-profile.calltrace.cycles-pp.ip_local_deliver_finish.__netif_receive_skb_one_core.process_backlog.__napi_poll.net_rx_action
9.82 ? 7% +2.1 11.94 ? 5% perf-profile.calltrace.cycles-pp.ip_protocol_deliver_rcu.ip_local_deliver_finish.__netif_receive_skb_one_core.process_backlog.__napi_poll
9.74 ? 7% +2.1 11.88 ? 5% perf-profile.calltrace.cycles-pp.sctp_rcv.ip_protocol_deliver_rcu.ip_local_deliver_finish.__netif_receive_skb_one_core.process_backlog
3.38 ? 7% +2.1 5.53 ? 5% perf-profile.calltrace.cycles-pp.sctp_skb_recv_datagram.sctp_recvmsg.inet_recvmsg.____sys_recvmsg.___sys_recvmsg
0.00 +2.4 2.38 ? 6% perf-profile.calltrace.cycles-pp.mem_cgroup_charge_skmem.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_sendmsg_to_asoc.sctp_sendmsg
0.00 +2.5 2.54 ? 6% perf-profile.calltrace.cycles-pp.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg
0.00 +2.5 2.55 ? 6% perf-profile.calltrace.cycles-pp.__sk_mem_schedule.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg.____sys_sendmsg
8.08 ? 7% +2.7 10.74 ? 5% perf-profile.calltrace.cycles-pp.sctp_assoc_bh_rcv.sctp_rcv.ip_protocol_deliver_rcu.ip_local_deliver_finish.__netif_receive_skb_one_core
7.63 ? 7% +2.8 10.44 ? 5% perf-profile.calltrace.cycles-pp.sctp_do_sm.sctp_assoc_bh_rcv.sctp_rcv.ip_protocol_deliver_rcu.ip_local_deliver_finish
15.25 ? 6% +2.9 18.11 ? 6% perf-profile.calltrace.cycles-pp.inet_recvmsg.____sys_recvmsg.___sys_recvmsg.__sys_recvmsg.do_syscall_64
15.15 ? 7% +2.9 18.03 ? 5% perf-profile.calltrace.cycles-pp.release_sock.sctp_sendmsg.sock_sendmsg.____sys_sendmsg.___sys_sendmsg
15.20 ? 6% +2.9 18.09 ? 6% perf-profile.calltrace.cycles-pp.sctp_recvmsg.inet_recvmsg.____sys_recvmsg.___sys_recvmsg.__sys_recvmsg
15.02 ? 7% +2.9 17.94 ? 6% perf-profile.calltrace.cycles-pp.__release_sock.release_sock.sctp_sendmsg.sock_sendmsg.____sys_sendmsg
14.96 ? 7% +2.9 17.90 ? 6% perf-profile.calltrace.cycles-pp.sctp_backlog_rcv.__release_sock.release_sock.sctp_sendmsg.sock_sendmsg
11.70 ? 7% +3.0 14.67 ? 6% perf-profile.calltrace.cycles-pp.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv.sctp_backlog_rcv.__release_sock
0.84 ? 10% +3.1 3.94 ? 5% perf-profile.calltrace.cycles-pp._raw_spin_lock_bh.lock_sock_nested.sctp_skb_recv_datagram.sctp_recvmsg.inet_recvmsg
0.96 ? 9% +3.1 4.06 ? 6% perf-profile.calltrace.cycles-pp.lock_sock_nested.sctp_skb_recv_datagram.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
0.64 ? 11% +3.2 3.82 ? 5% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_bh.lock_sock_nested.sctp_skb_recv_datagram.sctp_recvmsg
0.00 +3.9 3.90 ? 5% perf-profile.calltrace.cycles-pp.sctp_eat_data.sctp_sf_eat_data_6_2.sctp_do_sm.sctp_assoc_bh_rcv.sctp_rcv
0.00 +4.0 3.98 ? 5% perf-profile.calltrace.cycles-pp.sctp_sf_eat_data_6_2.sctp_do_sm.sctp_assoc_bh_rcv.sctp_rcv.ip_protocol_deliver_rcu
2.90 ? 7% +4.2 7.09 ? 6% perf-profile.calltrace.cycles-pp.sctp_outq_sack.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv.sctp_backlog_rcv
1.87 ? 7% +4.5 6.40 ? 6% perf-profile.calltrace.cycles-pp.sctp_chunk_put.sctp_outq_sack.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv
1.43 ? 7% +4.6 6.01 ? 6% perf-profile.calltrace.cycles-pp.consume_skb.sctp_chunk_put.sctp_outq_sack.sctp_cmd_interpreter.sctp_do_sm
1.78 ? 6% +4.6 6.42 ? 6% perf-profile.calltrace.cycles-pp.kfree_skb_reason.sctp_recvmsg.inet_recvmsg.____sys_recvmsg.___sys_recvmsg
0.00 +5.1 5.10 ? 5% perf-profile.calltrace.cycles-pp.__sk_mem_reduce_allocated.sctp_wfree.skb_release_head_state.consume_skb.sctp_chunk_put
0.17 ?141% +5.3 5.42 ? 6% perf-profile.calltrace.cycles-pp.skb_release_head_state.consume_skb.sctp_chunk_put.sctp_outq_sack.sctp_cmd_interpreter
0.00 +5.3 5.35 ? 6% perf-profile.calltrace.cycles-pp.sctp_wfree.skb_release_head_state.consume_skb.sctp_chunk_put.sctp_outq_sack
0.00 +5.5 5.51 ? 6% perf-profile.calltrace.cycles-pp.__sk_mem_reduce_allocated.skb_release_head_state.kfree_skb_reason.sctp_recvmsg.inet_recvmsg
0.00 +5.7 5.65 ? 6% perf-profile.calltrace.cycles-pp.skb_release_head_state.kfree_skb_reason.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
9.38 ? 6% -3.5 5.93 ? 6% perf-profile.children.cycles-pp.sctp_packet_pack
24.70 ? 6% -3.4 21.34 ? 6% perf-profile.children.cycles-pp.sctp_outq_flush
8.83 ? 6% -3.2 5.60 ? 6% perf-profile.children.cycles-pp.copy_user_enhanced_fast_string
8.77 ? 6% -3.2 5.58 ? 6% perf-profile.children.cycles-pp.memcpy_erms
5.88 ? 4% -3.0 2.92 ? 8% perf-profile.children.cycles-pp.intel_idle_irq
21.34 ? 6% -2.8 18.58 ? 6% perf-profile.children.cycles-pp.sctp_sendmsg_to_asoc
7.06 ? 6% -2.5 4.54 ? 7% perf-profile.children.cycles-pp.sctp_datamsg_from_user
13.52 ? 6% -2.5 11.00 ? 6% perf-profile.children.cycles-pp.sctp_primitive_SEND
6.53 ? 6% -2.5 4.06 ? 6% perf-profile.children.cycles-pp.skb_copy_datagram_iter
6.51 ? 6% -2.5 4.05 ? 6% perf-profile.children.cycles-pp.__skb_datagram_iter
5.68 ? 6% -2.1 3.54 ? 6% perf-profile.children.cycles-pp._copy_to_iter
5.52 ? 6% -2.1 3.45 ? 6% perf-profile.children.cycles-pp.copyout
3.33 ? 6% -1.4 1.92 ? 6% perf-profile.children.cycles-pp.__alloc_skb
3.38 ? 6% -1.3 2.07 ? 6% perf-profile.children.cycles-pp.__schedule
3.51 ? 6% -1.2 2.31 ? 7% perf-profile.children.cycles-pp.sctp_user_addto_chunk
2.67 ? 6% -1.2 1.47 ? 6% perf-profile.children.cycles-pp.skb_release_data
2.43 ? 6% -1.1 1.32 ? 6% perf-profile.children.cycles-pp.kmalloc_reserve
3.17 ? 6% -1.1 2.08 ? 7% perf-profile.children.cycles-pp._copy_from_iter
2.38 ? 6% -1.1 1.29 ? 6% perf-profile.children.cycles-pp.__kmalloc_node_track_caller
2.88 ? 7% -1.1 1.80 ? 7% perf-profile.children.cycles-pp.sctp_make_datafrag_empty
3.18 ? 6% -1.1 2.12 ? 6% perf-profile.children.cycles-pp.sctp_data_ready
3.02 ? 6% -1.0 1.98 ? 7% perf-profile.children.cycles-pp.copyin
3.28 ? 6% -1.0 2.25 ? 6% perf-profile.children.cycles-pp.sctp_ulpq_tail_event
2.26 ? 6% -1.0 1.22 ? 5% perf-profile.children.cycles-pp.kmalloc_large_node
2.62 ? 6% -1.0 1.60 ? 7% perf-profile.children.cycles-pp._sctp_make_chunk
2.90 ? 6% -1.0 1.89 ? 6% perf-profile.children.cycles-pp.__wake_up_common_lock
2.69 ? 6% -0.9 1.75 ? 6% perf-profile.children.cycles-pp.__wake_up_common
1.97 ? 6% -0.9 1.04 ? 6% perf-profile.children.cycles-pp.__alloc_pages
2.50 ? 6% -0.9 1.64 ? 6% perf-profile.children.cycles-pp.autoremove_wake_function
2.48 ? 6% -0.9 1.63 ? 6% perf-profile.children.cycles-pp.try_to_wake_up
1.67 ? 5% -0.8 0.86 ? 6% perf-profile.children.cycles-pp.get_page_from_freelist
2.26 ? 7% -0.8 1.48 ? 7% perf-profile.children.cycles-pp.sctp_outq_flush_data
1.86 ? 6% -0.8 1.10 ? 6% perf-profile.children.cycles-pp.schedule_timeout
1.76 ? 8% -0.7 1.02 ? 6% perf-profile.children.cycles-pp.sctp_ulpevent_free
2.08 ? 5% -0.7 1.33 ? 8% perf-profile.children.cycles-pp.kmem_cache_free
1.84 ? 6% -0.7 1.10 ? 6% perf-profile.children.cycles-pp.schedule
1.62 ? 6% -0.6 1.02 ? 6% perf-profile.children.cycles-pp.schedule_idle
1.20 ? 6% -0.6 0.62 ? 7% perf-profile.children.cycles-pp.free_unref_page
1.49 ? 5% -0.5 0.95 ? 6% perf-profile.children.cycles-pp._copy_from_user
1.03 ? 6% -0.5 0.54 ? 5% perf-profile.children.cycles-pp.rmqueue
1.06 ? 6% -0.5 0.58 ? 5% perf-profile.children.cycles-pp._raw_spin_lock_irqsave
1.18 ? 5% -0.4 0.74 ? 6% perf-profile.children.cycles-pp.__check_object_size
1.22 ? 4% -0.4 0.79 ? 8% perf-profile.children.cycles-pp.__rhashtable_lookup
0.97 ? 6% -0.4 0.57 ? 7% perf-profile.children.cycles-pp.__slab_free
0.94 ? 7% -0.4 0.54 ? 7% perf-profile.children.cycles-pp.dequeue_task_fair
1.18 ? 6% -0.4 0.78 ? 7% perf-profile.children.cycles-pp.ttwu_do_activate
0.82 ? 6% -0.4 0.43 ? 7% perf-profile.children.cycles-pp._raw_spin_lock
1.11 ? 6% -0.4 0.73 ? 7% perf-profile.children.cycles-pp.sendmsg_copy_msghdr
1.14 ? 6% -0.4 0.76 ? 7% perf-profile.children.cycles-pp.enqueue_task_fair
0.87 ? 7% -0.4 0.50 ? 7% perf-profile.children.cycles-pp.dequeue_entity
0.92 ? 11% -0.4 0.56 ? 6% perf-profile.children.cycles-pp.sctp_outq_select_transport
1.00 ? 5% -0.4 0.65 ? 6% perf-profile.children.cycles-pp.__copy_msghdr_from_user
0.82 ? 12% -0.3 0.49 ? 6% perf-profile.children.cycles-pp.sctp_packet_config
0.84 ? 6% -0.3 0.54 ? 8% perf-profile.children.cycles-pp.import_iovec
0.84 ? 6% -0.3 0.55 ? 7% perf-profile.children.cycles-pp.pick_next_task_fair
0.80 ? 4% -0.3 0.51 ? 9% perf-profile.children.cycles-pp.kfree
0.50 ? 6% -0.3 0.20 ? 7% perf-profile.children.cycles-pp.free_pcppages_bulk
0.80 ? 6% -0.3 0.51 ? 8% perf-profile.children.cycles-pp.__import_iovec
0.83 ? 6% -0.3 0.54 ? 7% perf-profile.children.cycles-pp.enqueue_entity
0.79 ? 7% -0.3 0.51 ? 7% perf-profile.children.cycles-pp.kmem_cache_alloc
0.70 ? 6% -0.3 0.42 ? 8% perf-profile.children.cycles-pp.update_load_avg
0.78 ? 7% -0.3 0.50 ? 10% perf-profile.children.cycles-pp.sctp_hash_cmp
0.73 ? 5% -0.3 0.47 ? 6% perf-profile.children.cycles-pp.__sctp_rcv_lookup
0.69 ? 6% -0.3 0.44 ? 6% perf-profile.children.cycles-pp.check_heap_object
0.47 ? 6% -0.3 0.22 ? 8% perf-profile.children.cycles-pp.__free_pages_ok
0.70 ? 4% -0.2 0.46 ? 6% perf-profile.children.cycles-pp.sctp_addrs_lookup_transport
0.68 ? 6% -0.2 0.43 ? 8% perf-profile.children.cycles-pp.iovec_from_user
0.71 ? 6% -0.2 0.47 ? 8% perf-profile.children.cycles-pp.sctp_chunkify
0.66 ? 5% -0.2 0.42 ? 8% perf-profile.children.cycles-pp.__might_resched
0.60 ? 5% -0.2 0.36 ? 4% perf-profile.children.cycles-pp.set_next_entity
0.74 ? 6% -0.2 0.51 ? 8% perf-profile.children.cycles-pp.sctp_endpoint_lookup_assoc
0.72 ? 6% -0.2 0.50 ? 9% perf-profile.children.cycles-pp.sctp_epaddr_lookup_transport
0.54 ? 6% -0.2 0.32 ? 10% perf-profile.children.cycles-pp.__dev_queue_xmit
0.44 ? 7% -0.2 0.23 ? 6% perf-profile.children.cycles-pp.sctp_inq_pop
0.61 ? 6% -0.2 0.40 ? 7% perf-profile.children.cycles-pp.copy_user_short_string
0.62 ? 13% -0.2 0.42 ? 5% perf-profile.children.cycles-pp.__skb_clone
0.52 ? 5% -0.2 0.32 ? 7% perf-profile.children.cycles-pp.__might_fault
0.46 ? 5% -0.2 0.27 ? 7% perf-profile.children.cycles-pp.simple_copy_to_iter
0.51 ? 7% -0.2 0.33 ? 4% perf-profile.children.cycles-pp.__entry_text_start
0.36 ? 7% -0.2 0.19 ? 8% perf-profile.children.cycles-pp.update_curr
0.56 ? 6% -0.2 0.38 ? 6% perf-profile.children.cycles-pp.__list_del_entry_valid
0.37 ? 7% -0.2 0.21 ? 5% perf-profile.children.cycles-pp.__mod_node_page_state
0.32 ? 7% -0.2 0.17 ? 8% perf-profile.children.cycles-pp.__zone_watermark_ok
0.42 ? 10% -0.2 0.27 ? 6% perf-profile.children.cycles-pp.select_task_rq
0.46 ? 17% -0.2 0.31 ? 5% perf-profile.children.cycles-pp.__copy_skb_header
0.46 ? 7% -0.1 0.31 ? 10% perf-profile.children.cycles-pp.sctp_packet_transmit_chunk
0.46 ? 7% -0.1 0.32 ? 4% perf-profile.children.cycles-pp.update_rq_clock
0.43 ? 5% -0.1 0.29 ? 10% perf-profile.children.cycles-pp.sctp_addto_chunk
0.41 ? 7% -0.1 0.27 ? 10% perf-profile.children.cycles-pp.sctp_packet_append_chunk
0.34 ? 6% -0.1 0.20 ? 9% perf-profile.children.cycles-pp.dev_hard_start_xmit
0.46 ? 8% -0.1 0.32 ? 6% perf-profile.children.cycles-pp.kmem_cache_alloc_node
0.40 ? 5% -0.1 0.27 ? 5% perf-profile.children.cycles-pp.sockfd_lookup_light
0.38 ? 6% -0.1 0.25 ? 6% perf-profile.children.cycles-pp.sctp_chunk_free
0.24 ? 14% -0.1 0.10 ? 11% perf-profile.children.cycles-pp.sctp_assoc_rwnd_increase
0.35 ? 6% -0.1 0.21 ? 6% perf-profile.children.cycles-pp.__list_add_valid
0.36 ? 7% -0.1 0.22 ? 10% perf-profile.children.cycles-pp.sctp_gen_sack
0.36 ? 5% -0.1 0.23 ? 9% perf-profile.children.cycles-pp.memset_erms
0.35 ? 8% -0.1 0.22 ? 10% perf-profile.children.cycles-pp.memcg_slab_free_hook
0.32 ? 6% -0.1 0.19 ? 9% perf-profile.children.cycles-pp.loopback_xmit
0.36 ? 5% -0.1 0.23 ? 8% perf-profile.children.cycles-pp.sock_kmalloc
0.36 ? 4% -0.1 0.24 ? 7% perf-profile.children.cycles-pp.__might_sleep
0.33 ? 8% -0.1 0.21 ? 7% perf-profile.children.cycles-pp.__virt_addr_valid
0.28 ? 8% -0.1 0.15 ? 8% perf-profile.children.cycles-pp.free_unref_page_commit
0.34 ? 4% -0.1 0.22 ? 6% perf-profile.children.cycles-pp.free_pcp_prepare
0.33 ? 7% -0.1 0.20 ? 11% perf-profile.children.cycles-pp.sctp_make_sack
0.30 ? 13% -0.1 0.18 ? 9% perf-profile.children.cycles-pp.sctp_v4_xmit
0.34 ? 5% -0.1 0.22 ? 6% perf-profile.children.cycles-pp.__fdget
0.35 ? 6% -0.1 0.23 ? 7% perf-profile.children.cycles-pp.move_addr_to_kernel
0.25 ? 5% -0.1 0.13 ? 6% perf-profile.children.cycles-pp.__free_one_page
0.40 ? 9% -0.1 0.29 ? 6% perf-profile.children.cycles-pp.aa_sk_perm
0.24 ? 6% -0.1 0.12 ? 8% perf-profile.children.cycles-pp.prepare_task_switch
0.32 ? 7% -0.1 0.21 ? 6% perf-profile.children.cycles-pp.sctp_check_transmitted
0.30 ? 6% -0.1 0.19 ? 7% perf-profile.children.cycles-pp.sctp_association_put
0.36 ? 7% -0.1 0.24 ? 7% perf-profile.children.cycles-pp.sock_recvmsg
0.26 ? 12% -0.1 0.16 ? 15% perf-profile.children.cycles-pp.ipv4_dst_check
0.22 ? 9% -0.1 0.11 ? 46% perf-profile.children.cycles-pp.accept_connections
0.31 ? 11% -0.1 0.20 ? 9% perf-profile.children.cycles-pp.select_task_rq_fair
0.22 ? 17% -0.1 0.11 ? 6% perf-profile.children.cycles-pp.dst_release
0.23 ? 6% -0.1 0.12 ? 7% perf-profile.children.cycles-pp.rmqueue_bulk
0.22 ? 6% -0.1 0.11 ? 12% perf-profile.children.cycles-pp.sock_wfree
0.28 ? 6% -0.1 0.18 ? 5% perf-profile.children.cycles-pp.send_sctp_stream_1toMany
0.32 ? 7% -0.1 0.23 ? 9% perf-profile.children.cycles-pp.security_socket_recvmsg
0.30 ? 7% -0.1 0.20 ? 9% perf-profile.children.cycles-pp.skb_set_owner_w
0.28 ? 7% -0.1 0.18 ? 8% perf-profile.children.cycles-pp.entry_SYSRETQ_unsafe_stack
0.27 ? 4% -0.1 0.17 ? 9% perf-profile.children.cycles-pp.sock_kfree_s
0.27 ? 10% -0.1 0.18 ? 11% perf-profile.children.cycles-pp.sctp_transport_hold
0.24 ? 6% -0.1 0.15 ? 5% perf-profile.children.cycles-pp.__switch_to
0.24 ? 7% -0.1 0.15 ? 6% perf-profile.children.cycles-pp.tick_nohz_idle_exit
0.21 ? 10% -0.1 0.12 ? 9% perf-profile.children.cycles-pp.recv_sctp_stream_1toMany
0.23 ? 7% -0.1 0.14 ? 7% perf-profile.children.cycles-pp.prepare_to_wait_exclusive
0.22 ? 9% -0.1 0.14 ? 9% perf-profile.children.cycles-pp.accept_connection
0.22 ? 9% -0.1 0.14 ? 9% perf-profile.children.cycles-pp.spawn_child
0.22 ? 9% -0.1 0.14 ? 9% perf-profile.children.cycles-pp.process_requests
0.23 ? 3% -0.1 0.14 ? 7% perf-profile.children.cycles-pp.__kmalloc
0.20 ? 9% -0.1 0.12 ? 10% perf-profile.children.cycles-pp.skb_clone
0.31 ? 10% -0.1 0.22 ? 13% perf-profile.children.cycles-pp.sctp_sched_fcfs_dequeue
0.29 ? 12% -0.1 0.21 ? 5% perf-profile.children.cycles-pp.sctp_transport_put
0.22 ? 8% -0.1 0.14 ? 7% perf-profile.children.cycles-pp.select_idle_sibling
0.22 ? 8% -0.1 0.14 ? 5% perf-profile.children.cycles-pp.__switch_to_asm
0.25 ? 10% -0.1 0.17 ? 11% perf-profile.children.cycles-pp.__sctp_packet_append_chunk
0.23 ? 5% -0.1 0.15 ? 10% perf-profile.children.cycles-pp.__cond_resched
0.20 ? 6% -0.1 0.13 ? 5% perf-profile.children.cycles-pp.available_idle_cpu
0.19 ? 9% -0.1 0.11 ? 6% perf-profile.children.cycles-pp.poll_idle
0.20 ? 6% -0.1 0.13 ? 9% perf-profile.children.cycles-pp.reweight_entity
0.20 ? 7% -0.1 0.13 ? 8% perf-profile.children.cycles-pp.sctp_make_control
0.25 ? 9% -0.1 0.18 ? 8% perf-profile.children.cycles-pp.syscall_exit_to_user_mode
0.20 ? 6% -0.1 0.12 ? 8% perf-profile.children.cycles-pp.__free_pages
0.26 ? 8% -0.1 0.19 ? 9% perf-profile.children.cycles-pp.security_socket_sendmsg
0.19 ? 8% -0.1 0.12 ? 13% perf-profile.children.cycles-pp.syscall_return_via_sysret
0.17 ? 9% -0.1 0.10 ? 14% perf-profile.children.cycles-pp.__mod_timer
0.16 ? 5% -0.1 0.09 ? 7% perf-profile.children.cycles-pp.___perf_sw_event
0.21 ? 6% -0.1 0.14 ? 9% perf-profile.children.cycles-pp.sctp_datamsg_put
0.21 ? 7% -0.1 0.14 ? 7% perf-profile.children.cycles-pp.sctp_association_hold
0.18 ? 10% -0.1 0.11 ? 4% perf-profile.children.cycles-pp.sctp_outq_flush_ctrl
0.16 ? 8% -0.1 0.09 ? 14% perf-profile.children.cycles-pp.__update_load_avg_cfs_rq
0.20 ? 4% -0.1 0.13 ? 9% perf-profile.children.cycles-pp.kmem_cache_alloc_trace
0.19 ? 5% -0.1 0.13 ? 8% perf-profile.children.cycles-pp.__update_load_avg_se
0.16 ? 5% -0.1 0.10 ? 9% perf-profile.children.cycles-pp.del_timer
0.14 ? 11% -0.1 0.08 ? 4% perf-profile.children.cycles-pp.check_new_pages
0.23 ? 7% -0.1 0.17 ? 10% perf-profile.children.cycles-pp.sctp_sched_dequeue_common
0.10 ? 5% -0.1 0.04 ? 71% perf-profile.children.cycles-pp.finish_task_switch
0.18 ? 7% -0.1 0.12 ? 6% perf-profile.children.cycles-pp.sctp_outq_tail
0.12 ? 4% -0.1 0.06 ? 13% perf-profile.children.cycles-pp.check_stack_object
0.13 ? 8% -0.1 0.08 ? 8% perf-profile.children.cycles-pp.__put_user_nocheck_4
0.16 ? 4% -0.1 0.10 ? 13% perf-profile.children.cycles-pp.__genradix_ptr
0.17 ? 6% -0.1 0.11 ? 6% perf-profile.children.cycles-pp.nr_iowait_cpu
0.16 ? 8% -0.1 0.11 ? 13% perf-profile.children.cycles-pp.sctp_sendmsg_parse
0.13 ? 7% -0.1 0.08 ? 12% perf-profile.children.cycles-pp.lock_timer_base
0.08 ? 12% -0.0 0.02 ? 99% perf-profile.children.cycles-pp.ip_local_out
0.14 ? 7% -0.0 0.09 ? 11% perf-profile.children.cycles-pp.memcg_slab_post_alloc_hook
0.14 ? 9% -0.0 0.09 ? 7% perf-profile.children.cycles-pp.switch_mm_irqs_off
0.14 ? 9% -0.0 0.09 ? 10% perf-profile.children.cycles-pp.sctp_datamsg_destroy
0.12 ? 20% -0.0 0.07 ? 21% perf-profile.children.cycles-pp.sctp_hash_key
0.11 ? 20% -0.0 0.06 ? 14% perf-profile.children.cycles-pp.ip_rcv
0.11 ? 8% -0.0 0.07 ? 7% perf-profile.children.cycles-pp.sctp_sock_rfree
0.24 ? 7% -0.0 0.19 ? 6% perf-profile.children.cycles-pp.sched_clock_cpu
0.12 ? 6% -0.0 0.08 ? 11% perf-profile.children.cycles-pp.__netif_rx
0.16 ? 8% -0.0 0.11 ? 10% perf-profile.children.cycles-pp.ttwu_do_wakeup
0.31 ? 6% -0.0 0.27 ? 5% perf-profile.children.cycles-pp.sctp_ulpevent_receive_data
0.13 ? 7% -0.0 0.09 ? 5% perf-profile.children.cycles-pp.pick_next_entity
0.12 ? 6% -0.0 0.08 ? 7% perf-profile.children.cycles-pp.__check_heap_object
0.12 ? 8% -0.0 0.08 ? 12% perf-profile.children.cycles-pp.netif_rx_internal
0.08 ? 10% -0.0 0.03 ? 70% perf-profile.children.cycles-pp.__wrgsbase_inactive
0.22 ? 8% -0.0 0.18 ? 6% perf-profile.children.cycles-pp.read_tsc
0.14 ? 8% -0.0 0.10 ? 9% perf-profile.children.cycles-pp.check_preempt_curr
0.10 ? 9% -0.0 0.06 ? 7% perf-profile.children.cycles-pp.rcu_all_qs
0.10 ? 10% -0.0 0.06 ? 13% perf-profile.children.cycles-pp.validate_xmit_skb
0.14 ? 8% -0.0 0.10 ? 9% perf-profile.children.cycles-pp.perf_trace_sched_wakeup_template
0.11 ? 6% -0.0 0.07 ? 9% perf-profile.children.cycles-pp.enqueue_to_backlog
0.09 ? 5% -0.0 0.05 ? 45% perf-profile.children.cycles-pp.__netif_receive_skb_core
0.12 ? 5% -0.0 0.08 ? 8% perf-profile.children.cycles-pp.__put_user_nocheck_8
0.12 ? 6% -0.0 0.08 ? 14% perf-profile.children.cycles-pp.sctp_chunk_abandoned
0.12 ? 5% -0.0 0.08 ? 8% perf-profile.children.cycles-pp.skb_put
0.20 ? 7% -0.0 0.17 ? 5% perf-profile.children.cycles-pp.native_sched_clock
0.09 ? 5% -0.0 0.05 ? 46% perf-profile.children.cycles-pp.__mod_lruvec_page_state
0.11 ? 9% -0.0 0.08 ? 6% perf-profile.children.cycles-pp.sctp_inet_skb_msgname
0.10 ? 8% -0.0 0.06 perf-profile.children.cycles-pp.tick_nohz_idle_enter
0.09 ? 7% -0.0 0.06 ? 6% perf-profile.children.cycles-pp.os_xsave
0.12 ? 9% -0.0 0.09 ? 8% perf-profile.children.cycles-pp.resched_curr
0.09 ? 10% -0.0 0.06 ? 9% perf-profile.children.cycles-pp.sctp_ulpevent_init
0.14 ? 10% -0.0 0.11 ? 5% perf-profile.children.cycles-pp.update_cfs_group
0.09 ? 10% -0.0 0.06 ? 8% perf-profile.children.cycles-pp.sctp_ulpq_order
0.08 ? 11% -0.0 0.05 ? 44% perf-profile.children.cycles-pp.perf_tp_event
0.07 ? 5% -0.0 0.04 ? 45% perf-profile.children.cycles-pp._raw_spin_unlock_irqrestore
0.10 ? 8% -0.0 0.07 ? 9% perf-profile.children.cycles-pp.syscall_exit_to_user_mode_prepare
0.07 ? 10% -0.0 0.04 ? 45% perf-profile.children.cycles-pp.__cgroup_account_cputime
0.08 ? 6% -0.0 0.06 ? 8% perf-profile.children.cycles-pp.sctp_tsnmap_check
0.00 +0.1 0.06 ? 9% perf-profile.children.cycles-pp.perf_mux_hrtimer_handler
0.00 +0.1 0.07 ? 26% perf-profile.children.cycles-pp.update_sd_lb_stats
0.00 +0.1 0.08 ? 22% perf-profile.children.cycles-pp.find_busiest_group
0.01 ?223% +0.1 0.09 ? 26% perf-profile.children.cycles-pp.load_balance
0.05 ? 46% +0.1 0.13 ? 4% perf-profile.children.cycles-pp.sctp_get_af_specific
0.00 +0.1 0.09 ? 14% perf-profile.children.cycles-pp.lapic_next_deadline
0.00 +0.1 0.09 ? 13% perf-profile.children.cycles-pp.native_irq_return_iret
0.00 +0.1 0.09 ? 39% perf-profile.children.cycles-pp.tsc_verify_tsc_adjust
0.00 +0.1 0.09 ? 31% perf-profile.children.cycles-pp.calc_global_load_tick
0.00 +0.1 0.10 ? 37% perf-profile.children.cycles-pp.arch_cpu_idle_enter
0.00 +0.1 0.10 ? 31% perf-profile.children.cycles-pp._raw_spin_trylock
0.00 +0.1 0.10 ? 25% perf-profile.children.cycles-pp.arch_scale_freq_tick
0.03 ?100% +0.1 0.15 ? 35% perf-profile.children.cycles-pp.rebalance_domains
0.00 +0.1 0.14 ? 21% perf-profile.children.cycles-pp.cgroup_rstat_updated
0.17 ? 15% +0.1 0.32 ? 19% perf-profile.children.cycles-pp.scheduler_tick
0.00 +0.2 0.16 ? 32% perf-profile.children.cycles-pp.tick_nohz_irq_exit
0.24 ? 11% +0.2 0.42 ? 20% perf-profile.children.cycles-pp.update_process_times
0.24 ? 11% +0.2 0.43 ? 20% perf-profile.children.cycles-pp.tick_sched_handle
0.09 ? 11% +0.2 0.29 ? 28% perf-profile.children.cycles-pp.__irq_exit_rcu
0.53 ? 7% +0.2 0.74 ? 16% perf-profile.children.cycles-pp.ktime_get
0.20 ? 10% +0.2 0.42 ? 21% perf-profile.children.cycles-pp.clockevents_program_event
0.06 ? 13% +0.3 0.31 ? 53% perf-profile.children.cycles-pp.timekeeping_max_deferment
0.00 +0.3 0.26 ? 12% perf-profile.children.cycles-pp.propagate_protected_usage
0.28 ? 8% +0.3 0.54 ? 19% perf-profile.children.cycles-pp.tick_sched_timer
0.55 ? 7% +0.3 0.84 ? 26% perf-profile.children.cycles-pp.menu_select
0.29 ? 7% +0.3 0.58 ? 39% perf-profile.children.cycles-pp.tick_nohz_get_sleep_length
0.20 ? 8% +0.3 0.50 ? 44% perf-profile.children.cycles-pp.tick_nohz_next_event
0.32 ? 6% +0.3 0.66 ? 17% perf-profile.children.cycles-pp.__hrtimer_run_queues
0.00 +0.5 0.52 ? 6% perf-profile.children.cycles-pp.mem_cgroup_uncharge_skmem
0.56 ? 7% +0.6 1.16 ? 17% perf-profile.children.cycles-pp.hrtimer_interrupt
0.56 ? 7% +0.6 1.17 ? 17% perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
0.00 +1.0 0.96 ? 4% perf-profile.children.cycles-pp.__mod_memcg_state
0.72 ? 6% +1.0 1.69 ? 20% perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
0.91 ? 6% +1.1 2.05 ? 16% perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
1.62 ? 9% +1.3 2.96 ? 5% perf-profile.children.cycles-pp.sctp_ulpevent_make_rcvmsg
10.76 ? 7% +1.8 12.59 ? 6% perf-profile.children.cycles-pp.__local_bh_enable_ip
10.64 ? 7% +1.9 12.51 ? 6% perf-profile.children.cycles-pp.do_softirq
10.41 ? 7% +2.0 12.37 ? 6% perf-profile.children.cycles-pp.net_rx_action
10.31 ? 7% +2.0 12.30 ? 6% perf-profile.children.cycles-pp.__napi_poll
10.29 ? 7% +2.0 12.30 ? 6% perf-profile.children.cycles-pp.process_backlog
10.66 ? 7% +2.1 12.73 ? 5% perf-profile.children.cycles-pp.__softirqentry_text_start
10.12 ? 7% +2.1 12.20 ? 6% perf-profile.children.cycles-pp.__netif_receive_skb_one_core
0.00 +2.1 2.10 ? 6% perf-profile.children.cycles-pp.page_counter_uncharge
3.40 ? 7% +2.1 5.54 ? 5% perf-profile.children.cycles-pp.sctp_skb_recv_datagram
0.00 +2.1 2.14 ? 5% perf-profile.children.cycles-pp.drain_stock
9.89 ? 7% +2.2 12.07 ? 6% perf-profile.children.cycles-pp.ip_local_deliver_finish
9.87 ? 7% +2.2 12.06 ? 6% perf-profile.children.cycles-pp.ip_protocol_deliver_rcu
0.00 +2.2 2.20 ? 5% perf-profile.children.cycles-pp.page_counter_try_charge
9.80 ? 7% +2.2 12.00 ? 6% perf-profile.children.cycles-pp.sctp_rcv
0.00 +2.4 2.36 ? 6% perf-profile.children.cycles-pp.refill_stock
15.40 ? 6% +2.8 18.21 ? 6% perf-profile.children.cycles-pp.sctp_recvmsg
1.10 ? 8% +2.8 3.94 ? 6% perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
15.25 ? 6% +2.9 18.12 ? 6% perf-profile.children.cycles-pp.inet_recvmsg
15.36 ? 6% +2.9 18.28 ? 6% perf-profile.children.cycles-pp.release_sock
1.36 ? 7% +3.0 4.35 ? 6% perf-profile.children.cycles-pp.lock_sock_nested
1.25 ? 8% +3.0 4.24 ? 6% perf-profile.children.cycles-pp._raw_spin_lock_bh
15.10 ? 6% +3.0 18.11 ? 6% perf-profile.children.cycles-pp.__release_sock
15.04 ? 6% +3.0 18.07 ? 6% perf-profile.children.cycles-pp.sctp_backlog_rcv
0.62 ? 6% +3.4 4.02 ? 6% perf-profile.children.cycles-pp.sctp_sf_eat_data_6_2
0.00 +3.4 3.41 ? 6% perf-profile.children.cycles-pp.try_charge_memcg
0.49 ? 6% +3.4 3.94 ? 6% perf-profile.children.cycles-pp.sctp_eat_data
19.74 ? 6% +3.7 23.44 ? 6% perf-profile.children.cycles-pp.sctp_assoc_bh_rcv
3.40 ? 7% +3.9 7.34 ? 6% perf-profile.children.cycles-pp.sctp_chunk_put
0.00 +4.0 4.04 ? 6% perf-profile.children.cycles-pp.mem_cgroup_charge_skmem
2.92 ? 6% +4.2 7.16 ? 6% perf-profile.children.cycles-pp.sctp_outq_sack
0.00 +4.3 4.29 ? 6% perf-profile.children.cycles-pp.__sk_mem_raise_allocated
0.00 +4.3 4.32 ? 6% perf-profile.children.cycles-pp.__sk_mem_schedule
1.99 ? 6% +4.4 6.40 ? 6% perf-profile.children.cycles-pp.consume_skb
1.78 ? 6% +4.6 6.42 ? 6% perf-profile.children.cycles-pp.kfree_skb_reason
0.37 ? 8% +5.0 5.40 ? 6% perf-profile.children.cycles-pp.sctp_wfree
0.87 ? 9% +10.3 11.20 ? 6% perf-profile.children.cycles-pp.skb_release_head_state
0.00 +10.7 10.66 ? 6% perf-profile.children.cycles-pp.__sk_mem_reduce_allocated
8.78 ? 6% -3.2 5.58 ? 6% perf-profile.self.cycles-pp.copy_user_enhanced_fast_string
8.70 ? 6% -3.2 5.54 ? 6% perf-profile.self.cycles-pp.memcpy_erms
0.96 ? 6% -0.4 0.57 ? 7% perf-profile.self.cycles-pp.__slab_free
0.83 ? 7% -0.3 0.54 ? 6% perf-profile.self.cycles-pp._raw_spin_lock_irqsave
0.70 ? 9% -0.3 0.41 ? 5% perf-profile.self.cycles-pp.sctp_chunk_put
0.83 ? 4% -0.2 0.59 ? 8% perf-profile.self.cycles-pp.kmem_cache_free
0.60 ? 6% -0.2 0.36 ? 8% perf-profile.self.cycles-pp.sctp_recvmsg
0.64 ? 5% -0.2 0.41 ? 7% perf-profile.self.cycles-pp.__might_resched
0.57 ? 6% -0.2 0.36 ? 8% perf-profile.self.cycles-pp._raw_spin_lock_bh
0.38 ? 7% -0.2 0.16 ? 9% perf-profile.self.cycles-pp.sctp_sendmsg_to_asoc
0.62 ? 6% -0.2 0.40 ? 6% perf-profile.self.cycles-pp._raw_spin_lock
0.59 ? 7% -0.2 0.39 ? 5% perf-profile.self.cycles-pp.__schedule
0.58 ? 6% -0.2 0.39 ? 7% perf-profile.self.cycles-pp.copy_user_short_string
0.50 ? 12% -0.2 0.31 ? 8% perf-profile.self.cycles-pp.sctp_packet_config
0.40 ? 8% -0.2 0.23 ? 9% perf-profile.self.cycles-pp.sctp_packet_pack
0.33 ? 8% -0.2 0.16 ? 6% perf-profile.self.cycles-pp.rmqueue
0.45 ? 3% -0.2 0.29 ? 11% perf-profile.self.cycles-pp.kfree
0.45 ? 8% -0.2 0.29 ? 8% perf-profile.self.cycles-pp.sctp_datamsg_from_user
0.32 ? 5% -0.2 0.16 ? 9% perf-profile.self.cycles-pp.get_page_from_freelist
0.37 ? 7% -0.2 0.21 ? 5% perf-profile.self.cycles-pp.__mod_node_page_state
0.52 ? 7% -0.2 0.36 ? 7% perf-profile.self.cycles-pp.__list_del_entry_valid
0.32 ? 6% -0.2 0.16 ? 8% perf-profile.self.cycles-pp.__zone_watermark_ok
0.41 ? 7% -0.2 0.26 ? 8% perf-profile.self.cycles-pp.kmem_cache_alloc
0.46 ? 17% -0.1 0.31 ? 4% perf-profile.self.cycles-pp.__copy_skb_header
0.35 ? 7% -0.1 0.20 ? 7% perf-profile.self.cycles-pp.update_load_avg
0.39 ? 5% -0.1 0.25 ? 9% perf-profile.self.cycles-pp.sctp_outq_flush_data
0.37 ? 7% -0.1 0.23 ? 5% perf-profile.self.cycles-pp.update_rq_clock
0.37 ? 6% -0.1 0.24 ? 7% perf-profile.self.cycles-pp.__skb_datagram_iter
0.23 ? 13% -0.1 0.10 ? 12% perf-profile.self.cycles-pp.sctp_assoc_rwnd_increase
0.38 ? 6% -0.1 0.25 ? 10% perf-profile.self.cycles-pp.sctp_sendmsg
0.35 ? 5% -0.1 0.22 ? 9% perf-profile.self.cycles-pp.memset_erms
0.37 ? 5% -0.1 0.24 ? 5% perf-profile.self.cycles-pp.set_next_entity
0.37 ? 6% -0.1 0.24 ? 6% perf-profile.self.cycles-pp.skb_release_data
0.36 ? 13% -0.1 0.23 ? 7% perf-profile.self.cycles-pp.entry_SYSCALL_64_after_hwframe
0.35 ? 5% -0.1 0.22 ? 7% perf-profile.self.cycles-pp.check_heap_object
0.31 ? 6% -0.1 0.18 ? 14% perf-profile.self.cycles-pp.__rhashtable_lookup
0.34 ? 7% -0.1 0.22 ? 10% perf-profile.self.cycles-pp.memcg_slab_free_hook
0.34 ? 5% -0.1 0.22 ? 6% perf-profile.self.cycles-pp.free_pcp_prepare
0.32 ? 8% -0.1 0.20 ? 7% perf-profile.self.cycles-pp.__virt_addr_valid
0.25 ? 7% -0.1 0.13 ? 10% perf-profile.self.cycles-pp.free_unref_page_commit
0.39 ? 6% -0.1 0.28 ? 9% perf-profile.self.cycles-pp.enqueue_entity
0.32 ? 6% -0.1 0.20 ? 6% perf-profile.self.cycles-pp.__list_add_valid
0.30 ? 13% -0.1 0.18 ? 11% perf-profile.self.cycles-pp.sctp_v4_xmit
0.32 ? 7% -0.1 0.20 ? 7% perf-profile.self.cycles-pp.__alloc_skb
0.33 ? 5% -0.1 0.22 ? 7% perf-profile.self.cycles-pp.__fdget
0.30 ? 6% -0.1 0.19 ? 7% perf-profile.self.cycles-pp.sctp_association_put
0.31 ? 5% -0.1 0.20 ? 7% perf-profile.self.cycles-pp.__might_sleep
0.21 ? 18% -0.1 0.10 ? 7% perf-profile.self.cycles-pp.dst_release
0.26 ? 13% -0.1 0.15 ? 13% perf-profile.self.cycles-pp.ipv4_dst_check
0.31 ? 6% -0.1 0.21 ? 7% perf-profile.self.cycles-pp.sctp_skb_recv_datagram
0.26 ? 9% -0.1 0.16 ? 13% perf-profile.self.cycles-pp.sctp_cmd_interpreter
0.21 ? 6% -0.1 0.11 ? 12% perf-profile.self.cycles-pp.sock_wfree
0.23 ? 7% -0.1 0.13 ? 6% perf-profile.self.cycles-pp.__alloc_pages
0.31 ? 6% -0.1 0.21 ? 9% perf-profile.self.cycles-pp.kmem_cache_alloc_node
0.24 ? 5% -0.1 0.14 ? 10% perf-profile.self.cycles-pp.__check_object_size
0.17 ? 7% -0.1 0.07 ? 12% perf-profile.self.cycles-pp.update_curr
0.18 ? 36% -0.1 0.08 ? 20% perf-profile.self.cycles-pp.sctp_cmp_addr_exact
0.31 ? 6% -0.1 0.22 ? 9% perf-profile.self.cycles-pp.enqueue_task_fair
0.27 ? 5% -0.1 0.18 ? 7% perf-profile.self.cycles-pp.send_sctp_stream_1toMany
0.27 ? 8% -0.1 0.18 ? 8% perf-profile.self.cycles-pp.entry_SYSRETQ_unsafe_stack
0.29 ? 6% -0.1 0.20 ? 8% perf-profile.self.cycles-pp.skb_set_owner_w
0.26 ? 10% -0.1 0.17 ? 11% perf-profile.self.cycles-pp.sctp_transport_hold
0.21 ? 10% -0.1 0.12 ? 8% perf-profile.self.cycles-pp.recv_sctp_stream_1toMany
0.24 ? 7% -0.1 0.15 ? 6% perf-profile.self.cycles-pp.__switch_to
0.25 ? 12% -0.1 0.17 ? 12% perf-profile.self.cycles-pp.sctp_rcv
0.24 ? 4% -0.1 0.16 ? 4% perf-profile.self.cycles-pp.__entry_text_start
0.28 ? 12% -0.1 0.20 ? 4% perf-profile.self.cycles-pp.sctp_transport_put
0.20 ? 5% -0.1 0.11 ? 6% perf-profile.self.cycles-pp.__wake_up_common
0.22 ? 6% -0.1 0.14 ? 5% perf-profile.self.cycles-pp.__switch_to_asm
0.22 ? 5% -0.1 0.14 ? 9% perf-profile.self.cycles-pp.consume_skb
0.21 ? 7% -0.1 0.13 ? 8% perf-profile.self.cycles-pp.recvmsg
0.20 ? 6% -0.1 0.13 ? 9% perf-profile.self.cycles-pp.reweight_entity
0.21 ? 6% -0.1 0.14 ? 10% perf-profile.self.cycles-pp.sctp_datamsg_put
0.20 ? 4% -0.1 0.13 ? 5% perf-profile.self.cycles-pp.available_idle_cpu
0.20 ? 6% -0.1 0.12 ? 6% perf-profile.self.cycles-pp.__free_pages
0.10 ? 16% -0.1 0.02 ? 99% perf-profile.self.cycles-pp.sctp_ulpq_tail_data
0.16 ? 9% -0.1 0.09 ? 14% perf-profile.self.cycles-pp.____sys_recvmsg
0.28 ? 9% -0.1 0.21 ? 7% perf-profile.self.cycles-pp.aa_sk_perm
0.18 ? 8% -0.1 0.11 ? 13% perf-profile.self.cycles-pp.syscall_return_via_sysret
0.10 ? 9% -0.1 0.03 ?100% perf-profile.self.cycles-pp.process_backlog
0.20 ? 7% -0.1 0.14 ? 7% perf-profile.self.cycles-pp.sctp_association_hold
0.20 ? 7% -0.1 0.13 ? 9% perf-profile.self.cycles-pp.sctp_hash_cmp
0.22 ? 8% -0.1 0.16 ? 4% perf-profile.self.cycles-pp.sctp_wfree
0.18 ? 7% -0.1 0.11 ? 6% perf-profile.self.cycles-pp._copy_from_user
0.17 ? 8% -0.1 0.10 ? 6% perf-profile.self.cycles-pp.poll_idle
0.16 ? 8% -0.1 0.09 ? 14% perf-profile.self.cycles-pp.__update_load_avg_cfs_rq
0.19 ? 6% -0.1 0.13 ? 10% perf-profile.self.cycles-pp.__update_load_avg_se
0.14 ? 7% -0.1 0.08 ? 8% perf-profile.self.cycles-pp.___perf_sw_event
0.16 ? 8% -0.1 0.10 ? 10% perf-profile.self.cycles-pp.try_to_wake_up
0.16 ? 5% -0.1 0.10 ? 5% perf-profile.self.cycles-pp.sctp_packet_append_chunk
0.20 ? 8% -0.1 0.14 ? 5% perf-profile.self.cycles-pp.sendmsg
0.14 ? 12% -0.1 0.08 ? 4% perf-profile.self.cycles-pp.check_new_pages
0.17 ? 11% -0.1 0.11 ? 15% perf-profile.self.cycles-pp.sctp_outq_flush
0.17 ? 8% -0.1 0.11 ? 9% perf-profile.self.cycles-pp.sctp_assoc_bh_rcv
0.17 ? 6% -0.1 0.11 ? 5% perf-profile.self.cycles-pp.nr_iowait_cpu
0.17 ? 7% -0.1 0.12 ? 9% perf-profile.self.cycles-pp.sctp_packet_transmit
0.12 ? 9% -0.1 0.06 ? 14% perf-profile.self.cycles-pp.dequeue_entity
0.15 ? 8% -0.1 0.10 ? 13% perf-profile.self.cycles-pp.____sys_sendmsg
0.15 ? 14% -0.1 0.10 ? 8% perf-profile.self.cycles-pp.sctp_ulpevent_free
0.18 ? 9% -0.1 0.12 ? 10% perf-profile.self.cycles-pp.sctp_do_sm
0.14 ? 4% -0.1 0.09 ? 10% perf-profile.self.cycles-pp.__genradix_ptr
0.13 ? 10% -0.1 0.08 ? 6% perf-profile.self.cycles-pp.__put_user_nocheck_4
0.15 ? 7% -0.1 0.10 ? 7% perf-profile.self.cycles-pp.__copy_msghdr_from_user
0.13 ? 10% -0.1 0.08 ? 12% perf-profile.self.cycles-pp.__import_iovec
0.16 ? 4% -0.0 0.11 ? 11% perf-profile.self.cycles-pp.__skb_clone
0.16 ? 8% -0.0 0.11 ? 13% perf-profile.self.cycles-pp.sctp_sendmsg_parse
0.13 ? 6% -0.0 0.08 ? 8% perf-profile.self.cycles-pp.do_syscall_64
0.17 ? 6% -0.0 0.12 ? 5% perf-profile.self.cycles-pp.do_idle
0.12 ? 17% -0.0 0.07 ? 18% perf-profile.self.cycles-pp.sctp_hash_key
0.14 ? 9% -0.0 0.09 ? 9% perf-profile.self.cycles-pp.sctp_check_transmitted
0.13 ? 5% -0.0 0.08 ? 7% perf-profile.self.cycles-pp.sock_kmalloc
0.11 ? 6% -0.0 0.06 ? 7% perf-profile.self.cycles-pp.sctp_sock_rfree
0.11 ? 6% -0.0 0.07 ? 7% perf-profile.self.cycles-pp.__kmalloc_node_track_caller
0.12 ? 9% -0.0 0.07 ? 10% perf-profile.self.cycles-pp.___sys_recvmsg
0.14 ? 8% -0.0 0.09 ? 7% perf-profile.self.cycles-pp.switch_mm_irqs_off
0.07 ? 10% -0.0 0.03 ?100% perf-profile.self.cycles-pp._sctp_make_chunk
0.22 ? 8% -0.0 0.18 ? 4% perf-profile.self.cycles-pp.read_tsc
0.12 ? 5% -0.0 0.07 ? 6% perf-profile.self.cycles-pp.__put_user_nocheck_8
0.13 ? 8% -0.0 0.08 ? 8% perf-profile.self.cycles-pp.sctp_inq_pop
0.12 ? 6% -0.0 0.08 ? 7% perf-profile.self.cycles-pp.__check_heap_object
0.12 ? 4% -0.0 0.08 ? 8% perf-profile.self.cycles-pp.__kmalloc
0.08 ? 8% -0.0 0.04 ? 71% perf-profile.self.cycles-pp.__mod_lruvec_page_state
0.12 ? 8% -0.0 0.08 ? 6% perf-profile.self.cycles-pp.__local_bh_enable_ip
0.12 ? 7% -0.0 0.08 ? 8% perf-profile.self.cycles-pp.sctp_sf_eat_data_6_2
0.12 ? 6% -0.0 0.07 ? 10% perf-profile.self.cycles-pp.__free_pages_ok
0.11 ? 8% -0.0 0.07 ? 7% perf-profile.self.cycles-pp._copy_to_iter
0.11 ? 6% -0.0 0.06 ? 7% perf-profile.self.cycles-pp.__ip_queue_xmit
0.08 ? 10% -0.0 0.03 ? 70% perf-profile.self.cycles-pp.__wrgsbase_inactive
0.13 ? 8% -0.0 0.09 ? 5% perf-profile.self.cycles-pp.pick_next_entity
0.11 ? 8% -0.0 0.07 ? 7% perf-profile.self.cycles-pp.select_task_rq
0.11 ? 7% -0.0 0.07 ? 8% perf-profile.self.cycles-pp.sctp_inet_skb_msgname
0.10 ? 9% -0.0 0.06 ? 11% perf-profile.self.cycles-pp.sctp_outq_sack
0.08 ? 5% -0.0 0.05 ? 45% perf-profile.self.cycles-pp.__netif_receive_skb_core
0.20 ? 8% -0.0 0.16 ? 5% perf-profile.self.cycles-pp.native_sched_clock
0.12 ? 9% -0.0 0.08 ? 10% perf-profile.self.cycles-pp.memcg_slab_post_alloc_hook
0.10 ? 7% -0.0 0.06 ? 14% perf-profile.self.cycles-pp.__cond_resched
0.12 ? 4% -0.0 0.08 ? 12% perf-profile.self.cycles-pp.kmem_cache_alloc_trace
0.10 ? 5% -0.0 0.06 ? 15% perf-profile.self.cycles-pp.check_stack_object
0.12 ? 6% -0.0 0.08 ? 7% perf-profile.self.cycles-pp.sctp_outq_tail
0.12 ? 8% -0.0 0.08 ? 12% perf-profile.self.cycles-pp.sctp_chunk_abandoned
0.10 ? 4% -0.0 0.06 ? 14% perf-profile.self.cycles-pp.iovec_from_user
0.09 ? 7% -0.0 0.06 ? 6% perf-profile.self.cycles-pp.os_xsave
0.13 ? 7% -0.0 0.09 ? 11% perf-profile.self.cycles-pp.sctp_sched_dequeue_common
0.07 ? 11% -0.0 0.04 ? 71% perf-profile.self.cycles-pp.move_addr_to_kernel
0.09 ? 10% -0.0 0.05 ? 8% perf-profile.self.cycles-pp.sctp_ulpevent_init
0.08 ? 10% -0.0 0.05 ? 45% perf-profile.self.cycles-pp.free_unref_page
0.10 ? 5% -0.0 0.06 ? 11% perf-profile.self.cycles-pp.__might_fault
0.11 ? 12% -0.0 0.08 ? 10% perf-profile.self.cycles-pp.__sctp_packet_append_chunk
0.07 ? 11% -0.0 0.03 ? 70% perf-profile.self.cycles-pp.sockfd_lookup_light
0.11 ? 7% -0.0 0.08 ? 9% perf-profile.self.cycles-pp.cpuidle_idle_call
0.12 ? 7% -0.0 0.09 ? 10% perf-profile.self.cycles-pp.ip_finish_output2
0.12 ? 8% -0.0 0.09 ? 8% perf-profile.self.cycles-pp.resched_curr
0.10 ? 5% -0.0 0.07 ? 10% perf-profile.self.cycles-pp.skb_put
0.09 ? 7% -0.0 0.06 ? 6% perf-profile.self.cycles-pp.sctp_outq_select_transport
0.08 ? 6% -0.0 0.04 ? 45% perf-profile.self.cycles-pp.__sys_recvmsg
0.08 ? 5% -0.0 0.05 ? 8% perf-profile.self.cycles-pp.sock_kfree_s
0.10 ? 5% -0.0 0.07 ? 14% perf-profile.self.cycles-pp.net_rx_action
0.09 ? 8% -0.0 0.06 ? 11% perf-profile.self.cycles-pp.sctp_chunkify
0.10 ? 9% -0.0 0.07 ? 10% perf-profile.self.cycles-pp.syscall_exit_to_user_mode_prepare
0.14 ? 10% -0.0 0.11 ? 5% perf-profile.self.cycles-pp.update_cfs_group
0.09 ? 8% -0.0 0.06 ? 9% perf-profile.self.cycles-pp.security_socket_recvmsg
0.08 ? 9% -0.0 0.05 perf-profile.self.cycles-pp.rcu_all_qs
0.07 ? 9% -0.0 0.04 ? 45% perf-profile.self.cycles-pp.sctp_tsnmap_check
0.08 ? 11% -0.0 0.06 ? 13% perf-profile.self.cycles-pp.sctp_addto_chunk
0.08 ? 10% -0.0 0.06 ? 13% perf-profile.self.cycles-pp.kmalloc_large_node
0.10 ? 8% +0.0 0.12 ? 6% perf-profile.self.cycles-pp.lock_sock_nested
0.00 +0.1 0.06 ? 9% perf-profile.self.cycles-pp.mem_cgroup_uncharge_skmem
0.00 +0.1 0.08 ? 11% perf-profile.self.cycles-pp.lapic_next_deadline
0.00 +0.1 0.09 ? 13% perf-profile.self.cycles-pp.native_irq_return_iret
0.04 ? 71% +0.1 0.12 ? 7% perf-profile.self.cycles-pp.sctp_get_af_specific
0.00 +0.1 0.09 ? 41% perf-profile.self.cycles-pp.tsc_verify_tsc_adjust
0.00 +0.1 0.09 ? 31% perf-profile.self.cycles-pp.calc_global_load_tick
0.00 +0.1 0.10 ? 31% perf-profile.self.cycles-pp._raw_spin_trylock
0.01 ?223% +0.1 0.10 ? 50% perf-profile.self.cycles-pp.tick_nohz_next_event
0.11 ? 9% +0.1 0.21 ? 10% perf-profile.self.cycles-pp.cpuidle_enter_state
0.00 +0.1 0.10 ? 25% perf-profile.self.cycles-pp.arch_scale_freq_tick
0.00 +0.1 0.13 ? 19% perf-profile.self.cycles-pp.cgroup_rstat_updated
0.00 +0.1 0.14 ? 10% perf-profile.self.cycles-pp.mem_cgroup_charge_skmem
0.00 +0.2 0.21 ? 8% perf-profile.self.cycles-pp.refill_stock
0.00 +0.2 0.25 ? 9% perf-profile.self.cycles-pp.__sk_mem_raise_allocated
0.31 ? 8% +0.3 0.56 ? 23% perf-profile.self.cycles-pp.ktime_get
0.06 ? 13% +0.3 0.31 ? 53% perf-profile.self.cycles-pp.timekeeping_max_deferment
0.00 +0.3 0.26 ? 11% perf-profile.self.cycles-pp.propagate_protected_usage
0.00 +0.8 0.82 ? 3% perf-profile.self.cycles-pp.__mod_memcg_state
0.00 +1.2 1.19 ? 7% perf-profile.self.cycles-pp.try_charge_memcg
0.00 +2.0 1.96 ? 6% perf-profile.self.cycles-pp.page_counter_uncharge
0.00 +2.1 2.07 ? 5% perf-profile.self.cycles-pp.page_counter_try_charge
1.09 ? 8% +2.8 3.92 ? 6% perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
0.29 ? 6% +3.5 3.81 ? 6% perf-profile.self.cycles-pp.sctp_eat_data
0.00 +7.8 7.76 ? 6% perf-profile.self.cycles-pp.__sk_mem_reduce_allocated


***************************************************************************************************
lkp-cpl-4sp1: 144 threads 4 sockets Intel(R) Xeon(R) Gold 5318H CPU @ 2.50GHz with 128G memory
=========================================================================================
cluster/compiler/cpufreq_governor/ip/kconfig/nr_threads/rootfs/runtime/send_size/tbox_group/test/testcase/ucode:
cs-localhost/gcc-11/performance/ipv4/x86_64-rhel-8.3/50%/debian-11.1-x86_64-20220510.cgz/300s/10K/lkp-cpl-4sp1/SCTP_STREAM_MANY/netperf/0x7002402

commit:
7c80b038d2 ("net: fix sk_wmem_schedule() and sk_rmem_schedule() errors")
4890b686f4 ("net: keep sk->sk_forward_alloc as small as possible")

7c80b038d23e1f4c 4890b686f4088c90432149bd6de
---------------- ---------------------------
%stddev %change %stddev
\ | \
9985 -73.7% 2624 netperf.Throughput_Mbps
718950 -73.7% 188942 netperf.Throughput_total_Mbps
68186 -71.8% 19239 ? 2% netperf.time.involuntary_context_switches
29139 ? 2% -16.8% 24247 netperf.time.minor_page_faults
3137 -68.1% 1000 netperf.time.percent_of_cpu_this_job_got
9264 -67.8% 2979 netperf.time.system_time
199.50 -77.7% 44.44 ? 2% netperf.time.user_time
316769 ? 2% +21.5% 384840 netperf.time.voluntary_context_switches
2.633e+09 -73.7% 6.919e+08 netperf.workload
29358 +42.4% 41821 uptime.idle
2.381e+10 +52.8% 3.639e+10 cpuidle..time
7.884e+08 -65.8% 2.694e+08 cpuidle..usage
56.10 +27.9 83.96 mpstat.cpu.all.idle%
0.80 -0.2 0.61 ? 5% mpstat.cpu.all.irq%
7.50 -4.2 3.29 mpstat.cpu.all.soft%
34.45 -22.6 11.86 mpstat.cpu.all.sys%
1.14 -0.9 0.27 mpstat.cpu.all.usr%
55.83 +49.3% 83.33 vmstat.cpu.id
5727924 ? 3% -43.6% 3229130 vmstat.memory.cache
64.67 ? 2% -65.2% 22.50 ? 2% vmstat.procs.r
4915230 -73.6% 1295498 vmstat.system.cs
298596 -3.1% 289395 vmstat.system.in
92833 ?101% -63.0% 34380 ? 35% numa-meminfo.node1.Inactive
92773 ?101% -62.9% 34380 ? 35% numa-meminfo.node1.Inactive(anon)
1468231 ? 19% -74.4% 376377 ? 16% numa-meminfo.node3.Active
1468231 ? 19% -74.4% 376377 ? 16% numa-meminfo.node3.Active(anon)
925228 ? 17% -95.4% 42887 ? 27% numa-meminfo.node3.Inactive
925228 ? 17% -95.4% 42887 ? 27% numa-meminfo.node3.Inactive(anon)
1162413 ? 17% -85.6% 167082 ? 21% numa-meminfo.node3.Mapped
2348690 ? 14% -83.8% 380889 ? 16% numa-meminfo.node3.Shmem
8.422e+08 -72.9% 2.286e+08 ? 5% numa-numastat.node0.local_node
8.425e+08 -72.8% 2.288e+08 ? 5% numa-numastat.node0.numa_hit
8.344e+08 -73.3% 2.226e+08 ? 5% numa-numastat.node1.local_node
8.341e+08 -73.3% 2.227e+08 ? 5% numa-numastat.node1.numa_hit
8.584e+08 -74.8% 2.163e+08 ? 10% numa-numastat.node2.local_node
8.583e+08 -74.8% 2.164e+08 ? 10% numa-numastat.node2.numa_hit
8.684e+08 -74.1% 2.251e+08 ? 6% numa-numastat.node3.local_node
8.681e+08 -74.1% 2.252e+08 ? 6% numa-numastat.node3.numa_hit
1860369 ? 5% -76.2% 442423 ? 3% meminfo.Active
1860257 ? 5% -76.2% 442314 ? 3% meminfo.Active(anon)
5599254 ? 3% -44.7% 3099170 meminfo.Cached
3542743 ? 6% -71.5% 1010644 meminfo.Committed_AS
1423628 ? 11% -76.4% 336110 meminfo.Inactive
1423448 ? 11% -76.4% 335928 meminfo.Inactive(anon)
1453354 ? 6% -83.5% 240203 ? 3% meminfo.Mapped
7494321 ? 2% -34.5% 4912267 meminfo.Memused
12458 -18.9% 10104 meminfo.PageTables
2948963 ? 7% -84.8% 449291 ? 3% meminfo.Shmem
11063817 -30.5% 7685444 meminfo.max_used_kB
8.425e+08 -72.8% 2.288e+08 ? 5% numa-vmstat.node0.numa_hit
8.422e+08 -72.9% 2.286e+08 ? 5% numa-vmstat.node0.numa_local
23213 ?101% -63.0% 8583 ? 35% numa-vmstat.node1.nr_inactive_anon
23213 ?101% -63.0% 8583 ? 35% numa-vmstat.node1.nr_zone_inactive_anon
8.341e+08 -73.3% 2.227e+08 ? 5% numa-vmstat.node1.numa_hit
8.344e+08 -73.3% 2.226e+08 ? 5% numa-vmstat.node1.numa_local
8.583e+08 -74.8% 2.164e+08 ? 10% numa-vmstat.node2.numa_hit
8.584e+08 -74.8% 2.163e+08 ? 10% numa-vmstat.node2.numa_local
366556 ? 19% -74.2% 94616 ? 16% numa-vmstat.node3.nr_active_anon
231387 ? 17% -95.4% 10658 ? 27% numa-vmstat.node3.nr_inactive_anon
291532 ? 17% -86.8% 38558 ? 21% numa-vmstat.node3.nr_mapped
586771 ? 14% -83.7% 95702 ? 16% numa-vmstat.node3.nr_shmem
366556 ? 19% -74.2% 94616 ? 16% numa-vmstat.node3.nr_zone_active_anon
231387 ? 17% -95.4% 10658 ? 27% numa-vmstat.node3.nr_zone_inactive_anon
8.681e+08 -74.1% 2.252e+08 ? 6% numa-vmstat.node3.numa_hit
8.684e+08 -74.1% 2.251e+08 ? 6% numa-vmstat.node3.numa_local
1495 -61.8% 571.00 turbostat.Avg_MHz
47.16 -29.8 17.36 turbostat.Busy%
3178 +3.7% 3295 turbostat.Bzy_MHz
7.464e+08 -73.4% 1.985e+08 turbostat.C1
13.74 -8.9 4.80 ? 2% turbostat.C1%
34154266 ? 14% +97.4% 67405025 turbostat.C1E
36.80 ? 21% +37.1 73.90 ? 3% turbostat.C1E%
52.81 +56.4% 82.60 turbostat.CPU%c1
58.83 ? 2% -20.1% 47.00 ? 2% turbostat.CoreTmp
0.20 -25.6% 0.15 turbostat.IPC
128.56 ? 3% -128.6 0.00 turbostat.PKG_%
5447540 ? 3% -83.7% 887322 turbostat.POLL
0.13 -0.1 0.03 turbostat.POLL%
58.83 ? 2% -19.8% 47.17 ? 2% turbostat.PkgTmp
591.77 -28.1% 425.61 turbostat.PkgWatt
8.17 +3.5% 8.46 turbostat.RAMWatt
465173 ? 5% -76.3% 110436 ? 3% proc-vmstat.nr_active_anon
83691 -1.7% 82260 proc-vmstat.nr_anon_pages
3086038 +2.1% 3150528 proc-vmstat.nr_dirty_background_threshold
6179622 +2.1% 6308761 proc-vmstat.nr_dirty_threshold
1399831 ? 3% -44.7% 774653 proc-vmstat.nr_file_pages
31038779 +2.1% 31684629 proc-vmstat.nr_free_pages
355768 ? 11% -76.4% 83981 proc-vmstat.nr_inactive_anon
363110 ? 6% -83.4% 60381 ? 3% proc-vmstat.nr_mapped
3114 -18.9% 2526 proc-vmstat.nr_page_table_pages
737257 ? 7% -84.8% 112182 ? 3% proc-vmstat.nr_shmem
33546 -3.9% 32233 proc-vmstat.nr_slab_reclaimable
465173 ? 5% -76.3% 110436 ? 3% proc-vmstat.nr_zone_active_anon
355768 ? 11% -76.4% 83981 proc-vmstat.nr_zone_inactive_anon
125868 ? 15% -29.5% 88747 ? 12% proc-vmstat.numa_hint_faults_local
3.403e+09 -73.8% 8.93e+08 proc-vmstat.numa_hit
3.403e+09 -73.8% 8.927e+08 proc-vmstat.numa_local
1698265 ? 8% -91.3% 146940 ? 2% proc-vmstat.pgactivate
9.031e+09 -73.7% 2.375e+09 proc-vmstat.pgalloc_normal
1809374 ? 2% -11.5% 1601059 ? 2% proc-vmstat.pgfault
9.031e+09 -73.7% 2.375e+09 proc-vmstat.pgfree
40961 +1.3% 41503 proc-vmstat.pgreuse
121388 ? 29% -96.2% 4632 ? 47% sched_debug.cfs_rq:/.MIN_vruntime.avg
3776647 ? 10% -91.9% 305894 ? 40% sched_debug.cfs_rq:/.MIN_vruntime.max
623786 ? 18% -94.3% 35468 ? 44% sched_debug.cfs_rq:/.MIN_vruntime.stddev
0.40 ? 7% -58.4% 0.17 ? 7% sched_debug.cfs_rq:/.h_nr_running.avg
0.45 -19.1% 0.37 ? 3% sched_debug.cfs_rq:/.h_nr_running.stddev
121388 ? 29% -96.2% 4632 ? 47% sched_debug.cfs_rq:/.max_vruntime.avg
3776647 ? 10% -91.9% 305894 ? 40% sched_debug.cfs_rq:/.max_vruntime.max
623786 ? 18% -94.3% 35468 ? 44% sched_debug.cfs_rq:/.max_vruntime.stddev
3455794 ? 6% -89.2% 374393 ? 9% sched_debug.cfs_rq:/.min_vruntime.avg
4153720 ? 7% -88.1% 494762 ? 7% sched_debug.cfs_rq:/.min_vruntime.max
2710531 ? 7% -91.1% 240412 ? 10% sched_debug.cfs_rq:/.min_vruntime.min
368488 ? 8% -85.9% 51911 ? 13% sched_debug.cfs_rq:/.min_vruntime.stddev
0.40 ? 7% -58.4% 0.17 ? 7% sched_debug.cfs_rq:/.nr_running.avg
0.45 -19.2% 0.37 ? 3% sched_debug.cfs_rq:/.nr_running.stddev
386.20 ? 3% -56.3% 168.94 sched_debug.cfs_rq:/.runnable_avg.avg
1215 ? 8% -18.7% 987.82 ? 4% sched_debug.cfs_rq:/.runnable_avg.max
340.61 -31.1% 234.69 ? 2% sched_debug.cfs_rq:/.runnable_avg.stddev
429174 ? 24% -103.8% -16476 sched_debug.cfs_rq:/.spread0.avg
1127012 ? 15% -90.8% 103881 ? 28% sched_debug.cfs_rq:/.spread0.max
-315952 -52.4% -150458 sched_debug.cfs_rq:/.spread0.min
368407 ? 8% -85.9% 51908 ? 13% sched_debug.cfs_rq:/.spread0.stddev
385.96 ? 3% -56.3% 168.81 sched_debug.cfs_rq:/.util_avg.avg
1215 ? 8% -18.7% 987.79 ? 4% sched_debug.cfs_rq:/.util_avg.max
340.53 -31.1% 234.60 ? 2% sched_debug.cfs_rq:/.util_avg.stddev
274.31 ? 7% -74.9% 68.97 ? 5% sched_debug.cfs_rq:/.util_est_enqueued.avg
981.12 -10.6% 877.51 ? 3% sched_debug.cfs_rq:/.util_est_enqueued.max
337.06 -47.8% 176.03 ? 3% sched_debug.cfs_rq:/.util_est_enqueued.stddev
457832 ? 5% +41.2% 646422 ? 5% sched_debug.cpu.avg_idle.avg
1762 ? 7% -42.8% 1009 ? 7% sched_debug.cpu.clock_task.stddev
2358 ? 5% -61.0% 920.92 ? 2% sched_debug.cpu.curr->pid.avg
2793 -22.1% 2176 ? 2% sched_debug.cpu.curr->pid.stddev
557670 ? 6% +8.4% 604409 ? 7% sched_debug.cpu.max_idle_balance_cost.max
0.00 ? 2% -27.5% 0.00 ? 5% sched_debug.cpu.next_balance.stddev
0.39 ? 5% -61.1% 0.15 ? 3% sched_debug.cpu.nr_running.avg
0.45 -22.2% 0.35 sched_debug.cpu.nr_running.stddev
5005510 ? 7% -75.6% 1221805 ? 10% sched_debug.cpu.nr_switches.avg
6928957 ? 8% -70.7% 2030568 ? 9% sched_debug.cpu.nr_switches.max
3447354 ? 12% -84.3% 541402 ? 17% sched_debug.cpu.nr_switches.min
629520 ? 7% -55.8% 278525 ? 6% sched_debug.cpu.nr_switches.stddev
3.074e+10 -71.6% 8.745e+09 perf-stat.i.branch-instructions
2.971e+08 ? 2% -70.3% 88205513 ? 4% perf-stat.i.branch-misses
1.05 ? 3% +5.6 6.66 perf-stat.i.cache-miss-rate%
30824055 ? 3% +98.8% 61267408 perf-stat.i.cache-misses
3.297e+09 -71.8% 9.296e+08 perf-stat.i.cache-references
4957450 -73.6% 1308421 perf-stat.i.context-switches
1.43 +33.1% 1.90 perf-stat.i.cpi
2.196e+11 -62.4% 8.259e+10 perf-stat.i.cpu-cycles
929.67 -64.3% 331.55 perf-stat.i.cpu-migrations
7494 ? 4% -82.0% 1351 perf-stat.i.cycles-between-cache-misses
0.01 ? 26% -0.0 0.00 perf-stat.i.dTLB-load-miss-rate%
2694006 ? 27% -90.1% 265691 perf-stat.i.dTLB-load-misses
4.398e+10 -71.5% 1.252e+10 perf-stat.i.dTLB-loads
0.00 ? 7% -0.0 0.00 ? 5% perf-stat.i.dTLB-store-miss-rate%
967059 ? 8% -84.3% 151835 ? 6% perf-stat.i.dTLB-store-misses
2.599e+10 -71.6% 7.377e+09 perf-stat.i.dTLB-stores
63.87 +10.9 74.76 perf-stat.i.iTLB-load-miss-rate%
1.972e+08 -70.5% 58088811 ? 5% perf-stat.i.iTLB-load-misses
1.126e+08 -82.7% 19440392 perf-stat.i.iTLB-loads
1.538e+11 -71.5% 4.38e+10 perf-stat.i.instructions
0.71 -24.8% 0.53 perf-stat.i.ipc
1.53 -62.4% 0.57 perf-stat.i.metric.GHz
909.37 -69.9% 273.61 perf-stat.i.metric.K/sec
722.30 -71.6% 205.33 perf-stat.i.metric.M/sec
5562 ? 3% -12.4% 4874 ? 2% perf-stat.i.minor-faults
8844727 ? 5% +30.4% 11534796 perf-stat.i.node-load-misses
623949 ? 11% +54.9% 966213 ? 2% perf-stat.i.node-loads
91.43 +5.4 96.84 perf-stat.i.node-store-miss-rate%
3461624 ? 2% +66.0% 5746681 perf-stat.i.node-store-misses
410920 ? 7% -49.7% 206502 ? 3% perf-stat.i.node-stores
5563 ? 3% -12.4% 4876 ? 2% perf-stat.i.page-faults
21.44 -1.0% 21.22 perf-stat.overall.MPKI
0.93 ? 4% +5.7 6.59 perf-stat.overall.cache-miss-rate%
1.43 +32.0% 1.89 perf-stat.overall.cpi
7136 ? 4% -81.1% 1348 perf-stat.overall.cycles-between-cache-misses
0.01 ? 27% -0.0 0.00 perf-stat.overall.dTLB-load-miss-rate%
0.00 ? 7% -0.0 0.00 ? 5% perf-stat.overall.dTLB-store-miss-rate%
63.66 +11.2 74.89 perf-stat.overall.iTLB-load-miss-rate%
0.70 -24.3% 0.53 perf-stat.overall.ipc
89.40 +7.1 96.53 perf-stat.overall.node-store-miss-rate%
17593 +8.6% 19108 perf-stat.overall.path-length
3.063e+10 -71.5% 8.716e+09 perf-stat.ps.branch-instructions
2.961e+08 ? 2% -70.3% 87907848 ? 4% perf-stat.ps.branch-misses
30717141 ? 3% +98.8% 61062871 perf-stat.ps.cache-misses
3.286e+09 -71.8% 9.265e+08 perf-stat.ps.cache-references
4940513 -73.6% 1304046 perf-stat.ps.context-switches
2.189e+11 -62.4% 8.231e+10 perf-stat.ps.cpu-cycles
926.60 -64.3% 330.61 perf-stat.ps.cpu-migrations
2684405 ? 27% -90.1% 264976 perf-stat.ps.dTLB-load-misses
4.383e+10 -71.5% 1.248e+10 perf-stat.ps.dTLB-loads
963689 ? 8% -84.3% 151363 ? 6% perf-stat.ps.dTLB-store-misses
2.591e+10 -71.6% 7.352e+09 perf-stat.ps.dTLB-stores
1.965e+08 -70.5% 57894987 ? 5% perf-stat.ps.iTLB-load-misses
1.122e+08 -82.7% 19375491 perf-stat.ps.iTLB-loads
1.532e+11 -71.5% 4.366e+10 perf-stat.ps.instructions
5536 ? 3% -12.3% 4857 ? 2% perf-stat.ps.minor-faults
8813664 ? 5% +30.4% 11496729 perf-stat.ps.node-load-misses
622201 ? 11% +54.8% 963024 ? 2% perf-stat.ps.node-loads
3450020 ? 2% +66.0% 5727738 perf-stat.ps.node-store-misses
409584 ? 8% -49.7% 206027 ? 3% perf-stat.ps.node-stores
5537 ? 3% -12.3% 4858 ? 2% perf-stat.ps.page-faults
4.632e+13 -71.5% 1.322e+13 perf-stat.total.instructions
9.54 ? 4% -6.1 3.40 ? 5% perf-profile.calltrace.cycles-pp.intel_idle_irq.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
9.17 ? 5% -6.1 3.05 ? 6% perf-profile.calltrace.cycles-pp.mwait_idle_with_hints.intel_idle_irq.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call
8.84 ? 4% -3.2 5.62 ? 4% perf-profile.calltrace.cycles-pp.sctp_do_sm.sctp_primitive_SEND.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg
8.63 ? 4% -3.1 5.53 ? 4% perf-profile.calltrace.cycles-pp.sctp_packet_pack.sctp_packet_transmit.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm
8.18 ? 4% -2.9 5.29 ? 4% perf-profile.calltrace.cycles-pp.memcpy_erms.sctp_packet_pack.sctp_packet_transmit.sctp_outq_flush.sctp_cmd_interpreter
6.49 ? 4% -2.5 3.99 ? 4% perf-profile.calltrace.cycles-pp.skb_copy_datagram_iter.sctp_recvmsg.inet_recvmsg.____sys_recvmsg.___sys_recvmsg
6.47 ? 3% -2.5 3.98 ? 4% perf-profile.calltrace.cycles-pp.__skb_datagram_iter.skb_copy_datagram_iter.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
21.33 ? 3% -2.5 18.86 ? 4% perf-profile.calltrace.cycles-pp.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg.____sys_sendmsg.___sys_sendmsg
13.50 ? 3% -2.5 11.04 ? 4% perf-profile.calltrace.cycles-pp.sctp_primitive_SEND.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg.____sys_sendmsg
6.89 ? 3% -2.4 4.50 ? 4% perf-profile.calltrace.cycles-pp.sctp_datamsg_from_user.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg.____sys_sendmsg
5.63 ? 3% -2.1 3.49 ? 4% perf-profile.calltrace.cycles-pp._copy_to_iter.__skb_datagram_iter.skb_copy_datagram_iter.sctp_recvmsg.inet_recvmsg
5.46 ? 3% -2.1 3.40 ? 4% perf-profile.calltrace.cycles-pp.copyout._copy_to_iter.__skb_datagram_iter.skb_copy_datagram_iter.sctp_recvmsg
5.42 ? 3% -2.0 3.38 ? 4% perf-profile.calltrace.cycles-pp.copy_user_enhanced_fast_string.copyout._copy_to_iter.__skb_datagram_iter.skb_copy_datagram_iter
11.96 ? 3% -1.9 10.08 ? 4% perf-profile.calltrace.cycles-pp.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm.sctp_primitive_SEND.sctp_sendmsg_to_asoc
9.08 ? 4% -1.6 7.48 ? 4% perf-profile.calltrace.cycles-pp.sctp_cmd_interpreter.sctp_do_sm.sctp_primitive_SEND.sctp_sendmsg_to_asoc.sctp_sendmsg
1.76 ? 4% -1.3 0.43 ? 44% perf-profile.calltrace.cycles-pp.__alloc_pages.kmalloc_large_node.__kmalloc_node_track_caller.kmalloc_reserve.__alloc_skb
10.30 ? 4% -1.2 9.11 ? 4% perf-profile.calltrace.cycles-pp.sctp_packet_transmit.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm.sctp_primitive_SEND
3.42 ? 3% -1.1 2.30 ? 4% perf-profile.calltrace.cycles-pp.sctp_user_addto_chunk.sctp_datamsg_from_user.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg
2.85 ? 3% -1.1 1.78 ? 4% perf-profile.calltrace.cycles-pp.sctp_make_datafrag_empty.sctp_datamsg_from_user.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg
1.79 ? 4% -1.0 0.75 ? 4% perf-profile.calltrace.cycles-pp.skb_release_data.kfree_skb_reason.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
3.08 ? 4% -1.0 2.07 ? 4% perf-profile.calltrace.cycles-pp.sctp_data_ready.sctp_ulpq_tail_event.sctp_ulpq_tail_data.sctp_cmd_interpreter.sctp_do_sm
3.08 ? 3% -1.0 2.08 ? 4% perf-profile.calltrace.cycles-pp._copy_from_iter.sctp_user_addto_chunk.sctp_datamsg_from_user.sctp_sendmsg_to_asoc.sctp_sendmsg
3.20 ? 4% -1.0 2.20 ? 4% perf-profile.calltrace.cycles-pp.sctp_ulpq_tail_event.sctp_ulpq_tail_data.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv
2.82 ? 4% -1.0 1.84 ? 4% perf-profile.calltrace.cycles-pp.__wake_up_common_lock.sctp_data_ready.sctp_ulpq_tail_event.sctp_ulpq_tail_data.sctp_cmd_interpreter
2.94 ? 3% -1.0 1.97 ? 4% perf-profile.calltrace.cycles-pp.copyin._copy_from_iter.sctp_user_addto_chunk.sctp_datamsg_from_user.sctp_sendmsg_to_asoc
2.42 ? 3% -1.0 1.46 ? 4% perf-profile.calltrace.cycles-pp._sctp_make_chunk.sctp_make_datafrag_empty.sctp_datamsg_from_user.sctp_sendmsg_to_asoc.sctp_sendmsg
2.90 ? 3% -0.9 1.95 ? 4% perf-profile.calltrace.cycles-pp.copy_user_enhanced_fast_string.copyin._copy_from_iter.sctp_user_addto_chunk.sctp_datamsg_from_user
11.68 ? 3% -0.9 10.75 ? 4% perf-profile.calltrace.cycles-pp.sctp_packet_transmit.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv
2.63 ? 4% -0.9 1.70 ? 4% perf-profile.calltrace.cycles-pp.__wake_up_common.__wake_up_common_lock.sctp_data_ready.sctp_ulpq_tail_event.sctp_ulpq_tail_data
2.46 ? 4% -0.9 1.59 ? 4% perf-profile.calltrace.cycles-pp.autoremove_wake_function.__wake_up_common.__wake_up_common_lock.sctp_data_ready.sctp_ulpq_tail_event
2.44 ? 4% -0.9 1.58 ? 4% perf-profile.calltrace.cycles-pp.try_to_wake_up.autoremove_wake_function.__wake_up_common.__wake_up_common_lock.sctp_data_ready
11.49 ? 3% -0.8 10.64 ? 4% perf-profile.calltrace.cycles-pp.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv.sctp_backlog_rcv
1.82 ? 3% -0.8 1.02 ? 4% perf-profile.calltrace.cycles-pp.sctp_ulpevent_free.sctp_recvmsg.inet_recvmsg.____sys_recvmsg.___sys_recvmsg
1.86 ? 4% -0.8 1.08 ? 4% perf-profile.calltrace.cycles-pp.__alloc_skb._sctp_make_chunk.sctp_make_datafrag_empty.sctp_datamsg_from_user.sctp_sendmsg_to_asoc
1.88 ? 3% -0.8 1.10 ? 5% perf-profile.calltrace.cycles-pp.schedule_timeout.sctp_skb_recv_datagram.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
1.84 ? 3% -0.8 1.09 ? 5% perf-profile.calltrace.cycles-pp.schedule.schedule_timeout.sctp_skb_recv_datagram.sctp_recvmsg.inet_recvmsg
1.81 ? 3% -0.7 1.07 ? 5% perf-profile.calltrace.cycles-pp.__schedule.schedule.schedule_timeout.sctp_skb_recv_datagram.sctp_recvmsg
1.68 ? 4% -0.7 0.98 ? 6% perf-profile.calltrace.cycles-pp.sctp_outq_flush_data.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm.sctp_primitive_SEND
1.65 ? 4% -0.7 0.99 ? 3% perf-profile.calltrace.cycles-pp.schedule_idle.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
1.61 ? 4% -0.6 0.96 ? 3% perf-profile.calltrace.cycles-pp.__schedule.schedule_idle.do_idle.cpu_startup_entry.start_secondary
1.30 ? 3% -0.6 0.69 ? 3% perf-profile.calltrace.cycles-pp.kmalloc_reserve.__alloc_skb._sctp_make_chunk.sctp_make_datafrag_empty.sctp_datamsg_from_user
1.27 ? 3% -0.6 0.67 ? 3% perf-profile.calltrace.cycles-pp.__kmalloc_node_track_caller.kmalloc_reserve.__alloc_skb._sctp_make_chunk.sctp_make_datafrag_empty
1.24 ? 3% -0.6 0.65 ? 3% perf-profile.calltrace.cycles-pp.kmalloc_large_node.__kmalloc_node_track_caller.kmalloc_reserve.__alloc_skb._sctp_make_chunk
0.97 ? 3% -0.5 0.46 ? 44% perf-profile.calltrace.cycles-pp.dequeue_task_fair.__schedule.schedule.schedule_timeout.sctp_skb_recv_datagram
1.02 ? 3% -0.5 0.57 ? 3% perf-profile.calltrace.cycles-pp.sctp_chunk_put.sctp_ulpevent_free.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
1.10 ? 4% -0.4 0.67 ? 5% perf-profile.calltrace.cycles-pp.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv.sctp_rcv
0.92 ? 3% -0.4 0.56 ? 6% perf-profile.calltrace.cycles-pp.skb_release_data.consume_skb.sctp_chunk_put.sctp_outq_sack.sctp_cmd_interpreter
1.04 ? 4% -0.3 0.71 ? 4% perf-profile.calltrace.cycles-pp.sendmsg_copy_msghdr.___sys_sendmsg.__sys_sendmsg.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.17 ?141% +0.5 0.70 ? 15% perf-profile.calltrace.cycles-pp.menu_select.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary
0.00 +0.9 0.91 ? 5% perf-profile.calltrace.cycles-pp.page_counter_uncharge.drain_stock.refill_stock.__sk_mem_reduce_allocated.skb_release_head_state
0.00 +0.9 0.93 ? 5% perf-profile.calltrace.cycles-pp.drain_stock.refill_stock.__sk_mem_reduce_allocated.skb_release_head_state.kfree_skb_reason
0.00 +1.0 0.95 ? 15% perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call
0.00 +1.0 1.03 ? 5% perf-profile.calltrace.cycles-pp.refill_stock.__sk_mem_reduce_allocated.skb_release_head_state.kfree_skb_reason.sctp_recvmsg
0.00 +1.0 1.04 ? 14% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
11.57 ? 3% +1.3 12.82 ? 4% perf-profile.calltrace.cycles-pp.sctp_assoc_bh_rcv.sctp_backlog_rcv.__release_sock.release_sock.sctp_sendmsg
0.00 +1.3 1.33 ? 4% perf-profile.calltrace.cycles-pp.page_counter_uncharge.drain_stock.refill_stock.__sk_mem_reduce_allocated.sctp_wfree
0.00 +1.4 1.36 ? 4% perf-profile.calltrace.cycles-pp.drain_stock.refill_stock.__sk_mem_reduce_allocated.sctp_wfree.skb_release_head_state
1.67 ? 3% +1.4 3.04 ? 4% perf-profile.calltrace.cycles-pp.sctp_ulpevent_make_rcvmsg.sctp_ulpq_tail_data.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv
11.39 ? 3% +1.4 12.77 ? 4% perf-profile.calltrace.cycles-pp.sctp_do_sm.sctp_assoc_bh_rcv.sctp_backlog_rcv.__release_sock.release_sock
0.00 +1.4 1.45 ? 4% perf-profile.calltrace.cycles-pp.refill_stock.__sk_mem_reduce_allocated.sctp_wfree.skb_release_head_state.consume_skb
0.00 +1.5 1.46 ? 5% perf-profile.calltrace.cycles-pp.try_charge_memcg.mem_cgroup_charge_skmem.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_ulpevent_make_rcvmsg
0.00 +1.7 1.74 ? 4% perf-profile.calltrace.cycles-pp.mem_cgroup_charge_skmem.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_ulpevent_make_rcvmsg.sctp_ulpq_tail_data
0.00 +1.8 1.83 ? 4% perf-profile.calltrace.cycles-pp.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_ulpevent_make_rcvmsg.sctp_ulpq_tail_data.sctp_cmd_interpreter
0.00 +1.8 1.84 ? 4% perf-profile.calltrace.cycles-pp.__sk_mem_schedule.sctp_ulpevent_make_rcvmsg.sctp_ulpq_tail_data.sctp_cmd_interpreter.sctp_do_sm
0.00 +2.0 1.97 ? 5% perf-profile.calltrace.cycles-pp.page_counter_try_charge.try_charge_memcg.mem_cgroup_charge_skmem.__sk_mem_raise_allocated.__sk_mem_schedule
18.94 ? 3% +2.1 21.02 ? 4% perf-profile.calltrace.cycles-pp.recvmsg
0.00 +2.1 2.14 ? 4% perf-profile.calltrace.cycles-pp.try_charge_memcg.mem_cgroup_charge_skmem.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_sendmsg_to_asoc
10.88 ? 4% +2.2 13.11 ? 4% perf-profile.calltrace.cycles-pp.__ip_queue_xmit.sctp_packet_transmit.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm
10.74 ? 4% +2.3 13.04 ? 4% perf-profile.calltrace.cycles-pp.ip_finish_output2.__ip_queue_xmit.sctp_packet_transmit.sctp_outq_flush.sctp_cmd_interpreter
18.22 ? 3% +2.3 20.55 ? 4% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.recvmsg
10.51 ? 4% +2.4 12.88 ? 4% perf-profile.calltrace.cycles-pp.__local_bh_enable_ip.ip_finish_output2.__ip_queue_xmit.sctp_packet_transmit.sctp_outq_flush
10.47 ? 4% +2.4 12.84 ? 4% perf-profile.calltrace.cycles-pp.do_softirq.__local_bh_enable_ip.ip_finish_output2.__ip_queue_xmit.sctp_packet_transmit
18.08 ? 3% +2.4 20.46 ? 4% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.recvmsg
10.42 ? 4% +2.4 12.81 ? 4% perf-profile.calltrace.cycles-pp.__softirqentry_text_start.do_softirq.__local_bh_enable_ip.ip_finish_output2.__ip_queue_xmit
10.25 ? 4% +2.4 12.70 ? 4% perf-profile.calltrace.cycles-pp.net_rx_action.__softirqentry_text_start.do_softirq.__local_bh_enable_ip.ip_finish_output2
17.84 ? 3% +2.5 20.32 ? 4% perf-profile.calltrace.cycles-pp.__sys_recvmsg.do_syscall_64.entry_SYSCALL_64_after_hwframe.recvmsg
10.15 ? 4% +2.5 12.64 ? 4% perf-profile.calltrace.cycles-pp.__napi_poll.net_rx_action.__softirqentry_text_start.do_softirq.__local_bh_enable_ip
10.13 ? 4% +2.5 12.63 ? 4% perf-profile.calltrace.cycles-pp.process_backlog.__napi_poll.net_rx_action.__softirqentry_text_start.do_softirq
0.00 +2.5 2.50 ? 4% perf-profile.calltrace.cycles-pp.mem_cgroup_charge_skmem.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_sendmsg_to_asoc.sctp_sendmsg
3.59 ? 3% +2.5 6.11 ? 5% perf-profile.calltrace.cycles-pp.sctp_skb_recv_datagram.sctp_recvmsg.inet_recvmsg.____sys_recvmsg.___sys_recvmsg
9.97 ? 4% +2.6 12.54 ? 4% perf-profile.calltrace.cycles-pp.__netif_receive_skb_one_core.process_backlog.__napi_poll.net_rx_action.__softirqentry_text_start
17.56 ? 3% +2.6 20.14 ? 4% perf-profile.calltrace.cycles-pp.___sys_recvmsg.__sys_recvmsg.do_syscall_64.entry_SYSCALL_64_after_hwframe.recvmsg
0.00 +2.7 2.66 ? 4% perf-profile.calltrace.cycles-pp.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg
0.00 +2.7 2.68 ? 4% perf-profile.calltrace.cycles-pp.__sk_mem_schedule.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg.____sys_sendmsg
9.70 ? 4% +2.7 12.40 ? 4% perf-profile.calltrace.cycles-pp.ip_local_deliver_finish.__netif_receive_skb_one_core.process_backlog.__napi_poll.net_rx_action
9.68 ? 4% +2.7 12.39 ? 4% perf-profile.calltrace.cycles-pp.ip_protocol_deliver_rcu.ip_local_deliver_finish.__netif_receive_skb_one_core.process_backlog.__napi_poll
9.61 ? 4% +2.7 12.33 ? 4% perf-profile.calltrace.cycles-pp.sctp_rcv.ip_protocol_deliver_rcu.ip_local_deliver_finish.__netif_receive_skb_one_core.process_backlog
16.65 ? 3% +2.9 19.58 ? 4% perf-profile.calltrace.cycles-pp.____sys_recvmsg.___sys_recvmsg.__sys_recvmsg.do_syscall_64.entry_SYSCALL_64_after_hwframe
8.06 ? 4% +3.1 11.19 ? 4% perf-profile.calltrace.cycles-pp.sctp_assoc_bh_rcv.sctp_rcv.ip_protocol_deliver_rcu.ip_local_deliver_finish.__netif_receive_skb_one_core
15.88 ? 3% +3.2 19.12 ? 4% perf-profile.calltrace.cycles-pp.inet_recvmsg.____sys_recvmsg.___sys_recvmsg.__sys_recvmsg.do_syscall_64
15.83 ? 3% +3.3 19.09 ? 4% perf-profile.calltrace.cycles-pp.sctp_recvmsg.inet_recvmsg.____sys_recvmsg.___sys_recvmsg.__sys_recvmsg
7.62 ? 3% +3.3 10.90 ? 4% perf-profile.calltrace.cycles-pp.sctp_do_sm.sctp_assoc_bh_rcv.sctp_rcv.ip_protocol_deliver_rcu.ip_local_deliver_finish
1.14 ? 3% +3.5 4.62 ? 5% perf-profile.calltrace.cycles-pp.lock_sock_nested.sctp_skb_recv_datagram.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
1.02 ? 3% +3.5 4.51 ? 5% perf-profile.calltrace.cycles-pp._raw_spin_lock_bh.lock_sock_nested.sctp_skb_recv_datagram.sctp_recvmsg.inet_recvmsg
0.83 ? 4% +3.6 4.38 ? 5% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_bh.lock_sock_nested.sctp_skb_recv_datagram.sctp_recvmsg
15.16 ? 3% +3.6 18.74 ? 5% perf-profile.calltrace.cycles-pp.release_sock.sctp_sendmsg.sock_sendmsg.____sys_sendmsg.___sys_sendmsg
15.04 ? 3% +3.6 18.66 ? 5% perf-profile.calltrace.cycles-pp.__release_sock.release_sock.sctp_sendmsg.sock_sendmsg.____sys_sendmsg
11.66 ? 3% +3.6 15.29 ? 5% perf-profile.calltrace.cycles-pp.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv.sctp_backlog_rcv.__release_sock
14.98 ? 3% +3.6 18.62 ? 5% perf-profile.calltrace.cycles-pp.sctp_backlog_rcv.__release_sock.release_sock.sctp_sendmsg.sock_sendmsg
0.00 +4.2 4.24 ? 4% perf-profile.calltrace.cycles-pp.sctp_eat_data.sctp_sf_eat_data_6_2.sctp_do_sm.sctp_assoc_bh_rcv.sctp_rcv
0.00 +4.3 4.33 ? 4% perf-profile.calltrace.cycles-pp.sctp_sf_eat_data_6_2.sctp_do_sm.sctp_assoc_bh_rcv.sctp_rcv.ip_protocol_deliver_rcu
2.88 ? 3% +4.7 7.54 ? 4% perf-profile.calltrace.cycles-pp.sctp_outq_sack.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv.sctp_backlog_rcv
23.81 ? 12% +4.8 28.57 ? 11% perf-profile.calltrace.cycles-pp.mwait_idle_with_hints.intel_idle.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call
23.81 ? 12% +4.8 28.58 ? 11% perf-profile.calltrace.cycles-pp.intel_idle.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
2.11 ? 4% +4.8 6.89 ? 4% perf-profile.calltrace.cycles-pp.kfree_skb_reason.sctp_recvmsg.inet_recvmsg.____sys_recvmsg.___sys_recvmsg
1.84 ? 3% +5.0 6.86 ? 4% perf-profile.calltrace.cycles-pp.sctp_chunk_put.sctp_outq_sack.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv
1.42 ? 3% +5.0 6.47 ? 4% perf-profile.calltrace.cycles-pp.consume_skb.sctp_chunk_put.sctp_outq_sack.sctp_cmd_interpreter.sctp_do_sm
0.00 +5.5 5.53 ? 4% perf-profile.calltrace.cycles-pp.__sk_mem_reduce_allocated.sctp_wfree.skb_release_head_state.consume_skb.sctp_chunk_put
0.00 +5.8 5.80 ? 4% perf-profile.calltrace.cycles-pp.sctp_wfree.skb_release_head_state.consume_skb.sctp_chunk_put.sctp_outq_sack
0.00 +5.9 5.87 ? 4% perf-profile.calltrace.cycles-pp.skb_release_head_state.consume_skb.sctp_chunk_put.sctp_outq_sack.sctp_cmd_interpreter
0.00 +6.0 5.99 ? 4% perf-profile.calltrace.cycles-pp.__sk_mem_reduce_allocated.skb_release_head_state.kfree_skb_reason.sctp_recvmsg.inet_recvmsg
0.00 +6.1 6.13 ? 4% perf-profile.calltrace.cycles-pp.skb_release_head_state.kfree_skb_reason.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
9.61 ? 5% -6.2 3.42 ? 5% perf-profile.children.cycles-pp.intel_idle_irq
9.16 ? 3% -3.3 5.87 ? 5% perf-profile.children.cycles-pp.sctp_packet_pack
8.74 ? 3% -3.1 5.61 ? 4% perf-profile.children.cycles-pp.copy_user_enhanced_fast_string
25.01 ? 3% -3.0 21.98 ? 4% perf-profile.children.cycles-pp.sctp_outq_flush
8.52 ? 3% -3.0 5.50 ? 5% perf-profile.children.cycles-pp.memcpy_erms
6.49 ? 3% -2.5 3.99 ? 4% perf-profile.children.cycles-pp.skb_copy_datagram_iter
6.47 ? 3% -2.5 3.98 ? 4% perf-profile.children.cycles-pp.__skb_datagram_iter
7.07 ? 3% -2.5 4.60 ? 4% perf-profile.children.cycles-pp.sctp_datamsg_from_user
21.61 ? 3% -2.4 19.19 ? 4% perf-profile.children.cycles-pp.sctp_sendmsg_to_asoc
13.72 ? 3% -2.4 11.31 ? 5% perf-profile.children.cycles-pp.sctp_primitive_SEND
5.63 ? 3% -2.1 3.50 ? 4% perf-profile.children.cycles-pp._copy_to_iter
5.46 ? 3% -2.1 3.40 ? 4% perf-profile.children.cycles-pp.copyout
22.37 ? 3% -2.0 20.39 ? 4% perf-profile.children.cycles-pp.sctp_packet_transmit
3.59 ? 3% -1.6 1.96 ? 4% perf-profile.children.cycles-pp.__alloc_skb
2.98 ? 3% -1.5 1.49 ? 5% perf-profile.children.cycles-pp.skb_release_data
3.46 ? 4% -1.4 2.06 ? 4% perf-profile.children.cycles-pp.__schedule
2.66 ? 3% -1.3 1.35 ? 4% perf-profile.children.cycles-pp.kmalloc_reserve
2.62 ? 3% -1.3 1.32 ? 4% perf-profile.children.cycles-pp.__kmalloc_node_track_caller
2.50 ? 3% -1.3 1.24 ? 4% perf-profile.children.cycles-pp.kmalloc_large_node
3.51 ? 3% -1.2 2.36 ? 4% perf-profile.children.cycles-pp.sctp_user_addto_chunk
2.20 ? 3% -1.1 1.07 ? 4% perf-profile.children.cycles-pp.__alloc_pages
2.93 ? 3% -1.1 1.82 ? 5% perf-profile.children.cycles-pp.sctp_make_datafrag_empty
2.70 ? 3% -1.1 1.62 ? 4% perf-profile.children.cycles-pp._sctp_make_chunk
3.17 ? 3% -1.0 2.12 ? 5% perf-profile.children.cycles-pp.sctp_data_ready
3.16 ? 3% -1.0 2.13 ? 4% perf-profile.children.cycles-pp._copy_from_iter
3.28 ? 3% -1.0 2.26 ? 5% perf-profile.children.cycles-pp.sctp_ulpq_tail_event
2.90 ? 3% -1.0 1.88 ? 5% perf-profile.children.cycles-pp.__wake_up_common_lock
1.89 ? 4% -1.0 0.88 ? 4% perf-profile.children.cycles-pp.get_page_from_freelist
3.02 ? 3% -1.0 2.02 ? 4% perf-profile.children.cycles-pp.copyin
2.47 ? 3% -1.0 1.49 ? 7% perf-profile.children.cycles-pp.sctp_outq_flush_data
2.70 ? 3% -0.9 1.75 ? 5% perf-profile.children.cycles-pp.__wake_up_common
2.53 ? 3% -0.9 1.64 ? 5% perf-profile.children.cycles-pp.autoremove_wake_function
2.51 ? 3% -0.9 1.63 ? 4% perf-profile.children.cycles-pp.try_to_wake_up
1.83 ? 3% -0.8 1.02 ? 4% perf-profile.children.cycles-pp.sctp_ulpevent_free
2.10 ? 3% -0.8 1.32 ? 6% perf-profile.children.cycles-pp.kmem_cache_free
1.89 ? 3% -0.8 1.12 ? 5% perf-profile.children.cycles-pp.schedule_timeout
1.39 ? 4% -0.8 0.63 ? 5% perf-profile.children.cycles-pp.free_unref_page
1.86 ? 3% -0.7 1.11 ? 4% perf-profile.children.cycles-pp.schedule
1.25 ? 4% -0.7 0.56 ? 5% perf-profile.children.cycles-pp.rmqueue
1.28 ? 4% -0.7 0.60 ? 5% perf-profile.children.cycles-pp._raw_spin_lock_irqsave
1.66 ? 4% -0.7 1.00 ? 4% perf-profile.children.cycles-pp.schedule_idle
0.98 ? 4% -0.6 0.43 ? 3% perf-profile.children.cycles-pp._raw_spin_lock
1.08 ? 4% -0.5 0.54 ? 4% perf-profile.children.cycles-pp.sctp_outq_select_transport
1.45 ? 4% -0.5 0.94 ? 4% perf-profile.children.cycles-pp._copy_from_user
0.98 ? 4% -0.5 0.48 ? 5% perf-profile.children.cycles-pp.sctp_packet_config
1.25 ? 12% -0.5 0.78 ? 5% perf-profile.children.cycles-pp.__rhashtable_lookup
1.21 ? 3% -0.5 0.73 ? 4% perf-profile.children.cycles-pp.__check_object_size
0.98 ? 3% -0.4 0.53 ? 7% perf-profile.children.cycles-pp.dequeue_task_fair
0.64 ? 5% -0.4 0.20 ? 5% perf-profile.children.cycles-pp.free_pcppages_bulk
1.18 ? 4% -0.4 0.78 ? 5% perf-profile.children.cycles-pp.ttwu_do_activate
0.62 ? 4% -0.4 0.22 ? 4% perf-profile.children.cycles-pp.__free_pages_ok
0.90 ? 3% -0.4 0.50 ? 6% perf-profile.children.cycles-pp.dequeue_entity
1.14 ? 3% -0.4 0.75 ? 4% perf-profile.children.cycles-pp.enqueue_task_fair
0.95 ? 4% -0.4 0.57 ? 6% perf-profile.children.cycles-pp.__slab_free
0.76 ? 3% -0.3 0.42 ? 3% perf-profile.children.cycles-pp.__skb_clone
1.07 ? 4% -0.3 0.73 ? 5% perf-profile.children.cycles-pp.sendmsg_copy_msghdr
0.76 ? 3% -0.3 0.42 ? 5% perf-profile.children.cycles-pp.update_load_avg
0.98 ? 4% -0.3 0.65 ? 3% perf-profile.children.cycles-pp.__copy_msghdr_from_user
0.81 ? 14% -0.3 0.49 ? 6% perf-profile.children.cycles-pp.sctp_hash_cmp
0.88 ? 4% -0.3 0.56 ? 4% perf-profile.children.cycles-pp.pick_next_task_fair
0.84 ? 3% -0.3 0.54 ? 5% perf-profile.children.cycles-pp.enqueue_entity
0.82 ? 3% -0.3 0.53 ? 5% perf-profile.children.cycles-pp.import_iovec
0.81 ? 2% -0.3 0.52 ? 6% perf-profile.children.cycles-pp.kfree
0.59 ? 3% -0.3 0.30 ? 4% perf-profile.children.cycles-pp.__copy_skb_header
0.78 ? 3% -0.3 0.51 ? 6% perf-profile.children.cycles-pp.__import_iovec
0.62 ? 5% -0.3 0.35 ? 5% perf-profile.children.cycles-pp.set_next_entity
0.76 ? 12% -0.3 0.50 ? 5% perf-profile.children.cycles-pp.sctp_endpoint_lookup_assoc
0.69 ? 4% -0.3 0.43 ? 5% perf-profile.children.cycles-pp.check_heap_object
0.77 ? 3% -0.3 0.52 ? 5% perf-profile.children.cycles-pp.kmem_cache_alloc
0.74 ? 12% -0.3 0.49 ? 5% perf-profile.children.cycles-pp.sctp_epaddr_lookup_transport
0.72 ? 10% -0.2 0.48 ? 6% perf-profile.children.cycles-pp.__sctp_rcv_lookup
0.66 ? 4% -0.2 0.42 ? 5% perf-profile.children.cycles-pp.__might_resched
0.70 ? 10% -0.2 0.46 ? 7% perf-profile.children.cycles-pp.sctp_addrs_lookup_transport
0.65 ? 2% -0.2 0.42 ? 5% perf-profile.children.cycles-pp.iovec_from_user
0.50 ? 4% -0.2 0.26 ? 4% perf-profile.children.cycles-pp.simple_copy_to_iter
0.46 ? 3% -0.2 0.23 ? 8% perf-profile.children.cycles-pp.sctp_inq_pop
0.41 ? 3% -0.2 0.19 ? 6% perf-profile.children.cycles-pp.update_curr
0.69 ? 3% -0.2 0.47 ? 5% perf-profile.children.cycles-pp.sctp_chunkify
0.55 ? 3% -0.2 0.34 ? 4% perf-profile.children.cycles-pp.__dev_queue_xmit
0.53 ? 4% -0.2 0.33 ? 6% perf-profile.children.cycles-pp.__might_fault
0.58 ? 4% -0.2 0.39 ? 5% perf-profile.children.cycles-pp.copy_user_short_string
0.46 ? 9% -0.2 0.28 ? 12% perf-profile.children.cycles-pp.select_task_rq
0.33 ? 4% -0.2 0.16 ? 7% perf-profile.children.cycles-pp.free_unref_page_commit
0.28 ? 5% -0.2 0.11 ? 9% perf-profile.children.cycles-pp.dst_release
0.32 ? 7% -0.2 0.15 ? 7% perf-profile.children.cycles-pp.ipv4_dst_check
0.38 ? 3% -0.2 0.22 ? 8% perf-profile.children.cycles-pp.__mod_node_page_state
0.44 ? 8% -0.2 0.28 ? 13% perf-profile.children.cycles-pp.sctp_packet_append_chunk
0.49 ? 6% -0.2 0.33 ? 11% perf-profile.children.cycles-pp.sctp_packet_transmit_chunk
0.54 ? 4% -0.2 0.39 ? 6% perf-profile.children.cycles-pp.__list_del_entry_valid
0.34 ? 4% -0.2 0.19 ? 8% perf-profile.children.cycles-pp.sctp_v4_xmit
0.42 ? 3% -0.2 0.27 ? 10% perf-profile.children.cycles-pp.sockfd_lookup_light
0.32 ? 4% -0.1 0.17 ? 4% perf-profile.children.cycles-pp.__zone_watermark_ok
0.47 ? 3% -0.1 0.32 ? 5% perf-profile.children.cycles-pp.__entry_text_start
0.28 ? 2% -0.1 0.13 ? 5% perf-profile.children.cycles-pp.__free_one_page
0.28 ? 5% -0.1 0.13 ? 4% perf-profile.children.cycles-pp.prepare_task_switch
0.43 ? 5% -0.1 0.29 ? 9% perf-profile.children.cycles-pp.sctp_addto_chunk
0.37 ? 4% -0.1 0.23 ? 6% perf-profile.children.cycles-pp.sctp_gen_sack
0.41 ? 4% -0.1 0.27 ? 8% perf-profile.children.cycles-pp.aa_sk_perm
0.35 ? 3% -0.1 0.22 ? 6% perf-profile.children.cycles-pp.__list_add_valid
0.34 ? 4% -0.1 0.20 ? 6% perf-profile.children.cycles-pp.sctp_make_sack
0.44 ? 4% -0.1 0.30 ? 6% perf-profile.children.cycles-pp.update_rq_clock
0.37 ? 4% -0.1 0.24 ? 7% perf-profile.children.cycles-pp.sock_recvmsg
0.34 ? 4% -0.1 0.22 ? 4% perf-profile.children.cycles-pp.sctp_check_transmitted
0.35 ? 3% -0.1 0.22 ? 6% perf-profile.children.cycles-pp.memcg_slab_free_hook
0.36 ? 4% -0.1 0.23 ? 8% perf-profile.children.cycles-pp.__fdget
0.34 ? 10% -0.1 0.22 ? 15% perf-profile.children.cycles-pp.select_task_rq_fair
0.44 ? 4% -0.1 0.32 ? 7% perf-profile.children.cycles-pp.kmem_cache_alloc_node
0.34 ? 5% -0.1 0.21 ? 7% perf-profile.children.cycles-pp.__virt_addr_valid
0.34 ? 3% -0.1 0.22 ? 5% perf-profile.children.cycles-pp.dev_hard_start_xmit
0.28 ? 14% -0.1 0.16 ? 6% perf-profile.children.cycles-pp.sctp_transport_hold
0.35 ? 2% -0.1 0.23 ? 7% perf-profile.children.cycles-pp.__might_sleep
0.38 ? 4% -0.1 0.26 ? 4% perf-profile.children.cycles-pp.sctp_chunk_free
0.31 ? 5% -0.1 0.20 ? 5% perf-profile.children.cycles-pp.loopback_xmit
0.33 ? 4% -0.1 0.22 ? 4% perf-profile.children.cycles-pp.free_pcp_prepare
0.33 ? 5% -0.1 0.22 ? 7% perf-profile.children.cycles-pp.security_socket_recvmsg
0.35 ? 2% -0.1 0.24 ? 4% perf-profile.children.cycles-pp.sock_kmalloc
0.30 ? 2% -0.1 0.20 ? 8% perf-profile.children.cycles-pp.sctp_association_put
0.22 ? 7% -0.1 0.12 ? 6% perf-profile.children.cycles-pp.sctp_assoc_rwnd_increase
0.34 ? 4% -0.1 0.23 ? 2% perf-profile.children.cycles-pp.move_addr_to_kernel
0.28 ? 4% -0.1 0.17 ? 4% perf-profile.children.cycles-pp.sock_kfree_s
0.23 ? 5% -0.1 0.13 ? 7% perf-profile.children.cycles-pp.rmqueue_bulk
0.31 ? 9% -0.1 0.21 ? 8% perf-profile.children.cycles-pp.sctp_transport_put
0.33 ? 6% -0.1 0.23 ? 12% perf-profile.children.cycles-pp.sctp_sched_fcfs_dequeue
0.28 ? 8% -0.1 0.18 ? 20% perf-profile.children.cycles-pp.__sctp_packet_append_chunk
0.34 ? 4% -0.1 0.24 ? 8% perf-profile.children.cycles-pp.memset_erms
0.19 ? 4% -0.1 0.10 ? 5% perf-profile.children.cycles-pp.__update_load_avg_cfs_rq
0.29 ? 4% -0.1 0.19 ? 7% perf-profile.children.cycles-pp.skb_set_owner_w
0.27 ? 4% -0.1 0.18 ? 5% perf-profile.children.cycles-pp.security_socket_sendmsg
0.24 ? 4% -0.1 0.14 ? 6% perf-profile.children.cycles-pp.__switch_to
0.26 ? 4% -0.1 0.17 ? 9% perf-profile.children.cycles-pp.entry_SYSRETQ_unsafe_stack
0.27 ? 2% -0.1 0.18 ? 5% perf-profile.children.cycles-pp.send_sctp_stream_1toMany
0.23 ? 4% -0.1 0.14 ? 5% perf-profile.children.cycles-pp.tick_nohz_idle_exit
0.20 ? 4% -0.1 0.11 ? 8% perf-profile.children.cycles-pp.sctp_outq_flush_ctrl
0.24 ? 7% -0.1 0.15 ? 14% perf-profile.children.cycles-pp.select_idle_sibling
0.26 ? 4% -0.1 0.17 ? 11% perf-profile.children.cycles-pp.syscall_exit_to_user_mode
0.22 ? 5% -0.1 0.14 ? 5% perf-profile.children.cycles-pp.__switch_to_asm
0.21 -0.1 0.13 ? 8% perf-profile.children.cycles-pp.accept_connection
0.21 -0.1 0.13 ? 8% perf-profile.children.cycles-pp.spawn_child
0.21 -0.1 0.13 ? 8% perf-profile.children.cycles-pp.process_requests
0.14 ? 5% -0.1 0.06 ? 11% perf-profile.children.cycles-pp.check_stack_object
0.20 ? 3% -0.1 0.12 ? 9% perf-profile.children.cycles-pp.sock_wfree
0.22 ? 4% -0.1 0.14 ? 3% perf-profile.children.cycles-pp.prepare_to_wait_exclusive
0.24 ? 5% -0.1 0.17 ? 9% perf-profile.children.cycles-pp.sctp_sched_dequeue_common
0.22 ? 3% -0.1 0.15 ? 2% perf-profile.children.cycles-pp.__kmalloc
0.21 ? 3% -0.1 0.13 ? 8% perf-profile.children.cycles-pp.sctp_make_control
0.20 ? 2% -0.1 0.12 ? 9% perf-profile.children.cycles-pp.recv_sctp_stream_1toMany
0.17 ? 6% -0.1 0.09 ? 5% perf-profile.children.cycles-pp.___perf_sw_event
0.14 ? 5% -0.1 0.06 ? 7% perf-profile.children.cycles-pp.sctp_sock_rfree
0.20 ? 4% -0.1 0.13 ? 7% perf-profile.children.cycles-pp.skb_clone
0.19 ? 30% -0.1 0.12 ? 5% perf-profile.children.cycles-pp.sctp_cmp_addr_exact
0.17 ? 4% -0.1 0.10 ? 8% perf-profile.children.cycles-pp.__mod_timer
0.20 ? 4% -0.1 0.13 ? 6% perf-profile.children.cycles-pp.available_idle_cpu
0.23 ? 7% -0.1 0.16 ? 3% perf-profile.children.cycles-pp.__cond_resched
0.21 ? 6% -0.1 0.14 ? 10% perf-profile.children.cycles-pp.sctp_datamsg_put
0.16 ? 4% -0.1 0.09 ? 5% perf-profile.children.cycles-pp.check_new_pages
0.14 ? 6% -0.1 0.08 ? 12% perf-profile.children.cycles-pp.sctp_inet_skb_msgname
0.14 ? 6% -0.1 0.08 ? 9% perf-profile.children.cycles-pp.ip_rcv
0.18 ? 3% -0.1 0.11 ? 9% perf-profile.children.cycles-pp.syscall_return_via_sysret
0.18 ? 4% -0.1 0.12 ? 9% perf-profile.children.cycles-pp.poll_idle
0.20 ? 6% -0.1 0.14 ? 6% perf-profile.children.cycles-pp.kmem_cache_alloc_trace
0.19 -0.1 0.13 ? 8% perf-profile.children.cycles-pp.sctp_outq_tail
0.18 ? 3% -0.1 0.12 ? 7% perf-profile.children.cycles-pp.reweight_entity
0.20 ? 4% -0.1 0.13 ? 11% perf-profile.children.cycles-pp.update_cfs_group
0.16 ? 5% -0.1 0.10 ? 4% perf-profile.children.cycles-pp.del_timer
0.16 ? 5% -0.1 0.10 ? 3% perf-profile.children.cycles-pp.sctp_sendmsg_parse
0.16 ? 5% -0.1 0.10 ? 9% perf-profile.children.cycles-pp.perf_trace_sched_wakeup_template
0.18 ? 2% -0.1 0.12 ? 4% perf-profile.children.cycles-pp.__free_pages
0.17 ? 4% -0.1 0.11 ? 5% perf-profile.children.cycles-pp.nr_iowait_cpu
0.14 ? 2% -0.1 0.09 ? 10% perf-profile.children.cycles-pp.switch_mm_irqs_off
0.11 ? 6% -0.1 0.05 ? 8% perf-profile.children.cycles-pp.__build_skb_around
0.19 ? 3% -0.1 0.14 ? 8% perf-profile.children.cycles-pp.__update_load_avg_se
0.16 ? 5% -0.1 0.11 ? 6% perf-profile.children.cycles-pp.ttwu_do_wakeup
0.13 ? 6% -0.1 0.08 ? 13% perf-profile.children.cycles-pp.task_tick_fair
0.21 ? 3% -0.0 0.16 ? 6% perf-profile.children.cycles-pp.sctp_association_hold
0.15 ? 6% -0.0 0.10 ? 8% perf-profile.children.cycles-pp.check_preempt_curr
0.07 ? 6% -0.0 0.02 ? 99% perf-profile.children.cycles-pp._raw_spin_unlock_irqrestore
0.10 ? 5% -0.0 0.05 ? 7% perf-profile.children.cycles-pp.finish_task_switch
0.14 ? 6% -0.0 0.10 ? 6% perf-profile.children.cycles-pp.__genradix_ptr
0.12 ? 4% -0.0 0.08 ? 8% perf-profile.children.cycles-pp.sctp_chunk_abandoned
0.07 ? 5% -0.0 0.02 ? 99% perf-profile.children.cycles-pp.ip_local_out
0.22 ? 6% -0.0 0.17 ? 8% perf-profile.children.cycles-pp.sched_clock_cpu
0.14 ? 5% -0.0 0.09 ? 6% perf-profile.children.cycles-pp.memcg_slab_post_alloc_hook
0.12 ? 6% -0.0 0.08 ? 7% perf-profile.children.cycles-pp.__netif_rx
0.20 ? 5% -0.0 0.16 ? 8% perf-profile.children.cycles-pp.read_tsc
0.12 ? 6% -0.0 0.08 ? 10% perf-profile.children.cycles-pp.lock_timer_base
0.19 ? 5% -0.0 0.15 ? 7% perf-profile.children.cycles-pp.native_sched_clock
0.13 ? 5% -0.0 0.09 ? 5% perf-profile.children.cycles-pp.pick_next_entity
0.12 ? 5% -0.0 0.08 ? 6% perf-profile.children.cycles-pp.netif_rx_internal
0.12 ? 7% -0.0 0.08 ? 10% perf-profile.children.cycles-pp.__put_user_nocheck_4
0.12 ? 4% -0.0 0.08 ? 6% perf-profile.children.cycles-pp.skb_put
0.11 ? 4% -0.0 0.07 ? 5% perf-profile.children.cycles-pp.__put_user_nocheck_8
0.12 ? 3% -0.0 0.08 ? 8% perf-profile.children.cycles-pp.resched_curr
0.11 ? 8% -0.0 0.07 ? 25% perf-profile.children.cycles-pp.sctp_chunk_assign_ssn
0.13 ? 4% -0.0 0.09 ? 6% perf-profile.children.cycles-pp.sctp_datamsg_destroy
0.07 ? 7% -0.0 0.03 ?100% perf-profile.children.cycles-pp.ip_rcv_core
0.11 ? 3% -0.0 0.07 ? 10% perf-profile.children.cycles-pp.enqueue_to_backlog
0.10 ? 5% -0.0 0.06 ? 13% perf-profile.children.cycles-pp.sctp_transport_reset_t3_rtx
0.09 ? 6% -0.0 0.05 ? 7% perf-profile.children.cycles-pp.__netif_receive_skb_core
0.10 ? 5% -0.0 0.06 ? 6% perf-profile.children.cycles-pp.validate_xmit_skb
0.11 ? 6% -0.0 0.07 ? 12% perf-profile.children.cycles-pp.syscall_exit_to_user_mode_prepare
0.06 ? 6% -0.0 0.02 ? 99% perf-profile.children.cycles-pp.entry_SYSCALL_64_safe_stack
0.10 ? 8% -0.0 0.06 ? 9% perf-profile.children.cycles-pp.sctp_ulpq_order
0.09 ? 6% -0.0 0.06 ? 8% perf-profile.children.cycles-pp.os_xsave
0.12 ? 5% -0.0 0.08 ? 8% perf-profile.children.cycles-pp.__check_heap_object
0.10 ? 5% -0.0 0.07 ? 10% perf-profile.children.cycles-pp.rcu_all_qs
0.08 ? 5% -0.0 0.06 ? 11% perf-profile.children.cycles-pp.tick_nohz_idle_enter
0.08 ? 5% -0.0 0.06 ? 8% perf-profile.children.cycles-pp.perf_tp_event
0.08 -0.0 0.06 ? 9% perf-profile.children.cycles-pp.__mod_lruvec_page_state
0.31 ? 3% -0.0 0.28 ? 4% perf-profile.children.cycles-pp.sctp_ulpevent_receive_data
0.10 ? 10% -0.0 0.08 ? 10% perf-profile.children.cycles-pp.rcu_idle_exit
0.08 ? 6% -0.0 0.05 ? 8% perf-profile.children.cycles-pp.sctp_tsnmap_check
0.08 ? 5% -0.0 0.06 ? 14% perf-profile.children.cycles-pp.sctp_ulpevent_init
0.07 ? 8% -0.0 0.05 perf-profile.children.cycles-pp.__wrgsbase_inactive
0.10 ? 5% +0.0 0.13 ? 10% perf-profile.children.cycles-pp.sctp_sockaddr_af
0.00 +0.1 0.05 perf-profile.children.cycles-pp.perf_mux_hrtimer_handler
0.19 ? 3% +0.1 0.24 ? 11% perf-profile.children.cycles-pp.scheduler_tick
0.00 +0.1 0.06 ? 7% perf-profile.children.cycles-pp.lapic_next_deadline
0.00 +0.1 0.07 ? 32% perf-profile.children.cycles-pp.arch_cpu_idle_enter
0.00 +0.1 0.07 ? 16% perf-profile.children.cycles-pp.native_irq_return_iret
0.00 +0.1 0.08 ? 22% perf-profile.children.cycles-pp.update_blocked_averages
0.00 +0.1 0.08 ? 22% perf-profile.children.cycles-pp.run_rebalance_domains
0.24 ? 3% +0.1 0.32 ? 9% perf-profile.children.cycles-pp.update_process_times
0.00 +0.1 0.08 ? 22% perf-profile.children.cycles-pp.update_sg_lb_stats
0.24 ? 4% +0.1 0.32 ? 9% perf-profile.children.cycles-pp.tick_sched_handle
0.00 +0.1 0.08 ? 41% perf-profile.children.cycles-pp._raw_spin_trylock
0.00 +0.1 0.09 ? 20% perf-profile.children.cycles-pp.update_sd_lb_stats
0.00 +0.1 0.09 ? 19% perf-profile.children.cycles-pp.find_busiest_group
0.04 ? 71% +0.1 0.14 ? 3% perf-profile.children.cycles-pp.sctp_get_af_specific
0.01 ?223% +0.1 0.11 ? 18% perf-profile.children.cycles-pp.load_balance
0.00 +0.1 0.11 ? 11% perf-profile.children.cycles-pp.tick_nohz_irq_exit
0.02 ?141% +0.1 0.14 ? 46% perf-profile.children.cycles-pp.rebalance_domains
0.45 ? 5% +0.1 0.58 ? 10% perf-profile.children.cycles-pp.ktime_get
0.00 +0.1 0.14 ? 17% perf-profile.children.cycles-pp.cgroup_rstat_updated
0.26 ? 4% +0.1 0.41 ? 12% perf-profile.children.cycles-pp.tick_sched_timer
0.15 ? 10% +0.2 0.32 ? 16% perf-profile.children.cycles-pp.clockevents_program_event
0.30 ? 4% +0.2 0.50 ? 10% perf-profile.children.cycles-pp.__hrtimer_run_queues
0.25 ? 4% +0.2 0.45 ? 20% perf-profile.children.cycles-pp.tick_nohz_get_sleep_length
0.00 +0.2 0.20 ? 16% perf-profile.children.cycles-pp.timekeeping_max_deferment
0.50 ? 2% +0.2 0.71 ? 15% perf-profile.children.cycles-pp.menu_select
0.16 ? 5% +0.2 0.38 ? 22% perf-profile.children.cycles-pp.tick_nohz_next_event
0.08 ? 8% +0.2 0.30 ? 23% perf-profile.children.cycles-pp.__irq_exit_rcu
0.00 +0.3 0.33 ? 12% perf-profile.children.cycles-pp.propagate_protected_usage
0.48 ? 5% +0.4 0.89 ? 11% perf-profile.children.cycles-pp.hrtimer_interrupt
0.48 ? 6% +0.4 0.90 ? 10% perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
0.00 +0.6 0.55 ? 3% perf-profile.children.cycles-pp.mem_cgroup_uncharge_skmem
0.61 ? 5% +0.7 1.35 ? 11% perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
1.00 ? 2% +0.8 1.82 ? 10% perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
0.00 +1.0 1.03 ? 4% perf-profile.children.cycles-pp.__mod_memcg_state
1.73 ? 2% +1.4 3.14 ? 4% perf-profile.children.cycles-pp.sctp_ulpevent_make_rcvmsg
19.09 ? 3% +2.0 21.11 ? 4% perf-profile.children.cycles-pp.recvmsg
11.51 ? 3% +2.2 13.66 ? 4% perf-profile.children.cycles-pp.__ip_queue_xmit
11.26 ? 3% +2.3 13.53 ? 4% perf-profile.children.cycles-pp.ip_finish_output2
0.00 +2.3 2.27 ? 5% perf-profile.children.cycles-pp.page_counter_uncharge
0.00 +2.3 2.32 ? 4% perf-profile.children.cycles-pp.drain_stock
0.00 +2.4 2.40 ? 5% perf-profile.children.cycles-pp.page_counter_try_charge
10.87 ? 3% +2.4 13.28 ? 4% perf-profile.children.cycles-pp.__local_bh_enable_ip
10.76 ? 3% +2.4 13.20 ? 4% perf-profile.children.cycles-pp.do_softirq
17.84 ? 3% +2.5 20.32 ? 4% perf-profile.children.cycles-pp.__sys_recvmsg
3.60 ? 3% +2.5 6.12 ? 5% perf-profile.children.cycles-pp.sctp_skb_recv_datagram
10.53 ? 3% +2.5 13.07 ? 4% perf-profile.children.cycles-pp.net_rx_action
0.00 +2.5 2.54 ? 4% perf-profile.children.cycles-pp.refill_stock
10.43 ? 3% +2.6 13.01 ? 4% perf-profile.children.cycles-pp.__napi_poll
17.57 ? 3% +2.6 20.15 ? 4% perf-profile.children.cycles-pp.___sys_recvmsg
10.41 ? 3% +2.6 12.99 ? 4% perf-profile.children.cycles-pp.process_backlog
10.77 ? 3% +2.7 13.43 ? 4% perf-profile.children.cycles-pp.__softirqentry_text_start
10.24 ? 3% +2.7 12.90 ? 4% perf-profile.children.cycles-pp.__netif_receive_skb_one_core
9.97 ? 3% +2.8 12.76 ? 4% perf-profile.children.cycles-pp.ip_local_deliver_finish
9.95 ? 3% +2.8 12.74 ? 4% perf-profile.children.cycles-pp.ip_protocol_deliver_rcu
9.89 ? 3% +2.8 12.69 ? 4% perf-profile.children.cycles-pp.sctp_rcv
1.71 ? 4% +2.9 4.57 ? 4% perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
16.66 ? 3% +2.9 19.59 ? 4% perf-profile.children.cycles-pp.____sys_recvmsg
16.01 ? 3% +3.2 19.21 ? 4% perf-profile.children.cycles-pp.sctp_recvmsg
15.89 ? 3% +3.2 19.12 ? 4% perf-profile.children.cycles-pp.inet_recvmsg
1.49 ? 3% +3.4 4.86 ? 4% perf-profile.children.cycles-pp._raw_spin_lock_bh
1.59 ? 3% +3.4 4.96 ? 4% perf-profile.children.cycles-pp.lock_sock_nested
15.60 ? 3% +3.6 19.22 ? 4% perf-profile.children.cycles-pp.release_sock
0.00 +3.7 3.69 ? 4% perf-profile.children.cycles-pp.try_charge_memcg
15.33 ? 3% +3.7 19.04 ? 4% perf-profile.children.cycles-pp.__release_sock
15.27 ? 3% +3.7 19.00 ? 4% perf-profile.children.cycles-pp.sctp_backlog_rcv
0.62 ? 5% +3.8 4.45 ? 4% perf-profile.children.cycles-pp.sctp_sf_eat_data_6_2
0.48 ? 5% +3.9 4.36 ? 5% perf-profile.children.cycles-pp.sctp_eat_data
0.00 +4.4 4.36 ? 4% perf-profile.children.cycles-pp.mem_cgroup_charge_skmem
3.46 ? 3% +4.4 7.91 ? 4% perf-profile.children.cycles-pp.sctp_chunk_put
20.07 ? 3% +4.6 24.66 ? 4% perf-profile.children.cycles-pp.sctp_assoc_bh_rcv
0.00 +4.6 4.60 ? 4% perf-profile.children.cycles-pp.__sk_mem_raise_allocated
0.00 +4.6 4.63 ? 4% perf-profile.children.cycles-pp.__sk_mem_schedule
24.02 ? 12% +4.7 28.76 ? 11% perf-profile.children.cycles-pp.intel_idle
2.97 ? 3% +4.8 7.75 ? 4% perf-profile.children.cycles-pp.sctp_outq_sack
2.11 ? 4% +4.8 6.90 ? 4% perf-profile.children.cycles-pp.kfree_skb_reason
2.04 ? 3% +4.9 6.98 ? 4% perf-profile.children.cycles-pp.consume_skb
0.37 ? 3% +5.6 5.94 ? 4% perf-profile.children.cycles-pp.sctp_wfree
0.95 ? 3% +11.3 12.23 ? 4% perf-profile.children.cycles-pp.skb_release_head_state
0.00 +11.7 11.67 ? 4% perf-profile.children.cycles-pp.__sk_mem_reduce_allocated
8.68 ? 3% -3.1 5.58 ? 4% perf-profile.self.cycles-pp.copy_user_enhanced_fast_string
8.45 ? 3% -3.0 5.46 ? 5% perf-profile.self.cycles-pp.memcpy_erms
0.94 ? 4% -0.4 0.57 ? 6% perf-profile.self.cycles-pp.__slab_free
0.71 ? 3% -0.3 0.42 ? 2% perf-profile.self.cycles-pp.sctp_chunk_put
0.58 ? 3% -0.3 0.30 ? 3% perf-profile.self.cycles-pp.__copy_skb_header
0.87 ? 3% -0.3 0.59 ? 6% perf-profile.self.cycles-pp.kmem_cache_free
0.58 ? 5% -0.3 0.30 ? 6% perf-profile.self.cycles-pp.sctp_packet_config
0.82 ? 3% -0.3 0.55 ? 5% perf-profile.self.cycles-pp._raw_spin_lock_irqsave
0.65 ? 4% -0.2 0.41 ? 5% perf-profile.self.cycles-pp.__might_resched
0.41 ? 4% -0.2 0.17 ? 8% perf-profile.self.cycles-pp.rmqueue
0.63 ? 4% -0.2 0.40 ? 3% perf-profile.self.cycles-pp._raw_spin_lock
0.59 ? 4% -0.2 0.36 ? 7% perf-profile.self.cycles-pp.sctp_recvmsg
0.39 ? 4% -0.2 0.18 ? 5% perf-profile.self.cycles-pp.sctp_sendmsg_to_asoc
0.58 ? 3% -0.2 0.36 ? 3% perf-profile.self.cycles-pp._raw_spin_lock_bh
0.43 ? 4% -0.2 0.24 ? 9% perf-profile.self.cycles-pp.sctp_packet_pack
0.39 ? 4% -0.2 0.20 ? 2% perf-profile.self.cycles-pp.update_load_avg
0.56 ? 4% -0.2 0.38 ? 5% perf-profile.self.cycles-pp.copy_user_short_string
0.27 ? 5% -0.2 0.10 ? 7% perf-profile.self.cycles-pp.dst_release
0.32 ? 3% -0.2 0.16 ? 7% perf-profile.self.cycles-pp.get_page_from_freelist
0.31 ? 7% -0.2 0.15 ? 6% perf-profile.self.cycles-pp.ipv4_dst_check
0.30 ? 4% -0.2 0.14 ? 7% perf-profile.self.cycles-pp.free_unref_page_commit
0.38 ? 3% -0.2 0.22 ? 8% perf-profile.self.cycles-pp.__mod_node_page_state
0.54 ? 3% -0.2 0.38 ? 5% perf-profile.self.cycles-pp.__schedule
0.47 ? 2% -0.2 0.31 ? 5% perf-profile.self.cycles-pp.kfree
0.34 ? 4% -0.2 0.18 ? 7% perf-profile.self.cycles-pp.sctp_v4_xmit
0.23 ? 4% -0.2 0.07 ? 10% perf-profile.self.cycles-pp.update_curr
0.32 ? 12% -0.2 0.17 ? 9% perf-profile.self.cycles-pp.__rhashtable_lookup
0.51 ? 4% -0.2 0.36 ? 6% perf-profile.self.cycles-pp.__list_del_entry_valid
0.32 ? 3% -0.2 0.17 ? 4% perf-profile.self.cycles-pp.__zone_watermark_ok
0.42 ? 3% -0.1 0.27 ? 5% perf-profile.self.cycles-pp.kmem_cache_alloc
0.40 ? 5% -0.1 0.26 ? 7% perf-profile.self.cycles-pp.sctp_outq_flush_data
0.28 ? 4% -0.1 0.15 ? 6% perf-profile.self.cycles-pp.__check_object_size
0.35 ? 4% -0.1 0.22 ? 9% perf-profile.self.cycles-pp.__fdget
0.42 ? 4% -0.1 0.29 ? 6% perf-profile.self.cycles-pp.sctp_datamsg_from_user
0.36 ? 5% -0.1 0.24 ? 5% perf-profile.self.cycles-pp.set_next_entity
0.34 ? 3% -0.1 0.21 ? 3% perf-profile.self.cycles-pp.check_heap_object
0.34 ? 3% -0.1 0.22 ? 7% perf-profile.self.cycles-pp.memcg_slab_free_hook
0.32 ? 2% -0.1 0.20 ? 5% perf-profile.self.cycles-pp.__list_add_valid
0.32 ? 5% -0.1 0.20 ? 8% perf-profile.self.cycles-pp.__virt_addr_valid
0.34 ? 5% -0.1 0.22 ? 4% perf-profile.self.cycles-pp.__skb_datagram_iter
0.34 ? 4% -0.1 0.22 ? 7% perf-profile.self.cycles-pp.update_rq_clock
0.37 ? 5% -0.1 0.25 ? 4% perf-profile.self.cycles-pp.sctp_sendmsg
0.27 ? 13% -0.1 0.16 ? 6% perf-profile.self.cycles-pp.sctp_transport_hold
0.35 ? 3% -0.1 0.24 ? 6% perf-profile.self.cycles-pp.skb_release_data
0.32 ? 4% -0.1 0.20 ? 7% perf-profile.self.cycles-pp.__alloc_skb
0.22 ? 7% -0.1 0.10 ? 7% perf-profile.self.cycles-pp.sctp_assoc_rwnd_increase
0.33 ? 5% -0.1 0.22 ? 4% perf-profile.self.cycles-pp.free_pcp_prepare
0.31 ? 4% -0.1 0.20 ? 7% perf-profile.self.cycles-pp.enqueue_task_fair
0.36 ? 3% -0.1 0.26 ? 3% perf-profile.self.cycles-pp.enqueue_entity
0.30 -0.1 0.20 ? 8% perf-profile.self.cycles-pp.sctp_association_put
0.24 ? 4% -0.1 0.14 ? 7% perf-profile.self.cycles-pp.__alloc_pages
0.30 ? 9% -0.1 0.20 ? 6% perf-profile.self.cycles-pp.sctp_transport_put
0.31 ? 3% -0.1 0.21 ? 6% perf-profile.self.cycles-pp.sctp_skb_recv_datagram
0.30 ? 4% -0.1 0.20 ? 8% perf-profile.self.cycles-pp.aa_sk_perm
0.33 ? 4% -0.1 0.23 ? 6% perf-profile.self.cycles-pp.entry_SYSCALL_64_after_hwframe
0.30 ? 3% -0.1 0.20 ? 7% perf-profile.self.cycles-pp.__might_sleep
0.18 ? 4% -0.1 0.09 ? 10% perf-profile.self.cycles-pp.____sys_recvmsg
0.19 ? 5% -0.1 0.09 ? 11% perf-profile.self.cycles-pp.sctp_ulpevent_free
0.32 ? 3% -0.1 0.23 ? 8% perf-profile.self.cycles-pp.memset_erms
0.30 ? 6% -0.1 0.21 ? 7% perf-profile.self.cycles-pp.kmem_cache_alloc_node
0.28 ? 3% -0.1 0.19 ? 6% perf-profile.self.cycles-pp.skb_set_owner_w
0.23 ? 6% -0.1 0.14 ? 7% perf-profile.self.cycles-pp.__switch_to
0.18 ? 4% -0.1 0.10 ? 5% perf-profile.self.cycles-pp.__update_load_avg_cfs_rq
0.26 ? 4% -0.1 0.17 ? 7% perf-profile.self.cycles-pp.sctp_cmd_interpreter
0.25 ? 4% -0.1 0.17 ? 7% perf-profile.self.cycles-pp.entry_SYSRETQ_unsafe_stack
0.26 -0.1 0.18 ? 5% perf-profile.self.cycles-pp.send_sctp_stream_1toMany
0.17 ? 32% -0.1 0.09 ? 6% perf-profile.self.cycles-pp.sctp_cmp_addr_exact
0.22 ? 5% -0.1 0.14 ? 7% perf-profile.self.cycles-pp.__switch_to_asm
0.20 ? 2% -0.1 0.12 ? 10% perf-profile.self.cycles-pp.recv_sctp_stream_1toMany
0.19 ? 3% -0.1 0.11 ? 9% perf-profile.self.cycles-pp.sock_wfree
0.10 ? 5% -0.1 0.02 ? 99% perf-profile.self.cycles-pp.process_backlog
0.10 ? 6% -0.1 0.02 ? 99% perf-profile.self.cycles-pp.__free_one_page
0.20 ? 4% -0.1 0.13 ? 5% perf-profile.self.cycles-pp.recvmsg
0.22 ? 3% -0.1 0.14 ? 7% perf-profile.self.cycles-pp.consume_skb
0.18 ? 3% -0.1 0.11 ? 10% perf-profile.self.cycles-pp._copy_from_user
0.15 ? 7% -0.1 0.08 ? 4% perf-profile.self.cycles-pp.___perf_sw_event
0.15 ? 4% -0.1 0.08 ? 5% perf-profile.self.cycles-pp.check_new_pages
0.20 ? 5% -0.1 0.13 ? 6% perf-profile.self.cycles-pp.available_idle_cpu
0.22 ? 3% -0.1 0.15 ? 5% perf-profile.self.cycles-pp.__entry_text_start
0.18 ? 2% -0.1 0.11 ? 7% perf-profile.self.cycles-pp.sctp_assoc_bh_rcv
0.13 ? 7% -0.1 0.06 ? 7% perf-profile.self.cycles-pp.sctp_sock_rfree
0.17 ? 4% -0.1 0.11 ? 10% perf-profile.self.cycles-pp.syscall_return_via_sysret
0.13 ? 5% -0.1 0.07 ? 13% perf-profile.self.cycles-pp.__free_pages_ok
0.20 ? 5% -0.1 0.14 ? 9% perf-profile.self.cycles-pp.sctp_datamsg_put
0.19 ? 3% -0.1 0.12 ? 12% perf-profile.self.cycles-pp.sendmsg
0.18 ? 4% -0.1 0.12 ? 7% perf-profile.self.cycles-pp.reweight_entity
0.14 ? 7% -0.1 0.07 ? 10% perf-profile.self.cycles-pp.sctp_inet_skb_msgname
0.23 ? 3% -0.1 0.17 ? 9% perf-profile.self.cycles-pp.sctp_wfree
0.19 ? 3% -0.1 0.13 ? 11% perf-profile.self.cycles-pp.update_cfs_group
0.17 ? 4% -0.1 0.11 ? 3% perf-profile.self.cycles-pp.nr_iowait_cpu
0.17 ? 7% -0.1 0.11 ? 9% perf-profile.self.cycles-pp.sctp_packet_transmit
0.08 ? 18% -0.1 0.02 ? 99% perf-profile.self.cycles-pp.sctp_ulpq_tail_data
0.17 ? 7% -0.1 0.11 ? 9% perf-profile.self.cycles-pp.sctp_packet_append_chunk
0.19 ? 4% -0.1 0.13 ? 8% perf-profile.self.cycles-pp.sctp_do_sm
0.18 ? 2% -0.1 0.12 ? 4% perf-profile.self.cycles-pp.__free_pages
0.17 ? 5% -0.1 0.12 ? 6% perf-profile.self.cycles-pp.__wake_up_common
0.16 ? 6% -0.1 0.10 ? 3% perf-profile.self.cycles-pp.sctp_sendmsg_parse
0.15 ? 6% -0.1 0.09 ? 13% perf-profile.self.cycles-pp.try_to_wake_up
0.10 ? 3% -0.1 0.04 ? 45% perf-profile.self.cycles-pp.__build_skb_around
0.15 ? 6% -0.1 0.09 ? 11% perf-profile.self.cycles-pp.sctp_check_transmitted
0.08 ? 14% -0.1 0.03 ?101% perf-profile.self.cycles-pp.select_task_rq_fair
0.12 ? 6% -0.1 0.06 ? 14% perf-profile.self.cycles-pp.dequeue_entity
0.19 ? 4% -0.1 0.14 ? 10% perf-profile.self.cycles-pp.sctp_hash_cmp
0.14 ? 4% -0.1 0.09 ? 10% perf-profile.self.cycles-pp.switch_mm_irqs_off
0.16 ? 4% -0.1 0.10 ? 7% perf-profile.self.cycles-pp.poll_idle
0.22 ? 8% -0.1 0.17 ? 10% perf-profile.self.cycles-pp.sctp_rcv
0.11 ? 4% -0.1 0.06 ? 11% perf-profile.self.cycles-pp.__ip_queue_xmit
0.12 ? 10% -0.1 0.08 ? 24% perf-profile.self.cycles-pp.__sctp_packet_append_chunk
0.13 ? 2% -0.0 0.08 ? 5% perf-profile.self.cycles-pp.sctp_inq_pop
0.18 ? 3% -0.0 0.13 ? 11% perf-profile.self.cycles-pp.__update_load_avg_se
0.14 ? 5% -0.0 0.09 ? 5% perf-profile.self.cycles-pp.do_syscall_64
0.11 ? 6% -0.0 0.06 ? 11% perf-profile.self.cycles-pp._copy_to_iter
0.09 ? 11% -0.0 0.04 ? 76% perf-profile.self.cycles-pp.sctp_chunk_assign_ssn
0.14 ? 3% -0.0 0.10 ? 5% perf-profile.self.cycles-pp.____sys_sendmsg
0.12 ? 4% -0.0 0.07 ? 6% perf-profile.self.cycles-pp.___sys_recvmsg
0.20 ? 2% -0.0 0.15 ? 8% perf-profile.self.cycles-pp.sctp_association_hold
0.17 ? 3% -0.0 0.12 ? 7% perf-profile.self.cycles-pp.__skb_clone
0.14 ? 3% -0.0 0.09 ? 7% perf-profile.self.cycles-pp.__copy_msghdr_from_user
0.13 ? 5% -0.0 0.08 ? 8% perf-profile.self.cycles-pp.sctp_sf_eat_data_6_2
0.13 ? 2% -0.0 0.09 ? 13% perf-profile.self.cycles-pp.sctp_outq_tail
0.20 ? 4% -0.0 0.16 ? 9% perf-profile.self.cycles-pp.read_tsc
0.13 ? 5% -0.0 0.08 ? 5% perf-profile.self.cycles-pp.__import_iovec
0.13 ? 6% -0.0 0.08 ? 5% perf-profile.self.cycles-pp.__genradix_ptr
0.11 ? 8% -0.0 0.07 ? 14% perf-profile.self.cycles-pp.select_task_rq
0.12 ? 3% -0.0 0.08 ? 10% perf-profile.self.cycles-pp.resched_curr
0.13 ? 5% -0.0 0.09 ? 5% perf-profile.self.cycles-pp.pick_next_entity
0.12 ? 6% -0.0 0.08 ? 12% perf-profile.self.cycles-pp.__put_user_nocheck_4
0.12 ? 6% -0.0 0.08 ? 6% perf-profile.self.cycles-pp.sctp_chunk_abandoned
0.10 ? 7% -0.0 0.06 ? 11% perf-profile.self.cycles-pp.syscall_exit_to_user_mode_prepare
0.07 ? 7% -0.0 0.02 ? 99% perf-profile.self.cycles-pp.ip_rcv_core
0.13 ? 7% -0.0 0.09 ? 7% perf-profile.self.cycles-pp.sctp_sched_dequeue_common
0.13 ? 7% -0.0 0.09 ? 6% perf-profile.self.cycles-pp.ip_finish_output2
0.12 ? 6% -0.0 0.08 ? 10% perf-profile.self.cycles-pp.memcg_slab_post_alloc_hook
0.08 ? 6% -0.0 0.04 ? 71% perf-profile.self.cycles-pp._sctp_make_chunk
0.10 ? 4% -0.0 0.06 ? 11% perf-profile.self.cycles-pp.net_rx_action
0.10 ? 5% -0.0 0.06 ? 8% perf-profile.self.cycles-pp.sctp_addto_chunk
0.09 ? 6% -0.0 0.05 ? 7% perf-profile.self.cycles-pp.__netif_receive_skb_core
0.12 ? 5% -0.0 0.08 ? 5% perf-profile.self.cycles-pp.__kmalloc
0.11 ? 5% -0.0 0.07 ? 5% perf-profile.self.cycles-pp.__put_user_nocheck_8
0.18 ? 6% -0.0 0.15 ? 7% perf-profile.self.cycles-pp.native_sched_clock
0.09 ? 5% -0.0 0.06 ? 8% perf-profile.self.cycles-pp.check_stack_object
0.07 -0.0 0.03 ? 70% perf-profile.self.cycles-pp.__mod_lruvec_page_state
0.16 ? 4% -0.0 0.12 ? 10% perf-profile.self.cycles-pp.sctp_outq_flush
0.10 ? 3% -0.0 0.06 ? 7% perf-profile.self.cycles-pp.skb_put
0.06 ? 6% -0.0 0.02 ? 99% perf-profile.self.cycles-pp.entry_SYSCALL_64_safe_stack
0.10 ? 4% -0.0 0.07 ? 15% perf-profile.self.cycles-pp.__might_fault
0.12 ? 5% -0.0 0.08 ? 5% perf-profile.self.cycles-pp.kmem_cache_alloc_trace
0.07 ? 8% -0.0 0.04 ? 71% perf-profile.self.cycles-pp.sockfd_lookup_light
0.10 ? 7% -0.0 0.07 ? 10% perf-profile.self.cycles-pp.__cond_resched
0.09 ? 6% -0.0 0.06 ? 8% perf-profile.self.cycles-pp.os_xsave
0.09 ? 6% -0.0 0.06 ? 8% perf-profile.self.cycles-pp.free_unref_page
0.09 ? 7% -0.0 0.06 ? 6% perf-profile.self.cycles-pp.security_socket_recvmsg
0.11 ? 4% -0.0 0.08 ? 7% perf-profile.self.cycles-pp.__local_bh_enable_ip
0.12 ? 4% -0.0 0.09 ? 12% perf-profile.self.cycles-pp.sock_kmalloc
0.09 ? 6% -0.0 0.06 ? 6% perf-profile.self.cycles-pp.sctp_user_addto_chunk
0.07 ? 6% -0.0 0.04 ? 44% perf-profile.self.cycles-pp.rcu_all_qs
0.10 ? 5% -0.0 0.06 ? 11% perf-profile.self.cycles-pp.sctp_outq_sack
0.09 ? 4% -0.0 0.06 ? 6% perf-profile.self.cycles-pp.sctp_outq_select_transport
0.10 ? 3% -0.0 0.07 ? 9% perf-profile.self.cycles-pp.__kmalloc_node_track_caller
0.08 ? 5% -0.0 0.05 ? 46% perf-profile.self.cycles-pp.sctp_ulpevent_init
0.12 ? 4% -0.0 0.08 ? 8% perf-profile.self.cycles-pp.__check_heap_object
0.07 ? 11% -0.0 0.04 ? 44% perf-profile.self.cycles-pp.loopback_xmit
0.14 ? 2% -0.0 0.12 ? 6% perf-profile.self.cycles-pp.do_idle
0.11 ? 4% -0.0 0.08 ? 12% perf-profile.self.cycles-pp.cpuidle_idle_call
0.09 -0.0 0.06 ? 11% perf-profile.self.cycles-pp._copy_from_iter
0.08 ? 5% -0.0 0.06 ? 13% perf-profile.self.cycles-pp.kmalloc_large_node
0.08 ? 5% -0.0 0.06 ? 9% perf-profile.self.cycles-pp.sctp_chunkify
0.09 ? 4% -0.0 0.06 ? 7% perf-profile.self.cycles-pp.iovec_from_user
0.14 ? 4% -0.0 0.12 ? 11% perf-profile.self.cycles-pp.sctp_ulpevent_make_rcvmsg
0.07 ? 6% -0.0 0.05 ? 7% perf-profile.self.cycles-pp.sctp_tsnmap_check
0.08 ? 6% -0.0 0.06 ? 9% perf-profile.self.cycles-pp.sock_kfree_s
0.08 ? 5% -0.0 0.06 ? 11% perf-profile.self.cycles-pp.skb_release_head_state
0.07 ? 8% -0.0 0.05 perf-profile.self.cycles-pp.__wrgsbase_inactive
0.09 ? 8% -0.0 0.07 ? 13% perf-profile.self.cycles-pp.__softirqentry_text_start
0.07 ? 5% -0.0 0.05 ? 8% perf-profile.self.cycles-pp.__sys_recvmsg
0.18 ? 4% +0.0 0.22 ? 12% perf-profile.self.cycles-pp.menu_select
0.00 +0.1 0.06 ? 19% perf-profile.self.cycles-pp.update_sg_lb_stats
0.00 +0.1 0.06 ? 7% perf-profile.self.cycles-pp.lapic_next_deadline
0.00 +0.1 0.07 ? 16% perf-profile.self.cycles-pp.native_irq_return_iret
0.09 ? 4% +0.1 0.16 ? 5% perf-profile.self.cycles-pp.cpuidle_enter_state
0.00 +0.1 0.08 ? 41% perf-profile.self.cycles-pp._raw_spin_trylock
0.00 +0.1 0.09 ? 54% perf-profile.self.cycles-pp.tick_nohz_next_event
0.01 ?223% +0.1 0.13 ? 5% perf-profile.self.cycles-pp.sctp_get_af_specific
0.00 +0.1 0.13 ? 6% perf-profile.self.cycles-pp.mem_cgroup_charge_skmem
0.00 +0.1 0.14 ? 19% perf-profile.self.cycles-pp.cgroup_rstat_updated
0.25 ? 6% +0.2 0.43 ? 14% perf-profile.self.cycles-pp.ktime_get
0.00 +0.2 0.20 ? 16% perf-profile.self.cycles-pp.timekeeping_max_deferment
0.00 +0.2 0.22 ? 6% perf-profile.self.cycles-pp.refill_stock
0.00 +0.3 0.25 ? 9% perf-profile.self.cycles-pp.__sk_mem_raise_allocated
0.00 +0.3 0.32 ? 12% perf-profile.self.cycles-pp.propagate_protected_usage
0.00 +0.9 0.89 ? 4% perf-profile.self.cycles-pp.__mod_memcg_state
0.00 +1.3 1.27 ? 5% perf-profile.self.cycles-pp.try_charge_memcg
0.00 +2.1 2.10 ? 5% perf-profile.self.cycles-pp.page_counter_uncharge
0.00 +2.2 2.23 ? 5% perf-profile.self.cycles-pp.page_counter_try_charge
1.70 ? 4% +2.8 4.54 ? 4% perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
0.29 ? 5% +4.0 4.24 ? 5% perf-profile.self.cycles-pp.sctp_eat_data
0.00 +8.6 8.55 ? 4% perf-profile.self.cycles-pp.__sk_mem_reduce_allocated





Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


--
0-DAY CI Kernel Test Service
https://01.org/lkp



Attachments:
(No filename) (157.14 kB)
config-5.19.0-rc1-00338-g4890b686f408 (166.42 kB)
job-script (8.47 kB)
job.yaml (5.89 kB)
reproduce (2.79 kB)
Download all attachments

2022-06-23 00:45:04

by Jakub Kicinski

[permalink] [raw]
Subject: Re: [net] 4890b686f4: netperf.Throughput_Mbps -69.4% regression

Could someone working on SCTP double check this is a real regression?
Feels like the regression reports are flowing at such rate its hard
to keep up.

On Sun, 19 Jun 2022 23:04:56 +0800 kernel test robot wrote:
> Greeting,
>
> FYI, we noticed a -69.4% regression of netperf.Throughput_Mbps due to commit:
>
>
> commit: 4890b686f4088c90432149bd6de567e621266fa2 ("net: keep sk->sk_forward_alloc as small as possible")
> https://git.kernel.org/cgit/linux/kernel/git/next/linux-next.git master
>
> in testcase: netperf
> on test machine: 144 threads 4 sockets Intel(R) Xeon(R) Gold 5318H CPU @ 2.50GHz with 128G memory
> with following parameters:
>
> ip: ipv4
> runtime: 300s
> nr_threads: 25%
> cluster: cs-localhost
> send_size: 10K
> test: SCTP_STREAM_MANY
> cpufreq_governor: performance
> ucode: 0x7002402
>
> test-description: Netperf is a benchmark that can be use to measure various aspect of networking performance.
> test-url: http://www.netperf.org/netperf/
>
> In addition to that, the commit also has significant impact on the following tests:
>
> +------------------+----------------------------------------------------------------------------------+
> | testcase: change | netperf: netperf.Throughput_Mbps -73.7% regression |
> | test machine | 144 threads 4 sockets Intel(R) Xeon(R) Gold 5318H CPU @ 2.50GHz with 128G memory |
> | test parameters | cluster=cs-localhost |
> | | cpufreq_governor=performance |
> | | ip=ipv4 |
> | | nr_threads=50% |
> | | runtime=300s |
> | | send_size=10K |
> | | test=SCTP_STREAM_MANY |
> | | ucode=0x7002402 |
> +------------------+----------------------------------------------------------------------------------+
>
>
> If you fix the issue, kindly add following tag
> Reported-by: kernel test robot <[email protected]>
>
>
> Details are as below:
> -------------------------------------------------------------------------------------------------->
>
>
> To reproduce:
>
> git clone https://github.com/intel/lkp-tests.git
> cd lkp-tests
> sudo bin/lkp install job.yaml # job file is attached in this email
> bin/lkp split-job --compatible job.yaml # generate the yaml file for lkp run
> sudo bin/lkp run generated-yaml-file
>
> # if come across any failure that blocks the test,
> # please remove ~/.lkp and /lkp dir to run from a clean state.
>
> =========================================================================================
> cluster/compiler/cpufreq_governor/ip/kconfig/nr_threads/rootfs/runtime/send_size/tbox_group/test/testcase/ucode:
> cs-localhost/gcc-11/performance/ipv4/x86_64-rhel-8.3/25%/debian-11.1-x86_64-20220510.cgz/300s/10K/lkp-cpl-4sp1/SCTP_STREAM_MANY/netperf/0x7002402
>
> commit:
> 7c80b038d2 ("net: fix sk_wmem_schedule() and sk_rmem_schedule() errors")
> 4890b686f4 ("net: keep sk->sk_forward_alloc as small as possible")
>
> 7c80b038d23e1f4c 4890b686f4088c90432149bd6de
> ---------------- ---------------------------
> %stddev %change %stddev
> \ | \
> 15855 -69.4% 4854 netperf.Throughput_Mbps
> 570788 -69.4% 174773 netperf.Throughput_total_Mbps
> 30543 -59.1% 12480 ± 3% netperf.time.involuntary_context_switches
> 16661 ± 2% -20.1% 13317 ± 2% netperf.time.minor_page_faults
> 2076 -58.1% 869.17 ± 2% netperf.time.percent_of_cpu_this_job_got
> 6118 -57.8% 2583 ± 2% netperf.time.system_time
> 143.71 ± 6% -72.3% 39.77 netperf.time.user_time
> 87371 ± 8% +81.1% 158253 netperf.time.voluntary_context_switches
> 2.09e+09 -69.4% 6.4e+08 netperf.workload
> 36264 +18.2% 42849 uptime.idle
> 3.058e+10 +22.0% 3.731e+10 cpuidle..time
> 6.536e+08 -60.7% 2.566e+08 cpuidle..usage
> 127819 ± 79% -82.3% 22609 ± 65% numa-meminfo.node3.Inactive
> 127764 ± 79% -82.4% 22440 ± 66% numa-meminfo.node3.Inactive(anon)
> 71.38 +14.7 86.09 mpstat.cpu.all.idle%
> 4.84 -2.1 2.78 ± 2% mpstat.cpu.all.soft%
> 22.32 -12.1 10.25 ± 2% mpstat.cpu.all.sys%
> 0.78 ± 3% -0.5 0.25 mpstat.cpu.all.usr%
> 71.00 +20.9% 85.83 vmstat.cpu.id
> 4121749 -23.6% 3150424 vmstat.memory.cache
> 43.00 ± 2% -55.0% 19.33 ± 2% vmstat.procs.r
> 3921783 -69.4% 1200263 vmstat.system.cs
> 292198 -1.3% 288478 vmstat.system.in
> 6.908e+08 ± 3% -70.1% 2.064e+08 ± 9% numa-numastat.node0.local_node
> 6.907e+08 ± 3% -70.1% 2.065e+08 ± 9% numa-numastat.node0.numa_hit
> 6.517e+08 ± 2% -65.5% 2.246e+08 ± 10% numa-numastat.node1.local_node
> 6.519e+08 ± 2% -65.5% 2.247e+08 ± 10% numa-numastat.node1.numa_hit
> 6.81e+08 ± 3% -72.1% 1.902e+08 ± 8% numa-numastat.node2.local_node
> 6.811e+08 ± 3% -72.0% 1.905e+08 ± 8% numa-numastat.node2.numa_hit
> 6.819e+08 ± 3% -69.9% 2.05e+08 ± 6% numa-numastat.node3.local_node
> 6.823e+08 ± 3% -69.9% 2.051e+08 ± 6% numa-numastat.node3.numa_hit
> 6.907e+08 ± 3% -70.1% 2.065e+08 ± 9% numa-vmstat.node0.numa_hit
> 6.908e+08 ± 3% -70.1% 2.064e+08 ± 9% numa-vmstat.node0.numa_local
> 6.519e+08 ± 2% -65.5% 2.247e+08 ± 10% numa-vmstat.node1.numa_hit
> 6.517e+08 ± 2% -65.5% 2.246e+08 ± 10% numa-vmstat.node1.numa_local
> 6.811e+08 ± 3% -72.0% 1.905e+08 ± 8% numa-vmstat.node2.numa_hit
> 6.81e+08 ± 3% -72.1% 1.902e+08 ± 8% numa-vmstat.node2.numa_local
> 31948 ± 79% -82.4% 5622 ± 66% numa-vmstat.node3.nr_inactive_anon
> 31948 ± 79% -82.4% 5622 ± 66% numa-vmstat.node3.nr_zone_inactive_anon
> 6.823e+08 ± 3% -69.9% 2.051e+08 ± 6% numa-vmstat.node3.numa_hit
> 6.819e+08 ± 3% -69.9% 2.05e+08 ± 6% numa-vmstat.node3.numa_local
> 1216292 ± 5% -70.1% 363637 ± 4% meminfo.Active
> 1215873 ± 5% -70.1% 363097 ± 4% meminfo.Active(anon)
> 3994598 -24.3% 3022264 meminfo.Cached
> 1878317 ± 4% -53.7% 870428 meminfo.Committed_AS
> 431272 ± 8% -28.0% 310461 meminfo.Inactive
> 428860 ± 9% -28.4% 306912 meminfo.Inactive(anon)
> 668273 ± 5% -71.3% 192010 ± 4% meminfo.Mapped
> 5793516 -17.1% 4802753 meminfo.Memused
> 8320 ± 2% -12.1% 7309 ± 2% meminfo.PageTables
> 1341918 ± 6% -72.5% 368590 ± 4% meminfo.Shmem
> 9796929 -25.1% 7337138 meminfo.max_used_kB
> 1029 -51.6% 498.17 ± 2% turbostat.Avg_MHz
> 31.32 -16.2 15.17 ± 2% turbostat.Busy%
> 5.953e+08 -69.1% 1.837e+08 ± 2% turbostat.C1
> 9.22 -5.2 4.04 ± 4% turbostat.C1%
> 68.58 +23.5% 84.67 turbostat.CPU%c1
> 57.67 -16.8% 48.00 ± 3% turbostat.CoreTmp
> 0.23 ± 2% -31.4% 0.16 turbostat.IPC
> 18.59 ± 30% -18.6 0.00 turbostat.PKG_%
> 4524200 ± 6% -81.5% 835280 turbostat.POLL
> 0.10 ± 3% -0.1 0.02 ± 17% turbostat.POLL%
> 57.33 -16.6% 47.83 ± 4% turbostat.PkgTmp
> 561.39 -25.9% 416.03 turbostat.PkgWatt
> 7.75 +6.8% 8.27 turbostat.RAMWatt
> 303881 ± 5% -70.2% 90609 ± 4% proc-vmstat.nr_active_anon
> 998594 -24.4% 755404 proc-vmstat.nr_file_pages
> 107254 ± 9% -28.4% 76741 proc-vmstat.nr_inactive_anon
> 167104 ± 5% -71.0% 48414 ± 4% proc-vmstat.nr_mapped
> 2081 ± 2% -12.1% 1828 ± 2% proc-vmstat.nr_page_table_pages
> 335422 ± 6% -72.6% 91983 ± 4% proc-vmstat.nr_shmem
> 32561 -1.7% 31994 proc-vmstat.nr_slab_reclaimable
> 303881 ± 5% -70.2% 90609 ± 4% proc-vmstat.nr_zone_active_anon
> 107254 ± 9% -28.4% 76741 proc-vmstat.nr_zone_inactive_anon
> 267554 ± 15% -25.4% 199518 ± 10% proc-vmstat.numa_hint_faults
> 82267 ± 20% -27.2% 59929 ± 13% proc-vmstat.numa_hint_faults_local
> 2.706e+09 -69.5% 8.266e+08 proc-vmstat.numa_hit
> 2.705e+09 -69.5% 8.262e+08 proc-vmstat.numa_local
> 489361 ± 9% -19.4% 394229 ± 6% proc-vmstat.numa_pte_updates
> 510140 ± 18% -75.3% 125829 ± 2% proc-vmstat.pgactivate
> 7.17e+09 -69.4% 2.197e+09 proc-vmstat.pgalloc_normal
> 1636476 ± 2% -11.5% 1448944 proc-vmstat.pgfault
> 7.17e+09 -69.4% 2.197e+09 proc-vmstat.pgfree
> 30967 ± 41% -88.7% 3488 ± 30% sched_debug.cfs_rq:/.MIN_vruntime.avg
> 1604384 ± 22% -83.6% 262510 ± 14% sched_debug.cfs_rq:/.MIN_vruntime.max
> 201146 ± 30% -86.1% 28034 ± 18% sched_debug.cfs_rq:/.MIN_vruntime.stddev
> 0.29 ± 8% -43.5% 0.16 ± 9% sched_debug.cfs_rq:/.h_nr_running.avg
> 0.42 ± 3% -15.1% 0.36 ± 3% sched_debug.cfs_rq:/.h_nr_running.stddev
> 30967 ± 41% -88.7% 3488 ± 30% sched_debug.cfs_rq:/.max_vruntime.avg
> 1604384 ± 22% -83.6% 262510 ± 14% sched_debug.cfs_rq:/.max_vruntime.max
> 201146 ± 30% -86.1% 28034 ± 18% sched_debug.cfs_rq:/.max_vruntime.stddev
> 1526268 ± 11% -80.6% 295647 ± 10% sched_debug.cfs_rq:/.min_vruntime.avg
> 2336190 ± 12% -79.4% 482042 ± 7% sched_debug.cfs_rq:/.min_vruntime.max
> 859659 ± 11% -85.8% 121714 ± 18% sched_debug.cfs_rq:/.min_vruntime.min
> 343902 ± 15% -78.7% 73263 ± 6% sched_debug.cfs_rq:/.min_vruntime.stddev
> 0.29 ± 8% -43.6% 0.16 ± 9% sched_debug.cfs_rq:/.nr_running.avg
> 0.42 ± 3% -15.1% 0.36 ± 3% sched_debug.cfs_rq:/.nr_running.stddev
> 312.39 ± 7% -46.1% 168.29 ± 5% sched_debug.cfs_rq:/.runnable_avg.avg
> 355.85 ± 2% -27.3% 258.86 ± 3% sched_debug.cfs_rq:/.runnable_avg.stddev
> 984535 ± 22% -82.1% 176253 ± 31% sched_debug.cfs_rq:/.spread0.max
> -491914 -62.6% -184050 sched_debug.cfs_rq:/.spread0.min
> 343876 ± 15% -78.7% 73261 ± 6% sched_debug.cfs_rq:/.spread0.stddev
> 312.25 ± 7% -46.1% 168.23 ± 5% sched_debug.cfs_rq:/.util_avg.avg
> 355.80 ± 2% -27.3% 258.80 ± 3% sched_debug.cfs_rq:/.util_avg.stddev
> 213.78 ± 8% -62.3% 80.50 ± 11% sched_debug.cfs_rq:/.util_est_enqueued.avg
> 331.99 ± 3% -37.4% 207.79 ± 4% sched_debug.cfs_rq:/.util_est_enqueued.stddev
> 630039 +19.0% 749877 ± 3% sched_debug.cpu.avg_idle.avg
> 4262 +45.9% 6219 ± 3% sched_debug.cpu.avg_idle.min
> 2345 ± 14% -35.1% 1522 ± 6% sched_debug.cpu.clock_task.stddev
> 1496 ± 5% -47.9% 779.72 ± 7% sched_debug.cpu.curr->pid.avg
> 2528 ± 3% -19.7% 2031 ± 3% sched_debug.cpu.curr->pid.stddev
> 0.00 ± 10% -23.0% 0.00 ± 4% sched_debug.cpu.next_balance.stddev
> 0.25 ± 6% -48.0% 0.13 ± 8% sched_debug.cpu.nr_running.avg
> 0.41 ± 2% -19.9% 0.33 ± 3% sched_debug.cpu.nr_running.stddev
> 3859540 ± 10% -70.7% 1132195 ± 10% sched_debug.cpu.nr_switches.avg
> 7113795 ± 11% -64.8% 2504199 ± 10% sched_debug.cpu.nr_switches.max
> 1618475 ± 15% -86.1% 224980 ± 30% sched_debug.cpu.nr_switches.min
> 1045624 ± 9% -57.2% 447501 ± 9% sched_debug.cpu.nr_switches.stddev
> 2.446e+10 -66.9% 8.087e+09 perf-stat.i.branch-instructions
> 2.356e+08 ± 2% -65.6% 80979566 ± 3% perf-stat.i.branch-misses
> 0.74 ± 8% +5.7 6.47 perf-stat.i.cache-miss-rate%
> 16166727 ± 10% +236.6% 54420281 ± 2% perf-stat.i.cache-misses
> 2.609e+09 -67.5% 8.493e+08 ± 2% perf-stat.i.cache-references
> 3956687 -69.4% 1211999 perf-stat.i.context-switches
> 1.23 +44.9% 1.79 perf-stat.i.cpi
> 1.51e+11 -52.3% 7.199e+10 ± 2% perf-stat.i.cpu-cycles
> 363.75 -38.7% 223.06 perf-stat.i.cpu-migrations
> 10558 ± 10% -87.4% 1326 perf-stat.i.cycles-between-cache-misses
> 652403 ±109% -72.7% 178258 ± 8% perf-stat.i.dTLB-load-misses
> 3.497e+10 -66.9% 1.158e+10 perf-stat.i.dTLB-loads
> 0.00 ± 3% +0.0 0.00 ± 11% perf-stat.i.dTLB-store-miss-rate%
> 119514 ± 5% -51.8% 57642 ± 9% perf-stat.i.dTLB-store-misses
> 2.067e+10 -66.9% 6.833e+09 perf-stat.i.dTLB-stores
> 74.98 +0.8 75.82 perf-stat.i.iTLB-load-miss-rate%
> 1.648e+08 ± 3% -67.2% 54063746 ± 4% perf-stat.i.iTLB-load-misses
> 54768215 -68.8% 17107737 perf-stat.i.iTLB-loads
> 1.223e+11 -66.9% 4.052e+10 perf-stat.i.instructions
> 0.81 -30.7% 0.56 perf-stat.i.ipc
> 1.05 -52.3% 0.50 ± 2% perf-stat.i.metric.GHz
> 453.64 -46.1% 244.53 perf-stat.i.metric.K/sec
> 574.31 -66.9% 189.94 perf-stat.i.metric.M/sec
> 4984 ± 2% -12.4% 4368 perf-stat.i.minor-faults
> 4306378 ± 13% +145.4% 10568372 ± 2% perf-stat.i.node-load-misses
> 267971 ± 20% +205.5% 818656 ± 4% perf-stat.i.node-loads
> 93.32 +4.7 98.02 perf-stat.i.node-store-miss-rate%
> 1642200 ± 14% +217.5% 5214268 ± 2% perf-stat.i.node-store-misses
> 198883 ± 16% -39.3% 120629 ± 9% perf-stat.i.node-stores
> 4985 ± 2% -12.4% 4369 perf-stat.i.page-faults
> 21.33 -1.7% 20.96 perf-stat.overall.MPKI
> 0.62 ± 10% +5.8 6.41 perf-stat.overall.cache-miss-rate%
> 1.23 +44.0% 1.78 perf-stat.overall.cpi
> 9426 ± 9% -86.0% 1323 perf-stat.overall.cycles-between-cache-misses
> 0.00 ± 4% +0.0 0.00 ± 9% perf-stat.overall.dTLB-store-miss-rate%
> 75.04 +0.9 75.95 perf-stat.overall.iTLB-load-miss-rate%
> 0.81 -30.5% 0.56 perf-stat.overall.ipc
> 89.11 +8.6 97.73 perf-stat.overall.node-store-miss-rate%
> 17633 +8.3% 19103 perf-stat.overall.path-length
> 2.438e+10 -66.9% 8.061e+09 perf-stat.ps.branch-instructions
> 2.348e+08 ± 2% -65.6% 80709139 ± 3% perf-stat.ps.branch-misses
> 16118295 ± 10% +236.5% 54240216 ± 2% perf-stat.ps.cache-misses
> 2.6e+09 -67.4% 8.464e+08 ± 2% perf-stat.ps.cache-references
> 3943074 -69.4% 1207953 perf-stat.ps.context-switches
> 1.505e+11 -52.3% 7.175e+10 ± 2% perf-stat.ps.cpu-cycles
> 362.72 -38.7% 222.48 perf-stat.ps.cpu-migrations
> 650701 ±109% -72.7% 177916 ± 8% perf-stat.ps.dTLB-load-misses
> 3.485e+10 -66.9% 1.155e+10 perf-stat.ps.dTLB-loads
> 119179 ± 5% -51.8% 57464 ± 9% perf-stat.ps.dTLB-store-misses
> 2.06e+10 -66.9% 6.81e+09 perf-stat.ps.dTLB-stores
> 1.642e+08 ± 3% -67.2% 53885104 ± 4% perf-stat.ps.iTLB-load-misses
> 54583301 -68.8% 17050690 perf-stat.ps.iTLB-loads
> 1.219e+11 -66.9% 4.039e+10 perf-stat.ps.instructions
> 4965 ± 2% -12.3% 4352 perf-stat.ps.minor-faults
> 4293134 ± 13% +145.4% 10534117 ± 2% perf-stat.ps.node-load-misses
> 267097 ± 20% +205.5% 816047 ± 4% perf-stat.ps.node-loads
> 1637287 ± 13% +217.4% 5197315 ± 2% perf-stat.ps.node-store-misses
> 198487 ± 15% -39.3% 120430 ± 9% perf-stat.ps.node-stores
> 4966 ± 2% -12.3% 4353 perf-stat.ps.page-faults
> 3.686e+13 -66.8% 1.223e+13 perf-stat.total.instructions
> 9.05 ± 7% -3.4 5.70 ± 6% perf-profile.calltrace.cycles-pp.sctp_packet_pack.sctp_packet_transmit.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm
> 8.94 ± 7% -3.2 5.69 ± 5% perf-profile.calltrace.cycles-pp.sctp_do_sm.sctp_primitive_SEND.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg
> 8.61 ± 7% -3.2 5.44 ± 6% perf-profile.calltrace.cycles-pp.memcpy_erms.sctp_packet_pack.sctp_packet_transmit.sctp_outq_flush.sctp_cmd_interpreter
> 5.69 ± 4% -3.0 2.68 ± 7% perf-profile.calltrace.cycles-pp.mwait_idle_with_hints.intel_idle_irq.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call
> 5.85 ± 4% -3.0 2.90 ± 8% perf-profile.calltrace.cycles-pp.intel_idle_irq.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
> 21.32 ± 6% -2.9 18.43 ± 5% perf-profile.calltrace.cycles-pp.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg.____sys_sendmsg.___sys_sendmsg
> 13.45 ± 7% -2.5 10.92 ± 5% perf-profile.calltrace.cycles-pp.sctp_primitive_SEND.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg.____sys_sendmsg
> 7.02 ± 7% -2.5 4.51 ± 6% perf-profile.calltrace.cycles-pp.sctp_datamsg_from_user.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg.____sys_sendmsg
> 6.53 ± 6% -2.5 4.06 ± 6% perf-profile.calltrace.cycles-pp.skb_copy_datagram_iter.sctp_recvmsg.inet_recvmsg.____sys_recvmsg.___sys_recvmsg
> 6.51 ± 6% -2.5 4.05 ± 6% perf-profile.calltrace.cycles-pp.__skb_datagram_iter.skb_copy_datagram_iter.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
> 5.68 ± 6% -2.1 3.54 ± 6% perf-profile.calltrace.cycles-pp._copy_to_iter.__skb_datagram_iter.skb_copy_datagram_iter.sctp_recvmsg.inet_recvmsg
> 5.52 ± 6% -2.1 3.45 ± 6% perf-profile.calltrace.cycles-pp.copyout._copy_to_iter.__skb_datagram_iter.skb_copy_datagram_iter.sctp_recvmsg
> 5.48 ± 6% -2.1 3.42 ± 6% perf-profile.calltrace.cycles-pp.copy_user_enhanced_fast_string.copyout._copy_to_iter.__skb_datagram_iter.skb_copy_datagram_iter
> 11.97 ± 7% -2.0 9.98 ± 5% perf-profile.calltrace.cycles-pp.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm.sctp_primitive_SEND.sctp_sendmsg_to_asoc
> 9.16 ± 7% -1.8 7.37 ± 5% perf-profile.calltrace.cycles-pp.sctp_cmd_interpreter.sctp_do_sm.sctp_primitive_SEND.sctp_sendmsg_to_asoc.sctp_sendmsg
> 10.42 ± 7% -1.4 9.00 ± 5% perf-profile.calltrace.cycles-pp.sctp_packet_transmit.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm.sctp_primitive_SEND
> 1.58 ± 6% -1.2 0.35 ± 70% perf-profile.calltrace.cycles-pp.__alloc_pages.kmalloc_large_node.__kmalloc_node_track_caller.kmalloc_reserve.__alloc_skb
> 3.49 ± 6% -1.2 2.30 ± 7% perf-profile.calltrace.cycles-pp.sctp_user_addto_chunk.sctp_datamsg_from_user.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg
> 3.15 ± 6% -1.1 2.06 ± 6% perf-profile.calltrace.cycles-pp._copy_from_iter.sctp_user_addto_chunk.sctp_datamsg_from_user.sctp_sendmsg_to_asoc.sctp_sendmsg
> 2.86 ± 7% -1.1 1.79 ± 7% perf-profile.calltrace.cycles-pp.sctp_make_datafrag_empty.sctp_datamsg_from_user.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg
> 3.16 ± 7% -1.1 2.11 ± 5% perf-profile.calltrace.cycles-pp.sctp_data_ready.sctp_ulpq_tail_event.sctp_ulpq_tail_data.sctp_cmd_interpreter.sctp_do_sm
> 3.00 ± 6% -1.0 1.96 ± 6% perf-profile.calltrace.cycles-pp.copyin._copy_from_iter.sctp_user_addto_chunk.sctp_datamsg_from_user.sctp_sendmsg_to_asoc
> 2.97 ± 6% -1.0 1.93 ± 6% perf-profile.calltrace.cycles-pp.copy_user_enhanced_fast_string.copyin._copy_from_iter.sctp_user_addto_chunk.sctp_datamsg_from_user
> 3.26 ± 7% -1.0 2.22 ± 5% perf-profile.calltrace.cycles-pp.sctp_ulpq_tail_event.sctp_ulpq_tail_data.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv
> 2.89 ± 7% -1.0 1.87 ± 5% perf-profile.calltrace.cycles-pp.__wake_up_common_lock.sctp_data_ready.sctp_ulpq_tail_event.sctp_ulpq_tail_data.sctp_cmd_interpreter
> 2.68 ± 6% -0.9 1.74 ± 5% perf-profile.calltrace.cycles-pp.__wake_up_common.__wake_up_common_lock.sctp_data_ready.sctp_ulpq_tail_event.sctp_ulpq_tail_data
> 2.40 ± 7% -0.9 1.46 ± 6% perf-profile.calltrace.cycles-pp._sctp_make_chunk.sctp_make_datafrag_empty.sctp_datamsg_from_user.sctp_sendmsg_to_asoc.sctp_sendmsg
> 2.49 ± 7% -0.9 1.62 ± 5% perf-profile.calltrace.cycles-pp.autoremove_wake_function.__wake_up_common.__wake_up_common_lock.sctp_data_ready.sctp_ulpq_tail_event
> 2.46 ± 6% -0.9 1.60 ± 5% perf-profile.calltrace.cycles-pp.try_to_wake_up.autoremove_wake_function.__wake_up_common.__wake_up_common_lock.sctp_data_ready
> 1.86 ± 6% -0.8 1.10 ± 6% perf-profile.calltrace.cycles-pp.schedule_timeout.sctp_skb_recv_datagram.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
> 1.81 ± 7% -0.8 1.06 ± 5% perf-profile.calltrace.cycles-pp.__alloc_skb._sctp_make_chunk.sctp_make_datafrag_empty.sctp_datamsg_from_user.sctp_sendmsg_to_asoc
> 1.76 ± 9% -0.7 1.01 ± 6% perf-profile.calltrace.cycles-pp.sctp_ulpevent_free.sctp_recvmsg.inet_recvmsg.____sys_recvmsg.___sys_recvmsg
> 1.82 ± 6% -0.7 1.08 ± 6% perf-profile.calltrace.cycles-pp.schedule.schedule_timeout.sctp_skb_recv_datagram.sctp_recvmsg.inet_recvmsg
> 1.50 ± 6% -0.7 0.76 ± 5% perf-profile.calltrace.cycles-pp.skb_release_data.kfree_skb_reason.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
> 1.79 ± 6% -0.7 1.06 ± 6% perf-profile.calltrace.cycles-pp.__schedule.schedule.schedule_timeout.sctp_skb_recv_datagram.sctp_recvmsg
> 1.61 ± 6% -0.6 1.01 ± 6% perf-profile.calltrace.cycles-pp.schedule_idle.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
> 0.86 ± 7% -0.6 0.26 ±100% perf-profile.calltrace.cycles-pp.dequeue_entity.dequeue_task_fair.__schedule.schedule.schedule_timeout
> 1.56 ± 6% -0.6 0.98 ± 6% perf-profile.calltrace.cycles-pp.__schedule.schedule_idle.do_idle.cpu_startup_entry.start_secondary
> 1.56 ± 8% -0.6 0.99 ± 6% perf-profile.calltrace.cycles-pp.sctp_outq_flush_data.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm.sctp_primitive_SEND
> 1.25 ± 7% -0.6 0.68 ± 5% perf-profile.calltrace.cycles-pp.kmalloc_reserve.__alloc_skb._sctp_make_chunk.sctp_make_datafrag_empty.sctp_datamsg_from_user
> 1.22 ± 7% -0.6 0.66 ± 5% perf-profile.calltrace.cycles-pp.__kmalloc_node_track_caller.kmalloc_reserve.__alloc_skb._sctp_make_chunk.sctp_make_datafrag_empty
> 1.19 ± 7% -0.5 0.64 ± 5% perf-profile.calltrace.cycles-pp.kmalloc_large_node.__kmalloc_node_track_caller.kmalloc_reserve.__alloc_skb._sctp_make_chunk
> 0.90 ± 7% -0.5 0.39 ± 70% perf-profile.calltrace.cycles-pp.skb_release_data.consume_skb.sctp_chunk_put.sctp_outq_sack.sctp_cmd_interpreter
> 0.94 ± 7% -0.5 0.46 ± 45% perf-profile.calltrace.cycles-pp.dequeue_task_fair.__schedule.schedule.schedule_timeout.sctp_skb_recv_datagram
> 0.74 ± 6% -0.5 0.27 ±100% perf-profile.calltrace.cycles-pp.sctp_endpoint_lookup_assoc.sctp_sendmsg.sock_sendmsg.____sys_sendmsg.___sys_sendmsg
> 0.98 ± 8% -0.4 0.57 ± 7% perf-profile.calltrace.cycles-pp.sctp_chunk_put.sctp_ulpevent_free.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
> 1.05 ± 8% -0.4 0.66 ± 7% perf-profile.calltrace.cycles-pp.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv.sctp_rcv
> 1.10 ± 7% -0.4 0.72 ± 6% perf-profile.calltrace.cycles-pp.sendmsg_copy_msghdr.___sys_sendmsg.__sys_sendmsg.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.46 ± 45% +0.4 0.83 ± 27% perf-profile.calltrace.cycles-pp.menu_select.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary
> 0.00 +0.8 0.84 ± 24% perf-profile.calltrace.cycles-pp.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state
> 0.00 +0.8 0.84 ± 24% perf-profile.calltrace.cycles-pp.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter
> 0.00 +0.9 0.85 ± 6% perf-profile.calltrace.cycles-pp.page_counter_uncharge.drain_stock.refill_stock.__sk_mem_reduce_allocated.skb_release_head_state
> 0.00 +0.9 0.87 ± 6% perf-profile.calltrace.cycles-pp.drain_stock.refill_stock.__sk_mem_reduce_allocated.skb_release_head_state.kfree_skb_reason
> 0.00 +1.0 0.98 ± 7% perf-profile.calltrace.cycles-pp.refill_stock.__sk_mem_reduce_allocated.skb_release_head_state.kfree_skb_reason.sctp_recvmsg
> 0.00 +1.2 1.24 ± 5% perf-profile.calltrace.cycles-pp.page_counter_uncharge.drain_stock.refill_stock.__sk_mem_reduce_allocated.sctp_wfree
> 0.00 +1.3 1.26 ± 5% perf-profile.calltrace.cycles-pp.drain_stock.refill_stock.__sk_mem_reduce_allocated.sctp_wfree.skb_release_head_state
> 0.00 +1.3 1.27 ± 19% perf-profile.calltrace.cycles-pp.try_charge_memcg.mem_cgroup_charge_skmem.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_ulpevent_make_rcvmsg
> 0.00 +1.3 1.29 ± 26% perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call
> 1.60 ± 10% +1.3 2.92 ± 5% perf-profile.calltrace.cycles-pp.sctp_ulpevent_make_rcvmsg.sctp_ulpq_tail_data.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv
> 0.00 +1.4 1.35 ± 5% perf-profile.calltrace.cycles-pp.refill_stock.__sk_mem_reduce_allocated.sctp_wfree.skb_release_head_state.consume_skb
> 0.00 +1.4 1.42 ± 25% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
> 0.00 +1.6 1.63 ± 5% perf-profile.calltrace.cycles-pp.mem_cgroup_charge_skmem.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_ulpevent_make_rcvmsg.sctp_ulpq_tail_data
> 0.00 +1.7 1.66 ± 19% perf-profile.calltrace.cycles-pp.page_counter_try_charge.try_charge_memcg.mem_cgroup_charge_skmem.__sk_mem_raise_allocated.__sk_mem_schedule
> 10.96 ± 7% +1.7 12.67 ± 5% perf-profile.calltrace.cycles-pp.__ip_queue_xmit.sctp_packet_transmit.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm
> 0.00 +1.7 1.72 ± 5% perf-profile.calltrace.cycles-pp.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_ulpevent_make_rcvmsg.sctp_ulpq_tail_data.sctp_cmd_interpreter
> 0.00 +1.7 1.73 ± 5% perf-profile.calltrace.cycles-pp.__sk_mem_schedule.sctp_ulpevent_make_rcvmsg.sctp_ulpq_tail_data.sctp_cmd_interpreter.sctp_do_sm
> 10.84 ± 7% +1.8 12.59 ± 5% perf-profile.calltrace.cycles-pp.ip_finish_output2.__ip_queue_xmit.sctp_packet_transmit.sctp_outq_flush.sctp_cmd_interpreter
> 10.63 ± 7% +1.8 12.42 ± 5% perf-profile.calltrace.cycles-pp.__local_bh_enable_ip.ip_finish_output2.__ip_queue_xmit.sctp_packet_transmit.sctp_outq_flush
> 10.58 ± 7% +1.8 12.40 ± 5% perf-profile.calltrace.cycles-pp.do_softirq.__local_bh_enable_ip.ip_finish_output2.__ip_queue_xmit.sctp_packet_transmit
> 10.53 ± 7% +1.8 12.36 ± 5% perf-profile.calltrace.cycles-pp.__softirqentry_text_start.do_softirq.__local_bh_enable_ip.ip_finish_output2.__ip_queue_xmit
> 10.36 ± 7% +1.9 12.25 ± 5% perf-profile.calltrace.cycles-pp.net_rx_action.__softirqentry_text_start.do_softirq.__local_bh_enable_ip.ip_finish_output2
> 10.26 ± 7% +1.9 12.18 ± 5% perf-profile.calltrace.cycles-pp.__napi_poll.net_rx_action.__softirqentry_text_start.do_softirq.__local_bh_enable_ip
> 10.23 ± 7% +1.9 12.17 ± 5% perf-profile.calltrace.cycles-pp.process_backlog.__napi_poll.net_rx_action.__softirqentry_text_start.do_softirq
> 10.07 ± 7% +2.0 12.08 ± 5% perf-profile.calltrace.cycles-pp.__netif_receive_skb_one_core.process_backlog.__napi_poll.net_rx_action.__softirqentry_text_start
> 0.00 +2.0 2.02 ± 6% perf-profile.calltrace.cycles-pp.try_charge_memcg.mem_cgroup_charge_skmem.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_sendmsg_to_asoc
> 9.84 ± 7% +2.1 11.95 ± 5% perf-profile.calltrace.cycles-pp.ip_local_deliver_finish.__netif_receive_skb_one_core.process_backlog.__napi_poll.net_rx_action
> 9.82 ± 7% +2.1 11.94 ± 5% perf-profile.calltrace.cycles-pp.ip_protocol_deliver_rcu.ip_local_deliver_finish.__netif_receive_skb_one_core.process_backlog.__napi_poll
> 9.74 ± 7% +2.1 11.88 ± 5% perf-profile.calltrace.cycles-pp.sctp_rcv.ip_protocol_deliver_rcu.ip_local_deliver_finish.__netif_receive_skb_one_core.process_backlog
> 3.38 ± 7% +2.1 5.53 ± 5% perf-profile.calltrace.cycles-pp.sctp_skb_recv_datagram.sctp_recvmsg.inet_recvmsg.____sys_recvmsg.___sys_recvmsg
> 0.00 +2.4 2.38 ± 6% perf-profile.calltrace.cycles-pp.mem_cgroup_charge_skmem.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_sendmsg_to_asoc.sctp_sendmsg
> 0.00 +2.5 2.54 ± 6% perf-profile.calltrace.cycles-pp.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg
> 0.00 +2.5 2.55 ± 6% perf-profile.calltrace.cycles-pp.__sk_mem_schedule.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg.____sys_sendmsg
> 8.08 ± 7% +2.7 10.74 ± 5% perf-profile.calltrace.cycles-pp.sctp_assoc_bh_rcv.sctp_rcv.ip_protocol_deliver_rcu.ip_local_deliver_finish.__netif_receive_skb_one_core
> 7.63 ± 7% +2.8 10.44 ± 5% perf-profile.calltrace.cycles-pp.sctp_do_sm.sctp_assoc_bh_rcv.sctp_rcv.ip_protocol_deliver_rcu.ip_local_deliver_finish
> 15.25 ± 6% +2.9 18.11 ± 6% perf-profile.calltrace.cycles-pp.inet_recvmsg.____sys_recvmsg.___sys_recvmsg.__sys_recvmsg.do_syscall_64
> 15.15 ± 7% +2.9 18.03 ± 5% perf-profile.calltrace.cycles-pp.release_sock.sctp_sendmsg.sock_sendmsg.____sys_sendmsg.___sys_sendmsg
> 15.20 ± 6% +2.9 18.09 ± 6% perf-profile.calltrace.cycles-pp.sctp_recvmsg.inet_recvmsg.____sys_recvmsg.___sys_recvmsg.__sys_recvmsg
> 15.02 ± 7% +2.9 17.94 ± 6% perf-profile.calltrace.cycles-pp.__release_sock.release_sock.sctp_sendmsg.sock_sendmsg.____sys_sendmsg
> 14.96 ± 7% +2.9 17.90 ± 6% perf-profile.calltrace.cycles-pp.sctp_backlog_rcv.__release_sock.release_sock.sctp_sendmsg.sock_sendmsg
> 11.70 ± 7% +3.0 14.67 ± 6% perf-profile.calltrace.cycles-pp.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv.sctp_backlog_rcv.__release_sock
> 0.84 ± 10% +3.1 3.94 ± 5% perf-profile.calltrace.cycles-pp._raw_spin_lock_bh.lock_sock_nested.sctp_skb_recv_datagram.sctp_recvmsg.inet_recvmsg
> 0.96 ± 9% +3.1 4.06 ± 6% perf-profile.calltrace.cycles-pp.lock_sock_nested.sctp_skb_recv_datagram.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
> 0.64 ± 11% +3.2 3.82 ± 5% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_bh.lock_sock_nested.sctp_skb_recv_datagram.sctp_recvmsg
> 0.00 +3.9 3.90 ± 5% perf-profile.calltrace.cycles-pp.sctp_eat_data.sctp_sf_eat_data_6_2.sctp_do_sm.sctp_assoc_bh_rcv.sctp_rcv
> 0.00 +4.0 3.98 ± 5% perf-profile.calltrace.cycles-pp.sctp_sf_eat_data_6_2.sctp_do_sm.sctp_assoc_bh_rcv.sctp_rcv.ip_protocol_deliver_rcu
> 2.90 ± 7% +4.2 7.09 ± 6% perf-profile.calltrace.cycles-pp.sctp_outq_sack.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv.sctp_backlog_rcv
> 1.87 ± 7% +4.5 6.40 ± 6% perf-profile.calltrace.cycles-pp.sctp_chunk_put.sctp_outq_sack.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv
> 1.43 ± 7% +4.6 6.01 ± 6% perf-profile.calltrace.cycles-pp.consume_skb.sctp_chunk_put.sctp_outq_sack.sctp_cmd_interpreter.sctp_do_sm
> 1.78 ± 6% +4.6 6.42 ± 6% perf-profile.calltrace.cycles-pp.kfree_skb_reason.sctp_recvmsg.inet_recvmsg.____sys_recvmsg.___sys_recvmsg
> 0.00 +5.1 5.10 ± 5% perf-profile.calltrace.cycles-pp.__sk_mem_reduce_allocated.sctp_wfree.skb_release_head_state.consume_skb.sctp_chunk_put
> 0.17 ±141% +5.3 5.42 ± 6% perf-profile.calltrace.cycles-pp.skb_release_head_state.consume_skb.sctp_chunk_put.sctp_outq_sack.sctp_cmd_interpreter
> 0.00 +5.3 5.35 ± 6% perf-profile.calltrace.cycles-pp.sctp_wfree.skb_release_head_state.consume_skb.sctp_chunk_put.sctp_outq_sack
> 0.00 +5.5 5.51 ± 6% perf-profile.calltrace.cycles-pp.__sk_mem_reduce_allocated.skb_release_head_state.kfree_skb_reason.sctp_recvmsg.inet_recvmsg
> 0.00 +5.7 5.65 ± 6% perf-profile.calltrace.cycles-pp.skb_release_head_state.kfree_skb_reason.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
> 9.38 ± 6% -3.5 5.93 ± 6% perf-profile.children.cycles-pp.sctp_packet_pack
> 24.70 ± 6% -3.4 21.34 ± 6% perf-profile.children.cycles-pp.sctp_outq_flush
> 8.83 ± 6% -3.2 5.60 ± 6% perf-profile.children.cycles-pp.copy_user_enhanced_fast_string
> 8.77 ± 6% -3.2 5.58 ± 6% perf-profile.children.cycles-pp.memcpy_erms
> 5.88 ± 4% -3.0 2.92 ± 8% perf-profile.children.cycles-pp.intel_idle_irq
> 21.34 ± 6% -2.8 18.58 ± 6% perf-profile.children.cycles-pp.sctp_sendmsg_to_asoc
> 7.06 ± 6% -2.5 4.54 ± 7% perf-profile.children.cycles-pp.sctp_datamsg_from_user
> 13.52 ± 6% -2.5 11.00 ± 6% perf-profile.children.cycles-pp.sctp_primitive_SEND
> 6.53 ± 6% -2.5 4.06 ± 6% perf-profile.children.cycles-pp.skb_copy_datagram_iter
> 6.51 ± 6% -2.5 4.05 ± 6% perf-profile.children.cycles-pp.__skb_datagram_iter
> 5.68 ± 6% -2.1 3.54 ± 6% perf-profile.children.cycles-pp._copy_to_iter
> 5.52 ± 6% -2.1 3.45 ± 6% perf-profile.children.cycles-pp.copyout
> 3.33 ± 6% -1.4 1.92 ± 6% perf-profile.children.cycles-pp.__alloc_skb
> 3.38 ± 6% -1.3 2.07 ± 6% perf-profile.children.cycles-pp.__schedule
> 3.51 ± 6% -1.2 2.31 ± 7% perf-profile.children.cycles-pp.sctp_user_addto_chunk
> 2.67 ± 6% -1.2 1.47 ± 6% perf-profile.children.cycles-pp.skb_release_data
> 2.43 ± 6% -1.1 1.32 ± 6% perf-profile.children.cycles-pp.kmalloc_reserve
> 3.17 ± 6% -1.1 2.08 ± 7% perf-profile.children.cycles-pp._copy_from_iter
> 2.38 ± 6% -1.1 1.29 ± 6% perf-profile.children.cycles-pp.__kmalloc_node_track_caller
> 2.88 ± 7% -1.1 1.80 ± 7% perf-profile.children.cycles-pp.sctp_make_datafrag_empty
> 3.18 ± 6% -1.1 2.12 ± 6% perf-profile.children.cycles-pp.sctp_data_ready
> 3.02 ± 6% -1.0 1.98 ± 7% perf-profile.children.cycles-pp.copyin
> 3.28 ± 6% -1.0 2.25 ± 6% perf-profile.children.cycles-pp.sctp_ulpq_tail_event
> 2.26 ± 6% -1.0 1.22 ± 5% perf-profile.children.cycles-pp.kmalloc_large_node
> 2.62 ± 6% -1.0 1.60 ± 7% perf-profile.children.cycles-pp._sctp_make_chunk
> 2.90 ± 6% -1.0 1.89 ± 6% perf-profile.children.cycles-pp.__wake_up_common_lock
> 2.69 ± 6% -0.9 1.75 ± 6% perf-profile.children.cycles-pp.__wake_up_common
> 1.97 ± 6% -0.9 1.04 ± 6% perf-profile.children.cycles-pp.__alloc_pages
> 2.50 ± 6% -0.9 1.64 ± 6% perf-profile.children.cycles-pp.autoremove_wake_function
> 2.48 ± 6% -0.9 1.63 ± 6% perf-profile.children.cycles-pp.try_to_wake_up
> 1.67 ± 5% -0.8 0.86 ± 6% perf-profile.children.cycles-pp.get_page_from_freelist
> 2.26 ± 7% -0.8 1.48 ± 7% perf-profile.children.cycles-pp.sctp_outq_flush_data
> 1.86 ± 6% -0.8 1.10 ± 6% perf-profile.children.cycles-pp.schedule_timeout
> 1.76 ± 8% -0.7 1.02 ± 6% perf-profile.children.cycles-pp.sctp_ulpevent_free
> 2.08 ± 5% -0.7 1.33 ± 8% perf-profile.children.cycles-pp.kmem_cache_free
> 1.84 ± 6% -0.7 1.10 ± 6% perf-profile.children.cycles-pp.schedule
> 1.62 ± 6% -0.6 1.02 ± 6% perf-profile.children.cycles-pp.schedule_idle
> 1.20 ± 6% -0.6 0.62 ± 7% perf-profile.children.cycles-pp.free_unref_page
> 1.49 ± 5% -0.5 0.95 ± 6% perf-profile.children.cycles-pp._copy_from_user
> 1.03 ± 6% -0.5 0.54 ± 5% perf-profile.children.cycles-pp.rmqueue
> 1.06 ± 6% -0.5 0.58 ± 5% perf-profile.children.cycles-pp._raw_spin_lock_irqsave
> 1.18 ± 5% -0.4 0.74 ± 6% perf-profile.children.cycles-pp.__check_object_size
> 1.22 ± 4% -0.4 0.79 ± 8% perf-profile.children.cycles-pp.__rhashtable_lookup
> 0.97 ± 6% -0.4 0.57 ± 7% perf-profile.children.cycles-pp.__slab_free
> 0.94 ± 7% -0.4 0.54 ± 7% perf-profile.children.cycles-pp.dequeue_task_fair
> 1.18 ± 6% -0.4 0.78 ± 7% perf-profile.children.cycles-pp.ttwu_do_activate
> 0.82 ± 6% -0.4 0.43 ± 7% perf-profile.children.cycles-pp._raw_spin_lock
> 1.11 ± 6% -0.4 0.73 ± 7% perf-profile.children.cycles-pp.sendmsg_copy_msghdr
> 1.14 ± 6% -0.4 0.76 ± 7% perf-profile.children.cycles-pp.enqueue_task_fair
> 0.87 ± 7% -0.4 0.50 ± 7% perf-profile.children.cycles-pp.dequeue_entity
> 0.92 ± 11% -0.4 0.56 ± 6% perf-profile.children.cycles-pp.sctp_outq_select_transport
> 1.00 ± 5% -0.4 0.65 ± 6% perf-profile.children.cycles-pp.__copy_msghdr_from_user
> 0.82 ± 12% -0.3 0.49 ± 6% perf-profile.children.cycles-pp.sctp_packet_config
> 0.84 ± 6% -0.3 0.54 ± 8% perf-profile.children.cycles-pp.import_iovec
> 0.84 ± 6% -0.3 0.55 ± 7% perf-profile.children.cycles-pp.pick_next_task_fair
> 0.80 ± 4% -0.3 0.51 ± 9% perf-profile.children.cycles-pp.kfree
> 0.50 ± 6% -0.3 0.20 ± 7% perf-profile.children.cycles-pp.free_pcppages_bulk
> 0.80 ± 6% -0.3 0.51 ± 8% perf-profile.children.cycles-pp.__import_iovec
> 0.83 ± 6% -0.3 0.54 ± 7% perf-profile.children.cycles-pp.enqueue_entity
> 0.79 ± 7% -0.3 0.51 ± 7% perf-profile.children.cycles-pp.kmem_cache_alloc
> 0.70 ± 6% -0.3 0.42 ± 8% perf-profile.children.cycles-pp.update_load_avg
> 0.78 ± 7% -0.3 0.50 ± 10% perf-profile.children.cycles-pp.sctp_hash_cmp
> 0.73 ± 5% -0.3 0.47 ± 6% perf-profile.children.cycles-pp.__sctp_rcv_lookup
> 0.69 ± 6% -0.3 0.44 ± 6% perf-profile.children.cycles-pp.check_heap_object
> 0.47 ± 6% -0.3 0.22 ± 8% perf-profile.children.cycles-pp.__free_pages_ok
> 0.70 ± 4% -0.2 0.46 ± 6% perf-profile.children.cycles-pp.sctp_addrs_lookup_transport
> 0.68 ± 6% -0.2 0.43 ± 8% perf-profile.children.cycles-pp.iovec_from_user
> 0.71 ± 6% -0.2 0.47 ± 8% perf-profile.children.cycles-pp.sctp_chunkify
> 0.66 ± 5% -0.2 0.42 ± 8% perf-profile.children.cycles-pp.__might_resched
> 0.60 ± 5% -0.2 0.36 ± 4% perf-profile.children.cycles-pp.set_next_entity
> 0.74 ± 6% -0.2 0.51 ± 8% perf-profile.children.cycles-pp.sctp_endpoint_lookup_assoc
> 0.72 ± 6% -0.2 0.50 ± 9% perf-profile.children.cycles-pp.sctp_epaddr_lookup_transport
> 0.54 ± 6% -0.2 0.32 ± 10% perf-profile.children.cycles-pp.__dev_queue_xmit
> 0.44 ± 7% -0.2 0.23 ± 6% perf-profile.children.cycles-pp.sctp_inq_pop
> 0.61 ± 6% -0.2 0.40 ± 7% perf-profile.children.cycles-pp.copy_user_short_string
> 0.62 ± 13% -0.2 0.42 ± 5% perf-profile.children.cycles-pp.__skb_clone
> 0.52 ± 5% -0.2 0.32 ± 7% perf-profile.children.cycles-pp.__might_fault
> 0.46 ± 5% -0.2 0.27 ± 7% perf-profile.children.cycles-pp.simple_copy_to_iter
> 0.51 ± 7% -0.2 0.33 ± 4% perf-profile.children.cycles-pp.__entry_text_start
> 0.36 ± 7% -0.2 0.19 ± 8% perf-profile.children.cycles-pp.update_curr
> 0.56 ± 6% -0.2 0.38 ± 6% perf-profile.children.cycles-pp.__list_del_entry_valid
> 0.37 ± 7% -0.2 0.21 ± 5% perf-profile.children.cycles-pp.__mod_node_page_state
> 0.32 ± 7% -0.2 0.17 ± 8% perf-profile.children.cycles-pp.__zone_watermark_ok
> 0.42 ± 10% -0.2 0.27 ± 6% perf-profile.children.cycles-pp.select_task_rq
> 0.46 ± 17% -0.2 0.31 ± 5% perf-profile.children.cycles-pp.__copy_skb_header
> 0.46 ± 7% -0.1 0.31 ± 10% perf-profile.children.cycles-pp.sctp_packet_transmit_chunk
> 0.46 ± 7% -0.1 0.32 ± 4% perf-profile.children.cycles-pp.update_rq_clock
> 0.43 ± 5% -0.1 0.29 ± 10% perf-profile.children.cycles-pp.sctp_addto_chunk
> 0.41 ± 7% -0.1 0.27 ± 10% perf-profile.children.cycles-pp.sctp_packet_append_chunk
> 0.34 ± 6% -0.1 0.20 ± 9% perf-profile.children.cycles-pp.dev_hard_start_xmit
> 0.46 ± 8% -0.1 0.32 ± 6% perf-profile.children.cycles-pp.kmem_cache_alloc_node
> 0.40 ± 5% -0.1 0.27 ± 5% perf-profile.children.cycles-pp.sockfd_lookup_light
> 0.38 ± 6% -0.1 0.25 ± 6% perf-profile.children.cycles-pp.sctp_chunk_free
> 0.24 ± 14% -0.1 0.10 ± 11% perf-profile.children.cycles-pp.sctp_assoc_rwnd_increase
> 0.35 ± 6% -0.1 0.21 ± 6% perf-profile.children.cycles-pp.__list_add_valid
> 0.36 ± 7% -0.1 0.22 ± 10% perf-profile.children.cycles-pp.sctp_gen_sack
> 0.36 ± 5% -0.1 0.23 ± 9% perf-profile.children.cycles-pp.memset_erms
> 0.35 ± 8% -0.1 0.22 ± 10% perf-profile.children.cycles-pp.memcg_slab_free_hook
> 0.32 ± 6% -0.1 0.19 ± 9% perf-profile.children.cycles-pp.loopback_xmit
> 0.36 ± 5% -0.1 0.23 ± 8% perf-profile.children.cycles-pp.sock_kmalloc
> 0.36 ± 4% -0.1 0.24 ± 7% perf-profile.children.cycles-pp.__might_sleep
> 0.33 ± 8% -0.1 0.21 ± 7% perf-profile.children.cycles-pp.__virt_addr_valid
> 0.28 ± 8% -0.1 0.15 ± 8% perf-profile.children.cycles-pp.free_unref_page_commit
> 0.34 ± 4% -0.1 0.22 ± 6% perf-profile.children.cycles-pp.free_pcp_prepare
> 0.33 ± 7% -0.1 0.20 ± 11% perf-profile.children.cycles-pp.sctp_make_sack
> 0.30 ± 13% -0.1 0.18 ± 9% perf-profile.children.cycles-pp.sctp_v4_xmit
> 0.34 ± 5% -0.1 0.22 ± 6% perf-profile.children.cycles-pp.__fdget
> 0.35 ± 6% -0.1 0.23 ± 7% perf-profile.children.cycles-pp.move_addr_to_kernel
> 0.25 ± 5% -0.1 0.13 ± 6% perf-profile.children.cycles-pp.__free_one_page
> 0.40 ± 9% -0.1 0.29 ± 6% perf-profile.children.cycles-pp.aa_sk_perm
> 0.24 ± 6% -0.1 0.12 ± 8% perf-profile.children.cycles-pp.prepare_task_switch
> 0.32 ± 7% -0.1 0.21 ± 6% perf-profile.children.cycles-pp.sctp_check_transmitted
> 0.30 ± 6% -0.1 0.19 ± 7% perf-profile.children.cycles-pp.sctp_association_put
> 0.36 ± 7% -0.1 0.24 ± 7% perf-profile.children.cycles-pp.sock_recvmsg
> 0.26 ± 12% -0.1 0.16 ± 15% perf-profile.children.cycles-pp.ipv4_dst_check
> 0.22 ± 9% -0.1 0.11 ± 46% perf-profile.children.cycles-pp.accept_connections
> 0.31 ± 11% -0.1 0.20 ± 9% perf-profile.children.cycles-pp.select_task_rq_fair
> 0.22 ± 17% -0.1 0.11 ± 6% perf-profile.children.cycles-pp.dst_release
> 0.23 ± 6% -0.1 0.12 ± 7% perf-profile.children.cycles-pp.rmqueue_bulk
> 0.22 ± 6% -0.1 0.11 ± 12% perf-profile.children.cycles-pp.sock_wfree
> 0.28 ± 6% -0.1 0.18 ± 5% perf-profile.children.cycles-pp.send_sctp_stream_1toMany
> 0.32 ± 7% -0.1 0.23 ± 9% perf-profile.children.cycles-pp.security_socket_recvmsg
> 0.30 ± 7% -0.1 0.20 ± 9% perf-profile.children.cycles-pp.skb_set_owner_w
> 0.28 ± 7% -0.1 0.18 ± 8% perf-profile.children.cycles-pp.entry_SYSRETQ_unsafe_stack
> 0.27 ± 4% -0.1 0.17 ± 9% perf-profile.children.cycles-pp.sock_kfree_s
> 0.27 ± 10% -0.1 0.18 ± 11% perf-profile.children.cycles-pp.sctp_transport_hold
> 0.24 ± 6% -0.1 0.15 ± 5% perf-profile.children.cycles-pp.__switch_to
> 0.24 ± 7% -0.1 0.15 ± 6% perf-profile.children.cycles-pp.tick_nohz_idle_exit
> 0.21 ± 10% -0.1 0.12 ± 9% perf-profile.children.cycles-pp.recv_sctp_stream_1toMany
> 0.23 ± 7% -0.1 0.14 ± 7% perf-profile.children.cycles-pp.prepare_to_wait_exclusive
> 0.22 ± 9% -0.1 0.14 ± 9% perf-profile.children.cycles-pp.accept_connection
> 0.22 ± 9% -0.1 0.14 ± 9% perf-profile.children.cycles-pp.spawn_child
> 0.22 ± 9% -0.1 0.14 ± 9% perf-profile.children.cycles-pp.process_requests
> 0.23 ± 3% -0.1 0.14 ± 7% perf-profile.children.cycles-pp.__kmalloc
> 0.20 ± 9% -0.1 0.12 ± 10% perf-profile.children.cycles-pp.skb_clone
> 0.31 ± 10% -0.1 0.22 ± 13% perf-profile.children.cycles-pp.sctp_sched_fcfs_dequeue
> 0.29 ± 12% -0.1 0.21 ± 5% perf-profile.children.cycles-pp.sctp_transport_put
> 0.22 ± 8% -0.1 0.14 ± 7% perf-profile.children.cycles-pp.select_idle_sibling
> 0.22 ± 8% -0.1 0.14 ± 5% perf-profile.children.cycles-pp.__switch_to_asm
> 0.25 ± 10% -0.1 0.17 ± 11% perf-profile.children.cycles-pp.__sctp_packet_append_chunk
> 0.23 ± 5% -0.1 0.15 ± 10% perf-profile.children.cycles-pp.__cond_resched
> 0.20 ± 6% -0.1 0.13 ± 5% perf-profile.children.cycles-pp.available_idle_cpu
> 0.19 ± 9% -0.1 0.11 ± 6% perf-profile.children.cycles-pp.poll_idle
> 0.20 ± 6% -0.1 0.13 ± 9% perf-profile.children.cycles-pp.reweight_entity
> 0.20 ± 7% -0.1 0.13 ± 8% perf-profile.children.cycles-pp.sctp_make_control
> 0.25 ± 9% -0.1 0.18 ± 8% perf-profile.children.cycles-pp.syscall_exit_to_user_mode
> 0.20 ± 6% -0.1 0.12 ± 8% perf-profile.children.cycles-pp.__free_pages
> 0.26 ± 8% -0.1 0.19 ± 9% perf-profile.children.cycles-pp.security_socket_sendmsg
> 0.19 ± 8% -0.1 0.12 ± 13% perf-profile.children.cycles-pp.syscall_return_via_sysret
> 0.17 ± 9% -0.1 0.10 ± 14% perf-profile.children.cycles-pp.__mod_timer
> 0.16 ± 5% -0.1 0.09 ± 7% perf-profile.children.cycles-pp.___perf_sw_event
> 0.21 ± 6% -0.1 0.14 ± 9% perf-profile.children.cycles-pp.sctp_datamsg_put
> 0.21 ± 7% -0.1 0.14 ± 7% perf-profile.children.cycles-pp.sctp_association_hold
> 0.18 ± 10% -0.1 0.11 ± 4% perf-profile.children.cycles-pp.sctp_outq_flush_ctrl
> 0.16 ± 8% -0.1 0.09 ± 14% perf-profile.children.cycles-pp.__update_load_avg_cfs_rq
> 0.20 ± 4% -0.1 0.13 ± 9% perf-profile.children.cycles-pp.kmem_cache_alloc_trace
> 0.19 ± 5% -0.1 0.13 ± 8% perf-profile.children.cycles-pp.__update_load_avg_se
> 0.16 ± 5% -0.1 0.10 ± 9% perf-profile.children.cycles-pp.del_timer
> 0.14 ± 11% -0.1 0.08 ± 4% perf-profile.children.cycles-pp.check_new_pages
> 0.23 ± 7% -0.1 0.17 ± 10% perf-profile.children.cycles-pp.sctp_sched_dequeue_common
> 0.10 ± 5% -0.1 0.04 ± 71% perf-profile.children.cycles-pp.finish_task_switch
> 0.18 ± 7% -0.1 0.12 ± 6% perf-profile.children.cycles-pp.sctp_outq_tail
> 0.12 ± 4% -0.1 0.06 ± 13% perf-profile.children.cycles-pp.check_stack_object
> 0.13 ± 8% -0.1 0.08 ± 8% perf-profile.children.cycles-pp.__put_user_nocheck_4
> 0.16 ± 4% -0.1 0.10 ± 13% perf-profile.children.cycles-pp.__genradix_ptr
> 0.17 ± 6% -0.1 0.11 ± 6% perf-profile.children.cycles-pp.nr_iowait_cpu
> 0.16 ± 8% -0.1 0.11 ± 13% perf-profile.children.cycles-pp.sctp_sendmsg_parse
> 0.13 ± 7% -0.1 0.08 ± 12% perf-profile.children.cycles-pp.lock_timer_base
> 0.08 ± 12% -0.0 0.02 ± 99% perf-profile.children.cycles-pp.ip_local_out
> 0.14 ± 7% -0.0 0.09 ± 11% perf-profile.children.cycles-pp.memcg_slab_post_alloc_hook
> 0.14 ± 9% -0.0 0.09 ± 7% perf-profile.children.cycles-pp.switch_mm_irqs_off
> 0.14 ± 9% -0.0 0.09 ± 10% perf-profile.children.cycles-pp.sctp_datamsg_destroy
> 0.12 ± 20% -0.0 0.07 ± 21% perf-profile.children.cycles-pp.sctp_hash_key
> 0.11 ± 20% -0.0 0.06 ± 14% perf-profile.children.cycles-pp.ip_rcv
> 0.11 ± 8% -0.0 0.07 ± 7% perf-profile.children.cycles-pp.sctp_sock_rfree
> 0.24 ± 7% -0.0 0.19 ± 6% perf-profile.children.cycles-pp.sched_clock_cpu
> 0.12 ± 6% -0.0 0.08 ± 11% perf-profile.children.cycles-pp.__netif_rx
> 0.16 ± 8% -0.0 0.11 ± 10% perf-profile.children.cycles-pp.ttwu_do_wakeup
> 0.31 ± 6% -0.0 0.27 ± 5% perf-profile.children.cycles-pp.sctp_ulpevent_receive_data
> 0.13 ± 7% -0.0 0.09 ± 5% perf-profile.children.cycles-pp.pick_next_entity
> 0.12 ± 6% -0.0 0.08 ± 7% perf-profile.children.cycles-pp.__check_heap_object
> 0.12 ± 8% -0.0 0.08 ± 12% perf-profile.children.cycles-pp.netif_rx_internal
> 0.08 ± 10% -0.0 0.03 ± 70% perf-profile.children.cycles-pp.__wrgsbase_inactive
> 0.22 ± 8% -0.0 0.18 ± 6% perf-profile.children.cycles-pp.read_tsc
> 0.14 ± 8% -0.0 0.10 ± 9% perf-profile.children.cycles-pp.check_preempt_curr
> 0.10 ± 9% -0.0 0.06 ± 7% perf-profile.children.cycles-pp.rcu_all_qs
> 0.10 ± 10% -0.0 0.06 ± 13% perf-profile.children.cycles-pp.validate_xmit_skb
> 0.14 ± 8% -0.0 0.10 ± 9% perf-profile.children.cycles-pp.perf_trace_sched_wakeup_template
> 0.11 ± 6% -0.0 0.07 ± 9% perf-profile.children.cycles-pp.enqueue_to_backlog
> 0.09 ± 5% -0.0 0.05 ± 45% perf-profile.children.cycles-pp.__netif_receive_skb_core
> 0.12 ± 5% -0.0 0.08 ± 8% perf-profile.children.cycles-pp.__put_user_nocheck_8
> 0.12 ± 6% -0.0 0.08 ± 14% perf-profile.children.cycles-pp.sctp_chunk_abandoned
> 0.12 ± 5% -0.0 0.08 ± 8% perf-profile.children.cycles-pp.skb_put
> 0.20 ± 7% -0.0 0.17 ± 5% perf-profile.children.cycles-pp.native_sched_clock
> 0.09 ± 5% -0.0 0.05 ± 46% perf-profile.children.cycles-pp.__mod_lruvec_page_state
> 0.11 ± 9% -0.0 0.08 ± 6% perf-profile.children.cycles-pp.sctp_inet_skb_msgname
> 0.10 ± 8% -0.0 0.06 perf-profile.children.cycles-pp.tick_nohz_idle_enter
> 0.09 ± 7% -0.0 0.06 ± 6% perf-profile.children.cycles-pp.os_xsave
> 0.12 ± 9% -0.0 0.09 ± 8% perf-profile.children.cycles-pp.resched_curr
> 0.09 ± 10% -0.0 0.06 ± 9% perf-profile.children.cycles-pp.sctp_ulpevent_init
> 0.14 ± 10% -0.0 0.11 ± 5% perf-profile.children.cycles-pp.update_cfs_group
> 0.09 ± 10% -0.0 0.06 ± 8% perf-profile.children.cycles-pp.sctp_ulpq_order
> 0.08 ± 11% -0.0 0.05 ± 44% perf-profile.children.cycles-pp.perf_tp_event
> 0.07 ± 5% -0.0 0.04 ± 45% perf-profile.children.cycles-pp._raw_spin_unlock_irqrestore
> 0.10 ± 8% -0.0 0.07 ± 9% perf-profile.children.cycles-pp.syscall_exit_to_user_mode_prepare
> 0.07 ± 10% -0.0 0.04 ± 45% perf-profile.children.cycles-pp.__cgroup_account_cputime
> 0.08 ± 6% -0.0 0.06 ± 8% perf-profile.children.cycles-pp.sctp_tsnmap_check
> 0.00 +0.1 0.06 ± 9% perf-profile.children.cycles-pp.perf_mux_hrtimer_handler
> 0.00 +0.1 0.07 ± 26% perf-profile.children.cycles-pp.update_sd_lb_stats
> 0.00 +0.1 0.08 ± 22% perf-profile.children.cycles-pp.find_busiest_group
> 0.01 ±223% +0.1 0.09 ± 26% perf-profile.children.cycles-pp.load_balance
> 0.05 ± 46% +0.1 0.13 ± 4% perf-profile.children.cycles-pp.sctp_get_af_specific
> 0.00 +0.1 0.09 ± 14% perf-profile.children.cycles-pp.lapic_next_deadline
> 0.00 +0.1 0.09 ± 13% perf-profile.children.cycles-pp.native_irq_return_iret
> 0.00 +0.1 0.09 ± 39% perf-profile.children.cycles-pp.tsc_verify_tsc_adjust
> 0.00 +0.1 0.09 ± 31% perf-profile.children.cycles-pp.calc_global_load_tick
> 0.00 +0.1 0.10 ± 37% perf-profile.children.cycles-pp.arch_cpu_idle_enter
> 0.00 +0.1 0.10 ± 31% perf-profile.children.cycles-pp._raw_spin_trylock
> 0.00 +0.1 0.10 ± 25% perf-profile.children.cycles-pp.arch_scale_freq_tick
> 0.03 ±100% +0.1 0.15 ± 35% perf-profile.children.cycles-pp.rebalance_domains
> 0.00 +0.1 0.14 ± 21% perf-profile.children.cycles-pp.cgroup_rstat_updated
> 0.17 ± 15% +0.1 0.32 ± 19% perf-profile.children.cycles-pp.scheduler_tick
> 0.00 +0.2 0.16 ± 32% perf-profile.children.cycles-pp.tick_nohz_irq_exit
> 0.24 ± 11% +0.2 0.42 ± 20% perf-profile.children.cycles-pp.update_process_times
> 0.24 ± 11% +0.2 0.43 ± 20% perf-profile.children.cycles-pp.tick_sched_handle
> 0.09 ± 11% +0.2 0.29 ± 28% perf-profile.children.cycles-pp.__irq_exit_rcu
> 0.53 ± 7% +0.2 0.74 ± 16% perf-profile.children.cycles-pp.ktime_get
> 0.20 ± 10% +0.2 0.42 ± 21% perf-profile.children.cycles-pp.clockevents_program_event
> 0.06 ± 13% +0.3 0.31 ± 53% perf-profile.children.cycles-pp.timekeeping_max_deferment
> 0.00 +0.3 0.26 ± 12% perf-profile.children.cycles-pp.propagate_protected_usage
> 0.28 ± 8% +0.3 0.54 ± 19% perf-profile.children.cycles-pp.tick_sched_timer
> 0.55 ± 7% +0.3 0.84 ± 26% perf-profile.children.cycles-pp.menu_select
> 0.29 ± 7% +0.3 0.58 ± 39% perf-profile.children.cycles-pp.tick_nohz_get_sleep_length
> 0.20 ± 8% +0.3 0.50 ± 44% perf-profile.children.cycles-pp.tick_nohz_next_event
> 0.32 ± 6% +0.3 0.66 ± 17% perf-profile.children.cycles-pp.__hrtimer_run_queues
> 0.00 +0.5 0.52 ± 6% perf-profile.children.cycles-pp.mem_cgroup_uncharge_skmem
> 0.56 ± 7% +0.6 1.16 ± 17% perf-profile.children.cycles-pp.hrtimer_interrupt
> 0.56 ± 7% +0.6 1.17 ± 17% perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
> 0.00 +1.0 0.96 ± 4% perf-profile.children.cycles-pp.__mod_memcg_state
> 0.72 ± 6% +1.0 1.69 ± 20% perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
> 0.91 ± 6% +1.1 2.05 ± 16% perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
> 1.62 ± 9% +1.3 2.96 ± 5% perf-profile.children.cycles-pp.sctp_ulpevent_make_rcvmsg
> 10.76 ± 7% +1.8 12.59 ± 6% perf-profile.children.cycles-pp.__local_bh_enable_ip
> 10.64 ± 7% +1.9 12.51 ± 6% perf-profile.children.cycles-pp.do_softirq
> 10.41 ± 7% +2.0 12.37 ± 6% perf-profile.children.cycles-pp.net_rx_action
> 10.31 ± 7% +2.0 12.30 ± 6% perf-profile.children.cycles-pp.__napi_poll
> 10.29 ± 7% +2.0 12.30 ± 6% perf-profile.children.cycles-pp.process_backlog
> 10.66 ± 7% +2.1 12.73 ± 5% perf-profile.children.cycles-pp.__softirqentry_text_start
> 10.12 ± 7% +2.1 12.20 ± 6% perf-profile.children.cycles-pp.__netif_receive_skb_one_core
> 0.00 +2.1 2.10 ± 6% perf-profile.children.cycles-pp.page_counter_uncharge
> 3.40 ± 7% +2.1 5.54 ± 5% perf-profile.children.cycles-pp.sctp_skb_recv_datagram
> 0.00 +2.1 2.14 ± 5% perf-profile.children.cycles-pp.drain_stock
> 9.89 ± 7% +2.2 12.07 ± 6% perf-profile.children.cycles-pp.ip_local_deliver_finish
> 9.87 ± 7% +2.2 12.06 ± 6% perf-profile.children.cycles-pp.ip_protocol_deliver_rcu
> 0.00 +2.2 2.20 ± 5% perf-profile.children.cycles-pp.page_counter_try_charge
> 9.80 ± 7% +2.2 12.00 ± 6% perf-profile.children.cycles-pp.sctp_rcv
> 0.00 +2.4 2.36 ± 6% perf-profile.children.cycles-pp.refill_stock
> 15.40 ± 6% +2.8 18.21 ± 6% perf-profile.children.cycles-pp.sctp_recvmsg
> 1.10 ± 8% +2.8 3.94 ± 6% perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
> 15.25 ± 6% +2.9 18.12 ± 6% perf-profile.children.cycles-pp.inet_recvmsg
> 15.36 ± 6% +2.9 18.28 ± 6% perf-profile.children.cycles-pp.release_sock
> 1.36 ± 7% +3.0 4.35 ± 6% perf-profile.children.cycles-pp.lock_sock_nested
> 1.25 ± 8% +3.0 4.24 ± 6% perf-profile.children.cycles-pp._raw_spin_lock_bh
> 15.10 ± 6% +3.0 18.11 ± 6% perf-profile.children.cycles-pp.__release_sock
> 15.04 ± 6% +3.0 18.07 ± 6% perf-profile.children.cycles-pp.sctp_backlog_rcv
> 0.62 ± 6% +3.4 4.02 ± 6% perf-profile.children.cycles-pp.sctp_sf_eat_data_6_2
> 0.00 +3.4 3.41 ± 6% perf-profile.children.cycles-pp.try_charge_memcg
> 0.49 ± 6% +3.4 3.94 ± 6% perf-profile.children.cycles-pp.sctp_eat_data
> 19.74 ± 6% +3.7 23.44 ± 6% perf-profile.children.cycles-pp.sctp_assoc_bh_rcv
> 3.40 ± 7% +3.9 7.34 ± 6% perf-profile.children.cycles-pp.sctp_chunk_put
> 0.00 +4.0 4.04 ± 6% perf-profile.children.cycles-pp.mem_cgroup_charge_skmem
> 2.92 ± 6% +4.2 7.16 ± 6% perf-profile.children.cycles-pp.sctp_outq_sack
> 0.00 +4.3 4.29 ± 6% perf-profile.children.cycles-pp.__sk_mem_raise_allocated
> 0.00 +4.3 4.32 ± 6% perf-profile.children.cycles-pp.__sk_mem_schedule
> 1.99 ± 6% +4.4 6.40 ± 6% perf-profile.children.cycles-pp.consume_skb
> 1.78 ± 6% +4.6 6.42 ± 6% perf-profile.children.cycles-pp.kfree_skb_reason
> 0.37 ± 8% +5.0 5.40 ± 6% perf-profile.children.cycles-pp.sctp_wfree
> 0.87 ± 9% +10.3 11.20 ± 6% perf-profile.children.cycles-pp.skb_release_head_state
> 0.00 +10.7 10.66 ± 6% perf-profile.children.cycles-pp.__sk_mem_reduce_allocated
> 8.78 ± 6% -3.2 5.58 ± 6% perf-profile.self.cycles-pp.copy_user_enhanced_fast_string
> 8.70 ± 6% -3.2 5.54 ± 6% perf-profile.self.cycles-pp.memcpy_erms
> 0.96 ± 6% -0.4 0.57 ± 7% perf-profile.self.cycles-pp.__slab_free
> 0.83 ± 7% -0.3 0.54 ± 6% perf-profile.self.cycles-pp._raw_spin_lock_irqsave
> 0.70 ± 9% -0.3 0.41 ± 5% perf-profile.self.cycles-pp.sctp_chunk_put
> 0.83 ± 4% -0.2 0.59 ± 8% perf-profile.self.cycles-pp.kmem_cache_free
> 0.60 ± 6% -0.2 0.36 ± 8% perf-profile.self.cycles-pp.sctp_recvmsg
> 0.64 ± 5% -0.2 0.41 ± 7% perf-profile.self.cycles-pp.__might_resched
> 0.57 ± 6% -0.2 0.36 ± 8% perf-profile.self.cycles-pp._raw_spin_lock_bh
> 0.38 ± 7% -0.2 0.16 ± 9% perf-profile.self.cycles-pp.sctp_sendmsg_to_asoc
> 0.62 ± 6% -0.2 0.40 ± 6% perf-profile.self.cycles-pp._raw_spin_lock
> 0.59 ± 7% -0.2 0.39 ± 5% perf-profile.self.cycles-pp.__schedule
> 0.58 ± 6% -0.2 0.39 ± 7% perf-profile.self.cycles-pp.copy_user_short_string
> 0.50 ± 12% -0.2 0.31 ± 8% perf-profile.self.cycles-pp.sctp_packet_config
> 0.40 ± 8% -0.2 0.23 ± 9% perf-profile.self.cycles-pp.sctp_packet_pack
> 0.33 ± 8% -0.2 0.16 ± 6% perf-profile.self.cycles-pp.rmqueue
> 0.45 ± 3% -0.2 0.29 ± 11% perf-profile.self.cycles-pp.kfree
> 0.45 ± 8% -0.2 0.29 ± 8% perf-profile.self.cycles-pp.sctp_datamsg_from_user
> 0.32 ± 5% -0.2 0.16 ± 9% perf-profile.self.cycles-pp.get_page_from_freelist
> 0.37 ± 7% -0.2 0.21 ± 5% perf-profile.self.cycles-pp.__mod_node_page_state
> 0.52 ± 7% -0.2 0.36 ± 7% perf-profile.self.cycles-pp.__list_del_entry_valid
> 0.32 ± 6% -0.2 0.16 ± 8% perf-profile.self.cycles-pp.__zone_watermark_ok
> 0.41 ± 7% -0.2 0.26 ± 8% perf-profile.self.cycles-pp.kmem_cache_alloc
> 0.46 ± 17% -0.1 0.31 ± 4% perf-profile.self.cycles-pp.__copy_skb_header
> 0.35 ± 7% -0.1 0.20 ± 7% perf-profile.self.cycles-pp.update_load_avg
> 0.39 ± 5% -0.1 0.25 ± 9% perf-profile.self.cycles-pp.sctp_outq_flush_data
> 0.37 ± 7% -0.1 0.23 ± 5% perf-profile.self.cycles-pp.update_rq_clock
> 0.37 ± 6% -0.1 0.24 ± 7% perf-profile.self.cycles-pp.__skb_datagram_iter
> 0.23 ± 13% -0.1 0.10 ± 12% perf-profile.self.cycles-pp.sctp_assoc_rwnd_increase
> 0.38 ± 6% -0.1 0.25 ± 10% perf-profile.self.cycles-pp.sctp_sendmsg
> 0.35 ± 5% -0.1 0.22 ± 9% perf-profile.self.cycles-pp.memset_erms
> 0.37 ± 5% -0.1 0.24 ± 5% perf-profile.self.cycles-pp.set_next_entity
> 0.37 ± 6% -0.1 0.24 ± 6% perf-profile.self.cycles-pp.skb_release_data
> 0.36 ± 13% -0.1 0.23 ± 7% perf-profile.self.cycles-pp.entry_SYSCALL_64_after_hwframe
> 0.35 ± 5% -0.1 0.22 ± 7% perf-profile.self.cycles-pp.check_heap_object
> 0.31 ± 6% -0.1 0.18 ± 14% perf-profile.self.cycles-pp.__rhashtable_lookup
> 0.34 ± 7% -0.1 0.22 ± 10% perf-profile.self.cycles-pp.memcg_slab_free_hook
> 0.34 ± 5% -0.1 0.22 ± 6% perf-profile.self.cycles-pp.free_pcp_prepare
> 0.32 ± 8% -0.1 0.20 ± 7% perf-profile.self.cycles-pp.__virt_addr_valid
> 0.25 ± 7% -0.1 0.13 ± 10% perf-profile.self.cycles-pp.free_unref_page_commit
> 0.39 ± 6% -0.1 0.28 ± 9% perf-profile.self.cycles-pp.enqueue_entity
> 0.32 ± 6% -0.1 0.20 ± 6% perf-profile.self.cycles-pp.__list_add_valid
> 0.30 ± 13% -0.1 0.18 ± 11% perf-profile.self.cycles-pp.sctp_v4_xmit
> 0.32 ± 7% -0.1 0.20 ± 7% perf-profile.self.cycles-pp.__alloc_skb
> 0.33 ± 5% -0.1 0.22 ± 7% perf-profile.self.cycles-pp.__fdget
> 0.30 ± 6% -0.1 0.19 ± 7% perf-profile.self.cycles-pp.sctp_association_put
> 0.31 ± 5% -0.1 0.20 ± 7% perf-profile.self.cycles-pp.__might_sleep
> 0.21 ± 18% -0.1 0.10 ± 7% perf-profile.self.cycles-pp.dst_release
> 0.26 ± 13% -0.1 0.15 ± 13% perf-profile.self.cycles-pp.ipv4_dst_check
> 0.31 ± 6% -0.1 0.21 ± 7% perf-profile.self.cycles-pp.sctp_skb_recv_datagram
> 0.26 ± 9% -0.1 0.16 ± 13% perf-profile.self.cycles-pp.sctp_cmd_interpreter
> 0.21 ± 6% -0.1 0.11 ± 12% perf-profile.self.cycles-pp.sock_wfree
> 0.23 ± 7% -0.1 0.13 ± 6% perf-profile.self.cycles-pp.__alloc_pages
> 0.31 ± 6% -0.1 0.21 ± 9% perf-profile.self.cycles-pp.kmem_cache_alloc_node
> 0.24 ± 5% -0.1 0.14 ± 10% perf-profile.self.cycles-pp.__check_object_size
> 0.17 ± 7% -0.1 0.07 ± 12% perf-profile.self.cycles-pp.update_curr
> 0.18 ± 36% -0.1 0.08 ± 20% perf-profile.self.cycles-pp.sctp_cmp_addr_exact
> 0.31 ± 6% -0.1 0.22 ± 9% perf-profile.self.cycles-pp.enqueue_task_fair
> 0.27 ± 5% -0.1 0.18 ± 7% perf-profile.self.cycles-pp.send_sctp_stream_1toMany
> 0.27 ± 8% -0.1 0.18 ± 8% perf-profile.self.cycles-pp.entry_SYSRETQ_unsafe_stack
> 0.29 ± 6% -0.1 0.20 ± 8% perf-profile.self.cycles-pp.skb_set_owner_w
> 0.26 ± 10% -0.1 0.17 ± 11% perf-profile.self.cycles-pp.sctp_transport_hold
> 0.21 ± 10% -0.1 0.12 ± 8% perf-profile.self.cycles-pp.recv_sctp_stream_1toMany
> 0.24 ± 7% -0.1 0.15 ± 6% perf-profile.self.cycles-pp.__switch_to
> 0.25 ± 12% -0.1 0.17 ± 12% perf-profile.self.cycles-pp.sctp_rcv
> 0.24 ± 4% -0.1 0.16 ± 4% perf-profile.self.cycles-pp.__entry_text_start
> 0.28 ± 12% -0.1 0.20 ± 4% perf-profile.self.cycles-pp.sctp_transport_put
> 0.20 ± 5% -0.1 0.11 ± 6% perf-profile.self.cycles-pp.__wake_up_common
> 0.22 ± 6% -0.1 0.14 ± 5% perf-profile.self.cycles-pp.__switch_to_asm
> 0.22 ± 5% -0.1 0.14 ± 9% perf-profile.self.cycles-pp.consume_skb
> 0.21 ± 7% -0.1 0.13 ± 8% perf-profile.self.cycles-pp.recvmsg
> 0.20 ± 6% -0.1 0.13 ± 9% perf-profile.self.cycles-pp.reweight_entity
> 0.21 ± 6% -0.1 0.14 ± 10% perf-profile.self.cycles-pp.sctp_datamsg_put
> 0.20 ± 4% -0.1 0.13 ± 5% perf-profile.self.cycles-pp.available_idle_cpu
> 0.20 ± 6% -0.1 0.12 ± 6% perf-profile.self.cycles-pp.__free_pages
> 0.10 ± 16% -0.1 0.02 ± 99% perf-profile.self.cycles-pp.sctp_ulpq_tail_data
> 0.16 ± 9% -0.1 0.09 ± 14% perf-profile.self.cycles-pp.____sys_recvmsg
> 0.28 ± 9% -0.1 0.21 ± 7% perf-profile.self.cycles-pp.aa_sk_perm
> 0.18 ± 8% -0.1 0.11 ± 13% perf-profile.self.cycles-pp.syscall_return_via_sysret
> 0.10 ± 9% -0.1 0.03 ±100% perf-profile.self.cycles-pp.process_backlog
> 0.20 ± 7% -0.1 0.14 ± 7% perf-profile.self.cycles-pp.sctp_association_hold
> 0.20 ± 7% -0.1 0.13 ± 9% perf-profile.self.cycles-pp.sctp_hash_cmp
> 0.22 ± 8% -0.1 0.16 ± 4% perf-profile.self.cycles-pp.sctp_wfree
> 0.18 ± 7% -0.1 0.11 ± 6% perf-profile.self.cycles-pp._copy_from_user
> 0.17 ± 8% -0.1 0.10 ± 6% perf-profile.self.cycles-pp.poll_idle
> 0.16 ± 8% -0.1 0.09 ± 14% perf-profile.self.cycles-pp.__update_load_avg_cfs_rq
> 0.19 ± 6% -0.1 0.13 ± 10% perf-profile.self.cycles-pp.__update_load_avg_se
> 0.14 ± 7% -0.1 0.08 ± 8% perf-profile.self.cycles-pp.___perf_sw_event
> 0.16 ± 8% -0.1 0.10 ± 10% perf-profile.self.cycles-pp.try_to_wake_up
> 0.16 ± 5% -0.1 0.10 ± 5% perf-profile.self.cycles-pp.sctp_packet_append_chunk
> 0.20 ± 8% -0.1 0.14 ± 5% perf-profile.self.cycles-pp.sendmsg
> 0.14 ± 12% -0.1 0.08 ± 4% perf-profile.self.cycles-pp.check_new_pages
> 0.17 ± 11% -0.1 0.11 ± 15% perf-profile.self.cycles-pp.sctp_outq_flush
> 0.17 ± 8% -0.1 0.11 ± 9% perf-profile.self.cycles-pp.sctp_assoc_bh_rcv
> 0.17 ± 6% -0.1 0.11 ± 5% perf-profile.self.cycles-pp.nr_iowait_cpu
> 0.17 ± 7% -0.1 0.12 ± 9% perf-profile.self.cycles-pp.sctp_packet_transmit
> 0.12 ± 9% -0.1 0.06 ± 14% perf-profile.self.cycles-pp.dequeue_entity
> 0.15 ± 8% -0.1 0.10 ± 13% perf-profile.self.cycles-pp.____sys_sendmsg
> 0.15 ± 14% -0.1 0.10 ± 8% perf-profile.self.cycles-pp.sctp_ulpevent_free
> 0.18 ± 9% -0.1 0.12 ± 10% perf-profile.self.cycles-pp.sctp_do_sm
> 0.14 ± 4% -0.1 0.09 ± 10% perf-profile.self.cycles-pp.__genradix_ptr
> 0.13 ± 10% -0.1 0.08 ± 6% perf-profile.self.cycles-pp.__put_user_nocheck_4
> 0.15 ± 7% -0.1 0.10 ± 7% perf-profile.self.cycles-pp.__copy_msghdr_from_user
> 0.13 ± 10% -0.1 0.08 ± 12% perf-profile.self.cycles-pp.__import_iovec
> 0.16 ± 4% -0.0 0.11 ± 11% perf-profile.self.cycles-pp.__skb_clone
> 0.16 ± 8% -0.0 0.11 ± 13% perf-profile.self.cycles-pp.sctp_sendmsg_parse
> 0.13 ± 6% -0.0 0.08 ± 8% perf-profile.self.cycles-pp.do_syscall_64
> 0.17 ± 6% -0.0 0.12 ± 5% perf-profile.self.cycles-pp.do_idle
> 0.12 ± 17% -0.0 0.07 ± 18% perf-profile.self.cycles-pp.sctp_hash_key
> 0.14 ± 9% -0.0 0.09 ± 9% perf-profile.self.cycles-pp.sctp_check_transmitted
> 0.13 ± 5% -0.0 0.08 ± 7% perf-profile.self.cycles-pp.sock_kmalloc
> 0.11 ± 6% -0.0 0.06 ± 7% perf-profile.self.cycles-pp.sctp_sock_rfree
> 0.11 ± 6% -0.0 0.07 ± 7% perf-profile.self.cycles-pp.__kmalloc_node_track_caller
> 0.12 ± 9% -0.0 0.07 ± 10% perf-profile.self.cycles-pp.___sys_recvmsg
> 0.14 ± 8% -0.0 0.09 ± 7% perf-profile.self.cycles-pp.switch_mm_irqs_off
> 0.07 ± 10% -0.0 0.03 ±100% perf-profile.self.cycles-pp._sctp_make_chunk
> 0.22 ± 8% -0.0 0.18 ± 4% perf-profile.self.cycles-pp.read_tsc
> 0.12 ± 5% -0.0 0.07 ± 6% perf-profile.self.cycles-pp.__put_user_nocheck_8
> 0.13 ± 8% -0.0 0.08 ± 8% perf-profile.self.cycles-pp.sctp_inq_pop
> 0.12 ± 6% -0.0 0.08 ± 7% perf-profile.self.cycles-pp.__check_heap_object
> 0.12 ± 4% -0.0 0.08 ± 8% perf-profile.self.cycles-pp.__kmalloc
> 0.08 ± 8% -0.0 0.04 ± 71% perf-profile.self.cycles-pp.__mod_lruvec_page_state
> 0.12 ± 8% -0.0 0.08 ± 6% perf-profile.self.cycles-pp.__local_bh_enable_ip
> 0.12 ± 7% -0.0 0.08 ± 8% perf-profile.self.cycles-pp.sctp_sf_eat_data_6_2
> 0.12 ± 6% -0.0 0.07 ± 10% perf-profile.self.cycles-pp.__free_pages_ok
> 0.11 ± 8% -0.0 0.07 ± 7% perf-profile.self.cycles-pp._copy_to_iter
> 0.11 ± 6% -0.0 0.06 ± 7% perf-profile.self.cycles-pp.__ip_queue_xmit
> 0.08 ± 10% -0.0 0.03 ± 70% perf-profile.self.cycles-pp.__wrgsbase_inactive
> 0.13 ± 8% -0.0 0.09 ± 5% perf-profile.self.cycles-pp.pick_next_entity
> 0.11 ± 8% -0.0 0.07 ± 7% perf-profile.self.cycles-pp.select_task_rq
> 0.11 ± 7% -0.0 0.07 ± 8% perf-profile.self.cycles-pp.sctp_inet_skb_msgname
> 0.10 ± 9% -0.0 0.06 ± 11% perf-profile.self.cycles-pp.sctp_outq_sack
> 0.08 ± 5% -0.0 0.05 ± 45% perf-profile.self.cycles-pp.__netif_receive_skb_core
> 0.20 ± 8% -0.0 0.16 ± 5% perf-profile.self.cycles-pp.native_sched_clock
> 0.12 ± 9% -0.0 0.08 ± 10% perf-profile.self.cycles-pp.memcg_slab_post_alloc_hook
> 0.10 ± 7% -0.0 0.06 ± 14% perf-profile.self.cycles-pp.__cond_resched
> 0.12 ± 4% -0.0 0.08 ± 12% perf-profile.self.cycles-pp.kmem_cache_alloc_trace
> 0.10 ± 5% -0.0 0.06 ± 15% perf-profile.self.cycles-pp.check_stack_object
> 0.12 ± 6% -0.0 0.08 ± 7% perf-profile.self.cycles-pp.sctp_outq_tail
> 0.12 ± 8% -0.0 0.08 ± 12% perf-profile.self.cycles-pp.sctp_chunk_abandoned
> 0.10 ± 4% -0.0 0.06 ± 14% perf-profile.self.cycles-pp.iovec_from_user
> 0.09 ± 7% -0.0 0.06 ± 6% perf-profile.self.cycles-pp.os_xsave
> 0.13 ± 7% -0.0 0.09 ± 11% perf-profile.self.cycles-pp.sctp_sched_dequeue_common
> 0.07 ± 11% -0.0 0.04 ± 71% perf-profile.self.cycles-pp.move_addr_to_kernel
> 0.09 ± 10% -0.0 0.05 ± 8% perf-profile.self.cycles-pp.sctp_ulpevent_init
> 0.08 ± 10% -0.0 0.05 ± 45% perf-profile.self.cycles-pp.free_unref_page
> 0.10 ± 5% -0.0 0.06 ± 11% perf-profile.self.cycles-pp.__might_fault
> 0.11 ± 12% -0.0 0.08 ± 10% perf-profile.self.cycles-pp.__sctp_packet_append_chunk
> 0.07 ± 11% -0.0 0.03 ± 70% perf-profile.self.cycles-pp.sockfd_lookup_light
> 0.11 ± 7% -0.0 0.08 ± 9% perf-profile.self.cycles-pp.cpuidle_idle_call
> 0.12 ± 7% -0.0 0.09 ± 10% perf-profile.self.cycles-pp.ip_finish_output2
> 0.12 ± 8% -0.0 0.09 ± 8% perf-profile.self.cycles-pp.resched_curr
> 0.10 ± 5% -0.0 0.07 ± 10% perf-profile.self.cycles-pp.skb_put
> 0.09 ± 7% -0.0 0.06 ± 6% perf-profile.self.cycles-pp.sctp_outq_select_transport
> 0.08 ± 6% -0.0 0.04 ± 45% perf-profile.self.cycles-pp.__sys_recvmsg
> 0.08 ± 5% -0.0 0.05 ± 8% perf-profile.self.cycles-pp.sock_kfree_s
> 0.10 ± 5% -0.0 0.07 ± 14% perf-profile.self.cycles-pp.net_rx_action
> 0.09 ± 8% -0.0 0.06 ± 11% perf-profile.self.cycles-pp.sctp_chunkify
> 0.10 ± 9% -0.0 0.07 ± 10% perf-profile.self.cycles-pp.syscall_exit_to_user_mode_prepare
> 0.14 ± 10% -0.0 0.11 ± 5% perf-profile.self.cycles-pp.update_cfs_group
> 0.09 ± 8% -0.0 0.06 ± 9% perf-profile.self.cycles-pp.security_socket_recvmsg
> 0.08 ± 9% -0.0 0.05 perf-profile.self.cycles-pp.rcu_all_qs
> 0.07 ± 9% -0.0 0.04 ± 45% perf-profile.self.cycles-pp.sctp_tsnmap_check
> 0.08 ± 11% -0.0 0.06 ± 13% perf-profile.self.cycles-pp.sctp_addto_chunk
> 0.08 ± 10% -0.0 0.06 ± 13% perf-profile.self.cycles-pp.kmalloc_large_node
> 0.10 ± 8% +0.0 0.12 ± 6% perf-profile.self.cycles-pp.lock_sock_nested
> 0.00 +0.1 0.06 ± 9% perf-profile.self.cycles-pp.mem_cgroup_uncharge_skmem
> 0.00 +0.1 0.08 ± 11% perf-profile.self.cycles-pp.lapic_next_deadline
> 0.00 +0.1 0.09 ± 13% perf-profile.self.cycles-pp.native_irq_return_iret
> 0.04 ± 71% +0.1 0.12 ± 7% perf-profile.self.cycles-pp.sctp_get_af_specific
> 0.00 +0.1 0.09 ± 41% perf-profile.self.cycles-pp.tsc_verify_tsc_adjust
> 0.00 +0.1 0.09 ± 31% perf-profile.self.cycles-pp.calc_global_load_tick
> 0.00 +0.1 0.10 ± 31% perf-profile.self.cycles-pp._raw_spin_trylock
> 0.01 ±223% +0.1 0.10 ± 50% perf-profile.self.cycles-pp.tick_nohz_next_event
> 0.11 ± 9% +0.1 0.21 ± 10% perf-profile.self.cycles-pp.cpuidle_enter_state
> 0.00 +0.1 0.10 ± 25% perf-profile.self.cycles-pp.arch_scale_freq_tick
> 0.00 +0.1 0.13 ± 19% perf-profile.self.cycles-pp.cgroup_rstat_updated
> 0.00 +0.1 0.14 ± 10% perf-profile.self.cycles-pp.mem_cgroup_charge_skmem
> 0.00 +0.2 0.21 ± 8% perf-profile.self.cycles-pp.refill_stock
> 0.00 +0.2 0.25 ± 9% perf-profile.self.cycles-pp.__sk_mem_raise_allocated
> 0.31 ± 8% +0.3 0.56 ± 23% perf-profile.self.cycles-pp.ktime_get
> 0.06 ± 13% +0.3 0.31 ± 53% perf-profile.self.cycles-pp.timekeeping_max_deferment
> 0.00 +0.3 0.26 ± 11% perf-profile.self.cycles-pp.propagate_protected_usage
> 0.00 +0.8 0.82 ± 3% perf-profile.self.cycles-pp.__mod_memcg_state
> 0.00 +1.2 1.19 ± 7% perf-profile.self.cycles-pp.try_charge_memcg
> 0.00 +2.0 1.96 ± 6% perf-profile.self.cycles-pp.page_counter_uncharge
> 0.00 +2.1 2.07 ± 5% perf-profile.self.cycles-pp.page_counter_try_charge
> 1.09 ± 8% +2.8 3.92 ± 6% perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
> 0.29 ± 6% +3.5 3.81 ± 6% perf-profile.self.cycles-pp.sctp_eat_data
> 0.00 +7.8 7.76 ± 6% perf-profile.self.cycles-pp.__sk_mem_reduce_allocated
>
>
> ***************************************************************************************************
> lkp-cpl-4sp1: 144 threads 4 sockets Intel(R) Xeon(R) Gold 5318H CPU @ 2.50GHz with 128G memory
> =========================================================================================
> cluster/compiler/cpufreq_governor/ip/kconfig/nr_threads/rootfs/runtime/send_size/tbox_group/test/testcase/ucode:
> cs-localhost/gcc-11/performance/ipv4/x86_64-rhel-8.3/50%/debian-11.1-x86_64-20220510.cgz/300s/10K/lkp-cpl-4sp1/SCTP_STREAM_MANY/netperf/0x7002402
>
> commit:
> 7c80b038d2 ("net: fix sk_wmem_schedule() and sk_rmem_schedule() errors")
> 4890b686f4 ("net: keep sk->sk_forward_alloc as small as possible")
>
> 7c80b038d23e1f4c 4890b686f4088c90432149bd6de
> ---------------- ---------------------------
> %stddev %change %stddev
> \ | \
> 9985 -73.7% 2624 netperf.Throughput_Mbps
> 718950 -73.7% 188942 netperf.Throughput_total_Mbps
> 68186 -71.8% 19239 ± 2% netperf.time.involuntary_context_switches
> 29139 ± 2% -16.8% 24247 netperf.time.minor_page_faults
> 3137 -68.1% 1000 netperf.time.percent_of_cpu_this_job_got
> 9264 -67.8% 2979 netperf.time.system_time
> 199.50 -77.7% 44.44 ± 2% netperf.time.user_time
> 316769 ± 2% +21.5% 384840 netperf.time.voluntary_context_switches
> 2.633e+09 -73.7% 6.919e+08 netperf.workload
> 29358 +42.4% 41821 uptime.idle
> 2.381e+10 +52.8% 3.639e+10 cpuidle..time
> 7.884e+08 -65.8% 2.694e+08 cpuidle..usage
> 56.10 +27.9 83.96 mpstat.cpu.all.idle%
> 0.80 -0.2 0.61 ± 5% mpstat.cpu.all.irq%
> 7.50 -4.2 3.29 mpstat.cpu.all.soft%
> 34.45 -22.6 11.86 mpstat.cpu.all.sys%
> 1.14 -0.9 0.27 mpstat.cpu.all.usr%
> 55.83 +49.3% 83.33 vmstat.cpu.id
> 5727924 ± 3% -43.6% 3229130 vmstat.memory.cache
> 64.67 ± 2% -65.2% 22.50 ± 2% vmstat.procs.r
> 4915230 -73.6% 1295498 vmstat.system.cs
> 298596 -3.1% 289395 vmstat.system.in
> 92833 ±101% -63.0% 34380 ± 35% numa-meminfo.node1.Inactive
> 92773 ±101% -62.9% 34380 ± 35% numa-meminfo.node1.Inactive(anon)
> 1468231 ± 19% -74.4% 376377 ± 16% numa-meminfo.node3.Active
> 1468231 ± 19% -74.4% 376377 ± 16% numa-meminfo.node3.Active(anon)
> 925228 ± 17% -95.4% 42887 ± 27% numa-meminfo.node3.Inactive
> 925228 ± 17% -95.4% 42887 ± 27% numa-meminfo.node3.Inactive(anon)
> 1162413 ± 17% -85.6% 167082 ± 21% numa-meminfo.node3.Mapped
> 2348690 ± 14% -83.8% 380889 ± 16% numa-meminfo.node3.Shmem
> 8.422e+08 -72.9% 2.286e+08 ± 5% numa-numastat.node0.local_node
> 8.425e+08 -72.8% 2.288e+08 ± 5% numa-numastat.node0.numa_hit
> 8.344e+08 -73.3% 2.226e+08 ± 5% numa-numastat.node1.local_node
> 8.341e+08 -73.3% 2.227e+08 ± 5% numa-numastat.node1.numa_hit
> 8.584e+08 -74.8% 2.163e+08 ± 10% numa-numastat.node2.local_node
> 8.583e+08 -74.8% 2.164e+08 ± 10% numa-numastat.node2.numa_hit
> 8.684e+08 -74.1% 2.251e+08 ± 6% numa-numastat.node3.local_node
> 8.681e+08 -74.1% 2.252e+08 ± 6% numa-numastat.node3.numa_hit
> 1860369 ± 5% -76.2% 442423 ± 3% meminfo.Active
> 1860257 ± 5% -76.2% 442314 ± 3% meminfo.Active(anon)
> 5599254 ± 3% -44.7% 3099170 meminfo.Cached
> 3542743 ± 6% -71.5% 1010644 meminfo.Committed_AS
> 1423628 ± 11% -76.4% 336110 meminfo.Inactive
> 1423448 ± 11% -76.4% 335928 meminfo.Inactive(anon)
> 1453354 ± 6% -83.5% 240203 ± 3% meminfo.Mapped
> 7494321 ± 2% -34.5% 4912267 meminfo.Memused
> 12458 -18.9% 10104 meminfo.PageTables
> 2948963 ± 7% -84.8% 449291 ± 3% meminfo.Shmem
> 11063817 -30.5% 7685444 meminfo.max_used_kB
> 8.425e+08 -72.8% 2.288e+08 ± 5% numa-vmstat.node0.numa_hit
> 8.422e+08 -72.9% 2.286e+08 ± 5% numa-vmstat.node0.numa_local
> 23213 ±101% -63.0% 8583 ± 35% numa-vmstat.node1.nr_inactive_anon
> 23213 ±101% -63.0% 8583 ± 35% numa-vmstat.node1.nr_zone_inactive_anon
> 8.341e+08 -73.3% 2.227e+08 ± 5% numa-vmstat.node1.numa_hit
> 8.344e+08 -73.3% 2.226e+08 ± 5% numa-vmstat.node1.numa_local
> 8.583e+08 -74.8% 2.164e+08 ± 10% numa-vmstat.node2.numa_hit
> 8.584e+08 -74.8% 2.163e+08 ± 10% numa-vmstat.node2.numa_local
> 366556 ± 19% -74.2% 94616 ± 16% numa-vmstat.node3.nr_active_anon
> 231387 ± 17% -95.4% 10658 ± 27% numa-vmstat.node3.nr_inactive_anon
> 291532 ± 17% -86.8% 38558 ± 21% numa-vmstat.node3.nr_mapped
> 586771 ± 14% -83.7% 95702 ± 16% numa-vmstat.node3.nr_shmem
> 366556 ± 19% -74.2% 94616 ± 16% numa-vmstat.node3.nr_zone_active_anon
> 231387 ± 17% -95.4% 10658 ± 27% numa-vmstat.node3.nr_zone_inactive_anon
> 8.681e+08 -74.1% 2.252e+08 ± 6% numa-vmstat.node3.numa_hit
> 8.684e+08 -74.1% 2.251e+08 ± 6% numa-vmstat.node3.numa_local
> 1495 -61.8% 571.00 turbostat.Avg_MHz
> 47.16 -29.8 17.36 turbostat.Busy%
> 3178 +3.7% 3295 turbostat.Bzy_MHz
> 7.464e+08 -73.4% 1.985e+08 turbostat.C1
> 13.74 -8.9 4.80 ± 2% turbostat.C1%
> 34154266 ± 14% +97.4% 67405025 turbostat.C1E
> 36.80 ± 21% +37.1 73.90 ± 3% turbostat.C1E%
> 52.81 +56.4% 82.60 turbostat.CPU%c1
> 58.83 ± 2% -20.1% 47.00 ± 2% turbostat.CoreTmp
> 0.20 -25.6% 0.15 turbostat.IPC
> 128.56 ± 3% -128.6 0.00 turbostat.PKG_%
> 5447540 ± 3% -83.7% 887322 turbostat.POLL
> 0.13 -0.1 0.03 turbostat.POLL%
> 58.83 ± 2% -19.8% 47.17 ± 2% turbostat.PkgTmp
> 591.77 -28.1% 425.61 turbostat.PkgWatt
> 8.17 +3.5% 8.46 turbostat.RAMWatt
> 465173 ± 5% -76.3% 110436 ± 3% proc-vmstat.nr_active_anon
> 83691 -1.7% 82260 proc-vmstat.nr_anon_pages
> 3086038 +2.1% 3150528 proc-vmstat.nr_dirty_background_threshold
> 6179622 +2.1% 6308761 proc-vmstat.nr_dirty_threshold
> 1399831 ± 3% -44.7% 774653 proc-vmstat.nr_file_pages
> 31038779 +2.1% 31684629 proc-vmstat.nr_free_pages
> 355768 ± 11% -76.4% 83981 proc-vmstat.nr_inactive_anon
> 363110 ± 6% -83.4% 60381 ± 3% proc-vmstat.nr_mapped
> 3114 -18.9% 2526 proc-vmstat.nr_page_table_pages
> 737257 ± 7% -84.8% 112182 ± 3% proc-vmstat.nr_shmem
> 33546 -3.9% 32233 proc-vmstat.nr_slab_reclaimable
> 465173 ± 5% -76.3% 110436 ± 3% proc-vmstat.nr_zone_active_anon
> 355768 ± 11% -76.4% 83981 proc-vmstat.nr_zone_inactive_anon
> 125868 ± 15% -29.5% 88747 ± 12% proc-vmstat.numa_hint_faults_local
> 3.403e+09 -73.8% 8.93e+08 proc-vmstat.numa_hit
> 3.403e+09 -73.8% 8.927e+08 proc-vmstat.numa_local
> 1698265 ± 8% -91.3% 146940 ± 2% proc-vmstat.pgactivate
> 9.031e+09 -73.7% 2.375e+09 proc-vmstat.pgalloc_normal
> 1809374 ± 2% -11.5% 1601059 ± 2% proc-vmstat.pgfault
> 9.031e+09 -73.7% 2.375e+09 proc-vmstat.pgfree
> 40961 +1.3% 41503 proc-vmstat.pgreuse
> 121388 ± 29% -96.2% 4632 ± 47% sched_debug.cfs_rq:/.MIN_vruntime.avg
> 3776647 ± 10% -91.9% 305894 ± 40% sched_debug.cfs_rq:/.MIN_vruntime.max
> 623786 ± 18% -94.3% 35468 ± 44% sched_debug.cfs_rq:/.MIN_vruntime.stddev
> 0.40 ± 7% -58.4% 0.17 ± 7% sched_debug.cfs_rq:/.h_nr_running.avg
> 0.45 -19.1% 0.37 ± 3% sched_debug.cfs_rq:/.h_nr_running.stddev
> 121388 ± 29% -96.2% 4632 ± 47% sched_debug.cfs_rq:/.max_vruntime.avg
> 3776647 ± 10% -91.9% 305894 ± 40% sched_debug.cfs_rq:/.max_vruntime.max
> 623786 ± 18% -94.3% 35468 ± 44% sched_debug.cfs_rq:/.max_vruntime.stddev
> 3455794 ± 6% -89.2% 374393 ± 9% sched_debug.cfs_rq:/.min_vruntime.avg
> 4153720 ± 7% -88.1% 494762 ± 7% sched_debug.cfs_rq:/.min_vruntime.max
> 2710531 ± 7% -91.1% 240412 ± 10% sched_debug.cfs_rq:/.min_vruntime.min
> 368488 ± 8% -85.9% 51911 ± 13% sched_debug.cfs_rq:/.min_vruntime.stddev
> 0.40 ± 7% -58.4% 0.17 ± 7% sched_debug.cfs_rq:/.nr_running.avg
> 0.45 -19.2% 0.37 ± 3% sched_debug.cfs_rq:/.nr_running.stddev
> 386.20 ± 3% -56.3% 168.94 sched_debug.cfs_rq:/.runnable_avg.avg
> 1215 ± 8% -18.7% 987.82 ± 4% sched_debug.cfs_rq:/.runnable_avg.max
> 340.61 -31.1% 234.69 ± 2% sched_debug.cfs_rq:/.runnable_avg.stddev
> 429174 ± 24% -103.8% -16476 sched_debug.cfs_rq:/.spread0.avg
> 1127012 ± 15% -90.8% 103881 ± 28% sched_debug.cfs_rq:/.spread0.max
> -315952 -52.4% -150458 sched_debug.cfs_rq:/.spread0.min
> 368407 ± 8% -85.9% 51908 ± 13% sched_debug.cfs_rq:/.spread0.stddev
> 385.96 ± 3% -56.3% 168.81 sched_debug.cfs_rq:/.util_avg.avg
> 1215 ± 8% -18.7% 987.79 ± 4% sched_debug.cfs_rq:/.util_avg.max
> 340.53 -31.1% 234.60 ± 2% sched_debug.cfs_rq:/.util_avg.stddev
> 274.31 ± 7% -74.9% 68.97 ± 5% sched_debug.cfs_rq:/.util_est_enqueued.avg
> 981.12 -10.6% 877.51 ± 3% sched_debug.cfs_rq:/.util_est_enqueued.max
> 337.06 -47.8% 176.03 ± 3% sched_debug.cfs_rq:/.util_est_enqueued.stddev
> 457832 ± 5% +41.2% 646422 ± 5% sched_debug.cpu.avg_idle.avg
> 1762 ± 7% -42.8% 1009 ± 7% sched_debug.cpu.clock_task.stddev
> 2358 ± 5% -61.0% 920.92 ± 2% sched_debug.cpu.curr->pid.avg
> 2793 -22.1% 2176 ± 2% sched_debug.cpu.curr->pid.stddev
> 557670 ± 6% +8.4% 604409 ± 7% sched_debug.cpu.max_idle_balance_cost.max
> 0.00 ± 2% -27.5% 0.00 ± 5% sched_debug.cpu.next_balance.stddev
> 0.39 ± 5% -61.1% 0.15 ± 3% sched_debug.cpu.nr_running.avg
> 0.45 -22.2% 0.35 sched_debug.cpu.nr_running.stddev
> 5005510 ± 7% -75.6% 1221805 ± 10% sched_debug.cpu.nr_switches.avg
> 6928957 ± 8% -70.7% 2030568 ± 9% sched_debug.cpu.nr_switches.max
> 3447354 ± 12% -84.3% 541402 ± 17% sched_debug.cpu.nr_switches.min
> 629520 ± 7% -55.8% 278525 ± 6% sched_debug.cpu.nr_switches.stddev
> 3.074e+10 -71.6% 8.745e+09 perf-stat.i.branch-instructions
> 2.971e+08 ± 2% -70.3% 88205513 ± 4% perf-stat.i.branch-misses
> 1.05 ± 3% +5.6 6.66 perf-stat.i.cache-miss-rate%
> 30824055 ± 3% +98.8% 61267408 perf-stat.i.cache-misses
> 3.297e+09 -71.8% 9.296e+08 perf-stat.i.cache-references
> 4957450 -73.6% 1308421 perf-stat.i.context-switches
> 1.43 +33.1% 1.90 perf-stat.i.cpi
> 2.196e+11 -62.4% 8.259e+10 perf-stat.i.cpu-cycles
> 929.67 -64.3% 331.55 perf-stat.i.cpu-migrations
> 7494 ± 4% -82.0% 1351 perf-stat.i.cycles-between-cache-misses
> 0.01 ± 26% -0.0 0.00 perf-stat.i.dTLB-load-miss-rate%
> 2694006 ± 27% -90.1% 265691 perf-stat.i.dTLB-load-misses
> 4.398e+10 -71.5% 1.252e+10 perf-stat.i.dTLB-loads
> 0.00 ± 7% -0.0 0.00 ± 5% perf-stat.i.dTLB-store-miss-rate%
> 967059 ± 8% -84.3% 151835 ± 6% perf-stat.i.dTLB-store-misses
> 2.599e+10 -71.6% 7.377e+09 perf-stat.i.dTLB-stores
> 63.87 +10.9 74.76 perf-stat.i.iTLB-load-miss-rate%
> 1.972e+08 -70.5% 58088811 ± 5% perf-stat.i.iTLB-load-misses
> 1.126e+08 -82.7% 19440392 perf-stat.i.iTLB-loads
> 1.538e+11 -71.5% 4.38e+10 perf-stat.i.instructions
> 0.71 -24.8% 0.53 perf-stat.i.ipc
> 1.53 -62.4% 0.57 perf-stat.i.metric.GHz
> 909.37 -69.9% 273.61 perf-stat.i.metric.K/sec
> 722.30 -71.6% 205.33 perf-stat.i.metric.M/sec
> 5562 ± 3% -12.4% 4874 ± 2% perf-stat.i.minor-faults
> 8844727 ± 5% +30.4% 11534796 perf-stat.i.node-load-misses
> 623949 ± 11% +54.9% 966213 ± 2% perf-stat.i.node-loads
> 91.43 +5.4 96.84 perf-stat.i.node-store-miss-rate%
> 3461624 ± 2% +66.0% 5746681 perf-stat.i.node-store-misses
> 410920 ± 7% -49.7% 206502 ± 3% perf-stat.i.node-stores
> 5563 ± 3% -12.4% 4876 ± 2% perf-stat.i.page-faults
> 21.44 -1.0% 21.22 perf-stat.overall.MPKI
> 0.93 ± 4% +5.7 6.59 perf-stat.overall.cache-miss-rate%
> 1.43 +32.0% 1.89 perf-stat.overall.cpi
> 7136 ± 4% -81.1% 1348 perf-stat.overall.cycles-between-cache-misses
> 0.01 ± 27% -0.0 0.00 perf-stat.overall.dTLB-load-miss-rate%
> 0.00 ± 7% -0.0 0.00 ± 5% perf-stat.overall.dTLB-store-miss-rate%
> 63.66 +11.2 74.89 perf-stat.overall.iTLB-load-miss-rate%
> 0.70 -24.3% 0.53 perf-stat.overall.ipc
> 89.40 +7.1 96.53 perf-stat.overall.node-store-miss-rate%
> 17593 +8.6% 19108 perf-stat.overall.path-length
> 3.063e+10 -71.5% 8.716e+09 perf-stat.ps.branch-instructions
> 2.961e+08 ± 2% -70.3% 87907848 ± 4% perf-stat.ps.branch-misses
> 30717141 ± 3% +98.8% 61062871 perf-stat.ps.cache-misses
> 3.286e+09 -71.8% 9.265e+08 perf-stat.ps.cache-references
> 4940513 -73.6% 1304046 perf-stat.ps.context-switches
> 2.189e+11 -62.4% 8.231e+10 perf-stat.ps.cpu-cycles
> 926.60 -64.3% 330.61 perf-stat.ps.cpu-migrations
> 2684405 ± 27% -90.1% 264976 perf-stat.ps.dTLB-load-misses
> 4.383e+10 -71.5% 1.248e+10 perf-stat.ps.dTLB-loads
> 963689 ± 8% -84.3% 151363 ± 6% perf-stat.ps.dTLB-store-misses
> 2.591e+10 -71.6% 7.352e+09 perf-stat.ps.dTLB-stores
> 1.965e+08 -70.5% 57894987 ± 5% perf-stat.ps.iTLB-load-misses
> 1.122e+08 -82.7% 19375491 perf-stat.ps.iTLB-loads
> 1.532e+11 -71.5% 4.366e+10 perf-stat.ps.instructions
> 5536 ± 3% -12.3% 4857 ± 2% perf-stat.ps.minor-faults
> 8813664 ± 5% +30.4% 11496729 perf-stat.ps.node-load-misses
> 622201 ± 11% +54.8% 963024 ± 2% perf-stat.ps.node-loads
> 3450020 ± 2% +66.0% 5727738 perf-stat.ps.node-store-misses
> 409584 ± 8% -49.7% 206027 ± 3% perf-stat.ps.node-stores
> 5537 ± 3% -12.3% 4858 ± 2% perf-stat.ps.page-faults
> 4.632e+13 -71.5% 1.322e+13 perf-stat.total.instructions
> 9.54 ± 4% -6.1 3.40 ± 5% perf-profile.calltrace.cycles-pp.intel_idle_irq.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
> 9.17 ± 5% -6.1 3.05 ± 6% perf-profile.calltrace.cycles-pp.mwait_idle_with_hints.intel_idle_irq.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call
> 8.84 ± 4% -3.2 5.62 ± 4% perf-profile.calltrace.cycles-pp.sctp_do_sm.sctp_primitive_SEND.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg
> 8.63 ± 4% -3.1 5.53 ± 4% perf-profile.calltrace.cycles-pp.sctp_packet_pack.sctp_packet_transmit.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm
> 8.18 ± 4% -2.9 5.29 ± 4% perf-profile.calltrace.cycles-pp.memcpy_erms.sctp_packet_pack.sctp_packet_transmit.sctp_outq_flush.sctp_cmd_interpreter
> 6.49 ± 4% -2.5 3.99 ± 4% perf-profile.calltrace.cycles-pp.skb_copy_datagram_iter.sctp_recvmsg.inet_recvmsg.____sys_recvmsg.___sys_recvmsg
> 6.47 ± 3% -2.5 3.98 ± 4% perf-profile.calltrace.cycles-pp.__skb_datagram_iter.skb_copy_datagram_iter.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
> 21.33 ± 3% -2.5 18.86 ± 4% perf-profile.calltrace.cycles-pp.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg.____sys_sendmsg.___sys_sendmsg
> 13.50 ± 3% -2.5 11.04 ± 4% perf-profile.calltrace.cycles-pp.sctp_primitive_SEND.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg.____sys_sendmsg
> 6.89 ± 3% -2.4 4.50 ± 4% perf-profile.calltrace.cycles-pp.sctp_datamsg_from_user.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg.____sys_sendmsg
> 5.63 ± 3% -2.1 3.49 ± 4% perf-profile.calltrace.cycles-pp._copy_to_iter.__skb_datagram_iter.skb_copy_datagram_iter.sctp_recvmsg.inet_recvmsg
> 5.46 ± 3% -2.1 3.40 ± 4% perf-profile.calltrace.cycles-pp.copyout._copy_to_iter.__skb_datagram_iter.skb_copy_datagram_iter.sctp_recvmsg
> 5.42 ± 3% -2.0 3.38 ± 4% perf-profile.calltrace.cycles-pp.copy_user_enhanced_fast_string.copyout._copy_to_iter.__skb_datagram_iter.skb_copy_datagram_iter
> 11.96 ± 3% -1.9 10.08 ± 4% perf-profile.calltrace.cycles-pp.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm.sctp_primitive_SEND.sctp_sendmsg_to_asoc
> 9.08 ± 4% -1.6 7.48 ± 4% perf-profile.calltrace.cycles-pp.sctp_cmd_interpreter.sctp_do_sm.sctp_primitive_SEND.sctp_sendmsg_to_asoc.sctp_sendmsg
> 1.76 ± 4% -1.3 0.43 ± 44% perf-profile.calltrace.cycles-pp.__alloc_pages.kmalloc_large_node.__kmalloc_node_track_caller.kmalloc_reserve.__alloc_skb
> 10.30 ± 4% -1.2 9.11 ± 4% perf-profile.calltrace.cycles-pp.sctp_packet_transmit.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm.sctp_primitive_SEND
> 3.42 ± 3% -1.1 2.30 ± 4% perf-profile.calltrace.cycles-pp.sctp_user_addto_chunk.sctp_datamsg_from_user.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg
> 2.85 ± 3% -1.1 1.78 ± 4% perf-profile.calltrace.cycles-pp.sctp_make_datafrag_empty.sctp_datamsg_from_user.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg
> 1.79 ± 4% -1.0 0.75 ± 4% perf-profile.calltrace.cycles-pp.skb_release_data.kfree_skb_reason.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
> 3.08 ± 4% -1.0 2.07 ± 4% perf-profile.calltrace.cycles-pp.sctp_data_ready.sctp_ulpq_tail_event.sctp_ulpq_tail_data.sctp_cmd_interpreter.sctp_do_sm
> 3.08 ± 3% -1.0 2.08 ± 4% perf-profile.calltrace.cycles-pp._copy_from_iter.sctp_user_addto_chunk.sctp_datamsg_from_user.sctp_sendmsg_to_asoc.sctp_sendmsg
> 3.20 ± 4% -1.0 2.20 ± 4% perf-profile.calltrace.cycles-pp.sctp_ulpq_tail_event.sctp_ulpq_tail_data.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv
> 2.82 ± 4% -1.0 1.84 ± 4% perf-profile.calltrace.cycles-pp.__wake_up_common_lock.sctp_data_ready.sctp_ulpq_tail_event.sctp_ulpq_tail_data.sctp_cmd_interpreter
> 2.94 ± 3% -1.0 1.97 ± 4% perf-profile.calltrace.cycles-pp.copyin._copy_from_iter.sctp_user_addto_chunk.sctp_datamsg_from_user.sctp_sendmsg_to_asoc
> 2.42 ± 3% -1.0 1.46 ± 4% perf-profile.calltrace.cycles-pp._sctp_make_chunk.sctp_make_datafrag_empty.sctp_datamsg_from_user.sctp_sendmsg_to_asoc.sctp_sendmsg
> 2.90 ± 3% -0.9 1.95 ± 4% perf-profile.calltrace.cycles-pp.copy_user_enhanced_fast_string.copyin._copy_from_iter.sctp_user_addto_chunk.sctp_datamsg_from_user
> 11.68 ± 3% -0.9 10.75 ± 4% perf-profile.calltrace.cycles-pp.sctp_packet_transmit.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv
> 2.63 ± 4% -0.9 1.70 ± 4% perf-profile.calltrace.cycles-pp.__wake_up_common.__wake_up_common_lock.sctp_data_ready.sctp_ulpq_tail_event.sctp_ulpq_tail_data
> 2.46 ± 4% -0.9 1.59 ± 4% perf-profile.calltrace.cycles-pp.autoremove_wake_function.__wake_up_common.__wake_up_common_lock.sctp_data_ready.sctp_ulpq_tail_event
> 2.44 ± 4% -0.9 1.58 ± 4% perf-profile.calltrace.cycles-pp.try_to_wake_up.autoremove_wake_function.__wake_up_common.__wake_up_common_lock.sctp_data_ready
> 11.49 ± 3% -0.8 10.64 ± 4% perf-profile.calltrace.cycles-pp.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv.sctp_backlog_rcv
> 1.82 ± 3% -0.8 1.02 ± 4% perf-profile.calltrace.cycles-pp.sctp_ulpevent_free.sctp_recvmsg.inet_recvmsg.____sys_recvmsg.___sys_recvmsg
> 1.86 ± 4% -0.8 1.08 ± 4% perf-profile.calltrace.cycles-pp.__alloc_skb._sctp_make_chunk.sctp_make_datafrag_empty.sctp_datamsg_from_user.sctp_sendmsg_to_asoc
> 1.88 ± 3% -0.8 1.10 ± 5% perf-profile.calltrace.cycles-pp.schedule_timeout.sctp_skb_recv_datagram.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
> 1.84 ± 3% -0.8 1.09 ± 5% perf-profile.calltrace.cycles-pp.schedule.schedule_timeout.sctp_skb_recv_datagram.sctp_recvmsg.inet_recvmsg
> 1.81 ± 3% -0.7 1.07 ± 5% perf-profile.calltrace.cycles-pp.__schedule.schedule.schedule_timeout.sctp_skb_recv_datagram.sctp_recvmsg
> 1.68 ± 4% -0.7 0.98 ± 6% perf-profile.calltrace.cycles-pp.sctp_outq_flush_data.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm.sctp_primitive_SEND
> 1.65 ± 4% -0.7 0.99 ± 3% perf-profile.calltrace.cycles-pp.schedule_idle.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
> 1.61 ± 4% -0.6 0.96 ± 3% perf-profile.calltrace.cycles-pp.__schedule.schedule_idle.do_idle.cpu_startup_entry.start_secondary
> 1.30 ± 3% -0.6 0.69 ± 3% perf-profile.calltrace.cycles-pp.kmalloc_reserve.__alloc_skb._sctp_make_chunk.sctp_make_datafrag_empty.sctp_datamsg_from_user
> 1.27 ± 3% -0.6 0.67 ± 3% perf-profile.calltrace.cycles-pp.__kmalloc_node_track_caller.kmalloc_reserve.__alloc_skb._sctp_make_chunk.sctp_make_datafrag_empty
> 1.24 ± 3% -0.6 0.65 ± 3% perf-profile.calltrace.cycles-pp.kmalloc_large_node.__kmalloc_node_track_caller.kmalloc_reserve.__alloc_skb._sctp_make_chunk
> 0.97 ± 3% -0.5 0.46 ± 44% perf-profile.calltrace.cycles-pp.dequeue_task_fair.__schedule.schedule.schedule_timeout.sctp_skb_recv_datagram
> 1.02 ± 3% -0.5 0.57 ± 3% perf-profile.calltrace.cycles-pp.sctp_chunk_put.sctp_ulpevent_free.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
> 1.10 ± 4% -0.4 0.67 ± 5% perf-profile.calltrace.cycles-pp.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv.sctp_rcv
> 0.92 ± 3% -0.4 0.56 ± 6% perf-profile.calltrace.cycles-pp.skb_release_data.consume_skb.sctp_chunk_put.sctp_outq_sack.sctp_cmd_interpreter
> 1.04 ± 4% -0.3 0.71 ± 4% perf-profile.calltrace.cycles-pp.sendmsg_copy_msghdr.___sys_sendmsg.__sys_sendmsg.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 0.17 ±141% +0.5 0.70 ± 15% perf-profile.calltrace.cycles-pp.menu_select.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary
> 0.00 +0.9 0.91 ± 5% perf-profile.calltrace.cycles-pp.page_counter_uncharge.drain_stock.refill_stock.__sk_mem_reduce_allocated.skb_release_head_state
> 0.00 +0.9 0.93 ± 5% perf-profile.calltrace.cycles-pp.drain_stock.refill_stock.__sk_mem_reduce_allocated.skb_release_head_state.kfree_skb_reason
> 0.00 +1.0 0.95 ± 15% perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call
> 0.00 +1.0 1.03 ± 5% perf-profile.calltrace.cycles-pp.refill_stock.__sk_mem_reduce_allocated.skb_release_head_state.kfree_skb_reason.sctp_recvmsg
> 0.00 +1.0 1.04 ± 14% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
> 11.57 ± 3% +1.3 12.82 ± 4% perf-profile.calltrace.cycles-pp.sctp_assoc_bh_rcv.sctp_backlog_rcv.__release_sock.release_sock.sctp_sendmsg
> 0.00 +1.3 1.33 ± 4% perf-profile.calltrace.cycles-pp.page_counter_uncharge.drain_stock.refill_stock.__sk_mem_reduce_allocated.sctp_wfree
> 0.00 +1.4 1.36 ± 4% perf-profile.calltrace.cycles-pp.drain_stock.refill_stock.__sk_mem_reduce_allocated.sctp_wfree.skb_release_head_state
> 1.67 ± 3% +1.4 3.04 ± 4% perf-profile.calltrace.cycles-pp.sctp_ulpevent_make_rcvmsg.sctp_ulpq_tail_data.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv
> 11.39 ± 3% +1.4 12.77 ± 4% perf-profile.calltrace.cycles-pp.sctp_do_sm.sctp_assoc_bh_rcv.sctp_backlog_rcv.__release_sock.release_sock
> 0.00 +1.4 1.45 ± 4% perf-profile.calltrace.cycles-pp.refill_stock.__sk_mem_reduce_allocated.sctp_wfree.skb_release_head_state.consume_skb
> 0.00 +1.5 1.46 ± 5% perf-profile.calltrace.cycles-pp.try_charge_memcg.mem_cgroup_charge_skmem.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_ulpevent_make_rcvmsg
> 0.00 +1.7 1.74 ± 4% perf-profile.calltrace.cycles-pp.mem_cgroup_charge_skmem.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_ulpevent_make_rcvmsg.sctp_ulpq_tail_data
> 0.00 +1.8 1.83 ± 4% perf-profile.calltrace.cycles-pp.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_ulpevent_make_rcvmsg.sctp_ulpq_tail_data.sctp_cmd_interpreter
> 0.00 +1.8 1.84 ± 4% perf-profile.calltrace.cycles-pp.__sk_mem_schedule.sctp_ulpevent_make_rcvmsg.sctp_ulpq_tail_data.sctp_cmd_interpreter.sctp_do_sm
> 0.00 +2.0 1.97 ± 5% perf-profile.calltrace.cycles-pp.page_counter_try_charge.try_charge_memcg.mem_cgroup_charge_skmem.__sk_mem_raise_allocated.__sk_mem_schedule
> 18.94 ± 3% +2.1 21.02 ± 4% perf-profile.calltrace.cycles-pp.recvmsg
> 0.00 +2.1 2.14 ± 4% perf-profile.calltrace.cycles-pp.try_charge_memcg.mem_cgroup_charge_skmem.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_sendmsg_to_asoc
> 10.88 ± 4% +2.2 13.11 ± 4% perf-profile.calltrace.cycles-pp.__ip_queue_xmit.sctp_packet_transmit.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm
> 10.74 ± 4% +2.3 13.04 ± 4% perf-profile.calltrace.cycles-pp.ip_finish_output2.__ip_queue_xmit.sctp_packet_transmit.sctp_outq_flush.sctp_cmd_interpreter
> 18.22 ± 3% +2.3 20.55 ± 4% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.recvmsg
> 10.51 ± 4% +2.4 12.88 ± 4% perf-profile.calltrace.cycles-pp.__local_bh_enable_ip.ip_finish_output2.__ip_queue_xmit.sctp_packet_transmit.sctp_outq_flush
> 10.47 ± 4% +2.4 12.84 ± 4% perf-profile.calltrace.cycles-pp.do_softirq.__local_bh_enable_ip.ip_finish_output2.__ip_queue_xmit.sctp_packet_transmit
> 18.08 ± 3% +2.4 20.46 ± 4% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.recvmsg
> 10.42 ± 4% +2.4 12.81 ± 4% perf-profile.calltrace.cycles-pp.__softirqentry_text_start.do_softirq.__local_bh_enable_ip.ip_finish_output2.__ip_queue_xmit
> 10.25 ± 4% +2.4 12.70 ± 4% perf-profile.calltrace.cycles-pp.net_rx_action.__softirqentry_text_start.do_softirq.__local_bh_enable_ip.ip_finish_output2
> 17.84 ± 3% +2.5 20.32 ± 4% perf-profile.calltrace.cycles-pp.__sys_recvmsg.do_syscall_64.entry_SYSCALL_64_after_hwframe.recvmsg
> 10.15 ± 4% +2.5 12.64 ± 4% perf-profile.calltrace.cycles-pp.__napi_poll.net_rx_action.__softirqentry_text_start.do_softirq.__local_bh_enable_ip
> 10.13 ± 4% +2.5 12.63 ± 4% perf-profile.calltrace.cycles-pp.process_backlog.__napi_poll.net_rx_action.__softirqentry_text_start.do_softirq
> 0.00 +2.5 2.50 ± 4% perf-profile.calltrace.cycles-pp.mem_cgroup_charge_skmem.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_sendmsg_to_asoc.sctp_sendmsg
> 3.59 ± 3% +2.5 6.11 ± 5% perf-profile.calltrace.cycles-pp.sctp_skb_recv_datagram.sctp_recvmsg.inet_recvmsg.____sys_recvmsg.___sys_recvmsg
> 9.97 ± 4% +2.6 12.54 ± 4% perf-profile.calltrace.cycles-pp.__netif_receive_skb_one_core.process_backlog.__napi_poll.net_rx_action.__softirqentry_text_start
> 17.56 ± 3% +2.6 20.14 ± 4% perf-profile.calltrace.cycles-pp.___sys_recvmsg.__sys_recvmsg.do_syscall_64.entry_SYSCALL_64_after_hwframe.recvmsg
> 0.00 +2.7 2.66 ± 4% perf-profile.calltrace.cycles-pp.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg
> 0.00 +2.7 2.68 ± 4% perf-profile.calltrace.cycles-pp.__sk_mem_schedule.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg.____sys_sendmsg
> 9.70 ± 4% +2.7 12.40 ± 4% perf-profile.calltrace.cycles-pp.ip_local_deliver_finish.__netif_receive_skb_one_core.process_backlog.__napi_poll.net_rx_action
> 9.68 ± 4% +2.7 12.39 ± 4% perf-profile.calltrace.cycles-pp.ip_protocol_deliver_rcu.ip_local_deliver_finish.__netif_receive_skb_one_core.process_backlog.__napi_poll
> 9.61 ± 4% +2.7 12.33 ± 4% perf-profile.calltrace.cycles-pp.sctp_rcv.ip_protocol_deliver_rcu.ip_local_deliver_finish.__netif_receive_skb_one_core.process_backlog
> 16.65 ± 3% +2.9 19.58 ± 4% perf-profile.calltrace.cycles-pp.____sys_recvmsg.___sys_recvmsg.__sys_recvmsg.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 8.06 ± 4% +3.1 11.19 ± 4% perf-profile.calltrace.cycles-pp.sctp_assoc_bh_rcv.sctp_rcv.ip_protocol_deliver_rcu.ip_local_deliver_finish.__netif_receive_skb_one_core
> 15.88 ± 3% +3.2 19.12 ± 4% perf-profile.calltrace.cycles-pp.inet_recvmsg.____sys_recvmsg.___sys_recvmsg.__sys_recvmsg.do_syscall_64
> 15.83 ± 3% +3.3 19.09 ± 4% perf-profile.calltrace.cycles-pp.sctp_recvmsg.inet_recvmsg.____sys_recvmsg.___sys_recvmsg.__sys_recvmsg
> 7.62 ± 3% +3.3 10.90 ± 4% perf-profile.calltrace.cycles-pp.sctp_do_sm.sctp_assoc_bh_rcv.sctp_rcv.ip_protocol_deliver_rcu.ip_local_deliver_finish
> 1.14 ± 3% +3.5 4.62 ± 5% perf-profile.calltrace.cycles-pp.lock_sock_nested.sctp_skb_recv_datagram.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
> 1.02 ± 3% +3.5 4.51 ± 5% perf-profile.calltrace.cycles-pp._raw_spin_lock_bh.lock_sock_nested.sctp_skb_recv_datagram.sctp_recvmsg.inet_recvmsg
> 0.83 ± 4% +3.6 4.38 ± 5% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_bh.lock_sock_nested.sctp_skb_recv_datagram.sctp_recvmsg
> 15.16 ± 3% +3.6 18.74 ± 5% perf-profile.calltrace.cycles-pp.release_sock.sctp_sendmsg.sock_sendmsg.____sys_sendmsg.___sys_sendmsg
> 15.04 ± 3% +3.6 18.66 ± 5% perf-profile.calltrace.cycles-pp.__release_sock.release_sock.sctp_sendmsg.sock_sendmsg.____sys_sendmsg
> 11.66 ± 3% +3.6 15.29 ± 5% perf-profile.calltrace.cycles-pp.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv.sctp_backlog_rcv.__release_sock
> 14.98 ± 3% +3.6 18.62 ± 5% perf-profile.calltrace.cycles-pp.sctp_backlog_rcv.__release_sock.release_sock.sctp_sendmsg.sock_sendmsg
> 0.00 +4.2 4.24 ± 4% perf-profile.calltrace.cycles-pp.sctp_eat_data.sctp_sf_eat_data_6_2.sctp_do_sm.sctp_assoc_bh_rcv.sctp_rcv
> 0.00 +4.3 4.33 ± 4% perf-profile.calltrace.cycles-pp.sctp_sf_eat_data_6_2.sctp_do_sm.sctp_assoc_bh_rcv.sctp_rcv.ip_protocol_deliver_rcu
> 2.88 ± 3% +4.7 7.54 ± 4% perf-profile.calltrace.cycles-pp.sctp_outq_sack.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv.sctp_backlog_rcv
> 23.81 ± 12% +4.8 28.57 ± 11% perf-profile.calltrace.cycles-pp.mwait_idle_with_hints.intel_idle.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call
> 23.81 ± 12% +4.8 28.58 ± 11% perf-profile.calltrace.cycles-pp.intel_idle.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
> 2.11 ± 4% +4.8 6.89 ± 4% perf-profile.calltrace.cycles-pp.kfree_skb_reason.sctp_recvmsg.inet_recvmsg.____sys_recvmsg.___sys_recvmsg
> 1.84 ± 3% +5.0 6.86 ± 4% perf-profile.calltrace.cycles-pp.sctp_chunk_put.sctp_outq_sack.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv
> 1.42 ± 3% +5.0 6.47 ± 4% perf-profile.calltrace.cycles-pp.consume_skb.sctp_chunk_put.sctp_outq_sack.sctp_cmd_interpreter.sctp_do_sm
> 0.00 +5.5 5.53 ± 4% perf-profile.calltrace.cycles-pp.__sk_mem_reduce_allocated.sctp_wfree.skb_release_head_state.consume_skb.sctp_chunk_put
> 0.00 +5.8 5.80 ± 4% perf-profile.calltrace.cycles-pp.sctp_wfree.skb_release_head_state.consume_skb.sctp_chunk_put.sctp_outq_sack
> 0.00 +5.9 5.87 ± 4% perf-profile.calltrace.cycles-pp.skb_release_head_state.consume_skb.sctp_chunk_put.sctp_outq_sack.sctp_cmd_interpreter
> 0.00 +6.0 5.99 ± 4% perf-profile.calltrace.cycles-pp.__sk_mem_reduce_allocated.skb_release_head_state.kfree_skb_reason.sctp_recvmsg.inet_recvmsg
> 0.00 +6.1 6.13 ± 4% perf-profile.calltrace.cycles-pp.skb_release_head_state.kfree_skb_reason.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
> 9.61 ± 5% -6.2 3.42 ± 5% perf-profile.children.cycles-pp.intel_idle_irq
> 9.16 ± 3% -3.3 5.87 ± 5% perf-profile.children.cycles-pp.sctp_packet_pack
> 8.74 ± 3% -3.1 5.61 ± 4% perf-profile.children.cycles-pp.copy_user_enhanced_fast_string
> 25.01 ± 3% -3.0 21.98 ± 4% perf-profile.children.cycles-pp.sctp_outq_flush
> 8.52 ± 3% -3.0 5.50 ± 5% perf-profile.children.cycles-pp.memcpy_erms
> 6.49 ± 3% -2.5 3.99 ± 4% perf-profile.children.cycles-pp.skb_copy_datagram_iter
> 6.47 ± 3% -2.5 3.98 ± 4% perf-profile.children.cycles-pp.__skb_datagram_iter
> 7.07 ± 3% -2.5 4.60 ± 4% perf-profile.children.cycles-pp.sctp_datamsg_from_user
> 21.61 ± 3% -2.4 19.19 ± 4% perf-profile.children.cycles-pp.sctp_sendmsg_to_asoc
> 13.72 ± 3% -2.4 11.31 ± 5% perf-profile.children.cycles-pp.sctp_primitive_SEND
> 5.63 ± 3% -2.1 3.50 ± 4% perf-profile.children.cycles-pp._copy_to_iter
> 5.46 ± 3% -2.1 3.40 ± 4% perf-profile.children.cycles-pp.copyout
> 22.37 ± 3% -2.0 20.39 ± 4% perf-profile.children.cycles-pp.sctp_packet_transmit
> 3.59 ± 3% -1.6 1.96 ± 4% perf-profile.children.cycles-pp.__alloc_skb
> 2.98 ± 3% -1.5 1.49 ± 5% perf-profile.children.cycles-pp.skb_release_data
> 3.46 ± 4% -1.4 2.06 ± 4% perf-profile.children.cycles-pp.__schedule
> 2.66 ± 3% -1.3 1.35 ± 4% perf-profile.children.cycles-pp.kmalloc_reserve
> 2.62 ± 3% -1.3 1.32 ± 4% perf-profile.children.cycles-pp.__kmalloc_node_track_caller
> 2.50 ± 3% -1.3 1.24 ± 4% perf-profile.children.cycles-pp.kmalloc_large_node
> 3.51 ± 3% -1.2 2.36 ± 4% perf-profile.children.cycles-pp.sctp_user_addto_chunk
> 2.20 ± 3% -1.1 1.07 ± 4% perf-profile.children.cycles-pp.__alloc_pages
> 2.93 ± 3% -1.1 1.82 ± 5% perf-profile.children.cycles-pp.sctp_make_datafrag_empty
> 2.70 ± 3% -1.1 1.62 ± 4% perf-profile.children.cycles-pp._sctp_make_chunk
> 3.17 ± 3% -1.0 2.12 ± 5% perf-profile.children.cycles-pp.sctp_data_ready
> 3.16 ± 3% -1.0 2.13 ± 4% perf-profile.children.cycles-pp._copy_from_iter
> 3.28 ± 3% -1.0 2.26 ± 5% perf-profile.children.cycles-pp.sctp_ulpq_tail_event
> 2.90 ± 3% -1.0 1.88 ± 5% perf-profile.children.cycles-pp.__wake_up_common_lock
> 1.89 ± 4% -1.0 0.88 ± 4% perf-profile.children.cycles-pp.get_page_from_freelist
> 3.02 ± 3% -1.0 2.02 ± 4% perf-profile.children.cycles-pp.copyin
> 2.47 ± 3% -1.0 1.49 ± 7% perf-profile.children.cycles-pp.sctp_outq_flush_data
> 2.70 ± 3% -0.9 1.75 ± 5% perf-profile.children.cycles-pp.__wake_up_common
> 2.53 ± 3% -0.9 1.64 ± 5% perf-profile.children.cycles-pp.autoremove_wake_function
> 2.51 ± 3% -0.9 1.63 ± 4% perf-profile.children.cycles-pp.try_to_wake_up
> 1.83 ± 3% -0.8 1.02 ± 4% perf-profile.children.cycles-pp.sctp_ulpevent_free
> 2.10 ± 3% -0.8 1.32 ± 6% perf-profile.children.cycles-pp.kmem_cache_free
> 1.89 ± 3% -0.8 1.12 ± 5% perf-profile.children.cycles-pp.schedule_timeout
> 1.39 ± 4% -0.8 0.63 ± 5% perf-profile.children.cycles-pp.free_unref_page
> 1.86 ± 3% -0.7 1.11 ± 4% perf-profile.children.cycles-pp.schedule
> 1.25 ± 4% -0.7 0.56 ± 5% perf-profile.children.cycles-pp.rmqueue
> 1.28 ± 4% -0.7 0.60 ± 5% perf-profile.children.cycles-pp._raw_spin_lock_irqsave
> 1.66 ± 4% -0.7 1.00 ± 4% perf-profile.children.cycles-pp.schedule_idle
> 0.98 ± 4% -0.6 0.43 ± 3% perf-profile.children.cycles-pp._raw_spin_lock
> 1.08 ± 4% -0.5 0.54 ± 4% perf-profile.children.cycles-pp.sctp_outq_select_transport
> 1.45 ± 4% -0.5 0.94 ± 4% perf-profile.children.cycles-pp._copy_from_user
> 0.98 ± 4% -0.5 0.48 ± 5% perf-profile.children.cycles-pp.sctp_packet_config
> 1.25 ± 12% -0.5 0.78 ± 5% perf-profile.children.cycles-pp.__rhashtable_lookup
> 1.21 ± 3% -0.5 0.73 ± 4% perf-profile.children.cycles-pp.__check_object_size
> 0.98 ± 3% -0.4 0.53 ± 7% perf-profile.children.cycles-pp.dequeue_task_fair
> 0.64 ± 5% -0.4 0.20 ± 5% perf-profile.children.cycles-pp.free_pcppages_bulk
> 1.18 ± 4% -0.4 0.78 ± 5% perf-profile.children.cycles-pp.ttwu_do_activate
> 0.62 ± 4% -0.4 0.22 ± 4% perf-profile.children.cycles-pp.__free_pages_ok
> 0.90 ± 3% -0.4 0.50 ± 6% perf-profile.children.cycles-pp.dequeue_entity
> 1.14 ± 3% -0.4 0.75 ± 4% perf-profile.children.cycles-pp.enqueue_task_fair
> 0.95 ± 4% -0.4 0.57 ± 6% perf-profile.children.cycles-pp.__slab_free
> 0.76 ± 3% -0.3 0.42 ± 3% perf-profile.children.cycles-pp.__skb_clone
> 1.07 ± 4% -0.3 0.73 ± 5% perf-profile.children.cycles-pp.sendmsg_copy_msghdr
> 0.76 ± 3% -0.3 0.42 ± 5% perf-profile.children.cycles-pp.update_load_avg
> 0.98 ± 4% -0.3 0.65 ± 3% perf-profile.children.cycles-pp.__copy_msghdr_from_user
> 0.81 ± 14% -0.3 0.49 ± 6% perf-profile.children.cycles-pp.sctp_hash_cmp
> 0.88 ± 4% -0.3 0.56 ± 4% perf-profile.children.cycles-pp.pick_next_task_fair
> 0.84 ± 3% -0.3 0.54 ± 5% perf-profile.children.cycles-pp.enqueue_entity
> 0.82 ± 3% -0.3 0.53 ± 5% perf-profile.children.cycles-pp.import_iovec
> 0.81 ± 2% -0.3 0.52 ± 6% perf-profile.children.cycles-pp.kfree
> 0.59 ± 3% -0.3 0.30 ± 4% perf-profile.children.cycles-pp.__copy_skb_header
> 0.78 ± 3% -0.3 0.51 ± 6% perf-profile.children.cycles-pp.__import_iovec
> 0.62 ± 5% -0.3 0.35 ± 5% perf-profile.children.cycles-pp.set_next_entity
> 0.76 ± 12% -0.3 0.50 ± 5% perf-profile.children.cycles-pp.sctp_endpoint_lookup_assoc
> 0.69 ± 4% -0.3 0.43 ± 5% perf-profile.children.cycles-pp.check_heap_object
> 0.77 ± 3% -0.3 0.52 ± 5% perf-profile.children.cycles-pp.kmem_cache_alloc
> 0.74 ± 12% -0.3 0.49 ± 5% perf-profile.children.cycles-pp.sctp_epaddr_lookup_transport
> 0.72 ± 10% -0.2 0.48 ± 6% perf-profile.children.cycles-pp.__sctp_rcv_lookup
> 0.66 ± 4% -0.2 0.42 ± 5% perf-profile.children.cycles-pp.__might_resched
> 0.70 ± 10% -0.2 0.46 ± 7% perf-profile.children.cycles-pp.sctp_addrs_lookup_transport
> 0.65 ± 2% -0.2 0.42 ± 5% perf-profile.children.cycles-pp.iovec_from_user
> 0.50 ± 4% -0.2 0.26 ± 4% perf-profile.children.cycles-pp.simple_copy_to_iter
> 0.46 ± 3% -0.2 0.23 ± 8% perf-profile.children.cycles-pp.sctp_inq_pop
> 0.41 ± 3% -0.2 0.19 ± 6% perf-profile.children.cycles-pp.update_curr
> 0.69 ± 3% -0.2 0.47 ± 5% perf-profile.children.cycles-pp.sctp_chunkify
> 0.55 ± 3% -0.2 0.34 ± 4% perf-profile.children.cycles-pp.__dev_queue_xmit
> 0.53 ± 4% -0.2 0.33 ± 6% perf-profile.children.cycles-pp.__might_fault
> 0.58 ± 4% -0.2 0.39 ± 5% perf-profile.children.cycles-pp.copy_user_short_string
> 0.46 ± 9% -0.2 0.28 ± 12% perf-profile.children.cycles-pp.select_task_rq
> 0.33 ± 4% -0.2 0.16 ± 7% perf-profile.children.cycles-pp.free_unref_page_commit
> 0.28 ± 5% -0.2 0.11 ± 9% perf-profile.children.cycles-pp.dst_release
> 0.32 ± 7% -0.2 0.15 ± 7% perf-profile.children.cycles-pp.ipv4_dst_check
> 0.38 ± 3% -0.2 0.22 ± 8% perf-profile.children.cycles-pp.__mod_node_page_state
> 0.44 ± 8% -0.2 0.28 ± 13% perf-profile.children.cycles-pp.sctp_packet_append_chunk
> 0.49 ± 6% -0.2 0.33 ± 11% perf-profile.children.cycles-pp.sctp_packet_transmit_chunk
> 0.54 ± 4% -0.2 0.39 ± 6% perf-profile.children.cycles-pp.__list_del_entry_valid
> 0.34 ± 4% -0.2 0.19 ± 8% perf-profile.children.cycles-pp.sctp_v4_xmit
> 0.42 ± 3% -0.2 0.27 ± 10% perf-profile.children.cycles-pp.sockfd_lookup_light
> 0.32 ± 4% -0.1 0.17 ± 4% perf-profile.children.cycles-pp.__zone_watermark_ok
> 0.47 ± 3% -0.1 0.32 ± 5% perf-profile.children.cycles-pp.__entry_text_start
> 0.28 ± 2% -0.1 0.13 ± 5% perf-profile.children.cycles-pp.__free_one_page
> 0.28 ± 5% -0.1 0.13 ± 4% perf-profile.children.cycles-pp.prepare_task_switch
> 0.43 ± 5% -0.1 0.29 ± 9% perf-profile.children.cycles-pp.sctp_addto_chunk
> 0.37 ± 4% -0.1 0.23 ± 6% perf-profile.children.cycles-pp.sctp_gen_sack
> 0.41 ± 4% -0.1 0.27 ± 8% perf-profile.children.cycles-pp.aa_sk_perm
> 0.35 ± 3% -0.1 0.22 ± 6% perf-profile.children.cycles-pp.__list_add_valid
> 0.34 ± 4% -0.1 0.20 ± 6% perf-profile.children.cycles-pp.sctp_make_sack
> 0.44 ± 4% -0.1 0.30 ± 6% perf-profile.children.cycles-pp.update_rq_clock
> 0.37 ± 4% -0.1 0.24 ± 7% perf-profile.children.cycles-pp.sock_recvmsg
> 0.34 ± 4% -0.1 0.22 ± 4% perf-profile.children.cycles-pp.sctp_check_transmitted
> 0.35 ± 3% -0.1 0.22 ± 6% perf-profile.children.cycles-pp.memcg_slab_free_hook
> 0.36 ± 4% -0.1 0.23 ± 8% perf-profile.children.cycles-pp.__fdget
> 0.34 ± 10% -0.1 0.22 ± 15% perf-profile.children.cycles-pp.select_task_rq_fair
> 0.44 ± 4% -0.1 0.32 ± 7% perf-profile.children.cycles-pp.kmem_cache_alloc_node
> 0.34 ± 5% -0.1 0.21 ± 7% perf-profile.children.cycles-pp.__virt_addr_valid
> 0.34 ± 3% -0.1 0.22 ± 5% perf-profile.children.cycles-pp.dev_hard_start_xmit
> 0.28 ± 14% -0.1 0.16 ± 6% perf-profile.children.cycles-pp.sctp_transport_hold
> 0.35 ± 2% -0.1 0.23 ± 7% perf-profile.children.cycles-pp.__might_sleep
> 0.38 ± 4% -0.1 0.26 ± 4% perf-profile.children.cycles-pp.sctp_chunk_free
> 0.31 ± 5% -0.1 0.20 ± 5% perf-profile.children.cycles-pp.loopback_xmit
> 0.33 ± 4% -0.1 0.22 ± 4% perf-profile.children.cycles-pp.free_pcp_prepare
> 0.33 ± 5% -0.1 0.22 ± 7% perf-profile.children.cycles-pp.security_socket_recvmsg
> 0.35 ± 2% -0.1 0.24 ± 4% perf-profile.children.cycles-pp.sock_kmalloc
> 0.30 ± 2% -0.1 0.20 ± 8% perf-profile.children.cycles-pp.sctp_association_put
> 0.22 ± 7% -0.1 0.12 ± 6% perf-profile.children.cycles-pp.sctp_assoc_rwnd_increase
> 0.34 ± 4% -0.1 0.23 ± 2% perf-profile.children.cycles-pp.move_addr_to_kernel
> 0.28 ± 4% -0.1 0.17 ± 4% perf-profile.children.cycles-pp.sock_kfree_s
> 0.23 ± 5% -0.1 0.13 ± 7% perf-profile.children.cycles-pp.rmqueue_bulk
> 0.31 ± 9% -0.1 0.21 ± 8% perf-profile.children.cycles-pp.sctp_transport_put
> 0.33 ± 6% -0.1 0.23 ± 12% perf-profile.children.cycles-pp.sctp_sched_fcfs_dequeue
> 0.28 ± 8% -0.1 0.18 ± 20% perf-profile.children.cycles-pp.__sctp_packet_append_chunk
> 0.34 ± 4% -0.1 0.24 ± 8% perf-profile.children.cycles-pp.memset_erms
> 0.19 ± 4% -0.1 0.10 ± 5% perf-profile.children.cycles-pp.__update_load_avg_cfs_rq
> 0.29 ± 4% -0.1 0.19 ± 7% perf-profile.children.cycles-pp.skb_set_owner_w
> 0.27 ± 4% -0.1 0.18 ± 5% perf-profile.children.cycles-pp.security_socket_sendmsg
> 0.24 ± 4% -0.1 0.14 ± 6% perf-profile.children.cycles-pp.__switch_to
> 0.26 ± 4% -0.1 0.17 ± 9% perf-profile.children.cycles-pp.entry_SYSRETQ_unsafe_stack
> 0.27 ± 2% -0.1 0.18 ± 5% perf-profile.children.cycles-pp.send_sctp_stream_1toMany
> 0.23 ± 4% -0.1 0.14 ± 5% perf-profile.children.cycles-pp.tick_nohz_idle_exit
> 0.20 ± 4% -0.1 0.11 ± 8% perf-profile.children.cycles-pp.sctp_outq_flush_ctrl
> 0.24 ± 7% -0.1 0.15 ± 14% perf-profile.children.cycles-pp.select_idle_sibling
> 0.26 ± 4% -0.1 0.17 ± 11% perf-profile.children.cycles-pp.syscall_exit_to_user_mode
> 0.22 ± 5% -0.1 0.14 ± 5% perf-profile.children.cycles-pp.__switch_to_asm
> 0.21 -0.1 0.13 ± 8% perf-profile.children.cycles-pp.accept_connection
> 0.21 -0.1 0.13 ± 8% perf-profile.children.cycles-pp.spawn_child
> 0.21 -0.1 0.13 ± 8% perf-profile.children.cycles-pp.process_requests
> 0.14 ± 5% -0.1 0.06 ± 11% perf-profile.children.cycles-pp.check_stack_object
> 0.20 ± 3% -0.1 0.12 ± 9% perf-profile.children.cycles-pp.sock_wfree
> 0.22 ± 4% -0.1 0.14 ± 3% perf-profile.children.cycles-pp.prepare_to_wait_exclusive
> 0.24 ± 5% -0.1 0.17 ± 9% perf-profile.children.cycles-pp.sctp_sched_dequeue_common
> 0.22 ± 3% -0.1 0.15 ± 2% perf-profile.children.cycles-pp.__kmalloc
> 0.21 ± 3% -0.1 0.13 ± 8% perf-profile.children.cycles-pp.sctp_make_control
> 0.20 ± 2% -0.1 0.12 ± 9% perf-profile.children.cycles-pp.recv_sctp_stream_1toMany
> 0.17 ± 6% -0.1 0.09 ± 5% perf-profile.children.cycles-pp.___perf_sw_event
> 0.14 ± 5% -0.1 0.06 ± 7% perf-profile.children.cycles-pp.sctp_sock_rfree
> 0.20 ± 4% -0.1 0.13 ± 7% perf-profile.children.cycles-pp.skb_clone
> 0.19 ± 30% -0.1 0.12 ± 5% perf-profile.children.cycles-pp.sctp_cmp_addr_exact
> 0.17 ± 4% -0.1 0.10 ± 8% perf-profile.children.cycles-pp.__mod_timer
> 0.20 ± 4% -0.1 0.13 ± 6% perf-profile.children.cycles-pp.available_idle_cpu
> 0.23 ± 7% -0.1 0.16 ± 3% perf-profile.children.cycles-pp.__cond_resched
> 0.21 ± 6% -0.1 0.14 ± 10% perf-profile.children.cycles-pp.sctp_datamsg_put
> 0.16 ± 4% -0.1 0.09 ± 5% perf-profile.children.cycles-pp.check_new_pages
> 0.14 ± 6% -0.1 0.08 ± 12% perf-profile.children.cycles-pp.sctp_inet_skb_msgname
> 0.14 ± 6% -0.1 0.08 ± 9% perf-profile.children.cycles-pp.ip_rcv
> 0.18 ± 3% -0.1 0.11 ± 9% perf-profile.children.cycles-pp.syscall_return_via_sysret
> 0.18 ± 4% -0.1 0.12 ± 9% perf-profile.children.cycles-pp.poll_idle
> 0.20 ± 6% -0.1 0.14 ± 6% perf-profile.children.cycles-pp.kmem_cache_alloc_trace
> 0.19 -0.1 0.13 ± 8% perf-profile.children.cycles-pp.sctp_outq_tail
> 0.18 ± 3% -0.1 0.12 ± 7% perf-profile.children.cycles-pp.reweight_entity
> 0.20 ± 4% -0.1 0.13 ± 11% perf-profile.children.cycles-pp.update_cfs_group
> 0.16 ± 5% -0.1 0.10 ± 4% perf-profile.children.cycles-pp.del_timer
> 0.16 ± 5% -0.1 0.10 ± 3% perf-profile.children.cycles-pp.sctp_sendmsg_parse
> 0.16 ± 5% -0.1 0.10 ± 9% perf-profile.children.cycles-pp.perf_trace_sched_wakeup_template
> 0.18 ± 2% -0.1 0.12 ± 4% perf-profile.children.cycles-pp.__free_pages
> 0.17 ± 4% -0.1 0.11 ± 5% perf-profile.children.cycles-pp.nr_iowait_cpu
> 0.14 ± 2% -0.1 0.09 ± 10% perf-profile.children.cycles-pp.switch_mm_irqs_off
> 0.11 ± 6% -0.1 0.05 ± 8% perf-profile.children.cycles-pp.__build_skb_around
> 0.19 ± 3% -0.1 0.14 ± 8% perf-profile.children.cycles-pp.__update_load_avg_se
> 0.16 ± 5% -0.1 0.11 ± 6% perf-profile.children.cycles-pp.ttwu_do_wakeup
> 0.13 ± 6% -0.1 0.08 ± 13% perf-profile.children.cycles-pp.task_tick_fair
> 0.21 ± 3% -0.0 0.16 ± 6% perf-profile.children.cycles-pp.sctp_association_hold
> 0.15 ± 6% -0.0 0.10 ± 8% perf-profile.children.cycles-pp.check_preempt_curr
> 0.07 ± 6% -0.0 0.02 ± 99% perf-profile.children.cycles-pp._raw_spin_unlock_irqrestore
> 0.10 ± 5% -0.0 0.05 ± 7% perf-profile.children.cycles-pp.finish_task_switch
> 0.14 ± 6% -0.0 0.10 ± 6% perf-profile.children.cycles-pp.__genradix_ptr
> 0.12 ± 4% -0.0 0.08 ± 8% perf-profile.children.cycles-pp.sctp_chunk_abandoned
> 0.07 ± 5% -0.0 0.02 ± 99% perf-profile.children.cycles-pp.ip_local_out
> 0.22 ± 6% -0.0 0.17 ± 8% perf-profile.children.cycles-pp.sched_clock_cpu
> 0.14 ± 5% -0.0 0.09 ± 6% perf-profile.children.cycles-pp.memcg_slab_post_alloc_hook
> 0.12 ± 6% -0.0 0.08 ± 7% perf-profile.children.cycles-pp.__netif_rx
> 0.20 ± 5% -0.0 0.16 ± 8% perf-profile.children.cycles-pp.read_tsc
> 0.12 ± 6% -0.0 0.08 ± 10% perf-profile.children.cycles-pp.lock_timer_base
> 0.19 ± 5% -0.0 0.15 ± 7% perf-profile.children.cycles-pp.native_sched_clock
> 0.13 ± 5% -0.0 0.09 ± 5% perf-profile.children.cycles-pp.pick_next_entity
> 0.12 ± 5% -0.0 0.08 ± 6% perf-profile.children.cycles-pp.netif_rx_internal
> 0.12 ± 7% -0.0 0.08 ± 10% perf-profile.children.cycles-pp.__put_user_nocheck_4
> 0.12 ± 4% -0.0 0.08 ± 6% perf-profile.children.cycles-pp.skb_put
> 0.11 ± 4% -0.0 0.07 ± 5% perf-profile.children.cycles-pp.__put_user_nocheck_8
> 0.12 ± 3% -0.0 0.08 ± 8% perf-profile.children.cycles-pp.resched_curr
> 0.11 ± 8% -0.0 0.07 ± 25% perf-profile.children.cycles-pp.sctp_chunk_assign_ssn
> 0.13 ± 4% -0.0 0.09 ± 6% perf-profile.children.cycles-pp.sctp_datamsg_destroy
> 0.07 ± 7% -0.0 0.03 ±100% perf-profile.children.cycles-pp.ip_rcv_core
> 0.11 ± 3% -0.0 0.07 ± 10% perf-profile.children.cycles-pp.enqueue_to_backlog
> 0.10 ± 5% -0.0 0.06 ± 13% perf-profile.children.cycles-pp.sctp_transport_reset_t3_rtx
> 0.09 ± 6% -0.0 0.05 ± 7% perf-profile.children.cycles-pp.__netif_receive_skb_core
> 0.10 ± 5% -0.0 0.06 ± 6% perf-profile.children.cycles-pp.validate_xmit_skb
> 0.11 ± 6% -0.0 0.07 ± 12% perf-profile.children.cycles-pp.syscall_exit_to_user_mode_prepare
> 0.06 ± 6% -0.0 0.02 ± 99% perf-profile.children.cycles-pp.entry_SYSCALL_64_safe_stack
> 0.10 ± 8% -0.0 0.06 ± 9% perf-profile.children.cycles-pp.sctp_ulpq_order
> 0.09 ± 6% -0.0 0.06 ± 8% perf-profile.children.cycles-pp.os_xsave
> 0.12 ± 5% -0.0 0.08 ± 8% perf-profile.children.cycles-pp.__check_heap_object
> 0.10 ± 5% -0.0 0.07 ± 10% perf-profile.children.cycles-pp.rcu_all_qs
> 0.08 ± 5% -0.0 0.06 ± 11% perf-profile.children.cycles-pp.tick_nohz_idle_enter
> 0.08 ± 5% -0.0 0.06 ± 8% perf-profile.children.cycles-pp.perf_tp_event
> 0.08 -0.0 0.06 ± 9% perf-profile.children.cycles-pp.__mod_lruvec_page_state
> 0.31 ± 3% -0.0 0.28 ± 4% perf-profile.children.cycles-pp.sctp_ulpevent_receive_data
> 0.10 ± 10% -0.0 0.08 ± 10% perf-profile.children.cycles-pp.rcu_idle_exit
> 0.08 ± 6% -0.0 0.05 ± 8% perf-profile.children.cycles-pp.sctp_tsnmap_check
> 0.08 ± 5% -0.0 0.06 ± 14% perf-profile.children.cycles-pp.sctp_ulpevent_init
> 0.07 ± 8% -0.0 0.05 perf-profile.children.cycles-pp.__wrgsbase_inactive
> 0.10 ± 5% +0.0 0.13 ± 10% perf-profile.children.cycles-pp.sctp_sockaddr_af
> 0.00 +0.1 0.05 perf-profile.children.cycles-pp.perf_mux_hrtimer_handler
> 0.19 ± 3% +0.1 0.24 ± 11% perf-profile.children.cycles-pp.scheduler_tick
> 0.00 +0.1 0.06 ± 7% perf-profile.children.cycles-pp.lapic_next_deadline
> 0.00 +0.1 0.07 ± 32% perf-profile.children.cycles-pp.arch_cpu_idle_enter
> 0.00 +0.1 0.07 ± 16% perf-profile.children.cycles-pp.native_irq_return_iret
> 0.00 +0.1 0.08 ± 22% perf-profile.children.cycles-pp.update_blocked_averages
> 0.00 +0.1 0.08 ± 22% perf-profile.children.cycles-pp.run_rebalance_domains
> 0.24 ± 3% +0.1 0.32 ± 9% perf-profile.children.cycles-pp.update_process_times
> 0.00 +0.1 0.08 ± 22% perf-profile.children.cycles-pp.update_sg_lb_stats
> 0.24 ± 4% +0.1 0.32 ± 9% perf-profile.children.cycles-pp.tick_sched_handle
> 0.00 +0.1 0.08 ± 41% perf-profile.children.cycles-pp._raw_spin_trylock
> 0.00 +0.1 0.09 ± 20% perf-profile.children.cycles-pp.update_sd_lb_stats
> 0.00 +0.1 0.09 ± 19% perf-profile.children.cycles-pp.find_busiest_group
> 0.04 ± 71% +0.1 0.14 ± 3% perf-profile.children.cycles-pp.sctp_get_af_specific
> 0.01 ±223% +0.1 0.11 ± 18% perf-profile.children.cycles-pp.load_balance
> 0.00 +0.1 0.11 ± 11% perf-profile.children.cycles-pp.tick_nohz_irq_exit
> 0.02 ±141% +0.1 0.14 ± 46% perf-profile.children.cycles-pp.rebalance_domains
> 0.45 ± 5% +0.1 0.58 ± 10% perf-profile.children.cycles-pp.ktime_get
> 0.00 +0.1 0.14 ± 17% perf-profile.children.cycles-pp.cgroup_rstat_updated
> 0.26 ± 4% +0.1 0.41 ± 12% perf-profile.children.cycles-pp.tick_sched_timer
> 0.15 ± 10% +0.2 0.32 ± 16% perf-profile.children.cycles-pp.clockevents_program_event
> 0.30 ± 4% +0.2 0.50 ± 10% perf-profile.children.cycles-pp.__hrtimer_run_queues
> 0.25 ± 4% +0.2 0.45 ± 20% perf-profile.children.cycles-pp.tick_nohz_get_sleep_length
> 0.00 +0.2 0.20 ± 16% perf-profile.children.cycles-pp.timekeeping_max_deferment
> 0.50 ± 2% +0.2 0.71 ± 15% perf-profile.children.cycles-pp.menu_select
> 0.16 ± 5% +0.2 0.38 ± 22% perf-profile.children.cycles-pp.tick_nohz_next_event
> 0.08 ± 8% +0.2 0.30 ± 23% perf-profile.children.cycles-pp.__irq_exit_rcu
> 0.00 +0.3 0.33 ± 12% perf-profile.children.cycles-pp.propagate_protected_usage
> 0.48 ± 5% +0.4 0.89 ± 11% perf-profile.children.cycles-pp.hrtimer_interrupt
> 0.48 ± 6% +0.4 0.90 ± 10% perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
> 0.00 +0.6 0.55 ± 3% perf-profile.children.cycles-pp.mem_cgroup_uncharge_skmem
> 0.61 ± 5% +0.7 1.35 ± 11% perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
> 1.00 ± 2% +0.8 1.82 ± 10% perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
> 0.00 +1.0 1.03 ± 4% perf-profile.children.cycles-pp.__mod_memcg_state
> 1.73 ± 2% +1.4 3.14 ± 4% perf-profile.children.cycles-pp.sctp_ulpevent_make_rcvmsg
> 19.09 ± 3% +2.0 21.11 ± 4% perf-profile.children.cycles-pp.recvmsg
> 11.51 ± 3% +2.2 13.66 ± 4% perf-profile.children.cycles-pp.__ip_queue_xmit
> 11.26 ± 3% +2.3 13.53 ± 4% perf-profile.children.cycles-pp.ip_finish_output2
> 0.00 +2.3 2.27 ± 5% perf-profile.children.cycles-pp.page_counter_uncharge
> 0.00 +2.3 2.32 ± 4% perf-profile.children.cycles-pp.drain_stock
> 0.00 +2.4 2.40 ± 5% perf-profile.children.cycles-pp.page_counter_try_charge
> 10.87 ± 3% +2.4 13.28 ± 4% perf-profile.children.cycles-pp.__local_bh_enable_ip
> 10.76 ± 3% +2.4 13.20 ± 4% perf-profile.children.cycles-pp.do_softirq
> 17.84 ± 3% +2.5 20.32 ± 4% perf-profile.children.cycles-pp.__sys_recvmsg
> 3.60 ± 3% +2.5 6.12 ± 5% perf-profile.children.cycles-pp.sctp_skb_recv_datagram
> 10.53 ± 3% +2.5 13.07 ± 4% perf-profile.children.cycles-pp.net_rx_action
> 0.00 +2.5 2.54 ± 4% perf-profile.children.cycles-pp.refill_stock
> 10.43 ± 3% +2.6 13.01 ± 4% perf-profile.children.cycles-pp.__napi_poll
> 17.57 ± 3% +2.6 20.15 ± 4% perf-profile.children.cycles-pp.___sys_recvmsg
> 10.41 ± 3% +2.6 12.99 ± 4% perf-profile.children.cycles-pp.process_backlog
> 10.77 ± 3% +2.7 13.43 ± 4% perf-profile.children.cycles-pp.__softirqentry_text_start
> 10.24 ± 3% +2.7 12.90 ± 4% perf-profile.children.cycles-pp.__netif_receive_skb_one_core
> 9.97 ± 3% +2.8 12.76 ± 4% perf-profile.children.cycles-pp.ip_local_deliver_finish
> 9.95 ± 3% +2.8 12.74 ± 4% perf-profile.children.cycles-pp.ip_protocol_deliver_rcu
> 9.89 ± 3% +2.8 12.69 ± 4% perf-profile.children.cycles-pp.sctp_rcv
> 1.71 ± 4% +2.9 4.57 ± 4% perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
> 16.66 ± 3% +2.9 19.59 ± 4% perf-profile.children.cycles-pp.____sys_recvmsg
> 16.01 ± 3% +3.2 19.21 ± 4% perf-profile.children.cycles-pp.sctp_recvmsg
> 15.89 ± 3% +3.2 19.12 ± 4% perf-profile.children.cycles-pp.inet_recvmsg
> 1.49 ± 3% +3.4 4.86 ± 4% perf-profile.children.cycles-pp._raw_spin_lock_bh
> 1.59 ± 3% +3.4 4.96 ± 4% perf-profile.children.cycles-pp.lock_sock_nested
> 15.60 ± 3% +3.6 19.22 ± 4% perf-profile.children.cycles-pp.release_sock
> 0.00 +3.7 3.69 ± 4% perf-profile.children.cycles-pp.try_charge_memcg
> 15.33 ± 3% +3.7 19.04 ± 4% perf-profile.children.cycles-pp.__release_sock
> 15.27 ± 3% +3.7 19.00 ± 4% perf-profile.children.cycles-pp.sctp_backlog_rcv
> 0.62 ± 5% +3.8 4.45 ± 4% perf-profile.children.cycles-pp.sctp_sf_eat_data_6_2
> 0.48 ± 5% +3.9 4.36 ± 5% perf-profile.children.cycles-pp.sctp_eat_data
> 0.00 +4.4 4.36 ± 4% perf-profile.children.cycles-pp.mem_cgroup_charge_skmem
> 3.46 ± 3% +4.4 7.91 ± 4% perf-profile.children.cycles-pp.sctp_chunk_put
> 20.07 ± 3% +4.6 24.66 ± 4% perf-profile.children.cycles-pp.sctp_assoc_bh_rcv
> 0.00 +4.6 4.60 ± 4% perf-profile.children.cycles-pp.__sk_mem_raise_allocated
> 0.00 +4.6 4.63 ± 4% perf-profile.children.cycles-pp.__sk_mem_schedule
> 24.02 ± 12% +4.7 28.76 ± 11% perf-profile.children.cycles-pp.intel_idle
> 2.97 ± 3% +4.8 7.75 ± 4% perf-profile.children.cycles-pp.sctp_outq_sack
> 2.11 ± 4% +4.8 6.90 ± 4% perf-profile.children.cycles-pp.kfree_skb_reason
> 2.04 ± 3% +4.9 6.98 ± 4% perf-profile.children.cycles-pp.consume_skb
> 0.37 ± 3% +5.6 5.94 ± 4% perf-profile.children.cycles-pp.sctp_wfree
> 0.95 ± 3% +11.3 12.23 ± 4% perf-profile.children.cycles-pp.skb_release_head_state
> 0.00 +11.7 11.67 ± 4% perf-profile.children.cycles-pp.__sk_mem_reduce_allocated
> 8.68 ± 3% -3.1 5.58 ± 4% perf-profile.self.cycles-pp.copy_user_enhanced_fast_string
> 8.45 ± 3% -3.0 5.46 ± 5% perf-profile.self.cycles-pp.memcpy_erms
> 0.94 ± 4% -0.4 0.57 ± 6% perf-profile.self.cycles-pp.__slab_free
> 0.71 ± 3% -0.3 0.42 ± 2% perf-profile.self.cycles-pp.sctp_chunk_put
> 0.58 ± 3% -0.3 0.30 ± 3% perf-profile.self.cycles-pp.__copy_skb_header
> 0.87 ± 3% -0.3 0.59 ± 6% perf-profile.self.cycles-pp.kmem_cache_free
> 0.58 ± 5% -0.3 0.30 ± 6% perf-profile.self.cycles-pp.sctp_packet_config
> 0.82 ± 3% -0.3 0.55 ± 5% perf-profile.self.cycles-pp._raw_spin_lock_irqsave
> 0.65 ± 4% -0.2 0.41 ± 5% perf-profile.self.cycles-pp.__might_resched
> 0.41 ± 4% -0.2 0.17 ± 8% perf-profile.self.cycles-pp.rmqueue
> 0.63 ± 4% -0.2 0.40 ± 3% perf-profile.self.cycles-pp._raw_spin_lock
> 0.59 ± 4% -0.2 0.36 ± 7% perf-profile.self.cycles-pp.sctp_recvmsg
> 0.39 ± 4% -0.2 0.18 ± 5% perf-profile.self.cycles-pp.sctp_sendmsg_to_asoc
> 0.58 ± 3% -0.2 0.36 ± 3% perf-profile.self.cycles-pp._raw_spin_lock_bh
> 0.43 ± 4% -0.2 0.24 ± 9% perf-profile.self.cycles-pp.sctp_packet_pack
> 0.39 ± 4% -0.2 0.20 ± 2% perf-profile.self.cycles-pp.update_load_avg
> 0.56 ± 4% -0.2 0.38 ± 5% perf-profile.self.cycles-pp.copy_user_short_string
> 0.27 ± 5% -0.2 0.10 ± 7% perf-profile.self.cycles-pp.dst_release
> 0.32 ± 3% -0.2 0.16 ± 7% perf-profile.self.cycles-pp.get_page_from_freelist
> 0.31 ± 7% -0.2 0.15 ± 6% perf-profile.self.cycles-pp.ipv4_dst_check
> 0.30 ± 4% -0.2 0.14 ± 7% perf-profile.self.cycles-pp.free_unref_page_commit
> 0.38 ± 3% -0.2 0.22 ± 8% perf-profile.self.cycles-pp.__mod_node_page_state
> 0.54 ± 3% -0.2 0.38 ± 5% perf-profile.self.cycles-pp.__schedule
> 0.47 ± 2% -0.2 0.31 ± 5% perf-profile.self.cycles-pp.kfree
> 0.34 ± 4% -0.2 0.18 ± 7% perf-profile.self.cycles-pp.sctp_v4_xmit
> 0.23 ± 4% -0.2 0.07 ± 10% perf-profile.self.cycles-pp.update_curr
> 0.32 ± 12% -0.2 0.17 ± 9% perf-profile.self.cycles-pp.__rhashtable_lookup
> 0.51 ± 4% -0.2 0.36 ± 6% perf-profile.self.cycles-pp.__list_del_entry_valid
> 0.32 ± 3% -0.2 0.17 ± 4% perf-profile.self.cycles-pp.__zone_watermark_ok
> 0.42 ± 3% -0.1 0.27 ± 5% perf-profile.self.cycles-pp.kmem_cache_alloc
> 0.40 ± 5% -0.1 0.26 ± 7% perf-profile.self.cycles-pp.sctp_outq_flush_data
> 0.28 ± 4% -0.1 0.15 ± 6% perf-profile.self.cycles-pp.__check_object_size
> 0.35 ± 4% -0.1 0.22 ± 9% perf-profile.self.cycles-pp.__fdget
> 0.42 ± 4% -0.1 0.29 ± 6% perf-profile.self.cycles-pp.sctp_datamsg_from_user
> 0.36 ± 5% -0.1 0.24 ± 5% perf-profile.self.cycles-pp.set_next_entity
> 0.34 ± 3% -0.1 0.21 ± 3% perf-profile.self.cycles-pp.check_heap_object
> 0.34 ± 3% -0.1 0.22 ± 7% perf-profile.self.cycles-pp.memcg_slab_free_hook
> 0.32 ± 2% -0.1 0.20 ± 5% perf-profile.self.cycles-pp.__list_add_valid
> 0.32 ± 5% -0.1 0.20 ± 8% perf-profile.self.cycles-pp.__virt_addr_valid
> 0.34 ± 5% -0.1 0.22 ± 4% perf-profile.self.cycles-pp.__skb_datagram_iter
> 0.34 ± 4% -0.1 0.22 ± 7% perf-profile.self.cycles-pp.update_rq_clock
> 0.37 ± 5% -0.1 0.25 ± 4% perf-profile.self.cycles-pp.sctp_sendmsg
> 0.27 ± 13% -0.1 0.16 ± 6% perf-profile.self.cycles-pp.sctp_transport_hold
> 0.35 ± 3% -0.1 0.24 ± 6% perf-profile.self.cycles-pp.skb_release_data
> 0.32 ± 4% -0.1 0.20 ± 7% perf-profile.self.cycles-pp.__alloc_skb
> 0.22 ± 7% -0.1 0.10 ± 7% perf-profile.self.cycles-pp.sctp_assoc_rwnd_increase
> 0.33 ± 5% -0.1 0.22 ± 4% perf-profile.self.cycles-pp.free_pcp_prepare
> 0.31 ± 4% -0.1 0.20 ± 7% perf-profile.self.cycles-pp.enqueue_task_fair
> 0.36 ± 3% -0.1 0.26 ± 3% perf-profile.self.cycles-pp.enqueue_entity
> 0.30 -0.1 0.20 ± 8% perf-profile.self.cycles-pp.sctp_association_put
> 0.24 ± 4% -0.1 0.14 ± 7% perf-profile.self.cycles-pp.__alloc_pages
> 0.30 ± 9% -0.1 0.20 ± 6% perf-profile.self.cycles-pp.sctp_transport_put
> 0.31 ± 3% -0.1 0.21 ± 6% perf-profile.self.cycles-pp.sctp_skb_recv_datagram
> 0.30 ± 4% -0.1 0.20 ± 8% perf-profile.self.cycles-pp.aa_sk_perm
> 0.33 ± 4% -0.1 0.23 ± 6% perf-profile.self.cycles-pp.entry_SYSCALL_64_after_hwframe
> 0.30 ± 3% -0.1 0.20 ± 7% perf-profile.self.cycles-pp.__might_sleep
> 0.18 ± 4% -0.1 0.09 ± 10% perf-profile.self.cycles-pp.____sys_recvmsg
> 0.19 ± 5% -0.1 0.09 ± 11% perf-profile.self.cycles-pp.sctp_ulpevent_free
> 0.32 ± 3% -0.1 0.23 ± 8% perf-profile.self.cycles-pp.memset_erms
> 0.30 ± 6% -0.1 0.21 ± 7% perf-profile.self.cycles-pp.kmem_cache_alloc_node
> 0.28 ± 3% -0.1 0.19 ± 6% perf-profile.self.cycles-pp.skb_set_owner_w
> 0.23 ± 6% -0.1 0.14 ± 7% perf-profile.self.cycles-pp.__switch_to
> 0.18 ± 4% -0.1 0.10 ± 5% perf-profile.self.cycles-pp.__update_load_avg_cfs_rq
> 0.26 ± 4% -0.1 0.17 ± 7% perf-profile.self.cycles-pp.sctp_cmd_interpreter
> 0.25 ± 4% -0.1 0.17 ± 7% perf-profile.self.cycles-pp.entry_SYSRETQ_unsafe_stack
> 0.26 -0.1 0.18 ± 5% perf-profile.self.cycles-pp.send_sctp_stream_1toMany
> 0.17 ± 32% -0.1 0.09 ± 6% perf-profile.self.cycles-pp.sctp_cmp_addr_exact
> 0.22 ± 5% -0.1 0.14 ± 7% perf-profile.self.cycles-pp.__switch_to_asm
> 0.20 ± 2% -0.1 0.12 ± 10% perf-profile.self.cycles-pp.recv_sctp_stream_1toMany
> 0.19 ± 3% -0.1 0.11 ± 9% perf-profile.self.cycles-pp.sock_wfree
> 0.10 ± 5% -0.1 0.02 ± 99% perf-profile.self.cycles-pp.process_backlog
> 0.10 ± 6% -0.1 0.02 ± 99% perf-profile.self.cycles-pp.__free_one_page
> 0.20 ± 4% -0.1 0.13 ± 5% perf-profile.self.cycles-pp.recvmsg
> 0.22 ± 3% -0.1 0.14 ± 7% perf-profile.self.cycles-pp.consume_skb
> 0.18 ± 3% -0.1 0.11 ± 10% perf-profile.self.cycles-pp._copy_from_user
> 0.15 ± 7% -0.1 0.08 ± 4% perf-profile.self.cycles-pp.___perf_sw_event
> 0.15 ± 4% -0.1 0.08 ± 5% perf-profile.self.cycles-pp.check_new_pages
> 0.20 ± 5% -0.1 0.13 ± 6% perf-profile.self.cycles-pp.available_idle_cpu
> 0.22 ± 3% -0.1 0.15 ± 5% perf-profile.self.cycles-pp.__entry_text_start
> 0.18 ± 2% -0.1 0.11 ± 7% perf-profile.self.cycles-pp.sctp_assoc_bh_rcv
> 0.13 ± 7% -0.1 0.06 ± 7% perf-profile.self.cycles-pp.sctp_sock_rfree
> 0.17 ± 4% -0.1 0.11 ± 10% perf-profile.self.cycles-pp.syscall_return_via_sysret
> 0.13 ± 5% -0.1 0.07 ± 13% perf-profile.self.cycles-pp.__free_pages_ok
> 0.20 ± 5% -0.1 0.14 ± 9% perf-profile.self.cycles-pp.sctp_datamsg_put
> 0.19 ± 3% -0.1 0.12 ± 12% perf-profile.self.cycles-pp.sendmsg
> 0.18 ± 4% -0.1 0.12 ± 7% perf-profile.self.cycles-pp.reweight_entity
> 0.14 ± 7% -0.1 0.07 ± 10% perf-profile.self.cycles-pp.sctp_inet_skb_msgname
> 0.23 ± 3% -0.1 0.17 ± 9% perf-profile.self.cycles-pp.sctp_wfree
> 0.19 ± 3% -0.1 0.13 ± 11% perf-profile.self.cycles-pp.update_cfs_group
> 0.17 ± 4% -0.1 0.11 ± 3% perf-profile.self.cycles-pp.nr_iowait_cpu
> 0.17 ± 7% -0.1 0.11 ± 9% perf-profile.self.cycles-pp.sctp_packet_transmit
> 0.08 ± 18% -0.1 0.02 ± 99% perf-profile.self.cycles-pp.sctp_ulpq_tail_data
> 0.17 ± 7% -0.1 0.11 ± 9% perf-profile.self.cycles-pp.sctp_packet_append_chunk
> 0.19 ± 4% -0.1 0.13 ± 8% perf-profile.self.cycles-pp.sctp_do_sm
> 0.18 ± 2% -0.1 0.12 ± 4% perf-profile.self.cycles-pp.__free_pages
> 0.17 ± 5% -0.1 0.12 ± 6% perf-profile.self.cycles-pp.__wake_up_common
> 0.16 ± 6% -0.1 0.10 ± 3% perf-profile.self.cycles-pp.sctp_sendmsg_parse
> 0.15 ± 6% -0.1 0.09 ± 13% perf-profile.self.cycles-pp.try_to_wake_up
> 0.10 ± 3% -0.1 0.04 ± 45% perf-profile.self.cycles-pp.__build_skb_around
> 0.15 ± 6% -0.1 0.09 ± 11% perf-profile.self.cycles-pp.sctp_check_transmitted
> 0.08 ± 14% -0.1 0.03 ±101% perf-profile.self.cycles-pp.select_task_rq_fair
> 0.12 ± 6% -0.1 0.06 ± 14% perf-profile.self.cycles-pp.dequeue_entity
> 0.19 ± 4% -0.1 0.14 ± 10% perf-profile.self.cycles-pp.sctp_hash_cmp
> 0.14 ± 4% -0.1 0.09 ± 10% perf-profile.self.cycles-pp.switch_mm_irqs_off
> 0.16 ± 4% -0.1 0.10 ± 7% perf-profile.self.cycles-pp.poll_idle
> 0.22 ± 8% -0.1 0.17 ± 10% perf-profile.self.cycles-pp.sctp_rcv
> 0.11 ± 4% -0.1 0.06 ± 11% perf-profile.self.cycles-pp.__ip_queue_xmit
> 0.12 ± 10% -0.1 0.08 ± 24% perf-profile.self.cycles-pp.__sctp_packet_append_chunk
> 0.13 ± 2% -0.0 0.08 ± 5% perf-profile.self.cycles-pp.sctp_inq_pop
> 0.18 ± 3% -0.0 0.13 ± 11% perf-profile.self.cycles-pp.__update_load_avg_se
> 0.14 ± 5% -0.0 0.09 ± 5% perf-profile.self.cycles-pp.do_syscall_64
> 0.11 ± 6% -0.0 0.06 ± 11% perf-profile.self.cycles-pp._copy_to_iter
> 0.09 ± 11% -0.0 0.04 ± 76% perf-profile.self.cycles-pp.sctp_chunk_assign_ssn
> 0.14 ± 3% -0.0 0.10 ± 5% perf-profile.self.cycles-pp.____sys_sendmsg
> 0.12 ± 4% -0.0 0.07 ± 6% perf-profile.self.cycles-pp.___sys_recvmsg
> 0.20 ± 2% -0.0 0.15 ± 8% perf-profile.self.cycles-pp.sctp_association_hold
> 0.17 ± 3% -0.0 0.12 ± 7% perf-profile.self.cycles-pp.__skb_clone
> 0.14 ± 3% -0.0 0.09 ± 7% perf-profile.self.cycles-pp.__copy_msghdr_from_user
> 0.13 ± 5% -0.0 0.08 ± 8% perf-profile.self.cycles-pp.sctp_sf_eat_data_6_2
> 0.13 ± 2% -0.0 0.09 ± 13% perf-profile.self.cycles-pp.sctp_outq_tail
> 0.20 ± 4% -0.0 0.16 ± 9% perf-profile.self.cycles-pp.read_tsc
> 0.13 ± 5% -0.0 0.08 ± 5% perf-profile.self.cycles-pp.__import_iovec
> 0.13 ± 6% -0.0 0.08 ± 5% perf-profile.self.cycles-pp.__genradix_ptr
> 0.11 ± 8% -0.0 0.07 ± 14% perf-profile.self.cycles-pp.select_task_rq
> 0.12 ± 3% -0.0 0.08 ± 10% perf-profile.self.cycles-pp.resched_curr
> 0.13 ± 5% -0.0 0.09 ± 5% perf-profile.self.cycles-pp.pick_next_entity
> 0.12 ± 6% -0.0 0.08 ± 12% perf-profile.self.cycles-pp.__put_user_nocheck_4
> 0.12 ± 6% -0.0 0.08 ± 6% perf-profile.self.cycles-pp.sctp_chunk_abandoned
> 0.10 ± 7% -0.0 0.06 ± 11% perf-profile.self.cycles-pp.syscall_exit_to_user_mode_prepare
> 0.07 ± 7% -0.0 0.02 ± 99% perf-profile.self.cycles-pp.ip_rcv_core
> 0.13 ± 7% -0.0 0.09 ± 7% perf-profile.self.cycles-pp.sctp_sched_dequeue_common
> 0.13 ± 7% -0.0 0.09 ± 6% perf-profile.self.cycles-pp.ip_finish_output2
> 0.12 ± 6% -0.0 0.08 ± 10% perf-profile.self.cycles-pp.memcg_slab_post_alloc_hook
> 0.08 ± 6% -0.0 0.04 ± 71% perf-profile.self.cycles-pp._sctp_make_chunk
> 0.10 ± 4% -0.0 0.06 ± 11% perf-profile.self.cycles-pp.net_rx_action
> 0.10 ± 5% -0.0 0.06 ± 8% perf-profile.self.cycles-pp.sctp_addto_chunk
> 0.09 ± 6% -0.0 0.05 ± 7% perf-profile.self.cycles-pp.__netif_receive_skb_core
> 0.12 ± 5% -0.0 0.08 ± 5% perf-profile.self.cycles-pp.__kmalloc
> 0.11 ± 5% -0.0 0.07 ± 5% perf-profile.self.cycles-pp.__put_user_nocheck_8
> 0.18 ± 6% -0.0 0.15 ± 7% perf-profile.self.cycles-pp.native_sched_clock
> 0.09 ± 5% -0.0 0.06 ± 8% perf-profile.self.cycles-pp.check_stack_object
> 0.07 -0.0 0.03 ± 70% perf-profile.self.cycles-pp.__mod_lruvec_page_state
> 0.16 ± 4% -0.0 0.12 ± 10% perf-profile.self.cycles-pp.sctp_outq_flush
> 0.10 ± 3% -0.0 0.06 ± 7% perf-profile.self.cycles-pp.skb_put
> 0.06 ± 6% -0.0 0.02 ± 99% perf-profile.self.cycles-pp.entry_SYSCALL_64_safe_stack
> 0.10 ± 4% -0.0 0.07 ± 15% perf-profile.self.cycles-pp.__might_fault
> 0.12 ± 5% -0.0 0.08 ± 5% perf-profile.self.cycles-pp.kmem_cache_alloc_trace
> 0.07 ± 8% -0.0 0.04 ± 71% perf-profile.self.cycles-pp.sockfd_lookup_light
> 0.10 ± 7% -0.0 0.07 ± 10% perf-profile.self.cycles-pp.__cond_resched
> 0.09 ± 6% -0.0 0.06 ± 8% perf-profile.self.cycles-pp.os_xsave
> 0.09 ± 6% -0.0 0.06 ± 8% perf-profile.self.cycles-pp.free_unref_page
> 0.09 ± 7% -0.0 0.06 ± 6% perf-profile.self.cycles-pp.security_socket_recvmsg
> 0.11 ± 4% -0.0 0.08 ± 7% perf-profile.self.cycles-pp.__local_bh_enable_ip
> 0.12 ± 4% -0.0 0.09 ± 12% perf-profile.self.cycles-pp.sock_kmalloc
> 0.09 ± 6% -0.0 0.06 ± 6% perf-profile.self.cycles-pp.sctp_user_addto_chunk
> 0.07 ± 6% -0.0 0.04 ± 44% perf-profile.self.cycles-pp.rcu_all_qs
> 0.10 ± 5% -0.0 0.06 ± 11% perf-profile.self.cycles-pp.sctp_outq_sack
> 0.09 ± 4% -0.0 0.06 ± 6% perf-profile.self.cycles-pp.sctp_outq_select_transport
> 0.10 ± 3% -0.0 0.07 ± 9% perf-profile.self.cycles-pp.__kmalloc_node_track_caller
> 0.08 ± 5% -0.0 0.05 ± 46% perf-profile.self.cycles-pp.sctp_ulpevent_init
> 0.12 ± 4% -0.0 0.08 ± 8% perf-profile.self.cycles-pp.__check_heap_object
> 0.07 ± 11% -0.0 0.04 ± 44% perf-profile.self.cycles-pp.loopback_xmit
> 0.14 ± 2% -0.0 0.12 ± 6% perf-profile.self.cycles-pp.do_idle
> 0.11 ± 4% -0.0 0.08 ± 12% perf-profile.self.cycles-pp.cpuidle_idle_call
> 0.09 -0.0 0.06 ± 11% perf-profile.self.cycles-pp._copy_from_iter
> 0.08 ± 5% -0.0 0.06 ± 13% perf-profile.self.cycles-pp.kmalloc_large_node
> 0.08 ± 5% -0.0 0.06 ± 9% perf-profile.self.cycles-pp.sctp_chunkify
> 0.09 ± 4% -0.0 0.06 ± 7% perf-profile.self.cycles-pp.iovec_from_user
> 0.14 ± 4% -0.0 0.12 ± 11% perf-profile.self.cycles-pp.sctp_ulpevent_make_rcvmsg
> 0.07 ± 6% -0.0 0.05 ± 7% perf-profile.self.cycles-pp.sctp_tsnmap_check
> 0.08 ± 6% -0.0 0.06 ± 9% perf-profile.self.cycles-pp.sock_kfree_s
> 0.08 ± 5% -0.0 0.06 ± 11% perf-profile.self.cycles-pp.skb_release_head_state
> 0.07 ± 8% -0.0 0.05 perf-profile.self.cycles-pp.__wrgsbase_inactive
> 0.09 ± 8% -0.0 0.07 ± 13% perf-profile.self.cycles-pp.__softirqentry_text_start
> 0.07 ± 5% -0.0 0.05 ± 8% perf-profile.self.cycles-pp.__sys_recvmsg
> 0.18 ± 4% +0.0 0.22 ± 12% perf-profile.self.cycles-pp.menu_select
> 0.00 +0.1 0.06 ± 19% perf-profile.self.cycles-pp.update_sg_lb_stats
> 0.00 +0.1 0.06 ± 7% perf-profile.self.cycles-pp.lapic_next_deadline
> 0.00 +0.1 0.07 ± 16% perf-profile.self.cycles-pp.native_irq_return_iret
> 0.09 ± 4% +0.1 0.16 ± 5% perf-profile.self.cycles-pp.cpuidle_enter_state
> 0.00 +0.1 0.08 ± 41% perf-profile.self.cycles-pp._raw_spin_trylock
> 0.00 +0.1 0.09 ± 54% perf-profile.self.cycles-pp.tick_nohz_next_event
> 0.01 ±223% +0.1 0.13 ± 5% perf-profile.self.cycles-pp.sctp_get_af_specific
> 0.00 +0.1 0.13 ± 6% perf-profile.self.cycles-pp.mem_cgroup_charge_skmem
> 0.00 +0.1 0.14 ± 19% perf-profile.self.cycles-pp.cgroup_rstat_updated
> 0.25 ± 6% +0.2 0.43 ± 14% perf-profile.self.cycles-pp.ktime_get
> 0.00 +0.2 0.20 ± 16% perf-profile.self.cycles-pp.timekeeping_max_deferment
> 0.00 +0.2 0.22 ± 6% perf-profile.self.cycles-pp.refill_stock
> 0.00 +0.3 0.25 ± 9% perf-profile.self.cycles-pp.__sk_mem_raise_allocated
> 0.00 +0.3 0.32 ± 12% perf-profile.self.cycles-pp.propagate_protected_usage
> 0.00 +0.9 0.89 ± 4% perf-profile.self.cycles-pp.__mod_memcg_state
> 0.00 +1.3 1.27 ± 5% perf-profile.self.cycles-pp.try_charge_memcg
> 0.00 +2.1 2.10 ± 5% perf-profile.self.cycles-pp.page_counter_uncharge
> 0.00 +2.2 2.23 ± 5% perf-profile.self.cycles-pp.page_counter_try_charge
> 1.70 ± 4% +2.8 4.54 ± 4% perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
> 0.29 ± 5% +4.0 4.24 ± 5% perf-profile.self.cycles-pp.sctp_eat_data
> 0.00 +8.6 8.55 ± 4% perf-profile.self.cycles-pp.__sk_mem_reduce_allocated
>
>
>
>
>
> Disclaimer:
> Results have been estimated based on internal Intel analysis and are provided
> for informational purposes only. Any difference in system hardware or software
> design or configuration may affect actual performance.
>
>

2022-06-23 05:04:18

by Xin Long

[permalink] [raw]
Subject: Re: [net] 4890b686f4: netperf.Throughput_Mbps -69.4% regression

Yes, I'm working on it. I couldn't see the regression in my env with
the 'reproduce' script attached.
I will try with lkp tomorrow.

Thanks.

On Wed, Jun 22, 2022 at 8:29 PM Jakub Kicinski <[email protected]> wrote:
>
> Could someone working on SCTP double check this is a real regression?
> Feels like the regression reports are flowing at such rate its hard
> to keep up.
>
> On Sun, 19 Jun 2022 23:04:56 +0800 kernel test robot wrote:
> > Greeting,
> >
> > FYI, we noticed a -69.4% regression of netperf.Throughput_Mbps due to commit:
> >
> >
> > commit: 4890b686f4088c90432149bd6de567e621266fa2 ("net: keep sk->sk_forward_alloc as small as possible")
> > https://git.kernel.org/cgit/linux/kernel/git/next/linux-next.git master
> >
> > in testcase: netperf
> > on test machine: 144 threads 4 sockets Intel(R) Xeon(R) Gold 5318H CPU @ 2.50GHz with 128G memory
> > with following parameters:
> >
> > ip: ipv4
> > runtime: 300s
> > nr_threads: 25%
> > cluster: cs-localhost
> > send_size: 10K
> > test: SCTP_STREAM_MANY
> > cpufreq_governor: performance
> > ucode: 0x7002402
> >
> > test-description: Netperf is a benchmark that can be use to measure various aspect of networking performance.
> > test-url: http://www.netperf.org/netperf/
> >
> > In addition to that, the commit also has significant impact on the following tests:
> >
> > +------------------+----------------------------------------------------------------------------------+
> > | testcase: change | netperf: netperf.Throughput_Mbps -73.7% regression |
> > | test machine | 144 threads 4 sockets Intel(R) Xeon(R) Gold 5318H CPU @ 2.50GHz with 128G memory |
> > | test parameters | cluster=cs-localhost |
> > | | cpufreq_governor=performance |
> > | | ip=ipv4 |
> > | | nr_threads=50% |
> > | | runtime=300s |
> > | | send_size=10K |
> > | | test=SCTP_STREAM_MANY |
> > | | ucode=0x7002402 |
> > +------------------+----------------------------------------------------------------------------------+
> >
> >
> > If you fix the issue, kindly add following tag
> > Reported-by: kernel test robot <[email protected]>
> >
> >
> > Details are as below:
> > -------------------------------------------------------------------------------------------------->
> >
> >
> > To reproduce:
> >
> > git clone https://github.com/intel/lkp-tests.git
> > cd lkp-tests
> > sudo bin/lkp install job.yaml # job file is attached in this email
> > bin/lkp split-job --compatible job.yaml # generate the yaml file for lkp run
> > sudo bin/lkp run generated-yaml-file
> >
> > # if come across any failure that blocks the test,
> > # please remove ~/.lkp and /lkp dir to run from a clean state.
> >
> > =========================================================================================
> > cluster/compiler/cpufreq_governor/ip/kconfig/nr_threads/rootfs/runtime/send_size/tbox_group/test/testcase/ucode:
> > cs-localhost/gcc-11/performance/ipv4/x86_64-rhel-8.3/25%/debian-11.1-x86_64-20220510.cgz/300s/10K/lkp-cpl-4sp1/SCTP_STREAM_MANY/netperf/0x7002402
> >
> > commit:
> > 7c80b038d2 ("net: fix sk_wmem_schedule() and sk_rmem_schedule() errors")
> > 4890b686f4 ("net: keep sk->sk_forward_alloc as small as possible")
> >
> > 7c80b038d23e1f4c 4890b686f4088c90432149bd6de
> > ---------------- ---------------------------
> > %stddev %change %stddev
> > \ | \
> > 15855 -69.4% 4854 netperf.Throughput_Mbps
> > 570788 -69.4% 174773 netperf.Throughput_total_Mbps
> > 30543 -59.1% 12480 ± 3% netperf.time.involuntary_context_switches
> > 16661 ± 2% -20.1% 13317 ± 2% netperf.time.minor_page_faults
> > 2076 -58.1% 869.17 ± 2% netperf.time.percent_of_cpu_this_job_got
> > 6118 -57.8% 2583 ± 2% netperf.time.system_time
> > 143.71 ± 6% -72.3% 39.77 netperf.time.user_time
> > 87371 ± 8% +81.1% 158253 netperf.time.voluntary_context_switches
> > 2.09e+09 -69.4% 6.4e+08 netperf.workload
> > 36264 +18.2% 42849 uptime.idle
> > 3.058e+10 +22.0% 3.731e+10 cpuidle..time
> > 6.536e+08 -60.7% 2.566e+08 cpuidle..usage
> > 127819 ± 79% -82.3% 22609 ± 65% numa-meminfo.node3.Inactive
> > 127764 ± 79% -82.4% 22440 ± 66% numa-meminfo.node3.Inactive(anon)
> > 71.38 +14.7 86.09 mpstat.cpu.all.idle%
> > 4.84 -2.1 2.78 ± 2% mpstat.cpu.all.soft%
> > 22.32 -12.1 10.25 ± 2% mpstat.cpu.all.sys%
> > 0.78 ± 3% -0.5 0.25 mpstat.cpu.all.usr%
> > 71.00 +20.9% 85.83 vmstat.cpu.id
> > 4121749 -23.6% 3150424 vmstat.memory.cache
> > 43.00 ± 2% -55.0% 19.33 ± 2% vmstat.procs.r
> > 3921783 -69.4% 1200263 vmstat.system.cs
> > 292198 -1.3% 288478 vmstat.system.in
> > 6.908e+08 ± 3% -70.1% 2.064e+08 ± 9% numa-numastat.node0.local_node
> > 6.907e+08 ± 3% -70.1% 2.065e+08 ± 9% numa-numastat.node0.numa_hit
> > 6.517e+08 ± 2% -65.5% 2.246e+08 ± 10% numa-numastat.node1.local_node
> > 6.519e+08 ± 2% -65.5% 2.247e+08 ± 10% numa-numastat.node1.numa_hit
> > 6.81e+08 ± 3% -72.1% 1.902e+08 ± 8% numa-numastat.node2.local_node
> > 6.811e+08 ± 3% -72.0% 1.905e+08 ± 8% numa-numastat.node2.numa_hit
> > 6.819e+08 ± 3% -69.9% 2.05e+08 ± 6% numa-numastat.node3.local_node
> > 6.823e+08 ± 3% -69.9% 2.051e+08 ± 6% numa-numastat.node3.numa_hit
> > 6.907e+08 ± 3% -70.1% 2.065e+08 ± 9% numa-vmstat.node0.numa_hit
> > 6.908e+08 ± 3% -70.1% 2.064e+08 ± 9% numa-vmstat.node0.numa_local
> > 6.519e+08 ± 2% -65.5% 2.247e+08 ± 10% numa-vmstat.node1.numa_hit
> > 6.517e+08 ± 2% -65.5% 2.246e+08 ± 10% numa-vmstat.node1.numa_local
> > 6.811e+08 ± 3% -72.0% 1.905e+08 ± 8% numa-vmstat.node2.numa_hit
> > 6.81e+08 ± 3% -72.1% 1.902e+08 ± 8% numa-vmstat.node2.numa_local
> > 31948 ± 79% -82.4% 5622 ± 66% numa-vmstat.node3.nr_inactive_anon
> > 31948 ± 79% -82.4% 5622 ± 66% numa-vmstat.node3.nr_zone_inactive_anon
> > 6.823e+08 ± 3% -69.9% 2.051e+08 ± 6% numa-vmstat.node3.numa_hit
> > 6.819e+08 ± 3% -69.9% 2.05e+08 ± 6% numa-vmstat.node3.numa_local
> > 1216292 ± 5% -70.1% 363637 ± 4% meminfo.Active
> > 1215873 ± 5% -70.1% 363097 ± 4% meminfo.Active(anon)
> > 3994598 -24.3% 3022264 meminfo.Cached
> > 1878317 ± 4% -53.7% 870428 meminfo.Committed_AS
> > 431272 ± 8% -28.0% 310461 meminfo.Inactive
> > 428860 ± 9% -28.4% 306912 meminfo.Inactive(anon)
> > 668273 ± 5% -71.3% 192010 ± 4% meminfo.Mapped
> > 5793516 -17.1% 4802753 meminfo.Memused
> > 8320 ± 2% -12.1% 7309 ± 2% meminfo.PageTables
> > 1341918 ± 6% -72.5% 368590 ± 4% meminfo.Shmem
> > 9796929 -25.1% 7337138 meminfo.max_used_kB
> > 1029 -51.6% 498.17 ± 2% turbostat.Avg_MHz
> > 31.32 -16.2 15.17 ± 2% turbostat.Busy%
> > 5.953e+08 -69.1% 1.837e+08 ± 2% turbostat.C1
> > 9.22 -5.2 4.04 ± 4% turbostat.C1%
> > 68.58 +23.5% 84.67 turbostat.CPU%c1
> > 57.67 -16.8% 48.00 ± 3% turbostat.CoreTmp
> > 0.23 ± 2% -31.4% 0.16 turbostat.IPC
> > 18.59 ± 30% -18.6 0.00 turbostat.PKG_%
> > 4524200 ± 6% -81.5% 835280 turbostat.POLL
> > 0.10 ± 3% -0.1 0.02 ± 17% turbostat.POLL%
> > 57.33 -16.6% 47.83 ± 4% turbostat.PkgTmp
> > 561.39 -25.9% 416.03 turbostat.PkgWatt
> > 7.75 +6.8% 8.27 turbostat.RAMWatt
> > 303881 ± 5% -70.2% 90609 ± 4% proc-vmstat.nr_active_anon
> > 998594 -24.4% 755404 proc-vmstat.nr_file_pages
> > 107254 ± 9% -28.4% 76741 proc-vmstat.nr_inactive_anon
> > 167104 ± 5% -71.0% 48414 ± 4% proc-vmstat.nr_mapped
> > 2081 ± 2% -12.1% 1828 ± 2% proc-vmstat.nr_page_table_pages
> > 335422 ± 6% -72.6% 91983 ± 4% proc-vmstat.nr_shmem
> > 32561 -1.7% 31994 proc-vmstat.nr_slab_reclaimable
> > 303881 ± 5% -70.2% 90609 ± 4% proc-vmstat.nr_zone_active_anon
> > 107254 ± 9% -28.4% 76741 proc-vmstat.nr_zone_inactive_anon
> > 267554 ± 15% -25.4% 199518 ± 10% proc-vmstat.numa_hint_faults
> > 82267 ± 20% -27.2% 59929 ± 13% proc-vmstat.numa_hint_faults_local
> > 2.706e+09 -69.5% 8.266e+08 proc-vmstat.numa_hit
> > 2.705e+09 -69.5% 8.262e+08 proc-vmstat.numa_local
> > 489361 ± 9% -19.4% 394229 ± 6% proc-vmstat.numa_pte_updates
> > 510140 ± 18% -75.3% 125829 ± 2% proc-vmstat.pgactivate
> > 7.17e+09 -69.4% 2.197e+09 proc-vmstat.pgalloc_normal
> > 1636476 ± 2% -11.5% 1448944 proc-vmstat.pgfault
> > 7.17e+09 -69.4% 2.197e+09 proc-vmstat.pgfree
> > 30967 ± 41% -88.7% 3488 ± 30% sched_debug.cfs_rq:/.MIN_vruntime.avg
> > 1604384 ± 22% -83.6% 262510 ± 14% sched_debug.cfs_rq:/.MIN_vruntime.max
> > 201146 ± 30% -86.1% 28034 ± 18% sched_debug.cfs_rq:/.MIN_vruntime.stddev
> > 0.29 ± 8% -43.5% 0.16 ± 9% sched_debug.cfs_rq:/.h_nr_running.avg
> > 0.42 ± 3% -15.1% 0.36 ± 3% sched_debug.cfs_rq:/.h_nr_running.stddev
> > 30967 ± 41% -88.7% 3488 ± 30% sched_debug.cfs_rq:/.max_vruntime.avg
> > 1604384 ± 22% -83.6% 262510 ± 14% sched_debug.cfs_rq:/.max_vruntime.max
> > 201146 ± 30% -86.1% 28034 ± 18% sched_debug.cfs_rq:/.max_vruntime.stddev
> > 1526268 ± 11% -80.6% 295647 ± 10% sched_debug.cfs_rq:/.min_vruntime.avg
> > 2336190 ± 12% -79.4% 482042 ± 7% sched_debug.cfs_rq:/.min_vruntime.max
> > 859659 ± 11% -85.8% 121714 ± 18% sched_debug.cfs_rq:/.min_vruntime.min
> > 343902 ± 15% -78.7% 73263 ± 6% sched_debug.cfs_rq:/.min_vruntime.stddev
> > 0.29 ± 8% -43.6% 0.16 ± 9% sched_debug.cfs_rq:/.nr_running.avg
> > 0.42 ± 3% -15.1% 0.36 ± 3% sched_debug.cfs_rq:/.nr_running.stddev
> > 312.39 ± 7% -46.1% 168.29 ± 5% sched_debug.cfs_rq:/.runnable_avg.avg
> > 355.85 ± 2% -27.3% 258.86 ± 3% sched_debug.cfs_rq:/.runnable_avg.stddev
> > 984535 ± 22% -82.1% 176253 ± 31% sched_debug.cfs_rq:/.spread0.max
> > -491914 -62.6% -184050 sched_debug.cfs_rq:/.spread0.min
> > 343876 ± 15% -78.7% 73261 ± 6% sched_debug.cfs_rq:/.spread0.stddev
> > 312.25 ± 7% -46.1% 168.23 ± 5% sched_debug.cfs_rq:/.util_avg.avg
> > 355.80 ± 2% -27.3% 258.80 ± 3% sched_debug.cfs_rq:/.util_avg.stddev
> > 213.78 ± 8% -62.3% 80.50 ± 11% sched_debug.cfs_rq:/.util_est_enqueued.avg
> > 331.99 ± 3% -37.4% 207.79 ± 4% sched_debug.cfs_rq:/.util_est_enqueued.stddev
> > 630039 +19.0% 749877 ± 3% sched_debug.cpu.avg_idle.avg
> > 4262 +45.9% 6219 ± 3% sched_debug.cpu.avg_idle.min
> > 2345 ± 14% -35.1% 1522 ± 6% sched_debug.cpu.clock_task.stddev
> > 1496 ± 5% -47.9% 779.72 ± 7% sched_debug.cpu.curr->pid.avg
> > 2528 ± 3% -19.7% 2031 ± 3% sched_debug.cpu.curr->pid.stddev
> > 0.00 ± 10% -23.0% 0.00 ± 4% sched_debug.cpu.next_balance.stddev
> > 0.25 ± 6% -48.0% 0.13 ± 8% sched_debug.cpu.nr_running.avg
> > 0.41 ± 2% -19.9% 0.33 ± 3% sched_debug.cpu.nr_running.stddev
> > 3859540 ± 10% -70.7% 1132195 ± 10% sched_debug.cpu.nr_switches.avg
> > 7113795 ± 11% -64.8% 2504199 ± 10% sched_debug.cpu.nr_switches.max
> > 1618475 ± 15% -86.1% 224980 ± 30% sched_debug.cpu.nr_switches.min
> > 1045624 ± 9% -57.2% 447501 ± 9% sched_debug.cpu.nr_switches.stddev
> > 2.446e+10 -66.9% 8.087e+09 perf-stat.i.branch-instructions
> > 2.356e+08 ± 2% -65.6% 80979566 ± 3% perf-stat.i.branch-misses
> > 0.74 ± 8% +5.7 6.47 perf-stat.i.cache-miss-rate%
> > 16166727 ± 10% +236.6% 54420281 ± 2% perf-stat.i.cache-misses
> > 2.609e+09 -67.5% 8.493e+08 ± 2% perf-stat.i.cache-references
> > 3956687 -69.4% 1211999 perf-stat.i.context-switches
> > 1.23 +44.9% 1.79 perf-stat.i.cpi
> > 1.51e+11 -52.3% 7.199e+10 ± 2% perf-stat.i.cpu-cycles
> > 363.75 -38.7% 223.06 perf-stat.i.cpu-migrations
> > 10558 ± 10% -87.4% 1326 perf-stat.i.cycles-between-cache-misses
> > 652403 ±109% -72.7% 178258 ± 8% perf-stat.i.dTLB-load-misses
> > 3.497e+10 -66.9% 1.158e+10 perf-stat.i.dTLB-loads
> > 0.00 ± 3% +0.0 0.00 ± 11% perf-stat.i.dTLB-store-miss-rate%
> > 119514 ± 5% -51.8% 57642 ± 9% perf-stat.i.dTLB-store-misses
> > 2.067e+10 -66.9% 6.833e+09 perf-stat.i.dTLB-stores
> > 74.98 +0.8 75.82 perf-stat.i.iTLB-load-miss-rate%
> > 1.648e+08 ± 3% -67.2% 54063746 ± 4% perf-stat.i.iTLB-load-misses
> > 54768215 -68.8% 17107737 perf-stat.i.iTLB-loads
> > 1.223e+11 -66.9% 4.052e+10 perf-stat.i.instructions
> > 0.81 -30.7% 0.56 perf-stat.i.ipc
> > 1.05 -52.3% 0.50 ± 2% perf-stat.i.metric.GHz
> > 453.64 -46.1% 244.53 perf-stat.i.metric.K/sec
> > 574.31 -66.9% 189.94 perf-stat.i.metric.M/sec
> > 4984 ± 2% -12.4% 4368 perf-stat.i.minor-faults
> > 4306378 ± 13% +145.4% 10568372 ± 2% perf-stat.i.node-load-misses
> > 267971 ± 20% +205.5% 818656 ± 4% perf-stat.i.node-loads
> > 93.32 +4.7 98.02 perf-stat.i.node-store-miss-rate%
> > 1642200 ± 14% +217.5% 5214268 ± 2% perf-stat.i.node-store-misses
> > 198883 ± 16% -39.3% 120629 ± 9% perf-stat.i.node-stores
> > 4985 ± 2% -12.4% 4369 perf-stat.i.page-faults
> > 21.33 -1.7% 20.96 perf-stat.overall.MPKI
> > 0.62 ± 10% +5.8 6.41 perf-stat.overall.cache-miss-rate%
> > 1.23 +44.0% 1.78 perf-stat.overall.cpi
> > 9426 ± 9% -86.0% 1323 perf-stat.overall.cycles-between-cache-misses
> > 0.00 ± 4% +0.0 0.00 ± 9% perf-stat.overall.dTLB-store-miss-rate%
> > 75.04 +0.9 75.95 perf-stat.overall.iTLB-load-miss-rate%
> > 0.81 -30.5% 0.56 perf-stat.overall.ipc
> > 89.11 +8.6 97.73 perf-stat.overall.node-store-miss-rate%
> > 17633 +8.3% 19103 perf-stat.overall.path-length
> > 2.438e+10 -66.9% 8.061e+09 perf-stat.ps.branch-instructions
> > 2.348e+08 ± 2% -65.6% 80709139 ± 3% perf-stat.ps.branch-misses
> > 16118295 ± 10% +236.5% 54240216 ± 2% perf-stat.ps.cache-misses
> > 2.6e+09 -67.4% 8.464e+08 ± 2% perf-stat.ps.cache-references
> > 3943074 -69.4% 1207953 perf-stat.ps.context-switches
> > 1.505e+11 -52.3% 7.175e+10 ± 2% perf-stat.ps.cpu-cycles
> > 362.72 -38.7% 222.48 perf-stat.ps.cpu-migrations
> > 650701 ±109% -72.7% 177916 ± 8% perf-stat.ps.dTLB-load-misses
> > 3.485e+10 -66.9% 1.155e+10 perf-stat.ps.dTLB-loads
> > 119179 ± 5% -51.8% 57464 ± 9% perf-stat.ps.dTLB-store-misses
> > 2.06e+10 -66.9% 6.81e+09 perf-stat.ps.dTLB-stores
> > 1.642e+08 ± 3% -67.2% 53885104 ± 4% perf-stat.ps.iTLB-load-misses
> > 54583301 -68.8% 17050690 perf-stat.ps.iTLB-loads
> > 1.219e+11 -66.9% 4.039e+10 perf-stat.ps.instructions
> > 4965 ± 2% -12.3% 4352 perf-stat.ps.minor-faults
> > 4293134 ± 13% +145.4% 10534117 ± 2% perf-stat.ps.node-load-misses
> > 267097 ± 20% +205.5% 816047 ± 4% perf-stat.ps.node-loads
> > 1637287 ± 13% +217.4% 5197315 ± 2% perf-stat.ps.node-store-misses
> > 198487 ± 15% -39.3% 120430 ± 9% perf-stat.ps.node-stores
> > 4966 ± 2% -12.3% 4353 perf-stat.ps.page-faults
> > 3.686e+13 -66.8% 1.223e+13 perf-stat.total.instructions
> > 9.05 ± 7% -3.4 5.70 ± 6% perf-profile.calltrace.cycles-pp.sctp_packet_pack.sctp_packet_transmit.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm
> > 8.94 ± 7% -3.2 5.69 ± 5% perf-profile.calltrace.cycles-pp.sctp_do_sm.sctp_primitive_SEND.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg
> > 8.61 ± 7% -3.2 5.44 ± 6% perf-profile.calltrace.cycles-pp.memcpy_erms.sctp_packet_pack.sctp_packet_transmit.sctp_outq_flush.sctp_cmd_interpreter
> > 5.69 ± 4% -3.0 2.68 ± 7% perf-profile.calltrace.cycles-pp.mwait_idle_with_hints.intel_idle_irq.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call
> > 5.85 ± 4% -3.0 2.90 ± 8% perf-profile.calltrace.cycles-pp.intel_idle_irq.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
> > 21.32 ± 6% -2.9 18.43 ± 5% perf-profile.calltrace.cycles-pp.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg.____sys_sendmsg.___sys_sendmsg
> > 13.45 ± 7% -2.5 10.92 ± 5% perf-profile.calltrace.cycles-pp.sctp_primitive_SEND.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg.____sys_sendmsg
> > 7.02 ± 7% -2.5 4.51 ± 6% perf-profile.calltrace.cycles-pp.sctp_datamsg_from_user.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg.____sys_sendmsg
> > 6.53 ± 6% -2.5 4.06 ± 6% perf-profile.calltrace.cycles-pp.skb_copy_datagram_iter.sctp_recvmsg.inet_recvmsg.____sys_recvmsg.___sys_recvmsg
> > 6.51 ± 6% -2.5 4.05 ± 6% perf-profile.calltrace.cycles-pp.__skb_datagram_iter.skb_copy_datagram_iter.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
> > 5.68 ± 6% -2.1 3.54 ± 6% perf-profile.calltrace.cycles-pp._copy_to_iter.__skb_datagram_iter.skb_copy_datagram_iter.sctp_recvmsg.inet_recvmsg
> > 5.52 ± 6% -2.1 3.45 ± 6% perf-profile.calltrace.cycles-pp.copyout._copy_to_iter.__skb_datagram_iter.skb_copy_datagram_iter.sctp_recvmsg
> > 5.48 ± 6% -2.1 3.42 ± 6% perf-profile.calltrace.cycles-pp.copy_user_enhanced_fast_string.copyout._copy_to_iter.__skb_datagram_iter.skb_copy_datagram_iter
> > 11.97 ± 7% -2.0 9.98 ± 5% perf-profile.calltrace.cycles-pp.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm.sctp_primitive_SEND.sctp_sendmsg_to_asoc
> > 9.16 ± 7% -1.8 7.37 ± 5% perf-profile.calltrace.cycles-pp.sctp_cmd_interpreter.sctp_do_sm.sctp_primitive_SEND.sctp_sendmsg_to_asoc.sctp_sendmsg
> > 10.42 ± 7% -1.4 9.00 ± 5% perf-profile.calltrace.cycles-pp.sctp_packet_transmit.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm.sctp_primitive_SEND
> > 1.58 ± 6% -1.2 0.35 ± 70% perf-profile.calltrace.cycles-pp.__alloc_pages.kmalloc_large_node.__kmalloc_node_track_caller.kmalloc_reserve.__alloc_skb
> > 3.49 ± 6% -1.2 2.30 ± 7% perf-profile.calltrace.cycles-pp.sctp_user_addto_chunk.sctp_datamsg_from_user.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg
> > 3.15 ± 6% -1.1 2.06 ± 6% perf-profile.calltrace.cycles-pp._copy_from_iter.sctp_user_addto_chunk.sctp_datamsg_from_user.sctp_sendmsg_to_asoc.sctp_sendmsg
> > 2.86 ± 7% -1.1 1.79 ± 7% perf-profile.calltrace.cycles-pp.sctp_make_datafrag_empty.sctp_datamsg_from_user.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg
> > 3.16 ± 7% -1.1 2.11 ± 5% perf-profile.calltrace.cycles-pp.sctp_data_ready.sctp_ulpq_tail_event.sctp_ulpq_tail_data.sctp_cmd_interpreter.sctp_do_sm
> > 3.00 ± 6% -1.0 1.96 ± 6% perf-profile.calltrace.cycles-pp.copyin._copy_from_iter.sctp_user_addto_chunk.sctp_datamsg_from_user.sctp_sendmsg_to_asoc
> > 2.97 ± 6% -1.0 1.93 ± 6% perf-profile.calltrace.cycles-pp.copy_user_enhanced_fast_string.copyin._copy_from_iter.sctp_user_addto_chunk.sctp_datamsg_from_user
> > 3.26 ± 7% -1.0 2.22 ± 5% perf-profile.calltrace.cycles-pp.sctp_ulpq_tail_event.sctp_ulpq_tail_data.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv
> > 2.89 ± 7% -1.0 1.87 ± 5% perf-profile.calltrace.cycles-pp.__wake_up_common_lock.sctp_data_ready.sctp_ulpq_tail_event.sctp_ulpq_tail_data.sctp_cmd_interpreter
> > 2.68 ± 6% -0.9 1.74 ± 5% perf-profile.calltrace.cycles-pp.__wake_up_common.__wake_up_common_lock.sctp_data_ready.sctp_ulpq_tail_event.sctp_ulpq_tail_data
> > 2.40 ± 7% -0.9 1.46 ± 6% perf-profile.calltrace.cycles-pp._sctp_make_chunk.sctp_make_datafrag_empty.sctp_datamsg_from_user.sctp_sendmsg_to_asoc.sctp_sendmsg
> > 2.49 ± 7% -0.9 1.62 ± 5% perf-profile.calltrace.cycles-pp.autoremove_wake_function.__wake_up_common.__wake_up_common_lock.sctp_data_ready.sctp_ulpq_tail_event
> > 2.46 ± 6% -0.9 1.60 ± 5% perf-profile.calltrace.cycles-pp.try_to_wake_up.autoremove_wake_function.__wake_up_common.__wake_up_common_lock.sctp_data_ready
> > 1.86 ± 6% -0.8 1.10 ± 6% perf-profile.calltrace.cycles-pp.schedule_timeout.sctp_skb_recv_datagram.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
> > 1.81 ± 7% -0.8 1.06 ± 5% perf-profile.calltrace.cycles-pp.__alloc_skb._sctp_make_chunk.sctp_make_datafrag_empty.sctp_datamsg_from_user.sctp_sendmsg_to_asoc
> > 1.76 ± 9% -0.7 1.01 ± 6% perf-profile.calltrace.cycles-pp.sctp_ulpevent_free.sctp_recvmsg.inet_recvmsg.____sys_recvmsg.___sys_recvmsg
> > 1.82 ± 6% -0.7 1.08 ± 6% perf-profile.calltrace.cycles-pp.schedule.schedule_timeout.sctp_skb_recv_datagram.sctp_recvmsg.inet_recvmsg
> > 1.50 ± 6% -0.7 0.76 ± 5% perf-profile.calltrace.cycles-pp.skb_release_data.kfree_skb_reason.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
> > 1.79 ± 6% -0.7 1.06 ± 6% perf-profile.calltrace.cycles-pp.__schedule.schedule.schedule_timeout.sctp_skb_recv_datagram.sctp_recvmsg
> > 1.61 ± 6% -0.6 1.01 ± 6% perf-profile.calltrace.cycles-pp.schedule_idle.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
> > 0.86 ± 7% -0.6 0.26 ±100% perf-profile.calltrace.cycles-pp.dequeue_entity.dequeue_task_fair.__schedule.schedule.schedule_timeout
> > 1.56 ± 6% -0.6 0.98 ± 6% perf-profile.calltrace.cycles-pp.__schedule.schedule_idle.do_idle.cpu_startup_entry.start_secondary
> > 1.56 ± 8% -0.6 0.99 ± 6% perf-profile.calltrace.cycles-pp.sctp_outq_flush_data.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm.sctp_primitive_SEND
> > 1.25 ± 7% -0.6 0.68 ± 5% perf-profile.calltrace.cycles-pp.kmalloc_reserve.__alloc_skb._sctp_make_chunk.sctp_make_datafrag_empty.sctp_datamsg_from_user
> > 1.22 ± 7% -0.6 0.66 ± 5% perf-profile.calltrace.cycles-pp.__kmalloc_node_track_caller.kmalloc_reserve.__alloc_skb._sctp_make_chunk.sctp_make_datafrag_empty
> > 1.19 ± 7% -0.5 0.64 ± 5% perf-profile.calltrace.cycles-pp.kmalloc_large_node.__kmalloc_node_track_caller.kmalloc_reserve.__alloc_skb._sctp_make_chunk
> > 0.90 ± 7% -0.5 0.39 ± 70% perf-profile.calltrace.cycles-pp.skb_release_data.consume_skb.sctp_chunk_put.sctp_outq_sack.sctp_cmd_interpreter
> > 0.94 ± 7% -0.5 0.46 ± 45% perf-profile.calltrace.cycles-pp.dequeue_task_fair.__schedule.schedule.schedule_timeout.sctp_skb_recv_datagram
> > 0.74 ± 6% -0.5 0.27 ±100% perf-profile.calltrace.cycles-pp.sctp_endpoint_lookup_assoc.sctp_sendmsg.sock_sendmsg.____sys_sendmsg.___sys_sendmsg
> > 0.98 ± 8% -0.4 0.57 ± 7% perf-profile.calltrace.cycles-pp.sctp_chunk_put.sctp_ulpevent_free.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
> > 1.05 ± 8% -0.4 0.66 ± 7% perf-profile.calltrace.cycles-pp.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv.sctp_rcv
> > 1.10 ± 7% -0.4 0.72 ± 6% perf-profile.calltrace.cycles-pp.sendmsg_copy_msghdr.___sys_sendmsg.__sys_sendmsg.do_syscall_64.entry_SYSCALL_64_after_hwframe
> > 0.46 ± 45% +0.4 0.83 ± 27% perf-profile.calltrace.cycles-pp.menu_select.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary
> > 0.00 +0.8 0.84 ± 24% perf-profile.calltrace.cycles-pp.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state
> > 0.00 +0.8 0.84 ± 24% perf-profile.calltrace.cycles-pp.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter
> > 0.00 +0.9 0.85 ± 6% perf-profile.calltrace.cycles-pp.page_counter_uncharge.drain_stock.refill_stock.__sk_mem_reduce_allocated.skb_release_head_state
> > 0.00 +0.9 0.87 ± 6% perf-profile.calltrace.cycles-pp.drain_stock.refill_stock.__sk_mem_reduce_allocated.skb_release_head_state.kfree_skb_reason
> > 0.00 +1.0 0.98 ± 7% perf-profile.calltrace.cycles-pp.refill_stock.__sk_mem_reduce_allocated.skb_release_head_state.kfree_skb_reason.sctp_recvmsg
> > 0.00 +1.2 1.24 ± 5% perf-profile.calltrace.cycles-pp.page_counter_uncharge.drain_stock.refill_stock.__sk_mem_reduce_allocated.sctp_wfree
> > 0.00 +1.3 1.26 ± 5% perf-profile.calltrace.cycles-pp.drain_stock.refill_stock.__sk_mem_reduce_allocated.sctp_wfree.skb_release_head_state
> > 0.00 +1.3 1.27 ± 19% perf-profile.calltrace.cycles-pp.try_charge_memcg.mem_cgroup_charge_skmem.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_ulpevent_make_rcvmsg
> > 0.00 +1.3 1.29 ± 26% perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call
> > 1.60 ± 10% +1.3 2.92 ± 5% perf-profile.calltrace.cycles-pp.sctp_ulpevent_make_rcvmsg.sctp_ulpq_tail_data.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv
> > 0.00 +1.4 1.35 ± 5% perf-profile.calltrace.cycles-pp.refill_stock.__sk_mem_reduce_allocated.sctp_wfree.skb_release_head_state.consume_skb
> > 0.00 +1.4 1.42 ± 25% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
> > 0.00 +1.6 1.63 ± 5% perf-profile.calltrace.cycles-pp.mem_cgroup_charge_skmem.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_ulpevent_make_rcvmsg.sctp_ulpq_tail_data
> > 0.00 +1.7 1.66 ± 19% perf-profile.calltrace.cycles-pp.page_counter_try_charge.try_charge_memcg.mem_cgroup_charge_skmem.__sk_mem_raise_allocated.__sk_mem_schedule
> > 10.96 ± 7% +1.7 12.67 ± 5% perf-profile.calltrace.cycles-pp.__ip_queue_xmit.sctp_packet_transmit.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm
> > 0.00 +1.7 1.72 ± 5% perf-profile.calltrace.cycles-pp.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_ulpevent_make_rcvmsg.sctp_ulpq_tail_data.sctp_cmd_interpreter
> > 0.00 +1.7 1.73 ± 5% perf-profile.calltrace.cycles-pp.__sk_mem_schedule.sctp_ulpevent_make_rcvmsg.sctp_ulpq_tail_data.sctp_cmd_interpreter.sctp_do_sm
> > 10.84 ± 7% +1.8 12.59 ± 5% perf-profile.calltrace.cycles-pp.ip_finish_output2.__ip_queue_xmit.sctp_packet_transmit.sctp_outq_flush.sctp_cmd_interpreter
> > 10.63 ± 7% +1.8 12.42 ± 5% perf-profile.calltrace.cycles-pp.__local_bh_enable_ip.ip_finish_output2.__ip_queue_xmit.sctp_packet_transmit.sctp_outq_flush
> > 10.58 ± 7% +1.8 12.40 ± 5% perf-profile.calltrace.cycles-pp.do_softirq.__local_bh_enable_ip.ip_finish_output2.__ip_queue_xmit.sctp_packet_transmit
> > 10.53 ± 7% +1.8 12.36 ± 5% perf-profile.calltrace.cycles-pp.__softirqentry_text_start.do_softirq.__local_bh_enable_ip.ip_finish_output2.__ip_queue_xmit
> > 10.36 ± 7% +1.9 12.25 ± 5% perf-profile.calltrace.cycles-pp.net_rx_action.__softirqentry_text_start.do_softirq.__local_bh_enable_ip.ip_finish_output2
> > 10.26 ± 7% +1.9 12.18 ± 5% perf-profile.calltrace.cycles-pp.__napi_poll.net_rx_action.__softirqentry_text_start.do_softirq.__local_bh_enable_ip
> > 10.23 ± 7% +1.9 12.17 ± 5% perf-profile.calltrace.cycles-pp.process_backlog.__napi_poll.net_rx_action.__softirqentry_text_start.do_softirq
> > 10.07 ± 7% +2.0 12.08 ± 5% perf-profile.calltrace.cycles-pp.__netif_receive_skb_one_core.process_backlog.__napi_poll.net_rx_action.__softirqentry_text_start
> > 0.00 +2.0 2.02 ± 6% perf-profile.calltrace.cycles-pp.try_charge_memcg.mem_cgroup_charge_skmem.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_sendmsg_to_asoc
> > 9.84 ± 7% +2.1 11.95 ± 5% perf-profile.calltrace.cycles-pp.ip_local_deliver_finish.__netif_receive_skb_one_core.process_backlog.__napi_poll.net_rx_action
> > 9.82 ± 7% +2.1 11.94 ± 5% perf-profile.calltrace.cycles-pp.ip_protocol_deliver_rcu.ip_local_deliver_finish.__netif_receive_skb_one_core.process_backlog.__napi_poll
> > 9.74 ± 7% +2.1 11.88 ± 5% perf-profile.calltrace.cycles-pp.sctp_rcv.ip_protocol_deliver_rcu.ip_local_deliver_finish.__netif_receive_skb_one_core.process_backlog
> > 3.38 ± 7% +2.1 5.53 ± 5% perf-profile.calltrace.cycles-pp.sctp_skb_recv_datagram.sctp_recvmsg.inet_recvmsg.____sys_recvmsg.___sys_recvmsg
> > 0.00 +2.4 2.38 ± 6% perf-profile.calltrace.cycles-pp.mem_cgroup_charge_skmem.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_sendmsg_to_asoc.sctp_sendmsg
> > 0.00 +2.5 2.54 ± 6% perf-profile.calltrace.cycles-pp.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg
> > 0.00 +2.5 2.55 ± 6% perf-profile.calltrace.cycles-pp.__sk_mem_schedule.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg.____sys_sendmsg
> > 8.08 ± 7% +2.7 10.74 ± 5% perf-profile.calltrace.cycles-pp.sctp_assoc_bh_rcv.sctp_rcv.ip_protocol_deliver_rcu.ip_local_deliver_finish.__netif_receive_skb_one_core
> > 7.63 ± 7% +2.8 10.44 ± 5% perf-profile.calltrace.cycles-pp.sctp_do_sm.sctp_assoc_bh_rcv.sctp_rcv.ip_protocol_deliver_rcu.ip_local_deliver_finish
> > 15.25 ± 6% +2.9 18.11 ± 6% perf-profile.calltrace.cycles-pp.inet_recvmsg.____sys_recvmsg.___sys_recvmsg.__sys_recvmsg.do_syscall_64
> > 15.15 ± 7% +2.9 18.03 ± 5% perf-profile.calltrace.cycles-pp.release_sock.sctp_sendmsg.sock_sendmsg.____sys_sendmsg.___sys_sendmsg
> > 15.20 ± 6% +2.9 18.09 ± 6% perf-profile.calltrace.cycles-pp.sctp_recvmsg.inet_recvmsg.____sys_recvmsg.___sys_recvmsg.__sys_recvmsg
> > 15.02 ± 7% +2.9 17.94 ± 6% perf-profile.calltrace.cycles-pp.__release_sock.release_sock.sctp_sendmsg.sock_sendmsg.____sys_sendmsg
> > 14.96 ± 7% +2.9 17.90 ± 6% perf-profile.calltrace.cycles-pp.sctp_backlog_rcv.__release_sock.release_sock.sctp_sendmsg.sock_sendmsg
> > 11.70 ± 7% +3.0 14.67 ± 6% perf-profile.calltrace.cycles-pp.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv.sctp_backlog_rcv.__release_sock
> > 0.84 ± 10% +3.1 3.94 ± 5% perf-profile.calltrace.cycles-pp._raw_spin_lock_bh.lock_sock_nested.sctp_skb_recv_datagram.sctp_recvmsg.inet_recvmsg
> > 0.96 ± 9% +3.1 4.06 ± 6% perf-profile.calltrace.cycles-pp.lock_sock_nested.sctp_skb_recv_datagram.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
> > 0.64 ± 11% +3.2 3.82 ± 5% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_bh.lock_sock_nested.sctp_skb_recv_datagram.sctp_recvmsg
> > 0.00 +3.9 3.90 ± 5% perf-profile.calltrace.cycles-pp.sctp_eat_data.sctp_sf_eat_data_6_2.sctp_do_sm.sctp_assoc_bh_rcv.sctp_rcv
> > 0.00 +4.0 3.98 ± 5% perf-profile.calltrace.cycles-pp.sctp_sf_eat_data_6_2.sctp_do_sm.sctp_assoc_bh_rcv.sctp_rcv.ip_protocol_deliver_rcu
> > 2.90 ± 7% +4.2 7.09 ± 6% perf-profile.calltrace.cycles-pp.sctp_outq_sack.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv.sctp_backlog_rcv
> > 1.87 ± 7% +4.5 6.40 ± 6% perf-profile.calltrace.cycles-pp.sctp_chunk_put.sctp_outq_sack.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv
> > 1.43 ± 7% +4.6 6.01 ± 6% perf-profile.calltrace.cycles-pp.consume_skb.sctp_chunk_put.sctp_outq_sack.sctp_cmd_interpreter.sctp_do_sm
> > 1.78 ± 6% +4.6 6.42 ± 6% perf-profile.calltrace.cycles-pp.kfree_skb_reason.sctp_recvmsg.inet_recvmsg.____sys_recvmsg.___sys_recvmsg
> > 0.00 +5.1 5.10 ± 5% perf-profile.calltrace.cycles-pp.__sk_mem_reduce_allocated.sctp_wfree.skb_release_head_state.consume_skb.sctp_chunk_put
> > 0.17 ±141% +5.3 5.42 ± 6% perf-profile.calltrace.cycles-pp.skb_release_head_state.consume_skb.sctp_chunk_put.sctp_outq_sack.sctp_cmd_interpreter
> > 0.00 +5.3 5.35 ± 6% perf-profile.calltrace.cycles-pp.sctp_wfree.skb_release_head_state.consume_skb.sctp_chunk_put.sctp_outq_sack
> > 0.00 +5.5 5.51 ± 6% perf-profile.calltrace.cycles-pp.__sk_mem_reduce_allocated.skb_release_head_state.kfree_skb_reason.sctp_recvmsg.inet_recvmsg
> > 0.00 +5.7 5.65 ± 6% perf-profile.calltrace.cycles-pp.skb_release_head_state.kfree_skb_reason.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
> > 9.38 ± 6% -3.5 5.93 ± 6% perf-profile.children.cycles-pp.sctp_packet_pack
> > 24.70 ± 6% -3.4 21.34 ± 6% perf-profile.children.cycles-pp.sctp_outq_flush
> > 8.83 ± 6% -3.2 5.60 ± 6% perf-profile.children.cycles-pp.copy_user_enhanced_fast_string
> > 8.77 ± 6% -3.2 5.58 ± 6% perf-profile.children.cycles-pp.memcpy_erms
> > 5.88 ± 4% -3.0 2.92 ± 8% perf-profile.children.cycles-pp.intel_idle_irq
> > 21.34 ± 6% -2.8 18.58 ± 6% perf-profile.children.cycles-pp.sctp_sendmsg_to_asoc
> > 7.06 ± 6% -2.5 4.54 ± 7% perf-profile.children.cycles-pp.sctp_datamsg_from_user
> > 13.52 ± 6% -2.5 11.00 ± 6% perf-profile.children.cycles-pp.sctp_primitive_SEND
> > 6.53 ± 6% -2.5 4.06 ± 6% perf-profile.children.cycles-pp.skb_copy_datagram_iter
> > 6.51 ± 6% -2.5 4.05 ± 6% perf-profile.children.cycles-pp.__skb_datagram_iter
> > 5.68 ± 6% -2.1 3.54 ± 6% perf-profile.children.cycles-pp._copy_to_iter
> > 5.52 ± 6% -2.1 3.45 ± 6% perf-profile.children.cycles-pp.copyout
> > 3.33 ± 6% -1.4 1.92 ± 6% perf-profile.children.cycles-pp.__alloc_skb
> > 3.38 ± 6% -1.3 2.07 ± 6% perf-profile.children.cycles-pp.__schedule
> > 3.51 ± 6% -1.2 2.31 ± 7% perf-profile.children.cycles-pp.sctp_user_addto_chunk
> > 2.67 ± 6% -1.2 1.47 ± 6% perf-profile.children.cycles-pp.skb_release_data
> > 2.43 ± 6% -1.1 1.32 ± 6% perf-profile.children.cycles-pp.kmalloc_reserve
> > 3.17 ± 6% -1.1 2.08 ± 7% perf-profile.children.cycles-pp._copy_from_iter
> > 2.38 ± 6% -1.1 1.29 ± 6% perf-profile.children.cycles-pp.__kmalloc_node_track_caller
> > 2.88 ± 7% -1.1 1.80 ± 7% perf-profile.children.cycles-pp.sctp_make_datafrag_empty
> > 3.18 ± 6% -1.1 2.12 ± 6% perf-profile.children.cycles-pp.sctp_data_ready
> > 3.02 ± 6% -1.0 1.98 ± 7% perf-profile.children.cycles-pp.copyin
> > 3.28 ± 6% -1.0 2.25 ± 6% perf-profile.children.cycles-pp.sctp_ulpq_tail_event
> > 2.26 ± 6% -1.0 1.22 ± 5% perf-profile.children.cycles-pp.kmalloc_large_node
> > 2.62 ± 6% -1.0 1.60 ± 7% perf-profile.children.cycles-pp._sctp_make_chunk
> > 2.90 ± 6% -1.0 1.89 ± 6% perf-profile.children.cycles-pp.__wake_up_common_lock
> > 2.69 ± 6% -0.9 1.75 ± 6% perf-profile.children.cycles-pp.__wake_up_common
> > 1.97 ± 6% -0.9 1.04 ± 6% perf-profile.children.cycles-pp.__alloc_pages
> > 2.50 ± 6% -0.9 1.64 ± 6% perf-profile.children.cycles-pp.autoremove_wake_function
> > 2.48 ± 6% -0.9 1.63 ± 6% perf-profile.children.cycles-pp.try_to_wake_up
> > 1.67 ± 5% -0.8 0.86 ± 6% perf-profile.children.cycles-pp.get_page_from_freelist
> > 2.26 ± 7% -0.8 1.48 ± 7% perf-profile.children.cycles-pp.sctp_outq_flush_data
> > 1.86 ± 6% -0.8 1.10 ± 6% perf-profile.children.cycles-pp.schedule_timeout
> > 1.76 ± 8% -0.7 1.02 ± 6% perf-profile.children.cycles-pp.sctp_ulpevent_free
> > 2.08 ± 5% -0.7 1.33 ± 8% perf-profile.children.cycles-pp.kmem_cache_free
> > 1.84 ± 6% -0.7 1.10 ± 6% perf-profile.children.cycles-pp.schedule
> > 1.62 ± 6% -0.6 1.02 ± 6% perf-profile.children.cycles-pp.schedule_idle
> > 1.20 ± 6% -0.6 0.62 ± 7% perf-profile.children.cycles-pp.free_unref_page
> > 1.49 ± 5% -0.5 0.95 ± 6% perf-profile.children.cycles-pp._copy_from_user
> > 1.03 ± 6% -0.5 0.54 ± 5% perf-profile.children.cycles-pp.rmqueue
> > 1.06 ± 6% -0.5 0.58 ± 5% perf-profile.children.cycles-pp._raw_spin_lock_irqsave
> > 1.18 ± 5% -0.4 0.74 ± 6% perf-profile.children.cycles-pp.__check_object_size
> > 1.22 ± 4% -0.4 0.79 ± 8% perf-profile.children.cycles-pp.__rhashtable_lookup
> > 0.97 ± 6% -0.4 0.57 ± 7% perf-profile.children.cycles-pp.__slab_free
> > 0.94 ± 7% -0.4 0.54 ± 7% perf-profile.children.cycles-pp.dequeue_task_fair
> > 1.18 ± 6% -0.4 0.78 ± 7% perf-profile.children.cycles-pp.ttwu_do_activate
> > 0.82 ± 6% -0.4 0.43 ± 7% perf-profile.children.cycles-pp._raw_spin_lock
> > 1.11 ± 6% -0.4 0.73 ± 7% perf-profile.children.cycles-pp.sendmsg_copy_msghdr
> > 1.14 ± 6% -0.4 0.76 ± 7% perf-profile.children.cycles-pp.enqueue_task_fair
> > 0.87 ± 7% -0.4 0.50 ± 7% perf-profile.children.cycles-pp.dequeue_entity
> > 0.92 ± 11% -0.4 0.56 ± 6% perf-profile.children.cycles-pp.sctp_outq_select_transport
> > 1.00 ± 5% -0.4 0.65 ± 6% perf-profile.children.cycles-pp.__copy_msghdr_from_user
> > 0.82 ± 12% -0.3 0.49 ± 6% perf-profile.children.cycles-pp.sctp_packet_config
> > 0.84 ± 6% -0.3 0.54 ± 8% perf-profile.children.cycles-pp.import_iovec
> > 0.84 ± 6% -0.3 0.55 ± 7% perf-profile.children.cycles-pp.pick_next_task_fair
> > 0.80 ± 4% -0.3 0.51 ± 9% perf-profile.children.cycles-pp.kfree
> > 0.50 ± 6% -0.3 0.20 ± 7% perf-profile.children.cycles-pp.free_pcppages_bulk
> > 0.80 ± 6% -0.3 0.51 ± 8% perf-profile.children.cycles-pp.__import_iovec
> > 0.83 ± 6% -0.3 0.54 ± 7% perf-profile.children.cycles-pp.enqueue_entity
> > 0.79 ± 7% -0.3 0.51 ± 7% perf-profile.children.cycles-pp.kmem_cache_alloc
> > 0.70 ± 6% -0.3 0.42 ± 8% perf-profile.children.cycles-pp.update_load_avg
> > 0.78 ± 7% -0.3 0.50 ± 10% perf-profile.children.cycles-pp.sctp_hash_cmp
> > 0.73 ± 5% -0.3 0.47 ± 6% perf-profile.children.cycles-pp.__sctp_rcv_lookup
> > 0.69 ± 6% -0.3 0.44 ± 6% perf-profile.children.cycles-pp.check_heap_object
> > 0.47 ± 6% -0.3 0.22 ± 8% perf-profile.children.cycles-pp.__free_pages_ok
> > 0.70 ± 4% -0.2 0.46 ± 6% perf-profile.children.cycles-pp.sctp_addrs_lookup_transport
> > 0.68 ± 6% -0.2 0.43 ± 8% perf-profile.children.cycles-pp.iovec_from_user
> > 0.71 ± 6% -0.2 0.47 ± 8% perf-profile.children.cycles-pp.sctp_chunkify
> > 0.66 ± 5% -0.2 0.42 ± 8% perf-profile.children.cycles-pp.__might_resched
> > 0.60 ± 5% -0.2 0.36 ± 4% perf-profile.children.cycles-pp.set_next_entity
> > 0.74 ± 6% -0.2 0.51 ± 8% perf-profile.children.cycles-pp.sctp_endpoint_lookup_assoc
> > 0.72 ± 6% -0.2 0.50 ± 9% perf-profile.children.cycles-pp.sctp_epaddr_lookup_transport
> > 0.54 ± 6% -0.2 0.32 ± 10% perf-profile.children.cycles-pp.__dev_queue_xmit
> > 0.44 ± 7% -0.2 0.23 ± 6% perf-profile.children.cycles-pp.sctp_inq_pop
> > 0.61 ± 6% -0.2 0.40 ± 7% perf-profile.children.cycles-pp.copy_user_short_string
> > 0.62 ± 13% -0.2 0.42 ± 5% perf-profile.children.cycles-pp.__skb_clone
> > 0.52 ± 5% -0.2 0.32 ± 7% perf-profile.children.cycles-pp.__might_fault
> > 0.46 ± 5% -0.2 0.27 ± 7% perf-profile.children.cycles-pp.simple_copy_to_iter
> > 0.51 ± 7% -0.2 0.33 ± 4% perf-profile.children.cycles-pp.__entry_text_start
> > 0.36 ± 7% -0.2 0.19 ± 8% perf-profile.children.cycles-pp.update_curr
> > 0.56 ± 6% -0.2 0.38 ± 6% perf-profile.children.cycles-pp.__list_del_entry_valid
> > 0.37 ± 7% -0.2 0.21 ± 5% perf-profile.children.cycles-pp.__mod_node_page_state
> > 0.32 ± 7% -0.2 0.17 ± 8% perf-profile.children.cycles-pp.__zone_watermark_ok
> > 0.42 ± 10% -0.2 0.27 ± 6% perf-profile.children.cycles-pp.select_task_rq
> > 0.46 ± 17% -0.2 0.31 ± 5% perf-profile.children.cycles-pp.__copy_skb_header
> > 0.46 ± 7% -0.1 0.31 ± 10% perf-profile.children.cycles-pp.sctp_packet_transmit_chunk
> > 0.46 ± 7% -0.1 0.32 ± 4% perf-profile.children.cycles-pp.update_rq_clock
> > 0.43 ± 5% -0.1 0.29 ± 10% perf-profile.children.cycles-pp.sctp_addto_chunk
> > 0.41 ± 7% -0.1 0.27 ± 10% perf-profile.children.cycles-pp.sctp_packet_append_chunk
> > 0.34 ± 6% -0.1 0.20 ± 9% perf-profile.children.cycles-pp.dev_hard_start_xmit
> > 0.46 ± 8% -0.1 0.32 ± 6% perf-profile.children.cycles-pp.kmem_cache_alloc_node
> > 0.40 ± 5% -0.1 0.27 ± 5% perf-profile.children.cycles-pp.sockfd_lookup_light
> > 0.38 ± 6% -0.1 0.25 ± 6% perf-profile.children.cycles-pp.sctp_chunk_free
> > 0.24 ± 14% -0.1 0.10 ± 11% perf-profile.children.cycles-pp.sctp_assoc_rwnd_increase
> > 0.35 ± 6% -0.1 0.21 ± 6% perf-profile.children.cycles-pp.__list_add_valid
> > 0.36 ± 7% -0.1 0.22 ± 10% perf-profile.children.cycles-pp.sctp_gen_sack
> > 0.36 ± 5% -0.1 0.23 ± 9% perf-profile.children.cycles-pp.memset_erms
> > 0.35 ± 8% -0.1 0.22 ± 10% perf-profile.children.cycles-pp.memcg_slab_free_hook
> > 0.32 ± 6% -0.1 0.19 ± 9% perf-profile.children.cycles-pp.loopback_xmit
> > 0.36 ± 5% -0.1 0.23 ± 8% perf-profile.children.cycles-pp.sock_kmalloc
> > 0.36 ± 4% -0.1 0.24 ± 7% perf-profile.children.cycles-pp.__might_sleep
> > 0.33 ± 8% -0.1 0.21 ± 7% perf-profile.children.cycles-pp.__virt_addr_valid
> > 0.28 ± 8% -0.1 0.15 ± 8% perf-profile.children.cycles-pp.free_unref_page_commit
> > 0.34 ± 4% -0.1 0.22 ± 6% perf-profile.children.cycles-pp.free_pcp_prepare
> > 0.33 ± 7% -0.1 0.20 ± 11% perf-profile.children.cycles-pp.sctp_make_sack
> > 0.30 ± 13% -0.1 0.18 ± 9% perf-profile.children.cycles-pp.sctp_v4_xmit
> > 0.34 ± 5% -0.1 0.22 ± 6% perf-profile.children.cycles-pp.__fdget
> > 0.35 ± 6% -0.1 0.23 ± 7% perf-profile.children.cycles-pp.move_addr_to_kernel
> > 0.25 ± 5% -0.1 0.13 ± 6% perf-profile.children.cycles-pp.__free_one_page
> > 0.40 ± 9% -0.1 0.29 ± 6% perf-profile.children.cycles-pp.aa_sk_perm
> > 0.24 ± 6% -0.1 0.12 ± 8% perf-profile.children.cycles-pp.prepare_task_switch
> > 0.32 ± 7% -0.1 0.21 ± 6% perf-profile.children.cycles-pp.sctp_check_transmitted
> > 0.30 ± 6% -0.1 0.19 ± 7% perf-profile.children.cycles-pp.sctp_association_put
> > 0.36 ± 7% -0.1 0.24 ± 7% perf-profile.children.cycles-pp.sock_recvmsg
> > 0.26 ± 12% -0.1 0.16 ± 15% perf-profile.children.cycles-pp.ipv4_dst_check
> > 0.22 ± 9% -0.1 0.11 ± 46% perf-profile.children.cycles-pp.accept_connections
> > 0.31 ± 11% -0.1 0.20 ± 9% perf-profile.children.cycles-pp.select_task_rq_fair
> > 0.22 ± 17% -0.1 0.11 ± 6% perf-profile.children.cycles-pp.dst_release
> > 0.23 ± 6% -0.1 0.12 ± 7% perf-profile.children.cycles-pp.rmqueue_bulk
> > 0.22 ± 6% -0.1 0.11 ± 12% perf-profile.children.cycles-pp.sock_wfree
> > 0.28 ± 6% -0.1 0.18 ± 5% perf-profile.children.cycles-pp.send_sctp_stream_1toMany
> > 0.32 ± 7% -0.1 0.23 ± 9% perf-profile.children.cycles-pp.security_socket_recvmsg
> > 0.30 ± 7% -0.1 0.20 ± 9% perf-profile.children.cycles-pp.skb_set_owner_w
> > 0.28 ± 7% -0.1 0.18 ± 8% perf-profile.children.cycles-pp.entry_SYSRETQ_unsafe_stack
> > 0.27 ± 4% -0.1 0.17 ± 9% perf-profile.children.cycles-pp.sock_kfree_s
> > 0.27 ± 10% -0.1 0.18 ± 11% perf-profile.children.cycles-pp.sctp_transport_hold
> > 0.24 ± 6% -0.1 0.15 ± 5% perf-profile.children.cycles-pp.__switch_to
> > 0.24 ± 7% -0.1 0.15 ± 6% perf-profile.children.cycles-pp.tick_nohz_idle_exit
> > 0.21 ± 10% -0.1 0.12 ± 9% perf-profile.children.cycles-pp.recv_sctp_stream_1toMany
> > 0.23 ± 7% -0.1 0.14 ± 7% perf-profile.children.cycles-pp.prepare_to_wait_exclusive
> > 0.22 ± 9% -0.1 0.14 ± 9% perf-profile.children.cycles-pp.accept_connection
> > 0.22 ± 9% -0.1 0.14 ± 9% perf-profile.children.cycles-pp.spawn_child
> > 0.22 ± 9% -0.1 0.14 ± 9% perf-profile.children.cycles-pp.process_requests
> > 0.23 ± 3% -0.1 0.14 ± 7% perf-profile.children.cycles-pp.__kmalloc
> > 0.20 ± 9% -0.1 0.12 ± 10% perf-profile.children.cycles-pp.skb_clone
> > 0.31 ± 10% -0.1 0.22 ± 13% perf-profile.children.cycles-pp.sctp_sched_fcfs_dequeue
> > 0.29 ± 12% -0.1 0.21 ± 5% perf-profile.children.cycles-pp.sctp_transport_put
> > 0.22 ± 8% -0.1 0.14 ± 7% perf-profile.children.cycles-pp.select_idle_sibling
> > 0.22 ± 8% -0.1 0.14 ± 5% perf-profile.children.cycles-pp.__switch_to_asm
> > 0.25 ± 10% -0.1 0.17 ± 11% perf-profile.children.cycles-pp.__sctp_packet_append_chunk
> > 0.23 ± 5% -0.1 0.15 ± 10% perf-profile.children.cycles-pp.__cond_resched
> > 0.20 ± 6% -0.1 0.13 ± 5% perf-profile.children.cycles-pp.available_idle_cpu
> > 0.19 ± 9% -0.1 0.11 ± 6% perf-profile.children.cycles-pp.poll_idle
> > 0.20 ± 6% -0.1 0.13 ± 9% perf-profile.children.cycles-pp.reweight_entity
> > 0.20 ± 7% -0.1 0.13 ± 8% perf-profile.children.cycles-pp.sctp_make_control
> > 0.25 ± 9% -0.1 0.18 ± 8% perf-profile.children.cycles-pp.syscall_exit_to_user_mode
> > 0.20 ± 6% -0.1 0.12 ± 8% perf-profile.children.cycles-pp.__free_pages
> > 0.26 ± 8% -0.1 0.19 ± 9% perf-profile.children.cycles-pp.security_socket_sendmsg
> > 0.19 ± 8% -0.1 0.12 ± 13% perf-profile.children.cycles-pp.syscall_return_via_sysret
> > 0.17 ± 9% -0.1 0.10 ± 14% perf-profile.children.cycles-pp.__mod_timer
> > 0.16 ± 5% -0.1 0.09 ± 7% perf-profile.children.cycles-pp.___perf_sw_event
> > 0.21 ± 6% -0.1 0.14 ± 9% perf-profile.children.cycles-pp.sctp_datamsg_put
> > 0.21 ± 7% -0.1 0.14 ± 7% perf-profile.children.cycles-pp.sctp_association_hold
> > 0.18 ± 10% -0.1 0.11 ± 4% perf-profile.children.cycles-pp.sctp_outq_flush_ctrl
> > 0.16 ± 8% -0.1 0.09 ± 14% perf-profile.children.cycles-pp.__update_load_avg_cfs_rq
> > 0.20 ± 4% -0.1 0.13 ± 9% perf-profile.children.cycles-pp.kmem_cache_alloc_trace
> > 0.19 ± 5% -0.1 0.13 ± 8% perf-profile.children.cycles-pp.__update_load_avg_se
> > 0.16 ± 5% -0.1 0.10 ± 9% perf-profile.children.cycles-pp.del_timer
> > 0.14 ± 11% -0.1 0.08 ± 4% perf-profile.children.cycles-pp.check_new_pages
> > 0.23 ± 7% -0.1 0.17 ± 10% perf-profile.children.cycles-pp.sctp_sched_dequeue_common
> > 0.10 ± 5% -0.1 0.04 ± 71% perf-profile.children.cycles-pp.finish_task_switch
> > 0.18 ± 7% -0.1 0.12 ± 6% perf-profile.children.cycles-pp.sctp_outq_tail
> > 0.12 ± 4% -0.1 0.06 ± 13% perf-profile.children.cycles-pp.check_stack_object
> > 0.13 ± 8% -0.1 0.08 ± 8% perf-profile.children.cycles-pp.__put_user_nocheck_4
> > 0.16 ± 4% -0.1 0.10 ± 13% perf-profile.children.cycles-pp.__genradix_ptr
> > 0.17 ± 6% -0.1 0.11 ± 6% perf-profile.children.cycles-pp.nr_iowait_cpu
> > 0.16 ± 8% -0.1 0.11 ± 13% perf-profile.children.cycles-pp.sctp_sendmsg_parse
> > 0.13 ± 7% -0.1 0.08 ± 12% perf-profile.children.cycles-pp.lock_timer_base
> > 0.08 ± 12% -0.0 0.02 ± 99% perf-profile.children.cycles-pp.ip_local_out
> > 0.14 ± 7% -0.0 0.09 ± 11% perf-profile.children.cycles-pp.memcg_slab_post_alloc_hook
> > 0.14 ± 9% -0.0 0.09 ± 7% perf-profile.children.cycles-pp.switch_mm_irqs_off
> > 0.14 ± 9% -0.0 0.09 ± 10% perf-profile.children.cycles-pp.sctp_datamsg_destroy
> > 0.12 ± 20% -0.0 0.07 ± 21% perf-profile.children.cycles-pp.sctp_hash_key
> > 0.11 ± 20% -0.0 0.06 ± 14% perf-profile.children.cycles-pp.ip_rcv
> > 0.11 ± 8% -0.0 0.07 ± 7% perf-profile.children.cycles-pp.sctp_sock_rfree
> > 0.24 ± 7% -0.0 0.19 ± 6% perf-profile.children.cycles-pp.sched_clock_cpu
> > 0.12 ± 6% -0.0 0.08 ± 11% perf-profile.children.cycles-pp.__netif_rx
> > 0.16 ± 8% -0.0 0.11 ± 10% perf-profile.children.cycles-pp.ttwu_do_wakeup
> > 0.31 ± 6% -0.0 0.27 ± 5% perf-profile.children.cycles-pp.sctp_ulpevent_receive_data
> > 0.13 ± 7% -0.0 0.09 ± 5% perf-profile.children.cycles-pp.pick_next_entity
> > 0.12 ± 6% -0.0 0.08 ± 7% perf-profile.children.cycles-pp.__check_heap_object
> > 0.12 ± 8% -0.0 0.08 ± 12% perf-profile.children.cycles-pp.netif_rx_internal
> > 0.08 ± 10% -0.0 0.03 ± 70% perf-profile.children.cycles-pp.__wrgsbase_inactive
> > 0.22 ± 8% -0.0 0.18 ± 6% perf-profile.children.cycles-pp.read_tsc
> > 0.14 ± 8% -0.0 0.10 ± 9% perf-profile.children.cycles-pp.check_preempt_curr
> > 0.10 ± 9% -0.0 0.06 ± 7% perf-profile.children.cycles-pp.rcu_all_qs
> > 0.10 ± 10% -0.0 0.06 ± 13% perf-profile.children.cycles-pp.validate_xmit_skb
> > 0.14 ± 8% -0.0 0.10 ± 9% perf-profile.children.cycles-pp.perf_trace_sched_wakeup_template
> > 0.11 ± 6% -0.0 0.07 ± 9% perf-profile.children.cycles-pp.enqueue_to_backlog
> > 0.09 ± 5% -0.0 0.05 ± 45% perf-profile.children.cycles-pp.__netif_receive_skb_core
> > 0.12 ± 5% -0.0 0.08 ± 8% perf-profile.children.cycles-pp.__put_user_nocheck_8
> > 0.12 ± 6% -0.0 0.08 ± 14% perf-profile.children.cycles-pp.sctp_chunk_abandoned
> > 0.12 ± 5% -0.0 0.08 ± 8% perf-profile.children.cycles-pp.skb_put
> > 0.20 ± 7% -0.0 0.17 ± 5% perf-profile.children.cycles-pp.native_sched_clock
> > 0.09 ± 5% -0.0 0.05 ± 46% perf-profile.children.cycles-pp.__mod_lruvec_page_state
> > 0.11 ± 9% -0.0 0.08 ± 6% perf-profile.children.cycles-pp.sctp_inet_skb_msgname
> > 0.10 ± 8% -0.0 0.06 perf-profile.children.cycles-pp.tick_nohz_idle_enter
> > 0.09 ± 7% -0.0 0.06 ± 6% perf-profile.children.cycles-pp.os_xsave
> > 0.12 ± 9% -0.0 0.09 ± 8% perf-profile.children.cycles-pp.resched_curr
> > 0.09 ± 10% -0.0 0.06 ± 9% perf-profile.children.cycles-pp.sctp_ulpevent_init
> > 0.14 ± 10% -0.0 0.11 ± 5% perf-profile.children.cycles-pp.update_cfs_group
> > 0.09 ± 10% -0.0 0.06 ± 8% perf-profile.children.cycles-pp.sctp_ulpq_order
> > 0.08 ± 11% -0.0 0.05 ± 44% perf-profile.children.cycles-pp.perf_tp_event
> > 0.07 ± 5% -0.0 0.04 ± 45% perf-profile.children.cycles-pp._raw_spin_unlock_irqrestore
> > 0.10 ± 8% -0.0 0.07 ± 9% perf-profile.children.cycles-pp.syscall_exit_to_user_mode_prepare
> > 0.07 ± 10% -0.0 0.04 ± 45% perf-profile.children.cycles-pp.__cgroup_account_cputime
> > 0.08 ± 6% -0.0 0.06 ± 8% perf-profile.children.cycles-pp.sctp_tsnmap_check
> > 0.00 +0.1 0.06 ± 9% perf-profile.children.cycles-pp.perf_mux_hrtimer_handler
> > 0.00 +0.1 0.07 ± 26% perf-profile.children.cycles-pp.update_sd_lb_stats
> > 0.00 +0.1 0.08 ± 22% perf-profile.children.cycles-pp.find_busiest_group
> > 0.01 ±223% +0.1 0.09 ± 26% perf-profile.children.cycles-pp.load_balance
> > 0.05 ± 46% +0.1 0.13 ± 4% perf-profile.children.cycles-pp.sctp_get_af_specific
> > 0.00 +0.1 0.09 ± 14% perf-profile.children.cycles-pp.lapic_next_deadline
> > 0.00 +0.1 0.09 ± 13% perf-profile.children.cycles-pp.native_irq_return_iret
> > 0.00 +0.1 0.09 ± 39% perf-profile.children.cycles-pp.tsc_verify_tsc_adjust
> > 0.00 +0.1 0.09 ± 31% perf-profile.children.cycles-pp.calc_global_load_tick
> > 0.00 +0.1 0.10 ± 37% perf-profile.children.cycles-pp.arch_cpu_idle_enter
> > 0.00 +0.1 0.10 ± 31% perf-profile.children.cycles-pp._raw_spin_trylock
> > 0.00 +0.1 0.10 ± 25% perf-profile.children.cycles-pp.arch_scale_freq_tick
> > 0.03 ±100% +0.1 0.15 ± 35% perf-profile.children.cycles-pp.rebalance_domains
> > 0.00 +0.1 0.14 ± 21% perf-profile.children.cycles-pp.cgroup_rstat_updated
> > 0.17 ± 15% +0.1 0.32 ± 19% perf-profile.children.cycles-pp.scheduler_tick
> > 0.00 +0.2 0.16 ± 32% perf-profile.children.cycles-pp.tick_nohz_irq_exit
> > 0.24 ± 11% +0.2 0.42 ± 20% perf-profile.children.cycles-pp.update_process_times
> > 0.24 ± 11% +0.2 0.43 ± 20% perf-profile.children.cycles-pp.tick_sched_handle
> > 0.09 ± 11% +0.2 0.29 ± 28% perf-profile.children.cycles-pp.__irq_exit_rcu
> > 0.53 ± 7% +0.2 0.74 ± 16% perf-profile.children.cycles-pp.ktime_get
> > 0.20 ± 10% +0.2 0.42 ± 21% perf-profile.children.cycles-pp.clockevents_program_event
> > 0.06 ± 13% +0.3 0.31 ± 53% perf-profile.children.cycles-pp.timekeeping_max_deferment
> > 0.00 +0.3 0.26 ± 12% perf-profile.children.cycles-pp.propagate_protected_usage
> > 0.28 ± 8% +0.3 0.54 ± 19% perf-profile.children.cycles-pp.tick_sched_timer
> > 0.55 ± 7% +0.3 0.84 ± 26% perf-profile.children.cycles-pp.menu_select
> > 0.29 ± 7% +0.3 0.58 ± 39% perf-profile.children.cycles-pp.tick_nohz_get_sleep_length
> > 0.20 ± 8% +0.3 0.50 ± 44% perf-profile.children.cycles-pp.tick_nohz_next_event
> > 0.32 ± 6% +0.3 0.66 ± 17% perf-profile.children.cycles-pp.__hrtimer_run_queues
> > 0.00 +0.5 0.52 ± 6% perf-profile.children.cycles-pp.mem_cgroup_uncharge_skmem
> > 0.56 ± 7% +0.6 1.16 ± 17% perf-profile.children.cycles-pp.hrtimer_interrupt
> > 0.56 ± 7% +0.6 1.17 ± 17% perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
> > 0.00 +1.0 0.96 ± 4% perf-profile.children.cycles-pp.__mod_memcg_state
> > 0.72 ± 6% +1.0 1.69 ± 20% perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
> > 0.91 ± 6% +1.1 2.05 ± 16% perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
> > 1.62 ± 9% +1.3 2.96 ± 5% perf-profile.children.cycles-pp.sctp_ulpevent_make_rcvmsg
> > 10.76 ± 7% +1.8 12.59 ± 6% perf-profile.children.cycles-pp.__local_bh_enable_ip
> > 10.64 ± 7% +1.9 12.51 ± 6% perf-profile.children.cycles-pp.do_softirq
> > 10.41 ± 7% +2.0 12.37 ± 6% perf-profile.children.cycles-pp.net_rx_action
> > 10.31 ± 7% +2.0 12.30 ± 6% perf-profile.children.cycles-pp.__napi_poll
> > 10.29 ± 7% +2.0 12.30 ± 6% perf-profile.children.cycles-pp.process_backlog
> > 10.66 ± 7% +2.1 12.73 ± 5% perf-profile.children.cycles-pp.__softirqentry_text_start
> > 10.12 ± 7% +2.1 12.20 ± 6% perf-profile.children.cycles-pp.__netif_receive_skb_one_core
> > 0.00 +2.1 2.10 ± 6% perf-profile.children.cycles-pp.page_counter_uncharge
> > 3.40 ± 7% +2.1 5.54 ± 5% perf-profile.children.cycles-pp.sctp_skb_recv_datagram
> > 0.00 +2.1 2.14 ± 5% perf-profile.children.cycles-pp.drain_stock
> > 9.89 ± 7% +2.2 12.07 ± 6% perf-profile.children.cycles-pp.ip_local_deliver_finish
> > 9.87 ± 7% +2.2 12.06 ± 6% perf-profile.children.cycles-pp.ip_protocol_deliver_rcu
> > 0.00 +2.2 2.20 ± 5% perf-profile.children.cycles-pp.page_counter_try_charge
> > 9.80 ± 7% +2.2 12.00 ± 6% perf-profile.children.cycles-pp.sctp_rcv
> > 0.00 +2.4 2.36 ± 6% perf-profile.children.cycles-pp.refill_stock
> > 15.40 ± 6% +2.8 18.21 ± 6% perf-profile.children.cycles-pp.sctp_recvmsg
> > 1.10 ± 8% +2.8 3.94 ± 6% perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
> > 15.25 ± 6% +2.9 18.12 ± 6% perf-profile.children.cycles-pp.inet_recvmsg
> > 15.36 ± 6% +2.9 18.28 ± 6% perf-profile.children.cycles-pp.release_sock
> > 1.36 ± 7% +3.0 4.35 ± 6% perf-profile.children.cycles-pp.lock_sock_nested
> > 1.25 ± 8% +3.0 4.24 ± 6% perf-profile.children.cycles-pp._raw_spin_lock_bh
> > 15.10 ± 6% +3.0 18.11 ± 6% perf-profile.children.cycles-pp.__release_sock
> > 15.04 ± 6% +3.0 18.07 ± 6% perf-profile.children.cycles-pp.sctp_backlog_rcv
> > 0.62 ± 6% +3.4 4.02 ± 6% perf-profile.children.cycles-pp.sctp_sf_eat_data_6_2
> > 0.00 +3.4 3.41 ± 6% perf-profile.children.cycles-pp.try_charge_memcg
> > 0.49 ± 6% +3.4 3.94 ± 6% perf-profile.children.cycles-pp.sctp_eat_data
> > 19.74 ± 6% +3.7 23.44 ± 6% perf-profile.children.cycles-pp.sctp_assoc_bh_rcv
> > 3.40 ± 7% +3.9 7.34 ± 6% perf-profile.children.cycles-pp.sctp_chunk_put
> > 0.00 +4.0 4.04 ± 6% perf-profile.children.cycles-pp.mem_cgroup_charge_skmem
> > 2.92 ± 6% +4.2 7.16 ± 6% perf-profile.children.cycles-pp.sctp_outq_sack
> > 0.00 +4.3 4.29 ± 6% perf-profile.children.cycles-pp.__sk_mem_raise_allocated
> > 0.00 +4.3 4.32 ± 6% perf-profile.children.cycles-pp.__sk_mem_schedule
> > 1.99 ± 6% +4.4 6.40 ± 6% perf-profile.children.cycles-pp.consume_skb
> > 1.78 ± 6% +4.6 6.42 ± 6% perf-profile.children.cycles-pp.kfree_skb_reason
> > 0.37 ± 8% +5.0 5.40 ± 6% perf-profile.children.cycles-pp.sctp_wfree
> > 0.87 ± 9% +10.3 11.20 ± 6% perf-profile.children.cycles-pp.skb_release_head_state
> > 0.00 +10.7 10.66 ± 6% perf-profile.children.cycles-pp.__sk_mem_reduce_allocated
> > 8.78 ± 6% -3.2 5.58 ± 6% perf-profile.self.cycles-pp.copy_user_enhanced_fast_string
> > 8.70 ± 6% -3.2 5.54 ± 6% perf-profile.self.cycles-pp.memcpy_erms
> > 0.96 ± 6% -0.4 0.57 ± 7% perf-profile.self.cycles-pp.__slab_free
> > 0.83 ± 7% -0.3 0.54 ± 6% perf-profile.self.cycles-pp._raw_spin_lock_irqsave
> > 0.70 ± 9% -0.3 0.41 ± 5% perf-profile.self.cycles-pp.sctp_chunk_put
> > 0.83 ± 4% -0.2 0.59 ± 8% perf-profile.self.cycles-pp.kmem_cache_free
> > 0.60 ± 6% -0.2 0.36 ± 8% perf-profile.self.cycles-pp.sctp_recvmsg
> > 0.64 ± 5% -0.2 0.41 ± 7% perf-profile.self.cycles-pp.__might_resched
> > 0.57 ± 6% -0.2 0.36 ± 8% perf-profile.self.cycles-pp._raw_spin_lock_bh
> > 0.38 ± 7% -0.2 0.16 ± 9% perf-profile.self.cycles-pp.sctp_sendmsg_to_asoc
> > 0.62 ± 6% -0.2 0.40 ± 6% perf-profile.self.cycles-pp._raw_spin_lock
> > 0.59 ± 7% -0.2 0.39 ± 5% perf-profile.self.cycles-pp.__schedule
> > 0.58 ± 6% -0.2 0.39 ± 7% perf-profile.self.cycles-pp.copy_user_short_string
> > 0.50 ± 12% -0.2 0.31 ± 8% perf-profile.self.cycles-pp.sctp_packet_config
> > 0.40 ± 8% -0.2 0.23 ± 9% perf-profile.self.cycles-pp.sctp_packet_pack
> > 0.33 ± 8% -0.2 0.16 ± 6% perf-profile.self.cycles-pp.rmqueue
> > 0.45 ± 3% -0.2 0.29 ± 11% perf-profile.self.cycles-pp.kfree
> > 0.45 ± 8% -0.2 0.29 ± 8% perf-profile.self.cycles-pp.sctp_datamsg_from_user
> > 0.32 ± 5% -0.2 0.16 ± 9% perf-profile.self.cycles-pp.get_page_from_freelist
> > 0.37 ± 7% -0.2 0.21 ± 5% perf-profile.self.cycles-pp.__mod_node_page_state
> > 0.52 ± 7% -0.2 0.36 ± 7% perf-profile.self.cycles-pp.__list_del_entry_valid
> > 0.32 ± 6% -0.2 0.16 ± 8% perf-profile.self.cycles-pp.__zone_watermark_ok
> > 0.41 ± 7% -0.2 0.26 ± 8% perf-profile.self.cycles-pp.kmem_cache_alloc
> > 0.46 ± 17% -0.1 0.31 ± 4% perf-profile.self.cycles-pp.__copy_skb_header
> > 0.35 ± 7% -0.1 0.20 ± 7% perf-profile.self.cycles-pp.update_load_avg
> > 0.39 ± 5% -0.1 0.25 ± 9% perf-profile.self.cycles-pp.sctp_outq_flush_data
> > 0.37 ± 7% -0.1 0.23 ± 5% perf-profile.self.cycles-pp.update_rq_clock
> > 0.37 ± 6% -0.1 0.24 ± 7% perf-profile.self.cycles-pp.__skb_datagram_iter
> > 0.23 ± 13% -0.1 0.10 ± 12% perf-profile.self.cycles-pp.sctp_assoc_rwnd_increase
> > 0.38 ± 6% -0.1 0.25 ± 10% perf-profile.self.cycles-pp.sctp_sendmsg
> > 0.35 ± 5% -0.1 0.22 ± 9% perf-profile.self.cycles-pp.memset_erms
> > 0.37 ± 5% -0.1 0.24 ± 5% perf-profile.self.cycles-pp.set_next_entity
> > 0.37 ± 6% -0.1 0.24 ± 6% perf-profile.self.cycles-pp.skb_release_data
> > 0.36 ± 13% -0.1 0.23 ± 7% perf-profile.self.cycles-pp.entry_SYSCALL_64_after_hwframe
> > 0.35 ± 5% -0.1 0.22 ± 7% perf-profile.self.cycles-pp.check_heap_object
> > 0.31 ± 6% -0.1 0.18 ± 14% perf-profile.self.cycles-pp.__rhashtable_lookup
> > 0.34 ± 7% -0.1 0.22 ± 10% perf-profile.self.cycles-pp.memcg_slab_free_hook
> > 0.34 ± 5% -0.1 0.22 ± 6% perf-profile.self.cycles-pp.free_pcp_prepare
> > 0.32 ± 8% -0.1 0.20 ± 7% perf-profile.self.cycles-pp.__virt_addr_valid
> > 0.25 ± 7% -0.1 0.13 ± 10% perf-profile.self.cycles-pp.free_unref_page_commit
> > 0.39 ± 6% -0.1 0.28 ± 9% perf-profile.self.cycles-pp.enqueue_entity
> > 0.32 ± 6% -0.1 0.20 ± 6% perf-profile.self.cycles-pp.__list_add_valid
> > 0.30 ± 13% -0.1 0.18 ± 11% perf-profile.self.cycles-pp.sctp_v4_xmit
> > 0.32 ± 7% -0.1 0.20 ± 7% perf-profile.self.cycles-pp.__alloc_skb
> > 0.33 ± 5% -0.1 0.22 ± 7% perf-profile.self.cycles-pp.__fdget
> > 0.30 ± 6% -0.1 0.19 ± 7% perf-profile.self.cycles-pp.sctp_association_put
> > 0.31 ± 5% -0.1 0.20 ± 7% perf-profile.self.cycles-pp.__might_sleep
> > 0.21 ± 18% -0.1 0.10 ± 7% perf-profile.self.cycles-pp.dst_release
> > 0.26 ± 13% -0.1 0.15 ± 13% perf-profile.self.cycles-pp.ipv4_dst_check
> > 0.31 ± 6% -0.1 0.21 ± 7% perf-profile.self.cycles-pp.sctp_skb_recv_datagram
> > 0.26 ± 9% -0.1 0.16 ± 13% perf-profile.self.cycles-pp.sctp_cmd_interpreter
> > 0.21 ± 6% -0.1 0.11 ± 12% perf-profile.self.cycles-pp.sock_wfree
> > 0.23 ± 7% -0.1 0.13 ± 6% perf-profile.self.cycles-pp.__alloc_pages
> > 0.31 ± 6% -0.1 0.21 ± 9% perf-profile.self.cycles-pp.kmem_cache_alloc_node
> > 0.24 ± 5% -0.1 0.14 ± 10% perf-profile.self.cycles-pp.__check_object_size
> > 0.17 ± 7% -0.1 0.07 ± 12% perf-profile.self.cycles-pp.update_curr
> > 0.18 ± 36% -0.1 0.08 ± 20% perf-profile.self.cycles-pp.sctp_cmp_addr_exact
> > 0.31 ± 6% -0.1 0.22 ± 9% perf-profile.self.cycles-pp.enqueue_task_fair
> > 0.27 ± 5% -0.1 0.18 ± 7% perf-profile.self.cycles-pp.send_sctp_stream_1toMany
> > 0.27 ± 8% -0.1 0.18 ± 8% perf-profile.self.cycles-pp.entry_SYSRETQ_unsafe_stack
> > 0.29 ± 6% -0.1 0.20 ± 8% perf-profile.self.cycles-pp.skb_set_owner_w
> > 0.26 ± 10% -0.1 0.17 ± 11% perf-profile.self.cycles-pp.sctp_transport_hold
> > 0.21 ± 10% -0.1 0.12 ± 8% perf-profile.self.cycles-pp.recv_sctp_stream_1toMany
> > 0.24 ± 7% -0.1 0.15 ± 6% perf-profile.self.cycles-pp.__switch_to
> > 0.25 ± 12% -0.1 0.17 ± 12% perf-profile.self.cycles-pp.sctp_rcv
> > 0.24 ± 4% -0.1 0.16 ± 4% perf-profile.self.cycles-pp.__entry_text_start
> > 0.28 ± 12% -0.1 0.20 ± 4% perf-profile.self.cycles-pp.sctp_transport_put
> > 0.20 ± 5% -0.1 0.11 ± 6% perf-profile.self.cycles-pp.__wake_up_common
> > 0.22 ± 6% -0.1 0.14 ± 5% perf-profile.self.cycles-pp.__switch_to_asm
> > 0.22 ± 5% -0.1 0.14 ± 9% perf-profile.self.cycles-pp.consume_skb
> > 0.21 ± 7% -0.1 0.13 ± 8% perf-profile.self.cycles-pp.recvmsg
> > 0.20 ± 6% -0.1 0.13 ± 9% perf-profile.self.cycles-pp.reweight_entity
> > 0.21 ± 6% -0.1 0.14 ± 10% perf-profile.self.cycles-pp.sctp_datamsg_put
> > 0.20 ± 4% -0.1 0.13 ± 5% perf-profile.self.cycles-pp.available_idle_cpu
> > 0.20 ± 6% -0.1 0.12 ± 6% perf-profile.self.cycles-pp.__free_pages
> > 0.10 ± 16% -0.1 0.02 ± 99% perf-profile.self.cycles-pp.sctp_ulpq_tail_data
> > 0.16 ± 9% -0.1 0.09 ± 14% perf-profile.self.cycles-pp.____sys_recvmsg
> > 0.28 ± 9% -0.1 0.21 ± 7% perf-profile.self.cycles-pp.aa_sk_perm
> > 0.18 ± 8% -0.1 0.11 ± 13% perf-profile.self.cycles-pp.syscall_return_via_sysret
> > 0.10 ± 9% -0.1 0.03 ±100% perf-profile.self.cycles-pp.process_backlog
> > 0.20 ± 7% -0.1 0.14 ± 7% perf-profile.self.cycles-pp.sctp_association_hold
> > 0.20 ± 7% -0.1 0.13 ± 9% perf-profile.self.cycles-pp.sctp_hash_cmp
> > 0.22 ± 8% -0.1 0.16 ± 4% perf-profile.self.cycles-pp.sctp_wfree
> > 0.18 ± 7% -0.1 0.11 ± 6% perf-profile.self.cycles-pp._copy_from_user
> > 0.17 ± 8% -0.1 0.10 ± 6% perf-profile.self.cycles-pp.poll_idle
> > 0.16 ± 8% -0.1 0.09 ± 14% perf-profile.self.cycles-pp.__update_load_avg_cfs_rq
> > 0.19 ± 6% -0.1 0.13 ± 10% perf-profile.self.cycles-pp.__update_load_avg_se
> > 0.14 ± 7% -0.1 0.08 ± 8% perf-profile.self.cycles-pp.___perf_sw_event
> > 0.16 ± 8% -0.1 0.10 ± 10% perf-profile.self.cycles-pp.try_to_wake_up
> > 0.16 ± 5% -0.1 0.10 ± 5% perf-profile.self.cycles-pp.sctp_packet_append_chunk
> > 0.20 ± 8% -0.1 0.14 ± 5% perf-profile.self.cycles-pp.sendmsg
> > 0.14 ± 12% -0.1 0.08 ± 4% perf-profile.self.cycles-pp.check_new_pages
> > 0.17 ± 11% -0.1 0.11 ± 15% perf-profile.self.cycles-pp.sctp_outq_flush
> > 0.17 ± 8% -0.1 0.11 ± 9% perf-profile.self.cycles-pp.sctp_assoc_bh_rcv
> > 0.17 ± 6% -0.1 0.11 ± 5% perf-profile.self.cycles-pp.nr_iowait_cpu
> > 0.17 ± 7% -0.1 0.12 ± 9% perf-profile.self.cycles-pp.sctp_packet_transmit
> > 0.12 ± 9% -0.1 0.06 ± 14% perf-profile.self.cycles-pp.dequeue_entity
> > 0.15 ± 8% -0.1 0.10 ± 13% perf-profile.self.cycles-pp.____sys_sendmsg
> > 0.15 ± 14% -0.1 0.10 ± 8% perf-profile.self.cycles-pp.sctp_ulpevent_free
> > 0.18 ± 9% -0.1 0.12 ± 10% perf-profile.self.cycles-pp.sctp_do_sm
> > 0.14 ± 4% -0.1 0.09 ± 10% perf-profile.self.cycles-pp.__genradix_ptr
> > 0.13 ± 10% -0.1 0.08 ± 6% perf-profile.self.cycles-pp.__put_user_nocheck_4
> > 0.15 ± 7% -0.1 0.10 ± 7% perf-profile.self.cycles-pp.__copy_msghdr_from_user
> > 0.13 ± 10% -0.1 0.08 ± 12% perf-profile.self.cycles-pp.__import_iovec
> > 0.16 ± 4% -0.0 0.11 ± 11% perf-profile.self.cycles-pp.__skb_clone
> > 0.16 ± 8% -0.0 0.11 ± 13% perf-profile.self.cycles-pp.sctp_sendmsg_parse
> > 0.13 ± 6% -0.0 0.08 ± 8% perf-profile.self.cycles-pp.do_syscall_64
> > 0.17 ± 6% -0.0 0.12 ± 5% perf-profile.self.cycles-pp.do_idle
> > 0.12 ± 17% -0.0 0.07 ± 18% perf-profile.self.cycles-pp.sctp_hash_key
> > 0.14 ± 9% -0.0 0.09 ± 9% perf-profile.self.cycles-pp.sctp_check_transmitted
> > 0.13 ± 5% -0.0 0.08 ± 7% perf-profile.self.cycles-pp.sock_kmalloc
> > 0.11 ± 6% -0.0 0.06 ± 7% perf-profile.self.cycles-pp.sctp_sock_rfree
> > 0.11 ± 6% -0.0 0.07 ± 7% perf-profile.self.cycles-pp.__kmalloc_node_track_caller
> > 0.12 ± 9% -0.0 0.07 ± 10% perf-profile.self.cycles-pp.___sys_recvmsg
> > 0.14 ± 8% -0.0 0.09 ± 7% perf-profile.self.cycles-pp.switch_mm_irqs_off
> > 0.07 ± 10% -0.0 0.03 ±100% perf-profile.self.cycles-pp._sctp_make_chunk
> > 0.22 ± 8% -0.0 0.18 ± 4% perf-profile.self.cycles-pp.read_tsc
> > 0.12 ± 5% -0.0 0.07 ± 6% perf-profile.self.cycles-pp.__put_user_nocheck_8
> > 0.13 ± 8% -0.0 0.08 ± 8% perf-profile.self.cycles-pp.sctp_inq_pop
> > 0.12 ± 6% -0.0 0.08 ± 7% perf-profile.self.cycles-pp.__check_heap_object
> > 0.12 ± 4% -0.0 0.08 ± 8% perf-profile.self.cycles-pp.__kmalloc
> > 0.08 ± 8% -0.0 0.04 ± 71% perf-profile.self.cycles-pp.__mod_lruvec_page_state
> > 0.12 ± 8% -0.0 0.08 ± 6% perf-profile.self.cycles-pp.__local_bh_enable_ip
> > 0.12 ± 7% -0.0 0.08 ± 8% perf-profile.self.cycles-pp.sctp_sf_eat_data_6_2
> > 0.12 ± 6% -0.0 0.07 ± 10% perf-profile.self.cycles-pp.__free_pages_ok
> > 0.11 ± 8% -0.0 0.07 ± 7% perf-profile.self.cycles-pp._copy_to_iter
> > 0.11 ± 6% -0.0 0.06 ± 7% perf-profile.self.cycles-pp.__ip_queue_xmit
> > 0.08 ± 10% -0.0 0.03 ± 70% perf-profile.self.cycles-pp.__wrgsbase_inactive
> > 0.13 ± 8% -0.0 0.09 ± 5% perf-profile.self.cycles-pp.pick_next_entity
> > 0.11 ± 8% -0.0 0.07 ± 7% perf-profile.self.cycles-pp.select_task_rq
> > 0.11 ± 7% -0.0 0.07 ± 8% perf-profile.self.cycles-pp.sctp_inet_skb_msgname
> > 0.10 ± 9% -0.0 0.06 ± 11% perf-profile.self.cycles-pp.sctp_outq_sack
> > 0.08 ± 5% -0.0 0.05 ± 45% perf-profile.self.cycles-pp.__netif_receive_skb_core
> > 0.20 ± 8% -0.0 0.16 ± 5% perf-profile.self.cycles-pp.native_sched_clock
> > 0.12 ± 9% -0.0 0.08 ± 10% perf-profile.self.cycles-pp.memcg_slab_post_alloc_hook
> > 0.10 ± 7% -0.0 0.06 ± 14% perf-profile.self.cycles-pp.__cond_resched
> > 0.12 ± 4% -0.0 0.08 ± 12% perf-profile.self.cycles-pp.kmem_cache_alloc_trace
> > 0.10 ± 5% -0.0 0.06 ± 15% perf-profile.self.cycles-pp.check_stack_object
> > 0.12 ± 6% -0.0 0.08 ± 7% perf-profile.self.cycles-pp.sctp_outq_tail
> > 0.12 ± 8% -0.0 0.08 ± 12% perf-profile.self.cycles-pp.sctp_chunk_abandoned
> > 0.10 ± 4% -0.0 0.06 ± 14% perf-profile.self.cycles-pp.iovec_from_user
> > 0.09 ± 7% -0.0 0.06 ± 6% perf-profile.self.cycles-pp.os_xsave
> > 0.13 ± 7% -0.0 0.09 ± 11% perf-profile.self.cycles-pp.sctp_sched_dequeue_common
> > 0.07 ± 11% -0.0 0.04 ± 71% perf-profile.self.cycles-pp.move_addr_to_kernel
> > 0.09 ± 10% -0.0 0.05 ± 8% perf-profile.self.cycles-pp.sctp_ulpevent_init
> > 0.08 ± 10% -0.0 0.05 ± 45% perf-profile.self.cycles-pp.free_unref_page
> > 0.10 ± 5% -0.0 0.06 ± 11% perf-profile.self.cycles-pp.__might_fault
> > 0.11 ± 12% -0.0 0.08 ± 10% perf-profile.self.cycles-pp.__sctp_packet_append_chunk
> > 0.07 ± 11% -0.0 0.03 ± 70% perf-profile.self.cycles-pp.sockfd_lookup_light
> > 0.11 ± 7% -0.0 0.08 ± 9% perf-profile.self.cycles-pp.cpuidle_idle_call
> > 0.12 ± 7% -0.0 0.09 ± 10% perf-profile.self.cycles-pp.ip_finish_output2
> > 0.12 ± 8% -0.0 0.09 ± 8% perf-profile.self.cycles-pp.resched_curr
> > 0.10 ± 5% -0.0 0.07 ± 10% perf-profile.self.cycles-pp.skb_put
> > 0.09 ± 7% -0.0 0.06 ± 6% perf-profile.self.cycles-pp.sctp_outq_select_transport
> > 0.08 ± 6% -0.0 0.04 ± 45% perf-profile.self.cycles-pp.__sys_recvmsg
> > 0.08 ± 5% -0.0 0.05 ± 8% perf-profile.self.cycles-pp.sock_kfree_s
> > 0.10 ± 5% -0.0 0.07 ± 14% perf-profile.self.cycles-pp.net_rx_action
> > 0.09 ± 8% -0.0 0.06 ± 11% perf-profile.self.cycles-pp.sctp_chunkify
> > 0.10 ± 9% -0.0 0.07 ± 10% perf-profile.self.cycles-pp.syscall_exit_to_user_mode_prepare
> > 0.14 ± 10% -0.0 0.11 ± 5% perf-profile.self.cycles-pp.update_cfs_group
> > 0.09 ± 8% -0.0 0.06 ± 9% perf-profile.self.cycles-pp.security_socket_recvmsg
> > 0.08 ± 9% -0.0 0.05 perf-profile.self.cycles-pp.rcu_all_qs
> > 0.07 ± 9% -0.0 0.04 ± 45% perf-profile.self.cycles-pp.sctp_tsnmap_check
> > 0.08 ± 11% -0.0 0.06 ± 13% perf-profile.self.cycles-pp.sctp_addto_chunk
> > 0.08 ± 10% -0.0 0.06 ± 13% perf-profile.self.cycles-pp.kmalloc_large_node
> > 0.10 ± 8% +0.0 0.12 ± 6% perf-profile.self.cycles-pp.lock_sock_nested
> > 0.00 +0.1 0.06 ± 9% perf-profile.self.cycles-pp.mem_cgroup_uncharge_skmem
> > 0.00 +0.1 0.08 ± 11% perf-profile.self.cycles-pp.lapic_next_deadline
> > 0.00 +0.1 0.09 ± 13% perf-profile.self.cycles-pp.native_irq_return_iret
> > 0.04 ± 71% +0.1 0.12 ± 7% perf-profile.self.cycles-pp.sctp_get_af_specific
> > 0.00 +0.1 0.09 ± 41% perf-profile.self.cycles-pp.tsc_verify_tsc_adjust
> > 0.00 +0.1 0.09 ± 31% perf-profile.self.cycles-pp.calc_global_load_tick
> > 0.00 +0.1 0.10 ± 31% perf-profile.self.cycles-pp._raw_spin_trylock
> > 0.01 ±223% +0.1 0.10 ± 50% perf-profile.self.cycles-pp.tick_nohz_next_event
> > 0.11 ± 9% +0.1 0.21 ± 10% perf-profile.self.cycles-pp.cpuidle_enter_state
> > 0.00 +0.1 0.10 ± 25% perf-profile.self.cycles-pp.arch_scale_freq_tick
> > 0.00 +0.1 0.13 ± 19% perf-profile.self.cycles-pp.cgroup_rstat_updated
> > 0.00 +0.1 0.14 ± 10% perf-profile.self.cycles-pp.mem_cgroup_charge_skmem
> > 0.00 +0.2 0.21 ± 8% perf-profile.self.cycles-pp.refill_stock
> > 0.00 +0.2 0.25 ± 9% perf-profile.self.cycles-pp.__sk_mem_raise_allocated
> > 0.31 ± 8% +0.3 0.56 ± 23% perf-profile.self.cycles-pp.ktime_get
> > 0.06 ± 13% +0.3 0.31 ± 53% perf-profile.self.cycles-pp.timekeeping_max_deferment
> > 0.00 +0.3 0.26 ± 11% perf-profile.self.cycles-pp.propagate_protected_usage
> > 0.00 +0.8 0.82 ± 3% perf-profile.self.cycles-pp.__mod_memcg_state
> > 0.00 +1.2 1.19 ± 7% perf-profile.self.cycles-pp.try_charge_memcg
> > 0.00 +2.0 1.96 ± 6% perf-profile.self.cycles-pp.page_counter_uncharge
> > 0.00 +2.1 2.07 ± 5% perf-profile.self.cycles-pp.page_counter_try_charge
> > 1.09 ± 8% +2.8 3.92 ± 6% perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
> > 0.29 ± 6% +3.5 3.81 ± 6% perf-profile.self.cycles-pp.sctp_eat_data
> > 0.00 +7.8 7.76 ± 6% perf-profile.self.cycles-pp.__sk_mem_reduce_allocated
> >
> >
> > ***************************************************************************************************
> > lkp-cpl-4sp1: 144 threads 4 sockets Intel(R) Xeon(R) Gold 5318H CPU @ 2.50GHz with 128G memory
> > =========================================================================================
> > cluster/compiler/cpufreq_governor/ip/kconfig/nr_threads/rootfs/runtime/send_size/tbox_group/test/testcase/ucode:
> > cs-localhost/gcc-11/performance/ipv4/x86_64-rhel-8.3/50%/debian-11.1-x86_64-20220510.cgz/300s/10K/lkp-cpl-4sp1/SCTP_STREAM_MANY/netperf/0x7002402
> >
> > commit:
> > 7c80b038d2 ("net: fix sk_wmem_schedule() and sk_rmem_schedule() errors")
> > 4890b686f4 ("net: keep sk->sk_forward_alloc as small as possible")
> >
> > 7c80b038d23e1f4c 4890b686f4088c90432149bd6de
> > ---------------- ---------------------------
> > %stddev %change %stddev
> > \ | \
> > 9985 -73.7% 2624 netperf.Throughput_Mbps
> > 718950 -73.7% 188942 netperf.Throughput_total_Mbps
> > 68186 -71.8% 19239 ± 2% netperf.time.involuntary_context_switches
> > 29139 ± 2% -16.8% 24247 netperf.time.minor_page_faults
> > 3137 -68.1% 1000 netperf.time.percent_of_cpu_this_job_got
> > 9264 -67.8% 2979 netperf.time.system_time
> > 199.50 -77.7% 44.44 ± 2% netperf.time.user_time
> > 316769 ± 2% +21.5% 384840 netperf.time.voluntary_context_switches
> > 2.633e+09 -73.7% 6.919e+08 netperf.workload
> > 29358 +42.4% 41821 uptime.idle
> > 2.381e+10 +52.8% 3.639e+10 cpuidle..time
> > 7.884e+08 -65.8% 2.694e+08 cpuidle..usage
> > 56.10 +27.9 83.96 mpstat.cpu.all.idle%
> > 0.80 -0.2 0.61 ± 5% mpstat.cpu.all.irq%
> > 7.50 -4.2 3.29 mpstat.cpu.all.soft%
> > 34.45 -22.6 11.86 mpstat.cpu.all.sys%
> > 1.14 -0.9 0.27 mpstat.cpu.all.usr%
> > 55.83 +49.3% 83.33 vmstat.cpu.id
> > 5727924 ± 3% -43.6% 3229130 vmstat.memory.cache
> > 64.67 ± 2% -65.2% 22.50 ± 2% vmstat.procs.r
> > 4915230 -73.6% 1295498 vmstat.system.cs
> > 298596 -3.1% 289395 vmstat.system.in
> > 92833 ±101% -63.0% 34380 ± 35% numa-meminfo.node1.Inactive
> > 92773 ±101% -62.9% 34380 ± 35% numa-meminfo.node1.Inactive(anon)
> > 1468231 ± 19% -74.4% 376377 ± 16% numa-meminfo.node3.Active
> > 1468231 ± 19% -74.4% 376377 ± 16% numa-meminfo.node3.Active(anon)
> > 925228 ± 17% -95.4% 42887 ± 27% numa-meminfo.node3.Inactive
> > 925228 ± 17% -95.4% 42887 ± 27% numa-meminfo.node3.Inactive(anon)
> > 1162413 ± 17% -85.6% 167082 ± 21% numa-meminfo.node3.Mapped
> > 2348690 ± 14% -83.8% 380889 ± 16% numa-meminfo.node3.Shmem
> > 8.422e+08 -72.9% 2.286e+08 ± 5% numa-numastat.node0.local_node
> > 8.425e+08 -72.8% 2.288e+08 ± 5% numa-numastat.node0.numa_hit
> > 8.344e+08 -73.3% 2.226e+08 ± 5% numa-numastat.node1.local_node
> > 8.341e+08 -73.3% 2.227e+08 ± 5% numa-numastat.node1.numa_hit
> > 8.584e+08 -74.8% 2.163e+08 ± 10% numa-numastat.node2.local_node
> > 8.583e+08 -74.8% 2.164e+08 ± 10% numa-numastat.node2.numa_hit
> > 8.684e+08 -74.1% 2.251e+08 ± 6% numa-numastat.node3.local_node
> > 8.681e+08 -74.1% 2.252e+08 ± 6% numa-numastat.node3.numa_hit
> > 1860369 ± 5% -76.2% 442423 ± 3% meminfo.Active
> > 1860257 ± 5% -76.2% 442314 ± 3% meminfo.Active(anon)
> > 5599254 ± 3% -44.7% 3099170 meminfo.Cached
> > 3542743 ± 6% -71.5% 1010644 meminfo.Committed_AS
> > 1423628 ± 11% -76.4% 336110 meminfo.Inactive
> > 1423448 ± 11% -76.4% 335928 meminfo.Inactive(anon)
> > 1453354 ± 6% -83.5% 240203 ± 3% meminfo.Mapped
> > 7494321 ± 2% -34.5% 4912267 meminfo.Memused
> > 12458 -18.9% 10104 meminfo.PageTables
> > 2948963 ± 7% -84.8% 449291 ± 3% meminfo.Shmem
> > 11063817 -30.5% 7685444 meminfo.max_used_kB
> > 8.425e+08 -72.8% 2.288e+08 ± 5% numa-vmstat.node0.numa_hit
> > 8.422e+08 -72.9% 2.286e+08 ± 5% numa-vmstat.node0.numa_local
> > 23213 ±101% -63.0% 8583 ± 35% numa-vmstat.node1.nr_inactive_anon
> > 23213 ±101% -63.0% 8583 ± 35% numa-vmstat.node1.nr_zone_inactive_anon
> > 8.341e+08 -73.3% 2.227e+08 ± 5% numa-vmstat.node1.numa_hit
> > 8.344e+08 -73.3% 2.226e+08 ± 5% numa-vmstat.node1.numa_local
> > 8.583e+08 -74.8% 2.164e+08 ± 10% numa-vmstat.node2.numa_hit
> > 8.584e+08 -74.8% 2.163e+08 ± 10% numa-vmstat.node2.numa_local
> > 366556 ± 19% -74.2% 94616 ± 16% numa-vmstat.node3.nr_active_anon
> > 231387 ± 17% -95.4% 10658 ± 27% numa-vmstat.node3.nr_inactive_anon
> > 291532 ± 17% -86.8% 38558 ± 21% numa-vmstat.node3.nr_mapped
> > 586771 ± 14% -83.7% 95702 ± 16% numa-vmstat.node3.nr_shmem
> > 366556 ± 19% -74.2% 94616 ± 16% numa-vmstat.node3.nr_zone_active_anon
> > 231387 ± 17% -95.4% 10658 ± 27% numa-vmstat.node3.nr_zone_inactive_anon
> > 8.681e+08 -74.1% 2.252e+08 ± 6% numa-vmstat.node3.numa_hit
> > 8.684e+08 -74.1% 2.251e+08 ± 6% numa-vmstat.node3.numa_local
> > 1495 -61.8% 571.00 turbostat.Avg_MHz
> > 47.16 -29.8 17.36 turbostat.Busy%
> > 3178 +3.7% 3295 turbostat.Bzy_MHz
> > 7.464e+08 -73.4% 1.985e+08 turbostat.C1
> > 13.74 -8.9 4.80 ± 2% turbostat.C1%
> > 34154266 ± 14% +97.4% 67405025 turbostat.C1E
> > 36.80 ± 21% +37.1 73.90 ± 3% turbostat.C1E%
> > 52.81 +56.4% 82.60 turbostat.CPU%c1
> > 58.83 ± 2% -20.1% 47.00 ± 2% turbostat.CoreTmp
> > 0.20 -25.6% 0.15 turbostat.IPC
> > 128.56 ± 3% -128.6 0.00 turbostat.PKG_%
> > 5447540 ± 3% -83.7% 887322 turbostat.POLL
> > 0.13 -0.1 0.03 turbostat.POLL%
> > 58.83 ± 2% -19.8% 47.17 ± 2% turbostat.PkgTmp
> > 591.77 -28.1% 425.61 turbostat.PkgWatt
> > 8.17 +3.5% 8.46 turbostat.RAMWatt
> > 465173 ± 5% -76.3% 110436 ± 3% proc-vmstat.nr_active_anon
> > 83691 -1.7% 82260 proc-vmstat.nr_anon_pages
> > 3086038 +2.1% 3150528 proc-vmstat.nr_dirty_background_threshold
> > 6179622 +2.1% 6308761 proc-vmstat.nr_dirty_threshold
> > 1399831 ± 3% -44.7% 774653 proc-vmstat.nr_file_pages
> > 31038779 +2.1% 31684629 proc-vmstat.nr_free_pages
> > 355768 ± 11% -76.4% 83981 proc-vmstat.nr_inactive_anon
> > 363110 ± 6% -83.4% 60381 ± 3% proc-vmstat.nr_mapped
> > 3114 -18.9% 2526 proc-vmstat.nr_page_table_pages
> > 737257 ± 7% -84.8% 112182 ± 3% proc-vmstat.nr_shmem
> > 33546 -3.9% 32233 proc-vmstat.nr_slab_reclaimable
> > 465173 ± 5% -76.3% 110436 ± 3% proc-vmstat.nr_zone_active_anon
> > 355768 ± 11% -76.4% 83981 proc-vmstat.nr_zone_inactive_anon
> > 125868 ± 15% -29.5% 88747 ± 12% proc-vmstat.numa_hint_faults_local
> > 3.403e+09 -73.8% 8.93e+08 proc-vmstat.numa_hit
> > 3.403e+09 -73.8% 8.927e+08 proc-vmstat.numa_local
> > 1698265 ± 8% -91.3% 146940 ± 2% proc-vmstat.pgactivate
> > 9.031e+09 -73.7% 2.375e+09 proc-vmstat.pgalloc_normal
> > 1809374 ± 2% -11.5% 1601059 ± 2% proc-vmstat.pgfault
> > 9.031e+09 -73.7% 2.375e+09 proc-vmstat.pgfree
> > 40961 +1.3% 41503 proc-vmstat.pgreuse
> > 121388 ± 29% -96.2% 4632 ± 47% sched_debug.cfs_rq:/.MIN_vruntime.avg
> > 3776647 ± 10% -91.9% 305894 ± 40% sched_debug.cfs_rq:/.MIN_vruntime.max
> > 623786 ± 18% -94.3% 35468 ± 44% sched_debug.cfs_rq:/.MIN_vruntime.stddev
> > 0.40 ± 7% -58.4% 0.17 ± 7% sched_debug.cfs_rq:/.h_nr_running.avg
> > 0.45 -19.1% 0.37 ± 3% sched_debug.cfs_rq:/.h_nr_running.stddev
> > 121388 ± 29% -96.2% 4632 ± 47% sched_debug.cfs_rq:/.max_vruntime.avg
> > 3776647 ± 10% -91.9% 305894 ± 40% sched_debug.cfs_rq:/.max_vruntime.max
> > 623786 ± 18% -94.3% 35468 ± 44% sched_debug.cfs_rq:/.max_vruntime.stddev
> > 3455794 ± 6% -89.2% 374393 ± 9% sched_debug.cfs_rq:/.min_vruntime.avg
> > 4153720 ± 7% -88.1% 494762 ± 7% sched_debug.cfs_rq:/.min_vruntime.max
> > 2710531 ± 7% -91.1% 240412 ± 10% sched_debug.cfs_rq:/.min_vruntime.min
> > 368488 ± 8% -85.9% 51911 ± 13% sched_debug.cfs_rq:/.min_vruntime.stddev
> > 0.40 ± 7% -58.4% 0.17 ± 7% sched_debug.cfs_rq:/.nr_running.avg
> > 0.45 -19.2% 0.37 ± 3% sched_debug.cfs_rq:/.nr_running.stddev
> > 386.20 ± 3% -56.3% 168.94 sched_debug.cfs_rq:/.runnable_avg.avg
> > 1215 ± 8% -18.7% 987.82 ± 4% sched_debug.cfs_rq:/.runnable_avg.max
> > 340.61 -31.1% 234.69 ± 2% sched_debug.cfs_rq:/.runnable_avg.stddev
> > 429174 ± 24% -103.8% -16476 sched_debug.cfs_rq:/.spread0.avg
> > 1127012 ± 15% -90.8% 103881 ± 28% sched_debug.cfs_rq:/.spread0.max
> > -315952 -52.4% -150458 sched_debug.cfs_rq:/.spread0.min
> > 368407 ± 8% -85.9% 51908 ± 13% sched_debug.cfs_rq:/.spread0.stddev
> > 385.96 ± 3% -56.3% 168.81 sched_debug.cfs_rq:/.util_avg.avg
> > 1215 ± 8% -18.7% 987.79 ± 4% sched_debug.cfs_rq:/.util_avg.max
> > 340.53 -31.1% 234.60 ± 2% sched_debug.cfs_rq:/.util_avg.stddev
> > 274.31 ± 7% -74.9% 68.97 ± 5% sched_debug.cfs_rq:/.util_est_enqueued.avg
> > 981.12 -10.6% 877.51 ± 3% sched_debug.cfs_rq:/.util_est_enqueued.max
> > 337.06 -47.8% 176.03 ± 3% sched_debug.cfs_rq:/.util_est_enqueued.stddev
> > 457832 ± 5% +41.2% 646422 ± 5% sched_debug.cpu.avg_idle.avg
> > 1762 ± 7% -42.8% 1009 ± 7% sched_debug.cpu.clock_task.stddev
> > 2358 ± 5% -61.0% 920.92 ± 2% sched_debug.cpu.curr->pid.avg
> > 2793 -22.1% 2176 ± 2% sched_debug.cpu.curr->pid.stddev
> > 557670 ± 6% +8.4% 604409 ± 7% sched_debug.cpu.max_idle_balance_cost.max
> > 0.00 ± 2% -27.5% 0.00 ± 5% sched_debug.cpu.next_balance.stddev
> > 0.39 ± 5% -61.1% 0.15 ± 3% sched_debug.cpu.nr_running.avg
> > 0.45 -22.2% 0.35 sched_debug.cpu.nr_running.stddev
> > 5005510 ± 7% -75.6% 1221805 ± 10% sched_debug.cpu.nr_switches.avg
> > 6928957 ± 8% -70.7% 2030568 ± 9% sched_debug.cpu.nr_switches.max
> > 3447354 ± 12% -84.3% 541402 ± 17% sched_debug.cpu.nr_switches.min
> > 629520 ± 7% -55.8% 278525 ± 6% sched_debug.cpu.nr_switches.stddev
> > 3.074e+10 -71.6% 8.745e+09 perf-stat.i.branch-instructions
> > 2.971e+08 ± 2% -70.3% 88205513 ± 4% perf-stat.i.branch-misses
> > 1.05 ± 3% +5.6 6.66 perf-stat.i.cache-miss-rate%
> > 30824055 ± 3% +98.8% 61267408 perf-stat.i.cache-misses
> > 3.297e+09 -71.8% 9.296e+08 perf-stat.i.cache-references
> > 4957450 -73.6% 1308421 perf-stat.i.context-switches
> > 1.43 +33.1% 1.90 perf-stat.i.cpi
> > 2.196e+11 -62.4% 8.259e+10 perf-stat.i.cpu-cycles
> > 929.67 -64.3% 331.55 perf-stat.i.cpu-migrations
> > 7494 ± 4% -82.0% 1351 perf-stat.i.cycles-between-cache-misses
> > 0.01 ± 26% -0.0 0.00 perf-stat.i.dTLB-load-miss-rate%
> > 2694006 ± 27% -90.1% 265691 perf-stat.i.dTLB-load-misses
> > 4.398e+10 -71.5% 1.252e+10 perf-stat.i.dTLB-loads
> > 0.00 ± 7% -0.0 0.00 ± 5% perf-stat.i.dTLB-store-miss-rate%
> > 967059 ± 8% -84.3% 151835 ± 6% perf-stat.i.dTLB-store-misses
> > 2.599e+10 -71.6% 7.377e+09 perf-stat.i.dTLB-stores
> > 63.87 +10.9 74.76 perf-stat.i.iTLB-load-miss-rate%
> > 1.972e+08 -70.5% 58088811 ± 5% perf-stat.i.iTLB-load-misses
> > 1.126e+08 -82.7% 19440392 perf-stat.i.iTLB-loads
> > 1.538e+11 -71.5% 4.38e+10 perf-stat.i.instructions
> > 0.71 -24.8% 0.53 perf-stat.i.ipc
> > 1.53 -62.4% 0.57 perf-stat.i.metric.GHz
> > 909.37 -69.9% 273.61 perf-stat.i.metric.K/sec
> > 722.30 -71.6% 205.33 perf-stat.i.metric.M/sec
> > 5562 ± 3% -12.4% 4874 ± 2% perf-stat.i.minor-faults
> > 8844727 ± 5% +30.4% 11534796 perf-stat.i.node-load-misses
> > 623949 ± 11% +54.9% 966213 ± 2% perf-stat.i.node-loads
> > 91.43 +5.4 96.84 perf-stat.i.node-store-miss-rate%
> > 3461624 ± 2% +66.0% 5746681 perf-stat.i.node-store-misses
> > 410920 ± 7% -49.7% 206502 ± 3% perf-stat.i.node-stores
> > 5563 ± 3% -12.4% 4876 ± 2% perf-stat.i.page-faults
> > 21.44 -1.0% 21.22 perf-stat.overall.MPKI
> > 0.93 ± 4% +5.7 6.59 perf-stat.overall.cache-miss-rate%
> > 1.43 +32.0% 1.89 perf-stat.overall.cpi
> > 7136 ± 4% -81.1% 1348 perf-stat.overall.cycles-between-cache-misses
> > 0.01 ± 27% -0.0 0.00 perf-stat.overall.dTLB-load-miss-rate%
> > 0.00 ± 7% -0.0 0.00 ± 5% perf-stat.overall.dTLB-store-miss-rate%
> > 63.66 +11.2 74.89 perf-stat.overall.iTLB-load-miss-rate%
> > 0.70 -24.3% 0.53 perf-stat.overall.ipc
> > 89.40 +7.1 96.53 perf-stat.overall.node-store-miss-rate%
> > 17593 +8.6% 19108 perf-stat.overall.path-length
> > 3.063e+10 -71.5% 8.716e+09 perf-stat.ps.branch-instructions
> > 2.961e+08 ± 2% -70.3% 87907848 ± 4% perf-stat.ps.branch-misses
> > 30717141 ± 3% +98.8% 61062871 perf-stat.ps.cache-misses
> > 3.286e+09 -71.8% 9.265e+08 perf-stat.ps.cache-references
> > 4940513 -73.6% 1304046 perf-stat.ps.context-switches
> > 2.189e+11 -62.4% 8.231e+10 perf-stat.ps.cpu-cycles
> > 926.60 -64.3% 330.61 perf-stat.ps.cpu-migrations
> > 2684405 ± 27% -90.1% 264976 perf-stat.ps.dTLB-load-misses
> > 4.383e+10 -71.5% 1.248e+10 perf-stat.ps.dTLB-loads
> > 963689 ± 8% -84.3% 151363 ± 6% perf-stat.ps.dTLB-store-misses
> > 2.591e+10 -71.6% 7.352e+09 perf-stat.ps.dTLB-stores
> > 1.965e+08 -70.5% 57894987 ± 5% perf-stat.ps.iTLB-load-misses
> > 1.122e+08 -82.7% 19375491 perf-stat.ps.iTLB-loads
> > 1.532e+11 -71.5% 4.366e+10 perf-stat.ps.instructions
> > 5536 ± 3% -12.3% 4857 ± 2% perf-stat.ps.minor-faults
> > 8813664 ± 5% +30.4% 11496729 perf-stat.ps.node-load-misses
> > 622201 ± 11% +54.8% 963024 ± 2% perf-stat.ps.node-loads
> > 3450020 ± 2% +66.0% 5727738 perf-stat.ps.node-store-misses
> > 409584 ± 8% -49.7% 206027 ± 3% perf-stat.ps.node-stores
> > 5537 ± 3% -12.3% 4858 ± 2% perf-stat.ps.page-faults
> > 4.632e+13 -71.5% 1.322e+13 perf-stat.total.instructions
> > 9.54 ± 4% -6.1 3.40 ± 5% perf-profile.calltrace.cycles-pp.intel_idle_irq.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
> > 9.17 ± 5% -6.1 3.05 ± 6% perf-profile.calltrace.cycles-pp.mwait_idle_with_hints.intel_idle_irq.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call
> > 8.84 ± 4% -3.2 5.62 ± 4% perf-profile.calltrace.cycles-pp.sctp_do_sm.sctp_primitive_SEND.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg
> > 8.63 ± 4% -3.1 5.53 ± 4% perf-profile.calltrace.cycles-pp.sctp_packet_pack.sctp_packet_transmit.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm
> > 8.18 ± 4% -2.9 5.29 ± 4% perf-profile.calltrace.cycles-pp.memcpy_erms.sctp_packet_pack.sctp_packet_transmit.sctp_outq_flush.sctp_cmd_interpreter
> > 6.49 ± 4% -2.5 3.99 ± 4% perf-profile.calltrace.cycles-pp.skb_copy_datagram_iter.sctp_recvmsg.inet_recvmsg.____sys_recvmsg.___sys_recvmsg
> > 6.47 ± 3% -2.5 3.98 ± 4% perf-profile.calltrace.cycles-pp.__skb_datagram_iter.skb_copy_datagram_iter.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
> > 21.33 ± 3% -2.5 18.86 ± 4% perf-profile.calltrace.cycles-pp.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg.____sys_sendmsg.___sys_sendmsg
> > 13.50 ± 3% -2.5 11.04 ± 4% perf-profile.calltrace.cycles-pp.sctp_primitive_SEND.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg.____sys_sendmsg
> > 6.89 ± 3% -2.4 4.50 ± 4% perf-profile.calltrace.cycles-pp.sctp_datamsg_from_user.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg.____sys_sendmsg
> > 5.63 ± 3% -2.1 3.49 ± 4% perf-profile.calltrace.cycles-pp._copy_to_iter.__skb_datagram_iter.skb_copy_datagram_iter.sctp_recvmsg.inet_recvmsg
> > 5.46 ± 3% -2.1 3.40 ± 4% perf-profile.calltrace.cycles-pp.copyout._copy_to_iter.__skb_datagram_iter.skb_copy_datagram_iter.sctp_recvmsg
> > 5.42 ± 3% -2.0 3.38 ± 4% perf-profile.calltrace.cycles-pp.copy_user_enhanced_fast_string.copyout._copy_to_iter.__skb_datagram_iter.skb_copy_datagram_iter
> > 11.96 ± 3% -1.9 10.08 ± 4% perf-profile.calltrace.cycles-pp.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm.sctp_primitive_SEND.sctp_sendmsg_to_asoc
> > 9.08 ± 4% -1.6 7.48 ± 4% perf-profile.calltrace.cycles-pp.sctp_cmd_interpreter.sctp_do_sm.sctp_primitive_SEND.sctp_sendmsg_to_asoc.sctp_sendmsg
> > 1.76 ± 4% -1.3 0.43 ± 44% perf-profile.calltrace.cycles-pp.__alloc_pages.kmalloc_large_node.__kmalloc_node_track_caller.kmalloc_reserve.__alloc_skb
> > 10.30 ± 4% -1.2 9.11 ± 4% perf-profile.calltrace.cycles-pp.sctp_packet_transmit.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm.sctp_primitive_SEND
> > 3.42 ± 3% -1.1 2.30 ± 4% perf-profile.calltrace.cycles-pp.sctp_user_addto_chunk.sctp_datamsg_from_user.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg
> > 2.85 ± 3% -1.1 1.78 ± 4% perf-profile.calltrace.cycles-pp.sctp_make_datafrag_empty.sctp_datamsg_from_user.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg
> > 1.79 ± 4% -1.0 0.75 ± 4% perf-profile.calltrace.cycles-pp.skb_release_data.kfree_skb_reason.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
> > 3.08 ± 4% -1.0 2.07 ± 4% perf-profile.calltrace.cycles-pp.sctp_data_ready.sctp_ulpq_tail_event.sctp_ulpq_tail_data.sctp_cmd_interpreter.sctp_do_sm
> > 3.08 ± 3% -1.0 2.08 ± 4% perf-profile.calltrace.cycles-pp._copy_from_iter.sctp_user_addto_chunk.sctp_datamsg_from_user.sctp_sendmsg_to_asoc.sctp_sendmsg
> > 3.20 ± 4% -1.0 2.20 ± 4% perf-profile.calltrace.cycles-pp.sctp_ulpq_tail_event.sctp_ulpq_tail_data.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv
> > 2.82 ± 4% -1.0 1.84 ± 4% perf-profile.calltrace.cycles-pp.__wake_up_common_lock.sctp_data_ready.sctp_ulpq_tail_event.sctp_ulpq_tail_data.sctp_cmd_interpreter
> > 2.94 ± 3% -1.0 1.97 ± 4% perf-profile.calltrace.cycles-pp.copyin._copy_from_iter.sctp_user_addto_chunk.sctp_datamsg_from_user.sctp_sendmsg_to_asoc
> > 2.42 ± 3% -1.0 1.46 ± 4% perf-profile.calltrace.cycles-pp._sctp_make_chunk.sctp_make_datafrag_empty.sctp_datamsg_from_user.sctp_sendmsg_to_asoc.sctp_sendmsg
> > 2.90 ± 3% -0.9 1.95 ± 4% perf-profile.calltrace.cycles-pp.copy_user_enhanced_fast_string.copyin._copy_from_iter.sctp_user_addto_chunk.sctp_datamsg_from_user
> > 11.68 ± 3% -0.9 10.75 ± 4% perf-profile.calltrace.cycles-pp.sctp_packet_transmit.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv
> > 2.63 ± 4% -0.9 1.70 ± 4% perf-profile.calltrace.cycles-pp.__wake_up_common.__wake_up_common_lock.sctp_data_ready.sctp_ulpq_tail_event.sctp_ulpq_tail_data
> > 2.46 ± 4% -0.9 1.59 ± 4% perf-profile.calltrace.cycles-pp.autoremove_wake_function.__wake_up_common.__wake_up_common_lock.sctp_data_ready.sctp_ulpq_tail_event
> > 2.44 ± 4% -0.9 1.58 ± 4% perf-profile.calltrace.cycles-pp.try_to_wake_up.autoremove_wake_function.__wake_up_common.__wake_up_common_lock.sctp_data_ready
> > 11.49 ± 3% -0.8 10.64 ± 4% perf-profile.calltrace.cycles-pp.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv.sctp_backlog_rcv
> > 1.82 ± 3% -0.8 1.02 ± 4% perf-profile.calltrace.cycles-pp.sctp_ulpevent_free.sctp_recvmsg.inet_recvmsg.____sys_recvmsg.___sys_recvmsg
> > 1.86 ± 4% -0.8 1.08 ± 4% perf-profile.calltrace.cycles-pp.__alloc_skb._sctp_make_chunk.sctp_make_datafrag_empty.sctp_datamsg_from_user.sctp_sendmsg_to_asoc
> > 1.88 ± 3% -0.8 1.10 ± 5% perf-profile.calltrace.cycles-pp.schedule_timeout.sctp_skb_recv_datagram.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
> > 1.84 ± 3% -0.8 1.09 ± 5% perf-profile.calltrace.cycles-pp.schedule.schedule_timeout.sctp_skb_recv_datagram.sctp_recvmsg.inet_recvmsg
> > 1.81 ± 3% -0.7 1.07 ± 5% perf-profile.calltrace.cycles-pp.__schedule.schedule.schedule_timeout.sctp_skb_recv_datagram.sctp_recvmsg
> > 1.68 ± 4% -0.7 0.98 ± 6% perf-profile.calltrace.cycles-pp.sctp_outq_flush_data.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm.sctp_primitive_SEND
> > 1.65 ± 4% -0.7 0.99 ± 3% perf-profile.calltrace.cycles-pp.schedule_idle.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify
> > 1.61 ± 4% -0.6 0.96 ± 3% perf-profile.calltrace.cycles-pp.__schedule.schedule_idle.do_idle.cpu_startup_entry.start_secondary
> > 1.30 ± 3% -0.6 0.69 ± 3% perf-profile.calltrace.cycles-pp.kmalloc_reserve.__alloc_skb._sctp_make_chunk.sctp_make_datafrag_empty.sctp_datamsg_from_user
> > 1.27 ± 3% -0.6 0.67 ± 3% perf-profile.calltrace.cycles-pp.__kmalloc_node_track_caller.kmalloc_reserve.__alloc_skb._sctp_make_chunk.sctp_make_datafrag_empty
> > 1.24 ± 3% -0.6 0.65 ± 3% perf-profile.calltrace.cycles-pp.kmalloc_large_node.__kmalloc_node_track_caller.kmalloc_reserve.__alloc_skb._sctp_make_chunk
> > 0.97 ± 3% -0.5 0.46 ± 44% perf-profile.calltrace.cycles-pp.dequeue_task_fair.__schedule.schedule.schedule_timeout.sctp_skb_recv_datagram
> > 1.02 ± 3% -0.5 0.57 ± 3% perf-profile.calltrace.cycles-pp.sctp_chunk_put.sctp_ulpevent_free.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
> > 1.10 ± 4% -0.4 0.67 ± 5% perf-profile.calltrace.cycles-pp.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv.sctp_rcv
> > 0.92 ± 3% -0.4 0.56 ± 6% perf-profile.calltrace.cycles-pp.skb_release_data.consume_skb.sctp_chunk_put.sctp_outq_sack.sctp_cmd_interpreter
> > 1.04 ± 4% -0.3 0.71 ± 4% perf-profile.calltrace.cycles-pp.sendmsg_copy_msghdr.___sys_sendmsg.__sys_sendmsg.do_syscall_64.entry_SYSCALL_64_after_hwframe
> > 0.17 ±141% +0.5 0.70 ± 15% perf-profile.calltrace.cycles-pp.menu_select.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary
> > 0.00 +0.9 0.91 ± 5% perf-profile.calltrace.cycles-pp.page_counter_uncharge.drain_stock.refill_stock.__sk_mem_reduce_allocated.skb_release_head_state
> > 0.00 +0.9 0.93 ± 5% perf-profile.calltrace.cycles-pp.drain_stock.refill_stock.__sk_mem_reduce_allocated.skb_release_head_state.kfree_skb_reason
> > 0.00 +1.0 0.95 ± 15% perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call
> > 0.00 +1.0 1.03 ± 5% perf-profile.calltrace.cycles-pp.refill_stock.__sk_mem_reduce_allocated.skb_release_head_state.kfree_skb_reason.sctp_recvmsg
> > 0.00 +1.0 1.04 ± 14% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
> > 11.57 ± 3% +1.3 12.82 ± 4% perf-profile.calltrace.cycles-pp.sctp_assoc_bh_rcv.sctp_backlog_rcv.__release_sock.release_sock.sctp_sendmsg
> > 0.00 +1.3 1.33 ± 4% perf-profile.calltrace.cycles-pp.page_counter_uncharge.drain_stock.refill_stock.__sk_mem_reduce_allocated.sctp_wfree
> > 0.00 +1.4 1.36 ± 4% perf-profile.calltrace.cycles-pp.drain_stock.refill_stock.__sk_mem_reduce_allocated.sctp_wfree.skb_release_head_state
> > 1.67 ± 3% +1.4 3.04 ± 4% perf-profile.calltrace.cycles-pp.sctp_ulpevent_make_rcvmsg.sctp_ulpq_tail_data.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv
> > 11.39 ± 3% +1.4 12.77 ± 4% perf-profile.calltrace.cycles-pp.sctp_do_sm.sctp_assoc_bh_rcv.sctp_backlog_rcv.__release_sock.release_sock
> > 0.00 +1.4 1.45 ± 4% perf-profile.calltrace.cycles-pp.refill_stock.__sk_mem_reduce_allocated.sctp_wfree.skb_release_head_state.consume_skb
> > 0.00 +1.5 1.46 ± 5% perf-profile.calltrace.cycles-pp.try_charge_memcg.mem_cgroup_charge_skmem.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_ulpevent_make_rcvmsg
> > 0.00 +1.7 1.74 ± 4% perf-profile.calltrace.cycles-pp.mem_cgroup_charge_skmem.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_ulpevent_make_rcvmsg.sctp_ulpq_tail_data
> > 0.00 +1.8 1.83 ± 4% perf-profile.calltrace.cycles-pp.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_ulpevent_make_rcvmsg.sctp_ulpq_tail_data.sctp_cmd_interpreter
> > 0.00 +1.8 1.84 ± 4% perf-profile.calltrace.cycles-pp.__sk_mem_schedule.sctp_ulpevent_make_rcvmsg.sctp_ulpq_tail_data.sctp_cmd_interpreter.sctp_do_sm
> > 0.00 +2.0 1.97 ± 5% perf-profile.calltrace.cycles-pp.page_counter_try_charge.try_charge_memcg.mem_cgroup_charge_skmem.__sk_mem_raise_allocated.__sk_mem_schedule
> > 18.94 ± 3% +2.1 21.02 ± 4% perf-profile.calltrace.cycles-pp.recvmsg
> > 0.00 +2.1 2.14 ± 4% perf-profile.calltrace.cycles-pp.try_charge_memcg.mem_cgroup_charge_skmem.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_sendmsg_to_asoc
> > 10.88 ± 4% +2.2 13.11 ± 4% perf-profile.calltrace.cycles-pp.__ip_queue_xmit.sctp_packet_transmit.sctp_outq_flush.sctp_cmd_interpreter.sctp_do_sm
> > 10.74 ± 4% +2.3 13.04 ± 4% perf-profile.calltrace.cycles-pp.ip_finish_output2.__ip_queue_xmit.sctp_packet_transmit.sctp_outq_flush.sctp_cmd_interpreter
> > 18.22 ± 3% +2.3 20.55 ± 4% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.recvmsg
> > 10.51 ± 4% +2.4 12.88 ± 4% perf-profile.calltrace.cycles-pp.__local_bh_enable_ip.ip_finish_output2.__ip_queue_xmit.sctp_packet_transmit.sctp_outq_flush
> > 10.47 ± 4% +2.4 12.84 ± 4% perf-profile.calltrace.cycles-pp.do_softirq.__local_bh_enable_ip.ip_finish_output2.__ip_queue_xmit.sctp_packet_transmit
> > 18.08 ± 3% +2.4 20.46 ± 4% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.recvmsg
> > 10.42 ± 4% +2.4 12.81 ± 4% perf-profile.calltrace.cycles-pp.__softirqentry_text_start.do_softirq.__local_bh_enable_ip.ip_finish_output2.__ip_queue_xmit
> > 10.25 ± 4% +2.4 12.70 ± 4% perf-profile.calltrace.cycles-pp.net_rx_action.__softirqentry_text_start.do_softirq.__local_bh_enable_ip.ip_finish_output2
> > 17.84 ± 3% +2.5 20.32 ± 4% perf-profile.calltrace.cycles-pp.__sys_recvmsg.do_syscall_64.entry_SYSCALL_64_after_hwframe.recvmsg
> > 10.15 ± 4% +2.5 12.64 ± 4% perf-profile.calltrace.cycles-pp.__napi_poll.net_rx_action.__softirqentry_text_start.do_softirq.__local_bh_enable_ip
> > 10.13 ± 4% +2.5 12.63 ± 4% perf-profile.calltrace.cycles-pp.process_backlog.__napi_poll.net_rx_action.__softirqentry_text_start.do_softirq
> > 0.00 +2.5 2.50 ± 4% perf-profile.calltrace.cycles-pp.mem_cgroup_charge_skmem.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_sendmsg_to_asoc.sctp_sendmsg
> > 3.59 ± 3% +2.5 6.11 ± 5% perf-profile.calltrace.cycles-pp.sctp_skb_recv_datagram.sctp_recvmsg.inet_recvmsg.____sys_recvmsg.___sys_recvmsg
> > 9.97 ± 4% +2.6 12.54 ± 4% perf-profile.calltrace.cycles-pp.__netif_receive_skb_one_core.process_backlog.__napi_poll.net_rx_action.__softirqentry_text_start
> > 17.56 ± 3% +2.6 20.14 ± 4% perf-profile.calltrace.cycles-pp.___sys_recvmsg.__sys_recvmsg.do_syscall_64.entry_SYSCALL_64_after_hwframe.recvmsg
> > 0.00 +2.7 2.66 ± 4% perf-profile.calltrace.cycles-pp.__sk_mem_raise_allocated.__sk_mem_schedule.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg
> > 0.00 +2.7 2.68 ± 4% perf-profile.calltrace.cycles-pp.__sk_mem_schedule.sctp_sendmsg_to_asoc.sctp_sendmsg.sock_sendmsg.____sys_sendmsg
> > 9.70 ± 4% +2.7 12.40 ± 4% perf-profile.calltrace.cycles-pp.ip_local_deliver_finish.__netif_receive_skb_one_core.process_backlog.__napi_poll.net_rx_action
> > 9.68 ± 4% +2.7 12.39 ± 4% perf-profile.calltrace.cycles-pp.ip_protocol_deliver_rcu.ip_local_deliver_finish.__netif_receive_skb_one_core.process_backlog.__napi_poll
> > 9.61 ± 4% +2.7 12.33 ± 4% perf-profile.calltrace.cycles-pp.sctp_rcv.ip_protocol_deliver_rcu.ip_local_deliver_finish.__netif_receive_skb_one_core.process_backlog
> > 16.65 ± 3% +2.9 19.58 ± 4% perf-profile.calltrace.cycles-pp.____sys_recvmsg.___sys_recvmsg.__sys_recvmsg.do_syscall_64.entry_SYSCALL_64_after_hwframe
> > 8.06 ± 4% +3.1 11.19 ± 4% perf-profile.calltrace.cycles-pp.sctp_assoc_bh_rcv.sctp_rcv.ip_protocol_deliver_rcu.ip_local_deliver_finish.__netif_receive_skb_one_core
> > 15.88 ± 3% +3.2 19.12 ± 4% perf-profile.calltrace.cycles-pp.inet_recvmsg.____sys_recvmsg.___sys_recvmsg.__sys_recvmsg.do_syscall_64
> > 15.83 ± 3% +3.3 19.09 ± 4% perf-profile.calltrace.cycles-pp.sctp_recvmsg.inet_recvmsg.____sys_recvmsg.___sys_recvmsg.__sys_recvmsg
> > 7.62 ± 3% +3.3 10.90 ± 4% perf-profile.calltrace.cycles-pp.sctp_do_sm.sctp_assoc_bh_rcv.sctp_rcv.ip_protocol_deliver_rcu.ip_local_deliver_finish
> > 1.14 ± 3% +3.5 4.62 ± 5% perf-profile.calltrace.cycles-pp.lock_sock_nested.sctp_skb_recv_datagram.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
> > 1.02 ± 3% +3.5 4.51 ± 5% perf-profile.calltrace.cycles-pp._raw_spin_lock_bh.lock_sock_nested.sctp_skb_recv_datagram.sctp_recvmsg.inet_recvmsg
> > 0.83 ± 4% +3.6 4.38 ± 5% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_bh.lock_sock_nested.sctp_skb_recv_datagram.sctp_recvmsg
> > 15.16 ± 3% +3.6 18.74 ± 5% perf-profile.calltrace.cycles-pp.release_sock.sctp_sendmsg.sock_sendmsg.____sys_sendmsg.___sys_sendmsg
> > 15.04 ± 3% +3.6 18.66 ± 5% perf-profile.calltrace.cycles-pp.__release_sock.release_sock.sctp_sendmsg.sock_sendmsg.____sys_sendmsg
> > 11.66 ± 3% +3.6 15.29 ± 5% perf-profile.calltrace.cycles-pp.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv.sctp_backlog_rcv.__release_sock
> > 14.98 ± 3% +3.6 18.62 ± 5% perf-profile.calltrace.cycles-pp.sctp_backlog_rcv.__release_sock.release_sock.sctp_sendmsg.sock_sendmsg
> > 0.00 +4.2 4.24 ± 4% perf-profile.calltrace.cycles-pp.sctp_eat_data.sctp_sf_eat_data_6_2.sctp_do_sm.sctp_assoc_bh_rcv.sctp_rcv
> > 0.00 +4.3 4.33 ± 4% perf-profile.calltrace.cycles-pp.sctp_sf_eat_data_6_2.sctp_do_sm.sctp_assoc_bh_rcv.sctp_rcv.ip_protocol_deliver_rcu
> > 2.88 ± 3% +4.7 7.54 ± 4% perf-profile.calltrace.cycles-pp.sctp_outq_sack.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv.sctp_backlog_rcv
> > 23.81 ± 12% +4.8 28.57 ± 11% perf-profile.calltrace.cycles-pp.mwait_idle_with_hints.intel_idle.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call
> > 23.81 ± 12% +4.8 28.58 ± 11% perf-profile.calltrace.cycles-pp.intel_idle.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
> > 2.11 ± 4% +4.8 6.89 ± 4% perf-profile.calltrace.cycles-pp.kfree_skb_reason.sctp_recvmsg.inet_recvmsg.____sys_recvmsg.___sys_recvmsg
> > 1.84 ± 3% +5.0 6.86 ± 4% perf-profile.calltrace.cycles-pp.sctp_chunk_put.sctp_outq_sack.sctp_cmd_interpreter.sctp_do_sm.sctp_assoc_bh_rcv
> > 1.42 ± 3% +5.0 6.47 ± 4% perf-profile.calltrace.cycles-pp.consume_skb.sctp_chunk_put.sctp_outq_sack.sctp_cmd_interpreter.sctp_do_sm
> > 0.00 +5.5 5.53 ± 4% perf-profile.calltrace.cycles-pp.__sk_mem_reduce_allocated.sctp_wfree.skb_release_head_state.consume_skb.sctp_chunk_put
> > 0.00 +5.8 5.80 ± 4% perf-profile.calltrace.cycles-pp.sctp_wfree.skb_release_head_state.consume_skb.sctp_chunk_put.sctp_outq_sack
> > 0.00 +5.9 5.87 ± 4% perf-profile.calltrace.cycles-pp.skb_release_head_state.consume_skb.sctp_chunk_put.sctp_outq_sack.sctp_cmd_interpreter
> > 0.00 +6.0 5.99 ± 4% perf-profile.calltrace.cycles-pp.__sk_mem_reduce_allocated.skb_release_head_state.kfree_skb_reason.sctp_recvmsg.inet_recvmsg
> > 0.00 +6.1 6.13 ± 4% perf-profile.calltrace.cycles-pp.skb_release_head_state.kfree_skb_reason.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
> > 9.61 ± 5% -6.2 3.42 ± 5% perf-profile.children.cycles-pp.intel_idle_irq
> > 9.16 ± 3% -3.3 5.87 ± 5% perf-profile.children.cycles-pp.sctp_packet_pack
> > 8.74 ± 3% -3.1 5.61 ± 4% perf-profile.children.cycles-pp.copy_user_enhanced_fast_string
> > 25.01 ± 3% -3.0 21.98 ± 4% perf-profile.children.cycles-pp.sctp_outq_flush
> > 8.52 ± 3% -3.0 5.50 ± 5% perf-profile.children.cycles-pp.memcpy_erms
> > 6.49 ± 3% -2.5 3.99 ± 4% perf-profile.children.cycles-pp.skb_copy_datagram_iter
> > 6.47 ± 3% -2.5 3.98 ± 4% perf-profile.children.cycles-pp.__skb_datagram_iter
> > 7.07 ± 3% -2.5 4.60 ± 4% perf-profile.children.cycles-pp.sctp_datamsg_from_user
> > 21.61 ± 3% -2.4 19.19 ± 4% perf-profile.children.cycles-pp.sctp_sendmsg_to_asoc
> > 13.72 ± 3% -2.4 11.31 ± 5% perf-profile.children.cycles-pp.sctp_primitive_SEND
> > 5.63 ± 3% -2.1 3.50 ± 4% perf-profile.children.cycles-pp._copy_to_iter
> > 5.46 ± 3% -2.1 3.40 ± 4% perf-profile.children.cycles-pp.copyout
> > 22.37 ± 3% -2.0 20.39 ± 4% perf-profile.children.cycles-pp.sctp_packet_transmit
> > 3.59 ± 3% -1.6 1.96 ± 4% perf-profile.children.cycles-pp.__alloc_skb
> > 2.98 ± 3% -1.5 1.49 ± 5% perf-profile.children.cycles-pp.skb_release_data
> > 3.46 ± 4% -1.4 2.06 ± 4% perf-profile.children.cycles-pp.__schedule
> > 2.66 ± 3% -1.3 1.35 ± 4% perf-profile.children.cycles-pp.kmalloc_reserve
> > 2.62 ± 3% -1.3 1.32 ± 4% perf-profile.children.cycles-pp.__kmalloc_node_track_caller
> > 2.50 ± 3% -1.3 1.24 ± 4% perf-profile.children.cycles-pp.kmalloc_large_node
> > 3.51 ± 3% -1.2 2.36 ± 4% perf-profile.children.cycles-pp.sctp_user_addto_chunk
> > 2.20 ± 3% -1.1 1.07 ± 4% perf-profile.children.cycles-pp.__alloc_pages
> > 2.93 ± 3% -1.1 1.82 ± 5% perf-profile.children.cycles-pp.sctp_make_datafrag_empty
> > 2.70 ± 3% -1.1 1.62 ± 4% perf-profile.children.cycles-pp._sctp_make_chunk
> > 3.17 ± 3% -1.0 2.12 ± 5% perf-profile.children.cycles-pp.sctp_data_ready
> > 3.16 ± 3% -1.0 2.13 ± 4% perf-profile.children.cycles-pp._copy_from_iter
> > 3.28 ± 3% -1.0 2.26 ± 5% perf-profile.children.cycles-pp.sctp_ulpq_tail_event
> > 2.90 ± 3% -1.0 1.88 ± 5% perf-profile.children.cycles-pp.__wake_up_common_lock
> > 1.89 ± 4% -1.0 0.88 ± 4% perf-profile.children.cycles-pp.get_page_from_freelist
> > 3.02 ± 3% -1.0 2.02 ± 4% perf-profile.children.cycles-pp.copyin
> > 2.47 ± 3% -1.0 1.49 ± 7% perf-profile.children.cycles-pp.sctp_outq_flush_data
> > 2.70 ± 3% -0.9 1.75 ± 5% perf-profile.children.cycles-pp.__wake_up_common
> > 2.53 ± 3% -0.9 1.64 ± 5% perf-profile.children.cycles-pp.autoremove_wake_function
> > 2.51 ± 3% -0.9 1.63 ± 4% perf-profile.children.cycles-pp.try_to_wake_up
> > 1.83 ± 3% -0.8 1.02 ± 4% perf-profile.children.cycles-pp.sctp_ulpevent_free
> > 2.10 ± 3% -0.8 1.32 ± 6% perf-profile.children.cycles-pp.kmem_cache_free
> > 1.89 ± 3% -0.8 1.12 ± 5% perf-profile.children.cycles-pp.schedule_timeout
> > 1.39 ± 4% -0.8 0.63 ± 5% perf-profile.children.cycles-pp.free_unref_page
> > 1.86 ± 3% -0.7 1.11 ± 4% perf-profile.children.cycles-pp.schedule
> > 1.25 ± 4% -0.7 0.56 ± 5% perf-profile.children.cycles-pp.rmqueue
> > 1.28 ± 4% -0.7 0.60 ± 5% perf-profile.children.cycles-pp._raw_spin_lock_irqsave
> > 1.66 ± 4% -0.7 1.00 ± 4% perf-profile.children.cycles-pp.schedule_idle
> > 0.98 ± 4% -0.6 0.43 ± 3% perf-profile.children.cycles-pp._raw_spin_lock
> > 1.08 ± 4% -0.5 0.54 ± 4% perf-profile.children.cycles-pp.sctp_outq_select_transport
> > 1.45 ± 4% -0.5 0.94 ± 4% perf-profile.children.cycles-pp._copy_from_user
> > 0.98 ± 4% -0.5 0.48 ± 5% perf-profile.children.cycles-pp.sctp_packet_config
> > 1.25 ± 12% -0.5 0.78 ± 5% perf-profile.children.cycles-pp.__rhashtable_lookup
> > 1.21 ± 3% -0.5 0.73 ± 4% perf-profile.children.cycles-pp.__check_object_size
> > 0.98 ± 3% -0.4 0.53 ± 7% perf-profile.children.cycles-pp.dequeue_task_fair
> > 0.64 ± 5% -0.4 0.20 ± 5% perf-profile.children.cycles-pp.free_pcppages_bulk
> > 1.18 ± 4% -0.4 0.78 ± 5% perf-profile.children.cycles-pp.ttwu_do_activate
> > 0.62 ± 4% -0.4 0.22 ± 4% perf-profile.children.cycles-pp.__free_pages_ok
> > 0.90 ± 3% -0.4 0.50 ± 6% perf-profile.children.cycles-pp.dequeue_entity
> > 1.14 ± 3% -0.4 0.75 ± 4% perf-profile.children.cycles-pp.enqueue_task_fair
> > 0.95 ± 4% -0.4 0.57 ± 6% perf-profile.children.cycles-pp.__slab_free
> > 0.76 ± 3% -0.3 0.42 ± 3% perf-profile.children.cycles-pp.__skb_clone
> > 1.07 ± 4% -0.3 0.73 ± 5% perf-profile.children.cycles-pp.sendmsg_copy_msghdr
> > 0.76 ± 3% -0.3 0.42 ± 5% perf-profile.children.cycles-pp.update_load_avg
> > 0.98 ± 4% -0.3 0.65 ± 3% perf-profile.children.cycles-pp.__copy_msghdr_from_user
> > 0.81 ± 14% -0.3 0.49 ± 6% perf-profile.children.cycles-pp.sctp_hash_cmp
> > 0.88 ± 4% -0.3 0.56 ± 4% perf-profile.children.cycles-pp.pick_next_task_fair
> > 0.84 ± 3% -0.3 0.54 ± 5% perf-profile.children.cycles-pp.enqueue_entity
> > 0.82 ± 3% -0.3 0.53 ± 5% perf-profile.children.cycles-pp.import_iovec
> > 0.81 ± 2% -0.3 0.52 ± 6% perf-profile.children.cycles-pp.kfree
> > 0.59 ± 3% -0.3 0.30 ± 4% perf-profile.children.cycles-pp.__copy_skb_header
> > 0.78 ± 3% -0.3 0.51 ± 6% perf-profile.children.cycles-pp.__import_iovec
> > 0.62 ± 5% -0.3 0.35 ± 5% perf-profile.children.cycles-pp.set_next_entity
> > 0.76 ± 12% -0.3 0.50 ± 5% perf-profile.children.cycles-pp.sctp_endpoint_lookup_assoc
> > 0.69 ± 4% -0.3 0.43 ± 5% perf-profile.children.cycles-pp.check_heap_object
> > 0.77 ± 3% -0.3 0.52 ± 5% perf-profile.children.cycles-pp.kmem_cache_alloc
> > 0.74 ± 12% -0.3 0.49 ± 5% perf-profile.children.cycles-pp.sctp_epaddr_lookup_transport
> > 0.72 ± 10% -0.2 0.48 ± 6% perf-profile.children.cycles-pp.__sctp_rcv_lookup
> > 0.66 ± 4% -0.2 0.42 ± 5% perf-profile.children.cycles-pp.__might_resched
> > 0.70 ± 10% -0.2 0.46 ± 7% perf-profile.children.cycles-pp.sctp_addrs_lookup_transport
> > 0.65 ± 2% -0.2 0.42 ± 5% perf-profile.children.cycles-pp.iovec_from_user
> > 0.50 ± 4% -0.2 0.26 ± 4% perf-profile.children.cycles-pp.simple_copy_to_iter
> > 0.46 ± 3% -0.2 0.23 ± 8% perf-profile.children.cycles-pp.sctp_inq_pop
> > 0.41 ± 3% -0.2 0.19 ± 6% perf-profile.children.cycles-pp.update_curr
> > 0.69 ± 3% -0.2 0.47 ± 5% perf-profile.children.cycles-pp.sctp_chunkify
> > 0.55 ± 3% -0.2 0.34 ± 4% perf-profile.children.cycles-pp.__dev_queue_xmit
> > 0.53 ± 4% -0.2 0.33 ± 6% perf-profile.children.cycles-pp.__might_fault
> > 0.58 ± 4% -0.2 0.39 ± 5% perf-profile.children.cycles-pp.copy_user_short_string
> > 0.46 ± 9% -0.2 0.28 ± 12% perf-profile.children.cycles-pp.select_task_rq
> > 0.33 ± 4% -0.2 0.16 ± 7% perf-profile.children.cycles-pp.free_unref_page_commit
> > 0.28 ± 5% -0.2 0.11 ± 9% perf-profile.children.cycles-pp.dst_release
> > 0.32 ± 7% -0.2 0.15 ± 7% perf-profile.children.cycles-pp.ipv4_dst_check
> > 0.38 ± 3% -0.2 0.22 ± 8% perf-profile.children.cycles-pp.__mod_node_page_state
> > 0.44 ± 8% -0.2 0.28 ± 13% perf-profile.children.cycles-pp.sctp_packet_append_chunk
> > 0.49 ± 6% -0.2 0.33 ± 11% perf-profile.children.cycles-pp.sctp_packet_transmit_chunk
> > 0.54 ± 4% -0.2 0.39 ± 6% perf-profile.children.cycles-pp.__list_del_entry_valid
> > 0.34 ± 4% -0.2 0.19 ± 8% perf-profile.children.cycles-pp.sctp_v4_xmit
> > 0.42 ± 3% -0.2 0.27 ± 10% perf-profile.children.cycles-pp.sockfd_lookup_light
> > 0.32 ± 4% -0.1 0.17 ± 4% perf-profile.children.cycles-pp.__zone_watermark_ok
> > 0.47 ± 3% -0.1 0.32 ± 5% perf-profile.children.cycles-pp.__entry_text_start
> > 0.28 ± 2% -0.1 0.13 ± 5% perf-profile.children.cycles-pp.__free_one_page
> > 0.28 ± 5% -0.1 0.13 ± 4% perf-profile.children.cycles-pp.prepare_task_switch
> > 0.43 ± 5% -0.1 0.29 ± 9% perf-profile.children.cycles-pp.sctp_addto_chunk
> > 0.37 ± 4% -0.1 0.23 ± 6% perf-profile.children.cycles-pp.sctp_gen_sack
> > 0.41 ± 4% -0.1 0.27 ± 8% perf-profile.children.cycles-pp.aa_sk_perm
> > 0.35 ± 3% -0.1 0.22 ± 6% perf-profile.children.cycles-pp.__list_add_valid
> > 0.34 ± 4% -0.1 0.20 ± 6% perf-profile.children.cycles-pp.sctp_make_sack
> > 0.44 ± 4% -0.1 0.30 ± 6% perf-profile.children.cycles-pp.update_rq_clock
> > 0.37 ± 4% -0.1 0.24 ± 7% perf-profile.children.cycles-pp.sock_recvmsg
> > 0.34 ± 4% -0.1 0.22 ± 4% perf-profile.children.cycles-pp.sctp_check_transmitted
> > 0.35 ± 3% -0.1 0.22 ± 6% perf-profile.children.cycles-pp.memcg_slab_free_hook
> > 0.36 ± 4% -0.1 0.23 ± 8% perf-profile.children.cycles-pp.__fdget
> > 0.34 ± 10% -0.1 0.22 ± 15% perf-profile.children.cycles-pp.select_task_rq_fair
> > 0.44 ± 4% -0.1 0.32 ± 7% perf-profile.children.cycles-pp.kmem_cache_alloc_node
> > 0.34 ± 5% -0.1 0.21 ± 7% perf-profile.children.cycles-pp.__virt_addr_valid
> > 0.34 ± 3% -0.1 0.22 ± 5% perf-profile.children.cycles-pp.dev_hard_start_xmit
> > 0.28 ± 14% -0.1 0.16 ± 6% perf-profile.children.cycles-pp.sctp_transport_hold
> > 0.35 ± 2% -0.1 0.23 ± 7% perf-profile.children.cycles-pp.__might_sleep
> > 0.38 ± 4% -0.1 0.26 ± 4% perf-profile.children.cycles-pp.sctp_chunk_free
> > 0.31 ± 5% -0.1 0.20 ± 5% perf-profile.children.cycles-pp.loopback_xmit
> > 0.33 ± 4% -0.1 0.22 ± 4% perf-profile.children.cycles-pp.free_pcp_prepare
> > 0.33 ± 5% -0.1 0.22 ± 7% perf-profile.children.cycles-pp.security_socket_recvmsg
> > 0.35 ± 2% -0.1 0.24 ± 4% perf-profile.children.cycles-pp.sock_kmalloc
> > 0.30 ± 2% -0.1 0.20 ± 8% perf-profile.children.cycles-pp.sctp_association_put
> > 0.22 ± 7% -0.1 0.12 ± 6% perf-profile.children.cycles-pp.sctp_assoc_rwnd_increase
> > 0.34 ± 4% -0.1 0.23 ± 2% perf-profile.children.cycles-pp.move_addr_to_kernel
> > 0.28 ± 4% -0.1 0.17 ± 4% perf-profile.children.cycles-pp.sock_kfree_s
> > 0.23 ± 5% -0.1 0.13 ± 7% perf-profile.children.cycles-pp.rmqueue_bulk
> > 0.31 ± 9% -0.1 0.21 ± 8% perf-profile.children.cycles-pp.sctp_transport_put
> > 0.33 ± 6% -0.1 0.23 ± 12% perf-profile.children.cycles-pp.sctp_sched_fcfs_dequeue
> > 0.28 ± 8% -0.1 0.18 ± 20% perf-profile.children.cycles-pp.__sctp_packet_append_chunk
> > 0.34 ± 4% -0.1 0.24 ± 8% perf-profile.children.cycles-pp.memset_erms
> > 0.19 ± 4% -0.1 0.10 ± 5% perf-profile.children.cycles-pp.__update_load_avg_cfs_rq
> > 0.29 ± 4% -0.1 0.19 ± 7% perf-profile.children.cycles-pp.skb_set_owner_w
> > 0.27 ± 4% -0.1 0.18 ± 5% perf-profile.children.cycles-pp.security_socket_sendmsg
> > 0.24 ± 4% -0.1 0.14 ± 6% perf-profile.children.cycles-pp.__switch_to
> > 0.26 ± 4% -0.1 0.17 ± 9% perf-profile.children.cycles-pp.entry_SYSRETQ_unsafe_stack
> > 0.27 ± 2% -0.1 0.18 ± 5% perf-profile.children.cycles-pp.send_sctp_stream_1toMany
> > 0.23 ± 4% -0.1 0.14 ± 5% perf-profile.children.cycles-pp.tick_nohz_idle_exit
> > 0.20 ± 4% -0.1 0.11 ± 8% perf-profile.children.cycles-pp.sctp_outq_flush_ctrl
> > 0.24 ± 7% -0.1 0.15 ± 14% perf-profile.children.cycles-pp.select_idle_sibling
> > 0.26 ± 4% -0.1 0.17 ± 11% perf-profile.children.cycles-pp.syscall_exit_to_user_mode
> > 0.22 ± 5% -0.1 0.14 ± 5% perf-profile.children.cycles-pp.__switch_to_asm
> > 0.21 -0.1 0.13 ± 8% perf-profile.children.cycles-pp.accept_connection
> > 0.21 -0.1 0.13 ± 8% perf-profile.children.cycles-pp.spawn_child
> > 0.21 -0.1 0.13 ± 8% perf-profile.children.cycles-pp.process_requests
> > 0.14 ± 5% -0.1 0.06 ± 11% perf-profile.children.cycles-pp.check_stack_object
> > 0.20 ± 3% -0.1 0.12 ± 9% perf-profile.children.cycles-pp.sock_wfree
> > 0.22 ± 4% -0.1 0.14 ± 3% perf-profile.children.cycles-pp.prepare_to_wait_exclusive
> > 0.24 ± 5% -0.1 0.17 ± 9% perf-profile.children.cycles-pp.sctp_sched_dequeue_common
> > 0.22 ± 3% -0.1 0.15 ± 2% perf-profile.children.cycles-pp.__kmalloc
> > 0.21 ± 3% -0.1 0.13 ± 8% perf-profile.children.cycles-pp.sctp_make_control
> > 0.20 ± 2% -0.1 0.12 ± 9% perf-profile.children.cycles-pp.recv_sctp_stream_1toMany
> > 0.17 ± 6% -0.1 0.09 ± 5% perf-profile.children.cycles-pp.___perf_sw_event
> > 0.14 ± 5% -0.1 0.06 ± 7% perf-profile.children.cycles-pp.sctp_sock_rfree
> > 0.20 ± 4% -0.1 0.13 ± 7% perf-profile.children.cycles-pp.skb_clone
> > 0.19 ± 30% -0.1 0.12 ± 5% perf-profile.children.cycles-pp.sctp_cmp_addr_exact
> > 0.17 ± 4% -0.1 0.10 ± 8% perf-profile.children.cycles-pp.__mod_timer
> > 0.20 ± 4% -0.1 0.13 ± 6% perf-profile.children.cycles-pp.available_idle_cpu
> > 0.23 ± 7% -0.1 0.16 ± 3% perf-profile.children.cycles-pp.__cond_resched
> > 0.21 ± 6% -0.1 0.14 ± 10% perf-profile.children.cycles-pp.sctp_datamsg_put
> > 0.16 ± 4% -0.1 0.09 ± 5% perf-profile.children.cycles-pp.check_new_pages
> > 0.14 ± 6% -0.1 0.08 ± 12% perf-profile.children.cycles-pp.sctp_inet_skb_msgname
> > 0.14 ± 6% -0.1 0.08 ± 9% perf-profile.children.cycles-pp.ip_rcv
> > 0.18 ± 3% -0.1 0.11 ± 9% perf-profile.children.cycles-pp.syscall_return_via_sysret
> > 0.18 ± 4% -0.1 0.12 ± 9% perf-profile.children.cycles-pp.poll_idle
> > 0.20 ± 6% -0.1 0.14 ± 6% perf-profile.children.cycles-pp.kmem_cache_alloc_trace
> > 0.19 -0.1 0.13 ± 8% perf-profile.children.cycles-pp.sctp_outq_tail
> > 0.18 ± 3% -0.1 0.12 ± 7% perf-profile.children.cycles-pp.reweight_entity
> > 0.20 ± 4% -0.1 0.13 ± 11% perf-profile.children.cycles-pp.update_cfs_group
> > 0.16 ± 5% -0.1 0.10 ± 4% perf-profile.children.cycles-pp.del_timer
> > 0.16 ± 5% -0.1 0.10 ± 3% perf-profile.children.cycles-pp.sctp_sendmsg_parse
> > 0.16 ± 5% -0.1 0.10 ± 9% perf-profile.children.cycles-pp.perf_trace_sched_wakeup_template
> > 0.18 ± 2% -0.1 0.12 ± 4% perf-profile.children.cycles-pp.__free_pages
> > 0.17 ± 4% -0.1 0.11 ± 5% perf-profile.children.cycles-pp.nr_iowait_cpu
> > 0.14 ± 2% -0.1 0.09 ± 10% perf-profile.children.cycles-pp.switch_mm_irqs_off
> > 0.11 ± 6% -0.1 0.05 ± 8% perf-profile.children.cycles-pp.__build_skb_around
> > 0.19 ± 3% -0.1 0.14 ± 8% perf-profile.children.cycles-pp.__update_load_avg_se
> > 0.16 ± 5% -0.1 0.11 ± 6% perf-profile.children.cycles-pp.ttwu_do_wakeup
> > 0.13 ± 6% -0.1 0.08 ± 13% perf-profile.children.cycles-pp.task_tick_fair
> > 0.21 ± 3% -0.0 0.16 ± 6% perf-profile.children.cycles-pp.sctp_association_hold
> > 0.15 ± 6% -0.0 0.10 ± 8% perf-profile.children.cycles-pp.check_preempt_curr
> > 0.07 ± 6% -0.0 0.02 ± 99% perf-profile.children.cycles-pp._raw_spin_unlock_irqrestore
> > 0.10 ± 5% -0.0 0.05 ± 7% perf-profile.children.cycles-pp.finish_task_switch
> > 0.14 ± 6% -0.0 0.10 ± 6% perf-profile.children.cycles-pp.__genradix_ptr
> > 0.12 ± 4% -0.0 0.08 ± 8% perf-profile.children.cycles-pp.sctp_chunk_abandoned
> > 0.07 ± 5% -0.0 0.02 ± 99% perf-profile.children.cycles-pp.ip_local_out
> > 0.22 ± 6% -0.0 0.17 ± 8% perf-profile.children.cycles-pp.sched_clock_cpu
> > 0.14 ± 5% -0.0 0.09 ± 6% perf-profile.children.cycles-pp.memcg_slab_post_alloc_hook
> > 0.12 ± 6% -0.0 0.08 ± 7% perf-profile.children.cycles-pp.__netif_rx
> > 0.20 ± 5% -0.0 0.16 ± 8% perf-profile.children.cycles-pp.read_tsc
> > 0.12 ± 6% -0.0 0.08 ± 10% perf-profile.children.cycles-pp.lock_timer_base
> > 0.19 ± 5% -0.0 0.15 ± 7% perf-profile.children.cycles-pp.native_sched_clock
> > 0.13 ± 5% -0.0 0.09 ± 5% perf-profile.children.cycles-pp.pick_next_entity
> > 0.12 ± 5% -0.0 0.08 ± 6% perf-profile.children.cycles-pp.netif_rx_internal
> > 0.12 ± 7% -0.0 0.08 ± 10% perf-profile.children.cycles-pp.__put_user_nocheck_4
> > 0.12 ± 4% -0.0 0.08 ± 6% perf-profile.children.cycles-pp.skb_put
> > 0.11 ± 4% -0.0 0.07 ± 5% perf-profile.children.cycles-pp.__put_user_nocheck_8
> > 0.12 ± 3% -0.0 0.08 ± 8% perf-profile.children.cycles-pp.resched_curr
> > 0.11 ± 8% -0.0 0.07 ± 25% perf-profile.children.cycles-pp.sctp_chunk_assign_ssn
> > 0.13 ± 4% -0.0 0.09 ± 6% perf-profile.children.cycles-pp.sctp_datamsg_destroy
> > 0.07 ± 7% -0.0 0.03 ±100% perf-profile.children.cycles-pp.ip_rcv_core
> > 0.11 ± 3% -0.0 0.07 ± 10% perf-profile.children.cycles-pp.enqueue_to_backlog
> > 0.10 ± 5% -0.0 0.06 ± 13% perf-profile.children.cycles-pp.sctp_transport_reset_t3_rtx
> > 0.09 ± 6% -0.0 0.05 ± 7% perf-profile.children.cycles-pp.__netif_receive_skb_core
> > 0.10 ± 5% -0.0 0.06 ± 6% perf-profile.children.cycles-pp.validate_xmit_skb
> > 0.11 ± 6% -0.0 0.07 ± 12% perf-profile.children.cycles-pp.syscall_exit_to_user_mode_prepare
> > 0.06 ± 6% -0.0 0.02 ± 99% perf-profile.children.cycles-pp.entry_SYSCALL_64_safe_stack
> > 0.10 ± 8% -0.0 0.06 ± 9% perf-profile.children.cycles-pp.sctp_ulpq_order
> > 0.09 ± 6% -0.0 0.06 ± 8% perf-profile.children.cycles-pp.os_xsave
> > 0.12 ± 5% -0.0 0.08 ± 8% perf-profile.children.cycles-pp.__check_heap_object
> > 0.10 ± 5% -0.0 0.07 ± 10% perf-profile.children.cycles-pp.rcu_all_qs
> > 0.08 ± 5% -0.0 0.06 ± 11% perf-profile.children.cycles-pp.tick_nohz_idle_enter
> > 0.08 ± 5% -0.0 0.06 ± 8% perf-profile.children.cycles-pp.perf_tp_event
> > 0.08 -0.0 0.06 ± 9% perf-profile.children.cycles-pp.__mod_lruvec_page_state
> > 0.31 ± 3% -0.0 0.28 ± 4% perf-profile.children.cycles-pp.sctp_ulpevent_receive_data
> > 0.10 ± 10% -0.0 0.08 ± 10% perf-profile.children.cycles-pp.rcu_idle_exit
> > 0.08 ± 6% -0.0 0.05 ± 8% perf-profile.children.cycles-pp.sctp_tsnmap_check
> > 0.08 ± 5% -0.0 0.06 ± 14% perf-profile.children.cycles-pp.sctp_ulpevent_init
> > 0.07 ± 8% -0.0 0.05 perf-profile.children.cycles-pp.__wrgsbase_inactive
> > 0.10 ± 5% +0.0 0.13 ± 10% perf-profile.children.cycles-pp.sctp_sockaddr_af
> > 0.00 +0.1 0.05 perf-profile.children.cycles-pp.perf_mux_hrtimer_handler
> > 0.19 ± 3% +0.1 0.24 ± 11% perf-profile.children.cycles-pp.scheduler_tick
> > 0.00 +0.1 0.06 ± 7% perf-profile.children.cycles-pp.lapic_next_deadline
> > 0.00 +0.1 0.07 ± 32% perf-profile.children.cycles-pp.arch_cpu_idle_enter
> > 0.00 +0.1 0.07 ± 16% perf-profile.children.cycles-pp.native_irq_return_iret
> > 0.00 +0.1 0.08 ± 22% perf-profile.children.cycles-pp.update_blocked_averages
> > 0.00 +0.1 0.08 ± 22% perf-profile.children.cycles-pp.run_rebalance_domains
> > 0.24 ± 3% +0.1 0.32 ± 9% perf-profile.children.cycles-pp.update_process_times
> > 0.00 +0.1 0.08 ± 22% perf-profile.children.cycles-pp.update_sg_lb_stats
> > 0.24 ± 4% +0.1 0.32 ± 9% perf-profile.children.cycles-pp.tick_sched_handle
> > 0.00 +0.1 0.08 ± 41% perf-profile.children.cycles-pp._raw_spin_trylock
> > 0.00 +0.1 0.09 ± 20% perf-profile.children.cycles-pp.update_sd_lb_stats
> > 0.00 +0.1 0.09 ± 19% perf-profile.children.cycles-pp.find_busiest_group
> > 0.04 ± 71% +0.1 0.14 ± 3% perf-profile.children.cycles-pp.sctp_get_af_specific
> > 0.01 ±223% +0.1 0.11 ± 18% perf-profile.children.cycles-pp.load_balance
> > 0.00 +0.1 0.11 ± 11% perf-profile.children.cycles-pp.tick_nohz_irq_exit
> > 0.02 ±141% +0.1 0.14 ± 46% perf-profile.children.cycles-pp.rebalance_domains
> > 0.45 ± 5% +0.1 0.58 ± 10% perf-profile.children.cycles-pp.ktime_get
> > 0.00 +0.1 0.14 ± 17% perf-profile.children.cycles-pp.cgroup_rstat_updated
> > 0.26 ± 4% +0.1 0.41 ± 12% perf-profile.children.cycles-pp.tick_sched_timer
> > 0.15 ± 10% +0.2 0.32 ± 16% perf-profile.children.cycles-pp.clockevents_program_event
> > 0.30 ± 4% +0.2 0.50 ± 10% perf-profile.children.cycles-pp.__hrtimer_run_queues
> > 0.25 ± 4% +0.2 0.45 ± 20% perf-profile.children.cycles-pp.tick_nohz_get_sleep_length
> > 0.00 +0.2 0.20 ± 16% perf-profile.children.cycles-pp.timekeeping_max_deferment
> > 0.50 ± 2% +0.2 0.71 ± 15% perf-profile.children.cycles-pp.menu_select
> > 0.16 ± 5% +0.2 0.38 ± 22% perf-profile.children.cycles-pp.tick_nohz_next_event
> > 0.08 ± 8% +0.2 0.30 ± 23% perf-profile.children.cycles-pp.__irq_exit_rcu
> > 0.00 +0.3 0.33 ± 12% perf-profile.children.cycles-pp.propagate_protected_usage
> > 0.48 ± 5% +0.4 0.89 ± 11% perf-profile.children.cycles-pp.hrtimer_interrupt
> > 0.48 ± 6% +0.4 0.90 ± 10% perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
> > 0.00 +0.6 0.55 ± 3% perf-profile.children.cycles-pp.mem_cgroup_uncharge_skmem
> > 0.61 ± 5% +0.7 1.35 ± 11% perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
> > 1.00 ± 2% +0.8 1.82 ± 10% perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
> > 0.00 +1.0 1.03 ± 4% perf-profile.children.cycles-pp.__mod_memcg_state
> > 1.73 ± 2% +1.4 3.14 ± 4% perf-profile.children.cycles-pp.sctp_ulpevent_make_rcvmsg
> > 19.09 ± 3% +2.0 21.11 ± 4% perf-profile.children.cycles-pp.recvmsg
> > 11.51 ± 3% +2.2 13.66 ± 4% perf-profile.children.cycles-pp.__ip_queue_xmit
> > 11.26 ± 3% +2.3 13.53 ± 4% perf-profile.children.cycles-pp.ip_finish_output2
> > 0.00 +2.3 2.27 ± 5% perf-profile.children.cycles-pp.page_counter_uncharge
> > 0.00 +2.3 2.32 ± 4% perf-profile.children.cycles-pp.drain_stock
> > 0.00 +2.4 2.40 ± 5% perf-profile.children.cycles-pp.page_counter_try_charge
> > 10.87 ± 3% +2.4 13.28 ± 4% perf-profile.children.cycles-pp.__local_bh_enable_ip
> > 10.76 ± 3% +2.4 13.20 ± 4% perf-profile.children.cycles-pp.do_softirq
> > 17.84 ± 3% +2.5 20.32 ± 4% perf-profile.children.cycles-pp.__sys_recvmsg
> > 3.60 ± 3% +2.5 6.12 ± 5% perf-profile.children.cycles-pp.sctp_skb_recv_datagram
> > 10.53 ± 3% +2.5 13.07 ± 4% perf-profile.children.cycles-pp.net_rx_action
> > 0.00 +2.5 2.54 ± 4% perf-profile.children.cycles-pp.refill_stock
> > 10.43 ± 3% +2.6 13.01 ± 4% perf-profile.children.cycles-pp.__napi_poll
> > 17.57 ± 3% +2.6 20.15 ± 4% perf-profile.children.cycles-pp.___sys_recvmsg
> > 10.41 ± 3% +2.6 12.99 ± 4% perf-profile.children.cycles-pp.process_backlog
> > 10.77 ± 3% +2.7 13.43 ± 4% perf-profile.children.cycles-pp.__softirqentry_text_start
> > 10.24 ± 3% +2.7 12.90 ± 4% perf-profile.children.cycles-pp.__netif_receive_skb_one_core
> > 9.97 ± 3% +2.8 12.76 ± 4% perf-profile.children.cycles-pp.ip_local_deliver_finish
> > 9.95 ± 3% +2.8 12.74 ± 4% perf-profile.children.cycles-pp.ip_protocol_deliver_rcu
> > 9.89 ± 3% +2.8 12.69 ± 4% perf-profile.children.cycles-pp.sctp_rcv
> > 1.71 ± 4% +2.9 4.57 ± 4% perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
> > 16.66 ± 3% +2.9 19.59 ± 4% perf-profile.children.cycles-pp.____sys_recvmsg
> > 16.01 ± 3% +3.2 19.21 ± 4% perf-profile.children.cycles-pp.sctp_recvmsg
> > 15.89 ± 3% +3.2 19.12 ± 4% perf-profile.children.cycles-pp.inet_recvmsg
> > 1.49 ± 3% +3.4 4.86 ± 4% perf-profile.children.cycles-pp._raw_spin_lock_bh
> > 1.59 ± 3% +3.4 4.96 ± 4% perf-profile.children.cycles-pp.lock_sock_nested
> > 15.60 ± 3% +3.6 19.22 ± 4% perf-profile.children.cycles-pp.release_sock
> > 0.00 +3.7 3.69 ± 4% perf-profile.children.cycles-pp.try_charge_memcg
> > 15.33 ± 3% +3.7 19.04 ± 4% perf-profile.children.cycles-pp.__release_sock
> > 15.27 ± 3% +3.7 19.00 ± 4% perf-profile.children.cycles-pp.sctp_backlog_rcv
> > 0.62 ± 5% +3.8 4.45 ± 4% perf-profile.children.cycles-pp.sctp_sf_eat_data_6_2
> > 0.48 ± 5% +3.9 4.36 ± 5% perf-profile.children.cycles-pp.sctp_eat_data
> > 0.00 +4.4 4.36 ± 4% perf-profile.children.cycles-pp.mem_cgroup_charge_skmem
> > 3.46 ± 3% +4.4 7.91 ± 4% perf-profile.children.cycles-pp.sctp_chunk_put
> > 20.07 ± 3% +4.6 24.66 ± 4% perf-profile.children.cycles-pp.sctp_assoc_bh_rcv
> > 0.00 +4.6 4.60 ± 4% perf-profile.children.cycles-pp.__sk_mem_raise_allocated
> > 0.00 +4.6 4.63 ± 4% perf-profile.children.cycles-pp.__sk_mem_schedule
> > 24.02 ± 12% +4.7 28.76 ± 11% perf-profile.children.cycles-pp.intel_idle
> > 2.97 ± 3% +4.8 7.75 ± 4% perf-profile.children.cycles-pp.sctp_outq_sack
> > 2.11 ± 4% +4.8 6.90 ± 4% perf-profile.children.cycles-pp.kfree_skb_reason
> > 2.04 ± 3% +4.9 6.98 ± 4% perf-profile.children.cycles-pp.consume_skb
> > 0.37 ± 3% +5.6 5.94 ± 4% perf-profile.children.cycles-pp.sctp_wfree
> > 0.95 ± 3% +11.3 12.23 ± 4% perf-profile.children.cycles-pp.skb_release_head_state
> > 0.00 +11.7 11.67 ± 4% perf-profile.children.cycles-pp.__sk_mem_reduce_allocated
> > 8.68 ± 3% -3.1 5.58 ± 4% perf-profile.self.cycles-pp.copy_user_enhanced_fast_string
> > 8.45 ± 3% -3.0 5.46 ± 5% perf-profile.self.cycles-pp.memcpy_erms
> > 0.94 ± 4% -0.4 0.57 ± 6% perf-profile.self.cycles-pp.__slab_free
> > 0.71 ± 3% -0.3 0.42 ± 2% perf-profile.self.cycles-pp.sctp_chunk_put
> > 0.58 ± 3% -0.3 0.30 ± 3% perf-profile.self.cycles-pp.__copy_skb_header
> > 0.87 ± 3% -0.3 0.59 ± 6% perf-profile.self.cycles-pp.kmem_cache_free
> > 0.58 ± 5% -0.3 0.30 ± 6% perf-profile.self.cycles-pp.sctp_packet_config
> > 0.82 ± 3% -0.3 0.55 ± 5% perf-profile.self.cycles-pp._raw_spin_lock_irqsave
> > 0.65 ± 4% -0.2 0.41 ± 5% perf-profile.self.cycles-pp.__might_resched
> > 0.41 ± 4% -0.2 0.17 ± 8% perf-profile.self.cycles-pp.rmqueue
> > 0.63 ± 4% -0.2 0.40 ± 3% perf-profile.self.cycles-pp._raw_spin_lock
> > 0.59 ± 4% -0.2 0.36 ± 7% perf-profile.self.cycles-pp.sctp_recvmsg
> > 0.39 ± 4% -0.2 0.18 ± 5% perf-profile.self.cycles-pp.sctp_sendmsg_to_asoc
> > 0.58 ± 3% -0.2 0.36 ± 3% perf-profile.self.cycles-pp._raw_spin_lock_bh
> > 0.43 ± 4% -0.2 0.24 ± 9% perf-profile.self.cycles-pp.sctp_packet_pack
> > 0.39 ± 4% -0.2 0.20 ± 2% perf-profile.self.cycles-pp.update_load_avg
> > 0.56 ± 4% -0.2 0.38 ± 5% perf-profile.self.cycles-pp.copy_user_short_string
> > 0.27 ± 5% -0.2 0.10 ± 7% perf-profile.self.cycles-pp.dst_release
> > 0.32 ± 3% -0.2 0.16 ± 7% perf-profile.self.cycles-pp.get_page_from_freelist
> > 0.31 ± 7% -0.2 0.15 ± 6% perf-profile.self.cycles-pp.ipv4_dst_check
> > 0.30 ± 4% -0.2 0.14 ± 7% perf-profile.self.cycles-pp.free_unref_page_commit
> > 0.38 ± 3% -0.2 0.22 ± 8% perf-profile.self.cycles-pp.__mod_node_page_state
> > 0.54 ± 3% -0.2 0.38 ± 5% perf-profile.self.cycles-pp.__schedule
> > 0.47 ± 2% -0.2 0.31 ± 5% perf-profile.self.cycles-pp.kfree
> > 0.34 ± 4% -0.2 0.18 ± 7% perf-profile.self.cycles-pp.sctp_v4_xmit
> > 0.23 ± 4% -0.2 0.07 ± 10% perf-profile.self.cycles-pp.update_curr
> > 0.32 ± 12% -0.2 0.17 ± 9% perf-profile.self.cycles-pp.__rhashtable_lookup
> > 0.51 ± 4% -0.2 0.36 ± 6% perf-profile.self.cycles-pp.__list_del_entry_valid
> > 0.32 ± 3% -0.2 0.17 ± 4% perf-profile.self.cycles-pp.__zone_watermark_ok
> > 0.42 ± 3% -0.1 0.27 ± 5% perf-profile.self.cycles-pp.kmem_cache_alloc
> > 0.40 ± 5% -0.1 0.26 ± 7% perf-profile.self.cycles-pp.sctp_outq_flush_data
> > 0.28 ± 4% -0.1 0.15 ± 6% perf-profile.self.cycles-pp.__check_object_size
> > 0.35 ± 4% -0.1 0.22 ± 9% perf-profile.self.cycles-pp.__fdget
> > 0.42 ± 4% -0.1 0.29 ± 6% perf-profile.self.cycles-pp.sctp_datamsg_from_user
> > 0.36 ± 5% -0.1 0.24 ± 5% perf-profile.self.cycles-pp.set_next_entity
> > 0.34 ± 3% -0.1 0.21 ± 3% perf-profile.self.cycles-pp.check_heap_object
> > 0.34 ± 3% -0.1 0.22 ± 7% perf-profile.self.cycles-pp.memcg_slab_free_hook
> > 0.32 ± 2% -0.1 0.20 ± 5% perf-profile.self.cycles-pp.__list_add_valid
> > 0.32 ± 5% -0.1 0.20 ± 8% perf-profile.self.cycles-pp.__virt_addr_valid
> > 0.34 ± 5% -0.1 0.22 ± 4% perf-profile.self.cycles-pp.__skb_datagram_iter
> > 0.34 ± 4% -0.1 0.22 ± 7% perf-profile.self.cycles-pp.update_rq_clock
> > 0.37 ± 5% -0.1 0.25 ± 4% perf-profile.self.cycles-pp.sctp_sendmsg
> > 0.27 ± 13% -0.1 0.16 ± 6% perf-profile.self.cycles-pp.sctp_transport_hold
> > 0.35 ± 3% -0.1 0.24 ± 6% perf-profile.self.cycles-pp.skb_release_data
> > 0.32 ± 4% -0.1 0.20 ± 7% perf-profile.self.cycles-pp.__alloc_skb
> > 0.22 ± 7% -0.1 0.10 ± 7% perf-profile.self.cycles-pp.sctp_assoc_rwnd_increase
> > 0.33 ± 5% -0.1 0.22 ± 4% perf-profile.self.cycles-pp.free_pcp_prepare
> > 0.31 ± 4% -0.1 0.20 ± 7% perf-profile.self.cycles-pp.enqueue_task_fair
> > 0.36 ± 3% -0.1 0.26 ± 3% perf-profile.self.cycles-pp.enqueue_entity
> > 0.30 -0.1 0.20 ± 8% perf-profile.self.cycles-pp.sctp_association_put
> > 0.24 ± 4% -0.1 0.14 ± 7% perf-profile.self.cycles-pp.__alloc_pages
> > 0.30 ± 9% -0.1 0.20 ± 6% perf-profile.self.cycles-pp.sctp_transport_put
> > 0.31 ± 3% -0.1 0.21 ± 6% perf-profile.self.cycles-pp.sctp_skb_recv_datagram
> > 0.30 ± 4% -0.1 0.20 ± 8% perf-profile.self.cycles-pp.aa_sk_perm
> > 0.33 ± 4% -0.1 0.23 ± 6% perf-profile.self.cycles-pp.entry_SYSCALL_64_after_hwframe
> > 0.30 ± 3% -0.1 0.20 ± 7% perf-profile.self.cycles-pp.__might_sleep
> > 0.18 ± 4% -0.1 0.09 ± 10% perf-profile.self.cycles-pp.____sys_recvmsg
> > 0.19 ± 5% -0.1 0.09 ± 11% perf-profile.self.cycles-pp.sctp_ulpevent_free
> > 0.32 ± 3% -0.1 0.23 ± 8% perf-profile.self.cycles-pp.memset_erms
> > 0.30 ± 6% -0.1 0.21 ± 7% perf-profile.self.cycles-pp.kmem_cache_alloc_node
> > 0.28 ± 3% -0.1 0.19 ± 6% perf-profile.self.cycles-pp.skb_set_owner_w
> > 0.23 ± 6% -0.1 0.14 ± 7% perf-profile.self.cycles-pp.__switch_to
> > 0.18 ± 4% -0.1 0.10 ± 5% perf-profile.self.cycles-pp.__update_load_avg_cfs_rq
> > 0.26 ± 4% -0.1 0.17 ± 7% perf-profile.self.cycles-pp.sctp_cmd_interpreter
> > 0.25 ± 4% -0.1 0.17 ± 7% perf-profile.self.cycles-pp.entry_SYSRETQ_unsafe_stack
> > 0.26 -0.1 0.18 ± 5% perf-profile.self.cycles-pp.send_sctp_stream_1toMany
> > 0.17 ± 32% -0.1 0.09 ± 6% perf-profile.self.cycles-pp.sctp_cmp_addr_exact
> > 0.22 ± 5% -0.1 0.14 ± 7% perf-profile.self.cycles-pp.__switch_to_asm
> > 0.20 ± 2% -0.1 0.12 ± 10% perf-profile.self.cycles-pp.recv_sctp_stream_1toMany
> > 0.19 ± 3% -0.1 0.11 ± 9% perf-profile.self.cycles-pp.sock_wfree
> > 0.10 ± 5% -0.1 0.02 ± 99% perf-profile.self.cycles-pp.process_backlog
> > 0.10 ± 6% -0.1 0.02 ± 99% perf-profile.self.cycles-pp.__free_one_page
> > 0.20 ± 4% -0.1 0.13 ± 5% perf-profile.self.cycles-pp.recvmsg
> > 0.22 ± 3% -0.1 0.14 ± 7% perf-profile.self.cycles-pp.consume_skb
> > 0.18 ± 3% -0.1 0.11 ± 10% perf-profile.self.cycles-pp._copy_from_user
> > 0.15 ± 7% -0.1 0.08 ± 4% perf-profile.self.cycles-pp.___perf_sw_event
> > 0.15 ± 4% -0.1 0.08 ± 5% perf-profile.self.cycles-pp.check_new_pages
> > 0.20 ± 5% -0.1 0.13 ± 6% perf-profile.self.cycles-pp.available_idle_cpu
> > 0.22 ± 3% -0.1 0.15 ± 5% perf-profile.self.cycles-pp.__entry_text_start
> > 0.18 ± 2% -0.1 0.11 ± 7% perf-profile.self.cycles-pp.sctp_assoc_bh_rcv
> > 0.13 ± 7% -0.1 0.06 ± 7% perf-profile.self.cycles-pp.sctp_sock_rfree
> > 0.17 ± 4% -0.1 0.11 ± 10% perf-profile.self.cycles-pp.syscall_return_via_sysret
> > 0.13 ± 5% -0.1 0.07 ± 13% perf-profile.self.cycles-pp.__free_pages_ok
> > 0.20 ± 5% -0.1 0.14 ± 9% perf-profile.self.cycles-pp.sctp_datamsg_put
> > 0.19 ± 3% -0.1 0.12 ± 12% perf-profile.self.cycles-pp.sendmsg
> > 0.18 ± 4% -0.1 0.12 ± 7% perf-profile.self.cycles-pp.reweight_entity
> > 0.14 ± 7% -0.1 0.07 ± 10% perf-profile.self.cycles-pp.sctp_inet_skb_msgname
> > 0.23 ± 3% -0.1 0.17 ± 9% perf-profile.self.cycles-pp.sctp_wfree
> > 0.19 ± 3% -0.1 0.13 ± 11% perf-profile.self.cycles-pp.update_cfs_group
> > 0.17 ± 4% -0.1 0.11 ± 3% perf-profile.self.cycles-pp.nr_iowait_cpu
> > 0.17 ± 7% -0.1 0.11 ± 9% perf-profile.self.cycles-pp.sctp_packet_transmit
> > 0.08 ± 18% -0.1 0.02 ± 99% perf-profile.self.cycles-pp.sctp_ulpq_tail_data
> > 0.17 ± 7% -0.1 0.11 ± 9% perf-profile.self.cycles-pp.sctp_packet_append_chunk
> > 0.19 ± 4% -0.1 0.13 ± 8% perf-profile.self.cycles-pp.sctp_do_sm
> > 0.18 ± 2% -0.1 0.12 ± 4% perf-profile.self.cycles-pp.__free_pages
> > 0.17 ± 5% -0.1 0.12 ± 6% perf-profile.self.cycles-pp.__wake_up_common
> > 0.16 ± 6% -0.1 0.10 ± 3% perf-profile.self.cycles-pp.sctp_sendmsg_parse
> > 0.15 ± 6% -0.1 0.09 ± 13% perf-profile.self.cycles-pp.try_to_wake_up
> > 0.10 ± 3% -0.1 0.04 ± 45% perf-profile.self.cycles-pp.__build_skb_around
> > 0.15 ± 6% -0.1 0.09 ± 11% perf-profile.self.cycles-pp.sctp_check_transmitted
> > 0.08 ± 14% -0.1 0.03 ±101% perf-profile.self.cycles-pp.select_task_rq_fair
> > 0.12 ± 6% -0.1 0.06 ± 14% perf-profile.self.cycles-pp.dequeue_entity
> > 0.19 ± 4% -0.1 0.14 ± 10% perf-profile.self.cycles-pp.sctp_hash_cmp
> > 0.14 ± 4% -0.1 0.09 ± 10% perf-profile.self.cycles-pp.switch_mm_irqs_off
> > 0.16 ± 4% -0.1 0.10 ± 7% perf-profile.self.cycles-pp.poll_idle
> > 0.22 ± 8% -0.1 0.17 ± 10% perf-profile.self.cycles-pp.sctp_rcv
> > 0.11 ± 4% -0.1 0.06 ± 11% perf-profile.self.cycles-pp.__ip_queue_xmit
> > 0.12 ± 10% -0.1 0.08 ± 24% perf-profile.self.cycles-pp.__sctp_packet_append_chunk
> > 0.13 ± 2% -0.0 0.08 ± 5% perf-profile.self.cycles-pp.sctp_inq_pop
> > 0.18 ± 3% -0.0 0.13 ± 11% perf-profile.self.cycles-pp.__update_load_avg_se
> > 0.14 ± 5% -0.0 0.09 ± 5% perf-profile.self.cycles-pp.do_syscall_64
> > 0.11 ± 6% -0.0 0.06 ± 11% perf-profile.self.cycles-pp._copy_to_iter
> > 0.09 ± 11% -0.0 0.04 ± 76% perf-profile.self.cycles-pp.sctp_chunk_assign_ssn
> > 0.14 ± 3% -0.0 0.10 ± 5% perf-profile.self.cycles-pp.____sys_sendmsg
> > 0.12 ± 4% -0.0 0.07 ± 6% perf-profile.self.cycles-pp.___sys_recvmsg
> > 0.20 ± 2% -0.0 0.15 ± 8% perf-profile.self.cycles-pp.sctp_association_hold
> > 0.17 ± 3% -0.0 0.12 ± 7% perf-profile.self.cycles-pp.__skb_clone
> > 0.14 ± 3% -0.0 0.09 ± 7% perf-profile.self.cycles-pp.__copy_msghdr_from_user
> > 0.13 ± 5% -0.0 0.08 ± 8% perf-profile.self.cycles-pp.sctp_sf_eat_data_6_2
> > 0.13 ± 2% -0.0 0.09 ± 13% perf-profile.self.cycles-pp.sctp_outq_tail
> > 0.20 ± 4% -0.0 0.16 ± 9% perf-profile.self.cycles-pp.read_tsc
> > 0.13 ± 5% -0.0 0.08 ± 5% perf-profile.self.cycles-pp.__import_iovec
> > 0.13 ± 6% -0.0 0.08 ± 5% perf-profile.self.cycles-pp.__genradix_ptr
> > 0.11 ± 8% -0.0 0.07 ± 14% perf-profile.self.cycles-pp.select_task_rq
> > 0.12 ± 3% -0.0 0.08 ± 10% perf-profile.self.cycles-pp.resched_curr
> > 0.13 ± 5% -0.0 0.09 ± 5% perf-profile.self.cycles-pp.pick_next_entity
> > 0.12 ± 6% -0.0 0.08 ± 12% perf-profile.self.cycles-pp.__put_user_nocheck_4
> > 0.12 ± 6% -0.0 0.08 ± 6% perf-profile.self.cycles-pp.sctp_chunk_abandoned
> > 0.10 ± 7% -0.0 0.06 ± 11% perf-profile.self.cycles-pp.syscall_exit_to_user_mode_prepare
> > 0.07 ± 7% -0.0 0.02 ± 99% perf-profile.self.cycles-pp.ip_rcv_core
> > 0.13 ± 7% -0.0 0.09 ± 7% perf-profile.self.cycles-pp.sctp_sched_dequeue_common
> > 0.13 ± 7% -0.0 0.09 ± 6% perf-profile.self.cycles-pp.ip_finish_output2
> > 0.12 ± 6% -0.0 0.08 ± 10% perf-profile.self.cycles-pp.memcg_slab_post_alloc_hook
> > 0.08 ± 6% -0.0 0.04 ± 71% perf-profile.self.cycles-pp._sctp_make_chunk
> > 0.10 ± 4% -0.0 0.06 ± 11% perf-profile.self.cycles-pp.net_rx_action
> > 0.10 ± 5% -0.0 0.06 ± 8% perf-profile.self.cycles-pp.sctp_addto_chunk
> > 0.09 ± 6% -0.0 0.05 ± 7% perf-profile.self.cycles-pp.__netif_receive_skb_core
> > 0.12 ± 5% -0.0 0.08 ± 5% perf-profile.self.cycles-pp.__kmalloc
> > 0.11 ± 5% -0.0 0.07 ± 5% perf-profile.self.cycles-pp.__put_user_nocheck_8
> > 0.18 ± 6% -0.0 0.15 ± 7% perf-profile.self.cycles-pp.native_sched_clock
> > 0.09 ± 5% -0.0 0.06 ± 8% perf-profile.self.cycles-pp.check_stack_object
> > 0.07 -0.0 0.03 ± 70% perf-profile.self.cycles-pp.__mod_lruvec_page_state
> > 0.16 ± 4% -0.0 0.12 ± 10% perf-profile.self.cycles-pp.sctp_outq_flush
> > 0.10 ± 3% -0.0 0.06 ± 7% perf-profile.self.cycles-pp.skb_put
> > 0.06 ± 6% -0.0 0.02 ± 99% perf-profile.self.cycles-pp.entry_SYSCALL_64_safe_stack
> > 0.10 ± 4% -0.0 0.07 ± 15% perf-profile.self.cycles-pp.__might_fault
> > 0.12 ± 5% -0.0 0.08 ± 5% perf-profile.self.cycles-pp.kmem_cache_alloc_trace
> > 0.07 ± 8% -0.0 0.04 ± 71% perf-profile.self.cycles-pp.sockfd_lookup_light
> > 0.10 ± 7% -0.0 0.07 ± 10% perf-profile.self.cycles-pp.__cond_resched
> > 0.09 ± 6% -0.0 0.06 ± 8% perf-profile.self.cycles-pp.os_xsave
> > 0.09 ± 6% -0.0 0.06 ± 8% perf-profile.self.cycles-pp.free_unref_page
> > 0.09 ± 7% -0.0 0.06 ± 6% perf-profile.self.cycles-pp.security_socket_recvmsg
> > 0.11 ± 4% -0.0 0.08 ± 7% perf-profile.self.cycles-pp.__local_bh_enable_ip
> > 0.12 ± 4% -0.0 0.09 ± 12% perf-profile.self.cycles-pp.sock_kmalloc
> > 0.09 ± 6% -0.0 0.06 ± 6% perf-profile.self.cycles-pp.sctp_user_addto_chunk
> > 0.07 ± 6% -0.0 0.04 ± 44% perf-profile.self.cycles-pp.rcu_all_qs
> > 0.10 ± 5% -0.0 0.06 ± 11% perf-profile.self.cycles-pp.sctp_outq_sack
> > 0.09 ± 4% -0.0 0.06 ± 6% perf-profile.self.cycles-pp.sctp_outq_select_transport
> > 0.10 ± 3% -0.0 0.07 ± 9% perf-profile.self.cycles-pp.__kmalloc_node_track_caller
> > 0.08 ± 5% -0.0 0.05 ± 46% perf-profile.self.cycles-pp.sctp_ulpevent_init
> > 0.12 ± 4% -0.0 0.08 ± 8% perf-profile.self.cycles-pp.__check_heap_object
> > 0.07 ± 11% -0.0 0.04 ± 44% perf-profile.self.cycles-pp.loopback_xmit
> > 0.14 ± 2% -0.0 0.12 ± 6% perf-profile.self.cycles-pp.do_idle
> > 0.11 ± 4% -0.0 0.08 ± 12% perf-profile.self.cycles-pp.cpuidle_idle_call
> > 0.09 -0.0 0.06 ± 11% perf-profile.self.cycles-pp._copy_from_iter
> > 0.08 ± 5% -0.0 0.06 ± 13% perf-profile.self.cycles-pp.kmalloc_large_node
> > 0.08 ± 5% -0.0 0.06 ± 9% perf-profile.self.cycles-pp.sctp_chunkify
> > 0.09 ± 4% -0.0 0.06 ± 7% perf-profile.self.cycles-pp.iovec_from_user
> > 0.14 ± 4% -0.0 0.12 ± 11% perf-profile.self.cycles-pp.sctp_ulpevent_make_rcvmsg
> > 0.07 ± 6% -0.0 0.05 ± 7% perf-profile.self.cycles-pp.sctp_tsnmap_check
> > 0.08 ± 6% -0.0 0.06 ± 9% perf-profile.self.cycles-pp.sock_kfree_s
> > 0.08 ± 5% -0.0 0.06 ± 11% perf-profile.self.cycles-pp.skb_release_head_state
> > 0.07 ± 8% -0.0 0.05 perf-profile.self.cycles-pp.__wrgsbase_inactive
> > 0.09 ± 8% -0.0 0.07 ± 13% perf-profile.self.cycles-pp.__softirqentry_text_start
> > 0.07 ± 5% -0.0 0.05 ± 8% perf-profile.self.cycles-pp.__sys_recvmsg
> > 0.18 ± 4% +0.0 0.22 ± 12% perf-profile.self.cycles-pp.menu_select
> > 0.00 +0.1 0.06 ± 19% perf-profile.self.cycles-pp.update_sg_lb_stats
> > 0.00 +0.1 0.06 ± 7% perf-profile.self.cycles-pp.lapic_next_deadline
> > 0.00 +0.1 0.07 ± 16% perf-profile.self.cycles-pp.native_irq_return_iret
> > 0.09 ± 4% +0.1 0.16 ± 5% perf-profile.self.cycles-pp.cpuidle_enter_state
> > 0.00 +0.1 0.08 ± 41% perf-profile.self.cycles-pp._raw_spin_trylock
> > 0.00 +0.1 0.09 ± 54% perf-profile.self.cycles-pp.tick_nohz_next_event
> > 0.01 ±223% +0.1 0.13 ± 5% perf-profile.self.cycles-pp.sctp_get_af_specific
> > 0.00 +0.1 0.13 ± 6% perf-profile.self.cycles-pp.mem_cgroup_charge_skmem
> > 0.00 +0.1 0.14 ± 19% perf-profile.self.cycles-pp.cgroup_rstat_updated
> > 0.25 ± 6% +0.2 0.43 ± 14% perf-profile.self.cycles-pp.ktime_get
> > 0.00 +0.2 0.20 ± 16% perf-profile.self.cycles-pp.timekeeping_max_deferment
> > 0.00 +0.2 0.22 ± 6% perf-profile.self.cycles-pp.refill_stock
> > 0.00 +0.3 0.25 ± 9% perf-profile.self.cycles-pp.__sk_mem_raise_allocated
> > 0.00 +0.3 0.32 ± 12% perf-profile.self.cycles-pp.propagate_protected_usage
> > 0.00 +0.9 0.89 ± 4% perf-profile.self.cycles-pp.__mod_memcg_state
> > 0.00 +1.3 1.27 ± 5% perf-profile.self.cycles-pp.try_charge_memcg
> > 0.00 +2.1 2.10 ± 5% perf-profile.self.cycles-pp.page_counter_uncharge
> > 0.00 +2.2 2.23 ± 5% perf-profile.self.cycles-pp.page_counter_try_charge
> > 1.70 ± 4% +2.8 4.54 ± 4% perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
> > 0.29 ± 5% +4.0 4.24 ± 5% perf-profile.self.cycles-pp.sctp_eat_data
> > 0.00 +8.6 8.55 ± 4% perf-profile.self.cycles-pp.__sk_mem_reduce_allocated
> >
> >
> >
> >
> >
> > Disclaimer:
> > Results have been estimated based on internal Intel analysis and are provided
> > for informational purposes only. Any difference in system hardware or software
> > design or configuration may affect actual performance.
> >
> >
>

2022-06-23 23:01:57

by Xin Long

[permalink] [raw]
Subject: Re: [net] 4890b686f4: netperf.Throughput_Mbps -69.4% regression

On Wed, Jun 22, 2022 at 11:08 PM Xin Long <[email protected]> wrote:
>
> Yes, I'm working on it. I couldn't see the regression in my env with
> the 'reproduce' script attached.
> I will try with lkp tomorrow.
>
> Thanks.
>
> On Wed, Jun 22, 2022 at 8:29 PM Jakub Kicinski <[email protected]> wrote:
> >
> > Could someone working on SCTP double check this is a real regression?
> > Feels like the regression reports are flowing at such rate its hard
> > to keep up.
> >
> > >
> > > commit:
> > > 7c80b038d2 ("net: fix sk_wmem_schedule() and sk_rmem_schedule() errors")
> > > 4890b686f4 ("net: keep sk->sk_forward_alloc as small as possible")
> > >
> > > 7c80b038d23e1f4c 4890b686f4088c90432149bd6de
> > > ---------------- ---------------------------
> > > %stddev %change %stddev
> > > \ | \
> > > 15855 -69.4% 4854 netperf.Throughput_Mbps
> > > 570788 -69.4% 174773 netperf.Throughput_total_Mbps
...
> > > 0.00 +5.1 5.10 ± 5% perf-profile.calltrace.cycles-pp.__sk_mem_reduce_allocated.sctp_wfree.skb_release_head_state.consume_skb.sctp_chunk_put
> > > 0.17 ±141% +5.3 5.42 ± 6% perf-profile.calltrace.cycles-pp.skb_release_head_state.consume_skb.sctp_chunk_put.sctp_outq_sack.sctp_cmd_interpreter
> > > 0.00 +5.3 5.35 ± 6% perf-profile.calltrace.cycles-pp.sctp_wfree.skb_release_head_state.consume_skb.sctp_chunk_put.sctp_outq_sack
> > > 0.00 +5.5 5.51 ± 6% perf-profile.calltrace.cycles-pp.__sk_mem_reduce_allocated.skb_release_head_state.kfree_skb_reason.sctp_recvmsg.inet_recvmsg
> > > 0.00 +5.7 5.65 ± 6% perf-profile.calltrace.cycles-pp.skb_release_head_state.kfree_skb_reason.sctp_recvmsg.inet_recvmsg.____sys_recvmsg
...
> > > 0.00 +4.0 4.04 ± 6% perf-profile.children.cycles-pp.mem_cgroup_charge_skmem
> > > 2.92 ± 6% +4.2 7.16 ± 6% perf-profile.children.cycles-pp.sctp_outq_sack
> > > 0.00 +4.3 4.29 ± 6% perf-profile.children.cycles-pp.__sk_mem_raise_allocated
> > > 0.00 +4.3 4.32 ± 6% perf-profile.children.cycles-pp.__sk_mem_schedule
> > > 1.99 ± 6% +4.4 6.40 ± 6% perf-profile.children.cycles-pp.consume_skb
> > > 1.78 ± 6% +4.6 6.42 ± 6% perf-profile.children.cycles-pp.kfree_skb_reason
> > > 0.37 ± 8% +5.0 5.40 ± 6% perf-profile.children.cycles-pp.sctp_wfree
> > > 0.87 ± 9% +10.3 11.20 ± 6% perf-profile.children.cycles-pp.skb_release_head_state
> > > 0.00 +10.7 10.66 ± 6% perf-profile.children.cycles-pp.__sk_mem_reduce_allocated
...
> > > 0.00 +1.2 1.19 ± 7% perf-profile.self.cycles-pp.try_charge_memcg
> > > 0.00 +2.0 1.96 ± 6% perf-profile.self.cycles-pp.page_counter_uncharge
> > > 0.00 +2.1 2.07 ± 5% perf-profile.self.cycles-pp.page_counter_try_charge
> > > 1.09 ± 8% +2.8 3.92 ± 6% perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
> > > 0.29 ± 6% +3.5 3.81 ± 6% perf-profile.self.cycles-pp.sctp_eat_data
> > > 0.00 +7.8 7.76 ± 6% perf-profile.self.cycles-pp.__sk_mem_reduce_allocated

From the perf data, we can see __sk_mem_reduce_allocated() is the one
using CPU the most more than before, and mem_cgroup APIs are also
called in this function. It means the mem cgroup must be enabled in
the test env, which may explain why I couldn't reproduce it.

The Commit 4890b686f4 ("net: keep sk->sk_forward_alloc as small as
possible") uses sk_mem_reclaim(checking reclaimable >= PAGE_SIZE) to
reclaim the memory, which is *more frequent* to call
__sk_mem_reduce_allocated() than before (checking reclaimable >=
SK_RECLAIM_THRESHOLD). It might be cheap when
mem_cgroup_sockets_enabled is false, but I'm not sure if it's still
cheap when mem_cgroup_sockets_enabled is true.

I think SCTP netperf could trigger this, as the CPU is the bottleneck
for SCTP netperf testing, which is more sensitive to the extra
function calls than TCP.

Can we re-run this testing without mem cgroup enabled?

Thanks.

2022-06-24 02:08:22

by Jakub Kicinski

[permalink] [raw]
Subject: Re: [net] 4890b686f4: netperf.Throughput_Mbps -69.4% regression

On Thu, 23 Jun 2022 18:50:07 -0400 Xin Long wrote:
> From the perf data, we can see __sk_mem_reduce_allocated() is the one
> using CPU the most more than before, and mem_cgroup APIs are also
> called in this function. It means the mem cgroup must be enabled in
> the test env, which may explain why I couldn't reproduce it.
>
> The Commit 4890b686f4 ("net: keep sk->sk_forward_alloc as small as
> possible") uses sk_mem_reclaim(checking reclaimable >= PAGE_SIZE) to
> reclaim the memory, which is *more frequent* to call
> __sk_mem_reduce_allocated() than before (checking reclaimable >=
> SK_RECLAIM_THRESHOLD). It might be cheap when
> mem_cgroup_sockets_enabled is false, but I'm not sure if it's still
> cheap when mem_cgroup_sockets_enabled is true.
>
> I think SCTP netperf could trigger this, as the CPU is the bottleneck
> for SCTP netperf testing, which is more sensitive to the extra
> function calls than TCP.
>
> Can we re-run this testing without mem cgroup enabled?

FWIW I defer to Eric, thanks a lot for double checking the report
and digging in!

2022-06-24 04:21:55

by Eric Dumazet

[permalink] [raw]
Subject: Re: [net] 4890b686f4: netperf.Throughput_Mbps -69.4% regression

On Fri, Jun 24, 2022 at 3:57 AM Jakub Kicinski <[email protected]> wrote:
>
> On Thu, 23 Jun 2022 18:50:07 -0400 Xin Long wrote:
> > From the perf data, we can see __sk_mem_reduce_allocated() is the one
> > using CPU the most more than before, and mem_cgroup APIs are also
> > called in this function. It means the mem cgroup must be enabled in
> > the test env, which may explain why I couldn't reproduce it.
> >
> > The Commit 4890b686f4 ("net: keep sk->sk_forward_alloc as small as
> > possible") uses sk_mem_reclaim(checking reclaimable >= PAGE_SIZE) to
> > reclaim the memory, which is *more frequent* to call
> > __sk_mem_reduce_allocated() than before (checking reclaimable >=
> > SK_RECLAIM_THRESHOLD). It might be cheap when
> > mem_cgroup_sockets_enabled is false, but I'm not sure if it's still
> > cheap when mem_cgroup_sockets_enabled is true.
> >
> > I think SCTP netperf could trigger this, as the CPU is the bottleneck
> > for SCTP netperf testing, which is more sensitive to the extra
> > function calls than TCP.
> >
> > Can we re-run this testing without mem cgroup enabled?
>
> FWIW I defer to Eric, thanks a lot for double checking the report
> and digging in!

I did tests with TCP + memcg and noticed a very small additional cost
in memcg functions,
because of suboptimal layout:

Extract of an internal Google bug, update from June 9th:

--------------------------------
I have noticed a minor false sharing to fetch (struct
mem_cgroup)->css.parent, at offset 0xc0,
because it shares the cache line containing struct mem_cgroup.memory,
at offset 0xd0

Ideally, memcg->socket_pressure and memcg->parent should sit in a read
mostly cache line.
-----------------------

But nothing that could explain a "-69.4% regression"

memcg has a very similar strategy of per-cpu reserves, with
MEMCG_CHARGE_BATCH being 32 pages per cpu.

It is not clear why SCTP with 10K writes would overflow this reserve constantly.

Presumably memcg experts will have to rework structure alignments to
make sure they can cope better
with more charge/uncharge operations, because we are not going back to
gigantic per-socket reserves,
this simply does not scale.

2022-06-24 04:25:21

by Eric Dumazet

[permalink] [raw]
Subject: Re: [net] 4890b686f4: netperf.Throughput_Mbps -69.4% regression

On Fri, Jun 24, 2022 at 6:13 AM Eric Dumazet <[email protected]> wrote:
>
> On Fri, Jun 24, 2022 at 3:57 AM Jakub Kicinski <[email protected]> wrote:
> >
> > On Thu, 23 Jun 2022 18:50:07 -0400 Xin Long wrote:
> > > From the perf data, we can see __sk_mem_reduce_allocated() is the one
> > > using CPU the most more than before, and mem_cgroup APIs are also
> > > called in this function. It means the mem cgroup must be enabled in
> > > the test env, which may explain why I couldn't reproduce it.
> > >
> > > The Commit 4890b686f4 ("net: keep sk->sk_forward_alloc as small as
> > > possible") uses sk_mem_reclaim(checking reclaimable >= PAGE_SIZE) to
> > > reclaim the memory, which is *more frequent* to call
> > > __sk_mem_reduce_allocated() than before (checking reclaimable >=
> > > SK_RECLAIM_THRESHOLD). It might be cheap when
> > > mem_cgroup_sockets_enabled is false, but I'm not sure if it's still
> > > cheap when mem_cgroup_sockets_enabled is true.
> > >
> > > I think SCTP netperf could trigger this, as the CPU is the bottleneck
> > > for SCTP netperf testing, which is more sensitive to the extra
> > > function calls than TCP.
> > >
> > > Can we re-run this testing without mem cgroup enabled?
> >
> > FWIW I defer to Eric, thanks a lot for double checking the report
> > and digging in!
>
> I did tests with TCP + memcg and noticed a very small additional cost
> in memcg functions,
> because of suboptimal layout:
>
> Extract of an internal Google bug, update from June 9th:
>
> --------------------------------
> I have noticed a minor false sharing to fetch (struct
> mem_cgroup)->css.parent, at offset 0xc0,
> because it shares the cache line containing struct mem_cgroup.memory,
> at offset 0xd0
>
> Ideally, memcg->socket_pressure and memcg->parent should sit in a read
> mostly cache line.
> -----------------------
>
> But nothing that could explain a "-69.4% regression"

I guess the test now hits memcg limits more often, forcing expensive reclaim,
and the memcg limits need some adjustments.

Overall, tests enabling memcg should probably need fine tuning, I will
defer to Intel folks.


>
> memcg has a very similar strategy of per-cpu reserves, with
> MEMCG_CHARGE_BATCH being 32 pages per cpu.
>
> It is not clear why SCTP with 10K writes would overflow this reserve constantly.
>
> Presumably memcg experts will have to rework structure alignments to
> make sure they can cope better
> with more charge/uncharge operations, because we are not going back to
> gigantic per-socket reserves,
> this simply does not scale.

2022-06-24 05:21:38

by Feng Tang

[permalink] [raw]
Subject: Re: [net] 4890b686f4: netperf.Throughput_Mbps -69.4% regression

Hi Eric,

On Fri, Jun 24, 2022 at 06:13:51AM +0200, Eric Dumazet wrote:
> On Fri, Jun 24, 2022 at 3:57 AM Jakub Kicinski <[email protected]> wrote:
> >
> > On Thu, 23 Jun 2022 18:50:07 -0400 Xin Long wrote:
> > > From the perf data, we can see __sk_mem_reduce_allocated() is the one
> > > using CPU the most more than before, and mem_cgroup APIs are also
> > > called in this function. It means the mem cgroup must be enabled in
> > > the test env, which may explain why I couldn't reproduce it.
> > >
> > > The Commit 4890b686f4 ("net: keep sk->sk_forward_alloc as small as
> > > possible") uses sk_mem_reclaim(checking reclaimable >= PAGE_SIZE) to
> > > reclaim the memory, which is *more frequent* to call
> > > __sk_mem_reduce_allocated() than before (checking reclaimable >=
> > > SK_RECLAIM_THRESHOLD). It might be cheap when
> > > mem_cgroup_sockets_enabled is false, but I'm not sure if it's still
> > > cheap when mem_cgroup_sockets_enabled is true.
> > >
> > > I think SCTP netperf could trigger this, as the CPU is the bottleneck
> > > for SCTP netperf testing, which is more sensitive to the extra
> > > function calls than TCP.
> > >
> > > Can we re-run this testing without mem cgroup enabled?
> >
> > FWIW I defer to Eric, thanks a lot for double checking the report
> > and digging in!
>
> I did tests with TCP + memcg and noticed a very small additional cost
> in memcg functions,
> because of suboptimal layout:
>
> Extract of an internal Google bug, update from June 9th:
>
> --------------------------------
> I have noticed a minor false sharing to fetch (struct
> mem_cgroup)->css.parent, at offset 0xc0,
> because it shares the cache line containing struct mem_cgroup.memory,
> at offset 0xd0
>
> Ideally, memcg->socket_pressure and memcg->parent should sit in a read
> mostly cache line.
> -----------------------
>
> But nothing that could explain a "-69.4% regression"

We can double check that.

> memcg has a very similar strategy of per-cpu reserves, with
> MEMCG_CHARGE_BATCH being 32 pages per cpu.

We have proposed patch to increase the batch numer for stats
update, which was not accepted as it hurts the accuracy and
the data is used by many tools.

> It is not clear why SCTP with 10K writes would overflow this reserve constantly.
>
> Presumably memcg experts will have to rework structure alignments to
> make sure they can cope better
> with more charge/uncharge operations, because we are not going back to
> gigantic per-socket reserves,
> this simply does not scale.

Yes, the memcg statitics and charge/unchage update is very sensitive
with the data alignemnt layout, and can easily trigger peformance
changes, as we've seen quite some similar cases in the past several
years.

One pattern we've seen is, even if a memcg stats updating or charge
function only takes about 2%~3% of the CPU cycles in perf-profile data,
once it got affected, the peformance change could be amplified to up to
60% or more.

Thanks,
Feng


2022-06-24 06:07:51

by Eric Dumazet

[permalink] [raw]
Subject: Re: [net] 4890b686f4: netperf.Throughput_Mbps -69.4% regression

On Fri, Jun 24, 2022 at 7:14 AM Feng Tang <[email protected]> wrote:
>
> Hi Eric,
>
> On Fri, Jun 24, 2022 at 06:13:51AM +0200, Eric Dumazet wrote:
> > On Fri, Jun 24, 2022 at 3:57 AM Jakub Kicinski <[email protected]> wrote:
> > >
> > > On Thu, 23 Jun 2022 18:50:07 -0400 Xin Long wrote:
> > > > From the perf data, we can see __sk_mem_reduce_allocated() is the one
> > > > using CPU the most more than before, and mem_cgroup APIs are also
> > > > called in this function. It means the mem cgroup must be enabled in
> > > > the test env, which may explain why I couldn't reproduce it.
> > > >
> > > > The Commit 4890b686f4 ("net: keep sk->sk_forward_alloc as small as
> > > > possible") uses sk_mem_reclaim(checking reclaimable >= PAGE_SIZE) to
> > > > reclaim the memory, which is *more frequent* to call
> > > > __sk_mem_reduce_allocated() than before (checking reclaimable >=
> > > > SK_RECLAIM_THRESHOLD). It might be cheap when
> > > > mem_cgroup_sockets_enabled is false, but I'm not sure if it's still
> > > > cheap when mem_cgroup_sockets_enabled is true.
> > > >
> > > > I think SCTP netperf could trigger this, as the CPU is the bottleneck
> > > > for SCTP netperf testing, which is more sensitive to the extra
> > > > function calls than TCP.
> > > >
> > > > Can we re-run this testing without mem cgroup enabled?
> > >
> > > FWIW I defer to Eric, thanks a lot for double checking the report
> > > and digging in!
> >
> > I did tests with TCP + memcg and noticed a very small additional cost
> > in memcg functions,
> > because of suboptimal layout:
> >
> > Extract of an internal Google bug, update from June 9th:
> >
> > --------------------------------
> > I have noticed a minor false sharing to fetch (struct
> > mem_cgroup)->css.parent, at offset 0xc0,
> > because it shares the cache line containing struct mem_cgroup.memory,
> > at offset 0xd0
> >
> > Ideally, memcg->socket_pressure and memcg->parent should sit in a read
> > mostly cache line.
> > -----------------------
> >
> > But nothing that could explain a "-69.4% regression"
>
> We can double check that.
>
> > memcg has a very similar strategy of per-cpu reserves, with
> > MEMCG_CHARGE_BATCH being 32 pages per cpu.
>
> We have proposed patch to increase the batch numer for stats
> update, which was not accepted as it hurts the accuracy and
> the data is used by many tools.
>
> > It is not clear why SCTP with 10K writes would overflow this reserve constantly.
> >
> > Presumably memcg experts will have to rework structure alignments to
> > make sure they can cope better
> > with more charge/uncharge operations, because we are not going back to
> > gigantic per-socket reserves,
> > this simply does not scale.
>
> Yes, the memcg statitics and charge/unchage update is very sensitive
> with the data alignemnt layout, and can easily trigger peformance
> changes, as we've seen quite some similar cases in the past several
> years.
>
> One pattern we've seen is, even if a memcg stats updating or charge
> function only takes about 2%~3% of the CPU cycles in perf-profile data,
> once it got affected, the peformance change could be amplified to up to
> 60% or more.
>

Reorganizing "struct mem_cgroup" to put "struct page_counter memory"
in a separate cache line would be beneficial.

Many low hanging fruits, assuming nobody will use __randomize_layout on it ;)

Also some fields are written even if their value is not changed.

diff --git a/mm/memcontrol.c b/mm/memcontrol.c
index abec50f31fe64100f4be5b029c7161b3a6077a74..53d9c1e581e78303ef73942e2b34338567987b74
100644
--- a/mm/memcontrol.c
+++ b/mm/memcontrol.c
@@ -7037,10 +7037,12 @@ bool mem_cgroup_charge_skmem(struct mem_cgroup
*memcg, unsigned int nr_pages,
struct page_counter *fail;

if (page_counter_try_charge(&memcg->tcpmem, nr_pages, &fail)) {
- memcg->tcpmem_pressure = 0;
+ if (READ_ONCE(memcg->tcpmem_pressure))
+ WRITE_ONCE(memcg->tcpmem_pressure, 0);
return true;
}
- memcg->tcpmem_pressure = 1;
+ if (!READ_ONCE(memcg->tcpmem_pressure))
+ WRITE_ONCE(memcg->tcpmem_pressure, 1);
if (gfp_mask & __GFP_NOFAIL) {
page_counter_charge(&memcg->tcpmem, nr_pages);
return true;

2022-06-24 06:09:19

by Feng Tang

[permalink] [raw]
Subject: Re: [net] 4890b686f4: netperf.Throughput_Mbps -69.4% regression

On Fri, Jun 24, 2022 at 07:45:00AM +0200, Eric Dumazet wrote:
> On Fri, Jun 24, 2022 at 7:14 AM Feng Tang <[email protected]> wrote:
> >
> > Hi Eric,
> >
> > On Fri, Jun 24, 2022 at 06:13:51AM +0200, Eric Dumazet wrote:
> > > On Fri, Jun 24, 2022 at 3:57 AM Jakub Kicinski <[email protected]> wrote:
> > > >
> > > > On Thu, 23 Jun 2022 18:50:07 -0400 Xin Long wrote:
> > > > > From the perf data, we can see __sk_mem_reduce_allocated() is the one
> > > > > using CPU the most more than before, and mem_cgroup APIs are also
> > > > > called in this function. It means the mem cgroup must be enabled in
> > > > > the test env, which may explain why I couldn't reproduce it.
> > > > >
> > > > > The Commit 4890b686f4 ("net: keep sk->sk_forward_alloc as small as
> > > > > possible") uses sk_mem_reclaim(checking reclaimable >= PAGE_SIZE) to
> > > > > reclaim the memory, which is *more frequent* to call
> > > > > __sk_mem_reduce_allocated() than before (checking reclaimable >=
> > > > > SK_RECLAIM_THRESHOLD). It might be cheap when
> > > > > mem_cgroup_sockets_enabled is false, but I'm not sure if it's still
> > > > > cheap when mem_cgroup_sockets_enabled is true.
> > > > >
> > > > > I think SCTP netperf could trigger this, as the CPU is the bottleneck
> > > > > for SCTP netperf testing, which is more sensitive to the extra
> > > > > function calls than TCP.
> > > > >
> > > > > Can we re-run this testing without mem cgroup enabled?
> > > >
> > > > FWIW I defer to Eric, thanks a lot for double checking the report
> > > > and digging in!
> > >
> > > I did tests with TCP + memcg and noticed a very small additional cost
> > > in memcg functions,
> > > because of suboptimal layout:
> > >
> > > Extract of an internal Google bug, update from June 9th:
> > >
> > > --------------------------------
> > > I have noticed a minor false sharing to fetch (struct
> > > mem_cgroup)->css.parent, at offset 0xc0,
> > > because it shares the cache line containing struct mem_cgroup.memory,
> > > at offset 0xd0
> > >
> > > Ideally, memcg->socket_pressure and memcg->parent should sit in a read
> > > mostly cache line.
> > > -----------------------
> > >
> > > But nothing that could explain a "-69.4% regression"
> >
> > We can double check that.
> >
> > > memcg has a very similar strategy of per-cpu reserves, with
> > > MEMCG_CHARGE_BATCH being 32 pages per cpu.
> >
> > We have proposed patch to increase the batch numer for stats
> > update, which was not accepted as it hurts the accuracy and
> > the data is used by many tools.
> >
> > > It is not clear why SCTP with 10K writes would overflow this reserve constantly.
> > >
> > > Presumably memcg experts will have to rework structure alignments to
> > > make sure they can cope better
> > > with more charge/uncharge operations, because we are not going back to
> > > gigantic per-socket reserves,
> > > this simply does not scale.
> >
> > Yes, the memcg statitics and charge/unchage update is very sensitive
> > with the data alignemnt layout, and can easily trigger peformance
> > changes, as we've seen quite some similar cases in the past several
> > years.
> >
> > One pattern we've seen is, even if a memcg stats updating or charge
> > function only takes about 2%~3% of the CPU cycles in perf-profile data,
> > once it got affected, the peformance change could be amplified to up to
> > 60% or more.
> >
>
> Reorganizing "struct mem_cgroup" to put "struct page_counter memory"
> in a separate cache line would be beneficial.

That may help.

And I also want to say the benchmarks(especially micro one) are very
sensitive to the layout of mem_cgroup. As the 'page_counter' is 112
bytes in size, I recently made a patch to make it cacheline aligned
(take 2 cachelines), which improved some hackbench/netperf test
cases, but caused huge (49%) drop for some vm-scalability tests.

> Many low hanging fruits, assuming nobody will use __randomize_layout on it ;)
>
> Also some fields are written even if their value is not changed.
>
> diff --git a/mm/memcontrol.c b/mm/memcontrol.c
> index abec50f31fe64100f4be5b029c7161b3a6077a74..53d9c1e581e78303ef73942e2b34338567987b74
> 100644
> --- a/mm/memcontrol.c
> +++ b/mm/memcontrol.c
> @@ -7037,10 +7037,12 @@ bool mem_cgroup_charge_skmem(struct mem_cgroup
> *memcg, unsigned int nr_pages,
> struct page_counter *fail;
>
> if (page_counter_try_charge(&memcg->tcpmem, nr_pages, &fail)) {
> - memcg->tcpmem_pressure = 0;
> + if (READ_ONCE(memcg->tcpmem_pressure))
> + WRITE_ONCE(memcg->tcpmem_pressure, 0);
> return true;
> }
> - memcg->tcpmem_pressure = 1;
> + if (!READ_ONCE(memcg->tcpmem_pressure))
> + WRITE_ONCE(memcg->tcpmem_pressure, 1);
> if (gfp_mask & __GFP_NOFAIL) {
> page_counter_charge(&memcg->tcpmem, nr_pages);
> return true;

I will also try this patch, which may take some time.

Thanks,
Feng

2022-06-24 06:17:04

by Eric Dumazet

[permalink] [raw]
Subject: Re: [net] 4890b686f4: netperf.Throughput_Mbps -69.4% regression

On Fri, Jun 24, 2022 at 8:01 AM Feng Tang <[email protected]> wrote:
>
> On Fri, Jun 24, 2022 at 07:45:00AM +0200, Eric Dumazet wrote:
> > On Fri, Jun 24, 2022 at 7:14 AM Feng Tang <[email protected]> wrote:
> > >
> > > Hi Eric,
> > >
> > > On Fri, Jun 24, 2022 at 06:13:51AM +0200, Eric Dumazet wrote:
> > > > On Fri, Jun 24, 2022 at 3:57 AM Jakub Kicinski <[email protected]> wrote:
> > > > >
> > > > > On Thu, 23 Jun 2022 18:50:07 -0400 Xin Long wrote:
> > > > > > From the perf data, we can see __sk_mem_reduce_allocated() is the one
> > > > > > using CPU the most more than before, and mem_cgroup APIs are also
> > > > > > called in this function. It means the mem cgroup must be enabled in
> > > > > > the test env, which may explain why I couldn't reproduce it.
> > > > > >
> > > > > > The Commit 4890b686f4 ("net: keep sk->sk_forward_alloc as small as
> > > > > > possible") uses sk_mem_reclaim(checking reclaimable >= PAGE_SIZE) to
> > > > > > reclaim the memory, which is *more frequent* to call
> > > > > > __sk_mem_reduce_allocated() than before (checking reclaimable >=
> > > > > > SK_RECLAIM_THRESHOLD). It might be cheap when
> > > > > > mem_cgroup_sockets_enabled is false, but I'm not sure if it's still
> > > > > > cheap when mem_cgroup_sockets_enabled is true.
> > > > > >
> > > > > > I think SCTP netperf could trigger this, as the CPU is the bottleneck
> > > > > > for SCTP netperf testing, which is more sensitive to the extra
> > > > > > function calls than TCP.
> > > > > >
> > > > > > Can we re-run this testing without mem cgroup enabled?
> > > > >
> > > > > FWIW I defer to Eric, thanks a lot for double checking the report
> > > > > and digging in!
> > > >
> > > > I did tests with TCP + memcg and noticed a very small additional cost
> > > > in memcg functions,
> > > > because of suboptimal layout:
> > > >
> > > > Extract of an internal Google bug, update from June 9th:
> > > >
> > > > --------------------------------
> > > > I have noticed a minor false sharing to fetch (struct
> > > > mem_cgroup)->css.parent, at offset 0xc0,
> > > > because it shares the cache line containing struct mem_cgroup.memory,
> > > > at offset 0xd0
> > > >
> > > > Ideally, memcg->socket_pressure and memcg->parent should sit in a read
> > > > mostly cache line.
> > > > -----------------------
> > > >
> > > > But nothing that could explain a "-69.4% regression"
> > >
> > > We can double check that.
> > >
> > > > memcg has a very similar strategy of per-cpu reserves, with
> > > > MEMCG_CHARGE_BATCH being 32 pages per cpu.
> > >
> > > We have proposed patch to increase the batch numer for stats
> > > update, which was not accepted as it hurts the accuracy and
> > > the data is used by many tools.
> > >
> > > > It is not clear why SCTP with 10K writes would overflow this reserve constantly.
> > > >
> > > > Presumably memcg experts will have to rework structure alignments to
> > > > make sure they can cope better
> > > > with more charge/uncharge operations, because we are not going back to
> > > > gigantic per-socket reserves,
> > > > this simply does not scale.
> > >
> > > Yes, the memcg statitics and charge/unchage update is very sensitive
> > > with the data alignemnt layout, and can easily trigger peformance
> > > changes, as we've seen quite some similar cases in the past several
> > > years.
> > >
> > > One pattern we've seen is, even if a memcg stats updating or charge
> > > function only takes about 2%~3% of the CPU cycles in perf-profile data,
> > > once it got affected, the peformance change could be amplified to up to
> > > 60% or more.
> > >
> >
> > Reorganizing "struct mem_cgroup" to put "struct page_counter memory"
> > in a separate cache line would be beneficial.
>
> That may help.
>
> And I also want to say the benchmarks(especially micro one) are very
> sensitive to the layout of mem_cgroup. As the 'page_counter' is 112
> bytes in size, I recently made a patch to make it cacheline aligned
> (take 2 cachelines), which improved some hackbench/netperf test
> cases, but caused huge (49%) drop for some vm-scalability tests.
>
> > Many low hanging fruits, assuming nobody will use __randomize_layout on it ;)
> >
> > Also some fields are written even if their value is not changed.
> >
> > diff --git a/mm/memcontrol.c b/mm/memcontrol.c
> > index abec50f31fe64100f4be5b029c7161b3a6077a74..53d9c1e581e78303ef73942e2b34338567987b74
> > 100644
> > --- a/mm/memcontrol.c
> > +++ b/mm/memcontrol.c
> > @@ -7037,10 +7037,12 @@ bool mem_cgroup_charge_skmem(struct mem_cgroup
> > *memcg, unsigned int nr_pages,
> > struct page_counter *fail;
> >
> > if (page_counter_try_charge(&memcg->tcpmem, nr_pages, &fail)) {
> > - memcg->tcpmem_pressure = 0;
> > + if (READ_ONCE(memcg->tcpmem_pressure))
> > + WRITE_ONCE(memcg->tcpmem_pressure, 0);
> > return true;
> > }
> > - memcg->tcpmem_pressure = 1;
> > + if (!READ_ONCE(memcg->tcpmem_pressure))
> > + WRITE_ONCE(memcg->tcpmem_pressure, 1);
> > if (gfp_mask & __GFP_NOFAIL) {
> > page_counter_charge(&memcg->tcpmem, nr_pages);
> > return true;
>
> I will also try this patch, which may take some time.

Note that applications can opt-in reserving memory for one socket,
using SO_RESERVE_MEM

This can be used for jobs with a controlled number of sockets, as this
will avoid many charge/uncharge operations.

2022-06-24 06:43:12

by Shakeel Butt

[permalink] [raw]
Subject: Re: [net] 4890b686f4: netperf.Throughput_Mbps -69.4% regression

CCing memcg folks.

The thread starts at
https://lore.kernel.org/all/20220619150456.GB34471@xsang-OptiPlex-9020/

On Thu, Jun 23, 2022 at 9:14 PM Eric Dumazet <[email protected]> wrote:
>
> On Fri, Jun 24, 2022 at 3:57 AM Jakub Kicinski <[email protected]> wrote:
> >
> > On Thu, 23 Jun 2022 18:50:07 -0400 Xin Long wrote:
> > > From the perf data, we can see __sk_mem_reduce_allocated() is the one
> > > using CPU the most more than before, and mem_cgroup APIs are also
> > > called in this function. It means the mem cgroup must be enabled in
> > > the test env, which may explain why I couldn't reproduce it.
> > >
> > > The Commit 4890b686f4 ("net: keep sk->sk_forward_alloc as small as
> > > possible") uses sk_mem_reclaim(checking reclaimable >= PAGE_SIZE) to
> > > reclaim the memory, which is *more frequent* to call
> > > __sk_mem_reduce_allocated() than before (checking reclaimable >=
> > > SK_RECLAIM_THRESHOLD). It might be cheap when
> > > mem_cgroup_sockets_enabled is false, but I'm not sure if it's still
> > > cheap when mem_cgroup_sockets_enabled is true.
> > >
> > > I think SCTP netperf could trigger this, as the CPU is the bottleneck
> > > for SCTP netperf testing, which is more sensitive to the extra
> > > function calls than TCP.
> > >
> > > Can we re-run this testing without mem cgroup enabled?
> >
> > FWIW I defer to Eric, thanks a lot for double checking the report
> > and digging in!
>
> I did tests with TCP + memcg and noticed a very small additional cost
> in memcg functions,
> because of suboptimal layout:
>
> Extract of an internal Google bug, update from June 9th:
>
> --------------------------------
> I have noticed a minor false sharing to fetch (struct
> mem_cgroup)->css.parent, at offset 0xc0,
> because it shares the cache line containing struct mem_cgroup.memory,
> at offset 0xd0
>
> Ideally, memcg->socket_pressure and memcg->parent should sit in a read
> mostly cache line.
> -----------------------
>
> But nothing that could explain a "-69.4% regression"
>
> memcg has a very similar strategy of per-cpu reserves, with
> MEMCG_CHARGE_BATCH being 32 pages per cpu.
>
> It is not clear why SCTP with 10K writes would overflow this reserve constantly.
>
> Presumably memcg experts will have to rework structure alignments to
> make sure they can cope better
> with more charge/uncharge operations, because we are not going back to
> gigantic per-socket reserves,
> this simply does not scale.

Yes I agree. As you pointed out there are fields which are mostly
read-only but sharing cache lines with fields which get updated and
definitely need work.

However can we first confirm if memcg charging is really the issue
here as I remember these intel lkp tests are configured to run in root
memcg and the kernel does not associate root memcg to any socket (see
mem_cgroup_sk_alloc()).

If these tests are running in non-root memcg, is this cgroup v1 or v2?
The memory counter and the 32 pages per cpu stock are only used on v2.
For v1, there is no per-cpu stock and there is a separate tcpmem page
counter and on v1 the network memory accounting has to be enabled
explicitly i.e. not enabled by default.

There is definite possibility of slowdown on v1 but let's first
confirm the memcg setup used for this testing environment.

Feng, can you please explain the memcg setup on these test machines
and if the tests are run in root or non-root memcg?

thanks,
Shakeel

2022-06-24 07:30:35

by Feng Tang

[permalink] [raw]
Subject: Re: [net] 4890b686f4: netperf.Throughput_Mbps -69.4% regression

On Thu, Jun 23, 2022 at 11:34:15PM -0700, Shakeel Butt wrote:
> CCing memcg folks.
>
> The thread starts at
> https://lore.kernel.org/all/20220619150456.GB34471@xsang-OptiPlex-9020/
>
> On Thu, Jun 23, 2022 at 9:14 PM Eric Dumazet <[email protected]> wrote:
> >
> > On Fri, Jun 24, 2022 at 3:57 AM Jakub Kicinski <[email protected]> wrote:
> > >
> > > On Thu, 23 Jun 2022 18:50:07 -0400 Xin Long wrote:
> > > > From the perf data, we can see __sk_mem_reduce_allocated() is the one
> > > > using CPU the most more than before, and mem_cgroup APIs are also
> > > > called in this function. It means the mem cgroup must be enabled in
> > > > the test env, which may explain why I couldn't reproduce it.
> > > >
> > > > The Commit 4890b686f4 ("net: keep sk->sk_forward_alloc as small as
> > > > possible") uses sk_mem_reclaim(checking reclaimable >= PAGE_SIZE) to
> > > > reclaim the memory, which is *more frequent* to call
> > > > __sk_mem_reduce_allocated() than before (checking reclaimable >=
> > > > SK_RECLAIM_THRESHOLD). It might be cheap when
> > > > mem_cgroup_sockets_enabled is false, but I'm not sure if it's still
> > > > cheap when mem_cgroup_sockets_enabled is true.
> > > >
> > > > I think SCTP netperf could trigger this, as the CPU is the bottleneck
> > > > for SCTP netperf testing, which is more sensitive to the extra
> > > > function calls than TCP.
> > > >
> > > > Can we re-run this testing without mem cgroup enabled?
> > >
> > > FWIW I defer to Eric, thanks a lot for double checking the report
> > > and digging in!
> >
> > I did tests with TCP + memcg and noticed a very small additional cost
> > in memcg functions,
> > because of suboptimal layout:
> >
> > Extract of an internal Google bug, update from June 9th:
> >
> > --------------------------------
> > I have noticed a minor false sharing to fetch (struct
> > mem_cgroup)->css.parent, at offset 0xc0,
> > because it shares the cache line containing struct mem_cgroup.memory,
> > at offset 0xd0
> >
> > Ideally, memcg->socket_pressure and memcg->parent should sit in a read
> > mostly cache line.
> > -----------------------
> >
> > But nothing that could explain a "-69.4% regression"
> >
> > memcg has a very similar strategy of per-cpu reserves, with
> > MEMCG_CHARGE_BATCH being 32 pages per cpu.
> >
> > It is not clear why SCTP with 10K writes would overflow this reserve constantly.
> >
> > Presumably memcg experts will have to rework structure alignments to
> > make sure they can cope better
> > with more charge/uncharge operations, because we are not going back to
> > gigantic per-socket reserves,
> > this simply does not scale.
>
> Yes I agree. As you pointed out there are fields which are mostly
> read-only but sharing cache lines with fields which get updated and
> definitely need work.
>
> However can we first confirm if memcg charging is really the issue
> here as I remember these intel lkp tests are configured to run in root
> memcg and the kernel does not associate root memcg to any socket (see
> mem_cgroup_sk_alloc()).
>
> If these tests are running in non-root memcg, is this cgroup v1 or v2?
> The memory counter and the 32 pages per cpu stock are only used on v2.
> For v1, there is no per-cpu stock and there is a separate tcpmem page
> counter and on v1 the network memory accounting has to be enabled
> explicitly i.e. not enabled by default.
>
> There is definite possibility of slowdown on v1 but let's first
> confirm the memcg setup used for this testing environment.
>
> Feng, can you please explain the memcg setup on these test machines
> and if the tests are run in root or non-root memcg?

I don't know the exact setup, Philip/Oliver from 0Day can correct me.

I logged into a test box which runs netperf test, and it seems to be
cgoup v1 and non-root memcg. The netperf tasks all sit in dir:
'/sys/fs/cgroup/memory/system.slice/lkp-bootstrap.service'

And the rootfs is a debian based rootfs

Thanks,
Feng


> thanks,
> Shakeel

2022-06-24 15:15:06

by Shakeel Butt

[permalink] [raw]
Subject: Re: [net] 4890b686f4: netperf.Throughput_Mbps -69.4% regression

On Fri, Jun 24, 2022 at 03:06:56PM +0800, Feng Tang wrote:
> On Thu, Jun 23, 2022 at 11:34:15PM -0700, Shakeel Butt wrote:
[...]
> >
> > Feng, can you please explain the memcg setup on these test machines
> > and if the tests are run in root or non-root memcg?
>
> I don't know the exact setup, Philip/Oliver from 0Day can correct me.
>
> I logged into a test box which runs netperf test, and it seems to be
> cgoup v1 and non-root memcg. The netperf tasks all sit in dir:
> '/sys/fs/cgroup/memory/system.slice/lkp-bootstrap.service'
>

Thanks Feng. Can you check the value of memory.kmem.tcp.max_usage_in_bytes
in /sys/fs/cgroup/memory/system.slice/lkp-bootstrap.service after making
sure that the netperf test has already run?

If this is non-zero then network memory accounting is enabled and the
slowdown is expected.

> And the rootfs is a debian based rootfs
>
> Thanks,
> Feng
>
>
> > thanks,
> > Shakeel

2022-06-25 02:39:49

by Feng Tang

[permalink] [raw]
Subject: Re: [net] 4890b686f4: netperf.Throughput_Mbps -69.4% regression

On Fri, Jun 24, 2022 at 02:43:58PM +0000, Shakeel Butt wrote:
> On Fri, Jun 24, 2022 at 03:06:56PM +0800, Feng Tang wrote:
> > On Thu, Jun 23, 2022 at 11:34:15PM -0700, Shakeel Butt wrote:
> [...]
> > >
> > > Feng, can you please explain the memcg setup on these test machines
> > > and if the tests are run in root or non-root memcg?
> >
> > I don't know the exact setup, Philip/Oliver from 0Day can correct me.
> >
> > I logged into a test box which runs netperf test, and it seems to be
> > cgoup v1 and non-root memcg. The netperf tasks all sit in dir:
> > '/sys/fs/cgroup/memory/system.slice/lkp-bootstrap.service'
> >
>
> Thanks Feng. Can you check the value of memory.kmem.tcp.max_usage_in_bytes
> in /sys/fs/cgroup/memory/system.slice/lkp-bootstrap.service after making
> sure that the netperf test has already run?

memory.kmem.tcp.max_usage_in_bytes:0

And here is more memcg stats (let me know if you want to check more)

/sys/fs/cgroup/memory/system.slice/lkp-bootstrap.service# grep . memory.*
memory.failcnt:0
memory.kmem.failcnt:0
memory.kmem.limit_in_bytes:9223372036854771712
memory.kmem.max_usage_in_bytes:47861760
memory.kmem.tcp.failcnt:0
memory.kmem.tcp.limit_in_bytes:9223372036854771712
memory.kmem.tcp.max_usage_in_bytes:0
memory.kmem.tcp.usage_in_bytes:0
memory.kmem.usage_in_bytes:40730624
memory.limit_in_bytes:9223372036854771712
memory.max_usage_in_bytes:642424832
memory.memsw.failcnt:0
memory.memsw.limit_in_bytes:9223372036854771712
memory.memsw.max_usage_in_bytes:642424832
memory.memsw.usage_in_bytes:639549440
memory.move_charge_at_immigrate:0
memory.numa_stat:total=144073 N0=124819 N1=19254
memory.numa_stat:file=0 N0=0 N1=0
memory.numa_stat:anon=77721 N0=58502 N1=19219
memory.numa_stat:unevictable=66352 N0=66317 N1=35
memory.numa_stat:hierarchical_total=144073 N0=124819 N1=19254
memory.numa_stat:hierarchical_file=0 N0=0 N1=0
memory.numa_stat:hierarchical_anon=77721 N0=58502 N1=19219
memory.numa_stat:hierarchical_unevictable=66352 N0=66317 N1=35
memory.oom_control:oom_kill_disable 0
memory.oom_control:under_oom 0
memory.oom_control:oom_kill 0
grep: memory.pressure_level: Invalid argument
memory.soft_limit_in_bytes:9223372036854771712
memory.stat:cache 282562560
memory.stat:rss 307884032
memory.stat:rss_huge 239075328
memory.stat:shmem 10784768
memory.stat:mapped_file 3444736
memory.stat:dirty 0
memory.stat:writeback 0
memory.stat:swap 0
memory.stat:pgpgin 1018918
memory.stat:pgpgout 932902
memory.stat:pgfault 2130513
memory.stat:pgmajfault 0
memory.stat:inactive_anon 310272000
memory.stat:active_anon 8073216
memory.stat:inactive_file 0
memory.stat:active_file 0
memory.stat:unevictable 271777792
memory.stat:hierarchical_memory_limit 9223372036854771712
memory.stat:hierarchical_memsw_limit 9223372036854771712
memory.stat:total_cache 282562560
memory.stat:total_rss 307884032
memory.stat:total_rss_huge 239075328
memory.stat:total_shmem 10784768
memory.stat:total_mapped_file 3444736
memory.stat:total_dirty 0
memory.stat:total_writeback 0
memory.stat:total_swap 0
memory.stat:total_pgpgin 1018918
memory.stat:total_pgpgout 932902
memory.stat:total_pgfault 2130513
memory.stat:total_pgmajfault 0
memory.stat:total_inactive_anon 310272000
memory.stat:total_active_anon 8073216
memory.stat:total_inactive_file 0
memory.stat:total_active_file 0
memory.stat:total_unevictable 271777792
memory.swappiness:60
memory.usage_in_bytes:639549440
memory.use_hierarchy:1

> If this is non-zero then network memory accounting is enabled and the
> slowdown is expected.

From the perf-profile data in original report, both
__sk_mem_raise_allocated() and __sk_mem_reduce_allocated() are called
much more often, which call memcg charge/uncharge functions.

IIUC, the call chain is:

__sk_mem_raise_allocated
sk_memory_allocated_add
mem_cgroup_charge_skmem
charge memcg->tcpmem (for cgroup v2)
try_charge memcg (for v1)

Also from Eric's one earlier commit log:

"
net: implement per-cpu reserves for memory_allocated
...
This means we are going to call sk_memory_allocated_add()
and sk_memory_allocated_sub() more often.
...
"

So this slowdown is related to the more calling of charge/uncharge?

Thanks,
Feng

> > And the rootfs is a debian based rootfs
> >
> > Thanks,
> > Feng
> >
> >
> > > thanks,
> > > Shakeel

2022-06-27 02:53:23

by Feng Tang

[permalink] [raw]
Subject: Re: [net] 4890b686f4: netperf.Throughput_Mbps -69.4% regression

On Sat, Jun 25, 2022 at 10:36:42AM +0800, Feng Tang wrote:
> On Fri, Jun 24, 2022 at 02:43:58PM +0000, Shakeel Butt wrote:
> > On Fri, Jun 24, 2022 at 03:06:56PM +0800, Feng Tang wrote:
> > > On Thu, Jun 23, 2022 at 11:34:15PM -0700, Shakeel Butt wrote:
> > [...]
> > > >
> > > > Feng, can you please explain the memcg setup on these test machines
> > > > and if the tests are run in root or non-root memcg?
> > >
> > > I don't know the exact setup, Philip/Oliver from 0Day can correct me.
> > >
> > > I logged into a test box which runs netperf test, and it seems to be
> > > cgoup v1 and non-root memcg. The netperf tasks all sit in dir:
> > > '/sys/fs/cgroup/memory/system.slice/lkp-bootstrap.service'
> > >
> >
> > Thanks Feng. Can you check the value of memory.kmem.tcp.max_usage_in_bytes
> > in /sys/fs/cgroup/memory/system.slice/lkp-bootstrap.service after making
> > sure that the netperf test has already run?
>
> memory.kmem.tcp.max_usage_in_bytes:0

Sorry, I made a mistake that in the original report from Oliver, it
was 'cgroup v2' with a 'debian-11.1' rootfs.

When you asked about cgroup info, I tried the job on another tbox, and
the original 'job.yaml' didn't work, so I kept the 'netperf' test
parameters and started a new job which somehow run with a 'debian-10.4'
rootfs and acutally run with cgroup v1.

And as you mentioned cgroup version does make a big difference, that
with v1, the regression is reduced to 1% ~ 5% on different generations
of test platforms. Eric mentioned they also got regression report,
but much smaller one, maybe it's due to the cgroup version?

Thanks,
Feng

> And here is more memcg stats (let me know if you want to check more)
>
> > If this is non-zero then network memory accounting is enabled and the
> > slowdown is expected.
>
> >From the perf-profile data in original report, both
> __sk_mem_raise_allocated() and __sk_mem_reduce_allocated() are called
> much more often, which call memcg charge/uncharge functions.
>
> IIUC, the call chain is:
>
> __sk_mem_raise_allocated
> sk_memory_allocated_add
> mem_cgroup_charge_skmem
> charge memcg->tcpmem (for cgroup v2)
> try_charge memcg (for v1)
>
> Also from Eric's one earlier commit log:
>
> "
> net: implement per-cpu reserves for memory_allocated
> ...
> This means we are going to call sk_memory_allocated_add()
> and sk_memory_allocated_sub() more often.
> ...
> "
>
> So this slowdown is related to the more calling of charge/uncharge?
>
> Thanks,
> Feng
>
> > > And the rootfs is a debian based rootfs
> > >
> > > Thanks,
> > > Feng
> > >
> > >
> > > > thanks,
> > > > Shakeel

2022-06-27 08:55:52

by Eric Dumazet

[permalink] [raw]
Subject: Re: [net] 4890b686f4: netperf.Throughput_Mbps -69.4% regression

On Mon, Jun 27, 2022 at 4:38 AM Feng Tang <[email protected]> wrote:
>
> On Sat, Jun 25, 2022 at 10:36:42AM +0800, Feng Tang wrote:
> > On Fri, Jun 24, 2022 at 02:43:58PM +0000, Shakeel Butt wrote:
> > > On Fri, Jun 24, 2022 at 03:06:56PM +0800, Feng Tang wrote:
> > > > On Thu, Jun 23, 2022 at 11:34:15PM -0700, Shakeel Butt wrote:
> > > [...]
> > > > >
> > > > > Feng, can you please explain the memcg setup on these test machines
> > > > > and if the tests are run in root or non-root memcg?
> > > >
> > > > I don't know the exact setup, Philip/Oliver from 0Day can correct me.
> > > >
> > > > I logged into a test box which runs netperf test, and it seems to be
> > > > cgoup v1 and non-root memcg. The netperf tasks all sit in dir:
> > > > '/sys/fs/cgroup/memory/system.slice/lkp-bootstrap.service'
> > > >
> > >
> > > Thanks Feng. Can you check the value of memory.kmem.tcp.max_usage_in_bytes
> > > in /sys/fs/cgroup/memory/system.slice/lkp-bootstrap.service after making
> > > sure that the netperf test has already run?
> >
> > memory.kmem.tcp.max_usage_in_bytes:0
>
> Sorry, I made a mistake that in the original report from Oliver, it
> was 'cgroup v2' with a 'debian-11.1' rootfs.
>
> When you asked about cgroup info, I tried the job on another tbox, and
> the original 'job.yaml' didn't work, so I kept the 'netperf' test
> parameters and started a new job which somehow run with a 'debian-10.4'
> rootfs and acutally run with cgroup v1.
>
> And as you mentioned cgroup version does make a big difference, that
> with v1, the regression is reduced to 1% ~ 5% on different generations
> of test platforms. Eric mentioned they also got regression report,
> but much smaller one, maybe it's due to the cgroup version?

This was using the current net-next tree.
Used recipe was something like:

Make sure cgroup2 is mounted or mount it by mount -t cgroup2 none $MOUNT_POINT.
Enable memory controller by echo +memory > $MOUNT_POINT/cgroup.subtree_control.
Create a cgroup by mkdir $MOUNT_POINT/job.
Jump into that cgroup by echo $$ > $MOUNT_POINT/job/cgroup.procs.

<Launch tests>

The regression was smaller than 1%, so considered noise compared to
the benefits of the bug fix.

>
> Thanks,
> Feng
>
> > And here is more memcg stats (let me know if you want to check more)
> >
> > > If this is non-zero then network memory accounting is enabled and the
> > > slowdown is expected.
> >
> > >From the perf-profile data in original report, both
> > __sk_mem_raise_allocated() and __sk_mem_reduce_allocated() are called
> > much more often, which call memcg charge/uncharge functions.
> >
> > IIUC, the call chain is:
> >
> > __sk_mem_raise_allocated
> > sk_memory_allocated_add
> > mem_cgroup_charge_skmem
> > charge memcg->tcpmem (for cgroup v2)
> > try_charge memcg (for v1)
> >
> > Also from Eric's one earlier commit log:
> >
> > "
> > net: implement per-cpu reserves for memory_allocated
> > ...
> > This means we are going to call sk_memory_allocated_add()
> > and sk_memory_allocated_sub() more often.
> > ...
> > "
> >
> > So this slowdown is related to the more calling of charge/uncharge?
> >
> > Thanks,
> > Feng
> >
> > > > And the rootfs is a debian based rootfs
> > > >
> > > > Thanks,
> > > > Feng
> > > >
> > > >
> > > > > thanks,
> > > > > Shakeel

2022-06-27 12:37:07

by Feng Tang

[permalink] [raw]
Subject: Re: [net] 4890b686f4: netperf.Throughput_Mbps -69.4% regression

On Mon, Jun 27, 2022 at 10:46:21AM +0200, Eric Dumazet wrote:
> On Mon, Jun 27, 2022 at 4:38 AM Feng Tang <[email protected]> wrote:
[snip]
> > > >
> > > > Thanks Feng. Can you check the value of memory.kmem.tcp.max_usage_in_bytes
> > > > in /sys/fs/cgroup/memory/system.slice/lkp-bootstrap.service after making
> > > > sure that the netperf test has already run?
> > >
> > > memory.kmem.tcp.max_usage_in_bytes:0
> >
> > Sorry, I made a mistake that in the original report from Oliver, it
> > was 'cgroup v2' with a 'debian-11.1' rootfs.
> >
> > When you asked about cgroup info, I tried the job on another tbox, and
> > the original 'job.yaml' didn't work, so I kept the 'netperf' test
> > parameters and started a new job which somehow run with a 'debian-10.4'
> > rootfs and acutally run with cgroup v1.
> >
> > And as you mentioned cgroup version does make a big difference, that
> > with v1, the regression is reduced to 1% ~ 5% on different generations
> > of test platforms. Eric mentioned they also got regression report,
> > but much smaller one, maybe it's due to the cgroup version?
>
> This was using the current net-next tree.
> Used recipe was something like:
>
> Make sure cgroup2 is mounted or mount it by mount -t cgroup2 none $MOUNT_POINT.
> Enable memory controller by echo +memory > $MOUNT_POINT/cgroup.subtree_control.
> Create a cgroup by mkdir $MOUNT_POINT/job.
> Jump into that cgroup by echo $$ > $MOUNT_POINT/job/cgroup.procs.
>
> <Launch tests>
>
> The regression was smaller than 1%, so considered noise compared to
> the benefits of the bug fix.

Yes, 1% is just around noise level for a microbenchmark.

I went check the original test data of Oliver's report, the tests was
run 6 rounds and the performance data is pretty stable (0Day's report
will show any std deviation bigger than 2%)

The test platform is a 4 sockets 72C/144T machine, and I run the
same job (nr_tasks = 25% * nr_cpus) on one CascadeLake AP (4 nodes)
and one Icelake 2 sockets platform, and saw 75% and 53% regresson on
them.

In the first email, there is a file named 'reproduce', it shows the
basic test process:

"
use 'performane' cpufre governor for all CPUs

netserver -4 -D
modprobe sctp
netperf -4 -H 127.0.0.1 -t SCTP_STREAM_MANY -c -C -l 300 -- -m 10K &
netperf -4 -H 127.0.0.1 -t SCTP_STREAM_MANY -c -C -l 300 -- -m 10K &
netperf -4 -H 127.0.0.1 -t SCTP_STREAM_MANY -c -C -l 300 -- -m 10K &
(repeat 36 times in total)
...

"

Which starts 36 (25% of nr_cpus) netperf clients. And the clients number
also matters, I tried to increase the client number from 36 to 72(50%),
and the regression is changed from 69.4% to 73.7%

Thanks,
Feng

> >
> > Thanks,
> > Feng

2022-06-27 14:38:46

by Eric Dumazet

[permalink] [raw]
Subject: Re: [net] 4890b686f4: netperf.Throughput_Mbps -69.4% regression

On Mon, Jun 27, 2022 at 2:34 PM Feng Tang <[email protected]> wrote:
>
> On Mon, Jun 27, 2022 at 10:46:21AM +0200, Eric Dumazet wrote:
> > On Mon, Jun 27, 2022 at 4:38 AM Feng Tang <[email protected]> wrote:
> [snip]
> > > > >
> > > > > Thanks Feng. Can you check the value of memory.kmem.tcp.max_usage_in_bytes
> > > > > in /sys/fs/cgroup/memory/system.slice/lkp-bootstrap.service after making
> > > > > sure that the netperf test has already run?
> > > >
> > > > memory.kmem.tcp.max_usage_in_bytes:0
> > >
> > > Sorry, I made a mistake that in the original report from Oliver, it
> > > was 'cgroup v2' with a 'debian-11.1' rootfs.
> > >
> > > When you asked about cgroup info, I tried the job on another tbox, and
> > > the original 'job.yaml' didn't work, so I kept the 'netperf' test
> > > parameters and started a new job which somehow run with a 'debian-10.4'
> > > rootfs and acutally run with cgroup v1.
> > >
> > > And as you mentioned cgroup version does make a big difference, that
> > > with v1, the regression is reduced to 1% ~ 5% on different generations
> > > of test platforms. Eric mentioned they also got regression report,
> > > but much smaller one, maybe it's due to the cgroup version?
> >
> > This was using the current net-next tree.
> > Used recipe was something like:
> >
> > Make sure cgroup2 is mounted or mount it by mount -t cgroup2 none $MOUNT_POINT.
> > Enable memory controller by echo +memory > $MOUNT_POINT/cgroup.subtree_control.
> > Create a cgroup by mkdir $MOUNT_POINT/job.
> > Jump into that cgroup by echo $$ > $MOUNT_POINT/job/cgroup.procs.
> >
> > <Launch tests>
> >
> > The regression was smaller than 1%, so considered noise compared to
> > the benefits of the bug fix.
>
> Yes, 1% is just around noise level for a microbenchmark.
>
> I went check the original test data of Oliver's report, the tests was
> run 6 rounds and the performance data is pretty stable (0Day's report
> will show any std deviation bigger than 2%)
>
> The test platform is a 4 sockets 72C/144T machine, and I run the
> same job (nr_tasks = 25% * nr_cpus) on one CascadeLake AP (4 nodes)
> and one Icelake 2 sockets platform, and saw 75% and 53% regresson on
> them.
>
> In the first email, there is a file named 'reproduce', it shows the
> basic test process:
>
> "
> use 'performane' cpufre governor for all CPUs
>
> netserver -4 -D
> modprobe sctp
> netperf -4 -H 127.0.0.1 -t SCTP_STREAM_MANY -c -C -l 300 -- -m 10K &
> netperf -4 -H 127.0.0.1 -t SCTP_STREAM_MANY -c -C -l 300 -- -m 10K &
> netperf -4 -H 127.0.0.1 -t SCTP_STREAM_MANY -c -C -l 300 -- -m 10K &
> (repeat 36 times in total)
> ...
>
> "
>
> Which starts 36 (25% of nr_cpus) netperf clients. And the clients number
> also matters, I tried to increase the client number from 36 to 72(50%),
> and the regression is changed from 69.4% to 73.7%"
>

This seems like a lot of opportunities for memcg folks :)

struct page_counter has poor field placement [1], and no per-cpu cache.

[1] "atomic_long_t usage" is sharing cache line with read mostly fields.

(struct mem_cgroup also has poor field placement, mainly because of
struct page_counter)

28.69% [kernel] [k] copy_user_enhanced_fast_string
16.13% [kernel] [k] intel_idle_irq
6.46% [kernel] [k] page_counter_try_charge
6.20% [kernel] [k] __sk_mem_reduce_allocated
5.68% [kernel] [k] try_charge_memcg
5.16% [kernel] [k] page_counter_cancel



> Thanks,
> Feng
>
> > >
> > > Thanks,
> > > Feng

2022-06-27 15:13:39

by Shakeel Butt

[permalink] [raw]
Subject: Re: [net] 4890b686f4: netperf.Throughput_Mbps -69.4% regression

On Mon, Jun 27, 2022 at 5:34 AM Feng Tang <[email protected]> wrote:
>
> On Mon, Jun 27, 2022 at 10:46:21AM +0200, Eric Dumazet wrote:
> > On Mon, Jun 27, 2022 at 4:38 AM Feng Tang <[email protected]> wrote:
> [snip]
> > > > >
> > > > > Thanks Feng. Can you check the value of memory.kmem.tcp.max_usage_in_bytes
> > > > > in /sys/fs/cgroup/memory/system.slice/lkp-bootstrap.service after making
> > > > > sure that the netperf test has already run?
> > > >
> > > > memory.kmem.tcp.max_usage_in_bytes:0
> > >
> > > Sorry, I made a mistake that in the original report from Oliver, it
> > > was 'cgroup v2' with a 'debian-11.1' rootfs.
> > >
> > > When you asked about cgroup info, I tried the job on another tbox, and
> > > the original 'job.yaml' didn't work, so I kept the 'netperf' test
> > > parameters and started a new job which somehow run with a 'debian-10.4'
> > > rootfs and acutally run with cgroup v1.
> > >
> > > And as you mentioned cgroup version does make a big difference, that
> > > with v1, the regression is reduced to 1% ~ 5% on different generations
> > > of test platforms. Eric mentioned they also got regression report,
> > > but much smaller one, maybe it's due to the cgroup version?
> >
> > This was using the current net-next tree.
> > Used recipe was something like:
> >
> > Make sure cgroup2 is mounted or mount it by mount -t cgroup2 none $MOUNT_POINT.
> > Enable memory controller by echo +memory > $MOUNT_POINT/cgroup.subtree_control.
> > Create a cgroup by mkdir $MOUNT_POINT/job.
> > Jump into that cgroup by echo $$ > $MOUNT_POINT/job/cgroup.procs.
> >
> > <Launch tests>
> >
> > The regression was smaller than 1%, so considered noise compared to
> > the benefits of the bug fix.
>
> Yes, 1% is just around noise level for a microbenchmark.
>
> I went check the original test data of Oliver's report, the tests was
> run 6 rounds and the performance data is pretty stable (0Day's report
> will show any std deviation bigger than 2%)
>
> The test platform is a 4 sockets 72C/144T machine, and I run the
> same job (nr_tasks = 25% * nr_cpus) on one CascadeLake AP (4 nodes)
> and one Icelake 2 sockets platform, and saw 75% and 53% regresson on
> them.
>
> In the first email, there is a file named 'reproduce', it shows the
> basic test process:
>
> "
> use 'performane' cpufre governor for all CPUs
>
> netserver -4 -D
> modprobe sctp
> netperf -4 -H 127.0.0.1 -t SCTP_STREAM_MANY -c -C -l 300 -- -m 10K &
> netperf -4 -H 127.0.0.1 -t SCTP_STREAM_MANY -c -C -l 300 -- -m 10K &
> netperf -4 -H 127.0.0.1 -t SCTP_STREAM_MANY -c -C -l 300 -- -m 10K &
> (repeat 36 times in total)
> ...
>
> "
>
> Which starts 36 (25% of nr_cpus) netperf clients. And the clients number
> also matters, I tried to increase the client number from 36 to 72(50%),
> and the regression is changed from 69.4% to 73.7%
>

Am I understanding correctly that this 69.4% (or 73.7%) regression is
with cgroup v2?

Eric did the experiments on v2 but on real hardware where the
performance impact was negligible.

BTW do you see similar regression for tcp as well or just sctp?

2022-06-27 15:17:16

by Feng Tang

[permalink] [raw]
Subject: Re: [net] 4890b686f4: netperf.Throughput_Mbps -69.4% regression

On Mon, Jun 27, 2022 at 04:07:55PM +0200, Eric Dumazet wrote:
> On Mon, Jun 27, 2022 at 2:34 PM Feng Tang <[email protected]> wrote:
> >
> > On Mon, Jun 27, 2022 at 10:46:21AM +0200, Eric Dumazet wrote:
> > > On Mon, Jun 27, 2022 at 4:38 AM Feng Tang <[email protected]> wrote:
> > [snip]
> > > > > >
> > > > > > Thanks Feng. Can you check the value of memory.kmem.tcp.max_usage_in_bytes
> > > > > > in /sys/fs/cgroup/memory/system.slice/lkp-bootstrap.service after making
> > > > > > sure that the netperf test has already run?
> > > > >
> > > > > memory.kmem.tcp.max_usage_in_bytes:0
> > > >
> > > > Sorry, I made a mistake that in the original report from Oliver, it
> > > > was 'cgroup v2' with a 'debian-11.1' rootfs.
> > > >
> > > > When you asked about cgroup info, I tried the job on another tbox, and
> > > > the original 'job.yaml' didn't work, so I kept the 'netperf' test
> > > > parameters and started a new job which somehow run with a 'debian-10.4'
> > > > rootfs and acutally run with cgroup v1.
> > > >
> > > > And as you mentioned cgroup version does make a big difference, that
> > > > with v1, the regression is reduced to 1% ~ 5% on different generations
> > > > of test platforms. Eric mentioned they also got regression report,
> > > > but much smaller one, maybe it's due to the cgroup version?
> > >
> > > This was using the current net-next tree.
> > > Used recipe was something like:
> > >
> > > Make sure cgroup2 is mounted or mount it by mount -t cgroup2 none $MOUNT_POINT.
> > > Enable memory controller by echo +memory > $MOUNT_POINT/cgroup.subtree_control.
> > > Create a cgroup by mkdir $MOUNT_POINT/job.
> > > Jump into that cgroup by echo $$ > $MOUNT_POINT/job/cgroup.procs.
> > >
> > > <Launch tests>
> > >
> > > The regression was smaller than 1%, so considered noise compared to
> > > the benefits of the bug fix.
> >
> > Yes, 1% is just around noise level for a microbenchmark.
> >
> > I went check the original test data of Oliver's report, the tests was
> > run 6 rounds and the performance data is pretty stable (0Day's report
> > will show any std deviation bigger than 2%)
> >
> > The test platform is a 4 sockets 72C/144T machine, and I run the
> > same job (nr_tasks = 25% * nr_cpus) on one CascadeLake AP (4 nodes)
> > and one Icelake 2 sockets platform, and saw 75% and 53% regresson on
> > them.
> >
> > In the first email, there is a file named 'reproduce', it shows the
> > basic test process:
> >
> > "
> > use 'performane' cpufre governor for all CPUs
> >
> > netserver -4 -D
> > modprobe sctp
> > netperf -4 -H 127.0.0.1 -t SCTP_STREAM_MANY -c -C -l 300 -- -m 10K &
> > netperf -4 -H 127.0.0.1 -t SCTP_STREAM_MANY -c -C -l 300 -- -m 10K &
> > netperf -4 -H 127.0.0.1 -t SCTP_STREAM_MANY -c -C -l 300 -- -m 10K &
> > (repeat 36 times in total)
> > ...
> >
> > "
> >
> > Which starts 36 (25% of nr_cpus) netperf clients. And the clients number
> > also matters, I tried to increase the client number from 36 to 72(50%),
> > and the regression is changed from 69.4% to 73.7%"
> >
>
> This seems like a lot of opportunities for memcg folks :)
>
> struct page_counter has poor field placement [1], and no per-cpu cache.
>
> [1] "atomic_long_t usage" is sharing cache line with read mostly fields.
>
> (struct mem_cgroup also has poor field placement, mainly because of
> struct page_counter)
>
> 28.69% [kernel] [k] copy_user_enhanced_fast_string
> 16.13% [kernel] [k] intel_idle_irq
> 6.46% [kernel] [k] page_counter_try_charge
> 6.20% [kernel] [k] __sk_mem_reduce_allocated
> 5.68% [kernel] [k] try_charge_memcg
> 5.16% [kernel] [k] page_counter_cancel

Yes, I also analyzed the perf-profile data, and made some layout changes
which could recover the changes from 69% to 40%.

7c80b038d23e1f4c 4890b686f4088c90432149bd6de 332b589c49656a45881bca4ecc0
---------------- --------------------------- ---------------------------
15722 -69.5% 4792 -40.8% 9300 netperf.Throughput_Mbps


diff --git a/include/linux/cgroup-defs.h b/include/linux/cgroup-defs.h
index 1bfcfb1af352..aa37bd39116c 100644
--- a/include/linux/cgroup-defs.h
+++ b/include/linux/cgroup-defs.h
@@ -179,14 +179,13 @@ struct cgroup_subsys_state {
atomic_t online_cnt;

/* percpu_ref killing and RCU release */
- struct work_struct destroy_work;
struct rcu_work destroy_rwork;
-
+ struct cgroup_subsys_state *parent;
+ struct work_struct destroy_work;
/*
* PI: the parent css. Placed here for cache proximity to following
* fields of the containing structure.
*/
- struct cgroup_subsys_state *parent;
};

/*
diff --git a/include/linux/memcontrol.h b/include/linux/memcontrol.h
index 9ecead1042b9..963b88ab9930 100644
--- a/include/linux/memcontrol.h
+++ b/include/linux/memcontrol.h
@@ -239,9 +239,6 @@ struct mem_cgroup {
/* Private memcg ID. Used to ID objects that outlive the cgroup */
struct mem_cgroup_id id;

- /* Accounted resources */
- struct page_counter memory; /* Both v1 & v2 */
-
union {
struct page_counter swap; /* v2 only */
struct page_counter memsw; /* v1 only */
@@ -251,6 +248,9 @@ struct mem_cgroup {
struct page_counter kmem; /* v1 only */
struct page_counter tcpmem; /* v1 only */

+ /* Accounted resources */
+ struct page_counter memory; /* Both v1 & v2 */
+
/* Range enforcement for interrupt charges */
struct work_struct high_work;

@@ -313,7 +313,6 @@ struct mem_cgroup {
atomic_long_t memory_events[MEMCG_NR_MEMORY_EVENTS];
atomic_long_t memory_events_local[MEMCG_NR_MEMORY_EVENTS];

- unsigned long socket_pressure;

/* Legacy tcp memory accounting */
bool tcpmem_active;
@@ -349,6 +348,7 @@ struct mem_cgroup {
#ifdef CONFIG_TRANSPARENT_HUGEPAGE
struct deferred_split deferred_split_queue;
#endif
+ unsigned long socket_pressure;

struct mem_cgroup_per_node *nodeinfo[];
};

And some of these are specific for network and may not be a universal
win, though I think the 'cgroup_subsys_state' could keep the
read-mostly 'parent' away from following written-mostly counters.

Btw, I tried your debug patch which compiled fail with 0Day's kbuild
system, but it did compile ok on my local machine.

Thanks,
Feng

>
> > Thanks,
> > Feng
> >
> > > >
> > > > Thanks,
> > > > Feng

2022-06-27 15:32:25

by Feng Tang

[permalink] [raw]
Subject: Re: [net] 4890b686f4: netperf.Throughput_Mbps -69.4% regression

On Mon, Jun 27, 2022 at 07:52:55AM -0700, Shakeel Butt wrote:
> On Mon, Jun 27, 2022 at 5:34 AM Feng Tang <[email protected]> wrote:
> > Yes, 1% is just around noise level for a microbenchmark.
> >
> > I went check the original test data of Oliver's report, the tests was
> > run 6 rounds and the performance data is pretty stable (0Day's report
> > will show any std deviation bigger than 2%)
> >
> > The test platform is a 4 sockets 72C/144T machine, and I run the
> > same job (nr_tasks = 25% * nr_cpus) on one CascadeLake AP (4 nodes)
> > and one Icelake 2 sockets platform, and saw 75% and 53% regresson on
> > them.
> >
> > In the first email, there is a file named 'reproduce', it shows the
> > basic test process:
> >
> > "
> > use 'performane' cpufre governor for all CPUs
> >
> > netserver -4 -D
> > modprobe sctp
> > netperf -4 -H 127.0.0.1 -t SCTP_STREAM_MANY -c -C -l 300 -- -m 10K &
> > netperf -4 -H 127.0.0.1 -t SCTP_STREAM_MANY -c -C -l 300 -- -m 10K &
> > netperf -4 -H 127.0.0.1 -t SCTP_STREAM_MANY -c -C -l 300 -- -m 10K &
> > (repeat 36 times in total)
> > ...
> >
> > "
> >
> > Which starts 36 (25% of nr_cpus) netperf clients. And the clients number
> > also matters, I tried to increase the client number from 36 to 72(50%),
> > and the regression is changed from 69.4% to 73.7%
> >
>
> Am I understanding correctly that this 69.4% (or 73.7%) regression is
> with cgroup v2?

Yes.

> Eric did the experiments on v2 but on real hardware where the
> performance impact was negligible.
>
> BTW do you see similar regression for tcp as well or just sctp?

Yes, I run TCP_SENDFILE case with 'send_size'==10K, it hits a
70%+ regressioin.

Thanks,
Feng

2022-06-27 17:04:03

by Shakeel Butt

[permalink] [raw]
Subject: Re: [net] 4890b686f4: netperf.Throughput_Mbps -69.4% regression

On Mon, Jun 27, 2022 at 8:25 AM Feng Tang <[email protected]> wrote:
>
> On Mon, Jun 27, 2022 at 07:52:55AM -0700, Shakeel Butt wrote:
> > On Mon, Jun 27, 2022 at 5:34 AM Feng Tang <[email protected]> wrote:
> > > Yes, 1% is just around noise level for a microbenchmark.
> > >
> > > I went check the original test data of Oliver's report, the tests was
> > > run 6 rounds and the performance data is pretty stable (0Day's report
> > > will show any std deviation bigger than 2%)
> > >
> > > The test platform is a 4 sockets 72C/144T machine, and I run the
> > > same job (nr_tasks = 25% * nr_cpus) on one CascadeLake AP (4 nodes)
> > > and one Icelake 2 sockets platform, and saw 75% and 53% regresson on
> > > them.
> > >
> > > In the first email, there is a file named 'reproduce', it shows the
> > > basic test process:
> > >
> > > "
> > > use 'performane' cpufre governor for all CPUs
> > >
> > > netserver -4 -D
> > > modprobe sctp
> > > netperf -4 -H 127.0.0.1 -t SCTP_STREAM_MANY -c -C -l 300 -- -m 10K &
> > > netperf -4 -H 127.0.0.1 -t SCTP_STREAM_MANY -c -C -l 300 -- -m 10K &
> > > netperf -4 -H 127.0.0.1 -t SCTP_STREAM_MANY -c -C -l 300 -- -m 10K &
> > > (repeat 36 times in total)
> > > ...
> > >
> > > "
> > >
> > > Which starts 36 (25% of nr_cpus) netperf clients. And the clients number
> > > also matters, I tried to increase the client number from 36 to 72(50%),
> > > and the regression is changed from 69.4% to 73.7%
> > >
> >
> > Am I understanding correctly that this 69.4% (or 73.7%) regression is
> > with cgroup v2?
>
> Yes.
>
> > Eric did the experiments on v2 but on real hardware where the
> > performance impact was negligible.
> >
> > BTW do you see similar regression for tcp as well or just sctp?
>
> Yes, I run TCP_SENDFILE case with 'send_size'==10K, it hits a
> 70%+ regressioin.
>

Thanks Feng. I think we should start with squeezing whatever we can
from layout changes and then try other approaches like increasing
batch size or something else. I can take a stab at this next week.

thanks,
Shakeel

2022-06-27 17:09:15

by Eric Dumazet

[permalink] [raw]
Subject: Re: [net] 4890b686f4: netperf.Throughput_Mbps -69.4% regression

On Mon, Jun 27, 2022 at 4:48 PM Feng Tang <[email protected]> wrote:
>
> On Mon, Jun 27, 2022 at 04:07:55PM +0200, Eric Dumazet wrote:
> > On Mon, Jun 27, 2022 at 2:34 PM Feng Tang <[email protected]> wrote:
> > >
> > > On Mon, Jun 27, 2022 at 10:46:21AM +0200, Eric Dumazet wrote:
> > > > On Mon, Jun 27, 2022 at 4:38 AM Feng Tang <[email protected]> wrote:
> > > [snip]
> > > > > > >
> > > > > > > Thanks Feng. Can you check the value of memory.kmem.tcp.max_usage_in_bytes
> > > > > > > in /sys/fs/cgroup/memory/system.slice/lkp-bootstrap.service after making
> > > > > > > sure that the netperf test has already run?
> > > > > >
> > > > > > memory.kmem.tcp.max_usage_in_bytes:0
> > > > >
> > > > > Sorry, I made a mistake that in the original report from Oliver, it
> > > > > was 'cgroup v2' with a 'debian-11.1' rootfs.
> > > > >
> > > > > When you asked about cgroup info, I tried the job on another tbox, and
> > > > > the original 'job.yaml' didn't work, so I kept the 'netperf' test
> > > > > parameters and started a new job which somehow run with a 'debian-10.4'
> > > > > rootfs and acutally run with cgroup v1.
> > > > >
> > > > > And as you mentioned cgroup version does make a big difference, that
> > > > > with v1, the regression is reduced to 1% ~ 5% on different generations
> > > > > of test platforms. Eric mentioned they also got regression report,
> > > > > but much smaller one, maybe it's due to the cgroup version?
> > > >
> > > > This was using the current net-next tree.
> > > > Used recipe was something like:
> > > >
> > > > Make sure cgroup2 is mounted or mount it by mount -t cgroup2 none $MOUNT_POINT.
> > > > Enable memory controller by echo +memory > $MOUNT_POINT/cgroup.subtree_control.
> > > > Create a cgroup by mkdir $MOUNT_POINT/job.
> > > > Jump into that cgroup by echo $$ > $MOUNT_POINT/job/cgroup.procs.
> > > >
> > > > <Launch tests>
> > > >
> > > > The regression was smaller than 1%, so considered noise compared to
> > > > the benefits of the bug fix.
> > >
> > > Yes, 1% is just around noise level for a microbenchmark.
> > >
> > > I went check the original test data of Oliver's report, the tests was
> > > run 6 rounds and the performance data is pretty stable (0Day's report
> > > will show any std deviation bigger than 2%)
> > >
> > > The test platform is a 4 sockets 72C/144T machine, and I run the
> > > same job (nr_tasks = 25% * nr_cpus) on one CascadeLake AP (4 nodes)
> > > and one Icelake 2 sockets platform, and saw 75% and 53% regresson on
> > > them.
> > >
> > > In the first email, there is a file named 'reproduce', it shows the
> > > basic test process:
> > >
> > > "
> > > use 'performane' cpufre governor for all CPUs
> > >
> > > netserver -4 -D
> > > modprobe sctp
> > > netperf -4 -H 127.0.0.1 -t SCTP_STREAM_MANY -c -C -l 300 -- -m 10K &
> > > netperf -4 -H 127.0.0.1 -t SCTP_STREAM_MANY -c -C -l 300 -- -m 10K &
> > > netperf -4 -H 127.0.0.1 -t SCTP_STREAM_MANY -c -C -l 300 -- -m 10K &
> > > (repeat 36 times in total)
> > > ...
> > >
> > > "
> > >
> > > Which starts 36 (25% of nr_cpus) netperf clients. And the clients number
> > > also matters, I tried to increase the client number from 36 to 72(50%),
> > > and the regression is changed from 69.4% to 73.7%"
> > >
> >
> > This seems like a lot of opportunities for memcg folks :)
> >
> > struct page_counter has poor field placement [1], and no per-cpu cache.
> >
> > [1] "atomic_long_t usage" is sharing cache line with read mostly fields.
> >
> > (struct mem_cgroup also has poor field placement, mainly because of
> > struct page_counter)
> >
> > 28.69% [kernel] [k] copy_user_enhanced_fast_string
> > 16.13% [kernel] [k] intel_idle_irq
> > 6.46% [kernel] [k] page_counter_try_charge
> > 6.20% [kernel] [k] __sk_mem_reduce_allocated
> > 5.68% [kernel] [k] try_charge_memcg
> > 5.16% [kernel] [k] page_counter_cancel
>
> Yes, I also analyzed the perf-profile data, and made some layout changes
> which could recover the changes from 69% to 40%.
>
> 7c80b038d23e1f4c 4890b686f4088c90432149bd6de 332b589c49656a45881bca4ecc0
> ---------------- --------------------------- ---------------------------
> 15722 -69.5% 4792 -40.8% 9300 netperf.Throughput_Mbps
>
>
> diff --git a/include/linux/cgroup-defs.h b/include/linux/cgroup-defs.h
> index 1bfcfb1af352..aa37bd39116c 100644
> --- a/include/linux/cgroup-defs.h
> +++ b/include/linux/cgroup-defs.h
> @@ -179,14 +179,13 @@ struct cgroup_subsys_state {
> atomic_t online_cnt;
>
> /* percpu_ref killing and RCU release */
> - struct work_struct destroy_work;
> struct rcu_work destroy_rwork;
> -
> + struct cgroup_subsys_state *parent;
> + struct work_struct destroy_work;
> /*
> * PI: the parent css. Placed here for cache proximity to following
> * fields of the containing structure.
> */
> - struct cgroup_subsys_state *parent;
> };
>
> /*
> diff --git a/include/linux/memcontrol.h b/include/linux/memcontrol.h
> index 9ecead1042b9..963b88ab9930 100644
> --- a/include/linux/memcontrol.h
> +++ b/include/linux/memcontrol.h
> @@ -239,9 +239,6 @@ struct mem_cgroup {
> /* Private memcg ID. Used to ID objects that outlive the cgroup */
> struct mem_cgroup_id id;
>
> - /* Accounted resources */
> - struct page_counter memory; /* Both v1 & v2 */
> -
> union {
> struct page_counter swap; /* v2 only */
> struct page_counter memsw; /* v1 only */
> @@ -251,6 +248,9 @@ struct mem_cgroup {
> struct page_counter kmem; /* v1 only */
> struct page_counter tcpmem; /* v1 only */
>
> + /* Accounted resources */
> + struct page_counter memory; /* Both v1 & v2 */
> +
> /* Range enforcement for interrupt charges */
> struct work_struct high_work;
>
> @@ -313,7 +313,6 @@ struct mem_cgroup {
> atomic_long_t memory_events[MEMCG_NR_MEMORY_EVENTS];
> atomic_long_t memory_events_local[MEMCG_NR_MEMORY_EVENTS];
>
> - unsigned long socket_pressure;
>
> /* Legacy tcp memory accounting */
> bool tcpmem_active;
> @@ -349,6 +348,7 @@ struct mem_cgroup {
> #ifdef CONFIG_TRANSPARENT_HUGEPAGE
> struct deferred_split deferred_split_queue;
> #endif
> + unsigned long socket_pressure;
>
> struct mem_cgroup_per_node *nodeinfo[];
> };
>

I simply did the following and got much better results.

But I am not sure if updates to ->usage are really needed that often...


diff --git a/include/linux/page_counter.h b/include/linux/page_counter.h
index 679591301994d316062f92b275efa2459a8349c9..e267be4ba849760117d9fd041e22c2a44658ab36
100644
--- a/include/linux/page_counter.h
+++ b/include/linux/page_counter.h
@@ -3,12 +3,15 @@
#define _LINUX_PAGE_COUNTER_H

#include <linux/atomic.h>
+#include <linux/cache.h>
#include <linux/kernel.h>
#include <asm/page.h>

struct page_counter {
- atomic_long_t usage;
- unsigned long min;
+ /* contended cache line. */
+ atomic_long_t usage ____cacheline_aligned_in_smp;
+
+ unsigned long min ____cacheline_aligned_in_smp;
unsigned long low;
unsigned long high;
unsigned long max;
@@ -27,12 +30,6 @@ struct page_counter {
unsigned long watermark;
unsigned long failcnt;

- /*
- * 'parent' is placed here to be far from 'usage' to reduce
- * cache false sharing, as 'usage' is written mostly while
- * parent is frequently read for cgroup's hierarchical
- * counting nature.
- */
struct page_counter *parent;
};



> And some of these are specific for network and may not be a universal
> win, though I think the 'cgroup_subsys_state' could keep the
> read-mostly 'parent' away from following written-mostly counters.
>
> Btw, I tried your debug patch which compiled fail with 0Day's kbuild
> system, but it did compile ok on my local machine.
>
> Thanks,
> Feng
>
> >
> > > Thanks,
> > > Feng
> > >
> > > > >
> > > > > Thanks,
> > > > > Feng

2022-06-27 17:26:35

by Shakeel Butt

[permalink] [raw]
Subject: Re: [net] 4890b686f4: netperf.Throughput_Mbps -69.4% regression

On Mon, Jun 27, 2022 at 9:26 AM Eric Dumazet <[email protected]> wrote:
>
[...]
> >
>
> I simply did the following and got much better results.
>
> But I am not sure if updates to ->usage are really needed that often...

I suspect we need to improve the per-cpu memcg stock usage here. Were
the updates mostly from uncharge path or charge path or that's
irrelevant?

I think doing full drain (i.e. drain_stock()) within __refill_stock()
when the local cache is larger than MEMCG_CHARGE_BATCH is not best.
Rather we should always keep at least MEMCG_CHARGE_BATCH for such
scenarios.

>
>
> diff --git a/include/linux/page_counter.h b/include/linux/page_counter.h
> index 679591301994d316062f92b275efa2459a8349c9..e267be4ba849760117d9fd041e22c2a44658ab36
> 100644
> --- a/include/linux/page_counter.h
> +++ b/include/linux/page_counter.h
> @@ -3,12 +3,15 @@
> #define _LINUX_PAGE_COUNTER_H
>
> #include <linux/atomic.h>
> +#include <linux/cache.h>
> #include <linux/kernel.h>
> #include <asm/page.h>
>
> struct page_counter {
> - atomic_long_t usage;
> - unsigned long min;
> + /* contended cache line. */
> + atomic_long_t usage ____cacheline_aligned_in_smp;
> +
> + unsigned long min ____cacheline_aligned_in_smp;

Do we need to align 'min' too?

> unsigned long low;
> unsigned long high;
> unsigned long max;
> @@ -27,12 +30,6 @@ struct page_counter {
> unsigned long watermark;
> unsigned long failcnt;
>
> - /*
> - * 'parent' is placed here to be far from 'usage' to reduce
> - * cache false sharing, as 'usage' is written mostly while
> - * parent is frequently read for cgroup's hierarchical
> - * counting nature.
> - */
> struct page_counter *parent;
> };
>
>
>

2022-06-28 01:58:32

by Roman Gushchin

[permalink] [raw]
Subject: Re: [net] 4890b686f4: netperf.Throughput_Mbps -69.4% regression

On Mon, Jun 27, 2022 at 09:48:01AM -0700, Shakeel Butt wrote:
> On Mon, Jun 27, 2022 at 9:26 AM Eric Dumazet <[email protected]> wrote:
> >
> [...]
> > >
> >
> > I simply did the following and got much better results.
> >
> > But I am not sure if updates to ->usage are really needed that often...
>
> I suspect we need to improve the per-cpu memcg stock usage here. Were
> the updates mostly from uncharge path or charge path or that's
> irrelevant?
>
> I think doing full drain (i.e. drain_stock()) within __refill_stock()
> when the local cache is larger than MEMCG_CHARGE_BATCH is not best.
> Rather we should always keep at least MEMCG_CHARGE_BATCH for such
> scenarios.

+1, really good point.

2022-06-28 04:41:56

by Feng Tang

[permalink] [raw]
Subject: Re: [net] 4890b686f4: netperf.Throughput_Mbps -69.4% regression

On Mon, Jun 27, 2022 at 06:25:59PM +0200, Eric Dumazet wrote:
> On Mon, Jun 27, 2022 at 4:48 PM Feng Tang <[email protected]> wrote:
> >
> > Yes, I also analyzed the perf-profile data, and made some layout changes
> > which could recover the changes from 69% to 40%.
> >
> > 7c80b038d23e1f4c 4890b686f4088c90432149bd6de 332b589c49656a45881bca4ecc0
> > ---------------- --------------------------- ---------------------------
> > 15722 -69.5% 4792 -40.8% 9300 netperf.Throughput_Mbps
> >
>
> I simply did the following and got much better results.
>
> But I am not sure if updates to ->usage are really needed that often...
>
>
> diff --git a/include/linux/page_counter.h b/include/linux/page_counter.h
> index 679591301994d316062f92b275efa2459a8349c9..e267be4ba849760117d9fd041e22c2a44658ab36
> 100644
> --- a/include/linux/page_counter.h
> +++ b/include/linux/page_counter.h
> @@ -3,12 +3,15 @@
> #define _LINUX_PAGE_COUNTER_H
>
> #include <linux/atomic.h>
> +#include <linux/cache.h>
> #include <linux/kernel.h>
> #include <asm/page.h>
>
> struct page_counter {
> - atomic_long_t usage;
> - unsigned long min;
> + /* contended cache line. */
> + atomic_long_t usage ____cacheline_aligned_in_smp;
> +
> + unsigned long min ____cacheline_aligned_in_smp;
> unsigned long low;
> unsigned long high;
> unsigned long max;
> @@ -27,12 +30,6 @@ struct page_counter {
> unsigned long watermark;
> unsigned long failcnt;
>
> - /*
> - * 'parent' is placed here to be far from 'usage' to reduce
> - * cache false sharing, as 'usage' is written mostly while
> - * parent is frequently read for cgroup's hierarchical
> - * counting nature.
> - */
> struct page_counter *parent;
> };

I just tested it, it does perform better (the 4th is with your patch),
some perf-profile data is also listed.

7c80b038d23e1f4c 4890b686f4088c90432149bd6de 332b589c49656a45881bca4ecc0 e719635902654380b23ffce908d
---------------- --------------------------- --------------------------- ---------------------------
15722 -69.5% 4792 -40.8% 9300 -27.9% 11341 netperf.Throughput_Mbps

0.00 +0.3 0.26 ± 5% +0.5 0.51 +1.3 1.27 ± 2%pp.self.__sk_mem_raise_allocated
0.00 +0.3 0.32 ± 15% +1.7 1.74 ± 2% +0.4 0.40 ± 2% pp.self.propagate_protected_usage
0.00 +0.8 0.82 ± 7% +0.9 0.90 +0.8 0.84 pp.self.__mod_memcg_state
0.00 +1.2 1.24 ± 4% +1.0 1.01 +1.4 1.44 pp.self.try_charge_memcg
0.00 +2.1 2.06 +2.1 2.13 +2.1 2.11 pp.self.page_counter_uncharge
0.00 +2.1 2.14 ± 4% +2.7 2.71 +2.6 2.60 ± 2% pp.self.page_counter_try_charge
1.12 ± 4% +3.1 4.24 +1.1 2.22 +1.4 2.51 pp.self.native_queued_spin_lock_slowpath
0.28 ± 9% +3.8 4.06 ± 4% +0.2 0.48 +0.4 0.68 pp.self.sctp_eat_data
0.00 +8.2 8.23 +0.8 0.83 +1.3 1.26 pp.self.__sk_mem_reduce_allocated

And the size of 'mem_cgroup' is increased from 4224 Bytes to 4608.

Another info is the perf hotspos are slightly different between
tcp and sctp test cases.

Thanks,
Feng

2022-07-01 15:57:36

by Shakeel Butt

[permalink] [raw]
Subject: Re: [net] 4890b686f4: netperf.Throughput_Mbps -69.4% regression

On Mon, Jun 27, 2022 at 8:49 PM Feng Tang <[email protected]> wrote:
>
> On Mon, Jun 27, 2022 at 06:25:59PM +0200, Eric Dumazet wrote:
> > On Mon, Jun 27, 2022 at 4:48 PM Feng Tang <[email protected]> wrote:
> > >
> > > Yes, I also analyzed the perf-profile data, and made some layout changes
> > > which could recover the changes from 69% to 40%.
> > >
> > > 7c80b038d23e1f4c 4890b686f4088c90432149bd6de 332b589c49656a45881bca4ecc0
> > > ---------------- --------------------------- ---------------------------
> > > 15722 -69.5% 4792 -40.8% 9300 netperf.Throughput_Mbps
> > >
> >
> > I simply did the following and got much better results.
> >
> > But I am not sure if updates to ->usage are really needed that often...
> >
> >
> > diff --git a/include/linux/page_counter.h b/include/linux/page_counter.h
> > index 679591301994d316062f92b275efa2459a8349c9..e267be4ba849760117d9fd041e22c2a44658ab36
> > 100644
> > --- a/include/linux/page_counter.h
> > +++ b/include/linux/page_counter.h
> > @@ -3,12 +3,15 @@
> > #define _LINUX_PAGE_COUNTER_H
> >
> > #include <linux/atomic.h>
> > +#include <linux/cache.h>
> > #include <linux/kernel.h>
> > #include <asm/page.h>
> >
> > struct page_counter {
> > - atomic_long_t usage;
> > - unsigned long min;
> > + /* contended cache line. */
> > + atomic_long_t usage ____cacheline_aligned_in_smp;
> > +
> > + unsigned long min ____cacheline_aligned_in_smp;
> > unsigned long low;
> > unsigned long high;
> > unsigned long max;
> > @@ -27,12 +30,6 @@ struct page_counter {
> > unsigned long watermark;
> > unsigned long failcnt;
> >
> > - /*
> > - * 'parent' is placed here to be far from 'usage' to reduce
> > - * cache false sharing, as 'usage' is written mostly while
> > - * parent is frequently read for cgroup's hierarchical
> > - * counting nature.
> > - */
> > struct page_counter *parent;
> > };
>
> I just tested it, it does perform better (the 4th is with your patch),
> some perf-profile data is also listed.
>
> 7c80b038d23e1f4c 4890b686f4088c90432149bd6de 332b589c49656a45881bca4ecc0 e719635902654380b23ffce908d
> ---------------- --------------------------- --------------------------- ---------------------------
> 15722 -69.5% 4792 -40.8% 9300 -27.9% 11341 netperf.Throughput_Mbps
>
> 0.00 +0.3 0.26 ± 5% +0.5 0.51 +1.3 1.27 ± 2%pp.self.__sk_mem_raise_allocated
> 0.00 +0.3 0.32 ± 15% +1.7 1.74 ± 2% +0.4 0.40 ± 2% pp.self.propagate_protected_usage
> 0.00 +0.8 0.82 ± 7% +0.9 0.90 +0.8 0.84 pp.self.__mod_memcg_state
> 0.00 +1.2 1.24 ± 4% +1.0 1.01 +1.4 1.44 pp.self.try_charge_memcg
> 0.00 +2.1 2.06 +2.1 2.13 +2.1 2.11 pp.self.page_counter_uncharge
> 0.00 +2.1 2.14 ± 4% +2.7 2.71 +2.6 2.60 ± 2% pp.self.page_counter_try_charge
> 1.12 ± 4% +3.1 4.24 +1.1 2.22 +1.4 2.51 pp.self.native_queued_spin_lock_slowpath
> 0.28 ± 9% +3.8 4.06 ± 4% +0.2 0.48 +0.4 0.68 pp.self.sctp_eat_data
> 0.00 +8.2 8.23 +0.8 0.83 +1.3 1.26 pp.self.__sk_mem_reduce_allocated
>
> And the size of 'mem_cgroup' is increased from 4224 Bytes to 4608.

Hi Feng, can you please try two more configurations? Take Eric's patch
of adding ____cacheline_aligned_in_smp in page_counter and for first
increase MEMCG_CHARGE_BATCH to 64 and for second increase it to 128.
Basically batch increases combined with Eric's patch.

2022-07-03 11:07:24

by Feng Tang

[permalink] [raw]
Subject: Re: [net] 4890b686f4: netperf.Throughput_Mbps -69.4% regression

Hi Shakeel,

On Fri, Jul 01, 2022 at 08:47:29AM -0700, Shakeel Butt wrote:
> On Mon, Jun 27, 2022 at 8:49 PM Feng Tang <[email protected]> wrote:
> > I just tested it, it does perform better (the 4th is with your patch),
> > some perf-profile data is also listed.
> >
> > 7c80b038d23e1f4c 4890b686f4088c90432149bd6de 332b589c49656a45881bca4ecc0 e719635902654380b23ffce908d
> > ---------------- --------------------------- --------------------------- ---------------------------
> > 15722 -69.5% 4792 -40.8% 9300 -27.9% 11341 netperf.Throughput_Mbps
> >
> > 0.00 +0.3 0.26 ± 5% +0.5 0.51 +1.3 1.27 ± 2%pp.self.__sk_mem_raise_allocated
> > 0.00 +0.3 0.32 ± 15% +1.7 1.74 ± 2% +0.4 0.40 ± 2% pp.self.propagate_protected_usage
> > 0.00 +0.8 0.82 ± 7% +0.9 0.90 +0.8 0.84 pp.self.__mod_memcg_state
> > 0.00 +1.2 1.24 ± 4% +1.0 1.01 +1.4 1.44 pp.self.try_charge_memcg
> > 0.00 +2.1 2.06 +2.1 2.13 +2.1 2.11 pp.self.page_counter_uncharge
> > 0.00 +2.1 2.14 ± 4% +2.7 2.71 +2.6 2.60 ± 2% pp.self.page_counter_try_charge
> > 1.12 ± 4% +3.1 4.24 +1.1 2.22 +1.4 2.51 pp.self.native_queued_spin_lock_slowpath
> > 0.28 ± 9% +3.8 4.06 ± 4% +0.2 0.48 +0.4 0.68 pp.self.sctp_eat_data
> > 0.00 +8.2 8.23 +0.8 0.83 +1.3 1.26 pp.self.__sk_mem_reduce_allocated
> >
> > And the size of 'mem_cgroup' is increased from 4224 Bytes to 4608.
>
> Hi Feng, can you please try two more configurations? Take Eric's patch
> of adding ____cacheline_aligned_in_smp in page_counter and for first
> increase MEMCG_CHARGE_BATCH to 64 and for second increase it to 128.
> Basically batch increases combined with Eric's patch.

With increasing batch to 128, the regression could be reduced to -12.4%.

Some more details with perf-profile data below:

7c80b038d23e1f4c 4890b686f4088c90432149bd6de Eric's patch Eric's patch + batch-64 Eric's patch + batch-128
---------------- --------------------------- --------------------------- --------------------------- ---------------------------
15722 -69.5% 4792 -27.9% 11341 -14.0% 13521 -12.4% 13772 netperf.Throughput_Mbps

0.05 +0.2 0.27 ± 18% +0.0 0.08 ± 6% -0.1 0.00 -0.0 0.03 ±100% pp.self.timekeeping_max_deferment
0.00 +0.3 0.26 ± 5% +1.3 1.27 ± 2% +1.8 1.82 ± 10% +2.0 1.96 ± 9% pp.self.__sk_mem_raise_allocated
0.00 +0.3 0.32 ± 15% +0.4 0.40 ± 2% +0.1 0.10 ± 5% +0.0 0.00 pp.self.propagate_protected_usage
0.00 +0.8 0.82 ± 7% +0.8 0.84 +0.5 0.48 +0.4 0.36 ± 2% pp.self.__mod_memcg_state
0.00 +1.2 1.24 ± 4% +1.4 1.44 +0.4 0.40 ± 3% +0.2 0.24 ± 6% pp.self.try_charge_memcg
0.00 +2.1 2.06 +2.1 2.11 +0.5 0.50 +0.2 0.18 ± 8% pp.self.page_counter_uncharge
0.00 +2.1 2.14 ± 4% +2.6 2.60 ± 2% +0.6 0.58 +0.2 0.20 pp.self.page_counter_try_charge
1.12 ± 4% +3.1 4.24 +1.4 2.51 +1.0 2.10 ± 2% +1.0 2.10 ± 9% pp.self.native_queued_spin_lock_slowpath
0.28 ± 9% +3.8 4.06 ± 4% +0.4 0.68 +0.6 0.90 ± 9% +0.7 1.00 ± 11% pp.self.sctp_eat_data
0.00 +8.2 8.23 +1.3 1.26 +1.7 1.72 ± 6% +2.0 1.95 ± 10% pp.self.__sk_mem_reduce_allocated

Thanks,
Feng

2022-07-03 23:09:38

by Roman Gushchin

[permalink] [raw]
Subject: Re: [net] 4890b686f4: netperf.Throughput_Mbps -69.4% regression

On Sun, Jul 03, 2022 at 06:43:53PM +0800, Feng Tang wrote:
> Hi Shakeel,
>
> On Fri, Jul 01, 2022 at 08:47:29AM -0700, Shakeel Butt wrote:
> > On Mon, Jun 27, 2022 at 8:49 PM Feng Tang <[email protected]> wrote:
> > > I just tested it, it does perform better (the 4th is with your patch),
> > > some perf-profile data is also listed.
> > >
> > > 7c80b038d23e1f4c 4890b686f4088c90432149bd6de 332b589c49656a45881bca4ecc0 e719635902654380b23ffce908d
> > > ---------------- --------------------------- --------------------------- ---------------------------
> > > 15722 -69.5% 4792 -40.8% 9300 -27.9% 11341 netperf.Throughput_Mbps
> > >
> > > 0.00 +0.3 0.26 ? 5% +0.5 0.51 +1.3 1.27 ? 2%pp.self.__sk_mem_raise_allocated
> > > 0.00 +0.3 0.32 ? 15% +1.7 1.74 ? 2% +0.4 0.40 ? 2% pp.self.propagate_protected_usage
> > > 0.00 +0.8 0.82 ? 7% +0.9 0.90 +0.8 0.84 pp.self.__mod_memcg_state
> > > 0.00 +1.2 1.24 ? 4% +1.0 1.01 +1.4 1.44 pp.self.try_charge_memcg
> > > 0.00 +2.1 2.06 +2.1 2.13 +2.1 2.11 pp.self.page_counter_uncharge
> > > 0.00 +2.1 2.14 ? 4% +2.7 2.71 +2.6 2.60 ? 2% pp.self.page_counter_try_charge
> > > 1.12 ? 4% +3.1 4.24 +1.1 2.22 +1.4 2.51 pp.self.native_queued_spin_lock_slowpath
> > > 0.28 ? 9% +3.8 4.06 ? 4% +0.2 0.48 +0.4 0.68 pp.self.sctp_eat_data
> > > 0.00 +8.2 8.23 +0.8 0.83 +1.3 1.26 pp.self.__sk_mem_reduce_allocated
> > >
> > > And the size of 'mem_cgroup' is increased from 4224 Bytes to 4608.
> >
> > Hi Feng, can you please try two more configurations? Take Eric's patch
> > of adding ____cacheline_aligned_in_smp in page_counter and for first
> > increase MEMCG_CHARGE_BATCH to 64 and for second increase it to 128.
> > Basically batch increases combined with Eric's patch.
>
> With increasing batch to 128, the regression could be reduced to -12.4%.

If we're going to bump it, I wonder if we should scale it dynamically depending
on the size of the memory cgroup?

Thanks!

2022-07-05 05:38:03

by Feng Tang

[permalink] [raw]
Subject: Re: [net] 4890b686f4: netperf.Throughput_Mbps -69.4% regression

On Sun, Jul 03, 2022 at 03:55:31PM -0700, Roman Gushchin wrote:
> On Sun, Jul 03, 2022 at 06:43:53PM +0800, Feng Tang wrote:
> > Hi Shakeel,
> >
> > On Fri, Jul 01, 2022 at 08:47:29AM -0700, Shakeel Butt wrote:
> > > On Mon, Jun 27, 2022 at 8:49 PM Feng Tang <[email protected]> wrote:
> > > > I just tested it, it does perform better (the 4th is with your patch),
> > > > some perf-profile data is also listed.
> > > >
> > > > 7c80b038d23e1f4c 4890b686f4088c90432149bd6de 332b589c49656a45881bca4ecc0 e719635902654380b23ffce908d
> > > > ---------------- --------------------------- --------------------------- ---------------------------
> > > > 15722 -69.5% 4792 -40.8% 9300 -27.9% 11341 netperf.Throughput_Mbps
> > > >
> > > > 0.00 +0.3 0.26 ± 5% +0.5 0.51 +1.3 1.27 ± 2%pp.self.__sk_mem_raise_allocated
> > > > 0.00 +0.3 0.32 ± 15% +1.7 1.74 ± 2% +0.4 0.40 ± 2% pp.self.propagate_protected_usage
> > > > 0.00 +0.8 0.82 ± 7% +0.9 0.90 +0.8 0.84 pp.self.__mod_memcg_state
> > > > 0.00 +1.2 1.24 ± 4% +1.0 1.01 +1.4 1.44 pp.self.try_charge_memcg
> > > > 0.00 +2.1 2.06 +2.1 2.13 +2.1 2.11 pp.self.page_counter_uncharge
> > > > 0.00 +2.1 2.14 ± 4% +2.7 2.71 +2.6 2.60 ± 2% pp.self.page_counter_try_charge
> > > > 1.12 ± 4% +3.1 4.24 +1.1 2.22 +1.4 2.51 pp.self.native_queued_spin_lock_slowpath
> > > > 0.28 ± 9% +3.8 4.06 ± 4% +0.2 0.48 +0.4 0.68 pp.self.sctp_eat_data
> > > > 0.00 +8.2 8.23 +0.8 0.83 +1.3 1.26 pp.self.__sk_mem_reduce_allocated
> > > >
> > > > And the size of 'mem_cgroup' is increased from 4224 Bytes to 4608.
> > >
> > > Hi Feng, can you please try two more configurations? Take Eric's patch
> > > of adding ____cacheline_aligned_in_smp in page_counter and for first
> > > increase MEMCG_CHARGE_BATCH to 64 and for second increase it to 128.
> > > Basically batch increases combined with Eric's patch.
> >
> > With increasing batch to 128, the regression could be reduced to -12.4%.
>
> If we're going to bump it, I wonder if we should scale it dynamically depending
> on the size of the memory cgroup?

I think it makes sense, or also make it a configurable parameter? From
the test reports of 0Day, these charging/counting play critical role
in performance (easy to see up to 60% performance effect). If user only
wants memcg for isolating things or doesn't care charging/stats, these
seem to be extra taxes.

For bumping to 64 or 128, universal improvement is expected with the
only concern of accuracy.

Thanks,
Feng

> Thanks!

2022-08-16 08:49:17

by kernel test robot

[permalink] [raw]
Subject: Re: [net] 4890b686f4: netperf.Throughput_Mbps -69.4% regression

Hi all,

now we noticed this commit has already merged into mainline, and in our tests
there is still similar regression. [1]

not sure if there is a plan to merge some of the solutions that have been
discussed in this thread? we'll very glad to test patches if there is a request

Thanks a lot!

[1]
=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/ip/runtime/nr_threads/cluster/send_size/test/cpufreq_governor/ucode:
lkp-icl-2sp4/netperf/debian-11.1-x86_64-20220510.cgz/x86_64-rhel-8.3/gcc-11/ipv4/300s/50%/cs-localhost/10K/SCTP_STREAM_MANY/performance/0xd000363

7c80b038d23e1f4c 4890b686f4088c90432149bd6de
---------------- ---------------------------
%stddev %change %stddev
\ | \
9078 -55.9% 4006 netperf.Throughput_Mbps
581006 -55.9% 256385 netperf.Throughput_total_Mbps
36715 -54.6% 16674 ? 4% netperf.time.involuntary_context_switches
1885 -50.2% 938.33 ? 3% netperf.time.percent_of_cpu_this_job_got
5533 -49.9% 2771 ? 2% netperf.time.system_time
152.13 -59.5% 61.61 ? 2% netperf.time.user_time
418171 ? 5% +89.4% 791954 ? 17% netperf.time.voluntary_context_switches
2.128e+09 -55.9% 9.389e+08 netperf.workload
30217 +17.8% 35608 uptime.idle
2.689e+10 +20.3% 3.234e+10 cpuidle..time
6.366e+08 -48.1% 3.305e+08 cpuidle..usage
70.26 +13.5 83.78 mpstat.cpu.all.idle%
4.46 -1.5 2.92 ? 3% mpstat.cpu.all.soft%
23.71 -11.6 12.16 ? 3% mpstat.cpu.all.sys%
0.89 -0.5 0.38 mpstat.cpu.all.usr%
1.392e+09 -57.5% 5.91e+08 ? 12% numa-numastat.node0.local_node
1.389e+09 -57.5% 5.906e+08 ? 12% numa-numastat.node0.numa_hit
1.369e+09 -54.5% 6.226e+08 ? 12% numa-numastat.node1.local_node
1.366e+09 -54.4% 6.222e+08 ? 12% numa-numastat.node1.numa_hit
36715 -54.6% 16674 ? 4% time.involuntary_context_switches
1885 -50.2% 938.33 ? 3% time.percent_of_cpu_this_job_got
5533 -49.9% 2771 ? 2% time.system_time
152.13 -59.5% 61.61 ? 2% time.user_time
418171 ? 5% +89.4% 791954 ? 17% time.voluntary_context_switches


On Tue, Jul 05, 2022 at 01:03:26PM +0800, Feng Tang wrote:
> On Sun, Jul 03, 2022 at 03:55:31PM -0700, Roman Gushchin wrote:
> > On Sun, Jul 03, 2022 at 06:43:53PM +0800, Feng Tang wrote:
> > > Hi Shakeel,
> > >
> > > On Fri, Jul 01, 2022 at 08:47:29AM -0700, Shakeel Butt wrote:
> > > > On Mon, Jun 27, 2022 at 8:49 PM Feng Tang <[email protected]> wrote:
> > > > > I just tested it, it does perform better (the 4th is with your patch),
> > > > > some perf-profile data is also listed.
> > > > >
> > > > > 7c80b038d23e1f4c 4890b686f4088c90432149bd6de 332b589c49656a45881bca4ecc0 e719635902654380b23ffce908d
> > > > > ---------------- --------------------------- --------------------------- ---------------------------
> > > > > 15722 -69.5% 4792 -40.8% 9300 -27.9% 11341 netperf.Throughput_Mbps
> > > > >
> > > > > 0.00 +0.3 0.26 ? 5% +0.5 0.51 +1.3 1.27 ? 2%pp.self.__sk_mem_raise_allocated
> > > > > 0.00 +0.3 0.32 ? 15% +1.7 1.74 ? 2% +0.4 0.40 ? 2% pp.self.propagate_protected_usage
> > > > > 0.00 +0.8 0.82 ? 7% +0.9 0.90 +0.8 0.84 pp.self.__mod_memcg_state
> > > > > 0.00 +1.2 1.24 ? 4% +1.0 1.01 +1.4 1.44 pp.self.try_charge_memcg
> > > > > 0.00 +2.1 2.06 +2.1 2.13 +2.1 2.11 pp.self.page_counter_uncharge
> > > > > 0.00 +2.1 2.14 ? 4% +2.7 2.71 +2.6 2.60 ? 2% pp.self.page_counter_try_charge
> > > > > 1.12 ? 4% +3.1 4.24 +1.1 2.22 +1.4 2.51 pp.self.native_queued_spin_lock_slowpath
> > > > > 0.28 ? 9% +3.8 4.06 ? 4% +0.2 0.48 +0.4 0.68 pp.self.sctp_eat_data
> > > > > 0.00 +8.2 8.23 +0.8 0.83 +1.3 1.26 pp.self.__sk_mem_reduce_allocated
> > > > >
> > > > > And the size of 'mem_cgroup' is increased from 4224 Bytes to 4608.
> > > >
> > > > Hi Feng, can you please try two more configurations? Take Eric's patch
> > > > of adding ____cacheline_aligned_in_smp in page_counter and for first
> > > > increase MEMCG_CHARGE_BATCH to 64 and for second increase it to 128.
> > > > Basically batch increases combined with Eric's patch.
> > >
> > > With increasing batch to 128, the regression could be reduced to -12.4%.
> >
> > If we're going to bump it, I wonder if we should scale it dynamically depending
> > on the size of the memory cgroup?
>
> I think it makes sense, or also make it a configurable parameter? From
> the test reports of 0Day, these charging/counting play critical role
> in performance (easy to see up to 60% performance effect). If user only
> wants memcg for isolating things or doesn't care charging/stats, these
> seem to be extra taxes.
>
> For bumping to 64 or 128, universal improvement is expected with the
> only concern of accuracy.
>
> Thanks,
> Feng
>
> > Thanks!

2022-08-16 16:53:42

by Shakeel Butt

[permalink] [raw]
Subject: Re: [net] 4890b686f4: netperf.Throughput_Mbps -69.4% regression

On Mon, Aug 15, 2022 at 10:53 PM Oliver Sang <[email protected]> wrote:
>
> Hi all,
>
> now we noticed this commit has already merged into mainline, and in our tests
> there is still similar regression. [1]
>
> not sure if there is a plan to merge some of the solutions that have been
> discussed in this thread? we'll very glad to test patches if there is a request
>
> Thanks a lot!

Hi Oliver, sorry for the delay. I will send out the patches in a day or two.

thanks,
Shakeel