2022-10-18 09:02:20

by kernel test robot

[permalink] [raw]
Subject: [mm] f35b5d7d67: will-it-scale.per_process_ops -95.5% regression

Hi Rik,

Please be noted that we reported
[mm] 96db82a66d: will-it-scale.per_process_ops -95.3% regression
at
https://lore.kernel.org/all/YwIoiIYo4qsYBcgd@xsang-OptiPlex-9020/
when this commit was on
https://git.kernel.org/pub/scm/linux/kernel/git/next/linux-next.git master
now we noticed the commit has been merged to mainline and we still
observed similar test result.

We are not sure if this is an expected performance change when
switching to huge page or if it could benefit other use cases, so we
report again FYI. Please feel free to ignore the report if the test
result is just as expected. Thanks.


Greeting,

FYI, we noticed a -95.5% regression of will-it-scale.per_process_ops due to commit:

commit: f35b5d7d676e59e401690b678cd3cfec5e785c23 ("mm: align larger anonymous mappings on THP boundaries")
https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master

in testcase: will-it-scale
on test machine: 128 threads 2 sockets Intel(R) Xeon(R) Gold 6338 CPU @ 2.00GHz (Ice Lake) with 256G memory
with following parameters:

nr_task: 100%
mode: process
test: malloc1
cpufreq_governor: performance

test-description: Will It Scale takes a testcase and runs it from 1 through to n parallel copies to see if the testcase will scale. It builds both a process and threads based test in order to see any differences between the two.
test-url: https://github.com/antonblanchard/will-it-scale

In addition to that, the commit also has significant impact on the following tests:

+------------------+----------------------------------------------------------------------------------------------------+
| testcase: change | stress-ng: stress-ng.numa.ops_per_sec -52.9% regression |
| test machine | 192 threads 4 sockets Intel(R) Xeon(R) Platinum 9242 CPU @ 2.30GHz (Cascade Lake) with 192G memory |
| test parameters | class=cpu |
| | cpufreq_governor=performance |
| | nr_threads=100% |
| | test=numa |
| | testtime=60s |
+------------------+----------------------------------------------------------------------------------------------------+
| testcase: change | will-it-scale: will-it-scale.per_process_ops -92.9% regression |
| test machine | 88 threads 2 sockets Intel(R) Xeon(R) Gold 6238M CPU @ 2.10GHz (Cascade Lake) with 128G memory |
| test parameters | cpufreq_governor=performance |
| | mode=process |
| | nr_task=16 |
| | test=malloc1 |
+------------------+----------------------------------------------------------------------------------------------------+


Details are as below:

=========================================================================================
compiler/cpufreq_governor/kconfig/mode/nr_task/rootfs/tbox_group/test/testcase:
gcc-11/performance/x86_64-rhel-8.3/process/100%/debian-11.1-x86_64-20220510.cgz/lkp-icl-2sp2/malloc1/will-it-scale

commit:
7b5a0b664e ("mm/page_ext: remove unused variable in offline_page_ext")
f35b5d7d67 ("mm: align larger anonymous mappings on THP boundaries")

7b5a0b664ebe2625 f35b5d7d676e59e401690b678cd
---------------- ---------------------------
%stddev %change %stddev
\ | \
2765078 -95.5% 124266 ? 4% will-it-scale.128.processes
21601 -95.5% 970.33 ? 4% will-it-scale.per_process_ops
2765078 -95.5% 124266 ? 4% will-it-scale.workload
1943 -3.5% 1874 vmstat.system.cs
0.68 +1.6 2.32 ? 4% mpstat.cpu.all.irq%
0.00 ? 3% +0.0 0.03 ? 4% mpstat.cpu.all.soft%
0.89 ? 2% -0.4 0.49 ? 2% mpstat.cpu.all.usr%
0.09 ? 4% -88.7% 0.01 turbostat.IPC
351.67 +11.2% 391.20 turbostat.PkgWatt
29.30 +299.5% 117.04 turbostat.RAMWatt
8.251e+08 -95.5% 37387027 numa-numastat.node0.local_node
8.252e+08 -95.5% 37467051 numa-numastat.node0.numa_hit
8.405e+08 -95.5% 38161942 ? 7% numa-numastat.node1.local_node
8.406e+08 -95.5% 38196663 ? 7% numa-numastat.node1.numa_hit
174409 ? 9% -21.4% 137126 ? 2% meminfo.Active
174071 ? 9% -21.4% 136806 ? 2% meminfo.Active(anon)
311891 +10.9% 346028 meminfo.AnonPages
343079 +42.0% 487127 meminfo.Inactive
342068 +42.1% 486072 meminfo.Inactive(anon)
69414 ? 2% +221.8% 223379 ? 2% meminfo.Mapped
204255 ? 8% +24.9% 255031 meminfo.Shmem
32528 ? 48% +147.6% 80547 ? 38% numa-meminfo.node0.AnonHugePages
92821 ? 23% +59.3% 147839 ? 28% numa-meminfo.node0.AnonPages
99694 ? 17% +56.9% 156414 ? 26% numa-meminfo.node0.Inactive
99136 ? 17% +57.7% 156290 ? 26% numa-meminfo.node0.Inactive(anon)
30838 ? 53% +134.0% 72155 ? 21% numa-meminfo.node0.Mapped
171865 ? 9% -22.7% 132920 ? 2% numa-meminfo.node1.Active
171791 ? 9% -22.7% 132730 ? 2% numa-meminfo.node1.Active(anon)
243260 ? 7% +36.0% 330799 ? 11% numa-meminfo.node1.Inactive
242807 ? 7% +35.9% 329868 ? 11% numa-meminfo.node1.Inactive(anon)
38681 ? 37% +291.2% 151319 ? 8% numa-meminfo.node1.Mapped
195654 ? 8% +27.6% 249732 numa-meminfo.node1.Shmem
23192 ? 23% +59.3% 36946 ? 28% numa-vmstat.node0.nr_anon_pages
24771 ? 17% +57.7% 39074 ? 26% numa-vmstat.node0.nr_inactive_anon
7625 ? 53% +136.5% 18031 ? 21% numa-vmstat.node0.nr_mapped
24771 ? 17% +57.8% 39085 ? 26% numa-vmstat.node0.nr_zone_inactive_anon
8.252e+08 -95.5% 37466761 numa-vmstat.node0.numa_hit
8.251e+08 -95.5% 37386737 numa-vmstat.node0.numa_local
43036 ? 9% -23.1% 33107 ? 2% numa-vmstat.node1.nr_active_anon
60590 ? 7% +36.1% 82475 ? 11% numa-vmstat.node1.nr_inactive_anon
9533 ? 38% +297.1% 37858 ? 8% numa-vmstat.node1.nr_mapped
48889 ? 8% +27.6% 62403 numa-vmstat.node1.nr_shmem
43036 ? 9% -23.1% 33107 ? 2% numa-vmstat.node1.nr_zone_active_anon
60589 ? 7% +36.0% 82430 ? 11% numa-vmstat.node1.nr_zone_inactive_anon
8.406e+08 -95.5% 38196529 ? 7% numa-vmstat.node1.numa_hit
8.405e+08 -95.5% 38161808 ? 7% numa-vmstat.node1.numa_local
43513 ? 9% -21.8% 34042 ? 2% proc-vmstat.nr_active_anon
77940 +11.0% 86526 proc-vmstat.nr_anon_pages
762952 +1.7% 775553 proc-vmstat.nr_file_pages
85507 +42.1% 121487 proc-vmstat.nr_inactive_anon
17361 ? 2% +221.5% 55823 ? 2% proc-vmstat.nr_mapped
3300 +4.6% 3452 proc-vmstat.nr_page_table_pages
51081 ? 8% +24.6% 63669 proc-vmstat.nr_shmem
43513 ? 9% -21.8% 34042 ? 2% proc-vmstat.nr_zone_active_anon
85507 +42.1% 121480 proc-vmstat.nr_zone_inactive_anon
23080 ? 20% +56.7% 36156 ? 11% proc-vmstat.numa_hint_faults
16266 ? 13% +75.2% 28496 ? 5% proc-vmstat.numa_hint_faults_local
1.666e+09 -95.5% 75751403 ? 3% proc-vmstat.numa_hit
63.17 ? 50% +2948.3% 1925 proc-vmstat.numa_huge_pte_updates
1.666e+09 -95.5% 75551968 ? 4% proc-vmstat.numa_local
176965 ? 9% +543.0% 1137910 proc-vmstat.numa_pte_updates
160517 ? 3% -14.3% 137522 ? 2% proc-vmstat.pgactivate
1.665e+09 -95.5% 75663487 ? 4% proc-vmstat.pgalloc_normal
8.332e+08 -95.4% 38289978 ? 4% proc-vmstat.pgfault
1.665e+09 -95.5% 75646557 ? 4% proc-vmstat.pgfree
18.00 +2.1e+08% 37369911 ? 4% proc-vmstat.thp_fault_alloc
1.46 ?223% +1.3e+06% 19552 sched_debug.cfs_rq:/.MIN_vruntime.avg
187.51 ?223% +1.3e+06% 2502777 sched_debug.cfs_rq:/.MIN_vruntime.max
16.51 ?223% +1.3e+06% 220350 sched_debug.cfs_rq:/.MIN_vruntime.stddev
233.78 ? 28% +45.4% 339.81 ? 29% sched_debug.cfs_rq:/.load_avg.max
1.46 ?223% +1.3e+06% 19552 sched_debug.cfs_rq:/.max_vruntime.avg
187.51 ?223% +1.3e+06% 2502777 sched_debug.cfs_rq:/.max_vruntime.max
16.51 ?223% +1.3e+06% 220350 sched_debug.cfs_rq:/.max_vruntime.stddev
20463200 -12.7% 17863314 sched_debug.cfs_rq:/.min_vruntime.min
227934 ? 6% +73.5% 395381 ? 6% sched_debug.cfs_rq:/.min_vruntime.stddev
557786 ? 6% +22.1% 680843 ? 7% sched_debug.cfs_rq:/.spread0.max
-668417 +343.1% -2961726 sched_debug.cfs_rq:/.spread0.min
227979 ? 6% +73.4% 395300 ? 6% sched_debug.cfs_rq:/.spread0.stddev
793.86 ? 3% -31.1% 546.72 ? 9% sched_debug.cfs_rq:/.util_avg.min
57.90 ? 8% +47.0% 85.09 ? 11% sched_debug.cfs_rq:/.util_avg.stddev
535.54 ? 3% -18.1% 438.80 sched_debug.cfs_rq:/.util_est_enqueued.avg
224.57 ? 5% -38.5% 138.22 ? 8% sched_debug.cfs_rq:/.util_est_enqueued.stddev
957251 +12.9% 1080580 ? 11% sched_debug.cpu.avg_idle.avg
10.65 ? 5% +303.8% 43.02 ? 21% sched_debug.cpu.clock.stddev
644.97 +46.0% 941.70 ? 5% sched_debug.cpu.clock_task.stddev
0.00 ? 13% +107.1% 0.00 ? 18% sched_debug.cpu.next_balance.stddev
1802 ? 10% -19.3% 1454 ? 13% sched_debug.cpu.nr_switches.min
3.28 ? 13% +8839.1% 293.65 ? 3% perf-stat.i.MPKI
1.936e+10 -81.9% 3.503e+09 perf-stat.i.branch-instructions
0.14 -0.0 0.13 ? 3% perf-stat.i.branch-miss-rate%
26013221 -82.5% 4556180 ? 2% perf-stat.i.branch-misses
5.62 ? 5% +46.1 51.68 ? 3% perf-stat.i.cache-miss-rate%
15513099 ? 8% +14150.9% 2.211e+09 perf-stat.i.cache-misses
2.787e+08 ? 13% +1437.2% 4.284e+09 ? 4% perf-stat.i.cache-references
1870 -3.6% 1803 perf-stat.i.context-switches
3.87 +474.5% 22.23 perf-stat.i.cpi
3.288e+11 -1.4% 3.242e+11 perf-stat.i.cpu-cycles
174.70 -10.3% 156.69 perf-stat.i.cpu-migrations
21352 ? 8% -99.3% 159.27 ? 17% perf-stat.i.cycles-between-cache-misses
0.01 -0.0 0.00 ? 11% perf-stat.i.dTLB-load-miss-rate%
2874326 ? 2% -94.7% 152528 ? 10% perf-stat.i.dTLB-load-misses
2.047e+10 -81.3% 3.825e+09 perf-stat.i.dTLB-loads
0.25 -0.2 0.06 perf-stat.i.dTLB-store-miss-rate%
19343669 -95.4% 891050 ? 4% perf-stat.i.dTLB-store-misses
7.829e+09 -80.1% 1.561e+09 ? 3% perf-stat.i.dTLB-stores
8.49e+10 -82.8% 1.463e+10 perf-stat.i.instructions
0.26 -82.2% 0.05 perf-stat.i.ipc
0.14 ? 38% +60.3% 0.22 ? 12% perf-stat.i.major-faults
2.57 -1.4% 2.53 perf-stat.i.metric.GHz
265.58 +203.4% 805.78 ? 15% perf-stat.i.metric.K/sec
374.50 -68.2% 119.04 perf-stat.i.metric.M/sec
2757302 -95.4% 126231 ? 4% perf-stat.i.minor-faults
92.63 +3.9 96.51 perf-stat.i.node-load-miss-rate%
3007607 ? 4% +285.4% 11591077 ? 6% perf-stat.i.node-load-misses
240194 ? 17% +71.9% 412981 ? 6% perf-stat.i.node-loads
97.87 -92.4 5.47 ? 7% perf-stat.i.node-store-miss-rate%
5503394 +2009.9% 1.161e+08 ? 7% perf-stat.i.node-store-misses
119412 ? 6% +1.7e+06% 2.041e+09 perf-stat.i.node-stores
2757302 -95.4% 126231 ? 4% perf-stat.i.page-faults
3.28 ? 13% +8826.2% 293.21 ? 3% perf-stat.overall.MPKI
0.13 -0.0 0.13 ? 2% perf-stat.overall.branch-miss-rate%
5.61 ? 5% +46.1 51.70 ? 3% perf-stat.overall.cache-miss-rate%
3.87 +473.3% 22.20 perf-stat.overall.cpi
21335 ? 8% -99.3% 146.65 perf-stat.overall.cycles-between-cache-misses
0.01 -0.0 0.00 ? 9% perf-stat.overall.dTLB-load-miss-rate%
0.25 -0.2 0.06 perf-stat.overall.dTLB-store-miss-rate%
0.26 -82.6% 0.05 perf-stat.overall.ipc
92.65 +4.0 96.63 perf-stat.overall.node-load-miss-rate%
97.88 -92.5 5.38 ? 8% perf-stat.overall.node-store-miss-rate%
9272709 +283.9% 35600802 ? 3% perf-stat.overall.path-length
1.929e+10 -81.9% 3.487e+09 perf-stat.ps.branch-instructions
25928796 -82.7% 4477103 ? 2% perf-stat.ps.branch-misses
15464091 ? 8% +14157.2% 2.205e+09 perf-stat.ps.cache-misses
2.778e+08 ? 13% +1437.1% 4.27e+09 ? 4% perf-stat.ps.cache-references
1865 -3.9% 1791 perf-stat.ps.context-switches
3.277e+11 -1.4% 3.233e+11 perf-stat.ps.cpu-cycles
174.25 -11.7% 153.93 perf-stat.ps.cpu-migrations
2866660 ? 2% -94.7% 151686 ? 10% perf-stat.ps.dTLB-load-misses
2.041e+10 -81.3% 3.808e+09 perf-stat.ps.dTLB-loads
19279774 -95.4% 888826 ? 4% perf-stat.ps.dTLB-store-misses
7.803e+09 -80.1% 1.555e+09 ? 3% perf-stat.ps.dTLB-stores
8.462e+10 -82.8% 1.456e+10 perf-stat.ps.instructions
0.14 ? 38% +56.7% 0.21 ? 14% perf-stat.ps.major-faults
2748185 -95.4% 125830 ? 4% perf-stat.ps.minor-faults
2998556 ? 4% +291.3% 11734146 ? 6% perf-stat.ps.node-load-misses
239400 ? 17% +70.8% 408868 ? 7% perf-stat.ps.node-loads
5485289 +2007.4% 1.156e+08 ? 7% perf-stat.ps.node-store-misses
119090 ? 6% +1.7e+06% 2.035e+09 perf-stat.ps.node-stores
2748185 -95.4% 125831 ? 4% perf-stat.ps.page-faults
2.564e+13 -82.8% 4.417e+12 perf-stat.total.instructions
95.23 -79.8 15.41 ? 6% perf-profile.calltrace.cycles-pp.__munmap
95.08 -79.7 15.40 ? 6% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__munmap
95.02 -79.6 15.39 ? 6% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
94.96 -79.6 15.37 ? 6% perf-profile.calltrace.cycles-pp.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
94.95 -79.6 15.37 ? 6% perf-profile.calltrace.cycles-pp.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
94.86 -79.5 15.35 ? 6% perf-profile.calltrace.cycles-pp.__do_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe
94.38 -79.2 15.22 ? 6% perf-profile.calltrace.cycles-pp.unmap_region.__do_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64
42.74 -42.7 0.00 perf-profile.calltrace.cycles-pp.lru_add_drain.unmap_region.__do_munmap.__vm_munmap.__x64_sys_munmap
42.74 -42.7 0.00 perf-profile.calltrace.cycles-pp.lru_add_drain_cpu.lru_add_drain.unmap_region.__do_munmap.__vm_munmap
42.72 -42.7 0.00 perf-profile.calltrace.cycles-pp.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain.unmap_region.__do_munmap
41.84 -41.8 0.00 perf-profile.calltrace.cycles-pp.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain.unmap_region
41.70 -41.7 0.00 perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain
41.62 -41.6 0.00 perf-profile.calltrace.cycles-pp.folio_lruvec_lock_irqsave.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region
41.55 -41.6 0.00 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu
41.52 -41.5 0.00 perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.release_pages.tlb_batch_pages_flush.tlb_finish_mmu
41.28 -41.3 0.00 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.release_pages.tlb_batch_pages_flush
46.93 -37.0 9.94 ? 6% perf-profile.calltrace.cycles-pp.tlb_finish_mmu.unmap_region.__do_munmap.__vm_munmap.__x64_sys_munmap
43.64 -33.8 9.84 ? 6% perf-profile.calltrace.cycles-pp.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region.__do_munmap.__vm_munmap
43.40 -33.6 9.81 ? 6% perf-profile.calltrace.cycles-pp.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region.__do_munmap
0.00 +0.6 0.56 ? 6% perf-profile.calltrace.cycles-pp.update_process_times.tick_sched_handle.tick_sched_timer.__hrtimer_run_queues.hrtimer_interrupt
0.00 +0.6 0.57 ? 6% perf-profile.calltrace.cycles-pp.tick_sched_handle.tick_sched_timer.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt
0.00 +0.6 0.60 ? 6% perf-profile.calltrace.cycles-pp.tick_sched_timer.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt
0.00 +0.7 0.67 ? 5% perf-profile.calltrace.cycles-pp.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
0.00 +0.7 0.74 ? 5% perf-profile.calltrace.cycles-pp.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.clear_page_erms
0.00 +0.8 0.77 ? 4% perf-profile.calltrace.cycles-pp.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.clear_page_erms.clear_huge_page
0.00 +0.8 0.81 ? 4% perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.clear_page_erms.clear_huge_page.__do_huge_pmd_anonymous_page
0.00 +1.1 1.09 ? 3% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.clear_page_erms.clear_huge_page.__do_huge_pmd_anonymous_page.__handle_mm_fault
0.00 +1.1 1.11 ? 3% perf-profile.calltrace.cycles-pp.free_pcp_prepare.free_unref_page.release_pages.tlb_batch_pages_flush.tlb_finish_mmu
3.60 ? 3% +1.5 5.08 ? 7% perf-profile.calltrace.cycles-pp.unmap_vmas.unmap_region.__do_munmap.__vm_munmap.__x64_sys_munmap
3.51 ? 3% +1.6 5.06 ? 7% perf-profile.calltrace.cycles-pp.unmap_page_range.unmap_vmas.unmap_region.__do_munmap.__vm_munmap
3.29 ? 3% +1.7 5.04 ? 7% perf-profile.calltrace.cycles-pp.zap_pmd_range.unmap_page_range.unmap_vmas.unmap_region.__do_munmap
0.00 +2.8 2.78 ? 2% perf-profile.calltrace.cycles-pp.__cond_resched.clear_huge_page.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault
0.00 +3.3 3.28 ? 2% perf-profile.calltrace.cycles-pp.__might_resched.clear_huge_page.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault
0.00 +4.2 4.21 ? 8% perf-profile.calltrace.cycles-pp._raw_spin_lock.free_pcppages_bulk.free_unref_page.zap_huge_pmd.zap_pmd_range
0.00 +4.2 4.21 ? 8% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.free_pcppages_bulk.free_unref_page.zap_huge_pmd
0.00 +4.3 4.34 ? 8% perf-profile.calltrace.cycles-pp.rmqueue_bulk.rmqueue.get_page_from_freelist.__alloc_pages.pte_alloc_one
0.00 +4.4 4.40 ? 8% perf-profile.calltrace.cycles-pp.rmqueue.get_page_from_freelist.__alloc_pages.pte_alloc_one.__do_huge_pmd_anonymous_page
0.00 +4.4 4.43 ? 8% perf-profile.calltrace.cycles-pp.rmqueue_bulk.rmqueue.get_page_from_freelist.__alloc_pages.__folio_alloc
0.00 +4.5 4.49 ? 8% perf-profile.calltrace.cycles-pp.get_page_from_freelist.__alloc_pages.pte_alloc_one.__do_huge_pmd_anonymous_page.__handle_mm_fault
0.00 +4.5 4.51 ? 8% perf-profile.calltrace.cycles-pp.rmqueue.get_page_from_freelist.__alloc_pages.__folio_alloc.vma_alloc_folio
0.00 +4.6 4.59 ? 8% perf-profile.calltrace.cycles-pp.__alloc_pages.pte_alloc_one.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault
0.00 +4.6 4.62 ? 8% perf-profile.calltrace.cycles-pp.pte_alloc_one.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
0.00 +4.6 4.63 ? 8% perf-profile.calltrace.cycles-pp.free_pcppages_bulk.free_unref_page.zap_huge_pmd.zap_pmd_range.unmap_page_range
0.00 +4.7 4.70 ? 7% perf-profile.calltrace.cycles-pp.get_page_from_freelist.__alloc_pages.__folio_alloc.vma_alloc_folio.do_huge_pmd_anonymous_page
0.00 +4.7 4.72 ? 7% perf-profile.calltrace.cycles-pp.__alloc_pages.__folio_alloc.vma_alloc_folio.do_huge_pmd_anonymous_page.__handle_mm_fault
0.00 +4.7 4.73 ? 7% perf-profile.calltrace.cycles-pp.__folio_alloc.vma_alloc_folio.do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault
0.00 +4.8 4.75 ? 7% perf-profile.calltrace.cycles-pp.vma_alloc_folio.do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
0.00 +4.8 4.76 ? 8% perf-profile.calltrace.cycles-pp.free_unref_page.zap_huge_pmd.zap_pmd_range.unmap_page_range.unmap_vmas
0.00 +4.8 4.82 ? 7% perf-profile.calltrace.cycles-pp.do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
0.00 +4.9 4.88 ? 7% perf-profile.calltrace.cycles-pp.zap_huge_pmd.zap_pmd_range.unmap_page_range.unmap_vmas.unmap_region
0.00 +8.2 8.22 ? 8% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.rmqueue_bulk.rmqueue.get_page_from_freelist
0.00 +8.2 8.23 ? 8% perf-profile.calltrace.cycles-pp._raw_spin_lock.rmqueue_bulk.rmqueue.get_page_from_freelist.__alloc_pages
0.00 +8.3 8.35 ? 8% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.free_pcppages_bulk.free_unref_page.release_pages
0.00 +8.3 8.35 ? 8% perf-profile.calltrace.cycles-pp._raw_spin_lock.free_pcppages_bulk.free_unref_page.release_pages.tlb_batch_pages_flush
0.00 +8.4 8.37 ? 8% perf-profile.calltrace.cycles-pp.free_pcppages_bulk.free_unref_page.release_pages.tlb_batch_pages_flush.tlb_finish_mmu
0.00 +9.6 9.60 ? 6% perf-profile.calltrace.cycles-pp.free_unref_page.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region
0.00 +65.5 65.48 ? 2% perf-profile.calltrace.cycles-pp.clear_page_erms.clear_huge_page.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault
0.00 +72.5 72.51 ? 2% perf-profile.calltrace.cycles-pp.clear_huge_page.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
0.00 +78.6 78.58 perf-profile.calltrace.cycles-pp.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
2.62 ? 3% +80.9 83.56 perf-profile.calltrace.cycles-pp.asm_exc_page_fault
2.60 ? 3% +81.0 83.56 perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault
2.58 ? 3% +81.0 83.57 perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
2.38 ? 3% +81.1 83.52 perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
2.26 ? 3% +81.2 83.45 perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
83.48 -83.4 0.06 ? 9% perf-profile.children.cycles-pp.folio_lruvec_lock_irqsave
83.28 -83.2 0.08 ? 8% perf-profile.children.cycles-pp._raw_spin_lock_irqsave
96.34 -80.3 16.09 ? 6% perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
96.27 -80.2 16.08 ? 6% perf-profile.children.cycles-pp.do_syscall_64
95.28 -79.9 15.41 ? 6% perf-profile.children.cycles-pp.__munmap
94.96 -79.6 15.37 ? 6% perf-profile.children.cycles-pp.__x64_sys_munmap
94.96 -79.6 15.37 ? 6% perf-profile.children.cycles-pp.__vm_munmap
94.87 -79.5 15.36 ? 6% perf-profile.children.cycles-pp.__do_munmap
94.39 -79.2 15.22 ? 6% perf-profile.children.cycles-pp.unmap_region
82.88 -62.0 20.90 ? 8% perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
42.78 -42.8 0.00 perf-profile.children.cycles-pp.lru_add_drain
42.76 -42.8 0.00 perf-profile.children.cycles-pp.lru_add_drain_cpu
42.75 -42.6 0.10 perf-profile.children.cycles-pp.folio_batch_move_lru
46.94 -37.0 9.94 ? 6% perf-profile.children.cycles-pp.tlb_finish_mmu
43.64 -33.8 9.84 ? 6% perf-profile.children.cycles-pp.tlb_batch_pages_flush
43.62 -33.8 9.82 ? 6% perf-profile.children.cycles-pp.release_pages
3.21 ? 4% -3.1 0.09 ? 5% perf-profile.children.cycles-pp.flush_tlb_mm_range
3.11 ? 4% -3.1 0.06 ? 8% perf-profile.children.cycles-pp.flush_tlb_func
3.00 ? 4% -3.0 0.03 ? 70% perf-profile.children.cycles-pp.native_flush_tlb_one_user
1.33 ? 3% -0.9 0.42 ? 5% perf-profile.children.cycles-pp.__mmap
1.10 ? 4% -0.7 0.39 ? 5% perf-profile.children.cycles-pp.vm_mmap_pgoff
0.79 ? 4% -0.7 0.09 ? 7% perf-profile.children.cycles-pp.uncharge_batch
0.97 ? 4% -0.6 0.36 ? 5% perf-profile.children.cycles-pp.do_mmap
0.65 ? 5% -0.6 0.06 ? 8% perf-profile.children.cycles-pp.page_counter_uncharge
0.64 ? 4% -0.6 0.05 ? 8% perf-profile.children.cycles-pp.free_pgd_range
0.62 ? 4% -0.6 0.05 perf-profile.children.cycles-pp.free_p4d_range
0.81 ? 3% -0.5 0.30 ? 6% perf-profile.children.cycles-pp.mmap_region
0.59 ? 4% -0.5 0.08 ? 6% perf-profile.children.cycles-pp.kmem_cache_alloc
0.66 ? 5% -0.5 0.16 ? 4% perf-profile.children.cycles-pp.__mod_lruvec_page_state
0.55 ? 3% -0.5 0.06 ? 9% perf-profile.children.cycles-pp.__anon_vma_prepare
0.44 ? 4% -0.4 0.06 ? 9% perf-profile.children.cycles-pp.lru_add_fn
0.42 ? 5% -0.3 0.12 ? 4% perf-profile.children.cycles-pp.free_pgtables
0.40 ? 4% -0.3 0.10 ? 9% perf-profile.children.cycles-pp.kmem_cache_free
0.41 ? 5% -0.3 0.11 ? 5% perf-profile.children.cycles-pp.unlink_anon_vmas
0.25 ? 5% -0.2 0.03 ? 70% perf-profile.children.cycles-pp.vm_area_alloc
0.42 ? 6% -0.2 0.26 ? 3% perf-profile.children.cycles-pp._raw_spin_unlock_irqrestore
0.16 ? 5% -0.1 0.02 ? 99% perf-profile.children.cycles-pp.__put_anon_vma
0.27 ? 4% -0.1 0.13 ? 3% perf-profile.children.cycles-pp.native_irq_return_iret
0.28 ? 5% -0.1 0.14 ? 7% perf-profile.children.cycles-pp.perf_event_mmap
0.18 ? 4% -0.1 0.07 ? 7% perf-profile.children.cycles-pp.page_add_new_anon_rmap
0.23 ? 3% -0.1 0.14 ? 7% perf-profile.children.cycles-pp.perf_event_mmap_event
0.18 ? 4% -0.1 0.08 ? 5% perf-profile.children.cycles-pp.__memcg_kmem_charge_page
0.16 ? 5% -0.1 0.07 perf-profile.children.cycles-pp.page_remove_rmap
0.13 ? 2% -0.1 0.05 ? 7% perf-profile.children.cycles-pp.get_unmapped_area
0.15 ? 5% -0.0 0.10 ? 9% perf-profile.children.cycles-pp.perf_iterate_sb
0.10 ? 5% -0.0 0.06 ? 9% perf-profile.children.cycles-pp.find_vma
0.39 ? 3% -0.0 0.35 ? 4% perf-profile.children.cycles-pp.rcu_all_qs
0.09 ? 5% -0.0 0.06 ? 9% perf-profile.children.cycles-pp.__perf_sw_event
0.15 ? 4% -0.0 0.13 ? 7% perf-profile.children.cycles-pp.__mem_cgroup_charge
0.07 -0.0 0.06 ? 8% perf-profile.children.cycles-pp.___perf_sw_event
0.12 ? 5% +0.0 0.14 ? 3% perf-profile.children.cycles-pp.__mod_lruvec_state
0.00 +0.1 0.05 perf-profile.children.cycles-pp.perf_output_sample
0.00 +0.1 0.05 perf-profile.children.cycles-pp.memcg_check_events
0.08 ? 4% +0.1 0.13 ? 5% perf-profile.children.cycles-pp.__mod_node_page_state
0.00 +0.1 0.06 ? 8% perf-profile.children.cycles-pp.__get_user_nocheck_8
0.00 +0.1 0.06 ? 6% perf-profile.children.cycles-pp.perf_callchain_user
0.00 +0.1 0.06 ? 7% perf-profile.children.cycles-pp.update_load_avg
0.00 +0.1 0.06 ? 7% perf-profile.children.cycles-pp.__orc_find
0.19 ? 4% +0.1 0.26 ? 6% perf-profile.children.cycles-pp.__list_del_entry_valid
0.11 ? 4% +0.1 0.18 ? 7% perf-profile.children.cycles-pp.unwind_next_frame
0.00 +0.1 0.08 ? 7% perf-profile.children.cycles-pp.__page_cache_release
0.02 ? 99% +0.1 0.11 ? 4% perf-profile.children.cycles-pp.folio_add_lru
0.00 +0.1 0.08 ? 13% perf-profile.children.cycles-pp.shmem_alloc_and_acct_folio
0.00 +0.1 0.08 ? 13% perf-profile.children.cycles-pp.shmem_alloc_folio
0.00 +0.1 0.09 ? 10% perf-profile.children.cycles-pp.__unwind_start
0.00 +0.1 0.09 ? 11% perf-profile.children.cycles-pp.shmem_write_begin
0.00 +0.1 0.09 ? 11% perf-profile.children.cycles-pp.shmem_getpage_gfp
0.00 +0.1 0.10 ? 6% perf-profile.children.cycles-pp.free_compound_page
0.00 +0.1 0.10 ? 6% perf-profile.children.cycles-pp.__mem_cgroup_uncharge
0.13 ? 4% +0.1 0.23 ? 7% perf-profile.children.cycles-pp.perf_callchain_kernel
0.13 ? 17% +0.1 0.24 ? 15% perf-profile.children.cycles-pp.cmd_record
0.13 ? 15% +0.1 0.24 ? 16% perf-profile.children.cycles-pp.__libc_start_main
0.13 ? 15% +0.1 0.24 ? 16% perf-profile.children.cycles-pp.main
0.13 ? 15% +0.1 0.24 ? 16% perf-profile.children.cycles-pp.run_builtin
0.04 ? 47% +0.1 0.16 ? 11% perf-profile.children.cycles-pp.generic_perform_write
0.04 ? 47% +0.1 0.16 ? 12% perf-profile.children.cycles-pp.generic_file_write_iter
0.04 ? 47% +0.1 0.16 ? 12% perf-profile.children.cycles-pp.__generic_file_write_iter
0.04 ? 47% +0.1 0.16 ? 14% perf-profile.children.cycles-pp.record__pushfn
0.04 ? 47% +0.1 0.16 ? 14% perf-profile.children.cycles-pp.__libc_write
0.04 ? 47% +0.1 0.16 ? 13% perf-profile.children.cycles-pp.vfs_write
0.04 ? 47% +0.1 0.16 ? 13% perf-profile.children.cycles-pp.perf_mmap__push
0.04 ? 47% +0.1 0.16 ? 13% perf-profile.children.cycles-pp.ksys_write
0.04 ? 47% +0.1 0.17 ? 12% perf-profile.children.cycles-pp.record__mmap_read_evlist
0.13 ? 17% +0.1 0.27 ? 17% perf-profile.children.cycles-pp.__cmd_record
0.15 ? 3% +0.2 0.30 ? 7% perf-profile.children.cycles-pp.get_perf_callchain
0.15 ? 3% +0.2 0.31 ? 6% perf-profile.children.cycles-pp.perf_callchain
0.16 ? 6% +0.2 0.33 ? 7% perf-profile.children.cycles-pp.perf_prepare_sample
0.17 ? 4% +0.2 0.41 ? 6% perf-profile.children.cycles-pp.perf_event_output_forward
0.17 ? 4% +0.2 0.41 ? 6% perf-profile.children.cycles-pp.__perf_event_overflow
0.00 +0.3 0.27 ? 8% perf-profile.children.cycles-pp.__free_one_page
0.18 ? 5% +0.3 0.44 ? 6% perf-profile.children.cycles-pp.perf_tp_event
0.18 ? 5% +0.3 0.46 ? 6% perf-profile.children.cycles-pp.perf_trace_sched_stat_runtime
0.19 ? 5% +0.3 0.53 ? 6% perf-profile.children.cycles-pp.update_curr
0.21 ? 3% +0.4 0.62 ? 5% perf-profile.children.cycles-pp.task_tick_fair
0.00 +0.4 0.42 ? 8% perf-profile.children.cycles-pp.check_new_pages
0.23 ? 3% +0.5 0.72 ? 5% perf-profile.children.cycles-pp.scheduler_tick
0.25 ? 4% +0.6 0.83 ? 5% perf-profile.children.cycles-pp.update_process_times
0.25 ? 4% +0.6 0.84 ? 5% perf-profile.children.cycles-pp.tick_sched_handle
0.26 ? 4% +0.6 0.87 ? 5% perf-profile.children.cycles-pp.tick_sched_timer
0.29 ? 3% +0.7 0.98 ? 4% perf-profile.children.cycles-pp.__hrtimer_run_queues
0.34 ? 4% +0.7 1.09 ? 4% perf-profile.children.cycles-pp.hrtimer_interrupt
0.36 ? 4% +0.8 1.12 ? 4% perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
0.38 ? 4% +0.8 1.18 ? 4% perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
0.44 ? 5% +1.1 1.51 ? 3% perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
0.00 +1.1 1.14 ? 2% perf-profile.children.cycles-pp.free_pcp_prepare
3.60 ? 3% +1.5 5.08 ? 7% perf-profile.children.cycles-pp.unmap_vmas
3.52 ? 3% +1.6 5.07 ? 7% perf-profile.children.cycles-pp.unmap_page_range
3.43 ? 3% +1.6 5.05 ? 7% perf-profile.children.cycles-pp.zap_pmd_range
0.74 ? 3% +1.7 2.42 ? 2% perf-profile.children.cycles-pp.__cond_resched
0.79 ? 4% +2.9 3.67 ? 2% perf-profile.children.cycles-pp.__might_resched
0.73 ? 3% +3.9 4.63 ? 8% perf-profile.children.cycles-pp.pte_alloc_one
0.32 ? 5% +4.5 4.84 ? 7% perf-profile.children.cycles-pp.vma_alloc_folio
0.27 ? 5% +4.5 4.82 ? 7% perf-profile.children.cycles-pp.__folio_alloc
0.00 +4.8 4.82 ? 7% perf-profile.children.cycles-pp.do_huge_pmd_anonymous_page
0.00 +4.9 4.88 ? 7% perf-profile.children.cycles-pp.zap_huge_pmd
0.75 ? 4% +8.7 9.41 ? 7% perf-profile.children.cycles-pp.__alloc_pages
0.16 ? 5% +8.8 9.00 ? 8% perf-profile.children.cycles-pp.rmqueue
0.00 +8.9 8.86 ? 8% perf-profile.children.cycles-pp.rmqueue_bulk
0.42 ? 5% +8.9 9.28 ? 7% perf-profile.children.cycles-pp.get_page_from_freelist
0.00 +13.0 13.02 ? 8% perf-profile.children.cycles-pp.free_pcppages_bulk
0.00 +14.4 14.36 ? 7% perf-profile.children.cycles-pp.free_unref_page
0.12 ? 3% +20.9 21.00 ? 8% perf-profile.children.cycles-pp._raw_spin_lock
0.18 ? 3% +65.9 66.08 ? 2% perf-profile.children.cycles-pp.clear_page_erms
0.00 +73.5 73.51 ? 2% perf-profile.children.cycles-pp.clear_huge_page
0.00 +78.6 78.58 perf-profile.children.cycles-pp.__do_huge_pmd_anonymous_page
2.65 ? 3% +81.0 83.62 perf-profile.children.cycles-pp.asm_exc_page_fault
2.61 ? 3% +81.0 83.59 perf-profile.children.cycles-pp.exc_page_fault
2.60 ? 3% +81.0 83.59 perf-profile.children.cycles-pp.do_user_addr_fault
2.39 ? 3% +81.1 83.53 perf-profile.children.cycles-pp.handle_mm_fault
2.27 ? 3% +81.2 83.46 perf-profile.children.cycles-pp.__handle_mm_fault
82.87 -62.0 20.90 ? 8% perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
2.98 ? 4% -2.9 0.03 ? 70% perf-profile.self.cycles-pp.native_flush_tlb_one_user
0.72 ? 3% -0.6 0.08 ? 10% perf-profile.self.cycles-pp.zap_pmd_range
0.50 ? 5% -0.5 0.03 ? 70% perf-profile.self.cycles-pp.page_counter_uncharge
0.41 ? 4% -0.3 0.06 ? 7% perf-profile.self.cycles-pp._raw_spin_lock_irqsave
0.27 ? 4% -0.1 0.13 ? 3% perf-profile.self.cycles-pp.native_irq_return_iret
0.20 ? 6% -0.1 0.08 ? 8% perf-profile.self.cycles-pp.kmem_cache_free
0.22 ? 3% -0.1 0.14 ? 3% perf-profile.self.cycles-pp.rcu_all_qs
0.08 ? 5% -0.0 0.07 ? 6% perf-profile.self.cycles-pp.try_charge_memcg
0.02 ?141% +0.1 0.07 ? 5% perf-profile.self.cycles-pp.unwind_next_frame
0.08 ? 6% +0.1 0.13 ? 4% perf-profile.self.cycles-pp.__mod_node_page_state
0.00 +0.1 0.06 ? 8% perf-profile.self.cycles-pp.__do_huge_pmd_anonymous_page
0.00 +0.1 0.06 ? 8% perf-profile.self.cycles-pp.page_counter_try_charge
0.00 +0.1 0.06 ? 9% perf-profile.self.cycles-pp.__orc_find
0.19 ? 3% +0.1 0.26 ? 8% perf-profile.self.cycles-pp.__list_del_entry_valid
0.08 ? 7% +0.1 0.19 ? 8% perf-profile.self.cycles-pp.get_page_from_freelist
0.00 +0.3 0.25 ? 8% perf-profile.self.cycles-pp.__free_one_page
0.00 +0.4 0.42 ? 8% perf-profile.self.cycles-pp.check_new_pages
0.00 +1.1 1.10 ? 2% perf-profile.self.cycles-pp.free_pcp_prepare
0.42 ? 4% +1.2 1.59 ? 2% perf-profile.self.cycles-pp.__cond_resched
0.00 +2.5 2.47 ? 2% perf-profile.self.cycles-pp.clear_huge_page
0.70 ? 4% +2.7 3.36 ? 2% perf-profile.self.cycles-pp.__might_resched
0.18 ? 4% +65.0 65.14 ? 2% perf-profile.self.cycles-pp.clear_page_erms


If you fix the issue, kindly add following tag
| Reported-by: kernel test robot <[email protected]>
| Link: https://lore.kernel.org/r/[email protected]


To reproduce:

git clone https://github.com/intel/lkp-tests.git
cd lkp-tests
sudo bin/lkp install job.yaml # job file is attached in this email
bin/lkp split-job --compatible job.yaml # generate the yaml file for lkp run
sudo bin/lkp run generated-yaml-file

# if come across any failure that blocks the test,
# please remove ~/.lkp and /lkp dir to run from a clean state.


Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


--
0-DAY CI Kernel Test Service
https://01.org/lkp


Attachments:
(No filename) (40.73 kB)
config-6.0.0-rc3-00017-gf35b5d7d676e (166.40 kB)
job-script (8.09 kB)
job.yaml (5.63 kB)
reproduce (358.00 B)
Download all attachments

2022-10-19 02:18:32

by Huang, Ying

[permalink] [raw]
Subject: Re: [mm] f35b5d7d67: will-it-scale.per_process_ops -95.5% regression

Hi, Yujie,

> 32528 48% +147.6% 80547 38% numa-meminfo.node0.AnonHugePages
> 92821 23% +59.3% 147839 28% numa-meminfo.node0.AnonPages

The Anon pages allocated are much more than the parent commit. This is
expected, because THP instead of normal page will be allocated for
aligned memory area.

> 95.23 -79.8 15.41 6% perf-profile.calltrace.cycles-pp.__munmap
> 95.08 -79.7 15.40 6% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__munmap
> 95.02 -79.6 15.39 6% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
> 94.96 -79.6 15.37 6% perf-profile.calltrace.cycles-pp.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
> 94.95 -79.6 15.37 6% perf-profile.calltrace.cycles-pp.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
> 94.86 -79.5 15.35 6% perf-profile.calltrace.cycles-pp.__do_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe
> 94.38 -79.2 15.22 6% perf-profile.calltrace.cycles-pp.unmap_region.__do_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64
> 42.74 -42.7 0.00 perf-profile.calltrace.cycles-pp.lru_add_drain.unmap_region.__do_munmap.__vm_munmap.__x64_sys_munmap
> 42.74 -42.7 0.00 perf-profile.calltrace.cycles-pp.lru_add_drain_cpu.lru_add_drain.unmap_region.__do_munmap.__vm_munmap
> 42.72 -42.7 0.00 perf-profile.calltrace.cycles-pp.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain.unmap_region.__do_munmap
> 41.84 -41.8 0.00 perf-profile.calltrace.cycles-pp.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain.unmap_region
> 41.70 -41.7 0.00 perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain
> 41.62 -41.6 0.00 perf-profile.calltrace.cycles-pp.folio_lruvec_lock_irqsave.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region
> 41.55 -41.6 0.00 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu
> 41.52 -41.5 0.00 perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.release_pages.tlb_batch_pages_flush.tlb_finish_mmu
> 41.28 -41.3 0.00 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.release_pages.tlb_batch_pages_flush

In the parent commit, most CPU cycles are used for contention on LRU lock.

> 0.00 +4.8 4.82 7% perf-profile.calltrace.cycles-pp.do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
> 0.00 +4.9 4.88 7% perf-profile.calltrace.cycles-pp.zap_huge_pmd.zap_pmd_range.unmap_page_range.unmap_vmas.unmap_region
> 0.00 +8.2 8.22 8% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.rmqueue_bulk.rmqueue.get_page_from_freelist
> 0.00 +8.2 8.23 8% perf-profile.calltrace.cycles-pp._raw_spin_lock.rmqueue_bulk.rmqueue.get_page_from_freelist.__alloc_pages
> 0.00 +8.3 8.35 8% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.free_pcppages_bulk.free_unref_page.release_pages
> 0.00 +8.3 8.35 8% perf-profile.calltrace.cycles-pp._raw_spin_lock.free_pcppages_bulk.free_unref_page.release_pages.tlb_batch_pages_flush
> 0.00 +8.4 8.37 8% perf-profile.calltrace.cycles-pp.free_pcppages_bulk.free_unref_page.release_pages.tlb_batch_pages_flush.tlb_finish_mmu
> 0.00 +9.6 9.60 6% perf-profile.calltrace.cycles-pp.free_unref_page.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region
> 0.00 +65.5 65.48 2% perf-profile.calltrace.cycles-pp.clear_page_erms.clear_huge_page.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault
> 0.00 +72.5 72.51 2% perf-profile.calltrace.cycles-pp.clear_huge_page.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault

With the commit, most CPU cycles are consumed for clear huge page. This
is expected. We allocate more pages, so, we need more cycles to clear
them.

Check the source code of test case (will-it-scale/malloc1), I found that
it will allocate some memory with malloc() then free it.

In the parent commit, because the virtual memory address isn't aligned
with 2M, normal page will be allocated. With the commit, THP will be
allocated, so more page clearing and less LRU lock contention. I think
this is the expected behavior of the commit. And the test case isn't so
popular (malloc() then free() but don't access the memory allocated). So
this regression isn't important. We can just ignore it.

Best Regards,
Huang, Ying

2022-10-20 04:40:29

by Nathan Chancellor

[permalink] [raw]
Subject: Re: [mm] f35b5d7d67: will-it-scale.per_process_ops -95.5% regression

Hi Ying,

On Wed, Oct 19, 2022 at 10:05:50AM +0800, Huang, Ying wrote:
> Hi, Yujie,
>
> > 32528 48% +147.6% 80547 38% numa-meminfo.node0.AnonHugePages
> > 92821 23% +59.3% 147839 28% numa-meminfo.node0.AnonPages
>
> The Anon pages allocated are much more than the parent commit. This is
> expected, because THP instead of normal page will be allocated for
> aligned memory area.
>
> > 95.23 -79.8 15.41 6% perf-profile.calltrace.cycles-pp.__munmap
> > 95.08 -79.7 15.40 6% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__munmap
> > 95.02 -79.6 15.39 6% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
> > 94.96 -79.6 15.37 6% perf-profile.calltrace.cycles-pp.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
> > 94.95 -79.6 15.37 6% perf-profile.calltrace.cycles-pp.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
> > 94.86 -79.5 15.35 6% perf-profile.calltrace.cycles-pp.__do_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe
> > 94.38 -79.2 15.22 6% perf-profile.calltrace.cycles-pp.unmap_region.__do_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64
> > 42.74 -42.7 0.00 perf-profile.calltrace.cycles-pp.lru_add_drain.unmap_region.__do_munmap.__vm_munmap.__x64_sys_munmap
> > 42.74 -42.7 0.00 perf-profile.calltrace.cycles-pp.lru_add_drain_cpu.lru_add_drain.unmap_region.__do_munmap.__vm_munmap
> > 42.72 -42.7 0.00 perf-profile.calltrace.cycles-pp.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain.unmap_region.__do_munmap
> > 41.84 -41.8 0.00 perf-profile.calltrace.cycles-pp.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain.unmap_region
> > 41.70 -41.7 0.00 perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain
> > 41.62 -41.6 0.00 perf-profile.calltrace.cycles-pp.folio_lruvec_lock_irqsave.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region
> > 41.55 -41.6 0.00 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu
> > 41.52 -41.5 0.00 perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.release_pages.tlb_batch_pages_flush.tlb_finish_mmu
> > 41.28 -41.3 0.00 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.release_pages.tlb_batch_pages_flush
>
> In the parent commit, most CPU cycles are used for contention on LRU lock.
>
> > 0.00 +4.8 4.82 7% perf-profile.calltrace.cycles-pp.do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
> > 0.00 +4.9 4.88 7% perf-profile.calltrace.cycles-pp.zap_huge_pmd.zap_pmd_range.unmap_page_range.unmap_vmas.unmap_region
> > 0.00 +8.2 8.22 8% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.rmqueue_bulk.rmqueue.get_page_from_freelist
> > 0.00 +8.2 8.23 8% perf-profile.calltrace.cycles-pp._raw_spin_lock.rmqueue_bulk.rmqueue.get_page_from_freelist.__alloc_pages
> > 0.00 +8.3 8.35 8% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.free_pcppages_bulk.free_unref_page.release_pages
> > 0.00 +8.3 8.35 8% perf-profile.calltrace.cycles-pp._raw_spin_lock.free_pcppages_bulk.free_unref_page.release_pages.tlb_batch_pages_flush
> > 0.00 +8.4 8.37 8% perf-profile.calltrace.cycles-pp.free_pcppages_bulk.free_unref_page.release_pages.tlb_batch_pages_flush.tlb_finish_mmu
> > 0.00 +9.6 9.60 6% perf-profile.calltrace.cycles-pp.free_unref_page.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region
> > 0.00 +65.5 65.48 2% perf-profile.calltrace.cycles-pp.clear_page_erms.clear_huge_page.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault
> > 0.00 +72.5 72.51 2% perf-profile.calltrace.cycles-pp.clear_huge_page.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
>
> With the commit, most CPU cycles are consumed for clear huge page. This
> is expected. We allocate more pages, so, we need more cycles to clear
> them.
>
> Check the source code of test case (will-it-scale/malloc1), I found that
> it will allocate some memory with malloc() then free it.
>
> In the parent commit, because the virtual memory address isn't aligned
> with 2M, normal page will be allocated. With the commit, THP will be
> allocated, so more page clearing and less LRU lock contention. I think
> this is the expected behavior of the commit. And the test case isn't so
> popular (malloc() then free() but don't access the memory allocated). So
> this regression isn't important. We can just ignore it.

For what it's worth, I just bisected a massive and visible performance
regression on my Threadripper 3990X workstation to commit f35b5d7d676e
("mm: align larger anonymous mappings on THP boundaries"), which seems
directly related to this report/analysis. I initially noticed this
because my full set of kernel builds against mainline went from 2 hours
and 20 minutes or so to over 3 hours. Zeroing in on x86_64 allmodconfig,
which I used for the bisect:

@ 7b5a0b664ebe ("mm/page_ext: remove unused variable in offline_page_ext"):

Benchmark 1: make -skj128 LLVM=1 allmodconfig all
Time (mean ± σ): 318.172 s ± 0.730 s [User: 31750.902 s, System: 4564.246 s]
Range (min … max): 317.332 s … 318.662 s 3 runs

@ f35b5d7d676e ("mm: align larger anonymous mappings on THP boundaries"):

Benchmark 1: make -skj128 LLVM=1 allmodconfig all
Time (mean ± σ): 406.688 s ± 0.676 s [User: 31819.526 s, System: 16327.022 s]
Range (min … max): 405.954 s … 407.284 s 3 run

That is a pretty big difference (27%), which is visible while doing a
lot of builds, only because of the extra system time. If there is any
way to improve this, it should certainly be considered.

For now, I'll just revert it locally.

Cheers,
Nathan

# bad: [aae703b02f92bde9264366c545e87cec451de471] Merge tag 'for-6.1-rc1-tag' of git://git.kernel.org/pub/scm/linux/kernel/git/kdave/linux
# good: [4fe89d07dcc2804c8b562f6c7896a45643d34b2f] Linux 6.0
git bisect start 'aae703b02f92bde9264366c545e87cec451de471' 'v6.0'
# good: [18fd049731e67651009f316195da9281b756f2cf] Merge tag 'arm64-upstream' of git://git.kernel.org/pub/scm/linux/kernel/git/arm64/linux
git bisect good 18fd049731e67651009f316195da9281b756f2cf
# good: [ab0c23b535f3f9d8345d8ad4c18c0a8594459d55] MAINTAINERS: add RISC-V's patchwork
git bisect good ab0c23b535f3f9d8345d8ad4c18c0a8594459d55
# bad: [f721d24e5dae8358b49b24399d27ba5d12a7e049] Merge tag 'pull-tmpfile' of git://git.kernel.org/pub/scm/linux/kernel/git/viro/vfs
git bisect bad f721d24e5dae8358b49b24399d27ba5d12a7e049
# good: [ada3bfb6492a6d0d3eca50f3b61315fe032efc72] Merge tag 'tpmdd-next-v6.1-rc1' of git://git.kernel.org/pub/scm/linux/kernel/git/jarkko/linux-tpmdd
git bisect good ada3bfb6492a6d0d3eca50f3b61315fe032efc72
# bad: [4e07acdda7fc23f5c4666e54961ef972a1195ffd] mm/hwpoison: add __init/__exit annotations to module init/exit funcs
git bisect bad 4e07acdda7fc23f5c4666e54961ef972a1195ffd
# bad: [000a449345bbb4ffbd880f7143b5fb4acac34121] radix tree test suite: add allocation counts and size to kmem_cache
git bisect bad 000a449345bbb4ffbd880f7143b5fb4acac34121
# bad: [47d55419951312d723de1b6ad53ee92948b8eace] btrfs: convert process_page_range() to use filemap_get_folios_contig()
git bisect bad 47d55419951312d723de1b6ad53ee92948b8eace
# bad: [4d86d4f7227c6f2acfbbbe0623d49865aa71b756] mm: add more BUILD_BUG_ONs to gfp_migratetype()
git bisect bad 4d86d4f7227c6f2acfbbbe0623d49865aa71b756
# bad: [816284a3d0e27828b5cc35f3cf539b0711939ce3] userfaultfd: update documentation to describe /dev/userfaultfd
git bisect bad 816284a3d0e27828b5cc35f3cf539b0711939ce3
# good: [be6667b0db97e10b2a6d57a906c2c8fd2b985e5e] selftests/vm: dedup hugepage allocation logic
git bisect good be6667b0db97e10b2a6d57a906c2c8fd2b985e5e
# bad: [2ace36f0f55777be8a871c370832527e1cd54b15] mm: memory-failure: cleanup try_to_split_thp_page()
git bisect bad 2ace36f0f55777be8a871c370832527e1cd54b15
# good: [9d0d946840075e0268f4f77fe39ba0f53e84c7c4] selftests/vm: add selftest to verify multi THP collapse
git bisect good 9d0d946840075e0268f4f77fe39ba0f53e84c7c4
# bad: [f35b5d7d676e59e401690b678cd3cfec5e785c23] mm: align larger anonymous mappings on THP boundaries
git bisect bad f35b5d7d676e59e401690b678cd3cfec5e785c23
# good: [7b5a0b664ebe2625965a0fdba2614c88c4b9bbc6] mm/page_ext: remove unused variable in offline_page_ext
git bisect good 7b5a0b664ebe2625965a0fdba2614c88c4b9bbc6
# first bad commit: [f35b5d7d676e59e401690b678cd3cfec5e785c23] mm: align larger anonymous mappings on THP boundaries

2022-10-20 05:26:02

by Huang, Ying

[permalink] [raw]
Subject: Re: [mm] f35b5d7d67: will-it-scale.per_process_ops -95.5% regression

Hi, Nathan,

Thanks for your information! That's valuable.

Nathan Chancellor <[email protected]> writes:

> Hi Ying,
>
> On Wed, Oct 19, 2022 at 10:05:50AM +0800, Huang, Ying wrote:
>> Hi, Yujie,
>>
>> > 32528 48% +147.6% 80547 38% numa-meminfo.node0.AnonHugePages
>> > 92821 23% +59.3% 147839 28% numa-meminfo.node0.AnonPages
>>
>> The Anon pages allocated are much more than the parent commit. This is
>> expected, because THP instead of normal page will be allocated for
>> aligned memory area.
>>
>> > 95.23 -79.8 15.41 6% perf-profile.calltrace.cycles-pp.__munmap
>> > 95.08 -79.7 15.40 6% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__munmap
>> > 95.02 -79.6 15.39 6% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
>> > 94.96 -79.6 15.37 6% perf-profile.calltrace.cycles-pp.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
>> > 94.95 -79.6 15.37 6% perf-profile.calltrace.cycles-pp.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
>> > 94.86 -79.5 15.35 6% perf-profile.calltrace.cycles-pp.__do_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe
>> > 94.38 -79.2 15.22 6% perf-profile.calltrace.cycles-pp.unmap_region.__do_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64
>> > 42.74 -42.7 0.00 perf-profile.calltrace.cycles-pp.lru_add_drain.unmap_region.__do_munmap.__vm_munmap.__x64_sys_munmap
>> > 42.74 -42.7 0.00 perf-profile.calltrace.cycles-pp.lru_add_drain_cpu.lru_add_drain.unmap_region.__do_munmap.__vm_munmap
>> > 42.72 -42.7 0.00 perf-profile.calltrace.cycles-pp.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain.unmap_region.__do_munmap
>> > 41.84 -41.8 0.00 perf-profile.calltrace.cycles-pp.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain.unmap_region
>> > 41.70 -41.7 0.00 perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain
>> > 41.62 -41.6 0.00 perf-profile.calltrace.cycles-pp.folio_lruvec_lock_irqsave.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region
>> > 41.55 -41.6 0.00
>> > perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu
>> > 41.52 -41.5 0.00 perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.release_pages.tlb_batch_pages_flush.tlb_finish_mmu
>> > 41.28 -41.3 0.00
>> > perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.release_pages.tlb_batch_pages_flush
>>
>> In the parent commit, most CPU cycles are used for contention on LRU lock.
>>
>> > 0.00 +4.8 4.82 7% perf-profile.calltrace.cycles-pp.do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
>> > 0.00 +4.9 4.88 7% perf-profile.calltrace.cycles-pp.zap_huge_pmd.zap_pmd_range.unmap_page_range.unmap_vmas.unmap_region
>> > 0.00 +8.2 8.22 8% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.rmqueue_bulk.rmqueue.get_page_from_freelist
>> > 0.00 +8.2 8.23 8% perf-profile.calltrace.cycles-pp._raw_spin_lock.rmqueue_bulk.rmqueue.get_page_from_freelist.__alloc_pages
>> > 0.00 +8.3 8.35 8% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.free_pcppages_bulk.free_unref_page.release_pages
>> > 0.00 +8.3 8.35 8% perf-profile.calltrace.cycles-pp._raw_spin_lock.free_pcppages_bulk.free_unref_page.release_pages.tlb_batch_pages_flush
>> > 0.00 +8.4 8.37 8% perf-profile.calltrace.cycles-pp.free_pcppages_bulk.free_unref_page.release_pages.tlb_batch_pages_flush.tlb_finish_mmu
>> > 0.00 +9.6 9.60 6% perf-profile.calltrace.cycles-pp.free_unref_page.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region
>> > 0.00 +65.5 65.48 2% perf-profile.calltrace.cycles-pp.clear_page_erms.clear_huge_page.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault
>> > 0.00 +72.5 72.51 2% perf-profile.calltrace.cycles-pp.clear_huge_page.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
>>
>> With the commit, most CPU cycles are consumed for clear huge page. This
>> is expected. We allocate more pages, so, we need more cycles to clear
>> them.
>>
>> Check the source code of test case (will-it-scale/malloc1), I found that
>> it will allocate some memory with malloc() then free it.
>>
>> In the parent commit, because the virtual memory address isn't aligned
>> with 2M, normal page will be allocated. With the commit, THP will be
>> allocated, so more page clearing and less LRU lock contention. I think
>> this is the expected behavior of the commit. And the test case isn't so
>> popular (malloc() then free() but don't access the memory allocated). So
>> this regression isn't important. We can just ignore it.
>
> For what it's worth, I just bisected a massive and visible performance
> regression on my Threadripper 3990X workstation to commit f35b5d7d676e
> ("mm: align larger anonymous mappings on THP boundaries"), which seems
> directly related to this report/analysis. I initially noticed this
> because my full set of kernel builds against mainline went from 2 hours
> and 20 minutes or so to over 3 hours. Zeroing in on x86_64 allmodconfig,
> which I used for the bisect:
>
> @ 7b5a0b664ebe ("mm/page_ext: remove unused variable in offline_page_ext"):
>
> Benchmark 1: make -skj128 LLVM=1 allmodconfig all
> Time (mean ± σ): 318.172 s ± 0.730 s [User: 31750.902 s, System: 4564.246 s]
> Range (min … max): 317.332 s … 318.662 s 3 runs
>
> @ f35b5d7d676e ("mm: align larger anonymous mappings on THP boundaries"):
>
> Benchmark 1: make -skj128 LLVM=1 allmodconfig all

Have you tried to build with gcc? Want to check whether is this clang
specific issue or not.

Best Regards,
Huang, Ying

> Time (mean ± σ): 406.688 s ± 0.676 s [User: 31819.526 s, System: 16327.022 s]
> Range (min … max): 405.954 s … 407.284 s 3 run
>
> That is a pretty big difference (27%), which is visible while doing a
> lot of builds, only because of the extra system time. If there is any
> way to improve this, it should certainly be considered.
>
> For now, I'll just revert it locally.
>
> Cheers,
> Nathan
>
> # bad: [aae703b02f92bde9264366c545e87cec451de471] Merge tag 'for-6.1-rc1-tag' of git://git.kernel.org/pub/scm/linux/kernel/git/kdave/linux
> # good: [4fe89d07dcc2804c8b562f6c7896a45643d34b2f] Linux 6.0
> git bisect start 'aae703b02f92bde9264366c545e87cec451de471' 'v6.0'
> # good: [18fd049731e67651009f316195da9281b756f2cf] Merge tag 'arm64-upstream' of git://git.kernel.org/pub/scm/linux/kernel/git/arm64/linux
> git bisect good 18fd049731e67651009f316195da9281b756f2cf
> # good: [ab0c23b535f3f9d8345d8ad4c18c0a8594459d55] MAINTAINERS: add RISC-V's patchwork
> git bisect good ab0c23b535f3f9d8345d8ad4c18c0a8594459d55
> # bad: [f721d24e5dae8358b49b24399d27ba5d12a7e049] Merge tag 'pull-tmpfile' of git://git.kernel.org/pub/scm/linux/kernel/git/viro/vfs
> git bisect bad f721d24e5dae8358b49b24399d27ba5d12a7e049
> # good: [ada3bfb6492a6d0d3eca50f3b61315fe032efc72] Merge tag 'tpmdd-next-v6.1-rc1' of git://git.kernel.org/pub/scm/linux/kernel/git/jarkko/linux-tpmdd
> git bisect good ada3bfb6492a6d0d3eca50f3b61315fe032efc72
> # bad: [4e07acdda7fc23f5c4666e54961ef972a1195ffd] mm/hwpoison: add __init/__exit annotations to module init/exit funcs
> git bisect bad 4e07acdda7fc23f5c4666e54961ef972a1195ffd
> # bad: [000a449345bbb4ffbd880f7143b5fb4acac34121] radix tree test suite: add allocation counts and size to kmem_cache
> git bisect bad 000a449345bbb4ffbd880f7143b5fb4acac34121
> # bad: [47d55419951312d723de1b6ad53ee92948b8eace] btrfs: convert process_page_range() to use filemap_get_folios_contig()
> git bisect bad 47d55419951312d723de1b6ad53ee92948b8eace
> # bad: [4d86d4f7227c6f2acfbbbe0623d49865aa71b756] mm: add more BUILD_BUG_ONs to gfp_migratetype()
> git bisect bad 4d86d4f7227c6f2acfbbbe0623d49865aa71b756
> # bad: [816284a3d0e27828b5cc35f3cf539b0711939ce3] userfaultfd: update documentation to describe /dev/userfaultfd
> git bisect bad 816284a3d0e27828b5cc35f3cf539b0711939ce3
> # good: [be6667b0db97e10b2a6d57a906c2c8fd2b985e5e] selftests/vm: dedup hugepage allocation logic
> git bisect good be6667b0db97e10b2a6d57a906c2c8fd2b985e5e
> # bad: [2ace36f0f55777be8a871c370832527e1cd54b15] mm: memory-failure: cleanup try_to_split_thp_page()
> git bisect bad 2ace36f0f55777be8a871c370832527e1cd54b15
> # good: [9d0d946840075e0268f4f77fe39ba0f53e84c7c4] selftests/vm: add selftest to verify multi THP collapse
> git bisect good 9d0d946840075e0268f4f77fe39ba0f53e84c7c4
> # bad: [f35b5d7d676e59e401690b678cd3cfec5e785c23] mm: align larger anonymous mappings on THP boundaries
> git bisect bad f35b5d7d676e59e401690b678cd3cfec5e785c23
> # good: [7b5a0b664ebe2625965a0fdba2614c88c4b9bbc6] mm/page_ext: remove unused variable in offline_page_ext
> git bisect good 7b5a0b664ebe2625965a0fdba2614c88c4b9bbc6
> # first bad commit: [f35b5d7d676e59e401690b678cd3cfec5e785c23] mm: align larger anonymous mappings on THP boundaries

2022-10-20 16:24:00

by Rik van Riel

[permalink] [raw]
Subject: Re: [mm] f35b5d7d67: will-it-scale.per_process_ops -95.5% regression

On Thu, 2022-10-20 at 13:07 +0800, Huang, Ying wrote:
>
> Nathan Chancellor <[email protected]> writes:
> >
> > For what it's worth, I just bisected a massive and visible
> > performance
> > regression on my Threadripper 3990X workstation to commit
> > f35b5d7d676e
> > ("mm: align larger anonymous mappings on THP boundaries"), which
> > seems
> > directly related to this report/analysis. I initially noticed this
> > because my full set of kernel builds against mainline went from 2
> > hours
> > and 20 minutes or so to over 3 hours. Zeroing in on x86_64
> > allmodconfig,
> > which I used for the bisect:
> >
> > @ 7b5a0b664ebe ("mm/page_ext: remove unused variable in
> > offline_page_ext"):
> >
> > Benchmark 1: make -skj128 LLVM=1 allmodconfig all
> >   Time (mean ± σ):     318.172 s ±  0.730 s    [User: 31750.902 s,
> > System: 4564.246 s]
> >   Range (min … max):   317.332 s … 318.662 s    3 runs
> >
> > @ f35b5d7d676e ("mm: align larger anonymous mappings on THP
> > boundaries"):
> >
> > Benchmark 1: make -skj128 LLVM=1 allmodconfig all
> > Time (mean ± σ): 406.688 s ± 0.676 s [User: 31819.526 s,
System: 16327.022 s]
> > Range (min … max): 405.954 s … 407.284 s 3 run
>
> Have you tried to build with gcc?  Want to check whether is this
> clang
> specific issue or not.

This may indeed be something LLVM specific. In previous tests,
GCC has generally seen a benefit from increased THP usage.
Many other applications also benefit from getting more THPs.

LLVM showing 10% system time before this change, and a whopping
30% system time after that change, suggests that LLVM is behaving
quite differently from GCC in some ways.

If we can figure out what these differences are, maybe we can
just fine tune the code to avoid this issue.

I'll try to play around with LLVM compilation a little bit next
week, to see if I can figure out what might be going on. I wonder
if LLVM is doing lots of mremap calls or something...

--
All Rights Reversed.


Attachments:
signature.asc (499.00 B)
This is a digitally signed message part

2022-10-20 16:52:55

by kernel test robot

[permalink] [raw]
Subject: Re: [mm] f35b5d7d67: will-it-scale.per_process_ops -95.5% regression

On Thu, Oct 20, 2022 at 01:07:09PM +0800, Huang, Ying wrote:
> Hi, Nathan,
>
> Thanks for your information! That's valuable.
>
> Nathan Chancellor <[email protected]> writes:
>
> > Hi Ying,
> >
> > On Wed, Oct 19, 2022 at 10:05:50AM +0800, Huang, Ying wrote:
> >> Hi, Yujie,
> >>
> >> > 32528 48% +147.6% 80547 38% numa-meminfo.node0.AnonHugePages
> >> > 92821 23% +59.3% 147839 28% numa-meminfo.node0.AnonPages
> >>
> >> The Anon pages allocated are much more than the parent commit. This is
> >> expected, because THP instead of normal page will be allocated for
> >> aligned memory area.
> >>
> >> > 95.23 -79.8 15.41 6% perf-profile.calltrace.cycles-pp.__munmap
> >> > 95.08 -79.7 15.40 6% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__munmap
> >> > 95.02 -79.6 15.39 6% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
> >> > 94.96 -79.6 15.37 6% perf-profile.calltrace.cycles-pp.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
> >> > 94.95 -79.6 15.37 6% perf-profile.calltrace.cycles-pp.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
> >> > 94.86 -79.5 15.35 6% perf-profile.calltrace.cycles-pp.__do_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe
> >> > 94.38 -79.2 15.22 6% perf-profile.calltrace.cycles-pp.unmap_region.__do_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64
> >> > 42.74 -42.7 0.00 perf-profile.calltrace.cycles-pp.lru_add_drain.unmap_region.__do_munmap.__vm_munmap.__x64_sys_munmap
> >> > 42.74 -42.7 0.00 perf-profile.calltrace.cycles-pp.lru_add_drain_cpu.lru_add_drain.unmap_region.__do_munmap.__vm_munmap
> >> > 42.72 -42.7 0.00 perf-profile.calltrace.cycles-pp.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain.unmap_region.__do_munmap
> >> > 41.84 -41.8 0.00 perf-profile.calltrace.cycles-pp.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain.unmap_region
> >> > 41.70 -41.7 0.00 perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain
> >> > 41.62 -41.6 0.00 perf-profile.calltrace.cycles-pp.folio_lruvec_lock_irqsave.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region
> >> > 41.55 -41.6 0.00
> >> > perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu
> >> > 41.52 -41.5 0.00 perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.release_pages.tlb_batch_pages_flush.tlb_finish_mmu
> >> > 41.28 -41.3 0.00
> >> > perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.release_pages.tlb_batch_pages_flush
> >>
> >> In the parent commit, most CPU cycles are used for contention on LRU lock.
> >>
> >> > 0.00 +4.8 4.82 7% perf-profile.calltrace.cycles-pp.do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
> >> > 0.00 +4.9 4.88 7% perf-profile.calltrace.cycles-pp.zap_huge_pmd.zap_pmd_range.unmap_page_range.unmap_vmas.unmap_region
> >> > 0.00 +8.2 8.22 8% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.rmqueue_bulk.rmqueue.get_page_from_freelist
> >> > 0.00 +8.2 8.23 8% perf-profile.calltrace.cycles-pp._raw_spin_lock.rmqueue_bulk.rmqueue.get_page_from_freelist.__alloc_pages
> >> > 0.00 +8.3 8.35 8% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.free_pcppages_bulk.free_unref_page.release_pages
> >> > 0.00 +8.3 8.35 8% perf-profile.calltrace.cycles-pp._raw_spin_lock.free_pcppages_bulk.free_unref_page.release_pages.tlb_batch_pages_flush
> >> > 0.00 +8.4 8.37 8% perf-profile.calltrace.cycles-pp.free_pcppages_bulk.free_unref_page.release_pages.tlb_batch_pages_flush.tlb_finish_mmu
> >> > 0.00 +9.6 9.60 6% perf-profile.calltrace.cycles-pp.free_unref_page.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region
> >> > 0.00 +65.5 65.48 2% perf-profile.calltrace.cycles-pp.clear_page_erms.clear_huge_page.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault
> >> > 0.00 +72.5 72.51 2% perf-profile.calltrace.cycles-pp.clear_huge_page.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
> >>
> >> With the commit, most CPU cycles are consumed for clear huge page. This
> >> is expected. We allocate more pages, so, we need more cycles to clear
> >> them.
> >>
> >> Check the source code of test case (will-it-scale/malloc1), I found that
> >> it will allocate some memory with malloc() then free it.
> >>
> >> In the parent commit, because the virtual memory address isn't aligned
> >> with 2M, normal page will be allocated. With the commit, THP will be
> >> allocated, so more page clearing and less LRU lock contention. I think
> >> this is the expected behavior of the commit. And the test case isn't so
> >> popular (malloc() then free() but don't access the memory allocated). So
> >> this regression isn't important. We can just ignore it.
> >
> > For what it's worth, I just bisected a massive and visible performance
> > regression on my Threadripper 3990X workstation to commit f35b5d7d676e
> > ("mm: align larger anonymous mappings on THP boundaries"), which seems
> > directly related to this report/analysis. I initially noticed this
> > because my full set of kernel builds against mainline went from 2 hours
> > and 20 minutes or so to over 3 hours. Zeroing in on x86_64 allmodconfig,
> > which I used for the bisect:
> >
> > @ 7b5a0b664ebe ("mm/page_ext: remove unused variable in offline_page_ext"):
> >
> > Benchmark 1: make -skj128 LLVM=1 allmodconfig all
> > Time (mean ± σ): 318.172 s ± 0.730 s [User: 31750.902 s, System: 4564.246 s]
> > Range (min … max): 317.332 s … 318.662 s 3 runs
> >
> > @ f35b5d7d676e ("mm: align larger anonymous mappings on THP boundaries"):
> >
> > Benchmark 1: make -skj128 LLVM=1 allmodconfig all
>
> Have you tried to build with gcc? Want to check whether is this clang
> specific issue or not.

We saw a 2.8% gain in kbuild time test with gcc. Please check below
test result for details:

=========================================================================================
build_kconfig/compiler/cpufreq_governor/kconfig/nr_task/rootfs/runtime/target/tbox_group/testcase/thp_defrag/thp_enabled:
allyesconfig/gcc-11/performance/x86_64-rhel-8.3/200%/debian-11.1-x86_64-20220510.cgz/600s/autoksyms_recursive/lkp-icl-2sp5/kbuild/always/always

commit:
7b5a0b664ebe ("mm/page_ext: remove unused variable in offline_page_ext")
f35b5d7d676e ("mm: align larger anonymous mappings on THP boundaries")

7b5a0b664ebe2625 f35b5d7d676e59e401690b678cd
---------------- ---------------------------
%stddev %change %stddev
\ | \
220.89 -2.8% 214.78 kbuild.buildtime_per_iteration
662.55 -2.7% 644.52 kbuild.time.elapsed_time
662.55 -2.7% 644.52 kbuild.time.elapsed_time.max
7918841 -2.3% 7740039 kbuild.time.involuntary_context_switches
55936 +6.4% 59540 kbuild.time.major_page_faults
2845254 +0.0% 2846257 kbuild.time.maximum_resident_set_size
1.534e+09 -43.2% 8.709e+08 kbuild.time.minor_page_faults
4096 +0.0% 4096 kbuild.time.page_size
7295 -2.1% 7145 kbuild.time.percent_of_cpu_this_job_got
8017 -29.0% 5690 kbuild.time.system_time
40321 +0.1% 40364 kbuild.time.user_time
8024178 +1.0% 8106814 kbuild.time.voluntary_context_switches
3.00 +0.0% 3.00 kbuild.workload
788.52 -1.6% 776.25 ± 2% dmesg.timestamp:last
429.90 -0.5% 427.75 kmsg.timestamp:last
3.464e+10 +0.2% 3.472e+10 cpuidle..time
71725422 -0.0% 71720918 cpuidle..usage
705.10 -2.5% 687.50 uptime.boot
39399 +0.3% 39504 uptime.idle
37.28 +0.3% 37.38 boot-time.boot
22.72 -0.3% 22.64 boot-time.dhcp
4254 +0.7% 4286 boot-time.idle
1.52 -0.0% 1.51 boot-time.smp_boot
40.87 +1.2 42.12 mpstat.cpu.all.idle%
1.96 ± 7% -0.0 1.95 ± 8% mpstat.cpu.all.irq%
0.19 ± 7% +0.0 0.19 ± 7% mpstat.cpu.all.soft%
9.32 -2.5 6.81 mpstat.cpu.all.sys%
47.65 +1.3 48.93 mpstat.cpu.all.usr%
0.00 -100.0% 0.00 numa-numastat.node0.interleave_hit
7.222e+08 -46.6% 3.857e+08 numa-numastat.node0.local_node
7.223e+08 -46.6% 3.858e+08 numa-numastat.node0.numa_hit
77202 ± 39% -64.3% 27589 ± 70% numa-numastat.node0.other_node
0.00 -100.0% 0.00 numa-numastat.node1.interleave_hit
7.104e+08 -45.5% 3.869e+08 numa-numastat.node1.local_node
7.105e+08 -45.5% 3.87e+08 numa-numastat.node1.numa_hit
38778 ± 78% +127.9% 88368 ± 21% numa-numastat.node1.other_node
662.55 -2.7% 644.52 time.elapsed_time
662.55 -2.7% 644.52 time.elapsed_time.max
7918841 -2.3% 7740039 time.involuntary_context_switches
55936 +6.4% 59540 time.major_page_faults
2845254 +0.0% 2846257 time.maximum_resident_set_size
1.534e+09 -43.2% 8.709e+08 time.minor_page_faults
4096 +0.0% 4096 time.page_size
7295 -2.1% 7145 time.percent_of_cpu_this_job_got
8017 -29.0% 5690 time.system_time
40321 +0.1% 40364 time.user_time
8024178 +1.0% 8106814 time.voluntary_context_switches
41.00 +2.4% 42.00 vmstat.cpu.id
11.00 -27.3% 8.00 vmstat.cpu.sy
47.00 +2.8% 48.33 vmstat.cpu.us
0.00 -100.0% 0.00 vmstat.io.bi
0.00 +6.7e+101% 0.67 ± 70% vmstat.io.bo
5.00 ± 28% +20.0% 6.00 ± 23% vmstat.memory.buff
6438472 +0.4% 6466124 vmstat.memory.cache
1.178e+08 -0.1% 1.177e+08 vmstat.memory.free
0.00 -100.0% 0.00 vmstat.procs.b
146.00 -2.7% 142.00 vmstat.procs.r
31058 +2.0% 31693 vmstat.system.cs
265989 -0.0% 265868 vmstat.system.in
1804 -6.4% 1689 turbostat.Avg_MHz
59.63 -1.2 58.48 turbostat.Busy%
3033 -4.6% 2895 turbostat.Bzy_MHz
56301 +178.2% 156603 ± 89% turbostat.C1
0.00 +0.1 0.11 ±134% turbostat.C1%
55119210 ± 25% -12.9% 48006244 ± 34% turbostat.C1E
26.32 ± 51% -8.8 17.49 ± 98% turbostat.C1E%
16462378 ± 88% +42.6% 23471160 ± 70% turbostat.C6
14.35 ± 94% +9.9 24.26 ± 70% turbostat.C6%
29.19 ± 48% -36.2% 18.62 ± 87% turbostat.CPU%c1
11.17 ±124% +105.0% 22.91 ± 70% turbostat.CPU%c6
60.33 ± 5% -1.7% 59.33 ± 5% turbostat.CoreTmp
0.25 +6.6% 0.27 turbostat.IPC
1.772e+08 -2.7% 1.725e+08 turbostat.IRQ
114.11 -2.7 111.37 turbostat.PKG_%
34096 ± 2% +4.6% 35653 ± 9% turbostat.POLL
3.22 ±141% +95.0% 6.29 ± 70% turbostat.Pkg%pc2
60.67 ± 4% -1.6% 59.67 ± 5% turbostat.PkgTmp
392.96 -2.3% 383.84 ± 2% turbostat.PkgWatt
78.55 +4.4% 81.98 turbostat.RAMWatt
2594 +0.0% 2594 turbostat.TSC_MHz
182786 -4.9% 173793 ± 2% meminfo.Active
182711 -5.0% 173527 ± 2% meminfo.Active(anon)
75.33 ± 21% +252.7% 265.67 ± 58% meminfo.Active(file)
414770 +904.5% 4166363 meminfo.AnonHugePages
5917906 +0.7% 5960021 meminfo.AnonPages
5.00 ± 28% +20.0% 6.00 ± 23% meminfo.Buffers
6178648 +0.4% 6202569 meminfo.Cached
65836880 +0.0% 65836880 meminfo.CommitLimit
6914586 -2.7% 6727090 meminfo.Committed_AS
1.265e+08 -0.6% 1.258e+08 meminfo.DirectMap1G
9768277 ± 14% +6.7% 10420224 ± 14% meminfo.DirectMap2M
746928 ± 15% +6.3% 794032 ± 3% meminfo.DirectMap4k
0.00 +2.7e+102% 2.67 ± 70% meminfo.Dirty
2048 +0.0% 2048 meminfo.Hugepagesize
5983334 +0.7% 6026071 meminfo.Inactive
5983186 +0.7% 6025838 meminfo.Inactive(anon)
147.33 ± 14% +57.9% 232.67 ± 27% meminfo.Inactive(file)
279848 +0.3% 280656 meminfo.KReclaimable
27804 -0.6% 27650 meminfo.KernelStack
110382 ± 2% -1.4% 108862 meminfo.Mapped
1.173e+08 -0.1% 1.172e+08 meminfo.MemAvailable
1.178e+08 -0.1% 1.177e+08 meminfo.MemFree
1.317e+08 +0.0% 1.317e+08 meminfo.MemTotal
13845459 +0.8% 13951748 meminfo.Memused
1266 -2.4% 1236 meminfo.Mlocked
36943 -1.7% 36301 meminfo.PageTables
69992 -0.2% 69887 meminfo.Percpu
279848 +0.3% 280656 meminfo.SReclaimable
305520 +0.4% 306812 meminfo.SUnreclaim
229589 ± 2% -3.3% 222046 meminfo.Shmem
585369 +0.4% 587468 meminfo.Slab
5948778 +0.5% 5979958 meminfo.Unevictable
1.374e+13 +0.0% 1.374e+13 meminfo.VmallocTotal
194761 -0.1% 194634 meminfo.VmallocUsed
19416429 +1.7% 19756098 meminfo.max_used_kB
6516 ± 18% -9.9% 5871 ± 14% numa-meminfo.node0.Active
6472 ± 19% -11.5% 5729 ± 14% numa-meminfo.node0.Active(anon)
43.67 ± 65% +222.9% 141.00 ± 94% numa-meminfo.node0.Active(file)
192829 ± 14% +960.5% 2044918 ± 2% numa-meminfo.node0.AnonHugePages
2946303 -0.7% 2926031 numa-meminfo.node0.AnonPages
6975049 ± 4% +1.9% 7107616 ± 2% numa-meminfo.node0.AnonPages.max
0.00 +1.3e+102% 1.33 ±141% numa-meminfo.node0.Dirty
2479087 ± 77% +46.3% 3627935 ± 42% numa-meminfo.node0.FilePages
2958399 -0.7% 2937077 numa-meminfo.node0.Inactive
2958313 -0.7% 2936942 numa-meminfo.node0.Inactive(anon)
86.67 ± 70% +55.4% 134.67 ± 45% numa-meminfo.node0.Inactive(file)
120799 ± 44% +33.7% 161521 ± 33% numa-meminfo.node0.KReclaimable
13904 ± 2% -2.1% 13607 ± 2% numa-meminfo.node0.KernelStack
36268 ± 64% +39.4% 50562 ± 42% numa-meminfo.node0.Mapped
59414836 ± 3% -2.0% 58226661 ± 2% numa-meminfo.node0.MemFree
65663956 +0.0% 65663956 numa-meminfo.node0.MemTotal
6249119 ± 30% +19.0% 7437293 ± 18% numa-meminfo.node0.MemUsed
545.00 ± 93% +51.4% 825.33 ± 70% numa-meminfo.node0.Mlocked
18693 -5.8% 17617 numa-meminfo.node0.PageTables
120799 ± 44% +33.7% 161521 ± 33% numa-meminfo.node0.SReclaimable
155553 ± 7% -0.0% 155524 ± 5% numa-meminfo.node0.SUnreclaim
9569 ± 17% -21.4% 7517 ± 11% numa-meminfo.node0.Shmem
276353 ± 19% +14.7% 317046 ± 15% numa-meminfo.node0.Slab
2469448 ± 77% +46.6% 3620137 ± 42% numa-meminfo.node0.Unevictable
176559 -4.7% 168248 ± 3% numa-meminfo.node1.Active
176527 -4.8% 168123 ± 2% numa-meminfo.node1.Active(anon)
31.33 ±141% +297.9% 124.67 ±109% numa-meminfo.node1.Active(file)
222472 ± 11% +840.3% 2092001 numa-meminfo.node1.AnonHugePages
2940819 +1.9% 2996244 numa-meminfo.node1.AnonPages
6775470 ± 2% +13.9% 7719184 ± 4% numa-meminfo.node1.AnonPages.max
0.00 +1.3e+102% 1.33 ±141% numa-meminfo.node1.Dirty
3703809 ± 51% -30.3% 2582256 ± 59% numa-meminfo.node1.FilePages
2994213 ± 2% +1.9% 3052599 numa-meminfo.node1.Inactive
2994152 ± 2% +1.9% 3052501 numa-meminfo.node1.Inactive(anon)
60.67 ±136% +61.0% 97.67 ±109% numa-meminfo.node1.Inactive(file)
159156 ± 33% -25.0% 119290 ± 45% numa-meminfo.node1.KReclaimable
13840 ± 2% +1.2% 14002 ± 2% numa-meminfo.node1.KernelStack
74373 ± 32% -20.2% 59373 ± 35% numa-meminfo.node1.Mapped
58441734 ± 3% +1.9% 59526117 ± 2% numa-meminfo.node1.MemFree
66009808 +0.0% 66009808 numa-meminfo.node1.MemTotal
7568072 ± 25% -14.3% 6483689 ± 22% numa-meminfo.node1.MemUsed
722.67 ± 70% -43.0% 412.00 ±141% numa-meminfo.node1.Mlocked
18101 ± 2% +2.1% 18485 numa-meminfo.node1.PageTables
159156 ± 33% -25.0% 119290 ± 45% numa-meminfo.node1.SReclaimable
149871 ± 7% +1.0% 151303 ± 4% numa-meminfo.node1.SUnreclaim
220726 -2.4% 215489 ± 2% numa-meminfo.node1.Shmem
309028 ± 17% -12.4% 270594 ± 18% numa-meminfo.node1.Slab
3482874 ± 54% -32.1% 2366476 ± 64% numa-meminfo.node1.Unevictable
339.67 ± 16% +7.0% 363.33 ± 3% proc-vmstat.direct_map_level2_splits
2.33 ± 53% +28.6% 3.00 ± 47% proc-vmstat.direct_map_level3_splits
45656 -4.9% 43403 ± 2% proc-vmstat.nr_active_anon
18.67 ± 20% +255.4% 66.33 ± 58% proc-vmstat.nr_active_file
1476514 +0.6% 1485684 proc-vmstat.nr_anon_pages
202.00 +903.5% 2027 proc-vmstat.nr_anon_transparent_hugepages
5.33 ±141% +2125.0% 118.67 ± 70% proc-vmstat.nr_dirtied
0.00 +6.7e+101% 0.67 ± 70% proc-vmstat.nr_dirty
2926273 -0.1% 2923659 proc-vmstat.nr_dirty_background_threshold
5859703 -0.1% 5854468 proc-vmstat.nr_dirty_threshold
1544951 +0.4% 1551178 proc-vmstat.nr_file_pages
29460263 -0.1% 29434012 proc-vmstat.nr_free_pages
1492541 +0.7% 1502535 proc-vmstat.nr_inactive_anon
36.67 ± 14% +58.2% 58.00 ± 27% proc-vmstat.nr_inactive_file
0.00 +3.3e+101% 0.33 ±141% proc-vmstat.nr_isolated_anon
27788 -0.5% 27654 proc-vmstat.nr_kernel_stack
27584 -0.5% 27458 proc-vmstat.nr_mapped
316.33 -2.5% 308.33 proc-vmstat.nr_mlock
9217 -1.6% 9067 proc-vmstat.nr_page_table_pages
57434 ± 2% -3.2% 55599 proc-vmstat.nr_shmem
69964 +0.3% 70176 proc-vmstat.nr_slab_reclaimable
76352 +0.5% 76758 proc-vmstat.nr_slab_unreclaimable
1487442 +0.5% 1495431 proc-vmstat.nr_unevictable
0.00 -100.0% 0.00 proc-vmstat.nr_writeback
6.67 ±141% +1680.0% 118.67 ± 70% proc-vmstat.nr_written
45656 -4.9% 43403 ± 2% proc-vmstat.nr_zone_active_anon
18.67 ± 20% +255.4% 66.33 ± 58% proc-vmstat.nr_zone_active_file
1492541 +0.7% 1502535 proc-vmstat.nr_zone_inactive_anon
36.67 ± 14% +58.2% 58.00 ± 27% proc-vmstat.nr_zone_inactive_file
1487442 +0.5% 1495431 proc-vmstat.nr_zone_unevictable
0.00 +6.7e+101% 0.67 ± 70% proc-vmstat.nr_zone_write_pending
664037 ± 15% +46.5% 972834 ± 33% proc-vmstat.numa_hint_faults
524692 ± 15% +39.8% 733659 ± 37% proc-vmstat.numa_hint_faults_local
1.433e+09 -46.1% 7.728e+08 proc-vmstat.numa_hit
9265 ± 4% +912.7% 93825 proc-vmstat.numa_huge_pte_updates
0.00 -100.0% 0.00 proc-vmstat.numa_interleave
1.433e+09 -46.1% 7.726e+08 proc-vmstat.numa_local
115980 -0.0% 115957 proc-vmstat.numa_other
222139 ± 49% +207.6% 683356 ± 6% proc-vmstat.numa_pages_migrated
6070902 ± 6% +711.1% 49242299 proc-vmstat.numa_pte_updates
1657653 -0.8% 1645015 proc-vmstat.pgactivate
0.00 -100.0% 0.00 proc-vmstat.pgalloc_dma32
1.432e+09 -46.1% 7.726e+08 proc-vmstat.pgalloc_normal
1.537e+09 -43.1% 8.745e+08 proc-vmstat.pgfault
1.43e+09 -46.1% 7.709e+08 proc-vmstat.pgfree
222139 ± 49% +207.6% 683356 ± 6% proc-vmstat.pgmigrate_success
0.00 +2.1e+104% 208.00 ± 70% proc-vmstat.pgpgin
29.33 ±141% +1554.5% 485.33 ± 70% proc-vmstat.pgpgout
4728028 -0.0% 4726238 proc-vmstat.pgreuse
240.67 ± 6% -6.9% 224.00 ± 5% proc-vmstat.thp_collapse_alloc
168.67 -92.9% 12.00 proc-vmstat.thp_deferred_split_page
40677 +3313.6% 1388568 proc-vmstat.thp_fault_alloc
164.33 ± 77% +436.5% 881.67 ± 13% proc-vmstat.thp_migration_success
1370 -11.6% 1211 proc-vmstat.thp_split_pmd
0.67 ± 70% -50.0% 0.33 ±141% proc-vmstat.thp_zero_page_alloc
5512262 -0.0% 5512262 proc-vmstat.unevictable_pgs_culled
608.00 -0.2% 607.00 proc-vmstat.unevictable_pgs_mlocked
4.00 ± 20% -25.0% 3.00 proc-vmstat.unevictable_pgs_munlocked
4.00 ± 20% -25.0% 3.00 proc-vmstat.unevictable_pgs_rescued
1628 ± 18% -11.6% 1439 ± 13% numa-vmstat.node0.nr_active_anon
10.67 ± 70% +228.1% 35.00 ± 94% numa-vmstat.node0.nr_active_file
734608 -0.3% 732747 numa-vmstat.node0.nr_anon_pages
93.67 ± 13% +966.5% 999.00 ± 2% numa-vmstat.node0.nr_anon_transparent_hugepages
0.00 +5.9e+103% 58.67 ±113% numa-vmstat.node0.nr_dirtied
0.00 +3.3e+101% 0.33 ±141% numa-vmstat.node0.nr_dirty
619437 ± 76% +46.4% 906779 ± 42% numa-vmstat.node0.nr_file_pages
14855848 ± 3% -2.0% 14555595 ± 2% numa-vmstat.node0.nr_free_pages
737673 -0.3% 735382 numa-vmstat.node0.nr_inactive_anon
21.67 ± 70% +53.8% 33.33 ± 46% numa-vmstat.node0.nr_inactive_file
0.00 +3.3e+101% 0.33 ±141% numa-vmstat.node0.nr_isolated_anon
13888 ± 2% -2.2% 13587 ± 2% numa-vmstat.node0.nr_kernel_stack
9041 ± 64% +40.2% 12677 ± 42% numa-vmstat.node0.nr_mapped
135.67 ± 93% +52.3% 206.67 ± 70% numa-vmstat.node0.nr_mlock
4658 -5.7% 4392 numa-vmstat.node0.nr_page_table_pages
2395 ± 17% -21.3% 1886 ± 10% numa-vmstat.node0.nr_shmem
30203 ± 44% +33.7% 40383 ± 33% numa-vmstat.node0.nr_slab_reclaimable
38902 ± 7% +0.0% 38911 ± 5% numa-vmstat.node0.nr_slab_unreclaimable
617024 ± 77% +46.6% 904823 ± 42% numa-vmstat.node0.nr_unevictable
0.00 +5.9e+103% 58.67 ±113% numa-vmstat.node0.nr_written
1629 ± 18% -11.7% 1439 ± 13% numa-vmstat.node0.nr_zone_active_anon
10.67 ± 70% +228.1% 35.00 ± 94% numa-vmstat.node0.nr_zone_active_file
737675 -0.3% 735362 numa-vmstat.node0.nr_zone_inactive_anon
21.67 ± 70% +53.8% 33.33 ± 46% numa-vmstat.node0.nr_zone_inactive_file
617023 ± 77% +46.6% 904823 ± 42% numa-vmstat.node0.nr_zone_unevictable
0.00 +3.3e+101% 0.33 ±141% numa-vmstat.node0.nr_zone_write_pending
7.223e+08 -46.6% 3.858e+08 numa-vmstat.node0.numa_hit
0.00 -100.0% 0.00 numa-vmstat.node0.numa_interleave
7.222e+08 -46.6% 3.857e+08 numa-vmstat.node0.numa_local
77202 ± 39% -64.3% 27589 ± 70% numa-vmstat.node0.numa_other
44125 -4.8% 42015 ± 3% numa-vmstat.node1.nr_active_anon
7.67 ±141% +300.0% 30.67 ±110% numa-vmstat.node1.nr_active_file
733960 +2.1% 749048 numa-vmstat.node1.nr_anon_pages
108.00 ± 12% +844.1% 1019 numa-vmstat.node1.nr_anon_transparent_hugepages
5.33 ±141% +1025.0% 60.00 ±114% numa-vmstat.node1.nr_dirtied
0.00 +3.3e+101% 0.33 ±141% numa-vmstat.node1.nr_dirty
926708 ± 51% -30.3% 645773 ± 59% numa-vmstat.node1.nr_file_pages
14610869 ± 3% +1.9% 14881195 ± 2% numa-vmstat.node1.nr_free_pages
747090 +2.2% 763248 numa-vmstat.node1.nr_inactive_anon
15.00 ±136% +62.2% 24.33 ±110% numa-vmstat.node1.nr_inactive_file
0.00 -100.0% 0.00 numa-vmstat.node1.nr_isolated_anon
13862 ± 2% +1.0% 14005 numa-vmstat.node1.nr_kernel_stack
18536 ± 32% -20.1% 14813 ± 36% numa-vmstat.node1.nr_mapped
180.67 ± 70% -43.0% 103.00 ±141% numa-vmstat.node1.nr_mlock
4525 ± 2% +2.0% 4617 numa-vmstat.node1.nr_page_table_pages
55065 -2.0% 53957 ± 2% numa-vmstat.node1.nr_shmem
39795 ± 33% -25.1% 29825 ± 45% numa-vmstat.node1.nr_slab_reclaimable
37487 ± 7% +1.0% 37863 ± 4% numa-vmstat.node1.nr_slab_unreclaimable
871589 ± 54% -32.1% 591742 ± 64% numa-vmstat.node1.nr_unevictable
6.67 ±141% +800.0% 60.00 ±114% numa-vmstat.node1.nr_written
44125 -4.8% 42014 ± 3% numa-vmstat.node1.nr_zone_active_anon
7.67 ±141% +300.0% 30.67 ±110% numa-vmstat.node1.nr_zone_active_file
747087 +2.2% 763250 numa-vmstat.node1.nr_zone_inactive_anon
15.00 ±136% +62.2% 24.33 ±110% numa-vmstat.node1.nr_zone_inactive_file
871589 ± 54% -32.1% 591742 ± 64% numa-vmstat.node1.nr_zone_unevictable
0.00 +3.3e+101% 0.33 ±141% numa-vmstat.node1.nr_zone_write_pending
7.105e+08 -45.5% 3.87e+08 numa-vmstat.node1.numa_hit
0.00 -100.0% 0.00 numa-vmstat.node1.numa_interleave
7.104e+08 -45.5% 3.869e+08 numa-vmstat.node1.numa_local
38778 ± 78% +127.9% 88368 ± 21% numa-vmstat.node1.numa_other
11.03 ± 26% +19.4% 13.17 ± 23% perf-stat.i.MPKI
3.859e+10 +0.8% 3.889e+10 perf-stat.i.branch-instructions
2.44 ± 13% +0.2 2.68 ± 13% perf-stat.i.branch-miss-rate%
9.766e+08 +2.9% 1.005e+09 perf-stat.i.branch-misses
22.01 ± 11% -1.1 20.93 ± 13% perf-stat.i.cache-miss-rate%
4.328e+08 +6.7% 4.617e+08 perf-stat.i.cache-misses
1.639e+09 +4.4% 1.71e+09 perf-stat.i.cache-references
30948 +2.4% 31684 perf-stat.i.context-switches
1.28 ± 6% +0.3% 1.29 ± 6% perf-stat.i.cpi
128012 -0.0% 128008 perf-stat.i.cpu-clock
2.26e+11 -6.3% 2.119e+11 perf-stat.i.cpu-cycles
2554 +3.3% 2638 perf-stat.i.cpu-migrations
1097 ± 12% +4.8% 1150 ± 9% perf-stat.i.cycles-between-cache-misses
0.18 ± 23% -0.0 0.15 ± 31% perf-stat.i.dTLB-load-miss-rate%
60757842 -57.8% 25667706 ± 3% perf-stat.i.dTLB-load-misses
4.847e+10 +0.7% 4.879e+10 perf-stat.i.dTLB-loads
0.13 ± 10% -0.0 0.09 ± 16% perf-stat.i.dTLB-store-miss-rate%
28490589 -48.3% 14736435 perf-stat.i.dTLB-store-misses
2.176e+10 +0.1% 2.179e+10 perf-stat.i.dTLB-stores
1.756e+11 +0.6% 1.766e+11 perf-stat.i.instructions
0.83 ± 6% -0.1% 0.83 ± 6% perf-stat.i.ipc
83.76 +9.6% 91.77 perf-stat.i.major-faults
1.77 -6.3% 1.66 perf-stat.i.metric.GHz
497.86 ± 12% -32.9% 334.04 ± 16% perf-stat.i.metric.K/sec
864.66 +0.7% 870.45 perf-stat.i.metric.M/sec
2284606 -41.7% 1331208 perf-stat.i.minor-faults
15.63 ± 2% +0.7 16.37 perf-stat.i.node-load-miss-rate%
8958056 +0.8% 9026272 perf-stat.i.node-load-misses
1.457e+08 -0.5% 1.45e+08 perf-stat.i.node-loads
4.67 ± 17% +0.1 4.74 ± 16% perf-stat.i.node-store-miss-rate%
1889607 -1.9% 1852810 perf-stat.i.node-store-misses
1.043e+08 +9.3% 1.14e+08 perf-stat.i.node-stores
2284690 -41.7% 1331300 perf-stat.i.page-faults
128012 -0.0% 128008 perf-stat.i.task-clock
9.34 +3.7% 9.68 perf-stat.overall.MPKI
2.53 +0.1 2.58 perf-stat.overall.branch-miss-rate%
26.42 +0.6 27.02 perf-stat.overall.cache-miss-rate%
1.29 -6.8% 1.20 perf-stat.overall.cpi
522.01 -12.1% 458.69 perf-stat.overall.cycles-between-cache-misses
0.13 -0.1 0.05 ± 3% perf-stat.overall.dTLB-load-miss-rate%
0.13 -0.1 0.07 perf-stat.overall.dTLB-store-miss-rate%
0.78 +7.3% 0.83 perf-stat.overall.ipc
5.78 +0.1 5.89 perf-stat.overall.node-load-miss-rate%
1.78 -0.2 1.60 perf-stat.overall.node-store-miss-rate%
3.918e+13 -2.2% 3.832e+13 perf-stat.overall.path-length
3.892e+10 +0.6% 3.915e+10 perf-stat.ps.branch-instructions
9.85e+08 +2.7% 1.012e+09 perf-stat.ps.branch-misses
4.367e+08 +6.5% 4.649e+08 perf-stat.ps.cache-misses
1.653e+09 +4.1% 1.721e+09 perf-stat.ps.cache-references
31105 +2.1% 31746 perf-stat.ps.context-switches
127807 -0.0% 127801 perf-stat.ps.cpu-clock
2.28e+11 -6.5% 2.133e+11 perf-stat.ps.cpu-cycles
2573 +3.2% 2654 perf-stat.ps.cpu-migrations
61292622 -57.9% 25818880 ± 3% perf-stat.ps.dTLB-load-misses
4.888e+10 +0.5% 4.911e+10 perf-stat.ps.dTLB-loads
28723007 -48.4% 14823680 perf-stat.ps.dTLB-store-misses
2.195e+10 -0.1% 2.193e+10 perf-stat.ps.dTLB-stores
1.771e+11 +0.4% 1.777e+11 perf-stat.ps.instructions
84.44 +9.3% 92.27 perf-stat.ps.major-faults
2302575 -41.9% 1338900 perf-stat.ps.minor-faults
9029295 +1.2% 9141253 perf-stat.ps.node-load-misses
1.471e+08 -0.7% 1.46e+08 perf-stat.ps.node-loads
1901728 -1.6% 1870808 perf-stat.ps.node-store-misses
1.052e+08 +9.1% 1.148e+08 perf-stat.ps.node-stores
2302659 -41.9% 1338992 perf-stat.ps.page-faults
127807 -0.0% 127801 perf-stat.ps.task-clock
1.175e+14 -2.2% 1.149e+14 perf-stat.total.instructions
58777 ± 24% -22.5% 45566 ± 53% sched_debug.cfs_rq:/.MIN_vruntime.avg
7072999 ± 32% -17.5% 5832509 ± 53% sched_debug.cfs_rq:/.MIN_vruntime.max
0.00 +0.0% 0.00 sched_debug.cfs_rq:/.MIN_vruntime.min
638221 ± 29% -19.5% 513508 ± 53% sched_debug.cfs_rq:/.MIN_vruntime.stddev
1.08 ± 3% +19.3% 1.29 ± 4% sched_debug.cfs_rq:/.h_nr_running.avg
2.35 ± 5% +26.5% 2.97 ± 4% sched_debug.cfs_rq:/.h_nr_running.max
0.52 ± 4% +10.0% 0.57 ± 8% sched_debug.cfs_rq:/.h_nr_running.min
0.46 +15.7% 0.53 sched_debug.cfs_rq:/.h_nr_running.stddev
18158 ± 5% -21.7% 14209 ± 11% sched_debug.cfs_rq:/.load.avg
542665 ± 13% -35.5% 349767 ± 56% sched_debug.cfs_rq:/.load.max
1730 ± 5% +15.4% 1996 ± 10% sched_debug.cfs_rq:/.load.min
71133 ± 7% -36.5% 45198 ± 38% sched_debug.cfs_rq:/.load.stddev
27.26 ± 17% -7.2% 25.30 ± 15% sched_debug.cfs_rq:/.load_avg.avg
668.31 ± 9% +0.6% 672.13 ± 15% sched_debug.cfs_rq:/.load_avg.max
1.72 ± 5% +26.2% 2.17 ± 9% sched_debug.cfs_rq:/.load_avg.min
101.88 ± 13% -5.6% 96.17 ± 18% sched_debug.cfs_rq:/.load_avg.stddev
58777 ± 24% -22.5% 45566 ± 53% sched_debug.cfs_rq:/.max_vruntime.avg
7072999 ± 32% -17.5% 5832510 ± 53% sched_debug.cfs_rq:/.max_vruntime.max
0.00 +0.0% 0.00 sched_debug.cfs_rq:/.max_vruntime.min
638221 ± 29% -19.5% 513508 ± 53% sched_debug.cfs_rq:/.max_vruntime.stddev
30192748 ± 3% -15.4% 25535393 sched_debug.cfs_rq:/.min_vruntime.avg
32157892 ± 4% -15.9% 27049639 sched_debug.cfs_rq:/.min_vruntime.max
28428586 ± 3% -15.4% 24036465 sched_debug.cfs_rq:/.min_vruntime.min
459798 ± 12% -16.4% 384573 ± 5% sched_debug.cfs_rq:/.min_vruntime.stddev
0.55 ± 3% +28.8% 0.71 ± 2% sched_debug.cfs_rq:/.nr_running.avg
1.35 ± 6% -8.4% 1.23 ± 10% sched_debug.cfs_rq:/.nr_running.max
0.52 ± 4% +10.0% 0.57 ± 8% sched_debug.cfs_rq:/.nr_running.min
0.15 -15.6% 0.13 ± 19% sched_debug.cfs_rq:/.nr_running.stddev
11.65 ± 35% -14.4% 9.97 ± 34% sched_debug.cfs_rq:/.removed.load_avg.avg
490.45 ± 18% +14.7% 562.37 ± 19% sched_debug.cfs_rq:/.removed.load_avg.max
68.74 ± 21% +2.1% 70.18 ± 26% sched_debug.cfs_rq:/.removed.load_avg.stddev
4.93 ± 31% -22.7% 3.81 ± 21% sched_debug.cfs_rq:/.removed.runnable_avg.avg
218.51 ± 26% +8.2% 236.53 ± 21% sched_debug.cfs_rq:/.removed.runnable_avg.max
29.13 ± 17% -6.1% 27.35 ± 19% sched_debug.cfs_rq:/.removed.runnable_avg.stddev
3.97 ± 36% -24.4% 3.00 ± 23% sched_debug.cfs_rq:/.removed.util_avg.avg
150.56 ± 29% +17.2% 176.50 ± 18% sched_debug.cfs_rq:/.removed.util_avg.max
21.53 ± 17% -2.9% 20.91 ± 16% sched_debug.cfs_rq:/.removed.util_avg.stddev
929.10 ± 4% +19.9% 1114 ± 4% sched_debug.cfs_rq:/.runnable_avg.avg
2098 ± 2% +12.5% 2360 ± 9% sched_debug.cfs_rq:/.runnable_avg.max
416.51 ± 11% +5.3% 438.47 ± 11% sched_debug.cfs_rq:/.runnable_avg.min
348.16 +12.8% 392.78 ± 5% sched_debug.cfs_rq:/.runnable_avg.stddev
0.01 ± 72% -44.8% 0.00 ±141% sched_debug.cfs_rq:/.spread.avg
0.78 ± 72% -44.8% 0.43 ±141% sched_debug.cfs_rq:/.spread.max
0.07 ± 72% -44.8% 0.04 ±141% sched_debug.cfs_rq:/.spread.stddev
-1923853 -22.5% -1490838 sched_debug.cfs_rq:/.spread0.avg
44889 ±370% -42.6% 25746 ±164% sched_debug.cfs_rq:/.spread0.max
-3688946 -18.9% -2991064 sched_debug.cfs_rq:/.spread0.min
459603 ± 12% -16.4% 384061 ± 5% sched_debug.cfs_rq:/.spread0.stddev
537.53 ± 3% +29.1% 694.09 ± 2% sched_debug.cfs_rq:/.util_avg.avg
1258 ± 5% +13.6% 1429 ± 5% sched_debug.cfs_rq:/.util_avg.max
210.10 ± 16% +3.8% 218.10 ± 2% sched_debug.cfs_rq:/.util_avg.min
160.50 ± 3% +17.7% 188.96 sched_debug.cfs_rq:/.util_avg.stddev
17.02 ± 26% +226.0% 55.50 ± 46% sched_debug.cfs_rq:/.util_est_enqueued.avg
555.21 ± 6% +31.4% 729.80 ± 18% sched_debug.cfs_rq:/.util_est_enqueued.max
72.57 ± 6% +64.0% 119.00 ± 13% sched_debug.cfs_rq:/.util_est_enqueued.stddev
750029 -7.7% 692057 ± 2% sched_debug.cpu.avg_idle.avg
1085988 ± 5% +0.3% 1088834 ± 4% sched_debug.cpu.avg_idle.max
104257 ± 29% +28.7% 134136 ± 25% sched_debug.cpu.avg_idle.min
190580 ± 5% +11.7% 212828 sched_debug.cpu.avg_idle.stddev
361058 ± 4% -14.0% 310652 sched_debug.cpu.clock.avg
361088 ± 4% -14.0% 310679 sched_debug.cpu.clock.max
361026 ± 4% -14.0% 310622 sched_debug.cpu.clock.min
18.10 ± 11% -6.3% 16.96 ± 8% sched_debug.cpu.clock.stddev
353859 ± 3% -13.9% 304664 sched_debug.cpu.clock_task.avg
354440 ± 3% -13.9% 305329 sched_debug.cpu.clock_task.max
347598 ± 4% -14.3% 297852 sched_debug.cpu.clock_task.min
670.49 ± 4% +16.2% 779.29 ± 13% sched_debug.cpu.clock_task.stddev
440487 ± 4% +41.1% 621446 ± 3% sched_debug.cpu.curr->pid.avg
922331 ± 3% -11.8% 813459 sched_debug.cpu.curr->pid.max
326811 ± 30% +33.5% 436169 ± 24% sched_debug.cpu.curr->pid.min
76483 ± 19% -20.7% 60619 ± 52% sched_debug.cpu.curr->pid.stddev
501232 +0.1% 501595 sched_debug.cpu.max_idle_balance_cost.avg
542187 ± 4% +0.5% 544947 ± 5% sched_debug.cpu.max_idle_balance_cost.max
500000 +0.0% 500000 sched_debug.cpu.max_idle_balance_cost.min
5329 ± 42% +18.5% 6316 ± 41% sched_debug.cpu.max_idle_balance_cost.stddev
4295 -0.0% 4294 sched_debug.cpu.next_balance.avg
4295 -0.0% 4294 sched_debug.cpu.next_balance.max
4295 -0.0% 4294 sched_debug.cpu.next_balance.min
0.00 ± 8% +45.5% 0.00 ± 29% sched_debug.cpu.next_balance.stddev
1.06 ± 3% +20.3% 1.28 ± 4% sched_debug.cpu.nr_running.avg
2.38 ± 6% +24.8% 2.97 ± 4% sched_debug.cpu.nr_running.max
0.52 ± 4% +10.0% 0.57 ± 8% sched_debug.cpu.nr_running.min
0.42 +18.8% 0.50 ± 2% sched_debug.cpu.nr_running.stddev
87519 ± 3% -12.1% 76971 sched_debug.cpu.nr_switches.avg
110370 ± 5% -9.3% 100152 ± 4% sched_debug.cpu.nr_switches.max
77628 ± 4% -12.3% 68061 sched_debug.cpu.nr_switches.min
5244 +1.0% 5296 ± 8% sched_debug.cpu.nr_switches.stddev
2.076e+09 ± 2% +1.2% 2.102e+09 sched_debug.cpu.nr_uninterruptible.avg
4.295e+09 +0.0% 4.295e+09 sched_debug.cpu.nr_uninterruptible.max
0.88 ± 44% -47.0% 0.47 ± 66% sched_debug.cpu.nr_uninterruptible.min
2.144e+09 +0.0% 2.144e+09 sched_debug.cpu.nr_uninterruptible.stddev
361026 ± 4% -14.0% 310622 sched_debug.cpu_clk
996147 +0.0% 996147 sched_debug.dl_rq:.dl_bw->bw.avg
996147 +0.0% 996147 sched_debug.dl_rq:.dl_bw->bw.max
996147 +0.0% 996147 sched_debug.dl_rq:.dl_bw->bw.min
4.295e+09 -0.0% 4.295e+09 sched_debug.jiffies
360299 ± 4% -14.0% 309897 sched_debug.ktime
950.00 +0.0% 950.00 sched_debug.rt_rq:/.rt_runtime.avg
950.00 +0.0% 950.00 sched_debug.rt_rq:/.rt_runtime.max
950.00 +0.0% 950.00 sched_debug.rt_rq:/.rt_runtime.min
361614 ± 4% -13.9% 311213 sched_debug.sched_clk
1.00 +0.0% 1.00 sched_debug.sched_clock_stable()
58611259 +0.0% 58611259 sched_debug.sysctl_sched.sysctl_sched_features
0.75 +0.0% 0.75 sched_debug.sysctl_sched.sysctl_sched_idle_min_granularity
24.00 +0.0% 24.00 sched_debug.sysctl_sched.sysctl_sched_latency
3.00 +0.0% 3.00 sched_debug.sysctl_sched.sysctl_sched_min_granularity
1.00 +0.0% 1.00 sched_debug.sysctl_sched.sysctl_sched_tunable_scaling
4.00 +0.0% 4.00 sched_debug.sysctl_sched.sysctl_sched_wakeup_granularity
494099 -0.0% 494074 slabinfo.Acpi-Operand.active_objs
8828 -0.0% 8826 slabinfo.Acpi-Operand.active_slabs
494405 -0.0% 494293 slabinfo.Acpi-Operand.num_objs
8828 -0.0% 8826 slabinfo.Acpi-Operand.num_slabs
5579 ± 3% -2.9% 5416 slabinfo.Acpi-Parse.active_objs
76.00 ± 4% -3.1% 73.67 slabinfo.Acpi-Parse.active_slabs
5579 ± 3% -2.9% 5416 slabinfo.Acpi-Parse.num_objs
76.00 ± 4% -3.1% 73.67 slabinfo.Acpi-Parse.num_slabs
1376 ± 13% +5.2% 1448 ± 4% slabinfo.Acpi-State.active_objs
26.67 ± 14% +5.0% 28.00 ± 5% slabinfo.Acpi-State.active_slabs
1376 ± 13% +5.2% 1448 ± 4% slabinfo.Acpi-State.num_objs
26.67 ± 14% +5.0% 28.00 ± 5% slabinfo.Acpi-State.num_slabs
36.00 +0.0% 36.00 slabinfo.DCCP.active_objs
2.00 +0.0% 2.00 slabinfo.DCCP.active_slabs
36.00 +0.0% 36.00 slabinfo.DCCP.num_objs
2.00 +0.0% 2.00 slabinfo.DCCP.num_slabs
34.00 +0.0% 34.00 slabinfo.DCCPv6.active_objs
2.00 +0.0% 2.00 slabinfo.DCCPv6.active_slabs
34.00 +0.0% 34.00 slabinfo.DCCPv6.num_objs
2.00 +0.0% 2.00 slabinfo.DCCPv6.num_slabs
256.00 +0.0% 256.00 slabinfo.RAW.active_objs
8.00 +0.0% 8.00 slabinfo.RAW.active_slabs
256.00 +0.0% 256.00 slabinfo.RAW.num_objs
8.00 +0.0% 8.00 slabinfo.RAW.num_slabs
190.67 ± 6% +4.5% 199.33 ± 6% slabinfo.RAWv6.active_objs
7.33 ± 6% +4.5% 7.67 ± 6% slabinfo.RAWv6.active_slabs
190.67 ± 6% +4.5% 199.33 ± 6% slabinfo.RAWv6.num_objs
7.33 ± 6% +4.5% 7.67 ± 6% slabinfo.RAWv6.num_slabs
56.00 +0.0% 56.00 slabinfo.TCP.active_objs
4.00 +0.0% 4.00 slabinfo.TCP.active_slabs
56.00 +0.0% 56.00 slabinfo.TCP.num_objs
4.00 +0.0% 4.00 slabinfo.TCP.num_slabs
39.00 +0.0% 39.00 slabinfo.TCPv6.active_objs
3.00 +0.0% 3.00 slabinfo.TCPv6.active_slabs
39.00 +0.0% 39.00 slabinfo.TCPv6.num_objs
3.00 +0.0% 3.00 slabinfo.TCPv6.num_slabs
300.00 -3.3% 290.00 ± 12% slabinfo.UDP.active_objs
10.00 -3.3% 9.67 ± 12% slabinfo.UDP.active_slabs
300.00 -3.3% 290.00 ± 12% slabinfo.UDP.num_objs
10.00 -3.3% 9.67 ± 12% slabinfo.UDP.num_slabs
141.67 ± 8% +5.9% 150.00 slabinfo.UDPv6.active_objs
5.67 ± 8% +5.9% 6.00 slabinfo.UDPv6.active_slabs
141.67 ± 8% +5.9% 150.00 slabinfo.UDPv6.num_objs
5.67 ± 8% +5.9% 6.00 slabinfo.UDPv6.num_slabs
2967 ± 2% +0.6% 2986 ± 2% slabinfo.UNIX.active_objs
92.00 ± 2% +0.7% 92.67 ± 2% slabinfo.UNIX.active_slabs
2967 ± 2% +0.6% 2986 ± 2% slabinfo.UNIX.num_objs
92.00 ± 2% +0.7% 92.67 ± 2% slabinfo.UNIX.num_slabs
23604 +3.5% 24423 slabinfo.anon_vma.active_objs
517.00 +3.5% 535.00 slabinfo.anon_vma.active_slabs
23802 +3.5% 24626 slabinfo.anon_vma.num_objs
517.00 +3.5% 535.00 slabinfo.anon_vma.num_slabs
27164 +3.2% 28025 slabinfo.anon_vma_chain.active_objs
431.67 +3.6% 447.00 slabinfo.anon_vma_chain.active_slabs
27646 +3.6% 28636 slabinfo.anon_vma_chain.num_objs
431.67 +3.6% 447.00 slabinfo.anon_vma_chain.num_slabs
126.67 ± 7% +15.8% 146.67 ± 6% slabinfo.bdev_cache.active_objs
6.33 ± 7% +15.8% 7.33 ± 6% slabinfo.bdev_cache.active_slabs
126.67 ± 7% +15.8% 146.67 ± 6% slabinfo.bdev_cache.num_objs
6.33 ± 7% +15.8% 7.33 ± 6% slabinfo.bdev_cache.num_slabs
256.00 +0.0% 256.00 slabinfo.bio-120.active_objs
4.00 +0.0% 4.00 slabinfo.bio-120.active_slabs
256.00 +0.0% 256.00 slabinfo.bio-120.num_objs
4.00 +0.0% 4.00 slabinfo.bio-120.num_slabs
713.33 ± 9% +15.3% 822.67 ± 16% slabinfo.bio-184.active_objs
16.67 ± 10% +14.0% 19.00 ± 17% slabinfo.bio-184.active_slabs
713.33 ± 9% +15.3% 822.67 ± 16% slabinfo.bio-184.num_objs
16.67 ± 10% +14.0% 19.00 ± 17% slabinfo.bio-184.num_slabs
64.00 +0.0% 64.00 slabinfo.bio-216.active_objs
1.00 +0.0% 1.00 slabinfo.bio-216.active_slabs
64.00 +0.0% 64.00 slabinfo.bio-216.num_objs
1.00 +0.0% 1.00 slabinfo.bio-216.num_slabs
128.00 +0.0% 128.00 slabinfo.bio-248.active_objs
2.00 +0.0% 2.00 slabinfo.bio-248.active_slabs
128.00 +0.0% 128.00 slabinfo.bio-248.num_objs
2.00 +0.0% 2.00 slabinfo.bio-248.num_slabs
139.33 ± 13% +62.9% 227.00 ± 31% slabinfo.bio-336.active_objs
3.00 +66.7% 5.00 ± 28% slabinfo.bio-336.active_slabs
139.33 ± 13% +62.9% 227.00 ± 31% slabinfo.bio-336.num_objs
3.00 +66.7% 5.00 ± 28% slabinfo.bio-336.num_slabs
160.00 ± 14% +28.3% 205.33 ± 11% slabinfo.biovec-128.active_objs
10.00 ± 14% +23.3% 12.33 ± 13% slabinfo.biovec-128.active_slabs
160.00 ± 14% +28.3% 205.33 ± 11% slabinfo.biovec-128.num_objs
10.00 ± 14% +23.3% 12.33 ± 13% slabinfo.biovec-128.num_slabs
394.67 ± 3% +10.8% 437.33 ± 20% slabinfo.biovec-64.active_objs
12.33 ± 3% +10.8% 13.67 ± 20% slabinfo.biovec-64.active_slabs
394.67 ± 3% +10.8% 437.33 ± 20% slabinfo.biovec-64.num_objs
12.33 ± 3% +10.8% 13.67 ± 20% slabinfo.biovec-64.num_slabs
79.67 ± 7% +24.3% 99.00 ± 13% slabinfo.biovec-max.active_objs
9.67 ± 4% +24.1% 12.00 ± 11% slabinfo.biovec-max.active_slabs
79.67 ± 7% +24.3% 99.00 ± 13% slabinfo.biovec-max.num_objs
9.67 ± 4% +24.1% 12.00 ± 11% slabinfo.biovec-max.num_slabs
38.33 ±141% +241.7% 131.00 ± 71% slabinfo.btrfs_free_space.active_objs
0.67 ±141% +350.0% 3.00 ± 72% slabinfo.btrfs_free_space.active_slabs
38.33 ±141% +241.7% 131.00 ± 71% slabinfo.btrfs_free_space.num_objs
0.67 ±141% +350.0% 3.00 ± 72% slabinfo.btrfs_free_space.num_slabs
108.00 +7.1% 115.67 ± 10% slabinfo.btrfs_inode.active_objs
4.00 -8.3% 3.67 ± 12% slabinfo.btrfs_inode.active_slabs
108.00 +7.1% 115.67 ± 10% slabinfo.btrfs_inode.num_objs
4.00 -8.3% 3.67 ± 12% slabinfo.btrfs_inode.num_slabs
0.00 +2.1e+103% 21.33 ± 70% slabinfo.btrfs_ordered_extent.active_objs
0.00 -100.0% 0.00 slabinfo.btrfs_ordered_extent.active_slabs
0.00 +2.1e+103% 21.33 ± 70% slabinfo.btrfs_ordered_extent.num_objs
0.00 -100.0% 0.00 slabinfo.btrfs_ordered_extent.num_slabs
259.33 ± 13% +22.4% 317.33 ± 27% slabinfo.buffer_head.active_objs
6.33 ± 7% +21.1% 7.67 ± 24% slabinfo.buffer_head.active_slabs
259.33 ± 13% +22.4% 317.33 ± 27% slabinfo.buffer_head.num_objs
6.33 ± 7% +21.1% 7.67 ± 24% slabinfo.buffer_head.num_slabs
35983 +2.8% 36974 slabinfo.cred_jar.active_objs
870.67 +3.4% 900.33 slabinfo.cred_jar.active_slabs
36604 +3.3% 37827 slabinfo.cred_jar.num_objs
870.67 +3.4% 900.33 slabinfo.cred_jar.num_slabs
39.00 +0.0% 39.00 slabinfo.dax_cache.active_objs
1.00 +0.0% 1.00 slabinfo.dax_cache.active_slabs
39.00 +0.0% 39.00 slabinfo.dax_cache.num_objs
1.00 +0.0% 1.00 slabinfo.dax_cache.num_slabs
265837 +0.4% 266770 slabinfo.dentry.active_objs
7248 +0.9% 7310 slabinfo.dentry.active_slabs
304475 +0.8% 307058 slabinfo.dentry.num_objs
7248 +0.9% 7310 slabinfo.dentry.num_slabs
30.00 +0.0% 30.00 slabinfo.dmaengine-unmap-128.active_objs
1.00 +0.0% 1.00 slabinfo.dmaengine-unmap-128.active_slabs
30.00 +0.0% 30.00 slabinfo.dmaengine-unmap-128.num_objs
1.00 +0.0% 1.00 slabinfo.dmaengine-unmap-128.num_slabs
1489 ± 4% +0.2% 1492 ± 3% slabinfo.dmaengine-unmap-16.active_objs
35.00 ± 4% -1.0% 34.67 ± 3% slabinfo.dmaengine-unmap-16.active_slabs
1489 ± 4% +0.2% 1492 ± 3% slabinfo.dmaengine-unmap-16.num_objs
35.00 ± 4% -1.0% 34.67 ± 3% slabinfo.dmaengine-unmap-16.num_slabs
15.00 +0.0% 15.00 slabinfo.dmaengine-unmap-256.active_objs
1.00 +0.0% 1.00 slabinfo.dmaengine-unmap-256.active_slabs
15.00 +0.0% 15.00 slabinfo.dmaengine-unmap-256.num_objs
1.00 +0.0% 1.00 slabinfo.dmaengine-unmap-256.num_slabs
26156 ± 2% -10.4% 23434 ± 4% slabinfo.ep_head.active_objs
102.00 ± 2% -10.8% 91.00 ± 4% slabinfo.ep_head.active_slabs
26156 ± 2% -10.4% 23434 ± 4% slabinfo.ep_head.num_objs
102.00 ± 2% -10.8% 91.00 ± 4% slabinfo.ep_head.num_slabs
2878 -3.9% 2765 slabinfo.file_lock_cache.active_objs
77.33 -3.9% 74.33 slabinfo.file_lock_cache.active_slabs
2878 -3.9% 2765 slabinfo.file_lock_cache.num_objs
77.33 -3.9% 74.33 slabinfo.file_lock_cache.num_slabs
6210 -0.6% 6171 slabinfo.files_cache.active_objs
134.67 -1.0% 133.33 slabinfo.files_cache.active_slabs
6210 -0.6% 6171 slabinfo.files_cache.num_objs
134.67 -1.0% 133.33 slabinfo.files_cache.num_slabs
35516 +3.4% 36720 ± 3% slabinfo.filp.active_objs
667.33 -0.1% 666.67 ± 2% slabinfo.filp.active_slabs
42724 -0.1% 42691 ± 2% slabinfo.filp.num_objs
667.33 -0.1% 666.67 ± 2% slabinfo.filp.num_slabs
2732 ± 11% +15.7% 3163 ± 8% slabinfo.fsnotify_mark_connector.active_objs
21.33 ± 11% +14.1% 24.33 ± 8% slabinfo.fsnotify_mark_connector.active_slabs
2732 ± 11% +15.7% 3163 ± 8% slabinfo.fsnotify_mark_connector.num_objs
21.33 ± 11% +14.1% 24.33 ± 8% slabinfo.fsnotify_mark_connector.num_slabs
53975 -0.2% 53890 slabinfo.ftrace_event_field.active_objs
635.00 -0.2% 634.00 slabinfo.ftrace_event_field.active_slabs
53975 -0.2% 53890 slabinfo.ftrace_event_field.num_objs
635.00 -0.2% 634.00 slabinfo.ftrace_event_field.num_slabs
92.33 ± 55% +105.8% 190.00 ± 49% slabinfo.fuse_request.active_objs
1.33 ± 35% +125.0% 3.00 ± 47% slabinfo.fuse_request.active_slabs
92.33 ± 55% +105.8% 190.00 ± 49% slabinfo.fuse_request.num_objs
1.33 ± 35% +125.0% 3.00 ± 47% slabinfo.fuse_request.num_slabs
98.00 +0.0% 98.00 slabinfo.hugetlbfs_inode_cache.active_objs
2.00 +0.0% 2.00 slabinfo.hugetlbfs_inode_cache.active_slabs
98.00 +0.0% 98.00 slabinfo.hugetlbfs_inode_cache.num_objs
2.00 +0.0% 2.00 slabinfo.hugetlbfs_inode_cache.num_slabs
237067 +0.2% 237441 slabinfo.inode_cache.active_objs
4740 +0.2% 4748 slabinfo.inode_cache.active_slabs
241752 +0.2% 242169 slabinfo.inode_cache.num_objs
4740 +0.2% 4748 slabinfo.inode_cache.num_slabs
170.33 ± 20% +28.6% 219.00 ± 27% slabinfo.ip_fib_alias.active_objs
2.33 ± 20% +28.6% 3.00 ± 27% slabinfo.ip_fib_alias.active_slabs
170.33 ± 20% +28.6% 219.00 ± 27% slabinfo.ip_fib_alias.num_objs
2.33 ± 20% +28.6% 3.00 ± 27% slabinfo.ip_fib_alias.num_slabs
198.33 ± 20% +28.6% 255.00 ± 27% slabinfo.ip_fib_trie.active_objs
2.33 ± 20% +28.6% 3.00 ± 27% slabinfo.ip_fib_trie.active_slabs
198.33 ± 20% +28.6% 255.00 ± 27% slabinfo.ip_fib_trie.num_objs
2.33 ± 20% +28.6% 3.00 ± 27% slabinfo.ip_fib_trie.num_slabs
113280 -0.1% 113216 slabinfo.kernfs_node_cache.active_objs
1769 -0.0% 1769 slabinfo.kernfs_node_cache.active_slabs
113280 -0.1% 113216 slabinfo.kernfs_node_cache.num_objs
1769 -0.0% 1769 slabinfo.kernfs_node_cache.num_slabs
4582 +0.2% 4589 slabinfo.khugepaged_mm_slot.active_objs
127.00 +0.0% 127.00 slabinfo.khugepaged_mm_slot.active_slabs
4582 +0.2% 4589 slabinfo.khugepaged_mm_slot.num_objs
127.00 +0.0% 127.00 slabinfo.khugepaged_mm_slot.num_slabs
11297 -1.9% 11083 slabinfo.kmalloc-128.active_objs
177.33 ± 2% -1.1% 175.33 slabinfo.kmalloc-128.active_slabs
11384 ± 2% -1.3% 11238 slabinfo.kmalloc-128.num_objs
177.33 ± 2% -1.1% 175.33 slabinfo.kmalloc-128.num_slabs
49512 -0.0% 49493 slabinfo.kmalloc-16.active_objs
193.67 -0.2% 193.33 slabinfo.kmalloc-16.active_slabs
49578 -0.2% 49493 slabinfo.kmalloc-16.num_objs
193.67 -0.2% 193.33 slabinfo.kmalloc-16.num_slabs
7798 +0.3% 7822 slabinfo.kmalloc-192.active_objs
185.67 +0.4% 186.33 slabinfo.kmalloc-192.active_slabs
7798 +0.4% 7826 slabinfo.kmalloc-192.num_objs
185.67 +0.4% 186.33 slabinfo.kmalloc-192.num_slabs
6385 ± 3% +0.6% 6421 ± 3% slabinfo.kmalloc-1k.active_objs
200.67 ± 4% +1.0% 202.67 ± 3% slabinfo.kmalloc-1k.active_slabs
6434 ± 4% +0.9% 6490 ± 3% slabinfo.kmalloc-1k.num_objs
200.67 ± 4% +1.0% 202.67 ± 3% slabinfo.kmalloc-1k.num_slabs
11745 -0.9% 11642 slabinfo.kmalloc-256.active_objs
187.00 -1.4% 184.33 slabinfo.kmalloc-256.active_slabs
11994 -1.5% 11808 ± 2% slabinfo.kmalloc-256.num_objs
187.00 -1.4% 184.33 slabinfo.kmalloc-256.num_slabs
5794 -1.2% 5723 slabinfo.kmalloc-2k.active_objs
369.33 -1.5% 363.67 slabinfo.kmalloc-2k.active_slabs
5918 -1.6% 5824 slabinfo.kmalloc-2k.num_objs
369.33 -1.5% 363.67 slabinfo.kmalloc-2k.num_slabs
93489 ± 2% -1.6% 92022 ± 3% slabinfo.kmalloc-32.active_objs
732.00 ± 2% -1.6% 720.00 ± 3% slabinfo.kmalloc-32.active_slabs
93754 ± 3% -1.6% 92218 ± 3% slabinfo.kmalloc-32.num_objs
732.00 ± 2% -1.6% 720.00 ± 3% slabinfo.kmalloc-32.num_slabs
1945 +0.3% 1950 slabinfo.kmalloc-4k.active_objs
248.00 +0.3% 248.67 slabinfo.kmalloc-4k.active_slabs
1987 +0.3% 1993 slabinfo.kmalloc-4k.num_objs
248.00 +0.3% 248.67 slabinfo.kmalloc-4k.num_slabs
26031 +1.7% 26484 slabinfo.kmalloc-512.active_objs
468.67 +7.1% 502.00 ± 2% slabinfo.kmalloc-512.active_slabs
30027 +7.0% 32143 ± 2% slabinfo.kmalloc-512.num_objs
468.67 +7.1% 502.00 ± 2% slabinfo.kmalloc-512.num_slabs
114083 -0.9% 113046 slabinfo.kmalloc-64.active_objs
1788 -1.0% 1770 slabinfo.kmalloc-64.active_slabs
114457 -1.0% 113312 slabinfo.kmalloc-64.num_objs
1788 -1.0% 1770 slabinfo.kmalloc-64.num_slabs
74273 -0.5% 73893 ± 2% slabinfo.kmalloc-8.active_objs
150.67 ± 2% +0.0% 150.67 slabinfo.kmalloc-8.active_slabs
77367 ± 2% +0.1% 77408 slabinfo.kmalloc-8.num_objs
150.67 ± 2% +0.0% 150.67 slabinfo.kmalloc-8.num_slabs
875.33 +0.0% 875.33 slabinfo.kmalloc-8k.active_objs
223.00 +0.0% 223.00 slabinfo.kmalloc-8k.active_slabs
894.33 +0.0% 894.33 slabinfo.kmalloc-8k.num_objs
223.00 +0.0% 223.00 slabinfo.kmalloc-8k.num_slabs
9911 -0.3% 9881 slabinfo.kmalloc-96.active_objs
238.00 -0.4% 237.00 slabinfo.kmalloc-96.active_slabs
10003 -0.3% 9974 slabinfo.kmalloc-96.num_objs
238.00 -0.4% 237.00 slabinfo.kmalloc-96.num_slabs
853.33 ± 9% +7.5% 917.33 ± 3% slabinfo.kmalloc-cg-128.active_objs
13.33 ± 9% +7.5% 14.33 ± 3% slabinfo.kmalloc-cg-128.active_slabs
853.33 ± 9% +7.5% 917.33 ± 3% slabinfo.kmalloc-cg-128.num_objs
13.33 ± 9% +7.5% 14.33 ± 3% slabinfo.kmalloc-cg-128.num_slabs
3136 ± 2% +1.5% 3184 ± 6% slabinfo.kmalloc-cg-16.active_objs
11.67 ± 4% +2.9% 12.00 ± 6% slabinfo.kmalloc-cg-16.active_slabs
3136 ± 2% +1.5% 3184 ± 6% slabinfo.kmalloc-cg-16.num_objs
11.67 ± 4% +2.9% 12.00 ± 6% slabinfo.kmalloc-cg-16.num_slabs
5371 +0.0% 5372 slabinfo.kmalloc-cg-192.active_objs
127.00 +0.0% 127.00 slabinfo.kmalloc-cg-192.active_slabs
5371 +0.0% 5372 slabinfo.kmalloc-cg-192.num_objs
127.00 +0.0% 127.00 slabinfo.kmalloc-cg-192.num_slabs
4531 -3.9% 4353 slabinfo.kmalloc-cg-1k.active_objs
141.00 -4.0% 135.33 slabinfo.kmalloc-cg-1k.active_slabs
4531 -3.9% 4353 slabinfo.kmalloc-cg-1k.num_objs
141.00 -4.0% 135.33 slabinfo.kmalloc-cg-1k.num_slabs
613.67 ± 4% +17.1% 718.33 ± 4% slabinfo.kmalloc-cg-256.active_objs
8.67 ± 5% +19.2% 10.33 ± 4% slabinfo.kmalloc-cg-256.active_slabs
613.67 ± 4% +17.1% 718.33 ± 4% slabinfo.kmalloc-cg-256.num_objs
8.67 ± 5% +19.2% 10.33 ± 4% slabinfo.kmalloc-cg-256.num_slabs
1317 +1.3% 1335 slabinfo.kmalloc-cg-2k.active_objs
82.00 +1.2% 83.00 slabinfo.kmalloc-cg-2k.active_slabs
1317 +1.3% 1335 slabinfo.kmalloc-cg-2k.num_objs
82.00 +1.2% 83.00 slabinfo.kmalloc-cg-2k.num_slabs
16374 +0.0% 16374 slabinfo.kmalloc-cg-32.active_objs
127.00 +0.0% 127.00 slabinfo.kmalloc-cg-32.active_slabs
16374 +0.0% 16374 slabinfo.kmalloc-cg-32.num_objs
127.00 +0.0% 127.00 slabinfo.kmalloc-cg-32.num_slabs
1040 -0.3% 1037 slabinfo.kmalloc-cg-4k.active_objs
130.00 -0.3% 129.67 slabinfo.kmalloc-cg-4k.active_slabs
1040 -0.3% 1037 slabinfo.kmalloc-cg-4k.num_objs
130.00 -0.3% 129.67 slabinfo.kmalloc-cg-4k.num_slabs
8192 -0.5% 8149 slabinfo.kmalloc-cg-512.active_objs
128.00 -0.5% 127.33 slabinfo.kmalloc-cg-512.active_slabs
8192 -0.5% 8149 slabinfo.kmalloc-cg-512.num_objs
128.00 -0.5% 127.33 slabinfo.kmalloc-cg-512.num_slabs
6462 +1.6% 6563 ± 2% slabinfo.kmalloc-cg-64.active_objs
100.33 +1.3% 101.67 ± 2% slabinfo.kmalloc-cg-64.active_slabs
6462 +1.6% 6563 ± 2% slabinfo.kmalloc-cg-64.num_objs
100.33 +1.3% 101.67 ± 2% slabinfo.kmalloc-cg-64.num_slabs
66047 +0.0% 66048 slabinfo.kmalloc-cg-8.active_objs
128.67 +0.3% 129.00 slabinfo.kmalloc-cg-8.active_slabs
66047 +0.0% 66048 slabinfo.kmalloc-cg-8.num_objs
128.67 +0.3% 129.00 slabinfo.kmalloc-cg-8.num_slabs
58.67 ± 3% -2.3% 57.33 ± 4% slabinfo.kmalloc-cg-8k.active_objs
14.67 ± 3% -6.8% 13.67 ± 3% slabinfo.kmalloc-cg-8k.active_slabs
58.67 ± 3% -2.3% 57.33 ± 4% slabinfo.kmalloc-cg-8k.num_objs
14.67 ± 3% -6.8% 13.67 ± 3% slabinfo.kmalloc-cg-8k.num_slabs
3337 +0.5% 3352 slabinfo.kmalloc-cg-96.active_objs
79.00 +0.4% 79.33 slabinfo.kmalloc-cg-96.active_slabs
3337 +0.5% 3352 slabinfo.kmalloc-cg-96.num_objs
79.00 +0.4% 79.33 slabinfo.kmalloc-cg-96.num_slabs
213.33 ± 14% +20.0% 256.00 slabinfo.kmalloc-rcl-128.active_objs
3.33 ± 14% +20.0% 4.00 slabinfo.kmalloc-rcl-128.active_slabs
213.33 ± 14% +20.0% 256.00 slabinfo.kmalloc-rcl-128.num_objs
3.33 ± 14% +20.0% 4.00 slabinfo.kmalloc-rcl-128.num_slabs
42.00 +0.0% 42.00 slabinfo.kmalloc-rcl-192.active_objs
1.00 +0.0% 1.00 slabinfo.kmalloc-rcl-192.active_slabs
42.00 +0.0% 42.00 slabinfo.kmalloc-rcl-192.num_objs
1.00 +0.0% 1.00 slabinfo.kmalloc-rcl-192.num_slabs
12059 +0.7% 12149 ± 2% slabinfo.kmalloc-rcl-64.active_objs
188.00 +0.9% 189.67 ± 2% slabinfo.kmalloc-rcl-64.active_slabs
12059 +0.8% 12150 ± 2% slabinfo.kmalloc-rcl-64.num_objs
188.00 +0.9% 189.67 ± 2% slabinfo.kmalloc-rcl-64.num_slabs
1671 +1.0% 1687 ± 2% slabinfo.kmalloc-rcl-96.active_objs
39.33 +0.8% 39.67 ± 2% slabinfo.kmalloc-rcl-96.active_slabs
1671 +1.0% 1687 ± 2% slabinfo.kmalloc-rcl-96.num_objs
39.33 +0.8% 39.67 ± 2% slabinfo.kmalloc-rcl-96.num_slabs
960.00 ± 5% +2.2% 981.33 ± 6% slabinfo.kmem_cache.active_objs
15.00 ± 5% +2.2% 15.33 ± 6% slabinfo.kmem_cache.active_slabs
960.00 ± 5% +2.2% 981.33 ± 6% slabinfo.kmem_cache.num_objs
15.00 ± 5% +2.2% 15.33 ± 6% slabinfo.kmem_cache.num_slabs
1138 ± 4% +1.9% 1159 ± 5% slabinfo.kmem_cache_node.active_objs
18.00 ± 4% +1.9% 18.33 ± 5% slabinfo.kmem_cache_node.active_slabs
1152 ± 4% +1.9% 1173 ± 5% slabinfo.kmem_cache_node.num_objs
18.00 ± 4% +1.9% 18.33 ± 5% slabinfo.kmem_cache_node.num_slabs
46174 -0.3% 46022 slabinfo.lsm_file_cache.active_objs
273.00 -0.2% 272.33 slabinfo.lsm_file_cache.active_slabs
46524 -0.3% 46407 slabinfo.lsm_file_cache.num_objs
273.00 -0.2% 272.33 slabinfo.lsm_file_cache.num_slabs
4223 +0.9% 4263 slabinfo.mm_struct.active_objs
140.33 +1.2% 142.00 slabinfo.mm_struct.active_slabs
4223 +1.0% 4266 slabinfo.mm_struct.num_objs
140.33 +1.2% 142.00 slabinfo.mm_struct.num_slabs
952.00 ± 2% +1.8% 969.00 ± 4% slabinfo.mnt_cache.active_objs
18.67 ± 2% +1.8% 19.00 ± 4% slabinfo.mnt_cache.active_slabs
952.00 ± 2% +1.8% 969.00 ± 4% slabinfo.mnt_cache.num_objs
18.67 ± 2% +1.8% 19.00 ± 4% slabinfo.mnt_cache.num_slabs
34.00 +0.0% 34.00 slabinfo.mqueue_inode_cache.active_objs
1.00 +0.0% 1.00 slabinfo.mqueue_inode_cache.active_slabs
34.00 +0.0% 34.00 slabinfo.mqueue_inode_cache.num_objs
1.00 +0.0% 1.00 slabinfo.mqueue_inode_cache.num_slabs
1758 ± 6% +5.3% 1851 ± 3% slabinfo.names_cache.active_objs
220.67 ± 6% +5.6% 233.00 ± 3% slabinfo.names_cache.active_slabs
1770 ± 6% +5.4% 1867 ± 3% slabinfo.names_cache.num_objs
220.67 ± 6% +5.6% 233.00 ± 3% slabinfo.names_cache.num_slabs
8.00 +0.0% 8.00 slabinfo.net_namespace.active_objs
1.00 +0.0% 1.00 slabinfo.net_namespace.active_slabs
8.00 +0.0% 8.00 slabinfo.net_namespace.num_objs
1.00 +0.0% 1.00 slabinfo.net_namespace.num_slabs
36.00 +0.0% 36.00 slabinfo.nfs_read_data.active_objs
1.00 +0.0% 1.00 slabinfo.nfs_read_data.active_slabs
36.00 +0.0% 36.00 slabinfo.nfs_read_data.num_objs
1.00 +0.0% 1.00 slabinfo.nfs_read_data.num_slabs
271.67 ± 9% +0.0% 271.67 ± 9% slabinfo.nsproxy.active_objs
4.33 ± 10% +0.0% 4.33 ± 10% slabinfo.nsproxy.active_slabs
271.67 ± 9% +0.0% 271.67 ± 9% slabinfo.nsproxy.num_objs
4.33 ± 10% +0.0% 4.33 ± 10% slabinfo.nsproxy.num_slabs
180.00 +0.0% 180.00 slabinfo.numa_policy.active_objs
3.00 +0.0% 3.00 slabinfo.numa_policy.active_slabs
180.00 +0.0% 180.00 slabinfo.numa_policy.num_objs
3.00 +0.0% 3.00 slabinfo.numa_policy.num_slabs
13045 -0.0% 13045 slabinfo.pde_opener.active_objs
127.00 +0.0% 127.00 slabinfo.pde_opener.active_slabs
13045 -0.0% 13045 slabinfo.pde_opener.num_objs
127.00 +0.0% 127.00 slabinfo.pde_opener.num_slabs
5212 +0.4% 5233 slabinfo.perf_event.active_objs
196.00 +0.3% 196.67 slabinfo.perf_event.active_slabs
5303 +0.4% 5324 slabinfo.perf_event.num_objs
196.00 +0.3% 196.67 slabinfo.perf_event.num_slabs
12380 ± 4% -0.4% 12329 ± 5% slabinfo.pid.active_objs
193.33 ± 4% -0.5% 192.33 ± 5% slabinfo.pid.active_slabs
12390 ± 4% -0.3% 12349 ± 5% slabinfo.pid.num_objs
193.33 ± 4% -0.5% 192.33 ± 5% slabinfo.pid.num_slabs
3111 ± 2% +7.7% 3350 ± 3% slabinfo.pool_workqueue.active_objs
48.67 ± 2% +6.8% 52.00 ± 3% slabinfo.pool_workqueue.active_slabs
3136 ± 2% +7.2% 3363 ± 3% slabinfo.pool_workqueue.num_objs
48.67 ± 2% +6.8% 52.00 ± 3% slabinfo.pool_workqueue.num_slabs
3556 +3.1% 3668 slabinfo.proc_dir_entry.active_objs
84.67 +3.1% 87.33 slabinfo.proc_dir_entry.active_slabs
3556 +3.1% 3668 slabinfo.proc_dir_entry.num_objs
84.67 +3.1% 87.33 slabinfo.proc_dir_entry.num_slabs
13396 -0.4% 13339 slabinfo.proc_inode_cache.active_objs
291.00 -0.3% 290.00 slabinfo.proc_inode_cache.active_slabs
13414 -0.4% 13367 slabinfo.proc_inode_cache.num_objs
291.00 -0.3% 290.00 slabinfo.proc_inode_cache.num_slabs
97227 +0.3% 97498 slabinfo.radix_tree_node.active_objs
1746 +0.2% 1750 slabinfo.radix_tree_node.active_slabs
97813 +0.2% 98049 slabinfo.radix_tree_node.num_objs
1746 +0.2% 1750 slabinfo.radix_tree_node.num_slabs
128.33 ± 12% +0.0% 128.33 ± 12% slabinfo.request_queue.active_objs
3.67 ± 12% +0.0% 3.67 ± 12% slabinfo.request_queue.active_slabs
128.33 ± 12% +0.0% 128.33 ± 12% slabinfo.request_queue.num_objs
3.67 ± 12% +0.0% 3.67 ± 12% slabinfo.request_queue.num_slabs
46.00 +0.0% 46.00 slabinfo.rpc_inode_cache.active_objs
1.00 +0.0% 1.00 slabinfo.rpc_inode_cache.active_slabs
46.00 +0.0% 46.00 slabinfo.rpc_inode_cache.num_objs
1.00 +0.0% 1.00 slabinfo.rpc_inode_cache.num_slabs
746.67 ± 4% +0.0% 746.67 ± 4% slabinfo.scsi_sense_cache.active_objs
11.67 ± 4% +0.0% 11.67 ± 4% slabinfo.scsi_sense_cache.active_slabs
746.67 ± 4% +0.0% 746.67 ± 4% slabinfo.scsi_sense_cache.num_objs
11.67 ± 4% +0.0% 11.67 ± 4% slabinfo.scsi_sense_cache.num_slabs
8704 +0.0% 8704 slabinfo.seq_file.active_objs
128.00 +0.0% 128.00 slabinfo.seq_file.active_slabs
8704 +0.0% 8704 slabinfo.seq_file.num_objs
128.00 +0.0% 128.00 slabinfo.seq_file.num_slabs
6965 -0.8% 6910 slabinfo.shmem_inode_cache.active_objs
165.33 -0.8% 164.00 slabinfo.shmem_inode_cache.active_slabs
6965 -0.8% 6910 slabinfo.shmem_inode_cache.num_objs
165.33 -0.8% 164.00 slabinfo.shmem_inode_cache.num_slabs
6477 +1.4% 6566 slabinfo.sighand_cache.active_objs
434.33 +1.3% 440.00 slabinfo.sighand_cache.active_slabs
6525 +1.3% 6611 slabinfo.sighand_cache.num_objs
434.33 +1.3% 440.00 slabinfo.sighand_cache.num_slabs
8339 -2.4% 8142 slabinfo.signal_cache.active_objs
301.00 -2.4% 293.67 slabinfo.signal_cache.active_slabs
8446 -2.4% 8244 slabinfo.signal_cache.num_objs
301.00 -2.4% 293.67 slabinfo.signal_cache.num_slabs
7070 ± 2% -1.2% 6985 slabinfo.sigqueue.active_objs
137.67 ± 2% -1.2% 136.00 slabinfo.sigqueue.active_slabs
7070 ± 2% -1.2% 6985 slabinfo.sigqueue.num_objs
137.67 ± 2% -1.2% 136.00 slabinfo.sigqueue.num_slabs
853.33 ± 7% +9.5% 934.00 ± 27% slabinfo.skbuff_fclone_cache.active_objs
13.33 ± 7% +7.5% 14.33 ± 25% slabinfo.skbuff_fclone_cache.active_slabs
853.33 ± 7% +9.5% 934.00 ± 27% slabinfo.skbuff_fclone_cache.num_objs
13.33 ± 7% +7.5% 14.33 ± 25% slabinfo.skbuff_fclone_cache.num_slabs
9984 +0.4% 10022 slabinfo.skbuff_head_cache.active_objs
157.33 +0.0% 157.33 slabinfo.skbuff_head_cache.active_slabs
10069 +0.0% 10069 slabinfo.skbuff_head_cache.num_objs
157.33 +0.0% 157.33 slabinfo.skbuff_head_cache.num_slabs
4409 ± 3% +1.0% 4454 ± 3% slabinfo.sock_inode_cache.active_objs
112.67 ± 3% +1.2% 114.00 ± 3% slabinfo.sock_inode_cache.active_slabs
4409 ± 3% +1.0% 4454 ± 3% slabinfo.sock_inode_cache.num_objs
112.67 ± 3% +1.2% 114.00 ± 3% slabinfo.sock_inode_cache.num_slabs
1416 ± 6% +3.7% 1469 ± 3% slabinfo.task_group.active_objs
30.33 ± 5% +3.3% 31.33 ± 3% slabinfo.task_group.active_slabs
1416 ± 6% +3.7% 1469 ± 3% slabinfo.task_group.num_objs
30.33 ± 5% +3.3% 31.33 ± 3% slabinfo.task_group.num_slabs
2661 -0.4% 2651 slabinfo.task_struct.active_objs
2665 -0.4% 2655 slabinfo.task_struct.active_slabs
2665 -0.4% 2655 slabinfo.task_struct.num_objs
2665 -0.4% 2655 slabinfo.task_struct.num_slabs
308.00 -4.5% 294.00 ± 6% slabinfo.taskstats.active_objs
7.00 -4.8% 6.67 ± 7% slabinfo.taskstats.active_slabs
308.00 -4.5% 294.00 ± 6% slabinfo.taskstats.num_objs
7.00 -4.8% 6.67 ± 7% slabinfo.taskstats.num_slabs
2722 ± 4% +3.7% 2822 slabinfo.trace_event_file.active_objs
59.00 ± 4% +4.0% 61.33 slabinfo.trace_event_file.active_slabs
2722 ± 4% +3.7% 2822 slabinfo.trace_event_file.num_objs
59.00 ± 4% +4.0% 61.33 slabinfo.trace_event_file.num_slabs
66.00 +0.0% 66.00 slabinfo.tw_sock_TCP.active_objs
1.00 +0.0% 1.00 slabinfo.tw_sock_TCP.active_slabs
66.00 +0.0% 66.00 slabinfo.tw_sock_TCP.num_objs
1.00 +0.0% 1.00 slabinfo.tw_sock_TCP.num_slabs
74.00 -16.7% 61.67 ± 28% slabinfo.uts_namespace.active_objs
2.00 -16.7% 1.67 ± 28% slabinfo.uts_namespace.active_slabs
74.00 -16.7% 61.67 ± 28% slabinfo.uts_namespace.num_objs
2.00 -16.7% 1.67 ± 28% slabinfo.uts_namespace.num_slabs
29597 -0.8% 29361 slabinfo.vmap_area.active_objs
580.33 -0.5% 577.67 slabinfo.vmap_area.active_slabs
37166 -0.4% 37007 slabinfo.vmap_area.num_objs
580.33 -0.5% 577.67 slabinfo.vmap_area.num_slabs
6.45 -3.4 3.07 ± 2% perf-profile.calltrace.cycles-pp.do_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
8.05 -2.8 5.27 ± 2% perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
6.69 ± 2% -1.6 5.08 ± 6% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
6.68 ± 2% -1.6 5.07 ± 6% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
3.24 ± 2% -1.6 1.64 ± 3% perf-profile.calltrace.cycles-pp.__x64_sys_exit_group.do_syscall_64.entry_SYSCALL_64_after_hwframe
3.24 ± 2% -1.6 1.64 ± 3% perf-profile.calltrace.cycles-pp.do_group_exit.__x64_sys_exit_group.do_syscall_64.entry_SYSCALL_64_after_hwframe
3.24 ± 2% -1.6 1.64 ± 3% perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64.entry_SYSCALL_64_after_hwframe
3.09 ± 2% -1.6 1.49 ± 3% perf-profile.calltrace.cycles-pp.exit_mm.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
3.07 ± 2% -1.6 1.48 ± 4% perf-profile.calltrace.cycles-pp.exit_mmap.__mmput.exit_mm.do_exit.do_group_exit
3.08 ± 2% -1.6 1.49 ± 4% perf-profile.calltrace.cycles-pp.__mmput.exit_mm.do_exit.do_group_exit.__x64_sys_exit_group
2.06 ± 2% -1.4 0.64 ± 2% perf-profile.calltrace.cycles-pp.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas.exit_mmap
2.08 ± 2% -1.4 0.67 ± 2% perf-profile.calltrace.cycles-pp.zap_pmd_range.unmap_page_range.unmap_vmas.exit_mmap.__mmput
2.09 ± 2% -1.4 0.68 ± 3% perf-profile.calltrace.cycles-pp.unmap_page_range.unmap_vmas.exit_mmap.__mmput.exit_mm
2.11 ± 2% -1.4 0.70 ± 2% perf-profile.calltrace.cycles-pp.unmap_vmas.exit_mmap.__mmput.exit_mm.do_exit
5.45 -1.2 4.28 perf-profile.calltrace.cycles-pp.asm_exc_page_fault
2.71 ± 2% -1.1 1.58 ± 3% perf-profile.calltrace.cycles-pp.vma_alloc_folio.do_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
2.63 -1.1 1.51 ± 2% perf-profile.calltrace.cycles-pp.__alloc_pages.__folio_alloc.vma_alloc_folio.do_anonymous_page.__handle_mm_fault
2.64 ± 2% -1.1 1.52 ± 3% perf-profile.calltrace.cycles-pp.__folio_alloc.vma_alloc_folio.do_anonymous_page.__handle_mm_fault.handle_mm_fault
2.51 -1.1 1.41 ± 3% perf-profile.calltrace.cycles-pp.get_page_from_freelist.__alloc_pages.__folio_alloc.vma_alloc_folio.do_anonymous_page
5.27 -1.1 4.22 perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault
1.05 ± 2% -1.1 0.00 perf-profile.calltrace.cycles-pp.asm_exc_page_fault.build_decl
5.24 -1.0 4.20 perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
4.93 -1.0 3.92 perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
0.98 ± 3% -1.0 0.00 perf-profile.calltrace.cycles-pp.tlb_batch_pages_flush.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas
0.95 ± 2% -1.0 0.00 perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.build_decl
0.95 ± 2% -1.0 0.00 perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.build_decl
0.90 ± 3% -0.9 0.00 perf-profile.calltrace.cycles-pp.release_pages.tlb_batch_pages_flush.zap_pte_range.zap_pmd_range.unmap_page_range
0.90 ± 2% -0.9 0.00 perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.build_decl
0.89 -0.9 0.00 perf-profile.calltrace.cycles-pp.__mem_cgroup_charge.do_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
0.87 -0.9 0.00 perf-profile.calltrace.cycles-pp.asm_exc_page_fault.make_node
0.79 -0.8 0.00 perf-profile.calltrace.cycles-pp.clear_page_erms.get_page_from_freelist.__alloc_pages.__folio_alloc.vma_alloc_folio
0.79 ± 3% -0.8 0.00 perf-profile.calltrace.cycles-pp.rmqueue.get_page_from_freelist.__alloc_pages.__folio_alloc.vma_alloc_folio
0.78 ± 2% -0.8 0.00 perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.make_node
0.78 ± 2% -0.8 0.00 perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.make_node
0.74 ± 2% -0.7 0.00 perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.make_node
0.68 -0.7 0.00 perf-profile.calltrace.cycles-pp.charge_memcg.__mem_cgroup_charge.do_anonymous_page.__handle_mm_fault.handle_mm_fault
0.60 ± 4% -0.6 0.00 perf-profile.calltrace.cycles-pp.rmqueue_bulk.rmqueue.get_page_from_freelist.__alloc_pages.__folio_alloc
0.59 ± 4% -0.6 0.00 perf-profile.calltrace.cycles-pp.free_unref_page_list.release_pages.tlb_batch_pages_flush.zap_pte_range.zap_pmd_range
0.55 -0.6 0.00 perf-profile.calltrace.cycles-pp.allocate_struct_function
0.55 -0.5 0.00 perf-profile.calltrace.cycles-pp.tick_sched_handle.tick_sched_timer.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt
0.55 -0.5 0.00 perf-profile.calltrace.cycles-pp.linemap_enter_macro
0.54 -0.5 0.00 perf-profile.calltrace.cycles-pp.update_process_times.tick_sched_handle.tick_sched_timer.__hrtimer_run_queues.hrtimer_interrupt
1.71 -0.5 1.19 perf-profile.calltrace.cycles-pp.make_node
0.67 ± 3% -0.5 0.18 ±141% perf-profile.calltrace.cycles-pp.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.exit_mmap.__mmput
2.03 ± 2% -0.5 1.58 perf-profile.calltrace.cycles-pp.build_decl
0.35 ± 70% -0.4 0.00 perf-profile.calltrace.cycles-pp.free_pcppages_bulk.free_unref_page_list.release_pages.tlb_batch_pages_flush.zap_pte_range
0.35 ± 70% -0.4 0.00 perf-profile.calltrace.cycles-pp.tlb_finish_mmu.unmap_region.__do_munmap.__vm_munmap.__x64_sys_munmap
0.34 ± 70% -0.3 0.00 perf-profile.calltrace.cycles-pp.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region.__do_munmap.__vm_munmap
0.34 ± 70% -0.3 0.00 perf-profile.calltrace.cycles-pp.wp_page_copy.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
0.33 ± 70% -0.3 0.00 perf-profile.calltrace.cycles-pp.lookup_name
1.78 -0.2 1.55 ± 3% perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.elf_read
1.86 ± 2% -0.2 1.64 ± 3% perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.elf_read
1.87 ± 2% -0.2 1.65 ± 3% perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.elf_read
2.09 ± 2% -0.2 1.89 ± 3% perf-profile.calltrace.cycles-pp.asm_exc_page_fault.elf_read
0.75 ± 3% -0.2 0.56 ± 5% perf-profile.calltrace.cycles-pp.tlb_finish_mmu.exit_mmap.__mmput.exit_mm.do_exit
0.74 ± 3% -0.2 0.55 ± 6% perf-profile.calltrace.cycles-pp.tlb_batch_pages_flush.tlb_finish_mmu.exit_mmap.__mmput.exit_mm
2.53 ± 2% -0.2 2.35 ± 3% perf-profile.calltrace.cycles-pp.elf_read
0.17 ±141% -0.2 0.00 perf-profile.calltrace.cycles-pp.release_pages.tlb_batch_pages_flush.tlb_finish_mmu.unmap_region.__do_munmap
0.78 ± 3% -0.1 0.64 ± 2% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__munmap
0.78 ± 3% -0.1 0.64 ± 2% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
0.78 ± 3% -0.1 0.64 ± 2% perf-profile.calltrace.cycles-pp.__munmap
0.78 ± 3% -0.1 0.64 ± 2% perf-profile.calltrace.cycles-pp.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
0.78 ± 3% -0.1 0.64 ± 2% perf-profile.calltrace.cycles-pp.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.__munmap
0.78 ± 2% -0.1 0.64 ± 2% perf-profile.calltrace.cycles-pp.__do_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.76 ± 2% -0.1 0.62 ± 2% perf-profile.calltrace.cycles-pp.unmap_region.__do_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64
2.49 -0.1 2.37 perf-profile.calltrace.cycles-pp.path_openat.do_filp_open.do_sys_openat2.__x64_sys_openat.do_syscall_64
1.19 -0.1 1.07 perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt
2.50 -0.1 2.39 perf-profile.calltrace.cycles-pp.do_filp_open.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe
1.12 -0.1 1.01 perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
0.87 -0.1 0.76 perf-profile.calltrace.cycles-pp.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
0.86 -0.1 0.75 perf-profile.calltrace.cycles-pp.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
1.41 ± 4% -0.1 1.31 perf-profile.calltrace.cycles-pp.read
1.77 -0.1 1.68 perf-profile.calltrace.cycles-pp.type_hash_canon
0.63 -0.1 0.53 perf-profile.calltrace.cycles-pp.build_function_type
1.37 ± 4% -0.1 1.28 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
1.34 ± 5% -0.1 1.25 perf-profile.calltrace.cycles-pp.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
1.37 ± 5% -0.1 1.28 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.read
0.74 -0.1 0.65 ± 2% perf-profile.calltrace.cycles-pp.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt
1.35 ± 5% -0.1 1.26 perf-profile.calltrace.cycles-pp.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
1.24 -0.1 1.17 perf-profile.calltrace.cycles-pp.link_path_walk.path_openat.do_filp_open.do_sys_openat2.__x64_sys_openat
1.82 -0.1 1.76 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.open64
1.81 -0.1 1.75 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.open64
0.61 ± 2% -0.1 0.54 ± 3% perf-profile.calltrace.cycles-pp.tick_sched_timer.__hrtimer_run_queues.hrtimer_interrupt.__sysvec_apic_timer_interrupt.sysvec_apic_timer_interrupt
1.86 -0.1 1.80 perf-profile.calltrace.cycles-pp.open64
1.72 -0.0 1.67 perf-profile.calltrace.cycles-pp.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe.open64
1.71 -0.0 1.66 perf-profile.calltrace.cycles-pp.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe.open64
1.14 ± 2% -0.0 1.10 perf-profile.calltrace.cycles-pp.filemap_read.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
1.09 -0.0 1.06 ± 3% perf-profile.calltrace.cycles-pp.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe
1.09 -0.0 1.05 ± 3% perf-profile.calltrace.cycles-pp.do_sys_openat2.__x64_sys_openat.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.85 ± 3% -0.0 0.82 perf-profile.calltrace.cycles-pp._copy_to_iter.copy_page_to_iter.filemap_read.vfs_read.ksys_read
0.71 ± 3% -0.0 0.68 perf-profile.calltrace.cycles-pp.copy_user_enhanced_fast_string.copyout._copy_to_iter.copy_page_to_iter.filemap_read
0.82 ± 3% -0.0 0.80 perf-profile.calltrace.cycles-pp.copyout._copy_to_iter.copy_page_to_iter.filemap_read.vfs_read
0.72 -0.0 0.70 ± 3% perf-profile.calltrace.cycles-pp.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
0.86 ± 3% -0.0 0.84 perf-profile.calltrace.cycles-pp.copy_page_to_iter.filemap_read.vfs_read.ksys_read.do_syscall_64
0.74 ± 2% +0.0 0.75 perf-profile.calltrace.cycles-pp.pop_scope
2.49 +0.0 2.51 ± 2% perf-profile.calltrace.cycles-pp._cpp_create_definition
0.60 +0.0 0.62 ± 4% perf-profile.calltrace.cycles-pp.bprm_execve.do_execveat_common.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.66 ± 2% +0.0 0.69 perf-profile.calltrace.cycles-pp.variably_modified_type_p
0.51 +0.0 0.54 ± 5% perf-profile.calltrace.cycles-pp.exec_binprm.bprm_execve.do_execveat_common.__x64_sys_execve.do_syscall_64
0.51 +0.0 0.54 ± 5% perf-profile.calltrace.cycles-pp.search_binary_handler.exec_binprm.bprm_execve.do_execveat_common.__x64_sys_execve
0.88 +0.1 0.93 ± 4% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.execve
0.88 +0.1 0.93 ± 4% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
0.88 +0.1 0.93 ± 4% perf-profile.calltrace.cycles-pp.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
0.89 +0.1 0.94 ± 3% perf-profile.calltrace.cycles-pp.execve
0.88 +0.1 0.93 ± 3% perf-profile.calltrace.cycles-pp.do_execveat_common.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
0.55 ± 2% +0.1 0.61 perf-profile.calltrace.cycles-pp._cpp_skip_block_comment
0.51 +0.1 0.58 ± 2% perf-profile.calltrace.cycles-pp.build_binary_op
0.51 +0.1 0.57 perf-profile.calltrace.cycles-pp.operand_equal_p
1.01 +0.1 1.08 perf-profile.calltrace.cycles-pp.size_binop_loc
0.56 ± 2% +0.1 0.63 ± 5% perf-profile.calltrace.cycles-pp.location_get_source_line
0.62 ± 5% +0.1 0.70 ± 2% perf-profile.calltrace.cycles-pp.df_bb_refs_record
0.64 +0.1 0.72 perf-profile.calltrace.cycles-pp.c_parser_declspecs
0.87 +0.1 0.96 perf-profile.calltrace.cycles-pp.cfree
0.58 ± 2% +0.1 0.69 perf-profile.calltrace.cycles-pp.layout_type
0.83 ± 3% +0.1 0.95 ± 9% perf-profile.calltrace.cycles-pp.linemap_location_in_system_header_p
0.00 +0.2 0.17 ±141% perf-profile.calltrace.cycles-pp.constrain_operands
0.00 +0.2 0.17 ±141% perf-profile.calltrace.cycles-pp.getenv
0.00 +0.2 0.17 ±141% perf-profile.calltrace.cycles-pp._cpp_handle_directive
1.94 +0.2 2.12 perf-profile.calltrace.cycles-pp.malloc
0.00 +0.2 0.19 ±141% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.remove_wait_queue.poll_freewait.do_select
0.00 +0.2 0.19 ±141% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.remove_wait_queue.poll_freewait.do_select.core_sys_select
0.87 +0.2 1.05 perf-profile.calltrace.cycles-pp.ggc_internal_alloc
0.00 +0.2 0.19 ±141% perf-profile.calltrace.cycles-pp.remove_wait_queue.poll_freewait.do_select.core_sys_select.do_pselect
0.00 +0.2 0.19 ±141% perf-profile.calltrace.cycles-pp.poll_freewait.do_select.core_sys_select.do_pselect.__x64_sys_pselect6
0.33 ± 70% +0.2 0.53 ± 5% perf-profile.calltrace.cycles-pp.load_elf_binary.search_binary_handler.exec_binprm.bprm_execve.do_execveat_common
3.59 ± 2% +0.2 3.80 perf-profile.calltrace.cycles-pp.htab_find_slot_with_hash
0.18 ±141% +0.2 0.40 ± 73% perf-profile.calltrace.cycles-pp.do_select.core_sys_select.do_pselect.__x64_sys_pselect6.do_syscall_64
0.19 ±141% +0.2 0.41 ± 72% perf-profile.calltrace.cycles-pp.__x64_sys_pselect6.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.19 ±141% +0.2 0.41 ± 72% perf-profile.calltrace.cycles-pp.do_pselect.__x64_sys_pselect6.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.19 ±141% +0.2 0.41 ± 72% perf-profile.calltrace.cycles-pp.core_sys_select.do_pselect.__x64_sys_pselect6.do_syscall_64.entry_SYSCALL_64_after_hwframe
1.02 +0.2 1.25 ± 3% perf-profile.calltrace.cycles-pp._cpp_lex_direct
0.00 +0.5 0.53 perf-profile.calltrace.cycles-pp.wi::force_to_size
0.00 +0.7 0.68 perf-profile.calltrace.cycles-pp.clear_page_erms.clear_huge_page.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault
0.00 +0.8 0.77 perf-profile.calltrace.cycles-pp.clear_huge_page.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
0.00 +0.8 0.79 perf-profile.calltrace.cycles-pp.__do_huge_pmd_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
6.09 +0.9 7.03 perf-profile.calltrace.cycles-pp.c_lex_with_flags
9.49 -5.8 3.72 ± 2% perf-profile.children.cycles-pp.do_anonymous_page
13.72 -4.3 9.38 perf-profile.children.cycles-pp.asm_exc_page_fault
12.62 -3.8 8.87 perf-profile.children.cycles-pp.exc_page_fault
12.57 -3.7 8.84 perf-profile.children.cycles-pp.do_user_addr_fault
11.96 -3.5 8.46 perf-profile.children.cycles-pp.handle_mm_fault
5.55 -3.3 2.22 ± 2% perf-profile.children.cycles-pp.vma_alloc_folio
11.51 -3.3 8.20 perf-profile.children.cycles-pp.__handle_mm_fault
5.59 -3.3 2.33 ± 2% perf-profile.children.cycles-pp.__alloc_pages
5.38 -3.3 2.12 ± 2% perf-profile.children.cycles-pp.__folio_alloc
5.26 -3.1 2.12 ± 2% perf-profile.children.cycles-pp.get_page_from_freelist
13.50 ± 2% -2.0 11.47 ± 2% perf-profile.children.cycles-pp.do_syscall_64
13.53 -2.0 11.50 ± 2% perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
3.18 ± 4% -2.0 1.19 ± 15% perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
2.51 ± 2% -1.6 0.87 ± 2% perf-profile.children.cycles-pp.rmqueue
3.27 ± 2% -1.6 1.67 ± 3% perf-profile.children.cycles-pp.do_exit
3.10 ± 2% -1.6 1.50 ± 4% perf-profile.children.cycles-pp.exit_mm
3.27 ± 2% -1.6 1.67 ± 3% perf-profile.children.cycles-pp.__x64_sys_exit_group
3.27 ± 2% -1.6 1.67 ± 3% perf-profile.children.cycles-pp.do_group_exit
3.20 ± 2% -1.6 1.63 ± 3% perf-profile.children.cycles-pp.__mmput
3.20 ± 2% -1.6 1.63 ± 3% perf-profile.children.cycles-pp.exit_mmap
2.38 ± 3% -1.5 0.89 ± 6% perf-profile.children.cycles-pp._raw_spin_lock
2.75 -1.5 1.26 perf-profile.children.cycles-pp.__mem_cgroup_charge
2.38 ± 2% -1.4 0.94 ± 2% perf-profile.children.cycles-pp.zap_pte_range
2.39 ± 2% -1.4 0.96 ± 2% perf-profile.children.cycles-pp.zap_pmd_range
2.40 ± 2% -1.4 0.98 ± 2% perf-profile.children.cycles-pp.unmap_page_range
2.43 ± 2% -1.4 1.01 ± 2% perf-profile.children.cycles-pp.unmap_vmas
1.91 ± 3% -1.3 0.58 ± 3% perf-profile.children.cycles-pp.rmqueue_bulk
2.27 ± 2% -1.2 1.08 ± 4% perf-profile.children.cycles-pp.tlb_batch_pages_flush
2.21 ± 2% -1.2 1.05 ± 4% perf-profile.children.cycles-pp.release_pages
2.08 -1.1 1.00 perf-profile.children.cycles-pp.charge_memcg
1.35 ± 2% -0.8 0.58 ± 4% perf-profile.children.cycles-pp.free_unref_page_list
1.23 ± 3% -0.7 0.56 ± 3% perf-profile.children.cycles-pp.free_pcppages_bulk
1.82 -0.6 1.21 perf-profile.children.cycles-pp.make_node
1.34 ± 7% -0.6 0.75 ± 26% perf-profile.children.cycles-pp._raw_spin_lock_irqsave
2.17 ± 2% -0.6 1.59 perf-profile.children.cycles-pp.build_decl
0.93 -0.6 0.38 ± 2% perf-profile.children.cycles-pp.folio_batch_move_lru
0.91 -0.6 0.36 ± 2% perf-profile.children.cycles-pp.folio_add_lru
0.99 -0.5 0.48 perf-profile.children.cycles-pp.try_charge_memcg
1.03 -0.5 0.52 perf-profile.children.cycles-pp.__list_del_entry_valid
0.57 ± 2% -0.4 0.13 ± 3% perf-profile.children.cycles-pp.folio_lruvec_lock_irqsave
0.76 -0.4 0.40 ± 3% perf-profile.children.cycles-pp.__free_one_page
0.88 -0.3 0.55 ± 4% perf-profile.children.cycles-pp.error_entry
0.51 ± 2% -0.3 0.18 ± 4% perf-profile.children.cycles-pp.get_mem_cgroup_from_mm
0.80 -0.3 0.50 ± 4% perf-profile.children.cycles-pp.sync_regs
1.32 ± 2% -0.3 1.05 ± 5% perf-profile.children.cycles-pp.tlb_finish_mmu
0.68 -0.3 0.42 ± 2% perf-profile.children.cycles-pp.native_irq_return_iret
2.61 -0.2 2.36 perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
2.44 -0.2 2.20 perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
1.87 -0.2 1.63 perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
1.84 -0.2 1.61 perf-profile.children.cycles-pp.hrtimer_interrupt
0.44 ± 2% -0.2 0.22 ± 3% perf-profile.children.cycles-pp.__mod_lruvec_page_state
0.49 -0.2 0.28 ± 3% perf-profile.children.cycles-pp.tree_cons
0.33 ± 2% -0.2 0.15 ± 3% perf-profile.children.cycles-pp.page_add_new_anon_rmap
1.57 -0.2 1.39 perf-profile.children.cycles-pp.__hrtimer_run_queues
0.36 -0.2 0.18 ± 2% perf-profile.children.cycles-pp.lru_add_fn
1.18 ± 2% -0.2 1.00 ± 2% perf-profile.children.cycles-pp.__do_munmap
2.74 ± 2% -0.2 2.58 ± 3% perf-profile.children.cycles-pp.elf_read
0.92 ± 3% -0.2 0.75 ± 2% perf-profile.children.cycles-pp.unmap_region
0.83 ± 12% -0.2 0.67 ± 33% perf-profile.children.cycles-pp.core_sys_select
0.84 ± 12% -0.2 0.67 ± 33% perf-profile.children.cycles-pp.__x64_sys_pselect6
0.84 ± 12% -0.2 0.67 ± 33% perf-profile.children.cycles-pp.do_pselect
0.82 ± 12% -0.2 0.66 ± 34% perf-profile.children.cycles-pp.do_select
0.72 ± 13% -0.2 0.57 ± 35% perf-profile.children.cycles-pp.poll_freewait
0.70 ± 12% -0.2 0.55 ± 35% perf-profile.children.cycles-pp.remove_wait_queue
0.46 -0.2 0.31 ± 4% perf-profile.children.cycles-pp.page_remove_rmap
0.28 ± 2% -0.2 0.13 perf-profile.children.cycles-pp.__mod_lruvec_state
0.33 ± 2% -0.2 0.18 perf-profile.children.cycles-pp.cgraph_node::create
0.35 ± 18% -0.1 0.21 ± 15% perf-profile.children.cycles-pp.pselect
0.91 ± 2% -0.1 0.76 ± 2% perf-profile.children.cycles-pp.__vm_munmap
0.78 ± 3% -0.1 0.64 ± 2% perf-profile.children.cycles-pp.__munmap
0.85 ± 2% -0.1 0.71 ± 3% perf-profile.children.cycles-pp.__x64_sys_munmap
1.29 ± 2% -0.1 1.16 ± 2% perf-profile.children.cycles-pp.tick_sched_timer
0.60 -0.1 0.47 perf-profile.children.cycles-pp.allocate_struct_function
1.17 ± 2% -0.1 1.04 perf-profile.children.cycles-pp.tick_sched_handle
0.23 ± 2% -0.1 0.10 perf-profile.children.cycles-pp.__mod_node_page_state
0.67 -0.1 0.54 perf-profile.children.cycles-pp.build_function_type
2.64 -0.1 2.52 perf-profile.children.cycles-pp.do_filp_open
2.63 -0.1 2.51 perf-profile.children.cycles-pp.path_openat
1.15 -0.1 1.03 perf-profile.children.cycles-pp.update_process_times
0.84 -0.1 0.72 perf-profile.children.cycles-pp.scheduler_tick
0.57 -0.1 0.45 ± 2% perf-profile.children.cycles-pp.linemap_enter_macro
0.65 -0.1 0.54 perf-profile.children.cycles-pp.task_tick_fair
0.31 -0.1 0.20 ± 2% perf-profile.children.cycles-pp.__perf_sw_event
1.78 -0.1 1.69 perf-profile.children.cycles-pp.type_hash_canon
0.37 -0.1 0.28 perf-profile.children.cycles-pp.build1
1.41 ± 4% -0.1 1.32 perf-profile.children.cycles-pp.read
2.90 -0.1 2.82 perf-profile.children.cycles-pp.__x64_sys_openat
2.89 -0.1 2.81 perf-profile.children.cycles-pp.do_sys_openat2
0.24 -0.1 0.16 ± 2% perf-profile.children.cycles-pp.___perf_sw_event
0.15 -0.1 0.07 ± 6% perf-profile.children.cycles-pp.free_pages_and_swap_cache
0.40 ± 3% -0.1 0.32 perf-profile.children.cycles-pp.build_call_array_loc
0.07 ± 6% -0.1 0.00 perf-profile.children.cycles-pp.mem_cgroup_charge_statistics
0.07 ± 12% -0.1 0.00 perf-profile.children.cycles-pp.__cgroup_throttle_swaprate
0.26 ± 4% -0.1 0.18 ± 5% perf-profile.children.cycles-pp.update_cfs_group
0.20 ± 2% -0.1 0.13 ± 3% perf-profile.children.cycles-pp.__mod_memcg_lruvec_state
0.14 -0.1 0.07 ± 11% perf-profile.children.cycles-pp.free_swap_cache
0.13 -0.1 0.06 perf-profile.children.cycles-pp.page_counter_try_charge
0.07 -0.1 0.00 perf-profile.children.cycles-pp.mem_cgroup_update_lru_size
0.07 -0.1 0.00 perf-profile.children.cycles-pp.__mod_zone_page_state
0.07 -0.1 0.00 perf-profile.children.cycles-pp.free_unref_page_commit
1.45 -0.1 1.38 ± 2% perf-profile.children.cycles-pp.link_path_walk
0.56 ± 2% -0.1 0.50 ± 3% perf-profile.children.cycles-pp.wp_page_copy
0.51 ± 2% -0.1 0.44 perf-profile.children.cycles-pp.lookup_name
1.53 ± 7% -0.1 1.47 ± 8% perf-profile.children.cycles-pp.vfs_read
1.05 -0.1 0.99 ± 2% perf-profile.children.cycles-pp.walk_component
0.06 -0.1 0.00 perf-profile.children.cycles-pp.__page_set_anon_rmap
0.15 ± 5% -0.1 0.09 ± 9% perf-profile.children.cycles-pp.__list_add_valid
0.13 ± 3% -0.1 0.07 ± 6% perf-profile.children.cycles-pp.build_string
0.30 ± 2% -0.1 0.24 ± 5% perf-profile.children.cycles-pp.add_stmt
0.06 ± 8% -0.1 0.00 perf-profile.children.cycles-pp.down_read_trylock
0.06 ± 8% -0.1 0.00 perf-profile.children.cycles-pp.blk_cgroup_congested
0.12 -0.1 0.06 ± 7% perf-profile.children.cycles-pp.build_tree_list
1.87 -0.1 1.81 perf-profile.children.cycles-pp.open64
1.53 ± 7% -0.1 1.47 ± 8% perf-profile.children.cycles-pp.ksys_read
0.47 ± 2% -0.1 0.42 perf-profile.children.cycles-pp.build2
0.41 ± 2% -0.1 0.36 ± 2% perf-profile.children.cycles-pp.pushdecl
0.05 -0.1 0.00 perf-profile.children.cycles-pp.security_file_open
0.05 -0.1 0.00 perf-profile.children.cycles-pp.output_addr_const
0.05 -0.1 0.00 perf-profile.children.cycles-pp.apparmor_file_open
0.05 -0.1 0.00 perf-profile.children.cycles-pp.enqueue_hrtimer
0.05 -0.1 0.00 perf-profile.children.cycles-pp.errseq_sample
0.05 -0.1 0.00 perf-profile.children.cycles-pp.uncharge_folio
0.48 ± 2% -0.0 0.43 ± 3% perf-profile.children.cycles-pp.dput
0.12 ± 4% -0.0 0.07 ± 7% perf-profile.children.cycles-pp._raw_spin_unlock_irqrestore
0.11 ± 4% -0.0 0.07 ± 7% perf-profile.children.cycles-pp.cgroup_rstat_updated
0.48 ± 2% -0.0 0.44 perf-profile.children.cycles-pp.exit_to_user_mode_loop
0.55 ± 2% -0.0 0.51 ± 2% perf-profile.children.cycles-pp.exit_to_user_mode_prepare
1.22 ± 3% -0.0 1.18 perf-profile.children.cycles-pp.filemap_read
0.10 ± 9% -0.0 0.06 ± 8% perf-profile.children.cycles-pp.__count_memcg_events
0.59 -0.0 0.55 ± 3% perf-profile.children.cycles-pp.d_alloc_parallel
0.11 -0.0 0.07 ± 6% perf-profile.children.cycles-pp.vmacache_find
0.36 ± 2% -0.0 0.32 ± 2% perf-profile.children.cycles-pp.task_work_run
0.20 ± 51% -0.0 0.16 ± 91% perf-profile.children.cycles-pp.__mutex_lock
0.15 ± 3% -0.0 0.11 ± 25% perf-profile.children.cycles-pp.__clone
0.05 -0.0 0.02 ±141% perf-profile.children.cycles-pp.copy_p4d_range
0.07 ± 7% -0.0 0.03 ± 70% perf-profile.children.cycles-pp.__pte_alloc
0.03 ± 70% -0.0 0.00 perf-profile.children.cycles-pp.create_tmp_var_raw
0.03 ± 70% -0.0 0.00 perf-profile.children.cycles-pp.__tlb_remove_page_size
0.03 ± 70% -0.0 0.00 perf-profile.children.cycles-pp.rwsem_spin_on_owner
0.03 ± 70% -0.0 0.00 perf-profile.children.cycles-pp.__update_load_avg_se
0.91 ± 3% -0.0 0.88 perf-profile.children.cycles-pp._copy_to_iter
0.87 ± 3% -0.0 0.84 perf-profile.children.cycles-pp.copyout
0.34 -0.0 0.31 perf-profile.children.cycles-pp.ht_lookup_with_hash
0.16 ± 5% -0.0 0.13 perf-profile.children.cycles-pp.security_file_free
0.16 ± 3% -0.0 0.13 ± 3% perf-profile.children.cycles-pp.apparmor_file_free_security
0.63 -0.0 0.61 ± 3% perf-profile.children.cycles-pp.__lookup_slow
0.18 -0.0 0.15 ± 3% perf-profile.children.cycles-pp.find_vma
0.18 ± 10% -0.0 0.16 ± 3% perf-profile.children.cycles-pp.copy_node
0.27 ± 4% -0.0 0.24 ± 3% perf-profile.children.cycles-pp.__fput
0.24 ± 3% -0.0 0.21 ± 3% perf-profile.children.cycles-pp.dentry_kill
0.18 ± 2% -0.0 0.15 perf-profile.children.cycles-pp.irqentry_exit_to_user_mode
0.17 ± 2% -0.0 0.15 ± 3% perf-profile.children.cycles-pp.lockref_put_or_lock
0.13 -0.0 0.10 ± 4% perf-profile.children.cycles-pp.ggc_internal_cleared_alloc
0.09 ± 5% -0.0 0.07 ± 7% perf-profile.children.cycles-pp.shallow_copy_rtx
0.09 ± 5% -0.0 0.06 perf-profile.children.cycles-pp.lru_add_drain
0.08 -0.0 0.05 ± 8% perf-profile.children.cycles-pp.__mem_cgroup_uncharge_list
0.08 -0.0 0.05 ± 8% perf-profile.children.cycles-pp.__irqentry_text_end
0.06 ± 13% -0.0 0.03 ± 70% perf-profile.children.cycles-pp.__lock_parent
0.10 ± 4% -0.0 0.08 ± 6% perf-profile.children.cycles-pp._raw_spin_trylock
0.25 ± 3% -0.0 0.22 ± 2% perf-profile.children.cycles-pp.security_file_alloc
0.96 ± 2% -0.0 0.93 perf-profile.children.cycles-pp.copy_user_enhanced_fast_string
0.92 ± 3% -0.0 0.89 perf-profile.children.cycles-pp.copy_page_to_iter
2.57 -0.0 2.54 ± 2% perf-profile.children.cycles-pp._cpp_create_definition
0.25 -0.0 0.23 ± 3% perf-profile.children.cycles-pp.vma_interval_tree_insert
0.19 -0.0 0.17 ± 2% perf-profile.children.cycles-pp.build_stmt
0.07 ± 6% -0.0 0.05 ± 74% perf-profile.children.cycles-pp.target_environment
0.19 ± 2% -0.0 0.16 ± 2% perf-profile.children.cycles-pp.apparmor_file_alloc_security
0.23 ± 45% -0.0 0.20 ± 67% perf-profile.children.cycles-pp.pipe_read
0.21 ± 47% -0.0 0.19 ± 69% perf-profile.children.cycles-pp.osq_lock
0.07 ± 6% -0.0 0.05 perf-profile.children.cycles-pp.ktime_get_update_offsets_now
0.07 ± 6% -0.0 0.05 perf-profile.children.cycles-pp.up_read
0.06 ± 8% -0.0 0.03 ± 70% perf-profile.children.cycles-pp.timerqueue_del
0.06 -0.0 0.04 ± 71% perf-profile.children.cycles-pp.__wake_up_common
0.07 ± 25% -0.0 0.04 ± 71% perf-profile.children.cycles-pp.run_builtin
0.07 ± 20% -0.0 0.05 ± 72% perf-profile.children.cycles-pp.__libc_start_main
0.65 -0.0 0.63 perf-profile.children.cycles-pp.__alloc_file
0.36 ± 2% -0.0 0.33 ± 3% perf-profile.children.cycles-pp.step_into
0.67 -0.0 0.65 perf-profile.children.cycles-pp.alloc_empty_file
0.45 -0.0 0.43 perf-profile.children.cycles-pp.lookup_fast
0.43 -0.0 0.41 perf-profile.children.cycles-pp.d_alloc
0.32 -0.0 0.30 ± 3% perf-profile.children.cycles-pp.__vma_adjust
0.30 ± 8% -0.0 0.28 ± 9% perf-profile.children.cycles-pp.check
0.19 ± 2% -0.0 0.17 ± 2% perf-profile.children.cycles-pp.update_load_avg
0.07 ± 6% -0.0 0.05 ± 8% perf-profile.children.cycles-pp.make_ssa_name_fn
0.08 -0.0 0.06 perf-profile.children.cycles-pp.lru_add_drain_cpu
0.06 ± 19% -0.0 0.04 ± 71% perf-profile.children.cycles-pp.cmd_record
0.06 ± 19% -0.0 0.04 ± 71% perf-profile.children.cycles-pp.__cmd_record
0.10 ± 8% -0.0 0.08 ± 10% perf-profile.children.cycles-pp.rtx_alloc
0.08 ± 5% -0.0 0.06 ± 7% perf-profile.children.cycles-pp.rwsem_down_write_slowpath
0.24 -0.0 0.22 ± 2% perf-profile.children.cycles-pp.fold_build2_loc
0.28 -0.0 0.27 perf-profile.children.cycles-pp.do_open
0.19 -0.0 0.17 ± 2% perf-profile.children.cycles-pp.do_dentry_open
0.18 ± 2% -0.0 0.17 ± 5% perf-profile.children.cycles-pp.select_task_rq_fair
0.13 -0.0 0.11 ± 4% perf-profile.children.cycles-pp.lockref_get_not_dead
0.13 ± 3% -0.0 0.11 perf-profile.children.cycles-pp.build_pointer_type_for_mode
0.05 -0.0 0.03 ± 70% perf-profile.children.cycles-pp.arch_scale_freq_tick
0.05 -0.0 0.03 ± 70% perf-profile.children.cycles-pp.user_path_at_empty
0.05 -0.0 0.03 ± 70% perf-profile.children.cycles-pp._cpp_parse_expr
0.05 -0.0 0.03 ± 70% perf-profile.children.cycles-pp.string_concat_db::record_string_concatenation
0.03 ± 70% -0.0 0.02 ±141% perf-profile.children.cycles-pp.mm_init
0.02 ±141% -0.0 0.00 perf-profile.children.cycles-pp.build4
0.02 ±141% -0.0 0.00 perf-profile.children.cycles-pp.policy_node
0.03 ± 70% -0.0 0.02 ±141% perf-profile.children.cycles-pp.enqueue_task_fair
0.02 ±141% -0.0 0.00 perf-profile.children.cycles-pp.record__pushfn
0.02 ±141% -0.0 0.00 perf-profile.children.cycles-pp.__libc_write
0.03 ± 70% -0.0 0.02 ±141% perf-profile.children.cycles-pp.up_write
0.02 ±141% -0.0 0.00 perf-profile.children.cycles-pp.memcg_check_events
0.02 ±141% -0.0 0.00 perf-profile.children.cycles-pp.make_edge
0.02 ±141% -0.0 0.00 perf-profile.children.cycles-pp.plus_constant
0.02 ±141% -0.0 0.00 perf-profile.children.cycles-pp.split_edge
0.02 ±141% -0.0 0.00 perf-profile.children.cycles-pp.scev_initialize
0.02 ±141% -0.0 0.00 perf-profile.children.cycles-pp.bitmap_ior_into
0.35 -0.0 0.34 ± 3% perf-profile.children.cycles-pp.next_uptodate_page
0.14 ± 3% -0.0 0.13 ± 7% perf-profile.children.cycles-pp.update_sg_wakeup_stats
0.14 ± 3% -0.0 0.13 ± 3% perf-profile.children.cycles-pp.__legitimize_path
0.08 -0.0 0.06 ± 7% perf-profile.children.cycles-pp.c_add_case_label
0.15 ± 3% -0.0 0.14 ± 3% perf-profile.children.cycles-pp.try_to_unlazy
0.15 ± 3% -0.0 0.14 ± 6% perf-profile.children.cycles-pp.find_idlest_group
0.40 ± 3% -0.0 0.38 ± 2% perf-profile.children.cycles-pp.syscall_exit_to_user_mode
0.33 ± 2% -0.0 0.31 perf-profile.children.cycles-pp.__split_vma
0.25 -0.0 0.24 ± 3% perf-profile.children.cycles-pp.__schedule
0.18 -0.0 0.17 ± 2% perf-profile.children.cycles-pp.update_curr
0.07 -0.0 0.06 ± 8% perf-profile.children.cycles-pp.handle_pte_fault
0.17 ± 7% -0.0 0.16 ± 5% perf-profile.children.cycles-pp.vfs_write
0.16 ± 7% -0.0 0.15 ± 5% perf-profile.children.cycles-pp.ksys_write
0.13 ± 3% -0.0 0.12 perf-profile.children.cycles-pp.build_pointer_type
0.12 ± 3% -0.0 0.11 perf-profile.children.cycles-pp.terminate_walk
0.11 ± 4% -0.0 0.09 ± 5% perf-profile.children.cycles-pp.__might_sleep
0.11 ± 4% -0.0 0.09 ± 5% perf-profile.children.cycles-pp.perf_mux_hrtimer_handler
0.08 -0.0 0.07 ± 14% perf-profile.children.cycles-pp.irqtime_account_irq
0.08 ± 6% -0.0 0.06 ± 7% perf-profile.children.cycles-pp.rwsem_optimistic_spin
0.10 ± 9% -0.0 0.09 ± 9% perf-profile.children.cycles-pp.generic_perform_write
0.69 -0.0 0.68 perf-profile.children.cycles-pp.mmap_region
0.32 -0.0 0.31 perf-profile.children.cycles-pp.__d_lookup_rcu
0.16 -0.0 0.15 ± 6% perf-profile.children.cycles-pp.find_idlest_cpu
0.22 ± 2% -0.0 0.21 ± 2% perf-profile.children.cycles-pp.schedule
0.21 ± 2% -0.0 0.20 ± 6% perf-profile.children.cycles-pp.copy_page
0.19 ± 2% -0.0 0.18 ± 2% perf-profile.children.cycles-pp.elf_map
0.17 ± 2% -0.0 0.16 ± 2% perf-profile.children.cycles-pp.filemap_get_pages
0.17 ± 5% -0.0 0.16 ± 2% perf-profile.children.cycles-pp.build_qualified_type
0.16 ± 2% -0.0 0.15 ± 3% perf-profile.children.cycles-pp.filemap_get_read_batch
0.14 -0.0 0.13 perf-profile.children.cycles-pp.__dentry_kill
0.13 -0.0 0.12 perf-profile.children.cycles-pp.build_external_ref
0.11 ± 7% -0.0 0.10 ± 8% perf-profile.children.cycles-pp.generic_file_write_iter
0.07 -0.0 0.06 perf-profile.children.cycles-pp.update_type_inheritance_graph
0.07 ± 7% -0.0 0.06 ± 8% perf-profile.children.cycles-pp.ktime_get
0.12 -0.0 0.11 perf-profile.children.cycles-pp.load_elf_interp
0.06 -0.0 0.05 perf-profile.children.cycles-pp._cpp_stack_include
0.06 -0.0 0.05 perf-profile.children.cycles-pp.update_rq_clock
0.46 -0.0 0.45 perf-profile.children.cycles-pp.ksys_mmap_pgoff
0.21 ± 3% -0.0 0.20 perf-profile.children.cycles-pp.__close
0.23 ± 3% -0.0 0.22 ± 5% perf-profile.children.cycles-pp.elf_create_section
0.10 -0.0 0.09 ± 5% perf-profile.children.cycles-pp.sched_exec
0.11 ± 4% -0.0 0.10 ± 8% perf-profile.children.cycles-pp.__generic_file_write_iter
0.09 -0.0 0.08 ± 5% perf-profile.children.cycles-pp.push_stmt_list
0.09 -0.0 0.08 ± 5% perf-profile.children.cycles-pp.build_type_inheritance_graph
0.07 ± 6% -0.0 0.07 ± 7% perf-profile.children.cycles-pp.down_read
0.07 ± 6% -0.0 0.07 ± 7% perf-profile.children.cycles-pp.hard_function_value
0.10 ± 4% -0.0 0.09 perf-profile.children.cycles-pp.rcu_sched_clock_irq
0.09 ± 5% -0.0 0.08 perf-profile.children.cycles-pp.write
0.07 -0.0 0.06 ± 7% perf-profile.children.cycles-pp.__mmdrop
0.07 -0.0 0.06 ± 7% perf-profile.children.cycles-pp.vma_merge
0.08 ± 6% -0.0 0.07 perf-profile.children.cycles-pp.vfs_fstat
0.07 -0.0 0.06 ± 7% perf-profile.children.cycles-pp.finish_task_switch
0.07 -0.0 0.06 ± 7% perf-profile.children.cycles-pp.do_wait
0.06 -0.0 0.05 ± 8% perf-profile.children.cycles-pp._raw_spin_lock_irq
0.06 -0.0 0.05 ± 8% perf-profile.children.cycles-pp.get_zeroed_page
0.43 -0.0 0.42 ± 4% perf-profile.children.cycles-pp.kernel_clone
0.34 -0.0 0.33 ± 4% perf-profile.children.cycles-pp.__do_sys_clone
0.26 ± 3% -0.0 0.25 perf-profile.children.cycles-pp.open_last_lookups
0.16 ± 3% -0.0 0.15 ± 5% perf-profile.children.cycles-pp.__slab_free
0.13 ± 3% -0.0 0.13 ± 3% perf-profile.children.cycles-pp.unlink_file_vma
0.08 ± 11% -0.0 0.08 ± 6% perf-profile.children.cycles-pp.tick_sched_do_timer
0.08 ± 5% -0.0 0.08 ± 6% perf-profile.children.cycles-pp.clockevents_program_event
0.18 -0.0 0.18 ± 2% perf-profile.children.cycles-pp.vma_link
0.12 ± 3% -0.0 0.12 perf-profile.children.cycles-pp.__get_user_pages_remote
0.12 ± 3% -0.0 0.12 perf-profile.children.cycles-pp.__get_user_pages
0.10 -0.0 0.10 ± 4% perf-profile.children.cycles-pp.vma_interval_tree_remove
0.10 -0.0 0.10 ± 4% perf-profile.children.cycles-pp.try_to_wake_up
0.10 ± 4% -0.0 0.09 ± 5% perf-profile.children.cycles-pp.__x64_sys_vfork
0.10 ± 4% -0.0 0.09 ± 5% perf-profile.children.cycles-pp.wake_up_new_task
0.07 ± 6% -0.0 0.07 perf-profile.children.cycles-pp.check_cpu_stall
0.07 ± 6% -0.0 0.07 perf-profile.children.cycles-pp.slab_pre_alloc_hook
0.07 ± 6% -0.0 0.07 perf-profile.children.cycles-pp.__memcg_kmem_charge_page
0.08 ± 10% -0.0 0.08 ± 16% perf-profile.children.cycles-pp.__wake_up_common_lock
0.07 -0.0 0.07 ± 7% perf-profile.children.cycles-pp.finish_record_layout
0.07 -0.0 0.07 ± 7% perf-profile.children.cycles-pp.do_faccessat
0.07 -0.0 0.07 ± 7% perf-profile.children.cycles-pp.copy_string_kernel
0.07 -0.0 0.07 ± 7% perf-profile.children.cycles-pp.finish_declspecs
0.06 ± 7% -0.0 0.06 perf-profile.children.cycles-pp.may_open
0.07 ± 6% -0.0 0.07 ± 11% perf-profile.children.cycles-pp.get_alias_set
0.06 ± 7% -0.0 0.06 perf-profile.children.cycles-pp.cgraph_edge::rebuild_edges
0.06 -0.0 0.06 ± 8% perf-profile.children.cycles-pp.touch_atime
0.02 ±141% -0.0 0.02 ±141% perf-profile.children.cycles-pp.record__mmap_read_evlist
0.17 ± 2% -0.0 0.16 ± 2% perf-profile.children.cycles-pp.__x64_sys_mprotect
0.13 -0.0 0.13 ± 3% perf-profile.children.cycles-pp.get_arg_page
0.12 ± 8% -0.0 0.11 ± 4% perf-profile.children.cycles-pp.cleanup_tree_cfg
0.08 ± 6% -0.0 0.07 ± 6% perf-profile.children.cycles-pp.wait4
0.05 ± 8% -0.0 0.05 perf-profile.children.cycles-pp.xas_start
0.06 ± 8% -0.0 0.05 ± 8% perf-profile.children.cycles-pp.__get_free_pages
0.06 ± 8% -0.0 0.05 ± 8% perf-profile.children.cycles-pp.cgraph_node::create_edge
0.05 ± 8% -0.0 0.05 perf-profile.children.cycles-pp.alloc_bprm
0.05 ± 8% -0.0 0.05 perf-profile.children.cycles-pp.gimple_copy
0.27 -0.0 0.26 perf-profile.children.cycles-pp.__do_sys_newstat
0.26 -0.0 0.25 ± 3% perf-profile.children.cycles-pp.set_source_range
0.31 ± 2% +0.0 0.31 ± 4% perf-profile.children.cycles-pp.copy_process
0.29 +0.0 0.29 perf-profile.children.cycles-pp.build_function_call_vec
0.24 ± 3% +0.0 0.24 ± 3% perf-profile.children.cycles-pp.run_ksoftirqd
0.24 ± 3% +0.0 0.24 ± 3% perf-profile.children.cycles-pp.free_pgtables
0.22 ± 3% +0.0 0.22 perf-profile.children.cycles-pp.fold_convert_loc
0.16 ± 2% +0.0 0.16 ± 2% perf-profile.children.cycles-pp.do_mprotect_pkey
0.14 +0.0 0.14 ± 5% perf-profile.children.cycles-pp.mprotect_fixup
0.14 ± 12% +0.0 0.14 ± 9% perf-profile.children.cycles-pp.main
0.11 +0.0 0.11 perf-profile.children.cycles-pp.perf_iterate_sb
0.11 +0.0 0.11 perf-profile.children.cycles-pp.__do_sys_newfstat
0.10 ± 4% +0.0 0.10 ± 4% perf-profile.children.cycles-pp.setlocale
0.09 +0.0 0.09 perf-profile.children.cycles-pp.pte_alloc_one
0.09 +0.0 0.09 perf-profile.children.cycles-pp.inode_permission
0.09 +0.0 0.09 perf-profile.children.cycles-pp.sizeof_pointer_memaccess_warning
0.08 ± 5% +0.0 0.08 ± 5% perf-profile.children.cycles-pp.__xstat64
0.08 +0.0 0.08 perf-profile.children.cycles-pp.warning_at
0.08 ± 5% +0.0 0.08 ± 5% perf-profile.children.cycles-pp.dup_task_struct
0.08 ± 6% +0.0 0.08 ± 6% perf-profile.children.cycles-pp.flush_tlb_mm_range
0.07 ± 6% +0.0 0.07 ± 6% perf-profile.children.cycles-pp.compute_fn_summary
0.07 +0.0 0.07 ± 11% perf-profile.children.cycles-pp.kernel_wait4
0.08 ± 10% +0.0 0.08 perf-profile.children.cycles-pp.execute_fixup_cfg
0.08 ± 12% +0.0 0.08 ± 6% perf-profile.children.cycles-pp.compute_may_aliases
0.07 +0.0 0.07 perf-profile.children.cycles-pp.mark_dfs_back_edges
0.06 +0.0 0.06 perf-profile.children.cycles-pp.down_write
0.06 +0.0 0.06 perf-profile.children.cycles-pp.mark_exp_read
0.05 ± 8% +0.0 0.05 ± 8% perf-profile.children.cycles-pp.lookup_open
0.06 +0.0 0.06 perf-profile.children.cycles-pp.finish_function
0.06 ± 8% +0.0 0.06 ± 8% perf-profile.children.cycles-pp.push_scope
0.06 +0.0 0.06 perf-profile.children.cycles-pp.__do_sys_brk
0.06 ± 7% +0.0 0.06 ± 7% perf-profile.children.cycles-pp.et_below
0.06 ± 8% +0.0 0.06 ± 8% perf-profile.children.cycles-pp.__close_nocancel
0.06 +0.0 0.06 perf-profile.children.cycles-pp.split_all_insns
0.05 ± 8% +0.0 0.05 ± 8% perf-profile.children.cycles-pp.finish_fault
0.05 ± 8% +0.0 0.05 ± 8% perf-profile.children.cycles-pp.ptep_clear_flush
0.05 +0.0 0.05 perf-profile.children.cycles-pp.force_fit_type
0.05 ± 8% +0.0 0.05 ± 8% perf-profile.children.cycles-pp.alloc_thread_stack_node
0.05 ± 8% +0.0 0.05 ± 8% perf-profile.children.cycles-pp.may_trap_p_1
0.05 +0.0 0.05 perf-profile.children.cycles-pp.fput
0.05 +0.0 0.05 perf-profile.children.cycles-pp.c_parser_next_tokens_start_declaration
0.05 +0.0 0.05 perf-profile.children.cycles-pp.xas_find
0.05 +0.0 0.05 perf-profile.children.cycles-pp.compute_record_mode
0.05 +0.0 0.05 perf-profile.children.cycles-pp.copy_page_range
0.02 ±141% +0.0 0.02 ±141% perf-profile.children.cycles-pp.folio_memcg_lock
0.05 +0.0 0.05 perf-profile.children.cycles-pp.pick_next_task_fair
0.05 +0.0 0.05 perf-profile.children.cycles-pp.do_brk_flags
0.02 ±141% +0.0 0.02 ±141% perf-profile.children.cycles-pp.c_finish_return
0.05 +0.0 0.05 perf-profile.children.cycles-pp.rest_of_decl_compilation
0.05 +0.0 0.05 perf-profile.children.cycles-pp.filp_close
0.05 +0.0 0.05 perf-profile.children.cycles-pp.find_comparison_dom_walker::before_dom_children
0.05 +0.0 0.05 perf-profile.children.cycles-pp.shorten_binary_op
0.04 ± 71% +0.0 0.04 ± 71% perf-profile.children.cycles-pp.pipe_write
0.02 ±141% +0.0 0.02 ±141% perf-profile.children.cycles-pp.perf_mmap__push
0.03 ± 70% +0.0 0.03 ± 70% perf-profile.children.cycles-pp.useless_type_conversion_p
0.02 ±141% +0.0 0.02 ±141% perf-profile.children.cycles-pp.pipe_poll
0.02 ±141% +0.0 0.02 ±141% perf-profile.children.cycles-pp.schedule_hrtimeout_range_clock
0.02 ±141% +0.0 0.02 ±141% perf-profile.children.cycles-pp.validate_branch
0.03 ± 70% +0.0 0.03 ± 70% perf-profile.children.cycles-pp.set_reg_attrs_for_decl_rtl
0.02 ±141% +0.0 0.02 ±141% perf-profile.children.cycles-pp.update_ssa
0.10 ± 8% +0.0 0.10 perf-profile.children.cycles-pp.xas_load
0.10 ± 8% +0.0 0.10 perf-profile.children.cycles-pp.finish_struct
0.29 +0.0 0.29 ± 2% perf-profile.children.cycles-pp.vfs_fstatat
0.27 ± 4% +0.0 0.28 ± 3% perf-profile.children.cycles-pp.smpboot_thread_fn
0.10 +0.0 0.10 ± 4% perf-profile.children.cycles-pp.__vfork
0.10 ± 4% +0.0 0.10 ± 8% perf-profile.children.cycles-pp.ira_init_register_move_cost
0.24 +0.0 0.24 ± 3% perf-profile.children.cycles-pp.build3
0.14 +0.0 0.14 ± 3% perf-profile.children.cycles-pp.cgraph_node::remove
0.15 ± 3% +0.0 0.15 perf-profile.children.cycles-pp.__fxstat64
0.13 +0.0 0.13 ± 3% perf-profile.children.cycles-pp.strncpy_from_user
0.13 +0.0 0.13 ± 3% perf-profile.children.cycles-pp.dup_mm
0.08 ± 6% +0.0 0.08 perf-profile.children.cycles-pp.wi::mul_internal
0.07 ± 6% +0.0 0.08 ± 6% perf-profile.children.cycles-pp.lra
0.06 ± 7% +0.0 0.07 ± 7% perf-profile.children.cycles-pp.reg_scan
0.05 ± 8% +0.0 0.06 ± 8% perf-profile.children.cycles-pp.file_free_rcu
0.05 ± 8% +0.0 0.06 ± 8% perf-profile.children.cycles-pp.do_open_execat
0.06 ± 7% +0.0 0.07 ± 7% perf-profile.children.cycles-pp.df_grow_reg_info
0.05 +0.0 0.05 ± 8% perf-profile.children.cycles-pp.cpp_spell_token
0.05 ± 8% +0.0 0.06 ± 8% perf-profile.children.cycles-pp.__x64_sys_close
0.05 ± 8% +0.0 0.06 ± 8% perf-profile.children.cycles-pp.__get_obj_cgroup_from_memcg
0.05 +0.0 0.05 ± 8% perf-profile.children.cycles-pp.expand_call
0.05 ± 8% +0.0 0.06 ± 8% perf-profile.children.cycles-pp.delete_basic_block
0.06 ± 8% +0.0 0.06 perf-profile.children.cycles-pp.flush_tlb_func
0.06 ± 8% +0.0 0.06 perf-profile.children.cycles-pp.schedule_block
0.06 ± 8% +0.0 0.06 perf-profile.children.cycles-pp.gimple_resimplify2
0.12 ± 3% +0.0 0.13 ± 3% perf-profile.children.cycles-pp.get_inner_reference
0.10 ± 9% +0.0 0.11 ± 8% perf-profile.children.cycles-pp.__check_object_size
0.10 ± 4% +0.0 0.10 perf-profile.children.cycles-pp.__open64_nocancel
0.09 ± 5% +0.0 0.09 perf-profile.children.cycles-pp.page_add_file_rmap
0.09 ± 5% +0.0 0.09 ± 9% perf-profile.children.cycles-pp.simple_lookup
0.08 ± 5% +0.0 0.09 ± 5% perf-profile.children.cycles-pp.post_order_compute
0.07 +0.0 0.07 ± 6% perf-profile.children.cycles-pp.eliminate_dom_walker::before_dom_children
0.07 ± 7% +0.0 0.07 perf-profile.children.cycles-pp._copy_from_user
0.07 ± 7% +0.0 0.07 perf-profile.children.cycles-pp.remove_vma
0.07 ± 7% +0.0 0.07 perf-profile.children.cycles-pp.dom_opt_dom_walker::before_dom_children
0.06 +0.0 0.06 ± 7% perf-profile.children.cycles-pp.memcpy@plt
0.06 +0.0 0.06 ± 7% perf-profile.children.cycles-pp.lvalue_p
0.07 ± 14% +0.0 0.07 perf-profile.children.cycles-pp.ipa_icf::sem_function::init
0.06 +0.0 0.06 ± 7% perf-profile.children.cycles-pp.make_pointer_declarator
0.06 +0.0 0.06 ± 7% perf-profile.children.cycles-pp.brk
0.06 +0.0 0.06 ± 7% perf-profile.children.cycles-pp.bitmap_clear_bit
0.06 +0.0 0.06 ± 7% perf-profile.children.cycles-pp.df_analyze
0.06 +0.0 0.06 ± 7% perf-profile.children.cycles-pp.estimate_num_insns
0.06 ± 13% +0.0 0.06 ± 7% perf-profile.children.cycles-pp.regstat_compute_calls_crossed
0.28 ± 3% +0.0 0.29 ± 3% perf-profile.children.cycles-pp.filename_lookup
0.28 ± 3% +0.0 0.29 ± 3% perf-profile.children.cycles-pp.path_lookupat
0.16 ± 6% +0.0 0.16 ± 5% perf-profile.children.cycles-pp.update_stmt_operands
0.97 +0.0 0.97 ± 3% perf-profile.children.cycles-pp.do_fault
0.78 +0.0 0.78 perf-profile.children.cycles-pp.do_mmap
0.81 +0.0 0.82 perf-profile.children.cycles-pp.vm_mmap_pgoff
0.14 ± 3% +0.0 0.15 perf-profile.children.cycles-pp.remove_unused_locals
0.13 ± 9% +0.0 0.13 ± 3% perf-profile.children.cycles-pp.rewrite_update_dom_walker::before_dom_children
0.10 +0.0 0.11 ± 11% perf-profile.children.cycles-pp.page_remove_file_rmap
0.10 ± 4% +0.0 0.10 ± 9% perf-profile.children.cycles-pp.vn_reference_lookup
0.08 ± 6% +0.0 0.08 ± 5% perf-profile.children.cycles-pp.__d_add
0.08 ± 6% +0.0 0.08 ± 5% perf-profile.children.cycles-pp.__might_fault
0.07 ± 6% +0.0 0.08 perf-profile.children.cycles-pp.c_common_truthvalue_conversion
0.05 +0.0 0.06 ± 8% perf-profile.children.cycles-pp.build_component_ref
0.05 ± 8% +0.0 0.06 perf-profile.children.cycles-pp.count
0.05 ± 8% +0.0 0.06 perf-profile.children.cycles-pp.excess_precision_type
0.05 ± 8% +0.0 0.06 perf-profile.children.cycles-pp.sd_add_dep
0.05 ± 8% +0.0 0.06 perf-profile.children.cycles-pp.bitmap_bit_p
0.12 ± 3% +0.0 0.13 perf-profile.children.cycles-pp.walk_stmt_load_store_addr_ops
0.12 ± 3% +0.0 0.13 perf-profile.children.cycles-pp.substitute_and_fold_engine::substitute_and_fold
0.10 ± 4% +0.0 0.11 perf-profile.children.cycles-pp.c_parse_file
0.10 ± 9% +0.0 0.10 ± 4% perf-profile.children.cycles-pp.mark_irreducible_loops
0.09 ± 5% +0.0 0.09 ± 5% perf-profile.children.cycles-pp._cpp_temp_token
0.08 +0.0 0.09 ± 5% perf-profile.children.cycles-pp.generic_simplify
0.07 +0.0 0.08 ± 6% perf-profile.children.cycles-pp.call_rcu
0.07 +0.0 0.08 ± 12% perf-profile.children.cycles-pp.build_asm_expr
0.06 ± 7% +0.0 0.07 perf-profile.children.cycles-pp.const_unop
0.07 +0.0 0.08 ± 6% perf-profile.children.cycles-pp.execute_update_addresses_taken
0.06 ± 7% +0.0 0.07 perf-profile.children.cycles-pp.flow_loops_find
0.06 +0.0 0.07 ± 7% perf-profile.children.cycles-pp.start_function
0.06 +0.0 0.07 ± 7% perf-profile.children.cycles-pp.rcu_all_qs
0.06 +0.0 0.07 ± 7% perf-profile.children.cycles-pp.simple_cst_equal
0.07 ± 11% +0.0 0.08 ± 6% perf-profile.children.cycles-pp.gimple_fold_stmt_to_constant_1
0.06 +0.0 0.07 ± 7% perf-profile.children.cycles-pp.compare_tree_int
0.06 +0.0 0.07 ± 7% perf-profile.children.cycles-pp.recompute_tree_invariant_for_addr_expr
0.06 +0.0 0.07 ± 7% perf-profile.children.cycles-pp.lra_assign
0.06 +0.0 0.07 ± 7% perf-profile.children.cycles-pp.protected_set_expr_location
0.06 ± 8% +0.0 0.06 ± 7% perf-profile.children.cycles-pp.lra_constraints
0.06 ± 8% +0.0 0.06 ± 7% perf-profile.children.cycles-pp.init_subregs_of_mode
0.12 ± 4% +0.0 0.12 ± 3% perf-profile.children.cycles-pp.dup_mmap
0.12 ± 6% +0.0 0.13 ± 3% perf-profile.children.cycles-pp.cleanup_cfg
0.07 ± 7% +0.0 0.07 ± 6% perf-profile.children.cycles-pp.lra_remat
0.26 +0.0 0.26 perf-profile.children.cycles-pp.vfs_statx
0.28 +0.0 0.29 perf-profile.children.cycles-pp.convert_lvalue_to_rvalue
0.22 ± 2% +0.0 0.23 ± 2% perf-profile.children.cycles-pp._cpp_pop_context
0.22 ± 2% +0.0 0.23 ± 2% perf-profile.children.cycles-pp._cpp_find_file
0.14 ± 3% +0.0 0.15 ± 3% perf-profile.children.cycles-pp.do_set_pte
0.14 +0.0 0.15 perf-profile.children.cycles-pp.wide_int_binop
0.10 +0.0 0.11 perf-profile.children.cycles-pp.c_token_starts_typename
0.07 ± 6% +0.0 0.08 ± 5% perf-profile.children.cycles-pp.extract_constrain_insn_cached
0.05 +0.0 0.06 perf-profile.children.cycles-pp.fold_ternary_loc
0.05 +0.0 0.06 perf-profile.children.cycles-pp.cgraph_node::finalize_function
0.05 +0.0 0.06 perf-profile.children.cycles-pp.build_c_cast
0.05 +0.0 0.06 perf-profile.children.cycles-pp.alloc_fd
0.05 +0.0 0.06 ± 13% perf-profile.children.cycles-pp.rtx_equal_for_cselib_1
0.05 +0.0 0.06 perf-profile.children.cycles-pp.sched_free_deps
0.11 +0.0 0.12 ± 6% perf-profile.children.cycles-pp.create_elf_tables
0.10 ± 4% +0.0 0.11 ± 4% perf-profile.children.cycles-pp.realloc
0.10 ± 4% +0.0 0.11 ± 4% perf-profile.children.cycles-pp.mod_objcg_state
0.08 ± 6% +0.0 0.09 ± 5% perf-profile.children.cycles-pp.hash_rtx_cb
0.07 +0.0 0.08 perf-profile.children.cycles-pp.shorten_compare
0.07 +0.0 0.08 perf-profile.children.cycles-pp.get_user_arg_ptr
0.07 +0.0 0.08 perf-profile.children.cycles-pp.arch_get_unmapped_area_topdown
0.05 ± 8% +0.0 0.06 ± 7% perf-profile.children.cycles-pp.avail_exprs_stack::lookup_avail_expr
0.62 +0.0 0.63 ± 2% perf-profile.children.cycles-pp.filemap_map_pages
0.26 +0.0 0.27 ± 5% perf-profile.children.cycles-pp.__libc_fork
0.22 ± 3% +0.0 0.23 perf-profile.children.cycles-pp.__d_alloc
0.21 ± 2% +0.0 0.22 ± 2% perf-profile.children.cycles-pp.check_function_arguments
0.17 +0.0 0.18 perf-profile.children.cycles-pp.declspecs_add_type
0.16 ± 2% +0.0 0.17 ± 2% perf-profile.children.cycles-pp.kmem_cache_alloc_lru
0.15 ± 3% +0.0 0.16 ± 2% perf-profile.children.cycles-pp.linemap_resolve_location
0.16 ± 3% +0.0 0.17 ± 2% perf-profile.children.cycles-pp.memcg_slab_post_alloc_hook
0.13 ± 6% +0.0 0.14 perf-profile.children.cycles-pp.cpp_classify_number
0.13 ± 3% +0.0 0.14 ± 3% perf-profile.children.cycles-pp.ssa_propagation_engine::ssa_propagate
0.11 ± 4% +0.0 0.12 ± 3% perf-profile.children.cycles-pp.linemap_lookup
0.10 ± 9% +0.0 0.11 ± 4% perf-profile.children.cycles-pp.inverted_post_order_compute
0.10 ± 9% +0.0 0.11 ± 4% perf-profile.children.cycles-pp.rebuild_jump_labels
0.08 ± 5% +0.0 0.09 ± 5% perf-profile.children.cycles-pp.make_location
0.06 +0.0 0.07 perf-profile.children.cycles-pp.check_heap_object
0.06 ± 7% +0.0 0.07 ± 6% perf-profile.children.cycles-pp.delete_trivially_dead_insns
0.06 ± 7% +0.0 0.07 ± 6% perf-profile.children.cycles-pp.ggc_free
0.06 ± 13% +0.0 0.07 perf-profile.children.cycles-pp.unshare_all_rtl_in_chain
0.12 ± 4% +0.0 0.13 ± 3% perf-profile.children.cycles-pp.extract_insn
0.66 +0.0 0.67 ± 2% perf-profile.children.cycles-pp.do_read_fault
0.40 +0.0 0.41 ± 2% perf-profile.children.cycles-pp.fold_unary_loc
0.19 ± 2% +0.0 0.21 ± 2% perf-profile.children.cycles-pp.begin_new_exec
0.14 ± 3% +0.0 0.16 ± 3% perf-profile.children.cycles-pp.ira_init
0.14 ± 3% +0.0 0.16 ± 3% perf-profile.children.cycles-pp.perf_event_mmap_event
0.10 ± 4% +0.0 0.12 ± 4% perf-profile.children.cycles-pp.get_attr_type
0.10 ± 4% +0.0 0.12 ± 4% perf-profile.children.cycles-pp.df_compact_blocks
0.07 +0.0 0.08 ± 5% perf-profile.children.cycles-pp.unlink_anon_vmas
0.05 +0.0 0.06 ± 7% perf-profile.children.cycles-pp.unmapped_area_topdown
0.05 ± 8% +0.0 0.07 ± 7% perf-profile.children.cycles-pp.lra_final_code_change
0.05 +0.0 0.06 ± 7% perf-profile.children.cycles-pp.__clear_user
0.16 +0.0 0.17 ± 7% perf-profile.children.cycles-pp.type_hash_canon_hash
0.15 +0.0 0.16 ± 2% perf-profile.children.cycles-pp.iterative_hash
0.15 ± 3% +0.0 0.17 ± 2% perf-profile.children.cycles-pp.exec_mmap
0.13 +0.0 0.14 ± 3% perf-profile.children.cycles-pp.rcu_cblist_dequeue
0.14 ± 5% +0.0 0.15 ± 6% perf-profile.children.cycles-pp.side_effects_p
0.13 ± 6% +0.0 0.14 ± 3% perf-profile.children.cycles-pp.schedule_insns
0.12 +0.0 0.13 ± 9% perf-profile.children.cycles-pp.ferror@plt
0.09 ± 5% +0.0 0.11 ± 4% perf-profile.children.cycles-pp.place_field
0.09 ± 5% +0.0 0.11 ± 4% perf-profile.children.cycles-pp.max_issue
0.09 ± 5% +0.0 0.11 ± 4% perf-profile.children.cycles-pp.add_dependence
0.08 +0.0 0.09 ± 5% perf-profile.children.cycles-pp.get_unmapped_area
0.08 ± 6% +0.0 0.09 perf-profile.children.cycles-pp.wi::divmod_internal
0.06 ± 7% +0.0 0.08 ± 6% perf-profile.children.cycles-pp.___slab_alloc
0.07 ± 7% +0.0 0.08 perf-profile.children.cycles-pp.get_obj_cgroup_from_current
0.06 ± 7% +0.0 0.08 ± 6% perf-profile.children.cycles-pp.rtx_equal_p
0.06 ± 8% +0.0 0.07 perf-profile.children.cycles-pp.__entry_text_start
0.06 ± 8% +0.0 0.07 perf-profile.children.cycles-pp.vm_unmapped_area
0.06 ± 8% +0.0 0.07 perf-profile.children.cycles-pp.c_build_qualified_type
0.11 ± 4% +0.0 0.13 ± 3% perf-profile.children.cycles-pp.pop_stmt_list
0.11 ± 4% +0.0 0.13 ± 3% perf-profile.children.cycles-pp.gimplify_expr
0.08 ± 5% +0.0 0.10 ± 4% perf-profile.children.cycles-pp.dom_walker::walk
0.06 +0.0 0.07 ± 6% perf-profile.children.cycles-pp.wi::lshift_large
0.33 ± 2% +0.0 0.35 perf-profile.children.cycles-pp.build_null_declspecs
0.38 ± 4% +0.0 0.40 ± 4% perf-profile.children.cycles-pp.df_insn_rescan
0.31 ± 2% +0.0 0.33 ± 2% perf-profile.children.cycles-pp.kthread
0.20 ± 2% +0.0 0.22 perf-profile.children.cycles-pp.getname_flags
0.15 +0.0 0.17 ± 2% perf-profile.children.cycles-pp.perf_event_mmap
0.14 ± 3% +0.0 0.16 perf-profile.children.cycles-pp.parser_build_binary_op
0.10 ± 4% +0.0 0.12 perf-profile.children.cycles-pp.walk_tree_1
0.09 +0.0 0.11 ± 4% perf-profile.children.cycles-pp.cpp_interpret_integer
0.09 +0.0 0.11 ± 8% perf-profile.children.cycles-pp.build_int_cst
0.07 +0.0 0.09 ± 5% perf-profile.children.cycles-pp.wi::zext_large
0.02 ±141% +0.0 0.03 ± 70% perf-profile.children.cycles-pp.update_min_vruntime
0.00 +0.0 0.02 ±141% perf-profile.children.cycles-pp.shorten_branches
0.00 +0.0 0.02 ±141% perf-profile.children.cycles-pp._cpp_get_buff
0.00 +0.0 0.02 ±141% perf-profile.children.cycles-pp.__rb_insert_augmented
0.00 +0.0 0.02 ±141% perf-profile.children.cycles-pp.lra_undo_inheritance
0.00 +0.0 0.02 ±141% perf-profile.children.cycles-pp.record_temporary_equivalences
0.00 +0.0 0.02 ±141% perf-profile.children.cycles-pp.add_wait_queue
0.00 +0.0 0.02 ±141% perf-profile.children.cycles-pp.int_fits_type_p
0.00 +0.0 0.02 ±141% perf-profile.children.cycles-pp.update_alignment_for_field
0.12 ± 3% +0.0 0.14 perf-profile.children.cycles-pp.malloc@plt
0.12 ± 3% +0.0 0.14 ± 5% perf-profile.children.cycles-pp.inchash::add_expr
0.12 ± 3% +0.0 0.14 ± 5% perf-profile.children.cycles-pp.warnings_for_convert_and_check
0.12 ± 4% +0.0 0.13 ± 3% perf-profile.children.cycles-pp.memchr@plt
0.12 ± 4% +0.0 0.13 ± 3% perf-profile.children.cycles-pp.sched_rgn_compute_dependencies
0.11 +0.0 0.13 ± 3% perf-profile.children.cycles-pp.fold
0.06 +0.0 0.08 ± 6% perf-profile.children.cycles-pp.size_int_kind
0.06 +0.0 0.08 ± 6% perf-profile.children.cycles-pp.rpo_vn_valueize
0.03 ± 70% +0.0 0.05 perf-profile.children.cycles-pp.native_flush_tlb_one_user
0.03 ± 70% +0.0 0.05 perf-profile.children.cycles-pp.c_process_expr_stmt
0.03 ± 70% +0.0 0.05 perf-profile.children.cycles-pp.__fsnotify_parent
0.03 ± 70% +0.0 0.05 perf-profile.children.cycles-pp.htab_hash_string
0.03 ± 70% +0.0 0.05 perf-profile.children.cycles-pp.recog
0.03 ± 70% +0.0 0.05 perf-profile.children.cycles-pp.__update_load_avg_cfs_rq
0.03 ± 70% +0.0 0.05 perf-profile.children.cycles-pp.delete_unreachable_blocks
0.03 ± 70% +0.0 0.05 perf-profile.children.cycles-pp.vm_area_dup
0.16 ± 3% +0.0 0.17 ± 2% perf-profile.children.cycles-pp._dl_addr
0.17 ± 7% +0.0 0.19 ± 4% perf-profile.children.cycles-pp.pre_and_rev_post_order_compute_fn
0.74 ± 3% +0.0 0.76 perf-profile.children.cycles-pp.pop_scope
0.49 +0.0 0.51 ± 2% perf-profile.children.cycles-pp.__irq_exit_rcu
0.36 ± 4% +0.0 0.38 ± 5% perf-profile.children.cycles-pp.gimple_call_flags
0.29 ± 4% +0.0 0.31 perf-profile.children.cycles-pp.calculate_dominance_info
0.21 ± 5% +0.0 0.23 ± 5% perf-profile.children.cycles-pp.regstat_compute_ri
0.19 ± 4% +0.0 0.21 ± 7% perf-profile.children.cycles-pp.get_ref_base_and_extent
0.19 ± 4% +0.0 0.21 ± 2% perf-profile.children.cycles-pp.htab_traverse
0.07 ± 7% +0.0 0.09 ± 5% perf-profile.children.cycles-pp.shift_arg_pages
0.03 ± 70% +0.0 0.05 ± 8% perf-profile.children.cycles-pp.parse_input_constraint
0.03 ± 70% +0.0 0.05 ± 8% perf-profile.children.cycles-pp.simplify_const_relational_operation
0.03 ± 70% +0.0 0.05 ± 8% perf-profile.children.cycles-pp.expand_expr_real_1
0.03 ± 70% +0.0 0.05 ± 8% perf-profile.children.cycles-pp.entry_SYSRETQ_unsafe_stack
0.42 ± 7% +0.0 0.44 perf-profile.children.cycles-pp.df_worklist_dataflow
0.34 ± 2% +0.0 0.36 ± 3% perf-profile.children.cycles-pp.ret_from_fork
0.17 ± 2% +0.0 0.19 ± 2% perf-profile.children.cycles-pp.strnlen_user
0.17 +0.0 0.19 ± 4% perf-profile.children.cycles-pp.free@plt
0.16 ± 3% +0.0 0.18 ± 2% perf-profile.children.cycles-pp.c_parser_declarator
0.24 ± 3% +0.0 0.26 perf-profile.children.cycles-pp.get_range_from_loc
0.21 ± 6% +0.0 0.23 ± 5% perf-profile.children.cycles-pp.single_def_use_dom_walker::before_dom_children
0.18 ± 7% +0.0 0.20 ± 10% perf-profile.children.cycles-pp.diagnostic_report_diagnostic
0.19 ± 4% +0.0 0.21 ± 5% perf-profile.children.cycles-pp._IO_ferror
0.17 ± 7% +0.0 0.20 ± 4% perf-profile.children.cycles-pp.cselib_lookup
0.13 ± 9% +0.0 0.16 ± 3% perf-profile.children.cycles-pp.lra_inheritance
0.08 +0.0 0.10 ± 4% perf-profile.children.cycles-pp.c_parser_skip_until_found
0.03 ± 70% +0.0 0.06 ± 8% perf-profile.children.cycles-pp.c_determine_visibility
0.03 ± 70% +0.0 0.06 ± 8% perf-profile.children.cycles-pp.insn_extract
0.24 +0.0 0.27 perf-profile.children.cycles-pp.convert
0.10 ± 4% +0.0 0.13 ± 3% perf-profile.children.cycles-pp.start_decl
0.08 ± 6% +0.0 0.10 perf-profile.children.cycles-pp.@plt
0.11 ± 4% +0.0 0.14 ± 3% perf-profile.children.cycles-pp.general_operand
0.34 +0.0 0.36 perf-profile.children.cycles-pp.verify_sequence_points
0.25 +0.0 0.27 ± 4% perf-profile.children.cycles-pp.__mmap
0.43 ± 6% +0.0 0.46 ± 2% perf-profile.children.cycles-pp.execute_one_pass
0.68 +0.0 0.70 ± 2% perf-profile.children.cycles-pp.__softirqentry_text_start
0.56 ± 2% +0.0 0.59 perf-profile.children.cycles-pp.rcu_do_batch
0.62 ± 2% +0.0 0.65 ± 2% perf-profile.children.cycles-pp.rcu_core
0.18 ± 2% +0.0 0.21 perf-profile.children.cycles-pp.layout_decl
0.18 ± 2% +0.0 0.21 perf-profile.children.cycles-pp.rtx_cost
0.14 ± 3% +0.0 0.16 ± 2% perf-profile.children.cycles-pp.build_unary_op
0.13 ± 3% +0.0 0.15 ± 3% perf-profile.children.cycles-pp.c_parser_require
0.22 ± 3% +0.0 0.25 perf-profile.children.cycles-pp.integer_zerop
0.07 +0.0 0.10 ± 4% perf-profile.children.cycles-pp.setup_arg_pages
0.11 ± 8% +0.0 0.14 perf-profile.children.cycles-pp.c_fully_fold
0.23 ± 4% +0.0 0.26 perf-profile.children.cycles-pp.__libc_calloc
0.04 ± 71% +0.0 0.07 ± 7% perf-profile.children.cycles-pp.account_user_time
0.47 +0.0 0.50 ± 2% perf-profile.children.cycles-pp.fold_binary_loc
0.27 ± 3% +0.0 0.30 perf-profile.children.cycles-pp.kmem_cache_free
0.00 +0.0 0.03 ± 70% perf-profile.children.cycles-pp.exp_equiv_p
0.00 +0.0 0.03 ± 70% perf-profile.children.cycles-pp.get_attr_memory
0.00 +0.0 0.03 ± 70% perf-profile.children.cycles-pp.__get_user_8
0.00 +0.0 0.03 ± 70% perf-profile.children.cycles-pp.declspecs_add_scspec
0.00 +0.0 0.03 ± 70% perf-profile.children.cycles-pp.worker_thread
0.00 +0.0 0.03 ± 70% perf-profile.children.cycles-pp.integer_onep
0.02 ±141% +0.0 0.05 perf-profile.children.cycles-pp.c_objc_common_truthvalue_conversion
0.02 ±141% +0.0 0.05 perf-profile.children.cycles-pp.c_common_signed_or_unsigned_type
0.02 ±141% +0.0 0.05 perf-profile.children.cycles-pp.strncmp@plt
0.02 ±141% +0.0 0.05 perf-profile.children.cycles-pp.bitmap_alloc
0.02 ±141% +0.0 0.05 perf-profile.children.cycles-pp.evrp_dom_walker::before_dom_children
0.02 ±141% +0.0 0.05 perf-profile.children.cycles-pp.const_binop
0.02 ±141% +0.0 0.05 perf-profile.children.cycles-pp.output_asm_insn
0.02 ±141% +0.0 0.05 perf-profile.children.cycles-pp.cselib_subst_to_values
0.32 +0.0 0.35 ± 2% perf-profile.children.cycles-pp.copy_strings
0.50 +0.0 0.54 ± 3% perf-profile.children.cycles-pp.getenv
0.02 ±141% +0.0 0.05 ± 8% perf-profile.children.cycles-pp.ira_build_conflicts
0.02 ±141% +0.0 0.05 ± 8% perf-profile.children.cycles-pp.vr_values::extract_range_from_stmt
0.00 +0.0 0.04 ± 71% perf-profile.children.cycles-pp.insn_default_length
0.97 +0.0 1.01 ± 4% perf-profile.children.cycles-pp.execve
0.29 ± 2% +0.0 0.33 ± 2% perf-profile.children.cycles-pp.default_conversion
0.26 ± 3% +0.0 0.31 ± 4% perf-profile.children.cycles-pp.get_combined_adhoc_loc
0.88 +0.0 0.92 ± 2% perf-profile.children.cycles-pp.bprm_execve
0.47 +0.0 0.51 ± 2% perf-profile.children.cycles-pp._cpp_clean_line
0.24 ± 5% +0.0 0.28 ± 2% perf-profile.children.cycles-pp.bitmap_set_bit
0.72 +0.0 0.77 ± 3% perf-profile.children.cycles-pp.exec_binprm
0.44 +0.0 0.48 perf-profile.children.cycles-pp.kmem_cache_alloc
0.41 ± 6% +0.0 0.45 ± 3% perf-profile.children.cycles-pp.init_alias_analysis
0.32 ± 2% +0.0 0.37 ± 2% perf-profile.children.cycles-pp.c_parser_consume_token
0.69 ± 3% +0.0 0.74 perf-profile.children.cycles-pp.variably_modified_type_p
0.28 ± 2% +0.0 0.33 ± 6% perf-profile.children.cycles-pp.fix_string_type
0.26 +0.0 0.31 ± 3% perf-profile.children.cycles-pp.linemap_line_start
0.00 +0.1 0.05 perf-profile.children.cycles-pp.__anon_vma_prepare
0.00 +0.1 0.05 perf-profile.children.cycles-pp.c_parser_peek_2nd_token
0.00 +0.1 0.05 perf-profile.children.cycles-pp.memcmp@plt
0.00 +0.1 0.05 perf-profile.children.cycles-pp.c_vla_unspec_p
0.00 +0.1 0.05 perf-profile.children.cycles-pp.pos_from_bit
0.71 +0.1 0.76 ± 2% perf-profile.children.cycles-pp.search_binary_handler
0.70 +0.1 0.75 ± 3% perf-profile.children.cycles-pp.load_elf_binary
0.08 +0.1 0.13 ± 3% perf-profile.children.cycles-pp.free_unref_page
0.00 +0.1 0.05 ± 8% perf-profile.children.cycles-pp.c_end_compound_stmt
0.14 ± 3% +0.1 0.20 ± 2% perf-profile.children.cycles-pp.__cond_resched
0.00 +0.1 0.06 ± 8% perf-profile.children.cycles-pp.free_pcp_prepare
0.00 +0.1 0.06 perf-profile.children.cycles-pp.check_new_pages
0.00 +0.1 0.06 perf-profile.children.cycles-pp.set_cfun
0.11 +0.1 0.17 ± 2% perf-profile.children.cycles-pp.get_parm_info
0.56 ± 2% +0.1 0.62 perf-profile.children.cycles-pp._cpp_skip_block_comment
0.52 +0.1 0.59 ± 2% perf-profile.children.cycles-pp.operand_equal_p
0.44 ± 2% +0.1 0.51 ± 4% perf-profile.children.cycles-pp._cpp_handle_directive
1.39 +0.1 1.46 ± 2% perf-profile.children.cycles-pp.__x64_sys_execve
0.54 +0.1 0.62 ± 2% perf-profile.children.cycles-pp.build_binary_op
1.38 +0.1 1.46 ± 2% perf-profile.children.cycles-pp.do_execveat_common
0.63 ± 5% +0.1 0.71 ± 2% perf-profile.children.cycles-pp.df_bb_refs_record
1.09 +0.1 1.17 perf-profile.children.cycles-pp.size_binop_loc
0.69 ± 2% +0.1 0.77 ± 5% perf-profile.children.cycles-pp.location_get_source_line
0.42 ± 5% +0.1 0.51 ± 4% perf-profile.children.cycles-pp.constrain_operands
0.00 +0.1 0.09 perf-profile.children.cycles-pp.do_huge_pmd_anonymous_page
0.71 +0.1 0.80 perf-profile.children.cycles-pp.c_parser_declspecs
0.42 ± 2% +0.1 0.52 perf-profile.children.cycles-pp.decl_attributes
0.19 ± 2% +0.1 0.29 perf-profile.children.cycles-pp.finish_decl
0.95 +0.1 1.06 perf-profile.children.cycles-pp.cfree
0.00 +0.1 0.12 ± 4% perf-profile.children.cycles-pp.init_ttree
0.62 ± 2% +0.1 0.74 ± 2% perf-profile.children.cycles-pp.layout_type
0.84 ± 2% +0.1 0.96 ± 9% perf-profile.children.cycles-pp.linemap_location_in_system_header_p
0.24 +0.2 0.40 perf-profile.children.cycles-pp.__might_resched
0.45 +0.2 0.62 perf-profile.children.cycles-pp.wi::force_to_size
0.89 +0.2 1.08 perf-profile.children.cycles-pp.ggc_internal_alloc
2.01 +0.2 2.21 perf-profile.children.cycles-pp.malloc
3.60 ± 2% +0.2 3.81 perf-profile.children.cycles-pp.htab_find_slot_with_hash
1.06 +0.2 1.29 ± 3% perf-profile.children.cycles-pp._cpp_lex_direct
2.58 +0.8 3.41 perf-profile.children.cycles-pp.clear_page_erms
6.26 +1.0 7.23 perf-profile.children.cycles-pp.c_lex_with_flags
0.04 ± 71% +2.5 2.58 perf-profile.children.cycles-pp.clear_huge_page
0.05 ± 70% +2.6 2.62 perf-profile.children.cycles-pp.__do_huge_pmd_anonymous_page
3.17 ± 4% -2.0 1.19 ± 15% perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
1.01 -0.5 0.51 perf-profile.self.cycles-pp.__list_del_entry_valid
1.01 ± 2% -0.5 0.51 perf-profile.self.cycles-pp.charge_memcg
0.83 -0.4 0.40 perf-profile.self.cycles-pp.try_charge_memcg
0.84 -0.4 0.49 ± 3% perf-profile.self.cycles-pp.zap_pte_range
0.71 -0.3 0.38 ± 2% perf-profile.self.cycles-pp.__free_one_page
0.49 ± 3% -0.3 0.17 ± 5% perf-profile.self.cycles-pp.get_mem_cgroup_from_mm
0.79 -0.3 0.49 ± 4% perf-profile.self.cycles-pp.sync_regs
0.68 -0.3 0.42 ± 2% perf-profile.self.cycles-pp.native_irq_return_iret
0.21 ± 4% -0.1 0.09 ± 5% perf-profile.self.cycles-pp.__mod_node_page_state
0.24 -0.1 0.12 perf-profile.self.cycles-pp.get_page_from_freelist
0.28 -0.1 0.17 ± 2% perf-profile.self.cycles-pp.__handle_mm_fault
0.28 ± 2% -0.1 0.18 ± 2% perf-profile.self.cycles-pp.release_pages
0.19 ± 4% -0.1 0.09 perf-profile.self.cycles-pp.page_remove_rmap
0.20 ± 4% -0.1 0.12 ± 4% perf-profile.self.cycles-pp.do_anonymous_page
0.16 ± 2% -0.1 0.08 perf-profile.self.cycles-pp.__mem_cgroup_charge
0.16 ± 5% -0.1 0.08 ± 6% perf-profile.self.cycles-pp.__mod_lruvec_page_state
1.72 -0.1 1.64 perf-profile.self.cycles-pp.type_hash_canon
0.18 ± 5% -0.1 0.10 ± 4% perf-profile.self.cycles-pp.__alloc_pages
0.16 ± 3% -0.1 0.08 ± 5% perf-profile.self.cycles-pp.lru_add_fn
0.26 ± 4% -0.1 0.18 ± 5% perf-profile.self.cycles-pp.update_cfs_group
0.49 -0.1 0.42 perf-profile.self.cycles-pp.lookup_name
0.13 ± 3% -0.1 0.06 ± 7% perf-profile.self.cycles-pp.free_swap_cache
0.16 ± 3% -0.1 0.09 ± 5% perf-profile.self.cycles-pp.rmqueue
0.17 ± 2% -0.1 0.11 perf-profile.self.cycles-pp.handle_mm_fault
0.06 ± 7% -0.1 0.00 perf-profile.self.cycles-pp.mem_cgroup_update_lru_size
0.06 ± 7% -0.1 0.00 perf-profile.self.cycles-pp.__count_memcg_events
0.18 ± 2% -0.1 0.12 ± 4% perf-profile.self.cycles-pp.___perf_sw_event
0.06 -0.1 0.00 perf-profile.self.cycles-pp.ktime_get_update_offsets_now
0.06 -0.1 0.00 perf-profile.self.cycles-pp.__mod_zone_page_state
0.06 -0.1 0.00 perf-profile.self.cycles-pp.free_unref_page_commit
0.11 ± 4% -0.1 0.05 perf-profile.self.cycles-pp.page_counter_try_charge
0.05 ± 8% -0.1 0.00 perf-profile.self.cycles-pp.folio_add_lru
0.05 ± 8% -0.1 0.00 perf-profile.self.cycles-pp.mem_cgroup_charge_statistics
0.05 ± 8% -0.1 0.00 perf-profile.self.cycles-pp.down_read_trylock
0.11 ± 4% -0.1 0.06 perf-profile.self.cycles-pp.folio_batch_move_lru
0.05 -0.1 0.00 perf-profile.self.cycles-pp.__mod_lruvec_state
0.05 -0.1 0.00 perf-profile.self.cycles-pp.__page_set_anon_rmap
0.05 -0.1 0.00 perf-profile.self.cycles-pp.output_addr_const
0.05 -0.1 0.00 perf-profile.self.cycles-pp.errseq_sample
0.14 ± 3% -0.1 0.09 ± 5% perf-profile.self.cycles-pp.__mod_memcg_lruvec_state
0.13 ± 6% -0.1 0.08 ± 10% perf-profile.self.cycles-pp.__list_add_valid
0.07 ± 7% -0.1 0.02 ±141% perf-profile.self.cycles-pp.up_read
0.10 ± 4% -0.0 0.05 ± 8% perf-profile.self.cycles-pp.rmqueue_bulk
0.42 -0.0 0.37 ± 5% perf-profile.self.cycles-pp._raw_spin_lock
0.06 ± 7% -0.0 0.02 ±141% perf-profile.self.cycles-pp.asm_exc_page_fault
0.08 -0.0 0.03 ± 70% perf-profile.self.cycles-pp.free_unref_page_list
0.14 ± 3% -0.0 0.10 ± 4% perf-profile.self.cycles-pp.do_user_addr_fault
0.11 ± 4% -0.0 0.06 ± 7% perf-profile.self.cycles-pp.vma_alloc_folio
0.10 ± 9% -0.0 0.06 perf-profile.self.cycles-pp.cgroup_rstat_updated
0.05 ± 8% -0.0 0.02 ±141% perf-profile.self.cycles-pp.irqtime_account_irq
0.09 -0.0 0.06 ± 8% perf-profile.self.cycles-pp.__perf_sw_event
0.05 -0.0 0.02 ±141% perf-profile.self.cycles-pp.string_concat_db::record_string_concatenation
0.03 ± 70% -0.0 0.00 perf-profile.self.cycles-pp.blk_cgroup_congested
0.03 ± 70% -0.0 0.00 perf-profile.self.cycles-pp.apparmor_file_open
0.03 ± 70% -0.0 0.00 perf-profile.self.cycles-pp.rwsem_spin_on_owner
0.03 ± 70% -0.0 0.00 perf-profile.self.cycles-pp.up_write
0.15 ± 6% -0.0 0.12 ± 3% perf-profile.self.cycles-pp.apparmor_file_free_security
0.10 -0.0 0.07 perf-profile.self.cycles-pp.vmacache_find
0.09 ± 5% -0.0 0.06 ± 8% perf-profile.self.cycles-pp.error_entry
0.11 -0.0 0.08 perf-profile.self.cycles-pp._raw_spin_lock_irqsave
0.08 -0.0 0.05 perf-profile.self.cycles-pp.__irqentry_text_end
0.31 -0.0 0.28 perf-profile.self.cycles-pp.ht_lookup_with_hash
0.33 ± 2% -0.0 0.30 ± 4% perf-profile.self.cycles-pp.build_call_array_loc
0.17 ± 2% -0.0 0.15 ± 3% perf-profile.self.cycles-pp.apparmor_file_alloc_security
0.10 -0.0 0.07 ± 6% perf-profile.self.cycles-pp._raw_spin_trylock
0.06 -0.0 0.03 ± 70% perf-profile.self.cycles-pp.handle_pte_fault
0.07 ± 6% -0.0 0.05 ± 74% perf-profile.self.cycles-pp.target_environment
0.21 ± 46% -0.0 0.18 ± 68% perf-profile.self.cycles-pp.osq_lock
0.35 -0.0 0.33 ± 3% perf-profile.self.cycles-pp.next_uptodate_page
0.32 ± 2% -0.0 0.30 perf-profile.self.cycles-pp.__d_lookup_rcu
0.05 ± 8% -0.0 0.03 ± 70% perf-profile.self.cycles-pp.ktime_get
0.25 -0.0 0.23 ± 2% perf-profile.self.cycles-pp.vma_interval_tree_insert
0.12 -0.0 0.10 perf-profile.self.cycles-pp.cgraph_node::create
0.05 -0.0 0.03 ± 70% perf-profile.self.cycles-pp.__vma_adjust
0.05 -0.0 0.03 ± 70% perf-profile.self.cycles-pp.build_tree_list
0.05 -0.0 0.03 ± 70% perf-profile.self.cycles-pp.arch_scale_freq_tick
0.05 -0.0 0.03 ± 70% perf-profile.self.cycles-pp._cpp_parse_expr
0.05 -0.0 0.03 ± 70% perf-profile.self.cycles-pp.find_comparison_dom_walker::before_dom_children
0.05 ± 8% -0.0 0.04 ± 71% perf-profile.self.cycles-pp.file_free_rcu
0.03 ± 70% -0.0 0.02 ±141% perf-profile.self.cycles-pp.hrtimer_interrupt
0.02 ±141% -0.0 0.00 perf-profile.self.cycles-pp.may_open
0.02 ±141% -0.0 0.00 perf-profile.self.cycles-pp.c_finish_return
0.03 ± 70% -0.0 0.02 ±141% perf-profile.self.cycles-pp.htab_hash_string
0.02 ±141% -0.0 0.00 perf-profile.self.cycles-pp.__hrtimer_run_queues
0.02 ±141% -0.0 0.00 perf-profile.self.cycles-pp.make_ssa_name_fn
0.02 ±141% -0.0 0.00 perf-profile.self.cycles-pp.update_ssa
0.02 ±141% -0.0 0.00 perf-profile.self.cycles-pp.bitmap_ior_into
0.13 -0.0 0.12 ± 4% perf-profile.self.cycles-pp.filemap_get_read_batch
0.13 -0.0 0.12 ± 4% perf-profile.self.cycles-pp.build_pointer_type
0.13 ± 3% -0.0 0.12 ± 6% perf-profile.self.cycles-pp.update_sg_wakeup_stats
0.12 ± 3% -0.0 0.11 perf-profile.self.cycles-pp.build_pointer_type_for_mode
0.09 ± 5% -0.0 0.08 perf-profile.self.cycles-pp.update_load_avg
0.73 ± 2% -0.0 0.72 perf-profile.self.cycles-pp.copy_user_enhanced_fast_string
0.21 ± 3% -0.0 0.20 ± 2% perf-profile.self.cycles-pp.__alloc_file
0.23 ± 9% -0.0 0.22 ± 11% perf-profile.self.cycles-pp.check
0.13 ± 3% -0.0 0.12 ± 4% perf-profile.self.cycles-pp.build_external_ref
0.16 ± 5% -0.0 0.15 ± 3% perf-profile.self.cycles-pp.build_qualified_type
0.09 -0.0 0.08 perf-profile.self.cycles-pp.__might_sleep
0.07 -0.0 0.06 perf-profile.self.cycles-pp.down_read
0.07 -0.0 0.06 perf-profile.self.cycles-pp.update_type_inheritance_graph
0.08 -0.0 0.07 perf-profile.self.cycles-pp.lockref_get_not_dead
0.08 ± 5% -0.0 0.07 ± 6% perf-profile.self.cycles-pp.update_curr
0.20 ± 2% -0.0 0.20 ± 6% perf-profile.self.cycles-pp.copy_page
0.09 -0.0 0.08 ± 5% perf-profile.self.cycles-pp.build_type_inheritance_graph
0.07 ± 12% -0.0 0.07 ± 7% perf-profile.self.cycles-pp.update_process_times
0.07 ± 7% -0.0 0.06 perf-profile.self.cycles-pp.finish_declspecs
0.24 ± 3% -0.0 0.24 perf-profile.self.cycles-pp.set_source_range
0.06 ± 8% -0.0 0.05 perf-profile.self.cycles-pp._cpp_stack_include
0.10 ± 4% -0.0 0.09 ± 5% perf-profile.self.cycles-pp.vma_interval_tree_remove
0.09 ± 5% -0.0 0.08 ± 5% perf-profile.self.cycles-pp.filemap_read
0.09 ± 5% -0.0 0.08 ± 5% perf-profile.self.cycles-pp.push_stmt_list
0.08 ± 6% -0.0 0.07 ± 12% perf-profile.self.cycles-pp.tick_sched_do_timer
0.07 ± 6% -0.0 0.07 perf-profile.self.cycles-pp.check_cpu_stall
0.06 ± 7% -0.0 0.06 perf-profile.self.cycles-pp.hard_function_value
0.20 ± 2% -0.0 0.19 ± 2% perf-profile.self.cycles-pp.tree_cons
0.15 ± 3% -0.0 0.15 ± 5% perf-profile.self.cycles-pp.__slab_free
0.06 ± 8% -0.0 0.05 ± 8% perf-profile.self.cycles-pp.inode_permission
0.11 ± 11% -0.0 0.11 ± 4% perf-profile.self.cycles-pp.cleanup_tree_cfg
0.37 ± 2% +0.0 0.37 ± 2% perf-profile.self.cycles-pp.linemap_enter_macro
0.28 +0.0 0.28 perf-profile.self.cycles-pp.build_function_call_vec
0.08 +0.0 0.08 perf-profile.self.cycles-pp.d_alloc_parallel
0.08 +0.0 0.08 perf-profile.self.cycles-pp.ggc_internal_cleared_alloc
0.07 +0.0 0.07 perf-profile.self.cycles-pp.shorten_compare
0.06 ± 7% +0.0 0.06 ± 7% perf-profile.self.cycles-pp.finish_record_layout
0.07 ± 6% +0.0 0.07 ± 6% perf-profile.self.cycles-pp.compute_fn_summary
0.08 ± 10% +0.0 0.08 perf-profile.self.cycles-pp.execute_fixup_cfg
0.07 ± 11% +0.0 0.07 ± 11% perf-profile.self.cycles-pp.get_alias_set
0.06 ± 7% +0.0 0.06 ± 7% perf-profile.self.cycles-pp.flow_loops_find
0.06 +0.0 0.06 perf-profile.self.cycles-pp.mark_exp_read
0.07 ± 14% +0.0 0.07 ± 7% perf-profile.self.cycles-pp.ipa_icf::sem_function::init
0.06 ± 8% +0.0 0.06 ± 8% perf-profile.self.cycles-pp.push_scope
0.06 ± 8% +0.0 0.06 ± 8% perf-profile.self.cycles-pp.schedule_insns
0.05 +0.0 0.05 perf-profile.self.cycles-pp.xas_load
0.05 +0.0 0.05 perf-profile.self.cycles-pp.finish_function
0.05 +0.0 0.05 perf-profile.self.cycles-pp._raw_spin_lock_irq
0.05 +0.0 0.05 perf-profile.self.cycles-pp.cgraph_node::create_edge
0.05 +0.0 0.05 perf-profile.self.cycles-pp.excess_precision_type
0.05 ± 8% +0.0 0.05 ± 8% perf-profile.self.cycles-pp.may_trap_p_1
0.05 +0.0 0.05 perf-profile.self.cycles-pp.c_parser_next_tokens_start_declaration
0.03 ± 70% +0.0 0.03 ± 70% perf-profile.self.cycles-pp.force_fit_type
0.03 ± 70% +0.0 0.03 ± 70% perf-profile.self.cycles-pp.useless_type_conversion_p
0.03 ± 70% +0.0 0.03 ± 70% perf-profile.self.cycles-pp.delete_unreachable_blocks
0.02 ±141% +0.0 0.02 ±141% perf-profile.self.cycles-pp.set_reg_attrs_for_decl_rtl
0.21 ± 2% +0.0 0.22 ± 2% perf-profile.self.cycles-pp.build1
0.20 ± 2% +0.0 0.21 ± 2% perf-profile.self.cycles-pp.fold_convert_loc
0.14 ± 3% +0.0 0.15 ± 3% perf-profile.self.cycles-pp.remove_unused_locals
0.09 +0.0 0.09 ± 10% perf-profile.self.cycles-pp.page_remove_file_rmap
0.08 +0.0 0.08 ± 5% perf-profile.self.cycles-pp._cpp_temp_token
0.05 ± 8% +0.0 0.06 ± 8% perf-profile.self.cycles-pp.start_function
0.05 +0.0 0.05 ± 8% perf-profile.self.cycles-pp.elf_create_section
0.05 ± 8% +0.0 0.06 ± 8% perf-profile.self.cycles-pp.perf_iterate_sb
0.05 ± 8% +0.0 0.06 ± 8% perf-profile.self.cycles-pp.shallow_copy_rtx
0.05 +0.0 0.05 ± 8% perf-profile.self.cycles-pp.c_add_case_label
0.05 +0.0 0.05 ± 8% perf-profile.self.cycles-pp.protected_set_expr_location
0.05 +0.0 0.05 ± 8% perf-profile.self.cycles-pp.cpp_spell_token
0.05 +0.0 0.05 ± 8% perf-profile.self.cycles-pp.fold_ternary_loc
0.05 +0.0 0.05 ± 8% perf-profile.self.cycles-pp.__get_obj_cgroup_from_memcg
0.05 +0.0 0.05 ± 8% perf-profile.self.cycles-pp.build_c_cast
0.05 ± 8% +0.0 0.06 ± 8% perf-profile.self.cycles-pp.delete_basic_block
0.06 ± 8% +0.0 0.06 perf-profile.self.cycles-pp.const_unop
0.06 ± 8% +0.0 0.06 perf-profile.self.cycles-pp.simple_cst_equal
0.06 ± 8% +0.0 0.06 perf-profile.self.cycles-pp.make_pointer_declarator
0.06 ± 8% +0.0 0.06 perf-profile.self.cycles-pp.compare_tree_int
0.06 ± 8% +0.0 0.06 perf-profile.self.cycles-pp.estimate_num_insns
0.12 ± 3% +0.0 0.13 ± 3% perf-profile.self.cycles-pp.walk_stmt_load_store_addr_ops
0.12 ± 4% +0.0 0.12 perf-profile.self.cycles-pp.get_inner_reference
0.12 ± 8% +0.0 0.12 perf-profile.self.cycles-pp.cleanup_cfg
0.09 ± 5% +0.0 0.10 ± 4% perf-profile.self.cycles-pp.ira_init_register_move_cost
0.09 ± 5% +0.0 0.09 perf-profile.self.cycles-pp.sizeof_pointer_memaccess_warning
0.08 ± 5% +0.0 0.09 ± 5% perf-profile.self.cycles-pp.post_order_compute
0.07 +0.0 0.07 ± 6% perf-profile.self.cycles-pp.c_common_truthvalue_conversion
0.07 ± 7% +0.0 0.07 perf-profile.self.cycles-pp.mark_dfs_back_edges
0.06 +0.0 0.06 ± 7% perf-profile.self.cycles-pp.lockref_put_or_lock
0.07 ± 11% +0.0 0.07 ± 6% perf-profile.self.cycles-pp.compute_may_aliases
0.07 ± 7% +0.0 0.07 perf-profile.self.cycles-pp.eliminate_dom_walker::before_dom_children
0.06 ± 13% +0.0 0.06 ± 7% perf-profile.self.cycles-pp.regstat_compute_calls_crossed
0.20 ± 2% +0.0 0.21 perf-profile.self.cycles-pp.check_function_arguments
0.14 ± 3% +0.0 0.14 ± 3% perf-profile.self.cycles-pp.cgraph_node::remove
0.14 ± 3% +0.0 0.14 ± 3% perf-profile.self.cycles-pp.wide_int_binop
0.11 ± 7% +0.0 0.12 ± 4% perf-profile.self.cycles-pp.extract_insn
0.05 +0.0 0.06 ± 8% perf-profile.self.cycles-pp.gimple_resimplify2
0.05 +0.0 0.06 ± 8% perf-profile.self.cycles-pp.rtx_equal_for_cselib_1
0.05 +0.0 0.06 ± 8% perf-profile.self.cycles-pp.bitmap_bit_p
0.05 ± 8% +0.0 0.06 perf-profile.self.cycles-pp.lvalue_p
0.05 ± 8% +0.0 0.06 perf-profile.self.cycles-pp.rtx_alloc
0.05 ± 8% +0.0 0.06 perf-profile.self.cycles-pp.split_all_insns
0.05 ± 8% +0.0 0.06 perf-profile.self.cycles-pp.schedule_block
0.12 ± 3% +0.0 0.13 perf-profile.self.cycles-pp.cpp_classify_number
0.10 ± 4% +0.0 0.11 perf-profile.self.cycles-pp.c_parse_file
0.10 ± 9% +0.0 0.10 ± 4% perf-profile.self.cycles-pp.mark_irreducible_loops
0.08 +0.0 0.09 ± 5% perf-profile.self.cycles-pp.realloc
0.07 +0.0 0.08 ± 6% perf-profile.self.cycles-pp.wi::mul_internal
0.08 ± 10% +0.0 0.09 ± 5% perf-profile.self.cycles-pp.dom_walker::walk
0.06 +0.0 0.07 ± 7% perf-profile.self.cycles-pp.memchr@plt
0.06 +0.0 0.07 ± 7% perf-profile.self.cycles-pp.ferror@plt
0.06 ± 7% +0.0 0.07 ± 11% perf-profile.self.cycles-pp.gimple_fold_stmt_to_constant_1
0.06 ± 7% +0.0 0.07 perf-profile.self.cycles-pp.dom_opt_dom_walker::before_dom_children
0.06 ± 7% +0.0 0.07 perf-profile.self.cycles-pp.delete_trivially_dead_insns
0.06 ± 7% +0.0 0.07 perf-profile.self.cycles-pp.ggc_free
0.06 +0.0 0.07 ± 7% perf-profile.self.cycles-pp.reg_scan
0.06 ± 13% +0.0 0.07 ± 7% perf-profile.self.cycles-pp.unshare_all_rtl_in_chain
0.06 ± 8% +0.0 0.06 ± 7% perf-profile.self.cycles-pp.et_below
0.06 ± 8% +0.0 0.06 ± 7% perf-profile.self.cycles-pp.lra_constraints
0.06 ± 8% +0.0 0.06 ± 7% perf-profile.self.cycles-pp.df_analyze
0.06 ± 8% +0.0 0.06 ± 14% perf-profile.self.cycles-pp.df_grow_reg_info
0.06 ± 8% +0.0 0.06 ± 7% perf-profile.self.cycles-pp.init_subregs_of_mode
0.22 +0.0 0.23 ± 2% perf-profile.self.cycles-pp._cpp_find_file
0.13 +0.0 0.14 ± 3% perf-profile.self.cycles-pp.rcu_cblist_dequeue
0.13 ± 10% +0.0 0.14 ± 3% perf-profile.self.cycles-pp.copy_node
0.09 ± 5% +0.0 0.10 perf-profile.self.cycles-pp.c_token_starts_typename
0.09 ± 5% +0.0 0.10 perf-profile.self.cycles-pp.max_issue
0.07 ± 14% +0.0 0.07 ± 6% perf-profile.self.cycles-pp.lra
0.85 +0.0 0.86 perf-profile.self.cycles-pp.make_node
0.15 ± 3% +0.0 0.16 ± 3% perf-profile.self.cycles-pp.kmem_cache_alloc
0.15 ± 3% +0.0 0.16 ± 6% perf-profile.self.cycles-pp.linemap_resolve_location
0.14 +0.0 0.15 perf-profile.self.cycles-pp.build_stmt
0.11 ± 4% +0.0 0.12 ± 8% perf-profile.self.cycles-pp.linemap_lookup
0.05 +0.0 0.06 perf-profile.self.cycles-pp.c_build_qualified_type
0.05 +0.0 0.06 perf-profile.self.cycles-pp.cgraph_node::finalize_function
0.10 ± 4% +0.0 0.11 ± 4% perf-profile.self.cycles-pp.memcg_slab_post_alloc_hook
0.10 ± 4% +0.0 0.11 ± 4% perf-profile.self.cycles-pp.malloc@plt
0.12 ± 10% +0.0 0.13 ± 3% perf-profile.self.cycles-pp.rewrite_update_dom_walker::before_dom_children
0.09 ± 5% +0.0 0.10 ± 4% perf-profile.self.cycles-pp.filemap_map_pages
0.09 ± 5% +0.0 0.10 ± 9% perf-profile.self.cycles-pp.vn_reference_lookup
0.09 ± 5% +0.0 0.10 ± 4% perf-profile.self.cycles-pp.add_dependence
0.08 ± 6% +0.0 0.09 ± 5% perf-profile.self.cycles-pp.mod_objcg_state
0.08 ± 6% +0.0 0.09 ± 5% perf-profile.self.cycles-pp.hash_rtx_cb
0.07 +0.0 0.08 perf-profile.self.cycles-pp.link_path_walk
0.07 +0.0 0.08 perf-profile.self.cycles-pp.generic_simplify
0.08 ± 6% +0.0 0.09 ± 5% perf-profile.self.cycles-pp.make_location
0.08 ± 6% +0.0 0.09 ± 5% perf-profile.self.cycles-pp.wi::divmod_internal
0.07 +0.0 0.08 ± 10% perf-profile.self.cycles-pp.@plt
0.05 ± 8% +0.0 0.06 ± 7% perf-profile.self.cycles-pp.lra_assign
0.05 ± 8% +0.0 0.06 ± 7% perf-profile.self.cycles-pp.extract_constrain_insn_cached
0.05 ± 8% +0.0 0.06 ± 7% perf-profile.self.cycles-pp.bitmap_clear_bit
0.05 ± 8% +0.0 0.06 ± 7% perf-profile.self.cycles-pp.avail_exprs_stack::lookup_avail_expr
0.05 ± 8% +0.0 0.06 ± 7% perf-profile.self.cycles-pp.lra_final_code_change
0.21 ± 2% +0.0 0.22 ± 2% perf-profile.self.cycles-pp._cpp_pop_context
0.16 +0.0 0.17 perf-profile.self.cycles-pp.declspecs_add_type
0.13 ± 7% +0.0 0.14 ± 3% perf-profile.self.cycles-pp.update_stmt_operands
0.13 ± 3% +0.0 0.14 ± 3% perf-profile.self.cycles-pp.ssa_propagation_engine::ssa_propagate
0.09 ± 5% +0.0 0.10 ± 4% perf-profile.self.cycles-pp.finish_struct
0.08 ± 5% +0.0 0.09 ± 5% perf-profile.self.cycles-pp.place_field
0.10 ± 9% +0.0 0.11 ± 4% perf-profile.self.cycles-pp.inverted_post_order_compute
0.09 ± 5% +0.0 0.10 ± 4% perf-profile.self.cycles-pp.get_attr_type
0.08 ± 5% +0.0 0.09 ± 5% perf-profile.self.cycles-pp.build_int_cst
0.06 +0.0 0.07 perf-profile.self.cycles-pp.find_vma
0.28 +0.0 0.29 perf-profile.self.cycles-pp.convert_lvalue_to_rvalue
0.73 ± 2% +0.0 0.74 perf-profile.self.cycles-pp.pop_scope
0.37 ± 3% +0.0 0.39 ± 4% perf-profile.self.cycles-pp.df_insn_rescan
0.18 ± 2% +0.0 0.20 ± 4% perf-profile.self.cycles-pp.add_stmt
0.19 +0.0 0.20 ± 2% perf-profile.self.cycles-pp.fold_build2_loc
0.19 ± 4% +0.0 0.20 ± 8% perf-profile.self.cycles-pp.get_ref_base_and_extent
0.10 ± 4% +0.0 0.12 ± 4% perf-profile.self.cycles-pp.pop_stmt_list
0.10 ± 4% +0.0 0.12 ± 4% perf-profile.self.cycles-pp.gimplify_expr
0.09 +0.0 0.10 ± 4% perf-profile.self.cycles-pp.cpp_interpret_integer
0.05 ± 8% +0.0 0.07 ± 7% perf-profile.self.cycles-pp.recompute_tree_invariant_for_addr_expr
0.05 +0.0 0.06 ± 7% perf-profile.self.cycles-pp.unmapped_area_topdown
0.15 ± 3% +0.0 0.16 ± 5% perf-profile.self.cycles-pp.type_hash_canon_hash
0.13 ± 3% +0.0 0.15 ± 3% perf-profile.self.cycles-pp.side_effects_p
0.12 +0.0 0.13 ± 3% perf-profile.self.cycles-pp.free@plt
0.12 ± 6% +0.0 0.13 ± 3% perf-profile.self.cycles-pp.inchash::add_expr
0.11 +0.0 0.12 ± 3% perf-profile.self.cycles-pp.ira_init
0.10 ± 8% +0.0 0.11 ± 4% perf-profile.self.cycles-pp.walk_tree_1
0.10 ± 8% +0.0 0.11 ± 4% perf-profile.self.cycles-pp.df_compact_blocks
0.09 ± 5% +0.0 0.11 ± 4% perf-profile.self.cycles-pp.rebuild_jump_labels
0.07 ± 7% +0.0 0.08 perf-profile.self.cycles-pp.main
0.06 ± 7% +0.0 0.08 ± 6% perf-profile.self.cycles-pp.rtx_equal_p
0.12 ± 8% +0.0 0.13 perf-profile.self.cycles-pp.substitute_and_fold_engine::substitute_and_fold
0.06 +0.0 0.07 ± 6% perf-profile.self.cycles-pp.wi::lshift_large
0.06 +0.0 0.07 ± 6% perf-profile.self.cycles-pp.lra_remat
0.38 +0.0 0.40 perf-profile.self.cycles-pp.fold_unary_loc
0.19 ± 4% +0.0 0.21 perf-profile.self.cycles-pp.htab_traverse
0.10 +0.0 0.12 ± 4% perf-profile.self.cycles-pp.fold
0.41 ± 6% +0.0 0.43 perf-profile.self.cycles-pp.df_worklist_dataflow
0.17 ± 8% +0.0 0.19 ± 5% perf-profile.self.cycles-pp.pre_and_rev_post_order_compute_fn
0.07 +0.0 0.09 ± 5% perf-profile.self.cycles-pp.wi::zext_large
0.02 ±141% +0.0 0.03 ± 70% perf-profile.self.cycles-pp.xas_start
0.02 ±141% +0.0 0.03 ± 70% perf-profile.self.cycles-pp.compute_record_mode
0.02 ±141% +0.0 0.03 ± 70% perf-profile.self.cycles-pp.__fsnotify_parent
0.02 ±141% +0.0 0.03 ± 70% perf-profile.self.cycles-pp.__update_load_avg_cfs_rq
0.02 ±141% +0.0 0.03 ± 70% perf-profile.self.cycles-pp.gimple_copy
0.02 ±141% +0.0 0.03 ± 70% perf-profile.self.cycles-pp.update_min_vruntime
0.00 +0.0 0.02 ±141% perf-profile.self.cycles-pp.shorten_branches
0.00 +0.0 0.02 ±141% perf-profile.self.cycles-pp.lra_undo_inheritance
0.00 +0.0 0.02 ±141% perf-profile.self.cycles-pp.build_asm_expr
0.00 +0.0 0.02 ±141% perf-profile.self.cycles-pp.call_rcu
0.00 +0.0 0.02 ±141% perf-profile.self.cycles-pp.c_end_compound_stmt
0.00 +0.0 0.02 ±141% perf-profile.self.cycles-pp.insn_default_length
0.00 +0.0 0.02 ±141% perf-profile.self.cycles-pp.const_binop
0.00 +0.0 0.02 ±141% perf-profile.self.cycles-pp.pos_from_bit
0.05 ± 8% +0.0 0.07 perf-profile.self.cycles-pp.rpo_vn_valueize
0.04 ± 71% +0.0 0.05 ± 8% perf-profile.self.cycles-pp.sd_add_dep
0.12 ± 3% +0.0 0.14 ± 5% perf-profile.self.cycles-pp.warnings_for_convert_and_check
0.11 ± 4% +0.0 0.13 ± 6% perf-profile.self.cycles-pp.sched_rgn_compute_dependencies
0.03 ± 70% +0.0 0.05 perf-profile.self.cycles-pp.__d_alloc
0.03 ± 70% +0.0 0.05 perf-profile.self.cycles-pp.strncpy_from_user
0.03 ± 70% +0.0 0.05 perf-profile.self.cycles-pp.down_write
0.03 ± 70% +0.0 0.05 perf-profile.self.cycles-pp.native_flush_tlb_one_user
0.03 ± 70% +0.0 0.05 perf-profile.self.cycles-pp.expand_call
0.03 ± 70% +0.0 0.05 perf-profile.self.cycles-pp.expand_expr_real_1
0.03 ± 70% +0.0 0.05 perf-profile.self.cycles-pp.entry_SYSRETQ_unsafe_stack
0.17 ± 2% +0.0 0.19 perf-profile.self.cycles-pp.strnlen_user
0.43 ± 5% +0.0 0.45 ± 2% perf-profile.self.cycles-pp.execute_one_pass
0.31 ± 3% +0.0 0.33 perf-profile.self.cycles-pp.build_null_declspecs
0.28 ± 4% +0.0 0.30 perf-profile.self.cycles-pp.calculate_dominance_info
0.19 ± 4% +0.0 0.21 perf-profile.self.cycles-pp.integer_zerop
0.14 ± 3% +0.0 0.16 ± 3% perf-profile.self.cycles-pp._dl_addr
0.18 ± 7% +0.0 0.20 ± 8% perf-profile.self.cycles-pp.diagnostic_report_diagnostic
0.15 ± 3% +0.0 0.17 ± 2% perf-profile.self.cycles-pp.c_parser_declarator
0.14 +0.0 0.16 ± 5% perf-profile.self.cycles-pp.iterative_hash
0.13 ± 3% +0.0 0.15 ± 8% perf-profile.self.cycles-pp._IO_ferror
0.13 ± 6% +0.0 0.15 ± 5% perf-profile.self.cycles-pp.lra_inheritance
0.12 ± 3% +0.0 0.14 ± 3% perf-profile.self.cycles-pp.parser_build_binary_op
0.11 ± 8% +0.0 0.13 ± 3% perf-profile.self.cycles-pp.general_operand
0.05 ± 8% +0.0 0.07 ± 6% perf-profile.self.cycles-pp.size_int_kind
0.35 ± 4% +0.0 0.37 ± 5% perf-profile.self.cycles-pp.gimple_call_flags
0.24 ± 3% +0.0 0.26 perf-profile.self.cycles-pp.get_range_from_loc
0.17 ± 5% +0.0 0.19 ± 5% perf-profile.self.cycles-pp.cselib_lookup
0.33 +0.0 0.35 perf-profile.self.cycles-pp.verify_sequence_points
0.19 ± 2% +0.0 0.22 ± 2% perf-profile.self.cycles-pp.kmem_cache_free
0.20 ± 6% +0.0 0.23 ± 5% perf-profile.self.cycles-pp.regstat_compute_ri
0.12 ± 3% +0.0 0.15 ± 3% perf-profile.self.cycles-pp.build_unary_op
0.08 +0.0 0.10 ± 4% perf-profile.self.cycles-pp.c_parser_skip_until_found
0.03 ± 70% +0.0 0.06 ± 8% perf-profile.self.cycles-pp.insn_extract
0.03 ± 70% +0.0 0.06 ± 8% perf-profile.self.cycles-pp.sched_free_deps
0.21 ± 4% +0.0 0.23 perf-profile.self.cycles-pp.__libc_calloc
0.22 ± 2% +0.0 0.25 ± 3% perf-profile.self.cycles-pp.convert
0.21 ± 4% +0.0 0.23 ± 5% perf-profile.self.cycles-pp.single_def_use_dom_walker::before_dom_children
0.18 ± 4% +0.0 0.21 ± 2% perf-profile.self.cycles-pp.layout_decl
0.10 +0.0 0.13 ± 3% perf-profile.self.cycles-pp.start_decl
0.18 ± 2% +0.0 0.20 ± 4% perf-profile.self.cycles-pp.get_combined_adhoc_loc
0.10 ± 9% +0.0 0.13 perf-profile.self.cycles-pp.c_fully_fold
0.44 +0.0 0.47 perf-profile.self.cycles-pp.fold_binary_loc
0.18 ± 4% +0.0 0.21 perf-profile.self.cycles-pp.rtx_cost
0.12 ± 3% +0.0 0.15 ± 3% perf-profile.self.cycles-pp.c_parser_require
0.28 ± 2% +0.0 0.31 ± 3% perf-profile.self.cycles-pp.default_conversion
0.00 +0.0 0.03 ± 70% perf-profile.self.cycles-pp.cselib_subst_to_values
0.00 +0.0 0.03 ± 70% perf-profile.self.cycles-pp.exp_equiv_p
0.00 +0.0 0.03 ± 70% perf-profile.self.cycles-pp.c_common_signed_or_unsigned_type
0.00 +0.0 0.03 ± 70% perf-profile.self.cycles-pp.__get_user_8
0.00 +0.0 0.03 ± 70% perf-profile.self.cycles-pp.declspecs_add_scspec
0.00 +0.0 0.03 ± 70% perf-profile.self.cycles-pp.c_process_expr_stmt
0.02 ±141% +0.0 0.05 perf-profile.self.cycles-pp.c_determine_visibility
0.02 ±141% +0.0 0.05 perf-profile.self.cycles-pp.parse_input_constraint
0.02 ±141% +0.0 0.05 perf-profile.self.cycles-pp.rest_of_decl_compilation
0.02 ±141% +0.0 0.05 perf-profile.self.cycles-pp.page_add_file_rmap
0.02 ±141% +0.0 0.05 perf-profile.self.cycles-pp.simplify_const_relational_operation
0.02 ±141% +0.0 0.05 perf-profile.self.cycles-pp.recog
0.02 ±141% +0.0 0.05 perf-profile.self.cycles-pp.bitmap_alloc
0.07 +0.0 0.11 ± 4% perf-profile.self.cycles-pp.__cond_resched
0.02 ±141% +0.0 0.05 ± 8% perf-profile.self.cycles-pp.ira_build_conflicts
0.49 +0.0 0.53 ± 2% perf-profile.self.cycles-pp.getenv
0.29 ± 2% +0.0 0.33 perf-profile.self.cycles-pp.pushdecl
0.24 +0.0 0.28 ± 3% perf-profile.self.cycles-pp.fix_string_type
0.24 +0.0 0.28 ± 3% perf-profile.self.cycles-pp.linemap_line_start
0.60 ± 3% +0.0 0.64 ± 2% perf-profile.self.cycles-pp.elf_read
0.23 ± 3% +0.0 0.27 ± 3% perf-profile.self.cycles-pp.bitmap_set_bit
0.43 +0.0 0.47 perf-profile.self.cycles-pp.build_function_type
0.33 ± 2% +0.0 0.37 perf-profile.self.cycles-pp.build2
0.30 ± 2% +0.0 0.34 perf-profile.self.cycles-pp.allocate_struct_function
0.40 ± 5% +0.0 0.45 ± 4% perf-profile.self.cycles-pp.init_alias_analysis
0.67 ± 3% +0.0 0.72 perf-profile.self.cycles-pp.variably_modified_type_p
0.30 +0.0 0.35 ± 2% perf-profile.self.cycles-pp.c_parser_consume_token
0.44 +0.0 0.49 ± 2% perf-profile.self.cycles-pp._cpp_clean_line
0.00 +0.1 0.05 perf-profile.self.cycles-pp.___slab_alloc
0.00 +0.1 0.05 perf-profile.self.cycles-pp.build_component_ref
0.00 +0.1 0.05 perf-profile.self.cycles-pp.shorten_binary_op
0.00 +0.1 0.05 perf-profile.self.cycles-pp.output_asm_insn
0.00 +0.1 0.05 perf-profile.self.cycles-pp.c_objc_common_truthvalue_conversion
0.00 +0.1 0.05 perf-profile.self.cycles-pp.c_parser_peek_2nd_token
0.00 +0.1 0.05 perf-profile.self.cycles-pp.free_pcp_prepare
0.15 ± 3% +0.1 0.21 ± 2% perf-profile.self.cycles-pp.build3
0.00 +0.1 0.06 perf-profile.self.cycles-pp.check_new_pages
0.00 +0.1 0.06 perf-profile.self.cycles-pp.set_cfun
0.54 ± 3% +0.1 0.61 perf-profile.self.cycles-pp._cpp_skip_block_comment
0.43 ± 2% +0.1 0.49 ± 4% perf-profile.self.cycles-pp._cpp_handle_directive
0.11 ± 4% +0.1 0.17 perf-profile.self.cycles-pp.get_parm_info
0.51 +0.1 0.57 perf-profile.self.cycles-pp.operand_equal_p
0.51 +0.1 0.58 ± 2% perf-profile.self.cycles-pp.build_binary_op
0.63 ± 2% +0.1 0.70 ± 5% perf-profile.self.cycles-pp.location_get_source_line
0.57 ± 6% +0.1 0.64 ± 3% perf-profile.self.cycles-pp.df_bb_refs_record
1.00 +0.1 1.07 perf-profile.self.cycles-pp.size_binop_loc
0.00 +0.1 0.08 perf-profile.self.cycles-pp.clear_huge_page
0.41 ± 4% +0.1 0.50 ± 4% perf-profile.self.cycles-pp.constrain_operands
0.41 ± 2% +0.1 0.49 perf-profile.self.cycles-pp.decl_attributes
0.82 +0.1 0.91 perf-profile.self.cycles-pp.cfree
0.66 +0.1 0.75 perf-profile.self.cycles-pp.c_parser_declspecs
0.17 +0.1 0.27 perf-profile.self.cycles-pp.finish_decl
0.58 ± 2% +0.1 0.70 ± 2% perf-profile.self.cycles-pp.layout_type
0.82 ± 3% +0.1 0.94 ± 9% perf-profile.self.cycles-pp.linemap_location_in_system_header_p
0.22 ± 2% +0.1 0.36 perf-profile.self.cycles-pp.__might_resched
1.01 ± 2% +0.2 1.16 perf-profile.self.cycles-pp.build_decl
0.42 +0.2 0.57 perf-profile.self.cycles-pp.wi::force_to_size
0.86 +0.2 1.03 perf-profile.self.cycles-pp.ggc_internal_alloc
1.87 +0.2 2.05 perf-profile.self.cycles-pp.malloc
3.54 ± 2% +0.2 3.75 perf-profile.self.cycles-pp.htab_find_slot_with_hash
1.02 +0.2 1.25 ± 3% perf-profile.self.cycles-pp._cpp_lex_direct
2.04 +0.3 2.29 ± 2% perf-profile.self.cycles-pp._cpp_create_definition
2.54 +0.8 3.36 perf-profile.self.cycles-pp.clear_page_erms
6.10 +1.0 7.05 perf-profile.self.cycles-pp.c_lex_with_flags


Best Regards,
Yujie

>
> Best Regards,
> Huang, Ying
>
> > Time (mean ± σ): 406.688 s ± 0.676 s [User: 31819.526 s, System: 16327.022 s]
> > Range (min … max): 405.954 s … 407.284 s 3 run
> >
> > That is a pretty big difference (27%), which is visible while doing a
> > lot of builds, only because of the extra system time. If there is any
> > way to improve this, it should certainly be considered.
> >
> > For now, I'll just revert it locally.
> >
> > Cheers,
> > Nathan
> >
> > # bad: [aae703b02f92bde9264366c545e87cec451de471] Merge tag 'for-6.1-rc1-tag' of git://git.kernel.org/pub/scm/linux/kernel/git/kdave/linux
> > # good: [4fe89d07dcc2804c8b562f6c7896a45643d34b2f] Linux 6.0
> > git bisect start 'aae703b02f92bde9264366c545e87cec451de471' 'v6.0'
> > # good: [18fd049731e67651009f316195da9281b756f2cf] Merge tag 'arm64-upstream' of git://git.kernel.org/pub/scm/linux/kernel/git/arm64/linux
> > git bisect good 18fd049731e67651009f316195da9281b756f2cf
> > # good: [ab0c23b535f3f9d8345d8ad4c18c0a8594459d55] MAINTAINERS: add RISC-V's patchwork
> > git bisect good ab0c23b535f3f9d8345d8ad4c18c0a8594459d55
> > # bad: [f721d24e5dae8358b49b24399d27ba5d12a7e049] Merge tag 'pull-tmpfile' of git://git.kernel.org/pub/scm/linux/kernel/git/viro/vfs
> > git bisect bad f721d24e5dae8358b49b24399d27ba5d12a7e049
> > # good: [ada3bfb6492a6d0d3eca50f3b61315fe032efc72] Merge tag 'tpmdd-next-v6.1-rc1' of git://git.kernel.org/pub/scm/linux/kernel/git/jarkko/linux-tpmdd
> > git bisect good ada3bfb6492a6d0d3eca50f3b61315fe032efc72
> > # bad: [4e07acdda7fc23f5c4666e54961ef972a1195ffd] mm/hwpoison: add __init/__exit annotations to module init/exit funcs
> > git bisect bad 4e07acdda7fc23f5c4666e54961ef972a1195ffd
> > # bad: [000a449345bbb4ffbd880f7143b5fb4acac34121] radix tree test suite: add allocation counts and size to kmem_cache
> > git bisect bad 000a449345bbb4ffbd880f7143b5fb4acac34121
> > # bad: [47d55419951312d723de1b6ad53ee92948b8eace] btrfs: convert process_page_range() to use filemap_get_folios_contig()
> > git bisect bad 47d55419951312d723de1b6ad53ee92948b8eace
> > # bad: [4d86d4f7227c6f2acfbbbe0623d49865aa71b756] mm: add more BUILD_BUG_ONs to gfp_migratetype()
> > git bisect bad 4d86d4f7227c6f2acfbbbe0623d49865aa71b756
> > # bad: [816284a3d0e27828b5cc35f3cf539b0711939ce3] userfaultfd: update documentation to describe /dev/userfaultfd
> > git bisect bad 816284a3d0e27828b5cc35f3cf539b0711939ce3
> > # good: [be6667b0db97e10b2a6d57a906c2c8fd2b985e5e] selftests/vm: dedup hugepage allocation logic
> > git bisect good be6667b0db97e10b2a6d57a906c2c8fd2b985e5e
> > # bad: [2ace36f0f55777be8a871c370832527e1cd54b15] mm: memory-failure: cleanup try_to_split_thp_page()
> > git bisect bad 2ace36f0f55777be8a871c370832527e1cd54b15
> > # good: [9d0d946840075e0268f4f77fe39ba0f53e84c7c4] selftests/vm: add selftest to verify multi THP collapse
> > git bisect good 9d0d946840075e0268f4f77fe39ba0f53e84c7c4
> > # bad: [f35b5d7d676e59e401690b678cd3cfec5e785c23] mm: align larger anonymous mappings on THP boundaries
> > git bisect bad f35b5d7d676e59e401690b678cd3cfec5e785c23
> > # good: [7b5a0b664ebe2625965a0fdba2614c88c4b9bbc6] mm/page_ext: remove unused variable in offline_page_ext
> > git bisect good 7b5a0b664ebe2625965a0fdba2614c88c4b9bbc6
> > # first bad commit: [f35b5d7d676e59e401690b678cd3cfec5e785c23] mm: align larger anonymous mappings on THP boundaries

2022-10-20 17:57:54

by Nathan Chancellor

[permalink] [raw]
Subject: Re: [mm] f35b5d7d67: will-it-scale.per_process_ops -95.5% regression

Hi Rik,

On Thu, Oct 20, 2022 at 11:28:16AM -0400, Rik van Riel wrote:
> On Thu, 2022-10-20 at 13:07 +0800, Huang, Ying wrote:
> >
> > Nathan Chancellor <[email protected]> writes:
> > >
> > > For what it's worth, I just bisected a massive and visible
> > > performance
> > > regression on my Threadripper 3990X workstation to commit
> > > f35b5d7d676e
> > > ("mm: align larger anonymous mappings on THP boundaries"), which
> > > seems
> > > directly related to this report/analysis. I initially noticed this
> > > because my full set of kernel builds against mainline went from 2
> > > hours
> > > and 20 minutes or so to over 3 hours. Zeroing in on x86_64
> > > allmodconfig,
> > > which I used for the bisect:
> > >
> > > @ 7b5a0b664ebe ("mm/page_ext: remove unused variable in
> > > offline_page_ext"):
> > >
> > > Benchmark 1: make -skj128 LLVM=1 allmodconfig all
> > >   Time (mean ± σ):     318.172 s ±  0.730 s    [User: 31750.902 s,
> > > System: 4564.246 s]
> > >   Range (min … max):   317.332 s … 318.662 s    3 runs
> > >
> > > @ f35b5d7d676e ("mm: align larger anonymous mappings on THP
> > > boundaries"):
> > >
> > > Benchmark 1: make -skj128 LLVM=1 allmodconfig all
> > > Time (mean ± σ): 406.688 s ± 0.676 s [User: 31819.526 s,
> System: 16327.022 s]
> > > Range (min … max): 405.954 s … 407.284 s 3 run
> >
> > Have you tried to build with gcc?  Want to check whether is this
> > clang
> > specific issue or not.
>
> This may indeed be something LLVM specific. In previous tests,
> GCC has generally seen a benefit from increased THP usage.
> Many other applications also benefit from getting more THPs.

Indeed, GCC builds actually appear to be slightly faster on my system now,
apologies for not trying that before reporting :/

7b5a0b664ebe:

Benchmark 1: make -skj128 allmodconfig all
Time (mean ± σ): 355.294 s ± 0.931 s [User: 33620.469 s, System: 6390.064 s]
Range (min … max): 354.571 s … 356.344 s 3 runs

f35b5d7d676e:

Benchmark 1: make -skj128 allmodconfig all
Time (mean ± σ): 347.400 s ± 2.029 s [User: 34389.724 s, System: 4603.175 s]
Range (min … max): 345.815 s … 349.686 s 3 runs

> LLVM showing 10% system time before this change, and a whopping
> 30% system time after that change, suggests that LLVM is behaving
> quite differently from GCC in some ways.

The above tests were done with GCC 12.2.0 from Arch Linux. The previous LLVM
tests were done with a self-compiled version of LLVM from the main branch
(16.0.0), optimized with BOLT [1]. To eliminate that as a source of issues, I
used my distribution's version of clang (14.0.6) and saw similar results as
before:

7b5a0b664ebe:

Benchmark 1: make -skj128 LLVM=/usr/bin/ allmodconfig all
Time (mean ± σ): 462.517 s ± 1.214 s [User: 48544.240 s, System: 5586.212 s]
Range (min … max): 461.115 s … 463.245 s 3 runs

f35b5d7d676e:

Benchmark 1: make -skj128 LLVM=/usr/bin/ allmodconfig all
Time (mean ± σ): 547.927 s ± 0.862 s [User: 47913.709 s, System: 17682.514 s]
Range (min … max): 547.429 s … 548.922 s 3 runs

> If we can figure out what these differences are, maybe we can
> just fine tune the code to avoid this issue.
>
> I'll try to play around with LLVM compilation a little bit next
> week, to see if I can figure out what might be going on. I wonder
> if LLVM is doing lots of mremap calls or something...

If there is any further information I can provide or patches I can test,
I am more than happy to do so.

[1]: https://github.com/llvm/llvm-project/tree/96552e73900176d65ee6650facae8d669d6f9498/bolt

Cheers,
Nathan

2022-11-28 06:48:54

by Nathan Chancellor

[permalink] [raw]
Subject: Re: [mm] f35b5d7d67: will-it-scale.per_process_ops -95.5% regression

Hi Rik,

On Thu, Oct 20, 2022 at 10:16:20AM -0700, Nathan Chancellor wrote:
> On Thu, Oct 20, 2022 at 11:28:16AM -0400, Rik van Riel wrote:
> > On Thu, 2022-10-20 at 13:07 +0800, Huang, Ying wrote:
> > >
> > > Nathan Chancellor <[email protected]> writes:
> > > >
> > > > For what it's worth, I just bisected a massive and visible
> > > > performance
> > > > regression on my Threadripper 3990X workstation to commit
> > > > f35b5d7d676e
> > > > ("mm: align larger anonymous mappings on THP boundaries"), which
> > > > seems
> > > > directly related to this report/analysis. I initially noticed this
> > > > because my full set of kernel builds against mainline went from 2
> > > > hours
> > > > and 20 minutes or so to over 3 hours. Zeroing in on x86_64
> > > > allmodconfig,
> > > > which I used for the bisect:
> > > >
> > > > @ 7b5a0b664ebe ("mm/page_ext: remove unused variable in
> > > > offline_page_ext"):
> > > >
> > > > Benchmark 1: make -skj128 LLVM=1 allmodconfig all
> > > >   Time (mean ± σ):     318.172 s ±  0.730 s    [User: 31750.902 s,
> > > > System: 4564.246 s]
> > > >   Range (min … max):   317.332 s … 318.662 s    3 runs
> > > >
> > > > @ f35b5d7d676e ("mm: align larger anonymous mappings on THP
> > > > boundaries"):
> > > >
> > > > Benchmark 1: make -skj128 LLVM=1 allmodconfig all
> > > > Time (mean ± σ): 406.688 s ± 0.676 s [User: 31819.526 s,
> > System: 16327.022 s]
> > > > Range (min … max): 405.954 s … 407.284 s 3 run
> > >
> > > Have you tried to build with gcc?  Want to check whether is this
> > > clang
> > > specific issue or not.
> >
> > This may indeed be something LLVM specific. In previous tests,
> > GCC has generally seen a benefit from increased THP usage.
> > Many other applications also benefit from getting more THPs.
>
> Indeed, GCC builds actually appear to be slightly faster on my system now,
> apologies for not trying that before reporting :/
>
> 7b5a0b664ebe:
>
> Benchmark 1: make -skj128 allmodconfig all
> Time (mean ± σ): 355.294 s ± 0.931 s [User: 33620.469 s, System: 6390.064 s]
> Range (min … max): 354.571 s … 356.344 s 3 runs
>
> f35b5d7d676e:
>
> Benchmark 1: make -skj128 allmodconfig all
> Time (mean ± σ): 347.400 s ± 2.029 s [User: 34389.724 s, System: 4603.175 s]
> Range (min … max): 345.815 s … 349.686 s 3 runs
>
> > LLVM showing 10% system time before this change, and a whopping
> > 30% system time after that change, suggests that LLVM is behaving
> > quite differently from GCC in some ways.
>
> The above tests were done with GCC 12.2.0 from Arch Linux. The previous LLVM
> tests were done with a self-compiled version of LLVM from the main branch
> (16.0.0), optimized with BOLT [1]. To eliminate that as a source of issues, I
> used my distribution's version of clang (14.0.6) and saw similar results as
> before:
>
> 7b5a0b664ebe:
>
> Benchmark 1: make -skj128 LLVM=/usr/bin/ allmodconfig all
> Time (mean ± σ): 462.517 s ± 1.214 s [User: 48544.240 s, System: 5586.212 s]
> Range (min … max): 461.115 s … 463.245 s 3 runs
>
> f35b5d7d676e:
>
> Benchmark 1: make -skj128 LLVM=/usr/bin/ allmodconfig all
> Time (mean ± σ): 547.927 s ± 0.862 s [User: 47913.709 s, System: 17682.514 s]
> Range (min … max): 547.429 s … 548.922 s 3 runs
>
> > If we can figure out what these differences are, maybe we can
> > just fine tune the code to avoid this issue.
> >
> > I'll try to play around with LLVM compilation a little bit next
> > week, to see if I can figure out what might be going on. I wonder
> > if LLVM is doing lots of mremap calls or something...
>
> If there is any further information I can provide or patches I can test,
> I am more than happy to do so.
>
> [1]: https://github.com/llvm/llvm-project/tree/96552e73900176d65ee6650facae8d669d6f9498/bolt

Was there ever a follow up to this report that I missed? I just
noticed that I am still reverting f35b5d7d676e in my mainline kernel.

Cheers,
Nathan

2022-11-29 09:15:16

by Thorsten Leemhuis

[permalink] [raw]
Subject: Re: [mm] f35b5d7d67: will-it-scale.per_process_ops -95.5% regression #forregzbot

[Note: this mail is primarily send for documentation purposes and/or for
regzbot, my Linux kernel regression tracking bot. That's why I removed
most or all folks from the list of recipients, but left any that looked
like a mailing lists. These mails usually contain '#forregzbot' in the
subject, to make them easy to spot and filter out.]

[TLDR: I'm adding this regression report to the list of tracked
regressions; all text from me you find below is based on a few templates
paragraphs you might have encountered already already in similar form.]

Hi, this is your Linux kernel regression tracker.

On 20.10.22 06:23, Nathan Chancellor wrote:
> [...]
> For what it's worth, I just bisected a massive and visible performance
> regression on my Threadripper 3990X workstation to commit f35b5d7d676e
> ("mm: align larger anonymous mappings on THP boundaries"), which seems
> directly related to this report/analysis. I initially noticed this
> because my full set of kernel builds against mainline went from 2 hours
> and 20 minutes or so to over 3 hours. Zeroing in on x86_64 allmodconfig,
> which I used for the bisect:

Thanks for the report. To be sure below issue doesn't fall through the
cracks unnoticed, I'm adding it to regzbot, my Linux kernel regression
tracking bot:

#regzbot ^introduced f35b5d7d676e59e4
#regzbot title performance: mm: building Linux with clang takes a lot
longer (~27 %)
#regzbot ignore-activity

This isn't a regression? This issue or a fix for it are already
discussed somewhere else? It was fixed already? You want to clarify when
the regression started to happen? Or point out I got the title or
something else totally wrong? Then just reply -- ideally with also
telling regzbot about it, as explained here:
https://linux-regtracking.leemhuis.info/tracked-regression/

Reminder for developers: When fixing the issue, add 'Link:' tags
pointing to the report (the mail this one replies to), as explained for
in the Linux kernel's documentation; above webpage explains why this is
important for tracked regressions.

Ciao, Thorsten (wearing his 'the Linux kernel's regression tracker' hat)

P.S.: As the Linux kernel's regression tracker I deal with a lot of
reports and sometimes miss something important when writing mails like
this. If that's the case here, don't hesitate to tell me in a public
reply, it's in everyone's interest to set the public record straight.

2022-12-01 19:27:05

by Thorsten Leemhuis

[permalink] [raw]
Subject: Re: [mm] f35b5d7d67: will-it-scale.per_process_ops -95.5% regression

Hi, this is your Linux kernel regression tracker.

On 28.11.22 07:40, Nathan Chancellor wrote:
> Hi Rik,

I wonder what we should do about below performance regression. Is
reverting the culprit now and reapplying it later together with a fix a
viable option? Or was anything done/is anybody doing something already
to address the problem and I just missed it?

Yang Shi, Andrew, what's your option on this? I ask you directly,
because it looks like Rik hasn't posted anything to lists archived on
lore during the last few weeks. :-/

Ciao, Thorsten (wearing his 'the Linux kernel's regression tracker' hat)

P.S.: As the Linux kernel's regression tracker I deal with a lot of
reports and sometimes miss something important when writing mails like
this. If that's the case here, don't hesitate to tell me in a public
reply, it's in everyone's interest to set the public record straight.

> On Thu, Oct 20, 2022 at 10:16:20AM -0700, Nathan Chancellor wrote:
>> On Thu, Oct 20, 2022 at 11:28:16AM -0400, Rik van Riel wrote:
>>> On Thu, 2022-10-20 at 13:07 +0800, Huang, Ying wrote:
>>>> Nathan Chancellor <[email protected]> writes:
>>>>>
>>>>> For what it's worth, I just bisected a massive and visible
>>>>> performance
>>>>> regression on my Threadripper 3990X workstation to commit
>>>>> f35b5d7d676e
>>>>> ("mm: align larger anonymous mappings on THP boundaries"), which
>>>>> seems
>>>>> directly related to this report/analysis. I initially noticed this
>>>>> because my full set of kernel builds against mainline went from 2
>>>>> hours
>>>>> and 20 minutes or so to over 3 hours. Zeroing in on x86_64
>>>>> allmodconfig,
>>>>> which I used for the bisect:
>>>>>
>>>>> @ 7b5a0b664ebe ("mm/page_ext: remove unused variable in
>>>>> offline_page_ext"):
>>>>>
>>>>> Benchmark 1: make -skj128 LLVM=1 allmodconfig all
>>>>>   Time (mean ± σ):     318.172 s ±  0.730 s    [User: 31750.902 s,
>>>>> System: 4564.246 s]
>>>>>   Range (min … max):   317.332 s … 318.662 s    3 runs
>>>>>
>>>>> @ f35b5d7d676e ("mm: align larger anonymous mappings on THP
>>>>> boundaries"):
>>>>>
>>>>> Benchmark 1: make -skj128 LLVM=1 allmodconfig all
>>>>> Time (mean ± σ): 406.688 s ± 0.676 s [User: 31819.526 s,
>>> System: 16327.022 s]
>>>>> Range (min … max): 405.954 s … 407.284 s 3 run
>>>>
>>>> Have you tried to build with gcc?  Want to check whether is this
>>>> clang
>>>> specific issue or not.
>>>
>>> This may indeed be something LLVM specific. In previous tests,
>>> GCC has generally seen a benefit from increased THP usage.
>>> Many other applications also benefit from getting more THPs.
>>
>> Indeed, GCC builds actually appear to be slightly faster on my system now,
>> apologies for not trying that before reporting :/
>>
>> 7b5a0b664ebe:
>>
>> Benchmark 1: make -skj128 allmodconfig all
>> Time (mean ± σ): 355.294 s ± 0.931 s [User: 33620.469 s, System: 6390.064 s]
>> Range (min … max): 354.571 s … 356.344 s 3 runs
>>
>> f35b5d7d676e:
>>
>> Benchmark 1: make -skj128 allmodconfig all
>> Time (mean ± σ): 347.400 s ± 2.029 s [User: 34389.724 s, System: 4603.175 s]
>> Range (min … max): 345.815 s … 349.686 s 3 runs
>>
>>> LLVM showing 10% system time before this change, and a whopping
>>> 30% system time after that change, suggests that LLVM is behaving
>>> quite differently from GCC in some ways.
>>
>> The above tests were done with GCC 12.2.0 from Arch Linux. The previous LLVM
>> tests were done with a self-compiled version of LLVM from the main branch
>> (16.0.0), optimized with BOLT [1]. To eliminate that as a source of issues, I
>> used my distribution's version of clang (14.0.6) and saw similar results as
>> before:
>>
>> 7b5a0b664ebe:
>>
>> Benchmark 1: make -skj128 LLVM=/usr/bin/ allmodconfig all
>> Time (mean ± σ): 462.517 s ± 1.214 s [User: 48544.240 s, System: 5586.212 s]
>> Range (min … max): 461.115 s … 463.245 s 3 runs
>>
>> f35b5d7d676e:
>>
>> Benchmark 1: make -skj128 LLVM=/usr/bin/ allmodconfig all
>> Time (mean ± σ): 547.927 s ± 0.862 s [User: 47913.709 s, System: 17682.514 s]
>> Range (min … max): 547.429 s … 548.922 s 3 runs
>>
>>> If we can figure out what these differences are, maybe we can
>>> just fine tune the code to avoid this issue.
>>>
>>> I'll try to play around with LLVM compilation a little bit next
>>> week, to see if I can figure out what might be going on. I wonder
>>> if LLVM is doing lots of mremap calls or something...
>>
>> If there is any further information I can provide or patches I can test,
>> I am more than happy to do so.
>>
>> [1]: https://github.com/llvm/llvm-project/tree/96552e73900176d65ee6650facae8d669d6f9498/bolt
>
> Was there ever a follow up to this report that I missed? I just
> noticed that I am still reverting f35b5d7d676e in my mainline kernel.
>
> Cheers,
> Nathan
>

#regzbot ignore-activity

2022-12-01 21:14:02

by Rik van Riel

[permalink] [raw]
Subject: Re: [mm] f35b5d7d67: will-it-scale.per_process_ops -95.5% regression

On Thu, 2022-12-01 at 19:33 +0100, Thorsten Leemhuis wrote:
> Hi, this is your Linux kernel regression tracker.
>
> On 28.11.22 07:40, Nathan Chancellor wrote:
> > Hi Rik,
>
> I wonder what we should do about below performance regression. Is
> reverting the culprit now and reapplying it later together with a fix
> a
> viable option? Or was anything done/is anybody doing something
> already
> to address the problem and I just missed it?

The changeset in question speeds up kernel compiles with
GCC, as well as the runtime speed of other programs, due
to being able to use THPs more. However, it slows down kernel
compiles with clang, due to ... something clang does.

I have not figured out what that something is yet.

I don't know if I have the wrong version of clang here,
but I have not seen any smoking gun at all when tracing
clang system calls. I see predominantly small mmap and
unmap calls, and nothing that even triggers 2MB alignment.

>
> Yang Shi, Andrew, what's your option on this? I ask you directly,
> because it looks like Rik hasn't posted anything to lists archived on
> lore during the last few weeks. :-/
>
> Ciao, Thorsten (wearing his 'the Linux kernel's regression tracker'
> hat)
>
> P.S.: As the Linux kernel's regression tracker I deal with a lot of
> reports and sometimes miss something important when writing mails
> like
> this. If that's the case here, don't hesitate to tell me in a public
> reply, it's in everyone's interest to set the public record straight.
>
> > On Thu, Oct 20, 2022 at 10:16:20AM -0700, Nathan Chancellor wrote:
> > > On Thu, Oct 20, 2022 at 11:28:16AM -0400, Rik van Riel wrote:
> > > > On Thu, 2022-10-20 at 13:07 +0800, Huang, Ying wrote:
> > > > > Nathan Chancellor <[email protected]> writes:
> > > > > >
> > > > > > For what it's worth, I just bisected a massive and visible
> > > > > > performance
> > > > > > regression on my Threadripper 3990X workstation to commit
> > > > > > f35b5d7d676e
> > > > > > ("mm: align larger anonymous mappings on THP boundaries"),
> > > > > > which
> > > > > > seems
> > > > > > directly related to this report/analysis. I initially
> > > > > > noticed this
> > > > > > because my full set of kernel builds against mainline went
> > > > > > from 2
> > > > > > hours
> > > > > > and 20 minutes or so to over 3 hours. Zeroing in on x86_64
> > > > > > allmodconfig,
> > > > > > which I used for the bisect:
> > > > > >
> > > > > > @ 7b5a0b664ebe ("mm/page_ext: remove unused variable in
> > > > > > offline_page_ext"):
> > > > > >
> > > > > > Benchmark 1: make -skj128 LLVM=1 allmodconfig all
> > > > > >   Time (mean ± σ):     318.172 s ±  0.730 s    [User:
> > > > > > 31750.902 s,
> > > > > > System: 4564.246 s]
> > > > > >   Range (min … max):   317.332 s … 318.662 s    3 runs
> > > > > >
> > > > > > @ f35b5d7d676e ("mm: align larger anonymous mappings on THP
> > > > > > boundaries"):
> > > > > >
> > > > > > Benchmark 1: make -skj128 LLVM=1 allmodconfig all
> > > > > >   Time (mean ± σ):     406.688 s ±  0.676 s    [User:
> > > > > > 31819.526 s,
> > > > System: 16327.022 s]
> > > > > >   Range (min … max):   405.954 s … 407.284 s    3 run
> > > > >
> > > > > Have you tried to build with gcc?  Want to check whether is
> > > > > this
> > > > > clang
> > > > > specific issue or not.
> > > >
> > > > This may indeed be something LLVM specific. In previous tests,
> > > > GCC has generally seen a benefit from increased THP usage.
> > > > Many other applications also benefit from getting more THPs.
> > >
> > > Indeed, GCC builds actually appear to be slightly faster on my
> > > system now,
> > > apologies for not trying that before reporting :/
> > >
> > > 7b5a0b664ebe:
> > >
> > > Benchmark 1: make -skj128 allmodconfig all
> > >   Time (mean ± σ):     355.294 s ±  0.931 s    [User: 33620.469
> > > s, System: 6390.064 s]
> > >   Range (min … max):   354.571 s … 356.344 s    3 runs
> > >
> > > f35b5d7d676e:
> > >
> > > Benchmark 1: make -skj128 allmodconfig all
> > >   Time (mean ± σ):     347.400 s ±  2.029 s    [User: 34389.724
> > > s, System: 4603.175 s]
> > >   Range (min … max):   345.815 s … 349.686 s    3 runs
> > >
> > > > LLVM showing 10% system time before this change, and a whopping
> > > > 30% system time after that change, suggests that LLVM is
> > > > behaving
> > > > quite differently from GCC in some ways.
> > >
> > > The above tests were done with GCC 12.2.0 from Arch Linux. The
> > > previous LLVM
> > > tests were done with a self-compiled version of LLVM from the
> > > main branch
> > > (16.0.0), optimized with BOLT [1]. To eliminate that as a source
> > > of issues, I
> > > used my distribution's version of clang (14.0.6) and saw similar
> > > results as
> > > before:
> > >
> > > 7b5a0b664ebe:
> > >
> > > Benchmark 1: make -skj128 LLVM=/usr/bin/ allmodconfig all
> > >   Time (mean ± σ):     462.517 s ±  1.214 s    [User: 48544.240
> > > s, System: 5586.212 s]
> > >   Range (min … max):   461.115 s … 463.245 s    3 runs
> > >
> > > f35b5d7d676e:
> > >
> > > Benchmark 1: make -skj128 LLVM=/usr/bin/ allmodconfig all
> > >   Time (mean ± σ):     547.927 s ±  0.862 s    [User: 47913.709
> > > s, System: 17682.514 s]
> > >   Range (min … max):   547.429 s … 548.922 s    3 runs
> > >
> > > > If we can figure out what these differences are, maybe we can
> > > > just fine tune the code to avoid this issue.
> > > >
> > > > I'll try to play around with LLVM compilation a little bit next
> > > > week, to see if I can figure out what might be going on. I
> > > > wonder
> > > > if LLVM is doing lots of mremap calls or something...
> > >
> > > If there is any further information I can provide or patches I
> > > can test,
> > > I am more than happy to do so.
> > >
> > > [1]:
> > > https://github.com/llvm/llvm-project/tree/96552e73900176d65ee6650facae8d669d6f9498/bolt
> >
> > Was there ever a follow up to this report that I missed? I just
> > noticed that I am still reverting f35b5d7d676e in my mainline
> > kernel.
> >
> > Cheers,
> > Nathan
> >
>
> #regzbot ignore-activity
>

--
All Rights Reversed.


Attachments:
signature.asc (499.00 B)
This is a digitally signed message part

2022-12-01 21:28:56

by Andrew Morton

[permalink] [raw]
Subject: Re: [mm] f35b5d7d67: will-it-scale.per_process_ops -95.5% regression

On Thu, 01 Dec 2022 15:29:41 -0500 Rik van Riel <[email protected]> wrote:

> On Thu, 2022-12-01 at 19:33 +0100, Thorsten Leemhuis wrote:
> > Hi, this is your Linux kernel regression tracker.
> >
> > On 28.11.22 07:40, Nathan Chancellor wrote:
> > > Hi Rik,
> >
> > I wonder what we should do about below performance regression. Is
> > reverting the culprit now and reapplying it later together with a fix
> > a
> > viable option? Or was anything done/is anybody doing something
> > already
> > to address the problem and I just missed it?
>
> The changeset in question speeds up kernel compiles with
> GCC, as well as the runtime speed of other programs, due
> to being able to use THPs more. However, it slows down kernel
> compiles with clang, due to ... something clang does.
>
> I have not figured out what that something is yet.
>
> I don't know if I have the wrong version of clang here,
> but I have not seen any smoking gun at all when tracing
> clang system calls. I see predominantly small mmap and
> unmap calls, and nothing that even triggers 2MB alignment.

2.8% speedup for gcc is nice. Massive slowdown in the malloc banchmark
and in LLVM/clang is very bad - we don't know what other userspace will
be so affected.

So I think we revert until this is fully understood.


2022-12-01 22:01:35

by Yang Shi

[permalink] [raw]
Subject: Re: [mm] f35b5d7d67: will-it-scale.per_process_ops -95.5% regression

On Thu, Dec 1, 2022 at 1:22 PM Andrew Morton <[email protected]> wrote:
>
> On Thu, 01 Dec 2022 15:29:41 -0500 Rik van Riel <[email protected]> wrote:
>
> > On Thu, 2022-12-01 at 19:33 +0100, Thorsten Leemhuis wrote:
> > > Hi, this is your Linux kernel regression tracker.
> > >
> > > On 28.11.22 07:40, Nathan Chancellor wrote:
> > > > Hi Rik,
> > >
> > > I wonder what we should do about below performance regression. Is
> > > reverting the culprit now and reapplying it later together with a fix
> > > a
> > > viable option? Or was anything done/is anybody doing something
> > > already
> > > to address the problem and I just missed it?
> >
> > The changeset in question speeds up kernel compiles with
> > GCC, as well as the runtime speed of other programs, due
> > to being able to use THPs more. However, it slows down kernel
> > compiles with clang, due to ... something clang does.
> >
> > I have not figured out what that something is yet.
> >
> > I don't know if I have the wrong version of clang here,
> > but I have not seen any smoking gun at all when tracing
> > clang system calls. I see predominantly small mmap and
> > unmap calls, and nothing that even triggers 2MB alignment.
>
> 2.8% speedup for gcc is nice. Massive slowdown in the malloc banchmark
> and in LLVM/clang is very bad - we don't know what other userspace will
> be so affected.
>
> So I think we revert until this is fully understood.

+1

>
>

2022-12-01 22:03:06

by Nathan Chancellor

[permalink] [raw]
Subject: Re: [mm] f35b5d7d67: will-it-scale.per_process_ops -95.5% regression

On Thu, Dec 01, 2022 at 03:29:41PM -0500, Rik van Riel wrote:
> On Thu, 2022-12-01 at 19:33 +0100, Thorsten Leemhuis wrote:
> > Hi, this is your Linux kernel regression tracker.
> >
> > On 28.11.22 07:40, Nathan Chancellor wrote:
> > > Hi Rik,
> >
> > I wonder what we should do about below performance regression. Is
> > reverting the culprit now and reapplying it later together with a fix
> > a
> > viable option? Or was anything done/is anybody doing something
> > already
> > to address the problem and I just missed it?
>
> The changeset in question speeds up kernel compiles with
> GCC, as well as the runtime speed of other programs, due
> to being able to use THPs more. However, it slows down kernel
> compiles with clang, due to ... something clang does.
>
> I have not figured out what that something is yet.
>
> I don't know if I have the wrong version of clang here,
> but I have not seen any smoking gun at all when tracing
> clang system calls. I see predominantly small mmap and
> unmap calls, and nothing that even triggers 2MB alignment.

Sorry about that :/ What version of clang are you trying to reproduce
with? I was able to see this with 14.x and 16.x, it is possible that
older versions do not do the thing that is causing this.

I cannot really be testing much on my main workstation but I will see if
I can reproduce this behavior on one of my other test systems or a
virtual machine. Once I do that, if you are still unable to reproduce
it, I can potentially try and help you debug this, although I will
likely need some hand holding.

Cheers,
Nathan

> > Yang Shi, Andrew, what's your option on this? I ask you directly,
> > because it looks like Rik hasn't posted anything to lists archived on
> > lore during the last few weeks. :-/
> >
> > Ciao, Thorsten (wearing his 'the Linux kernel's regression tracker'
> > hat)
> >
> > P.S.: As the Linux kernel's regression tracker I deal with a lot of
> > reports and sometimes miss something important when writing mails
> > like
> > this. If that's the case here, don't hesitate to tell me in a public
> > reply, it's in everyone's interest to set the public record straight.
> >
> > > On Thu, Oct 20, 2022 at 10:16:20AM -0700, Nathan Chancellor wrote:
> > > > On Thu, Oct 20, 2022 at 11:28:16AM -0400, Rik van Riel wrote:
> > > > > On Thu, 2022-10-20 at 13:07 +0800, Huang, Ying wrote:
> > > > > > Nathan Chancellor <[email protected]> writes:
> > > > > > >
> > > > > > > For what it's worth, I just bisected a massive and visible
> > > > > > > performance
> > > > > > > regression on my Threadripper 3990X workstation to commit
> > > > > > > f35b5d7d676e
> > > > > > > ("mm: align larger anonymous mappings on THP boundaries"),
> > > > > > > which
> > > > > > > seems
> > > > > > > directly related to this report/analysis. I initially
> > > > > > > noticed this
> > > > > > > because my full set of kernel builds against mainline went
> > > > > > > from 2
> > > > > > > hours
> > > > > > > and 20 minutes or so to over 3 hours. Zeroing in on x86_64
> > > > > > > allmodconfig,
> > > > > > > which I used for the bisect:
> > > > > > >
> > > > > > > @ 7b5a0b664ebe ("mm/page_ext: remove unused variable in
> > > > > > > offline_page_ext"):
> > > > > > >
> > > > > > > Benchmark 1: make -skj128 LLVM=1 allmodconfig all
> > > > > > >   Time (mean ± σ):     318.172 s ±  0.730 s    [User:
> > > > > > > 31750.902 s,
> > > > > > > System: 4564.246 s]
> > > > > > >   Range (min … max):   317.332 s … 318.662 s    3 runs
> > > > > > >
> > > > > > > @ f35b5d7d676e ("mm: align larger anonymous mappings on THP
> > > > > > > boundaries"):
> > > > > > >
> > > > > > > Benchmark 1: make -skj128 LLVM=1 allmodconfig all
> > > > > > >   Time (mean ± σ):     406.688 s ±  0.676 s    [User:
> > > > > > > 31819.526 s,
> > > > > System: 16327.022 s]
> > > > > > >   Range (min … max):   405.954 s … 407.284 s    3 run
> > > > > >
> > > > > > Have you tried to build with gcc?  Want to check whether is
> > > > > > this
> > > > > > clang
> > > > > > specific issue or not.
> > > > >
> > > > > This may indeed be something LLVM specific. In previous tests,
> > > > > GCC has generally seen a benefit from increased THP usage.
> > > > > Many other applications also benefit from getting more THPs.
> > > >
> > > > Indeed, GCC builds actually appear to be slightly faster on my
> > > > system now,
> > > > apologies for not trying that before reporting :/
> > > >
> > > > 7b5a0b664ebe:
> > > >
> > > > Benchmark 1: make -skj128 allmodconfig all
> > > >   Time (mean ± σ):     355.294 s ±  0.931 s    [User: 33620.469
> > > > s, System: 6390.064 s]
> > > >   Range (min … max):   354.571 s … 356.344 s    3 runs
> > > >
> > > > f35b5d7d676e:
> > > >
> > > > Benchmark 1: make -skj128 allmodconfig all
> > > >   Time (mean ± σ):     347.400 s ±  2.029 s    [User: 34389.724
> > > > s, System: 4603.175 s]
> > > >   Range (min … max):   345.815 s … 349.686 s    3 runs
> > > >
> > > > > LLVM showing 10% system time before this change, and a whopping
> > > > > 30% system time after that change, suggests that LLVM is
> > > > > behaving
> > > > > quite differently from GCC in some ways.
> > > >
> > > > The above tests were done with GCC 12.2.0 from Arch Linux. The
> > > > previous LLVM
> > > > tests were done with a self-compiled version of LLVM from the
> > > > main branch
> > > > (16.0.0), optimized with BOLT [1]. To eliminate that as a source
> > > > of issues, I
> > > > used my distribution's version of clang (14.0.6) and saw similar
> > > > results as
> > > > before:
> > > >
> > > > 7b5a0b664ebe:
> > > >
> > > > Benchmark 1: make -skj128 LLVM=/usr/bin/ allmodconfig all
> > > >   Time (mean ± σ):     462.517 s ±  1.214 s    [User: 48544.240
> > > > s, System: 5586.212 s]
> > > >   Range (min … max):   461.115 s … 463.245 s    3 runs
> > > >
> > > > f35b5d7d676e:
> > > >
> > > > Benchmark 1: make -skj128 LLVM=/usr/bin/ allmodconfig all
> > > >   Time (mean ± σ):     547.927 s ±  0.862 s    [User: 47913.709
> > > > s, System: 17682.514 s]
> > > >   Range (min … max):   547.429 s … 548.922 s    3 runs
> > > >
> > > > > If we can figure out what these differences are, maybe we can
> > > > > just fine tune the code to avoid this issue.
> > > > >
> > > > > I'll try to play around with LLVM compilation a little bit next
> > > > > week, to see if I can figure out what might be going on. I
> > > > > wonder
> > > > > if LLVM is doing lots of mremap calls or something...
> > > >
> > > > If there is any further information I can provide or patches I
> > > > can test,
> > > > I am more than happy to do so.
> > > >
> > > > [1]:
> > > > https://github.com/llvm/llvm-project/tree/96552e73900176d65ee6650facae8d669d6f9498/bolt
> > >
> > > Was there ever a follow up to this report that I missed? I just
> > > noticed that I am still reverting f35b5d7d676e in my mainline
> > > kernel.
> > >
> > > Cheers,
> > > Nathan
> > >
> >
> > #regzbot ignore-activity
> >
>
> --
> All Rights Reversed.


2022-12-02 07:48:33

by Thorsten Leemhuis

[permalink] [raw]
Subject: Re: [mm] f35b5d7d67: will-it-scale.per_process_ops -95.5% regression #forregzbot

On 29.11.22 09:59, Thorsten Leemhuis wrote:
> [Note: this mail is primarily send for documentation purposes and/or for
> regzbot, my Linux kernel regression tracking bot. That's why I removed
> most or all folks from the list of recipients, but left any that looked
> like a mailing lists. These mails usually contain '#forregzbot' in the
> subject, to make them easy to spot and filter out.]
>
> On 20.10.22 06:23, Nathan Chancellor wrote:
>> [...]
>> For what it's worth, I just bisected a massive and visible performance
>> regression on my Threadripper 3990X workstation to commit f35b5d7d676e
>> ("mm: align larger anonymous mappings on THP boundaries"), which seems
>> directly related to this report/analysis. I initially noticed this
>> because my full set of kernel builds against mainline went from 2 hours
>> and 20 minutes or so to over 3 hours. Zeroing in on x86_64 allmodconfig,
>> which I used for the bisect:
>
> Thanks for the report. To be sure below issue doesn't fall through the
> cracks unnoticed, I'm adding it to regzbot, my Linux kernel regression
> tracking bot:
>
> #regzbot ^introduced f35b5d7d676e59e4
> #regzbot title performance: mm: building Linux with clang takes a lot
> longer (~27 %)
> #regzbot ignore-activity

#regzbot fix: revert "mm: align larger anonymous mappings on THP boundaries"

Ciao, Thorsten (wearing his 'the Linux kernel's regression tracker' hat)

P.S.: As the Linux kernel's regression tracker I deal with a lot of
reports and sometimes miss something important when writing mails like
this. If that's the case here, don't hesitate to tell me in a public
reply, it's in everyone's interest to set the public record straight.

2022-12-02 09:22:12

by Thorsten Leemhuis

[permalink] [raw]
Subject: Re: [mm] f35b5d7d67: will-it-scale.per_process_ops -95.5% regression

On 01.12.22 22:22, Andrew Morton wrote:
> On Thu, 01 Dec 2022 15:29:41 -0500 Rik van Riel <[email protected]> wrote:
>> On Thu, 2022-12-01 at 19:33 +0100, Thorsten Leemhuis wrote:
>>> On 28.11.22 07:40, Nathan Chancellor wrote:
>>> I wonder what we should do about below performance regression. Is
>>> reverting the culprit now and reapplying it later together with a fix
>>> a
>>> viable option? Or was anything done/is anybody doing something
>>> already
>>> to address the problem and I just missed it?
>>
>> The changeset in question speeds up kernel compiles with
>> GCC, as well as the runtime speed of other programs, due
>> to being able to use THPs more. However, it slows down kernel
>> compiles with clang, due to ... something clang does.
>>
>> I have not figured out what that something is yet.
>>
>> I don't know if I have the wrong version of clang here,
>> but I have not seen any smoking gun at all when tracing
>> clang system calls. I see predominantly small mmap and
>> unmap calls, and nothing that even triggers 2MB alignment.
>
> 2.8% speedup for gcc is nice. Massive slowdown in the malloc banchmark
> and in LLVM/clang is very bad - we don't know what other userspace will
> be so affected.
>
> So I think we revert until this is fully understood.

Andrew, many thx for taking care of this. While at it let me please get
a small process issue of my chest:

What beverage of choice do I have to offer you to make you in the future
include 'Link:' tags linking to the report(s) when you add reverts like
the one for this issue?

My regression tracking bot heavily relies on them, that's why I care.
And our documentation (see [1]) also says that they should be used. But
the main reason why I ask in this particular case is different:

They in cases like afaics are especially helpful, as they make life a
whole lot easier for future code archeologists. And I guess that's not
something theoretical in this case, as I assume the patch that triggered
the issue will come back sooner or later -- and then those links will
help a lot to find this thread. Which is also why Linus really wants to
see them:
https://lore.kernel.org/all/CAHk-=wjMmSZzMJ3Xnskdg4+GGz=5p5p+GSYyFBTh0f-DgvdBWg@mail.gmail.com/
https://lore.kernel.org/all/CAHk-=wgs38ZrfPvy=nOwVkVzjpM3VFU1zobP37Fwd_h9iAD5JQ@mail.gmail.com/
https://lore.kernel.org/all/CAHk-=wjxzafG-=J8oT30s7upn4RhBs6TX-uVFZ5rME+L5_DoJA@mail.gmail.com/

Ciao, Thorsten

[1] see Documentation/process/submitting-patches.rst
(http://docs.kernel.org/process/submitting-patches.html) and
Documentation/process/5.Posting.rst
(https://docs.kernel.org/process/5.Posting.html)

2022-12-02 19:51:13

by Thorsten Leemhuis

[permalink] [raw]
Subject: Re: [mm] f35b5d7d67: will-it-scale.per_process_ops -95.5% regression

On 02.12.22 19:44, Andrew Morton wrote:
> On Fri, 2 Dec 2022 09:46:48 +0100 Thorsten Leemhuis <[email protected]> wrote:
>
>> What beverage of choice do I have to offer you to make you in the future
>> include 'Link:' tags linking to the report(s) when you add reverts like
>> the one for this issue?
>
> $10 supermarket pinot seems to be my tipple.

/me will make a note

> I can do that.

Great, many thx!

> Any Reported-by: requires a Link:, OK?

Yup, that's how it is, except for a few corner cases (say someone
reported something on IRC or on a hallway).

Ciao, Thorsten

2022-12-02 19:55:07

by Andrew Morton

[permalink] [raw]
Subject: Re: [mm] f35b5d7d67: will-it-scale.per_process_ops -95.5% regression

On Fri, 2 Dec 2022 09:46:48 +0100 Thorsten Leemhuis <[email protected]> wrote:

> What beverage of choice do I have to offer you to make you in the future
> include 'Link:' tags linking to the report(s) when you add reverts like
> the one for this issue?

$10 supermarket pinot seems to be my tipple.

I can do that. Any Reported-by: requires a Link:, OK?

2022-12-16 12:03:15

by Yin, Fengwei

[permalink] [raw]
Subject: Re: [mm] f35b5d7d67: will-it-scale.per_process_ops -95.5% regression



On 12/2/2022 4:29 AM, Rik van Riel wrote:
> On Thu, 2022-12-01 at 19:33 +0100, Thorsten Leemhuis wrote:
>> Hi, this is your Linux kernel regression tracker.
>>
>> On 28.11.22 07:40, Nathan Chancellor wrote:
>>> Hi Rik,
>>
>> I wonder what we should do about below performance regression. Is
>> reverting the culprit now and reapplying it later together with a fix
>> a
>> viable option? Or was anything done/is anybody doing something
>> already
>> to address the problem and I just missed it?
>
> The changeset in question speeds up kernel compiles with
> GCC, as well as the runtime speed of other programs, due
> to being able to use THPs more. However, it slows down kernel
> compiles with clang, due to ... something clang does.
>
> I have not figured out what that something is yet.
>
> I don't know if I have the wrong version of clang here,
> but I have not seen any smoking gun at all when tracing
> clang system calls. I see predominantly small mmap and
> unmap calls, and nothing that even triggers 2MB alignment.

There are some findings in my side. I not very sure that it's root
case and like to share here to call more eyes on it.
1. The regression is not related to clang. It's related with
ld.lld.
In my env, command
make LD=ld.lld -skj96 allmodconfig all
could reproduce the regression. But command
make CC=clang -skj96 allmodconfig all
can't reproduce the regression.

2. It looks like ld.lld calls madvise(MADV_DONTNEED) many times to
free memory. But the parameters "len" is not friendly to THP.

trace-bpfcc with do_madvise gave:
531607 531732 ld.lld do_madvise.part.0 start: 0x7feca9000000, len: 0x7fb000, behavior: 0x4
531607 531793 ld.lld do_madvise.part.0 start: 0x7fec86a00000, len: 0x7fb000, behavior: 0x4
531607 531767 ld.lld do_madvise.part.0 start: 0x7fec95000000, len: 0x7fb000, behavior: 0x4
531607 531703 ld.lld do_madvise.part.0 start: 0x7fecb3000000, len: 0x7fb000, behavior: 0x4
531607 531757 ld.lld do_madvise.part.0 start: 0x7fec9b400000, len: 0x7fb000, behavior: 0x4
531607 531701 ld.lld do_madvise.part.0 start: 0x7fecb4400000, len: 0x7fb000, behavior: 0x4
531607 531795 ld.lld do_madvise.part.0 start: 0x7fec85600000, len: 0x7fb000, behavior: 0x4
531607 531792 ld.lld do_madvise.part.0 start: 0x7fec87400000, len: 0x7fb000, behavior: 0x4
531607 531695 ld.lld do_madvise.part.0 start: 0x7fecbca00000, len: 0x7fb000, behavior: 0x4
531607 531760 ld.lld do_madvise.part.0 start: 0x7fec99600000, len: 0x7fb000, behavior: 0x4
531607 531706 ld.lld do_madvise.part.0 start: 0x7fecb1200000, len: 0x7fb000, behavior: 0x4
531607 531769 ld.lld do_madvise.part.0 start: 0x7fec94600000, len: 0x7fb000, behavior: 0x4
531607 531699 ld.lld do_madvise.part.0 start: 0x7fecb5800000, len: 0x7fb000, behavior: 0x4
531607 531738 ld.lld do_madvise.part.0 start: 0x7feca5e00000, len: 0x7fb000, behavior: 0x4

3. I captured the perf data for 50s when kernel build with ld.lld
was ongoing and got:

24.20% 0.27% cc1 [kernel.kallsyms] [k] asm_exc_page_fault
9.52%
asm_exc_page_fault
exc_page_fault
do_user_addr_fault
handle_mm_fault
__handle_mm_fault
do_huge_pmd_anonymous_page
vma_alloc_folio
__folio_alloc
__alloc_pages
get_page_from_freelist
clear_page_erms

14.85% 0.00% ld.lld [kernel.kallsyms] [k] entry_SYSCALL_64_after_hwframe
11.52%
entry_SYSCALL_64_after_hwframe
do_syscall_64
__x64_sys_madvise
do_madvise.part.0
zap_page_range
unmap_single_vma
unmap_page_range
page_remove_rmap
deferred_split_huge_page
__lock_text_start
native_queued_spin_lock_slowpath

My understanding is that the deferred_split_huge_page() here is kind of
matching len: 0x7fb000 in do_madvise() call. The end is not PMD size
aligned and trigger lock contention on deferred_split_queue.

With GNU ld, there is no any madvise(MADV_DONTNEED) called.


Regards
Yin, Fengwei

>
>>
>> Yang Shi, Andrew, what's your option on this? I ask you directly,
>> because it looks like Rik hasn't posted anything to lists archived on
>> lore during the last few weeks. :-/
>>
>> Ciao, Thorsten (wearing his 'the Linux kernel's regression tracker'
>> hat)
>>
>> P.S.: As the Linux kernel's regression tracker I deal with a lot of
>> reports and sometimes miss something important when writing mails
>> like
>> this. If that's the case here, don't hesitate to tell me in a public
>> reply, it's in everyone's interest to set the public record straight.
>>
>>> On Thu, Oct 20, 2022 at 10:16:20AM -0700, Nathan Chancellor wrote:
>>>> On Thu, Oct 20, 2022 at 11:28:16AM -0400, Rik van Riel wrote:
>>>>> On Thu, 2022-10-20 at 13:07 +0800, Huang, Ying wrote:
>>>>>> Nathan Chancellor <[email protected]> writes:
>>>>>>>
>>>>>>> For what it's worth, I just bisected a massive and visible
>>>>>>> performance
>>>>>>> regression on my Threadripper 3990X workstation to commit
>>>>>>> f35b5d7d676e
>>>>>>> ("mm: align larger anonymous mappings on THP boundaries"),
>>>>>>> which
>>>>>>> seems
>>>>>>> directly related to this report/analysis. I initially
>>>>>>> noticed this
>>>>>>> because my full set of kernel builds against mainline went
>>>>>>> from 2
>>>>>>> hours
>>>>>>> and 20 minutes or so to over 3 hours. Zeroing in on x86_64
>>>>>>> allmodconfig,
>>>>>>> which I used for the bisect:
>>>>>>>
>>>>>>> @ 7b5a0b664ebe ("mm/page_ext: remove unused variable in
>>>>>>> offline_page_ext"):
>>>>>>>
>>>>>>> Benchmark 1: make -skj128 LLVM=1 allmodconfig all
>>>>>>>   Time (mean ± σ):     318.172 s ±  0.730 s    [User:
>>>>>>> 31750.902 s,
>>>>>>> System: 4564.246 s]
>>>>>>>   Range (min … max):   317.332 s … 318.662 s    3 runs
>>>>>>>
>>>>>>> @ f35b5d7d676e ("mm: align larger anonymous mappings on THP
>>>>>>> boundaries"):
>>>>>>>
>>>>>>> Benchmark 1: make -skj128 LLVM=1 allmodconfig all
>>>>>>>   Time (mean ± σ):     406.688 s ±  0.676 s    [User:
>>>>>>> 31819.526 s,
>>>>> System: 16327.022 s]
>>>>>>>   Range (min … max):   405.954 s … 407.284 s    3 run
>>>>>>
>>>>>> Have you tried to build with gcc?  Want to check whether is
>>>>>> this
>>>>>> clang
>>>>>> specific issue or not.
>>>>>
>>>>> This may indeed be something LLVM specific. In previous tests,
>>>>> GCC has generally seen a benefit from increased THP usage.
>>>>> Many other applications also benefit from getting more THPs.
>>>>
>>>> Indeed, GCC builds actually appear to be slightly faster on my
>>>> system now,
>>>> apologies for not trying that before reporting :/
>>>>
>>>> 7b5a0b664ebe:
>>>>
>>>> Benchmark 1: make -skj128 allmodconfig all
>>>>   Time (mean ± σ):     355.294 s ±  0.931 s    [User: 33620.469
>>>> s, System: 6390.064 s]
>>>>   Range (min … max):   354.571 s … 356.344 s    3 runs
>>>>
>>>> f35b5d7d676e:
>>>>
>>>> Benchmark 1: make -skj128 allmodconfig all
>>>>   Time (mean ± σ):     347.400 s ±  2.029 s    [User: 34389.724
>>>> s, System: 4603.175 s]
>>>>   Range (min … max):   345.815 s … 349.686 s    3 runs
>>>>
>>>>> LLVM showing 10% system time before this change, and a whopping
>>>>> 30% system time after that change, suggests that LLVM is
>>>>> behaving
>>>>> quite differently from GCC in some ways.
>>>>
>>>> The above tests were done with GCC 12.2.0 from Arch Linux. The
>>>> previous LLVM
>>>> tests were done with a self-compiled version of LLVM from the
>>>> main branch
>>>> (16.0.0), optimized with BOLT [1]. To eliminate that as a source
>>>> of issues, I
>>>> used my distribution's version of clang (14.0.6) and saw similar
>>>> results as
>>>> before:
>>>>
>>>> 7b5a0b664ebe:
>>>>
>>>> Benchmark 1: make -skj128 LLVM=/usr/bin/ allmodconfig all
>>>>   Time (mean ± σ):     462.517 s ±  1.214 s    [User: 48544.240
>>>> s, System: 5586.212 s]
>>>>   Range (min … max):   461.115 s … 463.245 s    3 runs
>>>>
>>>> f35b5d7d676e:
>>>>
>>>> Benchmark 1: make -skj128 LLVM=/usr/bin/ allmodconfig all
>>>>   Time (mean ± σ):     547.927 s ±  0.862 s    [User: 47913.709
>>>> s, System: 17682.514 s]
>>>>   Range (min … max):   547.429 s … 548.922 s    3 runs
>>>>
>>>>> If we can figure out what these differences are, maybe we can
>>>>> just fine tune the code to avoid this issue.
>>>>>
>>>>> I'll try to play around with LLVM compilation a little bit next
>>>>> week, to see if I can figure out what might be going on. I
>>>>> wonder
>>>>> if LLVM is doing lots of mremap calls or something...
>>>>
>>>> If there is any further information I can provide or patches I
>>>> can test,
>>>> I am more than happy to do so.
>>>>
>>>> [1]:
>>>> https://github.com/llvm/llvm-project/tree/96552e73900176d65ee6650facae8d669d6f9498/bolt
>>>
>>> Was there ever a follow up to this report that I missed? I just
>>> noticed that I am still reverting f35b5d7d676e in my mainline
>>> kernel.
>>>
>>> Cheers,
>>> Nathan
>>>
>>
>> #regzbot ignore-activity
>>
>