2023-11-15 14:02:47

by kernel test robot

[permalink] [raw]
Subject: [linus:master] [mm] 12214eba19: vm-scalability.throughput 49.5% improvement



Hello,

kernel test robot noticed a 49.5% improvement of vm-scalability.throughput on:


commit: 12214eba1992642eee5813a9cc9f626e5b2d1815 ("mm: handle read faults under the VMA lock")
https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master

testcase: vm-scalability
test machine: 224 threads 4 sockets Intel(R) Xeon(R) Platinum 8380H CPU @ 2.90GHz (Cooper Lake) with 192G memory
parameters:

runtime: 300s
size: 2T
test: shm-pread-seq
cpufreq_governor: performance

test-description: The motivation behind this suite is to exercise functions and regions of the mm/ of the Linux kernel which are of interest to us.
test-url: https://git.kernel.org/cgit/linux/kernel/git/wfg/vm-scalability.git/





Details are as below:
-------------------------------------------------------------------------------------------------->


The kernel config and materials to reproduce are available at:
https://download.01.org/0day-ci/archive/20231115/[email protected]

=========================================================================================
compiler/cpufreq_governor/kconfig/rootfs/runtime/size/tbox_group/test/testcase:
gcc-12/performance/x86_64-rhel-8.3/debian-11.1-x86_64-20220510.cgz/300s/2T/lkp-cpl-4sp2/shm-pread-seq/vm-scalability

commit:
4de8c93a47 ("mm: handle COW faults under the VMA lock")
12214eba19 ("mm: handle read faults under the VMA lock")

4de8c93a4751e107 12214eba1992642eee5813a9cc9
---------------- ---------------------------
%stddev %change %stddev
\ | \
25.26 +80.3% 45.55 ? 4% vm-scalability.free_time
166309 +49.5% 248702 ? 2% vm-scalability.median
0.06 ? 17% -0.0 0.04 ? 28% vm-scalability.stddev%
37253200 +49.5% 55709403 ? 2% vm-scalability.throughput
329.93 +6.8% 352.34 vm-scalability.time.elapsed_time
329.93 +6.8% 352.34 vm-scalability.time.elapsed_time.max
204399 +15.0% 235056 ? 3% vm-scalability.time.involuntary_context_switches
37504320 +66.1% 62277600 ? 3% vm-scalability.time.maximum_resident_set_size
7.573e+08 +64.6% 1.247e+09 vm-scalability.time.minor_page_faults
60103 +4.7% 62934 vm-scalability.time.system_time
2677 +65.6% 4434 ? 3% vm-scalability.time.user_time
2.907e+08 +41.7% 4.119e+08 ? 2% vm-scalability.time.voluntary_context_switches
1.119e+10 +49.6% 1.674e+10 ? 2% vm-scalability.workload
2.561e+08 +42.2% 3.642e+08 ? 7% cpuidle..usage
1.905e+08 +17.7% 2.241e+08 ? 14% turbostat.C1
0.05 ? 9% +34.2% 0.06 ? 7% turbostat.IPC
3.103e+08 +21.1% 3.758e+08 ? 8% turbostat.IRQ
65396804 +113.9% 1.399e+08 ? 3% turbostat.POLL
4592731 +77.5% 8150136 ? 7% meminfo.Active
4592558 +77.5% 8149968 ? 7% meminfo.Active(anon)
199212 +16.2% 231560 ? 4% meminfo.KReclaimable
199212 +16.2% 231560 ? 4% meminfo.SReclaimable
38088314 +66.4% 63396415 ? 14% meminfo.max_used_kB
656.00 ? 4% +14.8% 753.38 perf-c2c.DRAM.local
19685 ? 4% +33.2% 26225 ? 2% perf-c2c.DRAM.remote
11307 ? 4% +24.5% 14073 ? 2% perf-c2c.HITM.local
10620 ? 4% +37.2% 14571 ? 2% perf-c2c.HITM.remote
21927 ? 4% +30.6% 28644 ? 2% perf-c2c.HITM.total
1274254 +51.6% 1931316 ? 16% sched_debug.cpu.nr_switches.avg
1330622 +52.2% 2024967 ? 16% sched_debug.cpu.nr_switches.max
1169737 +49.4% 1747634 ? 17% sched_debug.cpu.nr_switches.min
23313 ? 3% +56.5% 36484 ? 15% sched_debug.cpu.nr_switches.stddev
-601.85 +106.5% -1243 sched_debug.cpu.nr_uninterruptible.min
190.78 ? 10% +52.0% 290.04 ? 20% sched_debug.cpu.nr_uninterruptible.stddev
4361232 ? 3% +61.0% 7021897 ? 6% numa-numastat.node0.local_node
4423736 ? 3% +61.0% 7120683 ? 6% numa-numastat.node0.numa_hit
62510 ? 63% +58.0% 98777 ? 22% numa-numastat.node0.other_node
4186176 ? 3% +58.7% 6642324 ? 3% numa-numastat.node1.local_node
4275058 ? 3% +57.2% 6720332 ? 3% numa-numastat.node1.numa_hit
4281988 ? 2% +55.5% 6658287 ? 5% numa-numastat.node2.local_node
4377154 ? 2% +53.9% 6736235 ? 5% numa-numastat.node2.numa_hit
4970732 ? 2% +57.9% 7846573 ? 6% numa-numastat.node3.local_node
5072257 ? 2% +56.5% 7939683 ? 6% numa-numastat.node3.numa_hit
589678 ? 48% +138.8% 1408222 ? 29% numa-meminfo.node0.Active
589604 ? 48% +138.8% 1408190 ? 29% numa-meminfo.node0.Active(anon)
6000866 ? 20% +62.5% 9750086 ? 11% numa-meminfo.node0.FilePages
39362890 ? 3% -13.1% 34215245 ? 4% numa-meminfo.node0.MemFree
8676149 ? 15% +59.3% 13823794 ? 12% numa-meminfo.node0.MemUsed
5279360 ? 6% +69.0% 8923256 ? 17% numa-meminfo.node0.Shmem
2542157 ? 6% +64.7% 4186028 ? 10% numa-meminfo.node3.Active
2542137 ? 6% +64.7% 4185981 ? 10% numa-meminfo.node3.Active(anon)
148212 ? 27% +62.1% 240250 ? 13% numa-meminfo.node3.AnonPages.max
7275379 ? 4% +65.1% 12011702 ? 17% numa-meminfo.node3.FilePages
7134785 ? 3% +58.2% 11290575 ? 13% numa-meminfo.node3.Shmem
1147703 +77.4% 2036461 ? 7% proc-vmstat.nr_active_anon
49805 +16.2% 57887 ? 4% proc-vmstat.nr_slab_reclaimable
1147703 +77.4% 2036461 ? 7% proc-vmstat.nr_zone_active_anon
18151336 +57.1% 28520631 ? 3% proc-vmstat.numa_hit
17803254 +58.2% 28172774 ? 3% proc-vmstat.numa_local
9636459 +65.9% 15986667 ? 2% proc-vmstat.pgactivate
18367048 +56.5% 28743362 ? 3% proc-vmstat.pgalloc_normal
7.591e+08 +64.5% 1.249e+09 proc-vmstat.pgfault
17245013 +58.6% 27346493 ? 3% proc-vmstat.pgfree
178502 ? 2% +13.2% 202009 ? 5% proc-vmstat.pgreuse
3501024 ? 5% +20.8% 4230646 ? 28% proc-vmstat.unevictable_pgs_scanned
0.01 -28.6% 0.01 perf-sched.sch_delay.avg.ms.io_schedule.folio_wait_bit_common.shmem_get_folio_gfp.shmem_fault
0.01 ? 9% -19.8% 0.01 ? 7% perf-sched.sch_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
0.08 ? 7% -26.2% 0.06 ? 5% perf-sched.sch_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
78.73 ? 6% -20.7% 62.43 ? 7% perf-sched.sch_delay.max.ms.io_schedule.folio_wait_bit_common.shmem_get_folio_gfp.shmem_fault
0.01 -26.8% 0.01 ? 6% perf-sched.total_sch_delay.average.ms
0.65 -26.2% 0.48 ? 3% perf-sched.total_wait_and_delay.average.ms
3413581 +36.5% 4660142 ? 2% perf-sched.total_wait_and_delay.count.ms
0.64 -26.2% 0.47 ? 2% perf-sched.total_wait_time.average.ms
0.03 -29.6% 0.02 perf-sched.wait_and_delay.avg.ms.io_schedule.folio_wait_bit_common.shmem_get_folio_gfp.shmem_fault
3399862 +36.6% 4645123 ? 2% perf-sched.wait_and_delay.count.io_schedule.folio_wait_bit_common.shmem_get_folio_gfp.shmem_fault
0.16 ?198% -82.6% 0.03 ? 6% perf-sched.wait_time.avg.ms.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate.isra
0.02 -30.5% 0.02 perf-sched.wait_time.avg.ms.io_schedule.folio_wait_bit_common.shmem_get_folio_gfp.shmem_fault
147244 ? 48% +139.0% 351869 ? 29% numa-vmstat.node0.nr_active_anon
1499808 ? 20% +62.6% 2438217 ? 11% numa-vmstat.node0.nr_file_pages
9840486 ? 3% -13.1% 8551977 ? 4% numa-vmstat.node0.nr_free_pages
1319431 ? 6% +69.1% 2231510 ? 17% numa-vmstat.node0.nr_shmem
147241 ? 48% +139.0% 351866 ? 29% numa-vmstat.node0.nr_zone_active_anon
4423747 ? 3% +61.0% 7120789 ? 6% numa-vmstat.node0.numa_hit
4361243 ? 3% +61.0% 7022004 ? 6% numa-vmstat.node0.numa_local
62510 ? 63% +58.0% 98776 ? 22% numa-vmstat.node0.numa_other
4275031 ? 3% +57.2% 6720401 ? 3% numa-vmstat.node1.numa_hit
4186150 ? 3% +58.7% 6642392 ? 3% numa-vmstat.node1.numa_local
4377220 ? 2% +53.9% 6736168 ? 5% numa-vmstat.node2.numa_hit
4282054 ? 2% +55.5% 6658220 ? 5% numa-vmstat.node2.numa_local
635357 ? 6% +64.6% 1046089 ? 10% numa-vmstat.node3.nr_active_anon
1818439 ? 4% +65.2% 3003397 ? 17% numa-vmstat.node3.nr_file_pages
1783291 ? 3% +58.3% 2823116 ? 13% numa-vmstat.node3.nr_shmem
635354 ? 6% +64.6% 1046084 ? 10% numa-vmstat.node3.nr_zone_active_anon
5072172 ? 2% +56.5% 7939769 ? 6% numa-vmstat.node3.numa_hit
4970647 ? 2% +57.9% 7846659 ? 6% numa-vmstat.node3.numa_local
42.17 -3.8 38.39 ? 16% perf-stat.i.cache-miss-rate%
7.60 -4.8% 7.24 ? 5% perf-stat.i.cpi
49.99 +4.2 54.23 perf-stat.i.iTLB-load-miss-rate%
0.16 +40.4% 0.23 ? 2% perf-stat.i.ipc
0.09 ? 35% -53.9% 0.04 ? 59% perf-stat.i.major-faults
904.06 +13.2% 1023 ? 2% perf-stat.i.metric.K/sec
4.24 -22.9% 3.27 ? 2% perf-stat.overall.MPKI
0.38 -0.1 0.28 ? 2% perf-stat.overall.branch-miss-rate%
43.12 -2.6 40.52 perf-stat.overall.cache-miss-rate%
6.30 -29.9% 4.42 perf-stat.overall.cpi
1485 -9.1% 1350 perf-stat.overall.cycles-between-cache-misses
0.05 ? 2% -0.0 0.05 ? 2% perf-stat.overall.dTLB-load-miss-rate%
0.04 +0.0 0.04 perf-stat.overall.dTLB-store-miss-rate%
47.06 +2.5 49.59 perf-stat.overall.iTLB-load-miss-rate%
10804 +21.8% 13154 ? 3% perf-stat.overall.instructions-per-iTLB-miss
0.16 +42.7% 0.23 perf-stat.overall.ipc
3579 +3.8% 3716 perf-stat.overall.path-length
0.09 ? 34% -52.5% 0.04 ? 59% perf-stat.ps.major-faults
4.007e+13 +55.3% 6.222e+13 ? 2% perf-stat.total.instructions
63.05 -10.1 52.95 perf-profile.calltrace.cycles-pp.filemap_map_pages.do_read_fault.do_fault.__handle_mm_fault.handle_mm_fault
36.62 -9.6 27.00 perf-profile.calltrace.cycles-pp.next_uptodate_folio.filemap_map_pages.do_read_fault.do_fault.__handle_mm_fault
82.71 -3.4 79.30 perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
83.18 -3.3 79.84 perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.do_access
83.58 -3.3 80.25 perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.do_access
83.60 -3.3 80.28 perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.do_access
82.20 -3.3 78.88 perf-profile.calltrace.cycles-pp.do_read_fault.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
82.21 -3.3 78.89 perf-profile.calltrace.cycles-pp.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
83.97 -3.1 80.83 perf-profile.calltrace.cycles-pp.asm_exc_page_fault.do_access
15.72 -2.1 13.60 ? 5% perf-profile.calltrace.cycles-pp.asm_sysvec_call_function_single.acpi_safe_halt.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter
85.23 -1.6 83.61 perf-profile.calltrace.cycles-pp.do_access
8.88 -1.2 7.64 ? 4% perf-profile.calltrace.cycles-pp.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
0.99 +0.1 1.08 ? 3% perf-profile.calltrace.cycles-pp.folio_add_file_rmap_range.set_pte_range.filemap_map_pages.do_read_fault.do_fault
1.21 +0.1 1.32 ? 3% perf-profile.calltrace.cycles-pp.set_pte_range.filemap_map_pages.do_read_fault.do_fault.__handle_mm_fault
1.37 ? 3% +0.2 1.54 ? 3% perf-profile.calltrace.cycles-pp.schedule.io_schedule.folio_wait_bit_common.shmem_get_folio_gfp.shmem_fault
1.38 ? 3% +0.2 1.55 ? 3% perf-profile.calltrace.cycles-pp.io_schedule.folio_wait_bit_common.shmem_get_folio_gfp.shmem_fault.__do_fault
0.51 +0.3 0.78 perf-profile.calltrace.cycles-pp.finish_fault.do_read_fault.do_fault.__handle_mm_fault.handle_mm_fault
14.54 +0.5 15.00 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_wake_bit.filemap_map_pages.do_read_fault
14.96 +0.5 15.47 perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.folio_wake_bit.filemap_map_pages.do_read_fault.do_fault
0.00 +0.6 0.55 perf-profile.calltrace.cycles-pp.set_pte_range.finish_fault.do_read_fault.do_fault.__handle_mm_fault
16.57 +0.6 17.17 perf-profile.calltrace.cycles-pp.folio_wake_bit.filemap_map_pages.do_read_fault.do_fault.__handle_mm_fault
1.53 +0.7 2.25 ? 2% perf-profile.calltrace.cycles-pp.filemap_get_entry.shmem_get_folio_gfp.shmem_fault.__do_fault.do_read_fault
2.08 +1.4 3.50 ? 4% perf-profile.calltrace.cycles-pp.poll_idle.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
3.37 +2.2 5.59 ? 3% perf-profile.calltrace.cycles-pp.do_rw_once
6.18 +2.3 8.43 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irq.folio_wait_bit_common.shmem_get_folio_gfp.shmem_fault
6.44 +2.4 8.79 perf-profile.calltrace.cycles-pp._raw_spin_lock_irq.folio_wait_bit_common.shmem_get_folio_gfp.shmem_fault.__do_fault
5.34 +2.9 8.24 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_wake_bit.do_read_fault.do_fault
9.41 +2.9 12.33 perf-profile.calltrace.cycles-pp.folio_wait_bit_common.shmem_get_folio_gfp.shmem_fault.__do_fault.do_read_fault
5.46 +3.0 8.45 perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.folio_wake_bit.do_read_fault.do_fault.__handle_mm_fault
5.95 +3.3 9.21 perf-profile.calltrace.cycles-pp.folio_wake_bit.do_read_fault.do_fault.__handle_mm_fault.handle_mm_fault
11.51 +3.9 15.42 perf-profile.calltrace.cycles-pp.shmem_get_folio_gfp.shmem_fault.__do_fault.do_read_fault.do_fault
11.67 +3.9 15.62 perf-profile.calltrace.cycles-pp.shmem_fault.__do_fault.do_read_fault.do_fault.__handle_mm_fault
11.69 +4.0 15.64 perf-profile.calltrace.cycles-pp.__do_fault.do_read_fault.do_fault.__handle_mm_fault.handle_mm_fault
63.06 -10.1 52.95 perf-profile.children.cycles-pp.filemap_map_pages
36.64 -9.6 27.01 perf-profile.children.cycles-pp.next_uptodate_folio
82.71 -3.4 79.30 perf-profile.children.cycles-pp.__handle_mm_fault
83.19 -3.3 79.85 perf-profile.children.cycles-pp.handle_mm_fault
83.59 -3.3 80.26 perf-profile.children.cycles-pp.do_user_addr_fault
83.60 -3.3 80.28 perf-profile.children.cycles-pp.exc_page_fault
82.20 -3.3 78.89 perf-profile.children.cycles-pp.do_read_fault
82.21 -3.3 78.90 perf-profile.children.cycles-pp.do_fault
83.98 -3.1 80.85 perf-profile.children.cycles-pp.asm_exc_page_fault
86.13 -1.8 84.34 perf-profile.children.cycles-pp.do_access
8.91 -1.2 7.67 ? 4% perf-profile.children.cycles-pp.acpi_safe_halt
8.91 -1.2 7.67 ? 4% perf-profile.children.cycles-pp.acpi_idle_enter
8.28 -1.0 7.28 ? 4% perf-profile.children.cycles-pp.asm_sysvec_call_function_single
0.91 ? 6% -0.8 0.14 ? 4% perf-profile.children.cycles-pp.up_read
0.97 -0.2 0.81 perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
0.12 ? 3% -0.0 0.08 ? 4% perf-profile.children.cycles-pp.down_read_trylock
0.30 ? 2% -0.0 0.27 ? 3% perf-profile.children.cycles-pp.__pte_offset_map_lock
0.25 ? 2% -0.0 0.23 ? 5% perf-profile.children.cycles-pp.tick_sched_timer
0.23 ? 2% -0.0 0.21 ? 4% perf-profile.children.cycles-pp.tick_sched_handle
0.22 ? 2% -0.0 0.21 ? 4% perf-profile.children.cycles-pp.update_process_times
0.19 ? 2% -0.0 0.17 ? 3% perf-profile.children.cycles-pp.scheduler_tick
0.06 ? 5% -0.0 0.05 perf-profile.children.cycles-pp.tick_nohz_get_sleep_length
0.06 +0.0 0.07 perf-profile.children.cycles-pp.exit_to_user_mode_prepare
0.06 +0.0 0.07 ? 4% perf-profile.children.cycles-pp.update_blocked_averages
0.06 ? 5% +0.0 0.08 ? 5% perf-profile.children.cycles-pp.__switch_to
0.05 +0.0 0.07 ? 7% perf-profile.children.cycles-pp.__switch_to_asm
0.05 ? 6% +0.0 0.07 ? 4% perf-profile.children.cycles-pp.prepare_task_switch
0.09 ? 3% +0.0 0.11 ? 4% perf-profile.children.cycles-pp.llist_add_batch
0.06 ? 5% +0.0 0.08 ? 6% perf-profile.children.cycles-pp.llist_reverse_order
0.15 ? 2% +0.0 0.17 ? 2% perf-profile.children.cycles-pp.update_load_avg
0.06 +0.0 0.08 ? 5% perf-profile.children.cycles-pp.available_idle_cpu
0.05 ? 8% +0.0 0.08 ? 6% perf-profile.children.cycles-pp.wake_affine
0.08 +0.0 0.11 ? 4% perf-profile.children.cycles-pp.update_rq_clock_task
0.08 ? 4% +0.0 0.11 ? 5% perf-profile.children.cycles-pp.__list_del_entry_valid_or_report
0.03 ? 77% +0.0 0.06 ? 5% perf-profile.children.cycles-pp.xas_descend
0.09 ? 4% +0.0 0.12 ? 2% perf-profile.children.cycles-pp.select_task_rq_fair
0.09 ? 5% +0.0 0.13 ? 3% perf-profile.children.cycles-pp.select_task_rq
0.04 ? 37% +0.0 0.08 ? 4% perf-profile.children.cycles-pp.mas_walk
0.14 ? 3% +0.0 0.18 ? 3% perf-profile.children.cycles-pp.finish_task_switch
0.11 ? 4% +0.0 0.16 ? 3% perf-profile.children.cycles-pp.__smp_call_single_queue
0.02 ?129% +0.0 0.06 ? 6% perf-profile.children.cycles-pp.raw_spin_rq_lock_nested
0.18 +0.0 0.22 ? 3% perf-profile.children.cycles-pp.dequeue_entity
0.12 ? 4% +0.0 0.17 ? 3% perf-profile.children.cycles-pp.lock_vma_under_rcu
0.20 ? 2% +0.0 0.25 ? 2% perf-profile.children.cycles-pp.dequeue_task_fair
0.08 ? 5% +0.0 0.13 ? 6% perf-profile.children.cycles-pp.shmem_alloc_and_add_folio
0.16 +0.0 0.21 ? 3% perf-profile.children.cycles-pp.enqueue_entity
0.08 ? 5% +0.1 0.13 ? 3% perf-profile.children.cycles-pp.sync_regs
0.00 +0.1 0.05 perf-profile.children.cycles-pp.native_sched_clock
0.00 +0.1 0.05 perf-profile.children.cycles-pp.set_next_entity
0.51 +0.1 0.57 ? 2% perf-profile.children.cycles-pp.__sysvec_call_function_single
0.22 ? 2% +0.1 0.28 ? 4% perf-profile.children.cycles-pp.ttwu_queue_wakelist
0.26 +0.1 0.34 ? 4% perf-profile.children.cycles-pp._compound_head
0.60 +0.1 0.68 ? 2% perf-profile.children.cycles-pp.sysvec_call_function_single
0.48 ? 5% +0.1 0.56 ? 5% perf-profile.children.cycles-pp.___perf_sw_event
0.22 ? 2% +0.1 0.30 perf-profile.children.cycles-pp.enqueue_task_fair
0.22 ? 2% +0.1 0.31 perf-profile.children.cycles-pp.activate_task
0.14 ? 3% +0.1 0.23 ? 2% perf-profile.children.cycles-pp.folio_unlock
0.30 ? 2% +0.1 0.39 perf-profile.children.cycles-pp.schedule_idle
0.30 +0.1 0.40 perf-profile.children.cycles-pp.ttwu_do_activate
0.13 ? 3% +0.1 0.26 ? 2% perf-profile.children.cycles-pp.flush_smp_call_function_queue
0.45 +0.1 0.59 ? 3% perf-profile.children.cycles-pp.try_to_wake_up
0.44 +0.1 0.58 ? 2% perf-profile.children.cycles-pp.sched_ttwu_pending
0.30 ? 2% +0.1 0.44 perf-profile.children.cycles-pp.native_irq_return_iret
1.16 +0.1 1.31 ? 2% perf-profile.children.cycles-pp.wake_page_function
1.38 ? 3% +0.2 1.54 ? 3% perf-profile.children.cycles-pp.schedule
1.38 ? 3% +0.2 1.55 ? 3% perf-profile.children.cycles-pp.io_schedule
0.62 +0.2 0.80 ? 2% perf-profile.children.cycles-pp.__flush_smp_call_function_queue
1.40 +0.2 1.62 ? 2% perf-profile.children.cycles-pp.__wake_up_common
1.63 ? 2% +0.2 1.88 ? 3% perf-profile.children.cycles-pp.__schedule
1.32 +0.3 1.58 ? 2% perf-profile.children.cycles-pp.folio_add_file_rmap_range
0.51 +0.3 0.78 perf-profile.children.cycles-pp.finish_fault
1.57 +0.3 1.87 ? 2% perf-profile.children.cycles-pp.set_pte_range
1.53 +0.7 2.25 ? 2% perf-profile.children.cycles-pp.filemap_get_entry
2.09 +1.4 3.52 ? 4% perf-profile.children.cycles-pp.poll_idle
3.14 +2.1 5.20 ? 3% perf-profile.children.cycles-pp.do_rw_once
6.46 +2.4 8.82 perf-profile.children.cycles-pp._raw_spin_lock_irq
9.41 +2.9 12.33 perf-profile.children.cycles-pp.folio_wait_bit_common
20.55 +3.5 24.08 perf-profile.children.cycles-pp._raw_spin_lock_irqsave
22.53 +3.9 26.39 perf-profile.children.cycles-pp.folio_wake_bit
11.52 +3.9 15.42 perf-profile.children.cycles-pp.shmem_get_folio_gfp
11.67 +3.9 15.62 perf-profile.children.cycles-pp.shmem_fault
11.69 +4.0 15.65 perf-profile.children.cycles-pp.__do_fault
26.07 +5.6 31.69 perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
36.28 -9.6 26.64 perf-profile.self.cycles-pp.next_uptodate_folio
8.26 -1.3 7.00 ? 5% perf-profile.self.cycles-pp.acpi_safe_halt
8.35 -1.1 7.23 ? 2% perf-profile.self.cycles-pp.filemap_map_pages
0.91 ? 6% -0.8 0.14 ? 5% perf-profile.self.cycles-pp.up_read
0.47 ? 8% -0.1 0.37 ? 8% perf-profile.self.cycles-pp.__handle_mm_fault
0.12 ? 3% -0.0 0.08 ? 4% perf-profile.self.cycles-pp.down_read_trylock
0.05 +0.0 0.07 ? 7% perf-profile.self.cycles-pp.__switch_to_asm
0.06 +0.0 0.08 ? 5% perf-profile.self.cycles-pp.__switch_to
0.09 ? 3% +0.0 0.11 ? 4% perf-profile.self.cycles-pp.llist_add_batch
0.12 ? 2% +0.0 0.14 ? 3% perf-profile.self.cycles-pp.__flush_smp_call_function_queue
0.06 ? 5% +0.0 0.08 ? 6% perf-profile.self.cycles-pp.llist_reverse_order
0.06 +0.0 0.08 ? 4% perf-profile.self.cycles-pp.available_idle_cpu
0.06 +0.0 0.08 ? 5% perf-profile.self.cycles-pp.ttwu_queue_wakelist
0.06 ? 5% +0.0 0.09 ? 4% perf-profile.self.cycles-pp.newidle_balance
0.08 ? 4% +0.0 0.11 ? 5% perf-profile.self.cycles-pp.__list_del_entry_valid_or_report
0.06 ? 5% +0.0 0.09 ? 6% perf-profile.self.cycles-pp.update_rq_clock_task
0.06 +0.0 0.09 ? 3% perf-profile.self.cycles-pp.xas_load
0.06 ? 7% +0.0 0.09 ? 5% perf-profile.self.cycles-pp.enqueue_task_fair
0.18 ? 2% +0.0 0.21 ? 2% perf-profile.self.cycles-pp.__schedule
0.08 ? 5% +0.0 0.13 ? 3% perf-profile.self.cycles-pp.sync_regs
0.00 +0.1 0.05 perf-profile.self.cycles-pp.native_sched_clock
0.00 +0.1 0.06 ? 5% perf-profile.self.cycles-pp.xas_descend
0.00 +0.1 0.06 ? 8% perf-profile.self.cycles-pp.set_pte_range
0.00 +0.1 0.06 perf-profile.self.cycles-pp.enqueue_entity
0.26 +0.1 0.34 ? 4% perf-profile.self.cycles-pp._compound_head
0.24 ? 2% +0.1 0.31 ? 2% perf-profile.self.cycles-pp.__wake_up_common
0.14 ? 3% +0.1 0.23 perf-profile.self.cycles-pp.folio_unlock
0.28 ? 2% +0.1 0.38 perf-profile.self.cycles-pp._raw_spin_lock_irq
0.60 +0.1 0.72 perf-profile.self.cycles-pp.folio_wake_bit
0.30 ? 2% +0.1 0.44 perf-profile.self.cycles-pp.native_irq_return_iret
0.67 +0.2 0.83 perf-profile.self.cycles-pp._raw_spin_lock_irqsave
0.49 +0.2 0.69 ? 2% perf-profile.self.cycles-pp.shmem_get_folio_gfp
1.31 +0.3 1.57 ? 2% perf-profile.self.cycles-pp.folio_add_file_rmap_range
1.52 +0.4 1.90 perf-profile.self.cycles-pp.folio_wait_bit_common
1.48 +0.6 2.12 ? 2% perf-profile.self.cycles-pp.filemap_get_entry
1.57 +1.0 2.56 ? 2% perf-profile.self.cycles-pp.do_access
2.06 +1.4 3.47 ? 4% perf-profile.self.cycles-pp.poll_idle
2.42 +1.6 4.03 ? 3% perf-profile.self.cycles-pp.do_rw_once
26.07 +5.6 31.69 perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath




Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


--
0-DAY CI Kernel Test Service
https://github.com/intel/lkp-tests/wiki