* [linux-next:master] [mm] a6fde7add7: will-it-scale.per_thread_ops 176.7% improvement
@ 2025-07-23 6:12 kernel test robot
0 siblings, 0 replies; only message in thread
From: kernel test robot @ 2025-07-23 6:12 UTC (permalink / raw)
To: Barry Song
Cc: oe-lkp, lkp, Andrew Morton, Lorenzo Stoakes, Qi Zheng,
Liam R. Howlett, David Hildenbrand, Vlastimil Babka, Jann Horn,
Suren Baghdasaryan, Lokesh Gidra, Tangquan Zheng, Lance Yang,
linux-mm, oliver.sang
Hello,
kernel test robot noticed a 176.7% improvement of will-it-scale.per_thread_ops on:
commit: a6fde7add78d122f5e09cb6280f99c4b5ead7d56 ("mm: use per_vma lock for MADV_DONTNEED")
https://git.kernel.org/cgit/linux/kernel/git/next/linux-next.git master
testcase: will-it-scale
config: x86_64-rhel-9.4
compiler: gcc-12
test machine: 192 threads 4 sockets Intel(R) Xeon(R) Platinum 9242 CPU @ 2.30GHz (Cascade Lake) with 176G memory
parameters:
nr_task: 100%
mode: thread
test: tlb_flush2
cpufreq_governor: performance
In addition to that, the commit also has significant impact on the following tests:
+------------------+----------------------------------------------------------------+
| testcase: change | will-it-scale: will-it-scale.per_thread_ops 112.4% improvement |
| test machine | 104 threads 2 sockets (Skylake) with 192G memory |
| test parameters | cpufreq_governor=performance |
| | mode=thread |
| | nr_task=100% |
| | test=tlb_flush2 |
+------------------+----------------------------------------------------------------+
Details are as below:
-------------------------------------------------------------------------------------------------->
The kernel config and materials to reproduce are available at:
https://download.01.org/0day-ci/archive/20250723/202507230728.21b64efd-lkp@intel.com
=========================================================================================
compiler/cpufreq_governor/kconfig/mode/nr_task/rootfs/tbox_group/test/testcase:
gcc-12/performance/x86_64-rhel-9.4/thread/100%/debian-12-x86_64-20240206.cgz/lkp-csl-2sp10/tlb_flush2/will-it-scale
commit:
5e00e31867 ("userfaultfd: remove UFFD_CLOEXEC, UFFD_NONBLOCK, and UFFD_FLAGS_SET")
a6fde7add7 ("mm: use per_vma lock for MADV_DONTNEED")
5e00e31867d16e23 a6fde7add78d122f5e09cb6280f
---------------- ---------------------------
%stddev %change %stddev
\ | \
33172 -54.7% 15041 ± 6% uptime.idle
2.384e+10 -75.5% 5.852e+09 ± 7% cpuidle..time
2.244e+08 -88.1% 26677846 ± 3% cpuidle..usage
405945 +176.7% 1123225 will-it-scale.192.threads
42.80 -78.0% 9.40 ± 7% will-it-scale.192.threads_idle
2114 +176.7% 5849 will-it-scale.per_thread_ops
405945 +176.7% 1123225 will-it-scale.workload
723.50 ± 8% +34.0% 969.33 ± 3% perf-c2c.DRAM.local
25877 ± 3% +227.3% 84694 perf-c2c.DRAM.remote
13915 +154.7% 35449 perf-c2c.HITM.local
20352 ± 3% +256.9% 72638 perf-c2c.HITM.remote
34268 ± 2% +215.4% 108088 perf-c2c.HITM.total
43.37 -75.8% 10.48 ± 7% vmstat.cpu.id
56.32 +58.2% 89.07 vmstat.cpu.sy
66.58 +157.5% 171.45 vmstat.procs.r
594362 -94.2% 34588 ± 2% vmstat.system.cs
3910919 +96.4% 7681576 vmstat.system.in
2180450 -25.2% 1630420 meminfo.Active
2180450 -25.2% 1630420 meminfo.Active(anon)
4974053 -11.0% 4424807 meminfo.Cached
4042704 -13.7% 3488169 meminfo.Committed_AS
143355 ± 3% -31.0% 98964 meminfo.Mapped
1434003 -38.3% 884762 meminfo.Shmem
43.08 -33.2 9.90 ± 7% mpstat.cpu.all.idle%
22.91 -4.5 18.40 mpstat.cpu.all.irq%
0.16 -0.1 0.04 ± 3% mpstat.cpu.all.soft%
33.53 +37.7 71.22 mpstat.cpu.all.sys%
0.31 ± 6% +0.1 0.44 ± 8% mpstat.cpu.all.usr%
60.15 +55.5% 93.52 mpstat.max_utilization_pct
30748625 ± 3% +179.8% 86045170 numa-numastat.node0.local_node
30788809 ± 3% +179.7% 86123587 numa-numastat.node0.numa_hit
40287 ± 46% +95.1% 78587 ± 27% numa-numastat.node0.other_node
31393483 +173.9% 85988290 numa-numastat.node1.local_node
31478477 +173.4% 86057228 numa-numastat.node1.numa_hit
30878912 ± 4% +174.0% 84617170 numa-numastat.node2.local_node
30971736 ± 4% +173.4% 84686594 numa-numastat.node2.numa_hit
30733028 ± 4% +169.6% 82850466 numa-numastat.node3.local_node
30812231 ± 4% +169.2% 82932327 numa-numastat.node3.numa_hit
141120 ± 68% +125.7% 318471 ± 47% numa-meminfo.node0.Active
141120 ± 68% +125.7% 318471 ± 47% numa-meminfo.node0.Active(anon)
949762 ±124% -87.4% 120012 ± 55% numa-meminfo.node2.FilePages
20871 ±121% -94.9% 1054 ± 60% numa-meminfo.node2.Mapped
1367319 ± 21% -54.2% 625556 ± 18% numa-meminfo.node3.Active
1367319 ± 21% -54.2% 625556 ± 18% numa-meminfo.node3.Active(anon)
410132 ± 15% -52.6% 194262 ± 52% numa-meminfo.node3.AnonPages.max
58276 ± 24% -67.5% 18949 ± 22% numa-meminfo.node3.Mapped
1754799 ± 16% -35.5% 1131778 ± 12% numa-meminfo.node3.MemUsed
2883 ± 16% -25.6% 2146 ± 11% numa-meminfo.node3.PageTables
1041610 ± 27% -53.8% 481729 ± 22% numa-meminfo.node3.Shmem
545056 -25.2% 407612 proc-vmstat.nr_active_anon
1243447 -11.0% 1106212 proc-vmstat.nr_file_pages
35661 ± 2% -29.5% 25144 proc-vmstat.nr_mapped
358434 -38.3% 221200 proc-vmstat.nr_shmem
545056 -25.2% 407612 proc-vmstat.nr_zone_active_anon
55961 ± 22% -48.6% 28768 ± 31% proc-vmstat.numa_hint_faults_local
1.241e+08 +173.9% 3.398e+08 proc-vmstat.numa_hit
1.238e+08 +174.3% 3.395e+08 proc-vmstat.numa_local
1.242e+08 +173.8% 3.399e+08 proc-vmstat.pgalloc_normal
2.459e+08 +175.6% 6.776e+08 proc-vmstat.pgfault
1.237e+08 +174.5% 3.397e+08 proc-vmstat.pgfree
35272 ± 68% +125.7% 79623 ± 47% numa-vmstat.node0.nr_active_anon
35272 ± 68% +125.7% 79623 ± 47% numa-vmstat.node0.nr_zone_active_anon
30789775 ± 3% +179.7% 86130973 numa-vmstat.node0.numa_hit
30749591 ± 3% +179.8% 86052556 numa-vmstat.node0.numa_local
40287 ± 46% +95.1% 78587 ± 27% numa-vmstat.node0.numa_other
31478049 +173.4% 86064030 numa-vmstat.node1.numa_hit
31393056 +173.9% 85995094 numa-vmstat.node1.numa_local
237440 ±124% -87.4% 30004 ± 55% numa-vmstat.node2.nr_file_pages
5286 ±122% -95.8% 219.75 ± 66% numa-vmstat.node2.nr_mapped
30972241 ± 4% +173.4% 84692041 numa-vmstat.node2.numa_hit
30879418 ± 4% +174.0% 84622618 numa-vmstat.node2.numa_local
341715 ± 21% -54.2% 156460 ± 18% numa-vmstat.node3.nr_active_anon
14749 ± 25% -68.4% 4658 ± 24% numa-vmstat.node3.nr_mapped
721.34 ± 15% -25.6% 536.53 ± 11% numa-vmstat.node3.nr_page_table_pages
260294 ± 27% -53.7% 120499 ± 22% numa-vmstat.node3.nr_shmem
341714 ± 21% -54.2% 156459 ± 18% numa-vmstat.node3.nr_zone_active_anon
30811721 ± 4% +169.2% 82937335 numa-vmstat.node3.numa_hit
30732518 ± 4% +169.6% 82855474 numa-vmstat.node3.numa_local
6.40 ± 2% +58.3% 10.13 perf-stat.i.MPKI
6.954e+09 +90.6% 1.326e+10 perf-stat.i.branch-instructions
1.13 -0.2 0.95 perf-stat.i.branch-miss-rate%
79915661 +55.9% 1.246e+08 perf-stat.i.branch-misses
35.77 +5.8 41.57 perf-stat.i.cache-miss-rate%
1.974e+08 ± 2% +203.1% 5.984e+08 perf-stat.i.cache-misses
5.655e+08 ± 2% +154.4% 1.439e+09 perf-stat.i.cache-references
598927 -94.2% 34813 ± 2% perf-stat.i.context-switches
11.43 -20.4% 9.10 perf-stat.i.cpi
3.513e+11 +53.0% 5.375e+11 perf-stat.i.cpu-cycles
592.83 +111.1% 1251 perf-stat.i.cpu-migrations
1774 ± 2% -49.4% 898.01 perf-stat.i.cycles-between-cache-misses
3.103e+10 +90.1% 5.898e+10 perf-stat.i.instructions
0.09 +20.3% 0.11 perf-stat.i.ipc
11.60 +101.5% 23.37 perf-stat.i.metric.K/sec
813942 +175.6% 2243400 perf-stat.i.minor-faults
813942 +175.6% 2243400 perf-stat.i.page-faults
6.36 ± 2% +59.5% 10.15 perf-stat.overall.MPKI
1.15 -0.2 0.94 perf-stat.overall.branch-miss-rate%
34.89 +6.7 41.60 perf-stat.overall.cache-miss-rate%
11.32 -19.5% 9.11 perf-stat.overall.cpi
1780 ± 2% -49.6% 898.26 perf-stat.overall.cycles-between-cache-misses
0.09 +24.2% 0.11 perf-stat.overall.ipc
23055469 -31.3% 15833441 perf-stat.overall.path-length
6.931e+09 +90.5% 1.321e+10 perf-stat.ps.branch-instructions
79657784 +55.9% 1.242e+08 perf-stat.ps.branch-misses
1.967e+08 ± 2% +203.1% 5.963e+08 perf-stat.ps.cache-misses
5.639e+08 ± 2% +154.2% 1.433e+09 perf-stat.ps.cache-references
596817 -94.2% 34680 ± 2% perf-stat.ps.context-switches
3.501e+11 +53.0% 5.355e+11 perf-stat.ps.cpu-cycles
590.65 +110.9% 1245 perf-stat.ps.cpu-migrations
3.093e+10 +90.0% 5.877e+10 perf-stat.ps.instructions
811027 +175.6% 2235532 perf-stat.ps.minor-faults
811027 +175.6% 2235532 perf-stat.ps.page-faults
9.359e+12 +90.0% 1.778e+13 perf-stat.total.instructions
3568606 +373.7% 16903333 sched_debug.cfs_rq:/.avg_vruntime.avg
4465189 ± 4% +295.8% 17671644 sched_debug.cfs_rq:/.avg_vruntime.max
3129343 ± 2% +302.7% 12602026 sched_debug.cfs_rq:/.avg_vruntime.min
250285 ± 9% +258.8% 898123 ± 2% sched_debug.cfs_rq:/.avg_vruntime.stddev
0.32 ± 9% +104.6% 0.66 ± 3% sched_debug.cfs_rq:/.h_nr_queued.avg
0.45 ± 3% -11.8% 0.39 ± 4% sched_debug.cfs_rq:/.h_nr_queued.stddev
0.32 ± 9% +105.5% 0.66 ± 3% sched_debug.cfs_rq:/.h_nr_runnable.avg
1.25 ± 14% +28.9% 1.61 ± 7% sched_debug.cfs_rq:/.h_nr_runnable.max
0.44 ± 3% -11.1% 0.39 ± 4% sched_debug.cfs_rq:/.h_nr_runnable.stddev
2.81 ± 11% -80.2% 0.56 ± 87% sched_debug.cfs_rq:/.load_avg.min
3568606 +373.7% 16903333 sched_debug.cfs_rq:/.min_vruntime.avg
4465189 ± 4% +295.8% 17671644 sched_debug.cfs_rq:/.min_vruntime.max
3129343 ± 2% +302.7% 12602026 sched_debug.cfs_rq:/.min_vruntime.min
250285 ± 9% +258.8% 898123 ± 2% sched_debug.cfs_rq:/.min_vruntime.stddev
0.32 ± 9% +105.2% 0.66 ± 3% sched_debug.cfs_rq:/.nr_queued.avg
0.44 ± 3% -12.0% 0.39 ± 4% sched_debug.cfs_rq:/.nr_queued.stddev
328.85 ± 2% +102.4% 665.73 sched_debug.cfs_rq:/.runnable_avg.avg
859.00 ± 6% +77.0% 1520 ± 6% sched_debug.cfs_rq:/.runnable_avg.max
162.67 ± 11% -84.6% 24.97 ±144% sched_debug.cfs_rq:/.runnable_avg.min
83.72 ± 5% +95.0% 163.23 ± 8% sched_debug.cfs_rq:/.runnable_avg.stddev
328.43 ± 2% +102.2% 664.12 sched_debug.cfs_rq:/.util_avg.avg
852.03 ± 5% +73.1% 1474 ± 6% sched_debug.cfs_rq:/.util_avg.max
164.19 ± 10% -85.4% 24.03 ±143% sched_debug.cfs_rq:/.util_avg.min
82.62 ± 5% +92.9% 159.38 ± 8% sched_debug.cfs_rq:/.util_avg.stddev
113.06 ± 10% +362.8% 523.26 ± 4% sched_debug.cfs_rq:/.util_est.avg
500.78 ± 8% +141.7% 1210 ± 7% sched_debug.cfs_rq:/.util_est.max
159.89 ± 3% +89.6% 303.23 ± 3% sched_debug.cfs_rq:/.util_est.stddev
483053 ± 3% +77.1% 855327 sched_debug.cpu.avg_idle.avg
751868 ± 10% +40.4% 1055695 sched_debug.cpu.avg_idle.max
134891 ± 7% +107.8% 280294 ± 12% sched_debug.cpu.avg_idle.min
78767 ± 16% +89.2% 149010 sched_debug.cpu.avg_idle.stddev
6221 ± 9% -21.2% 4900 sched_debug.cpu.clock_task.stddev
2061 ± 8% +94.2% 4005 ± 3% sched_debug.cpu.curr->pid.avg
2655 ± 2% -13.8% 2288 ± 4% sched_debug.cpu.curr->pid.stddev
0.00 ± 41% +87.6% 0.00 ± 13% sched_debug.cpu.next_balance.stddev
0.32 ± 9% +104.2% 0.66 ± 3% sched_debug.cpu.nr_running.avg
0.45 ± 3% -11.6% 0.40 ± 4% sched_debug.cpu.nr_running.stddev
463856 -94.0% 27685 ± 2% sched_debug.cpu.nr_switches.avg
505561 ± 2% -90.2% 49365 ± 4% sched_debug.cpu.nr_switches.max
397023 ± 4% -94.3% 22498 ± 3% sched_debug.cpu.nr_switches.min
7607 ± 15% -45.5% 4149 ± 2% sched_debug.cpu.nr_switches.stddev
0.53 ± 5% -68.4% 0.17 ± 12% sched_debug.cpu.nr_uninterruptible.avg
165.69 ± 21% +2874.8% 4929 sched_debug.cpu.nr_uninterruptible.max
-153.56 +3099.7% -4913 sched_debug.cpu.nr_uninterruptible.min
22.58 ± 10% +6092.0% 1398 sched_debug.cpu.nr_uninterruptible.stddev
0.02 ± 86% +3043.1% 0.53 ± 53% perf-sched.sch_delay.avg.ms.__cond_resched.__alloc_frozen_pages_noprof.alloc_pages_mpol.vma_alloc_folio_noprof.alloc_anon_folio
0.03 ± 19% +1647.7% 0.58 ± 12% perf-sched.sch_delay.avg.ms.__cond_resched.__tlb_batch_free_encoded_pages.tlb_finish_mmu.do_madvise.__x64_sys_madvise
0.04 ± 89% +455.3% 0.20 ± 2% perf-sched.sch_delay.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
0.00 ±223% +7625.0% 0.10 ±124% perf-sched.sch_delay.avg.ms.__cond_resched.remove_vma.vms_complete_munmap_vmas.do_vmi_align_munmap.do_vmi_munmap
0.00 ± 39% +100.0% 0.01 ± 21% perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
0.08 ± 49% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.zap_page_range_single_batched.madvise_dontneed_free.madvise_vma_behavior.madvise_do_behavior
0.03 ±155% +2837.3% 0.77 ± 64% perf-sched.sch_delay.avg.ms.__cond_resched.zap_pmd_range.isra.0.unmap_page_range
0.05 ±105% +1070.2% 0.61 ± 30% perf-sched.sch_delay.avg.ms.__cond_resched.zap_pte_range.zap_pmd_range.isra.0
0.00 ± 10% +703.6% 0.04 ± 28% perf-sched.sch_delay.avg.ms.__vma_start_write.__split_vma.vms_gather_munmap_vmas.do_vmi_align_munmap
0.00 ± 10% +410.7% 0.02 ± 23% perf-sched.sch_delay.avg.ms.__vma_start_write.vma_expand.vma_merge_new_range.__mmap_region
0.01 ± 24% +604.9% 0.07 ±176% perf-sched.sch_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
0.01 ± 15% +361.4% 0.03 ± 43% perf-sched.sch_delay.avg.ms.anon_pipe_read.fifo_pipe_read.vfs_read.ksys_read
0.00 ± 28% +150.0% 0.00 ± 21% perf-sched.sch_delay.avg.ms.anon_pipe_read.vfs_read.ksys_read.do_syscall_64
0.01 ± 31% +289.7% 0.04 ± 62% perf-sched.sch_delay.avg.ms.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.11 ± 35% +304.0% 0.43 ± 45% perf-sched.sch_delay.avg.ms.exit_to_user_mode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
0.04 ±114% +1264.9% 0.52 ± 33% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
0.11 ± 52% +650.6% 0.79 ± 4% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
0.02 ±135% +1400.7% 0.34 ± 74% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown].[unknown]
0.01 ± 45% +703.2% 0.04 ± 39% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function.[unknown]
0.16 ± 20% +67.1% 0.26 ± 20% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown]
0.00 ± 28% +9300.0% 0.16 ± 50% perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.poll_schedule_timeout.constprop.0.do_poll
0.01 -100.0% 0.00 perf-sched.sch_delay.avg.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read.madvise_lock
0.01 ± 50% +353.5% 0.05 ± 21% perf-sched.sch_delay.avg.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read_killable.lock_mm_and_find_vma
0.01 ± 8% +752.9% 0.05 ± 12% perf-sched.sch_delay.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write_killable.__vm_munmap
0.01 ± 21% +292.3% 0.03 ± 15% perf-sched.sch_delay.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write_killable.vm_mmap_pgoff
0.01 ± 20% +1050.0% 0.08 ± 27% perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
0.01 ± 9% +206.8% 0.03 ± 69% perf-sched.sch_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
0.01 ± 19% +265.8% 0.04 ± 69% perf-sched.sch_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
0.01 ± 23% +1270.5% 0.10 ± 18% perf-sched.sch_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
0.06 ±130% +13233.7% 7.44 ± 91% perf-sched.sch_delay.max.ms.__cond_resched.__alloc_frozen_pages_noprof.alloc_pages_mpol.vma_alloc_folio_noprof.alloc_anon_folio
0.77 ± 23% +6200.8% 48.80 ± 96% perf-sched.sch_delay.max.ms.__cond_resched.__tlb_batch_free_encoded_pages.tlb_finish_mmu.do_madvise.__x64_sys_madvise
0.00 ±223% +1.1e+05% 1.42 ±138% perf-sched.sch_delay.max.ms.__cond_resched.remove_vma.vms_complete_munmap_vmas.do_vmi_align_munmap.do_vmi_munmap
0.07 ± 83% +116.5% 0.15 ± 23% perf-sched.sch_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
0.02 ± 52% -64.2% 0.01 ± 83% perf-sched.sch_delay.max.ms.__cond_resched.unmap_vmas.vms_clear_ptes.part.0
0.47 ± 29% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.zap_page_range_single_batched.madvise_dontneed_free.madvise_vma_behavior.madvise_do_behavior
0.06 ±183% +5567.6% 3.52 ± 45% perf-sched.sch_delay.max.ms.__cond_resched.zap_pmd_range.isra.0.unmap_page_range
0.15 ±107% +5160.3% 7.81 ± 94% perf-sched.sch_delay.max.ms.__cond_resched.zap_pte_range.zap_pmd_range.isra.0
0.02 ± 12% +1.6e+05% 25.23 ± 54% perf-sched.sch_delay.max.ms.__vma_start_write.__split_vma.vms_gather_munmap_vmas.do_vmi_align_munmap
0.02 ± 29% +1.6e+05% 24.02 ± 62% perf-sched.sch_delay.max.ms.__vma_start_write.vma_expand.vma_merge_new_range.__mmap_region
0.35 ± 63% +1071.6% 4.11 ±121% perf-sched.sch_delay.max.ms.anon_pipe_read.fifo_pipe_read.vfs_read.ksys_read
0.49 ± 57% +197.1% 1.47 ± 69% perf-sched.sch_delay.max.ms.anon_pipe_read.vfs_read.ksys_read.do_syscall_64
0.20 ± 98% +842.6% 1.88 ± 76% perf-sched.sch_delay.max.ms.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.78 ± 24% +721.1% 6.39 ± 44% perf-sched.sch_delay.max.ms.exit_to_user_mode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
0.23 ± 88% +2103.1% 5.05 ± 9% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
0.61 ± 71% +1074.1% 7.20 ± 72% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
0.03 ±144% +2054.5% 0.69 ± 58% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown].[unknown]
0.01 ± 45% +6350.0% 0.34 ± 65% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function.[unknown]
0.34 ± 26% +424.4% 1.79 ± 43% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown]
0.61 ± 27% +152.4% 1.54 ± 42% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
0.08 ± 59% +2324.9% 1.85 ± 82% perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.poll_schedule_timeout.constprop.0.do_poll
12.13 ± 10% -100.0% 0.00 perf-sched.sch_delay.max.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read.madvise_lock
0.37 ±118% +9135.1% 33.91 ± 33% perf-sched.sch_delay.max.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read_killable.lock_mm_and_find_vma
5.65 ± 45% +960.4% 59.96 ± 70% perf-sched.sch_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write_killable.__vm_munmap
4.13 ± 80% +495.1% 24.59 ± 55% perf-sched.sch_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write_killable.vm_mmap_pgoff
0.69 ±223% +1343.9% 9.96 ± 87% perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr
0.12 ± 47% +2736.5% 3.30 ± 13% perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
0.04 ± 60% +1121.5% 0.46 ±163% perf-sched.sch_delay.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
3.66 ± 93% +1131.1% 45.04 ± 71% perf-sched.sch_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
0.01 ± 39% +835.3% 0.08 ± 10% perf-sched.total_sch_delay.average.ms
2.24 +990.8% 24.40 perf-sched.total_wait_and_delay.average.ms
1210914 -93.1% 83432 ± 2% perf-sched.total_wait_and_delay.count.ms
3800 ± 13% +23.8% 4706 ± 6% perf-sched.total_wait_and_delay.max.ms
2.23 +991.2% 24.32 perf-sched.total_wait_time.average.ms
3800 ± 13% +23.8% 4706 ± 6% perf-sched.total_wait_time.max.ms
3.67 ± 8% +70.0% 6.24 ± 36% perf-sched.wait_and_delay.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
119.96 ± 2% +17.3% 140.76 ± 3% perf-sched.wait_and_delay.avg.ms.anon_pipe_read.fifo_pipe_read.vfs_read.ksys_read
66.65 ± 5% +697.3% 531.40 ± 13% perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range_clock.poll_schedule_timeout.constprop.0.do_poll
0.71 -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read.madvise_lock
0.43 +449.0% 2.37 ± 4% perf-sched.wait_and_delay.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write_killable.__vm_munmap
3.30 ± 2% +38.5% 4.57 perf-sched.wait_and_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
370.33 ± 3% -10.3% 332.23 ± 4% perf-sched.wait_and_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
2.00 ± 81% +575.0% 13.50 ± 24% perf-sched.wait_and_delay.count.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
370.00 ± 2% -12.7% 323.17 ± 3% perf-sched.wait_and_delay.count.anon_pipe_read.fifo_pipe_read.vfs_read.ksys_read
235.67 ± 4% -89.8% 24.00 ± 4% perf-sched.wait_and_delay.count.schedule_hrtimeout_range_clock.poll_schedule_timeout.constprop.0.do_poll
1174293 -100.0% 0.00 perf-sched.wait_and_delay.count.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read.madvise_lock
21019 +31.3% 27603 perf-sched.wait_and_delay.count.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write_killable.__vm_munmap
1513 ± 2% -27.0% 1105 perf-sched.wait_and_delay.count.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
1000 +196.8% 2970 ± 66% perf-sched.wait_and_delay.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
379.57 ±152% +430.7% 2014 ± 58% perf-sched.wait_and_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
21.98 ± 11% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read.madvise_lock
11.88 ± 11% +658.6% 90.10 ±107% perf-sched.wait_and_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write_killable.__vm_munmap
6.17 ± 21% +450.9% 34.00 ±105% perf-sched.wait_and_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
0.73 ± 11% +183.5% 2.07 ± 19% perf-sched.wait_time.avg.ms.__cond_resched.__alloc_frozen_pages_noprof.alloc_pages_mpol.vma_alloc_folio_noprof.alloc_anon_folio
0.76 ± 12% +167.2% 2.03 ± 5% perf-sched.wait_time.avg.ms.__cond_resched.__tlb_batch_free_encoded_pages.tlb_finish_mmu.do_madvise.__x64_sys_madvise
3.63 ± 8% +66.2% 6.04 ± 37% perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
0.00 ±223% +71650.0% 1.91 ± 27% perf-sched.wait_time.avg.ms.__cond_resched.remove_vma.vms_complete_munmap_vmas.do_vmi_align_munmap.do_vmi_munmap
0.82 ± 11% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.zap_page_range_single_batched.madvise_dontneed_free.madvise_vma_behavior.madvise_do_behavior
0.67 ± 49% +223.8% 2.16 ± 39% perf-sched.wait_time.avg.ms.__cond_resched.zap_pmd_range.isra.0.unmap_page_range
0.83 ± 27% +152.0% 2.08 ± 12% perf-sched.wait_time.avg.ms.__cond_resched.zap_pte_range.zap_pmd_range.isra.0
0.22 ± 20% +912.6% 2.24 ± 5% perf-sched.wait_time.avg.ms.__vma_start_write.__split_vma.vms_gather_munmap_vmas.do_vmi_align_munmap
0.27 ± 24% +625.8% 1.94 ± 6% perf-sched.wait_time.avg.ms.__vma_start_write.vma_expand.vma_merge_new_range.__mmap_region
119.96 ± 2% +17.3% 140.73 ± 3% perf-sched.wait_time.avg.ms.anon_pipe_read.fifo_pipe_read.vfs_read.ksys_read
0.44 ± 2% +261.4% 1.60 ±147% perf-sched.wait_time.avg.ms.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.45 ± 37% +164.6% 1.19 ± 41% perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
0.75 ± 11% +152.8% 1.90 ± 5% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
0.20 ± 36% +323.6% 0.84 ± 4% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
0.17 ± 17% +51.5% 0.26 ± 20% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown]
66.65 ± 5% +697.1% 531.24 ± 13% perf-sched.wait_time.avg.ms.schedule_hrtimeout_range_clock.poll_schedule_timeout.constprop.0.do_poll
0.70 -100.0% 0.00 perf-sched.wait_time.avg.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read.madvise_lock
0.60 ± 8% +193.4% 1.77 ± 4% perf-sched.wait_time.avg.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read_killable.lock_mm_and_find_vma
0.43 +445.2% 2.32 ± 4% perf-sched.wait_time.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write_killable.__vm_munmap
0.29 +885.3% 2.85 ± 3% perf-sched.wait_time.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write_killable.vm_mmap_pgoff
0.86 +26.1% 1.09 ± 6% perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
3.29 ± 2% +37.7% 4.53 perf-sched.wait_time.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
370.32 ± 3% -10.3% 332.13 ± 4% perf-sched.wait_time.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
0.94 ± 23% +940.8% 9.82 ± 59% perf-sched.wait_time.max.ms.__cond_resched.__alloc_frozen_pages_noprof.alloc_pages_mpol.vma_alloc_folio_noprof.alloc_anon_folio
2.95 ± 87% +1596.0% 49.96 ± 97% perf-sched.wait_time.max.ms.__cond_resched.__tlb_batch_free_encoded_pages.tlb_finish_mmu.do_madvise.__x64_sys_madvise
1000 +196.8% 2969 ± 66% perf-sched.wait_time.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
0.00 ±223% +2.4e+05% 6.47 ± 15% perf-sched.wait_time.max.ms.__cond_resched.remove_vma.vms_complete_munmap_vmas.do_vmi_align_munmap.do_vmi_munmap
379.54 ±152% +430.8% 2014 ± 58% perf-sched.wait_time.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
1.41 ± 12% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.zap_page_range_single_batched.madvise_dontneed_free.madvise_vma_behavior.madvise_do_behavior
0.75 ± 55% +642.4% 5.53 ± 26% perf-sched.wait_time.max.ms.__cond_resched.zap_pmd_range.isra.0.unmap_page_range
1.18 ± 17% +882.6% 11.63 ± 66% perf-sched.wait_time.max.ms.__cond_resched.zap_pte_range.zap_pmd_range.isra.0
2.11 ± 58% +1003.0% 23.25 ± 41% perf-sched.wait_time.max.ms.__vma_start_write.__split_vma.vms_gather_munmap_vmas.do_vmi_align_munmap
1.91 ± 76% +1255.0% 25.84 ± 65% perf-sched.wait_time.max.ms.__vma_start_write.vma_expand.vma_merge_new_range.__mmap_region
23.98 +714.1% 195.25 ±184% perf-sched.wait_time.max.ms.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe
1.50 ± 47% +557.6% 9.90 ± 21% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
0.89 ± 28% +765.6% 7.66 ± 65% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
0.43 ± 29% +318.3% 1.79 ± 43% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown]
0.67 ± 32% +287.2% 2.59 ± 60% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
13.59 ± 23% -100.0% 0.00 perf-sched.wait_time.max.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read.madvise_lock
1.57 ± 19% +1798.2% 29.87 ± 53% perf-sched.wait_time.max.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read_killable.lock_mm_and_find_vma
11.83 ± 11% +357.4% 54.13 ± 85% perf-sched.wait_time.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write_killable.__vm_munmap
10.35 ± 25% +179.6% 28.95 ± 56% perf-sched.wait_time.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write_killable.vm_mmap_pgoff
0.69 ±223% +5790.5% 40.64 ±144% perf-sched.wait_time.max.ms.schedule_timeout.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr
2.95 ± 3% +94.4% 5.74 ± 20% perf-sched.wait_time.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
41.56 -37.5 4.04 ± 7% perf-profile.calltrace.cycles-pp.common_startup_64
41.34 -37.3 4.02 ± 7% perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.common_startup_64
41.34 -37.3 4.02 ± 7% perf-profile.calltrace.cycles-pp.start_secondary.common_startup_64
41.32 -37.3 4.02 ± 7% perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.common_startup_64
39.04 -35.1 3.97 ± 7% perf-profile.calltrace.cycles-pp.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary.common_startup_64
38.29 -34.3 3.96 ± 7% perf-profile.calltrace.cycles-pp.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary
38.19 -34.2 3.96 ± 7% perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry
22.62 -22.6 0.00 perf-profile.calltrace.cycles-pp.asm_sysvec_call_function.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
22.52 -22.5 0.00 perf-profile.calltrace.cycles-pp.sysvec_call_function.asm_sysvec_call_function.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call
22.43 -22.4 0.00 perf-profile.calltrace.cycles-pp.__sysvec_call_function.sysvec_call_function.asm_sysvec_call_function.cpuidle_enter_state.cpuidle_enter
22.29 -22.3 0.00 perf-profile.calltrace.cycles-pp.__flush_smp_call_function_queue.__sysvec_call_function.sysvec_call_function.asm_sysvec_call_function.cpuidle_enter_state
22.58 -19.9 2.72 perf-profile.calltrace.cycles-pp.flush_tlb_func.__flush_smp_call_function_queue.__sysvec_call_function.sysvec_call_function.asm_sysvec_call_function
9.10 ± 4% -9.1 0.00 perf-profile.calltrace.cycles-pp.madvise_lock.do_madvise.__x64_sys_madvise.do_syscall_64.entry_SYSCALL_64_after_hwframe
9.08 ± 4% -9.1 0.00 perf-profile.calltrace.cycles-pp.down_read.madvise_lock.do_madvise.__x64_sys_madvise.do_syscall_64
11.40 -8.9 2.49 ± 8% perf-profile.calltrace.cycles-pp.intel_idle.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
8.69 ± 4% -8.7 0.00 perf-profile.calltrace.cycles-pp.rwsem_down_read_slowpath.down_read.madvise_lock.do_madvise.__x64_sys_madvise
6.90 ± 5% -6.9 0.00 perf-profile.calltrace.cycles-pp._raw_spin_lock_irq.rwsem_down_read_slowpath.down_read.madvise_lock.do_madvise
6.77 ± 5% -6.8 0.00 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irq.rwsem_down_read_slowpath.down_read.madvise_lock
0.80 ± 8% -0.5 0.30 ±100% perf-profile.calltrace.cycles-pp.lock_vma_under_rcu.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.testcase
0.88 +0.3 1.19 perf-profile.calltrace.cycles-pp.default_send_IPI_mask_sequence_phys.smp_call_function_many_cond.on_each_cpu_cond_mask.flush_tlb_mm_range.tlb_finish_mmu
0.64 ± 7% +0.6 1.24 ± 8% perf-profile.calltrace.cycles-pp.do_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
1.74 ± 7% +0.6 2.36 ± 9% perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.testcase
1.75 ± 7% +0.6 2.37 ± 9% perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.testcase
0.73 ± 9% +0.6 1.36 ± 9% perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
0.80 ± 10% +0.7 1.49 ± 10% perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.testcase
0.00 +0.7 0.71 ± 5% perf-profile.calltrace.cycles-pp.tlb_gather_mmu.do_madvise.__x64_sys_madvise.do_syscall_64.entry_SYSCALL_64_after_hwframe
1.88 ± 7% +0.7 2.63 ± 8% perf-profile.calltrace.cycles-pp.asm_exc_page_fault.testcase
2.00 ± 7% +0.9 2.86 ± 10% perf-profile.calltrace.cycles-pp.testcase
0.00 +0.9 0.86 ± 6% perf-profile.calltrace.cycles-pp.zap_pte_range.zap_pmd_range.unmap_page_range.zap_page_range_single_batched.madvise_dontneed_free
0.00 +0.9 0.87 ± 6% perf-profile.calltrace.cycles-pp.zap_pmd_range.unmap_page_range.zap_page_range_single_batched.madvise_dontneed_free.madvise_vma_behavior
0.89 ± 12% +0.9 1.80 ± 10% perf-profile.calltrace.cycles-pp.madvise_do_behavior.do_madvise.__x64_sys_madvise.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +0.9 0.91 ± 6% perf-profile.calltrace.cycles-pp.unmap_page_range.zap_page_range_single_batched.madvise_dontneed_free.madvise_vma_behavior.madvise_walk_vmas
0.00 +1.4 1.41 ± 10% perf-profile.calltrace.cycles-pp.zap_page_range_single_batched.madvise_dontneed_free.madvise_vma_behavior.madvise_walk_vmas.madvise_do_behavior
0.00 +1.4 1.44 ± 10% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.intel_idle_irq.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call
0.00 +1.4 1.44 ± 10% perf-profile.calltrace.cycles-pp.madvise_dontneed_free.madvise_vma_behavior.madvise_walk_vmas.madvise_do_behavior.do_madvise
0.00 +1.4 1.45 ± 10% perf-profile.calltrace.cycles-pp.madvise_vma_behavior.madvise_walk_vmas.madvise_do_behavior.do_madvise.__x64_sys_madvise
0.00 +1.8 1.78 ± 10% perf-profile.calltrace.cycles-pp.madvise_walk_vmas.madvise_do_behavior.do_madvise.__x64_sys_madvise.do_syscall_64
4.37 +7.7 12.10 perf-profile.calltrace.cycles-pp.llist_reverse_order.__flush_smp_call_function_queue.__sysvec_call_function.sysvec_call_function.asm_sysvec_call_function
4.94 +12.0 16.94 perf-profile.calltrace.cycles-pp.__sysvec_call_function.sysvec_call_function.asm_sysvec_call_function.smp_call_function_many_cond.on_each_cpu_cond_mask
4.84 +12.0 16.88 perf-profile.calltrace.cycles-pp.__flush_smp_call_function_queue.__sysvec_call_function.sysvec_call_function.asm_sysvec_call_function.smp_call_function_many_cond
5.05 +12.1 17.16 perf-profile.calltrace.cycles-pp.sysvec_call_function.asm_sysvec_call_function.smp_call_function_many_cond.on_each_cpu_cond_mask.flush_tlb_mm_range
6.42 +12.3 18.77 perf-profile.calltrace.cycles-pp.asm_sysvec_call_function.smp_call_function_many_cond.on_each_cpu_cond_mask.flush_tlb_mm_range.tlb_finish_mmu
55.56 +36.7 92.29 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__madvise
55.56 +36.7 92.30 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__madvise
55.46 +36.8 92.25 perf-profile.calltrace.cycles-pp.__x64_sys_madvise.do_syscall_64.entry_SYSCALL_64_after_hwframe.__madvise
55.46 +36.8 92.25 perf-profile.calltrace.cycles-pp.do_madvise.__x64_sys_madvise.do_syscall_64.entry_SYSCALL_64_after_hwframe.__madvise
55.73 +36.8 92.56 perf-profile.calltrace.cycles-pp.__madvise
43.09 +43.8 86.92 perf-profile.calltrace.cycles-pp.smp_call_function_many_cond.on_each_cpu_cond_mask.flush_tlb_mm_range.tlb_finish_mmu.do_madvise
43.77 +44.0 87.76 perf-profile.calltrace.cycles-pp.on_each_cpu_cond_mask.flush_tlb_mm_range.tlb_finish_mmu.do_madvise.__x64_sys_madvise
44.35 +44.1 88.41 perf-profile.calltrace.cycles-pp.flush_tlb_mm_range.tlb_finish_mmu.do_madvise.__x64_sys_madvise.do_syscall_64
44.90 +44.8 89.67 perf-profile.calltrace.cycles-pp.tlb_finish_mmu.do_madvise.__x64_sys_madvise.do_syscall_64.entry_SYSCALL_64_after_hwframe
41.56 -37.5 4.04 ± 7% perf-profile.children.cycles-pp.common_startup_64
41.56 -37.5 4.04 ± 7% perf-profile.children.cycles-pp.cpu_startup_entry
41.56 -37.5 4.04 ± 7% perf-profile.children.cycles-pp.do_idle
41.34 -37.3 4.02 ± 7% perf-profile.children.cycles-pp.start_secondary
39.25 -35.3 4.00 ± 7% perf-profile.children.cycles-pp.cpuidle_idle_call
38.49 -34.5 3.98 ± 7% perf-profile.children.cycles-pp.cpuidle_enter
38.49 -34.5 3.98 ± 7% perf-profile.children.cycles-pp.cpuidle_enter_state
24.11 -20.8 3.27 perf-profile.children.cycles-pp.flush_tlb_func
31.48 -12.6 18.93 perf-profile.children.cycles-pp.__flush_smp_call_function_queue
32.54 -12.3 20.20 perf-profile.children.cycles-pp.asm_sysvec_call_function
30.92 -11.9 19.04 perf-profile.children.cycles-pp.sysvec_call_function
30.66 -11.9 18.78 perf-profile.children.cycles-pp.__sysvec_call_function
9.10 ± 4% -9.1 0.00 perf-profile.children.cycles-pp.madvise_lock
9.10 ± 4% -9.1 0.00 perf-profile.children.cycles-pp.down_read
11.48 -9.0 2.51 ± 8% perf-profile.children.cycles-pp.intel_idle
8.73 ± 4% -8.7 0.00 perf-profile.children.cycles-pp.rwsem_down_read_slowpath
6.99 ± 5% -7.0 0.00 perf-profile.children.cycles-pp._raw_spin_lock_irq
6.91 ± 5% -6.8 0.14 ± 4% perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
2.43 ± 3% -2.3 0.10 ± 4% perf-profile.children.cycles-pp.__schedule
2.99 ± 7% -1.7 1.32 ± 5% perf-profile.children.cycles-pp.intel_idle_irq
0.96 ± 4% -0.9 0.06 ± 9% perf-profile.children.cycles-pp.schedule_preempt_disabled
0.97 ± 4% -0.9 0.07 perf-profile.children.cycles-pp.schedule
0.54 ± 8% -0.5 0.05 perf-profile.children.cycles-pp.__pick_next_task
0.96 ± 2% -0.5 0.48 perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
0.42 ± 11% -0.4 0.04 ± 44% perf-profile.children.cycles-pp.sched_balance_newidle
0.43 ± 10% -0.4 0.06 perf-profile.children.cycles-pp.sched_balance_rq
0.53 ± 2% -0.4 0.18 perf-profile.children.cycles-pp.__irq_exit_rcu
0.34 ± 8% -0.3 0.05 ± 7% perf-profile.children.cycles-pp.sched_balance_find_src_group
0.34 ± 8% -0.3 0.05 perf-profile.children.cycles-pp.update_sd_lb_stats
0.32 ± 7% -0.3 0.05 perf-profile.children.cycles-pp.update_sg_lb_stats
0.58 ± 2% -0.2 0.35 perf-profile.children.cycles-pp._raw_spin_lock
0.25 ± 4% -0.1 0.16 ± 2% perf-profile.children.cycles-pp.asm_sysvec_call_function_single
0.53 ± 2% -0.1 0.44 ± 2% perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
0.52 ± 2% -0.1 0.44 perf-profile.children.cycles-pp.hrtimer_interrupt
0.24 ± 5% -0.1 0.15 perf-profile.children.cycles-pp.__sysvec_call_function_single
0.24 ± 5% -0.1 0.16 ± 2% perf-profile.children.cycles-pp.sysvec_call_function_single
0.38 -0.1 0.30 perf-profile.children.cycles-pp.__munmap
0.14 ± 3% -0.1 0.06 perf-profile.children.cycles-pp.update_load_avg
0.38 -0.1 0.30 perf-profile.children.cycles-pp.__vm_munmap
0.38 -0.1 0.30 perf-profile.children.cycles-pp.__x64_sys_munmap
0.16 ± 2% -0.1 0.09 ± 5% perf-profile.children.cycles-pp.native_sched_clock
0.30 -0.1 0.24 perf-profile.children.cycles-pp.vms_complete_munmap_vmas
0.11 ± 4% -0.1 0.05 ± 7% perf-profile.children.cycles-pp.rwsem_down_write_slowpath
0.12 ± 4% -0.1 0.06 ± 8% perf-profile.children.cycles-pp.down_write_killable
0.31 -0.1 0.25 perf-profile.children.cycles-pp.do_vmi_align_munmap
0.31 -0.1 0.25 perf-profile.children.cycles-pp.do_vmi_munmap
0.45 ± 2% -0.0 0.40 perf-profile.children.cycles-pp.__hrtimer_run_queues
0.05 +0.0 0.06 perf-profile.children.cycles-pp.___perf_sw_event
0.05 +0.0 0.06 perf-profile.children.cycles-pp.rmqueue
0.56 +0.0 0.58 perf-profile.children.cycles-pp._find_next_bit
0.03 ± 70% +0.0 0.06 ± 6% perf-profile.children.cycles-pp.error_entry
0.13 ± 2% +0.0 0.16 ± 2% perf-profile.children.cycles-pp.irqtime_account_irq
0.05 +0.0 0.08 ± 4% perf-profile.children.cycles-pp.clear_page_erms
0.06 +0.0 0.09 ± 4% perf-profile.children.cycles-pp.free_pages_and_swap_cache
0.06 ± 7% +0.0 0.10 ± 3% perf-profile.children.cycles-pp.__tlb_batch_free_encoded_pages
0.11 +0.1 0.16 ± 2% perf-profile.children.cycles-pp.get_page_from_freelist
0.05 ± 8% +0.1 0.11 ± 4% perf-profile.children.cycles-pp.sync_regs
0.00 +0.1 0.05 ± 13% perf-profile.children.cycles-pp.try_charge_memcg
0.14 ± 3% +0.1 0.19 perf-profile.children.cycles-pp.native_flush_tlb_local
0.31 ± 2% +0.1 0.37 ± 2% perf-profile.children.cycles-pp.tick_nohz_handler
0.00 +0.1 0.06 perf-profile.children.cycles-pp.__page_cache_release
0.00 +0.1 0.06 ± 9% perf-profile.children.cycles-pp.charge_memcg
0.00 +0.1 0.06 perf-profile.children.cycles-pp.folio_add_new_anon_rmap
0.00 +0.1 0.06 perf-profile.children.cycles-pp.lru_add
0.08 ± 24% +0.1 0.14 ± 19% perf-profile.children.cycles-pp.__mem_cgroup_charge
0.12 ± 4% +0.1 0.19 ± 2% perf-profile.children.cycles-pp.__alloc_frozen_pages_noprof
0.27 +0.1 0.35 perf-profile.children.cycles-pp.update_process_times
0.14 ± 3% +0.1 0.21 ± 2% perf-profile.children.cycles-pp.alloc_pages_mpol
0.00 +0.1 0.08 ± 4% perf-profile.children.cycles-pp.__perf_sw_event
0.15 ± 3% +0.1 0.23 ± 2% perf-profile.children.cycles-pp.vma_alloc_folio_noprof
0.11 ± 3% +0.1 0.20 ± 2% perf-profile.children.cycles-pp.folios_put_refs
0.16 ± 4% +0.1 0.27 perf-profile.children.cycles-pp.sched_tick
0.11 ± 3% +0.1 0.23 perf-profile.children.cycles-pp.vms_clear_ptes
0.10 +0.1 0.23 perf-profile.children.cycles-pp.folio_batch_move_lru
0.06 +0.1 0.19 ± 2% perf-profile.children.cycles-pp.mm_needs_global_asid
0.00 +0.1 0.14 ± 3% perf-profile.children.cycles-pp.update_cfs_group
0.13 +0.1 0.28 perf-profile.children.cycles-pp.folio_add_lru
0.22 ± 3% +0.2 0.36 ± 2% perf-profile.children.cycles-pp.__pte_offset_map_lock
0.24 ± 8% +0.2 0.39 ± 6% perf-profile.children.cycles-pp.alloc_anon_folio
0.06 +0.2 0.23 ± 2% perf-profile.children.cycles-pp.task_tick_fair
0.42 ± 7% +0.2 0.65 ± 7% perf-profile.children.cycles-pp.native_irq_return_iret
0.91 +0.3 1.21 perf-profile.children.cycles-pp.default_send_IPI_mask_sequence_phys
0.12 ± 13% +0.4 0.54 ± 7% perf-profile.children.cycles-pp.flush_tlb_batched_pending
0.26 ± 7% +0.5 0.73 ± 6% perf-profile.children.cycles-pp.tlb_gather_mmu
0.43 ± 9% +0.5 0.92 ± 6% perf-profile.children.cycles-pp.unmap_page_range
0.38 ± 6% +0.5 0.87 ± 6% perf-profile.children.cycles-pp.zap_pte_range
0.39 ± 6% +0.5 0.88 ± 6% perf-profile.children.cycles-pp.zap_pmd_range
0.64 ± 7% +0.6 1.25 ± 8% perf-profile.children.cycles-pp.do_anonymous_page
1.75 ± 7% +0.6 2.36 ± 9% perf-profile.children.cycles-pp.do_user_addr_fault
1.76 ± 7% +0.6 2.38 ± 9% perf-profile.children.cycles-pp.exc_page_fault
0.73 ± 9% +0.6 1.36 ± 9% perf-profile.children.cycles-pp.__handle_mm_fault
0.80 ± 10% +0.7 1.50 ± 10% perf-profile.children.cycles-pp.handle_mm_fault
0.68 ± 12% +0.8 1.45 ± 10% perf-profile.children.cycles-pp.madvise_vma_behavior
0.66 ± 13% +0.8 1.44 ± 10% perf-profile.children.cycles-pp.madvise_dontneed_free
0.62 ± 13% +0.8 1.42 ± 11% perf-profile.children.cycles-pp.zap_page_range_single_batched
1.98 ± 6% +0.8 2.81 ± 8% perf-profile.children.cycles-pp.asm_exc_page_fault
0.90 ± 12% +0.9 1.80 ± 10% perf-profile.children.cycles-pp.madvise_do_behavior
2.15 ± 7% +1.0 3.16 ± 9% perf-profile.children.cycles-pp.testcase
0.00 +1.8 1.78 ± 10% perf-profile.children.cycles-pp.madvise_walk_vmas
5.40 +8.0 13.44 perf-profile.children.cycles-pp.llist_reverse_order
56.17 +36.5 92.67 perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
56.16 +36.5 92.66 perf-profile.children.cycles-pp.do_syscall_64
55.46 +36.8 92.25 perf-profile.children.cycles-pp.do_madvise
55.46 +36.8 92.25 perf-profile.children.cycles-pp.__x64_sys_madvise
55.74 +36.8 92.59 perf-profile.children.cycles-pp.__madvise
43.86 +44.1 87.98 perf-profile.children.cycles-pp.smp_call_function_many_cond
43.86 +44.1 87.98 perf-profile.children.cycles-pp.on_each_cpu_cond_mask
44.45 +44.2 88.63 perf-profile.children.cycles-pp.flush_tlb_mm_range
45.01 +44.9 89.91 perf-profile.children.cycles-pp.tlb_finish_mmu
23.91 -21.0 2.89 perf-profile.self.cycles-pp.flush_tlb_func
11.48 -9.0 2.51 ± 8% perf-profile.self.cycles-pp.intel_idle
6.91 ± 5% -6.8 0.12 ± 3% perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
1.63 ± 8% -0.4 1.24 ± 5% perf-profile.self.cycles-pp.intel_idle_irq
0.49 -0.3 0.20 perf-profile.self.cycles-pp._raw_spin_lock
0.64 ± 7% -0.2 0.49 ± 11% perf-profile.self.cycles-pp.lock_vma_under_rcu
0.16 ± 2% -0.1 0.09 ± 5% perf-profile.self.cycles-pp.native_sched_clock
0.50 ± 2% -0.1 0.44 perf-profile.self.cycles-pp._find_next_bit
0.10 ± 4% -0.0 0.08 ± 6% perf-profile.self.cycles-pp.asm_sysvec_call_function
0.00 +0.1 0.05 perf-profile.self.cycles-pp.irqtime_account_irq
0.01 ±223% +0.1 0.06 perf-profile.self.cycles-pp.folios_put_refs
0.05 ± 8% +0.1 0.11 ± 4% perf-profile.self.cycles-pp.sync_regs
0.14 ± 3% +0.1 0.19 perf-profile.self.cycles-pp.native_flush_tlb_local
0.00 +0.1 0.06 perf-profile.self.cycles-pp.error_entry
0.06 ± 6% +0.1 0.13 ± 3% perf-profile.self.cycles-pp.do_user_addr_fault
0.00 +0.1 0.07 perf-profile.self.cycles-pp.clear_page_erms
0.00 +0.1 0.07 perf-profile.self.cycles-pp.madvise_walk_vmas
0.03 ± 70% +0.1 0.14 ± 2% perf-profile.self.cycles-pp.mm_needs_global_asid
0.00 +0.1 0.14 ± 3% perf-profile.self.cycles-pp.update_cfs_group
0.16 ± 3% +0.2 0.34 ± 2% perf-profile.self.cycles-pp.testcase
0.76 +0.2 0.94 perf-profile.self.cycles-pp.default_send_IPI_mask_sequence_phys
0.42 ± 7% +0.2 0.65 ± 7% perf-profile.self.cycles-pp.native_irq_return_iret
0.17 ± 23% +0.2 0.42 ± 20% perf-profile.self.cycles-pp.zap_page_range_single_batched
0.20 ± 14% +0.3 0.48 ± 15% perf-profile.self.cycles-pp.flush_tlb_mm_range
0.10 ± 14% +0.3 0.44 ± 7% perf-profile.self.cycles-pp.flush_tlb_batched_pending
0.22 ± 6% +0.4 0.60 ± 6% perf-profile.self.cycles-pp.tlb_gather_mmu
0.42 ± 9% +0.6 0.98 ± 6% perf-profile.self.cycles-pp.tlb_finish_mmu
1.63 +0.6 2.25 perf-profile.self.cycles-pp.__flush_smp_call_function_queue
5.15 +8.3 13.44 perf-profile.self.cycles-pp.llist_reverse_order
36.88 +31.1 67.98 perf-profile.self.cycles-pp.smp_call_function_many_cond
***************************************************************************************************
lkp-skl-fpga01: 104 threads 2 sockets (Skylake) with 192G memory
=========================================================================================
compiler/cpufreq_governor/kconfig/mode/nr_task/rootfs/tbox_group/test/testcase:
gcc-12/performance/x86_64-rhel-9.4/thread/100%/debian-12-x86_64-20240206.cgz/lkp-skl-fpga01/tlb_flush2/will-it-scale
commit:
5e00e31867 ("userfaultfd: remove UFFD_CLOEXEC, UFFD_NONBLOCK, and UFFD_FLAGS_SET")
a6fde7add7 ("mm: use per_vma lock for MADV_DONTNEED")
5e00e31867d16e23 a6fde7add78d122f5e09cb6280f
---------------- ---------------------------
%stddev %change %stddev
\ | \
14925 -61.7% 5717 uptime.idle
1.16e+10 -81.0% 2.199e+09 ± 2% cpuidle..time
2.126e+08 -88.4% 24637059 cpuidle..usage
3245191 ± 33% -57.4% 1382593 ±116% numa-meminfo.node0.FilePages
7270056 ± 14% -26.4% 5353286 ± 28% numa-meminfo.node0.MemUsed
1170266 ± 24% -36.7% 741206 ± 24% numa-meminfo.node1.Active
1170266 ± 24% -36.7% 741206 ± 24% numa-meminfo.node1.Active(anon)
98106610 ± 5% +113.8% 2.098e+08 numa-numastat.node0.local_node
98156371 ± 5% +113.7% 2.098e+08 numa-numastat.node0.numa_hit
1.017e+08 ± 4% +109.5% 2.131e+08 numa-numastat.node1.local_node
1.017e+08 ± 4% +109.5% 2.132e+08 numa-numastat.node1.numa_hit
20547 +44.6% 29711 ± 16% perf-c2c.DRAM.remote
25526 +34.9% 34429 ± 16% perf-c2c.HITM.local
16981 +50.6% 25574 ± 16% perf-c2c.HITM.remote
42507 +41.2% 60004 ± 16% perf-c2c.HITM.total
660348 +112.4% 1402275 will-it-scale.104.threads
37.30 -83.1% 6.30 ± 2% will-it-scale.104.threads_idle
6348 +112.4% 13482 will-it-scale.per_thread_ops
660348 +112.4% 1402275 will-it-scale.workload
37.60 -31.0 6.63 ± 3% mpstat.cpu.all.idle%
0.14 -0.1 0.05 mpstat.cpu.all.soft%
41.16 +28.9 70.06 mpstat.cpu.all.sys%
2.18 +1.7 3.90 mpstat.cpu.all.usr%
65.29 +49.2% 97.40 mpstat.max_utilization_pct
37.91 -81.0% 7.19 ± 2% vmstat.cpu.id
59.91 +48.3% 88.87 vmstat.cpu.sy
61.87 ± 2% +56.4% 96.78 vmstat.procs.r
648283 -91.5% 55326 vmstat.system.cs
4410702 +75.1% 7721156 vmstat.system.in
1751586 -29.4% 1236767 ± 2% meminfo.Active
1751586 -29.4% 1236767 ± 2% meminfo.Active(anon)
4605219 -11.1% 4094953 meminfo.Cached
2806474 -18.3% 2293090 meminfo.Committed_AS
96403 -16.6% 80363 meminfo.Mapped
1065038 -47.9% 554780 ± 5% meminfo.Shmem
811327 ± 33% -57.4% 345637 ±116% numa-vmstat.node0.nr_file_pages
98156076 ± 5% +113.6% 2.097e+08 numa-vmstat.node0.numa_hit
98106315 ± 5% +113.7% 2.097e+08 numa-vmstat.node0.numa_local
292557 ± 24% -36.7% 185258 ± 24% numa-vmstat.node1.nr_active_anon
292557 ± 24% -36.7% 185258 ± 24% numa-vmstat.node1.nr_zone_active_anon
1.017e+08 ± 4% +109.4% 2.131e+08 numa-vmstat.node1.numa_hit
1.017e+08 ± 4% +109.5% 2.13e+08 numa-vmstat.node1.numa_local
437965 -29.4% 309243 ± 2% proc-vmstat.nr_active_anon
1151373 -11.1% 1023774 proc-vmstat.nr_file_pages
24519 -16.5% 20472 proc-vmstat.nr_mapped
266326 -47.9% 138730 ± 5% proc-vmstat.nr_shmem
437965 -29.4% 309243 ± 2% proc-vmstat.nr_zone_active_anon
1.998e+08 +111.6% 4.228e+08 proc-vmstat.numa_hit
1.997e+08 +111.6% 4.227e+08 proc-vmstat.numa_local
17545 ± 59% +89.3% 33209 ± 30% proc-vmstat.numa_pages_migrated
111435 ± 41% +81.2% 201867 ± 17% proc-vmstat.numa_pte_updates
1.999e+08 +111.5% 4.229e+08 proc-vmstat.pgalloc_normal
3.982e+08 +112.0% 8.443e+08 proc-vmstat.pgfault
1.996e+08 +111.8% 4.227e+08 proc-vmstat.pgfree
17545 ± 59% +89.3% 33209 ± 30% proc-vmstat.pgmigrate_success
6.15 +19.6% 7.35 perf-stat.i.MPKI
5.226e+09 +66.8% 8.718e+09 perf-stat.i.branch-instructions
2.74 +0.1 2.87 perf-stat.i.branch-miss-rate%
1.438e+08 +73.6% 2.496e+08 perf-stat.i.branch-misses
25.66 +1.6 27.25 perf-stat.i.cache-miss-rate%
1.537e+08 +99.1% 3.06e+08 perf-stat.i.cache-misses
6.048e+08 +85.8% 1.124e+09 perf-stat.i.cache-references
652603 -91.5% 55668 perf-stat.i.context-switches
7.71 -15.2% 6.54 perf-stat.i.cpi
1.923e+11 +41.5% 2.721e+11 perf-stat.i.cpu-cycles
565.76 +18.9% 672.78 perf-stat.i.cpu-migrations
1248 -28.8% 888.98 perf-stat.i.cycles-between-cache-misses
2.522e+10 +64.7% 4.152e+10 perf-stat.i.instructions
0.14 +14.3% 0.15 perf-stat.i.ipc
31.63 +70.0% 53.78 perf-stat.i.metric.K/sec
1318750 +112.1% 2796474 perf-stat.i.minor-faults
1318750 +112.1% 2796474 perf-stat.i.page-faults
6.09 +20.9% 7.37 perf-stat.overall.MPKI
2.75 +0.1 2.86 perf-stat.overall.branch-miss-rate%
25.40 +1.8 27.23 perf-stat.overall.cache-miss-rate%
7.62 -14.0% 6.55 perf-stat.overall.cpi
1251 -28.9% 889.43 perf-stat.overall.cycles-between-cache-misses
0.13 +16.3% 0.15 perf-stat.overall.ipc
11491863 -22.5% 8909088 perf-stat.overall.path-length
5.209e+09 +66.7% 8.685e+09 perf-stat.ps.branch-instructions
1.433e+08 +73.5% 2.487e+08 perf-stat.ps.branch-misses
1.532e+08 +99.0% 3.048e+08 perf-stat.ps.cache-misses
6.03e+08 +85.6% 1.12e+09 perf-stat.ps.cache-references
650365 -91.5% 55448 perf-stat.ps.context-switches
1.917e+11 +41.4% 2.711e+11 perf-stat.ps.cpu-cycles
563.80 +18.8% 669.81 perf-stat.ps.cpu-migrations
2.514e+10 +64.6% 4.137e+10 perf-stat.ps.instructions
1314147 +112.0% 2786592 perf-stat.ps.minor-faults
1314147 +112.0% 2786592 perf-stat.ps.page-faults
7.588e+12 +64.6% 1.249e+13 perf-stat.total.instructions
2697494 +219.7% 8623695 sched_debug.cfs_rq:/.avg_vruntime.avg
3062060 ± 2% +194.6% 9022225 sched_debug.cfs_rq:/.avg_vruntime.max
2522627 ± 2% +188.2% 7270351 sched_debug.cfs_rq:/.avg_vruntime.min
112005 ± 14% +229.0% 368480 sched_debug.cfs_rq:/.avg_vruntime.stddev
0.44 ± 3% +36.5% 0.59 ± 3% sched_debug.cfs_rq:/.h_nr_queued.avg
0.43 ± 3% +37.6% 0.59 ± 4% sched_debug.cfs_rq:/.h_nr_runnable.avg
2697494 +219.7% 8623696 sched_debug.cfs_rq:/.min_vruntime.avg
3062060 ± 2% +194.6% 9022225 sched_debug.cfs_rq:/.min_vruntime.max
2522627 ± 2% +188.2% 7270351 sched_debug.cfs_rq:/.min_vruntime.min
112005 ± 14% +229.0% 368480 sched_debug.cfs_rq:/.min_vruntime.stddev
0.42 ± 3% +37.8% 0.59 ± 3% sched_debug.cfs_rq:/.nr_queued.avg
400.27 +70.4% 682.01 sched_debug.cfs_rq:/.runnable_avg.avg
962.64 ± 4% +50.3% 1446 ± 8% sched_debug.cfs_rq:/.runnable_avg.max
109.90 ± 7% +50.6% 165.52 ± 6% sched_debug.cfs_rq:/.runnable_avg.stddev
400.04 +70.2% 680.87 sched_debug.cfs_rq:/.util_avg.avg
952.14 ± 4% +50.9% 1437 ± 8% sched_debug.cfs_rq:/.util_avg.max
108.76 ± 7% +50.6% 163.83 ± 6% sched_debug.cfs_rq:/.util_avg.stddev
174.98 ± 4% +153.1% 442.92 ± 4% sched_debug.cfs_rq:/.util_est.avg
674.39 ± 12% +72.8% 1165 ± 14% sched_debug.cfs_rq:/.util_est.max
199.68 +65.8% 330.97 ± 2% sched_debug.cfs_rq:/.util_est.stddev
285047 ± 2% +43.0% 407743 ± 3% sched_debug.cpu.avg_idle.avg
425414 ± 8% +136.3% 1005391 ± 2% sched_debug.cpu.avg_idle.max
90251 ± 5% -46.5% 48283 ± 9% sched_debug.cpu.avg_idle.min
60514 ± 15% +266.4% 221730 ± 6% sched_debug.cpu.avg_idle.stddev
3632 ± 14% +23.9% 4501 sched_debug.cpu.clock_task.stddev
1842 ± 4% +28.7% 2370 ± 3% sched_debug.cpu.curr->pid.avg
0.00 ± 10% +201.1% 0.00 ± 9% sched_debug.cpu.next_balance.stddev
0.43 ± 4% +37.4% 0.59 ± 3% sched_debug.cpu.nr_running.avg
930519 -91.5% 79276 sched_debug.cpu.nr_switches.avg
977386 -90.3% 95005 sched_debug.cpu.nr_switches.max
859951 ± 2% -91.2% 75446 sched_debug.cpu.nr_switches.min
12012 ± 12% -72.7% 3280 ± 5% sched_debug.cpu.nr_switches.stddev
0.44 ± 3% -38.3% 0.27 ± 8% sched_debug.cpu.nr_uninterruptible.avg
183.69 ± 23% +1850.6% 3583 sched_debug.cpu.nr_uninterruptible.max
-179.94 +1889.2% -3579 sched_debug.cpu.nr_uninterruptible.min
35.08 ± 12% +2354.3% 860.88 sched_debug.cpu.nr_uninterruptible.stddev
0.02 ± 81% +2385.8% 0.56 ± 20% perf-sched.sch_delay.avg.ms.__cond_resched.__alloc_frozen_pages_noprof.alloc_pages_mpol.vma_alloc_folio_noprof.alloc_anon_folio
0.01 ± 10% +5908.5% 0.59 ± 6% perf-sched.sch_delay.avg.ms.__cond_resched.__tlb_batch_free_encoded_pages.tlb_finish_mmu.do_madvise.__x64_sys_madvise
0.03 ± 24% +532.8% 0.19 ± 12% perf-sched.sch_delay.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
0.01 ± 31% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.down_read.madvise_lock.do_madvise.__x64_sys_madvise
0.00 ±223% +75533.3% 0.76 ±191% perf-sched.sch_delay.avg.ms.__cond_resched.remove_vma.vms_complete_munmap_vmas.do_vmi_align_munmap.do_vmi_munmap
0.04 ± 40% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.zap_page_range_single_batched.madvise_dontneed_free.madvise_vma_behavior.madvise_do_behavior
0.01 ± 14% +14584.2% 0.93 ± 73% perf-sched.sch_delay.avg.ms.__cond_resched.zap_pmd_range.isra.0.unmap_page_range
0.01 ± 72% +3310.7% 0.48 ± 51% perf-sched.sch_delay.avg.ms.__cond_resched.zap_pte_range.zap_pmd_range.isra.0
0.01 ± 27% +1037.7% 0.10 ±117% perf-sched.sch_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
0.00 ± 20% +209.1% 0.01 ± 31% perf-sched.sch_delay.avg.ms.anon_pipe_read.fifo_pipe_read.vfs_read.ksys_read
0.00 ± 14% +95.0% 0.01 ± 21% perf-sched.sch_delay.avg.ms.anon_pipe_read.vfs_read.ksys_read.do_syscall_64
0.02 ± 10% +26.7% 0.02 ± 11% perf-sched.sch_delay.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
0.02 ± 54% +1344.3% 0.26 ± 55% perf-sched.sch_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
0.01 ± 25% +165.6% 0.03 ± 39% perf-sched.sch_delay.avg.ms.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.02 ± 46% +1596.3% 0.30 ± 34% perf-sched.sch_delay.avg.ms.exit_to_user_mode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
0.01 ± 41% +203.5% 0.03 ± 24% perf-sched.sch_delay.avg.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
0.02 ± 37% +3505.2% 0.58 ± 17% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
0.02 ± 18% +3951.6% 0.63 ± 16% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
0.09 ± 50% +202.1% 0.27 ± 33% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function.[unknown]
0.01 ± 52% +2595.8% 0.22 ± 22% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown]
0.08 ± 56% +254.6% 0.27 ± 29% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
0.01 ± 21% +334.9% 0.05 ± 72% perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
0.00 ± 71% +1931.6% 0.06 ± 22% perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.poll_schedule_timeout.constprop.0.do_poll
0.01 ± 6% -100.0% 0.00 perf-sched.sch_delay.avg.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read.madvise_lock
0.01 ± 17% +456.5% 0.04 ± 10% perf-sched.sch_delay.avg.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read_killable.lock_mm_and_find_vma
0.01 +563.3% 0.03 ± 16% perf-sched.sch_delay.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write_killable.__vm_munmap
0.01 ± 15% +180.9% 0.02 ± 11% perf-sched.sch_delay.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write_killable.vm_mmap_pgoff
0.01 ± 14% +112.5% 0.02 ± 28% perf-sched.sch_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
0.00 ± 71% +12773.7% 0.41 ± 95% perf-sched.sch_delay.avg.ms.schedule_timeout.khugepaged_wait_work.khugepaged.kthread
0.01 ± 26% +198.2% 0.03 ± 24% perf-sched.sch_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
0.01 +588.1% 0.05 ±106% perf-sched.sch_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
0.03 ± 15% +77.1% 0.05 ± 10% perf-sched.sch_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
0.21 ± 97% +1950.0% 4.29 ± 8% perf-sched.sch_delay.max.ms.__cond_resched.__alloc_frozen_pages_noprof.alloc_pages_mpol.vma_alloc_folio_noprof.alloc_anon_folio
0.39 ± 58% +3818.5% 15.18 ± 49% perf-sched.sch_delay.max.ms.__cond_resched.__tlb_batch_free_encoded_pages.tlb_finish_mmu.do_madvise.__x64_sys_madvise
0.02 ±102% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.down_read.madvise_lock.do_madvise.__x64_sys_madvise
0.00 ±223% +1.3e+05% 1.29 ±140% perf-sched.sch_delay.max.ms.__cond_resched.remove_vma.vms_complete_munmap_vmas.do_vmi_align_munmap.do_vmi_munmap
0.02 ± 89% +171.0% 0.06 ± 29% perf-sched.sch_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
0.20 ± 42% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.zap_page_range_single_batched.madvise_dontneed_free.madvise_vma_behavior.madvise_do_behavior
0.01 ± 15% +47829.3% 3.28 ± 47% perf-sched.sch_delay.max.ms.__cond_resched.zap_pmd_range.isra.0.unmap_page_range
0.07 ±120% +5227.9% 3.78 ± 15% perf-sched.sch_delay.max.ms.__cond_resched.zap_pte_range.zap_pmd_range.isra.0
0.58 ±218% +3359.0% 19.97 ± 38% perf-sched.sch_delay.max.ms.__vma_start_write.__split_vma.vms_gather_munmap_vmas.do_vmi_align_munmap
1.16 ±139% +893.1% 11.53 ± 77% perf-sched.sch_delay.max.ms.__vma_start_write.vma_expand.vma_merge_new_range.__mmap_region
0.02 ± 75% +2692.5% 0.43 ±138% perf-sched.sch_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
0.23 ± 95% +383.4% 1.11 ± 44% perf-sched.sch_delay.max.ms.anon_pipe_read.fifo_pipe_read.vfs_read.ksys_read
0.36 ± 29% +199.6% 1.08 ± 35% perf-sched.sch_delay.max.ms.anon_pipe_read.vfs_read.ksys_read.do_syscall_64
0.11 ± 85% +1384.4% 1.70 ± 55% perf-sched.sch_delay.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
0.18 ±140% +2046.1% 3.87 ± 15% perf-sched.sch_delay.max.ms.exit_to_user_mode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
0.03 ± 70% +165.2% 0.07 ± 20% perf-sched.sch_delay.max.ms.irq_thread.kthread.ret_from_fork.ret_from_fork_asm
0.18 ± 52% +2490.2% 4.73 ± 6% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
0.08 ± 22% +7683.6% 5.84 ± 49% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
0.37 ± 62% +619.2% 2.65 ± 24% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function.[unknown]
0.01 ± 69% +7073.2% 0.98 ± 53% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown]
0.41 ± 50% +651.7% 3.05 ± 28% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
0.03 ± 97% +1278.5% 0.40 ±109% perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
0.27 ± 80% +878.6% 2.68 ± 44% perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.poll_schedule_timeout.constprop.0.do_poll
8.22 ± 21% -100.0% 0.00 perf-sched.sch_delay.max.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read.madvise_lock
0.20 ± 58% +12192.1% 25.12 ± 23% perf-sched.sch_delay.max.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read_killable.lock_mm_and_find_vma
4.01 ± 20% +539.5% 25.63 ± 34% perf-sched.sch_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write_killable.__vm_munmap
4.17 ± 23% +244.1% 14.35 ± 74% perf-sched.sch_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write_killable.vm_mmap_pgoff
0.00 ± 71% +12773.7% 0.41 ± 95% perf-sched.sch_delay.max.ms.schedule_timeout.khugepaged_wait_work.khugepaged.kthread
1.69 ±193% +376.3% 8.03 ± 56% perf-sched.sch_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
0.39 ± 70% +7744.0% 30.77 ±171% perf-sched.sch_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
0.01 ± 6% +721.6% 0.05 ± 7% perf-sched.total_sch_delay.average.ms
9.35 ± 15% +418.9% 48.50 ± 92% perf-sched.total_sch_delay.max.ms
1.72 +486.2% 10.09 ± 2% perf-sched.total_wait_and_delay.average.ms
933796 -87.2% 119247 perf-sched.total_wait_and_delay.count.ms
1.71 +485.4% 10.04 ± 2% perf-sched.total_wait_time.average.ms
61.83 ± 4% +127.3% 140.56 ± 12% perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range_clock.poll_schedule_timeout.constprop.0.do_poll
0.47 -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read.madvise_lock
0.21 ± 4% +529.1% 1.29 ± 3% perf-sched.wait_and_delay.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write_killable.__vm_munmap
0.17 ± 6% +516.8% 1.06 ± 5% perf-sched.wait_and_delay.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write_killable.vm_mmap_pgoff
3.75 +11.4% 4.18 perf-sched.wait_and_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
469.33 ± 5% -9.9% 422.67 ± 8% perf-sched.wait_and_delay.count.anon_pipe_read.fifo_pipe_read.vfs_read.ksys_read
2200 ± 2% -17.3% 1819 ± 7% perf-sched.wait_and_delay.count.anon_pipe_read.vfs_read.ksys_read.do_syscall_64
253.83 ± 4% -64.0% 91.50 ± 12% perf-sched.wait_and_delay.count.schedule_hrtimeout_range_clock.poll_schedule_timeout.constprop.0.do_poll
886768 -100.0% 0.00 perf-sched.wait_and_delay.count.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read.madvise_lock
5766 ± 4% +87.3% 10801 perf-sched.wait_and_delay.count.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write_killable.vm_mmap_pgoff
1330 -10.0% 1197 perf-sched.wait_and_delay.count.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
1370 ± 2% -22.4% 1062 ± 2% perf-sched.wait_and_delay.count.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
8.91 ± 11% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read.madvise_lock
6.57 ± 17% +290.0% 25.63 ± 34% perf-sched.wait_and_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write_killable.__vm_munmap
7.70 ± 4% +135.8% 18.14 ± 42% perf-sched.wait_and_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write_killable.vm_mmap_pgoff
5.94 ± 22% +184.6% 16.92 ± 62% perf-sched.wait_and_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
3519 ± 12% -45.8% 1909 ± 14% perf-sched.wait_and_delay.max.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
0.42 ± 9% +154.1% 1.07 ± 14% perf-sched.wait_time.avg.ms.__cond_resched.__alloc_frozen_pages_noprof.alloc_pages_mpol.vma_alloc_folio_noprof.alloc_anon_folio
0.44 +153.8% 1.12 ± 3% perf-sched.wait_time.avg.ms.__cond_resched.__tlb_batch_free_encoded_pages.tlb_finish_mmu.do_madvise.__x64_sys_madvise
0.45 ± 4% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.down_read.madvise_lock.do_madvise.__x64_sys_madvise
0.00 ±223% +68918.2% 1.27 ±103% perf-sched.wait_time.avg.ms.__cond_resched.remove_vma.vms_complete_munmap_vmas.do_vmi_align_munmap.do_vmi_munmap
0.49 ± 6% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.zap_page_range_single_batched.madvise_dontneed_free.madvise_vma_behavior.madvise_do_behavior
0.43 ± 11% +294.3% 1.70 ± 47% perf-sched.wait_time.avg.ms.__cond_resched.zap_pmd_range.isra.0.unmap_page_range
0.36 ± 23% +195.7% 1.08 ± 27% perf-sched.wait_time.avg.ms.__cond_resched.zap_pte_range.zap_pmd_range.isra.0
0.12 ± 17% +655.8% 0.88 ± 4% perf-sched.wait_time.avg.ms.__vma_start_write.__split_vma.vms_gather_munmap_vmas.do_vmi_align_munmap
0.13 ± 34% +528.4% 0.80 ± 8% perf-sched.wait_time.avg.ms.__vma_start_write.vma_expand.vma_merge_new_range.__mmap_region
4.18 ±201% +1368.5% 61.40 ± 55% perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
0.43 ± 4% +332.3% 1.85 ± 98% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
0.34 ± 15% +171.8% 0.94 ± 11% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
0.29 ± 32% +58.8% 0.46 ± 31% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
61.82 ± 4% +127.2% 140.49 ± 12% perf-sched.wait_time.avg.ms.schedule_hrtimeout_range_clock.poll_schedule_timeout.constprop.0.do_poll
0.47 -100.0% 0.00 perf-sched.wait_time.avg.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read.madvise_lock
0.38 ± 4% +106.3% 0.79 ± 6% perf-sched.wait_time.avg.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read_killable.lock_mm_and_find_vma
0.20 ± 4% +528.3% 1.26 ± 4% perf-sched.wait_time.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write_killable.__vm_munmap
0.16 ± 7% +532.3% 1.04 ± 5% perf-sched.wait_time.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write_killable.vm_mmap_pgoff
1.02 ± 2% +17.3% 1.20 perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
3.74 +11.0% 4.15 perf-sched.wait_time.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
0.78 ± 13% +596.9% 5.41 ± 17% perf-sched.wait_time.max.ms.__cond_resched.__alloc_frozen_pages_noprof.alloc_pages_mpol.vma_alloc_folio_noprof.alloc_anon_folio
2.03 ± 62% +451.4% 11.19 ± 45% perf-sched.wait_time.max.ms.__cond_resched.__tlb_batch_free_encoded_pages.tlb_finish_mmu.do_madvise.__x64_sys_madvise
0.58 ± 20% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.down_read.madvise_lock.do_madvise.__x64_sys_madvise
0.00 ±223% +1.4e+05% 2.50 ± 62% perf-sched.wait_time.max.ms.__cond_resched.remove_vma.vms_complete_munmap_vmas.do_vmi_align_munmap.do_vmi_munmap
0.79 ± 20% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.zap_page_range_single_batched.madvise_dontneed_free.madvise_vma_behavior.madvise_do_behavior
0.54 ± 22% +738.7% 4.51 ± 19% perf-sched.wait_time.max.ms.__cond_resched.zap_pmd_range.isra.0.unmap_page_range
0.55 ± 17% +870.1% 5.33 ± 20% perf-sched.wait_time.max.ms.__cond_resched.zap_pte_range.zap_pmd_range.isra.0
1.63 ± 71% +626.3% 11.81 ± 20% perf-sched.wait_time.max.ms.__vma_start_write.__split_vma.vms_gather_munmap_vmas.do_vmi_align_munmap
1.92 ± 76% +485.1% 11.26 ± 15% perf-sched.wait_time.max.ms.__vma_start_write.vma_expand.vma_merge_new_range.__mmap_region
38.18 +13.8% 43.45 ± 3% perf-sched.wait_time.max.ms.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe
169.52 ±219% +392.5% 834.84 ± 44% perf-sched.wait_time.max.ms.exit_to_user_mode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
1.04 ± 41% +16368.0% 171.38 ±216% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_exc_page_fault.[unknown]
0.81 ± 10% +674.5% 6.31 ± 13% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
0.69 ± 17% +466.7% 3.90 ± 41% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function.[unknown]
0.75 ± 20% +378.5% 3.57 ± 24% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
4.51 ± 2% -100.0% 0.00 perf-sched.wait_time.max.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read.madvise_lock
1.89 ± 66% +602.4% 13.28 ± 15% perf-sched.wait_time.max.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read_killable.lock_mm_and_find_vma
4.53 ± 3% +200.4% 13.61 ± 20% perf-sched.wait_time.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write_killable.__vm_munmap
4.36 ± 2% +187.3% 12.53 ± 14% perf-sched.wait_time.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write_killable.vm_mmap_pgoff
3.75 ± 3% +34.3% 5.03 ± 7% perf-sched.wait_time.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
5.17 +106.1% 10.66 ± 49% perf-sched.wait_time.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
3519 ± 12% -45.8% 1909 ± 14% perf-sched.wait_time.max.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
30.63 -27.7 2.97 ± 2% perf-profile.calltrace.cycles-pp.common_startup_64
30.33 -27.4 2.94 ± 2% perf-profile.calltrace.cycles-pp.start_secondary.common_startup_64
30.32 -27.4 2.94 ± 2% perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.common_startup_64
30.31 -27.4 2.94 ± 2% perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.common_startup_64
27.76 -25.0 2.78 ± 2% perf-profile.calltrace.cycles-pp.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary.common_startup_64
26.94 -24.2 2.75 ± 2% perf-profile.calltrace.cycles-pp.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry.start_secondary
26.80 -24.1 2.74 ± 2% perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle.cpu_startup_entry
11.17 ± 2% -11.2 0.00 perf-profile.calltrace.cycles-pp.asm_sysvec_call_function.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
11.04 ± 2% -11.0 0.00 perf-profile.calltrace.cycles-pp.sysvec_call_function.asm_sysvec_call_function.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call
10.90 ± 2% -10.9 0.00 perf-profile.calltrace.cycles-pp.__sysvec_call_function.sysvec_call_function.asm_sysvec_call_function.cpuidle_enter_state.cpuidle_enter
10.86 ± 2% -10.9 0.00 perf-profile.calltrace.cycles-pp.__flush_smp_call_function_queue.__sysvec_call_function.sysvec_call_function.asm_sysvec_call_function.cpuidle_enter_state
13.30 ± 2% -9.2 4.10 perf-profile.calltrace.cycles-pp.flush_tlb_func.__flush_smp_call_function_queue.__sysvec_call_function.sysvec_call_function.asm_sysvec_call_function
9.07 -9.0 0.08 ±223% perf-profile.calltrace.cycles-pp.intel_idle.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
8.80 ± 3% -8.8 0.00 perf-profile.calltrace.cycles-pp.madvise_lock.do_madvise.__x64_sys_madvise.do_syscall_64.entry_SYSCALL_64_after_hwframe
8.78 ± 3% -8.8 0.00 perf-profile.calltrace.cycles-pp.down_read.madvise_lock.do_madvise.__x64_sys_madvise.do_syscall_64
8.12 ± 3% -8.1 0.00 perf-profile.calltrace.cycles-pp.rwsem_down_read_slowpath.down_read.madvise_lock.do_madvise.__x64_sys_madvise
5.85 ± 4% -5.9 0.00 perf-profile.calltrace.cycles-pp._raw_spin_lock_irq.rwsem_down_read_slowpath.down_read.madvise_lock.do_madvise
5.72 ± 5% -5.7 0.00 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irq.rwsem_down_read_slowpath.down_read.madvise_lock
5.27 ± 2% -4.7 0.59 perf-profile.calltrace.cycles-pp.asm_sysvec_call_function.intel_idle_irq.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call
4.13 ± 3% -2.7 1.40 ± 2% perf-profile.calltrace.cycles-pp.intel_idle_irq.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call.do_idle
1.27 ± 5% -0.2 1.11 ± 6% perf-profile.calltrace.cycles-pp.lock_vma_under_rcu.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.testcase
0.59 ± 4% +0.3 0.84 ± 3% perf-profile.calltrace.cycles-pp.alloc_anon_folio.do_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
0.64 ± 6% +0.4 1.02 ± 9% perf-profile.calltrace.cycles-pp.zap_pte_range.zap_pmd_range.unmap_page_range.zap_page_range_single_batched.madvise_dontneed_free
0.66 ± 5% +0.4 1.06 ± 8% perf-profile.calltrace.cycles-pp.zap_pmd_range.unmap_page_range.zap_page_range_single_batched.madvise_dontneed_free.madvise_vma_behavior
0.00 +0.5 0.52 perf-profile.calltrace.cycles-pp.error_entry.testcase
0.00 +0.5 0.52 ± 2% perf-profile.calltrace.cycles-pp.sysvec_call_function.asm_sysvec_call_function.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
0.00 +0.5 0.53 perf-profile.calltrace.cycles-pp.alloc_pages_mpol.vma_alloc_folio_noprof.alloc_anon_folio.do_anonymous_page.__handle_mm_fault
0.00 +0.6 0.55 perf-profile.calltrace.cycles-pp.asm_sysvec_call_function.default_send_IPI_mask_sequence_phys.smp_call_function_many_cond.on_each_cpu_cond_mask.flush_tlb_mm_range
0.00 +0.6 0.55 perf-profile.calltrace.cycles-pp.__flush_smp_call_function_queue.__sysvec_call_function.sysvec_call_function.asm_sysvec_call_function.testcase
0.00 +0.6 0.56 perf-profile.calltrace.cycles-pp.__sysvec_call_function.sysvec_call_function.asm_sysvec_call_function.testcase
0.00 +0.6 0.56 ± 2% perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.smp_call_function_many_cond.on_each_cpu_cond_mask.flush_tlb_mm_range
0.00 +0.6 0.57 perf-profile.calltrace.cycles-pp._raw_spin_lock.__pte_offset_map_lock.do_anonymous_page.__handle_mm_fault.handle_mm_fault
0.00 +0.6 0.57 perf-profile.calltrace.cycles-pp.asm_sysvec_call_function.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.testcase
0.00 +0.6 0.58 perf-profile.calltrace.cycles-pp.sysvec_call_function.asm_sysvec_call_function.testcase
0.00 +0.6 0.59 ± 2% perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.smp_call_function_many_cond.on_each_cpu_cond_mask.flush_tlb_mm_range.tlb_finish_mmu
1.81 +0.6 2.40 perf-profile.calltrace.cycles-pp.default_send_IPI_mask_sequence_phys.smp_call_function_many_cond.on_each_cpu_cond_mask.flush_tlb_mm_range.tlb_finish_mmu
0.00 +0.6 0.60 ± 4% perf-profile.calltrace.cycles-pp.tlb_gather_mmu.do_madvise.__x64_sys_madvise.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +0.6 0.61 perf-profile.calltrace.cycles-pp.vma_alloc_folio_noprof.alloc_anon_folio.do_anonymous_page.__handle_mm_fault.handle_mm_fault
0.00 +0.6 0.64 perf-profile.calltrace.cycles-pp.__pte_offset_map_lock.do_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
0.00 +0.7 0.66 perf-profile.calltrace.cycles-pp.asm_sysvec_call_function.testcase
0.00 +0.7 0.68 perf-profile.calltrace.cycles-pp.folio_batch_move_lru.folio_add_lru.do_anonymous_page.__handle_mm_fault.handle_mm_fault
1.28 +0.7 1.99 perf-profile.calltrace.cycles-pp.__irqentry_text_end.testcase
0.00 +0.8 0.78 perf-profile.calltrace.cycles-pp.folio_add_lru.do_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
0.00 +0.8 0.82 perf-profile.calltrace.cycles-pp._find_next_bit.smp_call_function_many_cond.on_each_cpu_cond_mask.flush_tlb_mm_range.tlb_finish_mmu
1.75 ± 5% +0.8 2.58 ± 7% perf-profile.calltrace.cycles-pp.madvise_do_behavior.do_madvise.__x64_sys_madvise.do_syscall_64.entry_SYSCALL_64_after_hwframe
2.00 ± 3% +1.0 3.05 ± 4% perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
1.68 ± 4% +1.0 2.73 ± 4% perf-profile.calltrace.cycles-pp.do_anonymous_page.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
0.00 +1.2 1.16 ± 8% perf-profile.calltrace.cycles-pp.unmap_page_range.zap_page_range_single_batched.madvise_dontneed_free.madvise_vma_behavior.madvise_walk_vmas
2.24 ± 4% +1.2 3.42 ± 4% perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.testcase
4.14 ± 3% +1.6 5.73 ± 4% perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.testcase
4.19 ± 3% +1.6 5.80 ± 4% perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.testcase
0.00 +1.7 1.72 ± 8% perf-profile.calltrace.cycles-pp.zap_page_range_single_batched.madvise_dontneed_free.madvise_vma_behavior.madvise_walk_vmas.madvise_do_behavior
0.00 +1.8 1.79 ± 8% perf-profile.calltrace.cycles-pp.madvise_dontneed_free.madvise_vma_behavior.madvise_walk_vmas.madvise_do_behavior.do_madvise
0.00 +1.8 1.81 ± 7% perf-profile.calltrace.cycles-pp.madvise_vma_behavior.madvise_walk_vmas.madvise_do_behavior.do_madvise.__x64_sys_madvise
4.60 ± 3% +1.9 6.47 ± 3% perf-profile.calltrace.cycles-pp.asm_exc_page_fault.testcase
0.00 +2.5 2.53 ± 7% perf-profile.calltrace.cycles-pp.madvise_walk_vmas.madvise_do_behavior.do_madvise.__x64_sys_madvise.do_syscall_64
6.85 ± 2% +3.2 10.04 ± 2% perf-profile.calltrace.cycles-pp.testcase
4.33 +4.0 8.32 perf-profile.calltrace.cycles-pp.llist_reverse_order.__flush_smp_call_function_queue.__sysvec_call_function.sysvec_call_function.asm_sysvec_call_function
6.67 +8.0 14.63 perf-profile.calltrace.cycles-pp.__flush_smp_call_function_queue.__sysvec_call_function.sysvec_call_function.asm_sysvec_call_function.smp_call_function_many_cond
6.78 +8.0 14.79 perf-profile.calltrace.cycles-pp.__sysvec_call_function.sysvec_call_function.asm_sysvec_call_function.smp_call_function_many_cond.on_each_cpu_cond_mask
7.08 +8.2 15.29 perf-profile.calltrace.cycles-pp.sysvec_call_function.asm_sysvec_call_function.smp_call_function_many_cond.on_each_cpu_cond_mask.flush_tlb_mm_range
8.54 +9.2 17.76 perf-profile.calltrace.cycles-pp.asm_sysvec_call_function.smp_call_function_many_cond.on_each_cpu_cond_mask.flush_tlb_mm_range.tlb_finish_mmu
59.46 +23.9 83.35 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.__madvise
59.18 +23.9 83.10 perf-profile.calltrace.cycles-pp.do_madvise.__x64_sys_madvise.do_syscall_64.entry_SYSCALL_64_after_hwframe.__madvise
59.20 +23.9 83.12 perf-profile.calltrace.cycles-pp.__x64_sys_madvise.do_syscall_64.entry_SYSCALL_64_after_hwframe.__madvise
59.65 +24.0 83.66 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.__madvise
61.21 +24.7 85.94 perf-profile.calltrace.cycles-pp.__madvise
46.20 +31.8 77.95 perf-profile.calltrace.cycles-pp.on_each_cpu_cond_mask.flush_tlb_mm_range.tlb_finish_mmu.do_madvise.__x64_sys_madvise
45.45 +31.8 77.24 perf-profile.calltrace.cycles-pp.smp_call_function_many_cond.on_each_cpu_cond_mask.flush_tlb_mm_range.tlb_finish_mmu.do_madvise
46.86 +31.8 78.69 perf-profile.calltrace.cycles-pp.flush_tlb_mm_range.tlb_finish_mmu.do_madvise.__x64_sys_madvise.do_syscall_64
47.79 +32.0 79.84 perf-profile.calltrace.cycles-pp.tlb_finish_mmu.do_madvise.__x64_sys_madvise.do_syscall_64.entry_SYSCALL_64_after_hwframe
30.63 -27.7 2.97 ± 2% perf-profile.children.cycles-pp.common_startup_64
30.63 -27.7 2.97 ± 2% perf-profile.children.cycles-pp.cpu_startup_entry
30.62 -27.7 2.97 ± 2% perf-profile.children.cycles-pp.do_idle
30.33 -27.4 2.94 ± 2% perf-profile.children.cycles-pp.start_secondary
28.04 -25.2 2.80 ± 2% perf-profile.children.cycles-pp.cpuidle_idle_call
27.21 -24.4 2.78 ± 2% perf-profile.children.cycles-pp.cpuidle_enter
27.20 -24.4 2.78 ± 2% perf-profile.children.cycles-pp.cpuidle_enter_state
15.16 ± 2% -9.6 5.60 perf-profile.children.cycles-pp.flush_tlb_func
8.80 ± 3% -8.8 0.00 perf-profile.children.cycles-pp.madvise_lock
8.79 ± 3% -8.8 0.01 ±223% perf-profile.children.cycles-pp.down_read
9.18 -8.7 0.48 ± 3% perf-profile.children.cycles-pp.intel_idle
8.17 ± 3% -8.0 0.12 ± 4% perf-profile.children.cycles-pp.rwsem_down_read_slowpath
5.91 ± 4% -5.9 0.00 perf-profile.children.cycles-pp._raw_spin_lock_irq
5.98 ± 4% -5.5 0.43 ± 2% perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
24.46 -5.3 19.20 perf-profile.children.cycles-pp.__flush_smp_call_function_queue
26.95 -5.2 21.74 perf-profile.children.cycles-pp.asm_sysvec_call_function
23.53 -4.6 18.93 perf-profile.children.cycles-pp.__sysvec_call_function
24.21 -4.6 19.65 perf-profile.children.cycles-pp.sysvec_call_function
5.83 ± 2% -3.6 2.21 ± 2% perf-profile.children.cycles-pp.intel_idle_irq
2.93 ± 2% -2.7 0.24 ± 3% perf-profile.children.cycles-pp.__schedule
1.56 ± 3% -1.5 0.10 perf-profile.children.cycles-pp.schedule_idle
1.40 ± 2% -1.3 0.12 ± 4% perf-profile.children.cycles-pp.schedule_preempt_disabled
1.39 ± 2% -1.2 0.14 ± 2% perf-profile.children.cycles-pp.schedule
0.83 ± 3% -0.8 0.05 perf-profile.children.cycles-pp.flush_smp_call_function_queue
0.67 ± 3% -0.6 0.09 ± 4% perf-profile.children.cycles-pp.__pick_next_task
0.62 ± 3% -0.5 0.08 ± 5% perf-profile.children.cycles-pp.pick_next_task_fair
0.53 ± 3% -0.5 0.07 ± 5% perf-profile.children.cycles-pp.finish_task_switch
0.47 ± 5% -0.4 0.08 ± 6% perf-profile.children.cycles-pp.sched_balance_newidle
0.46 ± 5% -0.4 0.09 ± 5% perf-profile.children.cycles-pp.sched_balance_rq
0.38 ± 4% -0.3 0.07 perf-profile.children.cycles-pp.sched_balance_find_src_group
0.38 ± 4% -0.3 0.07 perf-profile.children.cycles-pp.update_sd_lb_stats
0.35 ± 5% -0.3 0.06 perf-profile.children.cycles-pp.update_sg_lb_stats
0.31 -0.2 0.09 perf-profile.children.cycles-pp.vm_mmap_pgoff
0.31 -0.2 0.10 perf-profile.children.cycles-pp.__mmap
0.25 -0.2 0.08 ± 7% perf-profile.children.cycles-pp.handle_softirqs
0.30 ± 3% -0.2 0.14 ± 3% perf-profile.children.cycles-pp._raw_spin_lock_irqsave
0.60 -0.1 0.47 perf-profile.children.cycles-pp.__vm_munmap
0.60 -0.1 0.47 perf-profile.children.cycles-pp.__x64_sys_munmap
0.61 -0.1 0.48 perf-profile.children.cycles-pp.__munmap
0.47 -0.1 0.36 perf-profile.children.cycles-pp.vms_complete_munmap_vmas
0.20 -0.1 0.10 ± 4% perf-profile.children.cycles-pp.down_write_killable
0.19 ± 2% -0.1 0.10 ± 5% perf-profile.children.cycles-pp.rwsem_down_write_slowpath
0.18 ± 2% -0.1 0.09 ± 4% perf-profile.children.cycles-pp.update_load_avg
0.49 -0.1 0.40 perf-profile.children.cycles-pp.do_vmi_align_munmap
0.56 -0.1 0.47 perf-profile.children.cycles-pp.__irq_exit_rcu
0.49 -0.1 0.40 perf-profile.children.cycles-pp.do_vmi_munmap
0.20 ± 2% -0.0 0.19 ± 2% perf-profile.children.cycles-pp.native_sched_clock
0.08 ± 5% -0.0 0.07 ± 5% perf-profile.children.cycles-pp.irqentry_enter
0.06 +0.0 0.07 perf-profile.children.cycles-pp.native_apic_mem_eoi
0.23 ± 2% +0.0 0.25 perf-profile.children.cycles-pp.sched_clock_cpu
0.05 ± 7% +0.0 0.09 perf-profile.children.cycles-pp.free_unref_folios
0.09 ± 5% +0.0 0.13 ± 3% perf-profile.children.cycles-pp.rmqueue
0.07 ± 6% +0.0 0.12 ± 4% perf-profile.children.cycles-pp.mod_memcg_lruvec_state
0.02 ± 99% +0.0 0.07 ± 7% perf-profile.children.cycles-pp.tmigr_requires_handle_remote
0.06 ± 14% +0.0 0.11 ± 11% perf-profile.children.cycles-pp.__folio_mod_stat
0.06 ± 9% +0.0 0.10 ± 3% perf-profile.children.cycles-pp.lru_gen_del_folio
0.06 ± 8% +0.0 0.10 perf-profile.children.cycles-pp.lru_gen_add_folio
0.08 ± 14% +0.0 0.12 ± 8% perf-profile.children.cycles-pp.count_memcg_events
0.12 ± 11% +0.1 0.17 ± 9% perf-profile.children.cycles-pp.__lruvec_stat_mod_folio
0.00 +0.1 0.05 perf-profile.children.cycles-pp.___pte_offset_map
0.00 +0.1 0.05 perf-profile.children.cycles-pp._raw_spin_trylock
0.20 ± 2% +0.1 0.25 ± 2% perf-profile.children.cycles-pp.zap_present_ptes
0.17 ± 4% +0.1 0.22 ± 2% perf-profile.children.cycles-pp.___perf_sw_event
0.01 ±223% +0.1 0.06 perf-profile.children.cycles-pp.irq_exit_rcu
0.01 ±223% +0.1 0.06 ± 6% perf-profile.children.cycles-pp.get_jiffies_update
0.75 +0.1 0.81 perf-profile.children.cycles-pp._raw_spin_lock
0.00 +0.1 0.06 ± 6% perf-profile.children.cycles-pp.__rmqueue_pcplist
0.00 +0.1 0.06 ± 11% perf-profile.children.cycles-pp.try_charge_memcg
0.10 ± 7% +0.1 0.16 ± 8% perf-profile.children.cycles-pp.folio_add_new_anon_rmap
0.00 +0.1 0.06 perf-profile.children.cycles-pp.__mem_cgroup_uncharge_folios
0.00 +0.1 0.06 perf-profile.children.cycles-pp.get_vma_policy
0.01 ±223% +0.1 0.07 ± 10% perf-profile.children.cycles-pp.charge_memcg
0.71 +0.1 0.77 perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
0.13 ± 3% +0.1 0.19 ± 3% perf-profile.children.cycles-pp.free_pages_and_swap_cache
0.07 ± 5% +0.1 0.13 ± 2% perf-profile.children.cycles-pp.folio_lruvec_lock_irqsave
0.01 ±223% +0.1 0.07 ± 6% perf-profile.children.cycles-pp.handle_pte_fault
0.09 ± 4% +0.1 0.16 ± 3% perf-profile.children.cycles-pp.lru_add
0.12 ± 3% +0.1 0.19 perf-profile.children.cycles-pp.entry_SYSCALL_64
0.00 +0.1 0.07 perf-profile.children.cycles-pp.unmap_vmas
0.15 ± 3% +0.1 0.23 ± 2% perf-profile.children.cycles-pp.clear_page_erms
0.14 ± 3% +0.1 0.22 ± 3% perf-profile.children.cycles-pp.__tlb_batch_free_encoded_pages
0.30 +0.1 0.37 perf-profile.children.cycles-pp.irqtime_account_irq
0.00 +0.1 0.08 ± 4% perf-profile.children.cycles-pp.down_read_killable
0.00 +0.1 0.09 ± 4% perf-profile.children.cycles-pp.lock_mm_and_find_vma
0.12 ± 4% +0.1 0.21 perf-profile.children.cycles-pp.__page_cache_release
0.17 ± 2% +0.1 0.27 ± 3% perf-profile.children.cycles-pp.__perf_sw_event
0.00 +0.1 0.11 ± 6% perf-profile.children.cycles-pp.update_cfs_group
0.26 +0.1 0.39 perf-profile.children.cycles-pp.get_page_from_freelist
0.22 ± 2% +0.1 0.36 perf-profile.children.cycles-pp.sync_regs
0.13 ± 3% +0.1 0.26 ± 2% perf-profile.children.cycles-pp.mm_needs_global_asid
0.32 +0.2 0.48 perf-profile.children.cycles-pp.syscall_return_via_sysret
0.31 ± 2% +0.2 0.48 perf-profile.children.cycles-pp.entry_SYSRETQ_unsafe_stack
0.34 +0.2 0.50 perf-profile.children.cycles-pp.__alloc_frozen_pages_noprof
0.16 ± 2% +0.2 0.32 perf-profile.children.cycles-pp.vms_clear_ptes
0.37 +0.2 0.54 perf-profile.children.cycles-pp.native_flush_tlb_local
0.36 ± 2% +0.2 0.54 perf-profile.children.cycles-pp.alloc_pages_mpol
0.05 +0.2 0.24 ± 4% perf-profile.children.cycles-pp.task_tick_fair
0.12 ± 3% +0.2 0.31 ± 3% perf-profile.children.cycles-pp.sched_tick
0.41 +0.2 0.61 perf-profile.children.cycles-pp.vma_alloc_folio_noprof
1.01 +0.2 1.21 perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
0.77 ± 2% +0.2 0.98 perf-profile.children.cycles-pp._find_next_bit
0.41 ± 5% +0.2 0.62 ± 4% perf-profile.children.cycles-pp.tlb_gather_mmu
0.30 ± 3% +0.2 0.51 perf-profile.children.cycles-pp.folios_put_refs
0.19 ± 11% +0.2 0.40 ± 5% perf-profile.children.cycles-pp.flush_tlb_batched_pending
0.22 ± 3% +0.2 0.46 ± 2% perf-profile.children.cycles-pp.update_process_times
0.35 ± 3% +0.2 0.58 ± 2% perf-profile.children.cycles-pp.__hrtimer_run_queues
0.26 ± 4% +0.2 0.50 ± 2% perf-profile.children.cycles-pp.tick_nohz_handler
0.42 ± 2% +0.2 0.66 perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
0.42 ± 2% +0.2 0.66 perf-profile.children.cycles-pp.hrtimer_interrupt
0.59 ± 4% +0.3 0.85 ± 2% perf-profile.children.cycles-pp.alloc_anon_folio
0.37 +0.3 0.68 perf-profile.children.cycles-pp.folio_batch_move_lru
0.77 +0.3 1.08 perf-profile.children.cycles-pp.native_irq_return_iret
0.53 +0.3 0.87 ± 2% perf-profile.children.cycles-pp.__pte_offset_map_lock
1.29 ± 5% +0.3 1.63 ± 7% perf-profile.children.cycles-pp.lock_vma_under_rcu
0.43 ± 2% +0.4 0.79 perf-profile.children.cycles-pp.folio_add_lru
0.74 +0.4 1.12 perf-profile.children.cycles-pp.error_entry
0.68 ± 5% +0.4 1.08 ± 9% perf-profile.children.cycles-pp.zap_pte_range
0.82 ± 4% +0.4 1.23 ± 7% perf-profile.children.cycles-pp.unmap_page_range
0.70 ± 5% +0.4 1.12 ± 7% perf-profile.children.cycles-pp.zap_pmd_range
1.24 ± 7% +0.6 1.82 ± 7% perf-profile.children.cycles-pp.madvise_vma_behavior
1.22 ± 7% +0.6 1.79 ± 8% perf-profile.children.cycles-pp.madvise_dontneed_free
1.88 +0.6 2.48 perf-profile.children.cycles-pp.default_send_IPI_mask_sequence_phys
1.10 ± 7% +0.6 1.73 ± 8% perf-profile.children.cycles-pp.zap_page_range_single_batched
1.34 +0.7 2.06 perf-profile.children.cycles-pp.__irqentry_text_end
1.75 ± 5% +0.8 2.58 ± 7% perf-profile.children.cycles-pp.madvise_do_behavior
2.02 ± 3% +1.0 3.06 ± 4% perf-profile.children.cycles-pp.__handle_mm_fault
1.68 ± 4% +1.1 2.74 ± 4% perf-profile.children.cycles-pp.do_anonymous_page
2.26 ± 4% +1.2 3.44 ± 4% perf-profile.children.cycles-pp.handle_mm_fault
4.18 ± 3% +1.6 5.78 ± 4% perf-profile.children.cycles-pp.do_user_addr_fault
4.20 ± 3% +1.6 5.81 ± 4% perf-profile.children.cycles-pp.exc_page_fault
4.90 ± 3% +2.0 6.95 ± 3% perf-profile.children.cycles-pp.asm_exc_page_fault
0.00 +2.5 2.54 ± 7% perf-profile.children.cycles-pp.madvise_walk_vmas
7.33 ± 2% +3.5 10.82 ± 2% perf-profile.children.cycles-pp.testcase
6.34 +4.5 10.81 perf-profile.children.cycles-pp.llist_reverse_order
60.44 +23.5 83.96 perf-profile.children.cycles-pp.do_syscall_64
60.63 +23.6 84.27 perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
59.18 +23.9 83.11 perf-profile.children.cycles-pp.do_madvise
59.20 +23.9 83.12 perf-profile.children.cycles-pp.__x64_sys_madvise
61.26 +24.8 86.01 perf-profile.children.cycles-pp.__madvise
46.32 +31.9 78.20 perf-profile.children.cycles-pp.smp_call_function_many_cond
46.32 +31.9 78.20 perf-profile.children.cycles-pp.on_each_cpu_cond_mask
46.99 +32.0 78.95 perf-profile.children.cycles-pp.flush_tlb_mm_range
47.93 +32.2 80.11 perf-profile.children.cycles-pp.tlb_finish_mmu
14.64 ± 2% -9.9 4.79 perf-profile.self.cycles-pp.flush_tlb_func
9.18 -8.7 0.48 ± 3% perf-profile.self.cycles-pp.intel_idle
5.95 ± 4% -5.6 0.37 ± 2% perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
3.14 ± 3% -1.2 1.96 ± 2% perf-profile.self.cycles-pp.intel_idle_irq
0.25 ± 5% -0.2 0.05 ± 7% perf-profile.self.cycles-pp.update_sg_lb_stats
0.55 ± 2% -0.1 0.41 perf-profile.self.cycles-pp._raw_spin_lock
0.42 ± 7% -0.1 0.36 ± 13% perf-profile.self.cycles-pp.mas_walk
0.23 ± 7% -0.0 0.19 ± 14% perf-profile.self.cycles-pp.__handle_mm_fault
0.20 ± 2% -0.0 0.18 ± 2% perf-profile.self.cycles-pp.native_sched_clock
0.08 -0.0 0.07 perf-profile.self.cycles-pp.sched_core_idle_cpu
0.06 +0.0 0.07 perf-profile.self.cycles-pp.do_syscall_64
0.05 +0.0 0.06 ± 6% perf-profile.self.cycles-pp.irqentry_enter
0.05 ± 8% +0.0 0.07 ± 7% perf-profile.self.cycles-pp.native_apic_mem_eoi
0.07 ± 7% +0.0 0.08 ± 4% perf-profile.self.cycles-pp.folios_put_refs
0.06 ± 8% +0.0 0.07 ± 6% perf-profile.self.cycles-pp.mod_memcg_lruvec_state
0.07 +0.0 0.09 perf-profile.self.cycles-pp.zap_present_ptes
0.10 ± 4% +0.0 0.12 ± 3% perf-profile.self.cycles-pp.asm_sysvec_call_function
0.06 ± 6% +0.0 0.08 perf-profile.self.cycles-pp.__irq_exit_rcu
0.12 ± 3% +0.0 0.15 ± 2% perf-profile.self.cycles-pp.sysvec_call_function
0.12 ± 4% +0.0 0.15 ± 4% perf-profile.self.cycles-pp.___perf_sw_event
0.05 +0.0 0.08 perf-profile.self.cycles-pp.__alloc_frozen_pages_noprof
0.05 ± 7% +0.0 0.08 ± 5% perf-profile.self.cycles-pp.folio_add_lru
0.10 ± 3% +0.0 0.14 ± 3% perf-profile.self.cycles-pp.irqtime_account_irq
0.05 ± 45% +0.0 0.08 ± 5% perf-profile.self.cycles-pp.count_memcg_events
0.08 ± 6% +0.0 0.12 perf-profile.self.cycles-pp.__madvise
0.00 +0.1 0.05 perf-profile.self.cycles-pp.__pte_offset_map_lock
0.00 +0.1 0.05 perf-profile.self.cycles-pp.asm_exc_page_fault
0.00 +0.1 0.05 perf-profile.self.cycles-pp.get_vma_policy
0.01 ±223% +0.1 0.06 ± 6% perf-profile.self.cycles-pp.get_jiffies_update
0.00 +0.1 0.05 ± 8% perf-profile.self.cycles-pp.lru_add
0.00 +0.1 0.05 ± 8% perf-profile.self.cycles-pp.zap_pte_range
0.13 ± 2% +0.1 0.18 ± 3% perf-profile.self.cycles-pp.clear_page_erms
0.11 ± 3% +0.1 0.17 perf-profile.self.cycles-pp.entry_SYSCALL_64
0.00 +0.1 0.07 ± 5% perf-profile.self.cycles-pp.lru_gen_del_folio
0.00 +0.1 0.07 perf-profile.self.cycles-pp.lru_gen_add_folio
0.64 ± 3% +0.1 0.74 perf-profile.self.cycles-pp._find_next_bit
0.10 ± 4% +0.1 0.20 ± 3% perf-profile.self.cycles-pp.mm_needs_global_asid
0.66 ± 7% +0.1 0.76 ± 4% perf-profile.self.cycles-pp.tlb_finish_mmu
0.00 +0.1 0.11 ± 6% perf-profile.self.cycles-pp.update_cfs_group
0.20 ± 3% +0.1 0.32 perf-profile.self.cycles-pp.entry_SYSCALL_64_after_hwframe
0.00 +0.1 0.13 ± 3% perf-profile.self.cycles-pp.madvise_walk_vmas
0.22 ± 2% +0.1 0.35 ± 2% perf-profile.self.cycles-pp.sync_regs
0.34 ± 5% +0.2 0.50 ± 4% perf-profile.self.cycles-pp.tlb_gather_mmu
0.80 ± 6% +0.2 0.96 ± 5% perf-profile.self.cycles-pp.lock_vma_under_rcu
0.31 ± 2% +0.2 0.47 perf-profile.self.cycles-pp.entry_SYSRETQ_unsafe_stack
0.19 ± 2% +0.2 0.35 perf-profile.self.cycles-pp.do_user_addr_fault
0.16 ± 10% +0.2 0.32 ± 6% perf-profile.self.cycles-pp.flush_tlb_batched_pending
0.32 +0.2 0.48 perf-profile.self.cycles-pp.syscall_return_via_sysret
0.36 +0.2 0.53 perf-profile.self.cycles-pp.native_flush_tlb_local
0.27 ± 16% +0.2 0.45 ± 9% perf-profile.self.cycles-pp.zap_page_range_single_batched
0.32 ± 4% +0.2 0.53 ± 7% perf-profile.self.cycles-pp.flush_tlb_mm_range
0.77 +0.3 1.08 perf-profile.self.cycles-pp.native_irq_return_iret
0.54 +0.3 0.88 perf-profile.self.cycles-pp.testcase
0.73 +0.4 1.11 perf-profile.self.cycles-pp.error_entry
1.49 +0.4 1.90 perf-profile.self.cycles-pp.default_send_IPI_mask_sequence_phys
2.32 +0.6 2.93 perf-profile.self.cycles-pp.__flush_smp_call_function_queue
1.34 +0.7 2.06 perf-profile.self.cycles-pp.__irqentry_text_end
6.26 +4.5 10.76 perf-profile.self.cycles-pp.llist_reverse_order
35.86 +21.9 57.76 perf-profile.self.cycles-pp.smp_call_function_many_cond
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
--
0-DAY CI Kernel Test Service
https://github.com/intel/lkp-tests/wiki
^ permalink raw reply [flat|nested] only message in thread
only message in thread, other threads:[~2025-07-23 6:13 UTC | newest]
Thread overview: (only message) (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2025-07-23 6:12 [linux-next:master] [mm] a6fde7add7: will-it-scale.per_thread_ops 176.7% improvement kernel test robot
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).