* [vbabka:slub-percpu-sheaves-v6r1] [maple_tree] 46bfd29ca2: stress-ng.pagemove.page_remaps_per_sec 7.8% regression
@ 2025-09-03 14:53 kernel test robot
0 siblings, 0 replies; only message in thread
From: kernel test robot @ 2025-09-03 14:53 UTC (permalink / raw)
To: Vlastimil Babka
Cc: oe-lkp, lkp, Suren Baghdasaryan, maple-tree, linux-mm,
oliver.sang
Hello,
kernel test robot noticed a 7.8% regression of stress-ng.pagemove.page_remaps_per_sec on:
commit: 46bfd29ca2ac25defc473db5682ea0dd8215814a ("maple_tree: use percpu sheaves for maple_node_cache")
https://git.kernel.org/cgit/linux/kernel/git/vbabka/linux.git slub-percpu-sheaves-v6r1
testcase: stress-ng
config: x86_64-rhel-9.4
compiler: gcc-12
test machine: 64 threads 2 sockets Intel(R) Xeon(R) Gold 6346 CPU @ 3.10GHz (Ice Lake) with 256G memory
parameters:
nr_threads: 100%
testtime: 60s
test: pagemove
cpufreq_governor: performance
If you fix the issue in a separate patch/commit (i.e. not just a new version of
the same patch/commit), kindly add following tags
| Reported-by: kernel test robot <oliver.sang@intel.com>
| Closes: https://lore.kernel.org/oe-lkp/202509032258.411b604d-lkp@intel.com
Details are as below:
-------------------------------------------------------------------------------------------------->
The kernel config and materials to reproduce are available at:
https://download.01.org/0day-ci/archive/20250903/202509032258.411b604d-lkp@intel.com
=========================================================================================
compiler/cpufreq_governor/kconfig/nr_threads/rootfs/tbox_group/test/testcase/testtime:
gcc-12/performance/x86_64-rhel-9.4/100%/debian-12-x86_64-20240206.cgz/lkp-icl-2sp7/pagemove/stress-ng/60s
commit:
ac43de991d ("tools: Add sheaves support to testing infrastructure")
46bfd29ca2 ("maple_tree: use percpu sheaves for maple_node_cache")
ac43de991d9a4a42 46bfd29ca2ac25defc473db5682
---------------- ---------------------------
%stddev %change %stddev
\ | \
0.15 ± 31% +8.7 8.90 ± 29% mpstat.cpu.all.soft%
26992 ± 31% -86.5% 3649 ± 18% vmstat.system.cs
304348 ± 17% +124.5% 683375 ± 23% meminfo.SUnreclaim
400437 ± 12% +94.6% 779201 ± 20% meminfo.Slab
38862 ± 20% +116.9% 84299 ± 24% numa-vmstat.node0.nr_slab_unreclaimable
36829 ± 15% +113.9% 78790 ± 25% numa-vmstat.node1.nr_slab_unreclaimable
8305 ± 22% +64.2% 13635 ± 20% perf-c2c.DRAM.local
5195 ± 13% +304.2% 20998 ± 5% perf-c2c.HITM.local
6550 ± 12% +238.2% 22155 ± 4% perf-c2c.HITM.total
157267 ± 20% +129.0% 360199 ± 25% numa-meminfo.node0.SUnreclaim
199417 ± 24% +97.7% 394339 ± 23% numa-meminfo.node0.Slab
148061 ± 15% +127.0% 336065 ± 26% numa-meminfo.node1.SUnreclaim
202003 ± 13% +96.9% 397753 ± 22% numa-meminfo.node1.Slab
76366 ± 17% +121.9% 169482 ± 23% proc-vmstat.nr_slab_unreclaimable
31732425 -6.4% 29713727 proc-vmstat.numa_hit
31663453 -6.4% 29647558 proc-vmstat.numa_local
60845467 -1.9% 59703569 proc-vmstat.pgalloc_normal
60410743 -1.7% 59357556 proc-vmstat.pgfree
20530240 -8.2% 18847954 stress-ng.pagemove.ops
342222 -8.2% 314179 stress-ng.pagemove.ops_per_sec
242134 -7.8% 223297 stress-ng.pagemove.page_remaps_per_sec
951112 -89.1% 103990 ± 7% stress-ng.time.involuntary_context_switches
5975 -8.1% 5493 stress-ng.time.percent_of_cpu_this_job_got
3499 -8.1% 3217 stress-ng.time.system_time
91.57 ± 2% -7.3% 84.85 stress-ng.time.user_time
28017 ± 32% -87.2% 3575 ± 18% perf-stat.i.context-switches
1969 ± 3% -7.4% 1823 ± 2% perf-stat.i.cycles-between-cache-misses
0.50 ± 2% +18.4% 0.59 ± 2% perf-stat.overall.MPKI
0.24 ± 4% +0.0 0.28 ± 4% perf-stat.overall.branch-miss-rate%
73.76 ± 2% +6.1 79.90 perf-stat.overall.cache-miss-rate%
1.00 +8.6% 1.09 perf-stat.overall.cpi
2003 ± 3% -8.3% 1837 ± 3% perf-stat.overall.cycles-between-cache-misses
1.00 -7.9% 0.92 perf-stat.overall.ipc
27584 ± 31% -87.2% 3524 ± 18% perf-stat.ps.context-switches
1.181e+13 -7.7% 1.09e+13 perf-stat.total.instructions
82208 ± 35% -83.3% 13717 ±223% sched_debug.cfs_rq:/.left_deadline.avg
1731609 ± 24% -82.9% 295452 ±223% sched_debug.cfs_rq:/.left_deadline.max
360700 ± 27% -82.8% 61870 ±223% sched_debug.cfs_rq:/.left_deadline.stddev
82206 ± 35% -83.3% 13717 ±223% sched_debug.cfs_rq:/.left_vruntime.avg
1731555 ± 24% -82.9% 295431 ±223% sched_debug.cfs_rq:/.left_vruntime.max
360691 ± 27% -82.8% 61868 ±223% sched_debug.cfs_rq:/.left_vruntime.stddev
316185 ±101% -94.7% 16803 ± 72% sched_debug.cfs_rq:/.load.avg
18547721 ±108% -98.6% 255633 ±149% sched_debug.cfs_rq:/.load.max
2319717 ±106% -98.3% 38982 ±145% sched_debug.cfs_rq:/.load.stddev
885.40 ± 42% -79.9% 178.37 ± 45% sched_debug.cfs_rq:/.load_avg.stddev
82206 ± 35% -83.3% 13717 ±223% sched_debug.cfs_rq:/.right_vruntime.avg
1731559 ± 24% -82.9% 295431 ±223% sched_debug.cfs_rq:/.right_vruntime.max
360691 ± 27% -82.8% 61869 ±223% sched_debug.cfs_rq:/.right_vruntime.stddev
18775 ± 22% -78.2% 4087 ± 19% sched_debug.cpu.nr_switches.avg
29179 ± 23% -44.8% 16108 ± 21% sched_debug.cpu.nr_switches.max
14138 ± 23% -89.8% 1448 ± 23% sched_debug.cpu.nr_switches.min
0.01 ± 20% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.__alloc_frozen_pages_noprof.alloc_pages_mpol.alloc_pages_noprof.get_free_pages_noprof
0.09 ± 15% +190.1% 0.27 ± 27% perf-sched.sch_delay.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
0.34 ± 10% +1689.9% 6.10 ± 27% perf-sched.sch_delay.avg.ms.__cond_resched.change_pmd_range.isra.0.change_pud_range
0.29 ± 13% +1932.3% 5.86 ± 19% perf-sched.sch_delay.avg.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
0.25 ± 9% +2276.9% 5.97 ± 5% perf-sched.sch_delay.avg.ms.__cond_resched.down_write.anon_vma_clone.__split_vma.vms_gather_munmap_vmas
0.34 ± 7% +1657.8% 6.00 ± 16% perf-sched.sch_delay.avg.ms.__cond_resched.down_write.anon_vma_clone.copy_vma.copy_vma_and_data
0.35 ± 5% +1974.1% 7.24 ± 12% perf-sched.sch_delay.avg.ms.__cond_resched.down_write.move_ptes.move_page_tables.copy_vma_and_data
0.39 ± 17% +1522.1% 6.25 ± 11% perf-sched.sch_delay.avg.ms.__cond_resched.down_write.unlink_anon_vmas.free_pgtables.vms_clear_ptes
0.29 ± 7% +1834.4% 5.57 ± 18% perf-sched.sch_delay.avg.ms.__cond_resched.down_write.vma_prepare.__split_vma.vms_gather_munmap_vmas
0.30 ± 11% +1905.2% 6.02 ± 21% perf-sched.sch_delay.avg.ms.__cond_resched.down_write.vma_prepare.commit_merge.vma_expand
0.49 ± 24% +863.8% 4.77 ± 63% perf-sched.sch_delay.avg.ms.__cond_resched.down_write_killable.do_mprotect_pkey.__x64_sys_mprotect.do_syscall_64
0.32 ± 11% +1739.1% 5.97 ± 20% perf-sched.sch_delay.avg.ms.__cond_resched.down_write_killable.do_mremap.__do_sys_mremap.do_syscall_64
0.05 ± 25% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.kfree_rcu_work.process_one_work.worker_thread.kthread
0.26 ± 10% +2559.4% 6.90 ± 22% perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.__split_vma
0.30 ± 8% +2062.2% 6.39 ± 13% perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.commit_merge
0.34 ± 32% +1754.3% 6.27 ± 43% perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.vma_link
0.32 ± 11% +1913.0% 6.49 ± 18% perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_dup.__split_vma.vms_gather_munmap_vmas
0.31 ± 10% +1967.2% 6.46 ± 19% perf-sched.sch_delay.avg.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_dup.copy_vma.copy_vma_and_data
0.33 ± 12% +1852.2% 6.49 ± 12% perf-sched.sch_delay.avg.ms.__cond_resched.move_page_tables.copy_vma_and_data.move_vma.remap_move
0.02 ± 25% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.process_one_work.worker_thread.kthread.ret_from_fork
0.39 ± 8% +1385.0% 5.86 ± 13% perf-sched.sch_delay.avg.ms.__cond_resched.remove_vma.vms_complete_munmap_vmas.do_vmi_align_munmap.do_vmi_munmap
0.04 ±119% +276.3% 0.13 ± 23% perf-sched.sch_delay.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
0.53 ± 29% +105.4% 1.10 ± 54% perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
0.24 ±129% -100.0% 0.00 perf-sched.sch_delay.avg.ms.__cond_resched.task_work_run.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
0.29 ± 14% +1898.2% 5.87 ± 22% perf-sched.sch_delay.avg.ms.__cond_resched.unmap_page_range.unmap_vmas.vms_clear_ptes.part
0.27 ± 8% +2310.4% 6.41 ± 9% perf-sched.sch_delay.avg.ms.__cond_resched.unmap_vmas.vms_clear_ptes.part.0
0.32 ± 19% +1824.1% 6.18 ± 48% perf-sched.sch_delay.avg.ms.__cond_resched.zap_pmd_range.isra.0.unmap_page_range
0.21 ± 7% +3360.3% 7.24 ± 16% perf-sched.sch_delay.avg.ms.__cond_resched.zap_pte_range.zap_pmd_range.isra.0
0.23 ± 8% +1503.0% 3.64 ± 12% perf-sched.sch_delay.avg.ms.exit_to_user_mode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
0.52 ± 30% +560.2% 3.45 ± 56% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
0.47 ± 53% +662.5% 3.59 ± 46% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown].[unknown]
0.02 ± 20% +191.8% 0.05 ± 29% perf-sched.sch_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
0.03 ± 19% +1119.2% 0.37 ± 25% perf-sched.sch_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
0.03 ± 2% +262.7% 0.12 ± 92% perf-sched.sch_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
0.02 ± 25% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.__alloc_frozen_pages_noprof.alloc_pages_mpol.alloc_pages_noprof.get_free_pages_noprof
3.10 ± 28% +1019.8% 34.72 ± 30% perf-sched.sch_delay.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
3.79 ± 22% +1135.3% 46.77 ± 14% perf-sched.sch_delay.max.ms.__cond_resched.change_pmd_range.isra.0.change_pud_range
4.36 ± 21% +1064.9% 50.74 ± 10% perf-sched.sch_delay.max.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
5.47 ± 17% +870.9% 53.14 ± 4% perf-sched.sch_delay.max.ms.__cond_resched.down_write.anon_vma_clone.__split_vma.vms_gather_munmap_vmas
4.69 ± 15% +925.3% 48.13 ± 6% perf-sched.sch_delay.max.ms.__cond_resched.down_write.anon_vma_clone.copy_vma.copy_vma_and_data
4.64 ± 21% +1020.2% 51.94 ± 4% perf-sched.sch_delay.max.ms.__cond_resched.down_write.move_ptes.move_page_tables.copy_vma_and_data
4.03 ± 13% +1051.4% 46.45 ± 19% perf-sched.sch_delay.max.ms.__cond_resched.down_write.unlink_anon_vmas.free_pgtables.vms_clear_ptes
5.06 ± 26% +876.5% 49.38 ± 10% perf-sched.sch_delay.max.ms.__cond_resched.down_write.vma_prepare.__split_vma.vms_gather_munmap_vmas
4.59 ± 33% +1037.7% 52.22 ± 4% perf-sched.sch_delay.max.ms.__cond_resched.down_write.vma_prepare.commit_merge.vma_expand
3.24 ± 14% +769.1% 28.18 ± 70% perf-sched.sch_delay.max.ms.__cond_resched.down_write_killable.do_mprotect_pkey.__x64_sys_mprotect.do_syscall_64
5.24 ± 32% +898.2% 52.33 ± 5% perf-sched.sch_delay.max.ms.__cond_resched.down_write_killable.do_mremap.__do_sys_mremap.do_syscall_64
2.33 ± 2% +40.3% 3.26 ± 37% perf-sched.sch_delay.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
1.54 ± 56% -99.3% 0.01 ±115% perf-sched.sch_delay.max.ms.__cond_resched.kfree_rcu_monitor.process_one_work.worker_thread.kthread
3.01 ± 67% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.kfree_rcu_work.process_one_work.worker_thread.kthread
5.97 ± 18% +793.3% 53.30 ± 5% perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.__split_vma
6.14 ± 32% +821.3% 56.60 ± 3% perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.commit_merge
2.27 ± 44% +1109.8% 27.49 ± 48% perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.vma_link
6.04 ± 22% +795.0% 54.04 perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_dup.__split_vma.vms_gather_munmap_vmas
5.43 ± 20% +874.7% 52.94 ± 3% perf-sched.sch_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_dup.copy_vma.copy_vma_and_data
8.25 ± 60% +562.2% 54.63 perf-sched.sch_delay.max.ms.__cond_resched.move_page_tables.copy_vma_and_data.move_vma.remap_move
0.97 ±101% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.process_one_work.worker_thread.kthread.ret_from_fork
5.50 ± 17% +830.6% 51.18 ± 6% perf-sched.sch_delay.max.ms.__cond_resched.remove_vma.vms_complete_munmap_vmas.do_vmi_align_munmap.do_vmi_munmap
0.21 ±142% +4630.1% 10.00 ± 44% perf-sched.sch_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
0.50 ±121% -100.0% 0.00 perf-sched.sch_delay.max.ms.__cond_resched.task_work_run.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
3.33 ± 10% +994.1% 36.47 ± 27% perf-sched.sch_delay.max.ms.__cond_resched.unmap_page_range.unmap_vmas.vms_clear_ptes.part
6.23 ± 12% +789.7% 55.47 perf-sched.sch_delay.max.ms.__cond_resched.unmap_vmas.vms_clear_ptes.part.0
3.97 ± 24% +812.2% 36.25 ± 42% perf-sched.sch_delay.max.ms.__cond_resched.zap_pmd_range.isra.0.unmap_page_range
5.95 ± 25% +791.9% 53.08 ± 7% perf-sched.sch_delay.max.ms.__cond_resched.zap_pte_range.zap_pmd_range.isra.0
1.97 ± 58% +263.6% 7.15 ± 52% perf-sched.sch_delay.max.ms.anon_pipe_read.fifo_pipe_read.vfs_read.ksys_read
7.12 ± 27% +668.8% 54.75 perf-sched.sch_delay.max.ms.exit_to_user_mode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
5.89 ± 44% +720.4% 48.36 ± 10% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
5.76 ± 41% +794.7% 51.50 ± 7% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown].[unknown]
0.92 ±117% +1009.3% 10.23 ± 49% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
2.97 ± 16% +155.7% 7.60 ± 84% perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
1.78 ± 69% +245.5% 6.16 ± 57% perf-sched.sch_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
3.92 ± 27% +228.4% 12.89 ± 13% perf-sched.sch_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
27.62 ± 71% -82.7% 4.77 ± 39% perf-sched.sch_delay.max.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
0.16 ± 8% +1850.6% 3.12 ± 9% perf-sched.total_sch_delay.average.ms
9.73 +582.2% 66.40 ± 4% perf-sched.total_wait_and_delay.average.ms
140704 -88.3% 16412 ± 4% perf-sched.total_wait_and_delay.count.ms
3638 ± 5% -19.8% 2917 ± 5% perf-sched.total_wait_and_delay.max.ms
9.57 ± 2% +561.0% 63.28 ± 4% perf-sched.total_wait_time.average.ms
3638 ± 5% -19.8% 2917 ± 5% perf-sched.total_wait_time.max.ms
39.58 ±154% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.__cond_resched.__alloc_frozen_pages_noprof.alloc_pages_mpol.alloc_pages_noprof.get_free_pages_noprof
10.91 ± 2% +25.1% 13.64 ± 9% perf-sched.wait_and_delay.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
0.50 ± 9% +2276.3% 11.95 ± 5% perf-sched.wait_and_delay.avg.ms.__cond_resched.down_write.anon_vma_clone.__split_vma.vms_gather_munmap_vmas
0.28 ±101% +4248.3% 12.04 ± 21% perf-sched.wait_and_delay.avg.ms.__cond_resched.down_write.vma_prepare.commit_merge.vma_expand
0.65 ± 11% +1739.6% 11.95 ± 20% perf-sched.wait_and_delay.avg.ms.__cond_resched.down_write_killable.do_mremap.__do_sys_mremap.do_syscall_64
0.52 ± 10% +2560.3% 13.80 ± 22% perf-sched.wait_and_delay.avg.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.__split_vma
0.59 ± 8% +2062.1% 12.78 ± 13% perf-sched.wait_and_delay.avg.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.commit_merge
0.64 ± 11% +1912.0% 12.97 ± 18% perf-sched.wait_and_delay.avg.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_dup.__split_vma.vms_gather_munmap_vmas
0.62 ± 10% +1967.2% 12.91 ± 19% perf-sched.wait_and_delay.avg.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_dup.copy_vma.copy_vma_and_data
0.66 ± 12% +1854.0% 12.99 ± 12% perf-sched.wait_and_delay.avg.ms.__cond_resched.move_page_tables.copy_vma_and_data.move_vma.remap_move
0.79 ± 8% +1384.4% 11.72 ± 13% perf-sched.wait_and_delay.avg.ms.__cond_resched.remove_vma.vms_complete_munmap_vmas.do_vmi_align_munmap.do_vmi_munmap
0.53 ± 8% +2309.6% 12.82 ± 9% perf-sched.wait_and_delay.avg.ms.__cond_resched.unmap_vmas.vms_clear_ptes.part.0
0.42 ± 7% +3361.8% 14.48 ± 16% perf-sched.wait_and_delay.avg.ms.__cond_resched.zap_pte_range.zap_pmd_range.isra.0
0.90 ± 5% +1294.4% 12.51 ± 11% perf-sched.wait_and_delay.avg.ms.exit_to_user_mode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
1.21 ± 31% +495.3% 7.19 ± 46% perf-sched.wait_and_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown].[unknown]
105.52 ± 8% +68.9% 178.18 ± 7% perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range_clock.poll_schedule_timeout.constprop.0.do_poll
522.24 ± 3% -67.5% 169.60 ± 15% perf-sched.wait_and_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
12.74 +4477.0% 583.16 ± 5% perf-sched.wait_and_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
3.83 ± 50% -100.0% 0.00 perf-sched.wait_and_delay.count.__cond_resched.__alloc_frozen_pages_noprof.alloc_pages_mpol.alloc_pages_noprof.get_free_pages_noprof
4875 ± 7% -92.9% 345.83 ± 11% perf-sched.wait_and_delay.count.__cond_resched.down_write.anon_vma_clone.__split_vma.vms_gather_munmap_vmas
1491 ± 5% -82.5% 261.67 ± 12% perf-sched.wait_and_delay.count.__cond_resched.down_write_killable.do_mremap.__do_sys_mremap.do_syscall_64
93.83 ± 43% -75.7% 22.83 ± 79% perf-sched.wait_and_delay.count.__cond_resched.generic_perform_write.shmem_file_write_iter.vfs_write.ksys_write
215.00 ± 13% -99.8% 0.33 ±141% perf-sched.wait_and_delay.count.__cond_resched.kfree_rcu_monitor.process_one_work.worker_thread.kthread
3319 ± 14% -94.6% 178.33 ± 11% perf-sched.wait_and_delay.count.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.__split_vma
6164 ± 4% -83.2% 1033 ± 10% perf-sched.wait_and_delay.count.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.commit_merge
2792 ± 5% -79.8% 563.83 ± 7% perf-sched.wait_and_delay.count.__cond_resched.kmem_cache_alloc_noprof.vm_area_dup.__split_vma.vms_gather_munmap_vmas
2914 ± 5% -82.5% 509.67 ± 9% perf-sched.wait_and_delay.count.__cond_resched.kmem_cache_alloc_noprof.vm_area_dup.copy_vma.copy_vma_and_data
4177 ± 6% -78.4% 900.83 ± 10% perf-sched.wait_and_delay.count.__cond_resched.move_page_tables.copy_vma_and_data.move_vma.remap_move
1261 ± 6% -76.2% 300.33 ± 11% perf-sched.wait_and_delay.count.__cond_resched.remove_vma.vms_complete_munmap_vmas.do_vmi_align_munmap.do_vmi_munmap
6.33 ± 98% +11578.9% 739.67 ± 12% perf-sched.wait_and_delay.count.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
15156 ± 2% -90.6% 1421 ± 5% perf-sched.wait_and_delay.count.__cond_resched.unmap_vmas.vms_clear_ptes.part.0
4087 ± 5% -94.5% 226.83 ± 8% perf-sched.wait_and_delay.count.__cond_resched.zap_pte_range.zap_pmd_range.isra.0
16960 -92.5% 1273 ± 19% perf-sched.wait_and_delay.count.exit_to_user_mode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
1756 ± 19% -54.4% 801.33 ± 82% perf-sched.wait_and_delay.count.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown].[unknown]
116.67 ± 5% -39.7% 70.33 ± 7% perf-sched.wait_and_delay.count.schedule_hrtimeout_range_clock.poll_schedule_timeout.constprop.0.do_poll
689.33 ± 9% +338.6% 3023 ± 14% perf-sched.wait_and_delay.count.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
65400 -99.3% 476.83 ± 7% perf-sched.wait_and_delay.count.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
151.10 ±202% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.__cond_resched.__alloc_frozen_pages_noprof.alloc_pages_mpol.alloc_pages_noprof.get_free_pages_noprof
1001 +22.7% 1227 ± 16% perf-sched.wait_and_delay.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
10.94 ± 17% +870.9% 106.27 ± 4% perf-sched.wait_and_delay.max.ms.__cond_resched.down_write.anon_vma_clone.__split_vma.vms_gather_munmap_vmas
3.94 ±113% +2553.3% 104.43 ± 4% perf-sched.wait_and_delay.max.ms.__cond_resched.down_write.vma_prepare.commit_merge.vma_expand
10.48 ± 32% +898.2% 104.65 ± 5% perf-sched.wait_and_delay.max.ms.__cond_resched.down_write_killable.do_mremap.__do_sys_mremap.do_syscall_64
11.93 ± 18% +793.3% 106.60 ± 5% perf-sched.wait_and_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.__split_vma
12.29 ± 32% +821.3% 113.20 ± 3% perf-sched.wait_and_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.commit_merge
12.07 ± 22% +795.0% 108.07 perf-sched.wait_and_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_dup.__split_vma.vms_gather_munmap_vmas
10.86 ± 20% +874.7% 105.88 ± 3% perf-sched.wait_and_delay.max.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_dup.copy_vma.copy_vma_and_data
16.50 ± 60% +562.2% 109.26 perf-sched.wait_and_delay.max.ms.__cond_resched.move_page_tables.copy_vma_and_data.move_vma.remap_move
11.00 ± 17% +830.6% 102.37 ± 6% perf-sched.wait_and_delay.max.ms.__cond_resched.remove_vma.vms_complete_munmap_vmas.do_vmi_align_munmap.do_vmi_munmap
12.47 ± 12% +789.7% 110.94 perf-sched.wait_and_delay.max.ms.__cond_resched.unmap_vmas.vms_clear_ptes.part.0
11.90 ± 25% +791.9% 106.17 ± 7% perf-sched.wait_and_delay.max.ms.__cond_resched.zap_pte_range.zap_pmd_range.isra.0
6.56 ± 26% +159.9% 17.04 ± 46% perf-sched.wait_and_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
3545 ± 9% -69.1% 1094 ± 7% perf-sched.wait_and_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
0.34 ± 10% +1689.9% 6.10 ± 27% perf-sched.wait_time.avg.ms.__cond_resched.change_pmd_range.isra.0.change_pud_range
0.29 ± 13% +1932.2% 5.86 ± 19% perf-sched.wait_time.avg.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
0.25 ± 9% +2276.9% 5.97 ± 5% perf-sched.wait_time.avg.ms.__cond_resched.down_write.anon_vma_clone.__split_vma.vms_gather_munmap_vmas
0.34 ± 7% +1657.8% 6.00 ± 16% perf-sched.wait_time.avg.ms.__cond_resched.down_write.anon_vma_clone.copy_vma.copy_vma_and_data
0.35 ± 5% +1974.1% 7.24 ± 12% perf-sched.wait_time.avg.ms.__cond_resched.down_write.move_ptes.move_page_tables.copy_vma_and_data
0.39 ± 17% +1522.1% 6.25 ± 11% perf-sched.wait_time.avg.ms.__cond_resched.down_write.unlink_anon_vmas.free_pgtables.vms_clear_ptes
0.29 ± 7% +1834.4% 5.57 ± 18% perf-sched.wait_time.avg.ms.__cond_resched.down_write.vma_prepare.__split_vma.vms_gather_munmap_vmas
0.30 ± 11% +1905.2% 6.02 ± 21% perf-sched.wait_time.avg.ms.__cond_resched.down_write.vma_prepare.commit_merge.vma_expand
0.49 ± 24% +863.8% 4.77 ± 63% perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.do_mprotect_pkey.__x64_sys_mprotect.do_syscall_64
0.32 ± 11% +1739.1% 5.97 ± 20% perf-sched.wait_time.avg.ms.__cond_resched.down_write_killable.do_mremap.__do_sys_mremap.do_syscall_64
11.93 ± 30% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.kfree_rcu_work.process_one_work.worker_thread.kthread
0.26 ± 10% +2559.4% 6.90 ± 22% perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.__split_vma
0.30 ± 8% +2062.2% 6.39 ± 13% perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.commit_merge
0.34 ± 32% +1754.3% 6.27 ± 43% perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.vma_link
0.32 ± 11% +1913.0% 6.49 ± 18% perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_dup.__split_vma.vms_gather_munmap_vmas
0.31 ± 10% +1967.2% 6.46 ± 19% perf-sched.wait_time.avg.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_dup.copy_vma.copy_vma_and_data
0.33 ± 12% +1852.2% 6.49 ± 12% perf-sched.wait_time.avg.ms.__cond_resched.move_page_tables.copy_vma_and_data.move_vma.remap_move
8.76 ± 42% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.process_one_work.worker_thread.kthread.ret_from_fork
0.39 ± 8% +1385.0% 5.86 ± 13% perf-sched.wait_time.avg.ms.__cond_resched.remove_vma.vms_complete_munmap_vmas.do_vmi_align_munmap.do_vmi_munmap
0.24 ±129% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.task_work_run.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
0.29 ± 14% +1898.2% 5.87 ± 22% perf-sched.wait_time.avg.ms.__cond_resched.unmap_page_range.unmap_vmas.vms_clear_ptes.part
0.27 ± 8% +2310.4% 6.41 ± 9% perf-sched.wait_time.avg.ms.__cond_resched.unmap_vmas.vms_clear_ptes.part.0
0.32 ± 19% +1824.1% 6.18 ± 48% perf-sched.wait_time.avg.ms.__cond_resched.zap_pmd_range.isra.0.unmap_page_range
0.21 ± 7% +3360.2% 7.24 ± 16% perf-sched.wait_time.avg.ms.__cond_resched.zap_pte_range.zap_pmd_range.isra.0
0.67 ± 7% +1223.9% 8.88 ± 13% perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
0.52 ± 30% +559.6% 3.45 ± 56% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
0.74 ± 32% +389.0% 3.61 ± 46% perf-sched.wait_time.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown].[unknown]
104.99 ± 8% +68.9% 177.29 ± 7% perf-sched.wait_time.avg.ms.schedule_hrtimeout_range_clock.poll_schedule_timeout.constprop.0.do_poll
522.21 ± 3% -67.6% 169.23 ± 15% perf-sched.wait_time.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
12.71 +4488.3% 583.04 ± 5% perf-sched.wait_time.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
1000 +22.0% 1220 ± 17% perf-sched.wait_time.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
3.79 ± 22% +1135.3% 46.77 ± 14% perf-sched.wait_time.max.ms.__cond_resched.change_pmd_range.isra.0.change_pud_range
4.36 ± 21% +1064.9% 50.74 ± 10% perf-sched.wait_time.max.ms.__cond_resched.change_pud_range.isra.0.change_protection_range
5.47 ± 17% +870.9% 53.14 ± 4% perf-sched.wait_time.max.ms.__cond_resched.down_write.anon_vma_clone.__split_vma.vms_gather_munmap_vmas
4.69 ± 15% +925.3% 48.13 ± 6% perf-sched.wait_time.max.ms.__cond_resched.down_write.anon_vma_clone.copy_vma.copy_vma_and_data
4.64 ± 21% +1020.2% 51.94 ± 4% perf-sched.wait_time.max.ms.__cond_resched.down_write.move_ptes.move_page_tables.copy_vma_and_data
4.03 ± 13% +1051.4% 46.45 ± 19% perf-sched.wait_time.max.ms.__cond_resched.down_write.unlink_anon_vmas.free_pgtables.vms_clear_ptes
5.06 ± 26% +876.5% 49.38 ± 10% perf-sched.wait_time.max.ms.__cond_resched.down_write.vma_prepare.__split_vma.vms_gather_munmap_vmas
4.59 ± 33% +1037.7% 52.22 ± 4% perf-sched.wait_time.max.ms.__cond_resched.down_write.vma_prepare.commit_merge.vma_expand
3.24 ± 14% +769.1% 28.18 ± 70% perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.do_mprotect_pkey.__x64_sys_mprotect.do_syscall_64
5.24 ± 32% +898.2% 52.33 ± 5% perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.do_mremap.__do_sys_mremap.do_syscall_64
2.33 ± 2% +40.3% 3.26 ± 37% perf-sched.wait_time.max.ms.__cond_resched.dput.__fput.__x64_sys_close.do_syscall_64
920.04 ±115% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.kfree_rcu_work.process_one_work.worker_thread.kthread
5.97 ± 18% +793.3% 53.30 ± 5% perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.__split_vma
6.14 ± 32% +821.3% 56.60 ± 3% perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.commit_merge
2.27 ± 44% +1109.8% 27.49 ± 48% perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.mas_alloc_nodes.mas_preallocate.vma_link
6.04 ± 22% +795.0% 54.04 perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_dup.__split_vma.vms_gather_munmap_vmas
5.43 ± 20% +874.7% 52.94 ± 3% perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.vm_area_dup.copy_vma.copy_vma_and_data
8.25 ± 60% +562.2% 54.63 perf-sched.wait_time.max.ms.__cond_resched.move_page_tables.copy_vma_and_data.move_vma.remap_move
681.18 ± 74% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.process_one_work.worker_thread.kthread.ret_from_fork
5.50 ± 17% +830.6% 51.18 ± 6% perf-sched.wait_time.max.ms.__cond_resched.remove_vma.vms_complete_munmap_vmas.do_vmi_align_munmap.do_vmi_munmap
0.50 ±121% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.task_work_run.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
3.33 ± 10% +994.1% 36.47 ± 27% perf-sched.wait_time.max.ms.__cond_resched.unmap_page_range.unmap_vmas.vms_clear_ptes.part
6.23 ± 12% +789.7% 55.47 perf-sched.wait_time.max.ms.__cond_resched.unmap_vmas.vms_clear_ptes.part.0
3.97 ± 24% +812.2% 36.25 ± 42% perf-sched.wait_time.max.ms.__cond_resched.zap_pmd_range.isra.0.unmap_page_range
5.95 ± 25% +791.9% 53.08 ± 7% perf-sched.wait_time.max.ms.__cond_resched.zap_pte_range.zap_pmd_range.isra.0
6.12 ± 40% +689.8% 48.36 ± 10% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
0.92 ±117% +1009.3% 10.23 ± 49% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown].[unknown]
5.73 ± 15% +117.2% 12.44 ± 45% perf-sched.wait_time.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
3545 ± 9% -69.1% 1094 ± 7% perf-sched.wait_time.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
--
0-DAY CI Kernel Test Service
https://github.com/intel/lkp-tests/wiki
^ permalink raw reply [flat|nested] only message in thread
only message in thread, other threads:[~2025-09-03 14:54 UTC | newest]
Thread overview: (only message) (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2025-09-03 14:53 [vbabka:slub-percpu-sheaves-v6r1] [maple_tree] 46bfd29ca2: stress-ng.pagemove.page_remaps_per_sec 7.8% regression kernel test robot
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).