* [linux-next:master] [nfs] b6dea6c7fe: fsmark.files_per_sec 17.6% regression
@ 2024-11-30 10:44 kernel test robot
0 siblings, 0 replies; only message in thread
From: kernel test robot @ 2024-11-30 10:44 UTC (permalink / raw)
To: Li Lingfeng; +Cc: oe-lkp, lkp, Trond Myklebust, linux-nfs, oliver.sang
Hello,
kernel test robot noticed a 17.6% regression of fsmark.files_per_sec on:
commit: b6dea6c7fe2d8187050f882fe6f872d30e495ffe ("nfs: pass flags to second superblock")
https://git.kernel.org/cgit/linux/kernel/git/next/linux-next.git master
[test failed on linux-next/master cfba9f07a1d6aeca38f47f1f472cfb0ba133d341]
testcase: fsmark
config: x86_64-rhel-9.4
compiler: gcc-12
test machine: 48 threads 2 sockets Intel(R) Xeon(R) CPU E5-2697 v2 @ 2.70GHz (Ivy Bridge-EP) with 64G memory
parameters:
iterations: 1x
nr_threads: 32t
disk: 1SSD
fs: xfs
fs2: nfsv4
filesize: 8K
test_size: 400M
sync_method: fsyncBeforeClose
nr_directories: 16d
nr_files_per_directory: 256fpd
cpufreq_governor: performance
In addition to that, the commit also has significant impact on the following tests:
+------------------+------------------------------------------------------------------------------------------------+
| testcase: change | fsmark: fsmark.files_per_sec 9.4% regression |
| test machine | 48 threads 2 sockets Intel(R) Xeon(R) CPU E5-2697 v2 @ 2.70GHz (Ivy Bridge-EP) with 64G memory |
| test parameters | cpufreq_governor=performance |
| | disk=1SSD |
| | filesize=9B |
| | fs2=nfsv4 |
| | fs=ext4 |
| | iterations=1x |
| | nr_directories=16d |
| | nr_files_per_directory=256fpd |
| | nr_threads=32t |
| | sync_method=fsyncBeforeClose |
| | test_size=400M |
+------------------+------------------------------------------------------------------------------------------------+
| testcase: change | fsmark: fsmark.files_per_sec 15.9% regression |
| test machine | 48 threads 2 sockets Intel(R) Xeon(R) CPU E5-2697 v2 @ 2.70GHz (Ivy Bridge-EP) with 64G memory |
| test parameters | cpufreq_governor=performance |
| | disk=1SSD |
| | filesize=9B |
| | fs2=nfsv4 |
| | fs=btrfs |
| | iterations=1x |
| | nr_directories=16d |
| | nr_files_per_directory=256fpd |
| | nr_threads=32t |
| | sync_method=fsyncBeforeClose |
| | test_size=400M |
+------------------+------------------------------------------------------------------------------------------------+
If you fix the issue in a separate patch/commit (i.e. not just a new version of
the same patch/commit), kindly add following tags
| Reported-by: kernel test robot <oliver.sang@intel.com>
| Closes: https://lore.kernel.org/oe-lkp/202411301633.3ed8df2-lkp@intel.com
Details are as below:
-------------------------------------------------------------------------------------------------->
The kernel config and materials to reproduce are available at:
https://download.01.org/0day-ci/archive/20241130/202411301633.3ed8df2-lkp@intel.com
=========================================================================================
compiler/cpufreq_governor/disk/filesize/fs2/fs/iterations/kconfig/nr_directories/nr_files_per_directory/nr_threads/rootfs/sync_method/tbox_group/test_size/testcase:
gcc-12/performance/1SSD/8K/nfsv4/xfs/1x/x86_64-rhel-9.4/16d/256fpd/32t/debian-12-x86_64-20240206.cgz/fsyncBeforeClose/lkp-ivb-2ep2/400M/fsmark
commit:
66f9dac907 ("Revert "nfs: don't reuse partially completed requests in nfs_lock_and_join_requests"")
b6dea6c7fe ("nfs: pass flags to second superblock")
66f9dac9077c9c06 b6dea6c7fe2d8187050f882fe6f
---------------- ---------------------------
%stddev %change %stddev
\ | \
5.937e+08 ± 2% +12.7% 6.69e+08 cpuidle..time
2360703 +75.8% 4150723 cpuidle..usage
381905 ± 4% +70.8% 652235 ± 2% numa-numastat.node0.local_node
410324 +65.9% 680670 ± 3% numa-numastat.node0.numa_hit
0.48 ±223% +2.8 3.27 ± 55% perf-profile.calltrace.cycles-pp.__mmput.exit_mm.do_exit.do_group_exit.__x64_sys_exit_group
0.48 ±223% +2.8 3.27 ± 55% perf-profile.calltrace.cycles-pp.exit_mm.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call
75.28 +5.2% 79.20 iostat.cpu.idle
16.87 -36.2% 10.76 iostat.cpu.iowait
5.30 ± 2% +42.1% 7.52 ± 2% iostat.cpu.system
19.34 -7.3 12.07 mpstat.cpu.all.iowait%
0.65 ± 3% +0.3 0.94 ± 2% mpstat.cpu.all.soft%
4.88 ± 3% +2.1 7.00 ± 3% mpstat.cpu.all.sys%
11.82 ± 3% +20.4% 14.24 mpstat.max_utilization_pct
72239710 ± 2% +55.2% 1.121e+08 fsmark.app_overhead
4389 ± 2% -17.6% 3614 fsmark.files_per_sec
54.67 ± 2% +106.1% 112.67 fsmark.time.percent_of_cpu_this_job_got
5.94 +160.1% 15.45 fsmark.time.system_time
285139 +307.7% 1162591 fsmark.time.voluntary_context_switches
368790 -47.8% 192655 ± 2% meminfo.Inactive
368790 -47.8% 192655 ± 2% meminfo.Inactive(file)
173587 -23.0% 133612 meminfo.KReclaimable
173587 -23.0% 133612 meminfo.SReclaimable
317373 -13.0% 276190 meminfo.Slab
16.92 -36.4% 10.76 vmstat.cpu.wa
58264 ± 3% -11.4% 51622 ± 2% vmstat.io.bo
8.45 ± 7% -37.6% 5.27 ± 7% vmstat.procs.b
269409 ± 2% +66.6% 448791 ± 3% vmstat.system.cs
61548 +16.2% 71538 vmstat.system.in
53681 -19.8% 43073 ± 8% numa-vmstat.node0.nr_inactive_file
53681 -19.8% 43073 ± 8% numa-vmstat.node0.nr_zone_inactive_file
410388 +66.1% 681472 ± 3% numa-vmstat.node0.numa_hit
381970 ± 4% +71.0% 653038 ± 2% numa-vmstat.node0.numa_local
60942 ± 22% -48.1% 31659 ± 28% numa-vmstat.node1.nr_file_pages
38775 ± 5% -87.9% 4703 ± 62% numa-vmstat.node1.nr_inactive_file
12374 ± 5% -60.4% 4902 ± 17% numa-vmstat.node1.nr_slab_reclaimable
38775 ± 5% -87.9% 4703 ± 62% numa-vmstat.node1.nr_zone_inactive_file
998853 -4.5% 954271 proc-vmstat.nr_file_pages
92299 ± 2% -48.3% 47689 ± 2% proc-vmstat.nr_inactive_file
43433 -23.4% 33269 proc-vmstat.nr_slab_reclaimable
92299 ± 2% -48.3% 47689 ± 2% proc-vmstat.nr_zone_inactive_file
646052 +46.8% 948130 proc-vmstat.numa_hit
596308 +50.7% 898452 proc-vmstat.numa_local
1067778 +28.3% 1369682 proc-vmstat.pgalloc_normal
791047 ± 7% +39.5% 1103617 ± 2% proc-vmstat.pgfree
214306 -19.8% 171874 ± 8% numa-meminfo.node0.Inactive
214306 -19.8% 171874 ± 8% numa-meminfo.node0.Inactive(file)
243433 ± 22% -48.0% 126616 ± 28% numa-meminfo.node1.FilePages
154772 ± 5% -87.9% 18798 ± 62% numa-meminfo.node1.Inactive
154772 ± 5% -87.9% 18798 ± 62% numa-meminfo.node1.Inactive(file)
49305 ± 5% -60.2% 19601 ± 17% numa-meminfo.node1.KReclaimable
1226203 ± 10% -21.5% 962004 ± 4% numa-meminfo.node1.MemUsed
49305 ± 5% -60.2% 19601 ± 17% numa-meminfo.node1.SReclaimable
113120 ± 4% -30.5% 78616 ± 6% numa-meminfo.node1.Slab
1.09 ± 2% +17.2% 1.28 ± 2% perf-stat.i.MPKI
2.083e+09 +19.9% 2.496e+09 perf-stat.i.branch-instructions
5.09 -0.6 4.46 perf-stat.i.branch-miss-rate%
1.047e+08 +6.2% 1.112e+08 perf-stat.i.branch-misses
10791820 ± 2% +39.9% 15097313 perf-stat.i.cache-misses
2.272e+08 +34.6% 3.058e+08 perf-stat.i.cache-references
326576 ± 2% +60.3% 523588 perf-stat.i.context-switches
1.62 +8.2% 1.75 perf-stat.i.cpi
1.584e+10 +30.3% 2.063e+10 perf-stat.i.cpu-cycles
2314 ± 5% +68.5% 3899 ± 2% perf-stat.i.cpu-migrations
1484 -7.9% 1366 perf-stat.i.cycles-between-cache-misses
1.002e+10 +18.8% 1.191e+10 perf-stat.i.instructions
0.64 -7.8% 0.59 perf-stat.i.ipc
6.98 ± 2% +58.4% 11.06 perf-stat.i.metric.K/sec
9083 ± 4% -12.8% 7919 ± 5% perf-stat.i.minor-faults
9084 ± 4% -12.8% 7919 ± 5% perf-stat.i.page-faults
1.08 ± 2% +17.7% 1.27 perf-stat.overall.MPKI
5.03 -0.6 4.46 perf-stat.overall.branch-miss-rate%
4.75 +0.2 4.94 perf-stat.overall.cache-miss-rate%
1.58 +9.6% 1.73 perf-stat.overall.cpi
1468 -6.9% 1366 perf-stat.overall.cycles-between-cache-misses
0.63 -8.8% 0.58 perf-stat.overall.ipc
1.923e+09 +21.2% 2.33e+09 perf-stat.ps.branch-instructions
96657387 +7.4% 1.038e+08 perf-stat.ps.branch-misses
9961712 ± 2% +41.5% 14091659 perf-stat.ps.cache-misses
2.097e+08 +36.1% 2.854e+08 perf-stat.ps.cache-references
301501 ± 2% +62.1% 488770 perf-stat.ps.context-switches
44319 +1.1% 44810 perf-stat.ps.cpu-clock
1.462e+10 +31.7% 1.926e+10 perf-stat.ps.cpu-cycles
2137 ± 5% +70.3% 3640 ± 2% perf-stat.ps.cpu-migrations
9.25e+09 +20.2% 1.112e+10 perf-stat.ps.instructions
44319 +1.1% 44810 perf-stat.ps.task-clock
1.206e+11 +38.7% 1.673e+11 perf-stat.total.instructions
***************************************************************************************************
lkp-ivb-2ep2: 48 threads 2 sockets Intel(R) Xeon(R) CPU E5-2697 v2 @ 2.70GHz (Ivy Bridge-EP) with 64G memory
=========================================================================================
compiler/cpufreq_governor/disk/filesize/fs2/fs/iterations/kconfig/nr_directories/nr_files_per_directory/nr_threads/rootfs/sync_method/tbox_group/test_size/testcase:
gcc-12/performance/1SSD/9B/nfsv4/ext4/1x/x86_64-rhel-9.4/16d/256fpd/32t/debian-12-x86_64-20240206.cgz/fsyncBeforeClose/lkp-ivb-2ep2/400M/fsmark
commit:
66f9dac907 ("Revert "nfs: don't reuse partially completed requests in nfs_lock_and_join_requests"")
b6dea6c7fe ("nfs: pass flags to second superblock")
66f9dac9077c9c06 b6dea6c7fe2d8187050f882fe6f
---------------- ---------------------------
%stddev %change %stddev
\ | \
5106420 +67.0% 8530260 cpuidle..usage
-7.83 +34.0% -10.50 sched_debug.cpu.nr_uninterruptible.min
74.15 +4.5% 77.46 iostat.cpu.idle
18.76 -31.4% 12.87 iostat.cpu.iowait
4.63 +49.8% 6.94 iostat.cpu.system
2.46 ± 2% +10.9% 2.73 iostat.cpu.user
693560 ± 10% +42.7% 989531 ± 10% numa-numastat.node0.local_node
707512 ± 9% +42.3% 1007109 ± 9% numa-numastat.node0.numa_hit
506340 ± 12% +85.9% 941327 ± 10% numa-numastat.node1.local_node
542074 ± 11% +79.6% 973549 ± 9% numa-numastat.node1.numa_hit
2.236e+08 +34.9% 3.016e+08 fsmark.app_overhead
2969 -9.4% 2692 fsmark.files_per_sec
40.17 +138.2% 95.67 fsmark.time.percent_of_cpu_this_job_got
12.75 +176.3% 35.23 fsmark.time.system_time
568715 +308.9% 2325201 fsmark.time.voluntary_context_switches
19.76 -6.3 13.49 mpstat.cpu.all.iowait%
0.59 +0.3 0.88 mpstat.cpu.all.soft%
3.90 +2.1 5.99 mpstat.cpu.all.sys%
2.38 ± 2% +0.3 2.71 mpstat.cpu.all.usr%
2.00 +1783.3% 37.67 mpstat.max_utilization.seconds
10.03 +37.5% 13.79 mpstat.max_utilization_pct
18.74 -31.0% 12.92 vmstat.cpu.wa
107144 -5.5% 101202 vmstat.io.bo
9.71 ± 3% -35.7% 6.24 ± 10% vmstat.procs.b
4.53 ± 5% +27.6% 5.78 ± 6% vmstat.procs.r
197002 +79.6% 353838 vmstat.system.cs
58233 +13.7% 66197 vmstat.system.in
853950 +11.7% 954196 meminfo.Active
853950 +11.7% 954196 meminfo.Active(anon)
215530 +11.0% 239344 meminfo.Buffers
625583 -28.1% 449956 meminfo.Inactive
625583 -28.1% 449956 meminfo.Inactive(file)
241001 -27.6% 174436 meminfo.KReclaimable
106011 ± 8% +49.3% 158268 ± 10% meminfo.Mapped
241001 -27.6% 174436 meminfo.SReclaimable
46597 ± 21% +201.0% 140273 ± 8% meminfo.Shmem
388729 -19.3% 313805 meminfo.Slab
82817 ± 40% +45.7% 120654 ± 7% numa-vmstat.node0.nr_active_anon
76526 ± 43% +41.7% 108400 ± 2% numa-vmstat.node0.nr_anon_pages
951.19 ± 20% -22.7% 735.13 ± 7% numa-vmstat.node0.nr_dirty
80529 ± 21% -37.8% 50085 ± 10% numa-vmstat.node0.nr_inactive_file
21948 ± 8% +24.0% 27212 ± 10% numa-vmstat.node0.nr_mapped
39688 ± 5% -25.1% 29725 ± 4% numa-vmstat.node0.nr_slab_reclaimable
21176 ± 2% -8.3% 19425 ± 3% numa-vmstat.node0.nr_slab_unreclaimable
82817 ± 40% +45.7% 120653 ± 7% numa-vmstat.node0.nr_zone_active_anon
80529 ± 21% -37.8% 50085 ± 10% numa-vmstat.node0.nr_zone_inactive_file
707539 ± 9% +42.5% 1008378 ± 9% numa-vmstat.node0.numa_hit
693587 ± 10% +42.9% 990800 ± 10% numa-vmstat.node0.numa_local
5107 ± 58% +162.7% 13416 ± 30% numa-vmstat.node1.nr_mapped
5650 ± 24% +315.7% 23492 ± 41% numa-vmstat.node1.nr_shmem
20564 ± 10% -32.2% 13935 ± 9% numa-vmstat.node1.nr_slab_reclaimable
541047 ± 11% +80.2% 974981 ± 9% numa-vmstat.node1.numa_hit
505313 ± 13% +86.6% 942759 ± 10% numa-vmstat.node1.numa_local
6.38 ± 46% -3.5 2.85 ±141% perf-profile.calltrace.cycles-pp.__cmd_record
6.38 ± 46% -3.5 2.85 ±141% perf-profile.calltrace.cycles-pp.perf_session__process_events.record__finish_output.__cmd_record
6.38 ± 46% -3.5 2.85 ±141% perf-profile.calltrace.cycles-pp.reader__read_event.perf_session__process_events.record__finish_output.__cmd_record
6.38 ± 46% -3.5 2.85 ±141% perf-profile.calltrace.cycles-pp.record__finish_output.__cmd_record
5.88 ± 35% -3.4 2.47 ±142% perf-profile.calltrace.cycles-pp.ordered_events__queue.process_simple.reader__read_event.perf_session__process_events.record__finish_output
5.88 ± 35% -3.4 2.47 ±142% perf-profile.calltrace.cycles-pp.queue_event.ordered_events__queue.process_simple.reader__read_event.perf_session__process_events
4.37 ± 56% -1.5 2.85 ±141% perf-profile.calltrace.cycles-pp.process_simple.reader__read_event.perf_session__process_events.record__finish_output.__cmd_record
7.90 ± 12% -5.0 2.85 ±141% perf-profile.children.cycles-pp.perf_session__process_events
7.90 ± 12% -5.0 2.85 ±141% perf-profile.children.cycles-pp.reader__read_event
7.90 ± 12% -5.0 2.85 ±141% perf-profile.children.cycles-pp.record__finish_output
5.88 ± 35% -3.4 2.47 ±142% perf-profile.children.cycles-pp.ordered_events__queue
5.88 ± 35% -3.4 2.47 ±142% perf-profile.children.cycles-pp.queue_event
5.88 ± 35% -3.0 2.85 ±141% perf-profile.children.cycles-pp.process_simple
5.74 ± 74% +5.4 11.18 ± 11% perf-profile.children.cycles-pp.syscall_exit_to_user_mode
31.02 ± 11% +10.6 41.66 ± 12% perf-profile.children.cycles-pp.do_syscall_64
31.29 ± 11% +10.7 42.01 ± 12% perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
5.47 ± 35% -3.3 2.14 ±141% perf-profile.self.cycles-pp.queue_event
331331 ± 40% +45.4% 481839 ± 6% numa-meminfo.node0.Active
331331 ± 40% +45.4% 481839 ± 6% numa-meminfo.node0.Active(anon)
36095 ± 34% +185.6% 103075 ± 61% numa-meminfo.node0.AnonHugePages
306688 ± 43% +41.3% 433493 ± 2% numa-meminfo.node0.AnonPages
352666 ± 34% +39.9% 493368 ± 2% numa-meminfo.node0.AnonPages.max
3781 ± 20% -22.7% 2921 ± 7% numa-meminfo.node0.Dirty
321823 ± 21% -37.9% 199875 ± 10% numa-meminfo.node0.Inactive
321823 ± 21% -37.9% 199875 ± 10% numa-meminfo.node0.Inactive(file)
158690 ± 5% -25.1% 118818 ± 4% numa-meminfo.node0.KReclaimable
86534 ± 7% +23.8% 107125 ± 10% numa-meminfo.node0.Mapped
158690 ± 5% -25.1% 118818 ± 4% numa-meminfo.node0.SReclaimable
84681 ± 2% -8.2% 77738 ± 3% numa-meminfo.node0.SUnreclaim
243372 ± 3% -19.2% 196556 ± 3% numa-meminfo.node0.Slab
546078 ± 22% -22.0% 425684 ± 3% numa-meminfo.node1.AnonPages.max
82272 ± 10% -32.3% 55659 ± 9% numa-meminfo.node1.KReclaimable
19884 ± 58% +162.6% 52218 ± 29% numa-meminfo.node1.Mapped
82272 ± 10% -32.3% 55659 ± 9% numa-meminfo.node1.SReclaimable
22229 ± 25% +316.9% 92680 ± 41% numa-meminfo.node1.Shmem
145380 ± 5% -19.3% 117350 ± 5% numa-meminfo.node1.Slab
213612 +11.8% 238855 proc-vmstat.nr_active_anon
1073968 -1.9% 1053722 proc-vmstat.nr_file_pages
156460 -28.0% 112595 proc-vmstat.nr_inactive_file
26834 ± 8% +49.6% 40137 ± 10% proc-vmstat.nr_mapped
11754 ± 21% +200.7% 35348 ± 8% proc-vmstat.nr_shmem
60226 -27.6% 43625 proc-vmstat.nr_slab_reclaimable
36950 -5.6% 34868 proc-vmstat.nr_slab_unreclaimable
212974 -1.3% 210204 proc-vmstat.nr_written
213612 +11.8% 238855 proc-vmstat.nr_zone_active_anon
156460 -28.0% 112595 proc-vmstat.nr_zone_inactive_file
1933 ±131% +2667.4% 53507 ± 10% proc-vmstat.numa_hint_faults
1881 ±136% +1483.1% 29778 ± 17% proc-vmstat.numa_hint_faults_local
1251477 +58.6% 1984851 proc-vmstat.numa_hit
1201787 +60.9% 1933842 proc-vmstat.numa_local
44.50 ± 97% +17611.6% 7881 ± 45% proc-vmstat.numa_pages_migrated
104657 ± 37% +114.9% 224925 ± 28% proc-vmstat.numa_pte_updates
1702749 +43.7% 2446144 proc-vmstat.pgalloc_normal
296195 ± 2% +22.1% 361669 ± 2% proc-vmstat.pgfault
1182870 ± 5% +53.0% 1809606 proc-vmstat.pgfree
44.50 ± 97% +17611.6% 7881 ± 45% proc-vmstat.pgmigrate_success
4105642 +2.2% 4195938 proc-vmstat.pgpgout
1.28 ± 4% +11.1% 1.43 perf-stat.i.MPKI
1.739e+09 +34.1% 2.331e+09 perf-stat.i.branch-instructions
6.12 -1.1 5.00 perf-stat.i.branch-miss-rate%
1.039e+08 +9.3% 1.136e+08 perf-stat.i.branch-misses
10490708 ± 4% +46.2% 15335649 perf-stat.i.cache-misses
1.682e+08 +46.6% 2.467e+08 perf-stat.i.cache-references
212706 +77.6% 377834 perf-stat.i.context-switches
1.61 +6.3% 1.71 perf-stat.i.cpi
1.309e+10 +40.3% 1.836e+10 perf-stat.i.cpu-cycles
2446 ± 11% +59.1% 3891 ± 3% perf-stat.i.cpu-migrations
1267 ± 4% -5.2% 1201 perf-stat.i.cycles-between-cache-misses
8.326e+09 +32.8% 1.106e+10 perf-stat.i.instructions
0.63 -5.3% 0.60 perf-stat.i.ipc
4.47 +77.3% 7.93 perf-stat.i.metric.K/sec
5993 ± 3% +21.2% 7266 ± 3% perf-stat.i.minor-faults
5994 ± 3% +21.2% 7266 ± 3% perf-stat.i.page-faults
1.26 ± 4% +10.1% 1.39 perf-stat.overall.MPKI
5.98 -1.1 4.87 perf-stat.overall.branch-miss-rate%
1.57 +5.6% 1.66 perf-stat.overall.cpi
1249 ± 4% -4.2% 1197 perf-stat.overall.cycles-between-cache-misses
0.64 -5.3% 0.60 perf-stat.overall.ipc
1.69e+09 +34.4% 2.271e+09 perf-stat.ps.branch-instructions
1.01e+08 +9.5% 1.106e+08 perf-stat.ps.branch-misses
10198918 ± 4% +46.5% 14941127 perf-stat.ps.cache-misses
1.636e+08 +47.0% 2.404e+08 perf-stat.ps.cache-references
206791 +78.0% 368107 perf-stat.ps.context-switches
1.272e+10 +40.6% 1.788e+10 perf-stat.ps.cpu-cycles
2378 ± 11% +59.4% 3791 ± 3% perf-stat.ps.cpu-migrations
8.094e+09 +33.1% 1.077e+10 perf-stat.ps.instructions
5824 ± 3% +21.5% 7077 ± 3% perf-stat.ps.minor-faults
5824 ± 3% +21.5% 7077 ± 3% perf-stat.ps.page-faults
2.924e+11 ± 2% +43.5% 4.197e+11 perf-stat.total.instructions
0.00 ±223% +1020.0% 0.01 ± 59% perf-sched.sch_delay.avg.ms.__cond_resched.down_read.walk_component.link_path_walk.part
0.00 ±223% +560.0% 0.01 ± 37% perf-sched.sch_delay.avg.ms.__cond_resched.jbd2_log_wait_commit.ext4_sync_file.ext4_buffered_write_iter.do_iter_readv_writev
0.01 ± 7% -26.4% 0.01 ± 21% perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
0.01 -20.0% 0.00 perf-sched.sch_delay.avg.ms.__lock_sock.lock_sock_nested.tcp_recvmsg.inet6_recvmsg
0.01 -20.0% 0.00 perf-sched.sch_delay.avg.ms.__lock_sock.lock_sock_nested.tcp_sendmsg.sock_sendmsg
0.01 -30.0% 0.00 ± 14% perf-sched.sch_delay.avg.ms.__lock_sock.lock_sock_nested.tcp_sock_set_cork.xs_tcp_send_request
0.01 -11.1% 0.01 perf-sched.sch_delay.avg.ms.io_schedule.folio_wait_bit_common.folio_wait_writeback.__filemap_fdatawait_range
0.01 ± 74% +84.8% 0.01 ± 17% perf-sched.sch_delay.avg.ms.io_schedule.rq_qos_wait.wbt_wait.__rq_qos_throttle
0.02 ± 39% +78.6% 0.03 ± 34% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown].[unknown]
0.01 ± 5% -30.2% 0.01 perf-sched.sch_delay.avg.ms.jbd2_log_wait_commit.ext4_nfs_commit_metadata.nfsd_create_setattr.nfsd4_create_file
0.01 -22.2% 0.01 perf-sched.sch_delay.avg.ms.rpc_wait_bit_killable.__wait_on_bit.out_of_line_wait_on_bit.__rpc_execute
0.01 -12.5% 0.01 perf-sched.sch_delay.avg.ms.rpc_wait_bit_killable.__wait_on_bit.out_of_line_wait_on_bit.nfs4_do_close
0.01 ± 4% -20.3% 0.01 ± 5% perf-sched.sch_delay.avg.ms.rpc_wait_bit_killable.__wait_on_bit.out_of_line_wait_on_bit.nfs4_run_open_task
0.01 ± 10% -44.6% 0.01 ± 7% perf-sched.sch_delay.avg.ms.schedule_preempt_disabled.__mutex_lock.constprop.0.svc_tcp_sendto
0.01 ± 14% -22.4% 0.01 ± 12% perf-sched.sch_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
0.01 ± 5% -50.0% 0.01 perf-sched.sch_delay.avg.ms.svc_recv.nfsd.kthread.ret_from_fork
0.01 ±117% +245.2% 0.02 ± 36% perf-sched.sch_delay.max.ms.__cond_resched.__release_sock.release_sock.tcp_recvmsg.inet6_recvmsg
0.00 ±223% +710.3% 0.04 ± 31% perf-sched.sch_delay.max.ms.__cond_resched.__rpc_execute.rpc_execute.rpc_run_task.nfs4_call_sync_sequence
0.00 ±223% +1520.0% 0.01 ± 75% perf-sched.sch_delay.max.ms.__cond_resched.down_read.walk_component.link_path_walk.part
0.00 ±223% +600.0% 0.01 ± 26% perf-sched.sch_delay.max.ms.__cond_resched.jbd2_log_wait_commit.ext4_sync_file.ext4_buffered_write_iter.do_iter_readv_writev
0.00 ±126% +359.1% 0.02 ± 73% perf-sched.sch_delay.max.ms.__cond_resched.lock_sock_nested.tcp_sendmsg.sock_sendmsg.svc_tcp_sendmsg
0.00 ±169% +769.2% 0.02 ± 80% perf-sched.sch_delay.max.ms.__cond_resched.mutex_lock.svc_tcp_sendto.svc_send.svc_handle_xprt
0.01 ± 85% +186.7% 0.03 ± 30% perf-sched.sch_delay.max.ms.__cond_resched.xs_stream_data_receive_workfn.process_one_work.worker_thread.kthread
0.06 ± 34% -33.1% 0.04 ± 32% perf-sched.sch_delay.max.ms.jbd2_log_wait_commit.ext4_nfs_commit_metadata.nfsd_create_setattr.nfsd_create_locked
3.59 ± 11% -52.9% 1.69 ± 75% perf-sched.sch_delay.max.ms.rpc_wait_bit_killable.__wait_on_bit.out_of_line_wait_on_bit.nfs4_do_close
0.09 ± 34% +3544.0% 3.30 ± 42% perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.__synchronize_srcu.part.0
0.01 ± 25% +50.0% 0.01 ± 14% perf-sched.sch_delay.max.ms.start_this_handle.jbd2_journal_start_reserved.__ext4_journal_start_reserved.ext4_convert_unwritten_io_end_vec
0.02 ± 51% +89.1% 0.03 ± 21% perf-sched.sch_delay.max.ms.wait_transaction_locked.add_transaction_credits.start_this_handle.jbd2__journal_start
0.01 ± 5% -16.3% 0.01 perf-sched.total_sch_delay.average.ms
2.35 ± 4% -40.7% 1.39 perf-sched.total_wait_and_delay.average.ms
473705 ± 6% +83.6% 869820 perf-sched.total_wait_and_delay.count.ms
2.34 ± 4% -40.8% 1.39 perf-sched.total_wait_time.average.ms
14.98 ± 6% -6.8% 13.95 perf-sched.wait_and_delay.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
0.14 ± 3% -48.0% 0.07 ± 2% perf-sched.wait_and_delay.avg.ms.__lock_sock.lock_sock_nested.tcp_recvmsg.inet6_recvmsg
0.11 ± 5% -53.7% 0.05 ± 3% perf-sched.wait_and_delay.avg.ms.__lock_sock.lock_sock_nested.tcp_sendmsg.sock_sendmsg
0.16 ± 4% -55.2% 0.07 ± 40% perf-sched.wait_and_delay.avg.ms.__lock_sock.lock_sock_nested.tcp_sock_set_cork.xs_tcp_send_request
21.65 ± 14% +50.9% 32.67 perf-sched.wait_and_delay.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call
1.74 ± 2% -36.1% 1.11 perf-sched.wait_and_delay.avg.ms.io_schedule.folio_wait_bit_common.folio_wait_writeback.__filemap_fdatawait_range
0.22 ± 3% -20.6% 0.18 ± 2% perf-sched.wait_and_delay.avg.ms.jbd2_log_wait_commit.ext4_nfs_commit_metadata.nfsd_create_setattr.nfsd4_create_file
1.00 -58.9% 0.41 perf-sched.wait_and_delay.avg.ms.rpc_wait_bit_killable.__wait_on_bit.out_of_line_wait_on_bit.__rpc_execute
3.31 ± 2% -33.1% 2.22 perf-sched.wait_and_delay.avg.ms.rpc_wait_bit_killable.__wait_on_bit.out_of_line_wait_on_bit.nfs4_do_close
1.99 ± 2% -53.0% 0.94 perf-sched.wait_and_delay.avg.ms.rpc_wait_bit_killable.__wait_on_bit.out_of_line_wait_on_bit.nfs4_run_open_task
91.01 ± 16% -50.9% 44.65 perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
246.90 ± 5% -39.0% 150.63 ± 3% perf-sched.wait_and_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
0.42 ± 3% -60.8% 0.16 ± 2% perf-sched.wait_and_delay.avg.ms.svc_recv.nfsd.kthread.ret_from_fork
0.83 ±107% +340.0% 3.67 ± 46% perf-sched.wait_and_delay.count.__cond_resched.mutex_lock.srcu_gp_end.process_srcu.process_one_work
24.83 ± 19% +59.1% 39.50 ± 16% perf-sched.wait_and_delay.count.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
62182 ± 7% +156.7% 159647 perf-sched.wait_and_delay.count.__lock_sock.lock_sock_nested.tcp_recvmsg.inet6_recvmsg
3956 ± 7% +206.9% 12142 ± 2% perf-sched.wait_and_delay.count.__lock_sock.lock_sock_nested.tcp_sendmsg.sock_sendmsg
8707 ± 6% +320.1% 36582 ± 2% perf-sched.wait_and_delay.count.__lock_sock.lock_sock_nested.tcp_sock_set_cork.xs_tcp_send_request
26809 ± 6% -9.6% 24244 perf-sched.wait_and_delay.count.io_schedule.folio_wait_bit_common.folio_wait_writeback.__filemap_fdatawait_range
26615 ± 7% -12.6% 23273 perf-sched.wait_and_delay.count.jbd2_log_wait_commit.ext4_nfs_commit_metadata.nfsd_create_setattr.nfsd4_create_file
26752 ± 6% -11.5% 23677 perf-sched.wait_and_delay.count.jbd2_log_wait_commit.ext4_sync_file.ext4_buffered_write_iter.do_iter_readv_writev
30657 ± 7% +600.8% 214838 perf-sched.wait_and_delay.count.rpc_wait_bit_killable.__wait_on_bit.out_of_line_wait_on_bit.__rpc_execute
13405 ± 6% -9.6% 12116 perf-sched.wait_and_delay.count.rpc_wait_bit_killable.__wait_on_bit.out_of_line_wait_on_bit.nfs4_do_close
13399 ± 6% -9.5% 12122 perf-sched.wait_and_delay.count.rpc_wait_bit_killable.__wait_on_bit.out_of_line_wait_on_bit.nfs4_run_open_task
207.17 ± 5% +74.8% 362.17 perf-sched.wait_and_delay.count.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
1532 ± 6% +88.8% 2892 ± 2% perf-sched.wait_and_delay.count.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
34990 ± 7% +154.4% 89028 ± 2% perf-sched.wait_and_delay.count.svc_recv.nfsd.kthread.ret_from_fork
192563 ± 6% +15.3% 222007 perf-sched.wait_and_delay.count.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
27.54 ±107% +2216.5% 638.01 ±179% perf-sched.wait_and_delay.max.ms.__lock_sock.lock_sock_nested.tcp_sock_set_cork.xs_tcp_send_request
60.18 ±105% -91.6% 5.05 perf-sched.wait_and_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
3455 ± 22% -48.1% 1792 ± 25% perf-sched.wait_and_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
203.43 ± 94% -95.2% 9.74 ± 7% perf-sched.wait_and_delay.max.ms.svc_recv.nfsd.kthread.ret_from_fork
14.95 ± 6% -6.8% 13.93 perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
0.00 ±223% +46408.3% 0.93 ± 21% perf-sched.wait_time.avg.ms.__cond_resched.down_read.walk_component.link_path_walk.part
0.30 ± 2% -16.0% 0.25 perf-sched.wait_time.avg.ms.__cond_resched.jbd2_journal_commit_transaction.kjournald2.kthread.ret_from_fork
0.01 ±223% +1045.5% 0.15 ± 70% perf-sched.wait_time.avg.ms.__cond_resched.jbd2_log_wait_commit.ext4_sync_file.ext4_buffered_write_iter.do_iter_readv_writev
0.01 ± 23% +52.4% 0.02 ± 19% perf-sched.wait_time.avg.ms.__cond_resched.xs_stream_data_receive_workfn.process_one_work.worker_thread.kthread
0.14 ± 3% -49.1% 0.07 ± 2% perf-sched.wait_time.avg.ms.__lock_sock.lock_sock_nested.tcp_recvmsg.inet6_recvmsg
0.11 ± 6% -55.1% 0.05 ± 4% perf-sched.wait_time.avg.ms.__lock_sock.lock_sock_nested.tcp_sendmsg.sock_sendmsg
0.15 ± 4% -56.1% 0.07 ± 43% perf-sched.wait_time.avg.ms.__lock_sock.lock_sock_nested.tcp_sock_set_cork.xs_tcp_send_request
21.65 ± 14% +50.9% 32.67 perf-sched.wait_time.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.x64_sys_call
1.73 ± 2% -36.2% 1.11 perf-sched.wait_time.avg.ms.io_schedule.folio_wait_bit_common.folio_wait_writeback.__filemap_fdatawait_range
0.20 ± 3% +17.3% 0.24 ± 3% perf-sched.wait_time.avg.ms.jbd2_journal_wait_updates.jbd2_journal_commit_transaction.kjournald2.kthread
0.22 ± 3% -20.5% 0.17 ± 2% perf-sched.wait_time.avg.ms.jbd2_log_wait_commit.ext4_nfs_commit_metadata.nfsd_create_setattr.nfsd4_create_file
0.19 ± 5% -19.3% 0.16 ± 13% perf-sched.wait_time.avg.ms.jbd2_log_wait_commit.ext4_nfs_commit_metadata.nfsd_create_setattr.nfsd_create_locked
0.99 ± 2% -59.2% 0.40 perf-sched.wait_time.avg.ms.rpc_wait_bit_killable.__wait_on_bit.out_of_line_wait_on_bit.__rpc_execute
3.30 ± 2% -33.1% 2.21 perf-sched.wait_time.avg.ms.rpc_wait_bit_killable.__wait_on_bit.out_of_line_wait_on_bit.nfs4_do_close
1.98 ± 2% -53.2% 0.93 perf-sched.wait_time.avg.ms.rpc_wait_bit_killable.__wait_on_bit.out_of_line_wait_on_bit.nfs4_run_open_task
91.01 ± 16% -50.9% 44.65 perf-sched.wait_time.avg.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
246.89 ± 5% -39.0% 150.61 ± 3% perf-sched.wait_time.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
0.41 ± 3% -61.1% 0.16 ± 2% perf-sched.wait_time.avg.ms.svc_recv.nfsd.kthread.ret_from_fork
3.64 ± 10% -65.9% 1.24 ±103% perf-sched.wait_time.max.ms.__cond_resched.__kmalloc_cache_noprof.nfs_get_lock_context.nfs_page_create_from_folio.nfs_writepage_setup
0.02 ±126% +1465.3% 0.26 ± 81% perf-sched.wait_time.max.ms.__cond_resched.__release_sock.release_sock.tcp_recvmsg.inet6_recvmsg
0.30 ± 95% +190.8% 0.88 ± 12% perf-sched.wait_time.max.ms.__cond_resched.__release_sock.release_sock.tcp_sendmsg.sock_sendmsg
0.31 ±223% +557.0% 2.06 ± 19% perf-sched.wait_time.max.ms.__cond_resched.__rpc_execute.rpc_execute.rpc_run_task.nfs4_call_sync_sequence
0.00 ±223% +54158.3% 1.09 ± 31% perf-sched.wait_time.max.ms.__cond_resched.down_read.walk_component.link_path_walk.part
0.01 ±223% +1262.3% 0.17 ± 73% perf-sched.wait_time.max.ms.__cond_resched.jbd2_log_wait_commit.ext4_sync_file.ext4_buffered_write_iter.do_iter_readv_writev
0.11 ±203% +600.3% 0.78 ± 41% perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_noprof.prepare_creds.nfsd_setuser.nfsd_setuser_and_check_port
0.02 ± 73% +2366.0% 0.44 ±161% perf-sched.wait_time.max.ms.__cond_resched.lock_sock_nested.tcp_recvmsg.inet6_recvmsg.sock_recvmsg
0.11 ±213% +698.2% 0.88 ± 6% perf-sched.wait_time.max.ms.__cond_resched.lock_sock_nested.tcp_sendmsg.sock_sendmsg.svc_tcp_sendmsg
0.01 ± 33% +451.2% 0.08 ± 74% perf-sched.wait_time.max.ms.__cond_resched.xs_stream_data_receive_workfn.process_one_work.worker_thread.kthread
27.53 ±107% +2217.8% 638.00 ±179% perf-sched.wait_time.max.ms.__lock_sock.lock_sock_nested.tcp_sock_set_cork.xs_tcp_send_request
0.46 ± 10% +218.9% 1.48 ± 71% perf-sched.wait_time.max.ms.jbd2_journal_wait_updates.jbd2_journal_commit_transaction.kjournald2.kthread
2.69 +11.4% 2.99 ± 6% perf-sched.wait_time.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
60.15 ±105% -91.7% 5.02 perf-sched.wait_time.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
3455 ± 22% -48.1% 1792 ± 25% perf-sched.wait_time.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
203.35 ± 94% -95.2% 9.67 ± 7% perf-sched.wait_time.max.ms.svc_recv.nfsd.kthread.ret_from_fork
***************************************************************************************************
lkp-ivb-2ep2: 48 threads 2 sockets Intel(R) Xeon(R) CPU E5-2697 v2 @ 2.70GHz (Ivy Bridge-EP) with 64G memory
=========================================================================================
compiler/cpufreq_governor/disk/filesize/fs2/fs/iterations/kconfig/nr_directories/nr_files_per_directory/nr_threads/rootfs/sync_method/tbox_group/test_size/testcase:
gcc-12/performance/1SSD/9B/nfsv4/btrfs/1x/x86_64-rhel-9.4/16d/256fpd/32t/debian-12-x86_64-20240206.cgz/fsyncBeforeClose/lkp-ivb-2ep2/400M/fsmark
commit:
66f9dac907 ("Revert "nfs: don't reuse partially completed requests in nfs_lock_and_join_requests"")
b6dea6c7fe ("nfs: pass flags to second superblock")
66f9dac9077c9c06 b6dea6c7fe2d8187050f882fe6f
---------------- ---------------------------
%stddev %change %stddev
\ | \
1.188e+09 +11.3% 1.322e+09 cpuidle..time
3929892 +91.6% 7529218 cpuidle..usage
69.88 +7.4% 75.03 iostat.cpu.idle
22.22 -34.0% 14.67 iostat.cpu.iowait
5.52 ± 3% +43.8% 7.94 iostat.cpu.system
786761 ± 9% +24.2% 977255 ± 8% numa-numastat.node0.local_node
804849 ± 9% +23.7% 995380 ± 7% numa-numastat.node0.numa_hit
916826 ± 8% +55.5% 1425647 ± 5% numa-numastat.node1.local_node
948417 ± 8% +53.7% 1457243 ± 4% numa-numastat.node1.numa_hit
23.86 -8.3 15.58 mpstat.cpu.all.iowait%
0.30 +0.0 0.33 mpstat.cpu.all.irq%
0.58 +0.3 0.93 mpstat.cpu.all.soft%
4.95 ± 3% +2.2 7.10 mpstat.cpu.all.sys%
11.40 ± 2% +23.4% 14.06 ± 2% mpstat.max_utilization_pct
1.358e+08 ± 2% +73.7% 2.359e+08 fsmark.app_overhead
4107 -15.9% 3456 fsmark.files_per_sec
6386 ± 9% -9.9% 5754 fsmark.time.maximum_resident_set_size
49.17 +128.5% 112.33 fsmark.time.percent_of_cpu_this_job_got
11.25 +185.3% 32.10 fsmark.time.system_time
575615 +300.3% 2304344 fsmark.time.voluntary_context_switches
13811 ± 2% -17.8% 11354 meminfo.Dirty
1073015 -12.6% 938340 meminfo.Inactive
1073015 -12.6% 938340 meminfo.Inactive(file)
219309 -30.5% 152451 meminfo.KReclaimable
219309 -30.5% 152451 meminfo.SReclaimable
387968 -19.8% 311283 meminfo.Slab
69.89 +7.4% 75.05 vmstat.cpu.id
22.21 -33.6% 14.75 vmstat.cpu.wa
12.73 ± 4% -39.9% 7.65 ± 6% vmstat.procs.b
4.99 ± 10% +21.9% 6.08 ± 5% vmstat.procs.r
216786 +92.3% 416945 vmstat.system.cs
53985 +16.2% 62753 vmstat.system.in
5358 ± 13% -19.4% 4316 ± 7% numa-meminfo.node0.Dirty
463057 ± 9% -25.2% 346146 ± 9% numa-meminfo.node0.Inactive
463057 ± 9% -25.2% 346146 ± 9% numa-meminfo.node0.Inactive(file)
8474 ± 8% -16.8% 7049 ± 6% numa-meminfo.node1.Dirty
90297 ± 23% -37.0% 56862 ± 7% numa-meminfo.node1.KReclaimable
90297 ± 23% -37.0% 56862 ± 7% numa-meminfo.node1.SReclaimable
169041 ± 12% -21.1% 133345 ± 2% numa-meminfo.node1.Slab
1339 ± 12% -19.4% 1079 ± 7% numa-vmstat.node0.nr_dirty
115964 ± 9% -25.6% 86318 ± 9% numa-vmstat.node0.nr_inactive_file
115964 ± 9% -25.6% 86318 ± 9% numa-vmstat.node0.nr_zone_inactive_file
1192 ± 12% -28.5% 852.16 ± 10% numa-vmstat.node0.nr_zone_write_pending
804755 ± 9% +23.5% 994145 ± 7% numa-vmstat.node0.numa_hit
786667 ± 9% +24.1% 976013 ± 8% numa-vmstat.node0.numa_local
338555 ± 7% +15.7% 391647 ± 5% numa-vmstat.node1.nr_dirtied
2120 ± 8% -16.7% 1766 ± 6% numa-vmstat.node1.nr_dirty
22647 ± 23% -37.3% 14193 ± 8% numa-vmstat.node1.nr_slab_reclaimable
338299 ± 8% +15.3% 390105 ± 5% numa-vmstat.node1.nr_written
2027 ± 10% -19.4% 1634 ± 8% numa-vmstat.node1.nr_zone_write_pending
947862 ± 8% +53.5% 1454904 ± 4% numa-vmstat.node1.numa_hit
916271 ± 8% +55.3% 1423309 ± 5% numa-vmstat.node1.numa_local
600332 +9.7% 658719 proc-vmstat.nr_dirtied
3457 -17.7% 2844 ± 2% proc-vmstat.nr_dirty
1175196 -2.8% 1141722 proc-vmstat.nr_file_pages
268553 -12.5% 234871 proc-vmstat.nr_inactive_file
54906 -30.5% 38145 proc-vmstat.nr_slab_reclaimable
42192 -5.9% 39710 proc-vmstat.nr_slab_unreclaimable
599793 +9.4% 656108 proc-vmstat.nr_written
268553 -12.5% 234871 proc-vmstat.nr_zone_inactive_file
3187 ± 3% -23.3% 2444 ± 5% proc-vmstat.nr_zone_write_pending
1754966 +39.9% 2454658 proc-vmstat.numa_hit
1709407 +41.7% 2421514 proc-vmstat.numa_local
2209705 +31.8% 2912839 proc-vmstat.pgalloc_normal
1655605 ± 3% +39.7% 2312933 proc-vmstat.pgfree
3264798 +14.5% 3739381 proc-vmstat.pgpgout
0.02 ± 97% +118.2% 0.04 ± 31% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
2.40 ±192% -98.4% 0.04 ±144% perf-sched.sch_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
0.11 ± 20% -64.1% 0.04 ± 91% perf-sched.sch_delay.max.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
0.01 ±223% +588.9% 0.05 ± 36% perf-sched.sch_delay.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
0.03 ± 88% +232.1% 0.10 ± 36% perf-sched.sch_delay.max.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
96.58 ± 63% +711.2% 783.54 ± 44% perf-sched.total_wait_and_delay.max.ms
62.71 ± 53% +1124.4% 767.88 ± 44% perf-sched.total_wait_time.max.ms
0.95 ± 60% +370.2% 4.47 ± 35% perf-sched.wait_and_delay.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
6.05 ± 77% -96.3% 0.22 ±223% perf-sched.wait_and_delay.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
0.87 ± 40% -82.8% 0.15 ±141% perf-sched.wait_and_delay.avg.ms.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe
1.67 ± 43% -90.2% 0.16 ±102% perf-sched.wait_and_delay.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
3.38 ± 13% +174.6% 9.29 ± 32% perf-sched.wait_and_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
8.07 ±109% +1338.7% 116.15 ± 23% perf-sched.wait_and_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
4.33 ± 52% -92.3% 0.33 ±223% perf-sched.wait_and_delay.count.devkmsg_read.vfs_read.ksys_read.do_syscall_64
1.17 ±104% +957.1% 12.33 ± 57% perf-sched.wait_and_delay.count.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
94.18 ± 68% +533.2% 596.36 ± 43% perf-sched.wait_and_delay.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
15.07 ±103% -97.1% 0.43 ±223% perf-sched.wait_and_delay.max.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
6.60 ± 99% -93.2% 0.45 ±141% perf-sched.wait_and_delay.max.ms.do_wait.kernel_wait4.do_syscall_64.entry_SYSCALL_64_after_hwframe
4.18 ± 8% +5513.8% 234.70 ± 49% perf-sched.wait_and_delay.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
61.50 ± 50% +957.2% 650.14 ± 46% perf-sched.wait_and_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
0.48 ± 60% +777.3% 4.17 ± 35% perf-sched.wait_time.avg.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
1.61 ± 44% -82.4% 0.28 ± 17% perf-sched.wait_time.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
3.02 ± 18% +206.6% 9.26 ± 32% perf-sched.wait_time.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
7.94 ±112% +1360.8% 116.00 ± 23% perf-sched.wait_time.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
47.09 ± 68% +1166.4% 596.35 ± 43% perf-sched.wait_time.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
0.03 ±105% +3091.2% 0.90 ±126% perf-sched.wait_time.max.ms.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown]
4.16 ± 8% +5536.1% 234.65 ± 49% perf-sched.wait_time.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
61.02 ± 51% +937.2% 632.92 ± 45% perf-sched.wait_time.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
1.32 +19.1% 1.57 perf-stat.i.MPKI
1.869e+09 +22.9% 2.298e+09 perf-stat.i.branch-instructions
5.44 -0.7 4.74 perf-stat.i.branch-miss-rate%
1.006e+08 +7.9% 1.086e+08 perf-stat.i.branch-misses
11843942 ± 2% +46.2% 17314122 perf-stat.i.cache-misses
1.886e+08 +46.1% 2.755e+08 perf-stat.i.cache-references
239152 +87.2% 447643 perf-stat.i.context-switches
1.58 +10.8% 1.75 perf-stat.i.cpi
1.418e+10 +35.9% 1.927e+10 perf-stat.i.cpu-cycles
1842 ± 2% +73.5% 3195 perf-stat.i.cpu-migrations
1202 ± 2% -7.5% 1112 perf-stat.i.cycles-between-cache-misses
9.107e+09 +21.9% 1.11e+10 perf-stat.i.instructions
0.65 -9.8% 0.58 perf-stat.i.ipc
5.03 +86.7% 9.40 perf-stat.i.metric.K/sec
5735 ± 4% -10.8% 5115 ± 3% perf-stat.i.minor-faults
5735 ± 4% -10.8% 5116 ± 3% perf-stat.i.page-faults
1.30 +19.9% 1.56 perf-stat.overall.MPKI
5.38 -0.7 4.73 perf-stat.overall.branch-miss-rate%
1.56 +11.5% 1.74 perf-stat.overall.cpi
1197 -7.1% 1113 perf-stat.overall.cycles-between-cache-misses
0.64 -10.3% 0.58 perf-stat.overall.ipc
1.799e+09 +23.6% 2.223e+09 perf-stat.ps.branch-instructions
96805423 +8.5% 1.051e+08 perf-stat.ps.branch-misses
11396718 ± 2% +47.0% 16750043 perf-stat.ps.cache-misses
1.815e+08 +46.9% 2.665e+08 perf-stat.ps.cache-references
230141 +88.2% 433076 perf-stat.ps.context-switches
1.364e+10 +36.6% 1.864e+10 perf-stat.ps.cpu-cycles
1772 ± 2% +74.4% 3091 perf-stat.ps.cpu-migrations
8.763e+09 +22.6% 1.074e+10 perf-stat.ps.instructions
5514 ± 4% -10.3% 4945 ± 3% perf-stat.ps.minor-faults
5514 ± 4% -10.3% 4945 ± 3% perf-stat.ps.page-faults
2.33e+11 +41.9% 3.305e+11 perf-stat.total.instructions
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
--
0-DAY CI Kernel Test Service
https://github.com/intel/lkp-tests/wiki
^ permalink raw reply [flat|nested] only message in thread
only message in thread, other threads:[~2024-11-30 10:44 UTC | newest]
Thread overview: (only message) (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2024-11-30 10:44 [linux-next:master] [nfs] b6dea6c7fe: fsmark.files_per_sec 17.6% regression kernel test robot
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox