From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1756719AbcEaIeo (ORCPT ); Tue, 31 May 2016 04:34:44 -0400 Received: from mga01.intel.com ([192.55.52.88]:60526 "EHLO mga01.intel.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1756224AbcEaIek (ORCPT ); Tue, 31 May 2016 04:34:40 -0400 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.26,395,1459839600"; d="scan'208";a="965626062" From: "Huang\, Ying" To: Ingo Molnar Cc: Peter Zijlstra , Peter Zijlstra , , Mike Galbraith , , Thomas Gleixner , "Linus Torvalds" Subject: Re: [LKP] [lkp] [sched/fair] 53d3bc773e: hackbench.throughput -32.9% regression References: <87inxud4ex.fsf@yhuang-dev.intel.com> Date: Tue, 31 May 2016 16:34:36 +0800 In-Reply-To: <87inxud4ex.fsf@yhuang-dev.intel.com> (kernel test robot's message of "Tue, 31 May 2016 16:20:54 +0800") Message-ID: <87eg8id3s3.fsf@yhuang-dev.intel.com> User-Agent: Gnus/5.13 (Gnus v5.13) Emacs/24.5 (gnu/linux) MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Transfer-Encoding: 8bit Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Hi, Ingo, Part of the regression has been recovered in v4.7-rc1 from -32.9% to -9.8%. But there is still some regression. Is it possible for fully restore it? Details are as below. Best Regards, Huang, Ying ========================================================================================= compiler/cpufreq_governor/ipc/kconfig/mode/nr_threads/rootfs/tbox_group/testcase: gcc-4.9/performance/socket/x86_64-rhel/threads/50%/debian-x86_64-2015-02-07.cgz/ivb42/hackbench commit: c5114626f33b62fa7595e57d87f33d9d1f8298a2 53d3bc773eaa7ab1cf63585e76af7ee869d5e709 v4.7-rc1 c5114626f33b62fa 53d3bc773eaa7ab1cf63585e76 v4.7-rc1 ---------------- -------------------------- -------------------------- %stddev %change %stddev %change %stddev \ | \ | \ 196590 ± 0% -32.9% 131963 ± 2% -9.8% 177231 ± 0% hackbench.throughput 602.66 ± 0% +2.8% 619.27 ± 2% +0.3% 604.66 ± 0% hackbench.time.elapsed_time 602.66 ± 0% +2.8% 619.27 ± 2% +0.3% 604.66 ± 0% hackbench.time.elapsed_time.max 1.76e+08 ± 3% +236.0% 5.914e+08 ± 2% -49.6% 88783232 ± 5% hackbench.time.involuntary_context_switches 208664 ± 2% +26.0% 262929 ± 3% +15.7% 241377 ± 0% hackbench.time.minor_page_faults 4401 ± 0% +5.7% 4650 ± 0% -8.1% 4043 ± 0% hackbench.time.percent_of_cpu_this_job_got 25256 ± 0% +10.2% 27842 ± 2% -7.7% 23311 ± 0% hackbench.time.system_time 1272 ± 0% -24.5% 961.37 ± 2% -10.4% 1140 ± 0% hackbench.time.user_time 7.64e+08 ± 1% +131.8% 1.771e+09 ± 2% -30.1% 5.339e+08 ± 2% hackbench.time.voluntary_context_switches 4051 ± 0% -39.9% 2434 ± 3% +57.8% 6393 ± 0% uptime.idle 4337715 ± 1% +7.3% 4654464 ± 2% -23.3% 3325346 ± 5% softirqs.RCU 2462880 ± 0% -35.6% 1585869 ± 5% +58.1% 3893988 ± 0% softirqs.SCHED 1766752 ± 1% +122.6% 3932589 ± 1% -25.6% 1313619 ± 1% vmstat.system.cs 249718 ± 2% +307.4% 1017398 ± 3% -40.4% 148723 ± 5% vmstat.system.in 1.76e+08 ± 3% +236.0% 5.914e+08 ± 2% -49.6% 88783232 ± 5% time.involuntary_context_switches 208664 ± 2% +26.0% 262929 ± 3% +15.7% 241377 ± 0% time.minor_page_faults 1272 ± 0% -24.5% 961.37 ± 2% -10.4% 1140 ± 0% time.user_time 7.64e+08 ± 1% +131.8% 1.771e+09 ± 2% -30.1% 5.339e+08 ± 2% time.voluntary_context_switches 177383 ± 0% +2.0% 180939 ± 0% -51.3% 86390 ± 1% meminfo.Active 102033 ± 0% -0.1% 101893 ± 1% -85.6% 14740 ± 0% meminfo.Active(file) 392558 ± 0% +0.0% 392612 ± 0% +22.6% 481411 ± 0% meminfo.Inactive 382911 ± 0% +0.0% 382923 ± 0% +23.2% 471792 ± 0% meminfo.Inactive(file) 143370 ± 0% -12.0% 126124 ± 1% -1.5% 141210 ± 0% meminfo.SUnreclaim 1136461 ± 3% +16.6% 1324662 ± 5% +15.9% 1316829 ± 1% numa-numastat.node0.local_node 1140216 ± 3% +16.2% 1324689 ± 5% +15.5% 1316830 ± 1% numa-numastat.node0.numa_hit 3755 ± 68% -99.3% 27.25 ± 94% -100.0% 1.25 ± 34% numa-numastat.node0.other_node 1098889 ± 4% +20.1% 1320211 ± 6% +16.4% 1278783 ± 1% numa-numastat.node1.local_node 1101996 ± 4% +20.5% 1327590 ± 6% +16.0% 1278783 ± 1% numa-numastat.node1.numa_hit 3106 ± 99% +137.5% 7379 ± 17% -100.0% 0.00 ± -1% numa-numastat.node1.other_node 7.18 ± 0% -50.2% 3.57 ± 43% +76.1% 12.64 ± 1% perf-profile.cycles-pp.call_cpuidle 8.09 ± 0% -44.7% 4.47 ± 38% +72.4% 13.95 ± 1% perf-profile.cycles-pp.cpu_startup_entry 7.17 ± 0% -50.3% 3.56 ± 43% +76.2% 12.63 ± 1% perf-profile.cycles-pp.cpuidle_enter 7.14 ± 0% -50.3% 3.55 ± 43% +76.1% 12.58 ± 1% perf-profile.cycles-pp.cpuidle_enter_state 7.11 ± 0% -50.6% 3.52 ± 43% +76.3% 12.54 ± 1% perf-profile.cycles-pp.intel_idle 8.00 ± 0% -44.5% 4.44 ± 38% +72.1% 13.77 ± 1% perf-profile.cycles-pp.start_secondary 92.32 ± 0% +5.4% 97.32 ± 0% -7.7% 85.26 ± 0% turbostat.%Busy 2763 ± 0% +5.4% 2912 ± 0% -7.7% 2551 ± 0% turbostat.Avg_MHz 7.48 ± 0% -66.5% 2.50 ± 7% +94.5% 14.54 ± 0% turbostat.CPU%c1 0.20 ± 2% -6.4% 0.18 ± 2% +2.6% 0.20 ± 3% turbostat.CPU%c6 180.03 ± 0% -1.3% 177.62 ± 0% -2.4% 175.63 ± 0% turbostat.CorWatt 209.86 ± 0% -0.8% 208.08 ± 0% -2.0% 205.64 ± 0% turbostat.PkgWatt 5.83 ± 0% +38.9% 8.10 ± 3% +12.7% 6.57 ± 1% turbostat.RAMWatt 1.658e+09 ± 0% -59.1% 6.784e+08 ± 7% +89.3% 3.138e+09 ± 0% cpuidle.C1-IVT.time 1.066e+08 ± 0% -40.3% 63661563 ± 6% +44.3% 1.539e+08 ± 0% cpuidle.C1-IVT.usage 26348635 ± 0% -86.8% 3471048 ± 15% +50.0% 39513523 ± 0% cpuidle.C1E-IVT.time 291620 ± 0% -85.1% 43352 ± 15% +28.8% 375730 ± 1% cpuidle.C1E-IVT.usage 54158643 ± 1% -88.5% 6254009 ± 14% +78.4% 96596486 ± 1% cpuidle.C3-IVT.time 482437 ± 1% -87.0% 62620 ± 16% +45.6% 702258 ± 1% cpuidle.C3-IVT.usage 5.028e+08 ± 0% -75.8% 1.219e+08 ± 8% +85.5% 9.327e+08 ± 1% cpuidle.C6-IVT.time 3805026 ± 0% -85.5% 552326 ± 16% +49.4% 5684182 ± 1% cpuidle.C6-IVT.usage 2766 ± 4% -51.4% 1344 ± 6% +10.0% 3042 ± 7% cpuidle.POLL.usage 49725 ± 4% +2.1% 50775 ± 3% -85.2% 7360 ± 0% numa-meminfo.node0.Active(file) 2228 ± 92% +137.1% 5285 ± 15% +118.7% 4874 ± 19% numa-meminfo.node0.AnonHugePages 197699 ± 2% +1.6% 200772 ± 0% +23.9% 245042 ± 0% numa-meminfo.node0.Inactive 192790 ± 1% -0.6% 191611 ± 0% +22.3% 235849 ± 0% numa-meminfo.node0.Inactive(file) 73589 ± 4% -12.5% 64393 ± 2% -1.3% 72664 ± 2% numa-meminfo.node0.SUnreclaim 27438 ± 83% +102.6% 55585 ± 6% +83.0% 50223 ± 0% numa-meminfo.node0.Shmem 101051 ± 3% -10.9% 90044 ± 2% -1.2% 99863 ± 2% numa-meminfo.node0.Slab 89204 ± 25% -25.3% 66594 ± 4% -77.6% 19954 ± 4% numa-meminfo.node1.Active 52306 ± 3% -2.3% 51117 ± 4% -85.9% 7380 ± 0% numa-meminfo.node1.Active(file) 194864 ± 2% -1.6% 191824 ± 1% +21.3% 236372 ± 0% numa-meminfo.node1.Inactive 4742 ± 86% -89.2% 511.75 ± 41% -90.9% 430.00 ± 60% numa-meminfo.node1.Inactive(anon) 190121 ± 1% +0.6% 191311 ± 1% +24.1% 235942 ± 0% numa-meminfo.node1.Inactive(file) 69844 ± 4% -11.8% 61579 ± 3% -1.9% 68521 ± 3% numa-meminfo.node1.SUnreclaim 12430 ± 4% +2.1% 12693 ± 3% -85.2% 1839 ± 0% numa-vmstat.node0.nr_active_file 48197 ± 1% -0.6% 47902 ± 0% +22.3% 58962 ± 0% numa-vmstat.node0.nr_inactive_file 6857 ± 83% +102.8% 13905 ± 6% +83.1% 12559 ± 0% numa-vmstat.node0.nr_shmem 18395 ± 4% -12.4% 16121 ± 2% -1.1% 18187 ± 2% numa-vmstat.node0.nr_slab_unreclaimable 675569 ± 3% +12.7% 761135 ± 4% +18.8% 802726 ± 4% numa-vmstat.node0.numa_local 71537 ± 5% -7.9% 65920 ± 2% -100.0% 0.25 ±173% numa-vmstat.node0.numa_other 13076 ± 3% -2.3% 12778 ± 4% -85.9% 1844 ± 0% numa-vmstat.node1.nr_active_file 1187 ± 86% -89.3% 127.50 ± 41% -91.0% 107.25 ± 60% numa-vmstat.node1.nr_inactive_anon 47530 ± 1% +0.6% 47827 ± 1% +24.1% 58985 ± 0% numa-vmstat.node1.nr_inactive_file 17456 ± 4% -11.7% 15405 ± 3% -1.9% 17127 ± 3% numa-vmstat.node1.nr_slab_unreclaimable 695848 ± 3% +14.9% 799683 ± 5% +4.7% 728368 ± 3% numa-vmstat.node1.numa_hit 677405 ± 4% +14.5% 775903 ± 6% +7.5% 728368 ± 3% numa-vmstat.node1.numa_local 18442 ± 19% +28.9% 23779 ± 5% -100.0% 0.00 ± -1% numa-vmstat.node1.numa_other 25508 ± 0% -0.1% 25473 ± 1% -85.6% 3684 ± 0% proc-vmstat.nr_active_file 95727 ± 0% +0.0% 95730 ± 0% +23.2% 117947 ± 0% proc-vmstat.nr_inactive_file 35841 ± 0% -12.0% 31543 ± 0% -1.5% 35298 ± 0% proc-vmstat.nr_slab_unreclaimable 154090 ± 2% +43.1% 220509 ± 3% +23.5% 190284 ± 0% proc-vmstat.numa_hint_faults 129240 ± 2% +47.4% 190543 ± 3% +15.1% 148733 ± 1% proc-vmstat.numa_hint_faults_local 2238386 ± 1% +18.4% 2649737 ± 2% +15.8% 2591197 ± 0% proc-vmstat.numa_hit 2232163 ± 1% +18.4% 2643105 ± 2% +16.1% 2591195 ± 0% proc-vmstat.numa_local 6223 ± 0% +6.6% 6632 ± 10% -100.0% 1.25 ± 34% proc-vmstat.numa_other 22315 ± 1% -21.0% 17625 ± 5% -0.4% 22234 ± 0% proc-vmstat.numa_pages_migrated 154533 ± 2% +45.6% 225071 ± 3% +25.7% 194235 ± 0% proc-vmstat.numa_pte_updates 14224 ± 0% +5.5% 15006 ± 3% -17.8% 11689 ± 0% proc-vmstat.pgactivate 382980 ± 2% +33.2% 510157 ± 4% +22.0% 467358 ± 0% proc-vmstat.pgalloc_dma32 7311738 ± 2% +37.2% 10029060 ± 2% +28.2% 9374740 ± 0% proc-vmstat.pgalloc_normal 7672040 ± 2% +37.1% 10519738 ± 2% +28.0% 9823026 ± 0% proc-vmstat.pgfree 22315 ± 1% -21.0% 17625 ± 5% -0.4% 22234 ± 0% proc-vmstat.pgmigrate_success 720.75 ± 3% -11.3% 639.50 ± 1% -29.2% 510.00 ± 0% slabinfo.RAW.active_objs 720.75 ± 3% -11.3% 639.50 ± 1% -29.2% 510.00 ± 0% slabinfo.RAW.num_objs 5487 ± 6% -12.6% 4797 ± 4% -100.0% 0.00 ± -1% slabinfo.UNIX.active_objs 164.50 ± 5% -12.3% 144.25 ± 4% -100.0% 0.00 ± -1% slabinfo.UNIX.active_slabs 5609 ± 5% -12.2% 4926 ± 4% -100.0% 0.00 ± -1% slabinfo.UNIX.num_objs 164.50 ± 5% -12.3% 144.25 ± 4% -100.0% 0.00 ± -1% slabinfo.UNIX.num_slabs 4362 ± 4% +14.6% 4998 ± 2% -3.2% 4223 ± 4% slabinfo.cred_jar.active_objs 4362 ± 4% +14.6% 4998 ± 2% -3.2% 4223 ± 4% slabinfo.cred_jar.num_objs 2904 ± 4% -2.7% 2825 ± 1% +56.5% 4545 ± 2% slabinfo.kmalloc-1024.active_objs 2935 ± 2% -0.5% 2920 ± 1% +57.8% 4633 ± 2% slabinfo.kmalloc-1024.num_objs 42525 ± 0% -41.6% 24824 ± 3% +7.3% 45621 ± 0% slabinfo.kmalloc-256.active_objs 845.50 ± 0% -42.9% 482.50 ± 3% +3.0% 870.50 ± 0% slabinfo.kmalloc-256.active_slabs 54124 ± 0% -42.9% 30920 ± 3% +3.0% 55755 ± 0% slabinfo.kmalloc-256.num_objs 845.50 ± 0% -42.9% 482.50 ± 3% +3.0% 870.50 ± 0% slabinfo.kmalloc-256.num_slabs 47204 ± 0% -37.9% 29335 ± 2% +6.6% 50334 ± 0% slabinfo.kmalloc-512.active_objs 915.25 ± 0% -39.8% 551.00 ± 3% +2.8% 940.50 ± 0% slabinfo.kmalloc-512.active_slabs 58599 ± 0% -39.8% 35300 ± 3% +2.8% 60224 ± 0% slabinfo.kmalloc-512.num_objs 915.25 ± 0% -39.8% 551.00 ± 3% +2.8% 940.50 ± 0% slabinfo.kmalloc-512.num_slabs 12443 ± 2% -20.1% 9944 ± 3% -6.5% 11639 ± 1% slabinfo.pid.active_objs 12443 ± 2% -20.1% 9944 ± 3% -6.5% 11639 ± 1% slabinfo.pid.num_objs 440.00 ± 5% -32.8% 295.75 ± 4% -11.7% 388.50 ± 7% slabinfo.taskstats.active_objs 440.00 ± 5% -32.8% 295.75 ± 4% -11.7% 388.50 ± 7% slabinfo.taskstats.num_objs 188235 ± 74% +62.9% 306699 ± 27% -98.6% 2627 ± 40% sched_debug.cfs_rq:/.MIN_vruntime.avg 7146629 ± 80% +27.7% 9122933 ± 36% -98.6% 98261 ± 36% sched_debug.cfs_rq:/.MIN_vruntime.max 1117852 ± 77% +44.7% 1617052 ± 31% -98.6% 15548 ± 37% sched_debug.cfs_rq:/.MIN_vruntime.stddev 61.52 ±116% -70.6% 18.11 ± 6% +1.2e+06% 718736 ± 1% sched_debug.cfs_rq:/.load.avg 2144 ±161% -96.3% 79.41 ± 48% +49309.2% 1059411 ± 3% sched_debug.cfs_rq:/.load.max 312.45 ±157% -94.8% 16.29 ± 33% +1.1e+05% 333106 ± 5% sched_debug.cfs_rq:/.load.stddev 20.46 ± 4% +9.0% 22.31 ± 6% +3004.0% 635.15 ± 1% sched_debug.cfs_rq:/.load_avg.avg 81.57 ± 32% +14.2% 93.18 ± 26% +1035.5% 926.18 ± 3% sched_debug.cfs_rq:/.load_avg.max 8.14 ± 5% -2.8% 7.91 ± 3% +2585.8% 218.52 ± 13% sched_debug.cfs_rq:/.load_avg.min 13.90 ± 29% +16.9% 16.25 ± 22% +1089.3% 165.34 ± 5% sched_debug.cfs_rq:/.load_avg.stddev 188235 ± 74% +62.9% 306699 ± 27% -98.6% 2627 ± 40% sched_debug.cfs_rq:/.max_vruntime.avg 7146629 ± 80% +27.7% 9122933 ± 36% -98.6% 98261 ± 36% sched_debug.cfs_rq:/.max_vruntime.max 1117852 ± 77% +44.7% 1617052 ± 31% -98.6% 15548 ± 37% sched_debug.cfs_rq:/.max_vruntime.stddev 29491781 ± 0% -4.8% 28074842 ± 1% -99.0% 295426 ± 0% sched_debug.cfs_rq:/.min_vruntime.avg 31241540 ± 0% -5.8% 29418054 ± 0% -99.0% 320734 ± 0% sched_debug.cfs_rq:/.min_vruntime.max 27849652 ± 0% -3.7% 26821072 ± 2% -99.0% 275550 ± 0% sched_debug.cfs_rq:/.min_vruntime.min 861989 ± 3% -20.2% 687639 ± 22% -98.3% 14586 ± 2% sched_debug.cfs_rq:/.min_vruntime.stddev 0.27 ± 5% -56.3% 0.12 ± 30% +27.5% 0.34 ± 6% sched_debug.cfs_rq:/.nr_running.stddev 16.51 ± 1% +9.5% 18.08 ± 3% +3343.1% 568.61 ± 2% sched_debug.cfs_rq:/.runnable_load_avg.avg 34.80 ± 13% +15.0% 40.02 ± 19% +2514.0% 909.57 ± 0% sched_debug.cfs_rq:/.runnable_load_avg.max 0.05 ±100% +7950.0% 3.66 ± 48% +3250.0% 1.52 ± 89% sched_debug.cfs_rq:/.runnable_load_avg.min 7.18 ± 9% -0.1% 7.18 ± 13% +3571.2% 263.68 ± 4% sched_debug.cfs_rq:/.runnable_load_avg.stddev -740916 ±-28% -158.5% 433310 ±120% -96.8% -23579 ± -5% sched_debug.cfs_rq:/.spread0.avg 1009940 ± 19% +75.8% 1775442 ± 30% -99.8% 1736 ±164% sched_debug.cfs_rq:/.spread0.max -2384171 ± -7% -65.7% -818684 ±-76% -98.2% -43456 ± -4% sched_debug.cfs_rq:/.spread0.min 862765 ± 3% -20.4% 686825 ± 22% -98.3% 14591 ± 2% sched_debug.cfs_rq:/.spread0.stddev 749.14 ± 1% +13.0% 846.34 ± 1% -41.1% 441.05 ± 5% sched_debug.cfs_rq:/.util_avg.min 51.66 ± 4% -36.3% 32.92 ± 5% +150.6% 129.46 ± 6% sched_debug.cfs_rq:/.util_avg.stddev 161202 ± 7% -41.7% 93997 ± 4% +147.7% 399342 ± 1% sched_debug.cpu.avg_idle.avg 595158 ± 6% -51.2% 290491 ± 22% +37.8% 820120 ± 0% sched_debug.cpu.avg_idle.max 7658 ± 51% +9.2% 8366 ± 26% +114.4% 16423 ± 31% sched_debug.cpu.avg_idle.min 132760 ± 8% -58.8% 54718 ± 19% +97.8% 262608 ± 0% sched_debug.cpu.avg_idle.stddev 11.40 ± 11% +111.0% 24.05 ± 16% -58.1% 4.78 ± 3% sched_debug.cpu.clock.stddev 11.40 ± 11% +111.0% 24.05 ± 16% -58.1% 4.78 ± 3% sched_debug.cpu.clock_task.stddev 16.59 ± 1% +7.7% 17.86 ± 2% +3099.8% 530.73 ± 2% sched_debug.cpu.cpu_load[0].avg 32.34 ± 2% +23.9% 40.07 ± 19% +2715.0% 910.41 ± 0% sched_debug.cpu.cpu_load[0].max 0.34 ±103% +520.0% 2.11 ± 67% +140.0% 0.82 ±110% sched_debug.cpu.cpu_load[0].min 6.87 ± 3% +8.0% 7.42 ± 13% +4228.9% 297.50 ± 3% sched_debug.cpu.cpu_load[0].stddev 16.56 ± 0% +8.1% 17.91 ± 2% +3703.9% 630.04 ± 1% sched_debug.cpu.cpu_load[1].avg 32.18 ± 2% +22.7% 39.50 ± 17% +2728.5% 910.25 ± 0% sched_debug.cpu.cpu_load[1].max 3.32 ± 8% +84.9% 6.14 ± 12% +5364.4% 181.32 ± 9% sched_debug.cpu.cpu_load[1].min 6.14 ± 5% +12.5% 6.91 ± 13% +2708.6% 172.56 ± 5% sched_debug.cpu.cpu_load[1].stddev 16.75 ± 1% +7.6% 18.02 ± 2% +3646.9% 627.69 ± 1% sched_debug.cpu.cpu_load[2].avg 33.25 ± 7% +16.5% 38.75 ± 14% +2634.1% 909.09 ± 0% sched_debug.cpu.cpu_load[2].max 5.39 ± 7% +36.3% 7.34 ± 4% +3547.3% 196.45 ± 11% sched_debug.cpu.cpu_load[2].min 5.95 ± 9% +11.8% 6.65 ± 11% +2752.1% 169.73 ± 5% sched_debug.cpu.cpu_load[2].stddev 17.17 ± 1% +6.1% 18.22 ± 2% +3552.1% 626.96 ± 1% sched_debug.cpu.cpu_load[3].avg 33.20 ± 7% +14.6% 38.05 ± 9% +2631.3% 906.93 ± 0% sched_debug.cpu.cpu_load[3].max 6.93 ± 7% +10.5% 7.66 ± 1% +2766.9% 198.73 ± 11% sched_debug.cpu.cpu_load[3].min 5.70 ± 9% +13.9% 6.49 ± 8% +2825.6% 166.73 ± 5% sched_debug.cpu.cpu_load[3].stddev 17.49 ± 0% +4.9% 18.36 ± 2% +3482.1% 626.64 ± 1% sched_debug.cpu.cpu_load[4].avg 33.18 ± 3% +14.0% 37.82 ± 5% +2615.8% 901.16 ± 0% sched_debug.cpu.cpu_load[4].max 7.66 ± 8% +0.9% 7.73 ± 1% +2568.8% 204.41 ± 11% sched_debug.cpu.cpu_load[4].min 5.56 ± 6% +16.2% 6.45 ± 6% +2814.9% 161.96 ± 6% sched_debug.cpu.cpu_load[4].stddev 16741 ± 0% -15.4% 14166 ± 2% -13.0% 14564 ± 2% sched_debug.cpu.curr->pid.avg 19196 ± 0% -18.3% 15690 ± 1% -4.9% 18255 ± 0% sched_debug.cpu.curr->pid.max 5174 ± 5% -55.4% 2305 ± 14% +19.3% 6173 ± 6% sched_debug.cpu.curr->pid.stddev 18.60 ± 5% -2.7% 18.10 ± 6% +3.9e+06% 717646 ± 2% sched_debug.cpu.load.avg 81.23 ± 48% -2.4% 79.30 ± 47% +1.3e+06% 1059340 ± 3% sched_debug.cpu.load.max 18.01 ± 28% -9.4% 16.32 ± 33% +1.9e+06% 333436 ± 5% sched_debug.cpu.load.stddev 0.00 ± 2% +29.8% 0.00 ± 33% +39.0% 0.00 ± 15% sched_debug.cpu.next_balance.stddev 1410 ± 1% -14.2% 1210 ± 6% +34.5% 1896 ± 1% sched_debug.cpu.nr_load_updates.stddev 9.95 ± 3% -14.5% 8.51 ± 5% -1.2% 9.83 ± 2% sched_debug.cpu.nr_running.avg 29.07 ± 2% -15.0% 24.70 ± 4% +37.5% 39.98 ± 1% sched_debug.cpu.nr_running.max 0.05 ±100% +850.0% 0.43 ± 37% -100.0% 0.00 ± -1% sched_debug.cpu.nr_running.min 7.64 ± 3% -23.0% 5.88 ± 2% +48.6% 11.36 ± 2% sched_debug.cpu.nr_running.stddev 10979930 ± 1% +123.3% 24518490 ± 2% -26.3% 8091669 ± 1% sched_debug.cpu.nr_switches.avg 12350130 ± 1% +117.5% 26856375 ± 2% -17.0% 10249081 ± 2% sched_debug.cpu.nr_switches.max 9594835 ± 2% +132.6% 22314436 ± 2% -31.0% 6620975 ± 2% sched_debug.cpu.nr_switches.min 769296 ± 1% +56.8% 1206190 ± 3% +54.6% 1189172 ± 1% sched_debug.cpu.nr_switches.stddev 8.30 ± 18% +32.9% 11.02 ± 15% +113.7% 17.73 ± 26% sched_debug.cpu.nr_uninterruptible.max 4.87 ± 15% +14.3% 5.57 ± 6% +97.2% 9.61 ± 29% sched_debug.cpu.nr_uninterruptible.stddev