From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1755149AbbI1Gth (ORCPT ); Mon, 28 Sep 2015 02:49:37 -0400 Received: from mga09.intel.com ([134.134.136.24]:60115 "EHLO mga09.intel.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1752067AbbI1Gtf (ORCPT ); Mon, 28 Sep 2015 02:49:35 -0400 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.17,601,1437462000"; d="yaml'?scan'208";a="814286878" From: kernel test robot Subject: [lkp] [nfsd] 4aac1bf05b: -2.9% fsmark.files_per_sec CC: lkp@01.org CC: LKML TO: Jeff Layton Date: Mon, 28 Sep 2015 14:49:32 +0800 Message-ID: <87h9mfhwcj.fsf@yhuang-dev.intel.com> User-Agent: Gnus/5.13 (Gnus v5.13) Emacs/24.5 (gnu/linux) MIME-Version: 1.0 Content-Type: multipart/mixed; boundary="=-=-=" Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org --=-=-= Content-Type: text/plain; charset=iso-8859-1 Content-Disposition: inline Content-Transfer-Encoding: quoted-printable FYI, we noticed the below changes on =3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D= =3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D= =3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D= =3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D tbox_group/testcase/rootfs/kconfig/compiler/cpufreq_governor/iterations/nr_= threads/disk/fs/fs2/filesize/test_size/sync_method/nr_directories/nr_files_= per_directory: lkp-ne04/fsmark/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/performa= nce/1x/32t/1HDD/xfs/nfsv4/5K/400M/fsyncBeforeClose/16d/256fpd commit:=20 cd2d35ff27c4fda9ba73b0aa84313e8e20ce4d2c 4aac1bf05b053a201a4b392dd9a684fb2b7e6103 cd2d35ff27c4fda9 4aac1bf05b053a201a4b392dd9=20 ---------------- --------------------------=20 %stddev %change %stddev \ | \=20=20 14415356 =B1 0% +2.6% 14788625 =B1 1% fsmark.app_overhead 441.60 =B1 0% -2.9% 428.80 =B1 0% fsmark.files_per_sec 185.78 =B1 0% +2.9% 191.26 =B1 0% fsmark.time.elapsed_time 185.78 =B1 0% +2.9% 191.26 =B1 0% fsmark.time.elapsed_time.= max 97472 =B1 0% -2.8% 94713 =B1 0% fsmark.time.involuntary_c= ontext_switches 3077117 =B1 95% +251.2% 10805440 =B1112% latency_stats.sum.nfs_wai= t_on_request.nfs_updatepage.nfs_write_end.generic_perform_write.__generic_f= ile_write_iter.generic_file_write_iter.nfs_file_write.__vfs_write.vfs_write= .SyS_write.entry_SYSCALL_64_fastpath 12999 =B1 0% +32.9% 17276 =B1 0% proc-vmstat.nr_slab_unrec= laimable 64568 =B1 4% -14.8% 55032 =B1 0% softirqs.RCU 51999 =B1 0% +32.9% 69111 =B1 0% meminfo.SUnreclaim 159615 =B1 0% +13.5% 181115 =B1 0% meminfo.Slab 3.75 =B1 0% +3.3% 3.88 =B1 1% turbostat.%Busy 77.25 =B1 0% +6.5% 82.25 =B1 0% turbostat.Avg_MHz 30813025 =B1 2% -14.5% 26338527 =B1 9% cpuidle.C1E-NHM.time 164180 =B1 0% -28.9% 116758 =B1 7% cpuidle.C1E-NHM.usage 1738 =B1 2% -81.2% 326.75 =B1 4% cpuidle.POLL.usage 29979 =B1 2% +44.3% 43273 =B1 4% numa-meminfo.node0.SUnrec= laim 94889 =B1 0% +19.8% 113668 =B1 2% numa-meminfo.node0.Slab 22033 =B1 3% +17.3% 25835 =B1 7% numa-meminfo.node1.SUnrec= laim 7404 =B1 1% -2.7% 7206 =B1 0% vmstat.io.bo 27121 =B1 0% -4.8% 25817 =B1 0% vmstat.system.cs 3025 =B1 0% -13.5% 2615 =B1 0% vmstat.system.in 50126 =B1 1% +11.5% 55893 =B1 1% numa-vmstat.node0.nr_dirt= ied 7494 =B1 2% +44.3% 10818 =B1 4% numa-vmstat.node0.nr_slab= _unreclaimable 50088 =B1 1% +11.6% 55900 =B1 1% numa-vmstat.node0.nr_writ= ten 5507 =B1 3% +17.3% 6458 =B1 7% numa-vmstat.node1.nr_slab= _unreclaimable 7164 =B1 2% +275.2% 26885 =B1 0% slabinfo.kmalloc-16.activ= e_objs 7164 =B1 2% +275.3% 26885 =B1 0% slabinfo.kmalloc-16.num_o= bjs 7367 =B1 1% +787.7% 65401 =B1 0% slabinfo.kmalloc-192.acti= ve_objs 179.00 =B1 1% +771.8% 1560 =B1 0% slabinfo.kmalloc-192.acti= ve_slabs 7537 =B1 1% +770.0% 65572 =B1 0% slabinfo.kmalloc-192.num_= objs 179.00 =B1 1% +771.8% 1560 =B1 0% slabinfo.kmalloc-192.num_= slabs 3631 =B1 7% +522.3% 22600 =B1 0% slabinfo.kmalloc-256.acti= ve_objs 145.50 =B1 4% +398.1% 724.75 =B1 0% slabinfo.kmalloc-256.acti= ve_slabs 4667 =B1 4% +397.3% 23210 =B1 0% slabinfo.kmalloc-256.num_= objs 145.50 =B1 4% +398.1% 724.75 =B1 0% slabinfo.kmalloc-256.num_= slabs 17448 =B1 2% +75.6% 30643 =B1 0% slabinfo.kmalloc-32.activ= e_objs 137.50 =B1 2% +76.5% 242.75 =B1 0% slabinfo.kmalloc-32.activ= e_slabs 17651 =B1 2% +76.4% 31139 =B1 0% slabinfo.kmalloc-32.num_o= bjs 137.50 =B1 2% +76.5% 242.75 =B1 0% slabinfo.kmalloc-32.num_s= labs 2387 =B1 3% -10.7% 2132 =B1 8% slabinfo.kmalloc-512.acti= ve_objs 491.25 =B1 3% +33.9% 658.00 =B1 11% slabinfo.numa_policy.acti= ve_objs 491.25 =B1 3% +33.9% 658.00 =B1 11% slabinfo.numa_policy.num_= objs 2128 =B1 9% +59.3% 3391 =B1 34% sched_debug.cfs_rq[10]:/.= exec_clock 18088 =B1 17% +47.0% 26582 =B1 29% sched_debug.cfs_rq[10]:/.= min_vruntime 4326 =B1 11% -22.2% 3368 =B1 18% sched_debug.cfs_rq[5]:/.e= xec_clock 1459 =B1 1% -10.8% 1302 =B1 3% sched_debug.cpu#0.nr_unin= terruptible 122217 =B1 7% -18.6% 99447 =B1 2% sched_debug.cpu#1.nr_swit= ches 122732 =B1 8% -18.5% 99972 =B1 2% sched_debug.cpu#1.sched_c= ount 45603 =B1 10% -20.1% 36442 =B1 2% sched_debug.cpu#1.sched_g= oidle 27004 =B1 3% -18.9% 21895 =B1 5% sched_debug.cpu#1.ttwu_lo= cal 15469 =B1 5% +17.2% 18132 =B1 6% sched_debug.cpu#10.nr_loa= d_updates 78564 =B1 8% +26.6% 99492 =B1 5% sched_debug.cpu#10.nr_swi= tches 78605 =B1 8% +26.7% 99557 =B1 4% sched_debug.cpu#10.sched_= count 27470 =B1 9% +24.7% 34268 =B1 7% sched_debug.cpu#10.sched_= goidle 38215 =B1 1% +37.4% 52499 =B1 13% sched_debug.cpu#10.ttwu_c= ount 14816 =B1 5% +22.8% 18196 =B1 2% sched_debug.cpu#10.ttwu_l= ocal 19690 =B1 21% -29.9% 13802 =B1 15% sched_debug.cpu#11.nr_swi= tches 54.25 =B1 2% -47.5% 28.50 =B1 25% sched_debug.cpu#11.nr_uni= nterruptible 19721 =B1 21% -29.9% 13828 =B1 15% sched_debug.cpu#11.sched_= count 14545 =B1 2% +15.4% 16779 =B1 4% sched_debug.cpu#12.nr_loa= d_updates 72087 =B1 11% +27.9% 92204 =B1 7% sched_debug.cpu#12.nr_swi= tches 72126 =B1 11% +28.1% 92422 =B1 7% sched_debug.cpu#12.sched_= count 25418 =B1 13% +24.4% 31626 =B1 7% sched_debug.cpu#12.sched_= goidle 33399 =B1 15% +38.5% 46255 =B1 13% sched_debug.cpu#12.ttwu_c= ount 51.25 =B1 10% -39.0% 31.25 =B1 21% sched_debug.cpu#13.nr_uni= nterruptible 2593 =B1 11% -21.8% 2028 =B1 10% sched_debug.cpu#13.ttwu_l= ocal 71266 =B1 3% +20.1% 85620 =B1 5% sched_debug.cpu#14.nr_swi= tches 71306 =B1 3% +20.4% 85827 =B1 5% sched_debug.cpu#14.sched_= count 24634 =B1 3% +18.8% 29259 =B1 4% sched_debug.cpu#14.sched_= goidle 34625 =B1 11% +19.9% 41506 =B1 11% sched_debug.cpu#14.ttwu_c= ount 13866 =B1 3% +20.6% 16726 =B1 5% sched_debug.cpu#14.ttwu_l= ocal 12683 =B1 4% -14.7% 10817 =B1 2% sched_debug.cpu#15.nr_loa= d_updates 49.75 =B1 6% -46.2% 26.75 =B1 28% sched_debug.cpu#15.nr_uni= nterruptible 3374 =B1 12% -28.1% 2427 =B1 18% sched_debug.cpu#15.ttwu_l= ocal 186563 =B1 5% -12.1% 163975 =B1 4% sched_debug.cpu#2.nr_swit= ches -1324 =B1 -2% -16.0% -1111 =B1 -1% sched_debug.cpu#2.nr_unin= terruptible 187499 =B1 5% -11.2% 166447 =B1 4% sched_debug.cpu#2.sched_c= ount 67465 =B1 7% -13.6% 58308 =B1 6% sched_debug.cpu#2.sched_g= oidle 36525 =B1 4% -14.6% 31193 =B1 1% sched_debug.cpu#2.ttwu_lo= cal 23697 =B1 5% -13.2% 20572 =B1 9% sched_debug.cpu#3.nr_load= _updates 128070 =B1 1% -22.9% 98687 =B1 5% sched_debug.cpu#3.nr_swit= ches 129859 =B1 2% -23.5% 99357 =B1 4% sched_debug.cpu#3.sched_c= ount 48833 =B1 1% -23.7% 37243 =B1 6% sched_debug.cpu#3.sched_g= oidle 61622 =B1 3% -24.2% 46694 =B1 5% sched_debug.cpu#3.ttwu_co= unt 27510 =B1 7% -20.6% 21840 =B1 8% sched_debug.cpu#3.ttwu_lo= cal 81675 =B1 7% -13.6% 70536 =B1 1% sched_debug.cpu#4.ttwu_co= unt 34076 =B1 3% -12.9% 29683 =B1 1% sched_debug.cpu#4.ttwu_lo= cal 124470 =B1 4% -14.1% 106865 =B1 8% sched_debug.cpu#5.sched_c= ount 62502 =B1 3% -20.8% 49519 =B1 9% sched_debug.cpu#5.ttwu_co= unt 26562 =B1 0% -17.7% 21853 =B1 10% sched_debug.cpu#5.ttwu_lo= cal 181661 =B1 10% -15.1% 154229 =B1 6% sched_debug.cpu#6.nr_swit= ches 181937 =B1 10% -13.5% 157379 =B1 6% sched_debug.cpu#6.sched_c= ount 66672 =B1 14% -16.6% 55632 =B1 9% sched_debug.cpu#6.sched_g= oidle 78296 =B1 2% -10.2% 70346 =B1 6% sched_debug.cpu#6.ttwu_co= unt 33536 =B1 1% -14.4% 28696 =B1 1% sched_debug.cpu#6.ttwu_lo= cal 131463 =B1 6% -17.0% 109140 =B1 4% sched_debug.cpu#7.nr_swit= ches -32.25 =B1-58% -100.8% 0.25 =B19467% sched_debug.cpu#7.nr_uni= nterruptible 133606 =B1 7% -17.2% 110671 =B1 4% sched_debug.cpu#7.sched_c= ount 50986 =B1 7% -16.6% 42525 =B1 6% sched_debug.cpu#7.sched_g= oidle 61388 =B1 2% -19.8% 49213 =B1 5% sched_debug.cpu#7.ttwu_co= unt 26637 =B1 2% -21.8% 20837 =B1 3% sched_debug.cpu#7.ttwu_lo= cal 12312 =B1 3% +9.4% 13474 =B1 4% sched_debug.cpu#8.nr_load= _updates 53.50 =B1 6% -44.9% 29.50 =B1 27% sched_debug.cpu#9.nr_unin= terruptible 2724 =B1 15% -23.7% 2078 =B1 26% sched_debug.cpu#9.ttwu_lo= cal lkp-ne04: Nehalem-EP Memory: 12G cpuidle.POLL.usage 1800 ++----------*-----------------------------*-----*-----*--------*--*-= -+ *..*..*..*. .*..*..*..*..*..*..*..*. *. *. *..*. = * 1600 ++ *. = | 1400 ++ = | | = | 1200 ++ = | | = | 1000 ++ = | | = | 800 ++ = | 600 ++ = | | = | 400 O+ = | | O O O O O O O O O O O O O O O O O O O O O O = | 200 ++------------------------------------------------------------------= -+ cpuidle.C1E-NHM.usage 190000 ++-----------------------------------------------------*----------= -+ | : : = | 180000 ++ : : = | 170000 ++ .*. : : = | | .*.. .*.. .*. .*.. .*..*..*..*. * *..*..*.= .* 160000 *+ *. *. *..*..*. *..*..*. = | 150000 ++ = | | = | 140000 ++ = | 130000 ++ O = | | = | 120000 ++ O O O O = | 110000 ++ O O O O O O O O O = | O O O O O O O O O = | 100000 ++----------------------------------------------------------------= -+ fsmark.files_per_sec 446 ++-------------------------------------------------------------------= -+ 444 ++ *.. = | | .. = | 442 *+.*..*..*..*..*..*.. *..*..*..*...*..*..*..*..*..*..* *..*..*.= .* 440 ++ .. = | 438 ++ * = | 436 ++ = | | = | 434 ++ = | 432 ++ = | 430 ++ = | 428 ++ O O O O O O O O O O O O O O O O O O = | | = | 426 O+ O O O O = | 424 ++-------------------------------------------------------------------= -+ fsmark.time.elapsed_time 193 ++-------------------------------------------------------------------= -+ O O O = | 192 ++ O O O O O O O O O O = | 191 ++ O O O O O O O O = | | O O = | 190 ++ = | 189 ++ = | | = | 188 ++ = | 187 ++ = | | .*.. = | 186 *+. .*..*..*..*..*. *.. .*.. ..*.. .*.. .*.. .*.. *..*..*.= .* 185 ++ *. *. *. *. *. *. .. = | | * = | 184 ++-------------------------------------------------------------------= -+ fsmark.time.elapsed_time.max 193 ++-------------------------------------------------------------------= -+ O O O = | 192 ++ O O O O O O O O O O = | 191 ++ O O O O O O O O = | | O O = | 190 ++ = | 189 ++ = | | = | 188 ++ = | 187 ++ = | | .*.. = | 186 *+. .*..*..*..*..*. *.. .*.. ..*.. .*.. .*.. .*.. *..*..*.= .* 185 ++ *. *. *. *. *. *. .. = | | * = | 184 ++-------------------------------------------------------------------= -+ fsmark.time.involuntary_context_switches 98500 ++-----------------------------------------------------------------= -+ 98000 ++ .* = | | *.. .*. + *.. .*..*..* = * 97500 ++. *..*..*. + .. *..*.*..*..*..*..*..*..*. + = +| 97000 *+ *..* + += | | * = | 96500 ++ = | 96000 ++ = | 95500 ++ = | | = | 95000 ++ O O O O O O O = | 94500 ++ O O O O O O O O O = | O O O O O = | 94000 ++ O O = | 93500 ++-----------------------------------------------------------------= -+ vmstat.system.in 3100 ++------------------------------------------------------------------= -+ 3050 *+. .*.. .*.. .*..*.. .*..*.. = | | .*..*.. .*..*. *. *. *.. .*..*. .*..*..*.= .* 3000 ++ *. *. *. *. = | 2950 ++ = | 2900 ++ = | 2850 ++ = | | = | 2800 ++ = | 2750 ++ = | 2700 ++ = | 2650 ++ O O = | | O O O O O O O O O O O = | 2600 O+ O O O O O O O O O = | 2550 ++------------------------------------------------------------------= -+ numa-vmstat.node0.nr_slab_unreclaimable 12000 ++-----------------------------------------------------------------= -+ | O = | 11000 ++ O O O O O O = | O O O O O = | | O O O O O O O O = | 10000 ++ O O O = | | = | 9000 ++ = | | = | 8000 ++ *.. = | | .*..*.. + .*.. = .* | .*.. .*.. .*.*. + *..*..*. *..*.= | 7000 *+.*. .*. .*..*. * = | | *..*. *..*. = | 6000 ++-----------------------------------------------------------------= -+ numa-vmstat.node0.nr_dirtied 58000 ++-----------------------------------------------------------------= -+ 57000 ++ O = | O O O O = | 56000 ++ O O O O O O = | 55000 ++ O O O O O O O O O O = | | O O = | 54000 ++ = | 53000 ++ = | 52000 ++ = | | *.. * = | 51000 +++ .*.. .*.. .. + .*.. *.= .| 50000 ++ *. *..*..*.. .* * + .*.. *. *.. .. = * * *.. .*..*. *. .. * = | 49000 ++ *. * = | 48000 ++-----------------------------------------------------------------= -+ numa-vmstat.node0.nr_written 58000 ++-----------------------------------------------------------------= -+ 57000 ++ O = | O O O O O = | 56000 ++ O O O O O O = | 55000 ++ O O O O O O O O O = | | O O = | 54000 ++ = | 53000 ++ = | 52000 ++ = | | * = | 51000 ++ *.. .*.. .*.. .. + * .*.. *.= .| 50000 ++. *. *..*..*.. .* * + .. + *. *.. .. = * * *.. .*..*. * + .. * = | 49000 ++ *. * = | 48000 ++-----------------------------------------------------------------= -+ numa-meminfo.node0.SUnreclaim 50000 ++-----------------------------------------------------------------= -+ | = | | O = | 45000 ++ O O O O O O = | O O O O O = | | O O O O O O O O = | 40000 ++ O O O = | | = | 35000 ++ = | | = | | *.. *.. = | 30000 ++ .. *.. .. *.. .*.. = .* *..*..*.. .*.. .*..*.* * *..*. *..*.= | | *..*..*. .*. = | 25000 ++-------------------*--*------------------------------------------= -+ proc-vmstat.nr_slab_unreclaimable 17500 ++-O--O-----O--O-------------------O--O--------O-----O-----O-----O-= -+ 17000 O+ O O O O O O O O O O O O = | | = | 16500 ++ = | 16000 ++ = | | = | 15500 ++ = | 15000 ++ = | 14500 ++ = | | = | 14000 ++ = | 13500 ++ = | | = | 13000 *+.*..*..*..*..*..*..*..*..*..*..*.*..*..*..*..*..*..*..*..*..*..*.= .* 12500 ++-----------------------------------------------------------------= -+ meminfo.Slab 185000 ++----------------------------------------------------------------= -+ | O O O O O = | 180000 O+ O O O O O O O O O O O O O O O O O = | | = | | = | 175000 ++ = | | = | 170000 ++ = | | = | 165000 ++ = | | = | | = | 160000 *+.*..*..*..*..*.*..*.. .*..*..*..*..*.. .*.. .*..*..*..*..*.= .* | *..*. *. * = | 155000 ++----------------------------------------------------------------= -+ meminfo.SUnreclaim 70000 ++-O--O-----O--O-------------------O--O--------O-----O-----O-----O-= -+ 68000 O+ O O O O O O O O O O O O = | | = | 66000 ++ = | 64000 ++ = | | = | 62000 ++ = | 60000 ++ = | 58000 ++ = | | = | 56000 ++ = | 54000 ++ = | | = | 52000 *+.*..*..*..*..*..*..*..*..*..*..*.*..*..*..*..*..*..*..*..*..*..*.= .* 50000 ++-----------------------------------------------------------------= -+ slabinfo.kmalloc-256.active_objs 25000 ++-----------------------------------------------------------------= -+ | O O O O O O O O O O O O O O O O O O O O O = | O O = | 20000 ++ = | | = | | = | 15000 ++ = | | = | 10000 ++ = | | = | | = | 5000 ++ = | *..*..*..*..*..*..*..*..*..*..*..*.*..*..*..*..*..*..*..*..*..*..*.= .* | = | 0 ++-----------------------------------------------------------------= -+ slabinfo.kmalloc-256.num_objs 24000 ++-O--O-----O--O--------O-----O--O-O--O--O--O--O--O--O-----O-----O-= -+ 22000 O+ O O O O O O = | | = | 20000 ++ = | 18000 ++ = | | = | 16000 ++ = | 14000 ++ = | 12000 ++ = | | = | 10000 ++ = | 8000 ++ = | | = | 6000 ++ .*.. .*..*..*.. .*.. .*.= .| 4000 *+-*--*--*--*--*--*--*--*--*--*--*-*-----*-----------*--*-----*----= -* slabinfo.kmalloc-256.active_slabs 800 ++-------------------------------------------------------------------= -+ O O O O O O O O O O O O O O O O O O O = | 700 ++ O O O O = | | = | 600 ++ = | | = | 500 ++ = | | = | 400 ++ = | | = | 300 ++ = | | = | 200 ++ = | *..*..*..*..*..*..*..*..*..*..*..*...*..*..*..*..*..*..*..*..*..*..*.= .* 100 ++-------------------------------------------------------------------= -+ slabinfo.kmalloc-256.num_slabs 800 ++-------------------------------------------------------------------= -+ O O O O O O O O O O O O O O O O O O O = | 700 ++ O O O O = | | = | 600 ++ = | | = | 500 ++ = | | = | 400 ++ = | | = | 300 ++ = | | = | 200 ++ = | *..*..*..*..*..*..*..*..*..*..*..*...*..*..*..*..*..*..*..*..*..*..*.= .* 100 ++-------------------------------------------------------------------= -+ slabinfo.kmalloc-192.active_objs 70000 ++-----------------------------------------------------------------= -+ O O O O O O O O O O O O O O O O O O O O O O O = | 60000 ++ = | | = | 50000 ++ = | | = | 40000 ++ = | | = | 30000 ++ = | | = | 20000 ++ = | | = | 10000 *+. .*.. .*.= .| | *..*..*..*..*..*..*..*..*..*..*.*..*..*..*..*. *..*..*..*. = * 0 ++-----------------------------------------------------------------= -+ slabinfo.kmalloc-192.num_objs 70000 ++-----------------------------------------------------------------= -+ O O O O O O O O O O O O O O O O O O O O O O O = | 60000 ++ = | | = | 50000 ++ = | | = | 40000 ++ = | | = | 30000 ++ = | | = | 20000 ++ = | | = | 10000 *+. .*.. .*..*..*. .*.. .*.. .*.. .*.. .*.= .* | *. *..*..*..*..*..*. *. *. *. *..*. *. = | 0 ++-----------------------------------------------------------------= -+ slabinfo.kmalloc-192.active_slabs 1600 O+-O--O--O--O--O--O--O--O--O--O--O--O--O--O--O--O--O--O--O--O--O--O-= -+ | = | 1400 ++ = | 1200 ++ = | | = | 1000 ++ = | | = | 800 ++ = | | = | 600 ++ = | 400 ++ = | | = | 200 *+.*..*..*..*..*..*..*..*..*..*..*..*..*..*..*..*..*..*..*..*..*..*.= .* | = | 0 ++------------------------------------------------------------------= -+ slabinfo.kmalloc-192.num_slabs 1600 O+-O--O--O--O--O--O--O--O--O--O--O--O--O--O--O--O--O--O--O--O--O--O-= -+ | = | 1400 ++ = | 1200 ++ = | | = | 1000 ++ = | | = | 800 ++ = | | = | 600 ++ = | 400 ++ = | | = | 200 *+.*..*..*..*..*..*..*..*..*..*..*..*..*..*..*..*..*..*..*..*..*..*.= .* | = | 0 ++------------------------------------------------------------------= -+ slabinfo.kmalloc-32.active_objs 55000 ++-----------------------------------------------------------------= -+ | O = | 50000 O+ O O O = | 45000 ++ = | | = | 40000 ++ = | | = | 35000 ++ = | | O O O O O O O O O = | 30000 ++ O O O O O O O O O = | 25000 ++ = | | = | 20000 ++ = | *..*..*..*.. .*..*..*..*..*..*..*.*..*..*..*.. .*..*..*..*..*..*.= .* 15000 ++----------*----------------------------------*-------------------= -+ slabinfo.kmalloc-32.num_objs 55000 ++-----------------------------------------------------------------= -+ O O O O = | 50000 ++ O = | 45000 ++ = | | = | 40000 ++ = | | = | 35000 ++ = | | O O O O O O O O O O O O O O O O = | 30000 ++ O O = | 25000 ++ = | | = | 20000 ++ = | *..*..*..*.. .*..*..*..*..*..*..*.*..*..*..*.. .*..*..*..*..*..*.= .* 15000 ++----------*----------------------------------*-------------------= -+ slabinfo.kmalloc-32.active_slabs 400 O+-O--O--O--O--------------------------------------------------------= -+ | = | 350 ++ = | | = | | = | 300 ++ = | | = | 250 ++ O O O O O O O O O O O O O O O O = | | O O = | 200 ++ = | | = | | = | 150 *+.*..*..*.. .*..*..*.. .*..*..*...*..*.. .*.. .*..*.. .*..*..*.= .* | *. *. *. *. *. = | 100 ++-------------------------------------------------------------------= -+ slabinfo.kmalloc-32.num_slabs 400 O+-O--O--O--O--------------------------------------------------------= -+ | = | 350 ++ = | | = | | = | 300 ++ = | | = | 250 ++ O O O O O O O O O O O O O O O O = | | O O = | 200 ++ = | | = | | = | 150 *+.*..*..*.. .*..*..*.. .*..*..*...*..*.. .*.. .*..*.. .*..*..*.= .* | *. *. *. *. *. = | 100 ++-------------------------------------------------------------------= -+ kmsg.usb_usb7:can_t_set_config___error 1 ++-------------------------------------------------------------------= -* | = | | = :| 0.8 ++ = :| | = :| | = :| 0.6 ++ := | | := | 0.4 ++ := | | : = | | : = | 0.2 ++ : = | | : = | | : = | 0 *+--*---*---*---*----*---*---*---*---*---*---*---*----*---*---*---*--= -+ [*] bisect-good sample [O] bisect-bad sample To reproduce: git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tes= ts.git cd lkp-tests bin/lkp install job.yaml # job file is attached in this email bin/lkp run job.yaml Disclaimer: Results have been estimated based on internal Intel analysis and are provid= ed for informational purposes only. Any difference in system hardware or softw= are design or configuration may affect actual performance. Thanks, Ying Huang --=-=-= Content-Type: text/plain; charset=ascii Content-Disposition: attachment; filename=job.yaml --- LKP_SERVER: inn LKP_CGI_PORT: 80 LKP_CIFS_PORT: 139 testcase: fsmark default-monitors: wait: activate-monitor kmsg: uptime: iostat: vmstat: numa-numastat: numa-vmstat: numa-meminfo: proc-vmstat: proc-stat: interval: 10 meminfo: slabinfo: interrupts: lock_stat: latency_stats: softirqs: bdi_dev_mapping: diskstats: nfsstat: cpuidle: cpufreq-stats: turbostat: pmeter: sched_debug: interval: 60 cpufreq_governor: performance default-watchdogs: oom-killer: watchdog: commit: 31e1a90a1e09e8e600f1b0b5dcb871c399448f81 model: Nehalem-EP memory: 12G hdd_partitions: "/dev/disk/by-id/ata-ST3500514NS_9WJ03EBA-part3" swap_partitions: "/dev/disk/by-id/ata-ST3120026AS_5MS07HA2-part2" rootfs_partition: "/dev/disk/by-id/ata-ST3500514NS_9WJ03EBA-part1" category: benchmark iterations: 1x nr_threads: 32t disk: 1HDD fs: xfs fs2: nfsv4 fsmark: filesize: 5K test_size: 400M sync_method: fsyncBeforeClose nr_directories: 16d nr_files_per_directory: 256fpd queue: cyclic testbox: lkp-ne04 tbox_group: lkp-ne04 kconfig: x86_64-rhel enqueue_time: 2015-09-16 01:53:44.450007854 +08:00 id: fa22bba34d95d9194b42a923ea8e320612db9c82 user: lkp compiler: gcc-4.9 head_commit: 31e1a90a1e09e8e600f1b0b5dcb871c399448f81 base_commit: 6ff33f3902c3b1c5d0db6b1e2c70b6d76fba357f branch: linux-devel/devel-hourly-2015091604 kernel: "/pkg/linux/x86_64-rhel/gcc-4.9/31e1a90a1e09e8e600f1b0b5dcb871c399448f81/vmlinuz-4.3.0-rc1-wl-ath-00944-g31e1a90" rootfs: debian-x86_64-2015-02-07.cgz result_root: "/result/fsmark/performance-1x-32t-1HDD-xfs-nfsv4-5K-400M-fsyncBeforeClose-16d-256fpd/lkp-ne04/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/31e1a90a1e09e8e600f1b0b5dcb871c399448f81/0" job_file: "/lkp/scheduled/lkp-ne04/cyclic_fsmark-performance-1x-32t-1HDD-xfs-nfsv4-5K-400M-fsyncBeforeClose-16d-256fpd-x86_64-rhel-CYCLIC_HEAD-31e1a90a1e09e8e600f1b0b5dcb871c399448f81-20150916-88681-jka6un-0.yaml" dequeue_time: 2015-09-16 05:10:38.959773945 +08:00 nr_cpu: "$(nproc)" max_uptime: 1073.42 initrd: "/osimage/debian/debian-x86_64-2015-02-07.cgz" bootloader_append: - root=/dev/ram0 - user=lkp - job=/lkp/scheduled/lkp-ne04/cyclic_fsmark-performance-1x-32t-1HDD-xfs-nfsv4-5K-400M-fsyncBeforeClose-16d-256fpd-x86_64-rhel-CYCLIC_HEAD-31e1a90a1e09e8e600f1b0b5dcb871c399448f81-20150916-88681-jka6un-0.yaml - ARCH=x86_64 - kconfig=x86_64-rhel - branch=linux-devel/devel-hourly-2015091604 - commit=31e1a90a1e09e8e600f1b0b5dcb871c399448f81 - BOOT_IMAGE=/pkg/linux/x86_64-rhel/gcc-4.9/31e1a90a1e09e8e600f1b0b5dcb871c399448f81/vmlinuz-4.3.0-rc1-wl-ath-00944-g31e1a90 - max_uptime=1073 - RESULT_ROOT=/result/fsmark/performance-1x-32t-1HDD-xfs-nfsv4-5K-400M-fsyncBeforeClose-16d-256fpd/lkp-ne04/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/31e1a90a1e09e8e600f1b0b5dcb871c399448f81/0 - LKP_SERVER=inn - |2- earlyprintk=ttyS0,115200 systemd.log_level=err debug apic=debug sysrq_always_enabled rcupdate.rcu_cpu_stall_timeout=100 panic=-1 softlockup_panic=1 nmi_watchdog=panic oops=panic load_ramdisk=2 prompt_ramdisk=0 console=ttyS0,115200 console=tty0 vga=normal rw lkp_initrd: "/lkp/lkp/lkp-x86_64.cgz" modules_initrd: "/pkg/linux/x86_64-rhel/gcc-4.9/31e1a90a1e09e8e600f1b0b5dcb871c399448f81/modules.cgz" bm_initrd: "/osimage/deps/debian-x86_64-2015-02-07.cgz/lkp.cgz,/osimage/deps/debian-x86_64-2015-02-07.cgz/run-ipconfig.cgz,/osimage/deps/debian-x86_64-2015-02-07.cgz/turbostat.cgz,/lkp/benchmarks/turbostat.cgz,/osimage/deps/debian-x86_64-2015-02-07.cgz/fs.cgz,/osimage/deps/debian-x86_64-2015-02-07.cgz/fs2.cgz,/lkp/benchmarks/fsmark.cgz" job_state: finished loadavg: 32.41 16.01 6.32 1/269 5937 start_time: '1442351486' end_time: '1442351678' version: "/lkp/lkp/.src-20150915-235159" --=-=-= Content-Type: text/plain; charset=ascii Content-Disposition: attachment; filename=reproduce echo performance > /sys/devices/system/cpu/cpu0/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu1/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu10/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu11/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu12/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu13/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu14/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu15/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu2/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu3/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu4/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu5/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu6/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu7/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu8/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu9/cpufreq/scaling_governor mkfs -t xfs /dev/sda3 mount -t xfs -o nobarrier,inode64 /dev/sda3 /fs/sda3 /etc/init.d/rpcbind start /etc/init.d/nfs-common start /etc/init.d/nfs-kernel-server start mount -t nfs -o vers=4 localhost:/fs/sda3 /nfs/sda3 ./fs_mark -d /nfs/sda3/1 -d /nfs/sda3/2 -d /nfs/sda3/3 -d /nfs/sda3/4 -d /nfs/sda3/5 -d /nfs/sda3/6 -d /nfs/sda3/7 -d /nfs/sda3/8 -d /nfs/sda3/9 -d /nfs/sda3/10 -d /nfs/sda3/11 -d /nfs/sda3/12 -d /nfs/sda3/13 -d /nfs/sda3/14 -d /nfs/sda3/15 -d /nfs/sda3/16 -d /nfs/sda3/17 -d /nfs/sda3/18 -d /nfs/sda3/19 -d /nfs/sda3/20 -d /nfs/sda3/21 -d /nfs/sda3/22 -d /nfs/sda3/23 -d /nfs/sda3/24 -d /nfs/sda3/25 -d /nfs/sda3/26 -d /nfs/sda3/27 -d /nfs/sda3/28 -d /nfs/sda3/29 -d /nfs/sda3/30 -d /nfs/sda3/31 -d /nfs/sda3/32 -D 16 -N 256 -n 2560 -L 1 -S 1 -s 5120 --=-=-=--