From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1753379AbcB2IuV (ORCPT ); Mon, 29 Feb 2016 03:50:21 -0500 Received: from mga03.intel.com ([134.134.136.65]:50259 "EHLO mga03.intel.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1752074AbcB2IuS (ORCPT ); Mon, 29 Feb 2016 03:50:18 -0500 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.22,520,1449561600"; d="yaml'?scan'208";a="926032990" From: kernel test robot Subject: [lkp] [ext4] 1f2d779fed: +34.0% fsmark.files_per_sec CC: lkp@01.org CC: LKML CC: Andreas Dilger CC: "Theodore Ts'o" TO: Lokesh Jaliminche Date: Mon, 29 Feb 2016 16:50:15 +0800 Message-ID: <87vb57yk6w.fsf@yhuang-dev.intel.com> User-Agent: Gnus/5.13 (Gnus v5.13) Emacs/24.5 (gnu/linux) MIME-Version: 1.0 Content-Type: multipart/mixed; boundary="=-=-=" Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org --=-=-= Content-Type: text/plain; charset=iso-8859-1 Content-Disposition: inline Content-Transfer-Encoding: quoted-printable FYI, we noticed the below changes on https://git.kernel.org/pub/scm/linux/kernel/git/next/linux-next.git master commit 1f2d779fed21806ffed84aa65617c6125f480e85 ("ext4: optimize group sear= ch for inode allocation") =3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D= =3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D= =3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D= =3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D compiler/cpufreq_governor/disk/filesize/fs/iterations/kconfig/nr_directorie= s/nr_files_per_directory/nr_threads/rootfs/sync_method/tbox_group/test_size= /testcase: gcc-4.9/performance/1HDD/9B/ext4/1x/x86_64-rhel/16d/256fpd/32t/debian-x86= _64-2015-02-07.cgz/fsyncBeforeClose/lkp-ne04/400M/fsmark commit:=20 29c6eaffc8868ef6fa71997d0ea507a02c52712c 1f2d779fed21806ffed84aa65617c6125f480e85 29c6eaffc8868ef6 1f2d779fed21806ffed84aa656=20 ---------------- --------------------------=20 %stddev %change %stddev \ | \=20=20 1479204 =B1 11% +47.7% 2184108 =B1 7% fsmark.app_overhead 148.78 =B1 3% +34.0% 199.38 =B1 2% fsmark.files_per_sec 693.79 =B1 3% -26.1% 512.52 =B1 2% fsmark.time.elapsed_time 693.79 =B1 3% -26.1% 512.52 =B1 2% fsmark.time.elapsed_time.= max 264.00 =B1 0% +11372.0% 30286 =B1 15% fsmark.time.file_system_i= nputs 2.00 =B1 0% +100.0% 4.00 =B1 0% fsmark.time.percent_of_cp= u_this_job_got 408852 =B1 1% +19.0% 486551 =B1 0% fsmark.time.voluntary_con= text_switches 42686 =B1 4% -9.2% 38753 =B1 1% softirqs.RCU 43495 =B1 3% -6.4% 40722 =B1 4% softirqs.SCHED 717.23 =B1 3% -24.2% 543.65 =B1 2% uptime.boot 9657 =B1 4% -28.5% 6902 =B1 3% uptime.idle 3817 =B1 1% +26.2% 4819 =B1 2% vmstat.system.cs 431.50 =B1 1% +31.4% 567.00 =B1 4% vmstat.system.in 620517 =B1 4% -23.2% 476762 =B1 1% numa-numastat.node0.local= _node 620519 =B1 4% -23.2% 476763 =B1 1% numa-numastat.node0.numa_= hit 783065 =B1 5% -24.2% 593500 =B1 3% numa-numastat.node1.local= _node 783066 =B1 5% -24.2% 593500 =B1 3% numa-numastat.node1.numa_= hit 36911353 =B1 8% +61.1% 59481022 =B1 5% cpuidle.C1E-NHM.time 9758 =B1 5% +41.5% 13810 =B1 2% cpuidle.C1E-NHM.usage 4.406e+08 =B1 5% +34.7% 5.934e+08 =B1 5% cpuidle.C3-NHM.time 112418 =B1 3% +50.3% 168929 =B1 1% cpuidle.C3-NHM.usage 1.049e+10 =B1 3% -29.3% 7.418e+09 =B1 3% cpuidle.C6-NHM.time 81277 =B1 3% -20.9% 64279 =B1 2% cpuidle.POLL.usage 1059 =B1 5% -8.8% 965.75 =B1 7% slabinfo.blkdev_requests.= num_objs 638.25 =B1 4% +34.7% 859.75 =B1 8% slabinfo.dquot.active_objs 638.50 =B1 4% +38.5% 884.25 =B1 7% slabinfo.dquot.num_objs 2026 =B1 13% -26.9% 1480 =B1 9% slabinfo.jbd2_journal_hea= d.active_objs 2061 =B1 12% -25.8% 1529 =B1 9% slabinfo.jbd2_journal_hea= d.num_objs 2780 =B1 2% +11.5% 3099 =B1 3% slabinfo.kmalloc-256.acti= ve_objs 0.82 =B1 2% +40.4% 1.16 =B1 4% turbostat.%Busy 22.25 =B1 3% +40.4% 31.25 =B1 6% turbostat.Avg_MHz 2.69 =B1 5% +49.3% 4.01 =B1 6% turbostat.CPU%c1 6.79 =B1 2% +73.8% 11.81 =B1 8% turbostat.CPU%c3 28.16 =B1 3% +21.5% 34.22 =B1 3% turbostat.Pkg%pc3 33.39 =B1 2% -41.9% 19.41 =B1 12% turbostat.Pkg%pc6 693.79 =B1 3% -26.1% 512.52 =B1 2% time.elapsed_time 693.79 =B1 3% -26.1% 512.52 =B1 2% time.elapsed_time.max 264.00 =B1 0% +11372.0% 30286 =B1 15% time.file_system_inputs 2.00 =B1 0% +100.0% 4.00 =B1 0% time.percent_of_cpu_this_= job_got 17.25 =B1 1% +21.9% 21.03 =B1 2% time.system_time 1.14 =B1 2% +33.8% 1.53 =B1 12% time.user_time 408852 =B1 1% +19.0% 486551 =B1 0% time.voluntary_context_sw= itches 1400397 =B1 3% -23.9% 1066153 =B1 2% proc-vmstat.numa_hit 1400394 =B1 3% -23.9% 1066151 =B1 2% proc-vmstat.numa_local 247146 =B1 4% -23.1% 190110 =B1 1% proc-vmstat.pgalloc_dma32 1235243 =B1 4% -23.6% 943649 =B1 2% proc-vmstat.pgalloc_normal 1461923 =B1 4% -25.7% 1086000 =B1 2% proc-vmstat.pgfault 1314409 =B1 4% -26.9% 961033 =B1 2% proc-vmstat.pgfree 132.00 =B1 0% +11365.9% 15135 =B1 15% proc-vmstat.pgpgin 3926356 =B1 18% -38.5% 2413806 =B1 16% proc-vmstat.pgpgout 141.50 =B1 16% +1078.3% 1667 =B1 52% numa-vmstat.node0.nr_shmem 22694 =B1 0% -17.7% 18679 =B1 11% numa-vmstat.node0.nr_slab= _reclaimable 6899 =B1 1% -12.3% 6049 =B1 1% numa-vmstat.node0.nr_slab= _unreclaimable 447024 =B1 4% -13.7% 385841 =B1 4% numa-vmstat.node0.numa_hit 384765 =B1 5% -15.8% 323917 =B1 4% numa-vmstat.node0.numa_lo= cal 63402 =B1 1% +19.2% 75581 =B1 11% numa-vmstat.node1.nr_file= _pages 2160 =B1 0% -71.6% 613.25 =B1143% numa-vmstat.node1.nr_inac= tive_anon 48098 =B1 2% +28.5% 61803 =B1 12% numa-vmstat.node1.nr_inac= tive_file 2373 =B1 0% -31.3% 1630 =B1 25% numa-vmstat.node1.nr_mapp= ed 2207 =B1 1% -70.1% 659.50 =B1133% numa-vmstat.node1.nr_shmem 5466 =B1 4% +78.8% 9772 =B1 24% numa-vmstat.node1.nr_slab= _reclaimable 4392 =B1 1% +20.0% 5272 =B1 1% numa-vmstat.node1.nr_slab= _unreclaimable 523487 =B1 1% -20.1% 418465 =B1 6% numa-vmstat.node1.numa_hit 519756 =B1 1% -20.2% 414604 =B1 6% numa-vmstat.node1.numa_lo= cal 90778 =B1 0% -17.7% 74720 =B1 11% numa-meminfo.node0.SRecla= imable 27598 =B1 1% -12.3% 24198 =B1 1% numa-meminfo.node0.SUnrec= laim 567.50 =B1 16% +1075.3% 6670 =B1 52% numa-meminfo.node0.Shmem 118377 =B1 0% -16.4% 98919 =B1 9% numa-meminfo.node0.Slab 253611 =B1 1% +19.2% 302328 =B1 11% numa-meminfo.node1.FilePa= ges 201041 =B1 2% +24.2% 249668 =B1 13% numa-meminfo.node1.Inacti= ve 8645 =B1 0% -71.6% 2454 =B1143% numa-meminfo.node1.Inacti= ve(anon) 192396 =B1 2% +28.5% 247214 =B1 12% numa-meminfo.node1.Inacti= ve(file) 2197 =B1 15% +25.2% 2752 =B1 7% numa-meminfo.node1.Kernel= Stack 9494 =B1 0% -31.3% 6524 =B1 25% numa-meminfo.node1.Mapped 345480 =B1 1% +20.3% 415555 =B1 10% numa-meminfo.node1.MemUsed 21865 =B1 4% +78.8% 39089 =B1 24% numa-meminfo.node1.SRecla= imable 17570 =B1 1% +20.0% 21091 =B1 1% numa-meminfo.node1.SUnrec= laim 8830 =B1 1% -70.1% 2640 =B1133% numa-meminfo.node1.Shmem 39436 =B1 2% +52.6% 60181 =B1 16% numa-meminfo.node1.Slab 25609 =B1155% -99.0% 247.00 =B1173% latency_stats.avg.do_get_= write_access.jbd2_journal_get_write_access.__ext4_journal_get_write_access.= ext4_reserve_inode_write.ext4_mark_inode_dirty.ext4_dirty_inode.__mark_inod= e_dirty.generic_update_time.file_update_time.__generic_file_write_iter.ext4= _file_write_iter.__vfs_write 34613 =B1 17% -93.2% 2366 =B1100% latency_stats.avg.submit_= bio_wait.blkdev_issue_flush.ext4_sync_file.vfs_fsync_range.do_fsync.SyS_fsy= nc.entry_SYSCALL_64_fastpath 62402 =B1 14% -100.0% 0.00 =B1 -1% latency_stats.avg.submit_= bio_wait.blkdev_issue_flush.jbd2_cleanup_journal_tail.jbd2_log_do_checkpoin= t.__jbd2_log_wait_for_space.add_transaction_credits.start_this_handle.jbd2_= _journal_start.__ext4_journal_start_sb.__ext4_new_inode.ext4_create.vfs_cre= ate 85292 =B1 16% -100.0% 0.00 =B1 -1% latency_stats.avg.submit_= bio_wait.blkdev_issue_flush.jbd2_cleanup_journal_tail.jbd2_log_do_checkpoin= t.__jbd2_log_wait_for_space.add_transaction_credits.start_this_handle.jbd2_= _journal_start.__ext4_journal_start_sb.ext4_da_write_begin.generic_perform_= write.__generic_file_write_iter 233836 =B1 9% -100.0% 0.00 =B1 -1% latency_stats.avg.submit_= bio_wait.blkdev_issue_flush.jbd2_cleanup_journal_tail.jbd2_log_do_checkpoin= t.__jbd2_log_wait_for_space.add_transaction_credits.start_this_handle.jbd2_= _journal_start.__ext4_journal_start_sb.ext4_dirty_inode.__mark_inode_dirty.= generic_update_time 98586 =B1 15% -100.0% 0.00 =B1 -1% latency_stats.avg.submit_= bio_wait.blkdev_issue_flush.jbd2_cleanup_journal_tail.jbd2_log_do_checkpoin= t.__jbd2_log_wait_for_space.add_transaction_credits.start_this_handle.jbd2_= _journal_start.__ext4_journal_start_sb.ext4_writepages.do_writepages.__file= map_fdatawrite_range 138727 =B1 20% -99.7% 356.50 =B1 34% latency_stats.avg.wait_tr= ansaction_locked.add_transaction_credits.start_this_handle.jbd2__journal_st= art.__ext4_journal_start_sb.ext4_dirty_inode.__mark_inode_dirty.generic_upd= ate_time.file_update_time.__generic_file_write_iter.ext4_file_write_iter.__= vfs_write 0.00 =B1 -1% +Inf% 7289 =B1 37% latency_stats.hits.do_get= _write_access.jbd2_journal_get_write_access.__ext4_journal_get_write_access= .ext4_reserve_inode_write.ext4_mark_inode_dirty.add_dirent_to_buf.ext4_dx_a= dd_entry.ext4_add_entry.ext4_add_nondir.ext4_create.vfs_create.path_openat 27954 =B1 1% -86.6% 3734 =B1 49% latency_stats.hits.ext4_m= b_initialize_context.ext4_mb_new_blocks.ext4_ext_map_blocks.ext4_map_blocks= .ext4_writepages.do_writepages.__filemap_fdatawrite_range.filemap_write_and= _wait_range.ext4_sync_file.vfs_fsync_range.do_fsync.SyS_fsync 290546 =B1 16% -85.7% 41420 =B1 36% latency_stats.max.do_get_= write_access.jbd2_journal_get_write_access.__ext4_journal_get_write_access.= __ext4_new_inode.ext4_create.vfs_create.path_openat.do_filp_open.do_sys_ope= n.SyS_open.entry_SYSCALL_64_fastpath 265287 =B1 9% -85.9% 37329 =B1 63% latency_stats.max.do_get_= write_access.jbd2_journal_get_write_access.__ext4_journal_get_write_access.= ext4_mb_mark_diskspace_used.ext4_mb_new_blocks.ext4_ext_map_blocks.ext4_map= _blocks.ext4_writepages.do_writepages.__filemap_fdatawrite_range.filemap_wr= ite_and_wait_range.ext4_sync_file 0.00 =B1 -1% +Inf% 16683 =B1 61% latency_stats.max.do_get_= write_access.jbd2_journal_get_write_access.__ext4_journal_get_write_access.= ext4_reserve_inode_write.ext4_mark_inode_dirty.add_dirent_to_buf.ext4_add_e= ntry.ext4_add_nondir.ext4_create.vfs_create.path_openat.do_filp_open 0.00 =B1 -1% +Inf% 29279 =B1 73% latency_stats.max.do_get_= write_access.jbd2_journal_get_write_access.__ext4_journal_get_write_access.= ext4_reserve_inode_write.ext4_mark_inode_dirty.add_dirent_to_buf.ext4_dx_ad= d_entry.ext4_add_entry.ext4_add_nondir.ext4_create.vfs_create.path_openat 63561 =B1165% -99.6% 267.75 =B1173% latency_stats.max.do_get_= write_access.jbd2_journal_get_write_access.__ext4_journal_get_write_access.= ext4_reserve_inode_write.ext4_mark_inode_dirty.ext4_dirty_inode.__mark_inod= e_dirty.generic_update_time.file_update_time.__generic_file_write_iter.ext4= _file_write_iter.__vfs_write 62202 =B1 74% -95.4% 2841 =B1 21% latency_stats.max.do_get_= write_access.jbd2_journal_get_write_access.__ext4_journal_get_write_access.= ext4_reserve_inode_write.ext4_mark_inode_dirty.ext4_dirty_inode.__mark_inod= e_dirty.generic_write_end.ext4_da_write_end.generic_perform_write.__generic= _file_write_iter.ext4_file_write_iter 0.00 =B1 -1% +Inf% 37841 =B1 39% latency_stats.max.do_get_= write_access.jbd2_journal_get_write_access.__ext4_journal_get_write_access.= ext4_reserve_inode_write.ext4_mark_inode_dirty.ext4_ext_tree_init.__ext4_ne= w_inode.ext4_create.vfs_create.path_openat.do_filp_open.do_sys_open 279828 =B1 21% -94.5% 15429 =B1 97% latency_stats.max.ext4_mb= _initialize_context.ext4_mb_new_blocks.ext4_ext_map_blocks.ext4_map_blocks.= ext4_writepages.do_writepages.__filemap_fdatawrite_range.filemap_write_and_= wait_range.ext4_sync_file.vfs_fsync_range.do_fsync.SyS_fsync 176558 =B1 21% -95.8% 7488 =B1102% latency_stats.max.submit_= bio_wait.blkdev_issue_flush.ext4_sync_file.vfs_fsync_range.do_fsync.SyS_fsy= nc.entry_SYSCALL_64_fastpath 381747 =B1 21% -100.0% 0.00 =B1 -1% latency_stats.max.submit_= bio_wait.blkdev_issue_flush.jbd2_cleanup_journal_tail.jbd2_log_do_checkpoin= t.__jbd2_log_wait_for_space.add_transaction_credits.start_this_handle.jbd2_= _journal_start.__ext4_journal_start_sb.__ext4_new_inode.ext4_create.vfs_cre= ate 165796 =B1 24% -100.0% 0.00 =B1 -1% latency_stats.max.submit_= bio_wait.blkdev_issue_flush.jbd2_cleanup_journal_tail.jbd2_log_do_checkpoin= t.__jbd2_log_wait_for_space.add_transaction_credits.start_this_handle.jbd2_= _journal_start.__ext4_journal_start_sb.ext4_da_write_begin.generic_perform_= write.__generic_file_write_iter 406817 =B1 14% -100.0% 0.00 =B1 -1% latency_stats.max.submit_= bio_wait.blkdev_issue_flush.jbd2_cleanup_journal_tail.jbd2_log_do_checkpoin= t.__jbd2_log_wait_for_space.add_transaction_credits.start_this_handle.jbd2_= _journal_start.__ext4_journal_start_sb.ext4_dirty_inode.__mark_inode_dirty.= generic_update_time 165247 =B1 40% -100.0% 0.00 =B1 -1% latency_stats.max.submit_= bio_wait.blkdev_issue_flush.jbd2_cleanup_journal_tail.jbd2_log_do_checkpoin= t.__jbd2_log_wait_for_space.add_transaction_credits.start_this_handle.jbd2_= _journal_start.__ext4_journal_start_sb.ext4_writepages.do_writepages.__file= map_fdatawrite_range 324551 =B1 12% -91.2% 28575 =B1 43% latency_stats.max.wait_tr= ansaction_locked.add_transaction_credits.start_this_handle.jbd2__journal_st= art.__ext4_journal_start_sb.__ext4_new_inode.ext4_create.vfs_create.path_op= enat.do_filp_open.do_sys_open.SyS_open 458237 =B1 15% -99.5% 2140 =B1 76% latency_stats.max.wait_tr= ansaction_locked.add_transaction_credits.start_this_handle.jbd2__journal_st= art.__ext4_journal_start_sb.ext4_dirty_inode.__mark_inode_dirty.generic_upd= ate_time.file_update_time.__generic_file_write_iter.ext4_file_write_iter.__= vfs_write 234367 =B1 47% -94.6% 12766 =B1 80% latency_stats.max.wait_tr= ansaction_locked.add_transaction_credits.start_this_handle.jbd2__journal_st= art.__ext4_journal_start_sb.ext4_writepages.do_writepages.__filemap_fdatawr= ite_range.filemap_write_and_wait_range.ext4_sync_file.vfs_fsync_range.do_fs= ync 0.00 =B1 -1% +Inf% 23646 =B1123% latency_stats.sum.do_get_= write_access.jbd2_journal_get_write_access.__ext4_journal_get_write_access.= ext4_mb_mark_diskspace_used.ext4_mb_new_blocks.ext4_ext_map_blocks.ext4_map= _blocks.ext4_getblk.ext4_bread.ext4_append.do_split.ext4_dx_add_entry 0.00 =B1 -1% +Inf% 30398 =B1 50% latency_stats.sum.do_get_= write_access.jbd2_journal_get_write_access.__ext4_journal_get_write_access.= ext4_mb_mark_diskspace_used.ext4_mb_new_blocks.ext4_ext_map_blocks.ext4_map= _blocks.ext4_getblk.ext4_bread.ext4_append.ext4_mkdir.vfs_mkdir 0.00 =B1 -1% +Inf% 56525 =B1105% latency_stats.sum.do_get_= write_access.jbd2_journal_get_write_access.__ext4_journal_get_write_access.= ext4_mb_mark_diskspace_used.ext4_mb_new_blocks.ext4_ext_map_blocks.ext4_map= _blocks.ext4_getblk.ext4_bread.ext4_append.make_indexed_dir.ext4_add_entry 41228828 =B1 7% -92.4% 3149423 =B1 67% latency_stats.sum.do_get_= write_access.jbd2_journal_get_write_access.__ext4_journal_get_write_access.= ext4_mb_mark_diskspace_used.ext4_mb_new_blocks.ext4_ext_map_blocks.ext4_map= _blocks.ext4_writepages.do_writepages.__filemap_fdatawrite_range.filemap_wr= ite_and_wait_range.ext4_sync_file 0.00 =B1 -1% +Inf% 1241212 =B1 55% latency_stats.sum.do_get_= write_access.jbd2_journal_get_write_access.__ext4_journal_get_write_access.= ext4_reserve_inode_write.ext4_mark_inode_dirty.add_dirent_to_buf.ext4_add_e= ntry.ext4_add_nondir.ext4_create.vfs_create.path_openat.do_filp_open 0.00 =B1 -1% +Inf% 2727180 =B1 46% latency_stats.sum.do_get_= write_access.jbd2_journal_get_write_access.__ext4_journal_get_write_access.= ext4_reserve_inode_write.ext4_mark_inode_dirty.add_dirent_to_buf.ext4_dx_ad= d_entry.ext4_add_entry.ext4_add_nondir.ext4_create.vfs_create.path_openat 0.00 =B1 -1% +Inf% 88155 =B1 72% latency_stats.sum.do_get_= write_access.jbd2_journal_get_write_access.__ext4_journal_get_write_access.= ext4_reserve_inode_write.ext4_mark_inode_dirty.ext4_dirty_inode.__mark_inod= e_dirty.ext4_mb_new_blocks.ext4_ext_map_blocks.ext4_map_blocks.ext4_getblk.= ext4_bread 200757 =B1159% -99.5% 988.25 =B1173% latency_stats.sum.do_get_= write_access.jbd2_journal_get_write_access.__ext4_journal_get_write_access.= ext4_reserve_inode_write.ext4_mark_inode_dirty.ext4_dirty_inode.__mark_inod= e_dirty.generic_update_time.file_update_time.__generic_file_write_iter.ext4= _file_write_iter.__vfs_write 1755344 =B1 9% -94.1% 103667 =B1 53% latency_stats.sum.do_get_= write_access.jbd2_journal_get_write_access.__ext4_journal_get_write_access.= ext4_reserve_inode_write.ext4_mark_inode_dirty.ext4_dirty_inode.__mark_inod= e_dirty.generic_write_end.ext4_da_write_end.generic_perform_write.__generic= _file_write_iter.ext4_file_write_iter 0.00 =B1 -1% +Inf% 4885655 =B1 35% latency_stats.sum.do_get_= write_access.jbd2_journal_get_write_access.__ext4_journal_get_write_access.= ext4_reserve_inode_write.ext4_mark_inode_dirty.ext4_ext_tree_init.__ext4_ne= w_inode.ext4_create.vfs_create.path_openat.do_filp_open.do_sys_open 0.00 =B1 -1% +Inf% 35213 =B1 40% latency_stats.sum.do_get_= write_access.jbd2_journal_get_write_access.__ext4_journal_get_write_access.= ext4_reserve_inode_write.ext4_mark_inode_dirty.ext4_ext_tree_init.__ext4_ne= w_inode.ext4_mkdir.vfs_mkdir.SyS_mkdir.entry_SYSCALL_64_fastpath 35021878 =B1 12% -97.3% 959921 =B1 59% latency_stats.sum.ext4_mb= _initialize_context.ext4_mb_new_blocks.ext4_ext_map_blocks.ext4_map_blocks.= ext4_writepages.do_writepages.__filemap_fdatawrite_range.filemap_write_and_= wait_range.ext4_sync_file.vfs_fsync_range.do_fsync.SyS_fsync 2610110 =B1 10% -99.6% 10320 =B1115% latency_stats.sum.submit_= bio_wait.blkdev_issue_flush.ext4_sync_file.vfs_fsync_range.do_fsync.SyS_fsy= nc.entry_SYSCALL_64_fastpath 7233874 =B1 35% -100.0% 0.00 =B1 -1% latency_stats.sum.submit_= bio_wait.blkdev_issue_flush.jbd2_cleanup_journal_tail.jbd2_log_do_checkpoin= t.__jbd2_log_wait_for_space.add_transaction_credits.start_this_handle.jbd2_= _journal_start.__ext4_journal_start_sb.__ext4_new_inode.ext4_create.vfs_cre= ate 1159776 =B1 50% -100.0% 0.00 =B1 -1% latency_stats.sum.submit_= bio_wait.blkdev_issue_flush.jbd2_cleanup_journal_tail.jbd2_log_do_checkpoin= t.__jbd2_log_wait_for_space.add_transaction_credits.start_this_handle.jbd2_= _journal_start.__ext4_journal_start_sb.ext4_da_write_begin.generic_perform_= write.__generic_file_write_iter 822013 =B1 35% -100.0% 0.00 =B1 -1% latency_stats.sum.submit_= bio_wait.blkdev_issue_flush.jbd2_cleanup_journal_tail.jbd2_log_do_checkpoin= t.__jbd2_log_wait_for_space.add_transaction_credits.start_this_handle.jbd2_= _journal_start.__ext4_journal_start_sb.ext4_dirty_inode.__mark_inode_dirty.= generic_update_time 810979 =B1 49% -100.0% 0.00 =B1 -1% latency_stats.sum.submit_= bio_wait.blkdev_issue_flush.jbd2_cleanup_journal_tail.jbd2_log_do_checkpoin= t.__jbd2_log_wait_for_space.add_transaction_credits.start_this_handle.jbd2_= _journal_start.__ext4_journal_start_sb.ext4_writepages.do_writepages.__file= map_fdatawrite_range 3757907 =B1 7% -99.8% 7957 =B1 37% latency_stats.sum.wait_tr= ansaction_locked.add_transaction_credits.start_this_handle.jbd2__journal_st= art.__ext4_journal_start_sb.ext4_dirty_inode.__mark_inode_dirty.generic_upd= ate_time.file_update_time.__generic_file_write_iter.ext4_file_write_iter.__= vfs_write 1020128 =B1 78% -94.6% 54940 =B1 73% latency_stats.sum.wait_tr= ansaction_locked.add_transaction_credits.start_this_handle.jbd2__journal_st= art.__ext4_journal_start_sb.ext4_writepages.do_writepages.__filemap_fdatawr= ite_range.filemap_write_and_wait_range.ext4_sync_file.vfs_fsync_range.do_fs= ync 5691 =B1 8% -22.7% 4398 =B1 14% sched_debug.cfs_rq:/.exec= _clock.0 854.81 =B1 10% +157.0% 2196 =B1 27% sched_debug.cfs_rq:/.exec= _clock.11 2946 =B1 3% -16.0% 2475 =B1 4% sched_debug.cfs_rq:/.exec= _clock.4 2075 =B1 6% -19.3% 1673 =B1 10% sched_debug.cfs_rq:/.exec= _clock.8 2.75 =B1 30% +318.2% 11.50 =B1 58% sched_debug.cfs_rq:/.load= _avg.1 5.25 =B1 75% +266.7% 19.25 =B1 20% sched_debug.cfs_rq:/.load= _avg.7 12.00 =B1 31% -70.8% 3.50 =B1 24% sched_debug.cfs_rq:/.load= _avg.8 1.25 =B1 34% +860.0% 12.00 =B1 33% sched_debug.cfs_rq:/.load= _avg.9 0.22 =B1 37% +116.7% 0.47 =B1 19% sched_debug.cfs_rq:/.load= _avg.min 6649 =B1 29% -31.8% 4537 =B1 15% sched_debug.cfs_rq:/.min_= vruntime.8 7953 =B1 4% -12.7% 6940 =B1 3% sched_debug.cfs_rq:/.min_= vruntime.avg 19991 =B1 6% -22.3% 15531 =B1 5% sched_debug.cfs_rq:/.min_= vruntime.max 4514 =B1 8% -25.9% 3343 =B1 9% sched_debug.cfs_rq:/.min_= vruntime.stddev 0.75 =B1110% +433.3% 4.00 =B1 30% sched_debug.cfs_rq:/.nr_s= pread_over.13 4514 =B1 8% -25.9% 3343 =B1 9% sched_debug.cfs_rq:/.spre= ad0.stddev 22.00 =B1 19% +104.5% 45.00 =B1 19% sched_debug.cfs_rq:/.util= _avg.2 361962 =B1 3% -24.9% 271884 =B1 0% sched_debug.cpu.clock.0 361962 =B1 3% -24.9% 271883 =B1 0% sched_debug.cpu.clock.1 361968 =B1 3% -24.9% 271888 =B1 0% sched_debug.cpu.clock.10 361969 =B1 3% -24.9% 271884 =B1 0% sched_debug.cpu.clock.11 361969 =B1 3% -24.9% 271889 =B1 0% sched_debug.cpu.clock.12 361963 =B1 3% -24.9% 271887 =B1 0% sched_debug.cpu.clock.13 361971 =B1 3% -24.9% 271890 =B1 0% sched_debug.cpu.clock.14 361966 =B1 3% -24.9% 271893 =B1 0% sched_debug.cpu.clock.15 361965 =B1 3% -24.9% 271886 =B1 0% sched_debug.cpu.clock.2 361945 =B1 3% -24.9% 271880 =B1 0% sched_debug.cpu.clock.3 361968 =B1 3% -24.9% 271885 =B1 0% sched_debug.cpu.clock.4 361964 =B1 3% -24.9% 271890 =B1 0% sched_debug.cpu.clock.5 361971 =B1 3% -24.9% 271886 =B1 0% sched_debug.cpu.clock.6 361967 =B1 3% -24.9% 271890 =B1 0% sched_debug.cpu.clock.7 361963 =B1 3% -24.9% 271886 =B1 0% sched_debug.cpu.clock.8 361965 =B1 3% -24.9% 271881 =B1 0% sched_debug.cpu.clock.9 361965 =B1 3% -24.9% 271886 =B1 0% sched_debug.cpu.clock.avg 361973 =B1 3% -24.9% 271893 =B1 0% sched_debug.cpu.clock.max 361913 =B1 3% -24.9% 271854 =B1 0% sched_debug.cpu.clock.min 15.52 =B1 18% -27.6% 11.24 =B1 25% sched_debug.cpu.clock.std= dev 361962 =B1 3% -24.9% 271884 =B1 0% sched_debug.cpu.clock_tas= k.0 361962 =B1 3% -24.9% 271883 =B1 0% sched_debug.cpu.clock_tas= k.1 361968 =B1 3% -24.9% 271888 =B1 0% sched_debug.cpu.clock_tas= k.10 361969 =B1 3% -24.9% 271884 =B1 0% sched_debug.cpu.clock_tas= k.11 361969 =B1 3% -24.9% 271889 =B1 0% sched_debug.cpu.clock_tas= k.12 361963 =B1 3% -24.9% 271887 =B1 0% sched_debug.cpu.clock_tas= k.13 361971 =B1 3% -24.9% 271890 =B1 0% sched_debug.cpu.clock_tas= k.14 361966 =B1 3% -24.9% 271893 =B1 0% sched_debug.cpu.clock_tas= k.15 361965 =B1 3% -24.9% 271886 =B1 0% sched_debug.cpu.clock_tas= k.2 361945 =B1 3% -24.9% 271880 =B1 0% sched_debug.cpu.clock_tas= k.3 361968 =B1 3% -24.9% 271885 =B1 0% sched_debug.cpu.clock_tas= k.4 361964 =B1 3% -24.9% 271890 =B1 0% sched_debug.cpu.clock_tas= k.5 361971 =B1 3% -24.9% 271886 =B1 0% sched_debug.cpu.clock_tas= k.6 361967 =B1 3% -24.9% 271890 =B1 0% sched_debug.cpu.clock_tas= k.7 361963 =B1 3% -24.9% 271886 =B1 0% sched_debug.cpu.clock_tas= k.8 361965 =B1 3% -24.9% 271881 =B1 0% sched_debug.cpu.clock_tas= k.9 361965 =B1 3% -24.9% 271886 =B1 0% sched_debug.cpu.clock_tas= k.avg 361973 =B1 3% -24.9% 271893 =B1 0% sched_debug.cpu.clock_tas= k.max 361913 =B1 3% -24.9% 271854 =B1 0% sched_debug.cpu.clock_tas= k.min 15.52 =B1 18% -27.6% 11.24 =B1 25% sched_debug.cpu.clock_tas= k.stddev 594.58 =B1 12% -30.7% 412.29 =B1 12% sched_debug.cpu.curr->pid= .avg 8265 =B1 6% -36.1% 5285 =B1 12% sched_debug.cpu.curr->pid= .max 2062 =B1 7% -34.6% 1348 =B1 11% sched_debug.cpu.curr->pid= .stddev 25768 =B1 4% -11.4% 22823 =B1 7% sched_debug.cpu.nr_load_u= pdates.6 7155 =B1 5% -21.9% 5586 =B1 14% sched_debug.cpu.nr_load_u= pdates.stddev 7911 =B1 9% +225.8% 25779 =B1 60% sched_debug.cpu.nr_switch= es.11 61043 =B1 5% -20.1% 48763 =B1 23% sched_debug.cpu.nr_switch= es.14 11503 =B1 41% +164.9% 30476 =B1 57% sched_debug.cpu.nr_switch= es.15 7955 =B1 3% +138.1% 18945 =B1 39% sched_debug.cpu.nr_switch= es.9 7498 =B1 6% +134.5% 17581 =B1 40% sched_debug.cpu.nr_switch= es.min 61163 =B1 7% -18.1% 50102 =B1 9% sched_debug.cpu.nr_switch= es.stddev -33543 =B1 -4% -41.2% -19708 =B1-38% sched_debug.cpu.nr_uninte= rruptible.0 3853 =B1 6% -52.0% 1849 =B1 41% sched_debug.cpu.nr_uninte= rruptible.10 86.75 =B1 59% +195.4% 256.25 =B1 22% sched_debug.cpu.nr_uninte= rruptible.11 3849 =B1 10% -48.7% 1974 =B1 45% sched_debug.cpu.nr_uninte= rruptible.12 83.25 =B1 66% +253.8% 294.50 =B1 31% sched_debug.cpu.nr_uninte= rruptible.13 3930 =B1 7% -48.4% 2029 =B1 42% sched_debug.cpu.nr_uninte= rruptible.14 77.00 =B1 35% +241.6% 263.00 =B1 25% sched_debug.cpu.nr_uninte= rruptible.15 4437 =B1 11% -55.8% 1959 =B1 55% sched_debug.cpu.nr_uninte= rruptible.2 4509 =B1 5% -51.1% 2206 =B1 51% sched_debug.cpu.nr_uninte= rruptible.4 4507 =B1 9% -50.6% 2226 =B1 51% sched_debug.cpu.nr_uninte= rruptible.6 7142 =B1 6% -29.1% 5065 =B1 29% sched_debug.cpu.nr_uninte= rruptible.8 87.25 =B1 42% +204.3% 265.50 =B1 24% sched_debug.cpu.nr_uninte= rruptible.9 7146 =B1 6% -29.1% 5065 =B1 29% sched_debug.cpu.nr_uninte= rruptible.max -33542 =B1 -4% -41.2% -19707 =B1-38% sched_debug.cpu.nr_uninte= rruptible.min 8953 =B1 4% -41.4% 5250 =B1 38% sched_debug.cpu.nr_uninte= rruptible.stddev 7958 =B1 9% +224.4% 25817 =B1 60% sched_debug.cpu.sched_cou= nt.11 11549 =B1 40% +169.2% 31095 =B1 59% sched_debug.cpu.sched_cou= nt.15 8003 =B1 3% +137.1% 18980 =B1 39% sched_debug.cpu.sched_cou= nt.9 7638 =B1 6% +130.6% 17616 =B1 40% sched_debug.cpu.sched_cou= nt.min 3539 =B1 9% +217.3% 11231 =B1 61% sched_debug.cpu.sched_goi= dle.11 5137 =B1 46% +159.0% 13303 =B1 59% sched_debug.cpu.sched_goi= dle.15 3543 =B1 4% +119.8% 7789 =B1 37% sched_debug.cpu.sched_goi= dle.9 3327 =B1 5% +112.9% 7083 =B1 37% sched_debug.cpu.sched_goi= dle.min 28088 =B1 7% -18.9% 22770 =B1 7% sched_debug.cpu.sched_goi= dle.stddev 40517 =B1 77% -70.7% 11888 =B1 56% sched_debug.cpu.ttwu_coun= t.15 119597 =B1 3% -30.9% 82689 =B1 19% sched_debug.cpu.ttwu_loca= l.0 4054 =B1 8% +44.6% 5862 =B1 23% sched_debug.cpu.ttwu_loca= l.1 5950 =B1 3% -32.2% 4036 =B1 21% sched_debug.cpu.ttwu_loca= l.10 1259 =B1 9% +154.6% 3206 =B1 43% sched_debug.cpu.ttwu_loca= l.13 1530 =B1 8% +126.2% 3461 =B1 40% sched_debug.cpu.ttwu_loca= l.15 8444 =B1 6% -25.1% 6328 =B1 12% sched_debug.cpu.ttwu_loca= l.2 3189 =B1 15% +53.0% 4878 =B1 31% sched_debug.cpu.ttwu_loca= l.3 6155 =B1 9% -20.3% 4905 =B1 6% sched_debug.cpu.ttwu_loca= l.4 6013 =B1 12% -32.9% 4033 =B1 8% sched_debug.cpu.ttwu_loca= l.6 9433 =B1 2% -19.1% 7632 =B1 8% sched_debug.cpu.ttwu_loca= l.8 1303 =B1 12% +131.5% 3016 =B1 38% sched_debug.cpu.ttwu_loca= l.9 11588 =B1 2% -18.7% 9421 =B1 5% sched_debug.cpu.ttwu_loca= l.avg 119620 =B1 3% -30.9% 82707 =B1 19% sched_debug.cpu.ttwu_loca= l.max 1185 =B1 12% +85.3% 2196 =B1 26% sched_debug.cpu.ttwu_loca= l.min 28011 =B1 3% -32.1% 19019 =B1 21% sched_debug.cpu.ttwu_loca= l.stddev 361971 =B1 3% -24.9% 271891 =B1 0% sched_debug.cpu_clk 360052 =B1 3% -24.9% 270234 =B1 0% sched_debug.ktime 361971 =B1 3% -24.9% 271891 =B1 0% sched_debug.sched_clk lkp-ne04: Nehalem-EP Memory: 12G turbostat._Busy 1.5 ++-------------------------------------------------------------------= -+ | O O = | 1.4 ++ O = | 1.3 ++ = | | O O O O = | 1.2 ++ O O O O = | | O O O O O = O 1.1 O+ O O O O O O = | | O O O = | 1 ++ = | | = | 0.9 ++ .*.. *. .*.. *.. = | 0.8 *+.*.*.. .*..*. *.*.. .*. .. *. *.*.. .. *.*..*..*.*.= .* | *..* *. *..* * = | 0.7 ++-------------------------------------------------------------------= -+ uptime.idle 10500 ++-----------------------------------------------------------------= -+ * *.. .*.. = | 10000 ++ *.. .. * *.*.. *..*.. .*..*. .*..* *. = | 9500 +++ + .* + .. + *.*..*.*. *. *.= .| | * * *..* * = * 9000 ++ = | 8500 ++ = | | = | 8000 ++ O O O O = | 7500 O+ O O O O O O O O O = | | O O O O O = O 7000 ++ O O O O = | 6500 ++ O O = | | = | 6000 ++-----------------------------------O-----------------------------= -+ fsmark.time.voluntary_context_switches 500000 ++----------------------------------------------------------------= -+ 490000 ++ O O O O O O O = | O O O O O O O O O O = | 480000 ++ O O O O O O O O O = O 470000 ++ = | | = | 460000 ++ = | 450000 ++ = | 440000 ++ = | | = | 430000 ++ = | 420000 ++ .*.. *.. *. = | *.. .*.. .*.*.. .*.*..*.*..*..* + .*..*.. .. *.. = | 410000 ++ * *.*. *. * * *.* *.*.= .| 400000 ++----------------------------------------------------------------= -* fsmark.time.file_system_inputs 40000 ++-----------------------------------------------------------------= -+ | O = | 35000 ++ = | 30000 ++ O O = | | O O O O O O O = | 25000 O+ O O O O O O O O O O O = O | O O = | 20000 ++ O O = | | = | 15000 ++ = | 10000 ++ = | | = | 5000 ++ = | | = | 0 *+-*-*--*-*--*--*-*--*--*-*--*-*--*--*-*--*-*--*--*-*--*--*-*--*-*-= -* proc-vmstat.pgpgin 20000 ++-----------------------------------------------------------------= -+ 18000 ++ O = | | = | 16000 ++ O O = | 14000 ++ O O O O O O O = | O O O O O O O O O O O O = O 12000 ++ O O = | 10000 ++ O O = | 8000 ++ = | | = | 6000 ++ = | 4000 ++ = | | = | 2000 ++ = | 0 *+-*-*--*-*--*--*-*--*--*-*--*-*--*--*-*--*-*--*--*-*--*--*-*--*-*-= -* proc-vmstat.pgalloc_normal 1.35e+06 ++--------------------------------------------------------------= -+ 1.3e+06 ++ * = | * *. *..* *. .. + = | 1.25e+06 ++ *.. .. *.. + + .*..*. .. * *.. = | 1.2e+06 +++ + .* *.*..*.*..* *..*.*..* * *= .| | * * = * 1.15e+06 ++ = | 1.1e+06 ++ = | 1.05e+06 ++ O O O O = | O O O O O O O O O O O O O = | 1e+06 ++ O O = O 950000 ++ O O O O = | | O O= | 900000 ++ O = | 850000 ++--------------------------------------------------------------= -+ [*] bisect-good sample [O] bisect-bad sample To reproduce: git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tes= ts.git cd lkp-tests bin/lkp install job.yaml # job file is attached in this email bin/lkp run job.yaml Disclaimer: Results have been estimated based on internal Intel analysis and are provid= ed for informational purposes only. Any difference in system hardware or softw= are design or configuration may affect actual performance. Thanks, Ying Huang --=-=-= Content-Type: text/plain; charset=ascii Content-Disposition: attachment; filename=job.yaml --- LKP_SERVER: inn LKP_CGI_PORT: 80 LKP_CIFS_PORT: 139 testcase: fsmark default-monitors: wait: activate-monitor kmsg: uptime: iostat: heartbeat: vmstat: numa-numastat: numa-vmstat: numa-meminfo: proc-vmstat: proc-stat: interval: 10 meminfo: slabinfo: interrupts: lock_stat: latency_stats: softirqs: bdi_dev_mapping: diskstats: nfsstat: cpuidle: cpufreq-stats: turbostat: pmeter: sched_debug: interval: 60 cpufreq_governor: performance default-watchdogs: oom-killer: watchdog: commit: 1f2d779fed21806ffed84aa65617c6125f480e85 model: Nehalem-EP memory: 12G hdd_partitions: "/dev/disk/by-id/ata-ST3500514NS_9WJ03EBA-part3" swap_partitions: "/dev/disk/by-id/ata-ST3120026AS_5MS07HA2-part2" rootfs_partition: "/dev/disk/by-id/ata-ST3500514NS_9WJ03EBA-part1" category: benchmark iterations: 1x nr_threads: 32t disk: 1HDD fs: ext4 fs2: fsmark: filesize: 9B test_size: 400M sync_method: fsyncBeforeClose nr_directories: 16d nr_files_per_directory: 256fpd queue: bisect testbox: lkp-ne04 tbox_group: lkp-ne04 kconfig: x86_64-rhel enqueue_time: 2016-02-28 17:23:12.481767277 +08:00 compiler: gcc-4.9 rootfs: debian-x86_64-2015-02-07.cgz id: 68aa8f695a0e8531c991807de120bb153d0c18de user: lkp head_commit: 1736b6892f32b4fc71f44c86f026bd5729e37e1c base_commit: 81f70ba233d5f660e1ea5fe23260ee323af5d53a branch: linux-devel/devel-hourly-2016022721 result_root: "/result/fsmark/performance-1x-32t-1HDD-ext4-9B-400M-fsyncBeforeClose-16d-256fpd/lkp-ne04/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/1f2d779fed21806ffed84aa65617c6125f480e85/0" job_file: "/lkp/scheduled/lkp-ne04/bisect_fsmark-performance-1x-32t-1HDD-ext4-9B-400M-fsyncBeforeClose-16d-256fpd-debian-x86_64-2015-02-07.cgz-x86_64-rhel-1f2d779fed21806ffed84aa65617c6125f480e85-20160228-126629-2r1e6c-0.yaml" nr_cpu: "$(nproc)" max_uptime: 2384.62 initrd: "/osimage/debian/debian-x86_64-2015-02-07.cgz" bootloader_append: - root=/dev/ram0 - user=lkp - job=/lkp/scheduled/lkp-ne04/bisect_fsmark-performance-1x-32t-1HDD-ext4-9B-400M-fsyncBeforeClose-16d-256fpd-debian-x86_64-2015-02-07.cgz-x86_64-rhel-1f2d779fed21806ffed84aa65617c6125f480e85-20160228-126629-2r1e6c-0.yaml - ARCH=x86_64 - kconfig=x86_64-rhel - branch=linux-devel/devel-hourly-2016022721 - commit=1f2d779fed21806ffed84aa65617c6125f480e85 - BOOT_IMAGE=/pkg/linux/x86_64-rhel/gcc-4.9/1f2d779fed21806ffed84aa65617c6125f480e85/vmlinuz-4.5.0-rc2-00026-g1f2d779 - max_uptime=2384 - RESULT_ROOT=/result/fsmark/performance-1x-32t-1HDD-ext4-9B-400M-fsyncBeforeClose-16d-256fpd/lkp-ne04/debian-x86_64-2015-02-07.cgz/x86_64-rhel/gcc-4.9/1f2d779fed21806ffed84aa65617c6125f480e85/0 - LKP_SERVER=inn - |2- earlyprintk=ttyS0,115200 systemd.log_level=err debug apic=debug sysrq_always_enabled rcupdate.rcu_cpu_stall_timeout=100 panic=-1 softlockup_panic=1 nmi_watchdog=panic oops=panic load_ramdisk=2 prompt_ramdisk=0 console=ttyS0,115200 console=tty0 vga=normal rw lkp_initrd: "/lkp/lkp/lkp-x86_64.cgz" modules_initrd: "/pkg/linux/x86_64-rhel/gcc-4.9/1f2d779fed21806ffed84aa65617c6125f480e85/modules.cgz" bm_initrd: "/osimage/deps/debian-x86_64-2015-02-07.cgz/lkp.cgz,/osimage/deps/debian-x86_64-2015-02-07.cgz/run-ipconfig.cgz,/osimage/deps/debian-x86_64-2015-02-07.cgz/turbostat.cgz,/lkp/benchmarks/turbostat.cgz,/osimage/deps/debian-x86_64-2015-02-07.cgz/fs.cgz,/osimage/deps/debian-x86_64-2015-02-07.cgz/fs2.cgz,/lkp/benchmarks/fsmark.cgz" linux_headers_initrd: "/pkg/linux/x86_64-rhel/gcc-4.9/1f2d779fed21806ffed84aa65617c6125f480e85/linux-headers.cgz" repeat_to: 2 kernel: "/pkg/linux/x86_64-rhel/gcc-4.9/1f2d779fed21806ffed84aa65617c6125f480e85/vmlinuz-4.5.0-rc2-00026-g1f2d779" dequeue_time: 2016-02-28 17:39:51.598790914 +08:00 job_state: finished loadavg: 30.66 26.71 13.96 2/234 13815 start_time: '1456652435' end_time: '1456652948' version: "/lkp/lkp/.src-20160226-194908" --=-=-= Content-Type: text/plain; charset=ascii Content-Disposition: attachment; filename=reproduce 2016-02-28 17:40:32 echo performance > /sys/devices/system/cpu/cpu0/cpufreq/scaling_governor 2016-02-28 17:40:32 echo performance > /sys/devices/system/cpu/cpu1/cpufreq/scaling_governor 2016-02-28 17:40:32 echo performance > /sys/devices/system/cpu/cpu10/cpufreq/scaling_governor 2016-02-28 17:40:32 echo performance > /sys/devices/system/cpu/cpu11/cpufreq/scaling_governor 2016-02-28 17:40:32 echo performance > /sys/devices/system/cpu/cpu12/cpufreq/scaling_governor 2016-02-28 17:40:32 echo performance > /sys/devices/system/cpu/cpu13/cpufreq/scaling_governor 2016-02-28 17:40:32 echo performance > /sys/devices/system/cpu/cpu14/cpufreq/scaling_governor 2016-02-28 17:40:32 echo performance > /sys/devices/system/cpu/cpu15/cpufreq/scaling_governor 2016-02-28 17:40:32 echo performance > /sys/devices/system/cpu/cpu2/cpufreq/scaling_governor 2016-02-28 17:40:32 echo performance > /sys/devices/system/cpu/cpu3/cpufreq/scaling_governor 2016-02-28 17:40:32 echo performance > /sys/devices/system/cpu/cpu4/cpufreq/scaling_governor 2016-02-28 17:40:32 echo performance > /sys/devices/system/cpu/cpu5/cpufreq/scaling_governor 2016-02-28 17:40:32 echo performance > /sys/devices/system/cpu/cpu6/cpufreq/scaling_governor 2016-02-28 17:40:32 echo performance > /sys/devices/system/cpu/cpu7/cpufreq/scaling_governor 2016-02-28 17:40:32 echo performance > /sys/devices/system/cpu/cpu8/cpufreq/scaling_governor 2016-02-28 17:40:32 echo performance > /sys/devices/system/cpu/cpu9/cpufreq/scaling_governor 2016-02-28 17:40:32 mkfs -t ext4 -q -F /dev/sda3 2016-02-28 17:40:34 mount -t ext4 /dev/sda3 /fs/sda3 2016-02-28 17:40:35 ./fs_mark -d /fs/sda3/1 -d /fs/sda3/2 -d /fs/sda3/3 -d /fs/sda3/4 -d /fs/sda3/5 -d /fs/sda3/6 -d /fs/sda3/7 -d /fs/sda3/8 -d /fs/sda3/9 -d /fs/sda3/10 -d /fs/sda3/11 -d /fs/sda3/12 -d /fs/sda3/13 -d /fs/sda3/14 -d /fs/sda3/15 -d /fs/sda3/16 -d /fs/sda3/17 -d /fs/sda3/18 -d /fs/sda3/19 -d /fs/sda3/20 -d /fs/sda3/21 -d /fs/sda3/22 -d /fs/sda3/23 -d /fs/sda3/24 -d /fs/sda3/25 -d /fs/sda3/26 -d /fs/sda3/27 -d /fs/sda3/28 -d /fs/sda3/29 -d /fs/sda3/30 -d /fs/sda3/31 -d /fs/sda3/32 -D 16 -N 256 -n 3200 -L 1 -S 1 -s 9 --=-=-=--