public inbox for linux-kernel@vger.kernel.org
 help / color / mirror / Atom feed
From: kernel test robot <oliver.sang@intel.com>
To: Benjamin Segall <bsegall@google.com>
Cc: <oe-lkp@lists.linux.dev>, <lkp@intel.com>,
	<linux-kernel@vger.kernel.org>,
	Peter Zijlstra <peterz@infradead.org>, <ying.huang@intel.com>,
	<feng.tang@intel.com>, <fengwei.yin@intel.com>,
	<aubrey.li@linux.intel.com>, <yu.c.chen@intel.com>,
	<oliver.sang@intel.com>
Subject: [linus:master] [sched/eevdf]  b01db23d59:  hackbench.throughput -3.4% regression
Date: Wed, 18 Oct 2023 22:55:09 +0800	[thread overview]
Message-ID: <202310182229.78d950b2-oliver.sang@intel.com> (raw)



Hello,

kernel test robot noticed a -3.4% regression of hackbench.throughput on:


commit: b01db23d5923a35023540edc4f0c5f019e11ac7d ("sched/eevdf: Fix pick_eevdf()")
https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master

testcase: hackbench
test machine: 128 threads 2 sockets Intel(R) Xeon(R) Gold 6338 CPU @ 2.00GHz (Ice Lake) with 256G memory
parameters:

	nr_threads: 50%
	iterations: 4
	mode: process
	ipc: pipe
	cpufreq_governor: performance




If you fix the issue in a separate patch/commit (i.e. not just a new version of
the same patch/commit), kindly add following tags
| Reported-by: kernel test robot <oliver.sang@intel.com>
| Closes: https://lore.kernel.org/oe-lkp/202310182229.78d950b2-oliver.sang@intel.com


Details are as below:
-------------------------------------------------------------------------------------------------->


The kernel config and materials to reproduce are available at:
https://download.01.org/0day-ci/archive/20231018/202310182229.78d950b2-oliver.sang@intel.com

=========================================================================================
compiler/cpufreq_governor/ipc/iterations/kconfig/mode/nr_threads/rootfs/tbox_group/testcase:
  gcc-12/performance/pipe/4/x86_64-rhel-8.3/process/50%/debian-11.1-x86_64-20220510.cgz/lkp-icl-2sp2/hackbench

commit: 
  8dafa9d0eb ("sched/eevdf: Fix min_deadline heap integrity")
  b01db23d59 ("sched/eevdf: Fix pick_eevdf()")

8dafa9d0eb1a1550 b01db23d5923a35023540edc4f0 
---------------- --------------------------- 
         %stddev     %change         %stddev
             \          |                \  
 4.958e+08 ±  4%     +11.8%  5.544e+08 ±  9%  cpuidle..time
     47.00 ± 54%     -74.1%      12.18 ± 87%  turbostat.IPC
   2127266           +10.3%    2347320        sched_debug.cpu.nr_switches.avg
     36.94 ±  5%     +10.7%      40.89 ±  3%  sched_debug.cpu.nr_uninterruptible.stddev
   8760010            +9.9%    9630735        vmstat.system.cs
   1087178            +2.8%    1117923        vmstat.system.in
      8690 ±  4%    +108.2%      18091 ± 33%  meminfo.Active
      8034 ±  4%    +117.0%      17435 ± 35%  meminfo.Active(anon)
    695914 ±  4%     -10.0%     626568 ±  3%  meminfo.Mapped
    559.67 ± 71%    +741.2%       4708 ± 11%  perf-c2c.DRAM.remote
      3119 ± 76%    +972.2%      33448 ±  7%  perf-c2c.HITM.local
    236.00 ± 86%    +240.7%     804.17 ± 19%  perf-c2c.HITM.remote
      3355 ± 76%    +920.8%      34252 ±  7%  perf-c2c.HITM.total
   1052797            -3.4%    1016737        hackbench.throughput
   1011957            -3.2%     979124        hackbench.throughput_avg
   1052797            -3.4%    1016737        hackbench.throughput_best
    917410            -3.0%     890080        hackbench.throughput_worst
  99425414           +14.5%  1.138e+08        hackbench.time.involuntary_context_switches
      6523            +3.2%       6728        hackbench.time.system_time
    752.72            +2.6%     772.26        hackbench.time.user_time
 4.408e+08           +13.9%  5.021e+08        hackbench.time.voluntary_context_switches
    538.27 ± 56%    +180.9%       1511 ± 52%  numa-vmstat.node0.nr_active_anon
     44253 ± 39%    +609.8%     314098 ±115%  numa-vmstat.node0.nr_inactive_anon
     34617 ±  9%     +83.5%      63505 ± 54%  numa-vmstat.node0.nr_mapped
      1543 ± 22%  +16311.4%     253377 ±141%  numa-vmstat.node0.nr_shmem
    538.27 ± 56%    +180.9%       1511 ± 52%  numa-vmstat.node0.nr_zone_active_anon
     44252 ± 39%    +609.8%     314096 ±115%  numa-vmstat.node0.nr_zone_inactive_anon
    138022 ±  9%     -32.6%      92993 ± 41%  numa-vmstat.node1.nr_mapped
     15216 ± 45%     -52.7%       7199 ± 84%  numa-vmstat.node1.nr_page_table_pages
      2553 ± 49%    +149.5%       6369 ± 50%  numa-meminfo.node0.Active
      2225 ± 58%    +171.5%       6041 ± 53%  numa-meminfo.node0.Active(anon)
    177208 ± 38%    +610.0%    1258121 ±115%  numa-meminfo.node0.Inactive
    177010 ± 39%    +610.6%    1257927 ±115%  numa-meminfo.node0.Inactive(anon)
    138899 ±  9%     +83.1%     254371 ± 54%  numa-meminfo.node0.Mapped
      6256 ± 23%  +16125.5%    1015094 ±141%  numa-meminfo.node0.Shmem
    556497 ±  8%     -33.2%     371757 ± 40%  numa-meminfo.node1.Mapped
     60668 ± 45%     -52.7%      28716 ± 83%  numa-meminfo.node1.PageTables
    267338 ± 13%     -23.1%     205462 ± 21%  numa-meminfo.node1.Slab
      2008 ±  5%    +117.7%       4372 ± 35%  proc-vmstat.nr_active_anon
    173611 ±  5%      -9.6%     157026 ±  3%  proc-vmstat.nr_mapped
      2008 ±  5%    +117.7%       4372 ± 35%  proc-vmstat.nr_zone_active_anon
 1.243e+08            -1.8%  1.221e+08        proc-vmstat.numa_hit
 1.242e+08            -1.8%   1.22e+08        proc-vmstat.numa_local
    427.50 ± 26%    +428.8%       2260 ± 15%  proc-vmstat.pgactivate
 1.244e+08            -1.8%  1.222e+08        proc-vmstat.pgalloc_normal
 1.227e+08            -2.0%  1.202e+08        proc-vmstat.pgfree
    550400            +3.7%     571008        proc-vmstat.unevictable_pgs_scanned
      7.27 ± 68%      -4.8        2.42 ±161%  perf-profile.calltrace.cycles-pp.seq_read_iter.seq_read.vfs_read.ksys_read.do_syscall_64
      7.27 ± 68%      -4.8        2.42 ±161%  perf-profile.calltrace.cycles-pp.seq_read.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
      6.65 ± 77%      -4.2        2.42 ±161%  perf-profile.calltrace.cycles-pp.proc_single_show.seq_read_iter.seq_read.vfs_read.ksys_read
      6.65 ± 77%      -4.2        2.42 ±161%  perf-profile.calltrace.cycles-pp.proc_pid_status.proc_single_show.seq_read_iter.seq_read.vfs_read
      5.03 ±100%      -4.1        0.93 ±223%  perf-profile.calltrace.cycles-pp.number.vsnprintf.seq_printf.show_interrupts.seq_read_iter
      2.08 ±223%     +11.2       13.29 ± 53%  perf-profile.calltrace.cycles-pp.perf_session__deliver_event.__ordered_events__flush.perf_session__process_user_event.reader__read_event.perf_session__process_events
      2.78 ±223%     +12.6       15.36 ± 54%  perf-profile.calltrace.cycles-pp.perf_session__process_user_event.reader__read_event.perf_session__process_events.record__finish_output.__cmd_record
      2.08 ±223%     +13.3       15.36 ± 54%  perf-profile.calltrace.cycles-pp.__ordered_events__flush.perf_session__process_user_event.reader__read_event.perf_session__process_events.record__finish_output
      7.27 ± 68%      -4.8        2.42 ±161%  perf-profile.children.cycles-pp.seq_read
      6.01 ± 75%      -4.4        1.59 ±144%  perf-profile.children.cycles-pp.number
      6.65 ± 77%      -4.2        2.42 ±161%  perf-profile.children.cycles-pp.proc_single_show
      6.65 ± 77%      -4.2        2.42 ±161%  perf-profile.children.cycles-pp.proc_pid_status
      2.08 ±223%     +11.2       13.29 ± 53%  perf-profile.children.cycles-pp.perf_session__deliver_event
      2.78 ±223%     +12.6       15.36 ± 54%  perf-profile.children.cycles-pp.perf_session__process_user_event
      2.78 ±223%     +12.6       15.36 ± 54%  perf-profile.children.cycles-pp.__ordered_events__flush
      6.01 ± 75%      -4.4        1.59 ±144%  perf-profile.self.cycles-pp.number
      0.39            +0.0        0.42        perf-stat.i.branch-miss-rate%
 1.983e+08            +5.2%  2.085e+08        perf-stat.i.branch-misses
   9200253           +10.1%   10127041        perf-stat.i.context-switches
    554899            +5.3%     584042        perf-stat.i.cpu-migrations
      0.04            +0.0        0.05        perf-stat.i.dTLB-load-miss-rate%
  33430860           +10.3%   36871351        perf-stat.i.dTLB-load-misses
      0.85            -0.9%       0.84        perf-stat.i.ipc
      9.94 ± 16%     -26.2%       7.34 ± 24%  perf-stat.i.major-faults
   4603385 ±  3%      +6.8%    4916292 ±  3%  perf-stat.i.node-stores
      0.39            +0.0        0.41        perf-stat.overall.branch-miss-rate%
      0.04            +0.0        0.05        perf-stat.overall.dTLB-load-miss-rate%
 1.947e+08            +5.3%  2.051e+08        perf-stat.ps.branch-misses
   9028654           +10.2%    9953106        perf-stat.ps.context-switches
    543976            +5.4%     573599        perf-stat.ps.cpu-migrations
  32914153           +10.4%   36349710        perf-stat.ps.dTLB-load-misses
   4512147 ±  3%      +7.0%    4827523 ±  3%  perf-stat.ps.node-stores
 1.588e+13            +3.3%   1.64e+13        perf-stat.total.instructions




Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


-- 
0-DAY CI Kernel Test Service
https://github.com/intel/lkp-tests/wiki


                 reply	other threads:[~2023-10-18 14:55 UTC|newest]

Thread overview: [no followups] expand[flat|nested]  mbox.gz  Atom feed

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=202310182229.78d950b2-oliver.sang@intel.com \
    --to=oliver.sang@intel.com \
    --cc=aubrey.li@linux.intel.com \
    --cc=bsegall@google.com \
    --cc=feng.tang@intel.com \
    --cc=fengwei.yin@intel.com \
    --cc=linux-kernel@vger.kernel.org \
    --cc=lkp@intel.com \
    --cc=oe-lkp@lists.linux.dev \
    --cc=peterz@infradead.org \
    --cc=ying.huang@intel.com \
    --cc=yu.c.chen@intel.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox