public inbox for linux-kernel@vger.kernel.org
 help / color / mirror / Atom feed
From: Jet Chen <jet.chen@intel.com>
To: "Paul E. McKenney" <paulmck@linux.vnet.ibm.com>
Cc: LKML <linux-kernel@vger.kernel.org>,
	lkp@01.org, Fengguang Wu <fengguang.wu@intel.com>
Subject: [rcu] 5057f55e543: -23.5% qperf.udp.recv_bw
Date: Tue, 03 Jun 2014 18:08:41 +0800	[thread overview]
Message-ID: <538D9EA9.6070406@intel.com> (raw)

[-- Attachment #1: Type: text/plain, Size: 6346 bytes --]

Hi Paul,

FYI, we noticed the below changes on

git://git.kernel.org/pub/scm/linux/kernel/git/paulmck/linux-rcu.git rcu/fixes
commit 5057f55e543b7859cfd26bc281291795eac93f8a ("rcu: Bind RCU grace-period kthreads if NO_HZ_FULL")

71a9b26963f8c2d  5057f55e543b7859cfd26bc28
---------------  -------------------------
 2.127e+09 ~ 0%     -23.5%  1.628e+09 ~ 4%  bens/qperf/600s
 2.127e+09 ~ 0%     -23.5%  1.628e+09 ~ 4%  TOTAL qperf.udp.recv_bw

71a9b26963f8c2d  5057f55e543b7859cfd26bc28
---------------  -------------------------
 2.128e+09 ~ 0%     -23.3%  1.633e+09 ~ 4%  bens/qperf/600s
 2.128e+09 ~ 0%     -23.3%  1.633e+09 ~ 4%  TOTAL qperf.udp.send_bw

71a9b26963f8c2d  5057f55e543b7859cfd26bc28
---------------  -------------------------
 2.101e+10 ~ 2%     -18.7%  1.707e+10 ~ 2%  bens/iperf/300s-tcp
 2.101e+10 ~ 2%     -18.7%  1.707e+10 ~ 2%  TOTAL iperf.tcp.sender.bps

71a9b26963f8c2d  5057f55e543b7859cfd26bc28
---------------  -------------------------
 2.101e+10 ~ 2%     -18.7%  1.707e+10 ~ 2%  bens/iperf/300s-tcp
 2.101e+10 ~ 2%     -18.7%  1.707e+10 ~ 2%  TOTAL iperf.tcp.receiver.bps

71a9b26963f8c2d  5057f55e543b7859cfd26bc28
---------------  -------------------------
 1.331e+09 ~ 2%      -5.8%  1.255e+09 ~ 2%  bens/qperf/600s
   2.4e+09 ~ 6%     -30.4%  1.671e+09 ~12%  brickland3/qperf/600s
 2.384e+09 ~ 7%     -12.1%  2.096e+09 ~ 3%  lkp-sb03/qperf/600s
 6.115e+09 ~ 5%     -17.9%  5.022e+09 ~ 6%  TOTAL qperf.sctp.bw

71a9b26963f8c2d  5057f55e543b7859cfd26bc28
---------------  -------------------------
  2.83e+09 ~ 1%     -12.5%  2.476e+09 ~ 3%  bens/qperf/600s
  2.83e+09 ~ 1%     -12.5%  2.476e+09 ~ 3%  TOTAL qperf.tcp.bw

71a9b26963f8c2d  5057f55e543b7859cfd26bc28
---------------  -------------------------
 2.272e+08 ~ 1%     -13.3%   1.97e+08 ~ 2%  bens/qperf/600s
 2.272e+08 ~ 1%     -13.3%   1.97e+08 ~ 2%  TOTAL proc-vmstat.pgalloc_dma32

71a9b26963f8c2d  5057f55e543b7859cfd26bc28
---------------  -------------------------
     53062 ~ 2%     -35.1%      34464 ~ 3%  bens/qperf/600s
    109531 ~13%     +46.9%     160928 ~ 5%  brickland3/qperf/600s
     67902 ~ 1%     +13.8%      77302 ~ 3%  lkp-sb03/qperf/600s
    230496 ~ 7%     +18.3%     272694 ~ 4%  TOTAL softirqs.RCU

71a9b26963f8c2d  5057f55e543b7859cfd26bc28
---------------  -------------------------
     80344 ~ 1%     -26.2%      59325 ~ 2%  bens/qperf/600s
     80344 ~ 1%     -26.2%      59325 ~ 2%  TOTAL softirqs.SCHED

71a9b26963f8c2d  5057f55e543b7859cfd26bc28
---------------  -------------------------
      1036 ~ 4%     -17.6%        853 ~ 4%  brickland3/qperf/600s
      1036 ~ 4%     -17.6%        853 ~ 4%  TOTAL proc-vmstat.nr_page_table_pages

71a9b26963f8c2d  5057f55e543b7859cfd26bc28
---------------  -------------------------
     48.12 ~ 0%     -11.7%      42.46 ~ 6%  brickland3/qperf/600s
     48.12 ~ 0%     -11.7%      42.46 ~ 6%  TOTAL turbostat.%pc2

71a9b26963f8c2d  5057f55e543b7859cfd26bc28
---------------  -------------------------
  74689352 ~ 1%     -13.3%   64771743 ~ 2%  bens/qperf/600s
  74689352 ~ 1%     -13.3%   64771743 ~ 2%  TOTAL proc-vmstat.pgalloc_normal

71a9b26963f8c2d  5057f55e543b7859cfd26bc28
---------------  -------------------------
 3.019e+08 ~ 1%     -13.3%  2.618e+08 ~ 2%  bens/qperf/600s
 3.019e+08 ~ 1%     -13.3%  2.618e+08 ~ 2%  TOTAL proc-vmstat.pgfree

71a9b26963f8c2d  5057f55e543b7859cfd26bc28
---------------  -------------------------
  23538414 ~ 0%     -12.9%   20506157 ~ 2%  bens/qperf/600s
  23538414 ~ 0%     -12.9%   20506157 ~ 2%  TOTAL proc-vmstat.numa_local

71a9b26963f8c2d  5057f55e543b7859cfd26bc28
---------------  -------------------------
  23538414 ~ 0%     -12.9%   20506157 ~ 2%  bens/qperf/600s
  23538414 ~ 0%     -12.9%   20506157 ~ 2%  TOTAL proc-vmstat.numa_hit

71a9b26963f8c2d  5057f55e543b7859cfd26bc28
---------------  -------------------------
     12789 ~ 1%     -10.9%      11391 ~ 2%  bens/qperf/600s
     12789 ~ 1%     -10.9%      11391 ~ 2%  TOTAL softirqs.HRTIMER

71a9b26963f8c2d  5057f55e543b7859cfd26bc28
---------------  -------------------------
    481253 ~ 0%      -8.9%     438624 ~ 0%  bens/qperf/600s
    481253 ~ 0%      -8.9%     438624 ~ 0%  TOTAL softirqs.TIMER

71a9b26963f8c2d  5057f55e543b7859cfd26bc28
---------------  -------------------------
      1297 ~33%    +565.9%       8640 ~ 7%  bens/iperf/300s-tcp
      2788 ~ 3%    +588.8%      19204 ~ 4%  bens/qperf/600s
      1191 ~ 5%   +1200.9%      15493 ~ 4%  brickland3/qperf/600s
      1135 ~26%   +1195.9%      14709 ~ 4%  lkp-sb03/qperf/600s
      6411 ~13%    +805.3%      58047 ~ 4%  TOTAL time.involuntary_context_switches

71a9b26963f8c2d  5057f55e543b7859cfd26bc28
---------------  -------------------------
     72398 ~ 1%      -5.4%      68503 ~ 0%  bens/qperf/600s
      8789 ~ 4%     +22.3%      10749 ~15%  lkp-sb03/qperf/600s
     81187 ~ 1%      -2.4%      79253 ~ 2%  TOTAL vmstat.system.in

71a9b26963f8c2d  5057f55e543b7859cfd26bc28
---------------  -------------------------
    141174 ~ 1%      -5.4%     133551 ~ 0%  bens/qperf/600s
    143982 ~ 1%      -4.4%     137600 ~ 0%  brickland3/qperf/600s
    285156 ~ 1%      -4.9%     271152 ~ 0%  TOTAL vmstat.system.cs

71a9b26963f8c2d  5057f55e543b7859cfd26bc28
---------------  -------------------------
  42351859 ~ 0%      -5.3%   40114932 ~ 0%  bens/qperf/600s
  43015383 ~ 1%      -4.4%   41143092 ~ 0%  brickland3/qperf/600s
  85367242 ~ 1%      -4.8%   81258025 ~ 0%  TOTAL time.voluntary_context_switches

71a9b26963f8c2d  5057f55e543b7859cfd26bc28
---------------  -------------------------
       146 ~ 0%      -2.2%        143 ~ 0%  bens/qperf/600s
       147 ~ 1%      -4.8%        140 ~ 1%  brickland3/qperf/600s
       293 ~ 0%      -3.5%        283 ~ 0%  TOTAL time.percent_of_cpu_this_job_got

71a9b26963f8c2d  5057f55e543b7859cfd26bc28
---------------  -------------------------
       872 ~ 0%      -2.3%        853 ~ 0%  bens/qperf/600s
       874 ~ 1%      -4.6%        834 ~ 1%  brickland3/qperf/600s
      1747 ~ 0%      -3.4%       1687 ~ 0%  TOTAL time.system_time


Legend:
	~XX%    - stddev percent
	[+-]XX% - change percent




Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.

Thanks,
Jet



[-- Attachment #2: reproduce --]
[-- Type: text/plain, Size: 324 bytes --]

echo performance > /sys/devices/system/cpu/cpu0/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu1/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu2/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu3/cpufreq/scaling_governor
./runtest.py unlink1 32 1 2 3 4

             reply	other threads:[~2014-06-03 10:09 UTC|newest]

Thread overview: 4+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2014-06-03 10:08 Jet Chen [this message]
2014-06-03 14:17 ` [rcu] 5057f55e543: -23.5% qperf.udp.recv_bw Paul E. McKenney
2014-06-04 12:33   ` Fengguang Wu
2014-06-04 22:17     ` Paul E. McKenney

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=538D9EA9.6070406@intel.com \
    --to=jet.chen@intel.com \
    --cc=fengguang.wu@intel.com \
    --cc=linux-kernel@vger.kernel.org \
    --cc=lkp@01.org \
    --cc=paulmck@linux.vnet.ibm.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox