public inbox for linux-kernel@vger.kernel.org
 help / color / mirror / Atom feed
* [LKP] [sched/core] 9edfbfed3f5: +88.2% hackbench.time.involuntary_context_switches
@ 2015-02-09  5:58 Huang Ying
  2015-02-09  8:31 ` Peter Zijlstra
  0 siblings, 1 reply; 4+ messages in thread
From: Huang Ying @ 2015-02-09  5:58 UTC (permalink / raw)
  To: Peter Zijlstra; +Cc: Ingo Molnar, LKML, LKP ML

[-- Attachment #1: Type: text/plain, Size: 4708 bytes --]

FYI, we noticed the below changes on

git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip.git sched/core
commit 9edfbfed3f544a7830d99b341f0c175995a02950 ("sched/core: Rework rq->clock update skips")


testbox/testcase/testparams: xps2/hackbench/performance-1600%-process-socket

cebde6d681aa45f9  9edfbfed3f544a7830d99b341f  
----------------  --------------------------  
         %stddev     %change         %stddev
             \          |                \  
   1839273 ±  6%     +88.2%    3462337 ±  4%  hackbench.time.involuntary_context_switches
  41965851 ±  5%      +5.6%   44307403 ±  1%  hackbench.time.voluntary_context_switches
       388 ± 39%     -58.6%        160 ± 10%  sched_debug.cfs_rq[1]:/.tg_load_contrib
     12957 ± 14%     -60.5%       5117 ± 11%  sched_debug.cfs_rq[2]:/.tg_load_avg
     30505 ± 14%     -57.7%      12905 ±  6%  sched_debug.cfs_rq[3]:/.tg_load_avg
      2790 ± 24%     -65.4%        964 ± 32%  sched_debug.cfs_rq[3]:/.blocked_load_avg
      2915 ± 23%     -62.2%       1101 ± 29%  sched_debug.cfs_rq[3]:/.tg_load_contrib
   1839273 ±  6%     +88.2%    3462337 ±  4%  time.involuntary_context_switches
      1474 ± 28%     -61.7%        565 ± 43%  sched_debug.cfs_rq[2]:/.tg_load_contrib
     11830 ± 15%     +63.0%      19285 ± 11%  sched_debug.cpu#4.sched_goidle
     19319 ± 29%     +91.1%      36913 ±  7%  sched_debug.cpu#3.sched_goidle
      5899 ± 31%     -35.6%       3801 ± 11%  sched_debug.cfs_rq[4]:/.blocked_load_avg
      5999 ± 30%     -34.5%       3929 ± 11%  sched_debug.cfs_rq[4]:/.tg_load_contrib
     37884 ± 13%     -33.5%      25207 ±  7%  sched_debug.cfs_rq[4]:/.tg_load_avg
    229547 ±  5%     +47.9%     339519 ±  5%  cpuidle.C1-NHM.usage
     35712 ±  3%     +31.7%      47036 ±  9%  cpuidle.C3-NHM.usage
      5010 ±  9%     -29.0%       3556 ± 20%  sched_debug.cfs_rq[6]:/.blocked_load_avg
      5139 ±  9%     -28.2%       3690 ± 19%  sched_debug.cfs_rq[6]:/.tg_load_contrib
     49568 ±  6%     +24.8%      61867 ±  7%  sched_debug.cpu#1.sched_goidle
     26369 ± 35%     -42.0%      15289 ± 29%  cpuidle.C6-NHM.usage
        18 ± 16%     +36.5%         25 ±  7%  sched_debug.cpu#4.nr_running
      1.41 ± 12%     -19.3%       1.14 ± 13%  perf-profile.cpu-cycles.sock_wfree.unix_destruct_scm.skb_release_head_state.skb_release_all.consume_skb
        25 ± 15%     +28.7%         32 ±  9%  sched_debug.cpu#3.nr_running
      1.63 ± 11%     -18.0%       1.34 ± 12%  perf-profile.cpu-cycles.unix_destruct_scm.skb_release_head_state.skb_release_all.consume_skb.unix_stream_recvmsg
      0.57 ±  8%      +9.6%       0.62 ±  5%  turbostat.CPU%c1
       148 ± 11%     -16.7%        123 ±  7%  sched_debug.cfs_rq[1]:/.load
       109 ±  6%     +17.1%        128 ±  6%  sched_debug.cpu#6.cpu_load[0]
      2.41 ±  8%     -13.3%       2.09 ± 11%  perf-profile.cpu-cycles.skb_release_head_state.skb_release_all.consume_skb.unix_stream_recvmsg.sock_aio_read
       147 ± 12%     -16.4%        123 ±  7%  sched_debug.cpu#1.load
       111 ±  5%     +15.4%        129 ±  5%  sched_debug.cpu#6.cpu_load[2]
       110 ±  5%     +14.9%        127 ±  5%  sched_debug.cfs_rq[6]:/.runnable_load_avg
       112 ±  5%     +14.5%        128 ±  4%  sched_debug.cpu#6.cpu_load[3]
       113 ±  5%     +13.2%        128 ±  3%  sched_debug.cpu#6.cpu_load[4]
    789953 ±  2%     -10.8%     704528 ±  4%  sched_debug.cpu#3.avg_idle
     15471 ±  5%      -7.7%      14278 ±  2%  sched_debug.cpu#5.curr->pid
   2675106 ± 10%     +16.2%    3109411 ±  1%  sched_debug.cpu#4.nr_switches
   2675140 ± 10%     +16.2%    3109440 ±  1%  sched_debug.cpu#4.sched_count
    155201 ±  5%     +14.6%     177901 ±  3%  softirqs.RCU
      8.64 ±  6%      -9.6%       7.82 ±  5%  perf-profile.cpu-cycles.skb_release_all.consume_skb.unix_stream_recvmsg.sock_aio_read.sock_aio_read
   2658351 ± 11%     +13.7%    3021564 ±  2%  sched_debug.cpu#5.sched_count
   2658326 ± 11%     +13.7%    3021539 ±  2%  sched_debug.cpu#5.nr_switches
     71443 ±  5%      +9.9%      78486 ±  0%  vmstat.system.cs
      8209 ±  5%      +7.3%       8805 ±  0%  vmstat.system.in

xps2: Nehalem
Memory: 4G

To reproduce:

        apt-get install ruby ruby-oj
        git clone git://git.kernel.org/pub/scm/linux/kernel/git/wfg/lkp-tests.git
        cd lkp-tests
        bin/setup-local job.yaml # the job file attached in this email
        bin/run-local   job.yaml


Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.


Thanks,
Huang, Ying


[-- Attachment #2: reproduce --]
[-- Type: text/plain, Size: 51 bytes --]

tbench_srv &
tbench 4 127.0.0.1
killall tbench_srv

[-- Attachment #3: Type: text/plain, Size: 89 bytes --]

_______________________________________________
LKP mailing list
LKP@linux.intel.com
\r

[-- Attachment #4: job.yaml --]
[-- Type: application/x-yaml, Size: 1685 bytes --]

^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: [LKP] [sched/core] 9edfbfed3f5: +88.2% hackbench.time.involuntary_context_switches
  2015-02-09  5:58 [LKP] [sched/core] 9edfbfed3f5: +88.2% hackbench.time.involuntary_context_switches Huang Ying
@ 2015-02-09  8:31 ` Peter Zijlstra
       [not found]   ` <CAC=cRTN3VKmgM4BYUrN=rMfg8stiP-F_oWR+2LSPFR9m1Wmfqg@mail.gmail.com>
  2015-02-10  7:26   ` Markus Pargmann
  0 siblings, 2 replies; 4+ messages in thread
From: Peter Zijlstra @ 2015-02-09  8:31 UTC (permalink / raw)
  To: Huang Ying; +Cc: Ingo Molnar, LKML, LKP ML

On Mon, Feb 09, 2015 at 01:58:33PM +0800, Huang Ying wrote:
> FYI, we noticed the below changes on
> 
> git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip.git sched/core
> commit 9edfbfed3f544a7830d99b341f0c175995a02950 ("sched/core: Rework rq->clock update skips")
> 
> 
> testbox/testcase/testparams: xps2/hackbench/performance-1600%-process-socket
> 
> cebde6d681aa45f9  9edfbfed3f544a7830d99b341f  
> ----------------  --------------------------  
>          %stddev     %change         %stddev
>              \          |                \  
>    1839273 ±  6%     +88.2%    3462337 ±  4%  hackbench.time.involuntary_context_switches
>   41965851 ±  5%      +5.6%   44307403 ±  1%  hackbench.time.voluntary_context_switches
>        388 ± 39%     -58.6%        160 ± 10%  sched_debug.cfs_rq[1]:/.tg_load_contrib
>      12957 ± 14%     -60.5%       5117 ± 11%  sched_debug.cfs_rq[2]:/.tg_load_avg
>      30505 ± 14%     -57.7%      12905 ±  6%  sched_debug.cfs_rq[3]:/.tg_load_avg
>       2790 ± 24%     -65.4%        964 ± 32%  sched_debug.cfs_rq[3]:/.blocked_load_avg
>       2915 ± 23%     -62.2%       1101 ± 29%  sched_debug.cfs_rq[3]:/.tg_load_contrib
>    1839273 ±  6%     +88.2%    3462337 ±  4%  time.involuntary_context_switches
>       1474 ± 28%     -61.7%        565 ± 43%  sched_debug.cfs_rq[2]:/.tg_load_contrib
>      11830 ± 15%     +63.0%      19285 ± 11%  sched_debug.cpu#4.sched_goidle
>      19319 ± 29%     +91.1%      36913 ±  7%  sched_debug.cpu#3.sched_goidle
>       5899 ± 31%     -35.6%       3801 ± 11%  sched_debug.cfs_rq[4]:/.blocked_load_avg
>       5999 ± 30%     -34.5%       3929 ± 11%  sched_debug.cfs_rq[4]:/.tg_load_contrib
>      37884 ± 13%     -33.5%      25207 ±  7%  sched_debug.cfs_rq[4]:/.tg_load_avg
>     229547 ±  5%     +47.9%     339519 ±  5%  cpuidle.C1-NHM.usage
>      35712 ±  3%     +31.7%      47036 ±  9%  cpuidle.C3-NHM.usage
>       5010 ±  9%     -29.0%       3556 ± 20%  sched_debug.cfs_rq[6]:/.blocked_load_avg
>       5139 ±  9%     -28.2%       3690 ± 19%  sched_debug.cfs_rq[6]:/.tg_load_contrib
>      49568 ±  6%     +24.8%      61867 ±  7%  sched_debug.cpu#1.sched_goidle
>      26369 ± 35%     -42.0%      15289 ± 29%  cpuidle.C6-NHM.usage
>         18 ± 16%     +36.5%         25 ±  7%  sched_debug.cpu#4.nr_running
>       1.41 ± 12%     -19.3%       1.14 ± 13%  perf-profile.cpu-cycles.sock_wfree.unix_destruct_scm.skb_release_head_state.skb_release_all.consume_skb
>         25 ± 15%     +28.7%         32 ±  9%  sched_debug.cpu#3.nr_running
>       1.63 ± 11%     -18.0%       1.34 ± 12%  perf-profile.cpu-cycles.unix_destruct_scm.skb_release_head_state.skb_release_all.consume_skb.unix_stream_recvmsg
>       0.57 ±  8%      +9.6%       0.62 ±  5%  turbostat.CPU%c1
>        148 ± 11%     -16.7%        123 ±  7%  sched_debug.cfs_rq[1]:/.load
>        109 ±  6%     +17.1%        128 ±  6%  sched_debug.cpu#6.cpu_load[0]
>       2.41 ±  8%     -13.3%       2.09 ± 11%  perf-profile.cpu-cycles.skb_release_head_state.skb_release_all.consume_skb.unix_stream_recvmsg.sock_aio_read
>        147 ± 12%     -16.4%        123 ±  7%  sched_debug.cpu#1.load
>        111 ±  5%     +15.4%        129 ±  5%  sched_debug.cpu#6.cpu_load[2]
>        110 ±  5%     +14.9%        127 ±  5%  sched_debug.cfs_rq[6]:/.runnable_load_avg
>        112 ±  5%     +14.5%        128 ±  4%  sched_debug.cpu#6.cpu_load[3]
>        113 ±  5%     +13.2%        128 ±  3%  sched_debug.cpu#6.cpu_load[4]
>     789953 ±  2%     -10.8%     704528 ±  4%  sched_debug.cpu#3.avg_idle
>      15471 ±  5%      -7.7%      14278 ±  2%  sched_debug.cpu#5.curr->pid
>    2675106 ± 10%     +16.2%    3109411 ±  1%  sched_debug.cpu#4.nr_switches
>    2675140 ± 10%     +16.2%    3109440 ±  1%  sched_debug.cpu#4.sched_count
>     155201 ±  5%     +14.6%     177901 ±  3%  softirqs.RCU
>       8.64 ±  6%      -9.6%       7.82 ±  5%  perf-profile.cpu-cycles.skb_release_all.consume_skb.unix_stream_recvmsg.sock_aio_read.sock_aio_read
>    2658351 ± 11%     +13.7%    3021564 ±  2%  sched_debug.cpu#5.sched_count
>    2658326 ± 11%     +13.7%    3021539 ±  2%  sched_debug.cpu#5.nr_switches
>      71443 ±  5%      +9.9%      78486 ±  0%  vmstat.system.cs
>       8209 ±  5%      +7.3%       8805 ±  0%  vmstat.system.in
> 

OK, so the interesting number is total runtime; I cannot find it.
Therefore I cannot say what if anything changed. This is just a bunch of
random numbers afaict.

> To reproduce:
> 
>         apt-get install ruby ruby-oj

Mostly likely, you requiring ruby just means I will not be reproducing.


^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: [LKP] [sched/core] 9edfbfed3f5: +88.2% hackbench.time.involuntary_context_switches
       [not found]   ` <CAC=cRTN3VKmgM4BYUrN=rMfg8stiP-F_oWR+2LSPFR9m1Wmfqg@mail.gmail.com>
@ 2015-02-09  9:27     ` Peter Zijlstra
  0 siblings, 0 replies; 4+ messages in thread
From: Peter Zijlstra @ 2015-02-09  9:27 UTC (permalink / raw)
  To: huang ying; +Cc: Huang Ying, Ingo Molnar, LKML, LKP ML, Wu, Fengguang

On Mon, Feb 09, 2015 at 04:47:07PM +0800, huang ying wrote:
> There are no distinguishable difference between the parent and the child
> for hackbench throughput number.
> 
> Usually you will not consider statistics such as involuntary context
> switches?

Only if there's a 'problem' with the primary performance metric (total
runtime in case of hackbench).

Once the primary metric shifts, you go look at what the cause of this
change might be, at that point things like # context switches etc.. are
interesting. As long as the primary performance metric is stable, meh.

As such; I would suggest _always_ reporting the primary metric for each
benchmark, preferably on top and not hidden somewhere in the mass of
numbers.

I now had to very carefully waste a few minutes of my time reading those
numbers to see if there was anything useful in; there wasn't.

^ permalink raw reply	[flat|nested] 4+ messages in thread

* Re: [LKP] [sched/core] 9edfbfed3f5: +88.2% hackbench.time.involuntary_context_switches
  2015-02-09  8:31 ` Peter Zijlstra
       [not found]   ` <CAC=cRTN3VKmgM4BYUrN=rMfg8stiP-F_oWR+2LSPFR9m1Wmfqg@mail.gmail.com>
@ 2015-02-10  7:26   ` Markus Pargmann
  1 sibling, 0 replies; 4+ messages in thread
From: Markus Pargmann @ 2015-02-10  7:26 UTC (permalink / raw)
  To: Peter Zijlstra; +Cc: Huang Ying, Ingo Molnar, LKML, LKP ML

[-- Attachment #1: Type: text/plain, Size: 5296 bytes --]

Hi,

On Mon, Feb 09, 2015 at 09:31:20AM +0100, Peter Zijlstra wrote:
> On Mon, Feb 09, 2015 at 01:58:33PM +0800, Huang Ying wrote:
> > FYI, we noticed the below changes on
> > 
> > git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip.git sched/core
> > commit 9edfbfed3f544a7830d99b341f0c175995a02950 ("sched/core: Rework rq->clock update skips")
> > 
> > 
> > testbox/testcase/testparams: xps2/hackbench/performance-1600%-process-socket
> > 
> > cebde6d681aa45f9  9edfbfed3f544a7830d99b341f  
> > ----------------  --------------------------  
> >          %stddev     %change         %stddev
> >              \          |                \  
> >    1839273 ±  6%     +88.2%    3462337 ±  4%  hackbench.time.involuntary_context_switches
> >   41965851 ±  5%      +5.6%   44307403 ±  1%  hackbench.time.voluntary_context_switches
> >        388 ± 39%     -58.6%        160 ± 10%  sched_debug.cfs_rq[1]:/.tg_load_contrib
> >      12957 ± 14%     -60.5%       5117 ± 11%  sched_debug.cfs_rq[2]:/.tg_load_avg
> >      30505 ± 14%     -57.7%      12905 ±  6%  sched_debug.cfs_rq[3]:/.tg_load_avg
> >       2790 ± 24%     -65.4%        964 ± 32%  sched_debug.cfs_rq[3]:/.blocked_load_avg
> >       2915 ± 23%     -62.2%       1101 ± 29%  sched_debug.cfs_rq[3]:/.tg_load_contrib
> >    1839273 ±  6%     +88.2%    3462337 ±  4%  time.involuntary_context_switches
> >       1474 ± 28%     -61.7%        565 ± 43%  sched_debug.cfs_rq[2]:/.tg_load_contrib
> >      11830 ± 15%     +63.0%      19285 ± 11%  sched_debug.cpu#4.sched_goidle
> >      19319 ± 29%     +91.1%      36913 ±  7%  sched_debug.cpu#3.sched_goidle
> >       5899 ± 31%     -35.6%       3801 ± 11%  sched_debug.cfs_rq[4]:/.blocked_load_avg
> >       5999 ± 30%     -34.5%       3929 ± 11%  sched_debug.cfs_rq[4]:/.tg_load_contrib
> >      37884 ± 13%     -33.5%      25207 ±  7%  sched_debug.cfs_rq[4]:/.tg_load_avg
> >     229547 ±  5%     +47.9%     339519 ±  5%  cpuidle.C1-NHM.usage
> >      35712 ±  3%     +31.7%      47036 ±  9%  cpuidle.C3-NHM.usage
> >       5010 ±  9%     -29.0%       3556 ± 20%  sched_debug.cfs_rq[6]:/.blocked_load_avg
> >       5139 ±  9%     -28.2%       3690 ± 19%  sched_debug.cfs_rq[6]:/.tg_load_contrib
> >      49568 ±  6%     +24.8%      61867 ±  7%  sched_debug.cpu#1.sched_goidle
> >      26369 ± 35%     -42.0%      15289 ± 29%  cpuidle.C6-NHM.usage
> >         18 ± 16%     +36.5%         25 ±  7%  sched_debug.cpu#4.nr_running
> >       1.41 ± 12%     -19.3%       1.14 ± 13%  perf-profile.cpu-cycles.sock_wfree.unix_destruct_scm.skb_release_head_state.skb_release_all.consume_skb
> >         25 ± 15%     +28.7%         32 ±  9%  sched_debug.cpu#3.nr_running
> >       1.63 ± 11%     -18.0%       1.34 ± 12%  perf-profile.cpu-cycles.unix_destruct_scm.skb_release_head_state.skb_release_all.consume_skb.unix_stream_recvmsg
> >       0.57 ±  8%      +9.6%       0.62 ±  5%  turbostat.CPU%c1
> >        148 ± 11%     -16.7%        123 ±  7%  sched_debug.cfs_rq[1]:/.load
> >        109 ±  6%     +17.1%        128 ±  6%  sched_debug.cpu#6.cpu_load[0]
> >       2.41 ±  8%     -13.3%       2.09 ± 11%  perf-profile.cpu-cycles.skb_release_head_state.skb_release_all.consume_skb.unix_stream_recvmsg.sock_aio_read
> >        147 ± 12%     -16.4%        123 ±  7%  sched_debug.cpu#1.load
> >        111 ±  5%     +15.4%        129 ±  5%  sched_debug.cpu#6.cpu_load[2]
> >        110 ±  5%     +14.9%        127 ±  5%  sched_debug.cfs_rq[6]:/.runnable_load_avg
> >        112 ±  5%     +14.5%        128 ±  4%  sched_debug.cpu#6.cpu_load[3]
> >        113 ±  5%     +13.2%        128 ±  3%  sched_debug.cpu#6.cpu_load[4]
> >     789953 ±  2%     -10.8%     704528 ±  4%  sched_debug.cpu#3.avg_idle
> >      15471 ±  5%      -7.7%      14278 ±  2%  sched_debug.cpu#5.curr->pid
> >    2675106 ± 10%     +16.2%    3109411 ±  1%  sched_debug.cpu#4.nr_switches
> >    2675140 ± 10%     +16.2%    3109440 ±  1%  sched_debug.cpu#4.sched_count
> >     155201 ±  5%     +14.6%     177901 ±  3%  softirqs.RCU
> >       8.64 ±  6%      -9.6%       7.82 ±  5%  perf-profile.cpu-cycles.skb_release_all.consume_skb.unix_stream_recvmsg.sock_aio_read.sock_aio_read
> >    2658351 ± 11%     +13.7%    3021564 ±  2%  sched_debug.cpu#5.sched_count
> >    2658326 ± 11%     +13.7%    3021539 ±  2%  sched_debug.cpu#5.nr_switches
> >      71443 ±  5%      +9.9%      78486 ±  0%  vmstat.system.cs
> >       8209 ±  5%      +7.3%       8805 ±  0%  vmstat.system.in
> > 
> 
> OK, so the interesting number is total runtime; I cannot find it.
> Therefore I cannot say what if anything changed. This is just a bunch of
> random numbers afaict.

The total runtime of hackbench in v3.19 compared to v3.15 to v3.18 is
shown here [1] (Group 5 -> linux_perf.hackbench). It is not the same
project so the profiling data is not related and not recorded. The
results should be reproducable by simply running hackbench with the same
options as cbenchsuite [2] did. The options are always listed in the
result browser [1] below the plots.

Best Regards,

Markus

[1] http://results.cbenchsuite.org/plots/2015-02-09__v3.15-v3.19-quad/detailed/
[2] http://cbenchsuite.org


[-- Attachment #2: Digital signature --]
[-- Type: application/pgp-signature, Size: 819 bytes --]

^ permalink raw reply	[flat|nested] 4+ messages in thread

end of thread, other threads:[~2015-02-10  7:26 UTC | newest]

Thread overview: 4+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2015-02-09  5:58 [LKP] [sched/core] 9edfbfed3f5: +88.2% hackbench.time.involuntary_context_switches Huang Ying
2015-02-09  8:31 ` Peter Zijlstra
     [not found]   ` <CAC=cRTN3VKmgM4BYUrN=rMfg8stiP-F_oWR+2LSPFR9m1Wmfqg@mail.gmail.com>
2015-02-09  9:27     ` Peter Zijlstra
2015-02-10  7:26   ` Markus Pargmann

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox