public inbox for linux-kernel@vger.kernel.org
 help / color / mirror / Atom feed
From: Christoph Hellwig <hch@lst.de>
To: Ye Xiaolong <xiaolong.ye@intel.com>
Cc: Christoph Hellwig <hch@lst.de>,
	Greg Kroah-Hartman <gregkh@linuxfoundation.org>,
	"Darrick J. Wong" <darrick.wong@oracle.com>,
	LKML <linux-kernel@vger.kernel.org>,
	Linus Torvalds <torvalds@linux-foundation.org>,
	lkp@01.org, viro@zeniv.linux.org.uk
Subject: Re: [lkp-robot] [fs]  3deb642f0d:  will-it-scale.per_process_ops -8.8% regression
Date: Thu, 28 Jun 2018 15:38:33 +0200	[thread overview]
Message-ID: <20180628133833.GA11790@lst.de> (raw)
In-Reply-To: <20180628003834.GH18756@yexl-desktop>

On Thu, Jun 28, 2018 at 08:38:34AM +0800, Ye Xiaolong wrote:
> Update the result:
> 
> testcase/path_params/tbox_group/run: will-it-scale/poll2-performance/lkp-sb03

So this looks like a huge improvement in the per process ops, but not
as large as the original regression, and no change in the per-thread
ops.

But the baseline already looks much lower, e.g. this shows an improvement
from 404611 to 424608 for the per-process ops, while the original showed
a regression from 501456 to 457120.  Are we measuring on different
hardware?   Did we gain new spectre mitigations elsewhere?  Either way
I'm going to send these patches out for review, but I'd like to understand
the numbers a bit more.

> 
> 894b8c000ae6c106  8fbedc19c94fd25a2b9b327015  
> ----------------  --------------------------  
>          %stddev      change         %stddev
>              \          |                \  
>     404611 ±  4%         5%     424608        will-it-scale.per_process_ops
>       1489 ± 21%        28%       1899 ± 18%  will-it-scale.time.voluntary_context_switches
>   45828560                    46155690        will-it-scale.workload
>       2337                        2342        will-it-scale.time.system_time
>        806                         806        will-it-scale.time.percent_of_cpu_this_job_got
>        310                         310        will-it-scale.time.elapsed_time
>        310                         310        will-it-scale.time.elapsed_time.max
>       4096                        4096        will-it-scale.time.page_size
>     233917                      233862        will-it-scale.per_thread_ops
>      17196                       17179        will-it-scale.time.minor_page_faults
>       9901                        9862        will-it-scale.time.maximum_resident_set_size
>      14705 ±  3%                 14397 ±  4%  will-it-scale.time.involuntary_context_switches
>        167                         163        will-it-scale.time.user_time
>       0.66 ± 25%       -17%       0.54        will-it-scale.scalability
>     120508 ± 15%        -7%     112098 ±  5%  interrupts.CAL:Function_call_interrupts
>       1670 ±  3%        10%       1845 ±  3%  vmstat.system.cs
>      32707                       32635        vmstat.system.in
>        121                         122        turbostat.CorWatt
>        149                         150        turbostat.PkgWatt
>       1573                        1573        turbostat.Avg_MHz
>      17.54 ± 19%                 17.77 ± 19%  boot-time.kernel_boot
>        824 ± 12%                   834 ± 12%  boot-time.idle
>      27.45 ± 12%                 27.69 ± 12%  boot-time.boot
>      16.96 ± 21%                 16.93 ± 21%  boot-time.dhcp
>       1489 ± 21%        28%       1899 ± 18%  time.voluntary_context_switches
>       2337                        2342        time.system_time
>        806                         806        time.percent_of_cpu_this_job_got
>        310                         310        time.elapsed_time
>        310                         310        time.elapsed_time.max
>       4096                        4096        time.page_size
>      17196                       17179        time.minor_page_faults
>       9901                        9862        time.maximum_resident_set_size
>      14705 ±  3%                 14397 ±  4%  time.involuntary_context_switches
>        167                         163        time.user_time
>      18320               6%      19506 ±  8%  proc-vmstat.nr_slab_unreclaimable
>       1518 ±  7%                  1558 ± 10%  proc-vmstat.numa_hint_faults
>       1387 ±  8%                  1421 ±  9%  proc-vmstat.numa_hint_faults_local
>       1873 ±  5%                  1917 ±  8%  proc-vmstat.numa_pte_updates
>      19987                       20005        proc-vmstat.nr_anon_pages
>       8464                        8471        proc-vmstat.nr_kernel_stack
>     309815                      310062        proc-vmstat.nr_file_pages
>      50828                       50828        proc-vmstat.nr_free_cma
>   16065590                    16064831        proc-vmstat.nr_free_pages
>    3194669                     3194517        proc-vmstat.nr_dirty_threshold
>    1595384                     1595308        proc-vmstat.nr_dirty_background_threshold
>     798886                      797937        proc-vmstat.pgfault
>       6510                        6499        proc-vmstat.nr_mapped
>     659089                      657491        proc-vmstat.numa_local
>     665458                      663786        proc-vmstat.numa_hit
>       1037                        1033        proc-vmstat.nr_page_table_pages
>     669923                      665906        proc-vmstat.pgfree
>     676982                      672385        proc-vmstat.pgalloc_normal
>       6368                        6294        proc-vmstat.numa_other
>      13013              -7%      12152 ± 11%  proc-vmstat.nr_slab_reclaimable
>   51213164 ± 18%        23%   63014695 ± 25%  perf-stat.node-loads
>   22096136 ± 28%        20%   26619357 ± 35%  perf-stat.node-load-misses
>  2.079e+08 ±  9%        12%  2.323e+08 ± 11%  perf-stat.cache-misses
>     515039 ±  3%        10%     568299 ±  3%  perf-stat.context-switches
>  3.283e+08 ± 22%        10%  3.622e+08 ±  5%  perf-stat.iTLB-loads
> 
> Thanks,
> Xiaolong
---end quoted text---

      reply	other threads:[~2018-06-28 13:38 UTC|newest]

Thread overview: 24+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2018-06-22  8:27 [lkp-robot] [fs] 3deb642f0d: will-it-scale.per_process_ops -8.8% regression kernel test robot
2018-06-22  9:25 ` Linus Torvalds
2018-06-22  9:56   ` Christoph Hellwig
2018-06-22 10:00     ` Christoph Hellwig
2018-06-22 11:01       ` Al Viro
2018-06-22 11:53         ` Christoph Hellwig
2018-06-22 11:56           ` Al Viro
2018-06-22 12:07             ` Christoph Hellwig
2018-06-22 12:17               ` Al Viro
2018-06-22 12:33                 ` Christoph Hellwig
2018-06-22 12:29                   ` Al Viro
2018-06-22 19:06         ` Sean Paul
2018-06-22 10:02     ` Linus Torvalds
2018-06-22 10:05       ` Linus Torvalds
2018-06-22 15:02 ` Christoph Hellwig
2018-06-22 15:14   ` Al Viro
2018-06-22 15:28     ` Christoph Hellwig
2018-06-22 16:18       ` Christoph Hellwig
2018-06-22 20:02         ` Al Viro
2018-06-23  7:15           ` Christoph Hellwig
2018-06-26  6:03   ` Ye Xiaolong
2018-06-27  7:07     ` Christoph Hellwig
2018-06-28  0:38       ` Ye Xiaolong
2018-06-28 13:38         ` Christoph Hellwig [this message]

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20180628133833.GA11790@lst.de \
    --to=hch@lst.de \
    --cc=darrick.wong@oracle.com \
    --cc=gregkh@linuxfoundation.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=lkp@01.org \
    --cc=torvalds@linux-foundation.org \
    --cc=viro@zeniv.linux.org.uk \
    --cc=xiaolong.ye@intel.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox