From: Jason Wang <jasowang@redhat.com>
To: "Lukáš Doktor" <ldoktor@redhat.com>,
"QEMU Developers" <qemu-devel@nongnu.org>
Cc: Charles Shih <cheshi@redhat.com>,
Aleksandar Markovic <aleksandar.qemu.devel@gmail.com>,
Stefan Hajnoczi <stefanha@redhat.com>
Subject: Re: Proposal for a regular upstream performance testing
Date: Thu, 26 Nov 2020 16:23:47 +0800 [thread overview]
Message-ID: <32b0753d-d5bd-4790-a88b-998b152534bd@redhat.com> (raw)
In-Reply-To: <3a664806-8aa3-feb4-fb30-303d303217a8@redhat.com>
On 2020/11/26 下午4:10, Lukáš Doktor wrote:
> Hello guys,
>
> I had been around qemu on the Avocado-vt side for quite some time and
> a while ago I shifted my focus on performance testing. Currently I am
> not aware of any upstream CI that would continuously monitor the
> upstream qemu performance and I'd like to change that. There is a lot
> to cover so please bear with me.
>
> Goal
> ====
>
> The goal of this initiative is to detect system-wide performance
> regressions as well as improvements early, ideally pin-point the
> individual commits and notify people that they should fix things. All
> in upstream and ideally with least human interaction possible.
>
> Unlike the recent work of Ahmed Karaman's
> https://ahmedkrmn.github.io/TCG-Continuous-Benchmarking/ my aim is on
> the system-wide performance inside the guest (like fio, uperf, ...)
>
> Tools
> =====
>
> In house we have several different tools used by various teams and I
> bet there are tons of other tools out there that can do that. I can
> not speak for all teams but over the time many teams at Red Hat have
> come to like pbench
> https://distributed-system-analysis.github.io/pbench/ to run the tests
> and produce machine readable results and use other tools (Ansible,
> scripts, ...) to provision the systems and to generate the comparisons.
>
> As for myself I used python for PoC and over the last year I pushed
> hard to turn it into a usable and sensible tool which I'd like to
> offer: https://run-perf.readthedocs.io/en/latest/ anyway I am open to
> suggestions and comparisons. As I am using it downstream to watch
> regressions I do plan on keep developing the tool as well as the
> pipelines (unless a better tool is found that would replace it or it's
> parts).
FYI, Intel has invented a lot on the 0-day Linux kernel automated
performance regression test: https://01.org/lkp. It's being actively
developed upstream.
It's powerful and tons of regressions were reported (and bisected).
I think it can use qemu somehow but I'm not sure. Maybe we can have a try.
Thanks
>
> How
> ===
>
> This is a tough question. Ideally this should be a standalone service
> that would only notify the author of the patch that caused the change
> with a bunch of useful data so they can either address the issue or
> just be aware of this change and mark it as expected.
>
> Ideally the community should have a way to also issue their custom
> builds in order to verify their patches so they can debug and address
> issues better than just commit to qemu-master.
>
> The problem with those is that we can not simply use travis/gitlab/...
> machines for running those tests, because we are measuring in-guest
> actual performance. We can't just stop the time when the machine
> decides to schedule another container/vm. I briefly checked the public
> bare-metal offerings like rackspace but these are most probably not
> sufficient either because (unless I'm wrong) they only give you a
> machine but it is not guaranteed that it will be the same machine the
> next time. If we are to compare the results we don't need just the
> same model, we really need the very same machine. Any change to the
> machine might lead to a significant difference (disk replacement, even
> firmware update...).
>
> Solution 1
> ----------
>
> Doing this for downstream builds I can start doing this for upstream
> as well. At this point I can offer a single pipeline watching only
> changes in qemu (downstream we are checking distro/kernel changes as
> well but that would require too much time at this point) on a single
> x86_64 machine. I can not offer a public access to the testing
> machine, not even checking custom builds (unless someone provides me a
> publicly available machine(s) that I would use for this). What I can
> offer is running the checks on the latest qemu master, publishing the
> reports, bisecting issues and notifying people about the changes. An
> example of a report can be found here:
> https://drive.google.com/file/d/1V2w7QpSuybNusUaGxnyT5zTUvtZDOfsb/view?usp=sharing
> a documentation of the format is here:
> https://run-perf.readthedocs.io/en/latest/scripts.html#html-results I
> can also attach the raw pbench results if needed (as well as details
> about the tests that were executed and the params and other details).
>
> Currently the covered scenarios would be a default libvirt machine
> with qcow2 storage and tuned libvirt machine (cpus, hugepages, numa,
> raw disk...) running fio, uperf and linpack on the latest GA RHEL. In
> the future I can add/tweak the scenarios as well as tests selection
> based on your feedback.
>
> Solution 2
> ----------
>
> I can offer a documentation:
> https://run-perf.readthedocs.io/en/latest/jenkins.html and someone can
> fork/inspire by it and setup the pipelines on their system, making it
> available to the outside world, add your custom scenarios and
> variants. Note the setup does not require Jenkins, it's just an
> example and could be easily turned into a cronjob or whatever you chose.
>
> Solution 3
> ----------
>
> You name it. I bet there are many other ways to perform system-wide
> performance testing.
>
> Regards,
> Lukáš
>
>
next prev parent reply other threads:[~2020-11-26 8:24 UTC|newest]
Thread overview: 22+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-11-26 8:10 Proposal for a regular upstream performance testing Lukáš Doktor
2020-11-26 8:23 ` Jason Wang [this message]
2020-11-26 9:43 ` Daniel P. Berrangé
2020-11-26 11:29 ` Lukáš Doktor
2020-11-30 13:23 ` Stefan Hajnoczi
2020-12-01 7:51 ` Lukáš Doktor
2020-11-26 10:17 ` Peter Maydell
2020-11-26 11:16 ` Lukáš Doktor
2020-11-30 13:25 ` Stefan Hajnoczi
2020-12-01 8:05 ` Lukáš Doktor
2020-12-01 10:22 ` Stefan Hajnoczi
2020-12-01 12:06 ` Lukáš Doktor
2020-12-01 12:35 ` Stefan Hajnoczi
2020-12-02 8:58 ` Chenqun (kuhn)
2020-12-02 8:23 ` Chenqun (kuhn)
2022-03-21 8:46 ` Lukáš Doktor
2022-03-21 9:42 ` Stefan Hajnoczi
2022-03-21 10:29 ` Lukáš Doktor
2022-03-22 15:05 ` Stefan Hajnoczi
2022-03-28 6:18 ` Lukáš Doktor
2022-03-28 9:57 ` Stefan Hajnoczi
2022-03-28 11:09 ` Lukáš Doktor
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=32b0753d-d5bd-4790-a88b-998b152534bd@redhat.com \
--to=jasowang@redhat.com \
--cc=aleksandar.qemu.devel@gmail.com \
--cc=cheshi@redhat.com \
--cc=ldoktor@redhat.com \
--cc=qemu-devel@nongnu.org \
--cc=stefanha@redhat.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).