From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-7.3 required=3.0 tests=BAYES_00,DKIMWL_WL_HIGH, DKIM_SIGNED,DKIM_VALID,DKIM_VALID_AU,HEADER_FROM_DIFFERENT_DOMAINS, MAILING_LIST_MULTI,NICE_REPLY_A,SPF_HELO_NONE,SPF_PASS,USER_AGENT_SANE_1 autolearn=no autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 7BFCEC6379D for ; Thu, 26 Nov 2020 08:24:56 +0000 (UTC) Received: from lists.gnu.org (lists.gnu.org [209.51.188.17]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPS id 8E12520DD4 for ; Thu, 26 Nov 2020 08:24:55 +0000 (UTC) Authentication-Results: mail.kernel.org; dkim=pass (1024-bit key) header.d=redhat.com header.i=@redhat.com header.b="KkS8r6ne" DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org 8E12520DD4 Authentication-Results: mail.kernel.org; dmarc=fail (p=none dis=none) header.from=redhat.com Authentication-Results: mail.kernel.org; spf=pass smtp.mailfrom=qemu-devel-bounces+qemu-devel=archiver.kernel.org@nongnu.org Received: from localhost ([::1]:60734 helo=lists1p.gnu.org) by lists.gnu.org with esmtp (Exim 4.90_1) (envelope-from ) id 1kiCaD-0005fi-NF for qemu-devel@archiver.kernel.org; Thu, 26 Nov 2020 03:24:53 -0500 Received: from eggs.gnu.org ([2001:470:142:3::10]:39772) by lists.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1kiCZN-0005GA-6G for qemu-devel@nongnu.org; Thu, 26 Nov 2020 03:24:01 -0500 Received: from us-smtp-delivery-124.mimecast.com ([63.128.21.124]:37596) by eggs.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_CBC_SHA1:256) (Exim 4.90_1) (envelope-from ) id 1kiCZJ-0006tN-6G for qemu-devel@nongnu.org; Thu, 26 Nov 2020 03:23:59 -0500 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1606379035; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=d7/l/lFMfndFuBKAqOtvuslUEhRFRYw3ucgr2xZwNIs=; b=KkS8r6nejoT0IMrdEbHRdfeyYVpAGTEOVM2hoKLwiIAN8UgLxZN0JysVPcLI21y9Vfs7SK n4UTa4AN97veW3RWyNUe6qskeW113aIhjDj2dMQVxweBVsEXNKUt28BHtn9ibkrTThMbZJ hW8luDzBtZwyJMabkSmEoml4yIbgarI= Received: from mimecast-mx01.redhat.com (mimecast-mx01.redhat.com [209.132.183.4]) (Using TLS) by relay.mimecast.com with ESMTP id us-mta-104-xFZDbwVuMQuEJDi_JUPjBQ-1; Thu, 26 Nov 2020 03:23:51 -0500 X-MC-Unique: xFZDbwVuMQuEJDi_JUPjBQ-1 Received: from smtp.corp.redhat.com (int-mx01.intmail.prod.int.phx2.redhat.com [10.5.11.11]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mimecast-mx01.redhat.com (Postfix) with ESMTPS id 97CDF3E754; Thu, 26 Nov 2020 08:23:50 +0000 (UTC) Received: from [10.72.13.213] (ovpn-13-213.pek2.redhat.com [10.72.13.213]) by smtp.corp.redhat.com (Postfix) with ESMTP id 79CFF60636; Thu, 26 Nov 2020 08:23:48 +0000 (UTC) Subject: Re: Proposal for a regular upstream performance testing To: =?UTF-8?B?THVrw6HFoSBEb2t0b3I=?= , QEMU Developers References: <3a664806-8aa3-feb4-fb30-303d303217a8@redhat.com> From: Jason Wang Message-ID: <32b0753d-d5bd-4790-a88b-998b152534bd@redhat.com> Date: Thu, 26 Nov 2020 16:23:47 +0800 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:68.0) Gecko/20100101 Thunderbird/68.10.0 MIME-Version: 1.0 In-Reply-To: <3a664806-8aa3-feb4-fb30-303d303217a8@redhat.com> X-Scanned-By: MIMEDefang 2.79 on 10.5.11.11 Authentication-Results: relay.mimecast.com; auth=pass smtp.auth=CUSA124A263 smtp.mailfrom=jasowang@redhat.com X-Mimecast-Spam-Score: 0 X-Mimecast-Originator: redhat.com Content-Type: text/plain; charset=utf-8; format=flowed Content-Transfer-Encoding: 8bit Content-Language: en-US Received-SPF: pass client-ip=63.128.21.124; envelope-from=jasowang@redhat.com; helo=us-smtp-delivery-124.mimecast.com X-Spam_score_int: -20 X-Spam_score: -2.1 X-Spam_bar: -- X-Spam_report: (-2.1 / 5.0 requ) BAYES_00=-1.9, DKIMWL_WL_HIGH=-0.001, DKIM_SIGNED=0.1, DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, DKIM_VALID_EF=-0.1, NICE_REPLY_A=-0.001, RCVD_IN_DNSWL_NONE=-0.0001, RCVD_IN_MSPIKE_H4=0.001, RCVD_IN_MSPIKE_WL=0.001, SPF_HELO_NONE=0.001, SPF_PASS=-0.001 autolearn=ham autolearn_force=no X-Spam_action: no action X-BeenThere: qemu-devel@nongnu.org X-Mailman-Version: 2.1.23 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: Charles Shih , Aleksandar Markovic , Stefan Hajnoczi Errors-To: qemu-devel-bounces+qemu-devel=archiver.kernel.org@nongnu.org Sender: "Qemu-devel" On 2020/11/26 下午4:10, Lukáš Doktor wrote: > Hello guys, > > I had been around qemu on the Avocado-vt side for quite some time and > a while ago I shifted my focus on performance testing. Currently I am > not aware of any upstream CI that would continuously monitor the > upstream qemu performance and I'd like to change that. There is a lot > to cover so please bear with me. > > Goal > ==== > > The goal of this initiative is to detect system-wide performance > regressions as well as improvements early, ideally pin-point the > individual commits and notify people that they should fix things. All > in upstream and ideally with least human interaction possible. > > Unlike the recent work of Ahmed Karaman's > https://ahmedkrmn.github.io/TCG-Continuous-Benchmarking/ my aim is on > the system-wide performance inside the guest (like fio, uperf, ...) > > Tools > ===== > > In house we have several different tools used by various teams and I > bet there are tons of other tools out there that can do that. I can > not speak for all teams but over the time many teams at Red Hat have > come to like pbench > https://distributed-system-analysis.github.io/pbench/ to run the tests > and produce machine readable results and use other tools (Ansible, > scripts, ...) to provision the systems and to generate the comparisons. > > As for myself I used python for PoC and over the last year I pushed > hard to turn it into a usable and sensible tool which I'd like to > offer: https://run-perf.readthedocs.io/en/latest/ anyway I am open to > suggestions and comparisons. As I am using it downstream to watch > regressions I do plan on keep developing the tool as well as the > pipelines (unless a better tool is found that would replace it or it's > parts). FYI, Intel has invented a lot on the 0-day Linux kernel automated performance regression test: https://01.org/lkp. It's being actively developed upstream. It's powerful and tons of regressions were reported (and bisected). I think it can use qemu somehow but I'm not sure. Maybe we can have a try. Thanks > > How > === > > This is a tough question. Ideally this should be a standalone service > that would only notify the author of the patch that caused the change > with a bunch of useful data so they can either address the issue or > just be aware of this change and mark it as expected. > > Ideally the community should have a way to also issue their custom > builds in order to verify their patches so they can debug and address > issues better than just commit to qemu-master. > > The problem with those is that we can not simply use travis/gitlab/... > machines for running those tests, because we are measuring in-guest > actual performance. We can't just stop the time when the machine > decides to schedule another container/vm. I briefly checked the public > bare-metal offerings like rackspace but these are most probably not > sufficient either because (unless I'm wrong) they only give you a > machine but it is not guaranteed that it will be the same machine the > next time. If we are to compare the results we don't need just the > same model, we really need the very same machine. Any change to the > machine might lead to a significant difference (disk replacement, even > firmware update...). > > Solution 1 > ---------- > > Doing this for downstream builds I can start doing this for upstream > as well. At this point I can offer a single pipeline watching only > changes in qemu (downstream we are checking distro/kernel changes as > well but that would require too much time at this point) on a single > x86_64 machine. I can not offer a public access to the testing > machine, not even checking custom builds (unless someone provides me a > publicly available machine(s) that I would use for this). What I can > offer is running the checks on the latest qemu master, publishing the > reports, bisecting issues and notifying people about the changes. An > example of a report can be found here: > https://drive.google.com/file/d/1V2w7QpSuybNusUaGxnyT5zTUvtZDOfsb/view?usp=sharing > a documentation of the format is here: > https://run-perf.readthedocs.io/en/latest/scripts.html#html-results I > can also attach the raw pbench results if needed (as well as details > about the tests that were executed and the params and other details). > > Currently the covered scenarios would be a default libvirt machine > with qcow2 storage and tuned libvirt machine (cpus, hugepages, numa, > raw disk...) running fio, uperf and linpack on the latest GA RHEL. In > the future I can add/tweak the scenarios as well as tests selection > based on your feedback. > > Solution 2 > ---------- > > I can offer a documentation: > https://run-perf.readthedocs.io/en/latest/jenkins.html and someone can > fork/inspire by it and setup the pipelines on their system, making it > available to the outside world, add your custom scenarios and > variants. Note the setup does not require Jenkins, it's just an > example and could be easily turned into a cronjob or whatever you chose. > > Solution 3 > ---------- > > You name it. I bet there are many other ways to perform system-wide > performance testing. > > Regards, > Lukáš > >