From: Thomas Zimmermann <tzimmermann@suse.de>
To: "Noralf Trønnes" <noralf@tronnes.org>, "Daniel Vetter" <daniel@ffwll.ch>
Cc: Stephen Rothwell <sfr@canb.auug.org.au>,
Feng Tang <feng.tang@intel.com>,
rong.a.chen@intel.com, michel@daenzer.net,
dri-devel <dri-devel@lists.freedesktop.org>,
ying.huang@intel.com, lkp@01.org
Subject: Re: [drm/mgag200] 90f479ae51: vm-scalability.median -18.8% regression
Date: Sun, 4 Aug 2019 20:39:19 +0200 [thread overview]
Message-ID: <14fdaaed-51c8-b270-b46b-cba7b5c4ba52@suse.de> (raw)
In-Reply-To: <1c0bf22b-2c69-6b45-f700-ed832a3a5c17@suse.de>
[-- Attachment #1.1.1: Type: text/plain, Size: 63604 bytes --]
Hi
I did some further analysis on this problem and found that the blinking
cursor affects performance of the vm-scalability test case.
I only have a 4-core machine, so scalability is not really testable. Yet
I see the effects of running vm-scalibility against drm-tip, a revert of
the mgag200 patch and the vmap fixes that I posted a few days ago.
After reverting the mgag200 patch, running the test as described in the
report
bin/lkp run job.yaml
gives results like
2019-08-02 19:34:37 ./case-anon-cow-seq-hugetlb
2019-08-02 19:34:37 ./usemem --runtime 300 -n 4 --prealloc --prefault
-O -U 815395225
917319627 bytes / 756534 usecs = 1184110 KB/s
917319627 bytes / 764675 usecs = 1171504 KB/s
917319627 bytes / 766414 usecs = 1168846 KB/s
917319627 bytes / 777990 usecs = 1151454 KB/s
Running the test against current drm-tip gives slightly worse results,
such as.
2019-08-03 19:17:06 ./case-anon-cow-seq-hugetlb
2019-08-03 19:17:06 ./usemem --runtime 300 -n 4 --prealloc --prefault
-O -U 815394406
917318700 bytes / 871607 usecs = 1027778 KB/s
917318700 bytes / 894173 usecs = 1001840 KB/s
917318700 bytes / 919694 usecs = 974040 KB/s
917318700 bytes / 923341 usecs = 970193 KB/s
The test puts out roughly one result per second. Strangely sending the
output to /dev/null can make results significantly worse.
bin/lkp run job.yaml > /dev/null
2019-08-03 19:23:04 ./case-anon-cow-seq-hugetlb
2019-08-03 19:23:04 ./usemem --runtime 300 -n 4 --prealloc --prefault
-O -U 815394406
917318700 bytes / 1207358 usecs = 741966 KB/s
917318700 bytes / 1210456 usecs = 740067 KB/s
917318700 bytes / 1216572 usecs = 736346 KB/s
917318700 bytes / 1239152 usecs = 722929 KB/s
I realized that there's still a blinking cursor on the screen, which I
disabled with
tput civis
or alternatively
echo 0 > /sys/devices/virtual/graphics/fbcon/cursor_blink
Running the the test now gives the original or even better results, such as
bin/lkp run job.yaml > /dev/null
2019-08-03 19:29:17 ./case-anon-cow-seq-hugetlb
2019-08-03 19:29:17 ./usemem --runtime 300 -n 4 --prealloc --prefault
-O -U 815394406
917318700 bytes / 659419 usecs = 1358497 KB/s
917318700 bytes / 659658 usecs = 1358005 KB/s
917318700 bytes / 659916 usecs = 1357474 KB/s
917318700 bytes / 660168 usecs = 1356956 KB/s
Rong, Feng, could you confirm this by disabling the cursor or blinking?
The difference between mgag200's original fbdev support and generic
fbdev emulation is generic fbdev's worker task that updates the VRAM
buffer from the shadow buffer. mgag200 does this immediately, but relies
on drm_can_sleep(), which is deprecated.
I think that the worker task interferes with the test case, as the
worker has been in fbdev emulation since forever and no performance
regressions have been reported so far.
So unless there's a report where this problem happens in a real-world
use case, I'd like to keep code as it is. And apparently there's always
the workaround of disabling the cursor blinking.
Best regards
Thomas
Am 30.07.19 um 19:50 schrieb Thomas Zimmermann:
> Am 29.07.19 um 11:51 schrieb kernel test robot:
>> Greeting,
>>
>> FYI, we noticed a -18.8% regression of vm-scalability.median due to commit:>
>>
>> commit: 90f479ae51afa45efab97afdde9b94b9660dd3e4 ("drm/mgag200: Replace struct mga_fbdev with generic framebuffer emulation")
>> https://kernel.googlesource.com/pub/scm/linux/kernel/git/next/linux-next.git master
>
> Daniel, Noralf, we may have to revert this patch.
>
> I expected some change in display performance, but not in VM. Since it's
> a server chipset, probably no one cares much about display performance.
> So that seemed like a good trade-off for re-using shared code.
>
> Part of the patch set is that the generic fb emulation now maps and
> unmaps the fbdev BO when updating the screen. I guess that's the cause
> of the performance regression. And it should be visible with other
> drivers as well if they use a shadow FB for fbdev emulation.
>
> The thing is that we'd need another generic fbdev emulation for ast and
> mgag200 that handles this issue properly.
>
> Best regards
> Thomas
>
>>
>> in testcase: vm-scalability
>> on test machine: 288 threads Intel(R) Xeon Phi(TM) CPU 7295 @ 1.50GHz with 80G memory
>> with following parameters:
>>
>> runtime: 300s
>> size: 8T
>> test: anon-cow-seq-hugetlb
>> cpufreq_governor: performance
>>
>> test-description: The motivation behind this suite is to exercise functions and regions of the mm/ of the Linux kernel which are of interest to us.
>> test-url: https://git.kernel.org/cgit/linux/kernel/git/wfg/vm-scalability.git/
>>
>>
>>
>> Details are as below:
>> -------------------------------------------------------------------------------------------------->
>>
>>
>> To reproduce:
>>
>> git clone https://github.com/intel/lkp-tests.git
>> cd lkp-tests
>> bin/lkp install job.yaml # job file is attached in this email
>> bin/lkp run job.yaml
>>
>> =========================================================================================
>> compiler/cpufreq_governor/kconfig/rootfs/runtime/size/tbox_group/test/testcase:
>> gcc-7/performance/x86_64-rhel-7.6/debian-x86_64-2019-05-14.cgz/300s/8T/lkp-knm01/anon-cow-seq-hugetlb/vm-scalability
>>
>> commit:
>> f1f8555dfb ("drm/bochs: Use shadow buffer for bochs framebuffer console")
>> 90f479ae51 ("drm/mgag200: Replace struct mga_fbdev with generic framebuffer emulation")
>>
>> f1f8555dfb9a70a2 90f479ae51afa45efab97afdde9
>> ---------------- ---------------------------
>> fail:runs %reproduction fail:runs
>> | | |
>> 2:4 -50% :4 dmesg.WARNING:at#for_ip_interrupt_entry/0x
>> :4 25% 1:4 dmesg.WARNING:at_ip___perf_sw_event/0x
>> :4 25% 1:4 dmesg.WARNING:at_ip__fsnotify_parent/0x
>> %stddev %change %stddev
>> \ | \
>> 43955 ± 2% -18.8% 35691 vm-scalability.median
>> 0.06 ± 7% +193.0% 0.16 ± 2% vm-scalability.median_stddev
>> 14906559 ± 2% -17.9% 12237079 vm-scalability.throughput
>> 87651 ± 2% -17.4% 72374 vm-scalability.time.involuntary_context_switches
>> 2086168 -23.6% 1594224 vm-scalability.time.minor_page_faults
>> 15082 ± 2% -10.4% 13517 vm-scalability.time.percent_of_cpu_this_job_got
>> 29987 -8.9% 27327 vm-scalability.time.system_time
>> 15755 -12.4% 13795 vm-scalability.time.user_time
>> 122011 -19.3% 98418 vm-scalability.time.voluntary_context_switches
>> 3.034e+09 -23.6% 2.318e+09 vm-scalability.workload
>> 242478 ± 12% +68.5% 408518 ± 23% cpuidle.POLL.time
>> 2788 ± 21% +117.4% 6062 ± 26% cpuidle.POLL.usage
>> 56653 ± 10% +64.4% 93144 ± 20% meminfo.Mapped
>> 120392 ± 7% +14.0% 137212 ± 4% meminfo.Shmem
>> 47221 ± 11% +77.1% 83634 ± 22% numa-meminfo.node0.Mapped
>> 120465 ± 7% +13.9% 137205 ± 4% numa-meminfo.node0.Shmem
>> 2885513 -16.5% 2409384 numa-numastat.node0.local_node
>> 2885471 -16.5% 2409354 numa-numastat.node0.numa_hit
>> 11813 ± 11% +76.3% 20824 ± 22% numa-vmstat.node0.nr_mapped
>> 30096 ± 7% +13.8% 34238 ± 4% numa-vmstat.node0.nr_shmem
>> 43.72 ± 2% +5.5 49.20 mpstat.cpu.all.idle%
>> 0.03 ± 4% +0.0 0.05 ± 6% mpstat.cpu.all.soft%
>> 19.51 -2.4 17.08 mpstat.cpu.all.usr%
>> 1012 -7.9% 932.75 turbostat.Avg_MHz
>> 32.38 ± 10% +25.8% 40.73 turbostat.CPU%c1
>> 145.51 -3.1% 141.01 turbostat.PkgWatt
>> 15.09 -19.2% 12.19 turbostat.RAMWatt
>> 43.50 ± 2% +13.2% 49.25 vmstat.cpu.id
>> 18.75 ± 2% -13.3% 16.25 ± 2% vmstat.cpu.us
>> 152.00 ± 2% -9.5% 137.50 vmstat.procs.r
>> 4800 -13.1% 4173 vmstat.system.cs
>> 156170 -11.9% 137594 slabinfo.anon_vma.active_objs
>> 3395 -11.9% 2991 slabinfo.anon_vma.active_slabs
>> 156190 -11.9% 137606 slabinfo.anon_vma.num_objs
>> 3395 -11.9% 2991 slabinfo.anon_vma.num_slabs
>> 1716 ± 5% +11.5% 1913 ± 8% slabinfo.dmaengine-unmap-16.active_objs
>> 1716 ± 5% +11.5% 1913 ± 8% slabinfo.dmaengine-unmap-16.num_objs
>> 1767 ± 2% -19.0% 1431 ± 2% slabinfo.hugetlbfs_inode_cache.active_objs
>> 1767 ± 2% -19.0% 1431 ± 2% slabinfo.hugetlbfs_inode_cache.num_objs
>> 3597 ± 5% -16.4% 3006 ± 3% slabinfo.skbuff_ext_cache.active_objs
>> 3597 ± 5% -16.4% 3006 ± 3% slabinfo.skbuff_ext_cache.num_objs
>> 1330122 -23.6% 1016557 proc-vmstat.htlb_buddy_alloc_success
>> 77214 ± 3% +6.4% 82128 ± 2% proc-vmstat.nr_active_anon
>> 67277 +2.9% 69246 proc-vmstat.nr_anon_pages
>> 218.50 ± 3% -10.6% 195.25 proc-vmstat.nr_dirtied
>> 288628 +1.4% 292755 proc-vmstat.nr_file_pages
>> 360.50 -2.7% 350.75 proc-vmstat.nr_inactive_file
>> 14225 ± 9% +63.8% 23304 ± 20% proc-vmstat.nr_mapped
>> 30109 ± 7% +13.8% 34259 ± 4% proc-vmstat.nr_shmem
>> 99870 -1.3% 98597 proc-vmstat.nr_slab_unreclaimable
>> 204.00 ± 4% -12.1% 179.25 proc-vmstat.nr_written
>> 77214 ± 3% +6.4% 82128 ± 2% proc-vmstat.nr_zone_active_anon
>> 360.50 -2.7% 350.75 proc-vmstat.nr_zone_inactive_file
>> 8810 ± 19% -66.1% 2987 ± 42% proc-vmstat.numa_hint_faults
>> 8810 ± 19% -66.1% 2987 ± 42% proc-vmstat.numa_hint_faults_local
>> 2904082 -16.4% 2427026 proc-vmstat.numa_hit
>> 2904081 -16.4% 2427025 proc-vmstat.numa_local
>> 6.828e+08 -23.5% 5.221e+08 proc-vmstat.pgalloc_normal
>> 2900008 -17.2% 2400195 proc-vmstat.pgfault
>> 6.827e+08 -23.5% 5.22e+08 proc-vmstat.pgfree
>> 1.635e+10 -17.0% 1.357e+10 perf-stat.i.branch-instructions
>> 1.53 ± 4% -0.1 1.45 ± 3% perf-stat.i.branch-miss-rate%
>> 2.581e+08 ± 3% -20.5% 2.051e+08 ± 2% perf-stat.i.branch-misses
>> 12.66 +1.1 13.78 perf-stat.i.cache-miss-rate%
>> 72720849 -12.0% 63958986 perf-stat.i.cache-misses
>> 5.766e+08 -18.6% 4.691e+08 perf-stat.i.cache-references
>> 4674 ± 2% -13.0% 4064 perf-stat.i.context-switches
>> 4.29 +12.5% 4.83 perf-stat.i.cpi
>> 2.573e+11 -7.4% 2.383e+11 perf-stat.i.cpu-cycles
>> 231.35 -21.5% 181.56 perf-stat.i.cpu-migrations
>> 3522 +4.4% 3677 perf-stat.i.cycles-between-cache-misses
>> 0.09 ± 13% +0.0 0.12 ± 5% perf-stat.i.iTLB-load-miss-rate%
>> 5.894e+10 -15.8% 4.961e+10 perf-stat.i.iTLB-loads
>> 5.901e+10 -15.8% 4.967e+10 perf-stat.i.instructions
>> 1291 ± 14% -21.8% 1010 perf-stat.i.instructions-per-iTLB-miss
>> 0.24 -11.0% 0.21 perf-stat.i.ipc
>> 9476 -17.5% 7821 perf-stat.i.minor-faults
>> 9478 -17.5% 7821 perf-stat.i.page-faults
>> 9.76 -3.6% 9.41 perf-stat.overall.MPKI
>> 1.59 ± 4% -0.1 1.52 perf-stat.overall.branch-miss-rate%
>> 12.61 +1.1 13.71 perf-stat.overall.cache-miss-rate%
>> 4.38 +10.5% 4.83 perf-stat.overall.cpi
>> 3557 +5.3% 3747 perf-stat.overall.cycles-between-cache-misses
>> 0.08 ± 12% +0.0 0.10 perf-stat.overall.iTLB-load-miss-rate%
>> 1268 ± 15% -23.0% 976.22 perf-stat.overall.instructions-per-iTLB-miss
>> 0.23 -9.5% 0.21 perf-stat.overall.ipc
>> 5815 +9.7% 6378 perf-stat.overall.path-length
>> 1.634e+10 -17.5% 1.348e+10 perf-stat.ps.branch-instructions
>> 2.595e+08 ± 3% -21.2% 2.043e+08 ± 2% perf-stat.ps.branch-misses
>> 72565205 -12.2% 63706339 perf-stat.ps.cache-misses
>> 5.754e+08 -19.2% 4.646e+08 perf-stat.ps.cache-references
>> 4640 ± 2% -12.5% 4060 perf-stat.ps.context-switches
>> 2.581e+11 -7.5% 2.387e+11 perf-stat.ps.cpu-cycles
>> 229.91 -22.0% 179.42 perf-stat.ps.cpu-migrations
>> 5.889e+10 -16.3% 4.927e+10 perf-stat.ps.iTLB-loads
>> 5.899e+10 -16.3% 4.938e+10 perf-stat.ps.instructions
>> 9388 -18.2% 7677 perf-stat.ps.minor-faults
>> 9389 -18.2% 7677 perf-stat.ps.page-faults
>> 1.764e+13 -16.2% 1.479e+13 perf-stat.total.instructions
>> 46803 ± 3% -18.8% 37982 ± 6% sched_debug.cfs_rq:/.exec_clock.min
>> 5320 ± 3% +23.7% 6581 ± 3% sched_debug.cfs_rq:/.exec_clock.stddev
>> 6737 ± 14% +58.1% 10649 ± 10% sched_debug.cfs_rq:/.load.avg
>> 587978 ± 17% +58.2% 930382 ± 9% sched_debug.cfs_rq:/.load.max
>> 46952 ± 16% +64.8% 77388 ± 11% sched_debug.cfs_rq:/.load.stddev
>> 7.12 ± 4% +49.1% 10.62 ± 6% sched_debug.cfs_rq:/.load_avg.avg
>> 474.40 ± 23% +67.5% 794.60 ± 10% sched_debug.cfs_rq:/.load_avg.max
>> 37.70 ± 11% +74.8% 65.90 ± 9% sched_debug.cfs_rq:/.load_avg.stddev
>> 13424269 ± 4% -15.6% 11328098 ± 2% sched_debug.cfs_rq:/.min_vruntime.avg
>> 15411275 ± 3% -12.4% 13505072 ± 2% sched_debug.cfs_rq:/.min_vruntime.max
>> 7939295 ± 6% -17.5% 6551322 ± 7% sched_debug.cfs_rq:/.min_vruntime.min
>> 21.44 ± 7% -56.1% 9.42 ± 4% sched_debug.cfs_rq:/.nr_spread_over.avg
>> 117.45 ± 11% -60.6% 46.30 ± 14% sched_debug.cfs_rq:/.nr_spread_over.max
>> 19.33 ± 8% -66.4% 6.49 ± 9% sched_debug.cfs_rq:/.nr_spread_over.stddev
>> 4.32 ± 15% +84.4% 7.97 ± 3% sched_debug.cfs_rq:/.runnable_load_avg.avg
>> 353.85 ± 29% +118.8% 774.35 ± 11% sched_debug.cfs_rq:/.runnable_load_avg.max
>> 27.30 ± 24% +118.5% 59.64 ± 9% sched_debug.cfs_rq:/.runnable_load_avg.stddev
>> 6729 ± 14% +58.2% 10644 ± 10% sched_debug.cfs_rq:/.runnable_weight.avg
>> 587978 ± 17% +58.2% 930382 ± 9% sched_debug.cfs_rq:/.runnable_weight.max
>> 46950 ± 16% +64.8% 77387 ± 11% sched_debug.cfs_rq:/.runnable_weight.stddev
>> 5305069 ± 4% -17.4% 4380376 ± 7% sched_debug.cfs_rq:/.spread0.avg
>> 7328745 ± 3% -9.9% 6600897 ± 3% sched_debug.cfs_rq:/.spread0.max
>> 2220837 ± 4% +55.8% 3460596 ± 5% sched_debug.cpu.avg_idle.avg
>> 4590666 ± 9% +76.8% 8117037 ± 15% sched_debug.cpu.avg_idle.max
>> 485052 ± 7% +80.3% 874679 ± 10% sched_debug.cpu.avg_idle.stddev
>> 561.50 ± 26% +37.7% 773.30 ± 15% sched_debug.cpu.clock.stddev
>> 561.50 ± 26% +37.7% 773.30 ± 15% sched_debug.cpu.clock_task.stddev
>> 3.20 ± 10% +109.6% 6.70 ± 3% sched_debug.cpu.cpu_load[0].avg
>> 309.10 ± 20% +150.3% 773.75 ± 12% sched_debug.cpu.cpu_load[0].max
>> 21.02 ± 14% +160.8% 54.80 ± 9% sched_debug.cpu.cpu_load[0].stddev
>> 3.19 ± 8% +109.8% 6.70 ± 3% sched_debug.cpu.cpu_load[1].avg
>> 299.75 ± 19% +158.0% 773.30 ± 12% sched_debug.cpu.cpu_load[1].max
>> 20.32 ± 12% +168.7% 54.62 ± 9% sched_debug.cpu.cpu_load[1].stddev
>> 3.20 ± 8% +109.1% 6.69 ± 4% sched_debug.cpu.cpu_load[2].avg
>> 288.90 ± 20% +167.0% 771.40 ± 12% sched_debug.cpu.cpu_load[2].max
>> 19.70 ± 12% +175.4% 54.27 ± 9% sched_debug.cpu.cpu_load[2].stddev
>> 3.16 ± 8% +110.9% 6.66 ± 6% sched_debug.cpu.cpu_load[3].avg
>> 275.50 ± 24% +178.4% 766.95 ± 12% sched_debug.cpu.cpu_load[3].max
>> 18.92 ± 15% +184.2% 53.77 ± 10% sched_debug.cpu.cpu_load[3].stddev
>> 3.08 ± 8% +115.7% 6.65 ± 7% sched_debug.cpu.cpu_load[4].avg
>> 263.55 ± 28% +188.7% 760.85 ± 12% sched_debug.cpu.cpu_load[4].max
>> 18.03 ± 18% +196.6% 53.46 ± 11% sched_debug.cpu.cpu_load[4].stddev
>> 14543 -9.6% 13150 sched_debug.cpu.curr->pid.max
>> 5293 ± 16% +74.7% 9248 ± 11% sched_debug.cpu.load.avg
>> 587978 ± 17% +58.2% 930382 ± 9% sched_debug.cpu.load.max
>> 40887 ± 19% +78.3% 72891 ± 9% sched_debug.cpu.load.stddev
>> 1141679 ± 4% +56.9% 1790907 ± 5% sched_debug.cpu.max_idle_balance_cost.avg
>> 2432100 ± 9% +72.6% 4196779 ± 13% sched_debug.cpu.max_idle_balance_cost.max
>> 745656 +29.3% 964170 ± 5% sched_debug.cpu.max_idle_balance_cost.min
>> 239032 ± 9% +81.9% 434806 ± 10% sched_debug.cpu.max_idle_balance_cost.stddev
>> 0.00 ± 27% +92.1% 0.00 ± 31% sched_debug.cpu.next_balance.stddev
>> 1030 ± 4% -10.4% 924.00 ± 2% sched_debug.cpu.nr_switches.min
>> 0.04 ± 26% +139.0% 0.09 ± 41% sched_debug.cpu.nr_uninterruptible.avg
>> 830.35 ± 6% -12.0% 730.50 ± 2% sched_debug.cpu.sched_count.min
>> 912.00 ± 2% -9.5% 825.38 sched_debug.cpu.ttwu_count.avg
>> 433.05 ± 3% -19.2% 350.05 ± 3% sched_debug.cpu.ttwu_count.min
>> 160.70 ± 3% -12.5% 140.60 ± 4% sched_debug.cpu.ttwu_local.min
>> 9072 ± 11% -36.4% 5767 ± 8% softirqs.CPU1.RCU
>> 12769 ± 5% +15.3% 14718 ± 3% softirqs.CPU101.SCHED
>> 13198 +11.5% 14717 ± 3% softirqs.CPU102.SCHED
>> 12981 ± 4% +13.9% 14788 ± 3% softirqs.CPU105.SCHED
>> 13486 ± 3% +11.8% 15071 ± 4% softirqs.CPU111.SCHED
>> 12794 ± 4% +14.1% 14601 ± 9% softirqs.CPU112.SCHED
>> 12999 ± 4% +10.1% 14314 ± 4% softirqs.CPU115.SCHED
>> 12844 ± 4% +10.6% 14202 ± 2% softirqs.CPU120.SCHED
>> 13336 ± 3% +9.4% 14585 ± 3% softirqs.CPU122.SCHED
>> 12639 ± 4% +20.2% 15195 softirqs.CPU123.SCHED
>> 13040 ± 5% +15.2% 15024 ± 5% softirqs.CPU126.SCHED
>> 13123 +15.1% 15106 ± 5% softirqs.CPU127.SCHED
>> 9188 ± 6% -35.7% 5911 ± 2% softirqs.CPU13.RCU
>> 13054 ± 3% +13.1% 14761 ± 5% softirqs.CPU130.SCHED
>> 13158 ± 2% +13.9% 14985 ± 5% softirqs.CPU131.SCHED
>> 12797 ± 6% +13.5% 14524 ± 3% softirqs.CPU133.SCHED
>> 12452 ± 5% +14.8% 14297 softirqs.CPU134.SCHED
>> 13078 ± 3% +10.4% 14439 ± 3% softirqs.CPU138.SCHED
>> 12617 ± 2% +14.5% 14442 ± 5% softirqs.CPU139.SCHED
>> 12974 ± 3% +13.7% 14752 ± 4% softirqs.CPU142.SCHED
>> 12579 ± 4% +19.1% 14983 ± 3% softirqs.CPU143.SCHED
>> 9122 ± 24% -44.6% 5053 ± 5% softirqs.CPU144.RCU
>> 13366 ± 2% +11.1% 14848 ± 3% softirqs.CPU149.SCHED
>> 13246 ± 2% +22.0% 16162 ± 7% softirqs.CPU150.SCHED
>> 13452 ± 3% +20.5% 16210 ± 7% softirqs.CPU151.SCHED
>> 13507 +10.1% 14869 softirqs.CPU156.SCHED
>> 13808 ± 3% +9.2% 15079 ± 4% softirqs.CPU157.SCHED
>> 13442 ± 2% +13.4% 15248 ± 4% softirqs.CPU160.SCHED
>> 13311 +12.1% 14920 ± 2% softirqs.CPU162.SCHED
>> 13544 ± 3% +8.5% 14695 ± 4% softirqs.CPU163.SCHED
>> 13648 ± 3% +11.2% 15179 ± 2% softirqs.CPU166.SCHED
>> 13404 ± 4% +12.5% 15079 ± 3% softirqs.CPU168.SCHED
>> 13421 ± 6% +16.0% 15568 ± 8% softirqs.CPU169.SCHED
>> 13115 ± 3% +23.1% 16139 ± 10% softirqs.CPU171.SCHED
>> 13424 ± 6% +10.4% 14822 ± 3% softirqs.CPU175.SCHED
>> 13274 ± 3% +13.7% 15087 ± 9% softirqs.CPU185.SCHED
>> 13409 ± 3% +12.3% 15063 ± 3% softirqs.CPU190.SCHED
>> 13181 ± 7% +13.4% 14946 ± 3% softirqs.CPU196.SCHED
>> 13578 ± 3% +10.9% 15061 softirqs.CPU197.SCHED
>> 13323 ± 5% +24.8% 16627 ± 6% softirqs.CPU198.SCHED
>> 14072 ± 2% +12.3% 15798 ± 7% softirqs.CPU199.SCHED
>> 12604 ± 13% +17.9% 14865 softirqs.CPU201.SCHED
>> 13380 ± 4% +14.8% 15356 ± 3% softirqs.CPU203.SCHED
>> 13481 ± 8% +14.2% 15390 ± 3% softirqs.CPU204.SCHED
>> 12921 ± 2% +13.8% 14710 ± 3% softirqs.CPU206.SCHED
>> 13468 +13.0% 15218 ± 2% softirqs.CPU208.SCHED
>> 13253 ± 2% +13.1% 14992 softirqs.CPU209.SCHED
>> 13319 ± 2% +14.3% 15225 ± 7% softirqs.CPU210.SCHED
>> 13673 ± 5% +16.3% 15895 ± 3% softirqs.CPU211.SCHED
>> 13290 +17.0% 15556 ± 5% softirqs.CPU212.SCHED
>> 13455 ± 4% +14.4% 15392 ± 3% softirqs.CPU213.SCHED
>> 13454 ± 4% +14.3% 15377 ± 3% softirqs.CPU215.SCHED
>> 13872 ± 7% +9.7% 15221 ± 5% softirqs.CPU220.SCHED
>> 13555 ± 4% +17.3% 15896 ± 5% softirqs.CPU222.SCHED
>> 13411 ± 4% +20.8% 16197 ± 6% softirqs.CPU223.SCHED
>> 8472 ± 21% -44.8% 4680 ± 3% softirqs.CPU224.RCU
>> 13141 ± 3% +16.2% 15265 ± 7% softirqs.CPU225.SCHED
>> 14084 ± 3% +8.2% 15242 ± 2% softirqs.CPU226.SCHED
>> 13528 ± 4% +11.3% 15063 ± 4% softirqs.CPU228.SCHED
>> 13218 ± 3% +16.3% 15377 ± 4% softirqs.CPU229.SCHED
>> 14031 ± 4% +10.2% 15467 ± 2% softirqs.CPU231.SCHED
>> 13770 ± 3% +14.0% 15700 ± 3% softirqs.CPU232.SCHED
>> 13456 ± 3% +12.3% 15105 ± 3% softirqs.CPU233.SCHED
>> 13137 ± 4% +13.5% 14909 ± 3% softirqs.CPU234.SCHED
>> 13318 ± 2% +14.7% 15280 ± 2% softirqs.CPU235.SCHED
>> 13690 ± 2% +13.7% 15563 ± 7% softirqs.CPU238.SCHED
>> 13771 ± 5% +20.8% 16634 ± 7% softirqs.CPU241.SCHED
>> 13317 ± 7% +19.5% 15919 ± 9% softirqs.CPU243.SCHED
>> 8234 ± 16% -43.9% 4616 ± 5% softirqs.CPU244.RCU
>> 13845 ± 6% +13.0% 15643 ± 3% softirqs.CPU244.SCHED
>> 13179 ± 3% +16.3% 15323 softirqs.CPU246.SCHED
>> 13754 +12.2% 15438 ± 3% softirqs.CPU248.SCHED
>> 13769 ± 4% +10.9% 15276 ± 2% softirqs.CPU252.SCHED
>> 13702 +10.5% 15147 ± 2% softirqs.CPU254.SCHED
>> 13315 ± 2% +12.5% 14980 ± 3% softirqs.CPU255.SCHED
>> 13785 ± 3% +12.9% 15568 ± 5% softirqs.CPU256.SCHED
>> 13307 ± 3% +15.0% 15298 ± 3% softirqs.CPU257.SCHED
>> 13864 ± 3% +10.5% 15313 ± 2% softirqs.CPU259.SCHED
>> 13879 ± 2% +11.4% 15465 softirqs.CPU261.SCHED
>> 13815 +13.6% 15687 ± 5% softirqs.CPU264.SCHED
>> 119574 ± 2% +11.8% 133693 ± 11% softirqs.CPU266.TIMER
>> 13688 +10.9% 15180 ± 6% softirqs.CPU267.SCHED
>> 11716 ± 4% +19.3% 13974 ± 8% softirqs.CPU27.SCHED
>> 13866 ± 3% +13.7% 15765 ± 4% softirqs.CPU271.SCHED
>> 13887 ± 5% +12.5% 15621 softirqs.CPU272.SCHED
>> 13383 ± 3% +19.8% 16031 ± 2% softirqs.CPU274.SCHED
>> 13347 +14.1% 15232 ± 3% softirqs.CPU275.SCHED
>> 12884 ± 2% +21.0% 15593 ± 4% softirqs.CPU276.SCHED
>> 13131 ± 5% +13.4% 14891 ± 5% softirqs.CPU277.SCHED
>> 12891 ± 2% +19.2% 15371 ± 4% softirqs.CPU278.SCHED
>> 13313 ± 4% +13.0% 15049 ± 2% softirqs.CPU279.SCHED
>> 13514 ± 3% +10.2% 14897 ± 2% softirqs.CPU280.SCHED
>> 13501 ± 3% +13.7% 15346 softirqs.CPU281.SCHED
>> 13261 +17.5% 15577 softirqs.CPU282.SCHED
>> 8076 ± 15% -43.7% 4546 ± 5% softirqs.CPU283.RCU
>> 13686 ± 3% +12.6% 15413 ± 2% softirqs.CPU284.SCHED
>> 13439 ± 2% +9.2% 14670 ± 4% softirqs.CPU285.SCHED
>> 8878 ± 9% -35.4% 5735 ± 4% softirqs.CPU35.RCU
>> 11690 ± 2% +13.6% 13274 ± 5% softirqs.CPU40.SCHED
>> 11714 ± 2% +19.3% 13975 ± 13% softirqs.CPU41.SCHED
>> 11763 +12.5% 13239 ± 4% softirqs.CPU45.SCHED
>> 11662 ± 2% +9.4% 12757 ± 3% softirqs.CPU46.SCHED
>> 11805 ± 2% +9.3% 12902 ± 2% softirqs.CPU50.SCHED
>> 12158 ± 3% +12.3% 13655 ± 8% softirqs.CPU55.SCHED
>> 11716 ± 4% +8.8% 12751 ± 3% softirqs.CPU58.SCHED
>> 11922 ± 2% +9.9% 13100 ± 4% softirqs.CPU64.SCHED
>> 9674 ± 17% -41.8% 5625 ± 6% softirqs.CPU66.RCU
>> 11818 +12.0% 13237 softirqs.CPU66.SCHED
>> 124682 ± 7% -6.1% 117088 ± 5% softirqs.CPU66.TIMER
>> 8637 ± 9% -34.0% 5700 ± 7% softirqs.CPU70.RCU
>> 11624 ± 2% +11.0% 12901 ± 2% softirqs.CPU70.SCHED
>> 12372 ± 2% +13.2% 14003 ± 3% softirqs.CPU71.SCHED
>> 9949 ± 25% -33.9% 6574 ± 31% softirqs.CPU72.RCU
>> 10392 ± 26% -35.1% 6745 ± 35% softirqs.CPU73.RCU
>> 12766 ± 3% +11.1% 14188 ± 3% softirqs.CPU76.SCHED
>> 12611 ± 2% +18.8% 14984 ± 5% softirqs.CPU78.SCHED
>> 12786 ± 3% +17.9% 15079 ± 7% softirqs.CPU79.SCHED
>> 11947 ± 4% +9.7% 13103 ± 4% softirqs.CPU8.SCHED
>> 13379 ± 7% +11.8% 14962 ± 4% softirqs.CPU83.SCHED
>> 13438 ± 5% +9.7% 14738 ± 2% softirqs.CPU84.SCHED
>> 12768 +19.4% 15241 ± 6% softirqs.CPU88.SCHED
>> 8604 ± 13% -39.3% 5222 ± 3% softirqs.CPU89.RCU
>> 13077 ± 2% +17.1% 15308 ± 7% softirqs.CPU89.SCHED
>> 11887 ± 3% +20.1% 14272 ± 5% softirqs.CPU9.SCHED
>> 12723 ± 3% +11.3% 14165 ± 4% softirqs.CPU90.SCHED
>> 8439 ± 12% -38.9% 5153 ± 4% softirqs.CPU91.RCU
>> 13429 ± 3% +10.3% 14806 ± 2% softirqs.CPU95.SCHED
>> 12852 ± 4% +10.3% 14174 ± 5% softirqs.CPU96.SCHED
>> 13010 ± 2% +14.4% 14888 ± 5% softirqs.CPU97.SCHED
>> 2315644 ± 4% -36.2% 1477200 ± 4% softirqs.RCU
>> 1572 ± 10% +63.9% 2578 ± 39% interrupts.CPU0.NMI:Non-maskable_interrupts
>> 1572 ± 10% +63.9% 2578 ± 39% interrupts.CPU0.PMI:Performance_monitoring_interrupts
>> 252.00 ± 11% -35.2% 163.25 ± 13% interrupts.CPU104.RES:Rescheduling_interrupts
>> 2738 ± 24% +52.4% 4173 ± 19% interrupts.CPU105.NMI:Non-maskable_interrupts
>> 2738 ± 24% +52.4% 4173 ± 19% interrupts.CPU105.PMI:Performance_monitoring_interrupts
>> 245.75 ± 19% -31.0% 169.50 ± 7% interrupts.CPU105.RES:Rescheduling_interrupts
>> 228.75 ± 13% -24.7% 172.25 ± 19% interrupts.CPU106.RES:Rescheduling_interrupts
>> 2243 ± 15% +66.3% 3730 ± 35% interrupts.CPU113.NMI:Non-maskable_interrupts
>> 2243 ± 15% +66.3% 3730 ± 35% interrupts.CPU113.PMI:Performance_monitoring_interrupts
>> 2703 ± 31% +67.0% 4514 ± 33% interrupts.CPU118.NMI:Non-maskable_interrupts
>> 2703 ± 31% +67.0% 4514 ± 33% interrupts.CPU118.PMI:Performance_monitoring_interrupts
>> 2613 ± 25% +42.2% 3715 ± 24% interrupts.CPU121.NMI:Non-maskable_interrupts
>> 2613 ± 25% +42.2% 3715 ± 24% interrupts.CPU121.PMI:Performance_monitoring_interrupts
>> 311.50 ± 23% -47.7% 163.00 ± 9% interrupts.CPU122.RES:Rescheduling_interrupts
>> 266.75 ± 19% -31.6% 182.50 ± 15% interrupts.CPU124.RES:Rescheduling_interrupts
>> 293.75 ± 33% -32.3% 198.75 ± 19% interrupts.CPU125.RES:Rescheduling_interrupts
>> 2601 ± 36% +43.2% 3724 ± 29% interrupts.CPU127.NMI:Non-maskable_interrupts
>> 2601 ± 36% +43.2% 3724 ± 29% interrupts.CPU127.PMI:Performance_monitoring_interrupts
>> 2258 ± 21% +68.2% 3797 ± 29% interrupts.CPU13.NMI:Non-maskable_interrupts
>> 2258 ± 21% +68.2% 3797 ± 29% interrupts.CPU13.PMI:Performance_monitoring_interrupts
>> 3338 ± 29% +54.6% 5160 ± 9% interrupts.CPU139.NMI:Non-maskable_interrupts
>> 3338 ± 29% +54.6% 5160 ± 9% interrupts.CPU139.PMI:Performance_monitoring_interrupts
>> 219.50 ± 27% -23.0% 169.00 ± 21% interrupts.CPU139.RES:Rescheduling_interrupts
>> 290.25 ± 25% -32.5% 196.00 ± 11% interrupts.CPU14.RES:Rescheduling_interrupts
>> 243.50 ± 4% -16.0% 204.50 ± 12% interrupts.CPU140.RES:Rescheduling_interrupts
>> 1797 ± 15% +135.0% 4223 ± 46% interrupts.CPU147.NMI:Non-maskable_interrupts
>> 1797 ± 15% +135.0% 4223 ± 46% interrupts.CPU147.PMI:Performance_monitoring_interrupts
>> 2537 ± 22% +89.6% 4812 ± 28% interrupts.CPU15.NMI:Non-maskable_interrupts
>> 2537 ± 22% +89.6% 4812 ± 28% interrupts.CPU15.PMI:Performance_monitoring_interrupts
>> 292.25 ± 34% -33.9% 193.25 ± 6% interrupts.CPU15.RES:Rescheduling_interrupts
>> 424.25 ± 37% -58.5% 176.25 ± 14% interrupts.CPU158.RES:Rescheduling_interrupts
>> 312.50 ± 42% -54.2% 143.00 ± 18% interrupts.CPU159.RES:Rescheduling_interrupts
>> 725.00 ±118% -75.7% 176.25 ± 14% interrupts.CPU163.RES:Rescheduling_interrupts
>> 2367 ± 6% +59.9% 3786 ± 24% interrupts.CPU177.NMI:Non-maskable_interrupts
>> 2367 ± 6% +59.9% 3786 ± 24% interrupts.CPU177.PMI:Performance_monitoring_interrupts
>> 239.50 ± 30% -46.6% 128.00 ± 14% interrupts.CPU179.RES:Rescheduling_interrupts
>> 320.75 ± 15% -24.0% 243.75 ± 20% interrupts.CPU20.RES:Rescheduling_interrupts
>> 302.50 ± 17% -47.2% 159.75 ± 8% interrupts.CPU200.RES:Rescheduling_interrupts
>> 2166 ± 5% +92.0% 4157 ± 40% interrupts.CPU207.NMI:Non-maskable_interrupts
>> 2166 ± 5% +92.0% 4157 ± 40% interrupts.CPU207.PMI:Performance_monitoring_interrupts
>> 217.00 ± 11% -34.6% 142.00 ± 12% interrupts.CPU214.RES:Rescheduling_interrupts
>> 2610 ± 36% +47.4% 3848 ± 35% interrupts.CPU215.NMI:Non-maskable_interrupts
>> 2610 ± 36% +47.4% 3848 ± 35% interrupts.CPU215.PMI:Performance_monitoring_interrupts
>> 2046 ± 13% +118.6% 4475 ± 43% interrupts.CPU22.NMI:Non-maskable_interrupts
>> 2046 ± 13% +118.6% 4475 ± 43% interrupts.CPU22.PMI:Performance_monitoring_interrupts
>> 289.50 ± 28% -41.1% 170.50 ± 8% interrupts.CPU22.RES:Rescheduling_interrupts
>> 2232 ± 6% +33.0% 2970 ± 24% interrupts.CPU221.NMI:Non-maskable_interrupts
>> 2232 ± 6% +33.0% 2970 ± 24% interrupts.CPU221.PMI:Performance_monitoring_interrupts
>> 4552 ± 12% -27.6% 3295 ± 15% interrupts.CPU222.NMI:Non-maskable_interrupts
>> 4552 ± 12% -27.6% 3295 ± 15% interrupts.CPU222.PMI:Performance_monitoring_interrupts
>> 2013 ± 15% +80.9% 3641 ± 27% interrupts.CPU226.NMI:Non-maskable_interrupts
>> 2013 ± 15% +80.9% 3641 ± 27% interrupts.CPU226.PMI:Performance_monitoring_interrupts
>> 2575 ± 49% +67.1% 4302 ± 34% interrupts.CPU227.NMI:Non-maskable_interrupts
>> 2575 ± 49% +67.1% 4302 ± 34% interrupts.CPU227.PMI:Performance_monitoring_interrupts
>> 248.00 ± 36% -36.3% 158.00 ± 19% interrupts.CPU228.RES:Rescheduling_interrupts
>> 2441 ± 24% +43.0% 3490 ± 30% interrupts.CPU23.NMI:Non-maskable_interrupts
>> 2441 ± 24% +43.0% 3490 ± 30% interrupts.CPU23.PMI:Performance_monitoring_interrupts
>> 404.25 ± 69% -65.5% 139.50 ± 17% interrupts.CPU236.RES:Rescheduling_interrupts
>> 566.50 ± 40% -73.6% 149.50 ± 31% interrupts.CPU237.RES:Rescheduling_interrupts
>> 243.50 ± 26% -37.1% 153.25 ± 21% interrupts.CPU248.RES:Rescheduling_interrupts
>> 258.25 ± 12% -53.5% 120.00 ± 18% interrupts.CPU249.RES:Rescheduling_interrupts
>> 2888 ± 27% +49.4% 4313 ± 30% interrupts.CPU253.NMI:Non-maskable_interrupts
>> 2888 ± 27% +49.4% 4313 ± 30% interrupts.CPU253.PMI:Performance_monitoring_interrupts
>> 2468 ± 44% +67.3% 4131 ± 37% interrupts.CPU256.NMI:Non-maskable_interrupts
>> 2468 ± 44% +67.3% 4131 ± 37% interrupts.CPU256.PMI:Performance_monitoring_interrupts
>> 425.00 ± 59% -60.3% 168.75 ± 34% interrupts.CPU258.RES:Rescheduling_interrupts
>> 1859 ± 16% +106.3% 3834 ± 44% interrupts.CPU268.NMI:Non-maskable_interrupts
>> 1859 ± 16% +106.3% 3834 ± 44% interrupts.CPU268.PMI:Performance_monitoring_interrupts
>> 2684 ± 28% +61.2% 4326 ± 36% interrupts.CPU269.NMI:Non-maskable_interrupts
>> 2684 ± 28% +61.2% 4326 ± 36% interrupts.CPU269.PMI:Performance_monitoring_interrupts
>> 2171 ± 6% +108.8% 4533 ± 20% interrupts.CPU270.NMI:Non-maskable_interrupts
>> 2171 ± 6% +108.8% 4533 ± 20% interrupts.CPU270.PMI:Performance_monitoring_interrupts
>> 2262 ± 14% +61.8% 3659 ± 37% interrupts.CPU273.NMI:Non-maskable_interrupts
>> 2262 ± 14% +61.8% 3659 ± 37% interrupts.CPU273.PMI:Performance_monitoring_interrupts
>> 2203 ± 11% +50.7% 3320 ± 38% interrupts.CPU279.NMI:Non-maskable_interrupts
>> 2203 ± 11% +50.7% 3320 ± 38% interrupts.CPU279.PMI:Performance_monitoring_interrupts
>> 2433 ± 17% +52.9% 3721 ± 25% interrupts.CPU280.NMI:Non-maskable_interrupts
>> 2433 ± 17% +52.9% 3721 ± 25% interrupts.CPU280.PMI:Performance_monitoring_interrupts
>> 2778 ± 33% +63.1% 4531 ± 36% interrupts.CPU283.NMI:Non-maskable_interrupts
>> 2778 ± 33% +63.1% 4531 ± 36% interrupts.CPU283.PMI:Performance_monitoring_interrupts
>> 331.75 ± 32% -39.8% 199.75 ± 17% interrupts.CPU29.RES:Rescheduling_interrupts
>> 2178 ± 22% +53.9% 3353 ± 31% interrupts.CPU3.NMI:Non-maskable_interrupts
>> 2178 ± 22% +53.9% 3353 ± 31% interrupts.CPU3.PMI:Performance_monitoring_interrupts
>> 298.50 ± 30% -39.7% 180.00 ± 6% interrupts.CPU34.RES:Rescheduling_interrupts
>> 2490 ± 3% +58.7% 3953 ± 28% interrupts.CPU35.NMI:Non-maskable_interrupts
>> 2490 ± 3% +58.7% 3953 ± 28% interrupts.CPU35.PMI:Performance_monitoring_interrupts
>> 270.50 ± 24% -31.1% 186.25 ± 3% interrupts.CPU36.RES:Rescheduling_interrupts
>> 2493 ± 7% +57.0% 3915 ± 27% interrupts.CPU43.NMI:Non-maskable_interrupts
>> 2493 ± 7% +57.0% 3915 ± 27% interrupts.CPU43.PMI:Performance_monitoring_interrupts
>> 286.75 ± 36% -32.4% 193.75 ± 7% interrupts.CPU45.RES:Rescheduling_interrupts
>> 259.00 ± 12% -23.6% 197.75 ± 13% interrupts.CPU46.RES:Rescheduling_interrupts
>> 244.00 ± 21% -35.6% 157.25 ± 11% interrupts.CPU47.RES:Rescheduling_interrupts
>> 230.00 ± 7% -21.3% 181.00 ± 11% interrupts.CPU48.RES:Rescheduling_interrupts
>> 281.00 ± 13% -27.4% 204.00 ± 15% interrupts.CPU53.RES:Rescheduling_interrupts
>> 256.75 ± 5% -18.4% 209.50 ± 12% interrupts.CPU54.RES:Rescheduling_interrupts
>> 2433 ± 9% +68.4% 4098 ± 35% interrupts.CPU58.NMI:Non-maskable_interrupts
>> 2433 ± 9% +68.4% 4098 ± 35% interrupts.CPU58.PMI:Performance_monitoring_interrupts
>> 316.00 ± 25% -41.4% 185.25 ± 13% interrupts.CPU59.RES:Rescheduling_interrupts
>> 2703 ± 38% +56.0% 4217 ± 31% interrupts.CPU60.NMI:Non-maskable_interrupts
>> 2703 ± 38% +56.0% 4217 ± 31% interrupts.CPU60.PMI:Performance_monitoring_interrupts
>> 2425 ± 16% +39.9% 3394 ± 27% interrupts.CPU61.NMI:Non-maskable_interrupts
>> 2425 ± 16% +39.9% 3394 ± 27% interrupts.CPU61.PMI:Performance_monitoring_interrupts
>> 2388 ± 18% +69.5% 4047 ± 29% interrupts.CPU66.NMI:Non-maskable_interrupts
>> 2388 ± 18% +69.5% 4047 ± 29% interrupts.CPU66.PMI:Performance_monitoring_interrupts
>> 2322 ± 11% +93.4% 4491 ± 35% interrupts.CPU67.NMI:Non-maskable_interrupts
>> 2322 ± 11% +93.4% 4491 ± 35% interrupts.CPU67.PMI:Performance_monitoring_interrupts
>> 319.00 ± 40% -44.7% 176.25 ± 9% interrupts.CPU67.RES:Rescheduling_interrupts
>> 2512 ± 8% +28.1% 3219 ± 25% interrupts.CPU70.NMI:Non-maskable_interrupts
>> 2512 ± 8% +28.1% 3219 ± 25% interrupts.CPU70.PMI:Performance_monitoring_interrupts
>> 2290 ± 39% +78.7% 4094 ± 28% interrupts.CPU74.NMI:Non-maskable_interrupts
>> 2290 ± 39% +78.7% 4094 ± 28% interrupts.CPU74.PMI:Performance_monitoring_interrupts
>> 2446 ± 40% +94.8% 4764 ± 23% interrupts.CPU75.NMI:Non-maskable_interrupts
>> 2446 ± 40% +94.8% 4764 ± 23% interrupts.CPU75.PMI:Performance_monitoring_interrupts
>> 426.75 ± 61% -67.7% 138.00 ± 8% interrupts.CPU75.RES:Rescheduling_interrupts
>> 192.50 ± 13% +45.6% 280.25 ± 45% interrupts.CPU76.RES:Rescheduling_interrupts
>> 274.25 ± 34% -42.2% 158.50 ± 34% interrupts.CPU77.RES:Rescheduling_interrupts
>> 2357 ± 9% +73.0% 4078 ± 23% interrupts.CPU78.NMI:Non-maskable_interrupts
>> 2357 ± 9% +73.0% 4078 ± 23% interrupts.CPU78.PMI:Performance_monitoring_interrupts
>> 348.50 ± 53% -47.3% 183.75 ± 29% interrupts.CPU80.RES:Rescheduling_interrupts
>> 2650 ± 43% +46.2% 3874 ± 36% interrupts.CPU84.NMI:Non-maskable_interrupts
>> 2650 ± 43% +46.2% 3874 ± 36% interrupts.CPU84.PMI:Performance_monitoring_interrupts
>> 2235 ± 10% +117.8% 4867 ± 10% interrupts.CPU90.NMI:Non-maskable_interrupts
>> 2235 ± 10% +117.8% 4867 ± 10% interrupts.CPU90.PMI:Performance_monitoring_interrupts
>> 2606 ± 33% +38.1% 3598 ± 21% interrupts.CPU92.NMI:Non-maskable_interrupts
>> 2606 ± 33% +38.1% 3598 ± 21% interrupts.CPU92.PMI:Performance_monitoring_interrupts
>> 408.75 ± 58% -56.8% 176.75 ± 25% interrupts.CPU92.RES:Rescheduling_interrupts
>> 399.00 ± 64% -63.6% 145.25 ± 16% interrupts.CPU93.RES:Rescheduling_interrupts
>> 314.75 ± 36% -44.2% 175.75 ± 13% interrupts.CPU94.RES:Rescheduling_interrupts
>> 191.00 ± 15% -29.1% 135.50 ± 9% interrupts.CPU97.RES:Rescheduling_interrupts
>> 94.00 ± 8% +50.0% 141.00 ± 12% interrupts.IWI:IRQ_work_interrupts
>> 841457 ± 7% +16.6% 980751 ± 3% interrupts.NMI:Non-maskable_interrupts
>> 841457 ± 7% +16.6% 980751 ± 3% interrupts.PMI:Performance_monitoring_interrupts
>> 12.75 ± 11% -4.1 8.67 ± 31% perf-profile.calltrace.cycles-pp.do_rw_once
>> 1.02 ± 16% -0.6 0.47 ± 59% perf-profile.calltrace.cycles-pp.sched_clock.sched_clock_cpu.cpuidle_enter_state.cpuidle_enter.do_idle
>> 1.10 ± 15% -0.4 0.66 ± 14% perf-profile.calltrace.cycles-pp.sched_clock_cpu.cpuidle_enter_state.cpuidle_enter.do_idle.cpu_startup_entry
>> 1.05 ± 16% -0.4 0.61 ± 14% perf-profile.calltrace.cycles-pp.native_sched_clock.sched_clock.sched_clock_cpu.cpuidle_enter_state.cpuidle_enter
>> 1.58 ± 4% +0.3 1.91 ± 7% perf-profile.calltrace.cycles-pp.__hrtimer_run_queues.hrtimer_interrupt.smp_apic_timer_interrupt.apic_timer_interrupt.copy_page
>> 0.79 ± 26% +0.5 1.27 ± 18% perf-profile.calltrace.cycles-pp.__x64_sys_exit_group.do_syscall_64.entry_SYSCALL_64_after_hwframe
>> 0.79 ± 26% +0.5 1.27 ± 18% perf-profile.calltrace.cycles-pp.do_group_exit.__x64_sys_exit_group.do_syscall_64.entry_SYSCALL_64_after_hwframe
>> 0.79 ± 26% +0.5 1.27 ± 18% perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64.entry_SYSCALL_64_after_hwframe
>> 2.11 ± 4% +0.5 2.60 ± 7% perf-profile.calltrace.cycles-pp.apic_timer_interrupt.osq_lock.__mutex_lock.hugetlb_fault.handle_mm_fault
>> 0.83 ± 26% +0.5 1.32 ± 18% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
>> 0.83 ± 26% +0.5 1.32 ± 18% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
>> 1.90 ± 5% +0.6 2.45 ± 7% perf-profile.calltrace.cycles-pp.hrtimer_interrupt.smp_apic_timer_interrupt.apic_timer_interrupt.copy_page.copy_subpage
>> 0.65 ± 62% +0.6 1.20 ± 15% perf-profile.calltrace.cycles-pp.alloc_fresh_huge_page.alloc_surplus_huge_page.alloc_huge_page.hugetlb_cow.hugetlb_fault
>> 0.60 ± 62% +0.6 1.16 ± 18% perf-profile.calltrace.cycles-pp.free_huge_page.release_pages.tlb_flush_mmu.tlb_finish_mmu.exit_mmap
>> 0.95 ± 17% +0.6 1.52 ± 8% perf-profile.calltrace.cycles-pp.__hrtimer_run_queues.hrtimer_interrupt.smp_apic_timer_interrupt.apic_timer_interrupt.mutex_spin_on_owner
>> 0.61 ± 62% +0.6 1.18 ± 18% perf-profile.calltrace.cycles-pp.release_pages.tlb_flush_mmu.tlb_finish_mmu.exit_mmap.mmput
>> 0.61 ± 62% +0.6 1.19 ± 19% perf-profile.calltrace.cycles-pp.tlb_finish_mmu.exit_mmap.mmput.do_exit.do_group_exit
>> 0.61 ± 62% +0.6 1.19 ± 19% perf-profile.calltrace.cycles-pp.tlb_flush_mmu.tlb_finish_mmu.exit_mmap.mmput.do_exit
>> 0.64 ± 61% +0.6 1.23 ± 18% perf-profile.calltrace.cycles-pp.mmput.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
>> 0.64 ± 61% +0.6 1.23 ± 18% perf-profile.calltrace.cycles-pp.exit_mmap.mmput.do_exit.do_group_exit.__x64_sys_exit_group
>> 1.30 ± 9% +0.6 1.92 ± 8% perf-profile.calltrace.cycles-pp.hrtimer_interrupt.smp_apic_timer_interrupt.apic_timer_interrupt.mutex_spin_on_owner.__mutex_lock
>> 0.19 ±173% +0.7 0.89 ± 20% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.free_huge_page.release_pages.tlb_flush_mmu
>> 0.19 ±173% +0.7 0.90 ± 20% perf-profile.calltrace.cycles-pp._raw_spin_lock.free_huge_page.release_pages.tlb_flush_mmu.tlb_finish_mmu
>> 0.00 +0.8 0.77 ± 30% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.prep_new_huge_page.alloc_fresh_huge_page.alloc_surplus_huge_page
>> 0.00 +0.8 0.78 ± 30% perf-profile.calltrace.cycles-pp._raw_spin_lock.prep_new_huge_page.alloc_fresh_huge_page.alloc_surplus_huge_page.alloc_huge_page
>> 0.00 +0.8 0.79 ± 29% perf-profile.calltrace.cycles-pp.prep_new_huge_page.alloc_fresh_huge_page.alloc_surplus_huge_page.alloc_huge_page.hugetlb_cow
>> 0.82 ± 67% +0.9 1.72 ± 22% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.alloc_huge_page.hugetlb_cow.hugetlb_fault
>> 0.84 ± 66% +0.9 1.74 ± 20% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.alloc_surplus_huge_page.alloc_huge_page.hugetlb_cow
>> 2.52 ± 6% +0.9 3.44 ± 9% perf-profile.calltrace.cycles-pp.smp_apic_timer_interrupt.apic_timer_interrupt.copy_page.copy_subpage.copy_user_huge_page
>> 0.83 ± 67% +0.9 1.75 ± 21% perf-profile.calltrace.cycles-pp._raw_spin_lock.alloc_huge_page.hugetlb_cow.hugetlb_fault.handle_mm_fault
>> 0.84 ± 66% +0.9 1.77 ± 20% perf-profile.calltrace.cycles-pp._raw_spin_lock.alloc_surplus_huge_page.alloc_huge_page.hugetlb_cow.hugetlb_fault
>> 1.64 ± 12% +1.0 2.67 ± 7% perf-profile.calltrace.cycles-pp.smp_apic_timer_interrupt.apic_timer_interrupt.mutex_spin_on_owner.__mutex_lock.hugetlb_fault
>> 1.65 ± 45% +1.3 2.99 ± 18% perf-profile.calltrace.cycles-pp.alloc_surplus_huge_page.alloc_huge_page.hugetlb_cow.hugetlb_fault.handle_mm_fault
>> 1.74 ± 13% +1.4 3.16 ± 6% perf-profile.calltrace.cycles-pp.apic_timer_interrupt.mutex_spin_on_owner.__mutex_lock.hugetlb_fault.handle_mm_fault
>> 2.56 ± 48% +2.2 4.81 ± 19% perf-profile.calltrace.cycles-pp.alloc_huge_page.hugetlb_cow.hugetlb_fault.handle_mm_fault.__do_page_fault
>> 12.64 ± 14% +3.6 16.20 ± 8% perf-profile.calltrace.cycles-pp.mutex_spin_on_owner.__mutex_lock.hugetlb_fault.handle_mm_fault.__do_page_fault
>> 2.97 ± 7% +3.8 6.74 ± 9% perf-profile.calltrace.cycles-pp.apic_timer_interrupt.copy_page.copy_subpage.copy_user_huge_page.hugetlb_cow
>> 19.99 ± 9% +4.1 24.05 ± 6% perf-profile.calltrace.cycles-pp.hugetlb_cow.hugetlb_fault.handle_mm_fault.__do_page_fault.do_page_fault
>> 1.37 ± 15% -0.5 0.83 ± 13% perf-profile.children.cycles-pp.sched_clock_cpu
>> 1.31 ± 16% -0.5 0.78 ± 13% perf-profile.children.cycles-pp.sched_clock
>> 1.29 ± 16% -0.5 0.77 ± 13% perf-profile.children.cycles-pp.native_sched_clock
>> 1.80 ± 2% -0.3 1.47 ± 10% perf-profile.children.cycles-pp.task_tick_fair
>> 0.73 ± 2% -0.2 0.54 ± 11% perf-profile.children.cycles-pp.update_curr
>> 0.42 ± 17% -0.2 0.27 ± 16% perf-profile.children.cycles-pp.account_process_tick
>> 0.73 ± 10% -0.2 0.58 ± 9% perf-profile.children.cycles-pp.rcu_sched_clock_irq
>> 0.27 ± 6% -0.1 0.14 ± 14% perf-profile.children.cycles-pp.__acct_update_integrals
>> 0.27 ± 18% -0.1 0.16 ± 13% perf-profile.children.cycles-pp.rcu_segcblist_ready_cbs
>> 0.40 ± 12% -0.1 0.30 ± 14% perf-profile.children.cycles-pp.__next_timer_interrupt
>> 0.47 ± 7% -0.1 0.39 ± 13% perf-profile.children.cycles-pp.update_rq_clock
>> 0.29 ± 12% -0.1 0.21 ± 15% perf-profile.children.cycles-pp.cpuidle_governor_latency_req
>> 0.21 ± 7% -0.1 0.14 ± 12% perf-profile.children.cycles-pp.account_system_index_time
>> 0.38 ± 2% -0.1 0.31 ± 12% perf-profile.children.cycles-pp.timerqueue_add
>> 0.26 ± 11% -0.1 0.20 ± 13% perf-profile.children.cycles-pp.find_next_bit
>> 0.23 ± 15% -0.1 0.17 ± 15% perf-profile.children.cycles-pp.rcu_dynticks_eqs_exit
>> 0.14 ± 8% -0.1 0.07 ± 14% perf-profile.children.cycles-pp.account_user_time
>> 0.17 ± 6% -0.0 0.12 ± 10% perf-profile.children.cycles-pp.cpuacct_charge
>> 0.18 ± 20% -0.0 0.13 ± 3% perf-profile.children.cycles-pp.irq_work_tick
>> 0.11 ± 13% -0.0 0.07 ± 25% perf-profile.children.cycles-pp.tick_sched_do_timer
>> 0.12 ± 10% -0.0 0.08 ± 15% perf-profile.children.cycles-pp.get_cpu_device
>> 0.07 ± 11% -0.0 0.04 ± 58% perf-profile.children.cycles-pp.raise_softirq
>> 0.12 ± 3% -0.0 0.09 ± 8% perf-profile.children.cycles-pp.write
>> 0.11 ± 13% +0.0 0.14 ± 8% perf-profile.children.cycles-pp.native_write_msr
>> 0.09 ± 9% +0.0 0.11 ± 7% perf-profile.children.cycles-pp.finish_task_switch
>> 0.10 ± 10% +0.0 0.13 ± 5% perf-profile.children.cycles-pp.schedule_idle
>> 0.07 ± 6% +0.0 0.10 ± 12% perf-profile.children.cycles-pp.__read_nocancel
>> 0.04 ± 58% +0.0 0.07 ± 15% perf-profile.children.cycles-pp.__free_pages_ok
>> 0.06 ± 7% +0.0 0.09 ± 13% perf-profile.children.cycles-pp.perf_read
>> 0.07 +0.0 0.11 ± 14% perf-profile.children.cycles-pp.perf_evsel__read_counter
>> 0.07 +0.0 0.11 ± 13% perf-profile.children.cycles-pp.cmd_stat
>> 0.07 +0.0 0.11 ± 13% perf-profile.children.cycles-pp.__run_perf_stat
>> 0.07 +0.0 0.11 ± 13% perf-profile.children.cycles-pp.process_interval
>> 0.07 +0.0 0.11 ± 13% perf-profile.children.cycles-pp.read_counters
>> 0.07 ± 22% +0.0 0.11 ± 19% perf-profile.children.cycles-pp.__handle_mm_fault
>> 0.07 ± 19% +0.1 0.13 ± 8% perf-profile.children.cycles-pp.rb_erase
>> 0.03 ±100% +0.1 0.09 ± 9% perf-profile.children.cycles-pp.smp_call_function_single
>> 0.01 ±173% +0.1 0.08 ± 11% perf-profile.children.cycles-pp.perf_event_read
>> 0.00 +0.1 0.07 ± 13% perf-profile.children.cycles-pp.__perf_event_read_value
>> 0.00 +0.1 0.07 ± 7% perf-profile.children.cycles-pp.__intel_pmu_enable_all
>> 0.08 ± 17% +0.1 0.15 ± 8% perf-profile.children.cycles-pp.native_apic_msr_eoi_write
>> 0.04 ±103% +0.1 0.13 ± 58% perf-profile.children.cycles-pp.shmem_getpage_gfp
>> 0.38 ± 14% +0.1 0.51 ± 6% perf-profile.children.cycles-pp.run_timer_softirq
>> 0.11 ± 4% +0.3 0.37 ± 32% perf-profile.children.cycles-pp.worker_thread
>> 0.20 ± 5% +0.3 0.48 ± 25% perf-profile.children.cycles-pp.ret_from_fork
>> 0.20 ± 4% +0.3 0.48 ± 25% perf-profile.children.cycles-pp.kthread
>> 0.00 +0.3 0.29 ± 38% perf-profile.children.cycles-pp.memcpy_erms
>> 0.00 +0.3 0.29 ± 38% perf-profile.children.cycles-pp.drm_fb_helper_dirty_work
>> 0.00 +0.3 0.31 ± 37% perf-profile.children.cycles-pp.process_one_work
>> 0.47 ± 48% +0.4 0.91 ± 19% perf-profile.children.cycles-pp.prep_new_huge_page
>> 0.70 ± 29% +0.5 1.16 ± 18% perf-profile.children.cycles-pp.free_huge_page
>> 0.73 ± 29% +0.5 1.19 ± 18% perf-profile.children.cycles-pp.tlb_flush_mmu
>> 0.72 ± 29% +0.5 1.18 ± 18% perf-profile.children.cycles-pp.release_pages
>> 0.73 ± 29% +0.5 1.19 ± 18% perf-profile.children.cycles-pp.tlb_finish_mmu
>> 0.76 ± 27% +0.5 1.23 ± 18% perf-profile.children.cycles-pp.exit_mmap
>> 0.77 ± 27% +0.5 1.24 ± 18% perf-profile.children.cycles-pp.mmput
>> 0.79 ± 26% +0.5 1.27 ± 18% perf-profile.children.cycles-pp.__x64_sys_exit_group
>> 0.79 ± 26% +0.5 1.27 ± 18% perf-profile.children.cycles-pp.do_group_exit
>> 0.79 ± 26% +0.5 1.27 ± 18% perf-profile.children.cycles-pp.do_exit
>> 1.28 ± 29% +0.5 1.76 ± 9% perf-profile.children.cycles-pp.perf_mux_hrtimer_handler
>> 0.77 ± 28% +0.5 1.26 ± 13% perf-profile.children.cycles-pp.alloc_fresh_huge_page
>> 1.53 ± 15% +0.7 2.26 ± 14% perf-profile.children.cycles-pp.do_syscall_64
>> 1.53 ± 15% +0.7 2.27 ± 14% perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
>> 1.13 ± 3% +0.9 2.07 ± 14% perf-profile.children.cycles-pp.interrupt_entry
>> 0.79 ± 9% +1.0 1.76 ± 5% perf-profile.children.cycles-pp.perf_event_task_tick
>> 1.71 ± 39% +1.4 3.08 ± 16% perf-profile.children.cycles-pp.alloc_surplus_huge_page
>> 2.66 ± 42% +2.3 4.94 ± 17% perf-profile.children.cycles-pp.alloc_huge_page
>> 2.89 ± 45% +2.7 5.54 ± 18% perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
>> 3.34 ± 35% +2.7 6.02 ± 17% perf-profile.children.cycles-pp._raw_spin_lock
>> 12.77 ± 14% +3.9 16.63 ± 7% perf-profile.children.cycles-pp.mutex_spin_on_owner
>> 20.12 ± 9% +4.0 24.16 ± 6% perf-profile.children.cycles-pp.hugetlb_cow
>> 15.40 ± 10% -3.6 11.84 ± 28% perf-profile.self.cycles-pp.do_rw_once
>> 4.02 ± 9% -1.3 2.73 ± 30% perf-profile.self.cycles-pp.do_access
>> 2.00 ± 14% -0.6 1.41 ± 13% perf-profile.self.cycles-pp.cpuidle_enter_state
>> 1.26 ± 16% -0.5 0.74 ± 13% perf-profile.self.cycles-pp.native_sched_clock
>> 0.42 ± 17% -0.2 0.27 ± 16% perf-profile.self.cycles-pp.account_process_tick
>> 0.27 ± 19% -0.2 0.12 ± 17% perf-profile.self.cycles-pp.timerqueue_del
>> 0.53 ± 3% -0.1 0.38 ± 11% perf-profile.self.cycles-pp.update_curr
>> 0.27 ± 6% -0.1 0.14 ± 14% perf-profile.self.cycles-pp.__acct_update_integrals
>> 0.27 ± 18% -0.1 0.16 ± 13% perf-profile.self.cycles-pp.rcu_segcblist_ready_cbs
>> 0.61 ± 4% -0.1 0.51 ± 8% perf-profile.self.cycles-pp.task_tick_fair
>> 0.20 ± 8% -0.1 0.12 ± 14% perf-profile.self.cycles-pp.account_system_index_time
>> 0.23 ± 15% -0.1 0.16 ± 17% perf-profile.self.cycles-pp.rcu_dynticks_eqs_exit
>> 0.25 ± 11% -0.1 0.18 ± 14% perf-profile.self.cycles-pp.find_next_bit
>> 0.10 ± 11% -0.1 0.03 ±100% perf-profile.self.cycles-pp.tick_sched_do_timer
>> 0.29 -0.1 0.23 ± 11% perf-profile.self.cycles-pp.timerqueue_add
>> 0.12 ± 10% -0.1 0.06 ± 17% perf-profile.self.cycles-pp.account_user_time
>> 0.22 ± 15% -0.1 0.16 ± 6% perf-profile.self.cycles-pp.scheduler_tick
>> 0.17 ± 6% -0.0 0.12 ± 10% perf-profile.self.cycles-pp.cpuacct_charge
>> 0.18 ± 20% -0.0 0.13 ± 3% perf-profile.self.cycles-pp.irq_work_tick
>> 0.07 ± 13% -0.0 0.03 ±100% perf-profile.self.cycles-pp.update_process_times
>> 0.12 ± 7% -0.0 0.08 ± 15% perf-profile.self.cycles-pp.get_cpu_device
>> 0.07 ± 11% -0.0 0.04 ± 58% perf-profile.self.cycles-pp.raise_softirq
>> 0.12 ± 11% -0.0 0.09 ± 7% perf-profile.self.cycles-pp.tick_nohz_get_sleep_length
>> 0.11 ± 11% +0.0 0.14 ± 6% perf-profile.self.cycles-pp.native_write_msr
>> 0.10 ± 5% +0.1 0.15 ± 8% perf-profile.self.cycles-pp.__remove_hrtimer
>> 0.07 ± 23% +0.1 0.13 ± 8% perf-profile.self.cycles-pp.rb_erase
>> 0.08 ± 17% +0.1 0.15 ± 7% perf-profile.self.cycles-pp.native_apic_msr_eoi_write
>> 0.00 +0.1 0.08 ± 10% perf-profile.self.cycles-pp.smp_call_function_single
>> 0.32 ± 17% +0.1 0.42 ± 7% perf-profile.self.cycles-pp.run_timer_softirq
>> 0.22 ± 5% +0.1 0.34 ± 4% perf-profile.self.cycles-pp.ktime_get_update_offsets_now
>> 0.45 ± 15% +0.2 0.60 ± 12% perf-profile.self.cycles-pp.rcu_irq_enter
>> 0.31 ± 8% +0.2 0.46 ± 16% perf-profile.self.cycles-pp.irq_enter
>> 0.29 ± 10% +0.2 0.44 ± 16% perf-profile.self.cycles-pp.apic_timer_interrupt
>> 0.71 ± 30% +0.2 0.92 ± 8% perf-profile.self.cycles-pp.perf_mux_hrtimer_handler
>> 0.00 +0.3 0.28 ± 37% perf-profile.self.cycles-pp.memcpy_erms
>> 1.12 ± 3% +0.9 2.02 ± 15% perf-profile.self.cycles-pp.interrupt_entry
>> 0.79 ± 9% +0.9 1.73 ± 5% perf-profile.self.cycles-pp.perf_event_task_tick
>> 2.49 ± 45% +2.1 4.55 ± 20% perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
>> 10.95 ± 15% +2.7 13.61 ± 8% perf-profile.self.cycles-pp.mutex_spin_on_owner
>>
>>
>>
>> vm-scalability.throughput
>>
>> 1.6e+07 +-+---------------------------------------------------------------+
>> |..+.+ +..+.+..+.+. +. +..+.+..+.+..+.+..+.+..+ + |
>> 1.4e+07 +-+ : : O O O O |
>> 1.2e+07 O-+O O O O O O O O O O O O O O O O O O
>> | : : O O O O |
>> 1e+07 +-+ : : |
>> | : : |
>> 8e+06 +-+ : : |
>> | : : |
>> 6e+06 +-+ : : |
>> 4e+06 +-+ : : |
>> | :: |
>> 2e+06 +-+ : |
>> | : |
>> 0 +-+---------------------------------------------------------------+
>>
>>
>> vm-scalability.time.minor_page_faults
>>
>> 2.5e+06 +-+---------------------------------------------------------------+
>> | |
>> |..+.+ +..+.+..+.+..+.+..+.+.. .+. .+.+..+.+..+.+..+.+..+ |
>> 2e+06 +-+ : : +. +. |
>> O O O: O O O O O O O O O O |
>> | : : O O O O O O O O O O O O O O
>> 1.5e+06 +-+ : : |
>> | : : |
>> 1e+06 +-+ : : |
>> | : : |
>> | : : |
>> 500000 +-+ : : |
>> | : |
>> | : |
>> 0 +-+---------------------------------------------------------------+
>>
>>
>> vm-scalability.workload
>>
>> 3.5e+09 +-+---------------------------------------------------------------+
>> | .+. .+.+.. .+.. |
>> 3e+09 +-+ + +..+.+..+.+..+.+. +..+.+..+.+..+.+..+.+..+ + |
>> | : : O O O |
>> 2.5e+09 O-+O O: O O O O O O O O O |
>> | : : O O O O O O O O O O O O
>> 2e+09 +-+ : : |
>> | : : |
>> 1.5e+09 +-+ : : |
>> | : : |
>> 1e+09 +-+ : : |
>> | : : |
>> 5e+08 +-+ : |
>> | : |
>> 0 +-+---------------------------------------------------------------+
>>
>>
>> [*] bisect-good sample
>> [O] bisect-bad sample
>>
>>
>>
>> Disclaimer:
>> Results have been estimated based on internal Intel analysis and are provided
>> for informational purposes only. Any difference in system hardware or software
>> design or configuration may affect actual performance.
>>
>>
>> Thanks,
>> Rong Chen
>>
>
--
Thomas Zimmermann
Graphics Driver Developer
SUSE Linux GmbH, Maxfeldstrasse 5, 90409 Nuernberg, Germany
GF: Felix Imendörffer, Mary Higgins, Sri Rasiah
HRB 21284 (AG Nürnberg)
[-- Attachment #1.2: OpenPGP digital signature --]
[-- Type: application/pgp-signature, Size: 488 bytes --]
[-- Attachment #2: Type: text/plain, Size: 159 bytes --]
_______________________________________________
dri-devel mailing list
dri-devel@lists.freedesktop.org
https://lists.freedesktop.org/mailman/listinfo/dri-devel
WARNING: multiple messages have this Message-ID (diff)
From: Thomas Zimmermann <tzimmermann@suse.de>
To: lkp@lists.01.org
Subject: Re: [drm/mgag200] 90f479ae51: vm-scalability.median -18.8% regression
Date: Sun, 04 Aug 2019 20:39:19 +0200 [thread overview]
Message-ID: <14fdaaed-51c8-b270-b46b-cba7b5c4ba52@suse.de> (raw)
In-Reply-To: <1c0bf22b-2c69-6b45-f700-ed832a3a5c17@suse.de>
[-- Attachment #1: Type: text/plain, Size: 64563 bytes --]
Hi
I did some further analysis on this problem and found that the blinking
cursor affects performance of the vm-scalability test case.
I only have a 4-core machine, so scalability is not really testable. Yet
I see the effects of running vm-scalibility against drm-tip, a revert of
the mgag200 patch and the vmap fixes that I posted a few days ago.
After reverting the mgag200 patch, running the test as described in the
report
bin/lkp run job.yaml
gives results like
2019-08-02 19:34:37 ./case-anon-cow-seq-hugetlb
2019-08-02 19:34:37 ./usemem --runtime 300 -n 4 --prealloc --prefault
-O -U 815395225
917319627 bytes / 756534 usecs = 1184110 KB/s
917319627 bytes / 764675 usecs = 1171504 KB/s
917319627 bytes / 766414 usecs = 1168846 KB/s
917319627 bytes / 777990 usecs = 1151454 KB/s
Running the test against current drm-tip gives slightly worse results,
such as.
2019-08-03 19:17:06 ./case-anon-cow-seq-hugetlb
2019-08-03 19:17:06 ./usemem --runtime 300 -n 4 --prealloc --prefault
-O -U 815394406
917318700 bytes / 871607 usecs = 1027778 KB/s
917318700 bytes / 894173 usecs = 1001840 KB/s
917318700 bytes / 919694 usecs = 974040 KB/s
917318700 bytes / 923341 usecs = 970193 KB/s
The test puts out roughly one result per second. Strangely sending the
output to /dev/null can make results significantly worse.
bin/lkp run job.yaml > /dev/null
2019-08-03 19:23:04 ./case-anon-cow-seq-hugetlb
2019-08-03 19:23:04 ./usemem --runtime 300 -n 4 --prealloc --prefault
-O -U 815394406
917318700 bytes / 1207358 usecs = 741966 KB/s
917318700 bytes / 1210456 usecs = 740067 KB/s
917318700 bytes / 1216572 usecs = 736346 KB/s
917318700 bytes / 1239152 usecs = 722929 KB/s
I realized that there's still a blinking cursor on the screen, which I
disabled with
tput civis
or alternatively
echo 0 > /sys/devices/virtual/graphics/fbcon/cursor_blink
Running the the test now gives the original or even better results, such as
bin/lkp run job.yaml > /dev/null
2019-08-03 19:29:17 ./case-anon-cow-seq-hugetlb
2019-08-03 19:29:17 ./usemem --runtime 300 -n 4 --prealloc --prefault
-O -U 815394406
917318700 bytes / 659419 usecs = 1358497 KB/s
917318700 bytes / 659658 usecs = 1358005 KB/s
917318700 bytes / 659916 usecs = 1357474 KB/s
917318700 bytes / 660168 usecs = 1356956 KB/s
Rong, Feng, could you confirm this by disabling the cursor or blinking?
The difference between mgag200's original fbdev support and generic
fbdev emulation is generic fbdev's worker task that updates the VRAM
buffer from the shadow buffer. mgag200 does this immediately, but relies
on drm_can_sleep(), which is deprecated.
I think that the worker task interferes with the test case, as the
worker has been in fbdev emulation since forever and no performance
regressions have been reported so far.
So unless there's a report where this problem happens in a real-world
use case, I'd like to keep code as it is. And apparently there's always
the workaround of disabling the cursor blinking.
Best regards
Thomas
Am 30.07.19 um 19:50 schrieb Thomas Zimmermann:
> Am 29.07.19 um 11:51 schrieb kernel test robot:
>> Greeting,
>>
>> FYI, we noticed a -18.8% regression of vm-scalability.median due to commit:>
>>
>> commit: 90f479ae51afa45efab97afdde9b94b9660dd3e4 ("drm/mgag200: Replace struct mga_fbdev with generic framebuffer emulation")
>> https://kernel.googlesource.com/pub/scm/linux/kernel/git/next/linux-next.git master
>
> Daniel, Noralf, we may have to revert this patch.
>
> I expected some change in display performance, but not in VM. Since it's
> a server chipset, probably no one cares much about display performance.
> So that seemed like a good trade-off for re-using shared code.
>
> Part of the patch set is that the generic fb emulation now maps and
> unmaps the fbdev BO when updating the screen. I guess that's the cause
> of the performance regression. And it should be visible with other
> drivers as well if they use a shadow FB for fbdev emulation.
>
> The thing is that we'd need another generic fbdev emulation for ast and
> mgag200 that handles this issue properly.
>
> Best regards
> Thomas
>
>>
>> in testcase: vm-scalability
>> on test machine: 288 threads Intel(R) Xeon Phi(TM) CPU 7295 @ 1.50GHz with 80G memory
>> with following parameters:
>>
>> runtime: 300s
>> size: 8T
>> test: anon-cow-seq-hugetlb
>> cpufreq_governor: performance
>>
>> test-description: The motivation behind this suite is to exercise functions and regions of the mm/ of the Linux kernel which are of interest to us.
>> test-url: https://git.kernel.org/cgit/linux/kernel/git/wfg/vm-scalability.git/
>>
>>
>>
>> Details are as below:
>> -------------------------------------------------------------------------------------------------->
>>
>>
>> To reproduce:
>>
>> git clone https://github.com/intel/lkp-tests.git
>> cd lkp-tests
>> bin/lkp install job.yaml # job file is attached in this email
>> bin/lkp run job.yaml
>>
>> =========================================================================================
>> compiler/cpufreq_governor/kconfig/rootfs/runtime/size/tbox_group/test/testcase:
>> gcc-7/performance/x86_64-rhel-7.6/debian-x86_64-2019-05-14.cgz/300s/8T/lkp-knm01/anon-cow-seq-hugetlb/vm-scalability
>>
>> commit:
>> f1f8555dfb ("drm/bochs: Use shadow buffer for bochs framebuffer console")
>> 90f479ae51 ("drm/mgag200: Replace struct mga_fbdev with generic framebuffer emulation")
>>
>> f1f8555dfb9a70a2 90f479ae51afa45efab97afdde9
>> ---------------- ---------------------------
>> fail:runs %reproduction fail:runs
>> | | |
>> 2:4 -50% :4 dmesg.WARNING:at#for_ip_interrupt_entry/0x
>> :4 25% 1:4 dmesg.WARNING:at_ip___perf_sw_event/0x
>> :4 25% 1:4 dmesg.WARNING:at_ip__fsnotify_parent/0x
>> %stddev %change %stddev
>> \ | \
>> 43955 ± 2% -18.8% 35691 vm-scalability.median
>> 0.06 ± 7% +193.0% 0.16 ± 2% vm-scalability.median_stddev
>> 14906559 ± 2% -17.9% 12237079 vm-scalability.throughput
>> 87651 ± 2% -17.4% 72374 vm-scalability.time.involuntary_context_switches
>> 2086168 -23.6% 1594224 vm-scalability.time.minor_page_faults
>> 15082 ± 2% -10.4% 13517 vm-scalability.time.percent_of_cpu_this_job_got
>> 29987 -8.9% 27327 vm-scalability.time.system_time
>> 15755 -12.4% 13795 vm-scalability.time.user_time
>> 122011 -19.3% 98418 vm-scalability.time.voluntary_context_switches
>> 3.034e+09 -23.6% 2.318e+09 vm-scalability.workload
>> 242478 ± 12% +68.5% 408518 ± 23% cpuidle.POLL.time
>> 2788 ± 21% +117.4% 6062 ± 26% cpuidle.POLL.usage
>> 56653 ± 10% +64.4% 93144 ± 20% meminfo.Mapped
>> 120392 ± 7% +14.0% 137212 ± 4% meminfo.Shmem
>> 47221 ± 11% +77.1% 83634 ± 22% numa-meminfo.node0.Mapped
>> 120465 ± 7% +13.9% 137205 ± 4% numa-meminfo.node0.Shmem
>> 2885513 -16.5% 2409384 numa-numastat.node0.local_node
>> 2885471 -16.5% 2409354 numa-numastat.node0.numa_hit
>> 11813 ± 11% +76.3% 20824 ± 22% numa-vmstat.node0.nr_mapped
>> 30096 ± 7% +13.8% 34238 ± 4% numa-vmstat.node0.nr_shmem
>> 43.72 ± 2% +5.5 49.20 mpstat.cpu.all.idle%
>> 0.03 ± 4% +0.0 0.05 ± 6% mpstat.cpu.all.soft%
>> 19.51 -2.4 17.08 mpstat.cpu.all.usr%
>> 1012 -7.9% 932.75 turbostat.Avg_MHz
>> 32.38 ± 10% +25.8% 40.73 turbostat.CPU%c1
>> 145.51 -3.1% 141.01 turbostat.PkgWatt
>> 15.09 -19.2% 12.19 turbostat.RAMWatt
>> 43.50 ± 2% +13.2% 49.25 vmstat.cpu.id
>> 18.75 ± 2% -13.3% 16.25 ± 2% vmstat.cpu.us
>> 152.00 ± 2% -9.5% 137.50 vmstat.procs.r
>> 4800 -13.1% 4173 vmstat.system.cs
>> 156170 -11.9% 137594 slabinfo.anon_vma.active_objs
>> 3395 -11.9% 2991 slabinfo.anon_vma.active_slabs
>> 156190 -11.9% 137606 slabinfo.anon_vma.num_objs
>> 3395 -11.9% 2991 slabinfo.anon_vma.num_slabs
>> 1716 ± 5% +11.5% 1913 ± 8% slabinfo.dmaengine-unmap-16.active_objs
>> 1716 ± 5% +11.5% 1913 ± 8% slabinfo.dmaengine-unmap-16.num_objs
>> 1767 ± 2% -19.0% 1431 ± 2% slabinfo.hugetlbfs_inode_cache.active_objs
>> 1767 ± 2% -19.0% 1431 ± 2% slabinfo.hugetlbfs_inode_cache.num_objs
>> 3597 ± 5% -16.4% 3006 ± 3% slabinfo.skbuff_ext_cache.active_objs
>> 3597 ± 5% -16.4% 3006 ± 3% slabinfo.skbuff_ext_cache.num_objs
>> 1330122 -23.6% 1016557 proc-vmstat.htlb_buddy_alloc_success
>> 77214 ± 3% +6.4% 82128 ± 2% proc-vmstat.nr_active_anon
>> 67277 +2.9% 69246 proc-vmstat.nr_anon_pages
>> 218.50 ± 3% -10.6% 195.25 proc-vmstat.nr_dirtied
>> 288628 +1.4% 292755 proc-vmstat.nr_file_pages
>> 360.50 -2.7% 350.75 proc-vmstat.nr_inactive_file
>> 14225 ± 9% +63.8% 23304 ± 20% proc-vmstat.nr_mapped
>> 30109 ± 7% +13.8% 34259 ± 4% proc-vmstat.nr_shmem
>> 99870 -1.3% 98597 proc-vmstat.nr_slab_unreclaimable
>> 204.00 ± 4% -12.1% 179.25 proc-vmstat.nr_written
>> 77214 ± 3% +6.4% 82128 ± 2% proc-vmstat.nr_zone_active_anon
>> 360.50 -2.7% 350.75 proc-vmstat.nr_zone_inactive_file
>> 8810 ± 19% -66.1% 2987 ± 42% proc-vmstat.numa_hint_faults
>> 8810 ± 19% -66.1% 2987 ± 42% proc-vmstat.numa_hint_faults_local
>> 2904082 -16.4% 2427026 proc-vmstat.numa_hit
>> 2904081 -16.4% 2427025 proc-vmstat.numa_local
>> 6.828e+08 -23.5% 5.221e+08 proc-vmstat.pgalloc_normal
>> 2900008 -17.2% 2400195 proc-vmstat.pgfault
>> 6.827e+08 -23.5% 5.22e+08 proc-vmstat.pgfree
>> 1.635e+10 -17.0% 1.357e+10 perf-stat.i.branch-instructions
>> 1.53 ± 4% -0.1 1.45 ± 3% perf-stat.i.branch-miss-rate%
>> 2.581e+08 ± 3% -20.5% 2.051e+08 ± 2% perf-stat.i.branch-misses
>> 12.66 +1.1 13.78 perf-stat.i.cache-miss-rate%
>> 72720849 -12.0% 63958986 perf-stat.i.cache-misses
>> 5.766e+08 -18.6% 4.691e+08 perf-stat.i.cache-references
>> 4674 ± 2% -13.0% 4064 perf-stat.i.context-switches
>> 4.29 +12.5% 4.83 perf-stat.i.cpi
>> 2.573e+11 -7.4% 2.383e+11 perf-stat.i.cpu-cycles
>> 231.35 -21.5% 181.56 perf-stat.i.cpu-migrations
>> 3522 +4.4% 3677 perf-stat.i.cycles-between-cache-misses
>> 0.09 ± 13% +0.0 0.12 ± 5% perf-stat.i.iTLB-load-miss-rate%
>> 5.894e+10 -15.8% 4.961e+10 perf-stat.i.iTLB-loads
>> 5.901e+10 -15.8% 4.967e+10 perf-stat.i.instructions
>> 1291 ± 14% -21.8% 1010 perf-stat.i.instructions-per-iTLB-miss
>> 0.24 -11.0% 0.21 perf-stat.i.ipc
>> 9476 -17.5% 7821 perf-stat.i.minor-faults
>> 9478 -17.5% 7821 perf-stat.i.page-faults
>> 9.76 -3.6% 9.41 perf-stat.overall.MPKI
>> 1.59 ± 4% -0.1 1.52 perf-stat.overall.branch-miss-rate%
>> 12.61 +1.1 13.71 perf-stat.overall.cache-miss-rate%
>> 4.38 +10.5% 4.83 perf-stat.overall.cpi
>> 3557 +5.3% 3747 perf-stat.overall.cycles-between-cache-misses
>> 0.08 ± 12% +0.0 0.10 perf-stat.overall.iTLB-load-miss-rate%
>> 1268 ± 15% -23.0% 976.22 perf-stat.overall.instructions-per-iTLB-miss
>> 0.23 -9.5% 0.21 perf-stat.overall.ipc
>> 5815 +9.7% 6378 perf-stat.overall.path-length
>> 1.634e+10 -17.5% 1.348e+10 perf-stat.ps.branch-instructions
>> 2.595e+08 ± 3% -21.2% 2.043e+08 ± 2% perf-stat.ps.branch-misses
>> 72565205 -12.2% 63706339 perf-stat.ps.cache-misses
>> 5.754e+08 -19.2% 4.646e+08 perf-stat.ps.cache-references
>> 4640 ± 2% -12.5% 4060 perf-stat.ps.context-switches
>> 2.581e+11 -7.5% 2.387e+11 perf-stat.ps.cpu-cycles
>> 229.91 -22.0% 179.42 perf-stat.ps.cpu-migrations
>> 5.889e+10 -16.3% 4.927e+10 perf-stat.ps.iTLB-loads
>> 5.899e+10 -16.3% 4.938e+10 perf-stat.ps.instructions
>> 9388 -18.2% 7677 perf-stat.ps.minor-faults
>> 9389 -18.2% 7677 perf-stat.ps.page-faults
>> 1.764e+13 -16.2% 1.479e+13 perf-stat.total.instructions
>> 46803 ± 3% -18.8% 37982 ± 6% sched_debug.cfs_rq:/.exec_clock.min
>> 5320 ± 3% +23.7% 6581 ± 3% sched_debug.cfs_rq:/.exec_clock.stddev
>> 6737 ± 14% +58.1% 10649 ± 10% sched_debug.cfs_rq:/.load.avg
>> 587978 ± 17% +58.2% 930382 ± 9% sched_debug.cfs_rq:/.load.max
>> 46952 ± 16% +64.8% 77388 ± 11% sched_debug.cfs_rq:/.load.stddev
>> 7.12 ± 4% +49.1% 10.62 ± 6% sched_debug.cfs_rq:/.load_avg.avg
>> 474.40 ± 23% +67.5% 794.60 ± 10% sched_debug.cfs_rq:/.load_avg.max
>> 37.70 ± 11% +74.8% 65.90 ± 9% sched_debug.cfs_rq:/.load_avg.stddev
>> 13424269 ± 4% -15.6% 11328098 ± 2% sched_debug.cfs_rq:/.min_vruntime.avg
>> 15411275 ± 3% -12.4% 13505072 ± 2% sched_debug.cfs_rq:/.min_vruntime.max
>> 7939295 ± 6% -17.5% 6551322 ± 7% sched_debug.cfs_rq:/.min_vruntime.min
>> 21.44 ± 7% -56.1% 9.42 ± 4% sched_debug.cfs_rq:/.nr_spread_over.avg
>> 117.45 ± 11% -60.6% 46.30 ± 14% sched_debug.cfs_rq:/.nr_spread_over.max
>> 19.33 ± 8% -66.4% 6.49 ± 9% sched_debug.cfs_rq:/.nr_spread_over.stddev
>> 4.32 ± 15% +84.4% 7.97 ± 3% sched_debug.cfs_rq:/.runnable_load_avg.avg
>> 353.85 ± 29% +118.8% 774.35 ± 11% sched_debug.cfs_rq:/.runnable_load_avg.max
>> 27.30 ± 24% +118.5% 59.64 ± 9% sched_debug.cfs_rq:/.runnable_load_avg.stddev
>> 6729 ± 14% +58.2% 10644 ± 10% sched_debug.cfs_rq:/.runnable_weight.avg
>> 587978 ± 17% +58.2% 930382 ± 9% sched_debug.cfs_rq:/.runnable_weight.max
>> 46950 ± 16% +64.8% 77387 ± 11% sched_debug.cfs_rq:/.runnable_weight.stddev
>> 5305069 ± 4% -17.4% 4380376 ± 7% sched_debug.cfs_rq:/.spread0.avg
>> 7328745 ± 3% -9.9% 6600897 ± 3% sched_debug.cfs_rq:/.spread0.max
>> 2220837 ± 4% +55.8% 3460596 ± 5% sched_debug.cpu.avg_idle.avg
>> 4590666 ± 9% +76.8% 8117037 ± 15% sched_debug.cpu.avg_idle.max
>> 485052 ± 7% +80.3% 874679 ± 10% sched_debug.cpu.avg_idle.stddev
>> 561.50 ± 26% +37.7% 773.30 ± 15% sched_debug.cpu.clock.stddev
>> 561.50 ± 26% +37.7% 773.30 ± 15% sched_debug.cpu.clock_task.stddev
>> 3.20 ± 10% +109.6% 6.70 ± 3% sched_debug.cpu.cpu_load[0].avg
>> 309.10 ± 20% +150.3% 773.75 ± 12% sched_debug.cpu.cpu_load[0].max
>> 21.02 ± 14% +160.8% 54.80 ± 9% sched_debug.cpu.cpu_load[0].stddev
>> 3.19 ± 8% +109.8% 6.70 ± 3% sched_debug.cpu.cpu_load[1].avg
>> 299.75 ± 19% +158.0% 773.30 ± 12% sched_debug.cpu.cpu_load[1].max
>> 20.32 ± 12% +168.7% 54.62 ± 9% sched_debug.cpu.cpu_load[1].stddev
>> 3.20 ± 8% +109.1% 6.69 ± 4% sched_debug.cpu.cpu_load[2].avg
>> 288.90 ± 20% +167.0% 771.40 ± 12% sched_debug.cpu.cpu_load[2].max
>> 19.70 ± 12% +175.4% 54.27 ± 9% sched_debug.cpu.cpu_load[2].stddev
>> 3.16 ± 8% +110.9% 6.66 ± 6% sched_debug.cpu.cpu_load[3].avg
>> 275.50 ± 24% +178.4% 766.95 ± 12% sched_debug.cpu.cpu_load[3].max
>> 18.92 ± 15% +184.2% 53.77 ± 10% sched_debug.cpu.cpu_load[3].stddev
>> 3.08 ± 8% +115.7% 6.65 ± 7% sched_debug.cpu.cpu_load[4].avg
>> 263.55 ± 28% +188.7% 760.85 ± 12% sched_debug.cpu.cpu_load[4].max
>> 18.03 ± 18% +196.6% 53.46 ± 11% sched_debug.cpu.cpu_load[4].stddev
>> 14543 -9.6% 13150 sched_debug.cpu.curr->pid.max
>> 5293 ± 16% +74.7% 9248 ± 11% sched_debug.cpu.load.avg
>> 587978 ± 17% +58.2% 930382 ± 9% sched_debug.cpu.load.max
>> 40887 ± 19% +78.3% 72891 ± 9% sched_debug.cpu.load.stddev
>> 1141679 ± 4% +56.9% 1790907 ± 5% sched_debug.cpu.max_idle_balance_cost.avg
>> 2432100 ± 9% +72.6% 4196779 ± 13% sched_debug.cpu.max_idle_balance_cost.max
>> 745656 +29.3% 964170 ± 5% sched_debug.cpu.max_idle_balance_cost.min
>> 239032 ± 9% +81.9% 434806 ± 10% sched_debug.cpu.max_idle_balance_cost.stddev
>> 0.00 ± 27% +92.1% 0.00 ± 31% sched_debug.cpu.next_balance.stddev
>> 1030 ± 4% -10.4% 924.00 ± 2% sched_debug.cpu.nr_switches.min
>> 0.04 ± 26% +139.0% 0.09 ± 41% sched_debug.cpu.nr_uninterruptible.avg
>> 830.35 ± 6% -12.0% 730.50 ± 2% sched_debug.cpu.sched_count.min
>> 912.00 ± 2% -9.5% 825.38 sched_debug.cpu.ttwu_count.avg
>> 433.05 ± 3% -19.2% 350.05 ± 3% sched_debug.cpu.ttwu_count.min
>> 160.70 ± 3% -12.5% 140.60 ± 4% sched_debug.cpu.ttwu_local.min
>> 9072 ± 11% -36.4% 5767 ± 8% softirqs.CPU1.RCU
>> 12769 ± 5% +15.3% 14718 ± 3% softirqs.CPU101.SCHED
>> 13198 +11.5% 14717 ± 3% softirqs.CPU102.SCHED
>> 12981 ± 4% +13.9% 14788 ± 3% softirqs.CPU105.SCHED
>> 13486 ± 3% +11.8% 15071 ± 4% softirqs.CPU111.SCHED
>> 12794 ± 4% +14.1% 14601 ± 9% softirqs.CPU112.SCHED
>> 12999 ± 4% +10.1% 14314 ± 4% softirqs.CPU115.SCHED
>> 12844 ± 4% +10.6% 14202 ± 2% softirqs.CPU120.SCHED
>> 13336 ± 3% +9.4% 14585 ± 3% softirqs.CPU122.SCHED
>> 12639 ± 4% +20.2% 15195 softirqs.CPU123.SCHED
>> 13040 ± 5% +15.2% 15024 ± 5% softirqs.CPU126.SCHED
>> 13123 +15.1% 15106 ± 5% softirqs.CPU127.SCHED
>> 9188 ± 6% -35.7% 5911 ± 2% softirqs.CPU13.RCU
>> 13054 ± 3% +13.1% 14761 ± 5% softirqs.CPU130.SCHED
>> 13158 ± 2% +13.9% 14985 ± 5% softirqs.CPU131.SCHED
>> 12797 ± 6% +13.5% 14524 ± 3% softirqs.CPU133.SCHED
>> 12452 ± 5% +14.8% 14297 softirqs.CPU134.SCHED
>> 13078 ± 3% +10.4% 14439 ± 3% softirqs.CPU138.SCHED
>> 12617 ± 2% +14.5% 14442 ± 5% softirqs.CPU139.SCHED
>> 12974 ± 3% +13.7% 14752 ± 4% softirqs.CPU142.SCHED
>> 12579 ± 4% +19.1% 14983 ± 3% softirqs.CPU143.SCHED
>> 9122 ± 24% -44.6% 5053 ± 5% softirqs.CPU144.RCU
>> 13366 ± 2% +11.1% 14848 ± 3% softirqs.CPU149.SCHED
>> 13246 ± 2% +22.0% 16162 ± 7% softirqs.CPU150.SCHED
>> 13452 ± 3% +20.5% 16210 ± 7% softirqs.CPU151.SCHED
>> 13507 +10.1% 14869 softirqs.CPU156.SCHED
>> 13808 ± 3% +9.2% 15079 ± 4% softirqs.CPU157.SCHED
>> 13442 ± 2% +13.4% 15248 ± 4% softirqs.CPU160.SCHED
>> 13311 +12.1% 14920 ± 2% softirqs.CPU162.SCHED
>> 13544 ± 3% +8.5% 14695 ± 4% softirqs.CPU163.SCHED
>> 13648 ± 3% +11.2% 15179 ± 2% softirqs.CPU166.SCHED
>> 13404 ± 4% +12.5% 15079 ± 3% softirqs.CPU168.SCHED
>> 13421 ± 6% +16.0% 15568 ± 8% softirqs.CPU169.SCHED
>> 13115 ± 3% +23.1% 16139 ± 10% softirqs.CPU171.SCHED
>> 13424 ± 6% +10.4% 14822 ± 3% softirqs.CPU175.SCHED
>> 13274 ± 3% +13.7% 15087 ± 9% softirqs.CPU185.SCHED
>> 13409 ± 3% +12.3% 15063 ± 3% softirqs.CPU190.SCHED
>> 13181 ± 7% +13.4% 14946 ± 3% softirqs.CPU196.SCHED
>> 13578 ± 3% +10.9% 15061 softirqs.CPU197.SCHED
>> 13323 ± 5% +24.8% 16627 ± 6% softirqs.CPU198.SCHED
>> 14072 ± 2% +12.3% 15798 ± 7% softirqs.CPU199.SCHED
>> 12604 ± 13% +17.9% 14865 softirqs.CPU201.SCHED
>> 13380 ± 4% +14.8% 15356 ± 3% softirqs.CPU203.SCHED
>> 13481 ± 8% +14.2% 15390 ± 3% softirqs.CPU204.SCHED
>> 12921 ± 2% +13.8% 14710 ± 3% softirqs.CPU206.SCHED
>> 13468 +13.0% 15218 ± 2% softirqs.CPU208.SCHED
>> 13253 ± 2% +13.1% 14992 softirqs.CPU209.SCHED
>> 13319 ± 2% +14.3% 15225 ± 7% softirqs.CPU210.SCHED
>> 13673 ± 5% +16.3% 15895 ± 3% softirqs.CPU211.SCHED
>> 13290 +17.0% 15556 ± 5% softirqs.CPU212.SCHED
>> 13455 ± 4% +14.4% 15392 ± 3% softirqs.CPU213.SCHED
>> 13454 ± 4% +14.3% 15377 ± 3% softirqs.CPU215.SCHED
>> 13872 ± 7% +9.7% 15221 ± 5% softirqs.CPU220.SCHED
>> 13555 ± 4% +17.3% 15896 ± 5% softirqs.CPU222.SCHED
>> 13411 ± 4% +20.8% 16197 ± 6% softirqs.CPU223.SCHED
>> 8472 ± 21% -44.8% 4680 ± 3% softirqs.CPU224.RCU
>> 13141 ± 3% +16.2% 15265 ± 7% softirqs.CPU225.SCHED
>> 14084 ± 3% +8.2% 15242 ± 2% softirqs.CPU226.SCHED
>> 13528 ± 4% +11.3% 15063 ± 4% softirqs.CPU228.SCHED
>> 13218 ± 3% +16.3% 15377 ± 4% softirqs.CPU229.SCHED
>> 14031 ± 4% +10.2% 15467 ± 2% softirqs.CPU231.SCHED
>> 13770 ± 3% +14.0% 15700 ± 3% softirqs.CPU232.SCHED
>> 13456 ± 3% +12.3% 15105 ± 3% softirqs.CPU233.SCHED
>> 13137 ± 4% +13.5% 14909 ± 3% softirqs.CPU234.SCHED
>> 13318 ± 2% +14.7% 15280 ± 2% softirqs.CPU235.SCHED
>> 13690 ± 2% +13.7% 15563 ± 7% softirqs.CPU238.SCHED
>> 13771 ± 5% +20.8% 16634 ± 7% softirqs.CPU241.SCHED
>> 13317 ± 7% +19.5% 15919 ± 9% softirqs.CPU243.SCHED
>> 8234 ± 16% -43.9% 4616 ± 5% softirqs.CPU244.RCU
>> 13845 ± 6% +13.0% 15643 ± 3% softirqs.CPU244.SCHED
>> 13179 ± 3% +16.3% 15323 softirqs.CPU246.SCHED
>> 13754 +12.2% 15438 ± 3% softirqs.CPU248.SCHED
>> 13769 ± 4% +10.9% 15276 ± 2% softirqs.CPU252.SCHED
>> 13702 +10.5% 15147 ± 2% softirqs.CPU254.SCHED
>> 13315 ± 2% +12.5% 14980 ± 3% softirqs.CPU255.SCHED
>> 13785 ± 3% +12.9% 15568 ± 5% softirqs.CPU256.SCHED
>> 13307 ± 3% +15.0% 15298 ± 3% softirqs.CPU257.SCHED
>> 13864 ± 3% +10.5% 15313 ± 2% softirqs.CPU259.SCHED
>> 13879 ± 2% +11.4% 15465 softirqs.CPU261.SCHED
>> 13815 +13.6% 15687 ± 5% softirqs.CPU264.SCHED
>> 119574 ± 2% +11.8% 133693 ± 11% softirqs.CPU266.TIMER
>> 13688 +10.9% 15180 ± 6% softirqs.CPU267.SCHED
>> 11716 ± 4% +19.3% 13974 ± 8% softirqs.CPU27.SCHED
>> 13866 ± 3% +13.7% 15765 ± 4% softirqs.CPU271.SCHED
>> 13887 ± 5% +12.5% 15621 softirqs.CPU272.SCHED
>> 13383 ± 3% +19.8% 16031 ± 2% softirqs.CPU274.SCHED
>> 13347 +14.1% 15232 ± 3% softirqs.CPU275.SCHED
>> 12884 ± 2% +21.0% 15593 ± 4% softirqs.CPU276.SCHED
>> 13131 ± 5% +13.4% 14891 ± 5% softirqs.CPU277.SCHED
>> 12891 ± 2% +19.2% 15371 ± 4% softirqs.CPU278.SCHED
>> 13313 ± 4% +13.0% 15049 ± 2% softirqs.CPU279.SCHED
>> 13514 ± 3% +10.2% 14897 ± 2% softirqs.CPU280.SCHED
>> 13501 ± 3% +13.7% 15346 softirqs.CPU281.SCHED
>> 13261 +17.5% 15577 softirqs.CPU282.SCHED
>> 8076 ± 15% -43.7% 4546 ± 5% softirqs.CPU283.RCU
>> 13686 ± 3% +12.6% 15413 ± 2% softirqs.CPU284.SCHED
>> 13439 ± 2% +9.2% 14670 ± 4% softirqs.CPU285.SCHED
>> 8878 ± 9% -35.4% 5735 ± 4% softirqs.CPU35.RCU
>> 11690 ± 2% +13.6% 13274 ± 5% softirqs.CPU40.SCHED
>> 11714 ± 2% +19.3% 13975 ± 13% softirqs.CPU41.SCHED
>> 11763 +12.5% 13239 ± 4% softirqs.CPU45.SCHED
>> 11662 ± 2% +9.4% 12757 ± 3% softirqs.CPU46.SCHED
>> 11805 ± 2% +9.3% 12902 ± 2% softirqs.CPU50.SCHED
>> 12158 ± 3% +12.3% 13655 ± 8% softirqs.CPU55.SCHED
>> 11716 ± 4% +8.8% 12751 ± 3% softirqs.CPU58.SCHED
>> 11922 ± 2% +9.9% 13100 ± 4% softirqs.CPU64.SCHED
>> 9674 ± 17% -41.8% 5625 ± 6% softirqs.CPU66.RCU
>> 11818 +12.0% 13237 softirqs.CPU66.SCHED
>> 124682 ± 7% -6.1% 117088 ± 5% softirqs.CPU66.TIMER
>> 8637 ± 9% -34.0% 5700 ± 7% softirqs.CPU70.RCU
>> 11624 ± 2% +11.0% 12901 ± 2% softirqs.CPU70.SCHED
>> 12372 ± 2% +13.2% 14003 ± 3% softirqs.CPU71.SCHED
>> 9949 ± 25% -33.9% 6574 ± 31% softirqs.CPU72.RCU
>> 10392 ± 26% -35.1% 6745 ± 35% softirqs.CPU73.RCU
>> 12766 ± 3% +11.1% 14188 ± 3% softirqs.CPU76.SCHED
>> 12611 ± 2% +18.8% 14984 ± 5% softirqs.CPU78.SCHED
>> 12786 ± 3% +17.9% 15079 ± 7% softirqs.CPU79.SCHED
>> 11947 ± 4% +9.7% 13103 ± 4% softirqs.CPU8.SCHED
>> 13379 ± 7% +11.8% 14962 ± 4% softirqs.CPU83.SCHED
>> 13438 ± 5% +9.7% 14738 ± 2% softirqs.CPU84.SCHED
>> 12768 +19.4% 15241 ± 6% softirqs.CPU88.SCHED
>> 8604 ± 13% -39.3% 5222 ± 3% softirqs.CPU89.RCU
>> 13077 ± 2% +17.1% 15308 ± 7% softirqs.CPU89.SCHED
>> 11887 ± 3% +20.1% 14272 ± 5% softirqs.CPU9.SCHED
>> 12723 ± 3% +11.3% 14165 ± 4% softirqs.CPU90.SCHED
>> 8439 ± 12% -38.9% 5153 ± 4% softirqs.CPU91.RCU
>> 13429 ± 3% +10.3% 14806 ± 2% softirqs.CPU95.SCHED
>> 12852 ± 4% +10.3% 14174 ± 5% softirqs.CPU96.SCHED
>> 13010 ± 2% +14.4% 14888 ± 5% softirqs.CPU97.SCHED
>> 2315644 ± 4% -36.2% 1477200 ± 4% softirqs.RCU
>> 1572 ± 10% +63.9% 2578 ± 39% interrupts.CPU0.NMI:Non-maskable_interrupts
>> 1572 ± 10% +63.9% 2578 ± 39% interrupts.CPU0.PMI:Performance_monitoring_interrupts
>> 252.00 ± 11% -35.2% 163.25 ± 13% interrupts.CPU104.RES:Rescheduling_interrupts
>> 2738 ± 24% +52.4% 4173 ± 19% interrupts.CPU105.NMI:Non-maskable_interrupts
>> 2738 ± 24% +52.4% 4173 ± 19% interrupts.CPU105.PMI:Performance_monitoring_interrupts
>> 245.75 ± 19% -31.0% 169.50 ± 7% interrupts.CPU105.RES:Rescheduling_interrupts
>> 228.75 ± 13% -24.7% 172.25 ± 19% interrupts.CPU106.RES:Rescheduling_interrupts
>> 2243 ± 15% +66.3% 3730 ± 35% interrupts.CPU113.NMI:Non-maskable_interrupts
>> 2243 ± 15% +66.3% 3730 ± 35% interrupts.CPU113.PMI:Performance_monitoring_interrupts
>> 2703 ± 31% +67.0% 4514 ± 33% interrupts.CPU118.NMI:Non-maskable_interrupts
>> 2703 ± 31% +67.0% 4514 ± 33% interrupts.CPU118.PMI:Performance_monitoring_interrupts
>> 2613 ± 25% +42.2% 3715 ± 24% interrupts.CPU121.NMI:Non-maskable_interrupts
>> 2613 ± 25% +42.2% 3715 ± 24% interrupts.CPU121.PMI:Performance_monitoring_interrupts
>> 311.50 ± 23% -47.7% 163.00 ± 9% interrupts.CPU122.RES:Rescheduling_interrupts
>> 266.75 ± 19% -31.6% 182.50 ± 15% interrupts.CPU124.RES:Rescheduling_interrupts
>> 293.75 ± 33% -32.3% 198.75 ± 19% interrupts.CPU125.RES:Rescheduling_interrupts
>> 2601 ± 36% +43.2% 3724 ± 29% interrupts.CPU127.NMI:Non-maskable_interrupts
>> 2601 ± 36% +43.2% 3724 ± 29% interrupts.CPU127.PMI:Performance_monitoring_interrupts
>> 2258 ± 21% +68.2% 3797 ± 29% interrupts.CPU13.NMI:Non-maskable_interrupts
>> 2258 ± 21% +68.2% 3797 ± 29% interrupts.CPU13.PMI:Performance_monitoring_interrupts
>> 3338 ± 29% +54.6% 5160 ± 9% interrupts.CPU139.NMI:Non-maskable_interrupts
>> 3338 ± 29% +54.6% 5160 ± 9% interrupts.CPU139.PMI:Performance_monitoring_interrupts
>> 219.50 ± 27% -23.0% 169.00 ± 21% interrupts.CPU139.RES:Rescheduling_interrupts
>> 290.25 ± 25% -32.5% 196.00 ± 11% interrupts.CPU14.RES:Rescheduling_interrupts
>> 243.50 ± 4% -16.0% 204.50 ± 12% interrupts.CPU140.RES:Rescheduling_interrupts
>> 1797 ± 15% +135.0% 4223 ± 46% interrupts.CPU147.NMI:Non-maskable_interrupts
>> 1797 ± 15% +135.0% 4223 ± 46% interrupts.CPU147.PMI:Performance_monitoring_interrupts
>> 2537 ± 22% +89.6% 4812 ± 28% interrupts.CPU15.NMI:Non-maskable_interrupts
>> 2537 ± 22% +89.6% 4812 ± 28% interrupts.CPU15.PMI:Performance_monitoring_interrupts
>> 292.25 ± 34% -33.9% 193.25 ± 6% interrupts.CPU15.RES:Rescheduling_interrupts
>> 424.25 ± 37% -58.5% 176.25 ± 14% interrupts.CPU158.RES:Rescheduling_interrupts
>> 312.50 ± 42% -54.2% 143.00 ± 18% interrupts.CPU159.RES:Rescheduling_interrupts
>> 725.00 ±118% -75.7% 176.25 ± 14% interrupts.CPU163.RES:Rescheduling_interrupts
>> 2367 ± 6% +59.9% 3786 ± 24% interrupts.CPU177.NMI:Non-maskable_interrupts
>> 2367 ± 6% +59.9% 3786 ± 24% interrupts.CPU177.PMI:Performance_monitoring_interrupts
>> 239.50 ± 30% -46.6% 128.00 ± 14% interrupts.CPU179.RES:Rescheduling_interrupts
>> 320.75 ± 15% -24.0% 243.75 ± 20% interrupts.CPU20.RES:Rescheduling_interrupts
>> 302.50 ± 17% -47.2% 159.75 ± 8% interrupts.CPU200.RES:Rescheduling_interrupts
>> 2166 ± 5% +92.0% 4157 ± 40% interrupts.CPU207.NMI:Non-maskable_interrupts
>> 2166 ± 5% +92.0% 4157 ± 40% interrupts.CPU207.PMI:Performance_monitoring_interrupts
>> 217.00 ± 11% -34.6% 142.00 ± 12% interrupts.CPU214.RES:Rescheduling_interrupts
>> 2610 ± 36% +47.4% 3848 ± 35% interrupts.CPU215.NMI:Non-maskable_interrupts
>> 2610 ± 36% +47.4% 3848 ± 35% interrupts.CPU215.PMI:Performance_monitoring_interrupts
>> 2046 ± 13% +118.6% 4475 ± 43% interrupts.CPU22.NMI:Non-maskable_interrupts
>> 2046 ± 13% +118.6% 4475 ± 43% interrupts.CPU22.PMI:Performance_monitoring_interrupts
>> 289.50 ± 28% -41.1% 170.50 ± 8% interrupts.CPU22.RES:Rescheduling_interrupts
>> 2232 ± 6% +33.0% 2970 ± 24% interrupts.CPU221.NMI:Non-maskable_interrupts
>> 2232 ± 6% +33.0% 2970 ± 24% interrupts.CPU221.PMI:Performance_monitoring_interrupts
>> 4552 ± 12% -27.6% 3295 ± 15% interrupts.CPU222.NMI:Non-maskable_interrupts
>> 4552 ± 12% -27.6% 3295 ± 15% interrupts.CPU222.PMI:Performance_monitoring_interrupts
>> 2013 ± 15% +80.9% 3641 ± 27% interrupts.CPU226.NMI:Non-maskable_interrupts
>> 2013 ± 15% +80.9% 3641 ± 27% interrupts.CPU226.PMI:Performance_monitoring_interrupts
>> 2575 ± 49% +67.1% 4302 ± 34% interrupts.CPU227.NMI:Non-maskable_interrupts
>> 2575 ± 49% +67.1% 4302 ± 34% interrupts.CPU227.PMI:Performance_monitoring_interrupts
>> 248.00 ± 36% -36.3% 158.00 ± 19% interrupts.CPU228.RES:Rescheduling_interrupts
>> 2441 ± 24% +43.0% 3490 ± 30% interrupts.CPU23.NMI:Non-maskable_interrupts
>> 2441 ± 24% +43.0% 3490 ± 30% interrupts.CPU23.PMI:Performance_monitoring_interrupts
>> 404.25 ± 69% -65.5% 139.50 ± 17% interrupts.CPU236.RES:Rescheduling_interrupts
>> 566.50 ± 40% -73.6% 149.50 ± 31% interrupts.CPU237.RES:Rescheduling_interrupts
>> 243.50 ± 26% -37.1% 153.25 ± 21% interrupts.CPU248.RES:Rescheduling_interrupts
>> 258.25 ± 12% -53.5% 120.00 ± 18% interrupts.CPU249.RES:Rescheduling_interrupts
>> 2888 ± 27% +49.4% 4313 ± 30% interrupts.CPU253.NMI:Non-maskable_interrupts
>> 2888 ± 27% +49.4% 4313 ± 30% interrupts.CPU253.PMI:Performance_monitoring_interrupts
>> 2468 ± 44% +67.3% 4131 ± 37% interrupts.CPU256.NMI:Non-maskable_interrupts
>> 2468 ± 44% +67.3% 4131 ± 37% interrupts.CPU256.PMI:Performance_monitoring_interrupts
>> 425.00 ± 59% -60.3% 168.75 ± 34% interrupts.CPU258.RES:Rescheduling_interrupts
>> 1859 ± 16% +106.3% 3834 ± 44% interrupts.CPU268.NMI:Non-maskable_interrupts
>> 1859 ± 16% +106.3% 3834 ± 44% interrupts.CPU268.PMI:Performance_monitoring_interrupts
>> 2684 ± 28% +61.2% 4326 ± 36% interrupts.CPU269.NMI:Non-maskable_interrupts
>> 2684 ± 28% +61.2% 4326 ± 36% interrupts.CPU269.PMI:Performance_monitoring_interrupts
>> 2171 ± 6% +108.8% 4533 ± 20% interrupts.CPU270.NMI:Non-maskable_interrupts
>> 2171 ± 6% +108.8% 4533 ± 20% interrupts.CPU270.PMI:Performance_monitoring_interrupts
>> 2262 ± 14% +61.8% 3659 ± 37% interrupts.CPU273.NMI:Non-maskable_interrupts
>> 2262 ± 14% +61.8% 3659 ± 37% interrupts.CPU273.PMI:Performance_monitoring_interrupts
>> 2203 ± 11% +50.7% 3320 ± 38% interrupts.CPU279.NMI:Non-maskable_interrupts
>> 2203 ± 11% +50.7% 3320 ± 38% interrupts.CPU279.PMI:Performance_monitoring_interrupts
>> 2433 ± 17% +52.9% 3721 ± 25% interrupts.CPU280.NMI:Non-maskable_interrupts
>> 2433 ± 17% +52.9% 3721 ± 25% interrupts.CPU280.PMI:Performance_monitoring_interrupts
>> 2778 ± 33% +63.1% 4531 ± 36% interrupts.CPU283.NMI:Non-maskable_interrupts
>> 2778 ± 33% +63.1% 4531 ± 36% interrupts.CPU283.PMI:Performance_monitoring_interrupts
>> 331.75 ± 32% -39.8% 199.75 ± 17% interrupts.CPU29.RES:Rescheduling_interrupts
>> 2178 ± 22% +53.9% 3353 ± 31% interrupts.CPU3.NMI:Non-maskable_interrupts
>> 2178 ± 22% +53.9% 3353 ± 31% interrupts.CPU3.PMI:Performance_monitoring_interrupts
>> 298.50 ± 30% -39.7% 180.00 ± 6% interrupts.CPU34.RES:Rescheduling_interrupts
>> 2490 ± 3% +58.7% 3953 ± 28% interrupts.CPU35.NMI:Non-maskable_interrupts
>> 2490 ± 3% +58.7% 3953 ± 28% interrupts.CPU35.PMI:Performance_monitoring_interrupts
>> 270.50 ± 24% -31.1% 186.25 ± 3% interrupts.CPU36.RES:Rescheduling_interrupts
>> 2493 ± 7% +57.0% 3915 ± 27% interrupts.CPU43.NMI:Non-maskable_interrupts
>> 2493 ± 7% +57.0% 3915 ± 27% interrupts.CPU43.PMI:Performance_monitoring_interrupts
>> 286.75 ± 36% -32.4% 193.75 ± 7% interrupts.CPU45.RES:Rescheduling_interrupts
>> 259.00 ± 12% -23.6% 197.75 ± 13% interrupts.CPU46.RES:Rescheduling_interrupts
>> 244.00 ± 21% -35.6% 157.25 ± 11% interrupts.CPU47.RES:Rescheduling_interrupts
>> 230.00 ± 7% -21.3% 181.00 ± 11% interrupts.CPU48.RES:Rescheduling_interrupts
>> 281.00 ± 13% -27.4% 204.00 ± 15% interrupts.CPU53.RES:Rescheduling_interrupts
>> 256.75 ± 5% -18.4% 209.50 ± 12% interrupts.CPU54.RES:Rescheduling_interrupts
>> 2433 ± 9% +68.4% 4098 ± 35% interrupts.CPU58.NMI:Non-maskable_interrupts
>> 2433 ± 9% +68.4% 4098 ± 35% interrupts.CPU58.PMI:Performance_monitoring_interrupts
>> 316.00 ± 25% -41.4% 185.25 ± 13% interrupts.CPU59.RES:Rescheduling_interrupts
>> 2703 ± 38% +56.0% 4217 ± 31% interrupts.CPU60.NMI:Non-maskable_interrupts
>> 2703 ± 38% +56.0% 4217 ± 31% interrupts.CPU60.PMI:Performance_monitoring_interrupts
>> 2425 ± 16% +39.9% 3394 ± 27% interrupts.CPU61.NMI:Non-maskable_interrupts
>> 2425 ± 16% +39.9% 3394 ± 27% interrupts.CPU61.PMI:Performance_monitoring_interrupts
>> 2388 ± 18% +69.5% 4047 ± 29% interrupts.CPU66.NMI:Non-maskable_interrupts
>> 2388 ± 18% +69.5% 4047 ± 29% interrupts.CPU66.PMI:Performance_monitoring_interrupts
>> 2322 ± 11% +93.4% 4491 ± 35% interrupts.CPU67.NMI:Non-maskable_interrupts
>> 2322 ± 11% +93.4% 4491 ± 35% interrupts.CPU67.PMI:Performance_monitoring_interrupts
>> 319.00 ± 40% -44.7% 176.25 ± 9% interrupts.CPU67.RES:Rescheduling_interrupts
>> 2512 ± 8% +28.1% 3219 ± 25% interrupts.CPU70.NMI:Non-maskable_interrupts
>> 2512 ± 8% +28.1% 3219 ± 25% interrupts.CPU70.PMI:Performance_monitoring_interrupts
>> 2290 ± 39% +78.7% 4094 ± 28% interrupts.CPU74.NMI:Non-maskable_interrupts
>> 2290 ± 39% +78.7% 4094 ± 28% interrupts.CPU74.PMI:Performance_monitoring_interrupts
>> 2446 ± 40% +94.8% 4764 ± 23% interrupts.CPU75.NMI:Non-maskable_interrupts
>> 2446 ± 40% +94.8% 4764 ± 23% interrupts.CPU75.PMI:Performance_monitoring_interrupts
>> 426.75 ± 61% -67.7% 138.00 ± 8% interrupts.CPU75.RES:Rescheduling_interrupts
>> 192.50 ± 13% +45.6% 280.25 ± 45% interrupts.CPU76.RES:Rescheduling_interrupts
>> 274.25 ± 34% -42.2% 158.50 ± 34% interrupts.CPU77.RES:Rescheduling_interrupts
>> 2357 ± 9% +73.0% 4078 ± 23% interrupts.CPU78.NMI:Non-maskable_interrupts
>> 2357 ± 9% +73.0% 4078 ± 23% interrupts.CPU78.PMI:Performance_monitoring_interrupts
>> 348.50 ± 53% -47.3% 183.75 ± 29% interrupts.CPU80.RES:Rescheduling_interrupts
>> 2650 ± 43% +46.2% 3874 ± 36% interrupts.CPU84.NMI:Non-maskable_interrupts
>> 2650 ± 43% +46.2% 3874 ± 36% interrupts.CPU84.PMI:Performance_monitoring_interrupts
>> 2235 ± 10% +117.8% 4867 ± 10% interrupts.CPU90.NMI:Non-maskable_interrupts
>> 2235 ± 10% +117.8% 4867 ± 10% interrupts.CPU90.PMI:Performance_monitoring_interrupts
>> 2606 ± 33% +38.1% 3598 ± 21% interrupts.CPU92.NMI:Non-maskable_interrupts
>> 2606 ± 33% +38.1% 3598 ± 21% interrupts.CPU92.PMI:Performance_monitoring_interrupts
>> 408.75 ± 58% -56.8% 176.75 ± 25% interrupts.CPU92.RES:Rescheduling_interrupts
>> 399.00 ± 64% -63.6% 145.25 ± 16% interrupts.CPU93.RES:Rescheduling_interrupts
>> 314.75 ± 36% -44.2% 175.75 ± 13% interrupts.CPU94.RES:Rescheduling_interrupts
>> 191.00 ± 15% -29.1% 135.50 ± 9% interrupts.CPU97.RES:Rescheduling_interrupts
>> 94.00 ± 8% +50.0% 141.00 ± 12% interrupts.IWI:IRQ_work_interrupts
>> 841457 ± 7% +16.6% 980751 ± 3% interrupts.NMI:Non-maskable_interrupts
>> 841457 ± 7% +16.6% 980751 ± 3% interrupts.PMI:Performance_monitoring_interrupts
>> 12.75 ± 11% -4.1 8.67 ± 31% perf-profile.calltrace.cycles-pp.do_rw_once
>> 1.02 ± 16% -0.6 0.47 ± 59% perf-profile.calltrace.cycles-pp.sched_clock.sched_clock_cpu.cpuidle_enter_state.cpuidle_enter.do_idle
>> 1.10 ± 15% -0.4 0.66 ± 14% perf-profile.calltrace.cycles-pp.sched_clock_cpu.cpuidle_enter_state.cpuidle_enter.do_idle.cpu_startup_entry
>> 1.05 ± 16% -0.4 0.61 ± 14% perf-profile.calltrace.cycles-pp.native_sched_clock.sched_clock.sched_clock_cpu.cpuidle_enter_state.cpuidle_enter
>> 1.58 ± 4% +0.3 1.91 ± 7% perf-profile.calltrace.cycles-pp.__hrtimer_run_queues.hrtimer_interrupt.smp_apic_timer_interrupt.apic_timer_interrupt.copy_page
>> 0.79 ± 26% +0.5 1.27 ± 18% perf-profile.calltrace.cycles-pp.__x64_sys_exit_group.do_syscall_64.entry_SYSCALL_64_after_hwframe
>> 0.79 ± 26% +0.5 1.27 ± 18% perf-profile.calltrace.cycles-pp.do_group_exit.__x64_sys_exit_group.do_syscall_64.entry_SYSCALL_64_after_hwframe
>> 0.79 ± 26% +0.5 1.27 ± 18% perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64.entry_SYSCALL_64_after_hwframe
>> 2.11 ± 4% +0.5 2.60 ± 7% perf-profile.calltrace.cycles-pp.apic_timer_interrupt.osq_lock.__mutex_lock.hugetlb_fault.handle_mm_fault
>> 0.83 ± 26% +0.5 1.32 ± 18% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
>> 0.83 ± 26% +0.5 1.32 ± 18% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
>> 1.90 ± 5% +0.6 2.45 ± 7% perf-profile.calltrace.cycles-pp.hrtimer_interrupt.smp_apic_timer_interrupt.apic_timer_interrupt.copy_page.copy_subpage
>> 0.65 ± 62% +0.6 1.20 ± 15% perf-profile.calltrace.cycles-pp.alloc_fresh_huge_page.alloc_surplus_huge_page.alloc_huge_page.hugetlb_cow.hugetlb_fault
>> 0.60 ± 62% +0.6 1.16 ± 18% perf-profile.calltrace.cycles-pp.free_huge_page.release_pages.tlb_flush_mmu.tlb_finish_mmu.exit_mmap
>> 0.95 ± 17% +0.6 1.52 ± 8% perf-profile.calltrace.cycles-pp.__hrtimer_run_queues.hrtimer_interrupt.smp_apic_timer_interrupt.apic_timer_interrupt.mutex_spin_on_owner
>> 0.61 ± 62% +0.6 1.18 ± 18% perf-profile.calltrace.cycles-pp.release_pages.tlb_flush_mmu.tlb_finish_mmu.exit_mmap.mmput
>> 0.61 ± 62% +0.6 1.19 ± 19% perf-profile.calltrace.cycles-pp.tlb_finish_mmu.exit_mmap.mmput.do_exit.do_group_exit
>> 0.61 ± 62% +0.6 1.19 ± 19% perf-profile.calltrace.cycles-pp.tlb_flush_mmu.tlb_finish_mmu.exit_mmap.mmput.do_exit
>> 0.64 ± 61% +0.6 1.23 ± 18% perf-profile.calltrace.cycles-pp.mmput.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
>> 0.64 ± 61% +0.6 1.23 ± 18% perf-profile.calltrace.cycles-pp.exit_mmap.mmput.do_exit.do_group_exit.__x64_sys_exit_group
>> 1.30 ± 9% +0.6 1.92 ± 8% perf-profile.calltrace.cycles-pp.hrtimer_interrupt.smp_apic_timer_interrupt.apic_timer_interrupt.mutex_spin_on_owner.__mutex_lock
>> 0.19 ±173% +0.7 0.89 ± 20% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.free_huge_page.release_pages.tlb_flush_mmu
>> 0.19 ±173% +0.7 0.90 ± 20% perf-profile.calltrace.cycles-pp._raw_spin_lock.free_huge_page.release_pages.tlb_flush_mmu.tlb_finish_mmu
>> 0.00 +0.8 0.77 ± 30% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.prep_new_huge_page.alloc_fresh_huge_page.alloc_surplus_huge_page
>> 0.00 +0.8 0.78 ± 30% perf-profile.calltrace.cycles-pp._raw_spin_lock.prep_new_huge_page.alloc_fresh_huge_page.alloc_surplus_huge_page.alloc_huge_page
>> 0.00 +0.8 0.79 ± 29% perf-profile.calltrace.cycles-pp.prep_new_huge_page.alloc_fresh_huge_page.alloc_surplus_huge_page.alloc_huge_page.hugetlb_cow
>> 0.82 ± 67% +0.9 1.72 ± 22% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.alloc_huge_page.hugetlb_cow.hugetlb_fault
>> 0.84 ± 66% +0.9 1.74 ± 20% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.alloc_surplus_huge_page.alloc_huge_page.hugetlb_cow
>> 2.52 ± 6% +0.9 3.44 ± 9% perf-profile.calltrace.cycles-pp.smp_apic_timer_interrupt.apic_timer_interrupt.copy_page.copy_subpage.copy_user_huge_page
>> 0.83 ± 67% +0.9 1.75 ± 21% perf-profile.calltrace.cycles-pp._raw_spin_lock.alloc_huge_page.hugetlb_cow.hugetlb_fault.handle_mm_fault
>> 0.84 ± 66% +0.9 1.77 ± 20% perf-profile.calltrace.cycles-pp._raw_spin_lock.alloc_surplus_huge_page.alloc_huge_page.hugetlb_cow.hugetlb_fault
>> 1.64 ± 12% +1.0 2.67 ± 7% perf-profile.calltrace.cycles-pp.smp_apic_timer_interrupt.apic_timer_interrupt.mutex_spin_on_owner.__mutex_lock.hugetlb_fault
>> 1.65 ± 45% +1.3 2.99 ± 18% perf-profile.calltrace.cycles-pp.alloc_surplus_huge_page.alloc_huge_page.hugetlb_cow.hugetlb_fault.handle_mm_fault
>> 1.74 ± 13% +1.4 3.16 ± 6% perf-profile.calltrace.cycles-pp.apic_timer_interrupt.mutex_spin_on_owner.__mutex_lock.hugetlb_fault.handle_mm_fault
>> 2.56 ± 48% +2.2 4.81 ± 19% perf-profile.calltrace.cycles-pp.alloc_huge_page.hugetlb_cow.hugetlb_fault.handle_mm_fault.__do_page_fault
>> 12.64 ± 14% +3.6 16.20 ± 8% perf-profile.calltrace.cycles-pp.mutex_spin_on_owner.__mutex_lock.hugetlb_fault.handle_mm_fault.__do_page_fault
>> 2.97 ± 7% +3.8 6.74 ± 9% perf-profile.calltrace.cycles-pp.apic_timer_interrupt.copy_page.copy_subpage.copy_user_huge_page.hugetlb_cow
>> 19.99 ± 9% +4.1 24.05 ± 6% perf-profile.calltrace.cycles-pp.hugetlb_cow.hugetlb_fault.handle_mm_fault.__do_page_fault.do_page_fault
>> 1.37 ± 15% -0.5 0.83 ± 13% perf-profile.children.cycles-pp.sched_clock_cpu
>> 1.31 ± 16% -0.5 0.78 ± 13% perf-profile.children.cycles-pp.sched_clock
>> 1.29 ± 16% -0.5 0.77 ± 13% perf-profile.children.cycles-pp.native_sched_clock
>> 1.80 ± 2% -0.3 1.47 ± 10% perf-profile.children.cycles-pp.task_tick_fair
>> 0.73 ± 2% -0.2 0.54 ± 11% perf-profile.children.cycles-pp.update_curr
>> 0.42 ± 17% -0.2 0.27 ± 16% perf-profile.children.cycles-pp.account_process_tick
>> 0.73 ± 10% -0.2 0.58 ± 9% perf-profile.children.cycles-pp.rcu_sched_clock_irq
>> 0.27 ± 6% -0.1 0.14 ± 14% perf-profile.children.cycles-pp.__acct_update_integrals
>> 0.27 ± 18% -0.1 0.16 ± 13% perf-profile.children.cycles-pp.rcu_segcblist_ready_cbs
>> 0.40 ± 12% -0.1 0.30 ± 14% perf-profile.children.cycles-pp.__next_timer_interrupt
>> 0.47 ± 7% -0.1 0.39 ± 13% perf-profile.children.cycles-pp.update_rq_clock
>> 0.29 ± 12% -0.1 0.21 ± 15% perf-profile.children.cycles-pp.cpuidle_governor_latency_req
>> 0.21 ± 7% -0.1 0.14 ± 12% perf-profile.children.cycles-pp.account_system_index_time
>> 0.38 ± 2% -0.1 0.31 ± 12% perf-profile.children.cycles-pp.timerqueue_add
>> 0.26 ± 11% -0.1 0.20 ± 13% perf-profile.children.cycles-pp.find_next_bit
>> 0.23 ± 15% -0.1 0.17 ± 15% perf-profile.children.cycles-pp.rcu_dynticks_eqs_exit
>> 0.14 ± 8% -0.1 0.07 ± 14% perf-profile.children.cycles-pp.account_user_time
>> 0.17 ± 6% -0.0 0.12 ± 10% perf-profile.children.cycles-pp.cpuacct_charge
>> 0.18 ± 20% -0.0 0.13 ± 3% perf-profile.children.cycles-pp.irq_work_tick
>> 0.11 ± 13% -0.0 0.07 ± 25% perf-profile.children.cycles-pp.tick_sched_do_timer
>> 0.12 ± 10% -0.0 0.08 ± 15% perf-profile.children.cycles-pp.get_cpu_device
>> 0.07 ± 11% -0.0 0.04 ± 58% perf-profile.children.cycles-pp.raise_softirq
>> 0.12 ± 3% -0.0 0.09 ± 8% perf-profile.children.cycles-pp.write
>> 0.11 ± 13% +0.0 0.14 ± 8% perf-profile.children.cycles-pp.native_write_msr
>> 0.09 ± 9% +0.0 0.11 ± 7% perf-profile.children.cycles-pp.finish_task_switch
>> 0.10 ± 10% +0.0 0.13 ± 5% perf-profile.children.cycles-pp.schedule_idle
>> 0.07 ± 6% +0.0 0.10 ± 12% perf-profile.children.cycles-pp.__read_nocancel
>> 0.04 ± 58% +0.0 0.07 ± 15% perf-profile.children.cycles-pp.__free_pages_ok
>> 0.06 ± 7% +0.0 0.09 ± 13% perf-profile.children.cycles-pp.perf_read
>> 0.07 +0.0 0.11 ± 14% perf-profile.children.cycles-pp.perf_evsel__read_counter
>> 0.07 +0.0 0.11 ± 13% perf-profile.children.cycles-pp.cmd_stat
>> 0.07 +0.0 0.11 ± 13% perf-profile.children.cycles-pp.__run_perf_stat
>> 0.07 +0.0 0.11 ± 13% perf-profile.children.cycles-pp.process_interval
>> 0.07 +0.0 0.11 ± 13% perf-profile.children.cycles-pp.read_counters
>> 0.07 ± 22% +0.0 0.11 ± 19% perf-profile.children.cycles-pp.__handle_mm_fault
>> 0.07 ± 19% +0.1 0.13 ± 8% perf-profile.children.cycles-pp.rb_erase
>> 0.03 ±100% +0.1 0.09 ± 9% perf-profile.children.cycles-pp.smp_call_function_single
>> 0.01 ±173% +0.1 0.08 ± 11% perf-profile.children.cycles-pp.perf_event_read
>> 0.00 +0.1 0.07 ± 13% perf-profile.children.cycles-pp.__perf_event_read_value
>> 0.00 +0.1 0.07 ± 7% perf-profile.children.cycles-pp.__intel_pmu_enable_all
>> 0.08 ± 17% +0.1 0.15 ± 8% perf-profile.children.cycles-pp.native_apic_msr_eoi_write
>> 0.04 ±103% +0.1 0.13 ± 58% perf-profile.children.cycles-pp.shmem_getpage_gfp
>> 0.38 ± 14% +0.1 0.51 ± 6% perf-profile.children.cycles-pp.run_timer_softirq
>> 0.11 ± 4% +0.3 0.37 ± 32% perf-profile.children.cycles-pp.worker_thread
>> 0.20 ± 5% +0.3 0.48 ± 25% perf-profile.children.cycles-pp.ret_from_fork
>> 0.20 ± 4% +0.3 0.48 ± 25% perf-profile.children.cycles-pp.kthread
>> 0.00 +0.3 0.29 ± 38% perf-profile.children.cycles-pp.memcpy_erms
>> 0.00 +0.3 0.29 ± 38% perf-profile.children.cycles-pp.drm_fb_helper_dirty_work
>> 0.00 +0.3 0.31 ± 37% perf-profile.children.cycles-pp.process_one_work
>> 0.47 ± 48% +0.4 0.91 ± 19% perf-profile.children.cycles-pp.prep_new_huge_page
>> 0.70 ± 29% +0.5 1.16 ± 18% perf-profile.children.cycles-pp.free_huge_page
>> 0.73 ± 29% +0.5 1.19 ± 18% perf-profile.children.cycles-pp.tlb_flush_mmu
>> 0.72 ± 29% +0.5 1.18 ± 18% perf-profile.children.cycles-pp.release_pages
>> 0.73 ± 29% +0.5 1.19 ± 18% perf-profile.children.cycles-pp.tlb_finish_mmu
>> 0.76 ± 27% +0.5 1.23 ± 18% perf-profile.children.cycles-pp.exit_mmap
>> 0.77 ± 27% +0.5 1.24 ± 18% perf-profile.children.cycles-pp.mmput
>> 0.79 ± 26% +0.5 1.27 ± 18% perf-profile.children.cycles-pp.__x64_sys_exit_group
>> 0.79 ± 26% +0.5 1.27 ± 18% perf-profile.children.cycles-pp.do_group_exit
>> 0.79 ± 26% +0.5 1.27 ± 18% perf-profile.children.cycles-pp.do_exit
>> 1.28 ± 29% +0.5 1.76 ± 9% perf-profile.children.cycles-pp.perf_mux_hrtimer_handler
>> 0.77 ± 28% +0.5 1.26 ± 13% perf-profile.children.cycles-pp.alloc_fresh_huge_page
>> 1.53 ± 15% +0.7 2.26 ± 14% perf-profile.children.cycles-pp.do_syscall_64
>> 1.53 ± 15% +0.7 2.27 ± 14% perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
>> 1.13 ± 3% +0.9 2.07 ± 14% perf-profile.children.cycles-pp.interrupt_entry
>> 0.79 ± 9% +1.0 1.76 ± 5% perf-profile.children.cycles-pp.perf_event_task_tick
>> 1.71 ± 39% +1.4 3.08 ± 16% perf-profile.children.cycles-pp.alloc_surplus_huge_page
>> 2.66 ± 42% +2.3 4.94 ± 17% perf-profile.children.cycles-pp.alloc_huge_page
>> 2.89 ± 45% +2.7 5.54 ± 18% perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
>> 3.34 ± 35% +2.7 6.02 ± 17% perf-profile.children.cycles-pp._raw_spin_lock
>> 12.77 ± 14% +3.9 16.63 ± 7% perf-profile.children.cycles-pp.mutex_spin_on_owner
>> 20.12 ± 9% +4.0 24.16 ± 6% perf-profile.children.cycles-pp.hugetlb_cow
>> 15.40 ± 10% -3.6 11.84 ± 28% perf-profile.self.cycles-pp.do_rw_once
>> 4.02 ± 9% -1.3 2.73 ± 30% perf-profile.self.cycles-pp.do_access
>> 2.00 ± 14% -0.6 1.41 ± 13% perf-profile.self.cycles-pp.cpuidle_enter_state
>> 1.26 ± 16% -0.5 0.74 ± 13% perf-profile.self.cycles-pp.native_sched_clock
>> 0.42 ± 17% -0.2 0.27 ± 16% perf-profile.self.cycles-pp.account_process_tick
>> 0.27 ± 19% -0.2 0.12 ± 17% perf-profile.self.cycles-pp.timerqueue_del
>> 0.53 ± 3% -0.1 0.38 ± 11% perf-profile.self.cycles-pp.update_curr
>> 0.27 ± 6% -0.1 0.14 ± 14% perf-profile.self.cycles-pp.__acct_update_integrals
>> 0.27 ± 18% -0.1 0.16 ± 13% perf-profile.self.cycles-pp.rcu_segcblist_ready_cbs
>> 0.61 ± 4% -0.1 0.51 ± 8% perf-profile.self.cycles-pp.task_tick_fair
>> 0.20 ± 8% -0.1 0.12 ± 14% perf-profile.self.cycles-pp.account_system_index_time
>> 0.23 ± 15% -0.1 0.16 ± 17% perf-profile.self.cycles-pp.rcu_dynticks_eqs_exit
>> 0.25 ± 11% -0.1 0.18 ± 14% perf-profile.self.cycles-pp.find_next_bit
>> 0.10 ± 11% -0.1 0.03 ±100% perf-profile.self.cycles-pp.tick_sched_do_timer
>> 0.29 -0.1 0.23 ± 11% perf-profile.self.cycles-pp.timerqueue_add
>> 0.12 ± 10% -0.1 0.06 ± 17% perf-profile.self.cycles-pp.account_user_time
>> 0.22 ± 15% -0.1 0.16 ± 6% perf-profile.self.cycles-pp.scheduler_tick
>> 0.17 ± 6% -0.0 0.12 ± 10% perf-profile.self.cycles-pp.cpuacct_charge
>> 0.18 ± 20% -0.0 0.13 ± 3% perf-profile.self.cycles-pp.irq_work_tick
>> 0.07 ± 13% -0.0 0.03 ±100% perf-profile.self.cycles-pp.update_process_times
>> 0.12 ± 7% -0.0 0.08 ± 15% perf-profile.self.cycles-pp.get_cpu_device
>> 0.07 ± 11% -0.0 0.04 ± 58% perf-profile.self.cycles-pp.raise_softirq
>> 0.12 ± 11% -0.0 0.09 ± 7% perf-profile.self.cycles-pp.tick_nohz_get_sleep_length
>> 0.11 ± 11% +0.0 0.14 ± 6% perf-profile.self.cycles-pp.native_write_msr
>> 0.10 ± 5% +0.1 0.15 ± 8% perf-profile.self.cycles-pp.__remove_hrtimer
>> 0.07 ± 23% +0.1 0.13 ± 8% perf-profile.self.cycles-pp.rb_erase
>> 0.08 ± 17% +0.1 0.15 ± 7% perf-profile.self.cycles-pp.native_apic_msr_eoi_write
>> 0.00 +0.1 0.08 ± 10% perf-profile.self.cycles-pp.smp_call_function_single
>> 0.32 ± 17% +0.1 0.42 ± 7% perf-profile.self.cycles-pp.run_timer_softirq
>> 0.22 ± 5% +0.1 0.34 ± 4% perf-profile.self.cycles-pp.ktime_get_update_offsets_now
>> 0.45 ± 15% +0.2 0.60 ± 12% perf-profile.self.cycles-pp.rcu_irq_enter
>> 0.31 ± 8% +0.2 0.46 ± 16% perf-profile.self.cycles-pp.irq_enter
>> 0.29 ± 10% +0.2 0.44 ± 16% perf-profile.self.cycles-pp.apic_timer_interrupt
>> 0.71 ± 30% +0.2 0.92 ± 8% perf-profile.self.cycles-pp.perf_mux_hrtimer_handler
>> 0.00 +0.3 0.28 ± 37% perf-profile.self.cycles-pp.memcpy_erms
>> 1.12 ± 3% +0.9 2.02 ± 15% perf-profile.self.cycles-pp.interrupt_entry
>> 0.79 ± 9% +0.9 1.73 ± 5% perf-profile.self.cycles-pp.perf_event_task_tick
>> 2.49 ± 45% +2.1 4.55 ± 20% perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
>> 10.95 ± 15% +2.7 13.61 ± 8% perf-profile.self.cycles-pp.mutex_spin_on_owner
>>
>>
>>
>> vm-scalability.throughput
>>
>> 1.6e+07 +-+---------------------------------------------------------------+
>> |..+.+ +..+.+..+.+. +. +..+.+..+.+..+.+..+.+..+ + |
>> 1.4e+07 +-+ : : O O O O |
>> 1.2e+07 O-+O O O O O O O O O O O O O O O O O O
>> | : : O O O O |
>> 1e+07 +-+ : : |
>> | : : |
>> 8e+06 +-+ : : |
>> | : : |
>> 6e+06 +-+ : : |
>> 4e+06 +-+ : : |
>> | :: |
>> 2e+06 +-+ : |
>> | : |
>> 0 +-+---------------------------------------------------------------+
>>
>>
>> vm-scalability.time.minor_page_faults
>>
>> 2.5e+06 +-+---------------------------------------------------------------+
>> | |
>> |..+.+ +..+.+..+.+..+.+..+.+.. .+. .+.+..+.+..+.+..+.+..+ |
>> 2e+06 +-+ : : +. +. |
>> O O O: O O O O O O O O O O |
>> | : : O O O O O O O O O O O O O O
>> 1.5e+06 +-+ : : |
>> | : : |
>> 1e+06 +-+ : : |
>> | : : |
>> | : : |
>> 500000 +-+ : : |
>> | : |
>> | : |
>> 0 +-+---------------------------------------------------------------+
>>
>>
>> vm-scalability.workload
>>
>> 3.5e+09 +-+---------------------------------------------------------------+
>> | .+. .+.+.. .+.. |
>> 3e+09 +-+ + +..+.+..+.+..+.+. +..+.+..+.+..+.+..+.+..+ + |
>> | : : O O O |
>> 2.5e+09 O-+O O: O O O O O O O O O |
>> | : : O O O O O O O O O O O O
>> 2e+09 +-+ : : |
>> | : : |
>> 1.5e+09 +-+ : : |
>> | : : |
>> 1e+09 +-+ : : |
>> | : : |
>> 5e+08 +-+ : |
>> | : |
>> 0 +-+---------------------------------------------------------------+
>>
>>
>> [*] bisect-good sample
>> [O] bisect-bad sample
>>
>>
>>
>> Disclaimer:
>> Results have been estimated based on internal Intel analysis and are provided
>> for informational purposes only. Any difference in system hardware or software
>> design or configuration may affect actual performance.
>>
>>
>> Thanks,
>> Rong Chen
>>
>
--
Thomas Zimmermann
Graphics Driver Developer
SUSE Linux GmbH, Maxfeldstrasse 5, 90409 Nuernberg, Germany
GF: Felix Imendörffer, Mary Higgins, Sri Rasiah
HRB 21284 (AG Nürnberg)
[-- Attachment #2: signature.asc --]
[-- Type: application/pgp-signature, Size: 488 bytes --]
next prev parent reply other threads:[~2019-08-04 18:39 UTC|newest]
Thread overview: 132+ messages / expand[flat|nested] mbox.gz Atom feed top
2019-07-29 9:51 [drm/mgag200] 90f479ae51: vm-scalability.median -18.8% regression kernel test robot
2019-07-29 9:51 ` kernel test robot
2019-07-30 17:50 ` Thomas Zimmermann
2019-07-30 17:50 ` Thomas Zimmermann
2019-07-30 18:12 ` Daniel Vetter
2019-07-30 18:12 ` Daniel Vetter
2019-07-30 18:50 ` Thomas Zimmermann
2019-07-30 18:50 ` Thomas Zimmermann
2019-07-30 18:59 ` Daniel Vetter
2019-07-30 18:59 ` Daniel Vetter
2019-07-30 20:26 ` Dave Airlie
2019-07-30 20:26 ` Dave Airlie
2019-07-31 8:13 ` Daniel Vetter
2019-07-31 8:13 ` Daniel Vetter
2019-07-31 9:25 ` [LKP] " Huang, Ying
2019-07-31 9:25 ` Huang, Ying
2019-07-31 10:12 ` [LKP] " Thomas Zimmermann
2019-07-31 10:12 ` Thomas Zimmermann
2019-07-31 10:21 ` [LKP] " Michel Dänzer
2019-08-01 6:19 ` Rong Chen
2019-08-01 6:19 ` Rong Chen
2019-08-01 8:37 ` [LKP] " Feng Tang
2019-08-01 8:37 ` Feng Tang
2019-08-01 9:59 ` [LKP] " Thomas Zimmermann
2019-08-01 9:59 ` Thomas Zimmermann
2019-08-01 11:25 ` [LKP] " Feng Tang
2019-08-01 11:25 ` Feng Tang
2019-08-01 11:58 ` [LKP] " Thomas Zimmermann
2019-08-01 11:58 ` Thomas Zimmermann
2019-08-02 7:11 ` [LKP] " Rong Chen
2019-08-02 7:11 ` Rong Chen
2019-08-02 8:23 ` [LKP] " Thomas Zimmermann
2019-08-02 8:23 ` Thomas Zimmermann
2019-08-02 9:20 ` [LKP] " Thomas Zimmermann
2019-08-02 9:20 ` Thomas Zimmermann
2019-08-01 9:57 ` [LKP] " Thomas Zimmermann
2019-08-01 9:57 ` Thomas Zimmermann
2019-08-01 13:30 ` [LKP] " Michel Dänzer
2019-08-02 8:17 ` Thomas Zimmermann
2019-08-02 8:17 ` Thomas Zimmermann
2019-07-31 10:10 ` Thomas Zimmermann
2019-07-31 10:10 ` Thomas Zimmermann
2019-08-02 9:11 ` Daniel Vetter
2019-08-02 9:11 ` Daniel Vetter
2019-08-02 9:26 ` Thomas Zimmermann
2019-08-02 9:26 ` Thomas Zimmermann
2019-08-04 18:39 ` Thomas Zimmermann [this message]
2019-08-04 18:39 ` Thomas Zimmermann
2019-08-05 7:02 ` Feng Tang
2019-08-05 7:02 ` Feng Tang
2019-08-05 7:28 ` Rong Chen
2019-08-05 10:25 ` Thomas Zimmermann
2019-08-05 10:25 ` Thomas Zimmermann
2019-08-06 12:59 ` [LKP] " Chen, Rong A
2019-08-06 12:59 ` Chen, Rong A
2019-08-07 10:42 ` [LKP] " Thomas Zimmermann
2019-08-07 10:42 ` Thomas Zimmermann
2019-08-09 8:12 ` [LKP] " Rong Chen
2019-08-09 8:12 ` Rong Chen
2019-08-12 7:25 ` [LKP] " Feng Tang
2019-08-12 7:25 ` Feng Tang
2019-08-13 9:36 ` [LKP] " Feng Tang
2019-08-13 9:36 ` Feng Tang
2019-08-13 9:36 ` [LKP] " Feng Tang
2019-08-16 6:55 ` Feng Tang
2019-08-16 6:55 ` Feng Tang
2019-08-22 17:25 ` [LKP] " Thomas Zimmermann
2019-08-22 17:25 ` Thomas Zimmermann
2019-08-22 17:25 ` [LKP] " Thomas Zimmermann
2019-08-22 20:02 ` Dave Airlie
2019-08-22 20:02 ` Dave Airlie
2019-08-23 9:54 ` [LKP] " Thomas Zimmermann
2019-08-23 9:54 ` Thomas Zimmermann
2019-08-23 9:54 ` [LKP] " Thomas Zimmermann
2019-08-24 5:16 ` Feng Tang
2019-08-24 5:16 ` Feng Tang
2019-08-24 5:16 ` [LKP] " Feng Tang
2019-08-26 10:50 ` Thomas Zimmermann
2019-08-26 10:50 ` Thomas Zimmermann
2019-08-27 12:33 ` [LKP] " Chen, Rong A
2019-08-27 12:33 ` Chen, Rong A
2019-08-27 12:33 ` [LKP] " Chen, Rong A
2019-08-27 17:16 ` Thomas Zimmermann
2019-08-27 17:16 ` Thomas Zimmermann
2019-08-28 9:37 ` [LKP] " Rong Chen
2019-08-28 9:37 ` Rong Chen
2019-08-28 10:51 ` [LKP] " Thomas Zimmermann
2019-08-28 10:51 ` Thomas Zimmermann
2019-09-04 6:27 ` [LKP] " Feng Tang
2019-09-04 6:27 ` Feng Tang
2019-09-04 6:53 ` [LKP] " Thomas Zimmermann
2019-09-04 6:53 ` Thomas Zimmermann
2019-09-04 8:11 ` [LKP] " Daniel Vetter
2019-09-04 8:11 ` Daniel Vetter
2019-09-04 8:35 ` [LKP] " Feng Tang
2019-09-04 8:35 ` Feng Tang
2019-09-04 8:43 ` [LKP] " Thomas Zimmermann
2019-09-04 8:43 ` Thomas Zimmermann
2019-09-04 14:30 ` [LKP] " Chen, Rong A
2019-09-04 14:30 ` Chen, Rong A
2019-09-04 9:17 ` [LKP] " Daniel Vetter
2019-09-04 9:17 ` Daniel Vetter
2019-09-04 11:15 ` [LKP] " Dave Airlie
2019-09-04 11:15 ` Dave Airlie
2019-09-04 11:20 ` [LKP] " Daniel Vetter
2019-09-04 11:20 ` Daniel Vetter
2019-09-04 11:20 ` [LKP] " Daniel Vetter
2019-09-05 6:59 ` Feng Tang
2019-09-05 6:59 ` Feng Tang
2019-09-05 10:37 ` [LKP] " Daniel Vetter
2019-09-05 10:37 ` Daniel Vetter
2019-09-05 10:48 ` [LKP] " Feng Tang
2019-09-05 10:48 ` Feng Tang
2019-09-05 10:48 ` [LKP] " Feng Tang
2019-09-09 14:12 ` Thomas Zimmermann
2019-09-09 14:12 ` Thomas Zimmermann
2019-09-09 14:12 ` [LKP] " Thomas Zimmermann
2019-09-16 9:06 ` Feng Tang
2019-09-16 9:06 ` Feng Tang
2019-09-17 8:48 ` [LKP] " Thomas Zimmermann
2019-09-17 8:48 ` Thomas Zimmermann
2019-09-17 8:48 ` [LKP] " Thomas Zimmermann
2019-08-05 10:22 ` Thomas Zimmermann
2019-08-05 10:22 ` Thomas Zimmermann
2019-08-05 12:52 ` Feng Tang
2019-08-05 12:52 ` Feng Tang
2020-01-06 13:19 ` Thomas Zimmermann
2020-01-06 13:19 ` Thomas Zimmermann
2020-01-08 2:25 ` Rong Chen
2020-01-08 2:28 ` Rong Chen
2020-01-08 5:20 ` Thomas Zimmermann
2020-01-08 5:20 ` Thomas Zimmermann
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=14fdaaed-51c8-b270-b46b-cba7b5c4ba52@suse.de \
--to=tzimmermann@suse.de \
--cc=daniel@ffwll.ch \
--cc=dri-devel@lists.freedesktop.org \
--cc=feng.tang@intel.com \
--cc=lkp@01.org \
--cc=michel@daenzer.net \
--cc=noralf@tronnes.org \
--cc=rong.a.chen@intel.com \
--cc=sfr@canb.auug.org.au \
--cc=ying.huang@intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.