From: kernel test robot <oliver.sang@intel.com>
To: "Matthew Wilcox (Oracle)" <willy@infradead.org>
Cc: <oe-lkp@lists.linux.dev>, <lkp@intel.com>, <linux-mm@kvack.org>,
<ying.huang@intel.com>, <feng.tang@intel.com>,
<fengwei.yin@intel.com>,
Andrew Morton <akpm@linux-foundation.org>,
"Matthew Wilcox (Oracle)" <willy@infradead.org>,
Suren Baghdasaryan <surenb@google.com>, <oliver.sang@intel.com>
Subject: Re: [PATCH v2 3/6] mm: Handle shared faults under the VMA lock
Date: Fri, 20 Oct 2023 21:23:42 +0800 [thread overview]
Message-ID: <202310201857.d7db939a-oliver.sang@intel.com> (raw)
In-Reply-To: <20231006195318.4087158-4-willy@infradead.org>
Hello,
kernel test robot noticed a 67.5% improvement of stress-ng.fault.minor_page_faults_per_sec on:
commit: c8b329d48e0dac7438168a1857c3f67d4e23fed0 ("[PATCH v2 3/6] mm: Handle shared faults under the VMA lock")
url: https://github.com/intel-lab-lkp/linux/commits/Matthew-Wilcox-Oracle/mm-Make-lock_folio_maybe_drop_mmap-VMA-lock-aware/20231007-035513
base: v6.6-rc4
patch link: https://lore.kernel.org/all/20231006195318.4087158-4-willy@infradead.org/
patch subject: [PATCH v2 3/6] mm: Handle shared faults under the VMA lock
testcase: stress-ng
test machine: 36 threads 1 sockets Intel(R) Core(TM) i9-10980XE CPU @ 3.00GHz (Cascade Lake) with 128G memory
parameters:
nr_threads: 1
disk: 1HDD
testtime: 60s
fs: ext4
class: os
test: fault
cpufreq_governor: performance
In addition to that, the commit also has significant impact on the following tests:
+------------------+----------------------------------------------------------------------------------------------------+
| testcase: change | will-it-scale: will-it-scale.per_thread_ops 274.8% improvement |
| test machine | 224 threads 4 sockets Intel(R) Xeon(R) Platinum 8380H CPU @ 2.90GHz (Cooper Lake) with 192G memory |
| test parameters | cpufreq_governor=performance |
| | mode=thread |
| | nr_task=50% |
| | test=page_fault3 |
+------------------+----------------------------------------------------------------------------------------------------+
Details are as below:
-------------------------------------------------------------------------------------------------->
The kernel config and materials to reproduce are available at:
https://download.01.org/0day-ci/archive/20231020/202310201857.d7db939a-oliver.sang@intel.com
=========================================================================================
class/compiler/cpufreq_governor/disk/fs/kconfig/nr_threads/rootfs/tbox_group/test/testcase/testtime:
os/gcc-12/performance/1HDD/ext4/x86_64-rhel-8.3/1/debian-11.1-x86_64-20220510.cgz/lkp-csl-d02/fault/stress-ng/60s
commit:
34611600bf ("mm: Call wp_page_copy() under the VMA lock")
c8b329d48e ("mm: Handle shared faults under the VMA lock")
34611600bfd1bf9f c8b329d48e0dac7438168a1857c
---------------- ---------------------------
%stddev %change %stddev
\ | \
157941 ± 6% +20.3% 190026 ± 11% meminfo.DirectMap4k
0.05 +0.0 0.05 perf-stat.i.dTLB-store-miss-rate%
51205 -100.0% 0.03 ± 81% perf-stat.i.major-faults
79003 +65.6% 130837 perf-stat.i.minor-faults
50394 -100.0% 0.03 ± 81% perf-stat.ps.major-faults
77754 +65.6% 128767 perf-stat.ps.minor-faults
53411 -100.0% 0.00 ±223% stress-ng.fault.major_page_faults_per_sec
80118 +67.5% 134204 stress-ng.fault.minor_page_faults_per_sec
1417 -4.7% 1350 stress-ng.fault.nanosecs_per_page_fault
3204300 -100.0% 0.33 ±141% stress-ng.time.major_page_faults
4815857 +67.3% 8059294 stress-ng.time.minor_page_faults
0.01 ± 68% +224.2% 0.03 ± 51% perf-sched.sch_delay.avg.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
0.55 ± 95% +368.6% 2.56 ± 35% perf-sched.sch_delay.max.ms.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
0.05 ± 70% +168.2% 0.12 ± 32% perf-sched.wait_time.avg.ms.__cond_resched.__ext4_handle_dirty_metadata.ext4_mb_clear_bb.ext4_remove_blocks.ext4_ext_rm_leaf
0.05 ± 73% +114.3% 0.10 ± 13% perf-sched.wait_time.avg.ms.__cond_resched.zap_pmd_range.isra.0.unmap_page_range
0.09 ± 78% +79.2% 0.17 ± 8% perf-sched.wait_time.max.ms.__cond_resched.__ext4_handle_dirty_metadata.ext4_delete_entry.__ext4_unlink.ext4_unlink
0.05 ± 70% +229.6% 0.15 ± 21% perf-sched.wait_time.max.ms.__cond_resched.__ext4_handle_dirty_metadata.ext4_mb_clear_bb.ext4_remove_blocks.ext4_ext_rm_leaf
0.03 ±151% +260.5% 0.12 ± 35% perf-sched.wait_time.max.ms.__cond_resched.down_write_killable.vm_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.03 ±100% +183.8% 0.10 ± 35% perf-sched.wait_time.max.ms.__cond_resched.ext4_journal_check_start.__ext4_journal_start_sb.ext4_alloc_file_blocks.isra
0.08 ± 79% +134.1% 0.18 ± 36% perf-sched.wait_time.max.ms.__cond_resched.zap_pmd_range.isra.0.unmap_page_range
11.65 -0.8 10.82 ± 2% perf-profile.calltrace.cycles-pp.stress_fault
9.42 -0.8 8.61 ± 2% perf-profile.calltrace.cycles-pp.asm_exc_page_fault.stress_fault
8.84 -0.8 8.07 ± 3% perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.stress_fault
8.74 -0.7 8.00 ± 3% perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.stress_fault
7.56 ± 2% -0.5 7.04 ± 3% perf-profile.calltrace.cycles-pp.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.stress_fault
6.99 ± 2% -0.5 6.51 ± 3% perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
11.10 -0.9 10.24 ± 2% perf-profile.children.cycles-pp.asm_exc_page_fault
12.38 -0.8 11.54 perf-profile.children.cycles-pp.stress_fault
8.92 -0.8 8.14 ± 3% perf-profile.children.cycles-pp.exc_page_fault
8.84 -0.8 8.07 ± 3% perf-profile.children.cycles-pp.do_user_addr_fault
7.63 ± 2% -0.5 7.09 ± 2% perf-profile.children.cycles-pp.handle_mm_fault
7.06 ± 2% -0.5 6.56 ± 3% perf-profile.children.cycles-pp.__handle_mm_fault
0.36 ± 8% -0.2 0.19 ± 8% perf-profile.children.cycles-pp.lock_mm_and_find_vma
1.46 ± 4% -0.1 1.33 ± 5% perf-profile.children.cycles-pp.page_cache_ra_unbounded
0.40 ± 5% -0.1 0.34 ± 8% perf-profile.children.cycles-pp.mas_next_slot
0.22 ± 13% -0.1 0.17 ± 14% perf-profile.children.cycles-pp.__mod_memcg_lruvec_state
0.03 ±100% +0.0 0.07 ± 5% perf-profile.children.cycles-pp.housekeeping_test_cpu
0.44 ± 4% -0.1 0.32 ± 13% perf-profile.self.cycles-pp.__handle_mm_fault
0.67 ± 9% -0.1 0.54 ± 10% perf-profile.self.cycles-pp.mtree_range_walk
0.58 ± 7% -0.1 0.49 ± 6% perf-profile.self.cycles-pp.percpu_counter_add_batch
0.16 ± 7% -0.1 0.10 ± 20% perf-profile.self.cycles-pp.madvise_cold_or_pageout_pte_range
0.39 ± 5% -0.1 0.33 ± 8% perf-profile.self.cycles-pp.mas_next_slot
0.26 ± 6% +0.0 0.29 ± 10% perf-profile.self.cycles-pp.filemap_fault
***************************************************************************************************
lkp-cpl-4sp2: 224 threads 4 sockets Intel(R) Xeon(R) Platinum 8380H CPU @ 2.90GHz (Cooper Lake) with 192G memory
=========================================================================================
compiler/cpufreq_governor/kconfig/mode/nr_task/rootfs/tbox_group/test/testcase:
gcc-12/performance/x86_64-rhel-8.3/thread/50%/debian-11.1-x86_64-20220510.cgz/lkp-cpl-4sp2/page_fault3/will-it-scale
commit:
34611600bf ("mm: Call wp_page_copy() under the VMA lock")
c8b329d48e ("mm: Handle shared faults under the VMA lock")
34611600bfd1bf9f c8b329d48e0dac7438168a1857c
---------------- ---------------------------
%stddev %change %stddev
\ | \
46289 +39.6% 64618 ± 2% uptime.idle
3.839e+10 +47.5% 5.663e+10 ± 2% cpuidle..time
44548500 +28.6% 57277226 ± 2% cpuidle..usage
244.33 ± 7% +92.0% 469.17 ± 14% perf-c2c.DRAM.local
563.00 ± 3% -60.7% 221.00 ± 15% perf-c2c.HITM.remote
554762 -20.3% 441916 meminfo.Inactive
554566 -20.3% 441725 meminfo.Inactive(anon)
7360875 +46.0% 10746773 ± 2% meminfo.Mapped
20123 +28.9% 25930 meminfo.PageTables
56.22 +46.8% 82.52 ± 2% vmstat.cpu.id
63.54 ± 8% -39.5% 38.45 ± 14% vmstat.procs.r
23694 -84.7% 3627 vmstat.system.cs
455148 ± 2% +123.1% 1015448 ± 7% vmstat.system.in
2882478 ± 2% +274.8% 10804264 ± 5% will-it-scale.112.threads
55.72 +47.6% 82.27 ± 2% will-it-scale.112.threads_idle
25736 ± 2% +274.8% 96466 ± 5% will-it-scale.per_thread_ops
2882478 ± 2% +274.8% 10804264 ± 5% will-it-scale.workload
55.97 +26.4 82.36 ± 2% mpstat.cpu.all.idle%
0.82 -0.1 0.70 ± 4% mpstat.cpu.all.irq%
0.11 ± 4% -0.1 0.05 ± 5% mpstat.cpu.all.soft%
42.51 -26.9 15.64 ± 11% mpstat.cpu.all.sys%
0.59 ± 17% +0.7 1.25 ± 41% mpstat.cpu.all.usr%
1841712 +44.8% 2666713 ± 3% numa-meminfo.node0.Mapped
5224 ± 3% +31.6% 6877 ± 4% numa-meminfo.node0.PageTables
1845678 +46.3% 2699787 ± 2% numa-meminfo.node1.Mapped
5064 ± 5% +23.0% 6231 ± 2% numa-meminfo.node1.PageTables
1826141 +47.6% 2694729 ± 2% numa-meminfo.node2.Mapped
4794 ± 2% +30.8% 6269 ± 3% numa-meminfo.node2.PageTables
1868742 ± 2% +44.9% 2708096 ± 3% numa-meminfo.node3.Mapped
5026 ± 5% +28.8% 6474 ± 4% numa-meminfo.node3.PageTables
1591430 ± 4% +70.8% 2718150 ± 3% numa-numastat.node0.local_node
1673574 ± 3% +68.6% 2821949 ± 3% numa-numastat.node0.numa_hit
1577936 ± 6% +74.8% 2757801 ± 2% numa-numastat.node1.local_node
1645522 ± 5% +73.0% 2847142 ± 3% numa-numastat.node1.numa_hit
1537208 ± 3% +77.3% 2725353 ± 2% numa-numastat.node2.local_node
1639749 ± 3% +71.4% 2811161 ± 2% numa-numastat.node2.numa_hit
1637504 ± 5% +72.8% 2829154 ± 5% numa-numastat.node3.local_node
1732850 ± 4% +67.2% 2898001 ± 3% numa-numastat.node3.numa_hit
1684 -59.8% 677.17 ± 13% turbostat.Avg_MHz
44.43 -26.6 17.86 ± 13% turbostat.Busy%
44289096 +28.7% 57018721 ± 2% turbostat.C1
56.21 +26.6 82.76 ± 2% turbostat.C1%
55.57 +47.8% 82.14 ± 2% turbostat.CPU%c1
0.01 +533.3% 0.06 ± 17% turbostat.IPC
2.014e+08 ± 3% +174.5% 5.527e+08 ± 7% turbostat.IRQ
43515 ± 3% +37.9% 59997 ± 3% turbostat.POLL
685.24 -22.4% 532.03 ± 3% turbostat.PkgWatt
17.33 +5.6% 18.30 turbostat.RAMWatt
458598 +45.2% 666035 ± 3% numa-vmstat.node0.nr_mapped
1305 ± 3% +32.0% 1723 ± 4% numa-vmstat.node0.nr_page_table_pages
1673564 ± 3% +68.6% 2822055 ± 3% numa-vmstat.node0.numa_hit
1591420 ± 4% +70.8% 2718256 ± 3% numa-vmstat.node0.numa_local
461362 +46.3% 674878 ± 2% numa-vmstat.node1.nr_mapped
1266 ± 4% +23.2% 1559 ± 2% numa-vmstat.node1.nr_page_table_pages
1645442 ± 5% +73.0% 2847103 ± 3% numa-vmstat.node1.numa_hit
1577856 ± 6% +74.8% 2757762 ± 2% numa-vmstat.node1.numa_local
456314 +47.5% 672973 ± 2% numa-vmstat.node2.nr_mapped
1198 ± 2% +31.0% 1569 ± 3% numa-vmstat.node2.nr_page_table_pages
1639701 ± 3% +71.4% 2811174 ± 2% numa-vmstat.node2.numa_hit
1537161 ± 3% +77.3% 2725366 ± 2% numa-vmstat.node2.numa_local
464153 +46.1% 677988 ± 3% numa-vmstat.node3.nr_mapped
1255 ± 5% +29.1% 1621 ± 4% numa-vmstat.node3.nr_page_table_pages
1732732 ± 4% +67.3% 2898025 ± 3% numa-vmstat.node3.numa_hit
1637386 ± 5% +72.8% 2829178 ± 5% numa-vmstat.node3.numa_local
104802 -2.5% 102214 proc-vmstat.nr_anon_pages
4433098 -1.0% 4389891 proc-vmstat.nr_file_pages
138599 -20.3% 110426 proc-vmstat.nr_inactive_anon
1842991 +45.8% 2687030 ± 2% proc-vmstat.nr_mapped
5030 +28.9% 6483 proc-vmstat.nr_page_table_pages
3710638 -1.2% 3667429 proc-vmstat.nr_shmem
138599 -20.3% 110426 proc-vmstat.nr_zone_inactive_anon
43540 ± 7% -82.6% 7576 ± 47% proc-vmstat.numa_hint_faults
26753 ± 10% -77.6% 5982 ± 56% proc-vmstat.numa_hint_faults_local
6693986 +70.0% 11381806 ± 3% proc-vmstat.numa_hit
6346365 +73.9% 11034009 ± 3% proc-vmstat.numa_local
21587 ± 31% -92.2% 1683 ± 58% proc-vmstat.numa_pages_migrated
197966 -81.7% 36131 ± 25% proc-vmstat.numa_pte_updates
3749632 -1.1% 3708618 proc-vmstat.pgactivate
6848722 +68.4% 11532638 ± 3% proc-vmstat.pgalloc_normal
8.677e+08 ± 2% +276.2% 3.265e+09 ± 5% proc-vmstat.pgfault
6646708 +72.1% 11436096 ± 3% proc-vmstat.pgfree
21587 ± 31% -92.2% 1683 ± 58% proc-vmstat.pgmigrate_success
54536 ± 8% -24.2% 41332 ± 3% proc-vmstat.pgreuse
6305732 -84.8% 961479 ± 36% sched_debug.cfs_rq:/.avg_vruntime.avg
10700237 -83.0% 1820191 ± 34% sched_debug.cfs_rq:/.avg_vruntime.max
1797215 ± 18% -93.8% 112003 ± 80% sched_debug.cfs_rq:/.avg_vruntime.min
1512854 ± 2% -75.4% 372673 ± 31% sched_debug.cfs_rq:/.avg_vruntime.stddev
0.13 ± 20% +62.6% 0.21 ± 26% sched_debug.cfs_rq:/.h_nr_running.avg
0.33 ± 8% +17.2% 0.39 ± 9% sched_debug.cfs_rq:/.h_nr_running.stddev
4781 ± 82% -100.0% 0.12 ±223% sched_debug.cfs_rq:/.left_vruntime.avg
804679 ± 78% -100.0% 27.67 ±223% sched_debug.cfs_rq:/.left_vruntime.max
61817 ± 80% -100.0% 1.84 ±223% sched_debug.cfs_rq:/.left_vruntime.stddev
2654 ± 21% +156.8% 6815 ± 18% sched_debug.cfs_rq:/.load.avg
6305732 -84.8% 961479 ± 36% sched_debug.cfs_rq:/.min_vruntime.avg
10700237 -83.0% 1820191 ± 34% sched_debug.cfs_rq:/.min_vruntime.max
1797215 ± 18% -93.8% 112003 ± 80% sched_debug.cfs_rq:/.min_vruntime.min
1512854 ± 2% -75.4% 372673 ± 31% sched_debug.cfs_rq:/.min_vruntime.stddev
0.13 ± 20% +63.4% 0.21 ± 26% sched_debug.cfs_rq:/.nr_running.avg
0.33 ± 7% +18.1% 0.39 ± 9% sched_debug.cfs_rq:/.nr_running.stddev
4781 ± 82% -100.0% 0.12 ±223% sched_debug.cfs_rq:/.right_vruntime.avg
804679 ± 78% -100.0% 27.67 ±223% sched_debug.cfs_rq:/.right_vruntime.max
61817 ± 80% -100.0% 1.84 ±223% sched_debug.cfs_rq:/.right_vruntime.stddev
495.58 ± 3% -56.6% 214.98 ± 24% sched_debug.cfs_rq:/.runnable_avg.avg
1096 ± 7% -13.4% 949.07 ± 3% sched_debug.cfs_rq:/.runnable_avg.max
359.89 -23.0% 277.09 ± 11% sched_debug.cfs_rq:/.runnable_avg.stddev
493.94 ± 3% -56.5% 214.69 ± 24% sched_debug.cfs_rq:/.util_avg.avg
359.20 -22.9% 276.81 ± 11% sched_debug.cfs_rq:/.util_avg.stddev
97.00 ± 24% +76.3% 171.06 ± 31% sched_debug.cfs_rq:/.util_est_enqueued.avg
1512762 ± 4% -35.1% 981444 sched_debug.cpu.avg_idle.avg
5146368 ± 10% -71.3% 1476288 ± 33% sched_debug.cpu.avg_idle.max
578157 ± 8% -68.8% 180178 ± 10% sched_debug.cpu.avg_idle.min
670957 ± 5% -83.7% 109591 ± 26% sched_debug.cpu.avg_idle.stddev
73.60 ± 11% -81.3% 13.79 ± 9% sched_debug.cpu.clock.stddev
650.52 ± 18% +58.1% 1028 ± 14% sched_debug.cpu.curr->pid.avg
1959 ± 7% +19.6% 2342 ± 6% sched_debug.cpu.curr->pid.stddev
924262 ± 3% -45.6% 502853 sched_debug.cpu.max_idle_balance_cost.avg
2799134 ± 10% -70.8% 817753 ± 35% sched_debug.cpu.max_idle_balance_cost.max
377335 ± 9% -93.4% 24979 ± 94% sched_debug.cpu.max_idle_balance_cost.stddev
0.00 ± 8% -59.7% 0.00 ± 55% sched_debug.cpu.next_balance.stddev
0.10 ± 17% +57.8% 0.15 ± 14% sched_debug.cpu.nr_running.avg
1.28 ± 6% -19.6% 1.03 ± 6% sched_debug.cpu.nr_running.max
0.29 ± 7% +19.1% 0.35 ± 6% sched_debug.cpu.nr_running.stddev
17163 -79.4% 3534 ± 5% sched_debug.cpu.nr_switches.avg
7523 ± 10% -87.2% 961.21 ± 12% sched_debug.cpu.nr_switches.min
0.33 ± 5% -18.7% 0.27 ± 6% sched_debug.cpu.nr_uninterruptible.avg
0.00 -100.0% 0.00 sched_debug.rt_rq:.rt_nr_migratory.avg
0.17 -100.0% 0.00 sched_debug.rt_rq:.rt_nr_migratory.max
0.01 -100.0% 0.00 sched_debug.rt_rq:.rt_nr_migratory.stddev
0.00 -100.0% 0.00 sched_debug.rt_rq:.rt_nr_running.avg
0.17 -100.0% 0.00 sched_debug.rt_rq:.rt_nr_running.max
0.01 -100.0% 0.00 sched_debug.rt_rq:.rt_nr_running.stddev
0.18 ± 34% -100.0% 0.00 sched_debug.rt_rq:.rt_time.avg
40.73 ± 34% -100.0% 0.00 sched_debug.rt_rq:.rt_time.max
2.72 ± 34% -100.0% 0.00 sched_debug.rt_rq:.rt_time.stddev
2.63e+09 +165.9% 6.995e+09 ± 5% perf-stat.i.branch-instructions
0.45 -0.2 0.22 ± 3% perf-stat.i.branch-miss-rate%
12246564 +25.8% 15409530 ± 3% perf-stat.i.branch-misses
40.05 ± 6% +5.9 45.95 perf-stat.i.cache-miss-rate%
23716 -85.2% 3516 perf-stat.i.context-switches
30.25 ± 2% -84.8% 4.58 ± 18% perf-stat.i.cpi
3.785e+11 -60.2% 1.507e+11 ± 13% perf-stat.i.cpu-cycles
270.31 -6.3% 253.41 perf-stat.i.cpu-migrations
9670 ± 38% -79.6% 1972 ± 22% perf-stat.i.cycles-between-cache-misses
0.03 ± 4% -0.0 0.01 ± 10% perf-stat.i.dTLB-load-miss-rate%
958512 ± 3% +20.5% 1154691 ± 5% perf-stat.i.dTLB-load-misses
3.15e+09 +172.1% 8.571e+09 ± 5% perf-stat.i.dTLB-loads
4.91 +1.6 6.54 perf-stat.i.dTLB-store-miss-rate%
87894742 ± 2% +276.4% 3.308e+08 ± 5% perf-stat.i.dTLB-store-misses
1.709e+09 +176.5% 4.725e+09 ± 5% perf-stat.i.dTLB-stores
78.59 +13.6 92.23 perf-stat.i.iTLB-load-miss-rate%
8890053 +168.7% 23884564 ± 6% perf-stat.i.iTLB-load-misses
2405390 -17.2% 1990833 perf-stat.i.iTLB-loads
1.257e+10 +164.2% 3.323e+10 ± 5% perf-stat.i.instructions
0.03 ± 4% +571.5% 0.23 ± 17% perf-stat.i.ipc
1.69 -60.1% 0.67 ± 13% perf-stat.i.metric.GHz
33.38 +175.7% 92.04 ± 5% perf-stat.i.metric.M/sec
2877597 ± 2% +274.4% 10773809 ± 5% perf-stat.i.minor-faults
86.68 -3.1 83.57 ± 2% perf-stat.i.node-load-miss-rate%
5637384 ± 17% +105.0% 11559372 ± 20% perf-stat.i.node-load-misses
857520 ± 10% +158.1% 2213133 ± 9% perf-stat.i.node-loads
46.44 -16.5 29.97 perf-stat.i.node-store-miss-rate%
2608818 +80.4% 4705017 ± 3% perf-stat.i.node-store-misses
3024158 ± 2% +264.6% 11026854 ± 5% perf-stat.i.node-stores
2877597 ± 2% +274.4% 10773809 ± 5% perf-stat.i.page-faults
0.47 -0.2 0.22 ± 3% perf-stat.overall.branch-miss-rate%
39.95 ± 6% +6.0 45.93 perf-stat.overall.cache-miss-rate%
30.11 ± 2% -84.8% 4.58 ± 18% perf-stat.overall.cpi
9647 ± 38% -79.6% 1971 ± 21% perf-stat.overall.cycles-between-cache-misses
0.03 ± 4% -0.0 0.01 ± 10% perf-stat.overall.dTLB-load-miss-rate%
4.89 +1.7 6.54 perf-stat.overall.dTLB-store-miss-rate%
78.70 +13.6 92.28 perf-stat.overall.iTLB-load-miss-rate%
0.03 ± 2% +579.6% 0.23 ± 17% perf-stat.overall.ipc
86.61 -3.0 83.59 ± 2% perf-stat.overall.node-load-miss-rate%
46.33 -16.4 29.93 perf-stat.overall.node-store-miss-rate%
1315354 -29.2% 931848 perf-stat.overall.path-length
2.621e+09 +166.0% 6.97e+09 ± 5% perf-stat.ps.branch-instructions
12198295 +25.8% 15340303 ± 3% perf-stat.ps.branch-misses
23623 -85.2% 3499 perf-stat.ps.context-switches
3.77e+11 -60.2% 1.502e+11 ± 13% perf-stat.ps.cpu-cycles
266.50 -5.2% 252.58 perf-stat.ps.cpu-migrations
961568 ± 4% +19.7% 1150857 ± 5% perf-stat.ps.dTLB-load-misses
3.138e+09 +172.1% 8.541e+09 ± 5% perf-stat.ps.dTLB-loads
87534988 ± 2% +276.7% 3.297e+08 ± 5% perf-stat.ps.dTLB-store-misses
1.702e+09 +176.6% 4.708e+09 ± 5% perf-stat.ps.dTLB-stores
8850726 +169.0% 23812332 ± 6% perf-stat.ps.iTLB-load-misses
2394294 -17.1% 1983791 perf-stat.ps.iTLB-loads
1.253e+10 +164.3% 3.311e+10 ± 5% perf-stat.ps.instructions
2865432 ± 2% +274.7% 10737502 ± 5% perf-stat.ps.minor-faults
5615782 ± 17% +105.2% 11521907 ± 20% perf-stat.ps.node-load-misses
856502 ± 10% +157.6% 2206185 ± 9% perf-stat.ps.node-loads
2598165 +80.5% 4689182 ± 3% perf-stat.ps.node-store-misses
3009612 ± 2% +265.1% 10987528 ± 5% perf-stat.ps.node-stores
2865432 ± 2% +274.7% 10737502 ± 5% perf-stat.ps.page-faults
3.791e+12 +165.5% 1.006e+13 ± 5% perf-stat.total.instructions
0.05 ± 17% -77.2% 0.01 ± 73% perf-sched.sch_delay.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
0.07 ± 34% -90.1% 0.01 ± 99% perf-sched.sch_delay.avg.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
0.00 ± 33% +377.8% 0.01 ± 9% perf-sched.sch_delay.avg.ms.__cond_resched.tlb_batch_pages_flush.zap_pte_range.zap_pmd_range.isra
0.08 ± 25% -90.3% 0.01 ± 12% perf-sched.sch_delay.avg.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
0.12 ± 64% -91.3% 0.01 ± 8% perf-sched.sch_delay.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
0.06 ± 37% -89.4% 0.01 ± 16% perf-sched.sch_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
0.03 ± 19% -80.0% 0.01 ± 34% perf-sched.sch_delay.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
0.21 ±100% -97.3% 0.01 ± 6% perf-sched.sch_delay.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
0.02 ± 57% -100.0% 0.00 perf-sched.sch_delay.avg.ms.rcu_gp_kthread.kthread.ret_from_fork.ret_from_fork_asm
0.04 ± 34% -88.0% 0.00 ± 15% perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
0.08 ± 21% -90.3% 0.01 ± 25% perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
0.06 ± 80% -100.0% 0.00 perf-sched.sch_delay.avg.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read_killable.lock_mm_and_find_vma
0.00 ± 19% +173.3% 0.01 ± 5% perf-sched.sch_delay.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write_killable.vm_mmap_pgoff
0.02 ± 10% -87.3% 0.00 perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
0.05 ± 16% -85.9% 0.01 ± 7% perf-sched.sch_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
0.03 ± 10% -74.0% 0.01 ± 27% perf-sched.sch_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
0.09 ± 40% -91.2% 0.01 ± 17% perf-sched.sch_delay.avg.ms.syslog_print.do_syslog.kmsg_read.vfs_read
0.01 ± 16% -74.6% 0.00 perf-sched.sch_delay.avg.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
0.12 ± 66% -93.3% 0.01 ± 19% perf-sched.sch_delay.max.ms.__cond_resched.__wait_for_common.affine_move_task.__set_cpus_allowed_ptr.__sched_setaffinity
0.79 ± 40% -96.9% 0.02 ±186% perf-sched.sch_delay.max.ms.__cond_resched.stop_one_cpu.sched_exec.bprm_execve.part
0.00 ± 10% +123.1% 0.01 ± 22% perf-sched.sch_delay.max.ms.__cond_resched.tlb_batch_pages_flush.zap_pte_range.zap_pmd_range.isra
0.13 ± 14% -92.2% 0.01 ± 27% perf-sched.sch_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
0.15 ± 54% -91.6% 0.01 ± 17% perf-sched.sch_delay.max.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
0.14 ± 43% -93.0% 0.01 ± 27% perf-sched.sch_delay.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
24.13 ±116% -99.9% 0.01 ± 15% perf-sched.sch_delay.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
0.06 ± 63% -100.0% 0.00 perf-sched.sch_delay.max.ms.rcu_gp_kthread.kthread.ret_from_fork.ret_from_fork_asm
0.22 ± 39% -95.5% 0.01 ± 15% perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
0.22 ± 60% -93.8% 0.01 ± 20% perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait
65.91 ± 71% -100.0% 0.00 perf-sched.sch_delay.max.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read_killable.lock_mm_and_find_vma
13.37 ±143% +703.9% 107.48 ± 64% perf-sched.sch_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write_killable.__vm_munmap
0.01 ± 31% +121.2% 0.02 ± 31% perf-sched.sch_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write_killable.vm_mmap_pgoff
0.28 ± 14% -96.4% 0.01 ± 29% perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
0.19 ± 16% -93.0% 0.01 ± 23% perf-sched.sch_delay.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork
0.16 ± 50% -93.2% 0.01 ± 27% perf-sched.sch_delay.max.ms.syslog_print.do_syslog.kmsg_read.vfs_read
0.10 ± 9% -94.3% 0.01 ± 37% perf-sched.sch_delay.max.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
31.83 ± 4% +503.4% 192.03 ± 4% perf-sched.total_wait_and_delay.average.ms
61361 -82.4% 10800 ± 5% perf-sched.total_wait_and_delay.count.ms
31.76 ± 4% +502.8% 191.45 ± 5% perf-sched.total_wait_time.average.ms
1.67 ± 13% +9097.0% 153.21 ± 4% perf-sched.wait_and_delay.avg.ms.__cond_resched.shmem_get_folio_gfp.shmem_fault.__do_fault.do_fault
32.29 ± 9% -25.8% 23.95 ± 22% perf-sched.wait_and_delay.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
1.57 ± 2% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read_killable.lock_mm_and_find_vma
1.11 ± 8% +12026.0% 134.28 ± 8% perf-sched.wait_and_delay.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write_killable.__vm_munmap
3.84 ± 5% +20.0% 4.61 ± 2% perf-sched.wait_and_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
610.00 ± 7% -30.8% 421.83 ± 15% perf-sched.wait_and_delay.count.__cond_resched.shmem_get_folio_gfp.shmem_fault.__do_fault.do_fault
51568 ± 2% -100.0% 0.00 perf-sched.wait_and_delay.count.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read_killable.lock_mm_and_find_vma
1146 ± 4% +51.3% 1734 ± 6% perf-sched.wait_and_delay.count.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write_killable.__vm_munmap
1226 ± 4% -11.6% 1084 ± 2% perf-sched.wait_and_delay.count.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
995.33 ± 3% -19.5% 801.33 ± 4% perf-sched.wait_and_delay.count.worker_thread.kthread.ret_from_fork.ret_from_fork_asm
33.51 ± 78% +547.7% 217.02 ± 2% perf-sched.wait_and_delay.max.ms.__cond_resched.shmem_get_folio_gfp.shmem_fault.__do_fault.do_fault
65.98 ± 70% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read_killable.lock_mm_and_find_vma
15.43 ±115% +1309.1% 217.47 perf-sched.wait_and_delay.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write_killable.__vm_munmap
1.47 ± 8% +10284.0% 153.06 ± 4% perf-sched.wait_time.avg.ms.__cond_resched.shmem_get_folio_gfp.shmem_fault.__do_fault.do_fault
0.01 ± 34% +1.4e+06% 179.68 ± 11% perf-sched.wait_time.avg.ms.__cond_resched.tlb_batch_pages_flush.zap_pte_range.zap_pmd_range.isra
5.95 ± 21% -73.5% 1.58 ± 8% perf-sched.wait_time.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
1.44 ± 6% +10218.7% 148.31 ± 9% perf-sched.wait_time.avg.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
2.93 ± 23% -100.0% 0.00 perf-sched.wait_time.avg.ms.rcu_gp_kthread.kthread.ret_from_fork.ret_from_fork_asm
1.51 ± 4% -100.0% 0.00 perf-sched.wait_time.avg.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read_killable.lock_mm_and_find_vma
1.08 ± 5% +12334.2% 134.16 ± 8% perf-sched.wait_time.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write_killable.__vm_munmap
0.38 ± 27% +47951.7% 182.68 ± 2% perf-sched.wait_time.avg.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write_killable.vm_mmap_pgoff
2.83 ± 16% -82.5% 0.49 ± 2% perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
3.82 ± 6% +20.4% 4.60 ± 2% perf-sched.wait_time.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
6.49 ± 18% -75.3% 1.60 ± 9% perf-sched.wait_time.avg.ms.syslog_print.do_syslog.kmsg_read.vfs_read
0.03 ±133% -99.4% 0.00 ±223% perf-sched.wait_time.avg.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
2.99 ± 14% +7148.4% 217.02 ± 2% perf-sched.wait_time.max.ms.__cond_resched.shmem_get_folio_gfp.shmem_fault.__do_fault.do_fault
0.06 ± 51% +3.5e+05% 209.30 ± 2% perf-sched.wait_time.max.ms.__cond_resched.tlb_batch_pages_flush.zap_pte_range.zap_pmd_range.isra
11.89 ± 21% -73.5% 3.16 ± 8% perf-sched.wait_time.max.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
2.62 ± 4% +7966.9% 211.41 ± 2% perf-sched.wait_time.max.ms.exit_to_user_mode_loop.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_exc_page_fault
4.63 ± 7% -100.0% 0.00 perf-sched.wait_time.max.ms.rcu_gp_kthread.kthread.ret_from_fork.ret_from_fork_asm
5.60 ± 74% -100.0% 0.00 perf-sched.wait_time.max.ms.schedule_preempt_disabled.rwsem_down_read_slowpath.down_read_killable.lock_mm_and_find_vma
4.03 ± 3% +5294.2% 217.47 perf-sched.wait_time.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write_killable.__vm_munmap
2.46 ± 25% +8701.7% 216.14 perf-sched.wait_time.max.ms.schedule_preempt_disabled.rwsem_down_write_slowpath.down_write_killable.vm_mmap_pgoff
16.44 ± 21% -93.5% 1.07 ± 3% perf-sched.wait_time.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
14.31 ± 59% -65.0% 5.01 perf-sched.wait_time.max.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
12.98 ± 18% -75.3% 3.20 ± 9% perf-sched.wait_time.max.ms.syslog_print.do_syslog.kmsg_read.vfs_read
0.63 ±151% -98.7% 0.01 ± 46% perf-sched.wait_time.max.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
24.35 ± 5% -24.3 0.00 perf-profile.calltrace.cycles-pp.lock_mm_and_find_vma.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.testcase
22.08 ± 2% -22.1 0.00 perf-profile.calltrace.cycles-pp.down_read_trylock.lock_mm_and_find_vma.do_user_addr_fault.exc_page_fault.asm_exc_page_fault
13.70 ± 2% -13.7 0.00 perf-profile.calltrace.cycles-pp.up_read.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.testcase
74.33 -12.7 61.66 perf-profile.calltrace.cycles-pp.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.testcase
74.37 -12.2 62.18 perf-profile.calltrace.cycles-pp.exc_page_fault.asm_exc_page_fault.testcase
3.61 ± 8% -2.7 0.89 ± 16% perf-profile.calltrace.cycles-pp.acpi_safe_halt.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter.cpuidle_idle_call
0.00 +0.7 0.71 ± 21% perf-profile.calltrace.cycles-pp.sysvec_apic_timer_interrupt.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_enter.cpuidle_enter_state
0.00 +0.7 0.71 ± 5% perf-profile.calltrace.cycles-pp.smp_call_function_many_cond.on_each_cpu_cond_mask.flush_tlb_mm_range.zap_pte_range.zap_pmd_range
0.00 +0.7 0.72 ± 5% perf-profile.calltrace.cycles-pp.on_each_cpu_cond_mask.flush_tlb_mm_range.zap_pte_range.zap_pmd_range.unmap_page_range
0.00 +0.7 0.73 ± 5% perf-profile.calltrace.cycles-pp.flush_tlb_mm_range.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas
0.00 +0.8 0.81 ± 15% perf-profile.calltrace.cycles-pp.filemap_get_entry.shmem_get_folio_gfp.shmem_fault.__do_fault.do_fault
0.71 ± 3% +1.2 1.90 ± 14% perf-profile.calltrace.cycles-pp.zap_pte_range.zap_pmd_range.unmap_page_range.unmap_vmas.unmap_region
0.00 +1.2 1.20 ± 17% perf-profile.calltrace.cycles-pp.shmem_get_folio_gfp.shmem_fault.__do_fault.do_fault.__handle_mm_fault
0.71 ± 3% +1.2 1.94 ± 14% perf-profile.calltrace.cycles-pp.unmap_vmas.unmap_region.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap
0.71 ± 3% +1.2 1.94 ± 14% perf-profile.calltrace.cycles-pp.unmap_page_range.unmap_vmas.unmap_region.do_vmi_align_munmap.do_vmi_munmap
0.71 ± 3% +1.2 1.94 ± 14% perf-profile.calltrace.cycles-pp.zap_pmd_range.unmap_page_range.unmap_vmas.unmap_region.do_vmi_align_munmap
0.00 +1.3 1.28 ± 27% perf-profile.calltrace.cycles-pp.set_pte_range.finish_fault.do_fault.__handle_mm_fault.handle_mm_fault
0.73 ± 3% +1.3 2.02 ± 14% perf-profile.calltrace.cycles-pp.do_vmi_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.73 ± 3% +1.3 2.02 ± 14% perf-profile.calltrace.cycles-pp.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap.__x64_sys_munmap.do_syscall_64
0.72 ± 3% +1.3 2.00 ± 14% perf-profile.calltrace.cycles-pp.unmap_region.do_vmi_align_munmap.do_vmi_munmap.__vm_munmap.__x64_sys_munmap
0.00 +1.4 1.36 ± 17% perf-profile.calltrace.cycles-pp.shmem_fault.__do_fault.do_fault.__handle_mm_fault.handle_mm_fault
0.00 +1.4 1.40 ± 18% perf-profile.calltrace.cycles-pp.sync_regs.asm_exc_page_fault.testcase
0.00 +1.4 1.42 ± 17% perf-profile.calltrace.cycles-pp.__do_fault.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
0.00 +1.7 1.68 ± 20% perf-profile.calltrace.cycles-pp.finish_fault.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault
0.00 +2.0 2.05 ± 14% perf-profile.calltrace.cycles-pp.__vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.00 +2.0 2.05 ± 14% perf-profile.calltrace.cycles-pp.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe
41.13 +2.1 43.18 perf-profile.calltrace.cycles-pp.asm_sysvec_apic_timer_interrupt.acpi_safe_halt.acpi_idle_enter.cpuidle_enter_state.cpuidle_enter
0.00 +2.1 2.08 ± 14% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
0.00 +2.1 2.08 ± 14% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
1.52 ± 13% +2.1 3.66 ± 14% perf-profile.calltrace.cycles-pp.do_fault.__handle_mm_fault.handle_mm_fault.do_user_addr_fault.exc_page_fault
3.32 ± 19% +19.0 22.27 ± 24% perf-profile.calltrace.cycles-pp.lock_vma_under_rcu.do_user_addr_fault.exc_page_fault.asm_exc_page_fault.testcase
24.35 ± 5% -24.4 0.00 perf-profile.children.cycles-pp.lock_mm_and_find_vma
22.17 ± 2% -22.0 0.16 ± 17% perf-profile.children.cycles-pp.down_read_trylock
14.60 ± 2% -14.5 0.14 ± 21% perf-profile.children.cycles-pp.up_read
74.37 -12.4 62.02 perf-profile.children.cycles-pp.do_user_addr_fault
74.39 -12.2 62.20 perf-profile.children.cycles-pp.exc_page_fault
75.34 -5.5 69.86 perf-profile.children.cycles-pp.asm_exc_page_fault
23.24 -0.9 22.38 perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt
0.44 ± 4% -0.2 0.28 ± 14% perf-profile.children.cycles-pp.scheduler_tick
0.20 ± 39% -0.1 0.08 ± 86% perf-profile.children.cycles-pp.x86_64_start_kernel
0.20 ± 39% -0.1 0.08 ± 86% perf-profile.children.cycles-pp.x86_64_start_reservations
0.20 ± 39% -0.1 0.08 ± 86% perf-profile.children.cycles-pp.start_kernel
0.20 ± 39% -0.1 0.08 ± 86% perf-profile.children.cycles-pp.arch_call_rest_init
0.20 ± 39% -0.1 0.08 ± 86% perf-profile.children.cycles-pp.rest_init
0.28 ± 4% -0.1 0.17 ± 19% perf-profile.children.cycles-pp._compound_head
0.47 ± 4% -0.1 0.37 ± 14% perf-profile.children.cycles-pp.update_process_times
0.47 ± 4% -0.1 0.37 ± 14% perf-profile.children.cycles-pp.tick_sched_handle
0.12 ± 13% -0.1 0.03 ±102% perf-profile.children.cycles-pp.load_balance
0.00 +0.1 0.07 ± 17% perf-profile.children.cycles-pp._raw_spin_trylock
0.00 +0.1 0.07 ± 10% perf-profile.children.cycles-pp.irqtime_account_irq
0.03 ± 70% +0.1 0.11 ± 19% perf-profile.children.cycles-pp.rebalance_domains
0.00 +0.1 0.08 ± 19% perf-profile.children.cycles-pp.__irqentry_text_end
0.00 +0.1 0.08 ± 17% perf-profile.children.cycles-pp.__count_memcg_events
0.00 +0.1 0.08 ± 18% perf-profile.children.cycles-pp.cgroup_rstat_updated
0.00 +0.1 0.10 ± 23% perf-profile.children.cycles-pp.folio_mark_dirty
0.00 +0.1 0.10 ± 18% perf-profile.children.cycles-pp.__pte_offset_map
0.08 ± 6% +0.1 0.18 ± 20% perf-profile.children.cycles-pp.__do_softirq
0.00 +0.1 0.11 ± 14% perf-profile.children.cycles-pp.pte_offset_map_nolock
0.53 ± 4% +0.1 0.65 ± 16% perf-profile.children.cycles-pp.__hrtimer_run_queues
0.00 +0.1 0.12 ± 28% perf-profile.children.cycles-pp.__mod_node_page_state
0.05 ± 8% +0.1 0.18 ± 19% perf-profile.children.cycles-pp._raw_spin_lock
0.00 +0.1 0.14 ± 22% perf-profile.children.cycles-pp.folio_unlock
0.00 +0.1 0.14 ± 20% perf-profile.children.cycles-pp.release_pages
0.02 ±141% +0.1 0.16 ± 36% perf-profile.children.cycles-pp.ktime_get
0.00 +0.1 0.14 ± 6% perf-profile.children.cycles-pp.native_flush_tlb_local
0.08 ± 5% +0.1 0.23 ± 16% perf-profile.children.cycles-pp.__irq_exit_rcu
0.01 ±223% +0.2 0.16 ± 24% perf-profile.children.cycles-pp.perf_mux_hrtimer_handler
0.00 +0.2 0.15 ± 17% perf-profile.children.cycles-pp.handle_pte_fault
0.00 +0.2 0.16 ±115% perf-profile.children.cycles-pp.menu_select
0.00 +0.2 0.16 ± 20% perf-profile.children.cycles-pp.__mod_memcg_lruvec_state
0.57 ± 4% +0.2 0.74 ± 15% perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt
0.57 ± 4% +0.2 0.74 ± 15% perf-profile.children.cycles-pp.hrtimer_interrupt
0.01 ±223% +0.2 0.18 ± 29% perf-profile.children.cycles-pp.inode_needs_update_time
0.00 +0.2 0.18 ± 18% perf-profile.children.cycles-pp.tlb_batch_pages_flush
0.00 +0.2 0.19 ± 23% perf-profile.children.cycles-pp.__mod_lruvec_state
0.01 ±223% +0.2 0.21 ± 27% perf-profile.children.cycles-pp.file_update_time
0.00 +0.2 0.22 ± 4% perf-profile.children.cycles-pp.llist_reverse_order
0.02 ±141% +0.2 0.26 ± 5% perf-profile.children.cycles-pp.flush_tlb_func
0.00 +0.2 0.25 ± 18% perf-profile.children.cycles-pp.error_entry
0.00 +0.3 0.26 ± 18% perf-profile.children.cycles-pp.__pte_offset_map_lock
0.15 ± 9% +0.3 0.44 ± 21% perf-profile.children.cycles-pp.mtree_range_walk
0.07 ± 9% +0.3 0.38 ± 5% perf-profile.children.cycles-pp.default_send_IPI_mask_sequence_phys
0.05 ± 7% +0.3 0.37 ± 15% perf-profile.children.cycles-pp.xas_descend
0.66 ± 3% +0.3 1.00 ± 14% perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt
0.05 ± 7% +0.4 0.43 ± 18% perf-profile.children.cycles-pp.folio_add_file_rmap_range
0.04 ± 44% +0.4 0.43 ± 19% perf-profile.children.cycles-pp.page_remove_rmap
0.06 ± 6% +0.4 0.45 ± 20% perf-profile.children.cycles-pp.__mod_lruvec_page_state
0.04 ± 44% +0.4 0.44 ± 20% perf-profile.children.cycles-pp.tlb_flush_rmaps
0.06 ± 16% +0.4 0.48 ± 24% perf-profile.children.cycles-pp.fault_dirty_shared_page
0.30 ± 3% +0.4 0.72 ± 5% perf-profile.children.cycles-pp.on_each_cpu_cond_mask
0.30 ± 3% +0.4 0.72 ± 5% perf-profile.children.cycles-pp.smp_call_function_many_cond
0.31 ± 2% +0.4 0.74 ± 5% perf-profile.children.cycles-pp.flush_tlb_mm_range
0.07 ± 6% +0.5 0.54 ± 14% perf-profile.children.cycles-pp.xas_load
0.13 ± 8% +0.6 0.75 ± 4% perf-profile.children.cycles-pp.__flush_smp_call_function_queue
0.11 ± 8% +0.6 0.75 ± 4% perf-profile.children.cycles-pp.__sysvec_call_function
0.12 ± 8% +0.7 0.82 ± 4% perf-profile.children.cycles-pp.sysvec_call_function
0.12 ± 4% +0.7 0.82 ± 15% perf-profile.children.cycles-pp.filemap_get_entry
0.08 ± 5% +0.8 0.86 ± 19% perf-profile.children.cycles-pp.___perf_sw_event
0.11 ± 6% +1.0 1.09 ± 20% perf-profile.children.cycles-pp.__perf_sw_event
0.18 ± 6% +1.0 1.21 ± 16% perf-profile.children.cycles-pp.shmem_get_folio_gfp
0.25 ± 46% +1.0 1.30 ± 27% perf-profile.children.cycles-pp.set_pte_range
0.18 ± 5% +1.2 1.36 ± 17% perf-profile.children.cycles-pp.shmem_fault
0.25 ± 5% +1.2 1.43 ± 17% perf-profile.children.cycles-pp.__do_fault
0.93 ± 4% +1.2 2.12 ± 14% perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
0.93 ± 4% +1.2 2.12 ± 14% perf-profile.children.cycles-pp.do_syscall_64
0.72 ± 3% +1.2 1.94 ± 14% perf-profile.children.cycles-pp.zap_pte_range
0.72 ± 3% +1.2 1.94 ± 14% perf-profile.children.cycles-pp.unmap_vmas
0.72 ± 3% +1.2 1.94 ± 14% perf-profile.children.cycles-pp.unmap_page_range
0.72 ± 3% +1.2 1.94 ± 14% perf-profile.children.cycles-pp.zap_pmd_range
0.44 ± 4% +1.2 1.67 perf-profile.children.cycles-pp.asm_sysvec_call_function
0.18 ± 19% +1.3 1.45 ± 17% perf-profile.children.cycles-pp.sync_regs
0.74 ± 3% +1.3 2.02 ± 14% perf-profile.children.cycles-pp.do_vmi_munmap
0.74 ± 3% +1.3 2.02 ± 14% perf-profile.children.cycles-pp.do_vmi_align_munmap
0.72 ± 3% +1.3 2.00 ± 14% perf-profile.children.cycles-pp.unmap_region
0.74 ± 3% +1.3 2.05 ± 14% perf-profile.children.cycles-pp.__vm_munmap
0.74 ± 3% +1.3 2.05 ± 14% perf-profile.children.cycles-pp.__x64_sys_munmap
0.31 ± 37% +1.4 1.70 ± 20% perf-profile.children.cycles-pp.finish_fault
1.53 ± 13% +2.2 3.68 ± 14% perf-profile.children.cycles-pp.do_fault
0.82 ± 24% +4.1 4.90 ± 33% perf-profile.children.cycles-pp.native_irq_return_iret
3.34 ± 19% +19.0 22.33 ± 24% perf-profile.children.cycles-pp.lock_vma_under_rcu
21.96 ± 2% -21.8 0.16 ± 17% perf-profile.self.cycles-pp.down_read_trylock
14.47 ± 2% -14.3 0.13 ± 22% perf-profile.self.cycles-pp.up_read
3.51 ± 9% -3.5 0.04 ±108% perf-profile.self.cycles-pp.asm_sysvec_apic_timer_interrupt
0.27 ± 4% -0.1 0.14 ± 19% perf-profile.self.cycles-pp._compound_head
0.12 ± 6% +0.0 0.14 ± 4% perf-profile.self.cycles-pp.llist_add_batch
0.00 +0.1 0.06 ± 14% perf-profile.self.cycles-pp.__mod_lruvec_state
0.00 +0.1 0.07 ± 17% perf-profile.self.cycles-pp._raw_spin_trylock
0.00 +0.1 0.07 ± 23% perf-profile.self.cycles-pp.__irqentry_text_end
0.00 +0.1 0.07 ± 18% perf-profile.self.cycles-pp.do_fault
0.00 +0.1 0.08 ± 17% perf-profile.self.cycles-pp.finish_fault
0.00 +0.1 0.08 ± 14% perf-profile.self.cycles-pp.cgroup_rstat_updated
0.00 +0.1 0.09 ± 20% perf-profile.self.cycles-pp.inode_needs_update_time
0.00 +0.1 0.09 ± 16% perf-profile.self.cycles-pp.__pte_offset_map_lock
0.00 +0.1 0.09 ± 20% perf-profile.self.cycles-pp.__pte_offset_map
0.00 +0.1 0.10 ± 16% perf-profile.self.cycles-pp.__mod_lruvec_page_state
0.00 +0.1 0.11 ± 22% perf-profile.self.cycles-pp.__mod_memcg_lruvec_state
0.00 +0.1 0.11 ± 31% perf-profile.self.cycles-pp.__mod_node_page_state
0.00 +0.1 0.11 ± 13% perf-profile.self.cycles-pp.flush_tlb_func
0.07 ± 5% +0.1 0.19 ± 12% perf-profile.self.cycles-pp.smp_call_function_many_cond
0.02 ±141% +0.1 0.14 ± 40% perf-profile.self.cycles-pp.ktime_get
0.00 +0.1 0.13 ± 17% perf-profile.self.cycles-pp.exc_page_fault
0.00 +0.1 0.13 ± 14% perf-profile.self.cycles-pp.xas_load
0.00 +0.1 0.13 ± 21% perf-profile.self.cycles-pp.folio_unlock
0.00 +0.1 0.14 ± 19% perf-profile.self.cycles-pp.release_pages
0.00 +0.1 0.14 ± 7% perf-profile.self.cycles-pp.native_flush_tlb_local
0.00 +0.1 0.14 ± 18% perf-profile.self.cycles-pp.shmem_fault
0.01 ±223% +0.2 0.18 ± 18% perf-profile.self.cycles-pp._raw_spin_lock
0.00 +0.2 0.17 ± 18% perf-profile.self.cycles-pp.set_pte_range
0.00 +0.2 0.19 ± 17% perf-profile.self.cycles-pp.folio_add_file_rmap_range
0.00 +0.2 0.22 ± 19% perf-profile.self.cycles-pp.page_remove_rmap
0.00 +0.2 0.22 ± 5% perf-profile.self.cycles-pp.llist_reverse_order
0.00 +0.2 0.24 ± 18% perf-profile.self.cycles-pp.error_entry
0.00 +0.2 0.24 ± 25% perf-profile.self.cycles-pp.__perf_sw_event
0.02 ± 99% +0.2 0.27 ± 16% perf-profile.self.cycles-pp.filemap_get_entry
0.00 +0.3 0.28 ± 5% perf-profile.self.cycles-pp.__flush_smp_call_function_queue
0.14 ± 7% +0.3 0.43 ± 21% perf-profile.self.cycles-pp.mtree_range_walk
0.00 +0.3 0.29 ± 13% perf-profile.self.cycles-pp.asm_exc_page_fault
0.07 ± 8% +0.3 0.38 ± 4% perf-profile.self.cycles-pp.default_send_IPI_mask_sequence_phys
0.04 ± 44% +0.3 0.35 ± 15% perf-profile.self.cycles-pp.xas_descend
0.00 +0.3 0.31 ± 19% perf-profile.self.cycles-pp.zap_pte_range
0.03 ± 70% +0.3 0.35 ± 19% perf-profile.self.cycles-pp.shmem_get_folio_gfp
0.01 ±223% +0.5 0.53 ± 13% perf-profile.self.cycles-pp.do_user_addr_fault
0.08 ± 6% +0.7 0.74 ± 19% perf-profile.self.cycles-pp.___perf_sw_event
0.18 ± 19% +1.3 1.44 ± 17% perf-profile.self.cycles-pp.sync_regs
19.02 +2.4 21.40 perf-profile.self.cycles-pp.acpi_safe_halt
0.27 ± 71% +2.4 2.72 ± 76% perf-profile.self.cycles-pp.handle_mm_fault
0.82 ± 24% +4.1 4.89 ± 33% perf-profile.self.cycles-pp.native_irq_return_iret
0.76 ± 3% +5.6 6.37 ± 18% perf-profile.self.cycles-pp.testcase
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
--
0-DAY CI Kernel Test Service
https://github.com/intel/lkp-tests/wiki
next prev parent reply other threads:[~2023-10-20 13:24 UTC|newest]
Thread overview: 16+ messages / expand[flat|nested] mbox.gz Atom feed top
2023-10-06 19:53 [PATCH v2 0/6] Handle more faults under the VMA lock Matthew Wilcox (Oracle)
2023-10-06 19:53 ` [PATCH v2 1/6] mm: Make lock_folio_maybe_drop_mmap() VMA lock aware Matthew Wilcox (Oracle)
2023-10-08 21:47 ` Suren Baghdasaryan
2023-10-06 19:53 ` [PATCH v2 2/6] mm: Call wp_page_copy() under the VMA lock Matthew Wilcox (Oracle)
2023-10-08 22:00 ` Suren Baghdasaryan
2023-10-06 19:53 ` [PATCH v2 3/6] mm: Handle shared faults " Matthew Wilcox (Oracle)
2023-10-08 22:01 ` Suren Baghdasaryan
2023-10-20 13:23 ` kernel test robot [this message]
2023-10-06 19:53 ` [PATCH v2 4/6] mm: Handle COW " Matthew Wilcox (Oracle)
2023-10-08 22:05 ` Suren Baghdasaryan
2023-10-20 13:18 ` kernel test robot
2023-10-06 19:53 ` [PATCH v2 5/6] mm: Handle read " Matthew Wilcox (Oracle)
2023-10-08 22:06 ` Suren Baghdasaryan
2023-10-20 9:55 ` kernel test robot
2023-10-06 19:53 ` [PATCH v2 6/6] mm: Handle write faults to RO pages " Matthew Wilcox (Oracle)
2023-10-08 22:07 ` Suren Baghdasaryan
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=202310201857.d7db939a-oliver.sang@intel.com \
--to=oliver.sang@intel.com \
--cc=akpm@linux-foundation.org \
--cc=feng.tang@intel.com \
--cc=fengwei.yin@intel.com \
--cc=linux-mm@kvack.org \
--cc=lkp@intel.com \
--cc=oe-lkp@lists.linux.dev \
--cc=surenb@google.com \
--cc=willy@infradead.org \
--cc=ying.huang@intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).