========================================================================================= compiler/cpufreq_governor/kconfig/mode/nr_task/rootfs/tbox_group/test/testcase: gcc-12/performance/x86_64-rhel-8.3/process/100%/debian-11.1-x86_64-20220510.cgz/lkp-cpl-4sp2/fallocate1/will-it-scale commit: e0bf1dc859fdd mm: memcg: move vmstats structs definition above flushing code 8d59d2214c236 mm: memcg: make stats flushing threshold per-memcg 0cba55e237ba6 mm: memcg: optimize parent iteration in memcg_rstat_updated() e0bf1dc859fdd08e 8d59d2214c2362e7a9d185d80b6 0cba55e237ba61489c0a29f7d27 ---------------- --------------------------- --------------------------- %stddev %change %stddev %change %stddev \ | \ | \ 763.24 -1.2% 754.22 -2.0% 748.22 turbostat.PkgWatt 4560 ± 2% -19.4% 3673 +7.1% 4882 ± 2% vmstat.system.cs 0.03 ± 3% -0.0 0.02 -0.0 0.03 ± 2% mpstat.cpu.all.soft% 0.13 ± 2% -0.0 0.10 ± 2% -0.0 0.13 mpstat.cpu.all.usr% 293.00 ± 10% +54.7% 453.17 ± 12% +7.6% 315.33 ± 3% perf-c2c.DRAM.local 3720 ± 3% +41.1% 5251 ± 3% +0.8% 3752 ± 3% perf-c2c.DRAM.remote 325.67 ± 5% -21.5% 255.50 ± 7% +2.5% 333.83 ± 3% perf-c2c.HITM.remote 5426049 -33.8% 3590953 +3.3% 5605429 will-it-scale.224.processes 24222 -33.8% 16030 +3.3% 25023 will-it-scale.per_process_ops 5426049 -33.8% 3590953 +3.3% 5605429 will-it-scale.workload 148965 ± 9% +4.5% 155664 ± 20% -14.2% 127883 ± 10% numa-meminfo.node0.Slab 41751 ± 62% -31.7% 28502 ±122% -66.6% 13962 ±108% numa-meminfo.node1.Active 41727 ± 62% -31.7% 28502 ±122% -66.6% 13948 ±108% numa-meminfo.node1.Active(anon) 69062 ± 38% -19.5% 55596 ± 63% -39.1% 42090 ± 35% numa-meminfo.node1.Shmem 355193 ± 3% +16.1% 412516 ± 4% -3.3% 343648 ± 4% sched_debug.cfs_rq:/.avg_vruntime.stddev 355191 ± 3% +16.1% 412513 ± 4% -3.2% 343648 ± 4% sched_debug.cfs_rq:/.min_vruntime.stddev 89.04 ± 9% +15.9% 103.22 ± 9% +3.2% 91.93 ± 11% sched_debug.cfs_rq:/.runnable_avg.stddev 4289 -13.9% 3693 +4.9% 4498 sched_debug.cpu.nr_switches.avg 2259 ± 3% -25.1% 1693 ± 2% +5.7% 2388 ± 5% sched_debug.cpu.nr_switches.min 44536 -5.9% 41918 +1.5% 45191 proc-vmstat.nr_slab_reclaimable 3.257e+09 -33.9% 2.153e+09 +3.3% 3.366e+09 proc-vmstat.numa_hit 3.256e+09 -33.9% 2.152e+09 +3.3% 3.365e+09 proc-vmstat.numa_local 10269 ± 45% +87.3% 19237 ± 14% +83.8% 18876 ± 39% proc-vmstat.numa_pages_migrated 3.257e+09 -33.9% 2.153e+09 +3.3% 3.365e+09 proc-vmstat.pgalloc_normal 3.257e+09 -33.9% 2.153e+09 +3.3% 3.365e+09 proc-vmstat.pgfree 10269 ± 45% +87.3% 19237 ± 14% +83.8% 18876 ± 39% proc-vmstat.pgmigrate_success 7.906e+08 ± 4% -32.9% 5.303e+08 ± 2% +3.5% 8.181e+08 ± 4% numa-numastat.node0.local_node 7.909e+08 ± 4% -32.9% 5.305e+08 ± 2% +3.5% 8.184e+08 ± 4% numa-numastat.node0.numa_hit 8.069e+08 ± 3% -33.6% 5.361e+08 ± 2% +6.0% 8.552e+08 ± 2% numa-numastat.node1.local_node 8.072e+08 ± 3% -33.6% 5.363e+08 ± 2% +6.0% 8.556e+08 ± 2% numa-numastat.node1.numa_hit 101456 -21.4% 79695 ± 38% -33.4% 67613 ± 38% numa-numastat.node1.other_node 8.276e+08 -34.1% 5.457e+08 ± 2% +2.8% 8.508e+08 numa-numastat.node2.local_node 8.278e+08 -34.1% 5.459e+08 ± 2% +2.8% 8.511e+08 numa-numastat.node2.numa_hit 8.31e+08 -35.0% 5.403e+08 +1.1% 8.406e+08 ± 3% numa-numastat.node3.local_node 8.314e+08 -35.0% 5.404e+08 +1.2% 8.409e+08 ± 3% numa-numastat.node3.numa_hit 7.909e+08 ± 4% -32.9% 5.305e+08 ± 2% +3.5% 8.184e+08 ± 4% numa-vmstat.node0.numa_hit 7.906e+08 ± 4% -32.9% 5.303e+08 ± 2% +3.5% 8.181e+08 ± 4% numa-vmstat.node0.numa_local 10428 ± 62% -31.6% 7130 ±122% -66.6% 3486 ±108% numa-vmstat.node1.nr_active_anon 17331 ± 38% -19.0% 14042 ± 63% -37.6% 10816 ± 33% numa-vmstat.node1.nr_shmem 10428 ± 62% -31.6% 7130 ±122% -66.6% 3486 ±108% numa-vmstat.node1.nr_zone_active_anon 8.072e+08 ± 3% -33.6% 5.363e+08 ± 2% +6.0% 8.556e+08 ± 2% numa-vmstat.node1.numa_hit 8.069e+08 ± 3% -33.6% 5.361e+08 ± 2% +6.0% 8.552e+08 ± 2% numa-vmstat.node1.numa_local 101455 -21.4% 79693 ± 38% -33.4% 67613 ± 38% numa-vmstat.node1.numa_other 8.278e+08 -34.1% 5.459e+08 ± 2% +2.8% 8.511e+08 numa-vmstat.node2.numa_hit 8.276e+08 -34.1% 5.457e+08 ± 2% +2.8% 8.508e+08 numa-vmstat.node2.numa_local 8.314e+08 -35.0% 5.404e+08 +1.2% 8.409e+08 ± 3% numa-vmstat.node3.numa_hit 8.31e+08 -35.0% 5.403e+08 +1.1% 8.406e+08 ± 3% numa-vmstat.node3.numa_local 0.10 ± 8% +135.1% 0.24 ± 10% +32.0% 0.13 ± 23% perf-sched.sch_delay.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64 0.04 ± 11% +42.4% 0.06 ± 16% +13.6% 0.04 ± 20% perf-sched.sch_delay.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64 0.06 ± 33% +112.1% 0.14 ± 25% -32.0% 0.04 ± 47% perf-sched.sch_delay.avg.ms.syslog_print.do_syslog.kmsg_read.vfs_read 0.06 ± 46% +447.4% 0.31 ± 92% +8.2% 0.06 ± 38% perf-sched.sch_delay.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 0.09 ± 33% +82.8% 0.16 ± 25% -21.4% 0.07 ± 53% perf-sched.sch_delay.max.ms.syslog_print.do_syslog.kmsg_read.vfs_read 0.03 ± 6% +32.9% 0.04 ± 7% -6.1% 0.03 ± 11% perf-sched.total_sch_delay.average.ms 139.63 ± 4% +21.7% 169.99 ± 3% -9.8% 125.97 ± 3% perf-sched.total_wait_and_delay.average.ms 31780 ± 8% -19.0% 25751 ± 14% -5.8% 29937 ± 14% perf-sched.total_wait_and_delay.count.ms 139.60 ± 4% +21.7% 169.95 ± 3% -9.8% 125.94 ± 3% perf-sched.total_wait_time.average.ms 0.18 ± 6% +19.2% 0.22 ± 21% -14.5% 0.16 ± 11% perf-sched.wait_and_delay.avg.ms.__cond_resched.shmem_inode_acct_blocks.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fallocate 3.52 ± 5% +13.4% 3.99 ± 2% -0.3% 3.51 ± 4% perf-sched.wait_and_delay.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64 0.45 ±223% +821.8% 4.15 ± 9% -100.0% 0.00 perf-sched.wait_and_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone 305.95 ± 7% +44.4% 441.73 ± 4% -14.7% 260.96 ± 4% perf-sched.wait_and_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 6913 ± 6% -16.5% 5771 ± 13% -0.4% 6884 ± 13% perf-sched.wait_and_delay.count.__cond_resched.shmem_fallocate.vfs_fallocate.__x64_sys_fallocate.do_syscall_64 1974 ± 11% -42.6% 1132 ± 16% -16.3% 1651 ± 17% perf-sched.wait_and_delay.count.__cond_resched.shmem_inode_acct_blocks.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fallocate 1602 ± 7% +2.7% 1646 ± 13% -14.3% 1373 ± 12% perf-sched.wait_and_delay.count.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread 9474 ± 11% -33.5% 6303 ± 13% +0.5% 9524 ± 15% perf-sched.wait_and_delay.count.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 2.19 ±223% +770.9% 19.04 ± 63% -100.0% 0.00 perf-sched.wait_and_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone 1233 ± 30% +163.1% 3245 ± 26% +0.9% 1243 ± 30% perf-sched.wait_and_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 0.18 ± 6% +19.2% 0.22 ± 21% -14.5% 0.16 ± 11% perf-sched.wait_time.avg.ms.__cond_resched.shmem_inode_acct_blocks.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fallocate 104.96 ± 11% +50.8% 158.31 ± 14% -22.0% 81.88 ± 6% perf-sched.wait_time.avg.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 0.02 ±186% +985.0% 0.18 ± 33% -100.0% 0.00 perf-sched.wait_time.avg.ms.__cond_resched.unmap_vmas.exit_mmap.__mmput.exit_mm 3.41 ± 5% +9.8% 3.75 ± 3% -1.2% 3.37 ± 4% perf-sched.wait_time.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64 2.38 ± 6% +65.7% 3.95 ± 9% +1.4% 2.42 ± 10% perf-sched.wait_time.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone 305.93 ± 7% +44.4% 441.71 ± 4% -14.7% 260.94 ± 4% perf-sched.wait_time.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 0.07 ± 12% +59.5% 0.11 ± 27% +28.5% 0.09 ± 44% perf-sched.wait_time.avg.ms.wait_for_partner.fifo_open.do_dentry_open.do_open 361.49 ± 10% +163.6% 952.71 ± 24% -10.2% 324.59 ± 12% perf-sched.wait_time.max.ms.__cond_resched.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 0.02 ±186% +1370.0% 0.24 ± 20% -100.0% 0.00 perf-sched.wait_time.max.ms.__cond_resched.unmap_vmas.exit_mmap.__mmput.exit_mm 1233 ± 30% +163.1% 3245 ± 26% +0.9% 1243 ± 30% perf-sched.wait_time.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm 1.51 +70.8% 2.58 -5.4% 1.43 ± 2% perf-stat.i.MPKI 1.364e+10 -19.8% 1.094e+10 +6.4% 1.451e+10 perf-stat.i.branch-instructions 0.29 -0.0 0.25 -0.0 0.27 perf-stat.i.branch-miss-rate% 39037567 -29.6% 27478165 +1.8% 39751409 perf-stat.i.branch-misses 26.72 +7.0 33.67 +0.7 27.42 perf-stat.i.cache-miss-rate% 97210743 +34.0% 1.302e+08 ± 2% +0.3% 97544414 ± 2% perf-stat.i.cache-misses 3.641e+08 +6.3% 3.868e+08 ± 2% -2.2% 3.559e+08 ± 2% perf-stat.i.cache-references 4452 ± 2% -20.0% 3561 +7.5% 4784 ± 2% perf-stat.i.context-switches 13.13 +27.5% 16.74 -5.7% 12.38 perf-stat.i.cpi 270.31 -1.6% 265.85 +0.4% 271.43 perf-stat.i.cpu-migrations 8711 -25.3% 6504 -0.3% 8685 ± 2% perf-stat.i.cycles-between-cache-misses 1.66e+10 -21.1% 1.31e+10 +6.8% 1.774e+10 perf-stat.i.dTLB-loads 7.758e+09 -31.1% 5.343e+09 +6.3% 8.251e+09 perf-stat.i.dTLB-stores 12549015 -38.5% 7719822 +0.5% 12615766 perf-stat.i.iTLB-load-misses 6.454e+10 -21.6% 5.06e+10 +6.1% 6.846e+10 perf-stat.i.instructions 5208 +29.5% 6745 +5.4% 5491 perf-stat.i.instructions-per-iTLB-miss 0.08 -21.6% 0.06 +6.1% 0.08 perf-stat.i.ipc 0.36 ± 6% -24.6% 0.27 ± 25% -23.7% 0.27 ± 25% perf-stat.i.major-faults 86.32 ± 2% +27.6% 110.14 +2.1% 88.10 perf-stat.i.metric.K/sec 171.24 -22.4% 132.88 +6.5% 182.34 perf-stat.i.metric.M/sec 14793159 +36.3% 20167559 +0.9% 14924992 perf-stat.i.node-load-misses 1101912 ± 7% +48.5% 1636628 ± 4% +2.6% 1130608 ± 9% perf-stat.i.node-loads 3340101 ± 2% -19.8% 2679120 +6.9% 3571816 perf-stat.i.node-store-misses 84773 ± 5% -20.6% 67339 ± 6% +2.0% 86484 ± 5% perf-stat.i.node-stores 1.51 +70.8% 2.57 -5.4% 1.42 ± 2% perf-stat.overall.MPKI 0.29 -0.0 0.25 -0.0 0.27 perf-stat.overall.branch-miss-rate% 26.69 +6.9 33.63 +0.7 27.39 perf-stat.overall.cache-miss-rate% 13.12 +27.5% 16.73 -5.7% 12.37 perf-stat.overall.cpi 8709 -25.3% 6503 ± 2% -0.3% 8682 ± 2% perf-stat.overall.cycles-between-cache-misses 5146 +27.5% 6563 +5.5% 5430 perf-stat.overall.instructions-per-iTLB-miss 0.08 -21.6% 0.06 +6.1% 0.08 perf-stat.overall.ipc 3581676 +18.4% 4239733 +2.6% 3673713 perf-stat.overall.path-length 1.359e+10 -19.8% 1.091e+10 +6.4% 1.446e+10 perf-stat.ps.branch-instructions 38876130 -29.7% 27341584 +1.8% 39577054 perf-stat.ps.branch-misses 96879835 +34.0% 1.298e+08 ± 2% +0.3% 97215764 ± 2% perf-stat.ps.cache-misses 3.63e+08 +6.3% 3.859e+08 ± 2% -2.2% 3.549e+08 ± 2% perf-stat.ps.cache-references 4434 ± 2% -20.0% 3547 +7.4% 4764 ± 2% perf-stat.ps.context-switches 268.37 -1.9% 263.37 +0.3% 269.14 perf-stat.ps.cpu-migrations 1.655e+10 -21.1% 1.305e+10 +6.8% 1.768e+10 perf-stat.ps.dTLB-loads 7.733e+09 -31.1% 5.325e+09 +6.3% 8.223e+09 perf-stat.ps.dTLB-stores 12499097 -38.5% 7684522 +0.5% 12563331 perf-stat.ps.iTLB-load-misses 6.433e+10 -21.6% 5.044e+10 +6.1% 6.823e+10 perf-stat.ps.instructions 0.34 ± 6% -25.9% 0.25 ± 25% -23.8% 0.26 ± 25% perf-stat.ps.major-faults 14743590 +36.3% 20098836 +0.9% 14874764 perf-stat.ps.node-load-misses 1098750 ± 7% +48.7% 1633532 ± 4% +2.7% 1128235 ± 9% perf-stat.ps.node-loads 3328886 ± 2% -19.8% 2670192 +6.9% 3559593 perf-stat.ps.node-store-misses 84559 ± 5% -20.6% 67163 ± 6% +1.9% 86147 ± 5% perf-stat.ps.node-stores 1.943e+13 -21.7% 1.522e+13 +6.0% 2.059e+13 perf-stat.total.instructions 9.91 ± 10% -3.8 6.10 ± 4% -1.4 8.53 ± 11% perf-profile.calltrace.cycles-pp.__mem_cgroup_charge.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fallocate.vfs_fallocate 4.47 ± 10% -2.3 2.19 ± 4% -0.6 3.84 ± 11% perf-profile.calltrace.cycles-pp.get_mem_cgroup_from_mm.__mem_cgroup_charge.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fallocate 58.11 -2.1 56.01 -1.1 57.01 perf-profile.calltrace.cycles-pp.fallocate64 58.02 -2.1 55.95 -1.1 56.91 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.fallocate64 58.00 -2.1 55.94 -1.1 56.90 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.fallocate64 57.96 -2.1 55.91 -1.1 56.85 perf-profile.calltrace.cycles-pp.__x64_sys_fallocate.do_syscall_64.entry_SYSCALL_64_after_hwframe.fallocate64 57.92 -2.0 55.89 -1.1 56.82 perf-profile.calltrace.cycles-pp.vfs_fallocate.__x64_sys_fallocate.do_syscall_64.entry_SYSCALL_64_after_hwframe.fallocate64 57.82 -2.0 55.83 -1.1 56.72 perf-profile.calltrace.cycles-pp.shmem_fallocate.vfs_fallocate.__x64_sys_fallocate.do_syscall_64.entry_SYSCALL_64_after_hwframe 57.47 -1.8 55.62 -1.1 56.40 perf-profile.calltrace.cycles-pp.shmem_get_folio_gfp.shmem_fallocate.vfs_fallocate.__x64_sys_fallocate.do_syscall_64 57.30 -1.8 55.53 -1.1 56.22 perf-profile.calltrace.cycles-pp.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fallocate.vfs_fallocate.__x64_sys_fallocate 2.17 ± 4% -1.0 1.14 ± 3% -0.1 2.06 ± 4% perf-profile.calltrace.cycles-pp.__mem_cgroup_uncharge_list.release_pages.__folio_batch_release.shmem_undo_range.shmem_setattr 3.08 ± 9% -0.9 2.19 ± 4% -0.4 2.64 ± 11% perf-profile.calltrace.cycles-pp.mem_cgroup_commit_charge.__mem_cgroup_charge.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fallocate 1.29 ± 6% -0.7 0.54 ± 4% -0.1 1.16 ± 6% perf-profile.calltrace.cycles-pp.uncharge_folio.__mem_cgroup_uncharge_list.release_pages.__folio_batch_release.shmem_undo_range 0.88 ± 2% -0.3 0.59 ± 2% +0.0 0.90 ± 2% perf-profile.calltrace.cycles-pp.uncharge_batch.__mem_cgroup_uncharge_list.release_pages.__folio_batch_release.shmem_undo_range 1.66 -0.0 1.63 ± 3% +0.0 1.69 perf-profile.calltrace.cycles-pp.lru_add_drain_cpu.__folio_batch_release.shmem_undo_range.shmem_setattr.notify_change 1.64 -0.0 1.62 ± 3% +0.0 1.68 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu 1.66 -0.0 1.63 ± 3% +0.0 1.69 perf-profile.calltrace.cycles-pp.folio_batch_move_lru.lru_add_drain_cpu.__folio_batch_release.shmem_undo_range.shmem_setattr 0.80 +0.1 0.86 ± 2% -0.0 0.78 ± 2% perf-profile.calltrace.cycles-pp.lru_add_fn.folio_batch_move_lru.folio_add_lru.shmem_alloc_and_add_folio.shmem_get_folio_gfp 0.61 ± 2% +0.1 0.74 ± 2% -0.0 0.58 ± 3% perf-profile.calltrace.cycles-pp.__mod_memcg_lruvec_state.lru_add_fn.folio_batch_move_lru.folio_add_lru.shmem_alloc_and_add_folio 1.65 ± 2% +0.2 1.85 +0.0 1.67 ± 2% perf-profile.calltrace.cycles-pp.__mod_memcg_lruvec_state.__mod_lruvec_page_state.filemap_unaccount_folio.__filemap_remove_folio.filemap_remove_folio 1.44 ± 3% +0.4 1.79 ± 3% -0.1 1.34 ± 4% perf-profile.calltrace.cycles-pp.__mod_memcg_lruvec_state.__mod_lruvec_page_state.shmem_add_to_page_cache.shmem_alloc_and_add_folio.shmem_get_folio_gfp 0.08 ±223% +0.5 0.60 ± 2% +0.1 0.17 ±141% perf-profile.calltrace.cycles-pp.__mod_memcg_lruvec_state.release_pages.__folio_batch_release.shmem_undo_range.shmem_setattr 0.00 +0.9 0.86 ± 4% +0.0 0.00 perf-profile.calltrace.cycles-pp.__count_memcg_events.mem_cgroup_commit_charge.__mem_cgroup_charge.shmem_alloc_and_add_folio.shmem_get_folio_gfp 41.70 +2.1 43.82 ± 2% +1.1 42.82 ± 2% perf-profile.calltrace.cycles-pp.ftruncate64 41.68 +2.1 43.81 ± 2% +1.1 42.80 ± 2% perf-profile.calltrace.cycles-pp.do_sys_ftruncate.do_syscall_64.entry_SYSCALL_64_after_hwframe.ftruncate64 41.68 +2.1 43.81 ± 2% +1.1 42.80 ± 2% perf-profile.calltrace.cycles-pp.do_truncate.do_sys_ftruncate.do_syscall_64.entry_SYSCALL_64_after_hwframe.ftruncate64 41.68 +2.1 43.80 ± 2% +1.1 42.80 ± 2% perf-profile.calltrace.cycles-pp.notify_change.do_truncate.do_sys_ftruncate.do_syscall_64.entry_SYSCALL_64_after_hwframe 41.69 +2.1 43.82 ± 2% +1.1 42.81 ± 2% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.ftruncate64 41.69 +2.1 43.82 ± 2% +1.1 42.81 ± 2% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.ftruncate64 41.67 +2.1 43.80 ± 2% +1.1 42.80 ± 2% perf-profile.calltrace.cycles-pp.shmem_setattr.notify_change.do_truncate.do_sys_ftruncate.do_syscall_64 41.67 +2.1 43.80 ± 2% +1.1 42.79 ± 2% perf-profile.calltrace.cycles-pp.shmem_undo_range.shmem_setattr.notify_change.do_truncate.do_sys_ftruncate 38.67 +2.3 40.97 ± 2% +1.0 39.68 ± 2% perf-profile.calltrace.cycles-pp.__folio_batch_release.shmem_undo_range.shmem_setattr.notify_change.do_truncate 36.98 +2.3 39.32 ± 2% +1.0 37.96 ± 2% perf-profile.calltrace.cycles-pp.release_pages.__folio_batch_release.shmem_undo_range.shmem_setattr.notify_change 44.10 +2.4 46.47 ± 2% +0.4 44.48 perf-profile.calltrace.cycles-pp.folio_add_lru.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fallocate.vfs_fallocate 44.04 +2.4 46.42 ± 2% +0.4 44.42 perf-profile.calltrace.cycles-pp.folio_batch_move_lru.folio_add_lru.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fallocate 42.89 +2.4 45.32 ± 2% +0.4 43.29 perf-profile.calltrace.cycles-pp.folio_lruvec_lock_irqsave.folio_batch_move_lru.folio_add_lru.shmem_alloc_and_add_folio.shmem_get_folio_gfp 42.87 +2.4 45.31 ± 2% +0.4 43.27 perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.folio_batch_move_lru.folio_add_lru.shmem_alloc_and_add_folio 42.84 +2.4 45.29 ± 2% +0.4 43.24 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.folio_batch_move_lru.folio_add_lru 33.96 +3.4 37.31 ± 2% +1.1 35.02 ± 2% perf-profile.calltrace.cycles-pp.folio_lruvec_lock_irqsave.release_pages.__folio_batch_release.shmem_undo_range.shmem_setattr 33.94 +3.4 37.30 ± 2% +1.1 35.00 ± 2% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.release_pages.__folio_batch_release.shmem_undo_range 33.92 +3.4 37.28 ± 2% +1.1 34.98 ± 2% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.release_pages.__folio_batch_release 9.93 ± 10% -3.8 6.10 ± 4% -1.4 8.54 ± 11% perf-profile.children.cycles-pp.__mem_cgroup_charge 4.48 ± 10% -2.3 2.20 ± 4% -0.6 3.84 ± 11% perf-profile.children.cycles-pp.get_mem_cgroup_from_mm 58.14 -2.1 56.03 -1.1 57.04 perf-profile.children.cycles-pp.fallocate64 57.96 -2.1 55.91 -1.1 56.85 perf-profile.children.cycles-pp.__x64_sys_fallocate 57.92 -2.0 55.89 -1.1 56.82 perf-profile.children.cycles-pp.vfs_fallocate 57.82 -2.0 55.83 -1.1 56.73 perf-profile.children.cycles-pp.shmem_fallocate 57.53 -1.8 55.69 -1.1 56.44 perf-profile.children.cycles-pp.shmem_get_folio_gfp 57.36 -1.8 55.60 -1.1 56.27 perf-profile.children.cycles-pp.shmem_alloc_and_add_folio 2.18 ± 4% -1.0 1.14 ± 3% -0.1 2.07 ± 4% perf-profile.children.cycles-pp.__mem_cgroup_uncharge_list 3.09 ± 9% -0.9 2.19 ± 4% -0.4 2.64 ± 11% perf-profile.children.cycles-pp.mem_cgroup_commit_charge 1.29 ± 6% -0.7 0.54 ± 4% -0.1 1.16 ± 6% perf-profile.children.cycles-pp.uncharge_folio 0.88 ± 2% -0.3 0.59 ± 2% +0.0 0.90 ± 2% perf-profile.children.cycles-pp.uncharge_batch 0.36 -0.1 0.22 ± 2% +0.0 0.36 perf-profile.children.cycles-pp.shmem_alloc_folio 0.36 ± 2% -0.1 0.23 ± 2% +0.0 0.37 perf-profile.children.cycles-pp.xas_store 0.32 ± 2% -0.1 0.20 ± 2% +0.0 0.32 perf-profile.children.cycles-pp.alloc_pages_mpol 0.27 ± 2% -0.1 0.16 ± 4% -0.0 0.27 ± 2% perf-profile.children.cycles-pp.shmem_inode_acct_blocks 0.27 -0.1 0.17 ± 3% +0.0 0.28 perf-profile.children.cycles-pp.__alloc_pages 0.37 ± 4% -0.1 0.29 +0.0 0.40 perf-profile.children.cycles-pp.page_counter_uncharge 0.18 -0.1 0.11 ± 4% +0.0 0.18 ± 3% perf-profile.children.cycles-pp.get_page_from_freelist 0.16 ± 3% -0.1 0.09 +0.0 0.16 ± 4% perf-profile.children.cycles-pp.xas_load 0.18 ± 2% -0.1 0.12 ± 4% +0.0 0.18 ± 2% perf-profile.children.cycles-pp.__mod_lruvec_state 0.14 ± 2% -0.1 0.09 +0.0 0.15 ± 3% perf-profile.children.cycles-pp._raw_spin_lock 0.18 ± 3% -0.1 0.13 ± 4% +0.0 0.19 ± 3% perf-profile.children.cycles-pp.try_charge_memcg 0.09 ± 10% -0.0 0.04 ± 73% +0.0 0.09 ± 12% perf-profile.children.cycles-pp._raw_spin_lock_irq 0.12 -0.0 0.07 +0.0 0.12 perf-profile.children.cycles-pp.__dquot_alloc_space 0.11 -0.0 0.06 ± 6% +0.0 0.11 perf-profile.children.cycles-pp.filemap_get_entry 0.13 ± 2% -0.0 0.09 ± 4% +0.0 0.14 perf-profile.children.cycles-pp.__mod_node_page_state 0.10 ± 3% -0.0 0.06 ± 9% +0.0 0.10 ± 4% perf-profile.children.cycles-pp.xas_descend 0.12 -0.0 0.08 +0.0 0.12 ± 4% perf-profile.children.cycles-pp.free_unref_page_list 0.11 ± 3% -0.0 0.07 +0.0 0.12 ± 4% perf-profile.children.cycles-pp.rmqueue 0.10 ± 35% -0.0 0.06 -0.0 0.08 ± 8% perf-profile.children.cycles-pp.cgroup_rstat_updated 0.10 -0.0 0.06 ± 7% +0.0 0.10 ± 4% perf-profile.children.cycles-pp.xas_clear_mark 0.18 -0.0 0.14 ± 3% +0.0 0.18 ± 2% perf-profile.children.cycles-pp.find_lock_entries 0.16 ± 4% -0.0 0.13 ± 2% +0.0 0.17 ± 6% perf-profile.children.cycles-pp.propagate_protected_usage 0.10 -0.0 0.07 ± 5% +0.0 0.10 ± 4% perf-profile.children.cycles-pp.truncate_cleanup_folio 0.08 ± 4% -0.0 0.05 ± 7% +0.0 0.08 perf-profile.children.cycles-pp.xas_init_marks 0.09 ± 4% -0.0 0.06 ± 7% +0.0 0.09 ± 5% perf-profile.children.cycles-pp.page_counter_try_charge 0.18 ± 2% -0.0 0.16 -0.0 0.18 perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt 0.14 ± 2% -0.0 0.13 ± 2% -0.0 0.14 perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt 0.14 ± 2% -0.0 0.13 ± 2% -0.0 0.14 perf-profile.children.cycles-pp.hrtimer_interrupt 0.09 -0.0 0.08 +0.0 0.09 perf-profile.children.cycles-pp.tick_sched_handle 0.00 +0.0 0.00 +0.1 0.06 ± 6% perf-profile.children.cycles-pp.mem_cgroup_update_lru_size 0.82 +0.1 0.87 -0.0 0.79 ± 2% perf-profile.children.cycles-pp.lru_add_fn 99.81 +0.1 99.89 +0.0 99.81 perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe 99.79 +0.1 99.88 +0.0 99.80 perf-profile.children.cycles-pp.do_syscall_64 0.51 +0.5 0.98 ± 3% +0.0 0.53 ± 5% perf-profile.children.cycles-pp.__count_memcg_events 4.21 +0.8 5.01 -0.1 4.12 perf-profile.children.cycles-pp.__mod_memcg_lruvec_state 41.68 +2.1 43.81 ± 2% +1.1 42.80 ± 2% perf-profile.children.cycles-pp.do_sys_ftruncate 41.70 +2.1 43.82 ± 2% +1.1 42.82 ± 2% perf-profile.children.cycles-pp.ftruncate64 41.68 +2.1 43.81 ± 2% +1.1 42.80 ± 2% perf-profile.children.cycles-pp.do_truncate 41.68 +2.1 43.80 ± 2% +1.1 42.80 ± 2% perf-profile.children.cycles-pp.notify_change 41.67 +2.1 43.80 ± 2% +1.1 42.80 ± 2% perf-profile.children.cycles-pp.shmem_setattr 41.67 +2.1 43.81 ± 2% +1.1 42.80 ± 2% perf-profile.children.cycles-pp.shmem_undo_range 38.67 +2.3 40.98 ± 2% +1.0 39.68 ± 2% perf-profile.children.cycles-pp.__folio_batch_release 37.07 +2.3 39.39 ± 2% +1.0 38.05 ± 2% perf-profile.children.cycles-pp.release_pages 45.77 +2.4 48.14 +0.4 46.17 perf-profile.children.cycles-pp.folio_batch_move_lru 44.14 +2.4 46.52 ± 2% +0.4 44.51 perf-profile.children.cycles-pp.folio_add_lru 78.55 +5.8 84.34 +1.5 80.04 perf-profile.children.cycles-pp.folio_lruvec_lock_irqsave 78.52 +5.8 84.31 +1.5 80.00 perf-profile.children.cycles-pp._raw_spin_lock_irqsave 78.48 +5.8 84.29 +1.5 79.96 perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath 4.47 ± 10% -2.3 2.19 ± 4% -0.6 3.83 ± 11% perf-profile.self.cycles-pp.get_mem_cgroup_from_mm 2.67 ± 11% -1.3 1.32 ± 4% -0.5 2.22 ± 12% perf-profile.self.cycles-pp.mem_cgroup_commit_charge 1.28 ± 6% -0.7 0.54 ± 4% -0.1 1.16 ± 6% perf-profile.self.cycles-pp.uncharge_folio 2.18 ± 11% -0.6 1.58 ± 5% -0.3 1.86 ± 12% perf-profile.self.cycles-pp.__mem_cgroup_charge 1.16 ± 8% -0.3 0.84 ± 10% +0.1 1.24 ± 14% perf-profile.self.cycles-pp.__mod_lruvec_page_state 0.38 ± 7% -0.2 0.18 ± 5% -0.0 0.35 ± 6% perf-profile.self.cycles-pp.uncharge_batch 0.24 ± 4% -0.1 0.16 ± 2% +0.0 0.24 perf-profile.self.cycles-pp.folio_batch_move_lru 0.18 ± 3% -0.1 0.11 ± 3% -0.0 0.18 perf-profile.self.cycles-pp.xas_store 0.19 -0.1 0.12 ± 3% -0.0 0.18 ± 2% perf-profile.self.cycles-pp.release_pages 0.14 ± 5% -0.1 0.08 ± 5% -0.0 0.14 ± 2% perf-profile.self.cycles-pp.lru_add_fn 0.23 ± 5% -0.1 0.17 ± 2% +0.0 0.25 ± 2% perf-profile.self.cycles-pp.page_counter_uncharge 0.11 -0.1 0.06 -0.0 0.10 ± 4% perf-profile.self.cycles-pp.shmem_fallocate 0.13 -0.0 0.08 ± 5% +0.0 0.13 ± 2% perf-profile.self.cycles-pp.__mod_node_page_state 0.14 ± 3% -0.0 0.09 +0.0 0.14 perf-profile.self.cycles-pp._raw_spin_lock 0.09 ± 5% -0.0 0.05 ± 7% +0.0 0.09 ± 5% perf-profile.self.cycles-pp.xas_descend 0.10 ± 3% -0.0 0.06 +0.0 0.10 ± 4% perf-profile.self.cycles-pp.shmem_add_to_page_cache 0.09 ± 37% -0.0 0.05 ± 7% -0.0 0.07 ± 5% perf-profile.self.cycles-pp.cgroup_rstat_updated 0.16 ± 4% -0.0 0.13 ± 2% +0.0 0.17 ± 6% perf-profile.self.cycles-pp.propagate_protected_usage 0.09 -0.0 0.06 +0.0 0.09 ± 5% perf-profile.self.cycles-pp.xas_clear_mark 0.09 ± 5% -0.0 0.06 ± 7% +0.0 0.09 ± 5% perf-profile.self.cycles-pp.try_charge_memcg 0.15 -0.0 0.12 ± 3% +0.0 0.15 ± 3% perf-profile.self.cycles-pp.find_lock_entries 0.07 ± 7% -0.0 0.05 +0.0 0.07 ± 6% perf-profile.self.cycles-pp.page_counter_try_charge 0.00 +0.0 0.00 +0.1 0.06 ± 8% perf-profile.self.cycles-pp.mem_cgroup_update_lru_size 0.50 ± 3% +0.5 0.97 ± 3% +0.0 0.52 ± 5% perf-profile.self.cycles-pp.__count_memcg_events 4.14 ± 2% +0.8 4.97 -0.1 4.06 perf-profile.self.cycles-pp.__mod_memcg_lruvec_state 78.48 +5.8 84.29 +1.5 79.96 perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath