From: kernel test robot <oliver.sang@intel.com>
To: Yosry Ahmed <yosryahmed@google.com>
Cc: <oe-lkp@lists.linux.dev>, <lkp@intel.com>,
<linux-kernel@vger.kernel.org>,
Andrew Morton <akpm@linux-foundation.org>,
kernel test robot <oliver.sang@intel.com>,
Shakeel Butt <shakeelb@google.com>,
Johannes Weiner <hannes@cmpxchg.org>,
Michal Hocko <mhocko@kernel.org>,
Muchun Song <muchun.song@linux.dev>,
Roman Gushchin <roman.gushchin@linux.dev>,
"Greg Thelen" <gthelen@google.com>, <cgroups@vger.kernel.org>,
<linux-mm@kvack.org>, <ying.huang@intel.com>,
<feng.tang@intel.com>, <fengwei.yin@intel.com>
Subject: [linus:master] [mm] 9cee7e8ef3: netperf.Throughput_Mbps 4.0% improvement
Date: Sun, 18 Feb 2024 21:16:24 +0800 [thread overview]
Message-ID: <202402182000.f21279e1-oliver.sang@intel.com> (raw)
hi, Yosry Ahmed,
we shared the performance impact of this commit in
https://lore.kernel.org/lkml/ZbDJsfsZt2ITyo61@xsang-OptiPlex-9020/
now we noticed the commit is merged in mainline, and we observed improvements
in other performance tests such like netperf and stress-ng.
the vm-scalability and will-it-scale results are also included FYI.
Hello,
kernel test robot noticed a 4.0% improvement of netperf.Throughput_Mbps on:
commit: 9cee7e8ef3e31ca25b40ca52b8585dc6935deff2 ("mm: memcg: optimize parent iteration in memcg_rstat_updated()")
https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master
testcase: netperf
test machine: 128 threads 2 sockets Intel(R) Xeon(R) Gold 6338 CPU @ 2.00GHz (Ice Lake) with 256G memory
parameters:
ip: ipv4
runtime: 300s
nr_threads: 200%
cluster: cs-localhost
send_size: 10K
test: TCP_SENDFILE
cpufreq_governor: performance
In addition to that, the commit also has significant impact on the following tests:
+------------------+----------------------------------------------------------------------------------------------------+
| testcase: change | stress-ng: stress-ng.metamix.ops_per_sec 4.1% improvement |
| test machine | 64 threads 2 sockets Intel(R) Xeon(R) Gold 6346 CPU @ 3.10GHz (Ice Lake) with 256G memory |
| test parameters | cpufreq_governor=performance |
| | disk=1HDD |
| | fs=xfs |
| | nr_threads=10% |
| | test=metamix |
| | testtime=60s |
+------------------+----------------------------------------------------------------------------------------------------+
| testcase: change | vm-scalability: vm-scalability.throughput 42.0% improvement |
| test machine | 224 threads 4 sockets Intel(R) Xeon(R) Platinum 8380H CPU @ 2.90GHz (Cooper Lake) with 192G memory |
| test parameters | cpufreq_governor=performance |
| | runtime=300s |
| | size=128G |
| | test=truncate |
+------------------+----------------------------------------------------------------------------------------------------+
| testcase: change | will-it-scale: will-it-scale.per_thread_ops 54.9% improvement |
| test machine | 224 threads 4 sockets Intel(R) Xeon(R) Platinum 8380H CPU @ 2.90GHz (Cooper Lake) with 192G memory |
| test parameters | cpufreq_governor=performance |
| | mode=thread |
| | nr_task=50% |
| | test=fallocate1 |
+------------------+----------------------------------------------------------------------------------------------------+
Details are as below:
-------------------------------------------------------------------------------------------------->
The kernel config and materials to reproduce are available at:
https://download.01.org/0day-ci/archive/20240218/202402182000.f21279e1-oliver.sang@intel.com
=========================================================================================
cluster/compiler/cpufreq_governor/ip/kconfig/nr_threads/rootfs/runtime/send_size/tbox_group/test/testcase:
cs-localhost/gcc-12/performance/ipv4/x86_64-rhel-8.3/200%/debian-11.1-x86_64-20220510.cgz/300s/10K/lkp-icl-2sp2/TCP_SENDFILE/netperf
commit:
67b8bcbaed ("nilfs2: fix data corruption in dsync block recovery for small block sizes")
9cee7e8ef3 ("mm: memcg: optimize parent iteration in memcg_rstat_updated()")
67b8bcbaed477787 9cee7e8ef3e31ca25b40ca52b85
---------------- ---------------------------
%stddev %change %stddev
\ | \
140392 ± 5% +9.2% 153362 ± 4% meminfo.DirectMap4k
772.17 ± 2% -19.0% 625.33 ± 4% perf-c2c.DRAM.remote
894.17 ± 3% -19.1% 723.17 ± 4% perf-c2c.HITM.local
-12.69 +55.8% -19.78 sched_debug.cpu.nr_uninterruptible.min
4.96 ± 8% +16.3% 5.77 ± 8% sched_debug.cpu.nr_uninterruptible.stddev
0.94 ± 2% -0.0 0.90 turbostat.C1%
34.22 -4.4% 32.70 ± 2% turbostat.RAMWatt
4939 +17.1% 5785 ± 6% perf-sched.total_wait_time.max.ms
1511 ± 32% -66.8% 502.34 ± 99% perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
4725 ± 7% +36.7% 6459 ± 25% perf-sched.wait_and_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
1506 ± 32% -66.6% 502.68 ± 99% perf-sched.wait_time.max.ms.schedule_hrtimeout_range_clock.do_poll.constprop.0.do_sys_poll
4725 ± 7% +22.1% 5771 ± 6% perf-sched.wait_time.max.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
9040 +4.0% 9402 netperf.ThroughputBoth_Mbps
2314243 +4.0% 2406983 netperf.ThroughputBoth_total_Mbps
9040 +4.0% 9402 netperf.Throughput_Mbps
2314243 +4.0% 2406983 netperf.Throughput_total_Mbps
1056 +3.9% 1098 netperf.time.user_time
15571304 +5.9% 16491497 netperf.time.voluntary_context_switches
8.475e+09 +4.0% 8.815e+09 netperf.workload
0.65 ± 2% -33.0% 0.44 ± 7% perf-stat.i.MPKI
4.977e+10 +3.2% 5.138e+10 perf-stat.i.branch-instructions
0.71 -0.0 0.68 perf-stat.i.branch-miss-rate%
20.77 ± 2% -3.3 17.49 ± 6% perf-stat.i.cache-miss-rate%
1.708e+08 ± 2% -30.9% 1.181e+08 ± 6% perf-stat.i.cache-misses
8.234e+08 -17.7% 6.776e+08 perf-stat.i.cache-references
1.25 -3.1% 1.21 perf-stat.i.cpi
1908 ± 2% +45.7% 2779 ± 7% perf-stat.i.cycles-between-cache-misses
7.258e+10 +3.1% 7.482e+10 perf-stat.i.dTLB-loads
4.018e+10 +3.2% 4.145e+10 perf-stat.i.dTLB-stores
2.608e+11 +3.2% 2.692e+11 perf-stat.i.instructions
0.80 +3.2% 0.83 perf-stat.i.ipc
1276 +3.0% 1315 perf-stat.i.metric.M/sec
15636176 ± 2% -19.5% 12582173 ± 5% perf-stat.i.node-load-misses
951084 ± 7% -39.3% 577496 ± 14% perf-stat.i.node-loads
48.91 ± 2% +5.6 54.54 ± 2% perf-stat.i.node-store-miss-rate%
0.66 ± 2% -33.0% 0.44 ± 7% perf-stat.overall.MPKI
0.70 -0.0 0.68 perf-stat.overall.branch-miss-rate%
20.75 ± 2% -3.3 17.43 ± 6% perf-stat.overall.cache-miss-rate%
1.25 -3.1% 1.21 perf-stat.overall.cpi
1903 ± 2% +45.3% 2766 ± 7% perf-stat.overall.cycles-between-cache-misses
0.80 +3.2% 0.83 perf-stat.overall.ipc
47.72 ± 3% +5.6 53.30 ± 3% perf-stat.overall.node-store-miss-rate%
4.961e+10 +3.2% 5.122e+10 perf-stat.ps.branch-instructions
1.703e+08 ± 2% -30.9% 1.177e+08 ± 6% perf-stat.ps.cache-misses
8.207e+08 -17.7% 6.754e+08 perf-stat.ps.cache-references
7.233e+10 +3.1% 7.457e+10 perf-stat.ps.dTLB-loads
4.005e+10 +3.2% 4.131e+10 perf-stat.ps.dTLB-stores
2.6e+11 +3.2% 2.683e+11 perf-stat.ps.instructions
15585093 ± 2% -19.5% 12543422 ± 5% perf-stat.ps.node-load-misses
947879 ± 7% -39.3% 575590 ± 14% perf-stat.ps.node-loads
7.848e+13 +3.1% 8.093e+13 perf-stat.total.instructions
3.80 ± 3% -2.1 1.71 ± 5% perf-profile.calltrace.cycles-pp.__mod_memcg_state.mem_cgroup_charge_skmem.__sk_mem_raise_allocated.__sk_mem_schedule.tcp_wmem_schedule
6.27 -2.0 4.24 ± 2% perf-profile.calltrace.cycles-pp.mem_cgroup_charge_skmem.__sk_mem_raise_allocated.__sk_mem_schedule.tcp_wmem_schedule.tcp_sendmsg_locked
46.14 -1.5 44.60 perf-profile.calltrace.cycles-pp.sock_sendmsg.splice_to_socket.direct_splice_actor.splice_direct_to_actor.do_splice_direct
44.24 -1.5 42.75 perf-profile.calltrace.cycles-pp.tcp_sendmsg.sock_sendmsg.splice_to_socket.direct_splice_actor.splice_direct_to_actor
11.92 -1.5 10.44 perf-profile.calltrace.cycles-pp.__sk_mem_raise_allocated.__sk_mem_schedule.tcp_wmem_schedule.tcp_sendmsg_locked.tcp_sendmsg
12.15 -1.5 10.68 perf-profile.calltrace.cycles-pp.tcp_wmem_schedule.tcp_sendmsg_locked.tcp_sendmsg.sock_sendmsg.splice_to_socket
12.02 -1.5 10.56 perf-profile.calltrace.cycles-pp.__sk_mem_schedule.tcp_wmem_schedule.tcp_sendmsg_locked.tcp_sendmsg.sock_sendmsg
50.08 -1.3 48.81 perf-profile.calltrace.cycles-pp.splice_to_socket.direct_splice_actor.splice_direct_to_actor.do_splice_direct.do_sendfile
50.52 -1.3 49.26 perf-profile.calltrace.cycles-pp.direct_splice_actor.splice_direct_to_actor.do_splice_direct.do_sendfile.__x64_sys_sendfile64
39.04 -1.2 37.88 perf-profile.calltrace.cycles-pp.tcp_sendmsg_locked.tcp_sendmsg.sock_sendmsg.splice_to_socket.direct_splice_actor
60.56 -0.8 59.77 perf-profile.calltrace.cycles-pp.splice_direct_to_actor.do_splice_direct.do_sendfile.__x64_sys_sendfile64.do_syscall_64
60.81 -0.8 60.02 perf-profile.calltrace.cycles-pp.do_splice_direct.do_sendfile.__x64_sys_sendfile64.do_syscall_64.entry_SYSCALL_64_after_hwframe
3.84 -0.7 3.14 perf-profile.calltrace.cycles-pp.tcp_try_rmem_schedule.tcp_data_queue.tcp_rcv_established.tcp_v4_do_rcv.tcp_v4_rcv
3.72 ± 5% -0.6 3.11 perf-profile.calltrace.cycles-pp.__sk_mem_schedule.tcp_try_rmem_schedule.tcp_data_queue.tcp_rcv_established.tcp_v4_do_rcv
64.97 -0.6 64.37 perf-profile.calltrace.cycles-pp.do_sendfile.__x64_sys_sendfile64.do_syscall_64.entry_SYSCALL_64_after_hwframe.sendfile
11.66 -0.5 11.15 perf-profile.calltrace.cycles-pp.ip_local_deliver_finish.__netif_receive_skb_one_core.process_backlog.__napi_poll.net_rx_action
11.61 -0.5 11.10 perf-profile.calltrace.cycles-pp.ip_protocol_deliver_rcu.ip_local_deliver_finish.__netif_receive_skb_one_core.process_backlog.__napi_poll
10.02 -0.5 9.52 perf-profile.calltrace.cycles-pp.tcp_v4_do_rcv.tcp_v4_rcv.ip_protocol_deliver_rcu.ip_local_deliver_finish.__netif_receive_skb_one_core
11.34 -0.5 10.85 perf-profile.calltrace.cycles-pp.tcp_v4_rcv.ip_protocol_deliver_rcu.ip_local_deliver_finish.__netif_receive_skb_one_core.process_backlog
9.51 -0.5 9.02 perf-profile.calltrace.cycles-pp.tcp_rcv_established.tcp_v4_do_rcv.tcp_v4_rcv.ip_protocol_deliver_rcu.ip_local_deliver_finish
12.61 -0.5 12.12 perf-profile.calltrace.cycles-pp.__netif_receive_skb_one_core.process_backlog.__napi_poll.net_rx_action.__do_softirq
13.00 -0.5 12.53 perf-profile.calltrace.cycles-pp.process_backlog.__napi_poll.net_rx_action.__do_softirq.do_softirq
13.06 -0.5 12.60 perf-profile.calltrace.cycles-pp.__napi_poll.net_rx_action.__do_softirq.do_softirq.__local_bh_enable_ip
14.22 -0.5 13.76 perf-profile.calltrace.cycles-pp.__local_bh_enable_ip.__dev_queue_xmit.ip_finish_output2.__ip_queue_xmit.__tcp_transmit_skb
14.11 -0.5 13.65 perf-profile.calltrace.cycles-pp.do_softirq.__local_bh_enable_ip.__dev_queue_xmit.ip_finish_output2.__ip_queue_xmit
13.30 -0.4 12.86 perf-profile.calltrace.cycles-pp.net_rx_action.__do_softirq.do_softirq.__local_bh_enable_ip.__dev_queue_xmit
13.96 -0.4 13.52 perf-profile.calltrace.cycles-pp.__do_softirq.do_softirq.__local_bh_enable_ip.__dev_queue_xmit.ip_finish_output2
1.88 -0.4 1.49 perf-profile.calltrace.cycles-pp.__sk_mem_reduce_allocated.tcp_recvmsg_locked.tcp_recvmsg.inet_recvmsg.sock_recvmsg
3.45 ± 5% -0.4 3.09 perf-profile.calltrace.cycles-pp.__sk_mem_raise_allocated.__sk_mem_schedule.tcp_try_rmem_schedule.tcp_data_queue.tcp_rcv_established
6.74 -0.3 6.39 perf-profile.calltrace.cycles-pp.tcp_data_queue.tcp_rcv_established.tcp_v4_do_rcv.tcp_v4_rcv.ip_protocol_deliver_rcu
3.53 -0.3 3.18 ± 2% perf-profile.calltrace.cycles-pp.__release_sock.release_sock.tcp_sendmsg.sock_sendmsg.splice_to_socket
4.11 -0.3 3.78 ± 2% perf-profile.calltrace.cycles-pp.release_sock.tcp_sendmsg.sock_sendmsg.splice_to_socket.direct_splice_actor
4.51 ± 2% -0.3 4.21 perf-profile.calltrace.cycles-pp.ip_finish_output2.__ip_queue_xmit.__tcp_transmit_skb.tcp_write_xmit.__tcp_push_pending_frames
2.43 ± 3% -0.3 2.15 perf-profile.calltrace.cycles-pp.mem_cgroup_charge_skmem.__sk_mem_raise_allocated.__sk_mem_schedule.tcp_try_rmem_schedule.tcp_data_queue
12.72 -0.3 12.46 perf-profile.calltrace.cycles-pp.__dev_queue_xmit.ip_finish_output2.__ip_queue_xmit.__tcp_transmit_skb.tcp_write_xmit
2.28 -0.2 2.04 ± 2% perf-profile.calltrace.cycles-pp.tcp_rcv_established.tcp_v4_do_rcv.__release_sock.release_sock.tcp_sendmsg
2.32 -0.2 2.08 ± 2% perf-profile.calltrace.cycles-pp.tcp_v4_do_rcv.__release_sock.release_sock.tcp_sendmsg.sock_sendmsg
3.50 ± 2% -0.2 3.32 perf-profile.calltrace.cycles-pp.__tcp_transmit_skb.tcp_write_xmit.__tcp_push_pending_frames.tcp_sendmsg_locked.tcp_sendmsg
3.22 ± 2% -0.2 3.03 perf-profile.calltrace.cycles-pp.__ip_queue_xmit.__tcp_transmit_skb.tcp_write_xmit.__tcp_push_pending_frames.tcp_sendmsg_locked
1.01 -0.2 0.83 perf-profile.calltrace.cycles-pp.tcp_ack.tcp_rcv_established.tcp_v4_do_rcv.__release_sock.release_sock
0.84 -0.2 0.67 perf-profile.calltrace.cycles-pp.tcp_clean_rtx_queue.tcp_ack.tcp_rcv_established.tcp_v4_do_rcv.__release_sock
4.10 -0.2 3.94 perf-profile.calltrace.cycles-pp.__tcp_transmit_skb.tcp_recvmsg_locked.tcp_recvmsg.inet_recvmsg.sock_recvmsg
1.91 ± 2% -0.2 1.76 ± 3% perf-profile.calltrace.cycles-pp.tcp_write_xmit.__tcp_push_pending_frames.tcp_rcv_established.tcp_v4_do_rcv.__release_sock
1.85 -0.2 1.70 ± 2% perf-profile.calltrace.cycles-pp.tcp_ack.tcp_rcv_established.tcp_v4_do_rcv.tcp_v4_rcv.ip_protocol_deliver_rcu
3.84 -0.2 3.69 perf-profile.calltrace.cycles-pp.__ip_queue_xmit.__tcp_transmit_skb.tcp_recvmsg_locked.tcp_recvmsg.inet_recvmsg
1.92 ± 2% -0.2 1.77 ± 3% perf-profile.calltrace.cycles-pp.__tcp_push_pending_frames.tcp_rcv_established.tcp_v4_do_rcv.__release_sock.release_sock
3.59 -0.2 3.44 perf-profile.calltrace.cycles-pp.ip_finish_output2.__ip_queue_xmit.__tcp_transmit_skb.tcp_recvmsg_locked.tcp_recvmsg
3.48 -0.1 3.34 perf-profile.calltrace.cycles-pp.__dev_queue_xmit.ip_finish_output2.__ip_queue_xmit.__tcp_transmit_skb.tcp_recvmsg_locked
1.72 ± 2% -0.1 1.58 ± 3% perf-profile.calltrace.cycles-pp.__tcp_transmit_skb.tcp_write_xmit.__tcp_push_pending_frames.tcp_rcv_established.tcp_v4_do_rcv
1.58 ± 2% -0.1 1.45 ± 3% perf-profile.calltrace.cycles-pp.__ip_queue_xmit.__tcp_transmit_skb.tcp_write_xmit.__tcp_push_pending_frames.tcp_rcv_established
0.55 -0.1 0.43 ± 44% perf-profile.calltrace.cycles-pp.ttwu_do_activate.try_to_wake_up.__wake_up_common.__wake_up_sync_key.sock_def_readable
1.83 -0.1 1.71 perf-profile.calltrace.cycles-pp.tcp_stream_alloc_skb.tcp_sendmsg_locked.tcp_sendmsg.sock_sendmsg.splice_to_socket
0.58 -0.0 0.55 perf-profile.calltrace.cycles-pp.schedule.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.sendfile
0.55 -0.0 0.53 ± 2% perf-profile.calltrace.cycles-pp.__schedule.schedule.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.59 -0.0 0.57 perf-profile.calltrace.cycles-pp.lock_sock_nested.tcp_sendmsg.sock_sendmsg.splice_to_socket.direct_splice_actor
0.64 +0.0 0.68 perf-profile.calltrace.cycles-pp.tcp_event_new_data_sent.tcp_write_xmit.tcp_sendmsg_locked.tcp_sendmsg.sock_sendmsg
0.78 +0.0 0.81 perf-profile.calltrace.cycles-pp._copy_from_user.__x64_sys_sendfile64.do_syscall_64.entry_SYSCALL_64_after_hwframe.sendfile
1.11 +0.0 1.14 perf-profile.calltrace.cycles-pp.tcp_send_mss.tcp_sendmsg_locked.tcp_sendmsg.sock_sendmsg.splice_to_socket
0.67 +0.0 0.70 perf-profile.calltrace.cycles-pp.atime_needs_update.touch_atime.splice_direct_to_actor.do_splice_direct.do_sendfile
0.82 +0.0 0.85 perf-profile.calltrace.cycles-pp.touch_atime.splice_direct_to_actor.do_splice_direct.do_sendfile.__x64_sys_sendfile64
0.66 +0.0 0.69 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64.sendfile.sendfile_tcp_stream.main.__libc_start_main
0.94 +0.0 0.98 perf-profile.calltrace.cycles-pp.__alloc_skb.tcp_stream_alloc_skb.tcp_sendmsg_locked.tcp_sendmsg.sock_sendmsg
0.52 +0.0 0.56 perf-profile.calltrace.cycles-pp.aa_sk_perm.security_socket_sendmsg.sock_sendmsg.splice_to_socket.direct_splice_actor
1.11 +0.0 1.16 perf-profile.calltrace.cycles-pp.rw_verify_area.do_sendfile.__x64_sys_sendfile64.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.70 +0.0 0.74 perf-profile.calltrace.cycles-pp.iov_iter_advance.iov_iter_extract_pages.skb_splice_from_iter.tcp_sendmsg_locked.tcp_sendmsg
0.95 ± 2% +0.0 1.00 ± 2% perf-profile.calltrace.cycles-pp.page_cache_pipe_buf_release.splice_to_socket.direct_splice_actor.splice_direct_to_actor.do_splice_direct
0.58 +0.1 0.64 perf-profile.calltrace.cycles-pp.atime_needs_update.touch_atime.filemap_splice_read.splice_direct_to_actor.do_splice_direct
1.21 +0.1 1.28 perf-profile.calltrace.cycles-pp.__fsnotify_parent.do_sendfile.__x64_sys_sendfile64.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.56 +0.1 0.62 perf-profile.calltrace.cycles-pp.netperf_sendfile.sendfile_tcp_stream.main.__libc_start_main
0.68 +0.1 0.75 perf-profile.calltrace.cycles-pp.touch_atime.filemap_splice_read.splice_direct_to_actor.do_splice_direct.do_sendfile
1.16 +0.1 1.24 perf-profile.calltrace.cycles-pp.release_pages.__folio_batch_release.filemap_splice_read.splice_direct_to_actor.do_splice_direct
1.62 +0.1 1.70 perf-profile.calltrace.cycles-pp.splice_folio_into_pipe.filemap_splice_read.splice_direct_to_actor.do_splice_direct.do_sendfile
1.57 +0.1 1.66 perf-profile.calltrace.cycles-pp.skb_append_pagefrags.skb_splice_from_iter.tcp_sendmsg_locked.tcp_sendmsg.sock_sendmsg
1.46 +0.1 1.54 perf-profile.calltrace.cycles-pp.__folio_batch_release.filemap_splice_read.splice_direct_to_actor.do_splice_direct.do_sendfile
2.42 +0.1 2.53 perf-profile.calltrace.cycles-pp.filemap_get_read_batch.filemap_get_pages.filemap_splice_read.splice_direct_to_actor.do_splice_direct
1.83 +0.1 1.95 perf-profile.calltrace.cycles-pp.iov_iter_extract_pages.skb_splice_from_iter.tcp_sendmsg_locked.tcp_sendmsg.sock_sendmsg
2.06 +0.1 2.19 perf-profile.calltrace.cycles-pp.simple_copy_to_iter.__skb_datagram_iter.skb_copy_datagram_iter.tcp_recvmsg_locked.tcp_recvmsg
0.72 ± 14% +0.1 0.85 perf-profile.calltrace.cycles-pp.skb_release_data.skb_attempt_defer_free.tcp_recvmsg_locked.tcp_recvmsg.inet_recvmsg
0.62 ± 12% +0.1 0.75 ± 4% perf-profile.calltrace.cycles-pp.drain_stock.refill_stock.__sk_mem_reduce_allocated.tcp_recvmsg_locked.tcp_recvmsg
2.76 +0.1 2.89 perf-profile.calltrace.cycles-pp.filemap_get_pages.filemap_splice_read.splice_direct_to_actor.do_splice_direct.do_sendfile
0.76 ± 14% +0.1 0.90 perf-profile.calltrace.cycles-pp.skb_attempt_defer_free.tcp_recvmsg_locked.tcp_recvmsg.inet_recvmsg.sock_recvmsg
0.74 ± 13% +0.1 0.88 ± 3% perf-profile.calltrace.cycles-pp.refill_stock.__sk_mem_reduce_allocated.tcp_recvmsg_locked.tcp_recvmsg.inet_recvmsg
0.66 ± 13% +0.1 0.80 ± 3% perf-profile.calltrace.cycles-pp.refill_stock.__sk_mem_reduce_allocated.tcp_clean_rtx_queue.tcp_ack.tcp_rcv_established
24.66 +0.2 24.83 perf-profile.calltrace.cycles-pp.tcp_recvmsg.inet_recvmsg.sock_recvmsg.__sys_recvfrom.__x64_sys_recvfrom
24.30 +0.2 24.47 perf-profile.calltrace.cycles-pp.tcp_recvmsg_locked.tcp_recvmsg.inet_recvmsg.sock_recvmsg.__sys_recvfrom
24.76 +0.2 24.94 perf-profile.calltrace.cycles-pp.inet_recvmsg.sock_recvmsg.__sys_recvfrom.__x64_sys_recvfrom.do_syscall_64
24.90 +0.2 25.07 perf-profile.calltrace.cycles-pp.sock_recvmsg.__sys_recvfrom.__x64_sys_recvfrom.do_syscall_64.entry_SYSCALL_64_after_hwframe
25.86 +0.2 26.03 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.recv.process_requests.spawn_child.accept_connection
1.60 ± 11% +0.2 1.78 perf-profile.calltrace.cycles-pp.__check_object_size.simple_copy_to_iter.__skb_datagram_iter.skb_copy_datagram_iter.tcp_recvmsg_locked
25.82 +0.2 26.00 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.recv.process_requests.spawn_child
1.01 ± 16% +0.2 1.20 perf-profile.calltrace.cycles-pp.check_heap_object.__check_object_size.simple_copy_to_iter.__skb_datagram_iter.skb_copy_datagram_iter
26.12 +0.2 26.30 perf-profile.calltrace.cycles-pp.recv.process_requests.spawn_child.accept_connection.accept_connections
0.35 ± 70% +0.2 0.55 ± 2% perf-profile.calltrace.cycles-pp.__virt_addr_valid.check_heap_object.__check_object_size.simple_copy_to_iter.__skb_datagram_iter
25.41 +0.2 25.61 perf-profile.calltrace.cycles-pp.__x64_sys_recvfrom.do_syscall_64.entry_SYSCALL_64_after_hwframe.recv.process_requests
25.36 +0.2 25.56 perf-profile.calltrace.cycles-pp.__sys_recvfrom.__x64_sys_recvfrom.do_syscall_64.entry_SYSCALL_64_after_hwframe.recv
0.52 ± 46% +0.2 0.73 ± 3% perf-profile.calltrace.cycles-pp.page_counter_uncharge.drain_stock.refill_stock.__sk_mem_reduce_allocated.tcp_clean_rtx_queue
0.52 ± 45% +0.2 0.74 ± 4% perf-profile.calltrace.cycles-pp.page_counter_uncharge.drain_stock.refill_stock.__sk_mem_reduce_allocated.tcp_recvmsg_locked
0.52 ± 46% +0.2 0.74 ± 4% perf-profile.calltrace.cycles-pp.drain_stock.refill_stock.__sk_mem_reduce_allocated.tcp_clean_rtx_queue.tcp_ack
5.10 +0.3 5.40 perf-profile.calltrace.cycles-pp.skb_splice_from_iter.tcp_sendmsg_locked.tcp_sendmsg.sock_sendmsg.splice_to_socket
10.48 +0.4 10.92 perf-profile.calltrace.cycles-pp._copy_to_iter.__skb_datagram_iter.skb_copy_datagram_iter.tcp_recvmsg_locked.tcp_recvmsg
8.16 +0.4 8.60 perf-profile.calltrace.cycles-pp.filemap_splice_read.splice_direct_to_actor.do_splice_direct.do_sendfile.__x64_sys_sendfile64
13.73 +0.6 14.36 perf-profile.calltrace.cycles-pp.__skb_datagram_iter.skb_copy_datagram_iter.tcp_recvmsg_locked.tcp_recvmsg.inet_recvmsg
13.84 +0.7 14.50 perf-profile.calltrace.cycles-pp.skb_copy_datagram_iter.tcp_recvmsg_locked.tcp_recvmsg.inet_recvmsg.sock_recvmsg
23.63 ± 16% +2.9 26.56 perf-profile.calltrace.cycles-pp.accept_connection.accept_connections.main.__libc_start_main
23.63 ± 16% +2.9 26.56 perf-profile.calltrace.cycles-pp.accept_connections.main.__libc_start_main
23.63 ± 16% +2.9 26.56 perf-profile.calltrace.cycles-pp.process_requests.spawn_child.accept_connection.accept_connections.main
23.63 ± 16% +2.9 26.56 perf-profile.calltrace.cycles-pp.spawn_child.accept_connection.accept_connections.main.__libc_start_main
6.60 ± 3% -3.6 3.00 ± 4% perf-profile.children.cycles-pp.__mod_memcg_state
9.75 -2.6 7.11 perf-profile.children.cycles-pp.mem_cgroup_charge_skmem
16.64 -1.8 14.81 perf-profile.children.cycles-pp.__sk_mem_raise_allocated
16.77 -1.8 14.94 perf-profile.children.cycles-pp.__sk_mem_schedule
12.25 -1.4 10.82 perf-profile.children.cycles-pp.tcp_wmem_schedule
46.67 -1.4 45.29 perf-profile.children.cycles-pp.sock_sendmsg
44.82 -1.3 43.49 perf-profile.children.cycles-pp.tcp_sendmsg
50.77 -1.2 49.59 perf-profile.children.cycles-pp.splice_to_socket
51.05 -1.2 49.88 perf-profile.children.cycles-pp.direct_splice_actor
39.54 -1.0 38.53 perf-profile.children.cycles-pp.tcp_sendmsg_locked
1.62 ± 3% -0.8 0.80 ± 3% perf-profile.children.cycles-pp.mem_cgroup_uncharge_skmem
12.44 -0.7 11.73 perf-profile.children.cycles-pp.tcp_v4_do_rcv
11.91 -0.7 11.21 perf-profile.children.cycles-pp.tcp_rcv_established
3.49 -0.7 2.81 perf-profile.children.cycles-pp.__sk_mem_reduce_allocated
61.00 -0.6 60.40 perf-profile.children.cycles-pp.splice_direct_to_actor
61.22 -0.6 60.62 perf-profile.children.cycles-pp.do_splice_direct
11.68 -0.5 11.21 perf-profile.children.cycles-pp.ip_protocol_deliver_rcu
11.72 -0.5 11.25 perf-profile.children.cycles-pp.ip_local_deliver_finish
12.68 -0.5 12.22 perf-profile.children.cycles-pp.__netif_receive_skb_one_core
11.45 -0.5 10.99 perf-profile.children.cycles-pp.tcp_v4_rcv
13.08 -0.4 12.64 perf-profile.children.cycles-pp.process_backlog
13.14 -0.4 12.71 perf-profile.children.cycles-pp.__napi_poll
14.19 -0.4 13.78 perf-profile.children.cycles-pp.do_softirq
14.53 -0.4 14.12 perf-profile.children.cycles-pp.__local_bh_enable_ip
13.38 -0.4 12.98 perf-profile.children.cycles-pp.net_rx_action
14.06 -0.4 13.67 perf-profile.children.cycles-pp.__do_softirq
65.51 -0.4 65.12 perf-profile.children.cycles-pp.do_sendfile
16.79 -0.4 16.41 perf-profile.children.cycles-pp.ip_finish_output2
19.45 -0.4 19.08 perf-profile.children.cycles-pp.__tcp_transmit_skb
16.36 -0.4 16.00 perf-profile.children.cycles-pp.__dev_queue_xmit
17.89 -0.4 17.53 perf-profile.children.cycles-pp.__ip_queue_xmit
4.16 -0.3 3.83 perf-profile.children.cycles-pp.tcp_try_rmem_schedule
3.67 -0.3 3.33 ± 2% perf-profile.children.cycles-pp.__release_sock
4.47 -0.3 4.14 ± 2% perf-profile.children.cycles-pp.release_sock
67.14 -0.3 66.81 perf-profile.children.cycles-pp.__x64_sys_sendfile64
6.84 -0.3 6.52 perf-profile.children.cycles-pp.tcp_data_queue
3.20 -0.3 2.88 perf-profile.children.cycles-pp.tcp_ack
2.60 -0.3 2.29 perf-profile.children.cycles-pp.tcp_clean_rtx_queue
7.57 -0.3 7.31 perf-profile.children.cycles-pp.__tcp_push_pending_frames
95.15 -0.1 95.00 perf-profile.children.cycles-pp.do_syscall_64
95.61 -0.1 95.48 perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
1.94 -0.1 1.84 perf-profile.children.cycles-pp.tcp_stream_alloc_skb
1.10 -0.1 1.04 perf-profile.children.cycles-pp.ttwu_do_activate
0.15 ± 2% -0.1 0.09 ± 5% perf-profile.children.cycles-pp.apparmor_socket_sendmsg
0.83 -0.1 0.78 perf-profile.children.cycles-pp.enqueue_task_fair
0.86 -0.1 0.81 perf-profile.children.cycles-pp.activate_task
0.43 -0.0 0.39 perf-profile.children.cycles-pp.enqueue_entity
0.11 -0.0 0.09 ± 5% perf-profile.children.cycles-pp.iov_iter_bvec
0.27 ± 3% -0.0 0.25 perf-profile.children.cycles-pp.pick_eevdf
0.35 ± 2% -0.0 0.33 perf-profile.children.cycles-pp.prepare_task_switch
0.16 ± 4% -0.0 0.14 ± 4% perf-profile.children.cycles-pp.check_preempt_wakeup_fair
0.74 -0.0 0.72 perf-profile.children.cycles-pp.dequeue_task_fair
0.09 -0.0 0.08 perf-profile.children.cycles-pp.rb_first
0.07 +0.0 0.08 perf-profile.children.cycles-pp.security_socket_recvmsg
0.08 +0.0 0.09 perf-profile.children.cycles-pp.tcp_event_data_recv
0.12 ± 3% +0.0 0.13 perf-profile.children.cycles-pp.tcp_rearm_rto
0.23 ± 2% +0.0 0.25 perf-profile.children.cycles-pp.tcp_rcv_space_adjust
0.16 ± 3% +0.0 0.18 perf-profile.children.cycles-pp.lock_timer_base
0.24 ± 2% +0.0 0.26 perf-profile.children.cycles-pp.validate_xmit_skb
0.23 ± 2% +0.0 0.24 perf-profile.children.cycles-pp.__slab_free
0.07 ± 5% +0.0 0.09 ± 4% perf-profile.children.cycles-pp.nf_hook_slow
0.16 ± 2% +0.0 0.18 ± 2% perf-profile.children.cycles-pp.do_splice_read
0.12 +0.0 0.14 ± 3% perf-profile.children.cycles-pp.netif_skb_features
0.68 +0.0 0.70 perf-profile.children.cycles-pp.sk_reset_timer
0.38 +0.0 0.40 perf-profile.children.cycles-pp.__netif_rx
0.40 +0.0 0.42 perf-profile.children.cycles-pp.tcp_mstamp_refresh
0.36 +0.0 0.38 perf-profile.children.cycles-pp.netif_rx_internal
0.58 +0.0 0.60 perf-profile.children.cycles-pp.xas_load
0.30 ± 2% +0.0 0.32 ± 2% perf-profile.children.cycles-pp.rcu_all_qs
0.17 ± 2% +0.0 0.19 ± 2% perf-profile.children.cycles-pp.tcp_queue_rcv
0.56 +0.0 0.58 perf-profile.children.cycles-pp.kmem_cache_free
0.36 +0.0 0.38 perf-profile.children.cycles-pp.page_cache_pipe_buf_confirm
0.24 +0.0 0.26 ± 2% perf-profile.children.cycles-pp.ip_output
0.18 ± 2% +0.0 0.21 perf-profile.children.cycles-pp.ip_rcv_core
0.20 +0.0 0.22 ± 2% perf-profile.children.cycles-pp.is_vmalloc_addr
1.04 +0.0 1.06 perf-profile.children.cycles-pp.dev_hard_start_xmit
0.50 +0.0 0.52 perf-profile.children.cycles-pp.__put_user_8
0.95 +0.0 0.98 perf-profile.children.cycles-pp.loopback_xmit
0.66 +0.0 0.68 perf-profile.children.cycles-pp.syscall_return_via_sysret
0.19 ± 2% +0.0 0.22 ± 2% perf-profile.children.cycles-pp.sockfd_lookup_light
0.72 +0.0 0.75 perf-profile.children.cycles-pp.read_tsc
0.82 +0.0 0.85 perf-profile.children.cycles-pp.tcp_event_new_data_sent
0.86 +0.0 0.89 perf-profile.children.cycles-pp._copy_from_user
0.90 +0.0 0.94 perf-profile.children.cycles-pp.security_file_permission
0.57 +0.0 0.61 perf-profile.children.cycles-pp._raw_spin_lock_irqsave
0.71 +0.0 0.75 perf-profile.children.cycles-pp.netperf_sendfile
0.77 +0.0 0.81 perf-profile.children.cycles-pp.entry_SYSCALL_64
1.15 +0.0 1.19 perf-profile.children.cycles-pp.tcp_send_mss
0.90 +0.0 0.95 perf-profile.children.cycles-pp.kmem_cache_alloc_node
0.89 +0.0 0.94 perf-profile.children.cycles-pp.__fdget
0.60 +0.0 0.65 perf-profile.children.cycles-pp.aa_sk_perm
0.57 +0.0 0.62 perf-profile.children.cycles-pp.current_time
1.22 +0.0 1.27 perf-profile.children.cycles-pp.skb_release_data
1.25 +0.1 1.30 perf-profile.children.cycles-pp.__alloc_skb
0.74 +0.1 0.80 perf-profile.children.cycles-pp.iov_iter_advance
0.86 +0.1 0.91 perf-profile.children.cycles-pp.skb_attempt_defer_free
1.18 +0.1 1.23 perf-profile.children.cycles-pp.rw_verify_area
0.99 +0.1 1.05 perf-profile.children.cycles-pp.page_cache_pipe_buf_release
1.10 +0.1 1.16 ± 2% perf-profile.children.cycles-pp.ktime_get
1.25 +0.1 1.32 perf-profile.children.cycles-pp.__fsnotify_parent
1.24 +0.1 1.31 perf-profile.children.cycles-pp.check_heap_object
1.21 +0.1 1.29 perf-profile.children.cycles-pp.release_pages
1.67 +0.1 1.76 perf-profile.children.cycles-pp.splice_folio_into_pipe
1.63 +0.1 1.72 perf-profile.children.cycles-pp.skb_append_pagefrags
1.52 +0.1 1.62 perf-profile.children.cycles-pp.__folio_batch_release
1.37 +0.1 1.47 perf-profile.children.cycles-pp.atime_needs_update
1.56 +0.1 1.68 perf-profile.children.cycles-pp.touch_atime
2.48 +0.1 2.60 perf-profile.children.cycles-pp.filemap_get_read_batch
1.98 +0.1 2.11 perf-profile.children.cycles-pp.__check_object_size
1.96 +0.1 2.08 perf-profile.children.cycles-pp.iov_iter_extract_pages
2.13 +0.1 2.26 perf-profile.children.cycles-pp.simple_copy_to_iter
2.80 +0.1 2.95 perf-profile.children.cycles-pp.filemap_get_pages
25.18 +0.2 25.35 perf-profile.children.cycles-pp.inet_recvmsg
24.74 +0.2 24.91 perf-profile.children.cycles-pp.tcp_recvmsg_locked
25.32 +0.2 25.49 perf-profile.children.cycles-pp.sock_recvmsg
25.10 +0.2 25.27 perf-profile.children.cycles-pp.tcp_recvmsg
26.38 +0.2 26.56 perf-profile.children.cycles-pp.accept_connection
26.38 +0.2 26.56 perf-profile.children.cycles-pp.accept_connections
26.38 +0.2 26.56 perf-profile.children.cycles-pp.process_requests
26.38 +0.2 26.56 perf-profile.children.cycles-pp.spawn_child
27.00 +0.2 27.18 perf-profile.children.cycles-pp.recv
25.83 +0.2 26.03 perf-profile.children.cycles-pp.__x64_sys_recvfrom
25.78 +0.2 25.98 perf-profile.children.cycles-pp.__sys_recvfrom
5.31 +0.3 5.62 perf-profile.children.cycles-pp.skb_splice_from_iter
10.52 +0.4 10.96 perf-profile.children.cycles-pp._copy_to_iter
8.40 +0.5 8.89 perf-profile.children.cycles-pp.filemap_splice_read
13.82 +0.6 14.47 perf-profile.children.cycles-pp.__skb_datagram_iter
13.85 +0.7 14.50 perf-profile.children.cycles-pp.skb_copy_datagram_iter
6.34 ± 3% -3.6 2.71 ± 5% perf-profile.self.cycles-pp.__mod_memcg_state
0.12 ± 4% -0.0 0.08 ± 6% perf-profile.self.cycles-pp.apparmor_socket_sendmsg
0.10 ± 4% -0.0 0.07 perf-profile.self.cycles-pp.iov_iter_bvec
0.80 -0.0 0.78 perf-profile.self.cycles-pp.sock_sendmsg
0.20 -0.0 0.18 ± 2% perf-profile.self.cycles-pp.pick_eevdf
0.11 ± 3% -0.0 0.09 perf-profile.self.cycles-pp.enqueue_task_fair
0.12 -0.0 0.10 ± 3% perf-profile.self.cycles-pp.sk_wait_data
0.20 ± 2% -0.0 0.18 ± 2% perf-profile.self.cycles-pp.release_sock
0.08 ± 6% -0.0 0.06 ± 6% perf-profile.self.cycles-pp.enqueue_entity
0.15 ± 5% -0.0 0.13 ± 2% perf-profile.self.cycles-pp.do_softirq
0.26 -0.0 0.24 perf-profile.self.cycles-pp.refill_stock
0.06 ± 6% -0.0 0.05 perf-profile.self.cycles-pp.rb_erase
0.12 -0.0 0.11 perf-profile.self.cycles-pp.avg_vruntime
0.61 +0.0 0.63 perf-profile.self.cycles-pp.mem_cgroup_charge_skmem
0.14 ± 2% +0.0 0.16 ± 3% perf-profile.self.cycles-pp.tcp_data_queue
0.18 ± 2% +0.0 0.20 ± 2% perf-profile.self.cycles-pp.simple_copy_to_iter
0.22 +0.0 0.24 perf-profile.self.cycles-pp.syscall_exit_to_user_mode
0.28 +0.0 0.29 perf-profile.self.cycles-pp.direct_splice_actor
0.21 ± 2% +0.0 0.23 ± 2% perf-profile.self.cycles-pp.__slab_free
0.17 ± 2% +0.0 0.19 ± 3% perf-profile.self.cycles-pp.tcp_send_mss
0.14 ± 2% +0.0 0.16 ± 3% perf-profile.self.cycles-pp.do_splice_read
0.22 ± 2% +0.0 0.24 ± 2% perf-profile.self.cycles-pp.net_rx_action
0.27 +0.0 0.29 perf-profile.self.cycles-pp.rw_verify_area
0.25 +0.0 0.27 perf-profile.self.cycles-pp.entry_SYSCALL_64
0.20 +0.0 0.22 ± 2% perf-profile.self.cycles-pp.process_backlog
0.21 +0.0 0.23 ± 2% perf-profile.self.cycles-pp.rcu_all_qs
0.12 ± 4% +0.0 0.14 ± 3% perf-profile.self.cycles-pp.lock_sock_nested
0.52 +0.0 0.54 perf-profile.self.cycles-pp.__virt_addr_valid
0.44 +0.0 0.46 perf-profile.self.cycles-pp.__schedule
0.48 +0.0 0.50 perf-profile.self.cycles-pp.check_heap_object
0.33 +0.0 0.35 perf-profile.self.cycles-pp.filemap_get_pages
0.16 ± 2% +0.0 0.18 ± 2% perf-profile.self.cycles-pp.is_vmalloc_addr
0.32 ± 2% +0.0 0.34 perf-profile.self.cycles-pp.page_cache_pipe_buf_confirm
0.56 +0.0 0.59 perf-profile.self.cycles-pp.sendfile
0.39 +0.0 0.42 perf-profile.self.cycles-pp.tcp_recvmsg_locked
0.47 +0.0 0.50 perf-profile.self.cycles-pp.entry_SYSCALL_64_after_hwframe
0.48 +0.0 0.51 perf-profile.self.cycles-pp.__put_user_8
0.18 ± 4% +0.0 0.20 perf-profile.self.cycles-pp.ip_rcv_core
0.24 ± 3% +0.0 0.26 perf-profile.self.cycles-pp.__sk_mem_reduce_allocated
0.40 +0.0 0.43 perf-profile.self.cycles-pp.current_time
0.69 +0.0 0.72 perf-profile.self.cycles-pp.sendfile_tcp_stream
0.68 +0.0 0.72 perf-profile.self.cycles-pp.read_tsc
0.47 +0.0 0.50 perf-profile.self.cycles-pp.aa_sk_perm
0.95 +0.0 0.98 perf-profile.self.cycles-pp.skb_release_data
0.84 +0.0 0.87 perf-profile.self.cycles-pp._copy_from_user
0.55 +0.0 0.59 perf-profile.self.cycles-pp._raw_spin_lock_irqsave
0.64 +0.0 0.67 perf-profile.self.cycles-pp.netperf_sendfile
0.61 +0.0 0.65 perf-profile.self.cycles-pp.__check_object_size
1.00 +0.0 1.04 perf-profile.self.cycles-pp.tcp_write_xmit
0.82 +0.0 0.87 perf-profile.self.cycles-pp.__fdget
0.62 +0.0 0.67 perf-profile.self.cycles-pp.atime_needs_update
0.68 +0.0 0.73 perf-profile.self.cycles-pp.iov_iter_advance
0.00 +0.1 0.05 perf-profile.self.cycles-pp.free_unref_page_list
0.98 +0.1 1.03 perf-profile.self.cycles-pp.__skb_datagram_iter
0.95 +0.1 1.00 perf-profile.self.cycles-pp.page_cache_pipe_buf_release
1.22 +0.1 1.28 perf-profile.self.cycles-pp.__fsnotify_parent
1.06 +0.1 1.13 perf-profile.self.cycles-pp.release_pages
1.24 +0.1 1.31 perf-profile.self.cycles-pp.tcp_sendmsg_locked
1.47 +0.1 1.54 perf-profile.self.cycles-pp.filemap_splice_read
1.19 +0.1 1.27 perf-profile.self.cycles-pp.iov_iter_extract_pages
1.58 +0.1 1.67 perf-profile.self.cycles-pp.splice_folio_into_pipe
1.54 +0.1 1.62 perf-profile.self.cycles-pp.skb_append_pagefrags
1.85 +0.1 1.95 perf-profile.self.cycles-pp.skb_splice_from_iter
1.90 +0.1 2.00 perf-profile.self.cycles-pp.filemap_get_read_batch
2.59 +0.1 2.72 perf-profile.self.cycles-pp.splice_to_socket
1.17 ± 5% +0.1 1.32 ± 3% perf-profile.self.cycles-pp.page_counter_uncharge
10.43 +0.4 10.87 perf-profile.self.cycles-pp._copy_to_iter
7.04 +0.7 7.76 ± 2% perf-profile.self.cycles-pp.__sk_mem_raise_allocated
***************************************************************************************************
lkp-icl-2sp8: 64 threads 2 sockets Intel(R) Xeon(R) Gold 6346 CPU @ 3.10GHz (Ice Lake) with 256G memory
=========================================================================================
compiler/cpufreq_governor/disk/fs/kconfig/nr_threads/rootfs/tbox_group/test/testcase/testtime:
gcc-12/performance/1HDD/xfs/x86_64-rhel-8.3/10%/debian-11.1-x86_64-20220510.cgz/lkp-icl-2sp8/metamix/stress-ng/60s
commit:
67b8bcbaed ("nilfs2: fix data corruption in dsync block recovery for small block sizes")
9cee7e8ef3 ("mm: memcg: optimize parent iteration in memcg_rstat_updated()")
67b8bcbaed477787 9cee7e8ef3e31ca25b40ca52b85
---------------- ---------------------------
%stddev %change %stddev
\ | \
1356 ± 8% +21.8% 1652 ± 16% sched_debug.cfs_rq:/.util_est.max
20.30 ± 8% +3.9 24.17 ± 9% turbostat.PKG_%
3152098 +4.1% 3281361 stress-ng.metamix.ops
52508 +4.1% 54686 stress-ng.metamix.ops_per_sec
15793876 +4.1% 16439912 stress-ng.time.minor_page_faults
218.90 +1.9% 223.16 stress-ng.time.user_time
7.965e+08 +4.3% 8.306e+08 proc-vmstat.numa_hit
7.967e+08 +4.3% 8.307e+08 proc-vmstat.numa_local
7.935e+08 +4.3% 8.276e+08 proc-vmstat.pgalloc_normal
16118636 +4.3% 16808308 proc-vmstat.pgfault
7.933e+08 +4.3% 8.274e+08 proc-vmstat.pgfree
7.913e+08 +4.3% 8.253e+08 proc-vmstat.unevictable_pgs_culled
0.04 ± 18% +69.1% 0.06 ± 16% perf-sched.sch_delay.avg.ms.__cond_resched.__alloc_pages.alloc_pages_mpol.folio_alloc.__filemap_get_folio
0.03 ± 86% +268.9% 0.09 ± 34% perf-sched.sch_delay.avg.ms.__cond_resched.down_write.generic_file_write_iter.vfs_write.ksys_write
0.03 ±118% +454.7% 0.18 ± 44% perf-sched.sch_delay.avg.ms.__cond_resched.dput.open_last_lookups.path_openat.do_filp_open
0.04 ± 9% +44.9% 0.06 ± 17% perf-sched.sch_delay.avg.ms.__cond_resched.generic_perform_write.generic_file_write_iter.vfs_write.ksys_write
0.07 ± 10% -51.1% 0.04 ± 8% perf-sched.sch_delay.avg.ms.__cond_resched.truncate_inode_pages_range.evict.do_unlinkat.__x64_sys_unlink
0.01 ± 63% +438.3% 0.04 ± 18% perf-sched.sch_delay.avg.ms.irqentry_exit_to_user_mode.asm_sysvec_apic_timer_interrupt.[unknown]
0.02 ± 17% +49.1% 0.03 ± 17% perf-sched.sch_delay.avg.ms.pipe_read.vfs_read.ksys_read.do_syscall_64
0.02 ± 13% +56.5% 0.03 ± 17% perf-sched.sch_delay.avg.ms.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
5.62 ± 37% -31.7% 3.84 ± 22% perf-sched.sch_delay.max.ms.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown]
9.83 ± 52% +128.3% 22.45 ± 42% perf-sched.wait_time.max.ms.__cond_resched.dput.path_put.user_statfs.__do_sys_statfs
11.49 ± 49% +144.8% 28.13 ± 44% perf-sched.wait_time.max.ms.__cond_resched.kmem_cache_alloc_lru.alloc_inode.new_inode.ramfs_get_inode
1.04 -4.9% 0.99 perf-stat.i.MPKI
2.484e+10 +3.4% 2.569e+10 perf-stat.i.branch-instructions
1.098e+08 +2.7% 1.127e+08 perf-stat.i.branch-misses
11.78 -0.5 11.32 perf-stat.i.cache-miss-rate%
1.123e+09 +2.3% 1.149e+09 perf-stat.i.cache-references
1.78 -3.4% 1.72 perf-stat.i.cpi
0.00 ± 4% -0.0 0.00 ± 7% perf-stat.i.dTLB-load-miss-rate%
943921 ± 4% -10.4% 845743 ± 7% perf-stat.i.dTLB-load-misses
3.127e+10 +3.4% 3.232e+10 perf-stat.i.dTLB-loads
2.362e+10 +4.2% 2.46e+10 perf-stat.i.dTLB-stores
1.265e+11 +3.5% 1.31e+11 perf-stat.i.instructions
0.56 +3.5% 0.58 perf-stat.i.ipc
1262 +3.6% 1308 perf-stat.i.metric.M/sec
38.08 -1.7 36.37 ± 2% perf-stat.i.node-load-miss-rate%
3411848 ± 2% -6.2% 3199316 ± 2% perf-stat.i.node-load-misses
3111347 ± 2% +3.7% 3226199 perf-stat.i.node-store-misses
1.05 -4.9% 0.99 perf-stat.overall.MPKI
11.79 -0.5 11.34 perf-stat.overall.cache-miss-rate%
1.78 -3.4% 1.72 perf-stat.overall.cpi
0.00 ± 4% -0.0 0.00 ± 7% perf-stat.overall.dTLB-load-miss-rate%
0.56 +3.5% 0.58 perf-stat.overall.ipc
37.14 ± 2% -1.7 35.49 ± 2% perf-stat.overall.node-load-miss-rate%
2.442e+10 +3.4% 2.525e+10 perf-stat.ps.branch-instructions
1.079e+08 +2.7% 1.108e+08 perf-stat.ps.branch-misses
1.104e+09 +2.3% 1.13e+09 perf-stat.ps.cache-references
935750 ± 4% -10.1% 841448 ± 7% perf-stat.ps.dTLB-load-misses
3.075e+10 +3.3% 3.178e+10 perf-stat.ps.dTLB-loads
2.323e+10 +4.1% 2.419e+10 perf-stat.ps.dTLB-stores
1.244e+11 +3.5% 1.288e+11 perf-stat.ps.instructions
3354652 ± 2% -6.2% 3145099 ± 2% perf-stat.ps.node-load-misses
3060027 ± 2% +3.7% 3172859 perf-stat.ps.node-store-misses
7.565e+12 +3.0% 7.793e+12 perf-stat.total.instructions
32.38 -0.6 31.75 perf-profile.calltrace.cycles-pp.filemap_add_folio.__filemap_get_folio.simple_write_begin.generic_perform_write.generic_file_write_iter
26.28 -0.6 25.67 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.release_pages.__folio_batch_release
28.78 -0.6 28.18 perf-profile.calltrace.cycles-pp.release_pages.__folio_batch_release.truncate_inode_pages_range.evict.do_unlinkat
26.40 -0.6 25.80 perf-profile.calltrace.cycles-pp.folio_lruvec_lock_irqsave.release_pages.__folio_batch_release.truncate_inode_pages_range.evict
26.36 -0.6 25.77 perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.release_pages.__folio_batch_release.truncate_inode_pages_range
28.82 -0.6 28.22 perf-profile.calltrace.cycles-pp.__folio_batch_release.truncate_inode_pages_range.evict.do_unlinkat.__x64_sys_unlink
34.84 -0.6 34.30 perf-profile.calltrace.cycles-pp.__filemap_get_folio.simple_write_begin.generic_perform_write.generic_file_write_iter.vfs_write
32.18 -0.5 31.68 perf-profile.calltrace.cycles-pp.evict.do_unlinkat.__x64_sys_unlink.do_syscall_64.entry_SYSCALL_64_after_hwframe
32.12 -0.5 31.63 perf-profile.calltrace.cycles-pp.truncate_inode_pages_range.evict.do_unlinkat.__x64_sys_unlink.do_syscall_64
32.40 -0.5 31.92 perf-profile.calltrace.cycles-pp.__x64_sys_unlink.do_syscall_64.entry_SYSCALL_64_after_hwframe.unlink
32.41 -0.5 31.93 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.unlink
32.41 -0.5 31.93 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.unlink
32.39 -0.5 31.91 perf-profile.calltrace.cycles-pp.do_unlinkat.__x64_sys_unlink.do_syscall_64.entry_SYSCALL_64_after_hwframe.unlink
32.42 -0.5 31.94 perf-profile.calltrace.cycles-pp.unlink
40.74 -0.4 40.32 perf-profile.calltrace.cycles-pp.simple_write_begin.generic_perform_write.generic_file_write_iter.vfs_write.ksys_write
24.93 -0.4 24.52 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.folio_batch_move_lru.folio_add_lru
25.05 -0.4 24.65 perf-profile.calltrace.cycles-pp.folio_lruvec_lock_irqsave.folio_batch_move_lru.folio_add_lru.filemap_add_folio.__filemap_get_folio
25.02 -0.4 24.62 perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.folio_batch_move_lru.folio_add_lru.filemap_add_folio
27.07 -0.4 26.68 perf-profile.calltrace.cycles-pp.folio_batch_move_lru.folio_add_lru.filemap_add_folio.__filemap_get_folio.simple_write_begin
27.25 -0.4 26.87 perf-profile.calltrace.cycles-pp.folio_add_lru.filemap_add_folio.__filemap_get_folio.simple_write_begin.generic_perform_write
45.36 -0.4 45.01 perf-profile.calltrace.cycles-pp.generic_file_write_iter.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
43.32 -0.3 42.98 perf-profile.calltrace.cycles-pp.generic_perform_write.generic_file_write_iter.vfs_write.ksys_write.do_syscall_64
47.13 -0.3 46.84 perf-profile.calltrace.cycles-pp.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
47.48 -0.3 47.20 perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write.stress_metamix
47.82 -0.3 47.55 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.write.stress_metamix
47.95 -0.3 47.68 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.write.stress_metamix
4.88 -0.3 4.62 perf-profile.calltrace.cycles-pp.__filemap_add_folio.filemap_add_folio.__filemap_get_folio.simple_write_begin.generic_perform_write
48.84 -0.2 48.62 perf-profile.calltrace.cycles-pp.write.stress_metamix
1.62 -0.1 1.49 perf-profile.calltrace.cycles-pp.__mem_cgroup_charge.__filemap_add_folio.filemap_add_folio.__filemap_get_folio.simple_write_begin
1.46 -0.1 1.34 ± 2% perf-profile.calltrace.cycles-pp.__lruvec_stat_mod_folio.__filemap_add_folio.filemap_add_folio.__filemap_get_folio.simple_write_begin
1.52 -0.0 1.50 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu
1.64 -0.0 1.62 perf-profile.calltrace.cycles-pp.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain.populate_vma_page_range.__mm_populate
1.53 -0.0 1.51 perf-profile.calltrace.cycles-pp.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu.lru_add_drain.populate_vma_page_range
0.68 -0.0 0.65 perf-profile.calltrace.cycles-pp.__file_remove_privs.__generic_file_write_iter.generic_file_write_iter.vfs_write.ksys_write
0.57 +0.0 0.59 perf-profile.calltrace.cycles-pp.mmap_region.do_mmap.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64
0.53 +0.0 0.56 perf-profile.calltrace.cycles-pp.rmqueue.get_page_from_freelist.__alloc_pages.alloc_pages_mpol.folio_alloc
0.78 +0.0 0.81 perf-profile.calltrace.cycles-pp.__fsnotify_parent.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
0.54 +0.0 0.57 perf-profile.calltrace.cycles-pp.free_unref_page_list.release_pages.__folio_batch_release.truncate_inode_pages_range.evict
0.73 +0.0 0.76 perf-profile.calltrace.cycles-pp.xas_store.delete_from_page_cache_batch.truncate_inode_pages_range.evict.do_unlinkat
0.86 +0.0 0.90 perf-profile.calltrace.cycles-pp.find_lock_entries.truncate_inode_pages_range.evict.do_unlinkat.__x64_sys_unlink
0.84 +0.0 0.88 perf-profile.calltrace.cycles-pp.get_page_from_freelist.__alloc_pages.alloc_pages_mpol.folio_alloc.__filemap_get_folio
0.80 +0.0 0.84 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64.llseek.stress_metamix
1.44 +0.0 1.48 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.llseek.stress_metamix
1.29 +0.0 1.34 perf-profile.calltrace.cycles-pp.__alloc_pages.alloc_pages_mpol.folio_alloc.__filemap_get_folio.simple_write_begin
1.16 +0.1 1.21 ± 2% perf-profile.calltrace.cycles-pp._copy_to_iter.copy_page_to_iter.filemap_read.vfs_read.ksys_read
1.35 +0.1 1.40 perf-profile.calltrace.cycles-pp.filemap_get_read_batch.filemap_get_pages.filemap_read.vfs_read.ksys_read
1.60 +0.1 1.65 perf-profile.calltrace.cycles-pp.alloc_pages_mpol.folio_alloc.__filemap_get_folio.simple_write_begin.generic_perform_write
0.66 ± 2% +0.1 0.71 perf-profile.calltrace.cycles-pp.rw_verify_area.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
1.26 +0.1 1.32 ± 2% perf-profile.calltrace.cycles-pp.copy_page_to_iter.filemap_read.vfs_read.ksys_read.do_syscall_64
1.54 +0.1 1.60 perf-profile.calltrace.cycles-pp.filemap_get_pages.filemap_read.vfs_read.ksys_read.do_syscall_64
1.68 +0.1 1.74 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.llseek.stress_metamix
1.72 +0.1 1.78 perf-profile.calltrace.cycles-pp.folio_alloc.__filemap_get_folio.simple_write_begin.generic_perform_write.generic_file_write_iter
3.99 +0.2 4.15 perf-profile.calltrace.cycles-pp.llseek.stress_metamix
4.40 +0.2 4.58 perf-profile.calltrace.cycles-pp.filemap_read.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
6.68 +0.3 6.99 perf-profile.calltrace.cycles-pp.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read
7.01 +0.3 7.33 perf-profile.calltrace.cycles-pp.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read.stress_metamix
7.36 +0.3 7.70 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.read.stress_metamix
7.48 +0.3 7.83 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.read.stress_metamix
8.40 +0.4 8.80 perf-profile.calltrace.cycles-pp.read.stress_metamix
62.98 +0.4 63.37 perf-profile.calltrace.cycles-pp.stress_metamix
53.00 -1.0 51.98 perf-profile.children.cycles-pp.folio_lruvec_lock_irqsave
52.86 -1.0 51.84 perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
52.97 -1.0 51.95 perf-profile.children.cycles-pp._raw_spin_lock_irqsave
32.40 -0.6 31.78 perf-profile.children.cycles-pp.filemap_add_folio
28.82 -0.6 28.22 perf-profile.children.cycles-pp.__folio_batch_release
29.09 -0.6 28.51 perf-profile.children.cycles-pp.release_pages
34.90 -0.5 34.35 perf-profile.children.cycles-pp.__filemap_get_folio
32.18 -0.5 31.68 perf-profile.children.cycles-pp.evict
32.14 -0.5 31.65 perf-profile.children.cycles-pp.truncate_inode_pages_range
32.40 -0.5 31.92 perf-profile.children.cycles-pp.__x64_sys_unlink
32.42 -0.5 31.94 perf-profile.children.cycles-pp.unlink
32.39 -0.5 31.91 perf-profile.children.cycles-pp.do_unlinkat
40.76 -0.4 40.34 perf-profile.children.cycles-pp.simple_write_begin
28.75 -0.4 28.34 perf-profile.children.cycles-pp.folio_batch_move_lru
27.27 -0.4 26.89 perf-profile.children.cycles-pp.folio_add_lru
45.41 -0.4 45.06 perf-profile.children.cycles-pp.generic_file_write_iter
43.41 -0.3 43.08 perf-profile.children.cycles-pp.generic_perform_write
93.64 -0.3 93.35 perf-profile.children.cycles-pp.do_syscall_64
47.22 -0.3 46.93 perf-profile.children.cycles-pp.vfs_write
47.54 -0.3 47.26 perf-profile.children.cycles-pp.ksys_write
94.03 -0.3 93.75 perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
5.05 -0.2 4.80 perf-profile.children.cycles-pp.__filemap_add_folio
1.24 ± 2% -0.2 1.00 ± 2% perf-profile.children.cycles-pp.__mod_memcg_lruvec_state
49.16 -0.2 48.95 perf-profile.children.cycles-pp.write
0.49 -0.2 0.30 ± 3% perf-profile.children.cycles-pp.__count_memcg_events
0.63 -0.2 0.47 ± 2% perf-profile.children.cycles-pp.mem_cgroup_commit_charge
2.12 -0.1 1.98 ± 2% perf-profile.children.cycles-pp.__lruvec_stat_mod_folio
1.68 -0.1 1.54 perf-profile.children.cycles-pp.__mem_cgroup_charge
0.44 -0.0 0.42 perf-profile.children.cycles-pp.security_inode_need_killpriv
0.11 ± 3% -0.0 0.10 ± 5% perf-profile.children.cycles-pp.xattr_resolve_name
0.23 +0.0 0.24 perf-profile.children.cycles-pp.free_unref_page_prepare
0.45 +0.0 0.47 perf-profile.children.cycles-pp.fault_in_readable
0.38 +0.0 0.40 perf-profile.children.cycles-pp.stress_hash_jenkin
0.52 +0.0 0.54 perf-profile.children.cycles-pp.fault_in_iov_iter_readable
0.39 ± 2% +0.0 0.42 ± 2% perf-profile.children.cycles-pp.truncate_cleanup_folio
0.31 +0.0 0.33 perf-profile.children.cycles-pp.try_charge_memcg
0.51 +0.0 0.54 ± 2% perf-profile.children.cycles-pp.do_vmi_munmap
0.22 ± 2% +0.0 0.25 ± 3% perf-profile.children.cycles-pp.mem_cgroup_update_lru_size
0.42 +0.0 0.44 perf-profile.children.cycles-pp.atime_needs_update
0.66 +0.0 0.68 perf-profile.children.cycles-pp.syscall_exit_to_user_mode
0.58 +0.0 0.60 perf-profile.children.cycles-pp.mmap_region
0.56 +0.0 0.59 perf-profile.children.cycles-pp.rmqueue
0.26 ± 4% +0.0 0.29 ± 3% perf-profile.children.cycles-pp.run_ksoftirqd
0.56 +0.0 0.59 perf-profile.children.cycles-pp.free_unref_page_list
0.49 +0.0 0.52 perf-profile.children.cycles-pp.touch_atime
0.33 ± 3% +0.0 0.36 ± 3% perf-profile.children.cycles-pp.kmem_cache_free
0.29 ± 5% +0.0 0.32 ± 3% perf-profile.children.cycles-pp.kthread
0.29 ± 5% +0.0 0.32 ± 3% perf-profile.children.cycles-pp.ret_from_fork
0.29 ± 5% +0.0 0.32 ± 3% perf-profile.children.cycles-pp.ret_from_fork_asm
0.41 +0.0 0.44 perf-profile.children.cycles-pp.syscall_return_via_sysret
1.42 +0.0 1.45 perf-profile.children.cycles-pp.xas_store
0.27 ± 3% +0.0 0.30 ± 3% perf-profile.children.cycles-pp.smpboot_thread_fn
0.88 +0.0 0.91 perf-profile.children.cycles-pp.get_page_from_freelist
0.61 ± 2% +0.0 0.64 perf-profile.children.cycles-pp.stress_metamix_file
0.37 ± 4% +0.0 0.40 ± 3% perf-profile.children.cycles-pp.rcu_do_batch
0.87 +0.0 0.91 perf-profile.children.cycles-pp.find_lock_entries
0.89 +0.0 0.93 perf-profile.children.cycles-pp.simple_write_end
0.38 ± 4% +0.0 0.41 ± 4% perf-profile.children.cycles-pp.rcu_core
0.40 ± 4% +0.0 0.43 ± 4% perf-profile.children.cycles-pp.__do_softirq
1.19 +0.0 1.24 perf-profile.children.cycles-pp.__fsnotify_parent
1.38 +0.0 1.42 perf-profile.children.cycles-pp.filemap_get_read_batch
1.17 +0.0 1.22 ± 2% perf-profile.children.cycles-pp._copy_to_iter
1.34 +0.1 1.39 perf-profile.children.cycles-pp.__alloc_pages
1.29 +0.1 1.34 perf-profile.children.cycles-pp.entry_SYSRETQ_unsafe_stack
1.63 +0.1 1.68 perf-profile.children.cycles-pp.alloc_pages_mpol
1.28 +0.1 1.34 ± 2% perf-profile.children.cycles-pp.copy_page_to_iter
1.75 +0.1 1.81 perf-profile.children.cycles-pp.folio_alloc
1.56 +0.1 1.62 perf-profile.children.cycles-pp.filemap_get_pages
1.65 +0.1 1.72 perf-profile.children.cycles-pp.entry_SYSCALL_64
4.03 +0.2 4.19 perf-profile.children.cycles-pp.llseek
4.45 +0.2 4.64 perf-profile.children.cycles-pp.filemap_read
6.72 +0.3 7.03 perf-profile.children.cycles-pp.vfs_read
7.05 +0.3 7.37 perf-profile.children.cycles-pp.ksys_read
62.98 +0.4 63.37 perf-profile.children.cycles-pp.stress_metamix
8.72 +0.4 9.12 perf-profile.children.cycles-pp.read
52.86 -1.0 51.84 perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
1.07 ± 2% -0.3 0.81 ± 2% perf-profile.self.cycles-pp.__mod_memcg_lruvec_state
0.44 -0.2 0.24 ± 3% perf-profile.self.cycles-pp.__count_memcg_events
0.09 +0.0 0.10 perf-profile.self.cycles-pp.get_pfnblock_flags_mask
0.25 ± 2% +0.0 0.26 perf-profile.self.cycles-pp.__filemap_get_folio
0.20 +0.0 0.22 ± 2% perf-profile.self.cycles-pp.delete_from_page_cache_batch
0.33 +0.0 0.35 perf-profile.self.cycles-pp.stress_hash_jenkin
0.21 +0.0 0.23 ± 2% perf-profile.self.cycles-pp.try_charge_memcg
0.49 +0.0 0.51 perf-profile.self.cycles-pp.xas_descend
0.21 ± 3% +0.0 0.23 ± 2% perf-profile.self.cycles-pp.mem_cgroup_update_lru_size
0.29 ± 3% +0.0 0.32 perf-profile.self.cycles-pp.rw_verify_area
0.65 +0.0 0.68 perf-profile.self.cycles-pp.llseek
0.51 +0.0 0.54 perf-profile.self.cycles-pp.stress_metamix_file
0.75 +0.0 0.78 perf-profile.self.cycles-pp.vfs_read
0.81 +0.0 0.84 perf-profile.self.cycles-pp.filemap_read
0.71 +0.0 0.74 perf-profile.self.cycles-pp.find_lock_entries
0.40 +0.0 0.43 ± 2% perf-profile.self.cycles-pp.syscall_return_via_sysret
0.72 +0.0 0.75 perf-profile.self.cycles-pp.do_syscall_64
0.77 +0.0 0.80 perf-profile.self.cycles-pp.release_pages
0.88 +0.0 0.92 perf-profile.self.cycles-pp.lru_add_fn
0.80 +0.0 0.84 perf-profile.self.cycles-pp.vfs_write
1.00 +0.0 1.04 perf-profile.self.cycles-pp.filemap_get_read_batch
1.14 +0.0 1.19 perf-profile.self.cycles-pp.__fsnotify_parent
1.15 +0.0 1.20 ± 2% perf-profile.self.cycles-pp._copy_to_iter
1.25 +0.1 1.30 perf-profile.self.cycles-pp.entry_SYSRETQ_unsafe_stack
***************************************************************************************************
lkp-cpl-4sp2: 224 threads 4 sockets Intel(R) Xeon(R) Platinum 8380H CPU @ 2.90GHz (Cooper Lake) with 192G memory
=========================================================================================
compiler/cpufreq_governor/kconfig/rootfs/runtime/size/tbox_group/test/testcase:
gcc-12/performance/x86_64-rhel-8.3/debian-11.1-x86_64-20220510.cgz/300s/128G/lkp-cpl-4sp2/truncate/vm-scalability
commit:
67b8bcbaed ("nilfs2: fix data corruption in dsync block recovery for small block sizes")
9cee7e8ef3 ("mm: memcg: optimize parent iteration in memcg_rstat_updated()")
67b8bcbaed477787 9cee7e8ef3e31ca25b40ca52b85
---------------- ---------------------------
%stddev %change %stddev
\ | \
5.129e+08 ± 2% +42.0% 7.286e+08 ± 2% vm-scalability.median
5.129e+08 ± 2% +42.0% 7.286e+08 ± 2% vm-scalability.throughput
3842 ± 9% -23.4% 2943 ± 2% vm-scalability.time.involuntary_context_switches
251.17 ± 3% -20.2% 200.50 ± 3% vm-scalability.time.percent_of_cpu_this_job_got
544.92 ± 2% -20.3% 434.06 ± 4% vm-scalability.time.system_time
1.17 ± 2% -0.2 0.94 ± 4% mpstat.cpu.all.sys%
55.67 ± 10% -21.6% 43.67 ± 11% perf-c2c.DRAM.remote
4.50 +10.4% 4.97 ± 7% perf-sched.wait_and_delay.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
4.50 +10.4% 4.97 ± 7% perf-sched.wait_time.avg.ms.schedule_timeout.rcu_gp_fqs_loop.rcu_gp_kthread.kthread
80.83 -10.1% 72.67 ± 2% turbostat.Avg_MHz
2.13 -0.2 1.91 turbostat.Busy%
18971204 ± 24% +52.8% 28988952 ± 23% numa-meminfo.node0.MemFree
16667888 ± 47% +77.3% 29546741 ± 14% numa-meminfo.node2.Inactive
16530544 ± 47% +77.9% 29415298 ± 15% numa-meminfo.node2.Inactive(file)
32514972 ± 26% -40.4% 19367348 ± 21% numa-meminfo.node3.FilePages
31946066 ± 27% -39.6% 19280221 ± 22% numa-meminfo.node3.Inactive
31785044 ± 27% -39.8% 19134640 ± 22% numa-meminfo.node3.Inactive(file)
16347998 ± 52% +80.4% 29486790 ± 14% numa-meminfo.node3.MemFree
33131649 ± 26% -39.7% 19992857 ± 20% numa-meminfo.node3.MemUsed
359118 ± 41% +115.0% 772100 ± 51% numa-numastat.node1.local_node
431596 ± 35% +101.6% 869942 ± 44% numa-numastat.node1.numa_hit
906620 ± 16% -42.5% 521019 ± 58% numa-numastat.node1.numa_miss
977834 ± 15% -36.7% 619153 ± 50% numa-numastat.node1.other_node
836149 ± 40% -70.2% 248916 ± 56% numa-numastat.node3.local_node
1689066 ± 62% -86.6% 225607 ±118% numa-numastat.node3.numa_foreign
942394 ± 36% -62.7% 351650 ± 40% numa-numastat.node3.numa_hit
415036 ± 83% +113.8% 887345 ± 15% numa-numastat.node3.numa_miss
521278 ± 65% +90.1% 990792 ± 14% numa-numastat.node3.other_node
0.69 ± 53% +0.5 1.15 ± 18% perf-profile.calltrace.cycles-pp.trigger_load_balance.update_process_times.tick_sched_handle.tick_nohz_highres_handler.__hrtimer_run_queues
0.08 ± 16% +0.0 0.10 ± 8% perf-profile.children.cycles-pp.irqtime_account_process_tick
0.13 ± 13% +0.1 0.18 ± 17% perf-profile.children.cycles-pp.get_cpu_device
0.26 ± 15% +0.1 0.33 ± 8% perf-profile.children.cycles-pp.rcu_core
0.21 ± 11% +0.1 0.29 ± 7% perf-profile.children.cycles-pp.cpuidle_governor_latency_req
0.76 ± 35% +0.4 1.16 ± 17% perf-profile.children.cycles-pp.trigger_load_balance
0.08 ± 16% +0.0 0.10 ± 8% perf-profile.self.cycles-pp.irqtime_account_process_tick
0.13 ± 13% +0.1 0.18 ± 17% perf-profile.self.cycles-pp.get_cpu_device
0.75 ± 35% +0.4 1.15 ± 18% perf-profile.self.cycles-pp.trigger_load_balance
1.512e+10 -8.3% 1.387e+10 ± 3% perf-stat.i.cpu-cycles
2609319 -2.8% 2535671 perf-stat.i.iTLB-loads
0.07 -8.1% 0.06 ± 3% perf-stat.i.metric.GHz
4.78 -7.9% 4.40 ± 3% perf-stat.overall.cpi
168.84 -8.4% 154.73 ± 3% perf-stat.overall.cycles-between-cache-misses
0.21 +8.6% 0.23 ± 3% perf-stat.overall.ipc
5.544e+08 -1.1% 5.484e+08 perf-stat.ps.branch-instructions
1.51e+10 -8.6% 1.381e+10 ± 3% perf-stat.ps.cpu-cycles
2596703 -2.8% 2523269 perf-stat.ps.iTLB-loads
4744339 ± 24% +52.7% 7243893 ± 24% numa-vmstat.node0.nr_free_pages
431693 ± 35% +101.6% 870245 ± 44% numa-vmstat.node1.numa_hit
359215 ± 41% +115.0% 772404 ± 51% numa-vmstat.node1.numa_local
906620 ± 16% -42.5% 521065 ± 58% numa-vmstat.node1.numa_miss
977834 ± 15% -36.7% 619199 ± 50% numa-vmstat.node1.numa_other
4134651 ± 47% +78.1% 7362301 ± 14% numa-vmstat.node2.nr_inactive_file
4134668 ± 47% +78.1% 7362321 ± 14% numa-vmstat.node2.nr_zone_inactive_file
8128688 ± 26% -40.4% 4844440 ± 21% numa-vmstat.node3.nr_file_pages
4087062 ± 52% +80.3% 7369033 ± 13% numa-vmstat.node3.nr_free_pages
7946196 ± 27% -39.8% 4786206 ± 22% numa-vmstat.node3.nr_inactive_file
7946213 ± 27% -39.8% 4786223 ± 22% numa-vmstat.node3.nr_zone_inactive_file
1689066 ± 62% -86.6% 225607 ±118% numa-vmstat.node3.numa_foreign
942361 ± 36% -62.6% 352113 ± 40% numa-vmstat.node3.numa_hit
836116 ± 40% -70.2% 249379 ± 56% numa-vmstat.node3.numa_local
415036 ± 83% +113.9% 887836 ± 15% numa-vmstat.node3.numa_miss
521278 ± 65% +90.2% 991283 ± 14% numa-vmstat.node3.numa_other
72.96 ± 68% -72.8% 19.85 ± 66% numa-vmstat.node3.workingset_nodes
***************************************************************************************************
lkp-cpl-4sp2: 224 threads 4 sockets Intel(R) Xeon(R) Platinum 8380H CPU @ 2.90GHz (Cooper Lake) with 192G memory
=========================================================================================
compiler/cpufreq_governor/kconfig/mode/nr_task/rootfs/tbox_group/test/testcase:
gcc-12/performance/x86_64-rhel-8.3/thread/50%/debian-11.1-x86_64-20220510.cgz/lkp-cpl-4sp2/fallocate1/will-it-scale
commit:
67b8bcbaed ("nilfs2: fix data corruption in dsync block recovery for small block sizes")
9cee7e8ef3 ("mm: memcg: optimize parent iteration in memcg_rstat_updated()")
67b8bcbaed477787 9cee7e8ef3e31ca25b40ca52b85
---------------- ---------------------------
%stddev %change %stddev
\ | \
0.07 ± 2% +0.0 0.09 ± 2% mpstat.cpu.all.usr%
2980 ± 8% +320.7% 12537 ±105% numa-meminfo.node0.Mapped
3605 +25.4% 4522 vmstat.system.cs
276.17 ± 16% -44.2% 154.00 ± 15% perf-c2c.DRAM.local
3338 ± 3% -31.1% 2300 ± 3% perf-c2c.DRAM.remote
0.02 +50.0% 0.03 turbostat.IPC
9174 ± 22% -62.1% 3476 ± 26% turbostat.POLL
19.05 -2.6% 18.56 turbostat.RAMWatt
2492160 +54.9% 3861385 will-it-scale.112.threads
22251 +54.9% 34476 will-it-scale.per_thread_ops
2492160 +54.9% 3861385 will-it-scale.workload
5794888 ± 5% -14.7% 4940830 ± 5% sched_debug.cfs_rq:/.avg_vruntime.stddev
5794888 ± 5% -14.7% 4940829 ± 5% sched_debug.cfs_rq:/.min_vruntime.stddev
14.21 ± 5% +34.6% 19.12 ± 12% sched_debug.cpu.clock.stddev
3677 +14.2% 4198 ± 5% sched_debug.cpu.nr_switches.avg
41405 +7.3% 44427 proc-vmstat.nr_slab_reclaimable
1.499e+09 +55.1% 2.325e+09 proc-vmstat.numa_hit
1.498e+09 +55.1% 2.324e+09 proc-vmstat.numa_local
100185 -3.4% 96743 ± 2% proc-vmstat.pgactivate
1.499e+09 +55.0% 2.324e+09 proc-vmstat.pgalloc_normal
1.499e+09 +55.0% 2.324e+09 proc-vmstat.pgfree
3.466e+08 ± 2% +40.2% 4.861e+08 ± 14% numa-numastat.node0.local_node
3.468e+08 ± 2% +40.2% 4.863e+08 ± 14% numa-numastat.node0.numa_hit
3.825e+08 ± 2% +60.6% 6.142e+08 ± 2% numa-numastat.node1.local_node
3.827e+08 ± 2% +60.5% 6.144e+08 ± 2% numa-numastat.node1.numa_hit
3.831e+08 ± 2% +62.1% 6.21e+08 ± 2% numa-numastat.node2.local_node
3.832e+08 ± 2% +62.1% 6.212e+08 ± 2% numa-numastat.node2.numa_hit
3.858e+08 ± 2% +56.2% 6.026e+08 ± 11% numa-numastat.node3.local_node
3.86e+08 ± 2% +56.2% 6.027e+08 ± 11% numa-numastat.node3.numa_hit
3.468e+08 ± 2% +40.2% 4.863e+08 ± 14% numa-vmstat.node0.numa_hit
3.467e+08 ± 2% +40.2% 4.86e+08 ± 14% numa-vmstat.node0.numa_local
3.828e+08 ± 2% +60.5% 6.144e+08 ± 2% numa-vmstat.node1.numa_hit
3.826e+08 ± 2% +60.5% 6.142e+08 ± 2% numa-vmstat.node1.numa_local
3.833e+08 ± 2% +62.1% 6.212e+08 ± 2% numa-vmstat.node2.numa_hit
3.832e+08 ± 2% +62.1% 6.21e+08 ± 2% numa-vmstat.node2.numa_local
3.861e+08 ± 2% +56.1% 6.027e+08 ± 11% numa-vmstat.node3.numa_hit
3.858e+08 ± 2% +56.2% 6.026e+08 ± 11% numa-vmstat.node3.numa_local
0.02 ± 57% +149.5% 0.04 ± 58% perf-sched.sch_delay.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
0.01 ± 6% +29.7% 0.01 perf-sched.sch_delay.avg.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
0.01 ± 17% +56.2% 0.01 ± 8% perf-sched.sch_delay.avg.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
0.01 ± 26% +91.4% 0.02 ± 28% perf-sched.sch_delay.max.ms.do_nanosleep.hrtimer_nanosleep.common_nsleep.__x64_sys_clock_nanosleep
0.02 ± 19% +78.2% 0.03 ± 16% perf-sched.sch_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
0.02 ± 25% +87.9% 0.03 ± 22% perf-sched.sch_delay.max.ms.wait_for_partner.fifo_open.do_dentry_open.do_open
168.11 -21.1% 132.56 ± 3% perf-sched.total_wait_and_delay.average.ms
13857 ± 4% +29.3% 17912 ± 2% perf-sched.total_wait_and_delay.count.ms
167.97 -21.2% 132.44 ± 3% perf-sched.total_wait_time.average.ms
65.81 ± 18% +35.1% 88.88 ± 15% perf-sched.wait_and_delay.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
423.14 -37.1% 266.13 ± 4% perf-sched.wait_and_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
1078 ± 7% +31.1% 1413 ± 10% perf-sched.wait_and_delay.count.__cond_resched.shmem_fallocate.vfs_fallocate.__x64_sys_fallocate.do_syscall_64
1012 ± 8% +149.3% 2523 ± 7% perf-sched.wait_and_delay.count.__cond_resched.shmem_inode_acct_blocks.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fallocate
1361 ± 8% +23.1% 1675 ± 7% perf-sched.wait_and_delay.count.__cond_resched.shmem_undo_range.shmem_setattr.notify_change.do_truncate
3600 ± 4% +61.1% 5799 ± 4% perf-sched.wait_and_delay.count.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
0.20 ± 13% +78.4% 0.36 ± 7% perf-sched.wait_and_delay.max.ms.__cond_resched.shmem_fallocate.vfs_fallocate.__x64_sys_fallocate.do_syscall_64
0.20 ± 15% +102.4% 0.41 ± 21% perf-sched.wait_and_delay.max.ms.__cond_resched.shmem_inode_acct_blocks.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fallocate
0.23 ± 25% +58.8% 0.36 ± 7% perf-sched.wait_and_delay.max.ms.__cond_resched.shmem_undo_range.shmem_setattr.notify_change.do_truncate
29.38 ± 8% +562.6% 194.68 ±185% perf-sched.wait_and_delay.max.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
3.95 ± 8% +17.7% 4.65 ± 6% perf-sched.wait_and_delay.max.ms.schedule_timeout.__wait_for_common.wait_for_completion_state.kernel_clone
1.55 ± 5% +15.4% 1.79 ± 3% perf-sched.wait_time.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
423.12 -37.1% 266.12 ± 4% perf-sched.wait_time.avg.ms.smpboot_thread_fn.kthread.ret_from_fork.ret_from_fork_asm
1.57 ± 7% +16.2% 1.82 ± 4% perf-sched.wait_time.avg.ms.syslog_print.do_syslog.kmsg_read.vfs_read
0.20 ± 13% +78.4% 0.36 ± 7% perf-sched.wait_time.max.ms.__cond_resched.shmem_fallocate.vfs_fallocate.__x64_sys_fallocate.do_syscall_64
0.20 ± 15% +102.4% 0.41 ± 21% perf-sched.wait_time.max.ms.__cond_resched.shmem_inode_acct_blocks.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fallocate
0.23 ± 25% +58.8% 0.36 ± 7% perf-sched.wait_time.max.ms.__cond_resched.shmem_undo_range.shmem_setattr.notify_change.do_truncate
3.11 ± 5% +15.4% 3.59 ± 3% perf-sched.wait_time.max.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64
29.38 ± 8% +562.7% 194.68 ±185% perf-sched.wait_time.max.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64
3.14 ± 7% +16.1% 3.65 ± 4% perf-sched.wait_time.max.ms.syslog_print.do_syslog.kmsg_read.vfs_read
5.05 ± 3% -47.5% 2.65 perf-stat.i.MPKI
6.255e+09 +33.9% 8.375e+09 perf-stat.i.branch-instructions
0.37 ± 2% +0.0 0.39 perf-stat.i.branch-miss-rate%
23190779 +41.9% 32908628 perf-stat.i.branch-misses
32.92 -4.9 28.01 perf-stat.i.cache-miss-rate%
1.497e+08 ± 2% -28.1% 1.076e+08 perf-stat.i.cache-misses
4.548e+08 -15.6% 3.837e+08 perf-stat.i.cache-references
3482 +26.6% 4408 perf-stat.i.context-switches
14.39 -27.0% 10.50 perf-stat.i.cpi
262.60 +0.9% 265.06 perf-stat.i.cpu-migrations
2854 ± 2% +39.0% 3968 perf-stat.i.cycles-between-cache-misses
7.783e+09 +36.5% 1.062e+10 perf-stat.i.dTLB-loads
3.804e+09 +51.9% 5.779e+09 perf-stat.i.dTLB-stores
80.49 +4.8 85.30 perf-stat.i.iTLB-load-miss-rate%
10924802 +41.4% 15443362 perf-stat.i.iTLB-load-misses
2.972e+10 +36.5% 4.057e+10 perf-stat.i.instructions
2749 ± 2% -4.1% 2636 perf-stat.i.instructions-per-iTLB-miss
0.07 +37.2% 0.10 perf-stat.i.ipc
120.11 -21.3% 94.54 ± 5% perf-stat.i.metric.K/sec
81.63 +37.5% 112.27 perf-stat.i.metric.M/sec
20471399 -31.0% 14134700 ± 2% perf-stat.i.node-load-misses
1500875 ± 19% -44.9% 827495 ± 10% perf-stat.i.node-loads
2312406 +29.6% 2997675 perf-stat.i.node-store-misses
5.04 ± 3% -47.3% 2.65 perf-stat.overall.MPKI
0.37 ± 2% +0.0 0.39 perf-stat.overall.branch-miss-rate%
32.90 -4.9 28.04 perf-stat.overall.cache-miss-rate%
14.37 -26.9% 10.50 perf-stat.overall.cpi
2854 ± 2% +38.7% 3958 perf-stat.overall.cycles-between-cache-misses
0.00 ± 12% -0.0 0.00 ± 11% perf-stat.overall.dTLB-store-miss-rate%
80.63 +4.8 85.46 perf-stat.overall.iTLB-load-miss-rate%
0.07 +36.8% 0.10 perf-stat.overall.ipc
3580231 -11.7% 3162678 perf-stat.overall.path-length
6.232e+09 +33.9% 8.346e+09 perf-stat.ps.branch-instructions
23162804 +41.7% 32833133 perf-stat.ps.branch-misses
1.491e+08 ± 2% -28.1% 1.072e+08 perf-stat.ps.cache-misses
4.532e+08 -15.6% 3.825e+08 perf-stat.ps.cache-references
3470 +26.6% 4393 perf-stat.ps.context-switches
7.754e+09 +36.5% 1.059e+10 perf-stat.ps.dTLB-loads
3.789e+09 +52.0% 5.758e+09 perf-stat.ps.dTLB-stores
10884687 +41.4% 15387479 perf-stat.ps.iTLB-load-misses
2.962e+10 +36.5% 4.043e+10 perf-stat.ps.instructions
20394582 -30.9% 14085516 ± 2% perf-stat.ps.node-load-misses
1497500 ± 19% -44.8% 827353 ± 10% perf-stat.ps.node-loads
2303483 +29.7% 2986868 perf-stat.ps.node-store-misses
8.923e+12 +36.9% 1.221e+13 perf-stat.total.instructions
21.45 ± 4% -7.9 13.52 ± 2% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.release_pages.__folio_batch_release
21.47 ± 4% -7.9 13.55 ± 2% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.release_pages.__folio_batch_release.shmem_undo_range
21.49 ± 4% -7.9 13.58 ± 2% perf-profile.calltrace.cycles-pp.folio_lruvec_lock_irqsave.release_pages.__folio_batch_release.shmem_undo_range.shmem_setattr
21.70 ± 8% -6.8 14.87 ± 5% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.folio_batch_move_lru.folio_add_lru
21.72 ± 8% -6.8 14.90 ± 5% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.folio_batch_move_lru.folio_add_lru.shmem_alloc_and_add_folio
21.74 ± 8% -6.8 14.93 ± 5% perf-profile.calltrace.cycles-pp.folio_lruvec_lock_irqsave.folio_batch_move_lru.folio_add_lru.shmem_alloc_and_add_folio.shmem_get_folio_gfp
23.54 ± 7% -6.7 16.85 ± 4% perf-profile.calltrace.cycles-pp.folio_batch_move_lru.folio_add_lru.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fallocate
23.59 ± 7% -6.7 16.90 ± 4% perf-profile.calltrace.cycles-pp.folio_add_lru.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fallocate.vfs_fallocate
30.62 ± 2% -5.5 25.14 perf-profile.calltrace.cycles-pp.shmem_undo_range.shmem_setattr.notify_change.do_truncate.do_sys_ftruncate
30.63 ± 2% -5.5 25.16 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.ftruncate64
30.63 ± 2% -5.5 25.16 perf-profile.calltrace.cycles-pp.do_sys_ftruncate.do_syscall_64.entry_SYSCALL_64_after_hwframe.ftruncate64
30.63 ± 2% -5.5 25.16 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.ftruncate64
30.62 ± 2% -5.5 25.16 perf-profile.calltrace.cycles-pp.do_truncate.do_sys_ftruncate.do_syscall_64.entry_SYSCALL_64_after_hwframe.ftruncate64
30.62 ± 2% -5.5 25.15 perf-profile.calltrace.cycles-pp.shmem_setattr.notify_change.do_truncate.do_sys_ftruncate.do_syscall_64
30.63 ± 2% -5.5 25.16 perf-profile.calltrace.cycles-pp.ftruncate64
30.62 ± 2% -5.5 25.16 perf-profile.calltrace.cycles-pp.notify_change.do_truncate.do_sys_ftruncate.do_syscall_64.entry_SYSCALL_64_after_hwframe
26.52 ± 2% -5.2 21.36 perf-profile.calltrace.cycles-pp.__folio_batch_release.shmem_undo_range.shmem_setattr.notify_change.do_truncate
25.28 ± 2% -4.8 20.46 perf-profile.calltrace.cycles-pp.release_pages.__folio_batch_release.shmem_undo_range.shmem_setattr.notify_change
4.03 ± 4% -1.0 3.02 ± 5% perf-profile.calltrace.cycles-pp.__mod_memcg_lruvec_state.__lruvec_stat_mod_folio.shmem_add_to_page_cache.shmem_alloc_and_add_folio.shmem_get_folio_gfp
1.56 ± 5% -0.7 0.82 ± 9% perf-profile.calltrace.cycles-pp.__count_memcg_events.mem_cgroup_commit_charge.__mem_cgroup_charge.shmem_alloc_and_add_folio.shmem_get_folio_gfp
4.99 ± 3% -0.7 4.32 ± 3% perf-profile.calltrace.cycles-pp.__lruvec_stat_mod_folio.shmem_add_to_page_cache.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fallocate
3.02 -0.6 2.37 ± 3% perf-profile.calltrace.cycles-pp.__mod_memcg_lruvec_state.__lruvec_stat_mod_folio.filemap_unaccount_folio.__filemap_remove_folio.filemap_remove_folio
5.21 ± 3% -0.6 4.65 ± 3% perf-profile.calltrace.cycles-pp.shmem_add_to_page_cache.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fallocate.vfs_fallocate
3.64 ± 2% -0.5 3.09 ± 5% perf-profile.calltrace.cycles-pp.__lruvec_stat_mod_folio.filemap_unaccount_folio.__filemap_remove_folio.filemap_remove_folio.truncate_inode_folio
3.64 ± 2% -0.5 3.10 ± 5% perf-profile.calltrace.cycles-pp.filemap_unaccount_folio.__filemap_remove_folio.filemap_remove_folio.truncate_inode_folio.shmem_undo_range
3.77 ± 2% -0.5 3.31 ± 4% perf-profile.calltrace.cycles-pp.__filemap_remove_folio.filemap_remove_folio.truncate_inode_folio.shmem_undo_range.shmem_setattr
3.86 -0.4 3.43 ± 4% perf-profile.calltrace.cycles-pp.filemap_remove_folio.truncate_inode_folio.shmem_undo_range.shmem_setattr.notify_change
3.94 -0.4 3.56 ± 4% perf-profile.calltrace.cycles-pp.truncate_inode_folio.shmem_undo_range.shmem_setattr.notify_change.do_truncate
1.19 ± 3% -0.3 0.86 ± 2% perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu.__folio_batch_release
1.19 ± 3% -0.3 0.86 ± 2% perf-profile.calltrace.cycles-pp.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu.__folio_batch_release.shmem_undo_range
1.19 ± 3% -0.3 0.86 ± 2% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.folio_lruvec_lock_irqsave.folio_batch_move_lru.lru_add_drain_cpu
1.21 ± 3% -0.3 0.88 ± 2% perf-profile.calltrace.cycles-pp.lru_add_drain_cpu.__folio_batch_release.shmem_undo_range.shmem_setattr.notify_change
1.21 ± 3% -0.3 0.88 ± 2% perf-profile.calltrace.cycles-pp.folio_batch_move_lru.lru_add_drain_cpu.__folio_batch_release.shmem_undo_range.shmem_setattr
0.93 -0.2 0.68 perf-profile.calltrace.cycles-pp.__mod_memcg_lruvec_state.release_pages.__folio_batch_release.shmem_undo_range.shmem_setattr
1.42 -0.2 1.26 ± 2% perf-profile.calltrace.cycles-pp.__mod_memcg_lruvec_state.lru_add_fn.folio_batch_move_lru.folio_add_lru.shmem_alloc_and_add_folio
0.00 +0.6 0.57 ± 3% perf-profile.calltrace.cycles-pp.page_counter_uncharge.uncharge_batch.__mem_cgroup_uncharge_list.release_pages.__folio_batch_release
1.02 ± 6% +0.8 1.80 perf-profile.calltrace.cycles-pp.uncharge_batch.__mem_cgroup_uncharge_list.release_pages.__folio_batch_release.shmem_undo_range
1.62 ± 12% +2.5 4.07 ± 4% perf-profile.calltrace.cycles-pp.uncharge_folio.__mem_cgroup_uncharge_list.release_pages.__folio_batch_release.shmem_undo_range
2.64 ± 9% +3.2 5.87 ± 3% perf-profile.calltrace.cycles-pp.__mem_cgroup_uncharge_list.release_pages.__folio_batch_release.shmem_undo_range.shmem_setattr
6.40 ± 10% +3.4 9.84 ± 3% perf-profile.calltrace.cycles-pp.mem_cgroup_commit_charge.__mem_cgroup_charge.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fallocate
48.46 +4.9 53.38 perf-profile.calltrace.cycles-pp.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fallocate.vfs_fallocate.__x64_sys_fallocate
48.53 +5.0 53.48 perf-profile.calltrace.cycles-pp.shmem_get_folio_gfp.shmem_fallocate.vfs_fallocate.__x64_sys_fallocate.do_syscall_64
48.72 +5.1 53.78 perf-profile.calltrace.cycles-pp.shmem_fallocate.vfs_fallocate.__x64_sys_fallocate.do_syscall_64.entry_SYSCALL_64_after_hwframe
48.77 +5.1 53.84 perf-profile.calltrace.cycles-pp.vfs_fallocate.__x64_sys_fallocate.do_syscall_64.entry_SYSCALL_64_after_hwframe.fallocate64
48.81 +5.1 53.91 perf-profile.calltrace.cycles-pp.__x64_sys_fallocate.do_syscall_64.entry_SYSCALL_64_after_hwframe.fallocate64
48.83 +5.1 53.94 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.fallocate64
48.84 +5.1 53.96 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.fallocate64
48.90 +5.1 54.05 perf-profile.calltrace.cycles-pp.fallocate64
6.97 ± 11% +6.2 13.17 ± 4% perf-profile.calltrace.cycles-pp.get_mem_cgroup_from_mm.__mem_cgroup_charge.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fallocate
19.24 ± 11% +12.0 31.20 ± 4% perf-profile.calltrace.cycles-pp.__mem_cgroup_charge.shmem_alloc_and_add_folio.shmem_get_folio_gfp.shmem_fallocate.vfs_fallocate
44.41 ± 5% -15.1 29.28 ± 3% perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
44.43 ± 5% -15.1 29.34 ± 3% perf-profile.children.cycles-pp._raw_spin_lock_irqsave
44.47 ± 5% -15.1 29.39 ± 3% perf-profile.children.cycles-pp.folio_lruvec_lock_irqsave
24.81 ± 7% -7.1 17.76 ± 4% perf-profile.children.cycles-pp.folio_batch_move_lru
23.62 ± 7% -6.7 16.92 ± 4% perf-profile.children.cycles-pp.folio_add_lru
30.63 ± 2% -5.5 25.16 perf-profile.children.cycles-pp.do_sys_ftruncate
30.62 ± 2% -5.5 25.15 perf-profile.children.cycles-pp.shmem_undo_range
30.62 ± 2% -5.5 25.16 perf-profile.children.cycles-pp.do_truncate
30.62 ± 2% -5.5 25.15 perf-profile.children.cycles-pp.shmem_setattr
30.62 ± 2% -5.5 25.16 perf-profile.children.cycles-pp.notify_change
30.63 ± 2% -5.5 25.16 perf-profile.children.cycles-pp.ftruncate64
26.52 ± 2% -5.2 21.36 perf-profile.children.cycles-pp.__folio_batch_release
25.35 ± 2% -4.8 20.54 perf-profile.children.cycles-pp.release_pages
9.44 -2.1 7.38 ± 3% perf-profile.children.cycles-pp.__mod_memcg_lruvec_state
8.64 ± 2% -1.2 7.42 ± 3% perf-profile.children.cycles-pp.__lruvec_stat_mod_folio
1.76 ± 5% -0.7 1.05 ± 7% perf-profile.children.cycles-pp.__count_memcg_events
5.23 ± 3% -0.6 4.67 ± 3% perf-profile.children.cycles-pp.shmem_add_to_page_cache
3.65 ± 2% -0.5 3.10 ± 5% perf-profile.children.cycles-pp.filemap_unaccount_folio
3.78 ± 2% -0.5 3.31 ± 4% perf-profile.children.cycles-pp.__filemap_remove_folio
3.86 ± 2% -0.4 3.44 ± 4% perf-profile.children.cycles-pp.filemap_remove_folio
3.94 -0.4 3.56 ± 4% perf-profile.children.cycles-pp.truncate_inode_folio
1.22 ± 3% -0.3 0.88 ± 2% perf-profile.children.cycles-pp.lru_add_drain_cpu
0.06 ± 11% +0.0 0.08 ± 6% perf-profile.children.cycles-pp._raw_spin_lock_irq
0.06 ± 13% +0.0 0.08 ± 4% perf-profile.children.cycles-pp.xas_alloc
0.06 ± 9% +0.0 0.08 ± 5% perf-profile.children.cycles-pp.kmem_cache_alloc_lru
0.05 ± 7% +0.0 0.08 ± 4% perf-profile.children.cycles-pp.kmem_cache_free
0.06 +0.0 0.09 ± 5% perf-profile.children.cycles-pp.xas_load
0.05 ± 7% +0.0 0.08 perf-profile.children.cycles-pp.truncate_cleanup_folio
0.06 ± 6% +0.0 0.09 perf-profile.children.cycles-pp.cgroup_rstat_updated
0.07 ± 11% +0.0 0.10 ± 5% perf-profile.children.cycles-pp.xas_create
0.05 ± 8% +0.0 0.09 ± 5% perf-profile.children.cycles-pp.__do_softirq
0.06 ± 7% +0.0 0.10 ± 4% perf-profile.children.cycles-pp.rmqueue
0.07 ± 7% +0.0 0.10 perf-profile.children.cycles-pp.__dquot_alloc_space
0.06 ± 7% +0.0 0.10 ± 3% perf-profile.children.cycles-pp.free_unref_page_list
0.05 ± 7% +0.0 0.09 ± 5% perf-profile.children.cycles-pp.rcu_core
0.05 ± 7% +0.0 0.09 ± 5% perf-profile.children.cycles-pp.rcu_do_batch
0.03 ±100% +0.0 0.06 ± 7% perf-profile.children.cycles-pp.__memcg_slab_pre_alloc_hook
0.09 ± 6% +0.0 0.13 ± 5% perf-profile.children.cycles-pp.page_counter_try_charge
0.02 ± 99% +0.0 0.07 ± 5% perf-profile.children.cycles-pp.__memcg_slab_free_hook
0.10 ± 4% +0.0 0.15 ± 4% perf-profile.children.cycles-pp._raw_spin_lock
0.00 +0.1 0.05 perf-profile.children.cycles-pp.__cond_resched
0.00 +0.1 0.05 perf-profile.children.cycles-pp.__folio_cancel_dirty
0.00 +0.1 0.05 perf-profile.children.cycles-pp.shmem_recalc_inode
0.00 +0.1 0.05 perf-profile.children.cycles-pp.xas_init_marks
0.00 +0.1 0.05 ± 7% perf-profile.children.cycles-pp.irq_exit_rcu
0.00 +0.1 0.05 ± 7% perf-profile.children.cycles-pp.memcg_check_events
0.00 +0.1 0.06 ± 9% perf-profile.children.cycles-pp.entry_SYSCALL_64
0.00 +0.1 0.06 ± 9% perf-profile.children.cycles-pp.obj_cgroup_uncharge_pages
0.01 ±223% +0.1 0.06 ± 7% perf-profile.children.cycles-pp.obj_cgroup_charge
0.00 +0.1 0.06 ± 8% perf-profile.children.cycles-pp.filemap_get_entry
0.00 +0.1 0.06 ± 8% perf-profile.children.cycles-pp.xas_descend
0.13 ± 4% +0.1 0.19 ± 2% perf-profile.children.cycles-pp.find_lock_entries
0.10 ± 4% +0.1 0.16 ± 5% perf-profile.children.cycles-pp.get_page_from_freelist
0.00 +0.1 0.06 ± 9% perf-profile.children.cycles-pp.folio_unlock
0.00 +0.1 0.06 ± 7% perf-profile.children.cycles-pp.xas_clear_mark
0.22 ± 5% +0.1 0.28 ± 5% perf-profile.children.cycles-pp.propagate_protected_usage
0.14 ± 2% +0.1 0.22 perf-profile.children.cycles-pp.shmem_inode_acct_blocks
0.15 ± 3% +0.1 0.24 ± 3% perf-profile.children.cycles-pp.__alloc_pages
0.10 ± 5% +0.1 0.19 ± 5% perf-profile.children.cycles-pp.__mod_node_page_state
0.17 ± 2% +0.1 0.27 ± 2% perf-profile.children.cycles-pp.xas_store
0.18 ± 7% +0.1 0.28 ± 4% perf-profile.children.cycles-pp.try_charge_memcg
0.18 ± 2% +0.1 0.29 ± 3% perf-profile.children.cycles-pp.alloc_pages_mpol
0.13 ± 3% +0.1 0.23 ± 4% perf-profile.children.cycles-pp.__mod_lruvec_state
0.20 ± 2% +0.1 0.32 ± 3% perf-profile.children.cycles-pp.shmem_alloc_folio
0.41 ± 4% +0.2 0.57 ± 3% perf-profile.children.cycles-pp.page_counter_uncharge
1.02 ± 6% +0.8 1.80 perf-profile.children.cycles-pp.uncharge_batch
1.62 ± 12% +2.5 4.07 ± 4% perf-profile.children.cycles-pp.uncharge_folio
2.64 ± 9% +3.2 5.87 ± 3% perf-profile.children.cycles-pp.__mem_cgroup_uncharge_list
6.42 ± 10% +3.4 9.85 ± 3% perf-profile.children.cycles-pp.mem_cgroup_commit_charge
48.51 +4.9 53.42 perf-profile.children.cycles-pp.shmem_alloc_and_add_folio
48.58 +4.9 53.53 perf-profile.children.cycles-pp.shmem_get_folio_gfp
48.72 +5.1 53.78 perf-profile.children.cycles-pp.shmem_fallocate
48.77 +5.1 53.85 perf-profile.children.cycles-pp.vfs_fallocate
48.81 +5.1 53.91 perf-profile.children.cycles-pp.__x64_sys_fallocate
48.93 +5.2 54.08 perf-profile.children.cycles-pp.fallocate64
6.98 ± 11% +6.2 13.18 ± 4% perf-profile.children.cycles-pp.get_mem_cgroup_from_mm
19.26 ± 11% +12.0 31.22 ± 4% perf-profile.children.cycles-pp.__mem_cgroup_charge
44.41 ± 5% -15.1 29.28 ± 3% perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath
9.40 -2.1 7.32 ± 3% perf-profile.self.cycles-pp.__mod_memcg_lruvec_state
1.75 ± 5% -0.7 1.04 ± 7% perf-profile.self.cycles-pp.__count_memcg_events
0.05 +0.0 0.07 ± 5% perf-profile.self.cycles-pp.cgroup_rstat_updated
0.06 +0.0 0.08 ± 5% perf-profile.self.cycles-pp.shmem_fallocate
0.06 ± 7% +0.0 0.09 ± 4% perf-profile.self.cycles-pp.page_counter_try_charge
0.06 ± 6% +0.0 0.10 ± 4% perf-profile.self.cycles-pp.shmem_add_to_page_cache
0.06 ± 6% +0.0 0.10 perf-profile.self.cycles-pp.xas_store
0.02 ± 99% +0.0 0.06 ± 7% perf-profile.self.cycles-pp._raw_spin_lock_irq
0.11 ± 5% +0.0 0.16 ± 2% perf-profile.self.cycles-pp.find_lock_entries
0.01 ±223% +0.0 0.06 ± 6% perf-profile.self.cycles-pp.obj_cgroup_charge
0.10 ± 3% +0.1 0.15 ± 4% perf-profile.self.cycles-pp._raw_spin_lock
0.00 +0.1 0.05 perf-profile.self.cycles-pp.fallocate64
0.00 +0.1 0.05 ± 7% perf-profile.self.cycles-pp.__dquot_alloc_space
0.00 +0.1 0.05 ± 8% perf-profile.self.cycles-pp.xas_descend
0.00 +0.1 0.06 ± 6% perf-profile.self.cycles-pp.folio_unlock
0.08 ± 5% +0.1 0.14 ± 3% perf-profile.self.cycles-pp.try_charge_memcg
0.00 +0.1 0.06 perf-profile.self.cycles-pp.__alloc_pages
0.00 +0.1 0.06 perf-profile.self.cycles-pp.xas_clear_mark
0.01 ±223% +0.1 0.07 ± 10% perf-profile.self.cycles-pp.get_page_from_freelist
0.13 +0.1 0.19 ± 3% perf-profile.self.cycles-pp.release_pages
0.00 +0.1 0.07 perf-profile.self.cycles-pp._raw_spin_lock_irqsave
0.21 ± 5% +0.1 0.28 ± 4% perf-profile.self.cycles-pp.propagate_protected_usage
0.10 ± 5% +0.1 0.18 ± 6% perf-profile.self.cycles-pp.lru_add_fn
0.09 ± 4% +0.1 0.18 ± 4% perf-profile.self.cycles-pp.__mod_node_page_state
0.22 ± 4% +0.1 0.33 ± 5% perf-profile.self.cycles-pp.page_counter_uncharge
0.18 ± 2% +0.1 0.31 ± 3% perf-profile.self.cycles-pp.folio_batch_move_lru
0.08 ± 36% +0.1 0.22 ± 9% perf-profile.self.cycles-pp.asm_sysvec_apic_timer_interrupt
1.48 ± 4% +0.4 1.86 ± 9% perf-profile.self.cycles-pp.__lruvec_stat_mod_folio
0.40 ± 11% +0.6 0.96 ± 3% perf-profile.self.cycles-pp.uncharge_batch
5.68 ± 12% +2.2 7.89 ± 4% perf-profile.self.cycles-pp.__mem_cgroup_charge
1.61 ± 12% +2.4 4.06 ± 4% perf-profile.self.cycles-pp.uncharge_folio
4.82 ± 12% +4.1 8.97 ± 4% perf-profile.self.cycles-pp.mem_cgroup_commit_charge
6.95 ± 11% +6.2 13.14 ± 4% perf-profile.self.cycles-pp.get_mem_cgroup_from_mm
Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.
--
0-DAY CI Kernel Test Service
https://github.com/intel/lkp-tests/wiki
reply other threads:[~2024-02-18 13:16 UTC|newest]
Thread overview: [no followups] expand[flat|nested] mbox.gz Atom feed
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=202402182000.f21279e1-oliver.sang@intel.com \
--to=oliver.sang@intel.com \
--cc=akpm@linux-foundation.org \
--cc=cgroups@vger.kernel.org \
--cc=feng.tang@intel.com \
--cc=fengwei.yin@intel.com \
--cc=gthelen@google.com \
--cc=hannes@cmpxchg.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=lkp@intel.com \
--cc=mhocko@kernel.org \
--cc=muchun.song@linux.dev \
--cc=oe-lkp@lists.linux.dev \
--cc=roman.gushchin@linux.dev \
--cc=shakeelb@google.com \
--cc=ying.huang@intel.com \
--cc=yosryahmed@google.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.