Greeting, FYI, we noticed a -4.6% regression of fsmark.files_per_sec due to commit: commit: a79b28c284fd910bb291dbf307a26f4d432e88f3 ("xfs: separate CIL commit record IO") https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master in testcase: fsmark on test machine: 96 threads 2 sockets Intel(R) Xeon(R) Gold 6252 CPU @ 2.10GHz with 512G memory with following parameters: iterations: 1x nr_threads: 32t disk: 1SSD fs: xfs filesize: 8K test_size: 400M sync_method: fsyncBeforeClose nr_directories: 16d nr_files_per_directory: 256fpd cpufreq_governor: performance ucode: 0x5003006 test-description: The fsmark is a file system benchmark to test synchronous write workloads, for example, mail servers workload. test-url: https://sourceforge.net/projects/fsmark/ In addition to that, the commit also has significant impact on the following tests: +------------------+---------------------------------------------------------------------------------+ | testcase: change | aim7: aim7.jobs-per-min 22.0% improvement | | test machine | 88 threads 2 sockets Intel(R) Xeon(R) Gold 6238M CPU @ 2.10GHz with 128G memory | | test parameters | cpufreq_governor=performance | | | disk=4BRD_12G | | | fs=xfs | | | load=300 | | | md=RAID0 | | | test=sync_disk_rw | | | ucode=0x5003006 | +------------------+---------------------------------------------------------------------------------+ If you fix the issue, kindly add following tag Reported-by: kernel test robot Details are as below: --------------------------------------------------------------------------------------------------> To reproduce: git clone https://github.com/intel/lkp-tests.git cd lkp-tests bin/lkp install job.yaml # job file is attached in this email bin/lkp split-job --compatible job.yaml # generate the yaml file for lkp run bin/lkp run generated-yaml-file ========================================================================================= compiler/cpufreq_governor/disk/filesize/fs/iterations/kconfig/nr_directories/nr_files_per_directory/nr_threads/rootfs/sync_method/tbox_group/test_size/testcase/ucode: gcc-9/performance/1SSD/8K/xfs/1x/x86_64-rhel-8.3/16d/256fpd/32t/debian-10.4-x86_64-20200603.cgz/fsyncBeforeClose/lkp-csl-2sp7/400M/fsmark/0x5003006 commit: 18842e0a4f ("xfs: Fix 64-bit division on 32-bit in xlog_state_switch_iclogs()") a79b28c284 ("xfs: separate CIL commit record IO") 18842e0a4f48564b a79b28c284fd910bb291dbf307a ---------------- --------------------------- %stddev %change %stddev \ | \ 16388 -4.6% 15631 ± 2% fsmark.files_per_sec 19379 ± 6% -31.7% 13238 ± 3% fsmark.time.involuntary_context_switches 294578 +11.5% 328546 fsmark.time.voluntary_context_switches 11335 ± 11% +67.3% 18968 ± 56% cpuidle.POLL.usage 2860 ±199% -98.4% 45.67 ±127% softirqs.CPU72.TIMER 114218 ± 7% -11.2% 101430 vmstat.io.bo 23503 ± 12% +19.7% 28140 ± 9% numa-vmstat.node0.nr_slab_unreclaimable 588.67 ± 39% -44.8% 325.17 ± 50% numa-vmstat.node1.nr_page_table_pages 94014 ± 12% +19.7% 112564 ± 9% numa-meminfo.node0.SUnreclaim 164603 ± 67% -70.4% 48754 ± 74% numa-meminfo.node1.Inactive 2357 ± 39% -44.8% 1301 ± 50% numa-meminfo.node1.PageTables 70708 +0.7% 71212 proc-vmstat.nr_inactive_anon 18040 -3.2% 17455 proc-vmstat.nr_kernel_stack 70708 +0.7% 71212 proc-vmstat.nr_zone_inactive_anon 370332 +2.0% 377771 proc-vmstat.pgalloc_normal 157090 ± 41% +34.6% 211411 proc-vmstat.pgfree 3271411 ± 3% -8.3% 3001095 ± 3% perf-stat.i.iTLB-load-misses 2245484 ± 53% -64.3% 802425 ± 66% perf-stat.i.node-load-misses 56.94 ± 24% -22.2 34.72 ± 29% perf-stat.i.node-store-miss-rate% 1090824 ± 57% -65.7% 374199 ± 70% perf-stat.i.node-store-misses 59.80 ± 26% -23.7 36.06 ± 32% perf-stat.overall.node-store-miss-rate% 0.99 ± 45% -48.8% 0.51 ± 31% perf-stat.ps.major-faults 1716083 ± 54% -64.0% 618497 ± 70% perf-stat.ps.node-load-misses 834091 ± 58% -65.4% 288523 ± 74% perf-stat.ps.node-store-misses 487.67 ± 17% -35.0% 317.17 ± 12% slabinfo.biovec-max.active_objs 487.67 ± 17% -35.0% 317.17 ± 12% slabinfo.biovec-max.num_objs 8026 ± 5% +60.8% 12901 ± 3% slabinfo.kmalloc-1k.active_objs 252.33 ± 5% +61.1% 406.50 ± 3% slabinfo.kmalloc-1k.active_slabs 8086 ± 5% +61.0% 13017 ± 3% slabinfo.kmalloc-1k.num_objs 252.33 ± 5% +61.1% 406.50 ± 3% slabinfo.kmalloc-1k.num_slabs 2465 ± 6% -21.0% 1946 ± 14% slabinfo.pool_workqueue.active_objs 2475 ± 6% -20.9% 1958 ± 14% slabinfo.pool_workqueue.num_objs 18532 ± 7% -12.6% 16189 slabinfo.xfs_ili.active_objs 18570 ± 7% -12.6% 16222 slabinfo.xfs_ili.num_objs 57483 ± 5% -10.4% 51530 ± 3% interrupts.CAL:Function_call_interrupts 818.17 ± 45% -38.5% 503.00 interrupts.CPU11.CAL:Function_call_interrupts 572.83 ± 10% -12.6% 500.67 interrupts.CPU15.CAL:Function_call_interrupts 667.17 ± 20% -29.4% 470.83 ± 14% interrupts.CPU17.CAL:Function_call_interrupts 623.17 ± 12% -18.1% 510.50 ± 3% interrupts.CPU18.CAL:Function_call_interrupts 588.83 ± 2% -14.2% 505.00 interrupts.CPU19.CAL:Function_call_interrupts 606.33 ± 7% -17.3% 501.17 interrupts.CPU21.CAL:Function_call_interrupts 907.00 ± 30% -32.8% 609.67 ± 17% interrupts.CPU25.CAL:Function_call_interrupts 588.67 ± 5% -12.2% 516.67 ± 3% interrupts.CPU3.CAL:Function_call_interrupts 604.00 ± 13% -16.0% 507.50 ± 5% interrupts.CPU31.CAL:Function_call_interrupts 573.50 ± 3% -16.0% 481.67 ± 15% interrupts.CPU4.CAL:Function_call_interrupts 617.17 ± 15% -17.7% 507.83 ± 4% interrupts.CPU44.CAL:Function_call_interrupts 595.00 ± 4% -13.4% 515.33 ± 4% interrupts.CPU49.CAL:Function_call_interrupts 572.17 ± 4% -8.5% 523.67 ± 5% interrupts.CPU52.CAL:Function_call_interrupts 581.83 ± 5% -11.6% 514.17 ± 4% interrupts.CPU53.CAL:Function_call_interrupts 578.67 -10.2% 519.50 ± 4% interrupts.CPU54.CAL:Function_call_interrupts 581.00 ± 3% -12.4% 508.67 interrupts.CPU56.CAL:Function_call_interrupts 582.17 ± 3% -11.8% 513.67 interrupts.CPU57.CAL:Function_call_interrupts 581.67 ± 4% -13.2% 504.83 interrupts.CPU59.CAL:Function_call_interrupts 630.83 ± 18% -20.5% 501.50 interrupts.CPU61.CAL:Function_call_interrupts 633.00 ± 26% -19.5% 509.67 ± 3% interrupts.CPU64.CAL:Function_call_interrupts 604.00 ± 14% -16.6% 503.83 ± 4% interrupts.CPU75.CAL:Function_call_interrupts 603.67 ± 12% -16.0% 507.33 ± 4% interrupts.CPU78.CAL:Function_call_interrupts 602.17 ± 13% -16.0% 506.00 ± 4% interrupts.CPU80.CAL:Function_call_interrupts 618.33 ± 13% -14.3% 530.17 ± 7% interrupts.CPU90.CAL:Function_call_interrupts 616.00 ± 13% -15.2% 522.67 ± 5% interrupts.CPU91.CAL:Function_call_interrupts 57.10 ± 15% -16.4 40.71 ± 26% perf-profile.calltrace.cycles-pp.secondary_startup_64_no_verify 57.10 ± 15% -16.4 40.71 ± 26% perf-profile.calltrace.cycles-pp.start_secondary.secondary_startup_64_no_verify 57.10 ± 15% -16.4 40.71 ± 26% perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify 57.10 ± 15% -16.4 40.71 ± 26% perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify 14.67 ± 60% -10.1 4.57 ±148% perf-profile.calltrace.cycles-pp.seq_read_iter.proc_reg_read_iter.new_sync_read.vfs_read.ksys_read 14.67 ± 60% -10.1 4.57 ±148% perf-profile.calltrace.cycles-pp.proc_reg_read_iter.new_sync_read.vfs_read.ksys_read.do_syscall_64 14.67 ± 60% -8.7 6.02 ±161% perf-profile.calltrace.cycles-pp.new_sync_read.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe 12.00 ± 86% -7.4 4.56 ±148% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.read 12.00 ± 86% -7.4 4.56 ±148% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.read 12.00 ± 86% -7.4 4.56 ±148% perf-profile.calltrace.cycles-pp.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read 12.00 ± 86% -7.4 4.56 ±148% perf-profile.calltrace.cycles-pp.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe.read 12.00 ± 86% -7.4 4.56 ±148% perf-profile.calltrace.cycles-pp.read 5.45 ±104% -4.7 0.72 ±223% perf-profile.calltrace.cycles-pp.arch_show_interrupts.seq_read_iter.proc_reg_read_iter.new_sync_read.vfs_read 6.14 ±108% -4.7 1.45 ±223% perf-profile.calltrace.cycles-pp.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe 6.14 ±108% -4.7 1.45 ±223% perf-profile.calltrace.cycles-pp.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe 4.75 ±104% -3.6 1.19 ±223% perf-profile.calltrace.cycles-pp.__x64_sys_getdents64.do_syscall_64.entry_SYSCALL_64_after_hwframe 4.75 ±104% -3.6 1.19 ±223% perf-profile.calltrace.cycles-pp.iterate_dir.__x64_sys_getdents64.do_syscall_64.entry_SYSCALL_64_after_hwframe 6.55 ±100% -2.7 3.84 ±143% perf-profile.calltrace.cycles-pp.show_interrupts.seq_read_iter.proc_reg_read_iter.new_sync_read.vfs_read 57.10 ± 15% -16.4 40.71 ± 26% perf-profile.children.cycles-pp.secondary_startup_64_no_verify 57.10 ± 15% -16.4 40.71 ± 26% perf-profile.children.cycles-pp.start_secondary 57.10 ± 15% -16.4 40.71 ± 26% perf-profile.children.cycles-pp.cpu_startup_entry 57.10 ± 15% -16.4 40.71 ± 26% perf-profile.children.cycles-pp.do_idle 18.14 ± 62% -12.8 5.29 ±155% perf-profile.children.cycles-pp.seq_read_iter 18.14 ± 62% -12.1 6.02 ±161% perf-profile.children.cycles-pp.ksys_read 18.14 ± 62% -12.1 6.02 ±161% perf-profile.children.cycles-pp.vfs_read 14.67 ± 60% -10.1 4.56 ±148% perf-profile.children.cycles-pp.proc_reg_read_iter 14.67 ± 60% -8.7 6.02 ±161% perf-profile.children.cycles-pp.new_sync_read 12.00 ± 86% -6.7 5.29 ±155% perf-profile.children.cycles-pp.read 5.45 ±104% -4.7 0.72 ±223% perf-profile.children.cycles-pp.arch_show_interrupts 5.75 ±105% -4.3 1.45 ±223% perf-profile.children.cycles-pp.vsnprintf 5.75 ±105% -3.6 2.17 ±223% perf-profile.children.cycles-pp.seq_vprintf 5.75 ±105% -3.6 2.17 ±223% perf-profile.children.cycles-pp.seq_printf 4.75 ±104% -3.6 1.19 ±223% perf-profile.children.cycles-pp.__x64_sys_getdents64 4.75 ±104% -3.6 1.19 ±223% perf-profile.children.cycles-pp.iterate_dir 6.55 ±100% -2.7 3.84 ±143% perf-profile.children.cycles-pp.show_interrupts 5.45 ±104% -5.4 0.00 perf-profile.self.cycles-pp.arch_show_interrupts fsmark.files_per_sec 17000 +-------------------------------------------------------------------+ |+.+++ + :++.++ + +.++++.++++.+ +.+ + .++++. + .++ +.+ + .+ + | 16500 |-+ + + ++.+ :: :: + ++ + + + ::+. | 16000 |-+ O O OO O OO+O OOO OO O O+O O O + +| | O O O O O O OOO O O OO OO O O| 15500 |O+ O O O O | 15000 |-+ O O O O O O O | | O | 14500 |-+ | 14000 |-+ O | | O | 13500 |-+ O | 13000 |-+ | | O | 12500 +-------------------------------------------------------------------+ fsmark.time.voluntary_context_switches 350000 +------------------------------------------------------------------+ | O O | 340000 |-+ | | | | O O O O O | 330000 |OO O OO OOO O O OOOO OOOO OOOO OOOOO OOOO OOO OO OO OO OO| | O O | 320000 |-+ O | | | 310000 |-+ | | | | | 300000 |-+ | |++.++++.++++.++++.++++.++++.++++.+++++.++++.++++.++++.++++.++++.++| 290000 +------------------------------------------------------------------+ [*] bisect-good sample [O] bisect-bad sample *************************************************************************************************** lkp-csl-2sp9: 88 threads 2 sockets Intel(R) Xeon(R) Gold 6238M CPU @ 2.10GHz with 128G memory ========================================================================================= compiler/cpufreq_governor/disk/fs/kconfig/load/md/rootfs/tbox_group/test/testcase/ucode: gcc-9/performance/4BRD_12G/xfs/x86_64-rhel-8.3/300/RAID0/debian-10.4-x86_64-20200603.cgz/lkp-csl-2sp9/sync_disk_rw/aim7/0x5003006 commit: 18842e0a4f ("xfs: Fix 64-bit division on 32-bit in xlog_state_switch_iclogs()") a79b28c284 ("xfs: separate CIL commit record IO") 18842e0a4f48564b a79b28c284fd910bb291dbf307a ---------------- --------------------------- %stddev %change %stddev \ | \ 13879 +22.0% 16929 aim7.jobs-per-min 129.73 -18.0% 106.37 aim7.time.elapsed_time 129.73 -18.0% 106.37 aim7.time.elapsed_time.max 1647556 +16.5% 1919576 aim7.time.involuntary_context_switches 41390 ± 3% -11.2% 36759 aim7.time.minor_page_faults 5461 -17.9% 4483 aim7.time.system_time 72997986 -10.5% 65359678 aim7.time.voluntary_context_switches 0.01 +0.0 0.02 ± 9% mpstat.cpu.all.iowait% 0.54 -0.1 0.48 ± 2% mpstat.cpu.all.usr% 2982166 -9.3% 2704857 numa-numastat.node1.local_node 3008831 -9.0% 2737175 numa-numastat.node1.numa_hit 58057 +10.9% 64387 ± 5% slabinfo.anon_vma_chain.active_objs 58142 +10.7% 64387 ± 5% slabinfo.anon_vma_chain.num_objs 168.11 -13.9% 144.77 uptime.boot 8419 -11.9% 7419 uptime.idle 558231 +21.8% 679782 vmstat.io.bo 3562019 -20.1% 2846094 vmstat.memory.cache 1098857 +12.1% 1231460 vmstat.system.cs 9.824e+08 +11.9% 1.099e+09 ± 2% cpuidle.C1.time 21821055 +11.0% 24216149 ± 2% cpuidle.C1.usage 3.872e+09 ± 10% -29.1% 2.745e+09 ± 15% cpuidle.C1E.time 46682234 -20.9% 36905264 ± 3% cpuidle.C1E.usage 1122822 +12.3% 1260895 cpuidle.POLL.usage 8.51 +3.1 11.59 turbostat.C1% 46681814 -20.9% 36904474 ± 3% turbostat.C1E 29377272 -16.7% 24460328 ± 3% turbostat.IRQ 64.33 -5.7% 60.67 turbostat.PkgTmp 55.96 +1.3% 56.68 turbostat.RAMWatt 335419 ± 2% -69.3% 102817 ± 8% meminfo.Active 335163 ± 2% -69.4% 102561 ± 8% meminfo.Active(anon) 131608 -13.7% 113590 meminfo.AnonHugePages 3403302 -20.7% 2700472 meminfo.Cached 1483499 -47.5% 778943 ± 2% meminfo.Committed_AS 977192 -48.9% 499429 ± 2% meminfo.Inactive 831379 -57.0% 357444 ± 3% meminfo.Inactive(anon) 364174 -83.5% 60025 ± 3% meminfo.Mapped 5288714 -13.5% 4573105 meminfo.Memused 888801 -78.6% 189797 ± 10% meminfo.Shmem 5365908 -13.0% 4670485 meminfo.max_used_kB 91803 ± 6% -78.3% 19879 ± 32% numa-meminfo.node0.Active 91632 ± 6% -78.5% 19709 ± 32% numa-meminfo.node0.Active(anon) 144688 ± 8% -70.1% 43249 ± 9% numa-meminfo.node0.Mapped 278561 ± 5% -79.3% 57643 ± 37% numa-meminfo.node0.Shmem 244509 -65.6% 84084 ± 2% numa-meminfo.node1.Active 244424 -65.6% 83999 ± 2% numa-meminfo.node1.Active(anon) 1475164 ± 64% -75.5% 361168 ± 29% numa-meminfo.node1.FilePages 555525 ± 16% -65.2% 193567 ± 32% numa-meminfo.node1.Inactive 482733 ± 19% -74.4% 123553 ± 50% numa-meminfo.node1.Inactive(anon) 221734 ± 5% -92.4% 16893 ± 11% numa-meminfo.node1.Mapped 2341429 ± 42% -51.0% 1147446 ± 25% numa-meminfo.node1.MemUsed 612619 ± 2% -78.0% 134759 ± 4% numa-meminfo.node1.Shmem 22915 ± 6% -78.5% 4927 ± 32% numa-vmstat.node0.nr_active_anon 36199 ± 8% -70.1% 10813 ± 9% numa-vmstat.node0.nr_mapped 69661 ± 5% -79.3% 14415 ± 36% numa-vmstat.node0.nr_shmem 22915 ± 6% -78.5% 4927 ± 32% numa-vmstat.node0.nr_zone_active_anon 7370 ± 5% -6.5% 6890 ± 3% numa-vmstat.node0.nr_zone_write_pending 61128 -65.6% 21003 ± 2% numa-vmstat.node1.nr_active_anon 368867 ± 64% -75.5% 90307 ± 29% numa-vmstat.node1.nr_file_pages 120723 ± 19% -74.4% 30895 ± 50% numa-vmstat.node1.nr_inactive_anon 55470 ± 6% -92.4% 4222 ± 11% numa-vmstat.node1.nr_mapped 153215 ± 2% -78.0% 33695 ± 4% numa-vmstat.node1.nr_shmem 61128 -65.6% 21002 ± 2% numa-vmstat.node1.nr_zone_active_anon 120723 ± 19% -74.4% 30894 ± 50% numa-vmstat.node1.nr_zone_inactive_anon 83793 ± 2% -69.4% 25645 ± 8% proc-vmstat.nr_active_anon 69383 -2.5% 67666 proc-vmstat.nr_anon_pages 850838 -20.7% 675125 proc-vmstat.nr_file_pages 207856 -57.0% 89363 ± 3% proc-vmstat.nr_inactive_anon 36450 -2.6% 35495 proc-vmstat.nr_inactive_file 91055 -83.5% 15006 ± 3% proc-vmstat.nr_mapped 222214 -78.6% 47456 ± 11% proc-vmstat.nr_shmem 83793 ± 2% -69.4% 25645 ± 8% proc-vmstat.nr_zone_active_anon 207856 -57.0% 89363 ± 3% proc-vmstat.nr_zone_inactive_anon 36450 -2.6% 35495 proc-vmstat.nr_zone_inactive_file 15021 ± 2% -5.8% 14149 proc-vmstat.nr_zone_write_pending 266167 -35.6% 171458 ± 11% proc-vmstat.numa_hint_faults 140448 ± 3% -37.3% 88003 ± 10% proc-vmstat.numa_hint_faults_local 5650544 -6.4% 5290494 ± 2% proc-vmstat.numa_hit 5570801 -6.5% 5210756 ± 2% proc-vmstat.numa_local 380199 -50.1% 189829 ± 8% proc-vmstat.numa_pte_updates 5717786 -5.7% 5391070 proc-vmstat.pgalloc_normal 808432 -25.4% 602788 ± 3% proc-vmstat.pgfault 29460 -13.0% 25626 ± 2% proc-vmstat.pgreuse 4.307e+09 +3.6% 4.464e+09 perf-stat.i.branch-instructions 39861920 +11.3% 44384422 perf-stat.i.branch-misses 23.30 +0.8 24.14 perf-stat.i.cache-miss-rate% 52855610 +17.7% 62197817 perf-stat.i.cache-misses 2.166e+08 +12.0% 2.426e+08 perf-stat.i.cache-references 1118434 +12.8% 1261487 perf-stat.i.context-switches 6.75 -3.9% 6.48 perf-stat.i.cpi 1.308e+11 +1.1% 1.322e+11 perf-stat.i.cpu-cycles 178676 +17.9% 210712 perf-stat.i.cpu-migrations 2593 ± 4% -12.7% 2265 ± 7% perf-stat.i.cycles-between-cache-misses 5.075e+09 +5.3% 5.342e+09 perf-stat.i.dTLB-loads 512902 ± 7% +17.5% 602445 ± 5% perf-stat.i.dTLB-store-misses 1.715e+09 +10.7% 1.899e+09 perf-stat.i.dTLB-stores 8968888 +12.4% 10078328 perf-stat.i.iTLB-load-misses 16507011 +15.0% 18977719 ± 3% perf-stat.i.iTLB-loads 1.925e+10 +4.6% 2.014e+10 perf-stat.i.instructions 2247 -4.8% 2139 perf-stat.i.instructions-per-iTLB-miss 1.49 +1.1% 1.50 perf-stat.i.metric.GHz 610.38 ± 2% +19.5% 729.31 ± 5% perf-stat.i.metric.K/sec 128.52 +5.6% 135.69 perf-stat.i.metric.M/sec 5939 -9.8% 5356 ± 2% perf-stat.i.minor-faults 19010898 +18.2% 22479012 perf-stat.i.node-load-misses 2079877 +11.3% 2314010 perf-stat.i.node-loads 8988994 +17.4% 10554292 perf-stat.i.node-store-misses 1813880 +13.9% 2065595 perf-stat.i.node-stores 5941 -9.8% 5359 ± 2% perf-stat.i.page-faults 11.25 +7.1% 12.05 perf-stat.overall.MPKI 0.93 +0.1 0.99 perf-stat.overall.branch-miss-rate% 24.41 +1.2 25.64 perf-stat.overall.cache-miss-rate% 6.79 -3.3% 6.57 perf-stat.overall.cpi 2474 -14.1% 2125 perf-stat.overall.cycles-between-cache-misses 2146 -6.9% 1998 perf-stat.overall.instructions-per-iTLB-miss 0.15 +3.4% 0.15 perf-stat.overall.ipc 4.274e+09 +3.5% 4.423e+09 perf-stat.ps.branch-instructions 39550311 +11.2% 43964111 perf-stat.ps.branch-misses 52450879 +17.5% 61628414 perf-stat.ps.cache-misses 2.149e+08 +11.9% 2.404e+08 perf-stat.ps.cache-references 1109849 +12.6% 1249872 perf-stat.ps.context-switches 177305 +17.8% 208787 perf-stat.ps.cpu-migrations 5.037e+09 +5.1% 5.293e+09 perf-stat.ps.dTLB-loads 508884 ± 7% +17.3% 597002 ± 5% perf-stat.ps.dTLB-store-misses 1.702e+09 +10.5% 1.881e+09 perf-stat.ps.dTLB-stores 8901107 +12.2% 9985139 perf-stat.ps.iTLB-load-misses 16380357 +14.8% 18802343 ± 3% perf-stat.ps.iTLB-loads 1.911e+10 +4.4% 1.995e+10 perf-stat.ps.instructions 5893 -10.0% 5301 ± 2% perf-stat.ps.minor-faults 18864869 +18.1% 22272948 perf-stat.ps.node-load-misses 2063954 +11.1% 2293240 perf-stat.ps.node-loads 8920023 +17.2% 10457333 perf-stat.ps.node-store-misses 1800221 +13.7% 2046553 perf-stat.ps.node-stores 5895 -10.0% 5304 ± 2% perf-stat.ps.page-faults 2.486e+12 -14.2% 2.134e+12 perf-stat.total.instructions 5157270 -6.3% 4834857 interrupts.CAL:Function_call_interrupts 258813 -19.8% 207639 ± 4% interrupts.CPU0.LOC:Local_timer_interrupts 258857 -20.0% 207130 ± 4% interrupts.CPU1.LOC:Local_timer_interrupts 258867 -19.9% 207436 ± 4% interrupts.CPU10.LOC:Local_timer_interrupts 258829 -19.8% 207511 ± 4% interrupts.CPU11.LOC:Local_timer_interrupts 258768 -19.8% 207624 ± 4% interrupts.CPU12.LOC:Local_timer_interrupts 258820 -19.8% 207594 ± 4% interrupts.CPU13.LOC:Local_timer_interrupts 258806 -19.8% 207625 ± 4% interrupts.CPU14.LOC:Local_timer_interrupts 258664 -19.6% 207905 ± 4% interrupts.CPU15.LOC:Local_timer_interrupts 6059 ± 2% -8.9% 5518 ± 2% interrupts.CPU15.RES:Rescheduling_interrupts 258859 -19.8% 207619 ± 4% interrupts.CPU16.LOC:Local_timer_interrupts 6175 ± 4% -10.3% 5540 ± 3% interrupts.CPU16.RES:Rescheduling_interrupts 258722 -19.7% 207647 ± 4% interrupts.CPU17.LOC:Local_timer_interrupts 258646 -19.7% 207580 ± 4% interrupts.CPU18.LOC:Local_timer_interrupts 258799 -19.8% 207564 ± 4% interrupts.CPU19.LOC:Local_timer_interrupts 258801 -19.8% 207563 ± 4% interrupts.CPU2.LOC:Local_timer_interrupts 258723 -19.8% 207565 ± 4% interrupts.CPU20.LOC:Local_timer_interrupts 258804 -19.8% 207615 ± 4% interrupts.CPU21.LOC:Local_timer_interrupts 57725 ± 2% -7.5% 53408 ± 4% interrupts.CPU22.CAL:Function_call_interrupts 258842 -20.3% 206199 ± 5% interrupts.CPU22.LOC:Local_timer_interrupts 57984 ± 2% -7.6% 53551 ± 3% interrupts.CPU23.CAL:Function_call_interrupts 258760 -20.3% 206156 ± 5% interrupts.CPU23.LOC:Local_timer_interrupts 58472 -8.1% 53738 ± 3% interrupts.CPU24.CAL:Function_call_interrupts 258969 -20.4% 206187 ± 5% interrupts.CPU24.LOC:Local_timer_interrupts 59121 -8.9% 53835 ± 3% interrupts.CPU25.CAL:Function_call_interrupts 258770 -20.3% 206220 ± 5% interrupts.CPU25.LOC:Local_timer_interrupts 59203 -8.8% 53976 ± 4% interrupts.CPU26.CAL:Function_call_interrupts 258771 -20.3% 206166 ± 5% interrupts.CPU26.LOC:Local_timer_interrupts 57731 ± 2% -7.9% 53145 ± 3% interrupts.CPU27.CAL:Function_call_interrupts 258765 -20.3% 206110 ± 5% interrupts.CPU27.LOC:Local_timer_interrupts 59022 -9.1% 53645 ± 4% interrupts.CPU28.CAL:Function_call_interrupts 259049 -20.5% 206059 ± 5% interrupts.CPU28.LOC:Local_timer_interrupts 58573 -8.2% 53783 ± 4% interrupts.CPU29.CAL:Function_call_interrupts 258649 -20.3% 206190 ± 5% interrupts.CPU29.LOC:Local_timer_interrupts 258909 -19.9% 207403 ± 4% interrupts.CPU3.LOC:Local_timer_interrupts 58830 -8.8% 53648 ± 3% interrupts.CPU30.CAL:Function_call_interrupts 258775 -20.3% 206235 ± 5% interrupts.CPU30.LOC:Local_timer_interrupts 58760 -8.4% 53810 ± 3% interrupts.CPU31.CAL:Function_call_interrupts 258780 -20.4% 206118 ± 5% interrupts.CPU31.LOC:Local_timer_interrupts 58551 -8.7% 53477 ± 3% interrupts.CPU32.CAL:Function_call_interrupts 258680 -20.3% 206191 ± 5% interrupts.CPU32.LOC:Local_timer_interrupts 58657 ± 2% -8.8% 53517 ± 4% interrupts.CPU33.CAL:Function_call_interrupts 258752 -20.4% 206003 ± 5% interrupts.CPU33.LOC:Local_timer_interrupts 58161 -7.4% 53868 ± 4% interrupts.CPU34.CAL:Function_call_interrupts 258778 -20.4% 206106 ± 5% interrupts.CPU34.LOC:Local_timer_interrupts 58619 -8.4% 53676 ± 4% interrupts.CPU35.CAL:Function_call_interrupts 258721 -20.3% 206175 ± 5% interrupts.CPU35.LOC:Local_timer_interrupts 58349 -7.6% 53907 ± 4% interrupts.CPU36.CAL:Function_call_interrupts 258753 -20.3% 206154 ± 5% interrupts.CPU36.LOC:Local_timer_interrupts 58714 -9.2% 53323 ± 4% interrupts.CPU37.CAL:Function_call_interrupts 258694 -20.2% 206499 ± 5% interrupts.CPU37.LOC:Local_timer_interrupts 58291 -7.7% 53799 ± 4% interrupts.CPU38.CAL:Function_call_interrupts 259023 -20.4% 206168 ± 5% interrupts.CPU38.LOC:Local_timer_interrupts 58924 -7.8% 54338 ± 4% interrupts.CPU39.CAL:Function_call_interrupts 258920 -20.4% 206187 ± 5% interrupts.CPU39.LOC:Local_timer_interrupts 258817 -19.8% 207532 ± 4% interrupts.CPU4.LOC:Local_timer_interrupts 5957 ± 3% -8.7% 5441 ± 3% interrupts.CPU4.RES:Rescheduling_interrupts 258718 -20.3% 206124 ± 5% interrupts.CPU40.LOC:Local_timer_interrupts 57580 ± 2% -7.3% 53377 ± 4% interrupts.CPU41.CAL:Function_call_interrupts 258739 -20.3% 206212 ± 5% interrupts.CPU41.LOC:Local_timer_interrupts 57848 ± 2% -7.7% 53401 ± 3% interrupts.CPU42.CAL:Function_call_interrupts 258729 -20.3% 206208 ± 5% interrupts.CPU42.LOC:Local_timer_interrupts 58375 -7.0% 54263 ± 4% interrupts.CPU43.CAL:Function_call_interrupts 258771 -20.3% 206256 ± 5% interrupts.CPU43.LOC:Local_timer_interrupts 258781 -19.8% 207611 ± 4% interrupts.CPU44.LOC:Local_timer_interrupts 258798 -19.8% 207429 ± 4% interrupts.CPU45.LOC:Local_timer_interrupts 258765 -19.8% 207525 ± 4% interrupts.CPU46.LOC:Local_timer_interrupts 5658 ± 2% -8.3% 5188 ± 2% interrupts.CPU46.RES:Rescheduling_interrupts 258787 -19.8% 207521 ± 4% interrupts.CPU47.LOC:Local_timer_interrupts 5631 -9.1% 5116 ± 2% interrupts.CPU47.RES:Rescheduling_interrupts 258813 -19.8% 207610 ± 4% interrupts.CPU48.LOC:Local_timer_interrupts 258805 -19.8% 207627 ± 4% interrupts.CPU49.LOC:Local_timer_interrupts 258833 -19.6% 207990 ± 4% interrupts.CPU5.LOC:Local_timer_interrupts 258740 -19.8% 207549 ± 4% interrupts.CPU50.LOC:Local_timer_interrupts 258775 -19.8% 207483 ± 4% interrupts.CPU51.LOC:Local_timer_interrupts 258790 -19.8% 207538 ± 4% interrupts.CPU52.LOC:Local_timer_interrupts 258802 -19.8% 207564 ± 4% interrupts.CPU53.LOC:Local_timer_interrupts 258827 -19.8% 207583 ± 4% interrupts.CPU54.LOC:Local_timer_interrupts 258601 -19.8% 207522 ± 4% interrupts.CPU55.LOC:Local_timer_interrupts 5690 ± 2% -7.7% 5251 ± 2% interrupts.CPU55.RES:Rescheduling_interrupts 258790 -19.8% 207540 ± 4% interrupts.CPU56.LOC:Local_timer_interrupts 258822 -19.8% 207524 ± 4% interrupts.CPU57.LOC:Local_timer_interrupts 258775 -19.8% 207552 ± 4% interrupts.CPU58.LOC:Local_timer_interrupts 258818 -19.8% 207591 ± 4% interrupts.CPU59.LOC:Local_timer_interrupts 258798 -19.8% 207591 ± 4% interrupts.CPU6.LOC:Local_timer_interrupts 258785 -19.8% 207634 ± 4% interrupts.CPU60.LOC:Local_timer_interrupts 258835 -19.8% 207540 ± 4% interrupts.CPU61.LOC:Local_timer_interrupts 5474 -8.1% 5032 ± 3% interrupts.CPU61.RES:Rescheduling_interrupts 258841 -19.8% 207593 ± 4% interrupts.CPU62.LOC:Local_timer_interrupts 258828 -19.8% 207511 ± 4% interrupts.CPU63.LOC:Local_timer_interrupts 258799 -19.8% 207562 ± 4% interrupts.CPU64.LOC:Local_timer_interrupts 258810 -19.8% 207574 ± 4% interrupts.CPU65.LOC:Local_timer_interrupts 58711 -8.7% 53620 ± 4% interrupts.CPU66.CAL:Function_call_interrupts 258759 -20.3% 206149 ± 5% interrupts.CPU66.LOC:Local_timer_interrupts 58760 -8.7% 53648 ± 4% interrupts.CPU67.CAL:Function_call_interrupts 258798 -20.3% 206243 ± 5% interrupts.CPU67.LOC:Local_timer_interrupts 5834 -8.3% 5348 ± 4% interrupts.CPU67.RES:Rescheduling_interrupts 58749 -8.8% 53569 ± 4% interrupts.CPU68.CAL:Function_call_interrupts 258762 -20.3% 206156 ± 5% interrupts.CPU68.LOC:Local_timer_interrupts 58503 -8.5% 53511 ± 4% interrupts.CPU69.CAL:Function_call_interrupts 258766 -20.3% 206179 ± 5% interrupts.CPU69.LOC:Local_timer_interrupts 258815 -19.8% 207607 ± 4% interrupts.CPU7.LOC:Local_timer_interrupts 58243 -9.0% 53008 ± 4% interrupts.CPU70.CAL:Function_call_interrupts 258760 -20.3% 206133 ± 5% interrupts.CPU70.LOC:Local_timer_interrupts 58155 -8.6% 53157 ± 4% interrupts.CPU71.CAL:Function_call_interrupts 258816 -20.3% 206150 ± 5% interrupts.CPU71.LOC:Local_timer_interrupts 58617 -8.4% 53666 ± 4% interrupts.CPU72.CAL:Function_call_interrupts 258800 -20.3% 206195 ± 5% interrupts.CPU72.LOC:Local_timer_interrupts 56202 ± 3% -5.3% 53235 ± 4% interrupts.CPU73.CAL:Function_call_interrupts 258813 -20.4% 206117 ± 5% interrupts.CPU73.LOC:Local_timer_interrupts 58430 -8.8% 53264 ± 4% interrupts.CPU74.CAL:Function_call_interrupts 258763 -20.3% 206118 ± 5% interrupts.CPU74.LOC:Local_timer_interrupts 58240 -8.7% 53145 ± 4% interrupts.CPU75.CAL:Function_call_interrupts 258774 -20.3% 206145 ± 5% interrupts.CPU75.LOC:Local_timer_interrupts 58193 -7.9% 53585 ± 4% interrupts.CPU76.CAL:Function_call_interrupts 258747 -20.3% 206179 ± 5% interrupts.CPU76.LOC:Local_timer_interrupts 6221 -14.0% 5352 ± 2% interrupts.CPU76.RES:Rescheduling_interrupts 59053 -9.8% 53280 ± 4% interrupts.CPU77.CAL:Function_call_interrupts 258776 -20.3% 206177 ± 5% interrupts.CPU77.LOC:Local_timer_interrupts 6068 -15.4% 5132 interrupts.CPU77.RES:Rescheduling_interrupts 57716 -8.0% 53124 ± 4% interrupts.CPU78.CAL:Function_call_interrupts 258754 -20.3% 206213 ± 5% interrupts.CPU78.LOC:Local_timer_interrupts 5830 -13.3% 5056 interrupts.CPU78.RES:Rescheduling_interrupts 58692 -8.6% 53667 ± 4% interrupts.CPU79.CAL:Function_call_interrupts 258794 -20.3% 206167 ± 5% interrupts.CPU79.LOC:Local_timer_interrupts 6035 -14.1% 5186 interrupts.CPU79.RES:Rescheduling_interrupts 258767 -19.8% 207598 ± 4% interrupts.CPU8.LOC:Local_timer_interrupts 58102 -8.1% 53421 ± 4% interrupts.CPU80.CAL:Function_call_interrupts 258638 -20.3% 206237 ± 5% interrupts.CPU80.LOC:Local_timer_interrupts 5924 ± 2% -11.9% 5217 ± 2% interrupts.CPU80.RES:Rescheduling_interrupts 58216 -9.3% 52787 ± 3% interrupts.CPU81.CAL:Function_call_interrupts 258674 -20.3% 206164 ± 5% interrupts.CPU81.LOC:Local_timer_interrupts 6053 -11.9% 5335 interrupts.CPU81.RES:Rescheduling_interrupts 58210 -8.7% 53122 ± 4% interrupts.CPU82.CAL:Function_call_interrupts 258680 -20.3% 206108 ± 5% interrupts.CPU82.LOC:Local_timer_interrupts 6617 ± 5% -18.7% 5379 interrupts.CPU82.RES:Rescheduling_interrupts 58532 -9.1% 53179 ± 4% interrupts.CPU83.CAL:Function_call_interrupts 258721 -20.3% 206164 ± 5% interrupts.CPU83.LOC:Local_timer_interrupts 5855 ± 2% -14.3% 5015 ± 2% interrupts.CPU83.RES:Rescheduling_interrupts 58049 -8.3% 53240 ± 4% interrupts.CPU84.CAL:Function_call_interrupts 258762 -20.4% 206103 ± 5% interrupts.CPU84.LOC:Local_timer_interrupts 58529 -8.7% 53457 ± 4% interrupts.CPU85.CAL:Function_call_interrupts 258746 -20.3% 206226 ± 5% interrupts.CPU85.LOC:Local_timer_interrupts 58118 -8.2% 53328 ± 4% interrupts.CPU86.CAL:Function_call_interrupts 258769 -20.4% 206092 ± 5% interrupts.CPU86.LOC:Local_timer_interrupts 258862 -20.3% 206240 ± 5% interrupts.CPU87.LOC:Local_timer_interrupts 258817 -19.8% 207486 ± 4% interrupts.CPU9.LOC:Local_timer_interrupts 5855 -9.1% 5321 interrupts.CPU9.RES:Rescheduling_interrupts 2099 -23.1% 1613 ± 8% interrupts.IWI:IRQ_work_interrupts 22773085 -20.1% 18204614 ± 4% interrupts.LOC:Local_timer_interrupts 18980 -11.0% 16886 ± 3% softirqs.CPU0.RCU 33504 -15.1% 28441 softirqs.CPU0.SCHED 17793 ± 7% -13.5% 15389 ± 5% softirqs.CPU1.RCU 31237 ± 4% -13.8% 26937 ± 6% softirqs.CPU1.SCHED 16570 ± 2% -12.4% 14507 ± 4% softirqs.CPU10.RCU 29665 -14.1% 25496 ± 2% softirqs.CPU10.SCHED 30185 -17.4% 24946 ± 2% softirqs.CPU11.SCHED 30314 ± 2% -15.8% 25525 ± 2% softirqs.CPU12.SCHED 16626 ± 2% -12.0% 14628 ± 4% softirqs.CPU13.RCU 29804 -15.7% 25129 softirqs.CPU13.SCHED 29843 -15.9% 25104 softirqs.CPU14.SCHED 29723 -14.0% 25574 ± 2% softirqs.CPU15.SCHED 16869 ± 2% -12.7% 14725 ± 4% softirqs.CPU16.RCU 30037 -15.9% 25247 ± 2% softirqs.CPU16.SCHED 16731 ± 3% -10.3% 15002 ± 5% softirqs.CPU17.RCU 29838 -16.1% 25032 softirqs.CPU17.SCHED 17111 ± 4% -13.0% 14891 ± 2% softirqs.CPU18.RCU 30271 -15.8% 25493 softirqs.CPU18.SCHED 29760 -14.1% 25557 ± 4% softirqs.CPU19.SCHED 31874 ± 2% -16.9% 26476 ± 2% softirqs.CPU2.SCHED 18355 ± 8% -18.4% 14976 ± 6% softirqs.CPU20.RCU 29939 -15.5% 25295 ± 2% softirqs.CPU20.SCHED 29891 -15.7% 25204 ± 2% softirqs.CPU21.SCHED 16564 -11.1% 14727 softirqs.CPU22.RCU 29167 ± 2% -13.1% 25349 softirqs.CPU22.SCHED 16200 ± 2% -10.0% 14581 ± 2% softirqs.CPU23.RCU 29737 ± 2% -15.4% 25156 softirqs.CPU23.SCHED 29658 -15.3% 25110 softirqs.CPU24.SCHED 16316 -11.0% 14527 ± 2% softirqs.CPU25.RCU 29893 -16.3% 25012 softirqs.CPU25.SCHED 16615 -11.4% 14715 softirqs.CPU26.RCU 29791 -16.5% 24880 softirqs.CPU26.SCHED 16479 -9.0% 14999 ± 5% softirqs.CPU27.RCU 29569 -14.1% 25401 softirqs.CPU27.SCHED 16763 ± 3% -11.5% 14836 softirqs.CPU28.RCU 30093 -16.1% 25253 softirqs.CPU28.SCHED 16407 -11.7% 14482 ± 2% softirqs.CPU29.RCU 29681 -15.3% 25135 softirqs.CPU29.SCHED 17138 ± 4% -13.0% 14913 ± 5% softirqs.CPU3.RCU 29927 -13.4% 25914 ± 2% softirqs.CPU3.SCHED 16520 -11.6% 14611 ± 2% softirqs.CPU30.RCU 29770 -16.3% 24908 softirqs.CPU30.SCHED 16676 ± 2% -12.4% 14606 ± 4% softirqs.CPU31.RCU 29802 -15.8% 25086 softirqs.CPU31.SCHED 16998 ± 3% -13.3% 14735 ± 2% softirqs.CPU32.RCU 30551 ± 4% -17.1% 25319 softirqs.CPU32.SCHED 16477 -10.7% 14720 softirqs.CPU33.RCU 29801 ± 2% -16.0% 25025 softirqs.CPU33.SCHED 16949 ± 3% -10.6% 15156 ± 4% softirqs.CPU34.RCU 29803 -15.6% 25147 softirqs.CPU34.SCHED 16807 ± 3% -11.8% 14821 softirqs.CPU35.RCU 30197 -16.3% 25273 softirqs.CPU35.SCHED 29782 -15.8% 25072 softirqs.CPU36.SCHED 16477 -10.3% 14772 softirqs.CPU37.RCU 29601 -13.9% 25498 softirqs.CPU37.SCHED 16904 ± 2% -13.0% 14701 ± 2% softirqs.CPU38.RCU 29795 ± 2% -16.6% 24842 softirqs.CPU38.SCHED 16908 ± 3% -14.5% 14453 ± 2% softirqs.CPU39.RCU 29966 -16.5% 25032 softirqs.CPU39.SCHED 29868 -14.2% 25620 softirqs.CPU4.SCHED 17225 ± 6% -16.1% 14454 ± 4% softirqs.CPU40.RCU 29566 -15.3% 25057 softirqs.CPU40.SCHED 16558 -13.1% 14393 ± 4% softirqs.CPU41.RCU 29721 -14.5% 25411 softirqs.CPU41.SCHED 16780 -13.3% 14543 ± 3% softirqs.CPU42.RCU 29813 -15.9% 25065 softirqs.CPU42.SCHED 16691 -11.4% 14793 ± 2% softirqs.CPU43.RCU 29058 ± 2% -15.9% 24424 softirqs.CPU43.SCHED 29016 -13.3% 25153 ± 2% softirqs.CPU44.SCHED 16236 -11.5% 14361 ± 4% softirqs.CPU45.RCU 29485 -13.4% 25525 softirqs.CPU45.SCHED 29643 -16.0% 24890 ± 2% softirqs.CPU46.SCHED 29747 -14.7% 25386 softirqs.CPU47.SCHED 29573 -14.2% 25365 softirqs.CPU48.SCHED 16929 -14.7% 14446 ± 4% softirqs.CPU49.RCU 29866 -15.5% 25238 ± 2% softirqs.CPU49.SCHED 30110 -15.5% 25444 ± 2% softirqs.CPU5.SCHED 16731 ± 4% -14.9% 14246 ± 4% softirqs.CPU50.RCU 29804 -14.9% 25359 ± 2% softirqs.CPU50.SCHED 16865 ± 5% -14.2% 14467 ± 6% softirqs.CPU51.RCU 29674 -15.9% 24950 ± 2% softirqs.CPU51.SCHED 16599 -11.1% 14751 ± 5% softirqs.CPU52.RCU 29968 -15.0% 25467 softirqs.CPU52.SCHED 29848 -15.2% 25326 ± 2% softirqs.CPU53.SCHED 16584 ± 3% -10.5% 14841 ± 7% softirqs.CPU54.RCU 29831 -14.9% 25379 ± 2% softirqs.CPU54.SCHED 16716 ± 4% -12.2% 14669 ± 5% softirqs.CPU55.RCU 29856 -15.6% 25207 ± 2% softirqs.CPU55.SCHED 16415 ± 2% -10.8% 14642 ± 4% softirqs.CPU56.RCU 29554 -14.9% 25148 softirqs.CPU56.SCHED 16554 ± 3% -12.6% 14467 ± 5% softirqs.CPU57.RCU 30048 -16.5% 25096 ± 2% softirqs.CPU57.SCHED 16407 ± 2% -9.8% 14803 ± 8% softirqs.CPU58.RCU 29805 -15.5% 25191 softirqs.CPU58.SCHED 16966 ± 6% -14.2% 14563 ± 4% softirqs.CPU59.RCU 29788 -14.6% 25446 ± 2% softirqs.CPU59.SCHED 17270 ± 5% -14.1% 14842 ± 5% softirqs.CPU6.RCU 29950 -14.1% 25724 ± 2% softirqs.CPU6.SCHED 16423 ± 4% -10.4% 14707 ± 6% softirqs.CPU60.RCU 29734 -13.7% 25658 ± 2% softirqs.CPU60.SCHED 16520 ± 2% -11.5% 14614 ± 5% softirqs.CPU61.RCU 29831 -15.3% 25259 ± 2% softirqs.CPU61.SCHED 29301 -13.6% 25312 softirqs.CPU62.SCHED 16495 ± 2% -12.0% 14520 ± 4% softirqs.CPU63.RCU 29708 -14.3% 25458 softirqs.CPU63.SCHED 16599 ± 2% -13.4% 14369 ± 3% softirqs.CPU64.RCU 29984 -16.0% 25189 softirqs.CPU64.SCHED 16851 ± 3% -10.8% 15031 ± 8% softirqs.CPU65.RCU 29863 -15.0% 25384 ± 2% softirqs.CPU65.SCHED 18595 ± 15% -20.2% 14841 ± 2% softirqs.CPU66.RCU 29553 -14.8% 25187 softirqs.CPU66.SCHED 30196 -15.7% 25465 ± 2% softirqs.CPU67.SCHED 16409 -9.6% 14827 softirqs.CPU68.RCU 30172 -16.8% 25097 softirqs.CPU68.SCHED 30371 ± 3% -15.4% 25695 ± 3% softirqs.CPU69.SCHED 16520 ± 2% -11.0% 14710 ± 5% softirqs.CPU7.RCU 29770 -15.4% 25176 ± 2% softirqs.CPU7.SCHED 16658 ± 2% -10.7% 14868 ± 4% softirqs.CPU70.RCU 29774 -15.6% 25123 softirqs.CPU70.SCHED 16370 -11.6% 14475 ± 4% softirqs.CPU71.RCU 29799 -16.1% 24998 softirqs.CPU71.SCHED 30299 -17.7% 24938 softirqs.CPU72.SCHED 29062 ± 3% -14.0% 24986 softirqs.CPU73.SCHED 16933 ± 5% -11.7% 14945 softirqs.CPU74.RCU 30213 ± 2% -16.7% 25171 softirqs.CPU74.SCHED 15997 -11.2% 14199 ± 4% softirqs.CPU75.RCU 29767 -15.3% 25220 softirqs.CPU75.SCHED 16015 -11.2% 14225 ± 3% softirqs.CPU76.RCU 29681 -15.3% 25132 softirqs.CPU76.SCHED 29659 -15.6% 25036 softirqs.CPU77.SCHED 15990 -12.1% 14048 ± 3% softirqs.CPU78.RCU 29814 -16.1% 25013 softirqs.CPU78.SCHED 16210 ± 2% -13.0% 14101 ± 3% softirqs.CPU79.RCU 29738 -15.8% 25042 softirqs.CPU79.SCHED 16905 ± 5% -14.1% 14527 ± 4% softirqs.CPU8.RCU 29852 -15.5% 25239 ± 2% softirqs.CPU8.SCHED 16792 ± 6% -14.9% 14293 ± 3% softirqs.CPU80.RCU 30105 -17.0% 24988 softirqs.CPU80.SCHED 16428 ± 2% -13.3% 14236 ± 3% softirqs.CPU81.RCU 29787 -16.0% 25019 softirqs.CPU81.SCHED 29496 -14.4% 25253 softirqs.CPU82.SCHED 29706 -14.9% 25293 softirqs.CPU83.SCHED 15801 -10.1% 14210 softirqs.CPU84.RCU 29653 -15.1% 25181 softirqs.CPU84.SCHED 16095 -12.9% 14020 softirqs.CPU85.RCU 29799 -15.0% 25323 softirqs.CPU85.SCHED 16123 -10.7% 14391 ± 2% softirqs.CPU86.RCU 29644 -14.0% 25502 ± 2% softirqs.CPU86.SCHED 16822 ± 2% -13.1% 14626 softirqs.CPU87.RCU 28356 -13.3% 24594 softirqs.CPU87.SCHED 17085 ± 3% -13.2% 14838 ± 6% softirqs.CPU9.RCU 29917 -14.3% 25653 ± 3% softirqs.CPU9.SCHED 1467534 -11.5% 1298751 ± 3% softirqs.RCU 2628721 -15.3% 2226374 softirqs.SCHED 20535 -9.4% 18603 softirqs.TIMER 11.00 -10.2 0.77 ± 4% perf-profile.calltrace.cycles-pp.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write.new_sync_write.vfs_write 7.10 -7.1 0.00 perf-profile.calltrace.cycles-pp._raw_spin_lock.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write.new_sync_write 7.08 -7.1 0.00 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write 68.11 -1.6 66.52 perf-profile.calltrace.cycles-pp.xfs_file_fsync.xfs_file_buffered_write.new_sync_write.vfs_write.ksys_write 70.73 -1.5 69.20 perf-profile.calltrace.cycles-pp.xfs_file_buffered_write.new_sync_write.vfs_write.ksys_write.do_syscall_64 70.75 -1.5 69.22 perf-profile.calltrace.cycles-pp.new_sync_write.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe 70.83 -1.5 69.32 perf-profile.calltrace.cycles-pp.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write 70.84 -1.5 69.33 perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write 70.93 -1.5 69.42 perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.write 70.92 -1.5 69.41 perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.write 71.09 -1.5 69.61 perf-profile.calltrace.cycles-pp.write 0.82 -0.1 0.71 perf-profile.calltrace.cycles-pp.xlog_ioend_work.process_one_work.worker_thread.kthread.ret_from_fork 1.01 -0.1 0.94 ± 2% perf-profile.calltrace.cycles-pp.xlog_cil_push_work.process_one_work.worker_thread.kthread.ret_from_fork 0.54 +0.0 0.57 ± 2% perf-profile.calltrace.cycles-pp.schedule_idle.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify 0.52 +0.0 0.55 ± 2% perf-profile.calltrace.cycles-pp.__schedule.schedule_idle.do_idle.cpu_startup_entry.start_secondary 0.62 +0.0 0.66 ± 2% perf-profile.calltrace.cycles-pp.xlog_state_do_callback.xlog_ioend_work.process_one_work.worker_thread.kthread 2.79 +0.1 2.87 ± 2% perf-profile.calltrace.cycles-pp.__flush_work.xlog_cil_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write 0.52 ± 2% +0.1 0.61 perf-profile.calltrace.cycles-pp.wait_for_completion.__flush_work.xlog_cil_force_lsn.xfs_log_force_lsn.xfs_file_fsync 0.56 +0.1 0.66 perf-profile.calltrace.cycles-pp.iomap_apply.iomap_file_buffered_write.xfs_file_buffered_write.new_sync_write.vfs_write 0.56 +0.1 0.66 perf-profile.calltrace.cycles-pp.iomap_file_buffered_write.xfs_file_buffered_write.new_sync_write.vfs_write.ksys_write 0.95 ± 4% +0.1 1.07 ± 5% perf-profile.calltrace.cycles-pp.brd_submit_bio.submit_bio_noacct.submit_bio.iomap_submit_ioend.xfs_vm_writepages 0.54 +0.1 0.66 perf-profile.calltrace.cycles-pp.complete.process_one_work.worker_thread.kthread.ret_from_fork 0.70 +0.2 0.85 perf-profile.calltrace.cycles-pp.md_submit_flush_data.process_one_work.worker_thread.kthread.ret_from_fork 1.16 ± 4% +0.2 1.32 ± 4% perf-profile.calltrace.cycles-pp.submit_bio.iomap_submit_ioend.xfs_vm_writepages.do_writepages.__filemap_fdatawrite_range 1.13 ± 4% +0.2 1.28 ± 4% perf-profile.calltrace.cycles-pp.submit_bio_noacct.submit_bio.iomap_submit_ioend.xfs_vm_writepages.do_writepages 1.17 ± 3% +0.2 1.33 ± 4% perf-profile.calltrace.cycles-pp.iomap_submit_ioend.xfs_vm_writepages.do_writepages.__filemap_fdatawrite_range.file_write_and_wait_range 4.28 +0.2 4.45 perf-profile.calltrace.cycles-pp.process_one_work.worker_thread.kthread.ret_from_fork 4.56 +0.2 4.75 perf-profile.calltrace.cycles-pp.worker_thread.kthread.ret_from_fork 4.56 +0.2 4.76 perf-profile.calltrace.cycles-pp.ret_from_fork 4.56 +0.2 4.76 perf-profile.calltrace.cycles-pp.kthread.ret_from_fork 1.96 ± 5% +0.3 2.22 ± 5% perf-profile.calltrace.cycles-pp.__filemap_fdatawrite_range.file_write_and_wait_range.xfs_file_fsync.xfs_file_buffered_write.new_sync_write 1.95 ± 5% +0.3 2.21 ± 5% perf-profile.calltrace.cycles-pp.do_writepages.__filemap_fdatawrite_range.file_write_and_wait_range.xfs_file_fsync.xfs_file_buffered_write 1.94 ± 5% +0.3 2.21 ± 5% perf-profile.calltrace.cycles-pp.xfs_vm_writepages.do_writepages.__filemap_fdatawrite_range.file_write_and_wait_range.xfs_file_fsync 2.09 ± 5% +0.3 2.36 ± 5% perf-profile.calltrace.cycles-pp.file_write_and_wait_range.xfs_file_fsync.xfs_file_buffered_write.new_sync_write.vfs_write 8.93 +0.4 9.33 perf-profile.calltrace.cycles-pp.remove_wait_queue.xlog_cil_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write 8.90 +0.4 9.30 perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.remove_wait_queue.xlog_cil_force_lsn.xfs_log_force_lsn.xfs_file_fsync 8.86 +0.4 9.27 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.remove_wait_queue.xlog_cil_force_lsn.xfs_log_force_lsn 0.17 ±141% +0.4 0.58 perf-profile.calltrace.cycles-pp.prepare_to_wait_event.md_flush_request.raid0_make_request.md_handle_request.md_submit_bio 0.00 +0.5 0.52 perf-profile.calltrace.cycles-pp.iomap_write_actor.iomap_apply.iomap_file_buffered_write.xfs_file_buffered_write.new_sync_write 0.00 +0.6 0.58 perf-profile.calltrace.cycles-pp.try_to_wake_up.swake_up_locked.complete.process_one_work.worker_thread 0.00 +0.6 0.61 perf-profile.calltrace.cycles-pp.swake_up_locked.complete.process_one_work.worker_thread.kthread 9.75 +0.8 10.59 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irqsave.remove_wait_queue.xlog_wait_on_iclog.__xfs_log_force_lsn 21.50 +1.1 22.59 perf-profile.calltrace.cycles-pp.intel_idle.cpuidle_enter_state.cpuidle_enter.do_idle.cpu_startup_entry 22.08 +1.1 23.19 perf-profile.calltrace.cycles-pp.cpuidle_enter.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify 22.05 +1.1 23.16 perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.do_idle.cpu_startup_entry.start_secondary 23.93 +1.2 25.15 perf-profile.calltrace.cycles-pp.secondary_startup_64_no_verify 23.63 +1.3 24.88 perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify 23.64 +1.3 24.90 perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.secondary_startup_64_no_verify 23.64 +1.3 24.90 perf-profile.calltrace.cycles-pp.start_secondary.secondary_startup_64_no_verify 21.00 +1.9 22.92 perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock_irq.md_flush_request.raid0_make_request.md_handle_request 21.09 +1.9 23.03 perf-profile.calltrace.cycles-pp._raw_spin_lock_irq.md_flush_request.raid0_make_request.md_handle_request.md_submit_bio 13.72 +2.0 15.74 ± 2% perf-profile.calltrace.cycles-pp.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write.new_sync_write 22.13 +2.1 24.24 perf-profile.calltrace.cycles-pp.md_flush_request.raid0_make_request.md_handle_request.md_submit_bio.submit_bio_noacct 22.16 +2.1 24.27 perf-profile.calltrace.cycles-pp.raid0_make_request.md_handle_request.md_submit_bio.submit_bio_noacct.submit_bio 22.22 +2.1 24.34 perf-profile.calltrace.cycles-pp.md_handle_request.md_submit_bio.submit_bio_noacct.submit_bio.submit_bio_wait 22.28 +2.1 24.41 perf-profile.calltrace.cycles-pp.md_submit_bio.submit_bio_noacct.submit_bio.submit_bio_wait.blkdev_issue_flush 22.34 +2.1 24.48 perf-profile.calltrace.cycles-pp.submit_bio.submit_bio_wait.blkdev_issue_flush.xfs_file_fsync.xfs_file_buffered_write 22.37 +2.1 24.51 perf-profile.calltrace.cycles-pp.submit_bio_wait.blkdev_issue_flush.xfs_file_fsync.xfs_file_buffered_write.new_sync_write 22.38 +2.1 24.53 perf-profile.calltrace.cycles-pp.blkdev_issue_flush.xfs_file_fsync.xfs_file_buffered_write.new_sync_write.vfs_write 22.33 +2.1 24.48 perf-profile.calltrace.cycles-pp.submit_bio_noacct.submit_bio.submit_bio_wait.blkdev_issue_flush.xfs_file_fsync 6.70 ± 2% +3.7 10.35 ± 2% perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.xlog_cil_force_lsn.xfs_log_force_lsn.xfs_file_fsync 6.80 ± 2% +3.7 10.46 ± 2% perf-profile.calltrace.cycles-pp._raw_spin_lock.xlog_cil_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write 18.82 +4.2 23.01 perf-profile.calltrace.cycles-pp.xlog_cil_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write.new_sync_write 6.05 +4.6 10.62 perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.remove_wait_queue.xlog_wait_on_iclog.__xfs_log_force_lsn.xfs_log_force_lsn 6.06 +4.6 10.65 perf-profile.calltrace.cycles-pp.remove_wait_queue.xlog_wait_on_iclog.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync 6.25 +4.7 10.97 perf-profile.calltrace.cycles-pp.xlog_wait_on_iclog.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write 32.55 +6.2 38.76 perf-profile.calltrace.cycles-pp.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write.new_sync_write.vfs_write 24.73 -8.2 16.50 ± 2% perf-profile.children.cycles-pp.__xfs_log_force_lsn 20.13 -3.4 16.69 perf-profile.children.cycles-pp._raw_spin_lock 66.06 -2.2 63.83 perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath 68.11 -1.6 66.52 perf-profile.children.cycles-pp.xfs_file_fsync 70.73 -1.5 69.20 perf-profile.children.cycles-pp.xfs_file_buffered_write 70.76 -1.5 69.23 perf-profile.children.cycles-pp.new_sync_write 70.84 -1.5 69.32 perf-profile.children.cycles-pp.vfs_write 70.85 -1.5 69.33 perf-profile.children.cycles-pp.ksys_write 71.12 -1.5 69.62 perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe 71.10 -1.5 69.61 perf-profile.children.cycles-pp.do_syscall_64 71.12 -1.5 69.64 perf-profile.children.cycles-pp.write 21.61 -0.9 20.73 perf-profile.children.cycles-pp.remove_wait_queue 23.44 -0.7 22.73 perf-profile.children.cycles-pp._raw_spin_lock_irqsave 0.62 -0.2 0.43 ± 2% perf-profile.children.cycles-pp.xlog_write 0.20 ± 6% -0.1 0.05 perf-profile.children.cycles-pp.xlog_state_done_syncing 0.82 -0.1 0.71 perf-profile.children.cycles-pp.xlog_ioend_work 0.39 -0.1 0.29 ± 2% perf-profile.children.cycles-pp.xlog_state_release_iclog 1.01 -0.1 0.94 ± 2% perf-profile.children.cycles-pp.xlog_cil_push_work 0.23 ± 11% -0.1 0.16 ± 24% perf-profile.children.cycles-pp.xlog_grant_add_space 0.24 ± 17% -0.1 0.17 ± 19% perf-profile.children.cycles-pp.xfs_log_ticket_ungrant 0.13 -0.1 0.07 perf-profile.children.cycles-pp.xlog_state_get_iclog_space 0.68 -0.0 0.64 ± 2% perf-profile.children.cycles-pp.asm_sysvec_apic_timer_interrupt 0.43 -0.0 0.40 ± 3% perf-profile.children.cycles-pp.__sysvec_apic_timer_interrupt 0.60 -0.0 0.57 ± 2% perf-profile.children.cycles-pp.sysvec_apic_timer_interrupt 0.42 -0.0 0.39 ± 2% perf-profile.children.cycles-pp.hrtimer_interrupt 0.05 +0.0 0.06 perf-profile.children.cycles-pp.__radix_tree_lookup 0.12 +0.0 0.13 perf-profile.children.cycles-pp.__update_load_avg_cfs_rq 0.08 +0.0 0.09 perf-profile.children.cycles-pp.iomap_set_page_dirty 0.08 +0.0 0.09 perf-profile.children.cycles-pp.__list_add_valid 0.06 +0.0 0.07 perf-profile.children.cycles-pp.ttwu_do_wakeup 0.09 +0.0 0.10 ± 4% perf-profile.children.cycles-pp.iomap_set_range_uptodate 0.07 ± 6% +0.0 0.09 ± 5% perf-profile.children.cycles-pp.copyin 0.21 ± 2% +0.0 0.22 perf-profile.children.cycles-pp.ttwu_queue_wakelist 0.09 ± 5% +0.0 0.11 ± 4% perf-profile.children.cycles-pp.xfs_btree_lookup 0.08 +0.0 0.09 ± 5% perf-profile.children.cycles-pp.iov_iter_copy_from_user_atomic 0.09 ± 5% +0.0 0.10 perf-profile.children.cycles-pp.llseek 0.10 ± 4% +0.0 0.11 ± 4% perf-profile.children.cycles-pp.queue_work_on 0.16 ± 5% +0.0 0.18 ± 2% perf-profile.children.cycles-pp.update_rq_clock 0.13 ± 3% +0.0 0.14 ± 3% perf-profile.children.cycles-pp.syscall_exit_to_user_mode 0.09 ± 5% +0.0 0.10 ± 4% perf-profile.children.cycles-pp.__queue_work 0.05 ± 8% +0.0 0.07 ± 11% perf-profile.children.cycles-pp.xfs_map_blocks 0.12 ± 3% +0.0 0.14 ± 5% perf-profile.children.cycles-pp.exit_to_user_mode_prepare 0.11 ± 4% +0.0 0.13 perf-profile.children.cycles-pp.pagecache_get_page 0.11 +0.0 0.13 ± 3% perf-profile.children.cycles-pp.set_task_cpu 0.19 ± 4% +0.0 0.21 perf-profile.children.cycles-pp.__list_del_entry_valid 0.07 ± 6% +0.0 0.09 ± 5% perf-profile.children.cycles-pp.insert_work 0.08 ± 6% +0.0 0.10 ± 4% perf-profile.children.cycles-pp.migrate_task_rq_fair 0.17 ± 2% +0.0 0.19 ± 2% perf-profile.children.cycles-pp.iomap_write_begin 0.16 ± 3% +0.0 0.18 ± 2% perf-profile.children.cycles-pp.perf_trace_sched_wakeup_template 0.12 +0.0 0.14 perf-profile.children.cycles-pp.grab_cache_page_write_begin 0.28 +0.0 0.30 ± 2% perf-profile.children.cycles-pp.pick_next_task_fair 0.11 ± 4% +0.0 0.13 perf-profile.children.cycles-pp.xfs_trans_committed_bulk 0.09 ± 5% +0.0 0.11 perf-profile.children.cycles-pp.xfs_buffered_write_iomap_begin 0.16 +0.0 0.19 ± 6% perf-profile.children.cycles-pp.update_cfs_group 0.14 ± 3% +0.0 0.17 ± 4% perf-profile.children.cycles-pp.__switch_to_asm 0.17 ± 4% +0.0 0.20 ± 7% perf-profile.children.cycles-pp.xfs_inode_item_format 0.15 +0.0 0.18 ± 4% perf-profile.children.cycles-pp.iomap_write_end 0.30 +0.0 0.33 perf-profile.children.cycles-pp.select_idle_cpu 0.19 ± 4% +0.0 0.22 ± 2% perf-profile.children.cycles-pp.xfs_bmap_add_extent_unwritten_real 0.17 ± 4% +0.0 0.20 ± 2% perf-profile.children.cycles-pp.xlog_cil_process_committed 0.17 ± 4% +0.0 0.20 ± 2% perf-profile.children.cycles-pp.xlog_cil_committed 1.25 +0.0 1.29 perf-profile.children.cycles-pp.__wake_up_common_lock 0.26 +0.0 0.30 perf-profile.children.cycles-pp.available_idle_cpu 0.19 ± 4% +0.0 0.23 ± 3% perf-profile.children.cycles-pp.xfs_bmapi_convert_unwritten 0.48 +0.0 0.51 ± 2% perf-profile.children.cycles-pp.dequeue_entity 0.02 ±141% +0.0 0.06 ± 8% perf-profile.children.cycles-pp.xfs_iext_lookup_extent 0.15 ± 6% +0.0 0.19 ± 4% perf-profile.children.cycles-pp.poll_idle 0.11 ± 4% +0.0 0.15 ± 12% perf-profile.children.cycles-pp.submit_bio_checks 0.62 +0.0 0.66 ± 2% perf-profile.children.cycles-pp.xlog_state_do_callback 0.54 +0.0 0.58 ± 2% perf-profile.children.cycles-pp.schedule_idle 0.23 ± 3% +0.0 0.27 perf-profile.children.cycles-pp.xfs_bmapi_write 0.38 +0.0 0.42 perf-profile.children.cycles-pp.xlog_state_clean_iclog 0.37 ± 3% +0.0 0.42 perf-profile.children.cycles-pp.sched_ttwu_pending 0.40 +0.0 0.45 perf-profile.children.cycles-pp.select_idle_sibling 0.59 +0.0 0.64 perf-profile.children.cycles-pp.dequeue_task_fair 0.53 +0.0 0.58 ± 2% perf-profile.children.cycles-pp.select_task_rq_fair 0.00 +0.1 0.05 perf-profile.children.cycles-pp.__pagevec_release 0.56 +0.1 0.61 ± 3% perf-profile.children.cycles-pp.update_load_avg 0.00 +0.1 0.05 ± 8% perf-profile.children.cycles-pp.kfree 0.00 +0.1 0.05 ± 8% perf-profile.children.cycles-pp.submit_flushes 0.00 +0.1 0.05 ± 8% perf-profile.children.cycles-pp.bio_alloc_bioset 0.41 +0.1 0.46 perf-profile.children.cycles-pp.brd_do_bvec 1.19 +0.1 1.24 perf-profile.children.cycles-pp.__wake_up_common 0.47 ± 2% +0.1 0.52 perf-profile.children.cycles-pp.enqueue_entity 0.59 +0.1 0.64 perf-profile.children.cycles-pp.enqueue_task_fair 0.61 ± 2% +0.1 0.68 perf-profile.children.cycles-pp.ttwu_do_activate 0.00 +0.1 0.07 ± 35% perf-profile.children.cycles-pp.blk_throtl_bio 0.44 +0.1 0.51 perf-profile.children.cycles-pp.flush_smp_call_function_from_idle 0.45 +0.1 0.52 perf-profile.children.cycles-pp.iomap_write_actor 0.44 +0.1 0.51 perf-profile.children.cycles-pp.schedule_timeout 2.79 +0.1 2.87 ± 2% perf-profile.children.cycles-pp.__flush_work 0.52 ± 2% +0.1 0.62 perf-profile.children.cycles-pp.wait_for_completion 0.39 ± 2% +0.1 0.49 perf-profile.children.cycles-pp.autoremove_wake_function 0.49 ± 3% +0.1 0.58 perf-profile.children.cycles-pp.prepare_to_wait_event 0.56 +0.1 0.66 perf-profile.children.cycles-pp.iomap_apply 0.56 +0.1 0.66 perf-profile.children.cycles-pp.iomap_file_buffered_write 1.32 +0.1 1.43 perf-profile.children.cycles-pp.schedule 0.49 +0.1 0.61 perf-profile.children.cycles-pp.swake_up_locked 0.54 +0.1 0.67 perf-profile.children.cycles-pp.complete 1.00 ± 4% +0.1 1.13 ± 5% perf-profile.children.cycles-pp.brd_submit_bio 0.71 +0.1 0.85 perf-profile.children.cycles-pp.md_submit_flush_data 1.83 +0.2 1.98 perf-profile.children.cycles-pp.__schedule 4.29 +0.2 4.45 perf-profile.children.cycles-pp.process_one_work 1.17 ± 3% +0.2 1.33 ± 4% perf-profile.children.cycles-pp.iomap_submit_ioend 1.80 +0.2 1.99 perf-profile.children.cycles-pp.try_to_wake_up 4.56 +0.2 4.76 perf-profile.children.cycles-pp.ret_from_fork 4.56 +0.2 4.76 perf-profile.children.cycles-pp.kthread 4.56 +0.2 4.76 perf-profile.children.cycles-pp.worker_thread 1.96 ± 5% +0.3 2.22 ± 5% perf-profile.children.cycles-pp.__filemap_fdatawrite_range 1.95 ± 5% +0.3 2.21 ± 5% perf-profile.children.cycles-pp.do_writepages 1.95 ± 5% +0.3 2.21 ± 5% perf-profile.children.cycles-pp.xfs_vm_writepages 2.09 ± 5% +0.3 2.36 ± 5% perf-profile.children.cycles-pp.file_write_and_wait_range 21.76 +1.1 22.82 perf-profile.children.cycles-pp.intel_idle 22.34 +1.1 23.43 perf-profile.children.cycles-pp.cpuidle_enter 22.34 +1.1 23.43 perf-profile.children.cycles-pp.cpuidle_enter_state 23.93 +1.2 25.15 perf-profile.children.cycles-pp.secondary_startup_64_no_verify 23.93 +1.2 25.15 perf-profile.children.cycles-pp.cpu_startup_entry 23.92 +1.2 25.15 perf-profile.children.cycles-pp.do_idle 23.64 +1.3 24.90 perf-profile.children.cycles-pp.start_secondary 10.12 +1.3 11.45 perf-profile.children.cycles-pp.xlog_wait_on_iclog 23.85 +2.0 25.87 perf-profile.children.cycles-pp._raw_spin_lock_irq 22.33 +2.1 24.47 perf-profile.children.cycles-pp.md_flush_request 22.37 +2.1 24.51 perf-profile.children.cycles-pp.submit_bio_wait 22.38 +2.1 24.53 perf-profile.children.cycles-pp.blkdev_issue_flush 22.43 +2.2 24.59 perf-profile.children.cycles-pp.raid0_make_request 22.53 +2.2 24.70 perf-profile.children.cycles-pp.md_handle_request 22.62 +2.2 24.80 perf-profile.children.cycles-pp.md_submit_bio 23.73 +2.3 26.07 perf-profile.children.cycles-pp.submit_bio 23.75 +2.3 26.09 perf-profile.children.cycles-pp.submit_bio_noacct 18.82 +4.2 23.01 perf-profile.children.cycles-pp.xlog_cil_force_lsn 32.56 +6.2 38.77 perf-profile.children.cycles-pp.xfs_log_force_lsn 65.90 -2.2 63.69 perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath 0.22 ± 9% -0.1 0.16 ± 24% perf-profile.self.cycles-pp.xlog_grant_add_space 0.21 ± 13% -0.1 0.15 ± 16% perf-profile.self.cycles-pp.xfs_log_ticket_ungrant 0.08 +0.0 0.09 perf-profile.self.cycles-pp.__list_add_valid 0.06 +0.0 0.07 perf-profile.self.cycles-pp.write 0.20 ± 2% +0.0 0.21 perf-profile.self.cycles-pp.menu_select 0.11 ± 4% +0.0 0.12 perf-profile.self.cycles-pp.xfs_log_commit_cil 0.09 +0.0 0.10 ± 4% perf-profile.self.cycles-pp.iomap_set_range_uptodate 0.07 ± 6% +0.0 0.09 ± 5% perf-profile.self.cycles-pp.insert_work 0.07 +0.0 0.08 ± 5% perf-profile.self.cycles-pp.memcpy_erms 0.07 ± 7% +0.0 0.08 perf-profile.self.cycles-pp.perf_trace_sched_wakeup_template 0.06 +0.0 0.07 ± 6% perf-profile.self.cycles-pp.flush_smp_call_function_from_idle 0.11 ± 4% +0.0 0.12 ± 3% perf-profile.self.cycles-pp.try_to_wake_up 0.09 ± 5% +0.0 0.10 ± 4% perf-profile.self.cycles-pp.__switch_to 0.08 ± 12% +0.0 0.09 ± 10% perf-profile.self.cycles-pp.xfs_inode_item_format 0.07 ± 6% +0.0 0.09 perf-profile.self.cycles-pp.xlog_cil_force_lsn 0.06 ± 8% +0.0 0.07 ± 6% perf-profile.self.cycles-pp.prepare_to_wait_event 0.16 +0.0 0.18 ± 5% perf-profile.self.cycles-pp.update_cfs_group 0.18 ± 2% +0.0 0.21 perf-profile.self.cycles-pp.__list_del_entry_valid 0.14 ± 3% +0.0 0.17 ± 4% perf-profile.self.cycles-pp.__switch_to_asm 0.36 ± 2% +0.0 0.40 perf-profile.self.cycles-pp.__schedule 0.33 +0.0 0.37 perf-profile.self.cycles-pp.brd_do_bvec 0.30 +0.0 0.33 ± 5% perf-profile.self.cycles-pp.update_load_avg 0.26 ± 3% +0.0 0.30 perf-profile.self.cycles-pp.available_idle_cpu 0.14 ± 6% +0.0 0.18 ± 5% perf-profile.self.cycles-pp.poll_idle 0.57 ± 3% +0.0 0.61 ± 2% perf-profile.self.cycles-pp._raw_spin_lock_irqsave 0.29 +0.0 0.34 ± 3% perf-profile.self.cycles-pp._raw_spin_lock_irq 0.00 +0.1 0.05 perf-profile.self.cycles-pp.kmem_cache_alloc 0.00 +0.1 0.05 ± 8% perf-profile.self.cycles-pp.kfree 0.00 +0.1 0.06 ± 8% perf-profile.self.cycles-pp.migrate_task_rq_fair 0.00 +0.1 0.06 ± 16% perf-profile.self.cycles-pp.percpu_counter_add_batch 21.76 +1.1 22.82 perf-profile.self.cycles-pp.intel_idle 0.04 ± 3% -13.2% 0.04 ± 4% perf-sched.sch_delay.avg.ms.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write 0.02 ± 17% -41.4% 0.01 ± 14% perf-sched.sch_delay.avg.ms.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64 0.24 ± 27% -62.7% 0.09 ± 64% perf-sched.sch_delay.avg.ms.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown] 0.11 ± 12% -59.5% 0.04 ± 16% perf-sched.sch_delay.avg.ms.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.09 -40.1% 0.05 ± 3% perf-sched.sch_delay.avg.ms.io_schedule.wait_on_page_bit.wait_on_page_writeback.__filemap_fdatawait_range 0.08 ± 2% -19.0% 0.07 perf-sched.sch_delay.avg.ms.md_flush_request.raid0_make_request.md_handle_request.md_submit_bio 0.01 ± 35% -100.0% 0.00 perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.__alloc_pages.pagecache_get_page.grab_cache_page_write_begin 0.01 ± 5% -23.8% 0.01 ± 4% perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.__flush_work.xlog_cil_force_lsn.xfs_log_force_lsn 0.09 ± 26% -69.2% 0.03 ±100% perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.__kmalloc.kmem_alloc.kmem_alloc_large 0.13 ± 68% -74.1% 0.03 ±102% perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.do_user_addr_fault.exc_page_fault.asm_exc_page_fault 0.07 ± 22% -41.1% 0.04 ± 36% perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.down_read.xfs_file_fsync.xfs_file_buffered_write 0.04 ± 15% -41.0% 0.03 ± 15% perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.down_write.xfs_ilock.xfs_buffered_write_iomap_begin 0.11 ± 58% -76.4% 0.03 ±107% perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.down_write.xfs_ilock.xfs_file_buffered_write 0.04 ± 28% -63.3% 0.01 ± 17% perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.down_write.xfs_ilock.xfs_vn_update_time 0.00 ± 28% -100.0% 0.00 perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff 0.26 -40.4% 0.15 perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.process_one_work.worker_thread.kthread 0.08 ± 26% -30.9% 0.06 ± 22% perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate 0.14 ± 16% -83.5% 0.02 ±101% perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.submit_bio_checks.submit_bio_noacct.raid0_make_request 0.03 ± 5% -19.0% 0.03 perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.submit_bio_checks.submit_bio_noacct.submit_bio 0.11 ± 3% -63.9% 0.04 ± 4% perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.wait_for_completion.__flush_work.xlog_cil_force_lsn 0.07 ± 6% -64.7% 0.02 ± 39% perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.wait_for_completion.stop_two_cpus.migrate_swap 0.37 ± 2% -19.9% 0.29 ± 2% perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.wait_for_completion_io_timeout.submit_bio_wait.blkdev_issue_flush 0.09 ± 29% -58.2% 0.04 ± 36% perf-sched.sch_delay.avg.ms.preempt_schedule_common.__cond_resched.write_cache_pages.iomap_writepages.xfs_vm_writepages 0.03 ± 21% -41.4% 0.02 ± 17% perf-sched.sch_delay.avg.ms.rcu_gp_kthread.kthread.ret_from_fork 0.03 ± 28% +80.9% 0.05 ± 17% perf-sched.sch_delay.avg.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_free_eofblocks 0.19 ± 20% -84.8% 0.03 ±110% perf-sched.sch_delay.avg.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_inactive_ifree 0.13 ± 13% -46.4% 0.07 ± 44% perf-sched.sch_delay.avg.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_remove 0.07 ± 13% -61.3% 0.03 ± 37% perf-sched.sch_delay.avg.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_trans_roll 0.19 ± 3% -36.5% 0.12 ± 7% perf-sched.sch_delay.avg.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_vn_update_time 0.12 ± 15% -74.7% 0.03 ±117% perf-sched.sch_delay.avg.ms.rwsem_down_write_slowpath.do_unlinkat.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.05 ± 11% -46.9% 0.03 ± 4% perf-sched.sch_delay.avg.ms.rwsem_down_write_slowpath.xlog_cil_push_work.process_one_work.worker_thread 0.00 ± 10% +28.6% 0.01 perf-sched.sch_delay.avg.ms.schedule_hrtimeout_range_clock.poll_schedule_timeout.constprop.0.do_select 0.05 ± 5% -52.9% 0.02 ± 4% perf-sched.sch_delay.avg.ms.schedule_timeout.__down.down.xfs_buf_lock 0.04 ± 57% -100.0% 0.00 perf-sched.sch_delay.avg.ms.schedule_timeout.__down.down.xlog_write_iclog 0.03 ± 31% -57.3% 0.01 ± 36% perf-sched.sch_delay.avg.ms.schedule_timeout.kcompactd.kthread.ret_from_fork 0.03 ± 8% -25.5% 0.02 ± 10% perf-sched.sch_delay.avg.ms.schedule_timeout.rcu_gp_kthread.kthread.ret_from_fork 0.06 -25.7% 0.04 perf-sched.sch_delay.avg.ms.schedule_timeout.wait_for_completion.__flush_work.xlog_cil_force_lsn 0.03 ± 7% -15.6% 0.02 ± 5% perf-sched.sch_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork 0.02 -16.7% 0.01 ± 3% perf-sched.sch_delay.avg.ms.worker_thread.kthread.ret_from_fork 0.02 ± 4% -27.5% 0.01 ± 3% perf-sched.sch_delay.avg.ms.xlog_cil_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write 0.00 -100.0% 0.00 perf-sched.sch_delay.avg.ms.xlog_state_get_iclog_space.xlog_write.xlog_cil_push_work.process_one_work 0.11 ± 2% -60.7% 0.04 ± 2% perf-sched.sch_delay.avg.ms.xlog_wait_on_iclog.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write 0.06 -11.5% 0.05 perf-sched.sch_delay.avg.ms.xlog_wait_on_iclog.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync 5.52 ± 28% -51.9% 2.66 ± 47% perf-sched.sch_delay.max.ms.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write 2.27 ± 6% -47.7% 1.19 ± 56% perf-sched.sch_delay.max.ms.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_reschedule_ipi.[unknown] 2.36 ± 6% -42.5% 1.36 ± 9% perf-sched.sch_delay.max.ms.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe 0.01 ± 56% -100.0% 0.00 perf-sched.sch_delay.max.ms.preempt_schedule_common.__cond_resched.__alloc_pages.pagecache_get_page.grab_cache_page_write_begin 0.22 ± 67% -88.5% 0.03 ± 12% perf-sched.sch_delay.max.ms.preempt_schedule_common.__cond_resched.__filemap_fdatawait_range.file_write_and_wait_range.xfs_file_fsync 2.00 ± 14% -35.0% 1.30 ± 30% perf-sched.sch_delay.max.ms.preempt_schedule_common.__cond_resched.__flush_work.xlog_cil_force_lsn.xfs_log_force_lsn 2.24 ± 20% -75.6% 0.55 ± 73% perf-sched.sch_delay.max.ms.preempt_schedule_common.__cond_resched.__kmalloc.kmem_alloc.kmem_alloc_large 0.74 ± 75% -85.2% 0.11 ± 82% perf-sched.sch_delay.max.ms.preempt_schedule_common.__cond_resched.do_user_addr_fault.exc_page_fault.asm_exc_page_fault 2.18 ± 8% -43.3% 1.24 ± 35% perf-sched.sch_delay.max.ms.preempt_schedule_common.__cond_resched.down_read.xfs_file_fsync.xfs_file_buffered_write 1.72 ± 39% -71.1% 0.50 ± 72% perf-sched.sch_delay.max.ms.preempt_schedule_common.__cond_resched.down_read.xfs_log_commit_cil.__xfs_trans_commit 2.17 ± 3% -42.6% 1.25 ± 9% perf-sched.sch_delay.max.ms.preempt_schedule_common.__cond_resched.down_write.xfs_ilock.xfs_buffered_write_iomap_begin 1.60 ± 62% -86.3% 0.22 ±106% perf-sched.sch_delay.max.ms.preempt_schedule_common.__cond_resched.down_write.xfs_ilock.xfs_file_buffered_write 1.38 ± 37% -47.5% 0.72 ± 14% perf-sched.sch_delay.max.ms.preempt_schedule_common.__cond_resched.down_write.xfs_ilock.xfs_vn_update_time 0.00 ± 28% -100.0% 0.00 perf-sched.sch_delay.max.ms.preempt_schedule_common.__cond_resched.down_write_killable.vm_mmap_pgoff.ksys_mmap_pgoff 1.97 ± 10% -30.2% 1.38 ± 5% perf-sched.sch_delay.max.ms.preempt_schedule_common.__cond_resched.mempool_alloc.md_submit_bio.submit_bio_noacct 1.85 ± 7% -31.2% 1.27 ± 2% perf-sched.sch_delay.max.ms.preempt_schedule_common.__cond_resched.process_one_work.worker_thread.kthread 1.59 ± 25% -89.9% 0.16 ± 93% perf-sched.sch_delay.max.ms.preempt_schedule_common.__cond_resched.submit_bio_checks.submit_bio_noacct.raid0_make_request 2.56 ± 9% -28.7% 1.82 ± 8% perf-sched.sch_delay.max.ms.preempt_schedule_common.__cond_resched.wait_for_completion.__flush_work.xlog_cil_force_lsn 2.31 ± 11% -62.5% 0.87 ± 15% perf-sched.sch_delay.max.ms.preempt_schedule_common.__cond_resched.write_cache_pages.iomap_writepages.xfs_vm_writepages 1.06 ± 20% -59.5% 0.43 ± 48% perf-sched.sch_delay.max.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_create 0.94 ± 40% -77.1% 0.22 ±129% perf-sched.sch_delay.max.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_inactive_ifree 1.61 ± 62% -51.7% 0.78 ± 15% perf-sched.sch_delay.max.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_trans_roll 0.62 ± 24% -75.1% 0.15 ±133% perf-sched.sch_delay.max.ms.rwsem_down_write_slowpath.do_unlinkat.do_syscall_64.entry_SYSCALL_64_after_hwframe 1.11 ± 9% -61.9% 0.42 ± 8% perf-sched.sch_delay.max.ms.rwsem_down_write_slowpath.xlog_cil_push_work.process_one_work.worker_thread 0.01 ± 8% +135.3% 0.01 ± 7% perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.poll_schedule_timeout.constprop.0.do_select 0.18 ± 28% -89.7% 0.02 ± 13% perf-sched.sch_delay.max.ms.schedule_hrtimeout_range_clock.poll_schedule_timeout.constprop.0.do_sys_poll 2.54 ± 42% -46.1% 1.37 ± 26% perf-sched.sch_delay.max.ms.schedule_timeout.__down.down.xfs_buf_lock 0.33 ± 78% -100.0% 0.00 perf-sched.sch_delay.max.ms.schedule_timeout.__down.down.xlog_write_iclog 0.70 ± 51% -78.2% 0.15 ± 65% perf-sched.sch_delay.max.ms.schedule_timeout.kcompactd.kthread.ret_from_fork 0.01 ± 14% -100.0% 0.00 perf-sched.sch_delay.max.ms.xlog_state_get_iclog_space.xlog_write.xlog_cil_push_work.process_one_work 4.73 ± 12% -46.7% 2.52 ± 61% perf-sched.sch_delay.max.ms.xlog_wait_on_iclog.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write 5.49 ± 20% +24.0% 6.80 ± 3% perf-sched.sch_delay.max.ms.xlog_wait_on_iclog.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync 0.06 -23.4% 0.05 perf-sched.total_sch_delay.average.ms 1.45 -11.8% 1.28 perf-sched.total_wait_and_delay.average.ms 4218600 -22.8% 3256636 ± 9% perf-sched.total_wait_and_delay.count.ms 8577 ± 4% -20.4% 6826 ± 10% perf-sched.total_wait_and_delay.max.ms 1.39 -11.3% 1.23 perf-sched.total_wait_time.average.ms 8577 ± 4% -20.4% 6826 ± 10% perf-sched.total_wait_time.max.ms 0.78 +36.9% 1.07 ± 2% perf-sched.wait_and_delay.avg.ms.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write 0.49 ±141% +1.3e+05% 618.51 ± 16% perf-sched.wait_and_delay.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64 788.02 ± 5% -32.2% 533.91 ± 4% perf-sched.wait_and_delay.avg.ms.do_nanosleep.hrtimer_nanosleep.__x64_sys_nanosleep.do_syscall_64 0.50 ±141% +1.2e+05% 618.53 ± 16% perf-sched.wait_and_delay.avg.ms.do_syslog.part.0.kmsg_read.vfs_read 272.43 -61.4% 105.13 ± 6% perf-sched.wait_and_delay.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64 0.34 -44.4% 0.19 ± 2% perf-sched.wait_and_delay.avg.ms.io_schedule.wait_on_page_bit.wait_on_page_writeback.__filemap_fdatawait_range 0.83 +13.0% 0.94 perf-sched.wait_and_delay.avg.ms.md_flush_request.raid0_make_request.md_handle_request.md_submit_bio 127.64 ± 22% +59.7% 203.80 ± 15% perf-sched.wait_and_delay.avg.ms.pipe_read.new_sync_read.vfs_read.ksys_read 20.10 ±101% -87.9% 2.43 ± 95% perf-sched.wait_and_delay.avg.ms.preempt_schedule_common.__cond_resched.__alloc_pages.alloc_pages_vma.shmem_alloc_page 4.68 ± 14% -79.3% 0.97 ±141% perf-sched.wait_and_delay.avg.ms.preempt_schedule_common.__cond_resched.pagecache_get_page.shmem_getpage_gfp.shmem_write_begin 5.34 ± 29% -42.3% 3.08 ± 10% perf-sched.wait_and_delay.avg.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_iomap_write_unwritten 3.68 -47.5% 1.93 ± 2% perf-sched.wait_and_delay.avg.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_trans_roll 213.02 ± 3% +33.6% 284.56 ± 21% perf-sched.wait_and_delay.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait 7.00 ± 2% -12.6% 6.12 ± 7% perf-sched.wait_and_delay.avg.ms.schedule_timeout.rcu_gp_kthread.kthread.ret_from_fork 0.54 -42.8% 0.31 perf-sched.wait_and_delay.avg.ms.schedule_timeout.wait_for_completion.__flush_work.xlog_cil_force_lsn 487.35 ± 3% -17.6% 401.38 ± 4% perf-sched.wait_and_delay.avg.ms.smpboot_thread_fn.kthread.ret_from_fork 4.28 ± 2% -23.3% 3.28 perf-sched.wait_and_delay.avg.ms.worker_thread.kthread.ret_from_fork 0.92 -12.7% 0.80 perf-sched.wait_and_delay.avg.ms.xlog_cil_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write 0.97 -71.8% 0.27 perf-sched.wait_and_delay.avg.ms.xlog_wait_on_iclog.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write 20.00 -26.7% 14.67 ± 12% perf-sched.wait_and_delay.count.__x64_sys_pause.do_syscall_64.entry_SYSCALL_64_after_hwframe.[unknown] 245148 -86.0% 34284 ± 10% perf-sched.wait_and_delay.count.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write 0.67 ±141% +1600.0% 11.33 ± 8% perf-sched.wait_and_delay.count.devkmsg_read.vfs_read.ksys_read.do_syscall_64 0.67 ±141% +1600.0% 11.33 ± 8% perf-sched.wait_and_delay.count.do_syslog.part.0.kmsg_read.vfs_read 247.00 +119.8% 543.00 ± 3% perf-sched.wait_and_delay.count.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64 168.00 ± 70% +178.4% 467.67 ± 4% perf-sched.wait_and_delay.count.do_wait.kernel_wait4.__do_sys_wait4.do_syscall_64 251.67 ± 12% -73.8% 66.00 ±141% perf-sched.wait_and_delay.count.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe 201465 -21.1% 158976 ± 9% perf-sched.wait_and_delay.count.io_schedule.wait_on_page_bit.wait_on_page_writeback.__filemap_fdatawait_range 808550 -16.0% 679238 ± 9% perf-sched.wait_and_delay.count.md_flush_request.raid0_make_request.md_handle_request.md_submit_bio 1320 ± 21% -56.4% 576.33 ± 20% perf-sched.wait_and_delay.count.pipe_read.new_sync_read.vfs_read.ksys_read 27.67 ± 14% -43.4% 15.67 ± 23% perf-sched.wait_and_delay.count.preempt_schedule_common.__cond_resched.generic_perform_write.__generic_file_write_iter.generic_file_write_iter 8.33 ± 11% -88.0% 1.00 ±141% perf-sched.wait_and_delay.count.preempt_schedule_common.__cond_resched.pagecache_get_page.shmem_getpage_gfp.shmem_write_begin 7471 -53.7% 3455 ± 10% perf-sched.wait_and_delay.count.preempt_schedule_common.__cond_resched.process_one_work.worker_thread.kthread 969.67 -20.1% 774.33 ± 12% perf-sched.wait_and_delay.count.preempt_schedule_common.__cond_resched.wait_for_completion.affine_move_task.__set_cpus_allowed_ptr 246.33 ± 7% -21.7% 193.00 ± 15% perf-sched.wait_and_delay.count.rcu_gp_kthread.kthread.ret_from_fork 118.33 ± 8% +18.9% 140.67 ± 4% perf-sched.wait_and_delay.count.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_trans_roll 1035 -37.9% 643.67 ± 10% perf-sched.wait_and_delay.count.schedule_hrtimeout_range_clock.poll_schedule_timeout.constprop.0.do_sys_poll 39.33 ± 2% -27.1% 28.67 ± 11% perf-sched.wait_and_delay.count.schedule_timeout.kcompactd.kthread.ret_from_fork 1217352 -15.4% 1030088 ± 9% perf-sched.wait_and_delay.count.schedule_timeout.wait_for_completion.__flush_work.xlog_cil_force_lsn 206.67 -25.2% 154.67 ± 9% perf-sched.wait_and_delay.count.schedule_timeout.xfsaild.kthread.ret_from_fork 1994 ± 3% -13.6% 1722 ± 13% perf-sched.wait_and_delay.count.smpboot_thread_fn.kthread.ret_from_fork 470869 -13.6% 406938 ± 9% perf-sched.wait_and_delay.count.worker_thread.kthread.ret_from_fork 517416 -27.7% 374206 ± 9% perf-sched.wait_and_delay.count.xlog_cil_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write 241356 -86.0% 33853 ± 10% perf-sched.wait_and_delay.count.xlog_wait_on_iclog.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write 8.75 ± 23% -44.0% 4.90 ± 16% perf-sched.wait_and_delay.max.ms.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write 0.98 ±141% +7e+05% 6824 ± 10% perf-sched.wait_and_delay.max.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64 0.99 ±141% +6.9e+05% 6824 ± 10% perf-sched.wait_and_delay.max.ms.do_syslog.part.0.kmsg_read.vfs_read 999.86 -66.7% 333.06 ±141% perf-sched.wait_and_delay.max.ms.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe 13.31 ± 16% -39.9% 8.00 ± 16% perf-sched.wait_and_delay.max.ms.io_schedule.wait_on_page_bit.wait_on_page_writeback.__filemap_fdatawait_range 90.59 ± 72% -72.0% 25.40 ± 12% perf-sched.wait_and_delay.max.ms.md_flush_request.raid0_make_request.md_handle_request.md_submit_bio 1013 +573.9% 6826 ± 10% perf-sched.wait_and_delay.max.ms.pipe_read.new_sync_read.vfs_read.ksys_read 85.23 ±123% -94.7% 4.51 ± 88% perf-sched.wait_and_delay.max.ms.preempt_schedule_common.__cond_resched.__alloc_pages.alloc_pages_vma.shmem_alloc_page 174.15 ± 55% -78.9% 36.75 ± 95% perf-sched.wait_and_delay.max.ms.preempt_schedule_common.__cond_resched.generic_perform_write.__generic_file_write_iter.generic_file_write_iter 13.38 ± 19% -84.0% 2.14 ±141% perf-sched.wait_and_delay.max.ms.preempt_schedule_common.__cond_resched.pagecache_get_page.shmem_getpage_gfp.shmem_write_begin 63.82 ± 16% +3174.1% 2089 ±132% perf-sched.wait_and_delay.max.ms.preempt_schedule_common.__cond_resched.process_one_work.worker_thread.kthread 6.62 ± 11% -20.9% 5.24 ± 2% perf-sched.wait_and_delay.max.ms.rcu_gp_kthread.kthread.ret_from_fork 500.62 +1181.3% 6414 ± 11% perf-sched.wait_and_delay.max.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait 8033 ± 10% -40.4% 4787 ± 22% perf-sched.wait_and_delay.max.ms.smpboot_thread_fn.kthread.ret_from_fork 8134 ± 4% -34.5% 5325 ± 7% perf-sched.wait_and_delay.max.ms.worker_thread.kthread.ret_from_fork 0.74 +39.6% 1.04 ± 2% perf-sched.wait_time.avg.ms.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write 1.32 ± 8% +46932.4% 618.48 ± 16% perf-sched.wait_time.avg.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64 787.97 ± 5% -32.2% 533.88 ± 4% perf-sched.wait_time.avg.ms.do_nanosleep.hrtimer_nanosleep.__x64_sys_nanosleep.do_syscall_64 1.33 ± 8% +46567.6% 618.50 ± 16% perf-sched.wait_time.avg.ms.do_syslog.part.0.kmsg_read.vfs_read 272.41 -61.4% 105.12 ± 6% perf-sched.wait_time.avg.ms.do_task_dead.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64 0.13 ± 24% -41.9% 0.08 ± 45% perf-sched.wait_time.avg.ms.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown] 0.25 -45.6% 0.14 perf-sched.wait_time.avg.ms.io_schedule.wait_on_page_bit.wait_on_page_writeback.__filemap_fdatawait_range 0.75 +16.4% 0.88 perf-sched.wait_time.avg.ms.md_flush_request.raid0_make_request.md_handle_request.md_submit_bio 127.63 ± 22% +59.7% 203.79 ± 15% perf-sched.wait_time.avg.ms.pipe_read.new_sync_read.vfs_read.ksys_read 20.10 ±101% -87.5% 2.51 ± 89% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.__alloc_pages.alloc_pages_vma.shmem_alloc_page 0.05 ±109% -100.0% 0.00 perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.__alloc_pages.pagecache_get_page.grab_cache_page_write_begin 0.04 ± 3% -33.6% 0.03 ± 8% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.__filemap_fdatawait_range.file_write_and_wait_range.xfs_file_fsync 0.70 -38.4% 0.43 perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.__flush_work.xlog_cil_force_lsn.xfs_log_force_lsn 1.40 ± 41% -66.0% 0.48 ± 86% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.do_user_addr_fault.exc_page_fault.asm_exc_page_fault 0.16 ± 12% -35.7% 0.10 ± 18% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.down_read.xfs_file_fsync.xfs_file_buffered_write 0.20 ± 37% -66.6% 0.07 ± 37% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.down_read.xfs_log_commit_cil.__xfs_trans_commit 0.15 ± 15% -39.6% 0.09 ± 6% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.down_write.xfs_ilock.xfs_buffered_write_iomap_begin 0.27 ± 45% -67.2% 0.09 ± 39% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.down_write.xfs_ilock.xfs_file_buffered_write 0.19 ± 16% -58.3% 0.08 ± 29% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.down_write.xfs_ilock.xfs_vn_update_time 0.22 ± 24% -35.1% 0.14 ± 33% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.iomap_write_actor.iomap_apply.iomap_file_buffered_write 0.23 ± 16% -21.8% 0.18 ± 6% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.kmem_cache_alloc.xfs_trans_alloc.xfs_vn_update_time 0.08 ± 96% +345.4% 0.34 ± 66% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.mempool_alloc.bio_alloc_bioset.submit_flushes 0.52 ± 2% +42.1% 0.73 ± 3% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.mempool_alloc.md_submit_bio.submit_bio_noacct 0.00 ±141% +14412.5% 0.39 ±113% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.mnt_want_write.do_unlinkat.do_syscall_64 4.68 ± 14% -76.7% 1.09 ±118% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.pagecache_get_page.shmem_getpage_gfp.shmem_write_begin 0.81 ± 15% -70.7% 0.24 ± 15% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate 0.25 ± 10% -62.7% 0.09 ± 61% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.submit_bio_checks.submit_bio_noacct.raid0_make_request 0.33 ± 6% -39.7% 0.20 ± 4% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.wait_for_completion.__flush_work.xlog_cil_force_lsn 0.75 ± 17% -61.3% 0.29 ± 22% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.wait_for_completion.stop_two_cpus.migrate_swap 1.07 +11.9% 1.19 perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.wait_for_completion_io_timeout.submit_bio_wait.blkdev_issue_flush 0.18 ± 25% -49.3% 0.09 ± 32% perf-sched.wait_time.avg.ms.preempt_schedule_common.__cond_resched.write_cache_pages.iomap_writepages.xfs_vm_writepages 0.22 ± 8% -31.9% 0.15 ± 25% perf-sched.wait_time.avg.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_bmapi_convert_delalloc 0.22 ± 15% -46.3% 0.12 ± 21% perf-sched.wait_time.avg.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_create 0.14 ± 8% -21.6% 0.11 ± 3% perf-sched.wait_time.avg.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_free_eofblocks 0.31 ± 15% -76.0% 0.08 ± 42% perf-sched.wait_time.avg.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_inactive_ifree 5.33 ± 29% -42.4% 3.07 ± 10% perf-sched.wait_time.avg.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_iomap_write_unwritten 0.34 ± 14% -33.2% 0.23 ± 23% perf-sched.wait_time.avg.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_remove 3.61 -47.3% 1.90 ± 2% perf-sched.wait_time.avg.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_trans_roll 0.50 ± 4% -42.0% 0.29 ± 7% perf-sched.wait_time.avg.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_vn_update_time 213.00 ± 3% +33.6% 284.53 ± 21% perf-sched.wait_time.avg.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait 0.65 ± 9% -28.3% 0.46 ± 8% perf-sched.wait_time.avg.ms.schedule_timeout.__down.down.xfs_buf_lock 0.33 ± 40% -100.0% 0.00 perf-sched.wait_time.avg.ms.schedule_timeout.__down.down.xlog_write_iclog 6.97 ± 2% -12.6% 6.09 ± 7% perf-sched.wait_time.avg.ms.schedule_timeout.rcu_gp_kthread.kthread.ret_from_fork 0.48 -44.9% 0.26 perf-sched.wait_time.avg.ms.schedule_timeout.wait_for_completion.__flush_work.xlog_cil_force_lsn 0.10 ± 20% -64.3% 0.04 ± 75% perf-sched.wait_time.avg.ms.schedule_timeout.wait_for_completion.stop_two_cpus.migrate_swap 487.33 ± 3% -17.6% 401.36 ± 4% perf-sched.wait_time.avg.ms.smpboot_thread_fn.kthread.ret_from_fork 4.26 ± 2% -23.4% 3.27 perf-sched.wait_time.avg.ms.worker_thread.kthread.ret_from_fork 0.90 -12.5% 0.79 perf-sched.wait_time.avg.ms.xlog_cil_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write 0.39 ±128% -100.0% 0.00 perf-sched.wait_time.avg.ms.xlog_state_get_iclog_space.xlog_write.xlog_cil_push_work.process_one_work 0.87 -73.2% 0.23 perf-sched.wait_time.avg.ms.xlog_wait_on_iclog.__xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write 5.18 ± 8% -35.3% 3.35 ± 2% perf-sched.wait_time.max.ms.__xfs_log_force_lsn.xfs_log_force_lsn.xfs_file_fsync.xfs_file_buffered_write 2.63 ± 8% +2.6e+05% 6824 ± 10% perf-sched.wait_time.max.ms.devkmsg_read.vfs_read.ksys_read.do_syscall_64 2.65 ± 8% +2.6e+05% 6824 ± 10% perf-sched.wait_time.max.ms.do_syslog.part.0.kmsg_read.vfs_read 3.16 ± 38% -49.3% 1.60 ± 72% perf-sched.wait_time.max.ms.exit_to_user_mode_prepare.irqentry_exit_to_user_mode.asm_sysvec_call_function_single.[unknown] 999.83 -66.2% 337.99 ±138% perf-sched.wait_time.max.ms.exit_to_user_mode_prepare.syscall_exit_to_user_mode.do_syscall_64.entry_SYSCALL_64_after_hwframe 7.37 ± 16% -41.0% 4.35 ± 5% perf-sched.wait_time.max.ms.io_schedule.wait_on_page_bit.wait_on_page_writeback.__filemap_fdatawait_range 90.51 ± 72% -71.9% 25.40 ± 12% perf-sched.wait_time.max.ms.md_flush_request.raid0_make_request.md_handle_request.md_submit_bio 1013 +573.9% 6826 ± 10% perf-sched.wait_time.max.ms.pipe_read.new_sync_read.vfs_read.ksys_read 85.23 ±123% -94.6% 4.59 ± 85% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.__alloc_pages.alloc_pages_vma.shmem_alloc_page 0.08 ±121% -100.0% 0.00 perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.__alloc_pages.pagecache_get_page.grab_cache_page_write_begin 0.34 ± 24% -84.0% 0.05 ± 15% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.__filemap_fdatawait_range.file_write_and_wait_range.xfs_file_fsync 4.02 ± 9% -44.5% 2.23 ± 36% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.__flush_work.xlog_cil_force_lsn.xfs_log_force_lsn 6.69 ± 65% -66.8% 2.22 ± 86% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.do_user_addr_fault.exc_page_fault.asm_exc_page_fault 2.84 -50.4% 1.41 ± 29% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.down_read.xfs_file_fsync.xfs_file_buffered_write 2.16 ± 34% -74.2% 0.56 ± 59% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.down_read.xfs_log_commit_cil.__xfs_trans_commit 3.30 ± 13% -42.0% 1.91 ± 23% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.down_write.xfs_ilock.xfs_buffered_write_iomap_begin 2.18 ± 19% -46.4% 1.17 ± 11% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.down_write.xfs_ilock.xfs_vn_update_time 174.15 ± 55% -78.9% 36.75 ± 95% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.generic_perform_write.__generic_file_write_iter.generic_file_write_iter 2.15 ± 22% -36.2% 1.37 ± 10% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.kmem_cache_alloc.xfs_trans_alloc.xfs_vn_update_time 1.62 ± 31% -55.2% 0.72 ± 57% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.mempool_alloc.bio_alloc_bioset.iomap_writepage_map 0.08 ± 96% +345.4% 0.34 ± 66% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.mempool_alloc.bio_alloc_bioset.submit_flushes 0.00 ±141% +14412.5% 0.39 ±113% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.mnt_want_write.do_unlinkat.do_syscall_64 13.38 ± 19% -81.9% 2.43 ±116% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.pagecache_get_page.shmem_getpage_gfp.shmem_write_begin 63.48 ± 17% +3191.3% 2089 ±132% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.process_one_work.worker_thread.kthread 21.36 ± 34% -63.4% 7.81 ± 26% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.stop_one_cpu.migrate_task_to.task_numa_migrate 0.85 ± 47% -49.3% 0.43 ± 69% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.stop_one_cpu.sched_exec.bprm_execve 1.98 ± 29% -53.5% 0.92 ± 66% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.submit_bio_checks.submit_bio_noacct.raid0_make_request 3.17 ± 7% -25.2% 2.37 ± 14% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.wait_for_completion.__flush_work.xlog_cil_force_lsn 25.72 ± 7% -45.8% 13.95 ± 8% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.wait_for_completion.stop_two_cpus.migrate_swap 3.07 ± 8% -56.4% 1.34 ± 21% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.write_cache_pages.iomap_writepages.xfs_vm_writepages 1.62 ± 49% -68.9% 0.50 ± 69% perf-sched.wait_time.max.ms.preempt_schedule_common.__cond_resched.xfs_trans_alloc.xfs_vn_update_time.file_update_time 1.23 ± 17% -25.0% 0.92 ± 26% perf-sched.wait_time.max.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_bmapi_convert_delalloc 1.40 ± 28% -54.1% 0.64 ± 14% perf-sched.wait_time.max.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_create 1.14 ± 34% -67.1% 0.37 ± 50% perf-sched.wait_time.max.ms.rwsem_down_read_slowpath.xfs_log_commit_cil.__xfs_trans_commit.xfs_inactive_ifree 0.40 ± 71% -70.6% 0.12 ± 95% perf-sched.wait_time.max.ms.rwsem_down_write_slowpath.path_openat.do_filp_open.do_sys_openat2 500.24 +1182.3% 6414 ± 11% perf-sched.wait_time.max.ms.schedule_hrtimeout_range_clock.ep_poll.do_epoll_wait.__x64_sys_epoll_wait 48.99 ± 43% -54.0% 22.53 ± 17% perf-sched.wait_time.max.ms.schedule_timeout.__down.down.xfs_buf_lock 1.29 ± 19% -100.0% 0.00 perf-sched.wait_time.max.ms.schedule_timeout.__down.down.xlog_write_iclog 0.10 ± 20% -64.3% 0.04 ± 75% perf-sched.wait_time.max.ms.schedule_timeout.wait_for_completion.stop_two_cpus.migrate_swap 8033 ± 10% -40.4% 4787 ± 22% perf-sched.wait_time.max.ms.smpboot_thread_fn.kthread.ret_from_fork 8134 ± 4% -34.5% 5325 ± 7% perf-sched.wait_time.max.ms.worker_thread.kthread.ret_from_fork 1.88 ±135% -100.0% 0.00 perf-sched.wait_time.max.ms.xlog_state_get_iclog_space.xlog_write.xlog_cil_push_work.process_one_work Disclaimer: Results have been estimated based on internal Intel analysis and are provided for informational purposes only. Any difference in system hardware or software design or configuration may affect actual performance. --- 0DAY/LKP+ Test Infrastructure Open Source Technology Center https://lists.01.org/hyperkitty/list/lkp@lists.01.org Intel Corporation Thanks, Oliver Sang