Greeting, FYI, we noticed a -12.3% regression of blogbench.write_score and a +9.6% improvement of blogbench.read_score due to commit: commit: 9092c71bb724dba2ecba849eae69e5c9d39bd3d2 ("mm: use sc->priority for slab shrink targets") https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master in testcase: blogbench on test machine: 16 threads Intel(R) Xeon(R) CPU D-1541 @ 2.10GHz with 8G memory with following parameters: disk: 1SSD fs: btrfs cpufreq_governor: performance test-description: Blogbench is a portable filesystem benchmark that tries to reproduce the load of a real-world busy file server. test-url: https://www.pureftpd.org/project/blogbench Details are as below: --------------------------------------------------------------------------------------------------> To reproduce: git clone https://github.com/intel/lkp-tests.git cd lkp-tests bin/lkp install job.yaml # job file is attached in this email bin/lkp run job.yaml ========================================================================================= compiler/cpufreq_governor/disk/fs/kconfig/rootfs/tbox_group/testcase: gcc-7/performance/1SSD/btrfs/x86_64-rhel-7.2/debian-x86_64-2016-08-31.cgz/lkp-bdw-de1/blogbench commit: fcb2b0c577 ("mm: show total hugetlb memory consumption in /proc/meminfo") 9092c71bb7 ("mm: use sc->priority for slab shrink targets") fcb2b0c577f145c7 9092c71bb724dba2ecba849eae ---------------- -------------------------- %stddev %change %stddev \ | \ 3256 -12.3% 2854 blogbench.write_score 1235237 ± 2% +9.6% 1354163 blogbench.read_score 28050912 -10.1% 25212230 blogbench.time.file_system_outputs 6481995 ± 3% +25.0% 8105320 ± 2% blogbench.time.involuntary_context_switches 906.00 +13.7% 1030 blogbench.time.percent_of_cpu_this_job_got 2552 +14.0% 2908 blogbench.time.system_time 173.80 +8.4% 188.32 blogbench.time.user_time 19353936 +3.6% 20045728 blogbench.time.voluntary_context_switches 8719514 +13.0% 9850451 softirqs.RCU 2.97 ± 5% -0.7 2.30 ± 3% mpstat.cpu.idle% 24.92 -6.5 18.46 mpstat.cpu.iowait% 0.65 ± 2% +0.1 0.75 mpstat.cpu.soft% 67.76 +6.7 74.45 mpstat.cpu.sys% 50206 -10.7% 44858 vmstat.io.bo 49.25 -9.1% 44.75 ± 2% vmstat.procs.b 224125 -1.8% 220135 vmstat.system.cs 48903 +10.7% 54134 vmstat.system.in 3460654 +10.8% 3834883 meminfo.Active 3380666 +11.0% 3752872 meminfo.Active(file) 1853849 -17.4% 1530415 meminfo.Inactive 1836507 -17.6% 1513054 meminfo.Inactive(file) 551311 -10.3% 494265 meminfo.SReclaimable 196525 -12.6% 171775 meminfo.SUnreclaim 747837 -10.9% 666040 meminfo.Slab 8.904e+08 -24.9% 6.683e+08 cpuidle.C1.time 22971020 -12.8% 20035820 cpuidle.C1.usage 2.518e+08 ± 3% -31.7% 1.72e+08 cpuidle.C1E.time 821393 ± 2% -33.3% 548003 cpuidle.C1E.usage 75460078 ± 2% -23.3% 57903768 ± 2% cpuidle.C3.time 136506 ± 3% -25.3% 101956 ± 3% cpuidle.C3.usage 56892498 ± 4% -23.3% 43608427 ± 4% cpuidle.C6.time 85034 ± 3% -33.9% 56184 ± 3% cpuidle.C6.usage 24373567 -24.5% 18395538 cpuidle.POLL.time 449033 ± 2% -10.8% 400493 cpuidle.POLL.usage 1832 +9.3% 2002 turbostat.Avg_MHz 22967645 -12.8% 20032521 turbostat.C1 18.43 -4.6 13.85 turbostat.C1% 821328 ± 2% -33.3% 547948 turbostat.C1E 5.21 ± 3% -1.6 3.56 turbostat.C1E% 136377 ± 3% -25.3% 101823 ± 3% turbostat.C3 1.56 ± 2% -0.4 1.20 ± 3% turbostat.C3% 84404 ± 3% -34.0% 55743 ± 3% turbostat.C6 1.17 ± 4% -0.3 0.90 ± 4% turbostat.C6% 25.93 -26.2% 19.14 turbostat.CPU%c1 0.12 ± 3% -19.1% 0.10 ± 9% turbostat.CPU%c3 14813304 +10.7% 16398388 turbostat.IRQ 38.19 +3.6% 39.56 turbostat.PkgWatt 4.51 +4.5% 4.71 turbostat.RAMWatt 8111200 ± 13% -63.2% 2986242 ± 48% proc-vmstat.compact_daemon_free_scanned 1026719 ± 30% -81.2% 193485 ± 30% proc-vmstat.compact_daemon_migrate_scanned 2444 ± 21% -63.3% 897.50 ± 20% proc-vmstat.compact_daemon_wake 8111200 ± 13% -63.2% 2986242 ± 48% proc-vmstat.compact_free_scanned 755491 ± 32% -81.6% 138856 ± 28% proc-vmstat.compact_isolated 1026719 ± 30% -81.2% 193485 ± 30% proc-vmstat.compact_migrate_scanned 137.75 ± 34% +2.8e+06% 3801062 ± 2% proc-vmstat.kswapd_inodesteal 6749 ± 20% -53.6% 3131 ± 12% proc-vmstat.kswapd_low_wmark_hit_quickly 844991 +11.2% 939487 proc-vmstat.nr_active_file 3900576 -10.5% 3490567 proc-vmstat.nr_dirtied 459789 -17.8% 377930 proc-vmstat.nr_inactive_file 137947 -10.3% 123720 proc-vmstat.nr_slab_reclaimable 49165 -12.6% 42989 proc-vmstat.nr_slab_unreclaimable 1382 ± 11% -26.2% 1020 ± 20% proc-vmstat.nr_writeback 3809266 -10.7% 3403350 proc-vmstat.nr_written 844489 +11.2% 938974 proc-vmstat.nr_zone_active_file 459855 -17.8% 378121 proc-vmstat.nr_zone_inactive_file 7055 ± 18% -52.0% 3389 ± 11% proc-vmstat.pageoutrun 33764911 ± 2% +21.3% 40946445 proc-vmstat.pgactivate 42044161 ± 2% +12.1% 47139065 proc-vmstat.pgdeactivate 92153 ± 20% -69.1% 28514 ± 24% proc-vmstat.pgmigrate_success 15212270 -10.7% 13591573 proc-vmstat.pgpgout 42053817 ± 2% +12.1% 47151755 proc-vmstat.pgrefill 11297 ±107% +1025.4% 127138 ± 21% proc-vmstat.pgscan_direct 19930162 -24.0% 15141439 proc-vmstat.pgscan_kswapd 19423629 -24.0% 14758807 proc-vmstat.pgsteal_kswapd 10868768 +184.8% 30950752 proc-vmstat.slabs_scanned 3361780 ± 3% -22.9% 2593327 ± 3% proc-vmstat.workingset_activate 4994722 ± 2% -43.2% 2835020 ± 2% proc-vmstat.workingset_refault 316427 -9.3% 286844 slabinfo.Acpi-Namespace.active_objs 3123 -9.4% 2829 slabinfo.Acpi-Namespace.active_slabs 318605 -9.4% 288623 slabinfo.Acpi-Namespace.num_objs 3123 -9.4% 2829 slabinfo.Acpi-Namespace.num_slabs 220514 -40.7% 130747 slabinfo.btrfs_delayed_node.active_objs 9751 -25.3% 7283 slabinfo.btrfs_delayed_node.active_slabs 263293 -25.3% 196669 slabinfo.btrfs_delayed_node.num_objs 9751 -25.3% 7283 slabinfo.btrfs_delayed_node.num_slabs 6383 ± 8% -12.0% 5615 ± 2% slabinfo.btrfs_delayed_ref_head.num_objs 9496 +15.5% 10969 slabinfo.btrfs_extent_buffer.active_objs 9980 +20.5% 12022 slabinfo.btrfs_extent_buffer.num_objs 260933 -10.7% 233136 slabinfo.btrfs_extent_map.active_objs 9392 -10.6% 8396 slabinfo.btrfs_extent_map.active_slabs 263009 -10.6% 235107 slabinfo.btrfs_extent_map.num_objs 9392 -10.6% 8396 slabinfo.btrfs_extent_map.num_slabs 271938 -10.3% 243802 slabinfo.btrfs_inode.active_objs 9804 -10.6% 8768 slabinfo.btrfs_inode.active_slabs 273856 -10.4% 245359 slabinfo.btrfs_inode.num_objs 9804 -10.6% 8768 slabinfo.btrfs_inode.num_slabs 7085 ± 5% -5.5% 6692 ± 2% slabinfo.btrfs_path.num_objs 311936 -16.4% 260797 slabinfo.dentry.active_objs 7803 -9.6% 7058 slabinfo.dentry.active_slabs 327759 -9.6% 296439 slabinfo.dentry.num_objs 7803 -9.6% 7058 slabinfo.dentry.num_slabs 2289 -23.3% 1755 ± 6% slabinfo.proc_inode_cache.active_objs 2292 -19.0% 1856 ± 6% slabinfo.proc_inode_cache.num_objs 261546 -12.3% 229485 slabinfo.radix_tree_node.active_objs 9404 -11.9% 8288 slabinfo.radix_tree_node.active_slabs 263347 -11.9% 232089 slabinfo.radix_tree_node.num_objs 9404 -11.9% 8288 slabinfo.radix_tree_node.num_slabs 1140424 ± 12% +40.2% 1598980 ± 14% sched_debug.cfs_rq:/.MIN_vruntime.max 790.55 +13.0% 893.20 ± 3% sched_debug.cfs_rq:/.exec_clock.stddev 1140425 ± 12% +40.2% 1598982 ± 14% sched_debug.cfs_rq:/.max_vruntime.max 0.83 ± 10% +21.5% 1.00 ± 8% sched_debug.cfs_rq:/.nr_running.avg 3.30 ± 99% +266.3% 12.09 ± 13% sched_debug.cfs_rq:/.removed.load_avg.avg 153.02 ± 97% +266.6% 560.96 ± 13% sched_debug.cfs_rq:/.removed.runnable_sum.avg 569.93 ±102% +173.2% 1556 ± 14% sched_debug.cfs_rq:/.removed.runnable_sum.stddev 1.42 ± 60% +501.5% 8.52 ± 34% sched_debug.cfs_rq:/.removed.util_avg.avg 19.88 ± 59% +288.9% 77.29 ± 16% sched_debug.cfs_rq:/.removed.util_avg.max 5.05 ± 58% +342.3% 22.32 ± 22% sched_debug.cfs_rq:/.removed.util_avg.stddev 791.44 ± 3% +47.7% 1168 ± 8% sched_debug.cfs_rq:/.util_avg.avg 1305 ± 6% +33.2% 1738 ± 5% sched_debug.cfs_rq:/.util_avg.max 450.25 ± 11% +66.2% 748.17 ± 14% sched_debug.cfs_rq:/.util_avg.min 220.82 ± 8% +21.1% 267.46 ± 5% sched_debug.cfs_rq:/.util_avg.stddev 363118 ± 11% -23.8% 276520 ± 11% sched_debug.cpu.avg_idle.avg 726003 ± 8% -30.8% 502313 ± 4% sched_debug.cpu.avg_idle.max 202629 ± 3% -32.2% 137429 ± 18% sched_debug.cpu.avg_idle.stddev 31.96 ± 28% +54.6% 49.42 ± 14% sched_debug.cpu.cpu_load[3].min 36.21 ± 25% +64.0% 59.38 ± 6% sched_debug.cpu.cpu_load[4].min 1007 ± 5% +20.7% 1216 ± 7% sched_debug.cpu.curr->pid.avg 4.50 ± 5% +14.8% 5.17 ± 5% sched_debug.cpu.nr_running.max 2476195 -11.8% 2185022 sched_debug.cpu.nr_switches.max 212888 -26.6% 156172 ± 3% sched_debug.cpu.nr_switches.stddev 3570 ± 2% -58.7% 1474 ± 2% sched_debug.cpu.nr_uninterruptible.max -803.67 -28.7% -573.38 sched_debug.cpu.nr_uninterruptible.min 1004 ± 2% -50.4% 498.55 ± 3% sched_debug.cpu.nr_uninterruptible.stddev 2478809 -11.7% 2189310 sched_debug.cpu.sched_count.max 214130 -26.5% 157298 ± 3% sched_debug.cpu.sched_count.stddev 489430 ± 2% -16.6% 408309 ± 2% sched_debug.cpu.sched_goidle.avg 724333 ± 2% -28.2% 520263 ± 2% sched_debug.cpu.sched_goidle.max 457611 -18.1% 374746 ± 3% sched_debug.cpu.sched_goidle.min 62957 ± 2% -47.4% 33138 ± 3% sched_debug.cpu.sched_goidle.stddev 676053 ± 2% -15.4% 571816 ± 2% sched_debug.cpu.ttwu_local.max 42669 ± 3% +22.3% 52198 sched_debug.cpu.ttwu_local.min 151873 ± 2% -18.3% 124118 ± 2% sched_debug.cpu.ttwu_local.stddev blogbench.write_score 3300 +-+------------------------------------------------------------------+ 3250 +-+ +. .+ +. .+ : : : +. .+ .+.+.+. .| |: +. .+ +.+.+.+ + + + : +. : : +. + +.+ + + | 3200 +-+ + +.+ + : + + : + + | 3150 +-+.+ ++ +.+ | 3100 +-+ | 3050 +-+ | | | 3000 +-+ | 2950 +-+ O O | 2900 +-O O O O | 2850 +-+ O O O O O O O OO O O O | | O O O O | 2800 O-+ O O | 2750 +-+------------------------------------------------------------------+ [*] bisect-good sample [O] bisect-bad sample Disclaimer: Results have been estimated based on internal Intel analysis and are provided for informational purposes only. Any difference in system hardware or software design or configuration may affect actual performance. Thanks, Xiaolong