* [sched/fair] caeb178c60f: +252.0% cpuidle.C1-SNB.time, +3.1% turbostat.Pkg_W @ 2014-08-21 14:01 ` Fengguang Wu 0 siblings, 0 replies; 6+ messages in thread From: Fengguang Wu @ 2014-08-21 14:01 UTC (permalink / raw) To: Rik van Riel; +Cc: Dave Hansen, LKML, lkp, Ingo Molnar [-- Attachment #1: Type: text/plain, Size: 11083 bytes --] Hi Rik, FYI, we noticed the below changes on git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip.git sched/core commit caeb178c60f4f93f1b45c0bc056b5cf6d217b67f ("sched/fair: Make update_sd_pick_busiest() return 'true' on a busier sd") testbox/testcase/testparams: lkp-sb03/nepim/300s-100%-tcp6 743cb1ff191f00f caeb178c60f4f93f1b45c0bc0 --------------- ------------------------- 29718911 ±45% +329.5% 1.277e+08 ±10% cpuidle.C1E-SNB.time 861 ±34% +1590.4% 14564 ±31% cpuidle.C3-SNB.usage 1.65e+08 ±20% +175.4% 4.544e+08 ±15% cpuidle.C1-SNB.time 24 ±41% +247.6% 86 ±23% numa-numastat.node1.other_node 27717 ±11% +98.7% 55085 ± 6% softirqs.RCU 180767 ±11% +86.7% 337416 ±10% cpuidle.C7-SNB.usage 104591 ±14% +77.4% 185581 ±10% cpuidle.C1E-SNB.usage 384 ±10% +33.3% 512 ±11% slabinfo.kmem_cache.num_objs 384 ±10% +33.3% 512 ±11% slabinfo.kmem_cache.active_objs 494 ± 8% +25.9% 622 ± 9% slabinfo.kmem_cache_node.active_objs 512 ± 7% +25.0% 640 ± 8% slabinfo.kmem_cache_node.num_objs 83427 ± 6% +10.3% 92028 ± 5% meminfo.DirectMap4k 9508 ± 1% +21.3% 11534 ± 7% slabinfo.kmalloc-512.active_objs 9838 ± 1% +20.5% 11852 ± 6% slabinfo.kmalloc-512.num_objs 53997 ± 6% +11.1% 59981 ± 4% numa-meminfo.node1.Slab 2662 ± 3% -9.0% 2424 ± 3% slabinfo.kmalloc-96.active_objs 2710 ± 3% -8.6% 2478 ± 3% slabinfo.kmalloc-96.num_objs 921 ±41% +3577.7% 33901 ±14% time.involuntary_context_switches 2371 ± 2% +15.5% 2739 ± 2% vmstat.system.in testbox/testcase/testparams: lkp-sb03/nepim/300s-100%-tcp 743cb1ff191f00f caeb178c60f4f93f1b45c0bc0 --------------- ------------------------- 20657207 ±31% +358.2% 94650352 ±18% cpuidle.C1E-SNB.time 29718911 ±45% +329.5% 1.277e+08 ±10% cpuidle.C1E-SNB.time 861 ±34% +1590.4% 14564 ±31% cpuidle.C3-SNB.usage 0.05 ±46% +812.5% 0.44 ±34% turbostat.%c3 1.12e+08 ±25% +364.8% 5.207e+08 ±15% cpuidle.C1-SNB.time 1.65e+08 ±20% +175.4% 4.544e+08 ±15% cpuidle.C1-SNB.time 35 ±19% +105.6% 72 ±28% numa-numastat.node1.other_node 24 ±41% +247.6% 86 ±23% numa-numastat.node1.other_node 43 ±22% +86.2% 80 ±26% numa-vmstat.node0.nr_dirtied 24576 ± 6% +113.9% 52574 ± 1% softirqs.RCU 27717 ±11% +98.7% 55085 ± 6% softirqs.RCU 211533 ± 6% +58.4% 334990 ± 8% cpuidle.C7-SNB.usage 180767 ±11% +86.7% 337416 ±10% cpuidle.C7-SNB.usage 77739 ±13% +52.9% 118876 ±18% cpuidle.C1E-SNB.usage 104591 ±14% +77.4% 185581 ±10% cpuidle.C1E-SNB.usage 32.09 ±14% -24.8% 24.12 ±18% turbostat.%pc2 9.04 ± 6% +41.6% 12.80 ± 6% turbostat.%c1 384 ±10% +33.3% 512 ±11% slabinfo.kmem_cache.num_objs 384 ±10% +33.3% 512 ±11% slabinfo.kmem_cache.active_objs 494 ± 8% +25.9% 622 ± 9% slabinfo.kmem_cache_node.active_objs 512 ± 7% +25.0% 640 ± 8% slabinfo.kmem_cache_node.num_objs 379 ± 9% +16.7% 443 ± 7% numa-vmstat.node0.nr_page_table_pages 83427 ± 6% +10.3% 92028 ± 5% meminfo.DirectMap4k 1579 ± 6% -15.3% 1338 ± 7% numa-meminfo.node1.PageTables 394 ± 6% -15.1% 334 ± 7% numa-vmstat.node1.nr_page_table_pages 1509 ± 7% +16.6% 1760 ± 7% numa-meminfo.node0.PageTables 12681 ± 1% -17.3% 10482 ±14% numa-meminfo.node1.AnonPages 3169 ± 1% -17.3% 2620 ±14% numa-vmstat.node1.nr_anon_pages 10171 ± 3% +10.9% 11283 ± 3% slabinfo.kmalloc-512.active_objs 9508 ± 1% +21.3% 11534 ± 7% slabinfo.kmalloc-512.active_objs 10481 ± 3% +10.9% 11620 ± 3% slabinfo.kmalloc-512.num_objs 9838 ± 1% +20.5% 11852 ± 6% slabinfo.kmalloc-512.num_objs 53997 ± 6% +11.1% 59981 ± 4% numa-meminfo.node1.Slab 5072 ± 1% +11.6% 5662 ± 3% slabinfo.kmalloc-2048.num_objs 4974 ± 1% +11.6% 5551 ± 3% slabinfo.kmalloc-2048.active_objs 12824 ± 2% -16.1% 10754 ±14% numa-meminfo.node1.Active(anon) 3205 ± 2% -16.2% 2687 ±14% numa-vmstat.node1.nr_active_anon 2662 ± 3% -9.0% 2424 ± 3% slabinfo.kmalloc-96.active_objs 2710 ± 3% -8.6% 2478 ± 3% slabinfo.kmalloc-96.num_objs 15791 ± 1% +15.2% 18192 ± 9% numa-meminfo.node0.AnonPages 3949 ± 1% +15.2% 4549 ± 9% numa-vmstat.node0.nr_anon_pages 13669 ± 1% -7.5% 12645 ± 2% slabinfo.kmalloc-16.num_objs 662 ±23% +4718.6% 31918 ±12% time.involuntary_context_switches 921 ±41% +3577.7% 33901 ±14% time.involuntary_context_switches 2463 ± 1% +13.1% 2786 ± 3% vmstat.system.in 2371 ± 2% +15.5% 2739 ± 2% vmstat.system.in 49.40 ± 2% +4.8% 51.79 ± 2% turbostat.Cor_W 77.48 ± 1% +3.1% 79.91 ± 1% turbostat.Pkg_W testbox/testcase/testparams: lkp-st02/dd-write/5m-11HDD-RAID5-cfq-xfs-1dd 743cb1ff191f00f caeb178c60f4f93f1b45c0bc0 --------------- ------------------------- 18571 ± 7% +31.4% 24396 ± 4% proc-vmstat.pgscan_direct_normal 39983 ± 2% +38.3% 55286 ± 0% perf-stat.cpu-migrations 4193962 ± 2% +20.9% 5072009 ± 3% perf-stat.iTLB-load-misses 4.568e+09 ± 2% -17.2% 3.781e+09 ± 1% perf-stat.L1-icache-load-misses 1.762e+10 ± 0% -7.8% 1.625e+10 ± 1% perf-stat.cache-references 1.408e+09 ± 1% -6.6% 1.315e+09 ± 1% perf-stat.branch-load-misses 1.407e+09 ± 1% -6.5% 1.316e+09 ± 1% perf-stat.branch-misses 6.839e+09 ± 1% +5.0% 7.185e+09 ± 2% perf-stat.LLC-loads 1.558e+10 ± 0% +3.5% 1.612e+10 ± 1% perf-stat.L1-dcache-load-misses 1.318e+12 ± 0% +3.4% 1.363e+12 ± 0% perf-stat.L1-icache-loads 2.979e+10 ± 1% +2.4% 3.051e+10 ± 0% perf-stat.L1-dcache-store-misses 1.893e+11 ± 0% +2.5% 1.94e+11 ± 0% perf-stat.branch-instructions 2.298e+11 ± 0% +2.7% 2.361e+11 ± 0% perf-stat.L1-dcache-stores 1.016e+12 ± 0% +2.6% 1.042e+12 ± 0% perf-stat.instructions 1.892e+11 ± 0% +2.5% 1.94e+11 ± 0% perf-stat.branch-loads 3.71e+11 ± 0% +2.4% 3.799e+11 ± 0% perf-stat.dTLB-loads 3.711e+11 ± 0% +2.3% 3.798e+11 ± 0% perf-stat.L1-dcache-loads 325768 ± 0% +2.7% 334461 ± 0% vmstat.io.bo 8083 ± 0% +2.4% 8278 ± 0% iostat.sdf.wrqm/s 8083 ± 0% +2.4% 8278 ± 0% iostat.sdk.wrqm/s 8082 ± 0% +2.4% 8276 ± 0% iostat.sdg.wrqm/s 32615 ± 0% +2.4% 33398 ± 0% iostat.sdf.wkB/s 32617 ± 0% +2.4% 33401 ± 0% iostat.sdk.wkB/s 32612 ± 0% +2.4% 33393 ± 0% iostat.sdg.wkB/s 8083 ± 0% +2.4% 8277 ± 0% iostat.sdl.wrqm/s 8083 ± 0% +2.4% 8276 ± 0% iostat.sdi.wrqm/s 8082 ± 0% +2.4% 8277 ± 0% iostat.sdc.wrqm/s 32614 ± 0% +2.4% 33396 ± 0% iostat.sdl.wkB/s 8083 ± 0% +2.4% 8278 ± 0% iostat.sde.wrqm/s 8082 ± 0% +2.4% 8277 ± 0% iostat.sdh.wrqm/s 8083 ± 0% +2.4% 8277 ± 0% iostat.sdd.wrqm/s 32614 ± 0% +2.4% 33393 ± 0% iostat.sdi.wkB/s 32611 ± 0% +2.4% 33395 ± 0% iostat.sdc.wkB/s 325759 ± 0% +2.4% 333577 ± 0% iostat.md0.wkB/s 1274 ± 0% +2.4% 1305 ± 0% iostat.md0.w/s 8082 ± 0% +2.4% 8277 ± 0% iostat.sdb.wrqm/s 32618 ± 0% +2.4% 33398 ± 0% iostat.sde.wkB/s 32612 ± 0% +2.4% 33395 ± 0% iostat.sdh.wkB/s 32618 ± 0% +2.4% 33397 ± 0% iostat.sdd.wkB/s 8084 ± 0% +2.4% 8278 ± 0% iostat.sdj.wrqm/s 32611 ± 0% +2.4% 33396 ± 0% iostat.sdb.wkB/s 32618 ± 0% +2.4% 33400 ± 0% iostat.sdj.wkB/s 2.3e+11 ± 0% +2.5% 2.357e+11 ± 0% perf-stat.dTLB-stores 4898 ± 0% +2.1% 5003 ± 0% vmstat.system.cs 1.017e+12 ± 0% +2.4% 1.042e+12 ± 0% perf-stat.iTLB-loads 1518279 ± 0% +2.1% 1549457 ± 0% perf-stat.context-switches 1.456e+12 ± 0% +1.4% 1.476e+12 ± 0% perf-stat.cpu-cycles 1.456e+12 ± 0% +1.3% 1.475e+12 ± 0% perf-stat.ref-cycles 1.819e+11 ± 0% +1.3% 1.843e+11 ± 0% perf-stat.bus-cycles lkp-sb03 is a Sandy Bridge-EP server. Memory: 64G Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Byte Order: Little Endian CPU(s): 32 On-line CPU(s) list: 0-31 Thread(s) per core: 2 Core(s) per socket: 8 Socket(s): 2 NUMA node(s): 2 Vendor ID: GenuineIntel CPU family: 6 Model: 45 Stepping: 6 CPU MHz: 3500.613 BogoMIPS: 5391.16 Virtualization: VT-x L1d cache: 32K L1i cache: 32K L2 cache: 256K L3 cache: 20480K NUMA node0 CPU(s): 0-7,16-23 NUMA node1 CPU(s): 8-15,24-31 lkp-st02 is Core2 Memory: 8G time.involuntary_context_switches 40000 O+------------------------------------------------------------------+ | O O O | 35000 ++O O O O O O | 30000 ++ O O O | | O O O | 25000 ++ O O | | | 20000 ++ | | | 15000 ++ | 10000 ++ | | | 5000 ++ | | .*. | 0 *+*--*-*-*-*--*-*-*-*--*-*-*-*--*-*-*--*-*-*-*--*---*-*--*-*-*-*--*-* [*] bisect-good sample [O] bisect-bad sample Disclaimer: Results have been estimated based on internal Intel analysis and are provided for informational purposes only. Any difference in system hardware or software design or configuration may affect actual performance. Thanks, Fengguang [-- Attachment #2: reproduce --] [-- Type: text/plain, Size: 2403 bytes --] echo performance > /sys/devices/system/cpu/cpu0/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu1/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu10/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu11/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu12/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu13/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu14/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu15/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu16/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu17/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu18/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu19/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu2/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu20/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu21/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu22/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu23/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu24/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu25/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu26/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu27/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu28/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu29/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu3/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu30/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu31/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu4/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu5/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu6/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu7/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu8/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu9/cpufreq/scaling_governor ./nepim ./nepim -c 127.0.0.1 -d -a 300 -n 32 ^ permalink raw reply [flat|nested] 6+ messages in thread
* [sched/fair] caeb178c60f: +252.0% cpuidle.C1-SNB.time, +3.1% turbostat.Pkg_W @ 2014-08-21 14:01 ` Fengguang Wu 0 siblings, 0 replies; 6+ messages in thread From: Fengguang Wu @ 2014-08-21 14:01 UTC (permalink / raw) To: lkp [-- Attachment #1: Type: text/plain, Size: 11282 bytes --] Hi Rik, FYI, we noticed the below changes on git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip.git sched/core commit caeb178c60f4f93f1b45c0bc056b5cf6d217b67f ("sched/fair: Make update_sd_pick_busiest() return 'true' on a busier sd") testbox/testcase/testparams: lkp-sb03/nepim/300s-100%-tcp6 743cb1ff191f00f caeb178c60f4f93f1b45c0bc0 --------------- ------------------------- 29718911 ±45% +329.5% 1.277e+08 ±10% cpuidle.C1E-SNB.time 861 ±34% +1590.4% 14564 ±31% cpuidle.C3-SNB.usage 1.65e+08 ±20% +175.4% 4.544e+08 ±15% cpuidle.C1-SNB.time 24 ±41% +247.6% 86 ±23% numa-numastat.node1.other_node 27717 ±11% +98.7% 55085 ± 6% softirqs.RCU 180767 ±11% +86.7% 337416 ±10% cpuidle.C7-SNB.usage 104591 ±14% +77.4% 185581 ±10% cpuidle.C1E-SNB.usage 384 ±10% +33.3% 512 ±11% slabinfo.kmem_cache.num_objs 384 ±10% +33.3% 512 ±11% slabinfo.kmem_cache.active_objs 494 ± 8% +25.9% 622 ± 9% slabinfo.kmem_cache_node.active_objs 512 ± 7% +25.0% 640 ± 8% slabinfo.kmem_cache_node.num_objs 83427 ± 6% +10.3% 92028 ± 5% meminfo.DirectMap4k 9508 ± 1% +21.3% 11534 ± 7% slabinfo.kmalloc-512.active_objs 9838 ± 1% +20.5% 11852 ± 6% slabinfo.kmalloc-512.num_objs 53997 ± 6% +11.1% 59981 ± 4% numa-meminfo.node1.Slab 2662 ± 3% -9.0% 2424 ± 3% slabinfo.kmalloc-96.active_objs 2710 ± 3% -8.6% 2478 ± 3% slabinfo.kmalloc-96.num_objs 921 ±41% +3577.7% 33901 ±14% time.involuntary_context_switches 2371 ± 2% +15.5% 2739 ± 2% vmstat.system.in testbox/testcase/testparams: lkp-sb03/nepim/300s-100%-tcp 743cb1ff191f00f caeb178c60f4f93f1b45c0bc0 --------------- ------------------------- 20657207 ±31% +358.2% 94650352 ±18% cpuidle.C1E-SNB.time 29718911 ±45% +329.5% 1.277e+08 ±10% cpuidle.C1E-SNB.time 861 ±34% +1590.4% 14564 ±31% cpuidle.C3-SNB.usage 0.05 ±46% +812.5% 0.44 ±34% turbostat.%c3 1.12e+08 ±25% +364.8% 5.207e+08 ±15% cpuidle.C1-SNB.time 1.65e+08 ±20% +175.4% 4.544e+08 ±15% cpuidle.C1-SNB.time 35 ±19% +105.6% 72 ±28% numa-numastat.node1.other_node 24 ±41% +247.6% 86 ±23% numa-numastat.node1.other_node 43 ±22% +86.2% 80 ±26% numa-vmstat.node0.nr_dirtied 24576 ± 6% +113.9% 52574 ± 1% softirqs.RCU 27717 ±11% +98.7% 55085 ± 6% softirqs.RCU 211533 ± 6% +58.4% 334990 ± 8% cpuidle.C7-SNB.usage 180767 ±11% +86.7% 337416 ±10% cpuidle.C7-SNB.usage 77739 ±13% +52.9% 118876 ±18% cpuidle.C1E-SNB.usage 104591 ±14% +77.4% 185581 ±10% cpuidle.C1E-SNB.usage 32.09 ±14% -24.8% 24.12 ±18% turbostat.%pc2 9.04 ± 6% +41.6% 12.80 ± 6% turbostat.%c1 384 ±10% +33.3% 512 ±11% slabinfo.kmem_cache.num_objs 384 ±10% +33.3% 512 ±11% slabinfo.kmem_cache.active_objs 494 ± 8% +25.9% 622 ± 9% slabinfo.kmem_cache_node.active_objs 512 ± 7% +25.0% 640 ± 8% slabinfo.kmem_cache_node.num_objs 379 ± 9% +16.7% 443 ± 7% numa-vmstat.node0.nr_page_table_pages 83427 ± 6% +10.3% 92028 ± 5% meminfo.DirectMap4k 1579 ± 6% -15.3% 1338 ± 7% numa-meminfo.node1.PageTables 394 ± 6% -15.1% 334 ± 7% numa-vmstat.node1.nr_page_table_pages 1509 ± 7% +16.6% 1760 ± 7% numa-meminfo.node0.PageTables 12681 ± 1% -17.3% 10482 ±14% numa-meminfo.node1.AnonPages 3169 ± 1% -17.3% 2620 ±14% numa-vmstat.node1.nr_anon_pages 10171 ± 3% +10.9% 11283 ± 3% slabinfo.kmalloc-512.active_objs 9508 ± 1% +21.3% 11534 ± 7% slabinfo.kmalloc-512.active_objs 10481 ± 3% +10.9% 11620 ± 3% slabinfo.kmalloc-512.num_objs 9838 ± 1% +20.5% 11852 ± 6% slabinfo.kmalloc-512.num_objs 53997 ± 6% +11.1% 59981 ± 4% numa-meminfo.node1.Slab 5072 ± 1% +11.6% 5662 ± 3% slabinfo.kmalloc-2048.num_objs 4974 ± 1% +11.6% 5551 ± 3% slabinfo.kmalloc-2048.active_objs 12824 ± 2% -16.1% 10754 ±14% numa-meminfo.node1.Active(anon) 3205 ± 2% -16.2% 2687 ±14% numa-vmstat.node1.nr_active_anon 2662 ± 3% -9.0% 2424 ± 3% slabinfo.kmalloc-96.active_objs 2710 ± 3% -8.6% 2478 ± 3% slabinfo.kmalloc-96.num_objs 15791 ± 1% +15.2% 18192 ± 9% numa-meminfo.node0.AnonPages 3949 ± 1% +15.2% 4549 ± 9% numa-vmstat.node0.nr_anon_pages 13669 ± 1% -7.5% 12645 ± 2% slabinfo.kmalloc-16.num_objs 662 ±23% +4718.6% 31918 ±12% time.involuntary_context_switches 921 ±41% +3577.7% 33901 ±14% time.involuntary_context_switches 2463 ± 1% +13.1% 2786 ± 3% vmstat.system.in 2371 ± 2% +15.5% 2739 ± 2% vmstat.system.in 49.40 ± 2% +4.8% 51.79 ± 2% turbostat.Cor_W 77.48 ± 1% +3.1% 79.91 ± 1% turbostat.Pkg_W testbox/testcase/testparams: lkp-st02/dd-write/5m-11HDD-RAID5-cfq-xfs-1dd 743cb1ff191f00f caeb178c60f4f93f1b45c0bc0 --------------- ------------------------- 18571 ± 7% +31.4% 24396 ± 4% proc-vmstat.pgscan_direct_normal 39983 ± 2% +38.3% 55286 ± 0% perf-stat.cpu-migrations 4193962 ± 2% +20.9% 5072009 ± 3% perf-stat.iTLB-load-misses 4.568e+09 ± 2% -17.2% 3.781e+09 ± 1% perf-stat.L1-icache-load-misses 1.762e+10 ± 0% -7.8% 1.625e+10 ± 1% perf-stat.cache-references 1.408e+09 ± 1% -6.6% 1.315e+09 ± 1% perf-stat.branch-load-misses 1.407e+09 ± 1% -6.5% 1.316e+09 ± 1% perf-stat.branch-misses 6.839e+09 ± 1% +5.0% 7.185e+09 ± 2% perf-stat.LLC-loads 1.558e+10 ± 0% +3.5% 1.612e+10 ± 1% perf-stat.L1-dcache-load-misses 1.318e+12 ± 0% +3.4% 1.363e+12 ± 0% perf-stat.L1-icache-loads 2.979e+10 ± 1% +2.4% 3.051e+10 ± 0% perf-stat.L1-dcache-store-misses 1.893e+11 ± 0% +2.5% 1.94e+11 ± 0% perf-stat.branch-instructions 2.298e+11 ± 0% +2.7% 2.361e+11 ± 0% perf-stat.L1-dcache-stores 1.016e+12 ± 0% +2.6% 1.042e+12 ± 0% perf-stat.instructions 1.892e+11 ± 0% +2.5% 1.94e+11 ± 0% perf-stat.branch-loads 3.71e+11 ± 0% +2.4% 3.799e+11 ± 0% perf-stat.dTLB-loads 3.711e+11 ± 0% +2.3% 3.798e+11 ± 0% perf-stat.L1-dcache-loads 325768 ± 0% +2.7% 334461 ± 0% vmstat.io.bo 8083 ± 0% +2.4% 8278 ± 0% iostat.sdf.wrqm/s 8083 ± 0% +2.4% 8278 ± 0% iostat.sdk.wrqm/s 8082 ± 0% +2.4% 8276 ± 0% iostat.sdg.wrqm/s 32615 ± 0% +2.4% 33398 ± 0% iostat.sdf.wkB/s 32617 ± 0% +2.4% 33401 ± 0% iostat.sdk.wkB/s 32612 ± 0% +2.4% 33393 ± 0% iostat.sdg.wkB/s 8083 ± 0% +2.4% 8277 ± 0% iostat.sdl.wrqm/s 8083 ± 0% +2.4% 8276 ± 0% iostat.sdi.wrqm/s 8082 ± 0% +2.4% 8277 ± 0% iostat.sdc.wrqm/s 32614 ± 0% +2.4% 33396 ± 0% iostat.sdl.wkB/s 8083 ± 0% +2.4% 8278 ± 0% iostat.sde.wrqm/s 8082 ± 0% +2.4% 8277 ± 0% iostat.sdh.wrqm/s 8083 ± 0% +2.4% 8277 ± 0% iostat.sdd.wrqm/s 32614 ± 0% +2.4% 33393 ± 0% iostat.sdi.wkB/s 32611 ± 0% +2.4% 33395 ± 0% iostat.sdc.wkB/s 325759 ± 0% +2.4% 333577 ± 0% iostat.md0.wkB/s 1274 ± 0% +2.4% 1305 ± 0% iostat.md0.w/s 8082 ± 0% +2.4% 8277 ± 0% iostat.sdb.wrqm/s 32618 ± 0% +2.4% 33398 ± 0% iostat.sde.wkB/s 32612 ± 0% +2.4% 33395 ± 0% iostat.sdh.wkB/s 32618 ± 0% +2.4% 33397 ± 0% iostat.sdd.wkB/s 8084 ± 0% +2.4% 8278 ± 0% iostat.sdj.wrqm/s 32611 ± 0% +2.4% 33396 ± 0% iostat.sdb.wkB/s 32618 ± 0% +2.4% 33400 ± 0% iostat.sdj.wkB/s 2.3e+11 ± 0% +2.5% 2.357e+11 ± 0% perf-stat.dTLB-stores 4898 ± 0% +2.1% 5003 ± 0% vmstat.system.cs 1.017e+12 ± 0% +2.4% 1.042e+12 ± 0% perf-stat.iTLB-loads 1518279 ± 0% +2.1% 1549457 ± 0% perf-stat.context-switches 1.456e+12 ± 0% +1.4% 1.476e+12 ± 0% perf-stat.cpu-cycles 1.456e+12 ± 0% +1.3% 1.475e+12 ± 0% perf-stat.ref-cycles 1.819e+11 ± 0% +1.3% 1.843e+11 ± 0% perf-stat.bus-cycles lkp-sb03 is a Sandy Bridge-EP server. Memory: 64G Architecture: x86_64 CPU op-mode(s): 32-bit, 64-bit Byte Order: Little Endian CPU(s): 32 On-line CPU(s) list: 0-31 Thread(s) per core: 2 Core(s) per socket: 8 Socket(s): 2 NUMA node(s): 2 Vendor ID: GenuineIntel CPU family: 6 Model: 45 Stepping: 6 CPU MHz: 3500.613 BogoMIPS: 5391.16 Virtualization: VT-x L1d cache: 32K L1i cache: 32K L2 cache: 256K L3 cache: 20480K NUMA node0 CPU(s): 0-7,16-23 NUMA node1 CPU(s): 8-15,24-31 lkp-st02 is Core2 Memory: 8G time.involuntary_context_switches 40000 O+------------------------------------------------------------------+ | O O O | 35000 ++O O O O O O | 30000 ++ O O O | | O O O | 25000 ++ O O | | | 20000 ++ | | | 15000 ++ | 10000 ++ | | | 5000 ++ | | .*. | 0 *+*--*-*-*-*--*-*-*-*--*-*-*-*--*-*-*--*-*-*-*--*---*-*--*-*-*-*--*-* [*] bisect-good sample [O] bisect-bad sample Disclaimer: Results have been estimated based on internal Intel analysis and are provided for informational purposes only. Any difference in system hardware or software design or configuration may affect actual performance. Thanks, Fengguang [-- Attachment #2: reproduce.ksh --] [-- Type: text/plain, Size: 2403 bytes --] echo performance > /sys/devices/system/cpu/cpu0/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu1/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu10/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu11/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu12/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu13/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu14/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu15/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu16/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu17/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu18/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu19/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu2/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu20/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu21/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu22/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu23/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu24/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu25/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu26/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu27/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu28/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu29/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu3/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu30/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu31/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu4/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu5/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu6/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu7/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu8/cpufreq/scaling_governor echo performance > /sys/devices/system/cpu/cpu9/cpufreq/scaling_governor ./nepim ./nepim -c 127.0.0.1 -d -a 300 -n 32 ^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: [sched/fair] caeb178c60f: +252.0% cpuidle.C1-SNB.time, +3.1% turbostat.Pkg_W 2014-08-21 14:01 ` Fengguang Wu @ 2014-08-21 14:16 ` Rik van Riel -1 siblings, 0 replies; 6+ messages in thread From: Rik van Riel @ 2014-08-21 14:16 UTC (permalink / raw) To: Fengguang Wu; +Cc: Dave Hansen, LKML, lkp, Ingo Molnar On 08/21/2014 10:01 AM, Fengguang Wu wrote: > Hi Rik, > > FYI, we noticed the below changes on > > git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip.git sched/core > commit caeb178c60f4f93f1b45c0bc056b5cf6d217b67f ("sched/fair: Make update_sd_pick_busiest() return 'true' on a busier sd") > > testbox/testcase/testparams: lkp-sb03/nepim/300s-100%-tcp6 Is this good or bad? The numbers suggest the xfs + raid5 workload is doing around 2.4% more IO to disk per second with this change in, and there is more CPU idle time in the system... For the tcp test, I see no throughput numbers, but I see more idle time as well as more time in turbo mode, and more softirqs, which could mean that more packets were handled. Does the patch introduce any performance issues, or did it simply trip up something in the statistics that your script noticed? > 743cb1ff191f00f caeb178c60f4f93f1b45c0bc0 > --------------- ------------------------- > 29718911 ±45% +329.5% 1.277e+08 ±10% cpuidle.C1E-SNB.time > 861 ±34% +1590.4% 14564 ±31% cpuidle.C3-SNB.usage > 1.65e+08 ±20% +175.4% 4.544e+08 ±15% cpuidle.C1-SNB.time > 24 ±41% +247.6% 86 ±23% numa-numastat.node1.other_node > 27717 ±11% +98.7% 55085 ± 6% softirqs.RCU > 180767 ±11% +86.7% 337416 ±10% cpuidle.C7-SNB.usage > 104591 ±14% +77.4% 185581 ±10% cpuidle.C1E-SNB.usage > 384 ±10% +33.3% 512 ±11% slabinfo.kmem_cache.num_objs > 384 ±10% +33.3% 512 ±11% slabinfo.kmem_cache.active_objs > 494 ± 8% +25.9% 622 ± 9% slabinfo.kmem_cache_node.active_objs > 512 ± 7% +25.0% 640 ± 8% slabinfo.kmem_cache_node.num_objs > 83427 ± 6% +10.3% 92028 ± 5% meminfo.DirectMap4k > 9508 ± 1% +21.3% 11534 ± 7% slabinfo.kmalloc-512.active_objs > 9838 ± 1% +20.5% 11852 ± 6% slabinfo.kmalloc-512.num_objs > 53997 ± 6% +11.1% 59981 ± 4% numa-meminfo.node1.Slab > 2662 ± 3% -9.0% 2424 ± 3% slabinfo.kmalloc-96.active_objs > 2710 ± 3% -8.6% 2478 ± 3% slabinfo.kmalloc-96.num_objs > 921 ±41% +3577.7% 33901 ±14% time.involuntary_context_switches > 2371 ± 2% +15.5% 2739 ± 2% vmstat.system.in > > testbox/testcase/testparams: lkp-sb03/nepim/300s-100%-tcp > > 743cb1ff191f00f caeb178c60f4f93f1b45c0bc0 > --------------- ------------------------- > 20657207 ±31% +358.2% 94650352 ±18% cpuidle.C1E-SNB.time > 29718911 ±45% +329.5% 1.277e+08 ±10% cpuidle.C1E-SNB.time > 861 ±34% +1590.4% 14564 ±31% cpuidle.C3-SNB.usage > 0.05 ±46% +812.5% 0.44 ±34% turbostat.%c3 > 1.12e+08 ±25% +364.8% 5.207e+08 ±15% cpuidle.C1-SNB.time > 1.65e+08 ±20% +175.4% 4.544e+08 ±15% cpuidle.C1-SNB.time > 35 ±19% +105.6% 72 ±28% numa-numastat.node1.other_node > 24 ±41% +247.6% 86 ±23% numa-numastat.node1.other_node > 43 ±22% +86.2% 80 ±26% numa-vmstat.node0.nr_dirtied > 24576 ± 6% +113.9% 52574 ± 1% softirqs.RCU > 27717 ±11% +98.7% 55085 ± 6% softirqs.RCU > 211533 ± 6% +58.4% 334990 ± 8% cpuidle.C7-SNB.usage > 180767 ±11% +86.7% 337416 ±10% cpuidle.C7-SNB.usage > 77739 ±13% +52.9% 118876 ±18% cpuidle.C1E-SNB.usage > 104591 ±14% +77.4% 185581 ±10% cpuidle.C1E-SNB.usage > 32.09 ±14% -24.8% 24.12 ±18% turbostat.%pc2 > 9.04 ± 6% +41.6% 12.80 ± 6% turbostat.%c1 > 384 ±10% +33.3% 512 ±11% slabinfo.kmem_cache.num_objs > 384 ±10% +33.3% 512 ±11% slabinfo.kmem_cache.active_objs > 494 ± 8% +25.9% 622 ± 9% slabinfo.kmem_cache_node.active_objs > 512 ± 7% +25.0% 640 ± 8% slabinfo.kmem_cache_node.num_objs > 379 ± 9% +16.7% 443 ± 7% numa-vmstat.node0.nr_page_table_pages > 83427 ± 6% +10.3% 92028 ± 5% meminfo.DirectMap4k > 1579 ± 6% -15.3% 1338 ± 7% numa-meminfo.node1.PageTables > 394 ± 6% -15.1% 334 ± 7% numa-vmstat.node1.nr_page_table_pages > 1509 ± 7% +16.6% 1760 ± 7% numa-meminfo.node0.PageTables > 12681 ± 1% -17.3% 10482 ±14% numa-meminfo.node1.AnonPages > 3169 ± 1% -17.3% 2620 ±14% numa-vmstat.node1.nr_anon_pages > 10171 ± 3% +10.9% 11283 ± 3% slabinfo.kmalloc-512.active_objs > 9508 ± 1% +21.3% 11534 ± 7% slabinfo.kmalloc-512.active_objs > 10481 ± 3% +10.9% 11620 ± 3% slabinfo.kmalloc-512.num_objs > 9838 ± 1% +20.5% 11852 ± 6% slabinfo.kmalloc-512.num_objs > 53997 ± 6% +11.1% 59981 ± 4% numa-meminfo.node1.Slab > 5072 ± 1% +11.6% 5662 ± 3% slabinfo.kmalloc-2048.num_objs > 4974 ± 1% +11.6% 5551 ± 3% slabinfo.kmalloc-2048.active_objs > 12824 ± 2% -16.1% 10754 ±14% numa-meminfo.node1.Active(anon) > 3205 ± 2% -16.2% 2687 ±14% numa-vmstat.node1.nr_active_anon > 2662 ± 3% -9.0% 2424 ± 3% slabinfo.kmalloc-96.active_objs > 2710 ± 3% -8.6% 2478 ± 3% slabinfo.kmalloc-96.num_objs > 15791 ± 1% +15.2% 18192 ± 9% numa-meminfo.node0.AnonPages > 3949 ± 1% +15.2% 4549 ± 9% numa-vmstat.node0.nr_anon_pages > 13669 ± 1% -7.5% 12645 ± 2% slabinfo.kmalloc-16.num_objs > 662 ±23% +4718.6% 31918 ±12% time.involuntary_context_switches > 921 ±41% +3577.7% 33901 ±14% time.involuntary_context_switches > 2463 ± 1% +13.1% 2786 ± 3% vmstat.system.in > 2371 ± 2% +15.5% 2739 ± 2% vmstat.system.in > 49.40 ± 2% +4.8% 51.79 ± 2% turbostat.Cor_W > 77.48 ± 1% +3.1% 79.91 ± 1% turbostat.Pkg_W > > testbox/testcase/testparams: lkp-st02/dd-write/5m-11HDD-RAID5-cfq-xfs-1dd > > 743cb1ff191f00f caeb178c60f4f93f1b45c0bc0 > --------------- ------------------------- > 18571 ± 7% +31.4% 24396 ± 4% proc-vmstat.pgscan_direct_normal > 39983 ± 2% +38.3% 55286 ± 0% perf-stat.cpu-migrations > 4193962 ± 2% +20.9% 5072009 ± 3% perf-stat.iTLB-load-misses > 4.568e+09 ± 2% -17.2% 3.781e+09 ± 1% perf-stat.L1-icache-load-misses > 1.762e+10 ± 0% -7.8% 1.625e+10 ± 1% perf-stat.cache-references > 1.408e+09 ± 1% -6.6% 1.315e+09 ± 1% perf-stat.branch-load-misses > 1.407e+09 ± 1% -6.5% 1.316e+09 ± 1% perf-stat.branch-misses > 6.839e+09 ± 1% +5.0% 7.185e+09 ± 2% perf-stat.LLC-loads > 1.558e+10 ± 0% +3.5% 1.612e+10 ± 1% perf-stat.L1-dcache-load-misses > 1.318e+12 ± 0% +3.4% 1.363e+12 ± 0% perf-stat.L1-icache-loads > 2.979e+10 ± 1% +2.4% 3.051e+10 ± 0% perf-stat.L1-dcache-store-misses > 1.893e+11 ± 0% +2.5% 1.94e+11 ± 0% perf-stat.branch-instructions > 2.298e+11 ± 0% +2.7% 2.361e+11 ± 0% perf-stat.L1-dcache-stores > 1.016e+12 ± 0% +2.6% 1.042e+12 ± 0% perf-stat.instructions > 1.892e+11 ± 0% +2.5% 1.94e+11 ± 0% perf-stat.branch-loads > 3.71e+11 ± 0% +2.4% 3.799e+11 ± 0% perf-stat.dTLB-loads > 3.711e+11 ± 0% +2.3% 3.798e+11 ± 0% perf-stat.L1-dcache-loads > 325768 ± 0% +2.7% 334461 ± 0% vmstat.io.bo > 8083 ± 0% +2.4% 8278 ± 0% iostat.sdf.wrqm/s > 8083 ± 0% +2.4% 8278 ± 0% iostat.sdk.wrqm/s > 8082 ± 0% +2.4% 8276 ± 0% iostat.sdg.wrqm/s > 32615 ± 0% +2.4% 33398 ± 0% iostat.sdf.wkB/s > 32617 ± 0% +2.4% 33401 ± 0% iostat.sdk.wkB/s > 32612 ± 0% +2.4% 33393 ± 0% iostat.sdg.wkB/s > 8083 ± 0% +2.4% 8277 ± 0% iostat.sdl.wrqm/s > 8083 ± 0% +2.4% 8276 ± 0% iostat.sdi.wrqm/s > 8082 ± 0% +2.4% 8277 ± 0% iostat.sdc.wrqm/s > 32614 ± 0% +2.4% 33396 ± 0% iostat.sdl.wkB/s > 8083 ± 0% +2.4% 8278 ± 0% iostat.sde.wrqm/s > 8082 ± 0% +2.4% 8277 ± 0% iostat.sdh.wrqm/s > 8083 ± 0% +2.4% 8277 ± 0% iostat.sdd.wrqm/s > 32614 ± 0% +2.4% 33393 ± 0% iostat.sdi.wkB/s > 32611 ± 0% +2.4% 33395 ± 0% iostat.sdc.wkB/s > 325759 ± 0% +2.4% 333577 ± 0% iostat.md0.wkB/s > 1274 ± 0% +2.4% 1305 ± 0% iostat.md0.w/s > 8082 ± 0% +2.4% 8277 ± 0% iostat.sdb.wrqm/s > 32618 ± 0% +2.4% 33398 ± 0% iostat.sde.wkB/s > 32612 ± 0% +2.4% 33395 ± 0% iostat.sdh.wkB/s > 32618 ± 0% +2.4% 33397 ± 0% iostat.sdd.wkB/s > 8084 ± 0% +2.4% 8278 ± 0% iostat.sdj.wrqm/s > 32611 ± 0% +2.4% 33396 ± 0% iostat.sdb.wkB/s > 32618 ± 0% +2.4% 33400 ± 0% iostat.sdj.wkB/s > 2.3e+11 ± 0% +2.5% 2.357e+11 ± 0% perf-stat.dTLB-stores > 4898 ± 0% +2.1% 5003 ± 0% vmstat.system.cs > 1.017e+12 ± 0% +2.4% 1.042e+12 ± 0% perf-stat.iTLB-loads > 1518279 ± 0% +2.1% 1549457 ± 0% perf-stat.context-switches > 1.456e+12 ± 0% +1.4% 1.476e+12 ± 0% perf-stat.cpu-cycles > 1.456e+12 ± 0% +1.3% 1.475e+12 ± 0% perf-stat.ref-cycles > 1.819e+11 ± 0% +1.3% 1.843e+11 ± 0% perf-stat.bus-cycles > > lkp-sb03 is a Sandy Bridge-EP server. > Memory: 64G > Architecture: x86_64 > CPU op-mode(s): 32-bit, 64-bit > Byte Order: Little Endian > CPU(s): 32 > On-line CPU(s) list: 0-31 > Thread(s) per core: 2 > Core(s) per socket: 8 > Socket(s): 2 > NUMA node(s): 2 > Vendor ID: GenuineIntel > CPU family: 6 > Model: 45 > Stepping: 6 > CPU MHz: 3500.613 > BogoMIPS: 5391.16 > Virtualization: VT-x > L1d cache: 32K > L1i cache: 32K > L2 cache: 256K > L3 cache: 20480K > NUMA node0 CPU(s): 0-7,16-23 > NUMA node1 CPU(s): 8-15,24-31 > > lkp-st02 is Core2 > Memory: 8G > > > > > time.involuntary_context_switches > > 40000 O+------------------------------------------------------------------+ > | O O O | > 35000 ++O O O O O O | > 30000 ++ O O O | > | O O O | > 25000 ++ O O | > | | > 20000 ++ | > | | > 15000 ++ | > 10000 ++ | > | | > 5000 ++ | > | .*. | > 0 *+*--*-*-*-*--*-*-*-*--*-*-*-*--*-*-*--*-*-*-*--*---*-*--*-*-*-*--*-* > > > [*] bisect-good sample > [O] bisect-bad sample > > > Disclaimer: > Results have been estimated based on internal Intel analysis and are provided > for informational purposes only. Any difference in system hardware or software > design or configuration may affect actual performance. > > Thanks, > Fengguang > ^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: [sched/fair] caeb178c60f: +252.0% cpuidle.C1-SNB.time, +3.1% turbostat.Pkg_W @ 2014-08-21 14:16 ` Rik van Riel 0 siblings, 0 replies; 6+ messages in thread From: Rik van Riel @ 2014-08-21 14:16 UTC (permalink / raw) To: lkp [-- Attachment #1: Type: text/plain, Size: 12235 bytes --] On 08/21/2014 10:01 AM, Fengguang Wu wrote: > Hi Rik, > > FYI, we noticed the below changes on > > git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip.git sched/core > commit caeb178c60f4f93f1b45c0bc056b5cf6d217b67f ("sched/fair: Make update_sd_pick_busiest() return 'true' on a busier sd") > > testbox/testcase/testparams: lkp-sb03/nepim/300s-100%-tcp6 Is this good or bad? The numbers suggest the xfs + raid5 workload is doing around 2.4% more IO to disk per second with this change in, and there is more CPU idle time in the system... For the tcp test, I see no throughput numbers, but I see more idle time as well as more time in turbo mode, and more softirqs, which could mean that more packets were handled. Does the patch introduce any performance issues, or did it simply trip up something in the statistics that your script noticed? > 743cb1ff191f00f caeb178c60f4f93f1b45c0bc0 > --------------- ------------------------- > 29718911 ±45% +329.5% 1.277e+08 ±10% cpuidle.C1E-SNB.time > 861 ±34% +1590.4% 14564 ±31% cpuidle.C3-SNB.usage > 1.65e+08 ±20% +175.4% 4.544e+08 ±15% cpuidle.C1-SNB.time > 24 ±41% +247.6% 86 ±23% numa-numastat.node1.other_node > 27717 ±11% +98.7% 55085 ± 6% softirqs.RCU > 180767 ±11% +86.7% 337416 ±10% cpuidle.C7-SNB.usage > 104591 ±14% +77.4% 185581 ±10% cpuidle.C1E-SNB.usage > 384 ±10% +33.3% 512 ±11% slabinfo.kmem_cache.num_objs > 384 ±10% +33.3% 512 ±11% slabinfo.kmem_cache.active_objs > 494 ± 8% +25.9% 622 ± 9% slabinfo.kmem_cache_node.active_objs > 512 ± 7% +25.0% 640 ± 8% slabinfo.kmem_cache_node.num_objs > 83427 ± 6% +10.3% 92028 ± 5% meminfo.DirectMap4k > 9508 ± 1% +21.3% 11534 ± 7% slabinfo.kmalloc-512.active_objs > 9838 ± 1% +20.5% 11852 ± 6% slabinfo.kmalloc-512.num_objs > 53997 ± 6% +11.1% 59981 ± 4% numa-meminfo.node1.Slab > 2662 ± 3% -9.0% 2424 ± 3% slabinfo.kmalloc-96.active_objs > 2710 ± 3% -8.6% 2478 ± 3% slabinfo.kmalloc-96.num_objs > 921 ±41% +3577.7% 33901 ±14% time.involuntary_context_switches > 2371 ± 2% +15.5% 2739 ± 2% vmstat.system.in > > testbox/testcase/testparams: lkp-sb03/nepim/300s-100%-tcp > > 743cb1ff191f00f caeb178c60f4f93f1b45c0bc0 > --------------- ------------------------- > 20657207 ±31% +358.2% 94650352 ±18% cpuidle.C1E-SNB.time > 29718911 ±45% +329.5% 1.277e+08 ±10% cpuidle.C1E-SNB.time > 861 ±34% +1590.4% 14564 ±31% cpuidle.C3-SNB.usage > 0.05 ±46% +812.5% 0.44 ±34% turbostat.%c3 > 1.12e+08 ±25% +364.8% 5.207e+08 ±15% cpuidle.C1-SNB.time > 1.65e+08 ±20% +175.4% 4.544e+08 ±15% cpuidle.C1-SNB.time > 35 ±19% +105.6% 72 ±28% numa-numastat.node1.other_node > 24 ±41% +247.6% 86 ±23% numa-numastat.node1.other_node > 43 ±22% +86.2% 80 ±26% numa-vmstat.node0.nr_dirtied > 24576 ± 6% +113.9% 52574 ± 1% softirqs.RCU > 27717 ±11% +98.7% 55085 ± 6% softirqs.RCU > 211533 ± 6% +58.4% 334990 ± 8% cpuidle.C7-SNB.usage > 180767 ±11% +86.7% 337416 ±10% cpuidle.C7-SNB.usage > 77739 ±13% +52.9% 118876 ±18% cpuidle.C1E-SNB.usage > 104591 ±14% +77.4% 185581 ±10% cpuidle.C1E-SNB.usage > 32.09 ±14% -24.8% 24.12 ±18% turbostat.%pc2 > 9.04 ± 6% +41.6% 12.80 ± 6% turbostat.%c1 > 384 ±10% +33.3% 512 ±11% slabinfo.kmem_cache.num_objs > 384 ±10% +33.3% 512 ±11% slabinfo.kmem_cache.active_objs > 494 ± 8% +25.9% 622 ± 9% slabinfo.kmem_cache_node.active_objs > 512 ± 7% +25.0% 640 ± 8% slabinfo.kmem_cache_node.num_objs > 379 ± 9% +16.7% 443 ± 7% numa-vmstat.node0.nr_page_table_pages > 83427 ± 6% +10.3% 92028 ± 5% meminfo.DirectMap4k > 1579 ± 6% -15.3% 1338 ± 7% numa-meminfo.node1.PageTables > 394 ± 6% -15.1% 334 ± 7% numa-vmstat.node1.nr_page_table_pages > 1509 ± 7% +16.6% 1760 ± 7% numa-meminfo.node0.PageTables > 12681 ± 1% -17.3% 10482 ±14% numa-meminfo.node1.AnonPages > 3169 ± 1% -17.3% 2620 ±14% numa-vmstat.node1.nr_anon_pages > 10171 ± 3% +10.9% 11283 ± 3% slabinfo.kmalloc-512.active_objs > 9508 ± 1% +21.3% 11534 ± 7% slabinfo.kmalloc-512.active_objs > 10481 ± 3% +10.9% 11620 ± 3% slabinfo.kmalloc-512.num_objs > 9838 ± 1% +20.5% 11852 ± 6% slabinfo.kmalloc-512.num_objs > 53997 ± 6% +11.1% 59981 ± 4% numa-meminfo.node1.Slab > 5072 ± 1% +11.6% 5662 ± 3% slabinfo.kmalloc-2048.num_objs > 4974 ± 1% +11.6% 5551 ± 3% slabinfo.kmalloc-2048.active_objs > 12824 ± 2% -16.1% 10754 ±14% numa-meminfo.node1.Active(anon) > 3205 ± 2% -16.2% 2687 ±14% numa-vmstat.node1.nr_active_anon > 2662 ± 3% -9.0% 2424 ± 3% slabinfo.kmalloc-96.active_objs > 2710 ± 3% -8.6% 2478 ± 3% slabinfo.kmalloc-96.num_objs > 15791 ± 1% +15.2% 18192 ± 9% numa-meminfo.node0.AnonPages > 3949 ± 1% +15.2% 4549 ± 9% numa-vmstat.node0.nr_anon_pages > 13669 ± 1% -7.5% 12645 ± 2% slabinfo.kmalloc-16.num_objs > 662 ±23% +4718.6% 31918 ±12% time.involuntary_context_switches > 921 ±41% +3577.7% 33901 ±14% time.involuntary_context_switches > 2463 ± 1% +13.1% 2786 ± 3% vmstat.system.in > 2371 ± 2% +15.5% 2739 ± 2% vmstat.system.in > 49.40 ± 2% +4.8% 51.79 ± 2% turbostat.Cor_W > 77.48 ± 1% +3.1% 79.91 ± 1% turbostat.Pkg_W > > testbox/testcase/testparams: lkp-st02/dd-write/5m-11HDD-RAID5-cfq-xfs-1dd > > 743cb1ff191f00f caeb178c60f4f93f1b45c0bc0 > --------------- ------------------------- > 18571 ± 7% +31.4% 24396 ± 4% proc-vmstat.pgscan_direct_normal > 39983 ± 2% +38.3% 55286 ± 0% perf-stat.cpu-migrations > 4193962 ± 2% +20.9% 5072009 ± 3% perf-stat.iTLB-load-misses > 4.568e+09 ± 2% -17.2% 3.781e+09 ± 1% perf-stat.L1-icache-load-misses > 1.762e+10 ± 0% -7.8% 1.625e+10 ± 1% perf-stat.cache-references > 1.408e+09 ± 1% -6.6% 1.315e+09 ± 1% perf-stat.branch-load-misses > 1.407e+09 ± 1% -6.5% 1.316e+09 ± 1% perf-stat.branch-misses > 6.839e+09 ± 1% +5.0% 7.185e+09 ± 2% perf-stat.LLC-loads > 1.558e+10 ± 0% +3.5% 1.612e+10 ± 1% perf-stat.L1-dcache-load-misses > 1.318e+12 ± 0% +3.4% 1.363e+12 ± 0% perf-stat.L1-icache-loads > 2.979e+10 ± 1% +2.4% 3.051e+10 ± 0% perf-stat.L1-dcache-store-misses > 1.893e+11 ± 0% +2.5% 1.94e+11 ± 0% perf-stat.branch-instructions > 2.298e+11 ± 0% +2.7% 2.361e+11 ± 0% perf-stat.L1-dcache-stores > 1.016e+12 ± 0% +2.6% 1.042e+12 ± 0% perf-stat.instructions > 1.892e+11 ± 0% +2.5% 1.94e+11 ± 0% perf-stat.branch-loads > 3.71e+11 ± 0% +2.4% 3.799e+11 ± 0% perf-stat.dTLB-loads > 3.711e+11 ± 0% +2.3% 3.798e+11 ± 0% perf-stat.L1-dcache-loads > 325768 ± 0% +2.7% 334461 ± 0% vmstat.io.bo > 8083 ± 0% +2.4% 8278 ± 0% iostat.sdf.wrqm/s > 8083 ± 0% +2.4% 8278 ± 0% iostat.sdk.wrqm/s > 8082 ± 0% +2.4% 8276 ± 0% iostat.sdg.wrqm/s > 32615 ± 0% +2.4% 33398 ± 0% iostat.sdf.wkB/s > 32617 ± 0% +2.4% 33401 ± 0% iostat.sdk.wkB/s > 32612 ± 0% +2.4% 33393 ± 0% iostat.sdg.wkB/s > 8083 ± 0% +2.4% 8277 ± 0% iostat.sdl.wrqm/s > 8083 ± 0% +2.4% 8276 ± 0% iostat.sdi.wrqm/s > 8082 ± 0% +2.4% 8277 ± 0% iostat.sdc.wrqm/s > 32614 ± 0% +2.4% 33396 ± 0% iostat.sdl.wkB/s > 8083 ± 0% +2.4% 8278 ± 0% iostat.sde.wrqm/s > 8082 ± 0% +2.4% 8277 ± 0% iostat.sdh.wrqm/s > 8083 ± 0% +2.4% 8277 ± 0% iostat.sdd.wrqm/s > 32614 ± 0% +2.4% 33393 ± 0% iostat.sdi.wkB/s > 32611 ± 0% +2.4% 33395 ± 0% iostat.sdc.wkB/s > 325759 ± 0% +2.4% 333577 ± 0% iostat.md0.wkB/s > 1274 ± 0% +2.4% 1305 ± 0% iostat.md0.w/s > 8082 ± 0% +2.4% 8277 ± 0% iostat.sdb.wrqm/s > 32618 ± 0% +2.4% 33398 ± 0% iostat.sde.wkB/s > 32612 ± 0% +2.4% 33395 ± 0% iostat.sdh.wkB/s > 32618 ± 0% +2.4% 33397 ± 0% iostat.sdd.wkB/s > 8084 ± 0% +2.4% 8278 ± 0% iostat.sdj.wrqm/s > 32611 ± 0% +2.4% 33396 ± 0% iostat.sdb.wkB/s > 32618 ± 0% +2.4% 33400 ± 0% iostat.sdj.wkB/s > 2.3e+11 ± 0% +2.5% 2.357e+11 ± 0% perf-stat.dTLB-stores > 4898 ± 0% +2.1% 5003 ± 0% vmstat.system.cs > 1.017e+12 ± 0% +2.4% 1.042e+12 ± 0% perf-stat.iTLB-loads > 1518279 ± 0% +2.1% 1549457 ± 0% perf-stat.context-switches > 1.456e+12 ± 0% +1.4% 1.476e+12 ± 0% perf-stat.cpu-cycles > 1.456e+12 ± 0% +1.3% 1.475e+12 ± 0% perf-stat.ref-cycles > 1.819e+11 ± 0% +1.3% 1.843e+11 ± 0% perf-stat.bus-cycles > > lkp-sb03 is a Sandy Bridge-EP server. > Memory: 64G > Architecture: x86_64 > CPU op-mode(s): 32-bit, 64-bit > Byte Order: Little Endian > CPU(s): 32 > On-line CPU(s) list: 0-31 > Thread(s) per core: 2 > Core(s) per socket: 8 > Socket(s): 2 > NUMA node(s): 2 > Vendor ID: GenuineIntel > CPU family: 6 > Model: 45 > Stepping: 6 > CPU MHz: 3500.613 > BogoMIPS: 5391.16 > Virtualization: VT-x > L1d cache: 32K > L1i cache: 32K > L2 cache: 256K > L3 cache: 20480K > NUMA node0 CPU(s): 0-7,16-23 > NUMA node1 CPU(s): 8-15,24-31 > > lkp-st02 is Core2 > Memory: 8G > > > > > time.involuntary_context_switches > > 40000 O+------------------------------------------------------------------+ > | O O O | > 35000 ++O O O O O O | > 30000 ++ O O O | > | O O O | > 25000 ++ O O | > | | > 20000 ++ | > | | > 15000 ++ | > 10000 ++ | > | | > 5000 ++ | > | .*. | > 0 *+*--*-*-*-*--*-*-*-*--*-*-*-*--*-*-*--*-*-*-*--*---*-*--*-*-*-*--*-* > > > [*] bisect-good sample > [O] bisect-bad sample > > > Disclaimer: > Results have been estimated based on internal Intel analysis and are provided > for informational purposes only. Any difference in system hardware or software > design or configuration may affect actual performance. > > Thanks, > Fengguang > ^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: [sched/fair] caeb178c60f: +252.0% cpuidle.C1-SNB.time, +3.1% turbostat.Pkg_W 2014-08-21 14:16 ` Rik van Riel @ 2014-08-21 15:00 ` Fengguang Wu -1 siblings, 0 replies; 6+ messages in thread From: Fengguang Wu @ 2014-08-21 15:00 UTC (permalink / raw) To: Rik van Riel; +Cc: Dave Hansen, LKML, lkp, Ingo Molnar On Thu, Aug 21, 2014 at 10:16:13AM -0400, Rik van Riel wrote: > On 08/21/2014 10:01 AM, Fengguang Wu wrote: > > Hi Rik, > > > > FYI, we noticed the below changes on > > > > git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip.git sched/core > > commit caeb178c60f4f93f1b45c0bc056b5cf6d217b67f ("sched/fair: Make update_sd_pick_busiest() return 'true' on a busier sd") > > > > testbox/testcase/testparams: lkp-sb03/nepim/300s-100%-tcp6 > > Is this good or bad? It seems mixed results. The throughput is 2.4% better in sequential write test, while the power consumption (turbostat.Pkg_W) increases by 3.1% in the nepim/300s-100%-tcp test. > The numbers suggest the xfs + raid5 workload is doing around 2.4% > more IO to disk per second with this change in, and there is more Right. > CPU idle time in the system... Sorry "cpuidle" is the monitor name. You can find its code here: https://git.kernel.org/cgit/linux/kernel/git/wfg/lkp-tests.git/tree/monitors/cpuidle "cpuidle.C1-SNB.time" means the time spend in C1 state. > For the tcp test, I see no throughput numbers, but I see more > idle time as well as more time in turbo mode, and more softirqs, > which could mean that more packets were handled. Again, "turbostat" is a monitor name. "turbostat.Pkg_W" means the CPU package watts reported by the turbostat tool. > Does the patch introduce any performance issues, or did it > simply trip up something in the statistics that your script > noticed? In normal LKP reports, only changed stats are listed. Here is the performance/power index comparison, which lists all performance/power related stats. The index is geometric average of all results. Baseline is 100 for 743cb1ff191f00f. 100 perf-index (the larger, the better) 98 power-index (the larger, the better) 743cb1ff191f00f caeb178c60f4f93f1b45c0bc0 testbox/testcase/testparams --------------- ------------------------- --------------------------- %stddev %change %stddev \ | / 691053 ± 4% -5.1% 656100 ± 4% lkp-sb03/nepim/300s-100%-tcp 570185 ± 7% +5.4% 600774 ± 4% lkp-sb03/nepim/300s-100%-tcp6 1261238 ± 5% -0.3% 1256875 ± 4% TOTAL nepim.tcp.avg.kbps_in 743cb1ff191f00f caeb178c60f4f93f1b45c0bc0 --------------- ------------------------- 691216 ± 4% -5.1% 656264 ± 4% lkp-sb03/nepim/300s-100%-tcp 570347 ± 7% +5.4% 600902 ± 4% lkp-sb03/nepim/300s-100%-tcp6 1261564 ± 5% -0.3% 1257167 ± 4% TOTAL nepim.tcp.avg.kbps_out 743cb1ff191f00f caeb178c60f4f93f1b45c0bc0 --------------- ------------------------- 77.48 ± 1% +3.1% 79.91 ± 1% lkp-sb03/nepim/300s-100%-tcp 79.69 ± 2% -0.6% 79.21 ± 1% lkp-sb03/nepim/300s-100%-tcp6 157.17 ± 2% +1.2% 159.13 ± 1% TOTAL turbostat.Pkg_W 743cb1ff191f00f caeb178c60f4f93f1b45c0bc0 --------------- ------------------------- 6.05 ± 1% +1.2% 6.12 ± 1% lkp-sb03/nepim/300s-100%-tcp 6.06 ± 0% +1.0% 6.12 ± 1% lkp-sb03/nepim/300s-100%-tcp6 12.11 ± 1% +1.1% 12.24 ± 1% TOTAL turbostat.%c0 743cb1ff191f00f caeb178c60f4f93f1b45c0bc0 --------------- ------------------------- 325759 ± 0% +2.4% 333577 ± 0% lkp-st02/dd-write/5m-11HDD-RAID5-cfq-xfs-1dd 325759 ± 0% +2.4% 333577 ± 0% TOTAL iostat.md0.wkB/s The nepim throughput numbers are not stable enough comparing to the change, so are not regarded as real changes in the original email. I will need to increase its test time to make it more stable.. Thanks, Fengguang > > 743cb1ff191f00f caeb178c60f4f93f1b45c0bc0 > > --------------- ------------------------- > > 29718911 ±45% +329.5% 1.277e+08 ±10% cpuidle.C1E-SNB.time > > 861 ±34% +1590.4% 14564 ±31% cpuidle.C3-SNB.usage > > 1.65e+08 ±20% +175.4% 4.544e+08 ±15% cpuidle.C1-SNB.time > > 24 ±41% +247.6% 86 ±23% numa-numastat.node1.other_node > > 27717 ±11% +98.7% 55085 ± 6% softirqs.RCU > > 180767 ±11% +86.7% 337416 ±10% cpuidle.C7-SNB.usage > > 104591 ±14% +77.4% 185581 ±10% cpuidle.C1E-SNB.usage > > 384 ±10% +33.3% 512 ±11% slabinfo.kmem_cache.num_objs > > 384 ±10% +33.3% 512 ±11% slabinfo.kmem_cache.active_objs > > 494 ± 8% +25.9% 622 ± 9% slabinfo.kmem_cache_node.active_objs > > 512 ± 7% +25.0% 640 ± 8% slabinfo.kmem_cache_node.num_objs > > 83427 ± 6% +10.3% 92028 ± 5% meminfo.DirectMap4k > > 9508 ± 1% +21.3% 11534 ± 7% slabinfo.kmalloc-512.active_objs > > 9838 ± 1% +20.5% 11852 ± 6% slabinfo.kmalloc-512.num_objs > > 53997 ± 6% +11.1% 59981 ± 4% numa-meminfo.node1.Slab > > 2662 ± 3% -9.0% 2424 ± 3% slabinfo.kmalloc-96.active_objs > > 2710 ± 3% -8.6% 2478 ± 3% slabinfo.kmalloc-96.num_objs > > 921 ±41% +3577.7% 33901 ±14% time.involuntary_context_switches > > 2371 ± 2% +15.5% 2739 ± 2% vmstat.system.in > > > > testbox/testcase/testparams: lkp-sb03/nepim/300s-100%-tcp > > > > 743cb1ff191f00f caeb178c60f4f93f1b45c0bc0 > > --------------- ------------------------- > > 20657207 ±31% +358.2% 94650352 ±18% cpuidle.C1E-SNB.time > > 29718911 ±45% +329.5% 1.277e+08 ±10% cpuidle.C1E-SNB.time > > 861 ±34% +1590.4% 14564 ±31% cpuidle.C3-SNB.usage > > 0.05 ±46% +812.5% 0.44 ±34% turbostat.%c3 > > 1.12e+08 ±25% +364.8% 5.207e+08 ±15% cpuidle.C1-SNB.time > > 1.65e+08 ±20% +175.4% 4.544e+08 ±15% cpuidle.C1-SNB.time > > 35 ±19% +105.6% 72 ±28% numa-numastat.node1.other_node > > 24 ±41% +247.6% 86 ±23% numa-numastat.node1.other_node > > 43 ±22% +86.2% 80 ±26% numa-vmstat.node0.nr_dirtied > > 24576 ± 6% +113.9% 52574 ± 1% softirqs.RCU > > 27717 ±11% +98.7% 55085 ± 6% softirqs.RCU > > 211533 ± 6% +58.4% 334990 ± 8% cpuidle.C7-SNB.usage > > 180767 ±11% +86.7% 337416 ±10% cpuidle.C7-SNB.usage > > 77739 ±13% +52.9% 118876 ±18% cpuidle.C1E-SNB.usage > > 104591 ±14% +77.4% 185581 ±10% cpuidle.C1E-SNB.usage > > 32.09 ±14% -24.8% 24.12 ±18% turbostat.%pc2 > > 9.04 ± 6% +41.6% 12.80 ± 6% turbostat.%c1 > > 384 ±10% +33.3% 512 ±11% slabinfo.kmem_cache.num_objs > > 384 ±10% +33.3% 512 ±11% slabinfo.kmem_cache.active_objs > > 494 ± 8% +25.9% 622 ± 9% slabinfo.kmem_cache_node.active_objs > > 512 ± 7% +25.0% 640 ± 8% slabinfo.kmem_cache_node.num_objs > > 379 ± 9% +16.7% 443 ± 7% numa-vmstat.node0.nr_page_table_pages > > 83427 ± 6% +10.3% 92028 ± 5% meminfo.DirectMap4k > > 1579 ± 6% -15.3% 1338 ± 7% numa-meminfo.node1.PageTables > > 394 ± 6% -15.1% 334 ± 7% numa-vmstat.node1.nr_page_table_pages > > 1509 ± 7% +16.6% 1760 ± 7% numa-meminfo.node0.PageTables > > 12681 ± 1% -17.3% 10482 ±14% numa-meminfo.node1.AnonPages > > 3169 ± 1% -17.3% 2620 ±14% numa-vmstat.node1.nr_anon_pages > > 10171 ± 3% +10.9% 11283 ± 3% slabinfo.kmalloc-512.active_objs > > 9508 ± 1% +21.3% 11534 ± 7% slabinfo.kmalloc-512.active_objs > > 10481 ± 3% +10.9% 11620 ± 3% slabinfo.kmalloc-512.num_objs > > 9838 ± 1% +20.5% 11852 ± 6% slabinfo.kmalloc-512.num_objs > > 53997 ± 6% +11.1% 59981 ± 4% numa-meminfo.node1.Slab > > 5072 ± 1% +11.6% 5662 ± 3% slabinfo.kmalloc-2048.num_objs > > 4974 ± 1% +11.6% 5551 ± 3% slabinfo.kmalloc-2048.active_objs > > 12824 ± 2% -16.1% 10754 ±14% numa-meminfo.node1.Active(anon) > > 3205 ± 2% -16.2% 2687 ±14% numa-vmstat.node1.nr_active_anon > > 2662 ± 3% -9.0% 2424 ± 3% slabinfo.kmalloc-96.active_objs > > 2710 ± 3% -8.6% 2478 ± 3% slabinfo.kmalloc-96.num_objs > > 15791 ± 1% +15.2% 18192 ± 9% numa-meminfo.node0.AnonPages > > 3949 ± 1% +15.2% 4549 ± 9% numa-vmstat.node0.nr_anon_pages > > 13669 ± 1% -7.5% 12645 ± 2% slabinfo.kmalloc-16.num_objs > > 662 ±23% +4718.6% 31918 ±12% time.involuntary_context_switches > > 921 ±41% +3577.7% 33901 ±14% time.involuntary_context_switches > > 2463 ± 1% +13.1% 2786 ± 3% vmstat.system.in > > 2371 ± 2% +15.5% 2739 ± 2% vmstat.system.in > > 49.40 ± 2% +4.8% 51.79 ± 2% turbostat.Cor_W > > 77.48 ± 1% +3.1% 79.91 ± 1% turbostat.Pkg_W > > > > testbox/testcase/testparams: lkp-st02/dd-write/5m-11HDD-RAID5-cfq-xfs-1dd > > > > 743cb1ff191f00f caeb178c60f4f93f1b45c0bc0 > > --------------- ------------------------- > > 18571 ± 7% +31.4% 24396 ± 4% proc-vmstat.pgscan_direct_normal > > 39983 ± 2% +38.3% 55286 ± 0% perf-stat.cpu-migrations > > 4193962 ± 2% +20.9% 5072009 ± 3% perf-stat.iTLB-load-misses > > 4.568e+09 ± 2% -17.2% 3.781e+09 ± 1% perf-stat.L1-icache-load-misses > > 1.762e+10 ± 0% -7.8% 1.625e+10 ± 1% perf-stat.cache-references > > 1.408e+09 ± 1% -6.6% 1.315e+09 ± 1% perf-stat.branch-load-misses > > 1.407e+09 ± 1% -6.5% 1.316e+09 ± 1% perf-stat.branch-misses > > 6.839e+09 ± 1% +5.0% 7.185e+09 ± 2% perf-stat.LLC-loads > > 1.558e+10 ± 0% +3.5% 1.612e+10 ± 1% perf-stat.L1-dcache-load-misses > > 1.318e+12 ± 0% +3.4% 1.363e+12 ± 0% perf-stat.L1-icache-loads > > 2.979e+10 ± 1% +2.4% 3.051e+10 ± 0% perf-stat.L1-dcache-store-misses > > 1.893e+11 ± 0% +2.5% 1.94e+11 ± 0% perf-stat.branch-instructions > > 2.298e+11 ± 0% +2.7% 2.361e+11 ± 0% perf-stat.L1-dcache-stores > > 1.016e+12 ± 0% +2.6% 1.042e+12 ± 0% perf-stat.instructions > > 1.892e+11 ± 0% +2.5% 1.94e+11 ± 0% perf-stat.branch-loads > > 3.71e+11 ± 0% +2.4% 3.799e+11 ± 0% perf-stat.dTLB-loads > > 3.711e+11 ± 0% +2.3% 3.798e+11 ± 0% perf-stat.L1-dcache-loads > > 325768 ± 0% +2.7% 334461 ± 0% vmstat.io.bo > > 8083 ± 0% +2.4% 8278 ± 0% iostat.sdf.wrqm/s > > 8083 ± 0% +2.4% 8278 ± 0% iostat.sdk.wrqm/s > > 8082 ± 0% +2.4% 8276 ± 0% iostat.sdg.wrqm/s > > 32615 ± 0% +2.4% 33398 ± 0% iostat.sdf.wkB/s > > 32617 ± 0% +2.4% 33401 ± 0% iostat.sdk.wkB/s > > 32612 ± 0% +2.4% 33393 ± 0% iostat.sdg.wkB/s > > 8083 ± 0% +2.4% 8277 ± 0% iostat.sdl.wrqm/s > > 8083 ± 0% +2.4% 8276 ± 0% iostat.sdi.wrqm/s > > 8082 ± 0% +2.4% 8277 ± 0% iostat.sdc.wrqm/s > > 32614 ± 0% +2.4% 33396 ± 0% iostat.sdl.wkB/s > > 8083 ± 0% +2.4% 8278 ± 0% iostat.sde.wrqm/s > > 8082 ± 0% +2.4% 8277 ± 0% iostat.sdh.wrqm/s > > 8083 ± 0% +2.4% 8277 ± 0% iostat.sdd.wrqm/s > > 32614 ± 0% +2.4% 33393 ± 0% iostat.sdi.wkB/s > > 32611 ± 0% +2.4% 33395 ± 0% iostat.sdc.wkB/s > > 325759 ± 0% +2.4% 333577 ± 0% iostat.md0.wkB/s > > 1274 ± 0% +2.4% 1305 ± 0% iostat.md0.w/s > > 8082 ± 0% +2.4% 8277 ± 0% iostat.sdb.wrqm/s > > 32618 ± 0% +2.4% 33398 ± 0% iostat.sde.wkB/s > > 32612 ± 0% +2.4% 33395 ± 0% iostat.sdh.wkB/s > > 32618 ± 0% +2.4% 33397 ± 0% iostat.sdd.wkB/s > > 8084 ± 0% +2.4% 8278 ± 0% iostat.sdj.wrqm/s > > 32611 ± 0% +2.4% 33396 ± 0% iostat.sdb.wkB/s > > 32618 ± 0% +2.4% 33400 ± 0% iostat.sdj.wkB/s > > 2.3e+11 ± 0% +2.5% 2.357e+11 ± 0% perf-stat.dTLB-stores > > 4898 ± 0% +2.1% 5003 ± 0% vmstat.system.cs > > 1.017e+12 ± 0% +2.4% 1.042e+12 ± 0% perf-stat.iTLB-loads > > 1518279 ± 0% +2.1% 1549457 ± 0% perf-stat.context-switches > > 1.456e+12 ± 0% +1.4% 1.476e+12 ± 0% perf-stat.cpu-cycles > > 1.456e+12 ± 0% +1.3% 1.475e+12 ± 0% perf-stat.ref-cycles > > 1.819e+11 ± 0% +1.3% 1.843e+11 ± 0% perf-stat.bus-cycles > > > > lkp-sb03 is a Sandy Bridge-EP server. > > Memory: 64G > > Architecture: x86_64 > > CPU op-mode(s): 32-bit, 64-bit > > Byte Order: Little Endian > > CPU(s): 32 > > On-line CPU(s) list: 0-31 > > Thread(s) per core: 2 > > Core(s) per socket: 8 > > Socket(s): 2 > > NUMA node(s): 2 > > Vendor ID: GenuineIntel > > CPU family: 6 > > Model: 45 > > Stepping: 6 > > CPU MHz: 3500.613 > > BogoMIPS: 5391.16 > > Virtualization: VT-x > > L1d cache: 32K > > L1i cache: 32K > > L2 cache: 256K > > L3 cache: 20480K > > NUMA node0 CPU(s): 0-7,16-23 > > NUMA node1 CPU(s): 8-15,24-31 > > > > lkp-st02 is Core2 > > Memory: 8G > > > > > > > > > > time.involuntary_context_switches > > > > 40000 O+------------------------------------------------------------------+ > > | O O O | > > 35000 ++O O O O O O | > > 30000 ++ O O O | > > | O O O | > > 25000 ++ O O | > > | | > > 20000 ++ | > > | | > > 15000 ++ | > > 10000 ++ | > > | | > > 5000 ++ | > > | .*. | > > 0 *+*--*-*-*-*--*-*-*-*--*-*-*-*--*-*-*--*-*-*-*--*---*-*--*-*-*-*--*-* > > > > > > [*] bisect-good sample > > [O] bisect-bad sample > > > > > > Disclaimer: > > Results have been estimated based on internal Intel analysis and are provided > > for informational purposes only. Any difference in system hardware or software > > design or configuration may affect actual performance. > > > > Thanks, > > Fengguang > > ^ permalink raw reply [flat|nested] 6+ messages in thread
* Re: [sched/fair] caeb178c60f: +252.0% cpuidle.C1-SNB.time, +3.1% turbostat.Pkg_W @ 2014-08-21 15:00 ` Fengguang Wu 0 siblings, 0 replies; 6+ messages in thread From: Fengguang Wu @ 2014-08-21 15:00 UTC (permalink / raw) To: lkp [-- Attachment #1: Type: text/plain, Size: 15527 bytes --] On Thu, Aug 21, 2014 at 10:16:13AM -0400, Rik van Riel wrote: > On 08/21/2014 10:01 AM, Fengguang Wu wrote: > > Hi Rik, > > > > FYI, we noticed the below changes on > > > > git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip.git sched/core > > commit caeb178c60f4f93f1b45c0bc056b5cf6d217b67f ("sched/fair: Make update_sd_pick_busiest() return 'true' on a busier sd") > > > > testbox/testcase/testparams: lkp-sb03/nepim/300s-100%-tcp6 > > Is this good or bad? It seems mixed results. The throughput is 2.4% better in sequential write test, while the power consumption (turbostat.Pkg_W) increases by 3.1% in the nepim/300s-100%-tcp test. > The numbers suggest the xfs + raid5 workload is doing around 2.4% > more IO to disk per second with this change in, and there is more Right. > CPU idle time in the system... Sorry "cpuidle" is the monitor name. You can find its code here: https://git.kernel.org/cgit/linux/kernel/git/wfg/lkp-tests.git/tree/monitors/cpuidle "cpuidle.C1-SNB.time" means the time spend in C1 state. > For the tcp test, I see no throughput numbers, but I see more > idle time as well as more time in turbo mode, and more softirqs, > which could mean that more packets were handled. Again, "turbostat" is a monitor name. "turbostat.Pkg_W" means the CPU package watts reported by the turbostat tool. > Does the patch introduce any performance issues, or did it > simply trip up something in the statistics that your script > noticed? In normal LKP reports, only changed stats are listed. Here is the performance/power index comparison, which lists all performance/power related stats. The index is geometric average of all results. Baseline is 100 for 743cb1ff191f00f. 100 perf-index (the larger, the better) 98 power-index (the larger, the better) 743cb1ff191f00f caeb178c60f4f93f1b45c0bc0 testbox/testcase/testparams --------------- ------------------------- --------------------------- %stddev %change %stddev \ | / 691053 ± 4% -5.1% 656100 ± 4% lkp-sb03/nepim/300s-100%-tcp 570185 ± 7% +5.4% 600774 ± 4% lkp-sb03/nepim/300s-100%-tcp6 1261238 ± 5% -0.3% 1256875 ± 4% TOTAL nepim.tcp.avg.kbps_in 743cb1ff191f00f caeb178c60f4f93f1b45c0bc0 --------------- ------------------------- 691216 ± 4% -5.1% 656264 ± 4% lkp-sb03/nepim/300s-100%-tcp 570347 ± 7% +5.4% 600902 ± 4% lkp-sb03/nepim/300s-100%-tcp6 1261564 ± 5% -0.3% 1257167 ± 4% TOTAL nepim.tcp.avg.kbps_out 743cb1ff191f00f caeb178c60f4f93f1b45c0bc0 --------------- ------------------------- 77.48 ± 1% +3.1% 79.91 ± 1% lkp-sb03/nepim/300s-100%-tcp 79.69 ± 2% -0.6% 79.21 ± 1% lkp-sb03/nepim/300s-100%-tcp6 157.17 ± 2% +1.2% 159.13 ± 1% TOTAL turbostat.Pkg_W 743cb1ff191f00f caeb178c60f4f93f1b45c0bc0 --------------- ------------------------- 6.05 ± 1% +1.2% 6.12 ± 1% lkp-sb03/nepim/300s-100%-tcp 6.06 ± 0% +1.0% 6.12 ± 1% lkp-sb03/nepim/300s-100%-tcp6 12.11 ± 1% +1.1% 12.24 ± 1% TOTAL turbostat.%c0 743cb1ff191f00f caeb178c60f4f93f1b45c0bc0 --------------- ------------------------- 325759 ± 0% +2.4% 333577 ± 0% lkp-st02/dd-write/5m-11HDD-RAID5-cfq-xfs-1dd 325759 ± 0% +2.4% 333577 ± 0% TOTAL iostat.md0.wkB/s The nepim throughput numbers are not stable enough comparing to the change, so are not regarded as real changes in the original email. I will need to increase its test time to make it more stable.. Thanks, Fengguang > > 743cb1ff191f00f caeb178c60f4f93f1b45c0bc0 > > --------------- ------------------------- > > 29718911 ±45% +329.5% 1.277e+08 ±10% cpuidle.C1E-SNB.time > > 861 ±34% +1590.4% 14564 ±31% cpuidle.C3-SNB.usage > > 1.65e+08 ±20% +175.4% 4.544e+08 ±15% cpuidle.C1-SNB.time > > 24 ±41% +247.6% 86 ±23% numa-numastat.node1.other_node > > 27717 ±11% +98.7% 55085 ± 6% softirqs.RCU > > 180767 ±11% +86.7% 337416 ±10% cpuidle.C7-SNB.usage > > 104591 ±14% +77.4% 185581 ±10% cpuidle.C1E-SNB.usage > > 384 ±10% +33.3% 512 ±11% slabinfo.kmem_cache.num_objs > > 384 ±10% +33.3% 512 ±11% slabinfo.kmem_cache.active_objs > > 494 ± 8% +25.9% 622 ± 9% slabinfo.kmem_cache_node.active_objs > > 512 ± 7% +25.0% 640 ± 8% slabinfo.kmem_cache_node.num_objs > > 83427 ± 6% +10.3% 92028 ± 5% meminfo.DirectMap4k > > 9508 ± 1% +21.3% 11534 ± 7% slabinfo.kmalloc-512.active_objs > > 9838 ± 1% +20.5% 11852 ± 6% slabinfo.kmalloc-512.num_objs > > 53997 ± 6% +11.1% 59981 ± 4% numa-meminfo.node1.Slab > > 2662 ± 3% -9.0% 2424 ± 3% slabinfo.kmalloc-96.active_objs > > 2710 ± 3% -8.6% 2478 ± 3% slabinfo.kmalloc-96.num_objs > > 921 ±41% +3577.7% 33901 ±14% time.involuntary_context_switches > > 2371 ± 2% +15.5% 2739 ± 2% vmstat.system.in > > > > testbox/testcase/testparams: lkp-sb03/nepim/300s-100%-tcp > > > > 743cb1ff191f00f caeb178c60f4f93f1b45c0bc0 > > --------------- ------------------------- > > 20657207 ±31% +358.2% 94650352 ±18% cpuidle.C1E-SNB.time > > 29718911 ±45% +329.5% 1.277e+08 ±10% cpuidle.C1E-SNB.time > > 861 ±34% +1590.4% 14564 ±31% cpuidle.C3-SNB.usage > > 0.05 ±46% +812.5% 0.44 ±34% turbostat.%c3 > > 1.12e+08 ±25% +364.8% 5.207e+08 ±15% cpuidle.C1-SNB.time > > 1.65e+08 ±20% +175.4% 4.544e+08 ±15% cpuidle.C1-SNB.time > > 35 ±19% +105.6% 72 ±28% numa-numastat.node1.other_node > > 24 ±41% +247.6% 86 ±23% numa-numastat.node1.other_node > > 43 ±22% +86.2% 80 ±26% numa-vmstat.node0.nr_dirtied > > 24576 ± 6% +113.9% 52574 ± 1% softirqs.RCU > > 27717 ±11% +98.7% 55085 ± 6% softirqs.RCU > > 211533 ± 6% +58.4% 334990 ± 8% cpuidle.C7-SNB.usage > > 180767 ±11% +86.7% 337416 ±10% cpuidle.C7-SNB.usage > > 77739 ±13% +52.9% 118876 ±18% cpuidle.C1E-SNB.usage > > 104591 ±14% +77.4% 185581 ±10% cpuidle.C1E-SNB.usage > > 32.09 ±14% -24.8% 24.12 ±18% turbostat.%pc2 > > 9.04 ± 6% +41.6% 12.80 ± 6% turbostat.%c1 > > 384 ±10% +33.3% 512 ±11% slabinfo.kmem_cache.num_objs > > 384 ±10% +33.3% 512 ±11% slabinfo.kmem_cache.active_objs > > 494 ± 8% +25.9% 622 ± 9% slabinfo.kmem_cache_node.active_objs > > 512 ± 7% +25.0% 640 ± 8% slabinfo.kmem_cache_node.num_objs > > 379 ± 9% +16.7% 443 ± 7% numa-vmstat.node0.nr_page_table_pages > > 83427 ± 6% +10.3% 92028 ± 5% meminfo.DirectMap4k > > 1579 ± 6% -15.3% 1338 ± 7% numa-meminfo.node1.PageTables > > 394 ± 6% -15.1% 334 ± 7% numa-vmstat.node1.nr_page_table_pages > > 1509 ± 7% +16.6% 1760 ± 7% numa-meminfo.node0.PageTables > > 12681 ± 1% -17.3% 10482 ±14% numa-meminfo.node1.AnonPages > > 3169 ± 1% -17.3% 2620 ±14% numa-vmstat.node1.nr_anon_pages > > 10171 ± 3% +10.9% 11283 ± 3% slabinfo.kmalloc-512.active_objs > > 9508 ± 1% +21.3% 11534 ± 7% slabinfo.kmalloc-512.active_objs > > 10481 ± 3% +10.9% 11620 ± 3% slabinfo.kmalloc-512.num_objs > > 9838 ± 1% +20.5% 11852 ± 6% slabinfo.kmalloc-512.num_objs > > 53997 ± 6% +11.1% 59981 ± 4% numa-meminfo.node1.Slab > > 5072 ± 1% +11.6% 5662 ± 3% slabinfo.kmalloc-2048.num_objs > > 4974 ± 1% +11.6% 5551 ± 3% slabinfo.kmalloc-2048.active_objs > > 12824 ± 2% -16.1% 10754 ±14% numa-meminfo.node1.Active(anon) > > 3205 ± 2% -16.2% 2687 ±14% numa-vmstat.node1.nr_active_anon > > 2662 ± 3% -9.0% 2424 ± 3% slabinfo.kmalloc-96.active_objs > > 2710 ± 3% -8.6% 2478 ± 3% slabinfo.kmalloc-96.num_objs > > 15791 ± 1% +15.2% 18192 ± 9% numa-meminfo.node0.AnonPages > > 3949 ± 1% +15.2% 4549 ± 9% numa-vmstat.node0.nr_anon_pages > > 13669 ± 1% -7.5% 12645 ± 2% slabinfo.kmalloc-16.num_objs > > 662 ±23% +4718.6% 31918 ±12% time.involuntary_context_switches > > 921 ±41% +3577.7% 33901 ±14% time.involuntary_context_switches > > 2463 ± 1% +13.1% 2786 ± 3% vmstat.system.in > > 2371 ± 2% +15.5% 2739 ± 2% vmstat.system.in > > 49.40 ± 2% +4.8% 51.79 ± 2% turbostat.Cor_W > > 77.48 ± 1% +3.1% 79.91 ± 1% turbostat.Pkg_W > > > > testbox/testcase/testparams: lkp-st02/dd-write/5m-11HDD-RAID5-cfq-xfs-1dd > > > > 743cb1ff191f00f caeb178c60f4f93f1b45c0bc0 > > --------------- ------------------------- > > 18571 ± 7% +31.4% 24396 ± 4% proc-vmstat.pgscan_direct_normal > > 39983 ± 2% +38.3% 55286 ± 0% perf-stat.cpu-migrations > > 4193962 ± 2% +20.9% 5072009 ± 3% perf-stat.iTLB-load-misses > > 4.568e+09 ± 2% -17.2% 3.781e+09 ± 1% perf-stat.L1-icache-load-misses > > 1.762e+10 ± 0% -7.8% 1.625e+10 ± 1% perf-stat.cache-references > > 1.408e+09 ± 1% -6.6% 1.315e+09 ± 1% perf-stat.branch-load-misses > > 1.407e+09 ± 1% -6.5% 1.316e+09 ± 1% perf-stat.branch-misses > > 6.839e+09 ± 1% +5.0% 7.185e+09 ± 2% perf-stat.LLC-loads > > 1.558e+10 ± 0% +3.5% 1.612e+10 ± 1% perf-stat.L1-dcache-load-misses > > 1.318e+12 ± 0% +3.4% 1.363e+12 ± 0% perf-stat.L1-icache-loads > > 2.979e+10 ± 1% +2.4% 3.051e+10 ± 0% perf-stat.L1-dcache-store-misses > > 1.893e+11 ± 0% +2.5% 1.94e+11 ± 0% perf-stat.branch-instructions > > 2.298e+11 ± 0% +2.7% 2.361e+11 ± 0% perf-stat.L1-dcache-stores > > 1.016e+12 ± 0% +2.6% 1.042e+12 ± 0% perf-stat.instructions > > 1.892e+11 ± 0% +2.5% 1.94e+11 ± 0% perf-stat.branch-loads > > 3.71e+11 ± 0% +2.4% 3.799e+11 ± 0% perf-stat.dTLB-loads > > 3.711e+11 ± 0% +2.3% 3.798e+11 ± 0% perf-stat.L1-dcache-loads > > 325768 ± 0% +2.7% 334461 ± 0% vmstat.io.bo > > 8083 ± 0% +2.4% 8278 ± 0% iostat.sdf.wrqm/s > > 8083 ± 0% +2.4% 8278 ± 0% iostat.sdk.wrqm/s > > 8082 ± 0% +2.4% 8276 ± 0% iostat.sdg.wrqm/s > > 32615 ± 0% +2.4% 33398 ± 0% iostat.sdf.wkB/s > > 32617 ± 0% +2.4% 33401 ± 0% iostat.sdk.wkB/s > > 32612 ± 0% +2.4% 33393 ± 0% iostat.sdg.wkB/s > > 8083 ± 0% +2.4% 8277 ± 0% iostat.sdl.wrqm/s > > 8083 ± 0% +2.4% 8276 ± 0% iostat.sdi.wrqm/s > > 8082 ± 0% +2.4% 8277 ± 0% iostat.sdc.wrqm/s > > 32614 ± 0% +2.4% 33396 ± 0% iostat.sdl.wkB/s > > 8083 ± 0% +2.4% 8278 ± 0% iostat.sde.wrqm/s > > 8082 ± 0% +2.4% 8277 ± 0% iostat.sdh.wrqm/s > > 8083 ± 0% +2.4% 8277 ± 0% iostat.sdd.wrqm/s > > 32614 ± 0% +2.4% 33393 ± 0% iostat.sdi.wkB/s > > 32611 ± 0% +2.4% 33395 ± 0% iostat.sdc.wkB/s > > 325759 ± 0% +2.4% 333577 ± 0% iostat.md0.wkB/s > > 1274 ± 0% +2.4% 1305 ± 0% iostat.md0.w/s > > 8082 ± 0% +2.4% 8277 ± 0% iostat.sdb.wrqm/s > > 32618 ± 0% +2.4% 33398 ± 0% iostat.sde.wkB/s > > 32612 ± 0% +2.4% 33395 ± 0% iostat.sdh.wkB/s > > 32618 ± 0% +2.4% 33397 ± 0% iostat.sdd.wkB/s > > 8084 ± 0% +2.4% 8278 ± 0% iostat.sdj.wrqm/s > > 32611 ± 0% +2.4% 33396 ± 0% iostat.sdb.wkB/s > > 32618 ± 0% +2.4% 33400 ± 0% iostat.sdj.wkB/s > > 2.3e+11 ± 0% +2.5% 2.357e+11 ± 0% perf-stat.dTLB-stores > > 4898 ± 0% +2.1% 5003 ± 0% vmstat.system.cs > > 1.017e+12 ± 0% +2.4% 1.042e+12 ± 0% perf-stat.iTLB-loads > > 1518279 ± 0% +2.1% 1549457 ± 0% perf-stat.context-switches > > 1.456e+12 ± 0% +1.4% 1.476e+12 ± 0% perf-stat.cpu-cycles > > 1.456e+12 ± 0% +1.3% 1.475e+12 ± 0% perf-stat.ref-cycles > > 1.819e+11 ± 0% +1.3% 1.843e+11 ± 0% perf-stat.bus-cycles > > > > lkp-sb03 is a Sandy Bridge-EP server. > > Memory: 64G > > Architecture: x86_64 > > CPU op-mode(s): 32-bit, 64-bit > > Byte Order: Little Endian > > CPU(s): 32 > > On-line CPU(s) list: 0-31 > > Thread(s) per core: 2 > > Core(s) per socket: 8 > > Socket(s): 2 > > NUMA node(s): 2 > > Vendor ID: GenuineIntel > > CPU family: 6 > > Model: 45 > > Stepping: 6 > > CPU MHz: 3500.613 > > BogoMIPS: 5391.16 > > Virtualization: VT-x > > L1d cache: 32K > > L1i cache: 32K > > L2 cache: 256K > > L3 cache: 20480K > > NUMA node0 CPU(s): 0-7,16-23 > > NUMA node1 CPU(s): 8-15,24-31 > > > > lkp-st02 is Core2 > > Memory: 8G > > > > > > > > > > time.involuntary_context_switches > > > > 40000 O+------------------------------------------------------------------+ > > | O O O | > > 35000 ++O O O O O O | > > 30000 ++ O O O | > > | O O O | > > 25000 ++ O O | > > | | > > 20000 ++ | > > | | > > 15000 ++ | > > 10000 ++ | > > | | > > 5000 ++ | > > | .*. | > > 0 *+*--*-*-*-*--*-*-*-*--*-*-*-*--*-*-*--*-*-*-*--*---*-*--*-*-*-*--*-* > > > > > > [*] bisect-good sample > > [O] bisect-bad sample > > > > > > Disclaimer: > > Results have been estimated based on internal Intel analysis and are provided > > for informational purposes only. Any difference in system hardware or software > > design or configuration may affect actual performance. > > > > Thanks, > > Fengguang > > ^ permalink raw reply [flat|nested] 6+ messages in thread
end of thread, other threads:[~2014-08-21 15:01 UTC | newest] Thread overview: 6+ messages (download: mbox.gz / follow: Atom feed) -- links below jump to the message on this page -- 2014-08-21 14:01 [sched/fair] caeb178c60f: +252.0% cpuidle.C1-SNB.time, +3.1% turbostat.Pkg_W Fengguang Wu 2014-08-21 14:01 ` Fengguang Wu 2014-08-21 14:16 ` Rik van Riel 2014-08-21 14:16 ` Rik van Riel 2014-08-21 15:00 ` Fengguang Wu 2014-08-21 15:00 ` Fengguang Wu
This is an external index of several public inboxes, see mirroring instructions on how to clone and mirror all data and code used by this external index.