All of lore.kernel.org
 help / color / mirror / Atom feed
From: Fengguang Wu <fengguang.wu@intel.com>
To: Rik van Riel <riel@redhat.com>
Cc: Dave Hansen <dave.hansen@intel.com>,
	LKML <linux-kernel@vger.kernel.org>,
	lkp@01.org, Ingo Molnar <mingo@elte.hu>
Subject: [sched/fair] caeb178c60f: +252.0% cpuidle.C1-SNB.time, +3.1% turbostat.Pkg_W
Date: Thu, 21 Aug 2014 22:01:34 +0800	[thread overview]
Message-ID: <20140821140134.GB19246@localhost> (raw)

[-- Attachment #1: Type: text/plain, Size: 11083 bytes --]

Hi Rik,

FYI, we noticed the below changes on

git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip.git sched/core
commit caeb178c60f4f93f1b45c0bc056b5cf6d217b67f ("sched/fair: Make update_sd_pick_busiest() return 'true' on a busier sd")

testbox/testcase/testparams: lkp-sb03/nepim/300s-100%-tcp6

743cb1ff191f00f  caeb178c60f4f93f1b45c0bc0
---------------  -------------------------
  29718911 ±45%    +329.5%  1.277e+08 ±10%  cpuidle.C1E-SNB.time
       861 ±34%   +1590.4%      14564 ±31%  cpuidle.C3-SNB.usage
  1.65e+08 ±20%    +175.4%  4.544e+08 ±15%  cpuidle.C1-SNB.time
        24 ±41%    +247.6%         86 ±23%  numa-numastat.node1.other_node
     27717 ±11%     +98.7%      55085 ± 6%  softirqs.RCU
    180767 ±11%     +86.7%     337416 ±10%  cpuidle.C7-SNB.usage
    104591 ±14%     +77.4%     185581 ±10%  cpuidle.C1E-SNB.usage
       384 ±10%     +33.3%        512 ±11%  slabinfo.kmem_cache.num_objs
       384 ±10%     +33.3%        512 ±11%  slabinfo.kmem_cache.active_objs
       494 ± 8%     +25.9%        622 ± 9%  slabinfo.kmem_cache_node.active_objs
       512 ± 7%     +25.0%        640 ± 8%  slabinfo.kmem_cache_node.num_objs
     83427 ± 6%     +10.3%      92028 ± 5%  meminfo.DirectMap4k
      9508 ± 1%     +21.3%      11534 ± 7%  slabinfo.kmalloc-512.active_objs
      9838 ± 1%     +20.5%      11852 ± 6%  slabinfo.kmalloc-512.num_objs
     53997 ± 6%     +11.1%      59981 ± 4%  numa-meminfo.node1.Slab
      2662 ± 3%      -9.0%       2424 ± 3%  slabinfo.kmalloc-96.active_objs
      2710 ± 3%      -8.6%       2478 ± 3%  slabinfo.kmalloc-96.num_objs
       921 ±41%   +3577.7%      33901 ±14%  time.involuntary_context_switches
      2371 ± 2%     +15.5%       2739 ± 2%  vmstat.system.in

testbox/testcase/testparams: lkp-sb03/nepim/300s-100%-tcp

743cb1ff191f00f  caeb178c60f4f93f1b45c0bc0
---------------  -------------------------
  20657207 ±31%    +358.2%   94650352 ±18%  cpuidle.C1E-SNB.time
  29718911 ±45%    +329.5%  1.277e+08 ±10%  cpuidle.C1E-SNB.time
       861 ±34%   +1590.4%      14564 ±31%  cpuidle.C3-SNB.usage
      0.05 ±46%    +812.5%       0.44 ±34%  turbostat.%c3
  1.12e+08 ±25%    +364.8%  5.207e+08 ±15%  cpuidle.C1-SNB.time
  1.65e+08 ±20%    +175.4%  4.544e+08 ±15%  cpuidle.C1-SNB.time
        35 ±19%    +105.6%         72 ±28%  numa-numastat.node1.other_node
        24 ±41%    +247.6%         86 ±23%  numa-numastat.node1.other_node
        43 ±22%     +86.2%         80 ±26%  numa-vmstat.node0.nr_dirtied
     24576 ± 6%    +113.9%      52574 ± 1%  softirqs.RCU
     27717 ±11%     +98.7%      55085 ± 6%  softirqs.RCU
    211533 ± 6%     +58.4%     334990 ± 8%  cpuidle.C7-SNB.usage
    180767 ±11%     +86.7%     337416 ±10%  cpuidle.C7-SNB.usage
     77739 ±13%     +52.9%     118876 ±18%  cpuidle.C1E-SNB.usage
    104591 ±14%     +77.4%     185581 ±10%  cpuidle.C1E-SNB.usage
     32.09 ±14%     -24.8%      24.12 ±18%  turbostat.%pc2
      9.04 ± 6%     +41.6%      12.80 ± 6%  turbostat.%c1
       384 ±10%     +33.3%        512 ±11%  slabinfo.kmem_cache.num_objs
       384 ±10%     +33.3%        512 ±11%  slabinfo.kmem_cache.active_objs
       494 ± 8%     +25.9%        622 ± 9%  slabinfo.kmem_cache_node.active_objs
       512 ± 7%     +25.0%        640 ± 8%  slabinfo.kmem_cache_node.num_objs
       379 ± 9%     +16.7%        443 ± 7%  numa-vmstat.node0.nr_page_table_pages
     83427 ± 6%     +10.3%      92028 ± 5%  meminfo.DirectMap4k
      1579 ± 6%     -15.3%       1338 ± 7%  numa-meminfo.node1.PageTables
       394 ± 6%     -15.1%        334 ± 7%  numa-vmstat.node1.nr_page_table_pages
      1509 ± 7%     +16.6%       1760 ± 7%  numa-meminfo.node0.PageTables
     12681 ± 1%     -17.3%      10482 ±14%  numa-meminfo.node1.AnonPages
      3169 ± 1%     -17.3%       2620 ±14%  numa-vmstat.node1.nr_anon_pages
     10171 ± 3%     +10.9%      11283 ± 3%  slabinfo.kmalloc-512.active_objs
      9508 ± 1%     +21.3%      11534 ± 7%  slabinfo.kmalloc-512.active_objs
     10481 ± 3%     +10.9%      11620 ± 3%  slabinfo.kmalloc-512.num_objs
      9838 ± 1%     +20.5%      11852 ± 6%  slabinfo.kmalloc-512.num_objs
     53997 ± 6%     +11.1%      59981 ± 4%  numa-meminfo.node1.Slab
      5072 ± 1%     +11.6%       5662 ± 3%  slabinfo.kmalloc-2048.num_objs
      4974 ± 1%     +11.6%       5551 ± 3%  slabinfo.kmalloc-2048.active_objs
     12824 ± 2%     -16.1%      10754 ±14%  numa-meminfo.node1.Active(anon)
      3205 ± 2%     -16.2%       2687 ±14%  numa-vmstat.node1.nr_active_anon
      2662 ± 3%      -9.0%       2424 ± 3%  slabinfo.kmalloc-96.active_objs
      2710 ± 3%      -8.6%       2478 ± 3%  slabinfo.kmalloc-96.num_objs
     15791 ± 1%     +15.2%      18192 ± 9%  numa-meminfo.node0.AnonPages
      3949 ± 1%     +15.2%       4549 ± 9%  numa-vmstat.node0.nr_anon_pages
     13669 ± 1%      -7.5%      12645 ± 2%  slabinfo.kmalloc-16.num_objs
       662 ±23%   +4718.6%      31918 ±12%  time.involuntary_context_switches
       921 ±41%   +3577.7%      33901 ±14%  time.involuntary_context_switches
      2463 ± 1%     +13.1%       2786 ± 3%  vmstat.system.in
      2371 ± 2%     +15.5%       2739 ± 2%  vmstat.system.in
     49.40 ± 2%      +4.8%      51.79 ± 2%  turbostat.Cor_W
     77.48 ± 1%      +3.1%      79.91 ± 1%  turbostat.Pkg_W

testbox/testcase/testparams: lkp-st02/dd-write/5m-11HDD-RAID5-cfq-xfs-1dd

743cb1ff191f00f  caeb178c60f4f93f1b45c0bc0
---------------  -------------------------
     18571 ± 7%     +31.4%      24396 ± 4%  proc-vmstat.pgscan_direct_normal
     39983 ± 2%     +38.3%      55286 ± 0%  perf-stat.cpu-migrations
   4193962 ± 2%     +20.9%    5072009 ± 3%  perf-stat.iTLB-load-misses
 4.568e+09 ± 2%     -17.2%  3.781e+09 ± 1%  perf-stat.L1-icache-load-misses
 1.762e+10 ± 0%      -7.8%  1.625e+10 ± 1%  perf-stat.cache-references
 1.408e+09 ± 1%      -6.6%  1.315e+09 ± 1%  perf-stat.branch-load-misses
 1.407e+09 ± 1%      -6.5%  1.316e+09 ± 1%  perf-stat.branch-misses
 6.839e+09 ± 1%      +5.0%  7.185e+09 ± 2%  perf-stat.LLC-loads
 1.558e+10 ± 0%      +3.5%  1.612e+10 ± 1%  perf-stat.L1-dcache-load-misses
 1.318e+12 ± 0%      +3.4%  1.363e+12 ± 0%  perf-stat.L1-icache-loads
 2.979e+10 ± 1%      +2.4%  3.051e+10 ± 0%  perf-stat.L1-dcache-store-misses
 1.893e+11 ± 0%      +2.5%   1.94e+11 ± 0%  perf-stat.branch-instructions
 2.298e+11 ± 0%      +2.7%  2.361e+11 ± 0%  perf-stat.L1-dcache-stores
 1.016e+12 ± 0%      +2.6%  1.042e+12 ± 0%  perf-stat.instructions
 1.892e+11 ± 0%      +2.5%   1.94e+11 ± 0%  perf-stat.branch-loads
  3.71e+11 ± 0%      +2.4%  3.799e+11 ± 0%  perf-stat.dTLB-loads
 3.711e+11 ± 0%      +2.3%  3.798e+11 ± 0%  perf-stat.L1-dcache-loads
    325768 ± 0%      +2.7%     334461 ± 0%  vmstat.io.bo
      8083 ± 0%      +2.4%       8278 ± 0%  iostat.sdf.wrqm/s
      8083 ± 0%      +2.4%       8278 ± 0%  iostat.sdk.wrqm/s
      8082 ± 0%      +2.4%       8276 ± 0%  iostat.sdg.wrqm/s
     32615 ± 0%      +2.4%      33398 ± 0%  iostat.sdf.wkB/s
     32617 ± 0%      +2.4%      33401 ± 0%  iostat.sdk.wkB/s
     32612 ± 0%      +2.4%      33393 ± 0%  iostat.sdg.wkB/s
      8083 ± 0%      +2.4%       8277 ± 0%  iostat.sdl.wrqm/s
      8083 ± 0%      +2.4%       8276 ± 0%  iostat.sdi.wrqm/s
      8082 ± 0%      +2.4%       8277 ± 0%  iostat.sdc.wrqm/s
     32614 ± 0%      +2.4%      33396 ± 0%  iostat.sdl.wkB/s
      8083 ± 0%      +2.4%       8278 ± 0%  iostat.sde.wrqm/s
      8082 ± 0%      +2.4%       8277 ± 0%  iostat.sdh.wrqm/s
      8083 ± 0%      +2.4%       8277 ± 0%  iostat.sdd.wrqm/s
     32614 ± 0%      +2.4%      33393 ± 0%  iostat.sdi.wkB/s
     32611 ± 0%      +2.4%      33395 ± 0%  iostat.sdc.wkB/s
    325759 ± 0%      +2.4%     333577 ± 0%  iostat.md0.wkB/s
      1274 ± 0%      +2.4%       1305 ± 0%  iostat.md0.w/s
      8082 ± 0%      +2.4%       8277 ± 0%  iostat.sdb.wrqm/s
     32618 ± 0%      +2.4%      33398 ± 0%  iostat.sde.wkB/s
     32612 ± 0%      +2.4%      33395 ± 0%  iostat.sdh.wkB/s
     32618 ± 0%      +2.4%      33397 ± 0%  iostat.sdd.wkB/s
      8084 ± 0%      +2.4%       8278 ± 0%  iostat.sdj.wrqm/s
     32611 ± 0%      +2.4%      33396 ± 0%  iostat.sdb.wkB/s
     32618 ± 0%      +2.4%      33400 ± 0%  iostat.sdj.wkB/s
   2.3e+11 ± 0%      +2.5%  2.357e+11 ± 0%  perf-stat.dTLB-stores
      4898 ± 0%      +2.1%       5003 ± 0%  vmstat.system.cs
 1.017e+12 ± 0%      +2.4%  1.042e+12 ± 0%  perf-stat.iTLB-loads
   1518279 ± 0%      +2.1%    1549457 ± 0%  perf-stat.context-switches
 1.456e+12 ± 0%      +1.4%  1.476e+12 ± 0%  perf-stat.cpu-cycles
 1.456e+12 ± 0%      +1.3%  1.475e+12 ± 0%  perf-stat.ref-cycles
 1.819e+11 ± 0%      +1.3%  1.843e+11 ± 0%  perf-stat.bus-cycles

lkp-sb03 is a Sandy Bridge-EP server.
Memory: 64G
Architecture:          x86_64
CPU op-mode(s):        32-bit, 64-bit
Byte Order:            Little Endian
CPU(s):                32
On-line CPU(s) list:   0-31
Thread(s) per core:    2
Core(s) per socket:    8
Socket(s):             2
NUMA node(s):          2
Vendor ID:             GenuineIntel
CPU family:            6
Model:                 45
Stepping:              6
CPU MHz:               3500.613
BogoMIPS:              5391.16
Virtualization:        VT-x
L1d cache:             32K
L1i cache:             32K
L2 cache:              256K
L3 cache:              20480K
NUMA node0 CPU(s):     0-7,16-23
NUMA node1 CPU(s):     8-15,24-31

lkp-st02 is Core2
Memory: 8G




                          time.involuntary_context_switches

  40000 O+------------------------------------------------------------------+
        |        O               O          O                               |
  35000 ++O  O O        O   O             O                                 |
  30000 ++         O                 O         O                            |
        |                 O    O   O                                        |
  25000 ++            O                 O                                   |
        |                                                                   |
  20000 ++                                                                  |
        |                                                                   |
  15000 ++                                                                  |
  10000 ++                                                                  |
        |                                                                   |
   5000 ++                                                                  |
        |                                                .*.                |
      0 *+*--*-*-*-*--*-*-*-*--*-*-*-*--*-*-*--*-*-*-*--*---*-*--*-*-*-*--*-*


	[*] bisect-good sample
	[O] bisect-bad  sample


Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.

Thanks,
Fengguang

[-- Attachment #2: reproduce --]
[-- Type: text/plain, Size: 2403 bytes --]

echo performance > /sys/devices/system/cpu/cpu0/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu1/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu10/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu11/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu12/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu13/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu14/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu15/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu16/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu17/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu18/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu19/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu2/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu20/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu21/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu22/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu23/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu24/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu25/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu26/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu27/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu28/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu29/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu3/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu30/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu31/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu4/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu5/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu6/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu7/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu8/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu9/cpufreq/scaling_governor
./nepim
./nepim -c 127.0.0.1 -d -a 300 -n 32

WARNING: multiple messages have this Message-ID (diff)
From: Fengguang Wu <fengguang.wu@intel.com>
To: lkp@lists.01.org
Subject: [sched/fair] caeb178c60f: +252.0% cpuidle.C1-SNB.time, +3.1% turbostat.Pkg_W
Date: Thu, 21 Aug 2014 22:01:34 +0800	[thread overview]
Message-ID: <20140821140134.GB19246@localhost> (raw)

[-- Attachment #1: Type: text/plain, Size: 11282 bytes --]

Hi Rik,

FYI, we noticed the below changes on

git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip.git sched/core
commit caeb178c60f4f93f1b45c0bc056b5cf6d217b67f ("sched/fair: Make update_sd_pick_busiest() return 'true' on a busier sd")

testbox/testcase/testparams: lkp-sb03/nepim/300s-100%-tcp6

743cb1ff191f00f  caeb178c60f4f93f1b45c0bc0
---------------  -------------------------
  29718911 ±45%    +329.5%  1.277e+08 ±10%  cpuidle.C1E-SNB.time
       861 ±34%   +1590.4%      14564 ±31%  cpuidle.C3-SNB.usage
  1.65e+08 ±20%    +175.4%  4.544e+08 ±15%  cpuidle.C1-SNB.time
        24 ±41%    +247.6%         86 ±23%  numa-numastat.node1.other_node
     27717 ±11%     +98.7%      55085 ± 6%  softirqs.RCU
    180767 ±11%     +86.7%     337416 ±10%  cpuidle.C7-SNB.usage
    104591 ±14%     +77.4%     185581 ±10%  cpuidle.C1E-SNB.usage
       384 ±10%     +33.3%        512 ±11%  slabinfo.kmem_cache.num_objs
       384 ±10%     +33.3%        512 ±11%  slabinfo.kmem_cache.active_objs
       494 ± 8%     +25.9%        622 ± 9%  slabinfo.kmem_cache_node.active_objs
       512 ± 7%     +25.0%        640 ± 8%  slabinfo.kmem_cache_node.num_objs
     83427 ± 6%     +10.3%      92028 ± 5%  meminfo.DirectMap4k
      9508 ± 1%     +21.3%      11534 ± 7%  slabinfo.kmalloc-512.active_objs
      9838 ± 1%     +20.5%      11852 ± 6%  slabinfo.kmalloc-512.num_objs
     53997 ± 6%     +11.1%      59981 ± 4%  numa-meminfo.node1.Slab
      2662 ± 3%      -9.0%       2424 ± 3%  slabinfo.kmalloc-96.active_objs
      2710 ± 3%      -8.6%       2478 ± 3%  slabinfo.kmalloc-96.num_objs
       921 ±41%   +3577.7%      33901 ±14%  time.involuntary_context_switches
      2371 ± 2%     +15.5%       2739 ± 2%  vmstat.system.in

testbox/testcase/testparams: lkp-sb03/nepim/300s-100%-tcp

743cb1ff191f00f  caeb178c60f4f93f1b45c0bc0
---------------  -------------------------
  20657207 ±31%    +358.2%   94650352 ±18%  cpuidle.C1E-SNB.time
  29718911 ±45%    +329.5%  1.277e+08 ±10%  cpuidle.C1E-SNB.time
       861 ±34%   +1590.4%      14564 ±31%  cpuidle.C3-SNB.usage
      0.05 ±46%    +812.5%       0.44 ±34%  turbostat.%c3
  1.12e+08 ±25%    +364.8%  5.207e+08 ±15%  cpuidle.C1-SNB.time
  1.65e+08 ±20%    +175.4%  4.544e+08 ±15%  cpuidle.C1-SNB.time
        35 ±19%    +105.6%         72 ±28%  numa-numastat.node1.other_node
        24 ±41%    +247.6%         86 ±23%  numa-numastat.node1.other_node
        43 ±22%     +86.2%         80 ±26%  numa-vmstat.node0.nr_dirtied
     24576 ± 6%    +113.9%      52574 ± 1%  softirqs.RCU
     27717 ±11%     +98.7%      55085 ± 6%  softirqs.RCU
    211533 ± 6%     +58.4%     334990 ± 8%  cpuidle.C7-SNB.usage
    180767 ±11%     +86.7%     337416 ±10%  cpuidle.C7-SNB.usage
     77739 ±13%     +52.9%     118876 ±18%  cpuidle.C1E-SNB.usage
    104591 ±14%     +77.4%     185581 ±10%  cpuidle.C1E-SNB.usage
     32.09 ±14%     -24.8%      24.12 ±18%  turbostat.%pc2
      9.04 ± 6%     +41.6%      12.80 ± 6%  turbostat.%c1
       384 ±10%     +33.3%        512 ±11%  slabinfo.kmem_cache.num_objs
       384 ±10%     +33.3%        512 ±11%  slabinfo.kmem_cache.active_objs
       494 ± 8%     +25.9%        622 ± 9%  slabinfo.kmem_cache_node.active_objs
       512 ± 7%     +25.0%        640 ± 8%  slabinfo.kmem_cache_node.num_objs
       379 ± 9%     +16.7%        443 ± 7%  numa-vmstat.node0.nr_page_table_pages
     83427 ± 6%     +10.3%      92028 ± 5%  meminfo.DirectMap4k
      1579 ± 6%     -15.3%       1338 ± 7%  numa-meminfo.node1.PageTables
       394 ± 6%     -15.1%        334 ± 7%  numa-vmstat.node1.nr_page_table_pages
      1509 ± 7%     +16.6%       1760 ± 7%  numa-meminfo.node0.PageTables
     12681 ± 1%     -17.3%      10482 ±14%  numa-meminfo.node1.AnonPages
      3169 ± 1%     -17.3%       2620 ±14%  numa-vmstat.node1.nr_anon_pages
     10171 ± 3%     +10.9%      11283 ± 3%  slabinfo.kmalloc-512.active_objs
      9508 ± 1%     +21.3%      11534 ± 7%  slabinfo.kmalloc-512.active_objs
     10481 ± 3%     +10.9%      11620 ± 3%  slabinfo.kmalloc-512.num_objs
      9838 ± 1%     +20.5%      11852 ± 6%  slabinfo.kmalloc-512.num_objs
     53997 ± 6%     +11.1%      59981 ± 4%  numa-meminfo.node1.Slab
      5072 ± 1%     +11.6%       5662 ± 3%  slabinfo.kmalloc-2048.num_objs
      4974 ± 1%     +11.6%       5551 ± 3%  slabinfo.kmalloc-2048.active_objs
     12824 ± 2%     -16.1%      10754 ±14%  numa-meminfo.node1.Active(anon)
      3205 ± 2%     -16.2%       2687 ±14%  numa-vmstat.node1.nr_active_anon
      2662 ± 3%      -9.0%       2424 ± 3%  slabinfo.kmalloc-96.active_objs
      2710 ± 3%      -8.6%       2478 ± 3%  slabinfo.kmalloc-96.num_objs
     15791 ± 1%     +15.2%      18192 ± 9%  numa-meminfo.node0.AnonPages
      3949 ± 1%     +15.2%       4549 ± 9%  numa-vmstat.node0.nr_anon_pages
     13669 ± 1%      -7.5%      12645 ± 2%  slabinfo.kmalloc-16.num_objs
       662 ±23%   +4718.6%      31918 ±12%  time.involuntary_context_switches
       921 ±41%   +3577.7%      33901 ±14%  time.involuntary_context_switches
      2463 ± 1%     +13.1%       2786 ± 3%  vmstat.system.in
      2371 ± 2%     +15.5%       2739 ± 2%  vmstat.system.in
     49.40 ± 2%      +4.8%      51.79 ± 2%  turbostat.Cor_W
     77.48 ± 1%      +3.1%      79.91 ± 1%  turbostat.Pkg_W

testbox/testcase/testparams: lkp-st02/dd-write/5m-11HDD-RAID5-cfq-xfs-1dd

743cb1ff191f00f  caeb178c60f4f93f1b45c0bc0
---------------  -------------------------
     18571 ± 7%     +31.4%      24396 ± 4%  proc-vmstat.pgscan_direct_normal
     39983 ± 2%     +38.3%      55286 ± 0%  perf-stat.cpu-migrations
   4193962 ± 2%     +20.9%    5072009 ± 3%  perf-stat.iTLB-load-misses
 4.568e+09 ± 2%     -17.2%  3.781e+09 ± 1%  perf-stat.L1-icache-load-misses
 1.762e+10 ± 0%      -7.8%  1.625e+10 ± 1%  perf-stat.cache-references
 1.408e+09 ± 1%      -6.6%  1.315e+09 ± 1%  perf-stat.branch-load-misses
 1.407e+09 ± 1%      -6.5%  1.316e+09 ± 1%  perf-stat.branch-misses
 6.839e+09 ± 1%      +5.0%  7.185e+09 ± 2%  perf-stat.LLC-loads
 1.558e+10 ± 0%      +3.5%  1.612e+10 ± 1%  perf-stat.L1-dcache-load-misses
 1.318e+12 ± 0%      +3.4%  1.363e+12 ± 0%  perf-stat.L1-icache-loads
 2.979e+10 ± 1%      +2.4%  3.051e+10 ± 0%  perf-stat.L1-dcache-store-misses
 1.893e+11 ± 0%      +2.5%   1.94e+11 ± 0%  perf-stat.branch-instructions
 2.298e+11 ± 0%      +2.7%  2.361e+11 ± 0%  perf-stat.L1-dcache-stores
 1.016e+12 ± 0%      +2.6%  1.042e+12 ± 0%  perf-stat.instructions
 1.892e+11 ± 0%      +2.5%   1.94e+11 ± 0%  perf-stat.branch-loads
  3.71e+11 ± 0%      +2.4%  3.799e+11 ± 0%  perf-stat.dTLB-loads
 3.711e+11 ± 0%      +2.3%  3.798e+11 ± 0%  perf-stat.L1-dcache-loads
    325768 ± 0%      +2.7%     334461 ± 0%  vmstat.io.bo
      8083 ± 0%      +2.4%       8278 ± 0%  iostat.sdf.wrqm/s
      8083 ± 0%      +2.4%       8278 ± 0%  iostat.sdk.wrqm/s
      8082 ± 0%      +2.4%       8276 ± 0%  iostat.sdg.wrqm/s
     32615 ± 0%      +2.4%      33398 ± 0%  iostat.sdf.wkB/s
     32617 ± 0%      +2.4%      33401 ± 0%  iostat.sdk.wkB/s
     32612 ± 0%      +2.4%      33393 ± 0%  iostat.sdg.wkB/s
      8083 ± 0%      +2.4%       8277 ± 0%  iostat.sdl.wrqm/s
      8083 ± 0%      +2.4%       8276 ± 0%  iostat.sdi.wrqm/s
      8082 ± 0%      +2.4%       8277 ± 0%  iostat.sdc.wrqm/s
     32614 ± 0%      +2.4%      33396 ± 0%  iostat.sdl.wkB/s
      8083 ± 0%      +2.4%       8278 ± 0%  iostat.sde.wrqm/s
      8082 ± 0%      +2.4%       8277 ± 0%  iostat.sdh.wrqm/s
      8083 ± 0%      +2.4%       8277 ± 0%  iostat.sdd.wrqm/s
     32614 ± 0%      +2.4%      33393 ± 0%  iostat.sdi.wkB/s
     32611 ± 0%      +2.4%      33395 ± 0%  iostat.sdc.wkB/s
    325759 ± 0%      +2.4%     333577 ± 0%  iostat.md0.wkB/s
      1274 ± 0%      +2.4%       1305 ± 0%  iostat.md0.w/s
      8082 ± 0%      +2.4%       8277 ± 0%  iostat.sdb.wrqm/s
     32618 ± 0%      +2.4%      33398 ± 0%  iostat.sde.wkB/s
     32612 ± 0%      +2.4%      33395 ± 0%  iostat.sdh.wkB/s
     32618 ± 0%      +2.4%      33397 ± 0%  iostat.sdd.wkB/s
      8084 ± 0%      +2.4%       8278 ± 0%  iostat.sdj.wrqm/s
     32611 ± 0%      +2.4%      33396 ± 0%  iostat.sdb.wkB/s
     32618 ± 0%      +2.4%      33400 ± 0%  iostat.sdj.wkB/s
   2.3e+11 ± 0%      +2.5%  2.357e+11 ± 0%  perf-stat.dTLB-stores
      4898 ± 0%      +2.1%       5003 ± 0%  vmstat.system.cs
 1.017e+12 ± 0%      +2.4%  1.042e+12 ± 0%  perf-stat.iTLB-loads
   1518279 ± 0%      +2.1%    1549457 ± 0%  perf-stat.context-switches
 1.456e+12 ± 0%      +1.4%  1.476e+12 ± 0%  perf-stat.cpu-cycles
 1.456e+12 ± 0%      +1.3%  1.475e+12 ± 0%  perf-stat.ref-cycles
 1.819e+11 ± 0%      +1.3%  1.843e+11 ± 0%  perf-stat.bus-cycles

lkp-sb03 is a Sandy Bridge-EP server.
Memory: 64G
Architecture:          x86_64
CPU op-mode(s):        32-bit, 64-bit
Byte Order:            Little Endian
CPU(s):                32
On-line CPU(s) list:   0-31
Thread(s) per core:    2
Core(s) per socket:    8
Socket(s):             2
NUMA node(s):          2
Vendor ID:             GenuineIntel
CPU family:            6
Model:                 45
Stepping:              6
CPU MHz:               3500.613
BogoMIPS:              5391.16
Virtualization:        VT-x
L1d cache:             32K
L1i cache:             32K
L2 cache:              256K
L3 cache:              20480K
NUMA node0 CPU(s):     0-7,16-23
NUMA node1 CPU(s):     8-15,24-31

lkp-st02 is Core2
Memory: 8G




                          time.involuntary_context_switches

  40000 O+------------------------------------------------------------------+
        |        O               O          O                               |
  35000 ++O  O O        O   O             O                                 |
  30000 ++         O                 O         O                            |
        |                 O    O   O                                        |
  25000 ++            O                 O                                   |
        |                                                                   |
  20000 ++                                                                  |
        |                                                                   |
  15000 ++                                                                  |
  10000 ++                                                                  |
        |                                                                   |
   5000 ++                                                                  |
        |                                                .*.                |
      0 *+*--*-*-*-*--*-*-*-*--*-*-*-*--*-*-*--*-*-*-*--*---*-*--*-*-*-*--*-*


	[*] bisect-good sample
	[O] bisect-bad  sample


Disclaimer:
Results have been estimated based on internal Intel analysis and are provided
for informational purposes only. Any difference in system hardware or software
design or configuration may affect actual performance.

Thanks,
Fengguang

[-- Attachment #2: reproduce.ksh --]
[-- Type: text/plain, Size: 2403 bytes --]

echo performance > /sys/devices/system/cpu/cpu0/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu1/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu10/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu11/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu12/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu13/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu14/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu15/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu16/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu17/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu18/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu19/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu2/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu20/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu21/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu22/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu23/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu24/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu25/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu26/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu27/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu28/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu29/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu3/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu30/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu31/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu4/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu5/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu6/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu7/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu8/cpufreq/scaling_governor
echo performance > /sys/devices/system/cpu/cpu9/cpufreq/scaling_governor
./nepim
./nepim -c 127.0.0.1 -d -a 300 -n 32

             reply	other threads:[~2014-08-21 14:02 UTC|newest]

Thread overview: 6+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2014-08-21 14:01 Fengguang Wu [this message]
2014-08-21 14:01 ` [sched/fair] caeb178c60f: +252.0% cpuidle.C1-SNB.time, +3.1% turbostat.Pkg_W Fengguang Wu
2014-08-21 14:16 ` Rik van Riel
2014-08-21 14:16   ` Rik van Riel
2014-08-21 15:00   ` Fengguang Wu
2014-08-21 15:00     ` Fengguang Wu

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20140821140134.GB19246@localhost \
    --to=fengguang.wu@intel.com \
    --cc=dave.hansen@intel.com \
    --cc=linux-kernel@vger.kernel.org \
    --cc=lkp@01.org \
    --cc=mingo@elte.hu \
    --cc=riel@redhat.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.