All of lore.kernel.org
 help / color / mirror / Atom feed
From: Fengguang Wu <fengguang.wu@intel.com>
To: Christoph Hellwig <hch@lst.de>
Cc: Dave Chinner <david@fromorbit.com>,
	Ye Xiaolong <xiaolong.ye@intel.com>,
	Linus Torvalds <torvalds@linux-foundation.org>,
	LKML <linux-kernel@vger.kernel.org>,
	Bob Peterson <rpeterso@redhat.com>, LKP <lkp@01.org>
Subject: Re: [LKP] [lkp] [xfs] 68a9f5e700: aim7.jobs-per-min -13.6% regression
Date: Mon, 15 Aug 2016 22:14:55 +0800	[thread overview]
Message-ID: <20160815141455.GA22903@wfg-t540p.sh.intel.com> (raw)
In-Reply-To: <20160814161724.GA20274@lst.de>

Hi Christoph,

On Sun, Aug 14, 2016 at 06:17:24PM +0200, Christoph Hellwig wrote:
>Snipping the long contest:
>
>I think there are three observations here:
>
> (1) removing the mark_page_accessed (which is the only significant
>     change in the parent commit)  hurts the
>     aim7/1BRD_48G-xfs-disk_rr-3000-performance/ivb44 test.
>     I'd still rather stick to the filemap version and let the
>     VM people sort it out.  How do the numbers for this test
>     look for XFS vs say ext4 and btrfs?
> (2) lots of additional spinlock contention in the new case.  A quick
>     check shows that I fat-fingered my rewrite so that we do
>     the xfs_inode_set_eofblocks_tag call now for the pure lookup
>     case, and pretty much all new cycles come from that.
> (3) Boy, are those xfs_inode_set_eofblocks_tag calls expensive, and
>     we're already doing way to many even without my little bug above.
>
>So I've force pushed a new version of the iomap-fixes branch with
>(2) fixed, and also a little patch to xfs_inode_set_eofblocks_tag a
>lot less expensive slotted in before that.  Would be good to see
>the numbers with that.

The aim7 1BRD tests finished and there are ups and downs, with overall
performance remain flat.

99091700659f4df9  74a242ad94d13436a1644c0b45  bf4dc6e4ecc2a3d042029319bc  testcase/testparams/testbox
----------------  --------------------------  --------------------------  ---------------------------
         %stddev     %change         %stddev     %change         %stddev
             \          |                \          |                \  
    159926                      157324                      158574        GEO-MEAN aim7.jobs-per-min
     70897               5%      74137               4%      73775        aim7/1BRD_48G-xfs-creat-clo-1500-performance/ivb44
    485217 ±  3%                492431                      477533        aim7/1BRD_48G-xfs-disk_rd-9000-performance/ivb44
    360451             -19%     292980             -17%     299377        aim7/1BRD_48G-xfs-disk_rr-3000-performance/ivb44
    338114                      338410               5%     354078        aim7/1BRD_48G-xfs-disk_rw-3000-performance/ivb44
     60130 ±  5%         4%      62438               5%      62923        aim7/1BRD_48G-xfs-disk_src-3000-performance/ivb44
    403144                      397790                      410648        aim7/1BRD_48G-xfs-disk_wrt-3000-performance/ivb44
     26327                       26534                       26128        aim7/1BRD_48G-xfs-sync_disk_rw-600-performance/ivb44

The new commit bf4dc6e ("xfs: rewrite and optimize the delalloc write
path") improves the aim7/1BRD_48G-xfs-disk_rw-3000-performance/ivb44
case by 5%. Here are the detailed numbers:

aim7/1BRD_48G-xfs-disk_rw-3000-performance/ivb44

74a242ad94d13436  bf4dc6e4ecc2a3d042029319bc
----------------  --------------------------
         %stddev     %change         %stddev
             \          |                \
    338410               5%     354078        aim7.jobs-per-min
    404390               8%     435117        aim7.time.voluntary_context_switches
      2502              -4%       2396        aim7.time.maximum_resident_set_size
     15018              -9%      13701        aim7.time.involuntary_context_switches
       900             -11%        801        aim7.time.system_time
     17432              11%      19365        vmstat.system.cs
     47736 ± 19%       -24%      36087        interrupts.CAL:Function_call_interrupts
   2129646              31%    2790638        proc-vmstat.pgalloc_dma32
    379503              13%     429384        numa-meminfo.node0.Dirty
     15018              -9%      13701        time.involuntary_context_switches
       900             -11%        801        time.system_time
      1560              10%       1716        slabinfo.mnt_cache.active_objs
      1560              10%       1716        slabinfo.mnt_cache.num_objs
     61.53               -4      57.45 ±  4%  perf-profile.cycles-pp.intel_idle.cpuidle_enter_state.cpuidle_enter.call_cpuidle.cpu_startup_entry
     61.63               -4      57.55 ±  4%  perf-profile.func.cycles-pp.intel_idle
   1007188 ± 16%       156%    2577911 ±  6%  numa-numastat.node0.numa_miss
   9662857 ±  4%       -13%    8420159 ±  3%  numa-numastat.node0.numa_foreign
   1008220 ± 16%       155%    2570630 ±  6%  numa-numastat.node1.numa_foreign
   9664033 ±  4%       -13%    8413184 ±  3%  numa-numastat.node1.numa_miss
  26519887 ±  3%        18%   31322674        cpuidle.C1-IVT.time
    122238              16%     142383        cpuidle.C1-IVT.usage
     46548              11%      51645        cpuidle.C1E-IVT.usage
  17253419              13%   19567582        cpuidle.C3-IVT.time
     86847              13%      98333        cpuidle.C3-IVT.usage
    482033 ± 12%       108%    1000665 ±  8%  numa-vmstat.node0.numa_miss
     94689              14%     107744        numa-vmstat.node0.nr_zone_write_pending
     94677              14%     107718        numa-vmstat.node0.nr_dirty
   3156643 ±  3%       -20%    2527460 ±  3%  numa-vmstat.node0.numa_foreign
    429288 ± 12%       129%     983053 ±  8%  numa-vmstat.node1.numa_foreign
   3104193 ±  3%       -19%    2510128        numa-vmstat.node1.numa_miss
      6.43 ±  5%        51%       9.70 ± 11%  turbostat.Pkg%pc2
      0.30              28%       0.38        turbostat.CPU%c3
      9.71                        9.92        turbostat.RAMWatt
       158                         154        turbostat.PkgWatt
       125              -3%        121        turbostat.CorWatt
      1141              -6%       1078        turbostat.Avg_MHz
     38.70              -6%      36.48        turbostat.%Busy
      5.03 ± 11%       -51%       2.46 ± 40%  turbostat.Pkg%pc6
      8.33 ± 48%        88%      15.67 ± 36%  sched_debug.cfs_rq:/.runnable_load_avg.max
      1947 ±  3%       -12%       1710 ±  7%  sched_debug.cfs_rq:/.spread0.stddev
      1936 ±  3%       -12%       1698 ±  8%  sched_debug.cfs_rq:/.min_vruntime.stddev
      2170 ± 10%       -14%       1863 ±  6%  sched_debug.cfs_rq:/.load_avg.max
    220926 ± 18%        37%     303192 ±  5%  sched_debug.cpu.avg_idle.stddev
      0.06 ± 13%       357%       0.28 ± 23%  sched_debug.rt_rq:/.rt_time.avg
      0.37 ± 10%       240%       1.25 ± 15%  sched_debug.rt_rq:/.rt_time.stddev
      2.54 ± 10%       160%       6.59 ± 10%  sched_debug.rt_rq:/.rt_time.max
      0.32 ± 19%        29%       0.42 ± 10%  perf-stat.dTLB-load-miss-rate
    964727               7%    1028830        perf-stat.context-switches
    176406               4%     184289        perf-stat.cpu-migrations
      0.29               4%       0.30        perf-stat.branch-miss-rate
 1.634e+09                   1.673e+09        perf-stat.node-store-misses
     23.60                       23.99        perf-stat.node-store-miss-rate
     40.01                       40.57        perf-stat.cache-miss-rate
      0.95              -8%       0.87        perf-stat.ipc
 3.203e+12              -9%  2.928e+12        perf-stat.cpu-cycles
 1.506e+09             -11%  1.345e+09        perf-stat.branch-misses
     50.64 ± 13%       -14%      43.45 ±  4%  perf-stat.iTLB-load-miss-rate
 5.285e+11             -14%  4.523e+11        perf-stat.branch-instructions
 3.042e+12             -16%  2.551e+12        perf-stat.instructions
 7.996e+11             -18%  6.584e+11        perf-stat.dTLB-loads
 5.569e+11 ±  4%       -18%  4.578e+11        perf-stat.dTLB-stores


Here are the detailed numbers for the slowed down case:

aim7/1BRD_48G-xfs-disk_rr-3000-performance/ivb44

99091700659f4df9  bf4dc6e4ecc2a3d042029319bc
----------------  --------------------------
         %stddev      change         %stddev
             \          |                \
    360451             -17%     299377        aim7.jobs-per-min
     12806             481%      74447        aim7.time.involuntary_context_switches
       755              44%       1086        aim7.time.system_time
     50.17              20%      60.36        aim7.time.elapsed_time
     50.17              20%      60.36        aim7.time.elapsed_time.max
    438148                      446012        aim7.time.voluntary_context_switches
     37798 ± 16%       780%     332583 ±  8%  interrupts.CAL:Function_call_interrupts
     78.82 ±  5%        18%      93.35 ±  5%  uptime.boot
      2847 ±  7%        11%       3160 ±  7%  uptime.idle
    147490 ±  8%        34%     197261 ±  3%  softirqs.RCU
    648159              29%     839283        softirqs.TIMER
    160830              10%     177144        softirqs.SCHED
   3845352 ±  4%        91%    7349133        numa-numastat.node0.numa_miss
   4686838 ±  5%        67%    7835640        numa-numastat.node0.numa_foreign
   3848455 ±  4%        91%    7352436        numa-numastat.node1.numa_foreign
   4689920 ±  5%        67%    7838734        numa-numastat.node1.numa_miss
     50.17              20%      60.36        time.elapsed_time.max
     12806             481%      74447        time.involuntary_context_switches
       755              44%       1086        time.system_time
     50.17              20%      60.36        time.elapsed_time
      1563              18%       1846        time.percent_of_cpu_this_job_got
     11699 ± 19%      3738%     449048        vmstat.io.bo
  18836969             -16%   15789996        vmstat.memory.free
        16              19%         19        vmstat.procs.r
     19377             459%     108364        vmstat.system.cs
     48255              11%      53537        vmstat.system.in
   2357299              25%    2951384        meminfo.Inactive(file)
   2366381              25%    2960468        meminfo.Inactive
   1575292              -9%    1429971        meminfo.Cached
  19342499             -17%   16100340        meminfo.MemFree
   1057904             -20%     842987        meminfo.Dirty
      1057              21%       1284        turbostat.Avg_MHz
     35.78              21%      43.24        turbostat.%Busy
      9.95              15%      11.47        turbostat.RAMWatt
        74 ±  5%        10%         81        turbostat.CoreTmp
        74 ±  4%        10%         81        turbostat.PkgTmp
       118               8%        128        turbostat.CorWatt
       151               7%        162        turbostat.PkgWatt
     29.06             -23%      22.39        turbostat.CPU%c6
       487 ± 89%      3e+04      26448 ± 57%  latency_stats.max.down.xfs_buf_lock._xfs_buf_find.xfs_buf_get_map.xfs_buf_read_map.xfs_trans_read_buf_map.xfs_read_agf.xfs_alloc_read_agf.xfs_alloc_fix_freelist.xfs_free_extent_fix_freelist.xfs_free_extent.xfs_trans_free_extent
      1823 ± 82%      2e+06    1913796 ± 38%  latency_stats.sum.down.xfs_buf_lock._xfs_buf_find.xfs_buf_get_map.xfs_buf_read_map.xfs_trans_read_buf_map.xfs_read_agf.xfs_alloc_read_agf.xfs_alloc_fix_freelist.xfs_free_extent_fix_freelist.xfs_free_extent.xfs_trans_free_extent
    208475 ± 43%      1e+06    1409494 ±  5%  latency_stats.sum.wait_on_page_bit.truncate_inode_pages_range.truncate_inode_pages_final.evict.iput.dentry_unlink_inode.__dentry_kill.dput.__fput.____fput.task_work_run.exit_to_usermode_loop
      6884 ± 73%      8e+04      90790 ±  9%  latency_stats.sum.call_rwsem_down_read_failed.xfs_log_commit_cil.__xfs_trans_commit.xfs_trans_commit.xfs_vn_update_time.file_update_time.xfs_file_aio_write_checks.xfs_file_buffered_aio_write.xfs_file_write_iter.__vfs_write.vfs_write.SyS_write
      1598 ± 20%      3e+04      35015 ± 27%  latency_stats.sum.call_rwsem_down_read_failed.xfs_log_commit_cil.__xfs_trans_commit.__xfs_trans_roll.xfs_trans_roll.xfs_itruncate_extents.xfs_free_eofblocks.xfs_release.xfs_file_release.__fput.____fput.task_work_run
      2006 ± 25%      3e+04      31143 ± 35%  latency_stats.sum.call_rwsem_down_read_failed.xfs_log_commit_cil.__xfs_trans_commit.__xfs_trans_roll.xfs_trans_roll.xfs_itruncate_extents.xfs_inactive_truncate.xfs_inactive.xfs_fs_destroy_inode.destroy_inode.evict.iput
        29 ±101%      1e+04      10214 ± 29%  latency_stats.sum.call_rwsem_down_read_failed.xfs_log_commit_cil.__xfs_trans_commit.__xfs_trans_roll.xfs_trans_roll.xfs_defer_trans_roll.xfs_defer_finish.xfs_itruncate_extents.xfs_inactive_truncate.xfs_inactive.xfs_fs_destroy_inode.destroy_inode
      1206 ± 51%      9e+03       9919 ± 25%  latency_stats.sum.call_rwsem_down_read_failed.xfs_log_commit_cil.__xfs_trans_commit.xfs_trans_commit.xfs_vn_update_time.touch_atime.generic_file_read_iter.xfs_file_buffered_aio_read.xfs_file_read_iter.__vfs_read.vfs_read.SyS_read
  29869205 ±  4%       -10%   26804569        cpuidle.C1-IVT.time
   5737726              39%    7952214        cpuidle.C1E-IVT.time
     51141              17%      59958        cpuidle.C1E-IVT.usage
  18377551              37%   25176426        cpuidle.C3-IVT.time
     96067              17%     112045        cpuidle.C3-IVT.usage
   1806811              12%    2024041        cpuidle.C6-IVT.usage
   1104420 ± 36%       204%    3361085 ± 27%  cpuidle.POLL.time
       281 ± 10%        20%        338        cpuidle.POLL.usage
      5.61 ± 11%       -0.5       5.12 ± 18%  perf-profile.cycles-pp.irq_exit.smp_apic_timer_interrupt.apic_timer_interrupt.cpuidle_enter.call_cpuidle
      5.85 ±  6%       -0.8       5.06 ± 15%  perf-profile.cycles-pp.hrtimer_interrupt.local_apic_timer_interrupt.smp_apic_timer_interrupt.apic_timer_interrupt.cpuidle_enter
      6.32 ±  6%       -0.9       5.42 ± 15%  perf-profile.cycles-pp.local_apic_timer_interrupt.smp_apic_timer_interrupt.apic_timer_interrupt.cpuidle_enter.call_cpuidle
     15.77 ±  8%         -2      13.83 ± 17%  perf-profile.cycles-pp.smp_apic_timer_interrupt.apic_timer_interrupt.cpuidle_enter.call_cpuidle.cpu_startup_entry
     16.04 ±  8%         -2      14.01 ± 15%  perf-profile.cycles-pp.apic_timer_interrupt.cpuidle_enter.call_cpuidle.cpu_startup_entry.start_secondary
     60.25 ±  4%         -7      53.03 ±  7%  perf-profile.cycles-pp.intel_idle.cpuidle_enter_state.cpuidle_enter.call_cpuidle.cpu_startup_entry
     60.41 ±  4%         -7      53.12 ±  7%  perf-profile.func.cycles-pp.intel_idle
   1174104              22%    1436859        numa-meminfo.node0.Inactive
   1167471              22%    1428271        numa-meminfo.node0.Inactive(file)
    770811              -9%     698147        numa-meminfo.node0.FilePages
  20707294             -12%   18281509 ±  6%  numa-meminfo.node0.Active
  20613745             -12%   18180987 ±  6%  numa-meminfo.node0.Active(file)
   9676639             -17%    8003627        numa-meminfo.node0.MemFree
    509906             -22%     396192        numa-meminfo.node0.Dirty
   1189539              28%    1524697        numa-meminfo.node1.Inactive(file)
   1191989              28%    1525194        numa-meminfo.node1.Inactive
    804508             -10%     727067        numa-meminfo.node1.FilePages
   9654540             -16%    8077810        numa-meminfo.node1.MemFree
    547956             -19%     441933        numa-meminfo.node1.Dirty
       396 ± 12%       485%       2320 ± 37%  slabinfo.bio-1.num_objs
       396 ± 12%       481%       2303 ± 37%  slabinfo.bio-1.active_objs
        73             140%        176 ± 14%  slabinfo.kmalloc-128.active_slabs
        73             140%        176 ± 14%  slabinfo.kmalloc-128.num_slabs
      4734              94%       9171 ± 11%  slabinfo.kmalloc-128.num_objs
      4734              88%       8917 ± 13%  slabinfo.kmalloc-128.active_objs
     16238             -10%      14552 ±  3%  slabinfo.kmalloc-256.active_objs
     17189             -13%      15033 ±  3%  slabinfo.kmalloc-256.num_objs
     20651              96%      40387 ± 17%  slabinfo.radix_tree_node.active_objs
       398              91%        761 ± 17%  slabinfo.radix_tree_node.active_slabs
       398              91%        761 ± 17%  slabinfo.radix_tree_node.num_slabs
     22313              91%      42650 ± 17%  slabinfo.radix_tree_node.num_objs
        32             638%        236 ± 28%  slabinfo.xfs_efd_item.active_slabs
        32             638%        236 ± 28%  slabinfo.xfs_efd_item.num_slabs
      1295             281%       4934 ± 23%  slabinfo.xfs_efd_item.num_objs
      1295             280%       4923 ± 23%  slabinfo.xfs_efd_item.active_objs
      1661              81%       3000 ± 42%  slabinfo.xfs_log_ticket.num_objs
      1661              78%       2952 ± 42%  slabinfo.xfs_log_ticket.active_objs
      2617              49%       3905 ± 30%  slabinfo.xfs_trans.num_objs
      2617              48%       3870 ± 31%  slabinfo.xfs_trans.active_objs
   1015933             567%    6779099        perf-stat.context-switches
 4.864e+08             126%  1.101e+09        perf-stat.node-load-misses
 1.179e+09             103%  2.399e+09        perf-stat.node-loads
      0.06 ± 34%        92%       0.12 ± 11%  perf-stat.dTLB-store-miss-rate
 2.985e+08 ± 32%        86%  5.542e+08 ± 11%  perf-stat.dTLB-store-misses
 2.551e+09 ± 15%        81%  4.625e+09 ± 13%  perf-stat.dTLB-load-misses
      0.39 ± 14%        66%       0.65 ± 13%  perf-stat.dTLB-load-miss-rate
  1.26e+09              60%  2.019e+09        perf-stat.node-store-misses
  46072661 ± 27%        49%   68472915        perf-stat.iTLB-loads
 2.738e+12 ±  4%        43%  3.916e+12        perf-stat.cpu-cycles
     21.48              32%      28.35        perf-stat.node-store-miss-rate
 1.612e+10 ±  3%        28%  2.066e+10        perf-stat.cache-references
 1.669e+09 ±  3%        24%  2.063e+09        perf-stat.branch-misses
 6.816e+09 ±  3%        20%  8.179e+09        perf-stat.cache-misses
    177699              18%     209145        perf-stat.cpu-migrations
      0.39              13%       0.44        perf-stat.branch-miss-rate
 4.606e+09              11%  5.102e+09        perf-stat.node-stores
 4.329e+11 ±  4%         9%  4.727e+11        perf-stat.branch-instructions
 6.458e+11               9%  7.046e+11        perf-stat.dTLB-loads
     29.19               8%      31.45        perf-stat.node-load-miss-rate
    286173               8%     308115        perf-stat.page-faults
    286191               8%     308109        perf-stat.minor-faults
  45084934               4%   47073719        perf-stat.iTLB-load-misses
     42.28              -6%      39.58        perf-stat.cache-miss-rate
     50.62 ± 16%       -19%      40.75        perf-stat.iTLB-load-miss-rate
      0.89             -28%       0.64        perf-stat.ipc
         2 ± 36%     4e+07%     970191        proc-vmstat.pgrotated
       150 ± 21%     1e+07%   15356485 ±  3%  proc-vmstat.nr_vmscan_immediate_reclaim
     76823 ± 35%     56899%   43788651        proc-vmstat.pgscan_direct
    153407 ± 19%      4483%    7031431        proc-vmstat.nr_written
    619699 ± 19%      4441%   28139689        proc-vmstat.pgpgout
   5342421            1061%   62050709        proc-vmstat.pgactivate
        47 ± 25%       354%        217        proc-vmstat.nr_pages_scanned
   8542963 ±  3%        78%   15182914        proc-vmstat.numa_miss
   8542963 ±  3%        78%   15182715        proc-vmstat.numa_foreign
   2820568              31%    3699073        proc-vmstat.pgalloc_dma32
    589234              25%     738160        proc-vmstat.nr_zone_inactive_file
    589240              25%     738155        proc-vmstat.nr_inactive_file
  61347830              13%   69522958        proc-vmstat.pgfree
    393711              -9%     356981        proc-vmstat.nr_file_pages
   4831749             -17%    4020131        proc-vmstat.nr_free_pages
  61252784             -18%   50183773        proc-vmstat.pgrefill
  61245420             -18%   50176301        proc-vmstat.pgdeactivate
    264397             -20%     210222        proc-vmstat.nr_zone_write_pending
    264367             -20%     210188        proc-vmstat.nr_dirty
  60420248             -39%   36646178        proc-vmstat.pgscan_kswapd
  60373976             -44%   33735064        proc-vmstat.pgsteal_kswapd
      1753             -98%         43 ± 18%  proc-vmstat.pageoutrun
      1095             -98%         25 ± 17%  proc-vmstat.kswapd_low_wmark_hit_quickly
       656 ±  3%       -98%         15 ± 24%  proc-vmstat.kswapd_high_wmark_hit_quickly
         0                     1136221        numa-vmstat.node0.workingset_refault
         0                     1136221        numa-vmstat.node0.workingset_activate
        23 ± 45%     1e+07%    2756907        numa-vmstat.node0.nr_vmscan_immediate_reclaim
     37618 ± 24%      3234%    1254165        numa-vmstat.node0.nr_written
   1346538 ±  4%       104%    2748439        numa-vmstat.node0.numa_miss
   1577620 ±  5%        80%    2842882        numa-vmstat.node0.numa_foreign
    291242              23%     357407        numa-vmstat.node0.nr_inactive_file
    291237              23%     357390        numa-vmstat.node0.nr_zone_inactive_file
  13961935              12%   15577331        numa-vmstat.node0.numa_local
  13961938              12%   15577332        numa-vmstat.node0.numa_hit
     39831              10%      43768        numa-vmstat.node0.nr_unevictable
     39831              10%      43768        numa-vmstat.node0.nr_zone_unevictable
    193467             -10%     174639        numa-vmstat.node0.nr_file_pages
   5147212             -12%    4542321 ±  6%  numa-vmstat.node0.nr_active_file
   5147237             -12%    4542325 ±  6%  numa-vmstat.node0.nr_zone_active_file
   2426129             -17%    2008637        numa-vmstat.node0.nr_free_pages
    128285             -23%      99206        numa-vmstat.node0.nr_zone_write_pending
    128259             -23%      99183        numa-vmstat.node0.nr_dirty
         0                     1190594        numa-vmstat.node1.workingset_refault
         0                     1190594        numa-vmstat.node1.workingset_activate
        21 ± 36%     1e+07%    3120425 ±  4%  numa-vmstat.node1.nr_vmscan_immediate_reclaim
     38541 ± 26%      3336%    1324185        numa-vmstat.node1.nr_written
   1316819 ±  4%       105%    2699075        numa-vmstat.node1.numa_foreign
   1547929 ±  4%        80%    2793491        numa-vmstat.node1.numa_miss
    296714              28%     381124        numa-vmstat.node1.nr_zone_inactive_file
    296714              28%     381123        numa-vmstat.node1.nr_inactive_file
  14311131              10%   15750908        numa-vmstat.node1.numa_hit
  14311130              10%   15750905        numa-vmstat.node1.numa_local
    201164             -10%     181742        numa-vmstat.node1.nr_file_pages
   2422825             -16%    2027750        numa-vmstat.node1.nr_free_pages
    137069             -19%     110501        numa-vmstat.node1.nr_zone_write_pending
    137069             -19%     110497        numa-vmstat.node1.nr_dirty
       737 ± 29%     27349%     202387        sched_debug.cfs_rq:/.min_vruntime.min
      3637 ± 20%      7919%     291675        sched_debug.cfs_rq:/.min_vruntime.avg
     11.00 ± 44%      4892%     549.17 ±  9%  sched_debug.cfs_rq:/.runnable_load_avg.max
      2.12 ± 36%      4853%     105.12 ±  5%  sched_debug.cfs_rq:/.runnable_load_avg.stddev
      1885 ±  6%      4189%      80870        sched_debug.cfs_rq:/.min_vruntime.stddev
      1896 ±  6%      4166%      80895        sched_debug.cfs_rq:/.spread0.stddev
     10774 ± 13%      4113%     453925        sched_debug.cfs_rq:/.min_vruntime.max
      1.02 ± 19%      2630%      27.72 ±  7%  sched_debug.cfs_rq:/.runnable_load_avg.avg
     63060 ± 45%       776%     552157        sched_debug.cfs_rq:/.load.max
     14442 ± 21%       590%      99615 ± 14%  sched_debug.cfs_rq:/.load.stddev
      8397 ±  9%       309%      34370 ± 12%  sched_debug.cfs_rq:/.load.avg
     46.02 ± 24%       176%     126.96 ±  6%  sched_debug.cfs_rq:/.util_avg.stddev
       817              19%        974 ±  3%  sched_debug.cfs_rq:/.util_avg.max
       721             -17%        600 ±  3%  sched_debug.cfs_rq:/.util_avg.avg
       595 ± 11%       -38%        371 ±  7%  sched_debug.cfs_rq:/.util_avg.min
      1484 ± 20%       -47%        792 ±  5%  sched_debug.cfs_rq:/.load_avg.min
      1798 ±  4%       -50%        903 ±  5%  sched_debug.cfs_rq:/.load_avg.avg
       322 ±  8%      7726%      25239 ±  8%  sched_debug.cpu.nr_switches.min
       969            7238%      71158        sched_debug.cpu.nr_switches.avg
      2.23 ± 40%      4650%     106.14 ±  4%  sched_debug.cpu.cpu_load[0].stddev
       943 ±  4%      3475%      33730 ±  3%  sched_debug.cpu.nr_switches.stddev
      0.87 ± 25%      3057%      27.46 ±  7%  sched_debug.cpu.cpu_load[0].avg
      5.43 ± 13%      2232%     126.61        sched_debug.cpu.nr_uninterruptible.stddev
      6131 ±  3%      2028%     130453        sched_debug.cpu.nr_switches.max
      1.58 ± 29%      1852%      30.90 ±  4%  sched_debug.cpu.cpu_load[4].avg
      2.00 ± 49%      1422%      30.44 ±  5%  sched_debug.cpu.cpu_load[3].avg
     63060 ± 45%      1053%     726920 ± 32%  sched_debug.cpu.load.max
     21.25 ± 44%       777%     186.33 ±  7%  sched_debug.cpu.nr_uninterruptible.max
     14419 ± 21%       731%     119865 ± 31%  sched_debug.cpu.load.stddev
      3586             381%      17262        sched_debug.cpu.nr_load_updates.min
      8286 ±  8%       364%      38414 ± 17%  sched_debug.cpu.load.avg
      5444             303%      21956        sched_debug.cpu.nr_load_updates.avg
      1156             231%       3827        sched_debug.cpu.nr_load_updates.stddev
      8603 ±  4%       222%      27662        sched_debug.cpu.nr_load_updates.max
      1410             165%       3735        sched_debug.cpu.curr->pid.max
     28742 ± 15%       120%      63101 ±  7%  sched_debug.cpu.clock.min
     28742 ± 15%       120%      63101 ±  7%  sched_debug.cpu.clock_task.min
     28748 ± 15%       120%      63107 ±  7%  sched_debug.cpu.clock.avg
     28748 ± 15%       120%      63107 ±  7%  sched_debug.cpu.clock_task.avg
     28751 ± 15%       120%      63113 ±  7%  sched_debug.cpu.clock.max
     28751 ± 15%       120%      63113 ±  7%  sched_debug.cpu.clock_task.max
       442 ± 11%        93%        854 ± 15%  sched_debug.cpu.curr->pid.avg
       618 ±  3%        72%       1065 ±  4%  sched_debug.cpu.curr->pid.stddev
      1.88 ± 11%        50%       2.83 ±  8%  sched_debug.cpu.clock.stddev
      1.88 ± 11%        50%       2.83 ±  8%  sched_debug.cpu.clock_task.stddev
      5.22 ±  9%       -55%       2.34 ± 23%  sched_debug.rt_rq:/.rt_time.max
      0.85             -55%       0.38 ± 28%  sched_debug.rt_rq:/.rt_time.stddev
      0.17             -56%       0.07 ± 33%  sched_debug.rt_rq:/.rt_time.avg
     27633 ± 16%       124%      61980 ±  8%  sched_debug.ktime
     28745 ± 15%       120%      63102 ±  7%  sched_debug.sched_clk
     28745 ± 15%       120%      63102 ±  7%  sched_debug.cpu_clk

Thanks,
Fengguang

WARNING: multiple messages have this Message-ID (diff)
From: Fengguang Wu <fengguang.wu@intel.com>
To: lkp@lists.01.org
Subject: Re: [xfs] 68a9f5e700: aim7.jobs-per-min -13.6% regression
Date: Mon, 15 Aug 2016 22:14:55 +0800	[thread overview]
Message-ID: <20160815141455.GA22903@wfg-t540p.sh.intel.com> (raw)
In-Reply-To: <20160814161724.GA20274@lst.de>

[-- Attachment #1: Type: text/plain, Size: 27403 bytes --]

Hi Christoph,

On Sun, Aug 14, 2016 at 06:17:24PM +0200, Christoph Hellwig wrote:
>Snipping the long contest:
>
>I think there are three observations here:
>
> (1) removing the mark_page_accessed (which is the only significant
>     change in the parent commit)  hurts the
>     aim7/1BRD_48G-xfs-disk_rr-3000-performance/ivb44 test.
>     I'd still rather stick to the filemap version and let the
>     VM people sort it out.  How do the numbers for this test
>     look for XFS vs say ext4 and btrfs?
> (2) lots of additional spinlock contention in the new case.  A quick
>     check shows that I fat-fingered my rewrite so that we do
>     the xfs_inode_set_eofblocks_tag call now for the pure lookup
>     case, and pretty much all new cycles come from that.
> (3) Boy, are those xfs_inode_set_eofblocks_tag calls expensive, and
>     we're already doing way to many even without my little bug above.
>
>So I've force pushed a new version of the iomap-fixes branch with
>(2) fixed, and also a little patch to xfs_inode_set_eofblocks_tag a
>lot less expensive slotted in before that.  Would be good to see
>the numbers with that.

The aim7 1BRD tests finished and there are ups and downs, with overall
performance remain flat.

99091700659f4df9  74a242ad94d13436a1644c0b45  bf4dc6e4ecc2a3d042029319bc  testcase/testparams/testbox
----------------  --------------------------  --------------------------  ---------------------------
         %stddev     %change         %stddev     %change         %stddev
             \          |                \          |                \  
    159926                      157324                      158574        GEO-MEAN aim7.jobs-per-min
     70897               5%      74137               4%      73775        aim7/1BRD_48G-xfs-creat-clo-1500-performance/ivb44
    485217 ±  3%                492431                      477533        aim7/1BRD_48G-xfs-disk_rd-9000-performance/ivb44
    360451             -19%     292980             -17%     299377        aim7/1BRD_48G-xfs-disk_rr-3000-performance/ivb44
    338114                      338410               5%     354078        aim7/1BRD_48G-xfs-disk_rw-3000-performance/ivb44
     60130 ±  5%         4%      62438               5%      62923        aim7/1BRD_48G-xfs-disk_src-3000-performance/ivb44
    403144                      397790                      410648        aim7/1BRD_48G-xfs-disk_wrt-3000-performance/ivb44
     26327                       26534                       26128        aim7/1BRD_48G-xfs-sync_disk_rw-600-performance/ivb44

The new commit bf4dc6e ("xfs: rewrite and optimize the delalloc write
path") improves the aim7/1BRD_48G-xfs-disk_rw-3000-performance/ivb44
case by 5%. Here are the detailed numbers:

aim7/1BRD_48G-xfs-disk_rw-3000-performance/ivb44

74a242ad94d13436  bf4dc6e4ecc2a3d042029319bc
----------------  --------------------------
         %stddev     %change         %stddev
             \          |                \
    338410               5%     354078        aim7.jobs-per-min
    404390               8%     435117        aim7.time.voluntary_context_switches
      2502              -4%       2396        aim7.time.maximum_resident_set_size
     15018              -9%      13701        aim7.time.involuntary_context_switches
       900             -11%        801        aim7.time.system_time
     17432              11%      19365        vmstat.system.cs
     47736 ± 19%       -24%      36087        interrupts.CAL:Function_call_interrupts
   2129646              31%    2790638        proc-vmstat.pgalloc_dma32
    379503              13%     429384        numa-meminfo.node0.Dirty
     15018              -9%      13701        time.involuntary_context_switches
       900             -11%        801        time.system_time
      1560              10%       1716        slabinfo.mnt_cache.active_objs
      1560              10%       1716        slabinfo.mnt_cache.num_objs
     61.53               -4      57.45 ±  4%  perf-profile.cycles-pp.intel_idle.cpuidle_enter_state.cpuidle_enter.call_cpuidle.cpu_startup_entry
     61.63               -4      57.55 ±  4%  perf-profile.func.cycles-pp.intel_idle
   1007188 ± 16%       156%    2577911 ±  6%  numa-numastat.node0.numa_miss
   9662857 ±  4%       -13%    8420159 ±  3%  numa-numastat.node0.numa_foreign
   1008220 ± 16%       155%    2570630 ±  6%  numa-numastat.node1.numa_foreign
   9664033 ±  4%       -13%    8413184 ±  3%  numa-numastat.node1.numa_miss
  26519887 ±  3%        18%   31322674        cpuidle.C1-IVT.time
    122238              16%     142383        cpuidle.C1-IVT.usage
     46548              11%      51645        cpuidle.C1E-IVT.usage
  17253419              13%   19567582        cpuidle.C3-IVT.time
     86847              13%      98333        cpuidle.C3-IVT.usage
    482033 ± 12%       108%    1000665 ±  8%  numa-vmstat.node0.numa_miss
     94689              14%     107744        numa-vmstat.node0.nr_zone_write_pending
     94677              14%     107718        numa-vmstat.node0.nr_dirty
   3156643 ±  3%       -20%    2527460 ±  3%  numa-vmstat.node0.numa_foreign
    429288 ± 12%       129%     983053 ±  8%  numa-vmstat.node1.numa_foreign
   3104193 ±  3%       -19%    2510128        numa-vmstat.node1.numa_miss
      6.43 ±  5%        51%       9.70 ± 11%  turbostat.Pkg%pc2
      0.30              28%       0.38        turbostat.CPU%c3
      9.71                        9.92        turbostat.RAMWatt
       158                         154        turbostat.PkgWatt
       125              -3%        121        turbostat.CorWatt
      1141              -6%       1078        turbostat.Avg_MHz
     38.70              -6%      36.48        turbostat.%Busy
      5.03 ± 11%       -51%       2.46 ± 40%  turbostat.Pkg%pc6
      8.33 ± 48%        88%      15.67 ± 36%  sched_debug.cfs_rq:/.runnable_load_avg.max
      1947 ±  3%       -12%       1710 ±  7%  sched_debug.cfs_rq:/.spread0.stddev
      1936 ±  3%       -12%       1698 ±  8%  sched_debug.cfs_rq:/.min_vruntime.stddev
      2170 ± 10%       -14%       1863 ±  6%  sched_debug.cfs_rq:/.load_avg.max
    220926 ± 18%        37%     303192 ±  5%  sched_debug.cpu.avg_idle.stddev
      0.06 ± 13%       357%       0.28 ± 23%  sched_debug.rt_rq:/.rt_time.avg
      0.37 ± 10%       240%       1.25 ± 15%  sched_debug.rt_rq:/.rt_time.stddev
      2.54 ± 10%       160%       6.59 ± 10%  sched_debug.rt_rq:/.rt_time.max
      0.32 ± 19%        29%       0.42 ± 10%  perf-stat.dTLB-load-miss-rate
    964727               7%    1028830        perf-stat.context-switches
    176406               4%     184289        perf-stat.cpu-migrations
      0.29               4%       0.30        perf-stat.branch-miss-rate
 1.634e+09                   1.673e+09        perf-stat.node-store-misses
     23.60                       23.99        perf-stat.node-store-miss-rate
     40.01                       40.57        perf-stat.cache-miss-rate
      0.95              -8%       0.87        perf-stat.ipc
 3.203e+12              -9%  2.928e+12        perf-stat.cpu-cycles
 1.506e+09             -11%  1.345e+09        perf-stat.branch-misses
     50.64 ± 13%       -14%      43.45 ±  4%  perf-stat.iTLB-load-miss-rate
 5.285e+11             -14%  4.523e+11        perf-stat.branch-instructions
 3.042e+12             -16%  2.551e+12        perf-stat.instructions
 7.996e+11             -18%  6.584e+11        perf-stat.dTLB-loads
 5.569e+11 ±  4%       -18%  4.578e+11        perf-stat.dTLB-stores


Here are the detailed numbers for the slowed down case:

aim7/1BRD_48G-xfs-disk_rr-3000-performance/ivb44

99091700659f4df9  bf4dc6e4ecc2a3d042029319bc
----------------  --------------------------
         %stddev      change         %stddev
             \          |                \
    360451             -17%     299377        aim7.jobs-per-min
     12806             481%      74447        aim7.time.involuntary_context_switches
       755              44%       1086        aim7.time.system_time
     50.17              20%      60.36        aim7.time.elapsed_time
     50.17              20%      60.36        aim7.time.elapsed_time.max
    438148                      446012        aim7.time.voluntary_context_switches
     37798 ± 16%       780%     332583 ±  8%  interrupts.CAL:Function_call_interrupts
     78.82 ±  5%        18%      93.35 ±  5%  uptime.boot
      2847 ±  7%        11%       3160 ±  7%  uptime.idle
    147490 ±  8%        34%     197261 ±  3%  softirqs.RCU
    648159              29%     839283        softirqs.TIMER
    160830              10%     177144        softirqs.SCHED
   3845352 ±  4%        91%    7349133        numa-numastat.node0.numa_miss
   4686838 ±  5%        67%    7835640        numa-numastat.node0.numa_foreign
   3848455 ±  4%        91%    7352436        numa-numastat.node1.numa_foreign
   4689920 ±  5%        67%    7838734        numa-numastat.node1.numa_miss
     50.17              20%      60.36        time.elapsed_time.max
     12806             481%      74447        time.involuntary_context_switches
       755              44%       1086        time.system_time
     50.17              20%      60.36        time.elapsed_time
      1563              18%       1846        time.percent_of_cpu_this_job_got
     11699 ± 19%      3738%     449048        vmstat.io.bo
  18836969             -16%   15789996        vmstat.memory.free
        16              19%         19        vmstat.procs.r
     19377             459%     108364        vmstat.system.cs
     48255              11%      53537        vmstat.system.in
   2357299              25%    2951384        meminfo.Inactive(file)
   2366381              25%    2960468        meminfo.Inactive
   1575292              -9%    1429971        meminfo.Cached
  19342499             -17%   16100340        meminfo.MemFree
   1057904             -20%     842987        meminfo.Dirty
      1057              21%       1284        turbostat.Avg_MHz
     35.78              21%      43.24        turbostat.%Busy
      9.95              15%      11.47        turbostat.RAMWatt
        74 ±  5%        10%         81        turbostat.CoreTmp
        74 ±  4%        10%         81        turbostat.PkgTmp
       118               8%        128        turbostat.CorWatt
       151               7%        162        turbostat.PkgWatt
     29.06             -23%      22.39        turbostat.CPU%c6
       487 ± 89%      3e+04      26448 ± 57%  latency_stats.max.down.xfs_buf_lock._xfs_buf_find.xfs_buf_get_map.xfs_buf_read_map.xfs_trans_read_buf_map.xfs_read_agf.xfs_alloc_read_agf.xfs_alloc_fix_freelist.xfs_free_extent_fix_freelist.xfs_free_extent.xfs_trans_free_extent
      1823 ± 82%      2e+06    1913796 ± 38%  latency_stats.sum.down.xfs_buf_lock._xfs_buf_find.xfs_buf_get_map.xfs_buf_read_map.xfs_trans_read_buf_map.xfs_read_agf.xfs_alloc_read_agf.xfs_alloc_fix_freelist.xfs_free_extent_fix_freelist.xfs_free_extent.xfs_trans_free_extent
    208475 ± 43%      1e+06    1409494 ±  5%  latency_stats.sum.wait_on_page_bit.truncate_inode_pages_range.truncate_inode_pages_final.evict.iput.dentry_unlink_inode.__dentry_kill.dput.__fput.____fput.task_work_run.exit_to_usermode_loop
      6884 ± 73%      8e+04      90790 ±  9%  latency_stats.sum.call_rwsem_down_read_failed.xfs_log_commit_cil.__xfs_trans_commit.xfs_trans_commit.xfs_vn_update_time.file_update_time.xfs_file_aio_write_checks.xfs_file_buffered_aio_write.xfs_file_write_iter.__vfs_write.vfs_write.SyS_write
      1598 ± 20%      3e+04      35015 ± 27%  latency_stats.sum.call_rwsem_down_read_failed.xfs_log_commit_cil.__xfs_trans_commit.__xfs_trans_roll.xfs_trans_roll.xfs_itruncate_extents.xfs_free_eofblocks.xfs_release.xfs_file_release.__fput.____fput.task_work_run
      2006 ± 25%      3e+04      31143 ± 35%  latency_stats.sum.call_rwsem_down_read_failed.xfs_log_commit_cil.__xfs_trans_commit.__xfs_trans_roll.xfs_trans_roll.xfs_itruncate_extents.xfs_inactive_truncate.xfs_inactive.xfs_fs_destroy_inode.destroy_inode.evict.iput
        29 ±101%      1e+04      10214 ± 29%  latency_stats.sum.call_rwsem_down_read_failed.xfs_log_commit_cil.__xfs_trans_commit.__xfs_trans_roll.xfs_trans_roll.xfs_defer_trans_roll.xfs_defer_finish.xfs_itruncate_extents.xfs_inactive_truncate.xfs_inactive.xfs_fs_destroy_inode.destroy_inode
      1206 ± 51%      9e+03       9919 ± 25%  latency_stats.sum.call_rwsem_down_read_failed.xfs_log_commit_cil.__xfs_trans_commit.xfs_trans_commit.xfs_vn_update_time.touch_atime.generic_file_read_iter.xfs_file_buffered_aio_read.xfs_file_read_iter.__vfs_read.vfs_read.SyS_read
  29869205 ±  4%       -10%   26804569        cpuidle.C1-IVT.time
   5737726              39%    7952214        cpuidle.C1E-IVT.time
     51141              17%      59958        cpuidle.C1E-IVT.usage
  18377551              37%   25176426        cpuidle.C3-IVT.time
     96067              17%     112045        cpuidle.C3-IVT.usage
   1806811              12%    2024041        cpuidle.C6-IVT.usage
   1104420 ± 36%       204%    3361085 ± 27%  cpuidle.POLL.time
       281 ± 10%        20%        338        cpuidle.POLL.usage
      5.61 ± 11%       -0.5       5.12 ± 18%  perf-profile.cycles-pp.irq_exit.smp_apic_timer_interrupt.apic_timer_interrupt.cpuidle_enter.call_cpuidle
      5.85 ±  6%       -0.8       5.06 ± 15%  perf-profile.cycles-pp.hrtimer_interrupt.local_apic_timer_interrupt.smp_apic_timer_interrupt.apic_timer_interrupt.cpuidle_enter
      6.32 ±  6%       -0.9       5.42 ± 15%  perf-profile.cycles-pp.local_apic_timer_interrupt.smp_apic_timer_interrupt.apic_timer_interrupt.cpuidle_enter.call_cpuidle
     15.77 ±  8%         -2      13.83 ± 17%  perf-profile.cycles-pp.smp_apic_timer_interrupt.apic_timer_interrupt.cpuidle_enter.call_cpuidle.cpu_startup_entry
     16.04 ±  8%         -2      14.01 ± 15%  perf-profile.cycles-pp.apic_timer_interrupt.cpuidle_enter.call_cpuidle.cpu_startup_entry.start_secondary
     60.25 ±  4%         -7      53.03 ±  7%  perf-profile.cycles-pp.intel_idle.cpuidle_enter_state.cpuidle_enter.call_cpuidle.cpu_startup_entry
     60.41 ±  4%         -7      53.12 ±  7%  perf-profile.func.cycles-pp.intel_idle
   1174104              22%    1436859        numa-meminfo.node0.Inactive
   1167471              22%    1428271        numa-meminfo.node0.Inactive(file)
    770811              -9%     698147        numa-meminfo.node0.FilePages
  20707294             -12%   18281509 ±  6%  numa-meminfo.node0.Active
  20613745             -12%   18180987 ±  6%  numa-meminfo.node0.Active(file)
   9676639             -17%    8003627        numa-meminfo.node0.MemFree
    509906             -22%     396192        numa-meminfo.node0.Dirty
   1189539              28%    1524697        numa-meminfo.node1.Inactive(file)
   1191989              28%    1525194        numa-meminfo.node1.Inactive
    804508             -10%     727067        numa-meminfo.node1.FilePages
   9654540             -16%    8077810        numa-meminfo.node1.MemFree
    547956             -19%     441933        numa-meminfo.node1.Dirty
       396 ± 12%       485%       2320 ± 37%  slabinfo.bio-1.num_objs
       396 ± 12%       481%       2303 ± 37%  slabinfo.bio-1.active_objs
        73             140%        176 ± 14%  slabinfo.kmalloc-128.active_slabs
        73             140%        176 ± 14%  slabinfo.kmalloc-128.num_slabs
      4734              94%       9171 ± 11%  slabinfo.kmalloc-128.num_objs
      4734              88%       8917 ± 13%  slabinfo.kmalloc-128.active_objs
     16238             -10%      14552 ±  3%  slabinfo.kmalloc-256.active_objs
     17189             -13%      15033 ±  3%  slabinfo.kmalloc-256.num_objs
     20651              96%      40387 ± 17%  slabinfo.radix_tree_node.active_objs
       398              91%        761 ± 17%  slabinfo.radix_tree_node.active_slabs
       398              91%        761 ± 17%  slabinfo.radix_tree_node.num_slabs
     22313              91%      42650 ± 17%  slabinfo.radix_tree_node.num_objs
        32             638%        236 ± 28%  slabinfo.xfs_efd_item.active_slabs
        32             638%        236 ± 28%  slabinfo.xfs_efd_item.num_slabs
      1295             281%       4934 ± 23%  slabinfo.xfs_efd_item.num_objs
      1295             280%       4923 ± 23%  slabinfo.xfs_efd_item.active_objs
      1661              81%       3000 ± 42%  slabinfo.xfs_log_ticket.num_objs
      1661              78%       2952 ± 42%  slabinfo.xfs_log_ticket.active_objs
      2617              49%       3905 ± 30%  slabinfo.xfs_trans.num_objs
      2617              48%       3870 ± 31%  slabinfo.xfs_trans.active_objs
   1015933             567%    6779099        perf-stat.context-switches
 4.864e+08             126%  1.101e+09        perf-stat.node-load-misses
 1.179e+09             103%  2.399e+09        perf-stat.node-loads
      0.06 ± 34%        92%       0.12 ± 11%  perf-stat.dTLB-store-miss-rate
 2.985e+08 ± 32%        86%  5.542e+08 ± 11%  perf-stat.dTLB-store-misses
 2.551e+09 ± 15%        81%  4.625e+09 ± 13%  perf-stat.dTLB-load-misses
      0.39 ± 14%        66%       0.65 ± 13%  perf-stat.dTLB-load-miss-rate
  1.26e+09              60%  2.019e+09        perf-stat.node-store-misses
  46072661 ± 27%        49%   68472915        perf-stat.iTLB-loads
 2.738e+12 ±  4%        43%  3.916e+12        perf-stat.cpu-cycles
     21.48              32%      28.35        perf-stat.node-store-miss-rate
 1.612e+10 ±  3%        28%  2.066e+10        perf-stat.cache-references
 1.669e+09 ±  3%        24%  2.063e+09        perf-stat.branch-misses
 6.816e+09 ±  3%        20%  8.179e+09        perf-stat.cache-misses
    177699              18%     209145        perf-stat.cpu-migrations
      0.39              13%       0.44        perf-stat.branch-miss-rate
 4.606e+09              11%  5.102e+09        perf-stat.node-stores
 4.329e+11 ±  4%         9%  4.727e+11        perf-stat.branch-instructions
 6.458e+11               9%  7.046e+11        perf-stat.dTLB-loads
     29.19               8%      31.45        perf-stat.node-load-miss-rate
    286173               8%     308115        perf-stat.page-faults
    286191               8%     308109        perf-stat.minor-faults
  45084934               4%   47073719        perf-stat.iTLB-load-misses
     42.28              -6%      39.58        perf-stat.cache-miss-rate
     50.62 ± 16%       -19%      40.75        perf-stat.iTLB-load-miss-rate
      0.89             -28%       0.64        perf-stat.ipc
         2 ± 36%     4e+07%     970191        proc-vmstat.pgrotated
       150 ± 21%     1e+07%   15356485 ±  3%  proc-vmstat.nr_vmscan_immediate_reclaim
     76823 ± 35%     56899%   43788651        proc-vmstat.pgscan_direct
    153407 ± 19%      4483%    7031431        proc-vmstat.nr_written
    619699 ± 19%      4441%   28139689        proc-vmstat.pgpgout
   5342421            1061%   62050709        proc-vmstat.pgactivate
        47 ± 25%       354%        217        proc-vmstat.nr_pages_scanned
   8542963 ±  3%        78%   15182914        proc-vmstat.numa_miss
   8542963 ±  3%        78%   15182715        proc-vmstat.numa_foreign
   2820568              31%    3699073        proc-vmstat.pgalloc_dma32
    589234              25%     738160        proc-vmstat.nr_zone_inactive_file
    589240              25%     738155        proc-vmstat.nr_inactive_file
  61347830              13%   69522958        proc-vmstat.pgfree
    393711              -9%     356981        proc-vmstat.nr_file_pages
   4831749             -17%    4020131        proc-vmstat.nr_free_pages
  61252784             -18%   50183773        proc-vmstat.pgrefill
  61245420             -18%   50176301        proc-vmstat.pgdeactivate
    264397             -20%     210222        proc-vmstat.nr_zone_write_pending
    264367             -20%     210188        proc-vmstat.nr_dirty
  60420248             -39%   36646178        proc-vmstat.pgscan_kswapd
  60373976             -44%   33735064        proc-vmstat.pgsteal_kswapd
      1753             -98%         43 ± 18%  proc-vmstat.pageoutrun
      1095             -98%         25 ± 17%  proc-vmstat.kswapd_low_wmark_hit_quickly
       656 ±  3%       -98%         15 ± 24%  proc-vmstat.kswapd_high_wmark_hit_quickly
         0                     1136221        numa-vmstat.node0.workingset_refault
         0                     1136221        numa-vmstat.node0.workingset_activate
        23 ± 45%     1e+07%    2756907        numa-vmstat.node0.nr_vmscan_immediate_reclaim
     37618 ± 24%      3234%    1254165        numa-vmstat.node0.nr_written
   1346538 ±  4%       104%    2748439        numa-vmstat.node0.numa_miss
   1577620 ±  5%        80%    2842882        numa-vmstat.node0.numa_foreign
    291242              23%     357407        numa-vmstat.node0.nr_inactive_file
    291237              23%     357390        numa-vmstat.node0.nr_zone_inactive_file
  13961935              12%   15577331        numa-vmstat.node0.numa_local
  13961938              12%   15577332        numa-vmstat.node0.numa_hit
     39831              10%      43768        numa-vmstat.node0.nr_unevictable
     39831              10%      43768        numa-vmstat.node0.nr_zone_unevictable
    193467             -10%     174639        numa-vmstat.node0.nr_file_pages
   5147212             -12%    4542321 ±  6%  numa-vmstat.node0.nr_active_file
   5147237             -12%    4542325 ±  6%  numa-vmstat.node0.nr_zone_active_file
   2426129             -17%    2008637        numa-vmstat.node0.nr_free_pages
    128285             -23%      99206        numa-vmstat.node0.nr_zone_write_pending
    128259             -23%      99183        numa-vmstat.node0.nr_dirty
         0                     1190594        numa-vmstat.node1.workingset_refault
         0                     1190594        numa-vmstat.node1.workingset_activate
        21 ± 36%     1e+07%    3120425 ±  4%  numa-vmstat.node1.nr_vmscan_immediate_reclaim
     38541 ± 26%      3336%    1324185        numa-vmstat.node1.nr_written
   1316819 ±  4%       105%    2699075        numa-vmstat.node1.numa_foreign
   1547929 ±  4%        80%    2793491        numa-vmstat.node1.numa_miss
    296714              28%     381124        numa-vmstat.node1.nr_zone_inactive_file
    296714              28%     381123        numa-vmstat.node1.nr_inactive_file
  14311131              10%   15750908        numa-vmstat.node1.numa_hit
  14311130              10%   15750905        numa-vmstat.node1.numa_local
    201164             -10%     181742        numa-vmstat.node1.nr_file_pages
   2422825             -16%    2027750        numa-vmstat.node1.nr_free_pages
    137069             -19%     110501        numa-vmstat.node1.nr_zone_write_pending
    137069             -19%     110497        numa-vmstat.node1.nr_dirty
       737 ± 29%     27349%     202387        sched_debug.cfs_rq:/.min_vruntime.min
      3637 ± 20%      7919%     291675        sched_debug.cfs_rq:/.min_vruntime.avg
     11.00 ± 44%      4892%     549.17 ±  9%  sched_debug.cfs_rq:/.runnable_load_avg.max
      2.12 ± 36%      4853%     105.12 ±  5%  sched_debug.cfs_rq:/.runnable_load_avg.stddev
      1885 ±  6%      4189%      80870        sched_debug.cfs_rq:/.min_vruntime.stddev
      1896 ±  6%      4166%      80895        sched_debug.cfs_rq:/.spread0.stddev
     10774 ± 13%      4113%     453925        sched_debug.cfs_rq:/.min_vruntime.max
      1.02 ± 19%      2630%      27.72 ±  7%  sched_debug.cfs_rq:/.runnable_load_avg.avg
     63060 ± 45%       776%     552157        sched_debug.cfs_rq:/.load.max
     14442 ± 21%       590%      99615 ± 14%  sched_debug.cfs_rq:/.load.stddev
      8397 ±  9%       309%      34370 ± 12%  sched_debug.cfs_rq:/.load.avg
     46.02 ± 24%       176%     126.96 ±  6%  sched_debug.cfs_rq:/.util_avg.stddev
       817              19%        974 ±  3%  sched_debug.cfs_rq:/.util_avg.max
       721             -17%        600 ±  3%  sched_debug.cfs_rq:/.util_avg.avg
       595 ± 11%       -38%        371 ±  7%  sched_debug.cfs_rq:/.util_avg.min
      1484 ± 20%       -47%        792 ±  5%  sched_debug.cfs_rq:/.load_avg.min
      1798 ±  4%       -50%        903 ±  5%  sched_debug.cfs_rq:/.load_avg.avg
       322 ±  8%      7726%      25239 ±  8%  sched_debug.cpu.nr_switches.min
       969            7238%      71158        sched_debug.cpu.nr_switches.avg
      2.23 ± 40%      4650%     106.14 ±  4%  sched_debug.cpu.cpu_load[0].stddev
       943 ±  4%      3475%      33730 ±  3%  sched_debug.cpu.nr_switches.stddev
      0.87 ± 25%      3057%      27.46 ±  7%  sched_debug.cpu.cpu_load[0].avg
      5.43 ± 13%      2232%     126.61        sched_debug.cpu.nr_uninterruptible.stddev
      6131 ±  3%      2028%     130453        sched_debug.cpu.nr_switches.max
      1.58 ± 29%      1852%      30.90 ±  4%  sched_debug.cpu.cpu_load[4].avg
      2.00 ± 49%      1422%      30.44 ±  5%  sched_debug.cpu.cpu_load[3].avg
     63060 ± 45%      1053%     726920 ± 32%  sched_debug.cpu.load.max
     21.25 ± 44%       777%     186.33 ±  7%  sched_debug.cpu.nr_uninterruptible.max
     14419 ± 21%       731%     119865 ± 31%  sched_debug.cpu.load.stddev
      3586             381%      17262        sched_debug.cpu.nr_load_updates.min
      8286 ±  8%       364%      38414 ± 17%  sched_debug.cpu.load.avg
      5444             303%      21956        sched_debug.cpu.nr_load_updates.avg
      1156             231%       3827        sched_debug.cpu.nr_load_updates.stddev
      8603 ±  4%       222%      27662        sched_debug.cpu.nr_load_updates.max
      1410             165%       3735        sched_debug.cpu.curr->pid.max
     28742 ± 15%       120%      63101 ±  7%  sched_debug.cpu.clock.min
     28742 ± 15%       120%      63101 ±  7%  sched_debug.cpu.clock_task.min
     28748 ± 15%       120%      63107 ±  7%  sched_debug.cpu.clock.avg
     28748 ± 15%       120%      63107 ±  7%  sched_debug.cpu.clock_task.avg
     28751 ± 15%       120%      63113 ±  7%  sched_debug.cpu.clock.max
     28751 ± 15%       120%      63113 ±  7%  sched_debug.cpu.clock_task.max
       442 ± 11%        93%        854 ± 15%  sched_debug.cpu.curr->pid.avg
       618 ±  3%        72%       1065 ±  4%  sched_debug.cpu.curr->pid.stddev
      1.88 ± 11%        50%       2.83 ±  8%  sched_debug.cpu.clock.stddev
      1.88 ± 11%        50%       2.83 ±  8%  sched_debug.cpu.clock_task.stddev
      5.22 ±  9%       -55%       2.34 ± 23%  sched_debug.rt_rq:/.rt_time.max
      0.85             -55%       0.38 ± 28%  sched_debug.rt_rq:/.rt_time.stddev
      0.17             -56%       0.07 ± 33%  sched_debug.rt_rq:/.rt_time.avg
     27633 ± 16%       124%      61980 ±  8%  sched_debug.ktime
     28745 ± 15%       120%      63102 ±  7%  sched_debug.sched_clk
     28745 ± 15%       120%      63102 ±  7%  sched_debug.cpu_clk

Thanks,
Fengguang

  parent reply	other threads:[~2016-08-15 14:15 UTC|newest]

Thread overview: 219+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2016-08-09 14:33 [lkp] [xfs] 68a9f5e700: aim7.jobs-per-min -13.6% regression kernel test robot
2016-08-09 14:33 ` kernel test robot
2016-08-10 18:24 ` [lkp] " Linus Torvalds
2016-08-10 18:24   ` Linus Torvalds
2016-08-10 23:08   ` [lkp] " Dave Chinner
2016-08-10 23:08     ` Dave Chinner
2016-08-10 23:51     ` [lkp] " Linus Torvalds
2016-08-10 23:51       ` Linus Torvalds
2016-08-10 23:58       ` [LKP] [lkp] " Huang, Ying
2016-08-10 23:58         ` Huang, Ying
2016-08-11  0:11         ` [LKP] [lkp] " Huang, Ying
2016-08-11  0:11           ` Huang, Ying
2016-08-11  0:23           ` [LKP] [lkp] " Linus Torvalds
2016-08-11  0:23             ` Linus Torvalds
2016-08-11  0:33             ` [LKP] [lkp] " Huang, Ying
2016-08-11  0:33               ` Huang, Ying
2016-08-11  1:00               ` [LKP] [lkp] " Linus Torvalds
2016-08-11  1:00                 ` Linus Torvalds
2016-08-11  4:46                 ` [LKP] [lkp] " Dave Chinner
2016-08-11  4:46                   ` Dave Chinner
2016-08-15 17:22                   ` [LKP] [lkp] " Huang, Ying
2016-08-15 17:22                     ` Huang, Ying
2016-08-16  0:08                     ` [LKP] [lkp] " Dave Chinner
2016-08-16  0:08                       ` Dave Chinner
2016-08-11 15:57                 ` [LKP] [lkp] " Christoph Hellwig
2016-08-11 15:57                   ` Christoph Hellwig
2016-08-11 16:55                   ` [LKP] [lkp] " Linus Torvalds
2016-08-11 16:55                     ` Linus Torvalds
2016-08-11 17:51                     ` [LKP] [lkp] " Huang, Ying
2016-08-11 17:51                       ` Huang, Ying
2016-08-11 19:51                       ` [LKP] [lkp] " Linus Torvalds
2016-08-11 19:51                         ` Linus Torvalds
2016-08-11 20:00                         ` [LKP] [lkp] " Christoph Hellwig
2016-08-11 20:00                           ` Christoph Hellwig
2016-08-11 20:35                           ` [LKP] [lkp] " Linus Torvalds
2016-08-11 20:35                             ` Linus Torvalds
2016-08-11 22:16                             ` [LKP] [lkp] " Al Viro
2016-08-11 22:16                               ` Al Viro
2016-08-11 22:30                               ` [LKP] [lkp] " Linus Torvalds
2016-08-11 22:30                                 ` Linus Torvalds
2016-08-11 21:16                           ` [LKP] [lkp] " Huang, Ying
2016-08-11 21:16                             ` Huang, Ying
2016-08-11 21:40                             ` [LKP] [lkp] " Linus Torvalds
2016-08-11 21:40                               ` Linus Torvalds
2016-08-11 22:08                               ` [LKP] [lkp] " Christoph Hellwig
2016-08-11 22:08                                 ` Christoph Hellwig
2016-08-12  0:54                     ` [LKP] [lkp] " Dave Chinner
2016-08-12  0:54                       ` Dave Chinner
2016-08-12  2:23                       ` [LKP] [lkp] " Dave Chinner
2016-08-12  2:23                         ` Dave Chinner
2016-08-12  2:32                         ` [LKP] [lkp] " Linus Torvalds
2016-08-12  2:32                           ` Linus Torvalds
2016-08-12  2:52                         ` [LKP] [lkp] " Christoph Hellwig
2016-08-12  2:52                           ` Christoph Hellwig
2016-08-12  3:20                           ` [LKP] [lkp] " Linus Torvalds
2016-08-12  3:20                             ` Linus Torvalds
2016-08-12  4:16                             ` [LKP] [lkp] " Dave Chinner
2016-08-12  4:16                               ` Dave Chinner
2016-08-12  5:02                               ` [LKP] [lkp] " Linus Torvalds
2016-08-12  5:02                                 ` Linus Torvalds
2016-08-12  6:04                                 ` [LKP] [lkp] " Dave Chinner
2016-08-12  6:04                                   ` Dave Chinner
2016-08-12  6:29                                   ` [LKP] [lkp] " Ye Xiaolong
2016-08-12  6:29                                     ` Ye Xiaolong
2016-08-12  8:51                                     ` [LKP] [lkp] " Ye Xiaolong
2016-08-12  8:51                                       ` Ye Xiaolong
2016-08-12 10:02                                       ` [LKP] [lkp] " Dave Chinner
2016-08-12 10:02                                         ` Dave Chinner
2016-08-12 10:43                                         ` Fengguang Wu
2016-08-12 10:43                                           ` Fengguang Wu
2016-08-13  0:30                                         ` [LKP] [lkp] " Christoph Hellwig
2016-08-13  0:30                                           ` Christoph Hellwig
2016-08-13 21:48                                           ` [LKP] [lkp] " Christoph Hellwig
2016-08-13 21:48                                             ` Christoph Hellwig
2016-08-13 22:07                                             ` [LKP] [lkp] " Fengguang Wu
2016-08-13 22:07                                               ` Fengguang Wu
2016-08-13 22:15                                               ` [LKP] [lkp] " Christoph Hellwig
2016-08-13 22:15                                                 ` Christoph Hellwig
2016-08-13 22:51                                                 ` [LKP] [lkp] " Fengguang Wu
2016-08-13 22:51                                                   ` Fengguang Wu
2016-08-14 14:50                                                   ` [LKP] [lkp] " Fengguang Wu
2016-08-14 14:50                                                     ` Fengguang Wu
2016-08-14 16:17                                                     ` [LKP] [lkp] " Christoph Hellwig
2016-08-14 16:17                                                       ` Christoph Hellwig
2016-08-14 23:46                                                       ` [LKP] [lkp] " Dave Chinner
2016-08-14 23:46                                                         ` Dave Chinner
2016-08-14 23:57                                                       ` [LKP] [lkp] " Fengguang Wu
2016-08-14 23:57                                                         ` Fengguang Wu
2016-08-15 14:14                                                       ` Fengguang Wu [this message]
2016-08-15 14:14                                                         ` Fengguang Wu
2016-08-15 21:22                                                         ` [LKP] [lkp] " Dave Chinner
2016-08-15 21:22                                                           ` Dave Chinner
2016-08-16 12:20                                                           ` [LKP] [lkp] " Fengguang Wu
2016-08-16 12:20                                                             ` Fengguang Wu
2016-08-15 20:30                                                       ` [LKP] [lkp] " Huang, Ying
2016-08-15 20:30                                                         ` Huang, Ying
2016-08-22 22:09                                                         ` [LKP] [lkp] " Huang, Ying
2016-08-22 22:09                                                           ` Huang, Ying
2016-09-26  6:25                                                           ` [LKP] [lkp] " Huang, Ying
2016-09-26  6:25                                                             ` Huang, Ying
2016-09-26 14:55                                                             ` [LKP] [lkp] " Christoph Hellwig
2016-09-26 14:55                                                               ` Christoph Hellwig
2016-09-27  0:52                                                               ` [LKP] [lkp] " Huang, Ying
2016-09-27  0:52                                                                 ` Huang, Ying
2016-08-16 13:25                                                       ` [LKP] [lkp] " Fengguang Wu
2016-08-16 13:25                                                         ` Fengguang Wu
2016-08-13 23:32                                           ` [LKP] [lkp] " Dave Chinner
2016-08-13 23:32                                             ` Dave Chinner
2016-08-12  2:27                       ` [LKP] [lkp] " Linus Torvalds
2016-08-12  2:27                         ` Linus Torvalds
2016-08-12  3:56                         ` [LKP] [lkp] " Dave Chinner
2016-08-12  3:56                           ` Dave Chinner
2016-08-12 18:03                           ` [LKP] [lkp] " Linus Torvalds
2016-08-12 18:03                             ` Linus Torvalds
2016-08-13 23:58                             ` [LKP] [lkp] " Fengguang Wu
2016-08-13 23:58                               ` Fengguang Wu
2016-08-15  0:48                             ` [LKP] [lkp] " Dave Chinner
2016-08-15  0:48                               ` Dave Chinner
2016-08-15  1:37                               ` [LKP] [lkp] " Linus Torvalds
2016-08-15  1:37                                 ` Linus Torvalds
2016-08-15  2:28                                 ` [LKP] [lkp] " Dave Chinner
2016-08-15  2:28                                   ` Dave Chinner
2016-08-15  2:53                                   ` [LKP] [lkp] " Linus Torvalds
2016-08-15  2:53                                     ` Linus Torvalds
2016-08-15  5:00                                     ` [LKP] [lkp] " Dave Chinner
2016-08-15  5:00                                       ` Dave Chinner
     [not found]                                       ` <CA+55aFwva2Xffai+Eqv1Jn_NGryk3YJ2i5JoHOQnbQv6qVPAsw@mail.gmail.com>
     [not found]                                         ` <CA+55aFy14nUnJQ_GdF=j8Fa9xiH70c6fY2G3q5HQ01+8z1z3qQ@mail.gmail.com>
2016-08-15  5:12                                           ` Linus Torvalds
2016-08-15 22:22                                             ` [LKP] [lkp] " Dave Chinner
2016-08-15 22:22                                               ` Dave Chinner
2016-08-15 22:42                                               ` [LKP] [lkp] " Dave Chinner
2016-08-15 22:42                                                 ` Dave Chinner
2016-08-15 23:20                                                 ` [LKP] [lkp] " Linus Torvalds
2016-08-15 23:20                                                   ` Linus Torvalds
2016-08-15 23:48                                                   ` [LKP] [lkp] " Linus Torvalds
2016-08-15 23:48                                                     ` Linus Torvalds
2016-08-16  0:44                                                     ` [LKP] [lkp] " Dave Chinner
2016-08-16  0:44                                                       ` Dave Chinner
2016-08-16 15:05                                                     ` [LKP] [lkp] " Mel Gorman
2016-08-16 15:05                                                       ` Mel Gorman
2016-08-16 17:47                                                       ` [LKP] [lkp] " Linus Torvalds
2016-08-16 17:47                                                         ` Linus Torvalds
2016-08-17 15:48                                                         ` [LKP] [lkp] " Michal Hocko
2016-08-17 15:48                                                           ` Michal Hocko
2016-08-17 16:42                                                           ` [LKP] [lkp] " Michal Hocko
2016-08-17 16:42                                                             ` Michal Hocko
2016-08-17 15:49                                                         ` [LKP] [lkp] " Mel Gorman
2016-08-17 15:49                                                           ` Mel Gorman
2016-08-18  0:45                                                           ` [LKP] [lkp] " Mel Gorman
2016-08-18  0:45                                                             ` Mel Gorman
2016-08-18  7:11                                                             ` [LKP] [lkp] " Dave Chinner
2016-08-18  7:11                                                               ` Dave Chinner
2016-08-18 13:24                                                               ` [LKP] [lkp] " Mel Gorman
2016-08-18 13:24                                                                 ` Mel Gorman
2016-08-18 17:55                                                                 ` [LKP] [lkp] " Linus Torvalds
2016-08-18 17:55                                                                   ` Linus Torvalds
2016-08-18 21:19                                                                   ` [LKP] [lkp] " Dave Chinner
2016-08-18 21:19                                                                     ` Dave Chinner
2016-08-18 22:25                                                                     ` [LKP] [lkp] " Linus Torvalds
2016-08-18 22:25                                                                       ` Linus Torvalds
2016-08-19  9:00                                                                       ` [LKP] [lkp] " Michal Hocko
2016-08-19  9:00                                                                         ` Michal Hocko
2016-08-19 10:49                                                                       ` [LKP] [lkp] " Mel Gorman
2016-08-19 10:49                                                                         ` Mel Gorman
2016-08-19 23:48                                                                         ` [LKP] [lkp] " Dave Chinner
2016-08-19 23:48                                                                           ` Dave Chinner
2016-08-20  1:08                                                                           ` [LKP] [lkp] " Linus Torvalds
2016-08-20  1:08                                                                             ` Linus Torvalds
2016-08-20 12:16                                                                           ` [LKP] [lkp] " Mel Gorman
2016-08-20 12:16                                                                             ` Mel Gorman
2016-08-19 15:08                                                               ` [LKP] [lkp] " Mel Gorman
2016-08-19 15:08                                                                 ` Mel Gorman
2016-09-01 23:32                                                                 ` [LKP] [lkp] " Dave Chinner
2016-09-01 23:32                                                                   ` Dave Chinner
2016-09-06 15:37                                                                   ` [LKP] [lkp] " Mel Gorman
2016-09-06 15:37                                                                     ` Mel Gorman
2016-09-06 15:52                                                                     ` [LKP] [lkp] " Huang, Ying
2016-09-06 15:52                                                                       ` Huang, Ying
2016-08-24 15:40                                                             ` [LKP] [lkp] " Huang, Ying
2016-08-24 15:40                                                               ` Huang, Ying
2016-08-25  9:37                                                               ` [LKP] [lkp] " Mel Gorman
2016-08-25  9:37                                                                 ` Mel Gorman
2016-08-18  2:44                                                           ` [LKP] [lkp] " Dave Chinner
2016-08-18  2:44                                                             ` Dave Chinner
2016-08-16  0:15                                                   ` [LKP] [lkp] " Linus Torvalds
2016-08-16  0:15                                                     ` Linus Torvalds
2016-08-16  0:38                                                     ` [LKP] [lkp] " Dave Chinner
2016-08-16  0:38                                                       ` Dave Chinner
2016-08-16  0:50                                                       ` [LKP] [lkp] " Linus Torvalds
2016-08-16  0:50                                                         ` Linus Torvalds
2016-08-16  0:19                                                   ` [LKP] [lkp] " Dave Chinner
2016-08-16  0:19                                                     ` Dave Chinner
2016-08-16  1:51                                                     ` [LKP] [lkp] " Linus Torvalds
2016-08-16  1:51                                                       ` Linus Torvalds
2016-08-16 22:02                                                       ` [LKP] [lkp] " Dave Chinner
2016-08-16 22:02                                                         ` Dave Chinner
2016-08-16 23:23                                                         ` [LKP] [lkp] " Linus Torvalds
2016-08-16 23:23                                                           ` Linus Torvalds
2016-08-15 23:01                                               ` [LKP] [lkp] " Linus Torvalds
2016-08-15 23:01                                                 ` Linus Torvalds
2016-08-16  0:17                                                 ` [LKP] [lkp] " Dave Chinner
2016-08-16  0:17                                                   ` Dave Chinner
2016-08-16  0:45                                                   ` [LKP] [lkp] " Linus Torvalds
2016-08-16  0:45                                                     ` Linus Torvalds
2016-08-15  5:03                                     ` [LKP] [lkp] " Ingo Molnar
2016-08-15  5:03                                       ` Ingo Molnar
2016-08-17 16:24                                       ` [LKP] [lkp] " Peter Zijlstra
2016-08-17 16:24                                         ` Peter Zijlstra
2016-08-15 12:58                             ` [LKP] [lkp] " Fengguang Wu
2016-08-15 12:58                               ` Fengguang Wu
2016-08-11  1:16               ` [LKP] [lkp] " Dave Chinner
2016-08-11  1:16                 ` Dave Chinner
2016-08-11  1:32                 ` [LKP] [lkp] " Dave Chinner
2016-08-11  1:32                   ` Dave Chinner
2016-08-11  2:36                   ` [LKP] [lkp] " Ye Xiaolong
2016-08-11  2:36                     ` Ye Xiaolong
2016-08-11  3:05                     ` [LKP] [lkp] " Dave Chinner
2016-08-11  3:05                       ` Dave Chinner
2016-08-12  1:26                 ` [LKP] [lkp] " Dave Chinner
2016-08-12  1:26                   ` Dave Chinner

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20160815141455.GA22903@wfg-t540p.sh.intel.com \
    --to=fengguang.wu@intel.com \
    --cc=david@fromorbit.com \
    --cc=hch@lst.de \
    --cc=linux-kernel@vger.kernel.org \
    --cc=lkp@01.org \
    --cc=rpeterso@redhat.com \
    --cc=torvalds@linux-foundation.org \
    --cc=xiaolong.ye@intel.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.