All of lore.kernel.org
 help / color / mirror / Atom feed
From: Haiyan Song <haiyanx.song@intel.com>
To: Laurent Dufour <ldufour@linux.vnet.ibm.com>
Cc: "akpm@linux-foundation.org" <akpm@linux-foundation.org>,
	"mhocko@kernel.org" <mhocko@kernel.org>,
	"peterz@infradead.org" <peterz@infradead.org>,
	"kirill@shutemov.name" <kirill@shutemov.name>,
	"ak@linux.intel.com" <ak@linux.intel.com>,
	"dave@stgolabs.net" <dave@stgolabs.net>,
	"jack@suse.cz" <jack@suse.cz>,
	Matthew Wilcox <willy@infradead.org>,
	"khandual@linux.vnet.ibm.com" <khandual@linux.vnet.ibm.com>,
	"aneesh.kumar@linux.vnet.ibm.com"
	<aneesh.kumar@linux.vnet.ibm.com>,
	"benh@kernel.crashing.org" <benh@kernel.crashing.org>,
	"mpe@ellerman.id.au" <mpe@ellerman.id.au>,
	"paulus@samba.org" <paulus@samba.org>,
	Thomas Gleixner <tglx@linutronix.de>,
	Ingo Molnar <mingo@redhat.com>, "hpa@zytor.com" <hpa@zytor.com>,
	Will Deacon <will.deacon@arm.com>,
	Sergey Senozhatsky <sergey.senozhatsky@gmail.com>,
	"sergey.senozhatsky.work@gmail.com"
	<sergey.senozhatsky.work@gmail.com>,
	Andrea Arcangeli <aarcange@redhat.com>,
	Alexei Starovoitov <alexei.starovoitov@gmail.com>,
	"Wang, Kemi" <kemi.wang@intel.com>,
	Daniel Jordan <daniel.m.jordan@oracle.com>,
	David Rientjes <rientjes@google.com>,
	Jerome Glisse <jglisse@redhat.com>,
	Ganesh Mahendran <opensource.ganesh@gmail.com>,
	Minchan Kim <minchan@kernel.org>,
	Punit Agrawal <punitagrawal@gmail.com>,
	vinayak menon <vinayakm.list@gmail.com>,
	Yang Shi <yang.shi@linux.alibaba.com>,
	"linux-kernel@vger.kernel.org" <linux-kernel@vger.kernel.org>,
	"linux-mm@kvack.org" <linux-mm@kvack.org>,
	"haren@linux.vnet.ibm.com" <haren@linux.vnet.ibm.com>,
	"npiggin@gmail.com" <npiggin@gmail.com>,
	"bsingharora@gmail.com" <bsingharora@gmail.com>,
	"paulmck@linux.vnet.ibm.com" <paulmck@linux.vnet.ibm.com>,
	Tim Chen <tim.c.chen@linux.intel.com>,
	"linuxppc-dev@lists.ozlabs.org" <linuxppc-dev@lists.ozlabs.org>,
	"x86@kernel.org" <x86@kernel.org>
Subject: Re: [PATCH v11 00/26] Speculative page faults
Date: Tue, 19 Jun 2018 17:16:25 +0800	[thread overview]
Message-ID: <20180619091625.jfj7qwueisddas2h@haiyan.lkp.sh.intel.com> (raw)
In-Reply-To: <b13302ab-da49-22a4-eda7-68a6528fb9e7@linux.vnet.ibm.com>

[-- Attachment #1: Type: text/plain, Size: 31691 bytes --]

On Mon, Jun 11, 2018 at 05:15:22PM +0200, Laurent Dufour wrote:

Hi Laurent,

For perf date tested on Intel 4s Skylake platform, here attached the compare result
between base and head commit in attachment, which include the perf-profile comparision information.

And also attached some perf-profile.json captured from test result for page_fault2 and page_fault3 for
checking the regression, thanks.


Best regards,
Haiyan Song



> Hi Haiyan,
> 
> I don't have access to the same hardware you ran the test on, but I give a try
> to those test on a Power8 system (2 sockets, 5 cores/s, 8 threads/c, 80 CPUs 32G).
> I run each will-it-scale test 10 times and compute the average.
> 
> test THP enabled		4.17.0-rc4-mm1	spf		delta
> page_fault3_threads		2697.7		2683.5		-0.53%
> page_fault2_threads		170660.6	169574.1	-0.64%
> context_switch1_threads		6915269.2	6877507.3	-0.55%
> context_switch1_processes	6478076.2	6529493.5	0.79%
> rk1				243391.2	238527.5	-2.00%
> 
> Test were launched with the arguments '-t 80 -s 5', only the average report is
> taken in account. Note that page size is 64K by default on ppc64.
> 
> It would be nice if you could capture some perf data to figure out why the
> page_fault2/3 are showing such a performance regression.
> 
> Thanks,
> Laurent.
> 
> On 11/06/2018 09:49, Song, HaiyanX wrote:
> > Hi Laurent,
> > 
> > Regression test for v11 patch serials have been run, some regression is found by LKP-tools (linux kernel performance)
> > tested on Intel 4s skylake platform. This time only test the cases which have been run and found regressions on
> > V9 patch serials.
> > 
> > The regression result is sorted by the metric will-it-scale.per_thread_ops.
> > branch: Laurent-Dufour/Speculative-page-faults/20180520-045126
> > commit id:
> >   head commit : a7a8993bfe3ccb54ad468b9f1799649e4ad1ff12
> >   base commit : ba98a1cdad71d259a194461b3a61471b49b14df1
> > Benchmark: will-it-scale
> > Download link: https://github.com/antonblanchard/will-it-scale/tree/master
> > 
> > Metrics:
> >   will-it-scale.per_process_ops=processes/nr_cpu
> >   will-it-scale.per_thread_ops=threads/nr_cpu
> >   test box: lkp-skl-4sp1(nr_cpu=192,memory=768G)
> > THP: enable / disable
> > nr_task:100%
> > 
> > 1. Regressions:
> > 
> > a). Enable THP
> > testcase                          base           change      head           metric
> > page_fault3/enable THP           10519          -20.5%       8368      will-it-scale.per_thread_ops
> > page_fault2/enalbe THP            8281          -18.8%       6728      will-it-scale.per_thread_ops
> > brk1/eanble THP                 998475           -2.2%     976893      will-it-scale.per_process_ops
> > context_switch1/enable THP      223910           -1.3%     220930      will-it-scale.per_process_ops
> > context_switch1/enable THP      233722           -1.0%     231288      will-it-scale.per_thread_ops
> > 
> > b). Disable THP
> > page_fault3/disable THP          10856          -23.1%       8344      will-it-scale.per_thread_ops
> > page_fault2/disable THP           8147          -18.8%       6613      will-it-scale.per_thread_ops
> > brk1/disable THP                   957           -7.9%        881      will-it-scale.per_thread_ops
> > context_switch1/disable THP     237006           -2.2%     231907      will-it-scale.per_thread_ops
> > brk1/disable THP                997317           -2.0%     977778      will-it-scale.per_process_ops
> > page_fault3/disable THP         467454           -1.8%     459251      will-it-scale.per_process_ops
> > context_switch1/disable THP     224431           -1.3%     221567      will-it-scale.per_process_ops
> > 
> > Notes: for the above  values of test result, the higher is better.
> > 
> > 2. Improvement: not found improvement based on the selected test cases.
> > 
> > 
> > Best regards
> > Haiyan Song
> > ________________________________________
> > From: owner-linux-mm@kvack.org [owner-linux-mm@kvack.org] on behalf of Laurent Dufour [ldufour@linux.vnet.ibm.com]
> > Sent: Monday, May 28, 2018 4:54 PM
> > To: Song, HaiyanX
> > Cc: akpm@linux-foundation.org; mhocko@kernel.org; peterz@infradead.org; kirill@shutemov.name; ak@linux.intel.com; dave@stgolabs.net; jack@suse.cz; Matthew Wilcox; khandual@linux.vnet.ibm.com; aneesh.kumar@linux.vnet.ibm.com; benh@kernel.crashing.org; mpe@ellerman.id.au; paulus@samba.org; Thomas Gleixner; Ingo Molnar; hpa@zytor.com; Will Deacon; Sergey Senozhatsky; sergey.senozhatsky.work@gmail.com; Andrea Arcangeli; Alexei Starovoitov; Wang, Kemi; Daniel Jordan; David Rientjes; Jerome Glisse; Ganesh Mahendran; Minchan Kim; Punit Agrawal; vinayak menon; Yang Shi; linux-kernel@vger.kernel.org; linux-mm@kvack.org; haren@linux.vnet.ibm.com; npiggin@gmail.com; bsingharora@gmail.com; paulmck@linux.vnet.ibm.com; Tim Chen; linuxppc-dev@lists.ozlabs.org; x86@kernel.org
> > Subject: Re: [PATCH v11 00/26] Speculative page faults
> > 
> > On 28/05/2018 10:22, Haiyan Song wrote:
> >> Hi Laurent,
> >>
> >> Yes, these tests are done on V9 patch.
> > 
> > Do you plan to give this V11 a run ?
> > 
> >>
> >>
> >> Best regards,
> >> Haiyan Song
> >>
> >> On Mon, May 28, 2018 at 09:51:34AM +0200, Laurent Dufour wrote:
> >>> On 28/05/2018 07:23, Song, HaiyanX wrote:
> >>>>
> >>>> Some regression and improvements is found by LKP-tools(linux kernel performance) on V9 patch series
> >>>> tested on Intel 4s Skylake platform.
> >>>
> >>> Hi,
> >>>
> >>> Thanks for reporting this benchmark results, but you mentioned the "V9 patch
> >>> series" while responding to the v11 header series...
> >>> Were these tests done on v9 or v11 ?
> >>>
> >>> Cheers,
> >>> Laurent.
> >>>
> >>>>
> >>>> The regression result is sorted by the metric will-it-scale.per_thread_ops.
> >>>> Branch: Laurent-Dufour/Speculative-page-faults/20180316-151833 (V9 patch series)
> >>>> Commit id:
> >>>>     base commit: d55f34411b1b126429a823d06c3124c16283231f
> >>>>     head commit: 0355322b3577eeab7669066df42c550a56801110
> >>>> Benchmark suite: will-it-scale
> >>>> Download link:
> >>>> https://github.com/antonblanchard/will-it-scale/tree/master/tests
> >>>> Metrics:
> >>>>     will-it-scale.per_process_ops=processes/nr_cpu
> >>>>     will-it-scale.per_thread_ops=threads/nr_cpu
> >>>> test box: lkp-skl-4sp1(nr_cpu=192,memory=768G)
> >>>> THP: enable / disable
> >>>> nr_task: 100%
> >>>>
> >>>> 1. Regressions:
> >>>> a) THP enabled:
> >>>> testcase                        base            change          head       metric
> >>>> page_fault3/ enable THP         10092           -17.5%          8323       will-it-scale.per_thread_ops
> >>>> page_fault2/ enable THP          8300           -17.2%          6869       will-it-scale.per_thread_ops
> >>>> brk1/ enable THP                  957.67         -7.6%           885       will-it-scale.per_thread_ops
> >>>> page_fault3/ enable THP        172821            -5.3%        163692       will-it-scale.per_process_ops
> >>>> signal1/ enable THP              9125            -3.2%          8834       will-it-scale.per_process_ops
> >>>>
> >>>> b) THP disabled:
> >>>> testcase                        base            change          head       metric
> >>>> page_fault3/ disable THP        10107           -19.1%          8180       will-it-scale.per_thread_ops
> >>>> page_fault2/ disable THP         8432           -17.8%          6931       will-it-scale.per_thread_ops
> >>>> context_switch1/ disable THP   215389            -6.8%        200776       will-it-scale.per_thread_ops
> >>>> brk1/ disable THP                 939.67         -6.6%           877.33    will-it-scale.per_thread_ops
> >>>> page_fault3/ disable THP       173145            -4.7%        165064       will-it-scale.per_process_ops
> >>>> signal1/ disable THP             9162            -3.9%          8802       will-it-scale.per_process_ops
> >>>>
> >>>> 2. Improvements:
> >>>> a) THP enabled:
> >>>> testcase                        base            change          head       metric
> >>>> malloc1/ enable THP               66.33        +469.8%           383.67    will-it-scale.per_thread_ops
> >>>> writeseek3/ enable THP          2531             +4.5%          2646       will-it-scale.per_thread_ops
> >>>> signal1/ enable THP              989.33          +2.8%          1016       will-it-scale.per_thread_ops
> >>>>
> >>>> b) THP disabled:
> >>>> testcase                        base            change          head       metric
> >>>> malloc1/ disable THP              90.33        +417.3%           467.33    will-it-scale.per_thread_ops
> >>>> read2/ disable THP             58934            +39.2%         82060       will-it-scale.per_thread_ops
> >>>> page_fault1/ disable THP        8607            +36.4%         11736       will-it-scale.per_thread_ops
> >>>> read1/ disable THP            314063            +12.7%        353934       will-it-scale.per_thread_ops
> >>>> writeseek3/ disable THP         2452            +12.5%          2759       will-it-scale.per_thread_ops
> >>>> signal1/ disable THP             971.33          +5.5%          1024       will-it-scale.per_thread_ops
> >>>>
> >>>> Notes: for above values in column "change", the higher value means that the related testcase result
> >>>> on head commit is better than that on base commit for this benchmark.
> >>>>
> >>>>
> >>>> Best regards
> >>>> Haiyan Song
> >>>>
> >>>> ________________________________________
> >>>> From: owner-linux-mm@kvack.org [owner-linux-mm@kvack.org] on behalf of Laurent Dufour [ldufour@linux.vnet.ibm.com]
> >>>> Sent: Thursday, May 17, 2018 7:06 PM
> >>>> To: akpm@linux-foundation.org; mhocko@kernel.org; peterz@infradead.org; kirill@shutemov.name; ak@linux.intel.com; dave@stgolabs.net; jack@suse.cz; Matthew Wilcox; khandual@linux.vnet.ibm.com; aneesh.kumar@linux.vnet.ibm.com; benh@kernel.crashing.org; mpe@ellerman.id.au; paulus@samba.org; Thomas Gleixner; Ingo Molnar; hpa@zytor.com; Will Deacon; Sergey Senozhatsky; sergey.senozhatsky.work@gmail.com; Andrea Arcangeli; Alexei Starovoitov; Wang, Kemi; Daniel Jordan; David Rientjes; Jerome Glisse; Ganesh Mahendran; Minchan Kim; Punit Agrawal; vinayak menon; Yang Shi
> >>>> Cc: linux-kernel@vger.kernel.org; linux-mm@kvack.org; haren@linux.vnet.ibm.com; npiggin@gmail.com; bsingharora@gmail.com; paulmck@linux.vnet.ibm.com; Tim Chen; linuxppc-dev@lists.ozlabs.org; x86@kernel.org
> >>>> Subject: [PATCH v11 00/26] Speculative page faults
> >>>>
> >>>> This is a port on kernel 4.17 of the work done by Peter Zijlstra to handle
> >>>> page fault without holding the mm semaphore [1].
> >>>>
> >>>> The idea is to try to handle user space page faults without holding the
> >>>> mmap_sem. This should allow better concurrency for massively threaded
> >>>> process since the page fault handler will not wait for other threads memory
> >>>> layout change to be done, assuming that this change is done in another part
> >>>> of the process's memory space. This type page fault is named speculative
> >>>> page fault. If the speculative page fault fails because of a concurrency is
> >>>> detected or because underlying PMD or PTE tables are not yet allocating, it
> >>>> is failing its processing and a classic page fault is then tried.
> >>>>
> >>>> The speculative page fault (SPF) has to look for the VMA matching the fault
> >>>> address without holding the mmap_sem, this is done by introducing a rwlock
> >>>> which protects the access to the mm_rb tree. Previously this was done using
> >>>> SRCU but it was introducing a lot of scheduling to process the VMA's
> >>>> freeing operation which was hitting the performance by 20% as reported by
> >>>> Kemi Wang [2]. Using a rwlock to protect access to the mm_rb tree is
> >>>> limiting the locking contention to these operations which are expected to
> >>>> be in a O(log n) order. In addition to ensure that the VMA is not freed in
> >>>> our back a reference count is added and 2 services (get_vma() and
> >>>> put_vma()) are introduced to handle the reference count. Once a VMA is
> >>>> fetched from the RB tree using get_vma(), it must be later freed using
> >>>> put_vma(). I can't see anymore the overhead I got while will-it-scale
> >>>> benchmark anymore.
> >>>>
> >>>> The VMA's attributes checked during the speculative page fault processing
> >>>> have to be protected against parallel changes. This is done by using a per
> >>>> VMA sequence lock. This sequence lock allows the speculative page fault
> >>>> handler to fast check for parallel changes in progress and to abort the
> >>>> speculative page fault in that case.
> >>>>
> >>>> Once the VMA has been found, the speculative page fault handler would check
> >>>> for the VMA's attributes to verify that the page fault has to be handled
> >>>> correctly or not. Thus, the VMA is protected through a sequence lock which
> >>>> allows fast detection of concurrent VMA changes. If such a change is
> >>>> detected, the speculative page fault is aborted and a *classic* page fault
> >>>> is tried.  VMA sequence lockings are added when VMA attributes which are
> >>>> checked during the page fault are modified.
> >>>>
> >>>> When the PTE is fetched, the VMA is checked to see if it has been changed,
> >>>> so once the page table is locked, the VMA is valid, so any other changes
> >>>> leading to touching this PTE will need to lock the page table, so no
> >>>> parallel change is possible at this time.
> >>>>
> >>>> The locking of the PTE is done with interrupts disabled, this allows
> >>>> checking for the PMD to ensure that there is not an ongoing collapsing
> >>>> operation. Since khugepaged is firstly set the PMD to pmd_none and then is
> >>>> waiting for the other CPU to have caught the IPI interrupt, if the pmd is
> >>>> valid at the time the PTE is locked, we have the guarantee that the
> >>>> collapsing operation will have to wait on the PTE lock to move forward.
> >>>> This allows the SPF handler to map the PTE safely. If the PMD value is
> >>>> different from the one recorded at the beginning of the SPF operation, the
> >>>> classic page fault handler will be called to handle the operation while
> >>>> holding the mmap_sem. As the PTE lock is done with the interrupts disabled,
> >>>> the lock is done using spin_trylock() to avoid dead lock when handling a
> >>>> page fault while a TLB invalidate is requested by another CPU holding the
> >>>> PTE.
> >>>>
> >>>> In pseudo code, this could be seen as:
> >>>>     speculative_page_fault()
> >>>>     {
> >>>>             vma = get_vma()
> >>>>             check vma sequence count
> >>>>             check vma's support
> >>>>             disable interrupt
> >>>>                   check pgd,p4d,...,pte
> >>>>                   save pmd and pte in vmf
> >>>>                   save vma sequence counter in vmf
> >>>>             enable interrupt
> >>>>             check vma sequence count
> >>>>             handle_pte_fault(vma)
> >>>>                     ..
> >>>>                     page = alloc_page()
> >>>>                     pte_map_lock()
> >>>>                             disable interrupt
> >>>>                                     abort if sequence counter has changed
> >>>>                                     abort if pmd or pte has changed
> >>>>                                     pte map and lock
> >>>>                             enable interrupt
> >>>>                     if abort
> >>>>                        free page
> >>>>                        abort
> >>>>                     ...
> >>>>     }
> >>>>
> >>>>     arch_fault_handler()
> >>>>     {
> >>>>             if (speculative_page_fault(&vma))
> >>>>                goto done
> >>>>     again:
> >>>>             lock(mmap_sem)
> >>>>             vma = find_vma();
> >>>>             handle_pte_fault(vma);
> >>>>             if retry
> >>>>                unlock(mmap_sem)
> >>>>                goto again;
> >>>>     done:
> >>>>             handle fault error
> >>>>     }
> >>>>
> >>>> Support for THP is not done because when checking for the PMD, we can be
> >>>> confused by an in progress collapsing operation done by khugepaged. The
> >>>> issue is that pmd_none() could be true either if the PMD is not already
> >>>> populated or if the underlying PTE are in the way to be collapsed. So we
> >>>> cannot safely allocate a PMD if pmd_none() is true.
> >>>>
> >>>> This series add a new software performance event named 'speculative-faults'
> >>>> or 'spf'. It counts the number of successful page fault event handled
> >>>> speculatively. When recording 'faults,spf' events, the faults one is
> >>>> counting the total number of page fault events while 'spf' is only counting
> >>>> the part of the faults processed speculatively.
> >>>>
> >>>> There are some trace events introduced by this series. They allow
> >>>> identifying why the page faults were not processed speculatively. This
> >>>> doesn't take in account the faults generated by a monothreaded process
> >>>> which directly processed while holding the mmap_sem. This trace events are
> >>>> grouped in a system named 'pagefault', they are:
> >>>>  - pagefault:spf_vma_changed : if the VMA has been changed in our back
> >>>>  - pagefault:spf_vma_noanon : the vma->anon_vma field was not yet set.
> >>>>  - pagefault:spf_vma_notsup : the VMA's type is not supported
> >>>>  - pagefault:spf_vma_access : the VMA's access right are not respected
> >>>>  - pagefault:spf_pmd_changed : the upper PMD pointer has changed in our
> >>>>    back.
> >>>>
> >>>> To record all the related events, the easier is to run perf with the
> >>>> following arguments :
> >>>> $ perf stat -e 'faults,spf,pagefault:*' <command>
> >>>>
> >>>> There is also a dedicated vmstat counter showing the number of successful
> >>>> page fault handled speculatively. I can be seen this way:
> >>>> $ grep speculative_pgfault /proc/vmstat
> >>>>
> >>>> This series builds on top of v4.16-mmotm-2018-04-13-17-28 and is functional
> >>>> on x86, PowerPC and arm64.
> >>>>
> >>>> ---------------------
> >>>> Real Workload results
> >>>>
> >>>> As mentioned in previous email, we did non official runs using a "popular
> >>>> in memory multithreaded database product" on 176 cores SMT8 Power system
> >>>> which showed a 30% improvements in the number of transaction processed per
> >>>> second. This run has been done on the v6 series, but changes introduced in
> >>>> this new version should not impact the performance boost seen.
> >>>>
> >>>> Here are the perf data captured during 2 of these runs on top of the v8
> >>>> series:
> >>>>                 vanilla         spf
> >>>> faults          89.418          101.364         +13%
> >>>> spf                n/a           97.989
> >>>>
> >>>> With the SPF kernel, most of the page fault were processed in a speculative
> >>>> way.
> >>>>
> >>>> Ganesh Mahendran had backported the series on top of a 4.9 kernel and gave
> >>>> it a try on an android device. He reported that the application launch time
> >>>> was improved in average by 6%, and for large applications (~100 threads) by
> >>>> 20%.
> >>>>
> >>>> Here are the launch time Ganesh mesured on Android 8.0 on top of a Qcom
> >>>> MSM845 (8 cores) with 6GB (the less is better):
> >>>>
> >>>> Application                             4.9     4.9+spf delta
> >>>> com.tencent.mm                          416     389     -7%
> >>>> com.eg.android.AlipayGphone             1135    986     -13%
> >>>> com.tencent.mtt                         455     454     0%
> >>>> com.qqgame.hlddz                        1497    1409    -6%
> >>>> com.autonavi.minimap                    711     701     -1%
> >>>> com.tencent.tmgp.sgame                  788     748     -5%
> >>>> com.immomo.momo                         501     487     -3%
> >>>> com.tencent.peng                        2145    2112    -2%
> >>>> com.smile.gifmaker                      491     461     -6%
> >>>> com.baidu.BaiduMap                      479     366     -23%
> >>>> com.taobao.taobao                       1341    1198    -11%
> >>>> com.baidu.searchbox                     333     314     -6%
> >>>> com.tencent.mobileqq                    394     384     -3%
> >>>> com.sina.weibo                          907     906     0%
> >>>> com.youku.phone                         816     731     -11%
> >>>> com.happyelements.AndroidAnimal.qq      763     717     -6%
> >>>> com.UCMobile                            415     411     -1%
> >>>> com.tencent.tmgp.ak                     1464    1431    -2%
> >>>> com.tencent.qqmusic                     336     329     -2%
> >>>> com.sankuai.meituan                     1661    1302    -22%
> >>>> com.netease.cloudmusic                  1193    1200    1%
> >>>> air.tv.douyu.android                    4257    4152    -2%
> >>>>
> >>>> ------------------
> >>>> Benchmarks results
> >>>>
> >>>> Base kernel is v4.17.0-rc4-mm1
> >>>> SPF is BASE + this series
> >>>>
> >>>> Kernbench:
> >>>> ----------
> >>>> Here are the results on a 16 CPUs X86 guest using kernbench on a 4.15
> >>>> kernel (kernel is build 5 times):
> >>>>
> >>>> Average Half load -j 8
> >>>>                  Run    (std deviation)
> >>>>                  BASE                   SPF
> >>>> Elapsed Time     1448.65 (5.72312)      1455.84 (4.84951)       0.50%
> >>>> User    Time     10135.4 (30.3699)      10148.8 (31.1252)       0.13%
> >>>> System  Time     900.47  (2.81131)      923.28  (7.52779)       2.53%
> >>>> Percent CPU      761.4   (1.14018)      760.2   (0.447214)      -0.16%
> >>>> Context Switches 85380   (3419.52)      84748   (1904.44)       -0.74%
> >>>> Sleeps           105064  (1240.96)      105074  (337.612)       0.01%
> >>>>
> >>>> Average Optimal load -j 16
> >>>>                  Run    (std deviation)
> >>>>                  BASE                   SPF
> >>>> Elapsed Time     920.528 (10.1212)      927.404 (8.91789)       0.75%
> >>>> User    Time     11064.8 (981.142)      11085   (990.897)       0.18%
> >>>> System  Time     979.904 (84.0615)      1001.14 (82.5523)       2.17%
> >>>> Percent CPU      1089.5  (345.894)      1086.1  (343.545)       -0.31%
> >>>> Context Switches 159488  (78156.4)      158223  (77472.1)       -0.79%
> >>>> Sleeps           110566  (5877.49)      110388  (5617.75)       -0.16%
> >>>>
> >>>>
> >>>> During a run on the SPF, perf events were captured:
> >>>>  Performance counter stats for '../kernbench -M':
> >>>>          526743764      faults
> >>>>                210      spf
> >>>>                  3      pagefault:spf_vma_changed
> >>>>                  0      pagefault:spf_vma_noanon
> >>>>               2278      pagefault:spf_vma_notsup
> >>>>                  0      pagefault:spf_vma_access
> >>>>                  0      pagefault:spf_pmd_changed
> >>>>
> >>>> Very few speculative page faults were recorded as most of the processes
> >>>> involved are monothreaded (sounds that on this architecture some threads
> >>>> were created during the kernel build processing).
> >>>>
> >>>> Here are the kerbench results on a 80 CPUs Power8 system:
> >>>>
> >>>> Average Half load -j 40
> >>>>                  Run    (std deviation)
> >>>>                  BASE                   SPF
> >>>> Elapsed Time     117.152 (0.774642)     117.166 (0.476057)      0.01%
> >>>> User    Time     4478.52 (24.7688)      4479.76 (9.08555)       0.03%
> >>>> System  Time     131.104 (0.720056)     134.04  (0.708414)      2.24%
> >>>> Percent CPU      3934    (19.7104)      3937.2  (19.0184)       0.08%
> >>>> Context Switches 92125.4 (576.787)      92581.6 (198.622)       0.50%
> >>>> Sleeps           317923  (652.499)      318469  (1255.59)       0.17%
> >>>>
> >>>> Average Optimal load -j 80
> >>>>                  Run    (std deviation)
> >>>>                  BASE                   SPF
> >>>> Elapsed Time     107.73  (0.632416)     107.31  (0.584936)      -0.39%
> >>>> User    Time     5869.86 (1466.72)      5871.71 (1467.27)       0.03%
> >>>> System  Time     153.728 (23.8573)      157.153 (24.3704)       2.23%
> >>>> Percent CPU      5418.6  (1565.17)      5436.7  (1580.91)       0.33%
> >>>> Context Switches 223861  (138865)       225032  (139632)        0.52%
> >>>> Sleeps           330529  (13495.1)      332001  (14746.2)       0.45%
> >>>>
> >>>> During a run on the SPF, perf events were captured:
> >>>>  Performance counter stats for '../kernbench -M':
> >>>>          116730856      faults
> >>>>                  0      spf
> >>>>                  3      pagefault:spf_vma_changed
> >>>>                  0      pagefault:spf_vma_noanon
> >>>>                476      pagefault:spf_vma_notsup
> >>>>                  0      pagefault:spf_vma_access
> >>>>                  0      pagefault:spf_pmd_changed
> >>>>
> >>>> Most of the processes involved are monothreaded so SPF is not activated but
> >>>> there is no impact on the performance.
> >>>>
> >>>> Ebizzy:
> >>>> -------
> >>>> The test is counting the number of records per second it can manage, the
> >>>> higher is the best. I run it like this 'ebizzy -mTt <nrcpus>'. To get
> >>>> consistent result I repeated the test 100 times and measure the average
> >>>> result. The number is the record processes per second, the higher is the
> >>>> best.
> >>>>
> >>>>                 BASE            SPF             delta
> >>>> 16 CPUs x86 VM  742.57          1490.24         100.69%
> >>>> 80 CPUs P8 node 13105.4         24174.23        84.46%
> >>>>
> >>>> Here are the performance counter read during a run on a 16 CPUs x86 VM:
> >>>>  Performance counter stats for './ebizzy -mTt 16':
> >>>>            1706379      faults
> >>>>            1674599      spf
> >>>>              30588      pagefault:spf_vma_changed
> >>>>                  0      pagefault:spf_vma_noanon
> >>>>                363      pagefault:spf_vma_notsup
> >>>>                  0      pagefault:spf_vma_access
> >>>>                  0      pagefault:spf_pmd_changed
> >>>>
> >>>> And the ones captured during a run on a 80 CPUs Power node:
> >>>>  Performance counter stats for './ebizzy -mTt 80':
> >>>>            1874773      faults
> >>>>            1461153      spf
> >>>>             413293      pagefault:spf_vma_changed
> >>>>                  0      pagefault:spf_vma_noanon
> >>>>                200      pagefault:spf_vma_notsup
> >>>>                  0      pagefault:spf_vma_access
> >>>>                  0      pagefault:spf_pmd_changed
> >>>>
> >>>> In ebizzy's case most of the page fault were handled in a speculative way,
> >>>> leading the ebizzy performance boost.
> >>>>
> >>>> ------------------
> >>>> Changes since v10 (https://lkml.org/lkml/2018/4/17/572):
> >>>>  - Accounted for all review feedbacks from Punit Agrawal, Ganesh Mahendran
> >>>>    and Minchan Kim, hopefully.
> >>>>  - Remove unneeded check on CONFIG_SPECULATIVE_PAGE_FAULT in
> >>>>    __do_page_fault().
> >>>>  - Loop in pte_spinlock() and pte_map_lock() when pte try lock fails
> >>>>    instead
> >>>>    of aborting the speculative page fault handling. Dropping the now
> >>>> useless
> >>>>    trace event pagefault:spf_pte_lock.
> >>>>  - No more try to reuse the fetched VMA during the speculative page fault
> >>>>    handling when retrying is needed. This adds a lot of complexity and
> >>>>    additional tests done didn't show a significant performance improvement.
> >>>>  - Convert IS_ENABLED(CONFIG_NUMA) back to #ifdef due to build error.
> >>>>
> >>>> [1] http://linux-kernel.2935.n7.nabble.com/RFC-PATCH-0-6-Another-go-at-speculative-page-faults-tt965642.html#none
> >>>> [2] https://patchwork.kernel.org/patch/9999687/
> >>>>
> >>>>
> >>>> Laurent Dufour (20):
> >>>>   mm: introduce CONFIG_SPECULATIVE_PAGE_FAULT
> >>>>   x86/mm: define ARCH_SUPPORTS_SPECULATIVE_PAGE_FAULT
> >>>>   powerpc/mm: set ARCH_SUPPORTS_SPECULATIVE_PAGE_FAULT
> >>>>   mm: introduce pte_spinlock for FAULT_FLAG_SPECULATIVE
> >>>>   mm: make pte_unmap_same compatible with SPF
> >>>>   mm: introduce INIT_VMA()
> >>>>   mm: protect VMA modifications using VMA sequence count
> >>>>   mm: protect mremap() against SPF hanlder
> >>>>   mm: protect SPF handler against anon_vma changes
> >>>>   mm: cache some VMA fields in the vm_fault structure
> >>>>   mm/migrate: Pass vm_fault pointer to migrate_misplaced_page()
> >>>>   mm: introduce __lru_cache_add_active_or_unevictable
> >>>>   mm: introduce __vm_normal_page()
> >>>>   mm: introduce __page_add_new_anon_rmap()
> >>>>   mm: protect mm_rb tree with a rwlock
> >>>>   mm: adding speculative page fault failure trace events
> >>>>   perf: add a speculative page fault sw event
> >>>>   perf tools: add support for the SPF perf event
> >>>>   mm: add speculative page fault vmstats
> >>>>   powerpc/mm: add speculative page fault
> >>>>
> >>>> Mahendran Ganesh (2):
> >>>>   arm64/mm: define ARCH_SUPPORTS_SPECULATIVE_PAGE_FAULT
> >>>>   arm64/mm: add speculative page fault
> >>>>
> >>>> Peter Zijlstra (4):
> >>>>   mm: prepare for FAULT_FLAG_SPECULATIVE
> >>>>   mm: VMA sequence count
> >>>>   mm: provide speculative fault infrastructure
> >>>>   x86/mm: add speculative pagefault handling
> >>>>
> >>>>  arch/arm64/Kconfig                    |   1 +
> >>>>  arch/arm64/mm/fault.c                 |  12 +
> >>>>  arch/powerpc/Kconfig                  |   1 +
> >>>>  arch/powerpc/mm/fault.c               |  16 +
> >>>>  arch/x86/Kconfig                      |   1 +
> >>>>  arch/x86/mm/fault.c                   |  27 +-
> >>>>  fs/exec.c                             |   2 +-
> >>>>  fs/proc/task_mmu.c                    |   5 +-
> >>>>  fs/userfaultfd.c                      |  17 +-
> >>>>  include/linux/hugetlb_inline.h        |   2 +-
> >>>>  include/linux/migrate.h               |   4 +-
> >>>>  include/linux/mm.h                    | 136 +++++++-
> >>>>  include/linux/mm_types.h              |   7 +
> >>>>  include/linux/pagemap.h               |   4 +-
> >>>>  include/linux/rmap.h                  |  12 +-
> >>>>  include/linux/swap.h                  |  10 +-
> >>>>  include/linux/vm_event_item.h         |   3 +
> >>>>  include/trace/events/pagefault.h      |  80 +++++
> >>>>  include/uapi/linux/perf_event.h       |   1 +
> >>>>  kernel/fork.c                         |   5 +-
> >>>>  mm/Kconfig                            |  22 ++
> >>>>  mm/huge_memory.c                      |   6 +-
> >>>>  mm/hugetlb.c                          |   2 +
> >>>>  mm/init-mm.c                          |   3 +
> >>>>  mm/internal.h                         |  20 ++
> >>>>  mm/khugepaged.c                       |   5 +
> >>>>  mm/madvise.c                          |   6 +-
> >>>>  mm/memory.c                           | 612 +++++++++++++++++++++++++++++-----
> >>>>  mm/mempolicy.c                        |  51 ++-
> >>>>  mm/migrate.c                          |   6 +-
> >>>>  mm/mlock.c                            |  13 +-
> >>>>  mm/mmap.c                             | 229 ++++++++++---
> >>>>  mm/mprotect.c                         |   4 +-
> >>>>  mm/mremap.c                           |  13 +
> >>>>  mm/nommu.c                            |   2 +-
> >>>>  mm/rmap.c                             |   5 +-
> >>>>  mm/swap.c                             |   6 +-
> >>>>  mm/swap_state.c                       |   8 +-
> >>>>  mm/vmstat.c                           |   5 +-
> >>>>  tools/include/uapi/linux/perf_event.h |   1 +
> >>>>  tools/perf/util/evsel.c               |   1 +
> >>>>  tools/perf/util/parse-events.c        |   4 +
> >>>>  tools/perf/util/parse-events.l        |   1 +
> >>>>  tools/perf/util/python.c              |   1 +
> >>>>  44 files changed, 1161 insertions(+), 211 deletions(-)
> >>>>  create mode 100644 include/trace/events/pagefault.h
> >>>>
> >>>> --
> >>>> 2.7.4
> >>>>
> >>>>
> >>>
> >>
> > 
> 

[-- Attachment #2: compare-result.txt --]
[-- Type: text/plain, Size: 185207 bytes --]

=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/nr_task/thp_enabled/test/cpufreq_governor:
  lkp-skl-4sp1/will-it-scale/debian-x86_64-2018-04-03.cgz/x86_64-rhel-7.2/gcc-7/100%/always/page_fault3/performance

commit: 
  ba98a1cdad71d259a194461b3a61471b49b14df1
  a7a8993bfe3ccb54ad468b9f1799649e4ad1ff12

ba98a1cdad71d259 a7a8993bfe3ccb54ad468b9f17 
---------------- -------------------------- 
       fail:runs  %reproduction    fail:runs
           |             |             |    
         44:3          -13%          43:3     perf-profile.calltrace.cycles-pp.error_entry
         22:3           -6%          22:3     perf-profile.calltrace.cycles-pp.sync_regs.error_entry
         44:3          -13%          44:3     perf-profile.children.cycles-pp.error_entry
         21:3           -7%          21:3     perf-profile.self.cycles-pp.error_entry
         %stddev     %change         %stddev
             \          |                \  
     10519 +-  3%     -20.5%       8368 +-  6%  will-it-scale.per_thread_ops
    118098           +11.2%     131287 +-  2%  will-it-scale.time.involuntary_context_switches
 6.084e+08 +-  3%     -20.4%  4.845e+08 +-  6%  will-it-scale.time.minor_page_faults
      7467            +5.0%       7841        will-it-scale.time.percent_of_cpu_this_job_got
     44922            +5.0%      47176        will-it-scale.time.system_time
   7126337 +-  3%     -15.4%    6025689 +-  6%  will-it-scale.time.voluntary_context_switches
  91905646            -1.3%   90673935        will-it-scale.workload
     27.15 +-  6%      -8.7%      24.80 +- 10%  boot-time.boot
   2516213 +-  6%      +8.3%    2726303        interrupts.CAL:Function_call_interrupts
    388.00 +-  9%     +60.2%     621.67 +- 20%  irq_exception_noise.softirq_nr
     11.28 +-  2%      -1.9        9.37 +-  4%  mpstat.cpu.idle%
     10065 +-140%    +243.4%      34559 +-  4%  numa-numastat.node0.other_node
     18739           -11.6%      16573 +-  3%  uptime.idle
     29406 +-  2%     -11.8%      25929 +-  5%  vmstat.system.cs
    329614 +-  8%     +17.0%     385618 +- 10%  meminfo.DirectMap4k
    237851           +21.2%     288160 +-  5%  meminfo.Inactive
    237615           +21.2%     287924 +-  5%  meminfo.Inactive(anon)
   7917847           -10.7%    7071860        softirqs.RCU
   4784181 +-  3%     -14.5%    4089039 +-  4%  softirqs.SCHED
  45666107 +-  7%     +12.9%   51535472 +-  3%  softirqs.TIMER
 2.617e+09 +-  2%     -13.9%  2.253e+09 +-  6%  cpuidle.C1E.time
   6688774 +-  2%     -12.8%    5835101 +-  5%  cpuidle.C1E.usage
 1.022e+10 +-  2%     -18.0%  8.376e+09 +-  3%  cpuidle.C6.time
  13440993 +-  2%     -16.3%   11243794 +-  4%  cpuidle.C6.usage
     54781 +- 16%     +37.5%      75347 +- 12%  numa-meminfo.node0.Inactive
     54705 +- 16%     +37.7%      75347 +- 12%  numa-meminfo.node0.Inactive(anon)
     52522           +35.0%      70886 +-  6%  numa-meminfo.node2.Inactive
     52443           +34.7%      70653 +-  6%  numa-meminfo.node2.Inactive(anon)
     31046 +-  6%     +30.3%      40457 +- 11%  numa-meminfo.node2.SReclaimable
     58563           +21.1%      70945 +-  6%  proc-vmstat.nr_inactive_anon
     58564           +21.1%      70947 +-  6%  proc-vmstat.nr_zone_inactive_anon
  69701118            -1.2%   68842151        proc-vmstat.pgalloc_normal
 2.765e+10            -1.3%  2.729e+10        proc-vmstat.pgfault
  69330418            -1.2%   68466824        proc-vmstat.pgfree
    118098           +11.2%     131287 +-  2%  time.involuntary_context_switches
 6.084e+08 +-  3%     -20.4%  4.845e+08 +-  6%  time.minor_page_faults
      7467            +5.0%       7841        time.percent_of_cpu_this_job_got
     44922            +5.0%      47176        time.system_time
   7126337 +-  3%     -15.4%    6025689 +-  6%  time.voluntary_context_switches
     13653 +- 16%     +33.5%      18225 +- 12%  numa-vmstat.node0.nr_inactive_anon
     13651 +- 16%     +33.5%      18224 +- 12%  numa-vmstat.node0.nr_zone_inactive_anon
     13069 +-  3%     +30.1%      17001 +-  4%  numa-vmstat.node2.nr_inactive_anon
    134.67 +- 42%     -49.5%      68.00 +- 31%  numa-vmstat.node2.nr_mlock
      7758 +-  6%     +30.4%      10112 +- 11%  numa-vmstat.node2.nr_slab_reclaimable
     13066 +-  3%     +30.1%      16998 +-  4%  numa-vmstat.node2.nr_zone_inactive_anon
      1039 +- 11%     -17.5%     857.33        slabinfo.Acpi-ParseExt.active_objs
      1039 +- 11%     -17.5%     857.33        slabinfo.Acpi-ParseExt.num_objs
      2566 +-  6%      -8.8%       2340 +-  5%  slabinfo.biovec-64.active_objs
      2566 +-  6%      -8.8%       2340 +-  5%  slabinfo.biovec-64.num_objs
    898.33 +-  3%      -9.5%     813.33 +-  3%  slabinfo.kmem_cache_node.active_objs
      1066 +-  2%      -8.0%     981.33 +-  3%  slabinfo.kmem_cache_node.num_objs
      1940            +2.3%       1984        turbostat.Avg_MHz
   6679037 +-  2%     -12.7%    5830270 +-  5%  turbostat.C1E
      2.25 +-  2%      -0.3        1.94 +-  6%  turbostat.C1E%
  13418115           -16.3%   11234510 +-  4%  turbostat.C6
      8.75 +-  2%      -1.6        7.18 +-  3%  turbostat.C6%
      5.99 +-  2%     -14.4%       5.13 +-  4%  turbostat.CPU%c1
      5.01 +-  3%     -20.1%       4.00 +-  4%  turbostat.CPU%c6
      1.77 +-  3%     -34.7%       1.15        turbostat.Pkg%pc2
 1.378e+13            +1.2%  1.394e+13        perf-stat.branch-instructions
      0.98            -0.0        0.94        perf-stat.branch-miss-rate%
 1.344e+11            -2.3%  1.313e+11        perf-stat.branch-misses
 1.076e+11            -1.8%  1.057e+11        perf-stat.cache-misses
 2.258e+11            -2.1%   2.21e+11        perf-stat.cache-references
  17788064 +-  2%     -11.9%   15674207 +-  6%  perf-stat.context-switches
 2.241e+14            +2.4%  2.294e+14        perf-stat.cpu-cycles
 1.929e+13            +2.2%  1.971e+13        perf-stat.dTLB-loads
      4.01            -0.2        3.83        perf-stat.dTLB-store-miss-rate%
 4.519e+11            -1.3%  4.461e+11        perf-stat.dTLB-store-misses
 1.082e+13            +3.6%  1.121e+13        perf-stat.dTLB-stores
  3.02e+10           +23.2%  3.721e+10 +-  3%  perf-stat.iTLB-load-misses
 2.721e+08 +-  8%      -8.8%  2.481e+08 +-  3%  perf-stat.iTLB-loads
 6.985e+13            +1.8%  7.111e+13        perf-stat.instructions
      2313           -17.2%       1914 +-  3%  perf-stat.instructions-per-iTLB-miss
 2.764e+10            -1.3%  2.729e+10        perf-stat.minor-faults
 1.421e+09 +-  2%     -16.4%  1.188e+09 +-  9%  perf-stat.node-load-misses
 1.538e+10            -9.3%  1.395e+10        perf-stat.node-loads
      9.75            +1.4       11.10        perf-stat.node-store-miss-rate%
 3.012e+09           +14.1%  3.437e+09        perf-stat.node-store-misses
 2.789e+10            -1.3%  2.753e+10        perf-stat.node-stores
 2.764e+10            -1.3%  2.729e+10        perf-stat.page-faults
    760059            +3.2%     784235        perf-stat.path-length
    193545 +- 25%     -57.8%      81757 +- 46%  sched_debug.cfs_rq:/.MIN_vruntime.avg
  26516863 +- 19%     -49.7%   13338070 +- 33%  sched_debug.cfs_rq:/.MIN_vruntime.max
   2202271 +- 21%     -53.2%    1029581 +- 38%  sched_debug.cfs_rq:/.MIN_vruntime.stddev
    193545 +- 25%     -57.8%      81757 +- 46%  sched_debug.cfs_rq:/.max_vruntime.avg
  26516863 +- 19%     -49.7%   13338070 +- 33%  sched_debug.cfs_rq:/.max_vruntime.max
   2202271 +- 21%     -53.2%    1029581 +- 38%  sched_debug.cfs_rq:/.max_vruntime.stddev
      0.32 +- 70%    +253.2%       1.14 +- 54%  sched_debug.cfs_rq:/.removed.load_avg.avg
      4.44 +- 70%    +120.7%       9.80 +- 27%  sched_debug.cfs_rq:/.removed.load_avg.stddev
     14.90 +- 70%    +251.0%      52.31 +- 53%  sched_debug.cfs_rq:/.removed.runnable_sum.avg
    205.71 +- 70%    +119.5%     451.60 +- 27%  sched_debug.cfs_rq:/.removed.runnable_sum.stddev
      0.16 +- 70%    +237.9%       0.54 +- 50%  sched_debug.cfs_rq:/.removed.util_avg.avg
      2.23 +- 70%    +114.2%       4.77 +- 24%  sched_debug.cfs_rq:/.removed.util_avg.stddev
    573.70 +-  5%      -9.7%     518.06 +-  6%  sched_debug.cfs_rq:/.util_avg.min
    114.87 +-  8%     +14.1%     131.04 +- 10%  sched_debug.cfs_rq:/.util_est_enqueued.avg
     64.42 +- 54%     -63.9%      23.27 +- 68%  sched_debug.cpu.cpu_load[1].max
      5.05 +- 48%     -55.2%       2.26 +- 51%  sched_debug.cpu.cpu_load[1].stddev
     57.58 +- 59%     -60.3%      22.88 +- 70%  sched_debug.cpu.cpu_load[2].max
     21019 +-  3%     -15.1%      17841 +-  6%  sched_debug.cpu.nr_switches.min
     20797 +-  3%     -15.0%      17670 +-  6%  sched_debug.cpu.sched_count.min
     10287 +-  3%     -15.1%       8736 +-  6%  sched_debug.cpu.sched_goidle.avg
     13693 +-  2%     -10.7%      12233 +-  5%  sched_debug.cpu.sched_goidle.max
      9976 +-  3%     -16.0%       8381 +-  7%  sched_debug.cpu.sched_goidle.min
      0.00 +- 26%     +98.9%       0.00 +- 28%  sched_debug.rt_rq:/.rt_time.min
      4230 +-141%    -100.0%       0.00        latency_stats.avg.trace_module_notify.notifier_call_chain.blocking_notifier_call_chain.do_init_module.load_module.__do_sys_finit_module.do_syscall_64.entry_SYSCALL_64_after_hwframe
     28498 +-141%    -100.0%       0.00        latency_stats.avg.perf_event_alloc.__do_sys_perf_event_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
      4065 +-138%     -92.2%     315.33 +- 91%  latency_stats.avg.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_lookup.nfs_lookup_revalidate.lookup_fast.walk_component.link_path_walk.path_lookupat.filename_lookup
      0.00       +3.6e+105%       3641 +-141%  latency_stats.avg.down.console_lock.console_device.tty_lookup_driver.tty_open.chrdev_open.do_dentry_open.path_openat.do_filp_open.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00       +2.5e+106%      25040 +-141%  latency_stats.avg.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_lookup.nfs_lookup.path_openat.do_filp_open.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00       +3.4e+106%      34015 +-141%  latency_stats.avg.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_get_acl.get_acl.posix_acl_create.nfs3_proc_create.nfs_create.path_openat.do_filp_open.do_sys_open
      0.00       +4.8e+106%      47686 +-141%  latency_stats.avg.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_do_create.nfs3_proc_create.nfs_create.path_openat.do_filp_open.do_sys_open.do_syscall_64
      4230 +-141%    -100.0%       0.00        latency_stats.max.trace_module_notify.notifier_call_chain.blocking_notifier_call_chain.do_init_module.load_module.__do_sys_finit_module.do_syscall_64.entry_SYSCALL_64_after_hwframe
     28498 +-141%    -100.0%       0.00        latency_stats.max.perf_event_alloc.__do_sys_perf_event_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
      4065 +-138%     -92.2%     315.33 +- 91%  latency_stats.max.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_lookup.nfs_lookup_revalidate.lookup_fast.walk_component.link_path_walk.path_lookupat.filename_lookup
      4254 +-134%     -88.0%     511.67 +- 90%  latency_stats.max.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_getattr.__nfs_revalidate_inode.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_lookupat
     43093 +- 35%     +76.6%      76099 +-115%  latency_stats.max.blk_execute_rq.scsi_execute.ioctl_internal_command.scsi_set_medium_removal.cdrom_release.[cdrom].sr_block_release.[sr_mod].__blkdev_put.blkdev_close.__fput.task_work_run.exit_to_usermode_loop.do_syscall_64
     24139 +- 70%    +228.5%      79285 +-105%  latency_stats.max.blk_execute_rq.scsi_execute.scsi_test_unit_ready.sr_check_events.[sr_mod].cdrom_check_events.[cdrom].sr_block_check_events.[sr_mod].disk_check_events.disk_clear_events.check_disk_change.sr_block_open.[sr_mod].__blkdev_get.blkdev_get
      0.00       +3.6e+105%       3641 +-141%  latency_stats.max.down.console_lock.console_device.tty_lookup_driver.tty_open.chrdev_open.do_dentry_open.path_openat.do_filp_open.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00       +2.5e+106%      25040 +-141%  latency_stats.max.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_lookup.nfs_lookup.path_openat.do_filp_open.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00       +3.4e+106%      34015 +-141%  latency_stats.max.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_get_acl.get_acl.posix_acl_create.nfs3_proc_create.nfs_create.path_openat.do_filp_open.do_sys_open
      0.00       +6.5e+106%      64518 +-141%  latency_stats.max.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_do_create.nfs3_proc_create.nfs_create.path_openat.do_filp_open.do_sys_open.do_syscall_64
      4230 +-141%    -100.0%       0.00        latency_stats.sum.trace_module_notify.notifier_call_chain.blocking_notifier_call_chain.do_init_module.load_module.__do_sys_finit_module.do_syscall_64.entry_SYSCALL_64_after_hwframe
     28498 +-141%    -100.0%       0.00        latency_stats.sum.perf_event_alloc.__do_sys_perf_event_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
      4065 +-138%     -92.2%     315.33 +- 91%  latency_stats.sum.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_lookup.nfs_lookup_revalidate.lookup_fast.walk_component.link_path_walk.path_lookupat.filename_lookup
     57884 +-  9%     +47.3%      85264 +-118%  latency_stats.sum.blk_execute_rq.scsi_execute.ioctl_internal_command.scsi_set_medium_removal.cdrom_release.[cdrom].sr_block_release.[sr_mod].__blkdev_put.blkdev_close.__fput.task_work_run.exit_to_usermode_loop.do_syscall_64
      0.00       +3.6e+105%       3641 +-141%  latency_stats.sum.down.console_lock.console_device.tty_lookup_driver.tty_open.chrdev_open.do_dentry_open.path_openat.do_filp_open.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00       +2.5e+106%      25040 +-141%  latency_stats.sum.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_lookup.nfs_lookup.path_openat.do_filp_open.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00       +3.4e+106%      34015 +-141%  latency_stats.sum.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_get_acl.get_acl.posix_acl_create.nfs3_proc_create.nfs_create.path_openat.do_filp_open.do_sys_open
      0.00       +9.5e+106%      95373 +-141%  latency_stats.sum.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_do_create.nfs3_proc_create.nfs_create.path_openat.do_filp_open.do_sys_open.do_syscall_64
     11.70           -11.7        0.00        perf-profile.calltrace.cycles-pp.__do_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault.do_page_fault
     11.52           -11.5        0.00        perf-profile.calltrace.cycles-pp.shmem_fault.__do_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
     10.44           -10.4        0.00        perf-profile.calltrace.cycles-pp.shmem_getpage_gfp.shmem_fault.__do_fault.__handle_mm_fault.handle_mm_fault
      9.83            -9.8        0.00        perf-profile.calltrace.cycles-pp.find_lock_entry.shmem_getpage_gfp.shmem_fault.__do_fault.__handle_mm_fault
      9.55            -9.5        0.00        perf-profile.calltrace.cycles-pp.finish_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault.do_page_fault
      9.35            -9.3        0.00        perf-profile.calltrace.cycles-pp.alloc_set_pte.finish_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
      6.81            -6.8        0.00        perf-profile.calltrace.cycles-pp.page_add_file_rmap.alloc_set_pte.finish_fault.__handle_mm_fault.handle_mm_fault
      7.71            -0.3        7.45        perf-profile.calltrace.cycles-pp.find_get_entry.find_lock_entry.shmem_getpage_gfp.shmem_fault.__do_fault
      0.59 +-  7%      -0.2        0.35 +- 70%  perf-profile.calltrace.cycles-pp.smp_apic_timer_interrupt.apic_timer_interrupt.__do_page_fault.do_page_fault.page_fault
      0.59 +-  7%      -0.2        0.35 +- 70%  perf-profile.calltrace.cycles-pp.apic_timer_interrupt.__do_page_fault.do_page_fault.page_fault
     10.41            -0.2       10.24        perf-profile.calltrace.cycles-pp.native_irq_return_iret
      7.68            -0.1        7.60        perf-profile.calltrace.cycles-pp.swapgs_restore_regs_and_return_to_usermode
      0.76            -0.1        0.70        perf-profile.calltrace.cycles-pp.down_read_trylock.__do_page_fault.do_page_fault.page_fault
      1.38            -0.0        1.34        perf-profile.calltrace.cycles-pp.do_page_fault
      1.05            -0.0        1.02        perf-profile.calltrace.cycles-pp.trace_graph_entry.do_page_fault
      0.92            +0.0        0.94        perf-profile.calltrace.cycles-pp.find_vma.__do_page_fault.do_page_fault.page_fault
      0.91            +0.0        0.93        perf-profile.calltrace.cycles-pp.vmacache_find.find_vma.__do_page_fault.do_page_fault.page_fault
      0.65            +0.0        0.67        perf-profile.calltrace.cycles-pp.set_page_dirty.unmap_page_range.unmap_vmas.unmap_region.do_munmap
      0.62            +0.0        0.66        perf-profile.calltrace.cycles-pp.page_mapping.find_lock_entry.shmem_getpage_gfp.shmem_fault.__do_fault
      4.15            +0.1        4.27        perf-profile.calltrace.cycles-pp.page_remove_rmap.unmap_page_range.unmap_vmas.unmap_region.do_munmap
     10.17            +0.2       10.39        perf-profile.calltrace.cycles-pp.munmap
      9.56            +0.2        9.78        perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.munmap
      9.56            +0.2        9.78        perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.munmap
      9.56            +0.2        9.78        perf-profile.calltrace.cycles-pp.unmap_region.do_munmap.vm_munmap.__x64_sys_munmap.do_syscall_64
      9.54            +0.2        9.76        perf-profile.calltrace.cycles-pp.unmap_page_range.unmap_vmas.unmap_region.do_munmap.vm_munmap
      9.54            +0.2        9.76        perf-profile.calltrace.cycles-pp.unmap_vmas.unmap_region.do_munmap.vm_munmap.__x64_sys_munmap
      9.56            +0.2        9.78        perf-profile.calltrace.cycles-pp.do_munmap.vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe
      9.56            +0.2        9.78        perf-profile.calltrace.cycles-pp.vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.munmap
      9.56            +0.2        9.78        perf-profile.calltrace.cycles-pp.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.munmap
      0.00            +0.6        0.56 +-  2%  perf-profile.calltrace.cycles-pp.lock_page_memcg.page_add_file_rmap.alloc_set_pte.finish_fault.handle_pte_fault
      0.00            +0.6        0.59        perf-profile.calltrace.cycles-pp.page_mapping.set_page_dirty.fault_dirty_shared_page.handle_pte_fault.__handle_mm_fault
      0.00            +0.6        0.60        perf-profile.calltrace.cycles-pp.current_time.file_update_time.handle_pte_fault.__handle_mm_fault.handle_mm_fault
      0.00            +0.7        0.68        perf-profile.calltrace.cycles-pp.___might_sleep.find_lock_entry.shmem_getpage_gfp.shmem_fault.__do_fault
      0.00            +0.7        0.74        perf-profile.calltrace.cycles-pp.unlock_page.fault_dirty_shared_page.handle_pte_fault.__handle_mm_fault.handle_mm_fault
      0.00            +0.8        0.80        perf-profile.calltrace.cycles-pp.set_page_dirty.fault_dirty_shared_page.handle_pte_fault.__handle_mm_fault.handle_mm_fault
      0.00            +0.9        0.88        perf-profile.calltrace.cycles-pp._raw_spin_lock.pte_map_lock.alloc_set_pte.finish_fault.handle_pte_fault
      0.00            +0.9        0.91        perf-profile.calltrace.cycles-pp.__set_page_dirty_no_writeback.fault_dirty_shared_page.handle_pte_fault.__handle_mm_fault.handle_mm_fault
      0.00            +1.3        1.27        perf-profile.calltrace.cycles-pp.pte_map_lock.alloc_set_pte.finish_fault.handle_pte_fault.__handle_mm_fault
      0.00            +1.3        1.30        perf-profile.calltrace.cycles-pp.file_update_time.handle_pte_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
      0.00            +2.8        2.76        perf-profile.calltrace.cycles-pp.fault_dirty_shared_page.handle_pte_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
      0.00            +6.8        6.81        perf-profile.calltrace.cycles-pp.page_add_file_rmap.alloc_set_pte.finish_fault.handle_pte_fault.__handle_mm_fault
      0.00            +9.4        9.39        perf-profile.calltrace.cycles-pp.alloc_set_pte.finish_fault.handle_pte_fault.__handle_mm_fault.handle_mm_fault
      0.00            +9.6        9.59        perf-profile.calltrace.cycles-pp.finish_fault.handle_pte_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
      0.00            +9.8        9.77        perf-profile.calltrace.cycles-pp.find_lock_entry.shmem_getpage_gfp.shmem_fault.__do_fault.handle_pte_fault
      0.00           +10.4       10.37        perf-profile.calltrace.cycles-pp.shmem_getpage_gfp.shmem_fault.__do_fault.handle_pte_fault.__handle_mm_fault
      0.00           +11.5       11.46        perf-profile.calltrace.cycles-pp.shmem_fault.__do_fault.handle_pte_fault.__handle_mm_fault.handle_mm_fault
      0.00           +11.6       11.60        perf-profile.calltrace.cycles-pp.__do_fault.handle_pte_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
      0.00           +26.6       26.62        perf-profile.calltrace.cycles-pp.handle_pte_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault.do_page_fault
      7.88            -0.3        7.61        perf-profile.children.cycles-pp.find_get_entry
      1.34 +-  8%      -0.2        1.16 +-  2%  perf-profile.children.cycles-pp.hrtimer_interrupt
     10.41            -0.2       10.24        perf-profile.children.cycles-pp.native_irq_return_iret
      0.38 +- 28%      -0.1        0.26 +-  4%  perf-profile.children.cycles-pp.tick_sched_timer
     11.80            -0.1       11.68        perf-profile.children.cycles-pp.__do_fault
      0.55 +- 15%      -0.1        0.43 +-  2%  perf-profile.children.cycles-pp.__hrtimer_run_queues
      0.60            -0.1        0.51        perf-profile.children.cycles-pp.pmd_devmap_trans_unstable
      0.38 +- 13%      -0.1        0.29 +-  4%  perf-profile.children.cycles-pp.ktime_get
      7.68            -0.1        7.60        perf-profile.children.cycles-pp.swapgs_restore_regs_and_return_to_usermode
      5.18            -0.1        5.12        perf-profile.children.cycles-pp.trace_graph_entry
      0.79            -0.1        0.73        perf-profile.children.cycles-pp.down_read_trylock
      7.83            -0.1        7.76        perf-profile.children.cycles-pp.sync_regs
      3.01            -0.1        2.94        perf-profile.children.cycles-pp.fault_dirty_shared_page
      1.02            -0.1        0.96        perf-profile.children.cycles-pp._raw_spin_lock
      4.66            -0.1        4.61        perf-profile.children.cycles-pp.prepare_ftrace_return
      0.37 +-  8%      -0.1        0.32 +-  3%  perf-profile.children.cycles-pp.current_kernel_time64
      5.26            -0.1        5.21        perf-profile.children.cycles-pp.ftrace_graph_caller
      0.66 +-  5%      -0.1        0.61        perf-profile.children.cycles-pp.current_time
      0.18 +-  5%      -0.0        0.15 +-  3%  perf-profile.children.cycles-pp.update_process_times
      0.27            -0.0        0.26        perf-profile.children.cycles-pp._cond_resched
      0.16            -0.0        0.15 +-  3%  perf-profile.children.cycles-pp.rcu_all_qs
      0.94            +0.0        0.95        perf-profile.children.cycles-pp.vmacache_find
      0.48            +0.0        0.50        perf-profile.children.cycles-pp.__mod_node_page_state
      0.17            +0.0        0.19 +-  2%  perf-profile.children.cycles-pp.__unlock_page_memcg
      1.07            +0.0        1.10        perf-profile.children.cycles-pp.find_vma
      0.79 +-  3%      +0.1        0.86 +-  2%  perf-profile.children.cycles-pp.lock_page_memcg
      4.29            +0.1        4.40        perf-profile.children.cycles-pp.page_remove_rmap
      1.39 +-  2%      +0.1        1.52        perf-profile.children.cycles-pp.file_update_time
      0.00            +0.2        0.16        perf-profile.children.cycles-pp.__vm_normal_page
      9.63            +0.2        9.84        perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
      9.63            +0.2        9.84        perf-profile.children.cycles-pp.do_syscall_64
      9.63            +0.2        9.84        perf-profile.children.cycles-pp.unmap_page_range
     10.17            +0.2       10.39        perf-profile.children.cycles-pp.munmap
      9.56            +0.2        9.78        perf-profile.children.cycles-pp.unmap_region
      9.56            +0.2        9.78        perf-profile.children.cycles-pp.do_munmap
      9.56            +0.2        9.78        perf-profile.children.cycles-pp.vm_munmap
      9.56            +0.2        9.78        perf-profile.children.cycles-pp.__x64_sys_munmap
      9.54            +0.2        9.77        perf-profile.children.cycles-pp.unmap_vmas
      1.01            +0.2        1.25        perf-profile.children.cycles-pp.___might_sleep
      0.00            +1.6        1.59        perf-profile.children.cycles-pp.pte_map_lock
      0.00           +26.9       26.89        perf-profile.children.cycles-pp.handle_pte_fault
      4.25            -1.0        3.24        perf-profile.self.cycles-pp.__handle_mm_fault
      1.42            -0.3        1.11        perf-profile.self.cycles-pp.alloc_set_pte
      4.87            -0.3        4.59        perf-profile.self.cycles-pp.find_get_entry
     10.41            -0.2       10.24        perf-profile.self.cycles-pp.native_irq_return_iret
      0.37 +- 13%      -0.1        0.28 +-  4%  perf-profile.self.cycles-pp.ktime_get
      0.60            -0.1        0.51        perf-profile.self.cycles-pp.pmd_devmap_trans_unstable
      7.50            -0.1        7.42        perf-profile.self.cycles-pp.swapgs_restore_regs_and_return_to_usermode
      7.83            -0.1        7.76        perf-profile.self.cycles-pp.sync_regs
      4.85            -0.1        4.79        perf-profile.self.cycles-pp.trace_graph_entry
      1.01            -0.1        0.95        perf-profile.self.cycles-pp._raw_spin_lock
      0.78            -0.1        0.73        perf-profile.self.cycles-pp.down_read_trylock
      0.36 +-  9%      -0.1        0.31 +-  4%  perf-profile.self.cycles-pp.current_kernel_time64
      0.28            -0.0        0.23 +-  2%  perf-profile.self.cycles-pp.__do_fault
      1.04            -0.0        1.00        perf-profile.self.cycles-pp.find_lock_entry
      0.30            -0.0        0.28 +-  3%  perf-profile.self.cycles-pp.fault_dirty_shared_page
      0.70            -0.0        0.67        perf-profile.self.cycles-pp.prepare_ftrace_return
      0.44            -0.0        0.42        perf-profile.self.cycles-pp.do_page_fault
      0.16            -0.0        0.14        perf-profile.self.cycles-pp.rcu_all_qs
      0.78            -0.0        0.77        perf-profile.self.cycles-pp.shmem_getpage_gfp
      0.20            -0.0        0.19        perf-profile.self.cycles-pp._cond_resched
      0.50            +0.0        0.51        perf-profile.self.cycles-pp.set_page_dirty
      0.93            +0.0        0.95        perf-profile.self.cycles-pp.vmacache_find
      0.36 +-  2%      +0.0        0.38        perf-profile.self.cycles-pp.__might_sleep
      0.47            +0.0        0.50        perf-profile.self.cycles-pp.__mod_node_page_state
      0.17            +0.0        0.19 +-  2%  perf-profile.self.cycles-pp.__unlock_page_memcg
      2.34            +0.0        2.38        perf-profile.self.cycles-pp.unmap_page_range
      0.78 +-  3%      +0.1        0.85 +-  2%  perf-profile.self.cycles-pp.lock_page_memcg
      2.17            +0.1        2.24        perf-profile.self.cycles-pp.__do_page_fault
      0.00            +0.2        0.16 +-  3%  perf-profile.self.cycles-pp.__vm_normal_page
      1.00            +0.2        1.24        perf-profile.self.cycles-pp.___might_sleep
      0.00            +0.7        0.70        perf-profile.self.cycles-pp.pte_map_lock
      0.00            +1.4        1.42 +-  2%  perf-profile.self.cycles-pp.handle_pte_fault

=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/nr_task/thp_enabled/test/cpufreq_governor:
  lkp-skl-4sp1/will-it-scale/debian-x86_64-2018-04-03.cgz/x86_64-rhel-7.2/gcc-7/100%/never/context_switch1/performance

commit: 
  ba98a1cdad71d259a194461b3a61471b49b14df1
  a7a8993bfe3ccb54ad468b9f1799649e4ad1ff12

ba98a1cdad71d259 a7a8993bfe3ccb54ad468b9f17 
---------------- -------------------------- 
       fail:runs  %reproduction    fail:runs
           |             |             |    
           :3           33%           1:3     dmesg.WARNING:at#for_ip_interrupt_entry/0x
          2:3          -67%            :3     kmsg.pstore:crypto_comp_decompress_failed,ret=
          2:3          -67%            :3     kmsg.pstore:decompression_failed
         %stddev     %change         %stddev
             \          |                \  
    224431            -1.3%     221567        will-it-scale.per_process_ops
    237006            -2.2%     231907        will-it-scale.per_thread_ops
 1.601e+09 +- 29%     -46.9%  8.501e+08 +- 12%  will-it-scale.time.involuntary_context_switches
      5429            -1.6%       5344        will-it-scale.time.user_time
  88596221            -1.7%   87067269        will-it-scale.workload
      6863 +-  6%      -9.7%       6200        boot-time.idle
    144908 +- 40%     -66.8%      48173 +- 93%  meminfo.CmaFree
      0.00 +- 70%      +0.0        0.00        mpstat.cpu.iowait%
    448336 +- 14%     -34.8%     292125 +-  3%  turbostat.C1
      7684 +-  6%      -9.5%       6957        uptime.idle
 1.601e+09 +- 29%     -46.9%  8.501e+08 +- 12%  time.involuntary_context_switches
      5429            -1.6%       5344        time.user_time
  44013162            -1.7%   43243125        vmstat.system.cs
    207684            -1.1%     205485        vmstat.system.in
   2217033 +- 15%     -15.8%    1866876 +-  2%  cpuidle.C1.time
    451218 +- 14%     -34.7%     294841 +-  2%  cpuidle.C1.usage
     24839 +- 10%     -19.9%      19896        cpuidle.POLL.time
      7656 +- 11%     -38.9%       4676 +-  8%  cpuidle.POLL.usage
      5.48 +- 49%     -67.3%       1.79 +-100%  irq_exception_noise.__do_page_fault.95th
      9.46 +- 21%     -58.2%       3.95 +- 64%  irq_exception_noise.__do_page_fault.99th
     35.67 +-  8%   +1394.4%     533.00 +- 96%  irq_exception_noise.irq_nr
     52109 +-  3%     -16.0%      43784 +-  4%  irq_exception_noise.softirq_time
     36226 +- 40%     -66.7%      12048 +- 93%  proc-vmstat.nr_free_cma
     25916            -1.0%      25659        proc-vmstat.nr_slab_reclaimable
     16279 +-  8%   +2646.1%     447053 +- 82%  proc-vmstat.pgalloc_movable
   2231117           -18.4%    1820828 +- 20%  proc-vmstat.pgalloc_normal
   1109316 +- 46%     -86.9%     145207 +-109%  numa-numastat.node1.local_node
   1114700 +- 45%     -84.5%     172877 +- 85%  numa-numastat.node1.numa_hit
      5523 +-140%    +402.8%      27768 +- 39%  numa-numastat.node1.other_node
     29013 +- 29%   +3048.1%     913379 +- 73%  numa-numastat.node3.local_node
     65032 +- 13%   +1335.1%     933270 +- 70%  numa-numastat.node3.numa_hit
     36018           -44.8%      19897 +- 75%  numa-numastat.node3.other_node
     12.79 +- 21%   +7739.1%       1002 +-136%  sched_debug.cpu.cpu_load[1].max
      1.82 +- 10%   +3901.1%      72.92 +-135%  sched_debug.cpu.cpu_load[1].stddev
      1.71 +-  4%   +5055.8%      88.08 +-137%  sched_debug.cpu.cpu_load[2].stddev
     12.33 +- 23%   +9061.9%       1129 +-139%  sched_debug.cpu.cpu_load[3].max
      1.78 +- 10%   +4514.8%      82.18 +-137%  sched_debug.cpu.cpu_load[3].stddev
      4692 +- 72%    +154.5%      11945 +- 29%  sched_debug.cpu.max_idle_balance_cost.stddev
     23979            -8.3%      21983        slabinfo.kmalloc-96.active_objs
      1358 +-  6%     -17.9%       1114 +-  3%  slabinfo.nsproxy.active_objs
      1358 +-  6%     -17.9%       1114 +-  3%  slabinfo.nsproxy.num_objs
     15229           +12.4%      17119        slabinfo.pde_opener.active_objs
     15229           +12.4%      17119        slabinfo.pde_opener.num_objs
     59541 +-  8%     -10.1%      53537 +-  8%  slabinfo.vm_area_struct.active_objs
     59612 +-  8%     -10.1%      53604 +-  8%  slabinfo.vm_area_struct.num_objs
 4.163e+13            -1.4%  4.105e+13        perf-stat.branch-instructions
 6.537e+11            -1.2%  6.459e+11        perf-stat.branch-misses
 2.667e+10            -1.7%  2.621e+10        perf-stat.context-switches
      1.21            +1.3%       1.22        perf-stat.cpi
    150508            -9.8%     135825 +-  3%  perf-stat.cpu-migrations
      5.75 +- 33%      +5.4       11.11 +- 26%  perf-stat.iTLB-load-miss-rate%
 3.619e+09 +- 36%    +100.9%  7.272e+09 +- 30%  perf-stat.iTLB-load-misses
 2.089e+14            -1.3%  2.062e+14        perf-stat.instructions
     64607 +- 29%     -50.5%      31964 +- 37%  perf-stat.instructions-per-iTLB-miss
      0.83            -1.3%       0.82        perf-stat.ipc
      3972 +-  4%     -14.7%       3388 +-  8%  numa-meminfo.node0.PageTables
    207919 +- 25%     -57.2%      88989 +- 74%  numa-meminfo.node1.Active
    207715 +- 26%     -57.3%      88785 +- 74%  numa-meminfo.node1.Active(anon)
    356529           -34.3%     234069 +-  2%  numa-meminfo.node1.FilePages
    789129 +-  5%     -19.8%     633161 +- 12%  numa-meminfo.node1.MemUsed
     34777 +-  8%     -48.2%      18010 +- 30%  numa-meminfo.node1.SReclaimable
     69641 +-  4%     -20.7%      55250 +- 12%  numa-meminfo.node1.SUnreclaim
    125526 +-  4%     -96.3%       4602 +- 41%  numa-meminfo.node1.Shmem
    104419           -29.8%      73261 +- 16%  numa-meminfo.node1.Slab
    103661 +- 17%     -72.0%      29029 +- 99%  numa-meminfo.node2.Active
    103661 +- 17%     -72.2%      28829 +-101%  numa-meminfo.node2.Active(anon)
    103564 +- 18%     -72.0%      29007 +-100%  numa-meminfo.node2.AnonPages
    671654 +-  7%     -14.6%     573598 +-  4%  numa-meminfo.node2.MemUsed
     44206 +-127%    +301.4%     177465 +- 42%  numa-meminfo.node3.Active
     44206 +-127%    +301.0%     177263 +- 42%  numa-meminfo.node3.Active(anon)
      8738           +12.2%       9805 +-  8%  numa-meminfo.node3.KernelStack
    603605 +-  9%     +27.8%     771554 +- 14%  numa-meminfo.node3.MemUsed
     14438 +-  6%    +122.9%      32181 +- 42%  numa-meminfo.node3.SReclaimable
      2786 +-137%   +3302.0%      94792 +- 71%  numa-meminfo.node3.Shmem
     71461 +-  7%     +45.2%     103771 +- 29%  numa-meminfo.node3.Slab
    247197 +-  4%      -7.8%     227843        numa-meminfo.node3.Unevictable
    991.67 +-  4%     -14.7%     846.00 +-  8%  numa-vmstat.node0.nr_page_table_pages
     51926 +- 26%     -57.3%      22196 +- 74%  numa-vmstat.node1.nr_active_anon
     89137           -34.4%      58516 +-  2%  numa-vmstat.node1.nr_file_pages
      1679 +-  5%     -10.8%       1498 +-  4%  numa-vmstat.node1.nr_mapped
     31386 +-  4%     -96.3%       1150 +- 41%  numa-vmstat.node1.nr_shmem
      8694 +-  8%     -48.2%       4502 +- 30%  numa-vmstat.node1.nr_slab_reclaimable
     17410 +-  4%     -20.7%      13812 +- 12%  numa-vmstat.node1.nr_slab_unreclaimable
     51926 +- 26%     -57.3%      22196 +- 74%  numa-vmstat.node1.nr_zone_active_anon
   1037174 +- 24%     -57.0%     446205 +- 35%  numa-vmstat.node1.numa_hit
    961611 +- 26%     -65.8%     328687 +- 50%  numa-vmstat.node1.numa_local
     75563 +- 44%     +55.5%     117517 +-  9%  numa-vmstat.node1.numa_other
     25914 +- 17%     -72.2%       7206 +-101%  numa-vmstat.node2.nr_active_anon
     25891 +- 18%     -72.0%       7251 +-100%  numa-vmstat.node2.nr_anon_pages
     25914 +- 17%     -72.2%       7206 +-101%  numa-vmstat.node2.nr_zone_active_anon
     11051 +-127%    +301.0%      44309 +- 42%  numa-vmstat.node3.nr_active_anon
     36227 +- 40%     -66.7%      12049 +- 93%  numa-vmstat.node3.nr_free_cma
      0.33 +-141%  +25000.0%      83.67 +- 81%  numa-vmstat.node3.nr_inactive_file
      8739           +12.2%       9806 +-  8%  numa-vmstat.node3.nr_kernel_stack
    696.67 +-137%   +3299.7%      23684 +- 71%  numa-vmstat.node3.nr_shmem
      3609 +-  6%    +122.9%       8044 +- 42%  numa-vmstat.node3.nr_slab_reclaimable
     61799 +-  4%      -7.8%      56960        numa-vmstat.node3.nr_unevictable
     11053 +-127%    +301.4%      44361 +- 42%  numa-vmstat.node3.nr_zone_active_anon
      0.33 +-141%  +25000.0%      83.67 +- 81%  numa-vmstat.node3.nr_zone_inactive_file
     61799 +-  4%      -7.8%      56960        numa-vmstat.node3.nr_zone_unevictable
    217951 +-  8%    +280.8%     829976 +- 65%  numa-vmstat.node3.numa_hit
     91303 +- 19%    +689.3%     720647 +- 77%  numa-vmstat.node3.numa_local
    126648           -13.7%     109329 +- 13%  numa-vmstat.node3.numa_other
      8.54            -0.1        8.40        perf-profile.calltrace.cycles-pp.dequeue_task_fair.__schedule.schedule.pipe_wait.pipe_read
      5.04            -0.1        4.94        perf-profile.calltrace.cycles-pp.__switch_to.read
      3.43            -0.1        3.35        perf-profile.calltrace.cycles-pp.syscall_return_via_sysret.write
      2.77            -0.1        2.72        perf-profile.calltrace.cycles-pp.reweight_entity.enqueue_task_fair.ttwu_do_activate.try_to_wake_up.autoremove_wake_function
      1.99            -0.0        1.94        perf-profile.calltrace.cycles-pp.copy_page_to_iter.pipe_read.__vfs_read.vfs_read.ksys_read
      0.60 +-  2%      -0.0        0.57 +-  2%  perf-profile.calltrace.cycles-pp.find_next_bit.cpumask_next_wrap.select_idle_sibling.select_task_rq_fair.try_to_wake_up
      0.81            -0.0        0.78        perf-profile.calltrace.cycles-pp.___perf_sw_event.__schedule.schedule.pipe_wait.pipe_read
      0.78            +0.0        0.80        perf-profile.calltrace.cycles-pp.__fdget_pos.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
      0.73            +0.0        0.75        perf-profile.calltrace.cycles-pp.__fget_light.__fdget_pos.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.92            +0.0        0.95        perf-profile.calltrace.cycles-pp.check_preempt_wakeup.check_preempt_curr.ttwu_do_wakeup.try_to_wake_up.autoremove_wake_function
      2.11            +0.0        2.15        perf-profile.calltrace.cycles-pp.security_file_permission.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
      7.00            -0.1        6.86        perf-profile.children.cycles-pp.syscall_return_via_sysret
      5.26            -0.1        5.14        perf-profile.children.cycles-pp.__switch_to
      5.65            -0.1        5.56        perf-profile.children.cycles-pp.reweight_entity
      2.17            -0.1        2.12        perf-profile.children.cycles-pp.copy_page_to_iter
      2.94            -0.0        2.90        perf-profile.children.cycles-pp.update_cfs_group
      3.11            -0.0        3.07        perf-profile.children.cycles-pp.pick_next_task_fair
      2.59            -0.0        2.55        perf-profile.children.cycles-pp.load_new_mm_cr3
      1.92            -0.0        1.88        perf-profile.children.cycles-pp._raw_spin_lock_irqsave
      1.11            -0.0        1.08 +-  2%  perf-profile.children.cycles-pp.find_next_bit
      0.59            -0.0        0.56        perf-profile.children.cycles-pp.finish_task_switch
      0.14 +- 15%      -0.0        0.11 +- 16%  perf-profile.children.cycles-pp.write@plt
      1.21            -0.0        1.18        perf-profile.children.cycles-pp.set_next_entity
      0.85            -0.0        0.82        perf-profile.children.cycles-pp.___perf_sw_event
      0.13 +-  3%      -0.0        0.11 +-  4%  perf-profile.children.cycles-pp.timespec_trunc
      0.47 +-  2%      -0.0        0.45        perf-profile.children.cycles-pp.anon_pipe_buf_release
      0.38 +-  2%      -0.0        0.36        perf-profile.children.cycles-pp.file_update_time
      0.74            -0.0        0.73        perf-profile.children.cycles-pp.copyout
      0.41 +-  2%      -0.0        0.39        perf-profile.children.cycles-pp.copy_user_enhanced_fast_string
      0.32            -0.0        0.30        perf-profile.children.cycles-pp.__x64_sys_read
      0.14            -0.0        0.12 +-  3%  perf-profile.children.cycles-pp.current_kernel_time64
      0.91            +0.0        0.92        perf-profile.children.cycles-pp.touch_atime
      0.40            +0.0        0.41        perf-profile.children.cycles-pp._cond_resched
      0.18 +-  2%      +0.0        0.20        perf-profile.children.cycles-pp.activate_task
      0.05            +0.0        0.07 +-  6%  perf-profile.children.cycles-pp.default_wake_function
      0.24            +0.0        0.27 +-  3%  perf-profile.children.cycles-pp.rcu_all_qs
      0.60 +-  2%      +0.0        0.64 +-  2%  perf-profile.children.cycles-pp.update_min_vruntime
      0.42 +-  4%      +0.0        0.46 +-  4%  perf-profile.children.cycles-pp.probe_sched_switch
      1.33            +0.0        1.38        perf-profile.children.cycles-pp.__fget_light
      0.53 +-  2%      +0.1        0.58        perf-profile.children.cycles-pp.entry_SYSCALL_64_stage2
      0.31            +0.1        0.36 +-  2%  perf-profile.children.cycles-pp.generic_pipe_buf_confirm
      4.35            +0.1        4.41        perf-profile.children.cycles-pp.switch_mm_irqs_off
      2.52            +0.1        2.58        perf-profile.children.cycles-pp.selinux_file_permission
      0.00            +0.1        0.07 +- 11%  perf-profile.children.cycles-pp.hrtick_update
      7.00            -0.1        6.86        perf-profile.self.cycles-pp.syscall_return_via_sysret
      5.26            -0.1        5.14        perf-profile.self.cycles-pp.__switch_to
      0.29            -0.1        0.19 +-  2%  perf-profile.self.cycles-pp.ksys_read
      1.49            -0.1        1.43        perf-profile.self.cycles-pp.dequeue_task_fair
      2.41            -0.1        2.35        perf-profile.self.cycles-pp.__schedule
      1.46            -0.0        1.41        perf-profile.self.cycles-pp.select_task_rq_fair
      2.94            -0.0        2.90        perf-profile.self.cycles-pp.update_cfs_group
      0.44            -0.0        0.40        perf-profile.self.cycles-pp.dequeue_entity
      0.48            -0.0        0.44        perf-profile.self.cycles-pp.finish_task_switch
      2.59            -0.0        2.55        perf-profile.self.cycles-pp.load_new_mm_cr3
      1.11            -0.0        1.08 +-  2%  perf-profile.self.cycles-pp.find_next_bit
      1.91            -0.0        1.88        perf-profile.self.cycles-pp._raw_spin_lock_irqsave
      0.78            -0.0        0.75        perf-profile.self.cycles-pp.___perf_sw_event
      0.14 +- 15%      -0.0        0.11 +- 16%  perf-profile.self.cycles-pp.write@plt
      0.37            -0.0        0.35 +-  2%  perf-profile.self.cycles-pp.__wake_up_common_lock
      0.20 +-  2%      -0.0        0.17 +-  2%  perf-profile.self.cycles-pp.__fdget_pos
      0.47 +-  2%      -0.0        0.44        perf-profile.self.cycles-pp.anon_pipe_buf_release
      0.87            -0.0        0.85        perf-profile.self.cycles-pp.copy_user_generic_unrolled
      0.13 +-  3%      -0.0        0.11 +-  4%  perf-profile.self.cycles-pp.timespec_trunc
      0.41 +-  2%      -0.0        0.39        perf-profile.self.cycles-pp.copy_user_enhanced_fast_string
      0.38            -0.0        0.36        perf-profile.self.cycles-pp.__wake_up_common
      0.32            -0.0        0.30        perf-profile.self.cycles-pp.__x64_sys_read
      0.14 +-  3%      -0.0        0.12 +-  3%  perf-profile.self.cycles-pp.current_kernel_time64
      0.30            -0.0        0.28        perf-profile.self.cycles-pp.set_next_entity
      0.28 +-  3%      +0.0        0.30        perf-profile.self.cycles-pp._cond_resched
      0.18 +-  2%      +0.0        0.20        perf-profile.self.cycles-pp.activate_task
      0.17 +-  2%      +0.0        0.19        perf-profile.self.cycles-pp.__might_fault
      0.05            +0.0        0.07 +-  6%  perf-profile.self.cycles-pp.default_wake_function
      0.17 +-  2%      +0.0        0.20        perf-profile.self.cycles-pp.ttwu_do_activate
      0.66            +0.0        0.69        perf-profile.self.cycles-pp.write
      0.24            +0.0        0.27 +-  3%  perf-profile.self.cycles-pp.rcu_all_qs
      0.67            +0.0        0.70        perf-profile.self.cycles-pp.entry_SYSCALL_64_after_hwframe
      0.60 +-  2%      +0.0        0.64 +-  2%  perf-profile.self.cycles-pp.update_min_vruntime
      0.42 +-  4%      +0.0        0.46 +-  4%  perf-profile.self.cycles-pp.probe_sched_switch
      1.33            +0.0        1.37        perf-profile.self.cycles-pp.__fget_light
      1.61            +0.0        1.66        perf-profile.self.cycles-pp.pipe_read
      0.53 +-  2%      +0.1        0.58        perf-profile.self.cycles-pp.entry_SYSCALL_64_stage2
      0.31            +0.1        0.36 +-  2%  perf-profile.self.cycles-pp.generic_pipe_buf_confirm
      1.04            +0.1        1.11        perf-profile.self.cycles-pp.pipe_write
      0.00            +0.1        0.07 +- 11%  perf-profile.self.cycles-pp.hrtick_update
      2.00            +0.1        2.08        perf-profile.self.cycles-pp.switch_mm_irqs_off

=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/nr_task/thp_enabled/test/cpufreq_governor:
  lkp-skl-4sp1/will-it-scale/debian-x86_64-2018-04-03.cgz/x86_64-rhel-7.2/gcc-7/100%/never/page_fault3/performance

commit: 
  ba98a1cdad71d259a194461b3a61471b49b14df1
  a7a8993bfe3ccb54ad468b9f1799649e4ad1ff12

ba98a1cdad71d259 a7a8993bfe3ccb54ad468b9f17 
---------------- -------------------------- 
       fail:runs  %reproduction    fail:runs
           |             |             |    
          1:3          -33%            :3     dmesg.WARNING:stack_going_in_the_wrong_direction?ip=file_update_time/0x
           :3           33%           1:3     stderr.mount.nfs:Connection_timed_out
         34:3         -401%          22:3     perf-profile.calltrace.cycles-pp.error_entry.testcase
         17:3         -207%          11:3     perf-profile.calltrace.cycles-pp.sync_regs.error_entry.testcase
         34:3         -404%          22:3     perf-profile.children.cycles-pp.error_entry
          0:3           -2%           0:3     perf-profile.children.cycles-pp.error_exit
         16:3         -196%          11:3     perf-profile.self.cycles-pp.error_entry
          0:3           -2%           0:3     perf-profile.self.cycles-pp.error_exit
         %stddev     %change         %stddev
             \          |                \  
    467454            -1.8%     459251        will-it-scale.per_process_ops
     10856 +-  4%     -23.1%       8344 +-  7%  will-it-scale.per_thread_ops
    118134 +-  2%     +11.7%     131943        will-it-scale.time.involuntary_context_switches
 6.277e+08 +-  4%     -23.1%  4.827e+08 +-  7%  will-it-scale.time.minor_page_faults
      7406            +5.8%       7839        will-it-scale.time.percent_of_cpu_this_job_got
     44526            +5.8%      47106        will-it-scale.time.system_time
   7351468 +-  5%     -18.3%    6009014 +-  7%  will-it-scale.time.voluntary_context_switches
  91835846            -2.2%   89778599        will-it-scale.workload
   2534640            +4.3%    2643005 +-  2%  interrupts.CAL:Function_call_interrupts
      2819 +-  5%     +22.9%       3464 +- 18%  kthread_noise.total_time
     30273 +-  4%     -12.7%      26415 +-  5%  vmstat.system.cs
      1.52 +-  2%     +15.2%       1.75 +-  2%  irq_exception_noise.__do_page_fault.99th
    296.67 +- 12%     -36.7%     187.67 +- 12%  irq_exception_noise.softirq_time
    230900 +-  3%     +30.3%     300925 +-  3%  meminfo.Inactive
    230184 +-  3%     +30.4%     300180 +-  3%  meminfo.Inactive(anon)
     11.62 +-  3%      -2.2        9.40 +-  5%  mpstat.cpu.idle%
      0.00 +- 14%      -0.0        0.00 +-  4%  mpstat.cpu.iowait%
   7992174           -11.1%    7101976 +-  3%  softirqs.RCU
   4973624 +-  2%     -12.9%    4333370 +-  2%  softirqs.SCHED
    118134 +-  2%     +11.7%     131943        time.involuntary_context_switches
 6.277e+08 +-  4%     -23.1%  4.827e+08 +-  7%  time.minor_page_faults
      7406            +5.8%       7839        time.percent_of_cpu_this_job_got
     44526            +5.8%      47106        time.system_time
   7351468 +-  5%     -18.3%    6009014 +-  7%  time.voluntary_context_switches
 2.702e+09 +-  5%     -16.7%  2.251e+09 +-  7%  cpuidle.C1E.time
   6834329 +-  5%     -15.8%    5756243 +-  7%  cpuidle.C1E.usage
 1.046e+10 +-  3%     -19.8%  8.389e+09 +-  4%  cpuidle.C6.time
  13961845 +-  3%     -19.3%   11265555 +-  4%  cpuidle.C6.usage
   1309307 +-  7%     -14.8%    1116168 +-  8%  cpuidle.POLL.time
     19774 +-  6%     -13.7%      17063 +-  7%  cpuidle.POLL.usage
      2523 +-  4%     -11.1%       2243 +-  4%  slabinfo.biovec-64.active_objs
      2523 +-  4%     -11.1%       2243 +-  4%  slabinfo.biovec-64.num_objs
      2610 +-  8%     -33.7%       1731 +- 22%  slabinfo.dmaengine-unmap-16.active_objs
      2610 +-  8%     -33.7%       1731 +- 22%  slabinfo.dmaengine-unmap-16.num_objs
      5118 +- 17%     -22.6%       3962 +-  9%  slabinfo.eventpoll_pwq.active_objs
      5118 +- 17%     -22.6%       3962 +-  9%  slabinfo.eventpoll_pwq.num_objs
      4583 +-  3%     -14.0%       3941 +-  4%  slabinfo.sock_inode_cache.active_objs
      4583 +-  3%     -14.0%       3941 +-  4%  slabinfo.sock_inode_cache.num_objs
      1933            +2.6%       1984        turbostat.Avg_MHz
   6832021 +-  5%     -15.8%    5754156 +-  7%  turbostat.C1E
      2.32 +-  5%      -0.4        1.94 +-  7%  turbostat.C1E%
  13954211 +-  3%     -19.3%   11259436 +-  4%  turbostat.C6
      8.97 +-  3%      -1.8        7.20 +-  4%  turbostat.C6%
      6.18 +-  4%     -17.1%       5.13 +-  5%  turbostat.CPU%c1
      5.12 +-  3%     -21.7%       4.01 +-  4%  turbostat.CPU%c6
      1.76 +-  2%     -34.7%       1.15 +-  2%  turbostat.Pkg%pc2
     57314 +-  4%     +30.4%      74717 +-  4%  proc-vmstat.nr_inactive_anon
     57319 +-  4%     +30.4%      74719 +-  4%  proc-vmstat.nr_zone_inactive_anon
     24415 +- 19%     -62.2%       9236 +-  7%  proc-vmstat.numa_hint_faults
  69661453            -1.8%   68405712        proc-vmstat.numa_hit
  69553390            -1.8%   68297790        proc-vmstat.numa_local
      8792 +- 29%     -92.6%     654.33 +- 23%  proc-vmstat.numa_pages_migrated
     40251 +- 32%     -76.5%       9474 +-  3%  proc-vmstat.numa_pte_updates
  69522532            -1.6%   68383074        proc-vmstat.pgalloc_normal
 2.762e+10            -2.2%  2.701e+10        proc-vmstat.pgfault
  68825100            -1.5%   67772256        proc-vmstat.pgfree
      8792 +- 29%     -92.6%     654.33 +- 23%  proc-vmstat.pgmigrate_success
     57992 +-  6%     +56.2%      90591 +-  3%  numa-meminfo.node0.Inactive
     57916 +-  6%     +56.3%      90513 +-  3%  numa-meminfo.node0.Inactive(anon)
     37285 +- 12%     +36.0%      50709 +-  5%  numa-meminfo.node0.SReclaimable
    110971 +-  8%     +22.7%     136209 +-  8%  numa-meminfo.node0.Slab
     23601 +- 55%    +559.5%     155651 +- 36%  numa-meminfo.node1.AnonPages
     62484 +- 12%     +17.5%      73417 +-  3%  numa-meminfo.node1.Inactive
     62323 +- 12%     +17.2%      73023 +-  4%  numa-meminfo.node1.Inactive(anon)
    109714 +- 63%     -85.6%      15832 +- 96%  numa-meminfo.node2.AnonPages
     52236 +- 13%     +22.7%      64074 +-  3%  numa-meminfo.node2.Inactive
     51922 +- 12%     +23.2%      63963 +-  3%  numa-meminfo.node2.Inactive(anon)
     60241 +- 11%     +21.9%      73442 +-  8%  numa-meminfo.node3.Inactive
     60077 +- 12%     +22.0%      73279 +-  8%  numa-meminfo.node3.Inactive(anon)
     14093 +-  6%     +55.9%      21977 +-  3%  numa-vmstat.node0.nr_inactive_anon
      9321 +- 12%     +36.0%      12675 +-  5%  numa-vmstat.node0.nr_slab_reclaimable
     14090 +-  6%     +56.0%      21977 +-  3%  numa-vmstat.node0.nr_zone_inactive_anon
      5900 +- 55%    +559.4%      38909 +- 36%  numa-vmstat.node1.nr_anon_pages
     15413 +- 12%     +14.8%      17688 +-  4%  numa-vmstat.node1.nr_inactive_anon
     15413 +- 12%     +14.8%      17688 +-  4%  numa-vmstat.node1.nr_zone_inactive_anon
     27430 +- 63%     -85.6%       3960 +- 96%  numa-vmstat.node2.nr_anon_pages
     12928 +- 12%     +20.0%      15508 +-  3%  numa-vmstat.node2.nr_inactive_anon
     12927 +- 12%     +20.0%      15507 +-  3%  numa-vmstat.node2.nr_zone_inactive_anon
      6229 +- 10%    +117.5%      13547 +- 44%  numa-vmstat.node3
     14669 +- 11%     +19.6%      17537 +-  7%  numa-vmstat.node3.nr_inactive_anon
     14674 +- 11%     +19.5%      17541 +-  7%  numa-vmstat.node3.nr_zone_inactive_anon
     24617 +-141%    -100.0%       0.00        latency_stats.avg.io_schedule.nfs_lock_and_join_requests.nfs_updatepage.nfs_write_end.generic_perform_write.nfs_file_write.__vfs_write.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
      5049 +-105%     -99.4%      28.33 +- 82%  latency_stats.avg.call_rwsem_down_write_failed.path_openat.do_filp_open.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
    152457 +- 27%    +233.6%     508656 +- 92%  latency_stats.avg.max
      0.00       +3.9e+107%     390767 +-141%  latency_stats.avg.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_getattr.__nfs_revalidate_inode.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_openat
     24617 +-141%    -100.0%       0.00        latency_stats.max.io_schedule.nfs_lock_and_join_requests.nfs_updatepage.nfs_write_end.generic_perform_write.nfs_file_write.__vfs_write.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
      4240 +-141%    -100.0%       0.00        latency_stats.max.call_rwsem_down_write_failed.do_unlinkat.do_syscall_64.entry_SYSCALL_64_after_hwframe
      8565 +- 70%     -99.1%      80.33 +-115%  latency_stats.max.call_rwsem_down_write_failed.path_openat.do_filp_open.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
    204835 +-  6%    +457.6%    1142244 +-114%  latency_stats.max.max
      0.00       +5.1e+105%       5057 +-141%  latency_stats.max.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_access.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_openat.do_filp_open
      0.00         +1e+108%     995083 +-141%  latency_stats.max.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_getattr.__nfs_revalidate_inode.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_openat
     13175 +-  4%    -100.0%       0.00        latency_stats.sum.io_schedule.__lock_page_or_retry.filemap_fault.__do_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault.do_page_fault.page_fault
     24617 +-141%    -100.0%       0.00        latency_stats.sum.io_schedule.nfs_lock_and_join_requests.nfs_updatepage.nfs_write_end.generic_perform_write.nfs_file_write.__vfs_write.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
      4260 +-141%    -100.0%       0.00        latency_stats.sum.call_rwsem_down_write_failed.do_unlinkat.do_syscall_64.entry_SYSCALL_64_after_hwframe
      8640 +- 70%     -97.5%     216.33 +-108%  latency_stats.sum.call_rwsem_down_write_failed.path_openat.do_filp_open.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
      6673 +- 89%     -92.8%     477.67 +- 74%  latency_stats.sum.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_getattr.__nfs_revalidate_inode.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_lookupat
      0.00       +4.2e+105%       4228 +-130%  latency_stats.sum.io_schedule.__lock_page_killable.__lock_page_or_retry.filemap_fault.__do_fault.handle_pte_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault.do_page_fault.page_fault
      0.00       +7.5e+105%       7450 +- 98%  latency_stats.sum.io_schedule.__lock_page_or_retry.filemap_fault.__do_fault.handle_pte_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault.do_page_fault.page_fault
      0.00       +1.3e+106%      13050 +-141%  latency_stats.sum.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_access.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_openat.do_filp_open
      0.00       +1.5e+110%  1.508e+08 +-141%  latency_stats.sum.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_getattr.__nfs_revalidate_inode.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_openat
      0.97            -0.0        0.94        perf-stat.branch-miss-rate%
 1.329e+11            -2.6%  1.294e+11        perf-stat.branch-misses
 2.254e+11            -1.9%   2.21e+11        perf-stat.cache-references
  18308779 +-  4%     -12.8%   15969618 +-  5%  perf-stat.context-switches
      3.20            +1.8%       3.26        perf-stat.cpi
 2.233e+14            +2.7%  2.293e+14        perf-stat.cpu-cycles
      4.01            -0.2        3.83        perf-stat.dTLB-store-miss-rate%
  4.51e+11            -2.2%   4.41e+11        perf-stat.dTLB-store-misses
  1.08e+13            +2.6%  1.109e+13        perf-stat.dTLB-stores
 3.158e+10 +-  5%     +16.8%  3.689e+10 +-  2%  perf-stat.iTLB-load-misses
      2214 +-  5%     -13.8%       1907 +-  2%  perf-stat.instructions-per-iTLB-miss
      0.31            -1.8%       0.31        perf-stat.ipc
 2.762e+10            -2.2%  2.701e+10        perf-stat.minor-faults
 1.535e+10           -11.2%  1.362e+10        perf-stat.node-loads
      9.75            +1.1       10.89        perf-stat.node-store-miss-rate%
 3.012e+09           +10.6%  3.332e+09 +-  2%  perf-stat.node-store-misses
 2.787e+10            -2.2%  2.725e+10        perf-stat.node-stores
 2.762e+10            -2.2%  2.701e+10        perf-stat.page-faults
    759458            +3.2%     783404        perf-stat.path-length
    246.39 +- 15%     -20.4%     196.12 +-  6%  sched_debug.cfs_rq:/.load_avg.max
      0.21 +-  3%      +9.0%       0.23 +-  4%  sched_debug.cfs_rq:/.nr_running.stddev
     16.64 +- 27%     +61.0%      26.79 +- 17%  sched_debug.cfs_rq:/.nr_spread_over.max
     75.15           -14.4%      64.30 +-  4%  sched_debug.cfs_rq:/.util_avg.stddev
    178.80 +-  3%     +25.4%     224.12 +-  7%  sched_debug.cfs_rq:/.util_est_enqueued.avg
      1075 +-  5%     -12.3%     943.36 +-  2%  sched_debug.cfs_rq:/.util_est_enqueued.max
   2093630 +- 27%     -36.1%    1337941 +- 16%  sched_debug.cpu.avg_idle.max
    297057 +- 11%     +37.8%     409294 +- 14%  sched_debug.cpu.avg_idle.min
    293240 +- 55%     -62.3%     110571 +- 13%  sched_debug.cpu.avg_idle.stddev
    770075 +-  9%     -19.3%     621136 +- 12%  sched_debug.cpu.max_idle_balance_cost.max
     48919 +- 46%     -66.9%      16190 +- 81%  sched_debug.cpu.max_idle_balance_cost.stddev
     21716 +-  5%     -16.8%      18061 +-  7%  sched_debug.cpu.nr_switches.min
     21519 +-  5%     -17.7%      17700 +-  7%  sched_debug.cpu.sched_count.min
     10586 +-  5%     -18.1%       8669 +-  7%  sched_debug.cpu.sched_goidle.avg
     14183 +-  3%     -17.6%      11693 +-  5%  sched_debug.cpu.sched_goidle.max
     10322 +-  5%     -18.6%       8407 +-  7%  sched_debug.cpu.sched_goidle.min
    400.99 +-  8%     -13.0%     348.75 +-  3%  sched_debug.cpu.sched_goidle.stddev
      5459 +-  8%     +10.0%       6006 +-  3%  sched_debug.cpu.ttwu_local.avg
      8.47 +- 42%    +345.8%      37.73 +- 77%  sched_debug.rt_rq:/.rt_time.max
      0.61 +- 42%    +343.0%       2.72 +- 77%  sched_debug.rt_rq:/.rt_time.stddev
     91.98           -30.9       61.11 +- 70%  perf-profile.calltrace.cycles-pp.testcase
      9.05            -9.1        0.00        perf-profile.calltrace.cycles-pp.__do_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault.do_page_fault
      8.91            -8.9        0.00        perf-profile.calltrace.cycles-pp.shmem_fault.__do_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
      8.06            -8.1        0.00        perf-profile.calltrace.cycles-pp.shmem_getpage_gfp.shmem_fault.__do_fault.__handle_mm_fault.handle_mm_fault
      7.59            -7.6        0.00        perf-profile.calltrace.cycles-pp.find_lock_entry.shmem_getpage_gfp.shmem_fault.__do_fault.__handle_mm_fault
      7.44            -7.4        0.00        perf-profile.calltrace.cycles-pp.finish_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault.do_page_fault
      7.28            -7.3        0.00        perf-profile.calltrace.cycles-pp.alloc_set_pte.finish_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
      5.31            -5.3        0.00        perf-profile.calltrace.cycles-pp.page_add_file_rmap.alloc_set_pte.finish_fault.__handle_mm_fault.handle_mm_fault
      8.08            -2.8        5.30 +- 70%  perf-profile.calltrace.cycles-pp.native_irq_return_iret.testcase
      5.95            -2.1        3.83 +- 70%  perf-profile.calltrace.cycles-pp.find_get_entry.find_lock_entry.shmem_getpage_gfp.shmem_fault.__do_fault
      5.95            -2.0        3.93 +- 70%  perf-profile.calltrace.cycles-pp.swapgs_restore_regs_and_return_to_usermode.testcase
      3.10            -1.1        2.01 +- 70%  perf-profile.calltrace.cycles-pp.__perf_sw_event.__do_page_fault.do_page_fault.page_fault.testcase
      2.36            -0.8        1.55 +- 70%  perf-profile.calltrace.cycles-pp.___perf_sw_event.__perf_sw_event.__do_page_fault.do_page_fault.page_fault
      1.08            -0.4        0.70 +- 70%  perf-profile.calltrace.cycles-pp.do_page_fault.testcase
      0.82            -0.3        0.54 +- 70%  perf-profile.calltrace.cycles-pp.trace_graph_entry.do_page_fault.testcase
      0.77            -0.3        0.50 +- 70%  perf-profile.calltrace.cycles-pp.ftrace_graph_caller.__do_page_fault.do_page_fault.page_fault.testcase
      0.59            -0.2        0.37 +- 70%  perf-profile.calltrace.cycles-pp.down_read_trylock.__do_page_fault.do_page_fault.page_fault.testcase
     91.98           -30.9       61.11 +- 70%  perf-profile.children.cycles-pp.testcase
      9.14            -3.2        5.99 +- 70%  perf-profile.children.cycles-pp.__do_fault
      8.20            -2.8        5.40 +- 70%  perf-profile.children.cycles-pp.shmem_getpage_gfp
      8.08            -2.8        5.31 +- 70%  perf-profile.children.cycles-pp.native_irq_return_iret
      6.08            -2.2        3.92 +- 70%  perf-profile.children.cycles-pp.find_get_entry
      6.08            -2.1        3.96 +- 70%  perf-profile.children.cycles-pp.sync_regs
      5.95            -2.0        3.93 +- 70%  perf-profile.children.cycles-pp.swapgs_restore_regs_and_return_to_usermode
      4.12            -1.4        2.73 +- 70%  perf-profile.children.cycles-pp.ftrace_graph_caller
      3.65            -1.2        2.42 +- 70%  perf-profile.children.cycles-pp.prepare_ftrace_return
      3.18            -1.1        2.07 +- 70%  perf-profile.children.cycles-pp.__perf_sw_event
      2.34            -0.8        1.52 +- 70%  perf-profile.children.cycles-pp.fault_dirty_shared_page
      0.80            -0.3        0.50 +- 70%  perf-profile.children.cycles-pp._raw_spin_lock
      0.76            -0.3        0.50 +- 70%  perf-profile.children.cycles-pp.tlb_flush_mmu_free
      0.61            -0.2        0.39 +- 70%  perf-profile.children.cycles-pp.down_read_trylock
      0.48 +-  2%      -0.2        0.28 +- 70%  perf-profile.children.cycles-pp.pmd_devmap_trans_unstable
      0.26 +-  6%      -0.1        0.15 +- 71%  perf-profile.children.cycles-pp.ktime_get
      0.20 +-  2%      -0.1        0.12 +- 70%  perf-profile.children.cycles-pp.perf_exclude_event
      0.22 +-  2%      -0.1        0.13 +- 70%  perf-profile.children.cycles-pp._cond_resched
      0.17            -0.1        0.11 +- 70%  perf-profile.children.cycles-pp.page_rmapping
      0.13            -0.1        0.07 +- 70%  perf-profile.children.cycles-pp.rcu_all_qs
      0.07            -0.0        0.04 +- 70%  perf-profile.children.cycles-pp.ftrace_lookup_ip
     22.36            -7.8       14.59 +- 70%  perf-profile.self.cycles-pp.testcase
      8.08            -2.8        5.31 +- 70%  perf-profile.self.cycles-pp.native_irq_return_iret
      6.08            -2.1        3.96 +- 70%  perf-profile.self.cycles-pp.sync_regs
      5.81            -2.0        3.84 +- 70%  perf-profile.self.cycles-pp.swapgs_restore_regs_and_return_to_usermode
      3.27            -1.6        1.65 +- 70%  perf-profile.self.cycles-pp.__handle_mm_fault
      3.79            -1.4        2.36 +- 70%  perf-profile.self.cycles-pp.find_get_entry
      3.80            -1.3        2.53 +- 70%  perf-profile.self.cycles-pp.trace_graph_entry
      1.10            -0.5        0.57 +- 70%  perf-profile.self.cycles-pp.alloc_set_pte
      1.24            -0.4        0.81 +- 70%  perf-profile.self.cycles-pp.shmem_fault
      0.80            -0.3        0.50 +- 70%  perf-profile.self.cycles-pp._raw_spin_lock
      0.81            -0.3        0.51 +- 70%  perf-profile.self.cycles-pp.find_lock_entry
      0.80 +-  2%      -0.3        0.51 +- 70%  perf-profile.self.cycles-pp.__perf_sw_event
      0.61            -0.2        0.38 +- 70%  perf-profile.self.cycles-pp.down_read_trylock
      0.60            -0.2        0.39 +- 70%  perf-profile.self.cycles-pp.shmem_getpage_gfp
      0.48            -0.2        0.27 +- 70%  perf-profile.self.cycles-pp.pmd_devmap_trans_unstable
      0.47            -0.2        0.30 +- 70%  perf-profile.self.cycles-pp.file_update_time
      0.34            -0.1        0.22 +- 70%  perf-profile.self.cycles-pp.do_page_fault
      0.22 +-  4%      -0.1        0.11 +- 70%  perf-profile.self.cycles-pp.__do_fault
      0.25 +-  5%      -0.1        0.14 +- 71%  perf-profile.self.cycles-pp.ktime_get
      0.21 +-  2%      -0.1        0.12 +- 70%  perf-profile.self.cycles-pp.finish_fault
      0.23 +-  2%      -0.1        0.14 +- 70%  perf-profile.self.cycles-pp.fault_dirty_shared_page
      0.22 +-  2%      -0.1        0.14 +- 70%  perf-profile.self.cycles-pp.prepare_exit_to_usermode
      0.20 +-  2%      -0.1        0.12 +- 70%  perf-profile.self.cycles-pp.perf_exclude_event
      0.16            -0.1        0.10 +- 70%  perf-profile.self.cycles-pp._cond_resched
      0.13            -0.1        0.07 +- 70%  perf-profile.self.cycles-pp.rcu_all_qs
      0.07            -0.0        0.04 +- 70%  perf-profile.self.cycles-pp.ftrace_lookup_ip

=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/nr_task/thp_enabled/test/cpufreq_governor:
  lkp-skl-4sp1/will-it-scale/debian-x86_64-2018-04-03.cgz/x86_64-rhel-7.2/gcc-7/100%/always/context_switch1/performance

commit: 
  ba98a1cdad71d259a194461b3a61471b49b14df1
  a7a8993bfe3ccb54ad468b9f1799649e4ad1ff12

ba98a1cdad71d259 a7a8993bfe3ccb54ad468b9f17 
---------------- -------------------------- 
       fail:runs  %reproduction    fail:runs
           |             |             |    
           :3           33%           1:3     dmesg.WARNING:at#for_ip_interrupt_entry/0x
           :3           33%           1:3     dmesg.WARNING:at#for_ip_ret_from_intr/0x
           :3           67%           2:3     kmsg.pstore:crypto_comp_decompress_failed,ret=
           :3           67%           2:3     kmsg.pstore:decompression_failed
         %stddev     %change         %stddev
             \          |                \  
    223910            -1.3%     220930        will-it-scale.per_process_ops
    233722            -1.0%     231288        will-it-scale.per_thread_ops
 6.001e+08 +- 13%     +31.4%  7.887e+08 +-  4%  will-it-scale.time.involuntary_context_switches
     18003 +-  4%     +10.9%      19956        will-it-scale.time.minor_page_faults
  1.29e+10            -2.5%  1.258e+10        will-it-scale.time.voluntary_context_switches
  87865617            -1.2%   86826277        will-it-scale.workload
   2880329 +-  2%      +5.4%    3034904        interrupts.CAL:Function_call_interrupts
   7695018           -23.3%    5905066 +-  8%  meminfo.DirectMap2M
      0.00 +- 39%      -0.0        0.00 +- 78%  mpstat.cpu.iowait%
      4621 +- 12%     +13.4%       5241        proc-vmstat.numa_hint_faults_local
    715714           +27.6%     913142 +- 13%  softirqs.SCHED
    515653 +-  6%     -20.0%     412650 +- 15%  turbostat.C1
  43643516            -1.2%   43127031        vmstat.system.cs
   2893393 +-  4%     -23.6%    2210524 +- 10%  cpuidle.C1.time
    518051 +-  6%     -19.9%     415081 +- 15%  cpuidle.C1.usage
     23.10           +22.9%      28.38 +-  9%  boot-time.boot
     18.38           +23.2%      22.64 +- 12%  boot-time.dhcp
      5216            +5.0%       5478 +-  2%  boot-time.idle
    963.76 +- 44%    +109.7%       2021 +- 34%  irq_exception_noise.__do_page_fault.sum
      6.33 +- 14%    +726.3%      52.33 +- 62%  irq_exception_noise.irq_time
     56524 +-  7%     -18.8%      45915 +-  4%  irq_exception_noise.softirq_time
 6.001e+08 +- 13%     +31.4%  7.887e+08 +-  4%  time.involuntary_context_switches
     18003 +-  4%     +10.9%      19956        time.minor_page_faults
  1.29e+10            -2.5%  1.258e+10        time.voluntary_context_switches
      1386 +-  7%     +15.4%       1600 +- 11%  slabinfo.scsi_sense_cache.active_objs
      1386 +-  7%     +15.4%       1600 +- 11%  slabinfo.scsi_sense_cache.num_objs
      1427 +-  5%      -8.9%       1299 +-  2%  slabinfo.task_group.active_objs
      1427 +-  5%      -8.9%       1299 +-  2%  slabinfo.task_group.num_objs
     65519 +- 12%     +20.6%      79014 +- 16%  numa-meminfo.node0.SUnreclaim
      8484           -11.9%       7475 +-  7%  numa-meminfo.node1.KernelStack
      9264 +- 26%     -33.7%       6146 +-  7%  numa-meminfo.node1.Mapped
      2138 +- 61%    +373.5%      10127 +- 92%  numa-meminfo.node3.Inactive
      2059 +- 61%    +387.8%      10046 +- 93%  numa-meminfo.node3.Inactive(anon)
     16379 +- 12%     +20.6%      19752 +- 16%  numa-vmstat.node0.nr_slab_unreclaimable
      8483           -11.9%       7474 +-  7%  numa-vmstat.node1.nr_kernel_stack
      6250 +- 29%     -42.8%       3575 +- 24%  numa-vmstat.node2
      3798 +- 17%     +63.7%       6218 +-  5%  numa-vmstat.node3
    543.00 +- 61%    +368.1%       2541 +- 91%  numa-vmstat.node3.nr_inactive_anon
    543.33 +- 61%    +367.8%       2541 +- 91%  numa-vmstat.node3.nr_zone_inactive_anon
 4.138e+13            -1.1%   4.09e+13        perf-stat.branch-instructions
 6.569e+11            -2.0%  6.441e+11        perf-stat.branch-misses
 2.645e+10            -1.2%  2.613e+10        perf-stat.context-switches
      1.21            +1.2%       1.23        perf-stat.cpi
    153343 +-  2%     -12.1%     134776        perf-stat.cpu-migrations
 5.966e+13            -1.3%  5.889e+13        perf-stat.dTLB-loads
 3.736e+13            -1.2%   3.69e+13        perf-stat.dTLB-stores
      5.85 +- 15%      +8.8       14.67 +-  9%  perf-stat.iTLB-load-miss-rate%
 3.736e+09 +- 17%    +161.3%   9.76e+09 +- 11%  perf-stat.iTLB-load-misses
 5.987e+10            -5.4%  5.667e+10        perf-stat.iTLB-loads
 2.079e+14            -1.2%  2.054e+14        perf-stat.instructions
     57547 +- 18%     -62.9%      21340 +- 11%  perf-stat.instructions-per-iTLB-miss
      0.82            -1.2%       0.81        perf-stat.ipc
  27502531 +-  8%      +9.5%   30122136 +-  3%  perf-stat.node-store-misses
      1449 +- 27%     -34.6%     948.85        sched_debug.cfs_rq:/.load.min
    319416 +-115%    -188.5%    -282549        sched_debug.cfs_rq:/.spread0.avg
    657044 +- 55%     -88.3%      76887 +- 23%  sched_debug.cfs_rq:/.spread0.max
  -1525243           +54.6%   -2357898        sched_debug.cfs_rq:/.spread0.min
    101614 +-  6%     +30.6%     132713 +- 19%  sched_debug.cpu.avg_idle.stddev
     11.54 +- 41%     -61.2%       4.48        sched_debug.cpu.cpu_load[1].avg
      1369 +- 67%     -98.5%      20.67 +- 48%  sched_debug.cpu.cpu_load[1].max
     99.29 +- 67%     -97.6%       2.35 +- 26%  sched_debug.cpu.cpu_load[1].stddev
      9.58 +- 38%     -55.2%       4.29        sched_debug.cpu.cpu_load[2].avg
      1024 +- 68%     -98.5%      15.27 +- 36%  sched_debug.cpu.cpu_load[2].max
     74.51 +- 67%     -97.3%       1.99 +- 15%  sched_debug.cpu.cpu_load[2].stddev
      7.37 +- 29%     -42.0%       4.28        sched_debug.cpu.cpu_load[3].avg
    600.58 +- 68%     -97.9%      12.48 +- 20%  sched_debug.cpu.cpu_load[3].max
     43.98 +- 66%     -95.8%       1.83 +-  5%  sched_debug.cpu.cpu_load[3].stddev
      5.95 +- 19%     -28.1%       4.28        sched_debug.cpu.cpu_load[4].avg
    325.39 +- 67%     -96.4%      11.67 +- 10%  sched_debug.cpu.cpu_load[4].max
     24.19 +- 65%     -92.5%       1.81 +-  3%  sched_debug.cpu.cpu_load[4].stddev
    907.23 +-  4%     -14.1%     779.70 +- 10%  sched_debug.cpu.nr_load_updates.stddev
      0.00 +- 83%    +122.5%       0.00        sched_debug.rt_rq:/.rt_time.min
      8.49 +-  2%      -0.3        8.21 +-  2%  perf-profile.calltrace.cycles-pp.dequeue_task_fair.__schedule.schedule.pipe_wait.pipe_read
     57.28            -0.3       57.01        perf-profile.calltrace.cycles-pp.read
      5.06            -0.2        4.85        perf-profile.calltrace.cycles-pp.select_task_rq_fair.try_to_wake_up.autoremove_wake_function.__wake_up_common.__wake_up_common_lock
      4.98            -0.2        4.78        perf-profile.calltrace.cycles-pp.__switch_to.read
      3.55            -0.2        3.39 +-  2%  perf-profile.calltrace.cycles-pp.syscall_return_via_sysret.read
      2.72            -0.1        2.60        perf-profile.calltrace.cycles-pp.reweight_entity.enqueue_task_fair.ttwu_do_activate.try_to_wake_up.autoremove_wake_function
      2.67            -0.1        2.57 +-  2%  perf-profile.calltrace.cycles-pp.reweight_entity.dequeue_task_fair.__schedule.schedule.pipe_wait
      3.40            -0.1        3.31        perf-profile.calltrace.cycles-pp.syscall_return_via_sysret.write
      3.77            -0.1        3.68        perf-profile.calltrace.cycles-pp.select_idle_sibling.select_task_rq_fair.try_to_wake_up.autoremove_wake_function.__wake_up_common
      1.95            -0.1        1.88        perf-profile.calltrace.cycles-pp.copy_page_to_iter.pipe_read.__vfs_read.vfs_read.ksys_read
      2.19            -0.1        2.13        perf-profile.calltrace.cycles-pp.__switch_to_asm.read
      1.30            -0.1        1.25        perf-profile.calltrace.cycles-pp.update_curr.reweight_entity.enqueue_task_fair.ttwu_do_activate.try_to_wake_up
      1.27            -0.1        1.22 +-  2%  perf-profile.calltrace.cycles-pp.update_curr.reweight_entity.dequeue_task_fair.__schedule.schedule
      2.29            -0.0        2.24        perf-profile.calltrace.cycles-pp.load_new_mm_cr3.switch_mm_irqs_off.__schedule.schedule.pipe_wait
      0.96            -0.0        0.92        perf-profile.calltrace.cycles-pp.__calc_delta.update_curr.reweight_entity.dequeue_task_fair.__schedule
      0.85            -0.0        0.81 +-  3%  perf-profile.calltrace.cycles-pp.cpumask_next_wrap.select_idle_sibling.select_task_rq_fair.try_to_wake_up.autoremove_wake_function
      1.63            -0.0        1.59        perf-profile.calltrace.cycles-pp.native_write_msr.read
      0.72            -0.0        0.69        perf-profile.calltrace.cycles-pp.copyout.copy_page_to_iter.pipe_read.__vfs_read.vfs_read
      0.65 +-  2%      -0.0        0.62        perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.try_to_wake_up.autoremove_wake_function.__wake_up_common.__wake_up_common_lock
      0.61            -0.0        0.58 +-  2%  perf-profile.calltrace.cycles-pp.find_next_bit.cpumask_next_wrap.select_idle_sibling.select_task_rq_fair.try_to_wake_up
      0.88            -0.0        0.85        perf-profile.calltrace.cycles-pp.touch_atime.pipe_read.__vfs_read.vfs_read.ksys_read
      0.80            -0.0        0.77 +-  2%  perf-profile.calltrace.cycles-pp.___perf_sw_event.__schedule.schedule.pipe_wait.pipe_read
      0.82            -0.0        0.79        perf-profile.calltrace.cycles-pp.prepare_to_wait.pipe_wait.pipe_read.__vfs_read.vfs_read
      0.72            -0.0        0.70        perf-profile.calltrace.cycles-pp.mutex_lock.pipe_write.__vfs_write.vfs_write.ksys_write
      0.56 +-  2%      -0.0        0.53        perf-profile.calltrace.cycles-pp.update_rq_clock.try_to_wake_up.autoremove_wake_function.__wake_up_common.__wake_up_common_lock
      0.83            -0.0        0.81        perf-profile.calltrace.cycles-pp.__wake_up_common_lock.pipe_read.__vfs_read.vfs_read.ksys_read
     42.40            +0.3       42.69        perf-profile.calltrace.cycles-pp.write
     31.80            +0.4       32.18        perf-profile.calltrace.cycles-pp.__vfs_read.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
     24.35            +0.5       24.84        perf-profile.calltrace.cycles-pp.pipe_wait.pipe_read.__vfs_read.vfs_read.ksys_read
     20.36            +0.6       20.92 +-  2%  perf-profile.calltrace.cycles-pp.try_to_wake_up.autoremove_wake_function.__wake_up_common.__wake_up_common_lock.pipe_write
     22.01            +0.6       22.58        perf-profile.calltrace.cycles-pp.schedule.pipe_wait.pipe_read.__vfs_read.vfs_read
     21.87            +0.6       22.46        perf-profile.calltrace.cycles-pp.__schedule.schedule.pipe_wait.pipe_read.__vfs_read
      3.15 +- 11%      +1.0        4.12 +- 14%  perf-profile.calltrace.cycles-pp.ttwu_do_wakeup.try_to_wake_up.autoremove_wake_function.__wake_up_common.__wake_up_common_lock
      1.07 +- 34%      +1.1        2.12 +- 31%  perf-profile.calltrace.cycles-pp.tracing_record_taskinfo_sched_switch.__schedule.schedule.pipe_wait.pipe_read
      0.66 +- 75%      +1.1        1.72 +- 37%  perf-profile.calltrace.cycles-pp.trace_save_cmdline.tracing_record_taskinfo.ttwu_do_wakeup.try_to_wake_up.autoremove_wake_function
      0.75 +- 74%      +1.1        1.88 +- 34%  perf-profile.calltrace.cycles-pp.tracing_record_taskinfo.ttwu_do_wakeup.try_to_wake_up.autoremove_wake_function.__wake_up_common
      0.69 +- 76%      +1.2        1.85 +- 36%  perf-profile.calltrace.cycles-pp.trace_save_cmdline.tracing_record_taskinfo_sched_switch.__schedule.schedule.pipe_wait
      8.73 +-  2%      -0.3        8.45        perf-profile.children.cycles-pp.dequeue_task_fair
     57.28            -0.3       57.01        perf-profile.children.cycles-pp.read
      6.95            -0.2        6.70        perf-profile.children.cycles-pp.syscall_return_via_sysret
      5.57            -0.2        5.35        perf-profile.children.cycles-pp.reweight_entity
      5.26            -0.2        5.05        perf-profile.children.cycles-pp.select_task_rq_fair
      5.19            -0.2        4.99        perf-profile.children.cycles-pp.__switch_to
      4.90            -0.2        4.73 +-  2%  perf-profile.children.cycles-pp.update_curr
      1.27            -0.1        1.13 +-  8%  perf-profile.children.cycles-pp.fsnotify
      3.92            -0.1        3.83        perf-profile.children.cycles-pp.select_idle_sibling
      2.01            -0.1        1.93        perf-profile.children.cycles-pp.__calc_delta
      2.14            -0.1        2.06        perf-profile.children.cycles-pp.copy_page_to_iter
      1.58            -0.1        1.51        perf-profile.children.cycles-pp._raw_spin_unlock_irqrestore
      2.90            -0.1        2.84        perf-profile.children.cycles-pp.update_cfs_group
      1.93            -0.1        1.87        perf-profile.children.cycles-pp._raw_spin_lock_irqsave
      2.35            -0.1        2.29        perf-profile.children.cycles-pp.__switch_to_asm
      1.33            -0.1        1.27 +-  3%  perf-profile.children.cycles-pp.cpumask_next_wrap
      2.57            -0.1        2.52        perf-profile.children.cycles-pp.load_new_mm_cr3
      1.53            -0.1        1.47 +-  2%  perf-profile.children.cycles-pp.__fdget_pos
      1.11            -0.0        1.07 +-  2%  perf-profile.children.cycles-pp.find_next_bit
      1.18            -0.0        1.14        perf-profile.children.cycles-pp.update_rq_clock
      0.88            -0.0        0.83        perf-profile.children.cycles-pp.copy_user_generic_unrolled
      1.70            -0.0        1.65        perf-profile.children.cycles-pp.native_write_msr
      0.97            -0.0        0.93 +-  2%  perf-profile.children.cycles-pp.account_entity_dequeue
      0.59            -0.0        0.56        perf-profile.children.cycles-pp.finish_task_switch
      0.91            -0.0        0.88        perf-profile.children.cycles-pp.touch_atime
      0.69            -0.0        0.65        perf-profile.children.cycles-pp.account_entity_enqueue
      2.13            -0.0        2.09        perf-profile.children.cycles-pp.mutex_lock
      0.32 +-  3%      -0.0        0.29 +-  4%  perf-profile.children.cycles-pp.__sb_start_write
      0.84            -0.0        0.81 +-  2%  perf-profile.children.cycles-pp.___perf_sw_event
      0.89            -0.0        0.87        perf-profile.children.cycles-pp.prepare_to_wait
      0.73            -0.0        0.71        perf-profile.children.cycles-pp.copyout
      0.31 +-  2%      -0.0        0.28 +-  3%  perf-profile.children.cycles-pp.__list_del_entry_valid
      0.46 +-  2%      -0.0        0.44        perf-profile.children.cycles-pp.anon_pipe_buf_release
      0.38            -0.0        0.36 +-  3%  perf-profile.children.cycles-pp.idle_cpu
      0.32            -0.0        0.30 +-  2%  perf-profile.children.cycles-pp.__x64_sys_read
      0.21 +-  2%      -0.0        0.20 +-  2%  perf-profile.children.cycles-pp.deactivate_task
      0.13            -0.0        0.12 +-  4%  perf-profile.children.cycles-pp.timespec_trunc
      0.09            -0.0        0.08        perf-profile.children.cycles-pp.iov_iter_init
      0.08            -0.0        0.07        perf-profile.children.cycles-pp.native_load_tls
      0.11 +-  4%      +0.0        0.12        perf-profile.children.cycles-pp.tick_sched_timer
      0.08 +-  5%      +0.0        0.10 +-  4%  perf-profile.children.cycles-pp.finish_wait
      0.38 +-  2%      +0.0        0.40 +-  2%  perf-profile.children.cycles-pp.file_update_time
      0.31            +0.0        0.33 +-  2%  perf-profile.children.cycles-pp.smp_apic_timer_interrupt
      0.24 +-  3%      +0.0        0.26 +-  3%  perf-profile.children.cycles-pp.rcu_all_qs
      0.39            +0.0        0.41        perf-profile.children.cycles-pp._cond_resched
      0.05            +0.0        0.07 +-  6%  perf-profile.children.cycles-pp.default_wake_function
      0.23 +-  2%      +0.0        0.26 +-  3%  perf-profile.children.cycles-pp.current_time
      0.30            +0.0        0.35 +-  2%  perf-profile.children.cycles-pp.generic_pipe_buf_confirm
      0.52            +0.1        0.58        perf-profile.children.cycles-pp.entry_SYSCALL_64_stage2
      0.00            +0.1        0.08 +-  5%  perf-profile.children.cycles-pp.hrtick_update
     42.40            +0.3       42.69        perf-profile.children.cycles-pp.write
     31.86            +0.4       32.26        perf-profile.children.cycles-pp.__vfs_read
     24.40            +0.5       24.89        perf-profile.children.cycles-pp.pipe_wait
     20.40            +0.6       20.96 +-  2%  perf-profile.children.cycles-pp.try_to_wake_up
     22.30            +0.6       22.89        perf-profile.children.cycles-pp.schedule
     22.22            +0.6       22.84        perf-profile.children.cycles-pp.__schedule
      0.99 +- 36%      +0.9        1.94 +- 32%  perf-profile.children.cycles-pp.tracing_record_taskinfo
      3.30 +- 10%      +1.0        4.27 +- 13%  perf-profile.children.cycles-pp.ttwu_do_wakeup
      1.14 +- 31%      +1.1        2.24 +- 29%  perf-profile.children.cycles-pp.tracing_record_taskinfo_sched_switch
      1.59 +- 46%      +2.0        3.60 +- 36%  perf-profile.children.cycles-pp.trace_save_cmdline
      6.95            -0.2        6.70        perf-profile.self.cycles-pp.syscall_return_via_sysret
      5.19            -0.2        4.99        perf-profile.self.cycles-pp.__switch_to
      1.27            -0.1        1.12 +-  8%  perf-profile.self.cycles-pp.fsnotify
      1.49            -0.1        1.36        perf-profile.self.cycles-pp.select_task_rq_fair
      2.47            -0.1        2.37 +-  2%  perf-profile.self.cycles-pp.reweight_entity
      0.29            -0.1        0.19 +-  2%  perf-profile.self.cycles-pp.ksys_read
      1.50            -0.1        1.42        perf-profile.self.cycles-pp._raw_spin_unlock_irqrestore
      2.01            -0.1        1.93        perf-profile.self.cycles-pp.__calc_delta
      1.93            -0.1        1.86        perf-profile.self.cycles-pp._raw_spin_lock_irqsave
      1.47            -0.1        1.40        perf-profile.self.cycles-pp.dequeue_task_fair
      2.90            -0.1        2.84        perf-profile.self.cycles-pp.update_cfs_group
      1.29            -0.1        1.23        perf-profile.self.cycles-pp.do_syscall_64
      2.57            -0.1        2.52        perf-profile.self.cycles-pp.load_new_mm_cr3
      2.28            -0.1        2.23        perf-profile.self.cycles-pp.__switch_to_asm
      1.80            -0.1        1.75        perf-profile.self.cycles-pp.select_idle_sibling
      1.11            -0.0        1.07 +-  2%  perf-profile.self.cycles-pp.find_next_bit
      0.87            -0.0        0.83        perf-profile.self.cycles-pp.copy_user_generic_unrolled
      0.43            -0.0        0.39 +-  2%  perf-profile.self.cycles-pp.dequeue_entity
      1.70            -0.0        1.65        perf-profile.self.cycles-pp.native_write_msr
      0.92            -0.0        0.88 +-  2%  perf-profile.self.cycles-pp.account_entity_dequeue
      0.48            -0.0        0.44        perf-profile.self.cycles-pp.finish_task_switch
      0.77            -0.0        0.74        perf-profile.self.cycles-pp.___perf_sw_event
      0.66            -0.0        0.63        perf-profile.self.cycles-pp.account_entity_enqueue
      0.46 +-  2%      -0.0        0.43 +-  2%  perf-profile.self.cycles-pp.anon_pipe_buf_release
      0.32 +-  3%      -0.0        0.29 +-  4%  perf-profile.self.cycles-pp.__sb_start_write
      0.31 +-  2%      -0.0        0.28 +-  3%  perf-profile.self.cycles-pp.__list_del_entry_valid
      0.38            -0.0        0.36 +-  3%  perf-profile.self.cycles-pp.idle_cpu
      0.19 +-  4%      -0.0        0.17 +-  2%  perf-profile.self.cycles-pp.__fdget_pos
      0.50            -0.0        0.48        perf-profile.self.cycles-pp.__atime_needs_update
      0.23 +-  2%      -0.0        0.21 +-  3%  perf-profile.self.cycles-pp.touch_atime
      0.31            -0.0        0.30        perf-profile.self.cycles-pp.__x64_sys_read
      0.21 +-  2%      -0.0        0.20 +-  2%  perf-profile.self.cycles-pp.deactivate_task
      0.21 +-  2%      -0.0        0.19        perf-profile.self.cycles-pp.check_preempt_curr
      0.40            -0.0        0.39        perf-profile.self.cycles-pp.autoremove_wake_function
      0.40            -0.0        0.38        perf-profile.self.cycles-pp.copy_user_enhanced_fast_string
      0.27            -0.0        0.26        perf-profile.self.cycles-pp.pipe_wait
      0.13            -0.0        0.12 +-  4%  perf-profile.self.cycles-pp.timespec_trunc
      0.22 +-  2%      -0.0        0.20 +-  2%  perf-profile.self.cycles-pp.put_prev_entity
      0.09            -0.0        0.08        perf-profile.self.cycles-pp.iov_iter_init
      0.08            -0.0        0.07        perf-profile.self.cycles-pp.native_load_tls
      0.11            -0.0        0.10        perf-profile.self.cycles-pp.schedule
      0.12 +-  4%      +0.0        0.13        perf-profile.self.cycles-pp.copyin
      0.08 +-  5%      +0.0        0.10 +-  4%  perf-profile.self.cycles-pp.finish_wait
      0.18            +0.0        0.20 +-  2%  perf-profile.self.cycles-pp.ttwu_do_activate
      0.28 +-  2%      +0.0        0.30 +-  2%  perf-profile.self.cycles-pp._cond_resched
      0.24 +-  3%      +0.0        0.26 +-  3%  perf-profile.self.cycles-pp.rcu_all_qs
      0.05            +0.0        0.07 +-  6%  perf-profile.self.cycles-pp.default_wake_function
      0.08 +- 14%      +0.0        0.11 +- 14%  perf-profile.self.cycles-pp.tracing_record_taskinfo_sched_switch
      0.51            +0.0        0.55 +-  4%  perf-profile.self.cycles-pp.vfs_write
      0.30            +0.0        0.35 +-  2%  perf-profile.self.cycles-pp.generic_pipe_buf_confirm
      0.52            +0.1        0.58        perf-profile.self.cycles-pp.entry_SYSCALL_64_stage2
      0.00            +0.1        0.08 +-  5%  perf-profile.self.cycles-pp.hrtick_update
      1.97            +0.1        2.07 +-  2%  perf-profile.self.cycles-pp.switch_mm_irqs_off
      1.59 +- 46%      +2.0        3.60 +- 36%  perf-profile.self.cycles-pp.trace_save_cmdline

=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/nr_task/thp_enabled/test/cpufreq_governor:
  lkp-skl-4sp1/will-it-scale/debian-x86_64-2018-04-03.cgz/x86_64-rhel-7.2/gcc-7/100%/never/brk1/performance

commit: 
  ba98a1cdad71d259a194461b3a61471b49b14df1
  a7a8993bfe3ccb54ad468b9f1799649e4ad1ff12

ba98a1cdad71d259 a7a8993bfe3ccb54ad468b9f17 
---------------- -------------------------- 
       fail:runs  %reproduction    fail:runs
           |             |             |    
           :3           33%           1:3     kmsg.pstore:crypto_comp_decompress_failed,ret=
           :3           33%           1:3     kmsg.pstore:decompression_failed
         %stddev     %change         %stddev
             \          |                \  
    997317            -2.0%     977778        will-it-scale.per_process_ops
    957.00            -7.9%     881.00 +-  3%  will-it-scale.per_thread_ops
     18.42 +-  3%      -8.2%      16.90        will-it-scale.time.user_time
 1.917e+08            -2.0%  1.879e+08        will-it-scale.workload
     18.42 +-  3%      -8.2%      16.90        time.user_time
      0.30 +- 11%     -36.7%       0.19 +- 11%  turbostat.Pkg%pc2
     57539 +- 51%    +140.6%     138439 +- 31%  meminfo.CmaFree
    410877 +- 11%     -22.1%     320082 +- 22%  meminfo.DirectMap4k
    343575 +- 27%     +71.3%     588703 +- 31%  numa-numastat.node0.local_node
    374176 +- 24%     +63.3%     611007 +- 27%  numa-numastat.node0.numa_hit
   1056347 +-  4%     -39.9%     634843 +- 38%  numa-numastat.node3.local_node
   1060682 +-  4%     -39.0%     646862 +- 35%  numa-numastat.node3.numa_hit
     14383 +- 51%    +140.6%      34608 +- 31%  proc-vmstat.nr_free_cma
    179.00            +2.4%     183.33        proc-vmstat.nr_inactive_file
    179.00            +2.4%     183.33        proc-vmstat.nr_zone_inactive_file
    564483 +-  3%     -38.0%     350064 +- 36%  proc-vmstat.pgalloc_movable
   1811959           +10.8%    2008488 +-  5%  proc-vmstat.pgalloc_normal
      7153 +- 42%     -94.0%     431.33 +-119%  latency_stats.max.pipe_write.__vfs_write.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
      6627 +-141%    +380.5%      31843 +-110%  latency_stats.max.call_rwsem_down_write_failed_killable.do_mprotect_pkey.__x64_sys_mprotect.do_syscall_64.entry_SYSCALL_64_after_hwframe
     15244 +- 31%     -99.9%      15.00 +-141%  latency_stats.sum.call_rwsem_down_read_failed.__do_page_fault.do_page_fault.page_fault.__get_user_8.exit_robust_list.mm_release.do_exit.do_group_exit.get_signal.do_signal.exit_to_usermode_loop
      4301 +-117%     -83.7%     700.33 +-  6%  latency_stats.sum.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_getattr.__nfs_revalidate_inode.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_lookupat
     12153 +- 28%     -83.1%       2056 +- 70%  latency_stats.sum.pipe_write.__vfs_write.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
      6772 +-141%   +1105.8%      81665 +-127%  latency_stats.sum.call_rwsem_down_write_failed_killable.do_mprotect_pkey.__x64_sys_mprotect.do_syscall_64.entry_SYSCALL_64_after_hwframe
 2.465e+13            -1.3%  2.434e+13        perf-stat.branch-instructions
 2.691e+11            -2.1%  2.635e+11        perf-stat.branch-misses
 3.402e+13            -1.4%  3.355e+13        perf-stat.dTLB-loads
 1.694e+13            +1.4%  1.718e+13        perf-stat.dTLB-stores
      1.75 +- 50%      +4.7        6.45 +- 11%  perf-stat.iTLB-load-miss-rate%
 4.077e+08 +- 48%    +232.3%  1.355e+09 +- 11%  perf-stat.iTLB-load-misses
  2.31e+10 +-  2%     -14.9%  1.965e+10 +-  3%  perf-stat.iTLB-loads
 1.163e+14            -1.6%  1.144e+14        perf-stat.instructions
    346171 +- 36%     -75.3%      85575 +- 11%  perf-stat.instructions-per-iTLB-miss
 6.174e+08 +-  2%      -9.5%  5.589e+08        perf-stat.node-store-misses
    595.00 +- 10%     +31.4%     782.00 +-  3%  slabinfo.Acpi-State.active_objs
    595.00 +- 10%     +31.4%     782.00 +-  3%  slabinfo.Acpi-State.num_objs
      2831 +-  3%     -14.0%       2434 +-  5%  slabinfo.avtab_node.active_objs
      2831 +-  3%     -14.0%       2434 +-  5%  slabinfo.avtab_node.num_objs
    934.00           -10.9%     832.33 +-  5%  slabinfo.inotify_inode_mark.active_objs
    934.00           -10.9%     832.33 +-  5%  slabinfo.inotify_inode_mark.num_objs
      1232 +-  4%     +13.4%       1397 +-  6%  slabinfo.nsproxy.active_objs
      1232 +-  4%     +13.4%       1397 +-  6%  slabinfo.nsproxy.num_objs
    499.67 +- 12%     +24.8%     623.67 +- 10%  slabinfo.secpath_cache.active_objs
    499.67 +- 12%     +24.8%     623.67 +- 10%  slabinfo.secpath_cache.num_objs
     31393 +- 84%    +220.1%     100477 +- 21%  numa-meminfo.node0.Active
     31393 +- 84%    +220.1%     100477 +- 21%  numa-meminfo.node0.Active(anon)
     30013 +- 85%    +232.1%      99661 +- 21%  numa-meminfo.node0.AnonPages
     21603 +- 34%     -85.0%       3237 +-100%  numa-meminfo.node0.Inactive
     21528 +- 34%     -85.0%       3237 +-100%  numa-meminfo.node0.Inactive(anon)
     10247 +- 35%     -46.4%       5495        numa-meminfo.node0.Mapped
     35388 +- 14%     -41.6%      20670 +- 15%  numa-meminfo.node0.SReclaimable
     22911 +- 29%     -82.3%       4057 +- 84%  numa-meminfo.node0.Shmem
    117387 +-  9%     -22.5%      90986 +- 12%  numa-meminfo.node0.Slab
     68863 +- 67%     +77.7%     122351 +- 13%  numa-meminfo.node1.Active
     68863 +- 67%     +77.7%     122351 +- 13%  numa-meminfo.node1.Active(anon)
    228376           +22.3%     279406 +- 17%  numa-meminfo.node1.FilePages
      1481 +-116%   +1062.1%      17218 +- 39%  numa-meminfo.node1.Inactive
      1481 +-116%   +1062.0%      17216 +- 39%  numa-meminfo.node1.Inactive(anon)
      6593 +-  2%     +11.7%       7367 +-  3%  numa-meminfo.node1.KernelStack
    596227 +-  8%     +18.0%     703748 +-  4%  numa-meminfo.node1.MemUsed
     15298 +- 12%     +88.5%      28843 +- 36%  numa-meminfo.node1.SReclaimable
     52718 +-  9%     +21.0%      63810 +- 11%  numa-meminfo.node1.SUnreclaim
      1808 +- 97%   +2723.8%      51054 +- 97%  numa-meminfo.node1.Shmem
     68017 +-  5%     +36.2%      92654 +- 18%  numa-meminfo.node1.Slab
    125541 +- 29%     -64.9%      44024 +- 98%  numa-meminfo.node3.Active
    125137 +- 29%     -65.0%      43823 +- 98%  numa-meminfo.node3.Active(anon)
     93173 +- 25%     -87.8%      11381 +- 20%  numa-meminfo.node3.AnonPages
      9150 +-  5%      -9.3%       8301 +-  8%  numa-meminfo.node3.KernelStack
      7848 +- 84%    +220.0%      25118 +- 21%  numa-vmstat.node0.nr_active_anon
      7503 +- 85%    +232.1%      24914 +- 21%  numa-vmstat.node0.nr_anon_pages
      5381 +- 34%     -85.0%     809.00 +-100%  numa-vmstat.node0.nr_inactive_anon
      2559 +- 35%     -46.4%       1372        numa-vmstat.node0.nr_mapped
      5727 +- 29%     -82.3%       1014 +- 84%  numa-vmstat.node0.nr_shmem
      8846 +- 14%     -41.6%       5167 +- 15%  numa-vmstat.node0.nr_slab_reclaimable
      7848 +- 84%    +220.0%      25118 +- 21%  numa-vmstat.node0.nr_zone_active_anon
      5381 +- 34%     -85.0%     809.00 +-100%  numa-vmstat.node0.nr_zone_inactive_anon
      4821 +-  2%     +30.3%       6283 +- 15%  numa-vmstat.node1
     17215 +- 67%     +77.7%      30591 +- 13%  numa-vmstat.node1.nr_active_anon
     57093           +22.3%      69850 +- 17%  numa-vmstat.node1.nr_file_pages
    370.00 +-116%   +1061.8%       4298 +- 39%  numa-vmstat.node1.nr_inactive_anon
      6593 +-  2%     +11.7%       7366 +-  3%  numa-vmstat.node1.nr_kernel_stack
    451.67 +- 97%   +2725.6%      12762 +- 97%  numa-vmstat.node1.nr_shmem
      3824 +- 12%     +88.6%       7211 +- 36%  numa-vmstat.node1.nr_slab_reclaimable
     13179 +-  9%     +21.0%      15952 +- 11%  numa-vmstat.node1.nr_slab_unreclaimable
     17215 +- 67%     +77.7%      30591 +- 13%  numa-vmstat.node1.nr_zone_active_anon
    370.00 +-116%   +1061.8%       4298 +- 39%  numa-vmstat.node1.nr_zone_inactive_anon
    364789 +- 12%     +62.8%     593926 +- 34%  numa-vmstat.node1.numa_hit
    239539 +- 19%     +95.4%     468113 +- 43%  numa-vmstat.node1.numa_local
     71.00 +- 28%     +42.3%     101.00        numa-vmstat.node2.nr_mlock
     31285 +- 29%     -65.0%      10960 +- 98%  numa-vmstat.node3.nr_active_anon
     23292 +- 25%     -87.8%       2844 +- 19%  numa-vmstat.node3.nr_anon_pages
     14339 +- 52%    +141.1%      34566 +- 32%  numa-vmstat.node3.nr_free_cma
      9151 +-  5%      -9.3%       8299 +-  8%  numa-vmstat.node3.nr_kernel_stack
     31305 +- 29%     -64.9%      10975 +- 98%  numa-vmstat.node3.nr_zone_active_anon
    930131 +-  3%     -35.9%     596006 +- 34%  numa-vmstat.node3.numa_hit
    836455 +-  3%     -40.9%     493947 +- 44%  numa-vmstat.node3.numa_local
     75182 +- 58%     -83.8%      12160 +-  2%  sched_debug.cfs_rq:/.load.max
      6.65 +-  5%     -10.6%       5.94 +-  6%  sched_debug.cfs_rq:/.load_avg.avg
      0.16 +-  7%     +22.6%       0.20 +- 12%  sched_debug.cfs_rq:/.nr_running.stddev
      5.58 +- 24%    +427.7%      29.42 +- 93%  sched_debug.cfs_rq:/.nr_spread_over.max
      0.54 +- 15%    +306.8%       2.19 +- 86%  sched_debug.cfs_rq:/.nr_spread_over.stddev
      1.05 +- 25%     -65.1%       0.37 +- 71%  sched_debug.cfs_rq:/.removed.load_avg.avg
      9.62 +- 11%     -50.7%       4.74 +- 70%  sched_debug.cfs_rq:/.removed.load_avg.stddev
     48.70 +- 25%     -65.1%      17.02 +- 71%  sched_debug.cfs_rq:/.removed.runnable_sum.avg
    444.31 +- 11%     -50.7%     219.26 +- 70%  sched_debug.cfs_rq:/.removed.runnable_sum.stddev
      0.47 +- 13%     -60.9%       0.19 +- 71%  sched_debug.cfs_rq:/.removed.util_avg.avg
      4.47 +-  4%     -46.5%       2.39 +- 70%  sched_debug.cfs_rq:/.removed.util_avg.stddev
      1.64 +-  7%     +22.1%       2.00 +- 13%  sched_debug.cfs_rq:/.runnable_load_avg.stddev
     74653 +- 59%     -84.4%      11676        sched_debug.cfs_rq:/.runnable_weight.max
   -119169          -491.3%     466350 +- 27%  sched_debug.cfs_rq:/.spread0.avg
    517161 +- 30%    +145.8%    1271292 +- 23%  sched_debug.cfs_rq:/.spread0.max
    624.79 +-  5%     -14.2%     535.76 +-  7%  sched_debug.cfs_rq:/.util_est_enqueued.avg
    247.91 +- 32%     -99.8%       0.48 +-  8%  sched_debug.cfs_rq:/.util_est_enqueued.min
    179704 +-  3%     +30.4%     234297 +- 16%  sched_debug.cpu.avg_idle.stddev
      1.56 +-  9%     +24.4%       1.94 +- 14%  sched_debug.cpu.cpu_load[0].stddev
      1.50 +-  6%     +27.7%       1.91 +- 14%  sched_debug.cpu.cpu_load[1].stddev
      1.45 +-  3%     +30.8%       1.90 +- 14%  sched_debug.cpu.cpu_load[2].stddev
      1.43 +-  3%     +36.1%       1.95 +- 11%  sched_debug.cpu.cpu_load[3].stddev
      1.55 +-  7%     +43.5%       2.22 +-  7%  sched_debug.cpu.cpu_load[4].stddev
     10004 +-  3%     -11.6%       8839 +-  3%  sched_debug.cpu.curr->pid.avg
      1146 +- 26%     +52.2%       1745 +-  7%  sched_debug.cpu.curr->pid.min
      3162 +-  6%     +25.4%       3966 +- 11%  sched_debug.cpu.curr->pid.stddev
    403738 +-  3%     -11.7%     356696 +-  7%  sched_debug.cpu.nr_switches.max
      0.08 +- 21%     +78.2%       0.14 +- 14%  sched_debug.cpu.nr_uninterruptible.avg
    404435 +-  3%     -11.8%     356732 +-  7%  sched_debug.cpu.sched_count.max
      4.17            -0.3        3.87        perf-profile.calltrace.cycles-pp.kmem_cache_alloc.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      2.40            -0.2        2.17        perf-profile.calltrace.cycles-pp.vma_compute_subtree_gap.__vma_link_rb.vma_link.do_brk_flags.__x64_sys_brk
      7.58            -0.2        7.36        perf-profile.calltrace.cycles-pp.perf_event_mmap.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
     15.00            -0.2       14.81        perf-profile.calltrace.cycles-pp.syscall_return_via_sysret.brk
      7.83            -0.2        7.66        perf-profile.calltrace.cycles-pp.unmap_vmas.unmap_region.do_munmap.__x64_sys_brk.do_syscall_64
     28.66            -0.1       28.51        perf-profile.calltrace.cycles-pp.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe.brk
      2.15            -0.1        2.03        perf-profile.calltrace.cycles-pp.vma_compute_subtree_gap.do_munmap.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      1.07            -0.1        0.99        perf-profile.calltrace.cycles-pp.memcpy_erms.strlcpy.perf_event_mmap.do_brk_flags.__x64_sys_brk
      1.03            -0.1        0.95        perf-profile.calltrace.cycles-pp.kmem_cache_free.remove_vma.do_munmap.__x64_sys_brk.do_syscall_64
      7.33            -0.1        7.25        perf-profile.calltrace.cycles-pp.unmap_page_range.unmap_vmas.unmap_region.do_munmap.__x64_sys_brk
      0.76            -0.1        0.69        perf-profile.calltrace.cycles-pp.__vm_enough_memory.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
     11.85            -0.1       11.77        perf-profile.calltrace.cycles-pp.unmap_region.do_munmap.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      1.64            -0.1        1.57        perf-profile.calltrace.cycles-pp.strlcpy.perf_event_mmap.do_brk_flags.__x64_sys_brk.do_syscall_64
      1.06            -0.1        0.99        perf-profile.calltrace.cycles-pp.__indirect_thunk_start.brk
      0.73            -0.1        0.67        perf-profile.calltrace.cycles-pp.sync_mm_rss.unmap_page_range.unmap_vmas.unmap_region.do_munmap
      4.59            -0.1        4.52        perf-profile.calltrace.cycles-pp.security_vm_enough_memory_mm.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      2.82            -0.1        2.76        perf-profile.calltrace.cycles-pp.selinux_vm_enough_memory.security_vm_enough_memory_mm.do_brk_flags.__x64_sys_brk.do_syscall_64
      2.89            -0.1        2.84        perf-profile.calltrace.cycles-pp.down_write_killable.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe.brk
      3.37            -0.1        3.32        perf-profile.calltrace.cycles-pp.get_unmapped_area.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      1.99            -0.0        1.94        perf-profile.calltrace.cycles-pp.cred_has_capability.selinux_vm_enough_memory.security_vm_enough_memory_mm.do_brk_flags.__x64_sys_brk
      2.32            -0.0        2.27        perf-profile.calltrace.cycles-pp.perf_iterate_sb.perf_event_mmap.do_brk_flags.__x64_sys_brk.do_syscall_64
      1.88            -0.0        1.84        perf-profile.calltrace.cycles-pp.security_mmap_addr.get_unmapped_area.do_brk_flags.__x64_sys_brk.do_syscall_64
      0.77            -0.0        0.73        perf-profile.calltrace.cycles-pp._raw_spin_lock.unmap_page_range.unmap_vmas.unmap_region.do_munmap
      1.62            -0.0        1.59        perf-profile.calltrace.cycles-pp.memset_erms.kmem_cache_alloc.do_brk_flags.__x64_sys_brk.do_syscall_64
      0.81            -0.0        0.79        perf-profile.calltrace.cycles-pp.___might_sleep.down_write_killable.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.66            -0.0        0.64        perf-profile.calltrace.cycles-pp.arch_get_unmapped_area_topdown.brk
      0.72            +0.0        0.74        perf-profile.calltrace.cycles-pp.do_munmap.brk
      0.90            +0.0        0.93        perf-profile.calltrace.cycles-pp.___might_sleep.unmap_page_range.unmap_vmas.unmap_region.do_munmap
      4.40            +0.1        4.47        perf-profile.calltrace.cycles-pp.find_vma.do_munmap.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      1.96            +0.1        2.09        perf-profile.calltrace.cycles-pp.vmacache_find.find_vma.do_munmap.__x64_sys_brk.do_syscall_64
      0.52 +-  2%      +0.2        0.68        perf-profile.calltrace.cycles-pp.__vma_link_rb.brk
      0.35 +- 70%      +0.2        0.54 +-  2%  perf-profile.calltrace.cycles-pp.find_vma.brk
      2.20            +0.3        2.50        perf-profile.calltrace.cycles-pp.remove_vma.do_munmap.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
     64.62            +0.3       64.94        perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.brk
     60.53            +0.4       60.92        perf-profile.calltrace.cycles-pp.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe.brk
     63.20            +0.4       63.60        perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.brk
      3.73            +0.5        4.26        perf-profile.calltrace.cycles-pp.vma_link.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00            +0.6        0.56        perf-profile.calltrace.cycles-pp.free_pgtables.unmap_region.do_munmap.__x64_sys_brk.do_syscall_64
     24.54            +0.6       25.14        perf-profile.calltrace.cycles-pp.do_munmap.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe.brk
      0.00            +0.6        0.64        perf-profile.calltrace.cycles-pp.put_vma.remove_vma.do_munmap.__x64_sys_brk.do_syscall_64
      0.71            +0.6        1.36        perf-profile.calltrace.cycles-pp.__vma_rb_erase.do_munmap.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00            +0.7        0.70        perf-profile.calltrace.cycles-pp._raw_write_lock.__vma_rb_erase.do_munmap.__x64_sys_brk.do_syscall_64
      3.10            +0.7        3.82        perf-profile.calltrace.cycles-pp.__vma_link_rb.vma_link.do_brk_flags.__x64_sys_brk.do_syscall_64
      0.00            +0.8        0.76        perf-profile.calltrace.cycles-pp._raw_write_lock.__vma_link_rb.vma_link.do_brk_flags.__x64_sys_brk
      0.00            +0.8        0.85        perf-profile.calltrace.cycles-pp.__vma_merge.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      5.09            -0.5        4.62        perf-profile.children.cycles-pp.vma_compute_subtree_gap
      4.54            -0.3        4.21        perf-profile.children.cycles-pp.kmem_cache_alloc
      8.11            -0.2        7.89        perf-profile.children.cycles-pp.perf_event_mmap
      8.05            -0.2        7.85        perf-profile.children.cycles-pp.unmap_vmas
     15.01            -0.2       14.81        perf-profile.children.cycles-pp.syscall_return_via_sysret
     29.20            -0.1       29.06        perf-profile.children.cycles-pp.do_brk_flags
      1.11            -0.1        1.00        perf-profile.children.cycles-pp.kmem_cache_free
     12.28            -0.1       12.17        perf-profile.children.cycles-pp.unmap_region
      7.83            -0.1        7.74        perf-profile.children.cycles-pp.unmap_page_range
      0.87 +-  3%      -0.1        0.79        perf-profile.children.cycles-pp.__vm_enough_memory
      1.29            -0.1        1.22        perf-profile.children.cycles-pp.__indirect_thunk_start
      1.81            -0.1        1.74        perf-profile.children.cycles-pp.strlcpy
      4.65            -0.1        4.58        perf-profile.children.cycles-pp.security_vm_enough_memory_mm
      3.08            -0.1        3.02        perf-profile.children.cycles-pp.down_write_killable
      2.88            -0.1        2.82        perf-profile.children.cycles-pp.selinux_vm_enough_memory
      0.73            -0.1        0.67        perf-profile.children.cycles-pp.sync_mm_rss
      3.65            -0.1        3.59        perf-profile.children.cycles-pp.get_unmapped_area
      2.26            -0.1        2.20        perf-profile.children.cycles-pp.cred_has_capability
      1.12            -0.1        1.07        perf-profile.children.cycles-pp.memcpy_erms
      0.39            -0.0        0.35        perf-profile.children.cycles-pp.__rb_insert_augmented
      2.52            -0.0        2.48        perf-profile.children.cycles-pp.perf_iterate_sb
      2.13            -0.0        2.09        perf-profile.children.cycles-pp.security_mmap_addr
      0.55 +-  2%      -0.0        0.52        perf-profile.children.cycles-pp.unmap_single_vma
      1.62            -0.0        1.59        perf-profile.children.cycles-pp.memset_erms
      0.13 +-  3%      -0.0        0.11 +-  4%  perf-profile.children.cycles-pp.__vma_link_file
      0.80            -0.0        0.77        perf-profile.children.cycles-pp._raw_spin_lock
      0.43            -0.0        0.41        perf-profile.children.cycles-pp.strlen
      0.07 +-  6%      -0.0        0.06 +-  8%  perf-profile.children.cycles-pp.should_failslab
      0.43            -0.0        0.42        perf-profile.children.cycles-pp.may_expand_vm
      0.15            +0.0        0.16        perf-profile.children.cycles-pp.__vma_link_list
      0.45            +0.0        0.47        perf-profile.children.cycles-pp.rcu_all_qs
      0.81            +0.1        0.89        perf-profile.children.cycles-pp.free_pgtables
      6.35            +0.1        6.49        perf-profile.children.cycles-pp.find_vma
      2.28            +0.2        2.45        perf-profile.children.cycles-pp.vmacache_find
     64.66            +0.3       64.98        perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
      2.42            +0.3        2.76        perf-profile.children.cycles-pp.remove_vma
     61.77            +0.4       62.13        perf-profile.children.cycles-pp.__x64_sys_brk
     63.40            +0.4       63.79        perf-profile.children.cycles-pp.do_syscall_64
      1.27            +0.4        1.72        perf-profile.children.cycles-pp.__vma_rb_erase
      4.02            +0.5        4.53        perf-profile.children.cycles-pp.vma_link
     25.26            +0.6       25.89        perf-profile.children.cycles-pp.do_munmap
      0.00            +0.7        0.70        perf-profile.children.cycles-pp.put_vma
      3.80            +0.7        4.53        perf-profile.children.cycles-pp.__vma_link_rb
      0.00            +1.2        1.24        perf-profile.children.cycles-pp.__vma_merge
      0.00            +1.5        1.51        perf-profile.children.cycles-pp._raw_write_lock
      5.07            -0.5        4.60        perf-profile.self.cycles-pp.vma_compute_subtree_gap
      0.59            -0.2        0.38        perf-profile.self.cycles-pp.remove_vma
     15.01            -0.2       14.81        perf-profile.self.cycles-pp.syscall_return_via_sysret
      3.15            -0.2        2.96        perf-profile.self.cycles-pp.do_munmap
      0.98            -0.1        0.87        perf-profile.self.cycles-pp.__vma_rb_erase
      1.10            -0.1        0.99        perf-profile.self.cycles-pp.kmem_cache_free
      0.68            -0.1        0.58        perf-profile.self.cycles-pp.__vm_enough_memory
      0.42            -0.1        0.33        perf-profile.self.cycles-pp.unmap_vmas
      3.62            -0.1        3.53        perf-profile.self.cycles-pp.perf_event_mmap
      1.41            -0.1        1.34        perf-profile.self.cycles-pp.entry_SYSCALL_64_after_hwframe
      1.29            -0.1        1.22        perf-profile.self.cycles-pp.__indirect_thunk_start
      0.73            -0.1        0.66        perf-profile.self.cycles-pp.sync_mm_rss
      2.96            -0.1        2.90        perf-profile.self.cycles-pp.__x64_sys_brk
      3.24            -0.1        3.19        perf-profile.self.cycles-pp.brk
      1.11            -0.0        1.07        perf-profile.self.cycles-pp.memcpy_erms
      0.53 +-  3%      -0.0        0.49 +-  2%  perf-profile.self.cycles-pp.vma_link
      0.73            -0.0        0.69        perf-profile.self.cycles-pp.unmap_region
      1.66            -0.0        1.61        perf-profile.self.cycles-pp.down_write_killable
      0.39            -0.0        0.35        perf-profile.self.cycles-pp.__rb_insert_augmented
      1.74            -0.0        1.71        perf-profile.self.cycles-pp.kmem_cache_alloc
      0.55 +-  2%      -0.0        0.52        perf-profile.self.cycles-pp.unmap_single_vma
      1.61            -0.0        1.59        perf-profile.self.cycles-pp.memset_erms
      0.80            -0.0        0.77        perf-profile.self.cycles-pp._raw_spin_lock
      0.13            -0.0        0.11 +-  4%  perf-profile.self.cycles-pp.__vma_link_file
      0.43            -0.0        0.41        perf-profile.self.cycles-pp.strlen
      0.07 +-  6%      -0.0        0.06 +-  8%  perf-profile.self.cycles-pp.should_failslab
      0.81            -0.0        0.79        perf-profile.self.cycles-pp.tlb_finish_mmu
      0.15            +0.0        0.16        perf-profile.self.cycles-pp.__vma_link_list
      0.45            +0.0        0.47        perf-profile.self.cycles-pp.rcu_all_qs
      0.71            +0.0        0.72        perf-profile.self.cycles-pp.strlcpy
      0.51            +0.1        0.56        perf-profile.self.cycles-pp.free_pgtables
      1.41            +0.1        1.48        perf-profile.self.cycles-pp.__vma_link_rb
      2.27            +0.2        2.44        perf-profile.self.cycles-pp.vmacache_find
      0.00            +0.7        0.69        perf-profile.self.cycles-pp.put_vma
      0.00            +1.2        1.23        perf-profile.self.cycles-pp.__vma_merge
      0.00            +1.5        1.50        perf-profile.self.cycles-pp._raw_write_lock

=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/nr_task/thp_enabled/test/cpufreq_governor:
  lkp-skl-4sp1/will-it-scale/debian-x86_64-2018-04-03.cgz/x86_64-rhel-7.2/gcc-7/100%/always/brk1/performance

commit: 
  ba98a1cdad71d259a194461b3a61471b49b14df1
  a7a8993bfe3ccb54ad468b9f1799649e4ad1ff12

ba98a1cdad71d259 a7a8993bfe3ccb54ad468b9f17 
---------------- -------------------------- 
       fail:runs  %reproduction    fail:runs
           |             |             |    
           :3           33%           1:3     dmesg.WARNING:stack_going_in_the_wrong_direction?ip=schedule_tail/0x
           :3           33%           1:3     kmsg.DHCP/BOOTP:Reply_not_for_us_on_eth#,op[#]xid[#]
         %stddev     %change         %stddev
             \          |                \  
    998475            -2.2%     976893        will-it-scale.per_process_ops
    625.87            -2.3%     611.42        will-it-scale.time.elapsed_time
    625.87            -2.3%     611.42        will-it-scale.time.elapsed_time.max
      8158            -1.9%       8000        will-it-scale.time.maximum_resident_set_size
     18.42 +-  2%     -11.9%      16.24        will-it-scale.time.user_time
  34349225 +- 13%     -14.5%   29371024 +- 17%  will-it-scale.time.voluntary_context_switches
 1.919e+08            -2.2%  1.877e+08        will-it-scale.workload
      1639 +- 23%     -18.4%       1337 +- 30%  meminfo.Mlocked
     17748 +- 82%    +103.1%      36051        numa-numastat.node3.other_node
  33410486 +- 14%     -14.8%   28449258 +- 18%  cpuidle.C1.usage
    698749 +- 15%     -18.0%     573307 +- 20%  cpuidle.POLL.usage
   3013702 +- 14%     -15.1%    2559405 +- 17%  softirqs.SCHED
  54361293 +-  2%     -19.0%   44044816 +-  2%  softirqs.TIMER
  33408303 +- 14%     -14.9%   28447123 +- 18%  turbostat.C1
      0.34 +- 16%     -52.0%       0.16 +- 15%  turbostat.Pkg%pc2
      1310 +- 74%    +412.1%       6710 +- 58%  irq_exception_noise.__do_page_fault.samples
      3209 +- 74%    +281.9%      12258 +- 53%  irq_exception_noise.__do_page_fault.sum
    600.67 +-132%     -96.0%      24.00 +- 23%  irq_exception_noise.irq_nr
     99557 +-  7%     -24.0%      75627 +-  7%  irq_exception_noise.softirq_nr
     41424 +-  9%     -24.6%      31253 +-  6%  irq_exception_noise.softirq_time
    625.87            -2.3%     611.42        time.elapsed_time
    625.87            -2.3%     611.42        time.elapsed_time.max
      8158            -1.9%       8000        time.maximum_resident_set_size
     18.42 +-  2%     -11.9%      16.24        time.user_time
  34349225 +- 13%     -14.5%   29371024 +- 17%  time.voluntary_context_switches
    988.00 +-  8%     +14.5%       1131 +-  2%  slabinfo.Acpi-ParseExt.active_objs
    988.00 +-  8%     +14.5%       1131 +-  2%  slabinfo.Acpi-ParseExt.num_objs
      2384 +-  3%     +21.1%       2888 +- 11%  slabinfo.pool_workqueue.active_objs
      2474 +-  2%     +20.4%       2979 +- 11%  slabinfo.pool_workqueue.num_objs
    490.33 +- 10%     -19.2%     396.00 +- 11%  slabinfo.secpath_cache.active_objs
    490.33 +- 10%     -19.2%     396.00 +- 11%  slabinfo.secpath_cache.num_objs
      1123 +-  7%     +14.2%       1282 +-  3%  slabinfo.skbuff_fclone_cache.active_objs
      1123 +-  7%     +14.2%       1282 +-  3%  slabinfo.skbuff_fclone_cache.num_objs
      1.09            -0.0        1.07        perf-stat.branch-miss-rate%
 2.691e+11            -2.4%  2.628e+11        perf-stat.branch-misses
  71981351 +- 12%     -13.8%   62013509 +- 16%  perf-stat.context-switches
 1.697e+13            +1.1%  1.715e+13        perf-stat.dTLB-stores
      2.36 +- 29%      +4.4        6.76 +- 11%  perf-stat.iTLB-load-miss-rate%
  5.21e+08 +- 28%    +194.8%  1.536e+09 +- 10%  perf-stat.iTLB-load-misses
    239983 +- 24%     -68.4%      75819 +- 11%  perf-stat.instructions-per-iTLB-miss
   3295653 +-  2%      -6.3%    3088753 +-  3%  perf-stat.node-stores
    606239            +1.1%     612799        perf-stat.path-length
      3755 +- 28%     -37.5%       2346 +- 52%  sched_debug.cfs_rq:/.exec_clock.stddev
     10.45 +-  4%     +24.3%      12.98 +- 18%  sched_debug.cfs_rq:/.load_avg.stddev
      6243 +- 46%     -38.6%       3831 +- 78%  sched_debug.cpu.load.stddev
    867.80 +-  7%     +25.3%       1087 +-  6%  sched_debug.cpu.nr_load_updates.stddev
    395898 +-  3%     -11.1%     352071 +-  7%  sched_debug.cpu.nr_switches.max
    -13.33           -21.1%     -10.52        sched_debug.cpu.nr_uninterruptible.min
    395674 +-  3%     -11.1%     351762 +-  7%  sched_debug.cpu.sched_count.max
     33152 +-  4%     -12.8%      28899        sched_debug.cpu.ttwu_count.min
      0.03 +- 20%     +77.7%       0.05 +- 15%  sched_debug.rt_rq:/.rt_time.max
     89523            +1.8%      91099        proc-vmstat.nr_active_anon
    409.67 +- 23%     -18.4%     334.33 +- 30%  proc-vmstat.nr_mlock
     89530            +1.8%      91117        proc-vmstat.nr_zone_active_anon
   2337130            -2.2%    2286775        proc-vmstat.numa_hit
   2229090            -2.3%    2178626        proc-vmstat.numa_local
      8460 +- 39%     -75.5%       2076 +- 53%  proc-vmstat.numa_pages_migrated
     28643 +- 55%     -83.5%       4727 +- 58%  proc-vmstat.numa_pte_updates
   2695806            -1.8%    2646639        proc-vmstat.pgfault
   2330191            -2.1%    2281197        proc-vmstat.pgfree
      8460 +- 39%     -75.5%       2076 +- 53%  proc-vmstat.pgmigrate_success
    237651 +-  2%     +31.3%     312092 +- 16%  numa-meminfo.node0.FilePages
      8059 +-  2%     +10.7%       8925 +-  7%  numa-meminfo.node0.KernelStack
      6830 +- 25%     +48.8%      10164 +- 35%  numa-meminfo.node0.Mapped
      1612 +- 21%     +70.0%       2740 +- 19%  numa-meminfo.node0.PageTables
     10772 +- 65%    +679.4%      83962 +- 59%  numa-meminfo.node0.Shmem
    163195 +- 15%     -36.9%     103036 +- 32%  numa-meminfo.node1.Active
    163195 +- 15%     -36.9%     103036 +- 32%  numa-meminfo.node1.Active(anon)
      1730 +-  4%     +33.9%       2317 +- 14%  numa-meminfo.node1.PageTables
     55778 +- 19%     +32.5%      73910 +-  8%  numa-meminfo.node1.SUnreclaim
      2671 +- 16%     -45.0%       1469 +- 15%  numa-meminfo.node2.PageTables
     61537 +- 13%     -17.7%      50647 +-  3%  numa-meminfo.node2.SUnreclaim
     48644 +- 94%    +149.8%     121499 +- 11%  numa-meminfo.node3.Active
     48440 +- 94%    +150.4%     121295 +- 11%  numa-meminfo.node3.Active(anon)
     11832 +- 79%     -91.5%       1008 +- 67%  numa-meminfo.node3.Inactive
     11597 +- 82%     -93.3%     772.00 +- 82%  numa-meminfo.node3.Inactive(anon)
     10389 +- 32%     -43.0%       5921 +-  6%  numa-meminfo.node3.Mapped
     33704 +- 24%     -44.2%      18792 +- 15%  numa-meminfo.node3.SReclaimable
    104733 +- 14%     -25.3%      78275 +-  8%  numa-meminfo.node3.Slab
    139329 +-133%     -99.8%     241.67 +- 79%  latency_stats.avg.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_do_create.nfs3_proc_create.nfs_create.path_openat.do_filp_open.do_sys_open.do_syscall_64
      5403 +-139%     -97.5%     137.67 +- 71%  latency_stats.avg.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_lookup.nfs_lookup.path_openat.do_filp_open.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
    165968 +-101%     -61.9%      63304 +- 58%  latency_stats.avg.max
     83.00        +12810.4%      10715 +-140%  latency_stats.avg.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_access.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_lookupat.filename_lookup
    102.67 +-  6%  +18845.5%      19450 +-140%  latency_stats.avg.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_getattr.__nfs_revalidate_inode.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_lookupat
    136.33 +- 16%  +25043.5%      34279 +-141%  latency_stats.avg.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_lookup.nfs_lookup.__lookup_slow.lookup_slow.walk_component.path_lookupat.filename_lookup
     18497 +-141%    -100.0%       0.00        latency_stats.max.call_rwsem_down_write_failed_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe
    140500 +-131%     -99.8%     247.00 +- 78%  latency_stats.max.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_do_create.nfs3_proc_create.nfs_create.path_openat.do_filp_open.do_sys_open.do_syscall_64
      5403 +-139%     -97.5%     137.67 +- 71%  latency_stats.max.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_lookup.nfs_lookup.path_openat.do_filp_open.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
     87.33 +-  5%  +23963.0%      21015 +-140%  latency_stats.max.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_access.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_lookupat.filename_lookup
    136.33 +- 16%  +25043.5%      34279 +-141%  latency_stats.max.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_lookup.nfs_lookup.__lookup_slow.lookup_slow.walk_component.path_lookupat.filename_lookup
    149.33 +- 14%  +25485.9%      38208 +-141%  latency_stats.max.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_getattr.__nfs_revalidate_inode.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_lookupat
     18761 +-141%    -100.0%       0.00        latency_stats.sum.call_rwsem_down_write_failed_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe
     23363 +-114%    -100.0%       0.00        latency_stats.sum.call_rwsem_down_read_failed.__do_page_fault.do_page_fault.page_fault.__get_user_8.exit_robust_list.mm_release.do_exit.do_group_exit.get_signal.do_signal.exit_to_usermode_loop
    144810 +-125%     -99.8%     326.67 +- 70%  latency_stats.sum.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_do_create.nfs3_proc_create.nfs_create.path_openat.do_filp_open.do_sys_open.do_syscall_64
      5403 +-139%     -97.5%     137.67 +- 71%  latency_stats.sum.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_lookup.nfs_lookup.path_openat.do_filp_open.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
     59698 +- 98%     -78.0%      13110 +-141%  latency_stats.sum.call_rwsem_down_read_failed.do_exit.do_group_exit.get_signal.do_signal.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
    166.33        +12768.5%      21404 +-140%  latency_stats.sum.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_access.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_lookupat.filename_lookup
    825.00 +-  6%  +18761.7%     155609 +-140%  latency_stats.sum.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_getattr.__nfs_revalidate_inode.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_lookupat
    136.33 +- 16%  +25043.5%      34279 +-141%  latency_stats.sum.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_lookup.nfs_lookup.__lookup_slow.lookup_slow.walk_component.path_lookupat.filename_lookup
     59412 +-  2%     +31.3%      78021 +- 16%  numa-vmstat.node0.nr_file_pages
      8059 +-  2%     +10.7%       8923 +-  7%  numa-vmstat.node0.nr_kernel_stack
      1701 +- 25%     +49.1%       2536 +- 35%  numa-vmstat.node0.nr_mapped
    402.33 +- 21%     +70.0%     684.00 +- 19%  numa-vmstat.node0.nr_page_table_pages
      2692 +- 65%    +679.5%      20988 +- 59%  numa-vmstat.node0.nr_shmem
    622587 +- 36%     +37.7%     857545 +- 13%  numa-vmstat.node0.numa_local
     40797 +- 15%     -36.9%      25757 +- 32%  numa-vmstat.node1.nr_active_anon
    432.00 +-  4%     +33.9%     578.33 +- 14%  numa-vmstat.node1.nr_page_table_pages
     13944 +- 19%     +32.5%      18477 +-  8%  numa-vmstat.node1.nr_slab_unreclaimable
     40797 +- 15%     -36.9%      25757 +- 32%  numa-vmstat.node1.nr_zone_active_anon
    625073 +- 26%     +29.4%     808657 +- 18%  numa-vmstat.node1.numa_hit
    503969 +- 34%     +39.2%     701446 +- 23%  numa-vmstat.node1.numa_local
    137.33 +- 40%     -49.0%      70.00 +- 29%  numa-vmstat.node2.nr_mlock
    667.67 +- 17%     -45.1%     366.33 +- 15%  numa-vmstat.node2.nr_page_table_pages
     15384 +- 13%     -17.7%      12662 +-  3%  numa-vmstat.node2.nr_slab_unreclaimable
     12114 +- 94%    +150.3%      30326 +- 11%  numa-vmstat.node3.nr_active_anon
      2887 +- 83%     -93.4%     190.00 +- 82%  numa-vmstat.node3.nr_inactive_anon
      2632 +- 30%     -39.2%       1600 +-  5%  numa-vmstat.node3.nr_mapped
    101.00           -30.0%      70.67 +- 29%  numa-vmstat.node3.nr_mlock
      8425 +- 24%     -44.2%       4697 +- 15%  numa-vmstat.node3.nr_slab_reclaimable
     12122 +- 94%    +150.3%      30346 +- 11%  numa-vmstat.node3.nr_zone_active_anon
      2887 +- 83%     -93.4%     190.00 +- 82%  numa-vmstat.node3.nr_zone_inactive_anon
    106945 +- 13%     +17.4%     125554        numa-vmstat.node3.numa_other
      4.17            -0.3        3.82        perf-profile.calltrace.cycles-pp.kmem_cache_alloc.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
     15.02            -0.3       14.77        perf-profile.calltrace.cycles-pp.syscall_return_via_sysret.brk
      2.42            -0.2        2.18        perf-profile.calltrace.cycles-pp.vma_compute_subtree_gap.__vma_link_rb.vma_link.do_brk_flags.__x64_sys_brk
      7.60            -0.2        7.39        perf-profile.calltrace.cycles-pp.perf_event_mmap.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      7.79            -0.2        7.63        perf-profile.calltrace.cycles-pp.unmap_vmas.unmap_region.do_munmap.__x64_sys_brk.do_syscall_64
      0.82 +-  9%      -0.1        0.68        perf-profile.calltrace.cycles-pp.__vm_enough_memory.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      2.13            -0.1        2.00        perf-profile.calltrace.cycles-pp.vma_compute_subtree_gap.do_munmap.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      1.05            -0.1        0.95        perf-profile.calltrace.cycles-pp.kmem_cache_free.remove_vma.do_munmap.__x64_sys_brk.do_syscall_64
      7.31            -0.1        7.21        perf-profile.calltrace.cycles-pp.unmap_page_range.unmap_vmas.unmap_region.do_munmap.__x64_sys_brk
      0.74            -0.1        0.67        perf-profile.calltrace.cycles-pp.sync_mm_rss.unmap_page_range.unmap_vmas.unmap_region.do_munmap
      1.06            -0.1        1.00        perf-profile.calltrace.cycles-pp.memcpy_erms.strlcpy.perf_event_mmap.do_brk_flags.__x64_sys_brk
      3.38            -0.1        3.33        perf-profile.calltrace.cycles-pp.get_unmapped_area.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      1.05            -0.0        1.00 +-  2%  perf-profile.calltrace.cycles-pp.__indirect_thunk_start.brk
      2.34            -0.0        2.29        perf-profile.calltrace.cycles-pp.perf_iterate_sb.perf_event_mmap.do_brk_flags.__x64_sys_brk.do_syscall_64
      1.64            -0.0        1.59        perf-profile.calltrace.cycles-pp.strlcpy.perf_event_mmap.do_brk_flags.__x64_sys_brk.do_syscall_64
      1.89            -0.0        1.86        perf-profile.calltrace.cycles-pp.security_mmap_addr.get_unmapped_area.do_brk_flags.__x64_sys_brk.do_syscall_64
      0.76            -0.0        0.73        perf-profile.calltrace.cycles-pp._raw_spin_lock.unmap_page_range.unmap_vmas.unmap_region.do_munmap
      0.57 +-  2%      -0.0        0.55        perf-profile.calltrace.cycles-pp.selinux_mmap_addr.security_mmap_addr.get_unmapped_area.do_brk_flags.__x64_sys_brk
      0.54 +-  2%      +0.0        0.56        perf-profile.calltrace.cycles-pp.do_brk_flags.brk
      0.72            +0.0        0.76 +-  2%  perf-profile.calltrace.cycles-pp.do_munmap.brk
      4.38            +0.1        4.43        perf-profile.calltrace.cycles-pp.find_vma.do_munmap.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      1.96            +0.1        2.04        perf-profile.calltrace.cycles-pp.vmacache_find.find_vma.do_munmap.__x64_sys_brk.do_syscall_64
      0.53            +0.2        0.68        perf-profile.calltrace.cycles-pp.__vma_link_rb.brk
      2.21            +0.3        2.51        perf-profile.calltrace.cycles-pp.remove_vma.do_munmap.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
     64.44            +0.5       64.90        perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.brk
     63.04            +0.5       63.54        perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.brk
     60.37            +0.5       60.88        perf-profile.calltrace.cycles-pp.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe.brk
      3.75            +0.5        4.29        perf-profile.calltrace.cycles-pp.vma_link.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00            +0.6        0.57        perf-profile.calltrace.cycles-pp.free_pgtables.unmap_region.do_munmap.__x64_sys_brk.do_syscall_64
      0.00            +0.6        0.64        perf-profile.calltrace.cycles-pp.put_vma.remove_vma.do_munmap.__x64_sys_brk.do_syscall_64
      0.72            +0.7        1.37        perf-profile.calltrace.cycles-pp.__vma_rb_erase.do_munmap.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
     24.42            +0.7       25.08        perf-profile.calltrace.cycles-pp.do_munmap.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe.brk
      0.00            +0.7        0.71        perf-profile.calltrace.cycles-pp._raw_write_lock.__vma_rb_erase.do_munmap.__x64_sys_brk.do_syscall_64
      3.12            +0.7        3.84        perf-profile.calltrace.cycles-pp.__vma_link_rb.vma_link.do_brk_flags.__x64_sys_brk.do_syscall_64
      0.00            +0.8        0.77        perf-profile.calltrace.cycles-pp._raw_write_lock.__vma_link_rb.vma_link.do_brk_flags.__x64_sys_brk
      0.00            +0.9        0.85        perf-profile.calltrace.cycles-pp.__vma_merge.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      5.10            -0.5        4.60        perf-profile.children.cycles-pp.vma_compute_subtree_gap
      4.53            -0.3        4.18        perf-profile.children.cycles-pp.kmem_cache_alloc
     15.03            -0.3       14.77        perf-profile.children.cycles-pp.syscall_return_via_sysret
      8.13            -0.2        7.92        perf-profile.children.cycles-pp.perf_event_mmap
      8.01            -0.2        7.81        perf-profile.children.cycles-pp.unmap_vmas
      0.97 +- 14%      -0.2        0.78        perf-profile.children.cycles-pp.__vm_enough_memory
      1.13            -0.1        1.00        perf-profile.children.cycles-pp.kmem_cache_free
      7.82            -0.1        7.70        perf-profile.children.cycles-pp.unmap_page_range
     12.23            -0.1       12.13        perf-profile.children.cycles-pp.unmap_region
      0.74            -0.1        0.67        perf-profile.children.cycles-pp.sync_mm_rss
      3.06            -0.1        3.00        perf-profile.children.cycles-pp.down_write_killable
      0.40 +-  2%      -0.1        0.34        perf-profile.children.cycles-pp.__rb_insert_augmented
      1.29            -0.1        1.23        perf-profile.children.cycles-pp.__indirect_thunk_start
      2.54            -0.1        2.49        perf-profile.children.cycles-pp.perf_iterate_sb
      3.66            -0.0        3.61        perf-profile.children.cycles-pp.get_unmapped_area
      1.80            -0.0        1.75        perf-profile.children.cycles-pp.strlcpy
      0.53 +-  2%      -0.0        0.49 +-  2%  perf-profile.children.cycles-pp.cap_capable
      1.57            -0.0        1.53        perf-profile.children.cycles-pp.arch_get_unmapped_area_topdown
      1.11            -0.0        1.08        perf-profile.children.cycles-pp.memcpy_erms
      0.13            -0.0        0.10        perf-profile.children.cycles-pp.__vma_link_file
      0.55            -0.0        0.52        perf-profile.children.cycles-pp.unmap_single_vma
      1.47            -0.0        1.44        perf-profile.children.cycles-pp.cap_vm_enough_memory
      2.14            -0.0        2.12        perf-profile.children.cycles-pp.security_mmap_addr
      0.32            -0.0        0.30        perf-profile.children.cycles-pp.userfaultfd_unmap_complete
      1.25            -0.0        1.23        perf-profile.children.cycles-pp.up_write
      0.50            -0.0        0.49        perf-profile.children.cycles-pp.userfaultfd_unmap_prep
      0.27            -0.0        0.26        perf-profile.children.cycles-pp.tlb_flush_mmu_free
      1.14            -0.0        1.12        perf-profile.children.cycles-pp.__might_sleep
      0.07            -0.0        0.06        perf-profile.children.cycles-pp.should_failslab
      0.72            +0.0        0.74        perf-profile.children.cycles-pp._cond_resched
      0.45            +0.0        0.47        perf-profile.children.cycles-pp.rcu_all_qs
      0.15 +-  3%      +0.0        0.17 +-  4%  perf-profile.children.cycles-pp.__vma_link_list
      0.15 +-  5%      +0.0        0.18 +-  5%  perf-profile.children.cycles-pp.tick_sched_timer
      0.05 +-  8%      +0.1        0.12 +- 17%  perf-profile.children.cycles-pp.perf_mux_hrtimer_handler
      0.80            +0.1        0.89        perf-profile.children.cycles-pp.free_pgtables
      0.22 +-  7%      +0.1        0.31 +-  9%  perf-profile.children.cycles-pp.__hrtimer_run_queues
      0.00            +0.1        0.11 +- 15%  perf-profile.children.cycles-pp.clockevents_program_event
      6.34            +0.1        6.47        perf-profile.children.cycles-pp.find_vma
      2.27            +0.1        2.40        perf-profile.children.cycles-pp.vmacache_find
      0.40 +-  4%      +0.2        0.58 +-  5%  perf-profile.children.cycles-pp.apic_timer_interrupt
      0.40 +-  4%      +0.2        0.58 +-  5%  perf-profile.children.cycles-pp.smp_apic_timer_interrupt
      0.37 +-  4%      +0.2        0.54 +-  5%  perf-profile.children.cycles-pp.hrtimer_interrupt
      0.00            +0.2        0.19 +- 12%  perf-profile.children.cycles-pp.ktime_get
      2.42            +0.3        2.77        perf-profile.children.cycles-pp.remove_vma
     64.49            +0.5       64.94        perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
      1.27            +0.5        1.73        perf-profile.children.cycles-pp.__vma_rb_erase
     61.62            +0.5       62.10        perf-profile.children.cycles-pp.__x64_sys_brk
     63.24            +0.5       63.74        perf-profile.children.cycles-pp.do_syscall_64
      4.03            +0.5        4.56        perf-profile.children.cycles-pp.vma_link
      0.00            +0.7        0.69        perf-profile.children.cycles-pp.put_vma
     25.13            +0.7       25.84        perf-profile.children.cycles-pp.do_munmap
      3.83            +0.7        4.56        perf-profile.children.cycles-pp.__vma_link_rb
      0.00            +1.2        1.25        perf-profile.children.cycles-pp.__vma_merge
      0.00            +1.5        1.53        perf-profile.children.cycles-pp._raw_write_lock
      5.08            -0.5        4.58        perf-profile.self.cycles-pp.vma_compute_subtree_gap
     15.03            -0.3       14.77        perf-profile.self.cycles-pp.syscall_return_via_sysret
      0.59            -0.2        0.39        perf-profile.self.cycles-pp.remove_vma
      0.72 +-  7%      -0.1        0.58        perf-profile.self.cycles-pp.__vm_enough_memory
      1.12            -0.1        0.99        perf-profile.self.cycles-pp.kmem_cache_free
      3.11            -0.1        2.99        perf-profile.self.cycles-pp.do_munmap
      0.99            -0.1        0.88        perf-profile.self.cycles-pp.__vma_rb_erase
      3.63            -0.1        3.52        perf-profile.self.cycles-pp.perf_event_mmap
      3.26            -0.1        3.17        perf-profile.self.cycles-pp.brk
      0.41 +-  2%      -0.1        0.33        perf-profile.self.cycles-pp.unmap_vmas
      0.74            -0.1        0.67        perf-profile.self.cycles-pp.sync_mm_rss
      1.75            -0.1        1.68        perf-profile.self.cycles-pp.kmem_cache_alloc
      0.40 +-  2%      -0.1        0.34        perf-profile.self.cycles-pp.__rb_insert_augmented
      1.29 +-  2%      -0.1        1.23        perf-profile.self.cycles-pp.__indirect_thunk_start
      0.73            -0.0        0.68 +-  2%  perf-profile.self.cycles-pp.unmap_region
      0.53            -0.0        0.49        perf-profile.self.cycles-pp.vma_link
      1.40            -0.0        1.35        perf-profile.self.cycles-pp.entry_SYSCALL_64_after_hwframe
      5.22            -0.0        5.18        perf-profile.self.cycles-pp.unmap_page_range
      0.53 +-  2%      -0.0        0.49 +-  2%  perf-profile.self.cycles-pp.cap_capable
      1.11            -0.0        1.07        perf-profile.self.cycles-pp.memcpy_erms
      1.86            -0.0        1.82        perf-profile.self.cycles-pp.perf_iterate_sb
      1.30            -0.0        1.27        perf-profile.self.cycles-pp.arch_get_unmapped_area_topdown
      0.13            -0.0        0.10        perf-profile.self.cycles-pp.__vma_link_file
      0.55            -0.0        0.52        perf-profile.self.cycles-pp.unmap_single_vma
      0.74            -0.0        0.72        perf-profile.self.cycles-pp.selinux_mmap_addr
      0.32            -0.0        0.30        perf-profile.self.cycles-pp.userfaultfd_unmap_complete
      1.13            -0.0        1.12        perf-profile.self.cycles-pp.__might_sleep
      1.24            -0.0        1.23        perf-profile.self.cycles-pp.up_write
      0.50            -0.0        0.49        perf-profile.self.cycles-pp.userfaultfd_unmap_prep
      0.27            -0.0        0.26        perf-profile.self.cycles-pp.tlb_flush_mmu_free
      0.07            -0.0        0.06        perf-profile.self.cycles-pp.should_failslab
      0.45            +0.0        0.47        perf-profile.self.cycles-pp.rcu_all_qs
      0.71            +0.0        0.73        perf-profile.self.cycles-pp.strlcpy
      0.15 +-  3%      +0.0        0.17 +-  4%  perf-profile.self.cycles-pp.__vma_link_list
      0.51            +0.1        0.57        perf-profile.self.cycles-pp.free_pgtables
      1.40            +0.1        1.49        perf-profile.self.cycles-pp.__vma_link_rb
      2.27            +0.1        2.39        perf-profile.self.cycles-pp.vmacache_find
      0.00            +0.2        0.18 +- 12%  perf-profile.self.cycles-pp.ktime_get
      0.00            +0.7        0.69        perf-profile.self.cycles-pp.put_vma
      0.00            +1.2        1.24        perf-profile.self.cycles-pp.__vma_merge
      0.00            +1.5        1.52        perf-profile.self.cycles-pp._raw_write_lock

=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/nr_task/thp_enabled/test/cpufreq_governor:
  lkp-skl-4sp1/will-it-scale/debian-x86_64-2018-04-03.cgz/x86_64-rhel-7.2/gcc-7/100%/always/page_fault2/performance

commit: 
  ba98a1cdad71d259a194461b3a61471b49b14df1
  a7a8993bfe3ccb54ad468b9f1799649e4ad1ff12

ba98a1cdad71d259 a7a8993bfe3ccb54ad468b9f17 
---------------- -------------------------- 
       fail:runs  %reproduction    fail:runs
           |             |             |    
           :3           33%           1:3     dmesg.WARNING:at#for_ip_native_iret/0x
          1:3          -33%            :3     dmesg.WARNING:stack_going_in_the_wrong_direction?ip=__schedule/0x
           :3           33%           1:3     dmesg.WARNING:stack_going_in_the_wrong_direction?ip=__slab_free/0x
          1:3          -33%            :3     kmsg.DHCP/BOOTP:Reply_not_for_us_on_eth#,op[#]xid[#]
          3:3         -100%            :3     kmsg.pstore:crypto_comp_decompress_failed,ret=
          3:3         -100%            :3     kmsg.pstore:decompression_failed
          2:3            4%           2:3     perf-profile.calltrace.cycles-pp.sync_regs.error_entry
          5:3            7%           5:3     perf-profile.calltrace.cycles-pp.error_entry
          5:3            7%           5:3     perf-profile.children.cycles-pp.error_entry
          2:3            3%           2:3     perf-profile.self.cycles-pp.error_entry
         %stddev     %change         %stddev
             \          |                \  
      8281 +-  2%     -18.8%       6728        will-it-scale.per_thread_ops
     92778 +-  2%     +17.6%     109080        will-it-scale.time.involuntary_context_switches
  21954366 +-  3%      +4.1%   22857988 +-  2%  will-it-scale.time.maximum_resident_set_size
  4.81e+08 +-  2%     -18.9%  3.899e+08        will-it-scale.time.minor_page_faults
      5804           +12.2%       6512        will-it-scale.time.percent_of_cpu_this_job_got
     34918           +12.2%      39193        will-it-scale.time.system_time
   5638528 +-  2%     -15.3%    4778392        will-it-scale.time.voluntary_context_switches
  15846405            -2.0%   15531034        will-it-scale.workload
   2818137            +1.5%    2861500        interrupts.CAL:Function_call_interrupts
      3.33 +- 28%     -60.0%       1.33 +- 93%  irq_exception_noise.irq_time
      2866           +23.9%       3552 +-  2%  kthread_noise.total_time
   5589674 +- 14%     +31.4%    7344810 +-  6%  meminfo.DirectMap2M
     31169           -16.9%      25906        uptime.idle
     25242 +-  4%     -14.2%      21654 +-  6%  vmstat.system.cs
      7055           -11.6%       6237        boot-time.idle
     21.12           +19.3%      25.19 +-  9%  boot-time.kernel_boot
     20.03 +-  2%      -3.7       16.38        mpstat.cpu.idle%
      0.00 +-  8%      -0.0        0.00 +-  4%  mpstat.cpu.iowait%
   7284147 +-  2%     -16.4%    6092495        softirqs.RCU
   5350756 +-  2%     -10.9%    4769417 +-  4%  softirqs.SCHED
     42933 +- 21%     -28.2%      30807 +-  7%  numa-meminfo.node2.SReclaimable
     63219 +- 13%     -16.6%      52717 +-  6%  numa-meminfo.node2.SUnreclaim
    106153 +- 16%     -21.3%      83525 +-  5%  numa-meminfo.node2.Slab
    247154 +-  4%      -7.6%     228415        numa-meminfo.node3.Unevictable
     11904 +-  4%     +17.1%      13945 +-  8%  numa-vmstat.node0
      2239 +- 22%     -26.6%       1644 +-  2%  numa-vmstat.node2.nr_mapped
     10728 +- 21%     -28.2%       7701 +-  7%  numa-vmstat.node2.nr_slab_reclaimable
     15803 +- 13%     -16.6%      13179 +-  6%  numa-vmstat.node2.nr_slab_unreclaimable
     61788 +-  4%      -7.6%      57103        numa-vmstat.node3.nr_unevictable
     61788 +-  4%      -7.6%      57103        numa-vmstat.node3.nr_zone_unevictable
     92778 +-  2%     +17.6%     109080        time.involuntary_context_switches
  21954366 +-  3%      +4.1%   22857988 +-  2%  time.maximum_resident_set_size
  4.81e+08 +-  2%     -18.9%  3.899e+08        time.minor_page_faults
      5804           +12.2%       6512        time.percent_of_cpu_this_job_got
     34918           +12.2%      39193        time.system_time
   5638528 +-  2%     -15.3%    4778392        time.voluntary_context_switches
   3942289 +-  2%     -10.5%    3528902 +-  2%  cpuidle.C1.time
    242290           -14.2%     207992        cpuidle.C1.usage
  1.64e+09 +-  2%     -15.7%  1.381e+09        cpuidle.C1E.time
   4621281 +-  2%     -14.7%    3939757        cpuidle.C1E.usage
 2.115e+10 +-  2%     -18.5%  1.723e+10        cpuidle.C6.time
  24771099 +-  2%     -18.0%   20305766        cpuidle.C6.usage
   1210810 +-  4%     -17.6%     997270 +-  2%  cpuidle.POLL.time
     18742 +-  3%     -17.0%      15559 +-  2%  cpuidle.POLL.usage
      4135 +-141%    -100.0%       0.00        latency_stats.avg.x86_reserve_hardware.x86_pmu_event_init.perf_try_init_event.perf_event_alloc.__do_sys_perf_event_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
     33249 +-129%    -100.0%       0.00        latency_stats.max.call_rwsem_down_read_failed.m_start.seq_read.__vfs_read.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
      4135 +-141%    -100.0%       0.00        latency_stats.max.x86_reserve_hardware.x86_pmu_event_init.perf_try_init_event.perf_event_alloc.__do_sys_perf_event_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
     65839 +-116%    -100.0%       0.00        latency_stats.sum.call_rwsem_down_read_failed.m_start.seq_read.__vfs_read.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
      4135 +-141%    -100.0%       0.00        latency_stats.sum.x86_reserve_hardware.x86_pmu_event_init.perf_try_init_event.perf_event_alloc.__do_sys_perf_event_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
      8387 +-122%     -90.9%     767.00 +- 13%  latency_stats.sum.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_getattr.__nfs_revalidate_inode.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_lookupat
    263970 +- 10%     -68.6%      82994 +-  3%  latency_stats.sum.do_syslog.kmsg_read.proc_reg_read.__vfs_read.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
      6173 +- 77%    +173.3%      16869 +- 98%  latency_stats.sum.pipe_write.__vfs_write.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
    101.33            -4.6%      96.67        proc-vmstat.nr_anon_transparent_hugepages
     39967            -1.8%      39241        proc-vmstat.nr_slab_reclaimable
     67166            -2.4%      65522        proc-vmstat.nr_slab_unreclaimable
    237743            -3.9%     228396        proc-vmstat.nr_unevictable
    237743            -3.9%     228396        proc-vmstat.nr_zone_unevictable
 4.807e+09            -2.0%   4.71e+09        proc-vmstat.numa_hit
 4.807e+09            -2.0%   4.71e+09        proc-vmstat.numa_local
 4.791e+09            -2.1%   4.69e+09        proc-vmstat.pgalloc_normal
 4.783e+09            -2.0%  4.685e+09        proc-vmstat.pgfault
 4.807e+09            -2.0%  4.709e+09        proc-vmstat.pgfree
      1753            +4.6%       1833        turbostat.Avg_MHz
    239445           -14.1%     205783        turbostat.C1
   4617105 +-  2%     -14.8%    3934693        turbostat.C1E
      1.40 +-  2%      -0.2        1.18        turbostat.C1E%
  24764661 +-  2%     -18.0%   20297643        turbostat.C6
     18.09 +-  2%      -3.4       14.74        turbostat.C6%
      7.53 +-  2%     -17.1%       6.24        turbostat.CPU%c1
     11.88 +-  2%     -19.1%       9.61        turbostat.CPU%c6
      7.62 +-  3%     -20.8%       6.04        turbostat.Pkg%pc2
    388.30            +1.5%     393.93        turbostat.PkgWatt
    390974 +-  8%     +35.8%     530867 +- 11%  sched_debug.cfs_rq:/.min_vruntime.stddev
  -1754042           +75.7%   -3081270        sched_debug.cfs_rq:/.spread0.min
    388140 +-  8%     +36.2%     528494 +- 11%  sched_debug.cfs_rq:/.spread0.stddev
    542.30 +-  3%     -10.0%     488.21 +-  3%  sched_debug.cfs_rq:/.util_avg.min
     53.35 +- 16%     +48.7%      79.35 +- 12%  sched_debug.cfs_rq:/.util_est_enqueued.avg
     30520 +-  6%     -15.2%      25883 +- 12%  sched_debug.cpu.nr_switches.avg
    473770 +- 27%     -37.4%     296623 +- 32%  sched_debug.cpu.nr_switches.max
     17077 +-  2%     -15.1%      14493        sched_debug.cpu.nr_switches.min
     30138 +-  6%     -15.0%      25606 +- 12%  sched_debug.cpu.sched_count.avg
    472345 +- 27%     -37.2%     296419 +- 32%  sched_debug.cpu.sched_count.max
     16858 +-  2%     -15.2%      14299        sched_debug.cpu.sched_count.min
      8358 +-  2%     -15.5%       7063        sched_debug.cpu.sched_goidle.avg
     12225           -13.6%      10565        sched_debug.cpu.sched_goidle.max
      8032 +-  2%     -16.0%       6749        sched_debug.cpu.sched_goidle.min
     14839 +-  6%     -15.3%      12568 +- 12%  sched_debug.cpu.ttwu_count.avg
    235115 +- 28%     -38.3%     145175 +- 31%  sched_debug.cpu.ttwu_count.max
      7627 +-  3%     -15.9%       6413 +-  2%  sched_debug.cpu.ttwu_count.min
    226299 +- 29%     -39.5%     136827 +- 32%  sched_debug.cpu.ttwu_local.max
      0.85            -0.0        0.81        perf-stat.branch-miss-rate%
 3.675e+10            -4.1%  3.523e+10        perf-stat.branch-misses
 4.052e+11            -2.3%  3.958e+11        perf-stat.cache-misses
 7.008e+11            -2.5%  6.832e+11        perf-stat.cache-references
  15320995 +-  4%     -14.3%   13136557 +-  6%  perf-stat.context-switches
      9.16            +4.8%       9.59        perf-stat.cpi
  2.03e+14            +4.6%  2.124e+14        perf-stat.cpu-cycles
     44508            -1.7%      43743        perf-stat.cpu-migrations
      1.30            -0.1        1.24        perf-stat.dTLB-store-miss-rate%
 4.064e+10            -3.5%  3.922e+10        perf-stat.dTLB-store-misses
 3.086e+12            +1.1%  3.119e+12        perf-stat.dTLB-stores
 3.611e+08 +-  6%      -8.5%  3.304e+08 +-  5%  perf-stat.iTLB-loads
      0.11            -4.6%       0.10        perf-stat.ipc
 4.783e+09            -2.0%  4.685e+09        perf-stat.minor-faults
      1.53 +-  2%      -0.3        1.22 +-  8%  perf-stat.node-load-miss-rate%
 1.389e+09 +-  3%     -22.1%  1.083e+09 +-  9%  perf-stat.node-load-misses
 8.922e+10            -1.9%   8.75e+10        perf-stat.node-loads
      5.06            +1.7        6.77 +-  3%  perf-stat.node-store-miss-rate%
 1.204e+09           +29.3%  1.556e+09 +-  3%  perf-stat.node-store-misses
 2.256e+10            -5.1%  2.142e+10 +-  2%  perf-stat.node-stores
 4.783e+09            -2.0%  4.685e+09        perf-stat.page-faults
   1399242            +1.9%    1425404        perf-stat.path-length
      1144 +-  8%     -13.6%     988.00 +-  8%  slabinfo.Acpi-ParseExt.active_objs
      1144 +-  8%     -13.6%     988.00 +-  8%  slabinfo.Acpi-ParseExt.num_objs
      1878 +- 17%     +29.0%       2422 +- 16%  slabinfo.dmaengine-unmap-16.active_objs
      1878 +- 17%     +29.0%       2422 +- 16%  slabinfo.dmaengine-unmap-16.num_objs
      1085 +-  5%     -24.1%     823.33 +-  9%  slabinfo.file_lock_cache.active_objs
      1085 +-  5%     -24.1%     823.33 +-  9%  slabinfo.file_lock_cache.num_objs
     61584 +-  4%     -16.6%      51381 +-  5%  slabinfo.filp.active_objs
    967.00 +-  4%     -16.5%     807.67 +-  5%  slabinfo.filp.active_slabs
     61908 +-  4%     -16.5%      51713 +-  5%  slabinfo.filp.num_objs
    967.00 +-  4%     -16.5%     807.67 +-  5%  slabinfo.filp.num_slabs
      1455           -15.4%       1232 +-  4%  slabinfo.nsproxy.active_objs
      1455           -15.4%       1232 +-  4%  slabinfo.nsproxy.num_objs
     84720 +-  6%     -18.3%      69210 +-  4%  slabinfo.pid.active_objs
      1324 +-  6%     -18.2%       1083 +-  4%  slabinfo.pid.active_slabs
     84820 +-  5%     -18.2%      69386 +-  4%  slabinfo.pid.num_objs
      1324 +-  6%     -18.2%       1083 +-  4%  slabinfo.pid.num_slabs
      2112 +- 18%     -26.3%       1557 +-  5%  slabinfo.scsi_sense_cache.active_objs
      2112 +- 18%     -26.3%       1557 +-  5%  slabinfo.scsi_sense_cache.num_objs
      5018 +-  5%      -7.6%       4635 +-  4%  slabinfo.sock_inode_cache.active_objs
      5018 +-  5%      -7.6%       4635 +-  4%  slabinfo.sock_inode_cache.num_objs
      1193 +-  4%     +13.8%       1358 +-  4%  slabinfo.task_group.active_objs
      1193 +-  4%     +13.8%       1358 +-  4%  slabinfo.task_group.num_objs
     62807 +-  3%     -14.4%      53757 +-  3%  slabinfo.vm_area_struct.active_objs
      1571 +-  3%     -12.1%       1381 +-  3%  slabinfo.vm_area_struct.active_slabs
     62877 +-  3%     -14.3%      53880 +-  3%  slabinfo.vm_area_struct.num_objs
      1571 +-  3%     -12.1%       1381 +-  3%  slabinfo.vm_area_struct.num_slabs
     47.45           -47.4        0.00        perf-profile.calltrace.cycles-pp.alloc_pages_vma.__handle_mm_fault.handle_mm_fault.__do_page_fault.do_page_fault
     47.16           -47.2        0.00        perf-profile.calltrace.cycles-pp.__alloc_pages_nodemask.alloc_pages_vma.__handle_mm_fault.handle_mm_fault.__do_page_fault
     46.99           -47.0        0.00        perf-profile.calltrace.cycles-pp.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_vma.__handle_mm_fault.handle_mm_fault
     44.95           -44.9        0.00        perf-profile.calltrace.cycles-pp._raw_spin_lock.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_vma.__handle_mm_fault
      7.42 +-  2%      -7.4        0.00        perf-profile.calltrace.cycles-pp.copy_page.__handle_mm_fault.handle_mm_fault.__do_page_fault.do_page_fault
      6.32 +- 10%      -6.3        0.00        perf-profile.calltrace.cycles-pp.finish_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault.do_page_fault
      6.28 +- 10%      -6.3        0.00        perf-profile.calltrace.cycles-pp.alloc_set_pte.finish_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
      0.00            +0.9        0.85 +- 11%  perf-profile.calltrace.cycles-pp._raw_spin_lock.pte_map_lock.alloc_set_pte.finish_fault.handle_pte_fault
      0.00            +0.9        0.92 +-  4%  perf-profile.calltrace.cycles-pp.__list_del_entry_valid.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_vma.handle_pte_fault
      0.00            +1.1        1.13 +-  7%  perf-profile.calltrace.cycles-pp.find_lock_entry.shmem_getpage_gfp.shmem_fault.__do_fault.handle_pte_fault
      0.00            +1.2        1.19 +-  7%  perf-profile.calltrace.cycles-pp.shmem_getpage_gfp.shmem_fault.__do_fault.handle_pte_fault.__handle_mm_fault
      0.00            +1.2        1.22 +-  5%  perf-profile.calltrace.cycles-pp.pte_map_lock.alloc_set_pte.finish_fault.handle_pte_fault.__handle_mm_fault
      0.00            +1.3        1.34 +-  7%  perf-profile.calltrace.cycles-pp.shmem_fault.__do_fault.handle_pte_fault.__handle_mm_fault.handle_mm_fault
      0.00            +1.4        1.36 +-  7%  perf-profile.calltrace.cycles-pp.__do_fault.handle_pte_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
      0.00            +4.5        4.54 +- 19%  perf-profile.calltrace.cycles-pp.pagevec_lru_move_fn.__lru_cache_add.alloc_set_pte.finish_fault.handle_pte_fault
      0.00            +4.6        4.64 +- 19%  perf-profile.calltrace.cycles-pp.__lru_cache_add.alloc_set_pte.finish_fault.handle_pte_fault.__handle_mm_fault
      0.00            +6.6        6.64 +- 15%  perf-profile.calltrace.cycles-pp.alloc_set_pte.finish_fault.handle_pte_fault.__handle_mm_fault.handle_mm_fault
      0.00            +6.7        6.68 +- 15%  perf-profile.calltrace.cycles-pp.finish_fault.handle_pte_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
      0.00            +7.5        7.54 +-  5%  perf-profile.calltrace.cycles-pp.copy_page.handle_pte_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
      0.00           +44.6       44.55 +-  3%  perf-profile.calltrace.cycles-pp._raw_spin_lock.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_vma.handle_pte_fault
      0.00           +46.6       46.63 +-  3%  perf-profile.calltrace.cycles-pp.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_vma.handle_pte_fault.__handle_mm_fault
      0.00           +46.8       46.81 +-  3%  perf-profile.calltrace.cycles-pp.__alloc_pages_nodemask.alloc_pages_vma.handle_pte_fault.__handle_mm_fault.handle_mm_fault
      0.00           +47.1       47.10 +-  3%  perf-profile.calltrace.cycles-pp.alloc_pages_vma.handle_pte_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
      0.00           +63.1       63.15        perf-profile.calltrace.cycles-pp.handle_pte_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault.do_page_fault
      0.39 +-  3%      +0.0        0.42 +-  3%  perf-profile.children.cycles-pp.radix_tree_lookup_slot
      0.21 +-  3%      +0.0        0.25 +-  5%  perf-profile.children.cycles-pp.__mod_node_page_state
      0.00            +0.1        0.06 +-  8%  perf-profile.children.cycles-pp.get_vma_policy
      0.00            +0.1        0.08 +-  5%  perf-profile.children.cycles-pp.__lru_cache_add_active_or_unevictable
      0.00            +0.2        0.18 +-  6%  perf-profile.children.cycles-pp.__page_add_new_anon_rmap
      0.00            +1.4        1.35 +-  5%  perf-profile.children.cycles-pp.pte_map_lock
      0.00           +63.2       63.21        perf-profile.children.cycles-pp.handle_pte_fault
      1.40 +-  2%      -0.4        1.03 +- 10%  perf-profile.self.cycles-pp._raw_spin_lock
      0.56 +-  3%      -0.2        0.35 +-  6%  perf-profile.self.cycles-pp.__handle_mm_fault
      0.22 +-  3%      -0.0        0.18 +-  7%  perf-profile.self.cycles-pp.alloc_set_pte
      0.09            +0.0        0.10 +-  4%  perf-profile.self.cycles-pp.vmacache_find
      0.39 +-  2%      +0.0        0.41 +-  3%  perf-profile.self.cycles-pp.__radix_tree_lookup
      0.18            +0.0        0.20 +-  6%  perf-profile.self.cycles-pp.mem_cgroup_charge_statistics
      0.17 +-  2%      +0.0        0.20 +-  7%  perf-profile.self.cycles-pp.___might_sleep
      0.33 +-  2%      +0.0        0.36 +-  6%  perf-profile.self.cycles-pp.handle_mm_fault
      0.20 +-  2%      +0.0        0.24 +-  3%  perf-profile.self.cycles-pp.__mod_node_page_state
      0.00            +0.1        0.05        perf-profile.self.cycles-pp.finish_fault
      0.00            +0.1        0.05        perf-profile.self.cycles-pp.get_vma_policy
      0.00            +0.1        0.08 +- 10%  perf-profile.self.cycles-pp.__lru_cache_add_active_or_unevictable
      0.00            +0.2        0.25 +-  5%  perf-profile.self.cycles-pp.handle_pte_fault
      0.00            +0.5        0.49 +-  8%  perf-profile.self.cycles-pp.pte_map_lock

=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/nr_task/thp_enabled/test/cpufreq_governor:
  lkp-skl-4sp1/will-it-scale/debian-x86_64-2018-04-03.cgz/x86_64-rhel-7.2/gcc-7/100%/never/page_fault2/performance

commit: 
  ba98a1cdad71d259a194461b3a61471b49b14df1
  a7a8993bfe3ccb54ad468b9f1799649e4ad1ff12

ba98a1cdad71d259 a7a8993bfe3ccb54ad468b9f17 
---------------- -------------------------- 
       fail:runs  %reproduction    fail:runs
           |             |             |    
          1:3          -33%            :3     kmsg.DHCP/BOOTP:Reply_not_for_us_on_eth#,op[#]xid[#]
           :3           33%           1:3     dmesg.WARNING:stack_going_in_the_wrong_direction?ip=sched_slice/0x
          1:3          -33%            :3     dmesg.WARNING:stack_going_in_the_wrong_direction?ip=schedule_tail/0x
          1:3           24%           2:3     perf-profile.calltrace.cycles-pp.sync_regs.error_entry
          3:3           46%           5:3     perf-profile.calltrace.cycles-pp.error_entry
          5:3           -9%           5:3     perf-profile.children.cycles-pp.error_entry
          2:3           -4%           2:3     perf-profile.self.cycles-pp.error_entry
         %stddev     %change         %stddev
             \          |                \  
      8147           -18.8%       6613        will-it-scale.per_thread_ops
     93113           +17.0%     108982        will-it-scale.time.involuntary_context_switches
 4.732e+08           -19.0%  3.833e+08        will-it-scale.time.minor_page_faults
      5854           +12.0%       6555        will-it-scale.time.percent_of_cpu_this_job_got
     35247           +12.1%      39495        will-it-scale.time.system_time
   5546661           -15.5%    4689314        will-it-scale.time.voluntary_context_switches
  15801637            -1.9%   15504487        will-it-scale.workload
      1.43 +- 11%     -59.7%       0.58 +- 28%  irq_exception_noise.__do_page_fault.min
      2811 +-  3%     +23.7%       3477 +-  3%  kthread_noise.total_time
    292776 +-  5%     +39.6%     408829 +- 21%  meminfo.DirectMap4k
     19.80            -3.7       16.12        mpstat.cpu.idle%
     29940           -14.5%      25593        uptime.idle
     24064 +-  3%      -8.5%      22016        vmstat.system.cs
     34.86            -1.9%      34.19        boot-time.boot
     26.95            -2.8%      26.19 +-  2%  boot-time.kernel_boot
   7190569 +-  2%     -15.2%    6100136 +-  3%  softirqs.RCU
   5513663           -13.8%    4751548        softirqs.SCHED
     18064 +-  2%     +24.3%      22461 +-  7%  numa-vmstat.node0.nr_slab_unreclaimable
      8507 +- 12%     -16.8%       7075 +-  4%  numa-vmstat.node2.nr_slab_reclaimable
     18719 +-  9%     -19.6%      15043 +-  4%  numa-vmstat.node3.nr_slab_unreclaimable
     72265 +-  2%     +24.3%      89855 +-  7%  numa-meminfo.node0.SUnreclaim
    115980 +-  4%     +22.6%     142233 +- 12%  numa-meminfo.node0.Slab
     34035 +- 12%     -16.8%      28307 +-  4%  numa-meminfo.node2.SReclaimable
     74888 +-  9%     -19.7%      60162 +-  4%  numa-meminfo.node3.SUnreclaim
     93113           +17.0%     108982        time.involuntary_context_switches
 4.732e+08           -19.0%  3.833e+08        time.minor_page_faults
      5854           +12.0%       6555        time.percent_of_cpu_this_job_got
     35247           +12.1%      39495        time.system_time
   5546661           -15.5%    4689314        time.voluntary_context_switches
 4.792e+09            -1.9%  4.699e+09        proc-vmstat.numa_hit
 4.791e+09            -1.9%  4.699e+09        proc-vmstat.numa_local
     40447 +- 11%     +13.2%      45804 +-  6%  proc-vmstat.pgactivate
 4.778e+09            -1.9%  4.688e+09        proc-vmstat.pgalloc_normal
 4.767e+09            -1.9%  4.675e+09        proc-vmstat.pgfault
 4.791e+09            -1.9%  4.699e+09        proc-vmstat.pgfree
    230178 +-  2%     -10.1%     206883 +-  3%  cpuidle.C1.usage
 1.617e+09           -15.0%  1.375e+09        cpuidle.C1E.time
   4514401           -14.1%    3878206        cpuidle.C1E.usage
 2.087e+10           -18.5%  1.701e+10        cpuidle.C6.time
  24458365           -18.0%   20045336        cpuidle.C6.usage
   1163758           -16.1%     976094 +-  4%  cpuidle.POLL.time
     17907           -14.6%      15294 +-  4%  cpuidle.POLL.usage
      1758            +4.5%       1838        turbostat.Avg_MHz
    227522 +-  2%     -10.2%     204426 +-  3%  turbostat.C1
   4512700           -14.2%    3873264        turbostat.C1E
      1.39            -0.2        1.18        turbostat.C1E%
  24452583           -18.0%   20039031        turbostat.C6
     17.85            -3.3       14.55        turbostat.C6%
      7.44           -16.8%       6.19        turbostat.CPU%c1
     11.72           -19.3%       9.45        turbostat.CPU%c6
      7.51           -21.3%       5.91        turbostat.Pkg%pc2
    389.33            +1.6%     395.59        turbostat.PkgWatt
    559.33 +- 13%     -17.9%     459.33 +- 20%  slabinfo.dmaengine-unmap-128.active_objs
    559.33 +- 13%     -17.9%     459.33 +- 20%  slabinfo.dmaengine-unmap-128.num_objs
     57734 +-  3%      -5.7%      54421 +-  4%  slabinfo.filp.active_objs
    905.67 +-  3%      -5.6%     854.67 +-  4%  slabinfo.filp.active_slabs
     57981 +-  3%      -5.6%      54720 +-  4%  slabinfo.filp.num_objs
    905.67 +-  3%      -5.6%     854.67 +-  4%  slabinfo.filp.num_slabs
      1378           -12.0%       1212 +-  7%  slabinfo.nsproxy.active_objs
      1378           -12.0%       1212 +-  7%  slabinfo.nsproxy.num_objs
    507.33 +-  7%     -26.8%     371.33 +-  2%  slabinfo.secpath_cache.active_objs
    507.33 +-  7%     -26.8%     371.33 +-  2%  slabinfo.secpath_cache.num_objs
      4788 +-  5%      -8.3%       4391 +-  2%  slabinfo.sock_inode_cache.active_objs
      4788 +-  5%      -8.3%       4391 +-  2%  slabinfo.sock_inode_cache.num_objs
      1431 +-  8%     -12.3%       1255 +-  3%  slabinfo.task_group.active_objs
      1431 +-  8%     -12.3%       1255 +-  3%  slabinfo.task_group.num_objs
      4.27 +- 17%     +27.0%       5.42 +-  7%  sched_debug.cfs_rq:/.runnable_load_avg.avg
     13.44 +- 62%     +73.6%      23.33 +- 24%  sched_debug.cfs_rq:/.runnable_load_avg.stddev
    772.55 +- 21%     -32.7%     520.27 +-  4%  sched_debug.cfs_rq:/.util_est_enqueued.max
      4.39 +- 15%     +29.0%       5.66 +- 11%  sched_debug.cpu.cpu_load[0].avg
    152.09 +- 72%     +83.9%     279.67 +- 33%  sched_debug.cpu.cpu_load[0].max
     13.84 +- 58%     +78.7%      24.72 +- 29%  sched_debug.cpu.cpu_load[0].stddev
      4.53 +- 14%     +25.8%       5.70 +- 10%  sched_debug.cpu.cpu_load[1].avg
    156.58 +- 66%     +76.6%     276.58 +- 33%  sched_debug.cpu.cpu_load[1].max
     14.02 +- 55%     +72.4%      24.17 +- 28%  sched_debug.cpu.cpu_load[1].stddev
      4.87 +- 11%     +17.3%       5.72 +-  9%  sched_debug.cpu.cpu_load[2].avg
      1.58 +-  2%     +13.5%       1.79 +-  6%  sched_debug.cpu.nr_running.max
     16694           -14.6%      14259        sched_debug.cpu.nr_switches.min
     31989 +- 13%     +20.6%      38584 +-  6%  sched_debug.cpu.nr_switches.stddev
     16505           -14.8%      14068        sched_debug.cpu.sched_count.min
     32084 +- 13%     +19.9%      38482 +-  6%  sched_debug.cpu.sched_count.stddev
      8185           -15.0%       6957        sched_debug.cpu.sched_goidle.avg
     12151 +-  2%     -13.5%      10507        sched_debug.cpu.sched_goidle.max
      7867           -15.7%       6631        sched_debug.cpu.sched_goidle.min
      7595           -16.1%       6375        sched_debug.cpu.ttwu_count.min
     15873 +- 13%     +21.2%      19239 +-  6%  sched_debug.cpu.ttwu_count.stddev
      5244 +- 17%     +17.0%       6134 +-  5%  sched_debug.cpu.ttwu_local.avg
     15646 +- 12%     +21.5%      19008 +-  6%  sched_debug.cpu.ttwu_local.stddev
      0.85            -0.0        0.81        perf-stat.branch-miss-rate%
 3.689e+10            -4.6%  3.518e+10        perf-stat.branch-misses
     57.39            +0.6       58.00        perf-stat.cache-miss-rate%
 4.014e+11            -1.2%  3.967e+11        perf-stat.cache-misses
 6.994e+11            -2.2%   6.84e+11        perf-stat.cache-references
  14605393 +-  3%      -8.5%   13369913        perf-stat.context-switches
      9.21            +4.5%       9.63        perf-stat.cpi
 2.037e+14            +4.6%   2.13e+14        perf-stat.cpu-cycles
     44424            -2.0%      43541        perf-stat.cpu-migrations
      1.29            -0.1        1.24        perf-stat.dTLB-store-miss-rate%
 4.018e+10            -2.8%  3.905e+10        perf-stat.dTLB-store-misses
 3.071e+12            +1.4%  3.113e+12        perf-stat.dTLB-stores
     93.04            +1.5       94.51        perf-stat.iTLB-load-miss-rate%
 4.946e+09           +19.3%  5.903e+09 +-  5%  perf-stat.iTLB-load-misses
 3.702e+08            -7.5%  3.423e+08 +-  2%  perf-stat.iTLB-loads
      4470           -15.9%       3760 +-  5%  perf-stat.instructions-per-iTLB-miss
      0.11            -4.3%       0.10        perf-stat.ipc
 4.767e+09            -1.9%  4.675e+09        perf-stat.minor-faults
      1.46 +-  4%      -0.1        1.33 +-  9%  perf-stat.node-load-miss-rate%
      4.91            +1.7        6.65 +-  2%  perf-stat.node-store-miss-rate%
 1.195e+09           +32.8%  1.587e+09 +-  2%  perf-stat.node-store-misses
 2.313e+10            -3.7%  2.227e+10        perf-stat.node-stores
 4.767e+09            -1.9%  4.675e+09        perf-stat.page-faults
   1399047            +2.0%    1427115        perf-stat.path-length
      8908 +- 73%    -100.0%       0.00        latency_stats.avg.call_rwsem_down_read_failed.m_start.seq_read.__vfs_read.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
      3604 +-141%    -100.0%       0.00        latency_stats.avg.call_rwsem_down_write_failed.do_unlinkat.do_syscall_64.entry_SYSCALL_64_after_hwframe
     61499 +-130%     -92.6%       4534 +- 16%  latency_stats.avg.expand_files.__alloc_fd.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
      4391 +-138%     -70.9%       1277 +-129%  latency_stats.avg.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_lookup.nfs_lookup_revalidate.lookup_fast.walk_component.link_path_walk.path_lookupat.filename_lookup
     67311 +-112%     -48.5%      34681 +- 36%  latency_stats.avg.max
      3956 +-138%    +320.4%      16635 +-140%  latency_stats.avg.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_getattr.__nfs_revalidate_inode.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_lookupat
    164.67 +- 30%   +7264.0%      12126 +-138%  latency_stats.avg.flush_work.fsnotify_destroy_group.inotify_release.__fput.task_work_run.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00       +5.4e+105%       5367 +-141%  latency_stats.avg.call_rwsem_down_write_failed.unlink_file_vma.free_pgtables.exit_mmap.mmput.flush_old_exec.load_elf_binary.search_binary_handler.do_execveat_common.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe
     36937 +-119%    -100.0%       0.00        latency_stats.max.call_rwsem_down_read_failed.m_start.seq_read.__vfs_read.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
      3604 +-141%    -100.0%       0.00        latency_stats.max.call_rwsem_down_write_failed.do_unlinkat.do_syscall_64.entry_SYSCALL_64_after_hwframe
     84146 +-107%     -72.5%      23171 +- 31%  latency_stats.max.expand_files.__alloc_fd.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
      4391 +-138%     -70.9%       1277 +-129%  latency_stats.max.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_lookup.nfs_lookup_revalidate.lookup_fast.walk_component.link_path_walk.path_lookupat.filename_lookup
      5817 +- 83%     -69.7%       1760 +- 67%  latency_stats.max.pipe_write.__vfs_write.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
      6720 +-137%   +1628.2%     116147 +-141%  latency_stats.max.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_getattr.__nfs_revalidate_inode.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_lookupat
    164.67 +- 30%   +7264.0%      12126 +-138%  latency_stats.max.flush_work.fsnotify_destroy_group.inotify_release.__fput.task_work_run.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00       +1.2e+106%      12153 +-141%  latency_stats.max.call_rwsem_down_write_failed.unlink_file_vma.free_pgtables.exit_mmap.mmput.flush_old_exec.load_elf_binary.search_binary_handler.do_execveat_common.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe
    110122 +-120%    -100.0%       0.00        latency_stats.sum.call_rwsem_down_read_failed.m_start.seq_read.__vfs_read.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
      3604 +-141%    -100.0%       0.00        latency_stats.sum.call_rwsem_down_write_failed.do_unlinkat.do_syscall_64.entry_SYSCALL_64_after_hwframe
  12078828 +-139%     -99.3%      89363 +- 29%  latency_stats.sum.expand_files.__alloc_fd.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
    144453 +-120%     -80.9%      27650 +- 19%  latency_stats.sum.poll_schedule_timeout.do_sys_poll.__x64_sys_poll.do_syscall_64.entry_SYSCALL_64_after_hwframe
      4391 +-138%     -70.9%       1277 +-129%  latency_stats.sum.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_lookup.nfs_lookup_revalidate.lookup_fast.walk_component.link_path_walk.path_lookupat.filename_lookup
      9438 +- 86%     -68.4%       2980 +- 35%  latency_stats.sum.pipe_write.__vfs_write.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
     31656 +-138%    +320.4%     133084 +-140%  latency_stats.sum.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_getattr.__nfs_revalidate_inode.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_lookupat
    164.67 +- 30%   +7264.0%      12126 +-138%  latency_stats.sum.flush_work.fsnotify_destroy_group.inotify_release.__fput.task_work_run.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00       +8.8e+105%       8760 +-141%  latency_stats.sum.msleep_interruptible.uart_wait_until_sent.tty_wait_until_sent.tty_port_close_start.tty_port_close.tty_release.__fput.task_work_run.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00       +1.3e+106%      12897 +-141%  latency_stats.sum.tty_wait_until_sent.tty_port_close_start.tty_port_close.tty_release.__fput.task_work_run.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00       +3.2e+106%      32207 +-141%  latency_stats.sum.call_rwsem_down_write_failed.unlink_file_vma.free_pgtables.exit_mmap.mmput.flush_old_exec.load_elf_binary.search_binary_handler.do_execveat_common.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe
     44.43 +-  3%     -44.4        0.00        perf-profile.calltrace.cycles-pp.alloc_pages_vma.__handle_mm_fault.handle_mm_fault.__do_page_fault.do_page_fault
     44.13 +-  3%     -44.1        0.00        perf-profile.calltrace.cycles-pp.__alloc_pages_nodemask.alloc_pages_vma.__handle_mm_fault.handle_mm_fault.__do_page_fault
     43.95 +-  3%     -43.9        0.00        perf-profile.calltrace.cycles-pp.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_vma.__handle_mm_fault.handle_mm_fault
     41.85 +-  4%     -41.9        0.00        perf-profile.calltrace.cycles-pp._raw_spin_lock.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_vma.__handle_mm_fault
      7.74 +-  8%      -7.7        0.00        perf-profile.calltrace.cycles-pp.copy_page.__handle_mm_fault.handle_mm_fault.__do_page_fault.do_page_fault
      7.19 +-  4%      -7.2        0.00        perf-profile.calltrace.cycles-pp.finish_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault.do_page_fault
      7.15 +-  4%      -7.2        0.00        perf-profile.calltrace.cycles-pp.alloc_set_pte.finish_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
      5.09 +-  3%      -5.1        0.00        perf-profile.calltrace.cycles-pp.__lru_cache_add.alloc_set_pte.finish_fault.__handle_mm_fault.handle_mm_fault
      4.99 +-  3%      -5.0        0.00        perf-profile.calltrace.cycles-pp.pagevec_lru_move_fn.__lru_cache_add.alloc_set_pte.finish_fault.__handle_mm_fault
      0.93 +-  6%      -0.1        0.81 +-  2%  perf-profile.calltrace.cycles-pp.find_get_entry.find_lock_entry.shmem_getpage_gfp.shmem_fault.__do_fault
      0.00            +0.8        0.84        perf-profile.calltrace.cycles-pp._raw_spin_lock.pte_map_lock.alloc_set_pte.finish_fault.handle_pte_fault
      0.00            +0.9        0.92 +-  3%  perf-profile.calltrace.cycles-pp.__list_del_entry_valid.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_vma.handle_pte_fault
      0.00            +1.1        1.08        perf-profile.calltrace.cycles-pp.find_lock_entry.shmem_getpage_gfp.shmem_fault.__do_fault.handle_pte_fault
      0.00            +1.1        1.14        perf-profile.calltrace.cycles-pp.shmem_getpage_gfp.shmem_fault.__do_fault.handle_pte_fault.__handle_mm_fault
      0.00            +1.2        1.17        perf-profile.calltrace.cycles-pp.pte_map_lock.alloc_set_pte.finish_fault.handle_pte_fault.__handle_mm_fault
      0.00            +1.3        1.29        perf-profile.calltrace.cycles-pp.shmem_fault.__do_fault.handle_pte_fault.__handle_mm_fault.handle_mm_fault
      0.00            +1.3        1.31        perf-profile.calltrace.cycles-pp.__do_fault.handle_pte_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
     61.62            +1.7       63.33        perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.__do_page_fault.do_page_fault.page_fault
     41.73 +-  4%      +3.0       44.75        perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_vma
      0.00            +4.6        4.55 +- 15%  perf-profile.calltrace.cycles-pp.pagevec_lru_move_fn.__lru_cache_add.alloc_set_pte.finish_fault.handle_pte_fault
      0.00            +4.6        4.65 +- 14%  perf-profile.calltrace.cycles-pp.__lru_cache_add.alloc_set_pte.finish_fault.handle_pte_fault.__handle_mm_fault
      0.00            +6.6        6.57 +- 10%  perf-profile.calltrace.cycles-pp.alloc_set_pte.finish_fault.handle_pte_fault.__handle_mm_fault.handle_mm_fault
      0.00            +6.6        6.61 +- 10%  perf-profile.calltrace.cycles-pp.finish_fault.handle_pte_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
      0.00            +7.2        7.25 +-  2%  perf-profile.calltrace.cycles-pp.copy_page.handle_pte_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
     41.41 +- 70%     +22.3       63.67        perf-profile.calltrace.cycles-pp.handle_mm_fault.__do_page_fault.do_page_fault.page_fault
     42.19 +- 70%     +22.6       64.75        perf-profile.calltrace.cycles-pp.__do_page_fault.do_page_fault.page_fault
     42.20 +- 70%     +22.6       64.76        perf-profile.calltrace.cycles-pp.do_page_fault.page_fault
     42.27 +- 70%     +22.6       64.86        perf-profile.calltrace.cycles-pp.page_fault
      0.00           +44.9       44.88        perf-profile.calltrace.cycles-pp._raw_spin_lock.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_vma.handle_pte_fault
      0.00           +46.9       46.92        perf-profile.calltrace.cycles-pp.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_vma.handle_pte_fault.__handle_mm_fault
      0.00           +47.1       47.10        perf-profile.calltrace.cycles-pp.__alloc_pages_nodemask.alloc_pages_vma.handle_pte_fault.__handle_mm_fault.handle_mm_fault
      0.00           +47.4       47.37        perf-profile.calltrace.cycles-pp.alloc_pages_vma.handle_pte_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
      0.00           +63.0       63.00        perf-profile.calltrace.cycles-pp.handle_pte_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault.do_page_fault
      0.97 +-  6%      -0.1        0.84 +-  2%  perf-profile.children.cycles-pp.find_get_entry
      1.23 +-  6%      -0.1        1.11        perf-profile.children.cycles-pp.find_lock_entry
      0.09 +- 10%      -0.0        0.07 +-  6%  perf-profile.children.cycles-pp.unlock_page
      0.19 +-  4%      +0.0        0.21 +-  2%  perf-profile.children.cycles-pp.mem_cgroup_charge_statistics
      0.21 +-  2%      +0.0        0.25        perf-profile.children.cycles-pp.__mod_node_page_state
      0.00            +0.1        0.05 +-  8%  perf-profile.children.cycles-pp.get_vma_policy
      0.00            +0.1        0.08        perf-profile.children.cycles-pp.__lru_cache_add_active_or_unevictable
      0.00            +0.2        0.18 +-  2%  perf-profile.children.cycles-pp.__page_add_new_anon_rmap
      0.00            +1.3        1.30        perf-profile.children.cycles-pp.pte_map_lock
     63.40            +1.6       64.97        perf-profile.children.cycles-pp.__do_page_fault
     63.19            +1.6       64.83        perf-profile.children.cycles-pp.do_page_fault
     61.69            +1.7       63.36        perf-profile.children.cycles-pp.__handle_mm_fault
     63.19            +1.7       64.86        perf-profile.children.cycles-pp.page_fault
     61.99            +1.7       63.70        perf-profile.children.cycles-pp.handle_mm_fault
     72.27            +2.2       74.52        perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
     67.51            +2.4       69.87        perf-profile.children.cycles-pp._raw_spin_lock
     44.49 +-  3%      +3.0       47.45        perf-profile.children.cycles-pp.alloc_pages_vma
     44.28 +-  3%      +3.0       47.26        perf-profile.children.cycles-pp.__alloc_pages_nodemask
     44.13 +-  3%      +3.0       47.12        perf-profile.children.cycles-pp.get_page_from_freelist
      0.00           +63.1       63.06        perf-profile.children.cycles-pp.handle_pte_fault
      1.46 +-  7%      -0.5        1.01        perf-profile.self.cycles-pp._raw_spin_lock
      0.58 +-  6%      -0.2        0.34        perf-profile.self.cycles-pp.__handle_mm_fault
      0.55 +-  6%      -0.1        0.44 +-  2%  perf-profile.self.cycles-pp.find_get_entry
      0.22 +-  5%      -0.1        0.16 +-  2%  perf-profile.self.cycles-pp.alloc_set_pte
      0.10 +-  8%      -0.0        0.08        perf-profile.self.cycles-pp.down_read_trylock
      0.09 +-  5%      -0.0        0.07        perf-profile.self.cycles-pp.unlock_page
      0.06            -0.0        0.05        perf-profile.self.cycles-pp.pmd_devmap_trans_unstable
      0.20 +-  2%      +0.0        0.24 +-  3%  perf-profile.self.cycles-pp.__mod_node_page_state
      0.00            +0.1        0.05        perf-profile.self.cycles-pp.finish_fault
      0.00            +0.1        0.05        perf-profile.self.cycles-pp.get_vma_policy
      0.00            +0.1        0.08 +-  6%  perf-profile.self.cycles-pp.__lru_cache_add_active_or_unevictable
      0.00            +0.2        0.25        perf-profile.self.cycles-pp.handle_pte_fault
      0.00            +0.5        0.46 +-  7%  perf-profile.self.cycles-pp.pte_map_lock
     72.26            +2.3       74.52        perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath

[-- Attachment #3: perf-profile.zip --]
[-- Type: application/zip, Size: 19025 bytes --]

WARNING: multiple messages have this Message-ID (diff)
From: Haiyan Song <haiyanx.song@intel.com>
To: Laurent Dufour <ldufour@linux.vnet.ibm.com>
Cc: "akpm@linux-foundation.org" <akpm@linux-foundation.org>,
	"mhocko@kernel.org" <mhocko@kernel.org>,
	"peterz@infradead.org" <peterz@infradead.org>,
	"kirill@shutemov.name" <kirill@shutemov.name>,
	"ak@linux.intel.com" <ak@linux.intel.com>,
	"dave@stgolabs.net" <dave@stgolabs.net>,
	"jack@suse.cz" <jack@suse.cz>,
	Matthew Wilcox <willy@infradead.org>,
	"khandual@linux.vnet.ibm.com" <khandual@linux.vnet.ibm.com>,
	"aneesh.kumar@linux.vnet.ibm.com"
	<aneesh.kumar@linux.vnet.ibm.com>,
	"benh@kernel.crashing.org" <benh@kernel.crashing.org>,
	"mpe@ellerman.id.au" <mpe@ellerman.id.au>,
	"paulus@samba.org" <paulus@samba.org>,
	Thomas Gleixner <tglx@linutronix.de>,
	Ingo Molnar <mingo@redhat.com>, "hpa@zytor.com" <hpa@zytor.com>,
	Will Deacon <will.deacon@arm.com>,
	Sergey Senozhatsky <sergey.senozhatsky@gmail.com>,
	"sergey.senozhatsky.work@gmail.com"
	<sergey.senozhatsky.work@gmail.com>,
	Andrea Arcangeli <aarcange@redhat.com>,
	Alexei Starovoitov <alexei.starovoitov@gmail.com>,
	"Wang, Kemi" <kemi.wang@intel.com>,
	Daniel Jordan <daniel.m.jordan@oracle.com>,
	David Rientjes <rientjes@google.com>,
	Jerome Glisse <jglisse@redhat.com>,
	Ganesh Mahendran <opensource.ganesh@gmail.com>,
	Minchan Kim <minchan@kernel.org>,
	Punit Agrawal <punitagrawal@gmail.com>,
	vinayak menon <vinayakm.list@gmail.com>,
	Yang Shi <yang.shi@linux.alibaba.com>,
	"linux-kernel@vger.kernel.org" <linux-kernel@vger.kernel.org>,
	"linux-mm@kvack.org" <linux-mm@kvack.org>,
	"haren@linux.vnet.ibm.com" <haren@linux.vnet.ibm.com>,
	"npiggin@gmail.com" <npiggin@gmail.com>,
	"bsingharora@gmail.com" <bsingharora@gmail.com>,
	"paulmck@linux.vnet.ibm.com" <paulmck@linux.vnet.ibm.com>,
	Tim Chen <tim.c.chen@linux.intel.com>,
	"linuxppc-dev@lists.ozlabs.org" <linuxppc-dev@lists.ozlabs.org>,
	"x86@kernel.org" <x86@kernel.org>
Subject: Re: [PATCH v11 00/26] Speculative page faults
Date: Tue, 19 Jun 2018 17:16:25 +0800	[thread overview]
Message-ID: <20180619091625.jfj7qwueisddas2h@haiyan.lkp.sh.intel.com> (raw)
In-Reply-To: <b13302ab-da49-22a4-eda7-68a6528fb9e7@linux.vnet.ibm.com>

[-- Attachment #1: Type: text/plain, Size: 31691 bytes --]

On Mon, Jun 11, 2018 at 05:15:22PM +0200, Laurent Dufour wrote:

Hi Laurent,

For perf date tested on Intel 4s Skylake platform, here attached the compare result
between base and head commit in attachment, which include the perf-profile comparision information.

And also attached some perf-profile.json captured from test result for page_fault2 and page_fault3 for
checking the regression, thanks.


Best regards,
Haiyan Song



> Hi Haiyan,
> 
> I don't have access to the same hardware you ran the test on, but I give a try
> to those test on a Power8 system (2 sockets, 5 cores/s, 8 threads/c, 80 CPUs 32G).
> I run each will-it-scale test 10 times and compute the average.
> 
> test THP enabled		4.17.0-rc4-mm1	spf		delta
> page_fault3_threads		2697.7		2683.5		-0.53%
> page_fault2_threads		170660.6	169574.1	-0.64%
> context_switch1_threads		6915269.2	6877507.3	-0.55%
> context_switch1_processes	6478076.2	6529493.5	0.79%
> rk1				243391.2	238527.5	-2.00%
> 
> Test were launched with the arguments '-t 80 -s 5', only the average report is
> taken in account. Note that page size is 64K by default on ppc64.
> 
> It would be nice if you could capture some perf data to figure out why the
> page_fault2/3 are showing such a performance regression.
> 
> Thanks,
> Laurent.
> 
> On 11/06/2018 09:49, Song, HaiyanX wrote:
> > Hi Laurent,
> > 
> > Regression test for v11 patch serials have been run, some regression is found by LKP-tools (linux kernel performance)
> > tested on Intel 4s skylake platform. This time only test the cases which have been run and found regressions on
> > V9 patch serials.
> > 
> > The regression result is sorted by the metric will-it-scale.per_thread_ops.
> > branch: Laurent-Dufour/Speculative-page-faults/20180520-045126
> > commit id:
> >   head commit : a7a8993bfe3ccb54ad468b9f1799649e4ad1ff12
> >   base commit : ba98a1cdad71d259a194461b3a61471b49b14df1
> > Benchmark: will-it-scale
> > Download link: https://github.com/antonblanchard/will-it-scale/tree/master
> > 
> > Metrics:
> >   will-it-scale.per_process_ops=processes/nr_cpu
> >   will-it-scale.per_thread_ops=threads/nr_cpu
> >   test box: lkp-skl-4sp1(nr_cpu=192,memory=768G)
> > THP: enable / disable
> > nr_task:100%
> > 
> > 1. Regressions:
> > 
> > a). Enable THP
> > testcase                          base           change      head           metric
> > page_fault3/enable THP           10519          -20.5%       8368      will-it-scale.per_thread_ops
> > page_fault2/enalbe THP            8281          -18.8%       6728      will-it-scale.per_thread_ops
> > brk1/eanble THP                 998475           -2.2%     976893      will-it-scale.per_process_ops
> > context_switch1/enable THP      223910           -1.3%     220930      will-it-scale.per_process_ops
> > context_switch1/enable THP      233722           -1.0%     231288      will-it-scale.per_thread_ops
> > 
> > b). Disable THP
> > page_fault3/disable THP          10856          -23.1%       8344      will-it-scale.per_thread_ops
> > page_fault2/disable THP           8147          -18.8%       6613      will-it-scale.per_thread_ops
> > brk1/disable THP                   957           -7.9%        881      will-it-scale.per_thread_ops
> > context_switch1/disable THP     237006           -2.2%     231907      will-it-scale.per_thread_ops
> > brk1/disable THP                997317           -2.0%     977778      will-it-scale.per_process_ops
> > page_fault3/disable THP         467454           -1.8%     459251      will-it-scale.per_process_ops
> > context_switch1/disable THP     224431           -1.3%     221567      will-it-scale.per_process_ops
> > 
> > Notes: for the above  values of test result, the higher is better.
> > 
> > 2. Improvement: not found improvement based on the selected test cases.
> > 
> > 
> > Best regards
> > Haiyan Song
> > ________________________________________
> > From: owner-linux-mm@kvack.org [owner-linux-mm@kvack.org] on behalf of Laurent Dufour [ldufour@linux.vnet.ibm.com]
> > Sent: Monday, May 28, 2018 4:54 PM
> > To: Song, HaiyanX
> > Cc: akpm@linux-foundation.org; mhocko@kernel.org; peterz@infradead.org; kirill@shutemov.name; ak@linux.intel.com; dave@stgolabs.net; jack@suse.cz; Matthew Wilcox; khandual@linux.vnet.ibm.com; aneesh.kumar@linux.vnet.ibm.com; benh@kernel.crashing.org; mpe@ellerman.id.au; paulus@samba.org; Thomas Gleixner; Ingo Molnar; hpa@zytor.com; Will Deacon; Sergey Senozhatsky; sergey.senozhatsky.work@gmail.com; Andrea Arcangeli; Alexei Starovoitov; Wang, Kemi; Daniel Jordan; David Rientjes; Jerome Glisse; Ganesh Mahendran; Minchan Kim; Punit Agrawal; vinayak menon; Yang Shi; linux-kernel@vger.kernel.org; linux-mm@kvack.org; haren@linux.vnet.ibm.com; npiggin@gmail.com; bsingharora@gmail.com; paulmck@linux.vnet.ibm.com; Tim Chen; linuxppc-dev@lists.ozlabs.org; x86@kernel.org
> > Subject: Re: [PATCH v11 00/26] Speculative page faults
> > 
> > On 28/05/2018 10:22, Haiyan Song wrote:
> >> Hi Laurent,
> >>
> >> Yes, these tests are done on V9 patch.
> > 
> > Do you plan to give this V11 a run ?
> > 
> >>
> >>
> >> Best regards,
> >> Haiyan Song
> >>
> >> On Mon, May 28, 2018 at 09:51:34AM +0200, Laurent Dufour wrote:
> >>> On 28/05/2018 07:23, Song, HaiyanX wrote:
> >>>>
> >>>> Some regression and improvements is found by LKP-tools(linux kernel performance) on V9 patch series
> >>>> tested on Intel 4s Skylake platform.
> >>>
> >>> Hi,
> >>>
> >>> Thanks for reporting this benchmark results, but you mentioned the "V9 patch
> >>> series" while responding to the v11 header series...
> >>> Were these tests done on v9 or v11 ?
> >>>
> >>> Cheers,
> >>> Laurent.
> >>>
> >>>>
> >>>> The regression result is sorted by the metric will-it-scale.per_thread_ops.
> >>>> Branch: Laurent-Dufour/Speculative-page-faults/20180316-151833 (V9 patch series)
> >>>> Commit id:
> >>>>     base commit: d55f34411b1b126429a823d06c3124c16283231f
> >>>>     head commit: 0355322b3577eeab7669066df42c550a56801110
> >>>> Benchmark suite: will-it-scale
> >>>> Download link:
> >>>> https://github.com/antonblanchard/will-it-scale/tree/master/tests
> >>>> Metrics:
> >>>>     will-it-scale.per_process_ops=processes/nr_cpu
> >>>>     will-it-scale.per_thread_ops=threads/nr_cpu
> >>>> test box: lkp-skl-4sp1(nr_cpu=192,memory=768G)
> >>>> THP: enable / disable
> >>>> nr_task: 100%
> >>>>
> >>>> 1. Regressions:
> >>>> a) THP enabled:
> >>>> testcase                        base            change          head       metric
> >>>> page_fault3/ enable THP         10092           -17.5%          8323       will-it-scale.per_thread_ops
> >>>> page_fault2/ enable THP          8300           -17.2%          6869       will-it-scale.per_thread_ops
> >>>> brk1/ enable THP                  957.67         -7.6%           885       will-it-scale.per_thread_ops
> >>>> page_fault3/ enable THP        172821            -5.3%        163692       will-it-scale.per_process_ops
> >>>> signal1/ enable THP              9125            -3.2%          8834       will-it-scale.per_process_ops
> >>>>
> >>>> b) THP disabled:
> >>>> testcase                        base            change          head       metric
> >>>> page_fault3/ disable THP        10107           -19.1%          8180       will-it-scale.per_thread_ops
> >>>> page_fault2/ disable THP         8432           -17.8%          6931       will-it-scale.per_thread_ops
> >>>> context_switch1/ disable THP   215389            -6.8%        200776       will-it-scale.per_thread_ops
> >>>> brk1/ disable THP                 939.67         -6.6%           877.33    will-it-scale.per_thread_ops
> >>>> page_fault3/ disable THP       173145            -4.7%        165064       will-it-scale.per_process_ops
> >>>> signal1/ disable THP             9162            -3.9%          8802       will-it-scale.per_process_ops
> >>>>
> >>>> 2. Improvements:
> >>>> a) THP enabled:
> >>>> testcase                        base            change          head       metric
> >>>> malloc1/ enable THP               66.33        +469.8%           383.67    will-it-scale.per_thread_ops
> >>>> writeseek3/ enable THP          2531             +4.5%          2646       will-it-scale.per_thread_ops
> >>>> signal1/ enable THP              989.33          +2.8%          1016       will-it-scale.per_thread_ops
> >>>>
> >>>> b) THP disabled:
> >>>> testcase                        base            change          head       metric
> >>>> malloc1/ disable THP              90.33        +417.3%           467.33    will-it-scale.per_thread_ops
> >>>> read2/ disable THP             58934            +39.2%         82060       will-it-scale.per_thread_ops
> >>>> page_fault1/ disable THP        8607            +36.4%         11736       will-it-scale.per_thread_ops
> >>>> read1/ disable THP            314063            +12.7%        353934       will-it-scale.per_thread_ops
> >>>> writeseek3/ disable THP         2452            +12.5%          2759       will-it-scale.per_thread_ops
> >>>> signal1/ disable THP             971.33          +5.5%          1024       will-it-scale.per_thread_ops
> >>>>
> >>>> Notes: for above values in column "change", the higher value means that the related testcase result
> >>>> on head commit is better than that on base commit for this benchmark.
> >>>>
> >>>>
> >>>> Best regards
> >>>> Haiyan Song
> >>>>
> >>>> ________________________________________
> >>>> From: owner-linux-mm@kvack.org [owner-linux-mm@kvack.org] on behalf of Laurent Dufour [ldufour@linux.vnet.ibm.com]
> >>>> Sent: Thursday, May 17, 2018 7:06 PM
> >>>> To: akpm@linux-foundation.org; mhocko@kernel.org; peterz@infradead.org; kirill@shutemov.name; ak@linux.intel.com; dave@stgolabs.net; jack@suse.cz; Matthew Wilcox; khandual@linux.vnet.ibm.com; aneesh.kumar@linux.vnet.ibm.com; benh@kernel.crashing.org; mpe@ellerman.id.au; paulus@samba.org; Thomas Gleixner; Ingo Molnar; hpa@zytor.com; Will Deacon; Sergey Senozhatsky; sergey.senozhatsky.work@gmail.com; Andrea Arcangeli; Alexei Starovoitov; Wang, Kemi; Daniel Jordan; David Rientjes; Jerome Glisse; Ganesh Mahendran; Minchan Kim; Punit Agrawal; vinayak menon; Yang Shi
> >>>> Cc: linux-kernel@vger.kernel.org; linux-mm@kvack.org; haren@linux.vnet.ibm.com; npiggin@gmail.com; bsingharora@gmail.com; paulmck@linux.vnet.ibm.com; Tim Chen; linuxppc-dev@lists.ozlabs.org; x86@kernel.org
> >>>> Subject: [PATCH v11 00/26] Speculative page faults
> >>>>
> >>>> This is a port on kernel 4.17 of the work done by Peter Zijlstra to handle
> >>>> page fault without holding the mm semaphore [1].
> >>>>
> >>>> The idea is to try to handle user space page faults without holding the
> >>>> mmap_sem. This should allow better concurrency for massively threaded
> >>>> process since the page fault handler will not wait for other threads memory
> >>>> layout change to be done, assuming that this change is done in another part
> >>>> of the process's memory space. This type page fault is named speculative
> >>>> page fault. If the speculative page fault fails because of a concurrency is
> >>>> detected or because underlying PMD or PTE tables are not yet allocating, it
> >>>> is failing its processing and a classic page fault is then tried.
> >>>>
> >>>> The speculative page fault (SPF) has to look for the VMA matching the fault
> >>>> address without holding the mmap_sem, this is done by introducing a rwlock
> >>>> which protects the access to the mm_rb tree. Previously this was done using
> >>>> SRCU but it was introducing a lot of scheduling to process the VMA's
> >>>> freeing operation which was hitting the performance by 20% as reported by
> >>>> Kemi Wang [2]. Using a rwlock to protect access to the mm_rb tree is
> >>>> limiting the locking contention to these operations which are expected to
> >>>> be in a O(log n) order. In addition to ensure that the VMA is not freed in
> >>>> our back a reference count is added and 2 services (get_vma() and
> >>>> put_vma()) are introduced to handle the reference count. Once a VMA is
> >>>> fetched from the RB tree using get_vma(), it must be later freed using
> >>>> put_vma(). I can't see anymore the overhead I got while will-it-scale
> >>>> benchmark anymore.
> >>>>
> >>>> The VMA's attributes checked during the speculative page fault processing
> >>>> have to be protected against parallel changes. This is done by using a per
> >>>> VMA sequence lock. This sequence lock allows the speculative page fault
> >>>> handler to fast check for parallel changes in progress and to abort the
> >>>> speculative page fault in that case.
> >>>>
> >>>> Once the VMA has been found, the speculative page fault handler would check
> >>>> for the VMA's attributes to verify that the page fault has to be handled
> >>>> correctly or not. Thus, the VMA is protected through a sequence lock which
> >>>> allows fast detection of concurrent VMA changes. If such a change is
> >>>> detected, the speculative page fault is aborted and a *classic* page fault
> >>>> is tried.  VMA sequence lockings are added when VMA attributes which are
> >>>> checked during the page fault are modified.
> >>>>
> >>>> When the PTE is fetched, the VMA is checked to see if it has been changed,
> >>>> so once the page table is locked, the VMA is valid, so any other changes
> >>>> leading to touching this PTE will need to lock the page table, so no
> >>>> parallel change is possible at this time.
> >>>>
> >>>> The locking of the PTE is done with interrupts disabled, this allows
> >>>> checking for the PMD to ensure that there is not an ongoing collapsing
> >>>> operation. Since khugepaged is firstly set the PMD to pmd_none and then is
> >>>> waiting for the other CPU to have caught the IPI interrupt, if the pmd is
> >>>> valid at the time the PTE is locked, we have the guarantee that the
> >>>> collapsing operation will have to wait on the PTE lock to move forward.
> >>>> This allows the SPF handler to map the PTE safely. If the PMD value is
> >>>> different from the one recorded at the beginning of the SPF operation, the
> >>>> classic page fault handler will be called to handle the operation while
> >>>> holding the mmap_sem. As the PTE lock is done with the interrupts disabled,
> >>>> the lock is done using spin_trylock() to avoid dead lock when handling a
> >>>> page fault while a TLB invalidate is requested by another CPU holding the
> >>>> PTE.
> >>>>
> >>>> In pseudo code, this could be seen as:
> >>>>     speculative_page_fault()
> >>>>     {
> >>>>             vma = get_vma()
> >>>>             check vma sequence count
> >>>>             check vma's support
> >>>>             disable interrupt
> >>>>                   check pgd,p4d,...,pte
> >>>>                   save pmd and pte in vmf
> >>>>                   save vma sequence counter in vmf
> >>>>             enable interrupt
> >>>>             check vma sequence count
> >>>>             handle_pte_fault(vma)
> >>>>                     ..
> >>>>                     page = alloc_page()
> >>>>                     pte_map_lock()
> >>>>                             disable interrupt
> >>>>                                     abort if sequence counter has changed
> >>>>                                     abort if pmd or pte has changed
> >>>>                                     pte map and lock
> >>>>                             enable interrupt
> >>>>                     if abort
> >>>>                        free page
> >>>>                        abort
> >>>>                     ...
> >>>>     }
> >>>>
> >>>>     arch_fault_handler()
> >>>>     {
> >>>>             if (speculative_page_fault(&vma))
> >>>>                goto done
> >>>>     again:
> >>>>             lock(mmap_sem)
> >>>>             vma = find_vma();
> >>>>             handle_pte_fault(vma);
> >>>>             if retry
> >>>>                unlock(mmap_sem)
> >>>>                goto again;
> >>>>     done:
> >>>>             handle fault error
> >>>>     }
> >>>>
> >>>> Support for THP is not done because when checking for the PMD, we can be
> >>>> confused by an in progress collapsing operation done by khugepaged. The
> >>>> issue is that pmd_none() could be true either if the PMD is not already
> >>>> populated or if the underlying PTE are in the way to be collapsed. So we
> >>>> cannot safely allocate a PMD if pmd_none() is true.
> >>>>
> >>>> This series add a new software performance event named 'speculative-faults'
> >>>> or 'spf'. It counts the number of successful page fault event handled
> >>>> speculatively. When recording 'faults,spf' events, the faults one is
> >>>> counting the total number of page fault events while 'spf' is only counting
> >>>> the part of the faults processed speculatively.
> >>>>
> >>>> There are some trace events introduced by this series. They allow
> >>>> identifying why the page faults were not processed speculatively. This
> >>>> doesn't take in account the faults generated by a monothreaded process
> >>>> which directly processed while holding the mmap_sem. This trace events are
> >>>> grouped in a system named 'pagefault', they are:
> >>>>  - pagefault:spf_vma_changed : if the VMA has been changed in our back
> >>>>  - pagefault:spf_vma_noanon : the vma->anon_vma field was not yet set.
> >>>>  - pagefault:spf_vma_notsup : the VMA's type is not supported
> >>>>  - pagefault:spf_vma_access : the VMA's access right are not respected
> >>>>  - pagefault:spf_pmd_changed : the upper PMD pointer has changed in our
> >>>>    back.
> >>>>
> >>>> To record all the related events, the easier is to run perf with the
> >>>> following arguments :
> >>>> $ perf stat -e 'faults,spf,pagefault:*' <command>
> >>>>
> >>>> There is also a dedicated vmstat counter showing the number of successful
> >>>> page fault handled speculatively. I can be seen this way:
> >>>> $ grep speculative_pgfault /proc/vmstat
> >>>>
> >>>> This series builds on top of v4.16-mmotm-2018-04-13-17-28 and is functional
> >>>> on x86, PowerPC and arm64.
> >>>>
> >>>> ---------------------
> >>>> Real Workload results
> >>>>
> >>>> As mentioned in previous email, we did non official runs using a "popular
> >>>> in memory multithreaded database product" on 176 cores SMT8 Power system
> >>>> which showed a 30% improvements in the number of transaction processed per
> >>>> second. This run has been done on the v6 series, but changes introduced in
> >>>> this new version should not impact the performance boost seen.
> >>>>
> >>>> Here are the perf data captured during 2 of these runs on top of the v8
> >>>> series:
> >>>>                 vanilla         spf
> >>>> faults          89.418          101.364         +13%
> >>>> spf                n/a           97.989
> >>>>
> >>>> With the SPF kernel, most of the page fault were processed in a speculative
> >>>> way.
> >>>>
> >>>> Ganesh Mahendran had backported the series on top of a 4.9 kernel and gave
> >>>> it a try on an android device. He reported that the application launch time
> >>>> was improved in average by 6%, and for large applications (~100 threads) by
> >>>> 20%.
> >>>>
> >>>> Here are the launch time Ganesh mesured on Android 8.0 on top of a Qcom
> >>>> MSM845 (8 cores) with 6GB (the less is better):
> >>>>
> >>>> Application                             4.9     4.9+spf delta
> >>>> com.tencent.mm                          416     389     -7%
> >>>> com.eg.android.AlipayGphone             1135    986     -13%
> >>>> com.tencent.mtt                         455     454     0%
> >>>> com.qqgame.hlddz                        1497    1409    -6%
> >>>> com.autonavi.minimap                    711     701     -1%
> >>>> com.tencent.tmgp.sgame                  788     748     -5%
> >>>> com.immomo.momo                         501     487     -3%
> >>>> com.tencent.peng                        2145    2112    -2%
> >>>> com.smile.gifmaker                      491     461     -6%
> >>>> com.baidu.BaiduMap                      479     366     -23%
> >>>> com.taobao.taobao                       1341    1198    -11%
> >>>> com.baidu.searchbox                     333     314     -6%
> >>>> com.tencent.mobileqq                    394     384     -3%
> >>>> com.sina.weibo                          907     906     0%
> >>>> com.youku.phone                         816     731     -11%
> >>>> com.happyelements.AndroidAnimal.qq      763     717     -6%
> >>>> com.UCMobile                            415     411     -1%
> >>>> com.tencent.tmgp.ak                     1464    1431    -2%
> >>>> com.tencent.qqmusic                     336     329     -2%
> >>>> com.sankuai.meituan                     1661    1302    -22%
> >>>> com.netease.cloudmusic                  1193    1200    1%
> >>>> air.tv.douyu.android                    4257    4152    -2%
> >>>>
> >>>> ------------------
> >>>> Benchmarks results
> >>>>
> >>>> Base kernel is v4.17.0-rc4-mm1
> >>>> SPF is BASE + this series
> >>>>
> >>>> Kernbench:
> >>>> ----------
> >>>> Here are the results on a 16 CPUs X86 guest using kernbench on a 4.15
> >>>> kernel (kernel is build 5 times):
> >>>>
> >>>> Average Half load -j 8
> >>>>                  Run    (std deviation)
> >>>>                  BASE                   SPF
> >>>> Elapsed Time     1448.65 (5.72312)      1455.84 (4.84951)       0.50%
> >>>> User    Time     10135.4 (30.3699)      10148.8 (31.1252)       0.13%
> >>>> System  Time     900.47  (2.81131)      923.28  (7.52779)       2.53%
> >>>> Percent CPU      761.4   (1.14018)      760.2   (0.447214)      -0.16%
> >>>> Context Switches 85380   (3419.52)      84748   (1904.44)       -0.74%
> >>>> Sleeps           105064  (1240.96)      105074  (337.612)       0.01%
> >>>>
> >>>> Average Optimal load -j 16
> >>>>                  Run    (std deviation)
> >>>>                  BASE                   SPF
> >>>> Elapsed Time     920.528 (10.1212)      927.404 (8.91789)       0.75%
> >>>> User    Time     11064.8 (981.142)      11085   (990.897)       0.18%
> >>>> System  Time     979.904 (84.0615)      1001.14 (82.5523)       2.17%
> >>>> Percent CPU      1089.5  (345.894)      1086.1  (343.545)       -0.31%
> >>>> Context Switches 159488  (78156.4)      158223  (77472.1)       -0.79%
> >>>> Sleeps           110566  (5877.49)      110388  (5617.75)       -0.16%
> >>>>
> >>>>
> >>>> During a run on the SPF, perf events were captured:
> >>>>  Performance counter stats for '../kernbench -M':
> >>>>          526743764      faults
> >>>>                210      spf
> >>>>                  3      pagefault:spf_vma_changed
> >>>>                  0      pagefault:spf_vma_noanon
> >>>>               2278      pagefault:spf_vma_notsup
> >>>>                  0      pagefault:spf_vma_access
> >>>>                  0      pagefault:spf_pmd_changed
> >>>>
> >>>> Very few speculative page faults were recorded as most of the processes
> >>>> involved are monothreaded (sounds that on this architecture some threads
> >>>> were created during the kernel build processing).
> >>>>
> >>>> Here are the kerbench results on a 80 CPUs Power8 system:
> >>>>
> >>>> Average Half load -j 40
> >>>>                  Run    (std deviation)
> >>>>                  BASE                   SPF
> >>>> Elapsed Time     117.152 (0.774642)     117.166 (0.476057)      0.01%
> >>>> User    Time     4478.52 (24.7688)      4479.76 (9.08555)       0.03%
> >>>> System  Time     131.104 (0.720056)     134.04  (0.708414)      2.24%
> >>>> Percent CPU      3934    (19.7104)      3937.2  (19.0184)       0.08%
> >>>> Context Switches 92125.4 (576.787)      92581.6 (198.622)       0.50%
> >>>> Sleeps           317923  (652.499)      318469  (1255.59)       0.17%
> >>>>
> >>>> Average Optimal load -j 80
> >>>>                  Run    (std deviation)
> >>>>                  BASE                   SPF
> >>>> Elapsed Time     107.73  (0.632416)     107.31  (0.584936)      -0.39%
> >>>> User    Time     5869.86 (1466.72)      5871.71 (1467.27)       0.03%
> >>>> System  Time     153.728 (23.8573)      157.153 (24.3704)       2.23%
> >>>> Percent CPU      5418.6  (1565.17)      5436.7  (1580.91)       0.33%
> >>>> Context Switches 223861  (138865)       225032  (139632)        0.52%
> >>>> Sleeps           330529  (13495.1)      332001  (14746.2)       0.45%
> >>>>
> >>>> During a run on the SPF, perf events were captured:
> >>>>  Performance counter stats for '../kernbench -M':
> >>>>          116730856      faults
> >>>>                  0      spf
> >>>>                  3      pagefault:spf_vma_changed
> >>>>                  0      pagefault:spf_vma_noanon
> >>>>                476      pagefault:spf_vma_notsup
> >>>>                  0      pagefault:spf_vma_access
> >>>>                  0      pagefault:spf_pmd_changed
> >>>>
> >>>> Most of the processes involved are monothreaded so SPF is not activated but
> >>>> there is no impact on the performance.
> >>>>
> >>>> Ebizzy:
> >>>> -------
> >>>> The test is counting the number of records per second it can manage, the
> >>>> higher is the best. I run it like this 'ebizzy -mTt <nrcpus>'. To get
> >>>> consistent result I repeated the test 100 times and measure the average
> >>>> result. The number is the record processes per second, the higher is the
> >>>> best.
> >>>>
> >>>>                 BASE            SPF             delta
> >>>> 16 CPUs x86 VM  742.57          1490.24         100.69%
> >>>> 80 CPUs P8 node 13105.4         24174.23        84.46%
> >>>>
> >>>> Here are the performance counter read during a run on a 16 CPUs x86 VM:
> >>>>  Performance counter stats for './ebizzy -mTt 16':
> >>>>            1706379      faults
> >>>>            1674599      spf
> >>>>              30588      pagefault:spf_vma_changed
> >>>>                  0      pagefault:spf_vma_noanon
> >>>>                363      pagefault:spf_vma_notsup
> >>>>                  0      pagefault:spf_vma_access
> >>>>                  0      pagefault:spf_pmd_changed
> >>>>
> >>>> And the ones captured during a run on a 80 CPUs Power node:
> >>>>  Performance counter stats for './ebizzy -mTt 80':
> >>>>            1874773      faults
> >>>>            1461153      spf
> >>>>             413293      pagefault:spf_vma_changed
> >>>>                  0      pagefault:spf_vma_noanon
> >>>>                200      pagefault:spf_vma_notsup
> >>>>                  0      pagefault:spf_vma_access
> >>>>                  0      pagefault:spf_pmd_changed
> >>>>
> >>>> In ebizzy's case most of the page fault were handled in a speculative way,
> >>>> leading the ebizzy performance boost.
> >>>>
> >>>> ------------------
> >>>> Changes since v10 (https://lkml.org/lkml/2018/4/17/572):
> >>>>  - Accounted for all review feedbacks from Punit Agrawal, Ganesh Mahendran
> >>>>    and Minchan Kim, hopefully.
> >>>>  - Remove unneeded check on CONFIG_SPECULATIVE_PAGE_FAULT in
> >>>>    __do_page_fault().
> >>>>  - Loop in pte_spinlock() and pte_map_lock() when pte try lock fails
> >>>>    instead
> >>>>    of aborting the speculative page fault handling. Dropping the now
> >>>> useless
> >>>>    trace event pagefault:spf_pte_lock.
> >>>>  - No more try to reuse the fetched VMA during the speculative page fault
> >>>>    handling when retrying is needed. This adds a lot of complexity and
> >>>>    additional tests done didn't show a significant performance improvement.
> >>>>  - Convert IS_ENABLED(CONFIG_NUMA) back to #ifdef due to build error.
> >>>>
> >>>> [1] http://linux-kernel.2935.n7.nabble.com/RFC-PATCH-0-6-Another-go-at-speculative-page-faults-tt965642.html#none
> >>>> [2] https://patchwork.kernel.org/patch/9999687/
> >>>>
> >>>>
> >>>> Laurent Dufour (20):
> >>>>   mm: introduce CONFIG_SPECULATIVE_PAGE_FAULT
> >>>>   x86/mm: define ARCH_SUPPORTS_SPECULATIVE_PAGE_FAULT
> >>>>   powerpc/mm: set ARCH_SUPPORTS_SPECULATIVE_PAGE_FAULT
> >>>>   mm: introduce pte_spinlock for FAULT_FLAG_SPECULATIVE
> >>>>   mm: make pte_unmap_same compatible with SPF
> >>>>   mm: introduce INIT_VMA()
> >>>>   mm: protect VMA modifications using VMA sequence count
> >>>>   mm: protect mremap() against SPF hanlder
> >>>>   mm: protect SPF handler against anon_vma changes
> >>>>   mm: cache some VMA fields in the vm_fault structure
> >>>>   mm/migrate: Pass vm_fault pointer to migrate_misplaced_page()
> >>>>   mm: introduce __lru_cache_add_active_or_unevictable
> >>>>   mm: introduce __vm_normal_page()
> >>>>   mm: introduce __page_add_new_anon_rmap()
> >>>>   mm: protect mm_rb tree with a rwlock
> >>>>   mm: adding speculative page fault failure trace events
> >>>>   perf: add a speculative page fault sw event
> >>>>   perf tools: add support for the SPF perf event
> >>>>   mm: add speculative page fault vmstats
> >>>>   powerpc/mm: add speculative page fault
> >>>>
> >>>> Mahendran Ganesh (2):
> >>>>   arm64/mm: define ARCH_SUPPORTS_SPECULATIVE_PAGE_FAULT
> >>>>   arm64/mm: add speculative page fault
> >>>>
> >>>> Peter Zijlstra (4):
> >>>>   mm: prepare for FAULT_FLAG_SPECULATIVE
> >>>>   mm: VMA sequence count
> >>>>   mm: provide speculative fault infrastructure
> >>>>   x86/mm: add speculative pagefault handling
> >>>>
> >>>>  arch/arm64/Kconfig                    |   1 +
> >>>>  arch/arm64/mm/fault.c                 |  12 +
> >>>>  arch/powerpc/Kconfig                  |   1 +
> >>>>  arch/powerpc/mm/fault.c               |  16 +
> >>>>  arch/x86/Kconfig                      |   1 +
> >>>>  arch/x86/mm/fault.c                   |  27 +-
> >>>>  fs/exec.c                             |   2 +-
> >>>>  fs/proc/task_mmu.c                    |   5 +-
> >>>>  fs/userfaultfd.c                      |  17 +-
> >>>>  include/linux/hugetlb_inline.h        |   2 +-
> >>>>  include/linux/migrate.h               |   4 +-
> >>>>  include/linux/mm.h                    | 136 +++++++-
> >>>>  include/linux/mm_types.h              |   7 +
> >>>>  include/linux/pagemap.h               |   4 +-
> >>>>  include/linux/rmap.h                  |  12 +-
> >>>>  include/linux/swap.h                  |  10 +-
> >>>>  include/linux/vm_event_item.h         |   3 +
> >>>>  include/trace/events/pagefault.h      |  80 +++++
> >>>>  include/uapi/linux/perf_event.h       |   1 +
> >>>>  kernel/fork.c                         |   5 +-
> >>>>  mm/Kconfig                            |  22 ++
> >>>>  mm/huge_memory.c                      |   6 +-
> >>>>  mm/hugetlb.c                          |   2 +
> >>>>  mm/init-mm.c                          |   3 +
> >>>>  mm/internal.h                         |  20 ++
> >>>>  mm/khugepaged.c                       |   5 +
> >>>>  mm/madvise.c                          |   6 +-
> >>>>  mm/memory.c                           | 612 +++++++++++++++++++++++++++++-----
> >>>>  mm/mempolicy.c                        |  51 ++-
> >>>>  mm/migrate.c                          |   6 +-
> >>>>  mm/mlock.c                            |  13 +-
> >>>>  mm/mmap.c                             | 229 ++++++++++---
> >>>>  mm/mprotect.c                         |   4 +-
> >>>>  mm/mremap.c                           |  13 +
> >>>>  mm/nommu.c                            |   2 +-
> >>>>  mm/rmap.c                             |   5 +-
> >>>>  mm/swap.c                             |   6 +-
> >>>>  mm/swap_state.c                       |   8 +-
> >>>>  mm/vmstat.c                           |   5 +-
> >>>>  tools/include/uapi/linux/perf_event.h |   1 +
> >>>>  tools/perf/util/evsel.c               |   1 +
> >>>>  tools/perf/util/parse-events.c        |   4 +
> >>>>  tools/perf/util/parse-events.l        |   1 +
> >>>>  tools/perf/util/python.c              |   1 +
> >>>>  44 files changed, 1161 insertions(+), 211 deletions(-)
> >>>>  create mode 100644 include/trace/events/pagefault.h
> >>>>
> >>>> --
> >>>> 2.7.4
> >>>>
> >>>>
> >>>
> >>
> > 
> 

[-- Attachment #2: compare-result.txt --]
[-- Type: text/plain, Size: 183477 bytes --]

=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/nr_task/thp_enabled/test/cpufreq_governor:
  lkp-skl-4sp1/will-it-scale/debian-x86_64-2018-04-03.cgz/x86_64-rhel-7.2/gcc-7/100%/always/page_fault3/performance

commit: 
  ba98a1cdad71d259a194461b3a61471b49b14df1
  a7a8993bfe3ccb54ad468b9f1799649e4ad1ff12

ba98a1cdad71d259 a7a8993bfe3ccb54ad468b9f17 
---------------- -------------------------- 
       fail:runs  %reproduction    fail:runs
           |             |             |    
         44:3          -13%          43:3     perf-profile.calltrace.cycles-pp.error_entry
         22:3           -6%          22:3     perf-profile.calltrace.cycles-pp.sync_regs.error_entry
         44:3          -13%          44:3     perf-profile.children.cycles-pp.error_entry
         21:3           -7%          21:3     perf-profile.self.cycles-pp.error_entry
         %stddev     %change         %stddev
             \          |                \  
     10519 ±  3%     -20.5%       8368 ±  6%  will-it-scale.per_thread_ops
    118098           +11.2%     131287 ±  2%  will-it-scale.time.involuntary_context_switches
 6.084e+08 ±  3%     -20.4%  4.845e+08 ±  6%  will-it-scale.time.minor_page_faults
      7467            +5.0%       7841        will-it-scale.time.percent_of_cpu_this_job_got
     44922            +5.0%      47176        will-it-scale.time.system_time
   7126337 ±  3%     -15.4%    6025689 ±  6%  will-it-scale.time.voluntary_context_switches
  91905646            -1.3%   90673935        will-it-scale.workload
     27.15 ±  6%      -8.7%      24.80 ± 10%  boot-time.boot
   2516213 ±  6%      +8.3%    2726303        interrupts.CAL:Function_call_interrupts
    388.00 ±  9%     +60.2%     621.67 ± 20%  irq_exception_noise.softirq_nr
     11.28 ±  2%      -1.9        9.37 ±  4%  mpstat.cpu.idle%
     10065 ±140%    +243.4%      34559 ±  4%  numa-numastat.node0.other_node
     18739           -11.6%      16573 ±  3%  uptime.idle
     29406 ±  2%     -11.8%      25929 ±  5%  vmstat.system.cs
    329614 ±  8%     +17.0%     385618 ± 10%  meminfo.DirectMap4k
    237851           +21.2%     288160 ±  5%  meminfo.Inactive
    237615           +21.2%     287924 ±  5%  meminfo.Inactive(anon)
   7917847           -10.7%    7071860        softirqs.RCU
   4784181 ±  3%     -14.5%    4089039 ±  4%  softirqs.SCHED
  45666107 ±  7%     +12.9%   51535472 ±  3%  softirqs.TIMER
 2.617e+09 ±  2%     -13.9%  2.253e+09 ±  6%  cpuidle.C1E.time
   6688774 ±  2%     -12.8%    5835101 ±  5%  cpuidle.C1E.usage
 1.022e+10 ±  2%     -18.0%  8.376e+09 ±  3%  cpuidle.C6.time
  13440993 ±  2%     -16.3%   11243794 ±  4%  cpuidle.C6.usage
     54781 ± 16%     +37.5%      75347 ± 12%  numa-meminfo.node0.Inactive
     54705 ± 16%     +37.7%      75347 ± 12%  numa-meminfo.node0.Inactive(anon)
     52522           +35.0%      70886 ±  6%  numa-meminfo.node2.Inactive
     52443           +34.7%      70653 ±  6%  numa-meminfo.node2.Inactive(anon)
     31046 ±  6%     +30.3%      40457 ± 11%  numa-meminfo.node2.SReclaimable
     58563           +21.1%      70945 ±  6%  proc-vmstat.nr_inactive_anon
     58564           +21.1%      70947 ±  6%  proc-vmstat.nr_zone_inactive_anon
  69701118            -1.2%   68842151        proc-vmstat.pgalloc_normal
 2.765e+10            -1.3%  2.729e+10        proc-vmstat.pgfault
  69330418            -1.2%   68466824        proc-vmstat.pgfree
    118098           +11.2%     131287 ±  2%  time.involuntary_context_switches
 6.084e+08 ±  3%     -20.4%  4.845e+08 ±  6%  time.minor_page_faults
      7467            +5.0%       7841        time.percent_of_cpu_this_job_got
     44922            +5.0%      47176        time.system_time
   7126337 ±  3%     -15.4%    6025689 ±  6%  time.voluntary_context_switches
     13653 ± 16%     +33.5%      18225 ± 12%  numa-vmstat.node0.nr_inactive_anon
     13651 ± 16%     +33.5%      18224 ± 12%  numa-vmstat.node0.nr_zone_inactive_anon
     13069 ±  3%     +30.1%      17001 ±  4%  numa-vmstat.node2.nr_inactive_anon
    134.67 ± 42%     -49.5%      68.00 ± 31%  numa-vmstat.node2.nr_mlock
      7758 ±  6%     +30.4%      10112 ± 11%  numa-vmstat.node2.nr_slab_reclaimable
     13066 ±  3%     +30.1%      16998 ±  4%  numa-vmstat.node2.nr_zone_inactive_anon
      1039 ± 11%     -17.5%     857.33        slabinfo.Acpi-ParseExt.active_objs
      1039 ± 11%     -17.5%     857.33        slabinfo.Acpi-ParseExt.num_objs
      2566 ±  6%      -8.8%       2340 ±  5%  slabinfo.biovec-64.active_objs
      2566 ±  6%      -8.8%       2340 ±  5%  slabinfo.biovec-64.num_objs
    898.33 ±  3%      -9.5%     813.33 ±  3%  slabinfo.kmem_cache_node.active_objs
      1066 ±  2%      -8.0%     981.33 ±  3%  slabinfo.kmem_cache_node.num_objs
      1940            +2.3%       1984        turbostat.Avg_MHz
   6679037 ±  2%     -12.7%    5830270 ±  5%  turbostat.C1E
      2.25 ±  2%      -0.3        1.94 ±  6%  turbostat.C1E%
  13418115           -16.3%   11234510 ±  4%  turbostat.C6
      8.75 ±  2%      -1.6        7.18 ±  3%  turbostat.C6%
      5.99 ±  2%     -14.4%       5.13 ±  4%  turbostat.CPU%c1
      5.01 ±  3%     -20.1%       4.00 ±  4%  turbostat.CPU%c6
      1.77 ±  3%     -34.7%       1.15        turbostat.Pkg%pc2
 1.378e+13            +1.2%  1.394e+13        perf-stat.branch-instructions
      0.98            -0.0        0.94        perf-stat.branch-miss-rate%
 1.344e+11            -2.3%  1.313e+11        perf-stat.branch-misses
 1.076e+11            -1.8%  1.057e+11        perf-stat.cache-misses
 2.258e+11            -2.1%   2.21e+11        perf-stat.cache-references
  17788064 ±  2%     -11.9%   15674207 ±  6%  perf-stat.context-switches
 2.241e+14            +2.4%  2.294e+14        perf-stat.cpu-cycles
 1.929e+13            +2.2%  1.971e+13        perf-stat.dTLB-loads
      4.01            -0.2        3.83        perf-stat.dTLB-store-miss-rate%
 4.519e+11            -1.3%  4.461e+11        perf-stat.dTLB-store-misses
 1.082e+13            +3.6%  1.121e+13        perf-stat.dTLB-stores
  3.02e+10           +23.2%  3.721e+10 ±  3%  perf-stat.iTLB-load-misses
 2.721e+08 ±  8%      -8.8%  2.481e+08 ±  3%  perf-stat.iTLB-loads
 6.985e+13            +1.8%  7.111e+13        perf-stat.instructions
      2313           -17.2%       1914 ±  3%  perf-stat.instructions-per-iTLB-miss
 2.764e+10            -1.3%  2.729e+10        perf-stat.minor-faults
 1.421e+09 ±  2%     -16.4%  1.188e+09 ±  9%  perf-stat.node-load-misses
 1.538e+10            -9.3%  1.395e+10        perf-stat.node-loads
      9.75            +1.4       11.10        perf-stat.node-store-miss-rate%
 3.012e+09           +14.1%  3.437e+09        perf-stat.node-store-misses
 2.789e+10            -1.3%  2.753e+10        perf-stat.node-stores
 2.764e+10            -1.3%  2.729e+10        perf-stat.page-faults
    760059            +3.2%     784235        perf-stat.path-length
    193545 ± 25%     -57.8%      81757 ± 46%  sched_debug.cfs_rq:/.MIN_vruntime.avg
  26516863 ± 19%     -49.7%   13338070 ± 33%  sched_debug.cfs_rq:/.MIN_vruntime.max
   2202271 ± 21%     -53.2%    1029581 ± 38%  sched_debug.cfs_rq:/.MIN_vruntime.stddev
    193545 ± 25%     -57.8%      81757 ± 46%  sched_debug.cfs_rq:/.max_vruntime.avg
  26516863 ± 19%     -49.7%   13338070 ± 33%  sched_debug.cfs_rq:/.max_vruntime.max
   2202271 ± 21%     -53.2%    1029581 ± 38%  sched_debug.cfs_rq:/.max_vruntime.stddev
      0.32 ± 70%    +253.2%       1.14 ± 54%  sched_debug.cfs_rq:/.removed.load_avg.avg
      4.44 ± 70%    +120.7%       9.80 ± 27%  sched_debug.cfs_rq:/.removed.load_avg.stddev
     14.90 ± 70%    +251.0%      52.31 ± 53%  sched_debug.cfs_rq:/.removed.runnable_sum.avg
    205.71 ± 70%    +119.5%     451.60 ± 27%  sched_debug.cfs_rq:/.removed.runnable_sum.stddev
      0.16 ± 70%    +237.9%       0.54 ± 50%  sched_debug.cfs_rq:/.removed.util_avg.avg
      2.23 ± 70%    +114.2%       4.77 ± 24%  sched_debug.cfs_rq:/.removed.util_avg.stddev
    573.70 ±  5%      -9.7%     518.06 ±  6%  sched_debug.cfs_rq:/.util_avg.min
    114.87 ±  8%     +14.1%     131.04 ± 10%  sched_debug.cfs_rq:/.util_est_enqueued.avg
     64.42 ± 54%     -63.9%      23.27 ± 68%  sched_debug.cpu.cpu_load[1].max
      5.05 ± 48%     -55.2%       2.26 ± 51%  sched_debug.cpu.cpu_load[1].stddev
     57.58 ± 59%     -60.3%      22.88 ± 70%  sched_debug.cpu.cpu_load[2].max
     21019 ±  3%     -15.1%      17841 ±  6%  sched_debug.cpu.nr_switches.min
     20797 ±  3%     -15.0%      17670 ±  6%  sched_debug.cpu.sched_count.min
     10287 ±  3%     -15.1%       8736 ±  6%  sched_debug.cpu.sched_goidle.avg
     13693 ±  2%     -10.7%      12233 ±  5%  sched_debug.cpu.sched_goidle.max
      9976 ±  3%     -16.0%       8381 ±  7%  sched_debug.cpu.sched_goidle.min
      0.00 ± 26%     +98.9%       0.00 ± 28%  sched_debug.rt_rq:/.rt_time.min
      4230 ±141%    -100.0%       0.00        latency_stats.avg.trace_module_notify.notifier_call_chain.blocking_notifier_call_chain.do_init_module.load_module.__do_sys_finit_module.do_syscall_64.entry_SYSCALL_64_after_hwframe
     28498 ±141%    -100.0%       0.00        latency_stats.avg.perf_event_alloc.__do_sys_perf_event_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
      4065 ±138%     -92.2%     315.33 ± 91%  latency_stats.avg.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_lookup.nfs_lookup_revalidate.lookup_fast.walk_component.link_path_walk.path_lookupat.filename_lookup
      0.00       +3.6e+105%       3641 ±141%  latency_stats.avg.down.console_lock.console_device.tty_lookup_driver.tty_open.chrdev_open.do_dentry_open.path_openat.do_filp_open.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00       +2.5e+106%      25040 ±141%  latency_stats.avg.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_lookup.nfs_lookup.path_openat.do_filp_open.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00       +3.4e+106%      34015 ±141%  latency_stats.avg.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_get_acl.get_acl.posix_acl_create.nfs3_proc_create.nfs_create.path_openat.do_filp_open.do_sys_open
      0.00       +4.8e+106%      47686 ±141%  latency_stats.avg.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_do_create.nfs3_proc_create.nfs_create.path_openat.do_filp_open.do_sys_open.do_syscall_64
      4230 ±141%    -100.0%       0.00        latency_stats.max.trace_module_notify.notifier_call_chain.blocking_notifier_call_chain.do_init_module.load_module.__do_sys_finit_module.do_syscall_64.entry_SYSCALL_64_after_hwframe
     28498 ±141%    -100.0%       0.00        latency_stats.max.perf_event_alloc.__do_sys_perf_event_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
      4065 ±138%     -92.2%     315.33 ± 91%  latency_stats.max.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_lookup.nfs_lookup_revalidate.lookup_fast.walk_component.link_path_walk.path_lookupat.filename_lookup
      4254 ±134%     -88.0%     511.67 ± 90%  latency_stats.max.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_getattr.__nfs_revalidate_inode.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_lookupat
     43093 ± 35%     +76.6%      76099 ±115%  latency_stats.max.blk_execute_rq.scsi_execute.ioctl_internal_command.scsi_set_medium_removal.cdrom_release.[cdrom].sr_block_release.[sr_mod].__blkdev_put.blkdev_close.__fput.task_work_run.exit_to_usermode_loop.do_syscall_64
     24139 ± 70%    +228.5%      79285 ±105%  latency_stats.max.blk_execute_rq.scsi_execute.scsi_test_unit_ready.sr_check_events.[sr_mod].cdrom_check_events.[cdrom].sr_block_check_events.[sr_mod].disk_check_events.disk_clear_events.check_disk_change.sr_block_open.[sr_mod].__blkdev_get.blkdev_get
      0.00       +3.6e+105%       3641 ±141%  latency_stats.max.down.console_lock.console_device.tty_lookup_driver.tty_open.chrdev_open.do_dentry_open.path_openat.do_filp_open.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00       +2.5e+106%      25040 ±141%  latency_stats.max.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_lookup.nfs_lookup.path_openat.do_filp_open.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00       +3.4e+106%      34015 ±141%  latency_stats.max.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_get_acl.get_acl.posix_acl_create.nfs3_proc_create.nfs_create.path_openat.do_filp_open.do_sys_open
      0.00       +6.5e+106%      64518 ±141%  latency_stats.max.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_do_create.nfs3_proc_create.nfs_create.path_openat.do_filp_open.do_sys_open.do_syscall_64
      4230 ±141%    -100.0%       0.00        latency_stats.sum.trace_module_notify.notifier_call_chain.blocking_notifier_call_chain.do_init_module.load_module.__do_sys_finit_module.do_syscall_64.entry_SYSCALL_64_after_hwframe
     28498 ±141%    -100.0%       0.00        latency_stats.sum.perf_event_alloc.__do_sys_perf_event_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
      4065 ±138%     -92.2%     315.33 ± 91%  latency_stats.sum.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_lookup.nfs_lookup_revalidate.lookup_fast.walk_component.link_path_walk.path_lookupat.filename_lookup
     57884 ±  9%     +47.3%      85264 ±118%  latency_stats.sum.blk_execute_rq.scsi_execute.ioctl_internal_command.scsi_set_medium_removal.cdrom_release.[cdrom].sr_block_release.[sr_mod].__blkdev_put.blkdev_close.__fput.task_work_run.exit_to_usermode_loop.do_syscall_64
      0.00       +3.6e+105%       3641 ±141%  latency_stats.sum.down.console_lock.console_device.tty_lookup_driver.tty_open.chrdev_open.do_dentry_open.path_openat.do_filp_open.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00       +2.5e+106%      25040 ±141%  latency_stats.sum.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_lookup.nfs_lookup.path_openat.do_filp_open.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00       +3.4e+106%      34015 ±141%  latency_stats.sum.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_get_acl.get_acl.posix_acl_create.nfs3_proc_create.nfs_create.path_openat.do_filp_open.do_sys_open
      0.00       +9.5e+106%      95373 ±141%  latency_stats.sum.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_do_create.nfs3_proc_create.nfs_create.path_openat.do_filp_open.do_sys_open.do_syscall_64
     11.70           -11.7        0.00        perf-profile.calltrace.cycles-pp.__do_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault.do_page_fault
     11.52           -11.5        0.00        perf-profile.calltrace.cycles-pp.shmem_fault.__do_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
     10.44           -10.4        0.00        perf-profile.calltrace.cycles-pp.shmem_getpage_gfp.shmem_fault.__do_fault.__handle_mm_fault.handle_mm_fault
      9.83            -9.8        0.00        perf-profile.calltrace.cycles-pp.find_lock_entry.shmem_getpage_gfp.shmem_fault.__do_fault.__handle_mm_fault
      9.55            -9.5        0.00        perf-profile.calltrace.cycles-pp.finish_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault.do_page_fault
      9.35            -9.3        0.00        perf-profile.calltrace.cycles-pp.alloc_set_pte.finish_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
      6.81            -6.8        0.00        perf-profile.calltrace.cycles-pp.page_add_file_rmap.alloc_set_pte.finish_fault.__handle_mm_fault.handle_mm_fault
      7.71            -0.3        7.45        perf-profile.calltrace.cycles-pp.find_get_entry.find_lock_entry.shmem_getpage_gfp.shmem_fault.__do_fault
      0.59 ±  7%      -0.2        0.35 ± 70%  perf-profile.calltrace.cycles-pp.smp_apic_timer_interrupt.apic_timer_interrupt.__do_page_fault.do_page_fault.page_fault
      0.59 ±  7%      -0.2        0.35 ± 70%  perf-profile.calltrace.cycles-pp.apic_timer_interrupt.__do_page_fault.do_page_fault.page_fault
     10.41            -0.2       10.24        perf-profile.calltrace.cycles-pp.native_irq_return_iret
      7.68            -0.1        7.60        perf-profile.calltrace.cycles-pp.swapgs_restore_regs_and_return_to_usermode
      0.76            -0.1        0.70        perf-profile.calltrace.cycles-pp.down_read_trylock.__do_page_fault.do_page_fault.page_fault
      1.38            -0.0        1.34        perf-profile.calltrace.cycles-pp.do_page_fault
      1.05            -0.0        1.02        perf-profile.calltrace.cycles-pp.trace_graph_entry.do_page_fault
      0.92            +0.0        0.94        perf-profile.calltrace.cycles-pp.find_vma.__do_page_fault.do_page_fault.page_fault
      0.91            +0.0        0.93        perf-profile.calltrace.cycles-pp.vmacache_find.find_vma.__do_page_fault.do_page_fault.page_fault
      0.65            +0.0        0.67        perf-profile.calltrace.cycles-pp.set_page_dirty.unmap_page_range.unmap_vmas.unmap_region.do_munmap
      0.62            +0.0        0.66        perf-profile.calltrace.cycles-pp.page_mapping.find_lock_entry.shmem_getpage_gfp.shmem_fault.__do_fault
      4.15            +0.1        4.27        perf-profile.calltrace.cycles-pp.page_remove_rmap.unmap_page_range.unmap_vmas.unmap_region.do_munmap
     10.17            +0.2       10.39        perf-profile.calltrace.cycles-pp.munmap
      9.56            +0.2        9.78        perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.munmap
      9.56            +0.2        9.78        perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.munmap
      9.56            +0.2        9.78        perf-profile.calltrace.cycles-pp.unmap_region.do_munmap.vm_munmap.__x64_sys_munmap.do_syscall_64
      9.54            +0.2        9.76        perf-profile.calltrace.cycles-pp.unmap_page_range.unmap_vmas.unmap_region.do_munmap.vm_munmap
      9.54            +0.2        9.76        perf-profile.calltrace.cycles-pp.unmap_vmas.unmap_region.do_munmap.vm_munmap.__x64_sys_munmap
      9.56            +0.2        9.78        perf-profile.calltrace.cycles-pp.do_munmap.vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe
      9.56            +0.2        9.78        perf-profile.calltrace.cycles-pp.vm_munmap.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.munmap
      9.56            +0.2        9.78        perf-profile.calltrace.cycles-pp.__x64_sys_munmap.do_syscall_64.entry_SYSCALL_64_after_hwframe.munmap
      0.00            +0.6        0.56 ±  2%  perf-profile.calltrace.cycles-pp.lock_page_memcg.page_add_file_rmap.alloc_set_pte.finish_fault.handle_pte_fault
      0.00            +0.6        0.59        perf-profile.calltrace.cycles-pp.page_mapping.set_page_dirty.fault_dirty_shared_page.handle_pte_fault.__handle_mm_fault
      0.00            +0.6        0.60        perf-profile.calltrace.cycles-pp.current_time.file_update_time.handle_pte_fault.__handle_mm_fault.handle_mm_fault
      0.00            +0.7        0.68        perf-profile.calltrace.cycles-pp.___might_sleep.find_lock_entry.shmem_getpage_gfp.shmem_fault.__do_fault
      0.00            +0.7        0.74        perf-profile.calltrace.cycles-pp.unlock_page.fault_dirty_shared_page.handle_pte_fault.__handle_mm_fault.handle_mm_fault
      0.00            +0.8        0.80        perf-profile.calltrace.cycles-pp.set_page_dirty.fault_dirty_shared_page.handle_pte_fault.__handle_mm_fault.handle_mm_fault
      0.00            +0.9        0.88        perf-profile.calltrace.cycles-pp._raw_spin_lock.pte_map_lock.alloc_set_pte.finish_fault.handle_pte_fault
      0.00            +0.9        0.91        perf-profile.calltrace.cycles-pp.__set_page_dirty_no_writeback.fault_dirty_shared_page.handle_pte_fault.__handle_mm_fault.handle_mm_fault
      0.00            +1.3        1.27        perf-profile.calltrace.cycles-pp.pte_map_lock.alloc_set_pte.finish_fault.handle_pte_fault.__handle_mm_fault
      0.00            +1.3        1.30        perf-profile.calltrace.cycles-pp.file_update_time.handle_pte_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
      0.00            +2.8        2.76        perf-profile.calltrace.cycles-pp.fault_dirty_shared_page.handle_pte_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
      0.00            +6.8        6.81        perf-profile.calltrace.cycles-pp.page_add_file_rmap.alloc_set_pte.finish_fault.handle_pte_fault.__handle_mm_fault
      0.00            +9.4        9.39        perf-profile.calltrace.cycles-pp.alloc_set_pte.finish_fault.handle_pte_fault.__handle_mm_fault.handle_mm_fault
      0.00            +9.6        9.59        perf-profile.calltrace.cycles-pp.finish_fault.handle_pte_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
      0.00            +9.8        9.77        perf-profile.calltrace.cycles-pp.find_lock_entry.shmem_getpage_gfp.shmem_fault.__do_fault.handle_pte_fault
      0.00           +10.4       10.37        perf-profile.calltrace.cycles-pp.shmem_getpage_gfp.shmem_fault.__do_fault.handle_pte_fault.__handle_mm_fault
      0.00           +11.5       11.46        perf-profile.calltrace.cycles-pp.shmem_fault.__do_fault.handle_pte_fault.__handle_mm_fault.handle_mm_fault
      0.00           +11.6       11.60        perf-profile.calltrace.cycles-pp.__do_fault.handle_pte_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
      0.00           +26.6       26.62        perf-profile.calltrace.cycles-pp.handle_pte_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault.do_page_fault
      7.88            -0.3        7.61        perf-profile.children.cycles-pp.find_get_entry
      1.34 ±  8%      -0.2        1.16 ±  2%  perf-profile.children.cycles-pp.hrtimer_interrupt
     10.41            -0.2       10.24        perf-profile.children.cycles-pp.native_irq_return_iret
      0.38 ± 28%      -0.1        0.26 ±  4%  perf-profile.children.cycles-pp.tick_sched_timer
     11.80            -0.1       11.68        perf-profile.children.cycles-pp.__do_fault
      0.55 ± 15%      -0.1        0.43 ±  2%  perf-profile.children.cycles-pp.__hrtimer_run_queues
      0.60            -0.1        0.51        perf-profile.children.cycles-pp.pmd_devmap_trans_unstable
      0.38 ± 13%      -0.1        0.29 ±  4%  perf-profile.children.cycles-pp.ktime_get
      7.68            -0.1        7.60        perf-profile.children.cycles-pp.swapgs_restore_regs_and_return_to_usermode
      5.18            -0.1        5.12        perf-profile.children.cycles-pp.trace_graph_entry
      0.79            -0.1        0.73        perf-profile.children.cycles-pp.down_read_trylock
      7.83            -0.1        7.76        perf-profile.children.cycles-pp.sync_regs
      3.01            -0.1        2.94        perf-profile.children.cycles-pp.fault_dirty_shared_page
      1.02            -0.1        0.96        perf-profile.children.cycles-pp._raw_spin_lock
      4.66            -0.1        4.61        perf-profile.children.cycles-pp.prepare_ftrace_return
      0.37 ±  8%      -0.1        0.32 ±  3%  perf-profile.children.cycles-pp.current_kernel_time64
      5.26            -0.1        5.21        perf-profile.children.cycles-pp.ftrace_graph_caller
      0.66 ±  5%      -0.1        0.61        perf-profile.children.cycles-pp.current_time
      0.18 ±  5%      -0.0        0.15 ±  3%  perf-profile.children.cycles-pp.update_process_times
      0.27            -0.0        0.26        perf-profile.children.cycles-pp._cond_resched
      0.16            -0.0        0.15 ±  3%  perf-profile.children.cycles-pp.rcu_all_qs
      0.94            +0.0        0.95        perf-profile.children.cycles-pp.vmacache_find
      0.48            +0.0        0.50        perf-profile.children.cycles-pp.__mod_node_page_state
      0.17            +0.0        0.19 ±  2%  perf-profile.children.cycles-pp.__unlock_page_memcg
      1.07            +0.0        1.10        perf-profile.children.cycles-pp.find_vma
      0.79 ±  3%      +0.1        0.86 ±  2%  perf-profile.children.cycles-pp.lock_page_memcg
      4.29            +0.1        4.40        perf-profile.children.cycles-pp.page_remove_rmap
      1.39 ±  2%      +0.1        1.52        perf-profile.children.cycles-pp.file_update_time
      0.00            +0.2        0.16        perf-profile.children.cycles-pp.__vm_normal_page
      9.63            +0.2        9.84        perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
      9.63            +0.2        9.84        perf-profile.children.cycles-pp.do_syscall_64
      9.63            +0.2        9.84        perf-profile.children.cycles-pp.unmap_page_range
     10.17            +0.2       10.39        perf-profile.children.cycles-pp.munmap
      9.56            +0.2        9.78        perf-profile.children.cycles-pp.unmap_region
      9.56            +0.2        9.78        perf-profile.children.cycles-pp.do_munmap
      9.56            +0.2        9.78        perf-profile.children.cycles-pp.vm_munmap
      9.56            +0.2        9.78        perf-profile.children.cycles-pp.__x64_sys_munmap
      9.54            +0.2        9.77        perf-profile.children.cycles-pp.unmap_vmas
      1.01            +0.2        1.25        perf-profile.children.cycles-pp.___might_sleep
      0.00            +1.6        1.59        perf-profile.children.cycles-pp.pte_map_lock
      0.00           +26.9       26.89        perf-profile.children.cycles-pp.handle_pte_fault
      4.25            -1.0        3.24        perf-profile.self.cycles-pp.__handle_mm_fault
      1.42            -0.3        1.11        perf-profile.self.cycles-pp.alloc_set_pte
      4.87            -0.3        4.59        perf-profile.self.cycles-pp.find_get_entry
     10.41            -0.2       10.24        perf-profile.self.cycles-pp.native_irq_return_iret
      0.37 ± 13%      -0.1        0.28 ±  4%  perf-profile.self.cycles-pp.ktime_get
      0.60            -0.1        0.51        perf-profile.self.cycles-pp.pmd_devmap_trans_unstable
      7.50            -0.1        7.42        perf-profile.self.cycles-pp.swapgs_restore_regs_and_return_to_usermode
      7.83            -0.1        7.76        perf-profile.self.cycles-pp.sync_regs
      4.85            -0.1        4.79        perf-profile.self.cycles-pp.trace_graph_entry
      1.01            -0.1        0.95        perf-profile.self.cycles-pp._raw_spin_lock
      0.78            -0.1        0.73        perf-profile.self.cycles-pp.down_read_trylock
      0.36 ±  9%      -0.1        0.31 ±  4%  perf-profile.self.cycles-pp.current_kernel_time64
      0.28            -0.0        0.23 ±  2%  perf-profile.self.cycles-pp.__do_fault
      1.04            -0.0        1.00        perf-profile.self.cycles-pp.find_lock_entry
      0.30            -0.0        0.28 ±  3%  perf-profile.self.cycles-pp.fault_dirty_shared_page
      0.70            -0.0        0.67        perf-profile.self.cycles-pp.prepare_ftrace_return
      0.44            -0.0        0.42        perf-profile.self.cycles-pp.do_page_fault
      0.16            -0.0        0.14        perf-profile.self.cycles-pp.rcu_all_qs
      0.78            -0.0        0.77        perf-profile.self.cycles-pp.shmem_getpage_gfp
      0.20            -0.0        0.19        perf-profile.self.cycles-pp._cond_resched
      0.50            +0.0        0.51        perf-profile.self.cycles-pp.set_page_dirty
      0.93            +0.0        0.95        perf-profile.self.cycles-pp.vmacache_find
      0.36 ±  2%      +0.0        0.38        perf-profile.self.cycles-pp.__might_sleep
      0.47            +0.0        0.50        perf-profile.self.cycles-pp.__mod_node_page_state
      0.17            +0.0        0.19 ±  2%  perf-profile.self.cycles-pp.__unlock_page_memcg
      2.34            +0.0        2.38        perf-profile.self.cycles-pp.unmap_page_range
      0.78 ±  3%      +0.1        0.85 ±  2%  perf-profile.self.cycles-pp.lock_page_memcg
      2.17            +0.1        2.24        perf-profile.self.cycles-pp.__do_page_fault
      0.00            +0.2        0.16 ±  3%  perf-profile.self.cycles-pp.__vm_normal_page
      1.00            +0.2        1.24        perf-profile.self.cycles-pp.___might_sleep
      0.00            +0.7        0.70        perf-profile.self.cycles-pp.pte_map_lock
      0.00            +1.4        1.42 ±  2%  perf-profile.self.cycles-pp.handle_pte_fault

=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/nr_task/thp_enabled/test/cpufreq_governor:
  lkp-skl-4sp1/will-it-scale/debian-x86_64-2018-04-03.cgz/x86_64-rhel-7.2/gcc-7/100%/never/context_switch1/performance

commit: 
  ba98a1cdad71d259a194461b3a61471b49b14df1
  a7a8993bfe3ccb54ad468b9f1799649e4ad1ff12

ba98a1cdad71d259 a7a8993bfe3ccb54ad468b9f17 
---------------- -------------------------- 
       fail:runs  %reproduction    fail:runs
           |             |             |    
           :3           33%           1:3     dmesg.WARNING:at#for_ip_interrupt_entry/0x
          2:3          -67%            :3     kmsg.pstore:crypto_comp_decompress_failed,ret=
          2:3          -67%            :3     kmsg.pstore:decompression_failed
         %stddev     %change         %stddev
             \          |                \  
    224431            -1.3%     221567        will-it-scale.per_process_ops
    237006            -2.2%     231907        will-it-scale.per_thread_ops
 1.601e+09 ± 29%     -46.9%  8.501e+08 ± 12%  will-it-scale.time.involuntary_context_switches
      5429            -1.6%       5344        will-it-scale.time.user_time
  88596221            -1.7%   87067269        will-it-scale.workload
      6863 ±  6%      -9.7%       6200        boot-time.idle
    144908 ± 40%     -66.8%      48173 ± 93%  meminfo.CmaFree
      0.00 ± 70%      +0.0        0.00        mpstat.cpu.iowait%
    448336 ± 14%     -34.8%     292125 ±  3%  turbostat.C1
      7684 ±  6%      -9.5%       6957        uptime.idle
 1.601e+09 ± 29%     -46.9%  8.501e+08 ± 12%  time.involuntary_context_switches
      5429            -1.6%       5344        time.user_time
  44013162            -1.7%   43243125        vmstat.system.cs
    207684            -1.1%     205485        vmstat.system.in
   2217033 ± 15%     -15.8%    1866876 ±  2%  cpuidle.C1.time
    451218 ± 14%     -34.7%     294841 ±  2%  cpuidle.C1.usage
     24839 ± 10%     -19.9%      19896        cpuidle.POLL.time
      7656 ± 11%     -38.9%       4676 ±  8%  cpuidle.POLL.usage
      5.48 ± 49%     -67.3%       1.79 ±100%  irq_exception_noise.__do_page_fault.95th
      9.46 ± 21%     -58.2%       3.95 ± 64%  irq_exception_noise.__do_page_fault.99th
     35.67 ±  8%   +1394.4%     533.00 ± 96%  irq_exception_noise.irq_nr
     52109 ±  3%     -16.0%      43784 ±  4%  irq_exception_noise.softirq_time
     36226 ± 40%     -66.7%      12048 ± 93%  proc-vmstat.nr_free_cma
     25916            -1.0%      25659        proc-vmstat.nr_slab_reclaimable
     16279 ±  8%   +2646.1%     447053 ± 82%  proc-vmstat.pgalloc_movable
   2231117           -18.4%    1820828 ± 20%  proc-vmstat.pgalloc_normal
   1109316 ± 46%     -86.9%     145207 ±109%  numa-numastat.node1.local_node
   1114700 ± 45%     -84.5%     172877 ± 85%  numa-numastat.node1.numa_hit
      5523 ±140%    +402.8%      27768 ± 39%  numa-numastat.node1.other_node
     29013 ± 29%   +3048.1%     913379 ± 73%  numa-numastat.node3.local_node
     65032 ± 13%   +1335.1%     933270 ± 70%  numa-numastat.node3.numa_hit
     36018           -44.8%      19897 ± 75%  numa-numastat.node3.other_node
     12.79 ± 21%   +7739.1%       1002 ±136%  sched_debug.cpu.cpu_load[1].max
      1.82 ± 10%   +3901.1%      72.92 ±135%  sched_debug.cpu.cpu_load[1].stddev
      1.71 ±  4%   +5055.8%      88.08 ±137%  sched_debug.cpu.cpu_load[2].stddev
     12.33 ± 23%   +9061.9%       1129 ±139%  sched_debug.cpu.cpu_load[3].max
      1.78 ± 10%   +4514.8%      82.18 ±137%  sched_debug.cpu.cpu_load[3].stddev
      4692 ± 72%    +154.5%      11945 ± 29%  sched_debug.cpu.max_idle_balance_cost.stddev
     23979            -8.3%      21983        slabinfo.kmalloc-96.active_objs
      1358 ±  6%     -17.9%       1114 ±  3%  slabinfo.nsproxy.active_objs
      1358 ±  6%     -17.9%       1114 ±  3%  slabinfo.nsproxy.num_objs
     15229           +12.4%      17119        slabinfo.pde_opener.active_objs
     15229           +12.4%      17119        slabinfo.pde_opener.num_objs
     59541 ±  8%     -10.1%      53537 ±  8%  slabinfo.vm_area_struct.active_objs
     59612 ±  8%     -10.1%      53604 ±  8%  slabinfo.vm_area_struct.num_objs
 4.163e+13            -1.4%  4.105e+13        perf-stat.branch-instructions
 6.537e+11            -1.2%  6.459e+11        perf-stat.branch-misses
 2.667e+10            -1.7%  2.621e+10        perf-stat.context-switches
      1.21            +1.3%       1.22        perf-stat.cpi
    150508            -9.8%     135825 ±  3%  perf-stat.cpu-migrations
      5.75 ± 33%      +5.4       11.11 ± 26%  perf-stat.iTLB-load-miss-rate%
 3.619e+09 ± 36%    +100.9%  7.272e+09 ± 30%  perf-stat.iTLB-load-misses
 2.089e+14            -1.3%  2.062e+14        perf-stat.instructions
     64607 ± 29%     -50.5%      31964 ± 37%  perf-stat.instructions-per-iTLB-miss
      0.83            -1.3%       0.82        perf-stat.ipc
      3972 ±  4%     -14.7%       3388 ±  8%  numa-meminfo.node0.PageTables
    207919 ± 25%     -57.2%      88989 ± 74%  numa-meminfo.node1.Active
    207715 ± 26%     -57.3%      88785 ± 74%  numa-meminfo.node1.Active(anon)
    356529           -34.3%     234069 ±  2%  numa-meminfo.node1.FilePages
    789129 ±  5%     -19.8%     633161 ± 12%  numa-meminfo.node1.MemUsed
     34777 ±  8%     -48.2%      18010 ± 30%  numa-meminfo.node1.SReclaimable
     69641 ±  4%     -20.7%      55250 ± 12%  numa-meminfo.node1.SUnreclaim
    125526 ±  4%     -96.3%       4602 ± 41%  numa-meminfo.node1.Shmem
    104419           -29.8%      73261 ± 16%  numa-meminfo.node1.Slab
    103661 ± 17%     -72.0%      29029 ± 99%  numa-meminfo.node2.Active
    103661 ± 17%     -72.2%      28829 ±101%  numa-meminfo.node2.Active(anon)
    103564 ± 18%     -72.0%      29007 ±100%  numa-meminfo.node2.AnonPages
    671654 ±  7%     -14.6%     573598 ±  4%  numa-meminfo.node2.MemUsed
     44206 ±127%    +301.4%     177465 ± 42%  numa-meminfo.node3.Active
     44206 ±127%    +301.0%     177263 ± 42%  numa-meminfo.node3.Active(anon)
      8738           +12.2%       9805 ±  8%  numa-meminfo.node3.KernelStack
    603605 ±  9%     +27.8%     771554 ± 14%  numa-meminfo.node3.MemUsed
     14438 ±  6%    +122.9%      32181 ± 42%  numa-meminfo.node3.SReclaimable
      2786 ±137%   +3302.0%      94792 ± 71%  numa-meminfo.node3.Shmem
     71461 ±  7%     +45.2%     103771 ± 29%  numa-meminfo.node3.Slab
    247197 ±  4%      -7.8%     227843        numa-meminfo.node3.Unevictable
    991.67 ±  4%     -14.7%     846.00 ±  8%  numa-vmstat.node0.nr_page_table_pages
     51926 ± 26%     -57.3%      22196 ± 74%  numa-vmstat.node1.nr_active_anon
     89137           -34.4%      58516 ±  2%  numa-vmstat.node1.nr_file_pages
      1679 ±  5%     -10.8%       1498 ±  4%  numa-vmstat.node1.nr_mapped
     31386 ±  4%     -96.3%       1150 ± 41%  numa-vmstat.node1.nr_shmem
      8694 ±  8%     -48.2%       4502 ± 30%  numa-vmstat.node1.nr_slab_reclaimable
     17410 ±  4%     -20.7%      13812 ± 12%  numa-vmstat.node1.nr_slab_unreclaimable
     51926 ± 26%     -57.3%      22196 ± 74%  numa-vmstat.node1.nr_zone_active_anon
   1037174 ± 24%     -57.0%     446205 ± 35%  numa-vmstat.node1.numa_hit
    961611 ± 26%     -65.8%     328687 ± 50%  numa-vmstat.node1.numa_local
     75563 ± 44%     +55.5%     117517 ±  9%  numa-vmstat.node1.numa_other
     25914 ± 17%     -72.2%       7206 ±101%  numa-vmstat.node2.nr_active_anon
     25891 ± 18%     -72.0%       7251 ±100%  numa-vmstat.node2.nr_anon_pages
     25914 ± 17%     -72.2%       7206 ±101%  numa-vmstat.node2.nr_zone_active_anon
     11051 ±127%    +301.0%      44309 ± 42%  numa-vmstat.node3.nr_active_anon
     36227 ± 40%     -66.7%      12049 ± 93%  numa-vmstat.node3.nr_free_cma
      0.33 ±141%  +25000.0%      83.67 ± 81%  numa-vmstat.node3.nr_inactive_file
      8739           +12.2%       9806 ±  8%  numa-vmstat.node3.nr_kernel_stack
    696.67 ±137%   +3299.7%      23684 ± 71%  numa-vmstat.node3.nr_shmem
      3609 ±  6%    +122.9%       8044 ± 42%  numa-vmstat.node3.nr_slab_reclaimable
     61799 ±  4%      -7.8%      56960        numa-vmstat.node3.nr_unevictable
     11053 ±127%    +301.4%      44361 ± 42%  numa-vmstat.node3.nr_zone_active_anon
      0.33 ±141%  +25000.0%      83.67 ± 81%  numa-vmstat.node3.nr_zone_inactive_file
     61799 ±  4%      -7.8%      56960        numa-vmstat.node3.nr_zone_unevictable
    217951 ±  8%    +280.8%     829976 ± 65%  numa-vmstat.node3.numa_hit
     91303 ± 19%    +689.3%     720647 ± 77%  numa-vmstat.node3.numa_local
    126648           -13.7%     109329 ± 13%  numa-vmstat.node3.numa_other
      8.54            -0.1        8.40        perf-profile.calltrace.cycles-pp.dequeue_task_fair.__schedule.schedule.pipe_wait.pipe_read
      5.04            -0.1        4.94        perf-profile.calltrace.cycles-pp.__switch_to.read
      3.43            -0.1        3.35        perf-profile.calltrace.cycles-pp.syscall_return_via_sysret.write
      2.77            -0.1        2.72        perf-profile.calltrace.cycles-pp.reweight_entity.enqueue_task_fair.ttwu_do_activate.try_to_wake_up.autoremove_wake_function
      1.99            -0.0        1.94        perf-profile.calltrace.cycles-pp.copy_page_to_iter.pipe_read.__vfs_read.vfs_read.ksys_read
      0.60 ±  2%      -0.0        0.57 ±  2%  perf-profile.calltrace.cycles-pp.find_next_bit.cpumask_next_wrap.select_idle_sibling.select_task_rq_fair.try_to_wake_up
      0.81            -0.0        0.78        perf-profile.calltrace.cycles-pp.___perf_sw_event.__schedule.schedule.pipe_wait.pipe_read
      0.78            +0.0        0.80        perf-profile.calltrace.cycles-pp.__fdget_pos.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
      0.73            +0.0        0.75        perf-profile.calltrace.cycles-pp.__fget_light.__fdget_pos.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.92            +0.0        0.95        perf-profile.calltrace.cycles-pp.check_preempt_wakeup.check_preempt_curr.ttwu_do_wakeup.try_to_wake_up.autoremove_wake_function
      2.11            +0.0        2.15        perf-profile.calltrace.cycles-pp.security_file_permission.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
      7.00            -0.1        6.86        perf-profile.children.cycles-pp.syscall_return_via_sysret
      5.26            -0.1        5.14        perf-profile.children.cycles-pp.__switch_to
      5.65            -0.1        5.56        perf-profile.children.cycles-pp.reweight_entity
      2.17            -0.1        2.12        perf-profile.children.cycles-pp.copy_page_to_iter
      2.94            -0.0        2.90        perf-profile.children.cycles-pp.update_cfs_group
      3.11            -0.0        3.07        perf-profile.children.cycles-pp.pick_next_task_fair
      2.59            -0.0        2.55        perf-profile.children.cycles-pp.load_new_mm_cr3
      1.92            -0.0        1.88        perf-profile.children.cycles-pp._raw_spin_lock_irqsave
      1.11            -0.0        1.08 ±  2%  perf-profile.children.cycles-pp.find_next_bit
      0.59            -0.0        0.56        perf-profile.children.cycles-pp.finish_task_switch
      0.14 ± 15%      -0.0        0.11 ± 16%  perf-profile.children.cycles-pp.write@plt
      1.21            -0.0        1.18        perf-profile.children.cycles-pp.set_next_entity
      0.85            -0.0        0.82        perf-profile.children.cycles-pp.___perf_sw_event
      0.13 ±  3%      -0.0        0.11 ±  4%  perf-profile.children.cycles-pp.timespec_trunc
      0.47 ±  2%      -0.0        0.45        perf-profile.children.cycles-pp.anon_pipe_buf_release
      0.38 ±  2%      -0.0        0.36        perf-profile.children.cycles-pp.file_update_time
      0.74            -0.0        0.73        perf-profile.children.cycles-pp.copyout
      0.41 ±  2%      -0.0        0.39        perf-profile.children.cycles-pp.copy_user_enhanced_fast_string
      0.32            -0.0        0.30        perf-profile.children.cycles-pp.__x64_sys_read
      0.14            -0.0        0.12 ±  3%  perf-profile.children.cycles-pp.current_kernel_time64
      0.91            +0.0        0.92        perf-profile.children.cycles-pp.touch_atime
      0.40            +0.0        0.41        perf-profile.children.cycles-pp._cond_resched
      0.18 ±  2%      +0.0        0.20        perf-profile.children.cycles-pp.activate_task
      0.05            +0.0        0.07 ±  6%  perf-profile.children.cycles-pp.default_wake_function
      0.24            +0.0        0.27 ±  3%  perf-profile.children.cycles-pp.rcu_all_qs
      0.60 ±  2%      +0.0        0.64 ±  2%  perf-profile.children.cycles-pp.update_min_vruntime
      0.42 ±  4%      +0.0        0.46 ±  4%  perf-profile.children.cycles-pp.probe_sched_switch
      1.33            +0.0        1.38        perf-profile.children.cycles-pp.__fget_light
      0.53 ±  2%      +0.1        0.58        perf-profile.children.cycles-pp.entry_SYSCALL_64_stage2
      0.31            +0.1        0.36 ±  2%  perf-profile.children.cycles-pp.generic_pipe_buf_confirm
      4.35            +0.1        4.41        perf-profile.children.cycles-pp.switch_mm_irqs_off
      2.52            +0.1        2.58        perf-profile.children.cycles-pp.selinux_file_permission
      0.00            +0.1        0.07 ± 11%  perf-profile.children.cycles-pp.hrtick_update
      7.00            -0.1        6.86        perf-profile.self.cycles-pp.syscall_return_via_sysret
      5.26            -0.1        5.14        perf-profile.self.cycles-pp.__switch_to
      0.29            -0.1        0.19 ±  2%  perf-profile.self.cycles-pp.ksys_read
      1.49            -0.1        1.43        perf-profile.self.cycles-pp.dequeue_task_fair
      2.41            -0.1        2.35        perf-profile.self.cycles-pp.__schedule
      1.46            -0.0        1.41        perf-profile.self.cycles-pp.select_task_rq_fair
      2.94            -0.0        2.90        perf-profile.self.cycles-pp.update_cfs_group
      0.44            -0.0        0.40        perf-profile.self.cycles-pp.dequeue_entity
      0.48            -0.0        0.44        perf-profile.self.cycles-pp.finish_task_switch
      2.59            -0.0        2.55        perf-profile.self.cycles-pp.load_new_mm_cr3
      1.11            -0.0        1.08 ±  2%  perf-profile.self.cycles-pp.find_next_bit
      1.91            -0.0        1.88        perf-profile.self.cycles-pp._raw_spin_lock_irqsave
      0.78            -0.0        0.75        perf-profile.self.cycles-pp.___perf_sw_event
      0.14 ± 15%      -0.0        0.11 ± 16%  perf-profile.self.cycles-pp.write@plt
      0.37            -0.0        0.35 ±  2%  perf-profile.self.cycles-pp.__wake_up_common_lock
      0.20 ±  2%      -0.0        0.17 ±  2%  perf-profile.self.cycles-pp.__fdget_pos
      0.47 ±  2%      -0.0        0.44        perf-profile.self.cycles-pp.anon_pipe_buf_release
      0.87            -0.0        0.85        perf-profile.self.cycles-pp.copy_user_generic_unrolled
      0.13 ±  3%      -0.0        0.11 ±  4%  perf-profile.self.cycles-pp.timespec_trunc
      0.41 ±  2%      -0.0        0.39        perf-profile.self.cycles-pp.copy_user_enhanced_fast_string
      0.38            -0.0        0.36        perf-profile.self.cycles-pp.__wake_up_common
      0.32            -0.0        0.30        perf-profile.self.cycles-pp.__x64_sys_read
      0.14 ±  3%      -0.0        0.12 ±  3%  perf-profile.self.cycles-pp.current_kernel_time64
      0.30            -0.0        0.28        perf-profile.self.cycles-pp.set_next_entity
      0.28 ±  3%      +0.0        0.30        perf-profile.self.cycles-pp._cond_resched
      0.18 ±  2%      +0.0        0.20        perf-profile.self.cycles-pp.activate_task
      0.17 ±  2%      +0.0        0.19        perf-profile.self.cycles-pp.__might_fault
      0.05            +0.0        0.07 ±  6%  perf-profile.self.cycles-pp.default_wake_function
      0.17 ±  2%      +0.0        0.20        perf-profile.self.cycles-pp.ttwu_do_activate
      0.66            +0.0        0.69        perf-profile.self.cycles-pp.write
      0.24            +0.0        0.27 ±  3%  perf-profile.self.cycles-pp.rcu_all_qs
      0.67            +0.0        0.70        perf-profile.self.cycles-pp.entry_SYSCALL_64_after_hwframe
      0.60 ±  2%      +0.0        0.64 ±  2%  perf-profile.self.cycles-pp.update_min_vruntime
      0.42 ±  4%      +0.0        0.46 ±  4%  perf-profile.self.cycles-pp.probe_sched_switch
      1.33            +0.0        1.37        perf-profile.self.cycles-pp.__fget_light
      1.61            +0.0        1.66        perf-profile.self.cycles-pp.pipe_read
      0.53 ±  2%      +0.1        0.58        perf-profile.self.cycles-pp.entry_SYSCALL_64_stage2
      0.31            +0.1        0.36 ±  2%  perf-profile.self.cycles-pp.generic_pipe_buf_confirm
      1.04            +0.1        1.11        perf-profile.self.cycles-pp.pipe_write
      0.00            +0.1        0.07 ± 11%  perf-profile.self.cycles-pp.hrtick_update
      2.00            +0.1        2.08        perf-profile.self.cycles-pp.switch_mm_irqs_off

=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/nr_task/thp_enabled/test/cpufreq_governor:
  lkp-skl-4sp1/will-it-scale/debian-x86_64-2018-04-03.cgz/x86_64-rhel-7.2/gcc-7/100%/never/page_fault3/performance

commit: 
  ba98a1cdad71d259a194461b3a61471b49b14df1
  a7a8993bfe3ccb54ad468b9f1799649e4ad1ff12

ba98a1cdad71d259 a7a8993bfe3ccb54ad468b9f17 
---------------- -------------------------- 
       fail:runs  %reproduction    fail:runs
           |             |             |    
          1:3          -33%            :3     dmesg.WARNING:stack_going_in_the_wrong_direction?ip=file_update_time/0x
           :3           33%           1:3     stderr.mount.nfs:Connection_timed_out
         34:3         -401%          22:3     perf-profile.calltrace.cycles-pp.error_entry.testcase
         17:3         -207%          11:3     perf-profile.calltrace.cycles-pp.sync_regs.error_entry.testcase
         34:3         -404%          22:3     perf-profile.children.cycles-pp.error_entry
          0:3           -2%           0:3     perf-profile.children.cycles-pp.error_exit
         16:3         -196%          11:3     perf-profile.self.cycles-pp.error_entry
          0:3           -2%           0:3     perf-profile.self.cycles-pp.error_exit
         %stddev     %change         %stddev
             \          |                \  
    467454            -1.8%     459251        will-it-scale.per_process_ops
     10856 ±  4%     -23.1%       8344 ±  7%  will-it-scale.per_thread_ops
    118134 ±  2%     +11.7%     131943        will-it-scale.time.involuntary_context_switches
 6.277e+08 ±  4%     -23.1%  4.827e+08 ±  7%  will-it-scale.time.minor_page_faults
      7406            +5.8%       7839        will-it-scale.time.percent_of_cpu_this_job_got
     44526            +5.8%      47106        will-it-scale.time.system_time
   7351468 ±  5%     -18.3%    6009014 ±  7%  will-it-scale.time.voluntary_context_switches
  91835846            -2.2%   89778599        will-it-scale.workload
   2534640            +4.3%    2643005 ±  2%  interrupts.CAL:Function_call_interrupts
      2819 ±  5%     +22.9%       3464 ± 18%  kthread_noise.total_time
     30273 ±  4%     -12.7%      26415 ±  5%  vmstat.system.cs
      1.52 ±  2%     +15.2%       1.75 ±  2%  irq_exception_noise.__do_page_fault.99th
    296.67 ± 12%     -36.7%     187.67 ± 12%  irq_exception_noise.softirq_time
    230900 ±  3%     +30.3%     300925 ±  3%  meminfo.Inactive
    230184 ±  3%     +30.4%     300180 ±  3%  meminfo.Inactive(anon)
     11.62 ±  3%      -2.2        9.40 ±  5%  mpstat.cpu.idle%
      0.00 ± 14%      -0.0        0.00 ±  4%  mpstat.cpu.iowait%
   7992174           -11.1%    7101976 ±  3%  softirqs.RCU
   4973624 ±  2%     -12.9%    4333370 ±  2%  softirqs.SCHED
    118134 ±  2%     +11.7%     131943        time.involuntary_context_switches
 6.277e+08 ±  4%     -23.1%  4.827e+08 ±  7%  time.minor_page_faults
      7406            +5.8%       7839        time.percent_of_cpu_this_job_got
     44526            +5.8%      47106        time.system_time
   7351468 ±  5%     -18.3%    6009014 ±  7%  time.voluntary_context_switches
 2.702e+09 ±  5%     -16.7%  2.251e+09 ±  7%  cpuidle.C1E.time
   6834329 ±  5%     -15.8%    5756243 ±  7%  cpuidle.C1E.usage
 1.046e+10 ±  3%     -19.8%  8.389e+09 ±  4%  cpuidle.C6.time
  13961845 ±  3%     -19.3%   11265555 ±  4%  cpuidle.C6.usage
   1309307 ±  7%     -14.8%    1116168 ±  8%  cpuidle.POLL.time
     19774 ±  6%     -13.7%      17063 ±  7%  cpuidle.POLL.usage
      2523 ±  4%     -11.1%       2243 ±  4%  slabinfo.biovec-64.active_objs
      2523 ±  4%     -11.1%       2243 ±  4%  slabinfo.biovec-64.num_objs
      2610 ±  8%     -33.7%       1731 ± 22%  slabinfo.dmaengine-unmap-16.active_objs
      2610 ±  8%     -33.7%       1731 ± 22%  slabinfo.dmaengine-unmap-16.num_objs
      5118 ± 17%     -22.6%       3962 ±  9%  slabinfo.eventpoll_pwq.active_objs
      5118 ± 17%     -22.6%       3962 ±  9%  slabinfo.eventpoll_pwq.num_objs
      4583 ±  3%     -14.0%       3941 ±  4%  slabinfo.sock_inode_cache.active_objs
      4583 ±  3%     -14.0%       3941 ±  4%  slabinfo.sock_inode_cache.num_objs
      1933            +2.6%       1984        turbostat.Avg_MHz
   6832021 ±  5%     -15.8%    5754156 ±  7%  turbostat.C1E
      2.32 ±  5%      -0.4        1.94 ±  7%  turbostat.C1E%
  13954211 ±  3%     -19.3%   11259436 ±  4%  turbostat.C6
      8.97 ±  3%      -1.8        7.20 ±  4%  turbostat.C6%
      6.18 ±  4%     -17.1%       5.13 ±  5%  turbostat.CPU%c1
      5.12 ±  3%     -21.7%       4.01 ±  4%  turbostat.CPU%c6
      1.76 ±  2%     -34.7%       1.15 ±  2%  turbostat.Pkg%pc2
     57314 ±  4%     +30.4%      74717 ±  4%  proc-vmstat.nr_inactive_anon
     57319 ±  4%     +30.4%      74719 ±  4%  proc-vmstat.nr_zone_inactive_anon
     24415 ± 19%     -62.2%       9236 ±  7%  proc-vmstat.numa_hint_faults
  69661453            -1.8%   68405712        proc-vmstat.numa_hit
  69553390            -1.8%   68297790        proc-vmstat.numa_local
      8792 ± 29%     -92.6%     654.33 ± 23%  proc-vmstat.numa_pages_migrated
     40251 ± 32%     -76.5%       9474 ±  3%  proc-vmstat.numa_pte_updates
  69522532            -1.6%   68383074        proc-vmstat.pgalloc_normal
 2.762e+10            -2.2%  2.701e+10        proc-vmstat.pgfault
  68825100            -1.5%   67772256        proc-vmstat.pgfree
      8792 ± 29%     -92.6%     654.33 ± 23%  proc-vmstat.pgmigrate_success
     57992 ±  6%     +56.2%      90591 ±  3%  numa-meminfo.node0.Inactive
     57916 ±  6%     +56.3%      90513 ±  3%  numa-meminfo.node0.Inactive(anon)
     37285 ± 12%     +36.0%      50709 ±  5%  numa-meminfo.node0.SReclaimable
    110971 ±  8%     +22.7%     136209 ±  8%  numa-meminfo.node0.Slab
     23601 ± 55%    +559.5%     155651 ± 36%  numa-meminfo.node1.AnonPages
     62484 ± 12%     +17.5%      73417 ±  3%  numa-meminfo.node1.Inactive
     62323 ± 12%     +17.2%      73023 ±  4%  numa-meminfo.node1.Inactive(anon)
    109714 ± 63%     -85.6%      15832 ± 96%  numa-meminfo.node2.AnonPages
     52236 ± 13%     +22.7%      64074 ±  3%  numa-meminfo.node2.Inactive
     51922 ± 12%     +23.2%      63963 ±  3%  numa-meminfo.node2.Inactive(anon)
     60241 ± 11%     +21.9%      73442 ±  8%  numa-meminfo.node3.Inactive
     60077 ± 12%     +22.0%      73279 ±  8%  numa-meminfo.node3.Inactive(anon)
     14093 ±  6%     +55.9%      21977 ±  3%  numa-vmstat.node0.nr_inactive_anon
      9321 ± 12%     +36.0%      12675 ±  5%  numa-vmstat.node0.nr_slab_reclaimable
     14090 ±  6%     +56.0%      21977 ±  3%  numa-vmstat.node0.nr_zone_inactive_anon
      5900 ± 55%    +559.4%      38909 ± 36%  numa-vmstat.node1.nr_anon_pages
     15413 ± 12%     +14.8%      17688 ±  4%  numa-vmstat.node1.nr_inactive_anon
     15413 ± 12%     +14.8%      17688 ±  4%  numa-vmstat.node1.nr_zone_inactive_anon
     27430 ± 63%     -85.6%       3960 ± 96%  numa-vmstat.node2.nr_anon_pages
     12928 ± 12%     +20.0%      15508 ±  3%  numa-vmstat.node2.nr_inactive_anon
     12927 ± 12%     +20.0%      15507 ±  3%  numa-vmstat.node2.nr_zone_inactive_anon
      6229 ± 10%    +117.5%      13547 ± 44%  numa-vmstat.node3
     14669 ± 11%     +19.6%      17537 ±  7%  numa-vmstat.node3.nr_inactive_anon
     14674 ± 11%     +19.5%      17541 ±  7%  numa-vmstat.node3.nr_zone_inactive_anon
     24617 ±141%    -100.0%       0.00        latency_stats.avg.io_schedule.nfs_lock_and_join_requests.nfs_updatepage.nfs_write_end.generic_perform_write.nfs_file_write.__vfs_write.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
      5049 ±105%     -99.4%      28.33 ± 82%  latency_stats.avg.call_rwsem_down_write_failed.path_openat.do_filp_open.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
    152457 ± 27%    +233.6%     508656 ± 92%  latency_stats.avg.max
      0.00       +3.9e+107%     390767 ±141%  latency_stats.avg.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_getattr.__nfs_revalidate_inode.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_openat
     24617 ±141%    -100.0%       0.00        latency_stats.max.io_schedule.nfs_lock_and_join_requests.nfs_updatepage.nfs_write_end.generic_perform_write.nfs_file_write.__vfs_write.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
      4240 ±141%    -100.0%       0.00        latency_stats.max.call_rwsem_down_write_failed.do_unlinkat.do_syscall_64.entry_SYSCALL_64_after_hwframe
      8565 ± 70%     -99.1%      80.33 ±115%  latency_stats.max.call_rwsem_down_write_failed.path_openat.do_filp_open.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
    204835 ±  6%    +457.6%    1142244 ±114%  latency_stats.max.max
      0.00       +5.1e+105%       5057 ±141%  latency_stats.max.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_access.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_openat.do_filp_open
      0.00         +1e+108%     995083 ±141%  latency_stats.max.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_getattr.__nfs_revalidate_inode.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_openat
     13175 ±  4%    -100.0%       0.00        latency_stats.sum.io_schedule.__lock_page_or_retry.filemap_fault.__do_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault.do_page_fault.page_fault
     24617 ±141%    -100.0%       0.00        latency_stats.sum.io_schedule.nfs_lock_and_join_requests.nfs_updatepage.nfs_write_end.generic_perform_write.nfs_file_write.__vfs_write.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
      4260 ±141%    -100.0%       0.00        latency_stats.sum.call_rwsem_down_write_failed.do_unlinkat.do_syscall_64.entry_SYSCALL_64_after_hwframe
      8640 ± 70%     -97.5%     216.33 ±108%  latency_stats.sum.call_rwsem_down_write_failed.path_openat.do_filp_open.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
      6673 ± 89%     -92.8%     477.67 ± 74%  latency_stats.sum.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_getattr.__nfs_revalidate_inode.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_lookupat
      0.00       +4.2e+105%       4228 ±130%  latency_stats.sum.io_schedule.__lock_page_killable.__lock_page_or_retry.filemap_fault.__do_fault.handle_pte_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault.do_page_fault.page_fault
      0.00       +7.5e+105%       7450 ± 98%  latency_stats.sum.io_schedule.__lock_page_or_retry.filemap_fault.__do_fault.handle_pte_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault.do_page_fault.page_fault
      0.00       +1.3e+106%      13050 ±141%  latency_stats.sum.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_access.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_openat.do_filp_open
      0.00       +1.5e+110%  1.508e+08 ±141%  latency_stats.sum.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_getattr.__nfs_revalidate_inode.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_openat
      0.97            -0.0        0.94        perf-stat.branch-miss-rate%
 1.329e+11            -2.6%  1.294e+11        perf-stat.branch-misses
 2.254e+11            -1.9%   2.21e+11        perf-stat.cache-references
  18308779 ±  4%     -12.8%   15969618 ±  5%  perf-stat.context-switches
      3.20            +1.8%       3.26        perf-stat.cpi
 2.233e+14            +2.7%  2.293e+14        perf-stat.cpu-cycles
      4.01            -0.2        3.83        perf-stat.dTLB-store-miss-rate%
  4.51e+11            -2.2%   4.41e+11        perf-stat.dTLB-store-misses
  1.08e+13            +2.6%  1.109e+13        perf-stat.dTLB-stores
 3.158e+10 ±  5%     +16.8%  3.689e+10 ±  2%  perf-stat.iTLB-load-misses
      2214 ±  5%     -13.8%       1907 ±  2%  perf-stat.instructions-per-iTLB-miss
      0.31            -1.8%       0.31        perf-stat.ipc
 2.762e+10            -2.2%  2.701e+10        perf-stat.minor-faults
 1.535e+10           -11.2%  1.362e+10        perf-stat.node-loads
      9.75            +1.1       10.89        perf-stat.node-store-miss-rate%
 3.012e+09           +10.6%  3.332e+09 ±  2%  perf-stat.node-store-misses
 2.787e+10            -2.2%  2.725e+10        perf-stat.node-stores
 2.762e+10            -2.2%  2.701e+10        perf-stat.page-faults
    759458            +3.2%     783404        perf-stat.path-length
    246.39 ± 15%     -20.4%     196.12 ±  6%  sched_debug.cfs_rq:/.load_avg.max
      0.21 ±  3%      +9.0%       0.23 ±  4%  sched_debug.cfs_rq:/.nr_running.stddev
     16.64 ± 27%     +61.0%      26.79 ± 17%  sched_debug.cfs_rq:/.nr_spread_over.max
     75.15           -14.4%      64.30 ±  4%  sched_debug.cfs_rq:/.util_avg.stddev
    178.80 ±  3%     +25.4%     224.12 ±  7%  sched_debug.cfs_rq:/.util_est_enqueued.avg
      1075 ±  5%     -12.3%     943.36 ±  2%  sched_debug.cfs_rq:/.util_est_enqueued.max
   2093630 ± 27%     -36.1%    1337941 ± 16%  sched_debug.cpu.avg_idle.max
    297057 ± 11%     +37.8%     409294 ± 14%  sched_debug.cpu.avg_idle.min
    293240 ± 55%     -62.3%     110571 ± 13%  sched_debug.cpu.avg_idle.stddev
    770075 ±  9%     -19.3%     621136 ± 12%  sched_debug.cpu.max_idle_balance_cost.max
     48919 ± 46%     -66.9%      16190 ± 81%  sched_debug.cpu.max_idle_balance_cost.stddev
     21716 ±  5%     -16.8%      18061 ±  7%  sched_debug.cpu.nr_switches.min
     21519 ±  5%     -17.7%      17700 ±  7%  sched_debug.cpu.sched_count.min
     10586 ±  5%     -18.1%       8669 ±  7%  sched_debug.cpu.sched_goidle.avg
     14183 ±  3%     -17.6%      11693 ±  5%  sched_debug.cpu.sched_goidle.max
     10322 ±  5%     -18.6%       8407 ±  7%  sched_debug.cpu.sched_goidle.min
    400.99 ±  8%     -13.0%     348.75 ±  3%  sched_debug.cpu.sched_goidle.stddev
      5459 ±  8%     +10.0%       6006 ±  3%  sched_debug.cpu.ttwu_local.avg
      8.47 ± 42%    +345.8%      37.73 ± 77%  sched_debug.rt_rq:/.rt_time.max
      0.61 ± 42%    +343.0%       2.72 ± 77%  sched_debug.rt_rq:/.rt_time.stddev
     91.98           -30.9       61.11 ± 70%  perf-profile.calltrace.cycles-pp.testcase
      9.05            -9.1        0.00        perf-profile.calltrace.cycles-pp.__do_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault.do_page_fault
      8.91            -8.9        0.00        perf-profile.calltrace.cycles-pp.shmem_fault.__do_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
      8.06            -8.1        0.00        perf-profile.calltrace.cycles-pp.shmem_getpage_gfp.shmem_fault.__do_fault.__handle_mm_fault.handle_mm_fault
      7.59            -7.6        0.00        perf-profile.calltrace.cycles-pp.find_lock_entry.shmem_getpage_gfp.shmem_fault.__do_fault.__handle_mm_fault
      7.44            -7.4        0.00        perf-profile.calltrace.cycles-pp.finish_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault.do_page_fault
      7.28            -7.3        0.00        perf-profile.calltrace.cycles-pp.alloc_set_pte.finish_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
      5.31            -5.3        0.00        perf-profile.calltrace.cycles-pp.page_add_file_rmap.alloc_set_pte.finish_fault.__handle_mm_fault.handle_mm_fault
      8.08            -2.8        5.30 ± 70%  perf-profile.calltrace.cycles-pp.native_irq_return_iret.testcase
      5.95            -2.1        3.83 ± 70%  perf-profile.calltrace.cycles-pp.find_get_entry.find_lock_entry.shmem_getpage_gfp.shmem_fault.__do_fault
      5.95            -2.0        3.93 ± 70%  perf-profile.calltrace.cycles-pp.swapgs_restore_regs_and_return_to_usermode.testcase
      3.10            -1.1        2.01 ± 70%  perf-profile.calltrace.cycles-pp.__perf_sw_event.__do_page_fault.do_page_fault.page_fault.testcase
      2.36            -0.8        1.55 ± 70%  perf-profile.calltrace.cycles-pp.___perf_sw_event.__perf_sw_event.__do_page_fault.do_page_fault.page_fault
      1.08            -0.4        0.70 ± 70%  perf-profile.calltrace.cycles-pp.do_page_fault.testcase
      0.82            -0.3        0.54 ± 70%  perf-profile.calltrace.cycles-pp.trace_graph_entry.do_page_fault.testcase
      0.77            -0.3        0.50 ± 70%  perf-profile.calltrace.cycles-pp.ftrace_graph_caller.__do_page_fault.do_page_fault.page_fault.testcase
      0.59            -0.2        0.37 ± 70%  perf-profile.calltrace.cycles-pp.down_read_trylock.__do_page_fault.do_page_fault.page_fault.testcase
     91.98           -30.9       61.11 ± 70%  perf-profile.children.cycles-pp.testcase
      9.14            -3.2        5.99 ± 70%  perf-profile.children.cycles-pp.__do_fault
      8.20            -2.8        5.40 ± 70%  perf-profile.children.cycles-pp.shmem_getpage_gfp
      8.08            -2.8        5.31 ± 70%  perf-profile.children.cycles-pp.native_irq_return_iret
      6.08            -2.2        3.92 ± 70%  perf-profile.children.cycles-pp.find_get_entry
      6.08            -2.1        3.96 ± 70%  perf-profile.children.cycles-pp.sync_regs
      5.95            -2.0        3.93 ± 70%  perf-profile.children.cycles-pp.swapgs_restore_regs_and_return_to_usermode
      4.12            -1.4        2.73 ± 70%  perf-profile.children.cycles-pp.ftrace_graph_caller
      3.65            -1.2        2.42 ± 70%  perf-profile.children.cycles-pp.prepare_ftrace_return
      3.18            -1.1        2.07 ± 70%  perf-profile.children.cycles-pp.__perf_sw_event
      2.34            -0.8        1.52 ± 70%  perf-profile.children.cycles-pp.fault_dirty_shared_page
      0.80            -0.3        0.50 ± 70%  perf-profile.children.cycles-pp._raw_spin_lock
      0.76            -0.3        0.50 ± 70%  perf-profile.children.cycles-pp.tlb_flush_mmu_free
      0.61            -0.2        0.39 ± 70%  perf-profile.children.cycles-pp.down_read_trylock
      0.48 ±  2%      -0.2        0.28 ± 70%  perf-profile.children.cycles-pp.pmd_devmap_trans_unstable
      0.26 ±  6%      -0.1        0.15 ± 71%  perf-profile.children.cycles-pp.ktime_get
      0.20 ±  2%      -0.1        0.12 ± 70%  perf-profile.children.cycles-pp.perf_exclude_event
      0.22 ±  2%      -0.1        0.13 ± 70%  perf-profile.children.cycles-pp._cond_resched
      0.17            -0.1        0.11 ± 70%  perf-profile.children.cycles-pp.page_rmapping
      0.13            -0.1        0.07 ± 70%  perf-profile.children.cycles-pp.rcu_all_qs
      0.07            -0.0        0.04 ± 70%  perf-profile.children.cycles-pp.ftrace_lookup_ip
     22.36            -7.8       14.59 ± 70%  perf-profile.self.cycles-pp.testcase
      8.08            -2.8        5.31 ± 70%  perf-profile.self.cycles-pp.native_irq_return_iret
      6.08            -2.1        3.96 ± 70%  perf-profile.self.cycles-pp.sync_regs
      5.81            -2.0        3.84 ± 70%  perf-profile.self.cycles-pp.swapgs_restore_regs_and_return_to_usermode
      3.27            -1.6        1.65 ± 70%  perf-profile.self.cycles-pp.__handle_mm_fault
      3.79            -1.4        2.36 ± 70%  perf-profile.self.cycles-pp.find_get_entry
      3.80            -1.3        2.53 ± 70%  perf-profile.self.cycles-pp.trace_graph_entry
      1.10            -0.5        0.57 ± 70%  perf-profile.self.cycles-pp.alloc_set_pte
      1.24            -0.4        0.81 ± 70%  perf-profile.self.cycles-pp.shmem_fault
      0.80            -0.3        0.50 ± 70%  perf-profile.self.cycles-pp._raw_spin_lock
      0.81            -0.3        0.51 ± 70%  perf-profile.self.cycles-pp.find_lock_entry
      0.80 ±  2%      -0.3        0.51 ± 70%  perf-profile.self.cycles-pp.__perf_sw_event
      0.61            -0.2        0.38 ± 70%  perf-profile.self.cycles-pp.down_read_trylock
      0.60            -0.2        0.39 ± 70%  perf-profile.self.cycles-pp.shmem_getpage_gfp
      0.48            -0.2        0.27 ± 70%  perf-profile.self.cycles-pp.pmd_devmap_trans_unstable
      0.47            -0.2        0.30 ± 70%  perf-profile.self.cycles-pp.file_update_time
      0.34            -0.1        0.22 ± 70%  perf-profile.self.cycles-pp.do_page_fault
      0.22 ±  4%      -0.1        0.11 ± 70%  perf-profile.self.cycles-pp.__do_fault
      0.25 ±  5%      -0.1        0.14 ± 71%  perf-profile.self.cycles-pp.ktime_get
      0.21 ±  2%      -0.1        0.12 ± 70%  perf-profile.self.cycles-pp.finish_fault
      0.23 ±  2%      -0.1        0.14 ± 70%  perf-profile.self.cycles-pp.fault_dirty_shared_page
      0.22 ±  2%      -0.1        0.14 ± 70%  perf-profile.self.cycles-pp.prepare_exit_to_usermode
      0.20 ±  2%      -0.1        0.12 ± 70%  perf-profile.self.cycles-pp.perf_exclude_event
      0.16            -0.1        0.10 ± 70%  perf-profile.self.cycles-pp._cond_resched
      0.13            -0.1        0.07 ± 70%  perf-profile.self.cycles-pp.rcu_all_qs
      0.07            -0.0        0.04 ± 70%  perf-profile.self.cycles-pp.ftrace_lookup_ip

=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/nr_task/thp_enabled/test/cpufreq_governor:
  lkp-skl-4sp1/will-it-scale/debian-x86_64-2018-04-03.cgz/x86_64-rhel-7.2/gcc-7/100%/always/context_switch1/performance

commit: 
  ba98a1cdad71d259a194461b3a61471b49b14df1
  a7a8993bfe3ccb54ad468b9f1799649e4ad1ff12

ba98a1cdad71d259 a7a8993bfe3ccb54ad468b9f17 
---------------- -------------------------- 
       fail:runs  %reproduction    fail:runs
           |             |             |    
           :3           33%           1:3     dmesg.WARNING:at#for_ip_interrupt_entry/0x
           :3           33%           1:3     dmesg.WARNING:at#for_ip_ret_from_intr/0x
           :3           67%           2:3     kmsg.pstore:crypto_comp_decompress_failed,ret=
           :3           67%           2:3     kmsg.pstore:decompression_failed
         %stddev     %change         %stddev
             \          |                \  
    223910            -1.3%     220930        will-it-scale.per_process_ops
    233722            -1.0%     231288        will-it-scale.per_thread_ops
 6.001e+08 ± 13%     +31.4%  7.887e+08 ±  4%  will-it-scale.time.involuntary_context_switches
     18003 ±  4%     +10.9%      19956        will-it-scale.time.minor_page_faults
  1.29e+10            -2.5%  1.258e+10        will-it-scale.time.voluntary_context_switches
  87865617            -1.2%   86826277        will-it-scale.workload
   2880329 ±  2%      +5.4%    3034904        interrupts.CAL:Function_call_interrupts
   7695018           -23.3%    5905066 ±  8%  meminfo.DirectMap2M
      0.00 ± 39%      -0.0        0.00 ± 78%  mpstat.cpu.iowait%
      4621 ± 12%     +13.4%       5241        proc-vmstat.numa_hint_faults_local
    715714           +27.6%     913142 ± 13%  softirqs.SCHED
    515653 ±  6%     -20.0%     412650 ± 15%  turbostat.C1
  43643516            -1.2%   43127031        vmstat.system.cs
   2893393 ±  4%     -23.6%    2210524 ± 10%  cpuidle.C1.time
    518051 ±  6%     -19.9%     415081 ± 15%  cpuidle.C1.usage
     23.10           +22.9%      28.38 ±  9%  boot-time.boot
     18.38           +23.2%      22.64 ± 12%  boot-time.dhcp
      5216            +5.0%       5478 ±  2%  boot-time.idle
    963.76 ± 44%    +109.7%       2021 ± 34%  irq_exception_noise.__do_page_fault.sum
      6.33 ± 14%    +726.3%      52.33 ± 62%  irq_exception_noise.irq_time
     56524 ±  7%     -18.8%      45915 ±  4%  irq_exception_noise.softirq_time
 6.001e+08 ± 13%     +31.4%  7.887e+08 ±  4%  time.involuntary_context_switches
     18003 ±  4%     +10.9%      19956        time.minor_page_faults
  1.29e+10            -2.5%  1.258e+10        time.voluntary_context_switches
      1386 ±  7%     +15.4%       1600 ± 11%  slabinfo.scsi_sense_cache.active_objs
      1386 ±  7%     +15.4%       1600 ± 11%  slabinfo.scsi_sense_cache.num_objs
      1427 ±  5%      -8.9%       1299 ±  2%  slabinfo.task_group.active_objs
      1427 ±  5%      -8.9%       1299 ±  2%  slabinfo.task_group.num_objs
     65519 ± 12%     +20.6%      79014 ± 16%  numa-meminfo.node0.SUnreclaim
      8484           -11.9%       7475 ±  7%  numa-meminfo.node1.KernelStack
      9264 ± 26%     -33.7%       6146 ±  7%  numa-meminfo.node1.Mapped
      2138 ± 61%    +373.5%      10127 ± 92%  numa-meminfo.node3.Inactive
      2059 ± 61%    +387.8%      10046 ± 93%  numa-meminfo.node3.Inactive(anon)
     16379 ± 12%     +20.6%      19752 ± 16%  numa-vmstat.node0.nr_slab_unreclaimable
      8483           -11.9%       7474 ±  7%  numa-vmstat.node1.nr_kernel_stack
      6250 ± 29%     -42.8%       3575 ± 24%  numa-vmstat.node2
      3798 ± 17%     +63.7%       6218 ±  5%  numa-vmstat.node3
    543.00 ± 61%    +368.1%       2541 ± 91%  numa-vmstat.node3.nr_inactive_anon
    543.33 ± 61%    +367.8%       2541 ± 91%  numa-vmstat.node3.nr_zone_inactive_anon
 4.138e+13            -1.1%   4.09e+13        perf-stat.branch-instructions
 6.569e+11            -2.0%  6.441e+11        perf-stat.branch-misses
 2.645e+10            -1.2%  2.613e+10        perf-stat.context-switches
      1.21            +1.2%       1.23        perf-stat.cpi
    153343 ±  2%     -12.1%     134776        perf-stat.cpu-migrations
 5.966e+13            -1.3%  5.889e+13        perf-stat.dTLB-loads
 3.736e+13            -1.2%   3.69e+13        perf-stat.dTLB-stores
      5.85 ± 15%      +8.8       14.67 ±  9%  perf-stat.iTLB-load-miss-rate%
 3.736e+09 ± 17%    +161.3%   9.76e+09 ± 11%  perf-stat.iTLB-load-misses
 5.987e+10            -5.4%  5.667e+10        perf-stat.iTLB-loads
 2.079e+14            -1.2%  2.054e+14        perf-stat.instructions
     57547 ± 18%     -62.9%      21340 ± 11%  perf-stat.instructions-per-iTLB-miss
      0.82            -1.2%       0.81        perf-stat.ipc
  27502531 ±  8%      +9.5%   30122136 ±  3%  perf-stat.node-store-misses
      1449 ± 27%     -34.6%     948.85        sched_debug.cfs_rq:/.load.min
    319416 ±115%    -188.5%    -282549        sched_debug.cfs_rq:/.spread0.avg
    657044 ± 55%     -88.3%      76887 ± 23%  sched_debug.cfs_rq:/.spread0.max
  -1525243           +54.6%   -2357898        sched_debug.cfs_rq:/.spread0.min
    101614 ±  6%     +30.6%     132713 ± 19%  sched_debug.cpu.avg_idle.stddev
     11.54 ± 41%     -61.2%       4.48        sched_debug.cpu.cpu_load[1].avg
      1369 ± 67%     -98.5%      20.67 ± 48%  sched_debug.cpu.cpu_load[1].max
     99.29 ± 67%     -97.6%       2.35 ± 26%  sched_debug.cpu.cpu_load[1].stddev
      9.58 ± 38%     -55.2%       4.29        sched_debug.cpu.cpu_load[2].avg
      1024 ± 68%     -98.5%      15.27 ± 36%  sched_debug.cpu.cpu_load[2].max
     74.51 ± 67%     -97.3%       1.99 ± 15%  sched_debug.cpu.cpu_load[2].stddev
      7.37 ± 29%     -42.0%       4.28        sched_debug.cpu.cpu_load[3].avg
    600.58 ± 68%     -97.9%      12.48 ± 20%  sched_debug.cpu.cpu_load[3].max
     43.98 ± 66%     -95.8%       1.83 ±  5%  sched_debug.cpu.cpu_load[3].stddev
      5.95 ± 19%     -28.1%       4.28        sched_debug.cpu.cpu_load[4].avg
    325.39 ± 67%     -96.4%      11.67 ± 10%  sched_debug.cpu.cpu_load[4].max
     24.19 ± 65%     -92.5%       1.81 ±  3%  sched_debug.cpu.cpu_load[4].stddev
    907.23 ±  4%     -14.1%     779.70 ± 10%  sched_debug.cpu.nr_load_updates.stddev
      0.00 ± 83%    +122.5%       0.00        sched_debug.rt_rq:/.rt_time.min
      8.49 ±  2%      -0.3        8.21 ±  2%  perf-profile.calltrace.cycles-pp.dequeue_task_fair.__schedule.schedule.pipe_wait.pipe_read
     57.28            -0.3       57.01        perf-profile.calltrace.cycles-pp.read
      5.06            -0.2        4.85        perf-profile.calltrace.cycles-pp.select_task_rq_fair.try_to_wake_up.autoremove_wake_function.__wake_up_common.__wake_up_common_lock
      4.98            -0.2        4.78        perf-profile.calltrace.cycles-pp.__switch_to.read
      3.55            -0.2        3.39 ±  2%  perf-profile.calltrace.cycles-pp.syscall_return_via_sysret.read
      2.72            -0.1        2.60        perf-profile.calltrace.cycles-pp.reweight_entity.enqueue_task_fair.ttwu_do_activate.try_to_wake_up.autoremove_wake_function
      2.67            -0.1        2.57 ±  2%  perf-profile.calltrace.cycles-pp.reweight_entity.dequeue_task_fair.__schedule.schedule.pipe_wait
      3.40            -0.1        3.31        perf-profile.calltrace.cycles-pp.syscall_return_via_sysret.write
      3.77            -0.1        3.68        perf-profile.calltrace.cycles-pp.select_idle_sibling.select_task_rq_fair.try_to_wake_up.autoremove_wake_function.__wake_up_common
      1.95            -0.1        1.88        perf-profile.calltrace.cycles-pp.copy_page_to_iter.pipe_read.__vfs_read.vfs_read.ksys_read
      2.19            -0.1        2.13        perf-profile.calltrace.cycles-pp.__switch_to_asm.read
      1.30            -0.1        1.25        perf-profile.calltrace.cycles-pp.update_curr.reweight_entity.enqueue_task_fair.ttwu_do_activate.try_to_wake_up
      1.27            -0.1        1.22 ±  2%  perf-profile.calltrace.cycles-pp.update_curr.reweight_entity.dequeue_task_fair.__schedule.schedule
      2.29            -0.0        2.24        perf-profile.calltrace.cycles-pp.load_new_mm_cr3.switch_mm_irqs_off.__schedule.schedule.pipe_wait
      0.96            -0.0        0.92        perf-profile.calltrace.cycles-pp.__calc_delta.update_curr.reweight_entity.dequeue_task_fair.__schedule
      0.85            -0.0        0.81 ±  3%  perf-profile.calltrace.cycles-pp.cpumask_next_wrap.select_idle_sibling.select_task_rq_fair.try_to_wake_up.autoremove_wake_function
      1.63            -0.0        1.59        perf-profile.calltrace.cycles-pp.native_write_msr.read
      0.72            -0.0        0.69        perf-profile.calltrace.cycles-pp.copyout.copy_page_to_iter.pipe_read.__vfs_read.vfs_read
      0.65 ±  2%      -0.0        0.62        perf-profile.calltrace.cycles-pp._raw_spin_lock_irqsave.try_to_wake_up.autoremove_wake_function.__wake_up_common.__wake_up_common_lock
      0.61            -0.0        0.58 ±  2%  perf-profile.calltrace.cycles-pp.find_next_bit.cpumask_next_wrap.select_idle_sibling.select_task_rq_fair.try_to_wake_up
      0.88            -0.0        0.85        perf-profile.calltrace.cycles-pp.touch_atime.pipe_read.__vfs_read.vfs_read.ksys_read
      0.80            -0.0        0.77 ±  2%  perf-profile.calltrace.cycles-pp.___perf_sw_event.__schedule.schedule.pipe_wait.pipe_read
      0.82            -0.0        0.79        perf-profile.calltrace.cycles-pp.prepare_to_wait.pipe_wait.pipe_read.__vfs_read.vfs_read
      0.72            -0.0        0.70        perf-profile.calltrace.cycles-pp.mutex_lock.pipe_write.__vfs_write.vfs_write.ksys_write
      0.56 ±  2%      -0.0        0.53        perf-profile.calltrace.cycles-pp.update_rq_clock.try_to_wake_up.autoremove_wake_function.__wake_up_common.__wake_up_common_lock
      0.83            -0.0        0.81        perf-profile.calltrace.cycles-pp.__wake_up_common_lock.pipe_read.__vfs_read.vfs_read.ksys_read
     42.40            +0.3       42.69        perf-profile.calltrace.cycles-pp.write
     31.80            +0.4       32.18        perf-profile.calltrace.cycles-pp.__vfs_read.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
     24.35            +0.5       24.84        perf-profile.calltrace.cycles-pp.pipe_wait.pipe_read.__vfs_read.vfs_read.ksys_read
     20.36            +0.6       20.92 ±  2%  perf-profile.calltrace.cycles-pp.try_to_wake_up.autoremove_wake_function.__wake_up_common.__wake_up_common_lock.pipe_write
     22.01            +0.6       22.58        perf-profile.calltrace.cycles-pp.schedule.pipe_wait.pipe_read.__vfs_read.vfs_read
     21.87            +0.6       22.46        perf-profile.calltrace.cycles-pp.__schedule.schedule.pipe_wait.pipe_read.__vfs_read
      3.15 ± 11%      +1.0        4.12 ± 14%  perf-profile.calltrace.cycles-pp.ttwu_do_wakeup.try_to_wake_up.autoremove_wake_function.__wake_up_common.__wake_up_common_lock
      1.07 ± 34%      +1.1        2.12 ± 31%  perf-profile.calltrace.cycles-pp.tracing_record_taskinfo_sched_switch.__schedule.schedule.pipe_wait.pipe_read
      0.66 ± 75%      +1.1        1.72 ± 37%  perf-profile.calltrace.cycles-pp.trace_save_cmdline.tracing_record_taskinfo.ttwu_do_wakeup.try_to_wake_up.autoremove_wake_function
      0.75 ± 74%      +1.1        1.88 ± 34%  perf-profile.calltrace.cycles-pp.tracing_record_taskinfo.ttwu_do_wakeup.try_to_wake_up.autoremove_wake_function.__wake_up_common
      0.69 ± 76%      +1.2        1.85 ± 36%  perf-profile.calltrace.cycles-pp.trace_save_cmdline.tracing_record_taskinfo_sched_switch.__schedule.schedule.pipe_wait
      8.73 ±  2%      -0.3        8.45        perf-profile.children.cycles-pp.dequeue_task_fair
     57.28            -0.3       57.01        perf-profile.children.cycles-pp.read
      6.95            -0.2        6.70        perf-profile.children.cycles-pp.syscall_return_via_sysret
      5.57            -0.2        5.35        perf-profile.children.cycles-pp.reweight_entity
      5.26            -0.2        5.05        perf-profile.children.cycles-pp.select_task_rq_fair
      5.19            -0.2        4.99        perf-profile.children.cycles-pp.__switch_to
      4.90            -0.2        4.73 ±  2%  perf-profile.children.cycles-pp.update_curr
      1.27            -0.1        1.13 ±  8%  perf-profile.children.cycles-pp.fsnotify
      3.92            -0.1        3.83        perf-profile.children.cycles-pp.select_idle_sibling
      2.01            -0.1        1.93        perf-profile.children.cycles-pp.__calc_delta
      2.14            -0.1        2.06        perf-profile.children.cycles-pp.copy_page_to_iter
      1.58            -0.1        1.51        perf-profile.children.cycles-pp._raw_spin_unlock_irqrestore
      2.90            -0.1        2.84        perf-profile.children.cycles-pp.update_cfs_group
      1.93            -0.1        1.87        perf-profile.children.cycles-pp._raw_spin_lock_irqsave
      2.35            -0.1        2.29        perf-profile.children.cycles-pp.__switch_to_asm
      1.33            -0.1        1.27 ±  3%  perf-profile.children.cycles-pp.cpumask_next_wrap
      2.57            -0.1        2.52        perf-profile.children.cycles-pp.load_new_mm_cr3
      1.53            -0.1        1.47 ±  2%  perf-profile.children.cycles-pp.__fdget_pos
      1.11            -0.0        1.07 ±  2%  perf-profile.children.cycles-pp.find_next_bit
      1.18            -0.0        1.14        perf-profile.children.cycles-pp.update_rq_clock
      0.88            -0.0        0.83        perf-profile.children.cycles-pp.copy_user_generic_unrolled
      1.70            -0.0        1.65        perf-profile.children.cycles-pp.native_write_msr
      0.97            -0.0        0.93 ±  2%  perf-profile.children.cycles-pp.account_entity_dequeue
      0.59            -0.0        0.56        perf-profile.children.cycles-pp.finish_task_switch
      0.91            -0.0        0.88        perf-profile.children.cycles-pp.touch_atime
      0.69            -0.0        0.65        perf-profile.children.cycles-pp.account_entity_enqueue
      2.13            -0.0        2.09        perf-profile.children.cycles-pp.mutex_lock
      0.32 ±  3%      -0.0        0.29 ±  4%  perf-profile.children.cycles-pp.__sb_start_write
      0.84            -0.0        0.81 ±  2%  perf-profile.children.cycles-pp.___perf_sw_event
      0.89            -0.0        0.87        perf-profile.children.cycles-pp.prepare_to_wait
      0.73            -0.0        0.71        perf-profile.children.cycles-pp.copyout
      0.31 ±  2%      -0.0        0.28 ±  3%  perf-profile.children.cycles-pp.__list_del_entry_valid
      0.46 ±  2%      -0.0        0.44        perf-profile.children.cycles-pp.anon_pipe_buf_release
      0.38            -0.0        0.36 ±  3%  perf-profile.children.cycles-pp.idle_cpu
      0.32            -0.0        0.30 ±  2%  perf-profile.children.cycles-pp.__x64_sys_read
      0.21 ±  2%      -0.0        0.20 ±  2%  perf-profile.children.cycles-pp.deactivate_task
      0.13            -0.0        0.12 ±  4%  perf-profile.children.cycles-pp.timespec_trunc
      0.09            -0.0        0.08        perf-profile.children.cycles-pp.iov_iter_init
      0.08            -0.0        0.07        perf-profile.children.cycles-pp.native_load_tls
      0.11 ±  4%      +0.0        0.12        perf-profile.children.cycles-pp.tick_sched_timer
      0.08 ±  5%      +0.0        0.10 ±  4%  perf-profile.children.cycles-pp.finish_wait
      0.38 ±  2%      +0.0        0.40 ±  2%  perf-profile.children.cycles-pp.file_update_time
      0.31            +0.0        0.33 ±  2%  perf-profile.children.cycles-pp.smp_apic_timer_interrupt
      0.24 ±  3%      +0.0        0.26 ±  3%  perf-profile.children.cycles-pp.rcu_all_qs
      0.39            +0.0        0.41        perf-profile.children.cycles-pp._cond_resched
      0.05            +0.0        0.07 ±  6%  perf-profile.children.cycles-pp.default_wake_function
      0.23 ±  2%      +0.0        0.26 ±  3%  perf-profile.children.cycles-pp.current_time
      0.30            +0.0        0.35 ±  2%  perf-profile.children.cycles-pp.generic_pipe_buf_confirm
      0.52            +0.1        0.58        perf-profile.children.cycles-pp.entry_SYSCALL_64_stage2
      0.00            +0.1        0.08 ±  5%  perf-profile.children.cycles-pp.hrtick_update
     42.40            +0.3       42.69        perf-profile.children.cycles-pp.write
     31.86            +0.4       32.26        perf-profile.children.cycles-pp.__vfs_read
     24.40            +0.5       24.89        perf-profile.children.cycles-pp.pipe_wait
     20.40            +0.6       20.96 ±  2%  perf-profile.children.cycles-pp.try_to_wake_up
     22.30            +0.6       22.89        perf-profile.children.cycles-pp.schedule
     22.22            +0.6       22.84        perf-profile.children.cycles-pp.__schedule
      0.99 ± 36%      +0.9        1.94 ± 32%  perf-profile.children.cycles-pp.tracing_record_taskinfo
      3.30 ± 10%      +1.0        4.27 ± 13%  perf-profile.children.cycles-pp.ttwu_do_wakeup
      1.14 ± 31%      +1.1        2.24 ± 29%  perf-profile.children.cycles-pp.tracing_record_taskinfo_sched_switch
      1.59 ± 46%      +2.0        3.60 ± 36%  perf-profile.children.cycles-pp.trace_save_cmdline
      6.95            -0.2        6.70        perf-profile.self.cycles-pp.syscall_return_via_sysret
      5.19            -0.2        4.99        perf-profile.self.cycles-pp.__switch_to
      1.27            -0.1        1.12 ±  8%  perf-profile.self.cycles-pp.fsnotify
      1.49            -0.1        1.36        perf-profile.self.cycles-pp.select_task_rq_fair
      2.47            -0.1        2.37 ±  2%  perf-profile.self.cycles-pp.reweight_entity
      0.29            -0.1        0.19 ±  2%  perf-profile.self.cycles-pp.ksys_read
      1.50            -0.1        1.42        perf-profile.self.cycles-pp._raw_spin_unlock_irqrestore
      2.01            -0.1        1.93        perf-profile.self.cycles-pp.__calc_delta
      1.93            -0.1        1.86        perf-profile.self.cycles-pp._raw_spin_lock_irqsave
      1.47            -0.1        1.40        perf-profile.self.cycles-pp.dequeue_task_fair
      2.90            -0.1        2.84        perf-profile.self.cycles-pp.update_cfs_group
      1.29            -0.1        1.23        perf-profile.self.cycles-pp.do_syscall_64
      2.57            -0.1        2.52        perf-profile.self.cycles-pp.load_new_mm_cr3
      2.28            -0.1        2.23        perf-profile.self.cycles-pp.__switch_to_asm
      1.80            -0.1        1.75        perf-profile.self.cycles-pp.select_idle_sibling
      1.11            -0.0        1.07 ±  2%  perf-profile.self.cycles-pp.find_next_bit
      0.87            -0.0        0.83        perf-profile.self.cycles-pp.copy_user_generic_unrolled
      0.43            -0.0        0.39 ±  2%  perf-profile.self.cycles-pp.dequeue_entity
      1.70            -0.0        1.65        perf-profile.self.cycles-pp.native_write_msr
      0.92            -0.0        0.88 ±  2%  perf-profile.self.cycles-pp.account_entity_dequeue
      0.48            -0.0        0.44        perf-profile.self.cycles-pp.finish_task_switch
      0.77            -0.0        0.74        perf-profile.self.cycles-pp.___perf_sw_event
      0.66            -0.0        0.63        perf-profile.self.cycles-pp.account_entity_enqueue
      0.46 ±  2%      -0.0        0.43 ±  2%  perf-profile.self.cycles-pp.anon_pipe_buf_release
      0.32 ±  3%      -0.0        0.29 ±  4%  perf-profile.self.cycles-pp.__sb_start_write
      0.31 ±  2%      -0.0        0.28 ±  3%  perf-profile.self.cycles-pp.__list_del_entry_valid
      0.38            -0.0        0.36 ±  3%  perf-profile.self.cycles-pp.idle_cpu
      0.19 ±  4%      -0.0        0.17 ±  2%  perf-profile.self.cycles-pp.__fdget_pos
      0.50            -0.0        0.48        perf-profile.self.cycles-pp.__atime_needs_update
      0.23 ±  2%      -0.0        0.21 ±  3%  perf-profile.self.cycles-pp.touch_atime
      0.31            -0.0        0.30        perf-profile.self.cycles-pp.__x64_sys_read
      0.21 ±  2%      -0.0        0.20 ±  2%  perf-profile.self.cycles-pp.deactivate_task
      0.21 ±  2%      -0.0        0.19        perf-profile.self.cycles-pp.check_preempt_curr
      0.40            -0.0        0.39        perf-profile.self.cycles-pp.autoremove_wake_function
      0.40            -0.0        0.38        perf-profile.self.cycles-pp.copy_user_enhanced_fast_string
      0.27            -0.0        0.26        perf-profile.self.cycles-pp.pipe_wait
      0.13            -0.0        0.12 ±  4%  perf-profile.self.cycles-pp.timespec_trunc
      0.22 ±  2%      -0.0        0.20 ±  2%  perf-profile.self.cycles-pp.put_prev_entity
      0.09            -0.0        0.08        perf-profile.self.cycles-pp.iov_iter_init
      0.08            -0.0        0.07        perf-profile.self.cycles-pp.native_load_tls
      0.11            -0.0        0.10        perf-profile.self.cycles-pp.schedule
      0.12 ±  4%      +0.0        0.13        perf-profile.self.cycles-pp.copyin
      0.08 ±  5%      +0.0        0.10 ±  4%  perf-profile.self.cycles-pp.finish_wait
      0.18            +0.0        0.20 ±  2%  perf-profile.self.cycles-pp.ttwu_do_activate
      0.28 ±  2%      +0.0        0.30 ±  2%  perf-profile.self.cycles-pp._cond_resched
      0.24 ±  3%      +0.0        0.26 ±  3%  perf-profile.self.cycles-pp.rcu_all_qs
      0.05            +0.0        0.07 ±  6%  perf-profile.self.cycles-pp.default_wake_function
      0.08 ± 14%      +0.0        0.11 ± 14%  perf-profile.self.cycles-pp.tracing_record_taskinfo_sched_switch
      0.51            +0.0        0.55 ±  4%  perf-profile.self.cycles-pp.vfs_write
      0.30            +0.0        0.35 ±  2%  perf-profile.self.cycles-pp.generic_pipe_buf_confirm
      0.52            +0.1        0.58        perf-profile.self.cycles-pp.entry_SYSCALL_64_stage2
      0.00            +0.1        0.08 ±  5%  perf-profile.self.cycles-pp.hrtick_update
      1.97            +0.1        2.07 ±  2%  perf-profile.self.cycles-pp.switch_mm_irqs_off
      1.59 ± 46%      +2.0        3.60 ± 36%  perf-profile.self.cycles-pp.trace_save_cmdline

=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/nr_task/thp_enabled/test/cpufreq_governor:
  lkp-skl-4sp1/will-it-scale/debian-x86_64-2018-04-03.cgz/x86_64-rhel-7.2/gcc-7/100%/never/brk1/performance

commit: 
  ba98a1cdad71d259a194461b3a61471b49b14df1
  a7a8993bfe3ccb54ad468b9f1799649e4ad1ff12

ba98a1cdad71d259 a7a8993bfe3ccb54ad468b9f17 
---------------- -------------------------- 
       fail:runs  %reproduction    fail:runs
           |             |             |    
           :3           33%           1:3     kmsg.pstore:crypto_comp_decompress_failed,ret=
           :3           33%           1:3     kmsg.pstore:decompression_failed
         %stddev     %change         %stddev
             \          |                \  
    997317            -2.0%     977778        will-it-scale.per_process_ops
    957.00            -7.9%     881.00 ±  3%  will-it-scale.per_thread_ops
     18.42 ±  3%      -8.2%      16.90        will-it-scale.time.user_time
 1.917e+08            -2.0%  1.879e+08        will-it-scale.workload
     18.42 ±  3%      -8.2%      16.90        time.user_time
      0.30 ± 11%     -36.7%       0.19 ± 11%  turbostat.Pkg%pc2
     57539 ± 51%    +140.6%     138439 ± 31%  meminfo.CmaFree
    410877 ± 11%     -22.1%     320082 ± 22%  meminfo.DirectMap4k
    343575 ± 27%     +71.3%     588703 ± 31%  numa-numastat.node0.local_node
    374176 ± 24%     +63.3%     611007 ± 27%  numa-numastat.node0.numa_hit
   1056347 ±  4%     -39.9%     634843 ± 38%  numa-numastat.node3.local_node
   1060682 ±  4%     -39.0%     646862 ± 35%  numa-numastat.node3.numa_hit
     14383 ± 51%    +140.6%      34608 ± 31%  proc-vmstat.nr_free_cma
    179.00            +2.4%     183.33        proc-vmstat.nr_inactive_file
    179.00            +2.4%     183.33        proc-vmstat.nr_zone_inactive_file
    564483 ±  3%     -38.0%     350064 ± 36%  proc-vmstat.pgalloc_movable
   1811959           +10.8%    2008488 ±  5%  proc-vmstat.pgalloc_normal
      7153 ± 42%     -94.0%     431.33 ±119%  latency_stats.max.pipe_write.__vfs_write.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
      6627 ±141%    +380.5%      31843 ±110%  latency_stats.max.call_rwsem_down_write_failed_killable.do_mprotect_pkey.__x64_sys_mprotect.do_syscall_64.entry_SYSCALL_64_after_hwframe
     15244 ± 31%     -99.9%      15.00 ±141%  latency_stats.sum.call_rwsem_down_read_failed.__do_page_fault.do_page_fault.page_fault.__get_user_8.exit_robust_list.mm_release.do_exit.do_group_exit.get_signal.do_signal.exit_to_usermode_loop
      4301 ±117%     -83.7%     700.33 ±  6%  latency_stats.sum.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_getattr.__nfs_revalidate_inode.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_lookupat
     12153 ± 28%     -83.1%       2056 ± 70%  latency_stats.sum.pipe_write.__vfs_write.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
      6772 ±141%   +1105.8%      81665 ±127%  latency_stats.sum.call_rwsem_down_write_failed_killable.do_mprotect_pkey.__x64_sys_mprotect.do_syscall_64.entry_SYSCALL_64_after_hwframe
 2.465e+13            -1.3%  2.434e+13        perf-stat.branch-instructions
 2.691e+11            -2.1%  2.635e+11        perf-stat.branch-misses
 3.402e+13            -1.4%  3.355e+13        perf-stat.dTLB-loads
 1.694e+13            +1.4%  1.718e+13        perf-stat.dTLB-stores
      1.75 ± 50%      +4.7        6.45 ± 11%  perf-stat.iTLB-load-miss-rate%
 4.077e+08 ± 48%    +232.3%  1.355e+09 ± 11%  perf-stat.iTLB-load-misses
  2.31e+10 ±  2%     -14.9%  1.965e+10 ±  3%  perf-stat.iTLB-loads
 1.163e+14            -1.6%  1.144e+14        perf-stat.instructions
    346171 ± 36%     -75.3%      85575 ± 11%  perf-stat.instructions-per-iTLB-miss
 6.174e+08 ±  2%      -9.5%  5.589e+08        perf-stat.node-store-misses
    595.00 ± 10%     +31.4%     782.00 ±  3%  slabinfo.Acpi-State.active_objs
    595.00 ± 10%     +31.4%     782.00 ±  3%  slabinfo.Acpi-State.num_objs
      2831 ±  3%     -14.0%       2434 ±  5%  slabinfo.avtab_node.active_objs
      2831 ±  3%     -14.0%       2434 ±  5%  slabinfo.avtab_node.num_objs
    934.00           -10.9%     832.33 ±  5%  slabinfo.inotify_inode_mark.active_objs
    934.00           -10.9%     832.33 ±  5%  slabinfo.inotify_inode_mark.num_objs
      1232 ±  4%     +13.4%       1397 ±  6%  slabinfo.nsproxy.active_objs
      1232 ±  4%     +13.4%       1397 ±  6%  slabinfo.nsproxy.num_objs
    499.67 ± 12%     +24.8%     623.67 ± 10%  slabinfo.secpath_cache.active_objs
    499.67 ± 12%     +24.8%     623.67 ± 10%  slabinfo.secpath_cache.num_objs
     31393 ± 84%    +220.1%     100477 ± 21%  numa-meminfo.node0.Active
     31393 ± 84%    +220.1%     100477 ± 21%  numa-meminfo.node0.Active(anon)
     30013 ± 85%    +232.1%      99661 ± 21%  numa-meminfo.node0.AnonPages
     21603 ± 34%     -85.0%       3237 ±100%  numa-meminfo.node0.Inactive
     21528 ± 34%     -85.0%       3237 ±100%  numa-meminfo.node0.Inactive(anon)
     10247 ± 35%     -46.4%       5495        numa-meminfo.node0.Mapped
     35388 ± 14%     -41.6%      20670 ± 15%  numa-meminfo.node0.SReclaimable
     22911 ± 29%     -82.3%       4057 ± 84%  numa-meminfo.node0.Shmem
    117387 ±  9%     -22.5%      90986 ± 12%  numa-meminfo.node0.Slab
     68863 ± 67%     +77.7%     122351 ± 13%  numa-meminfo.node1.Active
     68863 ± 67%     +77.7%     122351 ± 13%  numa-meminfo.node1.Active(anon)
    228376           +22.3%     279406 ± 17%  numa-meminfo.node1.FilePages
      1481 ±116%   +1062.1%      17218 ± 39%  numa-meminfo.node1.Inactive
      1481 ±116%   +1062.0%      17216 ± 39%  numa-meminfo.node1.Inactive(anon)
      6593 ±  2%     +11.7%       7367 ±  3%  numa-meminfo.node1.KernelStack
    596227 ±  8%     +18.0%     703748 ±  4%  numa-meminfo.node1.MemUsed
     15298 ± 12%     +88.5%      28843 ± 36%  numa-meminfo.node1.SReclaimable
     52718 ±  9%     +21.0%      63810 ± 11%  numa-meminfo.node1.SUnreclaim
      1808 ± 97%   +2723.8%      51054 ± 97%  numa-meminfo.node1.Shmem
     68017 ±  5%     +36.2%      92654 ± 18%  numa-meminfo.node1.Slab
    125541 ± 29%     -64.9%      44024 ± 98%  numa-meminfo.node3.Active
    125137 ± 29%     -65.0%      43823 ± 98%  numa-meminfo.node3.Active(anon)
     93173 ± 25%     -87.8%      11381 ± 20%  numa-meminfo.node3.AnonPages
      9150 ±  5%      -9.3%       8301 ±  8%  numa-meminfo.node3.KernelStack
      7848 ± 84%    +220.0%      25118 ± 21%  numa-vmstat.node0.nr_active_anon
      7503 ± 85%    +232.1%      24914 ± 21%  numa-vmstat.node0.nr_anon_pages
      5381 ± 34%     -85.0%     809.00 ±100%  numa-vmstat.node0.nr_inactive_anon
      2559 ± 35%     -46.4%       1372        numa-vmstat.node0.nr_mapped
      5727 ± 29%     -82.3%       1014 ± 84%  numa-vmstat.node0.nr_shmem
      8846 ± 14%     -41.6%       5167 ± 15%  numa-vmstat.node0.nr_slab_reclaimable
      7848 ± 84%    +220.0%      25118 ± 21%  numa-vmstat.node0.nr_zone_active_anon
      5381 ± 34%     -85.0%     809.00 ±100%  numa-vmstat.node0.nr_zone_inactive_anon
      4821 ±  2%     +30.3%       6283 ± 15%  numa-vmstat.node1
     17215 ± 67%     +77.7%      30591 ± 13%  numa-vmstat.node1.nr_active_anon
     57093           +22.3%      69850 ± 17%  numa-vmstat.node1.nr_file_pages
    370.00 ±116%   +1061.8%       4298 ± 39%  numa-vmstat.node1.nr_inactive_anon
      6593 ±  2%     +11.7%       7366 ±  3%  numa-vmstat.node1.nr_kernel_stack
    451.67 ± 97%   +2725.6%      12762 ± 97%  numa-vmstat.node1.nr_shmem
      3824 ± 12%     +88.6%       7211 ± 36%  numa-vmstat.node1.nr_slab_reclaimable
     13179 ±  9%     +21.0%      15952 ± 11%  numa-vmstat.node1.nr_slab_unreclaimable
     17215 ± 67%     +77.7%      30591 ± 13%  numa-vmstat.node1.nr_zone_active_anon
    370.00 ±116%   +1061.8%       4298 ± 39%  numa-vmstat.node1.nr_zone_inactive_anon
    364789 ± 12%     +62.8%     593926 ± 34%  numa-vmstat.node1.numa_hit
    239539 ± 19%     +95.4%     468113 ± 43%  numa-vmstat.node1.numa_local
     71.00 ± 28%     +42.3%     101.00        numa-vmstat.node2.nr_mlock
     31285 ± 29%     -65.0%      10960 ± 98%  numa-vmstat.node3.nr_active_anon
     23292 ± 25%     -87.8%       2844 ± 19%  numa-vmstat.node3.nr_anon_pages
     14339 ± 52%    +141.1%      34566 ± 32%  numa-vmstat.node3.nr_free_cma
      9151 ±  5%      -9.3%       8299 ±  8%  numa-vmstat.node3.nr_kernel_stack
     31305 ± 29%     -64.9%      10975 ± 98%  numa-vmstat.node3.nr_zone_active_anon
    930131 ±  3%     -35.9%     596006 ± 34%  numa-vmstat.node3.numa_hit
    836455 ±  3%     -40.9%     493947 ± 44%  numa-vmstat.node3.numa_local
     75182 ± 58%     -83.8%      12160 ±  2%  sched_debug.cfs_rq:/.load.max
      6.65 ±  5%     -10.6%       5.94 ±  6%  sched_debug.cfs_rq:/.load_avg.avg
      0.16 ±  7%     +22.6%       0.20 ± 12%  sched_debug.cfs_rq:/.nr_running.stddev
      5.58 ± 24%    +427.7%      29.42 ± 93%  sched_debug.cfs_rq:/.nr_spread_over.max
      0.54 ± 15%    +306.8%       2.19 ± 86%  sched_debug.cfs_rq:/.nr_spread_over.stddev
      1.05 ± 25%     -65.1%       0.37 ± 71%  sched_debug.cfs_rq:/.removed.load_avg.avg
      9.62 ± 11%     -50.7%       4.74 ± 70%  sched_debug.cfs_rq:/.removed.load_avg.stddev
     48.70 ± 25%     -65.1%      17.02 ± 71%  sched_debug.cfs_rq:/.removed.runnable_sum.avg
    444.31 ± 11%     -50.7%     219.26 ± 70%  sched_debug.cfs_rq:/.removed.runnable_sum.stddev
      0.47 ± 13%     -60.9%       0.19 ± 71%  sched_debug.cfs_rq:/.removed.util_avg.avg
      4.47 ±  4%     -46.5%       2.39 ± 70%  sched_debug.cfs_rq:/.removed.util_avg.stddev
      1.64 ±  7%     +22.1%       2.00 ± 13%  sched_debug.cfs_rq:/.runnable_load_avg.stddev
     74653 ± 59%     -84.4%      11676        sched_debug.cfs_rq:/.runnable_weight.max
   -119169          -491.3%     466350 ± 27%  sched_debug.cfs_rq:/.spread0.avg
    517161 ± 30%    +145.8%    1271292 ± 23%  sched_debug.cfs_rq:/.spread0.max
    624.79 ±  5%     -14.2%     535.76 ±  7%  sched_debug.cfs_rq:/.util_est_enqueued.avg
    247.91 ± 32%     -99.8%       0.48 ±  8%  sched_debug.cfs_rq:/.util_est_enqueued.min
    179704 ±  3%     +30.4%     234297 ± 16%  sched_debug.cpu.avg_idle.stddev
      1.56 ±  9%     +24.4%       1.94 ± 14%  sched_debug.cpu.cpu_load[0].stddev
      1.50 ±  6%     +27.7%       1.91 ± 14%  sched_debug.cpu.cpu_load[1].stddev
      1.45 ±  3%     +30.8%       1.90 ± 14%  sched_debug.cpu.cpu_load[2].stddev
      1.43 ±  3%     +36.1%       1.95 ± 11%  sched_debug.cpu.cpu_load[3].stddev
      1.55 ±  7%     +43.5%       2.22 ±  7%  sched_debug.cpu.cpu_load[4].stddev
     10004 ±  3%     -11.6%       8839 ±  3%  sched_debug.cpu.curr->pid.avg
      1146 ± 26%     +52.2%       1745 ±  7%  sched_debug.cpu.curr->pid.min
      3162 ±  6%     +25.4%       3966 ± 11%  sched_debug.cpu.curr->pid.stddev
    403738 ±  3%     -11.7%     356696 ±  7%  sched_debug.cpu.nr_switches.max
      0.08 ± 21%     +78.2%       0.14 ± 14%  sched_debug.cpu.nr_uninterruptible.avg
    404435 ±  3%     -11.8%     356732 ±  7%  sched_debug.cpu.sched_count.max
      4.17            -0.3        3.87        perf-profile.calltrace.cycles-pp.kmem_cache_alloc.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      2.40            -0.2        2.17        perf-profile.calltrace.cycles-pp.vma_compute_subtree_gap.__vma_link_rb.vma_link.do_brk_flags.__x64_sys_brk
      7.58            -0.2        7.36        perf-profile.calltrace.cycles-pp.perf_event_mmap.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
     15.00            -0.2       14.81        perf-profile.calltrace.cycles-pp.syscall_return_via_sysret.brk
      7.83            -0.2        7.66        perf-profile.calltrace.cycles-pp.unmap_vmas.unmap_region.do_munmap.__x64_sys_brk.do_syscall_64
     28.66            -0.1       28.51        perf-profile.calltrace.cycles-pp.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe.brk
      2.15            -0.1        2.03        perf-profile.calltrace.cycles-pp.vma_compute_subtree_gap.do_munmap.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      1.07            -0.1        0.99        perf-profile.calltrace.cycles-pp.memcpy_erms.strlcpy.perf_event_mmap.do_brk_flags.__x64_sys_brk
      1.03            -0.1        0.95        perf-profile.calltrace.cycles-pp.kmem_cache_free.remove_vma.do_munmap.__x64_sys_brk.do_syscall_64
      7.33            -0.1        7.25        perf-profile.calltrace.cycles-pp.unmap_page_range.unmap_vmas.unmap_region.do_munmap.__x64_sys_brk
      0.76            -0.1        0.69        perf-profile.calltrace.cycles-pp.__vm_enough_memory.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
     11.85            -0.1       11.77        perf-profile.calltrace.cycles-pp.unmap_region.do_munmap.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      1.64            -0.1        1.57        perf-profile.calltrace.cycles-pp.strlcpy.perf_event_mmap.do_brk_flags.__x64_sys_brk.do_syscall_64
      1.06            -0.1        0.99        perf-profile.calltrace.cycles-pp.__indirect_thunk_start.brk
      0.73            -0.1        0.67        perf-profile.calltrace.cycles-pp.sync_mm_rss.unmap_page_range.unmap_vmas.unmap_region.do_munmap
      4.59            -0.1        4.52        perf-profile.calltrace.cycles-pp.security_vm_enough_memory_mm.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      2.82            -0.1        2.76        perf-profile.calltrace.cycles-pp.selinux_vm_enough_memory.security_vm_enough_memory_mm.do_brk_flags.__x64_sys_brk.do_syscall_64
      2.89            -0.1        2.84        perf-profile.calltrace.cycles-pp.down_write_killable.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe.brk
      3.37            -0.1        3.32        perf-profile.calltrace.cycles-pp.get_unmapped_area.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      1.99            -0.0        1.94        perf-profile.calltrace.cycles-pp.cred_has_capability.selinux_vm_enough_memory.security_vm_enough_memory_mm.do_brk_flags.__x64_sys_brk
      2.32            -0.0        2.27        perf-profile.calltrace.cycles-pp.perf_iterate_sb.perf_event_mmap.do_brk_flags.__x64_sys_brk.do_syscall_64
      1.88            -0.0        1.84        perf-profile.calltrace.cycles-pp.security_mmap_addr.get_unmapped_area.do_brk_flags.__x64_sys_brk.do_syscall_64
      0.77            -0.0        0.73        perf-profile.calltrace.cycles-pp._raw_spin_lock.unmap_page_range.unmap_vmas.unmap_region.do_munmap
      1.62            -0.0        1.59        perf-profile.calltrace.cycles-pp.memset_erms.kmem_cache_alloc.do_brk_flags.__x64_sys_brk.do_syscall_64
      0.81            -0.0        0.79        perf-profile.calltrace.cycles-pp.___might_sleep.down_write_killable.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.66            -0.0        0.64        perf-profile.calltrace.cycles-pp.arch_get_unmapped_area_topdown.brk
      0.72            +0.0        0.74        perf-profile.calltrace.cycles-pp.do_munmap.brk
      0.90            +0.0        0.93        perf-profile.calltrace.cycles-pp.___might_sleep.unmap_page_range.unmap_vmas.unmap_region.do_munmap
      4.40            +0.1        4.47        perf-profile.calltrace.cycles-pp.find_vma.do_munmap.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      1.96            +0.1        2.09        perf-profile.calltrace.cycles-pp.vmacache_find.find_vma.do_munmap.__x64_sys_brk.do_syscall_64
      0.52 ±  2%      +0.2        0.68        perf-profile.calltrace.cycles-pp.__vma_link_rb.brk
      0.35 ± 70%      +0.2        0.54 ±  2%  perf-profile.calltrace.cycles-pp.find_vma.brk
      2.20            +0.3        2.50        perf-profile.calltrace.cycles-pp.remove_vma.do_munmap.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
     64.62            +0.3       64.94        perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.brk
     60.53            +0.4       60.92        perf-profile.calltrace.cycles-pp.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe.brk
     63.20            +0.4       63.60        perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.brk
      3.73            +0.5        4.26        perf-profile.calltrace.cycles-pp.vma_link.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00            +0.6        0.56        perf-profile.calltrace.cycles-pp.free_pgtables.unmap_region.do_munmap.__x64_sys_brk.do_syscall_64
     24.54            +0.6       25.14        perf-profile.calltrace.cycles-pp.do_munmap.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe.brk
      0.00            +0.6        0.64        perf-profile.calltrace.cycles-pp.put_vma.remove_vma.do_munmap.__x64_sys_brk.do_syscall_64
      0.71            +0.6        1.36        perf-profile.calltrace.cycles-pp.__vma_rb_erase.do_munmap.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00            +0.7        0.70        perf-profile.calltrace.cycles-pp._raw_write_lock.__vma_rb_erase.do_munmap.__x64_sys_brk.do_syscall_64
      3.10            +0.7        3.82        perf-profile.calltrace.cycles-pp.__vma_link_rb.vma_link.do_brk_flags.__x64_sys_brk.do_syscall_64
      0.00            +0.8        0.76        perf-profile.calltrace.cycles-pp._raw_write_lock.__vma_link_rb.vma_link.do_brk_flags.__x64_sys_brk
      0.00            +0.8        0.85        perf-profile.calltrace.cycles-pp.__vma_merge.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      5.09            -0.5        4.62        perf-profile.children.cycles-pp.vma_compute_subtree_gap
      4.54            -0.3        4.21        perf-profile.children.cycles-pp.kmem_cache_alloc
      8.11            -0.2        7.89        perf-profile.children.cycles-pp.perf_event_mmap
      8.05            -0.2        7.85        perf-profile.children.cycles-pp.unmap_vmas
     15.01            -0.2       14.81        perf-profile.children.cycles-pp.syscall_return_via_sysret
     29.20            -0.1       29.06        perf-profile.children.cycles-pp.do_brk_flags
      1.11            -0.1        1.00        perf-profile.children.cycles-pp.kmem_cache_free
     12.28            -0.1       12.17        perf-profile.children.cycles-pp.unmap_region
      7.83            -0.1        7.74        perf-profile.children.cycles-pp.unmap_page_range
      0.87 ±  3%      -0.1        0.79        perf-profile.children.cycles-pp.__vm_enough_memory
      1.29            -0.1        1.22        perf-profile.children.cycles-pp.__indirect_thunk_start
      1.81            -0.1        1.74        perf-profile.children.cycles-pp.strlcpy
      4.65            -0.1        4.58        perf-profile.children.cycles-pp.security_vm_enough_memory_mm
      3.08            -0.1        3.02        perf-profile.children.cycles-pp.down_write_killable
      2.88            -0.1        2.82        perf-profile.children.cycles-pp.selinux_vm_enough_memory
      0.73            -0.1        0.67        perf-profile.children.cycles-pp.sync_mm_rss
      3.65            -0.1        3.59        perf-profile.children.cycles-pp.get_unmapped_area
      2.26            -0.1        2.20        perf-profile.children.cycles-pp.cred_has_capability
      1.12            -0.1        1.07        perf-profile.children.cycles-pp.memcpy_erms
      0.39            -0.0        0.35        perf-profile.children.cycles-pp.__rb_insert_augmented
      2.52            -0.0        2.48        perf-profile.children.cycles-pp.perf_iterate_sb
      2.13            -0.0        2.09        perf-profile.children.cycles-pp.security_mmap_addr
      0.55 ±  2%      -0.0        0.52        perf-profile.children.cycles-pp.unmap_single_vma
      1.62            -0.0        1.59        perf-profile.children.cycles-pp.memset_erms
      0.13 ±  3%      -0.0        0.11 ±  4%  perf-profile.children.cycles-pp.__vma_link_file
      0.80            -0.0        0.77        perf-profile.children.cycles-pp._raw_spin_lock
      0.43            -0.0        0.41        perf-profile.children.cycles-pp.strlen
      0.07 ±  6%      -0.0        0.06 ±  8%  perf-profile.children.cycles-pp.should_failslab
      0.43            -0.0        0.42        perf-profile.children.cycles-pp.may_expand_vm
      0.15            +0.0        0.16        perf-profile.children.cycles-pp.__vma_link_list
      0.45            +0.0        0.47        perf-profile.children.cycles-pp.rcu_all_qs
      0.81            +0.1        0.89        perf-profile.children.cycles-pp.free_pgtables
      6.35            +0.1        6.49        perf-profile.children.cycles-pp.find_vma
      2.28            +0.2        2.45        perf-profile.children.cycles-pp.vmacache_find
     64.66            +0.3       64.98        perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
      2.42            +0.3        2.76        perf-profile.children.cycles-pp.remove_vma
     61.77            +0.4       62.13        perf-profile.children.cycles-pp.__x64_sys_brk
     63.40            +0.4       63.79        perf-profile.children.cycles-pp.do_syscall_64
      1.27            +0.4        1.72        perf-profile.children.cycles-pp.__vma_rb_erase
      4.02            +0.5        4.53        perf-profile.children.cycles-pp.vma_link
     25.26            +0.6       25.89        perf-profile.children.cycles-pp.do_munmap
      0.00            +0.7        0.70        perf-profile.children.cycles-pp.put_vma
      3.80            +0.7        4.53        perf-profile.children.cycles-pp.__vma_link_rb
      0.00            +1.2        1.24        perf-profile.children.cycles-pp.__vma_merge
      0.00            +1.5        1.51        perf-profile.children.cycles-pp._raw_write_lock
      5.07            -0.5        4.60        perf-profile.self.cycles-pp.vma_compute_subtree_gap
      0.59            -0.2        0.38        perf-profile.self.cycles-pp.remove_vma
     15.01            -0.2       14.81        perf-profile.self.cycles-pp.syscall_return_via_sysret
      3.15            -0.2        2.96        perf-profile.self.cycles-pp.do_munmap
      0.98            -0.1        0.87        perf-profile.self.cycles-pp.__vma_rb_erase
      1.10            -0.1        0.99        perf-profile.self.cycles-pp.kmem_cache_free
      0.68            -0.1        0.58        perf-profile.self.cycles-pp.__vm_enough_memory
      0.42            -0.1        0.33        perf-profile.self.cycles-pp.unmap_vmas
      3.62            -0.1        3.53        perf-profile.self.cycles-pp.perf_event_mmap
      1.41            -0.1        1.34        perf-profile.self.cycles-pp.entry_SYSCALL_64_after_hwframe
      1.29            -0.1        1.22        perf-profile.self.cycles-pp.__indirect_thunk_start
      0.73            -0.1        0.66        perf-profile.self.cycles-pp.sync_mm_rss
      2.96            -0.1        2.90        perf-profile.self.cycles-pp.__x64_sys_brk
      3.24            -0.1        3.19        perf-profile.self.cycles-pp.brk
      1.11            -0.0        1.07        perf-profile.self.cycles-pp.memcpy_erms
      0.53 ±  3%      -0.0        0.49 ±  2%  perf-profile.self.cycles-pp.vma_link
      0.73            -0.0        0.69        perf-profile.self.cycles-pp.unmap_region
      1.66            -0.0        1.61        perf-profile.self.cycles-pp.down_write_killable
      0.39            -0.0        0.35        perf-profile.self.cycles-pp.__rb_insert_augmented
      1.74            -0.0        1.71        perf-profile.self.cycles-pp.kmem_cache_alloc
      0.55 ±  2%      -0.0        0.52        perf-profile.self.cycles-pp.unmap_single_vma
      1.61            -0.0        1.59        perf-profile.self.cycles-pp.memset_erms
      0.80            -0.0        0.77        perf-profile.self.cycles-pp._raw_spin_lock
      0.13            -0.0        0.11 ±  4%  perf-profile.self.cycles-pp.__vma_link_file
      0.43            -0.0        0.41        perf-profile.self.cycles-pp.strlen
      0.07 ±  6%      -0.0        0.06 ±  8%  perf-profile.self.cycles-pp.should_failslab
      0.81            -0.0        0.79        perf-profile.self.cycles-pp.tlb_finish_mmu
      0.15            +0.0        0.16        perf-profile.self.cycles-pp.__vma_link_list
      0.45            +0.0        0.47        perf-profile.self.cycles-pp.rcu_all_qs
      0.71            +0.0        0.72        perf-profile.self.cycles-pp.strlcpy
      0.51            +0.1        0.56        perf-profile.self.cycles-pp.free_pgtables
      1.41            +0.1        1.48        perf-profile.self.cycles-pp.__vma_link_rb
      2.27            +0.2        2.44        perf-profile.self.cycles-pp.vmacache_find
      0.00            +0.7        0.69        perf-profile.self.cycles-pp.put_vma
      0.00            +1.2        1.23        perf-profile.self.cycles-pp.__vma_merge
      0.00            +1.5        1.50        perf-profile.self.cycles-pp._raw_write_lock

=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/nr_task/thp_enabled/test/cpufreq_governor:
  lkp-skl-4sp1/will-it-scale/debian-x86_64-2018-04-03.cgz/x86_64-rhel-7.2/gcc-7/100%/always/brk1/performance

commit: 
  ba98a1cdad71d259a194461b3a61471b49b14df1
  a7a8993bfe3ccb54ad468b9f1799649e4ad1ff12

ba98a1cdad71d259 a7a8993bfe3ccb54ad468b9f17 
---------------- -------------------------- 
       fail:runs  %reproduction    fail:runs
           |             |             |    
           :3           33%           1:3     dmesg.WARNING:stack_going_in_the_wrong_direction?ip=schedule_tail/0x
           :3           33%           1:3     kmsg.DHCP/BOOTP:Reply_not_for_us_on_eth#,op[#]xid[#]
         %stddev     %change         %stddev
             \          |                \  
    998475            -2.2%     976893        will-it-scale.per_process_ops
    625.87            -2.3%     611.42        will-it-scale.time.elapsed_time
    625.87            -2.3%     611.42        will-it-scale.time.elapsed_time.max
      8158            -1.9%       8000        will-it-scale.time.maximum_resident_set_size
     18.42 ±  2%     -11.9%      16.24        will-it-scale.time.user_time
  34349225 ± 13%     -14.5%   29371024 ± 17%  will-it-scale.time.voluntary_context_switches
 1.919e+08            -2.2%  1.877e+08        will-it-scale.workload
      1639 ± 23%     -18.4%       1337 ± 30%  meminfo.Mlocked
     17748 ± 82%    +103.1%      36051        numa-numastat.node3.other_node
  33410486 ± 14%     -14.8%   28449258 ± 18%  cpuidle.C1.usage
    698749 ± 15%     -18.0%     573307 ± 20%  cpuidle.POLL.usage
   3013702 ± 14%     -15.1%    2559405 ± 17%  softirqs.SCHED
  54361293 ±  2%     -19.0%   44044816 ±  2%  softirqs.TIMER
  33408303 ± 14%     -14.9%   28447123 ± 18%  turbostat.C1
      0.34 ± 16%     -52.0%       0.16 ± 15%  turbostat.Pkg%pc2
      1310 ± 74%    +412.1%       6710 ± 58%  irq_exception_noise.__do_page_fault.samples
      3209 ± 74%    +281.9%      12258 ± 53%  irq_exception_noise.__do_page_fault.sum
    600.67 ±132%     -96.0%      24.00 ± 23%  irq_exception_noise.irq_nr
     99557 ±  7%     -24.0%      75627 ±  7%  irq_exception_noise.softirq_nr
     41424 ±  9%     -24.6%      31253 ±  6%  irq_exception_noise.softirq_time
    625.87            -2.3%     611.42        time.elapsed_time
    625.87            -2.3%     611.42        time.elapsed_time.max
      8158            -1.9%       8000        time.maximum_resident_set_size
     18.42 ±  2%     -11.9%      16.24        time.user_time
  34349225 ± 13%     -14.5%   29371024 ± 17%  time.voluntary_context_switches
    988.00 ±  8%     +14.5%       1131 ±  2%  slabinfo.Acpi-ParseExt.active_objs
    988.00 ±  8%     +14.5%       1131 ±  2%  slabinfo.Acpi-ParseExt.num_objs
      2384 ±  3%     +21.1%       2888 ± 11%  slabinfo.pool_workqueue.active_objs
      2474 ±  2%     +20.4%       2979 ± 11%  slabinfo.pool_workqueue.num_objs
    490.33 ± 10%     -19.2%     396.00 ± 11%  slabinfo.secpath_cache.active_objs
    490.33 ± 10%     -19.2%     396.00 ± 11%  slabinfo.secpath_cache.num_objs
      1123 ±  7%     +14.2%       1282 ±  3%  slabinfo.skbuff_fclone_cache.active_objs
      1123 ±  7%     +14.2%       1282 ±  3%  slabinfo.skbuff_fclone_cache.num_objs
      1.09            -0.0        1.07        perf-stat.branch-miss-rate%
 2.691e+11            -2.4%  2.628e+11        perf-stat.branch-misses
  71981351 ± 12%     -13.8%   62013509 ± 16%  perf-stat.context-switches
 1.697e+13            +1.1%  1.715e+13        perf-stat.dTLB-stores
      2.36 ± 29%      +4.4        6.76 ± 11%  perf-stat.iTLB-load-miss-rate%
  5.21e+08 ± 28%    +194.8%  1.536e+09 ± 10%  perf-stat.iTLB-load-misses
    239983 ± 24%     -68.4%      75819 ± 11%  perf-stat.instructions-per-iTLB-miss
   3295653 ±  2%      -6.3%    3088753 ±  3%  perf-stat.node-stores
    606239            +1.1%     612799        perf-stat.path-length
      3755 ± 28%     -37.5%       2346 ± 52%  sched_debug.cfs_rq:/.exec_clock.stddev
     10.45 ±  4%     +24.3%      12.98 ± 18%  sched_debug.cfs_rq:/.load_avg.stddev
      6243 ± 46%     -38.6%       3831 ± 78%  sched_debug.cpu.load.stddev
    867.80 ±  7%     +25.3%       1087 ±  6%  sched_debug.cpu.nr_load_updates.stddev
    395898 ±  3%     -11.1%     352071 ±  7%  sched_debug.cpu.nr_switches.max
    -13.33           -21.1%     -10.52        sched_debug.cpu.nr_uninterruptible.min
    395674 ±  3%     -11.1%     351762 ±  7%  sched_debug.cpu.sched_count.max
     33152 ±  4%     -12.8%      28899        sched_debug.cpu.ttwu_count.min
      0.03 ± 20%     +77.7%       0.05 ± 15%  sched_debug.rt_rq:/.rt_time.max
     89523            +1.8%      91099        proc-vmstat.nr_active_anon
    409.67 ± 23%     -18.4%     334.33 ± 30%  proc-vmstat.nr_mlock
     89530            +1.8%      91117        proc-vmstat.nr_zone_active_anon
   2337130            -2.2%    2286775        proc-vmstat.numa_hit
   2229090            -2.3%    2178626        proc-vmstat.numa_local
      8460 ± 39%     -75.5%       2076 ± 53%  proc-vmstat.numa_pages_migrated
     28643 ± 55%     -83.5%       4727 ± 58%  proc-vmstat.numa_pte_updates
   2695806            -1.8%    2646639        proc-vmstat.pgfault
   2330191            -2.1%    2281197        proc-vmstat.pgfree
      8460 ± 39%     -75.5%       2076 ± 53%  proc-vmstat.pgmigrate_success
    237651 ±  2%     +31.3%     312092 ± 16%  numa-meminfo.node0.FilePages
      8059 ±  2%     +10.7%       8925 ±  7%  numa-meminfo.node0.KernelStack
      6830 ± 25%     +48.8%      10164 ± 35%  numa-meminfo.node0.Mapped
      1612 ± 21%     +70.0%       2740 ± 19%  numa-meminfo.node0.PageTables
     10772 ± 65%    +679.4%      83962 ± 59%  numa-meminfo.node0.Shmem
    163195 ± 15%     -36.9%     103036 ± 32%  numa-meminfo.node1.Active
    163195 ± 15%     -36.9%     103036 ± 32%  numa-meminfo.node1.Active(anon)
      1730 ±  4%     +33.9%       2317 ± 14%  numa-meminfo.node1.PageTables
     55778 ± 19%     +32.5%      73910 ±  8%  numa-meminfo.node1.SUnreclaim
      2671 ± 16%     -45.0%       1469 ± 15%  numa-meminfo.node2.PageTables
     61537 ± 13%     -17.7%      50647 ±  3%  numa-meminfo.node2.SUnreclaim
     48644 ± 94%    +149.8%     121499 ± 11%  numa-meminfo.node3.Active
     48440 ± 94%    +150.4%     121295 ± 11%  numa-meminfo.node3.Active(anon)
     11832 ± 79%     -91.5%       1008 ± 67%  numa-meminfo.node3.Inactive
     11597 ± 82%     -93.3%     772.00 ± 82%  numa-meminfo.node3.Inactive(anon)
     10389 ± 32%     -43.0%       5921 ±  6%  numa-meminfo.node3.Mapped
     33704 ± 24%     -44.2%      18792 ± 15%  numa-meminfo.node3.SReclaimable
    104733 ± 14%     -25.3%      78275 ±  8%  numa-meminfo.node3.Slab
    139329 ±133%     -99.8%     241.67 ± 79%  latency_stats.avg.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_do_create.nfs3_proc_create.nfs_create.path_openat.do_filp_open.do_sys_open.do_syscall_64
      5403 ±139%     -97.5%     137.67 ± 71%  latency_stats.avg.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_lookup.nfs_lookup.path_openat.do_filp_open.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
    165968 ±101%     -61.9%      63304 ± 58%  latency_stats.avg.max
     83.00        +12810.4%      10715 ±140%  latency_stats.avg.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_access.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_lookupat.filename_lookup
    102.67 ±  6%  +18845.5%      19450 ±140%  latency_stats.avg.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_getattr.__nfs_revalidate_inode.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_lookupat
    136.33 ± 16%  +25043.5%      34279 ±141%  latency_stats.avg.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_lookup.nfs_lookup.__lookup_slow.lookup_slow.walk_component.path_lookupat.filename_lookup
     18497 ±141%    -100.0%       0.00        latency_stats.max.call_rwsem_down_write_failed_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe
    140500 ±131%     -99.8%     247.00 ± 78%  latency_stats.max.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_do_create.nfs3_proc_create.nfs_create.path_openat.do_filp_open.do_sys_open.do_syscall_64
      5403 ±139%     -97.5%     137.67 ± 71%  latency_stats.max.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_lookup.nfs_lookup.path_openat.do_filp_open.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
     87.33 ±  5%  +23963.0%      21015 ±140%  latency_stats.max.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_access.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_lookupat.filename_lookup
    136.33 ± 16%  +25043.5%      34279 ±141%  latency_stats.max.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_lookup.nfs_lookup.__lookup_slow.lookup_slow.walk_component.path_lookupat.filename_lookup
    149.33 ± 14%  +25485.9%      38208 ±141%  latency_stats.max.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_getattr.__nfs_revalidate_inode.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_lookupat
     18761 ±141%    -100.0%       0.00        latency_stats.sum.call_rwsem_down_write_failed_killable.vm_mmap_pgoff.ksys_mmap_pgoff.do_syscall_64.entry_SYSCALL_64_after_hwframe
     23363 ±114%    -100.0%       0.00        latency_stats.sum.call_rwsem_down_read_failed.__do_page_fault.do_page_fault.page_fault.__get_user_8.exit_robust_list.mm_release.do_exit.do_group_exit.get_signal.do_signal.exit_to_usermode_loop
    144810 ±125%     -99.8%     326.67 ± 70%  latency_stats.sum.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_do_create.nfs3_proc_create.nfs_create.path_openat.do_filp_open.do_sys_open.do_syscall_64
      5403 ±139%     -97.5%     137.67 ± 71%  latency_stats.sum.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_lookup.nfs_lookup.path_openat.do_filp_open.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
     59698 ± 98%     -78.0%      13110 ±141%  latency_stats.sum.call_rwsem_down_read_failed.do_exit.do_group_exit.get_signal.do_signal.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
    166.33        +12768.5%      21404 ±140%  latency_stats.sum.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_access.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_lookupat.filename_lookup
    825.00 ±  6%  +18761.7%     155609 ±140%  latency_stats.sum.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_getattr.__nfs_revalidate_inode.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_lookupat
    136.33 ± 16%  +25043.5%      34279 ±141%  latency_stats.sum.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_lookup.nfs_lookup.__lookup_slow.lookup_slow.walk_component.path_lookupat.filename_lookup
     59412 ±  2%     +31.3%      78021 ± 16%  numa-vmstat.node0.nr_file_pages
      8059 ±  2%     +10.7%       8923 ±  7%  numa-vmstat.node0.nr_kernel_stack
      1701 ± 25%     +49.1%       2536 ± 35%  numa-vmstat.node0.nr_mapped
    402.33 ± 21%     +70.0%     684.00 ± 19%  numa-vmstat.node0.nr_page_table_pages
      2692 ± 65%    +679.5%      20988 ± 59%  numa-vmstat.node0.nr_shmem
    622587 ± 36%     +37.7%     857545 ± 13%  numa-vmstat.node0.numa_local
     40797 ± 15%     -36.9%      25757 ± 32%  numa-vmstat.node1.nr_active_anon
    432.00 ±  4%     +33.9%     578.33 ± 14%  numa-vmstat.node1.nr_page_table_pages
     13944 ± 19%     +32.5%      18477 ±  8%  numa-vmstat.node1.nr_slab_unreclaimable
     40797 ± 15%     -36.9%      25757 ± 32%  numa-vmstat.node1.nr_zone_active_anon
    625073 ± 26%     +29.4%     808657 ± 18%  numa-vmstat.node1.numa_hit
    503969 ± 34%     +39.2%     701446 ± 23%  numa-vmstat.node1.numa_local
    137.33 ± 40%     -49.0%      70.00 ± 29%  numa-vmstat.node2.nr_mlock
    667.67 ± 17%     -45.1%     366.33 ± 15%  numa-vmstat.node2.nr_page_table_pages
     15384 ± 13%     -17.7%      12662 ±  3%  numa-vmstat.node2.nr_slab_unreclaimable
     12114 ± 94%    +150.3%      30326 ± 11%  numa-vmstat.node3.nr_active_anon
      2887 ± 83%     -93.4%     190.00 ± 82%  numa-vmstat.node3.nr_inactive_anon
      2632 ± 30%     -39.2%       1600 ±  5%  numa-vmstat.node3.nr_mapped
    101.00           -30.0%      70.67 ± 29%  numa-vmstat.node3.nr_mlock
      8425 ± 24%     -44.2%       4697 ± 15%  numa-vmstat.node3.nr_slab_reclaimable
     12122 ± 94%    +150.3%      30346 ± 11%  numa-vmstat.node3.nr_zone_active_anon
      2887 ± 83%     -93.4%     190.00 ± 82%  numa-vmstat.node3.nr_zone_inactive_anon
    106945 ± 13%     +17.4%     125554        numa-vmstat.node3.numa_other
      4.17            -0.3        3.82        perf-profile.calltrace.cycles-pp.kmem_cache_alloc.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
     15.02            -0.3       14.77        perf-profile.calltrace.cycles-pp.syscall_return_via_sysret.brk
      2.42            -0.2        2.18        perf-profile.calltrace.cycles-pp.vma_compute_subtree_gap.__vma_link_rb.vma_link.do_brk_flags.__x64_sys_brk
      7.60            -0.2        7.39        perf-profile.calltrace.cycles-pp.perf_event_mmap.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      7.79            -0.2        7.63        perf-profile.calltrace.cycles-pp.unmap_vmas.unmap_region.do_munmap.__x64_sys_brk.do_syscall_64
      0.82 ±  9%      -0.1        0.68        perf-profile.calltrace.cycles-pp.__vm_enough_memory.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      2.13            -0.1        2.00        perf-profile.calltrace.cycles-pp.vma_compute_subtree_gap.do_munmap.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      1.05            -0.1        0.95        perf-profile.calltrace.cycles-pp.kmem_cache_free.remove_vma.do_munmap.__x64_sys_brk.do_syscall_64
      7.31            -0.1        7.21        perf-profile.calltrace.cycles-pp.unmap_page_range.unmap_vmas.unmap_region.do_munmap.__x64_sys_brk
      0.74            -0.1        0.67        perf-profile.calltrace.cycles-pp.sync_mm_rss.unmap_page_range.unmap_vmas.unmap_region.do_munmap
      1.06            -0.1        1.00        perf-profile.calltrace.cycles-pp.memcpy_erms.strlcpy.perf_event_mmap.do_brk_flags.__x64_sys_brk
      3.38            -0.1        3.33        perf-profile.calltrace.cycles-pp.get_unmapped_area.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      1.05            -0.0        1.00 ±  2%  perf-profile.calltrace.cycles-pp.__indirect_thunk_start.brk
      2.34            -0.0        2.29        perf-profile.calltrace.cycles-pp.perf_iterate_sb.perf_event_mmap.do_brk_flags.__x64_sys_brk.do_syscall_64
      1.64            -0.0        1.59        perf-profile.calltrace.cycles-pp.strlcpy.perf_event_mmap.do_brk_flags.__x64_sys_brk.do_syscall_64
      1.89            -0.0        1.86        perf-profile.calltrace.cycles-pp.security_mmap_addr.get_unmapped_area.do_brk_flags.__x64_sys_brk.do_syscall_64
      0.76            -0.0        0.73        perf-profile.calltrace.cycles-pp._raw_spin_lock.unmap_page_range.unmap_vmas.unmap_region.do_munmap
      0.57 ±  2%      -0.0        0.55        perf-profile.calltrace.cycles-pp.selinux_mmap_addr.security_mmap_addr.get_unmapped_area.do_brk_flags.__x64_sys_brk
      0.54 ±  2%      +0.0        0.56        perf-profile.calltrace.cycles-pp.do_brk_flags.brk
      0.72            +0.0        0.76 ±  2%  perf-profile.calltrace.cycles-pp.do_munmap.brk
      4.38            +0.1        4.43        perf-profile.calltrace.cycles-pp.find_vma.do_munmap.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      1.96            +0.1        2.04        perf-profile.calltrace.cycles-pp.vmacache_find.find_vma.do_munmap.__x64_sys_brk.do_syscall_64
      0.53            +0.2        0.68        perf-profile.calltrace.cycles-pp.__vma_link_rb.brk
      2.21            +0.3        2.51        perf-profile.calltrace.cycles-pp.remove_vma.do_munmap.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
     64.44            +0.5       64.90        perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.brk
     63.04            +0.5       63.54        perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.brk
     60.37            +0.5       60.88        perf-profile.calltrace.cycles-pp.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe.brk
      3.75            +0.5        4.29        perf-profile.calltrace.cycles-pp.vma_link.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00            +0.6        0.57        perf-profile.calltrace.cycles-pp.free_pgtables.unmap_region.do_munmap.__x64_sys_brk.do_syscall_64
      0.00            +0.6        0.64        perf-profile.calltrace.cycles-pp.put_vma.remove_vma.do_munmap.__x64_sys_brk.do_syscall_64
      0.72            +0.7        1.37        perf-profile.calltrace.cycles-pp.__vma_rb_erase.do_munmap.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
     24.42            +0.7       25.08        perf-profile.calltrace.cycles-pp.do_munmap.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe.brk
      0.00            +0.7        0.71        perf-profile.calltrace.cycles-pp._raw_write_lock.__vma_rb_erase.do_munmap.__x64_sys_brk.do_syscall_64
      3.12            +0.7        3.84        perf-profile.calltrace.cycles-pp.__vma_link_rb.vma_link.do_brk_flags.__x64_sys_brk.do_syscall_64
      0.00            +0.8        0.77        perf-profile.calltrace.cycles-pp._raw_write_lock.__vma_link_rb.vma_link.do_brk_flags.__x64_sys_brk
      0.00            +0.9        0.85        perf-profile.calltrace.cycles-pp.__vma_merge.do_brk_flags.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe
      5.10            -0.5        4.60        perf-profile.children.cycles-pp.vma_compute_subtree_gap
      4.53            -0.3        4.18        perf-profile.children.cycles-pp.kmem_cache_alloc
     15.03            -0.3       14.77        perf-profile.children.cycles-pp.syscall_return_via_sysret
      8.13            -0.2        7.92        perf-profile.children.cycles-pp.perf_event_mmap
      8.01            -0.2        7.81        perf-profile.children.cycles-pp.unmap_vmas
      0.97 ± 14%      -0.2        0.78        perf-profile.children.cycles-pp.__vm_enough_memory
      1.13            -0.1        1.00        perf-profile.children.cycles-pp.kmem_cache_free
      7.82            -0.1        7.70        perf-profile.children.cycles-pp.unmap_page_range
     12.23            -0.1       12.13        perf-profile.children.cycles-pp.unmap_region
      0.74            -0.1        0.67        perf-profile.children.cycles-pp.sync_mm_rss
      3.06            -0.1        3.00        perf-profile.children.cycles-pp.down_write_killable
      0.40 ±  2%      -0.1        0.34        perf-profile.children.cycles-pp.__rb_insert_augmented
      1.29            -0.1        1.23        perf-profile.children.cycles-pp.__indirect_thunk_start
      2.54            -0.1        2.49        perf-profile.children.cycles-pp.perf_iterate_sb
      3.66            -0.0        3.61        perf-profile.children.cycles-pp.get_unmapped_area
      1.80            -0.0        1.75        perf-profile.children.cycles-pp.strlcpy
      0.53 ±  2%      -0.0        0.49 ±  2%  perf-profile.children.cycles-pp.cap_capable
      1.57            -0.0        1.53        perf-profile.children.cycles-pp.arch_get_unmapped_area_topdown
      1.11            -0.0        1.08        perf-profile.children.cycles-pp.memcpy_erms
      0.13            -0.0        0.10        perf-profile.children.cycles-pp.__vma_link_file
      0.55            -0.0        0.52        perf-profile.children.cycles-pp.unmap_single_vma
      1.47            -0.0        1.44        perf-profile.children.cycles-pp.cap_vm_enough_memory
      2.14            -0.0        2.12        perf-profile.children.cycles-pp.security_mmap_addr
      0.32            -0.0        0.30        perf-profile.children.cycles-pp.userfaultfd_unmap_complete
      1.25            -0.0        1.23        perf-profile.children.cycles-pp.up_write
      0.50            -0.0        0.49        perf-profile.children.cycles-pp.userfaultfd_unmap_prep
      0.27            -0.0        0.26        perf-profile.children.cycles-pp.tlb_flush_mmu_free
      1.14            -0.0        1.12        perf-profile.children.cycles-pp.__might_sleep
      0.07            -0.0        0.06        perf-profile.children.cycles-pp.should_failslab
      0.72            +0.0        0.74        perf-profile.children.cycles-pp._cond_resched
      0.45            +0.0        0.47        perf-profile.children.cycles-pp.rcu_all_qs
      0.15 ±  3%      +0.0        0.17 ±  4%  perf-profile.children.cycles-pp.__vma_link_list
      0.15 ±  5%      +0.0        0.18 ±  5%  perf-profile.children.cycles-pp.tick_sched_timer
      0.05 ±  8%      +0.1        0.12 ± 17%  perf-profile.children.cycles-pp.perf_mux_hrtimer_handler
      0.80            +0.1        0.89        perf-profile.children.cycles-pp.free_pgtables
      0.22 ±  7%      +0.1        0.31 ±  9%  perf-profile.children.cycles-pp.__hrtimer_run_queues
      0.00            +0.1        0.11 ± 15%  perf-profile.children.cycles-pp.clockevents_program_event
      6.34            +0.1        6.47        perf-profile.children.cycles-pp.find_vma
      2.27            +0.1        2.40        perf-profile.children.cycles-pp.vmacache_find
      0.40 ±  4%      +0.2        0.58 ±  5%  perf-profile.children.cycles-pp.apic_timer_interrupt
      0.40 ±  4%      +0.2        0.58 ±  5%  perf-profile.children.cycles-pp.smp_apic_timer_interrupt
      0.37 ±  4%      +0.2        0.54 ±  5%  perf-profile.children.cycles-pp.hrtimer_interrupt
      0.00            +0.2        0.19 ± 12%  perf-profile.children.cycles-pp.ktime_get
      2.42            +0.3        2.77        perf-profile.children.cycles-pp.remove_vma
     64.49            +0.5       64.94        perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
      1.27            +0.5        1.73        perf-profile.children.cycles-pp.__vma_rb_erase
     61.62            +0.5       62.10        perf-profile.children.cycles-pp.__x64_sys_brk
     63.24            +0.5       63.74        perf-profile.children.cycles-pp.do_syscall_64
      4.03            +0.5        4.56        perf-profile.children.cycles-pp.vma_link
      0.00            +0.7        0.69        perf-profile.children.cycles-pp.put_vma
     25.13            +0.7       25.84        perf-profile.children.cycles-pp.do_munmap
      3.83            +0.7        4.56        perf-profile.children.cycles-pp.__vma_link_rb
      0.00            +1.2        1.25        perf-profile.children.cycles-pp.__vma_merge
      0.00            +1.5        1.53        perf-profile.children.cycles-pp._raw_write_lock
      5.08            -0.5        4.58        perf-profile.self.cycles-pp.vma_compute_subtree_gap
     15.03            -0.3       14.77        perf-profile.self.cycles-pp.syscall_return_via_sysret
      0.59            -0.2        0.39        perf-profile.self.cycles-pp.remove_vma
      0.72 ±  7%      -0.1        0.58        perf-profile.self.cycles-pp.__vm_enough_memory
      1.12            -0.1        0.99        perf-profile.self.cycles-pp.kmem_cache_free
      3.11            -0.1        2.99        perf-profile.self.cycles-pp.do_munmap
      0.99            -0.1        0.88        perf-profile.self.cycles-pp.__vma_rb_erase
      3.63            -0.1        3.52        perf-profile.self.cycles-pp.perf_event_mmap
      3.26            -0.1        3.17        perf-profile.self.cycles-pp.brk
      0.41 ±  2%      -0.1        0.33        perf-profile.self.cycles-pp.unmap_vmas
      0.74            -0.1        0.67        perf-profile.self.cycles-pp.sync_mm_rss
      1.75            -0.1        1.68        perf-profile.self.cycles-pp.kmem_cache_alloc
      0.40 ±  2%      -0.1        0.34        perf-profile.self.cycles-pp.__rb_insert_augmented
      1.29 ±  2%      -0.1        1.23        perf-profile.self.cycles-pp.__indirect_thunk_start
      0.73            -0.0        0.68 ±  2%  perf-profile.self.cycles-pp.unmap_region
      0.53            -0.0        0.49        perf-profile.self.cycles-pp.vma_link
      1.40            -0.0        1.35        perf-profile.self.cycles-pp.entry_SYSCALL_64_after_hwframe
      5.22            -0.0        5.18        perf-profile.self.cycles-pp.unmap_page_range
      0.53 ±  2%      -0.0        0.49 ±  2%  perf-profile.self.cycles-pp.cap_capable
      1.11            -0.0        1.07        perf-profile.self.cycles-pp.memcpy_erms
      1.86            -0.0        1.82        perf-profile.self.cycles-pp.perf_iterate_sb
      1.30            -0.0        1.27        perf-profile.self.cycles-pp.arch_get_unmapped_area_topdown
      0.13            -0.0        0.10        perf-profile.self.cycles-pp.__vma_link_file
      0.55            -0.0        0.52        perf-profile.self.cycles-pp.unmap_single_vma
      0.74            -0.0        0.72        perf-profile.self.cycles-pp.selinux_mmap_addr
      0.32            -0.0        0.30        perf-profile.self.cycles-pp.userfaultfd_unmap_complete
      1.13            -0.0        1.12        perf-profile.self.cycles-pp.__might_sleep
      1.24            -0.0        1.23        perf-profile.self.cycles-pp.up_write
      0.50            -0.0        0.49        perf-profile.self.cycles-pp.userfaultfd_unmap_prep
      0.27            -0.0        0.26        perf-profile.self.cycles-pp.tlb_flush_mmu_free
      0.07            -0.0        0.06        perf-profile.self.cycles-pp.should_failslab
      0.45            +0.0        0.47        perf-profile.self.cycles-pp.rcu_all_qs
      0.71            +0.0        0.73        perf-profile.self.cycles-pp.strlcpy
      0.15 ±  3%      +0.0        0.17 ±  4%  perf-profile.self.cycles-pp.__vma_link_list
      0.51            +0.1        0.57        perf-profile.self.cycles-pp.free_pgtables
      1.40            +0.1        1.49        perf-profile.self.cycles-pp.__vma_link_rb
      2.27            +0.1        2.39        perf-profile.self.cycles-pp.vmacache_find
      0.00            +0.2        0.18 ± 12%  perf-profile.self.cycles-pp.ktime_get
      0.00            +0.7        0.69        perf-profile.self.cycles-pp.put_vma
      0.00            +1.2        1.24        perf-profile.self.cycles-pp.__vma_merge
      0.00            +1.5        1.52        perf-profile.self.cycles-pp._raw_write_lock

=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/nr_task/thp_enabled/test/cpufreq_governor:
  lkp-skl-4sp1/will-it-scale/debian-x86_64-2018-04-03.cgz/x86_64-rhel-7.2/gcc-7/100%/always/page_fault2/performance

commit: 
  ba98a1cdad71d259a194461b3a61471b49b14df1
  a7a8993bfe3ccb54ad468b9f1799649e4ad1ff12

ba98a1cdad71d259 a7a8993bfe3ccb54ad468b9f17 
---------------- -------------------------- 
       fail:runs  %reproduction    fail:runs
           |             |             |    
           :3           33%           1:3     dmesg.WARNING:at#for_ip_native_iret/0x
          1:3          -33%            :3     dmesg.WARNING:stack_going_in_the_wrong_direction?ip=__schedule/0x
           :3           33%           1:3     dmesg.WARNING:stack_going_in_the_wrong_direction?ip=__slab_free/0x
          1:3          -33%            :3     kmsg.DHCP/BOOTP:Reply_not_for_us_on_eth#,op[#]xid[#]
          3:3         -100%            :3     kmsg.pstore:crypto_comp_decompress_failed,ret=
          3:3         -100%            :3     kmsg.pstore:decompression_failed
          2:3            4%           2:3     perf-profile.calltrace.cycles-pp.sync_regs.error_entry
          5:3            7%           5:3     perf-profile.calltrace.cycles-pp.error_entry
          5:3            7%           5:3     perf-profile.children.cycles-pp.error_entry
          2:3            3%           2:3     perf-profile.self.cycles-pp.error_entry
         %stddev     %change         %stddev
             \          |                \  
      8281 ±  2%     -18.8%       6728        will-it-scale.per_thread_ops
     92778 ±  2%     +17.6%     109080        will-it-scale.time.involuntary_context_switches
  21954366 ±  3%      +4.1%   22857988 ±  2%  will-it-scale.time.maximum_resident_set_size
  4.81e+08 ±  2%     -18.9%  3.899e+08        will-it-scale.time.minor_page_faults
      5804           +12.2%       6512        will-it-scale.time.percent_of_cpu_this_job_got
     34918           +12.2%      39193        will-it-scale.time.system_time
   5638528 ±  2%     -15.3%    4778392        will-it-scale.time.voluntary_context_switches
  15846405            -2.0%   15531034        will-it-scale.workload
   2818137            +1.5%    2861500        interrupts.CAL:Function_call_interrupts
      3.33 ± 28%     -60.0%       1.33 ± 93%  irq_exception_noise.irq_time
      2866           +23.9%       3552 ±  2%  kthread_noise.total_time
   5589674 ± 14%     +31.4%    7344810 ±  6%  meminfo.DirectMap2M
     31169           -16.9%      25906        uptime.idle
     25242 ±  4%     -14.2%      21654 ±  6%  vmstat.system.cs
      7055           -11.6%       6237        boot-time.idle
     21.12           +19.3%      25.19 ±  9%  boot-time.kernel_boot
     20.03 ±  2%      -3.7       16.38        mpstat.cpu.idle%
      0.00 ±  8%      -0.0        0.00 ±  4%  mpstat.cpu.iowait%
   7284147 ±  2%     -16.4%    6092495        softirqs.RCU
   5350756 ±  2%     -10.9%    4769417 ±  4%  softirqs.SCHED
     42933 ± 21%     -28.2%      30807 ±  7%  numa-meminfo.node2.SReclaimable
     63219 ± 13%     -16.6%      52717 ±  6%  numa-meminfo.node2.SUnreclaim
    106153 ± 16%     -21.3%      83525 ±  5%  numa-meminfo.node2.Slab
    247154 ±  4%      -7.6%     228415        numa-meminfo.node3.Unevictable
     11904 ±  4%     +17.1%      13945 ±  8%  numa-vmstat.node0
      2239 ± 22%     -26.6%       1644 ±  2%  numa-vmstat.node2.nr_mapped
     10728 ± 21%     -28.2%       7701 ±  7%  numa-vmstat.node2.nr_slab_reclaimable
     15803 ± 13%     -16.6%      13179 ±  6%  numa-vmstat.node2.nr_slab_unreclaimable
     61788 ±  4%      -7.6%      57103        numa-vmstat.node3.nr_unevictable
     61788 ±  4%      -7.6%      57103        numa-vmstat.node3.nr_zone_unevictable
     92778 ±  2%     +17.6%     109080        time.involuntary_context_switches
  21954366 ±  3%      +4.1%   22857988 ±  2%  time.maximum_resident_set_size
  4.81e+08 ±  2%     -18.9%  3.899e+08        time.minor_page_faults
      5804           +12.2%       6512        time.percent_of_cpu_this_job_got
     34918           +12.2%      39193        time.system_time
   5638528 ±  2%     -15.3%    4778392        time.voluntary_context_switches
   3942289 ±  2%     -10.5%    3528902 ±  2%  cpuidle.C1.time
    242290           -14.2%     207992        cpuidle.C1.usage
  1.64e+09 ±  2%     -15.7%  1.381e+09        cpuidle.C1E.time
   4621281 ±  2%     -14.7%    3939757        cpuidle.C1E.usage
 2.115e+10 ±  2%     -18.5%  1.723e+10        cpuidle.C6.time
  24771099 ±  2%     -18.0%   20305766        cpuidle.C6.usage
   1210810 ±  4%     -17.6%     997270 ±  2%  cpuidle.POLL.time
     18742 ±  3%     -17.0%      15559 ±  2%  cpuidle.POLL.usage
      4135 ±141%    -100.0%       0.00        latency_stats.avg.x86_reserve_hardware.x86_pmu_event_init.perf_try_init_event.perf_event_alloc.__do_sys_perf_event_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
     33249 ±129%    -100.0%       0.00        latency_stats.max.call_rwsem_down_read_failed.m_start.seq_read.__vfs_read.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
      4135 ±141%    -100.0%       0.00        latency_stats.max.x86_reserve_hardware.x86_pmu_event_init.perf_try_init_event.perf_event_alloc.__do_sys_perf_event_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
     65839 ±116%    -100.0%       0.00        latency_stats.sum.call_rwsem_down_read_failed.m_start.seq_read.__vfs_read.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
      4135 ±141%    -100.0%       0.00        latency_stats.sum.x86_reserve_hardware.x86_pmu_event_init.perf_try_init_event.perf_event_alloc.__do_sys_perf_event_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
      8387 ±122%     -90.9%     767.00 ± 13%  latency_stats.sum.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_getattr.__nfs_revalidate_inode.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_lookupat
    263970 ± 10%     -68.6%      82994 ±  3%  latency_stats.sum.do_syslog.kmsg_read.proc_reg_read.__vfs_read.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
      6173 ± 77%    +173.3%      16869 ± 98%  latency_stats.sum.pipe_write.__vfs_write.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
    101.33            -4.6%      96.67        proc-vmstat.nr_anon_transparent_hugepages
     39967            -1.8%      39241        proc-vmstat.nr_slab_reclaimable
     67166            -2.4%      65522        proc-vmstat.nr_slab_unreclaimable
    237743            -3.9%     228396        proc-vmstat.nr_unevictable
    237743            -3.9%     228396        proc-vmstat.nr_zone_unevictable
 4.807e+09            -2.0%   4.71e+09        proc-vmstat.numa_hit
 4.807e+09            -2.0%   4.71e+09        proc-vmstat.numa_local
 4.791e+09            -2.1%   4.69e+09        proc-vmstat.pgalloc_normal
 4.783e+09            -2.0%  4.685e+09        proc-vmstat.pgfault
 4.807e+09            -2.0%  4.709e+09        proc-vmstat.pgfree
      1753            +4.6%       1833        turbostat.Avg_MHz
    239445           -14.1%     205783        turbostat.C1
   4617105 ±  2%     -14.8%    3934693        turbostat.C1E
      1.40 ±  2%      -0.2        1.18        turbostat.C1E%
  24764661 ±  2%     -18.0%   20297643        turbostat.C6
     18.09 ±  2%      -3.4       14.74        turbostat.C6%
      7.53 ±  2%     -17.1%       6.24        turbostat.CPU%c1
     11.88 ±  2%     -19.1%       9.61        turbostat.CPU%c6
      7.62 ±  3%     -20.8%       6.04        turbostat.Pkg%pc2
    388.30            +1.5%     393.93        turbostat.PkgWatt
    390974 ±  8%     +35.8%     530867 ± 11%  sched_debug.cfs_rq:/.min_vruntime.stddev
  -1754042           +75.7%   -3081270        sched_debug.cfs_rq:/.spread0.min
    388140 ±  8%     +36.2%     528494 ± 11%  sched_debug.cfs_rq:/.spread0.stddev
    542.30 ±  3%     -10.0%     488.21 ±  3%  sched_debug.cfs_rq:/.util_avg.min
     53.35 ± 16%     +48.7%      79.35 ± 12%  sched_debug.cfs_rq:/.util_est_enqueued.avg
     30520 ±  6%     -15.2%      25883 ± 12%  sched_debug.cpu.nr_switches.avg
    473770 ± 27%     -37.4%     296623 ± 32%  sched_debug.cpu.nr_switches.max
     17077 ±  2%     -15.1%      14493        sched_debug.cpu.nr_switches.min
     30138 ±  6%     -15.0%      25606 ± 12%  sched_debug.cpu.sched_count.avg
    472345 ± 27%     -37.2%     296419 ± 32%  sched_debug.cpu.sched_count.max
     16858 ±  2%     -15.2%      14299        sched_debug.cpu.sched_count.min
      8358 ±  2%     -15.5%       7063        sched_debug.cpu.sched_goidle.avg
     12225           -13.6%      10565        sched_debug.cpu.sched_goidle.max
      8032 ±  2%     -16.0%       6749        sched_debug.cpu.sched_goidle.min
     14839 ±  6%     -15.3%      12568 ± 12%  sched_debug.cpu.ttwu_count.avg
    235115 ± 28%     -38.3%     145175 ± 31%  sched_debug.cpu.ttwu_count.max
      7627 ±  3%     -15.9%       6413 ±  2%  sched_debug.cpu.ttwu_count.min
    226299 ± 29%     -39.5%     136827 ± 32%  sched_debug.cpu.ttwu_local.max
      0.85            -0.0        0.81        perf-stat.branch-miss-rate%
 3.675e+10            -4.1%  3.523e+10        perf-stat.branch-misses
 4.052e+11            -2.3%  3.958e+11        perf-stat.cache-misses
 7.008e+11            -2.5%  6.832e+11        perf-stat.cache-references
  15320995 ±  4%     -14.3%   13136557 ±  6%  perf-stat.context-switches
      9.16            +4.8%       9.59        perf-stat.cpi
  2.03e+14            +4.6%  2.124e+14        perf-stat.cpu-cycles
     44508            -1.7%      43743        perf-stat.cpu-migrations
      1.30            -0.1        1.24        perf-stat.dTLB-store-miss-rate%
 4.064e+10            -3.5%  3.922e+10        perf-stat.dTLB-store-misses
 3.086e+12            +1.1%  3.119e+12        perf-stat.dTLB-stores
 3.611e+08 ±  6%      -8.5%  3.304e+08 ±  5%  perf-stat.iTLB-loads
      0.11            -4.6%       0.10        perf-stat.ipc
 4.783e+09            -2.0%  4.685e+09        perf-stat.minor-faults
      1.53 ±  2%      -0.3        1.22 ±  8%  perf-stat.node-load-miss-rate%
 1.389e+09 ±  3%     -22.1%  1.083e+09 ±  9%  perf-stat.node-load-misses
 8.922e+10            -1.9%   8.75e+10        perf-stat.node-loads
      5.06            +1.7        6.77 ±  3%  perf-stat.node-store-miss-rate%
 1.204e+09           +29.3%  1.556e+09 ±  3%  perf-stat.node-store-misses
 2.256e+10            -5.1%  2.142e+10 ±  2%  perf-stat.node-stores
 4.783e+09            -2.0%  4.685e+09        perf-stat.page-faults
   1399242            +1.9%    1425404        perf-stat.path-length
      1144 ±  8%     -13.6%     988.00 ±  8%  slabinfo.Acpi-ParseExt.active_objs
      1144 ±  8%     -13.6%     988.00 ±  8%  slabinfo.Acpi-ParseExt.num_objs
      1878 ± 17%     +29.0%       2422 ± 16%  slabinfo.dmaengine-unmap-16.active_objs
      1878 ± 17%     +29.0%       2422 ± 16%  slabinfo.dmaengine-unmap-16.num_objs
      1085 ±  5%     -24.1%     823.33 ±  9%  slabinfo.file_lock_cache.active_objs
      1085 ±  5%     -24.1%     823.33 ±  9%  slabinfo.file_lock_cache.num_objs
     61584 ±  4%     -16.6%      51381 ±  5%  slabinfo.filp.active_objs
    967.00 ±  4%     -16.5%     807.67 ±  5%  slabinfo.filp.active_slabs
     61908 ±  4%     -16.5%      51713 ±  5%  slabinfo.filp.num_objs
    967.00 ±  4%     -16.5%     807.67 ±  5%  slabinfo.filp.num_slabs
      1455           -15.4%       1232 ±  4%  slabinfo.nsproxy.active_objs
      1455           -15.4%       1232 ±  4%  slabinfo.nsproxy.num_objs
     84720 ±  6%     -18.3%      69210 ±  4%  slabinfo.pid.active_objs
      1324 ±  6%     -18.2%       1083 ±  4%  slabinfo.pid.active_slabs
     84820 ±  5%     -18.2%      69386 ±  4%  slabinfo.pid.num_objs
      1324 ±  6%     -18.2%       1083 ±  4%  slabinfo.pid.num_slabs
      2112 ± 18%     -26.3%       1557 ±  5%  slabinfo.scsi_sense_cache.active_objs
      2112 ± 18%     -26.3%       1557 ±  5%  slabinfo.scsi_sense_cache.num_objs
      5018 ±  5%      -7.6%       4635 ±  4%  slabinfo.sock_inode_cache.active_objs
      5018 ±  5%      -7.6%       4635 ±  4%  slabinfo.sock_inode_cache.num_objs
      1193 ±  4%     +13.8%       1358 ±  4%  slabinfo.task_group.active_objs
      1193 ±  4%     +13.8%       1358 ±  4%  slabinfo.task_group.num_objs
     62807 ±  3%     -14.4%      53757 ±  3%  slabinfo.vm_area_struct.active_objs
      1571 ±  3%     -12.1%       1381 ±  3%  slabinfo.vm_area_struct.active_slabs
     62877 ±  3%     -14.3%      53880 ±  3%  slabinfo.vm_area_struct.num_objs
      1571 ±  3%     -12.1%       1381 ±  3%  slabinfo.vm_area_struct.num_slabs
     47.45           -47.4        0.00        perf-profile.calltrace.cycles-pp.alloc_pages_vma.__handle_mm_fault.handle_mm_fault.__do_page_fault.do_page_fault
     47.16           -47.2        0.00        perf-profile.calltrace.cycles-pp.__alloc_pages_nodemask.alloc_pages_vma.__handle_mm_fault.handle_mm_fault.__do_page_fault
     46.99           -47.0        0.00        perf-profile.calltrace.cycles-pp.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_vma.__handle_mm_fault.handle_mm_fault
     44.95           -44.9        0.00        perf-profile.calltrace.cycles-pp._raw_spin_lock.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_vma.__handle_mm_fault
      7.42 ±  2%      -7.4        0.00        perf-profile.calltrace.cycles-pp.copy_page.__handle_mm_fault.handle_mm_fault.__do_page_fault.do_page_fault
      6.32 ± 10%      -6.3        0.00        perf-profile.calltrace.cycles-pp.finish_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault.do_page_fault
      6.28 ± 10%      -6.3        0.00        perf-profile.calltrace.cycles-pp.alloc_set_pte.finish_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
      0.00            +0.9        0.85 ± 11%  perf-profile.calltrace.cycles-pp._raw_spin_lock.pte_map_lock.alloc_set_pte.finish_fault.handle_pte_fault
      0.00            +0.9        0.92 ±  4%  perf-profile.calltrace.cycles-pp.__list_del_entry_valid.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_vma.handle_pte_fault
      0.00            +1.1        1.13 ±  7%  perf-profile.calltrace.cycles-pp.find_lock_entry.shmem_getpage_gfp.shmem_fault.__do_fault.handle_pte_fault
      0.00            +1.2        1.19 ±  7%  perf-profile.calltrace.cycles-pp.shmem_getpage_gfp.shmem_fault.__do_fault.handle_pte_fault.__handle_mm_fault
      0.00            +1.2        1.22 ±  5%  perf-profile.calltrace.cycles-pp.pte_map_lock.alloc_set_pte.finish_fault.handle_pte_fault.__handle_mm_fault
      0.00            +1.3        1.34 ±  7%  perf-profile.calltrace.cycles-pp.shmem_fault.__do_fault.handle_pte_fault.__handle_mm_fault.handle_mm_fault
      0.00            +1.4        1.36 ±  7%  perf-profile.calltrace.cycles-pp.__do_fault.handle_pte_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
      0.00            +4.5        4.54 ± 19%  perf-profile.calltrace.cycles-pp.pagevec_lru_move_fn.__lru_cache_add.alloc_set_pte.finish_fault.handle_pte_fault
      0.00            +4.6        4.64 ± 19%  perf-profile.calltrace.cycles-pp.__lru_cache_add.alloc_set_pte.finish_fault.handle_pte_fault.__handle_mm_fault
      0.00            +6.6        6.64 ± 15%  perf-profile.calltrace.cycles-pp.alloc_set_pte.finish_fault.handle_pte_fault.__handle_mm_fault.handle_mm_fault
      0.00            +6.7        6.68 ± 15%  perf-profile.calltrace.cycles-pp.finish_fault.handle_pte_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
      0.00            +7.5        7.54 ±  5%  perf-profile.calltrace.cycles-pp.copy_page.handle_pte_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
      0.00           +44.6       44.55 ±  3%  perf-profile.calltrace.cycles-pp._raw_spin_lock.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_vma.handle_pte_fault
      0.00           +46.6       46.63 ±  3%  perf-profile.calltrace.cycles-pp.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_vma.handle_pte_fault.__handle_mm_fault
      0.00           +46.8       46.81 ±  3%  perf-profile.calltrace.cycles-pp.__alloc_pages_nodemask.alloc_pages_vma.handle_pte_fault.__handle_mm_fault.handle_mm_fault
      0.00           +47.1       47.10 ±  3%  perf-profile.calltrace.cycles-pp.alloc_pages_vma.handle_pte_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
      0.00           +63.1       63.15        perf-profile.calltrace.cycles-pp.handle_pte_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault.do_page_fault
      0.39 ±  3%      +0.0        0.42 ±  3%  perf-profile.children.cycles-pp.radix_tree_lookup_slot
      0.21 ±  3%      +0.0        0.25 ±  5%  perf-profile.children.cycles-pp.__mod_node_page_state
      0.00            +0.1        0.06 ±  8%  perf-profile.children.cycles-pp.get_vma_policy
      0.00            +0.1        0.08 ±  5%  perf-profile.children.cycles-pp.__lru_cache_add_active_or_unevictable
      0.00            +0.2        0.18 ±  6%  perf-profile.children.cycles-pp.__page_add_new_anon_rmap
      0.00            +1.4        1.35 ±  5%  perf-profile.children.cycles-pp.pte_map_lock
      0.00           +63.2       63.21        perf-profile.children.cycles-pp.handle_pte_fault
      1.40 ±  2%      -0.4        1.03 ± 10%  perf-profile.self.cycles-pp._raw_spin_lock
      0.56 ±  3%      -0.2        0.35 ±  6%  perf-profile.self.cycles-pp.__handle_mm_fault
      0.22 ±  3%      -0.0        0.18 ±  7%  perf-profile.self.cycles-pp.alloc_set_pte
      0.09            +0.0        0.10 ±  4%  perf-profile.self.cycles-pp.vmacache_find
      0.39 ±  2%      +0.0        0.41 ±  3%  perf-profile.self.cycles-pp.__radix_tree_lookup
      0.18            +0.0        0.20 ±  6%  perf-profile.self.cycles-pp.mem_cgroup_charge_statistics
      0.17 ±  2%      +0.0        0.20 ±  7%  perf-profile.self.cycles-pp.___might_sleep
      0.33 ±  2%      +0.0        0.36 ±  6%  perf-profile.self.cycles-pp.handle_mm_fault
      0.20 ±  2%      +0.0        0.24 ±  3%  perf-profile.self.cycles-pp.__mod_node_page_state
      0.00            +0.1        0.05        perf-profile.self.cycles-pp.finish_fault
      0.00            +0.1        0.05        perf-profile.self.cycles-pp.get_vma_policy
      0.00            +0.1        0.08 ± 10%  perf-profile.self.cycles-pp.__lru_cache_add_active_or_unevictable
      0.00            +0.2        0.25 ±  5%  perf-profile.self.cycles-pp.handle_pte_fault
      0.00            +0.5        0.49 ±  8%  perf-profile.self.cycles-pp.pte_map_lock

=========================================================================================
tbox_group/testcase/rootfs/kconfig/compiler/nr_task/thp_enabled/test/cpufreq_governor:
  lkp-skl-4sp1/will-it-scale/debian-x86_64-2018-04-03.cgz/x86_64-rhel-7.2/gcc-7/100%/never/page_fault2/performance

commit: 
  ba98a1cdad71d259a194461b3a61471b49b14df1
  a7a8993bfe3ccb54ad468b9f1799649e4ad1ff12

ba98a1cdad71d259 a7a8993bfe3ccb54ad468b9f17 
---------------- -------------------------- 
       fail:runs  %reproduction    fail:runs
           |             |             |    
          1:3          -33%            :3     kmsg.DHCP/BOOTP:Reply_not_for_us_on_eth#,op[#]xid[#]
           :3           33%           1:3     dmesg.WARNING:stack_going_in_the_wrong_direction?ip=sched_slice/0x
          1:3          -33%            :3     dmesg.WARNING:stack_going_in_the_wrong_direction?ip=schedule_tail/0x
          1:3           24%           2:3     perf-profile.calltrace.cycles-pp.sync_regs.error_entry
          3:3           46%           5:3     perf-profile.calltrace.cycles-pp.error_entry
          5:3           -9%           5:3     perf-profile.children.cycles-pp.error_entry
          2:3           -4%           2:3     perf-profile.self.cycles-pp.error_entry
         %stddev     %change         %stddev
             \          |                \  
      8147           -18.8%       6613        will-it-scale.per_thread_ops
     93113           +17.0%     108982        will-it-scale.time.involuntary_context_switches
 4.732e+08           -19.0%  3.833e+08        will-it-scale.time.minor_page_faults
      5854           +12.0%       6555        will-it-scale.time.percent_of_cpu_this_job_got
     35247           +12.1%      39495        will-it-scale.time.system_time
   5546661           -15.5%    4689314        will-it-scale.time.voluntary_context_switches
  15801637            -1.9%   15504487        will-it-scale.workload
      1.43 ± 11%     -59.7%       0.58 ± 28%  irq_exception_noise.__do_page_fault.min
      2811 ±  3%     +23.7%       3477 ±  3%  kthread_noise.total_time
    292776 ±  5%     +39.6%     408829 ± 21%  meminfo.DirectMap4k
     19.80            -3.7       16.12        mpstat.cpu.idle%
     29940           -14.5%      25593        uptime.idle
     24064 ±  3%      -8.5%      22016        vmstat.system.cs
     34.86            -1.9%      34.19        boot-time.boot
     26.95            -2.8%      26.19 ±  2%  boot-time.kernel_boot
   7190569 ±  2%     -15.2%    6100136 ±  3%  softirqs.RCU
   5513663           -13.8%    4751548        softirqs.SCHED
     18064 ±  2%     +24.3%      22461 ±  7%  numa-vmstat.node0.nr_slab_unreclaimable
      8507 ± 12%     -16.8%       7075 ±  4%  numa-vmstat.node2.nr_slab_reclaimable
     18719 ±  9%     -19.6%      15043 ±  4%  numa-vmstat.node3.nr_slab_unreclaimable
     72265 ±  2%     +24.3%      89855 ±  7%  numa-meminfo.node0.SUnreclaim
    115980 ±  4%     +22.6%     142233 ± 12%  numa-meminfo.node0.Slab
     34035 ± 12%     -16.8%      28307 ±  4%  numa-meminfo.node2.SReclaimable
     74888 ±  9%     -19.7%      60162 ±  4%  numa-meminfo.node3.SUnreclaim
     93113           +17.0%     108982        time.involuntary_context_switches
 4.732e+08           -19.0%  3.833e+08        time.minor_page_faults
      5854           +12.0%       6555        time.percent_of_cpu_this_job_got
     35247           +12.1%      39495        time.system_time
   5546661           -15.5%    4689314        time.voluntary_context_switches
 4.792e+09            -1.9%  4.699e+09        proc-vmstat.numa_hit
 4.791e+09            -1.9%  4.699e+09        proc-vmstat.numa_local
     40447 ± 11%     +13.2%      45804 ±  6%  proc-vmstat.pgactivate
 4.778e+09            -1.9%  4.688e+09        proc-vmstat.pgalloc_normal
 4.767e+09            -1.9%  4.675e+09        proc-vmstat.pgfault
 4.791e+09            -1.9%  4.699e+09        proc-vmstat.pgfree
    230178 ±  2%     -10.1%     206883 ±  3%  cpuidle.C1.usage
 1.617e+09           -15.0%  1.375e+09        cpuidle.C1E.time
   4514401           -14.1%    3878206        cpuidle.C1E.usage
 2.087e+10           -18.5%  1.701e+10        cpuidle.C6.time
  24458365           -18.0%   20045336        cpuidle.C6.usage
   1163758           -16.1%     976094 ±  4%  cpuidle.POLL.time
     17907           -14.6%      15294 ±  4%  cpuidle.POLL.usage
      1758            +4.5%       1838        turbostat.Avg_MHz
    227522 ±  2%     -10.2%     204426 ±  3%  turbostat.C1
   4512700           -14.2%    3873264        turbostat.C1E
      1.39            -0.2        1.18        turbostat.C1E%
  24452583           -18.0%   20039031        turbostat.C6
     17.85            -3.3       14.55        turbostat.C6%
      7.44           -16.8%       6.19        turbostat.CPU%c1
     11.72           -19.3%       9.45        turbostat.CPU%c6
      7.51           -21.3%       5.91        turbostat.Pkg%pc2
    389.33            +1.6%     395.59        turbostat.PkgWatt
    559.33 ± 13%     -17.9%     459.33 ± 20%  slabinfo.dmaengine-unmap-128.active_objs
    559.33 ± 13%     -17.9%     459.33 ± 20%  slabinfo.dmaengine-unmap-128.num_objs
     57734 ±  3%      -5.7%      54421 ±  4%  slabinfo.filp.active_objs
    905.67 ±  3%      -5.6%     854.67 ±  4%  slabinfo.filp.active_slabs
     57981 ±  3%      -5.6%      54720 ±  4%  slabinfo.filp.num_objs
    905.67 ±  3%      -5.6%     854.67 ±  4%  slabinfo.filp.num_slabs
      1378           -12.0%       1212 ±  7%  slabinfo.nsproxy.active_objs
      1378           -12.0%       1212 ±  7%  slabinfo.nsproxy.num_objs
    507.33 ±  7%     -26.8%     371.33 ±  2%  slabinfo.secpath_cache.active_objs
    507.33 ±  7%     -26.8%     371.33 ±  2%  slabinfo.secpath_cache.num_objs
      4788 ±  5%      -8.3%       4391 ±  2%  slabinfo.sock_inode_cache.active_objs
      4788 ±  5%      -8.3%       4391 ±  2%  slabinfo.sock_inode_cache.num_objs
      1431 ±  8%     -12.3%       1255 ±  3%  slabinfo.task_group.active_objs
      1431 ±  8%     -12.3%       1255 ±  3%  slabinfo.task_group.num_objs
      4.27 ± 17%     +27.0%       5.42 ±  7%  sched_debug.cfs_rq:/.runnable_load_avg.avg
     13.44 ± 62%     +73.6%      23.33 ± 24%  sched_debug.cfs_rq:/.runnable_load_avg.stddev
    772.55 ± 21%     -32.7%     520.27 ±  4%  sched_debug.cfs_rq:/.util_est_enqueued.max
      4.39 ± 15%     +29.0%       5.66 ± 11%  sched_debug.cpu.cpu_load[0].avg
    152.09 ± 72%     +83.9%     279.67 ± 33%  sched_debug.cpu.cpu_load[0].max
     13.84 ± 58%     +78.7%      24.72 ± 29%  sched_debug.cpu.cpu_load[0].stddev
      4.53 ± 14%     +25.8%       5.70 ± 10%  sched_debug.cpu.cpu_load[1].avg
    156.58 ± 66%     +76.6%     276.58 ± 33%  sched_debug.cpu.cpu_load[1].max
     14.02 ± 55%     +72.4%      24.17 ± 28%  sched_debug.cpu.cpu_load[1].stddev
      4.87 ± 11%     +17.3%       5.72 ±  9%  sched_debug.cpu.cpu_load[2].avg
      1.58 ±  2%     +13.5%       1.79 ±  6%  sched_debug.cpu.nr_running.max
     16694           -14.6%      14259        sched_debug.cpu.nr_switches.min
     31989 ± 13%     +20.6%      38584 ±  6%  sched_debug.cpu.nr_switches.stddev
     16505           -14.8%      14068        sched_debug.cpu.sched_count.min
     32084 ± 13%     +19.9%      38482 ±  6%  sched_debug.cpu.sched_count.stddev
      8185           -15.0%       6957        sched_debug.cpu.sched_goidle.avg
     12151 ±  2%     -13.5%      10507        sched_debug.cpu.sched_goidle.max
      7867           -15.7%       6631        sched_debug.cpu.sched_goidle.min
      7595           -16.1%       6375        sched_debug.cpu.ttwu_count.min
     15873 ± 13%     +21.2%      19239 ±  6%  sched_debug.cpu.ttwu_count.stddev
      5244 ± 17%     +17.0%       6134 ±  5%  sched_debug.cpu.ttwu_local.avg
     15646 ± 12%     +21.5%      19008 ±  6%  sched_debug.cpu.ttwu_local.stddev
      0.85            -0.0        0.81        perf-stat.branch-miss-rate%
 3.689e+10            -4.6%  3.518e+10        perf-stat.branch-misses
     57.39            +0.6       58.00        perf-stat.cache-miss-rate%
 4.014e+11            -1.2%  3.967e+11        perf-stat.cache-misses
 6.994e+11            -2.2%   6.84e+11        perf-stat.cache-references
  14605393 ±  3%      -8.5%   13369913        perf-stat.context-switches
      9.21            +4.5%       9.63        perf-stat.cpi
 2.037e+14            +4.6%   2.13e+14        perf-stat.cpu-cycles
     44424            -2.0%      43541        perf-stat.cpu-migrations
      1.29            -0.1        1.24        perf-stat.dTLB-store-miss-rate%
 4.018e+10            -2.8%  3.905e+10        perf-stat.dTLB-store-misses
 3.071e+12            +1.4%  3.113e+12        perf-stat.dTLB-stores
     93.04            +1.5       94.51        perf-stat.iTLB-load-miss-rate%
 4.946e+09           +19.3%  5.903e+09 ±  5%  perf-stat.iTLB-load-misses
 3.702e+08            -7.5%  3.423e+08 ±  2%  perf-stat.iTLB-loads
      4470           -15.9%       3760 ±  5%  perf-stat.instructions-per-iTLB-miss
      0.11            -4.3%       0.10        perf-stat.ipc
 4.767e+09            -1.9%  4.675e+09        perf-stat.minor-faults
      1.46 ±  4%      -0.1        1.33 ±  9%  perf-stat.node-load-miss-rate%
      4.91            +1.7        6.65 ±  2%  perf-stat.node-store-miss-rate%
 1.195e+09           +32.8%  1.587e+09 ±  2%  perf-stat.node-store-misses
 2.313e+10            -3.7%  2.227e+10        perf-stat.node-stores
 4.767e+09            -1.9%  4.675e+09        perf-stat.page-faults
   1399047            +2.0%    1427115        perf-stat.path-length
      8908 ± 73%    -100.0%       0.00        latency_stats.avg.call_rwsem_down_read_failed.m_start.seq_read.__vfs_read.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
      3604 ±141%    -100.0%       0.00        latency_stats.avg.call_rwsem_down_write_failed.do_unlinkat.do_syscall_64.entry_SYSCALL_64_after_hwframe
     61499 ±130%     -92.6%       4534 ± 16%  latency_stats.avg.expand_files.__alloc_fd.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
      4391 ±138%     -70.9%       1277 ±129%  latency_stats.avg.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_lookup.nfs_lookup_revalidate.lookup_fast.walk_component.link_path_walk.path_lookupat.filename_lookup
     67311 ±112%     -48.5%      34681 ± 36%  latency_stats.avg.max
      3956 ±138%    +320.4%      16635 ±140%  latency_stats.avg.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_getattr.__nfs_revalidate_inode.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_lookupat
    164.67 ± 30%   +7264.0%      12126 ±138%  latency_stats.avg.flush_work.fsnotify_destroy_group.inotify_release.__fput.task_work_run.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00       +5.4e+105%       5367 ±141%  latency_stats.avg.call_rwsem_down_write_failed.unlink_file_vma.free_pgtables.exit_mmap.mmput.flush_old_exec.load_elf_binary.search_binary_handler.do_execveat_common.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe
     36937 ±119%    -100.0%       0.00        latency_stats.max.call_rwsem_down_read_failed.m_start.seq_read.__vfs_read.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
      3604 ±141%    -100.0%       0.00        latency_stats.max.call_rwsem_down_write_failed.do_unlinkat.do_syscall_64.entry_SYSCALL_64_after_hwframe
     84146 ±107%     -72.5%      23171 ± 31%  latency_stats.max.expand_files.__alloc_fd.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
      4391 ±138%     -70.9%       1277 ±129%  latency_stats.max.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_lookup.nfs_lookup_revalidate.lookup_fast.walk_component.link_path_walk.path_lookupat.filename_lookup
      5817 ± 83%     -69.7%       1760 ± 67%  latency_stats.max.pipe_write.__vfs_write.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
      6720 ±137%   +1628.2%     116147 ±141%  latency_stats.max.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_getattr.__nfs_revalidate_inode.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_lookupat
    164.67 ± 30%   +7264.0%      12126 ±138%  latency_stats.max.flush_work.fsnotify_destroy_group.inotify_release.__fput.task_work_run.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00       +1.2e+106%      12153 ±141%  latency_stats.max.call_rwsem_down_write_failed.unlink_file_vma.free_pgtables.exit_mmap.mmput.flush_old_exec.load_elf_binary.search_binary_handler.do_execveat_common.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe
    110122 ±120%    -100.0%       0.00        latency_stats.sum.call_rwsem_down_read_failed.m_start.seq_read.__vfs_read.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
      3604 ±141%    -100.0%       0.00        latency_stats.sum.call_rwsem_down_write_failed.do_unlinkat.do_syscall_64.entry_SYSCALL_64_after_hwframe
  12078828 ±139%     -99.3%      89363 ± 29%  latency_stats.sum.expand_files.__alloc_fd.do_sys_open.do_syscall_64.entry_SYSCALL_64_after_hwframe
    144453 ±120%     -80.9%      27650 ± 19%  latency_stats.sum.poll_schedule_timeout.do_sys_poll.__x64_sys_poll.do_syscall_64.entry_SYSCALL_64_after_hwframe
      4391 ±138%     -70.9%       1277 ±129%  latency_stats.sum.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_lookup.nfs_lookup_revalidate.lookup_fast.walk_component.link_path_walk.path_lookupat.filename_lookup
      9438 ± 86%     -68.4%       2980 ± 35%  latency_stats.sum.pipe_write.__vfs_write.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
     31656 ±138%    +320.4%     133084 ±140%  latency_stats.sum.rpc_wait_bit_killable.__rpc_execute.rpc_run_task.rpc_call_sync.nfs3_rpc_wrapper.nfs3_proc_getattr.__nfs_revalidate_inode.nfs_do_access.nfs_permission.inode_permission.link_path_walk.path_lookupat
    164.67 ± 30%   +7264.0%      12126 ±138%  latency_stats.sum.flush_work.fsnotify_destroy_group.inotify_release.__fput.task_work_run.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00       +8.8e+105%       8760 ±141%  latency_stats.sum.msleep_interruptible.uart_wait_until_sent.tty_wait_until_sent.tty_port_close_start.tty_port_close.tty_release.__fput.task_work_run.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00       +1.3e+106%      12897 ±141%  latency_stats.sum.tty_wait_until_sent.tty_port_close_start.tty_port_close.tty_release.__fput.task_work_run.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe
      0.00       +3.2e+106%      32207 ±141%  latency_stats.sum.call_rwsem_down_write_failed.unlink_file_vma.free_pgtables.exit_mmap.mmput.flush_old_exec.load_elf_binary.search_binary_handler.do_execveat_common.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe
     44.43 ±  3%     -44.4        0.00        perf-profile.calltrace.cycles-pp.alloc_pages_vma.__handle_mm_fault.handle_mm_fault.__do_page_fault.do_page_fault
     44.13 ±  3%     -44.1        0.00        perf-profile.calltrace.cycles-pp.__alloc_pages_nodemask.alloc_pages_vma.__handle_mm_fault.handle_mm_fault.__do_page_fault
     43.95 ±  3%     -43.9        0.00        perf-profile.calltrace.cycles-pp.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_vma.__handle_mm_fault.handle_mm_fault
     41.85 ±  4%     -41.9        0.00        perf-profile.calltrace.cycles-pp._raw_spin_lock.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_vma.__handle_mm_fault
      7.74 ±  8%      -7.7        0.00        perf-profile.calltrace.cycles-pp.copy_page.__handle_mm_fault.handle_mm_fault.__do_page_fault.do_page_fault
      7.19 ±  4%      -7.2        0.00        perf-profile.calltrace.cycles-pp.finish_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault.do_page_fault
      7.15 ±  4%      -7.2        0.00        perf-profile.calltrace.cycles-pp.alloc_set_pte.finish_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
      5.09 ±  3%      -5.1        0.00        perf-profile.calltrace.cycles-pp.__lru_cache_add.alloc_set_pte.finish_fault.__handle_mm_fault.handle_mm_fault
      4.99 ±  3%      -5.0        0.00        perf-profile.calltrace.cycles-pp.pagevec_lru_move_fn.__lru_cache_add.alloc_set_pte.finish_fault.__handle_mm_fault
      0.93 ±  6%      -0.1        0.81 ±  2%  perf-profile.calltrace.cycles-pp.find_get_entry.find_lock_entry.shmem_getpage_gfp.shmem_fault.__do_fault
      0.00            +0.8        0.84        perf-profile.calltrace.cycles-pp._raw_spin_lock.pte_map_lock.alloc_set_pte.finish_fault.handle_pte_fault
      0.00            +0.9        0.92 ±  3%  perf-profile.calltrace.cycles-pp.__list_del_entry_valid.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_vma.handle_pte_fault
      0.00            +1.1        1.08        perf-profile.calltrace.cycles-pp.find_lock_entry.shmem_getpage_gfp.shmem_fault.__do_fault.handle_pte_fault
      0.00            +1.1        1.14        perf-profile.calltrace.cycles-pp.shmem_getpage_gfp.shmem_fault.__do_fault.handle_pte_fault.__handle_mm_fault
      0.00            +1.2        1.17        perf-profile.calltrace.cycles-pp.pte_map_lock.alloc_set_pte.finish_fault.handle_pte_fault.__handle_mm_fault
      0.00            +1.3        1.29        perf-profile.calltrace.cycles-pp.shmem_fault.__do_fault.handle_pte_fault.__handle_mm_fault.handle_mm_fault
      0.00            +1.3        1.31        perf-profile.calltrace.cycles-pp.__do_fault.handle_pte_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
     61.62            +1.7       63.33        perf-profile.calltrace.cycles-pp.__handle_mm_fault.handle_mm_fault.__do_page_fault.do_page_fault.page_fault
     41.73 ±  4%      +3.0       44.75        perf-profile.calltrace.cycles-pp.native_queued_spin_lock_slowpath._raw_spin_lock.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_vma
      0.00            +4.6        4.55 ± 15%  perf-profile.calltrace.cycles-pp.pagevec_lru_move_fn.__lru_cache_add.alloc_set_pte.finish_fault.handle_pte_fault
      0.00            +4.6        4.65 ± 14%  perf-profile.calltrace.cycles-pp.__lru_cache_add.alloc_set_pte.finish_fault.handle_pte_fault.__handle_mm_fault
      0.00            +6.6        6.57 ± 10%  perf-profile.calltrace.cycles-pp.alloc_set_pte.finish_fault.handle_pte_fault.__handle_mm_fault.handle_mm_fault
      0.00            +6.6        6.61 ± 10%  perf-profile.calltrace.cycles-pp.finish_fault.handle_pte_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
      0.00            +7.2        7.25 ±  2%  perf-profile.calltrace.cycles-pp.copy_page.handle_pte_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
     41.41 ± 70%     +22.3       63.67        perf-profile.calltrace.cycles-pp.handle_mm_fault.__do_page_fault.do_page_fault.page_fault
     42.19 ± 70%     +22.6       64.75        perf-profile.calltrace.cycles-pp.__do_page_fault.do_page_fault.page_fault
     42.20 ± 70%     +22.6       64.76        perf-profile.calltrace.cycles-pp.do_page_fault.page_fault
     42.27 ± 70%     +22.6       64.86        perf-profile.calltrace.cycles-pp.page_fault
      0.00           +44.9       44.88        perf-profile.calltrace.cycles-pp._raw_spin_lock.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_vma.handle_pte_fault
      0.00           +46.9       46.92        perf-profile.calltrace.cycles-pp.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_vma.handle_pte_fault.__handle_mm_fault
      0.00           +47.1       47.10        perf-profile.calltrace.cycles-pp.__alloc_pages_nodemask.alloc_pages_vma.handle_pte_fault.__handle_mm_fault.handle_mm_fault
      0.00           +47.4       47.37        perf-profile.calltrace.cycles-pp.alloc_pages_vma.handle_pte_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault
      0.00           +63.0       63.00        perf-profile.calltrace.cycles-pp.handle_pte_fault.__handle_mm_fault.handle_mm_fault.__do_page_fault.do_page_fault
      0.97 ±  6%      -0.1        0.84 ±  2%  perf-profile.children.cycles-pp.find_get_entry
      1.23 ±  6%      -0.1        1.11        perf-profile.children.cycles-pp.find_lock_entry
      0.09 ± 10%      -0.0        0.07 ±  6%  perf-profile.children.cycles-pp.unlock_page
      0.19 ±  4%      +0.0        0.21 ±  2%  perf-profile.children.cycles-pp.mem_cgroup_charge_statistics
      0.21 ±  2%      +0.0        0.25        perf-profile.children.cycles-pp.__mod_node_page_state
      0.00            +0.1        0.05 ±  8%  perf-profile.children.cycles-pp.get_vma_policy
      0.00            +0.1        0.08        perf-profile.children.cycles-pp.__lru_cache_add_active_or_unevictable
      0.00            +0.2        0.18 ±  2%  perf-profile.children.cycles-pp.__page_add_new_anon_rmap
      0.00            +1.3        1.30        perf-profile.children.cycles-pp.pte_map_lock
     63.40            +1.6       64.97        perf-profile.children.cycles-pp.__do_page_fault
     63.19            +1.6       64.83        perf-profile.children.cycles-pp.do_page_fault
     61.69            +1.7       63.36        perf-profile.children.cycles-pp.__handle_mm_fault
     63.19            +1.7       64.86        perf-profile.children.cycles-pp.page_fault
     61.99            +1.7       63.70        perf-profile.children.cycles-pp.handle_mm_fault
     72.27            +2.2       74.52        perf-profile.children.cycles-pp.native_queued_spin_lock_slowpath
     67.51            +2.4       69.87        perf-profile.children.cycles-pp._raw_spin_lock
     44.49 ±  3%      +3.0       47.45        perf-profile.children.cycles-pp.alloc_pages_vma
     44.28 ±  3%      +3.0       47.26        perf-profile.children.cycles-pp.__alloc_pages_nodemask
     44.13 ±  3%      +3.0       47.12        perf-profile.children.cycles-pp.get_page_from_freelist
      0.00           +63.1       63.06        perf-profile.children.cycles-pp.handle_pte_fault
      1.46 ±  7%      -0.5        1.01        perf-profile.self.cycles-pp._raw_spin_lock
      0.58 ±  6%      -0.2        0.34        perf-profile.self.cycles-pp.__handle_mm_fault
      0.55 ±  6%      -0.1        0.44 ±  2%  perf-profile.self.cycles-pp.find_get_entry
      0.22 ±  5%      -0.1        0.16 ±  2%  perf-profile.self.cycles-pp.alloc_set_pte
      0.10 ±  8%      -0.0        0.08        perf-profile.self.cycles-pp.down_read_trylock
      0.09 ±  5%      -0.0        0.07        perf-profile.self.cycles-pp.unlock_page
      0.06            -0.0        0.05        perf-profile.self.cycles-pp.pmd_devmap_trans_unstable
      0.20 ±  2%      +0.0        0.24 ±  3%  perf-profile.self.cycles-pp.__mod_node_page_state
      0.00            +0.1        0.05        perf-profile.self.cycles-pp.finish_fault
      0.00            +0.1        0.05        perf-profile.self.cycles-pp.get_vma_policy
      0.00            +0.1        0.08 ±  6%  perf-profile.self.cycles-pp.__lru_cache_add_active_or_unevictable
      0.00            +0.2        0.25        perf-profile.self.cycles-pp.handle_pte_fault
      0.00            +0.5        0.46 ±  7%  perf-profile.self.cycles-pp.pte_map_lock
     72.26            +2.3       74.52        perf-profile.self.cycles-pp.native_queued_spin_lock_slowpath

[-- Attachment #3: perf-profile.zip --]
[-- Type: application/zip, Size: 19025 bytes --]

  reply	other threads:[~2018-06-19  9:16 UTC|newest]

Thread overview: 106+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2018-05-17 11:06 [PATCH v11 00/26] Speculative page faults Laurent Dufour
2018-05-17 11:06 ` [PATCH v11 01/26] mm: introduce CONFIG_SPECULATIVE_PAGE_FAULT Laurent Dufour
2018-05-17 16:36   ` Randy Dunlap
2018-05-17 17:19     ` Matthew Wilcox
2018-05-17 17:34       ` Randy Dunlap
2018-05-22 12:00         ` [FIX PATCH " Laurent Dufour
2018-05-22 11:44       ` [PATCH " Laurent Dufour
2018-05-22 11:47     ` Laurent Dufour
2018-05-17 11:06 ` [PATCH v11 02/26] x86/mm: define ARCH_SUPPORTS_SPECULATIVE_PAGE_FAULT Laurent Dufour
2018-05-17 11:06 ` [PATCH v11 03/26] powerpc/mm: set ARCH_SUPPORTS_SPECULATIVE_PAGE_FAULT Laurent Dufour
2018-05-17 11:06 ` [PATCH v11 04/26] arm64/mm: define ARCH_SUPPORTS_SPECULATIVE_PAGE_FAULT Laurent Dufour
2018-05-17 11:06 ` [PATCH v11 05/26] mm: prepare for FAULT_FLAG_SPECULATIVE Laurent Dufour
2018-05-17 11:06 ` [PATCH v11 06/26] mm: introduce pte_spinlock " Laurent Dufour
2018-05-17 11:06 ` [PATCH v11 07/26] mm: make pte_unmap_same compatible with SPF Laurent Dufour
2018-05-17 11:06 ` [PATCH v11 08/26] mm: introduce INIT_VMA() Laurent Dufour
2018-05-17 11:06 ` [PATCH v11 09/26] mm: VMA sequence count Laurent Dufour
2018-05-17 11:06 ` [PATCH v11 10/26] mm: protect VMA modifications using " Laurent Dufour
2018-11-05  7:04   ` vinayak menon
2018-11-05  7:04     ` vinayak menon
2018-11-05 18:22     ` Laurent Dufour
2018-11-05 18:22       ` Laurent Dufour
2018-11-05 18:22       ` Laurent Dufour
2018-11-06  9:28       ` Vinayak Menon
2018-11-06  9:28         ` Vinayak Menon
2018-11-06  9:28         ` Vinayak Menon
2018-05-17 11:06 ` [PATCH v11 11/26] mm: protect mremap() against SPF hanlder Laurent Dufour
2018-05-17 11:06 ` [PATCH v11 12/26] mm: protect SPF handler against anon_vma changes Laurent Dufour
2018-05-17 11:06 ` [PATCH v11 13/26] mm: cache some VMA fields in the vm_fault structure Laurent Dufour
2018-05-17 11:06 ` [PATCH v11 14/26] mm/migrate: Pass vm_fault pointer to migrate_misplaced_page() Laurent Dufour
2018-05-17 11:06 ` [PATCH v11 15/26] mm: introduce __lru_cache_add_active_or_unevictable Laurent Dufour
2018-05-17 11:06 ` [PATCH v11 16/26] mm: introduce __vm_normal_page() Laurent Dufour
2018-05-17 11:06 ` [PATCH v11 17/26] mm: introduce __page_add_new_anon_rmap() Laurent Dufour
2018-05-17 11:06 ` [PATCH v11 18/26] mm: protect mm_rb tree with a rwlock Laurent Dufour
2018-05-17 11:06 ` [PATCH v11 19/26] mm: provide speculative fault infrastructure Laurent Dufour
2018-07-24 14:26   ` zhong jiang
2018-07-24 14:26     ` zhong jiang
2018-07-24 16:10     ` Laurent Dufour
2018-07-25  9:04       ` zhong jiang
2018-07-25  9:04         ` zhong jiang
2018-07-25 10:44         ` Laurent Dufour
2018-07-25 11:23           ` zhong jiang
2018-07-25 11:23             ` zhong jiang
2018-05-17 11:06 ` [PATCH v11 20/26] mm: adding speculative page fault failure trace events Laurent Dufour
2018-05-17 11:06 ` [PATCH v11 21/26] perf: add a speculative page fault sw event Laurent Dufour
2018-05-17 11:06 ` [PATCH v11 22/26] perf tools: add support for the SPF perf event Laurent Dufour
2018-05-17 11:06 ` [PATCH v11 23/26] mm: add speculative page fault vmstats Laurent Dufour
2018-05-17 11:06 ` [PATCH v11 24/26] x86/mm: add speculative pagefault handling Laurent Dufour
2018-05-17 11:06 ` [PATCH v11 25/26] powerpc/mm: add speculative page fault Laurent Dufour
2018-05-17 11:06 ` [PATCH v11 26/26] arm64/mm: " Laurent Dufour
2018-05-28  5:23 ` [PATCH v11 00/26] Speculative page faults Song, HaiyanX
2018-05-28  5:23   ` Song, HaiyanX
2018-05-28  7:51   ` Laurent Dufour
2018-05-28  8:22     ` Haiyan Song
2018-05-28  8:54       ` Laurent Dufour
2018-05-28 11:04         ` Wang, Kemi
2018-05-28 11:04           ` Wang, Kemi
2018-06-11  7:49         ` Song, HaiyanX
2018-06-11  7:49           ` Song, HaiyanX
2018-06-11 15:15           ` Laurent Dufour
2018-06-19  9:16             ` Haiyan Song [this message]
2018-06-19  9:16               ` Haiyan Song
2018-07-02  8:59           ` Laurent Dufour
2018-07-04  3:23             ` Song, HaiyanX
2018-07-04  3:23               ` Song, HaiyanX
2018-07-04  7:51               ` Laurent Dufour
2018-07-04  7:51                 ` Laurent Dufour
2018-07-11 17:05                 ` Laurent Dufour
2018-07-11 17:05                   ` Laurent Dufour
2018-07-13  3:56                   ` Song, HaiyanX
2018-07-17  9:36                     ` Laurent Dufour
2018-07-17  9:36                       ` Laurent Dufour
2018-08-03  6:36                       ` Song, HaiyanX
2018-08-03  6:45                         ` Song, HaiyanX
2018-08-22 14:23                         ` Laurent Dufour
2018-08-22 14:23                           ` Laurent Dufour
2018-09-18  6:42                           ` Song, HaiyanX
2018-11-05 10:42 ` Balbir Singh
2018-11-05 10:42   ` Balbir Singh
2018-11-05 16:08   ` Laurent Dufour
2018-11-05 16:08     ` Laurent Dufour
2018-11-05 16:08     ` Laurent Dufour
2019-01-11 15:43 Vinayak Menon
2019-01-14 13:19 ` Vinayak Menon
2019-01-15  8:24   ` Laurent Dufour
2019-01-16 11:41     ` Vinayak Menon
2019-01-16 13:31       ` Laurent Dufour
2019-01-16 11:41     ` Vinayak Menon
2019-01-17 15:51       ` zhong jiang
2019-01-17 15:51         ` zhong jiang
2019-01-18  9:29         ` Laurent Dufour
2019-01-18 15:41           ` zhong jiang
2019-01-18 15:41             ` zhong jiang
2019-01-18 15:51             ` Laurent Dufour
2019-01-18 16:24         ` Laurent Dufour
2019-01-19 17:05           ` zhong jiang
2019-01-19 17:05             ` zhong jiang
2019-01-22 16:22           ` zhong jiang
2019-01-22 16:22             ` zhong jiang
2019-01-24  8:20             ` Laurent Dufour
2019-01-25 12:32               ` zhong jiang
2019-01-25 12:32                 ` zhong jiang
2019-01-28  8:59                 ` Laurent Dufour
2019-01-28 14:09                   ` zhong jiang
2019-01-28 14:09                     ` zhong jiang
2019-01-28 15:45                     ` Laurent Dufour
2019-01-29 15:40                       ` zhong jiang

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20180619091625.jfj7qwueisddas2h@haiyan.lkp.sh.intel.com \
    --to=haiyanx.song@intel.com \
    --cc=aarcange@redhat.com \
    --cc=ak@linux.intel.com \
    --cc=akpm@linux-foundation.org \
    --cc=alexei.starovoitov@gmail.com \
    --cc=aneesh.kumar@linux.vnet.ibm.com \
    --cc=benh@kernel.crashing.org \
    --cc=bsingharora@gmail.com \
    --cc=daniel.m.jordan@oracle.com \
    --cc=dave@stgolabs.net \
    --cc=haren@linux.vnet.ibm.com \
    --cc=hpa@zytor.com \
    --cc=jack@suse.cz \
    --cc=jglisse@redhat.com \
    --cc=kemi.wang@intel.com \
    --cc=khandual@linux.vnet.ibm.com \
    --cc=kirill@shutemov.name \
    --cc=ldufour@linux.vnet.ibm.com \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-mm@kvack.org \
    --cc=linuxppc-dev@lists.ozlabs.org \
    --cc=mhocko@kernel.org \
    --cc=minchan@kernel.org \
    --cc=mingo@redhat.com \
    --cc=mpe@ellerman.id.au \
    --cc=npiggin@gmail.com \
    --cc=opensource.ganesh@gmail.com \
    --cc=paulmck@linux.vnet.ibm.com \
    --cc=paulus@samba.org \
    --cc=peterz@infradead.org \
    --cc=punitagrawal@gmail.com \
    --cc=rientjes@google.com \
    --cc=sergey.senozhatsky.work@gmail.com \
    --cc=sergey.senozhatsky@gmail.com \
    --cc=tglx@linutronix.de \
    --cc=tim.c.chen@linux.intel.com \
    --cc=vinayakm.list@gmail.com \
    --cc=will.deacon@arm.com \
    --cc=willy@infradead.org \
    --cc=x86@kernel.org \
    --cc=yang.shi@linux.alibaba.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.