linux-kernel.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: Vincent Guittot <vincent.guittot@linaro.org>
To: Xing Zhengjun <zhengjun.xing@linux.intel.com>
Cc: kernel test robot <rong.a.chen@intel.com>,
	Ingo Molnar <mingo@kernel.org>,
	Mel Gorman <mgorman@techsingularity.net>,
	Peter Zijlstra <a.p.zijlstra@chello.nl>,
	Juri Lelli <juri.lelli@redhat.com>,
	Steven Rostedt <rostedt@goodmis.org>,
	Valentin Schneider <valentin.schneider@arm.com>,
	Phil Auld <pauld@redhat.com>, Hillf Danton <hdanton@sina.com>,
	LKML <linux-kernel@vger.kernel.org>,
	lkp@lists.01.org
Subject: Re: [LKP] [sched/fair] 070f5e860e: reaim.jobs_per_min -10.5% regression
Date: Fri, 12 Jun 2020 17:19:07 +0200	[thread overview]
Message-ID: <20200612151907.GA11611@vingu-book> (raw)
In-Reply-To: <f5d91003-800c-5439-3159-67b5f7b6b744@linux.intel.com>

Le vendredi 12 juin 2020 à 14:36:49 (+0800), Xing Zhengjun a écrit :
> Hi Vincent,
> 
>   We test the regression still existed in v5.7, do you have time to look at
> it? Thanks.
> 
>

The commit 070f5e860ee2 moveis some cases from the state "group has spare capacity" to
to the state "group is overloaded". Typically when util_avg decreases significantly
after a migration but the group is in fact still overloaded.
The current rule uses a fix threshold but has the disavantage of possibly including
some cases with spare capacity but a high runnable_avg (because of tasks running
simultaneously as an example).
It looks like this benchmark is impacted by moving such cases from has_spare_capacity
to is_overloaded. I have a patch in my backlog  that tries to fix the problem but I 
never sent it because I failed to find a benchmark that will benefit from it.

This patch moves back some cases from overloaded state to "has spare capacity" state.

Could you make it a try ?

---
 kernel/sched/fair.c | 12 ++++++++++--
 1 file changed, 10 insertions(+), 2 deletions(-)

diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c
index 0ed04d2a8959..c24f85969591 100644
--- a/kernel/sched/fair.c
+++ b/kernel/sched/fair.c
@@ -8213,10 +8213,14 @@ static inline int sg_imbalanced(struct sched_group *group)
 static inline bool
 group_has_capacity(unsigned int imbalance_pct, struct sg_lb_stats *sgs)
 {
+	unsigned long imb;
+
 	if (sgs->sum_nr_running < sgs->group_weight)
 		return true;

-	if ((sgs->group_capacity * imbalance_pct) <
+	imb = imbalance_pct-100;
+	imb = sgs->sum_nr_running * imb + 100;
+	if ((sgs->group_capacity * imb) <
 			(sgs->group_runnable * 100))
 		return false;

@@ -8238,6 +8242,8 @@ group_has_capacity(unsigned int imbalance_pct, struct sg_lb_stats *sgs)
 static inline bool
 group_is_overloaded(unsigned int imbalance_pct, struct sg_lb_stats *sgs)
 {
+	unsigned long imb;
+
 	if (sgs->sum_nr_running <= sgs->group_weight)
 		return false;

@@ -8245,7 +8251,9 @@ group_is_overloaded(unsigned int imbalance_pct, struct sg_lb_stats *sgs)
 			(sgs->group_util * imbalance_pct))
 		return true;

-	if ((sgs->group_capacity * imbalance_pct) <
+	imb = imbalance_pct-100;
+	imb = sgs->sum_nr_running * imb + 100;
+	if ((sgs->group_capacity * imb) <
 			(sgs->group_runnable * 100))
 		return true;

--
2.17.1



> =========================================================================================
> tbox_group/testcase/rootfs/kconfig/compiler/runtime/nr_task/debug-setup/test/cpufreq_governor/ucode:
> 
> lkp-ivb-d04/reaim/debian-x86_64-20191114.cgz/x86_64-rhel-7.6/gcc-7/300s/100%/test/five_sec/performance/0x21
> 
> commit:
>   9f68395333ad7f5bfe2f83473fed363d4229f11c
>   070f5e860ee2bf588c99ef7b4c202451faa48236
>   v5.7
> 
> 9f68395333ad7f5b 070f5e860ee2bf588c99ef7b4c2                        v5.7
> ---------------- --------------------------- ---------------------------
>          %stddev     %change         %stddev     %change         %stddev
>              \          |                \          |                \
>       0.69           -10.3%       0.62            -9.1%       0.62
> reaim.child_systime
>       0.62            -1.0%       0.61            +0.5%       0.62
> reaim.child_utime
>      66870           -10.0%      60187            -7.6%      61787
> reaim.jobs_per_min
>      16717           -10.0%      15046            -7.6%      15446
> reaim.jobs_per_min_child
>      97.84            -1.1%      96.75            -0.4%      97.43
> reaim.jti
>      72000           -10.8%      64216            -8.3%      66000
> reaim.max_jobs_per_min
>       0.36           +10.6%       0.40            +7.8%       0.39
> reaim.parent_time
>       1.58 ±  2%     +71.0%       2.70 ±  2%     +26.9%       2.01 ± 2%
> reaim.std_dev_percent
>       0.00 ±  5%    +110.4%       0.01 ±  3%     +48.8%       0.01 ± 7%
> reaim.std_dev_time
>      50800            -2.4%      49600            -1.6%      50000
> reaim.workload
> 
> 
> On 3/19/2020 10:38 AM, kernel test robot wrote:
> > Greeting,
> > 
> > FYI, we noticed a -10.5% regression of reaim.jobs_per_min due to commit:
> > 
> > 
> > commit: 070f5e860ee2bf588c99ef7b4c202451faa48236 ("sched/fair: Take into account runnable_avg to classify group")
> > https://git.kernel.org/cgit/linux/kernel/git/next/linux-next.git master
> > 
> > in testcase: reaim
> > on test machine: 4 threads Intel(R) Core(TM) i3-3220 CPU @ 3.30GHz with 4G memory
> > with following parameters:
> > 
> > 	runtime: 300s
> > 	nr_task: 100%
> > 	test: five_sec
> > 	cpufreq_governor: performance
> > 	ucode: 0x21
> > 
> > test-description: REAIM is an updated and improved version of AIM 7 benchmark.
> > test-url: https://sourceforge.net/projects/re-aim-7/
> > 
> > 
> > 
> > If you fix the issue, kindly add following tag
> > Reported-by: kernel test robot <rong.a.chen@intel.com>
> > 
> > 
> > Details are as below:
> > -------------------------------------------------------------------------------------------------->
> > 
> > 
> > To reproduce:
> > 
> >          git clone https://github.com/intel/lkp-tests.git
> >          cd lkp-tests
> >          bin/lkp install job.yaml  # job file is attached in this email
> >          bin/lkp run     job.yaml
> > 
> > =========================================================================================
> > compiler/cpufreq_governor/kconfig/nr_task/rootfs/runtime/tbox_group/test/testcase/ucode:
> >    gcc-7/performance/x86_64-rhel-7.6/100%/debian-x86_64-20191114.cgz/300s/lkp-ivb-d04/five_sec/reaim/0x21
> > 
> > commit:
> >    9f68395333 ("sched/pelt: Add a new runnable average signal")
> >    070f5e860e ("sched/fair: Take into account runnable_avg to classify group")
> > 
> > 9f68395333ad7f5b 070f5e860ee2bf588c99ef7b4c2
> > ---------------- ---------------------------
> >         fail:runs  %reproduction    fail:runs
> >             |             |             |
> >            4:4          -18%           3:4     perf-profile.children.cycles-pp.error_entry
> >            3:4          -12%           3:4     perf-profile.self.cycles-pp.error_entry
> >           %stddev     %change         %stddev
> >               \          |                \
> >        0.68           -10.4%       0.61        reaim.child_systime
> >       67235           -10.5%      60195        reaim.jobs_per_min
> >       16808           -10.5%      15048        reaim.jobs_per_min_child
> >       97.90            -1.2%      96.70        reaim.jti
> >       72000           -10.8%      64216        reaim.max_jobs_per_min
> >        0.36           +11.3%       0.40        reaim.parent_time
> >        1.56 ±  3%     +79.1%       2.80 ±  6%  reaim.std_dev_percent
> >        0.00 ±  7%    +145.9%       0.01 ±  9%  reaim.std_dev_time
> >      104276           -16.0%      87616        reaim.time.involuntary_context_switches
> >    15511157            -2.4%   15144312        reaim.time.minor_page_faults
> >       55.00            -7.3%      51.00        reaim.time.percent_of_cpu_this_job_got
> >       88.01           -12.4%      77.12        reaim.time.system_time
> >       79.97            -3.2%      77.38        reaim.time.user_time
> >      216380            -3.4%     208924        reaim.time.voluntary_context_switches
> >       50800            -2.4%      49600        reaim.workload
> >       30.40 ±  2%      -4.7%      28.97 ±  2%  boot-time.boot
> >        9.38            -0.7        8.66 ±  3%  mpstat.cpu.all.sys%
> >        7452            +7.5%       8014        vmstat.system.cs
> >     1457802 ± 16%     +49.3%    2176122 ± 13%  cpuidle.C1.time
> >    48523684           +43.4%   69570233 ± 22%  cpuidle.C1E.time
> >      806543 ±  2%     +20.7%     973406 ± 11%  cpuidle.C1E.usage
> >       14328 ±  6%     +14.5%      16410 ±  8%  cpuidle.POLL.time
> >       43300 ±  4%     +13.5%      49150 ±  5%  softirqs.CPU0.SCHED
> >      118751            -9.3%     107763        softirqs.CPU1.RCU
> >       41679 ±  3%     +14.1%      47546 ±  4%  softirqs.CPU1.SCHED
> >       42688 ±  3%     +12.3%      47931 ±  4%  softirqs.CPU2.SCHED
> >       41730 ±  2%     +17.7%      49115 ±  4%  softirqs.CPU3.SCHED
> >      169399           +14.4%     193744 ±  2%  softirqs.SCHED
> >        3419            +1.0%       3453        proc-vmstat.nr_kernel_stack
> >    16365616            -1.8%   16077850        proc-vmstat.numa_hit
> >    16365616            -1.8%   16077850        proc-vmstat.numa_local
> >       93908            -1.6%      92389        proc-vmstat.pgactivate
> >    16269664            -3.9%   15629529 ±  2%  proc-vmstat.pgalloc_normal
> >    15918803            -2.3%   15557936        proc-vmstat.pgfault
> >    16644610            -2.0%   16310898        proc-vmstat.pgfree
> >       20125 ±123%    +161.7%      52662 ± 30%  sched_debug.cfs_rq:/.load.min
> >      348749 ± 10%     -11.2%     309562 ± 11%  sched_debug.cfs_rq:/.load.stddev
> >        1096 ±  6%     -14.4%     938.42 ±  7%  sched_debug.cfs_rq:/.load_avg.max
> >      448.46 ±  8%     -17.5%     370.19 ± 10%  sched_debug.cfs_rq:/.load_avg.stddev
> >      117372           -10.2%     105432        sched_debug.cfs_rq:/.min_vruntime.avg
> >      135242 ±  4%      -9.2%     122811        sched_debug.cfs_rq:/.min_vruntime.max
> >        0.53 ±  8%     +17.6%       0.62 ±  6%  sched_debug.cfs_rq:/.nr_running.avg
> >       29.79 ± 30%     -51.0%      14.58 ± 35%  sched_debug.cfs_rq:/.nr_spread_over.max
> >       10.21 ± 34%     -59.7%       4.12 ± 52%  sched_debug.cfs_rq:/.nr_spread_over.stddev
> >       78.25 ± 40%   +3304.7%       2664 ± 94%  sched_debug.cpu.curr->pid.min
> >      294309 ±  2%     +34.3%     395172 ± 12%  sched_debug.cpu.nr_switches.min
> >        9.58 ± 35%     +84.8%      17.71 ± 40%  sched_debug.cpu.nr_uninterruptible.max
> >       -6.88          +120.6%     -15.17        sched_debug.cpu.nr_uninterruptible.min
> >        6.41 ± 30%     +95.2%      12.52 ± 33%  sched_debug.cpu.nr_uninterruptible.stddev
> >      286185           +33.4%     381734 ± 13%  sched_debug.cpu.sched_count.min
> >      180416           +11.0%     200247        sched_debug.cpu.sched_goidle.avg
> >      116264 ±  3%     +44.6%     168090 ± 15%  sched_debug.cpu.sched_goidle.min
> >      476.00 ±  8%     +92.4%     915.75 ±  3%  interrupts.CAL:Function_call_interrupts
> >      110.50 ± 24%    +101.1%     222.25 ±  4%  interrupts.CPU0.CAL:Function_call_interrupts
> >        1381 ± 29%     +23.7%       1709 ± 26%  interrupts.CPU0.NMI:Non-maskable_interrupts
> >        1381 ± 29%     +23.7%       1709 ± 26%  interrupts.CPU0.PMI:Performance_monitoring_interrupts
> >        3319 ±  9%     +50.4%       4991 ±  2%  interrupts.CPU0.RES:Rescheduling_interrupts
> >       41.25 ± 30%    +274.5%     154.50 ± 15%  interrupts.CPU0.TLB:TLB_shootdowns
> >      116.25 ± 23%     +96.1%     228.00 ± 16%  interrupts.CPU1.CAL:Function_call_interrupts
> >        1183 ± 10%     +43.1%       1692 ± 23%  interrupts.CPU1.NMI:Non-maskable_interrupts
> >        1183 ± 10%     +43.1%       1692 ± 23%  interrupts.CPU1.PMI:Performance_monitoring_interrupts
> >        3335 ±  7%     +60.4%       5350 ±  5%  interrupts.CPU1.RES:Rescheduling_interrupts
> >       36.25 ± 30%    +344.1%     161.00 ±  8%  interrupts.CPU1.TLB:TLB_shootdowns
> >      131.25 ± 11%     +81.1%     237.75 ± 11%  interrupts.CPU2.CAL:Function_call_interrupts
> >        3247 ±  2%     +62.4%       5274        interrupts.CPU2.RES:Rescheduling_interrupts
> >       34.50 ± 36%    +357.2%     157.75 ±  7%  interrupts.CPU2.TLB:TLB_shootdowns
> >      118.00 ± 13%     +93.0%     227.75 ±  9%  interrupts.CPU3.CAL:Function_call_interrupts
> >        3155 ±  4%     +68.7%       5322 ±  3%  interrupts.CPU3.RES:Rescheduling_interrupts
> >       38.50 ± 16%    +303.9%     155.50 ±  3%  interrupts.CPU3.TLB:TLB_shootdowns
> >       13057 ±  2%     +60.4%      20939        interrupts.RES:Rescheduling_interrupts
> >      150.50 ± 27%    +317.8%     628.75 ±  3%  interrupts.TLB:TLB_shootdowns
> >        2.00            +0.1        2.09 ±  3%  perf-stat.i.branch-miss-rate%
> >       10.26            +1.1       11.36 ±  7%  perf-stat.i.cache-miss-rate%
> >     2009706 ±  2%      +5.4%    2117525 ±  3%  perf-stat.i.cache-misses
> >    16867421            -4.5%   16106908        perf-stat.i.cache-references
> >        7514            +7.6%       8083        perf-stat.i.context-switches
> >        1.51            -3.0%       1.47        perf-stat.i.cpi
> >   2.523e+09 ±  3%      -8.8%  2.301e+09 ±  2%  perf-stat.i.cpu-cycles
> >      124.54          +157.8%     321.08        perf-stat.i.cpu-migrations
> >        1842 ± 10%     -18.6%       1498 ±  6%  perf-stat.i.cycles-between-cache-misses
> >      752585 ±  2%      -4.1%     721714        perf-stat.i.dTLB-store-misses
> >      590441            +2.7%     606399        perf-stat.i.iTLB-load-misses
> >       68766            +4.0%      71488 ±  2%  perf-stat.i.iTLB-loads
> >   1.847e+09 ±  3%      -4.7%   1.76e+09 ±  2%  perf-stat.i.instructions
> >        3490 ±  4%      -8.5%       3195 ±  3%  perf-stat.i.instructions-per-iTLB-miss
> >        0.68            +3.7%       0.70        perf-stat.i.ipc
> >       51861            -2.1%      50797        perf-stat.i.minor-faults
> >       51861            -2.1%      50797        perf-stat.i.page-faults
> >        2.68 ±  2%      +0.1        2.78        perf-stat.overall.branch-miss-rate%
> >       11.91            +1.2       13.14 ±  2%  perf-stat.overall.cache-miss-rate%
> >        1.37            -4.3%       1.31        perf-stat.overall.cpi
> >        1255           -13.4%       1087 ±  2%  perf-stat.overall.cycles-between-cache-misses
> >        3127 ±  3%      -7.2%       2901 ±  2%  perf-stat.overall.instructions-per-iTLB-miss
> >        0.73            +4.5%       0.76        perf-stat.overall.ipc
> >     2002763 ±  2%      +5.4%    2110303 ±  3%  perf-stat.ps.cache-misses
> >    16809816            -4.5%   16051656        perf-stat.ps.cache-references
> >        7489            +7.6%       8055        perf-stat.ps.context-switches
> >   2.514e+09 ±  3%      -8.8%  2.293e+09 ±  2%  perf-stat.ps.cpu-cycles
> >      124.12          +157.8%     319.95        perf-stat.ps.cpu-migrations
> >      750010 ±  2%      -4.1%     719223        perf-stat.ps.dTLB-store-misses
> >      588424            +2.7%     604314        perf-stat.ps.iTLB-load-misses
> >       68533            +4.0%      71246 ±  2%  perf-stat.ps.iTLB-loads
> >   1.841e+09 ±  3%      -4.7%  1.754e+09 ±  2%  perf-stat.ps.instructions
> >       51683            -2.1%      50622        perf-stat.ps.minor-faults
> >       51683            -2.1%      50622        perf-stat.ps.page-faults
> >   5.577e+11 ±  3%      -5.1%  5.292e+11 ±  2%  perf-stat.total.instructions
> >        7.35 ± 17%      -2.7        4.60 ± 10%  perf-profile.calltrace.cycles-pp.serial8250_console_putchar.uart_console_write.serial8250_console_write.console_unlock.vprintk_emit
> >        7.74 ± 20%      -2.7        5.00 ±  6%  perf-profile.calltrace.cycles-pp.wait_for_xmitr.serial8250_console_putchar.uart_console_write.serial8250_console_write.console_unlock
> >       10.14 ±  8%      -2.7        7.44 ±  6%  perf-profile.calltrace.cycles-pp.new_sync_write.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe
> >       10.66 ±  8%      -2.6        8.07 ±  8%  perf-profile.calltrace.cycles-pp.vfs_write.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write
> >        7.10 ± 17%      -2.4        4.69 ±  7%  perf-profile.calltrace.cycles-pp.write._fini
> >        7.10 ± 17%      -2.4        4.69 ±  7%  perf-profile.calltrace.cycles-pp._fini
> >        7.09 ± 17%      -2.4        4.69 ±  7%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.write._fini
> >        7.09 ± 17%      -2.4        4.69 ±  7%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.write._fini
> >        7.09 ± 17%      -2.4        4.69 ±  7%  perf-profile.calltrace.cycles-pp.ksys_write.do_syscall_64.entry_SYSCALL_64_after_hwframe.write._fini
> >        7.09 ± 17%      -2.4        4.69 ±  7%  perf-profile.calltrace.cycles-pp.devkmsg_write.new_sync_write.vfs_write.ksys_write.do_syscall_64
> >        7.09 ± 17%      -2.4        4.69 ±  7%  perf-profile.calltrace.cycles-pp.vprintk_emit.devkmsg_emit.devkmsg_write.new_sync_write.vfs_write
> >        7.09 ± 17%      -2.4        4.69 ±  7%  perf-profile.calltrace.cycles-pp.devkmsg_emit.devkmsg_write.new_sync_write.vfs_write.ksys_write
> >        6.20 ±  8%      -2.1        4.08 ±  5%  perf-profile.calltrace.cycles-pp.console_unlock.vprintk_emit.devkmsg_emit.devkmsg_write.new_sync_write
> >        5.15 ± 11%      -1.8        3.38 ±  4%  perf-profile.calltrace.cycles-pp.serial8250_console_write.console_unlock.vprintk_emit.devkmsg_emit.devkmsg_write
> >        5.05 ± 11%      -1.7        3.31 ±  3%  perf-profile.calltrace.cycles-pp.uart_console_write.serial8250_console_write.console_unlock.vprintk_emit.devkmsg_emit
> >        7.41 ± 10%      -1.1        6.29 ±  5%  perf-profile.calltrace.cycles-pp.__do_execve_file.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
> >        7.57 ± 11%      -1.1        6.46 ±  5%  perf-profile.calltrace.cycles-pp.execve
> >        7.46 ± 10%      -1.1        6.37 ±  5%  perf-profile.calltrace.cycles-pp.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
> >        7.46 ± 10%      -1.1        6.37 ±  5%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.execve
> >        7.46 ± 10%      -1.1        6.37 ±  5%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.execve
> >        7.03 ±  5%      -1.1        5.95 ± 10%  perf-profile.calltrace.cycles-pp.brk
> >        5.90 ±  7%      -0.9        4.98 ± 10%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.brk
> >        5.84 ±  7%      -0.9        4.93 ±  9%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.brk
> >       15.77 ±  2%      -0.9       14.88 ±  2%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe
> >       15.86 ±  2%      -0.9       14.97 ±  2%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe
> >        3.88 ±  6%      -0.9        2.99 ±  5%  perf-profile.calltrace.cycles-pp.kill
> >        1.70 ± 23%      -0.8        0.90 ± 10%  perf-profile.calltrace.cycles-pp.delay_tsc.wait_for_xmitr.serial8250_console_putchar.uart_console_write.serial8250_console_write
> >        4.88 ±  8%      -0.8        4.08 ±  8%  perf-profile.calltrace.cycles-pp.__x64_sys_brk.do_syscall_64.entry_SYSCALL_64_after_hwframe.brk
> >        2.39 ± 27%      -0.7        1.67 ±  5%  perf-profile.calltrace.cycles-pp.flush_old_exec.load_elf_binary.search_binary_handler.__do_execve_file.__x64_sys_execve
> >        2.29 ± 30%      -0.7        1.59 ±  5%  perf-profile.calltrace.cycles-pp.mmput.flush_old_exec.load_elf_binary.search_binary_handler.__do_execve_file
> >        2.27 ± 30%      -0.7        1.58 ±  5%  perf-profile.calltrace.cycles-pp.exit_mmap.mmput.flush_old_exec.load_elf_binary.search_binary_handler
> >        3.11 ±  5%      -0.6        2.47 ±  9%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.kill
> >        3.07 ±  5%      -0.6        2.45 ±  9%  perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.kill
> >        2.09 ± 18%      -0.4        1.67 ±  3%  perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.read
> >        2.82 ±  9%      -0.4        2.40 ± 12%  perf-profile.calltrace.cycles-pp.mmput.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64
> >        2.80 ±  9%      -0.4        2.38 ± 12%  perf-profile.calltrace.cycles-pp.exit_mmap.mmput.do_exit.do_group_exit.__x64_sys_exit_group
> >        1.11 ± 33%      -0.4        0.71 ± 10%  perf-profile.calltrace.cycles-pp.unmap_vmas.exit_mmap.mmput.flush_old_exec.load_elf_binary
> >        1.05 ± 15%      -0.4        0.69 ± 13%  perf-profile.calltrace.cycles-pp.vt_console_print.console_unlock.vprintk_emit.devkmsg_emit.devkmsg_write
> >        1.03 ± 17%      -0.4        0.68 ± 13%  perf-profile.calltrace.cycles-pp.lf.vt_console_print.console_unlock.vprintk_emit.devkmsg_emit
> >        1.03 ± 17%      -0.4        0.68 ± 13%  perf-profile.calltrace.cycles-pp.con_scroll.lf.vt_console_print.console_unlock.vprintk_emit
> >        1.03 ± 17%      -0.4        0.68 ± 13%  perf-profile.calltrace.cycles-pp.fbcon_scroll.con_scroll.lf.vt_console_print.console_unlock
> >        0.96 ± 16%      -0.3        0.66 ± 12%  perf-profile.calltrace.cycles-pp.fbcon_putcs.fbcon_redraw.fbcon_scroll.con_scroll.lf
> >        1.85 ±  4%      -0.3        1.58 ±  8%  perf-profile.calltrace.cycles-pp.alloc_pages_vma.handle_pte_fault.__handle_mm_fault.handle_mm_fault.do_page_fault
> >        0.89 ± 15%      -0.3        0.62 ± 12%  perf-profile.calltrace.cycles-pp.kill_pid_info.kill_something_info.__x64_sys_kill.do_syscall_64.entry_SYSCALL_64_after_hwframe
> >        1.67 ±  5%      -0.3        1.41 ±  9%  perf-profile.calltrace.cycles-pp.__alloc_pages_nodemask.alloc_pages_vma.handle_pte_fault.__handle_mm_fault.handle_mm_fault
> >        1.02 ±  7%      -0.3        0.77 ± 12%  perf-profile.calltrace.cycles-pp.do_signal.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe.kill
> >        0.94 ± 16%      -0.2        0.70 ±  5%  perf-profile.calltrace.cycles-pp.clear_page_erms.prep_new_page.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_vma
> >        0.98 ± 16%      -0.2        0.74 ±  7%  perf-profile.calltrace.cycles-pp.prep_new_page.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_vma.handle_pte_fault
> >        1.03 ±  6%      -0.2        0.79 ± 10%  perf-profile.calltrace.cycles-pp.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe.kill
> >        1.00 ± 10%      -0.2        0.77 ±  9%  perf-profile.calltrace.cycles-pp.new_sync_read.vfs_read.ksys_read.do_syscall_64.entry_SYSCALL_64_after_hwframe
> >        0.87 ± 10%      -0.2        0.66 ± 15%  perf-profile.calltrace.cycles-pp.shmem_file_read_iter.new_sync_read.vfs_read.ksys_read.do_syscall_64
> >        1.41 ±  3%      -0.2        1.23 ±  7%  perf-profile.calltrace.cycles-pp.get_page_from_freelist.__alloc_pages_nodemask.alloc_pages_vma.handle_pte_fault.__handle_mm_fault
> >        1.88 ±  5%      -0.1        1.73        perf-profile.calltrace.cycles-pp.__x64_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe
> >        1.87 ±  5%      -0.1        1.73        perf-profile.calltrace.cycles-pp._do_fork.__x64_sys_clone.do_syscall_64.entry_SYSCALL_64_after_hwframe
> >       10.34 ± 11%      +7.3       17.66 ±  8%  perf-profile.calltrace.cycles-pp.cpuidle_enter.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64
> >       10.18 ± 11%      +7.3       17.52 ±  8%  perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.do_idle.cpu_startup_entry.start_secondary
> >       11.32 ±  9%      +7.7       19.03 ±  8%  perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64
> >       11.32 ±  9%      +7.7       19.05 ±  8%  perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.secondary_startup_64
> >       11.32 ±  9%      +7.7       19.05 ±  8%  perf-profile.calltrace.cycles-pp.start_secondary.secondary_startup_64
> >       11.02 ±  5%      +8.1       19.14 ±  7%  perf-profile.calltrace.cycles-pp.intel_idle.cpuidle_enter_state.cpuidle_enter.do_idle.cpu_startup_entry
> >       16.04 ±  6%      +9.1       25.17 ±  8%  perf-profile.calltrace.cycles-pp.secondary_startup_64
> >       55.98            -7.0       48.94 ±  4%  perf-profile.children.cycles-pp.entry_SYSCALL_64_after_hwframe
> >       55.67            -7.0       48.67 ±  4%  perf-profile.children.cycles-pp.do_syscall_64
> >       10.60 ± 16%      -3.3        7.30 ±  8%  perf-profile.children.cycles-pp.vprintk_emit
> >       13.02 ±  7%      -3.0        9.99 ±  7%  perf-profile.children.cycles-pp.write
> >        9.92 ± 13%      -2.8        7.08 ±  4%  perf-profile.children.cycles-pp.console_unlock
> >       10.26 ±  8%      -2.7        7.53 ±  6%  perf-profile.children.cycles-pp.new_sync_write
> >       10.79 ±  8%      -2.6        8.18 ±  8%  perf-profile.children.cycles-pp.vfs_write
> >       10.95 ±  8%      -2.6        8.36 ±  8%  perf-profile.children.cycles-pp.ksys_write
> >        7.17 ± 16%      -2.5        4.69 ±  7%  perf-profile.children.cycles-pp.devkmsg_write
> >        7.17 ± 16%      -2.5        4.69 ±  7%  perf-profile.children.cycles-pp.devkmsg_emit
> >        8.65 ± 16%      -2.4        6.21 ±  4%  perf-profile.children.cycles-pp.serial8250_console_write
> >        8.53 ± 17%      -2.4        6.11 ±  4%  perf-profile.children.cycles-pp.uart_console_write
> >        7.13 ± 17%      -2.4        4.71 ±  6%  perf-profile.children.cycles-pp._fini
> >        8.46 ± 16%      -2.4        6.07 ±  4%  perf-profile.children.cycles-pp.wait_for_xmitr
> >        8.34 ± 16%      -2.4        5.97 ±  4%  perf-profile.children.cycles-pp.serial8250_console_putchar
> >        5.80 ± 16%      -1.6        4.21 ±  6%  perf-profile.children.cycles-pp.io_serial_in
> >        7.85 ± 10%      -1.2        6.67 ±  5%  perf-profile.children.cycles-pp.execve
> >        7.72 ± 11%      -1.2        6.55 ±  5%  perf-profile.children.cycles-pp.__do_execve_file
> >        5.19 ± 13%      -1.1        4.05 ±  8%  perf-profile.children.cycles-pp.mmput
> >        5.16 ± 13%      -1.1        4.03 ±  8%  perf-profile.children.cycles-pp.exit_mmap
> >        7.76 ± 10%      -1.1        6.64 ±  5%  perf-profile.children.cycles-pp.__x64_sys_execve
> >        7.11 ±  5%      -1.1        6.01 ± 10%  perf-profile.children.cycles-pp.brk
> >        3.92 ±  6%      -0.9        3.03 ±  5%  perf-profile.children.cycles-pp.kill
> >        2.63 ± 17%      -0.8        1.85        perf-profile.children.cycles-pp.delay_tsc
> >        4.89 ±  8%      -0.8        4.12 ±  8%  perf-profile.children.cycles-pp.__x64_sys_brk
> >        2.48 ± 27%      -0.7        1.74 ±  4%  perf-profile.children.cycles-pp.flush_old_exec
> >        3.02 ± 12%      -0.7        2.28 ± 12%  perf-profile.children.cycles-pp.unmap_page_range
> >        3.15 ± 11%      -0.7        2.40 ± 12%  perf-profile.children.cycles-pp.unmap_vmas
> >        2.25 ± 19%      -0.5        1.75 ± 11%  perf-profile.children.cycles-pp.unmap_region
> >        1.27 ± 11%      -0.4        0.86 ±  8%  perf-profile.children.cycles-pp.vt_console_print
> >        1.24 ± 12%      -0.4        0.85 ±  9%  perf-profile.children.cycles-pp.lf
> >        1.24 ± 12%      -0.4        0.85 ±  9%  perf-profile.children.cycles-pp.con_scroll
> >        1.24 ± 12%      -0.4        0.85 ±  9%  perf-profile.children.cycles-pp.fbcon_scroll
> >        1.79 ±  9%      -0.4        1.41 ±  4%  perf-profile.children.cycles-pp.release_pages
> >        1.22 ± 11%      -0.4        0.85 ±  9%  perf-profile.children.cycles-pp.fbcon_redraw
> >        1.17 ± 12%      -0.4        0.82 ± 10%  perf-profile.children.cycles-pp.fbcon_putcs
> >        1.16 ± 13%      -0.3        0.82 ± 10%  perf-profile.children.cycles-pp.bit_putcs
> >        0.90 ± 16%      -0.3        0.62 ± 12%  perf-profile.children.cycles-pp.kill_pid_info
> >        0.95 ± 10%      -0.3        0.68 ±  6%  perf-profile.children.cycles-pp.drm_fb_helper_cfb_imageblit
> >        0.95 ± 11%      -0.3        0.68 ±  6%  perf-profile.children.cycles-pp.cfb_imageblit
> >        1.24 ±  7%      -0.2        1.01 ±  6%  perf-profile.children.cycles-pp.new_sync_read
> >        0.71 ±  4%      -0.2        0.49 ± 23%  perf-profile.children.cycles-pp.___perf_sw_event
> >        0.55 ± 31%      -0.2        0.33 ± 16%  perf-profile.children.cycles-pp.unlink_anon_vmas
> >        0.89 ± 11%      -0.2        0.67 ± 15%  perf-profile.children.cycles-pp.shmem_file_read_iter
> >        0.60 ± 20%      -0.2        0.39 ± 20%  perf-profile.children.cycles-pp.__send_signal
> >        1.06 ±  6%      -0.2        0.85 ± 16%  perf-profile.children.cycles-pp.pagevec_lru_move_fn
> >        0.88            -0.2        0.68 ±  6%  perf-profile.children.cycles-pp.__perf_sw_event
> >        1.49 ±  5%      -0.2        1.29 ±  7%  perf-profile.children.cycles-pp.prepare_exit_to_usermode
> >        0.56 ± 12%      -0.2        0.37 ± 11%  perf-profile.children.cycles-pp.do_send_sig_info
> >        1.65 ±  8%      -0.2        1.47 ±  4%  perf-profile.children.cycles-pp.perf_event_mmap
> >        0.69 ±  2%      -0.2        0.52 ± 16%  perf-profile.children.cycles-pp.page_remove_rmap
> >        0.61 ±  5%      -0.2        0.44 ± 15%  perf-profile.children.cycles-pp.free_unref_page_list
> >        0.60 ±  6%      -0.2        0.43 ± 15%  perf-profile.children.cycles-pp.__vm_munmap
> >        0.77 ± 12%      -0.2        0.62 ± 12%  perf-profile.children.cycles-pp.__might_sleep
> >        0.39 ± 12%      -0.2        0.24 ± 18%  perf-profile.children.cycles-pp.time
> >        0.46 ± 14%      -0.1        0.34 ± 14%  perf-profile.children.cycles-pp.lru_add_drain_cpu
> >        0.57 ±  8%      -0.1        0.47 ± 14%  perf-profile.children.cycles-pp.shmem_undo_range
> >        0.41 ± 12%      -0.1        0.30 ± 15%  perf-profile.children.cycles-pp.copy_fpstate_to_sigframe
> >        0.76 ±  7%      -0.1        0.67 ±  8%  perf-profile.children.cycles-pp.__x64_sys_rt_sigreturn
> >        0.26 ± 16%      -0.1        0.17 ± 17%  perf-profile.children.cycles-pp.mark_page_accessed
> >        0.12 ± 47%      -0.1        0.04 ±103%  perf-profile.children.cycles-pp.sigaction
> >        0.23 ± 12%      -0.1        0.15 ± 11%  perf-profile.children.cycles-pp.__vm_enough_memory
> >        0.12 ± 18%      -0.1        0.05 ±106%  perf-profile.children.cycles-pp.__vsprintf_chk
> >        0.23 ± 20%      -0.1        0.17 ± 13%  perf-profile.children.cycles-pp.d_add
> >        0.13 ± 23%      -0.1        0.07 ± 58%  perf-profile.children.cycles-pp.fput_many
> >        0.13 ± 14%      -0.1        0.08 ± 24%  perf-profile.children.cycles-pp.vfs_unlink
> >        0.11 ± 20%      -0.0        0.07 ±  7%  perf-profile.children.cycles-pp.__update_load_avg_cfs_rq
> >        0.04 ± 63%      +0.0        0.08 ± 23%  perf-profile.children.cycles-pp.uncharge_page
> >        0.06 ± 22%      +0.0        0.10 ± 36%  perf-profile.children.cycles-pp.sched_exec
> >        0.44 ±  4%      +0.0        0.48 ±  4%  perf-profile.children.cycles-pp.close
> >        0.14 ± 22%      +0.1        0.21 ± 17%  perf-profile.children.cycles-pp.pick_next_task_fair
> >        0.10 ± 17%      +0.1        0.17 ± 23%  perf-profile.children.cycles-pp.__anon_vma_prepare
> >        0.00            +0.1        0.07 ± 24%  perf-profile.children.cycles-pp.update_sd_lb_stats
> >        0.07 ± 34%      +0.1        0.15 ± 42%  perf-profile.children.cycles-pp.file_free_rcu
> >        0.15 ± 27%      +0.1        0.23 ± 21%  perf-profile.children.cycles-pp.__strcasecmp
> >        0.20 ± 21%      +0.1        0.29 ±  8%  perf-profile.children.cycles-pp.__pte_alloc
> >        0.14 ± 47%      +0.1        0.23 ± 27%  perf-profile.children.cycles-pp.update_blocked_averages
> >        0.09 ± 44%      +0.1        0.19 ± 18%  perf-profile.children.cycles-pp.schedule_idle
> >        0.00            +0.1        0.10 ± 33%  perf-profile.children.cycles-pp.newidle_balance
> >        0.00            +0.1        0.10 ± 18%  perf-profile.children.cycles-pp.__vmalloc_node_range
> >        0.21 ± 15%      +0.1        0.32 ± 25%  perf-profile.children.cycles-pp.__wake_up_common
> >        0.63 ±  8%      +0.1        0.77 ±  6%  perf-profile.children.cycles-pp.rcu_do_batch
> >        0.76 ± 14%      +0.1        0.90 ±  9%  perf-profile.children.cycles-pp.rcu_core
> >        0.07 ± 90%      +0.2        0.27 ±109%  perf-profile.children.cycles-pp.security_mmap_addr
> >        0.46 ± 26%      +0.3        0.75 ± 13%  perf-profile.children.cycles-pp.__sched_text_start
> >       11.32 ±  9%      +7.7       19.05 ±  8%  perf-profile.children.cycles-pp.start_secondary
> >       11.03 ±  5%      +8.1       19.16 ±  7%  perf-profile.children.cycles-pp.intel_idle
> >       14.78 ±  6%      +8.5       23.24 ±  8%  perf-profile.children.cycles-pp.cpuidle_enter
> >       14.76 ±  6%      +8.5       23.24 ±  8%  perf-profile.children.cycles-pp.cpuidle_enter_state
> >       16.04 ±  6%      +9.1       25.17 ±  8%  perf-profile.children.cycles-pp.secondary_startup_64
> >       16.04 ±  6%      +9.1       25.17 ±  8%  perf-profile.children.cycles-pp.cpu_startup_entry
> >       16.04 ±  6%      +9.1       25.19 ±  8%  perf-profile.children.cycles-pp.do_idle
> >        5.79 ± 16%      -1.6        4.21 ±  6%  perf-profile.self.cycles-pp.io_serial_in
> >        2.62 ± 17%      -0.8        1.85        perf-profile.self.cycles-pp.delay_tsc
> >        5.11 ±  4%      -0.6        4.56 ±  5%  perf-profile.self.cycles-pp.do_syscall_64
> >        1.44 ±  6%      -0.3        1.15 ±  5%  perf-profile.self.cycles-pp.unmap_page_range
> >        0.94 ± 11%      -0.3        0.68 ±  6%  perf-profile.self.cycles-pp.cfb_imageblit
> >        0.65 ±  6%      -0.2        0.42 ± 23%  perf-profile.self.cycles-pp.___perf_sw_event
> >        1.42 ±  5%      -0.2        1.22 ±  7%  perf-profile.self.cycles-pp.prepare_exit_to_usermode
> >        0.65 ± 13%      -0.2        0.47 ±  9%  perf-profile.self.cycles-pp.do_page_fault
> >        0.65 ±  9%      -0.1        0.52 ±  5%  perf-profile.self.cycles-pp.release_pages
> >        0.24 ± 20%      -0.1        0.15 ± 16%  perf-profile.self.cycles-pp.mark_page_accessed
> >        0.16 ± 28%      -0.1        0.08 ± 69%  perf-profile.self.cycles-pp.free_unref_page_commit
> >        0.12 ± 24%      -0.1        0.04 ± 59%  perf-profile.self.cycles-pp.__do_munmap
> >        0.10 ± 24%      -0.0        0.06 ±  7%  perf-profile.self.cycles-pp.__update_load_avg_cfs_rq
> >        0.04 ± 57%      +0.0        0.07 ± 19%  perf-profile.self.cycles-pp.__sbrk
> >        0.04 ± 57%      +0.0        0.08 ± 23%  perf-profile.self.cycles-pp.update_load_avg
> >        0.04 ± 57%      +0.0        0.08 ± 23%  perf-profile.self.cycles-pp.uncharge_page
> >        0.26 ± 11%      +0.1        0.39 ± 12%  perf-profile.self.cycles-pp.copy_page
> >        0.49 ± 13%      +0.1        0.63 ± 13%  perf-profile.self.cycles-pp.get_page_from_freelist
> >       11.00 ±  5%      +8.1       19.14 ±  7%  perf-profile.self.cycles-pp.intel_idle
> > 
> > 
> >                               reaim.time.system_time
> >    90 +----------------------------------------------------------------------+
> >       |           +.   .+       ++.++                               .+       |
> >    88 |+.+ +.++. +  +++  ++.++.+     +.++.+++.+++.+++. +.++ .+     +  :.+++.+|
> >       |   +     +                                     +    +  ++.++   +      |
> >    86 |-+                                                                    |
> >       |                                                                      |
> >    84 |-+                                                                    |
> >       |                                                                      |
> >    82 |-+  O                                                                 |
> >       |O OO  OO O                                                            |
> >    80 |-+                                                                    |
> >       |          OO OOO  O   O  O  O                                         |
> >    78 |-+               O O O  O O  OO  O                                    |
> >       |                                O  OOO O                              |
> >    76 +----------------------------------------------------------------------+
> >                        reaim.time.percent_of_cpu_this_job_got
> >      55 +--------------------------------------------------------------------+
> >         |                                                                    |
> >    54.5 |-+                                                                  |
> >      54 |-+                                                                  |
> >         |                                                                    |
> >    53.5 |-+                                                                  |
> >         |                                                                    |
> >      53 |-+OOO OOO                                                           |
> >         |                                                                    |
> >    52.5 |-+                                                                  |
> >      52 |O+        OOO OOO OOO OOO OO                                        |
> >         |                                                                    |
> >    51.5 |-+                                                                  |
> >         |                                                                    |
> >      51 +--------------------------------------------------------------------+
> >                                    reaim.parent_time
> >    0.405 +-------------------------------------------------------------------+
> >      0.4 |-+                           O                                     |
> >          |                            O O OOO OO                             |
> >    0.395 |-+                                                                 |
> >     0.39 |-+                                                                 |
> >    0.385 |-+    OOOO OOO OOO OOO OOOO                                        |
> >     0.38 |O+OOO                                                              |
> >          |                                                                   |
> >    0.375 |-+                                                                 |
> >     0.37 |-+                                                                 |
> >    0.365 |-+                                                                 |
> >     0.36 |-+                                                                 |
> >          |             +.            .+ +.   .+ +.   +       +.     +++. + .+|
> >    0.355 |+.  +.+ ++.++  +++. ++.++++  +  +++  +  +++ + ++.++  + +. :   + +  |
> >     0.35 +-------------------------------------------------------------------+
> >                                  reaim.child_systime
> >    0.69 +--------------------------------------------------------------------+
> >         |    +.+    ++. ++.+  .++      ++   +++.+  .++ .+++. +     .+ :  + .+|
> >    0.68 |+.++   ++.+   +    ++           +.+     ++   +     + +.+++   +.+ +  |
> >    0.67 |-+                                                                  |
> >         |                                                                    |
> >    0.66 |-+                                                                  |
> >    0.65 |-+                                                                  |
> >         |    O                                                               |
> >    0.64 |O+O   OOO                                                           |
> >    0.63 |-+ O                                                                |
> >         |          O   O       O O  O                                        |
> >    0.62 |-+         OO  O  O    O  O   O                                     |
> >    0.61 |-+              O  OO       O  OO OOOO                              |
> >         |                                                                    |
> >     0.6 +--------------------------------------------------------------------+
> >                                   reaim.jobs_per_min
> >    69000 +-------------------------------------------------------------------+
> >    68000 |-.++   +        + .++ .+                    .++   + .+++.+         |
> >          |+   +.+ ++.+++.+ +   +  +++.+++.+++.+++.++++   +.+ +      + +.+ +. |
> >    67000 |-+                                                         +   +  +|
> >    66000 |-+                                                                 |
> >          |                                                                   |
> >    65000 |-+                                                                 |
> >    64000 |-+                                                                 |
> >    63000 |-+                                                                 |
> >          |O OOO O O      O O                                                 |
> >    62000 |-+     O O OOO  O  OOO OOOO                                        |
> >    61000 |-+                                                                 |
> >          |                              O OOO OO                             |
> >    60000 |-+                          OO                                     |
> >    59000 +-------------------------------------------------------------------+
> >                                reaim.jobs_per_min_child
> >    17500 +-------------------------------------------------------------------+
> >          |                                                                   |
> >    17000 |-.+++. +        + .+++.+                    .++   + .+++.+         |
> >          |+     + ++.+++.+ +      +++.+++.+++.+++.++++   +.+ +      + +.+++.+|
> >          |                                                           +       |
> >    16500 |-+                                                                 |
> >          |                                                                   |
> >    16000 |-+                                                                 |
> >          |                                                                   |
> >    15500 |O+OOO OOOO OOO OOO OOO OOOO                                        |
> >          |                                                                   |
> >          |                                O O OO                             |
> >    15000 |-+                          OOO  O                                 |
> >          |                                                                   |
> >    14500 +-------------------------------------------------------------------+
> >                                 reaim.max_jobs_per_min
> >    76000 +-------------------------------------------------------------------+
> >          |                                                                   |
> >    74000 |-+  +                                                 +  +         |
> >          |    ::                                                : ::         |
> >          |   : :                                               : :: :        |
> >    72000 |+.++  ++++.+++.+++.+++.++++.+++.+++.+++.++++.+++.+++.+ +  +++.+++.+|
> >          |                                                                   |
> >    70000 |-+                                                                 |
> >          |                                                                   |
> >    68000 |-+ O                                                               |
> >          |                                                                   |
> >          |                                                                   |
> >    66000 |O+O O OOOO OOO OOO OOO OOOO                                        |
> >          |                                                                   |
> >    64000 +-------------------------------------------------------------------+
> > [*] bisect-good sample
> > [O] bisect-bad  sample
> > 
> > 
> > 
> > Disclaimer:
> > Results have been estimated based on internal Intel analysis and are provided
> > for informational purposes only. Any difference in system hardware or software
> > design or configuration may affect actual performance.
> > 
> > 
> > Thanks,
> > Rong Chen
> > 
> > 
> > _______________________________________________
> > LKP mailing list -- lkp@lists.01.org
> > To unsubscribe send an email to lkp-leave@lists.01.org
> > 
> 
> -- 
> Zhengjun Xing

  parent reply	other threads:[~2020-06-12 15:19 UTC|newest]

Thread overview: 21+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2020-03-19  2:38 [sched/fair] 070f5e860e: reaim.jobs_per_min -10.5% regression kernel test robot
2020-06-12  6:36 ` [LKP] " Xing Zhengjun
2020-06-12  7:07   ` Vincent Guittot
2020-06-12 15:19   ` Vincent Guittot [this message]
2020-06-15  7:37     ` Xing Zhengjun
     [not found]   ` <20200612110616.20264-1-hdanton@sina.com>
2020-06-12 15:23     ` Vincent Guittot
2020-06-15  7:26     ` Xing Zhengjun
2020-06-15  8:10       ` Vincent Guittot
2020-06-16  3:17         ` Xing Zhengjun
2020-06-16  6:54           ` Vincent Guittot
2020-06-17  0:30             ` Xing Zhengjun
2020-06-17 14:57               ` Vincent Guittot
2020-06-17 16:25                 ` Vincent Guittot
2020-06-18  2:45                   ` Xing Zhengjun
2020-06-18 12:35                     ` Vincent Guittot
2020-06-19  5:01                       ` Xing Zhengjun
     [not found]                   ` <20200618082406.8292-1-hdanton@sina.com>
2020-06-19  5:10                     ` Xing Zhengjun
2020-06-19  4:55                 ` Xing Zhengjun
2020-06-19  7:15                   ` Vincent Guittot
2020-06-24  9:04                     ` Vincent Guittot
     [not found]       ` <20200615151030.6480-1-hdanton@sina.com>
2020-06-16  3:24         ` Xing Zhengjun

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20200612151907.GA11611@vingu-book \
    --to=vincent.guittot@linaro.org \
    --cc=a.p.zijlstra@chello.nl \
    --cc=hdanton@sina.com \
    --cc=juri.lelli@redhat.com \
    --cc=linux-kernel@vger.kernel.org \
    --cc=lkp@lists.01.org \
    --cc=mgorman@techsingularity.net \
    --cc=mingo@kernel.org \
    --cc=pauld@redhat.com \
    --cc=rong.a.chen@intel.com \
    --cc=rostedt@goodmis.org \
    --cc=valentin.schneider@arm.com \
    --cc=zhengjun.xing@linux.intel.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).