From: Feng Tang <feng.tang@intel.com> To: Eric Dumazet <edumazet@google.com> Cc: kernel test robot <rong.a.chen@intel.com>, "David S. Miller" <davem@davemloft.net>, Willy Tarreau <w@1wt.eu>, Yue Cao <ycao009@ucr.edu>, LKML <linux-kernel@vger.kernel.org>, Linus Torvalds <torvalds@linux-foundation.org>, lkp@lists.01.org Subject: Re: [LKP] [net] 19f92a030c: apachebench.requests_per_second -37.9% regression Date: Wed, 13 Nov 2019 18:35:01 +0800 [thread overview] Message-ID: <20191113103501.GD65640@shbuild999.sh.intel.com> (raw) In-Reply-To: <20191108083513.GB29418@shao2-debian> Hi Eric, On Fri, Nov 08, 2019 at 04:35:13PM +0800, kernel test robot wrote: > Greeting, > > FYI, we noticed a -37.9% regression of apachebench.requests_per_second due to commit: > > commit: 19f92a030ca6d772ab44b22ee6a01378a8cb32d4 ("net: increase SOMAXCONN to 4096") > https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master Any thought on this? The test is actually: sysctl -w net.ipv4.tcp_syncookies=0 enable_apache_mod auth_basic authn_core authn_file authz_core authz_host authz_user access_compat systemctl restart apache2 ab -k -q -t 300 -n 1000000 -c 4000 127.0.0.1/ And some info about apachebench result is: w/o patch: Connection Times (ms) min mean[+/-sd] median max Connect: 0 0 19.5 0 7145 Processing: 0 4 110.0 3 21647 Waiting: 0 2 92.4 1 21646 Total: 0 4 121.1 3 24762 w/ patch: Connection Times (ms) min mean[+/-sd] median max Connect: 0 0 43.2 0 7143 Processing: 0 19 640.4 3 38708 Waiting: 0 24 796.5 1 38708 Total: 0 19 657.5 3 39725 Thanks, Feng > > in testcase: apachebench > on test machine: 16 threads Intel(R) Xeon(R) CPU D-1541 @ 2.10GHz with 48G memory > with following parameters: > > runtime: 300s > concurrency: 4000 > cluster: cs-localhost > cpufreq_governor: performance > ucode: 0x7000019 > > test-description: apachebench is a tool for benchmarking your Apache Hypertext Transfer Protocol (HTTP) server. > test-url: https://httpd.apache.org/docs/2.4/programs/ab.html > > In addition to that, the commit also has significant impact on the following tests: > > +------------------+------------------------------------------------------------------+ > | testcase: change | apachebench: apachebench.requests_per_second -37.5% regression | > | test machine | 16 threads Intel(R) Xeon(R) CPU D-1541 @ 2.10GHz with 48G memory | > | test parameters | cluster=cs-localhost | > | | concurrency=8000 | > | | cpufreq_governor=performance | > | | runtime=300s | > | | ucode=0x7000019 | > +------------------+------------------------------------------------------------------+ > > > If you fix the issue, kindly add following tag > Reported-by: kernel test robot <rong.a.chen@intel.com> > > > Details are as below: > --------------------------------------------------------------------------------------------------> > > > To reproduce: > > git clone https://github.com/intel/lkp-tests.git > cd lkp-tests > bin/lkp install job.yaml # job file is attached in this email > bin/lkp run job.yaml > > ========================================================================================= > cluster/compiler/concurrency/cpufreq_governor/kconfig/rootfs/runtime/tbox_group/testcase/ucode: > cs-localhost/gcc-7/4000/performance/x86_64-rhel-7.6/debian-x86_64-2019-09-23.cgz/300s/lkp-bdw-de1/apachebench/0x7000019 > > commit: > 6d6f0383b6 ("netdevsim: Fix use-after-free during device dismantle") > 19f92a030c ("net: increase SOMAXCONN to 4096") > > 6d6f0383b697f004 19f92a030ca6d772ab44b22ee6a > ---------------- --------------------------- > %stddev %change %stddev > \ | \ > 22640 ± 4% +71.1% 38734 apachebench.connection_time.processing.max > 24701 +60.9% 39743 apachebench.connection_time.total.max > 22639 ± 4% +71.1% 38734 apachebench.connection_time.waiting.max > 24701 +15042.0 39743 apachebench.max_latency.100% > 40454 -37.9% 25128 apachebench.requests_per_second > 25.69 +58.8% 40.79 apachebench.time.elapsed_time > 25.69 +58.8% 40.79 apachebench.time.elapsed_time.max > 79.00 -37.0% 49.75 apachebench.time.percent_of_cpu_this_job_got > 98.88 +61.0% 159.18 apachebench.time_per_request > 434631 -37.9% 269889 apachebench.transfer_rate > 1.5e+08 ± 18% +109.5% 3.141e+08 ± 27% cpuidle.C3.time > 578957 ± 7% +64.1% 949934 ± 12% cpuidle.C3.usage > 79085 ± 4% +24.8% 98720 meminfo.AnonHugePages > 41176 +14.2% 47013 meminfo.PageTables > 69429 -34.9% 45222 meminfo.max_used_kB > 63.48 +12.7 76.15 mpstat.cpu.all.idle% > 2.42 ± 2% -0.9 1.56 mpstat.cpu.all.soft% > 15.30 -5.2 10.13 mpstat.cpu.all.sys% > 18.80 -6.6 12.16 mpstat.cpu.all.usr% > 65.00 +17.7% 76.50 vmstat.cpu.id > 17.00 -35.3% 11.00 vmstat.cpu.us > 7.00 ± 24% -50.0% 3.50 ± 14% vmstat.procs.r > 62957 -33.3% 42012 vmstat.system.cs > 33174 -1.4% 32693 vmstat.system.in > 5394 ± 5% +16.3% 6272 ± 6% sched_debug.cfs_rq:/.min_vruntime.stddev > 5396 ± 5% +16.3% 6275 ± 6% sched_debug.cfs_rq:/.spread0.stddev > 33982 ± 48% -83.3% 5676 ± 47% sched_debug.cpu.avg_idle.min > 26.75 ± 77% +169.8% 72.17 ± 41% sched_debug.cpu.sched_count.avg > 212.00 ± 90% +168.2% 568.50 ± 50% sched_debug.cpu.sched_count.max > 52.30 ± 89% +182.5% 147.73 ± 48% sched_debug.cpu.sched_count.stddev > 11.33 ± 80% +193.9% 33.30 ± 42% sched_debug.cpu.sched_goidle.avg > 104.50 ± 92% +170.6% 282.75 ± 50% sched_debug.cpu.sched_goidle.max > 26.18 ± 90% +183.9% 74.31 ± 48% sched_debug.cpu.sched_goidle.stddev > 959.00 -32.0% 652.00 turbostat.Avg_MHz > 39.01 -11.2 27.79 turbostat.Busy% > 1.46 ± 7% -0.5 0.96 ± 5% turbostat.C1% > 9.58 ± 4% -3.2 6.38 turbostat.C1E% > 578646 ± 7% +64.1% 949626 ± 12% turbostat.C3 > 940073 +51.1% 1420298 turbostat.IRQ > 2.20 ± 22% +159.7% 5.71 ± 12% turbostat.Pkg%pc2 > 31.22 -17.2% 25.86 turbostat.PkgWatt > 4.74 -7.5% 4.39 turbostat.RAMWatt > 93184 -1.6% 91678 proc-vmstat.nr_active_anon > 92970 -1.8% 91314 proc-vmstat.nr_anon_pages > 288405 +1.0% 291286 proc-vmstat.nr_file_pages > 8307 +6.3% 8831 proc-vmstat.nr_kernel_stack > 10315 +14.2% 11783 proc-vmstat.nr_page_table_pages > 21499 +6.0% 22798 proc-vmstat.nr_slab_unreclaimable > 284131 +1.0% 286977 proc-vmstat.nr_unevictable > 93184 -1.6% 91678 proc-vmstat.nr_zone_active_anon > 284131 +1.0% 286977 proc-vmstat.nr_zone_unevictable > 198874 ± 2% +43.7% 285772 ± 16% proc-vmstat.numa_hit > 198874 ± 2% +43.7% 285772 ± 16% proc-vmstat.numa_local > 249594 ± 3% +59.6% 398267 ± 13% proc-vmstat.pgalloc_normal > 1216885 +12.7% 1371283 ± 3% proc-vmstat.pgfault > 179705 ± 16% +82.9% 328634 ± 21% proc-vmstat.pgfree > 346.25 ± 5% +133.5% 808.50 ± 2% slabinfo.TCPv6.active_objs > 346.25 ± 5% +134.4% 811.75 ± 2% slabinfo.TCPv6.num_objs > 22966 +15.6% 26559 slabinfo.anon_vma.active_objs > 23091 +15.5% 26664 slabinfo.anon_vma.num_objs > 69747 +16.1% 81011 slabinfo.anon_vma_chain.active_objs > 1094 +15.9% 1269 slabinfo.anon_vma_chain.active_slabs > 70092 +15.9% 81259 slabinfo.anon_vma_chain.num_objs > 1094 +15.9% 1269 slabinfo.anon_vma_chain.num_slabs > 1649 +12.9% 1861 slabinfo.cred_jar.active_objs > 1649 +12.9% 1861 slabinfo.cred_jar.num_objs > 4924 +20.0% 5907 slabinfo.pid.active_objs > 4931 +19.9% 5912 slabinfo.pid.num_objs > 266.50 ± 3% +299.2% 1063 slabinfo.request_sock_TCP.active_objs > 266.50 ± 3% +299.2% 1063 slabinfo.request_sock_TCP.num_objs > 11.50 ± 4% +1700.0% 207.00 ± 4% slabinfo.tw_sock_TCPv6.active_objs > 11.50 ± 4% +1700.0% 207.00 ± 4% slabinfo.tw_sock_TCPv6.num_objs > 41682 +16.0% 48360 slabinfo.vm_area_struct.active_objs > 1046 +15.7% 1211 slabinfo.vm_area_struct.active_slabs > 41879 +15.7% 48468 slabinfo.vm_area_struct.num_objs > 1046 +15.7% 1211 slabinfo.vm_area_struct.num_slabs > 4276 ± 2% +10.0% 4705 ± 2% slabinfo.vmap_area.num_objs > 21.25 ± 27% +3438.8% 752.00 ± 83% interrupts.36:IR-PCI-MSI.2621443-edge.eth0-TxRx-2 > 21.25 ± 20% +1777.6% 399.00 ±155% interrupts.38:IR-PCI-MSI.2621445-edge.eth0-TxRx-4 > 54333 +54.3% 83826 interrupts.CPU0.LOC:Local_timer_interrupts > 54370 +54.6% 84072 interrupts.CPU1.LOC:Local_timer_interrupts > 21.25 ± 27% +3438.8% 752.00 ± 83% interrupts.CPU10.36:IR-PCI-MSI.2621443-edge.eth0-TxRx-2 > 54236 +54.7% 83925 interrupts.CPU10.LOC:Local_timer_interrupts > 54223 +54.3% 83655 interrupts.CPU11.LOC:Local_timer_interrupts > 377.75 ± 21% +27.1% 480.25 ± 10% interrupts.CPU11.RES:Rescheduling_interrupts > 21.25 ± 20% +1777.6% 399.00 ±155% interrupts.CPU12.38:IR-PCI-MSI.2621445-edge.eth0-TxRx-4 > 54279 +54.1% 83646 interrupts.CPU12.LOC:Local_timer_interrupts > 53683 +55.3% 83365 interrupts.CPU13.LOC:Local_timer_interrupts > 53887 +55.7% 83903 interrupts.CPU14.LOC:Local_timer_interrupts > 54156 +54.7% 83803 interrupts.CPU15.LOC:Local_timer_interrupts > 54041 +55.1% 83806 interrupts.CPU2.LOC:Local_timer_interrupts > 54042 +55.4% 83991 interrupts.CPU3.LOC:Local_timer_interrupts > 54081 +55.2% 83938 interrupts.CPU4.LOC:Local_timer_interrupts > 54322 +54.9% 84166 interrupts.CPU5.LOC:Local_timer_interrupts > 53586 +56.5% 83849 interrupts.CPU6.LOC:Local_timer_interrupts > 54049 +55.2% 83892 interrupts.CPU7.LOC:Local_timer_interrupts > 54056 +54.9% 83751 interrupts.CPU8.LOC:Local_timer_interrupts > 53862 +54.7% 83331 interrupts.CPU9.LOC:Local_timer_interrupts > 865212 +55.0% 1340925 interrupts.LOC:Local_timer_interrupts > 16477 ± 4% +32.2% 21779 ± 8% softirqs.CPU0.TIMER > 18508 ± 15% +39.9% 25891 ± 17% softirqs.CPU1.TIMER > 16625 ± 8% +21.3% 20166 ± 7% softirqs.CPU10.TIMER > 5906 ± 21% +62.5% 9597 ± 13% softirqs.CPU12.SCHED > 17474 ± 12% +29.4% 22610 ± 7% softirqs.CPU12.TIMER > 7680 ± 11% +20.4% 9246 ± 14% softirqs.CPU13.SCHED > 45558 ± 36% -37.8% 28320 ± 25% softirqs.CPU14.NET_RX > 15365 ± 4% +40.7% 21622 ± 5% softirqs.CPU14.TIMER > 8084 ± 4% +18.7% 9599 ± 12% softirqs.CPU15.RCU > 16433 ± 4% +41.2% 23203 ± 14% softirqs.CPU2.TIMER > 8436 ± 7% +19.9% 10117 ± 10% softirqs.CPU3.RCU > 15992 ± 3% +48.5% 23742 ± 18% softirqs.CPU3.TIMER > 17389 ± 14% +38.7% 24116 ± 11% softirqs.CPU4.TIMER > 17749 ± 13% +42.2% 25235 ± 15% softirqs.CPU5.TIMER > 16528 ± 9% +28.3% 21200 ± 2% softirqs.CPU6.TIMER > 8321 ± 8% +31.3% 10929 ± 5% softirqs.CPU7.RCU > 18024 ± 8% +28.8% 23212 ± 5% softirqs.CPU7.TIMER > 15717 ± 5% +27.1% 19983 ± 7% softirqs.CPU8.TIMER > 7383 ± 11% +30.5% 9632 ± 9% softirqs.CPU9.SCHED > 16342 ± 18% +41.0% 23037 ± 10% softirqs.CPU9.TIMER > 148013 +10.2% 163086 ± 2% softirqs.RCU > 112139 +28.0% 143569 softirqs.SCHED > 276690 ± 3% +30.4% 360747 ± 3% softirqs.TIMER > 1.453e+09 -36.2% 9.273e+08 perf-stat.i.branch-instructions > 67671486 -35.9% 43396843 ± 2% perf-stat.i.branch-misses > 5.188e+08 -35.3% 3.357e+08 perf-stat.i.cache-misses > 5.188e+08 -35.3% 3.357e+08 perf-stat.i.cache-references > 71149 -36.0% 45536 perf-stat.i.context-switches > 2.92 ± 6% +38.4% 4.04 ± 5% perf-stat.i.cpi > 1.581e+10 -33.8% 1.046e+10 perf-stat.i.cpu-cycles > 1957 ± 6% -35.4% 1264 ± 5% perf-stat.i.cpu-migrations > 40.76 ± 2% +33.0% 54.21 perf-stat.i.cycles-between-cache-misses > 0.64 ± 2% +0.2 0.86 ± 2% perf-stat.i.dTLB-load-miss-rate% > 10720126 ± 2% -34.0% 7071299 perf-stat.i.dTLB-load-misses > 2.05e+09 -35.8% 1.315e+09 perf-stat.i.dTLB-loads > 0.16 +0.0 0.18 ± 5% perf-stat.i.dTLB-store-miss-rate% > 1688635 -31.7% 1153595 perf-stat.i.dTLB-store-misses > 1.17e+09 -33.5% 7.777e+08 perf-stat.i.dTLB-stores > 61.00 ± 6% +8.7 69.70 ± 2% perf-stat.i.iTLB-load-miss-rate% > 13112146 ± 10% -38.5% 8061589 ± 13% perf-stat.i.iTLB-load-misses > 9827689 ± 3% -37.1% 6184316 perf-stat.i.iTLB-loads > 7e+09 -36.2% 4.469e+09 perf-stat.i.instructions > 0.42 ± 2% -22.2% 0.33 ± 3% perf-stat.i.ipc > 45909 -32.4% 31036 perf-stat.i.minor-faults > 45909 -32.4% 31036 perf-stat.i.page-faults > 2.26 +3.6% 2.34 perf-stat.overall.cpi > 30.47 +2.3% 31.16 perf-stat.overall.cycles-between-cache-misses > 0.44 -3.5% 0.43 perf-stat.overall.ipc > 1.398e+09 -35.3% 9.049e+08 perf-stat.ps.branch-instructions > 65124290 -35.0% 42353120 ± 2% perf-stat.ps.branch-misses > 4.993e+08 -34.4% 3.276e+08 perf-stat.ps.cache-misses > 4.993e+08 -34.4% 3.276e+08 perf-stat.ps.cache-references > 68469 -35.1% 44440 perf-stat.ps.context-switches > 1.521e+10 -32.9% 1.021e+10 perf-stat.ps.cpu-cycles > 1884 ± 6% -34.5% 1234 ± 5% perf-stat.ps.cpu-migrations > 10314734 ± 2% -33.1% 6899548 perf-stat.ps.dTLB-load-misses > 1.973e+09 -34.9% 1.283e+09 perf-stat.ps.dTLB-loads > 1624883 -30.7% 1125668 perf-stat.ps.dTLB-store-misses > 1.126e+09 -32.6% 7.589e+08 perf-stat.ps.dTLB-stores > 12617676 ± 10% -37.7% 7866427 ± 13% perf-stat.ps.iTLB-load-misses > 9458687 ± 3% -36.2% 6036594 perf-stat.ps.iTLB-loads > 6.736e+09 -35.3% 4.361e+09 perf-stat.ps.instructions > 44179 -31.4% 30288 perf-stat.ps.minor-faults > 30791 +1.4% 31223 perf-stat.ps.msec > 44179 -31.4% 30288 perf-stat.ps.page-faults > 21.96 ± 69% -22.0 0.00 perf-profile.calltrace.cycles-pp.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe > 21.96 ± 69% -22.0 0.00 perf-profile.calltrace.cycles-pp.__do_execve_file.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe > 18.63 ± 89% -18.6 0.00 perf-profile.calltrace.cycles-pp.search_binary_handler.__do_execve_file.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe > 18.63 ± 89% -18.6 0.00 perf-profile.calltrace.cycles-pp.load_elf_binary.search_binary_handler.__do_execve_file.__x64_sys_execve.do_syscall_64 > 13.15 ± 67% -8.2 5.00 ±173% perf-profile.calltrace.cycles-pp.intel_idle.cpuidle_enter_state.cpuidle_enter.do_idle.cpu_startup_entry > 8.81 ±133% -8.0 0.83 ±173% perf-profile.calltrace.cycles-pp.ret_from_fork > 13.15 ± 67% -7.3 5.83 ±173% perf-profile.calltrace.cycles-pp.secondary_startup_64 > 13.15 ± 67% -7.3 5.83 ±173% perf-profile.calltrace.cycles-pp.start_secondary.secondary_startup_64 > 13.15 ± 67% -7.3 5.83 ±173% perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.secondary_startup_64 > 13.15 ± 67% -7.3 5.83 ±173% perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64 > 13.15 ± 67% -7.3 5.83 ±173% perf-profile.calltrace.cycles-pp.cpuidle_enter.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64 > 13.15 ± 67% -7.3 5.83 ±173% perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.do_idle.cpu_startup_entry.start_secondary > 6.70 ±100% -6.7 0.00 perf-profile.calltrace.cycles-pp.__clear_user.load_elf_binary.search_binary_handler.__do_execve_file.__x64_sys_execve > 4.79 ±108% -4.8 0.00 perf-profile.calltrace.cycles-pp.free_pgtables.exit_mmap.mmput.do_exit.do_group_exit > 4.79 ±108% -4.8 0.00 perf-profile.calltrace.cycles-pp.unlink_file_vma.free_pgtables.exit_mmap.mmput.do_exit > 4.79 ±108% -4.0 0.83 ±173% perf-profile.calltrace.cycles-pp.mmput.do_exit.do_group_exit.get_signal.do_signal > 4.79 ±108% -4.0 0.83 ±173% perf-profile.calltrace.cycles-pp.exit_mmap.mmput.do_exit.do_group_exit.get_signal > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.ioctl.perf_evsel__enable.evsel__enable.evlist__enable > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.ioctl.perf_evsel__enable.evsel__enable > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp.event_function_call.perf_event_for_each_child._perf_ioctl.perf_ioctl.do_vfs_ioctl > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp.smp_call_function_single.event_function_call.perf_event_for_each_child._perf_ioctl.perf_ioctl > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp.__libc_start_main > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp.main.__libc_start_main > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp.run_builtin.main.__libc_start_main > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp.cmd_record.run_builtin.main.__libc_start_main > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp.evlist__enable.cmd_record.run_builtin.main.__libc_start_main > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp.evsel__enable.evlist__enable.cmd_record.run_builtin.main > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp.perf_evsel__enable.evsel__enable.evlist__enable.cmd_record.run_builtin > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp.ioctl.perf_evsel__enable.evsel__enable.evlist__enable.cmd_record > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp.__x64_sys_ioctl.do_syscall_64.entry_SYSCALL_64_after_hwframe.ioctl.perf_evsel__enable > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp.ksys_ioctl.__x64_sys_ioctl.do_syscall_64.entry_SYSCALL_64_after_hwframe.ioctl > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp.do_vfs_ioctl.ksys_ioctl.__x64_sys_ioctl.do_syscall_64.entry_SYSCALL_64_after_hwframe > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp.perf_ioctl.do_vfs_ioctl.ksys_ioctl.__x64_sys_ioctl.do_syscall_64 > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp._perf_ioctl.perf_ioctl.do_vfs_ioctl.ksys_ioctl.__x64_sys_ioctl > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp.perf_event_for_each_child._perf_ioctl.perf_ioctl.do_vfs_ioctl.ksys_ioctl > 4.79 ±108% +8.5 13.33 ±173% perf-profile.calltrace.cycles-pp.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe > 4.79 ±108% +8.5 13.33 ±173% perf-profile.calltrace.cycles-pp.do_signal.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe > 4.79 ±108% +8.5 13.33 ±173% perf-profile.calltrace.cycles-pp.get_signal.do_signal.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe > 4.79 ±108% +8.5 13.33 ±173% perf-profile.calltrace.cycles-pp.do_group_exit.get_signal.do_signal.exit_to_usermode_loop.do_syscall_64 > 4.79 ±108% +8.5 13.33 ±173% perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.get_signal.do_signal.exit_to_usermode_loop > 12.92 ±100% +12.1 25.00 ±173% perf-profile.calltrace.cycles-pp.__x64_sys_exit_group.do_syscall_64.entry_SYSCALL_64_after_hwframe > 12.92 ±100% +12.1 25.00 ±173% perf-profile.calltrace.cycles-pp.do_group_exit.__x64_sys_exit_group.do_syscall_64.entry_SYSCALL_64_after_hwframe > 12.92 ±100% +12.1 25.00 ±173% perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64.entry_SYSCALL_64_after_hwframe > 11.25 ±101% +13.8 25.00 ±173% perf-profile.calltrace.cycles-pp.mmput.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64 > 9.58 ±108% +15.4 25.00 ±173% perf-profile.calltrace.cycles-pp.exit_mmap.mmput.do_exit.do_group_exit.__x64_sys_exit_group > 21.96 ± 69% -22.0 0.00 perf-profile.children.cycles-pp.__x64_sys_execve > 21.96 ± 69% -22.0 0.00 perf-profile.children.cycles-pp.__do_execve_file > 18.63 ± 89% -18.6 0.00 perf-profile.children.cycles-pp.search_binary_handler > 18.63 ± 89% -18.6 0.00 perf-profile.children.cycles-pp.load_elf_binary > 13.15 ± 67% -8.2 5.00 ±173% perf-profile.children.cycles-pp.intel_idle > 8.81 ±133% -8.0 0.83 ±173% perf-profile.children.cycles-pp.ret_from_fork > 13.15 ± 67% -7.3 5.83 ±173% perf-profile.children.cycles-pp.secondary_startup_64 > 13.15 ± 67% -7.3 5.83 ±173% perf-profile.children.cycles-pp.start_secondary > 13.15 ± 67% -7.3 5.83 ±173% perf-profile.children.cycles-pp.cpu_startup_entry > 13.15 ± 67% -7.3 5.83 ±173% perf-profile.children.cycles-pp.do_idle > 13.15 ± 67% -7.3 5.83 ±173% perf-profile.children.cycles-pp.cpuidle_enter > 13.15 ± 67% -7.3 5.83 ±173% perf-profile.children.cycles-pp.cpuidle_enter_state > 6.70 ±100% -6.7 0.00 perf-profile.children.cycles-pp.__clear_user > 6.46 ±100% -6.5 0.00 perf-profile.children.cycles-pp.handle_mm_fault > 4.79 ±108% -4.8 0.00 perf-profile.children.cycles-pp.page_fault > 4.79 ±108% -4.8 0.00 perf-profile.children.cycles-pp.do_page_fault > 4.79 ±108% -4.8 0.00 perf-profile.children.cycles-pp.__do_page_fault > 4.79 ±108% -4.8 0.00 perf-profile.children.cycles-pp.free_pgtables > 4.79 ±108% -4.8 0.00 perf-profile.children.cycles-pp.unlink_file_vma > 5.24 ±112% -0.2 5.00 ±173% perf-profile.children.cycles-pp.__libc_start_main > 5.24 ±112% -0.2 5.00 ±173% perf-profile.children.cycles-pp.main > 5.24 ±112% -0.2 5.00 ±173% perf-profile.children.cycles-pp.run_builtin > 5.24 ±112% -0.2 5.00 ±173% perf-profile.children.cycles-pp.cmd_record > 5.24 ±112% -0.2 5.00 ±173% perf-profile.children.cycles-pp.evlist__enable > 5.24 ±112% -0.2 5.00 ±173% perf-profile.children.cycles-pp.evsel__enable > 5.24 ±112% -0.2 5.00 ±173% perf-profile.children.cycles-pp.perf_evsel__enable > 5.24 ±112% -0.2 5.00 ±173% perf-profile.children.cycles-pp.ioctl > 5.24 ±112% -0.2 5.00 ±173% perf-profile.children.cycles-pp.__x64_sys_ioctl > 5.24 ±112% -0.2 5.00 ±173% perf-profile.children.cycles-pp.ksys_ioctl > 5.24 ±112% -0.2 5.00 ±173% perf-profile.children.cycles-pp.do_vfs_ioctl > 5.24 ±112% -0.2 5.00 ±173% perf-profile.children.cycles-pp.perf_ioctl > 5.24 ±112% -0.2 5.00 ±173% perf-profile.children.cycles-pp._perf_ioctl > 5.24 ±112% -0.2 5.00 ±173% perf-profile.children.cycles-pp.perf_event_for_each_child > 4.79 ±108% +8.5 13.33 ±173% perf-profile.children.cycles-pp.exit_to_usermode_loop > 4.79 ±108% +8.5 13.33 ±173% perf-profile.children.cycles-pp.do_signal > 4.79 ±108% +8.5 13.33 ±173% perf-profile.children.cycles-pp.get_signal > 5.24 ±112% +8.9 14.17 ±173% perf-profile.children.cycles-pp.event_function_call > 5.24 ±112% +8.9 14.17 ±173% perf-profile.children.cycles-pp.smp_call_function_single > 12.92 ±100% +12.1 25.00 ±173% perf-profile.children.cycles-pp.__x64_sys_exit_group > 13.15 ± 67% -8.2 5.00 ±173% perf-profile.self.cycles-pp.intel_idle > 6.46 ±100% -6.5 0.00 perf-profile.self.cycles-pp.unmap_page_range > 5.24 ±112% +8.9 14.17 ±173% perf-profile.self.cycles-pp.smp_call_function_single >
WARNING: multiple messages have this Message-ID (diff)
From: Feng Tang <feng.tang@intel.com> To: lkp@lists.01.org Subject: Re: [net] 19f92a030c: apachebench.requests_per_second -37.9% regression Date: Wed, 13 Nov 2019 18:35:01 +0800 [thread overview] Message-ID: <20191113103501.GD65640@shbuild999.sh.intel.com> (raw) In-Reply-To: <20191108083513.GB29418@shao2-debian> [-- Attachment #1: Type: text/plain, Size: 27350 bytes --] Hi Eric, On Fri, Nov 08, 2019 at 04:35:13PM +0800, kernel test robot wrote: > Greeting, > > FYI, we noticed a -37.9% regression of apachebench.requests_per_second due to commit: > > commit: 19f92a030ca6d772ab44b22ee6a01378a8cb32d4 ("net: increase SOMAXCONN to 4096") > https://git.kernel.org/cgit/linux/kernel/git/torvalds/linux.git master Any thought on this? The test is actually: sysctl -w net.ipv4.tcp_syncookies=0 enable_apache_mod auth_basic authn_core authn_file authz_core authz_host authz_user access_compat systemctl restart apache2 ab -k -q -t 300 -n 1000000 -c 4000 127.0.0.1/ And some info about apachebench result is: w/o patch: Connection Times (ms) min mean[+/-sd] median max Connect: 0 0 19.5 0 7145 Processing: 0 4 110.0 3 21647 Waiting: 0 2 92.4 1 21646 Total: 0 4 121.1 3 24762 w/ patch: Connection Times (ms) min mean[+/-sd] median max Connect: 0 0 43.2 0 7143 Processing: 0 19 640.4 3 38708 Waiting: 0 24 796.5 1 38708 Total: 0 19 657.5 3 39725 Thanks, Feng > > in testcase: apachebench > on test machine: 16 threads Intel(R) Xeon(R) CPU D-1541 @ 2.10GHz with 48G memory > with following parameters: > > runtime: 300s > concurrency: 4000 > cluster: cs-localhost > cpufreq_governor: performance > ucode: 0x7000019 > > test-description: apachebench is a tool for benchmarking your Apache Hypertext Transfer Protocol (HTTP) server. > test-url: https://httpd.apache.org/docs/2.4/programs/ab.html > > In addition to that, the commit also has significant impact on the following tests: > > +------------------+------------------------------------------------------------------+ > | testcase: change | apachebench: apachebench.requests_per_second -37.5% regression | > | test machine | 16 threads Intel(R) Xeon(R) CPU D-1541 @ 2.10GHz with 48G memory | > | test parameters | cluster=cs-localhost | > | | concurrency=8000 | > | | cpufreq_governor=performance | > | | runtime=300s | > | | ucode=0x7000019 | > +------------------+------------------------------------------------------------------+ > > > If you fix the issue, kindly add following tag > Reported-by: kernel test robot <rong.a.chen@intel.com> > > > Details are as below: > --------------------------------------------------------------------------------------------------> > > > To reproduce: > > git clone https://github.com/intel/lkp-tests.git > cd lkp-tests > bin/lkp install job.yaml # job file is attached in this email > bin/lkp run job.yaml > > ========================================================================================= > cluster/compiler/concurrency/cpufreq_governor/kconfig/rootfs/runtime/tbox_group/testcase/ucode: > cs-localhost/gcc-7/4000/performance/x86_64-rhel-7.6/debian-x86_64-2019-09-23.cgz/300s/lkp-bdw-de1/apachebench/0x7000019 > > commit: > 6d6f0383b6 ("netdevsim: Fix use-after-free during device dismantle") > 19f92a030c ("net: increase SOMAXCONN to 4096") > > 6d6f0383b697f004 19f92a030ca6d772ab44b22ee6a > ---------------- --------------------------- > %stddev %change %stddev > \ | \ > 22640 ± 4% +71.1% 38734 apachebench.connection_time.processing.max > 24701 +60.9% 39743 apachebench.connection_time.total.max > 22639 ± 4% +71.1% 38734 apachebench.connection_time.waiting.max > 24701 +15042.0 39743 apachebench.max_latency.100% > 40454 -37.9% 25128 apachebench.requests_per_second > 25.69 +58.8% 40.79 apachebench.time.elapsed_time > 25.69 +58.8% 40.79 apachebench.time.elapsed_time.max > 79.00 -37.0% 49.75 apachebench.time.percent_of_cpu_this_job_got > 98.88 +61.0% 159.18 apachebench.time_per_request > 434631 -37.9% 269889 apachebench.transfer_rate > 1.5e+08 ± 18% +109.5% 3.141e+08 ± 27% cpuidle.C3.time > 578957 ± 7% +64.1% 949934 ± 12% cpuidle.C3.usage > 79085 ± 4% +24.8% 98720 meminfo.AnonHugePages > 41176 +14.2% 47013 meminfo.PageTables > 69429 -34.9% 45222 meminfo.max_used_kB > 63.48 +12.7 76.15 mpstat.cpu.all.idle% > 2.42 ± 2% -0.9 1.56 mpstat.cpu.all.soft% > 15.30 -5.2 10.13 mpstat.cpu.all.sys% > 18.80 -6.6 12.16 mpstat.cpu.all.usr% > 65.00 +17.7% 76.50 vmstat.cpu.id > 17.00 -35.3% 11.00 vmstat.cpu.us > 7.00 ± 24% -50.0% 3.50 ± 14% vmstat.procs.r > 62957 -33.3% 42012 vmstat.system.cs > 33174 -1.4% 32693 vmstat.system.in > 5394 ± 5% +16.3% 6272 ± 6% sched_debug.cfs_rq:/.min_vruntime.stddev > 5396 ± 5% +16.3% 6275 ± 6% sched_debug.cfs_rq:/.spread0.stddev > 33982 ± 48% -83.3% 5676 ± 47% sched_debug.cpu.avg_idle.min > 26.75 ± 77% +169.8% 72.17 ± 41% sched_debug.cpu.sched_count.avg > 212.00 ± 90% +168.2% 568.50 ± 50% sched_debug.cpu.sched_count.max > 52.30 ± 89% +182.5% 147.73 ± 48% sched_debug.cpu.sched_count.stddev > 11.33 ± 80% +193.9% 33.30 ± 42% sched_debug.cpu.sched_goidle.avg > 104.50 ± 92% +170.6% 282.75 ± 50% sched_debug.cpu.sched_goidle.max > 26.18 ± 90% +183.9% 74.31 ± 48% sched_debug.cpu.sched_goidle.stddev > 959.00 -32.0% 652.00 turbostat.Avg_MHz > 39.01 -11.2 27.79 turbostat.Busy% > 1.46 ± 7% -0.5 0.96 ± 5% turbostat.C1% > 9.58 ± 4% -3.2 6.38 turbostat.C1E% > 578646 ± 7% +64.1% 949626 ± 12% turbostat.C3 > 940073 +51.1% 1420298 turbostat.IRQ > 2.20 ± 22% +159.7% 5.71 ± 12% turbostat.Pkg%pc2 > 31.22 -17.2% 25.86 turbostat.PkgWatt > 4.74 -7.5% 4.39 turbostat.RAMWatt > 93184 -1.6% 91678 proc-vmstat.nr_active_anon > 92970 -1.8% 91314 proc-vmstat.nr_anon_pages > 288405 +1.0% 291286 proc-vmstat.nr_file_pages > 8307 +6.3% 8831 proc-vmstat.nr_kernel_stack > 10315 +14.2% 11783 proc-vmstat.nr_page_table_pages > 21499 +6.0% 22798 proc-vmstat.nr_slab_unreclaimable > 284131 +1.0% 286977 proc-vmstat.nr_unevictable > 93184 -1.6% 91678 proc-vmstat.nr_zone_active_anon > 284131 +1.0% 286977 proc-vmstat.nr_zone_unevictable > 198874 ± 2% +43.7% 285772 ± 16% proc-vmstat.numa_hit > 198874 ± 2% +43.7% 285772 ± 16% proc-vmstat.numa_local > 249594 ± 3% +59.6% 398267 ± 13% proc-vmstat.pgalloc_normal > 1216885 +12.7% 1371283 ± 3% proc-vmstat.pgfault > 179705 ± 16% +82.9% 328634 ± 21% proc-vmstat.pgfree > 346.25 ± 5% +133.5% 808.50 ± 2% slabinfo.TCPv6.active_objs > 346.25 ± 5% +134.4% 811.75 ± 2% slabinfo.TCPv6.num_objs > 22966 +15.6% 26559 slabinfo.anon_vma.active_objs > 23091 +15.5% 26664 slabinfo.anon_vma.num_objs > 69747 +16.1% 81011 slabinfo.anon_vma_chain.active_objs > 1094 +15.9% 1269 slabinfo.anon_vma_chain.active_slabs > 70092 +15.9% 81259 slabinfo.anon_vma_chain.num_objs > 1094 +15.9% 1269 slabinfo.anon_vma_chain.num_slabs > 1649 +12.9% 1861 slabinfo.cred_jar.active_objs > 1649 +12.9% 1861 slabinfo.cred_jar.num_objs > 4924 +20.0% 5907 slabinfo.pid.active_objs > 4931 +19.9% 5912 slabinfo.pid.num_objs > 266.50 ± 3% +299.2% 1063 slabinfo.request_sock_TCP.active_objs > 266.50 ± 3% +299.2% 1063 slabinfo.request_sock_TCP.num_objs > 11.50 ± 4% +1700.0% 207.00 ± 4% slabinfo.tw_sock_TCPv6.active_objs > 11.50 ± 4% +1700.0% 207.00 ± 4% slabinfo.tw_sock_TCPv6.num_objs > 41682 +16.0% 48360 slabinfo.vm_area_struct.active_objs > 1046 +15.7% 1211 slabinfo.vm_area_struct.active_slabs > 41879 +15.7% 48468 slabinfo.vm_area_struct.num_objs > 1046 +15.7% 1211 slabinfo.vm_area_struct.num_slabs > 4276 ± 2% +10.0% 4705 ± 2% slabinfo.vmap_area.num_objs > 21.25 ± 27% +3438.8% 752.00 ± 83% interrupts.36:IR-PCI-MSI.2621443-edge.eth0-TxRx-2 > 21.25 ± 20% +1777.6% 399.00 ±155% interrupts.38:IR-PCI-MSI.2621445-edge.eth0-TxRx-4 > 54333 +54.3% 83826 interrupts.CPU0.LOC:Local_timer_interrupts > 54370 +54.6% 84072 interrupts.CPU1.LOC:Local_timer_interrupts > 21.25 ± 27% +3438.8% 752.00 ± 83% interrupts.CPU10.36:IR-PCI-MSI.2621443-edge.eth0-TxRx-2 > 54236 +54.7% 83925 interrupts.CPU10.LOC:Local_timer_interrupts > 54223 +54.3% 83655 interrupts.CPU11.LOC:Local_timer_interrupts > 377.75 ± 21% +27.1% 480.25 ± 10% interrupts.CPU11.RES:Rescheduling_interrupts > 21.25 ± 20% +1777.6% 399.00 ±155% interrupts.CPU12.38:IR-PCI-MSI.2621445-edge.eth0-TxRx-4 > 54279 +54.1% 83646 interrupts.CPU12.LOC:Local_timer_interrupts > 53683 +55.3% 83365 interrupts.CPU13.LOC:Local_timer_interrupts > 53887 +55.7% 83903 interrupts.CPU14.LOC:Local_timer_interrupts > 54156 +54.7% 83803 interrupts.CPU15.LOC:Local_timer_interrupts > 54041 +55.1% 83806 interrupts.CPU2.LOC:Local_timer_interrupts > 54042 +55.4% 83991 interrupts.CPU3.LOC:Local_timer_interrupts > 54081 +55.2% 83938 interrupts.CPU4.LOC:Local_timer_interrupts > 54322 +54.9% 84166 interrupts.CPU5.LOC:Local_timer_interrupts > 53586 +56.5% 83849 interrupts.CPU6.LOC:Local_timer_interrupts > 54049 +55.2% 83892 interrupts.CPU7.LOC:Local_timer_interrupts > 54056 +54.9% 83751 interrupts.CPU8.LOC:Local_timer_interrupts > 53862 +54.7% 83331 interrupts.CPU9.LOC:Local_timer_interrupts > 865212 +55.0% 1340925 interrupts.LOC:Local_timer_interrupts > 16477 ± 4% +32.2% 21779 ± 8% softirqs.CPU0.TIMER > 18508 ± 15% +39.9% 25891 ± 17% softirqs.CPU1.TIMER > 16625 ± 8% +21.3% 20166 ± 7% softirqs.CPU10.TIMER > 5906 ± 21% +62.5% 9597 ± 13% softirqs.CPU12.SCHED > 17474 ± 12% +29.4% 22610 ± 7% softirqs.CPU12.TIMER > 7680 ± 11% +20.4% 9246 ± 14% softirqs.CPU13.SCHED > 45558 ± 36% -37.8% 28320 ± 25% softirqs.CPU14.NET_RX > 15365 ± 4% +40.7% 21622 ± 5% softirqs.CPU14.TIMER > 8084 ± 4% +18.7% 9599 ± 12% softirqs.CPU15.RCU > 16433 ± 4% +41.2% 23203 ± 14% softirqs.CPU2.TIMER > 8436 ± 7% +19.9% 10117 ± 10% softirqs.CPU3.RCU > 15992 ± 3% +48.5% 23742 ± 18% softirqs.CPU3.TIMER > 17389 ± 14% +38.7% 24116 ± 11% softirqs.CPU4.TIMER > 17749 ± 13% +42.2% 25235 ± 15% softirqs.CPU5.TIMER > 16528 ± 9% +28.3% 21200 ± 2% softirqs.CPU6.TIMER > 8321 ± 8% +31.3% 10929 ± 5% softirqs.CPU7.RCU > 18024 ± 8% +28.8% 23212 ± 5% softirqs.CPU7.TIMER > 15717 ± 5% +27.1% 19983 ± 7% softirqs.CPU8.TIMER > 7383 ± 11% +30.5% 9632 ± 9% softirqs.CPU9.SCHED > 16342 ± 18% +41.0% 23037 ± 10% softirqs.CPU9.TIMER > 148013 +10.2% 163086 ± 2% softirqs.RCU > 112139 +28.0% 143569 softirqs.SCHED > 276690 ± 3% +30.4% 360747 ± 3% softirqs.TIMER > 1.453e+09 -36.2% 9.273e+08 perf-stat.i.branch-instructions > 67671486 -35.9% 43396843 ± 2% perf-stat.i.branch-misses > 5.188e+08 -35.3% 3.357e+08 perf-stat.i.cache-misses > 5.188e+08 -35.3% 3.357e+08 perf-stat.i.cache-references > 71149 -36.0% 45536 perf-stat.i.context-switches > 2.92 ± 6% +38.4% 4.04 ± 5% perf-stat.i.cpi > 1.581e+10 -33.8% 1.046e+10 perf-stat.i.cpu-cycles > 1957 ± 6% -35.4% 1264 ± 5% perf-stat.i.cpu-migrations > 40.76 ± 2% +33.0% 54.21 perf-stat.i.cycles-between-cache-misses > 0.64 ± 2% +0.2 0.86 ± 2% perf-stat.i.dTLB-load-miss-rate% > 10720126 ± 2% -34.0% 7071299 perf-stat.i.dTLB-load-misses > 2.05e+09 -35.8% 1.315e+09 perf-stat.i.dTLB-loads > 0.16 +0.0 0.18 ± 5% perf-stat.i.dTLB-store-miss-rate% > 1688635 -31.7% 1153595 perf-stat.i.dTLB-store-misses > 1.17e+09 -33.5% 7.777e+08 perf-stat.i.dTLB-stores > 61.00 ± 6% +8.7 69.70 ± 2% perf-stat.i.iTLB-load-miss-rate% > 13112146 ± 10% -38.5% 8061589 ± 13% perf-stat.i.iTLB-load-misses > 9827689 ± 3% -37.1% 6184316 perf-stat.i.iTLB-loads > 7e+09 -36.2% 4.469e+09 perf-stat.i.instructions > 0.42 ± 2% -22.2% 0.33 ± 3% perf-stat.i.ipc > 45909 -32.4% 31036 perf-stat.i.minor-faults > 45909 -32.4% 31036 perf-stat.i.page-faults > 2.26 +3.6% 2.34 perf-stat.overall.cpi > 30.47 +2.3% 31.16 perf-stat.overall.cycles-between-cache-misses > 0.44 -3.5% 0.43 perf-stat.overall.ipc > 1.398e+09 -35.3% 9.049e+08 perf-stat.ps.branch-instructions > 65124290 -35.0% 42353120 ± 2% perf-stat.ps.branch-misses > 4.993e+08 -34.4% 3.276e+08 perf-stat.ps.cache-misses > 4.993e+08 -34.4% 3.276e+08 perf-stat.ps.cache-references > 68469 -35.1% 44440 perf-stat.ps.context-switches > 1.521e+10 -32.9% 1.021e+10 perf-stat.ps.cpu-cycles > 1884 ± 6% -34.5% 1234 ± 5% perf-stat.ps.cpu-migrations > 10314734 ± 2% -33.1% 6899548 perf-stat.ps.dTLB-load-misses > 1.973e+09 -34.9% 1.283e+09 perf-stat.ps.dTLB-loads > 1624883 -30.7% 1125668 perf-stat.ps.dTLB-store-misses > 1.126e+09 -32.6% 7.589e+08 perf-stat.ps.dTLB-stores > 12617676 ± 10% -37.7% 7866427 ± 13% perf-stat.ps.iTLB-load-misses > 9458687 ± 3% -36.2% 6036594 perf-stat.ps.iTLB-loads > 6.736e+09 -35.3% 4.361e+09 perf-stat.ps.instructions > 44179 -31.4% 30288 perf-stat.ps.minor-faults > 30791 +1.4% 31223 perf-stat.ps.msec > 44179 -31.4% 30288 perf-stat.ps.page-faults > 21.96 ± 69% -22.0 0.00 perf-profile.calltrace.cycles-pp.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe > 21.96 ± 69% -22.0 0.00 perf-profile.calltrace.cycles-pp.__do_execve_file.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe > 18.63 ± 89% -18.6 0.00 perf-profile.calltrace.cycles-pp.search_binary_handler.__do_execve_file.__x64_sys_execve.do_syscall_64.entry_SYSCALL_64_after_hwframe > 18.63 ± 89% -18.6 0.00 perf-profile.calltrace.cycles-pp.load_elf_binary.search_binary_handler.__do_execve_file.__x64_sys_execve.do_syscall_64 > 13.15 ± 67% -8.2 5.00 ±173% perf-profile.calltrace.cycles-pp.intel_idle.cpuidle_enter_state.cpuidle_enter.do_idle.cpu_startup_entry > 8.81 ±133% -8.0 0.83 ±173% perf-profile.calltrace.cycles-pp.ret_from_fork > 13.15 ± 67% -7.3 5.83 ±173% perf-profile.calltrace.cycles-pp.secondary_startup_64 > 13.15 ± 67% -7.3 5.83 ±173% perf-profile.calltrace.cycles-pp.start_secondary.secondary_startup_64 > 13.15 ± 67% -7.3 5.83 ±173% perf-profile.calltrace.cycles-pp.cpu_startup_entry.start_secondary.secondary_startup_64 > 13.15 ± 67% -7.3 5.83 ±173% perf-profile.calltrace.cycles-pp.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64 > 13.15 ± 67% -7.3 5.83 ±173% perf-profile.calltrace.cycles-pp.cpuidle_enter.do_idle.cpu_startup_entry.start_secondary.secondary_startup_64 > 13.15 ± 67% -7.3 5.83 ±173% perf-profile.calltrace.cycles-pp.cpuidle_enter_state.cpuidle_enter.do_idle.cpu_startup_entry.start_secondary > 6.70 ±100% -6.7 0.00 perf-profile.calltrace.cycles-pp.__clear_user.load_elf_binary.search_binary_handler.__do_execve_file.__x64_sys_execve > 4.79 ±108% -4.8 0.00 perf-profile.calltrace.cycles-pp.free_pgtables.exit_mmap.mmput.do_exit.do_group_exit > 4.79 ±108% -4.8 0.00 perf-profile.calltrace.cycles-pp.unlink_file_vma.free_pgtables.exit_mmap.mmput.do_exit > 4.79 ±108% -4.0 0.83 ±173% perf-profile.calltrace.cycles-pp.mmput.do_exit.do_group_exit.get_signal.do_signal > 4.79 ±108% -4.0 0.83 ±173% perf-profile.calltrace.cycles-pp.exit_mmap.mmput.do_exit.do_group_exit.get_signal > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp.entry_SYSCALL_64_after_hwframe.ioctl.perf_evsel__enable.evsel__enable.evlist__enable > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp.do_syscall_64.entry_SYSCALL_64_after_hwframe.ioctl.perf_evsel__enable.evsel__enable > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp.event_function_call.perf_event_for_each_child._perf_ioctl.perf_ioctl.do_vfs_ioctl > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp.smp_call_function_single.event_function_call.perf_event_for_each_child._perf_ioctl.perf_ioctl > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp.__libc_start_main > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp.main.__libc_start_main > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp.run_builtin.main.__libc_start_main > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp.cmd_record.run_builtin.main.__libc_start_main > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp.evlist__enable.cmd_record.run_builtin.main.__libc_start_main > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp.evsel__enable.evlist__enable.cmd_record.run_builtin.main > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp.perf_evsel__enable.evsel__enable.evlist__enable.cmd_record.run_builtin > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp.ioctl.perf_evsel__enable.evsel__enable.evlist__enable.cmd_record > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp.__x64_sys_ioctl.do_syscall_64.entry_SYSCALL_64_after_hwframe.ioctl.perf_evsel__enable > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp.ksys_ioctl.__x64_sys_ioctl.do_syscall_64.entry_SYSCALL_64_after_hwframe.ioctl > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp.do_vfs_ioctl.ksys_ioctl.__x64_sys_ioctl.do_syscall_64.entry_SYSCALL_64_after_hwframe > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp.perf_ioctl.do_vfs_ioctl.ksys_ioctl.__x64_sys_ioctl.do_syscall_64 > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp._perf_ioctl.perf_ioctl.do_vfs_ioctl.ksys_ioctl.__x64_sys_ioctl > 5.24 ±112% -0.2 5.00 ±173% perf-profile.calltrace.cycles-pp.perf_event_for_each_child._perf_ioctl.perf_ioctl.do_vfs_ioctl.ksys_ioctl > 4.79 ±108% +8.5 13.33 ±173% perf-profile.calltrace.cycles-pp.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe > 4.79 ±108% +8.5 13.33 ±173% perf-profile.calltrace.cycles-pp.do_signal.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe > 4.79 ±108% +8.5 13.33 ±173% perf-profile.calltrace.cycles-pp.get_signal.do_signal.exit_to_usermode_loop.do_syscall_64.entry_SYSCALL_64_after_hwframe > 4.79 ±108% +8.5 13.33 ±173% perf-profile.calltrace.cycles-pp.do_group_exit.get_signal.do_signal.exit_to_usermode_loop.do_syscall_64 > 4.79 ±108% +8.5 13.33 ±173% perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.get_signal.do_signal.exit_to_usermode_loop > 12.92 ±100% +12.1 25.00 ±173% perf-profile.calltrace.cycles-pp.__x64_sys_exit_group.do_syscall_64.entry_SYSCALL_64_after_hwframe > 12.92 ±100% +12.1 25.00 ±173% perf-profile.calltrace.cycles-pp.do_group_exit.__x64_sys_exit_group.do_syscall_64.entry_SYSCALL_64_after_hwframe > 12.92 ±100% +12.1 25.00 ±173% perf-profile.calltrace.cycles-pp.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64.entry_SYSCALL_64_after_hwframe > 11.25 ±101% +13.8 25.00 ±173% perf-profile.calltrace.cycles-pp.mmput.do_exit.do_group_exit.__x64_sys_exit_group.do_syscall_64 > 9.58 ±108% +15.4 25.00 ±173% perf-profile.calltrace.cycles-pp.exit_mmap.mmput.do_exit.do_group_exit.__x64_sys_exit_group > 21.96 ± 69% -22.0 0.00 perf-profile.children.cycles-pp.__x64_sys_execve > 21.96 ± 69% -22.0 0.00 perf-profile.children.cycles-pp.__do_execve_file > 18.63 ± 89% -18.6 0.00 perf-profile.children.cycles-pp.search_binary_handler > 18.63 ± 89% -18.6 0.00 perf-profile.children.cycles-pp.load_elf_binary > 13.15 ± 67% -8.2 5.00 ±173% perf-profile.children.cycles-pp.intel_idle > 8.81 ±133% -8.0 0.83 ±173% perf-profile.children.cycles-pp.ret_from_fork > 13.15 ± 67% -7.3 5.83 ±173% perf-profile.children.cycles-pp.secondary_startup_64 > 13.15 ± 67% -7.3 5.83 ±173% perf-profile.children.cycles-pp.start_secondary > 13.15 ± 67% -7.3 5.83 ±173% perf-profile.children.cycles-pp.cpu_startup_entry > 13.15 ± 67% -7.3 5.83 ±173% perf-profile.children.cycles-pp.do_idle > 13.15 ± 67% -7.3 5.83 ±173% perf-profile.children.cycles-pp.cpuidle_enter > 13.15 ± 67% -7.3 5.83 ±173% perf-profile.children.cycles-pp.cpuidle_enter_state > 6.70 ±100% -6.7 0.00 perf-profile.children.cycles-pp.__clear_user > 6.46 ±100% -6.5 0.00 perf-profile.children.cycles-pp.handle_mm_fault > 4.79 ±108% -4.8 0.00 perf-profile.children.cycles-pp.page_fault > 4.79 ±108% -4.8 0.00 perf-profile.children.cycles-pp.do_page_fault > 4.79 ±108% -4.8 0.00 perf-profile.children.cycles-pp.__do_page_fault > 4.79 ±108% -4.8 0.00 perf-profile.children.cycles-pp.free_pgtables > 4.79 ±108% -4.8 0.00 perf-profile.children.cycles-pp.unlink_file_vma > 5.24 ±112% -0.2 5.00 ±173% perf-profile.children.cycles-pp.__libc_start_main > 5.24 ±112% -0.2 5.00 ±173% perf-profile.children.cycles-pp.main > 5.24 ±112% -0.2 5.00 ±173% perf-profile.children.cycles-pp.run_builtin > 5.24 ±112% -0.2 5.00 ±173% perf-profile.children.cycles-pp.cmd_record > 5.24 ±112% -0.2 5.00 ±173% perf-profile.children.cycles-pp.evlist__enable > 5.24 ±112% -0.2 5.00 ±173% perf-profile.children.cycles-pp.evsel__enable > 5.24 ±112% -0.2 5.00 ±173% perf-profile.children.cycles-pp.perf_evsel__enable > 5.24 ±112% -0.2 5.00 ±173% perf-profile.children.cycles-pp.ioctl > 5.24 ±112% -0.2 5.00 ±173% perf-profile.children.cycles-pp.__x64_sys_ioctl > 5.24 ±112% -0.2 5.00 ±173% perf-profile.children.cycles-pp.ksys_ioctl > 5.24 ±112% -0.2 5.00 ±173% perf-profile.children.cycles-pp.do_vfs_ioctl > 5.24 ±112% -0.2 5.00 ±173% perf-profile.children.cycles-pp.perf_ioctl > 5.24 ±112% -0.2 5.00 ±173% perf-profile.children.cycles-pp._perf_ioctl > 5.24 ±112% -0.2 5.00 ±173% perf-profile.children.cycles-pp.perf_event_for_each_child > 4.79 ±108% +8.5 13.33 ±173% perf-profile.children.cycles-pp.exit_to_usermode_loop > 4.79 ±108% +8.5 13.33 ±173% perf-profile.children.cycles-pp.do_signal > 4.79 ±108% +8.5 13.33 ±173% perf-profile.children.cycles-pp.get_signal > 5.24 ±112% +8.9 14.17 ±173% perf-profile.children.cycles-pp.event_function_call > 5.24 ±112% +8.9 14.17 ±173% perf-profile.children.cycles-pp.smp_call_function_single > 12.92 ±100% +12.1 25.00 ±173% perf-profile.children.cycles-pp.__x64_sys_exit_group > 13.15 ± 67% -8.2 5.00 ±173% perf-profile.self.cycles-pp.intel_idle > 6.46 ±100% -6.5 0.00 perf-profile.self.cycles-pp.unmap_page_range > 5.24 ±112% +8.9 14.17 ±173% perf-profile.self.cycles-pp.smp_call_function_single >
next prev parent reply other threads:[~2019-11-13 10:35 UTC|newest] Thread overview: 10+ messages / expand[flat|nested] mbox.gz Atom feed top 2019-11-08 8:35 [net] 19f92a030c: apachebench.requests_per_second -37.9% regression kernel test robot 2019-11-08 8:35 ` kernel test robot 2019-11-13 10:35 ` Feng Tang [this message] 2019-11-13 10:35 ` Feng Tang 2019-11-13 14:33 ` [LKP] " Eric Dumazet 2019-11-13 14:33 ` Eric Dumazet 2019-11-13 16:12 ` [LKP] " Feng Tang 2019-11-13 16:12 ` Feng Tang 2019-11-13 16:46 ` Eric Dumazet 2019-11-13 17:21 ` [LKP] " Willy Tarreau
Reply instructions: You may reply publicly to this message via plain-text email using any one of the following methods: * Save the following mbox file, import it into your mail client, and reply-to-all from there: mbox Avoid top-posting and favor interleaved quoting: https://en.wikipedia.org/wiki/Posting_style#Interleaved_style * Reply using the --to, --cc, and --in-reply-to switches of git-send-email(1): git send-email \ --in-reply-to=20191113103501.GD65640@shbuild999.sh.intel.com \ --to=feng.tang@intel.com \ --cc=davem@davemloft.net \ --cc=edumazet@google.com \ --cc=linux-kernel@vger.kernel.org \ --cc=lkp@lists.01.org \ --cc=rong.a.chen@intel.com \ --cc=torvalds@linux-foundation.org \ --cc=w@1wt.eu \ --cc=ycao009@ucr.edu \ /path/to/YOUR_REPLY https://kernel.org/pub/software/scm/git/docs/git-send-email.html * If your mail client supports setting the In-Reply-To header via mailto: links, try the mailto: linkBe sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes, see mirroring instructions on how to clone and mirror all data and code used by this external index.