From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-1.0 required=3.0 tests=HEADER_FROM_DIFFERENT_DOMAINS, MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 902DFC282DD for ; Thu, 23 May 2019 23:43:31 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [209.132.180.67]) by mail.kernel.org (Postfix) with ESMTP id 4F8C42175B for ; Thu, 23 May 2019 23:43:31 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S2388658AbfEWXna (ORCPT ); Thu, 23 May 2019 19:43:30 -0400 Received: from outgoing-stata.csail.mit.edu ([128.30.2.210]:50927 "EHLO outgoing-stata.csail.mit.edu" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S2388297AbfEWXn3 (ORCPT ); Thu, 23 May 2019 19:43:29 -0400 Received: from [4.30.142.84] (helo=srivatsab-a01.vmware.com) by outgoing-stata.csail.mit.edu with esmtpsa (TLS1.2:RSA_AES_128_CBC_SHA1:128) (Exim 4.82) (envelope-from ) id 1hTxMo-000QQB-NK; Thu, 23 May 2019 19:43:22 -0400 Subject: Re: CFQ idling kills I/O performance on ext4 with blkio cgroup controller To: Paolo Valente Cc: linux-fsdevel@vger.kernel.org, linux-block , linux-ext4@vger.kernel.org, cgroups@vger.kernel.org, kernel list , Jens Axboe , Jan Kara , jmoyer@redhat.com, Theodore Ts'o , amakhalov@vmware.com, anishs@vmware.com, srivatsab@vmware.com References: <8d72fcf7-bbb4-2965-1a06-e9fc177a8938@csail.mit.edu> <1812E450-14EF-4D5A-8F31-668499E13652@linaro.org> <46c6a4be-f567-3621-2e16-0e341762b828@csail.mit.edu> <07D11833-8285-49C2-943D-E4C1D23E8859@linaro.org> <5B6570A2-541A-4CF8-98E0-979EA6E3717D@linaro.org> <2CB39B34-21EE-4A95-A073-8633CF2D187C@linaro.org> <0e3fdf31-70d9-26eb-7b42-2795d4b03722@csail.mit.edu> <686D6469-9DE7-4738-B92A-002144C3E63E@linaro.org> <01d55216-5718-767a-e1e6-aadc67b632f4@csail.mit.edu> <6FE0A98F-1E3D-4EF6-8B38-2C85741924A4@linaro.org> <2A58C239-EF3F-422B-8D87-E7A3B500C57C@linaro.org> From: "Srivatsa S. Bhat" Message-ID: Date: Thu, 23 May 2019 16:43:20 -0700 User-Agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.13; rv:60.0) Gecko/20100101 Thunderbird/60.6.1 MIME-Version: 1.0 In-Reply-To: <2A58C239-EF3F-422B-8D87-E7A3B500C57C@linaro.org> Content-Type: text/plain; charset=windows-1252 Content-Language: en-US Content-Transfer-Encoding: 7bit Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 5/23/19 10:22 AM, Paolo Valente wrote: > >> Il giorno 23 mag 2019, alle ore 11:19, Paolo Valente ha scritto: >> >>> Il giorno 23 mag 2019, alle ore 04:30, Srivatsa S. Bhat ha scritto: >>> [...] >>> Also, I'm very happy to run additional tests or experiments to help >>> track down this issue. So, please don't hesitate to let me know if >>> you'd like me to try anything else or get you additional traces etc. :) >>> >> >> Here's to you! :) I've attached a new small improvement that may >> reduce fluctuations (path to apply on top of the others, of course). >> Unfortunately, I don't expect this change to boost the throughput >> though. >> >> In contrast, I've thought of a solution that might be rather >> effective: making BFQ aware (heuristically) of trivial >> synchronizations between processes in different groups. This will >> require a little more work and time. >> > > Hi Srivatsa, > I'm back :) > > First, there was a mistake in the last patch I sent you, namely in > 0001-block-bfq-re-sample-req-service-times-when-possible.patch. > Please don't apply that patch at all. > > I've attached a new series of patches instead. The first patch in this > series is a fixed version of the faulty patch above (if I'm creating too > much confusion, I'll send you again all patches to apply on top of > mainline). > No problem, I got it :) > This series also implements the more effective idea I told you a few > hours ago. In my system, the loss is now around only 10%, even with > low_latency on. > When trying to run multiple dd tasks simultaneously, I get the kernel panic shown below (mainline is fine, without these patches). [ 568.232231] BUG: kernel NULL pointer dereference, address: 0000000000000024 [ 568.232257] #PF: supervisor read access in kernel mode [ 568.232273] #PF: error_code(0x0000) - not-present page [ 568.232289] PGD 0 P4D 0 [ 568.232299] Oops: 0000 [#1] SMP PTI [ 568.232312] CPU: 0 PID: 1029 Comm: dd Tainted: G E 5.1.0-io-dbg-4+ #6 [ 568.232334] Hardware name: VMware, Inc. VMware Virtual Platform/440BX Desktop Reference Platform, BIOS 6.00 04/05/2016 [ 568.232388] RIP: 0010:bfq_serv_to_charge+0x21/0x50 [ 568.232404] Code: ff e8 c3 5e bc ff 0f 1f 00 0f 1f 44 00 00 48 8b 86 20 01 00 00 55 48 89 e5 53 48 89 fb a8 40 75 09 83 be a0 01 00 00 01 76 09 <8b> 43 24 c1 e8 09 5b 5d c3 48 8b 7e 08 e8 5d fd ff ff 84 c0 75 ea [ 568.232473] RSP: 0018:ffffa73a42dab750 EFLAGS: 00010002 [ 568.232489] RAX: 0000000000001052 RBX: 0000000000000000 RCX: ffffa73a42dab7a0 [ 568.232510] RDX: ffffa73a42dab657 RSI: ffff8b7b6ba2ab70 RDI: 0000000000000000 [ 568.232530] RBP: ffffa73a42dab758 R08: 0000000000000000 R09: 0000000000000001 [ 568.232551] R10: 0000000000000000 R11: ffffa73a42dab7a0 R12: ffff8b7b6aed3800 [ 568.232571] R13: 0000000000000000 R14: 0000000000000000 R15: ffff8b7b6aed3800 [ 568.232592] FS: 00007fb5b0724540(0000) GS:ffff8b7b6f800000(0000) knlGS:0000000000000000 [ 568.232615] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033 [ 568.232632] CR2: 0000000000000024 CR3: 00000004266be002 CR4: 00000000001606f0 [ 568.232690] Call Trace: [ 568.232703] bfq_select_queue+0x781/0x1000 [ 568.232717] bfq_dispatch_request+0x1d7/0xd60 [ 568.232731] ? bfq_bfqq_handle_idle_busy_switch.isra.36+0x2cd/0xb20 [ 568.232751] blk_mq_do_dispatch_sched+0xa8/0xe0 [ 568.232765] blk_mq_sched_dispatch_requests+0xe3/0x150 [ 568.232783] __blk_mq_run_hw_queue+0x56/0x100 [ 568.232798] __blk_mq_delay_run_hw_queue+0x107/0x160 [ 568.232814] blk_mq_run_hw_queue+0x75/0x190 [ 568.232828] blk_mq_sched_insert_requests+0x7a/0x100 [ 568.232844] blk_mq_flush_plug_list+0x1d7/0x280 [ 568.232859] blk_flush_plug_list+0xc2/0xe0 [ 568.232872] blk_finish_plug+0x2c/0x40 [ 568.232886] ext4_writepages+0x592/0xe60 [ 568.233381] ? ext4_mark_iloc_dirty+0x52b/0x860 [ 568.233851] do_writepages+0x3c/0xd0 [ 568.234304] ? ext4_mark_inode_dirty+0x1a0/0x1a0 [ 568.234748] ? do_writepages+0x3c/0xd0 [ 568.235197] ? __generic_write_end+0x4e/0x80 [ 568.235644] __filemap_fdatawrite_range+0xa5/0xe0 [ 568.236089] ? __filemap_fdatawrite_range+0xa5/0xe0 [ 568.236533] ? ext4_da_write_end+0x13c/0x280 [ 568.236983] file_write_and_wait_range+0x5a/0xb0 [ 568.237407] ext4_sync_file+0x11e/0x3e0 [ 568.237819] vfs_fsync_range+0x48/0x80 [ 568.238217] ext4_file_write_iter+0x234/0x3d0 [ 568.238610] ? _cond_resched+0x19/0x40 [ 568.238982] new_sync_write+0x112/0x190 [ 568.239347] __vfs_write+0x29/0x40 [ 568.239705] vfs_write+0xb1/0x1a0 [ 568.240078] ksys_write+0x89/0xc0 [ 568.240428] __x64_sys_write+0x1a/0x20 [ 568.240771] do_syscall_64+0x5b/0x140 [ 568.241115] entry_SYSCALL_64_after_hwframe+0x49/0xbe [ 568.241456] RIP: 0033:0x7fb5b02325f4 [ 568.241787] Code: 00 f7 d8 64 89 02 48 c7 c0 ff ff ff ff eb b3 0f 1f 80 00 00 00 00 48 8d 05 09 11 2d 00 8b 00 85 c0 75 13 b8 01 00 00 00 0f 05 <48> 3d 00 f0 ff ff 77 54 f3 c3 66 90 41 54 55 49 89 d4 53 48 89 f5 [ 568.242842] RSP: 002b:00007ffcb12e2968 EFLAGS: 00000246 ORIG_RAX: 0000000000000001 [ 568.243220] RAX: ffffffffffffffda RBX: 0000000000000000 RCX: 00007fb5b02325f4 [ 568.243616] RDX: 0000000000000200 RSI: 000055698f2ad000 RDI: 0000000000000001 [ 568.244026] RBP: 0000000000000200 R08: 0000000000000004 R09: 0000000000000003 [ 568.244401] R10: 00007fb5b04feca0 R11: 0000000000000246 R12: 000055698f2ad000 [ 568.244775] R13: 0000000000000000 R14: 0000000000000000 R15: 000055698f2ad000 [ 568.245154] Modules linked in: xt_MASQUERADE(E) nf_conntrack_netlink(E) nfnetlink(E) xfrm_user(E) xfrm_algo(E) xt_addrtype(E) br_netfilter(E) bridge(E) stp(E) llc(E) overlay(E) vmw_vsock_vmci_transport(E) vsock(E) ip6table_filter(E) ip6_tables(E) xt_conntrack(E) iptable_mangle(E) iptable_nat(E) nf_nat(E) iptable_filter [ 568.248651] CR2: 0000000000000024 [ 568.249142] ---[ end trace 0ddd315e0a5bdfba ]--- Regards, Srivatsa VMware Photon OS