From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1753298AbdJaNvY (ORCPT ); Tue, 31 Oct 2017 09:51:24 -0400 Received: from bombadil.infradead.org ([65.50.211.133]:50094 "EHLO bombadil.infradead.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1751795AbdJaNvX (ORCPT ); Tue, 31 Oct 2017 09:51:23 -0400 Date: Tue, 31 Oct 2017 14:51:05 +0100 From: Peter Zijlstra To: Michal Hocko Cc: Byungchul Park , Dmitry Vyukov , syzbot , Andrew Morton , Dan Williams , Johannes Weiner , Jan Kara , jglisse@redhat.com, LKML , linux-mm@kvack.org, shli@fb.com, syzkaller-bugs@googlegroups.com, Thomas Gleixner , Vlastimil Babka , ying.huang@intel.com, kernel-team@lge.com Subject: Re: possible deadlock in lru_add_drain_all Message-ID: <20171031135104.rnlytzawi2xzuih3@hirez.programming.kicks-ass.net> References: <089e0825eec8955c1f055c83d476@google.com> <20171027093418.om5e566srz2ztsrk@dhcp22.suse.cz> <20171027134234.7dyx4oshjwd44vqx@dhcp22.suse.cz> <20171030082203.4xvq2af25shfci2z@dhcp22.suse.cz> <20171030100921.GA18085@X58A-UD3R> <20171030151009.ip4k7nwan7muouca@hirez.programming.kicks-ass.net> <20171031131333.pr2ophwd2bsvxc3l@dhcp22.suse.cz> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20171031131333.pr2ophwd2bsvxc3l@dhcp22.suse.cz> User-Agent: NeoMutt/20170609 (1.8.3) Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Tue, Oct 31, 2017 at 02:13:33PM +0100, Michal Hocko wrote: > On Mon 30-10-17 16:10:09, Peter Zijlstra wrote: > > However, that splat translates like: > > > > __cpuhp_setup_state() > > #0 cpus_read_lock() > > __cpuhp_setup_state_cpuslocked() > > #1 mutex_lock(&cpuhp_state_mutex) > > > > > > > > __cpuhp_state_add_instance() > > #2 mutex_lock(&cpuhp_state_mutex) > > this should be #1 right? Yes > > cpuhp_issue_call() > > cpuhp_invoke_ap_callback() > > #3 wait_for_completion() > > > > msr_device_create() > > ... > > #4 filename_create() > > #3 complete() > > > > > > > > do_splice() > > #4 file_start_write() > > do_splice_from() > > iter_file_splice_write() > > #5 pipe_lock() > > vfs_iter_write() > > ... > > #6 inode_lock() > > > > > > > > sys_fcntl() > > do_fcntl() > > shmem_fcntl() > > #5 inode_lock() And that #6 > > shmem_wait_for_pins() > > if (!scan) > > lru_add_drain_all() > > #0 cpus_read_lock() > > > > > > > > Which is an actual real deadlock, there is no mixing of up and down. > > thanks a lot, this made it more clear to me. It took a while to > actually see 0 -> 1 -> 3 -> 4 -> 5 -> 0 cycle. I have only focused > on lru_add_drain_all while it was holding the cpus lock. Yeah, these things are a pain to read, which is why I always construct something like the above first.