From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-2.5 required=3.0 tests=HEADER_FROM_DIFFERENT_DOMAINS, MAILING_LIST_MULTI,SPF_PASS,USER_AGENT_MUTT autolearn=unavailable autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 78E82C282C2 for ; Wed, 23 Jan 2019 18:02:39 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [209.132.180.67]) by mail.kernel.org (Postfix) with ESMTP id 4C8A221855 for ; Wed, 23 Jan 2019 18:02:39 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726178AbfAWSCd (ORCPT ); Wed, 23 Jan 2019 13:02:33 -0500 Received: from mx2.suse.de ([195.135.220.15]:41712 "EHLO mx1.suse.de" rhost-flags-OK-OK-OK-FAIL) by vger.kernel.org with ESMTP id S1725896AbfAWSCd (ORCPT ); Wed, 23 Jan 2019 13:02:33 -0500 X-Virus-Scanned: by amavisd-new at test-mx.suse.de Received: from relay2.suse.de (unknown [195.135.220.254]) by mx1.suse.de (Postfix) with ESMTP id E063DAFE7; Wed, 23 Jan 2019 18:02:30 +0000 (UTC) Received: by quack2.suse.cz (Postfix, from userid 1000) id 1B8BF1E3FF5; Wed, 23 Jan 2019 19:02:30 +0100 (CET) Date: Wed, 23 Jan 2019 19:02:30 +0100 From: Jan Kara To: Jerome Glisse Cc: Jan Kara , John Hubbard , Matthew Wilcox , Dave Chinner , Dan Williams , John Hubbard , Andrew Morton , Linux MM , tom@talpey.com, Al Viro , benve@cisco.com, Christoph Hellwig , Christopher Lameter , "Dalessandro, Dennis" , Doug Ledford , Jason Gunthorpe , Michal Hocko , mike.marciniszyn@intel.com, rcampbell@nvidia.com, Linux Kernel Mailing List , linux-fsdevel Subject: Re: [PATCH 1/2] mm: introduce put_user_page*(), placeholder versions Message-ID: <20190123180230.GN13149@quack2.suse.cz> References: <20190114145447.GJ13316@quack2.suse.cz> <20190114172124.GA3702@redhat.com> <20190115080759.GC29524@quack2.suse.cz> <20190116113819.GD26069@quack2.suse.cz> <20190116130813.GA3617@redhat.com> <20190117093047.GB9378@quack2.suse.cz> <20190117151759.GA3550@redhat.com> <20190122152459.GG13149@quack2.suse.cz> <20190122164613.GA3188@redhat.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20190122164613.GA3188@redhat.com> User-Agent: Mutt/1.10.1 (2018-07-13) Sender: linux-fsdevel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-fsdevel@vger.kernel.org On Tue 22-01-19 11:46:13, Jerome Glisse wrote: > On Tue, Jan 22, 2019 at 04:24:59PM +0100, Jan Kara wrote: > > On Thu 17-01-19 10:17:59, Jerome Glisse wrote: > > > On Thu, Jan 17, 2019 at 10:30:47AM +0100, Jan Kara wrote: > > > > On Wed 16-01-19 08:08:14, Jerome Glisse wrote: > > > > > On Wed, Jan 16, 2019 at 12:38:19PM +0100, Jan Kara wrote: > > > > > > On Tue 15-01-19 09:07:59, Jan Kara wrote: > > > > > > > Agreed. So with page lock it would actually look like: > > > > > > > > > > > > > > get_page_pin() > > > > > > > lock_page(page); > > > > > > > wait_for_stable_page(); > > > > > > > atomic_add(&page->_refcount, PAGE_PIN_BIAS); > > > > > > > unlock_page(page); > > > > > > > > > > > > > > And if we perform page_pinned() check under page lock, then if > > > > > > > page_pinned() returned false, we are sure page is not and will not be > > > > > > > pinned until we drop the page lock (and also until page writeback is > > > > > > > completed if needed). > > > > > > > > > > > > After some more though, why do we even need wait_for_stable_page() and > > > > > > lock_page() in get_page_pin()? > > > > > > > > > > > > During writepage page_mkclean() will write protect all page tables. So > > > > > > there can be no new writeable GUP pins until we unlock the page as all such > > > > > > GUPs will have to first go through fault and ->page_mkwrite() handler. And > > > > > > that will wait on page lock and do wait_for_stable_page() for us anyway. > > > > > > Am I just confused? > > > > > > > > > > Yeah with page lock it should synchronize on the pte but you still > > > > > need to check for writeback iirc the page is unlocked after file > > > > > system has queue up the write and thus the page can be unlock with > > > > > write back pending (and PageWriteback() == trye) and i am not sure > > > > > that in that states we can safely let anyone write to that page. I > > > > > am assuming that in some case the block device also expect stable > > > > > page content (RAID stuff). > > > > > > > > > > So the PageWriteback() test is not only for racing page_mkclean()/ > > > > > test_set_page_writeback() and GUP but also for pending write back. > > > > > > > > But this is prevented by wait_for_stable_page() that is already present in > > > > ->page_mkwrite() handlers. Look: > > > > > > > > ->writepage() > > > > /* Page is locked here */ > > > > clear_page_dirty_for_io(page) > > > > page_mkclean(page) > > > > -> page tables get writeprotected > > > > /* The following line will be added by our patches */ > > > > if (page_pinned(page)) -> bounce > > > > TestClearPageDirty(page) > > > > set_page_writeback(page); > > > > unlock_page(page); > > > > ...submit_io... > > > > > > > > IRQ > > > > - IO completion > > > > end_page_writeback() > > > > > > > > So if GUP happens before page_mkclean() writeprotects corresponding PTE > > > > (and these two actions are synchronized on the PTE lock), page_pinned() > > > > will see the increment and report the page as pinned. > > > > > > > > If GUP happens after page_mkclean() writeprotects corresponding PTE, it > > > > will fault: > > > > handle_mm_fault() > > > > do_wp_page() > > > > wp_page_shared() > > > > do_page_mkwrite() > > > > ->page_mkwrite() - that is block_page_mkwrite() or > > > > iomap_page_mkwrite() or whatever filesystem provides > > > > lock_page(page) > > > > ... prepare page ... > > > > wait_for_stable_page(page) -> this blocks until IO completes > > > > if someone cares about pages not being modified while under IO. > > > > > > The case i am worried is GUP see pte with write flag set but has not > > > lock the page yet (GUP is get pte first, then pte to page then lock > > > page), then it locks the page but the lock page can make it wait for a > > > racing page_mkclean()...write back that have not yet write protected > > > the pte the GUP just read. So by the time GUP has the page locked the > > > pte it read might no longer have the write flag set. Hence why you need > > > to also check for write back after taking the page lock. Alternatively > > > you could recheck the pte after a successful try_lock on the page. > > > > This isn't really possible. GUP does: > > > > get_user_pages() > > ... > > follow_page_mask() > > ... > > follow_page_pte() > > ptep = pte_offset_map_lock() > > check permissions and page sanity > > if (flags & FOLL_GET) > > get_page(page); -> this would become > > atomic_add(&page->_refcount, PAGE_PIN_BIAS); > > pte_unmap_unlock(ptep, ptl); > > > > page_mkclean() on the other hand grabs the same pte lock to change the pte > > to write-protected. So after page_mkclean() has modified the PTE we are > > racing on for access, we are sure to either see increased _refcount or get > > page fault from GUP. > > > > If we see increased _refcount, we bounce the page and are fine. If GUP > > faults, we will wait for page lock (so wait until page is prepared for IO > > and has PageWriteback set) while handling the fault, then enter > > ->page_mkwrite, which will do wait_for_stable_page() -> wait for > > outstanding writeback to complete. > > > > So I still conclude - no need for page lock in the GUP path at all AFAICT. > > In fact we rely on the very same page fault vs page writeback synchronization > > for normal user faults as well. And normal user mmap access is even nastier > > than GUP access because the CPU reads page tables without taking PTE lock. > > For the "slow" GUP path you are right you do not need a lock as the > page table lock give you the ordering. For the GUP fast path you > would either need the lock or the memory barrier with the test for > page write back. > > Maybe an easier thing is to convert GUP fast to try to take the page > table lock if it fails taking the page table lock then we fall back > to slow GUP path. Otherwise then we have the same garantee as the slow > path. You're right I was looking at the wrong place for GUP_fast() path. But I still don't think anything special (i.e. page lock or new barrier) is necessary. GUP_fast() takes care already now that it cannot race with page unmapping or write-protection (as there are other places in MM that rely on this). Look, gup_pte_range() has: if (!page_cache_get_speculative(head)) goto pte_unmap; if (unlikely(pte_val(pte) != pte_val(*ptep))) { put_page(head); goto pte_unmap; } So that page_cache_get_speculative() will become page_cache_pin_speculative() to increment refcount by PAGE_PIN_BIAS instead of 1. That is atomic ordered operation so it cannot be reordered with the following check that PTE stayed same. So once page_mkclean() write-protects PTE, there can be no new pins from GUP_fast() and we are sure all succeeding pins are visible in page->_refcount after page_mkclean() completes. Again this is nothing new, other mm code already relies on either seeing page->_refcount incremented or GUP fast bailing out (e.g. DAX relies on this). Although strictly speaking I'm not 100% sure what prevents page->_refcount load to be speculatively reordered before PTE update even in current places using this but there's so much stuff inbetween that there's probably something ;). But we could add smp_rmb() after page_mkclean() before changing page_pinned() for the peace of mind I guess. > The issue is that i am not sure if the page table directory page and > it's associated spinlock can go in bad state if the directory is being > freed (like a racing munmap). This would need to be check. A scheme > that might protect against that is to take the above lock of each level > before going down one level. Once you are down one level you can unlock > the above level. So at any point in time GUP fast holds the lock to a > current and valid directory and thus no one could race to remove it. > > GUP_fast() > gup_pgd_range() > if (p4d_try_map_lock()) { > gup_p4d_range() > if (pud_try_map_lock()) { > p4d_unlock(); > gup_pud_range(); > if (pmd_try_map_lock()) { > pud_unlock(); > gup_pmd_range(); > if (pte_try_map_lock()) { > pmd_unlock(); > // Do gup > } > } > } > } > > Maybe this is worse than taking the mmap_sem and checking for vma. > > > > > > > > That actually touches on another question I wanted to get opinions on. GUP > > > > > > can be for read and GUP can be for write (that is one of GUP flags). > > > > > > Filesystems with page cache generally have issues only with GUP for write > > > > > > as it can currently corrupt data, unexpectedly dirty page etc.. DAX & memory > > > > > > hotplug have issues with both (DAX cannot truncate page pinned in any way, > > > > > > memory hotplug will just loop in kernel until the page gets unpinned). So > > > > > > we probably want to track both types of GUP pins and page-cache based > > > > > > filesystems will take the hit even if they don't have to for read-pins? > > > > > > > > > > Yes the distinction between read and write would be nice. With the map > > > > > count solution you can only increment the mapcount for GUP(write=true). > > > > > > > > Well, but if we track only pins for write, DAX or memory hotplug will not > > > > be able to use this mechanism. So at this point I'm more leaning towards > > > > tracking all pins. It will cost some performance needlessly for read pins > > > > and filesystems using page cache when bouncing such pages but it's not like > > > > writeback of pinned pages is some performance critical operation... But I > > > > wanted to spell this out so that people are aware of this. > > > > > > No they would know for regular pin, it is just as page migrate code. If > > > the refcount + (extra_ref_by_the_code_checking) > mapcount then you know > > > someone has extra reference on your page. > > > > > > Those extra references are either some regular fs event taking place (some > > > code doing find_get_page for instance) or a GUP reference (wether it is a > > > write pin or a read pin). > > > > > > So the only issue is false positive, ie thinking the page is under GUP > > > while it has just elevated refcount because of some other regular fs/mm > > > event. To minimize false positive for a more accurate pin test (write or > > > read) you can enforce few thing: > > > > > > 1 - first page lock > > > 2 - then freeze the page with expected counted > > > > > > With that it should minimize false positive. In the end even with the bias > > > case you can also have false positive. > > > > So this is basically what the code is currently doing. And for DAX it works > > well since the page is being truncated and so essentially nobody is > > touching it. But for hotplug it doesn't work quite well - hotplug would > > like to return EBUSY to userspace when the page is pinned but retry if the > > page reference is just transient. > > I do not think there is anyway around transient refcount other > than periodicaly check. Maybe hot unplug (i am assuming we are > talking about unplug here) can set the reserved page flag and > we can change the page get ref to never inc refcount for page > with reserved flag. I see that they have been cleanup on going > around reserved page so it might or might not be possible. Well, there's no problem with transient refcount. The code is fine with retrying in that case. The problem is when the refcount actually is *not* transient and so hot-unplug ends up retrying forever (or for very long time). In that case it should rather bail with EBUSY but currently there's no way to distinguish transient from longer term pin. Honza -- Jan Kara SUSE Labs, CR