From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-11.1 required=3.0 tests=BAYES_00,DKIMWL_WL_HIGH, DKIM_SIGNED,DKIM_VALID,DKIM_VALID_AU,HEADER_FROM_DIFFERENT_DOMAINS, INCLUDES_PATCH,MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id E0E90C433DB for ; Mon, 1 Feb 2021 19:21:28 +0000 (UTC) Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by mail.kernel.org (Postfix) with ESMTP id 6443064DE3 for ; Mon, 1 Feb 2021 19:21:28 +0000 (UTC) DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org 6443064DE3 Authentication-Results: mail.kernel.org; dmarc=fail (p=none dis=none) header.from=redhat.com Authentication-Results: mail.kernel.org; spf=pass smtp.mailfrom=owner-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix) id C4B0D6B006C; Mon, 1 Feb 2021 14:21:27 -0500 (EST) Received: by kanga.kvack.org (Postfix, from userid 40) id BF9A76B006E; Mon, 1 Feb 2021 14:21:27 -0500 (EST) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id A9EC86B0070; Mon, 1 Feb 2021 14:21:27 -0500 (EST) X-Delivered-To: linux-mm@kvack.org Received: from forelay.hostedemail.com (smtprelay0246.hostedemail.com [216.40.44.246]) by kanga.kvack.org (Postfix) with ESMTP id 925876B006C for ; Mon, 1 Feb 2021 14:21:27 -0500 (EST) Received: from smtpin18.hostedemail.com (10.5.19.251.rfc1918.com [10.5.19.251]) by forelay04.hostedemail.com (Postfix) with ESMTP id 5B9EE1EE6 for ; Mon, 1 Feb 2021 19:21:27 +0000 (UTC) X-FDA: 77770667814.18.cover49_3414472275c4 Received: from filter.hostedemail.com (10.5.16.251.rfc1918.com [10.5.16.251]) by smtpin18.hostedemail.com (Postfix) with ESMTP id 329F3100ED3C7 for ; Mon, 1 Feb 2021 19:21:27 +0000 (UTC) X-HE-Tag: cover49_3414472275c4 X-Filterd-Recvd-Size: 11159 Received: from us-smtp-delivery-124.mimecast.com (us-smtp-delivery-124.mimecast.com [63.128.21.124]) by imf19.hostedemail.com (Postfix) with ESMTP for ; Mon, 1 Feb 2021 19:21:26 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1612207285; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: in-reply-to:in-reply-to:references:references; bh=Q2b/N50hu0sHq/kMoTQCMLs3X1tcdQtb/zYbLnwU7BM=; b=eQehGxMSD1Yt3VqGRbujaHOkUo9z0np61Q2s/CBT2VPsHx3jS9xCLMMiLM7V8l1AmLQWNr XWh+t03aF96ny7jCCh6JEGAsXeX41iF5qkA3U1YdT+gDyR1kXSNywRzAlXqKcn5EXUtp7l /qoI5ou6FFExyz+BRhLgSNYhQQJYP7s= Received: from mail-qk1-f197.google.com (mail-qk1-f197.google.com [209.85.222.197]) (Using TLS) by relay.mimecast.com with ESMTP id us-mta-266-6q6azRcdPCiDB6c86egYJg-1; Mon, 01 Feb 2021 14:21:24 -0500 X-MC-Unique: 6q6azRcdPCiDB6c86egYJg-1 Received: by mail-qk1-f197.google.com with SMTP id m64so14122434qke.12 for ; Mon, 01 Feb 2021 11:21:24 -0800 (PST) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:date:from:to:cc:subject:message-id:references :mime-version:content-disposition:in-reply-to; bh=Q2b/N50hu0sHq/kMoTQCMLs3X1tcdQtb/zYbLnwU7BM=; b=eTK+t637e1N2p0z+HvUQWSmOcyW73LSdw9QzxGmQ2k7x5Cu+TLkGM+VZyncyGSBnI7 lQzn4Mtjq346hV9/gHlPeMazfLoWT7FUp8Sr1jPv+9F4iU5gHOGrnsQdQBE+AJX5HnEw a4/wSfgA+ibIjqWSxCQppTH8R7tLtOt+I1XOkVURUUO+dvlZP+GJMAi0YHjY/C2zpxqM YpCAx6QzAs16NctFeMVgiUmxZFMQtAU4AbWXyqlTfjDlkvw9TnO2YJjjt8K2fACir1JX lou3CY5/IAzrnehFTxfsryno50iswDQh1s1UrC9idn9YCRm9oPNHo5cvIpGY92FldXZC OS9A== X-Gm-Message-State: AOAM531V9RQ/QBKjQsANIzrVpsRJZR4YEZ/HJcpHtKWwYp9qGmeoR6Lz 2SaJSuMUFaBXEYUawuzvgnjGRMOzBxq0Br8OIhTVf37C6w3FEHKuSrZAwVENk65lCwx0nStmdXM hb2JBrdyY34c= X-Received: by 2002:ac8:5cd0:: with SMTP id s16mr16186190qta.309.1612207283697; Mon, 01 Feb 2021 11:21:23 -0800 (PST) X-Google-Smtp-Source: ABdhPJwWmFLJGbGwctGCP9A/nYWy61WeOGTKC6d+y1Ba8IhsTF2WrQd3mliiW1G26lcHYKqEou/z+w== X-Received: by 2002:ac8:5cd0:: with SMTP id s16mr16186156qta.309.1612207283172; Mon, 01 Feb 2021 11:21:23 -0800 (PST) Received: from xz-x1 ([142.126.83.202]) by smtp.gmail.com with ESMTPSA id o5sm14755572qko.85.2021.02.01.11.21.21 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Mon, 01 Feb 2021 11:21:22 -0800 (PST) Date: Mon, 1 Feb 2021 14:21:20 -0500 From: Peter Xu To: Axel Rasmussen Cc: Alexander Viro , Alexey Dobriyan , Andrea Arcangeli , Andrew Morton , Anshuman Khandual , Catalin Marinas , Chinwen Chang , Huang Ying , Ingo Molnar , Jann Horn , Jerome Glisse , Lokesh Gidra , "Matthew Wilcox (Oracle)" , Michael Ellerman , Michal =?utf-8?Q?Koutn=C3=BD?= , Michel Lespinasse , Mike Kravetz , Mike Rapoport , Nicholas Piggin , Shaohua Li , Shawn Anastasio , Steven Rostedt , Steven Price , Vlastimil Babka , linux-kernel@vger.kernel.org, linux-fsdevel@vger.kernel.org, linux-mm@kvack.org, Adam Ruprecht , Cannon Matthews , "Dr . David Alan Gilbert" , David Rientjes , Oliver Upton Subject: Re: [PATCH v3 7/9] userfaultfd: add UFFDIO_CONTINUE ioctl Message-ID: <20210201192120.GG260413@xz-x1> References: <20210128224819.2651899-1-axelrasmussen@google.com> <20210128224819.2651899-8-axelrasmussen@google.com> MIME-Version: 1.0 In-Reply-To: <20210128224819.2651899-8-axelrasmussen@google.com> Authentication-Results: relay.mimecast.com; auth=pass smtp.auth=CUSA124A263 smtp.mailfrom=peterx@redhat.com X-Mimecast-Spam-Score: 0 X-Mimecast-Originator: redhat.com Content-Type: text/plain; charset=utf-8 Content-Disposition: inline X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: On Thu, Jan 28, 2021 at 02:48:17PM -0800, Axel Rasmussen wrote: > diff --git a/include/linux/hugetlb.h b/include/linux/hugetlb.h > index f94a35296618..79e1f0155afa 100644 > --- a/include/linux/hugetlb.h > +++ b/include/linux/hugetlb.h > @@ -135,11 +135,14 @@ void hugetlb_show_meminfo(void); > unsigned long hugetlb_total_pages(void); > vm_fault_t hugetlb_fault(struct mm_struct *mm, struct vm_area_struct *vma, > unsigned long address, unsigned int flags); > +#ifdef CONFIG_USERFAULTFD I'm confused why this is needed.. hugetlb_mcopy_atomic_pte() should only be called in userfaultfd.c, but if without uffd config set it won't compile either: obj-$(CONFIG_USERFAULTFD) += userfaultfd.o > int hugetlb_mcopy_atomic_pte(struct mm_struct *dst_mm, pte_t *dst_pte, > struct vm_area_struct *dst_vma, > unsigned long dst_addr, > unsigned long src_addr, > + enum mcopy_atomic_mode mode, > struct page **pagep); > +#endif > int hugetlb_reserve_pages(struct inode *inode, long from, long to, > struct vm_area_struct *vma, > vm_flags_t vm_flags); > diff --git a/include/linux/userfaultfd_k.h b/include/linux/userfaultfd_k.h > index fb9abaeb4194..2fcb686211e8 100644 > --- a/include/linux/userfaultfd_k.h > +++ b/include/linux/userfaultfd_k.h > @@ -37,6 +37,22 @@ extern int sysctl_unprivileged_userfaultfd; > > extern vm_fault_t handle_userfault(struct vm_fault *vmf, unsigned long reason); > > +/* > + * The mode of operation for __mcopy_atomic and its helpers. > + * > + * This is almost an implementation detail (mcopy_atomic below doesn't take this > + * as a parameter), but it's exposed here because memory-kind-specific > + * implementations (e.g. hugetlbfs) need to know the mode of operation. > + */ > +enum mcopy_atomic_mode { > + /* A normal copy_from_user into the destination range. */ > + MCOPY_ATOMIC_NORMAL, > + /* Don't copy; map the destination range to the zero page. */ > + MCOPY_ATOMIC_ZEROPAGE, > + /* Just setup the dst_vma, without modifying the underlying page(s). */ > + MCOPY_ATOMIC_CONTINUE, > +}; > + Maybe better to keep this to where it's used, e.g. hugetlb.h where we've defined hugetlb_mcopy_atomic_pte()? [...] > diff --git a/mm/hugetlb.c b/mm/hugetlb.c > index 6f9d8349f818..3d318ef3d180 100644 > --- a/mm/hugetlb.c > +++ b/mm/hugetlb.c > @@ -4647,6 +4647,7 @@ vm_fault_t hugetlb_fault(struct mm_struct *mm, struct vm_area_struct *vma, > return ret; > } > > +#ifdef CONFIG_USERFAULTFD So I feel like you added the header ifdef for this. IMHO we can drop both since that's what we have had. I agree maybe it's better to not compile that without CONFIG_USERFAULTFD but that may worth a standalone patch anyways. > /* > * Used by userfaultfd UFFDIO_COPY. Based on mcopy_atomic_pte with > * modifications for huge pages. > @@ -4656,6 +4657,7 @@ int hugetlb_mcopy_atomic_pte(struct mm_struct *dst_mm, > struct vm_area_struct *dst_vma, > unsigned long dst_addr, > unsigned long src_addr, > + enum mcopy_atomic_mode mode, > struct page **pagep) > { > struct address_space *mapping; > @@ -4668,7 +4670,10 @@ int hugetlb_mcopy_atomic_pte(struct mm_struct *dst_mm, > int ret; > struct page *page; > > - if (!*pagep) { > + mapping = dst_vma->vm_file->f_mapping; > + idx = vma_hugecache_offset(h, dst_vma, dst_addr); > + > + if (!*pagep && mode != MCOPY_ATOMIC_CONTINUE) { > ret = -ENOMEM; > page = alloc_huge_page(dst_vma, dst_addr, 0); > if (IS_ERR(page)) > @@ -4685,6 +4690,12 @@ int hugetlb_mcopy_atomic_pte(struct mm_struct *dst_mm, > /* don't free the page */ > goto out; > } > + } else if (mode == MCOPY_ATOMIC_CONTINUE) { > + ret = -EFAULT; > + page = find_lock_page(mapping, idx); > + *pagep = NULL; > + if (!page) > + goto out; > } else { > page = *pagep; > *pagep = NULL; I would write this as: if (mode == MCOPY_ATOMIC_CONTINUE) ... else if (!*pagep) ... else ... No strong opinion, but that'll look slightly cleaner to me. [...] > @@ -408,7 +407,7 @@ extern ssize_t __mcopy_atomic_hugetlb(struct mm_struct *dst_mm, > unsigned long dst_start, > unsigned long src_start, > unsigned long len, > - bool zeropage); > + enum mcopy_atomic_mode mode); > #endif /* CONFIG_HUGETLB_PAGE */ > > static __always_inline ssize_t mfill_atomic_pte(struct mm_struct *dst_mm, > @@ -417,7 +416,7 @@ static __always_inline ssize_t mfill_atomic_pte(struct mm_struct *dst_mm, > unsigned long dst_addr, > unsigned long src_addr, > struct page **page, > - bool zeropage, > + enum mcopy_atomic_mode mode, > bool wp_copy) > { > ssize_t err; > @@ -433,22 +432,38 @@ static __always_inline ssize_t mfill_atomic_pte(struct mm_struct *dst_mm, > * and not in the radix tree. > */ > if (!(dst_vma->vm_flags & VM_SHARED)) { > - if (!zeropage) > + switch (mode) { > + case MCOPY_ATOMIC_NORMAL: > err = mcopy_atomic_pte(dst_mm, dst_pmd, dst_vma, > dst_addr, src_addr, page, > wp_copy); > - else > + break; > + case MCOPY_ATOMIC_ZEROPAGE: > err = mfill_zeropage_pte(dst_mm, dst_pmd, > dst_vma, dst_addr); > + break; > + /* It only makes sense to CONTINUE for shared memory. */ > + case MCOPY_ATOMIC_CONTINUE: > + err = -EINVAL; > + break; > + } > } else { > VM_WARN_ON_ONCE(wp_copy); > - if (!zeropage) > + switch (mode) { > + case MCOPY_ATOMIC_NORMAL: > err = shmem_mcopy_atomic_pte(dst_mm, dst_pmd, > dst_vma, dst_addr, > src_addr, page); > - else > + break; > + case MCOPY_ATOMIC_ZEROPAGE: > err = shmem_mfill_zeropage_pte(dst_mm, dst_pmd, > dst_vma, dst_addr); > + break; > + case MCOPY_ATOMIC_CONTINUE: > + /* FIXME: Add minor fault interception for shmem. */ > + err = -EINVAL; > + break; > + } > } > > return err; The whole chunk above is not needed for hugetlbfs it seems - I'd avoid touching the anon/shmem code path until it's being supported. What you need is probably set zeropage as below in __mcopy_atomic(): zeropage = (mode == MCOPY_ATOMIC_ZEROPAGE); Before passing it over to mfill_atomic_pte(). As long as we reject UFFDIO_CONTINUE with !hugetlbfs correctly that'll be enough iiuc. Thanks, -- Peter Xu