From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-5.5 required=3.0 tests=BAYES_00, HEADER_FROM_DIFFERENT_DOMAINS,MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS, URIBL_BLOCKED,USER_AGENT_SANE_1 autolearn=no autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id C7CB1C433E1 for ; Fri, 31 Jul 2020 16:22:44 +0000 (UTC) Received: from ml01.01.org (ml01.01.org [198.145.21.10]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPS id 97D9222D04 for ; Fri, 31 Jul 2020 16:22:44 +0000 (UTC) DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org 97D9222D04 Authentication-Results: mail.kernel.org; dmarc=none (p=none dis=none) header.from=arm.com Authentication-Results: mail.kernel.org; spf=none smtp.mailfrom=linux-nvdimm-bounces@lists.01.org Received: from ml01.vlan13.01.org (localhost [IPv6:::1]) by ml01.01.org (Postfix) with ESMTP id 6AF2A12918A66; Fri, 31 Jul 2020 09:22:44 -0700 (PDT) Received-SPF: Pass (mailfrom) identity=mailfrom; client-ip=198.145.29.99; helo=mail.kernel.org; envelope-from=cmarinas@kernel.org; receiver= Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by ml01.01.org (Postfix) with ESMTPS id 8549612918A61 for ; Fri, 31 Jul 2020 09:22:42 -0700 (PDT) Received: from gaia (unknown [95.146.230.158]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPSA id 1B0052177B; Fri, 31 Jul 2020 16:22:36 +0000 (UTC) Date: Fri, 31 Jul 2020 17:22:34 +0100 From: Catalin Marinas To: Mark Rutland Subject: Re: [PATCH v2 3/7] mm: introduce memfd_secret system call to create "secret" memory areas Message-ID: <20200731162234.GF29569@gaia> References: <20200727162935.31714-1-rppt@kernel.org> <20200727162935.31714-4-rppt@kernel.org> <20200730162209.GB3128@gaia> <20200731142905.GA67415@C02TD0UTHF1T.local> MIME-Version: 1.0 Content-Disposition: inline In-Reply-To: <20200731142905.GA67415@C02TD0UTHF1T.local> User-Agent: Mutt/1.10.1 (2018-07-13) Message-ID-Hash: MBGGTEVDAZ4KERO536EKRS4YWEKGTOV2 X-Message-ID-Hash: MBGGTEVDAZ4KERO536EKRS4YWEKGTOV2 X-MailFrom: cmarinas@kernel.org X-Mailman-Rule-Hits: nonmember-moderation X-Mailman-Rule-Misses: dmarc-mitigation; no-senders; approved; emergency; loop; banned-address; member-moderation CC: Mike Rapoport , linux-kernel@vger.kernel.org, Alexander Viro , Andrew Morton , Andy Lutomirski , Arnd Bergmann , Borislav Petkov , Christopher Lameter , Dave Hansen , Elena Reshetova , "H. Peter Anvin" , Idan Yaniv , Ingo Molnar , James Bottomley , "Kirill A. Shutemov" , Matthew Wilcox , Mike Rapoport , Michael Kerrisk , Palmer Dabbelt , Paul Walmsley , Peter Zijlstra , Thomas Gleixner , Tycho Andersen , Will Deacon , linux-api@vger.kernel.org, linux-arch@vger.kernel.org, linux-arm-kernel@lists.infradead.org, linux-fsdevel @vger.kernel.org, linux-mm@kvack.org, linux-nvdimm@lists.01.org, linux-riscv@lists.infradead.org, x86@kernel.org X-Mailman-Version: 3.1.1 Precedence: list List-Id: "Linux-nvdimm developer list." Archived-At: List-Archive: List-Help: List-Post: List-Subscribe: List-Unsubscribe: Content-Type: text/plain; charset="us-ascii" Content-Transfer-Encoding: 7bit On Fri, Jul 31, 2020 at 03:29:05PM +0100, Mark Rutland wrote: > On Thu, Jul 30, 2020 at 05:22:10PM +0100, Catalin Marinas wrote: > > On Mon, Jul 27, 2020 at 07:29:31PM +0300, Mike Rapoport wrote: > > > +static int secretmem_mmap(struct file *file, struct vm_area_struct *vma) > > > +{ > > > + struct secretmem_ctx *ctx = file->private_data; > > > + unsigned long mode = ctx->mode; > > > + unsigned long len = vma->vm_end - vma->vm_start; > > > + > > > + if (!mode) > > > + return -EINVAL; > > > + > > > + if ((vma->vm_flags & (VM_SHARED | VM_MAYSHARE)) == 0) > > > + return -EINVAL; > > > + > > > + if (mlock_future_check(vma->vm_mm, vma->vm_flags | VM_LOCKED, len)) > > > + return -EAGAIN; > > > + > > > + switch (mode) { > > > + case SECRETMEM_UNCACHED: > > > + vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot); > > > + fallthrough; > > > + case SECRETMEM_EXCLUSIVE: > > > + vma->vm_ops = &secretmem_vm_ops; > > > + break; > > > + default: > > > + return -EINVAL; > > > + } > > > + > > > + vma->vm_flags |= VM_LOCKED; > > > + > > > + return 0; > > > +} > > > > I think the uncached mapping is not the right thing for arm/arm64. First > > of all, pgprot_noncached() gives us Strongly Ordered (Device memory) > > semantics together with not allowing unaligned accesses. I suspect the > > semantics are different on x86. > > > The second, more serious problem, is that I can't find any place where > > the caches are flushed for the page mapped on fault. When a page is > > allocated, assuming GFP_ZERO, only the caches are guaranteed to be > > zeroed. Exposing this subsequently to user space as uncached would allow > > the user to read stale data prior to zeroing. The arm64 > > set_direct_map_default_noflush() doesn't do any cache maintenance. > > It's also worth noting that in a virtual machine this is liable to be > either broken (with a potential loss of coherency if the host has a > cacheable alias as existing KVM hosts have), or pointless (if the host > uses S2FWB to upgrade Stage-1 attribues to cacheable as existing KVM > hosts also have). > > I think that trying to avoid the data caches creates many more problems > than it solves, and I don't think there's a strong justification for > trying to support that on arm64 to begin with, so I'd rather entirely > opt-out on supporting SECRETMEM_UNCACHED. Good point, I forgot the virtualisation aspect. So unless there is a hypervisor API to unmap it from the host memory, the uncached option isn't of much use on arm64. -- Catalin _______________________________________________ Linux-nvdimm mailing list -- linux-nvdimm@lists.01.org To unsubscribe send an email to linux-nvdimm-leave@lists.01.org From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-5.5 required=3.0 tests=BAYES_00, HEADER_FROM_DIFFERENT_DOMAINS,MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS, USER_AGENT_SANE_1 autolearn=no autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 5CFE2C433DF for ; Fri, 31 Jul 2020 16:22:44 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id 44E6122B40 for ; Fri, 31 Jul 2020 16:22:44 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1732227AbgGaQWn (ORCPT ); Fri, 31 Jul 2020 12:22:43 -0400 Received: from mail.kernel.org ([198.145.29.99]:34072 "EHLO mail.kernel.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1729074AbgGaQWm (ORCPT ); Fri, 31 Jul 2020 12:22:42 -0400 Received: from gaia (unknown [95.146.230.158]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPSA id 1B0052177B; Fri, 31 Jul 2020 16:22:36 +0000 (UTC) Date: Fri, 31 Jul 2020 17:22:34 +0100 From: Catalin Marinas To: Mark Rutland Cc: Mike Rapoport , linux-kernel@vger.kernel.org, Alexander Viro , Andrew Morton , Andy Lutomirski , Arnd Bergmann , Borislav Petkov , Christopher Lameter , Dan Williams , Dave Hansen , Elena Reshetova , "H. Peter Anvin" , Idan Yaniv , Ingo Molnar , James Bottomley , "Kirill A. Shutemov" , Matthew Wilcox , Mike Rapoport , Michael Kerrisk , Palmer Dabbelt , Paul Walmsley , Peter Zijlstra , Thomas Gleixner , Tycho Andersen , Will Deacon , linux-api@vger.kernel.org, linux-arch@vger.kernel.org, linux-arm-kernel@lists.infradead.org, linux-fsdevel@vger.kernel.org, linux-mm@kvack.org, linux-nvdimm@lists.01.org, linux-riscv@lists.infradead.org, x86@kernel.org Subject: Re: [PATCH v2 3/7] mm: introduce memfd_secret system call to create "secret" memory areas Message-ID: <20200731162234.GF29569@gaia> References: <20200727162935.31714-1-rppt@kernel.org> <20200727162935.31714-4-rppt@kernel.org> <20200730162209.GB3128@gaia> <20200731142905.GA67415@C02TD0UTHF1T.local> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20200731142905.GA67415@C02TD0UTHF1T.local> User-Agent: Mutt/1.10.1 (2018-07-13) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Fri, Jul 31, 2020 at 03:29:05PM +0100, Mark Rutland wrote: > On Thu, Jul 30, 2020 at 05:22:10PM +0100, Catalin Marinas wrote: > > On Mon, Jul 27, 2020 at 07:29:31PM +0300, Mike Rapoport wrote: > > > +static int secretmem_mmap(struct file *file, struct vm_area_struct *vma) > > > +{ > > > + struct secretmem_ctx *ctx = file->private_data; > > > + unsigned long mode = ctx->mode; > > > + unsigned long len = vma->vm_end - vma->vm_start; > > > + > > > + if (!mode) > > > + return -EINVAL; > > > + > > > + if ((vma->vm_flags & (VM_SHARED | VM_MAYSHARE)) == 0) > > > + return -EINVAL; > > > + > > > + if (mlock_future_check(vma->vm_mm, vma->vm_flags | VM_LOCKED, len)) > > > + return -EAGAIN; > > > + > > > + switch (mode) { > > > + case SECRETMEM_UNCACHED: > > > + vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot); > > > + fallthrough; > > > + case SECRETMEM_EXCLUSIVE: > > > + vma->vm_ops = &secretmem_vm_ops; > > > + break; > > > + default: > > > + return -EINVAL; > > > + } > > > + > > > + vma->vm_flags |= VM_LOCKED; > > > + > > > + return 0; > > > +} > > > > I think the uncached mapping is not the right thing for arm/arm64. First > > of all, pgprot_noncached() gives us Strongly Ordered (Device memory) > > semantics together with not allowing unaligned accesses. I suspect the > > semantics are different on x86. > > > The second, more serious problem, is that I can't find any place where > > the caches are flushed for the page mapped on fault. When a page is > > allocated, assuming GFP_ZERO, only the caches are guaranteed to be > > zeroed. Exposing this subsequently to user space as uncached would allow > > the user to read stale data prior to zeroing. The arm64 > > set_direct_map_default_noflush() doesn't do any cache maintenance. > > It's also worth noting that in a virtual machine this is liable to be > either broken (with a potential loss of coherency if the host has a > cacheable alias as existing KVM hosts have), or pointless (if the host > uses S2FWB to upgrade Stage-1 attribues to cacheable as existing KVM > hosts also have). > > I think that trying to avoid the data caches creates many more problems > than it solves, and I don't think there's a strong justification for > trying to support that on arm64 to begin with, so I'd rather entirely > opt-out on supporting SECRETMEM_UNCACHED. Good point, I forgot the virtualisation aspect. So unless there is a hypervisor API to unmap it from the host memory, the uncached option isn't of much use on arm64. -- Catalin From mboxrd@z Thu Jan 1 00:00:00 1970 From: Catalin Marinas Subject: Re: [PATCH v2 3/7] mm: introduce memfd_secret system call to create "secret" memory areas Date: Fri, 31 Jul 2020 17:22:34 +0100 Message-ID: <20200731162234.GF29569@gaia> References: <20200727162935.31714-1-rppt@kernel.org> <20200727162935.31714-4-rppt@kernel.org> <20200730162209.GB3128@gaia> <20200731142905.GA67415@C02TD0UTHF1T.local> Mime-Version: 1.0 Content-Type: text/plain; charset=us-ascii Return-path: Content-Disposition: inline In-Reply-To: <20200731142905.GA67415-NxAhncQ8SJbILnBEAk/BfazUEOm+Xw19@public.gmane.org> Sender: linux-api-owner-u79uwXL29TY76Z2rM5mHXA@public.gmane.org To: Mark Rutland Cc: Mike Rapoport , linux-kernel-u79uwXL29TY76Z2rM5mHXA@public.gmane.org, Alexander Viro , Andrew Morton , Andy Lutomirski , Arnd Bergmann , Borislav Petkov , Christopher Lameter , Dan Williams , Dave Hansen , Elena Reshetova , "H. Peter Anvin" , Idan Yaniv , Ingo Molnar , James Bottomley , "Kirill A. Shutemov" , Matthew Wilcox , Mike Rapoport , Michael Kerrisk , Palmer Dabbelt , Paul List-Id: linux-arch.vger.kernel.org On Fri, Jul 31, 2020 at 03:29:05PM +0100, Mark Rutland wrote: > On Thu, Jul 30, 2020 at 05:22:10PM +0100, Catalin Marinas wrote: > > On Mon, Jul 27, 2020 at 07:29:31PM +0300, Mike Rapoport wrote: > > > +static int secretmem_mmap(struct file *file, struct vm_area_struct *vma) > > > +{ > > > + struct secretmem_ctx *ctx = file->private_data; > > > + unsigned long mode = ctx->mode; > > > + unsigned long len = vma->vm_end - vma->vm_start; > > > + > > > + if (!mode) > > > + return -EINVAL; > > > + > > > + if ((vma->vm_flags & (VM_SHARED | VM_MAYSHARE)) == 0) > > > + return -EINVAL; > > > + > > > + if (mlock_future_check(vma->vm_mm, vma->vm_flags | VM_LOCKED, len)) > > > + return -EAGAIN; > > > + > > > + switch (mode) { > > > + case SECRETMEM_UNCACHED: > > > + vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot); > > > + fallthrough; > > > + case SECRETMEM_EXCLUSIVE: > > > + vma->vm_ops = &secretmem_vm_ops; > > > + break; > > > + default: > > > + return -EINVAL; > > > + } > > > + > > > + vma->vm_flags |= VM_LOCKED; > > > + > > > + return 0; > > > +} > > > > I think the uncached mapping is not the right thing for arm/arm64. First > > of all, pgprot_noncached() gives us Strongly Ordered (Device memory) > > semantics together with not allowing unaligned accesses. I suspect the > > semantics are different on x86. > > > The second, more serious problem, is that I can't find any place where > > the caches are flushed for the page mapped on fault. When a page is > > allocated, assuming GFP_ZERO, only the caches are guaranteed to be > > zeroed. Exposing this subsequently to user space as uncached would allow > > the user to read stale data prior to zeroing. The arm64 > > set_direct_map_default_noflush() doesn't do any cache maintenance. > > It's also worth noting that in a virtual machine this is liable to be > either broken (with a potential loss of coherency if the host has a > cacheable alias as existing KVM hosts have), or pointless (if the host > uses S2FWB to upgrade Stage-1 attribues to cacheable as existing KVM > hosts also have). > > I think that trying to avoid the data caches creates many more problems > than it solves, and I don't think there's a strong justification for > trying to support that on arm64 to begin with, so I'd rather entirely > opt-out on supporting SECRETMEM_UNCACHED. Good point, I forgot the virtualisation aspect. So unless there is a hypervisor API to unmap it from the host memory, the uncached option isn't of much use on arm64. -- Catalin From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-5.5 required=3.0 tests=BAYES_00,DKIMWL_WL_HIGH, DKIM_SIGNED,DKIM_VALID,HEADER_FROM_DIFFERENT_DOMAINS,MAILING_LIST_MULTI, SPF_HELO_NONE,SPF_PASS,USER_AGENT_SANE_1 autolearn=no autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 52AD0C433E0 for ; Fri, 31 Jul 2020 16:23:03 +0000 (UTC) Received: from merlin.infradead.org (merlin.infradead.org [205.233.59.134]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPS id 1A1882177B for ; Fri, 31 Jul 2020 16:23:03 +0000 (UTC) Authentication-Results: mail.kernel.org; dkim=pass (2048-bit key) header.d=lists.infradead.org header.i=@lists.infradead.org header.b="WIL5DEuU" DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org 1A1882177B Authentication-Results: mail.kernel.org; dmarc=none (p=none dis=none) header.from=arm.com Authentication-Results: mail.kernel.org; spf=none smtp.mailfrom=linux-riscv-bounces+linux-riscv=archiver.kernel.org@lists.infradead.org DKIM-Signature: v=1; a=rsa-sha256; q=dns/txt; c=relaxed/relaxed; d=lists.infradead.org; s=merlin.20170209; h=Sender:Content-Transfer-Encoding: Content-Type:Cc:List-Subscribe:List-Help:List-Post:List-Archive: List-Unsubscribe:List-Id:In-Reply-To:MIME-Version:References:Message-ID: Subject:To:From:Date:Reply-To:Content-ID:Content-Description:Resent-Date: Resent-From:Resent-Sender:Resent-To:Resent-Cc:Resent-Message-ID:List-Owner; bh=DetK4v3A3PNL00xy6j6hZG/8OV8jRD66q8idZJQGxGo=; b=WIL5DEuU18cchIEil/YW0b2B5 YfVvapb/Iij7UHUBj7SwyCOslIWH6Bf0+d1DBJsJEGlr46f/5kvfTpl0C2+Ao+e/FtPrHm7uEFY06 S6sMy+g2mI9mGuG9XpjOj/GjedF1zGEhxshLUs0k2fKZDUoT+cGxYDle+ar1mdRjYcJ0zVIm91Qxn RZK7Oqq+oFSzMGZf0OJX2geuZCbRNAIhJ9OWMYvInuMZi9QufLIr4b5Wf9N0IebRcAEmPXTo0Ub5N CsMy0L5Moky2WkGzpIbL7fGMbEFnhTGV/uDaDqR7HZNWkQWl8tGTS7zwyWqd4/RfDeP6fvMv6xmhs Y9/dpQ3Vw==; Received: from localhost ([::1] helo=merlin.infradead.org) by merlin.infradead.org with esmtp (Exim 4.92.3 #3 (Red Hat Linux)) id 1k1Xo0-0000iE-Nc; Fri, 31 Jul 2020 16:22:48 +0000 Received: from mail.kernel.org ([198.145.29.99]) by merlin.infradead.org with esmtps (Exim 4.92.3 #3 (Red Hat Linux)) id 1k1Xnv-0000gB-18; Fri, 31 Jul 2020 16:22:44 +0000 Received: from gaia (unknown [95.146.230.158]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPSA id 1B0052177B; Fri, 31 Jul 2020 16:22:36 +0000 (UTC) Date: Fri, 31 Jul 2020 17:22:34 +0100 From: Catalin Marinas To: Mark Rutland Subject: Re: [PATCH v2 3/7] mm: introduce memfd_secret system call to create "secret" memory areas Message-ID: <20200731162234.GF29569@gaia> References: <20200727162935.31714-1-rppt@kernel.org> <20200727162935.31714-4-rppt@kernel.org> <20200730162209.GB3128@gaia> <20200731142905.GA67415@C02TD0UTHF1T.local> MIME-Version: 1.0 Content-Disposition: inline In-Reply-To: <20200731142905.GA67415@C02TD0UTHF1T.local> User-Agent: Mutt/1.10.1 (2018-07-13) X-CRM114-Version: 20100106-BlameMichelson ( TRE 0.8.0 (BSD) ) MR-646709E3 X-CRM114-CacheID: sfid-20200731_122243_223219_437EBA74 X-CRM114-Status: GOOD ( 29.51 ) X-BeenThere: linux-riscv@lists.infradead.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: Peter Zijlstra , Dave Hansen , linux-mm@kvack.org, "H. Peter Anvin" , Christopher Lameter , Idan Yaniv , Thomas Gleixner , Elena Reshetova , linux-arch@vger.kernel.org, Tycho Andersen , linux-nvdimm@lists.01.org, Will Deacon , x86@kernel.org, Matthew Wilcox , Mike Rapoport , Ingo Molnar , Michael Kerrisk , Arnd Bergmann , James Bottomley , Borislav Petkov , Alexander Viro , Andy Lutomirski , Paul Walmsley , "Kirill A. Shutemov" , Dan Williams , linux-arm-kernel@lists.infradead.org, linux-api@vger.kernel.org, linux-kernel@vger.kernel.org, linux-riscv@lists.infradead.org, Palmer Dabbelt , linux-fsdevel@vger.kernel.org, Andrew Morton , Mike Rapoport Content-Type: text/plain; charset="us-ascii" Content-Transfer-Encoding: 7bit Sender: "linux-riscv" Errors-To: linux-riscv-bounces+linux-riscv=archiver.kernel.org@lists.infradead.org On Fri, Jul 31, 2020 at 03:29:05PM +0100, Mark Rutland wrote: > On Thu, Jul 30, 2020 at 05:22:10PM +0100, Catalin Marinas wrote: > > On Mon, Jul 27, 2020 at 07:29:31PM +0300, Mike Rapoport wrote: > > > +static int secretmem_mmap(struct file *file, struct vm_area_struct *vma) > > > +{ > > > + struct secretmem_ctx *ctx = file->private_data; > > > + unsigned long mode = ctx->mode; > > > + unsigned long len = vma->vm_end - vma->vm_start; > > > + > > > + if (!mode) > > > + return -EINVAL; > > > + > > > + if ((vma->vm_flags & (VM_SHARED | VM_MAYSHARE)) == 0) > > > + return -EINVAL; > > > + > > > + if (mlock_future_check(vma->vm_mm, vma->vm_flags | VM_LOCKED, len)) > > > + return -EAGAIN; > > > + > > > + switch (mode) { > > > + case SECRETMEM_UNCACHED: > > > + vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot); > > > + fallthrough; > > > + case SECRETMEM_EXCLUSIVE: > > > + vma->vm_ops = &secretmem_vm_ops; > > > + break; > > > + default: > > > + return -EINVAL; > > > + } > > > + > > > + vma->vm_flags |= VM_LOCKED; > > > + > > > + return 0; > > > +} > > > > I think the uncached mapping is not the right thing for arm/arm64. First > > of all, pgprot_noncached() gives us Strongly Ordered (Device memory) > > semantics together with not allowing unaligned accesses. I suspect the > > semantics are different on x86. > > > The second, more serious problem, is that I can't find any place where > > the caches are flushed for the page mapped on fault. When a page is > > allocated, assuming GFP_ZERO, only the caches are guaranteed to be > > zeroed. Exposing this subsequently to user space as uncached would allow > > the user to read stale data prior to zeroing. The arm64 > > set_direct_map_default_noflush() doesn't do any cache maintenance. > > It's also worth noting that in a virtual machine this is liable to be > either broken (with a potential loss of coherency if the host has a > cacheable alias as existing KVM hosts have), or pointless (if the host > uses S2FWB to upgrade Stage-1 attribues to cacheable as existing KVM > hosts also have). > > I think that trying to avoid the data caches creates many more problems > than it solves, and I don't think there's a strong justification for > trying to support that on arm64 to begin with, so I'd rather entirely > opt-out on supporting SECRETMEM_UNCACHED. Good point, I forgot the virtualisation aspect. So unless there is a hypervisor API to unmap it from the host memory, the uncached option isn't of much use on arm64. -- Catalin _______________________________________________ linux-riscv mailing list linux-riscv@lists.infradead.org http://lists.infradead.org/mailman/listinfo/linux-riscv From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-5.5 required=3.0 tests=BAYES_00,DKIMWL_WL_HIGH, DKIM_SIGNED,DKIM_VALID,HEADER_FROM_DIFFERENT_DOMAINS,MAILING_LIST_MULTI, SPF_HELO_NONE,SPF_PASS,URIBL_BLOCKED,USER_AGENT_SANE_1 autolearn=no autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 932E5C433DF for ; Fri, 31 Jul 2020 16:24:33 +0000 (UTC) Received: from merlin.infradead.org (merlin.infradead.org [205.233.59.134]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPS id 4542D2177B for ; Fri, 31 Jul 2020 16:24:33 +0000 (UTC) Authentication-Results: mail.kernel.org; dkim=pass (2048-bit key) header.d=lists.infradead.org header.i=@lists.infradead.org header.b="lLoRtSlO" DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org 4542D2177B Authentication-Results: mail.kernel.org; dmarc=none (p=none dis=none) header.from=arm.com Authentication-Results: mail.kernel.org; spf=none smtp.mailfrom=linux-arm-kernel-bounces+linux-arm-kernel=archiver.kernel.org@lists.infradead.org DKIM-Signature: v=1; a=rsa-sha256; q=dns/txt; c=relaxed/relaxed; d=lists.infradead.org; s=merlin.20170209; h=Sender:Content-Transfer-Encoding: Content-Type:Cc:List-Subscribe:List-Help:List-Post:List-Archive: List-Unsubscribe:List-Id:In-Reply-To:MIME-Version:References:Message-ID: Subject:To:From:Date:Reply-To:Content-ID:Content-Description:Resent-Date: Resent-From:Resent-Sender:Resent-To:Resent-Cc:Resent-Message-ID:List-Owner; bh=j58A3Ho577UNhS7UcsD/+BiCn+vZvNmK+H+amFtdrOo=; b=lLoRtSlOdJ+ZbnTW3BNEiVgFu 3nK5uHL+5mhNFYMKd/ZqDltG7mcrSJukrUXjpL1g1XLs4hTAV1bt18fS95HeQ+dLqPxhHa/iFJZBq SL58Y+nT54v83O5S4/KfzbeGvb/L8NctUgNIEW6HlondTyKZbXqgZfZbTq79hQN4hJdpOzzAfdiMx CrIQi1r89CWNgMTi40aivdWPcbw2Bn8L4dNFHP5Ho8B5KyCr+JZ7Fbk5cPNn8yipV34AygsKT1bu1 qeA0TbT3qT2phdR0ziJ4AtAmoG7Sl2gVhZq5xBxL02C1WdjzDtTM6OdPL78pE3OzQUFV5ktOw6s7k mEfraw4XQ==; Received: from localhost ([::1] helo=merlin.infradead.org) by merlin.infradead.org with esmtp (Exim 4.92.3 #3 (Red Hat Linux)) id 1k1Xny-0000hV-VC; Fri, 31 Jul 2020 16:22:47 +0000 Received: from mail.kernel.org ([198.145.29.99]) by merlin.infradead.org with esmtps (Exim 4.92.3 #3 (Red Hat Linux)) id 1k1Xnv-0000gB-18; Fri, 31 Jul 2020 16:22:44 +0000 Received: from gaia (unknown [95.146.230.158]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPSA id 1B0052177B; Fri, 31 Jul 2020 16:22:36 +0000 (UTC) Date: Fri, 31 Jul 2020 17:22:34 +0100 From: Catalin Marinas To: Mark Rutland Subject: Re: [PATCH v2 3/7] mm: introduce memfd_secret system call to create "secret" memory areas Message-ID: <20200731162234.GF29569@gaia> References: <20200727162935.31714-1-rppt@kernel.org> <20200727162935.31714-4-rppt@kernel.org> <20200730162209.GB3128@gaia> <20200731142905.GA67415@C02TD0UTHF1T.local> MIME-Version: 1.0 Content-Disposition: inline In-Reply-To: <20200731142905.GA67415@C02TD0UTHF1T.local> User-Agent: Mutt/1.10.1 (2018-07-13) X-CRM114-Version: 20100106-BlameMichelson ( TRE 0.8.0 (BSD) ) MR-646709E3 X-CRM114-CacheID: sfid-20200731_122243_223219_437EBA74 X-CRM114-Status: GOOD ( 29.51 ) X-BeenThere: linux-arm-kernel@lists.infradead.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: Peter Zijlstra , Dave Hansen , linux-mm@kvack.org, "H. Peter Anvin" , Christopher Lameter , Idan Yaniv , Thomas Gleixner , Elena Reshetova , linux-arch@vger.kernel.org, Tycho Andersen , linux-nvdimm@lists.01.org, Will Deacon , x86@kernel.org, Matthew Wilcox , Mike Rapoport , Ingo Molnar , Michael Kerrisk , Arnd Bergmann , James Bottomley , Borislav Petkov , Alexander Viro , Andy Lutomirski , Paul Walmsley , "Kirill A. Shutemov" , Dan Williams , linux-arm-kernel@lists.infradead.org, linux-api@vger.kernel.org, linux-kernel@vger.kernel.org, linux-riscv@lists.infradead.org, Palmer Dabbelt , linux-fsdevel@vger.kernel.org, Andrew Morton , Mike Rapoport Content-Type: text/plain; charset="us-ascii" Content-Transfer-Encoding: 7bit Sender: "linux-arm-kernel" Errors-To: linux-arm-kernel-bounces+linux-arm-kernel=archiver.kernel.org@lists.infradead.org On Fri, Jul 31, 2020 at 03:29:05PM +0100, Mark Rutland wrote: > On Thu, Jul 30, 2020 at 05:22:10PM +0100, Catalin Marinas wrote: > > On Mon, Jul 27, 2020 at 07:29:31PM +0300, Mike Rapoport wrote: > > > +static int secretmem_mmap(struct file *file, struct vm_area_struct *vma) > > > +{ > > > + struct secretmem_ctx *ctx = file->private_data; > > > + unsigned long mode = ctx->mode; > > > + unsigned long len = vma->vm_end - vma->vm_start; > > > + > > > + if (!mode) > > > + return -EINVAL; > > > + > > > + if ((vma->vm_flags & (VM_SHARED | VM_MAYSHARE)) == 0) > > > + return -EINVAL; > > > + > > > + if (mlock_future_check(vma->vm_mm, vma->vm_flags | VM_LOCKED, len)) > > > + return -EAGAIN; > > > + > > > + switch (mode) { > > > + case SECRETMEM_UNCACHED: > > > + vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot); > > > + fallthrough; > > > + case SECRETMEM_EXCLUSIVE: > > > + vma->vm_ops = &secretmem_vm_ops; > > > + break; > > > + default: > > > + return -EINVAL; > > > + } > > > + > > > + vma->vm_flags |= VM_LOCKED; > > > + > > > + return 0; > > > +} > > > > I think the uncached mapping is not the right thing for arm/arm64. First > > of all, pgprot_noncached() gives us Strongly Ordered (Device memory) > > semantics together with not allowing unaligned accesses. I suspect the > > semantics are different on x86. > > > The second, more serious problem, is that I can't find any place where > > the caches are flushed for the page mapped on fault. When a page is > > allocated, assuming GFP_ZERO, only the caches are guaranteed to be > > zeroed. Exposing this subsequently to user space as uncached would allow > > the user to read stale data prior to zeroing. The arm64 > > set_direct_map_default_noflush() doesn't do any cache maintenance. > > It's also worth noting that in a virtual machine this is liable to be > either broken (with a potential loss of coherency if the host has a > cacheable alias as existing KVM hosts have), or pointless (if the host > uses S2FWB to upgrade Stage-1 attribues to cacheable as existing KVM > hosts also have). > > I think that trying to avoid the data caches creates many more problems > than it solves, and I don't think there's a strong justification for > trying to support that on arm64 to begin with, so I'd rather entirely > opt-out on supporting SECRETMEM_UNCACHED. Good point, I forgot the virtualisation aspect. So unless there is a hypervisor API to unmap it from the host memory, the uncached option isn't of much use on arm64. -- Catalin _______________________________________________ linux-arm-kernel mailing list linux-arm-kernel@lists.infradead.org http://lists.infradead.org/mailman/listinfo/linux-arm-kernel