From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id 1C2B9ECAAD8 for ; Fri, 16 Sep 2022 17:46:24 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id 87DC28D0002; Fri, 16 Sep 2022 13:46:24 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id 806F88D0001; Fri, 16 Sep 2022 13:46:24 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id 6801E8D0002; Fri, 16 Sep 2022 13:46:24 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0010.hostedemail.com [216.40.44.10]) by kanga.kvack.org (Postfix) with ESMTP id 53AC78D0001 for ; Fri, 16 Sep 2022 13:46:24 -0400 (EDT) Received: from smtpin27.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay08.hostedemail.com (Postfix) with ESMTP id 2128B141BA2 for ; Fri, 16 Sep 2022 17:46:24 +0000 (UTC) X-FDA: 79918677888.27.8D39EB5 Received: from mail-pj1-f49.google.com (mail-pj1-f49.google.com [209.85.216.49]) by imf10.hostedemail.com (Postfix) with ESMTP id CEB7AC0097 for ; Fri, 16 Sep 2022 17:46:23 +0000 (UTC) Received: by mail-pj1-f49.google.com with SMTP id q9-20020a17090a178900b0020265d92ae3so376625pja.5 for ; Fri, 16 Sep 2022 10:46:23 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20210112; h=cc:to:subject:message-id:date:from:in-reply-to:references :mime-version:from:to:cc:subject:date; bh=bGlPiAdoAwCN/kk/sgLTt0tFCqAu0UJel6VuA5nrRzo=; b=kj+jNlpEBI1gx5UYRCwGlU0K4V/x/gxhqMZH2GL4znu+wsn4o8owU9yUOQi3EWfGAo 5388vFVL2ZB/KlkberyIdDGa2Ba0/dzaD8lIjeO08fZJ4RknTMi9duk4Siwj4ODXKyw0 dDnuxM6XvRptf3VVhf2TH8AmWjNZSuy05+5DjEq0DLAZsRlQhtZGIxX7FH6ChQu4jlH+ 3rWJ51kMHiJ/CmD6twKSuMOJ0wqNUftzvnxol4F/84oawrrhQBJtw8MhD+0hB5+aeTbf t/7giJUExIxlOXza4ArgVfIb4MvmH4Q776HlqQVnmWtCNBjM+BzYaRx/SYTjoer5UeyH VOYQ== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=cc:to:subject:message-id:date:from:in-reply-to:references :mime-version:x-gm-message-state:from:to:cc:subject:date; bh=bGlPiAdoAwCN/kk/sgLTt0tFCqAu0UJel6VuA5nrRzo=; b=RQQiQLA75OCG1/9Uv85GOiOAl97t0lGw34BaHFCh4c4alM2BYc/sjZ2/wk1sQUwAye iI6qIAxwA8p8T/Tjli0x/cmQwLDQ7b/qeEVfkGfcUCd6KQLOo/HPwPt5GxKOxHqCjpf/ 04xNgpPZ7qoZtAA8u6W+fytKz97lUndvG7SlfHBYPabImVp2cNErW1AaLp3DNQ/qAyjy VxRJ+9jUQ53jVy7C3U8iCgGsa6QL6dVA/2mqjLt+2rzQqAJJ+XaERDUNtyvVBSkIldcf uEvDQbk7z0kYXiu0muBvqk0sFvnjVJwRXSrvkmWqunSuN6SIE61zt9Qcof2T9arSnc4k 0tHQ== X-Gm-Message-State: ACrzQf1DBBGALyvyu4FvbIYS2j6b5jaZHNE59GtCdmp6RA35Ev42rRFN BK7rc/eLhrZaO1OrJtL5IQ4amaU/96OozI1Yms0= X-Google-Smtp-Source: AMsMyM4LUFoR1XGTgnimYOGHfm9KyHt1RuU0QXZaJq3t9UVoRP8j6DITjCtGkO9biE+VIJ1u0nOpi2SkMtELa6coTio= X-Received: by 2002:a17:90b:1b4c:b0:202:c05f:6e7b with SMTP id nv12-20020a17090b1b4c00b00202c05f6e7bmr6729830pjb.200.1663350382684; Fri, 16 Sep 2022 10:46:22 -0700 (PDT) MIME-Version: 1.0 References: <20220907144521.3115321-1-zokeefe@google.com> <20220907144521.3115321-2-zokeefe@google.com> In-Reply-To: <20220907144521.3115321-2-zokeefe@google.com> From: Yang Shi Date: Fri, 16 Sep 2022 10:46:09 -0700 Message-ID: Subject: Re: [PATCH mm-unstable v3 01/10] mm/shmem: add flag to enforce shmem THP in hugepage_vma_check() To: "Zach O'Keefe" Cc: linux-mm@kvack.org, Andrew Morton , linux-api@vger.kernel.org, Axel Rasmussen , James Houghton , Hugh Dickins , Miaohe Lin , David Hildenbrand , David Rientjes , Matthew Wilcox , Pasha Tatashin , Peter Xu , Rongwei Wang , SeongJae Park , Song Liu , Vlastimil Babka , Chris Kennelly , "Kirill A. Shutemov" , Minchan Kim , Patrick Xia Content-Type: text/plain; charset="UTF-8" ARC-Seal: i=1; s=arc-20220608; d=hostedemail.com; t=1663350383; a=rsa-sha256; cv=none; b=yEJBgts6xY6lpeFoPr4us5vVMXd0Frokfp9KVLzu8REf9hTggmcmvQRAJu2gREe67Qw2Uq JmzuqrIhDNVjgATvREJfgpHdJHHnxZ8ICl7qAMuZAy/AyU3Tpyg4GiNvWYSoq359KZLXFS hr6rTZdzQx3wiYxDo7Q2uswDk309wuA= ARC-Authentication-Results: i=1; imf10.hostedemail.com; dkim=pass header.d=gmail.com header.s=20210112 header.b=kj+jNlpE; dmarc=pass (policy=none) header.from=gmail.com; spf=pass (imf10.hostedemail.com: domain of shy828301@gmail.com designates 209.85.216.49 as permitted sender) smtp.mailfrom=shy828301@gmail.com ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=hostedemail.com; s=arc-20220608; t=1663350383; h=from:from:sender:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:mime-version: content-type:content-type:content-transfer-encoding: in-reply-to:in-reply-to:references:references:dkim-signature; bh=bGlPiAdoAwCN/kk/sgLTt0tFCqAu0UJel6VuA5nrRzo=; b=Ve2WjZwdbRej1dRSpv9uCvcVTLPSjjLAex/Ngufl6y7/KyeuMzZoRVHs6VN8p81b38HKfL FE5WMx1yPYLM0ZWD9Mq3CesHwkyILm2Spy/XS7FTVnl+wevuy4ytTXTGY2j5X2EMn4kqIb sdIfDvzyYTVgOUo00P7CLgEUhIX2JfM= X-Rspam-User: X-Rspamd-Server: rspam11 X-Rspamd-Queue-Id: CEB7AC0097 Authentication-Results: imf10.hostedemail.com; dkim=pass header.d=gmail.com header.s=20210112 header.b=kj+jNlpE; dmarc=pass (policy=none) header.from=gmail.com; spf=pass (imf10.hostedemail.com: domain of shy828301@gmail.com designates 209.85.216.49 as permitted sender) smtp.mailfrom=shy828301@gmail.com X-Stat-Signature: ge1t43ye1wowpkfdyh78c4pqh13y3qbb X-HE-Tag: 1663350383-825579 X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: On Wed, Sep 7, 2022 at 7:45 AM Zach O'Keefe wrote: > > Extend 'mm/thp: add flag to enforce sysfs THP in > hugepage_vma_check()' to shmem, allowing callers to ignore > /sys/kernel/transparent_hugepage/shmem_enabled and tmpfs huge= mount. > > This is intended to be used by MADV_COLLAPSE, and the rationale is > analogous to the anon/file case: MADV_COLLAPSE is not coupled to > directives that advise the kernel's decisions on when THPs should be > considered eligible. shmem/tmpfs always claims large folio support, > regardless of sysfs or mount options. > > Signed-off-by: Zach O'Keefe Reviewed-by: Yang Shi A nit below... > --- > include/linux/shmem_fs.h | 10 ++++++---- > mm/huge_memory.c | 2 +- > mm/shmem.c | 18 +++++++++--------- > 3 files changed, 16 insertions(+), 14 deletions(-) > > diff --git a/include/linux/shmem_fs.h b/include/linux/shmem_fs.h > index f24071e3c826..d500ea967dc7 100644 > --- a/include/linux/shmem_fs.h > +++ b/include/linux/shmem_fs.h > @@ -92,11 +92,13 @@ extern struct page *shmem_read_mapping_page_gfp(struct address_space *mapping, > extern void shmem_truncate_range(struct inode *inode, loff_t start, loff_t end); > int shmem_unuse(unsigned int type); > > -extern bool shmem_is_huge(struct vm_area_struct *vma, > - struct inode *inode, pgoff_t index); > -static inline bool shmem_huge_enabled(struct vm_area_struct *vma) > +extern bool shmem_is_huge(struct vm_area_struct *vma, struct inode *inode, > + pgoff_t index, bool shmem_huge_force); > +static inline bool shmem_huge_enabled(struct vm_area_struct *vma, > + bool shmem_huge_force) > { > - return shmem_is_huge(vma, file_inode(vma->vm_file), vma->vm_pgoff); > + return shmem_is_huge(vma, file_inode(vma->vm_file), vma->vm_pgoff, > + shmem_huge_force); > } > extern unsigned long shmem_swap_usage(struct vm_area_struct *vma); > extern unsigned long shmem_partial_swap_usage(struct address_space *mapping, > diff --git a/mm/huge_memory.c b/mm/huge_memory.c > index 7fa74b9749a6..53d170dac332 100644 > --- a/mm/huge_memory.c > +++ b/mm/huge_memory.c > @@ -119,7 +119,7 @@ bool hugepage_vma_check(struct vm_area_struct *vma, unsigned long vm_flags, > * own flags. > */ > if (!in_pf && shmem_file(vma->vm_file)) > - return shmem_huge_enabled(vma); > + return shmem_huge_enabled(vma, !enforce_sysfs); > > /* Enforce sysfs THP requirements as necessary */ > if (enforce_sysfs && > diff --git a/mm/shmem.c b/mm/shmem.c > index 99b7341bd0bf..47c42c566fd1 100644 > --- a/mm/shmem.c > +++ b/mm/shmem.c > @@ -461,20 +461,20 @@ static bool shmem_confirm_swap(struct address_space *mapping, > > static int shmem_huge __read_mostly = SHMEM_HUGE_NEVER; > > -bool shmem_is_huge(struct vm_area_struct *vma, > - struct inode *inode, pgoff_t index) > +bool shmem_is_huge(struct vm_area_struct *vma, struct inode *inode, > + pgoff_t index, bool shmem_huge_force) > { > loff_t i_size; > > if (!S_ISREG(inode->i_mode)) > return false; > - if (shmem_huge == SHMEM_HUGE_DENY) > - return false; > if (vma && ((vma->vm_flags & VM_NOHUGEPAGE) || > test_bit(MMF_DISABLE_THP, &vma->vm_mm->flags))) > return false; > - if (shmem_huge == SHMEM_HUGE_FORCE) > + if (shmem_huge == SHMEM_HUGE_FORCE || shmem_huge_force) shmem_huge_force means ignore all sysfs and mount options, so it seems better to have it test explicitly IMHO, like: if (shmem_huge_force) return true; if (shmem_huge == SHMEM_HUGE_FORCE) return true; > return true; > + if (shmem_huge == SHMEM_HUGE_DENY) > + return false; > > switch (SHMEM_SB(inode->i_sb)->huge) { > case SHMEM_HUGE_ALWAYS: > @@ -669,8 +669,8 @@ static long shmem_unused_huge_count(struct super_block *sb, > > #define shmem_huge SHMEM_HUGE_DENY > > -bool shmem_is_huge(struct vm_area_struct *vma, > - struct inode *inode, pgoff_t index) > +bool shmem_is_huge(struct vm_area_struct *vma, struct inode *inode, > + pgoff_t index, bool shmem_huge_force) > { > return false; > } > @@ -1056,7 +1056,7 @@ static int shmem_getattr(struct user_namespace *mnt_userns, > STATX_ATTR_NODUMP); > generic_fillattr(&init_user_ns, inode, stat); > > - if (shmem_is_huge(NULL, inode, 0)) > + if (shmem_is_huge(NULL, inode, 0, false)) > stat->blksize = HPAGE_PMD_SIZE; > > if (request_mask & STATX_BTIME) { > @@ -1888,7 +1888,7 @@ static int shmem_get_folio_gfp(struct inode *inode, pgoff_t index, > return 0; > } > > - if (!shmem_is_huge(vma, inode, index)) > + if (!shmem_is_huge(vma, inode, index, false)) > goto alloc_nohuge; > > huge_gfp = vma_thp_gfp_mask(vma); > -- > 2.37.2.789.g6183377224-goog >