linux-mm.kvack.org archive mirror
 help / color / mirror / Atom feed
From: David Hildenbrand <david@redhat.com>
To: Zi Yan <ziy@nvidia.com>, linux-mm@kvack.org
Cc: Matthew Wilcox <willy@infradead.org>,
	"Kirill A . Shutemov" <kirill.shutemov@linux.intel.com>,
	Roman Gushchin <guro@fb.com>,
	Andrew Morton <akpm@linux-foundation.org>,
	Yang Shi <shy828301@gmail.com>, Michal Hocko <mhocko@suse.com>,
	John Hubbard <jhubbard@nvidia.com>,
	Ralph Campbell <rcampbell@nvidia.com>,
	David Nellans <dnellans@nvidia.com>,
	Jason Gunthorpe <jgg@ziepe.ca>,
	David Rientjes <rientjes@google.com>,
	Vlastimil Babka <vbabka@suse.cz>,
	Mike Kravetz <mike.kravetz@oracle.com>,
	Song Liu <songliubraving@fb.com>
Subject: Re: [RFC PATCH v3 00/49] 1GB PUD THP support on x86_64
Date: Thu, 25 Feb 2021 12:02:29 +0100	[thread overview]
Message-ID: <c0caebd1-9da6-9147-b30e-cc8ae1121228@redhat.com> (raw)
In-Reply-To: <20210224223536.803765-1-zi.yan@sent.com>

On 24.02.21 23:35, Zi Yan wrote:
> From: Zi Yan <ziy@nvidia.com>
> 
> Hi all,
> 
> I have rebased my 1GB PUD THP support patches on v5.11-mmotm-2021-02-18-18-29
> and the code is available at
> https://github.com/x-y-z/linux-1gb-thp/tree/1gb_thp_v5.11-mmotm-2021-02-18-18-29
> if you want to give it a try. The actual 49 patches are not sent out with this
> cover letter. :)
> 
> Instead of asking for code review, I would like to discuss on the concerns I got
> from previous RFCs. I think there are two major ones:
> 
> 1. 1GB page allocation. Current implementation allocates 1GB pages from CMA
>     regions that are reserved at boot time like hugetlbfs. The concerns on
>     using CMA is that an educated guess is needed to avoid depleting kernel
>     memory in case CMA regions are set too large. Recently David Rientjes
>     proposes to use process_madvise() for hugepage collapse, which is an
>     alternative [1] but might not work for 1GB pages, since there is no way of

I see two core ideas of THP:

1) Transparent to the user: you get speedup without really caring 
*except* having to enable/disable the optimization sometimes manually 
(i.e., MADV_HUGEPAGE) -  because in corner cases (e.g., userfaultfd), 
it's not completely transparent and might have performance impacts. 
mprotect(), mmap(MAP_FIXED), mremap() work as expected.

2) Transparent to other subsystems of the kernel: the page size of the 
mapping is in base pages - we can split anytime on demand in case we 
cannot handle THP. In addition, no special requirements: no CMA, no 
movability restrictions, no swappability restrictions, ... most stuff 
works transparently by splitting.

Your current approach messes with 2). Your proposal here messes with 1).

Any kind of explicit placement by the user can silently get reverted any 
time. So process_madvise() would really only be useful in cases where a 
temporary split might get reverted later on by the os automatically - 
like we have for 2MB THP right now.

So process_madvise() is less likely to help if the system won't try 
collapsing automatically (more below).

>     _allocating_ a 1GB page to which collapse pages. I proposed a similar
>     approach at LSF/MM 2019, generating physically contiguous memory after pages
>     are allocated [2], which is usable for 1GB THPs. This approach does in-place
>     huge page promotion thus does not require page allocation.

I like the idea of forming a 1GB THP at a location where already 
consecutive pages allow for it. It can be applied generically - and both 
1) and 2) keep working as expected. Anytime there was a split, we can 
retry forming a THP later.

However, I don't follow how this is actually really feasible in big 
scale. You could only ever collapse into a 1GB THP if you happen to have 
1GB consecutive 2MB THP / 4k already. Sounds to me like this happens 
when the stars align.

-- 
Thanks,

David / dhildenb



  reply	other threads:[~2021-02-25 11:02 UTC|newest]

Thread overview: 15+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2021-02-24 22:35 [RFC PATCH v3 00/49] 1GB PUD THP support on x86_64 Zi Yan
2021-02-25 11:02 ` David Hildenbrand [this message]
2021-02-25 22:13   ` Zi Yan
2021-03-02  8:55     ` David Hildenbrand
2021-03-03 23:42       ` Zi Yan
2021-03-04  9:26         ` David Hildenbrand
2021-03-02  1:59 ` Roman Gushchin
2021-03-04 16:26   ` Zi Yan
2021-03-04 16:45     ` Roman Gushchin
2021-03-30 17:24       ` Zi Yan
2021-03-30 18:02         ` Roman Gushchin
2021-03-31  2:04           ` Zi Yan
2021-03-31  3:09           ` Matthew Wilcox
2021-03-31  3:32             ` Roman Gushchin
2021-03-31 14:48               ` Zi Yan

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=c0caebd1-9da6-9147-b30e-cc8ae1121228@redhat.com \
    --to=david@redhat.com \
    --cc=akpm@linux-foundation.org \
    --cc=dnellans@nvidia.com \
    --cc=guro@fb.com \
    --cc=jgg@ziepe.ca \
    --cc=jhubbard@nvidia.com \
    --cc=kirill.shutemov@linux.intel.com \
    --cc=linux-mm@kvack.org \
    --cc=mhocko@suse.com \
    --cc=mike.kravetz@oracle.com \
    --cc=rcampbell@nvidia.com \
    --cc=rientjes@google.com \
    --cc=shy828301@gmail.com \
    --cc=songliubraving@fb.com \
    --cc=vbabka@suse.cz \
    --cc=willy@infradead.org \
    --cc=ziy@nvidia.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).