linux-mm.kvack.org archive mirror
 help / color / mirror / Atom feed
From: Jiaqi Yan <jiaqiyan@google.com>
To: Yang Shi <shy828301@gmail.com>
Cc: kirill.shutemov@linux.intel.com, kirill@shutemov.name,
	 tongtiangen@huawei.com, tony.luck@intel.com,
	akpm@linux-foundation.org,  naoya.horiguchi@nec.com,
	linmiaohe@huawei.com, linux-mm@kvack.org,  osalvador@suse.de,
	wangkefeng.wang@huawei.com
Subject: Re: [PATCH v10 1/3] mm/khugepaged: recover from poisoned anonymous memory
Date: Mon, 27 Mar 2023 13:46:33 -0700	[thread overview]
Message-ID: <CACw3F50JE+Luo_wQEhScp+rBXRCk0SrSGUrfscWa7iKNDDUaHA@mail.gmail.com> (raw)
In-Reply-To: <CACw3F50oBji3YG9wWNynATgH-BuFzk2BSa4FEvrkzPboLbq3xA@mail.gmail.com>

On Fri, Mar 24, 2023 at 3:31 PM Jiaqi Yan <jiaqiyan@google.com> wrote:
>
> On Fri, Mar 24, 2023 at 1:11 PM Yang Shi <shy828301@gmail.com> wrote:
> >
> > On Fri, Mar 24, 2023 at 8:34 AM Jiaqi Yan <jiaqiyan@google.com> wrote:
> > >
> > > On Thu, Mar 23, 2023 at 2:38 PM Yang Shi <shy828301@gmail.com> wrote:
> > > >
> > > > On Sat, Mar 4, 2023 at 10:51 PM Jiaqi Yan <jiaqiyan@google.com> wrote:
> > > > >
> > > > > Make __collapse_huge_page_copy return whether copying anonymous pages
> > > > > succeeded, and make collapse_huge_page handle the return status.
> > > > >
> > > > > Break existing PTE scan loop into two for-loops. The first loop copies
> > > > > source pages into target huge page, and can fail gracefully when running
> > > > > into memory errors in source pages. If copying all pages succeeds, the
> > > > > second loop releases and clears up these normal pages. Otherwise, the
> > > > > second loop rolls back the page table and page states by:
> > > > > - re-establishing the original PTEs-to-PMD connection.
> > > > > - releasing source pages back to their LRU list.
> > > > >
> > > > > Tested manually:
> > > > > 0. Enable khugepaged on system under test.
> > > > > 1. Start a two-thread application. Each thread allocates a chunk of
> > > > >    non-huge anonymous memory buffer.
> > > > > 2. Pick 4 random buffer locations (2 in each thread) and inject
> > > > >    uncorrectable memory errors at corresponding physical addresses.
> > > > > 3. Signal both threads to make their memory buffer collapsible, i.e.
> > > > >    calling madvise(MADV_HUGEPAGE).
> > > > > 4. Wait and check kernel log: khugepaged is able to recover from poisoned
> > > > >    pages and skips collapsing them.
> > > > > 5. Signal both threads to inspect their buffer contents and make sure no
> > > > >    data corruption.
> > > > >
> > > > > Signed-off-by: Jiaqi Yan <jiaqiyan@google.com>
> > > > > ---
> > > > >  include/trace/events/huge_memory.h |   3 +-
> > > > >  mm/khugepaged.c                    | 148 ++++++++++++++++++++++++-----
> > > > >  2 files changed, 128 insertions(+), 23 deletions(-)
> > > > >
> > > > > diff --git a/include/trace/events/huge_memory.h b/include/trace/events/huge_memory.h
> > > > > index 3e6fb05852f9a..46cce509957ba 100644
> > > > > --- a/include/trace/events/huge_memory.h
> > > > > +++ b/include/trace/events/huge_memory.h
> > > > > @@ -36,7 +36,8 @@
> > > > >         EM( SCAN_ALLOC_HUGE_PAGE_FAIL,  "alloc_huge_page_failed")       \
> > > > >         EM( SCAN_CGROUP_CHARGE_FAIL,    "ccgroup_charge_failed")        \
> > > > >         EM( SCAN_TRUNCATED,             "truncated")                    \
> > > > > -       EMe(SCAN_PAGE_HAS_PRIVATE,      "page_has_private")             \
> > > > > +       EM( SCAN_PAGE_HAS_PRIVATE,      "page_has_private")             \
> > > > > +       EMe(SCAN_COPY_MC,               "copy_poisoned_page")           \
> > > > >
> > > > >  #undef EM
> > > > >  #undef EMe
> > > > > diff --git a/mm/khugepaged.c b/mm/khugepaged.c
> > > > > index 27956d4404134..c3c217f6ebc6e 100644
> > > > > --- a/mm/khugepaged.c
> > > > > +++ b/mm/khugepaged.c
> > > > > @@ -19,6 +19,7 @@
> > > > >  #include <linux/page_table_check.h>
> > > > >  #include <linux/swapops.h>
> > > > >  #include <linux/shmem_fs.h>
> > > > > +#include <linux/kmsan.h>
> > > > >
> > > > >  #include <asm/tlb.h>
> > > > >  #include <asm/pgalloc.h>
> > > > > @@ -55,6 +56,7 @@ enum scan_result {
> > > > >         SCAN_CGROUP_CHARGE_FAIL,
> > > > >         SCAN_TRUNCATED,
> > > > >         SCAN_PAGE_HAS_PRIVATE,
> > > > > +       SCAN_COPY_MC,
> > > > >  };
> > > > >
> > > > >  #define CREATE_TRACE_POINTS
> > > > > @@ -681,47 +683,47 @@ static int __collapse_huge_page_isolate(struct vm_area_struct *vma,
> > > > >         return result;
> > > > >  }
> > > > >
> > > > > -static void __collapse_huge_page_copy(pte_t *pte, struct page *page,
> > > > > -                                     struct vm_area_struct *vma,
> > > > > -                                     unsigned long address,
> > > > > -                                     spinlock_t *ptl,
> > > > > -                                     struct list_head *compound_pagelist)
> > > > > +static void __collapse_huge_page_copy_succeeded(pte_t *pte,
> > > > > +                                               pmd_t *pmd,
> > > > > +                                               struct vm_area_struct *vma,
> > > > > +                                               unsigned long address,
> > > > > +                                               spinlock_t *pte_ptl,
> > > > > +                                               struct list_head *compound_pagelist)
> > > > >  {
> > > > >         struct page *src_page, *tmp;
> > > > >         pte_t *_pte;
> > > > > -       for (_pte = pte; _pte < pte + HPAGE_PMD_NR;
> > > > > -                               _pte++, page++, address += PAGE_SIZE) {
> > > > > -               pte_t pteval = *_pte;
> > > > > +       pte_t pteval;
> > > > > +       unsigned long _address;
> > > > >
> > > > > +       for (_pte = pte, _address = address; _pte < pte + HPAGE_PMD_NR;
> > > > > +            _pte++, _address += PAGE_SIZE) {
> > > > > +               pteval = *_pte;
> > > > >                 if (pte_none(pteval) || is_zero_pfn(pte_pfn(pteval))) {
> > > > > -                       clear_user_highpage(page, address);
> > > > >                         add_mm_counter(vma->vm_mm, MM_ANONPAGES, 1);
> > > > >                         if (is_zero_pfn(pte_pfn(pteval))) {
> > > > >                                 /*
> > > > > -                                * ptl mostly unnecessary.
> > > > > +                                * pte_ptl mostly unnecessary.
> > > > >                                  */
> > > > > -                               spin_lock(ptl);
> > > > > -                               ptep_clear(vma->vm_mm, address, _pte);
> > > > > -                               spin_unlock(ptl);
> > > > > +                               spin_lock(pte_ptl);
> > > >
> > > > Why did you have to rename ptl to pte_ptl? It seems unnecessary.
> > >
> > > Thanks, I will use `ptl` in the next version.
> > >
> > > >
> > > > > +                               pte_clear(vma->vm_mm, _address, _pte);

A bug here: calling pte_clear is wrong; we should use ptep_clear. V11
will make sure it is fixed.

BTW, __collapse_huge_page_copy_succeeded can drop _address and use
address directly.

> > > > > +                               spin_unlock(pte_ptl);
> > > > >                         }
> > > > >                 } else {
> > > > >                         src_page = pte_page(pteval);
> > > > > -                       copy_user_highpage(page, src_page, address, vma);
> > > > >                         if (!PageCompound(src_page))
> > > > >                                 release_pte_page(src_page);
> > > > >                         /*
> > > > > -                        * ptl mostly unnecessary, but preempt has to
> > > > > -                        * be disabled to update the per-cpu stats
> > > > > +                        * pte_ptl mostly unnecessary, but preempt has
> > > > > +                        * to be disabled to update the per-cpu stats
> > > > >                          * inside page_remove_rmap().
> > > > >                          */
> > > > > -                       spin_lock(ptl);
> > > > > -                       ptep_clear(vma->vm_mm, address, _pte);
> > > > > +                       spin_lock(pte_ptl);
> > > > > +                       ptep_clear(vma->vm_mm, _address, _pte);
> > > > >                         page_remove_rmap(src_page, vma, false);
> > > > > -                       spin_unlock(ptl);
> > > > > +                       spin_unlock(pte_ptl);
> > > > >                         free_page_and_swap_cache(src_page);
> > > > >                 }
> > > > >         }
> > > > > -
> > > > >         list_for_each_entry_safe(src_page, tmp, compound_pagelist, lru) {
> > > > >                 list_del(&src_page->lru);
> > > > >                 mod_node_page_state(page_pgdat(src_page),
> > > > > @@ -733,6 +735,104 @@ static void __collapse_huge_page_copy(pte_t *pte, struct page *page,
> > > > >         }
> > > > >  }
> > > > >
> > > > > +static void __collapse_huge_page_copy_failed(pte_t *pte,
> > > > > +                                            pmd_t *pmd,
> > > > > +                                            pmd_t orig_pmd,
> > > > > +                                            struct vm_area_struct *vma,
> > > > > +                                            unsigned long address,
> > > > > +                                            struct list_head *compound_pagelist)
> > > > > +{
> > > > > +       struct page *src_page, *tmp;
> > > > > +       pte_t *_pte;
> > > > > +       pte_t pteval;
> > > > > +       unsigned long _address;
> > > > > +       spinlock_t *pmd_ptl;
> > > > > +
> > > > > +       /*
> > > > > +        * Re-establish the PMD to point to the original page table
> > > > > +        * entry. Restoring PMD needs to be done prior to releasing
> > > > > +        * pages. Since pages are still isolated and locked here,
> > > > > +        * acquiring anon_vma_lock_write is unnecessary.
> > > > > +        */
> > > > > +       pmd_ptl = pmd_lock(vma->vm_mm, pmd);
> > > > > +       pmd_populate(vma->vm_mm, pmd, pmd_pgtable(orig_pmd));
> > > > > +       spin_unlock(pmd_ptl);
> > > > > +       /*
> > > > > +        * Release both raw and compound pages isolated
> > > > > +        * in __collapse_huge_page_isolate.
> > > > > +        */
> > > >
> > > > It looks like the below code could be replaced by release_pte_pages()
> > > > with advancing _pte to (pte + HPAGE_PMD_NR - 1).
> > > >
> > >
> > > Yeah, but I think _pte should be (pte + HPAGE_PMR_NR) because _pte is
> > > decremented before comparison in release_pte_pages(pte, _pte,
> > > compound_pagelist):
> > >
> > >     while (--_pte >= pte) {...}
> > >
> > > Advancing _pte to (pte + HPAGE_PMD_NR - 1) may leave the last page not released.
> >
> > Yeah, good catch. I think it is because the only user of
> > release_pte_pages() is __collapse_huge_page_isolate(). Once the loop
> > in it is done _pte is pte + HPAGE_PMD_NR.
> >
>
> Thanks for confirming ;)
> Since you have reviewed other 2 commits, I will soon send out V11 for
> your comments on this commit.
>
>
> > >
> > > >
> > > > > +       for (_pte = pte, _address = address; _pte < pte + HPAGE_PMD_NR;
> > > > > +               _pte++, _address += PAGE_SIZE) {
> > > > > +               pteval = *_pte;
> > > > > +               if (pte_none(pteval) || is_zero_pfn(pte_pfn(pteval)))
> > > > > +                       continue;
> > > > > +               src_page = pte_page(pteval);
> > > > > +               if (!PageCompound(src_page))
> > > > > +                       release_pte_page(src_page);
> > > > > +       }
> > > > > +       list_for_each_entry_safe(src_page, tmp, compound_pagelist, lru) {
> > > > > +               list_del(&src_page->lru);
> > > > > +               release_pte_page(src_page);
> > > > > +       }
> > > > > +}
> > > > > +
> > > > > +/*
> > > > > + * __collapse_huge_page_copy - attempts to copy memory contents from raw
> > > > > + * pages to a hugepage. Cleans up the raw pages if copying succeeds;
> > > > > + * otherwise restores the original page table and releases isolated raw pages.
> > > > > + * Returns SCAN_SUCCEED if copying succeeds, otherwise returns SCAN_COPY_MC.
> > > > > + *
> > > > > + * @pte: starting of the PTEs to copy from
> > > > > + * @page: the new hugepage to copy contents to
> > > > > + * @pmd: pointer to the new hugepage's PMD
> > > > > + * @orig_pmd: the original raw pages' PMD
> > > > > + * @vma: the original raw pages' virtual memory area
> > > > > + * @address: starting address to copy
> > > > > + * @pte_ptl: lock on raw pages' PTEs
> > > > > + * @compound_pagelist: list that stores compound pages
> > > > > + */
> > > > > +static int __collapse_huge_page_copy(pte_t *pte,
> > > > > +                                    struct page *page,
> > > > > +                                    pmd_t *pmd,
> > > > > +                                    pmd_t orig_pmd,
> > > > > +                                    struct vm_area_struct *vma,
> > > > > +                                    unsigned long address,
> > > > > +                                    spinlock_t *pte_ptl,
> > > > > +                                    struct list_head *compound_pagelist)
> > > > > +{
> > > > > +       struct page *src_page;
> > > > > +       pte_t *_pte;
> > > > > +       pte_t pteval;
> > > > > +       unsigned long _address;
> > > > > +       int result = SCAN_SUCCEED;
> > > > > +
> > > > > +       /*
> > > > > +        * Copying pages' contents is subject to memory poison at any iteration.
> > > > > +        */
> > > > > +       for (_pte = pte, _address = address; _pte < pte + HPAGE_PMD_NR;
> > > > > +            _pte++, page++, _address += PAGE_SIZE) {
> > > > > +               pteval = *_pte;
> > > > > +               if (pte_none(pteval) || is_zero_pfn(pte_pfn(pteval))) {
> > > > > +                       clear_user_highpage(page, _address);
> > > > > +                       continue;
> > > > > +               }
> > > > > +               src_page = pte_page(pteval);
> > > > > +               if (copy_mc_user_highpage(page, src_page, _address, vma) > 0) {
> > > > > +                       result = SCAN_COPY_MC;
> > > > > +                       break;
> > > > > +               }
> > > > > +       }
> > > > > +
> > > > > +       if (likely(result == SCAN_SUCCEED))
> > > > > +               __collapse_huge_page_copy_succeeded(pte, pmd, vma, address,
> > > > > +                                                   pte_ptl, compound_pagelist);
> > > > > +       else
> > > > > +               __collapse_huge_page_copy_failed(pte, pmd, orig_pmd, vma,
> > > > > +                                                address, compound_pagelist);
> > > > > +
> > > > > +       return result;
> > > > > +}
> > > > > +
> > > > >  static void khugepaged_alloc_sleep(void)
> > > > >  {
> > > > >         DEFINE_WAIT(wait);
> > > > > @@ -1106,9 +1206,13 @@ static int collapse_huge_page(struct mm_struct *mm, unsigned long address,
> > > > >          */
> > > > >         anon_vma_unlock_write(vma->anon_vma);
> > > > >
> > > > > -       __collapse_huge_page_copy(pte, hpage, vma, address, pte_ptl,
> > > > > -                                 &compound_pagelist);
> > > > > +       result = __collapse_huge_page_copy(pte, hpage, pmd, _pmd,
> > > > > +                                          vma, address, pte_ptl,
> > > > > +                                          &compound_pagelist);
> > > > >         pte_unmap(pte);
> > > > > +       if (unlikely(result != SCAN_SUCCEED))
> > > > > +               goto out_up_write;
> > > > > +
> > > > >         /*
> > > > >          * spin_lock() below is not the equivalent of smp_wmb(), but
> > > > >          * the smp_wmb() inside __SetPageUptodate() can be reused to
> > > > > --
> > > > > 2.40.0.rc0.216.gc4246ad0f0-goog
> > > > >


  reply	other threads:[~2023-03-27 20:46 UTC|newest]

Thread overview: 17+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2023-03-05  6:51 [PATCH v10 0/3] Memory poison recovery in khugepaged collapsing Jiaqi Yan
2023-03-05  6:51 ` [PATCH v10 1/3] mm/khugepaged: recover from poisoned anonymous memory Jiaqi Yan
2023-03-20 14:42   ` Jiaqi Yan
2023-03-21  0:12     ` Yang Shi
2023-03-23 21:37   ` Yang Shi
2023-03-24 15:34     ` Jiaqi Yan
2023-03-24 20:11       ` Yang Shi
2023-03-24 22:31         ` Jiaqi Yan
2023-03-27 20:46           ` Jiaqi Yan [this message]
2023-03-05  6:51 ` [PATCH v10 2/3] mm/hwpoison: introduce copy_mc_highpage Jiaqi Yan
2023-03-05  6:56   ` Jiaqi Yan
2023-03-24 20:24   ` Yang Shi
2023-03-05  6:51 ` [PATCH v10 3/3] mm/khugepaged: recover from poisoned file-backed memory Jiaqi Yan
2023-03-24 21:15   ` Yang Shi
2023-03-24 22:54     ` Jiaqi Yan
2023-03-25  0:39       ` Hugh Dickins
2023-03-27 21:15         ` Jiaqi Yan

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=CACw3F50JE+Luo_wQEhScp+rBXRCk0SrSGUrfscWa7iKNDDUaHA@mail.gmail.com \
    --to=jiaqiyan@google.com \
    --cc=akpm@linux-foundation.org \
    --cc=kirill.shutemov@linux.intel.com \
    --cc=kirill@shutemov.name \
    --cc=linmiaohe@huawei.com \
    --cc=linux-mm@kvack.org \
    --cc=naoya.horiguchi@nec.com \
    --cc=osalvador@suse.de \
    --cc=shy828301@gmail.com \
    --cc=tongtiangen@huawei.com \
    --cc=tony.luck@intel.com \
    --cc=wangkefeng.wang@huawei.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).