All of lore.kernel.org
 help / color / mirror / Atom feed
* + mm-reclaim-madv_free-pages.patch added to -mm tree
@ 2017-03-01  0:32 akpm
  2017-03-03  2:52 ` Minchan Kim
  0 siblings, 1 reply; 11+ messages in thread
From: akpm @ 2017-03-01  0:32 UTC (permalink / raw)
  To: shli, hannes, hillf.zj, hughd, mgorman, mhocko, minchan, riel,
	mm-commits


The patch titled
     Subject: mm: reclaim MADV_FREE pages
has been added to the -mm tree.  Its filename is
     mm-reclaim-madv_free-pages.patch

This patch should soon appear at
    http://ozlabs.org/~akpm/mmots/broken-out/mm-reclaim-madv_free-pages.patch
and later at
    http://ozlabs.org/~akpm/mmotm/broken-out/mm-reclaim-madv_free-pages.patch

Before you just go and hit "reply", please:
   a) Consider who else should be cc'ed
   b) Prefer to cc a suitable mailing list as well
   c) Ideally: find the original patch on the mailing list and do a
      reply-to-all to that, adding suitable additional cc's

*** Remember to use Documentation/SubmitChecklist when testing your code ***

The -mm tree is included into linux-next and is updated
there every 3-4 working days

------------------------------------------------------
From: Shaohua Li <shli@fb.com>
Subject: mm: reclaim MADV_FREE pages

When memory pressure is high, we free MADV_FREE pages.  If the pages are
not dirty in pte, the pages could be freed immediately.  Otherwise we
can't reclaim them.  We put the pages back to anonumous LRU list (by
setting SwapBacked flag) and the pages will be reclaimed in normal swapout
way.

We use normal page reclaim policy.  Since MADV_FREE pages are put into
inactive file list, such pages and inactive file pages are reclaimed
according to their age.  This is expected, because we don't want to
reclaim too many MADV_FREE pages before used once pages.

Based on Minchan's original patch

Link: http://lkml.kernel.org/r/14b8eb1d3f6bf6cc492833f183ac8c304e560484.1487965799.git.shli@fb.com
Signed-off-by: Shaohua Li <shli@fb.com>
Acked-by: Minchan Kim <minchan@kernel.org>
Acked-by: Michal Hocko <mhocko@suse.com>
Acked-by: Johannes Weiner <hannes@cmpxchg.org>
Acked-by: Hillf Danton <hillf.zj@alibaba-inc.com>
Cc: Hugh Dickins <hughd@google.com>
Cc: Rik van Riel <riel@redhat.com>
Cc: Mel Gorman <mgorman@techsingularity.net>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---

 include/linux/rmap.h |    2 +-
 mm/huge_memory.c     |    2 ++
 mm/madvise.c         |    1 +
 mm/rmap.c            |   40 +++++++++++++++++-----------------------
 mm/vmscan.c          |   34 ++++++++++++++++++++++------------
 5 files changed, 43 insertions(+), 36 deletions(-)

diff -puN include/linux/rmap.h~mm-reclaim-madv_free-pages include/linux/rmap.h
--- a/include/linux/rmap.h~mm-reclaim-madv_free-pages
+++ a/include/linux/rmap.h
@@ -298,6 +298,6 @@ static inline int page_mkclean(struct pa
 #define SWAP_AGAIN	1
 #define SWAP_FAIL	2
 #define SWAP_MLOCK	3
-#define SWAP_LZFREE	4
+#define SWAP_DIRTY	4
 
 #endif	/* _LINUX_RMAP_H */
diff -puN mm/huge_memory.c~mm-reclaim-madv_free-pages mm/huge_memory.c
--- a/mm/huge_memory.c~mm-reclaim-madv_free-pages
+++ a/mm/huge_memory.c
@@ -1571,6 +1571,8 @@ bool madvise_free_huge_pmd(struct mmu_ga
 		set_pmd_at(mm, addr, pmd, orig_pmd);
 		tlb_remove_pmd_tlb_entry(tlb, pmd, addr);
 	}
+
+	mark_page_lazyfree(page);
 	ret = true;
 out:
 	spin_unlock(ptl);
diff -puN mm/madvise.c~mm-reclaim-madv_free-pages mm/madvise.c
--- a/mm/madvise.c~mm-reclaim-madv_free-pages
+++ a/mm/madvise.c
@@ -413,6 +413,7 @@ static int madvise_free_pte_range(pmd_t
 			set_pte_at(mm, addr, pte, ptent);
 			tlb_remove_tlb_entry(tlb, pte, addr);
 		}
+		mark_page_lazyfree(page);
 	}
 out:
 	if (nr_swap) {
diff -puN mm/rmap.c~mm-reclaim-madv_free-pages mm/rmap.c
--- a/mm/rmap.c~mm-reclaim-madv_free-pages
+++ a/mm/rmap.c
@@ -1281,11 +1281,6 @@ void page_remove_rmap(struct page *page,
 	 */
 }
 
-struct rmap_private {
-	enum ttu_flags flags;
-	int lazyfreed;
-};
-
 /*
  * @arg: enum ttu_flags will be passed to this argument
  */
@@ -1301,8 +1296,7 @@ static int try_to_unmap_one(struct page
 	pte_t pteval;
 	struct page *subpage;
 	int ret = SWAP_AGAIN;
-	struct rmap_private *rp = arg;
-	enum ttu_flags flags = rp->flags;
+	enum ttu_flags flags = (enum ttu_flags)arg;
 
 	/* munlock has nothing to gain from examining un-locked vmas */
 	if ((flags & TTU_MUNLOCK) && !(vma->vm_flags & VM_LOCKED))
@@ -1419,11 +1413,21 @@ static int try_to_unmap_one(struct page
 			VM_BUG_ON_PAGE(!PageSwapCache(page) && PageSwapBacked(page),
 				page);
 
-			if (!PageDirty(page)) {
+			/*
+			 * swapin page could be clean, it has data stored in
+			 * swap. We can't silently discard it without setting
+			 * swap entry in the page table.
+			 */
+			if (!PageDirty(page) && !PageSwapCache(page)) {
 				/* It's a freeable page by MADV_FREE */
 				dec_mm_counter(mm, MM_ANONPAGES);
-				rp->lazyfreed++;
 				goto discard;
+			} else if (!PageSwapBacked(page)) {
+				/* dirty MADV_FREE page */
+				set_pte_at(mm, address, pvmw.pte, pteval);
+				ret = SWAP_DIRTY;
+				page_vma_mapped_walk_done(&pvmw);
+				break;
 			}
 
 			if (swap_duplicate(entry) < 0) {
@@ -1491,18 +1495,15 @@ static int page_mapcount_is_zero(struct
  * SWAP_AGAIN	- we missed a mapping, try again later
  * SWAP_FAIL	- the page is unswappable
  * SWAP_MLOCK	- page is mlocked.
+ * SWAP_DIRTY	- page is dirty MADV_FREE page
  */
 int try_to_unmap(struct page *page, enum ttu_flags flags)
 {
 	int ret;
-	struct rmap_private rp = {
-		.flags = flags,
-		.lazyfreed = 0,
-	};
 
 	struct rmap_walk_control rwc = {
 		.rmap_one = try_to_unmap_one,
-		.arg = &rp,
+		.arg = (void *)flags,
 		.done = page_mapcount_is_zero,
 		.anon_lock = page_lock_anon_vma_read,
 	};
@@ -1523,11 +1524,8 @@ int try_to_unmap(struct page *page, enum
 	else
 		ret = rmap_walk(page, &rwc);
 
-	if (ret != SWAP_MLOCK && !page_mapcount(page)) {
+	if (ret != SWAP_MLOCK && !page_mapcount(page))
 		ret = SWAP_SUCCESS;
-		if (rp.lazyfreed && !PageDirty(page))
-			ret = SWAP_LZFREE;
-	}
 	return ret;
 }
 
@@ -1554,14 +1552,10 @@ static int page_not_mapped(struct page *
 int try_to_munlock(struct page *page)
 {
 	int ret;
-	struct rmap_private rp = {
-		.flags = TTU_MUNLOCK,
-		.lazyfreed = 0,
-	};
 
 	struct rmap_walk_control rwc = {
 		.rmap_one = try_to_unmap_one,
-		.arg = &rp,
+		.arg = (void *)TTU_MUNLOCK,
 		.done = page_not_mapped,
 		.anon_lock = page_lock_anon_vma_read,
 
diff -puN mm/vmscan.c~mm-reclaim-madv_free-pages mm/vmscan.c
--- a/mm/vmscan.c~mm-reclaim-madv_free-pages
+++ a/mm/vmscan.c
@@ -905,7 +905,8 @@ static void page_check_dirty_writeback(s
 	 * Anonymous pages are not handled by flushers and must be written
 	 * from reclaim context. Do not stall reclaim based on them
 	 */
-	if (!page_is_file_cache(page)) {
+	if (!page_is_file_cache(page) ||
+	    (PageAnon(page) && !PageSwapBacked(page))) {
 		*dirty = false;
 		*writeback = false;
 		return;
@@ -986,7 +987,8 @@ static unsigned long shrink_page_list(st
 			goto keep_locked;
 
 		/* Double the slab pressure for mapped and swapcache pages */
-		if (page_mapped(page) || PageSwapCache(page))
+		if ((page_mapped(page) || PageSwapCache(page)) &&
+		    !(PageAnon(page) && !PageSwapBacked(page)))
 			sc->nr_scanned++;
 
 		may_enter_fs = (sc->gfp_mask & __GFP_FS) ||
@@ -1112,8 +1114,10 @@ static unsigned long shrink_page_list(st
 		/*
 		 * Anonymous process memory has backing store?
 		 * Try to allocate it some swap space here.
+		 * Lazyfree page could be freed directly
 		 */
-		if (PageAnon(page) && !PageSwapCache(page)) {
+		if (PageAnon(page) && PageSwapBacked(page) &&
+		    !PageSwapCache(page)) {
 			if (!(sc->gfp_mask & __GFP_IO))
 				goto keep_locked;
 			if (!add_to_swap(page, page_list))
@@ -1134,9 +1138,12 @@ static unsigned long shrink_page_list(st
 		 * The page is mapped into the page tables of one or more
 		 * processes. Try to unmap it here.
 		 */
-		if (page_mapped(page) && mapping) {
+		if (page_mapped(page)) {
 			switch (ret = try_to_unmap(page,
 				ttu_flags | TTU_BATCH_FLUSH)) {
+			case SWAP_DIRTY:
+				SetPageSwapBacked(page);
+				/* fall through */
 			case SWAP_FAIL:
 				nr_unmap_fail++;
 				goto activate_locked;
@@ -1144,8 +1151,6 @@ static unsigned long shrink_page_list(st
 				goto keep_locked;
 			case SWAP_MLOCK:
 				goto cull_mlocked;
-			case SWAP_LZFREE:
-				goto lazyfree;
 			case SWAP_SUCCESS:
 				; /* try to free the page below */
 			}
@@ -1257,10 +1262,18 @@ static unsigned long shrink_page_list(st
 			}
 		}
 
-lazyfree:
-		if (!mapping || !__remove_mapping(mapping, page, true))
-			goto keep_locked;
+		if (PageAnon(page) && !PageSwapBacked(page)) {
+			/* follow __remove_mapping for reference */
+			if (!page_ref_freeze(page, 1))
+				goto keep_locked;
+			if (PageDirty(page)) {
+				page_ref_unfreeze(page, 1);
+				goto keep_locked;
+			}
 
+			count_vm_event(PGLAZYFREED);
+		} else if (!mapping || !__remove_mapping(mapping, page, true))
+			goto keep_locked;
 		/*
 		 * At this point, we have no other references and there is
 		 * no way to pick any more up (removed from LRU, removed
@@ -1270,9 +1283,6 @@ lazyfree:
 		 */
 		__ClearPageLocked(page);
 free_it:
-		if (ret == SWAP_LZFREE)
-			count_vm_event(PGLAZYFREED);

^ permalink raw reply	[flat|nested] 11+ messages in thread

* Re: + mm-reclaim-madv_free-pages.patch added to -mm tree
  2017-03-01  0:32 + mm-reclaim-madv_free-pages.patch added to -mm tree akpm
@ 2017-03-03  2:52 ` Minchan Kim
  2017-03-03 15:18   ` Johannes Weiner
  2017-03-03 16:02   ` Shaohua Li
  0 siblings, 2 replies; 11+ messages in thread
From: Minchan Kim @ 2017-03-03  2:52 UTC (permalink / raw)
  To: linux-kernel
  Cc: shli, hannes, hillf.zj, hughd, mgorman, mhocko, riel, mm-commits

Hi,

On Tue, Feb 28, 2017 at 04:32:38PM -0800, akpm@linux-foundation.org wrote:
> 
> The patch titled
>      Subject: mm: reclaim MADV_FREE pages
> has been added to the -mm tree.  Its filename is
>      mm-reclaim-madv_free-pages.patch
> 
> This patch should soon appear at
>     http://ozlabs.org/~akpm/mmots/broken-out/mm-reclaim-madv_free-pages.patch
> and later at
>     http://ozlabs.org/~akpm/mmotm/broken-out/mm-reclaim-madv_free-pages.patch
> 
> Before you just go and hit "reply", please:
>    a) Consider who else should be cc'ed
>    b) Prefer to cc a suitable mailing list as well
>    c) Ideally: find the original patch on the mailing list and do a
>       reply-to-all to that, adding suitable additional cc's
> 
> *** Remember to use Documentation/SubmitChecklist when testing your code ***
> 
> The -mm tree is included into linux-next and is updated
> there every 3-4 working days
> 
> ------------------------------------------------------
> From: Shaohua Li <shli@fb.com>
> Subject: mm: reclaim MADV_FREE pages
> 
> When memory pressure is high, we free MADV_FREE pages.  If the pages are
> not dirty in pte, the pages could be freed immediately.  Otherwise we
> can't reclaim them.  We put the pages back to anonumous LRU list (by
> setting SwapBacked flag) and the pages will be reclaimed in normal swapout
> way.
> 
> We use normal page reclaim policy.  Since MADV_FREE pages are put into
> inactive file list, such pages and inactive file pages are reclaimed
> according to their age.  This is expected, because we don't want to
> reclaim too many MADV_FREE pages before used once pages.
> 
> Based on Minchan's original patch
> 
> Link: http://lkml.kernel.org/r/14b8eb1d3f6bf6cc492833f183ac8c304e560484.1487965799.git.shli@fb.com
> Signed-off-by: Shaohua Li <shli@fb.com>
> Acked-by: Minchan Kim <minchan@kernel.org>
> Acked-by: Michal Hocko <mhocko@suse.com>
> Acked-by: Johannes Weiner <hannes@cmpxchg.org>
> Acked-by: Hillf Danton <hillf.zj@alibaba-inc.com>
> Cc: Hugh Dickins <hughd@google.com>
> Cc: Rik van Riel <riel@redhat.com>
> Cc: Mel Gorman <mgorman@techsingularity.net>
> Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
> ---

< snip >

> @@ -1419,11 +1413,21 @@ static int try_to_unmap_one(struct page
>  			VM_BUG_ON_PAGE(!PageSwapCache(page) && PageSwapBacked(page),
>  				page);
>  
> -			if (!PageDirty(page)) {
> +			/*
> +			 * swapin page could be clean, it has data stored in
> +			 * swap. We can't silently discard it without setting
> +			 * swap entry in the page table.
> +			 */
> +			if (!PageDirty(page) && !PageSwapCache(page)) {
>  				/* It's a freeable page by MADV_FREE */
>  				dec_mm_counter(mm, MM_ANONPAGES);
> -				rp->lazyfreed++;
>  				goto discard;
> +			} else if (!PageSwapBacked(page)) {
> +				/* dirty MADV_FREE page */
> +				set_pte_at(mm, address, pvmw.pte, pteval);
> +				ret = SWAP_DIRTY;
> +				page_vma_mapped_walk_done(&pvmw);
> +				break;
>  			}

There is no point to make this logic complicated with clean swapin-page.

Andrew,
Could you fold below patch into the mm-reclaim-madv_free-pages.patch
if others are not against?

Thanks.

>From 0c28f6560fbc4e65da4f4a8cc4664ab9f7b11cf3 Mon Sep 17 00:00:00 2001
From: Minchan Kim <minchan@kernel.org>
Date: Fri, 3 Mar 2017 11:42:52 +0900
Subject: [PATCH] mm: clean up lazyfree page handling

We can make it simple to understand without need to be aware of
clean-swapin page.
This patch just clean up lazyfree page handling in try_to_unmap_one.

Signed-off-by: Minchan Kim <minchan@kernel.org>
---
 mm/rmap.c | 22 +++++++++++-----------
 1 file changed, 11 insertions(+), 11 deletions(-)

diff --git a/mm/rmap.c b/mm/rmap.c
index bb45712..f7eab40 100644
--- a/mm/rmap.c
+++ b/mm/rmap.c
@@ -1413,17 +1413,17 @@ static int try_to_unmap_one(struct page *page, struct vm_area_struct *vma,
 			VM_BUG_ON_PAGE(!PageSwapCache(page) && PageSwapBacked(page),
 				page);
 
-			/*
-			 * swapin page could be clean, it has data stored in
-			 * swap. We can't silently discard it without setting
-			 * swap entry in the page table.
-			 */
-			if (!PageDirty(page) && !PageSwapCache(page)) {
-				/* It's a freeable page by MADV_FREE */
-				dec_mm_counter(mm, MM_ANONPAGES);
-				goto discard;
-			} else if (!PageSwapBacked(page)) {
-				/* dirty MADV_FREE page */
+			/* MADV_FREE page check */
+			if (!PageSwapBacked(page)) {
+				if (!PageDirty(page)) {
+					dec_mm_counter(mm, MM_ANONPAGES);
+					goto discard;
+				}
+
+				/*
+				 * If the page was redirtied, it cannot be
+				 * discarded. Remap the page to page table.
+				 */
 				set_pte_at(mm, address, pvmw.pte, pteval);
 				ret = SWAP_DIRTY;
 				page_vma_mapped_walk_done(&pvmw);
-- 
2.7.4

^ permalink raw reply related	[flat|nested] 11+ messages in thread

* Re: + mm-reclaim-madv_free-pages.patch added to -mm tree
  2017-03-03  2:52 ` Minchan Kim
@ 2017-03-03 15:18   ` Johannes Weiner
  2017-03-06  3:03     ` Minchan Kim
  2017-03-03 16:02   ` Shaohua Li
  1 sibling, 1 reply; 11+ messages in thread
From: Johannes Weiner @ 2017-03-03 15:18 UTC (permalink / raw)
  To: Minchan Kim
  Cc: linux-kernel, shli, hillf.zj, hughd, mgorman, mhocko, riel, mm-commits

On Fri, Mar 03, 2017 at 11:52:37AM +0900, Minchan Kim wrote:
> On Tue, Feb 28, 2017 at 04:32:38PM -0800, akpm@linux-foundation.org wrote:
> > 
> > The patch titled
> >      Subject: mm: reclaim MADV_FREE pages
> > has been added to the -mm tree.  Its filename is
> >      mm-reclaim-madv_free-pages.patch
> > 
> > This patch should soon appear at
> >     http://ozlabs.org/~akpm/mmots/broken-out/mm-reclaim-madv_free-pages.patch
> > and later at
> >     http://ozlabs.org/~akpm/mmotm/broken-out/mm-reclaim-madv_free-pages.patch
> > 
> > Before you just go and hit "reply", please:
> >    a) Consider who else should be cc'ed
> >    b) Prefer to cc a suitable mailing list as well
> >    c) Ideally: find the original patch on the mailing list and do a
> >       reply-to-all to that, adding suitable additional cc's
> > 
> > *** Remember to use Documentation/SubmitChecklist when testing your code ***
> > 
> > The -mm tree is included into linux-next and is updated
> > there every 3-4 working days
> > 
> > ------------------------------------------------------
> > From: Shaohua Li <shli@fb.com>
> > Subject: mm: reclaim MADV_FREE pages
> > 
> > When memory pressure is high, we free MADV_FREE pages.  If the pages are
> > not dirty in pte, the pages could be freed immediately.  Otherwise we
> > can't reclaim them.  We put the pages back to anonumous LRU list (by
> > setting SwapBacked flag) and the pages will be reclaimed in normal swapout
> > way.
> > 
> > We use normal page reclaim policy.  Since MADV_FREE pages are put into
> > inactive file list, such pages and inactive file pages are reclaimed
> > according to their age.  This is expected, because we don't want to
> > reclaim too many MADV_FREE pages before used once pages.
> > 
> > Based on Minchan's original patch
> > 
> > Link: http://lkml.kernel.org/r/14b8eb1d3f6bf6cc492833f183ac8c304e560484.1487965799.git.shli@fb.com
> > Signed-off-by: Shaohua Li <shli@fb.com>
> > Acked-by: Minchan Kim <minchan@kernel.org>
> > Acked-by: Michal Hocko <mhocko@suse.com>
> > Acked-by: Johannes Weiner <hannes@cmpxchg.org>
> > Acked-by: Hillf Danton <hillf.zj@alibaba-inc.com>
> > Cc: Hugh Dickins <hughd@google.com>
> > Cc: Rik van Riel <riel@redhat.com>
> > Cc: Mel Gorman <mgorman@techsingularity.net>
> > Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
> > ---
> 
> < snip >
> 
> > @@ -1419,11 +1413,21 @@ static int try_to_unmap_one(struct page
> >  			VM_BUG_ON_PAGE(!PageSwapCache(page) && PageSwapBacked(page),
> >  				page);
> >  
> > -			if (!PageDirty(page)) {
> > +			/*
> > +			 * swapin page could be clean, it has data stored in
> > +			 * swap. We can't silently discard it without setting
> > +			 * swap entry in the page table.
> > +			 */
> > +			if (!PageDirty(page) && !PageSwapCache(page)) {
> >  				/* It's a freeable page by MADV_FREE */
> >  				dec_mm_counter(mm, MM_ANONPAGES);
> > -				rp->lazyfreed++;
> >  				goto discard;
> > +			} else if (!PageSwapBacked(page)) {
> > +				/* dirty MADV_FREE page */
> > +				set_pte_at(mm, address, pvmw.pte, pteval);
> > +				ret = SWAP_DIRTY;
> > +				page_vma_mapped_walk_done(&pvmw);
> > +				break;
> >  			}
> 
> There is no point to make this logic complicated with clean swapin-page.
> 
> Andrew,
> Could you fold below patch into the mm-reclaim-madv_free-pages.patch
> if others are not against?
> 
> Thanks.
> 
> From 0c28f6560fbc4e65da4f4a8cc4664ab9f7b11cf3 Mon Sep 17 00:00:00 2001
> From: Minchan Kim <minchan@kernel.org>
> Date: Fri, 3 Mar 2017 11:42:52 +0900
> Subject: [PATCH] mm: clean up lazyfree page handling
> 
> We can make it simple to understand without need to be aware of
> clean-swapin page.
> This patch just clean up lazyfree page handling in try_to_unmap_one.
> 
> Signed-off-by: Minchan Kim <minchan@kernel.org>

Agreed, this is a litle easier to follow.

Acked-by: Johannes Weiner <hannes@cmpxchg.org>

> ---
>  mm/rmap.c | 22 +++++++++++-----------
>  1 file changed, 11 insertions(+), 11 deletions(-)
> 
> diff --git a/mm/rmap.c b/mm/rmap.c
> index bb45712..f7eab40 100644
> --- a/mm/rmap.c
> +++ b/mm/rmap.c
> @@ -1413,17 +1413,17 @@ static int try_to_unmap_one(struct page *page, struct vm_area_struct *vma,
>  			VM_BUG_ON_PAGE(!PageSwapCache(page) && PageSwapBacked(page),
>  				page);

Since you're removing the PageSwapCache() check and we're now assuming
that !swapbacked is not in the swapcache, can you modify this to check
PageSwapBacked(page) != PageSwapCache(page)?

Better yet, change it into a warning and SWAP_FAIL.

^ permalink raw reply	[flat|nested] 11+ messages in thread

* Re: + mm-reclaim-madv_free-pages.patch added to -mm tree
  2017-03-03  2:52 ` Minchan Kim
  2017-03-03 15:18   ` Johannes Weiner
@ 2017-03-03 16:02   ` Shaohua Li
  1 sibling, 0 replies; 11+ messages in thread
From: Shaohua Li @ 2017-03-03 16:02 UTC (permalink / raw)
  To: Minchan Kim
  Cc: linux-kernel, hannes, hillf.zj, hughd, mgorman, mhocko, riel, mm-commits

On Fri, Mar 03, 2017 at 11:52:37AM +0900, Minchan Kim wrote:
> Hi,
> 
> On Tue, Feb 28, 2017 at 04:32:38PM -0800, akpm@linux-foundation.org wrote:
> > 
> > The patch titled
> >      Subject: mm: reclaim MADV_FREE pages
> > has been added to the -mm tree.  Its filename is
> >      mm-reclaim-madv_free-pages.patch
> > 
> > This patch should soon appear at
> >     https://urldefense.proofpoint.com/v2/url?u=http-3A__ozlabs.org_-7Eakpm_mmots_broken-2Dout_mm-2Dreclaim-2Dmadv-5Ffree-2Dpages.patch&d=DwIBAg&c=5VD0RTtNlTh3ycd41b3MUw&r=X13hAPkxmvBro1Ug8vcKHw&m=oMKTke-r00qQPnpEwo0Cn43po669gckvribhR9LQWf0&s=wzX_7IDavzBkFEMl7HDTMwrFXo1skB35mu0CjZmfFOg&e= 
> > and later at
> >     https://urldefense.proofpoint.com/v2/url?u=http-3A__ozlabs.org_-7Eakpm_mmotm_broken-2Dout_mm-2Dreclaim-2Dmadv-5Ffree-2Dpages.patch&d=DwIBAg&c=5VD0RTtNlTh3ycd41b3MUw&r=X13hAPkxmvBro1Ug8vcKHw&m=oMKTke-r00qQPnpEwo0Cn43po669gckvribhR9LQWf0&s=eUYtR1qeasKBGvuNWsZP6jE4XwoKwMb4CdKQU7OXGYM&e= 
> > 
> > Before you just go and hit "reply", please:
> >    a) Consider who else should be cc'ed
> >    b) Prefer to cc a suitable mailing list as well
> >    c) Ideally: find the original patch on the mailing list and do a
> >       reply-to-all to that, adding suitable additional cc's
> > 
> > *** Remember to use Documentation/SubmitChecklist when testing your code ***
> > 
> > The -mm tree is included into linux-next and is updated
> > there every 3-4 working days
> > 
> > ------------------------------------------------------
> > From: Shaohua Li <shli@fb.com>
> > Subject: mm: reclaim MADV_FREE pages
> > 
> > When memory pressure is high, we free MADV_FREE pages.  If the pages are
> > not dirty in pte, the pages could be freed immediately.  Otherwise we
> > can't reclaim them.  We put the pages back to anonumous LRU list (by
> > setting SwapBacked flag) and the pages will be reclaimed in normal swapout
> > way.
> > 
> > We use normal page reclaim policy.  Since MADV_FREE pages are put into
> > inactive file list, such pages and inactive file pages are reclaimed
> > according to their age.  This is expected, because we don't want to
> > reclaim too many MADV_FREE pages before used once pages.
> > 
> > Based on Minchan's original patch
> > 
> > Link: http://lkml.kernel.org/r/14b8eb1d3f6bf6cc492833f183ac8c304e560484.1487965799.git.shli@fb.com
> > Signed-off-by: Shaohua Li <shli@fb.com>
> > Acked-by: Minchan Kim <minchan@kernel.org>
> > Acked-by: Michal Hocko <mhocko@suse.com>
> > Acked-by: Johannes Weiner <hannes@cmpxchg.org>
> > Acked-by: Hillf Danton <hillf.zj@alibaba-inc.com>
> > Cc: Hugh Dickins <hughd@google.com>
> > Cc: Rik van Riel <riel@redhat.com>
> > Cc: Mel Gorman <mgorman@techsingularity.net>
> > Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
> > ---
> 
> < snip >
> 
> > @@ -1419,11 +1413,21 @@ static int try_to_unmap_one(struct page
> >  			VM_BUG_ON_PAGE(!PageSwapCache(page) && PageSwapBacked(page),
> >  				page);
> >  
> > -			if (!PageDirty(page)) {
> > +			/*
> > +			 * swapin page could be clean, it has data stored in
> > +			 * swap. We can't silently discard it without setting
> > +			 * swap entry in the page table.
> > +			 */
> > +			if (!PageDirty(page) && !PageSwapCache(page)) {
> >  				/* It's a freeable page by MADV_FREE */
> >  				dec_mm_counter(mm, MM_ANONPAGES);
> > -				rp->lazyfreed++;
> >  				goto discard;
> > +			} else if (!PageSwapBacked(page)) {
> > +				/* dirty MADV_FREE page */
> > +				set_pte_at(mm, address, pvmw.pte, pteval);
> > +				ret = SWAP_DIRTY;
> > +				page_vma_mapped_walk_done(&pvmw);
> > +				break;
> >  			}
> 
> There is no point to make this logic complicated with clean swapin-page.
> 
> Andrew,
> Could you fold below patch into the mm-reclaim-madv_free-pages.patch
> if others are not against?
> 
> Thanks.
> 
> From 0c28f6560fbc4e65da4f4a8cc4664ab9f7b11cf3 Mon Sep 17 00:00:00 2001
> From: Minchan Kim <minchan@kernel.org>
> Date: Fri, 3 Mar 2017 11:42:52 +0900
> Subject: [PATCH] mm: clean up lazyfree page handling
> 
> We can make it simple to understand without need to be aware of
> clean-swapin page.
> This patch just clean up lazyfree page handling in try_to_unmap_one.

Looks good, thanks!

Reviewed-by: Shaohua Li <shli@fb.com>
 
> Signed-off-by: Minchan Kim <minchan@kernel.org>
> ---
>  mm/rmap.c | 22 +++++++++++-----------
>  1 file changed, 11 insertions(+), 11 deletions(-)
> 
> diff --git a/mm/rmap.c b/mm/rmap.c
> index bb45712..f7eab40 100644
> --- a/mm/rmap.c
> +++ b/mm/rmap.c
> @@ -1413,17 +1413,17 @@ static int try_to_unmap_one(struct page *page, struct vm_area_struct *vma,
>  			VM_BUG_ON_PAGE(!PageSwapCache(page) && PageSwapBacked(page),
>  				page);
>  
> -			/*
> -			 * swapin page could be clean, it has data stored in
> -			 * swap. We can't silently discard it without setting
> -			 * swap entry in the page table.
> -			 */
> -			if (!PageDirty(page) && !PageSwapCache(page)) {
> -				/* It's a freeable page by MADV_FREE */
> -				dec_mm_counter(mm, MM_ANONPAGES);
> -				goto discard;
> -			} else if (!PageSwapBacked(page)) {
> -				/* dirty MADV_FREE page */
> +			/* MADV_FREE page check */
> +			if (!PageSwapBacked(page)) {
> +				if (!PageDirty(page)) {
> +					dec_mm_counter(mm, MM_ANONPAGES);
> +					goto discard;
> +				}
> +
> +				/*
> +				 * If the page was redirtied, it cannot be
> +				 * discarded. Remap the page to page table.
> +				 */
>  				set_pte_at(mm, address, pvmw.pte, pteval);
>  				ret = SWAP_DIRTY;
>  				page_vma_mapped_walk_done(&pvmw);
> -- 
> 2.7.4
> 

^ permalink raw reply	[flat|nested] 11+ messages in thread

* Re: + mm-reclaim-madv_free-pages.patch added to -mm tree
  2017-03-03 15:18   ` Johannes Weiner
@ 2017-03-06  3:03     ` Minchan Kim
  2017-03-06 15:49       ` Johannes Weiner
  0 siblings, 1 reply; 11+ messages in thread
From: Minchan Kim @ 2017-03-06  3:03 UTC (permalink / raw)
  To: Johannes Weiner
  Cc: linux-kernel, shli, hillf.zj, hughd, mgorman, mhocko, riel, mm-commits

On Fri, Mar 03, 2017 at 10:18:51AM -0500, Johannes Weiner wrote:
> On Fri, Mar 03, 2017 at 11:52:37AM +0900, Minchan Kim wrote:
> > On Tue, Feb 28, 2017 at 04:32:38PM -0800, akpm@linux-foundation.org wrote:
> > > 
> > > The patch titled
> > >      Subject: mm: reclaim MADV_FREE pages
> > > has been added to the -mm tree.  Its filename is
> > >      mm-reclaim-madv_free-pages.patch
> > > 
> > > This patch should soon appear at
> > >     http://ozlabs.org/~akpm/mmots/broken-out/mm-reclaim-madv_free-pages.patch
> > > and later at
> > >     http://ozlabs.org/~akpm/mmotm/broken-out/mm-reclaim-madv_free-pages.patch
> > > 
> > > Before you just go and hit "reply", please:
> > >    a) Consider who else should be cc'ed
> > >    b) Prefer to cc a suitable mailing list as well
> > >    c) Ideally: find the original patch on the mailing list and do a
> > >       reply-to-all to that, adding suitable additional cc's
> > > 
> > > *** Remember to use Documentation/SubmitChecklist when testing your code ***
> > > 
> > > The -mm tree is included into linux-next and is updated
> > > there every 3-4 working days
> > > 
> > > ------------------------------------------------------
> > > From: Shaohua Li <shli@fb.com>
> > > Subject: mm: reclaim MADV_FREE pages
> > > 
> > > When memory pressure is high, we free MADV_FREE pages.  If the pages are
> > > not dirty in pte, the pages could be freed immediately.  Otherwise we
> > > can't reclaim them.  We put the pages back to anonumous LRU list (by
> > > setting SwapBacked flag) and the pages will be reclaimed in normal swapout
> > > way.
> > > 
> > > We use normal page reclaim policy.  Since MADV_FREE pages are put into
> > > inactive file list, such pages and inactive file pages are reclaimed
> > > according to their age.  This is expected, because we don't want to
> > > reclaim too many MADV_FREE pages before used once pages.
> > > 
> > > Based on Minchan's original patch
> > > 
> > > Link: http://lkml.kernel.org/r/14b8eb1d3f6bf6cc492833f183ac8c304e560484.1487965799.git.shli@fb.com
> > > Signed-off-by: Shaohua Li <shli@fb.com>
> > > Acked-by: Minchan Kim <minchan@kernel.org>
> > > Acked-by: Michal Hocko <mhocko@suse.com>
> > > Acked-by: Johannes Weiner <hannes@cmpxchg.org>
> > > Acked-by: Hillf Danton <hillf.zj@alibaba-inc.com>
> > > Cc: Hugh Dickins <hughd@google.com>
> > > Cc: Rik van Riel <riel@redhat.com>
> > > Cc: Mel Gorman <mgorman@techsingularity.net>
> > > Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
> > > ---
> > 
> > < snip >
> > 
> > > @@ -1419,11 +1413,21 @@ static int try_to_unmap_one(struct page
> > >  			VM_BUG_ON_PAGE(!PageSwapCache(page) && PageSwapBacked(page),
> > >  				page);
> > >  
> > > -			if (!PageDirty(page)) {
> > > +			/*
> > > +			 * swapin page could be clean, it has data stored in
> > > +			 * swap. We can't silently discard it without setting
> > > +			 * swap entry in the page table.
> > > +			 */
> > > +			if (!PageDirty(page) && !PageSwapCache(page)) {
> > >  				/* It's a freeable page by MADV_FREE */
> > >  				dec_mm_counter(mm, MM_ANONPAGES);
> > > -				rp->lazyfreed++;
> > >  				goto discard;
> > > +			} else if (!PageSwapBacked(page)) {
> > > +				/* dirty MADV_FREE page */
> > > +				set_pte_at(mm, address, pvmw.pte, pteval);
> > > +				ret = SWAP_DIRTY;
> > > +				page_vma_mapped_walk_done(&pvmw);
> > > +				break;
> > >  			}
> > 
> > There is no point to make this logic complicated with clean swapin-page.
> > 
> > Andrew,
> > Could you fold below patch into the mm-reclaim-madv_free-pages.patch
> > if others are not against?
> > 
> > Thanks.
> > 
> > From 0c28f6560fbc4e65da4f4a8cc4664ab9f7b11cf3 Mon Sep 17 00:00:00 2001
> > From: Minchan Kim <minchan@kernel.org>
> > Date: Fri, 3 Mar 2017 11:42:52 +0900
> > Subject: [PATCH] mm: clean up lazyfree page handling
> > 
> > We can make it simple to understand without need to be aware of
> > clean-swapin page.
> > This patch just clean up lazyfree page handling in try_to_unmap_one.
> > 
> > Signed-off-by: Minchan Kim <minchan@kernel.org>
> 
> Agreed, this is a litle easier to follow.
> 
> Acked-by: Johannes Weiner <hannes@cmpxchg.org>

Thanks, Johannes.

> 
> > ---
> >  mm/rmap.c | 22 +++++++++++-----------
> >  1 file changed, 11 insertions(+), 11 deletions(-)
> > 
> > diff --git a/mm/rmap.c b/mm/rmap.c
> > index bb45712..f7eab40 100644
> > --- a/mm/rmap.c
> > +++ b/mm/rmap.c
> > @@ -1413,17 +1413,17 @@ static int try_to_unmap_one(struct page *page, struct vm_area_struct *vma,
> >  			VM_BUG_ON_PAGE(!PageSwapCache(page) && PageSwapBacked(page),
> >  				page);
> 
> Since you're removing the PageSwapCache() check and we're now assuming
> that !swapbacked is not in the swapcache, can you modify this to check
> PageSwapBacked(page) != PageSwapCache(page)?
> 
> Better yet, change it into a warning and SWAP_FAIL.

Maybe, what you wanted is

 !!PageSwapBacked(page) != !!PageSwapCache(page)

Personally, I prefer && style rather than equation expression
in this case.

How about this?
If others are not against, I will resend it to Andrew with
Acked|Reviewed-by all I got until now.

Thanks.

commit 118cfee42600
Author: Minchan Kim <minchan@kernel.org>
Date:   Sat Mar 4 01:01:38 2017 +0000

    mm: clean up lazyfree page handling
    
    We can make it simple to understand without need to be aware of
    clean-swapin page.
    This patch just clean up lazyfree page handling in try_to_unmap_one.
    
    Link: http://lkml.kernel.org/r/20170303025237.GB3503@bbox
    Signed-off-by: Minchan Kim <minchan@kernel.org>
    Cc: Shaohua Li <shli@fb.com>
    Cc: Michal Hocko <mhocko@suse.com>
    Cc: Johannes Weiner <hannes@cmpxchg.org>
    Cc: Hillf Danton <hillf.zj@alibaba-inc.com>
    Cc: Hugh Dickins <hughd@google.com>
    Cc: Rik van Riel <riel@redhat.com>
    Cc: Mel Gorman <mgorman@techsingularity.net>
    Signed-off-by: Andrew Morton <akpm@linux-foundation.org>

diff --git a/mm/rmap.c b/mm/rmap.c
index 3d86036d96ec..1377f7b0361e 100644
--- a/mm/rmap.c
+++ b/mm/rmap.c
@@ -1413,20 +1413,24 @@ static int try_to_unmap_one(struct page *page, struct vm_area_struct *vma,
 			 * Store the swap location in the pte.
 			 * See handle_pte_fault() ...
 			 */
-			VM_BUG_ON_PAGE(!PageSwapCache(page) && PageSwapBacked(page),
-				page);
+			if (VM_WARN_ON_ONCE(PageSwapBacked(page) &&
+						!PageSwapCache(page))) {
+				ret = SWAP_FAIL;
+				page_vma_mapped_walk_done(&pvmw);
+				break;
+			}
 
-			/*
-			 * swapin page could be clean, it has data stored in
-			 * swap. We can't silently discard it without setting
-			 * swap entry in the page table.
-			 */
-			if (!PageDirty(page) && !PageSwapCache(page)) {
-				/* It's a freeable page by MADV_FREE */
-				dec_mm_counter(mm, MM_ANONPAGES);
-				goto discard;
-			} else if (!PageSwapBacked(page)) {
-				/* dirty MADV_FREE page */
+			/* MADV_FREE page check */
+			if (!PageSwapBacked(page)) {
+				if (!PageDirty(page)) {
+					dec_mm_counter(mm, MM_ANONPAGES);
+					goto discard;
+				}
+
+				/*
+				 * If the page was redirtied, it cannot be
+				 * discarded. Remap the page to page table.
+				 */
 				set_pte_at(mm, address, pvmw.pte, pteval);
 				ret = SWAP_DIRTY;
 				page_vma_mapped_walk_done(&pvmw);

^ permalink raw reply related	[flat|nested] 11+ messages in thread

* Re: + mm-reclaim-madv_free-pages.patch added to -mm tree
  2017-03-06  3:03     ` Minchan Kim
@ 2017-03-06 15:49       ` Johannes Weiner
  2017-03-07  5:46         ` Minchan Kim
  2017-03-07  5:55         ` Minchan Kim
  0 siblings, 2 replies; 11+ messages in thread
From: Johannes Weiner @ 2017-03-06 15:49 UTC (permalink / raw)
  To: Minchan Kim
  Cc: linux-kernel, shli, hillf.zj, hughd, mgorman, mhocko, riel, mm-commits

On Mon, Mar 06, 2017 at 12:03:44PM +0900, Minchan Kim wrote:
> On Fri, Mar 03, 2017 at 10:18:51AM -0500, Johannes Weiner wrote:
> > On Fri, Mar 03, 2017 at 11:52:37AM +0900, Minchan Kim wrote:
> > > On Tue, Feb 28, 2017 at 04:32:38PM -0800, akpm@linux-foundation.org wrote:
> > > > 
> > > > The patch titled
> > > >      Subject: mm: reclaim MADV_FREE pages
> > > > has been added to the -mm tree.  Its filename is
> > > >      mm-reclaim-madv_free-pages.patch
> > > > 
> > > > This patch should soon appear at
> > > >     http://ozlabs.org/~akpm/mmots/broken-out/mm-reclaim-madv_free-pages.patch
> > > > and later at
> > > >     http://ozlabs.org/~akpm/mmotm/broken-out/mm-reclaim-madv_free-pages.patch
> > > > 
> > > > Before you just go and hit "reply", please:
> > > >    a) Consider who else should be cc'ed
> > > >    b) Prefer to cc a suitable mailing list as well
> > > >    c) Ideally: find the original patch on the mailing list and do a
> > > >       reply-to-all to that, adding suitable additional cc's
> > > > 
> > > > *** Remember to use Documentation/SubmitChecklist when testing your code ***
> > > > 
> > > > The -mm tree is included into linux-next and is updated
> > > > there every 3-4 working days
> > > > 
> > > > ------------------------------------------------------
> > > > From: Shaohua Li <shli@fb.com>
> > > > Subject: mm: reclaim MADV_FREE pages
> > > > 
> > > > When memory pressure is high, we free MADV_FREE pages.  If the pages are
> > > > not dirty in pte, the pages could be freed immediately.  Otherwise we
> > > > can't reclaim them.  We put the pages back to anonumous LRU list (by
> > > > setting SwapBacked flag) and the pages will be reclaimed in normal swapout
> > > > way.
> > > > 
> > > > We use normal page reclaim policy.  Since MADV_FREE pages are put into
> > > > inactive file list, such pages and inactive file pages are reclaimed
> > > > according to their age.  This is expected, because we don't want to
> > > > reclaim too many MADV_FREE pages before used once pages.
> > > > 
> > > > Based on Minchan's original patch
> > > > 
> > > > Link: http://lkml.kernel.org/r/14b8eb1d3f6bf6cc492833f183ac8c304e560484.1487965799.git.shli@fb.com
> > > > Signed-off-by: Shaohua Li <shli@fb.com>
> > > > Acked-by: Minchan Kim <minchan@kernel.org>
> > > > Acked-by: Michal Hocko <mhocko@suse.com>
> > > > Acked-by: Johannes Weiner <hannes@cmpxchg.org>
> > > > Acked-by: Hillf Danton <hillf.zj@alibaba-inc.com>
> > > > Cc: Hugh Dickins <hughd@google.com>
> > > > Cc: Rik van Riel <riel@redhat.com>
> > > > Cc: Mel Gorman <mgorman@techsingularity.net>
> > > > Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
> > > > ---
> > > 
> > > < snip >
> > > 
> > > > @@ -1419,11 +1413,21 @@ static int try_to_unmap_one(struct page
> > > >  			VM_BUG_ON_PAGE(!PageSwapCache(page) && PageSwapBacked(page),
> > > >  				page);
> > > >  
> > > > -			if (!PageDirty(page)) {
> > > > +			/*
> > > > +			 * swapin page could be clean, it has data stored in
> > > > +			 * swap. We can't silently discard it without setting
> > > > +			 * swap entry in the page table.
> > > > +			 */
> > > > +			if (!PageDirty(page) && !PageSwapCache(page)) {
> > > >  				/* It's a freeable page by MADV_FREE */
> > > >  				dec_mm_counter(mm, MM_ANONPAGES);
> > > > -				rp->lazyfreed++;
> > > >  				goto discard;
> > > > +			} else if (!PageSwapBacked(page)) {
> > > > +				/* dirty MADV_FREE page */
> > > > +				set_pte_at(mm, address, pvmw.pte, pteval);
> > > > +				ret = SWAP_DIRTY;
> > > > +				page_vma_mapped_walk_done(&pvmw);
> > > > +				break;
> > > >  			}
> > > 
> > > There is no point to make this logic complicated with clean swapin-page.
> > > 
> > > Andrew,
> > > Could you fold below patch into the mm-reclaim-madv_free-pages.patch
> > > if others are not against?
> > > 
> > > Thanks.
> > > 
> > > From 0c28f6560fbc4e65da4f4a8cc4664ab9f7b11cf3 Mon Sep 17 00:00:00 2001
> > > From: Minchan Kim <minchan@kernel.org>
> > > Date: Fri, 3 Mar 2017 11:42:52 +0900
> > > Subject: [PATCH] mm: clean up lazyfree page handling
> > > 
> > > We can make it simple to understand without need to be aware of
> > > clean-swapin page.
> > > This patch just clean up lazyfree page handling in try_to_unmap_one.
> > > 
> > > Signed-off-by: Minchan Kim <minchan@kernel.org>
> > 
> > Agreed, this is a litle easier to follow.
> > 
> > Acked-by: Johannes Weiner <hannes@cmpxchg.org>
> 
> Thanks, Johannes.
> 
> > 
> > > ---
> > >  mm/rmap.c | 22 +++++++++++-----------
> > >  1 file changed, 11 insertions(+), 11 deletions(-)
> > > 
> > > diff --git a/mm/rmap.c b/mm/rmap.c
> > > index bb45712..f7eab40 100644
> > > --- a/mm/rmap.c
> > > +++ b/mm/rmap.c
> > > @@ -1413,17 +1413,17 @@ static int try_to_unmap_one(struct page *page, struct vm_area_struct *vma,
> > >  			VM_BUG_ON_PAGE(!PageSwapCache(page) && PageSwapBacked(page),
> > >  				page);
> > 
> > Since you're removing the PageSwapCache() check and we're now assuming
> > that !swapbacked is not in the swapcache, can you modify this to check
> > PageSwapBacked(page) != PageSwapCache(page)?
> > 
> > Better yet, change it into a warning and SWAP_FAIL.
> 
> Maybe, what you wanted is
> 
>  !!PageSwapBacked(page) != !!PageSwapCache(page)

Those testers return 0 or 1 ints, on x86 even bool.

> Personally, I prefer && style rather than equation expression
> in this case.

> @@ -1413,20 +1413,24 @@ static int try_to_unmap_one(struct page *page, struct vm_area_struct *vma,
>  			 * Store the swap location in the pte.
>  			 * See handle_pte_fault() ...
>  			 */
> -			VM_BUG_ON_PAGE(!PageSwapCache(page) && PageSwapBacked(page),
> -				page);
> +			if (VM_WARN_ON_ONCE(PageSwapBacked(page) &&
> +						!PageSwapCache(page))) {
> +				ret = SWAP_FAIL;

But you're not adding the !swapbacked && swapcache case?

> +				page_vma_mapped_walk_done(&pvmw);
> +				break;
> +			}

[...]

> -			/*
> -			 * swapin page could be clean, it has data stored in
> -			 * swap. We can't silently discard it without setting
> -			 * swap entry in the page table.
> -			 */
> -			if (!PageDirty(page) && !PageSwapCache(page)) {
> -				/* It's a freeable page by MADV_FREE */
> -				dec_mm_counter(mm, MM_ANONPAGES);
> -				goto discard;
> -			} else if (!PageSwapBacked(page)) {
> -				/* dirty MADV_FREE page */
> +			/* MADV_FREE page check */
> +			if (!PageSwapBacked(page)) {
> +				if (!PageDirty(page)) {
> +					dec_mm_counter(mm, MM_ANONPAGES);
> +					goto discard;
> +				}

Andrew already has this, you might want to send the warning changes as
a separate patch on top of this one.

^ permalink raw reply	[flat|nested] 11+ messages in thread

* Re: + mm-reclaim-madv_free-pages.patch added to -mm tree
  2017-03-06 15:49       ` Johannes Weiner
@ 2017-03-07  5:46         ` Minchan Kim
  2017-03-07  5:55         ` Minchan Kim
  1 sibling, 0 replies; 11+ messages in thread
From: Minchan Kim @ 2017-03-07  5:46 UTC (permalink / raw)
  To: Johannes Weiner
  Cc: linux-kernel, shli, hillf.zj, hughd, mgorman, mhocko, riel, mm-commits

On Mon, Mar 06, 2017 at 10:49:06AM -0500, Johannes Weiner wrote:
> On Mon, Mar 06, 2017 at 12:03:44PM +0900, Minchan Kim wrote:
> > On Fri, Mar 03, 2017 at 10:18:51AM -0500, Johannes Weiner wrote:
> > > On Fri, Mar 03, 2017 at 11:52:37AM +0900, Minchan Kim wrote:
> > > > On Tue, Feb 28, 2017 at 04:32:38PM -0800, akpm@linux-foundation.org wrote:
> > > > > 
> > > > > The patch titled
> > > > >      Subject: mm: reclaim MADV_FREE pages
> > > > > has been added to the -mm tree.  Its filename is
> > > > >      mm-reclaim-madv_free-pages.patch
> > > > > 
> > > > > This patch should soon appear at
> > > > >     http://ozlabs.org/~akpm/mmots/broken-out/mm-reclaim-madv_free-pages.patch
> > > > > and later at
> > > > >     http://ozlabs.org/~akpm/mmotm/broken-out/mm-reclaim-madv_free-pages.patch
> > > > > 
> > > > > Before you just go and hit "reply", please:
> > > > >    a) Consider who else should be cc'ed
> > > > >    b) Prefer to cc a suitable mailing list as well
> > > > >    c) Ideally: find the original patch on the mailing list and do a
> > > > >       reply-to-all to that, adding suitable additional cc's
> > > > > 
> > > > > *** Remember to use Documentation/SubmitChecklist when testing your code ***
> > > > > 
> > > > > The -mm tree is included into linux-next and is updated
> > > > > there every 3-4 working days
> > > > > 
> > > > > ------------------------------------------------------
> > > > > From: Shaohua Li <shli@fb.com>
> > > > > Subject: mm: reclaim MADV_FREE pages
> > > > > 
> > > > > When memory pressure is high, we free MADV_FREE pages.  If the pages are
> > > > > not dirty in pte, the pages could be freed immediately.  Otherwise we
> > > > > can't reclaim them.  We put the pages back to anonumous LRU list (by
> > > > > setting SwapBacked flag) and the pages will be reclaimed in normal swapout
> > > > > way.
> > > > > 
> > > > > We use normal page reclaim policy.  Since MADV_FREE pages are put into
> > > > > inactive file list, such pages and inactive file pages are reclaimed
> > > > > according to their age.  This is expected, because we don't want to
> > > > > reclaim too many MADV_FREE pages before used once pages.
> > > > > 
> > > > > Based on Minchan's original patch
> > > > > 
> > > > > Link: http://lkml.kernel.org/r/14b8eb1d3f6bf6cc492833f183ac8c304e560484.1487965799.git.shli@fb.com
> > > > > Signed-off-by: Shaohua Li <shli@fb.com>
> > > > > Acked-by: Minchan Kim <minchan@kernel.org>
> > > > > Acked-by: Michal Hocko <mhocko@suse.com>
> > > > > Acked-by: Johannes Weiner <hannes@cmpxchg.org>
> > > > > Acked-by: Hillf Danton <hillf.zj@alibaba-inc.com>
> > > > > Cc: Hugh Dickins <hughd@google.com>
> > > > > Cc: Rik van Riel <riel@redhat.com>
> > > > > Cc: Mel Gorman <mgorman@techsingularity.net>
> > > > > Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
> > > > > ---
> > > > 
> > > > < snip >
> > > > 
> > > > > @@ -1419,11 +1413,21 @@ static int try_to_unmap_one(struct page
> > > > >  			VM_BUG_ON_PAGE(!PageSwapCache(page) && PageSwapBacked(page),
> > > > >  				page);
> > > > >  
> > > > > -			if (!PageDirty(page)) {
> > > > > +			/*
> > > > > +			 * swapin page could be clean, it has data stored in
> > > > > +			 * swap. We can't silently discard it without setting
> > > > > +			 * swap entry in the page table.
> > > > > +			 */
> > > > > +			if (!PageDirty(page) && !PageSwapCache(page)) {
> > > > >  				/* It's a freeable page by MADV_FREE */
> > > > >  				dec_mm_counter(mm, MM_ANONPAGES);
> > > > > -				rp->lazyfreed++;
> > > > >  				goto discard;
> > > > > +			} else if (!PageSwapBacked(page)) {
> > > > > +				/* dirty MADV_FREE page */
> > > > > +				set_pte_at(mm, address, pvmw.pte, pteval);
> > > > > +				ret = SWAP_DIRTY;
> > > > > +				page_vma_mapped_walk_done(&pvmw);
> > > > > +				break;
> > > > >  			}
> > > > 
> > > > There is no point to make this logic complicated with clean swapin-page.
> > > > 
> > > > Andrew,
> > > > Could you fold below patch into the mm-reclaim-madv_free-pages.patch
> > > > if others are not against?
> > > > 
> > > > Thanks.
> > > > 
> > > > From 0c28f6560fbc4e65da4f4a8cc4664ab9f7b11cf3 Mon Sep 17 00:00:00 2001
> > > > From: Minchan Kim <minchan@kernel.org>
> > > > Date: Fri, 3 Mar 2017 11:42:52 +0900
> > > > Subject: [PATCH] mm: clean up lazyfree page handling
> > > > 
> > > > We can make it simple to understand without need to be aware of
> > > > clean-swapin page.
> > > > This patch just clean up lazyfree page handling in try_to_unmap_one.
> > > > 
> > > > Signed-off-by: Minchan Kim <minchan@kernel.org>
> > > 
> > > Agreed, this is a litle easier to follow.
> > > 
> > > Acked-by: Johannes Weiner <hannes@cmpxchg.org>
> > 
> > Thanks, Johannes.
> > 
> > > 
> > > > ---
> > > >  mm/rmap.c | 22 +++++++++++-----------
> > > >  1 file changed, 11 insertions(+), 11 deletions(-)
> > > > 
> > > > diff --git a/mm/rmap.c b/mm/rmap.c
> > > > index bb45712..f7eab40 100644
> > > > --- a/mm/rmap.c
> > > > +++ b/mm/rmap.c
> > > > @@ -1413,17 +1413,17 @@ static int try_to_unmap_one(struct page *page, struct vm_area_struct *vma,
> > > >  			VM_BUG_ON_PAGE(!PageSwapCache(page) && PageSwapBacked(page),
> > > >  				page);
> > > 
> > > Since you're removing the PageSwapCache() check and we're now assuming
> > > that !swapbacked is not in the swapcache, can you modify this to check
> > > PageSwapBacked(page) != PageSwapCache(page)?
> > > 
> > > Better yet, change it into a warning and SWAP_FAIL.
> > 
> > Maybe, what you wanted is
> > 
> >  !!PageSwapBacked(page) != !!PageSwapCache(page)
> 
> Those testers return 0 or 1 ints, on x86 even bool.

You're right.
It seems PageXXX() was not in old day. Hmm, but seem to have corrected.

> 
> > Personally, I prefer && style rather than equation expression
> > in this case.
> 
> > @@ -1413,20 +1413,24 @@ static int try_to_unmap_one(struct page *page, struct vm_area_struct *vma,
> >  			 * Store the swap location in the pte.
> >  			 * See handle_pte_fault() ...
> >  			 */
> > -			VM_BUG_ON_PAGE(!PageSwapCache(page) && PageSwapBacked(page),
> > -				page);
> > +			if (VM_WARN_ON_ONCE(PageSwapBacked(page) &&
> > +						!PageSwapCache(page))) {
> > +				ret = SWAP_FAIL;
> 
> But you're not adding the !swapbacked && swapcache case?

PageSwapcache checks already includes PageSwapBacked check so practically,
it would be pointless at this moment. But yes, it would be bad on relying
on the current implemenation as considering the future.

As a bonus from side-effect, seeing the warning always only means
it is from PageSwapbacked && !swapcache so we don't need to introduce
VM_WARN_ON_ONCE_PAGE at this moment. :)

I will resend it.

Thanks for the review!

^ permalink raw reply	[flat|nested] 11+ messages in thread

* Re: + mm-reclaim-madv_free-pages.patch added to -mm tree
  2017-03-06 15:49       ` Johannes Weiner
  2017-03-07  5:46         ` Minchan Kim
@ 2017-03-07  5:55         ` Minchan Kim
  2017-03-07  9:58           ` Michal Hocko
  2017-03-07 16:59           ` Johannes Weiner
  1 sibling, 2 replies; 11+ messages in thread
From: Minchan Kim @ 2017-03-07  5:55 UTC (permalink / raw)
  To: Johannes Weiner, Andrew Morton
  Cc: linux-kernel, shli, hillf.zj, hughd, mgorman, mhocko, riel, mm-commits

On Mon, Mar 06, 2017 at 10:49:06AM -0500, Johannes Weiner wrote:

< snip >

> > @@ -1413,20 +1413,24 @@ static int try_to_unmap_one(struct page *page, struct vm_area_struct *vma,
> >  			 * Store the swap location in the pte.
> >  			 * See handle_pte_fault() ...
> >  			 */
> > -			VM_BUG_ON_PAGE(!PageSwapCache(page) && PageSwapBacked(page),
> > -				page);
> > +			if (VM_WARN_ON_ONCE(PageSwapBacked(page) &&
> > +						!PageSwapCache(page))) {
> > +				ret = SWAP_FAIL;
> 
> But you're not adding the !swapbacked && swapcache case?
> 
> > +				page_vma_mapped_walk_done(&pvmw);
> > +				break;
> > +			}
> 
> [...]
> 
> > -			/*
> > -			 * swapin page could be clean, it has data stored in
> > -			 * swap. We can't silently discard it without setting
> > -			 * swap entry in the page table.
> > -			 */
> > -			if (!PageDirty(page) && !PageSwapCache(page)) {
> > -				/* It's a freeable page by MADV_FREE */
> > -				dec_mm_counter(mm, MM_ANONPAGES);
> > -				goto discard;
> > -			} else if (!PageSwapBacked(page)) {
> > -				/* dirty MADV_FREE page */
> > +			/* MADV_FREE page check */
> > +			if (!PageSwapBacked(page)) {
> > +				if (!PageDirty(page)) {
> > +					dec_mm_counter(mm, MM_ANONPAGES);
> > +					goto discard;
> > +				}
> 
> Andrew already has this, you might want to send the warning changes as
> a separate patch on top of this one.

Here it goes.

>From d42d296950c3bbce74afddcff307fa18eef305fe Mon Sep 17 00:00:00 2001
From: Minchan Kim <minchan@kernel.org>
Date: Tue, 7 Mar 2017 14:48:37 +0900
Subject: [PATCH] mm: fix lazyfree bug on check in try_to_unmap_one

If a page is swapbacked, it means it should be in swapcache
in try_to_unmap_one's path.

If a page is !swapbacked, it mean it shouldn't be in swapcache
in try_to_unmap_one's path.

Check both two cases all at once and if it fails, warn and
return SWAP_FAIL. Such bug never mean we should shut down
the kernel.

Suggested-by: Johannes Weiner <hannes@cmpxchg.org>
Signed-off-by: Minchan Kim <minchan@kernel.org>
---
 mm/rmap.c | 9 +++++++--
 1 file changed, 7 insertions(+), 2 deletions(-)

diff --git a/mm/rmap.c b/mm/rmap.c
index 35acb83..9925f32 100644
--- a/mm/rmap.c
+++ b/mm/rmap.c
@@ -1413,8 +1413,13 @@ static int try_to_unmap_one(struct page *page, struct vm_area_struct *vma,
 			 * Store the swap location in the pte.
 			 * See handle_pte_fault() ...
 			 */
-			VM_BUG_ON_PAGE(!PageSwapCache(page) && PageSwapBacked(page),
-				page);
+			if (VM_WARN_ON_ONCE(PageSwapBacked(page) !=
+						PageSwapCache(page))) {
+				ret = SWAP_FAIL;
+				page_vma_mapped_walk_done(&pvmw);
+				break;
+
+			}
 
 			/* MADV_FREE page check */
 			if (!PageSwapBacked(page)) {
-- 
2.7.4

^ permalink raw reply related	[flat|nested] 11+ messages in thread

* Re: + mm-reclaim-madv_free-pages.patch added to -mm tree
  2017-03-07  5:55         ` Minchan Kim
@ 2017-03-07  9:58           ` Michal Hocko
  2017-03-07 16:59           ` Johannes Weiner
  1 sibling, 0 replies; 11+ messages in thread
From: Michal Hocko @ 2017-03-07  9:58 UTC (permalink / raw)
  To: Minchan Kim
  Cc: Johannes Weiner, Andrew Morton, linux-kernel, shli, hillf.zj,
	hughd, mgorman, riel, mm-commits

On Tue 07-03-17 14:55:51, Minchan Kim wrote:
[...]
> >From d42d296950c3bbce74afddcff307fa18eef305fe Mon Sep 17 00:00:00 2001
> From: Minchan Kim <minchan@kernel.org>
> Date: Tue, 7 Mar 2017 14:48:37 +0900
> Subject: [PATCH] mm: fix lazyfree bug on check in try_to_unmap_one
> 
> If a page is swapbacked, it means it should be in swapcache
> in try_to_unmap_one's path.
> 
> If a page is !swapbacked, it mean it shouldn't be in swapcache
> in try_to_unmap_one's path.
> 
> Check both two cases all at once and if it fails, warn and
> return SWAP_FAIL. Such bug never mean we should shut down
> the kernel.
> 
> Suggested-by: Johannes Weiner <hannes@cmpxchg.org>
> Signed-off-by: Minchan Kim <minchan@kernel.org>

looks good to me
Acked-by: Michal Hocko <mhocko@suse.com>

> ---
>  mm/rmap.c | 9 +++++++--
>  1 file changed, 7 insertions(+), 2 deletions(-)
> 
> diff --git a/mm/rmap.c b/mm/rmap.c
> index 35acb83..9925f32 100644
> --- a/mm/rmap.c
> +++ b/mm/rmap.c
> @@ -1413,8 +1413,13 @@ static int try_to_unmap_one(struct page *page, struct vm_area_struct *vma,
>  			 * Store the swap location in the pte.
>  			 * See handle_pte_fault() ...
>  			 */
> -			VM_BUG_ON_PAGE(!PageSwapCache(page) && PageSwapBacked(page),
> -				page);
> +			if (VM_WARN_ON_ONCE(PageSwapBacked(page) !=
> +						PageSwapCache(page))) {
> +				ret = SWAP_FAIL;
> +				page_vma_mapped_walk_done(&pvmw);
> +				break;
> +
> +			}
>  
>  			/* MADV_FREE page check */
>  			if (!PageSwapBacked(page)) {
> -- 
> 2.7.4
> 

-- 
Michal Hocko
SUSE Labs

^ permalink raw reply	[flat|nested] 11+ messages in thread

* Re: + mm-reclaim-madv_free-pages.patch added to -mm tree
  2017-03-07  5:55         ` Minchan Kim
  2017-03-07  9:58           ` Michal Hocko
@ 2017-03-07 16:59           ` Johannes Weiner
  1 sibling, 0 replies; 11+ messages in thread
From: Johannes Weiner @ 2017-03-07 16:59 UTC (permalink / raw)
  To: Minchan Kim
  Cc: Andrew Morton, linux-kernel, shli, hillf.zj, hughd, mgorman,
	mhocko, riel, mm-commits

On Tue, Mar 07, 2017 at 02:55:51PM +0900, Minchan Kim wrote:
> From d42d296950c3bbce74afddcff307fa18eef305fe Mon Sep 17 00:00:00 2001
> From: Minchan Kim <minchan@kernel.org>
> Date: Tue, 7 Mar 2017 14:48:37 +0900
> Subject: [PATCH] mm: fix lazyfree bug on check in try_to_unmap_one
> 
> If a page is swapbacked, it means it should be in swapcache
> in try_to_unmap_one's path.
> 
> If a page is !swapbacked, it mean it shouldn't be in swapcache
> in try_to_unmap_one's path.
> 
> Check both two cases all at once and if it fails, warn and
> return SWAP_FAIL. Such bug never mean we should shut down
> the kernel.
> 
> Suggested-by: Johannes Weiner <hannes@cmpxchg.org>
> Signed-off-by: Minchan Kim <minchan@kernel.org>

Acked-by: Johannes Weiner <hannes@cmpxchg.org>

Thanks Minchan

^ permalink raw reply	[flat|nested] 11+ messages in thread

* + mm-reclaim-madv_free-pages.patch added to -mm tree
@ 2017-03-07 22:57 akpm
  0 siblings, 0 replies; 11+ messages in thread
From: akpm @ 2017-03-07 22:57 UTC (permalink / raw)
  To: shli, hannes, hillf.zj, hughd, mgorman, mhocko, minchan, riel,
	mm-commits


The patch titled
     Subject: mm: reclaim MADV_FREE pages
has been added to the -mm tree.  Its filename is
     mm-reclaim-madv_free-pages.patch

This patch should soon appear at
    http://ozlabs.org/~akpm/mmots/broken-out/mm-reclaim-madv_free-pages.patch
and later at
    http://ozlabs.org/~akpm/mmotm/broken-out/mm-reclaim-madv_free-pages.patch

Before you just go and hit "reply", please:
   a) Consider who else should be cc'ed
   b) Prefer to cc a suitable mailing list as well
   c) Ideally: find the original patch on the mailing list and do a
      reply-to-all to that, adding suitable additional cc's

*** Remember to use Documentation/SubmitChecklist when testing your code ***

The -mm tree is included into linux-next and is updated
there every 3-4 working days

------------------------------------------------------
From: Shaohua Li <shli@fb.com>
Subject: mm: reclaim MADV_FREE pages

When memory pressure is high, we free MADV_FREE pages.  If the pages are
not dirty in pte, the pages could be freed immediately.  Otherwise we
can't reclaim them.  We put the pages back to anonumous LRU list (by
setting SwapBacked flag) and the pages will be reclaimed in normal swapout
way.

We use normal page reclaim policy.  Since MADV_FREE pages are put into
inactive file list, such pages and inactive file pages are reclaimed
according to their age.  This is expected, because we don't want to
reclaim too many MADV_FREE pages before used once pages.

Based on Minchan's original patch

Link: http://lkml.kernel.org/r/14b8eb1d3f6bf6cc492833f183ac8c304e560484.1487965799.git.shli@fb.com
Signed-off-by: Shaohua Li <shli@fb.com>
Acked-by: Minchan Kim <minchan@kernel.org>
Acked-by: Michal Hocko <mhocko@suse.com>
Acked-by: Johannes Weiner <hannes@cmpxchg.org>
Acked-by: Hillf Danton <hillf.zj@alibaba-inc.com>
Cc: Hugh Dickins <hughd@google.com>
Cc: Rik van Riel <riel@redhat.com>
Cc: Mel Gorman <mgorman@techsingularity.net>
Signed-off-by: Andrew Morton <>
---

 include/linux/rmap.h |    2 +-
 mm/huge_memory.c     |    2 ++
 mm/madvise.c         |    1 +
 mm/rmap.c            |   40 +++++++++++++++++-----------------------
 mm/vmscan.c          |   34 ++++++++++++++++++++++------------
 5 files changed, 43 insertions(+), 36 deletions(-)

diff -puN include/linux/rmap.h~mm-reclaim-madv_free-pages include/linux/rmap.h
--- a/include/linux/rmap.h~mm-reclaim-madv_free-pages
+++ a/include/linux/rmap.h
@@ -298,6 +298,6 @@ static inline int page_mkclean(struct pa
 #define SWAP_AGAIN	1
 #define SWAP_FAIL	2
 #define SWAP_MLOCK	3
-#define SWAP_LZFREE	4
+#define SWAP_DIRTY	4
 
 #endif	/* _LINUX_RMAP_H */
diff -puN mm/huge_memory.c~mm-reclaim-madv_free-pages mm/huge_memory.c
--- a/mm/huge_memory.c~mm-reclaim-madv_free-pages
+++ a/mm/huge_memory.c
@@ -1573,6 +1573,8 @@ bool madvise_free_huge_pmd(struct mmu_ga
 		set_pmd_at(mm, addr, pmd, orig_pmd);
 		tlb_remove_pmd_tlb_entry(tlb, pmd, addr);
 	}
+
+	mark_page_lazyfree(page);
 	ret = true;
 out:
 	spin_unlock(ptl);
diff -puN mm/madvise.c~mm-reclaim-madv_free-pages mm/madvise.c
--- a/mm/madvise.c~mm-reclaim-madv_free-pages
+++ a/mm/madvise.c
@@ -413,6 +413,7 @@ static int madvise_free_pte_range(pmd_t
 			set_pte_at(mm, addr, pte, ptent);
 			tlb_remove_tlb_entry(tlb, pte, addr);
 		}
+		mark_page_lazyfree(page);
 	}
 out:
 	if (nr_swap) {
diff -puN mm/rmap.c~mm-reclaim-madv_free-pages mm/rmap.c
--- a/mm/rmap.c~mm-reclaim-madv_free-pages
+++ a/mm/rmap.c
@@ -1283,11 +1283,6 @@ void page_remove_rmap(struct page *page,
 	 */
 }
 
-struct rmap_private {
-	enum ttu_flags flags;
-	int lazyfreed;
-};
-
 /*
  * @arg: enum ttu_flags will be passed to this argument
  */
@@ -1303,8 +1298,7 @@ static int try_to_unmap_one(struct page
 	pte_t pteval;
 	struct page *subpage;
 	int ret = SWAP_AGAIN;
-	struct rmap_private *rp = arg;
-	enum ttu_flags flags = rp->flags;
+	enum ttu_flags flags = (enum ttu_flags)arg;
 
 	/* munlock has nothing to gain from examining un-locked vmas */
 	if ((flags & TTU_MUNLOCK) && !(vma->vm_flags & VM_LOCKED))
@@ -1422,11 +1416,21 @@ static int try_to_unmap_one(struct page
 			VM_BUG_ON_PAGE(!PageSwapCache(page) && PageSwapBacked(page),
 				page);
 
-			if (!PageDirty(page)) {
+			/*
+			 * swapin page could be clean, it has data stored in
+			 * swap. We can't silently discard it without setting
+			 * swap entry in the page table.
+			 */
+			if (!PageDirty(page) && !PageSwapCache(page)) {
 				/* It's a freeable page by MADV_FREE */
 				dec_mm_counter(mm, MM_ANONPAGES);
-				rp->lazyfreed++;
 				goto discard;
+			} else if (!PageSwapBacked(page)) {
+				/* dirty MADV_FREE page */
+				set_pte_at(mm, address, pvmw.pte, pteval);
+				ret = SWAP_DIRTY;
+				page_vma_mapped_walk_done(&pvmw);
+				break;
 			}
 
 			if (swap_duplicate(entry) < 0) {
@@ -1494,18 +1498,15 @@ static int page_mapcount_is_zero(struct
  * SWAP_AGAIN	- we missed a mapping, try again later
  * SWAP_FAIL	- the page is unswappable
  * SWAP_MLOCK	- page is mlocked.
+ * SWAP_DIRTY	- page is dirty MADV_FREE page
  */
 int try_to_unmap(struct page *page, enum ttu_flags flags)
 {
 	int ret;
-	struct rmap_private rp = {
-		.flags = flags,
-		.lazyfreed = 0,
-	};
 
 	struct rmap_walk_control rwc = {
 		.rmap_one = try_to_unmap_one,
-		.arg = &rp,
+		.arg = (void *)flags,
 		.done = page_mapcount_is_zero,
 		.anon_lock = page_lock_anon_vma_read,
 	};
@@ -1526,11 +1527,8 @@ int try_to_unmap(struct page *page, enum
 	else
 		ret = rmap_walk(page, &rwc);
 
-	if (ret != SWAP_MLOCK && !page_mapcount(page)) {
+	if (ret != SWAP_MLOCK && !page_mapcount(page))
 		ret = SWAP_SUCCESS;
-		if (rp.lazyfreed && !PageDirty(page))
-			ret = SWAP_LZFREE;
-	}
 	return ret;
 }
 
@@ -1557,14 +1555,10 @@ static int page_not_mapped(struct page *
 int try_to_munlock(struct page *page)
 {
 	int ret;
-	struct rmap_private rp = {
-		.flags = TTU_MUNLOCK,
-		.lazyfreed = 0,
-	};
 
 	struct rmap_walk_control rwc = {
 		.rmap_one = try_to_unmap_one,
-		.arg = &rp,
+		.arg = (void *)TTU_MUNLOCK,
 		.done = page_not_mapped,
 		.anon_lock = page_lock_anon_vma_read,
 
diff -puN mm/vmscan.c~mm-reclaim-madv_free-pages mm/vmscan.c
--- a/mm/vmscan.c~mm-reclaim-madv_free-pages
+++ a/mm/vmscan.c
@@ -906,7 +906,8 @@ static void page_check_dirty_writeback(s
 	 * Anonymous pages are not handled by flushers and must be written
 	 * from reclaim context. Do not stall reclaim based on them
 	 */
-	if (!page_is_file_cache(page)) {
+	if (!page_is_file_cache(page) ||
+	    (PageAnon(page) && !PageSwapBacked(page))) {
 		*dirty = false;
 		*writeback = false;
 		return;
@@ -987,7 +988,8 @@ static unsigned long shrink_page_list(st
 			goto keep_locked;
 
 		/* Double the slab pressure for mapped and swapcache pages */
-		if (page_mapped(page) || PageSwapCache(page))
+		if ((page_mapped(page) || PageSwapCache(page)) &&
+		    !(PageAnon(page) && !PageSwapBacked(page)))
 			sc->nr_scanned++;
 
 		may_enter_fs = (sc->gfp_mask & __GFP_FS) ||
@@ -1113,8 +1115,10 @@ static unsigned long shrink_page_list(st
 		/*
 		 * Anonymous process memory has backing store?
 		 * Try to allocate it some swap space here.
+		 * Lazyfree page could be freed directly
 		 */
-		if (PageAnon(page) && !PageSwapCache(page)) {
+		if (PageAnon(page) && PageSwapBacked(page) &&
+		    !PageSwapCache(page)) {
 			if (!(sc->gfp_mask & __GFP_IO))
 				goto keep_locked;
 			if (!add_to_swap(page, page_list))
@@ -1135,9 +1139,12 @@ static unsigned long shrink_page_list(st
 		 * The page is mapped into the page tables of one or more
 		 * processes. Try to unmap it here.
 		 */
-		if (page_mapped(page) && mapping) {
+		if (page_mapped(page)) {
 			switch (ret = try_to_unmap(page,
 				ttu_flags | TTU_BATCH_FLUSH)) {
+			case SWAP_DIRTY:
+				SetPageSwapBacked(page);
+				/* fall through */
 			case SWAP_FAIL:
 				nr_unmap_fail++;
 				goto activate_locked;
@@ -1145,8 +1152,6 @@ static unsigned long shrink_page_list(st
 				goto keep_locked;
 			case SWAP_MLOCK:
 				goto cull_mlocked;
-			case SWAP_LZFREE:
-				goto lazyfree;
 			case SWAP_SUCCESS:
 				; /* try to free the page below */
 			}
@@ -1258,10 +1263,18 @@ static unsigned long shrink_page_list(st
 			}
 		}
 
-lazyfree:
-		if (!mapping || !__remove_mapping(mapping, page, true))
-			goto keep_locked;
+		if (PageAnon(page) && !PageSwapBacked(page)) {
+			/* follow __remove_mapping for reference */
+			if (!page_ref_freeze(page, 1))
+				goto keep_locked;
+			if (PageDirty(page)) {
+				page_ref_unfreeze(page, 1);
+				goto keep_locked;
+			}
 
+			count_vm_event(PGLAZYFREED);
+		} else if (!mapping || !__remove_mapping(mapping, page, true))
+			goto keep_locked;
 		/*
 		 * At this point, we have no other references and there is
 		 * no way to pick any more up (removed from LRU, removed
@@ -1271,9 +1284,6 @@ lazyfree:
 		 */
 		__ClearPageLocked(page);
 free_it:
-		if (ret == SWAP_LZFREE)
-			count_vm_event(PGLAZYFREED);

^ permalink raw reply	[flat|nested] 11+ messages in thread

end of thread, other threads:[~2017-03-07 23:06 UTC | newest]

Thread overview: 11+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2017-03-01  0:32 + mm-reclaim-madv_free-pages.patch added to -mm tree akpm
2017-03-03  2:52 ` Minchan Kim
2017-03-03 15:18   ` Johannes Weiner
2017-03-06  3:03     ` Minchan Kim
2017-03-06 15:49       ` Johannes Weiner
2017-03-07  5:46         ` Minchan Kim
2017-03-07  5:55         ` Minchan Kim
2017-03-07  9:58           ` Michal Hocko
2017-03-07 16:59           ` Johannes Weiner
2017-03-03 16:02   ` Shaohua Li
2017-03-07 22:57 akpm

This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.