From: Alex Shi <alex.shi@linux.alibaba.com> To: Alexander Duyck <alexander.duyck@gmail.com> Cc: Andrew Morton <akpm@linux-foundation.org>, Mel Gorman <mgorman@techsingularity.net>, Tejun Heo <tj@kernel.org>, Hugh Dickins <hughd@google.com>, Konstantin Khlebnikov <khlebnikov@yandex-team.ru>, Daniel Jordan <daniel.m.jordan@oracle.com>, Yang Shi <yang.shi@linux.alibaba.com>, Matthew Wilcox <willy@infradead.org>, Johannes Weiner <hannes@cmpxchg.org>, kbuild test robot <lkp@intel.com>, linux-mm <linux-mm@kvack.org>, LKML <linux-kernel@vger.kernel.org>, cgroups@vger.kernel.org, Shakeel Butt <shakeelb@google.com>, Joonsoo Kim <iamjoonsoo.kim@lge.com>, Wei Yang <richard.weiyang@gmail.com>, "Kirill A. Shutemov" <kirill@shutemov.name>, Rong Chen <rong.a.chen@intel.com>, Michal Hocko <mhocko@kernel.org>, Vladimir Davydov <vdavydov.dev@gmail.com> Subject: Re: [PATCH v17 17/21] mm/lru: replace pgdat lru_lock with lruvec lock Date: Tue, 28 Jul 2020 23:39:31 +0800 [thread overview] Message-ID: <1fd45e69-3a50-aae8-bcc4-47d891a5e263@linux.alibaba.com> (raw) In-Reply-To: <CAKgT0UdaW4Rf43yULhQBuP07vQgmoPbaWHGKv1Z7fEPP6jH83w@mail.gmail.com> 在 2020/7/28 上午7:34, Alexander Duyck 写道: > It might make more sense to look at modifying > compact_unlock_should_abort and compact_lock_irqsave (which always > returns true so should probably be a void) to address the deficiencies > they have that make them unusable for you. One of possible reuse for the func compact_unlock_should_abort, could be like the following, the locked parameter reused different in 2 places. but, it's seems no this style usage in kernel, isn't it? Thanks Alex From 41d5ce6562f20f74bc6ac2db83e226ac28d56e90 Mon Sep 17 00:00:00 2001 From: Alex Shi <alex.shi@linux.alibaba.com> Date: Tue, 28 Jul 2020 21:19:32 +0800 Subject: [PATCH] compaction polishing Signed-off-by: Alex Shi <alex.shi@linux.alibaba.com> --- mm/compaction.c | 71 ++++++++++++++++++++++++--------------------------------- 1 file changed, 30 insertions(+), 41 deletions(-) diff --git a/mm/compaction.c b/mm/compaction.c index c28a43481f01..36fce988de3e 100644 --- a/mm/compaction.c +++ b/mm/compaction.c @@ -479,20 +479,20 @@ static bool test_and_set_skip(struct compact_control *cc, struct page *page, * * Always returns true which makes it easier to track lock state in callers. */ -static bool compact_lock_irqsave(spinlock_t *lock, unsigned long *flags, +static void compact_lock_irqsave(spinlock_t *lock, unsigned long *flags, struct compact_control *cc) __acquires(lock) { /* Track if the lock is contended in async mode */ if (cc->mode == MIGRATE_ASYNC && !cc->contended) { if (spin_trylock_irqsave(lock, *flags)) - return true; + return; cc->contended = true; } spin_lock_irqsave(lock, *flags); - return true; + return; } /* @@ -511,11 +511,11 @@ static bool compact_lock_irqsave(spinlock_t *lock, unsigned long *flags, * scheduled) */ static bool compact_unlock_should_abort(spinlock_t *lock, - unsigned long flags, bool *locked, struct compact_control *cc) + unsigned long flags, void **locked, struct compact_control *cc) { if (*locked) { spin_unlock_irqrestore(lock, flags); - *locked = false; + *locked = NULL; } if (fatal_signal_pending(current)) { @@ -543,7 +543,7 @@ static unsigned long isolate_freepages_block(struct compact_control *cc, int nr_scanned = 0, total_isolated = 0; struct page *cursor; unsigned long flags = 0; - bool locked = false; + struct compact_control *locked = NULL; unsigned long blockpfn = *start_pfn; unsigned int order; @@ -565,7 +565,7 @@ static unsigned long isolate_freepages_block(struct compact_control *cc, */ if (!(blockpfn % SWAP_CLUSTER_MAX) && compact_unlock_should_abort(&cc->zone->lock, flags, - &locked, cc)) + (void**)&locked, cc)) break; nr_scanned++; @@ -599,8 +599,8 @@ static unsigned long isolate_freepages_block(struct compact_control *cc, * recheck as well. */ if (!locked) { - locked = compact_lock_irqsave(&cc->zone->lock, - &flags, cc); + compact_lock_irqsave(&cc->zone->lock, &flags, cc); + locked = cc; /* Recheck this is a buddy page under lock */ if (!PageBuddy(page)) @@ -787,7 +787,7 @@ static bool too_many_isolated(pg_data_t *pgdat) unsigned long nr_scanned = 0, nr_isolated = 0; struct lruvec *lruvec; unsigned long flags = 0; - struct lruvec *locked_lruvec = NULL; + struct lruvec *locked = NULL; struct page *page = NULL, *valid_page = NULL; unsigned long start_pfn = low_pfn; bool skip_on_failure = false; @@ -847,21 +847,11 @@ static bool too_many_isolated(pg_data_t *pgdat) * contention, to give chance to IRQs. Abort completely if * a fatal signal is pending. */ - if (!(low_pfn % SWAP_CLUSTER_MAX)) { - if (locked_lruvec) { - unlock_page_lruvec_irqrestore(locked_lruvec, - flags); - locked_lruvec = NULL; - } - - if (fatal_signal_pending(current)) { - cc->contended = true; - - low_pfn = 0; - goto fatal_pending; - } - - cond_resched(); + if (!(low_pfn % SWAP_CLUSTER_MAX) + && compact_unlock_should_abort(&locked->lru_lock, flags, + (void**)&locked, cc)) { + low_pfn = 0; + goto fatal_pending; } if (!pfn_valid_within(low_pfn)) @@ -932,9 +922,9 @@ static bool too_many_isolated(pg_data_t *pgdat) */ if (unlikely(__PageMovable(page)) && !PageIsolated(page)) { - if (locked_lruvec) { - unlock_page_lruvec_irqrestore(locked_lruvec, flags); - locked_lruvec = NULL; + if (locked) { + unlock_page_lruvec_irqrestore(locked, flags); + locked = NULL; } if (!isolate_movable_page(page, isolate_mode)) @@ -979,13 +969,13 @@ static bool too_many_isolated(pg_data_t *pgdat) lruvec = mem_cgroup_page_lruvec(page, pgdat); /* If we already hold the lock, we can skip some rechecking */ - if (lruvec != locked_lruvec) { - if (locked_lruvec) - unlock_page_lruvec_irqrestore(locked_lruvec, + if (lruvec != locked) { + if (locked) + unlock_page_lruvec_irqrestore(locked, flags); compact_lock_irqsave(&lruvec->lru_lock, &flags, cc); - locked_lruvec = lruvec; + locked = lruvec; rcu_read_unlock(); lruvec_memcg_debug(lruvec, page); @@ -1041,9 +1031,9 @@ static bool too_many_isolated(pg_data_t *pgdat) isolate_fail_put: /* Avoid potential deadlock in freeing page under lru_lock */ - if (locked_lruvec) { - unlock_page_lruvec_irqrestore(locked_lruvec, flags); - locked_lruvec = NULL; + if (locked) { + unlock_page_lruvec_irqrestore(locked, flags); + locked = NULL; } put_page(page); @@ -1057,10 +1047,9 @@ static bool too_many_isolated(pg_data_t *pgdat) * page anyway. */ if (nr_isolated) { - if (locked_lruvec) { - unlock_page_lruvec_irqrestore(locked_lruvec, - flags); - locked_lruvec = NULL; + if (locked) { + unlock_page_lruvec_irqrestore(locked, flags); + locked = NULL; } putback_movable_pages(&cc->migratepages); cc->nr_migratepages = 0; @@ -1087,8 +1076,8 @@ static bool too_many_isolated(pg_data_t *pgdat) page = NULL; isolate_abort: - if (locked_lruvec) - unlock_page_lruvec_irqrestore(locked_lruvec, flags); + if (locked) + unlock_page_lruvec_irqrestore(locked, flags); if (page) { SetPageLRU(page); put_page(page); -- 1.8.3.1
WARNING: multiple messages have this Message-ID (diff)
From: Alex Shi <alex.shi-KPsoFbNs7GizrGE5bRqYAgC/G2K4zDHf@public.gmane.org> To: Alexander Duyck <alexander.duyck-Re5JQEeQqe8AvxtiuMwx3w@public.gmane.org> Cc: Andrew Morton <akpm-de/tnXTf+JLsfHDXvbKv3WD2FQJk+8+b@public.gmane.org>, Mel Gorman <mgorman-3eNAlZScCAx27rWaFMvyedHuzzzSOjJt@public.gmane.org>, Tejun Heo <tj-DgEjT+Ai2ygdnm+yROfE0A@public.gmane.org>, Hugh Dickins <hughd-hpIqsD4AKlfQT0dZR+AlfA@public.gmane.org>, Konstantin Khlebnikov <khlebnikov-XoJtRXgx1JseBXzfvpsJ4g@public.gmane.org>, Daniel Jordan <daniel.m.jordan-QHcLZuEGTsvQT0dZR+AlfA@public.gmane.org>, Yang Shi <yang.shi-KPsoFbNs7GizrGE5bRqYAgC/G2K4zDHf@public.gmane.org>, Matthew Wilcox <willy-wEGCiKHe2LqWVfeAwA7xHQ@public.gmane.org>, Johannes Weiner <hannes-druUgvl0LCNAfugRpC6u6w@public.gmane.org>, kbuild test robot <lkp-ral2JQCrhuEAvxtiuMwx3w@public.gmane.org>, linux-mm <linux-mm-Bw31MaZKKs3YtjvyW6yDsg@public.gmane.org>, LKML <linux-kernel-u79uwXL29TY76Z2rM5mHXA@public.gmane.org>, cgroups-u79uwXL29TY76Z2rM5mHXA@public.gmane.org, Shakeel Butt <shakeelb-hpIqsD4AKlfQT0dZR+AlfA@public.gmane.org>, Joonsoo Kim <iamjoonsoo.kim-Hm3cg6mZ9cc@public.gmane.org>, Wei Yang <richard.weiyang-Re5JQEeQqe8AvxtiuMwx3w@public.gmane.org>, "Kirill A. Shutemov" <kirill-oKw7cIdHH8eLwutG50LtGA@public.gmane.org>, Rong Chen <rong.a.chen-ral2JQCrhuEAvxtiuMwx3w@public.gmane.org>, Michal Hocko <mhocko-DgEjT+Ai2ygdnm+yROfE0A@public.gmane.org>, Vladimir Davydov <vdavydov.dev-Re5JQEeQqe8AvxtiuMwx3w@public.gmane.org> Subject: Re: [PATCH v17 17/21] mm/lru: replace pgdat lru_lock with lruvec lock Date: Tue, 28 Jul 2020 23:39:31 +0800 [thread overview] Message-ID: <1fd45e69-3a50-aae8-bcc4-47d891a5e263@linux.alibaba.com> (raw) In-Reply-To: <CAKgT0UdaW4Rf43yULhQBuP07vQgmoPbaWHGKv1Z7fEPP6jH83w-JsoAwUIsXosN+BqQ9rBEUg@public.gmane.org> 在 2020/7/28 上午7:34, Alexander Duyck 写道: > It might make more sense to look at modifying > compact_unlock_should_abort and compact_lock_irqsave (which always > returns true so should probably be a void) to address the deficiencies > they have that make them unusable for you. One of possible reuse for the func compact_unlock_should_abort, could be like the following, the locked parameter reused different in 2 places. but, it's seems no this style usage in kernel, isn't it? Thanks Alex From 41d5ce6562f20f74bc6ac2db83e226ac28d56e90 Mon Sep 17 00:00:00 2001 From: Alex Shi <alex.shi-KPsoFbNs7GizrGE5bRqYAgC/G2K4zDHf@public.gmane.org> Date: Tue, 28 Jul 2020 21:19:32 +0800 Subject: [PATCH] compaction polishing Signed-off-by: Alex Shi <alex.shi-KPsoFbNs7GizrGE5bRqYAgC/G2K4zDHf@public.gmane.org> --- mm/compaction.c | 71 ++++++++++++++++++++++++--------------------------------- 1 file changed, 30 insertions(+), 41 deletions(-) diff --git a/mm/compaction.c b/mm/compaction.c index c28a43481f01..36fce988de3e 100644 --- a/mm/compaction.c +++ b/mm/compaction.c @@ -479,20 +479,20 @@ static bool test_and_set_skip(struct compact_control *cc, struct page *page, * * Always returns true which makes it easier to track lock state in callers. */ -static bool compact_lock_irqsave(spinlock_t *lock, unsigned long *flags, +static void compact_lock_irqsave(spinlock_t *lock, unsigned long *flags, struct compact_control *cc) __acquires(lock) { /* Track if the lock is contended in async mode */ if (cc->mode == MIGRATE_ASYNC && !cc->contended) { if (spin_trylock_irqsave(lock, *flags)) - return true; + return; cc->contended = true; } spin_lock_irqsave(lock, *flags); - return true; + return; } /* @@ -511,11 +511,11 @@ static bool compact_lock_irqsave(spinlock_t *lock, unsigned long *flags, * scheduled) */ static bool compact_unlock_should_abort(spinlock_t *lock, - unsigned long flags, bool *locked, struct compact_control *cc) + unsigned long flags, void **locked, struct compact_control *cc) { if (*locked) { spin_unlock_irqrestore(lock, flags); - *locked = false; + *locked = NULL; } if (fatal_signal_pending(current)) { @@ -543,7 +543,7 @@ static unsigned long isolate_freepages_block(struct compact_control *cc, int nr_scanned = 0, total_isolated = 0; struct page *cursor; unsigned long flags = 0; - bool locked = false; + struct compact_control *locked = NULL; unsigned long blockpfn = *start_pfn; unsigned int order; @@ -565,7 +565,7 @@ static unsigned long isolate_freepages_block(struct compact_control *cc, */ if (!(blockpfn % SWAP_CLUSTER_MAX) && compact_unlock_should_abort(&cc->zone->lock, flags, - &locked, cc)) + (void**)&locked, cc)) break; nr_scanned++; @@ -599,8 +599,8 @@ static unsigned long isolate_freepages_block(struct compact_control *cc, * recheck as well. */ if (!locked) { - locked = compact_lock_irqsave(&cc->zone->lock, - &flags, cc); + compact_lock_irqsave(&cc->zone->lock, &flags, cc); + locked = cc; /* Recheck this is a buddy page under lock */ if (!PageBuddy(page)) @@ -787,7 +787,7 @@ static bool too_many_isolated(pg_data_t *pgdat) unsigned long nr_scanned = 0, nr_isolated = 0; struct lruvec *lruvec; unsigned long flags = 0; - struct lruvec *locked_lruvec = NULL; + struct lruvec *locked = NULL; struct page *page = NULL, *valid_page = NULL; unsigned long start_pfn = low_pfn; bool skip_on_failure = false; @@ -847,21 +847,11 @@ static bool too_many_isolated(pg_data_t *pgdat) * contention, to give chance to IRQs. Abort completely if * a fatal signal is pending. */ - if (!(low_pfn % SWAP_CLUSTER_MAX)) { - if (locked_lruvec) { - unlock_page_lruvec_irqrestore(locked_lruvec, - flags); - locked_lruvec = NULL; - } - - if (fatal_signal_pending(current)) { - cc->contended = true; - - low_pfn = 0; - goto fatal_pending; - } - - cond_resched(); + if (!(low_pfn % SWAP_CLUSTER_MAX) + && compact_unlock_should_abort(&locked->lru_lock, flags, + (void**)&locked, cc)) { + low_pfn = 0; + goto fatal_pending; } if (!pfn_valid_within(low_pfn)) @@ -932,9 +922,9 @@ static bool too_many_isolated(pg_data_t *pgdat) */ if (unlikely(__PageMovable(page)) && !PageIsolated(page)) { - if (locked_lruvec) { - unlock_page_lruvec_irqrestore(locked_lruvec, flags); - locked_lruvec = NULL; + if (locked) { + unlock_page_lruvec_irqrestore(locked, flags); + locked = NULL; } if (!isolate_movable_page(page, isolate_mode)) @@ -979,13 +969,13 @@ static bool too_many_isolated(pg_data_t *pgdat) lruvec = mem_cgroup_page_lruvec(page, pgdat); /* If we already hold the lock, we can skip some rechecking */ - if (lruvec != locked_lruvec) { - if (locked_lruvec) - unlock_page_lruvec_irqrestore(locked_lruvec, + if (lruvec != locked) { + if (locked) + unlock_page_lruvec_irqrestore(locked, flags); compact_lock_irqsave(&lruvec->lru_lock, &flags, cc); - locked_lruvec = lruvec; + locked = lruvec; rcu_read_unlock(); lruvec_memcg_debug(lruvec, page); @@ -1041,9 +1031,9 @@ static bool too_many_isolated(pg_data_t *pgdat) isolate_fail_put: /* Avoid potential deadlock in freeing page under lru_lock */ - if (locked_lruvec) { - unlock_page_lruvec_irqrestore(locked_lruvec, flags); - locked_lruvec = NULL; + if (locked) { + unlock_page_lruvec_irqrestore(locked, flags); + locked = NULL; } put_page(page); @@ -1057,10 +1047,9 @@ static bool too_many_isolated(pg_data_t *pgdat) * page anyway. */ if (nr_isolated) { - if (locked_lruvec) { - unlock_page_lruvec_irqrestore(locked_lruvec, - flags); - locked_lruvec = NULL; + if (locked) { + unlock_page_lruvec_irqrestore(locked, flags); + locked = NULL; } putback_movable_pages(&cc->migratepages); cc->nr_migratepages = 0; @@ -1087,8 +1076,8 @@ static bool too_many_isolated(pg_data_t *pgdat) page = NULL; isolate_abort: - if (locked_lruvec) - unlock_page_lruvec_irqrestore(locked_lruvec, flags); + if (locked) + unlock_page_lruvec_irqrestore(locked, flags); if (page) { SetPageLRU(page); put_page(page); -- 1.8.3.1
next prev parent reply other threads:[~2020-07-28 15:39 UTC|newest] Thread overview: 209+ messages / expand[flat|nested] mbox.gz Atom feed top 2020-07-25 12:59 [PATCH v17 00/21] per memcg lru lock Alex Shi 2020-07-25 12:59 ` Alex Shi 2020-07-25 12:59 ` [PATCH v17 01/21] mm/vmscan: remove unnecessary lruvec adding Alex Shi 2020-08-06 3:47 ` Alex Shi 2020-08-06 3:47 ` Alex Shi 2020-07-25 12:59 ` [PATCH v17 02/21] mm/page_idle: no unlikely double check for idle page counting Alex Shi 2020-07-25 12:59 ` [PATCH v17 03/21] mm/compaction: correct the comments of compact_defer_shift Alex Shi 2020-07-27 17:29 ` Alexander Duyck 2020-07-27 17:29 ` Alexander Duyck 2020-07-28 11:59 ` Alex Shi 2020-07-28 11:59 ` Alex Shi 2020-07-28 14:17 ` Alexander Duyck 2020-07-28 14:17 ` Alexander Duyck 2020-07-28 14:17 ` Alexander Duyck 2020-07-25 12:59 ` [PATCH v17 04/21] mm/compaction: rename compact_deferred as compact_should_defer Alex Shi 2020-07-25 12:59 ` Alex Shi 2020-07-25 12:59 ` [PATCH v17 05/21] mm/thp: move lru_add_page_tail func to huge_memory.c Alex Shi 2020-07-25 12:59 ` Alex Shi 2020-07-25 12:59 ` [PATCH v17 06/21] mm/thp: clean up lru_add_page_tail Alex Shi 2020-07-25 12:59 ` [PATCH v17 07/21] mm/thp: remove code path which never got into Alex Shi 2020-07-25 12:59 ` [PATCH v17 08/21] mm/thp: narrow lru locking Alex Shi 2020-07-25 12:59 ` [PATCH v17 09/21] mm/memcg: add debug checking in lock_page_memcg Alex Shi 2020-07-25 12:59 ` Alex Shi 2020-07-25 12:59 ` [PATCH v17 10/21] mm/swap: fold vm event PGROTATED into pagevec_move_tail_fn Alex Shi 2020-07-25 12:59 ` [PATCH v17 11/21] mm/lru: move lru_lock holding in func lru_note_cost_page Alex Shi 2020-08-05 21:18 ` Alexander Duyck 2020-08-05 21:18 ` Alexander Duyck 2020-08-05 21:18 ` Alexander Duyck 2020-07-25 12:59 ` [PATCH v17 12/21] mm/lru: move lock into lru_note_cost Alex Shi 2020-07-25 12:59 ` Alex Shi 2020-07-25 12:59 ` [PATCH v17 13/21] mm/lru: introduce TestClearPageLRU Alex Shi 2020-07-25 12:59 ` Alex Shi 2020-07-29 3:53 ` Alex Shi 2020-07-29 3:53 ` Alex Shi 2020-08-05 22:43 ` Alexander Duyck 2020-08-05 22:43 ` Alexander Duyck 2020-08-06 1:54 ` Alex Shi 2020-08-06 1:54 ` Alex Shi 2020-08-06 14:41 ` Alexander Duyck 2020-08-06 14:41 ` Alexander Duyck 2020-07-25 12:59 ` [PATCH v17 14/21] mm/compaction: do page isolation first in compaction Alex Shi 2020-08-04 21:35 ` Alexander Duyck 2020-08-04 21:35 ` Alexander Duyck 2020-08-04 21:35 ` Alexander Duyck 2020-08-06 18:38 ` Alexander Duyck 2020-08-06 18:38 ` Alexander Duyck 2020-08-06 18:38 ` Alexander Duyck 2020-08-07 3:24 ` Alex Shi 2020-08-07 3:24 ` Alex Shi 2020-08-07 14:51 ` Alexander Duyck 2020-08-07 14:51 ` Alexander Duyck 2020-08-07 14:51 ` Alexander Duyck 2020-08-10 13:10 ` Alex Shi 2020-08-10 13:10 ` Alex Shi 2020-08-10 14:41 ` Alexander Duyck 2020-08-10 14:41 ` Alexander Duyck 2020-08-10 14:41 ` Alexander Duyck 2020-08-11 8:22 ` Alex Shi 2020-08-11 14:47 ` Alexander Duyck 2020-08-11 14:47 ` Alexander Duyck 2020-08-11 14:47 ` Alexander Duyck 2020-08-12 11:43 ` Alex Shi 2020-08-12 12:16 ` Alex Shi 2020-08-12 12:16 ` Alex Shi 2020-08-12 16:51 ` Alexander Duyck 2020-08-12 16:51 ` Alexander Duyck 2020-08-12 16:51 ` Alexander Duyck 2020-08-13 1:46 ` Alex Shi 2020-08-13 2:17 ` Alexander Duyck 2020-08-13 2:17 ` Alexander Duyck 2020-08-13 2:17 ` Alexander Duyck 2020-08-13 3:52 ` Alex Shi 2020-08-13 3:52 ` Alex Shi 2020-08-13 4:02 ` [RFC PATCH 0/3] " Alexander Duyck 2020-08-13 4:02 ` Alexander Duyck 2020-08-13 4:02 ` [RFC PATCH 1/3] mm: Drop locked from isolate_migratepages_block Alexander Duyck 2020-08-13 4:02 ` Alexander Duyck 2020-08-13 6:56 ` Alex Shi 2020-08-13 6:56 ` Alex Shi 2020-08-13 14:32 ` Alexander Duyck 2020-08-13 14:32 ` Alexander Duyck 2020-08-13 14:32 ` Alexander Duyck 2020-08-14 7:25 ` Alex Shi 2020-08-14 7:25 ` Alex Shi 2020-08-13 7:44 ` Alex Shi 2020-08-13 7:44 ` Alex Shi 2020-08-13 14:26 ` Alexander Duyck 2020-08-13 14:26 ` Alexander Duyck 2020-08-13 4:02 ` [RFC PATCH 2/3] mm: Drop use of test_and_set_skip in favor of just setting skip Alexander Duyck 2020-08-13 4:02 ` Alexander Duyck 2020-08-14 7:19 ` Alex Shi 2020-08-14 7:19 ` Alex Shi 2020-08-14 14:24 ` Alexander Duyck 2020-08-14 14:24 ` Alexander Duyck 2020-08-14 14:24 ` Alexander Duyck 2020-08-14 21:15 ` Alexander Duyck 2020-08-14 21:15 ` Alexander Duyck 2020-08-14 21:15 ` Alexander Duyck 2020-08-15 9:49 ` Alex Shi 2020-08-17 15:38 ` Alexander Duyck 2020-08-17 15:38 ` Alexander Duyck 2020-08-17 15:38 ` Alexander Duyck 2020-08-18 6:50 ` Alex Shi 2020-08-18 6:50 ` Alex Shi 2020-08-13 4:02 ` [RFC PATCH 3/3] mm: Identify compound pages sooner in isolate_migratepages_block Alexander Duyck 2020-08-13 4:02 ` Alexander Duyck 2020-08-14 7:20 ` Alex Shi 2020-08-14 7:20 ` Alex Shi 2020-08-17 22:58 ` [PATCH v17 14/21] mm/compaction: do page isolation first in compaction Alexander Duyck 2020-08-17 22:58 ` Alexander Duyck 2020-08-17 22:58 ` Alexander Duyck 2020-07-25 12:59 ` [PATCH v17 15/21] mm/thp: add tail pages into lru anyway in split_huge_page() Alex Shi 2020-07-25 12:59 ` Alex Shi 2020-07-25 12:59 ` [PATCH v17 16/21] mm/swap: serialize memcg changes in pagevec_lru_move_fn Alex Shi 2020-07-25 12:59 ` [PATCH v17 17/21] mm/lru: replace pgdat lru_lock with lruvec lock Alex Shi 2020-07-25 12:59 ` Alex Shi 2020-07-27 23:34 ` Alexander Duyck 2020-07-27 23:34 ` Alexander Duyck 2020-07-27 23:34 ` Alexander Duyck 2020-07-28 7:15 ` Alex Shi 2020-07-28 7:15 ` Alex Shi 2020-07-28 11:19 ` Alex Shi 2020-07-28 11:19 ` Alex Shi 2020-07-28 14:54 ` Alexander Duyck 2020-07-28 14:54 ` Alexander Duyck 2020-07-28 14:54 ` Alexander Duyck 2020-07-29 1:00 ` Alex Shi 2020-07-29 1:27 ` Alexander Duyck 2020-07-29 1:27 ` Alexander Duyck 2020-07-29 1:27 ` Alexander Duyck 2020-07-29 2:27 ` Alex Shi 2020-07-29 2:27 ` Alex Shi 2020-07-28 15:39 ` Alex Shi [this message] 2020-07-28 15:39 ` Alex Shi 2020-07-28 15:55 ` Alexander Duyck 2020-07-28 15:55 ` Alexander Duyck 2020-07-28 15:55 ` Alexander Duyck 2020-07-29 0:48 ` Alex Shi 2020-07-29 3:54 ` Alex Shi 2020-08-06 7:41 ` Alex Shi 2020-08-06 7:41 ` Alex Shi 2020-07-25 12:59 ` [PATCH v17 18/21] mm/lru: introduce the relock_page_lruvec function Alex Shi 2020-07-29 17:52 ` Alexander Duyck 2020-07-29 17:52 ` Alexander Duyck 2020-07-29 17:52 ` Alexander Duyck 2020-07-30 6:08 ` Alex Shi 2020-07-30 6:08 ` Alex Shi 2020-07-31 14:20 ` Alexander Duyck 2020-07-31 14:20 ` Alexander Duyck 2020-07-31 14:20 ` Alexander Duyck 2020-07-31 21:14 ` [PATCH RFC] mm: Add function for testing if the current lruvec lock is valid alexander.h.duyck 2020-07-31 21:14 ` alexander.h.duyck-ral2JQCrhuEAvxtiuMwx3w 2020-07-31 23:54 ` Alex Shi 2020-07-31 23:54 ` Alex Shi 2020-08-02 18:20 ` Alexander Duyck 2020-08-02 18:20 ` Alexander Duyck 2020-08-02 18:20 ` Alexander Duyck 2020-08-04 6:13 ` Alex Shi 2020-08-04 6:13 ` Alex Shi 2020-07-25 12:59 ` [PATCH v17 19/21] mm/vmscan: use relock for move_pages_to_lru Alex Shi 2020-08-03 22:49 ` Alexander Duyck 2020-08-03 22:49 ` Alexander Duyck 2020-08-04 6:23 ` Alex Shi 2020-08-04 6:23 ` Alex Shi 2020-07-25 12:59 ` [PATCH v17 20/21] mm/pgdat: remove pgdat lru_lock Alex Shi 2020-08-03 22:42 ` Alexander Duyck 2020-08-03 22:42 ` Alexander Duyck 2020-08-03 22:42 ` Alexander Duyck 2020-08-03 22:45 ` Alexander Duyck 2020-08-03 22:45 ` Alexander Duyck 2020-08-03 22:45 ` Alexander Duyck 2020-08-04 6:22 ` Alex Shi 2020-08-04 6:22 ` Alex Shi 2020-07-25 12:59 ` [PATCH v17 21/21] mm/lru: revise the comments of lru_lock Alex Shi 2020-08-03 22:37 ` Alexander Duyck 2020-08-03 22:37 ` Alexander Duyck 2020-08-04 10:04 ` Alex Shi 2020-08-04 10:04 ` Alex Shi 2020-08-04 14:29 ` Alexander Duyck 2020-08-04 14:29 ` Alexander Duyck 2020-08-04 14:29 ` Alexander Duyck 2020-08-06 1:39 ` Alex Shi 2020-08-06 1:39 ` Alex Shi 2020-08-06 16:27 ` Alexander Duyck 2020-08-06 16:27 ` Alexander Duyck 2020-08-06 16:27 ` Alexander Duyck 2020-07-27 5:40 ` [PATCH v17 00/21] per memcg lru lock Alex Shi 2020-07-27 5:40 ` Alex Shi 2020-07-29 14:49 ` Alex Shi 2020-07-29 14:49 ` Alex Shi 2020-07-29 18:06 ` Hugh Dickins 2020-07-29 18:06 ` Hugh Dickins 2020-07-30 2:16 ` Alex Shi 2020-07-30 2:16 ` Alex Shi 2020-08-03 15:07 ` Michal Hocko 2020-08-03 15:07 ` Michal Hocko 2020-08-04 6:14 ` Alex Shi 2020-08-04 6:14 ` Alex Shi 2020-07-31 21:31 ` Alexander Duyck 2020-07-31 21:31 ` Alexander Duyck 2020-07-31 21:31 ` Alexander Duyck 2020-08-04 8:36 ` Alex Shi 2020-08-04 8:36 ` Alex Shi 2020-08-04 8:36 ` Alex Shi 2020-08-04 8:36 ` Alex Shi 2020-08-04 8:37 ` Alex Shi 2020-08-04 8:37 ` Alex Shi 2020-08-04 8:37 ` Alex Shi 2020-08-04 8:37 ` Alex Shi
Reply instructions: You may reply publicly to this message via plain-text email using any one of the following methods: * Save the following mbox file, import it into your mail client, and reply-to-all from there: mbox Avoid top-posting and favor interleaved quoting: https://en.wikipedia.org/wiki/Posting_style#Interleaved_style * Reply using the --to, --cc, and --in-reply-to switches of git-send-email(1): git send-email \ --in-reply-to=1fd45e69-3a50-aae8-bcc4-47d891a5e263@linux.alibaba.com \ --to=alex.shi@linux.alibaba.com \ --cc=akpm@linux-foundation.org \ --cc=alexander.duyck@gmail.com \ --cc=cgroups@vger.kernel.org \ --cc=daniel.m.jordan@oracle.com \ --cc=hannes@cmpxchg.org \ --cc=hughd@google.com \ --cc=iamjoonsoo.kim@lge.com \ --cc=khlebnikov@yandex-team.ru \ --cc=kirill@shutemov.name \ --cc=linux-kernel@vger.kernel.org \ --cc=linux-mm@kvack.org \ --cc=lkp@intel.com \ --cc=mgorman@techsingularity.net \ --cc=mhocko@kernel.org \ --cc=richard.weiyang@gmail.com \ --cc=rong.a.chen@intel.com \ --cc=shakeelb@google.com \ --cc=tj@kernel.org \ --cc=vdavydov.dev@gmail.com \ --cc=willy@infradead.org \ --cc=yang.shi@linux.alibaba.com \ /path/to/YOUR_REPLY https://kernel.org/pub/software/scm/git/docs/git-send-email.html * If your mail client supports setting the In-Reply-To header via mailto: links, try the mailto: linkBe sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes, see mirroring instructions on how to clone and mirror all data and code used by this external index.