All of lore.kernel.org
 help / color / mirror / Atom feed
From: Waiman Long <longman@redhat.com>
To: Muchun Song <songmuchun@bytedance.com>,
	hannes@cmpxchg.org, mhocko@kernel.org, roman.gushchin@linux.dev,
	shakeelb@google.com
Cc: cgroups@vger.kernel.org, linux-mm@kvack.org,
	linux-kernel@vger.kernel.org, duanxiongchun@bytedance.com
Subject: Re: [PATCH v4 04/11] mm: vmscan: rework move_pages_to_lru()
Date: Tue, 24 May 2022 15:52:22 -0400	[thread overview]
Message-ID: <78de6197-7de6-9fe7-9567-1321c06c6e9b@redhat.com> (raw)
In-Reply-To: <20220524060551.80037-5-songmuchun@bytedance.com>

On 5/24/22 02:05, Muchun Song wrote:
> In the later patch, we will reparent the LRU pages. The pages moved to
> appropriate LRU list can be reparented during the process of the
> move_pages_to_lru(). So holding a lruvec lock by the caller is wrong, we
> should use the more general interface of folio_lruvec_relock_irq() to
> acquire the correct lruvec lock.
>
> Signed-off-by: Muchun Song <songmuchun@bytedance.com>
> ---
>   mm/vmscan.c | 49 +++++++++++++++++++++++++------------------------
>   1 file changed, 25 insertions(+), 24 deletions(-)
>
> diff --git a/mm/vmscan.c b/mm/vmscan.c
> index 1678802e03e7..761d5e0dd78d 100644
> --- a/mm/vmscan.c
> +++ b/mm/vmscan.c
> @@ -2230,23 +2230,28 @@ static int too_many_isolated(struct pglist_data *pgdat, int file,
>    * move_pages_to_lru() moves pages from private @list to appropriate LRU list.
>    * On return, @list is reused as a list of pages to be freed by the caller.
>    *
> - * Returns the number of pages moved to the given lruvec.
> + * Returns the number of pages moved to the appropriate LRU list.
> + *
> + * Note: The caller must not hold any lruvec lock.
>    */
> -static unsigned int move_pages_to_lru(struct lruvec *lruvec,
> -				      struct list_head *list)
> +static unsigned int move_pages_to_lru(struct list_head *list)
>   {
> -	int nr_pages, nr_moved = 0;
> +	int nr_moved = 0;
> +	struct lruvec *lruvec = NULL;
>   	LIST_HEAD(pages_to_free);
> -	struct page *page;
>   
>   	while (!list_empty(list)) {
> -		page = lru_to_page(list);
> +		int nr_pages;
> +		struct folio *folio = lru_to_folio(list);
> +		struct page *page = &folio->page;
> +
> +		lruvec = folio_lruvec_relock_irq(folio, lruvec);
>   		VM_BUG_ON_PAGE(PageLRU(page), page);
>   		list_del(&page->lru);
>   		if (unlikely(!page_evictable(page))) {
> -			spin_unlock_irq(&lruvec->lru_lock);
> +			unlock_page_lruvec_irq(lruvec);
>   			putback_lru_page(page);
> -			spin_lock_irq(&lruvec->lru_lock);
> +			lruvec = NULL;
>   			continue;
>   		}
>   
> @@ -2267,20 +2272,16 @@ static unsigned int move_pages_to_lru(struct lruvec *lruvec,
>   			__clear_page_lru_flags(page);
>   
>   			if (unlikely(PageCompound(page))) {
> -				spin_unlock_irq(&lruvec->lru_lock);
> +				unlock_page_lruvec_irq(lruvec);
>   				destroy_compound_page(page);
> -				spin_lock_irq(&lruvec->lru_lock);
> +				lruvec = NULL;
>   			} else
>   				list_add(&page->lru, &pages_to_free);
>   
>   			continue;
>   		}
>   
> -		/*
> -		 * All pages were isolated from the same lruvec (and isolation
> -		 * inhibits memcg migration).
> -		 */
> -		VM_BUG_ON_PAGE(!folio_matches_lruvec(page_folio(page), lruvec), page);
> +		VM_BUG_ON_PAGE(!folio_matches_lruvec(folio, lruvec), page);
>   		add_page_to_lru_list(page, lruvec);
>   		nr_pages = thp_nr_pages(page);
>   		nr_moved += nr_pages;
> @@ -2288,6 +2289,8 @@ static unsigned int move_pages_to_lru(struct lruvec *lruvec,
>   			workingset_age_nonresident(lruvec, nr_pages);
>   	}
>   
> +	if (lruvec)
> +		unlock_page_lruvec_irq(lruvec);
>   	/*
>   	 * To save our caller's stack, now use input list for pages to free.
>   	 */
> @@ -2359,16 +2362,16 @@ shrink_inactive_list(unsigned long nr_to_scan, struct lruvec *lruvec,
>   
>   	nr_reclaimed = shrink_page_list(&page_list, pgdat, sc, &stat, false);
>   
> -	spin_lock_irq(&lruvec->lru_lock);
> -	move_pages_to_lru(lruvec, &page_list);
> +	move_pages_to_lru(&page_list);
>   
> +	local_irq_disable();
>   	__mod_node_page_state(pgdat, NR_ISOLATED_ANON + file, -nr_taken);
>   	item = current_is_kswapd() ? PGSTEAL_KSWAPD : PGSTEAL_DIRECT;
>   	if (!cgroup_reclaim(sc))
>   		__count_vm_events(item, nr_reclaimed);
>   	__count_memcg_events(lruvec_memcg(lruvec), item, nr_reclaimed);
>   	__count_vm_events(PGSTEAL_ANON + file, nr_reclaimed);
> -	spin_unlock_irq(&lruvec->lru_lock);
> +	local_irq_enable();
>   
>   	lru_note_cost(lruvec, file, stat.nr_pageout);
>   	mem_cgroup_uncharge_list(&page_list);
> @@ -2498,18 +2501,16 @@ static void shrink_active_list(unsigned long nr_to_scan,
>   	/*
>   	 * Move pages back to the lru list.
>   	 */
> -	spin_lock_irq(&lruvec->lru_lock);
> -
> -	nr_activate = move_pages_to_lru(lruvec, &l_active);
> -	nr_deactivate = move_pages_to_lru(lruvec, &l_inactive);
> +	nr_activate = move_pages_to_lru(&l_active);
> +	nr_deactivate = move_pages_to_lru(&l_inactive);
>   	/* Keep all free pages in l_active list */
>   	list_splice(&l_inactive, &l_active);
>   
> +	local_irq_disable();
>   	__count_vm_events(PGDEACTIVATE, nr_deactivate);
>   	__count_memcg_events(lruvec_memcg(lruvec), PGDEACTIVATE, nr_deactivate);
> -
>   	__mod_node_page_state(pgdat, NR_ISOLATED_ANON + file, -nr_taken);
> -	spin_unlock_irq(&lruvec->lru_lock);
> +	local_irq_enable();
>   
>   	mem_cgroup_uncharge_list(&l_active);
>   	free_unref_page_list(&l_active);

Note that the RT engineers will likely change the 
local_irq_disable()/local_irq_enable() to 
local_lock_irq()/local_unlock_irq().

Cheers,
Longman


WARNING: multiple messages have this Message-ID (diff)
From: Waiman Long <longman-H+wXaHxf7aLQT0dZR+AlfA@public.gmane.org>
To: Muchun Song <songmuchun-EC8Uxl6Npydl57MIdRCFDg@public.gmane.org>,
	hannes-druUgvl0LCNAfugRpC6u6w@public.gmane.org,
	mhocko-DgEjT+Ai2ygdnm+yROfE0A@public.gmane.org,
	roman.gushchin-fxUVXftIFDnyG1zEObXtfA@public.gmane.org,
	shakeelb-hpIqsD4AKlfQT0dZR+AlfA@public.gmane.org
Cc: cgroups-u79uwXL29TY76Z2rM5mHXA@public.gmane.org,
	linux-mm-Bw31MaZKKs3YtjvyW6yDsg@public.gmane.org,
	linux-kernel-u79uwXL29TY76Z2rM5mHXA@public.gmane.org,
	duanxiongchun-EC8Uxl6Npydl57MIdRCFDg@public.gmane.org
Subject: Re: [PATCH v4 04/11] mm: vmscan: rework move_pages_to_lru()
Date: Tue, 24 May 2022 15:52:22 -0400	[thread overview]
Message-ID: <78de6197-7de6-9fe7-9567-1321c06c6e9b@redhat.com> (raw)
In-Reply-To: <20220524060551.80037-5-songmuchun-EC8Uxl6Npydl57MIdRCFDg@public.gmane.org>

On 5/24/22 02:05, Muchun Song wrote:
> In the later patch, we will reparent the LRU pages. The pages moved to
> appropriate LRU list can be reparented during the process of the
> move_pages_to_lru(). So holding a lruvec lock by the caller is wrong, we
> should use the more general interface of folio_lruvec_relock_irq() to
> acquire the correct lruvec lock.
>
> Signed-off-by: Muchun Song <songmuchun-EC8Uxl6Npydl57MIdRCFDg@public.gmane.org>
> ---
>   mm/vmscan.c | 49 +++++++++++++++++++++++++------------------------
>   1 file changed, 25 insertions(+), 24 deletions(-)
>
> diff --git a/mm/vmscan.c b/mm/vmscan.c
> index 1678802e03e7..761d5e0dd78d 100644
> --- a/mm/vmscan.c
> +++ b/mm/vmscan.c
> @@ -2230,23 +2230,28 @@ static int too_many_isolated(struct pglist_data *pgdat, int file,
>    * move_pages_to_lru() moves pages from private @list to appropriate LRU list.
>    * On return, @list is reused as a list of pages to be freed by the caller.
>    *
> - * Returns the number of pages moved to the given lruvec.
> + * Returns the number of pages moved to the appropriate LRU list.
> + *
> + * Note: The caller must not hold any lruvec lock.
>    */
> -static unsigned int move_pages_to_lru(struct lruvec *lruvec,
> -				      struct list_head *list)
> +static unsigned int move_pages_to_lru(struct list_head *list)
>   {
> -	int nr_pages, nr_moved = 0;
> +	int nr_moved = 0;
> +	struct lruvec *lruvec = NULL;
>   	LIST_HEAD(pages_to_free);
> -	struct page *page;
>   
>   	while (!list_empty(list)) {
> -		page = lru_to_page(list);
> +		int nr_pages;
> +		struct folio *folio = lru_to_folio(list);
> +		struct page *page = &folio->page;
> +
> +		lruvec = folio_lruvec_relock_irq(folio, lruvec);
>   		VM_BUG_ON_PAGE(PageLRU(page), page);
>   		list_del(&page->lru);
>   		if (unlikely(!page_evictable(page))) {
> -			spin_unlock_irq(&lruvec->lru_lock);
> +			unlock_page_lruvec_irq(lruvec);
>   			putback_lru_page(page);
> -			spin_lock_irq(&lruvec->lru_lock);
> +			lruvec = NULL;
>   			continue;
>   		}
>   
> @@ -2267,20 +2272,16 @@ static unsigned int move_pages_to_lru(struct lruvec *lruvec,
>   			__clear_page_lru_flags(page);
>   
>   			if (unlikely(PageCompound(page))) {
> -				spin_unlock_irq(&lruvec->lru_lock);
> +				unlock_page_lruvec_irq(lruvec);
>   				destroy_compound_page(page);
> -				spin_lock_irq(&lruvec->lru_lock);
> +				lruvec = NULL;
>   			} else
>   				list_add(&page->lru, &pages_to_free);
>   
>   			continue;
>   		}
>   
> -		/*
> -		 * All pages were isolated from the same lruvec (and isolation
> -		 * inhibits memcg migration).
> -		 */
> -		VM_BUG_ON_PAGE(!folio_matches_lruvec(page_folio(page), lruvec), page);
> +		VM_BUG_ON_PAGE(!folio_matches_lruvec(folio, lruvec), page);
>   		add_page_to_lru_list(page, lruvec);
>   		nr_pages = thp_nr_pages(page);
>   		nr_moved += nr_pages;
> @@ -2288,6 +2289,8 @@ static unsigned int move_pages_to_lru(struct lruvec *lruvec,
>   			workingset_age_nonresident(lruvec, nr_pages);
>   	}
>   
> +	if (lruvec)
> +		unlock_page_lruvec_irq(lruvec);
>   	/*
>   	 * To save our caller's stack, now use input list for pages to free.
>   	 */
> @@ -2359,16 +2362,16 @@ shrink_inactive_list(unsigned long nr_to_scan, struct lruvec *lruvec,
>   
>   	nr_reclaimed = shrink_page_list(&page_list, pgdat, sc, &stat, false);
>   
> -	spin_lock_irq(&lruvec->lru_lock);
> -	move_pages_to_lru(lruvec, &page_list);
> +	move_pages_to_lru(&page_list);
>   
> +	local_irq_disable();
>   	__mod_node_page_state(pgdat, NR_ISOLATED_ANON + file, -nr_taken);
>   	item = current_is_kswapd() ? PGSTEAL_KSWAPD : PGSTEAL_DIRECT;
>   	if (!cgroup_reclaim(sc))
>   		__count_vm_events(item, nr_reclaimed);
>   	__count_memcg_events(lruvec_memcg(lruvec), item, nr_reclaimed);
>   	__count_vm_events(PGSTEAL_ANON + file, nr_reclaimed);
> -	spin_unlock_irq(&lruvec->lru_lock);
> +	local_irq_enable();
>   
>   	lru_note_cost(lruvec, file, stat.nr_pageout);
>   	mem_cgroup_uncharge_list(&page_list);
> @@ -2498,18 +2501,16 @@ static void shrink_active_list(unsigned long nr_to_scan,
>   	/*
>   	 * Move pages back to the lru list.
>   	 */
> -	spin_lock_irq(&lruvec->lru_lock);
> -
> -	nr_activate = move_pages_to_lru(lruvec, &l_active);
> -	nr_deactivate = move_pages_to_lru(lruvec, &l_inactive);
> +	nr_activate = move_pages_to_lru(&l_active);
> +	nr_deactivate = move_pages_to_lru(&l_inactive);
>   	/* Keep all free pages in l_active list */
>   	list_splice(&l_inactive, &l_active);
>   
> +	local_irq_disable();
>   	__count_vm_events(PGDEACTIVATE, nr_deactivate);
>   	__count_memcg_events(lruvec_memcg(lruvec), PGDEACTIVATE, nr_deactivate);
> -
>   	__mod_node_page_state(pgdat, NR_ISOLATED_ANON + file, -nr_taken);
> -	spin_unlock_irq(&lruvec->lru_lock);
> +	local_irq_enable();
>   
>   	mem_cgroup_uncharge_list(&l_active);
>   	free_unref_page_list(&l_active);

Note that the RT engineers will likely change the 
local_irq_disable()/local_irq_enable() to 
local_lock_irq()/local_unlock_irq().

Cheers,
Longman


  parent reply	other threads:[~2022-05-24 19:52 UTC|newest]

Thread overview: 93+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2022-05-24  6:05 [PATCH v4 00/11] Use obj_cgroup APIs to charge the LRU pages Muchun Song
2022-05-24  6:05 ` Muchun Song
2022-05-24  6:05 ` [PATCH v4 01/11] mm: memcontrol: prepare objcg API for non-kmem usage Muchun Song
2022-05-24  6:05   ` Muchun Song
2022-05-24 19:01   ` Johannes Weiner
2022-05-24 19:01     ` Johannes Weiner
2022-05-25  8:46     ` Muchun Song
2022-05-25  8:46       ` Muchun Song
2022-05-25  2:36   ` Roman Gushchin
2022-05-25  2:36     ` Roman Gushchin
2022-05-25  7:57     ` Muchun Song
2022-05-25  7:57       ` Muchun Song
2022-05-25 12:37       ` Johannes Weiner
2022-05-25 12:37         ` Johannes Weiner
2022-05-25 13:08         ` Muchun Song
2022-05-25 13:08           ` Muchun Song
2022-05-24  6:05 ` [PATCH v4 02/11] mm: memcontrol: introduce compact_folio_lruvec_lock_irqsave Muchun Song
2022-05-24  6:05   ` Muchun Song
2022-05-24 19:22   ` Johannes Weiner
2022-05-24 19:22     ` Johannes Weiner
2022-05-25  9:38     ` Muchun Song
2022-05-25  9:38       ` Muchun Song
2022-05-24  6:05 ` [PATCH v4 03/11] mm: memcontrol: make lruvec lock safe when LRU pages are reparented Muchun Song
2022-05-24  6:05   ` Muchun Song
2022-05-24 19:23   ` Waiman Long
2022-05-25 10:20     ` Muchun Song
2022-05-25 10:20       ` Muchun Song
2022-05-25 14:59       ` Waiman Long
2022-05-25 14:59         ` Waiman Long
2022-05-24 19:27   ` Johannes Weiner
2022-05-24 19:27     ` Johannes Weiner
2022-05-25  9:53     ` Muchun Song
2022-05-25  9:53       ` Muchun Song
2022-05-25 12:30       ` Johannes Weiner
2022-05-25 12:30         ` Johannes Weiner
2022-05-25 13:03         ` Muchun Song
2022-05-25 13:03           ` Muchun Song
2022-05-25 14:48           ` Johannes Weiner
2022-05-25 14:48             ` Johannes Weiner
2022-05-25 15:38             ` Muchun Song
2022-05-25 15:38               ` Muchun Song
2022-05-26 20:17               ` Waiman Long
2022-05-26 20:17                 ` Waiman Long
2022-05-27  2:55                 ` Muchun Song
2022-05-27  2:55                   ` Muchun Song
2022-05-24  6:05 ` [PATCH v4 04/11] mm: vmscan: rework move_pages_to_lru() Muchun Song
2022-05-24  6:05   ` Muchun Song
2022-05-24 19:38   ` Johannes Weiner
2022-05-24 19:38     ` Johannes Weiner
2022-05-25 11:38     ` Muchun Song
2022-05-25 11:38       ` Muchun Song
2022-05-24 19:52   ` Waiman Long [this message]
2022-05-24 19:52     ` Waiman Long
2022-05-25 11:43     ` Muchun Song
2022-05-25 11:43       ` Muchun Song
2022-05-25  2:43   ` Roman Gushchin
2022-05-25  2:43     ` Roman Gushchin
2022-05-25 11:41     ` Muchun Song
2022-05-25 11:41       ` Muchun Song
2022-05-24  6:05 ` [PATCH v4 05/11] mm: thp: introduce folio_split_queue_lock{_irqsave}() Muchun Song
2022-05-24  6:05   ` Muchun Song
2022-05-24  6:05 ` [PATCH v4 06/11] mm: thp: make split queue lock safe when LRU pages are reparented Muchun Song
2022-05-24  6:05   ` Muchun Song
2022-05-25  2:54   ` Roman Gushchin
2022-05-25  2:54     ` Roman Gushchin
2022-05-25 11:44     ` Muchun Song
2022-05-25 11:44       ` Muchun Song
2022-05-24  6:05 ` [PATCH v4 07/11] mm: memcontrol: make all the callers of {folio,page}_memcg() safe Muchun Song
2022-05-24  6:05   ` Muchun Song
2022-05-25  3:03   ` Roman Gushchin
2022-05-25  3:03     ` Roman Gushchin
2022-05-25 11:51     ` Muchun Song
2022-05-25 11:51       ` Muchun Song
2022-05-24  6:05 ` [PATCH v4 08/11] mm: memcontrol: introduce memcg_reparent_ops Muchun Song
2022-05-24  6:05   ` Muchun Song
2022-05-24  6:05 ` [PATCH v4 09/11] mm: memcontrol: use obj_cgroup APIs to charge the LRU pages Muchun Song
2022-05-24  6:05   ` Muchun Song
2022-05-24 12:29   ` kernel test robot
2022-05-24 18:16   ` kernel test robot
2022-05-24 18:16     ` kernel test robot
2022-05-25  7:14   ` [mm] bec0ae1210: WARNING:possible_recursive_locking_detected kernel test robot
2022-05-25  7:14     ` kernel test robot
2022-05-24  6:05 ` [PATCH v4 10/11] mm: lru: add VM_BUG_ON_FOLIO to lru maintenance function Muchun Song
2022-05-24  6:05   ` Muchun Song
2022-05-24 19:44   ` Johannes Weiner
2022-05-24 19:44     ` Johannes Weiner
2022-05-25 11:59     ` Muchun Song
2022-05-25 11:59       ` Muchun Song
2022-05-25  2:40   ` Roman Gushchin
2022-05-25  2:40     ` Roman Gushchin
2022-05-25 11:58     ` Muchun Song
2022-05-24  6:05 ` [PATCH v4 11/11] mm: lru: use lruvec lock to serialize memcg changes Muchun Song
2022-05-24  6:05   ` Muchun Song

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=78de6197-7de6-9fe7-9567-1321c06c6e9b@redhat.com \
    --to=longman@redhat.com \
    --cc=cgroups@vger.kernel.org \
    --cc=duanxiongchun@bytedance.com \
    --cc=hannes@cmpxchg.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-mm@kvack.org \
    --cc=mhocko@kernel.org \
    --cc=roman.gushchin@linux.dev \
    --cc=shakeelb@google.com \
    --cc=songmuchun@bytedance.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.