From: Steven Price <steven.price@arm.com>
To: Chinwen Chang <chinwen.chang@mediatek.com>,
Matthias Brugger <matthias.bgg@gmail.com>,
Michel Lespinasse <walken@google.com>,
Andrew Morton <akpm@linux-foundation.org>,
Vlastimil Babka <vbabka@suse.cz>,
Daniel Jordan <daniel.m.jordan@oracle.com>,
Davidlohr Bueso <dbueso@suse.de>,
Alexey Dobriyan <adobriyan@gmail.com>,
"Matthew Wilcox (Oracle)" <willy@infradead.org>,
Jason Gunthorpe <jgg@ziepe.ca>, Song Liu <songliubraving@fb.com>,
Jimmy Assarsson <jimmyassarsson@gmail.com>,
Huang Ying <ying.huang@intel.com>,
Daniel Kiss <daniel.kiss@arm.com>,
Laurent Dufour <ldufour@linux.ibm.com>
Cc: linux-kernel@vger.kernel.org,
linux-arm-kernel@lists.infradead.org,
linux-mediatek@lists.infradead.org,
linux-fsdevel@vger.kernel.org, wsd_upstream@mediatek.com
Subject: Re: [PATCH v3 2/3] mm: smaps*: extend smap_gather_stats to support specified beginning
Date: Mon, 17 Aug 2020 09:38:37 +0100 [thread overview]
Message-ID: <5f4eb892-4980-8245-b93d-5358ffa01abe@arm.com> (raw)
In-Reply-To: <1597472419-32314-3-git-send-email-chinwen.chang@mediatek.com>
On 15/08/2020 07:20, Chinwen Chang wrote:
> Extend smap_gather_stats to support indicated beginning address at
> which it should start gathering. To achieve the goal, we add a new
> parameter @start assigned by the caller and try to refactor it for
> simplicity.
>
> If @start is 0, it will use the range of @vma for gathering.
>
> Change since v2:
> - This is a new change to make the retry behavior of smaps_rollup
> - more complete as suggested by Michel [1]
>
> [1] https://lore.kernel.org/lkml/CANN689FtCsC71cjAjs0GPspOhgo_HRj+diWsoU1wr98YPktgWg@mail.gmail.com/
>
> Signed-off-by: Chinwen Chang <chinwen.chang@mediatek.com>
> CC: Michel Lespinasse <walken@google.com>
> CC: Steven Price <steven.price@arm.com>
LGTM
Reviewed-by: Steven Price <steven.price@arm.com>
Steve
> ---
> fs/proc/task_mmu.c | 30 ++++++++++++++++++++++--------
> 1 file changed, 22 insertions(+), 8 deletions(-)
>
> diff --git a/fs/proc/task_mmu.c b/fs/proc/task_mmu.c
> index dbda449..76e623a 100644
> --- a/fs/proc/task_mmu.c
> +++ b/fs/proc/task_mmu.c
> @@ -723,9 +723,21 @@ static int smaps_hugetlb_range(pte_t *pte, unsigned long hmask,
> .pte_hole = smaps_pte_hole,
> };
>
> +/*
> + * Gather mem stats from @vma with the indicated beginning
> + * address @start, and keep them in @mss.
> + *
> + * Use vm_start of @vma as the beginning address if @start is 0.
> + */
> static void smap_gather_stats(struct vm_area_struct *vma,
> - struct mem_size_stats *mss)
> + struct mem_size_stats *mss, unsigned long start)
> {
> + const struct mm_walk_ops *ops = &smaps_walk_ops;
> +
> + /* Invalid start */
> + if (start >= vma->vm_end)
> + return;
> +
> #ifdef CONFIG_SHMEM
> /* In case of smaps_rollup, reset the value from previous vma */
> mss->check_shmem_swap = false;
> @@ -742,18 +754,20 @@ static void smap_gather_stats(struct vm_area_struct *vma,
> */
> unsigned long shmem_swapped = shmem_swap_usage(vma);
>
> - if (!shmem_swapped || (vma->vm_flags & VM_SHARED) ||
> - !(vma->vm_flags & VM_WRITE)) {
> + if (!start && (!shmem_swapped || (vma->vm_flags & VM_SHARED) ||
> + !(vma->vm_flags & VM_WRITE))) {
> mss->swap += shmem_swapped;
> } else {
> mss->check_shmem_swap = true;
> - walk_page_vma(vma, &smaps_shmem_walk_ops, mss);
> - return;
> + ops = &smaps_shmem_walk_ops;
> }
> }
> #endif
> /* mmap_lock is held in m_start */
> - walk_page_vma(vma, &smaps_walk_ops, mss);
> + if (!start)
> + walk_page_vma(vma, ops, mss);
> + else
> + walk_page_range(vma->vm_mm, start, vma->vm_end, ops, mss);
> }
>
> #define SEQ_PUT_DEC(str, val) \
> @@ -805,7 +819,7 @@ static int show_smap(struct seq_file *m, void *v)
>
> memset(&mss, 0, sizeof(mss));
>
> - smap_gather_stats(vma, &mss);
> + smap_gather_stats(vma, &mss, 0);
>
> show_map_vma(m, vma);
>
> @@ -854,7 +868,7 @@ static int show_smaps_rollup(struct seq_file *m, void *v)
> hold_task_mempolicy(priv);
>
> for (vma = priv->mm->mmap; vma; vma = vma->vm_next) {
> - smap_gather_stats(vma, &mss);
> + smap_gather_stats(vma, &mss, 0);
> last_vma_end = vma->vm_end;
> }
>
>
next parent reply other threads:[~2020-08-17 8:38 UTC|newest]
Thread overview: 3+ messages / expand[flat|nested] mbox.gz Atom feed top
[not found] <1597472419-32314-1-git-send-email-chinwen.chang@mediatek.com>
[not found] ` <1597472419-32314-3-git-send-email-chinwen.chang@mediatek.com>
2020-08-17 8:38 ` Steven Price [this message]
[not found] ` <1597472419-32314-4-git-send-email-chinwen.chang@mediatek.com>
2020-08-17 8:38 ` [PATCH v3 3/3] mm: proc: smaps_rollup: do not stall write attempts on mmap_lock Steven Price
2020-08-17 9:15 ` Chinwen Chang
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=5f4eb892-4980-8245-b93d-5358ffa01abe@arm.com \
--to=steven.price@arm.com \
--cc=adobriyan@gmail.com \
--cc=akpm@linux-foundation.org \
--cc=chinwen.chang@mediatek.com \
--cc=daniel.kiss@arm.com \
--cc=daniel.m.jordan@oracle.com \
--cc=dbueso@suse.de \
--cc=jgg@ziepe.ca \
--cc=jimmyassarsson@gmail.com \
--cc=ldufour@linux.ibm.com \
--cc=linux-arm-kernel@lists.infradead.org \
--cc=linux-fsdevel@vger.kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mediatek@lists.infradead.org \
--cc=matthias.bgg@gmail.com \
--cc=songliubraving@fb.com \
--cc=vbabka@suse.cz \
--cc=walken@google.com \
--cc=willy@infradead.org \
--cc=wsd_upstream@mediatek.com \
--cc=ying.huang@intel.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).