From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-3.8 required=3.0 tests=BAYES_00, HEADER_FROM_DIFFERENT_DOMAINS,MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS autolearn=no autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 86551C433B4 for ; Wed, 14 Apr 2021 08:27:40 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id 52620613C1 for ; Wed, 14 Apr 2021 08:27:40 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S229803AbhDNI17 (ORCPT ); Wed, 14 Apr 2021 04:27:59 -0400 Received: from mga18.intel.com ([134.134.136.126]:24956 "EHLO mga18.intel.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229467AbhDNI1y (ORCPT ); Wed, 14 Apr 2021 04:27:54 -0400 IronPort-SDR: gj56hEk6t/z9wBlpjYXHCnE5N92jjbsikdwmF2Oxe0QFyAeykTu5C6C3/LO2v7KPUOaH7p0Jmf jJOKqyxf8HjA== X-IronPort-AV: E=McAfee;i="6200,9189,9953"; a="182100361" X-IronPort-AV: E=Sophos;i="5.82,221,1613462400"; d="scan'208";a="182100361" Received: from orsmga008.jf.intel.com ([10.7.209.65]) by orsmga106.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 14 Apr 2021 01:27:32 -0700 IronPort-SDR: 14SQdVLTCk4i+9pQiSFX9sYdGYvZAlpQ1aaayf9l30dCbDnh/Odfq6h/GxWxQN25SOgX9jLpl7 tb49LF92dXug== X-IronPort-AV: E=Sophos;i="5.82,221,1613462400"; d="scan'208";a="424627039" Received: from yhuang6-desk1.sh.intel.com (HELO yhuang6-desk1.ccr.corp.intel.com) ([10.239.13.1]) by orsmga008-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 14 Apr 2021 01:27:24 -0700 From: "Huang, Ying" To: Yu Zhao Cc: Rik van Riel , Dave Chinner , Jens Axboe , SeongJae Park , Linux-MM , Andi Kleen , Andrew Morton , Benjamin Manes , Dave Hansen , Hillf Danton , Johannes Weiner , Jonathan Corbet , Joonsoo Kim , Matthew Wilcox , Mel Gorman , Miaohe Lin , Michael Larabel , Michal Hocko , Michel Lespinasse , Roman Gushchin , Rong Chen , SeongJae Park , Tim Chen , Vlastimil Babka , Yang Shi , Zi Yan , linux-kernel , lkp@lists.01.org, Kernel Page Reclaim v2 Subject: Re: [PATCH v2 00/16] Multigenerational LRU Framework References: <20210413075155.32652-1-sjpark@amazon.de> <3ddd4f8a-8e51-662b-df11-a63a0e75b2bc@kernel.dk> <20210413231436.GF63242@dread.disaster.area> <87tuo9qtmd.fsf@yhuang6-desk1.ccr.corp.intel.com> Date: Wed, 14 Apr 2021 16:27:22 +0800 In-Reply-To: (Yu Zhao's message of "Wed, 14 Apr 2021 01:58:56 -0600") Message-ID: <87lf9lqnit.fsf@yhuang6-desk1.ccr.corp.intel.com> User-Agent: Gnus/5.13 (Gnus v5.13) Emacs/27.1 (gnu/linux) MIME-Version: 1.0 Content-Type: text/plain; charset=ascii Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Yu Zhao writes: > On Wed, Apr 14, 2021 at 12:15 AM Huang, Ying wrote: >> >> Yu Zhao writes: >> >> > On Tue, Apr 13, 2021 at 8:30 PM Rik van Riel wrote: >> >> >> >> On Wed, 2021-04-14 at 09:14 +1000, Dave Chinner wrote: >> >> > On Tue, Apr 13, 2021 at 10:13:24AM -0600, Jens Axboe wrote: >> >> > >> >> > > The initial posting of this patchset did no better, in fact it did >> >> > > a bit >> >> > > worse. Performance dropped to the same levels and kswapd was using >> >> > > as >> >> > > much CPU as before, but on top of that we also got excessive >> >> > > swapping. >> >> > > Not at a high rate, but 5-10MB/sec continually. >> >> > > >> >> > > I had some back and forths with Yu Zhao and tested a few new >> >> > > revisions, >> >> > > and the current series does much better in this regard. Performance >> >> > > still dips a bit when page cache fills, but not nearly as much, and >> >> > > kswapd is using less CPU than before. >> >> > >> >> > Profiles would be interesting, because it sounds to me like reclaim >> >> > *might* be batching page cache removal better (e.g. fewer, larger >> >> > batches) and so spending less time contending on the mapping tree >> >> > lock... >> >> > >> >> > IOWs, I suspect this result might actually be a result of less lock >> >> > contention due to a change in batch processing characteristics of >> >> > the new algorithm rather than it being a "better" algorithm... >> >> >> >> That seems quite likely to me, given the issues we have >> >> had with virtual scan reclaim algorithms in the past. >> > >> > Hi Rik, >> > >> > Let paste the code so we can move beyond the "batching" hypothesis: >> > >> > static int __remove_mapping(struct address_space *mapping, struct page >> > *page, >> > bool reclaimed, struct mem_cgroup *target_memcg) >> > { >> > unsigned long flags; >> > int refcount; >> > void *shadow = NULL; >> > >> > BUG_ON(!PageLocked(page)); >> > BUG_ON(mapping != page_mapping(page)); >> > >> > xa_lock_irqsave(&mapping->i_pages, flags); >> > >> >> SeongJae, what is this algorithm supposed to do when faced >> >> with situations like this: >> > >> > I'll assume the questions were directed at me, not SeongJae. >> > >> >> 1) Running on a system with 8 NUMA nodes, and >> >> memory >> >> pressure in one of those nodes. >> >> 2) Running PostgresQL or Oracle, with hundreds of >> >> processes mapping the same (very large) shared >> >> memory segment. >> >> >> >> How do you keep your algorithm from falling into the worst >> >> case virtual scanning scenarios that were crippling the >> >> 2.4 kernel 15+ years ago on systems with just a few GB of >> >> memory? >> > >> > There is a fundamental shift: that time we were scanning for cold pages, >> > and nowadays we are scanning for hot pages. >> > >> > I'd be surprised if scanning for cold pages didn't fall apart, because it'd >> > find most of the entries accessed, if they are present at all. >> > >> > Scanning for hot pages, on the other hand, is way better. Let me just >> > reiterate: >> > 1) It will not scan page tables from processes that have been sleeping >> > since the last scan. >> > 2) It will not scan PTE tables under non-leaf PMD entries that do not >> > have the accessed bit set, when >> > CONFIG_HAVE_ARCH_PARENT_PMD_YOUNG=y. >> > 3) It will not zigzag between the PGD table and the same PMD or PTE >> > table spanning multiple VMAs. In other words, it finishes all the >> > VMAs with the range of the same PMD or PTE table before it returns >> > to the PGD table. This optimizes workloads that have large numbers >> > of tiny VMAs, especially when CONFIG_PGTABLE_LEVELS=5. >> > >> > So the cost is roughly proportional to the number of referenced pages it >> > discovers. If there is no memory pressure, no scanning at all. For a system >> > under heavy memory pressure, most of the pages are referenced (otherwise >> > why would it be under memory pressure?), and if we use the rmap, we need to >> > scan a lot of pages anyway. Why not just scan them all? >> >> This may be not the case. For rmap scanning, it's possible to scan only >> a small portion of memory. But with the page table scanning, you need >> to scan almost all (I understand you have some optimization as above). > > Hi Ying, > > Let's take a step back. > > For the sake of discussion, when does the scanning have to happen? Can > we agree that the simplest answer is when we have evicted all inactive > pages? > > If so, my next question is who's filled in the memory space previously > occupied by those inactive pages? Newly faulted in pages, right? They > have the accessed bit set, and we can't evict them without scanning > them first, would you agree? > > And there are also existing active pages, and they were protected from > eviction. But now we need to deactivate some of them. Do you think > whether they'd have been used or not since the last scan? (Remember > they were active.) > > You mentioned "a small portion" and "almost all". How do you interpret > them in terms of these steps? > > Intuitively, "a small portion" and "almost all" seem right. But our > observations from *millions* of machines say the ratio of > pgscan_kswapd to pgsteal_kswapd is well over 7 when anon percentage is >>90%. Unlikely streaming files, it doesn't make sense to "stream" anon > memory. What I said is that it is "POSSIBLE" to scan only a small portion of memory. Whether and in which cases to do that depends on the policy we choose. I didn't say we have chosen to do that for all cases. >> As Rik shown in the test case above, there may be memory pressure on >> only one of 8 NUMA nodes (because of NUMA binding?). Then ramp scanning >> only needs to scan pages in this node, while the page table scanning may >> need to scan pages in other nodes too. > > Yes, and this is on my to-do list in the patchset: > > To-do List > ========== > KVM Optimization > ---------------- > Support shadow page table scanning. > > NUMA Optimization > ----------------- > Support NUMA policies and per-node RSS counters. > > We only can move forward one step at a time. Fair? You don't need to implement that now definitely. But we can discuss the possible solution now. Note that it's possible that only some processes are bound to some NUMA nodes, while other processes aren't bound. Best Regards, Huang, Ying From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-3.8 required=3.0 tests=BAYES_00, HEADER_FROM_DIFFERENT_DOMAINS,MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS autolearn=no autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 05F31C433ED for ; Wed, 14 Apr 2021 08:27:37 +0000 (UTC) Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by mail.kernel.org (Postfix) with ESMTP id 96239613C1 for ; Wed, 14 Apr 2021 08:27:36 +0000 (UTC) DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org 96239613C1 Authentication-Results: mail.kernel.org; dmarc=fail (p=none dis=none) header.from=intel.com Authentication-Results: mail.kernel.org; spf=pass smtp.mailfrom=owner-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix) id 2104D6B0036; Wed, 14 Apr 2021 04:27:36 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id 1BF866B006C; Wed, 14 Apr 2021 04:27:36 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id 0886D6B0078; Wed, 14 Apr 2021 04:27:36 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from forelay.hostedemail.com (smtprelay0047.hostedemail.com [216.40.44.47]) by kanga.kvack.org (Postfix) with ESMTP id E08D26B0036 for ; Wed, 14 Apr 2021 04:27:35 -0400 (EDT) Received: from smtpin38.hostedemail.com (10.5.19.251.rfc1918.com [10.5.19.251]) by forelay01.hostedemail.com (Postfix) with ESMTP id 9913C181DD422 for ; Wed, 14 Apr 2021 08:27:35 +0000 (UTC) X-FDA: 78030293670.38.09B8D5A Received: from mga05.intel.com (mga05.intel.com [192.55.52.43]) by imf07.hostedemail.com (Postfix) with ESMTP id 7690FA0003A5 for ; Wed, 14 Apr 2021 08:27:34 +0000 (UTC) IronPort-SDR: uidlI8ZJ+bZEmwhuKq1567Fs4N8VFHa0fYrwhQK0bixOEl1RrclS4EaQRJ/pVtTYheFg2kQ5Dj KmX5sji48cPQ== X-IronPort-AV: E=McAfee;i="6200,9189,9953"; a="279906527" X-IronPort-AV: E=Sophos;i="5.82,221,1613462400"; d="scan'208";a="279906527" Received: from orsmga008.jf.intel.com ([10.7.209.65]) by fmsmga105.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 14 Apr 2021 01:27:33 -0700 IronPort-SDR: 14SQdVLTCk4i+9pQiSFX9sYdGYvZAlpQ1aaayf9l30dCbDnh/Odfq6h/GxWxQN25SOgX9jLpl7 tb49LF92dXug== X-IronPort-AV: E=Sophos;i="5.82,221,1613462400"; d="scan'208";a="424627039" Received: from yhuang6-desk1.sh.intel.com (HELO yhuang6-desk1.ccr.corp.intel.com) ([10.239.13.1]) by orsmga008-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 14 Apr 2021 01:27:24 -0700 From: "Huang, Ying" To: Yu Zhao Cc: Rik van Riel , Dave Chinner , Jens Axboe , SeongJae Park , Linux-MM , Andi Kleen , Andrew Morton , Benjamin Manes , Dave Hansen , Hillf Danton , Johannes Weiner , Jonathan Corbet , Joonsoo Kim , Matthew Wilcox , Mel Gorman , Miaohe Lin , Michael Larabel , Michal Hocko , Michel Lespinasse , Roman Gushchin , Rong Chen , SeongJae Park , Tim Chen , Vlastimil Babka , Yang Shi , Zi Yan , linux-kernel , lkp@lists.01.org, Kernel Page Reclaim v2 Subject: Re: [PATCH v2 00/16] Multigenerational LRU Framework References: <20210413075155.32652-1-sjpark@amazon.de> <3ddd4f8a-8e51-662b-df11-a63a0e75b2bc@kernel.dk> <20210413231436.GF63242@dread.disaster.area> <87tuo9qtmd.fsf@yhuang6-desk1.ccr.corp.intel.com> Date: Wed, 14 Apr 2021 16:27:22 +0800 In-Reply-To: (Yu Zhao's message of "Wed, 14 Apr 2021 01:58:56 -0600") Message-ID: <87lf9lqnit.fsf@yhuang6-desk1.ccr.corp.intel.com> User-Agent: Gnus/5.13 (Gnus v5.13) Emacs/27.1 (gnu/linux) MIME-Version: 1.0 Content-Type: text/plain; charset=ascii X-Rspamd-Server: rspam01 X-Rspamd-Queue-Id: 7690FA0003A5 X-Stat-Signature: 66siejrupmtip6z44ihp4jqkibsexmij Received-SPF: none (intel.com>: No applicable sender policy available) receiver=imf07; identity=mailfrom; envelope-from=""; helo=mga05.intel.com; client-ip=192.55.52.43 X-HE-DKIM-Result: none/none X-HE-Tag: 1618388854-881866 X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: Yu Zhao writes: > On Wed, Apr 14, 2021 at 12:15 AM Huang, Ying wrote: >> >> Yu Zhao writes: >> >> > On Tue, Apr 13, 2021 at 8:30 PM Rik van Riel wrote: >> >> >> >> On Wed, 2021-04-14 at 09:14 +1000, Dave Chinner wrote: >> >> > On Tue, Apr 13, 2021 at 10:13:24AM -0600, Jens Axboe wrote: >> >> > >> >> > > The initial posting of this patchset did no better, in fact it did >> >> > > a bit >> >> > > worse. Performance dropped to the same levels and kswapd was using >> >> > > as >> >> > > much CPU as before, but on top of that we also got excessive >> >> > > swapping. >> >> > > Not at a high rate, but 5-10MB/sec continually. >> >> > > >> >> > > I had some back and forths with Yu Zhao and tested a few new >> >> > > revisions, >> >> > > and the current series does much better in this regard. Performance >> >> > > still dips a bit when page cache fills, but not nearly as much, and >> >> > > kswapd is using less CPU than before. >> >> > >> >> > Profiles would be interesting, because it sounds to me like reclaim >> >> > *might* be batching page cache removal better (e.g. fewer, larger >> >> > batches) and so spending less time contending on the mapping tree >> >> > lock... >> >> > >> >> > IOWs, I suspect this result might actually be a result of less lock >> >> > contention due to a change in batch processing characteristics of >> >> > the new algorithm rather than it being a "better" algorithm... >> >> >> >> That seems quite likely to me, given the issues we have >> >> had with virtual scan reclaim algorithms in the past. >> > >> > Hi Rik, >> > >> > Let paste the code so we can move beyond the "batching" hypothesis: >> > >> > static int __remove_mapping(struct address_space *mapping, struct page >> > *page, >> > bool reclaimed, struct mem_cgroup *target_memcg) >> > { >> > unsigned long flags; >> > int refcount; >> > void *shadow = NULL; >> > >> > BUG_ON(!PageLocked(page)); >> > BUG_ON(mapping != page_mapping(page)); >> > >> > xa_lock_irqsave(&mapping->i_pages, flags); >> > >> >> SeongJae, what is this algorithm supposed to do when faced >> >> with situations like this: >> > >> > I'll assume the questions were directed at me, not SeongJae. >> > >> >> 1) Running on a system with 8 NUMA nodes, and >> >> memory >> >> pressure in one of those nodes. >> >> 2) Running PostgresQL or Oracle, with hundreds of >> >> processes mapping the same (very large) shared >> >> memory segment. >> >> >> >> How do you keep your algorithm from falling into the worst >> >> case virtual scanning scenarios that were crippling the >> >> 2.4 kernel 15+ years ago on systems with just a few GB of >> >> memory? >> > >> > There is a fundamental shift: that time we were scanning for cold pages, >> > and nowadays we are scanning for hot pages. >> > >> > I'd be surprised if scanning for cold pages didn't fall apart, because it'd >> > find most of the entries accessed, if they are present at all. >> > >> > Scanning for hot pages, on the other hand, is way better. Let me just >> > reiterate: >> > 1) It will not scan page tables from processes that have been sleeping >> > since the last scan. >> > 2) It will not scan PTE tables under non-leaf PMD entries that do not >> > have the accessed bit set, when >> > CONFIG_HAVE_ARCH_PARENT_PMD_YOUNG=y. >> > 3) It will not zigzag between the PGD table and the same PMD or PTE >> > table spanning multiple VMAs. In other words, it finishes all the >> > VMAs with the range of the same PMD or PTE table before it returns >> > to the PGD table. This optimizes workloads that have large numbers >> > of tiny VMAs, especially when CONFIG_PGTABLE_LEVELS=5. >> > >> > So the cost is roughly proportional to the number of referenced pages it >> > discovers. If there is no memory pressure, no scanning at all. For a system >> > under heavy memory pressure, most of the pages are referenced (otherwise >> > why would it be under memory pressure?), and if we use the rmap, we need to >> > scan a lot of pages anyway. Why not just scan them all? >> >> This may be not the case. For rmap scanning, it's possible to scan only >> a small portion of memory. But with the page table scanning, you need >> to scan almost all (I understand you have some optimization as above). > > Hi Ying, > > Let's take a step back. > > For the sake of discussion, when does the scanning have to happen? Can > we agree that the simplest answer is when we have evicted all inactive > pages? > > If so, my next question is who's filled in the memory space previously > occupied by those inactive pages? Newly faulted in pages, right? They > have the accessed bit set, and we can't evict them without scanning > them first, would you agree? > > And there are also existing active pages, and they were protected from > eviction. But now we need to deactivate some of them. Do you think > whether they'd have been used or not since the last scan? (Remember > they were active.) > > You mentioned "a small portion" and "almost all". How do you interpret > them in terms of these steps? > > Intuitively, "a small portion" and "almost all" seem right. But our > observations from *millions* of machines say the ratio of > pgscan_kswapd to pgsteal_kswapd is well over 7 when anon percentage is >>90%. Unlikely streaming files, it doesn't make sense to "stream" anon > memory. What I said is that it is "POSSIBLE" to scan only a small portion of memory. Whether and in which cases to do that depends on the policy we choose. I didn't say we have chosen to do that for all cases. >> As Rik shown in the test case above, there may be memory pressure on >> only one of 8 NUMA nodes (because of NUMA binding?). Then ramp scanning >> only needs to scan pages in this node, while the page table scanning may >> need to scan pages in other nodes too. > > Yes, and this is on my to-do list in the patchset: > > To-do List > ========== > KVM Optimization > ---------------- > Support shadow page table scanning. > > NUMA Optimization > ----------------- > Support NUMA policies and per-node RSS counters. > > We only can move forward one step at a time. Fair? You don't need to implement that now definitely. But we can discuss the possible solution now. Note that it's possible that only some processes are bound to some NUMA nodes, while other processes aren't bound. Best Regards, Huang, Ying From mboxrd@z Thu Jan 1 00:00:00 1970 Content-Type: multipart/mixed; boundary="===============3013686095564550690==" MIME-Version: 1.0 From: Huang, Ying To: lkp@lists.01.org Subject: Re: [PATCH v2 00/16] Multigenerational LRU Framework Date: Wed, 14 Apr 2021 16:27:22 +0800 Message-ID: <87lf9lqnit.fsf@yhuang6-desk1.ccr.corp.intel.com> In-Reply-To: List-Id: --===============3013686095564550690== Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: quoted-printable Yu Zhao writes: > On Wed, Apr 14, 2021 at 12:15 AM Huang, Ying wro= te: >> >> Yu Zhao writes: >> >> > On Tue, Apr 13, 2021 at 8:30 PM Rik van Riel wrot= e: >> >> >> >> On Wed, 2021-04-14 at 09:14 +1000, Dave Chinner wrote: >> >> > On Tue, Apr 13, 2021 at 10:13:24AM -0600, Jens Axboe wrote: >> >> > >> >> > > The initial posting of this patchset did no better, in fact it did >> >> > > a bit >> >> > > worse. Performance dropped to the same levels and kswapd was using >> >> > > as >> >> > > much CPU as before, but on top of that we also got excessive >> >> > > swapping. >> >> > > Not at a high rate, but 5-10MB/sec continually. >> >> > > >> >> > > I had some back and forths with Yu Zhao and tested a few new >> >> > > revisions, >> >> > > and the current series does much better in this regard. Performan= ce >> >> > > still dips a bit when page cache fills, but not nearly as much, a= nd >> >> > > kswapd is using less CPU than before. >> >> > >> >> > Profiles would be interesting, because it sounds to me like reclaim >> >> > *might* be batching page cache removal better (e.g. fewer, larger >> >> > batches) and so spending less time contending on the mapping tree >> >> > lock... >> >> > >> >> > IOWs, I suspect this result might actually be a result of less lock >> >> > contention due to a change in batch processing characteristics of >> >> > the new algorithm rather than it being a "better" algorithm... >> >> >> >> That seems quite likely to me, given the issues we have >> >> had with virtual scan reclaim algorithms in the past. >> > >> > Hi Rik, >> > >> > Let paste the code so we can move beyond the "batching" hypothesis: >> > >> > static int __remove_mapping(struct address_space *mapping, struct page >> > *page, >> > bool reclaimed, struct mem_cgroup *target_= memcg) >> > { >> > unsigned long flags; >> > int refcount; >> > void *shadow =3D NULL; >> > >> > BUG_ON(!PageLocked(page)); >> > BUG_ON(mapping !=3D page_mapping(page)); >> > >> > xa_lock_irqsave(&mapping->i_pages, flags); >> > >> >> SeongJae, what is this algorithm supposed to do when faced >> >> with situations like this: >> > >> > I'll assume the questions were directed at me, not SeongJae. >> > >> >> 1) Running on a system with 8 NUMA nodes, and >> >> memory >> >> pressure in one of those nodes. >> >> 2) Running PostgresQL or Oracle, with hundreds of >> >> processes mapping the same (very large) shared >> >> memory segment. >> >> >> >> How do you keep your algorithm from falling into the worst >> >> case virtual scanning scenarios that were crippling the >> >> 2.4 kernel 15+ years ago on systems with just a few GB of >> >> memory? >> > >> > There is a fundamental shift: that time we were scanning for cold page= s, >> > and nowadays we are scanning for hot pages. >> > >> > I'd be surprised if scanning for cold pages didn't fall apart, because= it'd >> > find most of the entries accessed, if they are present at all. >> > >> > Scanning for hot pages, on the other hand, is way better. Let me just >> > reiterate: >> > 1) It will not scan page tables from processes that have been sleeping >> > since the last scan. >> > 2) It will not scan PTE tables under non-leaf PMD entries that do not >> > have the accessed bit set, when >> > CONFIG_HAVE_ARCH_PARENT_PMD_YOUNG=3Dy. >> > 3) It will not zigzag between the PGD table and the same PMD or PTE >> > table spanning multiple VMAs. In other words, it finishes all the >> > VMAs with the range of the same PMD or PTE table before it returns >> > to the PGD table. This optimizes workloads that have large numbers >> > of tiny VMAs, especially when CONFIG_PGTABLE_LEVELS=3D5. >> > >> > So the cost is roughly proportional to the number of referenced pages = it >> > discovers. If there is no memory pressure, no scanning at all. For a s= ystem >> > under heavy memory pressure, most of the pages are referenced (otherwi= se >> > why would it be under memory pressure?), and if we use the rmap, we ne= ed to >> > scan a lot of pages anyway. Why not just scan them all? >> >> This may be not the case. For rmap scanning, it's possible to scan only >> a small portion of memory. But with the page table scanning, you need >> to scan almost all (I understand you have some optimization as above). > > Hi Ying, > > Let's take a step back. > > For the sake of discussion, when does the scanning have to happen? Can > we agree that the simplest answer is when we have evicted all inactive > pages? > > If so, my next question is who's filled in the memory space previously > occupied by those inactive pages? Newly faulted in pages, right? They > have the accessed bit set, and we can't evict them without scanning > them first, would you agree? > > And there are also existing active pages, and they were protected from > eviction. But now we need to deactivate some of them. Do you think > whether they'd have been used or not since the last scan? (Remember > they were active.) > > You mentioned "a small portion" and "almost all". How do you interpret > them in terms of these steps? > > Intuitively, "a small portion" and "almost all" seem right. But our > observations from *millions* of machines say the ratio of > pgscan_kswapd to pgsteal_kswapd is well over 7 when anon percentage is >>90%. Unlikely streaming files, it doesn't make sense to "stream" anon > memory. What I said is that it is "POSSIBLE" to scan only a small portion of memory. Whether and in which cases to do that depends on the policy we choose. I didn't say we have chosen to do that for all cases. >> As Rik shown in the test case above, there may be memory pressure on >> only one of 8 NUMA nodes (because of NUMA binding?). Then ramp scanning >> only needs to scan pages in this node, while the page table scanning may >> need to scan pages in other nodes too. > > Yes, and this is on my to-do list in the patchset: > > To-do List > =3D=3D=3D=3D=3D=3D=3D=3D=3D=3D > KVM Optimization > ---------------- > Support shadow page table scanning. > > NUMA Optimization > ----------------- > Support NUMA policies and per-node RSS counters. > > We only can move forward one step at a time. Fair? You don't need to implement that now definitely. But we can discuss the possible solution now. Note that it's possible that only some processes are bound to some NUMA nodes, while other processes aren't bound. Best Regards, Huang, Ying --===============3013686095564550690==--