From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-13.2 required=3.0 tests=BAYES_00,DKIMWL_WL_MED, DKIM_SIGNED,DKIM_VALID,DKIM_VALID_AU,HEADER_FROM_DIFFERENT_DOMAINS, MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS,URIBL_BLOCKED,USER_IN_DEF_DKIM_WL autolearn=no autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id B3B3CC433ED for ; Thu, 8 Apr 2021 17:19:10 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id 86C3F60FEA for ; Thu, 8 Apr 2021 17:19:10 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S231480AbhDHRTT (ORCPT ); Thu, 8 Apr 2021 13:19:19 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:33348 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S232388AbhDHRTM (ORCPT ); Thu, 8 Apr 2021 13:19:12 -0400 Received: from mail-lf1-x133.google.com (mail-lf1-x133.google.com [IPv6:2a00:1450:4864:20::133]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 5BD7FC061761 for ; Thu, 8 Apr 2021 10:19:00 -0700 (PDT) Received: by mail-lf1-x133.google.com with SMTP id j18so5312955lfg.5 for ; Thu, 08 Apr 2021 10:19:00 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20161025; h=mime-version:references:in-reply-to:from:date:message-id:subject:to :cc; bh=0gEpNulq0JouzND7X8G/z4bZ8lZh7ERFU8FSsc0Dzac=; b=noL0zUmUFj2zwRUlYv3RquGNc8nwvDeqlE7uolCFVwehPyaRUjY3VAKKdGmRjzbZLl fD/PB2RNqDPIm2ezoQBU50070lVbfVsBOPRUfF+ZdyFQ+u6wHf9KMr7fBrLDaTEI1FbC XnC8zE6YqHXvJ/ZndiMiAa/1OhHSxtZ5S2Qa5UuUIrwGZtS96hvH34Wr1Io9GkW4/YgG zP/vAfPyeqU0PLkig/L+1rfZZmne+DWF9FAvwzjvjDLs6bwdvCIEtJyrdhIzts+iS2OA TCT04r2xZHuzfYbs0VIAhgUPfg2fj8fWboitjtol+M2nlDBeCyOsQessN0yhxxmJ/rQH Ix3A== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc; bh=0gEpNulq0JouzND7X8G/z4bZ8lZh7ERFU8FSsc0Dzac=; b=J6vjazsE+RdnJ6tag5XiOb5c5MQllzjitp9U7OWPcbkU/w1Sxn5dLNszXbjj7W8iRf sFZcUNcqjl7T+zZXb7WEZguuatCxY9+2eSBFNQV/j7ufnd0XHbv6EMZjMl4U1EVSoQvd /DWgDJckRSZHe+PMT0o3RMkevdw/Zl9mOSjAAcfMGPRWYwm7uDP9UrhCujg6Y27wrpKE 7/wT6Ew8zIEm2YnoaDF1NzxdaKkLFD/zUFOtH1riJCnouAyFGgdLUfBTIRit4oDeGZGG CZ1GX7YAL05IIU5x1FcgwSyXrt+IG6PwRRYSNVGKLE6iCHpMgvmi72gw7qUymQK4b57o UwNQ== X-Gm-Message-State: AOAM532Qu0XVYshjRgPgLPSxrT2pkyj+T1q5uhCXDxOHTUHSjCVP0ewY AesoJXotqgQ4Ri3LTOOk/Z/INN4gmpNCFZ5+gjhzYA== X-Google-Smtp-Source: ABdhPJxrQIVn50yQwVDP1KBGEaeW1VjAaTwKlqTcXbP8XbkcSZjoaC/A0UKnyxmh/feBPbCrnpCocgTLcYJ/w/IEW9M= X-Received: by 2002:a05:6512:3703:: with SMTP id z3mr7163066lfr.358.1617902338550; Thu, 08 Apr 2021 10:18:58 -0700 (PDT) MIME-Version: 1.0 References: In-Reply-To: From: Shakeel Butt Date: Thu, 8 Apr 2021 10:18:47 -0700 Message-ID: Subject: Re: [RFC PATCH v1 00/11] Manage the top tier memory in a tiered memory To: Tim Chen Cc: Michal Hocko , Johannes Weiner , Andrew Morton , Dave Hansen , Ying Huang , Dan Williams , David Rientjes , Linux MM , Cgroups , LKML Content-Type: text/plain; charset="UTF-8" Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Hi Tim, On Mon, Apr 5, 2021 at 11:08 AM Tim Chen wrote: > > Traditionally, all memory is DRAM. Some DRAM might be closer/faster than > others NUMA wise, but a byte of media has about the same cost whether it > is close or far. But, with new memory tiers such as Persistent Memory > (PMEM). there is a choice between fast/expensive DRAM and slow/cheap > PMEM. > > The fast/expensive memory lives in the top tier of the memory hierachy. > > Previously, the patchset > [PATCH 00/10] [v7] Migrate Pages in lieu of discard > https://lore.kernel.org/linux-mm/20210401183216.443C4443@viggo.jf.intel.com/ > provides a mechanism to demote cold pages from DRAM node into PMEM. > > And the patchset > [PATCH 0/6] [RFC v6] NUMA balancing: optimize memory placement for memory tiering system > https://lore.kernel.org/linux-mm/20210311081821.138467-1-ying.huang@intel.com/ > provides a mechanism to promote hot pages in PMEM to the DRAM node > leveraging autonuma. > > The two patchsets together keep the hot pages in DRAM and colder pages > in PMEM. Thanks for working on this as this is becoming more and more important particularly in the data centers where memory is a big portion of the cost. I see you have responded to Michal and I will add my more specific response there. Here I wanted to give my high level concern regarding using v1's soft limit like semantics for top tier memory. This patch series aims to distribute/partition top tier memory between jobs of different priorities. We want high priority jobs to have preferential access to the top tier memory and we don't want low priority jobs to hog the top tier memory. Using v1's soft limit like behavior can potentially cause high priority jobs to stall to make enough space on top tier memory on their allocation path and I think this patchset is aiming to reduce that impact by making kswapd do that work. However I think the more concerning issue is the low priority job hogging the top tier memory. The possible ways the low priority job can hog the top tier memory are by allocating non-movable memory or by mlocking the memory. (Oh there is also pinning the memory but I don't know if there is a user api to pin memory?) For the mlocked memory, you need to either modify the reclaim code or use a different mechanism for demoting cold memory. Basically I am saying we should put the upfront control (limit) on the usage of top tier memory by the jobs.