linux-kernel.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
From: David Rientjes <rientjes@google.com>
To: Christoph Hellwig <hch@lst.de>, Tom Lendacky <thomas.lendacky@amd.com>
Cc: "Singh, Brijesh" <brijesh.singh@amd.com>,
	"Grimm, Jon" <jon.grimm@amd.com>, Joerg Roedel <joro@8bytes.org>,
	baekhw@google.com,
	"linux-kernel@vger.kernel.org" <linux-kernel@vger.kernel.org>,
	"iommu@lists.linux-foundation.org"
	<iommu@lists.linux-foundation.org>
Subject: [rfc 4/6] dma-remap: dynamically expanding atomic pools
Date: Sun, 1 Mar 2020 16:05:18 -0800 (PST)	[thread overview]
Message-ID: <alpine.DEB.2.21.2003011537440.213582@chino.kir.corp.google.com> (raw)
In-Reply-To: <alpine.DEB.2.21.2003011535510.213582@chino.kir.corp.google.com>

When an atomic pool becomes fully depleted because it is now relied upon
for all non-blocking allocations through the DMA API, allow background
expansion of each pool by a kworker.

When an atomic pool has less than the default size of memory left, kick
off a kworker to dynamically expand the pool in the background.  The pool
is doubled in size.

This allows the default size to be kept quite low when one or more of the
atomic pools is not used.

Also switch over some node ids to the more appropriate NUMA_NO_NODE.

Signed-off-by: David Rientjes <rientjes@google.com>
---
 kernel/dma/remap.c | 79 ++++++++++++++++++++++++++++++++++------------
 1 file changed, 58 insertions(+), 21 deletions(-)

diff --git a/kernel/dma/remap.c b/kernel/dma/remap.c
--- a/kernel/dma/remap.c
+++ b/kernel/dma/remap.c
@@ -10,6 +10,7 @@
 #include <linux/genalloc.h>
 #include <linux/slab.h>
 #include <linux/vmalloc.h>
+#include <linux/workqueue.h>
 
 struct page **dma_common_find_pages(void *cpu_addr)
 {
@@ -104,7 +105,10 @@ static struct gen_pool *atomic_pool_dma32 __ro_after_init;
 static struct gen_pool *atomic_pool_normal __ro_after_init;
 
 #define DEFAULT_DMA_COHERENT_POOL_SIZE  SZ_256K
-static size_t atomic_pool_size __initdata = DEFAULT_DMA_COHERENT_POOL_SIZE;
+static size_t atomic_pool_size = DEFAULT_DMA_COHERENT_POOL_SIZE;
+
+/* Dynamic background expansion when the atomic pool is near capacity */
+struct work_struct atomic_pool_work;
 
 static int __init early_coherent_pool(char *p)
 {
@@ -113,14 +117,14 @@ static int __init early_coherent_pool(char *p)
 }
 early_param("coherent_pool", early_coherent_pool);
 
-static int __init __dma_atomic_pool_init(struct gen_pool **pool,
-					 size_t pool_size, gfp_t gfp)
+static int atomic_pool_expand(struct gen_pool *pool, size_t pool_size,
+			      gfp_t gfp)
 {
-	const unsigned int order = get_order(pool_size);
 	const unsigned long nr_pages = pool_size >> PAGE_SHIFT;
+	const unsigned int order = get_order(pool_size);
 	struct page *page;
 	void *addr;
-	int ret;
+	int ret = -ENOMEM;
 
 	if (dev_get_cma_area(NULL))
 		page = dma_alloc_from_contiguous(NULL, nr_pages, order, false);
@@ -131,38 +135,67 @@ static int __init __dma_atomic_pool_init(struct gen_pool **pool,
 
 	arch_dma_prep_coherent(page, pool_size);
 
-	*pool = gen_pool_create(PAGE_SHIFT, -1);
-	if (!*pool)
-		goto free_page;
-
 	addr = dma_common_contiguous_remap(page, pool_size,
 					   pgprot_dmacoherent(PAGE_KERNEL),
 					   __builtin_return_address(0));
 	if (!addr)
-		goto destroy_genpool;
+		goto free_page;
 
-	ret = gen_pool_add_virt(*pool, (unsigned long)addr, page_to_phys(page),
-				pool_size, -1);
+	ret = gen_pool_add_virt(pool, (unsigned long)addr, page_to_phys(page),
+				pool_size, NUMA_NO_NODE);
 	if (ret)
 		goto remove_mapping;
-	gen_pool_set_algo(*pool, gen_pool_first_fit_order_align, NULL);
 
-	pr_info("DMA: preallocated %zu KiB %pGg pool for atomic allocations\n",
-		pool_size >> 10, &gfp);
 	return 0;
 
 remove_mapping:
 	dma_common_free_remap(addr, pool_size);
-destroy_genpool:
-	gen_pool_destroy(*pool);
-	*pool = NULL;
 free_page:
 	if (!dma_release_from_contiguous(NULL, page, nr_pages))
 		__free_pages(page, order);
 out:
-	pr_err("DMA: failed to allocate %zu KiB %pGg pool for atomic allocation\n",
-		atomic_pool_size >> 10, &gfp);
-	return -ENOMEM;
+	return ret;
+}
+
+static void atomic_pool_resize(struct gen_pool *pool, gfp_t gfp)
+{
+	if (pool && gen_pool_avail(pool) < atomic_pool_size)
+		atomic_pool_expand(pool, gen_pool_size(pool), gfp);
+}
+
+static void atomic_pool_work_fn(struct work_struct *work)
+{
+	if (IS_ENABLED(CONFIG_ZONE_DMA))
+		atomic_pool_resize(atomic_pool, GFP_DMA);
+	if (IS_ENABLED(CONFIG_ZONE_DMA32))
+		atomic_pool_resize(atomic_pool_dma32, GFP_DMA32);
+	atomic_pool_resize(atomic_pool_normal, GFP_KERNEL);
+}
+
+static int __init __dma_atomic_pool_init(struct gen_pool **pool,
+					 size_t pool_size, gfp_t gfp)
+{
+	int ret;
+
+	*pool = gen_pool_create(PAGE_SHIFT, NUMA_NO_NODE);
+	if (!*pool)
+		return -ENOMEM;
+
+	gen_pool_set_algo(*pool, gen_pool_first_fit_order_align, NULL);
+
+	ret = atomic_pool_expand(*pool, pool_size, gfp);
+	if (ret) {
+		gen_pool_destroy(*pool);
+		*pool = NULL;
+		pr_err("DMA: failed to allocate %zu KiB %pGg pool for atomic allocation\n",
+		       atomic_pool_size >> 10, &gfp);
+		return ret;
+	}
+
+
+	pr_info("DMA: preallocated %zu KiB %pGg pool for atomic allocations\n",
+		pool_size >> 10, &gfp);
+	return 0;
 }
 
 static int __init dma_atomic_pool_init(void)
@@ -170,6 +203,8 @@ static int __init dma_atomic_pool_init(void)
 	int ret = 0;
 	int err;
 
+	INIT_WORK(&atomic_pool_work, atomic_pool_work_fn);
+
 	ret = __dma_atomic_pool_init(&atomic_pool_normal, atomic_pool_size,
 				     GFP_KERNEL);
 	if (IS_ENABLED(CONFIG_ZONE_DMA)) {
@@ -231,6 +266,8 @@ void *dma_alloc_from_pool(struct device *dev, size_t size,
 		ptr = (void *)val;
 		memset(ptr, 0, size);
 	}
+	if (gen_pool_avail(pool) < atomic_pool_size)
+		schedule_work(&atomic_pool_work);
 
 	return ptr;
 }

  parent reply	other threads:[~2020-03-02  0:05 UTC|newest]

Thread overview: 37+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2020-01-01  1:54 [rfc] dma-mapping: preallocate unencrypted DMA atomic pool David Rientjes
2020-01-06 17:34 ` Robin Murphy
2020-01-07 10:54   ` Christoph Hellwig
2020-01-07 19:57     ` David Rientjes
2020-01-09 14:31       ` Christoph Hellwig
2020-01-09 18:58         ` David Rientjes
2020-01-17 15:28 ` Tom Lendacky
2020-02-28  9:27   ` David Rientjes
2020-03-02  0:05     ` [rfc 0/6] unencrypted atomic DMA pools with dynamic expansion David Rientjes
2020-03-02  0:05       ` [rfc 1/6] dma-mapping: pass device to atomic allocation functions David Rientjes
2020-03-02  0:05       ` [rfc 2/6] dma-remap: add additional atomic pools to map to gfp mask David Rientjes
2020-03-05 15:42         ` Christoph Hellwig
2020-03-02  0:05       ` [rfc 3/6] dma-remap: wire up the atomic pools depending on " David Rientjes
2020-03-05 15:43         ` Christoph Hellwig
2020-03-02  0:05       ` David Rientjes [this message]
2020-03-03 22:29         ` [rfc 4/6] dma-remap: dynamically expanding atomic pools David Rientjes
2020-03-02  0:05       ` [rfc 5/6] dma-direct: atomic allocations must come from unencrypted pools David Rientjes
2020-03-05 15:44         ` Christoph Hellwig
2020-03-07  0:36           ` David Rientjes
2020-03-10 18:47             ` Christoph Hellwig
2020-03-02  0:05       ` [rfc 6/6] dma-remap: double the default DMA coherent pool size David Rientjes
2020-03-05 15:45         ` Christoph Hellwig
2020-04-08 21:21       ` [rfc v2 0/6] unencrypted atomic DMA pools with dynamic expansion David Rientjes
2020-04-08 21:21         ` [rfc v2 1/6] dma-remap: separate DMA atomic pools from direct remap code David Rientjes
2020-04-14  6:35           ` Christoph Hellwig
2020-04-08 21:21         ` [rfc v2 2/6] dma-pool: add additional coherent pools to map to gfp mask David Rientjes
2020-04-08 21:21         ` [rfc v2 3/6] dma-pool: dynamically expanding atomic pools David Rientjes
2020-04-08 21:21         ` [rfc v2 4/6] dma-direct: atomic allocations must come from atomic coherent pools David Rientjes
2020-04-09 21:24           ` Tom Lendacky
2020-04-14 19:18             ` David Rientjes
2020-04-14  6:43           ` Christoph Hellwig
2020-04-14 19:20             ` David Rientjes
2020-04-08 21:21         ` [rfc v2 5/6] x86/mm: unencrypted non-blocking DMA allocations use " David Rientjes
2020-04-08 21:21         ` [rfc v2 6/6] dma-pool: scale the default DMA coherent pool size with memory capacity David Rientjes
     [not found]         ` <20200410145520.17864-1-hdanton@sina.com>
2020-04-10 19:37           ` [rfc v2 3/6] dma-pool: dynamically expanding atomic pools David Rientjes
2020-04-14  6:44             ` Christoph Hellwig
2020-04-14 19:23               ` David Rientjes

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=alpine.DEB.2.21.2003011537440.213582@chino.kir.corp.google.com \
    --to=rientjes@google.com \
    --cc=baekhw@google.com \
    --cc=brijesh.singh@amd.com \
    --cc=hch@lst.de \
    --cc=iommu@lists.linux-foundation.org \
    --cc=jon.grimm@amd.com \
    --cc=joro@8bytes.org \
    --cc=linux-kernel@vger.kernel.org \
    --cc=thomas.lendacky@amd.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).