linux-kernel.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
* [PATCH 0/13] x86: unifying ways to handle multiple sets of dma mapping ops
@ 2009-01-05 14:47 FUJITA Tomonori
  2009-01-05 14:47 ` [PATCH 1/8] add map_page and unmap_page to struct dma_mapping_ops FUJITA Tomonori
  2009-01-05 15:00 ` [PATCH 0/13] x86: unifying ways to handle multiple sets of dma mapping ops FUJITA Tomonori
  0 siblings, 2 replies; 16+ messages in thread
From: FUJITA Tomonori @ 2009-01-05 14:47 UTC (permalink / raw)
  To: mingo; +Cc: linux-kernel, fujita.tomonori

This patchset is the second part of the unification of ways to handle
multiple sets of dma mapping API. The whole work consists of three
patchset. This is for X86 and can be applied independently (against
tip/master).

I've submitted the first part (for IA64):

http://marc.info/?l=linux-kernel&m=123116676006794&w=2

dma_mapping_ops (or dma_ops) struct is used to handle multiple sets of
dma mapping API by X86, SPARC, and POWER. IA64 also handle multiple
sets of dma mapping API but in a very different way (some define
magic).

X86 and IA64 share VT-d and SWIOTLB code. We need several workarounds
for it because of the deference of ways to handle multiple sets of dma
mapping API (e.g., X86 people can't freely change struct
dma_mapping_ops in x86's dma-mapping.h now because it could break
IA64).  Seems POWER will use SWIOTLB code soon. I think that it's time
to unify ways to handle multiple sets of dma mapping API. After
applying the whole work, we have struct dma_map_ops
include/linux/dma-mapping.h (I also dream of changing all the archs to
use SWIOTLB in order to remove the bounce code in the block and
network stacks...).

This patchset doesn't include major changes, just converting x86's
dma_mapping_ops to use map_page and unmap_page instead of map_single
and unmap_single. Currently, x86's dma_mapping_ops uses physical
address as a map_single's argument. But it's confusing since
dma_map_single uses a virtual address argument. So I chose POWER
dma_mapping_ops scheme, which uses map_page to handle dma_map_single.

=
 arch/x86/include/asm/dma-mapping.h |   23 ++++++++++++++---------
 arch/x86/kernel/amd_iommu.c        |   16 ++++++++++------
 arch/x86/kernel/pci-calgary_64.c   |   23 +++++++++++++----------
 arch/x86/kernel/pci-gart_64.c      |   20 ++++++++++++--------
 arch/x86/kernel/pci-nommu.c        |   16 ++++++++--------
 arch/x86/kernel/pci-swiotlb_64.c   |   22 ++++++++++++++++------
 drivers/pci/intel-iommu.c          |   26 ++++++++++++++++++++++----
 7 files changed, 95 insertions(+), 51 deletions(-)




^ permalink raw reply	[flat|nested] 16+ messages in thread

* [PATCH 1/8] add map_page and unmap_page to struct dma_mapping_ops
  2009-01-05 14:47 [PATCH 0/13] x86: unifying ways to handle multiple sets of dma mapping ops FUJITA Tomonori
@ 2009-01-05 14:47 ` FUJITA Tomonori
  2009-01-05 14:47   ` [PATCH 2/8] swiotlb: add map_page and unmap_page FUJITA Tomonori
  2009-01-05 17:26   ` [PATCH 1/8] add map_page and unmap_page to struct dma_mapping_ops Joerg Roedel
  2009-01-05 15:00 ` [PATCH 0/13] x86: unifying ways to handle multiple sets of dma mapping ops FUJITA Tomonori
  1 sibling, 2 replies; 16+ messages in thread
From: FUJITA Tomonori @ 2009-01-05 14:47 UTC (permalink / raw)
  To: mingo; +Cc: linux-kernel, FUJITA Tomonori

This patch adds map_page and unmap_page to struct dma_mapping_ops.

This is a preparation of struct dma_mapping_ops unification. We use
map_page and unmap_page instead of map_single and unmap_single.

We will remove map_single and unmap_single hooks in the last patch in
this patchset.

Signed-off-by: FUJITA Tomonori <fujita.tomonori@lab.ntt.co.jp>
---
 arch/x86/include/asm/dma-mapping.h |    8 ++++++++
 1 files changed, 8 insertions(+), 0 deletions(-)

diff --git a/arch/x86/include/asm/dma-mapping.h b/arch/x86/include/asm/dma-mapping.h
index e93265c..2f89d2e 100644
--- a/arch/x86/include/asm/dma-mapping.h
+++ b/arch/x86/include/asm/dma-mapping.h
@@ -8,6 +8,7 @@
 
 #include <linux/kmemcheck.h>
 #include <linux/scatterlist.h>
+#include <linux/dma-attrs.h>
 #include <asm/io.h>
 #include <asm/swiotlb.h>
 #include <asm-generic/dma-coherent.h>
@@ -51,6 +52,13 @@ struct dma_mapping_ops {
 	void            (*unmap_sg)(struct device *hwdev,
 				struct scatterlist *sg, int nents,
 				int direction);
+	dma_addr_t	(*map_page)(struct device *dev, struct page *page,
+				    unsigned long offset, size_t size,
+				    enum dma_data_direction dir,
+				    struct dma_attrs *attrs);
+	void 		(*unmap_page)(struct device *dev, dma_addr_t dma_handle,
+				      size_t size, enum dma_data_direction dir,
+				      struct dma_attrs *attrs);
 	int             (*dma_supported)(struct device *hwdev, u64 mask);
 	int		is_phys;
 };
-- 
1.6.0.6


^ permalink raw reply related	[flat|nested] 16+ messages in thread

* [PATCH 2/8] swiotlb: add map_page and unmap_page
  2009-01-05 14:47 ` [PATCH 1/8] add map_page and unmap_page to struct dma_mapping_ops FUJITA Tomonori
@ 2009-01-05 14:47   ` FUJITA Tomonori
  2009-01-05 14:47     ` [PATCH 3/8] gart: " FUJITA Tomonori
  2009-01-05 17:26   ` [PATCH 1/8] add map_page and unmap_page to struct dma_mapping_ops Joerg Roedel
  1 sibling, 1 reply; 16+ messages in thread
From: FUJITA Tomonori @ 2009-01-05 14:47 UTC (permalink / raw)
  To: mingo; +Cc: linux-kernel, FUJITA Tomonori

This is a preparation of struct dma_mapping_ops unification. We use
map_page and unmap_page instead of map_single and unmap_single.

This is sorta temporary workaround. We will move them to lib/swiotlb.c
to enable x86's swiotlb code to directly use them.

We will remove map_single and unmap_single hooks in the last patch in
this patchset.

Signed-off-by: FUJITA Tomonori <fujita.tomonori@lab.ntt.co.jp>
---
 arch/x86/kernel/pci-swiotlb_64.c |   19 +++++++++++++++++++
 1 files changed, 19 insertions(+), 0 deletions(-)

diff --git a/arch/x86/kernel/pci-swiotlb_64.c b/arch/x86/kernel/pci-swiotlb_64.c
index d59c917..d1c0366 100644
--- a/arch/x86/kernel/pci-swiotlb_64.c
+++ b/arch/x86/kernel/pci-swiotlb_64.c
@@ -45,6 +45,23 @@ swiotlb_map_single_phys(struct device *hwdev, phys_addr_t paddr, size_t size,
 	return swiotlb_map_single(hwdev, phys_to_virt(paddr), size, direction);
 }
 
+/* these will be moved to lib/swiotlb.c later on */
+
+static dma_addr_t swiotlb_map_page(struct device *dev, struct page *page,
+				   unsigned long offset, size_t size,
+				   enum dma_data_direction dir,
+				   struct dma_attrs *attrs)
+{
+	return swiotlb_map_single(dev, page_address(page) + offset, size, dir);
+}
+
+static void swiotlb_unmap_page(struct device *dev, dma_addr_t dma_handle,
+			       size_t size, enum dma_data_direction dir,
+			       struct dma_attrs *attrs)
+{
+	swiotlb_unmap_single(dev, dma_handle, size, dir);
+}
+
 static void *x86_swiotlb_alloc_coherent(struct device *hwdev, size_t size,
 					dma_addr_t *dma_handle, gfp_t flags)
 {
@@ -71,6 +88,8 @@ struct dma_mapping_ops swiotlb_dma_ops = {
 	.sync_sg_for_device = swiotlb_sync_sg_for_device,
 	.map_sg = swiotlb_map_sg,
 	.unmap_sg = swiotlb_unmap_sg,
+	.map_page = swiotlb_map_page,
+	.unmap_page = swiotlb_unmap_page,
 	.dma_supported = NULL,
 };
 
-- 
1.6.0.6


^ permalink raw reply related	[flat|nested] 16+ messages in thread

* [PATCH 3/8] gart: add map_page and unmap_page
  2009-01-05 14:47   ` [PATCH 2/8] swiotlb: add map_page and unmap_page FUJITA Tomonori
@ 2009-01-05 14:47     ` FUJITA Tomonori
  2009-01-05 14:47       ` [PATCH 4/8] calgary: " FUJITA Tomonori
  0 siblings, 1 reply; 16+ messages in thread
From: FUJITA Tomonori @ 2009-01-05 14:47 UTC (permalink / raw)
  To: mingo; +Cc: linux-kernel, FUJITA Tomonori

This is a preparation of struct dma_mapping_ops unification. We use
map_page and unmap_page instead of map_single and unmap_single.

We will remove map_single and unmap_single hooks in the last patch in
this patchset.

Signed-off-by: FUJITA Tomonori <fujita.tomonori@lab.ntt.co.jp>
---
 arch/x86/kernel/pci-gart_64.c |   27 +++++++++++++++++++++++----
 1 files changed, 23 insertions(+), 4 deletions(-)

diff --git a/arch/x86/kernel/pci-gart_64.c b/arch/x86/kernel/pci-gart_64.c
index 00c2bcd..e49c6dd 100644
--- a/arch/x86/kernel/pci-gart_64.c
+++ b/arch/x86/kernel/pci-gart_64.c
@@ -255,10 +255,13 @@ static dma_addr_t dma_map_area(struct device *dev, dma_addr_t phys_mem,
 }
 
 /* Map a single area into the IOMMU */
-static dma_addr_t
-gart_map_single(struct device *dev, phys_addr_t paddr, size_t size, int dir)
+static dma_addr_t gart_map_page(struct device *dev, struct page *page,
+				unsigned long offset, size_t size,
+				enum dma_data_direction dir,
+				struct dma_attrs *attrs)
 {
 	unsigned long bus;
+	phys_addr_t paddr = page_to_phys(page) + offset;
 
 	if (!dev)
 		dev = &x86_dma_fallback_dev;
@@ -272,11 +275,19 @@ gart_map_single(struct device *dev, phys_addr_t paddr, size_t size, int dir)
 	return bus;
 }
 
+static dma_addr_t gart_map_single(struct device *dev, phys_addr_t paddr,
+				  size_t size, int dir)
+{
+	return gart_map_page(dev, pfn_to_page(paddr >> PAGE_SHIFT),
+			     paddr & ~PAGE_MASK, size, dir, NULL);
+}
+
 /*
  * Free a DMA mapping.
  */
-static void gart_unmap_single(struct device *dev, dma_addr_t dma_addr,
-			      size_t size, int direction)
+static void gart_unmap_page(struct device *dev, dma_addr_t dma_addr,
+			    size_t size, enum dma_data_direction dir,
+			    struct dma_attrs *attrs)
 {
 	unsigned long iommu_page;
 	int npages;
@@ -295,6 +306,12 @@ static void gart_unmap_single(struct device *dev, dma_addr_t dma_addr,
 	free_iommu(iommu_page, npages);
 }
 
+static void gart_unmap_single(struct device *dev, dma_addr_t dma_addr,
+			      size_t size, int direction)
+{
+	gart_unmap_page(dev, dma_addr, size, direction, NULL);
+}
+
 /*
  * Wrapper for pci_unmap_single working with scatterlists.
  */
@@ -712,6 +729,8 @@ static struct dma_mapping_ops gart_dma_ops = {
 	.unmap_single			= gart_unmap_single,
 	.map_sg				= gart_map_sg,
 	.unmap_sg			= gart_unmap_sg,
+	.map_page			= gart_map_page,
+	.unmap_page			= gart_unmap_page,
 	.alloc_coherent			= gart_alloc_coherent,
 	.free_coherent			= gart_free_coherent,
 };
-- 
1.6.0.6


^ permalink raw reply related	[flat|nested] 16+ messages in thread

* [PATCH 4/8] calgary: add map_page and unmap_page
  2009-01-05 14:47     ` [PATCH 3/8] gart: " FUJITA Tomonori
@ 2009-01-05 14:47       ` FUJITA Tomonori
  2009-01-05 14:47         ` [PATCH 5/8] AMD IOMMU: " FUJITA Tomonori
  2009-01-06  8:32         ` [PATCH 4/8] calgary: add map_page and unmap_page Muli Ben-Yehuda
  0 siblings, 2 replies; 16+ messages in thread
From: FUJITA Tomonori @ 2009-01-05 14:47 UTC (permalink / raw)
  To: mingo; +Cc: linux-kernel, FUJITA Tomonori, Muli Ben-Yehuda

This is a preparation of struct dma_mapping_ops unification. We use
map_page and unmap_page instead of map_single and unmap_single.

We will remove map_single and unmap_single hooks in the last patch in
this patchset.

Signed-off-by: FUJITA Tomonori <fujita.tomonori@lab.ntt.co.jp>
Cc: Muli Ben-Yehuda <muli@il.ibm.com>
---
 arch/x86/kernel/pci-calgary_64.c |   35 +++++++++++++++++++++++++++--------
 1 files changed, 27 insertions(+), 8 deletions(-)

diff --git a/arch/x86/kernel/pci-calgary_64.c b/arch/x86/kernel/pci-calgary_64.c
index d28bbdc..e33cfcf 100644
--- a/arch/x86/kernel/pci-calgary_64.c
+++ b/arch/x86/kernel/pci-calgary_64.c
@@ -445,10 +445,12 @@ error:
 	return 0;
 }
 
-static dma_addr_t calgary_map_single(struct device *dev, phys_addr_t paddr,
-	size_t size, int direction)
+static dma_addr_t calgary_map_page(struct device *dev, struct page *page,
+				   unsigned long offset, size_t size,
+				   enum dma_data_direction dir,
+				   struct dma_attrs *attrs)
 {
-	void *vaddr = phys_to_virt(paddr);
+	void *vaddr = page_address(page) + offset;
 	unsigned long uaddr;
 	unsigned int npages;
 	struct iommu_table *tbl = find_iommu_table(dev);
@@ -456,17 +458,32 @@ static dma_addr_t calgary_map_single(struct device *dev, phys_addr_t paddr,
 	uaddr = (unsigned long)vaddr;
 	npages = iommu_num_pages(uaddr, size, PAGE_SIZE);
 
-	return iommu_alloc(dev, tbl, vaddr, npages, direction);
+	return iommu_alloc(dev, tbl, vaddr, npages, dir);
 }
 
-static void calgary_unmap_single(struct device *dev, dma_addr_t dma_handle,
-	size_t size, int direction)
+static dma_addr_t calgary_map_single(struct device *dev, phys_addr_t paddr,
+				     size_t size, int direction)
+{
+	return calgary_map_page(dev, pfn_to_page(paddr >> PAGE_SHIFT),
+				paddr & ~PAGE_MASK, size,
+				direction, NULL);
+}
+
+static void calgary_unmap_page(struct device *dev, dma_addr_t dma_addr,
+			       size_t size, enum dma_data_direction dir,
+			       struct dma_attrs *attrs)
 {
 	struct iommu_table *tbl = find_iommu_table(dev);
 	unsigned int npages;
 
-	npages = iommu_num_pages(dma_handle, size, PAGE_SIZE);
-	iommu_free(tbl, dma_handle, npages);
+	npages = iommu_num_pages(dma_addr, size, PAGE_SIZE);
+	iommu_free(tbl, dma_addr, npages);
+}
+
+static void calgary_unmap_single(struct device *dev, dma_addr_t dma_handle,
+				 size_t size, int direction)
+{
+	calgary_unmap_page(dev, dma_handle, size, direction, NULL);
 }
 
 static void* calgary_alloc_coherent(struct device *dev, size_t size,
@@ -522,6 +539,8 @@ static struct dma_mapping_ops calgary_dma_ops = {
 	.unmap_single = calgary_unmap_single,
 	.map_sg = calgary_map_sg,
 	.unmap_sg = calgary_unmap_sg,
+	.map_page = calgary_map_page,
+	.unmap_page = calgary_unmap_page,
 };
 
 static inline void __iomem * busno_to_bbar(unsigned char num)
-- 
1.6.0.6


^ permalink raw reply related	[flat|nested] 16+ messages in thread

* [PATCH 5/8] AMD IOMMU: add map_page and unmap_page
  2009-01-05 14:47       ` [PATCH 4/8] calgary: " FUJITA Tomonori
@ 2009-01-05 14:47         ` FUJITA Tomonori
  2009-01-05 14:47           ` [PATCH 6/8] intel-iommu: " FUJITA Tomonori
  2009-01-06  8:32         ` [PATCH 4/8] calgary: add map_page and unmap_page Muli Ben-Yehuda
  1 sibling, 1 reply; 16+ messages in thread
From: FUJITA Tomonori @ 2009-01-05 14:47 UTC (permalink / raw)
  To: mingo; +Cc: linux-kernel, FUJITA Tomonori, Joerg Roedel

This is a preparation of struct dma_mapping_ops unification. We use
map_page and unmap_page instead of map_single and unmap_single.

We will remove map_single and unmap_single hooks in the last patch in
this patchset.

Signed-off-by: FUJITA Tomonori <fujita.tomonori@lab.ntt.co.jp>
Cc: Joerg Roedel <joerg.roedel@amd.com>
---
 arch/x86/kernel/amd_iommu.c |   27 +++++++++++++++++++++++----
 1 files changed, 23 insertions(+), 4 deletions(-)

diff --git a/arch/x86/kernel/amd_iommu.c b/arch/x86/kernel/amd_iommu.c
index 5113c08..8570441 100644
--- a/arch/x86/kernel/amd_iommu.c
+++ b/arch/x86/kernel/amd_iommu.c
@@ -22,6 +22,7 @@
 #include <linux/bitops.h>
 #include <linux/debugfs.h>
 #include <linux/scatterlist.h>
+#include <linux/dma-mapping.h>
 #include <linux/iommu-helper.h>
 #ifdef CONFIG_IOMMU_API
 #include <linux/iommu.h>
@@ -1297,8 +1298,10 @@ static void __unmap_single(struct amd_iommu *iommu,
 /*
  * The exported map_single function for dma_ops.
  */
-static dma_addr_t map_single(struct device *dev, phys_addr_t paddr,
-			     size_t size, int dir)
+static dma_addr_t map_page(struct device *dev, struct page *page,
+			   unsigned long offset, size_t size,
+			   enum dma_data_direction dir,
+			   struct dma_attrs *attrs)
 {
 	unsigned long flags;
 	struct amd_iommu *iommu;
@@ -1306,6 +1309,7 @@ static dma_addr_t map_single(struct device *dev, phys_addr_t paddr,
 	u16 devid;
 	dma_addr_t addr;
 	u64 dma_mask;
+	phys_addr_t paddr = page_to_phys(page) + offset;
 
 	INC_STATS_COUNTER(cnt_map_single);
 
@@ -1337,11 +1341,18 @@ out:
 	return addr;
 }
 
+static dma_addr_t map_single(struct device *dev, phys_addr_t paddr,
+			     size_t size, int dir)
+{
+	return map_page(dev, pfn_to_page(paddr >> PAGE_SHIFT),
+			paddr & ~PAGE_MASK, size, dir, NULL);
+}
+
 /*
  * The exported unmap_single function for dma_ops.
  */
-static void unmap_single(struct device *dev, dma_addr_t dma_addr,
-			 size_t size, int dir)
+static void unmap_page(struct device *dev, dma_addr_t dma_addr, size_t size,
+		       enum dma_data_direction dir, struct dma_attrs *attrs)
 {
 	unsigned long flags;
 	struct amd_iommu *iommu;
@@ -1367,6 +1378,12 @@ static void unmap_single(struct device *dev, dma_addr_t dma_addr,
 	spin_unlock_irqrestore(&domain->lock, flags);
 }
 
+static void unmap_single(struct device *dev, dma_addr_t dma_addr,
+			 size_t size, int dir)
+{
+	return unmap_page(dev, dma_addr, size, dir, NULL);
+}
+
 /*
  * This is a special map_sg function which is used if we should map a
  * device which is not handled by an AMD IOMMU in the system.
@@ -1649,6 +1666,8 @@ static struct dma_mapping_ops amd_iommu_dma_ops = {
 	.free_coherent = free_coherent,
 	.map_single = map_single,
 	.unmap_single = unmap_single,
+	.map_page = map_page,
+	.unmap_page = unmap_page,
 	.map_sg = map_sg,
 	.unmap_sg = unmap_sg,
 	.dma_supported = amd_iommu_dma_supported,
-- 
1.6.0.6


^ permalink raw reply related	[flat|nested] 16+ messages in thread

* [PATCH 6/8] intel-iommu: add map_page and unmap_page
  2009-01-05 14:47         ` [PATCH 5/8] AMD IOMMU: " FUJITA Tomonori
@ 2009-01-05 14:47           ` FUJITA Tomonori
  2009-01-05 14:47             ` [PATCH 7/8] pci-nommu: add map_page FUJITA Tomonori
  0 siblings, 1 reply; 16+ messages in thread
From: FUJITA Tomonori @ 2009-01-05 14:47 UTC (permalink / raw)
  To: mingo; +Cc: linux-kernel, FUJITA Tomonori, David Woodhouse

This is a preparation of struct dma_mapping_ops unification. We use
map_page and unmap_page instead of map_single and unmap_single.

This uses a temporary workaround, ifdef X86_64 to avoid IA64
build. The workaround will be removed after the unification. Well,
changing x86's struct dma_mapping_ops could break IA64. It's just
wrong. It's one of problems that this patchset fixes.

We will remove map_single and unmap_single hooks in the last patch in
this patchset.

Signed-off-by: FUJITA Tomonori <fujita.tomonori@lab.ntt.co.jp>
Cc: David Woodhouse <dwmw2@infradead.org>
---
 drivers/pci/intel-iommu.c |   24 ++++++++++++++++++++++--
 1 files changed, 22 insertions(+), 2 deletions(-)

diff --git a/drivers/pci/intel-iommu.c b/drivers/pci/intel-iommu.c
index 235fb7a..60258ec 100644
--- a/drivers/pci/intel-iommu.c
+++ b/drivers/pci/intel-iommu.c
@@ -2273,6 +2273,15 @@ error:
 	return 0;
 }
 
+static dma_addr_t intel_map_page(struct device *dev, struct page *page,
+				 unsigned long offset, size_t size,
+				 enum dma_data_direction dir,
+				 struct dma_attrs *attrs)
+{
+	return __intel_map_single(dev, page_to_phys(page) + offset, size,
+				  dir, to_pci_dev(dev)->dma_mask);
+}
+
 dma_addr_t intel_map_single(struct device *hwdev, phys_addr_t paddr,
 			    size_t size, int dir)
 {
@@ -2341,8 +2350,9 @@ static void add_unmap(struct dmar_domain *dom, struct iova *iova)
 	spin_unlock_irqrestore(&async_umap_flush_lock, flags);
 }
 
-void intel_unmap_single(struct device *dev, dma_addr_t dev_addr, size_t size,
-			int dir)
+static void intel_unmap_page(struct device *dev, dma_addr_t dev_addr,
+			     size_t size, enum dma_data_direction dir,
+			     struct dma_attrs *attrs)
 {
 	struct pci_dev *pdev = to_pci_dev(dev);
 	struct dmar_domain *domain;
@@ -2386,6 +2396,12 @@ void intel_unmap_single(struct device *dev, dma_addr_t dev_addr, size_t size,
 	}
 }
 
+void intel_unmap_single(struct device *dev, dma_addr_t dev_addr, size_t size,
+			int dir)
+{
+	intel_unmap_page(dev, dev_addr, size, dir, NULL);
+}
+
 void *intel_alloc_coherent(struct device *hwdev, size_t size,
 			   dma_addr_t *dma_handle, gfp_t flags)
 {
@@ -2570,6 +2586,10 @@ static struct dma_mapping_ops intel_dma_ops = {
 	.unmap_single = intel_unmap_single,
 	.map_sg = intel_map_sg,
 	.unmap_sg = intel_unmap_sg,
+#ifdef CONFIG_X86_64
+	.map_page = intel_map_page,
+	.unmap_page = intel_unmap_page,
+#endif
 };
 
 static inline int iommu_domain_cache_init(void)
-- 
1.6.0.6


^ permalink raw reply related	[flat|nested] 16+ messages in thread

* [PATCH 7/8] pci-nommu: add map_page
  2009-01-05 14:47           ` [PATCH 6/8] intel-iommu: " FUJITA Tomonori
@ 2009-01-05 14:47             ` FUJITA Tomonori
  2009-01-05 14:47               ` [PATCH 8/8] remove map_single and unmap_single in struct dma_mapping_ops FUJITA Tomonori
  0 siblings, 1 reply; 16+ messages in thread
From: FUJITA Tomonori @ 2009-01-05 14:47 UTC (permalink / raw)
  To: mingo; +Cc: linux-kernel, FUJITA Tomonori

This is a preparation of struct dma_mapping_ops unification. We use
map_page and unmap_page instead of map_single and unmap_single.

We will remove map_single hook in the last patch in this patchset.

Signed-off-by: FUJITA Tomonori <fujita.tomonori@lab.ntt.co.jp>
---
 arch/x86/kernel/pci-nommu.c |   20 ++++++++++++++------
 1 files changed, 14 insertions(+), 6 deletions(-)

diff --git a/arch/x86/kernel/pci-nommu.c b/arch/x86/kernel/pci-nommu.c
index c70ab5a..5a73a82 100644
--- a/arch/x86/kernel/pci-nommu.c
+++ b/arch/x86/kernel/pci-nommu.c
@@ -25,18 +25,25 @@ check_addr(char *name, struct device *hwdev, dma_addr_t bus, size_t size)
 	return 1;
 }
 
-static dma_addr_t
-nommu_map_single(struct device *hwdev, phys_addr_t paddr, size_t size,
-	       int direction)
+static dma_addr_t nommu_map_page(struct device *dev, struct page *page,
+				 unsigned long offset, size_t size,
+				 enum dma_data_direction dir,
+				 struct dma_attrs *attrs)
 {
-	dma_addr_t bus = paddr;
+	dma_addr_t bus = page_to_phys(page) + offset;
 	WARN_ON(size == 0);
-	if (!check_addr("map_single", hwdev, bus, size))
-				return bad_dma_address;
+	if (!check_addr("map_single", dev, bus, size))
+		return bad_dma_address;
 	flush_write_buffers();
 	return bus;
 }
 
+static dma_addr_t nommu_map_single(struct device *hwdev, phys_addr_t paddr,
+				   size_t size, int direction)
+{
+	return nommu_map_page(hwdev, pfn_to_page(paddr >> PAGE_SHIFT),
+			      paddr & ~PAGE_MASK, size, direction, NULL);
+}
 
 /* Map a set of buffers described by scatterlist in streaming
  * mode for DMA.  This is the scatter-gather version of the
@@ -83,6 +90,7 @@ struct dma_mapping_ops nommu_dma_ops = {
 	.free_coherent = nommu_free_coherent,
 	.map_single = nommu_map_single,
 	.map_sg = nommu_map_sg,
+	.map_page = nommu_map_page,
 	.is_phys = 1,
 };
 
-- 
1.6.0.6


^ permalink raw reply related	[flat|nested] 16+ messages in thread

* [PATCH 8/8] remove map_single and unmap_single in struct dma_mapping_ops
  2009-01-05 14:47             ` [PATCH 7/8] pci-nommu: add map_page FUJITA Tomonori
@ 2009-01-05 14:47               ` FUJITA Tomonori
  2009-01-05 18:00                 ` Joerg Roedel
  0 siblings, 1 reply; 16+ messages in thread
From: FUJITA Tomonori @ 2009-01-05 14:47 UTC (permalink / raw)
  To: mingo; +Cc: linux-kernel, FUJITA Tomonori

This patch converts dma_map_single and dma_unmap_single to use
map_page and unmap_page respectively and removes unnecessary
map_single and unmap_single in struct dma_mapping_ops.

This leaves intel-iommu's dma_map_single and dma_unmap_single since
IA64 uses them. They will be removed after the unification.

Signed-off-by: FUJITA Tomonori <fujita.tomonori@lab.ntt.co.jp>
---
 arch/x86/include/asm/dma-mapping.h |   15 ++++++---------
 arch/x86/kernel/amd_iommu.c        |   15 ---------------
 arch/x86/kernel/pci-calgary_64.c   |   16 ----------------
 arch/x86/kernel/pci-gart_64.c      |   19 ++-----------------
 arch/x86/kernel/pci-nommu.c        |    8 --------
 arch/x86/kernel/pci-swiotlb_64.c   |    9 ---------
 drivers/pci/intel-iommu.c          |    2 --
 7 files changed, 8 insertions(+), 76 deletions(-)

diff --git a/arch/x86/include/asm/dma-mapping.h b/arch/x86/include/asm/dma-mapping.h
index 2f89d2e..a0cb867 100644
--- a/arch/x86/include/asm/dma-mapping.h
+++ b/arch/x86/include/asm/dma-mapping.h
@@ -25,10 +25,6 @@ struct dma_mapping_ops {
 				dma_addr_t *dma_handle, gfp_t gfp);
 	void            (*free_coherent)(struct device *dev, size_t size,
 				void *vaddr, dma_addr_t dma_handle);
-	dma_addr_t      (*map_single)(struct device *hwdev, phys_addr_t ptr,
-				size_t size, int direction);
-	void            (*unmap_single)(struct device *dev, dma_addr_t addr,
-				size_t size, int direction);
 	void            (*sync_single_for_cpu)(struct device *hwdev,
 				dma_addr_t dma_handle, size_t size,
 				int direction);
@@ -105,7 +101,9 @@ dma_map_single(struct device *hwdev, void *ptr, size_t size,
 
 	BUG_ON(!valid_dma_direction(direction));
 	kmemcheck_mark_initialized(ptr, size);
-	return ops->map_single(hwdev, virt_to_phys(ptr), size, direction);
+	return ops->map_page(hwdev, virt_to_page(ptr),
+			     (unsigned long)ptr & ~PAGE_MASK, size,
+			     direction, NULL);
 }
 
 static inline void
@@ -115,8 +113,8 @@ dma_unmap_single(struct device *dev, dma_addr_t addr, size_t size,
 	struct dma_mapping_ops *ops = get_dma_ops(dev);
 
 	BUG_ON(!valid_dma_direction(direction));
-	if (ops->unmap_single)
-		ops->unmap_single(dev, addr, size, direction);
+	if (ops->unmap_page)
+		ops->unmap_page(dev, addr, size, direction, NULL);
 }
 
 static inline int
@@ -223,8 +221,7 @@ static inline dma_addr_t dma_map_page(struct device *dev, struct page *page,
 	struct dma_mapping_ops *ops = get_dma_ops(dev);
 
 	BUG_ON(!valid_dma_direction(direction));
-	return ops->map_single(dev, page_to_phys(page) + offset,
-			       size, direction);
+	return ops->map_page(dev, page, offset, size, direction, NULL);
 }
 
 static inline void dma_unmap_page(struct device *dev, dma_addr_t addr,
diff --git a/arch/x86/kernel/amd_iommu.c b/arch/x86/kernel/amd_iommu.c
index 8570441..a5dedb6 100644
--- a/arch/x86/kernel/amd_iommu.c
+++ b/arch/x86/kernel/amd_iommu.c
@@ -1341,13 +1341,6 @@ out:
 	return addr;
 }
 
-static dma_addr_t map_single(struct device *dev, phys_addr_t paddr,
-			     size_t size, int dir)
-{
-	return map_page(dev, pfn_to_page(paddr >> PAGE_SHIFT),
-			paddr & ~PAGE_MASK, size, dir, NULL);
-}
-
 /*
  * The exported unmap_single function for dma_ops.
  */
@@ -1378,12 +1371,6 @@ static void unmap_page(struct device *dev, dma_addr_t dma_addr, size_t size,
 	spin_unlock_irqrestore(&domain->lock, flags);
 }
 
-static void unmap_single(struct device *dev, dma_addr_t dma_addr,
-			 size_t size, int dir)
-{
-	return unmap_page(dev, dma_addr, size, dir, NULL);
-}
-
 /*
  * This is a special map_sg function which is used if we should map a
  * device which is not handled by an AMD IOMMU in the system.
@@ -1664,8 +1651,6 @@ static void prealloc_protection_domains(void)
 static struct dma_mapping_ops amd_iommu_dma_ops = {
 	.alloc_coherent = alloc_coherent,
 	.free_coherent = free_coherent,
-	.map_single = map_single,
-	.unmap_single = unmap_single,
 	.map_page = map_page,
 	.unmap_page = unmap_page,
 	.map_sg = map_sg,
diff --git a/arch/x86/kernel/pci-calgary_64.c b/arch/x86/kernel/pci-calgary_64.c
index e33cfcf..756138b 100644
--- a/arch/x86/kernel/pci-calgary_64.c
+++ b/arch/x86/kernel/pci-calgary_64.c
@@ -461,14 +461,6 @@ static dma_addr_t calgary_map_page(struct device *dev, struct page *page,
 	return iommu_alloc(dev, tbl, vaddr, npages, dir);
 }
 
-static dma_addr_t calgary_map_single(struct device *dev, phys_addr_t paddr,
-				     size_t size, int direction)
-{
-	return calgary_map_page(dev, pfn_to_page(paddr >> PAGE_SHIFT),
-				paddr & ~PAGE_MASK, size,
-				direction, NULL);
-}
-
 static void calgary_unmap_page(struct device *dev, dma_addr_t dma_addr,
 			       size_t size, enum dma_data_direction dir,
 			       struct dma_attrs *attrs)
@@ -480,12 +472,6 @@ static void calgary_unmap_page(struct device *dev, dma_addr_t dma_addr,
 	iommu_free(tbl, dma_addr, npages);
 }
 
-static void calgary_unmap_single(struct device *dev, dma_addr_t dma_handle,
-				 size_t size, int direction)
-{
-	calgary_unmap_page(dev, dma_handle, size, direction, NULL);
-}
-
 static void* calgary_alloc_coherent(struct device *dev, size_t size,
 	dma_addr_t *dma_handle, gfp_t flag)
 {
@@ -535,8 +521,6 @@ static void calgary_free_coherent(struct device *dev, size_t size,
 static struct dma_mapping_ops calgary_dma_ops = {
 	.alloc_coherent = calgary_alloc_coherent,
 	.free_coherent = calgary_free_coherent,
-	.map_single = calgary_map_single,
-	.unmap_single = calgary_unmap_single,
 	.map_sg = calgary_map_sg,
 	.unmap_sg = calgary_unmap_sg,
 	.map_page = calgary_map_page,
diff --git a/arch/x86/kernel/pci-gart_64.c b/arch/x86/kernel/pci-gart_64.c
index e49c6dd..9c557c0 100644
--- a/arch/x86/kernel/pci-gart_64.c
+++ b/arch/x86/kernel/pci-gart_64.c
@@ -275,13 +275,6 @@ static dma_addr_t gart_map_page(struct device *dev, struct page *page,
 	return bus;
 }
 
-static dma_addr_t gart_map_single(struct device *dev, phys_addr_t paddr,
-				  size_t size, int dir)
-{
-	return gart_map_page(dev, pfn_to_page(paddr >> PAGE_SHIFT),
-			     paddr & ~PAGE_MASK, size, dir, NULL);
-}
-
 /*
  * Free a DMA mapping.
  */
@@ -306,12 +299,6 @@ static void gart_unmap_page(struct device *dev, dma_addr_t dma_addr,
 	free_iommu(iommu_page, npages);
 }
 
-static void gart_unmap_single(struct device *dev, dma_addr_t dma_addr,
-			      size_t size, int direction)
-{
-	gart_unmap_page(dev, dma_addr, size, direction, NULL);
-}
-
 /*
  * Wrapper for pci_unmap_single working with scatterlists.
  */
@@ -324,7 +311,7 @@ gart_unmap_sg(struct device *dev, struct scatterlist *sg, int nents, int dir)
 	for_each_sg(sg, s, nents, i) {
 		if (!s->dma_length || !s->length)
 			break;
-		gart_unmap_single(dev, s->dma_address, s->dma_length, dir);
+		gart_unmap_page(dev, s->dma_address, s->dma_length, dir, NULL);
 	}
 }
 
@@ -538,7 +525,7 @@ static void
 gart_free_coherent(struct device *dev, size_t size, void *vaddr,
 		   dma_addr_t dma_addr)
 {
-	gart_unmap_single(dev, dma_addr, size, DMA_BIDIRECTIONAL);
+	gart_unmap_page(dev, dma_addr, size, DMA_BIDIRECTIONAL, NULL);
 	free_pages((unsigned long)vaddr, get_order(size));
 }
 
@@ -725,8 +712,6 @@ static __init int init_k8_gatt(struct agp_kern_info *info)
 }
 
 static struct dma_mapping_ops gart_dma_ops = {
-	.map_single			= gart_map_single,
-	.unmap_single			= gart_unmap_single,
 	.map_sg				= gart_map_sg,
 	.unmap_sg			= gart_unmap_sg,
 	.map_page			= gart_map_page,
diff --git a/arch/x86/kernel/pci-nommu.c b/arch/x86/kernel/pci-nommu.c
index 5a73a82..d42b69c 100644
--- a/arch/x86/kernel/pci-nommu.c
+++ b/arch/x86/kernel/pci-nommu.c
@@ -38,13 +38,6 @@ static dma_addr_t nommu_map_page(struct device *dev, struct page *page,
 	return bus;
 }
 
-static dma_addr_t nommu_map_single(struct device *hwdev, phys_addr_t paddr,
-				   size_t size, int direction)
-{
-	return nommu_map_page(hwdev, pfn_to_page(paddr >> PAGE_SHIFT),
-			      paddr & ~PAGE_MASK, size, direction, NULL);
-}
-
 /* Map a set of buffers described by scatterlist in streaming
  * mode for DMA.  This is the scatter-gather version of the
  * above pci_map_single interface.  Here the scatter gather list
@@ -88,7 +81,6 @@ static void nommu_free_coherent(struct device *dev, size_t size, void *vaddr,
 struct dma_mapping_ops nommu_dma_ops = {
 	.alloc_coherent = dma_generic_alloc_coherent,
 	.free_coherent = nommu_free_coherent,
-	.map_single = nommu_map_single,
 	.map_sg = nommu_map_sg,
 	.map_page = nommu_map_page,
 	.is_phys = 1,
diff --git a/arch/x86/kernel/pci-swiotlb_64.c b/arch/x86/kernel/pci-swiotlb_64.c
index d1c0366..3ae354c 100644
--- a/arch/x86/kernel/pci-swiotlb_64.c
+++ b/arch/x86/kernel/pci-swiotlb_64.c
@@ -38,13 +38,6 @@ int __weak swiotlb_arch_range_needs_mapping(void *ptr, size_t size)
 	return 0;
 }
 
-static dma_addr_t
-swiotlb_map_single_phys(struct device *hwdev, phys_addr_t paddr, size_t size,
-			int direction)
-{
-	return swiotlb_map_single(hwdev, phys_to_virt(paddr), size, direction);
-}
-
 /* these will be moved to lib/swiotlb.c later on */
 
 static dma_addr_t swiotlb_map_page(struct device *dev, struct page *page,
@@ -78,8 +71,6 @@ struct dma_mapping_ops swiotlb_dma_ops = {
 	.mapping_error = swiotlb_dma_mapping_error,
 	.alloc_coherent = x86_swiotlb_alloc_coherent,
 	.free_coherent = swiotlb_free_coherent,
-	.map_single = swiotlb_map_single_phys,
-	.unmap_single = swiotlb_unmap_single,
 	.sync_single_for_cpu = swiotlb_sync_single_for_cpu,
 	.sync_single_for_device = swiotlb_sync_single_for_device,
 	.sync_single_range_for_cpu = swiotlb_sync_single_range_for_cpu,
diff --git a/drivers/pci/intel-iommu.c b/drivers/pci/intel-iommu.c
index 60258ec..da273e4 100644
--- a/drivers/pci/intel-iommu.c
+++ b/drivers/pci/intel-iommu.c
@@ -2582,8 +2582,6 @@ int intel_map_sg(struct device *hwdev, struct scatterlist *sglist, int nelems,
 static struct dma_mapping_ops intel_dma_ops = {
 	.alloc_coherent = intel_alloc_coherent,
 	.free_coherent = intel_free_coherent,
-	.map_single = intel_map_single,
-	.unmap_single = intel_unmap_single,
 	.map_sg = intel_map_sg,
 	.unmap_sg = intel_unmap_sg,
 #ifdef CONFIG_X86_64
-- 
1.6.0.6


^ permalink raw reply related	[flat|nested] 16+ messages in thread

* Re: [PATCH 0/13] x86: unifying ways to handle multiple sets of dma mapping ops
  2009-01-05 14:47 [PATCH 0/13] x86: unifying ways to handle multiple sets of dma mapping ops FUJITA Tomonori
  2009-01-05 14:47 ` [PATCH 1/8] add map_page and unmap_page to struct dma_mapping_ops FUJITA Tomonori
@ 2009-01-05 15:00 ` FUJITA Tomonori
  1 sibling, 0 replies; 16+ messages in thread
From: FUJITA Tomonori @ 2009-01-05 15:00 UTC (permalink / raw)
  To: mingo; +Cc: linux-kernel, fujita.tomonori

On Mon,  5 Jan 2009 23:47:20 +0900
FUJITA Tomonori <fujita.tomonori@lab.ntt.co.jp> wrote:

> This patchset is the second part of the unification of ways to handle
> multiple sets of dma mapping API. The whole work consists of three
> patchset. This is for X86 and can be applied independently (against
> tip/master).

Oops, I messed up the subject, should have been:

[PATCH 0/8] x86: unifying ways to handle multiple sets of dma mapping ops

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: [PATCH 1/8] add map_page and unmap_page to struct dma_mapping_ops
  2009-01-05 14:47 ` [PATCH 1/8] add map_page and unmap_page to struct dma_mapping_ops FUJITA Tomonori
  2009-01-05 14:47   ` [PATCH 2/8] swiotlb: add map_page and unmap_page FUJITA Tomonori
@ 2009-01-05 17:26   ` Joerg Roedel
  2009-01-05 17:28     ` Joerg Roedel
  1 sibling, 1 reply; 16+ messages in thread
From: Joerg Roedel @ 2009-01-05 17:26 UTC (permalink / raw)
  To: FUJITA Tomonori; +Cc: mingo, linux-kernel

On Mon, Jan 05, 2009 at 11:47:21PM +0900, FUJITA Tomonori wrote:
> This patch adds map_page and unmap_page to struct dma_mapping_ops.
> 
> This is a preparation of struct dma_mapping_ops unification. We use
> map_page and unmap_page instead of map_single and unmap_single.
> 
> We will remove map_single and unmap_single hooks in the last patch in
> this patchset.
> 
> Signed-off-by: FUJITA Tomonori <fujita.tomonori@lab.ntt.co.jp>
> ---
>  arch/x86/include/asm/dma-mapping.h |    8 ++++++++
>  1 files changed, 8 insertions(+), 0 deletions(-)
> 
> diff --git a/arch/x86/include/asm/dma-mapping.h b/arch/x86/include/asm/dma-mapping.h
> index e93265c..2f89d2e 100644
> --- a/arch/x86/include/asm/dma-mapping.h
> +++ b/arch/x86/include/asm/dma-mapping.h
> @@ -8,6 +8,7 @@
>  
>  #include <linux/kmemcheck.h>
>  #include <linux/scatterlist.h>
> +#include <linux/dma-attrs.h>
>  #include <asm/io.h>
>  #include <asm/swiotlb.h>
>  #include <asm-generic/dma-coherent.h>
> @@ -51,6 +52,13 @@ struct dma_mapping_ops {
>  	void            (*unmap_sg)(struct device *hwdev,
>  				struct scatterlist *sg, int nents,
>  				int direction);
> +	dma_addr_t	(*map_page)(struct device *dev, struct page *page,
> +				    unsigned long offset, size_t size,
> +				    enum dma_data_direction dir,
> +				    struct dma_attrs *attrs);
> +	void 		(*unmap_page)(struct device *dev, dma_addr_t dma_handle,
> +				      size_t size, enum dma_data_direction dir,
> +				      struct dma_attrs *attrs);

Why do we need an offset into the page? The name suggests that this
function maps a whole page so the offset should be irrelevant.

Joerg

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: [PATCH 1/8] add map_page and unmap_page to struct dma_mapping_ops
  2009-01-05 17:26   ` [PATCH 1/8] add map_page and unmap_page to struct dma_mapping_ops Joerg Roedel
@ 2009-01-05 17:28     ` Joerg Roedel
  0 siblings, 0 replies; 16+ messages in thread
From: Joerg Roedel @ 2009-01-05 17:28 UTC (permalink / raw)
  To: FUJITA Tomonori; +Cc: mingo, linux-kernel

On Mon, Jan 05, 2009 at 06:26:08PM +0100, Joerg Roedel wrote:
> On Mon, Jan 05, 2009 at 11:47:21PM +0900, FUJITA Tomonori wrote:
> > This patch adds map_page and unmap_page to struct dma_mapping_ops.
> > 
> > This is a preparation of struct dma_mapping_ops unification. We use
> > map_page and unmap_page instead of map_single and unmap_single.
> > 
> > We will remove map_single and unmap_single hooks in the last patch in
> > this patchset.
> > 
> > Signed-off-by: FUJITA Tomonori <fujita.tomonori@lab.ntt.co.jp>
> > ---
> >  arch/x86/include/asm/dma-mapping.h |    8 ++++++++
> >  1 files changed, 8 insertions(+), 0 deletions(-)
> > 
> > diff --git a/arch/x86/include/asm/dma-mapping.h b/arch/x86/include/asm/dma-mapping.h
> > index e93265c..2f89d2e 100644
> > --- a/arch/x86/include/asm/dma-mapping.h
> > +++ b/arch/x86/include/asm/dma-mapping.h
> > @@ -8,6 +8,7 @@
> >  
> >  #include <linux/kmemcheck.h>
> >  #include <linux/scatterlist.h>
> > +#include <linux/dma-attrs.h>
> >  #include <asm/io.h>
> >  #include <asm/swiotlb.h>
> >  #include <asm-generic/dma-coherent.h>
> > @@ -51,6 +52,13 @@ struct dma_mapping_ops {
> >  	void            (*unmap_sg)(struct device *hwdev,
> >  				struct scatterlist *sg, int nents,
> >  				int direction);
> > +	dma_addr_t	(*map_page)(struct device *dev, struct page *page,
> > +				    unsigned long offset, size_t size,
> > +				    enum dma_data_direction dir,
> > +				    struct dma_attrs *attrs);
> > +	void 		(*unmap_page)(struct device *dev, dma_addr_t dma_handle,
> > +				      size_t size, enum dma_data_direction dir,
> > +				      struct dma_attrs *attrs);
> 
> Why do we need an offset into the page? The name suggests that this
> function maps a whole page so the offset should be irrelevant.

Ah, just saw it. Forget this stupid question :)


^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: [PATCH 8/8] remove map_single and unmap_single in struct dma_mapping_ops
  2009-01-05 14:47               ` [PATCH 8/8] remove map_single and unmap_single in struct dma_mapping_ops FUJITA Tomonori
@ 2009-01-05 18:00                 ` Joerg Roedel
  2009-01-06 10:22                   ` FUJITA Tomonori
  0 siblings, 1 reply; 16+ messages in thread
From: Joerg Roedel @ 2009-01-05 18:00 UTC (permalink / raw)
  To: FUJITA Tomonori; +Cc: mingo, linux-kernel

Is it the right way to implement map_single in terms of map_page? Doing
this you optimize for the map_page case. But a grep in drivers/ shows:

linux/drivers $ grep -r _map_page *|wc -l
126
linux/drivers $ grep -r _map_single *|wc -l
613

There are a lot more users of map_single than of map_page. I think its
better to optimize for the map_single case and implement map_page in
terms of map_single.

Joerg

On Mon, Jan 05, 2009 at 11:47:28PM +0900, FUJITA Tomonori wrote:
> This patch converts dma_map_single and dma_unmap_single to use
> map_page and unmap_page respectively and removes unnecessary
> map_single and unmap_single in struct dma_mapping_ops.
> 
> This leaves intel-iommu's dma_map_single and dma_unmap_single since
> IA64 uses them. They will be removed after the unification.
> 
> Signed-off-by: FUJITA Tomonori <fujita.tomonori@lab.ntt.co.jp>
> ---
>  arch/x86/include/asm/dma-mapping.h |   15 ++++++---------
>  arch/x86/kernel/amd_iommu.c        |   15 ---------------
>  arch/x86/kernel/pci-calgary_64.c   |   16 ----------------
>  arch/x86/kernel/pci-gart_64.c      |   19 ++-----------------
>  arch/x86/kernel/pci-nommu.c        |    8 --------
>  arch/x86/kernel/pci-swiotlb_64.c   |    9 ---------
>  drivers/pci/intel-iommu.c          |    2 --
>  7 files changed, 8 insertions(+), 76 deletions(-)
> 
> diff --git a/arch/x86/include/asm/dma-mapping.h b/arch/x86/include/asm/dma-mapping.h
> index 2f89d2e..a0cb867 100644
> --- a/arch/x86/include/asm/dma-mapping.h
> +++ b/arch/x86/include/asm/dma-mapping.h
> @@ -25,10 +25,6 @@ struct dma_mapping_ops {
>  				dma_addr_t *dma_handle, gfp_t gfp);
>  	void            (*free_coherent)(struct device *dev, size_t size,
>  				void *vaddr, dma_addr_t dma_handle);
> -	dma_addr_t      (*map_single)(struct device *hwdev, phys_addr_t ptr,
> -				size_t size, int direction);
> -	void            (*unmap_single)(struct device *dev, dma_addr_t addr,
> -				size_t size, int direction);
>  	void            (*sync_single_for_cpu)(struct device *hwdev,
>  				dma_addr_t dma_handle, size_t size,
>  				int direction);
> @@ -105,7 +101,9 @@ dma_map_single(struct device *hwdev, void *ptr, size_t size,
>  
>  	BUG_ON(!valid_dma_direction(direction));
>  	kmemcheck_mark_initialized(ptr, size);
> -	return ops->map_single(hwdev, virt_to_phys(ptr), size, direction);
> +	return ops->map_page(hwdev, virt_to_page(ptr),
> +			     (unsigned long)ptr & ~PAGE_MASK, size,
> +			     direction, NULL);
>  }
>  
>  static inline void
> @@ -115,8 +113,8 @@ dma_unmap_single(struct device *dev, dma_addr_t addr, size_t size,
>  	struct dma_mapping_ops *ops = get_dma_ops(dev);
>  
>  	BUG_ON(!valid_dma_direction(direction));
> -	if (ops->unmap_single)
> -		ops->unmap_single(dev, addr, size, direction);
> +	if (ops->unmap_page)
> +		ops->unmap_page(dev, addr, size, direction, NULL);
>  }
>  
>  static inline int
> @@ -223,8 +221,7 @@ static inline dma_addr_t dma_map_page(struct device *dev, struct page *page,
>  	struct dma_mapping_ops *ops = get_dma_ops(dev);
>  
>  	BUG_ON(!valid_dma_direction(direction));
> -	return ops->map_single(dev, page_to_phys(page) + offset,
> -			       size, direction);
> +	return ops->map_page(dev, page, offset, size, direction, NULL);
>  }
>  
>  static inline void dma_unmap_page(struct device *dev, dma_addr_t addr,
> diff --git a/arch/x86/kernel/amd_iommu.c b/arch/x86/kernel/amd_iommu.c
> index 8570441..a5dedb6 100644
> --- a/arch/x86/kernel/amd_iommu.c
> +++ b/arch/x86/kernel/amd_iommu.c
> @@ -1341,13 +1341,6 @@ out:
>  	return addr;
>  }
>  
> -static dma_addr_t map_single(struct device *dev, phys_addr_t paddr,
> -			     size_t size, int dir)
> -{
> -	return map_page(dev, pfn_to_page(paddr >> PAGE_SHIFT),
> -			paddr & ~PAGE_MASK, size, dir, NULL);
> -}
> -
>  /*
>   * The exported unmap_single function for dma_ops.
>   */
> @@ -1378,12 +1371,6 @@ static void unmap_page(struct device *dev, dma_addr_t dma_addr, size_t size,
>  	spin_unlock_irqrestore(&domain->lock, flags);
>  }
>  
> -static void unmap_single(struct device *dev, dma_addr_t dma_addr,
> -			 size_t size, int dir)
> -{
> -	return unmap_page(dev, dma_addr, size, dir, NULL);
> -}
> -
>  /*
>   * This is a special map_sg function which is used if we should map a
>   * device which is not handled by an AMD IOMMU in the system.
> @@ -1664,8 +1651,6 @@ static void prealloc_protection_domains(void)
>  static struct dma_mapping_ops amd_iommu_dma_ops = {
>  	.alloc_coherent = alloc_coherent,
>  	.free_coherent = free_coherent,
> -	.map_single = map_single,
> -	.unmap_single = unmap_single,
>  	.map_page = map_page,
>  	.unmap_page = unmap_page,
>  	.map_sg = map_sg,
> diff --git a/arch/x86/kernel/pci-calgary_64.c b/arch/x86/kernel/pci-calgary_64.c
> index e33cfcf..756138b 100644
> --- a/arch/x86/kernel/pci-calgary_64.c
> +++ b/arch/x86/kernel/pci-calgary_64.c
> @@ -461,14 +461,6 @@ static dma_addr_t calgary_map_page(struct device *dev, struct page *page,
>  	return iommu_alloc(dev, tbl, vaddr, npages, dir);
>  }
>  
> -static dma_addr_t calgary_map_single(struct device *dev, phys_addr_t paddr,
> -				     size_t size, int direction)
> -{
> -	return calgary_map_page(dev, pfn_to_page(paddr >> PAGE_SHIFT),
> -				paddr & ~PAGE_MASK, size,
> -				direction, NULL);
> -}
> -
>  static void calgary_unmap_page(struct device *dev, dma_addr_t dma_addr,
>  			       size_t size, enum dma_data_direction dir,
>  			       struct dma_attrs *attrs)
> @@ -480,12 +472,6 @@ static void calgary_unmap_page(struct device *dev, dma_addr_t dma_addr,
>  	iommu_free(tbl, dma_addr, npages);
>  }
>  
> -static void calgary_unmap_single(struct device *dev, dma_addr_t dma_handle,
> -				 size_t size, int direction)
> -{
> -	calgary_unmap_page(dev, dma_handle, size, direction, NULL);
> -}
> -
>  static void* calgary_alloc_coherent(struct device *dev, size_t size,
>  	dma_addr_t *dma_handle, gfp_t flag)
>  {
> @@ -535,8 +521,6 @@ static void calgary_free_coherent(struct device *dev, size_t size,
>  static struct dma_mapping_ops calgary_dma_ops = {
>  	.alloc_coherent = calgary_alloc_coherent,
>  	.free_coherent = calgary_free_coherent,
> -	.map_single = calgary_map_single,
> -	.unmap_single = calgary_unmap_single,
>  	.map_sg = calgary_map_sg,
>  	.unmap_sg = calgary_unmap_sg,
>  	.map_page = calgary_map_page,
> diff --git a/arch/x86/kernel/pci-gart_64.c b/arch/x86/kernel/pci-gart_64.c
> index e49c6dd..9c557c0 100644
> --- a/arch/x86/kernel/pci-gart_64.c
> +++ b/arch/x86/kernel/pci-gart_64.c
> @@ -275,13 +275,6 @@ static dma_addr_t gart_map_page(struct device *dev, struct page *page,
>  	return bus;
>  }
>  
> -static dma_addr_t gart_map_single(struct device *dev, phys_addr_t paddr,
> -				  size_t size, int dir)
> -{
> -	return gart_map_page(dev, pfn_to_page(paddr >> PAGE_SHIFT),
> -			     paddr & ~PAGE_MASK, size, dir, NULL);
> -}
> -
>  /*
>   * Free a DMA mapping.
>   */
> @@ -306,12 +299,6 @@ static void gart_unmap_page(struct device *dev, dma_addr_t dma_addr,
>  	free_iommu(iommu_page, npages);
>  }
>  
> -static void gart_unmap_single(struct device *dev, dma_addr_t dma_addr,
> -			      size_t size, int direction)
> -{
> -	gart_unmap_page(dev, dma_addr, size, direction, NULL);
> -}
> -
>  /*
>   * Wrapper for pci_unmap_single working with scatterlists.
>   */
> @@ -324,7 +311,7 @@ gart_unmap_sg(struct device *dev, struct scatterlist *sg, int nents, int dir)
>  	for_each_sg(sg, s, nents, i) {
>  		if (!s->dma_length || !s->length)
>  			break;
> -		gart_unmap_single(dev, s->dma_address, s->dma_length, dir);
> +		gart_unmap_page(dev, s->dma_address, s->dma_length, dir, NULL);
>  	}
>  }
>  
> @@ -538,7 +525,7 @@ static void
>  gart_free_coherent(struct device *dev, size_t size, void *vaddr,
>  		   dma_addr_t dma_addr)
>  {
> -	gart_unmap_single(dev, dma_addr, size, DMA_BIDIRECTIONAL);
> +	gart_unmap_page(dev, dma_addr, size, DMA_BIDIRECTIONAL, NULL);
>  	free_pages((unsigned long)vaddr, get_order(size));
>  }
>  
> @@ -725,8 +712,6 @@ static __init int init_k8_gatt(struct agp_kern_info *info)
>  }
>  
>  static struct dma_mapping_ops gart_dma_ops = {
> -	.map_single			= gart_map_single,
> -	.unmap_single			= gart_unmap_single,
>  	.map_sg				= gart_map_sg,
>  	.unmap_sg			= gart_unmap_sg,
>  	.map_page			= gart_map_page,
> diff --git a/arch/x86/kernel/pci-nommu.c b/arch/x86/kernel/pci-nommu.c
> index 5a73a82..d42b69c 100644
> --- a/arch/x86/kernel/pci-nommu.c
> +++ b/arch/x86/kernel/pci-nommu.c
> @@ -38,13 +38,6 @@ static dma_addr_t nommu_map_page(struct device *dev, struct page *page,
>  	return bus;
>  }
>  
> -static dma_addr_t nommu_map_single(struct device *hwdev, phys_addr_t paddr,
> -				   size_t size, int direction)
> -{
> -	return nommu_map_page(hwdev, pfn_to_page(paddr >> PAGE_SHIFT),
> -			      paddr & ~PAGE_MASK, size, direction, NULL);
> -}
> -
>  /* Map a set of buffers described by scatterlist in streaming
>   * mode for DMA.  This is the scatter-gather version of the
>   * above pci_map_single interface.  Here the scatter gather list
> @@ -88,7 +81,6 @@ static void nommu_free_coherent(struct device *dev, size_t size, void *vaddr,
>  struct dma_mapping_ops nommu_dma_ops = {
>  	.alloc_coherent = dma_generic_alloc_coherent,
>  	.free_coherent = nommu_free_coherent,
> -	.map_single = nommu_map_single,
>  	.map_sg = nommu_map_sg,
>  	.map_page = nommu_map_page,
>  	.is_phys = 1,
> diff --git a/arch/x86/kernel/pci-swiotlb_64.c b/arch/x86/kernel/pci-swiotlb_64.c
> index d1c0366..3ae354c 100644
> --- a/arch/x86/kernel/pci-swiotlb_64.c
> +++ b/arch/x86/kernel/pci-swiotlb_64.c
> @@ -38,13 +38,6 @@ int __weak swiotlb_arch_range_needs_mapping(void *ptr, size_t size)
>  	return 0;
>  }
>  
> -static dma_addr_t
> -swiotlb_map_single_phys(struct device *hwdev, phys_addr_t paddr, size_t size,
> -			int direction)
> -{
> -	return swiotlb_map_single(hwdev, phys_to_virt(paddr), size, direction);
> -}
> -
>  /* these will be moved to lib/swiotlb.c later on */
>  
>  static dma_addr_t swiotlb_map_page(struct device *dev, struct page *page,
> @@ -78,8 +71,6 @@ struct dma_mapping_ops swiotlb_dma_ops = {
>  	.mapping_error = swiotlb_dma_mapping_error,
>  	.alloc_coherent = x86_swiotlb_alloc_coherent,
>  	.free_coherent = swiotlb_free_coherent,
> -	.map_single = swiotlb_map_single_phys,
> -	.unmap_single = swiotlb_unmap_single,
>  	.sync_single_for_cpu = swiotlb_sync_single_for_cpu,
>  	.sync_single_for_device = swiotlb_sync_single_for_device,
>  	.sync_single_range_for_cpu = swiotlb_sync_single_range_for_cpu,
> diff --git a/drivers/pci/intel-iommu.c b/drivers/pci/intel-iommu.c
> index 60258ec..da273e4 100644
> --- a/drivers/pci/intel-iommu.c
> +++ b/drivers/pci/intel-iommu.c
> @@ -2582,8 +2582,6 @@ int intel_map_sg(struct device *hwdev, struct scatterlist *sglist, int nelems,
>  static struct dma_mapping_ops intel_dma_ops = {
>  	.alloc_coherent = intel_alloc_coherent,
>  	.free_coherent = intel_free_coherent,
> -	.map_single = intel_map_single,
> -	.unmap_single = intel_unmap_single,
>  	.map_sg = intel_map_sg,
>  	.unmap_sg = intel_unmap_sg,
>  #ifdef CONFIG_X86_64
> -- 
> 1.6.0.6
> 
> --
> To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
> the body of a message to majordomo@vger.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> Please read the FAQ at  http://www.tux.org/lkml/

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: [PATCH 4/8] calgary: add map_page and unmap_page
  2009-01-05 14:47       ` [PATCH 4/8] calgary: " FUJITA Tomonori
  2009-01-05 14:47         ` [PATCH 5/8] AMD IOMMU: " FUJITA Tomonori
@ 2009-01-06  8:32         ` Muli Ben-Yehuda
  2009-01-06 13:07           ` Ingo Molnar
  1 sibling, 1 reply; 16+ messages in thread
From: Muli Ben-Yehuda @ 2009-01-06  8:32 UTC (permalink / raw)
  To: FUJITA Tomonori; +Cc: mingo, linux-kernel

On Mon, Jan 05, 2009 at 11:47:24PM +0900, FUJITA Tomonori wrote:
> This is a preparation of struct dma_mapping_ops unification. We use
> map_page and unmap_page instead of map_single and unmap_single.
> 
> We will remove map_single and unmap_single hooks in the last patch in
> this patchset.
> 
> Signed-off-by: FUJITA Tomonori <fujita.tomonori@lab.ntt.co.jp>
> Cc: Muli Ben-Yehuda <muli@il.ibm.com>

Calgary bits look fine.

Acked-by: Muli Ben-Yehuda <muli@il.ibm.com>

Cheers,
Muli
-- 
SYSTOR 2009---The Israeli Experimental Systems Conference
May 4-6, 2009, Haifa, Israel
http://www.haifa.il.ibm.com/conferences/systor2009/

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: [PATCH 8/8] remove map_single and unmap_single in struct dma_mapping_ops
  2009-01-05 18:00                 ` Joerg Roedel
@ 2009-01-06 10:22                   ` FUJITA Tomonori
  0 siblings, 0 replies; 16+ messages in thread
From: FUJITA Tomonori @ 2009-01-06 10:22 UTC (permalink / raw)
  To: joro; +Cc: fujita.tomonori, mingo, linux-kernel

On Mon, 5 Jan 2009 19:00:38 +0100
Joerg Roedel <joro@8bytes.org> wrote:

> Is it the right way to implement map_single in terms of map_page? Doing
> this you optimize for the map_page case. But a grep in drivers/ shows:
> 
> linux/drivers $ grep -r _map_page *|wc -l
> 126
> linux/drivers $ grep -r _map_single *|wc -l
> 613

The comparison is irrelevant since dma_map_page and dma_map_single
have different purposes.

If passing virtual memory address to an IOMMU is enough (and
convenient), then drivers use dma_map_single.

For some purposes, drivers need to pass a page frame and use
dma_map_page (or dma_map_sg).

We could have two hooks in dma_map_ops struct for dma_map_single and
dma_map_page. Say, we have map_single and map_page hooks. But the
map_page hook can be used to support both dma_map_single and
dma_map_page. Note that the map_single hook can't do that since it use
a virtual address as an argument. That's why I have only the map_page
hook in dma_map_ops struct.

As X86 does now, we could have map_single hook that use a physical
address to handle both dma_map_single and dma_map_page. However, it's
confusing since it means that the arguments of dma_map_single and its
hook (map_single) is inconsistent.


> There are a lot more users of map_single than of map_page. I think its
> better to optimize for the map_single case and implement map_page in
> terms of map_single.

As I wrote above, it doesn't make sense.

^ permalink raw reply	[flat|nested] 16+ messages in thread

* Re: [PATCH 4/8] calgary: add map_page and unmap_page
  2009-01-06  8:32         ` [PATCH 4/8] calgary: add map_page and unmap_page Muli Ben-Yehuda
@ 2009-01-06 13:07           ` Ingo Molnar
  0 siblings, 0 replies; 16+ messages in thread
From: Ingo Molnar @ 2009-01-06 13:07 UTC (permalink / raw)
  To: Muli Ben-Yehuda; +Cc: FUJITA Tomonori, linux-kernel


* Muli Ben-Yehuda <muli@il.ibm.com> wrote:

> On Mon, Jan 05, 2009 at 11:47:24PM +0900, FUJITA Tomonori wrote:
> > This is a preparation of struct dma_mapping_ops unification. We use
> > map_page and unmap_page instead of map_single and unmap_single.
> > 
> > We will remove map_single and unmap_single hooks in the last patch in
> > this patchset.
> > 
> > Signed-off-by: FUJITA Tomonori <fujita.tomonori@lab.ntt.co.jp>
> > Cc: Muli Ben-Yehuda <muli@il.ibm.com>
> 
> Calgary bits look fine.
> 
> Acked-by: Muli Ben-Yehuda <muli@il.ibm.com>

thanks, added your ack to the changelog.

	Ingo

^ permalink raw reply	[flat|nested] 16+ messages in thread

end of thread, other threads:[~2009-01-06 13:08 UTC | newest]

Thread overview: 16+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2009-01-05 14:47 [PATCH 0/13] x86: unifying ways to handle multiple sets of dma mapping ops FUJITA Tomonori
2009-01-05 14:47 ` [PATCH 1/8] add map_page and unmap_page to struct dma_mapping_ops FUJITA Tomonori
2009-01-05 14:47   ` [PATCH 2/8] swiotlb: add map_page and unmap_page FUJITA Tomonori
2009-01-05 14:47     ` [PATCH 3/8] gart: " FUJITA Tomonori
2009-01-05 14:47       ` [PATCH 4/8] calgary: " FUJITA Tomonori
2009-01-05 14:47         ` [PATCH 5/8] AMD IOMMU: " FUJITA Tomonori
2009-01-05 14:47           ` [PATCH 6/8] intel-iommu: " FUJITA Tomonori
2009-01-05 14:47             ` [PATCH 7/8] pci-nommu: add map_page FUJITA Tomonori
2009-01-05 14:47               ` [PATCH 8/8] remove map_single and unmap_single in struct dma_mapping_ops FUJITA Tomonori
2009-01-05 18:00                 ` Joerg Roedel
2009-01-06 10:22                   ` FUJITA Tomonori
2009-01-06  8:32         ` [PATCH 4/8] calgary: add map_page and unmap_page Muli Ben-Yehuda
2009-01-06 13:07           ` Ingo Molnar
2009-01-05 17:26   ` [PATCH 1/8] add map_page and unmap_page to struct dma_mapping_ops Joerg Roedel
2009-01-05 17:28     ` Joerg Roedel
2009-01-05 15:00 ` [PATCH 0/13] x86: unifying ways to handle multiple sets of dma mapping ops FUJITA Tomonori

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).