linux-kernel.vger.kernel.org archive mirror
 help / color / mirror / Atom feed
* [PATCH 0/7 v2] extract a simple dmaengine library from shdma.c
@ 2012-01-26 14:56 Guennadi Liakhovetski
  2012-01-26 14:56 ` [PATCH 1/7 v2] dmaengine: add a simple dma library Guennadi Liakhovetski
                   ` (6 more replies)
  0 siblings, 7 replies; 33+ messages in thread
From: Guennadi Liakhovetski @ 2012-01-26 14:56 UTC (permalink / raw)
  To: linux-kernel
  Cc: linux-sh, Vinod Koul, Magnus Damm, Yoshihiro Shimoda, linux-mmc,
	alsa-devel, linux-serial, Paul Mundt

This is v2 of the simple DMA dmaengine library. It addresses Paul's 
comments to v1 - thanks, adds some comments.

Guennadi Liakhovetski (7):
  dmaengine: add a simple dma library
  dma: shdma: prepare for simple DMA conversion
  mmc: sh_mmcif: remove unneeded struct sh_mmcif_dma, prepare for
    simple DMA
  mmc: sh_mobile_sdhi: prepare for conversion to simple DMA
  serial: sh-sci: prepare for conversion to simple DMA
  ASoC: SIU: prepare for conversion to simple DMA
  dma: shdma: convert to the simple DMA library

 drivers/dma/Kconfig               |    4 +
 drivers/dma/Makefile              |    1 +
 drivers/dma/dma-simple.c          |  873 ++++++++++++++++++++++++++++
 drivers/dma/shdma.c               | 1138 +++++++++----------------------------
 drivers/dma/shdma.h               |   45 +-
 drivers/mmc/host/sh_mmcif.c       |   24 +-
 drivers/mmc/host/sh_mobile_sdhi.c |    8 +-
 drivers/tty/serial/sh-sci.c       |    8 +-
 include/linux/dma-simple.h        |  124 ++++
 include/linux/mmc/sh_mmcif.h      |    8 +-
 include/linux/sh_dma.h            |   39 +-
 sound/soc/sh/siu_pcm.c            |    8 +-
 12 files changed, 1334 insertions(+), 946 deletions(-)
 create mode 100644 drivers/dma/dma-simple.c
 create mode 100644 include/linux/dma-simple.h

-- 
1.7.2.5

Thanks
Guennadi
---
Guennadi Liakhovetski, Ph.D.
Freelance Open-Source Software Developer
http://www.open-technology.de/

^ permalink raw reply	[flat|nested] 33+ messages in thread

* [PATCH 1/7 v2] dmaengine: add a simple dma library
  2012-01-26 14:56 [PATCH 0/7 v2] extract a simple dmaengine library from shdma.c Guennadi Liakhovetski
@ 2012-01-26 14:56 ` Guennadi Liakhovetski
  2012-01-26 17:34   ` [alsa-devel] " Sascha Hauer
                     ` (3 more replies)
  2012-01-26 14:56 ` [PATCH 2/7 v2] dma: shdma: prepare for simple DMA conversion Guennadi Liakhovetski
                   ` (5 subsequent siblings)
  6 siblings, 4 replies; 33+ messages in thread
From: Guennadi Liakhovetski @ 2012-01-26 14:56 UTC (permalink / raw)
  To: linux-kernel
  Cc: linux-sh, Vinod Koul, Magnus Damm, Yoshihiro Shimoda, linux-mmc,
	alsa-devel, linux-serial, Paul Mundt

This patch adds a library of functions, helping to implement dmaengine
drivers for hardware, unable to handle scatter-gather lists natively.
The first version of this driver only supports memcpy and slave DMA
operation.

Signed-off-by: Guennadi Liakhovetski <g.liakhovetski@gmx.de>
---

v2:

1. switch from using a tasklet to threaded IRQ, which allowed to
2. remove lock / unlock inline functions
3. remove __devinit, __devexit annotations

 drivers/dma/Kconfig        |    3 +
 drivers/dma/Makefile       |    1 +
 drivers/dma/dma-simple.c   |  873 ++++++++++++++++++++++++++++++++++++++++++++
 include/linux/dma-simple.h |  124 +++++++
 4 files changed, 1001 insertions(+), 0 deletions(-)
 create mode 100644 drivers/dma/dma-simple.c
 create mode 100644 include/linux/dma-simple.h

diff --git a/drivers/dma/Kconfig b/drivers/dma/Kconfig
index f1a2749..f7c583e 100644
--- a/drivers/dma/Kconfig
+++ b/drivers/dma/Kconfig
@@ -149,6 +149,9 @@ config TXX9_DMAC
 	  Support the TXx9 SoC internal DMA controller.  This can be
 	  integrated in chips such as the Toshiba TX4927/38/39.
 
+config DMA_SIMPLE
+	tristate
+
 config SH_DMAE
 	tristate "Renesas SuperH DMAC support"
 	depends on (SUPERH && SH_DMA) || (ARM && ARCH_SHMOBILE)
diff --git a/drivers/dma/Makefile b/drivers/dma/Makefile
index 009a222..d63f773 100644
--- a/drivers/dma/Makefile
+++ b/drivers/dma/Makefile
@@ -2,6 +2,7 @@ ccflags-$(CONFIG_DMADEVICES_DEBUG)  := -DDEBUG
 ccflags-$(CONFIG_DMADEVICES_VDEBUG) += -DVERBOSE_DEBUG
 
 obj-$(CONFIG_DMA_ENGINE) += dmaengine.o
+obj-$(CONFIG_DMA_SIMPLE) += dma-simple.o
 obj-$(CONFIG_NET_DMA) += iovlock.o
 obj-$(CONFIG_INTEL_MID_DMAC) += intel_mid_dma.o
 obj-$(CONFIG_DMATEST) += dmatest.o
diff --git a/drivers/dma/dma-simple.c b/drivers/dma/dma-simple.c
new file mode 100644
index 0000000..49d8f7d
--- /dev/null
+++ b/drivers/dma/dma-simple.c
@@ -0,0 +1,873 @@
+/*
+ * Simple dmaengine driver library
+ *
+ * extracted from shdma.c
+ *
+ * Copyright (C) 2011-2012 Guennadi Liakhovetski <g.liakhovetski@gmx.de>
+ * Copyright (C) 2009 Nobuhiro Iwamatsu <iwamatsu.nobuhiro@renesas.com>
+ * Copyright (C) 2009 Renesas Solutions, Inc. All rights reserved.
+ * Copyright (C) 2007 Freescale Semiconductor, Inc. All rights reserved.
+ *
+ * This is free software; you can redistribute it and/or modify
+ * it under the terms of version 2 of the GNU General Public License as
+ * published by the Free Software Foundation.
+ */
+
+#include <linux/delay.h>
+#include <linux/dma-simple.h>
+#include <linux/dmaengine.h>
+#include <linux/init.h>
+#include <linux/interrupt.h>
+#include <linux/module.h>
+#include <linux/pm_runtime.h>
+#include <linux/slab.h>
+#include <linux/spinlock.h>
+
+/* DMA descriptor control */
+enum simple_desc_status {
+	DESC_IDLE,
+	DESC_PREPARED,
+	DESC_SUBMITTED,
+	DESC_COMPLETED,	/* completed, have to call callback */
+	DESC_WAITING,	/* callback called, waiting for ack / re-submit */
+};
+
+#define NR_DESCS_PER_CHANNEL 32
+
+#define to_simple_chan(c) container_of(c, struct dma_simple_chan, dma_chan)
+#define to_simple_dev(d) container_of(d, struct dma_simple_dev, dma_dev)
+
+/*
+ * For slave DMA we assume, that there is a finite number of DMA slaves in the
+ * system, and that each such slave can only use a finite number of channels.
+ * We use slave channel IDs to make sure, that no such slave channel ID is
+ * allocated more than once.
+ */
+static unsigned int slave_num = 256;
+module_param(slave_num, uint, 0444);
+
+/* A bitmask with slave_num bits */
+static unsigned long *simple_slave_used;
+
+/* Called under spin_lock_irq(&schan->chan_lock") */
+static void simple_chan_xfer_ld_queue(struct dma_simple_chan *schan)
+{
+	struct dma_simple_dev *sdev = to_simple_dev(schan->dma_chan.device);
+	const struct dma_simple_ops *ops = sdev->ops;
+	struct dma_simple_desc *sdesc;
+
+	/* DMA work check */
+	if (ops->channel_busy(schan))
+		return;
+
+	/* Find the first not transferred descriptor */
+	list_for_each_entry(sdesc, &schan->ld_queue, node)
+		if (sdesc->mark == DESC_SUBMITTED) {
+			ops->start_xfer(schan, sdesc);
+			break;
+		}
+}
+
+static dma_cookie_t simple_tx_submit(struct dma_async_tx_descriptor *tx)
+{
+	struct dma_simple_desc *chunk, *c, *desc =
+		container_of(tx, struct dma_simple_desc, async_tx),
+		*last = desc;
+	struct dma_simple_chan *schan = to_simple_chan(tx->chan);
+	struct dma_simple_slave *slave = tx->chan->private;
+	dma_async_tx_callback callback = tx->callback;
+	dma_cookie_t cookie;
+	bool power_up;
+
+	spin_lock_irq(&schan->chan_lock);
+
+	power_up = list_empty(&schan->ld_queue);
+
+	cookie = schan->dma_chan.cookie + 1;
+	if (cookie < 0)
+		cookie = 1;
+
+	schan->dma_chan.cookie = cookie;
+	tx->cookie = cookie;
+
+	/* Mark all chunks of this descriptor as submitted, move to the queue */
+	list_for_each_entry_safe(chunk, c, desc->node.prev, node) {
+		/*
+		 * All chunks are on the global ld_free, so, we have to find
+		 * the end of the chain ourselves
+		 */
+		if (chunk != desc && (chunk->mark == DESC_IDLE ||
+				      chunk->async_tx.cookie > 0 ||
+				      chunk->async_tx.cookie == -EBUSY ||
+				      &chunk->node == &schan->ld_free))
+			break;
+		chunk->mark = DESC_SUBMITTED;
+		/* Callback goes to the last chunk */
+		chunk->async_tx.callback = NULL;
+		chunk->cookie = cookie;
+		list_move_tail(&chunk->node, &schan->ld_queue);
+		last = chunk;
+
+		dev_dbg(schan->dev, "submit #%d@%p on %d\n",
+			tx->cookie, &last->async_tx, schan->id);
+	}
+
+	last->async_tx.callback = callback;
+	last->async_tx.callback_param = tx->callback_param;
+
+	if (power_up) {
+		int ret;
+		schan->pm_state = DMA_SIMPLE_PM_BUSY;
+
+		ret = pm_runtime_get(schan->dev);
+
+		spin_unlock_irq(&schan->chan_lock);
+		if (ret < 0)
+			dev_err(schan->dev, "%s(): GET = %d\n", __func__, ret);
+
+		pm_runtime_barrier(schan->dev);
+
+		spin_lock_irq(&schan->chan_lock);
+
+		/* Have we been reset, while waiting? */
+		if (schan->pm_state != DMA_SIMPLE_PM_ESTABLISHED) {
+			struct dma_simple_dev *sdev =
+				to_simple_dev(schan->dma_chan.device);
+			const struct dma_simple_ops *ops = sdev->ops;
+			dev_dbg(schan->dev, "Bring up channel %d\n",
+				schan->id);
+			/*
+			 * TODO: .xfer_setup() might fail on some platforms.
+			 * Make it int then, on error remove chunks from the
+			 * queue again
+			 */
+			ops->setup_xfer(schan, slave);
+
+			if (schan->pm_state == DMA_SIMPLE_PM_PENDING)
+				simple_chan_xfer_ld_queue(schan);
+			schan->pm_state = DMA_SIMPLE_PM_ESTABLISHED;
+		}
+	} else {
+		/*
+		 * Tell .device_issue_pending() not to run the queue, interrupts
+		 * will do it anyway
+		 */
+		schan->pm_state = DMA_SIMPLE_PM_PENDING;
+	}
+
+	spin_unlock_irq(&schan->chan_lock);
+
+	return cookie;
+}
+
+/* Called with desc_lock held */
+static struct dma_simple_desc *simple_get_desc(struct dma_simple_chan *schan)
+{
+	struct dma_simple_desc *sdesc;
+
+	list_for_each_entry(sdesc, &schan->ld_free, node)
+		if (sdesc->mark != DESC_PREPARED) {
+			BUG_ON(sdesc->mark != DESC_IDLE);
+			list_del(&sdesc->node);
+			return sdesc;
+		}
+
+	return NULL;
+}
+
+static int simple_alloc_chan_resources(struct dma_chan *chan)
+{
+	struct dma_simple_chan *schan = to_simple_chan(chan);
+	struct dma_simple_dev *sdev = to_simple_dev(schan->dma_chan.device);
+	const struct dma_simple_ops *ops = sdev->ops;
+	struct dma_simple_desc *desc;
+	struct dma_simple_slave *slave = chan->private;
+	int ret, i;
+
+	/*
+	 * This relies on the guarantee from dmaengine that alloc_chan_resources
+	 * never runs concurrently with itself or free_chan_resources.
+	 */
+	if (slave) {
+		if (test_and_set_bit(slave->slave_id, simple_slave_used)) {
+			ret = -EBUSY;
+			goto etestused;
+		}
+
+		ret = ops->set_slave(schan, slave);
+		if (ret < 0)
+			goto esetslave;
+	}
+
+	schan->desc = kcalloc(NR_DESCS_PER_CHANNEL,
+			      sdev->desc_size, GFP_KERNEL);
+	if (!schan->desc) {
+		ret = -ENOMEM;
+		goto edescalloc;
+	}
+	schan->desc_num = NR_DESCS_PER_CHANNEL;
+
+	for (i = 0; i < NR_DESCS_PER_CHANNEL; i++) {
+		desc = ops->embedded_desc(schan->desc, i);
+		dma_async_tx_descriptor_init(&desc->async_tx,
+					     &schan->dma_chan);
+		desc->async_tx.tx_submit = simple_tx_submit;
+		desc->mark = DESC_IDLE;
+
+		list_add(&desc->node, &schan->ld_free);
+	}
+
+	return NR_DESCS_PER_CHANNEL;
+
+edescalloc:
+	if (slave)
+esetslave:
+		clear_bit(slave->slave_id, simple_slave_used);
+etestused:
+	chan->private = NULL;
+	return ret;
+}
+
+static dma_async_tx_callback __ld_cleanup(struct dma_simple_chan *schan, bool all)
+{
+	struct dma_simple_desc *desc, *_desc;
+	/* Is the "exposed" head of a chain acked? */
+	bool head_acked = false;
+	dma_cookie_t cookie = 0;
+	dma_async_tx_callback callback = NULL;
+	void *param = NULL;
+	unsigned long flags;
+
+	spin_lock_irqsave(&schan->chan_lock, flags);
+	list_for_each_entry_safe(desc, _desc, &schan->ld_queue, node) {
+		struct dma_async_tx_descriptor *tx = &desc->async_tx;
+
+		BUG_ON(tx->cookie > 0 && tx->cookie != desc->cookie);
+		BUG_ON(desc->mark != DESC_SUBMITTED &&
+		       desc->mark != DESC_COMPLETED &&
+		       desc->mark != DESC_WAITING);
+
+		/*
+		 * queue is ordered, and we use this loop to (1) clean up all
+		 * completed descriptors, and to (2) update descriptor flags of
+		 * any chunks in a (partially) completed chain
+		 */
+		if (!all && desc->mark == DESC_SUBMITTED &&
+		    desc->cookie != cookie)
+			break;
+
+		if (tx->cookie > 0)
+			cookie = tx->cookie;
+
+		if (desc->mark == DESC_COMPLETED && desc->chunks == 1) {
+			if (schan->completed_cookie != desc->cookie - 1)
+				dev_dbg(schan->dev,
+					"Completing cookie %d, expected %d\n",
+					desc->cookie,
+					schan->completed_cookie + 1);
+			schan->completed_cookie = desc->cookie;
+		}
+
+		/* Call callback on the last chunk */
+		if (desc->mark == DESC_COMPLETED && tx->callback) {
+			desc->mark = DESC_WAITING;
+			callback = tx->callback;
+			param = tx->callback_param;
+			dev_dbg(schan->dev, "descriptor #%d@%p on %d callback\n",
+				tx->cookie, tx, schan->id);
+			BUG_ON(desc->chunks != 1);
+			break;
+		}
+
+		if (tx->cookie > 0 || tx->cookie == -EBUSY) {
+			if (desc->mark == DESC_COMPLETED) {
+				BUG_ON(tx->cookie < 0);
+				desc->mark = DESC_WAITING;
+			}
+			head_acked = async_tx_test_ack(tx);
+		} else {
+			switch (desc->mark) {
+			case DESC_COMPLETED:
+				desc->mark = DESC_WAITING;
+				/* Fall through */
+			case DESC_WAITING:
+				if (head_acked)
+					async_tx_ack(&desc->async_tx);
+			}
+		}
+
+		dev_dbg(schan->dev, "descriptor %p #%d completed.\n",
+			tx, tx->cookie);
+
+		if (((desc->mark == DESC_COMPLETED ||
+		      desc->mark == DESC_WAITING) &&
+		     async_tx_test_ack(&desc->async_tx)) || all) {
+			/* Remove from ld_queue list */
+			desc->mark = DESC_IDLE;
+
+			list_move(&desc->node, &schan->ld_free);
+
+			if (list_empty(&schan->ld_queue)) {
+				dev_dbg(schan->dev, "Bring down channel %d\n", schan->id);
+				pm_runtime_put(schan->dev);
+				schan->pm_state = DMA_SIMPLE_PM_ESTABLISHED;
+			}
+		}
+	}
+
+	if (all && !callback)
+		/*
+		 * Terminating and the loop completed normally: forgive
+		 * uncompleted cookies
+		 */
+		schan->completed_cookie = schan->dma_chan.cookie;
+
+	spin_unlock_irqrestore(&schan->chan_lock, flags);
+
+	if (callback)
+		callback(param);
+
+	return callback;
+}
+
+/*
+ * simple_chan_ld_cleanup - Clean up link descriptors
+ *
+ * Clean up the ld_queue of DMA channel.
+ */
+static void simple_chan_ld_cleanup(struct dma_simple_chan *schan, bool all)
+{
+	while (__ld_cleanup(schan, all))
+		;
+}
+
+/*
+ * simple_free_chan_resources - Free all resources of the channel.
+ */
+static void simple_free_chan_resources(struct dma_chan *chan)
+{
+	struct dma_simple_chan *schan = to_simple_chan(chan);
+	struct dma_simple_dev *sdev = to_simple_dev(chan->device);
+	const struct dma_simple_ops *ops = sdev->ops;
+	LIST_HEAD(list);
+
+	/* Protect against ISR */
+	spin_lock_irq(&schan->chan_lock);
+	ops->halt_channel(schan);
+	spin_unlock_irq(&schan->chan_lock);
+
+	/* Now no new interrupts will occur */
+
+	/* Prepared and not submitted descriptors can still be on the queue */
+	if (!list_empty(&schan->ld_queue))
+		simple_chan_ld_cleanup(schan, true);
+
+	if (chan->private) {
+		/* The caller is holding dma_list_mutex */
+		struct dma_simple_slave *slave = chan->private;
+		clear_bit(slave->slave_id, simple_slave_used);
+		chan->private = NULL;
+	}
+
+	spin_lock_irq(&schan->chan_lock);
+
+	list_splice_init(&schan->ld_free, &list);
+	schan->desc_num = 0;
+
+	spin_unlock_irq(&schan->chan_lock);
+
+	kfree(schan->desc);
+}
+
+/**
+ * simple_add_desc - get, set up and return one transfer descriptor
+ * @schan:	DMA channel
+ * @flags:	DMA transfer flags
+ * @dst:	destination DMA address, incremented when direction equals
+ *		DMA_DEV_TO_MEM or DMA_MEM_TO_MEM
+ * @src:	source DMA address, incremented when direction equals
+ *		DMA_MEM_TO_DEV or DMA_MEM_TO_MEM
+ * @len:	DMA transfer length
+ * @first:	if NULL, set to the current descriptor and cookie set to -EBUSY
+ * @direction:	needed for slave DMA to decide which address to keep constant,
+ *		equals DMA_MEM_TO_MEM for MEMCPY
+ * Returns 0 or an error
+ * Locks: called with desc_lock held
+ */
+static struct dma_simple_desc *simple_add_desc(struct dma_simple_chan *schan,
+	unsigned long flags, dma_addr_t *dst, dma_addr_t *src, size_t *len,
+	struct dma_simple_desc **first, enum dma_transfer_direction direction)
+{
+	struct dma_simple_dev *sdev = to_simple_dev(schan->dma_chan.device);
+	const struct dma_simple_ops *ops = sdev->ops;
+	struct dma_simple_desc *new;
+	size_t copy_size = *len;
+
+	if (!copy_size)
+		return NULL;
+
+	/* Allocate the link descriptor from the free list */
+	new = simple_get_desc(schan);
+	if (!new) {
+		dev_err(schan->dev, "No free link descriptor available\n");
+		return NULL;
+	}
+
+	ops->desc_setup(schan, new, *src, *dst, &copy_size);
+
+	if (!*first) {
+		/* First desc */
+		new->async_tx.cookie = -EBUSY;
+		*first = new;
+	} else {
+		/* Other desc - invisible to the user */
+		new->async_tx.cookie = -EINVAL;
+	}
+
+	dev_dbg(schan->dev,
+		"chaining (%u/%u)@%x -> %x with %p, cookie %d\n",
+		copy_size, *len, *src, *dst, &new->async_tx,
+		new->async_tx.cookie);
+
+	new->mark = DESC_PREPARED;
+	new->async_tx.flags = flags;
+	new->direction = direction;
+
+	*len -= copy_size;
+	if (direction == DMA_MEM_TO_MEM || direction == DMA_MEM_TO_DEV)
+		*src += copy_size;
+	if (direction == DMA_MEM_TO_MEM || direction == DMA_DEV_TO_MEM)
+		*dst += copy_size;
+
+	return new;
+}
+
+/*
+ * simple_prep_sg - prepare transfer descriptors from an SG list
+ *
+ * Common routine for public (MEMCPY) and slave DMA. The MEMCPY case is also
+ * converted to scatter-gather to guarantee consistent locking and a correct
+ * list manipulation. For slave DMA direction carries the usual meaning, and,
+ * logically, the SG list is RAM and the addr variable contains slave address,
+ * e.g., the FIFO I/O register. For MEMCPY direction equals DMA_MEM_TO_MEM
+ * and the SG list contains only one element and points at the source buffer.
+ */
+static struct dma_async_tx_descriptor *simple_prep_sg(struct dma_simple_chan *schan,
+	struct scatterlist *sgl, unsigned int sg_len, dma_addr_t *addr,
+	enum dma_transfer_direction direction, unsigned long flags)
+{
+	struct scatterlist *sg;
+	struct dma_simple_desc *first = NULL, *new = NULL /* compiler... */;
+	LIST_HEAD(tx_list);
+	int chunks = 0;
+	unsigned long irq_flags;
+	int i;
+
+	for_each_sg(sgl, sg, sg_len, i)
+		chunks += DIV_ROUND_UP(sg_dma_len(sg), schan->max_xfer_len);
+
+	/* Have to lock the whole loop to protect against concurrent release */
+	spin_lock_irqsave(&schan->chan_lock, irq_flags);
+
+	/*
+	 * Chaining:
+	 * first descriptor is what user is dealing with in all API calls, its
+	 *	cookie is at first set to -EBUSY, at tx-submit to a positive
+	 *	number
+	 * if more than one chunk is needed further chunks have cookie = -EINVAL
+	 * the last chunk, if not equal to the first, has cookie = -ENOSPC
+	 * all chunks are linked onto the tx_list head with their .node heads
+	 *	only during this function, then they are immediately spliced
+	 *	back onto the free list in form of a chain
+	 */
+	for_each_sg(sgl, sg, sg_len, i) {
+		dma_addr_t sg_addr = sg_dma_address(sg);
+		size_t len = sg_dma_len(sg);
+
+		if (!len)
+			goto err_get_desc;
+
+		do {
+			dev_dbg(schan->dev, "Add SG #%d@%p[%d], dma %llx\n",
+				i, sg, len, (unsigned long long)sg_addr);
+
+			if (direction == DMA_DEV_TO_MEM)
+				new = simple_add_desc(schan, flags,
+						&sg_addr, addr, &len, &first,
+						direction);
+			else
+				new = simple_add_desc(schan, flags,
+						addr, &sg_addr, &len, &first,
+						direction);
+			if (!new)
+				goto err_get_desc;
+
+			new->chunks = chunks--;
+			list_add_tail(&new->node, &tx_list);
+		} while (len);
+	}
+
+	if (new != first)
+		new->async_tx.cookie = -ENOSPC;
+
+	/* Put them back on the free list, so, they don't get lost */
+	list_splice_tail(&tx_list, &schan->ld_free);
+
+	spin_unlock_irqrestore(&schan->chan_lock, irq_flags);
+
+	return &first->async_tx;
+
+err_get_desc:
+	list_for_each_entry(new, &tx_list, node)
+		new->mark = DESC_IDLE;
+	list_splice(&tx_list, &schan->ld_free);
+
+	spin_unlock_irqrestore(&schan->chan_lock, irq_flags);
+
+	return NULL;
+}
+
+static struct dma_async_tx_descriptor *simple_prep_memcpy(
+	struct dma_chan *chan, dma_addr_t dma_dest, dma_addr_t dma_src,
+	size_t len, unsigned long flags)
+{
+	struct dma_simple_chan *schan = to_simple_chan(chan);
+	struct scatterlist sg;
+
+	if (!chan || !len)
+		return NULL;
+
+	BUG_ON(!schan->desc_num);
+
+	sg_init_table(&sg, 1);
+	sg_set_page(&sg, pfn_to_page(PFN_DOWN(dma_src)), len,
+		    offset_in_page(dma_src));
+	sg_dma_address(&sg) = dma_src;
+	sg_dma_len(&sg) = len;
+
+	return simple_prep_sg(schan, &sg, 1, &dma_dest, DMA_MEM_TO_MEM, flags);
+}
+
+static struct dma_async_tx_descriptor *simple_prep_slave_sg(
+	struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len,
+	enum dma_transfer_direction direction, unsigned long flags)
+{
+	struct dma_simple_chan *schan = to_simple_chan(chan);
+	struct dma_simple_dev *sdev = to_simple_dev(schan->dma_chan.device);
+	const struct dma_simple_ops *ops = sdev->ops;
+	struct dma_simple_slave *slave = chan->private;
+	dma_addr_t slave_addr;
+
+	if (!chan)
+		return NULL;
+
+	BUG_ON(!schan->desc_num);
+
+	/* Someone calling slave DMA on a generic channel? */
+	if (!slave || !sg_len) {
+		dev_warn(schan->dev, "%s: bad parameter: %p, %d, %d\n",
+			 __func__, slave, sg_len, slave ? slave->slave_id : -1);
+		return NULL;
+	}
+
+	slave_addr = ops->slave_addr(schan);
+
+	return simple_prep_sg(schan, sgl, sg_len, &slave_addr,
+			      direction, flags);
+}
+
+static int simple_control(struct dma_chan *chan, enum dma_ctrl_cmd cmd,
+			  unsigned long arg)
+{
+	struct dma_simple_chan *schan = to_simple_chan(chan);
+	struct dma_simple_dev *sdev = to_simple_dev(chan->device);
+	const struct dma_simple_ops *ops = sdev->ops;
+	unsigned long flags;
+
+	/* Only supports DMA_TERMINATE_ALL */
+	if (cmd != DMA_TERMINATE_ALL)
+		return -ENXIO;
+
+	if (!chan)
+		return -EINVAL;
+
+	spin_lock_irqsave(&schan->chan_lock, flags);
+
+	ops->halt_channel(schan);
+
+	spin_unlock_irqrestore(&schan->chan_lock, flags);
+
+	simple_chan_ld_cleanup(schan, true);
+
+	return 0;
+}
+
+static void simple_issue_pending(struct dma_chan *chan)
+{
+	struct dma_simple_chan *schan = to_simple_chan(chan);
+
+	spin_lock_irq(&schan->chan_lock);
+	if (schan->pm_state == DMA_SIMPLE_PM_ESTABLISHED)
+		simple_chan_xfer_ld_queue(schan);
+	else
+		schan->pm_state = DMA_SIMPLE_PM_PENDING;
+	spin_unlock_irq(&schan->chan_lock);
+}
+
+static enum dma_status simple_tx_status(struct dma_chan *chan,
+					dma_cookie_t cookie,
+					struct dma_tx_state *txstate)
+{
+	struct dma_simple_chan *schan = to_simple_chan(chan);
+	dma_cookie_t last_used;
+	dma_cookie_t last_complete;
+	enum dma_status status;
+	unsigned long flags;
+
+	simple_chan_ld_cleanup(schan, false);
+
+	/* First read completed cookie to avoid a skew */
+	last_complete = schan->completed_cookie;
+	rmb();
+	last_used = chan->cookie;
+	BUG_ON(last_complete < 0);
+	dma_set_tx_state(txstate, last_complete, last_used, 0);
+
+	spin_lock_irqsave(&schan->chan_lock, flags);
+
+	status = dma_async_is_complete(cookie, last_complete, last_used);
+
+	/*
+	 * If we don't find cookie on the queue, it has been aborted and we have
+	 * to report error
+	 */
+	if (status != DMA_SUCCESS) {
+		struct dma_simple_desc *sdesc;
+		status = DMA_ERROR;
+		list_for_each_entry(sdesc, &schan->ld_queue, node)
+			if (sdesc->cookie == cookie) {
+				status = DMA_IN_PROGRESS;
+				break;
+			}
+	}
+
+	spin_unlock_irqrestore(&schan->chan_lock, flags);
+
+	return status;
+}
+
+/* Called from error IRQ or NMI */
+bool dma_simple_reset(struct dma_simple_dev *sdev)
+{
+	const struct dma_simple_ops *ops = sdev->ops;
+	struct dma_simple_chan *schan;
+	unsigned int handled = 0;
+	int i;
+
+	/* Reset all channels */
+	dma_simple_for_each_chan(schan, sdev, i) {
+		struct dma_simple_desc *sdesc;
+		LIST_HEAD(dl);
+
+		if (!schan)
+			continue;
+
+		spin_lock(&schan->chan_lock);
+
+		/* Stop the channel */
+		ops->halt_channel(schan);
+
+		list_splice_init(&schan->ld_queue, &dl);
+
+		if (!list_empty(&dl)) {
+			dev_dbg(schan->dev, "Bring down channel %d\n", schan->id);
+			pm_runtime_put(schan->dev);
+		}
+		schan->pm_state = DMA_SIMPLE_PM_ESTABLISHED;
+
+		spin_unlock(&schan->chan_lock);
+
+		/* Complete all  */
+		list_for_each_entry(sdesc, &dl, node) {
+			struct dma_async_tx_descriptor *tx = &sdesc->async_tx;
+			sdesc->mark = DESC_IDLE;
+			if (tx->callback)
+				tx->callback(tx->callback_param);
+		}
+
+		spin_lock(&schan->chan_lock);
+		list_splice(&dl, &schan->ld_free);
+		spin_unlock(&schan->chan_lock);
+
+		handled++;
+	}
+
+	return !!handled;
+}
+EXPORT_SYMBOL(dma_simple_reset);
+
+static irqreturn_t chan_irq(int irq, void *dev)
+{
+	struct dma_simple_chan *schan = dev;
+	const struct dma_simple_ops *ops =
+		to_simple_dev(schan->dma_chan.device)->ops;
+	irqreturn_t ret;
+
+	spin_lock(&schan->chan_lock);
+
+	ret = ops->chan_irq(schan, irq) ? IRQ_WAKE_THREAD : IRQ_NONE;
+
+	spin_unlock(&schan->chan_lock);
+
+	return ret;
+}
+
+static irqreturn_t chan_irqt(int irq, void *dev)
+{
+	struct dma_simple_chan *schan = dev;
+	const struct dma_simple_ops *ops =
+		to_simple_dev(schan->dma_chan.device)->ops;
+	struct dma_simple_desc *sdesc;
+
+	spin_lock_irq(&schan->chan_lock);
+	list_for_each_entry(sdesc, &schan->ld_queue, node) {
+		if (sdesc->mark == DESC_SUBMITTED &&
+		    ops->desc_completed(schan, sdesc)) {
+			dev_dbg(schan->dev, "done #%d@%p\n",
+				sdesc->async_tx.cookie, &sdesc->async_tx);
+			sdesc->mark = DESC_COMPLETED;
+			break;
+		}
+	}
+	/* Next desc */
+	simple_chan_xfer_ld_queue(schan);
+	spin_unlock_irq(&schan->chan_lock);
+
+	simple_chan_ld_cleanup(schan, false);
+
+	return IRQ_HANDLED;
+}
+
+int dma_simple_request_irq(struct dma_simple_chan *schan, int irq,
+			   unsigned long flags, const char *name)
+{
+	int ret = request_threaded_irq(irq, chan_irq, chan_irqt,
+				       flags, name, schan);
+
+	schan->irq = ret < 0 ? ret : irq;
+
+	return ret;
+}
+EXPORT_SYMBOL(dma_simple_request_irq);
+
+void dma_simple_free_irq(struct dma_simple_chan *schan)
+{
+	if (schan->irq >= 0)
+		free_irq(schan->irq, schan);
+}
+EXPORT_SYMBOL(dma_simple_free_irq);
+
+void dma_simple_chan_probe(struct dma_simple_dev *sdev,
+			   struct dma_simple_chan *schan, int id)
+{
+	schan->pm_state = DMA_SIMPLE_PM_ESTABLISHED;
+
+	/* reference struct dma_device */
+	schan->dma_chan.device = &sdev->dma_dev;
+
+	schan->dev = sdev->dma_dev.dev;
+	schan->id = id;
+
+	if (!schan->max_xfer_len)
+		schan->max_xfer_len = PAGE_SIZE;
+
+	spin_lock_init(&schan->chan_lock);
+
+	/* Init descripter manage list */
+	INIT_LIST_HEAD(&schan->ld_queue);
+	INIT_LIST_HEAD(&schan->ld_free);
+
+	/* Add the channel to DMA device channel list */
+	list_add_tail(&schan->dma_chan.device_node,
+			&sdev->dma_dev.channels);
+	sdev->schan[sdev->dma_dev.chancnt++] = schan;
+}
+EXPORT_SYMBOL(dma_simple_chan_probe);
+
+void dma_simple_chan_remove(struct dma_simple_chan *schan)
+{
+	list_del(&schan->dma_chan.device_node);
+}
+EXPORT_SYMBOL(dma_simple_chan_remove);
+
+int dma_simple_init(struct device *dev, struct dma_simple_dev *sdev,
+		    int chan_num)
+{
+	struct dma_device *dma_dev = &sdev->dma_dev;
+
+	/*
+	 * Require all call-backs for now, they can trivially be made optional
+	 * later as required
+	 */
+	if (!sdev->ops ||
+	    !sdev->desc_size ||
+	    !sdev->ops->embedded_desc ||
+	    !sdev->ops->start_xfer ||
+	    !sdev->ops->setup_xfer ||
+	    !sdev->ops->set_slave ||
+	    !sdev->ops->desc_setup ||
+	    !sdev->ops->slave_addr ||
+	    !sdev->ops->channel_busy ||
+	    !sdev->ops->halt_channel ||
+	    !sdev->ops->desc_completed)
+		return -EINVAL;
+
+	sdev->schan = kcalloc(chan_num, sizeof(*sdev->schan), GFP_KERNEL);
+	if (!sdev->schan)
+		return -ENOMEM;
+
+	INIT_LIST_HEAD(&dma_dev->channels);
+
+	/* Common and MEMCPY operations */
+	dma_dev->device_alloc_chan_resources
+		= simple_alloc_chan_resources;
+	dma_dev->device_free_chan_resources = simple_free_chan_resources;
+	dma_dev->device_prep_dma_memcpy = simple_prep_memcpy;
+	dma_dev->device_tx_status = simple_tx_status;
+	dma_dev->device_issue_pending = simple_issue_pending;
+
+	/* Compulsory for DMA_SLAVE fields */
+	dma_dev->device_prep_slave_sg = simple_prep_slave_sg;
+	dma_dev->device_control = simple_control;
+
+	dma_dev->dev = dev;
+
+	return 0;
+}
+EXPORT_SYMBOL(dma_simple_init);
+
+void dma_simple_cleanup(struct dma_simple_dev *sdev)
+{
+	kfree(sdev->schan);
+}
+EXPORT_SYMBOL(dma_simple_cleanup);
+
+static int __init dma_simple_enter(void)
+{
+	simple_slave_used = kzalloc(DIV_ROUND_UP(slave_num, BITS_PER_LONG) *
+				    sizeof(long), GFP_KERNEL);
+	if (!simple_slave_used)
+		return -ENOMEM;
+	return 0;
+}
+module_init(dma_simple_enter);
+
+static void __exit dma_simple_exit(void)
+{
+	kfree(simple_slave_used);
+}
+module_exit(dma_simple_exit);
+
+MODULE_LICENSE("GPL v2");
+MODULE_DESCRIPTION("Simple dmaengine driver library");
+MODULE_AUTHOR("Guennadi Liakhovetski <g.liakhovetski@gmx.de>");
diff --git a/include/linux/dma-simple.h b/include/linux/dma-simple.h
new file mode 100644
index 0000000..5336674
--- /dev/null
+++ b/include/linux/dma-simple.h
@@ -0,0 +1,124 @@
+/*
+ * Simple dmaengine driver library
+ *
+ * extracted from shdma.c and headers
+ *
+ * Copyright (C) 2011-2012 Guennadi Liakhovetski <g.liakhovetski@gmx.de>
+ * Copyright (C) 2009 Nobuhiro Iwamatsu <iwamatsu.nobuhiro@renesas.com>
+ * Copyright (C) 2009 Renesas Solutions, Inc. All rights reserved.
+ * Copyright (C) 2007 Freescale Semiconductor, Inc. All rights reserved.
+ *
+ * This is free software; you can redistribute it and/or modify
+ * it under the terms of version 2 of the GNU General Public License as
+ * published by the Free Software Foundation.
+ */
+
+#ifndef DMA_SIMPLE_H
+#define DMA_SIMPLE_H
+
+#include <linux/dmaengine.h>
+#include <linux/interrupt.h>
+#include <linux/list.h>
+#include <linux/types.h>
+
+/**
+ * dma_simple_pm_state - DMA channel PM state
+ * DMA_SIMPLE_PM_ESTABLISHED:	either idle or during data transfer
+ * DMA_SIMPLE_PM_BUSY:		during the transfer preparation, when we have to
+ *				drop the lock temporarily
+ * DMA_SIMPLE_PM_PENDING:	transfers pending
+ */
+enum dma_simple_pm_state {
+	DMA_SIMPLE_PM_ESTABLISHED,
+	DMA_SIMPLE_PM_BUSY,
+	DMA_SIMPLE_PM_PENDING,
+};
+
+struct device;
+
+/*
+ * Drivers, using this library are expected to embed struct dma_simple_dev,
+ * struct dma_simple_chan, struct dma_simple_desc, and struct dma_simple_slave
+ * in their respective device, channel, descriptor and slave objects.
+ */
+
+struct dma_simple_slave {
+	unsigned int slave_id;
+};
+
+struct dma_simple_desc {
+	struct list_head node;
+	struct dma_async_tx_descriptor async_tx;
+	enum dma_transfer_direction direction;
+	dma_cookie_t cookie;
+	int chunks;
+	int mark;
+};
+
+struct dma_simple_chan {
+	dma_cookie_t completed_cookie;	/* The maximum cookie completed */
+	spinlock_t chan_lock;		/* Channel operation lock */
+	struct list_head ld_queue;	/* Link descriptors queue */
+	struct list_head ld_free;	/* Free link descriptors */
+	struct dma_chan dma_chan;	/* DMA channel */
+	struct device *dev;		/* Channel device */
+	void *desc;			/* buffer for descriptor array */
+	int desc_num;			/* desc count */
+	size_t max_xfer_len;		/* max transfer length */
+	int id;				/* Raw id of this channel */
+	int irq;			/* Channel IRQ */
+	enum dma_simple_pm_state pm_state;
+};
+
+/**
+ * struct dma_simple_ops - simple DMA driver operations
+ * desc_completed:	return true, if this is the descriptor, that just has
+ *			completed (atomic)
+ * halt_channel:	stop DMA channel operation (atomic)
+ * channel_busy:	return true, if the channel is busy (atomic)
+ * slave_addr:		return slave DMA address
+ * desc_setup:		set up the hardware specific descriptor portion (atomic)
+ * set_slave:		bind channel to a slave
+ * setup_xfer:		configure channel hardware for operation (atomic)
+ * start_xfer:		start the DMA transfer (atomic)
+ * embedded_desc:	return Nth struct dma_simple_desc pointer from the
+ *			descriptor array
+ * chan_irq:		process channel IRQ, return true if a transfer has
+ *			completed (atomic)
+ */
+struct dma_simple_ops {
+	bool (*desc_completed)(struct dma_simple_chan *, struct dma_simple_desc *);
+	void (*halt_channel)(struct dma_simple_chan *);
+	bool (*channel_busy)(struct dma_simple_chan *);
+	dma_addr_t (*slave_addr)(struct dma_simple_chan *);
+	int (*desc_setup)(struct dma_simple_chan *, struct dma_simple_desc *,
+			  dma_addr_t, dma_addr_t, size_t *);
+	int (*set_slave)(struct dma_simple_chan *, struct dma_simple_slave *);
+	void (*setup_xfer)(struct dma_simple_chan *, struct dma_simple_slave *);
+	void (*start_xfer)(struct dma_simple_chan *, struct dma_simple_desc *);
+	struct dma_simple_desc *(*embedded_desc)(void *, int);
+	bool (*chan_irq)(struct dma_simple_chan *, int);
+};
+
+struct dma_simple_dev {
+	struct dma_device dma_dev;
+	struct dma_simple_chan **schan;
+	const struct dma_simple_ops *ops;
+	size_t desc_size;
+};
+
+#define dma_simple_for_each_chan(c, d, i) for (i = 0, c = (d)->schan[0]; \
+				i < (d)->dma_dev.chancnt; c = (d)->schan[++i])
+
+int dma_simple_request_irq(struct dma_simple_chan *, int,
+			   unsigned long, const char *);
+void dma_simple_free_irq(struct dma_simple_chan *);
+bool dma_simple_reset(struct dma_simple_dev *sdev);
+void dma_simple_chan_probe(struct dma_simple_dev *sdev,
+			   struct dma_simple_chan *schan, int id);
+void dma_simple_chan_remove(struct dma_simple_chan *schan);
+int dma_simple_init(struct device *dev, struct dma_simple_dev *sdev,
+		    int chan_num);
+void dma_simple_cleanup(struct dma_simple_dev *sdev);
+
+#endif
-- 
1.7.2.5


^ permalink raw reply related	[flat|nested] 33+ messages in thread

* [PATCH 2/7 v2] dma: shdma: prepare for simple DMA conversion
  2012-01-26 14:56 [PATCH 0/7 v2] extract a simple dmaengine library from shdma.c Guennadi Liakhovetski
  2012-01-26 14:56 ` [PATCH 1/7 v2] dmaengine: add a simple dma library Guennadi Liakhovetski
@ 2012-01-26 14:56 ` Guennadi Liakhovetski
  2012-01-26 14:56 ` [PATCH 3/7 v2] mmc: sh_mmcif: remove unneeded struct sh_mmcif_dma, prepare for simple DMA Guennadi Liakhovetski
                   ` (4 subsequent siblings)
  6 siblings, 0 replies; 33+ messages in thread
From: Guennadi Liakhovetski @ 2012-01-26 14:56 UTC (permalink / raw)
  To: linux-kernel
  Cc: linux-sh, Vinod Koul, Magnus Damm, Yoshihiro Shimoda, linux-mmc,
	alsa-devel, linux-serial, Paul Mundt

By placing an anonymous union at the top of struct sh_dmae_slave we can
transparently prepare all drivers for the upcoming simple DMA conversion.

Signed-off-by: Guennadi Liakhovetski <g.liakhovetski@gmx.de>
---
 include/linux/sh_dma.h |    8 ++++++--
 1 files changed, 6 insertions(+), 2 deletions(-)

diff --git a/include/linux/sh_dma.h b/include/linux/sh_dma.h
index 425450b..ed2aa1e 100644
--- a/include/linux/sh_dma.h
+++ b/include/linux/sh_dma.h
@@ -10,12 +10,16 @@
 #ifndef SH_DMA_H
 #define SH_DMA_H
 
-#include <linux/list.h>
+#include <linux/dma-simple.h>
 #include <linux/dmaengine.h>
+#include <linux/list.h>
 
 /* Used by slave DMA clients to request DMA to/from a specific peripheral */
 struct sh_dmae_slave {
-	unsigned int			slave_id; /* Set by the platform */
+	union {
+		unsigned int		slave_id; /* Set by the platform */
+		struct dma_simple_slave	simple_slave;
+	};
 	struct device			*dma_dev; /* Set by the platform */
 	const struct sh_dmae_slave_config	*config;  /* Set by the driver */
 };
-- 
1.7.2.5


^ permalink raw reply related	[flat|nested] 33+ messages in thread

* [PATCH 3/7 v2] mmc: sh_mmcif: remove unneeded struct sh_mmcif_dma, prepare for simple DMA
  2012-01-26 14:56 [PATCH 0/7 v2] extract a simple dmaengine library from shdma.c Guennadi Liakhovetski
  2012-01-26 14:56 ` [PATCH 1/7 v2] dmaengine: add a simple dma library Guennadi Liakhovetski
  2012-01-26 14:56 ` [PATCH 2/7 v2] dma: shdma: prepare for simple DMA conversion Guennadi Liakhovetski
@ 2012-01-26 14:56 ` Guennadi Liakhovetski
  2012-01-26 14:56 ` [PATCH 4/7 v2] mmc: sh_mobile_sdhi: prepare for conversion to " Guennadi Liakhovetski
                   ` (3 subsequent siblings)
  6 siblings, 0 replies; 33+ messages in thread
From: Guennadi Liakhovetski @ 2012-01-26 14:56 UTC (permalink / raw)
  To: linux-kernel
  Cc: linux-sh, Vinod Koul, Magnus Damm, Yoshihiro Shimoda, linux-mmc,
	alsa-devel, linux-serial, Paul Mundt

Now that all users have been updated to use the embedded in struct
sh_mmcif_plat_data DMA slave IDs, struct sh_mmcif_dma is no longer needed
and can be removed. This also makes preparation for simple DMA conversion
easier.

Signed-off-by: Guennadi Liakhovetski <g.liakhovetski@gmx.de>
---
 drivers/mmc/host/sh_mmcif.c  |   24 ++++++++++--------------
 include/linux/mmc/sh_mmcif.h |    8 +-------
 2 files changed, 11 insertions(+), 21 deletions(-)

diff --git a/drivers/mmc/host/sh_mmcif.c b/drivers/mmc/host/sh_mmcif.c
index 2ad6c81..06a7cdc 100644
--- a/drivers/mmc/host/sh_mmcif.c
+++ b/drivers/mmc/host/sh_mmcif.c
@@ -384,31 +384,27 @@ static void sh_mmcif_request_dma(struct sh_mmcif_host *host,
 	host->dma_active = false;
 
 	/* We can only either use DMA for both Tx and Rx or not use it at all */
-	if (pdata->dma) {
-		dev_warn(&host->pd->dev,
-			 "Update your platform to use embedded DMA slave IDs\n");
-		tx = &pdata->dma->chan_priv_tx;
-		rx = &pdata->dma->chan_priv_rx;
-	} else {
-		tx = &host->dma_slave_tx;
-		tx->slave_id = pdata->slave_id_tx;
-		rx = &host->dma_slave_rx;
-		rx->slave_id = pdata->slave_id_rx;
-	}
-	if (tx->slave_id > 0 && rx->slave_id > 0) {
+	tx = &host->dma_slave_tx;
+	tx->simple_slave.slave_id = pdata->slave_id_tx;
+	rx = &host->dma_slave_rx;
+	rx->simple_slave.slave_id = pdata->slave_id_rx;
+
+	if (tx->simple_slave.slave_id > 0 && rx->simple_slave.slave_id > 0) {
 		dma_cap_mask_t mask;
 
 		dma_cap_zero(mask);
 		dma_cap_set(DMA_SLAVE, mask);
 
-		host->chan_tx = dma_request_channel(mask, sh_mmcif_filter, tx);
+		host->chan_tx = dma_request_channel(mask, sh_mmcif_filter,
+						    &tx->simple_slave);
 		dev_dbg(&host->pd->dev, "%s: TX: got channel %p\n", __func__,
 			host->chan_tx);
 
 		if (!host->chan_tx)
 			return;
 
-		host->chan_rx = dma_request_channel(mask, sh_mmcif_filter, rx);
+		host->chan_rx = dma_request_channel(mask, sh_mmcif_filter,
+						    &rx->simple_slave);
 		dev_dbg(&host->pd->dev, "%s: RX: got channel %p\n", __func__,
 			host->chan_rx);
 
diff --git a/include/linux/mmc/sh_mmcif.h b/include/linux/mmc/sh_mmcif.h
index 04ff452..b36caa9 100644
--- a/include/linux/mmc/sh_mmcif.h
+++ b/include/linux/mmc/sh_mmcif.h
@@ -32,17 +32,11 @@
  * 1111 : Peripheral clock (sup_pclk set '1')
  */
 
-struct sh_mmcif_dma {
-	struct sh_dmae_slave chan_priv_tx;
-	struct sh_dmae_slave chan_priv_rx;
-};
-
 struct sh_mmcif_plat_data {
 	void (*set_pwr)(struct platform_device *pdev, int state);
 	void (*down_pwr)(struct platform_device *pdev);
 	int (*get_cd)(struct platform_device *pdef);
-	struct sh_mmcif_dma	*dma;		/* Deprecated. Instead */
-	unsigned int		slave_id_tx;	/* use embedded slave_id_[tr]x */
+	unsigned int		slave_id_tx;	/* embedded slave_id_[tr]x */
 	unsigned int		slave_id_rx;
 	u8			sup_pclk;	/* 1 :SH7757, 0: SH7724/SH7372 */
 	unsigned long		caps;
-- 
1.7.2.5


^ permalink raw reply related	[flat|nested] 33+ messages in thread

* [PATCH 4/7 v2] mmc: sh_mobile_sdhi: prepare for conversion to simple DMA
  2012-01-26 14:56 [PATCH 0/7 v2] extract a simple dmaengine library from shdma.c Guennadi Liakhovetski
                   ` (2 preceding siblings ...)
  2012-01-26 14:56 ` [PATCH 3/7 v2] mmc: sh_mmcif: remove unneeded struct sh_mmcif_dma, prepare for simple DMA Guennadi Liakhovetski
@ 2012-01-26 14:56 ` Guennadi Liakhovetski
  2012-01-26 14:56 ` [PATCH 5/7 v2] serial: sh-sci: " Guennadi Liakhovetski
                   ` (2 subsequent siblings)
  6 siblings, 0 replies; 33+ messages in thread
From: Guennadi Liakhovetski @ 2012-01-26 14:56 UTC (permalink / raw)
  To: linux-kernel
  Cc: linux-sh, Vinod Koul, Magnus Damm, Yoshihiro Shimoda, linux-mmc,
	alsa-devel, linux-serial, Paul Mundt

Signed-off-by: Guennadi Liakhovetski <g.liakhovetski@gmx.de>
---
 drivers/mmc/host/sh_mobile_sdhi.c |    8 ++++----
 1 files changed, 4 insertions(+), 4 deletions(-)

diff --git a/drivers/mmc/host/sh_mobile_sdhi.c b/drivers/mmc/host/sh_mobile_sdhi.c
index f91c3aa..2c3a9d3 100644
--- a/drivers/mmc/host/sh_mobile_sdhi.c
+++ b/drivers/mmc/host/sh_mobile_sdhi.c
@@ -130,10 +130,10 @@ static int __devinit sh_mobile_sdhi_probe(struct platform_device *pdev)
 		mmc_data->cd_gpio = p->cd_gpio;
 
 		if (p->dma_slave_tx > 0 && p->dma_slave_rx > 0) {
-			priv->param_tx.slave_id = p->dma_slave_tx;
-			priv->param_rx.slave_id = p->dma_slave_rx;
-			priv->dma_priv.chan_priv_tx = &priv->param_tx;
-			priv->dma_priv.chan_priv_rx = &priv->param_rx;
+			priv->param_tx.simple_slave.slave_id = p->dma_slave_tx;
+			priv->param_rx.simple_slave.slave_id = p->dma_slave_rx;
+			priv->dma_priv.chan_priv_tx = &priv->param_tx.simple_slave;
+			priv->dma_priv.chan_priv_rx = &priv->param_rx.simple_slave;
 			priv->dma_priv.alignment_shift = 1; /* 2-byte alignment */
 			mmc_data->dma = &priv->dma_priv;
 		}
-- 
1.7.2.5


^ permalink raw reply related	[flat|nested] 33+ messages in thread

* [PATCH 5/7 v2] serial: sh-sci: prepare for conversion to simple DMA
  2012-01-26 14:56 [PATCH 0/7 v2] extract a simple dmaengine library from shdma.c Guennadi Liakhovetski
                   ` (3 preceding siblings ...)
  2012-01-26 14:56 ` [PATCH 4/7 v2] mmc: sh_mobile_sdhi: prepare for conversion to " Guennadi Liakhovetski
@ 2012-01-26 14:56 ` Guennadi Liakhovetski
  2012-01-26 14:56 ` [PATCH 6/7 v2] ASoC: SIU: " Guennadi Liakhovetski
  2012-01-26 14:56 ` [PATCH 7/7 v2] dma: shdma: convert to the simple DMA library Guennadi Liakhovetski
  6 siblings, 0 replies; 33+ messages in thread
From: Guennadi Liakhovetski @ 2012-01-26 14:56 UTC (permalink / raw)
  To: linux-kernel
  Cc: linux-sh, Vinod Koul, Magnus Damm, Yoshihiro Shimoda, linux-mmc,
	alsa-devel, linux-serial, Paul Mundt

Signed-off-by: Guennadi Liakhovetski <g.liakhovetski@gmx.de>
---
 drivers/tty/serial/sh-sci.c |    8 ++++----
 1 files changed, 4 insertions(+), 4 deletions(-)

diff --git a/drivers/tty/serial/sh-sci.c b/drivers/tty/serial/sh-sci.c
index 7508579..8b99aac 100644
--- a/drivers/tty/serial/sh-sci.c
+++ b/drivers/tty/serial/sh-sci.c
@@ -1574,9 +1574,9 @@ static bool filter(struct dma_chan *chan, void *slave)
 	struct sh_dmae_slave *param = slave;
 
 	dev_dbg(chan->device->dev, "%s: slave ID %d\n", __func__,
-		param->slave_id);
+		param->simple_slave.slave_id);
 
-	chan->private = param;
+	chan->private = &param->simple_slave;
 	return true;
 }
 
@@ -1615,7 +1615,7 @@ static void sci_request_dma(struct uart_port *port)
 	param = &s->param_tx;
 
 	/* Slave ID, e.g., SHDMA_SLAVE_SCIF0_TX */
-	param->slave_id = s->cfg->dma_slave_tx;
+	param->simple_slave.slave_id = s->cfg->dma_slave_tx;
 
 	s->cookie_tx = -EINVAL;
 	chan = dma_request_channel(mask, filter, param);
@@ -1643,7 +1643,7 @@ static void sci_request_dma(struct uart_port *port)
 	param = &s->param_rx;
 
 	/* Slave ID, e.g., SHDMA_SLAVE_SCIF0_RX */
-	param->slave_id = s->cfg->dma_slave_rx;
+	param->simple_slave.slave_id = s->cfg->dma_slave_rx;
 
 	chan = dma_request_channel(mask, filter, param);
 	dev_dbg(port->dev, "%s: RX: got channel %p\n", __func__, chan);
-- 
1.7.2.5


^ permalink raw reply related	[flat|nested] 33+ messages in thread

* [PATCH 6/7 v2] ASoC: SIU: prepare for conversion to simple DMA
  2012-01-26 14:56 [PATCH 0/7 v2] extract a simple dmaengine library from shdma.c Guennadi Liakhovetski
                   ` (4 preceding siblings ...)
  2012-01-26 14:56 ` [PATCH 5/7 v2] serial: sh-sci: " Guennadi Liakhovetski
@ 2012-01-26 14:56 ` Guennadi Liakhovetski
  2012-02-06  2:33   ` Vinod Koul
  2012-01-26 14:56 ` [PATCH 7/7 v2] dma: shdma: convert to the simple DMA library Guennadi Liakhovetski
  6 siblings, 1 reply; 33+ messages in thread
From: Guennadi Liakhovetski @ 2012-01-26 14:56 UTC (permalink / raw)
  To: linux-kernel
  Cc: linux-sh, Vinod Koul, Magnus Damm, Yoshihiro Shimoda, linux-mmc,
	alsa-devel, linux-serial, Paul Mundt

Signed-off-by: Guennadi Liakhovetski <g.liakhovetski@gmx.de>
---
 sound/soc/sh/siu_pcm.c |    8 ++++----
 1 files changed, 4 insertions(+), 4 deletions(-)

diff --git a/sound/soc/sh/siu_pcm.c b/sound/soc/sh/siu_pcm.c
index 0193e59..414bb35 100644
--- a/sound/soc/sh/siu_pcm.c
+++ b/sound/soc/sh/siu_pcm.c
@@ -330,12 +330,12 @@ static bool filter(struct dma_chan *chan, void *slave)
 {
 	struct sh_dmae_slave *param = slave;
 
-	pr_debug("%s: slave ID %d\n", __func__, param->slave_id);
+	pr_debug("%s: slave ID %d\n", __func__, param->simple_slave.slave_id);
 
 	if (unlikely(param->dma_dev != chan->device->dev))
 		return false;
 
-	chan->private = param;
+	chan->private = &param->simple_slave;
 	return true;
 }
 
@@ -360,12 +360,12 @@ static int siu_pcm_open(struct snd_pcm_substream *ss)
 	if (ss->stream == SNDRV_PCM_STREAM_PLAYBACK) {
 		siu_stream = &port_info->playback;
 		param = &siu_stream->param;
-		param->slave_id = port ? pdata->dma_slave_tx_b :
+		param->simple_slave.slave_id = port ? pdata->dma_slave_tx_b :
 			pdata->dma_slave_tx_a;
 	} else {
 		siu_stream = &port_info->capture;
 		param = &siu_stream->param;
-		param->slave_id = port ? pdata->dma_slave_rx_b :
+		param->simple_slave.slave_id = port ? pdata->dma_slave_rx_b :
 			pdata->dma_slave_rx_a;
 	}
 
-- 
1.7.2.5


^ permalink raw reply related	[flat|nested] 33+ messages in thread

* [PATCH 7/7 v2] dma: shdma: convert to the simple DMA library
  2012-01-26 14:56 [PATCH 0/7 v2] extract a simple dmaengine library from shdma.c Guennadi Liakhovetski
                   ` (5 preceding siblings ...)
  2012-01-26 14:56 ` [PATCH 6/7 v2] ASoC: SIU: " Guennadi Liakhovetski
@ 2012-01-26 14:56 ` Guennadi Liakhovetski
  6 siblings, 0 replies; 33+ messages in thread
From: Guennadi Liakhovetski @ 2012-01-26 14:56 UTC (permalink / raw)
  To: linux-kernel
  Cc: linux-sh, Vinod Koul, Magnus Damm, Yoshihiro Shimoda, linux-mmc,
	alsa-devel, linux-serial, Paul Mundt

The simple DMA library has originally been extracted from the shdma driver,
which now can be converted to actually use it.

Signed-off-by: Guennadi Liakhovetski <g.liakhovetski@gmx.de>
---

v2: update IRQ handling to match v2 of the simple DMA library

 drivers/dma/Kconfig    |    1 +
 drivers/dma/shdma.c    | 1138 ++++++++++++------------------------------------
 drivers/dma/shdma.h    |   45 +-
 include/linux/sh_dma.h |   39 +--
 4 files changed, 308 insertions(+), 915 deletions(-)

diff --git a/drivers/dma/Kconfig b/drivers/dma/Kconfig
index f7c583e..2696552 100644
--- a/drivers/dma/Kconfig
+++ b/drivers/dma/Kconfig
@@ -157,6 +157,7 @@ config SH_DMAE
 	depends on (SUPERH && SH_DMA) || (ARM && ARCH_SHMOBILE)
 	depends on !SH_DMA_API
 	select DMA_ENGINE
+	select DMA_SIMPLE
 	help
 	  Enable support for the Renesas SuperH DMA controllers.
 
diff --git a/drivers/dma/shdma.c b/drivers/dma/shdma.c
index 812fd76..08011e8 100644
--- a/drivers/dma/shdma.c
+++ b/drivers/dma/shdma.c
@@ -3,6 +3,7 @@
  *
  * base is drivers/dma/flsdma.c
  *
+ * Copyright (C) 2011-2012 Guennadi Liakhovetski <g.liakhovetski@gmx.de>
  * Copyright (C) 2009 Nobuhiro Iwamatsu <iwamatsu.nobuhiro@renesas.com>
  * Copyright (C) 2009 Renesas Solutions, Inc. All rights reserved.
  * Copyright (C) 2007 Freescale Semiconductor, Inc. All rights reserved.
@@ -32,18 +33,12 @@
 #include <linux/rculist.h>
 #include "shdma.h"
 
-/* DMA descriptor control */
-enum sh_dmae_desc_status {
-	DESC_IDLE,
-	DESC_PREPARED,
-	DESC_SUBMITTED,
-	DESC_COMPLETED,	/* completed, have to call callback */
-	DESC_WAITING,	/* callback called, waiting for ack / re-submit */
-};
+#define SH_DMAE_DRV_NAME "sh-dma-engine"
 
-#define NR_DESCS_PER_CHANNEL 32
 /* Default MEMCPY transfer size = 2^2 = 4 bytes */
 #define LOG2_DEFAULT_XFER_SIZE	2
+#define SH_DMA_SLAVE_NUMBER 256
+#define SH_DMA_TCR_MAX (16 * 1024 * 1024 - 1)
 
 /*
  * Used for write-side mutual exclusion for the global device list,
@@ -52,18 +47,12 @@ enum sh_dmae_desc_status {
 static DEFINE_SPINLOCK(sh_dmae_lock);
 static LIST_HEAD(sh_dmae_devices);
 
-/* A bitmask with bits enough for enum sh_dmae_slave_chan_id */
-static unsigned long sh_dmae_slave_used[BITS_TO_LONGS(SH_DMA_SLAVE_NUMBER)];
-
-static void sh_dmae_chan_ld_cleanup(struct sh_dmae_chan *sh_chan, bool all);
-static void sh_chan_xfer_ld_queue(struct sh_dmae_chan *sh_chan);
-
 static void chclr_write(struct sh_dmae_chan *sh_dc, u32 data)
 {
 	struct sh_dmae_device *shdev = to_sh_dev(sh_dc);
 
 	__raw_writel(data, shdev->chan_reg +
-		     shdev->pdata->channel[sh_dc->id].chclr_offset);
+		     shdev->pdata->channel[sh_dc->simple_chan.id].chclr_offset);
 }
 
 static void sh_dmae_writel(struct sh_dmae_chan *sh_dc, u32 data, u32 reg)
@@ -153,11 +142,11 @@ static int sh_dmae_rst(struct sh_dmae_device *shdev)
 	spin_unlock_irqrestore(&sh_dmae_lock, flags);
 
 	if (dmaor & (DMAOR_AE | DMAOR_NMIF)) {
-		dev_warn(shdev->common.dev, "Can't initialize DMAOR.\n");
+		dev_warn(shdev->simple_dev.dma_dev.dev, "Can't initialize DMAOR.\n");
 		return -EIO;
 	}
 	if (shdev->pdata->dmaor_init & ~dmaor)
-		dev_warn(shdev->common.dev,
+		dev_warn(shdev->simple_dev.dma_dev.dev,
 			 "DMAOR=0x%x hasn't latched the initial value 0x%x.\n",
 			 dmaor, shdev->pdata->dmaor_init);
 	return 0;
@@ -222,15 +211,6 @@ static void dmae_start(struct sh_dmae_chan *sh_chan)
 	chcr_write(sh_chan, chcr & ~CHCR_TE);
 }
 
-static void dmae_halt(struct sh_dmae_chan *sh_chan)
-{
-	struct sh_dmae_device *shdev = to_sh_dev(sh_chan);
-	u32 chcr = chcr_read(sh_chan);
-
-	chcr &= ~(CHCR_DE | CHCR_TE | shdev->chcr_ie_bit);
-	chcr_write(sh_chan, chcr);
-}
-
 static void dmae_init(struct sh_dmae_chan *sh_chan)
 {
 	/*
@@ -259,7 +239,7 @@ static int dmae_set_dmars(struct sh_dmae_chan *sh_chan, u16 val)
 {
 	struct sh_dmae_device *shdev = to_sh_dev(sh_chan);
 	struct sh_dmae_pdata *pdata = shdev->pdata;
-	const struct sh_dmae_channel *chan_pdata = &pdata->channel[sh_chan->id];
+	const struct sh_dmae_channel *chan_pdata = &pdata->channel[sh_chan->simple_chan.id];
 	u16 __iomem *addr = shdev->dmars;
 	unsigned int shift = chan_pdata->dmars_bit;
 
@@ -280,720 +260,142 @@ static int dmae_set_dmars(struct sh_dmae_chan *sh_chan, u16 val)
 	return 0;
 }
 
-static dma_cookie_t sh_dmae_tx_submit(struct dma_async_tx_descriptor *tx)
+static void sh_dmae_start_xfer(struct dma_simple_chan *schan,
+			       struct dma_simple_desc *sdesc)
 {
-	struct sh_desc *desc = tx_to_sh_desc(tx), *chunk, *last = desc, *c;
-	struct sh_dmae_chan *sh_chan = to_sh_chan(tx->chan);
-	struct sh_dmae_slave *param = tx->chan->private;
-	dma_async_tx_callback callback = tx->callback;
-	dma_cookie_t cookie;
-	bool power_up;
-
-	spin_lock_irq(&sh_chan->desc_lock);
-
-	if (list_empty(&sh_chan->ld_queue))
-		power_up = true;
-	else
-		power_up = false;
-
-	cookie = sh_chan->common.cookie;
-	cookie++;
-	if (cookie < 0)
-		cookie = 1;
-
-	sh_chan->common.cookie = cookie;
-	tx->cookie = cookie;
-
-	/* Mark all chunks of this descriptor as submitted, move to the queue */
-	list_for_each_entry_safe(chunk, c, desc->node.prev, node) {
-		/*
-		 * All chunks are on the global ld_free, so, we have to find
-		 * the end of the chain ourselves
-		 */
-		if (chunk != desc && (chunk->mark == DESC_IDLE ||
-				      chunk->async_tx.cookie > 0 ||
-				      chunk->async_tx.cookie == -EBUSY ||
-				      &chunk->node == &sh_chan->ld_free))
-			break;
-		chunk->mark = DESC_SUBMITTED;
-		/* Callback goes to the last chunk */
-		chunk->async_tx.callback = NULL;
-		chunk->cookie = cookie;
-		list_move_tail(&chunk->node, &sh_chan->ld_queue);
-		last = chunk;
-	}
-
-	last->async_tx.callback = callback;
-	last->async_tx.callback_param = tx->callback_param;
-
-	dev_dbg(sh_chan->dev, "submit #%d@%p on %d: %x[%d] -> %x\n",
-		tx->cookie, &last->async_tx, sh_chan->id,
-		desc->hw.sar, desc->hw.tcr, desc->hw.dar);
-
-	if (power_up) {
-		sh_chan->pm_state = DMAE_PM_BUSY;
-
-		pm_runtime_get(sh_chan->dev);
-
-		spin_unlock_irq(&sh_chan->desc_lock);
-
-		pm_runtime_barrier(sh_chan->dev);
-
-		spin_lock_irq(&sh_chan->desc_lock);
-
-		/* Have we been reset, while waiting? */
-		if (sh_chan->pm_state != DMAE_PM_ESTABLISHED) {
-			dev_dbg(sh_chan->dev, "Bring up channel %d\n",
-				sh_chan->id);
-			if (param) {
-				const struct sh_dmae_slave_config *cfg =
-					param->config;
-
-				dmae_set_dmars(sh_chan, cfg->mid_rid);
-				dmae_set_chcr(sh_chan, cfg->chcr);
-			} else {
-				dmae_init(sh_chan);
-			}
-
-			if (sh_chan->pm_state == DMAE_PM_PENDING)
-				sh_chan_xfer_ld_queue(sh_chan);
-			sh_chan->pm_state = DMAE_PM_ESTABLISHED;
-		}
-	} else {
-		sh_chan->pm_state = DMAE_PM_PENDING;
-	}
-
-	spin_unlock_irq(&sh_chan->desc_lock);
-
-	return cookie;
+	struct sh_dmae_chan *sh_chan = container_of(schan, struct sh_dmae_chan,
+						    simple_chan);
+	struct sh_dmae_desc *sh_desc = container_of(sdesc,
+					struct sh_dmae_desc, simple_desc);
+	dev_dbg(sh_chan->simple_chan.dev, "Queue #%d to %d: %u@%x -> %x\n",
+		sdesc->async_tx.cookie, sh_chan->simple_chan.id,
+		sh_desc->hw.tcr, sh_desc->hw.sar, sh_desc->hw.dar);
+	/* Get the ld start address from ld_queue */
+	dmae_set_reg(sh_chan, &sh_desc->hw);
+	dmae_start(sh_chan);
 }
 
-/* Called with desc_lock held */
-static struct sh_desc *sh_dmae_get_desc(struct sh_dmae_chan *sh_chan)
+static bool sh_dmae_channel_busy(struct dma_simple_chan *schan)
 {
-	struct sh_desc *desc;
-
-	list_for_each_entry(desc, &sh_chan->ld_free, node)
-		if (desc->mark != DESC_PREPARED) {
-			BUG_ON(desc->mark != DESC_IDLE);
-			list_del(&desc->node);
-			return desc;
-		}
-
-	return NULL;
+	struct sh_dmae_chan *sh_chan = container_of(schan, struct sh_dmae_chan,
+						    simple_chan);
+	return dmae_is_busy(sh_chan);
 }
 
-static const struct sh_dmae_slave_config *sh_dmae_find_slave(
-	struct sh_dmae_chan *sh_chan, struct sh_dmae_slave *param)
+static void sh_dmae_setup_xfer(struct dma_simple_chan *schan,
+			       struct dma_simple_slave *sslave)
 {
-	struct sh_dmae_device *shdev = to_sh_dev(sh_chan);
-	struct sh_dmae_pdata *pdata = shdev->pdata;
-	int i;
+	struct sh_dmae_chan *sh_chan = container_of(schan, struct sh_dmae_chan,
+						    simple_chan);
 
-	if (param->slave_id >= SH_DMA_SLAVE_NUMBER)
-		return NULL;
-
-	for (i = 0; i < pdata->slave_num; i++)
-		if (pdata->slave[i].slave_id == param->slave_id)
-			return pdata->slave + i;
-
-	return NULL;
-}
+	if (sslave) {
+		struct sh_dmae_slave *slave = container_of(sslave,
+					struct sh_dmae_slave, simple_slave);
+		const struct sh_dmae_slave_config *cfg =
+			slave->config;
 
-static int sh_dmae_alloc_chan_resources(struct dma_chan *chan)
-{
-	struct sh_dmae_chan *sh_chan = to_sh_chan(chan);
-	struct sh_desc *desc;
-	struct sh_dmae_slave *param = chan->private;
-	int ret;
-
-	/*
-	 * This relies on the guarantee from dmaengine that alloc_chan_resources
-	 * never runs concurrently with itself or free_chan_resources.
-	 */
-	if (param) {
-		const struct sh_dmae_slave_config *cfg;
-
-		cfg = sh_dmae_find_slave(sh_chan, param);
-		if (!cfg) {
-			ret = -EINVAL;
-			goto efindslave;
-		}
-
-		if (test_and_set_bit(param->slave_id, sh_dmae_slave_used)) {
-			ret = -EBUSY;
-			goto etestused;
-		}
-
-		param->config = cfg;
-	}
-
-	while (sh_chan->descs_allocated < NR_DESCS_PER_CHANNEL) {
-		desc = kzalloc(sizeof(struct sh_desc), GFP_KERNEL);
-		if (!desc)
-			break;
-		dma_async_tx_descriptor_init(&desc->async_tx,
-					&sh_chan->common);
-		desc->async_tx.tx_submit = sh_dmae_tx_submit;
-		desc->mark = DESC_IDLE;
-
-		list_add(&desc->node, &sh_chan->ld_free);
-		sh_chan->descs_allocated++;
-	}
-
-	if (!sh_chan->descs_allocated) {
-		ret = -ENOMEM;
-		goto edescalloc;
-	}
-
-	return sh_chan->descs_allocated;
-
-edescalloc:
-	if (param)
-		clear_bit(param->slave_id, sh_dmae_slave_used);
-etestused:
-efindslave:
-	chan->private = NULL;
-	return ret;
-}
-
-/*
- * sh_dma_free_chan_resources - Free all resources of the channel.
- */
-static void sh_dmae_free_chan_resources(struct dma_chan *chan)
-{
-	struct sh_dmae_chan *sh_chan = to_sh_chan(chan);
-	struct sh_desc *desc, *_desc;
-	LIST_HEAD(list);
-
-	/* Protect against ISR */
-	spin_lock_irq(&sh_chan->desc_lock);
-	dmae_halt(sh_chan);
-	spin_unlock_irq(&sh_chan->desc_lock);
-
-	/* Now no new interrupts will occur */
-
-	/* Prepared and not submitted descriptors can still be on the queue */
-	if (!list_empty(&sh_chan->ld_queue))
-		sh_dmae_chan_ld_cleanup(sh_chan, true);
-
-	if (chan->private) {
-		/* The caller is holding dma_list_mutex */
-		struct sh_dmae_slave *param = chan->private;
-		clear_bit(param->slave_id, sh_dmae_slave_used);
-		chan->private = NULL;
-	}
-
-	spin_lock_irq(&sh_chan->desc_lock);
-
-	list_splice_init(&sh_chan->ld_free, &list);
-	sh_chan->descs_allocated = 0;
-
-	spin_unlock_irq(&sh_chan->desc_lock);
-
-	list_for_each_entry_safe(desc, _desc, &list, node)
-		kfree(desc);
-}
-
-/**
- * sh_dmae_add_desc - get, set up and return one transfer descriptor
- * @sh_chan:	DMA channel
- * @flags:	DMA transfer flags
- * @dest:	destination DMA address, incremented when direction equals
- *		DMA_DEV_TO_MEM
- * @src:	source DMA address, incremented when direction equals
- *		DMA_MEM_TO_DEV
- * @len:	DMA transfer length
- * @first:	if NULL, set to the current descriptor and cookie set to -EBUSY
- * @direction:	needed for slave DMA to decide which address to keep constant,
- *		equals DMA_MEM_TO_MEM for MEMCPY
- * Returns 0 or an error
- * Locks: called with desc_lock held
- */
-static struct sh_desc *sh_dmae_add_desc(struct sh_dmae_chan *sh_chan,
-	unsigned long flags, dma_addr_t *dest, dma_addr_t *src, size_t *len,
-	struct sh_desc **first, enum dma_transfer_direction direction)
-{
-	struct sh_desc *new;
-	size_t copy_size;
-
-	if (!*len)
-		return NULL;
-
-	/* Allocate the link descriptor from the free list */
-	new = sh_dmae_get_desc(sh_chan);
-	if (!new) {
-		dev_err(sh_chan->dev, "No free link descriptor available\n");
-		return NULL;
-	}
-
-	copy_size = min(*len, (size_t)SH_DMA_TCR_MAX + 1);
-
-	new->hw.sar = *src;
-	new->hw.dar = *dest;
-	new->hw.tcr = copy_size;
-
-	if (!*first) {
-		/* First desc */
-		new->async_tx.cookie = -EBUSY;
-		*first = new;
+		dmae_set_dmars(sh_chan, cfg->mid_rid);
+		dmae_set_chcr(sh_chan, cfg->chcr);
 	} else {
-		/* Other desc - invisible to the user */
-		new->async_tx.cookie = -EINVAL;
+		dmae_init(sh_chan);
 	}
-
-	dev_dbg(sh_chan->dev,
-		"chaining (%u/%u)@%x -> %x with %p, cookie %d, shift %d\n",
-		copy_size, *len, *src, *dest, &new->async_tx,
-		new->async_tx.cookie, sh_chan->xmit_shift);
-
-	new->mark = DESC_PREPARED;
-	new->async_tx.flags = flags;
-	new->direction = direction;
-
-	*len -= copy_size;
-	if (direction == DMA_MEM_TO_MEM || direction == DMA_MEM_TO_DEV)
-		*src += copy_size;
-	if (direction == DMA_MEM_TO_MEM || direction == DMA_DEV_TO_MEM)
-		*dest += copy_size;
-
-	return new;
 }
 
-/*
- * sh_dmae_prep_sg - prepare transfer descriptors from an SG list
- *
- * Common routine for public (MEMCPY) and slave DMA. The MEMCPY case is also
- * converted to scatter-gather to guarantee consistent locking and a correct
- * list manipulation. For slave DMA direction carries the usual meaning, and,
- * logically, the SG list is RAM and the addr variable contains slave address,
- * e.g., the FIFO I/O register. For MEMCPY direction equals DMA_MEM_TO_MEM
- * and the SG list contains only one element and points at the source buffer.
- */
-static struct dma_async_tx_descriptor *sh_dmae_prep_sg(struct sh_dmae_chan *sh_chan,
-	struct scatterlist *sgl, unsigned int sg_len, dma_addr_t *addr,
-	enum dma_transfer_direction direction, unsigned long flags)
+static const struct sh_dmae_slave_config *dmae_find_slave(
+	struct sh_dmae_chan *sh_chan, struct sh_dmae_slave *slave)
 {
-	struct scatterlist *sg;
-	struct sh_desc *first = NULL, *new = NULL /* compiler... */;
-	LIST_HEAD(tx_list);
-	int chunks = 0;
-	unsigned long irq_flags;
+	struct sh_dmae_device *shdev = to_sh_dev(sh_chan);
+	struct sh_dmae_pdata *pdata = shdev->pdata;
+	const struct sh_dmae_slave_config *cfg;
 	int i;
 
-	if (!sg_len)
+	if (slave->simple_slave.slave_id >= SH_DMA_SLAVE_NUMBER)
 		return NULL;
 
-	for_each_sg(sgl, sg, sg_len, i)
-		chunks += (sg_dma_len(sg) + SH_DMA_TCR_MAX) /
-			(SH_DMA_TCR_MAX + 1);
-
-	/* Have to lock the whole loop to protect against concurrent release */
-	spin_lock_irqsave(&sh_chan->desc_lock, irq_flags);
-
-	/*
-	 * Chaining:
-	 * first descriptor is what user is dealing with in all API calls, its
-	 *	cookie is at first set to -EBUSY, at tx-submit to a positive
-	 *	number
-	 * if more than one chunk is needed further chunks have cookie = -EINVAL
-	 * the last chunk, if not equal to the first, has cookie = -ENOSPC
-	 * all chunks are linked onto the tx_list head with their .node heads
-	 *	only during this function, then they are immediately spliced
-	 *	back onto the free list in form of a chain
-	 */
-	for_each_sg(sgl, sg, sg_len, i) {
-		dma_addr_t sg_addr = sg_dma_address(sg);
-		size_t len = sg_dma_len(sg);
-
-		if (!len)
-			goto err_get_desc;
-
-		do {
-			dev_dbg(sh_chan->dev, "Add SG #%d@%p[%d], dma %llx\n",
-				i, sg, len, (unsigned long long)sg_addr);
-
-			if (direction == DMA_DEV_TO_MEM)
-				new = sh_dmae_add_desc(sh_chan, flags,
-						&sg_addr, addr, &len, &first,
-						direction);
-			else
-				new = sh_dmae_add_desc(sh_chan, flags,
-						addr, &sg_addr, &len, &first,
-						direction);
-			if (!new)
-				goto err_get_desc;
-
-			new->chunks = chunks--;
-			list_add_tail(&new->node, &tx_list);
-		} while (len);
-	}
-
-	if (new != first)
-		new->async_tx.cookie = -ENOSPC;
-
-	/* Put them back on the free list, so, they don't get lost */
-	list_splice_tail(&tx_list, &sh_chan->ld_free);
-
-	spin_unlock_irqrestore(&sh_chan->desc_lock, irq_flags);
-
-	return &first->async_tx;
-
-err_get_desc:
-	list_for_each_entry(new, &tx_list, node)
-		new->mark = DESC_IDLE;
-	list_splice(&tx_list, &sh_chan->ld_free);
-
-	spin_unlock_irqrestore(&sh_chan->desc_lock, irq_flags);
+	for (i = 0, cfg = pdata->slave; i < pdata->slave_num; i++, cfg++)
+		if (cfg->slave_id == slave->simple_slave.slave_id)
+			return cfg;
 
 	return NULL;
 }
 
-static struct dma_async_tx_descriptor *sh_dmae_prep_memcpy(
-	struct dma_chan *chan, dma_addr_t dma_dest, dma_addr_t dma_src,
-	size_t len, unsigned long flags)
-{
-	struct sh_dmae_chan *sh_chan;
-	struct scatterlist sg;
-
-	if (!chan || !len)
-		return NULL;
-
-	sh_chan = to_sh_chan(chan);
-
-	sg_init_table(&sg, 1);
-	sg_set_page(&sg, pfn_to_page(PFN_DOWN(dma_src)), len,
-		    offset_in_page(dma_src));
-	sg_dma_address(&sg) = dma_src;
-	sg_dma_len(&sg) = len;
-
-	return sh_dmae_prep_sg(sh_chan, &sg, 1, &dma_dest, DMA_MEM_TO_MEM,
-			       flags);
-}
-
-static struct dma_async_tx_descriptor *sh_dmae_prep_slave_sg(
-	struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len,
-	enum dma_transfer_direction direction, unsigned long flags)
+static int sh_dmae_set_slave(struct dma_simple_chan *schan,
+			     struct dma_simple_slave *sslave)
 {
-	struct sh_dmae_slave *param;
-	struct sh_dmae_chan *sh_chan;
-	dma_addr_t slave_addr;
-
-	if (!chan)
-		return NULL;
-
-	sh_chan = to_sh_chan(chan);
-	param = chan->private;
-
-	/* Someone calling slave DMA on a public channel? */
-	if (!param || !sg_len) {
-		dev_warn(sh_chan->dev, "%s: bad parameter: %p, %d, %d\n",
-			 __func__, param, sg_len, param ? param->slave_id : -1);
-		return NULL;
-	}
-
-	slave_addr = param->config->addr;
-
-	/*
-	 * if (param != NULL), this is a successfully requested slave channel,
-	 * therefore param->config != NULL too.
-	 */
-	return sh_dmae_prep_sg(sh_chan, sgl, sg_len, &slave_addr,
-			       direction, flags);
-}
-
-static int sh_dmae_control(struct dma_chan *chan, enum dma_ctrl_cmd cmd,
-			   unsigned long arg)
-{
-	struct sh_dmae_chan *sh_chan = to_sh_chan(chan);
-	unsigned long flags;
-
-	/* Only supports DMA_TERMINATE_ALL */
-	if (cmd != DMA_TERMINATE_ALL)
-		return -ENXIO;
-
-	if (!chan)
-		return -EINVAL;
-
-	spin_lock_irqsave(&sh_chan->desc_lock, flags);
-	dmae_halt(sh_chan);
-
-	if (!list_empty(&sh_chan->ld_queue)) {
-		/* Record partial transfer */
-		struct sh_desc *desc = list_entry(sh_chan->ld_queue.next,
-						  struct sh_desc, node);
-		desc->partial = (desc->hw.tcr - sh_dmae_readl(sh_chan, TCR)) <<
-			sh_chan->xmit_shift;
-	}
-	spin_unlock_irqrestore(&sh_chan->desc_lock, flags);
+	struct sh_dmae_chan *sh_chan = container_of(schan, struct sh_dmae_chan,
+						    simple_chan);
+	struct sh_dmae_slave *slave = container_of(sslave, struct sh_dmae_slave,
+						   simple_slave);
+	const struct sh_dmae_slave_config *cfg = dmae_find_slave(sh_chan, slave);
+	if (!cfg)
+		return -ENODEV;
 
-	sh_dmae_chan_ld_cleanup(sh_chan, true);
+	slave->config = cfg;
 
 	return 0;
 }
 
-static dma_async_tx_callback __ld_cleanup(struct sh_dmae_chan *sh_chan, bool all)
+static void dmae_halt(struct sh_dmae_chan *sh_chan)
 {
-	struct sh_desc *desc, *_desc;
-	/* Is the "exposed" head of a chain acked? */
-	bool head_acked = false;
-	dma_cookie_t cookie = 0;
-	dma_async_tx_callback callback = NULL;
-	void *param = NULL;
-	unsigned long flags;
-
-	spin_lock_irqsave(&sh_chan->desc_lock, flags);
-	list_for_each_entry_safe(desc, _desc, &sh_chan->ld_queue, node) {
-		struct dma_async_tx_descriptor *tx = &desc->async_tx;
-
-		BUG_ON(tx->cookie > 0 && tx->cookie != desc->cookie);
-		BUG_ON(desc->mark != DESC_SUBMITTED &&
-		       desc->mark != DESC_COMPLETED &&
-		       desc->mark != DESC_WAITING);
-
-		/*
-		 * queue is ordered, and we use this loop to (1) clean up all
-		 * completed descriptors, and to (2) update descriptor flags of
-		 * any chunks in a (partially) completed chain
-		 */
-		if (!all && desc->mark == DESC_SUBMITTED &&
-		    desc->cookie != cookie)
-			break;
-
-		if (tx->cookie > 0)
-			cookie = tx->cookie;
-
-		if (desc->mark == DESC_COMPLETED && desc->chunks == 1) {
-			if (sh_chan->completed_cookie != desc->cookie - 1)
-				dev_dbg(sh_chan->dev,
-					"Completing cookie %d, expected %d\n",
-					desc->cookie,
-					sh_chan->completed_cookie + 1);
-			sh_chan->completed_cookie = desc->cookie;
-		}
-
-		/* Call callback on the last chunk */
-		if (desc->mark == DESC_COMPLETED && tx->callback) {
-			desc->mark = DESC_WAITING;
-			callback = tx->callback;
-			param = tx->callback_param;
-			dev_dbg(sh_chan->dev, "descriptor #%d@%p on %d callback\n",
-				tx->cookie, tx, sh_chan->id);
-			BUG_ON(desc->chunks != 1);
-			break;
-		}
-
-		if (tx->cookie > 0 || tx->cookie == -EBUSY) {
-			if (desc->mark == DESC_COMPLETED) {
-				BUG_ON(tx->cookie < 0);
-				desc->mark = DESC_WAITING;
-			}
-			head_acked = async_tx_test_ack(tx);
-		} else {
-			switch (desc->mark) {
-			case DESC_COMPLETED:
-				desc->mark = DESC_WAITING;
-				/* Fall through */
-			case DESC_WAITING:
-				if (head_acked)
-					async_tx_ack(&desc->async_tx);
-			}
-		}
-
-		dev_dbg(sh_chan->dev, "descriptor %p #%d completed.\n",
-			tx, tx->cookie);
-
-		if (((desc->mark == DESC_COMPLETED ||
-		      desc->mark == DESC_WAITING) &&
-		     async_tx_test_ack(&desc->async_tx)) || all) {
-			/* Remove from ld_queue list */
-			desc->mark = DESC_IDLE;
-
-			list_move(&desc->node, &sh_chan->ld_free);
-
-			if (list_empty(&sh_chan->ld_queue)) {
-				dev_dbg(sh_chan->dev, "Bring down channel %d\n", sh_chan->id);
-				pm_runtime_put(sh_chan->dev);
-			}
-		}
-	}
-
-	if (all && !callback)
-		/*
-		 * Terminating and the loop completed normally: forgive
-		 * uncompleted cookies
-		 */
-		sh_chan->completed_cookie = sh_chan->common.cookie;
-
-	spin_unlock_irqrestore(&sh_chan->desc_lock, flags);
-
-	if (callback)
-		callback(param);
+	struct sh_dmae_device *shdev = to_sh_dev(sh_chan);
+	u32 chcr = chcr_read(sh_chan);
 
-	return callback;
+	chcr &= ~(CHCR_DE | CHCR_TE | shdev->chcr_ie_bit);
+	chcr_write(sh_chan, chcr);
 }
 
-/*
- * sh_chan_ld_cleanup - Clean up link descriptors
- *
- * This function cleans up the ld_queue of DMA channel.
- */
-static void sh_dmae_chan_ld_cleanup(struct sh_dmae_chan *sh_chan, bool all)
+static int sh_dmae_desc_setup(struct dma_simple_chan *schan,
+			      struct dma_simple_desc *sdesc,
+			      dma_addr_t src, dma_addr_t dst, size_t *len)
 {
-	while (__ld_cleanup(sh_chan, all))
-		;
-}
+	struct sh_dmae_desc *sh_desc = container_of(sdesc,
+					struct sh_dmae_desc, simple_desc);
 
-/* Called under spin_lock_irq(&sh_chan->desc_lock) */
-static void sh_chan_xfer_ld_queue(struct sh_dmae_chan *sh_chan)
-{
-	struct sh_desc *desc;
+	if (*len > schan->max_xfer_len)
+		*len = schan->max_xfer_len;
 
-	/* DMA work check */
-	if (dmae_is_busy(sh_chan))
-		return;
-
-	/* Find the first not transferred descriptor */
-	list_for_each_entry(desc, &sh_chan->ld_queue, node)
-		if (desc->mark == DESC_SUBMITTED) {
-			dev_dbg(sh_chan->dev, "Queue #%d to %d: %u@%x -> %x\n",
-				desc->async_tx.cookie, sh_chan->id,
-				desc->hw.tcr, desc->hw.sar, desc->hw.dar);
-			/* Get the ld start address from ld_queue */
-			dmae_set_reg(sh_chan, &desc->hw);
-			dmae_start(sh_chan);
-			break;
-		}
-}
+	sh_desc->hw.sar = src;
+	sh_desc->hw.dar = dst;
+	sh_desc->hw.tcr = *len;
 
-static void sh_dmae_memcpy_issue_pending(struct dma_chan *chan)
-{
-	struct sh_dmae_chan *sh_chan = to_sh_chan(chan);
-
-	spin_lock_irq(&sh_chan->desc_lock);
-	if (sh_chan->pm_state == DMAE_PM_ESTABLISHED)
-		sh_chan_xfer_ld_queue(sh_chan);
-	else
-		sh_chan->pm_state = DMAE_PM_PENDING;
-	spin_unlock_irq(&sh_chan->desc_lock);
+	return 0;
 }
 
-static enum dma_status sh_dmae_tx_status(struct dma_chan *chan,
-					dma_cookie_t cookie,
-					struct dma_tx_state *txstate)
+static void sh_dmae_halt(struct dma_simple_chan *schan)
 {
-	struct sh_dmae_chan *sh_chan = to_sh_chan(chan);
-	dma_cookie_t last_used;
-	dma_cookie_t last_complete;
-	enum dma_status status;
-	unsigned long flags;
-
-	sh_dmae_chan_ld_cleanup(sh_chan, false);
-
-	/* First read completed cookie to avoid a skew */
-	last_complete = sh_chan->completed_cookie;
-	rmb();
-	last_used = chan->cookie;
-	BUG_ON(last_complete < 0);
-	dma_set_tx_state(txstate, last_complete, last_used, 0);
-
-	spin_lock_irqsave(&sh_chan->desc_lock, flags);
-
-	status = dma_async_is_complete(cookie, last_complete, last_used);
-
-	/*
-	 * If we don't find cookie on the queue, it has been aborted and we have
-	 * to report error
-	 */
-	if (status != DMA_SUCCESS) {
-		struct sh_desc *desc;
-		status = DMA_ERROR;
-		list_for_each_entry(desc, &sh_chan->ld_queue, node)
-			if (desc->cookie == cookie) {
-				status = DMA_IN_PROGRESS;
-				break;
-			}
-	}
-
-	spin_unlock_irqrestore(&sh_chan->desc_lock, flags);
-
-	return status;
+	struct sh_dmae_chan *sh_chan = container_of(schan, struct sh_dmae_chan,
+						    simple_chan);
+	dmae_halt(sh_chan);
 }
 
-static irqreturn_t sh_dmae_interrupt(int irq, void *data)
+static bool sh_dmae_chan_irq(struct dma_simple_chan *schan, int irq)
 {
-	irqreturn_t ret = IRQ_NONE;
-	struct sh_dmae_chan *sh_chan = data;
-	u32 chcr;
-
-	spin_lock(&sh_chan->desc_lock);
+	struct sh_dmae_chan *sh_chan = container_of(schan, struct sh_dmae_chan,
+						    simple_chan);
 
-	chcr = chcr_read(sh_chan);
-
-	if (chcr & CHCR_TE) {
-		/* DMA stop */
-		dmae_halt(sh_chan);
-
-		ret = IRQ_HANDLED;
-		tasklet_schedule(&sh_chan->tasklet);
-	}
+	if (!(chcr_read(sh_chan) & CHCR_TE))
+		return false;
 
-	spin_unlock(&sh_chan->desc_lock);
+	/* DMA stop */
+	dmae_halt(sh_chan);
 
-	return ret;
+	return true;
 }
 
 /* Called from error IRQ or NMI */
 static bool sh_dmae_reset(struct sh_dmae_device *shdev)
 {
-	unsigned int handled = 0;
-	int i;
+	bool ret;
 
 	/* halt the dma controller */
 	sh_dmae_ctl_stop(shdev);
 
 	/* We cannot detect, which channel caused the error, have to reset all */
-	for (i = 0; i < SH_DMAC_MAX_CHANNELS; i++) {
-		struct sh_dmae_chan *sh_chan = shdev->chan[i];
-		struct sh_desc *desc;
-		LIST_HEAD(dl);
-
-		if (!sh_chan)
-			continue;
-
-		spin_lock(&sh_chan->desc_lock);
-
-		/* Stop the channel */
-		dmae_halt(sh_chan);
-
-		list_splice_init(&sh_chan->ld_queue, &dl);
-
-		if (!list_empty(&dl)) {
-			dev_dbg(sh_chan->dev, "Bring down channel %d\n", sh_chan->id);
-			pm_runtime_put(sh_chan->dev);
-		}
-		sh_chan->pm_state = DMAE_PM_ESTABLISHED;
-
-		spin_unlock(&sh_chan->desc_lock);
-
-		/* Complete all  */
-		list_for_each_entry(desc, &dl, node) {
-			struct dma_async_tx_descriptor *tx = &desc->async_tx;
-			desc->mark = DESC_IDLE;
-			if (tx->callback)
-				tx->callback(tx->callback_param);
-		}
-
-		spin_lock(&sh_chan->desc_lock);
-		list_splice(&dl, &sh_chan->ld_free);
-		spin_unlock(&sh_chan->desc_lock);
-
-		handled++;
-	}
+	ret = dma_simple_reset(&shdev->simple_dev);
 
 	sh_dmae_rst(shdev);
 
-	return !!handled;
+	return ret;
 }
 
 static irqreturn_t sh_dmae_err(int irq, void *data)
@@ -1003,35 +405,24 @@ static irqreturn_t sh_dmae_err(int irq, void *data)
 	if (!(dmaor_read(shdev) & DMAOR_AE))
 		return IRQ_NONE;
 
-	sh_dmae_reset(data);
+	sh_dmae_reset(shdev);
 	return IRQ_HANDLED;
 }
 
-static void dmae_do_tasklet(unsigned long data)
+static bool sh_dmae_desc_completed(struct dma_simple_chan *schan,
+				   struct dma_simple_desc *sdesc)
 {
-	struct sh_dmae_chan *sh_chan = (struct sh_dmae_chan *)data;
-	struct sh_desc *desc;
+	struct sh_dmae_chan *sh_chan = container_of(schan,
+					struct sh_dmae_chan, simple_chan);
+	struct sh_dmae_desc *sh_desc = container_of(sdesc,
+					struct sh_dmae_desc, simple_desc);
 	u32 sar_buf = sh_dmae_readl(sh_chan, SAR);
 	u32 dar_buf = sh_dmae_readl(sh_chan, DAR);
 
-	spin_lock_irq(&sh_chan->desc_lock);
-	list_for_each_entry(desc, &sh_chan->ld_queue, node) {
-		if (desc->mark == DESC_SUBMITTED &&
-		    ((desc->direction == DMA_DEV_TO_MEM &&
-		      (desc->hw.dar + desc->hw.tcr) == dar_buf) ||
-		     (desc->hw.sar + desc->hw.tcr) == sar_buf)) {
-			dev_dbg(sh_chan->dev, "done #%d@%p dst %u\n",
-				desc->async_tx.cookie, &desc->async_tx,
-				desc->hw.dar);
-			desc->mark = DESC_COMPLETED;
-			break;
-		}
-	}
-	/* Next desc */
-	sh_chan_xfer_ld_queue(sh_chan);
-	spin_unlock_irq(&sh_chan->desc_lock);
-
-	sh_dmae_chan_ld_cleanup(sh_chan, false);
+	return	(sdesc->direction == DMA_DEV_TO_MEM &&
+		 (sh_desc->hw.dar + sh_desc->hw.tcr) == dar_buf) ||
+		(sdesc->direction != DMA_DEV_TO_MEM &&
+		 (sh_desc->hw.sar + sh_desc->hw.tcr) == sar_buf);
 }
 
 static bool sh_dmae_nmi_notify(struct sh_dmae_device *shdev)
@@ -1085,96 +476,174 @@ static struct notifier_block sh_dmae_nmi_notifier __read_mostly = {
 static int __devinit sh_dmae_chan_probe(struct sh_dmae_device *shdev, int id,
 					int irq, unsigned long flags)
 {
-	int err;
 	const struct sh_dmae_channel *chan_pdata = &shdev->pdata->channel[id];
-	struct platform_device *pdev = to_platform_device(shdev->common.dev);
-	struct sh_dmae_chan *new_sh_chan;
+	struct dma_simple_dev *sdev = &shdev->simple_dev;
+	struct platform_device *pdev = to_platform_device(sdev->dma_dev.dev);
+	struct sh_dmae_chan *sh_chan;
+	struct dma_simple_chan *schan;
+	int err;
 
-	/* alloc channel */
-	new_sh_chan = kzalloc(sizeof(struct sh_dmae_chan), GFP_KERNEL);
-	if (!new_sh_chan) {
-		dev_err(shdev->common.dev,
+	sh_chan = kzalloc(sizeof(struct sh_dmae_chan), GFP_KERNEL);
+	if (!sh_chan) {
+		dev_err(sdev->dma_dev.dev,
 			"No free memory for allocating dma channels!\n");
 		return -ENOMEM;
 	}
 
-	new_sh_chan->pm_state = DMAE_PM_ESTABLISHED;
+	schan = &sh_chan->simple_chan;
+	schan->max_xfer_len = SH_DMA_TCR_MAX + 1;
 
-	/* reference struct dma_device */
-	new_sh_chan->common.device = &shdev->common;
+	dma_simple_chan_probe(sdev, schan, id);
 
-	new_sh_chan->dev = shdev->common.dev;
-	new_sh_chan->id = id;
-	new_sh_chan->irq = irq;
-	new_sh_chan->base = shdev->chan_reg + chan_pdata->offset / sizeof(u32);
-
-	/* Init DMA tasklet */
-	tasklet_init(&new_sh_chan->tasklet, dmae_do_tasklet,
-			(unsigned long)new_sh_chan);
-
-	spin_lock_init(&new_sh_chan->desc_lock);
-
-	/* Init descripter manage list */
-	INIT_LIST_HEAD(&new_sh_chan->ld_queue);
-	INIT_LIST_HEAD(&new_sh_chan->ld_free);
-
-	/* Add the channel to DMA device channel list */
-	list_add_tail(&new_sh_chan->common.device_node,
-			&shdev->common.channels);
-	shdev->common.chancnt++;
+	sh_chan->base = shdev->chan_reg + chan_pdata->offset / sizeof(u32);
 
+	/* set up channel irq */
 	if (pdev->id >= 0)
-		snprintf(new_sh_chan->dev_id, sizeof(new_sh_chan->dev_id),
-			 "sh-dmae%d.%d", pdev->id, new_sh_chan->id);
+		snprintf(sh_chan->dev_id, sizeof(sh_chan->dev_id),
+			 "sh-dmae%d.%d", pdev->id, id);
 	else
-		snprintf(new_sh_chan->dev_id, sizeof(new_sh_chan->dev_id),
-			 "sh-dma%d", new_sh_chan->id);
+		snprintf(sh_chan->dev_id, sizeof(sh_chan->dev_id),
+			 "sh-dma%d", id);
 
-	/* set up channel irq */
-	err = request_irq(irq, &sh_dmae_interrupt, flags,
-			  new_sh_chan->dev_id, new_sh_chan);
+	err = dma_simple_request_irq(schan, irq, flags, sh_chan->dev_id);
 	if (err) {
-		dev_err(shdev->common.dev, "DMA channel %d request_irq error "
-			"with return %d\n", id, err);
+		dev_err(sdev->dma_dev.dev,
+			"DMA channel %d request_irq error %d\n",
+			id, err);
 		goto err_no_irq;
 	}
 
-	shdev->chan[id] = new_sh_chan;
+	shdev->chan[id] = sh_chan;
 	return 0;
 
 err_no_irq:
 	/* remove from dmaengine device node */
-	list_del(&new_sh_chan->common.device_node);
-	kfree(new_sh_chan);
+	dma_simple_chan_remove(schan);
+	kfree(sh_chan);
 	return err;
 }
 
 static void sh_dmae_chan_remove(struct sh_dmae_device *shdev)
 {
+	struct dma_device *dma_dev = &shdev->simple_dev.dma_dev;
+	struct dma_simple_chan *schan;
 	int i;
 
-	for (i = shdev->common.chancnt - 1 ; i >= 0 ; i--) {
-		if (shdev->chan[i]) {
-			struct sh_dmae_chan *sh_chan = shdev->chan[i];
+	dma_simple_for_each_chan(schan, &shdev->simple_dev, i) {
+		struct sh_dmae_chan *sh_chan = container_of(schan,
+					struct sh_dmae_chan, simple_chan);
+		BUG_ON(!schan);
 
-			free_irq(sh_chan->irq, sh_chan);
+		dma_simple_free_irq(&sh_chan->simple_chan);
 
-			list_del(&sh_chan->common.device_node);
-			kfree(sh_chan);
-			shdev->chan[i] = NULL;
+		dma_simple_chan_remove(schan);
+		kfree(sh_chan);
+	}
+	dma_dev->chancnt = 0;
+}
+
+static void sh_dmae_shutdown(struct platform_device *pdev)
+{
+	struct sh_dmae_device *shdev = platform_get_drvdata(pdev);
+	sh_dmae_ctl_stop(shdev);
+}
+
+static int sh_dmae_runtime_suspend(struct device *dev)
+{
+	return 0;
+}
+
+static int sh_dmae_runtime_resume(struct device *dev)
+{
+	struct sh_dmae_device *shdev = dev_get_drvdata(dev);
+
+	return sh_dmae_rst(shdev);
+}
+
+#ifdef CONFIG_PM
+static int sh_dmae_suspend(struct device *dev)
+{
+	return 0;
+}
+
+static int sh_dmae_resume(struct device *dev)
+{
+	struct sh_dmae_device *shdev = dev_get_drvdata(dev);
+	int i, ret;
+
+	ret = sh_dmae_rst(shdev);
+	if (ret < 0)
+		dev_err(dev, "Failed to reset!\n");
+
+	for (i = 0; i < shdev->pdata->channel_num; i++) {
+		struct sh_dmae_chan *sh_chan = shdev->chan[i];
+		struct sh_dmae_slave *param = sh_chan->simple_chan.dma_chan.private;
+
+		if (!sh_chan->simple_chan.desc_num)
+			continue;
+
+		if (param) {
+			const struct sh_dmae_slave_config *cfg = param->config;
+			dmae_set_dmars(sh_chan, cfg->mid_rid);
+			dmae_set_chcr(sh_chan, cfg->chcr);
+		} else {
+			dmae_init(sh_chan);
 		}
 	}
-	shdev->common.chancnt = 0;
+
+	return 0;
+}
+#else
+#define sh_dmae_suspend NULL
+#define sh_dmae_resume NULL
+#endif
+
+const struct dev_pm_ops sh_dmae_pm = {
+	.suspend		= sh_dmae_suspend,
+	.resume			= sh_dmae_resume,
+	.runtime_suspend	= sh_dmae_runtime_suspend,
+	.runtime_resume		= sh_dmae_runtime_resume,
+};
+
+static dma_addr_t sh_dmae_slave_addr(struct dma_simple_chan *schan)
+{
+	struct sh_dmae_slave *param = schan->dma_chan.private;
+
+	/*
+	 * Implicit BUG_ON(!param)
+	 * if (param != NULL), this is a successfully requested slave channel,
+	 * therefore param->config != NULL too.
+	 */
+	return param->config->addr;
+}
+
+static struct dma_simple_desc *sh_dmae_embedded_desc(void *buf, int i)
+{
+	return &((struct sh_dmae_desc *)buf)[i].simple_desc;
 }
 
-static int __init sh_dmae_probe(struct platform_device *pdev)
+static const struct dma_simple_ops sh_dmae_simple_ops = {
+	.desc_completed = sh_dmae_desc_completed,
+	.halt_channel = sh_dmae_halt,
+	.channel_busy = sh_dmae_channel_busy,
+	.slave_addr = sh_dmae_slave_addr,
+	.desc_setup = sh_dmae_desc_setup,
+	.set_slave = sh_dmae_set_slave,
+	.setup_xfer = sh_dmae_setup_xfer,
+	.start_xfer = sh_dmae_start_xfer,
+	.embedded_desc = sh_dmae_embedded_desc,
+	.chan_irq = sh_dmae_chan_irq,
+};
+
+static int __devinit sh_dmae_probe(struct platform_device *pdev)
 {
 	struct sh_dmae_pdata *pdata = pdev->dev.platform_data;
 	unsigned long irqflags = IRQF_DISABLED,
-		chan_flag[SH_DMAC_MAX_CHANNELS] = {};
-	int errirq, chan_irq[SH_DMAC_MAX_CHANNELS];
+		chan_flag[SH_DMAE_MAX_CHANNELS] = {};
+	int errirq, chan_irq[SH_DMAE_MAX_CHANNELS];
 	int err, i, irq_cnt = 0, irqres = 0, irq_cap = 0;
 	struct sh_dmae_device *shdev;
+	struct dma_device *dma_dev;
 	struct resource *chan, *dmars, *errirq_res, *chanirq_res;
 
 	/* get platform data */
@@ -1222,6 +691,8 @@ static int __init sh_dmae_probe(struct platform_device *pdev)
 		goto ealloc;
 	}
 
+	dma_dev = &shdev->simple_dev.dma_dev;
+
 	shdev->chan_reg = ioremap(chan->start, resource_size(chan));
 	if (!shdev->chan_reg)
 		goto emapchan;
@@ -1231,8 +702,23 @@ static int __init sh_dmae_probe(struct platform_device *pdev)
 			goto emapdmars;
 	}
 
+	if (!pdata->slave_only)
+		dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask);
+	if (pdata->slave && pdata->slave_num)
+		dma_cap_set(DMA_SLAVE, dma_dev->cap_mask);
+
+	/* Default transfer size of 32 bytes requires 32-byte alignment */
+	dma_dev->copy_align = LOG2_DEFAULT_XFER_SIZE;
+
+	shdev->simple_dev.ops = &sh_dmae_simple_ops;
+	shdev->simple_dev.desc_size = sizeof(struct sh_dmae_desc);
+	err = dma_simple_init(&pdev->dev, &shdev->simple_dev,
+			      pdata->channel_num);
+	if (err < 0)
+		goto esimple;
+
 	/* platform data */
-	shdev->pdata = pdata;
+	shdev->pdata = pdev->dev.platform_data;
 
 	if (pdata->chcr_offset)
 		shdev->chcr_offset = pdata->chcr_offset;
@@ -1246,10 +732,10 @@ static int __init sh_dmae_probe(struct platform_device *pdev)
 
 	platform_set_drvdata(pdev, shdev);
 
-	shdev->common.dev = &pdev->dev;
-
 	pm_runtime_enable(&pdev->dev);
-	pm_runtime_get_sync(&pdev->dev);
+	err = pm_runtime_get_sync(&pdev->dev);
+	if (err < 0)
+		dev_err(&pdev->dev, "%s(): GET = %d\n", __func__, err);
 
 	spin_lock_irq(&sh_dmae_lock);
 	list_add_tail_rcu(&shdev->node, &sh_dmae_devices);
@@ -1260,27 +746,6 @@ static int __init sh_dmae_probe(struct platform_device *pdev)
 	if (err)
 		goto rst_err;
 
-	INIT_LIST_HEAD(&shdev->common.channels);
-
-	if (!pdata->slave_only)
-		dma_cap_set(DMA_MEMCPY, shdev->common.cap_mask);
-	if (pdata->slave && pdata->slave_num)
-		dma_cap_set(DMA_SLAVE, shdev->common.cap_mask);
-
-	shdev->common.device_alloc_chan_resources
-		= sh_dmae_alloc_chan_resources;
-	shdev->common.device_free_chan_resources = sh_dmae_free_chan_resources;
-	shdev->common.device_prep_dma_memcpy = sh_dmae_prep_memcpy;
-	shdev->common.device_tx_status = sh_dmae_tx_status;
-	shdev->common.device_issue_pending = sh_dmae_memcpy_issue_pending;
-
-	/* Compulsory for DMA_SLAVE fields */
-	shdev->common.device_prep_slave_sg = sh_dmae_prep_slave_sg;
-	shdev->common.device_control = sh_dmae_control;
-
-	/* Default transfer size of 32 bytes requires 32-byte alignment */
-	shdev->common.copy_align = LOG2_DEFAULT_XFER_SIZE;
-
 #if defined(CONFIG_CPU_SH4) || defined(CONFIG_ARCH_SHMOBILE)
 	chanirq_res = platform_get_resource(pdev, IORESOURCE_IRQ, 1);
 
@@ -1312,7 +777,7 @@ static int __init sh_dmae_probe(struct platform_device *pdev)
 	    !platform_get_resource(pdev, IORESOURCE_IRQ, 1)) {
 		/* Special case - all multiplexed */
 		for (; irq_cnt < pdata->channel_num; irq_cnt++) {
-			if (irq_cnt < SH_DMAC_MAX_CHANNELS) {
+			if (irq_cnt < SH_DMAE_MAX_CHANNELS) {
 				chan_irq[irq_cnt] = chanirq_res->start;
 				chan_flag[irq_cnt] = IRQF_SHARED;
 			} else {
@@ -1323,7 +788,7 @@ static int __init sh_dmae_probe(struct platform_device *pdev)
 	} else {
 		do {
 			for (i = chanirq_res->start; i <= chanirq_res->end; i++) {
-				if (irq_cnt >= SH_DMAC_MAX_CHANNELS) {
+				if (irq_cnt >= SH_DMAE_MAX_CHANNELS) {
 					irq_cap = 1;
 					break;
 				}
@@ -1339,7 +804,7 @@ static int __init sh_dmae_probe(struct platform_device *pdev)
 				chan_irq[irq_cnt++] = i;
 			}
 
-			if (irq_cnt >= SH_DMAC_MAX_CHANNELS)
+			if (irq_cnt >= SH_DMAE_MAX_CHANNELS)
 				break;
 
 			chanirq_res = platform_get_resource(pdev,
@@ -1357,14 +822,19 @@ static int __init sh_dmae_probe(struct platform_device *pdev)
 	if (irq_cap)
 		dev_notice(&pdev->dev, "Attempting to register %d DMA "
 			   "channels when a maximum of %d are supported.\n",
-			   pdata->channel_num, SH_DMAC_MAX_CHANNELS);
+			   pdata->channel_num, SH_DMAE_MAX_CHANNELS);
 
 	pm_runtime_put(&pdev->dev);
 
-	dma_async_device_register(&shdev->common);
+	err = dma_async_device_register(&shdev->simple_dev.dma_dev);
+	if (err < 0)
+		goto edmadevreg;
 
 	return err;
 
+edmadevreg:
+	pm_runtime_get(&pdev->dev);
+
 chan_probe_err:
 	sh_dmae_chan_remove(shdev);
 
@@ -1380,10 +850,11 @@ rst_err:
 	pm_runtime_put(&pdev->dev);
 	pm_runtime_disable(&pdev->dev);
 
+	platform_set_drvdata(pdev, NULL);
+	dma_simple_cleanup(&shdev->simple_dev);
+esimple:
 	if (dmars)
 		iounmap(shdev->dmars);
-
-	platform_set_drvdata(pdev, NULL);
 emapdmars:
 	iounmap(shdev->chan_reg);
 	synchronize_rcu();
@@ -1398,13 +869,14 @@ ermrdmars:
 	return err;
 }
 
-static int __exit sh_dmae_remove(struct platform_device *pdev)
+static int __devexit sh_dmae_remove(struct platform_device *pdev)
 {
 	struct sh_dmae_device *shdev = platform_get_drvdata(pdev);
+	struct dma_device *dma_dev = &shdev->simple_dev.dma_dev;
 	struct resource *res;
 	int errirq = platform_get_irq(pdev, 0);
 
-	dma_async_device_unregister(&shdev->common);
+	dma_async_device_unregister(dma_dev);
 
 	if (errirq > 0)
 		free_irq(errirq, shdev);
@@ -1413,11 +885,11 @@ static int __exit sh_dmae_remove(struct platform_device *pdev)
 	list_del_rcu(&shdev->node);
 	spin_unlock_irq(&sh_dmae_lock);
 
-	/* channel data remove */
-	sh_dmae_chan_remove(shdev);
-
 	pm_runtime_disable(&pdev->dev);
 
+	sh_dmae_chan_remove(shdev);
+	dma_simple_cleanup(&shdev->simple_dev);
+
 	if (shdev->dmars)
 		iounmap(shdev->dmars);
 	iounmap(shdev->chan_reg);
@@ -1437,77 +909,14 @@ static int __exit sh_dmae_remove(struct platform_device *pdev)
 	return 0;
 }
 
-static void sh_dmae_shutdown(struct platform_device *pdev)
-{
-	struct sh_dmae_device *shdev = platform_get_drvdata(pdev);
-	sh_dmae_ctl_stop(shdev);
-}
-
-static int sh_dmae_runtime_suspend(struct device *dev)
-{
-	return 0;
-}
-
-static int sh_dmae_runtime_resume(struct device *dev)
-{
-	struct sh_dmae_device *shdev = dev_get_drvdata(dev);
-
-	return sh_dmae_rst(shdev);
-}
-
-#ifdef CONFIG_PM
-static int sh_dmae_suspend(struct device *dev)
-{
-	return 0;
-}
-
-static int sh_dmae_resume(struct device *dev)
-{
-	struct sh_dmae_device *shdev = dev_get_drvdata(dev);
-	int i, ret;
-
-	ret = sh_dmae_rst(shdev);
-	if (ret < 0)
-		dev_err(dev, "Failed to reset!\n");
-
-	for (i = 0; i < shdev->pdata->channel_num; i++) {
-		struct sh_dmae_chan *sh_chan = shdev->chan[i];
-		struct sh_dmae_slave *param = sh_chan->common.private;
-
-		if (!sh_chan->descs_allocated)
-			continue;
-
-		if (param) {
-			const struct sh_dmae_slave_config *cfg = param->config;
-			dmae_set_dmars(sh_chan, cfg->mid_rid);
-			dmae_set_chcr(sh_chan, cfg->chcr);
-		} else {
-			dmae_init(sh_chan);
-		}
-	}
-
-	return 0;
-}
-#else
-#define sh_dmae_suspend NULL
-#define sh_dmae_resume NULL
-#endif
-
-const struct dev_pm_ops sh_dmae_pm = {
-	.suspend		= sh_dmae_suspend,
-	.resume			= sh_dmae_resume,
-	.runtime_suspend	= sh_dmae_runtime_suspend,
-	.runtime_resume		= sh_dmae_runtime_resume,
-};
-
 static struct platform_driver sh_dmae_driver = {
-	.remove		= __exit_p(sh_dmae_remove),
-	.shutdown	= sh_dmae_shutdown,
-	.driver = {
+	.driver 	= {
 		.owner	= THIS_MODULE,
-		.name	= "sh-dma-engine",
 		.pm	= &sh_dmae_pm,
+		.name	= SH_DMAE_DRV_NAME,
 	},
+	.remove		= __devexit_p(sh_dmae_remove),
+	.shutdown	= sh_dmae_shutdown,
 };
 
 static int __init sh_dmae_init(void)
@@ -1531,5 +940,6 @@ module_exit(sh_dmae_exit);
 
 MODULE_AUTHOR("Nobuhiro Iwamatsu <iwamatsu.nobuhiro@renesas.com>");
 MODULE_DESCRIPTION("Renesas SH DMA Engine driver");
-MODULE_LICENSE("GPL");
-MODULE_ALIAS("platform:sh-dma-engine");
+MODULE_LICENSE("GPL v2");
+MODULE_VERSION("0.1.0");
+MODULE_ALIAS("platform:" SH_DMAE_DRV_NAME);
diff --git a/drivers/dma/shdma.h b/drivers/dma/shdma.h
index 2b55a27..8bffc19 100644
--- a/drivers/dma/shdma.h
+++ b/drivers/dma/shdma.h
@@ -13,43 +13,27 @@
 #ifndef __DMA_SHDMA_H
 #define __DMA_SHDMA_H
 
+#include <linux/dma-simple.h>
 #include <linux/dmaengine.h>
 #include <linux/interrupt.h>
 #include <linux/list.h>
 
-#define SH_DMAC_MAX_CHANNELS 20
-#define SH_DMA_SLAVE_NUMBER 256
-#define SH_DMA_TCR_MAX 0x00FFFFFF	/* 16MB */
+#define SH_DMAE_MAX_CHANNELS 20
+#define SH_DMAE_TCR_MAX 0x00FFFFFF	/* 16MB */
 
 struct device;
 
-enum dmae_pm_state {
-	DMAE_PM_ESTABLISHED,
-	DMAE_PM_BUSY,
-	DMAE_PM_PENDING,
-};
-
 struct sh_dmae_chan {
-	dma_cookie_t completed_cookie;	/* The maximum cookie completed */
-	spinlock_t desc_lock;		/* Descriptor operation lock */
-	struct list_head ld_queue;	/* Link descriptors queue */
-	struct list_head ld_free;	/* Link descriptors free */
-	struct dma_chan common;		/* DMA common channel */
-	struct device *dev;		/* Channel device */
-	struct tasklet_struct tasklet;	/* Tasklet */
-	int descs_allocated;		/* desc count */
+	struct dma_simple_chan simple_chan;
 	int xmit_shift;			/* log_2(bytes_per_xfer) */
-	int irq;
-	int id;				/* Raw id of this channel */
 	u32 __iomem *base;
 	char dev_id[16];		/* unique name per DMAC of channel */
 	int pm_error;
-	enum dmae_pm_state pm_state;
 };
 
 struct sh_dmae_device {
-	struct dma_device common;
-	struct sh_dmae_chan *chan[SH_DMAC_MAX_CHANNELS];
+	struct dma_simple_dev simple_dev;
+	struct sh_dmae_chan *chan[SH_DMAE_MAX_CHANNELS];
 	struct sh_dmae_pdata *pdata;
 	struct list_head node;
 	u32 __iomem *chan_reg;
@@ -58,10 +42,21 @@ struct sh_dmae_device {
 	u32 chcr_ie_bit;
 };
 
-#define to_sh_chan(chan) container_of(chan, struct sh_dmae_chan, common)
+struct sh_dmae_regs {
+	u32 sar; /* SAR / source address */
+	u32 dar; /* DAR / destination address */
+	u32 tcr; /* TCR / transfer count */
+};
+
+struct sh_dmae_desc {
+	struct sh_dmae_regs hw;
+	struct dma_simple_desc simple_desc;
+};
+
+#define to_sh_chan(chan) container_of(chan, struct sh_dmae_chan, simple_chan)
 #define to_sh_desc(lh) container_of(lh, struct sh_desc, node)
 #define tx_to_sh_desc(tx) container_of(tx, struct sh_desc, async_tx)
-#define to_sh_dev(chan) container_of(chan->common.device,\
-				     struct sh_dmae_device, common)
+#define to_sh_dev(chan) container_of(chan->simple_chan.dma_chan.device,\
+				     struct sh_dmae_device, simple_dev.dma_dev)
 
 #endif	/* __DMA_SHDMA_H */
diff --git a/include/linux/sh_dma.h b/include/linux/sh_dma.h
index ed2aa1e..4f22466 100644
--- a/include/linux/sh_dma.h
+++ b/include/linux/sh_dma.h
@@ -13,34 +13,14 @@
 #include <linux/dma-simple.h>
 #include <linux/dmaengine.h>
 #include <linux/list.h>
+#include <linux/types.h>
 
-/* Used by slave DMA clients to request DMA to/from a specific peripheral */
-struct sh_dmae_slave {
-	union {
-		unsigned int		slave_id; /* Set by the platform */
-		struct dma_simple_slave	simple_slave;
-	};
-	struct device			*dma_dev; /* Set by the platform */
-	const struct sh_dmae_slave_config	*config;  /* Set by the driver */
-};
-
-struct sh_dmae_regs {
-	u32 sar; /* SAR / source address */
-	u32 dar; /* DAR / destination address */
-	u32 tcr; /* TCR / transfer count */
-};
-
-struct sh_desc {
-	struct sh_dmae_regs hw;
-	struct list_head node;
-	struct dma_async_tx_descriptor async_tx;
-	enum dma_transfer_direction direction;
-	dma_cookie_t cookie;
-	size_t partial;
-	int chunks;
-	int mark;
-};
+struct device;
 
+/*
+ * Supplied by platforms to specify, how a DMA channel has to be configured for
+ * a certain peripheral
+ */
 struct sh_dmae_slave_config {
 	unsigned int			slave_id;
 	dma_addr_t			addr;
@@ -48,6 +28,13 @@ struct sh_dmae_slave_config {
 	char				mid_rid;
 };
 
+/* Used by slave DMA clients to request DMA to/from a specific peripheral */
+struct sh_dmae_slave {
+	struct dma_simple_slave		simple_slave;	/* Set by the platform */
+	struct device			*dma_dev;	/* Set by the platform */
+	const struct sh_dmae_slave_config *config;	/* Set by the driver */
+};
+
 struct sh_dmae_channel {
 	unsigned int	offset;
 	unsigned int	dmars;
-- 
1.7.2.5


^ permalink raw reply related	[flat|nested] 33+ messages in thread

* Re: [alsa-devel] [PATCH 1/7 v2] dmaengine: add a simple dma library
  2012-01-26 14:56 ` [PATCH 1/7 v2] dmaengine: add a simple dma library Guennadi Liakhovetski
@ 2012-01-26 17:34   ` Sascha Hauer
  2012-01-26 21:07     ` Guennadi Liakhovetski
  2012-01-27  8:37   ` Shimoda, Yoshihiro
                     ` (2 subsequent siblings)
  3 siblings, 1 reply; 33+ messages in thread
From: Sascha Hauer @ 2012-01-26 17:34 UTC (permalink / raw)
  To: Guennadi Liakhovetski
  Cc: linux-kernel, alsa-devel, linux-sh, Vinod Koul, Magnus Damm,
	Yoshihiro Shimoda, linux-mmc, Paul Mundt, linux-serial

Hi Guennadi,

On Thu, Jan 26, 2012 at 03:56:18PM +0100, Guennadi Liakhovetski wrote:
> This patch adds a library of functions, helping to implement dmaengine
> drivers for hardware, unable to handle scatter-gather lists natively.
> The first version of this driver only supports memcpy and slave DMA
> operation.
> 
> Signed-off-by: Guennadi Liakhovetski <g.liakhovetski@gmx.de>
> ---
> 

[...]

> +
> +/*
> + * For slave DMA we assume, that there is a finite number of DMA slaves in the
> + * system, and that each such slave can only use a finite number of channels.
> + * We use slave channel IDs to make sure, that no such slave channel ID is
> + * allocated more than once.
> + */
> +static unsigned int slave_num = 256;
> +module_param(slave_num, uint, 0444);
> +
> +/* A bitmask with slave_num bits */
> +static unsigned long *simple_slave_used;

You never check that the slave ids passed into this code are
within the range of slave_num. Given that this is a user changeable
value this is a bit flawy.

Sascha

-- 
Pengutronix e.K.                           |                             |
Industrial Linux Solutions                 | http://www.pengutronix.de/  |
Peiner Str. 6-8, 31137 Hildesheim, Germany | Phone: +49-5121-206917-0    |
Amtsgericht Hildesheim, HRA 2686           | Fax:   +49-5121-206917-5555 |

^ permalink raw reply	[flat|nested] 33+ messages in thread

* Re: [alsa-devel] [PATCH 1/7 v2] dmaengine: add a simple dma library
  2012-01-26 17:34   ` [alsa-devel] " Sascha Hauer
@ 2012-01-26 21:07     ` Guennadi Liakhovetski
  0 siblings, 0 replies; 33+ messages in thread
From: Guennadi Liakhovetski @ 2012-01-26 21:07 UTC (permalink / raw)
  To: Sascha Hauer
  Cc: linux-kernel, alsa-devel, linux-sh, Vinod Koul, Magnus Damm,
	Yoshihiro Shimoda, linux-mmc, Paul Mundt, linux-serial

Hi Sascha

On Thu, 26 Jan 2012, Sascha Hauer wrote:

> Hi Guennadi,
> 
> On Thu, Jan 26, 2012 at 03:56:18PM +0100, Guennadi Liakhovetski wrote:
> > This patch adds a library of functions, helping to implement dmaengine
> > drivers for hardware, unable to handle scatter-gather lists natively.
> > The first version of this driver only supports memcpy and slave DMA
> > operation.
> > 
> > Signed-off-by: Guennadi Liakhovetski <g.liakhovetski@gmx.de>
> > ---
> > 
> 
> [...]
> 
> > +
> > +/*
> > + * For slave DMA we assume, that there is a finite number of DMA slaves in the
> > + * system, and that each such slave can only use a finite number of channels.
> > + * We use slave channel IDs to make sure, that no such slave channel ID is
> > + * allocated more than once.
> > + */
> > +static unsigned int slave_num = 256;
> > +module_param(slave_num, uint, 0444);
> > +
> > +/* A bitmask with slave_num bits */
> > +static unsigned long *simple_slave_used;
> 
> You never check that the slave ids passed into this code are
> within the range of slave_num. Given that this is a user changeable
> value this is a bit flawy.

Right, never trust the user;-) I'll fix that.

Thanks
Guennadi
---
Guennadi Liakhovetski, Ph.D.
Freelance Open-Source Software Developer
http://www.open-technology.de/

^ permalink raw reply	[flat|nested] 33+ messages in thread

* Re: [PATCH 1/7 v2] dmaengine: add a simple dma library
  2012-01-26 14:56 ` [PATCH 1/7 v2] dmaengine: add a simple dma library Guennadi Liakhovetski
  2012-01-26 17:34   ` [alsa-devel] " Sascha Hauer
@ 2012-01-27  8:37   ` Shimoda, Yoshihiro
  2012-01-27  8:48     ` Guennadi Liakhovetski
  2012-01-30  8:32   ` [PATCH 1/7 v2] dmaengine: add a simple dma library Vinod Koul
  2012-02-06  2:54   ` Vinod Koul
  3 siblings, 1 reply; 33+ messages in thread
From: Shimoda, Yoshihiro @ 2012-01-27  8:37 UTC (permalink / raw)
  To: Guennadi Liakhovetski
  Cc: linux-kernel, linux-sh, Vinod Koul, Magnus Damm, linux-mmc,
	alsa-devel, linux-serial, Paul Mundt

Hello Guennadi-san,

2012/01/26 23:56, Guennadi Liakhovetski wrote:
> This patch adds a library of functions, helping to implement dmaengine
> drivers for hardware, unable to handle scatter-gather lists natively.
> The first version of this driver only supports memcpy and slave DMA
> operation.
> 
> Signed-off-by: Guennadi Liakhovetski <g.liakhovetski@gmx.de>
> ---
> 
> v2:
> 
> 1. switch from using a tasklet to threaded IRQ, which allowed to

I tested this driver on 2 environment, but it could not work correctly.
 - renesas_usbhs driver with shdma driver on the mackerel board
 - renesas_usbhs driver with experimental SUDMAC driver

In this case, should we modify the renesas_usbhs driver?
For reference, if I changed the threaded IRQ to tasklet,
the 2 environment worked correctly:
==============
diff --git a/drivers/dma/dma-simple.c b/drivers/dma/dma-simple.c
index 49d8f7d..65a5170 100644
--- a/drivers/dma/dma-simple.c
+++ b/drivers/dma/dma-simple.c
@@ -715,16 +715,18 @@ static irqreturn_t chan_irq(int irq, void *dev)

 	spin_lock(&schan->chan_lock);

-	ret = ops->chan_irq(schan, irq) ? IRQ_WAKE_THREAD : IRQ_NONE;
+	ret = ops->chan_irq(schan, irq) ? IRQ_HANDLED : IRQ_NONE;
+	if (ret == IRQ_HANDLED)
+		tasklet_schedule(&schan->tasklet);

 	spin_unlock(&schan->chan_lock);

 	return ret;
 }

-static irqreturn_t chan_irqt(int irq, void *dev)
+static void chan_irqt(unsigned long dev)
 {
-	struct dma_simple_chan *schan = dev;
+	struct dma_simple_chan *schan = (struct dma_simple_chan *)dev;
 	const struct dma_simple_ops *ops =
 		to_simple_dev(schan->dma_chan.device)->ops;
 	struct dma_simple_desc *sdesc;
@@ -744,15 +746,13 @@ static irqreturn_t chan_irqt(int irq, void *dev)
 	spin_unlock_irq(&schan->chan_lock);

 	simple_chan_ld_cleanup(schan, false);
-
-	return IRQ_HANDLED;
 }

 int dma_simple_request_irq(struct dma_simple_chan *schan, int irq,
 			   unsigned long flags, const char *name)
 {
-	int ret = request_threaded_irq(irq, chan_irq, chan_irqt,
-				       flags, name, schan);
+	int ret = request_irq(irq, chan_irq, flags, name, schan);
+	tasklet_init(&schan->tasklet, chan_irqt, (unsigned long)schan);

 	schan->irq = ret < 0 ? ret : irq;

diff --git a/include/linux/dma-simple.h b/include/linux/dma-simple.h
index 5336674..beb1489 100644
--- a/include/linux/dma-simple.h
+++ b/include/linux/dma-simple.h
@@ -68,6 +68,7 @@ struct dma_simple_chan {
 	int id;				/* Raw id of this channel */
 	int irq;			/* Channel IRQ */
 	enum dma_simple_pm_state pm_state;
+	struct tasklet_struct tasklet;
 };

 /**
==============

Best regards,
Yoshihiro Shimoda

^ permalink raw reply related	[flat|nested] 33+ messages in thread

* Re: [PATCH 1/7 v2] dmaengine: add a simple dma library
  2012-01-27  8:37   ` Shimoda, Yoshihiro
@ 2012-01-27  8:48     ` Guennadi Liakhovetski
  2012-01-31 11:03       ` Shimoda, Yoshihiro
  0 siblings, 1 reply; 33+ messages in thread
From: Guennadi Liakhovetski @ 2012-01-27  8:48 UTC (permalink / raw)
  To: Shimoda, Yoshihiro
  Cc: linux-kernel, linux-sh, Vinod Koul, Magnus Damm, linux-mmc,
	alsa-devel, linux-serial, Paul Mundt

Hi Shimoda-san

On Fri, 27 Jan 2012, Shimoda, Yoshihiro wrote:

> Hello Guennadi-san,
> 
> 2012/01/26 23:56, Guennadi Liakhovetski wrote:
> > This patch adds a library of functions, helping to implement dmaengine
> > drivers for hardware, unable to handle scatter-gather lists natively.
> > The first version of this driver only supports memcpy and slave DMA
> > operation.
> > 
> > Signed-off-by: Guennadi Liakhovetski <g.liakhovetski@gmx.de>
> > ---
> > 
> > v2:
> > 
> > 1. switch from using a tasklet to threaded IRQ, which allowed to
> 
> I tested this driver on 2 environment, but it could not work correctly.
>  - renesas_usbhs driver with shdma driver on the mackerel board
>  - renesas_usbhs driver with experimental SUDMAC driver
> 
> In this case, should we modify the renesas_usbhs driver?

Yes, you do. Please, see the respective version of the shdma patch. It 
doesn't request channel IRQs any more directly, instead, it uses a 
dma-simple wrapper function. Also operations change a bit.

Thanks
Guennadi

> For reference, if I changed the threaded IRQ to tasklet,
> the 2 environment worked correctly:
> ==============
> diff --git a/drivers/dma/dma-simple.c b/drivers/dma/dma-simple.c
> index 49d8f7d..65a5170 100644
> --- a/drivers/dma/dma-simple.c
> +++ b/drivers/dma/dma-simple.c
> @@ -715,16 +715,18 @@ static irqreturn_t chan_irq(int irq, void *dev)
> 
>  	spin_lock(&schan->chan_lock);
> 
> -	ret = ops->chan_irq(schan, irq) ? IRQ_WAKE_THREAD : IRQ_NONE;
> +	ret = ops->chan_irq(schan, irq) ? IRQ_HANDLED : IRQ_NONE;
> +	if (ret == IRQ_HANDLED)
> +		tasklet_schedule(&schan->tasklet);
> 
>  	spin_unlock(&schan->chan_lock);
> 
>  	return ret;
>  }
> 
> -static irqreturn_t chan_irqt(int irq, void *dev)
> +static void chan_irqt(unsigned long dev)
>  {
> -	struct dma_simple_chan *schan = dev;
> +	struct dma_simple_chan *schan = (struct dma_simple_chan *)dev;
>  	const struct dma_simple_ops *ops =
>  		to_simple_dev(schan->dma_chan.device)->ops;
>  	struct dma_simple_desc *sdesc;
> @@ -744,15 +746,13 @@ static irqreturn_t chan_irqt(int irq, void *dev)
>  	spin_unlock_irq(&schan->chan_lock);
> 
>  	simple_chan_ld_cleanup(schan, false);
> -
> -	return IRQ_HANDLED;
>  }
> 
>  int dma_simple_request_irq(struct dma_simple_chan *schan, int irq,
>  			   unsigned long flags, const char *name)
>  {
> -	int ret = request_threaded_irq(irq, chan_irq, chan_irqt,
> -				       flags, name, schan);
> +	int ret = request_irq(irq, chan_irq, flags, name, schan);
> +	tasklet_init(&schan->tasklet, chan_irqt, (unsigned long)schan);
> 
>  	schan->irq = ret < 0 ? ret : irq;
> 
> diff --git a/include/linux/dma-simple.h b/include/linux/dma-simple.h
> index 5336674..beb1489 100644
> --- a/include/linux/dma-simple.h
> +++ b/include/linux/dma-simple.h
> @@ -68,6 +68,7 @@ struct dma_simple_chan {
>  	int id;				/* Raw id of this channel */
>  	int irq;			/* Channel IRQ */
>  	enum dma_simple_pm_state pm_state;
> +	struct tasklet_struct tasklet;
>  };
> 
>  /**
> ==============
> 
> Best regards,
> Yoshihiro Shimoda
> 

---
Guennadi Liakhovetski, Ph.D.
Freelance Open-Source Software Developer
http://www.open-technology.de/

^ permalink raw reply	[flat|nested] 33+ messages in thread

* Re: [PATCH 1/7 v2] dmaengine: add a simple dma library
  2012-01-26 14:56 ` [PATCH 1/7 v2] dmaengine: add a simple dma library Guennadi Liakhovetski
  2012-01-26 17:34   ` [alsa-devel] " Sascha Hauer
  2012-01-27  8:37   ` Shimoda, Yoshihiro
@ 2012-01-30  8:32   ` Vinod Koul
  2012-01-30  9:34     ` Guennadi Liakhovetski
  2012-02-06  2:54   ` Vinod Koul
  3 siblings, 1 reply; 33+ messages in thread
From: Vinod Koul @ 2012-01-30  8:32 UTC (permalink / raw)
  To: Guennadi Liakhovetski
  Cc: linux-kernel, linux-sh, Magnus Damm, Yoshihiro Shimoda,
	linux-mmc, alsa-devel, linux-serial, Paul Mundt

On Thu, 2012-01-26 at 15:56 +0100, Guennadi Liakhovetski wrote:
> This patch adds a library of functions, helping to implement dmaengine
> drivers for hardware, unable to handle scatter-gather lists natively.
> The first version of this driver only supports memcpy and slave DMA
> operation.
> 
> Signed-off-by: Guennadi Liakhovetski <g.liakhovetski@gmx.de>
> ---
> 
> v2:
> 
> 1. switch from using a tasklet to threaded IRQ, which allowed to
> 2. remove lock / unlock inline functions
> 3. remove __devinit, __devexit annotations
Sorry to join the discussion late, was on vacation, travel, long
weekend...

I don't still comprehend the need for a library on top of dmaengine
which gain is just a library between clients and dmacs. Surely we don't
want to write another abstraction on top of one provided?

If the question is to handle scatter-gather even if the hardware doesn't
have the capability, then why don't add that in dmaengine itself rather
than one more layer?

-- 
~Vinod


^ permalink raw reply	[flat|nested] 33+ messages in thread

* Re: [PATCH 1/7 v2] dmaengine: add a simple dma library
  2012-01-30  8:32   ` [PATCH 1/7 v2] dmaengine: add a simple dma library Vinod Koul
@ 2012-01-30  9:34     ` Guennadi Liakhovetski
  2012-01-31  6:41       ` Vinod Koul
  0 siblings, 1 reply; 33+ messages in thread
From: Guennadi Liakhovetski @ 2012-01-30  9:34 UTC (permalink / raw)
  To: Vinod Koul
  Cc: linux-kernel, linux-sh, Magnus Damm, Yoshihiro Shimoda,
	linux-mmc, alsa-devel, linux-serial, Paul Mundt

Hi Vinod

Thanks for your email.

On Mon, 30 Jan 2012, Vinod Koul wrote:

> On Thu, 2012-01-26 at 15:56 +0100, Guennadi Liakhovetski wrote:
> > This patch adds a library of functions, helping to implement dmaengine
> > drivers for hardware, unable to handle scatter-gather lists natively.
> > The first version of this driver only supports memcpy and slave DMA
> > operation.
> > 
> > Signed-off-by: Guennadi Liakhovetski <g.liakhovetski@gmx.de>
> > ---
> > 
> > v2:
> > 
> > 1. switch from using a tasklet to threaded IRQ, which allowed to
> > 2. remove lock / unlock inline functions
> > 3. remove __devinit, __devexit annotations
> Sorry to join the discussion late, was on vacation, travel, long
> weekend...
> 
> I don't still comprehend the need for a library on top of dmaengine
> which gain is just a library between clients and dmacs. Surely we don't
> want to write another abstraction on top of one provided?
> 
> If the question is to handle scatter-gather even if the hardware doesn't
> have the capability, then why don't add that in dmaengine itself rather
> than one more layer?

Well, yes, adding new abstraction layers is always a decision, that has to 
be well justified. In this case it does at least make the life easier for 
two sh-mobile drivers: shdma and the new SUDMAC driver.

However, I did name the library in a generic way without reference to sh, 
assuming, that it might with time become useful for other architectures 
too. The reasons why I prefered to keep it as an optional addition to 
dmaengine core, instead of tightly integrating it with it are, that (1) I 
did not want to add useless code to drivers, that do not need it, (2) I am 
not sure if and when this library will become useful for other drivers: 
apart from sh I am only familiar with one more dmaengine driver: 
ipu/ipu_idmac.c, and that one supports scatter-gather lists in a limited 
way and has some further peculiarities, that would likely make it a bad 
match for the simple DMA library, (3) keeping it separate makes its 
further development easier.

OTOH, I'm certainly fine with a tighter library integration with the 
dmaengine core. I think, it still would be better to keep it in a separate 
file and only build it if needed, right? This woult also simplify code 
debugging and further development. I can remove the "simple" notation, 
which does make it look like an additional abstraction layer, and replace 
it with, say, sgsoft (scatter-gather software implementation)? A more 
interesting question is what to do with struct dma_simple_dev, struct 
dma_simple_chan, struct dma_simple_desc, that embed struct dma_device, 
struct dma_chan and struct dma_async_tx_descriptor respectively. I don't 
think we want to merge all the additions from those wrapping structs back 
into their dmaengine counterparts?

How would you like to do this? Don't you think, it would be good to allow 
both: either implement a dmaengine driver directly, exactly as all drivers 
are doing now, or use the additional helper library for suitable (simple) 
hardware types? I see it similar to I2C, where you either implement an I2C 
driver directly, or you use the bitbanging abstraction for simpler 
hardware.

Thanks
Guennadi
---
Guennadi Liakhovetski, Ph.D.
Freelance Open-Source Software Developer
http://www.open-technology.de/

^ permalink raw reply	[flat|nested] 33+ messages in thread

* Re: [PATCH 1/7 v2] dmaengine: add a simple dma library
  2012-01-30  9:34     ` Guennadi Liakhovetski
@ 2012-01-31  6:41       ` Vinod Koul
  2012-01-31  8:59         ` Guennadi Liakhovetski
  0 siblings, 1 reply; 33+ messages in thread
From: Vinod Koul @ 2012-01-31  6:41 UTC (permalink / raw)
  To: Guennadi Liakhovetski
  Cc: linux-kernel, linux-sh, Magnus Damm, Yoshihiro Shimoda,
	linux-mmc, alsa-devel, linux-serial, Paul Mundt

On Mon, 2012-01-30 at 10:34 +0100, Guennadi Liakhovetski wrote:
> > I don't still comprehend the need for a library on top of dmaengine
> > which gain is just a library between clients and dmacs. Surely we don't
> > want to write another abstraction on top of one provided?
> > 
> > If the question is to handle scatter-gather even if the hardware doesn't
> > have the capability, then why don't add that in dmaengine itself rather
> > than one more layer?
> 
> Well, yes, adding new abstraction layers is always a decision, that has to 
> be well justified. In this case it does at least make the life easier for 
> two sh-mobile drivers: shdma and the new SUDMAC driver.
> 
> However, I did name the library in a generic way without reference to sh, 
> assuming, that it might with time become useful for other architectures 
> too. The reasons why I prefered to keep it as an optional addition to 
> dmaengine core, instead of tightly integrating it with it are, that (1) I 
> did not want to add useless code to drivers, that do not need it, 
So are we sure that only sh-mobile drivers need this capablity?
Btw does you hardware only support single transfers and no sg support,
would this remain the same in future?
> (2) I am 
> not sure if and when this library will become useful for other drivers: 
> apart from sh I am only familiar with one more dmaengine driver: 
> ipu/ipu_idmac.c, and that one supports scatter-gather lists in a limited 
> way and has some further peculiarities, that would likely make it a bad 
> match for the simple DMA library,
typically the dmacs will support this in some form or other, so your
point is valid :)
>  (3) keeping it separate makes its 
> further development easier.
> 
> OTOH, I'm certainly fine with a tighter library integration with the 
> dmaengine core. I think, it still would be better to keep it in a separate 
> file and only build it if needed, right? This woult also simplify code 
> debugging and further development. I can remove the "simple" notation, 
> which does make it look like an additional abstraction layer, and replace 
> it with, say, sgsoft (scatter-gather software implementation)?
that would be more apt :)
>  A more 
> interesting question is what to do with struct dma_simple_dev, struct 
> dma_simple_chan, struct dma_simple_desc, that embed struct dma_device, 
> struct dma_chan and struct dma_async_tx_descriptor respectively. I don't 
> think we want to merge all the additions from those wrapping structs back 
> into their dmaengine counterparts?
Sure they should be kept separate. I like the wrapping, this keeps it
simple.
> 
> How would you like to do this? Don't you think, it would be good to allow 
> both: either implement a dmaengine driver directly, exactly as all drivers 
> are doing now, or use the additional helper library for suitable (simple) 
> hardware types? I see it similar to I2C, where you either implement an I2C 
> driver directly, or you use the bitbanging abstraction for simpler 
> hardware.
I think it would be good to have both, this can be used by folks who
don't have sg support available.

-- 
~Vinod


^ permalink raw reply	[flat|nested] 33+ messages in thread

* Re: [PATCH 1/7 v2] dmaengine: add a simple dma library
  2012-01-31  6:41       ` Vinod Koul
@ 2012-01-31  8:59         ` Guennadi Liakhovetski
  2012-02-01  5:38           ` Vinod Koul
  0 siblings, 1 reply; 33+ messages in thread
From: Guennadi Liakhovetski @ 2012-01-31  8:59 UTC (permalink / raw)
  To: Vinod Koul
  Cc: linux-kernel, linux-sh, Magnus Damm, Yoshihiro Shimoda,
	linux-mmc, alsa-devel, linux-serial, Paul Mundt

Hi Vinod

Thanks for your reply.

On Tue, 31 Jan 2012, Vinod Koul wrote:

> On Mon, 2012-01-30 at 10:34 +0100, Guennadi Liakhovetski wrote:
> > > I don't still comprehend the need for a library on top of dmaengine
> > > which gain is just a library between clients and dmacs. Surely we don't
> > > want to write another abstraction on top of one provided?
> > > 
> > > If the question is to handle scatter-gather even if the hardware doesn't
> > > have the capability, then why don't add that in dmaengine itself rather
> > > than one more layer?
> > 
> > Well, yes, adding new abstraction layers is always a decision, that has to 
> > be well justified. In this case it does at least make the life easier for 
> > two sh-mobile drivers: shdma and the new SUDMAC driver.
> > 
> > However, I did name the library in a generic way without reference to sh, 
> > assuming, that it might with time become useful for other architectures 
> > too. The reasons why I prefered to keep it as an optional addition to 
> > dmaengine core, instead of tightly integrating it with it are, that (1) I 
> > did not want to add useless code to drivers, that do not need it, 
> So are we sure that only sh-mobile drivers need this capablity?

If I was sure about that, I'd call the library shmobile-dma or something 
similar:-)

> Btw does you hardware only support single transfers and no sg support,

Well, the controller does have some repeat- and reload-modes, but they are 
currently not supported by the driver and from the documentation I cannot 
easily understand, how useful those modes could be to implement SG.

> would this remain the same in future?

Who knows?...

> > (2) I am 
> > not sure if and when this library will become useful for other drivers: 
> > apart from sh I am only familiar with one more dmaengine driver: 
> > ipu/ipu_idmac.c, and that one supports scatter-gather lists in a limited 
> > way and has some further peculiarities, that would likely make it a bad 
> > match for the simple DMA library,
> typically the dmacs will support this in some form or other, so your
> point is valid :)
> >  (3) keeping it separate makes its 
> > further development easier.
> > 
> > OTOH, I'm certainly fine with a tighter library integration with the 
> > dmaengine core. I think, it still would be better to keep it in a separate 
> > file and only build it if needed, right? This woult also simplify code 
> > debugging and further development. I can remove the "simple" notation, 
> > which does make it look like an additional abstraction layer, and replace 
> > it with, say, sgsoft (scatter-gather software implementation)?
> that would be more apt :)
> >  A more 
> > interesting question is what to do with struct dma_simple_dev, struct 
> > dma_simple_chan, struct dma_simple_desc, that embed struct dma_device, 
> > struct dma_chan and struct dma_async_tx_descriptor respectively. I don't 
> > think we want to merge all the additions from those wrapping structs back 
> > into their dmaengine counterparts?
> Sure they should be kept separate. I like the wrapping, this keeps it
> simple.
> > 
> > How would you like to do this? Don't you think, it would be good to allow 
> > both: either implement a dmaengine driver directly, exactly as all drivers 
> > are doing now, or use the additional helper library for suitable (simple) 
> > hardware types? I see it similar to I2C, where you either implement an I2C 
> > driver directly, or you use the bitbanging abstraction for simpler 
> > hardware.
> I think it would be good to have both, this can be used by folks who
> don't have sg support available.

Ok, so, should I just rename the driver to sgsoft? Is this the only 
change, that you'd like to see?

Thanks
Guennadi
---
Guennadi Liakhovetski, Ph.D.
Freelance Open-Source Software Developer
http://www.open-technology.de/

^ permalink raw reply	[flat|nested] 33+ messages in thread

* Re: [PATCH 1/7 v2] dmaengine: add a simple dma library
  2012-01-27  8:48     ` Guennadi Liakhovetski
@ 2012-01-31 11:03       ` Shimoda, Yoshihiro
  2012-02-01  0:52         ` Guennadi Liakhovetski
  2012-02-02 22:19         ` Guennadi Liakhovetski
  0 siblings, 2 replies; 33+ messages in thread
From: Shimoda, Yoshihiro @ 2012-01-31 11:03 UTC (permalink / raw)
  To: Guennadi Liakhovetski
  Cc: linux-kernel, linux-sh, Vinod Koul, Magnus Damm, linux-mmc,
	alsa-devel, linux-serial, Paul Mundt

Hi Guennadi-san,

2012/01/27 17:48, Guennadi Liakhovetski wrote:
> Hi Shimoda-san
> 
> On Fri, 27 Jan 2012, Shimoda, Yoshihiro wrote:
[ snip ]
>>>
>>> 1. switch from using a tasklet to threaded IRQ, which allowed to
>>
>> I tested this driver on 2 environment, but it could not work correctly.
>>  - renesas_usbhs driver with shdma driver on the mackerel board
>>  - renesas_usbhs driver with experimental SUDMAC driver
>>
>> In this case, should we modify the renesas_usbhs driver?
> 
> Yes, you do. Please, see the respective version of the shdma patch. It 
> doesn't request channel IRQs any more directly, instead, it uses a 
> dma-simple wrapper function. Also operations change a bit.

Thank you for your comment.

I investigaed the issue. I found out the renesas_usbhs driver may
call tx_submit() in the callback() of the dma-simple finally.
In this case, the value of power_up in the simple_tx_submit is 0,
the start_xfer() is not called. And then, even if the ld_queue is
not empty, the DMA doesn't start.


So, if I add codes like the following in the dma-simple.c,
the renesas_usbhs driver can work correctly without the driver changes.
If you think the code is good, would you merge it to your code?
======= chan_irqt() =======
	simple_chan_ld_cleanup(schan, false);

+	spin_lock_irq(&schan->chan_lock);
+	/* Next desc */
+	simple_chan_xfer_ld_queue(schan);
+	spin_unlock_irq(&schan->chan_lock);

	return IRQ_HANDLED;
==============

Best regards,
Yoshihiro Shimoda

> Thanks
> Guennadi
> 
>> For reference, if I changed the threaded IRQ to tasklet,
>> the 2 environment worked correctly:
>> ==============
>> diff --git a/drivers/dma/dma-simple.c b/drivers/dma/dma-simple.c
>> index 49d8f7d..65a5170 100644
>> --- a/drivers/dma/dma-simple.c
>> +++ b/drivers/dma/dma-simple.c
>> @@ -715,16 +715,18 @@ static irqreturn_t chan_irq(int irq, void *dev)
>>
>>  	spin_lock(&schan->chan_lock);
>>
>> -	ret = ops->chan_irq(schan, irq) ? IRQ_WAKE_THREAD : IRQ_NONE;
>> +	ret = ops->chan_irq(schan, irq) ? IRQ_HANDLED : IRQ_NONE;
>> +	if (ret == IRQ_HANDLED)
>> +		tasklet_schedule(&schan->tasklet);
>>
>>  	spin_unlock(&schan->chan_lock);
>>
>>  	return ret;
>>  }
>>
>> -static irqreturn_t chan_irqt(int irq, void *dev)
>> +static void chan_irqt(unsigned long dev)
>>  {
>> -	struct dma_simple_chan *schan = dev;
>> +	struct dma_simple_chan *schan = (struct dma_simple_chan *)dev;
>>  	const struct dma_simple_ops *ops =
>>  		to_simple_dev(schan->dma_chan.device)->ops;
>>  	struct dma_simple_desc *sdesc;
>> @@ -744,15 +746,13 @@ static irqreturn_t chan_irqt(int irq, void *dev)
>>  	spin_unlock_irq(&schan->chan_lock);
>>
>>  	simple_chan_ld_cleanup(schan, false);
>> -
>> -	return IRQ_HANDLED;
>>  }
>>
>>  int dma_simple_request_irq(struct dma_simple_chan *schan, int irq,
>>  			   unsigned long flags, const char *name)
>>  {
>> -	int ret = request_threaded_irq(irq, chan_irq, chan_irqt,
>> -				       flags, name, schan);
>> +	int ret = request_irq(irq, chan_irq, flags, name, schan);
>> +	tasklet_init(&schan->tasklet, chan_irqt, (unsigned long)schan);
>>
>>  	schan->irq = ret < 0 ? ret : irq;
>>
>> diff --git a/include/linux/dma-simple.h b/include/linux/dma-simple.h
>> index 5336674..beb1489 100644
>> --- a/include/linux/dma-simple.h
>> +++ b/include/linux/dma-simple.h
>> @@ -68,6 +68,7 @@ struct dma_simple_chan {
>>  	int id;				/* Raw id of this channel */
>>  	int irq;			/* Channel IRQ */
>>  	enum dma_simple_pm_state pm_state;
>> +	struct tasklet_struct tasklet;
>>  };
>>
>>  /**
>> ==============
>>
>> Best regards,
>> Yoshihiro Shimoda
>>
> 
> ---
> Guennadi Liakhovetski, Ph.D.
> Freelance Open-Source Software Developer
> http://www.open-technology.de/
> 


-- 
Yoshihiro Shimoda <yoshihiro.shimoda.uh@renesas.com>
EC No.

^ permalink raw reply	[flat|nested] 33+ messages in thread

* Re: [PATCH 1/7 v2] dmaengine: add a simple dma library
  2012-01-31 11:03       ` Shimoda, Yoshihiro
@ 2012-02-01  0:52         ` Guennadi Liakhovetski
  2012-02-01  2:47           ` Shimoda, Yoshihiro
  2012-02-02 22:19         ` Guennadi Liakhovetski
  1 sibling, 1 reply; 33+ messages in thread
From: Guennadi Liakhovetski @ 2012-02-01  0:52 UTC (permalink / raw)
  To: Shimoda, Yoshihiro
  Cc: linux-kernel, linux-sh, Vinod Koul, Magnus Damm, linux-mmc,
	alsa-devel, linux-serial, Paul Mundt

Hi Shimoda-san

On Tue, 31 Jan 2012, Shimoda, Yoshihiro wrote:

> Hi Guennadi-san,
> 
> 2012/01/27 17:48, Guennadi Liakhovetski wrote:
> > Hi Shimoda-san
> > 
> > On Fri, 27 Jan 2012, Shimoda, Yoshihiro wrote:
> [ snip ]
> >>>
> >>> 1. switch from using a tasklet to threaded IRQ, which allowed to
> >>
> >> I tested this driver on 2 environment, but it could not work correctly.
> >>  - renesas_usbhs driver with shdma driver on the mackerel board
> >>  - renesas_usbhs driver with experimental SUDMAC driver
> >>
> >> In this case, should we modify the renesas_usbhs driver?
> > 
> > Yes, you do. Please, see the respective version of the shdma patch. It 
> > doesn't request channel IRQs any more directly, instead, it uses a 
> > dma-simple wrapper function. Also operations change a bit.
> 
> Thank you for your comment.
> 
> I investigaed the issue. I found out the renesas_usbhs driver may
> call tx_submit() in the callback()

Well, actually, this is not something, that shdma officially supports ATM. 
Originally it was prohibited by the API, but later a correction has been 
added to Documentation/dmaengine.txt, explicitly allowing this. So, good, 
if this works for you, but watch out for possible issues.

> of the dma-simple finally.
> In this case, the value of power_up in the simple_tx_submit is 0,
> the start_xfer() is not called. And then, even if the ld_queue is
> not empty, the DMA doesn't start.

This is correct. As soon as all the transfers, currently queued in the 
shdma driver, has been processed and completed, to restart DMA you need to 
call dma_async_issue_pending().

> So, if I add codes like the following in the dma-simple.c,
> the renesas_usbhs driver can work correctly without the driver changes.
> If you think the code is good, would you merge it to your code?
> ======= chan_irqt() =======
> 	simple_chan_ld_cleanup(schan, false);
> 
> +	spin_lock_irq(&schan->chan_lock);
> +	/* Next desc */
> +	simple_chan_xfer_ld_queue(schan);
> +	spin_unlock_irq(&schan->chan_lock);
> 
> 	return IRQ_HANDLED;
> ==============

Therefore, no, this shouldn't be needed.

Thanks
Guennadi

> 
> Best regards,
> Yoshihiro Shimoda
> 
> > Thanks
> > Guennadi
> > 
> >> For reference, if I changed the threaded IRQ to tasklet,
> >> the 2 environment worked correctly:
> >> ==============
> >> diff --git a/drivers/dma/dma-simple.c b/drivers/dma/dma-simple.c
> >> index 49d8f7d..65a5170 100644
> >> --- a/drivers/dma/dma-simple.c
> >> +++ b/drivers/dma/dma-simple.c
> >> @@ -715,16 +715,18 @@ static irqreturn_t chan_irq(int irq, void *dev)
> >>
> >>  	spin_lock(&schan->chan_lock);
> >>
> >> -	ret = ops->chan_irq(schan, irq) ? IRQ_WAKE_THREAD : IRQ_NONE;
> >> +	ret = ops->chan_irq(schan, irq) ? IRQ_HANDLED : IRQ_NONE;
> >> +	if (ret == IRQ_HANDLED)
> >> +		tasklet_schedule(&schan->tasklet);
> >>
> >>  	spin_unlock(&schan->chan_lock);
> >>
> >>  	return ret;
> >>  }
> >>
> >> -static irqreturn_t chan_irqt(int irq, void *dev)
> >> +static void chan_irqt(unsigned long dev)
> >>  {
> >> -	struct dma_simple_chan *schan = dev;
> >> +	struct dma_simple_chan *schan = (struct dma_simple_chan *)dev;
> >>  	const struct dma_simple_ops *ops =
> >>  		to_simple_dev(schan->dma_chan.device)->ops;
> >>  	struct dma_simple_desc *sdesc;
> >> @@ -744,15 +746,13 @@ static irqreturn_t chan_irqt(int irq, void *dev)
> >>  	spin_unlock_irq(&schan->chan_lock);
> >>
> >>  	simple_chan_ld_cleanup(schan, false);
> >> -
> >> -	return IRQ_HANDLED;
> >>  }
> >>
> >>  int dma_simple_request_irq(struct dma_simple_chan *schan, int irq,
> >>  			   unsigned long flags, const char *name)
> >>  {
> >> -	int ret = request_threaded_irq(irq, chan_irq, chan_irqt,
> >> -				       flags, name, schan);
> >> +	int ret = request_irq(irq, chan_irq, flags, name, schan);
> >> +	tasklet_init(&schan->tasklet, chan_irqt, (unsigned long)schan);
> >>
> >>  	schan->irq = ret < 0 ? ret : irq;
> >>
> >> diff --git a/include/linux/dma-simple.h b/include/linux/dma-simple.h
> >> index 5336674..beb1489 100644
> >> --- a/include/linux/dma-simple.h
> >> +++ b/include/linux/dma-simple.h
> >> @@ -68,6 +68,7 @@ struct dma_simple_chan {
> >>  	int id;				/* Raw id of this channel */
> >>  	int irq;			/* Channel IRQ */
> >>  	enum dma_simple_pm_state pm_state;
> >> +	struct tasklet_struct tasklet;
> >>  };
> >>
> >>  /**
> >> ==============
> >>
> >> Best regards,
> >> Yoshihiro Shimoda
> >>
> > 
> > ---
> > Guennadi Liakhovetski, Ph.D.
> > Freelance Open-Source Software Developer
> > http://www.open-technology.de/
> > 
> 
> 
> -- 
> Yoshihiro Shimoda <yoshihiro.shimoda.uh@renesas.com>
> EC No.
> 

---
Guennadi Liakhovetski, Ph.D.
Freelance Open-Source Software Developer
http://www.open-technology.de/

^ permalink raw reply	[flat|nested] 33+ messages in thread

* Re: [PATCH 1/7 v2] dmaengine: add a simple dma library
  2012-02-01  0:52         ` Guennadi Liakhovetski
@ 2012-02-01  2:47           ` Shimoda, Yoshihiro
  0 siblings, 0 replies; 33+ messages in thread
From: Shimoda, Yoshihiro @ 2012-02-01  2:47 UTC (permalink / raw)
  To: Guennadi Liakhovetski
  Cc: linux-kernel, linux-sh, Vinod Koul, Magnus Damm, linux-mmc,
	alsa-devel, linux-serial, Paul Mundt

Hi Guennadi-san,

2012/02/01 9:52, Guennadi Liakhovetski wrote:
> Hi Shimoda-san
> 
> On Tue, 31 Jan 2012, Shimoda, Yoshihiro wrote:
[ snip ]
>> I investigaed the issue. I found out the renesas_usbhs driver may
>> call tx_submit() in the callback()
> 
> Well, actually, this is not something, that shdma officially supports ATM. 
> Originally it was prohibited by the API, but later a correction has been 
> added to Documentation/dmaengine.txt, explicitly allowing this. So, good, 
> if this works for you, but watch out for possible issues.
>
>> of the dma-simple finally.
>> In this case, the value of power_up in the simple_tx_submit is 0,
>> the start_xfer() is not called. And then, even if the ld_queue is
>> not empty, the DMA doesn't start.
> 
> This is correct. As soon as all the transfers, currently queued in the 
> shdma driver, has been processed and completed, to restart DMA you need to 
> call dma_async_issue_pending().
> 
>> So, if I add codes like the following in the dma-simple.c,
>> the renesas_usbhs driver can work correctly without the driver changes.
>> If you think the code is good, would you merge it to your code?
>> ======= chan_irqt() =======
>> 	simple_chan_ld_cleanup(schan, false);
>>
>> +	spin_lock_irq(&schan->chan_lock);
>> +	/* Next desc */
>> +	simple_chan_xfer_ld_queue(schan);
>> +	spin_unlock_irq(&schan->chan_lock);
>>
>> 	return IRQ_HANDLED;
>> ==============
> 
> Therefore, no, this shouldn't be needed.

Thank you very much for your comment.
I understood it. I will investigate the renesas_usbhs driver more.

Best regards,
Yoshihiro Shimoda

> Thanks
> Guennadi
> 
>>
>> Best regards,
>> Yoshihiro Shimoda
>>
>>> Thanks
>>> Guennadi
>>>
>>>> For reference, if I changed the threaded IRQ to tasklet,
>>>> the 2 environment worked correctly:
>>>> ==============
>>>> diff --git a/drivers/dma/dma-simple.c b/drivers/dma/dma-simple.c
>>>> index 49d8f7d..65a5170 100644
>>>> --- a/drivers/dma/dma-simple.c
>>>> +++ b/drivers/dma/dma-simple.c
>>>> @@ -715,16 +715,18 @@ static irqreturn_t chan_irq(int irq, void *dev)
>>>>
>>>>  	spin_lock(&schan->chan_lock);
>>>>
>>>> -	ret = ops->chan_irq(schan, irq) ? IRQ_WAKE_THREAD : IRQ_NONE;
>>>> +	ret = ops->chan_irq(schan, irq) ? IRQ_HANDLED : IRQ_NONE;
>>>> +	if (ret == IRQ_HANDLED)
>>>> +		tasklet_schedule(&schan->tasklet);
>>>>
>>>>  	spin_unlock(&schan->chan_lock);
>>>>
>>>>  	return ret;
>>>>  }
>>>>
>>>> -static irqreturn_t chan_irqt(int irq, void *dev)
>>>> +static void chan_irqt(unsigned long dev)
>>>>  {
>>>> -	struct dma_simple_chan *schan = dev;
>>>> +	struct dma_simple_chan *schan = (struct dma_simple_chan *)dev;
>>>>  	const struct dma_simple_ops *ops =
>>>>  		to_simple_dev(schan->dma_chan.device)->ops;
>>>>  	struct dma_simple_desc *sdesc;
>>>> @@ -744,15 +746,13 @@ static irqreturn_t chan_irqt(int irq, void *dev)
>>>>  	spin_unlock_irq(&schan->chan_lock);
>>>>
>>>>  	simple_chan_ld_cleanup(schan, false);
>>>> -
>>>> -	return IRQ_HANDLED;
>>>>  }
>>>>
>>>>  int dma_simple_request_irq(struct dma_simple_chan *schan, int irq,
>>>>  			   unsigned long flags, const char *name)
>>>>  {
>>>> -	int ret = request_threaded_irq(irq, chan_irq, chan_irqt,
>>>> -				       flags, name, schan);
>>>> +	int ret = request_irq(irq, chan_irq, flags, name, schan);
>>>> +	tasklet_init(&schan->tasklet, chan_irqt, (unsigned long)schan);
>>>>
>>>>  	schan->irq = ret < 0 ? ret : irq;
>>>>
>>>> diff --git a/include/linux/dma-simple.h b/include/linux/dma-simple.h
>>>> index 5336674..beb1489 100644
>>>> --- a/include/linux/dma-simple.h
>>>> +++ b/include/linux/dma-simple.h
>>>> @@ -68,6 +68,7 @@ struct dma_simple_chan {
>>>>  	int id;				/* Raw id of this channel */
>>>>  	int irq;			/* Channel IRQ */
>>>>  	enum dma_simple_pm_state pm_state;
>>>> +	struct tasklet_struct tasklet;
>>>>  };
>>>>
>>>>  /**
>>>> ==============
>>>>
>>>> Best regards,
>>>> Yoshihiro Shimoda
>>>>
>>>
>>> ---
>>> Guennadi Liakhovetski, Ph.D.
>>> Freelance Open-Source Software Developer
>>> http://www.open-technology.de/
>>>
>>
>>
>> -- 
>> Yoshihiro Shimoda <yoshihiro.shimoda.uh@renesas.com>
>> EC No.
>>
> 
> ---
> Guennadi Liakhovetski, Ph.D.
> Freelance Open-Source Software Developer
> http://www.open-technology.de/
> 


-- 
Yoshihiro Shimoda <yoshihiro.shimoda.uh@renesas.com>
EC No.

^ permalink raw reply	[flat|nested] 33+ messages in thread

* Re: [PATCH 1/7 v2] dmaengine: add a simple dma library
  2012-01-31  8:59         ` Guennadi Liakhovetski
@ 2012-02-01  5:38           ` Vinod Koul
  2012-02-03 16:38             ` Guennadi Liakhovetski
  0 siblings, 1 reply; 33+ messages in thread
From: Vinod Koul @ 2012-02-01  5:38 UTC (permalink / raw)
  To: Guennadi Liakhovetski
  Cc: linux-kernel, linux-sh, Magnus Damm, Yoshihiro Shimoda,
	linux-mmc, alsa-devel, linux-serial, Paul Mundt

On Tue, 2012-01-31 at 09:59 +0100, Guennadi Liakhovetski wrote:
> Hi Vinod
> 
> Thanks for your reply.
> 
> On Tue, 31 Jan 2012, Vinod Koul wrote:
> 
> > On Mon, 2012-01-30 at 10:34 +0100, Guennadi Liakhovetski wrote:
> > > > I don't still comprehend the need for a library on top of dmaengine
> > > > which gain is just a library between clients and dmacs. Surely we don't
> > > > want to write another abstraction on top of one provided?
> > > > 
> > > > If the question is to handle scatter-gather even if the hardware doesn't
> > > > have the capability, then why don't add that in dmaengine itself rather
> > > > than one more layer?
> > > 
> > > Well, yes, adding new abstraction layers is always a decision, that has to 
> > > be well justified. In this case it does at least make the life easier for 
> > > two sh-mobile drivers: shdma and the new SUDMAC driver.
> > > 
> > > However, I did name the library in a generic way without reference to sh, 
> > > assuming, that it might with time become useful for other architectures 
> > > too. The reasons why I prefered to keep it as an optional addition to 
> > > dmaengine core, instead of tightly integrating it with it are, that (1) I 
> > > did not want to add useless code to drivers, that do not need it, 
> > So are we sure that only sh-mobile drivers need this capablity?
> 
> If I was sure about that, I'd call the library shmobile-dma or something 
> similar:-)
> 
> > Btw does you hardware only support single transfers and no sg support,
> 
> Well, the controller does have some repeat- and reload-modes, but they are 
> currently not supported by the driver and from the documentation I cannot 
> easily understand, how useful those modes could be to implement SG.
> 
> > would this remain the same in future?
> 
> Who knows?...
> 
> > > (2) I am 
> > > not sure if and when this library will become useful for other drivers: 
> > > apart from sh I am only familiar with one more dmaengine driver: 
> > > ipu/ipu_idmac.c, and that one supports scatter-gather lists in a limited 
> > > way and has some further peculiarities, that would likely make it a bad 
> > > match for the simple DMA library,
> > typically the dmacs will support this in some form or other, so your
> > point is valid :)
> > >  (3) keeping it separate makes its 
> > > further development easier.
> > > 
> > > OTOH, I'm certainly fine with a tighter library integration with the 
> > > dmaengine core. I think, it still would be better to keep it in a separate 
> > > file and only build it if needed, right? This woult also simplify code 
> > > debugging and further development. I can remove the "simple" notation, 
> > > which does make it look like an additional abstraction layer, and replace 
> > > it with, say, sgsoft (scatter-gather software implementation)?
> > that would be more apt :)
> > >  A more 
> > > interesting question is what to do with struct dma_simple_dev, struct 
> > > dma_simple_chan, struct dma_simple_desc, that embed struct dma_device, 
> > > struct dma_chan and struct dma_async_tx_descriptor respectively. I don't 
> > > think we want to merge all the additions from those wrapping structs back 
> > > into their dmaengine counterparts?
> > Sure they should be kept separate. I like the wrapping, this keeps it
> > simple.
> > > 
> > > How would you like to do this? Don't you think, it would be good to allow 
> > > both: either implement a dmaengine driver directly, exactly as all drivers 
> > > are doing now, or use the additional helper library for suitable (simple) 
> > > hardware types? I see it similar to I2C, where you either implement an I2C 
> > > driver directly, or you use the bitbanging abstraction for simpler 
> > > hardware.
> > I think it would be good to have both, this can be used by folks who
> > don't have sg support available.
> 
> Ok, so, should I just rename the driver to sgsoft? Is this the only 
> change, that you'd like to see?
That would be one :)
Also, I would review the other patch patches by today/tomorrow, you can
add any changes in next version as well

-- 
~Vinod


^ permalink raw reply	[flat|nested] 33+ messages in thread

* Re: [PATCH 1/7 v2] dmaengine: add a simple dma library
  2012-01-31 11:03       ` Shimoda, Yoshihiro
  2012-02-01  0:52         ` Guennadi Liakhovetski
@ 2012-02-02 22:19         ` Guennadi Liakhovetski
  2012-02-03  8:47           ` Shimoda, Yoshihiro
  1 sibling, 1 reply; 33+ messages in thread
From: Guennadi Liakhovetski @ 2012-02-02 22:19 UTC (permalink / raw)
  To: Shimoda, Yoshihiro
  Cc: linux-kernel, linux-sh, Vinod Koul, Magnus Damm, linux-mmc,
	alsa-devel, linux-serial, Paul Mundt

Shimoda-san

On Tue, 31 Jan 2012, Shimoda, Yoshihiro wrote:

> Hi Guennadi-san,
> 
> 2012/01/27 17:48, Guennadi Liakhovetski wrote:
> > Hi Shimoda-san
> > 
> > On Fri, 27 Jan 2012, Shimoda, Yoshihiro wrote:
> [ snip ]
> >>>
> >>> 1. switch from using a tasklet to threaded IRQ, which allowed to
> >>
> >> I tested this driver on 2 environment, but it could not work correctly.
> >>  - renesas_usbhs driver with shdma driver on the mackerel board
> >>  - renesas_usbhs driver with experimental SUDMAC driver
> >>
> >> In this case, should we modify the renesas_usbhs driver?
> > 
> > Yes, you do. Please, see the respective version of the shdma patch. It 
> > doesn't request channel IRQs any more directly, instead, it uses a 
> > dma-simple wrapper function. Also operations change a bit.
> 
> Thank you for your comment.
> 
> I investigaed the issue. I found out the renesas_usbhs driver may
> call tx_submit() in the callback() of the dma-simple finally.

Sorry, in my first reply to this your email I misread the renesas_usbhs 
for an issue with your new SUDMAC driver. Since this is not the case and 
my patch seems to be causing a regression, I'll look at it.

Thanks
Guennadi

> In this case, the value of power_up in the simple_tx_submit is 0,
> the start_xfer() is not called. And then, even if the ld_queue is
> not empty, the DMA doesn't start.
> 
> 
> So, if I add codes like the following in the dma-simple.c,
> the renesas_usbhs driver can work correctly without the driver changes.
> If you think the code is good, would you merge it to your code?
> ======= chan_irqt() =======
> 	simple_chan_ld_cleanup(schan, false);
> 
> +	spin_lock_irq(&schan->chan_lock);
> +	/* Next desc */
> +	simple_chan_xfer_ld_queue(schan);
> +	spin_unlock_irq(&schan->chan_lock);
> 
> 	return IRQ_HANDLED;
> ==============
> 
> Best regards,
> Yoshihiro Shimoda
> 
> > Thanks
> > Guennadi
> > 
> >> For reference, if I changed the threaded IRQ to tasklet,
> >> the 2 environment worked correctly:
> >> ==============
> >> diff --git a/drivers/dma/dma-simple.c b/drivers/dma/dma-simple.c
> >> index 49d8f7d..65a5170 100644
> >> --- a/drivers/dma/dma-simple.c
> >> +++ b/drivers/dma/dma-simple.c
> >> @@ -715,16 +715,18 @@ static irqreturn_t chan_irq(int irq, void *dev)
> >>
> >>  	spin_lock(&schan->chan_lock);
> >>
> >> -	ret = ops->chan_irq(schan, irq) ? IRQ_WAKE_THREAD : IRQ_NONE;
> >> +	ret = ops->chan_irq(schan, irq) ? IRQ_HANDLED : IRQ_NONE;
> >> +	if (ret == IRQ_HANDLED)
> >> +		tasklet_schedule(&schan->tasklet);
> >>
> >>  	spin_unlock(&schan->chan_lock);
> >>
> >>  	return ret;
> >>  }
> >>
> >> -static irqreturn_t chan_irqt(int irq, void *dev)
> >> +static void chan_irqt(unsigned long dev)
> >>  {
> >> -	struct dma_simple_chan *schan = dev;
> >> +	struct dma_simple_chan *schan = (struct dma_simple_chan *)dev;
> >>  	const struct dma_simple_ops *ops =
> >>  		to_simple_dev(schan->dma_chan.device)->ops;
> >>  	struct dma_simple_desc *sdesc;
> >> @@ -744,15 +746,13 @@ static irqreturn_t chan_irqt(int irq, void *dev)
> >>  	spin_unlock_irq(&schan->chan_lock);
> >>
> >>  	simple_chan_ld_cleanup(schan, false);
> >> -
> >> -	return IRQ_HANDLED;
> >>  }
> >>
> >>  int dma_simple_request_irq(struct dma_simple_chan *schan, int irq,
> >>  			   unsigned long flags, const char *name)
> >>  {
> >> -	int ret = request_threaded_irq(irq, chan_irq, chan_irqt,
> >> -				       flags, name, schan);
> >> +	int ret = request_irq(irq, chan_irq, flags, name, schan);
> >> +	tasklet_init(&schan->tasklet, chan_irqt, (unsigned long)schan);
> >>
> >>  	schan->irq = ret < 0 ? ret : irq;
> >>
> >> diff --git a/include/linux/dma-simple.h b/include/linux/dma-simple.h
> >> index 5336674..beb1489 100644
> >> --- a/include/linux/dma-simple.h
> >> +++ b/include/linux/dma-simple.h
> >> @@ -68,6 +68,7 @@ struct dma_simple_chan {
> >>  	int id;				/* Raw id of this channel */
> >>  	int irq;			/* Channel IRQ */
> >>  	enum dma_simple_pm_state pm_state;
> >> +	struct tasklet_struct tasklet;
> >>  };
> >>
> >>  /**
> >> ==============
> >>
> >> Best regards,
> >> Yoshihiro Shimoda
> >>
> > 
> > ---
> > Guennadi Liakhovetski, Ph.D.
> > Freelance Open-Source Software Developer
> > http://www.open-technology.de/
> > 
> 
> 
> -- 
> Yoshihiro Shimoda <yoshihiro.shimoda.uh@renesas.com>
> EC No.
> 

---
Guennadi Liakhovetski, Ph.D.
Freelance Open-Source Software Developer
http://www.open-technology.de/

^ permalink raw reply	[flat|nested] 33+ messages in thread

* Re: [PATCH 1/7 v2] dmaengine: add a simple dma library
  2012-02-02 22:19         ` Guennadi Liakhovetski
@ 2012-02-03  8:47           ` Shimoda, Yoshihiro
  2012-02-03 10:21             ` Guennadi Liakhovetski
  2012-02-03 15:43             ` [PATCH/RFC] usb: fix renesas_usbhs to not schedule in atomic context Guennadi Liakhovetski
  0 siblings, 2 replies; 33+ messages in thread
From: Shimoda, Yoshihiro @ 2012-02-03  8:47 UTC (permalink / raw)
  To: Guennadi Liakhovetski
  Cc: linux-kernel, linux-sh, Vinod Koul, Magnus Damm, linux-mmc,
	alsa-devel, linux-serial, Paul Mundt

Hi Guennadi-san,

2012/02/03 7:19, Guennadi Liakhovetski wrote:
[ snip ]
>>
>> Thank you for your comment.
>>
>> I investigaed the issue. I found out the renesas_usbhs driver may
>> call tx_submit() in the callback() of the dma-simple finally.
> 
> Sorry, in my first reply to this your email I misread the renesas_usbhs 
> for an issue with your new SUDMAC driver. Since this is not the case and 
> my patch seems to be causing a regression, I'll look at it.

Thank you for the comment.

I found out that if the "power_up" is 0 in the sample_tx_submit(),
the pm_state will be changed to "PENDING".
And then, even if the renesas_usbhs calls dma_async_issue_pending(),
the dma-simple driver doesn't call the simple_chan_xfer_ld_queue().
This is because the driver will call the simple_chan_xfer_ld_queue()
when the pm_state is "ESTABLISHED" only.

So I guess that we have to modify the simple_issue_pending().

Best regards,
Yoshihiro Shimoda

> Thanks
> Guennadi
> 
> ---
> Guennadi Liakhovetski, Ph.D.
> Freelance Open-Source Software Developer
> http://www.open-technology.de/
> 
-- 
Yoshihiro Shimoda <yoshihiro.shimoda.uh@renesas.com>

^ permalink raw reply	[flat|nested] 33+ messages in thread

* Re: [PATCH 1/7 v2] dmaengine: add a simple dma library
  2012-02-03  8:47           ` Shimoda, Yoshihiro
@ 2012-02-03 10:21             ` Guennadi Liakhovetski
  2012-02-03 15:43             ` [PATCH/RFC] usb: fix renesas_usbhs to not schedule in atomic context Guennadi Liakhovetski
  1 sibling, 0 replies; 33+ messages in thread
From: Guennadi Liakhovetski @ 2012-02-03 10:21 UTC (permalink / raw)
  To: Shimoda, Yoshihiro
  Cc: linux-kernel, linux-sh, Vinod Koul, Magnus Damm, linux-mmc,
	alsa-devel, linux-serial, Paul Mundt

On Fri, 3 Feb 2012, Shimoda, Yoshihiro wrote:

> Hi Guennadi-san,
> 
> 2012/02/03 7:19, Guennadi Liakhovetski wrote:
> [ snip ]
> >>
> >> Thank you for your comment.
> >>
> >> I investigaed the issue. I found out the renesas_usbhs driver may
> >> call tx_submit() in the callback() of the dma-simple finally.
> > 
> > Sorry, in my first reply to this your email I misread the renesas_usbhs 
> > for an issue with your new SUDMAC driver. Since this is not the case and 
> > my patch seems to be causing a regression, I'll look at it.
> 
> Thank you for the comment.
> 
> I found out that if the "power_up" is 0 in the sample_tx_submit(),
> the pm_state will be changed to "PENDING".
> And then, even if the renesas_usbhs calls dma_async_issue_pending(),
> the dma-simple driver doesn't call the simple_chan_xfer_ld_queue().
> This is because the driver will call the simple_chan_xfer_ld_queue()
> when the pm_state is "ESTABLISHED" only.
> 
> So I guess that we have to modify the simple_issue_pending().

Ok, I looked at this code, there's now a comment in simple_tx_submit():

	if (power_up) {
		...
	} else {
		/*
		 * Tell .device_issue_pending() not to run the queue, interrupts
		 * will do it anyway
		 */
		schan->pm_state = DMA_SIMPLE_PM_PENDING;
	}

And that's exactly what should be happening: an interrupt should trigger 
the IRQ thread should call simple_chan_xfer_ld_queue(), so, looks like the 
interrupt is not coming? It's also strange, why and how it worked without 
dma-simple? I don't think I changed that with the transition... 
Investigating.

Thanks
Guennadi
---
Guennadi Liakhovetski, Ph.D.
Freelance Open-Source Software Developer
http://www.open-technology.de/

^ permalink raw reply	[flat|nested] 33+ messages in thread

* [PATCH/RFC] usb: fix renesas_usbhs to not schedule in atomic context
  2012-02-03  8:47           ` Shimoda, Yoshihiro
  2012-02-03 10:21             ` Guennadi Liakhovetski
@ 2012-02-03 15:43             ` Guennadi Liakhovetski
  2012-02-05 14:54               ` Felipe Balbi
  2012-02-06  8:52               ` Shimoda, Yoshihiro
  1 sibling, 2 replies; 33+ messages in thread
From: Guennadi Liakhovetski @ 2012-02-03 15:43 UTC (permalink / raw)
  To: Shimoda, Yoshihiro
  Cc: linux-kernel, linux-sh, Vinod Koul, Magnus Damm, linux-mmc,
	alsa-devel, linux-serial, Paul Mundt, linux-usb

The current renesas_usbhs driver triggers

BUG: scheduling while atomic: ksoftirqd/0/3/0x00000102

with enabled CONFIG_DEBUG_ATOMIC_SLEEP, by submitting DMA transfers from 
an atomic (tasklet) context, which is not supported by the shdma dmaengine 
driver. Fix it by switching to a work. Also simplify some list 
manipulations.

Signed-off-by: Guennadi Liakhovetski <g.liakhovetski@gmx.de>
---

Shimoda-san, this is the problem, that you were observing. However, it 
exists with the present version of shdma just as well as with the new one 
- on top of the simple DMA library. I marked it an RFC because (1) I only 
lightly tested it with the gadget device on mackerel with the mass storage 
gadget, and (2) I am somewhat concerned about races. Currently the work 
function runs with no locking and there are no usual cancel_work_sync() 
points in the patch. However, it has also been like this before with the 
tasklet implementation, which is not much better, and it looks like there 
are no asynchronous operations on the same packets like timeouts. Only 
asynchronous events, that I can think about are things like unloading the 
driver or unplugging the cable, but these have been there before too. It 
would become worse on SMP, I think. Comments welcome.

diff --git a/drivers/usb/renesas_usbhs/fifo.c b/drivers/usb/renesas_usbhs/fifo.c
index 72339bd..4d739ec 100644
--- a/drivers/usb/renesas_usbhs/fifo.c
+++ b/drivers/usb/renesas_usbhs/fifo.c
@@ -75,8 +75,7 @@ void usbhs_pkt_push(struct usbhs_pipe *pipe, struct usbhs_pkt *pkt,
 		pipe->handler = &usbhsf_null_handler;
 	}
 
-	list_del_init(&pkt->node);
-	list_add_tail(&pkt->node, &pipe->list);
+	list_move_tail(&pkt->node, &pipe->list);
 
 	/*
 	 * each pkt must hold own handler.
@@ -106,7 +105,7 @@ static struct usbhs_pkt *__usbhsf_pkt_get(struct usbhs_pipe *pipe)
 	if (list_empty(&pipe->list))
 		return NULL;
 
-	return list_entry(pipe->list.next, struct usbhs_pkt, node);
+	return list_first_entry(&pipe->list, struct usbhs_pkt, node);
 }
 
 struct usbhs_pkt *usbhs_pkt_pop(struct usbhs_pipe *pipe, struct usbhs_pkt *pkt)
@@ -762,9 +761,9 @@ static int __usbhsf_dma_map_ctrl(struct usbhs_pkt *pkt, int map)
 }
 
 static void usbhsf_dma_complete(void *arg);
-static void usbhsf_dma_prepare_tasklet(unsigned long data)
+static void xfer_work(struct work_struct *work)
 {
-	struct usbhs_pkt *pkt = (struct usbhs_pkt *)data;
+	struct usbhs_pkt *pkt = container_of(work, struct usbhs_pkt, work);
 	struct usbhs_pipe *pipe = pkt->pipe;
 	struct usbhs_fifo *fifo = usbhs_pipe_to_fifo(pipe);
 	struct usbhs_priv *priv = usbhs_pipe_to_priv(pipe);
@@ -844,11 +843,8 @@ static int usbhsf_dma_prepare_push(struct usbhs_pkt *pkt, int *is_done)
 
 	pkt->trans = len;
 
-	tasklet_init(&fifo->tasklet,
-		     usbhsf_dma_prepare_tasklet,
-		     (unsigned long)pkt);
-
-	tasklet_schedule(&fifo->tasklet);
+	INIT_WORK(&pkt->work, xfer_work);
+	schedule_work(&pkt->work);
 
 	return 0;
 
@@ -938,11 +934,8 @@ static int usbhsf_dma_try_pop(struct usbhs_pkt *pkt, int *is_done)
 
 	pkt->trans = len;
 
-	tasklet_init(&fifo->tasklet,
-		     usbhsf_dma_prepare_tasklet,
-		     (unsigned long)pkt);
-
-	tasklet_schedule(&fifo->tasklet);
+	INIT_WORK(&pkt->work, xfer_work);
+	schedule_work(&pkt->work);
 
 	return 0;
 
diff --git a/drivers/usb/renesas_usbhs/fifo.h b/drivers/usb/renesas_usbhs/fifo.h
index f68609c..c31731a 100644
--- a/drivers/usb/renesas_usbhs/fifo.h
+++ b/drivers/usb/renesas_usbhs/fifo.h
@@ -19,6 +19,7 @@
 
 #include <linux/interrupt.h>
 #include <linux/sh_dma.h>
+#include <linux/workqueue.h>
 #include <asm/dma.h>
 #include "pipe.h"
 
@@ -31,7 +32,6 @@ struct usbhs_fifo {
 	u32 ctr;	/* xFIFOCTR */
 
 	struct usbhs_pipe	*pipe;
-	struct tasklet_struct	tasklet;
 
 	struct dma_chan		*tx_chan;
 	struct dma_chan		*rx_chan;
@@ -53,6 +53,7 @@ struct usbhs_pkt {
 	struct usbhs_pkt_handle *handler;
 	void (*done)(struct usbhs_priv *priv,
 		     struct usbhs_pkt *pkt);
+	struct work_struct work;
 	dma_addr_t dma;
 	void *buf;
 	int length;

^ permalink raw reply related	[flat|nested] 33+ messages in thread

* Re: [PATCH 1/7 v2] dmaengine: add a simple dma library
  2012-02-01  5:38           ` Vinod Koul
@ 2012-02-03 16:38             ` Guennadi Liakhovetski
  0 siblings, 0 replies; 33+ messages in thread
From: Guennadi Liakhovetski @ 2012-02-03 16:38 UTC (permalink / raw)
  To: Vinod Koul
  Cc: linux-kernel, linux-sh, Magnus Damm, Yoshihiro Shimoda,
	linux-mmc, alsa-devel, linux-serial, Paul Mundt

Hi Vinod

On Wed, 1 Feb 2012, Vinod Koul wrote:

> On Tue, 2012-01-31 at 09:59 +0100, Guennadi Liakhovetski wrote:
> > Hi Vinod
> > 
> > Thanks for your reply.
> > 
> > On Tue, 31 Jan 2012, Vinod Koul wrote:
> > 
> > > On Mon, 2012-01-30 at 10:34 +0100, Guennadi Liakhovetski wrote:

[snip]

> > > > How would you like to do this? Don't you think, it would be good to allow 
> > > > both: either implement a dmaengine driver directly, exactly as all drivers 
> > > > are doing now, or use the additional helper library for suitable (simple) 
> > > > hardware types? I see it similar to I2C, where you either implement an I2C 
> > > > driver directly, or you use the bitbanging abstraction for simpler 
> > > > hardware.
> > > I think it would be good to have both, this can be used by folks who
> > > don't have sg support available.
> > 
> > Ok, so, should I just rename the driver to sgsoft? Is this the only 
> > change, that you'd like to see?
> That would be one :)
> Also, I would review the other patch patches by today/tomorrow, you can
> add any changes in next version as well

Do I understand it right, that I'm waiting for your review before 
submitting the next version?

Thanks
Guennadi
---
Guennadi Liakhovetski, Ph.D.
Freelance Open-Source Software Developer
http://www.open-technology.de/

^ permalink raw reply	[flat|nested] 33+ messages in thread

* Re: [PATCH/RFC] usb: fix renesas_usbhs to not schedule in atomic context
  2012-02-03 15:43             ` [PATCH/RFC] usb: fix renesas_usbhs to not schedule in atomic context Guennadi Liakhovetski
@ 2012-02-05 14:54               ` Felipe Balbi
  2012-02-06 10:11                 ` Guennadi Liakhovetski
  2012-02-06  8:52               ` Shimoda, Yoshihiro
  1 sibling, 1 reply; 33+ messages in thread
From: Felipe Balbi @ 2012-02-05 14:54 UTC (permalink / raw)
  To: Guennadi Liakhovetski
  Cc: Shimoda, Yoshihiro, linux-kernel, linux-sh, Vinod Koul,
	Magnus Damm, linux-mmc, alsa-devel, linux-serial, Paul Mundt,
	linux-usb

[-- Attachment #1: Type: text/plain, Size: 2450 bytes --]

Hi,

On Fri, Feb 03, 2012 at 04:43:20PM +0100, Guennadi Liakhovetski wrote:
> The current renesas_usbhs driver triggers
> 
> BUG: scheduling while atomic: ksoftirqd/0/3/0x00000102
> 
> with enabled CONFIG_DEBUG_ATOMIC_SLEEP, by submitting DMA transfers from 
> an atomic (tasklet) context, which is not supported by the shdma dmaengine 
> driver. Fix it by switching to a work. Also simplify some list 
> manipulations.

you are doing much more than what you say. Also, instead of using a
workqueue, have you considered using threaded_irqs ?

(I didn't go over the driver again to see if it makes sense to use
threaded_irqs in this case, but doesn't hurt asking)

> Shimoda-san, this is the problem, that you were observing. However, it 
> exists with the present version of shdma just as well as with the new one 
> - on top of the simple DMA library. I marked it an RFC because (1) I only 
> lightly tested it with the gadget device on mackerel with the mass storage 
> gadget, and (2) I am somewhat concerned about races. Currently the work 
> function runs with no locking and there are no usual cancel_work_sync() 
> points in the patch. However, it has also been like this before with the 
> tasklet implementation, which is not much better, and it looks like there 
> are no asynchronous operations on the same packets like timeouts. Only 
> asynchronous events, that I can think about are things like unloading the 
> driver or unplugging the cable, but these have been there before too. It 
> would become worse on SMP, I think. Comments welcome.
> 
> diff --git a/drivers/usb/renesas_usbhs/fifo.c b/drivers/usb/renesas_usbhs/fifo.c
> index 72339bd..4d739ec 100644
> --- a/drivers/usb/renesas_usbhs/fifo.c
> +++ b/drivers/usb/renesas_usbhs/fifo.c
> @@ -75,8 +75,7 @@ void usbhs_pkt_push(struct usbhs_pipe *pipe, struct usbhs_pkt *pkt,
>  		pipe->handler = &usbhsf_null_handler;
>  	}
>  
> -	list_del_init(&pkt->node);
> -	list_add_tail(&pkt->node, &pipe->list);
> +	list_move_tail(&pkt->node, &pipe->list);
>  
>  	/*
>  	 * each pkt must hold own handler.
> @@ -106,7 +105,7 @@ static struct usbhs_pkt *__usbhsf_pkt_get(struct usbhs_pipe *pipe)
>  	if (list_empty(&pipe->list))
>  		return NULL;
>  
> -	return list_entry(pipe->list.next, struct usbhs_pkt, node);
> +	return list_first_entry(&pipe->list, struct usbhs_pkt, node);

these two hunks are not part of $SUBJECT

-- 
balbi

[-- Attachment #2: Digital signature --]
[-- Type: application/pgp-signature, Size: 836 bytes --]

^ permalink raw reply	[flat|nested] 33+ messages in thread

* Re: [PATCH 6/7 v2] ASoC: SIU: prepare for conversion to simple DMA
  2012-01-26 14:56 ` [PATCH 6/7 v2] ASoC: SIU: " Guennadi Liakhovetski
@ 2012-02-06  2:33   ` Vinod Koul
  2012-02-06  9:11     ` Guennadi Liakhovetski
  0 siblings, 1 reply; 33+ messages in thread
From: Vinod Koul @ 2012-02-06  2:33 UTC (permalink / raw)
  To: Guennadi Liakhovetski
  Cc: linux-kernel, linux-sh, Magnus Damm, Yoshihiro Shimoda,
	linux-mmc, alsa-devel, linux-serial, Paul Mundt

On Thu, 2012-01-26 at 15:56 +0100, Guennadi Liakhovetski wrote:
> Signed-off-by: Guennadi Liakhovetski <g.liakhovetski@gmx.de>
> ---
>  sound/soc/sh/siu_pcm.c |    8 ++++----
>  1 files changed, 4 insertions(+), 4 deletions(-)
> 
> diff --git a/sound/soc/sh/siu_pcm.c b/sound/soc/sh/siu_pcm.c
> index 0193e59..414bb35 100644
> --- a/sound/soc/sh/siu_pcm.c
> +++ b/sound/soc/sh/siu_pcm.c
> @@ -330,12 +330,12 @@ static bool filter(struct dma_chan *chan, void *slave)
>  {
>  	struct sh_dmae_slave *param = slave;
>  
> -	pr_debug("%s: slave ID %d\n", __func__, param->slave_id);
> +	pr_debug("%s: slave ID %d\n", __func__, param->simple_slave.slave_id);
>  
>  	if (unlikely(param->dma_dev != chan->device->dev))
>  		return false;
>  
> -	chan->private = param;
> +	chan->private = &param->simple_slave;
While at it, Can you fix the sh-mobile driver to NOT use chan->private?
>  	return true;
>  }
>  
> @@ -360,12 +360,12 @@ static int siu_pcm_open(struct snd_pcm_substream *ss)
>  	if (ss->stream == SNDRV_PCM_STREAM_PLAYBACK) {
>  		siu_stream = &port_info->playback;
>  		param = &siu_stream->param;
> -		param->slave_id = port ? pdata->dma_slave_tx_b :
> +		param->simple_slave.slave_id = port ? pdata->dma_slave_tx_b :
>  			pdata->dma_slave_tx_a;
>  	} else {
>  		siu_stream = &port_info->capture;
>  		param = &siu_stream->param;
> -		param->slave_id = port ? pdata->dma_slave_rx_b :
> +		param->simple_slave.slave_id = port ? pdata->dma_slave_rx_b :
>  			pdata->dma_slave_rx_a;
>  	}
>  


-- 
~Vinod


^ permalink raw reply	[flat|nested] 33+ messages in thread

* Re: [PATCH 1/7 v2] dmaengine: add a simple dma library
  2012-01-26 14:56 ` [PATCH 1/7 v2] dmaengine: add a simple dma library Guennadi Liakhovetski
                     ` (2 preceding siblings ...)
  2012-01-30  8:32   ` [PATCH 1/7 v2] dmaengine: add a simple dma library Vinod Koul
@ 2012-02-06  2:54   ` Vinod Koul
  2012-02-06  9:53     ` Guennadi Liakhovetski
  3 siblings, 1 reply; 33+ messages in thread
From: Vinod Koul @ 2012-02-06  2:54 UTC (permalink / raw)
  To: Guennadi Liakhovetski
  Cc: linux-kernel, linux-sh, Magnus Damm, Yoshihiro Shimoda,
	linux-mmc, alsa-devel, linux-serial, Paul Mundt

On Thu, 2012-01-26 at 15:56 +0100, Guennadi Liakhovetski wrote:
> This patch adds a library of functions, helping to implement dmaengine
> drivers for hardware, unable to handle scatter-gather lists natively.
> The first version of this driver only supports memcpy and slave DMA
> operation.
> 
> Signed-off-by: Guennadi Liakhovetski <g.liakhovetski@gmx.de>
> ---
> 
> v2:
> 
> 1. switch from using a tasklet to threaded IRQ, which allowed to
...?
> 2. remove lock / unlock inline functions
> 3. remove __devinit, __devexit annotations
> 
>  drivers/dma/Kconfig        |    3 +
>  drivers/dma/Makefile       |    1 +
>  drivers/dma/dma-simple.c   |  873 ++++++++++++++++++++++++++++++++++++++++++++
>  include/linux/dma-simple.h |  124 +++++++
>  4 files changed, 1001 insertions(+), 0 deletions(-)
>  create mode 100644 drivers/dma/dma-simple.c
>  create mode 100644 include/linux/dma-simple.h
> 
> diff --git a/drivers/dma/Kconfig b/drivers/dma/Kconfig
> index f1a2749..f7c583e 100644
> --- a/drivers/dma/Kconfig
> +++ b/drivers/dma/Kconfig
> @@ -149,6 +149,9 @@ config TXX9_DMAC
>  	  Support the TXx9 SoC internal DMA controller.  This can be
>  	  integrated in chips such as the Toshiba TX4927/38/39.
>  
> +config DMA_SIMPLE
> +	tristate
> +
>  config SH_DMAE
>  	tristate "Renesas SuperH DMAC support"
>  	depends on (SUPERH && SH_DMA) || (ARM && ARCH_SHMOBILE)
> diff --git a/drivers/dma/Makefile b/drivers/dma/Makefile
> index 009a222..d63f773 100644
> --- a/drivers/dma/Makefile
> +++ b/drivers/dma/Makefile
> @@ -2,6 +2,7 @@ ccflags-$(CONFIG_DMADEVICES_DEBUG)  := -DDEBUG
>  ccflags-$(CONFIG_DMADEVICES_VDEBUG) += -DVERBOSE_DEBUG
>  
>  obj-$(CONFIG_DMA_ENGINE) += dmaengine.o
> +obj-$(CONFIG_DMA_SIMPLE) += dma-simple.o
>  obj-$(CONFIG_NET_DMA) += iovlock.o
>  obj-$(CONFIG_INTEL_MID_DMAC) += intel_mid_dma.o
>  obj-$(CONFIG_DMATEST) += dmatest.o
> diff --git a/drivers/dma/dma-simple.c b/drivers/dma/dma-simple.c
> new file mode 100644
> index 0000000..49d8f7d
> --- /dev/null
> +++ b/drivers/dma/dma-simple.c
> @@ -0,0 +1,873 @@
> +/*
> + * Simple dmaengine driver library
> + *
> + * extracted from shdma.c
> + *
> + * Copyright (C) 2011-2012 Guennadi Liakhovetski <g.liakhovetski@gmx.de>
> + * Copyright (C) 2009 Nobuhiro Iwamatsu <iwamatsu.nobuhiro@renesas.com>
> + * Copyright (C) 2009 Renesas Solutions, Inc. All rights reserved.
> + * Copyright (C) 2007 Freescale Semiconductor, Inc. All rights reserved.
> + *
> + * This is free software; you can redistribute it and/or modify
> + * it under the terms of version 2 of the GNU General Public License as
> + * published by the Free Software Foundation.
> + */
> +
> +#include <linux/delay.h>
> +#include <linux/dma-simple.h>
> +#include <linux/dmaengine.h>
> +#include <linux/init.h>
> +#include <linux/interrupt.h>
> +#include <linux/module.h>
> +#include <linux/pm_runtime.h>
> +#include <linux/slab.h>
> +#include <linux/spinlock.h>
> +
> +/* DMA descriptor control */
> +enum simple_desc_status {
> +	DESC_IDLE,
> +	DESC_PREPARED,
> +	DESC_SUBMITTED,
> +	DESC_COMPLETED,	/* completed, have to call callback */
> +	DESC_WAITING,	/* callback called, waiting for ack / re-submit */
> +};
why do you need to keep track of descriptor status?
> +
> +#define NR_DESCS_PER_CHANNEL 32
> +
> +#define to_simple_chan(c) container_of(c, struct dma_simple_chan, dma_chan)
> +#define to_simple_dev(d) container_of(d, struct dma_simple_dev, dma_dev)
> +
> +/*
> + * For slave DMA we assume, that there is a finite number of DMA slaves in the
> + * system, and that each such slave can only use a finite number of channels.
> + * We use slave channel IDs to make sure, that no such slave channel ID is
> + * allocated more than once.
> + */
> +static unsigned int slave_num = 256;
> +module_param(slave_num, uint, 0444);
> +
> +/* A bitmask with slave_num bits */
> +static unsigned long *simple_slave_used;
> +
> +/* Called under spin_lock_irq(&schan->chan_lock") */
> +static void simple_chan_xfer_ld_queue(struct dma_simple_chan *schan)
> +{
> +	struct dma_simple_dev *sdev = to_simple_dev(schan->dma_chan.device);
> +	const struct dma_simple_ops *ops = sdev->ops;
> +	struct dma_simple_desc *sdesc;
> +
> +	/* DMA work check */
> +	if (ops->channel_busy(schan))
> +		return;
> +
> +	/* Find the first not transferred descriptor */
> +	list_for_each_entry(sdesc, &schan->ld_queue, node)
> +		if (sdesc->mark == DESC_SUBMITTED) {
> +			ops->start_xfer(schan, sdesc);
> +			break;
> +		}
> +}
> +
> +static dma_cookie_t simple_tx_submit(struct dma_async_tx_descriptor *tx)
> +{
> +	struct dma_simple_desc *chunk, *c, *desc =
> +		container_of(tx, struct dma_simple_desc, async_tx),
> +		*last = desc;
> +	struct dma_simple_chan *schan = to_simple_chan(tx->chan);
> +	struct dma_simple_slave *slave = tx->chan->private;
prive is masked depricated, os this needs to be removed.
Any slave config should be extracted from dma_lsave_config only... Do
you anything more than which is provided there??

> +	dma_async_tx_callback callback = tx->callback;
> +	dma_cookie_t cookie;
> +	bool power_up;
> +
> +	spin_lock_irq(&schan->chan_lock);
> +
> +	power_up = list_empty(&schan->ld_queue);
> +
> +	cookie = schan->dma_chan.cookie + 1;
> +	if (cookie < 0)
> +		cookie = 1;
> +
> +	schan->dma_chan.cookie = cookie;
> +	tx->cookie = cookie;
> +
> +	/* Mark all chunks of this descriptor as submitted, move to the queue */
> +	list_for_each_entry_safe(chunk, c, desc->node.prev, node) {
> +		/*
> +		 * All chunks are on the global ld_free, so, we have to find
> +		 * the end of the chain ourselves
> +		 */
> +		if (chunk != desc && (chunk->mark == DESC_IDLE ||
> +				      chunk->async_tx.cookie > 0 ||
> +				      chunk->async_tx.cookie == -EBUSY ||
> +				      &chunk->node == &schan->ld_free))
> +			break;
> +		chunk->mark = DESC_SUBMITTED;
> +		/* Callback goes to the last chunk */
> +		chunk->async_tx.callback = NULL;
> +		chunk->cookie = cookie;
> +		list_move_tail(&chunk->node, &schan->ld_queue);
> +		last = chunk;
> +
> +		dev_dbg(schan->dev, "submit #%d@%p on %d\n",
> +			tx->cookie, &last->async_tx, schan->id);
> +	}
> +
> +	last->async_tx.callback = callback;
> +	last->async_tx.callback_param = tx->callback_param;
> +
> +	if (power_up) {
> +		int ret;
> +		schan->pm_state = DMA_SIMPLE_PM_BUSY;
> +
> +		ret = pm_runtime_get(schan->dev);
> +
> +		spin_unlock_irq(&schan->chan_lock);
> +		if (ret < 0)
> +			dev_err(schan->dev, "%s(): GET = %d\n", __func__, ret);
> +
> +		pm_runtime_barrier(schan->dev);
> +
> +		spin_lock_irq(&schan->chan_lock);
> +
> +		/* Have we been reset, while waiting? */
> +		if (schan->pm_state != DMA_SIMPLE_PM_ESTABLISHED) {
> +			struct dma_simple_dev *sdev =
> +				to_simple_dev(schan->dma_chan.device);
> +			const struct dma_simple_ops *ops = sdev->ops;
> +			dev_dbg(schan->dev, "Bring up channel %d\n",
> +				schan->id);
> +			/*
> +			 * TODO: .xfer_setup() might fail on some platforms.
> +			 * Make it int then, on error remove chunks from the
> +			 * queue again
> +			 */
> +			ops->setup_xfer(schan, slave);
> +
> +			if (schan->pm_state == DMA_SIMPLE_PM_PENDING)
> +				simple_chan_xfer_ld_queue(schan);
> +			schan->pm_state = DMA_SIMPLE_PM_ESTABLISHED;
> +		}
> +	} else {
> +		/*
> +		 * Tell .device_issue_pending() not to run the queue, interrupts
> +		 * will do it anyway
> +		 */
> +		schan->pm_state = DMA_SIMPLE_PM_PENDING;
> +	}
> +
> +	spin_unlock_irq(&schan->chan_lock);
> +
> +	return cookie;
> +}
> +
> +/* Called with desc_lock held */
> +static struct dma_simple_desc *simple_get_desc(struct dma_simple_chan *schan)
> +{
> +	struct dma_simple_desc *sdesc;
> +
> +	list_for_each_entry(sdesc, &schan->ld_free, node)
> +		if (sdesc->mark != DESC_PREPARED) {
> +			BUG_ON(sdesc->mark != DESC_IDLE);
> +			list_del(&sdesc->node);
> +			return sdesc;
> +		}
> +
> +	return NULL;
> +}
> +
> +static int simple_alloc_chan_resources(struct dma_chan *chan)
> +{
> +	struct dma_simple_chan *schan = to_simple_chan(chan);
> +	struct dma_simple_dev *sdev = to_simple_dev(schan->dma_chan.device);
> +	const struct dma_simple_ops *ops = sdev->ops;
> +	struct dma_simple_desc *desc;
> +	struct dma_simple_slave *slave = chan->private;
> +	int ret, i;
> +
> +	/*
> +	 * This relies on the guarantee from dmaengine that alloc_chan_resources
> +	 * never runs concurrently with itself or free_chan_resources.
> +	 */
> +	if (slave) {
> +		if (test_and_set_bit(slave->slave_id, simple_slave_used)) {
> +			ret = -EBUSY;
> +			goto etestused;
> +		}
> +
> +		ret = ops->set_slave(schan, slave);
> +		if (ret < 0)
> +			goto esetslave;
> +	}
> +
> +	schan->desc = kcalloc(NR_DESCS_PER_CHANNEL,
> +			      sdev->desc_size, GFP_KERNEL);
> +	if (!schan->desc) {
> +		ret = -ENOMEM;
> +		goto edescalloc;
> +	}
> +	schan->desc_num = NR_DESCS_PER_CHANNEL;
> +
> +	for (i = 0; i < NR_DESCS_PER_CHANNEL; i++) {
> +		desc = ops->embedded_desc(schan->desc, i);
> +		dma_async_tx_descriptor_init(&desc->async_tx,
> +					     &schan->dma_chan);
> +		desc->async_tx.tx_submit = simple_tx_submit;
> +		desc->mark = DESC_IDLE;
> +
> +		list_add(&desc->node, &schan->ld_free);
> +	}
> +
> +	return NR_DESCS_PER_CHANNEL;
> +
> +edescalloc:
> +	if (slave)
> +esetslave:
> +		clear_bit(slave->slave_id, simple_slave_used);
> +etestused:
> +	chan->private = NULL;
> +	return ret;
> +}
Typically chan allocation involves some kind of hand shaking between the
client and dmac which typically is arch specfic. If we want to make this
a truly independent library, then I think we should move allocation to
driver and let them allocate the required channel. The library by
definition is to _help_ for sg transfers so it should just create a
library of APIs to call which manage the sg transfer when not supported
by dmac.
> +
> +static dma_async_tx_callback __ld_cleanup(struct dma_simple_chan *schan, bool all)
> +{
> +	struct dma_simple_desc *desc, *_desc;
> +	/* Is the "exposed" head of a chain acked? */
> +	bool head_acked = false;
> +	dma_cookie_t cookie = 0;
> +	dma_async_tx_callback callback = NULL;
> +	void *param = NULL;
> +	unsigned long flags;
> +
> +	spin_lock_irqsave(&schan->chan_lock, flags);
> +	list_for_each_entry_safe(desc, _desc, &schan->ld_queue, node) {
> +		struct dma_async_tx_descriptor *tx = &desc->async_tx;
> +
> +		BUG_ON(tx->cookie > 0 && tx->cookie != desc->cookie);
> +		BUG_ON(desc->mark != DESC_SUBMITTED &&
> +		       desc->mark != DESC_COMPLETED &&
> +		       desc->mark != DESC_WAITING);
> +
> +		/*
> +		 * queue is ordered, and we use this loop to (1) clean up all
> +		 * completed descriptors, and to (2) update descriptor flags of
> +		 * any chunks in a (partially) completed chain
> +		 */
> +		if (!all && desc->mark == DESC_SUBMITTED &&
> +		    desc->cookie != cookie)
> +			break;
> +
> +		if (tx->cookie > 0)
> +			cookie = tx->cookie;
> +
> +		if (desc->mark == DESC_COMPLETED && desc->chunks == 1) {
> +			if (schan->completed_cookie != desc->cookie - 1)
> +				dev_dbg(schan->dev,
> +					"Completing cookie %d, expected %d\n",
> +					desc->cookie,
> +					schan->completed_cookie + 1);
> +			schan->completed_cookie = desc->cookie;
> +		}
> +
> +		/* Call callback on the last chunk */
> +		if (desc->mark == DESC_COMPLETED && tx->callback) {
> +			desc->mark = DESC_WAITING;
> +			callback = tx->callback;
> +			param = tx->callback_param;
> +			dev_dbg(schan->dev, "descriptor #%d@%p on %d callback\n",
> +				tx->cookie, tx, schan->id);
> +			BUG_ON(desc->chunks != 1);
> +			break;
> +		}
> +
> +		if (tx->cookie > 0 || tx->cookie == -EBUSY) {
> +			if (desc->mark == DESC_COMPLETED) {
> +				BUG_ON(tx->cookie < 0);
> +				desc->mark = DESC_WAITING;
> +			}
> +			head_acked = async_tx_test_ack(tx);
> +		} else {
> +			switch (desc->mark) {
> +			case DESC_COMPLETED:
> +				desc->mark = DESC_WAITING;
> +				/* Fall through */
> +			case DESC_WAITING:
> +				if (head_acked)
> +					async_tx_ack(&desc->async_tx);
> +			}
> +		}
> +
> +		dev_dbg(schan->dev, "descriptor %p #%d completed.\n",
> +			tx, tx->cookie);
> +
> +		if (((desc->mark == DESC_COMPLETED ||
> +		      desc->mark == DESC_WAITING) &&
> +		     async_tx_test_ack(&desc->async_tx)) || all) {
> +			/* Remove from ld_queue list */
> +			desc->mark = DESC_IDLE;
> +
> +			list_move(&desc->node, &schan->ld_free);
> +
> +			if (list_empty(&schan->ld_queue)) {
> +				dev_dbg(schan->dev, "Bring down channel %d\n", schan->id);
> +				pm_runtime_put(schan->dev);
> +				schan->pm_state = DMA_SIMPLE_PM_ESTABLISHED;
> +			}
> +		}
> +	}
> +
> +	if (all && !callback)
> +		/*
> +		 * Terminating and the loop completed normally: forgive
> +		 * uncompleted cookies
> +		 */
> +		schan->completed_cookie = schan->dma_chan.cookie;
> +
> +	spin_unlock_irqrestore(&schan->chan_lock, flags);
> +
> +	if (callback)
> +		callback(param);
> +
> +	return callback;
> +}
> +
> +/*
> + * simple_chan_ld_cleanup - Clean up link descriptors
> + *
> + * Clean up the ld_queue of DMA channel.
> + */
> +static void simple_chan_ld_cleanup(struct dma_simple_chan *schan, bool all)
> +{
> +	while (__ld_cleanup(schan, all))
> +		;
> +}
> +
> +/*
> + * simple_free_chan_resources - Free all resources of the channel.
> + */
> +static void simple_free_chan_resources(struct dma_chan *chan)
> +{
> +	struct dma_simple_chan *schan = to_simple_chan(chan);
> +	struct dma_simple_dev *sdev = to_simple_dev(chan->device);
> +	const struct dma_simple_ops *ops = sdev->ops;
> +	LIST_HEAD(list);
> +
> +	/* Protect against ISR */
> +	spin_lock_irq(&schan->chan_lock);
> +	ops->halt_channel(schan);
> +	spin_unlock_irq(&schan->chan_lock);
> +
> +	/* Now no new interrupts will occur */
> +
> +	/* Prepared and not submitted descriptors can still be on the queue */
> +	if (!list_empty(&schan->ld_queue))
> +		simple_chan_ld_cleanup(schan, true);
> +
> +	if (chan->private) {
> +		/* The caller is holding dma_list_mutex */
> +		struct dma_simple_slave *slave = chan->private;
> +		clear_bit(slave->slave_id, simple_slave_used);
> +		chan->private = NULL;
> +	}
> +
> +	spin_lock_irq(&schan->chan_lock);
> +
> +	list_splice_init(&schan->ld_free, &list);
> +	schan->desc_num = 0;
> +
> +	spin_unlock_irq(&schan->chan_lock);
> +
> +	kfree(schan->desc);
> +}
> +
> +/**
> + * simple_add_desc - get, set up and return one transfer descriptor
> + * @schan:	DMA channel
> + * @flags:	DMA transfer flags
> + * @dst:	destination DMA address, incremented when direction equals
> + *		DMA_DEV_TO_MEM or DMA_MEM_TO_MEM
> + * @src:	source DMA address, incremented when direction equals
> + *		DMA_MEM_TO_DEV or DMA_MEM_TO_MEM
> + * @len:	DMA transfer length
> + * @first:	if NULL, set to the current descriptor and cookie set to -EBUSY
> + * @direction:	needed for slave DMA to decide which address to keep constant,
> + *		equals DMA_MEM_TO_MEM for MEMCPY
> + * Returns 0 or an error
> + * Locks: called with desc_lock held
> + */
> +static struct dma_simple_desc *simple_add_desc(struct dma_simple_chan *schan,
> +	unsigned long flags, dma_addr_t *dst, dma_addr_t *src, size_t *len,
> +	struct dma_simple_desc **first, enum dma_transfer_direction direction)
> +{
> +	struct dma_simple_dev *sdev = to_simple_dev(schan->dma_chan.device);
> +	const struct dma_simple_ops *ops = sdev->ops;
> +	struct dma_simple_desc *new;
> +	size_t copy_size = *len;
> +
> +	if (!copy_size)
> +		return NULL;
> +
> +	/* Allocate the link descriptor from the free list */
> +	new = simple_get_desc(schan);
> +	if (!new) {
> +		dev_err(schan->dev, "No free link descriptor available\n");
> +		return NULL;
> +	}
> +
> +	ops->desc_setup(schan, new, *src, *dst, &copy_size);
> +
> +	if (!*first) {
> +		/* First desc */
> +		new->async_tx.cookie = -EBUSY;
> +		*first = new;
> +	} else {
> +		/* Other desc - invisible to the user */
> +		new->async_tx.cookie = -EINVAL;
> +	}
> +
> +	dev_dbg(schan->dev,
> +		"chaining (%u/%u)@%x -> %x with %p, cookie %d\n",
> +		copy_size, *len, *src, *dst, &new->async_tx,
> +		new->async_tx.cookie);
> +
> +	new->mark = DESC_PREPARED;
> +	new->async_tx.flags = flags;
> +	new->direction = direction;
> +
> +	*len -= copy_size;
> +	if (direction == DMA_MEM_TO_MEM || direction == DMA_MEM_TO_DEV)
> +		*src += copy_size;
> +	if (direction == DMA_MEM_TO_MEM || direction == DMA_DEV_TO_MEM)
> +		*dst += copy_size;
> +
> +	return new;
> +}
> +
> +/*
> + * simple_prep_sg - prepare transfer descriptors from an SG list
> + *
> + * Common routine for public (MEMCPY) and slave DMA. The MEMCPY case is also
> + * converted to scatter-gather to guarantee consistent locking and a correct
> + * list manipulation. For slave DMA direction carries the usual meaning, and,
> + * logically, the SG list is RAM and the addr variable contains slave address,
> + * e.g., the FIFO I/O register. For MEMCPY direction equals DMA_MEM_TO_MEM
> + * and the SG list contains only one element and points at the source buffer.
> + */
> +static struct dma_async_tx_descriptor *simple_prep_sg(struct dma_simple_chan *schan,
> +	struct scatterlist *sgl, unsigned int sg_len, dma_addr_t *addr,
> +	enum dma_transfer_direction direction, unsigned long flags)
> +{
> +	struct scatterlist *sg;
> +	struct dma_simple_desc *first = NULL, *new = NULL /* compiler... */;
> +	LIST_HEAD(tx_list);
> +	int chunks = 0;
> +	unsigned long irq_flags;
> +	int i;
> +
> +	for_each_sg(sgl, sg, sg_len, i)
> +		chunks += DIV_ROUND_UP(sg_dma_len(sg), schan->max_xfer_len);
> +
> +	/* Have to lock the whole loop to protect against concurrent release */
> +	spin_lock_irqsave(&schan->chan_lock, irq_flags);
> +
> +	/*
> +	 * Chaining:
> +	 * first descriptor is what user is dealing with in all API calls, its
> +	 *	cookie is at first set to -EBUSY, at tx-submit to a positive
> +	 *	number
> +	 * if more than one chunk is needed further chunks have cookie = -EINVAL
> +	 * the last chunk, if not equal to the first, has cookie = -ENOSPC
> +	 * all chunks are linked onto the tx_list head with their .node heads
> +	 *	only during this function, then they are immediately spliced
> +	 *	back onto the free list in form of a chain
> +	 */
> +	for_each_sg(sgl, sg, sg_len, i) {
> +		dma_addr_t sg_addr = sg_dma_address(sg);
> +		size_t len = sg_dma_len(sg);
> +
> +		if (!len)
> +			goto err_get_desc;
> +
> +		do {
> +			dev_dbg(schan->dev, "Add SG #%d@%p[%d], dma %llx\n",
> +				i, sg, len, (unsigned long long)sg_addr);
> +
> +			if (direction == DMA_DEV_TO_MEM)
> +				new = simple_add_desc(schan, flags,
> +						&sg_addr, addr, &len, &first,
> +						direction);
> +			else
> +				new = simple_add_desc(schan, flags,
> +						addr, &sg_addr, &len, &first,
> +						direction);
> +			if (!new)
> +				goto err_get_desc;
> +
> +			new->chunks = chunks--;
> +			list_add_tail(&new->node, &tx_list);
> +		} while (len);
> +	}
> +
> +	if (new != first)
> +		new->async_tx.cookie = -ENOSPC;
> +
> +	/* Put them back on the free list, so, they don't get lost */
> +	list_splice_tail(&tx_list, &schan->ld_free);
> +
> +	spin_unlock_irqrestore(&schan->chan_lock, irq_flags);
> +
> +	return &first->async_tx;
> +
> +err_get_desc:
> +	list_for_each_entry(new, &tx_list, node)
> +		new->mark = DESC_IDLE;
> +	list_splice(&tx_list, &schan->ld_free);
> +
> +	spin_unlock_irqrestore(&schan->chan_lock, irq_flags);
> +
> +	return NULL;
> +}
> +
> +static struct dma_async_tx_descriptor *simple_prep_memcpy(
> +	struct dma_chan *chan, dma_addr_t dma_dest, dma_addr_t dma_src,
> +	size_t len, unsigned long flags)
> +{
> +	struct dma_simple_chan *schan = to_simple_chan(chan);
> +	struct scatterlist sg;
> +
> +	if (!chan || !len)
> +		return NULL;
> +
> +	BUG_ON(!schan->desc_num);
> +
> +	sg_init_table(&sg, 1);
> +	sg_set_page(&sg, pfn_to_page(PFN_DOWN(dma_src)), len,
> +		    offset_in_page(dma_src));
> +	sg_dma_address(&sg) = dma_src;
> +	sg_dma_len(&sg) = len;
> +
> +	return simple_prep_sg(schan, &sg, 1, &dma_dest, DMA_MEM_TO_MEM, flags);
> +}
mempcy is a single transfer why should this go thru library?
got sg_memcpy yes, but otherwise NO

> +
> +static struct dma_async_tx_descriptor *simple_prep_slave_sg(
> +	struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len,
> +	enum dma_transfer_direction direction, unsigned long flags)
> +{
> +	struct dma_simple_chan *schan = to_simple_chan(chan);
> +	struct dma_simple_dev *sdev = to_simple_dev(schan->dma_chan.device);
> +	const struct dma_simple_ops *ops = sdev->ops;
> +	struct dma_simple_slave *slave = chan->private;
> +	dma_addr_t slave_addr;
> +
> +	if (!chan)
> +		return NULL;
> +
> +	BUG_ON(!schan->desc_num);
> +
> +	/* Someone calling slave DMA on a generic channel? */
> +	if (!slave || !sg_len) {
> +		dev_warn(schan->dev, "%s: bad parameter: %p, %d, %d\n",
> +			 __func__, slave, sg_len, slave ? slave->slave_id : -1);
> +		return NULL;
> +	}
> +
> +	slave_addr = ops->slave_addr(schan);
> +
> +	return simple_prep_sg(schan, sgl, sg_len, &slave_addr,
> +			      direction, flags);
> +}
> +
> +static int simple_control(struct dma_chan *chan, enum dma_ctrl_cmd cmd,
> +			  unsigned long arg)
> +{
> +	struct dma_simple_chan *schan = to_simple_chan(chan);
> +	struct dma_simple_dev *sdev = to_simple_dev(chan->device);
> +	const struct dma_simple_ops *ops = sdev->ops;
> +	unsigned long flags;
> +
> +	/* Only supports DMA_TERMINATE_ALL */
> +	if (cmd != DMA_TERMINATE_ALL)
> +		return -ENXIO;
nope, you should check from respective dmac...

> +
> +	if (!chan)
> +		return -EINVAL;
> +
> +	spin_lock_irqsave(&schan->chan_lock, flags);
> +
> +	ops->halt_channel(schan);
> +
> +	spin_unlock_irqrestore(&schan->chan_lock, flags);
> +
> +	simple_chan_ld_cleanup(schan, true);
> +
> +	return 0;
> +}
> +
> +static void simple_issue_pending(struct dma_chan *chan)
> +{
> +	struct dma_simple_chan *schan = to_simple_chan(chan);
> +
> +	spin_lock_irq(&schan->chan_lock);
> +	if (schan->pm_state == DMA_SIMPLE_PM_ESTABLISHED)
> +		simple_chan_xfer_ld_queue(schan);
> +	else
> +		schan->pm_state = DMA_SIMPLE_PM_PENDING;
> +	spin_unlock_irq(&schan->chan_lock);
> +}
> +
> +static enum dma_status simple_tx_status(struct dma_chan *chan,
> +					dma_cookie_t cookie,
> +					struct dma_tx_state *txstate)
> +{
> +	struct dma_simple_chan *schan = to_simple_chan(chan);
> +	dma_cookie_t last_used;
> +	dma_cookie_t last_complete;
> +	enum dma_status status;
> +	unsigned long flags;
> +
> +	simple_chan_ld_cleanup(schan, false);
> +
> +	/* First read completed cookie to avoid a skew */
> +	last_complete = schan->completed_cookie;
> +	rmb();
> +	last_used = chan->cookie;
> +	BUG_ON(last_complete < 0);
> +	dma_set_tx_state(txstate, last_complete, last_used, 0);
> +
> +	spin_lock_irqsave(&schan->chan_lock, flags);
> +
> +	status = dma_async_is_complete(cookie, last_complete, last_used);
> +
> +	/*
> +	 * If we don't find cookie on the queue, it has been aborted and we have
> +	 * to report error
> +	 */
> +	if (status != DMA_SUCCESS) {
> +		struct dma_simple_desc *sdesc;
> +		status = DMA_ERROR;
> +		list_for_each_entry(sdesc, &schan->ld_queue, node)
> +			if (sdesc->cookie == cookie) {
> +				status = DMA_IN_PROGRESS;
> +				break;
> +			}
> +	}
> +
> +	spin_unlock_irqrestore(&schan->chan_lock, flags);
> +
> +	return status;
> +}
> +
> +/* Called from error IRQ or NMI */
> +bool dma_simple_reset(struct dma_simple_dev *sdev)
> +{
> +	const struct dma_simple_ops *ops = sdev->ops;
> +	struct dma_simple_chan *schan;
> +	unsigned int handled = 0;
> +	int i;
> +
> +	/* Reset all channels */
> +	dma_simple_for_each_chan(schan, sdev, i) {
> +		struct dma_simple_desc *sdesc;
> +		LIST_HEAD(dl);
> +
> +		if (!schan)
> +			continue;
> +
> +		spin_lock(&schan->chan_lock);
> +
> +		/* Stop the channel */
> +		ops->halt_channel(schan);
> +
> +		list_splice_init(&schan->ld_queue, &dl);
> +
> +		if (!list_empty(&dl)) {
> +			dev_dbg(schan->dev, "Bring down channel %d\n", schan->id);
> +			pm_runtime_put(schan->dev);
> +		}
> +		schan->pm_state = DMA_SIMPLE_PM_ESTABLISHED;
> +
> +		spin_unlock(&schan->chan_lock);
> +
> +		/* Complete all  */
> +		list_for_each_entry(sdesc, &dl, node) {
> +			struct dma_async_tx_descriptor *tx = &sdesc->async_tx;
> +			sdesc->mark = DESC_IDLE;
> +			if (tx->callback)
> +				tx->callback(tx->callback_param);
> +		}
> +
> +		spin_lock(&schan->chan_lock);
> +		list_splice(&dl, &schan->ld_free);
> +		spin_unlock(&schan->chan_lock);
> +
> +		handled++;
> +	}
> +
> +	return !!handled;
> +}
> +EXPORT_SYMBOL(dma_simple_reset);
> +
> +static irqreturn_t chan_irq(int irq, void *dev)
> +{
> +	struct dma_simple_chan *schan = dev;
> +	const struct dma_simple_ops *ops =
> +		to_simple_dev(schan->dma_chan.device)->ops;
> +	irqreturn_t ret;
> +
> +	spin_lock(&schan->chan_lock);
> +
> +	ret = ops->chan_irq(schan, irq) ? IRQ_WAKE_THREAD : IRQ_NONE;
> +
> +	spin_unlock(&schan->chan_lock);
> +
> +	return ret;
> +}
> +
> +static irqreturn_t chan_irqt(int irq, void *dev)
> +{
> +	struct dma_simple_chan *schan = dev;
> +	const struct dma_simple_ops *ops =
> +		to_simple_dev(schan->dma_chan.device)->ops;
> +	struct dma_simple_desc *sdesc;
> +
> +	spin_lock_irq(&schan->chan_lock);
> +	list_for_each_entry(sdesc, &schan->ld_queue, node) {
> +		if (sdesc->mark == DESC_SUBMITTED &&
> +		    ops->desc_completed(schan, sdesc)) {
> +			dev_dbg(schan->dev, "done #%d@%p\n",
> +				sdesc->async_tx.cookie, &sdesc->async_tx);
> +			sdesc->mark = DESC_COMPLETED;
> +			break;
> +		}
> +	}
> +	/* Next desc */
> +	simple_chan_xfer_ld_queue(schan);
> +	spin_unlock_irq(&schan->chan_lock);
> +
> +	simple_chan_ld_cleanup(schan, false);
> +
> +	return IRQ_HANDLED;
> +}
> +
> +int dma_simple_request_irq(struct dma_simple_chan *schan, int irq,
> +			   unsigned long flags, const char *name)
> +{
> +	int ret = request_threaded_irq(irq, chan_irq, chan_irqt,
> +				       flags, name, schan);
> +
> +	schan->irq = ret < 0 ? ret : irq;
> +
> +	return ret;
> +}
> +EXPORT_SYMBOL(dma_simple_request_irq);
> +
> +void dma_simple_free_irq(struct dma_simple_chan *schan)
> +{
> +	if (schan->irq >= 0)
> +		free_irq(schan->irq, schan);
> +}
> +EXPORT_SYMBOL(dma_simple_free_irq);
why would you use the irq here?? That should be handled by respective
dmac. Also, the library should just setup callbacks for descriptor and
use them to manage the descriptors for sg mode.
> +
> +void dma_simple_chan_probe(struct dma_simple_dev *sdev,
> +			   struct dma_simple_chan *schan, int id)
> +{
> +	schan->pm_state = DMA_SIMPLE_PM_ESTABLISHED;
> +
> +	/* reference struct dma_device */
> +	schan->dma_chan.device = &sdev->dma_dev;
> +
> +	schan->dev = sdev->dma_dev.dev;
> +	schan->id = id;
> +
> +	if (!schan->max_xfer_len)
> +		schan->max_xfer_len = PAGE_SIZE;
> +
> +	spin_lock_init(&schan->chan_lock);
> +
> +	/* Init descripter manage list */
> +	INIT_LIST_HEAD(&schan->ld_queue);
> +	INIT_LIST_HEAD(&schan->ld_free);
> +
> +	/* Add the channel to DMA device channel list */
> +	list_add_tail(&schan->dma_chan.device_node,
> +			&sdev->dma_dev.channels);
> +	sdev->schan[sdev->dma_dev.chancnt++] = schan;
> +}
> +EXPORT_SYMBOL(dma_simple_chan_probe);
> +
> +void dma_simple_chan_remove(struct dma_simple_chan *schan)
> +{
> +	list_del(&schan->dma_chan.device_node);
> +}
> +EXPORT_SYMBOL(dma_simple_chan_remove);
> +
> +int dma_simple_init(struct device *dev, struct dma_simple_dev *sdev,
> +		    int chan_num)
> +{
> +	struct dma_device *dma_dev = &sdev->dma_dev;
> +
> +	/*
> +	 * Require all call-backs for now, they can trivially be made optional
> +	 * later as required
> +	 */
> +	if (!sdev->ops ||
> +	    !sdev->desc_size ||
> +	    !sdev->ops->embedded_desc ||
> +	    !sdev->ops->start_xfer ||
> +	    !sdev->ops->setup_xfer ||
> +	    !sdev->ops->set_slave ||
> +	    !sdev->ops->desc_setup ||
> +	    !sdev->ops->slave_addr ||
> +	    !sdev->ops->channel_busy ||
> +	    !sdev->ops->halt_channel ||
> +	    !sdev->ops->desc_completed)
> +		return -EINVAL;
> +
> +	sdev->schan = kcalloc(chan_num, sizeof(*sdev->schan), GFP_KERNEL);
> +	if (!sdev->schan)
> +		return -ENOMEM;
> +
> +	INIT_LIST_HEAD(&dma_dev->channels);
> +
> +	/* Common and MEMCPY operations */
> +	dma_dev->device_alloc_chan_resources
> +		= simple_alloc_chan_resources;
> +	dma_dev->device_free_chan_resources = simple_free_chan_resources;
> +	dma_dev->device_prep_dma_memcpy = simple_prep_memcpy;
> +	dma_dev->device_tx_status = simple_tx_status;
> +	dma_dev->device_issue_pending = simple_issue_pending;
> +
> +	/* Compulsory for DMA_SLAVE fields */
> +	dma_dev->device_prep_slave_sg = simple_prep_slave_sg;
> +	dma_dev->device_control = simple_control;
> +
> +	dma_dev->dev = dev;
> +
> +	return 0;
> +}
> +EXPORT_SYMBOL(dma_simple_init);
> +
> +void dma_simple_cleanup(struct dma_simple_dev *sdev)
> +{
> +	kfree(sdev->schan);
> +}
> +EXPORT_SYMBOL(dma_simple_cleanup);
> +
> +static int __init dma_simple_enter(void)
> +{
> +	simple_slave_used = kzalloc(DIV_ROUND_UP(slave_num, BITS_PER_LONG) *
> +				    sizeof(long), GFP_KERNEL);
> +	if (!simple_slave_used)
> +		return -ENOMEM;
> +	return 0;
> +}
> +module_init(dma_simple_enter);
> +
> +static void __exit dma_simple_exit(void)
> +{
> +	kfree(simple_slave_used);
> +}
> +module_exit(dma_simple_exit);
> +
> +MODULE_LICENSE("GPL v2");
> +MODULE_DESCRIPTION("Simple dmaengine driver library");
> +MODULE_AUTHOR("Guennadi Liakhovetski <g.liakhovetski@gmx.de>");
> diff --git a/include/linux/dma-simple.h b/include/linux/dma-simple.h
> new file mode 100644
> index 0000000..5336674
> --- /dev/null
> +++ b/include/linux/dma-simple.h
> @@ -0,0 +1,124 @@
> +/*
> + * Simple dmaengine driver library
> + *
> + * extracted from shdma.c and headers
> + *
> + * Copyright (C) 2011-2012 Guennadi Liakhovetski <g.liakhovetski@gmx.de>
> + * Copyright (C) 2009 Nobuhiro Iwamatsu <iwamatsu.nobuhiro@renesas.com>
> + * Copyright (C) 2009 Renesas Solutions, Inc. All rights reserved.
> + * Copyright (C) 2007 Freescale Semiconductor, Inc. All rights reserved.
> + *
> + * This is free software; you can redistribute it and/or modify
> + * it under the terms of version 2 of the GNU General Public License as
> + * published by the Free Software Foundation.
> + */
> +
> +#ifndef DMA_SIMPLE_H
> +#define DMA_SIMPLE_H
> +
> +#include <linux/dmaengine.h>
> +#include <linux/interrupt.h>
> +#include <linux/list.h>
> +#include <linux/types.h>
> +
> +/**
> + * dma_simple_pm_state - DMA channel PM state
> + * DMA_SIMPLE_PM_ESTABLISHED:	either idle or during data transfer
> + * DMA_SIMPLE_PM_BUSY:		during the transfer preparation, when we have to
> + *				drop the lock temporarily
> + * DMA_SIMPLE_PM_PENDING:	transfers pending
> + */
> +enum dma_simple_pm_state {
> +	DMA_SIMPLE_PM_ESTABLISHED,
> +	DMA_SIMPLE_PM_BUSY,
> +	DMA_SIMPLE_PM_PENDING,
> +};
> +
> +struct device;
> +
> +/*
> + * Drivers, using this library are expected to embed struct dma_simple_dev,
> + * struct dma_simple_chan, struct dma_simple_desc, and struct dma_simple_slave
> + * in their respective device, channel, descriptor and slave objects.
> + */
> +
> +struct dma_simple_slave {
> +	unsigned int slave_id;
> +};
> +
> +struct dma_simple_desc {
> +	struct list_head node;
> +	struct dma_async_tx_descriptor async_tx;
> +	enum dma_transfer_direction direction;
> +	dma_cookie_t cookie;
> +	int chunks;
> +	int mark;
> +};
> +
> +struct dma_simple_chan {
> +	dma_cookie_t completed_cookie;	/* The maximum cookie completed */
> +	spinlock_t chan_lock;		/* Channel operation lock */
> +	struct list_head ld_queue;	/* Link descriptors queue */
> +	struct list_head ld_free;	/* Free link descriptors */
> +	struct dma_chan dma_chan;	/* DMA channel */
> +	struct device *dev;		/* Channel device */
> +	void *desc;			/* buffer for descriptor array */
> +	int desc_num;			/* desc count */
> +	size_t max_xfer_len;		/* max transfer length */
> +	int id;				/* Raw id of this channel */
> +	int irq;			/* Channel IRQ */
> +	enum dma_simple_pm_state pm_state;
> +};
> +
> +/**
> + * struct dma_simple_ops - simple DMA driver operations
> + * desc_completed:	return true, if this is the descriptor, that just has
> + *			completed (atomic)
> + * halt_channel:	stop DMA channel operation (atomic)
> + * channel_busy:	return true, if the channel is busy (atomic)
> + * slave_addr:		return slave DMA address
> + * desc_setup:		set up the hardware specific descriptor portion (atomic)
> + * set_slave:		bind channel to a slave
> + * setup_xfer:		configure channel hardware for operation (atomic)
> + * start_xfer:		start the DMA transfer (atomic)
> + * embedded_desc:	return Nth struct dma_simple_desc pointer from the
> + *			descriptor array
> + * chan_irq:		process channel IRQ, return true if a transfer has
> + *			completed (atomic)
> + */
> +struct dma_simple_ops {
> +	bool (*desc_completed)(struct dma_simple_chan *, struct dma_simple_desc *);
> +	void (*halt_channel)(struct dma_simple_chan *);
> +	bool (*channel_busy)(struct dma_simple_chan *);
> +	dma_addr_t (*slave_addr)(struct dma_simple_chan *);
> +	int (*desc_setup)(struct dma_simple_chan *, struct dma_simple_desc *,
> +			  dma_addr_t, dma_addr_t, size_t *);
> +	int (*set_slave)(struct dma_simple_chan *, struct dma_simple_slave *);
> +	void (*setup_xfer)(struct dma_simple_chan *, struct dma_simple_slave *);
> +	void (*start_xfer)(struct dma_simple_chan *, struct dma_simple_desc *);
> +	struct dma_simple_desc *(*embedded_desc)(void *, int);
> +	bool (*chan_irq)(struct dma_simple_chan *, int);
> +};
again so many callbacks... are they really required!!
> +
> +struct dma_simple_dev {
> +	struct dma_device dma_dev;
> +	struct dma_simple_chan **schan;
> +	const struct dma_simple_ops *ops;
> +	size_t desc_size;
> +};
> +
> +#define dma_simple_for_each_chan(c, d, i) for (i = 0, c = (d)->schan[0]; \
> +				i < (d)->dma_dev.chancnt; c = (d)->schan[++i])
> +
> +int dma_simple_request_irq(struct dma_simple_chan *, int,
> +			   unsigned long, const char *);
> +void dma_simple_free_irq(struct dma_simple_chan *);
> +bool dma_simple_reset(struct dma_simple_dev *sdev);
> +void dma_simple_chan_probe(struct dma_simple_dev *sdev,
> +			   struct dma_simple_chan *schan, int id);
> +void dma_simple_chan_remove(struct dma_simple_chan *schan);
> +int dma_simple_init(struct device *dev, struct dma_simple_dev *sdev,
> +		    int chan_num);
> +void dma_simple_cleanup(struct dma_simple_dev *sdev);
> +
> +#endif

Now I am confused on the intent of this library. It was proposed for
helping dmacs like sh-mobile to support sg transfers in software which
are not supported by hardware, but it seems this library is doing _much_
more.
IMHO, it should get inserted between dmaengine APIs and client driver,
_only_ for sg transfers. The channel allocation etc belong to dmac.
Further, the library should get notified by dmac based on the callbacks
set for descriptor and then should call native driver while submitting
the next one in queue...
Rest of the stuff (if required) would not be generic and probably should
be in arch specific directory.


-- 
~Vinod


^ permalink raw reply	[flat|nested] 33+ messages in thread

* Re: [PATCH/RFC] usb: fix renesas_usbhs to not schedule in atomic context
  2012-02-03 15:43             ` [PATCH/RFC] usb: fix renesas_usbhs to not schedule in atomic context Guennadi Liakhovetski
  2012-02-05 14:54               ` Felipe Balbi
@ 2012-02-06  8:52               ` Shimoda, Yoshihiro
  1 sibling, 0 replies; 33+ messages in thread
From: Shimoda, Yoshihiro @ 2012-02-06  8:52 UTC (permalink / raw)
  To: Guennadi Liakhovetski
  Cc: linux-kernel, linux-sh, Vinod Koul, Magnus Damm, linux-mmc,
	alsa-devel, linux-serial, Paul Mundt, linux-usb

Hi Guennadi-san,

2012/02/04 0:43, Guennadi Liakhovetski wrote:
> The current renesas_usbhs driver triggers
> 
> BUG: scheduling while atomic: ksoftirqd/0/3/0x00000102
> 
> with enabled CONFIG_DEBUG_ATOMIC_SLEEP, by submitting DMA transfers from 
> an atomic (tasklet) context, which is not supported by the shdma dmaengine 
> driver. Fix it by switching to a work. Also simplify some list 
> manipulations.
> 
> Signed-off-by: Guennadi Liakhovetski <g.liakhovetski@gmx.de>
> ---
> 
> Shimoda-san, this is the problem, that you were observing. However, it 
> exists with the present version of shdma just as well as with the new one 
> - on top of the simple DMA library. I marked it an RFC because (1) I only 
> lightly tested it with the gadget device on mackerel with the mass storage 
> gadget, and (2) I am somewhat concerned about races. Currently the work 
> function runs with no locking and there are no usual cancel_work_sync() 
> points in the patch. However, it has also been like this before with the 
> tasklet implementation, which is not much better, and it looks like there 
> are no asynchronous operations on the same packets like timeouts. Only 
> asynchronous events, that I can think about are things like unloading the 
> driver or unplugging the cable, but these have been there before too. It 
> would become worse on SMP, I think. Comments welcome.

Thank you for the patch and comments.
I also tested this patch using the g_mass_storage and g_ether.
About your points, I think so, too.
So, I will write such a code.

Best regards,
Yoshihiro Shimoda

^ permalink raw reply	[flat|nested] 33+ messages in thread

* Re: [PATCH 6/7 v2] ASoC: SIU: prepare for conversion to simple DMA
  2012-02-06  2:33   ` Vinod Koul
@ 2012-02-06  9:11     ` Guennadi Liakhovetski
  0 siblings, 0 replies; 33+ messages in thread
From: Guennadi Liakhovetski @ 2012-02-06  9:11 UTC (permalink / raw)
  To: Vinod Koul
  Cc: linux-kernel, linux-sh, Magnus Damm, Yoshihiro Shimoda,
	linux-mmc, alsa-devel, linux-serial, Paul Mundt

Hi Vinod

On Mon, 6 Feb 2012, Vinod Koul wrote:

> On Thu, 2012-01-26 at 15:56 +0100, Guennadi Liakhovetski wrote:
> > Signed-off-by: Guennadi Liakhovetski <g.liakhovetski@gmx.de>
> > ---
> >  sound/soc/sh/siu_pcm.c |    8 ++++----
> >  1 files changed, 4 insertions(+), 4 deletions(-)
> > 
> > diff --git a/sound/soc/sh/siu_pcm.c b/sound/soc/sh/siu_pcm.c
> > index 0193e59..414bb35 100644
> > --- a/sound/soc/sh/siu_pcm.c
> > +++ b/sound/soc/sh/siu_pcm.c
> > @@ -330,12 +330,12 @@ static bool filter(struct dma_chan *chan, void *slave)
> >  {
> >  	struct sh_dmae_slave *param = slave;
> >  
> > -	pr_debug("%s: slave ID %d\n", __func__, param->slave_id);
> > +	pr_debug("%s: slave ID %d\n", __func__, param->simple_slave.slave_id);
> >  
> >  	if (unlikely(param->dma_dev != chan->device->dev))
> >  		return false;
> >  
> > -	chan->private = param;
> > +	chan->private = &param->simple_slave;
> While at it, Can you fix the sh-mobile driver to NOT use chan->private?

This is an independent change and for it I need my recent RFC PATCH 
"[PATCH/RFC] dmaengine: add a slave parameter to __dma_request_channel()"

http://marc.info/?l=linux-kernel&m=132811019930675&w=2

Thanks
Guennadi

> >  	return true;
> >  }
> >  
> > @@ -360,12 +360,12 @@ static int siu_pcm_open(struct snd_pcm_substream *ss)
> >  	if (ss->stream == SNDRV_PCM_STREAM_PLAYBACK) {
> >  		siu_stream = &port_info->playback;
> >  		param = &siu_stream->param;
> > -		param->slave_id = port ? pdata->dma_slave_tx_b :
> > +		param->simple_slave.slave_id = port ? pdata->dma_slave_tx_b :
> >  			pdata->dma_slave_tx_a;
> >  	} else {
> >  		siu_stream = &port_info->capture;
> >  		param = &siu_stream->param;
> > -		param->slave_id = port ? pdata->dma_slave_rx_b :
> > +		param->simple_slave.slave_id = port ? pdata->dma_slave_rx_b :
> >  			pdata->dma_slave_rx_a;
> >  	}
> >  
> 
> 
> -- 
> ~Vinod
> 

---
Guennadi Liakhovetski, Ph.D.
Freelance Open-Source Software Developer
http://www.open-technology.de/

^ permalink raw reply	[flat|nested] 33+ messages in thread

* Re: [PATCH 1/7 v2] dmaengine: add a simple dma library
  2012-02-06  2:54   ` Vinod Koul
@ 2012-02-06  9:53     ` Guennadi Liakhovetski
  0 siblings, 0 replies; 33+ messages in thread
From: Guennadi Liakhovetski @ 2012-02-06  9:53 UTC (permalink / raw)
  To: Vinod Koul
  Cc: linux-kernel, linux-sh, Magnus Damm, Yoshihiro Shimoda,
	linux-mmc, alsa-devel, linux-serial, Paul Mundt

On Mon, 6 Feb 2012, Vinod Koul wrote:

> On Thu, 2012-01-26 at 15:56 +0100, Guennadi Liakhovetski wrote:
> > This patch adds a library of functions, helping to implement dmaengine
> > drivers for hardware, unable to handle scatter-gather lists natively.
> > The first version of this driver only supports memcpy and slave DMA
> > operation.
> > 
> > Signed-off-by: Guennadi Liakhovetski <g.liakhovetski@gmx.de>
> > ---
> > 
> > v2:
> > 
> > 1. switch from using a tasklet to threaded IRQ, which allowed to
> ...?

Sorry, what exactly is your question here? The unfinished sentence? It is 
finished below in item 2, so, it should read like "...allowed to remove 
lock..."

> > 2. remove lock / unlock inline functions
> > 3. remove __devinit, __devexit annotations
> > 
> >  drivers/dma/Kconfig        |    3 +
> >  drivers/dma/Makefile       |    1 +
> >  drivers/dma/dma-simple.c   |  873 ++++++++++++++++++++++++++++++++++++++++++++
> >  include/linux/dma-simple.h |  124 +++++++
> >  4 files changed, 1001 insertions(+), 0 deletions(-)
> >  create mode 100644 drivers/dma/dma-simple.c
> >  create mode 100644 include/linux/dma-simple.h
> > 
> > diff --git a/drivers/dma/Kconfig b/drivers/dma/Kconfig
> > index f1a2749..f7c583e 100644
> > --- a/drivers/dma/Kconfig
> > +++ b/drivers/dma/Kconfig
> > @@ -149,6 +149,9 @@ config TXX9_DMAC
> >  	  Support the TXx9 SoC internal DMA controller.  This can be
> >  	  integrated in chips such as the Toshiba TX4927/38/39.
> >  
> > +config DMA_SIMPLE
> > +	tristate
> > +
> >  config SH_DMAE
> >  	tristate "Renesas SuperH DMAC support"
> >  	depends on (SUPERH && SH_DMA) || (ARM && ARCH_SHMOBILE)
> > diff --git a/drivers/dma/Makefile b/drivers/dma/Makefile
> > index 009a222..d63f773 100644
> > --- a/drivers/dma/Makefile
> > +++ b/drivers/dma/Makefile
> > @@ -2,6 +2,7 @@ ccflags-$(CONFIG_DMADEVICES_DEBUG)  := -DDEBUG
> >  ccflags-$(CONFIG_DMADEVICES_VDEBUG) += -DVERBOSE_DEBUG
> >  
> >  obj-$(CONFIG_DMA_ENGINE) += dmaengine.o
> > +obj-$(CONFIG_DMA_SIMPLE) += dma-simple.o
> >  obj-$(CONFIG_NET_DMA) += iovlock.o
> >  obj-$(CONFIG_INTEL_MID_DMAC) += intel_mid_dma.o
> >  obj-$(CONFIG_DMATEST) += dmatest.o
> > diff --git a/drivers/dma/dma-simple.c b/drivers/dma/dma-simple.c
> > new file mode 100644
> > index 0000000..49d8f7d
> > --- /dev/null
> > +++ b/drivers/dma/dma-simple.c
> > @@ -0,0 +1,873 @@
> > +/*
> > + * Simple dmaengine driver library
> > + *
> > + * extracted from shdma.c
> > + *
> > + * Copyright (C) 2011-2012 Guennadi Liakhovetski <g.liakhovetski@gmx.de>
> > + * Copyright (C) 2009 Nobuhiro Iwamatsu <iwamatsu.nobuhiro@renesas.com>
> > + * Copyright (C) 2009 Renesas Solutions, Inc. All rights reserved.
> > + * Copyright (C) 2007 Freescale Semiconductor, Inc. All rights reserved.
> > + *
> > + * This is free software; you can redistribute it and/or modify
> > + * it under the terms of version 2 of the GNU General Public License as
> > + * published by the Free Software Foundation.
> > + */
> > +
> > +#include <linux/delay.h>
> > +#include <linux/dma-simple.h>
> > +#include <linux/dmaengine.h>
> > +#include <linux/init.h>
> > +#include <linux/interrupt.h>
> > +#include <linux/module.h>
> > +#include <linux/pm_runtime.h>
> > +#include <linux/slab.h>
> > +#include <linux/spinlock.h>
> > +
> > +/* DMA descriptor control */
> > +enum simple_desc_status {
> > +	DESC_IDLE,
> > +	DESC_PREPARED,
> > +	DESC_SUBMITTED,
> > +	DESC_COMPLETED,	/* completed, have to call callback */
> > +	DESC_WAITING,	/* callback called, waiting for ack / re-submit */
> > +};
> why do you need to keep track of descriptor status?

Because descriptors in different states can be present on the same queue 
and you have to differentiate between them, when traversing the list.

> > +
> > +#define NR_DESCS_PER_CHANNEL 32
> > +
> > +#define to_simple_chan(c) container_of(c, struct dma_simple_chan, dma_chan)
> > +#define to_simple_dev(d) container_of(d, struct dma_simple_dev, dma_dev)
> > +
> > +/*
> > + * For slave DMA we assume, that there is a finite number of DMA slaves in the
> > + * system, and that each such slave can only use a finite number of channels.
> > + * We use slave channel IDs to make sure, that no such slave channel ID is
> > + * allocated more than once.
> > + */
> > +static unsigned int slave_num = 256;
> > +module_param(slave_num, uint, 0444);
> > +
> > +/* A bitmask with slave_num bits */
> > +static unsigned long *simple_slave_used;
> > +
> > +/* Called under spin_lock_irq(&schan->chan_lock") */
> > +static void simple_chan_xfer_ld_queue(struct dma_simple_chan *schan)
> > +{
> > +	struct dma_simple_dev *sdev = to_simple_dev(schan->dma_chan.device);
> > +	const struct dma_simple_ops *ops = sdev->ops;
> > +	struct dma_simple_desc *sdesc;
> > +
> > +	/* DMA work check */
> > +	if (ops->channel_busy(schan))
> > +		return;
> > +
> > +	/* Find the first not transferred descriptor */
> > +	list_for_each_entry(sdesc, &schan->ld_queue, node)
> > +		if (sdesc->mark == DESC_SUBMITTED) {
> > +			ops->start_xfer(schan, sdesc);
> > +			break;
> > +		}
> > +}
> > +
> > +static dma_cookie_t simple_tx_submit(struct dma_async_tx_descriptor *tx)
> > +{
> > +	struct dma_simple_desc *chunk, *c, *desc =
> > +		container_of(tx, struct dma_simple_desc, async_tx),
> > +		*last = desc;
> > +	struct dma_simple_chan *schan = to_simple_chan(tx->chan);
> > +	struct dma_simple_slave *slave = tx->chan->private;
> prive is masked depricated, os this needs to be removed.

Right, it would be best to first merge patch "[PATCH/RFC] dmaengine: add a 
slave parameter to __dma_request_channel()" and then port this library on 
top of it, then private will not be used any more.

> Any slave config should be extracted from dma_lsave_config only... Do
> you anything more than which is provided there??

I don't think the dmaengine_slave_config() API is very well suitable for 
our situation. The problem is, that on sh-mobile not all DMA controllers 
support all functions. E.g., sh7372 has two dedicated USB DMA controllers, 
that otherwise are fully compatible with other DMA controllers on that 
platforms. If a client requests a channel for a USB slave and gets back a 
channel on one of other DMAC instances, issuing dmaengine_slave_config() 
with a USB configuration, obviously, will not work. Similarly, if a client 
would try to allocate a non-USB channel on a USB controller. So, it is 
best to be able to decide at dma_request_channel() time, whether and from 
which controller this slave channel request can be satisfied.

> > +	dma_async_tx_callback callback = tx->callback;
> > +	dma_cookie_t cookie;
> > +	bool power_up;
> > +
> > +	spin_lock_irq(&schan->chan_lock);
> > +
> > +	power_up = list_empty(&schan->ld_queue);
> > +
> > +	cookie = schan->dma_chan.cookie + 1;
> > +	if (cookie < 0)
> > +		cookie = 1;
> > +
> > +	schan->dma_chan.cookie = cookie;
> > +	tx->cookie = cookie;
> > +
> > +	/* Mark all chunks of this descriptor as submitted, move to the queue */
> > +	list_for_each_entry_safe(chunk, c, desc->node.prev, node) {
> > +		/*
> > +		 * All chunks are on the global ld_free, so, we have to find
> > +		 * the end of the chain ourselves
> > +		 */
> > +		if (chunk != desc && (chunk->mark == DESC_IDLE ||
> > +				      chunk->async_tx.cookie > 0 ||
> > +				      chunk->async_tx.cookie == -EBUSY ||
> > +				      &chunk->node == &schan->ld_free))
> > +			break;
> > +		chunk->mark = DESC_SUBMITTED;
> > +		/* Callback goes to the last chunk */
> > +		chunk->async_tx.callback = NULL;
> > +		chunk->cookie = cookie;
> > +		list_move_tail(&chunk->node, &schan->ld_queue);
> > +		last = chunk;
> > +
> > +		dev_dbg(schan->dev, "submit #%d@%p on %d\n",
> > +			tx->cookie, &last->async_tx, schan->id);
> > +	}
> > +
> > +	last->async_tx.callback = callback;
> > +	last->async_tx.callback_param = tx->callback_param;
> > +
> > +	if (power_up) {
> > +		int ret;
> > +		schan->pm_state = DMA_SIMPLE_PM_BUSY;
> > +
> > +		ret = pm_runtime_get(schan->dev);
> > +
> > +		spin_unlock_irq(&schan->chan_lock);
> > +		if (ret < 0)
> > +			dev_err(schan->dev, "%s(): GET = %d\n", __func__, ret);
> > +
> > +		pm_runtime_barrier(schan->dev);
> > +
> > +		spin_lock_irq(&schan->chan_lock);
> > +
> > +		/* Have we been reset, while waiting? */
> > +		if (schan->pm_state != DMA_SIMPLE_PM_ESTABLISHED) {
> > +			struct dma_simple_dev *sdev =
> > +				to_simple_dev(schan->dma_chan.device);
> > +			const struct dma_simple_ops *ops = sdev->ops;
> > +			dev_dbg(schan->dev, "Bring up channel %d\n",
> > +				schan->id);
> > +			/*
> > +			 * TODO: .xfer_setup() might fail on some platforms.
> > +			 * Make it int then, on error remove chunks from the
> > +			 * queue again
> > +			 */
> > +			ops->setup_xfer(schan, slave);
> > +
> > +			if (schan->pm_state == DMA_SIMPLE_PM_PENDING)
> > +				simple_chan_xfer_ld_queue(schan);
> > +			schan->pm_state = DMA_SIMPLE_PM_ESTABLISHED;
> > +		}
> > +	} else {
> > +		/*
> > +		 * Tell .device_issue_pending() not to run the queue, interrupts
> > +		 * will do it anyway
> > +		 */
> > +		schan->pm_state = DMA_SIMPLE_PM_PENDING;
> > +	}
> > +
> > +	spin_unlock_irq(&schan->chan_lock);
> > +
> > +	return cookie;
> > +}
> > +
> > +/* Called with desc_lock held */
> > +static struct dma_simple_desc *simple_get_desc(struct dma_simple_chan *schan)
> > +{
> > +	struct dma_simple_desc *sdesc;
> > +
> > +	list_for_each_entry(sdesc, &schan->ld_free, node)
> > +		if (sdesc->mark != DESC_PREPARED) {
> > +			BUG_ON(sdesc->mark != DESC_IDLE);
> > +			list_del(&sdesc->node);
> > +			return sdesc;
> > +		}
> > +
> > +	return NULL;
> > +}
> > +
> > +static int simple_alloc_chan_resources(struct dma_chan *chan)
> > +{
> > +	struct dma_simple_chan *schan = to_simple_chan(chan);
> > +	struct dma_simple_dev *sdev = to_simple_dev(schan->dma_chan.device);
> > +	const struct dma_simple_ops *ops = sdev->ops;
> > +	struct dma_simple_desc *desc;
> > +	struct dma_simple_slave *slave = chan->private;
> > +	int ret, i;
> > +
> > +	/*
> > +	 * This relies on the guarantee from dmaengine that alloc_chan_resources
> > +	 * never runs concurrently with itself or free_chan_resources.
> > +	 */
> > +	if (slave) {
> > +		if (test_and_set_bit(slave->slave_id, simple_slave_used)) {
> > +			ret = -EBUSY;
> > +			goto etestused;
> > +		}
> > +
> > +		ret = ops->set_slave(schan, slave);
> > +		if (ret < 0)
> > +			goto esetslave;
> > +	}
> > +
> > +	schan->desc = kcalloc(NR_DESCS_PER_CHANNEL,
> > +			      sdev->desc_size, GFP_KERNEL);
> > +	if (!schan->desc) {
> > +		ret = -ENOMEM;
> > +		goto edescalloc;
> > +	}
> > +	schan->desc_num = NR_DESCS_PER_CHANNEL;
> > +
> > +	for (i = 0; i < NR_DESCS_PER_CHANNEL; i++) {
> > +		desc = ops->embedded_desc(schan->desc, i);
> > +		dma_async_tx_descriptor_init(&desc->async_tx,
> > +					     &schan->dma_chan);
> > +		desc->async_tx.tx_submit = simple_tx_submit;
> > +		desc->mark = DESC_IDLE;
> > +
> > +		list_add(&desc->node, &schan->ld_free);
> > +	}
> > +
> > +	return NR_DESCS_PER_CHANNEL;
> > +
> > +edescalloc:
> > +	if (slave)
> > +esetslave:
> > +		clear_bit(slave->slave_id, simple_slave_used);
> > +etestused:
> > +	chan->private = NULL;
> > +	return ret;
> > +}
> Typically chan allocation involves some kind of hand shaking between the
> client and dmac which typically is arch specfic. If we want to make this
> a truly independent library, then I think we should move allocation to
> driver and let them allocate the required channel. The library by
> definition is to _help_ for sg transfers so it should just create a
> library of APIs to call which manage the sg transfer when not supported
> by dmac.

Again, that's also something, that should be handled by the proposed 
patch. With it any additional information, required to configure the 
controller and / or the channel for the slave operation is passed already 
to the allocation routine. Then, hopefully, no additional handshaking 
would be needed.

> > +
> > +static dma_async_tx_callback __ld_cleanup(struct dma_simple_chan *schan, bool all)
> > +{
> > +	struct dma_simple_desc *desc, *_desc;
> > +	/* Is the "exposed" head of a chain acked? */
> > +	bool head_acked = false;
> > +	dma_cookie_t cookie = 0;
> > +	dma_async_tx_callback callback = NULL;
> > +	void *param = NULL;
> > +	unsigned long flags;
> > +
> > +	spin_lock_irqsave(&schan->chan_lock, flags);
> > +	list_for_each_entry_safe(desc, _desc, &schan->ld_queue, node) {
> > +		struct dma_async_tx_descriptor *tx = &desc->async_tx;
> > +
> > +		BUG_ON(tx->cookie > 0 && tx->cookie != desc->cookie);
> > +		BUG_ON(desc->mark != DESC_SUBMITTED &&
> > +		       desc->mark != DESC_COMPLETED &&
> > +		       desc->mark != DESC_WAITING);
> > +
> > +		/*
> > +		 * queue is ordered, and we use this loop to (1) clean up all
> > +		 * completed descriptors, and to (2) update descriptor flags of
> > +		 * any chunks in a (partially) completed chain
> > +		 */
> > +		if (!all && desc->mark == DESC_SUBMITTED &&
> > +		    desc->cookie != cookie)
> > +			break;
> > +
> > +		if (tx->cookie > 0)
> > +			cookie = tx->cookie;
> > +
> > +		if (desc->mark == DESC_COMPLETED && desc->chunks == 1) {
> > +			if (schan->completed_cookie != desc->cookie - 1)
> > +				dev_dbg(schan->dev,
> > +					"Completing cookie %d, expected %d\n",
> > +					desc->cookie,
> > +					schan->completed_cookie + 1);
> > +			schan->completed_cookie = desc->cookie;
> > +		}
> > +
> > +		/* Call callback on the last chunk */
> > +		if (desc->mark == DESC_COMPLETED && tx->callback) {
> > +			desc->mark = DESC_WAITING;
> > +			callback = tx->callback;
> > +			param = tx->callback_param;
> > +			dev_dbg(schan->dev, "descriptor #%d@%p on %d callback\n",
> > +				tx->cookie, tx, schan->id);
> > +			BUG_ON(desc->chunks != 1);
> > +			break;
> > +		}
> > +
> > +		if (tx->cookie > 0 || tx->cookie == -EBUSY) {
> > +			if (desc->mark == DESC_COMPLETED) {
> > +				BUG_ON(tx->cookie < 0);
> > +				desc->mark = DESC_WAITING;
> > +			}
> > +			head_acked = async_tx_test_ack(tx);
> > +		} else {
> > +			switch (desc->mark) {
> > +			case DESC_COMPLETED:
> > +				desc->mark = DESC_WAITING;
> > +				/* Fall through */
> > +			case DESC_WAITING:
> > +				if (head_acked)
> > +					async_tx_ack(&desc->async_tx);
> > +			}
> > +		}
> > +
> > +		dev_dbg(schan->dev, "descriptor %p #%d completed.\n",
> > +			tx, tx->cookie);
> > +
> > +		if (((desc->mark == DESC_COMPLETED ||
> > +		      desc->mark == DESC_WAITING) &&
> > +		     async_tx_test_ack(&desc->async_tx)) || all) {
> > +			/* Remove from ld_queue list */
> > +			desc->mark = DESC_IDLE;
> > +
> > +			list_move(&desc->node, &schan->ld_free);
> > +
> > +			if (list_empty(&schan->ld_queue)) {
> > +				dev_dbg(schan->dev, "Bring down channel %d\n", schan->id);
> > +				pm_runtime_put(schan->dev);
> > +				schan->pm_state = DMA_SIMPLE_PM_ESTABLISHED;
> > +			}
> > +		}
> > +	}
> > +
> > +	if (all && !callback)
> > +		/*
> > +		 * Terminating and the loop completed normally: forgive
> > +		 * uncompleted cookies
> > +		 */
> > +		schan->completed_cookie = schan->dma_chan.cookie;
> > +
> > +	spin_unlock_irqrestore(&schan->chan_lock, flags);
> > +
> > +	if (callback)
> > +		callback(param);
> > +
> > +	return callback;
> > +}
> > +
> > +/*
> > + * simple_chan_ld_cleanup - Clean up link descriptors
> > + *
> > + * Clean up the ld_queue of DMA channel.
> > + */
> > +static void simple_chan_ld_cleanup(struct dma_simple_chan *schan, bool all)
> > +{
> > +	while (__ld_cleanup(schan, all))
> > +		;
> > +}
> > +
> > +/*
> > + * simple_free_chan_resources - Free all resources of the channel.
> > + */
> > +static void simple_free_chan_resources(struct dma_chan *chan)
> > +{
> > +	struct dma_simple_chan *schan = to_simple_chan(chan);
> > +	struct dma_simple_dev *sdev = to_simple_dev(chan->device);
> > +	const struct dma_simple_ops *ops = sdev->ops;
> > +	LIST_HEAD(list);
> > +
> > +	/* Protect against ISR */
> > +	spin_lock_irq(&schan->chan_lock);
> > +	ops->halt_channel(schan);
> > +	spin_unlock_irq(&schan->chan_lock);
> > +
> > +	/* Now no new interrupts will occur */
> > +
> > +	/* Prepared and not submitted descriptors can still be on the queue */
> > +	if (!list_empty(&schan->ld_queue))
> > +		simple_chan_ld_cleanup(schan, true);
> > +
> > +	if (chan->private) {
> > +		/* The caller is holding dma_list_mutex */
> > +		struct dma_simple_slave *slave = chan->private;
> > +		clear_bit(slave->slave_id, simple_slave_used);
> > +		chan->private = NULL;
> > +	}
> > +
> > +	spin_lock_irq(&schan->chan_lock);
> > +
> > +	list_splice_init(&schan->ld_free, &list);
> > +	schan->desc_num = 0;
> > +
> > +	spin_unlock_irq(&schan->chan_lock);
> > +
> > +	kfree(schan->desc);
> > +}
> > +
> > +/**
> > + * simple_add_desc - get, set up and return one transfer descriptor
> > + * @schan:	DMA channel
> > + * @flags:	DMA transfer flags
> > + * @dst:	destination DMA address, incremented when direction equals
> > + *		DMA_DEV_TO_MEM or DMA_MEM_TO_MEM
> > + * @src:	source DMA address, incremented when direction equals
> > + *		DMA_MEM_TO_DEV or DMA_MEM_TO_MEM
> > + * @len:	DMA transfer length
> > + * @first:	if NULL, set to the current descriptor and cookie set to -EBUSY
> > + * @direction:	needed for slave DMA to decide which address to keep constant,
> > + *		equals DMA_MEM_TO_MEM for MEMCPY
> > + * Returns 0 or an error
> > + * Locks: called with desc_lock held
> > + */
> > +static struct dma_simple_desc *simple_add_desc(struct dma_simple_chan *schan,
> > +	unsigned long flags, dma_addr_t *dst, dma_addr_t *src, size_t *len,
> > +	struct dma_simple_desc **first, enum dma_transfer_direction direction)
> > +{
> > +	struct dma_simple_dev *sdev = to_simple_dev(schan->dma_chan.device);
> > +	const struct dma_simple_ops *ops = sdev->ops;
> > +	struct dma_simple_desc *new;
> > +	size_t copy_size = *len;
> > +
> > +	if (!copy_size)
> > +		return NULL;
> > +
> > +	/* Allocate the link descriptor from the free list */
> > +	new = simple_get_desc(schan);
> > +	if (!new) {
> > +		dev_err(schan->dev, "No free link descriptor available\n");
> > +		return NULL;
> > +	}
> > +
> > +	ops->desc_setup(schan, new, *src, *dst, &copy_size);
> > +
> > +	if (!*first) {
> > +		/* First desc */
> > +		new->async_tx.cookie = -EBUSY;
> > +		*first = new;
> > +	} else {
> > +		/* Other desc - invisible to the user */
> > +		new->async_tx.cookie = -EINVAL;
> > +	}
> > +
> > +	dev_dbg(schan->dev,
> > +		"chaining (%u/%u)@%x -> %x with %p, cookie %d\n",
> > +		copy_size, *len, *src, *dst, &new->async_tx,
> > +		new->async_tx.cookie);
> > +
> > +	new->mark = DESC_PREPARED;
> > +	new->async_tx.flags = flags;
> > +	new->direction = direction;
> > +
> > +	*len -= copy_size;
> > +	if (direction == DMA_MEM_TO_MEM || direction == DMA_MEM_TO_DEV)
> > +		*src += copy_size;
> > +	if (direction == DMA_MEM_TO_MEM || direction == DMA_DEV_TO_MEM)
> > +		*dst += copy_size;
> > +
> > +	return new;
> > +}
> > +
> > +/*
> > + * simple_prep_sg - prepare transfer descriptors from an SG list
> > + *
> > + * Common routine for public (MEMCPY) and slave DMA. The MEMCPY case is also
> > + * converted to scatter-gather to guarantee consistent locking and a correct
> > + * list manipulation. For slave DMA direction carries the usual meaning, and,
> > + * logically, the SG list is RAM and the addr variable contains slave address,
> > + * e.g., the FIFO I/O register. For MEMCPY direction equals DMA_MEM_TO_MEM
> > + * and the SG list contains only one element and points at the source buffer.
> > + */
> > +static struct dma_async_tx_descriptor *simple_prep_sg(struct dma_simple_chan *schan,
> > +	struct scatterlist *sgl, unsigned int sg_len, dma_addr_t *addr,
> > +	enum dma_transfer_direction direction, unsigned long flags)
> > +{
> > +	struct scatterlist *sg;
> > +	struct dma_simple_desc *first = NULL, *new = NULL /* compiler... */;
> > +	LIST_HEAD(tx_list);
> > +	int chunks = 0;
> > +	unsigned long irq_flags;
> > +	int i;
> > +
> > +	for_each_sg(sgl, sg, sg_len, i)
> > +		chunks += DIV_ROUND_UP(sg_dma_len(sg), schan->max_xfer_len);
> > +
> > +	/* Have to lock the whole loop to protect against concurrent release */
> > +	spin_lock_irqsave(&schan->chan_lock, irq_flags);
> > +
> > +	/*
> > +	 * Chaining:
> > +	 * first descriptor is what user is dealing with in all API calls, its
> > +	 *	cookie is at first set to -EBUSY, at tx-submit to a positive
> > +	 *	number
> > +	 * if more than one chunk is needed further chunks have cookie = -EINVAL
> > +	 * the last chunk, if not equal to the first, has cookie = -ENOSPC
> > +	 * all chunks are linked onto the tx_list head with their .node heads
> > +	 *	only during this function, then they are immediately spliced
> > +	 *	back onto the free list in form of a chain
> > +	 */
> > +	for_each_sg(sgl, sg, sg_len, i) {
> > +		dma_addr_t sg_addr = sg_dma_address(sg);
> > +		size_t len = sg_dma_len(sg);
> > +
> > +		if (!len)
> > +			goto err_get_desc;
> > +
> > +		do {
> > +			dev_dbg(schan->dev, "Add SG #%d@%p[%d], dma %llx\n",
> > +				i, sg, len, (unsigned long long)sg_addr);
> > +
> > +			if (direction == DMA_DEV_TO_MEM)
> > +				new = simple_add_desc(schan, flags,
> > +						&sg_addr, addr, &len, &first,
> > +						direction);
> > +			else
> > +				new = simple_add_desc(schan, flags,
> > +						addr, &sg_addr, &len, &first,
> > +						direction);
> > +			if (!new)
> > +				goto err_get_desc;
> > +
> > +			new->chunks = chunks--;
> > +			list_add_tail(&new->node, &tx_list);
> > +		} while (len);
> > +	}
> > +
> > +	if (new != first)
> > +		new->async_tx.cookie = -ENOSPC;
> > +
> > +	/* Put them back on the free list, so, they don't get lost */
> > +	list_splice_tail(&tx_list, &schan->ld_free);
> > +
> > +	spin_unlock_irqrestore(&schan->chan_lock, irq_flags);
> > +
> > +	return &first->async_tx;
> > +
> > +err_get_desc:
> > +	list_for_each_entry(new, &tx_list, node)
> > +		new->mark = DESC_IDLE;
> > +	list_splice(&tx_list, &schan->ld_free);
> > +
> > +	spin_unlock_irqrestore(&schan->chan_lock, irq_flags);
> > +
> > +	return NULL;
> > +}
> > +
> > +static struct dma_async_tx_descriptor *simple_prep_memcpy(
> > +	struct dma_chan *chan, dma_addr_t dma_dest, dma_addr_t dma_src,
> > +	size_t len, unsigned long flags)
> > +{
> > +	struct dma_simple_chan *schan = to_simple_chan(chan);
> > +	struct scatterlist sg;
> > +
> > +	if (!chan || !len)
> > +		return NULL;
> > +
> > +	BUG_ON(!schan->desc_num);
> > +
> > +	sg_init_table(&sg, 1);
> > +	sg_set_page(&sg, pfn_to_page(PFN_DOWN(dma_src)), len,
> > +		    offset_in_page(dma_src));
> > +	sg_dma_address(&sg) = dma_src;
> > +	sg_dma_len(&sg) = len;
> > +
> > +	return simple_prep_sg(schan, &sg, 1, &dma_dest, DMA_MEM_TO_MEM, flags);
> > +}
> mempcy is a single transfer why should this go thru library?
> got sg_memcpy yes, but otherwise NO

This allows to unify the transfer (descriptor) handling also for cases, 
when the user is requesting too large a transfer, that has to be split 
internally in the driver into several transfers.

> > +
> > +static struct dma_async_tx_descriptor *simple_prep_slave_sg(
> > +	struct dma_chan *chan, struct scatterlist *sgl, unsigned int sg_len,
> > +	enum dma_transfer_direction direction, unsigned long flags)
> > +{
> > +	struct dma_simple_chan *schan = to_simple_chan(chan);
> > +	struct dma_simple_dev *sdev = to_simple_dev(schan->dma_chan.device);
> > +	const struct dma_simple_ops *ops = sdev->ops;
> > +	struct dma_simple_slave *slave = chan->private;
> > +	dma_addr_t slave_addr;
> > +
> > +	if (!chan)
> > +		return NULL;
> > +
> > +	BUG_ON(!schan->desc_num);
> > +
> > +	/* Someone calling slave DMA on a generic channel? */
> > +	if (!slave || !sg_len) {
> > +		dev_warn(schan->dev, "%s: bad parameter: %p, %d, %d\n",
> > +			 __func__, slave, sg_len, slave ? slave->slave_id : -1);
> > +		return NULL;
> > +	}
> > +
> > +	slave_addr = ops->slave_addr(schan);
> > +
> > +	return simple_prep_sg(schan, sgl, sg_len, &slave_addr,
> > +			      direction, flags);
> > +}
> > +
> > +static int simple_control(struct dma_chan *chan, enum dma_ctrl_cmd cmd,
> > +			  unsigned long arg)
> > +{
> > +	struct dma_simple_chan *schan = to_simple_chan(chan);
> > +	struct dma_simple_dev *sdev = to_simple_dev(chan->device);
> > +	const struct dma_simple_ops *ops = sdev->ops;
> > +	unsigned long flags;
> > +
> > +	/* Only supports DMA_TERMINATE_ALL */
> > +	if (cmd != DMA_TERMINATE_ALL)
> > +		return -ENXIO;
> nope, you should check from respective dmac...

Well, right, some drivers might implement more. So, our choice is either 
to preemptively prepare code to handle those, or wait until such drivers 
surface and wish to use this library, then we can extend it to handle 
those too.

> > +
> > +	if (!chan)
> > +		return -EINVAL;
> > +
> > +	spin_lock_irqsave(&schan->chan_lock, flags);
> > +
> > +	ops->halt_channel(schan);
> > +
> > +	spin_unlock_irqrestore(&schan->chan_lock, flags);
> > +
> > +	simple_chan_ld_cleanup(schan, true);
> > +
> > +	return 0;
> > +}
> > +
> > +static void simple_issue_pending(struct dma_chan *chan)
> > +{
> > +	struct dma_simple_chan *schan = to_simple_chan(chan);
> > +
> > +	spin_lock_irq(&schan->chan_lock);
> > +	if (schan->pm_state == DMA_SIMPLE_PM_ESTABLISHED)
> > +		simple_chan_xfer_ld_queue(schan);
> > +	else
> > +		schan->pm_state = DMA_SIMPLE_PM_PENDING;
> > +	spin_unlock_irq(&schan->chan_lock);
> > +}
> > +
> > +static enum dma_status simple_tx_status(struct dma_chan *chan,
> > +					dma_cookie_t cookie,
> > +					struct dma_tx_state *txstate)
> > +{
> > +	struct dma_simple_chan *schan = to_simple_chan(chan);
> > +	dma_cookie_t last_used;
> > +	dma_cookie_t last_complete;
> > +	enum dma_status status;
> > +	unsigned long flags;
> > +
> > +	simple_chan_ld_cleanup(schan, false);
> > +
> > +	/* First read completed cookie to avoid a skew */
> > +	last_complete = schan->completed_cookie;
> > +	rmb();
> > +	last_used = chan->cookie;
> > +	BUG_ON(last_complete < 0);
> > +	dma_set_tx_state(txstate, last_complete, last_used, 0);
> > +
> > +	spin_lock_irqsave(&schan->chan_lock, flags);
> > +
> > +	status = dma_async_is_complete(cookie, last_complete, last_used);
> > +
> > +	/*
> > +	 * If we don't find cookie on the queue, it has been aborted and we have
> > +	 * to report error
> > +	 */
> > +	if (status != DMA_SUCCESS) {
> > +		struct dma_simple_desc *sdesc;
> > +		status = DMA_ERROR;
> > +		list_for_each_entry(sdesc, &schan->ld_queue, node)
> > +			if (sdesc->cookie == cookie) {
> > +				status = DMA_IN_PROGRESS;
> > +				break;
> > +			}
> > +	}
> > +
> > +	spin_unlock_irqrestore(&schan->chan_lock, flags);
> > +
> > +	return status;
> > +}
> > +
> > +/* Called from error IRQ or NMI */
> > +bool dma_simple_reset(struct dma_simple_dev *sdev)
> > +{
> > +	const struct dma_simple_ops *ops = sdev->ops;
> > +	struct dma_simple_chan *schan;
> > +	unsigned int handled = 0;
> > +	int i;
> > +
> > +	/* Reset all channels */
> > +	dma_simple_for_each_chan(schan, sdev, i) {
> > +		struct dma_simple_desc *sdesc;
> > +		LIST_HEAD(dl);
> > +
> > +		if (!schan)
> > +			continue;
> > +
> > +		spin_lock(&schan->chan_lock);
> > +
> > +		/* Stop the channel */
> > +		ops->halt_channel(schan);
> > +
> > +		list_splice_init(&schan->ld_queue, &dl);
> > +
> > +		if (!list_empty(&dl)) {
> > +			dev_dbg(schan->dev, "Bring down channel %d\n", schan->id);
> > +			pm_runtime_put(schan->dev);
> > +		}
> > +		schan->pm_state = DMA_SIMPLE_PM_ESTABLISHED;
> > +
> > +		spin_unlock(&schan->chan_lock);
> > +
> > +		/* Complete all  */
> > +		list_for_each_entry(sdesc, &dl, node) {
> > +			struct dma_async_tx_descriptor *tx = &sdesc->async_tx;
> > +			sdesc->mark = DESC_IDLE;
> > +			if (tx->callback)
> > +				tx->callback(tx->callback_param);
> > +		}
> > +
> > +		spin_lock(&schan->chan_lock);
> > +		list_splice(&dl, &schan->ld_free);
> > +		spin_unlock(&schan->chan_lock);
> > +
> > +		handled++;
> > +	}
> > +
> > +	return !!handled;
> > +}
> > +EXPORT_SYMBOL(dma_simple_reset);
> > +
> > +static irqreturn_t chan_irq(int irq, void *dev)
> > +{
> > +	struct dma_simple_chan *schan = dev;
> > +	const struct dma_simple_ops *ops =
> > +		to_simple_dev(schan->dma_chan.device)->ops;
> > +	irqreturn_t ret;
> > +
> > +	spin_lock(&schan->chan_lock);
> > +
> > +	ret = ops->chan_irq(schan, irq) ? IRQ_WAKE_THREAD : IRQ_NONE;
> > +
> > +	spin_unlock(&schan->chan_lock);
> > +
> > +	return ret;
> > +}
> > +
> > +static irqreturn_t chan_irqt(int irq, void *dev)
> > +{
> > +	struct dma_simple_chan *schan = dev;
> > +	const struct dma_simple_ops *ops =
> > +		to_simple_dev(schan->dma_chan.device)->ops;
> > +	struct dma_simple_desc *sdesc;
> > +
> > +	spin_lock_irq(&schan->chan_lock);
> > +	list_for_each_entry(sdesc, &schan->ld_queue, node) {
> > +		if (sdesc->mark == DESC_SUBMITTED &&
> > +		    ops->desc_completed(schan, sdesc)) {
> > +			dev_dbg(schan->dev, "done #%d@%p\n",
> > +				sdesc->async_tx.cookie, &sdesc->async_tx);
> > +			sdesc->mark = DESC_COMPLETED;
> > +			break;
> > +		}
> > +	}
> > +	/* Next desc */
> > +	simple_chan_xfer_ld_queue(schan);
> > +	spin_unlock_irq(&schan->chan_lock);
> > +
> > +	simple_chan_ld_cleanup(schan, false);
> > +
> > +	return IRQ_HANDLED;
> > +}
> > +
> > +int dma_simple_request_irq(struct dma_simple_chan *schan, int irq,
> > +			   unsigned long flags, const char *name)
> > +{
> > +	int ret = request_threaded_irq(irq, chan_irq, chan_irqt,
> > +				       flags, name, schan);
> > +
> > +	schan->irq = ret < 0 ? ret : irq;
> > +
> > +	return ret;
> > +}
> > +EXPORT_SYMBOL(dma_simple_request_irq);
> > +
> > +void dma_simple_free_irq(struct dma_simple_chan *schan)
> > +{
> > +	if (schan->irq >= 0)
> > +		free_irq(schan->irq, schan);
> > +}
> > +EXPORT_SYMBOL(dma_simple_free_irq);
> why would you use the irq here?? That should be handled by respective
> dmac. Also, the library should just setup callbacks for descriptor and
> use them to manage the descriptors for sg mode.

Please, see below.

> > +
> > +void dma_simple_chan_probe(struct dma_simple_dev *sdev,
> > +			   struct dma_simple_chan *schan, int id)
> > +{
> > +	schan->pm_state = DMA_SIMPLE_PM_ESTABLISHED;
> > +
> > +	/* reference struct dma_device */
> > +	schan->dma_chan.device = &sdev->dma_dev;
> > +
> > +	schan->dev = sdev->dma_dev.dev;
> > +	schan->id = id;
> > +
> > +	if (!schan->max_xfer_len)
> > +		schan->max_xfer_len = PAGE_SIZE;
> > +
> > +	spin_lock_init(&schan->chan_lock);
> > +
> > +	/* Init descripter manage list */
> > +	INIT_LIST_HEAD(&schan->ld_queue);
> > +	INIT_LIST_HEAD(&schan->ld_free);
> > +
> > +	/* Add the channel to DMA device channel list */
> > +	list_add_tail(&schan->dma_chan.device_node,
> > +			&sdev->dma_dev.channels);
> > +	sdev->schan[sdev->dma_dev.chancnt++] = schan;
> > +}
> > +EXPORT_SYMBOL(dma_simple_chan_probe);
> > +
> > +void dma_simple_chan_remove(struct dma_simple_chan *schan)
> > +{
> > +	list_del(&schan->dma_chan.device_node);
> > +}
> > +EXPORT_SYMBOL(dma_simple_chan_remove);
> > +
> > +int dma_simple_init(struct device *dev, struct dma_simple_dev *sdev,
> > +		    int chan_num)
> > +{
> > +	struct dma_device *dma_dev = &sdev->dma_dev;
> > +
> > +	/*
> > +	 * Require all call-backs for now, they can trivially be made optional
> > +	 * later as required
> > +	 */
> > +	if (!sdev->ops ||
> > +	    !sdev->desc_size ||
> > +	    !sdev->ops->embedded_desc ||
> > +	    !sdev->ops->start_xfer ||
> > +	    !sdev->ops->setup_xfer ||
> > +	    !sdev->ops->set_slave ||
> > +	    !sdev->ops->desc_setup ||
> > +	    !sdev->ops->slave_addr ||
> > +	    !sdev->ops->channel_busy ||
> > +	    !sdev->ops->halt_channel ||
> > +	    !sdev->ops->desc_completed)
> > +		return -EINVAL;
> > +
> > +	sdev->schan = kcalloc(chan_num, sizeof(*sdev->schan), GFP_KERNEL);
> > +	if (!sdev->schan)
> > +		return -ENOMEM;
> > +
> > +	INIT_LIST_HEAD(&dma_dev->channels);
> > +
> > +	/* Common and MEMCPY operations */
> > +	dma_dev->device_alloc_chan_resources
> > +		= simple_alloc_chan_resources;
> > +	dma_dev->device_free_chan_resources = simple_free_chan_resources;
> > +	dma_dev->device_prep_dma_memcpy = simple_prep_memcpy;
> > +	dma_dev->device_tx_status = simple_tx_status;
> > +	dma_dev->device_issue_pending = simple_issue_pending;
> > +
> > +	/* Compulsory for DMA_SLAVE fields */
> > +	dma_dev->device_prep_slave_sg = simple_prep_slave_sg;
> > +	dma_dev->device_control = simple_control;
> > +
> > +	dma_dev->dev = dev;
> > +
> > +	return 0;
> > +}
> > +EXPORT_SYMBOL(dma_simple_init);
> > +
> > +void dma_simple_cleanup(struct dma_simple_dev *sdev)
> > +{
> > +	kfree(sdev->schan);
> > +}
> > +EXPORT_SYMBOL(dma_simple_cleanup);
> > +
> > +static int __init dma_simple_enter(void)
> > +{
> > +	simple_slave_used = kzalloc(DIV_ROUND_UP(slave_num, BITS_PER_LONG) *
> > +				    sizeof(long), GFP_KERNEL);
> > +	if (!simple_slave_used)
> > +		return -ENOMEM;
> > +	return 0;
> > +}
> > +module_init(dma_simple_enter);
> > +
> > +static void __exit dma_simple_exit(void)
> > +{
> > +	kfree(simple_slave_used);
> > +}
> > +module_exit(dma_simple_exit);
> > +
> > +MODULE_LICENSE("GPL v2");
> > +MODULE_DESCRIPTION("Simple dmaengine driver library");
> > +MODULE_AUTHOR("Guennadi Liakhovetski <g.liakhovetski@gmx.de>");
> > diff --git a/include/linux/dma-simple.h b/include/linux/dma-simple.h
> > new file mode 100644
> > index 0000000..5336674
> > --- /dev/null
> > +++ b/include/linux/dma-simple.h
> > @@ -0,0 +1,124 @@
> > +/*
> > + * Simple dmaengine driver library
> > + *
> > + * extracted from shdma.c and headers
> > + *
> > + * Copyright (C) 2011-2012 Guennadi Liakhovetski <g.liakhovetski@gmx.de>
> > + * Copyright (C) 2009 Nobuhiro Iwamatsu <iwamatsu.nobuhiro@renesas.com>
> > + * Copyright (C) 2009 Renesas Solutions, Inc. All rights reserved.
> > + * Copyright (C) 2007 Freescale Semiconductor, Inc. All rights reserved.
> > + *
> > + * This is free software; you can redistribute it and/or modify
> > + * it under the terms of version 2 of the GNU General Public License as
> > + * published by the Free Software Foundation.
> > + */
> > +
> > +#ifndef DMA_SIMPLE_H
> > +#define DMA_SIMPLE_H
> > +
> > +#include <linux/dmaengine.h>
> > +#include <linux/interrupt.h>
> > +#include <linux/list.h>
> > +#include <linux/types.h>
> > +
> > +/**
> > + * dma_simple_pm_state - DMA channel PM state
> > + * DMA_SIMPLE_PM_ESTABLISHED:	either idle or during data transfer
> > + * DMA_SIMPLE_PM_BUSY:		during the transfer preparation, when we have to
> > + *				drop the lock temporarily
> > + * DMA_SIMPLE_PM_PENDING:	transfers pending
> > + */
> > +enum dma_simple_pm_state {
> > +	DMA_SIMPLE_PM_ESTABLISHED,
> > +	DMA_SIMPLE_PM_BUSY,
> > +	DMA_SIMPLE_PM_PENDING,
> > +};
> > +
> > +struct device;
> > +
> > +/*
> > + * Drivers, using this library are expected to embed struct dma_simple_dev,
> > + * struct dma_simple_chan, struct dma_simple_desc, and struct dma_simple_slave
> > + * in their respective device, channel, descriptor and slave objects.
> > + */
> > +
> > +struct dma_simple_slave {
> > +	unsigned int slave_id;
> > +};
> > +
> > +struct dma_simple_desc {
> > +	struct list_head node;
> > +	struct dma_async_tx_descriptor async_tx;
> > +	enum dma_transfer_direction direction;
> > +	dma_cookie_t cookie;
> > +	int chunks;
> > +	int mark;
> > +};
> > +
> > +struct dma_simple_chan {
> > +	dma_cookie_t completed_cookie;	/* The maximum cookie completed */
> > +	spinlock_t chan_lock;		/* Channel operation lock */
> > +	struct list_head ld_queue;	/* Link descriptors queue */
> > +	struct list_head ld_free;	/* Free link descriptors */
> > +	struct dma_chan dma_chan;	/* DMA channel */
> > +	struct device *dev;		/* Channel device */
> > +	void *desc;			/* buffer for descriptor array */
> > +	int desc_num;			/* desc count */
> > +	size_t max_xfer_len;		/* max transfer length */
> > +	int id;				/* Raw id of this channel */
> > +	int irq;			/* Channel IRQ */
> > +	enum dma_simple_pm_state pm_state;
> > +};
> > +
> > +/**
> > + * struct dma_simple_ops - simple DMA driver operations
> > + * desc_completed:	return true, if this is the descriptor, that just has
> > + *			completed (atomic)
> > + * halt_channel:	stop DMA channel operation (atomic)
> > + * channel_busy:	return true, if the channel is busy (atomic)
> > + * slave_addr:		return slave DMA address
> > + * desc_setup:		set up the hardware specific descriptor portion (atomic)
> > + * set_slave:		bind channel to a slave
> > + * setup_xfer:		configure channel hardware for operation (atomic)
> > + * start_xfer:		start the DMA transfer (atomic)
> > + * embedded_desc:	return Nth struct dma_simple_desc pointer from the
> > + *			descriptor array
> > + * chan_irq:		process channel IRQ, return true if a transfer has
> > + *			completed (atomic)
> > + */
> > +struct dma_simple_ops {
> > +	bool (*desc_completed)(struct dma_simple_chan *, struct dma_simple_desc *);
> > +	void (*halt_channel)(struct dma_simple_chan *);
> > +	bool (*channel_busy)(struct dma_simple_chan *);
> > +	dma_addr_t (*slave_addr)(struct dma_simple_chan *);
> > +	int (*desc_setup)(struct dma_simple_chan *, struct dma_simple_desc *,
> > +			  dma_addr_t, dma_addr_t, size_t *);
> > +	int (*set_slave)(struct dma_simple_chan *, struct dma_simple_slave *);
> > +	void (*setup_xfer)(struct dma_simple_chan *, struct dma_simple_slave *);
> > +	void (*start_xfer)(struct dma_simple_chan *, struct dma_simple_desc *);
> > +	struct dma_simple_desc *(*embedded_desc)(void *, int);
> > +	bool (*chan_irq)(struct dma_simple_chan *, int);
> > +};
> again so many callbacks... are they really required!!

Yes, they are all used, therefore they are required.

> > +
> > +struct dma_simple_dev {
> > +	struct dma_device dma_dev;
> > +	struct dma_simple_chan **schan;
> > +	const struct dma_simple_ops *ops;
> > +	size_t desc_size;
> > +};
> > +
> > +#define dma_simple_for_each_chan(c, d, i) for (i = 0, c = (d)->schan[0]; \
> > +				i < (d)->dma_dev.chancnt; c = (d)->schan[++i])
> > +
> > +int dma_simple_request_irq(struct dma_simple_chan *, int,
> > +			   unsigned long, const char *);
> > +void dma_simple_free_irq(struct dma_simple_chan *);
> > +bool dma_simple_reset(struct dma_simple_dev *sdev);
> > +void dma_simple_chan_probe(struct dma_simple_dev *sdev,
> > +			   struct dma_simple_chan *schan, int id);
> > +void dma_simple_chan_remove(struct dma_simple_chan *schan);
> > +int dma_simple_init(struct device *dev, struct dma_simple_dev *sdev,
> > +		    int chan_num);
> > +void dma_simple_cleanup(struct dma_simple_dev *sdev);
> > +
> > +#endif
> 
> Now I am confused on the intent of this library. It was proposed for
> helping dmacs like sh-mobile to support sg transfers in software which
> are not supported by hardware, but it seems this library is doing _much_
> more.
> IMHO, it should get inserted between dmaengine APIs and client driver,
> _only_ for sg transfers. The channel allocation etc belong to dmac.
> Further, the library should get notified by dmac based on the callbacks
> set for descriptor and then should call native driver while submitting
> the next one in queue...
> Rest of the stuff (if required) would not be generic and probably should
> be in arch specific directory.

Ok, let me explain a bit more the intensions of this library. You're 
right, it is indeed doing more than just descriptor list manipulations. 
But the list handling is the most complex part of the library, which is 
why I advertised it as a library for aiding in that.

As a matter of fact, this library appeared when an attempt has been made 
to extend the shdma library to support the SUDMAC controller:

http://marc.info/?l=linux-sh&m=132626708503808&w=2

As you can see there, the SUDMAC hardware is completely incompatible with 
the original sh-mobile DMAC engines, but the SUDMAC code was able to reuse 
the driver to 99% by only replacing hardware-specific parts. So, instead 
of doing that I proposed to extract the generic code to a library and only 
provide hardware-specific bits to handle DMAC and SUDMAC. Since descriptor 
management is the largest and most complex part of the library, that's 
also how I described it.

I think, it would be good to preserve the library design at large as is, 
maybe updating its description to more precisely explain what it does, 
port it on top of the slave-parameter in channel allocation patch, maybe 
add some cosmetic improvements. If you think as it stands it is not 
generic enough, because it takes too much freedom away from individual 
drivers, we can make a step back and make it sh-mobile specific to be used 
only by shdma and sudmac, and then see, whether any other drivers will 
want to use it and how it will then have to be adjusted.

Thanks
Guennadi
---
Guennadi Liakhovetski, Ph.D.
Freelance Open-Source Software Developer
http://www.open-technology.de/

^ permalink raw reply	[flat|nested] 33+ messages in thread

* Re: [PATCH/RFC] usb: fix renesas_usbhs to not schedule in atomic context
  2012-02-05 14:54               ` Felipe Balbi
@ 2012-02-06 10:11                 ` Guennadi Liakhovetski
  2012-02-06 10:31                   ` Felipe Balbi
  0 siblings, 1 reply; 33+ messages in thread
From: Guennadi Liakhovetski @ 2012-02-06 10:11 UTC (permalink / raw)
  To: Felipe Balbi
  Cc: Shimoda, Yoshihiro, linux-kernel, linux-sh, Vinod Koul,
	Magnus Damm, linux-mmc, alsa-devel, linux-serial, Paul Mundt,
	linux-usb

Hi Felipe

On Sun, 5 Feb 2012, Felipe Balbi wrote:

> Hi,
> 
> On Fri, Feb 03, 2012 at 04:43:20PM +0100, Guennadi Liakhovetski wrote:
> > The current renesas_usbhs driver triggers
> > 
> > BUG: scheduling while atomic: ksoftirqd/0/3/0x00000102
> > 
> > with enabled CONFIG_DEBUG_ATOMIC_SLEEP, by submitting DMA transfers from 
> > an atomic (tasklet) context, which is not supported by the shdma dmaengine 
> > driver. Fix it by switching to a work. Also simplify some list 
> > manipulations.
> 
> you are doing much more than what you say.

Are those two list macro changes what you refer to as "a lot?" ;-) You're 
right in principle, they are not directly related to the purpose of this 
patch, they are just something that occurred to me, while tracking down 
DMA packets. But yes, it can be extracted to a separate cosmetic patch...

> Also, instead of using a
> workqueue, have you considered using threaded_irqs ?
> 
> (I didn't go over the driver again to see if it makes sense to use
> threaded_irqs in this case, but doesn't hurt asking)

>From a first glance these tasklets are not directly enough related to 
IRQs, so, doing that is either impossible, or would require a _much_ 
deeper change to the driver and _this_ would indeed be a much bigger 
change than just fixing the Oops.

Thanks
Guennadi

> > Shimoda-san, this is the problem, that you were observing. However, it 
> > exists with the present version of shdma just as well as with the new one 
> > - on top of the simple DMA library. I marked it an RFC because (1) I only 
> > lightly tested it with the gadget device on mackerel with the mass storage 
> > gadget, and (2) I am somewhat concerned about races. Currently the work 
> > function runs with no locking and there are no usual cancel_work_sync() 
> > points in the patch. However, it has also been like this before with the 
> > tasklet implementation, which is not much better, and it looks like there 
> > are no asynchronous operations on the same packets like timeouts. Only 
> > asynchronous events, that I can think about are things like unloading the 
> > driver or unplugging the cable, but these have been there before too. It 
> > would become worse on SMP, I think. Comments welcome.
> > 
> > diff --git a/drivers/usb/renesas_usbhs/fifo.c b/drivers/usb/renesas_usbhs/fifo.c
> > index 72339bd..4d739ec 100644
> > --- a/drivers/usb/renesas_usbhs/fifo.c
> > +++ b/drivers/usb/renesas_usbhs/fifo.c
> > @@ -75,8 +75,7 @@ void usbhs_pkt_push(struct usbhs_pipe *pipe, struct usbhs_pkt *pkt,
> >  		pipe->handler = &usbhsf_null_handler;
> >  	}
> >  
> > -	list_del_init(&pkt->node);
> > -	list_add_tail(&pkt->node, &pipe->list);
> > +	list_move_tail(&pkt->node, &pipe->list);
> >  
> >  	/*
> >  	 * each pkt must hold own handler.
> > @@ -106,7 +105,7 @@ static struct usbhs_pkt *__usbhsf_pkt_get(struct usbhs_pipe *pipe)
> >  	if (list_empty(&pipe->list))
> >  		return NULL;
> >  
> > -	return list_entry(pipe->list.next, struct usbhs_pkt, node);
> > +	return list_first_entry(&pipe->list, struct usbhs_pkt, node);
> 
> these two hunks are not part of $SUBJECT
> 
> -- 
> balbi
> 

---
Guennadi Liakhovetski, Ph.D.
Freelance Open-Source Software Developer
http://www.open-technology.de/

^ permalink raw reply	[flat|nested] 33+ messages in thread

* Re: [PATCH/RFC] usb: fix renesas_usbhs to not schedule in atomic context
  2012-02-06 10:11                 ` Guennadi Liakhovetski
@ 2012-02-06 10:31                   ` Felipe Balbi
  0 siblings, 0 replies; 33+ messages in thread
From: Felipe Balbi @ 2012-02-06 10:31 UTC (permalink / raw)
  To: Guennadi Liakhovetski
  Cc: Felipe Balbi, Shimoda, Yoshihiro, linux-kernel, linux-sh,
	Vinod Koul, Magnus Damm, linux-mmc, alsa-devel, linux-serial,
	Paul Mundt, linux-usb

[-- Attachment #1: Type: text/plain, Size: 1585 bytes --]

Hi,

On Mon, Feb 06, 2012 at 11:11:45AM +0100, Guennadi Liakhovetski wrote:
> Hi Felipe
> 
> On Sun, 5 Feb 2012, Felipe Balbi wrote:
> 
> > Hi,
> > 
> > On Fri, Feb 03, 2012 at 04:43:20PM +0100, Guennadi Liakhovetski wrote:
> > > The current renesas_usbhs driver triggers
> > > 
> > > BUG: scheduling while atomic: ksoftirqd/0/3/0x00000102
> > > 
> > > with enabled CONFIG_DEBUG_ATOMIC_SLEEP, by submitting DMA transfers from 
> > > an atomic (tasklet) context, which is not supported by the shdma dmaengine 
> > > driver. Fix it by switching to a work. Also simplify some list 
> > > manipulations.
> > 
> > you are doing much more than what you say.
> 
> Are those two list macro changes what you refer to as "a lot?" ;-) You're 
> right in principle, they are not directly related to the purpose of this 
> patch, they are just something that occurred to me, while tracking down 
> DMA packets. But yes, it can be extracted to a separate cosmetic patch...

please do so ;-)

> > Also, instead of using a
> > workqueue, have you considered using threaded_irqs ?
> > 
> > (I didn't go over the driver again to see if it makes sense to use
> > threaded_irqs in this case, but doesn't hurt asking)
> 
> From a first glance these tasklets are not directly enough related to 
> IRQs, so, doing that is either impossible, or would require a _much_ 
> deeper change to the driver and _this_ would indeed be a much bigger 
> change than just fixing the Oops.

I see.. so please just split the list changes to separate patch and
resend.

-- 
balbi

[-- Attachment #2: Digital signature --]
[-- Type: application/pgp-signature, Size: 836 bytes --]

^ permalink raw reply	[flat|nested] 33+ messages in thread

end of thread, other threads:[~2012-02-06 10:32 UTC | newest]

Thread overview: 33+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2012-01-26 14:56 [PATCH 0/7 v2] extract a simple dmaengine library from shdma.c Guennadi Liakhovetski
2012-01-26 14:56 ` [PATCH 1/7 v2] dmaengine: add a simple dma library Guennadi Liakhovetski
2012-01-26 17:34   ` [alsa-devel] " Sascha Hauer
2012-01-26 21:07     ` Guennadi Liakhovetski
2012-01-27  8:37   ` Shimoda, Yoshihiro
2012-01-27  8:48     ` Guennadi Liakhovetski
2012-01-31 11:03       ` Shimoda, Yoshihiro
2012-02-01  0:52         ` Guennadi Liakhovetski
2012-02-01  2:47           ` Shimoda, Yoshihiro
2012-02-02 22:19         ` Guennadi Liakhovetski
2012-02-03  8:47           ` Shimoda, Yoshihiro
2012-02-03 10:21             ` Guennadi Liakhovetski
2012-02-03 15:43             ` [PATCH/RFC] usb: fix renesas_usbhs to not schedule in atomic context Guennadi Liakhovetski
2012-02-05 14:54               ` Felipe Balbi
2012-02-06 10:11                 ` Guennadi Liakhovetski
2012-02-06 10:31                   ` Felipe Balbi
2012-02-06  8:52               ` Shimoda, Yoshihiro
2012-01-30  8:32   ` [PATCH 1/7 v2] dmaengine: add a simple dma library Vinod Koul
2012-01-30  9:34     ` Guennadi Liakhovetski
2012-01-31  6:41       ` Vinod Koul
2012-01-31  8:59         ` Guennadi Liakhovetski
2012-02-01  5:38           ` Vinod Koul
2012-02-03 16:38             ` Guennadi Liakhovetski
2012-02-06  2:54   ` Vinod Koul
2012-02-06  9:53     ` Guennadi Liakhovetski
2012-01-26 14:56 ` [PATCH 2/7 v2] dma: shdma: prepare for simple DMA conversion Guennadi Liakhovetski
2012-01-26 14:56 ` [PATCH 3/7 v2] mmc: sh_mmcif: remove unneeded struct sh_mmcif_dma, prepare for simple DMA Guennadi Liakhovetski
2012-01-26 14:56 ` [PATCH 4/7 v2] mmc: sh_mobile_sdhi: prepare for conversion to " Guennadi Liakhovetski
2012-01-26 14:56 ` [PATCH 5/7 v2] serial: sh-sci: " Guennadi Liakhovetski
2012-01-26 14:56 ` [PATCH 6/7 v2] ASoC: SIU: " Guennadi Liakhovetski
2012-02-06  2:33   ` Vinod Koul
2012-02-06  9:11     ` Guennadi Liakhovetski
2012-01-26 14:56 ` [PATCH 7/7 v2] dma: shdma: convert to the simple DMA library Guennadi Liakhovetski

This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).