diff options
Diffstat (limited to 'drivers/dma/imx-dma.c')
| -rw-r--r-- | drivers/dma/imx-dma.c | 1138 | 
1 files changed, 984 insertions, 154 deletions
diff --git a/drivers/dma/imx-dma.c b/drivers/dma/imx-dma.c index f629e4961af..286660a12cc 100644 --- a/drivers/dma/imx-dma.c +++ b/drivers/dma/imx-dma.c @@ -5,6 +5,7 @@   * found on i.MX1/21/27   *   * Copyright 2010 Sascha Hauer, Pengutronix <s.hauer@pengutronix.de> + * Copyright 2012 Javier Martin, Vista Silicon <javier.martin@vista-silicon.com>   *   * The code contained herein is licensed under the GNU General Public   * License. You may obtain a copy of the GNU General Public License @@ -13,6 +14,7 @@   * http://www.opensource.org/licenses/gpl-license.html   * http://www.gnu.org/copyleft/gpl.html   */ +#include <linux/err.h>  #include <linux/init.h>  #include <linux/types.h>  #include <linux/mm.h> @@ -22,72 +24,649 @@  #include <linux/dma-mapping.h>  #include <linux/slab.h>  #include <linux/platform_device.h> +#include <linux/clk.h>  #include <linux/dmaengine.h> +#include <linux/module.h> +#include <linux/of_device.h> +#include <linux/of_dma.h>  #include <asm/irq.h> -#include <mach/dma-v1.h> -#include <mach/hardware.h> +#include <linux/platform_data/dma-imx.h> + +#include "dmaengine.h" +#define IMXDMA_MAX_CHAN_DESCRIPTORS	16 +#define IMX_DMA_CHANNELS  16 + +#define IMX_DMA_2D_SLOTS	2 +#define IMX_DMA_2D_SLOT_A	0 +#define IMX_DMA_2D_SLOT_B	1 + +#define IMX_DMA_LENGTH_LOOP	((unsigned int)-1) +#define IMX_DMA_MEMSIZE_32	(0 << 4) +#define IMX_DMA_MEMSIZE_8	(1 << 4) +#define IMX_DMA_MEMSIZE_16	(2 << 4) +#define IMX_DMA_TYPE_LINEAR	(0 << 10) +#define IMX_DMA_TYPE_2D		(1 << 10) +#define IMX_DMA_TYPE_FIFO	(2 << 10) + +#define IMX_DMA_ERR_BURST     (1 << 0) +#define IMX_DMA_ERR_REQUEST   (1 << 1) +#define IMX_DMA_ERR_TRANSFER  (1 << 2) +#define IMX_DMA_ERR_BUFFER    (1 << 3) +#define IMX_DMA_ERR_TIMEOUT   (1 << 4) + +#define DMA_DCR     0x00		/* Control Register */ +#define DMA_DISR    0x04		/* Interrupt status Register */ +#define DMA_DIMR    0x08		/* Interrupt mask Register */ +#define DMA_DBTOSR  0x0c		/* Burst timeout status Register */ +#define DMA_DRTOSR  0x10		/* Request timeout Register */ +#define DMA_DSESR   0x14		/* Transfer Error Status Register */ +#define DMA_DBOSR   0x18		/* Buffer overflow status Register */ +#define DMA_DBTOCR  0x1c		/* Burst timeout control Register */ +#define DMA_WSRA    0x40		/* W-Size Register A */ +#define DMA_XSRA    0x44		/* X-Size Register A */ +#define DMA_YSRA    0x48		/* Y-Size Register A */ +#define DMA_WSRB    0x4c		/* W-Size Register B */ +#define DMA_XSRB    0x50		/* X-Size Register B */ +#define DMA_YSRB    0x54		/* Y-Size Register B */ +#define DMA_SAR(x)  (0x80 + ((x) << 6))	/* Source Address Registers */ +#define DMA_DAR(x)  (0x84 + ((x) << 6))	/* Destination Address Registers */ +#define DMA_CNTR(x) (0x88 + ((x) << 6))	/* Count Registers */ +#define DMA_CCR(x)  (0x8c + ((x) << 6))	/* Control Registers */ +#define DMA_RSSR(x) (0x90 + ((x) << 6))	/* Request source select Registers */ +#define DMA_BLR(x)  (0x94 + ((x) << 6))	/* Burst length Registers */ +#define DMA_RTOR(x) (0x98 + ((x) << 6))	/* Request timeout Registers */ +#define DMA_BUCR(x) (0x98 + ((x) << 6))	/* Bus Utilization Registers */ +#define DMA_CCNR(x) (0x9C + ((x) << 6))	/* Channel counter Registers */ + +#define DCR_DRST           (1<<1) +#define DCR_DEN            (1<<0) +#define DBTOCR_EN          (1<<15) +#define DBTOCR_CNT(x)      ((x) & 0x7fff) +#define CNTR_CNT(x)        ((x) & 0xffffff) +#define CCR_ACRPT          (1<<14) +#define CCR_DMOD_LINEAR    (0x0 << 12) +#define CCR_DMOD_2D        (0x1 << 12) +#define CCR_DMOD_FIFO      (0x2 << 12) +#define CCR_DMOD_EOBFIFO   (0x3 << 12) +#define CCR_SMOD_LINEAR    (0x0 << 10) +#define CCR_SMOD_2D        (0x1 << 10) +#define CCR_SMOD_FIFO      (0x2 << 10) +#define CCR_SMOD_EOBFIFO   (0x3 << 10) +#define CCR_MDIR_DEC       (1<<9) +#define CCR_MSEL_B         (1<<8) +#define CCR_DSIZ_32        (0x0 << 6) +#define CCR_DSIZ_8         (0x1 << 6) +#define CCR_DSIZ_16        (0x2 << 6) +#define CCR_SSIZ_32        (0x0 << 4) +#define CCR_SSIZ_8         (0x1 << 4) +#define CCR_SSIZ_16        (0x2 << 4) +#define CCR_REN            (1<<3) +#define CCR_RPT            (1<<2) +#define CCR_FRC            (1<<1) +#define CCR_CEN            (1<<0) +#define RTOR_EN            (1<<15) +#define RTOR_CLK           (1<<14) +#define RTOR_PSC           (1<<13) + +enum  imxdma_prep_type { +	IMXDMA_DESC_MEMCPY, +	IMXDMA_DESC_INTERLEAVED, +	IMXDMA_DESC_SLAVE_SG, +	IMXDMA_DESC_CYCLIC, +}; + +struct imx_dma_2d_config { +	u16		xsr; +	u16		ysr; +	u16		wsr; +	int		count; +}; + +struct imxdma_desc { +	struct list_head		node; +	struct dma_async_tx_descriptor	desc; +	enum dma_status			status; +	dma_addr_t			src; +	dma_addr_t			dest; +	size_t				len; +	enum dma_transfer_direction	direction; +	enum imxdma_prep_type		type; +	/* For memcpy and interleaved */ +	unsigned int			config_port; +	unsigned int			config_mem; +	/* For interleaved transfers */ +	unsigned int			x; +	unsigned int			y; +	unsigned int			w; +	/* For slave sg and cyclic */ +	struct scatterlist		*sg; +	unsigned int			sgcount; +};  struct imxdma_channel { +	int				hw_chaining; +	struct timer_list		watchdog;  	struct imxdma_engine		*imxdma;  	unsigned int			channel; -	unsigned int			imxdma_channel; +	struct tasklet_struct		dma_tasklet; +	struct list_head		ld_free; +	struct list_head		ld_queue; +	struct list_head		ld_active; +	int				descs_allocated;  	enum dma_slave_buswidth		word_size;  	dma_addr_t			per_address;  	u32				watermark_level;  	struct dma_chan			chan; -	spinlock_t			lock;  	struct dma_async_tx_descriptor	desc; -	dma_cookie_t			last_completed;  	enum dma_status			status;  	int				dma_request;  	struct scatterlist		*sg_list; +	u32				ccr_from_device; +	u32				ccr_to_device; +	bool				enabled_2d; +	int				slot_2d;  }; -#define MAX_DMA_CHANNELS 8 +enum imx_dma_type { +	IMX1_DMA, +	IMX21_DMA, +	IMX27_DMA, +};  struct imxdma_engine {  	struct device			*dev; +	struct device_dma_parameters	dma_parms;  	struct dma_device		dma_device; -	struct imxdma_channel		channel[MAX_DMA_CHANNELS]; +	void __iomem			*base; +	struct clk			*dma_ahb; +	struct clk			*dma_ipg; +	spinlock_t			lock; +	struct imx_dma_2d_config	slots_2d[IMX_DMA_2D_SLOTS]; +	struct imxdma_channel		channel[IMX_DMA_CHANNELS]; +	enum imx_dma_type		devtype; +}; + +struct imxdma_filter_data { +	struct imxdma_engine	*imxdma; +	int			 request; +}; + +static struct platform_device_id imx_dma_devtype[] = { +	{ +		.name = "imx1-dma", +		.driver_data = IMX1_DMA, +	}, { +		.name = "imx21-dma", +		.driver_data = IMX21_DMA, +	}, { +		.name = "imx27-dma", +		.driver_data = IMX27_DMA, +	}, { +		/* sentinel */ +	} +}; +MODULE_DEVICE_TABLE(platform, imx_dma_devtype); + +static const struct of_device_id imx_dma_of_dev_id[] = { +	{ +		.compatible = "fsl,imx1-dma", +		.data = &imx_dma_devtype[IMX1_DMA], +	}, { +		.compatible = "fsl,imx21-dma", +		.data = &imx_dma_devtype[IMX21_DMA], +	}, { +		.compatible = "fsl,imx27-dma", +		.data = &imx_dma_devtype[IMX27_DMA], +	}, { +		/* sentinel */ +	}  }; +MODULE_DEVICE_TABLE(of, imx_dma_of_dev_id); + +static inline int is_imx1_dma(struct imxdma_engine *imxdma) +{ +	return imxdma->devtype == IMX1_DMA; +} + +static inline int is_imx21_dma(struct imxdma_engine *imxdma) +{ +	return imxdma->devtype == IMX21_DMA; +} + +static inline int is_imx27_dma(struct imxdma_engine *imxdma) +{ +	return imxdma->devtype == IMX27_DMA; +}  static struct imxdma_channel *to_imxdma_chan(struct dma_chan *chan)  {  	return container_of(chan, struct imxdma_channel, chan);  } -static void imxdma_handle(struct imxdma_channel *imxdmac) +static inline bool imxdma_chan_is_doing_cyclic(struct imxdma_channel *imxdmac) +{ +	struct imxdma_desc *desc; + +	if (!list_empty(&imxdmac->ld_active)) { +		desc = list_first_entry(&imxdmac->ld_active, struct imxdma_desc, +					node); +		if (desc->type == IMXDMA_DESC_CYCLIC) +			return true; +	} +	return false; +} + + + +static void imx_dmav1_writel(struct imxdma_engine *imxdma, unsigned val, +			     unsigned offset) +{ +	__raw_writel(val, imxdma->base + offset); +} + +static unsigned imx_dmav1_readl(struct imxdma_engine *imxdma, unsigned offset) +{ +	return __raw_readl(imxdma->base + offset); +} + +static int imxdma_hw_chain(struct imxdma_channel *imxdmac) +{ +	struct imxdma_engine *imxdma = imxdmac->imxdma; + +	if (is_imx27_dma(imxdma)) +		return imxdmac->hw_chaining; +	else +		return 0; +} + +/* + * imxdma_sg_next - prepare next chunk for scatter-gather DMA emulation + */ +static inline int imxdma_sg_next(struct imxdma_desc *d)  { -	if (imxdmac->desc.callback) -		imxdmac->desc.callback(imxdmac->desc.callback_param); -	imxdmac->last_completed = imxdmac->desc.cookie; +	struct imxdma_channel *imxdmac = to_imxdma_chan(d->desc.chan); +	struct imxdma_engine *imxdma = imxdmac->imxdma; +	struct scatterlist *sg = d->sg; +	unsigned long now; + +	now = min(d->len, sg_dma_len(sg)); +	if (d->len != IMX_DMA_LENGTH_LOOP) +		d->len -= now; + +	if (d->direction == DMA_DEV_TO_MEM) +		imx_dmav1_writel(imxdma, sg->dma_address, +				 DMA_DAR(imxdmac->channel)); +	else +		imx_dmav1_writel(imxdma, sg->dma_address, +				 DMA_SAR(imxdmac->channel)); + +	imx_dmav1_writel(imxdma, now, DMA_CNTR(imxdmac->channel)); + +	dev_dbg(imxdma->dev, " %s channel: %d dst 0x%08x, src 0x%08x, " +		"size 0x%08x\n", __func__, imxdmac->channel, +		 imx_dmav1_readl(imxdma, DMA_DAR(imxdmac->channel)), +		 imx_dmav1_readl(imxdma, DMA_SAR(imxdmac->channel)), +		 imx_dmav1_readl(imxdma, DMA_CNTR(imxdmac->channel))); + +	return now;  } -static void imxdma_irq_handler(int channel, void *data) +static void imxdma_enable_hw(struct imxdma_desc *d)  { -	struct imxdma_channel *imxdmac = data; +	struct imxdma_channel *imxdmac = to_imxdma_chan(d->desc.chan); +	struct imxdma_engine *imxdma = imxdmac->imxdma; +	int channel = imxdmac->channel; +	unsigned long flags; + +	dev_dbg(imxdma->dev, "%s channel %d\n", __func__, channel); + +	local_irq_save(flags); + +	imx_dmav1_writel(imxdma, 1 << channel, DMA_DISR); +	imx_dmav1_writel(imxdma, imx_dmav1_readl(imxdma, DMA_DIMR) & +			 ~(1 << channel), DMA_DIMR); +	imx_dmav1_writel(imxdma, imx_dmav1_readl(imxdma, DMA_CCR(channel)) | +			 CCR_CEN | CCR_ACRPT, DMA_CCR(channel)); + +	if (!is_imx1_dma(imxdma) && +			d->sg && imxdma_hw_chain(imxdmac)) { +		d->sg = sg_next(d->sg); +		if (d->sg) { +			u32 tmp; +			imxdma_sg_next(d); +			tmp = imx_dmav1_readl(imxdma, DMA_CCR(channel)); +			imx_dmav1_writel(imxdma, tmp | CCR_RPT | CCR_ACRPT, +					 DMA_CCR(channel)); +		} +	} -	imxdmac->status = DMA_SUCCESS; -	imxdma_handle(imxdmac); +	local_irq_restore(flags);  } -static void imxdma_err_handler(int channel, void *data, int error) +static void imxdma_disable_hw(struct imxdma_channel *imxdmac)  { -	struct imxdma_channel *imxdmac = data; +	struct imxdma_engine *imxdma = imxdmac->imxdma; +	int channel = imxdmac->channel; +	unsigned long flags; + +	dev_dbg(imxdma->dev, "%s channel %d\n", __func__, channel); -	imxdmac->status = DMA_ERROR; -	imxdma_handle(imxdmac); +	if (imxdma_hw_chain(imxdmac)) +		del_timer(&imxdmac->watchdog); + +	local_irq_save(flags); +	imx_dmav1_writel(imxdma, imx_dmav1_readl(imxdma, DMA_DIMR) | +			 (1 << channel), DMA_DIMR); +	imx_dmav1_writel(imxdma, imx_dmav1_readl(imxdma, DMA_CCR(channel)) & +			 ~CCR_CEN, DMA_CCR(channel)); +	imx_dmav1_writel(imxdma, 1 << channel, DMA_DISR); +	local_irq_restore(flags);  } -static void imxdma_progression(int channel, void *data, -		struct scatterlist *sg) +static void imxdma_watchdog(unsigned long data)  { -	struct imxdma_channel *imxdmac = data; +	struct imxdma_channel *imxdmac = (struct imxdma_channel *)data; +	struct imxdma_engine *imxdma = imxdmac->imxdma; +	int channel = imxdmac->channel; + +	imx_dmav1_writel(imxdma, 0, DMA_CCR(channel)); + +	/* Tasklet watchdog error handler */ +	tasklet_schedule(&imxdmac->dma_tasklet); +	dev_dbg(imxdma->dev, "channel %d: watchdog timeout!\n", +		imxdmac->channel); +} + +static irqreturn_t imxdma_err_handler(int irq, void *dev_id) +{ +	struct imxdma_engine *imxdma = dev_id; +	unsigned int err_mask; +	int i, disr; +	int errcode; + +	disr = imx_dmav1_readl(imxdma, DMA_DISR); + +	err_mask = imx_dmav1_readl(imxdma, DMA_DBTOSR) | +		   imx_dmav1_readl(imxdma, DMA_DRTOSR) | +		   imx_dmav1_readl(imxdma, DMA_DSESR)  | +		   imx_dmav1_readl(imxdma, DMA_DBOSR); + +	if (!err_mask) +		return IRQ_HANDLED; + +	imx_dmav1_writel(imxdma, disr & err_mask, DMA_DISR); + +	for (i = 0; i < IMX_DMA_CHANNELS; i++) { +		if (!(err_mask & (1 << i))) +			continue; +		errcode = 0; + +		if (imx_dmav1_readl(imxdma, DMA_DBTOSR) & (1 << i)) { +			imx_dmav1_writel(imxdma, 1 << i, DMA_DBTOSR); +			errcode |= IMX_DMA_ERR_BURST; +		} +		if (imx_dmav1_readl(imxdma, DMA_DRTOSR) & (1 << i)) { +			imx_dmav1_writel(imxdma, 1 << i, DMA_DRTOSR); +			errcode |= IMX_DMA_ERR_REQUEST; +		} +		if (imx_dmav1_readl(imxdma, DMA_DSESR) & (1 << i)) { +			imx_dmav1_writel(imxdma, 1 << i, DMA_DSESR); +			errcode |= IMX_DMA_ERR_TRANSFER; +		} +		if (imx_dmav1_readl(imxdma, DMA_DBOSR) & (1 << i)) { +			imx_dmav1_writel(imxdma, 1 << i, DMA_DBOSR); +			errcode |= IMX_DMA_ERR_BUFFER; +		} +		/* Tasklet error handler */ +		tasklet_schedule(&imxdma->channel[i].dma_tasklet); + +		dev_warn(imxdma->dev, +			 "DMA timeout on channel %d -%s%s%s%s\n", i, +			 errcode & IMX_DMA_ERR_BURST ?    " burst" : "", +			 errcode & IMX_DMA_ERR_REQUEST ?  " request" : "", +			 errcode & IMX_DMA_ERR_TRANSFER ? " transfer" : "", +			 errcode & IMX_DMA_ERR_BUFFER ?   " buffer" : ""); +	} +	return IRQ_HANDLED; +} + +static void dma_irq_handle_channel(struct imxdma_channel *imxdmac) +{ +	struct imxdma_engine *imxdma = imxdmac->imxdma; +	int chno = imxdmac->channel; +	struct imxdma_desc *desc; +	unsigned long flags; + +	spin_lock_irqsave(&imxdma->lock, flags); +	if (list_empty(&imxdmac->ld_active)) { +		spin_unlock_irqrestore(&imxdma->lock, flags); +		goto out; +	} + +	desc = list_first_entry(&imxdmac->ld_active, +				struct imxdma_desc, +				node); +	spin_unlock_irqrestore(&imxdma->lock, flags); + +	if (desc->sg) { +		u32 tmp; +		desc->sg = sg_next(desc->sg); + +		if (desc->sg) { +			imxdma_sg_next(desc); + +			tmp = imx_dmav1_readl(imxdma, DMA_CCR(chno)); + +			if (imxdma_hw_chain(imxdmac)) { +				/* FIXME: The timeout should probably be +				 * configurable +				 */ +				mod_timer(&imxdmac->watchdog, +					jiffies + msecs_to_jiffies(500)); + +				tmp |= CCR_CEN | CCR_RPT | CCR_ACRPT; +				imx_dmav1_writel(imxdma, tmp, DMA_CCR(chno)); +			} else { +				imx_dmav1_writel(imxdma, tmp & ~CCR_CEN, +						 DMA_CCR(chno)); +				tmp |= CCR_CEN; +			} + +			imx_dmav1_writel(imxdma, tmp, DMA_CCR(chno)); + +			if (imxdma_chan_is_doing_cyclic(imxdmac)) +				/* Tasklet progression */ +				tasklet_schedule(&imxdmac->dma_tasklet); + +			return; +		} + +		if (imxdma_hw_chain(imxdmac)) { +			del_timer(&imxdmac->watchdog); +			return; +		} +	} + +out: +	imx_dmav1_writel(imxdma, 0, DMA_CCR(chno)); +	/* Tasklet irq */ +	tasklet_schedule(&imxdmac->dma_tasklet); +} + +static irqreturn_t dma_irq_handler(int irq, void *dev_id) +{ +	struct imxdma_engine *imxdma = dev_id; +	int i, disr; + +	if (!is_imx1_dma(imxdma)) +		imxdma_err_handler(irq, dev_id); + +	disr = imx_dmav1_readl(imxdma, DMA_DISR); + +	dev_dbg(imxdma->dev, "%s called, disr=0x%08x\n", __func__, disr); + +	imx_dmav1_writel(imxdma, disr, DMA_DISR); +	for (i = 0; i < IMX_DMA_CHANNELS; i++) { +		if (disr & (1 << i)) +			dma_irq_handle_channel(&imxdma->channel[i]); +	} + +	return IRQ_HANDLED; +} + +static int imxdma_xfer_desc(struct imxdma_desc *d) +{ +	struct imxdma_channel *imxdmac = to_imxdma_chan(d->desc.chan); +	struct imxdma_engine *imxdma = imxdmac->imxdma; +	int slot = -1; +	int i; + +	/* Configure and enable */ +	switch (d->type) { +	case IMXDMA_DESC_INTERLEAVED: +		/* Try to get a free 2D slot */ +		for (i = 0; i < IMX_DMA_2D_SLOTS; i++) { +			if ((imxdma->slots_2d[i].count > 0) && +			((imxdma->slots_2d[i].xsr != d->x) || +			(imxdma->slots_2d[i].ysr != d->y) || +			(imxdma->slots_2d[i].wsr != d->w))) +				continue; +			slot = i; +			break; +		} +		if (slot < 0) +			return -EBUSY; + +		imxdma->slots_2d[slot].xsr = d->x; +		imxdma->slots_2d[slot].ysr = d->y; +		imxdma->slots_2d[slot].wsr = d->w; +		imxdma->slots_2d[slot].count++; + +		imxdmac->slot_2d = slot; +		imxdmac->enabled_2d = true; + +		if (slot == IMX_DMA_2D_SLOT_A) { +			d->config_mem &= ~CCR_MSEL_B; +			d->config_port &= ~CCR_MSEL_B; +			imx_dmav1_writel(imxdma, d->x, DMA_XSRA); +			imx_dmav1_writel(imxdma, d->y, DMA_YSRA); +			imx_dmav1_writel(imxdma, d->w, DMA_WSRA); +		} else { +			d->config_mem |= CCR_MSEL_B; +			d->config_port |= CCR_MSEL_B; +			imx_dmav1_writel(imxdma, d->x, DMA_XSRB); +			imx_dmav1_writel(imxdma, d->y, DMA_YSRB); +			imx_dmav1_writel(imxdma, d->w, DMA_WSRB); +		} +		/* +		 * We fall-through here intentionally, since a 2D transfer is +		 * similar to MEMCPY just adding the 2D slot configuration. +		 */ +	case IMXDMA_DESC_MEMCPY: +		imx_dmav1_writel(imxdma, d->src, DMA_SAR(imxdmac->channel)); +		imx_dmav1_writel(imxdma, d->dest, DMA_DAR(imxdmac->channel)); +		imx_dmav1_writel(imxdma, d->config_mem | (d->config_port << 2), +			 DMA_CCR(imxdmac->channel)); + +		imx_dmav1_writel(imxdma, d->len, DMA_CNTR(imxdmac->channel)); + +		dev_dbg(imxdma->dev, +			"%s channel: %d dest=0x%08llx src=0x%08llx dma_length=%zu\n", +			__func__, imxdmac->channel, +			(unsigned long long)d->dest, +			(unsigned long long)d->src, d->len); + +		break; +	/* Cyclic transfer is the same as slave_sg with special sg configuration. */ +	case IMXDMA_DESC_CYCLIC: +	case IMXDMA_DESC_SLAVE_SG: +		if (d->direction == DMA_DEV_TO_MEM) { +			imx_dmav1_writel(imxdma, imxdmac->per_address, +					 DMA_SAR(imxdmac->channel)); +			imx_dmav1_writel(imxdma, imxdmac->ccr_from_device, +					 DMA_CCR(imxdmac->channel)); + +			dev_dbg(imxdma->dev, +				"%s channel: %d sg=%p sgcount=%d total length=%zu dev_addr=0x%08llx (dev2mem)\n", +				__func__, imxdmac->channel, +				d->sg, d->sgcount, d->len, +				(unsigned long long)imxdmac->per_address); +		} else if (d->direction == DMA_MEM_TO_DEV) { +			imx_dmav1_writel(imxdma, imxdmac->per_address, +					 DMA_DAR(imxdmac->channel)); +			imx_dmav1_writel(imxdma, imxdmac->ccr_to_device, +					 DMA_CCR(imxdmac->channel)); + +			dev_dbg(imxdma->dev, +				"%s channel: %d sg=%p sgcount=%d total length=%zu dev_addr=0x%08llx (mem2dev)\n", +				__func__, imxdmac->channel, +				d->sg, d->sgcount, d->len, +				(unsigned long long)imxdmac->per_address); +		} else { +			dev_err(imxdma->dev, "%s channel: %d bad dma mode\n", +				__func__, imxdmac->channel); +			return -EINVAL; +		} + +		imxdma_sg_next(d); + +		break; +	default: +		return -EINVAL; +	} +	imxdma_enable_hw(d); +	return 0; +} + +static void imxdma_tasklet(unsigned long data) +{ +	struct imxdma_channel *imxdmac = (void *)data; +	struct imxdma_engine *imxdma = imxdmac->imxdma; +	struct imxdma_desc *desc; +	unsigned long flags; + +	spin_lock_irqsave(&imxdma->lock, flags); + +	if (list_empty(&imxdmac->ld_active)) { +		/* Someone might have called terminate all */ +		spin_unlock_irqrestore(&imxdma->lock, flags); +		return; +	} +	desc = list_first_entry(&imxdmac->ld_active, struct imxdma_desc, node); + +	/* If we are dealing with a cyclic descriptor, keep it on ld_active +	 * and dont mark the descriptor as complete. +	 * Only in non-cyclic cases it would be marked as complete +	 */ +	if (imxdma_chan_is_doing_cyclic(imxdmac)) +		goto out; +	else +		dma_cookie_complete(&desc->desc); + +	/* Free 2D slot if it was an interleaved transfer */ +	if (imxdmac->enabled_2d) { +		imxdma->slots_2d[imxdmac->slot_2d].count--; +		imxdmac->enabled_2d = false; +	} + +	list_move_tail(imxdmac->ld_active.next, &imxdmac->ld_free); + +	if (!list_empty(&imxdmac->ld_queue)) { +		desc = list_first_entry(&imxdmac->ld_queue, struct imxdma_desc, +					node); +		list_move_tail(imxdmac->ld_queue.next, &imxdmac->ld_active); +		if (imxdma_xfer_desc(desc) < 0) +			dev_warn(imxdma->dev, "%s: channel: %d couldn't xfer desc\n", +				 __func__, imxdmac->channel); +	} +out: +	spin_unlock_irqrestore(&imxdma->lock, flags); + +	if (desc->desc.callback) +		desc->desc.callback(desc->desc.callback_param); -	imxdmac->status = DMA_SUCCESS; -	imxdma_handle(imxdmac);  }  static int imxdma_control(struct dma_chan *chan, enum dma_ctrl_cmd cmd, @@ -95,16 +674,21 @@ static int imxdma_control(struct dma_chan *chan, enum dma_ctrl_cmd cmd,  {  	struct imxdma_channel *imxdmac = to_imxdma_chan(chan);  	struct dma_slave_config *dmaengine_cfg = (void *)arg; -	int ret; +	struct imxdma_engine *imxdma = imxdmac->imxdma; +	unsigned long flags;  	unsigned int mode = 0;  	switch (cmd) {  	case DMA_TERMINATE_ALL: -		imxdmac->status = DMA_ERROR; -		imx_dma_disable(imxdmac->imxdma_channel); +		imxdma_disable_hw(imxdmac); + +		spin_lock_irqsave(&imxdma->lock, flags); +		list_splice_tail_init(&imxdmac->ld_active, &imxdmac->ld_free); +		list_splice_tail_init(&imxdmac->ld_queue, &imxdmac->ld_free); +		spin_unlock_irqrestore(&imxdma->lock, flags);  		return 0;  	case DMA_SLAVE_CONFIG: -		if (dmaengine_cfg->direction == DMA_FROM_DEVICE) { +		if (dmaengine_cfg->direction == DMA_DEV_TO_MEM) {  			imxdmac->per_address = dmaengine_cfg->src_addr;  			imxdmac->watermark_level = dmaengine_cfg->src_maxburst;  			imxdmac->word_size = dmaengine_cfg->src_addr_width; @@ -126,15 +710,21 @@ static int imxdma_control(struct dma_chan *chan, enum dma_ctrl_cmd cmd,  			mode = IMX_DMA_MEMSIZE_32;  			break;  		} -		ret = imx_dma_config_channel(imxdmac->imxdma_channel, -				mode | IMX_DMA_TYPE_FIFO, -				IMX_DMA_MEMSIZE_32 | IMX_DMA_TYPE_LINEAR, -				imxdmac->dma_request, 1); -		if (ret) -			return ret; +		imxdmac->hw_chaining = 0; -		imx_dma_config_burstlen(imxdmac->imxdma_channel, imxdmac->watermark_level); +		imxdmac->ccr_from_device = (mode | IMX_DMA_TYPE_FIFO) | +			((IMX_DMA_MEMSIZE_32 | IMX_DMA_TYPE_LINEAR) << 2) | +			CCR_REN; +		imxdmac->ccr_to_device = +			(IMX_DMA_MEMSIZE_32 | IMX_DMA_TYPE_LINEAR) | +			((mode | IMX_DMA_TYPE_FIFO) << 2) | CCR_REN; +		imx_dmav1_writel(imxdma, imxdmac->dma_request, +				 DMA_RSSR(imxdmac->channel)); + +		/* Set burst length */ +		imx_dmav1_writel(imxdma, imxdmac->watermark_level * +				imxdmac->word_size, DMA_BLR(imxdmac->channel));  		return 0;  	default: @@ -148,43 +738,20 @@ static enum dma_status imxdma_tx_status(struct dma_chan *chan,  					    dma_cookie_t cookie,  					    struct dma_tx_state *txstate)  { -	struct imxdma_channel *imxdmac = to_imxdma_chan(chan); -	dma_cookie_t last_used; -	enum dma_status ret; - -	last_used = chan->cookie; - -	ret = dma_async_is_complete(cookie, imxdmac->last_completed, last_used); -	dma_set_tx_state(txstate, imxdmac->last_completed, last_used, 0); - -	return ret; -} - -static dma_cookie_t imxdma_assign_cookie(struct imxdma_channel *imxdma) -{ -	dma_cookie_t cookie = imxdma->chan.cookie; - -	if (++cookie < 0) -		cookie = 1; - -	imxdma->chan.cookie = cookie; -	imxdma->desc.cookie = cookie; - -	return cookie; +	return dma_cookie_status(chan, cookie, txstate);  }  static dma_cookie_t imxdma_tx_submit(struct dma_async_tx_descriptor *tx)  {  	struct imxdma_channel *imxdmac = to_imxdma_chan(tx->chan); +	struct imxdma_engine *imxdma = imxdmac->imxdma;  	dma_cookie_t cookie; +	unsigned long flags; -	spin_lock_irq(&imxdmac->lock); - -	cookie = imxdma_assign_cookie(imxdmac); - -	imx_dma_enable(imxdmac->imxdma_channel); - -	spin_unlock_irq(&imxdmac->lock); +	spin_lock_irqsave(&imxdma->lock, flags); +	list_move_tail(imxdmac->ld_free.next, &imxdmac->ld_queue); +	cookie = dma_cookie_assign(tx); +	spin_unlock_irqrestore(&imxdma->lock, flags);  	return cookie;  } @@ -194,91 +761,132 @@ static int imxdma_alloc_chan_resources(struct dma_chan *chan)  	struct imxdma_channel *imxdmac = to_imxdma_chan(chan);  	struct imx_dma_data *data = chan->private; -	imxdmac->dma_request = data->dma_request; +	if (data != NULL) +		imxdmac->dma_request = data->dma_request; + +	while (imxdmac->descs_allocated < IMXDMA_MAX_CHAN_DESCRIPTORS) { +		struct imxdma_desc *desc; -	dma_async_tx_descriptor_init(&imxdmac->desc, chan); -	imxdmac->desc.tx_submit = imxdma_tx_submit; -	/* txd.flags will be overwritten in prep funcs */ -	imxdmac->desc.flags = DMA_CTRL_ACK; +		desc = kzalloc(sizeof(*desc), GFP_KERNEL); +		if (!desc) +			break; +		__memzero(&desc->desc, sizeof(struct dma_async_tx_descriptor)); +		dma_async_tx_descriptor_init(&desc->desc, chan); +		desc->desc.tx_submit = imxdma_tx_submit; +		/* txd.flags will be overwritten in prep funcs */ +		desc->desc.flags = DMA_CTRL_ACK; +		desc->status = DMA_COMPLETE; + +		list_add_tail(&desc->node, &imxdmac->ld_free); +		imxdmac->descs_allocated++; +	} -	imxdmac->status = DMA_SUCCESS; +	if (!imxdmac->descs_allocated) +		return -ENOMEM; -	return 0; +	return imxdmac->descs_allocated;  }  static void imxdma_free_chan_resources(struct dma_chan *chan)  {  	struct imxdma_channel *imxdmac = to_imxdma_chan(chan); +	struct imxdma_engine *imxdma = imxdmac->imxdma; +	struct imxdma_desc *desc, *_desc; +	unsigned long flags; + +	spin_lock_irqsave(&imxdma->lock, flags); -	imx_dma_disable(imxdmac->imxdma_channel); +	imxdma_disable_hw(imxdmac); +	list_splice_tail_init(&imxdmac->ld_active, &imxdmac->ld_free); +	list_splice_tail_init(&imxdmac->ld_queue, &imxdmac->ld_free); -	if (imxdmac->sg_list) { -		kfree(imxdmac->sg_list); -		imxdmac->sg_list = NULL; +	spin_unlock_irqrestore(&imxdma->lock, flags); + +	list_for_each_entry_safe(desc, _desc, &imxdmac->ld_free, node) { +		kfree(desc); +		imxdmac->descs_allocated--;  	} +	INIT_LIST_HEAD(&imxdmac->ld_free); + +	kfree(imxdmac->sg_list); +	imxdmac->sg_list = NULL;  }  static struct dma_async_tx_descriptor *imxdma_prep_slave_sg(  		struct dma_chan *chan, struct scatterlist *sgl, -		unsigned int sg_len, enum dma_data_direction direction, -		unsigned long flags) +		unsigned int sg_len, enum dma_transfer_direction direction, +		unsigned long flags, void *context)  {  	struct imxdma_channel *imxdmac = to_imxdma_chan(chan);  	struct scatterlist *sg; -	int i, ret, dma_length = 0; -	unsigned int dmamode; +	int i, dma_length = 0; +	struct imxdma_desc *desc; -	if (imxdmac->status == DMA_IN_PROGRESS) +	if (list_empty(&imxdmac->ld_free) || +	    imxdma_chan_is_doing_cyclic(imxdmac))  		return NULL; -	imxdmac->status = DMA_IN_PROGRESS; +	desc = list_first_entry(&imxdmac->ld_free, struct imxdma_desc, node);  	for_each_sg(sgl, sg, sg_len, i) { -		dma_length += sg->length; +		dma_length += sg_dma_len(sg);  	} -	if (direction == DMA_FROM_DEVICE) -		dmamode = DMA_MODE_READ; -	else -		dmamode = DMA_MODE_WRITE; - -	ret = imx_dma_setup_sg(imxdmac->imxdma_channel, sgl, sg_len, -		 dma_length, imxdmac->per_address, dmamode); -	if (ret) +	switch (imxdmac->word_size) { +	case DMA_SLAVE_BUSWIDTH_4_BYTES: +		if (sg_dma_len(sgl) & 3 || sgl->dma_address & 3) +			return NULL; +		break; +	case DMA_SLAVE_BUSWIDTH_2_BYTES: +		if (sg_dma_len(sgl) & 1 || sgl->dma_address & 1) +			return NULL; +		break; +	case DMA_SLAVE_BUSWIDTH_1_BYTE: +		break; +	default:  		return NULL; +	} -	return &imxdmac->desc; +	desc->type = IMXDMA_DESC_SLAVE_SG; +	desc->sg = sgl; +	desc->sgcount = sg_len; +	desc->len = dma_length; +	desc->direction = direction; +	if (direction == DMA_DEV_TO_MEM) { +		desc->src = imxdmac->per_address; +	} else { +		desc->dest = imxdmac->per_address; +	} +	desc->desc.callback = NULL; +	desc->desc.callback_param = NULL; + +	return &desc->desc;  }  static struct dma_async_tx_descriptor *imxdma_prep_dma_cyclic(  		struct dma_chan *chan, dma_addr_t dma_addr, size_t buf_len, -		size_t period_len, enum dma_data_direction direction) +		size_t period_len, enum dma_transfer_direction direction, +		unsigned long flags, void *context)  {  	struct imxdma_channel *imxdmac = to_imxdma_chan(chan);  	struct imxdma_engine *imxdma = imxdmac->imxdma; -	int i, ret; +	struct imxdma_desc *desc; +	int i;  	unsigned int periods = buf_len / period_len; -	unsigned int dmamode; -	dev_dbg(imxdma->dev, "%s channel: %d buf_len=%d period_len=%d\n", +	dev_dbg(imxdma->dev, "%s channel: %d buf_len=%zu period_len=%zu\n",  			__func__, imxdmac->channel, buf_len, period_len); -	if (imxdmac->status == DMA_IN_PROGRESS) +	if (list_empty(&imxdmac->ld_free) || +	    imxdma_chan_is_doing_cyclic(imxdmac))  		return NULL; -	imxdmac->status = DMA_IN_PROGRESS; -	ret = imx_dma_setup_progression_handler(imxdmac->imxdma_channel, -			imxdma_progression); -	if (ret) { -		dev_err(imxdma->dev, "Failed to setup the DMA handler\n"); -		return NULL; -	} +	desc = list_first_entry(&imxdmac->ld_free, struct imxdma_desc, node); -	if (imxdmac->sg_list) -		kfree(imxdmac->sg_list); +	kfree(imxdmac->sg_list);  	imxdmac->sg_list = kcalloc(periods + 1, -			sizeof(struct scatterlist), GFP_KERNEL); +			sizeof(struct scatterlist), GFP_ATOMIC);  	if (!imxdmac->sg_list)  		return NULL; @@ -288,76 +896,285 @@ static struct dma_async_tx_descriptor *imxdma_prep_dma_cyclic(  		imxdmac->sg_list[i].page_link = 0;  		imxdmac->sg_list[i].offset = 0;  		imxdmac->sg_list[i].dma_address = dma_addr; -		imxdmac->sg_list[i].length = period_len; +		sg_dma_len(&imxdmac->sg_list[i]) = period_len;  		dma_addr += period_len;  	}  	/* close the loop */  	imxdmac->sg_list[periods].offset = 0; -	imxdmac->sg_list[periods].length = 0; +	sg_dma_len(&imxdmac->sg_list[periods]) = 0;  	imxdmac->sg_list[periods].page_link =  		((unsigned long)imxdmac->sg_list | 0x01) & ~0x02; -	if (direction == DMA_FROM_DEVICE) -		dmamode = DMA_MODE_READ; -	else -		dmamode = DMA_MODE_WRITE; +	desc->type = IMXDMA_DESC_CYCLIC; +	desc->sg = imxdmac->sg_list; +	desc->sgcount = periods; +	desc->len = IMX_DMA_LENGTH_LOOP; +	desc->direction = direction; +	if (direction == DMA_DEV_TO_MEM) { +		desc->src = imxdmac->per_address; +	} else { +		desc->dest = imxdmac->per_address; +	} +	desc->desc.callback = NULL; +	desc->desc.callback_param = NULL; + +	return &desc->desc; +} + +static struct dma_async_tx_descriptor *imxdma_prep_dma_memcpy( +	struct dma_chan *chan, dma_addr_t dest, +	dma_addr_t src, size_t len, unsigned long flags) +{ +	struct imxdma_channel *imxdmac = to_imxdma_chan(chan); +	struct imxdma_engine *imxdma = imxdmac->imxdma; +	struct imxdma_desc *desc; -	ret = imx_dma_setup_sg(imxdmac->imxdma_channel, imxdmac->sg_list, periods, -		 IMX_DMA_LENGTH_LOOP, imxdmac->per_address, dmamode); -	if (ret) +	dev_dbg(imxdma->dev, "%s channel: %d src=0x%llx dst=0x%llx len=%zu\n", +		__func__, imxdmac->channel, (unsigned long long)src, +		(unsigned long long)dest, len); + +	if (list_empty(&imxdmac->ld_free) || +	    imxdma_chan_is_doing_cyclic(imxdmac))  		return NULL; -	return &imxdmac->desc; +	desc = list_first_entry(&imxdmac->ld_free, struct imxdma_desc, node); + +	desc->type = IMXDMA_DESC_MEMCPY; +	desc->src = src; +	desc->dest = dest; +	desc->len = len; +	desc->direction = DMA_MEM_TO_MEM; +	desc->config_port = IMX_DMA_MEMSIZE_32 | IMX_DMA_TYPE_LINEAR; +	desc->config_mem = IMX_DMA_MEMSIZE_32 | IMX_DMA_TYPE_LINEAR; +	desc->desc.callback = NULL; +	desc->desc.callback_param = NULL; + +	return &desc->desc; +} + +static struct dma_async_tx_descriptor *imxdma_prep_dma_interleaved( +	struct dma_chan *chan, struct dma_interleaved_template *xt, +	unsigned long flags) +{ +	struct imxdma_channel *imxdmac = to_imxdma_chan(chan); +	struct imxdma_engine *imxdma = imxdmac->imxdma; +	struct imxdma_desc *desc; + +	dev_dbg(imxdma->dev, "%s channel: %d src_start=0x%llx dst_start=0x%llx\n" +		"   src_sgl=%s dst_sgl=%s numf=%zu frame_size=%zu\n", __func__, +		imxdmac->channel, (unsigned long long)xt->src_start, +		(unsigned long long) xt->dst_start, +		xt->src_sgl ? "true" : "false", xt->dst_sgl ? "true" : "false", +		xt->numf, xt->frame_size); + +	if (list_empty(&imxdmac->ld_free) || +	    imxdma_chan_is_doing_cyclic(imxdmac)) +		return NULL; + +	if (xt->frame_size != 1 || xt->numf <= 0 || xt->dir != DMA_MEM_TO_MEM) +		return NULL; + +	desc = list_first_entry(&imxdmac->ld_free, struct imxdma_desc, node); + +	desc->type = IMXDMA_DESC_INTERLEAVED; +	desc->src = xt->src_start; +	desc->dest = xt->dst_start; +	desc->x = xt->sgl[0].size; +	desc->y = xt->numf; +	desc->w = xt->sgl[0].icg + desc->x; +	desc->len = desc->x * desc->y; +	desc->direction = DMA_MEM_TO_MEM; +	desc->config_port = IMX_DMA_MEMSIZE_32; +	desc->config_mem = IMX_DMA_MEMSIZE_32; +	if (xt->src_sgl) +		desc->config_mem |= IMX_DMA_TYPE_2D; +	if (xt->dst_sgl) +		desc->config_port |= IMX_DMA_TYPE_2D; +	desc->desc.callback = NULL; +	desc->desc.callback_param = NULL; + +	return &desc->desc;  }  static void imxdma_issue_pending(struct dma_chan *chan)  { -	/* -	 * Nothing to do. We only have a single descriptor -	 */ +	struct imxdma_channel *imxdmac = to_imxdma_chan(chan); +	struct imxdma_engine *imxdma = imxdmac->imxdma; +	struct imxdma_desc *desc; +	unsigned long flags; + +	spin_lock_irqsave(&imxdma->lock, flags); +	if (list_empty(&imxdmac->ld_active) && +	    !list_empty(&imxdmac->ld_queue)) { +		desc = list_first_entry(&imxdmac->ld_queue, +					struct imxdma_desc, node); + +		if (imxdma_xfer_desc(desc) < 0) { +			dev_warn(imxdma->dev, +				 "%s: channel: %d couldn't issue DMA xfer\n", +				 __func__, imxdmac->channel); +		} else { +			list_move_tail(imxdmac->ld_queue.next, +				       &imxdmac->ld_active); +		} +	} +	spin_unlock_irqrestore(&imxdma->lock, flags);  } -static int __init imxdma_probe(struct platform_device *pdev) +static bool imxdma_filter_fn(struct dma_chan *chan, void *param) +{ +	struct imxdma_filter_data *fdata = param; +	struct imxdma_channel *imxdma_chan = to_imxdma_chan(chan); + +	if (chan->device->dev != fdata->imxdma->dev) +		return false; + +	imxdma_chan->dma_request = fdata->request; +	chan->private = NULL; + +	return true; +} + +static struct dma_chan *imxdma_xlate(struct of_phandle_args *dma_spec, +						struct of_dma *ofdma)  { +	int count = dma_spec->args_count; +	struct imxdma_engine *imxdma = ofdma->of_dma_data; +	struct imxdma_filter_data fdata = { +		.imxdma = imxdma, +	}; + +	if (count != 1) +		return NULL; + +	fdata.request = dma_spec->args[0]; + +	return dma_request_channel(imxdma->dma_device.cap_mask, +					imxdma_filter_fn, &fdata); +} + +static int __init imxdma_probe(struct platform_device *pdev) +	{  	struct imxdma_engine *imxdma; +	struct resource *res; +	const struct of_device_id *of_id;  	int ret, i; +	int irq, irq_err; -	imxdma = kzalloc(sizeof(*imxdma), GFP_KERNEL); +	of_id = of_match_device(imx_dma_of_dev_id, &pdev->dev); +	if (of_id) +		pdev->id_entry = of_id->data; + +	imxdma = devm_kzalloc(&pdev->dev, sizeof(*imxdma), GFP_KERNEL);  	if (!imxdma)  		return -ENOMEM; +	imxdma->dev = &pdev->dev; +	imxdma->devtype = pdev->id_entry->driver_data; + +	res = platform_get_resource(pdev, IORESOURCE_MEM, 0); +	imxdma->base = devm_ioremap_resource(&pdev->dev, res); +	if (IS_ERR(imxdma->base)) +		return PTR_ERR(imxdma->base); + +	irq = platform_get_irq(pdev, 0); +	if (irq < 0) +		return irq; + +	imxdma->dma_ipg = devm_clk_get(&pdev->dev, "ipg"); +	if (IS_ERR(imxdma->dma_ipg)) +		return PTR_ERR(imxdma->dma_ipg); + +	imxdma->dma_ahb = devm_clk_get(&pdev->dev, "ahb"); +	if (IS_ERR(imxdma->dma_ahb)) +		return PTR_ERR(imxdma->dma_ahb); + +	clk_prepare_enable(imxdma->dma_ipg); +	clk_prepare_enable(imxdma->dma_ahb); + +	/* reset DMA module */ +	imx_dmav1_writel(imxdma, DCR_DRST, DMA_DCR); + +	if (is_imx1_dma(imxdma)) { +		ret = devm_request_irq(&pdev->dev, irq, +				       dma_irq_handler, 0, "DMA", imxdma); +		if (ret) { +			dev_warn(imxdma->dev, "Can't register IRQ for DMA\n"); +			goto err; +		} + +		irq_err = platform_get_irq(pdev, 1); +		if (irq_err < 0) { +			ret = irq_err; +			goto err; +		} + +		ret = devm_request_irq(&pdev->dev, irq_err, +				       imxdma_err_handler, 0, "DMA", imxdma); +		if (ret) { +			dev_warn(imxdma->dev, "Can't register ERRIRQ for DMA\n"); +			goto err; +		} +	} + +	/* enable DMA module */ +	imx_dmav1_writel(imxdma, DCR_DEN, DMA_DCR); + +	/* clear all interrupts */ +	imx_dmav1_writel(imxdma, (1 << IMX_DMA_CHANNELS) - 1, DMA_DISR); + +	/* disable interrupts */ +	imx_dmav1_writel(imxdma, (1 << IMX_DMA_CHANNELS) - 1, DMA_DIMR); +  	INIT_LIST_HEAD(&imxdma->dma_device.channels); +	dma_cap_set(DMA_SLAVE, imxdma->dma_device.cap_mask); +	dma_cap_set(DMA_CYCLIC, imxdma->dma_device.cap_mask); +	dma_cap_set(DMA_MEMCPY, imxdma->dma_device.cap_mask); +	dma_cap_set(DMA_INTERLEAVE, imxdma->dma_device.cap_mask); + +	/* Initialize 2D global parameters */ +	for (i = 0; i < IMX_DMA_2D_SLOTS; i++) +		imxdma->slots_2d[i].count = 0; + +	spin_lock_init(&imxdma->lock); +  	/* Initialize channel parameters */ -	for (i = 0; i < MAX_DMA_CHANNELS; i++) { +	for (i = 0; i < IMX_DMA_CHANNELS; i++) {  		struct imxdma_channel *imxdmac = &imxdma->channel[i]; -		imxdmac->imxdma_channel = imx_dma_request_by_prio("dmaengine", -				DMA_PRIO_MEDIUM); -		if ((int)imxdmac->channel < 0) { -			ret = -ENODEV; -			goto err_init; +		if (!is_imx1_dma(imxdma)) { +			ret = devm_request_irq(&pdev->dev, irq + i, +					dma_irq_handler, 0, "DMA", imxdma); +			if (ret) { +				dev_warn(imxdma->dev, "Can't register IRQ %d " +					 "for DMA channel %d\n", +					 irq + i, i); +				goto err; +			} +			init_timer(&imxdmac->watchdog); +			imxdmac->watchdog.function = &imxdma_watchdog; +			imxdmac->watchdog.data = (unsigned long)imxdmac;  		} -		imx_dma_setup_handlers(imxdmac->imxdma_channel, -		       imxdma_irq_handler, imxdma_err_handler, imxdmac); -  		imxdmac->imxdma = imxdma; -		spin_lock_init(&imxdmac->lock); -		dma_cap_set(DMA_SLAVE, imxdma->dma_device.cap_mask); -		dma_cap_set(DMA_CYCLIC, imxdma->dma_device.cap_mask); +		INIT_LIST_HEAD(&imxdmac->ld_queue); +		INIT_LIST_HEAD(&imxdmac->ld_free); +		INIT_LIST_HEAD(&imxdmac->ld_active); +		tasklet_init(&imxdmac->dma_tasklet, imxdma_tasklet, +			     (unsigned long)imxdmac);  		imxdmac->chan.device = &imxdma->dma_device; -		imxdmac->chan.chan_id = i; +		dma_cookie_init(&imxdmac->chan);  		imxdmac->channel = i;  		/* Add the channel to the DMAC list */ -		list_add_tail(&imxdmac->chan.device_node, &imxdma->dma_device.channels); +		list_add_tail(&imxdmac->chan.device_node, +			      &imxdma->dma_device.channels);  	} -	imxdma->dev = &pdev->dev;  	imxdma->dma_device.dev = &pdev->dev;  	imxdma->dma_device.device_alloc_chan_resources = imxdma_alloc_chan_resources; @@ -365,43 +1182,53 @@ static int __init imxdma_probe(struct platform_device *pdev)  	imxdma->dma_device.device_tx_status = imxdma_tx_status;  	imxdma->dma_device.device_prep_slave_sg = imxdma_prep_slave_sg;  	imxdma->dma_device.device_prep_dma_cyclic = imxdma_prep_dma_cyclic; +	imxdma->dma_device.device_prep_dma_memcpy = imxdma_prep_dma_memcpy; +	imxdma->dma_device.device_prep_interleaved_dma = imxdma_prep_dma_interleaved;  	imxdma->dma_device.device_control = imxdma_control;  	imxdma->dma_device.device_issue_pending = imxdma_issue_pending;  	platform_set_drvdata(pdev, imxdma); +	imxdma->dma_device.copy_align = 2; /* 2^2 = 4 bytes alignment */ +	imxdma->dma_device.dev->dma_parms = &imxdma->dma_parms; +	dma_set_max_seg_size(imxdma->dma_device.dev, 0xffffff); +  	ret = dma_async_device_register(&imxdma->dma_device);  	if (ret) {  		dev_err(&pdev->dev, "unable to register\n"); -		goto err_init; +		goto err;  	} -	return 0; - -err_init: -	while (i-- >= 0) { -		struct imxdma_channel *imxdmac = &imxdma->channel[i]; -		imx_dma_free(imxdmac->imxdma_channel); +	if (pdev->dev.of_node) { +		ret = of_dma_controller_register(pdev->dev.of_node, +				imxdma_xlate, imxdma); +		if (ret) { +			dev_err(&pdev->dev, "unable to register of_dma_controller\n"); +			goto err_of_dma_controller; +		}  	} -	kfree(imxdma); +	return 0; + +err_of_dma_controller: +	dma_async_device_unregister(&imxdma->dma_device); +err: +	clk_disable_unprepare(imxdma->dma_ipg); +	clk_disable_unprepare(imxdma->dma_ahb);  	return ret;  } -static int __exit imxdma_remove(struct platform_device *pdev) +static int imxdma_remove(struct platform_device *pdev)  {  	struct imxdma_engine *imxdma = platform_get_drvdata(pdev); -	int i;          dma_async_device_unregister(&imxdma->dma_device); -	for (i = 0; i < MAX_DMA_CHANNELS; i++) { -		struct imxdma_channel *imxdmac = &imxdma->channel[i]; - -		 imx_dma_free(imxdmac->imxdma_channel); -	} +	if (pdev->dev.of_node) +		of_dma_controller_free(pdev->dev.of_node); -        kfree(imxdma); +	clk_disable_unprepare(imxdma->dma_ipg); +	clk_disable_unprepare(imxdma->dma_ahb);          return 0;  } @@ -409,8 +1236,11 @@ static int __exit imxdma_remove(struct platform_device *pdev)  static struct platform_driver imxdma_driver = {  	.driver		= {  		.name	= "imx-dma", +		.owner	= THIS_MODULE, +		.of_match_table = imx_dma_of_dev_id,  	}, -	.remove		= __exit_p(imxdma_remove), +	.id_table	= imx_dma_devtype, +	.remove		= imxdma_remove,  };  static int __init imxdma_module_init(void)  | 
