xref: /linux/drivers/dma/uniphier-xdmac.c (revision c771600c6af14749609b49565ffb4cac2959710d)
1667b9251SKunihiko Hayashi // SPDX-License-Identifier: GPL-2.0
2667b9251SKunihiko Hayashi /*
3667b9251SKunihiko Hayashi  * External DMA controller driver for UniPhier SoCs
4667b9251SKunihiko Hayashi  * Copyright 2019 Socionext Inc.
5667b9251SKunihiko Hayashi  * Author: Kunihiko Hayashi <hayashi.kunihiko@socionext.com>
6667b9251SKunihiko Hayashi  */
7667b9251SKunihiko Hayashi 
8667b9251SKunihiko Hayashi #include <linux/bitops.h>
9667b9251SKunihiko Hayashi #include <linux/bitfield.h>
10667b9251SKunihiko Hayashi #include <linux/iopoll.h>
11667b9251SKunihiko Hayashi #include <linux/module.h>
12667b9251SKunihiko Hayashi #include <linux/of.h>
13667b9251SKunihiko Hayashi #include <linux/of_dma.h>
14667b9251SKunihiko Hayashi #include <linux/platform_device.h>
157999096fSHerbert Xu #include <linux/slab.h>
16667b9251SKunihiko Hayashi 
17667b9251SKunihiko Hayashi #include "dmaengine.h"
18667b9251SKunihiko Hayashi #include "virt-dma.h"
19667b9251SKunihiko Hayashi 
20667b9251SKunihiko Hayashi #define XDMAC_CH_WIDTH		0x100
21667b9251SKunihiko Hayashi 
22667b9251SKunihiko Hayashi #define XDMAC_TFA		0x08
23667b9251SKunihiko Hayashi #define XDMAC_TFA_MCNT_MASK	GENMASK(23, 16)
24667b9251SKunihiko Hayashi #define XDMAC_TFA_MASK		GENMASK(5, 0)
25667b9251SKunihiko Hayashi #define XDMAC_SADM		0x10
26667b9251SKunihiko Hayashi #define XDMAC_SADM_STW_MASK	GENMASK(25, 24)
27667b9251SKunihiko Hayashi #define XDMAC_SADM_SAM		BIT(4)
28667b9251SKunihiko Hayashi #define XDMAC_SADM_SAM_FIXED	XDMAC_SADM_SAM
29667b9251SKunihiko Hayashi #define XDMAC_SADM_SAM_INC	0
30667b9251SKunihiko Hayashi #define XDMAC_DADM		0x14
31667b9251SKunihiko Hayashi #define XDMAC_DADM_DTW_MASK	XDMAC_SADM_STW_MASK
32667b9251SKunihiko Hayashi #define XDMAC_DADM_DAM		XDMAC_SADM_SAM
33667b9251SKunihiko Hayashi #define XDMAC_DADM_DAM_FIXED	XDMAC_SADM_SAM_FIXED
34667b9251SKunihiko Hayashi #define XDMAC_DADM_DAM_INC	XDMAC_SADM_SAM_INC
35667b9251SKunihiko Hayashi #define XDMAC_EXSAD		0x18
36667b9251SKunihiko Hayashi #define XDMAC_EXDAD		0x1c
37667b9251SKunihiko Hayashi #define XDMAC_SAD		0x20
38667b9251SKunihiko Hayashi #define XDMAC_DAD		0x24
39667b9251SKunihiko Hayashi #define XDMAC_ITS		0x28
40667b9251SKunihiko Hayashi #define XDMAC_ITS_MASK		GENMASK(25, 0)
41667b9251SKunihiko Hayashi #define XDMAC_TNUM		0x2c
42667b9251SKunihiko Hayashi #define XDMAC_TNUM_MASK		GENMASK(15, 0)
43667b9251SKunihiko Hayashi #define XDMAC_TSS		0x30
44667b9251SKunihiko Hayashi #define XDMAC_TSS_REQ		BIT(0)
45667b9251SKunihiko Hayashi #define XDMAC_IEN		0x34
46667b9251SKunihiko Hayashi #define XDMAC_IEN_ERRIEN	BIT(1)
47667b9251SKunihiko Hayashi #define XDMAC_IEN_ENDIEN	BIT(0)
48667b9251SKunihiko Hayashi #define XDMAC_STAT		0x40
49667b9251SKunihiko Hayashi #define XDMAC_STAT_TENF		BIT(0)
50667b9251SKunihiko Hayashi #define XDMAC_IR		0x44
51667b9251SKunihiko Hayashi #define XDMAC_IR_ERRF		BIT(1)
52667b9251SKunihiko Hayashi #define XDMAC_IR_ENDF		BIT(0)
53667b9251SKunihiko Hayashi #define XDMAC_ID		0x48
54667b9251SKunihiko Hayashi #define XDMAC_ID_ERRIDF		BIT(1)
55667b9251SKunihiko Hayashi #define XDMAC_ID_ENDIDF		BIT(0)
56667b9251SKunihiko Hayashi 
57667b9251SKunihiko Hayashi #define XDMAC_MAX_CHANS		16
58667b9251SKunihiko Hayashi #define XDMAC_INTERVAL_CLKS	20
59667b9251SKunihiko Hayashi #define XDMAC_MAX_WORDS		XDMAC_TNUM_MASK
60667b9251SKunihiko Hayashi 
61667b9251SKunihiko Hayashi /* cut lower bit for maintain alignment of maximum transfer size */
62667b9251SKunihiko Hayashi #define XDMAC_MAX_WORD_SIZE	(XDMAC_ITS_MASK & ~GENMASK(3, 0))
63667b9251SKunihiko Hayashi 
64667b9251SKunihiko Hayashi #define UNIPHIER_XDMAC_BUSWIDTHS \
65667b9251SKunihiko Hayashi 	(BIT(DMA_SLAVE_BUSWIDTH_1_BYTE) | \
66667b9251SKunihiko Hayashi 	 BIT(DMA_SLAVE_BUSWIDTH_2_BYTES) | \
67667b9251SKunihiko Hayashi 	 BIT(DMA_SLAVE_BUSWIDTH_4_BYTES) | \
68667b9251SKunihiko Hayashi 	 BIT(DMA_SLAVE_BUSWIDTH_8_BYTES))
69667b9251SKunihiko Hayashi 
70667b9251SKunihiko Hayashi struct uniphier_xdmac_desc_node {
71667b9251SKunihiko Hayashi 	dma_addr_t src;
72667b9251SKunihiko Hayashi 	dma_addr_t dst;
73667b9251SKunihiko Hayashi 	u32 burst_size;
74667b9251SKunihiko Hayashi 	u32 nr_burst;
75667b9251SKunihiko Hayashi };
76667b9251SKunihiko Hayashi 
77667b9251SKunihiko Hayashi struct uniphier_xdmac_desc {
78667b9251SKunihiko Hayashi 	struct virt_dma_desc vd;
79667b9251SKunihiko Hayashi 
80667b9251SKunihiko Hayashi 	unsigned int nr_node;
81667b9251SKunihiko Hayashi 	unsigned int cur_node;
82667b9251SKunihiko Hayashi 	enum dma_transfer_direction dir;
835a67a8f9SKees Cook 	struct uniphier_xdmac_desc_node nodes[] __counted_by(nr_node);
84667b9251SKunihiko Hayashi };
85667b9251SKunihiko Hayashi 
86667b9251SKunihiko Hayashi struct uniphier_xdmac_chan {
87667b9251SKunihiko Hayashi 	struct virt_dma_chan vc;
88667b9251SKunihiko Hayashi 	struct uniphier_xdmac_device *xdev;
89667b9251SKunihiko Hayashi 	struct uniphier_xdmac_desc *xd;
90667b9251SKunihiko Hayashi 	void __iomem *reg_ch_base;
91667b9251SKunihiko Hayashi 	struct dma_slave_config	sconfig;
92667b9251SKunihiko Hayashi 	int id;
93667b9251SKunihiko Hayashi 	unsigned int req_factor;
94667b9251SKunihiko Hayashi };
95667b9251SKunihiko Hayashi 
96667b9251SKunihiko Hayashi struct uniphier_xdmac_device {
97667b9251SKunihiko Hayashi 	struct dma_device ddev;
98667b9251SKunihiko Hayashi 	void __iomem *reg_base;
99667b9251SKunihiko Hayashi 	int nr_chans;
1007935de86SKees Cook 	struct uniphier_xdmac_chan channels[] __counted_by(nr_chans);
101667b9251SKunihiko Hayashi };
102667b9251SKunihiko Hayashi 
103667b9251SKunihiko Hayashi static struct uniphier_xdmac_chan *
to_uniphier_xdmac_chan(struct virt_dma_chan * vc)104667b9251SKunihiko Hayashi to_uniphier_xdmac_chan(struct virt_dma_chan *vc)
105667b9251SKunihiko Hayashi {
106667b9251SKunihiko Hayashi 	return container_of(vc, struct uniphier_xdmac_chan, vc);
107667b9251SKunihiko Hayashi }
108667b9251SKunihiko Hayashi 
109667b9251SKunihiko Hayashi static struct uniphier_xdmac_desc *
to_uniphier_xdmac_desc(struct virt_dma_desc * vd)110667b9251SKunihiko Hayashi to_uniphier_xdmac_desc(struct virt_dma_desc *vd)
111667b9251SKunihiko Hayashi {
112667b9251SKunihiko Hayashi 	return container_of(vd, struct uniphier_xdmac_desc, vd);
113667b9251SKunihiko Hayashi }
114667b9251SKunihiko Hayashi 
115667b9251SKunihiko Hayashi /* xc->vc.lock must be held by caller */
116667b9251SKunihiko Hayashi static struct uniphier_xdmac_desc *
uniphier_xdmac_next_desc(struct uniphier_xdmac_chan * xc)117667b9251SKunihiko Hayashi uniphier_xdmac_next_desc(struct uniphier_xdmac_chan *xc)
118667b9251SKunihiko Hayashi {
119667b9251SKunihiko Hayashi 	struct virt_dma_desc *vd;
120667b9251SKunihiko Hayashi 
121667b9251SKunihiko Hayashi 	vd = vchan_next_desc(&xc->vc);
122667b9251SKunihiko Hayashi 	if (!vd)
123667b9251SKunihiko Hayashi 		return NULL;
124667b9251SKunihiko Hayashi 
125667b9251SKunihiko Hayashi 	list_del(&vd->node);
126667b9251SKunihiko Hayashi 
127667b9251SKunihiko Hayashi 	return to_uniphier_xdmac_desc(vd);
128667b9251SKunihiko Hayashi }
129667b9251SKunihiko Hayashi 
130667b9251SKunihiko Hayashi /* xc->vc.lock must be held by caller */
uniphier_xdmac_chan_start(struct uniphier_xdmac_chan * xc,struct uniphier_xdmac_desc * xd)131667b9251SKunihiko Hayashi static void uniphier_xdmac_chan_start(struct uniphier_xdmac_chan *xc,
132667b9251SKunihiko Hayashi 				      struct uniphier_xdmac_desc *xd)
133667b9251SKunihiko Hayashi {
134105a8c52SKunihiko Hayashi 	u32 src_mode, src_width;
135105a8c52SKunihiko Hayashi 	u32 dst_mode, dst_width;
136105a8c52SKunihiko Hayashi 	dma_addr_t src_addr, dst_addr;
137667b9251SKunihiko Hayashi 	u32 val, its, tnum;
138667b9251SKunihiko Hayashi 	enum dma_slave_buswidth buswidth;
139667b9251SKunihiko Hayashi 
140667b9251SKunihiko Hayashi 	src_addr = xd->nodes[xd->cur_node].src;
141667b9251SKunihiko Hayashi 	dst_addr = xd->nodes[xd->cur_node].dst;
142667b9251SKunihiko Hayashi 	its      = xd->nodes[xd->cur_node].burst_size;
143667b9251SKunihiko Hayashi 	tnum     = xd->nodes[xd->cur_node].nr_burst;
144667b9251SKunihiko Hayashi 
145667b9251SKunihiko Hayashi 	/*
146667b9251SKunihiko Hayashi 	 * The width of MEM side must be 4 or 8 bytes, that does not
147667b9251SKunihiko Hayashi 	 * affect that of DEV side and transfer size.
148667b9251SKunihiko Hayashi 	 */
149667b9251SKunihiko Hayashi 	if (xd->dir == DMA_DEV_TO_MEM) {
150667b9251SKunihiko Hayashi 		src_mode = XDMAC_SADM_SAM_FIXED;
151667b9251SKunihiko Hayashi 		buswidth = xc->sconfig.src_addr_width;
152667b9251SKunihiko Hayashi 	} else {
153667b9251SKunihiko Hayashi 		src_mode = XDMAC_SADM_SAM_INC;
154667b9251SKunihiko Hayashi 		buswidth = DMA_SLAVE_BUSWIDTH_8_BYTES;
155667b9251SKunihiko Hayashi 	}
156667b9251SKunihiko Hayashi 	src_width = FIELD_PREP(XDMAC_SADM_STW_MASK, __ffs(buswidth));
157667b9251SKunihiko Hayashi 
158667b9251SKunihiko Hayashi 	if (xd->dir == DMA_MEM_TO_DEV) {
159667b9251SKunihiko Hayashi 		dst_mode = XDMAC_DADM_DAM_FIXED;
160667b9251SKunihiko Hayashi 		buswidth = xc->sconfig.dst_addr_width;
161667b9251SKunihiko Hayashi 	} else {
162667b9251SKunihiko Hayashi 		dst_mode = XDMAC_DADM_DAM_INC;
163667b9251SKunihiko Hayashi 		buswidth = DMA_SLAVE_BUSWIDTH_8_BYTES;
164667b9251SKunihiko Hayashi 	}
165667b9251SKunihiko Hayashi 	dst_width = FIELD_PREP(XDMAC_DADM_DTW_MASK, __ffs(buswidth));
166667b9251SKunihiko Hayashi 
167667b9251SKunihiko Hayashi 	/* setup transfer factor */
168667b9251SKunihiko Hayashi 	val = FIELD_PREP(XDMAC_TFA_MCNT_MASK, XDMAC_INTERVAL_CLKS);
169667b9251SKunihiko Hayashi 	val |= FIELD_PREP(XDMAC_TFA_MASK, xc->req_factor);
170667b9251SKunihiko Hayashi 	writel(val, xc->reg_ch_base + XDMAC_TFA);
171667b9251SKunihiko Hayashi 
172667b9251SKunihiko Hayashi 	/* setup the channel */
173667b9251SKunihiko Hayashi 	writel(lower_32_bits(src_addr), xc->reg_ch_base + XDMAC_SAD);
174667b9251SKunihiko Hayashi 	writel(upper_32_bits(src_addr), xc->reg_ch_base + XDMAC_EXSAD);
175667b9251SKunihiko Hayashi 
176667b9251SKunihiko Hayashi 	writel(lower_32_bits(dst_addr), xc->reg_ch_base + XDMAC_DAD);
177667b9251SKunihiko Hayashi 	writel(upper_32_bits(dst_addr), xc->reg_ch_base + XDMAC_EXDAD);
178667b9251SKunihiko Hayashi 
179667b9251SKunihiko Hayashi 	src_mode |= src_width;
180667b9251SKunihiko Hayashi 	dst_mode |= dst_width;
181667b9251SKunihiko Hayashi 	writel(src_mode, xc->reg_ch_base + XDMAC_SADM);
182667b9251SKunihiko Hayashi 	writel(dst_mode, xc->reg_ch_base + XDMAC_DADM);
183667b9251SKunihiko Hayashi 
184667b9251SKunihiko Hayashi 	writel(its, xc->reg_ch_base + XDMAC_ITS);
185667b9251SKunihiko Hayashi 	writel(tnum, xc->reg_ch_base + XDMAC_TNUM);
186667b9251SKunihiko Hayashi 
187667b9251SKunihiko Hayashi 	/* enable interrupt */
188667b9251SKunihiko Hayashi 	writel(XDMAC_IEN_ENDIEN | XDMAC_IEN_ERRIEN,
189667b9251SKunihiko Hayashi 	       xc->reg_ch_base + XDMAC_IEN);
190667b9251SKunihiko Hayashi 
191667b9251SKunihiko Hayashi 	/* start XDMAC */
192667b9251SKunihiko Hayashi 	val = readl(xc->reg_ch_base + XDMAC_TSS);
193667b9251SKunihiko Hayashi 	val |= XDMAC_TSS_REQ;
194667b9251SKunihiko Hayashi 	writel(val, xc->reg_ch_base + XDMAC_TSS);
195667b9251SKunihiko Hayashi }
196667b9251SKunihiko Hayashi 
197667b9251SKunihiko Hayashi /* xc->vc.lock must be held by caller */
uniphier_xdmac_chan_stop(struct uniphier_xdmac_chan * xc)198667b9251SKunihiko Hayashi static int uniphier_xdmac_chan_stop(struct uniphier_xdmac_chan *xc)
199667b9251SKunihiko Hayashi {
200667b9251SKunihiko Hayashi 	u32 val;
201667b9251SKunihiko Hayashi 
202667b9251SKunihiko Hayashi 	/* disable interrupt */
203667b9251SKunihiko Hayashi 	val = readl(xc->reg_ch_base + XDMAC_IEN);
204667b9251SKunihiko Hayashi 	val &= ~(XDMAC_IEN_ENDIEN | XDMAC_IEN_ERRIEN);
205667b9251SKunihiko Hayashi 	writel(val, xc->reg_ch_base + XDMAC_IEN);
206667b9251SKunihiko Hayashi 
207667b9251SKunihiko Hayashi 	/* stop XDMAC */
208667b9251SKunihiko Hayashi 	val = readl(xc->reg_ch_base + XDMAC_TSS);
209667b9251SKunihiko Hayashi 	val &= ~XDMAC_TSS_REQ;
210667b9251SKunihiko Hayashi 	writel(0, xc->reg_ch_base + XDMAC_TSS);
211667b9251SKunihiko Hayashi 
212667b9251SKunihiko Hayashi 	/* wait until transfer is stopped */
21355f24c27SKunihiko Hayashi 	return readl_poll_timeout_atomic(xc->reg_ch_base + XDMAC_STAT, val,
214667b9251SKunihiko Hayashi 					 !(val & XDMAC_STAT_TENF), 100, 1000);
215667b9251SKunihiko Hayashi }
216667b9251SKunihiko Hayashi 
217667b9251SKunihiko Hayashi /* xc->vc.lock must be held by caller */
uniphier_xdmac_start(struct uniphier_xdmac_chan * xc)218667b9251SKunihiko Hayashi static void uniphier_xdmac_start(struct uniphier_xdmac_chan *xc)
219667b9251SKunihiko Hayashi {
220667b9251SKunihiko Hayashi 	struct uniphier_xdmac_desc *xd;
221667b9251SKunihiko Hayashi 
222667b9251SKunihiko Hayashi 	xd = uniphier_xdmac_next_desc(xc);
223667b9251SKunihiko Hayashi 	if (xd)
224667b9251SKunihiko Hayashi 		uniphier_xdmac_chan_start(xc, xd);
225667b9251SKunihiko Hayashi 
226667b9251SKunihiko Hayashi 	/* set desc to chan regardless of xd is null */
227667b9251SKunihiko Hayashi 	xc->xd = xd;
228667b9251SKunihiko Hayashi }
229667b9251SKunihiko Hayashi 
uniphier_xdmac_chan_irq(struct uniphier_xdmac_chan * xc)230667b9251SKunihiko Hayashi static void uniphier_xdmac_chan_irq(struct uniphier_xdmac_chan *xc)
231667b9251SKunihiko Hayashi {
232667b9251SKunihiko Hayashi 	u32 stat;
233667b9251SKunihiko Hayashi 	int ret;
234667b9251SKunihiko Hayashi 
235667b9251SKunihiko Hayashi 	spin_lock(&xc->vc.lock);
236667b9251SKunihiko Hayashi 
237667b9251SKunihiko Hayashi 	stat = readl(xc->reg_ch_base + XDMAC_ID);
238667b9251SKunihiko Hayashi 
239667b9251SKunihiko Hayashi 	if (stat & XDMAC_ID_ERRIDF) {
240667b9251SKunihiko Hayashi 		ret = uniphier_xdmac_chan_stop(xc);
241667b9251SKunihiko Hayashi 		if (ret)
242667b9251SKunihiko Hayashi 			dev_err(xc->xdev->ddev.dev,
243667b9251SKunihiko Hayashi 				"DMA transfer error with aborting issue\n");
244667b9251SKunihiko Hayashi 		else
245667b9251SKunihiko Hayashi 			dev_err(xc->xdev->ddev.dev,
246667b9251SKunihiko Hayashi 				"DMA transfer error\n");
247667b9251SKunihiko Hayashi 
248667b9251SKunihiko Hayashi 	} else if ((stat & XDMAC_ID_ENDIDF) && xc->xd) {
249667b9251SKunihiko Hayashi 		xc->xd->cur_node++;
250667b9251SKunihiko Hayashi 		if (xc->xd->cur_node >= xc->xd->nr_node) {
251667b9251SKunihiko Hayashi 			vchan_cookie_complete(&xc->xd->vd);
252667b9251SKunihiko Hayashi 			uniphier_xdmac_start(xc);
253667b9251SKunihiko Hayashi 		} else {
254667b9251SKunihiko Hayashi 			uniphier_xdmac_chan_start(xc, xc->xd);
255667b9251SKunihiko Hayashi 		}
256667b9251SKunihiko Hayashi 	}
257667b9251SKunihiko Hayashi 
258667b9251SKunihiko Hayashi 	/* write bits to clear */
259667b9251SKunihiko Hayashi 	writel(stat, xc->reg_ch_base + XDMAC_IR);
260667b9251SKunihiko Hayashi 
261667b9251SKunihiko Hayashi 	spin_unlock(&xc->vc.lock);
262667b9251SKunihiko Hayashi }
263667b9251SKunihiko Hayashi 
uniphier_xdmac_irq_handler(int irq,void * dev_id)264667b9251SKunihiko Hayashi static irqreturn_t uniphier_xdmac_irq_handler(int irq, void *dev_id)
265667b9251SKunihiko Hayashi {
266667b9251SKunihiko Hayashi 	struct uniphier_xdmac_device *xdev = dev_id;
267667b9251SKunihiko Hayashi 	int i;
268667b9251SKunihiko Hayashi 
269667b9251SKunihiko Hayashi 	for (i = 0; i < xdev->nr_chans; i++)
270667b9251SKunihiko Hayashi 		uniphier_xdmac_chan_irq(&xdev->channels[i]);
271667b9251SKunihiko Hayashi 
272667b9251SKunihiko Hayashi 	return IRQ_HANDLED;
273667b9251SKunihiko Hayashi }
274667b9251SKunihiko Hayashi 
uniphier_xdmac_free_chan_resources(struct dma_chan * chan)275667b9251SKunihiko Hayashi static void uniphier_xdmac_free_chan_resources(struct dma_chan *chan)
276667b9251SKunihiko Hayashi {
277667b9251SKunihiko Hayashi 	vchan_free_chan_resources(to_virt_chan(chan));
278667b9251SKunihiko Hayashi }
279667b9251SKunihiko Hayashi 
280667b9251SKunihiko Hayashi static struct dma_async_tx_descriptor *
uniphier_xdmac_prep_dma_memcpy(struct dma_chan * chan,dma_addr_t dst,dma_addr_t src,size_t len,unsigned long flags)281667b9251SKunihiko Hayashi uniphier_xdmac_prep_dma_memcpy(struct dma_chan *chan, dma_addr_t dst,
282667b9251SKunihiko Hayashi 			       dma_addr_t src, size_t len, unsigned long flags)
283667b9251SKunihiko Hayashi {
284667b9251SKunihiko Hayashi 	struct virt_dma_chan *vc = to_virt_chan(chan);
285667b9251SKunihiko Hayashi 	struct uniphier_xdmac_desc *xd;
286667b9251SKunihiko Hayashi 	unsigned int nr;
287667b9251SKunihiko Hayashi 	size_t burst_size, tlen;
288667b9251SKunihiko Hayashi 	int i;
289667b9251SKunihiko Hayashi 
290667b9251SKunihiko Hayashi 	if (len > XDMAC_MAX_WORD_SIZE * XDMAC_MAX_WORDS)
291667b9251SKunihiko Hayashi 		return NULL;
292667b9251SKunihiko Hayashi 
293667b9251SKunihiko Hayashi 	nr = 1 + len / XDMAC_MAX_WORD_SIZE;
294667b9251SKunihiko Hayashi 
295667b9251SKunihiko Hayashi 	xd = kzalloc(struct_size(xd, nodes, nr), GFP_NOWAIT);
296667b9251SKunihiko Hayashi 	if (!xd)
297667b9251SKunihiko Hayashi 		return NULL;
2985a67a8f9SKees Cook 	xd->nr_node = nr;
299667b9251SKunihiko Hayashi 
300667b9251SKunihiko Hayashi 	for (i = 0; i < nr; i++) {
301667b9251SKunihiko Hayashi 		burst_size = min_t(size_t, len, XDMAC_MAX_WORD_SIZE);
302667b9251SKunihiko Hayashi 		xd->nodes[i].src = src;
303667b9251SKunihiko Hayashi 		xd->nodes[i].dst = dst;
304667b9251SKunihiko Hayashi 		xd->nodes[i].burst_size = burst_size;
305667b9251SKunihiko Hayashi 		xd->nodes[i].nr_burst = len / burst_size;
306667b9251SKunihiko Hayashi 		tlen = rounddown(len, burst_size);
307667b9251SKunihiko Hayashi 		src += tlen;
308667b9251SKunihiko Hayashi 		dst += tlen;
309667b9251SKunihiko Hayashi 		len -= tlen;
310667b9251SKunihiko Hayashi 	}
311667b9251SKunihiko Hayashi 
312667b9251SKunihiko Hayashi 	xd->dir = DMA_MEM_TO_MEM;
313667b9251SKunihiko Hayashi 	xd->cur_node = 0;
314667b9251SKunihiko Hayashi 
315667b9251SKunihiko Hayashi 	return vchan_tx_prep(vc, &xd->vd, flags);
316667b9251SKunihiko Hayashi }
317667b9251SKunihiko Hayashi 
318667b9251SKunihiko Hayashi static struct dma_async_tx_descriptor *
uniphier_xdmac_prep_slave_sg(struct dma_chan * chan,struct scatterlist * sgl,unsigned int sg_len,enum dma_transfer_direction direction,unsigned long flags,void * context)319667b9251SKunihiko Hayashi uniphier_xdmac_prep_slave_sg(struct dma_chan *chan, struct scatterlist *sgl,
320667b9251SKunihiko Hayashi 			     unsigned int sg_len,
321667b9251SKunihiko Hayashi 			     enum dma_transfer_direction direction,
322667b9251SKunihiko Hayashi 			     unsigned long flags, void *context)
323667b9251SKunihiko Hayashi {
324667b9251SKunihiko Hayashi 	struct virt_dma_chan *vc = to_virt_chan(chan);
325667b9251SKunihiko Hayashi 	struct uniphier_xdmac_chan *xc = to_uniphier_xdmac_chan(vc);
326667b9251SKunihiko Hayashi 	struct uniphier_xdmac_desc *xd;
327667b9251SKunihiko Hayashi 	struct scatterlist *sg;
328667b9251SKunihiko Hayashi 	enum dma_slave_buswidth buswidth;
329667b9251SKunihiko Hayashi 	u32 maxburst;
330667b9251SKunihiko Hayashi 	int i;
331667b9251SKunihiko Hayashi 
332667b9251SKunihiko Hayashi 	if (!is_slave_direction(direction))
333667b9251SKunihiko Hayashi 		return NULL;
334667b9251SKunihiko Hayashi 
335667b9251SKunihiko Hayashi 	if (direction == DMA_DEV_TO_MEM) {
336667b9251SKunihiko Hayashi 		buswidth = xc->sconfig.src_addr_width;
337667b9251SKunihiko Hayashi 		maxburst = xc->sconfig.src_maxburst;
338667b9251SKunihiko Hayashi 	} else {
339667b9251SKunihiko Hayashi 		buswidth = xc->sconfig.dst_addr_width;
340667b9251SKunihiko Hayashi 		maxburst = xc->sconfig.dst_maxburst;
341667b9251SKunihiko Hayashi 	}
342667b9251SKunihiko Hayashi 
343667b9251SKunihiko Hayashi 	if (!maxburst)
344667b9251SKunihiko Hayashi 		maxburst = 1;
345667b9251SKunihiko Hayashi 	if (maxburst > xc->xdev->ddev.max_burst) {
346667b9251SKunihiko Hayashi 		dev_err(xc->xdev->ddev.dev,
347667b9251SKunihiko Hayashi 			"Exceed maximum number of burst words\n");
348667b9251SKunihiko Hayashi 		return NULL;
349667b9251SKunihiko Hayashi 	}
350667b9251SKunihiko Hayashi 
351667b9251SKunihiko Hayashi 	xd = kzalloc(struct_size(xd, nodes, sg_len), GFP_NOWAIT);
352667b9251SKunihiko Hayashi 	if (!xd)
353667b9251SKunihiko Hayashi 		return NULL;
3545a67a8f9SKees Cook 	xd->nr_node = sg_len;
355667b9251SKunihiko Hayashi 
356667b9251SKunihiko Hayashi 	for_each_sg(sgl, sg, sg_len, i) {
357667b9251SKunihiko Hayashi 		xd->nodes[i].src = (direction == DMA_DEV_TO_MEM)
358667b9251SKunihiko Hayashi 			? xc->sconfig.src_addr : sg_dma_address(sg);
359667b9251SKunihiko Hayashi 		xd->nodes[i].dst = (direction == DMA_MEM_TO_DEV)
360667b9251SKunihiko Hayashi 			? xc->sconfig.dst_addr : sg_dma_address(sg);
361667b9251SKunihiko Hayashi 		xd->nodes[i].burst_size = maxburst * buswidth;
362667b9251SKunihiko Hayashi 		xd->nodes[i].nr_burst =
363667b9251SKunihiko Hayashi 			sg_dma_len(sg) / xd->nodes[i].burst_size;
364667b9251SKunihiko Hayashi 
365667b9251SKunihiko Hayashi 		/*
366667b9251SKunihiko Hayashi 		 * Currently transfer that size doesn't align the unit size
367667b9251SKunihiko Hayashi 		 * (the number of burst words * bus-width) is not allowed,
368667b9251SKunihiko Hayashi 		 * because the driver does not support the way to transfer
369667b9251SKunihiko Hayashi 		 * residue size. As a matter of fact, in order to transfer
370667b9251SKunihiko Hayashi 		 * arbitrary size, 'src_maxburst' or 'dst_maxburst' of
371667b9251SKunihiko Hayashi 		 * dma_slave_config must be 1.
372667b9251SKunihiko Hayashi 		 */
373667b9251SKunihiko Hayashi 		if (sg_dma_len(sg) % xd->nodes[i].burst_size) {
374667b9251SKunihiko Hayashi 			dev_err(xc->xdev->ddev.dev,
375667b9251SKunihiko Hayashi 				"Unaligned transfer size: %d", sg_dma_len(sg));
376667b9251SKunihiko Hayashi 			kfree(xd);
377667b9251SKunihiko Hayashi 			return NULL;
378667b9251SKunihiko Hayashi 		}
379667b9251SKunihiko Hayashi 
380667b9251SKunihiko Hayashi 		if (xd->nodes[i].nr_burst > XDMAC_MAX_WORDS) {
381667b9251SKunihiko Hayashi 			dev_err(xc->xdev->ddev.dev,
382667b9251SKunihiko Hayashi 				"Exceed maximum transfer size");
383667b9251SKunihiko Hayashi 			kfree(xd);
384667b9251SKunihiko Hayashi 			return NULL;
385667b9251SKunihiko Hayashi 		}
386667b9251SKunihiko Hayashi 	}
387667b9251SKunihiko Hayashi 
388667b9251SKunihiko Hayashi 	xd->dir = direction;
389667b9251SKunihiko Hayashi 	xd->cur_node = 0;
390667b9251SKunihiko Hayashi 
391667b9251SKunihiko Hayashi 	return vchan_tx_prep(vc, &xd->vd, flags);
392667b9251SKunihiko Hayashi }
393667b9251SKunihiko Hayashi 
uniphier_xdmac_slave_config(struct dma_chan * chan,struct dma_slave_config * config)394667b9251SKunihiko Hayashi static int uniphier_xdmac_slave_config(struct dma_chan *chan,
395667b9251SKunihiko Hayashi 				       struct dma_slave_config *config)
396667b9251SKunihiko Hayashi {
397667b9251SKunihiko Hayashi 	struct virt_dma_chan *vc = to_virt_chan(chan);
398667b9251SKunihiko Hayashi 	struct uniphier_xdmac_chan *xc = to_uniphier_xdmac_chan(vc);
399667b9251SKunihiko Hayashi 
400667b9251SKunihiko Hayashi 	memcpy(&xc->sconfig, config, sizeof(*config));
401667b9251SKunihiko Hayashi 
402667b9251SKunihiko Hayashi 	return 0;
403667b9251SKunihiko Hayashi }
404667b9251SKunihiko Hayashi 
uniphier_xdmac_terminate_all(struct dma_chan * chan)405667b9251SKunihiko Hayashi static int uniphier_xdmac_terminate_all(struct dma_chan *chan)
406667b9251SKunihiko Hayashi {
407667b9251SKunihiko Hayashi 	struct virt_dma_chan *vc = to_virt_chan(chan);
408667b9251SKunihiko Hayashi 	struct uniphier_xdmac_chan *xc = to_uniphier_xdmac_chan(vc);
409667b9251SKunihiko Hayashi 	unsigned long flags;
410667b9251SKunihiko Hayashi 	int ret = 0;
411667b9251SKunihiko Hayashi 	LIST_HEAD(head);
412667b9251SKunihiko Hayashi 
413667b9251SKunihiko Hayashi 	spin_lock_irqsave(&vc->lock, flags);
414667b9251SKunihiko Hayashi 
415667b9251SKunihiko Hayashi 	if (xc->xd) {
416667b9251SKunihiko Hayashi 		vchan_terminate_vdesc(&xc->xd->vd);
417667b9251SKunihiko Hayashi 		xc->xd = NULL;
418667b9251SKunihiko Hayashi 		ret = uniphier_xdmac_chan_stop(xc);
419667b9251SKunihiko Hayashi 	}
420667b9251SKunihiko Hayashi 
421667b9251SKunihiko Hayashi 	vchan_get_all_descriptors(vc, &head);
422667b9251SKunihiko Hayashi 
423667b9251SKunihiko Hayashi 	spin_unlock_irqrestore(&vc->lock, flags);
424667b9251SKunihiko Hayashi 
425667b9251SKunihiko Hayashi 	vchan_dma_desc_free_list(vc, &head);
426667b9251SKunihiko Hayashi 
427667b9251SKunihiko Hayashi 	return ret;
428667b9251SKunihiko Hayashi }
429667b9251SKunihiko Hayashi 
uniphier_xdmac_synchronize(struct dma_chan * chan)430667b9251SKunihiko Hayashi static void uniphier_xdmac_synchronize(struct dma_chan *chan)
431667b9251SKunihiko Hayashi {
432667b9251SKunihiko Hayashi 	vchan_synchronize(to_virt_chan(chan));
433667b9251SKunihiko Hayashi }
434667b9251SKunihiko Hayashi 
uniphier_xdmac_issue_pending(struct dma_chan * chan)435667b9251SKunihiko Hayashi static void uniphier_xdmac_issue_pending(struct dma_chan *chan)
436667b9251SKunihiko Hayashi {
437667b9251SKunihiko Hayashi 	struct virt_dma_chan *vc = to_virt_chan(chan);
438667b9251SKunihiko Hayashi 	struct uniphier_xdmac_chan *xc = to_uniphier_xdmac_chan(vc);
439667b9251SKunihiko Hayashi 	unsigned long flags;
440667b9251SKunihiko Hayashi 
441667b9251SKunihiko Hayashi 	spin_lock_irqsave(&vc->lock, flags);
442667b9251SKunihiko Hayashi 
443667b9251SKunihiko Hayashi 	if (vchan_issue_pending(vc) && !xc->xd)
444667b9251SKunihiko Hayashi 		uniphier_xdmac_start(xc);
445667b9251SKunihiko Hayashi 
446667b9251SKunihiko Hayashi 	spin_unlock_irqrestore(&vc->lock, flags);
447667b9251SKunihiko Hayashi }
448667b9251SKunihiko Hayashi 
uniphier_xdmac_desc_free(struct virt_dma_desc * vd)449667b9251SKunihiko Hayashi static void uniphier_xdmac_desc_free(struct virt_dma_desc *vd)
450667b9251SKunihiko Hayashi {
451667b9251SKunihiko Hayashi 	kfree(to_uniphier_xdmac_desc(vd));
452667b9251SKunihiko Hayashi }
453667b9251SKunihiko Hayashi 
uniphier_xdmac_chan_init(struct uniphier_xdmac_device * xdev,int ch)454667b9251SKunihiko Hayashi static void uniphier_xdmac_chan_init(struct uniphier_xdmac_device *xdev,
455667b9251SKunihiko Hayashi 				     int ch)
456667b9251SKunihiko Hayashi {
457667b9251SKunihiko Hayashi 	struct uniphier_xdmac_chan *xc = &xdev->channels[ch];
458667b9251SKunihiko Hayashi 
459667b9251SKunihiko Hayashi 	xc->xdev = xdev;
460667b9251SKunihiko Hayashi 	xc->reg_ch_base = xdev->reg_base + XDMAC_CH_WIDTH * ch;
461667b9251SKunihiko Hayashi 	xc->vc.desc_free = uniphier_xdmac_desc_free;
462667b9251SKunihiko Hayashi 
463667b9251SKunihiko Hayashi 	vchan_init(&xc->vc, &xdev->ddev);
464667b9251SKunihiko Hayashi }
465667b9251SKunihiko Hayashi 
of_dma_uniphier_xlate(struct of_phandle_args * dma_spec,struct of_dma * ofdma)466667b9251SKunihiko Hayashi static struct dma_chan *of_dma_uniphier_xlate(struct of_phandle_args *dma_spec,
467667b9251SKunihiko Hayashi 					      struct of_dma *ofdma)
468667b9251SKunihiko Hayashi {
469667b9251SKunihiko Hayashi 	struct uniphier_xdmac_device *xdev = ofdma->of_dma_data;
470667b9251SKunihiko Hayashi 	int chan_id = dma_spec->args[0];
471667b9251SKunihiko Hayashi 
472667b9251SKunihiko Hayashi 	if (chan_id >= xdev->nr_chans)
473667b9251SKunihiko Hayashi 		return NULL;
474667b9251SKunihiko Hayashi 
475667b9251SKunihiko Hayashi 	xdev->channels[chan_id].id = chan_id;
476667b9251SKunihiko Hayashi 	xdev->channels[chan_id].req_factor = dma_spec->args[1];
477667b9251SKunihiko Hayashi 
478667b9251SKunihiko Hayashi 	return dma_get_slave_channel(&xdev->channels[chan_id].vc.chan);
479667b9251SKunihiko Hayashi }
480667b9251SKunihiko Hayashi 
uniphier_xdmac_probe(struct platform_device * pdev)481667b9251SKunihiko Hayashi static int uniphier_xdmac_probe(struct platform_device *pdev)
482667b9251SKunihiko Hayashi {
483667b9251SKunihiko Hayashi 	struct uniphier_xdmac_device *xdev;
484667b9251SKunihiko Hayashi 	struct device *dev = &pdev->dev;
485667b9251SKunihiko Hayashi 	struct dma_device *ddev;
486667b9251SKunihiko Hayashi 	int irq;
487667b9251SKunihiko Hayashi 	int nr_chans;
488667b9251SKunihiko Hayashi 	int i, ret;
489667b9251SKunihiko Hayashi 
490667b9251SKunihiko Hayashi 	if (of_property_read_u32(dev->of_node, "dma-channels", &nr_chans))
491667b9251SKunihiko Hayashi 		return -EINVAL;
492667b9251SKunihiko Hayashi 	if (nr_chans > XDMAC_MAX_CHANS)
493667b9251SKunihiko Hayashi 		nr_chans = XDMAC_MAX_CHANS;
494667b9251SKunihiko Hayashi 
495667b9251SKunihiko Hayashi 	xdev = devm_kzalloc(dev, struct_size(xdev, channels, nr_chans),
496667b9251SKunihiko Hayashi 			    GFP_KERNEL);
497667b9251SKunihiko Hayashi 	if (!xdev)
498667b9251SKunihiko Hayashi 		return -ENOMEM;
499667b9251SKunihiko Hayashi 
500667b9251SKunihiko Hayashi 	xdev->nr_chans = nr_chans;
501667b9251SKunihiko Hayashi 	xdev->reg_base = devm_platform_ioremap_resource(pdev, 0);
502667b9251SKunihiko Hayashi 	if (IS_ERR(xdev->reg_base))
503667b9251SKunihiko Hayashi 		return PTR_ERR(xdev->reg_base);
504667b9251SKunihiko Hayashi 
505667b9251SKunihiko Hayashi 	ddev = &xdev->ddev;
506667b9251SKunihiko Hayashi 	ddev->dev = dev;
507667b9251SKunihiko Hayashi 	dma_cap_zero(ddev->cap_mask);
508667b9251SKunihiko Hayashi 	dma_cap_set(DMA_MEMCPY, ddev->cap_mask);
509667b9251SKunihiko Hayashi 	dma_cap_set(DMA_SLAVE, ddev->cap_mask);
510667b9251SKunihiko Hayashi 	ddev->src_addr_widths = UNIPHIER_XDMAC_BUSWIDTHS;
511667b9251SKunihiko Hayashi 	ddev->dst_addr_widths = UNIPHIER_XDMAC_BUSWIDTHS;
512667b9251SKunihiko Hayashi 	ddev->directions = BIT(DMA_DEV_TO_MEM) | BIT(DMA_MEM_TO_DEV) |
513667b9251SKunihiko Hayashi 			   BIT(DMA_MEM_TO_MEM);
514667b9251SKunihiko Hayashi 	ddev->residue_granularity = DMA_RESIDUE_GRANULARITY_BURST;
515667b9251SKunihiko Hayashi 	ddev->max_burst = XDMAC_MAX_WORDS;
516667b9251SKunihiko Hayashi 	ddev->device_free_chan_resources = uniphier_xdmac_free_chan_resources;
517667b9251SKunihiko Hayashi 	ddev->device_prep_dma_memcpy = uniphier_xdmac_prep_dma_memcpy;
518667b9251SKunihiko Hayashi 	ddev->device_prep_slave_sg = uniphier_xdmac_prep_slave_sg;
519667b9251SKunihiko Hayashi 	ddev->device_config = uniphier_xdmac_slave_config;
520667b9251SKunihiko Hayashi 	ddev->device_terminate_all = uniphier_xdmac_terminate_all;
521667b9251SKunihiko Hayashi 	ddev->device_synchronize = uniphier_xdmac_synchronize;
522667b9251SKunihiko Hayashi 	ddev->device_tx_status = dma_cookie_status;
523667b9251SKunihiko Hayashi 	ddev->device_issue_pending = uniphier_xdmac_issue_pending;
524667b9251SKunihiko Hayashi 	INIT_LIST_HEAD(&ddev->channels);
525667b9251SKunihiko Hayashi 
526667b9251SKunihiko Hayashi 	for (i = 0; i < nr_chans; i++)
527667b9251SKunihiko Hayashi 		uniphier_xdmac_chan_init(xdev, i);
528667b9251SKunihiko Hayashi 
529667b9251SKunihiko Hayashi 	irq = platform_get_irq(pdev, 0);
5300950c7fdSVinod Koul 	if (irq < 0)
531667b9251SKunihiko Hayashi 		return irq;
532667b9251SKunihiko Hayashi 
533667b9251SKunihiko Hayashi 	ret = devm_request_irq(dev, irq, uniphier_xdmac_irq_handler,
534667b9251SKunihiko Hayashi 			       IRQF_SHARED, "xdmac", xdev);
535667b9251SKunihiko Hayashi 	if (ret) {
536667b9251SKunihiko Hayashi 		dev_err(dev, "Failed to request IRQ\n");
537667b9251SKunihiko Hayashi 		return ret;
538667b9251SKunihiko Hayashi 	}
539667b9251SKunihiko Hayashi 
540667b9251SKunihiko Hayashi 	ret = dma_async_device_register(ddev);
541667b9251SKunihiko Hayashi 	if (ret) {
542667b9251SKunihiko Hayashi 		dev_err(dev, "Failed to register XDMA device\n");
543667b9251SKunihiko Hayashi 		return ret;
544667b9251SKunihiko Hayashi 	}
545667b9251SKunihiko Hayashi 
546667b9251SKunihiko Hayashi 	ret = of_dma_controller_register(dev->of_node,
547667b9251SKunihiko Hayashi 					 of_dma_uniphier_xlate, xdev);
548667b9251SKunihiko Hayashi 	if (ret) {
549667b9251SKunihiko Hayashi 		dev_err(dev, "Failed to register XDMA controller\n");
550667b9251SKunihiko Hayashi 		goto out_unregister_dmac;
551667b9251SKunihiko Hayashi 	}
552667b9251SKunihiko Hayashi 
553667b9251SKunihiko Hayashi 	platform_set_drvdata(pdev, xdev);
554667b9251SKunihiko Hayashi 
555667b9251SKunihiko Hayashi 	dev_info(&pdev->dev, "UniPhier XDMAC driver (%d channels)\n",
556667b9251SKunihiko Hayashi 		 nr_chans);
557667b9251SKunihiko Hayashi 
558667b9251SKunihiko Hayashi 	return 0;
559667b9251SKunihiko Hayashi 
560667b9251SKunihiko Hayashi out_unregister_dmac:
561667b9251SKunihiko Hayashi 	dma_async_device_unregister(ddev);
562667b9251SKunihiko Hayashi 
563667b9251SKunihiko Hayashi 	return ret;
564667b9251SKunihiko Hayashi }
565667b9251SKunihiko Hayashi 
uniphier_xdmac_remove(struct platform_device * pdev)566ead0e402SUwe Kleine-König static void uniphier_xdmac_remove(struct platform_device *pdev)
567667b9251SKunihiko Hayashi {
568667b9251SKunihiko Hayashi 	struct uniphier_xdmac_device *xdev = platform_get_drvdata(pdev);
569667b9251SKunihiko Hayashi 	struct dma_device *ddev = &xdev->ddev;
570667b9251SKunihiko Hayashi 	struct dma_chan *chan;
571667b9251SKunihiko Hayashi 	int ret;
572667b9251SKunihiko Hayashi 
573667b9251SKunihiko Hayashi 	/*
574667b9251SKunihiko Hayashi 	 * Before reaching here, almost all descriptors have been freed by the
575667b9251SKunihiko Hayashi 	 * ->device_free_chan_resources() hook. However, each channel might
576667b9251SKunihiko Hayashi 	 * be still holding one descriptor that was on-flight at that moment.
577667b9251SKunihiko Hayashi 	 * Terminate it to make sure this hardware is no longer running. Then,
578667b9251SKunihiko Hayashi 	 * free the channel resources once again to avoid memory leak.
579667b9251SKunihiko Hayashi 	 */
580667b9251SKunihiko Hayashi 	list_for_each_entry(chan, &ddev->channels, device_node) {
581667b9251SKunihiko Hayashi 		ret = dmaengine_terminate_sync(chan);
582ead0e402SUwe Kleine-König 		if (ret) {
583ead0e402SUwe Kleine-König 			/*
584ead0e402SUwe Kleine-König 			 * This results in resource leakage and maybe also
585ead0e402SUwe Kleine-König 			 * use-after-free errors as e.g. *xdev is kfreed.
586ead0e402SUwe Kleine-König 			 */
587ead0e402SUwe Kleine-König 			dev_alert(&pdev->dev, "Failed to terminate channel %d (%pe)\n",
588ead0e402SUwe Kleine-König 				  chan->chan_id, ERR_PTR(ret));
589ead0e402SUwe Kleine-König 			return;
590ead0e402SUwe Kleine-König 		}
591667b9251SKunihiko Hayashi 		uniphier_xdmac_free_chan_resources(chan);
592667b9251SKunihiko Hayashi 	}
593667b9251SKunihiko Hayashi 
594667b9251SKunihiko Hayashi 	of_dma_controller_free(pdev->dev.of_node);
595667b9251SKunihiko Hayashi 	dma_async_device_unregister(ddev);
596667b9251SKunihiko Hayashi }
597667b9251SKunihiko Hayashi 
598667b9251SKunihiko Hayashi static const struct of_device_id uniphier_xdmac_match[] = {
599667b9251SKunihiko Hayashi 	{ .compatible = "socionext,uniphier-xdmac" },
600667b9251SKunihiko Hayashi 	{ /* sentinel */ }
601667b9251SKunihiko Hayashi };
602667b9251SKunihiko Hayashi MODULE_DEVICE_TABLE(of, uniphier_xdmac_match);
603667b9251SKunihiko Hayashi 
604667b9251SKunihiko Hayashi static struct platform_driver uniphier_xdmac_driver = {
605667b9251SKunihiko Hayashi 	.probe = uniphier_xdmac_probe,
606*76355c25SUwe Kleine-König 	.remove = uniphier_xdmac_remove,
607667b9251SKunihiko Hayashi 	.driver = {
608667b9251SKunihiko Hayashi 		.name = "uniphier-xdmac",
609667b9251SKunihiko Hayashi 		.of_match_table = uniphier_xdmac_match,
610667b9251SKunihiko Hayashi 	},
611667b9251SKunihiko Hayashi };
612667b9251SKunihiko Hayashi module_platform_driver(uniphier_xdmac_driver);
613667b9251SKunihiko Hayashi 
614667b9251SKunihiko Hayashi MODULE_AUTHOR("Kunihiko Hayashi <hayashi.kunihiko@socionext.com>");
615667b9251SKunihiko Hayashi MODULE_DESCRIPTION("UniPhier external DMA controller driver");
616667b9251SKunihiko Hayashi MODULE_LICENSE("GPL v2");
617