Linux Audio

Check our new training course

Loading...
Note: File does not exist in v3.1.
  1// SPDX-License-Identifier: GPL-2.0-or-later
  2/*
  3 * DMA driver for STMicroelectronics STi FDMA controller
  4 *
  5 * Copyright (C) 2014 STMicroelectronics
  6 *
  7 * Author: Ludovic Barre <Ludovic.barre@st.com>
  8 *	   Peter Griffin <peter.griffin@linaro.org>
  9 */
 10
 11#include <linux/init.h>
 12#include <linux/module.h>
 13#include <linux/of_device.h>
 14#include <linux/of_dma.h>
 15#include <linux/platform_device.h>
 16#include <linux/interrupt.h>
 17#include <linux/remoteproc.h>
 18
 19#include "st_fdma.h"
 20
 21static inline struct st_fdma_chan *to_st_fdma_chan(struct dma_chan *c)
 22{
 23	return container_of(c, struct st_fdma_chan, vchan.chan);
 24}
 25
 26static struct st_fdma_desc *to_st_fdma_desc(struct virt_dma_desc *vd)
 27{
 28	return container_of(vd, struct st_fdma_desc, vdesc);
 29}
 30
 31static int st_fdma_dreq_get(struct st_fdma_chan *fchan)
 32{
 33	struct st_fdma_dev *fdev = fchan->fdev;
 34	u32 req_line_cfg = fchan->cfg.req_line;
 35	u32 dreq_line;
 36	int try = 0;
 37
 38	/*
 39	 * dreq_mask is shared for n channels of fdma, so all accesses must be
 40	 * atomic. if the dreq_mask is changed between ffz and set_bit,
 41	 * we retry
 42	 */
 43	do {
 44		if (fdev->dreq_mask == ~0L) {
 45			dev_err(fdev->dev, "No req lines available\n");
 46			return -EINVAL;
 47		}
 48
 49		if (try || req_line_cfg >= ST_FDMA_NR_DREQS) {
 50			dev_err(fdev->dev, "Invalid or used req line\n");
 51			return -EINVAL;
 52		} else {
 53			dreq_line = req_line_cfg;
 54		}
 55
 56		try++;
 57	} while (test_and_set_bit(dreq_line, &fdev->dreq_mask));
 58
 59	dev_dbg(fdev->dev, "get dreq_line:%d mask:%#lx\n",
 60		dreq_line, fdev->dreq_mask);
 61
 62	return dreq_line;
 63}
 64
 65static void st_fdma_dreq_put(struct st_fdma_chan *fchan)
 66{
 67	struct st_fdma_dev *fdev = fchan->fdev;
 68
 69	dev_dbg(fdev->dev, "put dreq_line:%#x\n", fchan->dreq_line);
 70	clear_bit(fchan->dreq_line, &fdev->dreq_mask);
 71}
 72
 73static void st_fdma_xfer_desc(struct st_fdma_chan *fchan)
 74{
 75	struct virt_dma_desc *vdesc;
 76	unsigned long nbytes, ch_cmd, cmd;
 77
 78	vdesc = vchan_next_desc(&fchan->vchan);
 79	if (!vdesc)
 80		return;
 81
 82	fchan->fdesc = to_st_fdma_desc(vdesc);
 83	nbytes = fchan->fdesc->node[0].desc->nbytes;
 84	cmd = FDMA_CMD_START(fchan->vchan.chan.chan_id);
 85	ch_cmd = fchan->fdesc->node[0].pdesc | FDMA_CH_CMD_STA_START;
 86
 87	/* start the channel for the descriptor */
 88	fnode_write(fchan, nbytes, FDMA_CNTN_OFST);
 89	fchan_write(fchan, ch_cmd, FDMA_CH_CMD_OFST);
 90	writel(cmd,
 91		fchan->fdev->slim_rproc->peri + FDMA_CMD_SET_OFST);
 92
 93	dev_dbg(fchan->fdev->dev, "start chan:%d\n", fchan->vchan.chan.chan_id);
 94}
 95
 96static void st_fdma_ch_sta_update(struct st_fdma_chan *fchan,
 97				  unsigned long int_sta)
 98{
 99	unsigned long ch_sta, ch_err;
100	int ch_id = fchan->vchan.chan.chan_id;
101	struct st_fdma_dev *fdev = fchan->fdev;
102
103	ch_sta = fchan_read(fchan, FDMA_CH_CMD_OFST);
104	ch_err = ch_sta & FDMA_CH_CMD_ERR_MASK;
105	ch_sta &= FDMA_CH_CMD_STA_MASK;
106
107	if (int_sta & FDMA_INT_STA_ERR) {
108		dev_warn(fdev->dev, "chan:%d, error:%ld\n", ch_id, ch_err);
109		fchan->status = DMA_ERROR;
110		return;
111	}
112
113	switch (ch_sta) {
114	case FDMA_CH_CMD_STA_PAUSED:
115		fchan->status = DMA_PAUSED;
116		break;
117
118	case FDMA_CH_CMD_STA_RUNNING:
119		fchan->status = DMA_IN_PROGRESS;
120		break;
121	}
122}
123
124static irqreturn_t st_fdma_irq_handler(int irq, void *dev_id)
125{
126	struct st_fdma_dev *fdev = dev_id;
127	irqreturn_t ret = IRQ_NONE;
128	struct st_fdma_chan *fchan = &fdev->chans[0];
129	unsigned long int_sta, clr;
130
131	int_sta = fdma_read(fdev, FDMA_INT_STA_OFST);
132	clr = int_sta;
133
134	for (; int_sta != 0 ; int_sta >>= 2, fchan++) {
135		if (!(int_sta & (FDMA_INT_STA_CH | FDMA_INT_STA_ERR)))
136			continue;
137
138		spin_lock(&fchan->vchan.lock);
139		st_fdma_ch_sta_update(fchan, int_sta);
140
141		if (fchan->fdesc) {
142			if (!fchan->fdesc->iscyclic) {
143				list_del(&fchan->fdesc->vdesc.node);
144				vchan_cookie_complete(&fchan->fdesc->vdesc);
145				fchan->fdesc = NULL;
146				fchan->status = DMA_COMPLETE;
147			} else {
148				vchan_cyclic_callback(&fchan->fdesc->vdesc);
149			}
150
151			/* Start the next descriptor (if available) */
152			if (!fchan->fdesc)
153				st_fdma_xfer_desc(fchan);
154		}
155
156		spin_unlock(&fchan->vchan.lock);
157		ret = IRQ_HANDLED;
158	}
159
160	fdma_write(fdev, clr, FDMA_INT_CLR_OFST);
161
162	return ret;
163}
164
165static struct dma_chan *st_fdma_of_xlate(struct of_phandle_args *dma_spec,
166					 struct of_dma *ofdma)
167{
168	struct st_fdma_dev *fdev = ofdma->of_dma_data;
169	struct dma_chan *chan;
170	struct st_fdma_chan *fchan;
171	int ret;
172
173	if (dma_spec->args_count < 1)
174		return ERR_PTR(-EINVAL);
175
176	if (fdev->dma_device.dev->of_node != dma_spec->np)
177		return ERR_PTR(-EINVAL);
178
179	ret = rproc_boot(fdev->slim_rproc->rproc);
180	if (ret == -ENOENT)
181		return ERR_PTR(-EPROBE_DEFER);
182	else if (ret)
183		return ERR_PTR(ret);
184
185	chan = dma_get_any_slave_channel(&fdev->dma_device);
186	if (!chan)
187		goto err_chan;
188
189	fchan = to_st_fdma_chan(chan);
190
191	fchan->cfg.of_node = dma_spec->np;
192	fchan->cfg.req_line = dma_spec->args[0];
193	fchan->cfg.req_ctrl = 0;
194	fchan->cfg.type = ST_FDMA_TYPE_FREE_RUN;
195
196	if (dma_spec->args_count > 1)
197		fchan->cfg.req_ctrl = dma_spec->args[1]
198			& FDMA_REQ_CTRL_CFG_MASK;
199
200	if (dma_spec->args_count > 2)
201		fchan->cfg.type = dma_spec->args[2];
202
203	if (fchan->cfg.type == ST_FDMA_TYPE_FREE_RUN) {
204		fchan->dreq_line = 0;
205	} else {
206		fchan->dreq_line = st_fdma_dreq_get(fchan);
207		if (IS_ERR_VALUE(fchan->dreq_line)) {
208			chan = ERR_PTR(fchan->dreq_line);
209			goto err_chan;
210		}
211	}
212
213	dev_dbg(fdev->dev, "xlate req_line:%d type:%d req_ctrl:%#lx\n",
214		fchan->cfg.req_line, fchan->cfg.type, fchan->cfg.req_ctrl);
215
216	return chan;
217
218err_chan:
219	rproc_shutdown(fdev->slim_rproc->rproc);
220	return chan;
221
222}
223
224static void st_fdma_free_desc(struct virt_dma_desc *vdesc)
225{
226	struct st_fdma_desc *fdesc;
227	int i;
228
229	fdesc = to_st_fdma_desc(vdesc);
230	for (i = 0; i < fdesc->n_nodes; i++)
231		dma_pool_free(fdesc->fchan->node_pool, fdesc->node[i].desc,
232			      fdesc->node[i].pdesc);
233	kfree(fdesc);
234}
235
236static struct st_fdma_desc *st_fdma_alloc_desc(struct st_fdma_chan *fchan,
237					       int sg_len)
238{
239	struct st_fdma_desc *fdesc;
240	int i;
241
242	fdesc = kzalloc(struct_size(fdesc, node, sg_len), GFP_NOWAIT);
243	if (!fdesc)
244		return NULL;
245
246	fdesc->fchan = fchan;
247	fdesc->n_nodes = sg_len;
248	for (i = 0; i < sg_len; i++) {
249		fdesc->node[i].desc = dma_pool_alloc(fchan->node_pool,
250				GFP_NOWAIT, &fdesc->node[i].pdesc);
251		if (!fdesc->node[i].desc)
252			goto err;
253	}
254	return fdesc;
255
256err:
257	while (--i >= 0)
258		dma_pool_free(fchan->node_pool, fdesc->node[i].desc,
259			      fdesc->node[i].pdesc);
260	kfree(fdesc);
261	return NULL;
262}
263
264static int st_fdma_alloc_chan_res(struct dma_chan *chan)
265{
266	struct st_fdma_chan *fchan = to_st_fdma_chan(chan);
267
268	/* Create the dma pool for descriptor allocation */
269	fchan->node_pool = dma_pool_create(dev_name(&chan->dev->device),
270					    fchan->fdev->dev,
271					    sizeof(struct st_fdma_hw_node),
272					    __alignof__(struct st_fdma_hw_node),
273					    0);
274
275	if (!fchan->node_pool) {
276		dev_err(fchan->fdev->dev, "unable to allocate desc pool\n");
277		return -ENOMEM;
278	}
279
280	dev_dbg(fchan->fdev->dev, "alloc ch_id:%d type:%d\n",
281		fchan->vchan.chan.chan_id, fchan->cfg.type);
282
283	return 0;
284}
285
286static void st_fdma_free_chan_res(struct dma_chan *chan)
287{
288	struct st_fdma_chan *fchan = to_st_fdma_chan(chan);
289	struct rproc *rproc = fchan->fdev->slim_rproc->rproc;
290	unsigned long flags;
291
292	dev_dbg(fchan->fdev->dev, "%s: freeing chan:%d\n",
293		__func__, fchan->vchan.chan.chan_id);
294
295	if (fchan->cfg.type != ST_FDMA_TYPE_FREE_RUN)
296		st_fdma_dreq_put(fchan);
297
298	spin_lock_irqsave(&fchan->vchan.lock, flags);
299	fchan->fdesc = NULL;
300	spin_unlock_irqrestore(&fchan->vchan.lock, flags);
301
302	dma_pool_destroy(fchan->node_pool);
303	fchan->node_pool = NULL;
304	memset(&fchan->cfg, 0, sizeof(struct st_fdma_cfg));
305
306	rproc_shutdown(rproc);
307}
308
309static struct dma_async_tx_descriptor *st_fdma_prep_dma_memcpy(
310	struct dma_chan *chan,	dma_addr_t dst, dma_addr_t src,
311	size_t len, unsigned long flags)
312{
313	struct st_fdma_chan *fchan;
314	struct st_fdma_desc *fdesc;
315	struct st_fdma_hw_node *hw_node;
316
317	if (!len)
318		return NULL;
319
320	fchan = to_st_fdma_chan(chan);
321
322	/* We only require a single descriptor */
323	fdesc = st_fdma_alloc_desc(fchan, 1);
324	if (!fdesc) {
325		dev_err(fchan->fdev->dev, "no memory for desc\n");
326		return NULL;
327	}
328
329	hw_node = fdesc->node[0].desc;
330	hw_node->next = 0;
331	hw_node->control = FDMA_NODE_CTRL_REQ_MAP_FREE_RUN;
332	hw_node->control |= FDMA_NODE_CTRL_SRC_INCR;
333	hw_node->control |= FDMA_NODE_CTRL_DST_INCR;
334	hw_node->control |= FDMA_NODE_CTRL_INT_EON;
335	hw_node->nbytes = len;
336	hw_node->saddr = src;
337	hw_node->daddr = dst;
338	hw_node->generic.length = len;
339	hw_node->generic.sstride = 0;
340	hw_node->generic.dstride = 0;
341
342	return vchan_tx_prep(&fchan->vchan, &fdesc->vdesc, flags);
343}
344
345static int config_reqctrl(struct st_fdma_chan *fchan,
346			  enum dma_transfer_direction direction)
347{
348	u32 maxburst = 0, addr = 0;
349	enum dma_slave_buswidth width;
350	int ch_id = fchan->vchan.chan.chan_id;
351	struct st_fdma_dev *fdev = fchan->fdev;
352
353	switch (direction) {
354
355	case DMA_DEV_TO_MEM:
356		fchan->cfg.req_ctrl &= ~FDMA_REQ_CTRL_WNR;
357		maxburst = fchan->scfg.src_maxburst;
358		width = fchan->scfg.src_addr_width;
359		addr = fchan->scfg.src_addr;
360		break;
361
362	case DMA_MEM_TO_DEV:
363		fchan->cfg.req_ctrl |= FDMA_REQ_CTRL_WNR;
364		maxburst = fchan->scfg.dst_maxburst;
365		width = fchan->scfg.dst_addr_width;
366		addr = fchan->scfg.dst_addr;
367		break;
368
369	default:
370		return -EINVAL;
371	}
372
373	fchan->cfg.req_ctrl &= ~FDMA_REQ_CTRL_OPCODE_MASK;
374
375	switch (width) {
376
377	case DMA_SLAVE_BUSWIDTH_1_BYTE:
378		fchan->cfg.req_ctrl |= FDMA_REQ_CTRL_OPCODE_LD_ST1;
379		break;
380
381	case DMA_SLAVE_BUSWIDTH_2_BYTES:
382		fchan->cfg.req_ctrl |= FDMA_REQ_CTRL_OPCODE_LD_ST2;
383		break;
384
385	case DMA_SLAVE_BUSWIDTH_4_BYTES:
386		fchan->cfg.req_ctrl |= FDMA_REQ_CTRL_OPCODE_LD_ST4;
387		break;
388
389	case DMA_SLAVE_BUSWIDTH_8_BYTES:
390		fchan->cfg.req_ctrl |= FDMA_REQ_CTRL_OPCODE_LD_ST8;
391		break;
392
393	default:
394		return -EINVAL;
395	}
396
397	fchan->cfg.req_ctrl &= ~FDMA_REQ_CTRL_NUM_OPS_MASK;
398	fchan->cfg.req_ctrl |= FDMA_REQ_CTRL_NUM_OPS(maxburst-1);
399	dreq_write(fchan, fchan->cfg.req_ctrl, FDMA_REQ_CTRL_OFST);
400
401	fchan->cfg.dev_addr = addr;
402	fchan->cfg.dir = direction;
403
404	dev_dbg(fdev->dev, "chan:%d config_reqctrl:%#x req_ctrl:%#lx\n",
405		ch_id, addr, fchan->cfg.req_ctrl);
406
407	return 0;
408}
409
410static void fill_hw_node(struct st_fdma_hw_node *hw_node,
411			struct st_fdma_chan *fchan,
412			enum dma_transfer_direction direction)
413{
414	if (direction == DMA_MEM_TO_DEV) {
415		hw_node->control |= FDMA_NODE_CTRL_SRC_INCR;
416		hw_node->control |= FDMA_NODE_CTRL_DST_STATIC;
417		hw_node->daddr = fchan->cfg.dev_addr;
418	} else {
419		hw_node->control |= FDMA_NODE_CTRL_SRC_STATIC;
420		hw_node->control |= FDMA_NODE_CTRL_DST_INCR;
421		hw_node->saddr = fchan->cfg.dev_addr;
422	}
423
424	hw_node->generic.sstride = 0;
425	hw_node->generic.dstride = 0;
426}
427
428static inline struct st_fdma_chan *st_fdma_prep_common(struct dma_chan *chan,
429		size_t len, enum dma_transfer_direction direction)
430{
431	struct st_fdma_chan *fchan;
432
433	if (!chan || !len)
434		return NULL;
435
436	fchan = to_st_fdma_chan(chan);
437
438	if (!is_slave_direction(direction)) {
439		dev_err(fchan->fdev->dev, "bad direction?\n");
440		return NULL;
441	}
442
443	return fchan;
444}
445
446static struct dma_async_tx_descriptor *st_fdma_prep_dma_cyclic(
447		struct dma_chan *chan, dma_addr_t buf_addr, size_t len,
448		size_t period_len, enum dma_transfer_direction direction,
449		unsigned long flags)
450{
451	struct st_fdma_chan *fchan;
452	struct st_fdma_desc *fdesc;
453	int sg_len, i;
454
455	fchan = st_fdma_prep_common(chan, len, direction);
456	if (!fchan)
457		return NULL;
458
459	if (!period_len)
460		return NULL;
461
462	if (config_reqctrl(fchan, direction)) {
463		dev_err(fchan->fdev->dev, "bad width or direction\n");
464		return NULL;
465	}
466
467	/* the buffer length must be a multiple of period_len */
468	if (len % period_len != 0) {
469		dev_err(fchan->fdev->dev, "len is not multiple of period\n");
470		return NULL;
471	}
472
473	sg_len = len / period_len;
474	fdesc = st_fdma_alloc_desc(fchan, sg_len);
475	if (!fdesc) {
476		dev_err(fchan->fdev->dev, "no memory for desc\n");
477		return NULL;
478	}
479
480	fdesc->iscyclic = true;
481
482	for (i = 0; i < sg_len; i++) {
483		struct st_fdma_hw_node *hw_node = fdesc->node[i].desc;
484
485		hw_node->next = fdesc->node[(i + 1) % sg_len].pdesc;
486
487		hw_node->control =
488			FDMA_NODE_CTRL_REQ_MAP_DREQ(fchan->dreq_line);
489		hw_node->control |= FDMA_NODE_CTRL_INT_EON;
490
491		fill_hw_node(hw_node, fchan, direction);
492
493		if (direction == DMA_MEM_TO_DEV)
494			hw_node->saddr = buf_addr + (i * period_len);
495		else
496			hw_node->daddr = buf_addr + (i * period_len);
497
498		hw_node->nbytes = period_len;
499		hw_node->generic.length = period_len;
500	}
501
502	return vchan_tx_prep(&fchan->vchan, &fdesc->vdesc, flags);
503}
504
505static struct dma_async_tx_descriptor *st_fdma_prep_slave_sg(
506		struct dma_chan *chan, struct scatterlist *sgl,
507		unsigned int sg_len, enum dma_transfer_direction direction,
508		unsigned long flags, void *context)
509{
510	struct st_fdma_chan *fchan;
511	struct st_fdma_desc *fdesc;
512	struct st_fdma_hw_node *hw_node;
513	struct scatterlist *sg;
514	int i;
515
516	fchan = st_fdma_prep_common(chan, sg_len, direction);
517	if (!fchan)
518		return NULL;
519
520	if (!sgl)
521		return NULL;
522
523	fdesc = st_fdma_alloc_desc(fchan, sg_len);
524	if (!fdesc) {
525		dev_err(fchan->fdev->dev, "no memory for desc\n");
526		return NULL;
527	}
528
529	fdesc->iscyclic = false;
530
531	for_each_sg(sgl, sg, sg_len, i) {
532		hw_node = fdesc->node[i].desc;
533
534		hw_node->next = fdesc->node[(i + 1) % sg_len].pdesc;
535		hw_node->control = FDMA_NODE_CTRL_REQ_MAP_DREQ(fchan->dreq_line);
536
537		fill_hw_node(hw_node, fchan, direction);
538
539		if (direction == DMA_MEM_TO_DEV)
540			hw_node->saddr = sg_dma_address(sg);
541		else
542			hw_node->daddr = sg_dma_address(sg);
543
544		hw_node->nbytes = sg_dma_len(sg);
545		hw_node->generic.length = sg_dma_len(sg);
546	}
547
548	/* interrupt at end of last node */
549	hw_node->control |= FDMA_NODE_CTRL_INT_EON;
550
551	return vchan_tx_prep(&fchan->vchan, &fdesc->vdesc, flags);
552}
553
554static size_t st_fdma_desc_residue(struct st_fdma_chan *fchan,
555				   struct virt_dma_desc *vdesc,
556				   bool in_progress)
557{
558	struct st_fdma_desc *fdesc = fchan->fdesc;
559	size_t residue = 0;
560	dma_addr_t cur_addr = 0;
561	int i;
562
563	if (in_progress) {
564		cur_addr = fchan_read(fchan, FDMA_CH_CMD_OFST);
565		cur_addr &= FDMA_CH_CMD_DATA_MASK;
566	}
567
568	for (i = fchan->fdesc->n_nodes - 1 ; i >= 0; i--) {
569		if (cur_addr == fdesc->node[i].pdesc) {
570			residue += fnode_read(fchan, FDMA_CNTN_OFST);
571			break;
572		}
573		residue += fdesc->node[i].desc->nbytes;
574	}
575
576	return residue;
577}
578
579static enum dma_status st_fdma_tx_status(struct dma_chan *chan,
580					 dma_cookie_t cookie,
581					 struct dma_tx_state *txstate)
582{
583	struct st_fdma_chan *fchan = to_st_fdma_chan(chan);
584	struct virt_dma_desc *vd;
585	enum dma_status ret;
586	unsigned long flags;
587
588	ret = dma_cookie_status(chan, cookie, txstate);
589	if (ret == DMA_COMPLETE || !txstate)
590		return ret;
591
592	spin_lock_irqsave(&fchan->vchan.lock, flags);
593	vd = vchan_find_desc(&fchan->vchan, cookie);
594	if (fchan->fdesc && cookie == fchan->fdesc->vdesc.tx.cookie)
595		txstate->residue = st_fdma_desc_residue(fchan, vd, true);
596	else if (vd)
597		txstate->residue = st_fdma_desc_residue(fchan, vd, false);
598	else
599		txstate->residue = 0;
600
601	spin_unlock_irqrestore(&fchan->vchan.lock, flags);
602
603	return ret;
604}
605
606static void st_fdma_issue_pending(struct dma_chan *chan)
607{
608	struct st_fdma_chan *fchan = to_st_fdma_chan(chan);
609	unsigned long flags;
610
611	spin_lock_irqsave(&fchan->vchan.lock, flags);
612
613	if (vchan_issue_pending(&fchan->vchan) && !fchan->fdesc)
614		st_fdma_xfer_desc(fchan);
615
616	spin_unlock_irqrestore(&fchan->vchan.lock, flags);
617}
618
619static int st_fdma_pause(struct dma_chan *chan)
620{
621	unsigned long flags;
622	struct st_fdma_chan *fchan = to_st_fdma_chan(chan);
623	int ch_id = fchan->vchan.chan.chan_id;
624	unsigned long cmd = FDMA_CMD_PAUSE(ch_id);
625
626	dev_dbg(fchan->fdev->dev, "pause chan:%d\n", ch_id);
627
628	spin_lock_irqsave(&fchan->vchan.lock, flags);
629	if (fchan->fdesc)
630		fdma_write(fchan->fdev, cmd, FDMA_CMD_SET_OFST);
631	spin_unlock_irqrestore(&fchan->vchan.lock, flags);
632
633	return 0;
634}
635
636static int st_fdma_resume(struct dma_chan *chan)
637{
638	unsigned long flags;
639	unsigned long val;
640	struct st_fdma_chan *fchan = to_st_fdma_chan(chan);
641	int ch_id = fchan->vchan.chan.chan_id;
642
643	dev_dbg(fchan->fdev->dev, "resume chan:%d\n", ch_id);
644
645	spin_lock_irqsave(&fchan->vchan.lock, flags);
646	if (fchan->fdesc) {
647		val = fchan_read(fchan, FDMA_CH_CMD_OFST);
648		val &= FDMA_CH_CMD_DATA_MASK;
649		fchan_write(fchan, val, FDMA_CH_CMD_OFST);
650	}
651	spin_unlock_irqrestore(&fchan->vchan.lock, flags);
652
653	return 0;
654}
655
656static int st_fdma_terminate_all(struct dma_chan *chan)
657{
658	unsigned long flags;
659	LIST_HEAD(head);
660	struct st_fdma_chan *fchan = to_st_fdma_chan(chan);
661	int ch_id = fchan->vchan.chan.chan_id;
662	unsigned long cmd = FDMA_CMD_PAUSE(ch_id);
663
664	dev_dbg(fchan->fdev->dev, "terminate chan:%d\n", ch_id);
665
666	spin_lock_irqsave(&fchan->vchan.lock, flags);
667	fdma_write(fchan->fdev, cmd, FDMA_CMD_SET_OFST);
668	fchan->fdesc = NULL;
669	vchan_get_all_descriptors(&fchan->vchan, &head);
670	spin_unlock_irqrestore(&fchan->vchan.lock, flags);
671	vchan_dma_desc_free_list(&fchan->vchan, &head);
672
673	return 0;
674}
675
676static int st_fdma_slave_config(struct dma_chan *chan,
677				struct dma_slave_config *slave_cfg)
678{
679	struct st_fdma_chan *fchan = to_st_fdma_chan(chan);
680
681	memcpy(&fchan->scfg, slave_cfg, sizeof(fchan->scfg));
682	return 0;
683}
684
685static const struct st_fdma_driverdata fdma_mpe31_stih407_11 = {
686	.name = "STiH407",
687	.id = 0,
688};
689
690static const struct st_fdma_driverdata fdma_mpe31_stih407_12 = {
691	.name = "STiH407",
692	.id = 1,
693};
694
695static const struct st_fdma_driverdata fdma_mpe31_stih407_13 = {
696	.name = "STiH407",
697	.id = 2,
698};
699
700static const struct of_device_id st_fdma_match[] = {
701	{ .compatible = "st,stih407-fdma-mpe31-11"
702	  , .data = &fdma_mpe31_stih407_11 },
703	{ .compatible = "st,stih407-fdma-mpe31-12"
704	  , .data = &fdma_mpe31_stih407_12 },
705	{ .compatible = "st,stih407-fdma-mpe31-13"
706	  , .data = &fdma_mpe31_stih407_13 },
707	{},
708};
709MODULE_DEVICE_TABLE(of, st_fdma_match);
710
711static int st_fdma_parse_dt(struct platform_device *pdev,
712			const struct st_fdma_driverdata *drvdata,
713			struct st_fdma_dev *fdev)
714{
715	snprintf(fdev->fw_name, FW_NAME_SIZE, "fdma_%s_%d.elf",
716		drvdata->name, drvdata->id);
717
718	return of_property_read_u32(pdev->dev.of_node, "dma-channels",
719				    &fdev->nr_channels);
720}
721#define FDMA_DMA_BUSWIDTHS	(BIT(DMA_SLAVE_BUSWIDTH_1_BYTE) | \
722				 BIT(DMA_SLAVE_BUSWIDTH_2_BYTES) | \
723				 BIT(DMA_SLAVE_BUSWIDTH_3_BYTES) | \
724				 BIT(DMA_SLAVE_BUSWIDTH_4_BYTES))
725
726static void st_fdma_free(struct st_fdma_dev *fdev)
727{
728	struct st_fdma_chan *fchan;
729	int i;
730
731	for (i = 0; i < fdev->nr_channels; i++) {
732		fchan = &fdev->chans[i];
733		list_del(&fchan->vchan.chan.device_node);
734		tasklet_kill(&fchan->vchan.task);
735	}
736}
737
738static int st_fdma_probe(struct platform_device *pdev)
739{
740	struct st_fdma_dev *fdev;
741	const struct of_device_id *match;
742	struct device_node *np = pdev->dev.of_node;
743	const struct st_fdma_driverdata *drvdata;
744	int ret, i;
745
746	match = of_match_device((st_fdma_match), &pdev->dev);
747	if (!match || !match->data) {
748		dev_err(&pdev->dev, "No device match found\n");
749		return -ENODEV;
750	}
751
752	drvdata = match->data;
753
754	fdev = devm_kzalloc(&pdev->dev, sizeof(*fdev), GFP_KERNEL);
755	if (!fdev)
756		return -ENOMEM;
757
758	ret = st_fdma_parse_dt(pdev, drvdata, fdev);
759	if (ret) {
760		dev_err(&pdev->dev, "unable to find platform data\n");
761		goto err;
762	}
763
764	fdev->chans = devm_kcalloc(&pdev->dev, fdev->nr_channels,
765				   sizeof(struct st_fdma_chan), GFP_KERNEL);
766	if (!fdev->chans)
767		return -ENOMEM;
768
769	fdev->dev = &pdev->dev;
770	fdev->drvdata = drvdata;
771	platform_set_drvdata(pdev, fdev);
772
773	fdev->irq = platform_get_irq(pdev, 0);
774	if (fdev->irq < 0)
775		return -EINVAL;
776
777	ret = devm_request_irq(&pdev->dev, fdev->irq, st_fdma_irq_handler, 0,
778			       dev_name(&pdev->dev), fdev);
779	if (ret) {
780		dev_err(&pdev->dev, "Failed to request irq (%d)\n", ret);
781		goto err;
782	}
783
784	fdev->slim_rproc = st_slim_rproc_alloc(pdev, fdev->fw_name);
785	if (IS_ERR(fdev->slim_rproc)) {
786		ret = PTR_ERR(fdev->slim_rproc);
787		dev_err(&pdev->dev, "slim_rproc_alloc failed (%d)\n", ret);
788		goto err;
789	}
790
791	/* Initialise list of FDMA channels */
792	INIT_LIST_HEAD(&fdev->dma_device.channels);
793	for (i = 0; i < fdev->nr_channels; i++) {
794		struct st_fdma_chan *fchan = &fdev->chans[i];
795
796		fchan->fdev = fdev;
797		fchan->vchan.desc_free = st_fdma_free_desc;
798		vchan_init(&fchan->vchan, &fdev->dma_device);
799	}
800
801	/* Initialise the FDMA dreq (reserve 0 & 31 for FDMA use) */
802	fdev->dreq_mask = BIT(0) | BIT(31);
803
804	dma_cap_set(DMA_SLAVE, fdev->dma_device.cap_mask);
805	dma_cap_set(DMA_CYCLIC, fdev->dma_device.cap_mask);
806	dma_cap_set(DMA_MEMCPY, fdev->dma_device.cap_mask);
807
808	fdev->dma_device.dev = &pdev->dev;
809	fdev->dma_device.device_alloc_chan_resources = st_fdma_alloc_chan_res;
810	fdev->dma_device.device_free_chan_resources = st_fdma_free_chan_res;
811	fdev->dma_device.device_prep_dma_cyclic	= st_fdma_prep_dma_cyclic;
812	fdev->dma_device.device_prep_slave_sg = st_fdma_prep_slave_sg;
813	fdev->dma_device.device_prep_dma_memcpy = st_fdma_prep_dma_memcpy;
814	fdev->dma_device.device_tx_status = st_fdma_tx_status;
815	fdev->dma_device.device_issue_pending = st_fdma_issue_pending;
816	fdev->dma_device.device_terminate_all = st_fdma_terminate_all;
817	fdev->dma_device.device_config = st_fdma_slave_config;
818	fdev->dma_device.device_pause = st_fdma_pause;
819	fdev->dma_device.device_resume = st_fdma_resume;
820
821	fdev->dma_device.src_addr_widths = FDMA_DMA_BUSWIDTHS;
822	fdev->dma_device.dst_addr_widths = FDMA_DMA_BUSWIDTHS;
823	fdev->dma_device.directions = BIT(DMA_DEV_TO_MEM) | BIT(DMA_MEM_TO_DEV);
824	fdev->dma_device.residue_granularity = DMA_RESIDUE_GRANULARITY_BURST;
825
826	ret = dmaenginem_async_device_register(&fdev->dma_device);
827	if (ret) {
828		dev_err(&pdev->dev,
829			"Failed to register DMA device (%d)\n", ret);
830		goto err_rproc;
831	}
832
833	ret = of_dma_controller_register(np, st_fdma_of_xlate, fdev);
834	if (ret) {
835		dev_err(&pdev->dev,
836			"Failed to register controller (%d)\n", ret);
837		goto err_rproc;
838	}
839
840	dev_info(&pdev->dev, "ST FDMA engine driver, irq:%d\n", fdev->irq);
841
842	return 0;
843
844err_rproc:
845	st_fdma_free(fdev);
846	st_slim_rproc_put(fdev->slim_rproc);
847err:
848	return ret;
849}
850
851static int st_fdma_remove(struct platform_device *pdev)
852{
853	struct st_fdma_dev *fdev = platform_get_drvdata(pdev);
854
855	devm_free_irq(&pdev->dev, fdev->irq, fdev);
856	st_slim_rproc_put(fdev->slim_rproc);
857	of_dma_controller_free(pdev->dev.of_node);
858
859	return 0;
860}
861
862static struct platform_driver st_fdma_platform_driver = {
863	.driver = {
864		.name = DRIVER_NAME,
865		.of_match_table = st_fdma_match,
866	},
867	.probe = st_fdma_probe,
868	.remove = st_fdma_remove,
869};
870module_platform_driver(st_fdma_platform_driver);
871
872MODULE_LICENSE("GPL v2");
873MODULE_DESCRIPTION("STMicroelectronics FDMA engine driver");
874MODULE_AUTHOR("Ludovic.barre <Ludovic.barre@st.com>");
875MODULE_AUTHOR("Peter Griffin <peter.griffin@linaro.org>");
876MODULE_ALIAS("platform: " DRIVER_NAME);