Linux 6.9-rc1
[linux-2.6-microblaze.git] / drivers / dma / st_fdma.c
1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3  * DMA driver for STMicroelectronics STi FDMA controller
4  *
5  * Copyright (C) 2014 STMicroelectronics
6  *
7  * Author: Ludovic Barre <Ludovic.barre@st.com>
8  *         Peter Griffin <peter.griffin@linaro.org>
9  */
10
11 #include <linux/init.h>
12 #include <linux/module.h>
13 #include <linux/of.h>
14 #include <linux/of_dma.h>
15 #include <linux/platform_device.h>
16 #include <linux/property.h>
17 #include <linux/interrupt.h>
18 #include <linux/remoteproc.h>
19 #include <linux/slab.h>
20
21 #include "st_fdma.h"
22
23 static inline struct st_fdma_chan *to_st_fdma_chan(struct dma_chan *c)
24 {
25         return container_of(c, struct st_fdma_chan, vchan.chan);
26 }
27
28 static struct st_fdma_desc *to_st_fdma_desc(struct virt_dma_desc *vd)
29 {
30         return container_of(vd, struct st_fdma_desc, vdesc);
31 }
32
33 static int st_fdma_dreq_get(struct st_fdma_chan *fchan)
34 {
35         struct st_fdma_dev *fdev = fchan->fdev;
36         u32 req_line_cfg = fchan->cfg.req_line;
37         u32 dreq_line;
38         int try = 0;
39
40         /*
41          * dreq_mask is shared for n channels of fdma, so all accesses must be
42          * atomic. if the dreq_mask is changed between ffz and set_bit,
43          * we retry
44          */
45         do {
46                 if (fdev->dreq_mask == ~0L) {
47                         dev_err(fdev->dev, "No req lines available\n");
48                         return -EINVAL;
49                 }
50
51                 if (try || req_line_cfg >= ST_FDMA_NR_DREQS) {
52                         dev_err(fdev->dev, "Invalid or used req line\n");
53                         return -EINVAL;
54                 } else {
55                         dreq_line = req_line_cfg;
56                 }
57
58                 try++;
59         } while (test_and_set_bit(dreq_line, &fdev->dreq_mask));
60
61         dev_dbg(fdev->dev, "get dreq_line:%d mask:%#lx\n",
62                 dreq_line, fdev->dreq_mask);
63
64         return dreq_line;
65 }
66
67 static void st_fdma_dreq_put(struct st_fdma_chan *fchan)
68 {
69         struct st_fdma_dev *fdev = fchan->fdev;
70
71         dev_dbg(fdev->dev, "put dreq_line:%#x\n", fchan->dreq_line);
72         clear_bit(fchan->dreq_line, &fdev->dreq_mask);
73 }
74
75 static void st_fdma_xfer_desc(struct st_fdma_chan *fchan)
76 {
77         struct virt_dma_desc *vdesc;
78         unsigned long nbytes, ch_cmd, cmd;
79
80         vdesc = vchan_next_desc(&fchan->vchan);
81         if (!vdesc)
82                 return;
83
84         fchan->fdesc = to_st_fdma_desc(vdesc);
85         nbytes = fchan->fdesc->node[0].desc->nbytes;
86         cmd = FDMA_CMD_START(fchan->vchan.chan.chan_id);
87         ch_cmd = fchan->fdesc->node[0].pdesc | FDMA_CH_CMD_STA_START;
88
89         /* start the channel for the descriptor */
90         fnode_write(fchan, nbytes, FDMA_CNTN_OFST);
91         fchan_write(fchan, ch_cmd, FDMA_CH_CMD_OFST);
92         writel(cmd,
93                 fchan->fdev->slim_rproc->peri + FDMA_CMD_SET_OFST);
94
95         dev_dbg(fchan->fdev->dev, "start chan:%d\n", fchan->vchan.chan.chan_id);
96 }
97
98 static void st_fdma_ch_sta_update(struct st_fdma_chan *fchan,
99                                   unsigned long int_sta)
100 {
101         unsigned long ch_sta, ch_err;
102         int ch_id = fchan->vchan.chan.chan_id;
103         struct st_fdma_dev *fdev = fchan->fdev;
104
105         ch_sta = fchan_read(fchan, FDMA_CH_CMD_OFST);
106         ch_err = ch_sta & FDMA_CH_CMD_ERR_MASK;
107         ch_sta &= FDMA_CH_CMD_STA_MASK;
108
109         if (int_sta & FDMA_INT_STA_ERR) {
110                 dev_warn(fdev->dev, "chan:%d, error:%ld\n", ch_id, ch_err);
111                 fchan->status = DMA_ERROR;
112                 return;
113         }
114
115         switch (ch_sta) {
116         case FDMA_CH_CMD_STA_PAUSED:
117                 fchan->status = DMA_PAUSED;
118                 break;
119
120         case FDMA_CH_CMD_STA_RUNNING:
121                 fchan->status = DMA_IN_PROGRESS;
122                 break;
123         }
124 }
125
126 static irqreturn_t st_fdma_irq_handler(int irq, void *dev_id)
127 {
128         struct st_fdma_dev *fdev = dev_id;
129         irqreturn_t ret = IRQ_NONE;
130         struct st_fdma_chan *fchan = &fdev->chans[0];
131         unsigned long int_sta, clr;
132
133         int_sta = fdma_read(fdev, FDMA_INT_STA_OFST);
134         clr = int_sta;
135
136         for (; int_sta != 0 ; int_sta >>= 2, fchan++) {
137                 if (!(int_sta & (FDMA_INT_STA_CH | FDMA_INT_STA_ERR)))
138                         continue;
139
140                 spin_lock(&fchan->vchan.lock);
141                 st_fdma_ch_sta_update(fchan, int_sta);
142
143                 if (fchan->fdesc) {
144                         if (!fchan->fdesc->iscyclic) {
145                                 list_del(&fchan->fdesc->vdesc.node);
146                                 vchan_cookie_complete(&fchan->fdesc->vdesc);
147                                 fchan->fdesc = NULL;
148                                 fchan->status = DMA_COMPLETE;
149                         } else {
150                                 vchan_cyclic_callback(&fchan->fdesc->vdesc);
151                         }
152
153                         /* Start the next descriptor (if available) */
154                         if (!fchan->fdesc)
155                                 st_fdma_xfer_desc(fchan);
156                 }
157
158                 spin_unlock(&fchan->vchan.lock);
159                 ret = IRQ_HANDLED;
160         }
161
162         fdma_write(fdev, clr, FDMA_INT_CLR_OFST);
163
164         return ret;
165 }
166
167 static struct dma_chan *st_fdma_of_xlate(struct of_phandle_args *dma_spec,
168                                          struct of_dma *ofdma)
169 {
170         struct st_fdma_dev *fdev = ofdma->of_dma_data;
171         struct dma_chan *chan;
172         struct st_fdma_chan *fchan;
173         int ret;
174
175         if (dma_spec->args_count < 1)
176                 return ERR_PTR(-EINVAL);
177
178         if (fdev->dma_device.dev->of_node != dma_spec->np)
179                 return ERR_PTR(-EINVAL);
180
181         ret = rproc_boot(fdev->slim_rproc->rproc);
182         if (ret == -ENOENT)
183                 return ERR_PTR(-EPROBE_DEFER);
184         else if (ret)
185                 return ERR_PTR(ret);
186
187         chan = dma_get_any_slave_channel(&fdev->dma_device);
188         if (!chan)
189                 goto err_chan;
190
191         fchan = to_st_fdma_chan(chan);
192
193         fchan->cfg.of_node = dma_spec->np;
194         fchan->cfg.req_line = dma_spec->args[0];
195         fchan->cfg.req_ctrl = 0;
196         fchan->cfg.type = ST_FDMA_TYPE_FREE_RUN;
197
198         if (dma_spec->args_count > 1)
199                 fchan->cfg.req_ctrl = dma_spec->args[1]
200                         & FDMA_REQ_CTRL_CFG_MASK;
201
202         if (dma_spec->args_count > 2)
203                 fchan->cfg.type = dma_spec->args[2];
204
205         if (fchan->cfg.type == ST_FDMA_TYPE_FREE_RUN) {
206                 fchan->dreq_line = 0;
207         } else {
208                 fchan->dreq_line = st_fdma_dreq_get(fchan);
209                 if (IS_ERR_VALUE(fchan->dreq_line)) {
210                         chan = ERR_PTR(fchan->dreq_line);
211                         goto err_chan;
212                 }
213         }
214
215         dev_dbg(fdev->dev, "xlate req_line:%d type:%d req_ctrl:%#lx\n",
216                 fchan->cfg.req_line, fchan->cfg.type, fchan->cfg.req_ctrl);
217
218         return chan;
219
220 err_chan:
221         rproc_shutdown(fdev->slim_rproc->rproc);
222         return chan;
223
224 }
225
226 static void st_fdma_free_desc(struct virt_dma_desc *vdesc)
227 {
228         struct st_fdma_desc *fdesc;
229         int i;
230
231         fdesc = to_st_fdma_desc(vdesc);
232         for (i = 0; i < fdesc->n_nodes; i++)
233                 dma_pool_free(fdesc->fchan->node_pool, fdesc->node[i].desc,
234                               fdesc->node[i].pdesc);
235         kfree(fdesc);
236 }
237
238 static struct st_fdma_desc *st_fdma_alloc_desc(struct st_fdma_chan *fchan,
239                                                int sg_len)
240 {
241         struct st_fdma_desc *fdesc;
242         int i;
243
244         fdesc = kzalloc(struct_size(fdesc, node, sg_len), GFP_NOWAIT);
245         if (!fdesc)
246                 return NULL;
247
248         fdesc->fchan = fchan;
249         fdesc->n_nodes = sg_len;
250         for (i = 0; i < sg_len; i++) {
251                 fdesc->node[i].desc = dma_pool_alloc(fchan->node_pool,
252                                 GFP_NOWAIT, &fdesc->node[i].pdesc);
253                 if (!fdesc->node[i].desc)
254                         goto err;
255         }
256         return fdesc;
257
258 err:
259         while (--i >= 0)
260                 dma_pool_free(fchan->node_pool, fdesc->node[i].desc,
261                               fdesc->node[i].pdesc);
262         kfree(fdesc);
263         return NULL;
264 }
265
266 static int st_fdma_alloc_chan_res(struct dma_chan *chan)
267 {
268         struct st_fdma_chan *fchan = to_st_fdma_chan(chan);
269
270         /* Create the dma pool for descriptor allocation */
271         fchan->node_pool = dma_pool_create(dev_name(&chan->dev->device),
272                                             fchan->fdev->dev,
273                                             sizeof(struct st_fdma_hw_node),
274                                             __alignof__(struct st_fdma_hw_node),
275                                             0);
276
277         if (!fchan->node_pool) {
278                 dev_err(fchan->fdev->dev, "unable to allocate desc pool\n");
279                 return -ENOMEM;
280         }
281
282         dev_dbg(fchan->fdev->dev, "alloc ch_id:%d type:%d\n",
283                 fchan->vchan.chan.chan_id, fchan->cfg.type);
284
285         return 0;
286 }
287
288 static void st_fdma_free_chan_res(struct dma_chan *chan)
289 {
290         struct st_fdma_chan *fchan = to_st_fdma_chan(chan);
291         struct rproc *rproc = fchan->fdev->slim_rproc->rproc;
292         unsigned long flags;
293
294         dev_dbg(fchan->fdev->dev, "%s: freeing chan:%d\n",
295                 __func__, fchan->vchan.chan.chan_id);
296
297         if (fchan->cfg.type != ST_FDMA_TYPE_FREE_RUN)
298                 st_fdma_dreq_put(fchan);
299
300         spin_lock_irqsave(&fchan->vchan.lock, flags);
301         fchan->fdesc = NULL;
302         spin_unlock_irqrestore(&fchan->vchan.lock, flags);
303
304         dma_pool_destroy(fchan->node_pool);
305         fchan->node_pool = NULL;
306         memset(&fchan->cfg, 0, sizeof(struct st_fdma_cfg));
307
308         rproc_shutdown(rproc);
309 }
310
311 static struct dma_async_tx_descriptor *st_fdma_prep_dma_memcpy(
312         struct dma_chan *chan,  dma_addr_t dst, dma_addr_t src,
313         size_t len, unsigned long flags)
314 {
315         struct st_fdma_chan *fchan;
316         struct st_fdma_desc *fdesc;
317         struct st_fdma_hw_node *hw_node;
318
319         if (!len)
320                 return NULL;
321
322         fchan = to_st_fdma_chan(chan);
323
324         /* We only require a single descriptor */
325         fdesc = st_fdma_alloc_desc(fchan, 1);
326         if (!fdesc) {
327                 dev_err(fchan->fdev->dev, "no memory for desc\n");
328                 return NULL;
329         }
330
331         hw_node = fdesc->node[0].desc;
332         hw_node->next = 0;
333         hw_node->control = FDMA_NODE_CTRL_REQ_MAP_FREE_RUN;
334         hw_node->control |= FDMA_NODE_CTRL_SRC_INCR;
335         hw_node->control |= FDMA_NODE_CTRL_DST_INCR;
336         hw_node->control |= FDMA_NODE_CTRL_INT_EON;
337         hw_node->nbytes = len;
338         hw_node->saddr = src;
339         hw_node->daddr = dst;
340         hw_node->generic.length = len;
341         hw_node->generic.sstride = 0;
342         hw_node->generic.dstride = 0;
343
344         return vchan_tx_prep(&fchan->vchan, &fdesc->vdesc, flags);
345 }
346
347 static int config_reqctrl(struct st_fdma_chan *fchan,
348                           enum dma_transfer_direction direction)
349 {
350         u32 maxburst = 0, addr = 0;
351         enum dma_slave_buswidth width;
352         int ch_id = fchan->vchan.chan.chan_id;
353         struct st_fdma_dev *fdev = fchan->fdev;
354
355         switch (direction) {
356
357         case DMA_DEV_TO_MEM:
358                 fchan->cfg.req_ctrl &= ~FDMA_REQ_CTRL_WNR;
359                 maxburst = fchan->scfg.src_maxburst;
360                 width = fchan->scfg.src_addr_width;
361                 addr = fchan->scfg.src_addr;
362                 break;
363
364         case DMA_MEM_TO_DEV:
365                 fchan->cfg.req_ctrl |= FDMA_REQ_CTRL_WNR;
366                 maxburst = fchan->scfg.dst_maxburst;
367                 width = fchan->scfg.dst_addr_width;
368                 addr = fchan->scfg.dst_addr;
369                 break;
370
371         default:
372                 return -EINVAL;
373         }
374
375         fchan->cfg.req_ctrl &= ~FDMA_REQ_CTRL_OPCODE_MASK;
376
377         switch (width) {
378
379         case DMA_SLAVE_BUSWIDTH_1_BYTE:
380                 fchan->cfg.req_ctrl |= FDMA_REQ_CTRL_OPCODE_LD_ST1;
381                 break;
382
383         case DMA_SLAVE_BUSWIDTH_2_BYTES:
384                 fchan->cfg.req_ctrl |= FDMA_REQ_CTRL_OPCODE_LD_ST2;
385                 break;
386
387         case DMA_SLAVE_BUSWIDTH_4_BYTES:
388                 fchan->cfg.req_ctrl |= FDMA_REQ_CTRL_OPCODE_LD_ST4;
389                 break;
390
391         case DMA_SLAVE_BUSWIDTH_8_BYTES:
392                 fchan->cfg.req_ctrl |= FDMA_REQ_CTRL_OPCODE_LD_ST8;
393                 break;
394
395         default:
396                 return -EINVAL;
397         }
398
399         fchan->cfg.req_ctrl &= ~FDMA_REQ_CTRL_NUM_OPS_MASK;
400         fchan->cfg.req_ctrl |= FDMA_REQ_CTRL_NUM_OPS(maxburst-1);
401         dreq_write(fchan, fchan->cfg.req_ctrl, FDMA_REQ_CTRL_OFST);
402
403         fchan->cfg.dev_addr = addr;
404         fchan->cfg.dir = direction;
405
406         dev_dbg(fdev->dev, "chan:%d config_reqctrl:%#x req_ctrl:%#lx\n",
407                 ch_id, addr, fchan->cfg.req_ctrl);
408
409         return 0;
410 }
411
412 static void fill_hw_node(struct st_fdma_hw_node *hw_node,
413                         struct st_fdma_chan *fchan,
414                         enum dma_transfer_direction direction)
415 {
416         if (direction == DMA_MEM_TO_DEV) {
417                 hw_node->control |= FDMA_NODE_CTRL_SRC_INCR;
418                 hw_node->control |= FDMA_NODE_CTRL_DST_STATIC;
419                 hw_node->daddr = fchan->cfg.dev_addr;
420         } else {
421                 hw_node->control |= FDMA_NODE_CTRL_SRC_STATIC;
422                 hw_node->control |= FDMA_NODE_CTRL_DST_INCR;
423                 hw_node->saddr = fchan->cfg.dev_addr;
424         }
425
426         hw_node->generic.sstride = 0;
427         hw_node->generic.dstride = 0;
428 }
429
430 static inline struct st_fdma_chan *st_fdma_prep_common(struct dma_chan *chan,
431                 size_t len, enum dma_transfer_direction direction)
432 {
433         struct st_fdma_chan *fchan;
434
435         if (!chan || !len)
436                 return NULL;
437
438         fchan = to_st_fdma_chan(chan);
439
440         if (!is_slave_direction(direction)) {
441                 dev_err(fchan->fdev->dev, "bad direction?\n");
442                 return NULL;
443         }
444
445         return fchan;
446 }
447
448 static struct dma_async_tx_descriptor *st_fdma_prep_dma_cyclic(
449                 struct dma_chan *chan, dma_addr_t buf_addr, size_t len,
450                 size_t period_len, enum dma_transfer_direction direction,
451                 unsigned long flags)
452 {
453         struct st_fdma_chan *fchan;
454         struct st_fdma_desc *fdesc;
455         int sg_len, i;
456
457         fchan = st_fdma_prep_common(chan, len, direction);
458         if (!fchan)
459                 return NULL;
460
461         if (!period_len)
462                 return NULL;
463
464         if (config_reqctrl(fchan, direction)) {
465                 dev_err(fchan->fdev->dev, "bad width or direction\n");
466                 return NULL;
467         }
468
469         /* the buffer length must be a multiple of period_len */
470         if (len % period_len != 0) {
471                 dev_err(fchan->fdev->dev, "len is not multiple of period\n");
472                 return NULL;
473         }
474
475         sg_len = len / period_len;
476         fdesc = st_fdma_alloc_desc(fchan, sg_len);
477         if (!fdesc) {
478                 dev_err(fchan->fdev->dev, "no memory for desc\n");
479                 return NULL;
480         }
481
482         fdesc->iscyclic = true;
483
484         for (i = 0; i < sg_len; i++) {
485                 struct st_fdma_hw_node *hw_node = fdesc->node[i].desc;
486
487                 hw_node->next = fdesc->node[(i + 1) % sg_len].pdesc;
488
489                 hw_node->control =
490                         FDMA_NODE_CTRL_REQ_MAP_DREQ(fchan->dreq_line);
491                 hw_node->control |= FDMA_NODE_CTRL_INT_EON;
492
493                 fill_hw_node(hw_node, fchan, direction);
494
495                 if (direction == DMA_MEM_TO_DEV)
496                         hw_node->saddr = buf_addr + (i * period_len);
497                 else
498                         hw_node->daddr = buf_addr + (i * period_len);
499
500                 hw_node->nbytes = period_len;
501                 hw_node->generic.length = period_len;
502         }
503
504         return vchan_tx_prep(&fchan->vchan, &fdesc->vdesc, flags);
505 }
506
507 static struct dma_async_tx_descriptor *st_fdma_prep_slave_sg(
508                 struct dma_chan *chan, struct scatterlist *sgl,
509                 unsigned int sg_len, enum dma_transfer_direction direction,
510                 unsigned long flags, void *context)
511 {
512         struct st_fdma_chan *fchan;
513         struct st_fdma_desc *fdesc;
514         struct st_fdma_hw_node *hw_node;
515         struct scatterlist *sg;
516         int i;
517
518         fchan = st_fdma_prep_common(chan, sg_len, direction);
519         if (!fchan)
520                 return NULL;
521
522         if (!sgl)
523                 return NULL;
524
525         fdesc = st_fdma_alloc_desc(fchan, sg_len);
526         if (!fdesc) {
527                 dev_err(fchan->fdev->dev, "no memory for desc\n");
528                 return NULL;
529         }
530
531         fdesc->iscyclic = false;
532
533         for_each_sg(sgl, sg, sg_len, i) {
534                 hw_node = fdesc->node[i].desc;
535
536                 hw_node->next = fdesc->node[(i + 1) % sg_len].pdesc;
537                 hw_node->control = FDMA_NODE_CTRL_REQ_MAP_DREQ(fchan->dreq_line);
538
539                 fill_hw_node(hw_node, fchan, direction);
540
541                 if (direction == DMA_MEM_TO_DEV)
542                         hw_node->saddr = sg_dma_address(sg);
543                 else
544                         hw_node->daddr = sg_dma_address(sg);
545
546                 hw_node->nbytes = sg_dma_len(sg);
547                 hw_node->generic.length = sg_dma_len(sg);
548         }
549
550         /* interrupt at end of last node */
551         hw_node->control |= FDMA_NODE_CTRL_INT_EON;
552
553         return vchan_tx_prep(&fchan->vchan, &fdesc->vdesc, flags);
554 }
555
556 static size_t st_fdma_desc_residue(struct st_fdma_chan *fchan,
557                                    struct virt_dma_desc *vdesc,
558                                    bool in_progress)
559 {
560         struct st_fdma_desc *fdesc = fchan->fdesc;
561         size_t residue = 0;
562         dma_addr_t cur_addr = 0;
563         int i;
564
565         if (in_progress) {
566                 cur_addr = fchan_read(fchan, FDMA_CH_CMD_OFST);
567                 cur_addr &= FDMA_CH_CMD_DATA_MASK;
568         }
569
570         for (i = fchan->fdesc->n_nodes - 1 ; i >= 0; i--) {
571                 if (cur_addr == fdesc->node[i].pdesc) {
572                         residue += fnode_read(fchan, FDMA_CNTN_OFST);
573                         break;
574                 }
575                 residue += fdesc->node[i].desc->nbytes;
576         }
577
578         return residue;
579 }
580
581 static enum dma_status st_fdma_tx_status(struct dma_chan *chan,
582                                          dma_cookie_t cookie,
583                                          struct dma_tx_state *txstate)
584 {
585         struct st_fdma_chan *fchan = to_st_fdma_chan(chan);
586         struct virt_dma_desc *vd;
587         enum dma_status ret;
588         unsigned long flags;
589
590         ret = dma_cookie_status(chan, cookie, txstate);
591         if (ret == DMA_COMPLETE || !txstate)
592                 return ret;
593
594         spin_lock_irqsave(&fchan->vchan.lock, flags);
595         vd = vchan_find_desc(&fchan->vchan, cookie);
596         if (fchan->fdesc && cookie == fchan->fdesc->vdesc.tx.cookie)
597                 txstate->residue = st_fdma_desc_residue(fchan, vd, true);
598         else if (vd)
599                 txstate->residue = st_fdma_desc_residue(fchan, vd, false);
600         else
601                 txstate->residue = 0;
602
603         spin_unlock_irqrestore(&fchan->vchan.lock, flags);
604
605         return ret;
606 }
607
608 static void st_fdma_issue_pending(struct dma_chan *chan)
609 {
610         struct st_fdma_chan *fchan = to_st_fdma_chan(chan);
611         unsigned long flags;
612
613         spin_lock_irqsave(&fchan->vchan.lock, flags);
614
615         if (vchan_issue_pending(&fchan->vchan) && !fchan->fdesc)
616                 st_fdma_xfer_desc(fchan);
617
618         spin_unlock_irqrestore(&fchan->vchan.lock, flags);
619 }
620
621 static int st_fdma_pause(struct dma_chan *chan)
622 {
623         unsigned long flags;
624         struct st_fdma_chan *fchan = to_st_fdma_chan(chan);
625         int ch_id = fchan->vchan.chan.chan_id;
626         unsigned long cmd = FDMA_CMD_PAUSE(ch_id);
627
628         dev_dbg(fchan->fdev->dev, "pause chan:%d\n", ch_id);
629
630         spin_lock_irqsave(&fchan->vchan.lock, flags);
631         if (fchan->fdesc)
632                 fdma_write(fchan->fdev, cmd, FDMA_CMD_SET_OFST);
633         spin_unlock_irqrestore(&fchan->vchan.lock, flags);
634
635         return 0;
636 }
637
638 static int st_fdma_resume(struct dma_chan *chan)
639 {
640         unsigned long flags;
641         unsigned long val;
642         struct st_fdma_chan *fchan = to_st_fdma_chan(chan);
643         int ch_id = fchan->vchan.chan.chan_id;
644
645         dev_dbg(fchan->fdev->dev, "resume chan:%d\n", ch_id);
646
647         spin_lock_irqsave(&fchan->vchan.lock, flags);
648         if (fchan->fdesc) {
649                 val = fchan_read(fchan, FDMA_CH_CMD_OFST);
650                 val &= FDMA_CH_CMD_DATA_MASK;
651                 fchan_write(fchan, val, FDMA_CH_CMD_OFST);
652         }
653         spin_unlock_irqrestore(&fchan->vchan.lock, flags);
654
655         return 0;
656 }
657
658 static int st_fdma_terminate_all(struct dma_chan *chan)
659 {
660         unsigned long flags;
661         LIST_HEAD(head);
662         struct st_fdma_chan *fchan = to_st_fdma_chan(chan);
663         int ch_id = fchan->vchan.chan.chan_id;
664         unsigned long cmd = FDMA_CMD_PAUSE(ch_id);
665
666         dev_dbg(fchan->fdev->dev, "terminate chan:%d\n", ch_id);
667
668         spin_lock_irqsave(&fchan->vchan.lock, flags);
669         fdma_write(fchan->fdev, cmd, FDMA_CMD_SET_OFST);
670         fchan->fdesc = NULL;
671         vchan_get_all_descriptors(&fchan->vchan, &head);
672         spin_unlock_irqrestore(&fchan->vchan.lock, flags);
673         vchan_dma_desc_free_list(&fchan->vchan, &head);
674
675         return 0;
676 }
677
678 static int st_fdma_slave_config(struct dma_chan *chan,
679                                 struct dma_slave_config *slave_cfg)
680 {
681         struct st_fdma_chan *fchan = to_st_fdma_chan(chan);
682
683         memcpy(&fchan->scfg, slave_cfg, sizeof(fchan->scfg));
684         return 0;
685 }
686
687 static const struct st_fdma_driverdata fdma_mpe31_stih407_11 = {
688         .name = "STiH407",
689         .id = 0,
690 };
691
692 static const struct st_fdma_driverdata fdma_mpe31_stih407_12 = {
693         .name = "STiH407",
694         .id = 1,
695 };
696
697 static const struct st_fdma_driverdata fdma_mpe31_stih407_13 = {
698         .name = "STiH407",
699         .id = 2,
700 };
701
702 static const struct of_device_id st_fdma_match[] = {
703         { .compatible = "st,stih407-fdma-mpe31-11"
704           , .data = &fdma_mpe31_stih407_11 },
705         { .compatible = "st,stih407-fdma-mpe31-12"
706           , .data = &fdma_mpe31_stih407_12 },
707         { .compatible = "st,stih407-fdma-mpe31-13"
708           , .data = &fdma_mpe31_stih407_13 },
709         {},
710 };
711 MODULE_DEVICE_TABLE(of, st_fdma_match);
712
713 static int st_fdma_parse_dt(struct platform_device *pdev,
714                         const struct st_fdma_driverdata *drvdata,
715                         struct st_fdma_dev *fdev)
716 {
717         snprintf(fdev->fw_name, FW_NAME_SIZE, "fdma_%s_%d.elf",
718                 drvdata->name, drvdata->id);
719
720         return of_property_read_u32(pdev->dev.of_node, "dma-channels",
721                                     &fdev->nr_channels);
722 }
723 #define FDMA_DMA_BUSWIDTHS      (BIT(DMA_SLAVE_BUSWIDTH_1_BYTE) | \
724                                  BIT(DMA_SLAVE_BUSWIDTH_2_BYTES) | \
725                                  BIT(DMA_SLAVE_BUSWIDTH_3_BYTES) | \
726                                  BIT(DMA_SLAVE_BUSWIDTH_4_BYTES))
727
728 static void st_fdma_free(struct st_fdma_dev *fdev)
729 {
730         struct st_fdma_chan *fchan;
731         int i;
732
733         for (i = 0; i < fdev->nr_channels; i++) {
734                 fchan = &fdev->chans[i];
735                 list_del(&fchan->vchan.chan.device_node);
736                 tasklet_kill(&fchan->vchan.task);
737         }
738 }
739
740 static int st_fdma_probe(struct platform_device *pdev)
741 {
742         struct st_fdma_dev *fdev;
743         struct device_node *np = pdev->dev.of_node;
744         const struct st_fdma_driverdata *drvdata;
745         int ret, i;
746
747         drvdata = device_get_match_data(&pdev->dev);
748
749         fdev = devm_kzalloc(&pdev->dev, sizeof(*fdev), GFP_KERNEL);
750         if (!fdev)
751                 return -ENOMEM;
752
753         ret = st_fdma_parse_dt(pdev, drvdata, fdev);
754         if (ret) {
755                 dev_err(&pdev->dev, "unable to find platform data\n");
756                 goto err;
757         }
758
759         fdev->chans = devm_kcalloc(&pdev->dev, fdev->nr_channels,
760                                    sizeof(struct st_fdma_chan), GFP_KERNEL);
761         if (!fdev->chans)
762                 return -ENOMEM;
763
764         fdev->dev = &pdev->dev;
765         fdev->drvdata = drvdata;
766         platform_set_drvdata(pdev, fdev);
767
768         fdev->irq = platform_get_irq(pdev, 0);
769         if (fdev->irq < 0)
770                 return -EINVAL;
771
772         ret = devm_request_irq(&pdev->dev, fdev->irq, st_fdma_irq_handler, 0,
773                                dev_name(&pdev->dev), fdev);
774         if (ret) {
775                 dev_err(&pdev->dev, "Failed to request irq (%d)\n", ret);
776                 goto err;
777         }
778
779         fdev->slim_rproc = st_slim_rproc_alloc(pdev, fdev->fw_name);
780         if (IS_ERR(fdev->slim_rproc)) {
781                 ret = PTR_ERR(fdev->slim_rproc);
782                 dev_err(&pdev->dev, "slim_rproc_alloc failed (%d)\n", ret);
783                 goto err;
784         }
785
786         /* Initialise list of FDMA channels */
787         INIT_LIST_HEAD(&fdev->dma_device.channels);
788         for (i = 0; i < fdev->nr_channels; i++) {
789                 struct st_fdma_chan *fchan = &fdev->chans[i];
790
791                 fchan->fdev = fdev;
792                 fchan->vchan.desc_free = st_fdma_free_desc;
793                 vchan_init(&fchan->vchan, &fdev->dma_device);
794         }
795
796         /* Initialise the FDMA dreq (reserve 0 & 31 for FDMA use) */
797         fdev->dreq_mask = BIT(0) | BIT(31);
798
799         dma_cap_set(DMA_SLAVE, fdev->dma_device.cap_mask);
800         dma_cap_set(DMA_CYCLIC, fdev->dma_device.cap_mask);
801         dma_cap_set(DMA_MEMCPY, fdev->dma_device.cap_mask);
802
803         fdev->dma_device.dev = &pdev->dev;
804         fdev->dma_device.device_alloc_chan_resources = st_fdma_alloc_chan_res;
805         fdev->dma_device.device_free_chan_resources = st_fdma_free_chan_res;
806         fdev->dma_device.device_prep_dma_cyclic = st_fdma_prep_dma_cyclic;
807         fdev->dma_device.device_prep_slave_sg = st_fdma_prep_slave_sg;
808         fdev->dma_device.device_prep_dma_memcpy = st_fdma_prep_dma_memcpy;
809         fdev->dma_device.device_tx_status = st_fdma_tx_status;
810         fdev->dma_device.device_issue_pending = st_fdma_issue_pending;
811         fdev->dma_device.device_terminate_all = st_fdma_terminate_all;
812         fdev->dma_device.device_config = st_fdma_slave_config;
813         fdev->dma_device.device_pause = st_fdma_pause;
814         fdev->dma_device.device_resume = st_fdma_resume;
815
816         fdev->dma_device.src_addr_widths = FDMA_DMA_BUSWIDTHS;
817         fdev->dma_device.dst_addr_widths = FDMA_DMA_BUSWIDTHS;
818         fdev->dma_device.directions = BIT(DMA_DEV_TO_MEM) | BIT(DMA_MEM_TO_DEV);
819         fdev->dma_device.residue_granularity = DMA_RESIDUE_GRANULARITY_BURST;
820
821         ret = dmaenginem_async_device_register(&fdev->dma_device);
822         if (ret) {
823                 dev_err(&pdev->dev,
824                         "Failed to register DMA device (%d)\n", ret);
825                 goto err_rproc;
826         }
827
828         ret = of_dma_controller_register(np, st_fdma_of_xlate, fdev);
829         if (ret) {
830                 dev_err(&pdev->dev,
831                         "Failed to register controller (%d)\n", ret);
832                 goto err_rproc;
833         }
834
835         dev_info(&pdev->dev, "ST FDMA engine driver, irq:%d\n", fdev->irq);
836
837         return 0;
838
839 err_rproc:
840         st_fdma_free(fdev);
841         st_slim_rproc_put(fdev->slim_rproc);
842 err:
843         return ret;
844 }
845
846 static void st_fdma_remove(struct platform_device *pdev)
847 {
848         struct st_fdma_dev *fdev = platform_get_drvdata(pdev);
849
850         devm_free_irq(&pdev->dev, fdev->irq, fdev);
851         st_slim_rproc_put(fdev->slim_rproc);
852         of_dma_controller_free(pdev->dev.of_node);
853 }
854
855 static struct platform_driver st_fdma_platform_driver = {
856         .driver = {
857                 .name = DRIVER_NAME,
858                 .of_match_table = st_fdma_match,
859         },
860         .probe = st_fdma_probe,
861         .remove_new = st_fdma_remove,
862 };
863 module_platform_driver(st_fdma_platform_driver);
864
865 MODULE_LICENSE("GPL v2");
866 MODULE_DESCRIPTION("STMicroelectronics FDMA engine driver");
867 MODULE_AUTHOR("Ludovic.barre <Ludovic.barre@st.com>");
868 MODULE_AUTHOR("Peter Griffin <peter.griffin@linaro.org>");
869 MODULE_ALIAS("platform:" DRIVER_NAME);