Merge branch 'next' into for-linus
[linux-2.6-microblaze.git] / drivers / dma / mv_xor.c
1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  * offload engine driver for the Marvell XOR engine
4  * Copyright (C) 2007, 2008, Marvell International Ltd.
5  */
6
7 #include <linux/init.h>
8 #include <linux/slab.h>
9 #include <linux/delay.h>
10 #include <linux/dma-mapping.h>
11 #include <linux/spinlock.h>
12 #include <linux/interrupt.h>
13 #include <linux/of_device.h>
14 #include <linux/platform_device.h>
15 #include <linux/memory.h>
16 #include <linux/clk.h>
17 #include <linux/of.h>
18 #include <linux/of_irq.h>
19 #include <linux/irqdomain.h>
20 #include <linux/cpumask.h>
21 #include <linux/platform_data/dma-mv_xor.h>
22
23 #include "dmaengine.h"
24 #include "mv_xor.h"
25
26 enum mv_xor_type {
27         XOR_ORION,
28         XOR_ARMADA_38X,
29         XOR_ARMADA_37XX,
30 };
31
32 enum mv_xor_mode {
33         XOR_MODE_IN_REG,
34         XOR_MODE_IN_DESC,
35 };
36
37 static void mv_xor_issue_pending(struct dma_chan *chan);
38
39 #define to_mv_xor_chan(chan)            \
40         container_of(chan, struct mv_xor_chan, dmachan)
41
42 #define to_mv_xor_slot(tx)              \
43         container_of(tx, struct mv_xor_desc_slot, async_tx)
44
45 #define mv_chan_to_devp(chan)           \
46         ((chan)->dmadev.dev)
47
48 static void mv_desc_init(struct mv_xor_desc_slot *desc,
49                          dma_addr_t addr, u32 byte_count,
50                          enum dma_ctrl_flags flags)
51 {
52         struct mv_xor_desc *hw_desc = desc->hw_desc;
53
54         hw_desc->status = XOR_DESC_DMA_OWNED;
55         hw_desc->phy_next_desc = 0;
56         /* Enable end-of-descriptor interrupts only for DMA_PREP_INTERRUPT */
57         hw_desc->desc_command = (flags & DMA_PREP_INTERRUPT) ?
58                                 XOR_DESC_EOD_INT_EN : 0;
59         hw_desc->phy_dest_addr = addr;
60         hw_desc->byte_count = byte_count;
61 }
62
63 static void mv_desc_set_mode(struct mv_xor_desc_slot *desc)
64 {
65         struct mv_xor_desc *hw_desc = desc->hw_desc;
66
67         switch (desc->type) {
68         case DMA_XOR:
69         case DMA_INTERRUPT:
70                 hw_desc->desc_command |= XOR_DESC_OPERATION_XOR;
71                 break;
72         case DMA_MEMCPY:
73                 hw_desc->desc_command |= XOR_DESC_OPERATION_MEMCPY;
74                 break;
75         default:
76                 BUG();
77                 return;
78         }
79 }
80
81 static void mv_desc_set_next_desc(struct mv_xor_desc_slot *desc,
82                                   u32 next_desc_addr)
83 {
84         struct mv_xor_desc *hw_desc = desc->hw_desc;
85         BUG_ON(hw_desc->phy_next_desc);
86         hw_desc->phy_next_desc = next_desc_addr;
87 }
88
89 static void mv_desc_set_src_addr(struct mv_xor_desc_slot *desc,
90                                  int index, dma_addr_t addr)
91 {
92         struct mv_xor_desc *hw_desc = desc->hw_desc;
93         hw_desc->phy_src_addr[mv_phy_src_idx(index)] = addr;
94         if (desc->type == DMA_XOR)
95                 hw_desc->desc_command |= (1 << index);
96 }
97
98 static u32 mv_chan_get_current_desc(struct mv_xor_chan *chan)
99 {
100         return readl_relaxed(XOR_CURR_DESC(chan));
101 }
102
103 static void mv_chan_set_next_descriptor(struct mv_xor_chan *chan,
104                                         u32 next_desc_addr)
105 {
106         writel_relaxed(next_desc_addr, XOR_NEXT_DESC(chan));
107 }
108
109 static void mv_chan_unmask_interrupts(struct mv_xor_chan *chan)
110 {
111         u32 val = readl_relaxed(XOR_INTR_MASK(chan));
112         val |= XOR_INTR_MASK_VALUE << (chan->idx * 16);
113         writel_relaxed(val, XOR_INTR_MASK(chan));
114 }
115
116 static u32 mv_chan_get_intr_cause(struct mv_xor_chan *chan)
117 {
118         u32 intr_cause = readl_relaxed(XOR_INTR_CAUSE(chan));
119         intr_cause = (intr_cause >> (chan->idx * 16)) & 0xFFFF;
120         return intr_cause;
121 }
122
123 static void mv_chan_clear_eoc_cause(struct mv_xor_chan *chan)
124 {
125         u32 val;
126
127         val = XOR_INT_END_OF_DESC | XOR_INT_END_OF_CHAIN | XOR_INT_STOPPED;
128         val = ~(val << (chan->idx * 16));
129         dev_dbg(mv_chan_to_devp(chan), "%s, val 0x%08x\n", __func__, val);
130         writel_relaxed(val, XOR_INTR_CAUSE(chan));
131 }
132
133 static void mv_chan_clear_err_status(struct mv_xor_chan *chan)
134 {
135         u32 val = 0xFFFF0000 >> (chan->idx * 16);
136         writel_relaxed(val, XOR_INTR_CAUSE(chan));
137 }
138
139 static void mv_chan_set_mode(struct mv_xor_chan *chan,
140                              u32 op_mode)
141 {
142         u32 config = readl_relaxed(XOR_CONFIG(chan));
143
144         config &= ~0x7;
145         config |= op_mode;
146
147 #if defined(__BIG_ENDIAN)
148         config |= XOR_DESCRIPTOR_SWAP;
149 #else
150         config &= ~XOR_DESCRIPTOR_SWAP;
151 #endif
152
153         writel_relaxed(config, XOR_CONFIG(chan));
154 }
155
156 static void mv_chan_activate(struct mv_xor_chan *chan)
157 {
158         dev_dbg(mv_chan_to_devp(chan), " activate chan.\n");
159
160         /* writel ensures all descriptors are flushed before activation */
161         writel(BIT(0), XOR_ACTIVATION(chan));
162 }
163
164 static char mv_chan_is_busy(struct mv_xor_chan *chan)
165 {
166         u32 state = readl_relaxed(XOR_ACTIVATION(chan));
167
168         state = (state >> 4) & 0x3;
169
170         return (state == 1) ? 1 : 0;
171 }
172
173 /*
174  * mv_chan_start_new_chain - program the engine to operate on new
175  * chain headed by sw_desc
176  * Caller must hold &mv_chan->lock while calling this function
177  */
178 static void mv_chan_start_new_chain(struct mv_xor_chan *mv_chan,
179                                     struct mv_xor_desc_slot *sw_desc)
180 {
181         dev_dbg(mv_chan_to_devp(mv_chan), "%s %d: sw_desc %p\n",
182                 __func__, __LINE__, sw_desc);
183
184         /* set the hardware chain */
185         mv_chan_set_next_descriptor(mv_chan, sw_desc->async_tx.phys);
186
187         mv_chan->pending++;
188         mv_xor_issue_pending(&mv_chan->dmachan);
189 }
190
191 static dma_cookie_t
192 mv_desc_run_tx_complete_actions(struct mv_xor_desc_slot *desc,
193                                 struct mv_xor_chan *mv_chan,
194                                 dma_cookie_t cookie)
195 {
196         BUG_ON(desc->async_tx.cookie < 0);
197
198         if (desc->async_tx.cookie > 0) {
199                 cookie = desc->async_tx.cookie;
200
201                 dma_descriptor_unmap(&desc->async_tx);
202                 /* call the callback (must not sleep or submit new
203                  * operations to this channel)
204                  */
205                 dmaengine_desc_get_callback_invoke(&desc->async_tx, NULL);
206         }
207
208         /* run dependent operations */
209         dma_run_dependencies(&desc->async_tx);
210
211         return cookie;
212 }
213
214 static int
215 mv_chan_clean_completed_slots(struct mv_xor_chan *mv_chan)
216 {
217         struct mv_xor_desc_slot *iter, *_iter;
218
219         dev_dbg(mv_chan_to_devp(mv_chan), "%s %d\n", __func__, __LINE__);
220         list_for_each_entry_safe(iter, _iter, &mv_chan->completed_slots,
221                                  node) {
222
223                 if (async_tx_test_ack(&iter->async_tx)) {
224                         list_move_tail(&iter->node, &mv_chan->free_slots);
225                         if (!list_empty(&iter->sg_tx_list)) {
226                                 list_splice_tail_init(&iter->sg_tx_list,
227                                                         &mv_chan->free_slots);
228                         }
229                 }
230         }
231         return 0;
232 }
233
234 static int
235 mv_desc_clean_slot(struct mv_xor_desc_slot *desc,
236                    struct mv_xor_chan *mv_chan)
237 {
238         dev_dbg(mv_chan_to_devp(mv_chan), "%s %d: desc %p flags %d\n",
239                 __func__, __LINE__, desc, desc->async_tx.flags);
240
241         /* the client is allowed to attach dependent operations
242          * until 'ack' is set
243          */
244         if (!async_tx_test_ack(&desc->async_tx)) {
245                 /* move this slot to the completed_slots */
246                 list_move_tail(&desc->node, &mv_chan->completed_slots);
247                 if (!list_empty(&desc->sg_tx_list)) {
248                         list_splice_tail_init(&desc->sg_tx_list,
249                                               &mv_chan->completed_slots);
250                 }
251         } else {
252                 list_move_tail(&desc->node, &mv_chan->free_slots);
253                 if (!list_empty(&desc->sg_tx_list)) {
254                         list_splice_tail_init(&desc->sg_tx_list,
255                                               &mv_chan->free_slots);
256                 }
257         }
258
259         return 0;
260 }
261
262 /* This function must be called with the mv_xor_chan spinlock held */
263 static void mv_chan_slot_cleanup(struct mv_xor_chan *mv_chan)
264 {
265         struct mv_xor_desc_slot *iter, *_iter;
266         dma_cookie_t cookie = 0;
267         int busy = mv_chan_is_busy(mv_chan);
268         u32 current_desc = mv_chan_get_current_desc(mv_chan);
269         int current_cleaned = 0;
270         struct mv_xor_desc *hw_desc;
271
272         dev_dbg(mv_chan_to_devp(mv_chan), "%s %d\n", __func__, __LINE__);
273         dev_dbg(mv_chan_to_devp(mv_chan), "current_desc %x\n", current_desc);
274         mv_chan_clean_completed_slots(mv_chan);
275
276         /* free completed slots from the chain starting with
277          * the oldest descriptor
278          */
279
280         list_for_each_entry_safe(iter, _iter, &mv_chan->chain,
281                                  node) {
282
283                 /* clean finished descriptors */
284                 hw_desc = iter->hw_desc;
285                 if (hw_desc->status & XOR_DESC_SUCCESS) {
286                         cookie = mv_desc_run_tx_complete_actions(iter, mv_chan,
287                                                                  cookie);
288
289                         /* done processing desc, clean slot */
290                         mv_desc_clean_slot(iter, mv_chan);
291
292                         /* break if we did cleaned the current */
293                         if (iter->async_tx.phys == current_desc) {
294                                 current_cleaned = 1;
295                                 break;
296                         }
297                 } else {
298                         if (iter->async_tx.phys == current_desc) {
299                                 current_cleaned = 0;
300                                 break;
301                         }
302                 }
303         }
304
305         if ((busy == 0) && !list_empty(&mv_chan->chain)) {
306                 if (current_cleaned) {
307                         /*
308                          * current descriptor cleaned and removed, run
309                          * from list head
310                          */
311                         iter = list_entry(mv_chan->chain.next,
312                                           struct mv_xor_desc_slot,
313                                           node);
314                         mv_chan_start_new_chain(mv_chan, iter);
315                 } else {
316                         if (!list_is_last(&iter->node, &mv_chan->chain)) {
317                                 /*
318                                  * descriptors are still waiting after
319                                  * current, trigger them
320                                  */
321                                 iter = list_entry(iter->node.next,
322                                                   struct mv_xor_desc_slot,
323                                                   node);
324                                 mv_chan_start_new_chain(mv_chan, iter);
325                         } else {
326                                 /*
327                                  * some descriptors are still waiting
328                                  * to be cleaned
329                                  */
330                                 tasklet_schedule(&mv_chan->irq_tasklet);
331                         }
332                 }
333         }
334
335         if (cookie > 0)
336                 mv_chan->dmachan.completed_cookie = cookie;
337 }
338
339 static void mv_xor_tasklet(unsigned long data)
340 {
341         struct mv_xor_chan *chan = (struct mv_xor_chan *) data;
342
343         spin_lock(&chan->lock);
344         mv_chan_slot_cleanup(chan);
345         spin_unlock(&chan->lock);
346 }
347
348 static struct mv_xor_desc_slot *
349 mv_chan_alloc_slot(struct mv_xor_chan *mv_chan)
350 {
351         struct mv_xor_desc_slot *iter;
352
353         spin_lock_bh(&mv_chan->lock);
354
355         if (!list_empty(&mv_chan->free_slots)) {
356                 iter = list_first_entry(&mv_chan->free_slots,
357                                         struct mv_xor_desc_slot,
358                                         node);
359
360                 list_move_tail(&iter->node, &mv_chan->allocated_slots);
361
362                 spin_unlock_bh(&mv_chan->lock);
363
364                 /* pre-ack descriptor */
365                 async_tx_ack(&iter->async_tx);
366                 iter->async_tx.cookie = -EBUSY;
367
368                 return iter;
369
370         }
371
372         spin_unlock_bh(&mv_chan->lock);
373
374         /* try to free some slots if the allocation fails */
375         tasklet_schedule(&mv_chan->irq_tasklet);
376
377         return NULL;
378 }
379
380 /************************ DMA engine API functions ****************************/
381 static dma_cookie_t
382 mv_xor_tx_submit(struct dma_async_tx_descriptor *tx)
383 {
384         struct mv_xor_desc_slot *sw_desc = to_mv_xor_slot(tx);
385         struct mv_xor_chan *mv_chan = to_mv_xor_chan(tx->chan);
386         struct mv_xor_desc_slot *old_chain_tail;
387         dma_cookie_t cookie;
388         int new_hw_chain = 1;
389
390         dev_dbg(mv_chan_to_devp(mv_chan),
391                 "%s sw_desc %p: async_tx %p\n",
392                 __func__, sw_desc, &sw_desc->async_tx);
393
394         spin_lock_bh(&mv_chan->lock);
395         cookie = dma_cookie_assign(tx);
396
397         if (list_empty(&mv_chan->chain))
398                 list_move_tail(&sw_desc->node, &mv_chan->chain);
399         else {
400                 new_hw_chain = 0;
401
402                 old_chain_tail = list_entry(mv_chan->chain.prev,
403                                             struct mv_xor_desc_slot,
404                                             node);
405                 list_move_tail(&sw_desc->node, &mv_chan->chain);
406
407                 dev_dbg(mv_chan_to_devp(mv_chan), "Append to last desc %pa\n",
408                         &old_chain_tail->async_tx.phys);
409
410                 /* fix up the hardware chain */
411                 mv_desc_set_next_desc(old_chain_tail, sw_desc->async_tx.phys);
412
413                 /* if the channel is not busy */
414                 if (!mv_chan_is_busy(mv_chan)) {
415                         u32 current_desc = mv_chan_get_current_desc(mv_chan);
416                         /*
417                          * and the curren desc is the end of the chain before
418                          * the append, then we need to start the channel
419                          */
420                         if (current_desc == old_chain_tail->async_tx.phys)
421                                 new_hw_chain = 1;
422                 }
423         }
424
425         if (new_hw_chain)
426                 mv_chan_start_new_chain(mv_chan, sw_desc);
427
428         spin_unlock_bh(&mv_chan->lock);
429
430         return cookie;
431 }
432
433 /* returns the number of allocated descriptors */
434 static int mv_xor_alloc_chan_resources(struct dma_chan *chan)
435 {
436         void *virt_desc;
437         dma_addr_t dma_desc;
438         int idx;
439         struct mv_xor_chan *mv_chan = to_mv_xor_chan(chan);
440         struct mv_xor_desc_slot *slot = NULL;
441         int num_descs_in_pool = MV_XOR_POOL_SIZE/MV_XOR_SLOT_SIZE;
442
443         /* Allocate descriptor slots */
444         idx = mv_chan->slots_allocated;
445         while (idx < num_descs_in_pool) {
446                 slot = kzalloc(sizeof(*slot), GFP_KERNEL);
447                 if (!slot) {
448                         dev_info(mv_chan_to_devp(mv_chan),
449                                  "channel only initialized %d descriptor slots",
450                                  idx);
451                         break;
452                 }
453                 virt_desc = mv_chan->dma_desc_pool_virt;
454                 slot->hw_desc = virt_desc + idx * MV_XOR_SLOT_SIZE;
455
456                 dma_async_tx_descriptor_init(&slot->async_tx, chan);
457                 slot->async_tx.tx_submit = mv_xor_tx_submit;
458                 INIT_LIST_HEAD(&slot->node);
459                 INIT_LIST_HEAD(&slot->sg_tx_list);
460                 dma_desc = mv_chan->dma_desc_pool;
461                 slot->async_tx.phys = dma_desc + idx * MV_XOR_SLOT_SIZE;
462                 slot->idx = idx++;
463
464                 spin_lock_bh(&mv_chan->lock);
465                 mv_chan->slots_allocated = idx;
466                 list_add_tail(&slot->node, &mv_chan->free_slots);
467                 spin_unlock_bh(&mv_chan->lock);
468         }
469
470         dev_dbg(mv_chan_to_devp(mv_chan),
471                 "allocated %d descriptor slots\n",
472                 mv_chan->slots_allocated);
473
474         return mv_chan->slots_allocated ? : -ENOMEM;
475 }
476
477 /*
478  * Check if source or destination is an PCIe/IO address (non-SDRAM) and add
479  * a new MBus window if necessary. Use a cache for these check so that
480  * the MMIO mapped registers don't have to be accessed for this check
481  * to speed up this process.
482  */
483 static int mv_xor_add_io_win(struct mv_xor_chan *mv_chan, u32 addr)
484 {
485         struct mv_xor_device *xordev = mv_chan->xordev;
486         void __iomem *base = mv_chan->mmr_high_base;
487         u32 win_enable;
488         u32 size;
489         u8 target, attr;
490         int ret;
491         int i;
492
493         /* Nothing needs to get done for the Armada 3700 */
494         if (xordev->xor_type == XOR_ARMADA_37XX)
495                 return 0;
496
497         /*
498          * Loop over the cached windows to check, if the requested area
499          * is already mapped. If this the case, nothing needs to be done
500          * and we can return.
501          */
502         for (i = 0; i < WINDOW_COUNT; i++) {
503                 if (addr >= xordev->win_start[i] &&
504                     addr <= xordev->win_end[i]) {
505                         /* Window is already mapped */
506                         return 0;
507                 }
508         }
509
510         /*
511          * The window is not mapped, so we need to create the new mapping
512          */
513
514         /* If no IO window is found that addr has to be located in SDRAM */
515         ret = mvebu_mbus_get_io_win_info(addr, &size, &target, &attr);
516         if (ret < 0)
517                 return 0;
518
519         /*
520          * Mask the base addr 'addr' according to 'size' read back from the
521          * MBus window. Otherwise we might end up with an address located
522          * somewhere in the middle of this area here.
523          */
524         size -= 1;
525         addr &= ~size;
526
527         /*
528          * Reading one of both enabled register is enough, as they are always
529          * programmed to the identical values
530          */
531         win_enable = readl(base + WINDOW_BAR_ENABLE(0));
532
533         /* Set 'i' to the first free window to write the new values to */
534         i = ffs(~win_enable) - 1;
535         if (i >= WINDOW_COUNT)
536                 return -ENOMEM;
537
538         writel((addr & 0xffff0000) | (attr << 8) | target,
539                base + WINDOW_BASE(i));
540         writel(size & 0xffff0000, base + WINDOW_SIZE(i));
541
542         /* Fill the caching variables for later use */
543         xordev->win_start[i] = addr;
544         xordev->win_end[i] = addr + size;
545
546         win_enable |= (1 << i);
547         win_enable |= 3 << (16 + (2 * i));
548         writel(win_enable, base + WINDOW_BAR_ENABLE(0));
549         writel(win_enable, base + WINDOW_BAR_ENABLE(1));
550
551         return 0;
552 }
553
554 static struct dma_async_tx_descriptor *
555 mv_xor_prep_dma_xor(struct dma_chan *chan, dma_addr_t dest, dma_addr_t *src,
556                     unsigned int src_cnt, size_t len, unsigned long flags)
557 {
558         struct mv_xor_chan *mv_chan = to_mv_xor_chan(chan);
559         struct mv_xor_desc_slot *sw_desc;
560         int ret;
561
562         if (unlikely(len < MV_XOR_MIN_BYTE_COUNT))
563                 return NULL;
564
565         BUG_ON(len > MV_XOR_MAX_BYTE_COUNT);
566
567         dev_dbg(mv_chan_to_devp(mv_chan),
568                 "%s src_cnt: %d len: %zu dest %pad flags: %ld\n",
569                 __func__, src_cnt, len, &dest, flags);
570
571         /* Check if a new window needs to get added for 'dest' */
572         ret = mv_xor_add_io_win(mv_chan, dest);
573         if (ret)
574                 return NULL;
575
576         sw_desc = mv_chan_alloc_slot(mv_chan);
577         if (sw_desc) {
578                 sw_desc->type = DMA_XOR;
579                 sw_desc->async_tx.flags = flags;
580                 mv_desc_init(sw_desc, dest, len, flags);
581                 if (mv_chan->op_in_desc == XOR_MODE_IN_DESC)
582                         mv_desc_set_mode(sw_desc);
583                 while (src_cnt--) {
584                         /* Check if a new window needs to get added for 'src' */
585                         ret = mv_xor_add_io_win(mv_chan, src[src_cnt]);
586                         if (ret)
587                                 return NULL;
588                         mv_desc_set_src_addr(sw_desc, src_cnt, src[src_cnt]);
589                 }
590         }
591
592         dev_dbg(mv_chan_to_devp(mv_chan),
593                 "%s sw_desc %p async_tx %p \n",
594                 __func__, sw_desc, &sw_desc->async_tx);
595         return sw_desc ? &sw_desc->async_tx : NULL;
596 }
597
598 static struct dma_async_tx_descriptor *
599 mv_xor_prep_dma_memcpy(struct dma_chan *chan, dma_addr_t dest, dma_addr_t src,
600                 size_t len, unsigned long flags)
601 {
602         /*
603          * A MEMCPY operation is identical to an XOR operation with only
604          * a single source address.
605          */
606         return mv_xor_prep_dma_xor(chan, dest, &src, 1, len, flags);
607 }
608
609 static struct dma_async_tx_descriptor *
610 mv_xor_prep_dma_interrupt(struct dma_chan *chan, unsigned long flags)
611 {
612         struct mv_xor_chan *mv_chan = to_mv_xor_chan(chan);
613         dma_addr_t src, dest;
614         size_t len;
615
616         src = mv_chan->dummy_src_addr;
617         dest = mv_chan->dummy_dst_addr;
618         len = MV_XOR_MIN_BYTE_COUNT;
619
620         /*
621          * We implement the DMA_INTERRUPT operation as a minimum sized
622          * XOR operation with a single dummy source address.
623          */
624         return mv_xor_prep_dma_xor(chan, dest, &src, 1, len, flags);
625 }
626
627 static void mv_xor_free_chan_resources(struct dma_chan *chan)
628 {
629         struct mv_xor_chan *mv_chan = to_mv_xor_chan(chan);
630         struct mv_xor_desc_slot *iter, *_iter;
631         int in_use_descs = 0;
632
633         spin_lock_bh(&mv_chan->lock);
634
635         mv_chan_slot_cleanup(mv_chan);
636
637         list_for_each_entry_safe(iter, _iter, &mv_chan->chain,
638                                         node) {
639                 in_use_descs++;
640                 list_move_tail(&iter->node, &mv_chan->free_slots);
641         }
642         list_for_each_entry_safe(iter, _iter, &mv_chan->completed_slots,
643                                  node) {
644                 in_use_descs++;
645                 list_move_tail(&iter->node, &mv_chan->free_slots);
646         }
647         list_for_each_entry_safe(iter, _iter, &mv_chan->allocated_slots,
648                                  node) {
649                 in_use_descs++;
650                 list_move_tail(&iter->node, &mv_chan->free_slots);
651         }
652         list_for_each_entry_safe_reverse(
653                 iter, _iter, &mv_chan->free_slots, node) {
654                 list_del(&iter->node);
655                 kfree(iter);
656                 mv_chan->slots_allocated--;
657         }
658
659         dev_dbg(mv_chan_to_devp(mv_chan), "%s slots_allocated %d\n",
660                 __func__, mv_chan->slots_allocated);
661         spin_unlock_bh(&mv_chan->lock);
662
663         if (in_use_descs)
664                 dev_err(mv_chan_to_devp(mv_chan),
665                         "freeing %d in use descriptors!\n", in_use_descs);
666 }
667
668 /**
669  * mv_xor_status - poll the status of an XOR transaction
670  * @chan: XOR channel handle
671  * @cookie: XOR transaction identifier
672  * @txstate: XOR transactions state holder (or NULL)
673  */
674 static enum dma_status mv_xor_status(struct dma_chan *chan,
675                                           dma_cookie_t cookie,
676                                           struct dma_tx_state *txstate)
677 {
678         struct mv_xor_chan *mv_chan = to_mv_xor_chan(chan);
679         enum dma_status ret;
680
681         ret = dma_cookie_status(chan, cookie, txstate);
682         if (ret == DMA_COMPLETE)
683                 return ret;
684
685         spin_lock_bh(&mv_chan->lock);
686         mv_chan_slot_cleanup(mv_chan);
687         spin_unlock_bh(&mv_chan->lock);
688
689         return dma_cookie_status(chan, cookie, txstate);
690 }
691
692 static void mv_chan_dump_regs(struct mv_xor_chan *chan)
693 {
694         u32 val;
695
696         val = readl_relaxed(XOR_CONFIG(chan));
697         dev_err(mv_chan_to_devp(chan), "config       0x%08x\n", val);
698
699         val = readl_relaxed(XOR_ACTIVATION(chan));
700         dev_err(mv_chan_to_devp(chan), "activation   0x%08x\n", val);
701
702         val = readl_relaxed(XOR_INTR_CAUSE(chan));
703         dev_err(mv_chan_to_devp(chan), "intr cause   0x%08x\n", val);
704
705         val = readl_relaxed(XOR_INTR_MASK(chan));
706         dev_err(mv_chan_to_devp(chan), "intr mask    0x%08x\n", val);
707
708         val = readl_relaxed(XOR_ERROR_CAUSE(chan));
709         dev_err(mv_chan_to_devp(chan), "error cause  0x%08x\n", val);
710
711         val = readl_relaxed(XOR_ERROR_ADDR(chan));
712         dev_err(mv_chan_to_devp(chan), "error addr   0x%08x\n", val);
713 }
714
715 static void mv_chan_err_interrupt_handler(struct mv_xor_chan *chan,
716                                           u32 intr_cause)
717 {
718         if (intr_cause & XOR_INT_ERR_DECODE) {
719                 dev_dbg(mv_chan_to_devp(chan), "ignoring address decode error\n");
720                 return;
721         }
722
723         dev_err(mv_chan_to_devp(chan), "error on chan %d. intr cause 0x%08x\n",
724                 chan->idx, intr_cause);
725
726         mv_chan_dump_regs(chan);
727         WARN_ON(1);
728 }
729
730 static irqreturn_t mv_xor_interrupt_handler(int irq, void *data)
731 {
732         struct mv_xor_chan *chan = data;
733         u32 intr_cause = mv_chan_get_intr_cause(chan);
734
735         dev_dbg(mv_chan_to_devp(chan), "intr cause %x\n", intr_cause);
736
737         if (intr_cause & XOR_INTR_ERRORS)
738                 mv_chan_err_interrupt_handler(chan, intr_cause);
739
740         tasklet_schedule(&chan->irq_tasklet);
741
742         mv_chan_clear_eoc_cause(chan);
743
744         return IRQ_HANDLED;
745 }
746
747 static void mv_xor_issue_pending(struct dma_chan *chan)
748 {
749         struct mv_xor_chan *mv_chan = to_mv_xor_chan(chan);
750
751         if (mv_chan->pending >= MV_XOR_THRESHOLD) {
752                 mv_chan->pending = 0;
753                 mv_chan_activate(mv_chan);
754         }
755 }
756
757 /*
758  * Perform a transaction to verify the HW works.
759  */
760
761 static int mv_chan_memcpy_self_test(struct mv_xor_chan *mv_chan)
762 {
763         int i, ret;
764         void *src, *dest;
765         dma_addr_t src_dma, dest_dma;
766         struct dma_chan *dma_chan;
767         dma_cookie_t cookie;
768         struct dma_async_tx_descriptor *tx;
769         struct dmaengine_unmap_data *unmap;
770         int err = 0;
771
772         src = kmalloc(PAGE_SIZE, GFP_KERNEL);
773         if (!src)
774                 return -ENOMEM;
775
776         dest = kzalloc(PAGE_SIZE, GFP_KERNEL);
777         if (!dest) {
778                 kfree(src);
779                 return -ENOMEM;
780         }
781
782         /* Fill in src buffer */
783         for (i = 0; i < PAGE_SIZE; i++)
784                 ((u8 *) src)[i] = (u8)i;
785
786         dma_chan = &mv_chan->dmachan;
787         if (mv_xor_alloc_chan_resources(dma_chan) < 1) {
788                 err = -ENODEV;
789                 goto out;
790         }
791
792         unmap = dmaengine_get_unmap_data(dma_chan->device->dev, 2, GFP_KERNEL);
793         if (!unmap) {
794                 err = -ENOMEM;
795                 goto free_resources;
796         }
797
798         src_dma = dma_map_page(dma_chan->device->dev, virt_to_page(src),
799                                offset_in_page(src), PAGE_SIZE,
800                                DMA_TO_DEVICE);
801         unmap->addr[0] = src_dma;
802
803         ret = dma_mapping_error(dma_chan->device->dev, src_dma);
804         if (ret) {
805                 err = -ENOMEM;
806                 goto free_resources;
807         }
808         unmap->to_cnt = 1;
809
810         dest_dma = dma_map_page(dma_chan->device->dev, virt_to_page(dest),
811                                 offset_in_page(dest), PAGE_SIZE,
812                                 DMA_FROM_DEVICE);
813         unmap->addr[1] = dest_dma;
814
815         ret = dma_mapping_error(dma_chan->device->dev, dest_dma);
816         if (ret) {
817                 err = -ENOMEM;
818                 goto free_resources;
819         }
820         unmap->from_cnt = 1;
821         unmap->len = PAGE_SIZE;
822
823         tx = mv_xor_prep_dma_memcpy(dma_chan, dest_dma, src_dma,
824                                     PAGE_SIZE, 0);
825         if (!tx) {
826                 dev_err(dma_chan->device->dev,
827                         "Self-test cannot prepare operation, disabling\n");
828                 err = -ENODEV;
829                 goto free_resources;
830         }
831
832         cookie = mv_xor_tx_submit(tx);
833         if (dma_submit_error(cookie)) {
834                 dev_err(dma_chan->device->dev,
835                         "Self-test submit error, disabling\n");
836                 err = -ENODEV;
837                 goto free_resources;
838         }
839
840         mv_xor_issue_pending(dma_chan);
841         async_tx_ack(tx);
842         msleep(1);
843
844         if (mv_xor_status(dma_chan, cookie, NULL) !=
845             DMA_COMPLETE) {
846                 dev_err(dma_chan->device->dev,
847                         "Self-test copy timed out, disabling\n");
848                 err = -ENODEV;
849                 goto free_resources;
850         }
851
852         dma_sync_single_for_cpu(dma_chan->device->dev, dest_dma,
853                                 PAGE_SIZE, DMA_FROM_DEVICE);
854         if (memcmp(src, dest, PAGE_SIZE)) {
855                 dev_err(dma_chan->device->dev,
856                         "Self-test copy failed compare, disabling\n");
857                 err = -ENODEV;
858                 goto free_resources;
859         }
860
861 free_resources:
862         dmaengine_unmap_put(unmap);
863         mv_xor_free_chan_resources(dma_chan);
864 out:
865         kfree(src);
866         kfree(dest);
867         return err;
868 }
869
870 #define MV_XOR_NUM_SRC_TEST 4 /* must be <= 15 */
871 static int
872 mv_chan_xor_self_test(struct mv_xor_chan *mv_chan)
873 {
874         int i, src_idx, ret;
875         struct page *dest;
876         struct page *xor_srcs[MV_XOR_NUM_SRC_TEST];
877         dma_addr_t dma_srcs[MV_XOR_NUM_SRC_TEST];
878         dma_addr_t dest_dma;
879         struct dma_async_tx_descriptor *tx;
880         struct dmaengine_unmap_data *unmap;
881         struct dma_chan *dma_chan;
882         dma_cookie_t cookie;
883         u8 cmp_byte = 0;
884         u32 cmp_word;
885         int err = 0;
886         int src_count = MV_XOR_NUM_SRC_TEST;
887
888         for (src_idx = 0; src_idx < src_count; src_idx++) {
889                 xor_srcs[src_idx] = alloc_page(GFP_KERNEL);
890                 if (!xor_srcs[src_idx]) {
891                         while (src_idx--)
892                                 __free_page(xor_srcs[src_idx]);
893                         return -ENOMEM;
894                 }
895         }
896
897         dest = alloc_page(GFP_KERNEL);
898         if (!dest) {
899                 while (src_idx--)
900                         __free_page(xor_srcs[src_idx]);
901                 return -ENOMEM;
902         }
903
904         /* Fill in src buffers */
905         for (src_idx = 0; src_idx < src_count; src_idx++) {
906                 u8 *ptr = page_address(xor_srcs[src_idx]);
907                 for (i = 0; i < PAGE_SIZE; i++)
908                         ptr[i] = (1 << src_idx);
909         }
910
911         for (src_idx = 0; src_idx < src_count; src_idx++)
912                 cmp_byte ^= (u8) (1 << src_idx);
913
914         cmp_word = (cmp_byte << 24) | (cmp_byte << 16) |
915                 (cmp_byte << 8) | cmp_byte;
916
917         memset(page_address(dest), 0, PAGE_SIZE);
918
919         dma_chan = &mv_chan->dmachan;
920         if (mv_xor_alloc_chan_resources(dma_chan) < 1) {
921                 err = -ENODEV;
922                 goto out;
923         }
924
925         unmap = dmaengine_get_unmap_data(dma_chan->device->dev, src_count + 1,
926                                          GFP_KERNEL);
927         if (!unmap) {
928                 err = -ENOMEM;
929                 goto free_resources;
930         }
931
932         /* test xor */
933         for (i = 0; i < src_count; i++) {
934                 unmap->addr[i] = dma_map_page(dma_chan->device->dev, xor_srcs[i],
935                                               0, PAGE_SIZE, DMA_TO_DEVICE);
936                 dma_srcs[i] = unmap->addr[i];
937                 ret = dma_mapping_error(dma_chan->device->dev, unmap->addr[i]);
938                 if (ret) {
939                         err = -ENOMEM;
940                         goto free_resources;
941                 }
942                 unmap->to_cnt++;
943         }
944
945         unmap->addr[src_count] = dma_map_page(dma_chan->device->dev, dest, 0, PAGE_SIZE,
946                                       DMA_FROM_DEVICE);
947         dest_dma = unmap->addr[src_count];
948         ret = dma_mapping_error(dma_chan->device->dev, unmap->addr[src_count]);
949         if (ret) {
950                 err = -ENOMEM;
951                 goto free_resources;
952         }
953         unmap->from_cnt = 1;
954         unmap->len = PAGE_SIZE;
955
956         tx = mv_xor_prep_dma_xor(dma_chan, dest_dma, dma_srcs,
957                                  src_count, PAGE_SIZE, 0);
958         if (!tx) {
959                 dev_err(dma_chan->device->dev,
960                         "Self-test cannot prepare operation, disabling\n");
961                 err = -ENODEV;
962                 goto free_resources;
963         }
964
965         cookie = mv_xor_tx_submit(tx);
966         if (dma_submit_error(cookie)) {
967                 dev_err(dma_chan->device->dev,
968                         "Self-test submit error, disabling\n");
969                 err = -ENODEV;
970                 goto free_resources;
971         }
972
973         mv_xor_issue_pending(dma_chan);
974         async_tx_ack(tx);
975         msleep(8);
976
977         if (mv_xor_status(dma_chan, cookie, NULL) !=
978             DMA_COMPLETE) {
979                 dev_err(dma_chan->device->dev,
980                         "Self-test xor timed out, disabling\n");
981                 err = -ENODEV;
982                 goto free_resources;
983         }
984
985         dma_sync_single_for_cpu(dma_chan->device->dev, dest_dma,
986                                 PAGE_SIZE, DMA_FROM_DEVICE);
987         for (i = 0; i < (PAGE_SIZE / sizeof(u32)); i++) {
988                 u32 *ptr = page_address(dest);
989                 if (ptr[i] != cmp_word) {
990                         dev_err(dma_chan->device->dev,
991                                 "Self-test xor failed compare, disabling. index %d, data %x, expected %x\n",
992                                 i, ptr[i], cmp_word);
993                         err = -ENODEV;
994                         goto free_resources;
995                 }
996         }
997
998 free_resources:
999         dmaengine_unmap_put(unmap);
1000         mv_xor_free_chan_resources(dma_chan);
1001 out:
1002         src_idx = src_count;
1003         while (src_idx--)
1004                 __free_page(xor_srcs[src_idx]);
1005         __free_page(dest);
1006         return err;
1007 }
1008
1009 static int mv_xor_channel_remove(struct mv_xor_chan *mv_chan)
1010 {
1011         struct dma_chan *chan, *_chan;
1012         struct device *dev = mv_chan->dmadev.dev;
1013
1014         dma_async_device_unregister(&mv_chan->dmadev);
1015
1016         dma_free_coherent(dev, MV_XOR_POOL_SIZE,
1017                           mv_chan->dma_desc_pool_virt, mv_chan->dma_desc_pool);
1018         dma_unmap_single(dev, mv_chan->dummy_src_addr,
1019                          MV_XOR_MIN_BYTE_COUNT, DMA_FROM_DEVICE);
1020         dma_unmap_single(dev, mv_chan->dummy_dst_addr,
1021                          MV_XOR_MIN_BYTE_COUNT, DMA_TO_DEVICE);
1022
1023         list_for_each_entry_safe(chan, _chan, &mv_chan->dmadev.channels,
1024                                  device_node) {
1025                 list_del(&chan->device_node);
1026         }
1027
1028         free_irq(mv_chan->irq, mv_chan);
1029
1030         return 0;
1031 }
1032
1033 static struct mv_xor_chan *
1034 mv_xor_channel_add(struct mv_xor_device *xordev,
1035                    struct platform_device *pdev,
1036                    int idx, dma_cap_mask_t cap_mask, int irq)
1037 {
1038         int ret = 0;
1039         struct mv_xor_chan *mv_chan;
1040         struct dma_device *dma_dev;
1041
1042         mv_chan = devm_kzalloc(&pdev->dev, sizeof(*mv_chan), GFP_KERNEL);
1043         if (!mv_chan)
1044                 return ERR_PTR(-ENOMEM);
1045
1046         mv_chan->idx = idx;
1047         mv_chan->irq = irq;
1048         if (xordev->xor_type == XOR_ORION)
1049                 mv_chan->op_in_desc = XOR_MODE_IN_REG;
1050         else
1051                 mv_chan->op_in_desc = XOR_MODE_IN_DESC;
1052
1053         dma_dev = &mv_chan->dmadev;
1054         dma_dev->dev = &pdev->dev;
1055         mv_chan->xordev = xordev;
1056
1057         /*
1058          * These source and destination dummy buffers are used to implement
1059          * a DMA_INTERRUPT operation as a minimum-sized XOR operation.
1060          * Hence, we only need to map the buffers at initialization-time.
1061          */
1062         mv_chan->dummy_src_addr = dma_map_single(dma_dev->dev,
1063                 mv_chan->dummy_src, MV_XOR_MIN_BYTE_COUNT, DMA_FROM_DEVICE);
1064         mv_chan->dummy_dst_addr = dma_map_single(dma_dev->dev,
1065                 mv_chan->dummy_dst, MV_XOR_MIN_BYTE_COUNT, DMA_TO_DEVICE);
1066
1067         /* allocate coherent memory for hardware descriptors
1068          * note: writecombine gives slightly better performance, but
1069          * requires that we explicitly flush the writes
1070          */
1071         mv_chan->dma_desc_pool_virt =
1072           dma_alloc_wc(&pdev->dev, MV_XOR_POOL_SIZE, &mv_chan->dma_desc_pool,
1073                        GFP_KERNEL);
1074         if (!mv_chan->dma_desc_pool_virt)
1075                 return ERR_PTR(-ENOMEM);
1076
1077         /* discover transaction capabilites from the platform data */
1078         dma_dev->cap_mask = cap_mask;
1079
1080         INIT_LIST_HEAD(&dma_dev->channels);
1081
1082         /* set base routines */
1083         dma_dev->device_alloc_chan_resources = mv_xor_alloc_chan_resources;
1084         dma_dev->device_free_chan_resources = mv_xor_free_chan_resources;
1085         dma_dev->device_tx_status = mv_xor_status;
1086         dma_dev->device_issue_pending = mv_xor_issue_pending;
1087
1088         /* set prep routines based on capability */
1089         if (dma_has_cap(DMA_INTERRUPT, dma_dev->cap_mask))
1090                 dma_dev->device_prep_dma_interrupt = mv_xor_prep_dma_interrupt;
1091         if (dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask))
1092                 dma_dev->device_prep_dma_memcpy = mv_xor_prep_dma_memcpy;
1093         if (dma_has_cap(DMA_XOR, dma_dev->cap_mask)) {
1094                 dma_dev->max_xor = 8;
1095                 dma_dev->device_prep_dma_xor = mv_xor_prep_dma_xor;
1096         }
1097
1098         mv_chan->mmr_base = xordev->xor_base;
1099         mv_chan->mmr_high_base = xordev->xor_high_base;
1100         tasklet_init(&mv_chan->irq_tasklet, mv_xor_tasklet, (unsigned long)
1101                      mv_chan);
1102
1103         /* clear errors before enabling interrupts */
1104         mv_chan_clear_err_status(mv_chan);
1105
1106         ret = request_irq(mv_chan->irq, mv_xor_interrupt_handler,
1107                           0, dev_name(&pdev->dev), mv_chan);
1108         if (ret)
1109                 goto err_free_dma;
1110
1111         mv_chan_unmask_interrupts(mv_chan);
1112
1113         if (mv_chan->op_in_desc == XOR_MODE_IN_DESC)
1114                 mv_chan_set_mode(mv_chan, XOR_OPERATION_MODE_IN_DESC);
1115         else
1116                 mv_chan_set_mode(mv_chan, XOR_OPERATION_MODE_XOR);
1117
1118         spin_lock_init(&mv_chan->lock);
1119         INIT_LIST_HEAD(&mv_chan->chain);
1120         INIT_LIST_HEAD(&mv_chan->completed_slots);
1121         INIT_LIST_HEAD(&mv_chan->free_slots);
1122         INIT_LIST_HEAD(&mv_chan->allocated_slots);
1123         mv_chan->dmachan.device = dma_dev;
1124         dma_cookie_init(&mv_chan->dmachan);
1125
1126         list_add_tail(&mv_chan->dmachan.device_node, &dma_dev->channels);
1127
1128         if (dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask)) {
1129                 ret = mv_chan_memcpy_self_test(mv_chan);
1130                 dev_dbg(&pdev->dev, "memcpy self test returned %d\n", ret);
1131                 if (ret)
1132                         goto err_free_irq;
1133         }
1134
1135         if (dma_has_cap(DMA_XOR, dma_dev->cap_mask)) {
1136                 ret = mv_chan_xor_self_test(mv_chan);
1137                 dev_dbg(&pdev->dev, "xor self test returned %d\n", ret);
1138                 if (ret)
1139                         goto err_free_irq;
1140         }
1141
1142         dev_info(&pdev->dev, "Marvell XOR (%s): ( %s%s%s)\n",
1143                  mv_chan->op_in_desc ? "Descriptor Mode" : "Registers Mode",
1144                  dma_has_cap(DMA_XOR, dma_dev->cap_mask) ? "xor " : "",
1145                  dma_has_cap(DMA_MEMCPY, dma_dev->cap_mask) ? "cpy " : "",
1146                  dma_has_cap(DMA_INTERRUPT, dma_dev->cap_mask) ? "intr " : "");
1147
1148         ret = dma_async_device_register(dma_dev);
1149         if (ret)
1150                 goto err_free_irq;
1151
1152         return mv_chan;
1153
1154 err_free_irq:
1155         free_irq(mv_chan->irq, mv_chan);
1156 err_free_dma:
1157         dma_free_coherent(&pdev->dev, MV_XOR_POOL_SIZE,
1158                           mv_chan->dma_desc_pool_virt, mv_chan->dma_desc_pool);
1159         return ERR_PTR(ret);
1160 }
1161
1162 static void
1163 mv_xor_conf_mbus_windows(struct mv_xor_device *xordev,
1164                          const struct mbus_dram_target_info *dram)
1165 {
1166         void __iomem *base = xordev->xor_high_base;
1167         u32 win_enable = 0;
1168         int i;
1169
1170         for (i = 0; i < 8; i++) {
1171                 writel(0, base + WINDOW_BASE(i));
1172                 writel(0, base + WINDOW_SIZE(i));
1173                 if (i < 4)
1174                         writel(0, base + WINDOW_REMAP_HIGH(i));
1175         }
1176
1177         for (i = 0; i < dram->num_cs; i++) {
1178                 const struct mbus_dram_window *cs = dram->cs + i;
1179
1180                 writel((cs->base & 0xffff0000) |
1181                        (cs->mbus_attr << 8) |
1182                        dram->mbus_dram_target_id, base + WINDOW_BASE(i));
1183                 writel((cs->size - 1) & 0xffff0000, base + WINDOW_SIZE(i));
1184
1185                 /* Fill the caching variables for later use */
1186                 xordev->win_start[i] = cs->base;
1187                 xordev->win_end[i] = cs->base + cs->size - 1;
1188
1189                 win_enable |= (1 << i);
1190                 win_enable |= 3 << (16 + (2 * i));
1191         }
1192
1193         writel(win_enable, base + WINDOW_BAR_ENABLE(0));
1194         writel(win_enable, base + WINDOW_BAR_ENABLE(1));
1195         writel(0, base + WINDOW_OVERRIDE_CTRL(0));
1196         writel(0, base + WINDOW_OVERRIDE_CTRL(1));
1197 }
1198
1199 static void
1200 mv_xor_conf_mbus_windows_a3700(struct mv_xor_device *xordev)
1201 {
1202         void __iomem *base = xordev->xor_high_base;
1203         u32 win_enable = 0;
1204         int i;
1205
1206         for (i = 0; i < 8; i++) {
1207                 writel(0, base + WINDOW_BASE(i));
1208                 writel(0, base + WINDOW_SIZE(i));
1209                 if (i < 4)
1210                         writel(0, base + WINDOW_REMAP_HIGH(i));
1211         }
1212         /*
1213          * For Armada3700 open default 4GB Mbus window. The dram
1214          * related configuration are done at AXIS level.
1215          */
1216         writel(0xffff0000, base + WINDOW_SIZE(0));
1217         win_enable |= 1;
1218         win_enable |= 3 << 16;
1219
1220         writel(win_enable, base + WINDOW_BAR_ENABLE(0));
1221         writel(win_enable, base + WINDOW_BAR_ENABLE(1));
1222         writel(0, base + WINDOW_OVERRIDE_CTRL(0));
1223         writel(0, base + WINDOW_OVERRIDE_CTRL(1));
1224 }
1225
1226 /*
1227  * Since this XOR driver is basically used only for RAID5, we don't
1228  * need to care about synchronizing ->suspend with DMA activity,
1229  * because the DMA engine will naturally be quiet due to the block
1230  * devices being suspended.
1231  */
1232 static int mv_xor_suspend(struct platform_device *pdev, pm_message_t state)
1233 {
1234         struct mv_xor_device *xordev = platform_get_drvdata(pdev);
1235         int i;
1236
1237         for (i = 0; i < MV_XOR_MAX_CHANNELS; i++) {
1238                 struct mv_xor_chan *mv_chan = xordev->channels[i];
1239
1240                 if (!mv_chan)
1241                         continue;
1242
1243                 mv_chan->saved_config_reg =
1244                         readl_relaxed(XOR_CONFIG(mv_chan));
1245                 mv_chan->saved_int_mask_reg =
1246                         readl_relaxed(XOR_INTR_MASK(mv_chan));
1247         }
1248
1249         return 0;
1250 }
1251
1252 static int mv_xor_resume(struct platform_device *dev)
1253 {
1254         struct mv_xor_device *xordev = platform_get_drvdata(dev);
1255         const struct mbus_dram_target_info *dram;
1256         int i;
1257
1258         for (i = 0; i < MV_XOR_MAX_CHANNELS; i++) {
1259                 struct mv_xor_chan *mv_chan = xordev->channels[i];
1260
1261                 if (!mv_chan)
1262                         continue;
1263
1264                 writel_relaxed(mv_chan->saved_config_reg,
1265                                XOR_CONFIG(mv_chan));
1266                 writel_relaxed(mv_chan->saved_int_mask_reg,
1267                                XOR_INTR_MASK(mv_chan));
1268         }
1269
1270         if (xordev->xor_type == XOR_ARMADA_37XX) {
1271                 mv_xor_conf_mbus_windows_a3700(xordev);
1272                 return 0;
1273         }
1274
1275         dram = mv_mbus_dram_info();
1276         if (dram)
1277                 mv_xor_conf_mbus_windows(xordev, dram);
1278
1279         return 0;
1280 }
1281
1282 static const struct of_device_id mv_xor_dt_ids[] = {
1283         { .compatible = "marvell,orion-xor", .data = (void *)XOR_ORION },
1284         { .compatible = "marvell,armada-380-xor", .data = (void *)XOR_ARMADA_38X },
1285         { .compatible = "marvell,armada-3700-xor", .data = (void *)XOR_ARMADA_37XX },
1286         {},
1287 };
1288
1289 static unsigned int mv_xor_engine_count;
1290
1291 static int mv_xor_probe(struct platform_device *pdev)
1292 {
1293         const struct mbus_dram_target_info *dram;
1294         struct mv_xor_device *xordev;
1295         struct mv_xor_platform_data *pdata = dev_get_platdata(&pdev->dev);
1296         struct resource *res;
1297         unsigned int max_engines, max_channels;
1298         int i, ret;
1299
1300         dev_notice(&pdev->dev, "Marvell shared XOR driver\n");
1301
1302         xordev = devm_kzalloc(&pdev->dev, sizeof(*xordev), GFP_KERNEL);
1303         if (!xordev)
1304                 return -ENOMEM;
1305
1306         res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
1307         if (!res)
1308                 return -ENODEV;
1309
1310         xordev->xor_base = devm_ioremap(&pdev->dev, res->start,
1311                                         resource_size(res));
1312         if (!xordev->xor_base)
1313                 return -EBUSY;
1314
1315         res = platform_get_resource(pdev, IORESOURCE_MEM, 1);
1316         if (!res)
1317                 return -ENODEV;
1318
1319         xordev->xor_high_base = devm_ioremap(&pdev->dev, res->start,
1320                                              resource_size(res));
1321         if (!xordev->xor_high_base)
1322                 return -EBUSY;
1323
1324         platform_set_drvdata(pdev, xordev);
1325
1326
1327         /*
1328          * We need to know which type of XOR device we use before
1329          * setting up. In non-dt case it can only be the legacy one.
1330          */
1331         xordev->xor_type = XOR_ORION;
1332         if (pdev->dev.of_node) {
1333                 const struct of_device_id *of_id =
1334                         of_match_device(mv_xor_dt_ids,
1335                                         &pdev->dev);
1336
1337                 xordev->xor_type = (uintptr_t)of_id->data;
1338         }
1339
1340         /*
1341          * (Re-)program MBUS remapping windows if we are asked to.
1342          */
1343         if (xordev->xor_type == XOR_ARMADA_37XX) {
1344                 mv_xor_conf_mbus_windows_a3700(xordev);
1345         } else {
1346                 dram = mv_mbus_dram_info();
1347                 if (dram)
1348                         mv_xor_conf_mbus_windows(xordev, dram);
1349         }
1350
1351         /* Not all platforms can gate the clock, so it is not
1352          * an error if the clock does not exists.
1353          */
1354         xordev->clk = clk_get(&pdev->dev, NULL);
1355         if (!IS_ERR(xordev->clk))
1356                 clk_prepare_enable(xordev->clk);
1357
1358         /*
1359          * We don't want to have more than one channel per CPU in
1360          * order for async_tx to perform well. So we limit the number
1361          * of engines and channels so that we take into account this
1362          * constraint. Note that we also want to use channels from
1363          * separate engines when possible.  For dual-CPU Armada 3700
1364          * SoC with single XOR engine allow using its both channels.
1365          */
1366         max_engines = num_present_cpus();
1367         if (xordev->xor_type == XOR_ARMADA_37XX)
1368                 max_channels =  num_present_cpus();
1369         else
1370                 max_channels = min_t(unsigned int,
1371                                      MV_XOR_MAX_CHANNELS,
1372                                      DIV_ROUND_UP(num_present_cpus(), 2));
1373
1374         if (mv_xor_engine_count >= max_engines)
1375                 return 0;
1376
1377         if (pdev->dev.of_node) {
1378                 struct device_node *np;
1379                 int i = 0;
1380
1381                 for_each_child_of_node(pdev->dev.of_node, np) {
1382                         struct mv_xor_chan *chan;
1383                         dma_cap_mask_t cap_mask;
1384                         int irq;
1385
1386                         if (i >= max_channels)
1387                                 continue;
1388
1389                         dma_cap_zero(cap_mask);
1390                         dma_cap_set(DMA_MEMCPY, cap_mask);
1391                         dma_cap_set(DMA_XOR, cap_mask);
1392                         dma_cap_set(DMA_INTERRUPT, cap_mask);
1393
1394                         irq = irq_of_parse_and_map(np, 0);
1395                         if (!irq) {
1396                                 ret = -ENODEV;
1397                                 goto err_channel_add;
1398                         }
1399
1400                         chan = mv_xor_channel_add(xordev, pdev, i,
1401                                                   cap_mask, irq);
1402                         if (IS_ERR(chan)) {
1403                                 ret = PTR_ERR(chan);
1404                                 irq_dispose_mapping(irq);
1405                                 goto err_channel_add;
1406                         }
1407
1408                         xordev->channels[i] = chan;
1409                         i++;
1410                 }
1411         } else if (pdata && pdata->channels) {
1412                 for (i = 0; i < max_channels; i++) {
1413                         struct mv_xor_channel_data *cd;
1414                         struct mv_xor_chan *chan;
1415                         int irq;
1416
1417                         cd = &pdata->channels[i];
1418                         irq = platform_get_irq(pdev, i);
1419                         if (irq < 0) {
1420                                 ret = irq;
1421                                 goto err_channel_add;
1422                         }
1423
1424                         chan = mv_xor_channel_add(xordev, pdev, i,
1425                                                   cd->cap_mask, irq);
1426                         if (IS_ERR(chan)) {
1427                                 ret = PTR_ERR(chan);
1428                                 goto err_channel_add;
1429                         }
1430
1431                         xordev->channels[i] = chan;
1432                 }
1433         }
1434
1435         return 0;
1436
1437 err_channel_add:
1438         for (i = 0; i < MV_XOR_MAX_CHANNELS; i++)
1439                 if (xordev->channels[i]) {
1440                         mv_xor_channel_remove(xordev->channels[i]);
1441                         if (pdev->dev.of_node)
1442                                 irq_dispose_mapping(xordev->channels[i]->irq);
1443                 }
1444
1445         if (!IS_ERR(xordev->clk)) {
1446                 clk_disable_unprepare(xordev->clk);
1447                 clk_put(xordev->clk);
1448         }
1449
1450         return ret;
1451 }
1452
1453 static struct platform_driver mv_xor_driver = {
1454         .probe          = mv_xor_probe,
1455         .suspend        = mv_xor_suspend,
1456         .resume         = mv_xor_resume,
1457         .driver         = {
1458                 .name           = MV_XOR_NAME,
1459                 .of_match_table = of_match_ptr(mv_xor_dt_ids),
1460         },
1461 };
1462
1463 builtin_platform_driver(mv_xor_driver);
1464
1465 /*
1466 MODULE_AUTHOR("Saeed Bishara <saeed@marvell.com>");
1467 MODULE_DESCRIPTION("DMA engine driver for Marvell's XOR engine");
1468 MODULE_LICENSE("GPL");
1469 */