1 // SPDX-License-Identifier: GPL-2.0-or-later
3 * arch/powerpc/sysdev/qe_lib/qe_ic.c
5 * Copyright (C) 2006 Freescale Semiconductor, Inc. All rights reserved.
7 * Author: Li Yang <leoli@freescale.com>
8 * Based on code from Shlomi Gridish <gridish@freescale.com>
10 * QUICC ENGINE Interrupt Controller
13 #include <linux/of_irq.h>
14 #include <linux/of_address.h>
15 #include <linux/kernel.h>
16 #include <linux/init.h>
17 #include <linux/errno.h>
18 #include <linux/irq.h>
19 #include <linux/reboot.h>
20 #include <linux/slab.h>
21 #include <linux/stddef.h>
22 #include <linux/sched.h>
23 #include <linux/signal.h>
24 #include <linux/device.h>
25 #include <linux/spinlock.h>
26 #include <linux/platform_device.h>
29 #include <soc/fsl/qe/qe.h>
31 #define NR_QE_IC_INTS 64
33 /* QE IC registers offset */
34 #define QEIC_CICR 0x00
35 #define QEIC_CIVEC 0x04
36 #define QEIC_CIPXCC 0x10
37 #define QEIC_CIPYCC 0x14
38 #define QEIC_CIPWCC 0x18
39 #define QEIC_CIPZCC 0x1c
40 #define QEIC_CIMR 0x20
41 #define QEIC_CRIMR 0x24
42 #define QEIC_CIPRTA 0x30
43 #define QEIC_CIPRTB 0x34
44 #define QEIC_CHIVEC 0x60
47 /* Control registers offset */
50 /* The remapper for this QEIC */
51 struct irq_domain *irqhost;
53 /* The "linux" controller struct */
54 struct irq_chip hc_irq;
56 /* VIRQ numbers of QE high/low irqs */
62 * QE interrupt controller internal structure
65 /* Location of this source at the QIMR register */
68 /* Mask register offset */
72 * For grouped interrupts sources - the interrupt code as
73 * appears at the group priority register
77 /* Group priority register offset */
81 static DEFINE_RAW_SPINLOCK(qe_ic_lock);
83 static struct qe_ic_info qe_ic_info[] = {
86 .mask_reg = QEIC_CIMR,
88 .pri_reg = QEIC_CIPWCC,
92 .mask_reg = QEIC_CIMR,
94 .pri_reg = QEIC_CIPWCC,
98 .mask_reg = QEIC_CIMR,
100 .pri_reg = QEIC_CIPWCC,
104 .mask_reg = QEIC_CIMR,
106 .pri_reg = QEIC_CIPZCC,
110 .mask_reg = QEIC_CIMR,
112 .pri_reg = QEIC_CIPZCC,
116 .mask_reg = QEIC_CIMR,
118 .pri_reg = QEIC_CIPZCC,
122 .mask_reg = QEIC_CIMR,
124 .pri_reg = QEIC_CIPZCC,
128 .mask_reg = QEIC_CIMR,
130 .pri_reg = QEIC_CIPZCC,
134 .mask_reg = QEIC_CIMR,
136 .pri_reg = QEIC_CIPZCC,
140 .mask_reg = QEIC_CRIMR,
142 .pri_reg = QEIC_CIPRTA,
146 .mask_reg = QEIC_CRIMR,
148 .pri_reg = QEIC_CIPRTB,
152 .mask_reg = QEIC_CRIMR,
154 .pri_reg = QEIC_CIPRTB,
158 .mask_reg = QEIC_CRIMR,
160 .pri_reg = QEIC_CIPRTB,
164 .mask_reg = QEIC_CRIMR,
166 .pri_reg = QEIC_CIPRTB,
170 .mask_reg = QEIC_CIMR,
172 .pri_reg = QEIC_CIPXCC,
176 .mask_reg = QEIC_CIMR,
178 .pri_reg = QEIC_CIPXCC,
182 .mask_reg = QEIC_CIMR,
184 .pri_reg = QEIC_CIPXCC,
188 .mask_reg = QEIC_CIMR,
190 .pri_reg = QEIC_CIPXCC,
194 .mask_reg = QEIC_CIMR,
196 .pri_reg = QEIC_CIPXCC,
200 .mask_reg = QEIC_CIMR,
202 .pri_reg = QEIC_CIPYCC,
206 .mask_reg = QEIC_CIMR,
208 .pri_reg = QEIC_CIPYCC,
212 .mask_reg = QEIC_CIMR,
214 .pri_reg = QEIC_CIPYCC,
218 .mask_reg = QEIC_CIMR,
220 .pri_reg = QEIC_CIPYCC,
224 static inline u32 qe_ic_read(__be32 __iomem *base, unsigned int reg)
226 return ioread32be(base + (reg >> 2));
229 static inline void qe_ic_write(__be32 __iomem *base, unsigned int reg,
232 iowrite32be(value, base + (reg >> 2));
235 static inline struct qe_ic *qe_ic_from_irq(unsigned int virq)
237 return irq_get_chip_data(virq);
240 static inline struct qe_ic *qe_ic_from_irq_data(struct irq_data *d)
242 return irq_data_get_irq_chip_data(d);
245 static void qe_ic_unmask_irq(struct irq_data *d)
247 struct qe_ic *qe_ic = qe_ic_from_irq_data(d);
248 unsigned int src = irqd_to_hwirq(d);
252 raw_spin_lock_irqsave(&qe_ic_lock, flags);
254 temp = qe_ic_read(qe_ic->regs, qe_ic_info[src].mask_reg);
255 qe_ic_write(qe_ic->regs, qe_ic_info[src].mask_reg,
256 temp | qe_ic_info[src].mask);
258 raw_spin_unlock_irqrestore(&qe_ic_lock, flags);
261 static void qe_ic_mask_irq(struct irq_data *d)
263 struct qe_ic *qe_ic = qe_ic_from_irq_data(d);
264 unsigned int src = irqd_to_hwirq(d);
268 raw_spin_lock_irqsave(&qe_ic_lock, flags);
270 temp = qe_ic_read(qe_ic->regs, qe_ic_info[src].mask_reg);
271 qe_ic_write(qe_ic->regs, qe_ic_info[src].mask_reg,
272 temp & ~qe_ic_info[src].mask);
274 /* Flush the above write before enabling interrupts; otherwise,
275 * spurious interrupts will sometimes happen. To be 100% sure
276 * that the write has reached the device before interrupts are
277 * enabled, the mask register would have to be read back; however,
278 * this is not required for correctness, only to avoid wasting
279 * time on a large number of spurious interrupts. In testing,
280 * a sync reduced the observed spurious interrupts to zero.
284 raw_spin_unlock_irqrestore(&qe_ic_lock, flags);
287 static struct irq_chip qe_ic_irq_chip = {
289 .irq_unmask = qe_ic_unmask_irq,
290 .irq_mask = qe_ic_mask_irq,
291 .irq_mask_ack = qe_ic_mask_irq,
294 static int qe_ic_host_match(struct irq_domain *h, struct device_node *node,
295 enum irq_domain_bus_token bus_token)
297 /* Exact match, unless qe_ic node is NULL */
298 struct device_node *of_node = irq_domain_get_of_node(h);
299 return of_node == NULL || of_node == node;
302 static int qe_ic_host_map(struct irq_domain *h, unsigned int virq,
305 struct qe_ic *qe_ic = h->host_data;
306 struct irq_chip *chip;
308 if (hw >= ARRAY_SIZE(qe_ic_info)) {
309 pr_err("%s: Invalid hw irq number for QEIC\n", __func__);
313 if (qe_ic_info[hw].mask == 0) {
314 printk(KERN_ERR "Can't map reserved IRQ\n");
318 chip = &qe_ic->hc_irq;
320 irq_set_chip_data(virq, qe_ic);
321 irq_set_status_flags(virq, IRQ_LEVEL);
323 irq_set_chip_and_handler(virq, chip, handle_level_irq);
328 static const struct irq_domain_ops qe_ic_host_ops = {
329 .match = qe_ic_host_match,
330 .map = qe_ic_host_map,
331 .xlate = irq_domain_xlate_onetwocell,
334 /* Return an interrupt vector or 0 if no interrupt is pending. */
335 static unsigned int qe_ic_get_low_irq(struct qe_ic *qe_ic)
339 BUG_ON(qe_ic == NULL);
341 /* get the interrupt source vector. */
342 irq = qe_ic_read(qe_ic->regs, QEIC_CIVEC) >> 26;
347 return irq_linear_revmap(qe_ic->irqhost, irq);
350 /* Return an interrupt vector or 0 if no interrupt is pending. */
351 static unsigned int qe_ic_get_high_irq(struct qe_ic *qe_ic)
355 BUG_ON(qe_ic == NULL);
357 /* get the interrupt source vector. */
358 irq = qe_ic_read(qe_ic->regs, QEIC_CHIVEC) >> 26;
363 return irq_linear_revmap(qe_ic->irqhost, irq);
366 static void qe_ic_cascade_low(struct irq_desc *desc)
368 struct qe_ic *qe_ic = irq_desc_get_handler_data(desc);
369 unsigned int cascade_irq = qe_ic_get_low_irq(qe_ic);
370 struct irq_chip *chip = irq_desc_get_chip(desc);
372 if (cascade_irq != 0)
373 generic_handle_irq(cascade_irq);
376 chip->irq_eoi(&desc->irq_data);
379 static void qe_ic_cascade_high(struct irq_desc *desc)
381 struct qe_ic *qe_ic = irq_desc_get_handler_data(desc);
382 unsigned int cascade_irq = qe_ic_get_high_irq(qe_ic);
383 struct irq_chip *chip = irq_desc_get_chip(desc);
385 if (cascade_irq != 0)
386 generic_handle_irq(cascade_irq);
389 chip->irq_eoi(&desc->irq_data);
392 static void qe_ic_cascade_muxed_mpic(struct irq_desc *desc)
394 struct qe_ic *qe_ic = irq_desc_get_handler_data(desc);
395 unsigned int cascade_irq;
396 struct irq_chip *chip = irq_desc_get_chip(desc);
398 cascade_irq = qe_ic_get_high_irq(qe_ic);
399 if (cascade_irq == 0)
400 cascade_irq = qe_ic_get_low_irq(qe_ic);
402 if (cascade_irq != 0)
403 generic_handle_irq(cascade_irq);
405 chip->irq_eoi(&desc->irq_data);
408 static int qe_ic_init(struct platform_device *pdev)
410 struct device *dev = &pdev->dev;
411 void (*low_handler)(struct irq_desc *desc);
412 void (*high_handler)(struct irq_desc *desc);
414 struct resource *res;
415 struct device_node *node = pdev->dev.of_node;
417 res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
419 dev_err(dev, "no memory resource defined\n");
423 qe_ic = devm_kzalloc(dev, sizeof(*qe_ic), GFP_KERNEL);
427 qe_ic->regs = devm_ioremap(dev, res->start, resource_size(res));
428 if (qe_ic->regs == NULL) {
429 dev_err(dev, "failed to ioremap() registers\n");
433 qe_ic->hc_irq = qe_ic_irq_chip;
435 qe_ic->virq_high = platform_get_irq(pdev, 0);
436 qe_ic->virq_low = platform_get_irq(pdev, 1);
438 if (qe_ic->virq_low <= 0)
441 if (qe_ic->virq_high > 0 && qe_ic->virq_high != qe_ic->virq_low) {
442 low_handler = qe_ic_cascade_low;
443 high_handler = qe_ic_cascade_high;
445 low_handler = qe_ic_cascade_muxed_mpic;
449 qe_ic->irqhost = irq_domain_add_linear(node, NR_QE_IC_INTS,
450 &qe_ic_host_ops, qe_ic);
451 if (qe_ic->irqhost == NULL) {
452 dev_err(dev, "failed to add irq domain\n");
456 qe_ic_write(qe_ic->regs, QEIC_CICR, 0);
458 irq_set_handler_data(qe_ic->virq_low, qe_ic);
459 irq_set_chained_handler(qe_ic->virq_low, low_handler);
462 irq_set_handler_data(qe_ic->virq_high, qe_ic);
463 irq_set_chained_handler(qe_ic->virq_high, high_handler);
467 static const struct of_device_id qe_ic_ids[] = {
468 { .compatible = "fsl,qe-ic"},
473 static struct platform_driver qe_ic_driver =
477 .of_match_table = qe_ic_ids,
482 static int __init qe_ic_of_init(void)
484 platform_driver_register(&qe_ic_driver);
487 subsys_initcall(qe_ic_of_init);