blob: f8f45871429325a84e4e59412f1d0879ed9edf47 [file] [log] [blame]
b.liue9582032025-04-17 19:18:16 +08001// SPDX-License-Identifier: GPL-2.0-only
2/*
3 * linux/arch/arm/mach-mmp/irq.c
4 *
5 * Generic IRQ handling, GPIO IRQ demultiplexing, etc.
6 * Copyright (C) 2008 - 2012 Marvell Technology Group Ltd.
7 *
8 * Author: Bin Yang <bin.yang@marvell.com>
9 * Haojian Zhuang <haojian.zhuang@gmail.com>
10 */
11
12#include <linux/module.h>
13#include <linux/init.h>
14#include <linux/irq.h>
15#include <linux/irqchip.h>
16#include <linux/irqchip/chained_irq.h>
17#include <linux/irqdomain.h>
18#include <linux/io.h>
19#include <linux/ioport.h>
20#include <linux/of_address.h>
21#include <linux/of_irq.h>
22
23#include <asm/exception.h>
24#include <asm/hardirq.h>
25
26/* reduce the NR to save memory */
27#define MAX_ICU_NR 2/* 16 */
28
29#define PJ1_INT_SEL 0x10c
30#define PJ4_INT_SEL 0x104
31
32#define ICU_CFG2_OFFSET (0x154 - 0x100)
33#define ICU_BASE1_NR_IRQS (64)
34
35/* bit fields in PJ1_INT_SEL and PJ4_INT_SEL */
36#define SEL_INT_PENDING (1 << 6)
37#define SEL_INT_NUM_MASK 0x3f
38
39#define SEL_INT_PENDING_ASR1086 (1 << 7)
40#define SEL_INT_NUM_MASK_ASR1806 0x7f
41
42
43#define MMP2_ICU_INT_ROUTE_PJ4_IRQ (1 << 5)
44#define MMP2_ICU_INT_ROUTE_PJ4_FIQ (1 << 6)
45
46#define ICU_IRQ_ROUTE_TO_AP (1 << 6)
47#define ICU_INT_STATUS0 0x128
48#define ICU_INT_STATUS1 0x12C
49#define ICU_INT_STATUS2 0x14C
50#define ICU_INT_STATUS3 0x150
51
52#define ICU_AP_INT_SEL 0x10c
53#define ICU_CP_INT_SEL 0x104
54
55struct icu_chip_data {
56 int nr_irqs;
57 unsigned int virq_base;
58 unsigned int cascade_irq;
59 void __iomem *reg_status;
60 void __iomem *reg_mask;
61 void __iomem *reg_base2;
62 unsigned int conf_enable;
63 unsigned int conf_disable;
64 unsigned int conf_mask;
65 unsigned int conf2_mask;
66 unsigned int clr_mfp_irq_base;
67 unsigned int clr_mfp_hwirq;
68 struct irq_domain *domain;
69};
70
71struct mmp_intc_conf {
72 unsigned int conf_enable;
73 unsigned int conf_disable;
74 unsigned int conf_mask;
75 unsigned int conf2_mask;
76};
77
78static void __iomem *mmp_icu_base;
79static void __iomem *mmp_icu2_base;
80static struct icu_chip_data icu_data[MAX_ICU_NR];
81static int max_icu_nr;
82static int g_nr_irqs;
83
84extern void mmp2_clear_pmic_int(void);
85
86void __iomem *icu_get_base_addr(void)
87{
88 return mmp_icu_base;
89}
90
91void __iomem *icu_get_base_addr2(void)
92{
93 return mmp_icu_base + 0x14C;
94}
95
96void icu_dump_status(void)
97{
98 pr_emerg("ICU_STS0: 0x%08x\n", readl_relaxed(mmp_icu_base + ICU_INT_STATUS0));
99 pr_emerg("ICU_STS1: 0x%08x\n", readl_relaxed(mmp_icu_base + ICU_INT_STATUS1));
100
101 if (g_nr_irqs > 64)
102 pr_emerg("ICU_STS2: 0x%08x\n", readl_relaxed(mmp_icu_base + ICU_INT_STATUS2));
103 if (g_nr_irqs > 96)
104 pr_emerg("ICU_STS3: 0x%08x\n", readl_relaxed(mmp_icu_base + ICU_INT_STATUS3));
105}
106
107static void icu_mask_ack_irq(struct irq_data *d)
108{
109 struct irq_domain *domain = d->domain;
110 struct icu_chip_data *data = (struct icu_chip_data *)domain->host_data;
111 int hwirq;
112 u32 r;
113
114 hwirq = d->irq - data->virq_base;
115 if (data == &icu_data[0]) {
116 if (hwirq < ICU_BASE1_NR_IRQS) {
117 r = readl_relaxed(mmp_icu_base + (hwirq << 2));
118 r &= ~data->conf_mask;
119 r |= data->conf_disable;
120 writel_relaxed(r, mmp_icu_base + (hwirq << 2));
121 } else {
122 r = readl_relaxed(icu_data[0].reg_base2 + (hwirq << 2));
123 r &= ~data->conf_mask;
124 r |= data->conf_disable;
125 writel_relaxed(r, icu_data[0].reg_base2 + (hwirq << 2));
126 }
127 } else {
128#ifdef CONFIG_CPU_MMP2
129 if ((data->virq_base == data->clr_mfp_irq_base)
130 && (hwirq == data->clr_mfp_hwirq))
131 mmp2_clear_pmic_int();
132#endif
133 r = readl_relaxed(data->reg_mask) | (1 << hwirq);
134 writel_relaxed(r, data->reg_mask);
135 }
136
137}
138
139static void icu_mask_irq(struct irq_data *d)
140{
141 struct irq_domain *domain = d->domain;
142 struct icu_chip_data *data = (struct icu_chip_data *)domain->host_data;
143 int hwirq;
144 u32 r;
145
146 hwirq = d->irq - data->virq_base;
147 if (data == &icu_data[0]) {
148 if (hwirq < ICU_BASE1_NR_IRQS) {
149 r = readl_relaxed(mmp_icu_base + (hwirq << 2));
150 r &= ~data->conf_mask;
151 r |= data->conf_disable;
152 writel_relaxed(r, mmp_icu_base + (hwirq << 2));
153
154 if (data->conf2_mask) {
155 /*
156 * ICU1 (above) only controls PJ4 MP1; if using SMP,
157 * we need to also mask the MP2 and MM cores via ICU2.
158 */
159 r = readl_relaxed(mmp_icu2_base + (hwirq << 2));
160 r &= ~data->conf2_mask;
161 writel_relaxed(r, mmp_icu2_base + (hwirq << 2));
162 }
163 } else {
164 r = readl_relaxed(icu_data[0].reg_base2 + (hwirq << 2));
165 r &= ~data->conf_mask;
166 r |= data->conf_disable;
167 writel_relaxed(r, icu_data[0].reg_base2 + (hwirq << 2));
168
169 if (data->conf2_mask) {
170 /*
171 * ICU1 (above) only controls PJ4 MP1; if using SMP,
172 * we need to also mask the MP2 and MM cores via ICU2.
173 */
174 r = readl_relaxed(icu_data[0].reg_base2 + (hwirq << 2));
175 r &= ~data->conf2_mask;
176 writel_relaxed(r, icu_data[0].reg_base2 + (hwirq << 2));
177 }
178 }
179 } else {
180 r = readl_relaxed(data->reg_mask) | (1 << hwirq);
181 writel_relaxed(r, data->reg_mask);
182 }
183}
184
185static void icu_unmask_irq(struct irq_data *d)
186{
187 struct irq_domain *domain = d->domain;
188 struct icu_chip_data *data = (struct icu_chip_data *)domain->host_data;
189 int hwirq;
190 u32 r;
191
192 hwirq = d->irq - data->virq_base;
193 if (data == &icu_data[0]) {
194 if (hwirq < ICU_BASE1_NR_IRQS) {
195 r = readl_relaxed(mmp_icu_base + (hwirq << 2));
196 r &= ~data->conf_mask;
197 r |= data->conf_enable;
198 writel_relaxed(r, mmp_icu_base + (hwirq << 2));
199 } else {
200 r = readl_relaxed(icu_data[0].reg_base2 + (hwirq << 2));
201 r &= ~data->conf_mask;
202 r |= data->conf_enable;
203 writel_relaxed(r, icu_data[0].reg_base2 + (hwirq << 2));
204 }
205 } else {
206 r = readl_relaxed(data->reg_mask) & ~(1 << hwirq);
207 writel_relaxed(r, data->reg_mask);
208 }
209}
210
211struct irq_chip icu_irq_chip = {
212 .name = "icu_irq",
213 .irq_mask = icu_mask_irq,
214 .irq_mask_ack = icu_mask_ack_irq,
215 .irq_unmask = icu_unmask_irq,
216};
217#ifndef CONFIG_CPU_ASR18XX
218static void icu_mux_irq_demux(struct irq_desc *desc)
219{
220 unsigned int irq = irq_desc_get_irq(desc);
221 struct irq_chip *chip = irq_desc_get_chip(desc);
222 struct irq_domain *domain;
223 struct icu_chip_data *data;
224 int i;
225 unsigned long mask, status, n;
226
227 chained_irq_enter(chip, desc);
228
229 for (i = 1; i < max_icu_nr; i++) {
230 if (irq == icu_data[i].cascade_irq) {
231 domain = icu_data[i].domain;
232 data = (struct icu_chip_data *)domain->host_data;
233 break;
234 }
235 }
236 if (i >= max_icu_nr) {
237 pr_err("Spurious irq %d in MMP INTC\n", irq);
238 goto out;
239 }
240
241 mask = readl_relaxed(data->reg_mask);
242 while (1) {
243 status = readl_relaxed(data->reg_status) & ~mask;
244 if (status == 0)
245 break;
246 for_each_set_bit(n, &status, BITS_PER_LONG) {
247 generic_handle_irq(icu_data[i].virq_base + n);
248 }
249 }
250
251out:
252 chained_irq_exit(chip, desc);
253}
254#endif
255
256static int mmp_irq_domain_map(struct irq_domain *d, unsigned int irq,
257 irq_hw_number_t hw)
258{
259 irq_set_chip_and_handler(irq, &icu_irq_chip, handle_level_irq);
260 return 0;
261}
262
263static int mmp_irq_domain_xlate(struct irq_domain *d, struct device_node *node,
264 const u32 *intspec, unsigned int intsize,
265 unsigned long *out_hwirq,
266 unsigned int *out_type)
267{
268 *out_hwirq = intspec[0];
269 return 0;
270}
271
272static const struct irq_domain_ops mmp_irq_domain_ops = {
273 .map = mmp_irq_domain_map,
274 .xlate = mmp_irq_domain_xlate,
275};
276
277static const struct mmp_intc_conf mmp_conf = {
278 .conf_enable = 0x51,
279 .conf_disable = 0x0,
280 .conf_mask = 0x7f,
281};
282#ifndef CONFIG_CPU_ASR18XX
283static const struct mmp_intc_conf mmp2_conf = {
284 .conf_enable = 0x20,
285 .conf_disable = 0x0,
286 .conf_mask = MMP2_ICU_INT_ROUTE_PJ4_IRQ |
287 MMP2_ICU_INT_ROUTE_PJ4_FIQ,
288};
289
290static struct mmp_intc_conf mmp3_conf = {
291 .conf_enable = 0x20,
292 .conf_disable = 0x0,
293 .conf_mask = MMP2_ICU_INT_ROUTE_PJ4_IRQ |
294 MMP2_ICU_INT_ROUTE_PJ4_FIQ,
295 .conf2_mask = 0xf0,
296};
297#endif
298
299static void __exception_irq_entry mmp_handle_irq(struct pt_regs *regs)
300{
301 int hwirq;
302 u32 regval;
303
304 hwirq = readl_relaxed(mmp_icu_base + PJ1_INT_SEL);
305 if (g_nr_irqs <= ICU_BASE1_NR_IRQS) {
306 if (!(hwirq & SEL_INT_PENDING))
307 return;
308 hwirq &= SEL_INT_NUM_MASK;
309 } else {
310 if (!(hwirq & SEL_INT_PENDING_ASR1086))
311 return;
312 hwirq &= SEL_INT_NUM_MASK_ASR1806;
313 }
314
315 if (hwirq < ICU_BASE1_NR_IRQS)
316 regval = readl_relaxed(mmp_icu_base + (hwirq << 2));
317 else
318 regval = readl_relaxed(icu_data[0].reg_base2 + (hwirq << 2));
319
320 if (unlikely(!(regval & ICU_IRQ_ROUTE_TO_AP))) {
321 printk("error irq: %d, cfg: 0x%x, pending: 0x%x, 0x%x, status: 0x%x 0x%x\n",
322 hwirq, regval, readl_relaxed(mmp_icu_base + ICU_CP_INT_SEL),
323 readl_relaxed(mmp_icu_base + ICU_AP_INT_SEL),
324 readl_relaxed(mmp_icu_base + ICU_INT_STATUS0),
325 readl_relaxed(mmp_icu_base + ICU_INT_STATUS1));
326 return;
327 }
328 handle_domain_irq(icu_data[0].domain, hwirq, regs);
329}
330
331#ifndef CONFIG_CPU_ASR18XX
332static void __exception_irq_entry mmp2_handle_irq(struct pt_regs *regs)
333{
334 int hwirq;
335
336 hwirq = readl_relaxed(mmp_icu_base + PJ4_INT_SEL);
337 if (!(hwirq & SEL_INT_PENDING))
338 return;
339 hwirq &= SEL_INT_NUM_MASK;
340 handle_domain_irq(icu_data[0].domain, hwirq, regs);
341}
342#endif
343
344/* MMP (ARMv5) */
345void __init icu_init_irq(void)
346{
347 int irq;
348
349 max_icu_nr = 1;
350 mmp_icu_base = ioremap(0xd4282000, 0x1000);
351 icu_data[0].conf_enable = mmp_conf.conf_enable;
352 icu_data[0].conf_disable = mmp_conf.conf_disable;
353 icu_data[0].conf_mask = mmp_conf.conf_mask;
354 icu_data[0].nr_irqs = 64;
355 icu_data[0].virq_base = 0;
356 icu_data[0].domain = irq_domain_add_legacy(NULL, 64, 0, 0,
357 &irq_domain_simple_ops,
358 &icu_data[0]);
359 for (irq = 0; irq < 64; irq++) {
360 icu_mask_irq(irq_get_irq_data(irq));
361 irq_set_chip_and_handler(irq, &icu_irq_chip, handle_level_irq);
362 }
363 irq_set_default_host(icu_data[0].domain);
364 set_handle_irq(mmp_handle_irq);
365}
366#ifndef CONFIG_CPU_ASR18XX
367/* MMP2 (ARMv7) */
368void __init mmp2_init_icu(void)
369{
370 int irq, end;
371
372 max_icu_nr = 8;
373 mmp_icu_base = ioremap(0xd4282000, 0x1000);
374 icu_data[0].conf_enable = mmp2_conf.conf_enable;
375 icu_data[0].conf_disable = mmp2_conf.conf_disable;
376 icu_data[0].conf_mask = mmp2_conf.conf_mask;
377 icu_data[0].nr_irqs = 64;
378 icu_data[0].virq_base = 0;
379 icu_data[0].domain = irq_domain_add_legacy(NULL, 64, 0, 0,
380 &irq_domain_simple_ops,
381 &icu_data[0]);
382 icu_data[1].reg_status = mmp_icu_base + 0x150;
383 icu_data[1].reg_mask = mmp_icu_base + 0x168;
384 icu_data[1].clr_mfp_irq_base = icu_data[0].virq_base +
385 icu_data[0].nr_irqs;
386 icu_data[1].clr_mfp_hwirq = 1; /* offset to IRQ_MMP2_PMIC_BASE */
387 icu_data[1].nr_irqs = 2;
388 icu_data[1].cascade_irq = 4;
389 icu_data[1].virq_base = icu_data[0].virq_base + icu_data[0].nr_irqs;
390 icu_data[1].domain = irq_domain_add_legacy(NULL, icu_data[1].nr_irqs,
391 icu_data[1].virq_base, 0,
392 &irq_domain_simple_ops,
393 &icu_data[1]);
394 icu_data[2].reg_status = mmp_icu_base + 0x154;
395 icu_data[2].reg_mask = mmp_icu_base + 0x16c;
396 icu_data[2].nr_irqs = 2;
397 icu_data[2].cascade_irq = 5;
398 icu_data[2].virq_base = icu_data[1].virq_base + icu_data[1].nr_irqs;
399 icu_data[2].domain = irq_domain_add_legacy(NULL, icu_data[2].nr_irqs,
400 icu_data[2].virq_base, 0,
401 &irq_domain_simple_ops,
402 &icu_data[2]);
403 icu_data[3].reg_status = mmp_icu_base + 0x180;
404 icu_data[3].reg_mask = mmp_icu_base + 0x17c;
405 icu_data[3].nr_irqs = 3;
406 icu_data[3].cascade_irq = 9;
407 icu_data[3].virq_base = icu_data[2].virq_base + icu_data[2].nr_irqs;
408 icu_data[3].domain = irq_domain_add_legacy(NULL, icu_data[3].nr_irqs,
409 icu_data[3].virq_base, 0,
410 &irq_domain_simple_ops,
411 &icu_data[3]);
412 icu_data[4].reg_status = mmp_icu_base + 0x158;
413 icu_data[4].reg_mask = mmp_icu_base + 0x170;
414 icu_data[4].nr_irqs = 5;
415 icu_data[4].cascade_irq = 17;
416 icu_data[4].virq_base = icu_data[3].virq_base + icu_data[3].nr_irqs;
417 icu_data[4].domain = irq_domain_add_legacy(NULL, icu_data[4].nr_irqs,
418 icu_data[4].virq_base, 0,
419 &irq_domain_simple_ops,
420 &icu_data[4]);
421 icu_data[5].reg_status = mmp_icu_base + 0x15c;
422 icu_data[5].reg_mask = mmp_icu_base + 0x174;
423 icu_data[5].nr_irqs = 15;
424 icu_data[5].cascade_irq = 35;
425 icu_data[5].virq_base = icu_data[4].virq_base + icu_data[4].nr_irqs;
426 icu_data[5].domain = irq_domain_add_legacy(NULL, icu_data[5].nr_irqs,
427 icu_data[5].virq_base, 0,
428 &irq_domain_simple_ops,
429 &icu_data[5]);
430 icu_data[6].reg_status = mmp_icu_base + 0x160;
431 icu_data[6].reg_mask = mmp_icu_base + 0x178;
432 icu_data[6].nr_irqs = 2;
433 icu_data[6].cascade_irq = 51;
434 icu_data[6].virq_base = icu_data[5].virq_base + icu_data[5].nr_irqs;
435 icu_data[6].domain = irq_domain_add_legacy(NULL, icu_data[6].nr_irqs,
436 icu_data[6].virq_base, 0,
437 &irq_domain_simple_ops,
438 &icu_data[6]);
439 icu_data[7].reg_status = mmp_icu_base + 0x188;
440 icu_data[7].reg_mask = mmp_icu_base + 0x184;
441 icu_data[7].nr_irqs = 2;
442 icu_data[7].cascade_irq = 55;
443 icu_data[7].virq_base = icu_data[6].virq_base + icu_data[6].nr_irqs;
444 icu_data[7].domain = irq_domain_add_legacy(NULL, icu_data[7].nr_irqs,
445 icu_data[7].virq_base, 0,
446 &irq_domain_simple_ops,
447 &icu_data[7]);
448 end = icu_data[7].virq_base + icu_data[7].nr_irqs;
449 for (irq = 0; irq < end; irq++) {
450 icu_mask_irq(irq_get_irq_data(irq));
451 if (irq == icu_data[1].cascade_irq ||
452 irq == icu_data[2].cascade_irq ||
453 irq == icu_data[3].cascade_irq ||
454 irq == icu_data[4].cascade_irq ||
455 irq == icu_data[5].cascade_irq ||
456 irq == icu_data[6].cascade_irq ||
457 irq == icu_data[7].cascade_irq) {
458 irq_set_chip(irq, &icu_irq_chip);
459 irq_set_chained_handler(irq, icu_mux_irq_demux);
460 } else {
461 irq_set_chip_and_handler(irq, &icu_irq_chip,
462 handle_level_irq);
463 }
464 }
465 irq_set_default_host(icu_data[0].domain);
466 set_handle_irq(mmp2_handle_irq);
467}
468#endif
469#ifdef CONFIG_OF
470static int __init mmp_init_bases(struct device_node *node)
471{
472 int ret, nr_irqs, irq, i = 0;
473
474 ret = of_property_read_u32(node, "mrvl,intc-nr-irqs", &nr_irqs);
475 if (ret) {
476 pr_err("Not found mrvl,intc-nr-irqs property\n");
477 return ret;
478 }
479 g_nr_irqs = nr_irqs;
480
481 mmp_icu_base = of_iomap(node, 0);
482 if (!mmp_icu_base) {
483 pr_err("Failed to get interrupt controller register\n");
484 return -ENOMEM;
485 }
486
487 if (g_nr_irqs > ICU_BASE1_NR_IRQS)
488 icu_data[0].reg_base2 = mmp_icu_base + ICU_CFG2_OFFSET;
489 else
490 icu_data[0].reg_base2 = 0;
491
492 icu_data[0].virq_base = 0;
493 icu_data[0].domain = irq_domain_add_linear(node, nr_irqs,
494 &mmp_irq_domain_ops,
495 &icu_data[0]);
496 for (irq = 0; irq < nr_irqs; irq++) {
497 ret = irq_create_mapping(icu_data[0].domain, irq);
498 if (!ret) {
499 pr_err("Failed to mapping hwirq\n");
500 goto err;
501 }
502 if (!irq)
503 icu_data[0].virq_base = ret;
504 }
505 icu_data[0].nr_irqs = nr_irqs;
506 return 0;
507err:
508 if (icu_data[0].virq_base) {
509 for (i = 0; i < irq; i++)
510 irq_dispose_mapping(icu_data[0].virq_base + i);
511 }
512 irq_domain_remove(icu_data[0].domain);
513 iounmap(mmp_icu_base);
514 return -EINVAL;
515}
516
517static int __init mmp_of_init(struct device_node *node,
518 struct device_node *parent)
519{
520 int ret;
521
522 ret = mmp_init_bases(node);
523 if (ret < 0)
524 return ret;
525
526 icu_data[0].conf_enable = mmp_conf.conf_enable;
527 icu_data[0].conf_disable = mmp_conf.conf_disable;
528 icu_data[0].conf_mask = mmp_conf.conf_mask;
529 set_handle_irq(mmp_handle_irq);
530 max_icu_nr = 1;
531 return 0;
532}
533IRQCHIP_DECLARE(mmp_intc, "mrvl,mmp-intc", mmp_of_init);
534
535#ifndef CONFIG_CPU_ASR18XX
536static int __init mmp2_of_init(struct device_node *node,
537 struct device_node *parent)
538{
539 int ret;
540
541 ret = mmp_init_bases(node);
542 if (ret < 0)
543 return ret;
544
545 icu_data[0].conf_enable = mmp2_conf.conf_enable;
546 icu_data[0].conf_disable = mmp2_conf.conf_disable;
547 icu_data[0].conf_mask = mmp2_conf.conf_mask;
548 set_handle_irq(mmp2_handle_irq);
549 max_icu_nr = 1;
550 return 0;
551}
552IRQCHIP_DECLARE(mmp2_intc, "mrvl,mmp2-intc", mmp2_of_init);
553
554static int __init mmp3_of_init(struct device_node *node,
555 struct device_node *parent)
556{
557 int ret;
558
559 mmp_icu2_base = of_iomap(node, 1);
560 if (!mmp_icu2_base) {
561 pr_err("Failed to get interrupt controller register #2\n");
562 return -ENODEV;
563 }
564
565 ret = mmp_init_bases(node);
566 if (ret < 0) {
567 iounmap(mmp_icu2_base);
568 return ret;
569 }
570
571 icu_data[0].conf_enable = mmp3_conf.conf_enable;
572 icu_data[0].conf_disable = mmp3_conf.conf_disable;
573 icu_data[0].conf_mask = mmp3_conf.conf_mask;
574 icu_data[0].conf2_mask = mmp3_conf.conf2_mask;
575
576 if (!parent) {
577 /* This is the main interrupt controller. */
578 set_handle_irq(mmp2_handle_irq);
579 }
580
581 max_icu_nr = 1;
582 return 0;
583}
584IRQCHIP_DECLARE(mmp3_intc, "marvell,mmp3-intc", mmp3_of_init);
585
586static int __init mmp2_mux_of_init(struct device_node *node,
587 struct device_node *parent)
588{
589 int i, ret, irq, j = 0;
590 u32 nr_irqs, mfp_irq;
591 u32 reg[4];
592
593 if (!parent)
594 return -ENODEV;
595
596 i = max_icu_nr;
597 ret = of_property_read_u32(node, "mrvl,intc-nr-irqs",
598 &nr_irqs);
599 if (ret) {
600 pr_err("Not found mrvl,intc-nr-irqs property\n");
601 return -EINVAL;
602 }
603
604 /*
605 * For historical reasons, the "regs" property of the
606 * mrvl,mmp2-mux-intc is not a regular "regs" property containing
607 * addresses on the parent bus, but offsets from the intc's base.
608 * That is why we can't use of_address_to_resource() here.
609 */
610 ret = of_property_read_variable_u32_array(node, "reg", reg,
611 ARRAY_SIZE(reg),
612 ARRAY_SIZE(reg));
613 if (ret < 0) {
614 pr_err("Not found reg property\n");
615 return -EINVAL;
616 }
617 icu_data[i].reg_status = mmp_icu_base + reg[0];
618 icu_data[i].reg_mask = mmp_icu_base + reg[2];
619 icu_data[i].cascade_irq = irq_of_parse_and_map(node, 0);
620 if (!icu_data[i].cascade_irq)
621 return -EINVAL;
622
623 icu_data[i].virq_base = 0;
624 icu_data[i].domain = irq_domain_add_linear(node, nr_irqs,
625 &mmp_irq_domain_ops,
626 &icu_data[i]);
627 for (irq = 0; irq < nr_irqs; irq++) {
628 ret = irq_create_mapping(icu_data[i].domain, irq);
629 if (!ret) {
630 pr_err("Failed to mapping hwirq\n");
631 goto err;
632 }
633 if (!irq)
634 icu_data[i].virq_base = ret;
635 }
636 icu_data[i].nr_irqs = nr_irqs;
637 if (!of_property_read_u32(node, "mrvl,clr-mfp-irq",
638 &mfp_irq)) {
639 icu_data[i].clr_mfp_irq_base = icu_data[i].virq_base;
640 icu_data[i].clr_mfp_hwirq = mfp_irq;
641 }
642 irq_set_chained_handler(icu_data[i].cascade_irq,
643 icu_mux_irq_demux);
644 max_icu_nr++;
645 return 0;
646err:
647 if (icu_data[i].virq_base) {
648 for (j = 0; j < irq; j++)
649 irq_dispose_mapping(icu_data[i].virq_base + j);
650 }
651 irq_domain_remove(icu_data[i].domain);
652 return -EINVAL;
653}
654IRQCHIP_DECLARE(mmp2_mux_intc, "mrvl,mmp2-mux-intc", mmp2_mux_of_init);
655#endif
656#endif