Linux Audio

Check our new training course

Loading...
v3.15
  1/*
  2 *  linux/arch/arm/mach-mmp/irq.c
  3 *
  4 *  Generic IRQ handling, GPIO IRQ demultiplexing, etc.
  5 *  Copyright (C) 2008 - 2012 Marvell Technology Group Ltd.
  6 *
  7 *  Author:	Bin Yang <bin.yang@marvell.com>
  8 *              Haojian Zhuang <haojian.zhuang@gmail.com>
  9 *
 10 *  This program is free software; you can redistribute it and/or modify
 11 *  it under the terms of the GNU General Public License version 2 as
 12 *  published by the Free Software Foundation.
 13 */
 14
 15#include <linux/module.h>
 16#include <linux/init.h>
 17#include <linux/irq.h>
 
 18#include <linux/irqdomain.h>
 19#include <linux/io.h>
 20#include <linux/ioport.h>
 21#include <linux/of_address.h>
 22#include <linux/of_irq.h>
 23
 24#include <asm/exception.h>
 25#include <asm/hardirq.h>
 26
 27#include "irqchip.h"
 28
 29#define MAX_ICU_NR		16
 30
 31#define PJ1_INT_SEL		0x10c
 32#define PJ4_INT_SEL		0x104
 33
 34/* bit fields in PJ1_INT_SEL and PJ4_INT_SEL */
 35#define SEL_INT_PENDING		(1 << 6)
 36#define SEL_INT_NUM_MASK	0x3f
 37
 38struct icu_chip_data {
 39	int			nr_irqs;
 40	unsigned int		virq_base;
 41	unsigned int		cascade_irq;
 42	void __iomem		*reg_status;
 43	void __iomem		*reg_mask;
 44	unsigned int		conf_enable;
 45	unsigned int		conf_disable;
 46	unsigned int		conf_mask;
 47	unsigned int		clr_mfp_irq_base;
 48	unsigned int		clr_mfp_hwirq;
 49	struct irq_domain	*domain;
 50};
 51
 52struct mmp_intc_conf {
 53	unsigned int	conf_enable;
 54	unsigned int	conf_disable;
 55	unsigned int	conf_mask;
 56};
 57
 58static void __iomem *mmp_icu_base;
 59static struct icu_chip_data icu_data[MAX_ICU_NR];
 60static int max_icu_nr;
 61
 62extern void mmp2_clear_pmic_int(void);
 63
 64static void icu_mask_ack_irq(struct irq_data *d)
 65{
 66	struct irq_domain *domain = d->domain;
 67	struct icu_chip_data *data = (struct icu_chip_data *)domain->host_data;
 68	int hwirq;
 69	u32 r;
 70
 71	hwirq = d->irq - data->virq_base;
 72	if (data == &icu_data[0]) {
 73		r = readl_relaxed(mmp_icu_base + (hwirq << 2));
 74		r &= ~data->conf_mask;
 75		r |= data->conf_disable;
 76		writel_relaxed(r, mmp_icu_base + (hwirq << 2));
 77	} else {
 78#ifdef CONFIG_CPU_MMP2
 79		if ((data->virq_base == data->clr_mfp_irq_base)
 80			&& (hwirq == data->clr_mfp_hwirq))
 81			mmp2_clear_pmic_int();
 82#endif
 83		r = readl_relaxed(data->reg_mask) | (1 << hwirq);
 84		writel_relaxed(r, data->reg_mask);
 85	}
 86}
 87
 88static void icu_mask_irq(struct irq_data *d)
 89{
 90	struct irq_domain *domain = d->domain;
 91	struct icu_chip_data *data = (struct icu_chip_data *)domain->host_data;
 92	int hwirq;
 93	u32 r;
 94
 95	hwirq = d->irq - data->virq_base;
 96	if (data == &icu_data[0]) {
 97		r = readl_relaxed(mmp_icu_base + (hwirq << 2));
 98		r &= ~data->conf_mask;
 99		r |= data->conf_disable;
100		writel_relaxed(r, mmp_icu_base + (hwirq << 2));
101	} else {
102		r = readl_relaxed(data->reg_mask) | (1 << hwirq);
103		writel_relaxed(r, data->reg_mask);
104	}
105}
106
107static void icu_unmask_irq(struct irq_data *d)
108{
109	struct irq_domain *domain = d->domain;
110	struct icu_chip_data *data = (struct icu_chip_data *)domain->host_data;
111	int hwirq;
112	u32 r;
113
114	hwirq = d->irq - data->virq_base;
115	if (data == &icu_data[0]) {
116		r = readl_relaxed(mmp_icu_base + (hwirq << 2));
117		r &= ~data->conf_mask;
118		r |= data->conf_enable;
119		writel_relaxed(r, mmp_icu_base + (hwirq << 2));
120	} else {
121		r = readl_relaxed(data->reg_mask) & ~(1 << hwirq);
122		writel_relaxed(r, data->reg_mask);
123	}
124}
125
126struct irq_chip icu_irq_chip = {
127	.name		= "icu_irq",
128	.irq_mask	= icu_mask_irq,
129	.irq_mask_ack	= icu_mask_ack_irq,
130	.irq_unmask	= icu_unmask_irq,
131};
132
133static void icu_mux_irq_demux(unsigned int irq, struct irq_desc *desc)
134{
 
135	struct irq_domain *domain;
136	struct icu_chip_data *data;
137	int i;
138	unsigned long mask, status, n;
139
140	for (i = 1; i < max_icu_nr; i++) {
141		if (irq == icu_data[i].cascade_irq) {
142			domain = icu_data[i].domain;
143			data = (struct icu_chip_data *)domain->host_data;
144			break;
145		}
146	}
147	if (i >= max_icu_nr) {
148		pr_err("Spurious irq %d in MMP INTC\n", irq);
149		return;
150	}
151
152	mask = readl_relaxed(data->reg_mask);
153	while (1) {
154		status = readl_relaxed(data->reg_status) & ~mask;
155		if (status == 0)
156			break;
157		for_each_set_bit(n, &status, BITS_PER_LONG) {
158			generic_handle_irq(icu_data[i].virq_base + n);
159		}
160	}
161}
162
163static int mmp_irq_domain_map(struct irq_domain *d, unsigned int irq,
164			      irq_hw_number_t hw)
165{
166	irq_set_chip_and_handler(irq, &icu_irq_chip, handle_level_irq);
167	set_irq_flags(irq, IRQF_VALID);
168	return 0;
169}
170
171static int mmp_irq_domain_xlate(struct irq_domain *d, struct device_node *node,
172				const u32 *intspec, unsigned int intsize,
173				unsigned long *out_hwirq,
174				unsigned int *out_type)
175{
176	*out_hwirq = intspec[0];
177	return 0;
178}
179
180const struct irq_domain_ops mmp_irq_domain_ops = {
181	.map		= mmp_irq_domain_map,
182	.xlate		= mmp_irq_domain_xlate,
183};
184
185static struct mmp_intc_conf mmp_conf = {
186	.conf_enable	= 0x51,
187	.conf_disable	= 0x0,
188	.conf_mask	= 0x7f,
189};
190
191static struct mmp_intc_conf mmp2_conf = {
192	.conf_enable	= 0x20,
193	.conf_disable	= 0x0,
194	.conf_mask	= 0x7f,
195};
196
197static void __exception_irq_entry mmp_handle_irq(struct pt_regs *regs)
198{
199	int irq, hwirq;
200
201	hwirq = readl_relaxed(mmp_icu_base + PJ1_INT_SEL);
202	if (!(hwirq & SEL_INT_PENDING))
203		return;
204	hwirq &= SEL_INT_NUM_MASK;
205	irq = irq_find_mapping(icu_data[0].domain, hwirq);
206	handle_IRQ(irq, regs);
207}
208
209static void __exception_irq_entry mmp2_handle_irq(struct pt_regs *regs)
210{
211	int irq, hwirq;
212
213	hwirq = readl_relaxed(mmp_icu_base + PJ4_INT_SEL);
214	if (!(hwirq & SEL_INT_PENDING))
215		return;
216	hwirq &= SEL_INT_NUM_MASK;
217	irq = irq_find_mapping(icu_data[0].domain, hwirq);
218	handle_IRQ(irq, regs);
219}
220
221/* MMP (ARMv5) */
222void __init icu_init_irq(void)
223{
224	int irq;
225
226	max_icu_nr = 1;
227	mmp_icu_base = ioremap(0xd4282000, 0x1000);
228	icu_data[0].conf_enable = mmp_conf.conf_enable;
229	icu_data[0].conf_disable = mmp_conf.conf_disable;
230	icu_data[0].conf_mask = mmp_conf.conf_mask;
231	icu_data[0].nr_irqs = 64;
232	icu_data[0].virq_base = 0;
233	icu_data[0].domain = irq_domain_add_legacy(NULL, 64, 0, 0,
234						   &irq_domain_simple_ops,
235						   &icu_data[0]);
236	for (irq = 0; irq < 64; irq++) {
237		icu_mask_irq(irq_get_irq_data(irq));
238		irq_set_chip_and_handler(irq, &icu_irq_chip, handle_level_irq);
239		set_irq_flags(irq, IRQF_VALID);
240	}
241	irq_set_default_host(icu_data[0].domain);
242	set_handle_irq(mmp_handle_irq);
243}
244
245/* MMP2 (ARMv7) */
246void __init mmp2_init_icu(void)
247{
248	int irq, end;
249
250	max_icu_nr = 8;
251	mmp_icu_base = ioremap(0xd4282000, 0x1000);
252	icu_data[0].conf_enable = mmp2_conf.conf_enable;
253	icu_data[0].conf_disable = mmp2_conf.conf_disable;
254	icu_data[0].conf_mask = mmp2_conf.conf_mask;
255	icu_data[0].nr_irqs = 64;
256	icu_data[0].virq_base = 0;
257	icu_data[0].domain = irq_domain_add_legacy(NULL, 64, 0, 0,
258						   &irq_domain_simple_ops,
259						   &icu_data[0]);
260	icu_data[1].reg_status = mmp_icu_base + 0x150;
261	icu_data[1].reg_mask = mmp_icu_base + 0x168;
262	icu_data[1].clr_mfp_irq_base = icu_data[0].virq_base +
263				icu_data[0].nr_irqs;
264	icu_data[1].clr_mfp_hwirq = 1;		/* offset to IRQ_MMP2_PMIC_BASE */
265	icu_data[1].nr_irqs = 2;
266	icu_data[1].cascade_irq = 4;
267	icu_data[1].virq_base = icu_data[0].virq_base + icu_data[0].nr_irqs;
268	icu_data[1].domain = irq_domain_add_legacy(NULL, icu_data[1].nr_irqs,
269						   icu_data[1].virq_base, 0,
270						   &irq_domain_simple_ops,
271						   &icu_data[1]);
272	icu_data[2].reg_status = mmp_icu_base + 0x154;
273	icu_data[2].reg_mask = mmp_icu_base + 0x16c;
274	icu_data[2].nr_irqs = 2;
275	icu_data[2].cascade_irq = 5;
276	icu_data[2].virq_base = icu_data[1].virq_base + icu_data[1].nr_irqs;
277	icu_data[2].domain = irq_domain_add_legacy(NULL, icu_data[2].nr_irqs,
278						   icu_data[2].virq_base, 0,
279						   &irq_domain_simple_ops,
280						   &icu_data[2]);
281	icu_data[3].reg_status = mmp_icu_base + 0x180;
282	icu_data[3].reg_mask = mmp_icu_base + 0x17c;
283	icu_data[3].nr_irqs = 3;
284	icu_data[3].cascade_irq = 9;
285	icu_data[3].virq_base = icu_data[2].virq_base + icu_data[2].nr_irqs;
286	icu_data[3].domain = irq_domain_add_legacy(NULL, icu_data[3].nr_irqs,
287						   icu_data[3].virq_base, 0,
288						   &irq_domain_simple_ops,
289						   &icu_data[3]);
290	icu_data[4].reg_status = mmp_icu_base + 0x158;
291	icu_data[4].reg_mask = mmp_icu_base + 0x170;
292	icu_data[4].nr_irqs = 5;
293	icu_data[4].cascade_irq = 17;
294	icu_data[4].virq_base = icu_data[3].virq_base + icu_data[3].nr_irqs;
295	icu_data[4].domain = irq_domain_add_legacy(NULL, icu_data[4].nr_irqs,
296						   icu_data[4].virq_base, 0,
297						   &irq_domain_simple_ops,
298						   &icu_data[4]);
299	icu_data[5].reg_status = mmp_icu_base + 0x15c;
300	icu_data[5].reg_mask = mmp_icu_base + 0x174;
301	icu_data[5].nr_irqs = 15;
302	icu_data[5].cascade_irq = 35;
303	icu_data[5].virq_base = icu_data[4].virq_base + icu_data[4].nr_irqs;
304	icu_data[5].domain = irq_domain_add_legacy(NULL, icu_data[5].nr_irqs,
305						   icu_data[5].virq_base, 0,
306						   &irq_domain_simple_ops,
307						   &icu_data[5]);
308	icu_data[6].reg_status = mmp_icu_base + 0x160;
309	icu_data[6].reg_mask = mmp_icu_base + 0x178;
310	icu_data[6].nr_irqs = 2;
311	icu_data[6].cascade_irq = 51;
312	icu_data[6].virq_base = icu_data[5].virq_base + icu_data[5].nr_irqs;
313	icu_data[6].domain = irq_domain_add_legacy(NULL, icu_data[6].nr_irqs,
314						   icu_data[6].virq_base, 0,
315						   &irq_domain_simple_ops,
316						   &icu_data[6]);
317	icu_data[7].reg_status = mmp_icu_base + 0x188;
318	icu_data[7].reg_mask = mmp_icu_base + 0x184;
319	icu_data[7].nr_irqs = 2;
320	icu_data[7].cascade_irq = 55;
321	icu_data[7].virq_base = icu_data[6].virq_base + icu_data[6].nr_irqs;
322	icu_data[7].domain = irq_domain_add_legacy(NULL, icu_data[7].nr_irqs,
323						   icu_data[7].virq_base, 0,
324						   &irq_domain_simple_ops,
325						   &icu_data[7]);
326	end = icu_data[7].virq_base + icu_data[7].nr_irqs;
327	for (irq = 0; irq < end; irq++) {
328		icu_mask_irq(irq_get_irq_data(irq));
329		if (irq == icu_data[1].cascade_irq ||
330		    irq == icu_data[2].cascade_irq ||
331		    irq == icu_data[3].cascade_irq ||
332		    irq == icu_data[4].cascade_irq ||
333		    irq == icu_data[5].cascade_irq ||
334		    irq == icu_data[6].cascade_irq ||
335		    irq == icu_data[7].cascade_irq) {
336			irq_set_chip(irq, &icu_irq_chip);
337			irq_set_chained_handler(irq, icu_mux_irq_demux);
338		} else {
339			irq_set_chip_and_handler(irq, &icu_irq_chip,
340						 handle_level_irq);
341		}
342		set_irq_flags(irq, IRQF_VALID);
343	}
344	irq_set_default_host(icu_data[0].domain);
345	set_handle_irq(mmp2_handle_irq);
346}
347
348#ifdef CONFIG_OF
349static int __init mmp_init_bases(struct device_node *node)
350{
351	int ret, nr_irqs, irq, i = 0;
352
353	ret = of_property_read_u32(node, "mrvl,intc-nr-irqs", &nr_irqs);
354	if (ret) {
355		pr_err("Not found mrvl,intc-nr-irqs property\n");
356		return ret;
357	}
358
359	mmp_icu_base = of_iomap(node, 0);
360	if (!mmp_icu_base) {
361		pr_err("Failed to get interrupt controller register\n");
362		return -ENOMEM;
363	}
364
365	icu_data[0].virq_base = 0;
366	icu_data[0].domain = irq_domain_add_linear(node, nr_irqs,
367						   &mmp_irq_domain_ops,
368						   &icu_data[0]);
369	for (irq = 0; irq < nr_irqs; irq++) {
370		ret = irq_create_mapping(icu_data[0].domain, irq);
371		if (!ret) {
372			pr_err("Failed to mapping hwirq\n");
373			goto err;
374		}
375		if (!irq)
376			icu_data[0].virq_base = ret;
377	}
378	icu_data[0].nr_irqs = nr_irqs;
379	return 0;
380err:
381	if (icu_data[0].virq_base) {
382		for (i = 0; i < irq; i++)
383			irq_dispose_mapping(icu_data[0].virq_base + i);
384	}
385	irq_domain_remove(icu_data[0].domain);
386	iounmap(mmp_icu_base);
387	return -EINVAL;
388}
389
390static int __init mmp_of_init(struct device_node *node,
391			      struct device_node *parent)
392{
393	int ret;
394
395	ret = mmp_init_bases(node);
396	if (ret < 0)
397		return ret;
398
399	icu_data[0].conf_enable = mmp_conf.conf_enable;
400	icu_data[0].conf_disable = mmp_conf.conf_disable;
401	icu_data[0].conf_mask = mmp_conf.conf_mask;
402	irq_set_default_host(icu_data[0].domain);
403	set_handle_irq(mmp_handle_irq);
404	max_icu_nr = 1;
405	return 0;
406}
407IRQCHIP_DECLARE(mmp_intc, "mrvl,mmp-intc", mmp_of_init);
408
409static int __init mmp2_of_init(struct device_node *node,
410			       struct device_node *parent)
411{
412	int ret;
413
414	ret = mmp_init_bases(node);
415	if (ret < 0)
416		return ret;
417
418	icu_data[0].conf_enable = mmp2_conf.conf_enable;
419	icu_data[0].conf_disable = mmp2_conf.conf_disable;
420	icu_data[0].conf_mask = mmp2_conf.conf_mask;
421	irq_set_default_host(icu_data[0].domain);
422	set_handle_irq(mmp2_handle_irq);
423	max_icu_nr = 1;
424	return 0;
425}
426IRQCHIP_DECLARE(mmp2_intc, "mrvl,mmp2-intc", mmp2_of_init);
427
428static int __init mmp2_mux_of_init(struct device_node *node,
429				   struct device_node *parent)
430{
431	struct resource res;
432	int i, ret, irq, j = 0;
433	u32 nr_irqs, mfp_irq;
434
435	if (!parent)
436		return -ENODEV;
437
438	i = max_icu_nr;
439	ret = of_property_read_u32(node, "mrvl,intc-nr-irqs",
440				   &nr_irqs);
441	if (ret) {
442		pr_err("Not found mrvl,intc-nr-irqs property\n");
443		return -EINVAL;
444	}
445	ret = of_address_to_resource(node, 0, &res);
446	if (ret < 0) {
447		pr_err("Not found reg property\n");
448		return -EINVAL;
449	}
450	icu_data[i].reg_status = mmp_icu_base + res.start;
451	ret = of_address_to_resource(node, 1, &res);
452	if (ret < 0) {
453		pr_err("Not found reg property\n");
454		return -EINVAL;
455	}
456	icu_data[i].reg_mask = mmp_icu_base + res.start;
457	icu_data[i].cascade_irq = irq_of_parse_and_map(node, 0);
458	if (!icu_data[i].cascade_irq)
459		return -EINVAL;
460
461	icu_data[i].virq_base = 0;
462	icu_data[i].domain = irq_domain_add_linear(node, nr_irqs,
463						   &mmp_irq_domain_ops,
464						   &icu_data[i]);
465	for (irq = 0; irq < nr_irqs; irq++) {
466		ret = irq_create_mapping(icu_data[i].domain, irq);
467		if (!ret) {
468			pr_err("Failed to mapping hwirq\n");
469			goto err;
470		}
471		if (!irq)
472			icu_data[i].virq_base = ret;
473	}
474	icu_data[i].nr_irqs = nr_irqs;
475	if (!of_property_read_u32(node, "mrvl,clr-mfp-irq",
476				  &mfp_irq)) {
477		icu_data[i].clr_mfp_irq_base = icu_data[i].virq_base;
478		icu_data[i].clr_mfp_hwirq = mfp_irq;
479	}
480	irq_set_chained_handler(icu_data[i].cascade_irq,
481				icu_mux_irq_demux);
482	max_icu_nr++;
483	return 0;
484err:
485	if (icu_data[i].virq_base) {
486		for (j = 0; j < irq; j++)
487			irq_dispose_mapping(icu_data[i].virq_base + j);
488	}
489	irq_domain_remove(icu_data[i].domain);
490	return -EINVAL;
491}
492IRQCHIP_DECLARE(mmp2_mux_intc, "mrvl,mmp2-mux-intc", mmp2_mux_of_init);
493#endif
v4.10.11
  1/*
  2 *  linux/arch/arm/mach-mmp/irq.c
  3 *
  4 *  Generic IRQ handling, GPIO IRQ demultiplexing, etc.
  5 *  Copyright (C) 2008 - 2012 Marvell Technology Group Ltd.
  6 *
  7 *  Author:	Bin Yang <bin.yang@marvell.com>
  8 *              Haojian Zhuang <haojian.zhuang@gmail.com>
  9 *
 10 *  This program is free software; you can redistribute it and/or modify
 11 *  it under the terms of the GNU General Public License version 2 as
 12 *  published by the Free Software Foundation.
 13 */
 14
 15#include <linux/module.h>
 16#include <linux/init.h>
 17#include <linux/irq.h>
 18#include <linux/irqchip.h>
 19#include <linux/irqdomain.h>
 20#include <linux/io.h>
 21#include <linux/ioport.h>
 22#include <linux/of_address.h>
 23#include <linux/of_irq.h>
 24
 25#include <asm/exception.h>
 26#include <asm/hardirq.h>
 27
 
 
 28#define MAX_ICU_NR		16
 29
 30#define PJ1_INT_SEL		0x10c
 31#define PJ4_INT_SEL		0x104
 32
 33/* bit fields in PJ1_INT_SEL and PJ4_INT_SEL */
 34#define SEL_INT_PENDING		(1 << 6)
 35#define SEL_INT_NUM_MASK	0x3f
 36
 37struct icu_chip_data {
 38	int			nr_irqs;
 39	unsigned int		virq_base;
 40	unsigned int		cascade_irq;
 41	void __iomem		*reg_status;
 42	void __iomem		*reg_mask;
 43	unsigned int		conf_enable;
 44	unsigned int		conf_disable;
 45	unsigned int		conf_mask;
 46	unsigned int		clr_mfp_irq_base;
 47	unsigned int		clr_mfp_hwirq;
 48	struct irq_domain	*domain;
 49};
 50
 51struct mmp_intc_conf {
 52	unsigned int	conf_enable;
 53	unsigned int	conf_disable;
 54	unsigned int	conf_mask;
 55};
 56
 57static void __iomem *mmp_icu_base;
 58static struct icu_chip_data icu_data[MAX_ICU_NR];
 59static int max_icu_nr;
 60
 61extern void mmp2_clear_pmic_int(void);
 62
 63static void icu_mask_ack_irq(struct irq_data *d)
 64{
 65	struct irq_domain *domain = d->domain;
 66	struct icu_chip_data *data = (struct icu_chip_data *)domain->host_data;
 67	int hwirq;
 68	u32 r;
 69
 70	hwirq = d->irq - data->virq_base;
 71	if (data == &icu_data[0]) {
 72		r = readl_relaxed(mmp_icu_base + (hwirq << 2));
 73		r &= ~data->conf_mask;
 74		r |= data->conf_disable;
 75		writel_relaxed(r, mmp_icu_base + (hwirq << 2));
 76	} else {
 77#ifdef CONFIG_CPU_MMP2
 78		if ((data->virq_base == data->clr_mfp_irq_base)
 79			&& (hwirq == data->clr_mfp_hwirq))
 80			mmp2_clear_pmic_int();
 81#endif
 82		r = readl_relaxed(data->reg_mask) | (1 << hwirq);
 83		writel_relaxed(r, data->reg_mask);
 84	}
 85}
 86
 87static void icu_mask_irq(struct irq_data *d)
 88{
 89	struct irq_domain *domain = d->domain;
 90	struct icu_chip_data *data = (struct icu_chip_data *)domain->host_data;
 91	int hwirq;
 92	u32 r;
 93
 94	hwirq = d->irq - data->virq_base;
 95	if (data == &icu_data[0]) {
 96		r = readl_relaxed(mmp_icu_base + (hwirq << 2));
 97		r &= ~data->conf_mask;
 98		r |= data->conf_disable;
 99		writel_relaxed(r, mmp_icu_base + (hwirq << 2));
100	} else {
101		r = readl_relaxed(data->reg_mask) | (1 << hwirq);
102		writel_relaxed(r, data->reg_mask);
103	}
104}
105
106static void icu_unmask_irq(struct irq_data *d)
107{
108	struct irq_domain *domain = d->domain;
109	struct icu_chip_data *data = (struct icu_chip_data *)domain->host_data;
110	int hwirq;
111	u32 r;
112
113	hwirq = d->irq - data->virq_base;
114	if (data == &icu_data[0]) {
115		r = readl_relaxed(mmp_icu_base + (hwirq << 2));
116		r &= ~data->conf_mask;
117		r |= data->conf_enable;
118		writel_relaxed(r, mmp_icu_base + (hwirq << 2));
119	} else {
120		r = readl_relaxed(data->reg_mask) & ~(1 << hwirq);
121		writel_relaxed(r, data->reg_mask);
122	}
123}
124
125struct irq_chip icu_irq_chip = {
126	.name		= "icu_irq",
127	.irq_mask	= icu_mask_irq,
128	.irq_mask_ack	= icu_mask_ack_irq,
129	.irq_unmask	= icu_unmask_irq,
130};
131
132static void icu_mux_irq_demux(struct irq_desc *desc)
133{
134	unsigned int irq = irq_desc_get_irq(desc);
135	struct irq_domain *domain;
136	struct icu_chip_data *data;
137	int i;
138	unsigned long mask, status, n;
139
140	for (i = 1; i < max_icu_nr; i++) {
141		if (irq == icu_data[i].cascade_irq) {
142			domain = icu_data[i].domain;
143			data = (struct icu_chip_data *)domain->host_data;
144			break;
145		}
146	}
147	if (i >= max_icu_nr) {
148		pr_err("Spurious irq %d in MMP INTC\n", irq);
149		return;
150	}
151
152	mask = readl_relaxed(data->reg_mask);
153	while (1) {
154		status = readl_relaxed(data->reg_status) & ~mask;
155		if (status == 0)
156			break;
157		for_each_set_bit(n, &status, BITS_PER_LONG) {
158			generic_handle_irq(icu_data[i].virq_base + n);
159		}
160	}
161}
162
163static int mmp_irq_domain_map(struct irq_domain *d, unsigned int irq,
164			      irq_hw_number_t hw)
165{
166	irq_set_chip_and_handler(irq, &icu_irq_chip, handle_level_irq);
 
167	return 0;
168}
169
170static int mmp_irq_domain_xlate(struct irq_domain *d, struct device_node *node,
171				const u32 *intspec, unsigned int intsize,
172				unsigned long *out_hwirq,
173				unsigned int *out_type)
174{
175	*out_hwirq = intspec[0];
176	return 0;
177}
178
179const struct irq_domain_ops mmp_irq_domain_ops = {
180	.map		= mmp_irq_domain_map,
181	.xlate		= mmp_irq_domain_xlate,
182};
183
184static struct mmp_intc_conf mmp_conf = {
185	.conf_enable	= 0x51,
186	.conf_disable	= 0x0,
187	.conf_mask	= 0x7f,
188};
189
190static struct mmp_intc_conf mmp2_conf = {
191	.conf_enable	= 0x20,
192	.conf_disable	= 0x0,
193	.conf_mask	= 0x7f,
194};
195
196static void __exception_irq_entry mmp_handle_irq(struct pt_regs *regs)
197{
198	int hwirq;
199
200	hwirq = readl_relaxed(mmp_icu_base + PJ1_INT_SEL);
201	if (!(hwirq & SEL_INT_PENDING))
202		return;
203	hwirq &= SEL_INT_NUM_MASK;
204	handle_domain_irq(icu_data[0].domain, hwirq, regs);
 
205}
206
207static void __exception_irq_entry mmp2_handle_irq(struct pt_regs *regs)
208{
209	int hwirq;
210
211	hwirq = readl_relaxed(mmp_icu_base + PJ4_INT_SEL);
212	if (!(hwirq & SEL_INT_PENDING))
213		return;
214	hwirq &= SEL_INT_NUM_MASK;
215	handle_domain_irq(icu_data[0].domain, hwirq, regs);
 
216}
217
218/* MMP (ARMv5) */
219void __init icu_init_irq(void)
220{
221	int irq;
222
223	max_icu_nr = 1;
224	mmp_icu_base = ioremap(0xd4282000, 0x1000);
225	icu_data[0].conf_enable = mmp_conf.conf_enable;
226	icu_data[0].conf_disable = mmp_conf.conf_disable;
227	icu_data[0].conf_mask = mmp_conf.conf_mask;
228	icu_data[0].nr_irqs = 64;
229	icu_data[0].virq_base = 0;
230	icu_data[0].domain = irq_domain_add_legacy(NULL, 64, 0, 0,
231						   &irq_domain_simple_ops,
232						   &icu_data[0]);
233	for (irq = 0; irq < 64; irq++) {
234		icu_mask_irq(irq_get_irq_data(irq));
235		irq_set_chip_and_handler(irq, &icu_irq_chip, handle_level_irq);
 
236	}
237	irq_set_default_host(icu_data[0].domain);
238	set_handle_irq(mmp_handle_irq);
239}
240
241/* MMP2 (ARMv7) */
242void __init mmp2_init_icu(void)
243{
244	int irq, end;
245
246	max_icu_nr = 8;
247	mmp_icu_base = ioremap(0xd4282000, 0x1000);
248	icu_data[0].conf_enable = mmp2_conf.conf_enable;
249	icu_data[0].conf_disable = mmp2_conf.conf_disable;
250	icu_data[0].conf_mask = mmp2_conf.conf_mask;
251	icu_data[0].nr_irqs = 64;
252	icu_data[0].virq_base = 0;
253	icu_data[0].domain = irq_domain_add_legacy(NULL, 64, 0, 0,
254						   &irq_domain_simple_ops,
255						   &icu_data[0]);
256	icu_data[1].reg_status = mmp_icu_base + 0x150;
257	icu_data[1].reg_mask = mmp_icu_base + 0x168;
258	icu_data[1].clr_mfp_irq_base = icu_data[0].virq_base +
259				icu_data[0].nr_irqs;
260	icu_data[1].clr_mfp_hwirq = 1;		/* offset to IRQ_MMP2_PMIC_BASE */
261	icu_data[1].nr_irqs = 2;
262	icu_data[1].cascade_irq = 4;
263	icu_data[1].virq_base = icu_data[0].virq_base + icu_data[0].nr_irqs;
264	icu_data[1].domain = irq_domain_add_legacy(NULL, icu_data[1].nr_irqs,
265						   icu_data[1].virq_base, 0,
266						   &irq_domain_simple_ops,
267						   &icu_data[1]);
268	icu_data[2].reg_status = mmp_icu_base + 0x154;
269	icu_data[2].reg_mask = mmp_icu_base + 0x16c;
270	icu_data[2].nr_irqs = 2;
271	icu_data[2].cascade_irq = 5;
272	icu_data[2].virq_base = icu_data[1].virq_base + icu_data[1].nr_irqs;
273	icu_data[2].domain = irq_domain_add_legacy(NULL, icu_data[2].nr_irqs,
274						   icu_data[2].virq_base, 0,
275						   &irq_domain_simple_ops,
276						   &icu_data[2]);
277	icu_data[3].reg_status = mmp_icu_base + 0x180;
278	icu_data[3].reg_mask = mmp_icu_base + 0x17c;
279	icu_data[3].nr_irqs = 3;
280	icu_data[3].cascade_irq = 9;
281	icu_data[3].virq_base = icu_data[2].virq_base + icu_data[2].nr_irqs;
282	icu_data[3].domain = irq_domain_add_legacy(NULL, icu_data[3].nr_irqs,
283						   icu_data[3].virq_base, 0,
284						   &irq_domain_simple_ops,
285						   &icu_data[3]);
286	icu_data[4].reg_status = mmp_icu_base + 0x158;
287	icu_data[4].reg_mask = mmp_icu_base + 0x170;
288	icu_data[4].nr_irqs = 5;
289	icu_data[4].cascade_irq = 17;
290	icu_data[4].virq_base = icu_data[3].virq_base + icu_data[3].nr_irqs;
291	icu_data[4].domain = irq_domain_add_legacy(NULL, icu_data[4].nr_irqs,
292						   icu_data[4].virq_base, 0,
293						   &irq_domain_simple_ops,
294						   &icu_data[4]);
295	icu_data[5].reg_status = mmp_icu_base + 0x15c;
296	icu_data[5].reg_mask = mmp_icu_base + 0x174;
297	icu_data[5].nr_irqs = 15;
298	icu_data[5].cascade_irq = 35;
299	icu_data[5].virq_base = icu_data[4].virq_base + icu_data[4].nr_irqs;
300	icu_data[5].domain = irq_domain_add_legacy(NULL, icu_data[5].nr_irqs,
301						   icu_data[5].virq_base, 0,
302						   &irq_domain_simple_ops,
303						   &icu_data[5]);
304	icu_data[6].reg_status = mmp_icu_base + 0x160;
305	icu_data[6].reg_mask = mmp_icu_base + 0x178;
306	icu_data[6].nr_irqs = 2;
307	icu_data[6].cascade_irq = 51;
308	icu_data[6].virq_base = icu_data[5].virq_base + icu_data[5].nr_irqs;
309	icu_data[6].domain = irq_domain_add_legacy(NULL, icu_data[6].nr_irqs,
310						   icu_data[6].virq_base, 0,
311						   &irq_domain_simple_ops,
312						   &icu_data[6]);
313	icu_data[7].reg_status = mmp_icu_base + 0x188;
314	icu_data[7].reg_mask = mmp_icu_base + 0x184;
315	icu_data[7].nr_irqs = 2;
316	icu_data[7].cascade_irq = 55;
317	icu_data[7].virq_base = icu_data[6].virq_base + icu_data[6].nr_irqs;
318	icu_data[7].domain = irq_domain_add_legacy(NULL, icu_data[7].nr_irqs,
319						   icu_data[7].virq_base, 0,
320						   &irq_domain_simple_ops,
321						   &icu_data[7]);
322	end = icu_data[7].virq_base + icu_data[7].nr_irqs;
323	for (irq = 0; irq < end; irq++) {
324		icu_mask_irq(irq_get_irq_data(irq));
325		if (irq == icu_data[1].cascade_irq ||
326		    irq == icu_data[2].cascade_irq ||
327		    irq == icu_data[3].cascade_irq ||
328		    irq == icu_data[4].cascade_irq ||
329		    irq == icu_data[5].cascade_irq ||
330		    irq == icu_data[6].cascade_irq ||
331		    irq == icu_data[7].cascade_irq) {
332			irq_set_chip(irq, &icu_irq_chip);
333			irq_set_chained_handler(irq, icu_mux_irq_demux);
334		} else {
335			irq_set_chip_and_handler(irq, &icu_irq_chip,
336						 handle_level_irq);
337		}
 
338	}
339	irq_set_default_host(icu_data[0].domain);
340	set_handle_irq(mmp2_handle_irq);
341}
342
343#ifdef CONFIG_OF
344static int __init mmp_init_bases(struct device_node *node)
345{
346	int ret, nr_irqs, irq, i = 0;
347
348	ret = of_property_read_u32(node, "mrvl,intc-nr-irqs", &nr_irqs);
349	if (ret) {
350		pr_err("Not found mrvl,intc-nr-irqs property\n");
351		return ret;
352	}
353
354	mmp_icu_base = of_iomap(node, 0);
355	if (!mmp_icu_base) {
356		pr_err("Failed to get interrupt controller register\n");
357		return -ENOMEM;
358	}
359
360	icu_data[0].virq_base = 0;
361	icu_data[0].domain = irq_domain_add_linear(node, nr_irqs,
362						   &mmp_irq_domain_ops,
363						   &icu_data[0]);
364	for (irq = 0; irq < nr_irqs; irq++) {
365		ret = irq_create_mapping(icu_data[0].domain, irq);
366		if (!ret) {
367			pr_err("Failed to mapping hwirq\n");
368			goto err;
369		}
370		if (!irq)
371			icu_data[0].virq_base = ret;
372	}
373	icu_data[0].nr_irqs = nr_irqs;
374	return 0;
375err:
376	if (icu_data[0].virq_base) {
377		for (i = 0; i < irq; i++)
378			irq_dispose_mapping(icu_data[0].virq_base + i);
379	}
380	irq_domain_remove(icu_data[0].domain);
381	iounmap(mmp_icu_base);
382	return -EINVAL;
383}
384
385static int __init mmp_of_init(struct device_node *node,
386			      struct device_node *parent)
387{
388	int ret;
389
390	ret = mmp_init_bases(node);
391	if (ret < 0)
392		return ret;
393
394	icu_data[0].conf_enable = mmp_conf.conf_enable;
395	icu_data[0].conf_disable = mmp_conf.conf_disable;
396	icu_data[0].conf_mask = mmp_conf.conf_mask;
397	irq_set_default_host(icu_data[0].domain);
398	set_handle_irq(mmp_handle_irq);
399	max_icu_nr = 1;
400	return 0;
401}
402IRQCHIP_DECLARE(mmp_intc, "mrvl,mmp-intc", mmp_of_init);
403
404static int __init mmp2_of_init(struct device_node *node,
405			       struct device_node *parent)
406{
407	int ret;
408
409	ret = mmp_init_bases(node);
410	if (ret < 0)
411		return ret;
412
413	icu_data[0].conf_enable = mmp2_conf.conf_enable;
414	icu_data[0].conf_disable = mmp2_conf.conf_disable;
415	icu_data[0].conf_mask = mmp2_conf.conf_mask;
416	irq_set_default_host(icu_data[0].domain);
417	set_handle_irq(mmp2_handle_irq);
418	max_icu_nr = 1;
419	return 0;
420}
421IRQCHIP_DECLARE(mmp2_intc, "mrvl,mmp2-intc", mmp2_of_init);
422
423static int __init mmp2_mux_of_init(struct device_node *node,
424				   struct device_node *parent)
425{
426	struct resource res;
427	int i, ret, irq, j = 0;
428	u32 nr_irqs, mfp_irq;
429
430	if (!parent)
431		return -ENODEV;
432
433	i = max_icu_nr;
434	ret = of_property_read_u32(node, "mrvl,intc-nr-irqs",
435				   &nr_irqs);
436	if (ret) {
437		pr_err("Not found mrvl,intc-nr-irqs property\n");
438		return -EINVAL;
439	}
440	ret = of_address_to_resource(node, 0, &res);
441	if (ret < 0) {
442		pr_err("Not found reg property\n");
443		return -EINVAL;
444	}
445	icu_data[i].reg_status = mmp_icu_base + res.start;
446	ret = of_address_to_resource(node, 1, &res);
447	if (ret < 0) {
448		pr_err("Not found reg property\n");
449		return -EINVAL;
450	}
451	icu_data[i].reg_mask = mmp_icu_base + res.start;
452	icu_data[i].cascade_irq = irq_of_parse_and_map(node, 0);
453	if (!icu_data[i].cascade_irq)
454		return -EINVAL;
455
456	icu_data[i].virq_base = 0;
457	icu_data[i].domain = irq_domain_add_linear(node, nr_irqs,
458						   &mmp_irq_domain_ops,
459						   &icu_data[i]);
460	for (irq = 0; irq < nr_irqs; irq++) {
461		ret = irq_create_mapping(icu_data[i].domain, irq);
462		if (!ret) {
463			pr_err("Failed to mapping hwirq\n");
464			goto err;
465		}
466		if (!irq)
467			icu_data[i].virq_base = ret;
468	}
469	icu_data[i].nr_irqs = nr_irqs;
470	if (!of_property_read_u32(node, "mrvl,clr-mfp-irq",
471				  &mfp_irq)) {
472		icu_data[i].clr_mfp_irq_base = icu_data[i].virq_base;
473		icu_data[i].clr_mfp_hwirq = mfp_irq;
474	}
475	irq_set_chained_handler(icu_data[i].cascade_irq,
476				icu_mux_irq_demux);
477	max_icu_nr++;
478	return 0;
479err:
480	if (icu_data[i].virq_base) {
481		for (j = 0; j < irq; j++)
482			irq_dispose_mapping(icu_data[i].virq_base + j);
483	}
484	irq_domain_remove(icu_data[i].domain);
485	return -EINVAL;
486}
487IRQCHIP_DECLARE(mmp2_mux_intc, "mrvl,mmp2-mux-intc", mmp2_mux_of_init);
488#endif