Linux Audio

Check our new training course

Loading...
v6.8
  1// SPDX-License-Identifier: GPL-2.0-only
  2/*
  3 *  linux/arch/arm/mach-mmp/irq.c
  4 *
  5 *  Generic IRQ handling, GPIO IRQ demultiplexing, etc.
  6 *  Copyright (C) 2008 - 2012 Marvell Technology Group Ltd.
  7 *
  8 *  Author:	Bin Yang <bin.yang@marvell.com>
  9 *              Haojian Zhuang <haojian.zhuang@gmail.com>
 
 
 
 
 10 */
 11
 12#include <linux/module.h>
 13#include <linux/init.h>
 14#include <linux/irq.h>
 15#include <linux/irqchip.h>
 16#include <linux/irqchip/chained_irq.h>
 17#include <linux/irqdomain.h>
 18#include <linux/io.h>
 19#include <linux/ioport.h>
 20#include <linux/of_address.h>
 21#include <linux/of_irq.h>
 22
 23#include <asm/exception.h>
 24#include <asm/hardirq.h>
 25
 
 
 26#define MAX_ICU_NR		16
 27
 28#define PJ1_INT_SEL		0x10c
 29#define PJ4_INT_SEL		0x104
 30
 31/* bit fields in PJ1_INT_SEL and PJ4_INT_SEL */
 32#define SEL_INT_PENDING		(1 << 6)
 33#define SEL_INT_NUM_MASK	0x3f
 34
 35#define MMP2_ICU_INT_ROUTE_PJ4_IRQ	(1 << 5)
 36#define MMP2_ICU_INT_ROUTE_PJ4_FIQ	(1 << 6)
 37
 38struct icu_chip_data {
 39	int			nr_irqs;
 40	unsigned int		virq_base;
 41	unsigned int		cascade_irq;
 42	void __iomem		*reg_status;
 43	void __iomem		*reg_mask;
 44	unsigned int		conf_enable;
 45	unsigned int		conf_disable;
 46	unsigned int		conf_mask;
 47	unsigned int		conf2_mask;
 48	unsigned int		clr_mfp_irq_base;
 49	unsigned int		clr_mfp_hwirq;
 50	struct irq_domain	*domain;
 51};
 52
 53struct mmp_intc_conf {
 54	unsigned int	conf_enable;
 55	unsigned int	conf_disable;
 56	unsigned int	conf_mask;
 57	unsigned int	conf2_mask;
 58};
 59
 60static void __iomem *mmp_icu_base;
 61static void __iomem *mmp_icu2_base;
 62static struct icu_chip_data icu_data[MAX_ICU_NR];
 63static int max_icu_nr;
 64
 65extern void mmp2_clear_pmic_int(void);
 66
 67static void icu_mask_ack_irq(struct irq_data *d)
 68{
 69	struct irq_domain *domain = d->domain;
 70	struct icu_chip_data *data = (struct icu_chip_data *)domain->host_data;
 71	int hwirq;
 72	u32 r;
 73
 74	hwirq = d->irq - data->virq_base;
 75	if (data == &icu_data[0]) {
 76		r = readl_relaxed(mmp_icu_base + (hwirq << 2));
 77		r &= ~data->conf_mask;
 78		r |= data->conf_disable;
 79		writel_relaxed(r, mmp_icu_base + (hwirq << 2));
 80	} else {
 81#ifdef CONFIG_CPU_MMP2
 82		if ((data->virq_base == data->clr_mfp_irq_base)
 83			&& (hwirq == data->clr_mfp_hwirq))
 84			mmp2_clear_pmic_int();
 85#endif
 86		r = readl_relaxed(data->reg_mask) | (1 << hwirq);
 87		writel_relaxed(r, data->reg_mask);
 88	}
 89}
 90
 91static void icu_mask_irq(struct irq_data *d)
 92{
 93	struct irq_domain *domain = d->domain;
 94	struct icu_chip_data *data = (struct icu_chip_data *)domain->host_data;
 95	int hwirq;
 96	u32 r;
 97
 98	hwirq = d->irq - data->virq_base;
 99	if (data == &icu_data[0]) {
100		r = readl_relaxed(mmp_icu_base + (hwirq << 2));
101		r &= ~data->conf_mask;
102		r |= data->conf_disable;
103		writel_relaxed(r, mmp_icu_base + (hwirq << 2));
104
105		if (data->conf2_mask) {
106			/*
107			 * ICU1 (above) only controls PJ4 MP1; if using SMP,
108			 * we need to also mask the MP2 and MM cores via ICU2.
109			 */
110			r = readl_relaxed(mmp_icu2_base + (hwirq << 2));
111			r &= ~data->conf2_mask;
112			writel_relaxed(r, mmp_icu2_base + (hwirq << 2));
113		}
114	} else {
115		r = readl_relaxed(data->reg_mask) | (1 << hwirq);
116		writel_relaxed(r, data->reg_mask);
117	}
118}
119
120static void icu_unmask_irq(struct irq_data *d)
121{
122	struct irq_domain *domain = d->domain;
123	struct icu_chip_data *data = (struct icu_chip_data *)domain->host_data;
124	int hwirq;
125	u32 r;
126
127	hwirq = d->irq - data->virq_base;
128	if (data == &icu_data[0]) {
129		r = readl_relaxed(mmp_icu_base + (hwirq << 2));
130		r &= ~data->conf_mask;
131		r |= data->conf_enable;
132		writel_relaxed(r, mmp_icu_base + (hwirq << 2));
133	} else {
134		r = readl_relaxed(data->reg_mask) & ~(1 << hwirq);
135		writel_relaxed(r, data->reg_mask);
136	}
137}
138
139struct irq_chip icu_irq_chip = {
140	.name		= "icu_irq",
141	.irq_mask	= icu_mask_irq,
142	.irq_mask_ack	= icu_mask_ack_irq,
143	.irq_unmask	= icu_unmask_irq,
144};
145
146static void icu_mux_irq_demux(struct irq_desc *desc)
147{
148	unsigned int irq = irq_desc_get_irq(desc);
149	struct irq_chip *chip = irq_desc_get_chip(desc);
150	struct irq_domain *domain;
151	struct icu_chip_data *data;
152	int i;
153	unsigned long mask, status, n;
154
155	chained_irq_enter(chip, desc);
156
157	for (i = 1; i < max_icu_nr; i++) {
158		if (irq == icu_data[i].cascade_irq) {
159			domain = icu_data[i].domain;
160			data = (struct icu_chip_data *)domain->host_data;
161			break;
162		}
163	}
164	if (i >= max_icu_nr) {
165		pr_err("Spurious irq %d in MMP INTC\n", irq);
166		goto out;
167	}
168
169	mask = readl_relaxed(data->reg_mask);
170	while (1) {
171		status = readl_relaxed(data->reg_status) & ~mask;
172		if (status == 0)
173			break;
174		for_each_set_bit(n, &status, BITS_PER_LONG) {
175			generic_handle_irq(icu_data[i].virq_base + n);
176		}
177	}
178
179out:
180	chained_irq_exit(chip, desc);
181}
182
183static int mmp_irq_domain_map(struct irq_domain *d, unsigned int irq,
184			      irq_hw_number_t hw)
185{
186	irq_set_chip_and_handler(irq, &icu_irq_chip, handle_level_irq);
 
187	return 0;
188}
189
190static int mmp_irq_domain_xlate(struct irq_domain *d, struct device_node *node,
191				const u32 *intspec, unsigned int intsize,
192				unsigned long *out_hwirq,
193				unsigned int *out_type)
194{
195	*out_hwirq = intspec[0];
196	return 0;
197}
198
199static const struct irq_domain_ops mmp_irq_domain_ops = {
200	.map		= mmp_irq_domain_map,
201	.xlate		= mmp_irq_domain_xlate,
202};
203
204static const struct mmp_intc_conf mmp_conf = {
205	.conf_enable	= 0x51,
206	.conf_disable	= 0x0,
207	.conf_mask	= 0x7f,
208};
209
210static const struct mmp_intc_conf mmp2_conf = {
211	.conf_enable	= 0x20,
212	.conf_disable	= 0x0,
213	.conf_mask	= MMP2_ICU_INT_ROUTE_PJ4_IRQ |
214			  MMP2_ICU_INT_ROUTE_PJ4_FIQ,
215};
216
217static struct mmp_intc_conf mmp3_conf = {
218	.conf_enable	= 0x20,
219	.conf_disable	= 0x0,
220	.conf_mask	= MMP2_ICU_INT_ROUTE_PJ4_IRQ |
221			  MMP2_ICU_INT_ROUTE_PJ4_FIQ,
222	.conf2_mask	= 0xf0,
223};
224
225static void __exception_irq_entry mmp_handle_irq(struct pt_regs *regs)
226{
227	int hwirq;
228
229	hwirq = readl_relaxed(mmp_icu_base + PJ1_INT_SEL);
230	if (!(hwirq & SEL_INT_PENDING))
231		return;
232	hwirq &= SEL_INT_NUM_MASK;
233	generic_handle_domain_irq(icu_data[0].domain, hwirq);
 
234}
235
236static void __exception_irq_entry mmp2_handle_irq(struct pt_regs *regs)
237{
238	int hwirq;
239
240	hwirq = readl_relaxed(mmp_icu_base + PJ4_INT_SEL);
241	if (!(hwirq & SEL_INT_PENDING))
242		return;
243	hwirq &= SEL_INT_NUM_MASK;
244	generic_handle_domain_irq(icu_data[0].domain, hwirq);
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
245}
246
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
247static int __init mmp_init_bases(struct device_node *node)
248{
249	int ret, nr_irqs, irq, i = 0;
250
251	ret = of_property_read_u32(node, "mrvl,intc-nr-irqs", &nr_irqs);
252	if (ret) {
253		pr_err("Not found mrvl,intc-nr-irqs property\n");
254		return ret;
255	}
256
257	mmp_icu_base = of_iomap(node, 0);
258	if (!mmp_icu_base) {
259		pr_err("Failed to get interrupt controller register\n");
260		return -ENOMEM;
261	}
262
263	icu_data[0].virq_base = 0;
264	icu_data[0].domain = irq_domain_add_linear(node, nr_irqs,
265						   &mmp_irq_domain_ops,
266						   &icu_data[0]);
267	for (irq = 0; irq < nr_irqs; irq++) {
268		ret = irq_create_mapping(icu_data[0].domain, irq);
269		if (!ret) {
270			pr_err("Failed to mapping hwirq\n");
271			goto err;
272		}
273		if (!irq)
274			icu_data[0].virq_base = ret;
275	}
276	icu_data[0].nr_irqs = nr_irqs;
277	return 0;
278err:
279	if (icu_data[0].virq_base) {
280		for (i = 0; i < irq; i++)
281			irq_dispose_mapping(icu_data[0].virq_base + i);
282	}
283	irq_domain_remove(icu_data[0].domain);
284	iounmap(mmp_icu_base);
285	return -EINVAL;
286}
287
288static int __init mmp_of_init(struct device_node *node,
289			      struct device_node *parent)
290{
291	int ret;
292
293	ret = mmp_init_bases(node);
294	if (ret < 0)
295		return ret;
296
297	icu_data[0].conf_enable = mmp_conf.conf_enable;
298	icu_data[0].conf_disable = mmp_conf.conf_disable;
299	icu_data[0].conf_mask = mmp_conf.conf_mask;
 
300	set_handle_irq(mmp_handle_irq);
301	max_icu_nr = 1;
302	return 0;
303}
304IRQCHIP_DECLARE(mmp_intc, "mrvl,mmp-intc", mmp_of_init);
305
306static int __init mmp2_of_init(struct device_node *node,
307			       struct device_node *parent)
308{
309	int ret;
310
311	ret = mmp_init_bases(node);
312	if (ret < 0)
313		return ret;
314
315	icu_data[0].conf_enable = mmp2_conf.conf_enable;
316	icu_data[0].conf_disable = mmp2_conf.conf_disable;
317	icu_data[0].conf_mask = mmp2_conf.conf_mask;
 
318	set_handle_irq(mmp2_handle_irq);
319	max_icu_nr = 1;
320	return 0;
321}
322IRQCHIP_DECLARE(mmp2_intc, "mrvl,mmp2-intc", mmp2_of_init);
323
324static int __init mmp3_of_init(struct device_node *node,
325			       struct device_node *parent)
326{
327	int ret;
328
329	mmp_icu2_base = of_iomap(node, 1);
330	if (!mmp_icu2_base) {
331		pr_err("Failed to get interrupt controller register #2\n");
332		return -ENODEV;
333	}
334
335	ret = mmp_init_bases(node);
336	if (ret < 0) {
337		iounmap(mmp_icu2_base);
338		return ret;
339	}
340
341	icu_data[0].conf_enable = mmp3_conf.conf_enable;
342	icu_data[0].conf_disable = mmp3_conf.conf_disable;
343	icu_data[0].conf_mask = mmp3_conf.conf_mask;
344	icu_data[0].conf2_mask = mmp3_conf.conf2_mask;
345
346	if (!parent) {
347		/* This is the main interrupt controller. */
348		set_handle_irq(mmp2_handle_irq);
349	}
350
351	max_icu_nr = 1;
352	return 0;
353}
354IRQCHIP_DECLARE(mmp3_intc, "marvell,mmp3-intc", mmp3_of_init);
355
356static int __init mmp2_mux_of_init(struct device_node *node,
357				   struct device_node *parent)
358{
 
359	int i, ret, irq, j = 0;
360	u32 nr_irqs, mfp_irq;
361	u32 reg[4];
362
363	if (!parent)
364		return -ENODEV;
365
366	i = max_icu_nr;
367	ret = of_property_read_u32(node, "mrvl,intc-nr-irqs",
368				   &nr_irqs);
369	if (ret) {
370		pr_err("Not found mrvl,intc-nr-irqs property\n");
371		return -EINVAL;
372	}
373
374	/*
375	 * For historical reasons, the "regs" property of the
376	 * mrvl,mmp2-mux-intc is not a regular "regs" property containing
377	 * addresses on the parent bus, but offsets from the intc's base.
378	 * That is why we can't use of_address_to_resource() here.
379	 */
380	ret = of_property_read_variable_u32_array(node, "reg", reg,
381						  ARRAY_SIZE(reg),
382						  ARRAY_SIZE(reg));
383	if (ret < 0) {
384		pr_err("Not found reg property\n");
385		return -EINVAL;
386	}
387	icu_data[i].reg_status = mmp_icu_base + reg[0];
388	icu_data[i].reg_mask = mmp_icu_base + reg[2];
 
 
 
 
 
389	icu_data[i].cascade_irq = irq_of_parse_and_map(node, 0);
390	if (!icu_data[i].cascade_irq)
391		return -EINVAL;
392
393	icu_data[i].virq_base = 0;
394	icu_data[i].domain = irq_domain_add_linear(node, nr_irqs,
395						   &mmp_irq_domain_ops,
396						   &icu_data[i]);
397	for (irq = 0; irq < nr_irqs; irq++) {
398		ret = irq_create_mapping(icu_data[i].domain, irq);
399		if (!ret) {
400			pr_err("Failed to mapping hwirq\n");
401			goto err;
402		}
403		if (!irq)
404			icu_data[i].virq_base = ret;
405	}
406	icu_data[i].nr_irqs = nr_irqs;
407	if (!of_property_read_u32(node, "mrvl,clr-mfp-irq",
408				  &mfp_irq)) {
409		icu_data[i].clr_mfp_irq_base = icu_data[i].virq_base;
410		icu_data[i].clr_mfp_hwirq = mfp_irq;
411	}
412	irq_set_chained_handler(icu_data[i].cascade_irq,
413				icu_mux_irq_demux);
414	max_icu_nr++;
415	return 0;
416err:
417	if (icu_data[i].virq_base) {
418		for (j = 0; j < irq; j++)
419			irq_dispose_mapping(icu_data[i].virq_base + j);
420	}
421	irq_domain_remove(icu_data[i].domain);
422	return -EINVAL;
423}
424IRQCHIP_DECLARE(mmp2_mux_intc, "mrvl,mmp2-mux-intc", mmp2_mux_of_init);
v3.15
 
  1/*
  2 *  linux/arch/arm/mach-mmp/irq.c
  3 *
  4 *  Generic IRQ handling, GPIO IRQ demultiplexing, etc.
  5 *  Copyright (C) 2008 - 2012 Marvell Technology Group Ltd.
  6 *
  7 *  Author:	Bin Yang <bin.yang@marvell.com>
  8 *              Haojian Zhuang <haojian.zhuang@gmail.com>
  9 *
 10 *  This program is free software; you can redistribute it and/or modify
 11 *  it under the terms of the GNU General Public License version 2 as
 12 *  published by the Free Software Foundation.
 13 */
 14
 15#include <linux/module.h>
 16#include <linux/init.h>
 17#include <linux/irq.h>
 
 
 18#include <linux/irqdomain.h>
 19#include <linux/io.h>
 20#include <linux/ioport.h>
 21#include <linux/of_address.h>
 22#include <linux/of_irq.h>
 23
 24#include <asm/exception.h>
 25#include <asm/hardirq.h>
 26
 27#include "irqchip.h"
 28
 29#define MAX_ICU_NR		16
 30
 31#define PJ1_INT_SEL		0x10c
 32#define PJ4_INT_SEL		0x104
 33
 34/* bit fields in PJ1_INT_SEL and PJ4_INT_SEL */
 35#define SEL_INT_PENDING		(1 << 6)
 36#define SEL_INT_NUM_MASK	0x3f
 37
 
 
 
 38struct icu_chip_data {
 39	int			nr_irqs;
 40	unsigned int		virq_base;
 41	unsigned int		cascade_irq;
 42	void __iomem		*reg_status;
 43	void __iomem		*reg_mask;
 44	unsigned int		conf_enable;
 45	unsigned int		conf_disable;
 46	unsigned int		conf_mask;
 
 47	unsigned int		clr_mfp_irq_base;
 48	unsigned int		clr_mfp_hwirq;
 49	struct irq_domain	*domain;
 50};
 51
 52struct mmp_intc_conf {
 53	unsigned int	conf_enable;
 54	unsigned int	conf_disable;
 55	unsigned int	conf_mask;
 
 56};
 57
 58static void __iomem *mmp_icu_base;
 
 59static struct icu_chip_data icu_data[MAX_ICU_NR];
 60static int max_icu_nr;
 61
 62extern void mmp2_clear_pmic_int(void);
 63
 64static void icu_mask_ack_irq(struct irq_data *d)
 65{
 66	struct irq_domain *domain = d->domain;
 67	struct icu_chip_data *data = (struct icu_chip_data *)domain->host_data;
 68	int hwirq;
 69	u32 r;
 70
 71	hwirq = d->irq - data->virq_base;
 72	if (data == &icu_data[0]) {
 73		r = readl_relaxed(mmp_icu_base + (hwirq << 2));
 74		r &= ~data->conf_mask;
 75		r |= data->conf_disable;
 76		writel_relaxed(r, mmp_icu_base + (hwirq << 2));
 77	} else {
 78#ifdef CONFIG_CPU_MMP2
 79		if ((data->virq_base == data->clr_mfp_irq_base)
 80			&& (hwirq == data->clr_mfp_hwirq))
 81			mmp2_clear_pmic_int();
 82#endif
 83		r = readl_relaxed(data->reg_mask) | (1 << hwirq);
 84		writel_relaxed(r, data->reg_mask);
 85	}
 86}
 87
 88static void icu_mask_irq(struct irq_data *d)
 89{
 90	struct irq_domain *domain = d->domain;
 91	struct icu_chip_data *data = (struct icu_chip_data *)domain->host_data;
 92	int hwirq;
 93	u32 r;
 94
 95	hwirq = d->irq - data->virq_base;
 96	if (data == &icu_data[0]) {
 97		r = readl_relaxed(mmp_icu_base + (hwirq << 2));
 98		r &= ~data->conf_mask;
 99		r |= data->conf_disable;
100		writel_relaxed(r, mmp_icu_base + (hwirq << 2));
 
 
 
 
 
 
 
 
 
 
101	} else {
102		r = readl_relaxed(data->reg_mask) | (1 << hwirq);
103		writel_relaxed(r, data->reg_mask);
104	}
105}
106
107static void icu_unmask_irq(struct irq_data *d)
108{
109	struct irq_domain *domain = d->domain;
110	struct icu_chip_data *data = (struct icu_chip_data *)domain->host_data;
111	int hwirq;
112	u32 r;
113
114	hwirq = d->irq - data->virq_base;
115	if (data == &icu_data[0]) {
116		r = readl_relaxed(mmp_icu_base + (hwirq << 2));
117		r &= ~data->conf_mask;
118		r |= data->conf_enable;
119		writel_relaxed(r, mmp_icu_base + (hwirq << 2));
120	} else {
121		r = readl_relaxed(data->reg_mask) & ~(1 << hwirq);
122		writel_relaxed(r, data->reg_mask);
123	}
124}
125
126struct irq_chip icu_irq_chip = {
127	.name		= "icu_irq",
128	.irq_mask	= icu_mask_irq,
129	.irq_mask_ack	= icu_mask_ack_irq,
130	.irq_unmask	= icu_unmask_irq,
131};
132
133static void icu_mux_irq_demux(unsigned int irq, struct irq_desc *desc)
134{
 
 
135	struct irq_domain *domain;
136	struct icu_chip_data *data;
137	int i;
138	unsigned long mask, status, n;
139
 
 
140	for (i = 1; i < max_icu_nr; i++) {
141		if (irq == icu_data[i].cascade_irq) {
142			domain = icu_data[i].domain;
143			data = (struct icu_chip_data *)domain->host_data;
144			break;
145		}
146	}
147	if (i >= max_icu_nr) {
148		pr_err("Spurious irq %d in MMP INTC\n", irq);
149		return;
150	}
151
152	mask = readl_relaxed(data->reg_mask);
153	while (1) {
154		status = readl_relaxed(data->reg_status) & ~mask;
155		if (status == 0)
156			break;
157		for_each_set_bit(n, &status, BITS_PER_LONG) {
158			generic_handle_irq(icu_data[i].virq_base + n);
159		}
160	}
 
 
 
161}
162
163static int mmp_irq_domain_map(struct irq_domain *d, unsigned int irq,
164			      irq_hw_number_t hw)
165{
166	irq_set_chip_and_handler(irq, &icu_irq_chip, handle_level_irq);
167	set_irq_flags(irq, IRQF_VALID);
168	return 0;
169}
170
171static int mmp_irq_domain_xlate(struct irq_domain *d, struct device_node *node,
172				const u32 *intspec, unsigned int intsize,
173				unsigned long *out_hwirq,
174				unsigned int *out_type)
175{
176	*out_hwirq = intspec[0];
177	return 0;
178}
179
180const struct irq_domain_ops mmp_irq_domain_ops = {
181	.map		= mmp_irq_domain_map,
182	.xlate		= mmp_irq_domain_xlate,
183};
184
185static struct mmp_intc_conf mmp_conf = {
186	.conf_enable	= 0x51,
187	.conf_disable	= 0x0,
188	.conf_mask	= 0x7f,
189};
190
191static struct mmp_intc_conf mmp2_conf = {
192	.conf_enable	= 0x20,
193	.conf_disable	= 0x0,
194	.conf_mask	= 0x7f,
 
 
 
 
 
 
 
 
 
195};
196
197static void __exception_irq_entry mmp_handle_irq(struct pt_regs *regs)
198{
199	int irq, hwirq;
200
201	hwirq = readl_relaxed(mmp_icu_base + PJ1_INT_SEL);
202	if (!(hwirq & SEL_INT_PENDING))
203		return;
204	hwirq &= SEL_INT_NUM_MASK;
205	irq = irq_find_mapping(icu_data[0].domain, hwirq);
206	handle_IRQ(irq, regs);
207}
208
209static void __exception_irq_entry mmp2_handle_irq(struct pt_regs *regs)
210{
211	int irq, hwirq;
212
213	hwirq = readl_relaxed(mmp_icu_base + PJ4_INT_SEL);
214	if (!(hwirq & SEL_INT_PENDING))
215		return;
216	hwirq &= SEL_INT_NUM_MASK;
217	irq = irq_find_mapping(icu_data[0].domain, hwirq);
218	handle_IRQ(irq, regs);
219}
220
221/* MMP (ARMv5) */
222void __init icu_init_irq(void)
223{
224	int irq;
225
226	max_icu_nr = 1;
227	mmp_icu_base = ioremap(0xd4282000, 0x1000);
228	icu_data[0].conf_enable = mmp_conf.conf_enable;
229	icu_data[0].conf_disable = mmp_conf.conf_disable;
230	icu_data[0].conf_mask = mmp_conf.conf_mask;
231	icu_data[0].nr_irqs = 64;
232	icu_data[0].virq_base = 0;
233	icu_data[0].domain = irq_domain_add_legacy(NULL, 64, 0, 0,
234						   &irq_domain_simple_ops,
235						   &icu_data[0]);
236	for (irq = 0; irq < 64; irq++) {
237		icu_mask_irq(irq_get_irq_data(irq));
238		irq_set_chip_and_handler(irq, &icu_irq_chip, handle_level_irq);
239		set_irq_flags(irq, IRQF_VALID);
240	}
241	irq_set_default_host(icu_data[0].domain);
242	set_handle_irq(mmp_handle_irq);
243}
244
245/* MMP2 (ARMv7) */
246void __init mmp2_init_icu(void)
247{
248	int irq, end;
249
250	max_icu_nr = 8;
251	mmp_icu_base = ioremap(0xd4282000, 0x1000);
252	icu_data[0].conf_enable = mmp2_conf.conf_enable;
253	icu_data[0].conf_disable = mmp2_conf.conf_disable;
254	icu_data[0].conf_mask = mmp2_conf.conf_mask;
255	icu_data[0].nr_irqs = 64;
256	icu_data[0].virq_base = 0;
257	icu_data[0].domain = irq_domain_add_legacy(NULL, 64, 0, 0,
258						   &irq_domain_simple_ops,
259						   &icu_data[0]);
260	icu_data[1].reg_status = mmp_icu_base + 0x150;
261	icu_data[1].reg_mask = mmp_icu_base + 0x168;
262	icu_data[1].clr_mfp_irq_base = icu_data[0].virq_base +
263				icu_data[0].nr_irqs;
264	icu_data[1].clr_mfp_hwirq = 1;		/* offset to IRQ_MMP2_PMIC_BASE */
265	icu_data[1].nr_irqs = 2;
266	icu_data[1].cascade_irq = 4;
267	icu_data[1].virq_base = icu_data[0].virq_base + icu_data[0].nr_irqs;
268	icu_data[1].domain = irq_domain_add_legacy(NULL, icu_data[1].nr_irqs,
269						   icu_data[1].virq_base, 0,
270						   &irq_domain_simple_ops,
271						   &icu_data[1]);
272	icu_data[2].reg_status = mmp_icu_base + 0x154;
273	icu_data[2].reg_mask = mmp_icu_base + 0x16c;
274	icu_data[2].nr_irqs = 2;
275	icu_data[2].cascade_irq = 5;
276	icu_data[2].virq_base = icu_data[1].virq_base + icu_data[1].nr_irqs;
277	icu_data[2].domain = irq_domain_add_legacy(NULL, icu_data[2].nr_irqs,
278						   icu_data[2].virq_base, 0,
279						   &irq_domain_simple_ops,
280						   &icu_data[2]);
281	icu_data[3].reg_status = mmp_icu_base + 0x180;
282	icu_data[3].reg_mask = mmp_icu_base + 0x17c;
283	icu_data[3].nr_irqs = 3;
284	icu_data[3].cascade_irq = 9;
285	icu_data[3].virq_base = icu_data[2].virq_base + icu_data[2].nr_irqs;
286	icu_data[3].domain = irq_domain_add_legacy(NULL, icu_data[3].nr_irqs,
287						   icu_data[3].virq_base, 0,
288						   &irq_domain_simple_ops,
289						   &icu_data[3]);
290	icu_data[4].reg_status = mmp_icu_base + 0x158;
291	icu_data[4].reg_mask = mmp_icu_base + 0x170;
292	icu_data[4].nr_irqs = 5;
293	icu_data[4].cascade_irq = 17;
294	icu_data[4].virq_base = icu_data[3].virq_base + icu_data[3].nr_irqs;
295	icu_data[4].domain = irq_domain_add_legacy(NULL, icu_data[4].nr_irqs,
296						   icu_data[4].virq_base, 0,
297						   &irq_domain_simple_ops,
298						   &icu_data[4]);
299	icu_data[5].reg_status = mmp_icu_base + 0x15c;
300	icu_data[5].reg_mask = mmp_icu_base + 0x174;
301	icu_data[5].nr_irqs = 15;
302	icu_data[5].cascade_irq = 35;
303	icu_data[5].virq_base = icu_data[4].virq_base + icu_data[4].nr_irqs;
304	icu_data[5].domain = irq_domain_add_legacy(NULL, icu_data[5].nr_irqs,
305						   icu_data[5].virq_base, 0,
306						   &irq_domain_simple_ops,
307						   &icu_data[5]);
308	icu_data[6].reg_status = mmp_icu_base + 0x160;
309	icu_data[6].reg_mask = mmp_icu_base + 0x178;
310	icu_data[6].nr_irqs = 2;
311	icu_data[6].cascade_irq = 51;
312	icu_data[6].virq_base = icu_data[5].virq_base + icu_data[5].nr_irqs;
313	icu_data[6].domain = irq_domain_add_legacy(NULL, icu_data[6].nr_irqs,
314						   icu_data[6].virq_base, 0,
315						   &irq_domain_simple_ops,
316						   &icu_data[6]);
317	icu_data[7].reg_status = mmp_icu_base + 0x188;
318	icu_data[7].reg_mask = mmp_icu_base + 0x184;
319	icu_data[7].nr_irqs = 2;
320	icu_data[7].cascade_irq = 55;
321	icu_data[7].virq_base = icu_data[6].virq_base + icu_data[6].nr_irqs;
322	icu_data[7].domain = irq_domain_add_legacy(NULL, icu_data[7].nr_irqs,
323						   icu_data[7].virq_base, 0,
324						   &irq_domain_simple_ops,
325						   &icu_data[7]);
326	end = icu_data[7].virq_base + icu_data[7].nr_irqs;
327	for (irq = 0; irq < end; irq++) {
328		icu_mask_irq(irq_get_irq_data(irq));
329		if (irq == icu_data[1].cascade_irq ||
330		    irq == icu_data[2].cascade_irq ||
331		    irq == icu_data[3].cascade_irq ||
332		    irq == icu_data[4].cascade_irq ||
333		    irq == icu_data[5].cascade_irq ||
334		    irq == icu_data[6].cascade_irq ||
335		    irq == icu_data[7].cascade_irq) {
336			irq_set_chip(irq, &icu_irq_chip);
337			irq_set_chained_handler(irq, icu_mux_irq_demux);
338		} else {
339			irq_set_chip_and_handler(irq, &icu_irq_chip,
340						 handle_level_irq);
341		}
342		set_irq_flags(irq, IRQF_VALID);
343	}
344	irq_set_default_host(icu_data[0].domain);
345	set_handle_irq(mmp2_handle_irq);
346}
347
348#ifdef CONFIG_OF
349static int __init mmp_init_bases(struct device_node *node)
350{
351	int ret, nr_irqs, irq, i = 0;
352
353	ret = of_property_read_u32(node, "mrvl,intc-nr-irqs", &nr_irqs);
354	if (ret) {
355		pr_err("Not found mrvl,intc-nr-irqs property\n");
356		return ret;
357	}
358
359	mmp_icu_base = of_iomap(node, 0);
360	if (!mmp_icu_base) {
361		pr_err("Failed to get interrupt controller register\n");
362		return -ENOMEM;
363	}
364
365	icu_data[0].virq_base = 0;
366	icu_data[0].domain = irq_domain_add_linear(node, nr_irqs,
367						   &mmp_irq_domain_ops,
368						   &icu_data[0]);
369	for (irq = 0; irq < nr_irqs; irq++) {
370		ret = irq_create_mapping(icu_data[0].domain, irq);
371		if (!ret) {
372			pr_err("Failed to mapping hwirq\n");
373			goto err;
374		}
375		if (!irq)
376			icu_data[0].virq_base = ret;
377	}
378	icu_data[0].nr_irqs = nr_irqs;
379	return 0;
380err:
381	if (icu_data[0].virq_base) {
382		for (i = 0; i < irq; i++)
383			irq_dispose_mapping(icu_data[0].virq_base + i);
384	}
385	irq_domain_remove(icu_data[0].domain);
386	iounmap(mmp_icu_base);
387	return -EINVAL;
388}
389
390static int __init mmp_of_init(struct device_node *node,
391			      struct device_node *parent)
392{
393	int ret;
394
395	ret = mmp_init_bases(node);
396	if (ret < 0)
397		return ret;
398
399	icu_data[0].conf_enable = mmp_conf.conf_enable;
400	icu_data[0].conf_disable = mmp_conf.conf_disable;
401	icu_data[0].conf_mask = mmp_conf.conf_mask;
402	irq_set_default_host(icu_data[0].domain);
403	set_handle_irq(mmp_handle_irq);
404	max_icu_nr = 1;
405	return 0;
406}
407IRQCHIP_DECLARE(mmp_intc, "mrvl,mmp-intc", mmp_of_init);
408
409static int __init mmp2_of_init(struct device_node *node,
410			       struct device_node *parent)
411{
412	int ret;
413
414	ret = mmp_init_bases(node);
415	if (ret < 0)
416		return ret;
417
418	icu_data[0].conf_enable = mmp2_conf.conf_enable;
419	icu_data[0].conf_disable = mmp2_conf.conf_disable;
420	icu_data[0].conf_mask = mmp2_conf.conf_mask;
421	irq_set_default_host(icu_data[0].domain);
422	set_handle_irq(mmp2_handle_irq);
423	max_icu_nr = 1;
424	return 0;
425}
426IRQCHIP_DECLARE(mmp2_intc, "mrvl,mmp2-intc", mmp2_of_init);
427
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
428static int __init mmp2_mux_of_init(struct device_node *node,
429				   struct device_node *parent)
430{
431	struct resource res;
432	int i, ret, irq, j = 0;
433	u32 nr_irqs, mfp_irq;
 
434
435	if (!parent)
436		return -ENODEV;
437
438	i = max_icu_nr;
439	ret = of_property_read_u32(node, "mrvl,intc-nr-irqs",
440				   &nr_irqs);
441	if (ret) {
442		pr_err("Not found mrvl,intc-nr-irqs property\n");
443		return -EINVAL;
444	}
445	ret = of_address_to_resource(node, 0, &res);
 
 
 
 
 
 
 
 
 
446	if (ret < 0) {
447		pr_err("Not found reg property\n");
448		return -EINVAL;
449	}
450	icu_data[i].reg_status = mmp_icu_base + res.start;
451	ret = of_address_to_resource(node, 1, &res);
452	if (ret < 0) {
453		pr_err("Not found reg property\n");
454		return -EINVAL;
455	}
456	icu_data[i].reg_mask = mmp_icu_base + res.start;
457	icu_data[i].cascade_irq = irq_of_parse_and_map(node, 0);
458	if (!icu_data[i].cascade_irq)
459		return -EINVAL;
460
461	icu_data[i].virq_base = 0;
462	icu_data[i].domain = irq_domain_add_linear(node, nr_irqs,
463						   &mmp_irq_domain_ops,
464						   &icu_data[i]);
465	for (irq = 0; irq < nr_irqs; irq++) {
466		ret = irq_create_mapping(icu_data[i].domain, irq);
467		if (!ret) {
468			pr_err("Failed to mapping hwirq\n");
469			goto err;
470		}
471		if (!irq)
472			icu_data[i].virq_base = ret;
473	}
474	icu_data[i].nr_irqs = nr_irqs;
475	if (!of_property_read_u32(node, "mrvl,clr-mfp-irq",
476				  &mfp_irq)) {
477		icu_data[i].clr_mfp_irq_base = icu_data[i].virq_base;
478		icu_data[i].clr_mfp_hwirq = mfp_irq;
479	}
480	irq_set_chained_handler(icu_data[i].cascade_irq,
481				icu_mux_irq_demux);
482	max_icu_nr++;
483	return 0;
484err:
485	if (icu_data[i].virq_base) {
486		for (j = 0; j < irq; j++)
487			irq_dispose_mapping(icu_data[i].virq_base + j);
488	}
489	irq_domain_remove(icu_data[i].domain);
490	return -EINVAL;
491}
492IRQCHIP_DECLARE(mmp2_mux_intc, "mrvl,mmp2-mux-intc", mmp2_mux_of_init);
493#endif