Linux Audio

Check our new training course

Loading...
v6.8
  1/*
  2 * linux/arch/arm/mach-omap2/irq.c
  3 *
  4 * Interrupt handler for OMAP2 boards.
  5 *
  6 * Copyright (C) 2005 Nokia Corporation
  7 * Author: Paul Mundt <paul.mundt@nokia.com>
  8 *
  9 * This file is subject to the terms and conditions of the GNU General Public
 10 * License. See the file "COPYING" in the main directory of this archive
 11 * for more details.
 12 */
 13#include <linux/kernel.h>
 14#include <linux/module.h>
 15#include <linux/init.h>
 16#include <linux/interrupt.h>
 17#include <linux/io.h>
 18
 19#include <asm/exception.h>
 20#include <linux/irqchip.h>
 21#include <linux/irqdomain.h>
 22#include <linux/of.h>
 23#include <linux/of_address.h>
 24#include <linux/of_irq.h>
 25
 26#include <linux/irqchip/irq-omap-intc.h>
 27
 28/* selected INTC register offsets */
 29
 30#define INTC_REVISION		0x0000
 31#define INTC_SYSCONFIG		0x0010
 32#define INTC_SYSSTATUS		0x0014
 33#define INTC_SIR		0x0040
 34#define INTC_CONTROL		0x0048
 35#define INTC_PROTECTION		0x004C
 36#define INTC_IDLE		0x0050
 37#define INTC_THRESHOLD		0x0068
 38#define INTC_MIR0		0x0084
 39#define INTC_MIR_CLEAR0		0x0088
 40#define INTC_MIR_SET0		0x008c
 41#define INTC_PENDING_IRQ0	0x0098
 42#define INTC_PENDING_IRQ1	0x00b8
 43#define INTC_PENDING_IRQ2	0x00d8
 44#define INTC_PENDING_IRQ3	0x00f8
 45#define INTC_ILR0		0x0100
 46
 47#define ACTIVEIRQ_MASK		0x7f	/* omap2/3 active interrupt bits */
 48#define SPURIOUSIRQ_MASK	(0x1ffffff << 7)
 49#define INTCPS_NR_ILR_REGS	128
 50#define INTCPS_NR_MIR_REGS	4
 51
 52#define INTC_IDLE_FUNCIDLE	(1 << 0)
 53#define INTC_IDLE_TURBO		(1 << 1)
 54
 55#define INTC_PROTECTION_ENABLE	(1 << 0)
 56
 57struct omap_intc_regs {
 58	u32 sysconfig;
 59	u32 protection;
 60	u32 idle;
 61	u32 threshold;
 62	u32 ilr[INTCPS_NR_ILR_REGS];
 63	u32 mir[INTCPS_NR_MIR_REGS];
 64};
 65static struct omap_intc_regs intc_context;
 66
 67static struct irq_domain *domain;
 68static void __iomem *omap_irq_base;
 69static int omap_nr_pending;
 70static int omap_nr_irqs;
 71
 72static void intc_writel(u32 reg, u32 val)
 73{
 74	writel_relaxed(val, omap_irq_base + reg);
 75}
 76
 77static u32 intc_readl(u32 reg)
 78{
 79	return readl_relaxed(omap_irq_base + reg);
 80}
 81
 82void omap_intc_save_context(void)
 83{
 84	int i;
 85
 86	intc_context.sysconfig =
 87		intc_readl(INTC_SYSCONFIG);
 88	intc_context.protection =
 89		intc_readl(INTC_PROTECTION);
 90	intc_context.idle =
 91		intc_readl(INTC_IDLE);
 92	intc_context.threshold =
 93		intc_readl(INTC_THRESHOLD);
 94
 95	for (i = 0; i < omap_nr_irqs; i++)
 96		intc_context.ilr[i] =
 97			intc_readl((INTC_ILR0 + 0x4 * i));
 98	for (i = 0; i < INTCPS_NR_MIR_REGS; i++)
 99		intc_context.mir[i] =
100			intc_readl(INTC_MIR0 + (0x20 * i));
101}
102
103void omap_intc_restore_context(void)
104{
105	int i;
106
107	intc_writel(INTC_SYSCONFIG, intc_context.sysconfig);
108	intc_writel(INTC_PROTECTION, intc_context.protection);
109	intc_writel(INTC_IDLE, intc_context.idle);
110	intc_writel(INTC_THRESHOLD, intc_context.threshold);
111
112	for (i = 0; i < omap_nr_irqs; i++)
113		intc_writel(INTC_ILR0 + 0x4 * i,
114				intc_context.ilr[i]);
115
116	for (i = 0; i < INTCPS_NR_MIR_REGS; i++)
117		intc_writel(INTC_MIR0 + 0x20 * i,
118			intc_context.mir[i]);
119	/* MIRs are saved and restore with other PRCM registers */
120}
121
122void omap3_intc_prepare_idle(void)
123{
124	/*
125	 * Disable autoidle as it can stall interrupt controller,
126	 * cf. errata ID i540 for 3430 (all revisions up to 3.1.x)
127	 */
128	intc_writel(INTC_SYSCONFIG, 0);
129	intc_writel(INTC_IDLE, INTC_IDLE_TURBO);
130}
131
132void omap3_intc_resume_idle(void)
133{
134	/* Re-enable autoidle */
135	intc_writel(INTC_SYSCONFIG, 1);
136	intc_writel(INTC_IDLE, 0);
137}
138
139/* XXX: FIQ and additional INTC support (only MPU at the moment) */
140static void omap_ack_irq(struct irq_data *d)
141{
142	intc_writel(INTC_CONTROL, 0x1);
143}
144
145static void omap_mask_ack_irq(struct irq_data *d)
146{
147	irq_gc_mask_disable_reg(d);
148	omap_ack_irq(d);
149}
150
151static void __init omap_irq_soft_reset(void)
152{
153	unsigned long tmp;
154
155	tmp = intc_readl(INTC_REVISION) & 0xff;
156
157	pr_info("IRQ: Found an INTC at 0x%p (revision %ld.%ld) with %d interrupts\n",
158		omap_irq_base, tmp >> 4, tmp & 0xf, omap_nr_irqs);
159
160	tmp = intc_readl(INTC_SYSCONFIG);
161	tmp |= 1 << 1;	/* soft reset */
162	intc_writel(INTC_SYSCONFIG, tmp);
163
164	while (!(intc_readl(INTC_SYSSTATUS) & 0x1))
165		/* Wait for reset to complete */;
166
167	/* Enable autoidle */
168	intc_writel(INTC_SYSCONFIG, 1 << 0);
169}
170
171int omap_irq_pending(void)
172{
173	int i;
174
175	for (i = 0; i < omap_nr_pending; i++)
176		if (intc_readl(INTC_PENDING_IRQ0 + (0x20 * i)))
177			return 1;
178	return 0;
179}
180
181void omap3_intc_suspend(void)
182{
183	/* A pending interrupt would prevent OMAP from entering suspend */
184	omap_ack_irq(NULL);
185}
186
187static int __init omap_alloc_gc_of(struct irq_domain *d, void __iomem *base)
188{
189	int ret;
190	int i;
191
192	ret = irq_alloc_domain_generic_chips(d, 32, 1, "INTC",
193			handle_level_irq, IRQ_NOREQUEST | IRQ_NOPROBE,
194			IRQ_LEVEL, 0);
195	if (ret) {
196		pr_warn("Failed to allocate irq chips\n");
197		return ret;
198	}
199
200	for (i = 0; i < omap_nr_pending; i++) {
201		struct irq_chip_generic *gc;
202		struct irq_chip_type *ct;
203
204		gc = irq_get_domain_generic_chip(d, 32 * i);
205		gc->reg_base = base;
206		ct = gc->chip_types;
207
208		ct->type = IRQ_TYPE_LEVEL_MASK;
209
210		ct->chip.irq_ack = omap_mask_ack_irq;
211		ct->chip.irq_mask = irq_gc_mask_disable_reg;
212		ct->chip.irq_unmask = irq_gc_unmask_enable_reg;
213
214		ct->chip.flags |= IRQCHIP_SKIP_SET_WAKE;
215
216		ct->regs.enable = INTC_MIR_CLEAR0 + 32 * i;
217		ct->regs.disable = INTC_MIR_SET0 + 32 * i;
218	}
219
220	return 0;
221}
222
223static void __init omap_alloc_gc_legacy(void __iomem *base,
224		unsigned int irq_start, unsigned int num)
225{
226	struct irq_chip_generic *gc;
227	struct irq_chip_type *ct;
228
229	gc = irq_alloc_generic_chip("INTC", 1, irq_start, base,
230			handle_level_irq);
231	ct = gc->chip_types;
232	ct->chip.irq_ack = omap_mask_ack_irq;
233	ct->chip.irq_mask = irq_gc_mask_disable_reg;
234	ct->chip.irq_unmask = irq_gc_unmask_enable_reg;
235	ct->chip.flags |= IRQCHIP_SKIP_SET_WAKE;
236
237	ct->regs.enable = INTC_MIR_CLEAR0;
238	ct->regs.disable = INTC_MIR_SET0;
239	irq_setup_generic_chip(gc, IRQ_MSK(num), IRQ_GC_INIT_MASK_CACHE,
240			IRQ_NOREQUEST | IRQ_NOPROBE, 0);
241}
242
243static int __init omap_init_irq_of(struct device_node *node)
244{
245	int ret;
246
247	omap_irq_base = of_iomap(node, 0);
248	if (WARN_ON(!omap_irq_base))
249		return -ENOMEM;
250
251	domain = irq_domain_add_linear(node, omap_nr_irqs,
252			&irq_generic_chip_ops, NULL);
253
254	omap_irq_soft_reset();
255
256	ret = omap_alloc_gc_of(domain, omap_irq_base);
257	if (ret < 0)
258		irq_domain_remove(domain);
259
260	return ret;
261}
262
263static int __init omap_init_irq_legacy(u32 base, struct device_node *node)
264{
265	int j, irq_base;
266
267	omap_irq_base = ioremap(base, SZ_4K);
268	if (WARN_ON(!omap_irq_base))
269		return -ENOMEM;
270
271	irq_base = irq_alloc_descs(-1, 0, omap_nr_irqs, 0);
272	if (irq_base < 0) {
273		pr_warn("Couldn't allocate IRQ numbers\n");
274		irq_base = 0;
275	}
276
277	domain = irq_domain_add_legacy(node, omap_nr_irqs, irq_base, 0,
278			&irq_domain_simple_ops, NULL);
279
280	omap_irq_soft_reset();
281
282	for (j = 0; j < omap_nr_irqs; j += 32)
283		omap_alloc_gc_legacy(omap_irq_base + j, j + irq_base, 32);
284
285	return 0;
286}
287
288static void __init omap_irq_enable_protection(void)
289{
290	u32 reg;
291
292	reg = intc_readl(INTC_PROTECTION);
293	reg |= INTC_PROTECTION_ENABLE;
294	intc_writel(INTC_PROTECTION, reg);
295}
296
297static int __init omap_init_irq(u32 base, struct device_node *node)
298{
299	int ret;
300
301	/*
302	 * FIXME legacy OMAP DMA driver sitting under arch/arm/plat-omap/dma.c
303	 * depends is still not ready for linear IRQ domains; because of that
304	 * we need to temporarily "blacklist" OMAP2 and OMAP3 devices from using
305	 * linear IRQ Domain until that driver is finally fixed.
306	 */
307	if (of_device_is_compatible(node, "ti,omap2-intc") ||
308			of_device_is_compatible(node, "ti,omap3-intc")) {
309		struct resource res;
310
311		if (of_address_to_resource(node, 0, &res))
312			return -ENOMEM;
313
314		base = res.start;
315		ret = omap_init_irq_legacy(base, node);
316	} else if (node) {
317		ret = omap_init_irq_of(node);
318	} else {
319		ret = omap_init_irq_legacy(base, NULL);
320	}
321
322	if (ret == 0)
323		omap_irq_enable_protection();
324
325	return ret;
326}
327
328static asmlinkage void __exception_irq_entry
329omap_intc_handle_irq(struct pt_regs *regs)
330{
331	extern unsigned long irq_err_count;
332	u32 irqnr;
333
334	irqnr = intc_readl(INTC_SIR);
335
336	/*
337	 * A spurious IRQ can result if interrupt that triggered the
338	 * sorting is no longer active during the sorting (10 INTC
339	 * functional clock cycles after interrupt assertion). Or a
340	 * change in interrupt mask affected the result during sorting
341	 * time. There is no special handling required except ignoring
342	 * the SIR register value just read and retrying.
343	 * See section 6.2.5 of AM335x TRM Literature Number: SPRUH73K
344	 *
345	 * Many a times, a spurious interrupt situation has been fixed
346	 * by adding a flush for the posted write acking the IRQ in
347	 * the device driver. Typically, this is going be the device
348	 * driver whose interrupt was handled just before the spurious
349	 * IRQ occurred. Pay attention to those device drivers if you
350	 * run into hitting the spurious IRQ condition below.
351	 */
352	if (unlikely((irqnr & SPURIOUSIRQ_MASK) == SPURIOUSIRQ_MASK)) {
353		pr_err_once("%s: spurious irq!\n", __func__);
354		irq_err_count++;
355		omap_ack_irq(NULL);
356		return;
357	}
358
359	irqnr &= ACTIVEIRQ_MASK;
360	generic_handle_domain_irq(domain, irqnr);
361}
362
363static int __init intc_of_init(struct device_node *node,
364			     struct device_node *parent)
365{
366	int ret;
367
368	omap_nr_pending = 3;
369	omap_nr_irqs = 96;
370
371	if (WARN_ON(!node))
372		return -ENODEV;
373
374	if (of_device_is_compatible(node, "ti,dm814-intc") ||
375	    of_device_is_compatible(node, "ti,dm816-intc") ||
376	    of_device_is_compatible(node, "ti,am33xx-intc")) {
377		omap_nr_irqs = 128;
378		omap_nr_pending = 4;
379	}
380
381	ret = omap_init_irq(-1, of_node_get(node));
382	if (ret < 0)
383		return ret;
384
385	set_handle_irq(omap_intc_handle_irq);
386
387	return 0;
388}
389
390IRQCHIP_DECLARE(omap2_intc, "ti,omap2-intc", intc_of_init);
391IRQCHIP_DECLARE(omap3_intc, "ti,omap3-intc", intc_of_init);
392IRQCHIP_DECLARE(dm814x_intc, "ti,dm814-intc", intc_of_init);
393IRQCHIP_DECLARE(dm816x_intc, "ti,dm816-intc", intc_of_init);
394IRQCHIP_DECLARE(am33xx_intc, "ti,am33xx-intc", intc_of_init);
v5.4
  1/*
  2 * linux/arch/arm/mach-omap2/irq.c
  3 *
  4 * Interrupt handler for OMAP2 boards.
  5 *
  6 * Copyright (C) 2005 Nokia Corporation
  7 * Author: Paul Mundt <paul.mundt@nokia.com>
  8 *
  9 * This file is subject to the terms and conditions of the GNU General Public
 10 * License. See the file "COPYING" in the main directory of this archive
 11 * for more details.
 12 */
 13#include <linux/kernel.h>
 14#include <linux/module.h>
 15#include <linux/init.h>
 16#include <linux/interrupt.h>
 17#include <linux/io.h>
 18
 19#include <asm/exception.h>
 20#include <linux/irqchip.h>
 21#include <linux/irqdomain.h>
 22#include <linux/of.h>
 23#include <linux/of_address.h>
 24#include <linux/of_irq.h>
 25
 26#include <linux/irqchip/irq-omap-intc.h>
 27
 28/* selected INTC register offsets */
 29
 30#define INTC_REVISION		0x0000
 31#define INTC_SYSCONFIG		0x0010
 32#define INTC_SYSSTATUS		0x0014
 33#define INTC_SIR		0x0040
 34#define INTC_CONTROL		0x0048
 35#define INTC_PROTECTION		0x004C
 36#define INTC_IDLE		0x0050
 37#define INTC_THRESHOLD		0x0068
 38#define INTC_MIR0		0x0084
 39#define INTC_MIR_CLEAR0		0x0088
 40#define INTC_MIR_SET0		0x008c
 41#define INTC_PENDING_IRQ0	0x0098
 42#define INTC_PENDING_IRQ1	0x00b8
 43#define INTC_PENDING_IRQ2	0x00d8
 44#define INTC_PENDING_IRQ3	0x00f8
 45#define INTC_ILR0		0x0100
 46
 47#define ACTIVEIRQ_MASK		0x7f	/* omap2/3 active interrupt bits */
 48#define SPURIOUSIRQ_MASK	(0x1ffffff << 7)
 49#define INTCPS_NR_ILR_REGS	128
 50#define INTCPS_NR_MIR_REGS	4
 51
 52#define INTC_IDLE_FUNCIDLE	(1 << 0)
 53#define INTC_IDLE_TURBO		(1 << 1)
 54
 55#define INTC_PROTECTION_ENABLE	(1 << 0)
 56
 57struct omap_intc_regs {
 58	u32 sysconfig;
 59	u32 protection;
 60	u32 idle;
 61	u32 threshold;
 62	u32 ilr[INTCPS_NR_ILR_REGS];
 63	u32 mir[INTCPS_NR_MIR_REGS];
 64};
 65static struct omap_intc_regs intc_context;
 66
 67static struct irq_domain *domain;
 68static void __iomem *omap_irq_base;
 69static int omap_nr_pending;
 70static int omap_nr_irqs;
 71
 72static void intc_writel(u32 reg, u32 val)
 73{
 74	writel_relaxed(val, omap_irq_base + reg);
 75}
 76
 77static u32 intc_readl(u32 reg)
 78{
 79	return readl_relaxed(omap_irq_base + reg);
 80}
 81
 82void omap_intc_save_context(void)
 83{
 84	int i;
 85
 86	intc_context.sysconfig =
 87		intc_readl(INTC_SYSCONFIG);
 88	intc_context.protection =
 89		intc_readl(INTC_PROTECTION);
 90	intc_context.idle =
 91		intc_readl(INTC_IDLE);
 92	intc_context.threshold =
 93		intc_readl(INTC_THRESHOLD);
 94
 95	for (i = 0; i < omap_nr_irqs; i++)
 96		intc_context.ilr[i] =
 97			intc_readl((INTC_ILR0 + 0x4 * i));
 98	for (i = 0; i < INTCPS_NR_MIR_REGS; i++)
 99		intc_context.mir[i] =
100			intc_readl(INTC_MIR0 + (0x20 * i));
101}
102
103void omap_intc_restore_context(void)
104{
105	int i;
106
107	intc_writel(INTC_SYSCONFIG, intc_context.sysconfig);
108	intc_writel(INTC_PROTECTION, intc_context.protection);
109	intc_writel(INTC_IDLE, intc_context.idle);
110	intc_writel(INTC_THRESHOLD, intc_context.threshold);
111
112	for (i = 0; i < omap_nr_irqs; i++)
113		intc_writel(INTC_ILR0 + 0x4 * i,
114				intc_context.ilr[i]);
115
116	for (i = 0; i < INTCPS_NR_MIR_REGS; i++)
117		intc_writel(INTC_MIR0 + 0x20 * i,
118			intc_context.mir[i]);
119	/* MIRs are saved and restore with other PRCM registers */
120}
121
122void omap3_intc_prepare_idle(void)
123{
124	/*
125	 * Disable autoidle as it can stall interrupt controller,
126	 * cf. errata ID i540 for 3430 (all revisions up to 3.1.x)
127	 */
128	intc_writel(INTC_SYSCONFIG, 0);
129	intc_writel(INTC_IDLE, INTC_IDLE_TURBO);
130}
131
132void omap3_intc_resume_idle(void)
133{
134	/* Re-enable autoidle */
135	intc_writel(INTC_SYSCONFIG, 1);
136	intc_writel(INTC_IDLE, 0);
137}
138
139/* XXX: FIQ and additional INTC support (only MPU at the moment) */
140static void omap_ack_irq(struct irq_data *d)
141{
142	intc_writel(INTC_CONTROL, 0x1);
143}
144
145static void omap_mask_ack_irq(struct irq_data *d)
146{
147	irq_gc_mask_disable_reg(d);
148	omap_ack_irq(d);
149}
150
151static void __init omap_irq_soft_reset(void)
152{
153	unsigned long tmp;
154
155	tmp = intc_readl(INTC_REVISION) & 0xff;
156
157	pr_info("IRQ: Found an INTC at 0x%p (revision %ld.%ld) with %d interrupts\n",
158		omap_irq_base, tmp >> 4, tmp & 0xf, omap_nr_irqs);
159
160	tmp = intc_readl(INTC_SYSCONFIG);
161	tmp |= 1 << 1;	/* soft reset */
162	intc_writel(INTC_SYSCONFIG, tmp);
163
164	while (!(intc_readl(INTC_SYSSTATUS) & 0x1))
165		/* Wait for reset to complete */;
166
167	/* Enable autoidle */
168	intc_writel(INTC_SYSCONFIG, 1 << 0);
169}
170
171int omap_irq_pending(void)
172{
173	int i;
174
175	for (i = 0; i < omap_nr_pending; i++)
176		if (intc_readl(INTC_PENDING_IRQ0 + (0x20 * i)))
177			return 1;
178	return 0;
179}
180
181void omap3_intc_suspend(void)
182{
183	/* A pending interrupt would prevent OMAP from entering suspend */
184	omap_ack_irq(NULL);
185}
186
187static int __init omap_alloc_gc_of(struct irq_domain *d, void __iomem *base)
188{
189	int ret;
190	int i;
191
192	ret = irq_alloc_domain_generic_chips(d, 32, 1, "INTC",
193			handle_level_irq, IRQ_NOREQUEST | IRQ_NOPROBE,
194			IRQ_LEVEL, 0);
195	if (ret) {
196		pr_warn("Failed to allocate irq chips\n");
197		return ret;
198	}
199
200	for (i = 0; i < omap_nr_pending; i++) {
201		struct irq_chip_generic *gc;
202		struct irq_chip_type *ct;
203
204		gc = irq_get_domain_generic_chip(d, 32 * i);
205		gc->reg_base = base;
206		ct = gc->chip_types;
207
208		ct->type = IRQ_TYPE_LEVEL_MASK;
209
210		ct->chip.irq_ack = omap_mask_ack_irq;
211		ct->chip.irq_mask = irq_gc_mask_disable_reg;
212		ct->chip.irq_unmask = irq_gc_unmask_enable_reg;
213
214		ct->chip.flags |= IRQCHIP_SKIP_SET_WAKE;
215
216		ct->regs.enable = INTC_MIR_CLEAR0 + 32 * i;
217		ct->regs.disable = INTC_MIR_SET0 + 32 * i;
218	}
219
220	return 0;
221}
222
223static void __init omap_alloc_gc_legacy(void __iomem *base,
224		unsigned int irq_start, unsigned int num)
225{
226	struct irq_chip_generic *gc;
227	struct irq_chip_type *ct;
228
229	gc = irq_alloc_generic_chip("INTC", 1, irq_start, base,
230			handle_level_irq);
231	ct = gc->chip_types;
232	ct->chip.irq_ack = omap_mask_ack_irq;
233	ct->chip.irq_mask = irq_gc_mask_disable_reg;
234	ct->chip.irq_unmask = irq_gc_unmask_enable_reg;
235	ct->chip.flags |= IRQCHIP_SKIP_SET_WAKE;
236
237	ct->regs.enable = INTC_MIR_CLEAR0;
238	ct->regs.disable = INTC_MIR_SET0;
239	irq_setup_generic_chip(gc, IRQ_MSK(num), IRQ_GC_INIT_MASK_CACHE,
240			IRQ_NOREQUEST | IRQ_NOPROBE, 0);
241}
242
243static int __init omap_init_irq_of(struct device_node *node)
244{
245	int ret;
246
247	omap_irq_base = of_iomap(node, 0);
248	if (WARN_ON(!omap_irq_base))
249		return -ENOMEM;
250
251	domain = irq_domain_add_linear(node, omap_nr_irqs,
252			&irq_generic_chip_ops, NULL);
253
254	omap_irq_soft_reset();
255
256	ret = omap_alloc_gc_of(domain, omap_irq_base);
257	if (ret < 0)
258		irq_domain_remove(domain);
259
260	return ret;
261}
262
263static int __init omap_init_irq_legacy(u32 base, struct device_node *node)
264{
265	int j, irq_base;
266
267	omap_irq_base = ioremap(base, SZ_4K);
268	if (WARN_ON(!omap_irq_base))
269		return -ENOMEM;
270
271	irq_base = irq_alloc_descs(-1, 0, omap_nr_irqs, 0);
272	if (irq_base < 0) {
273		pr_warn("Couldn't allocate IRQ numbers\n");
274		irq_base = 0;
275	}
276
277	domain = irq_domain_add_legacy(node, omap_nr_irqs, irq_base, 0,
278			&irq_domain_simple_ops, NULL);
279
280	omap_irq_soft_reset();
281
282	for (j = 0; j < omap_nr_irqs; j += 32)
283		omap_alloc_gc_legacy(omap_irq_base + j, j + irq_base, 32);
284
285	return 0;
286}
287
288static void __init omap_irq_enable_protection(void)
289{
290	u32 reg;
291
292	reg = intc_readl(INTC_PROTECTION);
293	reg |= INTC_PROTECTION_ENABLE;
294	intc_writel(INTC_PROTECTION, reg);
295}
296
297static int __init omap_init_irq(u32 base, struct device_node *node)
298{
299	int ret;
300
301	/*
302	 * FIXME legacy OMAP DMA driver sitting under arch/arm/plat-omap/dma.c
303	 * depends is still not ready for linear IRQ domains; because of that
304	 * we need to temporarily "blacklist" OMAP2 and OMAP3 devices from using
305	 * linear IRQ Domain until that driver is finally fixed.
306	 */
307	if (of_device_is_compatible(node, "ti,omap2-intc") ||
308			of_device_is_compatible(node, "ti,omap3-intc")) {
309		struct resource res;
310
311		if (of_address_to_resource(node, 0, &res))
312			return -ENOMEM;
313
314		base = res.start;
315		ret = omap_init_irq_legacy(base, node);
316	} else if (node) {
317		ret = omap_init_irq_of(node);
318	} else {
319		ret = omap_init_irq_legacy(base, NULL);
320	}
321
322	if (ret == 0)
323		omap_irq_enable_protection();
324
325	return ret;
326}
327
328static asmlinkage void __exception_irq_entry
329omap_intc_handle_irq(struct pt_regs *regs)
330{
331	extern unsigned long irq_err_count;
332	u32 irqnr;
333
334	irqnr = intc_readl(INTC_SIR);
335
336	/*
337	 * A spurious IRQ can result if interrupt that triggered the
338	 * sorting is no longer active during the sorting (10 INTC
339	 * functional clock cycles after interrupt assertion). Or a
340	 * change in interrupt mask affected the result during sorting
341	 * time. There is no special handling required except ignoring
342	 * the SIR register value just read and retrying.
343	 * See section 6.2.5 of AM335x TRM Literature Number: SPRUH73K
344	 *
345	 * Many a times, a spurious interrupt situation has been fixed
346	 * by adding a flush for the posted write acking the IRQ in
347	 * the device driver. Typically, this is going be the device
348	 * driver whose interrupt was handled just before the spurious
349	 * IRQ occurred. Pay attention to those device drivers if you
350	 * run into hitting the spurious IRQ condition below.
351	 */
352	if (unlikely((irqnr & SPURIOUSIRQ_MASK) == SPURIOUSIRQ_MASK)) {
353		pr_err_once("%s: spurious irq!\n", __func__);
354		irq_err_count++;
355		omap_ack_irq(NULL);
356		return;
357	}
358
359	irqnr &= ACTIVEIRQ_MASK;
360	handle_domain_irq(domain, irqnr, regs);
361}
362
363static int __init intc_of_init(struct device_node *node,
364			     struct device_node *parent)
365{
366	int ret;
367
368	omap_nr_pending = 3;
369	omap_nr_irqs = 96;
370
371	if (WARN_ON(!node))
372		return -ENODEV;
373
374	if (of_device_is_compatible(node, "ti,dm814-intc") ||
375	    of_device_is_compatible(node, "ti,dm816-intc") ||
376	    of_device_is_compatible(node, "ti,am33xx-intc")) {
377		omap_nr_irqs = 128;
378		omap_nr_pending = 4;
379	}
380
381	ret = omap_init_irq(-1, of_node_get(node));
382	if (ret < 0)
383		return ret;
384
385	set_handle_irq(omap_intc_handle_irq);
386
387	return 0;
388}
389
390IRQCHIP_DECLARE(omap2_intc, "ti,omap2-intc", intc_of_init);
391IRQCHIP_DECLARE(omap3_intc, "ti,omap3-intc", intc_of_init);
392IRQCHIP_DECLARE(dm814x_intc, "ti,dm814-intc", intc_of_init);
393IRQCHIP_DECLARE(dm816x_intc, "ti,dm816-intc", intc_of_init);
394IRQCHIP_DECLARE(am33xx_intc, "ti,am33xx-intc", intc_of_init);