Linux Audio

Check our new training course

Loading...
v4.17
  1// SPDX-License-Identifier: GPL-2.0
  2/*
  3 * Copyright (C) 2009 Daniel Hellstrom (daniel@gaisler.com) Aeroflex Gaisler AB
  4 * Copyright (C) 2009 Konrad Eisele (konrad@gaisler.com) Aeroflex Gaisler AB
  5 */
  6
  7#include <linux/kernel.h>
 
  8#include <linux/errno.h>
  9#include <linux/mutex.h>
 10#include <linux/of.h>
 11#include <linux/of_platform.h>
 12#include <linux/interrupt.h>
 13#include <linux/of_device.h>
 14#include <linux/clocksource.h>
 15#include <linux/clockchips.h>
 16
 17#include <asm/oplib.h>
 18#include <asm/timer.h>
 19#include <asm/prom.h>
 20#include <asm/leon.h>
 21#include <asm/leon_amba.h>
 22#include <asm/traps.h>
 23#include <asm/cacheflush.h>
 24#include <asm/smp.h>
 25#include <asm/setup.h>
 26
 27#include "kernel.h"
 28#include "prom.h"
 29#include "irq.h"
 30
 31struct leon3_irqctrl_regs_map *leon3_irqctrl_regs; /* interrupt controller base address */
 32struct leon3_gptimer_regs_map *leon3_gptimer_regs; /* timer controller base address */
 33
 34int leondebug_irq_disable;
 35int leon_debug_irqout;
 36static volatile u32 dummy_master_l10_counter;
 37unsigned long amba_system_id;
 38static DEFINE_SPINLOCK(leon_irq_lock);
 39
 40static unsigned long leon3_gptimer_idx; /* Timer Index (0..6) within Timer Core */
 41static unsigned long leon3_gptimer_ackmask; /* For clearing pending bit */
 42unsigned long leon3_gptimer_irq; /* interrupt controller irq number */
 
 
 43unsigned int sparc_leon_eirq;
 44#define LEON_IMASK(cpu) (&leon3_irqctrl_regs->mask[cpu])
 45#define LEON_IACK (&leon3_irqctrl_regs->iclear)
 46#define LEON_DO_ACK_HW 1
 47
 48/* Return the last ACKed IRQ by the Extended IRQ controller. It has already
 49 * been (automatically) ACKed when the CPU takes the trap.
 50 */
 51static inline unsigned int leon_eirq_get(int cpu)
 52{
 53	return LEON3_BYPASS_LOAD_PA(&leon3_irqctrl_regs->intid[cpu]) & 0x1f;
 54}
 55
 56/* Handle one or multiple IRQs from the extended interrupt controller */
 57static void leon_handle_ext_irq(struct irq_desc *desc)
 58{
 59	unsigned int eirq;
 60	struct irq_bucket *p;
 61	int cpu = sparc_leon3_cpuid();
 62
 63	eirq = leon_eirq_get(cpu);
 64	p = irq_map[eirq];
 65	if ((eirq & 0x10) && p && p->irq) /* bit4 tells if IRQ happened */
 66		generic_handle_irq(p->irq);
 67}
 68
 69/* The extended IRQ controller has been found, this function registers it */
 70static void leon_eirq_setup(unsigned int eirq)
 71{
 72	unsigned long mask, oldmask;
 73	unsigned int veirq;
 74
 75	if (eirq < 1 || eirq > 0xf) {
 76		printk(KERN_ERR "LEON EXT IRQ NUMBER BAD: %d\n", eirq);
 77		return;
 78	}
 79
 80	veirq = leon_build_device_irq(eirq, leon_handle_ext_irq, "extirq", 0);
 81
 82	/*
 83	 * Unmask the Extended IRQ, the IRQs routed through the Ext-IRQ
 84	 * controller have a mask-bit of their own, so this is safe.
 85	 */
 86	irq_link(veirq);
 87	mask = 1 << eirq;
 88	oldmask = LEON3_BYPASS_LOAD_PA(LEON_IMASK(boot_cpu_id));
 89	LEON3_BYPASS_STORE_PA(LEON_IMASK(boot_cpu_id), (oldmask | mask));
 90	sparc_leon_eirq = eirq;
 91}
 92
 93unsigned long leon_get_irqmask(unsigned int irq)
 94{
 95	unsigned long mask;
 96
 97	if (!irq || ((irq > 0xf) && !sparc_leon_eirq)
 98	    || ((irq > 0x1f) && sparc_leon_eirq)) {
 99		printk(KERN_ERR
100		       "leon_get_irqmask: false irq number: %d\n", irq);
101		mask = 0;
102	} else {
103		mask = LEON_HARD_INT(irq);
104	}
105	return mask;
106}
107
108#ifdef CONFIG_SMP
109static int irq_choose_cpu(const struct cpumask *affinity)
110{
111	cpumask_t mask;
112
113	cpumask_and(&mask, cpu_online_mask, affinity);
114	if (cpumask_equal(&mask, cpu_online_mask) || cpumask_empty(&mask))
115		return boot_cpu_id;
116	else
117		return cpumask_first(&mask);
118}
119#else
120#define irq_choose_cpu(affinity) boot_cpu_id
121#endif
122
123static int leon_set_affinity(struct irq_data *data, const struct cpumask *dest,
124			     bool force)
125{
126	unsigned long mask, oldmask, flags;
127	int oldcpu, newcpu;
128
129	mask = (unsigned long)data->chip_data;
130	oldcpu = irq_choose_cpu(irq_data_get_affinity_mask(data));
131	newcpu = irq_choose_cpu(dest);
132
133	if (oldcpu == newcpu)
134		goto out;
135
136	/* unmask on old CPU first before enabling on the selected CPU */
137	spin_lock_irqsave(&leon_irq_lock, flags);
138	oldmask = LEON3_BYPASS_LOAD_PA(LEON_IMASK(oldcpu));
139	LEON3_BYPASS_STORE_PA(LEON_IMASK(oldcpu), (oldmask & ~mask));
140	oldmask = LEON3_BYPASS_LOAD_PA(LEON_IMASK(newcpu));
141	LEON3_BYPASS_STORE_PA(LEON_IMASK(newcpu), (oldmask | mask));
142	spin_unlock_irqrestore(&leon_irq_lock, flags);
143out:
144	return IRQ_SET_MASK_OK;
145}
146
147static void leon_unmask_irq(struct irq_data *data)
148{
149	unsigned long mask, oldmask, flags;
150	int cpu;
151
152	mask = (unsigned long)data->chip_data;
153	cpu = irq_choose_cpu(irq_data_get_affinity_mask(data));
154	spin_lock_irqsave(&leon_irq_lock, flags);
155	oldmask = LEON3_BYPASS_LOAD_PA(LEON_IMASK(cpu));
156	LEON3_BYPASS_STORE_PA(LEON_IMASK(cpu), (oldmask | mask));
157	spin_unlock_irqrestore(&leon_irq_lock, flags);
158}
159
160static void leon_mask_irq(struct irq_data *data)
161{
162	unsigned long mask, oldmask, flags;
163	int cpu;
164
165	mask = (unsigned long)data->chip_data;
166	cpu = irq_choose_cpu(irq_data_get_affinity_mask(data));
167	spin_lock_irqsave(&leon_irq_lock, flags);
168	oldmask = LEON3_BYPASS_LOAD_PA(LEON_IMASK(cpu));
169	LEON3_BYPASS_STORE_PA(LEON_IMASK(cpu), (oldmask & ~mask));
170	spin_unlock_irqrestore(&leon_irq_lock, flags);
171}
172
173static unsigned int leon_startup_irq(struct irq_data *data)
174{
175	irq_link(data->irq);
176	leon_unmask_irq(data);
177	return 0;
178}
179
180static void leon_shutdown_irq(struct irq_data *data)
181{
182	leon_mask_irq(data);
183	irq_unlink(data->irq);
184}
185
186/* Used by external level sensitive IRQ handlers on the LEON: ACK IRQ ctrl */
187static void leon_eoi_irq(struct irq_data *data)
188{
189	unsigned long mask = (unsigned long)data->chip_data;
190
191	if (mask & LEON_DO_ACK_HW)
192		LEON3_BYPASS_STORE_PA(LEON_IACK, mask & ~LEON_DO_ACK_HW);
193}
194
195static struct irq_chip leon_irq = {
196	.name			= "leon",
197	.irq_startup		= leon_startup_irq,
198	.irq_shutdown		= leon_shutdown_irq,
199	.irq_mask		= leon_mask_irq,
200	.irq_unmask		= leon_unmask_irq,
201	.irq_eoi		= leon_eoi_irq,
202	.irq_set_affinity	= leon_set_affinity,
203};
204
205/*
206 * Build a LEON IRQ for the edge triggered LEON IRQ controller:
207 *  Edge (normal) IRQ           - handle_simple_irq, ack=DON'T-CARE, never ack
208 *  Level IRQ (PCI|Level-GPIO)  - handle_fasteoi_irq, ack=1, ack after ISR
209 *  Per-CPU Edge                - handle_percpu_irq, ack=0
210 */
211unsigned int leon_build_device_irq(unsigned int real_irq,
212				    irq_flow_handler_t flow_handler,
213				    const char *name, int do_ack)
214{
215	unsigned int irq;
216	unsigned long mask;
217	struct irq_desc *desc;
218
219	irq = 0;
220	mask = leon_get_irqmask(real_irq);
221	if (mask == 0)
222		goto out;
223
224	irq = irq_alloc(real_irq, real_irq);
225	if (irq == 0)
226		goto out;
227
228	if (do_ack)
229		mask |= LEON_DO_ACK_HW;
230
231	desc = irq_to_desc(irq);
232	if (!desc || !desc->handle_irq || desc->handle_irq == handle_bad_irq) {
233		irq_set_chip_and_handler_name(irq, &leon_irq,
234					      flow_handler, name);
235		irq_set_chip_data(irq, (void *)mask);
236	}
237
238out:
239	return irq;
240}
241
242static unsigned int _leon_build_device_irq(struct platform_device *op,
243					   unsigned int real_irq)
244{
245	return leon_build_device_irq(real_irq, handle_simple_irq, "edge", 0);
246}
247
248void leon_update_virq_handling(unsigned int virq,
249			      irq_flow_handler_t flow_handler,
250			      const char *name, int do_ack)
251{
252	unsigned long mask = (unsigned long)irq_get_chip_data(virq);
253
254	mask &= ~LEON_DO_ACK_HW;
255	if (do_ack)
256		mask |= LEON_DO_ACK_HW;
257
258	irq_set_chip_and_handler_name(virq, &leon_irq,
259				      flow_handler, name);
260	irq_set_chip_data(virq, (void *)mask);
261}
262
263static u32 leon_cycles_offset(void)
264{
265	u32 rld, val, ctrl, off;
266
267	rld = LEON3_BYPASS_LOAD_PA(&leon3_gptimer_regs->e[leon3_gptimer_idx].rld);
268	val = LEON3_BYPASS_LOAD_PA(&leon3_gptimer_regs->e[leon3_gptimer_idx].val);
269	ctrl = LEON3_BYPASS_LOAD_PA(&leon3_gptimer_regs->e[leon3_gptimer_idx].ctrl);
270	if (LEON3_GPTIMER_CTRL_ISPENDING(ctrl)) {
271		val = LEON3_BYPASS_LOAD_PA(&leon3_gptimer_regs->e[leon3_gptimer_idx].val);
272		off = 2 * rld - val;
273	} else {
274		off = rld - val;
275	}
276
277	return off;
278}
279
280#ifdef CONFIG_SMP
281
282/* smp clockevent irq */
283static irqreturn_t leon_percpu_timer_ce_interrupt(int irq, void *unused)
284{
285	struct clock_event_device *ce;
286	int cpu = smp_processor_id();
287
288	leon_clear_profile_irq(cpu);
289
290	if (cpu == boot_cpu_id)
291		timer_interrupt(irq, NULL);
292
293	ce = &per_cpu(sparc32_clockevent, cpu);
294
295	irq_enter();
296	if (ce->event_handler)
297		ce->event_handler(ce);
298	irq_exit();
299
300	return IRQ_HANDLED;
301}
302
303#endif /* CONFIG_SMP */
304
305void __init leon_init_timers(void)
306{
307	int irq, eirq;
308	struct device_node *rootnp, *np, *nnp;
309	struct property *pp;
310	int len;
311	int icsel;
312	int ampopts;
313	int err;
314	u32 config;
315	u32 ctrl;
316
317	sparc_config.get_cycles_offset = leon_cycles_offset;
318	sparc_config.cs_period = 1000000 / HZ;
319	sparc_config.features |= FEAT_L10_CLOCKSOURCE;
320
321#ifndef CONFIG_SMP
322	sparc_config.features |= FEAT_L10_CLOCKEVENT;
323#endif
324
325	leondebug_irq_disable = 0;
326	leon_debug_irqout = 0;
327	master_l10_counter = (u32 __iomem *)&dummy_master_l10_counter;
328	dummy_master_l10_counter = 0;
329
330	rootnp = of_find_node_by_path("/ambapp0");
331	if (!rootnp)
332		goto bad;
333
334	/* Find System ID: GRLIB build ID and optional CHIP ID */
335	pp = of_find_property(rootnp, "systemid", &len);
336	if (pp)
337		amba_system_id = *(unsigned long *)pp->value;
338
339	/* Find IRQMP IRQ Controller Registers base adr otherwise bail out */
340	np = of_find_node_by_name(rootnp, "GAISLER_IRQMP");
341	if (!np) {
342		np = of_find_node_by_name(rootnp, "01_00d");
343		if (!np)
344			goto bad;
345	}
346	pp = of_find_property(np, "reg", &len);
347	if (!pp)
348		goto bad;
349	leon3_irqctrl_regs = *(struct leon3_irqctrl_regs_map **)pp->value;
350
351	/* Find GPTIMER Timer Registers base address otherwise bail out. */
352	nnp = rootnp;
 
 
 
 
 
 
 
353
354retry:
355	np = of_find_node_by_name(nnp, "GAISLER_GPTIMER");
356	if (!np) {
357		np = of_find_node_by_name(nnp, "01_011");
358		if (!np)
359			goto bad;
360	}
361
362	ampopts = 0;
363	pp = of_find_property(np, "ampopts", &len);
364	if (pp) {
365		ampopts = *(int *)pp->value;
366		if (ampopts == 0) {
367			/* Skip this instance, resource already
368			 * allocated by other OS */
369			nnp = np;
370			goto retry;
371		}
372	}
373
374	/* Select Timer-Instance on Timer Core. Default is zero */
375	leon3_gptimer_idx = ampopts & 0x7;
376
377	pp = of_find_property(np, "reg", &len);
378	if (pp)
379		leon3_gptimer_regs = *(struct leon3_gptimer_regs_map **)
380					pp->value;
381	pp = of_find_property(np, "interrupts", &len);
382	if (pp)
383		leon3_gptimer_irq = *(unsigned int *)pp->value;
 
384
385	if (!(leon3_gptimer_regs && leon3_irqctrl_regs && leon3_gptimer_irq))
386		goto bad;
387
388	ctrl = LEON3_BYPASS_LOAD_PA(&leon3_gptimer_regs->e[leon3_gptimer_idx].ctrl);
389	LEON3_BYPASS_STORE_PA(&leon3_gptimer_regs->e[leon3_gptimer_idx].ctrl,
390			      ctrl | LEON3_GPTIMER_CTRL_PENDING);
391	ctrl = LEON3_BYPASS_LOAD_PA(&leon3_gptimer_regs->e[leon3_gptimer_idx].ctrl);
392
393	if ((ctrl & LEON3_GPTIMER_CTRL_PENDING) != 0)
394		leon3_gptimer_ackmask = ~LEON3_GPTIMER_CTRL_PENDING;
395	else
396		leon3_gptimer_ackmask = ~0;
397
398	LEON3_BYPASS_STORE_PA(&leon3_gptimer_regs->e[leon3_gptimer_idx].val, 0);
399	LEON3_BYPASS_STORE_PA(&leon3_gptimer_regs->e[leon3_gptimer_idx].rld,
400				(((1000000 / HZ) - 1)));
401	LEON3_BYPASS_STORE_PA(
402			&leon3_gptimer_regs->e[leon3_gptimer_idx].ctrl, 0);
403
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
404	/*
405	 * The IRQ controller may (if implemented) consist of multiple
406	 * IRQ controllers, each mapped on a 4Kb boundary.
407	 * Each CPU may be routed to different IRQCTRLs, however
408	 * we assume that all CPUs (in SMP system) is routed to the
409	 * same IRQ Controller, and for non-SMP only one IRQCTRL is
410	 * accessed anyway.
411	 * In AMP systems, Linux must run on CPU0 for the time being.
412	 */
413	icsel = LEON3_BYPASS_LOAD_PA(&leon3_irqctrl_regs->icsel[boot_cpu_id/8]);
414	icsel = (icsel >> ((7 - (boot_cpu_id&0x7)) * 4)) & 0xf;
415	leon3_irqctrl_regs += icsel;
416
417	/* Mask all IRQs on boot-cpu IRQ controller */
418	LEON3_BYPASS_STORE_PA(&leon3_irqctrl_regs->mask[boot_cpu_id], 0);
419
420	/* Probe extended IRQ controller */
421	eirq = (LEON3_BYPASS_LOAD_PA(&leon3_irqctrl_regs->mpstatus)
422		>> 16) & 0xf;
423	if (eirq != 0)
424		leon_eirq_setup(eirq);
425
 
 
 
 
 
 
 
426#ifdef CONFIG_SMP
427	{
428		unsigned long flags;
429
430		/*
431		 * In SMP, sun4m adds a IPI handler to IRQ trap handler that
432		 * LEON never must take, sun4d and LEON overwrites the branch
433		 * with a NOP.
434		 */
435		local_irq_save(flags);
436		patchme_maybe_smp_msg[0] = 0x01000000; /* NOP out the branch */
437		local_ops->cache_all();
438		local_irq_restore(flags);
439	}
440#endif
441
442	config = LEON3_BYPASS_LOAD_PA(&leon3_gptimer_regs->config);
443	if (config & (1 << LEON3_GPTIMER_SEPIRQ))
444		leon3_gptimer_irq += leon3_gptimer_idx;
445	else if ((config & LEON3_GPTIMER_TIMERS) > 1)
446		pr_warn("GPTIMER uses shared irqs, using other timers of the same core will fail.\n");
447
448#ifdef CONFIG_SMP
449	/* Install per-cpu IRQ handler for broadcasted ticker */
450	irq = leon_build_device_irq(leon3_gptimer_irq, handle_percpu_irq,
451				    "per-cpu", 0);
452	err = request_irq(irq, leon_percpu_timer_ce_interrupt,
453			  IRQF_PERCPU | IRQF_TIMER, "timer", NULL);
454#else
455	irq = _leon_build_device_irq(NULL, leon3_gptimer_irq);
456	err = request_irq(irq, timer_interrupt, IRQF_TIMER, "timer", NULL);
457#endif
458	if (err) {
459		pr_err("Unable to attach timer IRQ%d\n", irq);
460		prom_halt();
461	}
462	LEON3_BYPASS_STORE_PA(&leon3_gptimer_regs->e[leon3_gptimer_idx].ctrl,
 
463			      LEON3_GPTIMER_EN |
464			      LEON3_GPTIMER_RL |
465			      LEON3_GPTIMER_LD |
466			      LEON3_GPTIMER_IRQEN);
 
467	return;
468bad:
469	printk(KERN_ERR "No Timer/irqctrl found\n");
470	BUG();
471	return;
472}
473
474static void leon_clear_clock_irq(void)
475{
476	u32 ctrl;
477
478	ctrl = LEON3_BYPASS_LOAD_PA(&leon3_gptimer_regs->e[leon3_gptimer_idx].ctrl);
479	LEON3_BYPASS_STORE_PA(&leon3_gptimer_regs->e[leon3_gptimer_idx].ctrl,
480			      ctrl & leon3_gptimer_ackmask);
481}
482
483static void leon_load_profile_irq(int cpu, unsigned int limit)
484{
 
485}
486
487void __init leon_trans_init(struct device_node *dp)
488{
489	if (strcmp(dp->type, "cpu") == 0 && strcmp(dp->name, "<NULL>") == 0) {
490		struct property *p;
491		p = of_find_property(dp, "mid", (void *)0);
492		if (p) {
493			int mid;
494			dp->name = prom_early_alloc(5 + 1);
495			memcpy(&mid, p->value, p->length);
496			sprintf((char *)dp->name, "cpu%.2d", mid);
497		}
498	}
499}
500
 
 
 
 
 
 
 
 
 
 
 
501#ifdef CONFIG_SMP
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
502void leon_clear_profile_irq(int cpu)
503{
504}
505
506void leon_enable_irq_cpu(unsigned int irq_nr, unsigned int cpu)
507{
508	unsigned long mask, flags, *addr;
509	mask = leon_get_irqmask(irq_nr);
510	spin_lock_irqsave(&leon_irq_lock, flags);
511	addr = (unsigned long *)LEON_IMASK(cpu);
512	LEON3_BYPASS_STORE_PA(addr, (LEON3_BYPASS_LOAD_PA(addr) | mask));
513	spin_unlock_irqrestore(&leon_irq_lock, flags);
514}
515
516#endif
517
518void __init leon_init_IRQ(void)
519{
520	sparc_config.init_timers      = leon_init_timers;
521	sparc_config.build_device_irq = _leon_build_device_irq;
522	sparc_config.clock_rate       = 1000000;
523	sparc_config.clear_clock_irq  = leon_clear_clock_irq;
524	sparc_config.load_profile_irq = leon_load_profile_irq;
 
 
 
 
 
 
 
 
 
 
 
 
 
 
525}
v3.1
 
  1/*
  2 * Copyright (C) 2009 Daniel Hellstrom (daniel@gaisler.com) Aeroflex Gaisler AB
  3 * Copyright (C) 2009 Konrad Eisele (konrad@gaisler.com) Aeroflex Gaisler AB
  4 */
  5
  6#include <linux/kernel.h>
  7#include <linux/module.h>
  8#include <linux/errno.h>
  9#include <linux/mutex.h>
 10#include <linux/of.h>
 11#include <linux/of_platform.h>
 12#include <linux/interrupt.h>
 13#include <linux/of_device.h>
 
 
 14
 15#include <asm/oplib.h>
 16#include <asm/timer.h>
 17#include <asm/prom.h>
 18#include <asm/leon.h>
 19#include <asm/leon_amba.h>
 20#include <asm/traps.h>
 21#include <asm/cacheflush.h>
 22#include <asm/smp.h>
 23#include <asm/setup.h>
 24
 
 25#include "prom.h"
 26#include "irq.h"
 27
 28struct leon3_irqctrl_regs_map *leon3_irqctrl_regs; /* interrupt controller base address */
 29struct leon3_gptimer_regs_map *leon3_gptimer_regs; /* timer controller base address */
 30
 31int leondebug_irq_disable;
 32int leon_debug_irqout;
 33static int dummy_master_l10_counter;
 34unsigned long amba_system_id;
 35static DEFINE_SPINLOCK(leon_irq_lock);
 36
 
 
 37unsigned long leon3_gptimer_irq; /* interrupt controller irq number */
 38unsigned long leon3_gptimer_idx; /* Timer Index (0..6) within Timer Core */
 39int leon3_ticker_irq; /* Timer ticker IRQ */
 40unsigned int sparc_leon_eirq;
 41#define LEON_IMASK(cpu) (&leon3_irqctrl_regs->mask[cpu])
 42#define LEON_IACK (&leon3_irqctrl_regs->iclear)
 43#define LEON_DO_ACK_HW 1
 44
 45/* Return the last ACKed IRQ by the Extended IRQ controller. It has already
 46 * been (automatically) ACKed when the CPU takes the trap.
 47 */
 48static inline unsigned int leon_eirq_get(int cpu)
 49{
 50	return LEON3_BYPASS_LOAD_PA(&leon3_irqctrl_regs->intid[cpu]) & 0x1f;
 51}
 52
 53/* Handle one or multiple IRQs from the extended interrupt controller */
 54static void leon_handle_ext_irq(unsigned int irq, struct irq_desc *desc)
 55{
 56	unsigned int eirq;
 
 57	int cpu = sparc_leon3_cpuid();
 58
 59	eirq = leon_eirq_get(cpu);
 60	if ((eirq & 0x10) && irq_map[eirq]->irq) /* bit4 tells if IRQ happened */
 61		generic_handle_irq(irq_map[eirq]->irq);
 
 62}
 63
 64/* The extended IRQ controller has been found, this function registers it */
 65void leon_eirq_setup(unsigned int eirq)
 66{
 67	unsigned long mask, oldmask;
 68	unsigned int veirq;
 69
 70	if (eirq < 1 || eirq > 0xf) {
 71		printk(KERN_ERR "LEON EXT IRQ NUMBER BAD: %d\n", eirq);
 72		return;
 73	}
 74
 75	veirq = leon_build_device_irq(eirq, leon_handle_ext_irq, "extirq", 0);
 76
 77	/*
 78	 * Unmask the Extended IRQ, the IRQs routed through the Ext-IRQ
 79	 * controller have a mask-bit of their own, so this is safe.
 80	 */
 81	irq_link(veirq);
 82	mask = 1 << eirq;
 83	oldmask = LEON3_BYPASS_LOAD_PA(LEON_IMASK(boot_cpu_id));
 84	LEON3_BYPASS_STORE_PA(LEON_IMASK(boot_cpu_id), (oldmask | mask));
 85	sparc_leon_eirq = eirq;
 86}
 87
 88static inline unsigned long get_irqmask(unsigned int irq)
 89{
 90	unsigned long mask;
 91
 92	if (!irq || ((irq > 0xf) && !sparc_leon_eirq)
 93	    || ((irq > 0x1f) && sparc_leon_eirq)) {
 94		printk(KERN_ERR
 95		       "leon_get_irqmask: false irq number: %d\n", irq);
 96		mask = 0;
 97	} else {
 98		mask = LEON_HARD_INT(irq);
 99	}
100	return mask;
101}
102
103#ifdef CONFIG_SMP
104static int irq_choose_cpu(const struct cpumask *affinity)
105{
106	cpumask_t mask;
107
108	cpus_and(mask, cpu_online_map, *affinity);
109	if (cpus_equal(mask, cpu_online_map) || cpus_empty(mask))
110		return boot_cpu_id;
111	else
112		return first_cpu(mask);
113}
114#else
115#define irq_choose_cpu(affinity) boot_cpu_id
116#endif
117
118static int leon_set_affinity(struct irq_data *data, const struct cpumask *dest,
119			     bool force)
120{
121	unsigned long mask, oldmask, flags;
122	int oldcpu, newcpu;
123
124	mask = (unsigned long)data->chip_data;
125	oldcpu = irq_choose_cpu(data->affinity);
126	newcpu = irq_choose_cpu(dest);
127
128	if (oldcpu == newcpu)
129		goto out;
130
131	/* unmask on old CPU first before enabling on the selected CPU */
132	spin_lock_irqsave(&leon_irq_lock, flags);
133	oldmask = LEON3_BYPASS_LOAD_PA(LEON_IMASK(oldcpu));
134	LEON3_BYPASS_STORE_PA(LEON_IMASK(oldcpu), (oldmask & ~mask));
135	oldmask = LEON3_BYPASS_LOAD_PA(LEON_IMASK(newcpu));
136	LEON3_BYPASS_STORE_PA(LEON_IMASK(newcpu), (oldmask | mask));
137	spin_unlock_irqrestore(&leon_irq_lock, flags);
138out:
139	return IRQ_SET_MASK_OK;
140}
141
142static void leon_unmask_irq(struct irq_data *data)
143{
144	unsigned long mask, oldmask, flags;
145	int cpu;
146
147	mask = (unsigned long)data->chip_data;
148	cpu = irq_choose_cpu(data->affinity);
149	spin_lock_irqsave(&leon_irq_lock, flags);
150	oldmask = LEON3_BYPASS_LOAD_PA(LEON_IMASK(cpu));
151	LEON3_BYPASS_STORE_PA(LEON_IMASK(cpu), (oldmask | mask));
152	spin_unlock_irqrestore(&leon_irq_lock, flags);
153}
154
155static void leon_mask_irq(struct irq_data *data)
156{
157	unsigned long mask, oldmask, flags;
158	int cpu;
159
160	mask = (unsigned long)data->chip_data;
161	cpu = irq_choose_cpu(data->affinity);
162	spin_lock_irqsave(&leon_irq_lock, flags);
163	oldmask = LEON3_BYPASS_LOAD_PA(LEON_IMASK(cpu));
164	LEON3_BYPASS_STORE_PA(LEON_IMASK(cpu), (oldmask & ~mask));
165	spin_unlock_irqrestore(&leon_irq_lock, flags);
166}
167
168static unsigned int leon_startup_irq(struct irq_data *data)
169{
170	irq_link(data->irq);
171	leon_unmask_irq(data);
172	return 0;
173}
174
175static void leon_shutdown_irq(struct irq_data *data)
176{
177	leon_mask_irq(data);
178	irq_unlink(data->irq);
179}
180
181/* Used by external level sensitive IRQ handlers on the LEON: ACK IRQ ctrl */
182static void leon_eoi_irq(struct irq_data *data)
183{
184	unsigned long mask = (unsigned long)data->chip_data;
185
186	if (mask & LEON_DO_ACK_HW)
187		LEON3_BYPASS_STORE_PA(LEON_IACK, mask & ~LEON_DO_ACK_HW);
188}
189
190static struct irq_chip leon_irq = {
191	.name			= "leon",
192	.irq_startup		= leon_startup_irq,
193	.irq_shutdown		= leon_shutdown_irq,
194	.irq_mask		= leon_mask_irq,
195	.irq_unmask		= leon_unmask_irq,
196	.irq_eoi		= leon_eoi_irq,
197	.irq_set_affinity	= leon_set_affinity,
198};
199
200/*
201 * Build a LEON IRQ for the edge triggered LEON IRQ controller:
202 *  Edge (normal) IRQ           - handle_simple_irq, ack=DONT-CARE, never ack
203 *  Level IRQ (PCI|Level-GPIO)  - handle_fasteoi_irq, ack=1, ack after ISR
204 *  Per-CPU Edge                - handle_percpu_irq, ack=0
205 */
206unsigned int leon_build_device_irq(unsigned int real_irq,
207				    irq_flow_handler_t flow_handler,
208				    const char *name, int do_ack)
209{
210	unsigned int irq;
211	unsigned long mask;
 
212
213	irq = 0;
214	mask = get_irqmask(real_irq);
215	if (mask == 0)
216		goto out;
217
218	irq = irq_alloc(real_irq, real_irq);
219	if (irq == 0)
220		goto out;
221
222	if (do_ack)
223		mask |= LEON_DO_ACK_HW;
224
225	irq_set_chip_and_handler_name(irq, &leon_irq,
226				      flow_handler, name);
227	irq_set_chip_data(irq, (void *)mask);
 
 
 
228
229out:
230	return irq;
231}
232
233static unsigned int _leon_build_device_irq(struct platform_device *op,
234					   unsigned int real_irq)
235{
236	return leon_build_device_irq(real_irq, handle_simple_irq, "edge", 0);
237}
238
239void leon_update_virq_handling(unsigned int virq,
240			      irq_flow_handler_t flow_handler,
241			      const char *name, int do_ack)
242{
243	unsigned long mask = (unsigned long)irq_get_chip_data(virq);
244
245	mask &= ~LEON_DO_ACK_HW;
246	if (do_ack)
247		mask |= LEON_DO_ACK_HW;
248
249	irq_set_chip_and_handler_name(virq, &leon_irq,
250				      flow_handler, name);
251	irq_set_chip_data(virq, (void *)mask);
252}
253
254void __init leon_init_timers(irq_handler_t counter_fn)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
255{
256	int irq, eirq;
257	struct device_node *rootnp, *np, *nnp;
258	struct property *pp;
259	int len;
260	int icsel;
261	int ampopts;
262	int err;
 
 
 
 
 
 
 
 
 
 
263
264	leondebug_irq_disable = 0;
265	leon_debug_irqout = 0;
266	master_l10_counter = (unsigned int *)&dummy_master_l10_counter;
267	dummy_master_l10_counter = 0;
268
269	rootnp = of_find_node_by_path("/ambapp0");
270	if (!rootnp)
271		goto bad;
272
273	/* Find System ID: GRLIB build ID and optional CHIP ID */
274	pp = of_find_property(rootnp, "systemid", &len);
275	if (pp)
276		amba_system_id = *(unsigned long *)pp->value;
277
278	/* Find IRQMP IRQ Controller Registers base adr otherwise bail out */
279	np = of_find_node_by_name(rootnp, "GAISLER_IRQMP");
280	if (!np) {
281		np = of_find_node_by_name(rootnp, "01_00d");
282		if (!np)
283			goto bad;
284	}
285	pp = of_find_property(np, "reg", &len);
286	if (!pp)
287		goto bad;
288	leon3_irqctrl_regs = *(struct leon3_irqctrl_regs_map **)pp->value;
289
290	/* Find GPTIMER Timer Registers base address otherwise bail out. */
291	nnp = rootnp;
292	do {
293		np = of_find_node_by_name(nnp, "GAISLER_GPTIMER");
294		if (!np) {
295			np = of_find_node_by_name(nnp, "01_011");
296			if (!np)
297				goto bad;
298		}
299
300		ampopts = 0;
301		pp = of_find_property(np, "ampopts", &len);
302		if (pp) {
303			ampopts = *(int *)pp->value;
304			if (ampopts == 0) {
305				/* Skip this instance, resource already
306				 * allocated by other OS */
307				nnp = np;
308				continue;
309			}
 
 
 
 
 
 
 
310		}
 
311
312		/* Select Timer-Instance on Timer Core. Default is zero */
313		leon3_gptimer_idx = ampopts & 0x7;
314
315		pp = of_find_property(np, "reg", &len);
316		if (pp)
317			leon3_gptimer_regs = *(struct leon3_gptimer_regs_map **)
318						pp->value;
319		pp = of_find_property(np, "interrupts", &len);
320		if (pp)
321			leon3_gptimer_irq = *(unsigned int *)pp->value;
322	} while (0);
323
324	if (!(leon3_gptimer_regs && leon3_irqctrl_regs && leon3_gptimer_irq))
325		goto bad;
326
 
 
 
 
 
 
 
 
 
 
327	LEON3_BYPASS_STORE_PA(&leon3_gptimer_regs->e[leon3_gptimer_idx].val, 0);
328	LEON3_BYPASS_STORE_PA(&leon3_gptimer_regs->e[leon3_gptimer_idx].rld,
329				(((1000000 / HZ) - 1)));
330	LEON3_BYPASS_STORE_PA(
331			&leon3_gptimer_regs->e[leon3_gptimer_idx].ctrl, 0);
332
333#ifdef CONFIG_SMP
334	leon3_ticker_irq = leon3_gptimer_irq + 1 + leon3_gptimer_idx;
335
336	if (!(LEON3_BYPASS_LOAD_PA(&leon3_gptimer_regs->config) &
337	      (1<<LEON3_GPTIMER_SEPIRQ))) {
338		printk(KERN_ERR "timer not configured with separate irqs\n");
339		BUG();
340	}
341
342	LEON3_BYPASS_STORE_PA(&leon3_gptimer_regs->e[leon3_gptimer_idx+1].val,
343				0);
344	LEON3_BYPASS_STORE_PA(&leon3_gptimer_regs->e[leon3_gptimer_idx+1].rld,
345				(((1000000/HZ) - 1)));
346	LEON3_BYPASS_STORE_PA(&leon3_gptimer_regs->e[leon3_gptimer_idx+1].ctrl,
347				0);
348#endif
349
350	/*
351	 * The IRQ controller may (if implemented) consist of multiple
352	 * IRQ controllers, each mapped on a 4Kb boundary.
353	 * Each CPU may be routed to different IRQCTRLs, however
354	 * we assume that all CPUs (in SMP system) is routed to the
355	 * same IRQ Controller, and for non-SMP only one IRQCTRL is
356	 * accessed anyway.
357	 * In AMP systems, Linux must run on CPU0 for the time being.
358	 */
359	icsel = LEON3_BYPASS_LOAD_PA(&leon3_irqctrl_regs->icsel[boot_cpu_id/8]);
360	icsel = (icsel >> ((7 - (boot_cpu_id&0x7)) * 4)) & 0xf;
361	leon3_irqctrl_regs += icsel;
362
363	/* Mask all IRQs on boot-cpu IRQ controller */
364	LEON3_BYPASS_STORE_PA(&leon3_irqctrl_regs->mask[boot_cpu_id], 0);
365
366	/* Probe extended IRQ controller */
367	eirq = (LEON3_BYPASS_LOAD_PA(&leon3_irqctrl_regs->mpstatus)
368		>> 16) & 0xf;
369	if (eirq != 0)
370		leon_eirq_setup(eirq);
371
372	irq = _leon_build_device_irq(NULL, leon3_gptimer_irq+leon3_gptimer_idx);
373	err = request_irq(irq, counter_fn, IRQF_TIMER, "timer", NULL);
374	if (err) {
375		printk(KERN_ERR "unable to attach timer IRQ%d\n", irq);
376		prom_halt();
377	}
378
379#ifdef CONFIG_SMP
380	{
381		unsigned long flags;
382
383		/*
384		 * In SMP, sun4m adds a IPI handler to IRQ trap handler that
385		 * LEON never must take, sun4d and LEON overwrites the branch
386		 * with a NOP.
387		 */
388		local_irq_save(flags);
389		patchme_maybe_smp_msg[0] = 0x01000000; /* NOP out the branch */
390		local_flush_cache_all();
391		local_irq_restore(flags);
392	}
393#endif
394
395	LEON3_BYPASS_STORE_PA(&leon3_gptimer_regs->e[leon3_gptimer_idx].ctrl,
396			      LEON3_GPTIMER_EN |
397			      LEON3_GPTIMER_RL |
398			      LEON3_GPTIMER_LD |
399			      LEON3_GPTIMER_IRQEN);
400
401#ifdef CONFIG_SMP
402	/* Install per-cpu IRQ handler for broadcasted ticker */
403	irq = leon_build_device_irq(leon3_ticker_irq, handle_percpu_irq,
404				    "per-cpu", 0);
405	err = request_irq(irq, leon_percpu_timer_interrupt,
406			  IRQF_PERCPU | IRQF_TIMER, "ticker",
407			  NULL);
 
 
 
408	if (err) {
409		printk(KERN_ERR "unable to attach ticker IRQ%d\n", irq);
410		prom_halt();
411	}
412
413	LEON3_BYPASS_STORE_PA(&leon3_gptimer_regs->e[leon3_gptimer_idx+1].ctrl,
414			      LEON3_GPTIMER_EN |
415			      LEON3_GPTIMER_RL |
416			      LEON3_GPTIMER_LD |
417			      LEON3_GPTIMER_IRQEN);
418#endif
419	return;
420bad:
421	printk(KERN_ERR "No Timer/irqctrl found\n");
422	BUG();
423	return;
424}
425
426void leon_clear_clock_irq(void)
427{
 
 
 
 
 
428}
429
430void leon_load_profile_irq(int cpu, unsigned int limit)
431{
432	BUG();
433}
434
435void __init leon_trans_init(struct device_node *dp)
436{
437	if (strcmp(dp->type, "cpu") == 0 && strcmp(dp->name, "<NULL>") == 0) {
438		struct property *p;
439		p = of_find_property(dp, "mid", (void *)0);
440		if (p) {
441			int mid;
442			dp->name = prom_early_alloc(5 + 1);
443			memcpy(&mid, p->value, p->length);
444			sprintf((char *)dp->name, "cpu%.2d", mid);
445		}
446	}
447}
448
449void __initdata (*prom_amba_init)(struct device_node *dp, struct device_node ***nextp) = 0;
450
451void __init leon_node_init(struct device_node *dp, struct device_node ***nextp)
452{
453	if (prom_amba_init &&
454	    strcmp(dp->type, "ambapp") == 0 &&
455	    strcmp(dp->name, "ambapp0") == 0) {
456		prom_amba_init(dp, nextp);
457	}
458}
459
460#ifdef CONFIG_SMP
461
462void leon_set_cpu_int(int cpu, int level)
463{
464	unsigned long mask;
465	mask = get_irqmask(level);
466	LEON3_BYPASS_STORE_PA(&leon3_irqctrl_regs->force[cpu], mask);
467}
468
469static void leon_clear_ipi(int cpu, int level)
470{
471	unsigned long mask;
472	mask = get_irqmask(level);
473	LEON3_BYPASS_STORE_PA(&leon3_irqctrl_regs->force[cpu], mask<<16);
474}
475
476static void leon_set_udt(int cpu)
477{
478}
479
480void leon_clear_profile_irq(int cpu)
481{
482}
483
484void leon_enable_irq_cpu(unsigned int irq_nr, unsigned int cpu)
485{
486	unsigned long mask, flags, *addr;
487	mask = get_irqmask(irq_nr);
488	spin_lock_irqsave(&leon_irq_lock, flags);
489	addr = (unsigned long *)LEON_IMASK(cpu);
490	LEON3_BYPASS_STORE_PA(addr, (LEON3_BYPASS_LOAD_PA(addr) | mask));
491	spin_unlock_irqrestore(&leon_irq_lock, flags);
492}
493
494#endif
495
496void __init leon_init_IRQ(void)
497{
498	sparc_irq_config.init_timers      = leon_init_timers;
499	sparc_irq_config.build_device_irq = _leon_build_device_irq;
500
501	BTFIXUPSET_CALL(clear_clock_irq, leon_clear_clock_irq,
502			BTFIXUPCALL_NORM);
503	BTFIXUPSET_CALL(load_profile_irq, leon_load_profile_irq,
504			BTFIXUPCALL_NOP);
505
506#ifdef CONFIG_SMP
507	BTFIXUPSET_CALL(set_cpu_int, leon_set_cpu_int, BTFIXUPCALL_NORM);
508	BTFIXUPSET_CALL(clear_cpu_int, leon_clear_ipi, BTFIXUPCALL_NORM);
509	BTFIXUPSET_CALL(set_irq_udt, leon_set_udt, BTFIXUPCALL_NORM);
510#endif
511
512}
513
514void __init leon_init(void)
515{
516	of_pdt_build_more = &leon_node_init;
517}