Linux Audio

Check our new training course

Loading...
v5.4
  1// SPDX-License-Identifier: GPL-2.0
  2/*
  3 * Low-Level PCI Support for the SH7780
  4 *
  5 *  Copyright (C) 2005 - 2010  Paul Mundt
 
 
 
 
  6 */
  7#include <linux/types.h>
  8#include <linux/kernel.h>
  9#include <linux/init.h>
 10#include <linux/pci.h>
 11#include <linux/interrupt.h>
 12#include <linux/timer.h>
 13#include <linux/irq.h>
 14#include <linux/errno.h>
 15#include <linux/delay.h>
 16#include <linux/log2.h>
 17#include "pci-sh4.h"
 18#include <asm/mmu.h>
 19#include <linux/sizes.h>
 20
 21#if defined(CONFIG_CPU_BIG_ENDIAN)
 22# define PCICR_ENDIANNESS SH4_PCICR_BSWP
 23#else
 24# define PCICR_ENDIANNESS 0
 25#endif
 26
 27
 28static struct resource sh7785_pci_resources[] = {
 29	{
 30		.name	= "PCI IO",
 31		.start	= 0x1000,
 32		.end	= SZ_4M - 1,
 33		.flags	= IORESOURCE_IO,
 34	}, {
 35		.name	= "PCI MEM 0",
 36		.start	= 0xfd000000,
 37		.end	= 0xfd000000 + SZ_16M - 1,
 38		.flags	= IORESOURCE_MEM,
 39	}, {
 40		.name	= "PCI MEM 1",
 41		.start	= 0x10000000,
 42		.end	= 0x10000000 + SZ_64M - 1,
 43		.flags	= IORESOURCE_MEM,
 44	}, {
 45		/*
 46		 * 32-bit only resources must be last.
 47		 */
 48		.name	= "PCI MEM 2",
 49		.start	= 0xc0000000,
 50		.end	= 0xc0000000 + SZ_512M - 1,
 51		.flags	= IORESOURCE_MEM | IORESOURCE_MEM_32BIT,
 52	},
 53};
 54
 55static struct pci_channel sh7780_pci_controller = {
 56	.pci_ops	= &sh4_pci_ops,
 57	.resources	= sh7785_pci_resources,
 58	.nr_resources	= ARRAY_SIZE(sh7785_pci_resources),
 59	.io_offset	= 0,
 60	.mem_offset	= 0,
 61	.io_map_base	= 0xfe200000,
 62	.serr_irq	= evt2irq(0xa00),
 63	.err_irq	= evt2irq(0xaa0),
 64};
 65
 66struct pci_errors {
 67	unsigned int	mask;
 68	const char	*str;
 69} pci_arbiter_errors[] = {
 70	{ SH4_PCIAINT_MBKN,	"master broken" },
 71	{ SH4_PCIAINT_TBTO,	"target bus time out" },
 72	{ SH4_PCIAINT_MBTO,	"master bus time out" },
 73	{ SH4_PCIAINT_TABT,	"target abort" },
 74	{ SH4_PCIAINT_MABT,	"master abort" },
 75	{ SH4_PCIAINT_RDPE,	"read data parity error" },
 76	{ SH4_PCIAINT_WDPE,	"write data parity error" },
 77}, pci_interrupt_errors[] = {
 78	{ SH4_PCIINT_MLCK,	"master lock error" },
 79	{ SH4_PCIINT_TABT,	"target-target abort" },
 80	{ SH4_PCIINT_TRET,	"target retry time out" },
 81	{ SH4_PCIINT_MFDE,	"master function disable error" },
 82	{ SH4_PCIINT_PRTY,	"address parity error" },
 83	{ SH4_PCIINT_SERR,	"SERR" },
 84	{ SH4_PCIINT_TWDP,	"data parity error for target write" },
 85	{ SH4_PCIINT_TRDP,	"PERR detected for target read" },
 86	{ SH4_PCIINT_MTABT,	"target abort for master" },
 87	{ SH4_PCIINT_MMABT,	"master abort for master" },
 88	{ SH4_PCIINT_MWPD,	"master write data parity error" },
 89	{ SH4_PCIINT_MRPD,	"master read data parity error" },
 90};
 91
 92static irqreturn_t sh7780_pci_err_irq(int irq, void *dev_id)
 93{
 94	struct pci_channel *hose = dev_id;
 95	unsigned long addr;
 96	unsigned int status;
 97	unsigned int cmd;
 98	int i;
 99
100	addr = __raw_readl(hose->reg_base + SH4_PCIALR);
101
102	/*
103	 * Handle status errors.
104	 */
105	status = __raw_readw(hose->reg_base + PCI_STATUS);
106	if (status & (PCI_STATUS_PARITY |
107		      PCI_STATUS_DETECTED_PARITY |
108		      PCI_STATUS_SIG_TARGET_ABORT |
109		      PCI_STATUS_REC_TARGET_ABORT |
110		      PCI_STATUS_REC_MASTER_ABORT)) {
111		cmd = pcibios_handle_status_errors(addr, status, hose);
112		if (likely(cmd))
113			__raw_writew(cmd, hose->reg_base + PCI_STATUS);
114	}
115
116	/*
117	 * Handle arbiter errors.
118	 */
119	status = __raw_readl(hose->reg_base + SH4_PCIAINT);
120	for (i = cmd = 0; i < ARRAY_SIZE(pci_arbiter_errors); i++) {
121		if (status & pci_arbiter_errors[i].mask) {
122			printk(KERN_DEBUG "PCI: %s, addr=%08lx\n",
123			       pci_arbiter_errors[i].str, addr);
124			cmd |= pci_arbiter_errors[i].mask;
125		}
126	}
127	__raw_writel(cmd, hose->reg_base + SH4_PCIAINT);
128
129	/*
130	 * Handle the remaining PCI errors.
131	 */
132	status = __raw_readl(hose->reg_base + SH4_PCIINT);
133	for (i = cmd = 0; i < ARRAY_SIZE(pci_interrupt_errors); i++) {
134		if (status & pci_interrupt_errors[i].mask) {
135			printk(KERN_DEBUG "PCI: %s, addr=%08lx\n",
136			       pci_interrupt_errors[i].str, addr);
137			cmd |= pci_interrupt_errors[i].mask;
138		}
139	}
140	__raw_writel(cmd, hose->reg_base + SH4_PCIINT);
141
142	return IRQ_HANDLED;
143}
144
145static irqreturn_t sh7780_pci_serr_irq(int irq, void *dev_id)
146{
147	struct pci_channel *hose = dev_id;
148
149	printk(KERN_DEBUG "PCI: system error received: ");
150	pcibios_report_status(PCI_STATUS_SIG_SYSTEM_ERROR, 1);
151	printk("\n");
152
153	/* Deassert SERR */
154	__raw_writel(SH4_PCIINTM_SDIM, hose->reg_base + SH4_PCIINTM);
155
156	/* Back off the IRQ for awhile */
157	disable_irq_nosync(irq);
158	hose->serr_timer.expires = jiffies + HZ;
159	add_timer(&hose->serr_timer);
160
161	return IRQ_HANDLED;
162}
163
164static int __init sh7780_pci_setup_irqs(struct pci_channel *hose)
165{
166	int ret;
167
168	/* Clear out PCI arbiter IRQs */
169	__raw_writel(0, hose->reg_base + SH4_PCIAINT);
170
171	/* Clear all error conditions */
172	__raw_writew(PCI_STATUS_DETECTED_PARITY  | \
173		     PCI_STATUS_SIG_SYSTEM_ERROR | \
174		     PCI_STATUS_REC_MASTER_ABORT | \
175		     PCI_STATUS_REC_TARGET_ABORT | \
176		     PCI_STATUS_SIG_TARGET_ABORT | \
177		     PCI_STATUS_PARITY, hose->reg_base + PCI_STATUS);
178
179	ret = request_irq(hose->serr_irq, sh7780_pci_serr_irq, 0,
180			  "PCI SERR interrupt", hose);
181	if (unlikely(ret)) {
182		printk(KERN_ERR "PCI: Failed hooking SERR IRQ\n");
183		return ret;
184	}
185
186	/*
187	 * The PCI ERR IRQ needs to be IRQF_SHARED since all of the power
188	 * down IRQ vectors are routed through the ERR IRQ vector. We
189	 * only request_irq() once as there is only a single masking
190	 * source for multiple events.
191	 */
192	ret = request_irq(hose->err_irq, sh7780_pci_err_irq, IRQF_SHARED,
193			  "PCI ERR interrupt", hose);
194	if (unlikely(ret)) {
195		free_irq(hose->serr_irq, hose);
196		return ret;
197	}
198
199	/* Unmask all of the arbiter IRQs. */
200	__raw_writel(SH4_PCIAINT_MBKN | SH4_PCIAINT_TBTO | SH4_PCIAINT_MBTO | \
201		     SH4_PCIAINT_TABT | SH4_PCIAINT_MABT | SH4_PCIAINT_RDPE | \
202		     SH4_PCIAINT_WDPE, hose->reg_base + SH4_PCIAINTM);
203
204	/* Unmask all of the PCI IRQs */
205	__raw_writel(SH4_PCIINTM_TTADIM  | SH4_PCIINTM_TMTOIM  | \
206		     SH4_PCIINTM_MDEIM   | SH4_PCIINTM_APEDIM  | \
207		     SH4_PCIINTM_SDIM    | SH4_PCIINTM_DPEITWM | \
208		     SH4_PCIINTM_PEDITRM | SH4_PCIINTM_TADIMM  | \
209		     SH4_PCIINTM_MADIMM  | SH4_PCIINTM_MWPDIM  | \
210		     SH4_PCIINTM_MRDPEIM, hose->reg_base + SH4_PCIINTM);
211
212	return ret;
213}
214
215static inline void __init sh7780_pci_teardown_irqs(struct pci_channel *hose)
216{
217	free_irq(hose->err_irq, hose);
218	free_irq(hose->serr_irq, hose);
219}
220
221static void __init sh7780_pci66_init(struct pci_channel *hose)
222{
223	unsigned int tmp;
224
225	if (!pci_is_66mhz_capable(hose, 0, 0))
226		return;
227
228	/* Enable register access */
229	tmp = __raw_readl(hose->reg_base + SH4_PCICR);
230	tmp |= SH4_PCICR_PREFIX;
231	__raw_writel(tmp, hose->reg_base + SH4_PCICR);
232
233	/* Enable 66MHz operation */
234	tmp = __raw_readw(hose->reg_base + PCI_STATUS);
235	tmp |= PCI_STATUS_66MHZ;
236	__raw_writew(tmp, hose->reg_base + PCI_STATUS);
237
238	/* Done */
239	tmp = __raw_readl(hose->reg_base + SH4_PCICR);
240	tmp |= SH4_PCICR_PREFIX | SH4_PCICR_CFIN;
241	__raw_writel(tmp, hose->reg_base + SH4_PCICR);
242}
243
244static int __init sh7780_pci_init(void)
245{
246	struct pci_channel *chan = &sh7780_pci_controller;
247	phys_addr_t memphys;
248	size_t memsize;
249	unsigned int id;
250	const char *type;
251	int ret, i;
252
253	printk(KERN_NOTICE "PCI: Starting initialization.\n");
254
255	chan->reg_base = 0xfe040000;
256
257	/* Enable CPU access to the PCIC registers. */
258	__raw_writel(PCIECR_ENBL, PCIECR);
259
260	/* Reset */
261	__raw_writel(SH4_PCICR_PREFIX | SH4_PCICR_PRST | PCICR_ENDIANNESS,
262		     chan->reg_base + SH4_PCICR);
263
264	/*
265	 * Wait for it to come back up. The spec says to allow for up to
266	 * 1 second after toggling the reset pin, but in practice 100ms
267	 * is more than enough.
268	 */
269	mdelay(100);
270
271	id = __raw_readw(chan->reg_base + PCI_VENDOR_ID);
272	if (id != PCI_VENDOR_ID_RENESAS) {
273		printk(KERN_ERR "PCI: Unknown vendor ID 0x%04x.\n", id);
274		return -ENODEV;
275	}
276
277	id = __raw_readw(chan->reg_base + PCI_DEVICE_ID);
278	type = (id == PCI_DEVICE_ID_RENESAS_SH7763) ? "SH7763" :
279	       (id == PCI_DEVICE_ID_RENESAS_SH7780) ? "SH7780" :
280	       (id == PCI_DEVICE_ID_RENESAS_SH7781) ? "SH7781" :
281	       (id == PCI_DEVICE_ID_RENESAS_SH7785) ? "SH7785" :
282					  NULL;
283	if (unlikely(!type)) {
284		printk(KERN_ERR "PCI: Found an unsupported Renesas host "
285		       "controller, device id 0x%04x.\n", id);
286		return -EINVAL;
287	}
288
289	printk(KERN_NOTICE "PCI: Found a Renesas %s host "
290	       "controller, revision %d.\n", type,
291	       __raw_readb(chan->reg_base + PCI_REVISION_ID));
292
293	/*
294	 * Now throw it in to register initialization mode and
295	 * start the real work.
296	 */
297	__raw_writel(SH4_PCICR_PREFIX | PCICR_ENDIANNESS,
298		     chan->reg_base + SH4_PCICR);
299
300	memphys = __pa(memory_start);
301	memsize = roundup_pow_of_two(memory_end - memory_start);
302
303	/*
304	 * If there's more than 512MB of memory, we need to roll over to
305	 * LAR1/LSR1.
306	 */
307	if (memsize > SZ_512M) {
308		__raw_writel(memphys + SZ_512M, chan->reg_base + SH4_PCILAR1);
309		__raw_writel((((memsize - SZ_512M) - SZ_1M) & 0x1ff00000) | 1,
310			     chan->reg_base + SH4_PCILSR1);
311		memsize = SZ_512M;
312	} else {
313		/*
314		 * Otherwise just zero it out and disable it.
315		 */
316		__raw_writel(0, chan->reg_base + SH4_PCILAR1);
317		__raw_writel(0, chan->reg_base + SH4_PCILSR1);
318	}
319
320	/*
321	 * LAR0/LSR0 covers up to the first 512MB, which is enough to
322	 * cover all of lowmem on most platforms.
323	 */
324	__raw_writel(memphys, chan->reg_base + SH4_PCILAR0);
325	__raw_writel(((memsize - SZ_1M) & 0x1ff00000) | 1,
326		     chan->reg_base + SH4_PCILSR0);
327
328	/*
329	 * Hook up the ERR and SERR IRQs.
330	 */
331	ret = sh7780_pci_setup_irqs(chan);
332	if (unlikely(ret))
333		return ret;
334
335	/*
336	 * Disable the cache snoop controller for non-coherent DMA.
337	 */
338	__raw_writel(0, chan->reg_base + SH7780_PCICSCR0);
339	__raw_writel(0, chan->reg_base + SH7780_PCICSAR0);
340	__raw_writel(0, chan->reg_base + SH7780_PCICSCR1);
341	__raw_writel(0, chan->reg_base + SH7780_PCICSAR1);
342
343	/*
344	 * Setup the memory BARs
345	 */
346	for (i = 1; i < chan->nr_resources; i++) {
347		struct resource *res = chan->resources + i;
348		resource_size_t size;
349
350		if (unlikely(res->flags & IORESOURCE_IO))
351			continue;
352
353		/*
354		 * Make sure we're in the right physical addressing mode
355		 * for dealing with the resource.
356		 */
357		if ((res->flags & IORESOURCE_MEM_32BIT) && __in_29bit_mode()) {
358			chan->nr_resources--;
359			continue;
360		}
361
362		size = resource_size(res);
363
364		/*
365		 * The MBMR mask is calculated in units of 256kB, which
366		 * keeps things pretty simple.
367		 */
368		__raw_writel(((roundup_pow_of_two(size) / SZ_256K) - 1) << 18,
369			     chan->reg_base + SH7780_PCIMBMR(i - 1));
370		__raw_writel(res->start, chan->reg_base + SH7780_PCIMBR(i - 1));
371	}
372
373	/*
374	 * And I/O.
375	 */
376	__raw_writel(0, chan->reg_base + PCI_BASE_ADDRESS_0);
377	__raw_writel(0, chan->reg_base + SH7780_PCIIOBR);
378	__raw_writel(0, chan->reg_base + SH7780_PCIIOBMR);
379
380	__raw_writew(PCI_COMMAND_SERR   | PCI_COMMAND_WAIT   | \
381		     PCI_COMMAND_PARITY | PCI_COMMAND_MASTER | \
382		     PCI_COMMAND_MEMORY, chan->reg_base + PCI_COMMAND);
383
384	/*
385	 * Initialization mode complete, release the control register and
386	 * enable round robin mode to stop device overruns/starvation.
387	 */
388	__raw_writel(SH4_PCICR_PREFIX | SH4_PCICR_CFIN | SH4_PCICR_FTO |
389		     PCICR_ENDIANNESS,
390		     chan->reg_base + SH4_PCICR);
391
392	ret = register_pci_controller(chan);
393	if (unlikely(ret))
394		goto err;
395
396	sh7780_pci66_init(chan);
397
398	printk(KERN_NOTICE "PCI: Running at %dMHz.\n",
399	       (__raw_readw(chan->reg_base + PCI_STATUS) & PCI_STATUS_66MHZ) ?
400	       66 : 33);
401
402	return 0;
403
404err:
405	sh7780_pci_teardown_irqs(chan);
406	return ret;
407}
408arch_initcall(sh7780_pci_init);
v3.5.6
 
  1/*
  2 * Low-Level PCI Support for the SH7780
  3 *
  4 *  Copyright (C) 2005 - 2010  Paul Mundt
  5 *
  6 * This file is subject to the terms and conditions of the GNU General Public
  7 * License.  See the file "COPYING" in the main directory of this archive
  8 * for more details.
  9 */
 10#include <linux/types.h>
 11#include <linux/kernel.h>
 12#include <linux/init.h>
 13#include <linux/pci.h>
 14#include <linux/interrupt.h>
 15#include <linux/timer.h>
 16#include <linux/irq.h>
 17#include <linux/errno.h>
 18#include <linux/delay.h>
 19#include <linux/log2.h>
 20#include "pci-sh4.h"
 21#include <asm/mmu.h>
 22#include <asm/sizes.h>
 23
 24#if defined(CONFIG_CPU_BIG_ENDIAN)
 25# define PCICR_ENDIANNESS SH4_PCICR_BSWP
 26#else
 27# define PCICR_ENDIANNESS 0
 28#endif
 29
 30
 31static struct resource sh7785_pci_resources[] = {
 32	{
 33		.name	= "PCI IO",
 34		.start	= 0x1000,
 35		.end	= SZ_4M - 1,
 36		.flags	= IORESOURCE_IO,
 37	}, {
 38		.name	= "PCI MEM 0",
 39		.start	= 0xfd000000,
 40		.end	= 0xfd000000 + SZ_16M - 1,
 41		.flags	= IORESOURCE_MEM,
 42	}, {
 43		.name	= "PCI MEM 1",
 44		.start	= 0x10000000,
 45		.end	= 0x10000000 + SZ_64M - 1,
 46		.flags	= IORESOURCE_MEM,
 47	}, {
 48		/*
 49		 * 32-bit only resources must be last.
 50		 */
 51		.name	= "PCI MEM 2",
 52		.start	= 0xc0000000,
 53		.end	= 0xc0000000 + SZ_512M - 1,
 54		.flags	= IORESOURCE_MEM | IORESOURCE_MEM_32BIT,
 55	},
 56};
 57
 58static struct pci_channel sh7780_pci_controller = {
 59	.pci_ops	= &sh4_pci_ops,
 60	.resources	= sh7785_pci_resources,
 61	.nr_resources	= ARRAY_SIZE(sh7785_pci_resources),
 62	.io_offset	= 0,
 63	.mem_offset	= 0,
 64	.io_map_base	= 0xfe200000,
 65	.serr_irq	= evt2irq(0xa00),
 66	.err_irq	= evt2irq(0xaa0),
 67};
 68
 69struct pci_errors {
 70	unsigned int	mask;
 71	const char	*str;
 72} pci_arbiter_errors[] = {
 73	{ SH4_PCIAINT_MBKN,	"master broken" },
 74	{ SH4_PCIAINT_TBTO,	"target bus time out" },
 75	{ SH4_PCIAINT_MBTO,	"master bus time out" },
 76	{ SH4_PCIAINT_TABT,	"target abort" },
 77	{ SH4_PCIAINT_MABT,	"master abort" },
 78	{ SH4_PCIAINT_RDPE,	"read data parity error" },
 79	{ SH4_PCIAINT_WDPE,	"write data parity error" },
 80}, pci_interrupt_errors[] = {
 81	{ SH4_PCIINT_MLCK,	"master lock error" },
 82	{ SH4_PCIINT_TABT,	"target-target abort" },
 83	{ SH4_PCIINT_TRET,	"target retry time out" },
 84	{ SH4_PCIINT_MFDE,	"master function disable error" },
 85	{ SH4_PCIINT_PRTY,	"address parity error" },
 86	{ SH4_PCIINT_SERR,	"SERR" },
 87	{ SH4_PCIINT_TWDP,	"data parity error for target write" },
 88	{ SH4_PCIINT_TRDP,	"PERR detected for target read" },
 89	{ SH4_PCIINT_MTABT,	"target abort for master" },
 90	{ SH4_PCIINT_MMABT,	"master abort for master" },
 91	{ SH4_PCIINT_MWPD,	"master write data parity error" },
 92	{ SH4_PCIINT_MRPD,	"master read data parity error" },
 93};
 94
 95static irqreturn_t sh7780_pci_err_irq(int irq, void *dev_id)
 96{
 97	struct pci_channel *hose = dev_id;
 98	unsigned long addr;
 99	unsigned int status;
100	unsigned int cmd;
101	int i;
102
103	addr = __raw_readl(hose->reg_base + SH4_PCIALR);
104
105	/*
106	 * Handle status errors.
107	 */
108	status = __raw_readw(hose->reg_base + PCI_STATUS);
109	if (status & (PCI_STATUS_PARITY |
110		      PCI_STATUS_DETECTED_PARITY |
111		      PCI_STATUS_SIG_TARGET_ABORT |
112		      PCI_STATUS_REC_TARGET_ABORT |
113		      PCI_STATUS_REC_MASTER_ABORT)) {
114		cmd = pcibios_handle_status_errors(addr, status, hose);
115		if (likely(cmd))
116			__raw_writew(cmd, hose->reg_base + PCI_STATUS);
117	}
118
119	/*
120	 * Handle arbiter errors.
121	 */
122	status = __raw_readl(hose->reg_base + SH4_PCIAINT);
123	for (i = cmd = 0; i < ARRAY_SIZE(pci_arbiter_errors); i++) {
124		if (status & pci_arbiter_errors[i].mask) {
125			printk(KERN_DEBUG "PCI: %s, addr=%08lx\n",
126			       pci_arbiter_errors[i].str, addr);
127			cmd |= pci_arbiter_errors[i].mask;
128		}
129	}
130	__raw_writel(cmd, hose->reg_base + SH4_PCIAINT);
131
132	/*
133	 * Handle the remaining PCI errors.
134	 */
135	status = __raw_readl(hose->reg_base + SH4_PCIINT);
136	for (i = cmd = 0; i < ARRAY_SIZE(pci_interrupt_errors); i++) {
137		if (status & pci_interrupt_errors[i].mask) {
138			printk(KERN_DEBUG "PCI: %s, addr=%08lx\n",
139			       pci_interrupt_errors[i].str, addr);
140			cmd |= pci_interrupt_errors[i].mask;
141		}
142	}
143	__raw_writel(cmd, hose->reg_base + SH4_PCIINT);
144
145	return IRQ_HANDLED;
146}
147
148static irqreturn_t sh7780_pci_serr_irq(int irq, void *dev_id)
149{
150	struct pci_channel *hose = dev_id;
151
152	printk(KERN_DEBUG "PCI: system error received: ");
153	pcibios_report_status(PCI_STATUS_SIG_SYSTEM_ERROR, 1);
154	printk("\n");
155
156	/* Deassert SERR */
157	__raw_writel(SH4_PCIINTM_SDIM, hose->reg_base + SH4_PCIINTM);
158
159	/* Back off the IRQ for awhile */
160	disable_irq_nosync(irq);
161	hose->serr_timer.expires = jiffies + HZ;
162	add_timer(&hose->serr_timer);
163
164	return IRQ_HANDLED;
165}
166
167static int __init sh7780_pci_setup_irqs(struct pci_channel *hose)
168{
169	int ret;
170
171	/* Clear out PCI arbiter IRQs */
172	__raw_writel(0, hose->reg_base + SH4_PCIAINT);
173
174	/* Clear all error conditions */
175	__raw_writew(PCI_STATUS_DETECTED_PARITY  | \
176		     PCI_STATUS_SIG_SYSTEM_ERROR | \
177		     PCI_STATUS_REC_MASTER_ABORT | \
178		     PCI_STATUS_REC_TARGET_ABORT | \
179		     PCI_STATUS_SIG_TARGET_ABORT | \
180		     PCI_STATUS_PARITY, hose->reg_base + PCI_STATUS);
181
182	ret = request_irq(hose->serr_irq, sh7780_pci_serr_irq, 0,
183			  "PCI SERR interrupt", hose);
184	if (unlikely(ret)) {
185		printk(KERN_ERR "PCI: Failed hooking SERR IRQ\n");
186		return ret;
187	}
188
189	/*
190	 * The PCI ERR IRQ needs to be IRQF_SHARED since all of the power
191	 * down IRQ vectors are routed through the ERR IRQ vector. We
192	 * only request_irq() once as there is only a single masking
193	 * source for multiple events.
194	 */
195	ret = request_irq(hose->err_irq, sh7780_pci_err_irq, IRQF_SHARED,
196			  "PCI ERR interrupt", hose);
197	if (unlikely(ret)) {
198		free_irq(hose->serr_irq, hose);
199		return ret;
200	}
201
202	/* Unmask all of the arbiter IRQs. */
203	__raw_writel(SH4_PCIAINT_MBKN | SH4_PCIAINT_TBTO | SH4_PCIAINT_MBTO | \
204		     SH4_PCIAINT_TABT | SH4_PCIAINT_MABT | SH4_PCIAINT_RDPE | \
205		     SH4_PCIAINT_WDPE, hose->reg_base + SH4_PCIAINTM);
206
207	/* Unmask all of the PCI IRQs */
208	__raw_writel(SH4_PCIINTM_TTADIM  | SH4_PCIINTM_TMTOIM  | \
209		     SH4_PCIINTM_MDEIM   | SH4_PCIINTM_APEDIM  | \
210		     SH4_PCIINTM_SDIM    | SH4_PCIINTM_DPEITWM | \
211		     SH4_PCIINTM_PEDITRM | SH4_PCIINTM_TADIMM  | \
212		     SH4_PCIINTM_MADIMM  | SH4_PCIINTM_MWPDIM  | \
213		     SH4_PCIINTM_MRDPEIM, hose->reg_base + SH4_PCIINTM);
214
215	return ret;
216}
217
218static inline void __init sh7780_pci_teardown_irqs(struct pci_channel *hose)
219{
220	free_irq(hose->err_irq, hose);
221	free_irq(hose->serr_irq, hose);
222}
223
224static void __init sh7780_pci66_init(struct pci_channel *hose)
225{
226	unsigned int tmp;
227
228	if (!pci_is_66mhz_capable(hose, 0, 0))
229		return;
230
231	/* Enable register access */
232	tmp = __raw_readl(hose->reg_base + SH4_PCICR);
233	tmp |= SH4_PCICR_PREFIX;
234	__raw_writel(tmp, hose->reg_base + SH4_PCICR);
235
236	/* Enable 66MHz operation */
237	tmp = __raw_readw(hose->reg_base + PCI_STATUS);
238	tmp |= PCI_STATUS_66MHZ;
239	__raw_writew(tmp, hose->reg_base + PCI_STATUS);
240
241	/* Done */
242	tmp = __raw_readl(hose->reg_base + SH4_PCICR);
243	tmp |= SH4_PCICR_PREFIX | SH4_PCICR_CFIN;
244	__raw_writel(tmp, hose->reg_base + SH4_PCICR);
245}
246
247static int __init sh7780_pci_init(void)
248{
249	struct pci_channel *chan = &sh7780_pci_controller;
250	phys_addr_t memphys;
251	size_t memsize;
252	unsigned int id;
253	const char *type;
254	int ret, i;
255
256	printk(KERN_NOTICE "PCI: Starting initialization.\n");
257
258	chan->reg_base = 0xfe040000;
259
260	/* Enable CPU access to the PCIC registers. */
261	__raw_writel(PCIECR_ENBL, PCIECR);
262
263	/* Reset */
264	__raw_writel(SH4_PCICR_PREFIX | SH4_PCICR_PRST | PCICR_ENDIANNESS,
265		     chan->reg_base + SH4_PCICR);
266
267	/*
268	 * Wait for it to come back up. The spec says to allow for up to
269	 * 1 second after toggling the reset pin, but in practice 100ms
270	 * is more than enough.
271	 */
272	mdelay(100);
273
274	id = __raw_readw(chan->reg_base + PCI_VENDOR_ID);
275	if (id != PCI_VENDOR_ID_RENESAS) {
276		printk(KERN_ERR "PCI: Unknown vendor ID 0x%04x.\n", id);
277		return -ENODEV;
278	}
279
280	id = __raw_readw(chan->reg_base + PCI_DEVICE_ID);
281	type = (id == PCI_DEVICE_ID_RENESAS_SH7763) ? "SH7763" :
282	       (id == PCI_DEVICE_ID_RENESAS_SH7780) ? "SH7780" :
283	       (id == PCI_DEVICE_ID_RENESAS_SH7781) ? "SH7781" :
284	       (id == PCI_DEVICE_ID_RENESAS_SH7785) ? "SH7785" :
285					  NULL;
286	if (unlikely(!type)) {
287		printk(KERN_ERR "PCI: Found an unsupported Renesas host "
288		       "controller, device id 0x%04x.\n", id);
289		return -EINVAL;
290	}
291
292	printk(KERN_NOTICE "PCI: Found a Renesas %s host "
293	       "controller, revision %d.\n", type,
294	       __raw_readb(chan->reg_base + PCI_REVISION_ID));
295
296	/*
297	 * Now throw it in to register initialization mode and
298	 * start the real work.
299	 */
300	__raw_writel(SH4_PCICR_PREFIX | PCICR_ENDIANNESS,
301		     chan->reg_base + SH4_PCICR);
302
303	memphys = __pa(memory_start);
304	memsize = roundup_pow_of_two(memory_end - memory_start);
305
306	/*
307	 * If there's more than 512MB of memory, we need to roll over to
308	 * LAR1/LSR1.
309	 */
310	if (memsize > SZ_512M) {
311		__raw_writel(memphys + SZ_512M, chan->reg_base + SH4_PCILAR1);
312		__raw_writel((((memsize - SZ_512M) - SZ_1M) & 0x1ff00000) | 1,
313			     chan->reg_base + SH4_PCILSR1);
314		memsize = SZ_512M;
315	} else {
316		/*
317		 * Otherwise just zero it out and disable it.
318		 */
319		__raw_writel(0, chan->reg_base + SH4_PCILAR1);
320		__raw_writel(0, chan->reg_base + SH4_PCILSR1);
321	}
322
323	/*
324	 * LAR0/LSR0 covers up to the first 512MB, which is enough to
325	 * cover all of lowmem on most platforms.
326	 */
327	__raw_writel(memphys, chan->reg_base + SH4_PCILAR0);
328	__raw_writel(((memsize - SZ_1M) & 0x1ff00000) | 1,
329		     chan->reg_base + SH4_PCILSR0);
330
331	/*
332	 * Hook up the ERR and SERR IRQs.
333	 */
334	ret = sh7780_pci_setup_irqs(chan);
335	if (unlikely(ret))
336		return ret;
337
338	/*
339	 * Disable the cache snoop controller for non-coherent DMA.
340	 */
341	__raw_writel(0, chan->reg_base + SH7780_PCICSCR0);
342	__raw_writel(0, chan->reg_base + SH7780_PCICSAR0);
343	__raw_writel(0, chan->reg_base + SH7780_PCICSCR1);
344	__raw_writel(0, chan->reg_base + SH7780_PCICSAR1);
345
346	/*
347	 * Setup the memory BARs
348	 */
349	for (i = 1; i < chan->nr_resources; i++) {
350		struct resource *res = chan->resources + i;
351		resource_size_t size;
352
353		if (unlikely(res->flags & IORESOURCE_IO))
354			continue;
355
356		/*
357		 * Make sure we're in the right physical addressing mode
358		 * for dealing with the resource.
359		 */
360		if ((res->flags & IORESOURCE_MEM_32BIT) && __in_29bit_mode()) {
361			chan->nr_resources--;
362			continue;
363		}
364
365		size = resource_size(res);
366
367		/*
368		 * The MBMR mask is calculated in units of 256kB, which
369		 * keeps things pretty simple.
370		 */
371		__raw_writel(((roundup_pow_of_two(size) / SZ_256K) - 1) << 18,
372			     chan->reg_base + SH7780_PCIMBMR(i - 1));
373		__raw_writel(res->start, chan->reg_base + SH7780_PCIMBR(i - 1));
374	}
375
376	/*
377	 * And I/O.
378	 */
379	__raw_writel(0, chan->reg_base + PCI_BASE_ADDRESS_0);
380	__raw_writel(0, chan->reg_base + SH7780_PCIIOBR);
381	__raw_writel(0, chan->reg_base + SH7780_PCIIOBMR);
382
383	__raw_writew(PCI_COMMAND_SERR   | PCI_COMMAND_WAIT   | \
384		     PCI_COMMAND_PARITY | PCI_COMMAND_MASTER | \
385		     PCI_COMMAND_MEMORY, chan->reg_base + PCI_COMMAND);
386
387	/*
388	 * Initialization mode complete, release the control register and
389	 * enable round robin mode to stop device overruns/starvation.
390	 */
391	__raw_writel(SH4_PCICR_PREFIX | SH4_PCICR_CFIN | SH4_PCICR_FTO |
392		     PCICR_ENDIANNESS,
393		     chan->reg_base + SH4_PCICR);
394
395	ret = register_pci_controller(chan);
396	if (unlikely(ret))
397		goto err;
398
399	sh7780_pci66_init(chan);
400
401	printk(KERN_NOTICE "PCI: Running at %dMHz.\n",
402	       (__raw_readw(chan->reg_base + PCI_STATUS) & PCI_STATUS_66MHZ) ?
403	       66 : 33);
404
405	return 0;
406
407err:
408	sh7780_pci_teardown_irqs(chan);
409	return ret;
410}
411arch_initcall(sh7780_pci_init);