Loading...
1/*
2 * linux/arch/arm/kernel/bios32.c
3 *
4 * PCI bios-type initialisation for PCI machines
5 *
6 * Bits taken from various places.
7 */
8#include <linux/export.h>
9#include <linux/kernel.h>
10#include <linux/pci.h>
11#include <linux/slab.h>
12#include <linux/init.h>
13#include <linux/io.h>
14
15#include <asm/mach-types.h>
16#include <asm/mach/map.h>
17#include <asm/mach/pci.h>
18
19static int debug_pci;
20
21/*
22 * We can't use pci_get_device() here since we are
23 * called from interrupt context.
24 */
25static void pcibios_bus_report_status(struct pci_bus *bus, u_int status_mask, int warn)
26{
27 struct pci_dev *dev;
28
29 list_for_each_entry(dev, &bus->devices, bus_list) {
30 u16 status;
31
32 /*
33 * ignore host bridge - we handle
34 * that separately
35 */
36 if (dev->bus->number == 0 && dev->devfn == 0)
37 continue;
38
39 pci_read_config_word(dev, PCI_STATUS, &status);
40 if (status == 0xffff)
41 continue;
42
43 if ((status & status_mask) == 0)
44 continue;
45
46 /* clear the status errors */
47 pci_write_config_word(dev, PCI_STATUS, status & status_mask);
48
49 if (warn)
50 printk("(%s: %04X) ", pci_name(dev), status);
51 }
52
53 list_for_each_entry(dev, &bus->devices, bus_list)
54 if (dev->subordinate)
55 pcibios_bus_report_status(dev->subordinate, status_mask, warn);
56}
57
58void pcibios_report_status(u_int status_mask, int warn)
59{
60 struct pci_bus *bus;
61
62 list_for_each_entry(bus, &pci_root_buses, node)
63 pcibios_bus_report_status(bus, status_mask, warn);
64}
65
66/*
67 * We don't use this to fix the device, but initialisation of it.
68 * It's not the correct use for this, but it works.
69 * Note that the arbiter/ISA bridge appears to be buggy, specifically in
70 * the following area:
71 * 1. park on CPU
72 * 2. ISA bridge ping-pong
73 * 3. ISA bridge master handling of target RETRY
74 *
75 * Bug 3 is responsible for the sound DMA grinding to a halt. We now
76 * live with bug 2.
77 */
78static void pci_fixup_83c553(struct pci_dev *dev)
79{
80 /*
81 * Set memory region to start at address 0, and enable IO
82 */
83 pci_write_config_dword(dev, PCI_BASE_ADDRESS_0, PCI_BASE_ADDRESS_SPACE_MEMORY);
84 pci_write_config_word(dev, PCI_COMMAND, PCI_COMMAND_IO);
85
86 dev->resource[0].end -= dev->resource[0].start;
87 dev->resource[0].start = 0;
88
89 /*
90 * All memory requests from ISA to be channelled to PCI
91 */
92 pci_write_config_byte(dev, 0x48, 0xff);
93
94 /*
95 * Enable ping-pong on bus master to ISA bridge transactions.
96 * This improves the sound DMA substantially. The fixed
97 * priority arbiter also helps (see below).
98 */
99 pci_write_config_byte(dev, 0x42, 0x01);
100
101 /*
102 * Enable PCI retry
103 */
104 pci_write_config_byte(dev, 0x40, 0x22);
105
106 /*
107 * We used to set the arbiter to "park on last master" (bit
108 * 1 set), but unfortunately the CyberPro does not park the
109 * bus. We must therefore park on CPU. Unfortunately, this
110 * may trigger yet another bug in the 553.
111 */
112 pci_write_config_byte(dev, 0x83, 0x02);
113
114 /*
115 * Make the ISA DMA request lowest priority, and disable
116 * rotating priorities completely.
117 */
118 pci_write_config_byte(dev, 0x80, 0x11);
119 pci_write_config_byte(dev, 0x81, 0x00);
120
121 /*
122 * Route INTA input to IRQ 11, and set IRQ11 to be level
123 * sensitive.
124 */
125 pci_write_config_word(dev, 0x44, 0xb000);
126 outb(0x08, 0x4d1);
127}
128DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_WINBOND, PCI_DEVICE_ID_WINBOND_83C553, pci_fixup_83c553);
129
130static void pci_fixup_unassign(struct pci_dev *dev)
131{
132 dev->resource[0].end -= dev->resource[0].start;
133 dev->resource[0].start = 0;
134}
135DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_WINBOND2, PCI_DEVICE_ID_WINBOND2_89C940F, pci_fixup_unassign);
136
137/*
138 * Prevent the PCI layer from seeing the resources allocated to this device
139 * if it is the host bridge by marking it as such. These resources are of
140 * no consequence to the PCI layer (they are handled elsewhere).
141 */
142static void pci_fixup_dec21285(struct pci_dev *dev)
143{
144 int i;
145
146 if (dev->devfn == 0) {
147 dev->class &= 0xff;
148 dev->class |= PCI_CLASS_BRIDGE_HOST << 8;
149 for (i = 0; i < PCI_NUM_RESOURCES; i++) {
150 dev->resource[i].start = 0;
151 dev->resource[i].end = 0;
152 dev->resource[i].flags = 0;
153 }
154 }
155}
156DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_DEC, PCI_DEVICE_ID_DEC_21285, pci_fixup_dec21285);
157
158/*
159 * PCI IDE controllers use non-standard I/O port decoding, respect it.
160 */
161static void pci_fixup_ide_bases(struct pci_dev *dev)
162{
163 struct resource *r;
164 int i;
165
166 if ((dev->class >> 8) != PCI_CLASS_STORAGE_IDE)
167 return;
168
169 for (i = 0; i < PCI_NUM_RESOURCES; i++) {
170 r = dev->resource + i;
171 if ((r->start & ~0x80) == 0x374) {
172 r->start |= 2;
173 r->end = r->start;
174 }
175 }
176}
177DECLARE_PCI_FIXUP_HEADER(PCI_ANY_ID, PCI_ANY_ID, pci_fixup_ide_bases);
178
179/*
180 * Put the DEC21142 to sleep
181 */
182static void pci_fixup_dec21142(struct pci_dev *dev)
183{
184 pci_write_config_dword(dev, 0x40, 0x80000000);
185}
186DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_DEC, PCI_DEVICE_ID_DEC_21142, pci_fixup_dec21142);
187
188/*
189 * The CY82C693 needs some rather major fixups to ensure that it does
190 * the right thing. Idea from the Alpha people, with a few additions.
191 *
192 * We ensure that the IDE base registers are set to 1f0/3f4 for the
193 * primary bus, and 170/374 for the secondary bus. Also, hide them
194 * from the PCI subsystem view as well so we won't try to perform
195 * our own auto-configuration on them.
196 *
197 * In addition, we ensure that the PCI IDE interrupts are routed to
198 * IRQ 14 and IRQ 15 respectively.
199 *
200 * The above gets us to a point where the IDE on this device is
201 * functional. However, The CY82C693U _does not work_ in bus
202 * master mode without locking the PCI bus solid.
203 */
204static void pci_fixup_cy82c693(struct pci_dev *dev)
205{
206 if ((dev->class >> 8) == PCI_CLASS_STORAGE_IDE) {
207 u32 base0, base1;
208
209 if (dev->class & 0x80) { /* primary */
210 base0 = 0x1f0;
211 base1 = 0x3f4;
212 } else { /* secondary */
213 base0 = 0x170;
214 base1 = 0x374;
215 }
216
217 pci_write_config_dword(dev, PCI_BASE_ADDRESS_0,
218 base0 | PCI_BASE_ADDRESS_SPACE_IO);
219 pci_write_config_dword(dev, PCI_BASE_ADDRESS_1,
220 base1 | PCI_BASE_ADDRESS_SPACE_IO);
221
222 dev->resource[0].start = 0;
223 dev->resource[0].end = 0;
224 dev->resource[0].flags = 0;
225
226 dev->resource[1].start = 0;
227 dev->resource[1].end = 0;
228 dev->resource[1].flags = 0;
229 } else if (PCI_FUNC(dev->devfn) == 0) {
230 /*
231 * Setup IDE IRQ routing.
232 */
233 pci_write_config_byte(dev, 0x4b, 14);
234 pci_write_config_byte(dev, 0x4c, 15);
235
236 /*
237 * Disable FREQACK handshake, enable USB.
238 */
239 pci_write_config_byte(dev, 0x4d, 0x41);
240
241 /*
242 * Enable PCI retry, and PCI post-write buffer.
243 */
244 pci_write_config_byte(dev, 0x44, 0x17);
245
246 /*
247 * Enable ISA master and DMA post write buffering.
248 */
249 pci_write_config_byte(dev, 0x45, 0x03);
250 }
251}
252DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_CONTAQ, PCI_DEVICE_ID_CONTAQ_82C693, pci_fixup_cy82c693);
253
254static void pci_fixup_it8152(struct pci_dev *dev)
255{
256 int i;
257 /* fixup for ITE 8152 devices */
258 /* FIXME: add defines for class 0x68000 and 0x80103 */
259 if ((dev->class >> 8) == PCI_CLASS_BRIDGE_HOST ||
260 dev->class == 0x68000 ||
261 dev->class == 0x80103) {
262 for (i = 0; i < PCI_NUM_RESOURCES; i++) {
263 dev->resource[i].start = 0;
264 dev->resource[i].end = 0;
265 dev->resource[i].flags = 0;
266 }
267 }
268}
269DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_ITE, PCI_DEVICE_ID_ITE_8152, pci_fixup_it8152);
270
271/*
272 * If the bus contains any of these devices, then we must not turn on
273 * parity checking of any kind. Currently this is CyberPro 20x0 only.
274 */
275static inline int pdev_bad_for_parity(struct pci_dev *dev)
276{
277 return ((dev->vendor == PCI_VENDOR_ID_INTERG &&
278 (dev->device == PCI_DEVICE_ID_INTERG_2000 ||
279 dev->device == PCI_DEVICE_ID_INTERG_2010)) ||
280 (dev->vendor == PCI_VENDOR_ID_ITE &&
281 dev->device == PCI_DEVICE_ID_ITE_8152));
282
283}
284
285/*
286 * pcibios_fixup_bus - Called after each bus is probed,
287 * but before its children are examined.
288 */
289void pcibios_fixup_bus(struct pci_bus *bus)
290{
291 struct pci_dev *dev;
292 u16 features = PCI_COMMAND_SERR | PCI_COMMAND_PARITY | PCI_COMMAND_FAST_BACK;
293
294 /*
295 * Walk the devices on this bus, working out what we can
296 * and can't support.
297 */
298 list_for_each_entry(dev, &bus->devices, bus_list) {
299 u16 status;
300
301 pci_read_config_word(dev, PCI_STATUS, &status);
302
303 /*
304 * If any device on this bus does not support fast back
305 * to back transfers, then the bus as a whole is not able
306 * to support them. Having fast back to back transfers
307 * on saves us one PCI cycle per transaction.
308 */
309 if (!(status & PCI_STATUS_FAST_BACK))
310 features &= ~PCI_COMMAND_FAST_BACK;
311
312 if (pdev_bad_for_parity(dev))
313 features &= ~(PCI_COMMAND_SERR | PCI_COMMAND_PARITY);
314
315 switch (dev->class >> 8) {
316 case PCI_CLASS_BRIDGE_PCI:
317 pci_read_config_word(dev, PCI_BRIDGE_CONTROL, &status);
318 status |= PCI_BRIDGE_CTL_PARITY|PCI_BRIDGE_CTL_MASTER_ABORT;
319 status &= ~(PCI_BRIDGE_CTL_BUS_RESET|PCI_BRIDGE_CTL_FAST_BACK);
320 pci_write_config_word(dev, PCI_BRIDGE_CONTROL, status);
321 break;
322
323 case PCI_CLASS_BRIDGE_CARDBUS:
324 pci_read_config_word(dev, PCI_CB_BRIDGE_CONTROL, &status);
325 status |= PCI_CB_BRIDGE_CTL_PARITY|PCI_CB_BRIDGE_CTL_MASTER_ABORT;
326 pci_write_config_word(dev, PCI_CB_BRIDGE_CONTROL, status);
327 break;
328 }
329 }
330
331 /*
332 * Now walk the devices again, this time setting them up.
333 */
334 list_for_each_entry(dev, &bus->devices, bus_list) {
335 u16 cmd;
336
337 pci_read_config_word(dev, PCI_COMMAND, &cmd);
338 cmd |= features;
339 pci_write_config_word(dev, PCI_COMMAND, cmd);
340
341 pci_write_config_byte(dev, PCI_CACHE_LINE_SIZE,
342 L1_CACHE_BYTES >> 2);
343 }
344
345 /*
346 * Propagate the flags to the PCI bridge.
347 */
348 if (bus->self && bus->self->hdr_type == PCI_HEADER_TYPE_BRIDGE) {
349 if (features & PCI_COMMAND_FAST_BACK)
350 bus->bridge_ctl |= PCI_BRIDGE_CTL_FAST_BACK;
351 if (features & PCI_COMMAND_PARITY)
352 bus->bridge_ctl |= PCI_BRIDGE_CTL_PARITY;
353 }
354
355 /*
356 * Report what we did for this bus
357 */
358 printk(KERN_INFO "PCI: bus%d: Fast back to back transfers %sabled\n",
359 bus->number, (features & PCI_COMMAND_FAST_BACK) ? "en" : "dis");
360}
361EXPORT_SYMBOL(pcibios_fixup_bus);
362
363void pcibios_add_bus(struct pci_bus *bus)
364{
365 struct pci_sys_data *sys = bus->sysdata;
366 if (sys->add_bus)
367 sys->add_bus(bus);
368}
369
370void pcibios_remove_bus(struct pci_bus *bus)
371{
372 struct pci_sys_data *sys = bus->sysdata;
373 if (sys->remove_bus)
374 sys->remove_bus(bus);
375}
376
377/*
378 * Swizzle the device pin each time we cross a bridge. If a platform does
379 * not provide a swizzle function, we perform the standard PCI swizzling.
380 *
381 * The default swizzling walks up the bus tree one level at a time, applying
382 * the standard swizzle function at each step, stopping when it finds the PCI
383 * root bus. This will return the slot number of the bridge device on the
384 * root bus and the interrupt pin on that device which should correspond
385 * with the downstream device interrupt.
386 *
387 * Platforms may override this, in which case the slot and pin returned
388 * depend entirely on the platform code. However, please note that the
389 * PCI standard swizzle is implemented on plug-in cards and Cardbus based
390 * PCI extenders, so it can not be ignored.
391 */
392static u8 pcibios_swizzle(struct pci_dev *dev, u8 *pin)
393{
394 struct pci_sys_data *sys = dev->sysdata;
395 int slot, oldpin = *pin;
396
397 if (sys->swizzle)
398 slot = sys->swizzle(dev, pin);
399 else
400 slot = pci_common_swizzle(dev, pin);
401
402 if (debug_pci)
403 printk("PCI: %s swizzling pin %d => pin %d slot %d\n",
404 pci_name(dev), oldpin, *pin, slot);
405
406 return slot;
407}
408
409/*
410 * Map a slot/pin to an IRQ.
411 */
412static int pcibios_map_irq(const struct pci_dev *dev, u8 slot, u8 pin)
413{
414 struct pci_sys_data *sys = dev->sysdata;
415 int irq = -1;
416
417 if (sys->map_irq)
418 irq = sys->map_irq(dev, slot, pin);
419
420 if (debug_pci)
421 printk("PCI: %s mapping slot %d pin %d => irq %d\n",
422 pci_name(dev), slot, pin, irq);
423
424 return irq;
425}
426
427static int pcibios_init_resources(int busnr, struct pci_sys_data *sys)
428{
429 int ret;
430 struct pci_host_bridge_window *window;
431
432 if (list_empty(&sys->resources)) {
433 pci_add_resource_offset(&sys->resources,
434 &iomem_resource, sys->mem_offset);
435 }
436
437 list_for_each_entry(window, &sys->resources, list) {
438 if (resource_type(window->res) == IORESOURCE_IO)
439 return 0;
440 }
441
442 sys->io_res.start = (busnr * SZ_64K) ? : pcibios_min_io;
443 sys->io_res.end = (busnr + 1) * SZ_64K - 1;
444 sys->io_res.flags = IORESOURCE_IO;
445 sys->io_res.name = sys->io_res_name;
446 sprintf(sys->io_res_name, "PCI%d I/O", busnr);
447
448 ret = request_resource(&ioport_resource, &sys->io_res);
449 if (ret) {
450 pr_err("PCI: unable to allocate I/O port region (%d)\n", ret);
451 return ret;
452 }
453 pci_add_resource_offset(&sys->resources, &sys->io_res,
454 sys->io_offset);
455
456 return 0;
457}
458
459static void pcibios_init_hw(struct device *parent, struct hw_pci *hw,
460 struct list_head *head)
461{
462 struct pci_sys_data *sys = NULL;
463 int ret;
464 int nr, busnr;
465
466 for (nr = busnr = 0; nr < hw->nr_controllers; nr++) {
467 sys = kzalloc(sizeof(struct pci_sys_data), GFP_KERNEL);
468 if (!sys)
469 panic("PCI: unable to allocate sys data!");
470
471#ifdef CONFIG_PCI_DOMAINS
472 sys->domain = hw->domain;
473#endif
474 sys->busnr = busnr;
475 sys->swizzle = hw->swizzle;
476 sys->map_irq = hw->map_irq;
477 sys->align_resource = hw->align_resource;
478 sys->add_bus = hw->add_bus;
479 sys->remove_bus = hw->remove_bus;
480 INIT_LIST_HEAD(&sys->resources);
481
482 if (hw->private_data)
483 sys->private_data = hw->private_data[nr];
484
485 ret = hw->setup(nr, sys);
486
487 if (ret > 0) {
488 ret = pcibios_init_resources(nr, sys);
489 if (ret) {
490 kfree(sys);
491 break;
492 }
493
494 if (hw->scan)
495 sys->bus = hw->scan(nr, sys);
496 else
497 sys->bus = pci_scan_root_bus(parent, sys->busnr,
498 hw->ops, sys, &sys->resources);
499
500 if (!sys->bus)
501 panic("PCI: unable to scan bus!");
502
503 busnr = sys->bus->busn_res.end + 1;
504
505 list_add(&sys->node, head);
506 } else {
507 kfree(sys);
508 if (ret < 0)
509 break;
510 }
511 }
512}
513
514void pci_common_init_dev(struct device *parent, struct hw_pci *hw)
515{
516 struct pci_sys_data *sys;
517 LIST_HEAD(head);
518
519 pci_add_flags(PCI_REASSIGN_ALL_RSRC);
520 if (hw->preinit)
521 hw->preinit();
522 pcibios_init_hw(parent, hw, &head);
523 if (hw->postinit)
524 hw->postinit();
525
526 pci_fixup_irqs(pcibios_swizzle, pcibios_map_irq);
527
528 list_for_each_entry(sys, &head, node) {
529 struct pci_bus *bus = sys->bus;
530
531 if (!pci_has_flag(PCI_PROBE_ONLY)) {
532 /*
533 * Size the bridge windows.
534 */
535 pci_bus_size_bridges(bus);
536
537 /*
538 * Assign resources.
539 */
540 pci_bus_assign_resources(bus);
541 }
542
543 /*
544 * Tell drivers about devices found.
545 */
546 pci_bus_add_devices(bus);
547 }
548}
549
550#ifndef CONFIG_PCI_HOST_ITE8152
551void pcibios_set_master(struct pci_dev *dev)
552{
553 /* No special bus mastering setup handling */
554}
555#endif
556
557char * __init pcibios_setup(char *str)
558{
559 if (!strcmp(str, "debug")) {
560 debug_pci = 1;
561 return NULL;
562 } else if (!strcmp(str, "firmware")) {
563 pci_add_flags(PCI_PROBE_ONLY);
564 return NULL;
565 }
566 return str;
567}
568
569/*
570 * From arch/i386/kernel/pci-i386.c:
571 *
572 * We need to avoid collisions with `mirrored' VGA ports
573 * and other strange ISA hardware, so we always want the
574 * addresses to be allocated in the 0x000-0x0ff region
575 * modulo 0x400.
576 *
577 * Why? Because some silly external IO cards only decode
578 * the low 10 bits of the IO address. The 0x00-0xff region
579 * is reserved for motherboard devices that decode all 16
580 * bits, so it's ok to allocate at, say, 0x2800-0x28ff,
581 * but we want to try to avoid allocating at 0x2900-0x2bff
582 * which might be mirrored at 0x0100-0x03ff..
583 */
584resource_size_t pcibios_align_resource(void *data, const struct resource *res,
585 resource_size_t size, resource_size_t align)
586{
587 struct pci_dev *dev = data;
588 struct pci_sys_data *sys = dev->sysdata;
589 resource_size_t start = res->start;
590
591 if (res->flags & IORESOURCE_IO && start & 0x300)
592 start = (start + 0x3ff) & ~0x3ff;
593
594 start = (start + align - 1) & ~(align - 1);
595
596 if (sys->align_resource)
597 return sys->align_resource(dev, res, start, size, align);
598
599 return start;
600}
601
602/**
603 * pcibios_enable_device - Enable I/O and memory.
604 * @dev: PCI device to be enabled
605 */
606int pcibios_enable_device(struct pci_dev *dev, int mask)
607{
608 if (pci_has_flag(PCI_PROBE_ONLY))
609 return 0;
610
611 return pci_enable_resources(dev, mask);
612}
613
614int pci_mmap_page_range(struct pci_dev *dev, struct vm_area_struct *vma,
615 enum pci_mmap_state mmap_state, int write_combine)
616{
617 struct pci_sys_data *root = dev->sysdata;
618 unsigned long phys;
619
620 if (mmap_state == pci_mmap_io) {
621 return -EINVAL;
622 } else {
623 phys = vma->vm_pgoff + (root->mem_offset >> PAGE_SHIFT);
624 }
625
626 /*
627 * Mark this as IO
628 */
629 vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot);
630
631 if (remap_pfn_range(vma, vma->vm_start, phys,
632 vma->vm_end - vma->vm_start,
633 vma->vm_page_prot))
634 return -EAGAIN;
635
636 return 0;
637}
638
639void __init pci_map_io_early(unsigned long pfn)
640{
641 struct map_desc pci_io_desc = {
642 .virtual = PCI_IO_VIRT_BASE,
643 .type = MT_DEVICE,
644 .length = SZ_64K,
645 };
646
647 pci_io_desc.pfn = pfn;
648 iotable_init(&pci_io_desc, 1);
649}
1// SPDX-License-Identifier: GPL-2.0
2/*
3 * linux/arch/arm/kernel/bios32.c
4 *
5 * PCI bios-type initialisation for PCI machines
6 *
7 * Bits taken from various places.
8 */
9#include <linux/export.h>
10#include <linux/kernel.h>
11#include <linux/pci.h>
12#include <linux/slab.h>
13#include <linux/init.h>
14#include <linux/io.h>
15
16#include <asm/mach-types.h>
17#include <asm/mach/map.h>
18#include <asm/mach/pci.h>
19
20static int debug_pci;
21
22/*
23 * We can't use pci_get_device() here since we are
24 * called from interrupt context.
25 */
26static void pcibios_bus_report_status(struct pci_bus *bus, u_int status_mask, int warn)
27{
28 struct pci_dev *dev;
29
30 list_for_each_entry(dev, &bus->devices, bus_list) {
31 u16 status;
32
33 /*
34 * ignore host bridge - we handle
35 * that separately
36 */
37 if (dev->bus->number == 0 && dev->devfn == 0)
38 continue;
39
40 pci_read_config_word(dev, PCI_STATUS, &status);
41 if (status == 0xffff)
42 continue;
43
44 if ((status & status_mask) == 0)
45 continue;
46
47 /* clear the status errors */
48 pci_write_config_word(dev, PCI_STATUS, status & status_mask);
49
50 if (warn)
51 printk("(%s: %04X) ", pci_name(dev), status);
52 }
53
54 list_for_each_entry(dev, &bus->devices, bus_list)
55 if (dev->subordinate)
56 pcibios_bus_report_status(dev->subordinate, status_mask, warn);
57}
58
59void pcibios_report_status(u_int status_mask, int warn)
60{
61 struct pci_bus *bus;
62
63 list_for_each_entry(bus, &pci_root_buses, node)
64 pcibios_bus_report_status(bus, status_mask, warn);
65}
66
67/*
68 * We don't use this to fix the device, but initialisation of it.
69 * It's not the correct use for this, but it works.
70 * Note that the arbiter/ISA bridge appears to be buggy, specifically in
71 * the following area:
72 * 1. park on CPU
73 * 2. ISA bridge ping-pong
74 * 3. ISA bridge master handling of target RETRY
75 *
76 * Bug 3 is responsible for the sound DMA grinding to a halt. We now
77 * live with bug 2.
78 */
79static void pci_fixup_83c553(struct pci_dev *dev)
80{
81 /*
82 * Set memory region to start at address 0, and enable IO
83 */
84 pci_write_config_dword(dev, PCI_BASE_ADDRESS_0, PCI_BASE_ADDRESS_SPACE_MEMORY);
85 pci_write_config_word(dev, PCI_COMMAND, PCI_COMMAND_IO);
86
87 dev->resource[0].end -= dev->resource[0].start;
88 dev->resource[0].start = 0;
89
90 /*
91 * All memory requests from ISA to be channelled to PCI
92 */
93 pci_write_config_byte(dev, 0x48, 0xff);
94
95 /*
96 * Enable ping-pong on bus master to ISA bridge transactions.
97 * This improves the sound DMA substantially. The fixed
98 * priority arbiter also helps (see below).
99 */
100 pci_write_config_byte(dev, 0x42, 0x01);
101
102 /*
103 * Enable PCI retry
104 */
105 pci_write_config_byte(dev, 0x40, 0x22);
106
107 /*
108 * We used to set the arbiter to "park on last master" (bit
109 * 1 set), but unfortunately the CyberPro does not park the
110 * bus. We must therefore park on CPU. Unfortunately, this
111 * may trigger yet another bug in the 553.
112 */
113 pci_write_config_byte(dev, 0x83, 0x02);
114
115 /*
116 * Make the ISA DMA request lowest priority, and disable
117 * rotating priorities completely.
118 */
119 pci_write_config_byte(dev, 0x80, 0x11);
120 pci_write_config_byte(dev, 0x81, 0x00);
121
122 /*
123 * Route INTA input to IRQ 11, and set IRQ11 to be level
124 * sensitive.
125 */
126 pci_write_config_word(dev, 0x44, 0xb000);
127 outb(0x08, 0x4d1);
128}
129DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_WINBOND, PCI_DEVICE_ID_WINBOND_83C553, pci_fixup_83c553);
130
131static void pci_fixup_unassign(struct pci_dev *dev)
132{
133 dev->resource[0].end -= dev->resource[0].start;
134 dev->resource[0].start = 0;
135}
136DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_WINBOND2, PCI_DEVICE_ID_WINBOND2_89C940F, pci_fixup_unassign);
137
138/*
139 * Prevent the PCI layer from seeing the resources allocated to this device
140 * if it is the host bridge by marking it as such. These resources are of
141 * no consequence to the PCI layer (they are handled elsewhere).
142 */
143static void pci_fixup_dec21285(struct pci_dev *dev)
144{
145 int i;
146
147 if (dev->devfn == 0) {
148 dev->class &= 0xff;
149 dev->class |= PCI_CLASS_BRIDGE_HOST << 8;
150 for (i = 0; i < PCI_NUM_RESOURCES; i++) {
151 dev->resource[i].start = 0;
152 dev->resource[i].end = 0;
153 dev->resource[i].flags = 0;
154 }
155 }
156}
157DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_DEC, PCI_DEVICE_ID_DEC_21285, pci_fixup_dec21285);
158
159/*
160 * PCI IDE controllers use non-standard I/O port decoding, respect it.
161 */
162static void pci_fixup_ide_bases(struct pci_dev *dev)
163{
164 struct resource *r;
165 int i;
166
167 if ((dev->class >> 8) != PCI_CLASS_STORAGE_IDE)
168 return;
169
170 for (i = 0; i < PCI_NUM_RESOURCES; i++) {
171 r = dev->resource + i;
172 if ((r->start & ~0x80) == 0x374) {
173 r->start |= 2;
174 r->end = r->start;
175 }
176 }
177}
178DECLARE_PCI_FIXUP_HEADER(PCI_ANY_ID, PCI_ANY_ID, pci_fixup_ide_bases);
179
180/*
181 * Put the DEC21142 to sleep
182 */
183static void pci_fixup_dec21142(struct pci_dev *dev)
184{
185 pci_write_config_dword(dev, 0x40, 0x80000000);
186}
187DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_DEC, PCI_DEVICE_ID_DEC_21142, pci_fixup_dec21142);
188
189/*
190 * The CY82C693 needs some rather major fixups to ensure that it does
191 * the right thing. Idea from the Alpha people, with a few additions.
192 *
193 * We ensure that the IDE base registers are set to 1f0/3f4 for the
194 * primary bus, and 170/374 for the secondary bus. Also, hide them
195 * from the PCI subsystem view as well so we won't try to perform
196 * our own auto-configuration on them.
197 *
198 * In addition, we ensure that the PCI IDE interrupts are routed to
199 * IRQ 14 and IRQ 15 respectively.
200 *
201 * The above gets us to a point where the IDE on this device is
202 * functional. However, The CY82C693U _does not work_ in bus
203 * master mode without locking the PCI bus solid.
204 */
205static void pci_fixup_cy82c693(struct pci_dev *dev)
206{
207 if ((dev->class >> 8) == PCI_CLASS_STORAGE_IDE) {
208 u32 base0, base1;
209
210 if (dev->class & 0x80) { /* primary */
211 base0 = 0x1f0;
212 base1 = 0x3f4;
213 } else { /* secondary */
214 base0 = 0x170;
215 base1 = 0x374;
216 }
217
218 pci_write_config_dword(dev, PCI_BASE_ADDRESS_0,
219 base0 | PCI_BASE_ADDRESS_SPACE_IO);
220 pci_write_config_dword(dev, PCI_BASE_ADDRESS_1,
221 base1 | PCI_BASE_ADDRESS_SPACE_IO);
222
223 dev->resource[0].start = 0;
224 dev->resource[0].end = 0;
225 dev->resource[0].flags = 0;
226
227 dev->resource[1].start = 0;
228 dev->resource[1].end = 0;
229 dev->resource[1].flags = 0;
230 } else if (PCI_FUNC(dev->devfn) == 0) {
231 /*
232 * Setup IDE IRQ routing.
233 */
234 pci_write_config_byte(dev, 0x4b, 14);
235 pci_write_config_byte(dev, 0x4c, 15);
236
237 /*
238 * Disable FREQACK handshake, enable USB.
239 */
240 pci_write_config_byte(dev, 0x4d, 0x41);
241
242 /*
243 * Enable PCI retry, and PCI post-write buffer.
244 */
245 pci_write_config_byte(dev, 0x44, 0x17);
246
247 /*
248 * Enable ISA master and DMA post write buffering.
249 */
250 pci_write_config_byte(dev, 0x45, 0x03);
251 }
252}
253DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_CONTAQ, PCI_DEVICE_ID_CONTAQ_82C693, pci_fixup_cy82c693);
254
255static void pci_fixup_it8152(struct pci_dev *dev)
256{
257 int i;
258 /* fixup for ITE 8152 devices */
259 /* FIXME: add defines for class 0x68000 and 0x80103 */
260 if ((dev->class >> 8) == PCI_CLASS_BRIDGE_HOST ||
261 dev->class == 0x68000 ||
262 dev->class == 0x80103) {
263 for (i = 0; i < PCI_NUM_RESOURCES; i++) {
264 dev->resource[i].start = 0;
265 dev->resource[i].end = 0;
266 dev->resource[i].flags = 0;
267 }
268 }
269}
270DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_ITE, PCI_DEVICE_ID_ITE_8152, pci_fixup_it8152);
271
272/*
273 * If the bus contains any of these devices, then we must not turn on
274 * parity checking of any kind. Currently this is CyberPro 20x0 only.
275 */
276static inline int pdev_bad_for_parity(struct pci_dev *dev)
277{
278 return ((dev->vendor == PCI_VENDOR_ID_INTERG &&
279 (dev->device == PCI_DEVICE_ID_INTERG_2000 ||
280 dev->device == PCI_DEVICE_ID_INTERG_2010)) ||
281 (dev->vendor == PCI_VENDOR_ID_ITE &&
282 dev->device == PCI_DEVICE_ID_ITE_8152));
283
284}
285
286/*
287 * pcibios_fixup_bus - Called after each bus is probed,
288 * but before its children are examined.
289 */
290void pcibios_fixup_bus(struct pci_bus *bus)
291{
292 struct pci_dev *dev;
293 u16 features = PCI_COMMAND_SERR | PCI_COMMAND_PARITY | PCI_COMMAND_FAST_BACK;
294
295 /*
296 * Walk the devices on this bus, working out what we can
297 * and can't support.
298 */
299 list_for_each_entry(dev, &bus->devices, bus_list) {
300 u16 status;
301
302 pci_read_config_word(dev, PCI_STATUS, &status);
303
304 /*
305 * If any device on this bus does not support fast back
306 * to back transfers, then the bus as a whole is not able
307 * to support them. Having fast back to back transfers
308 * on saves us one PCI cycle per transaction.
309 */
310 if (!(status & PCI_STATUS_FAST_BACK))
311 features &= ~PCI_COMMAND_FAST_BACK;
312
313 if (pdev_bad_for_parity(dev))
314 features &= ~(PCI_COMMAND_SERR | PCI_COMMAND_PARITY);
315
316 switch (dev->class >> 8) {
317 case PCI_CLASS_BRIDGE_PCI:
318 pci_read_config_word(dev, PCI_BRIDGE_CONTROL, &status);
319 status |= PCI_BRIDGE_CTL_PARITY|PCI_BRIDGE_CTL_MASTER_ABORT;
320 status &= ~(PCI_BRIDGE_CTL_BUS_RESET|PCI_BRIDGE_CTL_FAST_BACK);
321 pci_write_config_word(dev, PCI_BRIDGE_CONTROL, status);
322 break;
323
324 case PCI_CLASS_BRIDGE_CARDBUS:
325 pci_read_config_word(dev, PCI_CB_BRIDGE_CONTROL, &status);
326 status |= PCI_CB_BRIDGE_CTL_PARITY|PCI_CB_BRIDGE_CTL_MASTER_ABORT;
327 pci_write_config_word(dev, PCI_CB_BRIDGE_CONTROL, status);
328 break;
329 }
330 }
331
332 /*
333 * Now walk the devices again, this time setting them up.
334 */
335 list_for_each_entry(dev, &bus->devices, bus_list) {
336 u16 cmd;
337
338 pci_read_config_word(dev, PCI_COMMAND, &cmd);
339 cmd |= features;
340 pci_write_config_word(dev, PCI_COMMAND, cmd);
341
342 pci_write_config_byte(dev, PCI_CACHE_LINE_SIZE,
343 L1_CACHE_BYTES >> 2);
344 }
345
346 /*
347 * Propagate the flags to the PCI bridge.
348 */
349 if (bus->self && bus->self->hdr_type == PCI_HEADER_TYPE_BRIDGE) {
350 if (features & PCI_COMMAND_FAST_BACK)
351 bus->bridge_ctl |= PCI_BRIDGE_CTL_FAST_BACK;
352 if (features & PCI_COMMAND_PARITY)
353 bus->bridge_ctl |= PCI_BRIDGE_CTL_PARITY;
354 }
355
356 /*
357 * Report what we did for this bus
358 */
359 pr_info("PCI: bus%d: Fast back to back transfers %sabled\n",
360 bus->number, (features & PCI_COMMAND_FAST_BACK) ? "en" : "dis");
361}
362EXPORT_SYMBOL(pcibios_fixup_bus);
363
364/*
365 * Swizzle the device pin each time we cross a bridge. If a platform does
366 * not provide a swizzle function, we perform the standard PCI swizzling.
367 *
368 * The default swizzling walks up the bus tree one level at a time, applying
369 * the standard swizzle function at each step, stopping when it finds the PCI
370 * root bus. This will return the slot number of the bridge device on the
371 * root bus and the interrupt pin on that device which should correspond
372 * with the downstream device interrupt.
373 *
374 * Platforms may override this, in which case the slot and pin returned
375 * depend entirely on the platform code. However, please note that the
376 * PCI standard swizzle is implemented on plug-in cards and Cardbus based
377 * PCI extenders, so it can not be ignored.
378 */
379static u8 pcibios_swizzle(struct pci_dev *dev, u8 *pin)
380{
381 struct pci_sys_data *sys = dev->sysdata;
382 int slot, oldpin = *pin;
383
384 if (sys->swizzle)
385 slot = sys->swizzle(dev, pin);
386 else
387 slot = pci_common_swizzle(dev, pin);
388
389 if (debug_pci)
390 printk("PCI: %s swizzling pin %d => pin %d slot %d\n",
391 pci_name(dev), oldpin, *pin, slot);
392
393 return slot;
394}
395
396/*
397 * Map a slot/pin to an IRQ.
398 */
399static int pcibios_map_irq(const struct pci_dev *dev, u8 slot, u8 pin)
400{
401 struct pci_sys_data *sys = dev->sysdata;
402 int irq = -1;
403
404 if (sys->map_irq)
405 irq = sys->map_irq(dev, slot, pin);
406
407 if (debug_pci)
408 printk("PCI: %s mapping slot %d pin %d => irq %d\n",
409 pci_name(dev), slot, pin, irq);
410
411 return irq;
412}
413
414static int pcibios_init_resource(int busnr, struct pci_sys_data *sys,
415 int io_optional)
416{
417 int ret;
418 struct resource_entry *window;
419
420 if (list_empty(&sys->resources)) {
421 pci_add_resource_offset(&sys->resources,
422 &iomem_resource, sys->mem_offset);
423 }
424
425 /*
426 * If a platform says I/O port support is optional, we don't add
427 * the default I/O space. The platform is responsible for adding
428 * any I/O space it needs.
429 */
430 if (io_optional)
431 return 0;
432
433 resource_list_for_each_entry(window, &sys->resources)
434 if (resource_type(window->res) == IORESOURCE_IO)
435 return 0;
436
437 sys->io_res.start = (busnr * SZ_64K) ? : pcibios_min_io;
438 sys->io_res.end = (busnr + 1) * SZ_64K - 1;
439 sys->io_res.flags = IORESOURCE_IO;
440 sys->io_res.name = sys->io_res_name;
441 sprintf(sys->io_res_name, "PCI%d I/O", busnr);
442
443 ret = request_resource(&ioport_resource, &sys->io_res);
444 if (ret) {
445 pr_err("PCI: unable to allocate I/O port region (%d)\n", ret);
446 return ret;
447 }
448 pci_add_resource_offset(&sys->resources, &sys->io_res,
449 sys->io_offset);
450
451 return 0;
452}
453
454static void pcibios_init_hw(struct device *parent, struct hw_pci *hw,
455 struct list_head *head)
456{
457 struct pci_sys_data *sys = NULL;
458 int ret;
459 int nr, busnr;
460
461 for (nr = busnr = 0; nr < hw->nr_controllers; nr++) {
462 struct pci_host_bridge *bridge;
463
464 bridge = pci_alloc_host_bridge(sizeof(struct pci_sys_data));
465 if (WARN(!bridge, "PCI: unable to allocate bridge!"))
466 break;
467
468 sys = pci_host_bridge_priv(bridge);
469
470 sys->busnr = busnr;
471 sys->swizzle = hw->swizzle;
472 sys->map_irq = hw->map_irq;
473 INIT_LIST_HEAD(&sys->resources);
474
475 if (hw->private_data)
476 sys->private_data = hw->private_data[nr];
477
478 ret = hw->setup(nr, sys);
479
480 if (ret > 0) {
481
482 ret = pcibios_init_resource(nr, sys, hw->io_optional);
483 if (ret) {
484 pci_free_host_bridge(bridge);
485 break;
486 }
487
488 bridge->map_irq = pcibios_map_irq;
489 bridge->swizzle_irq = pcibios_swizzle;
490
491 if (hw->scan)
492 ret = hw->scan(nr, bridge);
493 else {
494 list_splice_init(&sys->resources,
495 &bridge->windows);
496 bridge->dev.parent = parent;
497 bridge->sysdata = sys;
498 bridge->busnr = sys->busnr;
499 bridge->ops = hw->ops;
500 bridge->msi = hw->msi_ctrl;
501 bridge->align_resource =
502 hw->align_resource;
503
504 ret = pci_scan_root_bus_bridge(bridge);
505 }
506
507 if (WARN(ret < 0, "PCI: unable to scan bus!")) {
508 pci_free_host_bridge(bridge);
509 break;
510 }
511
512 sys->bus = bridge->bus;
513
514 busnr = sys->bus->busn_res.end + 1;
515
516 list_add(&sys->node, head);
517 } else {
518 pci_free_host_bridge(bridge);
519 if (ret < 0)
520 break;
521 }
522 }
523}
524
525void pci_common_init_dev(struct device *parent, struct hw_pci *hw)
526{
527 struct pci_sys_data *sys;
528 LIST_HEAD(head);
529
530 pci_add_flags(PCI_REASSIGN_ALL_BUS);
531 if (hw->preinit)
532 hw->preinit();
533 pcibios_init_hw(parent, hw, &head);
534 if (hw->postinit)
535 hw->postinit();
536
537 list_for_each_entry(sys, &head, node) {
538 struct pci_bus *bus = sys->bus;
539
540 /*
541 * We insert PCI resources into the iomem_resource and
542 * ioport_resource trees in either pci_bus_claim_resources()
543 * or pci_bus_assign_resources().
544 */
545 if (pci_has_flag(PCI_PROBE_ONLY)) {
546 pci_bus_claim_resources(bus);
547 } else {
548 struct pci_bus *child;
549
550 pci_bus_size_bridges(bus);
551 pci_bus_assign_resources(bus);
552
553 list_for_each_entry(child, &bus->children, node)
554 pcie_bus_configure_settings(child);
555 }
556
557 pci_bus_add_devices(bus);
558 }
559}
560
561#ifndef CONFIG_PCI_HOST_ITE8152
562void pcibios_set_master(struct pci_dev *dev)
563{
564 /* No special bus mastering setup handling */
565}
566#endif
567
568char * __init pcibios_setup(char *str)
569{
570 if (!strcmp(str, "debug")) {
571 debug_pci = 1;
572 return NULL;
573 }
574 return str;
575}
576
577/*
578 * From arch/i386/kernel/pci-i386.c:
579 *
580 * We need to avoid collisions with `mirrored' VGA ports
581 * and other strange ISA hardware, so we always want the
582 * addresses to be allocated in the 0x000-0x0ff region
583 * modulo 0x400.
584 *
585 * Why? Because some silly external IO cards only decode
586 * the low 10 bits of the IO address. The 0x00-0xff region
587 * is reserved for motherboard devices that decode all 16
588 * bits, so it's ok to allocate at, say, 0x2800-0x28ff,
589 * but we want to try to avoid allocating at 0x2900-0x2bff
590 * which might be mirrored at 0x0100-0x03ff..
591 */
592resource_size_t pcibios_align_resource(void *data, const struct resource *res,
593 resource_size_t size, resource_size_t align)
594{
595 struct pci_dev *dev = data;
596 resource_size_t start = res->start;
597 struct pci_host_bridge *host_bridge;
598
599 if (res->flags & IORESOURCE_IO && start & 0x300)
600 start = (start + 0x3ff) & ~0x3ff;
601
602 start = (start + align - 1) & ~(align - 1);
603
604 host_bridge = pci_find_host_bridge(dev->bus);
605
606 if (host_bridge->align_resource)
607 return host_bridge->align_resource(dev, res,
608 start, size, align);
609
610 return start;
611}
612
613void __init pci_map_io_early(unsigned long pfn)
614{
615 struct map_desc pci_io_desc = {
616 .virtual = PCI_IO_VIRT_BASE,
617 .type = MT_DEVICE,
618 .length = SZ_64K,
619 };
620
621 pci_io_desc.pfn = pfn;
622 iotable_init(&pci_io_desc, 1);
623}