Loading...
1// SPDX-License-Identifier: GPL-2.0
2/*
3 * PCIe driver for Marvell Armada 370 and Armada XP SoCs
4 *
5 * Author: Thomas Petazzoni <thomas.petazzoni@free-electrons.com>
6 */
7
8#include <linux/kernel.h>
9#include <linux/module.h>
10#include <linux/pci.h>
11#include <linux/bitfield.h>
12#include <linux/clk.h>
13#include <linux/delay.h>
14#include <linux/gpio/consumer.h>
15#include <linux/init.h>
16#include <linux/irqchip/chained_irq.h>
17#include <linux/irqdomain.h>
18#include <linux/mbus.h>
19#include <linux/slab.h>
20#include <linux/platform_device.h>
21#include <linux/of_address.h>
22#include <linux/of_irq.h>
23#include <linux/of_pci.h>
24#include <linux/of_platform.h>
25
26#include "../pci.h"
27#include "../pci-bridge-emul.h"
28
29/*
30 * PCIe unit register offsets.
31 */
32#define PCIE_DEV_ID_OFF 0x0000
33#define PCIE_CMD_OFF 0x0004
34#define PCIE_DEV_REV_OFF 0x0008
35#define PCIE_BAR_LO_OFF(n) (0x0010 + ((n) << 3))
36#define PCIE_BAR_HI_OFF(n) (0x0014 + ((n) << 3))
37#define PCIE_SSDEV_ID_OFF 0x002c
38#define PCIE_CAP_PCIEXP 0x0060
39#define PCIE_CAP_PCIERR_OFF 0x0100
40#define PCIE_BAR_CTRL_OFF(n) (0x1804 + (((n) - 1) * 4))
41#define PCIE_WIN04_CTRL_OFF(n) (0x1820 + ((n) << 4))
42#define PCIE_WIN04_BASE_OFF(n) (0x1824 + ((n) << 4))
43#define PCIE_WIN04_REMAP_OFF(n) (0x182c + ((n) << 4))
44#define PCIE_WIN5_CTRL_OFF 0x1880
45#define PCIE_WIN5_BASE_OFF 0x1884
46#define PCIE_WIN5_REMAP_OFF 0x188c
47#define PCIE_CONF_ADDR_OFF 0x18f8
48#define PCIE_CONF_ADDR_EN 0x80000000
49#define PCIE_CONF_REG(r) ((((r) & 0xf00) << 16) | ((r) & 0xfc))
50#define PCIE_CONF_BUS(b) (((b) & 0xff) << 16)
51#define PCIE_CONF_DEV(d) (((d) & 0x1f) << 11)
52#define PCIE_CONF_FUNC(f) (((f) & 0x7) << 8)
53#define PCIE_CONF_ADDR(bus, devfn, where) \
54 (PCIE_CONF_BUS(bus) | PCIE_CONF_DEV(PCI_SLOT(devfn)) | \
55 PCIE_CONF_FUNC(PCI_FUNC(devfn)) | PCIE_CONF_REG(where) | \
56 PCIE_CONF_ADDR_EN)
57#define PCIE_CONF_DATA_OFF 0x18fc
58#define PCIE_INT_CAUSE_OFF 0x1900
59#define PCIE_INT_UNMASK_OFF 0x1910
60#define PCIE_INT_INTX(i) BIT(24+i)
61#define PCIE_INT_PM_PME BIT(28)
62#define PCIE_INT_ALL_MASK GENMASK(31, 0)
63#define PCIE_CTRL_OFF 0x1a00
64#define PCIE_CTRL_X1_MODE 0x0001
65#define PCIE_CTRL_RC_MODE BIT(1)
66#define PCIE_CTRL_MASTER_HOT_RESET BIT(24)
67#define PCIE_STAT_OFF 0x1a04
68#define PCIE_STAT_BUS 0xff00
69#define PCIE_STAT_DEV 0x1f0000
70#define PCIE_STAT_LINK_DOWN BIT(0)
71#define PCIE_SSPL_OFF 0x1a0c
72#define PCIE_SSPL_VALUE_SHIFT 0
73#define PCIE_SSPL_VALUE_MASK GENMASK(7, 0)
74#define PCIE_SSPL_SCALE_SHIFT 8
75#define PCIE_SSPL_SCALE_MASK GENMASK(9, 8)
76#define PCIE_SSPL_ENABLE BIT(16)
77#define PCIE_RC_RTSTA 0x1a14
78#define PCIE_DEBUG_CTRL 0x1a60
79#define PCIE_DEBUG_SOFT_RESET BIT(20)
80
81struct mvebu_pcie_port;
82
83/* Structure representing all PCIe interfaces */
84struct mvebu_pcie {
85 struct platform_device *pdev;
86 struct mvebu_pcie_port *ports;
87 struct resource io;
88 struct resource realio;
89 struct resource mem;
90 int nports;
91};
92
93struct mvebu_pcie_window {
94 phys_addr_t base;
95 phys_addr_t remap;
96 size_t size;
97};
98
99/* Structure representing one PCIe interface */
100struct mvebu_pcie_port {
101 char *name;
102 void __iomem *base;
103 u32 port;
104 u32 lane;
105 bool is_x4;
106 int devfn;
107 unsigned int mem_target;
108 unsigned int mem_attr;
109 unsigned int io_target;
110 unsigned int io_attr;
111 struct clk *clk;
112 struct gpio_desc *reset_gpio;
113 char *reset_name;
114 struct pci_bridge_emul bridge;
115 struct device_node *dn;
116 struct mvebu_pcie *pcie;
117 struct mvebu_pcie_window memwin;
118 struct mvebu_pcie_window iowin;
119 u32 saved_pcie_stat;
120 struct resource regs;
121 u8 slot_power_limit_value;
122 u8 slot_power_limit_scale;
123 struct irq_domain *intx_irq_domain;
124 raw_spinlock_t irq_lock;
125 int intx_irq;
126};
127
128static inline void mvebu_writel(struct mvebu_pcie_port *port, u32 val, u32 reg)
129{
130 writel(val, port->base + reg);
131}
132
133static inline u32 mvebu_readl(struct mvebu_pcie_port *port, u32 reg)
134{
135 return readl(port->base + reg);
136}
137
138static inline bool mvebu_has_ioport(struct mvebu_pcie_port *port)
139{
140 return port->io_target != -1 && port->io_attr != -1;
141}
142
143static bool mvebu_pcie_link_up(struct mvebu_pcie_port *port)
144{
145 return !(mvebu_readl(port, PCIE_STAT_OFF) & PCIE_STAT_LINK_DOWN);
146}
147
148static u8 mvebu_pcie_get_local_bus_nr(struct mvebu_pcie_port *port)
149{
150 return (mvebu_readl(port, PCIE_STAT_OFF) & PCIE_STAT_BUS) >> 8;
151}
152
153static void mvebu_pcie_set_local_bus_nr(struct mvebu_pcie_port *port, int nr)
154{
155 u32 stat;
156
157 stat = mvebu_readl(port, PCIE_STAT_OFF);
158 stat &= ~PCIE_STAT_BUS;
159 stat |= nr << 8;
160 mvebu_writel(port, stat, PCIE_STAT_OFF);
161}
162
163static void mvebu_pcie_set_local_dev_nr(struct mvebu_pcie_port *port, int nr)
164{
165 u32 stat;
166
167 stat = mvebu_readl(port, PCIE_STAT_OFF);
168 stat &= ~PCIE_STAT_DEV;
169 stat |= nr << 16;
170 mvebu_writel(port, stat, PCIE_STAT_OFF);
171}
172
173static void mvebu_pcie_disable_wins(struct mvebu_pcie_port *port)
174{
175 int i;
176
177 mvebu_writel(port, 0, PCIE_BAR_LO_OFF(0));
178 mvebu_writel(port, 0, PCIE_BAR_HI_OFF(0));
179
180 for (i = 1; i < 3; i++) {
181 mvebu_writel(port, 0, PCIE_BAR_CTRL_OFF(i));
182 mvebu_writel(port, 0, PCIE_BAR_LO_OFF(i));
183 mvebu_writel(port, 0, PCIE_BAR_HI_OFF(i));
184 }
185
186 for (i = 0; i < 5; i++) {
187 mvebu_writel(port, 0, PCIE_WIN04_CTRL_OFF(i));
188 mvebu_writel(port, 0, PCIE_WIN04_BASE_OFF(i));
189 mvebu_writel(port, 0, PCIE_WIN04_REMAP_OFF(i));
190 }
191
192 mvebu_writel(port, 0, PCIE_WIN5_CTRL_OFF);
193 mvebu_writel(port, 0, PCIE_WIN5_BASE_OFF);
194 mvebu_writel(port, 0, PCIE_WIN5_REMAP_OFF);
195}
196
197/*
198 * Setup PCIE BARs and Address Decode Wins:
199 * BAR[0] -> internal registers (needed for MSI)
200 * BAR[1] -> covers all DRAM banks
201 * BAR[2] -> Disabled
202 * WIN[0-3] -> DRAM bank[0-3]
203 */
204static void mvebu_pcie_setup_wins(struct mvebu_pcie_port *port)
205{
206 const struct mbus_dram_target_info *dram;
207 u32 size;
208 int i;
209
210 dram = mv_mbus_dram_info();
211
212 /* First, disable and clear BARs and windows. */
213 mvebu_pcie_disable_wins(port);
214
215 /* Setup windows for DDR banks. Count total DDR size on the fly. */
216 size = 0;
217 for (i = 0; i < dram->num_cs; i++) {
218 const struct mbus_dram_window *cs = dram->cs + i;
219
220 mvebu_writel(port, cs->base & 0xffff0000,
221 PCIE_WIN04_BASE_OFF(i));
222 mvebu_writel(port, 0, PCIE_WIN04_REMAP_OFF(i));
223 mvebu_writel(port,
224 ((cs->size - 1) & 0xffff0000) |
225 (cs->mbus_attr << 8) |
226 (dram->mbus_dram_target_id << 4) | 1,
227 PCIE_WIN04_CTRL_OFF(i));
228
229 size += cs->size;
230 }
231
232 /* Round up 'size' to the nearest power of two. */
233 if ((size & (size - 1)) != 0)
234 size = 1 << fls(size);
235
236 /* Setup BAR[1] to all DRAM banks. */
237 mvebu_writel(port, dram->cs[0].base, PCIE_BAR_LO_OFF(1));
238 mvebu_writel(port, 0, PCIE_BAR_HI_OFF(1));
239 mvebu_writel(port, ((size - 1) & 0xffff0000) | 1,
240 PCIE_BAR_CTRL_OFF(1));
241
242 /*
243 * Point BAR[0] to the device's internal registers.
244 */
245 mvebu_writel(port, round_down(port->regs.start, SZ_1M), PCIE_BAR_LO_OFF(0));
246 mvebu_writel(port, 0, PCIE_BAR_HI_OFF(0));
247}
248
249static void mvebu_pcie_setup_hw(struct mvebu_pcie_port *port)
250{
251 u32 ctrl, lnkcap, cmd, dev_rev, unmask, sspl;
252
253 /* Setup PCIe controller to Root Complex mode. */
254 ctrl = mvebu_readl(port, PCIE_CTRL_OFF);
255 ctrl |= PCIE_CTRL_RC_MODE;
256 mvebu_writel(port, ctrl, PCIE_CTRL_OFF);
257
258 /*
259 * Set Maximum Link Width to X1 or X4 in Root Port's PCIe Link
260 * Capability register. This register is defined by PCIe specification
261 * as read-only but this mvebu controller has it as read-write and must
262 * be set to number of SerDes PCIe lanes (1 or 4). If this register is
263 * not set correctly then link with endpoint card is not established.
264 */
265 lnkcap = mvebu_readl(port, PCIE_CAP_PCIEXP + PCI_EXP_LNKCAP);
266 lnkcap &= ~PCI_EXP_LNKCAP_MLW;
267 lnkcap |= FIELD_PREP(PCI_EXP_LNKCAP_MLW, port->is_x4 ? 4 : 1);
268 mvebu_writel(port, lnkcap, PCIE_CAP_PCIEXP + PCI_EXP_LNKCAP);
269
270 /* Disable Root Bridge I/O space, memory space and bus mastering. */
271 cmd = mvebu_readl(port, PCIE_CMD_OFF);
272 cmd &= ~(PCI_COMMAND_IO | PCI_COMMAND_MEMORY | PCI_COMMAND_MASTER);
273 mvebu_writel(port, cmd, PCIE_CMD_OFF);
274
275 /*
276 * Change Class Code of PCI Bridge device to PCI Bridge (0x6004)
277 * because default value is Memory controller (0x5080).
278 *
279 * Note that this mvebu PCI Bridge does not have compliant Type 1
280 * Configuration Space. Header Type is reported as Type 0 and it
281 * has format of Type 0 config space.
282 *
283 * Moreover Type 0 BAR registers (ranges 0x10 - 0x28 and 0x30 - 0x34)
284 * have the same format in Marvell's specification as in PCIe
285 * specification, but their meaning is totally different and they do
286 * different things: they are aliased into internal mvebu registers
287 * (e.g. PCIE_BAR_LO_OFF) and these should not be changed or
288 * reconfigured by pci device drivers.
289 *
290 * Therefore driver uses emulation of PCI Bridge which emulates
291 * access to configuration space via internal mvebu registers or
292 * emulated configuration buffer. Driver access these PCI Bridge
293 * directly for simplification, but these registers can be accessed
294 * also via standard mvebu way for accessing PCI config space.
295 */
296 dev_rev = mvebu_readl(port, PCIE_DEV_REV_OFF);
297 dev_rev &= ~0xffffff00;
298 dev_rev |= PCI_CLASS_BRIDGE_PCI_NORMAL << 8;
299 mvebu_writel(port, dev_rev, PCIE_DEV_REV_OFF);
300
301 /* Point PCIe unit MBUS decode windows to DRAM space. */
302 mvebu_pcie_setup_wins(port);
303
304 /*
305 * Program Root Port to automatically send Set_Slot_Power_Limit
306 * PCIe Message when changing status from Dl_Down to Dl_Up and valid
307 * slot power limit was specified.
308 */
309 sspl = mvebu_readl(port, PCIE_SSPL_OFF);
310 sspl &= ~(PCIE_SSPL_VALUE_MASK | PCIE_SSPL_SCALE_MASK | PCIE_SSPL_ENABLE);
311 if (port->slot_power_limit_value) {
312 sspl |= port->slot_power_limit_value << PCIE_SSPL_VALUE_SHIFT;
313 sspl |= port->slot_power_limit_scale << PCIE_SSPL_SCALE_SHIFT;
314 sspl |= PCIE_SSPL_ENABLE;
315 }
316 mvebu_writel(port, sspl, PCIE_SSPL_OFF);
317
318 /* Mask all interrupt sources. */
319 mvebu_writel(port, ~PCIE_INT_ALL_MASK, PCIE_INT_UNMASK_OFF);
320
321 /* Clear all interrupt causes. */
322 mvebu_writel(port, ~PCIE_INT_ALL_MASK, PCIE_INT_CAUSE_OFF);
323
324 /* Check if "intx" interrupt was specified in DT. */
325 if (port->intx_irq > 0)
326 return;
327
328 /*
329 * Fallback code when "intx" interrupt was not specified in DT:
330 * Unmask all legacy INTx interrupts as driver does not provide a way
331 * for masking and unmasking of individual legacy INTx interrupts.
332 * Legacy INTx are reported via one shared GIC source and therefore
333 * kernel cannot distinguish which individual legacy INTx was triggered.
334 * These interrupts are shared, so it should not cause any issue. Just
335 * performance penalty as every PCIe interrupt handler needs to be
336 * called when some interrupt is triggered.
337 */
338 unmask = mvebu_readl(port, PCIE_INT_UNMASK_OFF);
339 unmask |= PCIE_INT_INTX(0) | PCIE_INT_INTX(1) |
340 PCIE_INT_INTX(2) | PCIE_INT_INTX(3);
341 mvebu_writel(port, unmask, PCIE_INT_UNMASK_OFF);
342}
343
344static struct mvebu_pcie_port *mvebu_pcie_find_port(struct mvebu_pcie *pcie,
345 struct pci_bus *bus,
346 int devfn);
347
348static int mvebu_pcie_child_rd_conf(struct pci_bus *bus, u32 devfn, int where,
349 int size, u32 *val)
350{
351 struct mvebu_pcie *pcie = bus->sysdata;
352 struct mvebu_pcie_port *port;
353 void __iomem *conf_data;
354
355 port = mvebu_pcie_find_port(pcie, bus, devfn);
356 if (!port)
357 return PCIBIOS_DEVICE_NOT_FOUND;
358
359 if (!mvebu_pcie_link_up(port))
360 return PCIBIOS_DEVICE_NOT_FOUND;
361
362 conf_data = port->base + PCIE_CONF_DATA_OFF;
363
364 mvebu_writel(port, PCIE_CONF_ADDR(bus->number, devfn, where),
365 PCIE_CONF_ADDR_OFF);
366
367 switch (size) {
368 case 1:
369 *val = readb_relaxed(conf_data + (where & 3));
370 break;
371 case 2:
372 *val = readw_relaxed(conf_data + (where & 2));
373 break;
374 case 4:
375 *val = readl_relaxed(conf_data);
376 break;
377 default:
378 return PCIBIOS_BAD_REGISTER_NUMBER;
379 }
380
381 return PCIBIOS_SUCCESSFUL;
382}
383
384static int mvebu_pcie_child_wr_conf(struct pci_bus *bus, u32 devfn,
385 int where, int size, u32 val)
386{
387 struct mvebu_pcie *pcie = bus->sysdata;
388 struct mvebu_pcie_port *port;
389 void __iomem *conf_data;
390
391 port = mvebu_pcie_find_port(pcie, bus, devfn);
392 if (!port)
393 return PCIBIOS_DEVICE_NOT_FOUND;
394
395 if (!mvebu_pcie_link_up(port))
396 return PCIBIOS_DEVICE_NOT_FOUND;
397
398 conf_data = port->base + PCIE_CONF_DATA_OFF;
399
400 mvebu_writel(port, PCIE_CONF_ADDR(bus->number, devfn, where),
401 PCIE_CONF_ADDR_OFF);
402
403 switch (size) {
404 case 1:
405 writeb(val, conf_data + (where & 3));
406 break;
407 case 2:
408 writew(val, conf_data + (where & 2));
409 break;
410 case 4:
411 writel(val, conf_data);
412 break;
413 default:
414 return PCIBIOS_BAD_REGISTER_NUMBER;
415 }
416
417 return PCIBIOS_SUCCESSFUL;
418}
419
420static struct pci_ops mvebu_pcie_child_ops = {
421 .read = mvebu_pcie_child_rd_conf,
422 .write = mvebu_pcie_child_wr_conf,
423};
424
425/*
426 * Remove windows, starting from the largest ones to the smallest
427 * ones.
428 */
429static void mvebu_pcie_del_windows(struct mvebu_pcie_port *port,
430 phys_addr_t base, size_t size)
431{
432 while (size) {
433 size_t sz = 1 << (fls(size) - 1);
434
435 mvebu_mbus_del_window(base, sz);
436 base += sz;
437 size -= sz;
438 }
439}
440
441/*
442 * MBus windows can only have a power of two size, but PCI BARs do not
443 * have this constraint. Therefore, we have to split the PCI BAR into
444 * areas each having a power of two size. We start from the largest
445 * one (i.e highest order bit set in the size).
446 */
447static int mvebu_pcie_add_windows(struct mvebu_pcie_port *port,
448 unsigned int target, unsigned int attribute,
449 phys_addr_t base, size_t size,
450 phys_addr_t remap)
451{
452 size_t size_mapped = 0;
453
454 while (size) {
455 size_t sz = 1 << (fls(size) - 1);
456 int ret;
457
458 ret = mvebu_mbus_add_window_remap_by_id(target, attribute, base,
459 sz, remap);
460 if (ret) {
461 phys_addr_t end = base + sz - 1;
462
463 dev_err(&port->pcie->pdev->dev,
464 "Could not create MBus window at [mem %pa-%pa]: %d\n",
465 &base, &end, ret);
466 mvebu_pcie_del_windows(port, base - size_mapped,
467 size_mapped);
468 return ret;
469 }
470
471 size -= sz;
472 size_mapped += sz;
473 base += sz;
474 if (remap != MVEBU_MBUS_NO_REMAP)
475 remap += sz;
476 }
477
478 return 0;
479}
480
481static int mvebu_pcie_set_window(struct mvebu_pcie_port *port,
482 unsigned int target, unsigned int attribute,
483 const struct mvebu_pcie_window *desired,
484 struct mvebu_pcie_window *cur)
485{
486 int ret;
487
488 if (desired->base == cur->base && desired->remap == cur->remap &&
489 desired->size == cur->size)
490 return 0;
491
492 if (cur->size != 0) {
493 mvebu_pcie_del_windows(port, cur->base, cur->size);
494 cur->size = 0;
495 cur->base = 0;
496
497 /*
498 * If something tries to change the window while it is enabled
499 * the change will not be done atomically. That would be
500 * difficult to do in the general case.
501 */
502 }
503
504 if (desired->size == 0)
505 return 0;
506
507 ret = mvebu_pcie_add_windows(port, target, attribute, desired->base,
508 desired->size, desired->remap);
509 if (ret) {
510 cur->size = 0;
511 cur->base = 0;
512 return ret;
513 }
514
515 *cur = *desired;
516 return 0;
517}
518
519static int mvebu_pcie_handle_iobase_change(struct mvebu_pcie_port *port)
520{
521 struct mvebu_pcie_window desired = {};
522 struct pci_bridge_emul_conf *conf = &port->bridge.conf;
523
524 /* Are the new iobase/iolimit values invalid? */
525 if (conf->iolimit < conf->iobase ||
526 le16_to_cpu(conf->iolimitupper) < le16_to_cpu(conf->iobaseupper))
527 return mvebu_pcie_set_window(port, port->io_target, port->io_attr,
528 &desired, &port->iowin);
529
530 /*
531 * We read the PCI-to-PCI bridge emulated registers, and
532 * calculate the base address and size of the address decoding
533 * window to setup, according to the PCI-to-PCI bridge
534 * specifications. iobase is the bus address, port->iowin_base
535 * is the CPU address.
536 */
537 desired.remap = ((conf->iobase & 0xF0) << 8) |
538 (le16_to_cpu(conf->iobaseupper) << 16);
539 desired.base = port->pcie->io.start + desired.remap;
540 desired.size = ((0xFFF | ((conf->iolimit & 0xF0) << 8) |
541 (le16_to_cpu(conf->iolimitupper) << 16)) -
542 desired.remap) +
543 1;
544
545 return mvebu_pcie_set_window(port, port->io_target, port->io_attr, &desired,
546 &port->iowin);
547}
548
549static int mvebu_pcie_handle_membase_change(struct mvebu_pcie_port *port)
550{
551 struct mvebu_pcie_window desired = {.remap = MVEBU_MBUS_NO_REMAP};
552 struct pci_bridge_emul_conf *conf = &port->bridge.conf;
553
554 /* Are the new membase/memlimit values invalid? */
555 if (le16_to_cpu(conf->memlimit) < le16_to_cpu(conf->membase))
556 return mvebu_pcie_set_window(port, port->mem_target, port->mem_attr,
557 &desired, &port->memwin);
558
559 /*
560 * We read the PCI-to-PCI bridge emulated registers, and
561 * calculate the base address and size of the address decoding
562 * window to setup, according to the PCI-to-PCI bridge
563 * specifications.
564 */
565 desired.base = ((le16_to_cpu(conf->membase) & 0xFFF0) << 16);
566 desired.size = (((le16_to_cpu(conf->memlimit) & 0xFFF0) << 16) | 0xFFFFF) -
567 desired.base + 1;
568
569 return mvebu_pcie_set_window(port, port->mem_target, port->mem_attr, &desired,
570 &port->memwin);
571}
572
573static pci_bridge_emul_read_status_t
574mvebu_pci_bridge_emul_base_conf_read(struct pci_bridge_emul *bridge,
575 int reg, u32 *value)
576{
577 struct mvebu_pcie_port *port = bridge->data;
578
579 switch (reg) {
580 case PCI_COMMAND:
581 *value = mvebu_readl(port, PCIE_CMD_OFF);
582 break;
583
584 case PCI_PRIMARY_BUS: {
585 /*
586 * From the whole 32bit register we support reading from HW only
587 * secondary bus number which is mvebu local bus number.
588 * Other bits are retrieved only from emulated config buffer.
589 */
590 __le32 *cfgspace = (__le32 *)&bridge->conf;
591 u32 val = le32_to_cpu(cfgspace[PCI_PRIMARY_BUS / 4]);
592 val &= ~0xff00;
593 val |= mvebu_pcie_get_local_bus_nr(port) << 8;
594 *value = val;
595 break;
596 }
597
598 case PCI_INTERRUPT_LINE: {
599 /*
600 * From the whole 32bit register we support reading from HW only
601 * one bit: PCI_BRIDGE_CTL_BUS_RESET.
602 * Other bits are retrieved only from emulated config buffer.
603 */
604 __le32 *cfgspace = (__le32 *)&bridge->conf;
605 u32 val = le32_to_cpu(cfgspace[PCI_INTERRUPT_LINE / 4]);
606 if (mvebu_readl(port, PCIE_CTRL_OFF) & PCIE_CTRL_MASTER_HOT_RESET)
607 val |= PCI_BRIDGE_CTL_BUS_RESET << 16;
608 else
609 val &= ~(PCI_BRIDGE_CTL_BUS_RESET << 16);
610 *value = val;
611 break;
612 }
613
614 default:
615 return PCI_BRIDGE_EMUL_NOT_HANDLED;
616 }
617
618 return PCI_BRIDGE_EMUL_HANDLED;
619}
620
621static pci_bridge_emul_read_status_t
622mvebu_pci_bridge_emul_pcie_conf_read(struct pci_bridge_emul *bridge,
623 int reg, u32 *value)
624{
625 struct mvebu_pcie_port *port = bridge->data;
626
627 switch (reg) {
628 case PCI_EXP_DEVCAP:
629 *value = mvebu_readl(port, PCIE_CAP_PCIEXP + PCI_EXP_DEVCAP);
630 break;
631
632 case PCI_EXP_DEVCTL:
633 *value = mvebu_readl(port, PCIE_CAP_PCIEXP + PCI_EXP_DEVCTL);
634 break;
635
636 case PCI_EXP_LNKCAP:
637 /*
638 * PCIe requires that the Clock Power Management capability bit
639 * is hard-wired to zero for downstream ports but HW returns 1.
640 * Additionally enable Data Link Layer Link Active Reporting
641 * Capable bit as DL_Active indication is provided too.
642 */
643 *value = (mvebu_readl(port, PCIE_CAP_PCIEXP + PCI_EXP_LNKCAP) &
644 ~PCI_EXP_LNKCAP_CLKPM) | PCI_EXP_LNKCAP_DLLLARC;
645 break;
646
647 case PCI_EXP_LNKCTL:
648 /* DL_Active indication is provided via PCIE_STAT_OFF */
649 *value = mvebu_readl(port, PCIE_CAP_PCIEXP + PCI_EXP_LNKCTL) |
650 (mvebu_pcie_link_up(port) ?
651 (PCI_EXP_LNKSTA_DLLLA << 16) : 0);
652 break;
653
654 case PCI_EXP_SLTCTL: {
655 u16 slotctl = le16_to_cpu(bridge->pcie_conf.slotctl);
656 u16 slotsta = le16_to_cpu(bridge->pcie_conf.slotsta);
657 u32 val = 0;
658 /*
659 * When slot power limit was not specified in DT then
660 * ASPL_DISABLE bit is stored only in emulated config space.
661 * Otherwise reflect status of PCIE_SSPL_ENABLE bit in HW.
662 */
663 if (!port->slot_power_limit_value)
664 val |= slotctl & PCI_EXP_SLTCTL_ASPL_DISABLE;
665 else if (!(mvebu_readl(port, PCIE_SSPL_OFF) & PCIE_SSPL_ENABLE))
666 val |= PCI_EXP_SLTCTL_ASPL_DISABLE;
667 /* This callback is 32-bit and in high bits is slot status. */
668 val |= slotsta << 16;
669 *value = val;
670 break;
671 }
672
673 case PCI_EXP_RTSTA:
674 *value = mvebu_readl(port, PCIE_RC_RTSTA);
675 break;
676
677 case PCI_EXP_DEVCAP2:
678 *value = mvebu_readl(port, PCIE_CAP_PCIEXP + PCI_EXP_DEVCAP2);
679 break;
680
681 case PCI_EXP_DEVCTL2:
682 *value = mvebu_readl(port, PCIE_CAP_PCIEXP + PCI_EXP_DEVCTL2);
683 break;
684
685 case PCI_EXP_LNKCTL2:
686 *value = mvebu_readl(port, PCIE_CAP_PCIEXP + PCI_EXP_LNKCTL2);
687 break;
688
689 default:
690 return PCI_BRIDGE_EMUL_NOT_HANDLED;
691 }
692
693 return PCI_BRIDGE_EMUL_HANDLED;
694}
695
696static pci_bridge_emul_read_status_t
697mvebu_pci_bridge_emul_ext_conf_read(struct pci_bridge_emul *bridge,
698 int reg, u32 *value)
699{
700 struct mvebu_pcie_port *port = bridge->data;
701
702 switch (reg) {
703 case 0:
704 case PCI_ERR_UNCOR_STATUS:
705 case PCI_ERR_UNCOR_MASK:
706 case PCI_ERR_UNCOR_SEVER:
707 case PCI_ERR_COR_STATUS:
708 case PCI_ERR_COR_MASK:
709 case PCI_ERR_CAP:
710 case PCI_ERR_HEADER_LOG+0:
711 case PCI_ERR_HEADER_LOG+4:
712 case PCI_ERR_HEADER_LOG+8:
713 case PCI_ERR_HEADER_LOG+12:
714 case PCI_ERR_ROOT_COMMAND:
715 case PCI_ERR_ROOT_STATUS:
716 case PCI_ERR_ROOT_ERR_SRC:
717 *value = mvebu_readl(port, PCIE_CAP_PCIERR_OFF + reg);
718 break;
719
720 default:
721 return PCI_BRIDGE_EMUL_NOT_HANDLED;
722 }
723
724 return PCI_BRIDGE_EMUL_HANDLED;
725}
726
727static void
728mvebu_pci_bridge_emul_base_conf_write(struct pci_bridge_emul *bridge,
729 int reg, u32 old, u32 new, u32 mask)
730{
731 struct mvebu_pcie_port *port = bridge->data;
732 struct pci_bridge_emul_conf *conf = &bridge->conf;
733
734 switch (reg) {
735 case PCI_COMMAND:
736 mvebu_writel(port, new, PCIE_CMD_OFF);
737 break;
738
739 case PCI_IO_BASE:
740 if ((mask & 0xffff) && mvebu_has_ioport(port) &&
741 mvebu_pcie_handle_iobase_change(port)) {
742 /* On error disable IO range */
743 conf->iobase &= ~0xf0;
744 conf->iolimit &= ~0xf0;
745 conf->iobase |= 0xf0;
746 conf->iobaseupper = cpu_to_le16(0x0000);
747 conf->iolimitupper = cpu_to_le16(0x0000);
748 }
749 break;
750
751 case PCI_MEMORY_BASE:
752 if (mvebu_pcie_handle_membase_change(port)) {
753 /* On error disable mem range */
754 conf->membase = cpu_to_le16(le16_to_cpu(conf->membase) & ~0xfff0);
755 conf->memlimit = cpu_to_le16(le16_to_cpu(conf->memlimit) & ~0xfff0);
756 conf->membase = cpu_to_le16(le16_to_cpu(conf->membase) | 0xfff0);
757 }
758 break;
759
760 case PCI_IO_BASE_UPPER16:
761 if (mvebu_has_ioport(port) &&
762 mvebu_pcie_handle_iobase_change(port)) {
763 /* On error disable IO range */
764 conf->iobase &= ~0xf0;
765 conf->iolimit &= ~0xf0;
766 conf->iobase |= 0xf0;
767 conf->iobaseupper = cpu_to_le16(0x0000);
768 conf->iolimitupper = cpu_to_le16(0x0000);
769 }
770 break;
771
772 case PCI_PRIMARY_BUS:
773 if (mask & 0xff00)
774 mvebu_pcie_set_local_bus_nr(port, conf->secondary_bus);
775 break;
776
777 case PCI_INTERRUPT_LINE:
778 if (mask & (PCI_BRIDGE_CTL_BUS_RESET << 16)) {
779 u32 ctrl = mvebu_readl(port, PCIE_CTRL_OFF);
780 if (new & (PCI_BRIDGE_CTL_BUS_RESET << 16))
781 ctrl |= PCIE_CTRL_MASTER_HOT_RESET;
782 else
783 ctrl &= ~PCIE_CTRL_MASTER_HOT_RESET;
784 mvebu_writel(port, ctrl, PCIE_CTRL_OFF);
785 }
786 break;
787
788 default:
789 break;
790 }
791}
792
793static void
794mvebu_pci_bridge_emul_pcie_conf_write(struct pci_bridge_emul *bridge,
795 int reg, u32 old, u32 new, u32 mask)
796{
797 struct mvebu_pcie_port *port = bridge->data;
798
799 switch (reg) {
800 case PCI_EXP_DEVCTL:
801 mvebu_writel(port, new, PCIE_CAP_PCIEXP + PCI_EXP_DEVCTL);
802 break;
803
804 case PCI_EXP_LNKCTL:
805 /*
806 * PCIe requires that the Enable Clock Power Management bit
807 * is hard-wired to zero for downstream ports but HW allows
808 * to change it.
809 */
810 new &= ~PCI_EXP_LNKCTL_CLKREQ_EN;
811
812 mvebu_writel(port, new, PCIE_CAP_PCIEXP + PCI_EXP_LNKCTL);
813 break;
814
815 case PCI_EXP_SLTCTL:
816 /*
817 * Allow to change PCIE_SSPL_ENABLE bit only when slot power
818 * limit was specified in DT and configured into HW.
819 */
820 if ((mask & PCI_EXP_SLTCTL_ASPL_DISABLE) &&
821 port->slot_power_limit_value) {
822 u32 sspl = mvebu_readl(port, PCIE_SSPL_OFF);
823 if (new & PCI_EXP_SLTCTL_ASPL_DISABLE)
824 sspl &= ~PCIE_SSPL_ENABLE;
825 else
826 sspl |= PCIE_SSPL_ENABLE;
827 mvebu_writel(port, sspl, PCIE_SSPL_OFF);
828 }
829 break;
830
831 case PCI_EXP_RTSTA:
832 /*
833 * PME Status bit in Root Status Register (PCIE_RC_RTSTA)
834 * is read-only and can be cleared only by writing 0b to the
835 * Interrupt Cause RW0C register (PCIE_INT_CAUSE_OFF). So
836 * clear PME via Interrupt Cause.
837 */
838 if (new & PCI_EXP_RTSTA_PME)
839 mvebu_writel(port, ~PCIE_INT_PM_PME, PCIE_INT_CAUSE_OFF);
840 break;
841
842 case PCI_EXP_DEVCTL2:
843 mvebu_writel(port, new, PCIE_CAP_PCIEXP + PCI_EXP_DEVCTL2);
844 break;
845
846 case PCI_EXP_LNKCTL2:
847 mvebu_writel(port, new, PCIE_CAP_PCIEXP + PCI_EXP_LNKCTL2);
848 break;
849
850 default:
851 break;
852 }
853}
854
855static void
856mvebu_pci_bridge_emul_ext_conf_write(struct pci_bridge_emul *bridge,
857 int reg, u32 old, u32 new, u32 mask)
858{
859 struct mvebu_pcie_port *port = bridge->data;
860
861 switch (reg) {
862 /* These are W1C registers, so clear other bits */
863 case PCI_ERR_UNCOR_STATUS:
864 case PCI_ERR_COR_STATUS:
865 case PCI_ERR_ROOT_STATUS:
866 new &= mask;
867 fallthrough;
868
869 case PCI_ERR_UNCOR_MASK:
870 case PCI_ERR_UNCOR_SEVER:
871 case PCI_ERR_COR_MASK:
872 case PCI_ERR_CAP:
873 case PCI_ERR_HEADER_LOG+0:
874 case PCI_ERR_HEADER_LOG+4:
875 case PCI_ERR_HEADER_LOG+8:
876 case PCI_ERR_HEADER_LOG+12:
877 case PCI_ERR_ROOT_COMMAND:
878 case PCI_ERR_ROOT_ERR_SRC:
879 mvebu_writel(port, new, PCIE_CAP_PCIERR_OFF + reg);
880 break;
881
882 default:
883 break;
884 }
885}
886
887static const struct pci_bridge_emul_ops mvebu_pci_bridge_emul_ops = {
888 .read_base = mvebu_pci_bridge_emul_base_conf_read,
889 .write_base = mvebu_pci_bridge_emul_base_conf_write,
890 .read_pcie = mvebu_pci_bridge_emul_pcie_conf_read,
891 .write_pcie = mvebu_pci_bridge_emul_pcie_conf_write,
892 .read_ext = mvebu_pci_bridge_emul_ext_conf_read,
893 .write_ext = mvebu_pci_bridge_emul_ext_conf_write,
894};
895
896/*
897 * Initialize the configuration space of the PCI-to-PCI bridge
898 * associated with the given PCIe interface.
899 */
900static int mvebu_pci_bridge_emul_init(struct mvebu_pcie_port *port)
901{
902 unsigned int bridge_flags = PCI_BRIDGE_EMUL_NO_PREFMEM_FORWARD;
903 struct pci_bridge_emul *bridge = &port->bridge;
904 u32 dev_id = mvebu_readl(port, PCIE_DEV_ID_OFF);
905 u32 dev_rev = mvebu_readl(port, PCIE_DEV_REV_OFF);
906 u32 ssdev_id = mvebu_readl(port, PCIE_SSDEV_ID_OFF);
907 u32 pcie_cap = mvebu_readl(port, PCIE_CAP_PCIEXP);
908 u8 pcie_cap_ver = ((pcie_cap >> 16) & PCI_EXP_FLAGS_VERS);
909
910 bridge->conf.vendor = cpu_to_le16(dev_id & 0xffff);
911 bridge->conf.device = cpu_to_le16(dev_id >> 16);
912 bridge->conf.class_revision = cpu_to_le32(dev_rev & 0xff);
913
914 if (mvebu_has_ioport(port)) {
915 /* We support 32 bits I/O addressing */
916 bridge->conf.iobase = PCI_IO_RANGE_TYPE_32;
917 bridge->conf.iolimit = PCI_IO_RANGE_TYPE_32;
918 } else {
919 bridge_flags |= PCI_BRIDGE_EMUL_NO_IO_FORWARD;
920 }
921
922 /*
923 * Older mvebu hardware provides PCIe Capability structure only in
924 * version 1. New hardware provides it in version 2.
925 * Enable slot support which is emulated.
926 */
927 bridge->pcie_conf.cap = cpu_to_le16(pcie_cap_ver | PCI_EXP_FLAGS_SLOT);
928
929 /*
930 * Set Presence Detect State bit permanently as there is no support for
931 * unplugging PCIe card from the slot. Assume that PCIe card is always
932 * connected in slot.
933 *
934 * Set physical slot number to port+1 as mvebu ports are indexed from
935 * zero and zero value is reserved for ports within the same silicon
936 * as Root Port which is not mvebu case.
937 *
938 * Also set correct slot power limit.
939 */
940 bridge->pcie_conf.slotcap = cpu_to_le32(
941 FIELD_PREP(PCI_EXP_SLTCAP_SPLV, port->slot_power_limit_value) |
942 FIELD_PREP(PCI_EXP_SLTCAP_SPLS, port->slot_power_limit_scale) |
943 FIELD_PREP(PCI_EXP_SLTCAP_PSN, port->port+1));
944 bridge->pcie_conf.slotsta = cpu_to_le16(PCI_EXP_SLTSTA_PDS);
945
946 bridge->subsystem_vendor_id = ssdev_id & 0xffff;
947 bridge->subsystem_id = ssdev_id >> 16;
948 bridge->has_pcie = true;
949 bridge->pcie_start = PCIE_CAP_PCIEXP;
950 bridge->data = port;
951 bridge->ops = &mvebu_pci_bridge_emul_ops;
952
953 return pci_bridge_emul_init(bridge, bridge_flags);
954}
955
956static inline struct mvebu_pcie *sys_to_pcie(struct pci_sys_data *sys)
957{
958 return sys->private_data;
959}
960
961static struct mvebu_pcie_port *mvebu_pcie_find_port(struct mvebu_pcie *pcie,
962 struct pci_bus *bus,
963 int devfn)
964{
965 int i;
966
967 for (i = 0; i < pcie->nports; i++) {
968 struct mvebu_pcie_port *port = &pcie->ports[i];
969
970 if (!port->base)
971 continue;
972
973 if (bus->number == 0 && port->devfn == devfn)
974 return port;
975 if (bus->number != 0 &&
976 bus->number >= port->bridge.conf.secondary_bus &&
977 bus->number <= port->bridge.conf.subordinate_bus)
978 return port;
979 }
980
981 return NULL;
982}
983
984/* PCI configuration space write function */
985static int mvebu_pcie_wr_conf(struct pci_bus *bus, u32 devfn,
986 int where, int size, u32 val)
987{
988 struct mvebu_pcie *pcie = bus->sysdata;
989 struct mvebu_pcie_port *port;
990
991 port = mvebu_pcie_find_port(pcie, bus, devfn);
992 if (!port)
993 return PCIBIOS_DEVICE_NOT_FOUND;
994
995 return pci_bridge_emul_conf_write(&port->bridge, where, size, val);
996}
997
998/* PCI configuration space read function */
999static int mvebu_pcie_rd_conf(struct pci_bus *bus, u32 devfn, int where,
1000 int size, u32 *val)
1001{
1002 struct mvebu_pcie *pcie = bus->sysdata;
1003 struct mvebu_pcie_port *port;
1004
1005 port = mvebu_pcie_find_port(pcie, bus, devfn);
1006 if (!port)
1007 return PCIBIOS_DEVICE_NOT_FOUND;
1008
1009 return pci_bridge_emul_conf_read(&port->bridge, where, size, val);
1010}
1011
1012static struct pci_ops mvebu_pcie_ops = {
1013 .read = mvebu_pcie_rd_conf,
1014 .write = mvebu_pcie_wr_conf,
1015};
1016
1017static void mvebu_pcie_intx_irq_mask(struct irq_data *d)
1018{
1019 struct mvebu_pcie_port *port = d->domain->host_data;
1020 irq_hw_number_t hwirq = irqd_to_hwirq(d);
1021 unsigned long flags;
1022 u32 unmask;
1023
1024 raw_spin_lock_irqsave(&port->irq_lock, flags);
1025 unmask = mvebu_readl(port, PCIE_INT_UNMASK_OFF);
1026 unmask &= ~PCIE_INT_INTX(hwirq);
1027 mvebu_writel(port, unmask, PCIE_INT_UNMASK_OFF);
1028 raw_spin_unlock_irqrestore(&port->irq_lock, flags);
1029}
1030
1031static void mvebu_pcie_intx_irq_unmask(struct irq_data *d)
1032{
1033 struct mvebu_pcie_port *port = d->domain->host_data;
1034 irq_hw_number_t hwirq = irqd_to_hwirq(d);
1035 unsigned long flags;
1036 u32 unmask;
1037
1038 raw_spin_lock_irqsave(&port->irq_lock, flags);
1039 unmask = mvebu_readl(port, PCIE_INT_UNMASK_OFF);
1040 unmask |= PCIE_INT_INTX(hwirq);
1041 mvebu_writel(port, unmask, PCIE_INT_UNMASK_OFF);
1042 raw_spin_unlock_irqrestore(&port->irq_lock, flags);
1043}
1044
1045static struct irq_chip intx_irq_chip = {
1046 .name = "mvebu-INTx",
1047 .irq_mask = mvebu_pcie_intx_irq_mask,
1048 .irq_unmask = mvebu_pcie_intx_irq_unmask,
1049};
1050
1051static int mvebu_pcie_intx_irq_map(struct irq_domain *h,
1052 unsigned int virq, irq_hw_number_t hwirq)
1053{
1054 struct mvebu_pcie_port *port = h->host_data;
1055
1056 irq_set_status_flags(virq, IRQ_LEVEL);
1057 irq_set_chip_and_handler(virq, &intx_irq_chip, handle_level_irq);
1058 irq_set_chip_data(virq, port);
1059
1060 return 0;
1061}
1062
1063static const struct irq_domain_ops mvebu_pcie_intx_irq_domain_ops = {
1064 .map = mvebu_pcie_intx_irq_map,
1065 .xlate = irq_domain_xlate_onecell,
1066};
1067
1068static int mvebu_pcie_init_irq_domain(struct mvebu_pcie_port *port)
1069{
1070 struct device *dev = &port->pcie->pdev->dev;
1071 struct device_node *pcie_intc_node;
1072
1073 raw_spin_lock_init(&port->irq_lock);
1074
1075 pcie_intc_node = of_get_next_child(port->dn, NULL);
1076 if (!pcie_intc_node) {
1077 dev_err(dev, "No PCIe Intc node found for %s\n", port->name);
1078 return -ENODEV;
1079 }
1080
1081 port->intx_irq_domain = irq_domain_add_linear(pcie_intc_node, PCI_NUM_INTX,
1082 &mvebu_pcie_intx_irq_domain_ops,
1083 port);
1084 of_node_put(pcie_intc_node);
1085 if (!port->intx_irq_domain) {
1086 dev_err(dev, "Failed to get INTx IRQ domain for %s\n", port->name);
1087 return -ENOMEM;
1088 }
1089
1090 return 0;
1091}
1092
1093static void mvebu_pcie_irq_handler(struct irq_desc *desc)
1094{
1095 struct mvebu_pcie_port *port = irq_desc_get_handler_data(desc);
1096 struct irq_chip *chip = irq_desc_get_chip(desc);
1097 struct device *dev = &port->pcie->pdev->dev;
1098 u32 cause, unmask, status;
1099 int i;
1100
1101 chained_irq_enter(chip, desc);
1102
1103 cause = mvebu_readl(port, PCIE_INT_CAUSE_OFF);
1104 unmask = mvebu_readl(port, PCIE_INT_UNMASK_OFF);
1105 status = cause & unmask;
1106
1107 /* Process legacy INTx interrupts */
1108 for (i = 0; i < PCI_NUM_INTX; i++) {
1109 if (!(status & PCIE_INT_INTX(i)))
1110 continue;
1111
1112 if (generic_handle_domain_irq(port->intx_irq_domain, i) == -EINVAL)
1113 dev_err_ratelimited(dev, "unexpected INT%c IRQ\n", (char)i+'A');
1114 }
1115
1116 chained_irq_exit(chip, desc);
1117}
1118
1119static int mvebu_pcie_map_irq(const struct pci_dev *dev, u8 slot, u8 pin)
1120{
1121 /* Interrupt support on mvebu emulated bridges is not implemented yet */
1122 if (dev->bus->number == 0)
1123 return 0; /* Proper return code 0 == NO_IRQ */
1124
1125 return of_irq_parse_and_map_pci(dev, slot, pin);
1126}
1127
1128static resource_size_t mvebu_pcie_align_resource(struct pci_dev *dev,
1129 const struct resource *res,
1130 resource_size_t start,
1131 resource_size_t size,
1132 resource_size_t align)
1133{
1134 if (dev->bus->number != 0)
1135 return start;
1136
1137 /*
1138 * On the PCI-to-PCI bridge side, the I/O windows must have at
1139 * least a 64 KB size and the memory windows must have at
1140 * least a 1 MB size. Moreover, MBus windows need to have a
1141 * base address aligned on their size, and their size must be
1142 * a power of two. This means that if the BAR doesn't have a
1143 * power of two size, several MBus windows will actually be
1144 * created. We need to ensure that the biggest MBus window
1145 * (which will be the first one) is aligned on its size, which
1146 * explains the rounddown_pow_of_two() being done here.
1147 */
1148 if (res->flags & IORESOURCE_IO)
1149 return round_up(start, max_t(resource_size_t, SZ_64K,
1150 rounddown_pow_of_two(size)));
1151 else if (res->flags & IORESOURCE_MEM)
1152 return round_up(start, max_t(resource_size_t, SZ_1M,
1153 rounddown_pow_of_two(size)));
1154 else
1155 return start;
1156}
1157
1158static void __iomem *mvebu_pcie_map_registers(struct platform_device *pdev,
1159 struct device_node *np,
1160 struct mvebu_pcie_port *port)
1161{
1162 int ret = 0;
1163
1164 ret = of_address_to_resource(np, 0, &port->regs);
1165 if (ret)
1166 return (void __iomem *)ERR_PTR(ret);
1167
1168 return devm_ioremap_resource(&pdev->dev, &port->regs);
1169}
1170
1171#define DT_FLAGS_TO_TYPE(flags) (((flags) >> 24) & 0x03)
1172#define DT_TYPE_IO 0x1
1173#define DT_TYPE_MEM32 0x2
1174#define DT_CPUADDR_TO_TARGET(cpuaddr) (((cpuaddr) >> 56) & 0xFF)
1175#define DT_CPUADDR_TO_ATTR(cpuaddr) (((cpuaddr) >> 48) & 0xFF)
1176
1177static int mvebu_get_tgt_attr(struct device_node *np, int devfn,
1178 unsigned long type,
1179 unsigned int *tgt,
1180 unsigned int *attr)
1181{
1182 const int na = 3, ns = 2;
1183 const __be32 *range;
1184 int rlen, nranges, rangesz, pna, i;
1185
1186 *tgt = -1;
1187 *attr = -1;
1188
1189 range = of_get_property(np, "ranges", &rlen);
1190 if (!range)
1191 return -EINVAL;
1192
1193 pna = of_n_addr_cells(np);
1194 rangesz = pna + na + ns;
1195 nranges = rlen / sizeof(__be32) / rangesz;
1196
1197 for (i = 0; i < nranges; i++, range += rangesz) {
1198 u32 flags = of_read_number(range, 1);
1199 u32 slot = of_read_number(range + 1, 1);
1200 u64 cpuaddr = of_read_number(range + na, pna);
1201 unsigned long rtype;
1202
1203 if (DT_FLAGS_TO_TYPE(flags) == DT_TYPE_IO)
1204 rtype = IORESOURCE_IO;
1205 else if (DT_FLAGS_TO_TYPE(flags) == DT_TYPE_MEM32)
1206 rtype = IORESOURCE_MEM;
1207 else
1208 continue;
1209
1210 if (slot == PCI_SLOT(devfn) && type == rtype) {
1211 *tgt = DT_CPUADDR_TO_TARGET(cpuaddr);
1212 *attr = DT_CPUADDR_TO_ATTR(cpuaddr);
1213 return 0;
1214 }
1215 }
1216
1217 return -ENOENT;
1218}
1219
1220static int mvebu_pcie_suspend(struct device *dev)
1221{
1222 struct mvebu_pcie *pcie;
1223 int i;
1224
1225 pcie = dev_get_drvdata(dev);
1226 for (i = 0; i < pcie->nports; i++) {
1227 struct mvebu_pcie_port *port = pcie->ports + i;
1228 if (!port->base)
1229 continue;
1230 port->saved_pcie_stat = mvebu_readl(port, PCIE_STAT_OFF);
1231 }
1232
1233 return 0;
1234}
1235
1236static int mvebu_pcie_resume(struct device *dev)
1237{
1238 struct mvebu_pcie *pcie;
1239 int i;
1240
1241 pcie = dev_get_drvdata(dev);
1242 for (i = 0; i < pcie->nports; i++) {
1243 struct mvebu_pcie_port *port = pcie->ports + i;
1244 if (!port->base)
1245 continue;
1246 mvebu_writel(port, port->saved_pcie_stat, PCIE_STAT_OFF);
1247 mvebu_pcie_setup_hw(port);
1248 }
1249
1250 return 0;
1251}
1252
1253static void mvebu_pcie_port_clk_put(void *data)
1254{
1255 struct mvebu_pcie_port *port = data;
1256
1257 clk_put(port->clk);
1258}
1259
1260static int mvebu_pcie_parse_port(struct mvebu_pcie *pcie,
1261 struct mvebu_pcie_port *port, struct device_node *child)
1262{
1263 struct device *dev = &pcie->pdev->dev;
1264 u32 slot_power_limit;
1265 int ret;
1266 u32 num_lanes;
1267
1268 port->pcie = pcie;
1269
1270 if (of_property_read_u32(child, "marvell,pcie-port", &port->port)) {
1271 dev_warn(dev, "ignoring %pOF, missing pcie-port property\n",
1272 child);
1273 goto skip;
1274 }
1275
1276 if (of_property_read_u32(child, "marvell,pcie-lane", &port->lane))
1277 port->lane = 0;
1278
1279 if (!of_property_read_u32(child, "num-lanes", &num_lanes) && num_lanes == 4)
1280 port->is_x4 = true;
1281
1282 port->name = devm_kasprintf(dev, GFP_KERNEL, "pcie%d.%d", port->port,
1283 port->lane);
1284 if (!port->name) {
1285 ret = -ENOMEM;
1286 goto err;
1287 }
1288
1289 port->devfn = of_pci_get_devfn(child);
1290 if (port->devfn < 0)
1291 goto skip;
1292 if (PCI_FUNC(port->devfn) != 0) {
1293 dev_err(dev, "%s: invalid function number, must be zero\n",
1294 port->name);
1295 goto skip;
1296 }
1297
1298 ret = mvebu_get_tgt_attr(dev->of_node, port->devfn, IORESOURCE_MEM,
1299 &port->mem_target, &port->mem_attr);
1300 if (ret < 0) {
1301 dev_err(dev, "%s: cannot get tgt/attr for mem window\n",
1302 port->name);
1303 goto skip;
1304 }
1305
1306 if (resource_size(&pcie->io) != 0) {
1307 mvebu_get_tgt_attr(dev->of_node, port->devfn, IORESOURCE_IO,
1308 &port->io_target, &port->io_attr);
1309 } else {
1310 port->io_target = -1;
1311 port->io_attr = -1;
1312 }
1313
1314 /*
1315 * Old DT bindings do not contain "intx" interrupt
1316 * so do not fail probing driver when interrupt does not exist.
1317 */
1318 port->intx_irq = of_irq_get_byname(child, "intx");
1319 if (port->intx_irq == -EPROBE_DEFER) {
1320 ret = port->intx_irq;
1321 goto err;
1322 }
1323 if (port->intx_irq <= 0) {
1324 dev_warn(dev, "%s: legacy INTx interrupts cannot be masked individually, "
1325 "%pOF does not contain intx interrupt\n",
1326 port->name, child);
1327 }
1328
1329 port->reset_name = devm_kasprintf(dev, GFP_KERNEL, "%s-reset",
1330 port->name);
1331 if (!port->reset_name) {
1332 ret = -ENOMEM;
1333 goto err;
1334 }
1335
1336 port->reset_gpio = devm_fwnode_gpiod_get(dev, of_fwnode_handle(child),
1337 "reset", GPIOD_OUT_HIGH,
1338 port->name);
1339 ret = PTR_ERR_OR_ZERO(port->reset_gpio);
1340 if (ret) {
1341 if (ret != -ENOENT)
1342 goto err;
1343 /* reset gpio is optional */
1344 port->reset_gpio = NULL;
1345 devm_kfree(dev, port->reset_name);
1346 port->reset_name = NULL;
1347 }
1348
1349 slot_power_limit = of_pci_get_slot_power_limit(child,
1350 &port->slot_power_limit_value,
1351 &port->slot_power_limit_scale);
1352 if (slot_power_limit)
1353 dev_info(dev, "%s: Slot power limit %u.%uW\n",
1354 port->name,
1355 slot_power_limit / 1000,
1356 (slot_power_limit / 100) % 10);
1357
1358 port->clk = of_clk_get_by_name(child, NULL);
1359 if (IS_ERR(port->clk)) {
1360 dev_err(dev, "%s: cannot get clock\n", port->name);
1361 goto skip;
1362 }
1363
1364 ret = devm_add_action(dev, mvebu_pcie_port_clk_put, port);
1365 if (ret < 0) {
1366 clk_put(port->clk);
1367 goto err;
1368 }
1369
1370 return 1;
1371
1372skip:
1373 ret = 0;
1374
1375 /* In the case of skipping, we need to free these */
1376 devm_kfree(dev, port->reset_name);
1377 port->reset_name = NULL;
1378 devm_kfree(dev, port->name);
1379 port->name = NULL;
1380
1381err:
1382 return ret;
1383}
1384
1385/*
1386 * Power up a PCIe port. PCIe requires the refclk to be stable for 100µs
1387 * prior to releasing PERST. See table 2-4 in section 2.6.2 AC Specifications
1388 * of the PCI Express Card Electromechanical Specification, 1.1.
1389 */
1390static int mvebu_pcie_powerup(struct mvebu_pcie_port *port)
1391{
1392 int ret;
1393
1394 ret = clk_prepare_enable(port->clk);
1395 if (ret < 0)
1396 return ret;
1397
1398 if (port->reset_gpio) {
1399 u32 reset_udelay = PCI_PM_D3COLD_WAIT * 1000;
1400
1401 of_property_read_u32(port->dn, "reset-delay-us",
1402 &reset_udelay);
1403
1404 udelay(100);
1405
1406 gpiod_set_value_cansleep(port->reset_gpio, 0);
1407 msleep(reset_udelay / 1000);
1408 }
1409
1410 return 0;
1411}
1412
1413/*
1414 * Power down a PCIe port. Strictly, PCIe requires us to place the card
1415 * in D3hot state before asserting PERST#.
1416 */
1417static void mvebu_pcie_powerdown(struct mvebu_pcie_port *port)
1418{
1419 gpiod_set_value_cansleep(port->reset_gpio, 1);
1420
1421 clk_disable_unprepare(port->clk);
1422}
1423
1424/*
1425 * devm_of_pci_get_host_bridge_resources() only sets up translateable resources,
1426 * so we need extra resource setup parsing our special DT properties encoding
1427 * the MEM and IO apertures.
1428 */
1429static int mvebu_pcie_parse_request_resources(struct mvebu_pcie *pcie)
1430{
1431 struct device *dev = &pcie->pdev->dev;
1432 struct pci_host_bridge *bridge = pci_host_bridge_from_priv(pcie);
1433 int ret;
1434
1435 /* Get the PCIe memory aperture */
1436 mvebu_mbus_get_pcie_mem_aperture(&pcie->mem);
1437 if (resource_size(&pcie->mem) == 0) {
1438 dev_err(dev, "invalid memory aperture size\n");
1439 return -EINVAL;
1440 }
1441
1442 pcie->mem.name = "PCI MEM";
1443 pci_add_resource(&bridge->windows, &pcie->mem);
1444 ret = devm_request_resource(dev, &iomem_resource, &pcie->mem);
1445 if (ret)
1446 return ret;
1447
1448 /* Get the PCIe IO aperture */
1449 mvebu_mbus_get_pcie_io_aperture(&pcie->io);
1450
1451 if (resource_size(&pcie->io) != 0) {
1452 pcie->realio.flags = pcie->io.flags;
1453 pcie->realio.start = PCIBIOS_MIN_IO;
1454 pcie->realio.end = min_t(resource_size_t,
1455 IO_SPACE_LIMIT - SZ_64K,
1456 resource_size(&pcie->io) - 1);
1457 pcie->realio.name = "PCI I/O";
1458
1459 ret = devm_pci_remap_iospace(dev, &pcie->realio, pcie->io.start);
1460 if (ret)
1461 return ret;
1462
1463 pci_add_resource(&bridge->windows, &pcie->realio);
1464 ret = devm_request_resource(dev, &ioport_resource, &pcie->realio);
1465 if (ret)
1466 return ret;
1467 }
1468
1469 return 0;
1470}
1471
1472static int mvebu_pcie_probe(struct platform_device *pdev)
1473{
1474 struct device *dev = &pdev->dev;
1475 struct mvebu_pcie *pcie;
1476 struct pci_host_bridge *bridge;
1477 struct device_node *np = dev->of_node;
1478 struct device_node *child;
1479 int num, i, ret;
1480
1481 bridge = devm_pci_alloc_host_bridge(dev, sizeof(struct mvebu_pcie));
1482 if (!bridge)
1483 return -ENOMEM;
1484
1485 pcie = pci_host_bridge_priv(bridge);
1486 pcie->pdev = pdev;
1487 platform_set_drvdata(pdev, pcie);
1488
1489 ret = mvebu_pcie_parse_request_resources(pcie);
1490 if (ret)
1491 return ret;
1492
1493 num = of_get_available_child_count(np);
1494
1495 pcie->ports = devm_kcalloc(dev, num, sizeof(*pcie->ports), GFP_KERNEL);
1496 if (!pcie->ports)
1497 return -ENOMEM;
1498
1499 i = 0;
1500 for_each_available_child_of_node(np, child) {
1501 struct mvebu_pcie_port *port = &pcie->ports[i];
1502
1503 ret = mvebu_pcie_parse_port(pcie, port, child);
1504 if (ret < 0) {
1505 of_node_put(child);
1506 return ret;
1507 } else if (ret == 0) {
1508 continue;
1509 }
1510
1511 port->dn = child;
1512 i++;
1513 }
1514 pcie->nports = i;
1515
1516 for (i = 0; i < pcie->nports; i++) {
1517 struct mvebu_pcie_port *port = &pcie->ports[i];
1518 int irq = port->intx_irq;
1519
1520 child = port->dn;
1521 if (!child)
1522 continue;
1523
1524 ret = mvebu_pcie_powerup(port);
1525 if (ret < 0)
1526 continue;
1527
1528 port->base = mvebu_pcie_map_registers(pdev, child, port);
1529 if (IS_ERR(port->base)) {
1530 dev_err(dev, "%s: cannot map registers\n", port->name);
1531 port->base = NULL;
1532 mvebu_pcie_powerdown(port);
1533 continue;
1534 }
1535
1536 ret = mvebu_pci_bridge_emul_init(port);
1537 if (ret < 0) {
1538 dev_err(dev, "%s: cannot init emulated bridge\n",
1539 port->name);
1540 devm_iounmap(dev, port->base);
1541 port->base = NULL;
1542 mvebu_pcie_powerdown(port);
1543 continue;
1544 }
1545
1546 if (irq > 0) {
1547 ret = mvebu_pcie_init_irq_domain(port);
1548 if (ret) {
1549 dev_err(dev, "%s: cannot init irq domain\n",
1550 port->name);
1551 pci_bridge_emul_cleanup(&port->bridge);
1552 devm_iounmap(dev, port->base);
1553 port->base = NULL;
1554 mvebu_pcie_powerdown(port);
1555 continue;
1556 }
1557 irq_set_chained_handler_and_data(irq,
1558 mvebu_pcie_irq_handler,
1559 port);
1560 }
1561
1562 /*
1563 * PCIe topology exported by mvebu hw is quite complicated. In
1564 * reality has something like N fully independent host bridges
1565 * where each host bridge has one PCIe Root Port (which acts as
1566 * PCI Bridge device). Each host bridge has its own independent
1567 * internal registers, independent access to PCI config space,
1568 * independent interrupt lines, independent window and memory
1569 * access configuration. But additionally there is some kind of
1570 * peer-to-peer support between PCIe devices behind different
1571 * host bridges limited just to forwarding of memory and I/O
1572 * transactions (forwarding of error messages and config cycles
1573 * is not supported). So we could say there are N independent
1574 * PCIe Root Complexes.
1575 *
1576 * For this kind of setup DT should have been structured into
1577 * N independent PCIe controllers / host bridges. But instead
1578 * structure in past was defined to put PCIe Root Ports of all
1579 * host bridges into one bus zero, like in classic multi-port
1580 * Root Complex setup with just one host bridge.
1581 *
1582 * This means that pci-mvebu.c driver provides "virtual" bus 0
1583 * on which registers all PCIe Root Ports (PCI Bridge devices)
1584 * specified in DT by their BDF addresses and virtually routes
1585 * PCI config access of each PCI bridge device to specific PCIe
1586 * host bridge.
1587 *
1588 * Normally PCI Bridge should choose between Type 0 and Type 1
1589 * config requests based on primary and secondary bus numbers
1590 * configured on the bridge itself. But because mvebu PCI Bridge
1591 * does not have registers for primary and secondary bus numbers
1592 * in its config space, it determinates type of config requests
1593 * via its own custom way.
1594 *
1595 * There are two options how mvebu determinate type of config
1596 * request.
1597 *
1598 * 1. If Secondary Bus Number Enable bit is not set or is not
1599 * available (applies for pre-XP PCIe controllers) then Type 0
1600 * is used if target bus number equals Local Bus Number (bits
1601 * [15:8] in register 0x1a04) and target device number differs
1602 * from Local Device Number (bits [20:16] in register 0x1a04).
1603 * Type 1 is used if target bus number differs from Local Bus
1604 * Number. And when target bus number equals Local Bus Number
1605 * and target device equals Local Device Number then request is
1606 * routed to Local PCI Bridge (PCIe Root Port).
1607 *
1608 * 2. If Secondary Bus Number Enable bit is set (bit 7 in
1609 * register 0x1a2c) then mvebu hw determinate type of config
1610 * request like compliant PCI Bridge based on primary bus number
1611 * which is configured via Local Bus Number (bits [15:8] in
1612 * register 0x1a04) and secondary bus number which is configured
1613 * via Secondary Bus Number (bits [7:0] in register 0x1a2c).
1614 * Local PCI Bridge (PCIe Root Port) is available on primary bus
1615 * as device with Local Device Number (bits [20:16] in register
1616 * 0x1a04).
1617 *
1618 * Secondary Bus Number Enable bit is disabled by default and
1619 * option 2. is not available on pre-XP PCIe controllers. Hence
1620 * this driver always use option 1.
1621 *
1622 * Basically it means that primary and secondary buses shares
1623 * one virtual number configured via Local Bus Number bits and
1624 * Local Device Number bits determinates if accessing primary
1625 * or secondary bus. Set Local Device Number to 1 and redirect
1626 * all writes of PCI Bridge Secondary Bus Number register to
1627 * Local Bus Number (bits [15:8] in register 0x1a04).
1628 *
1629 * So when accessing devices on buses behind secondary bus
1630 * number it would work correctly. And also when accessing
1631 * device 0 at secondary bus number via config space would be
1632 * correctly routed to secondary bus. Due to issues described
1633 * in mvebu_pcie_setup_hw(), PCI Bridges at primary bus (zero)
1634 * are not accessed directly via PCI config space but rarher
1635 * indirectly via kernel emulated PCI bridge driver.
1636 */
1637 mvebu_pcie_setup_hw(port);
1638 mvebu_pcie_set_local_dev_nr(port, 1);
1639 mvebu_pcie_set_local_bus_nr(port, 0);
1640 }
1641
1642 bridge->sysdata = pcie;
1643 bridge->ops = &mvebu_pcie_ops;
1644 bridge->child_ops = &mvebu_pcie_child_ops;
1645 bridge->align_resource = mvebu_pcie_align_resource;
1646 bridge->map_irq = mvebu_pcie_map_irq;
1647
1648 return pci_host_probe(bridge);
1649}
1650
1651static void mvebu_pcie_remove(struct platform_device *pdev)
1652{
1653 struct mvebu_pcie *pcie = platform_get_drvdata(pdev);
1654 struct pci_host_bridge *bridge = pci_host_bridge_from_priv(pcie);
1655 u32 cmd, sspl;
1656 int i;
1657
1658 /* Remove PCI bus with all devices. */
1659 pci_lock_rescan_remove();
1660 pci_stop_root_bus(bridge->bus);
1661 pci_remove_root_bus(bridge->bus);
1662 pci_unlock_rescan_remove();
1663
1664 for (i = 0; i < pcie->nports; i++) {
1665 struct mvebu_pcie_port *port = &pcie->ports[i];
1666 int irq = port->intx_irq;
1667
1668 if (!port->base)
1669 continue;
1670
1671 /* Disable Root Bridge I/O space, memory space and bus mastering. */
1672 cmd = mvebu_readl(port, PCIE_CMD_OFF);
1673 cmd &= ~(PCI_COMMAND_IO | PCI_COMMAND_MEMORY | PCI_COMMAND_MASTER);
1674 mvebu_writel(port, cmd, PCIE_CMD_OFF);
1675
1676 /* Mask all interrupt sources. */
1677 mvebu_writel(port, ~PCIE_INT_ALL_MASK, PCIE_INT_UNMASK_OFF);
1678
1679 /* Clear all interrupt causes. */
1680 mvebu_writel(port, ~PCIE_INT_ALL_MASK, PCIE_INT_CAUSE_OFF);
1681
1682 if (irq > 0)
1683 irq_set_chained_handler_and_data(irq, NULL, NULL);
1684
1685 /* Remove IRQ domains. */
1686 if (port->intx_irq_domain)
1687 irq_domain_remove(port->intx_irq_domain);
1688
1689 /* Free config space for emulated root bridge. */
1690 pci_bridge_emul_cleanup(&port->bridge);
1691
1692 /* Disable sending Set_Slot_Power_Limit PCIe Message. */
1693 sspl = mvebu_readl(port, PCIE_SSPL_OFF);
1694 sspl &= ~(PCIE_SSPL_VALUE_MASK | PCIE_SSPL_SCALE_MASK | PCIE_SSPL_ENABLE);
1695 mvebu_writel(port, sspl, PCIE_SSPL_OFF);
1696
1697 /* Disable and clear BARs and windows. */
1698 mvebu_pcie_disable_wins(port);
1699
1700 /* Delete PCIe IO and MEM windows. */
1701 if (port->iowin.size)
1702 mvebu_pcie_del_windows(port, port->iowin.base, port->iowin.size);
1703 if (port->memwin.size)
1704 mvebu_pcie_del_windows(port, port->memwin.base, port->memwin.size);
1705
1706 /* Power down card and disable clocks. Must be the last step. */
1707 mvebu_pcie_powerdown(port);
1708 }
1709}
1710
1711static const struct of_device_id mvebu_pcie_of_match_table[] = {
1712 { .compatible = "marvell,armada-xp-pcie", },
1713 { .compatible = "marvell,armada-370-pcie", },
1714 { .compatible = "marvell,dove-pcie", },
1715 { .compatible = "marvell,kirkwood-pcie", },
1716 {},
1717};
1718
1719static const struct dev_pm_ops mvebu_pcie_pm_ops = {
1720 NOIRQ_SYSTEM_SLEEP_PM_OPS(mvebu_pcie_suspend, mvebu_pcie_resume)
1721};
1722
1723static struct platform_driver mvebu_pcie_driver = {
1724 .driver = {
1725 .name = "mvebu-pcie",
1726 .of_match_table = mvebu_pcie_of_match_table,
1727 .pm = &mvebu_pcie_pm_ops,
1728 },
1729 .probe = mvebu_pcie_probe,
1730 .remove = mvebu_pcie_remove,
1731};
1732module_platform_driver(mvebu_pcie_driver);
1733
1734MODULE_AUTHOR("Thomas Petazzoni <thomas.petazzoni@bootlin.com>");
1735MODULE_AUTHOR("Pali Rohár <pali@kernel.org>");
1736MODULE_DESCRIPTION("Marvell EBU PCIe controller");
1737MODULE_LICENSE("GPL v2");
1// SPDX-License-Identifier: GPL-2.0
2/*
3 * PCIe driver for Marvell Armada 370 and Armada XP SoCs
4 *
5 * Author: Thomas Petazzoni <thomas.petazzoni@free-electrons.com>
6 */
7
8#include <linux/kernel.h>
9#include <linux/pci.h>
10#include <linux/clk.h>
11#include <linux/delay.h>
12#include <linux/gpio.h>
13#include <linux/init.h>
14#include <linux/mbus.h>
15#include <linux/msi.h>
16#include <linux/slab.h>
17#include <linux/platform_device.h>
18#include <linux/of_address.h>
19#include <linux/of_irq.h>
20#include <linux/of_gpio.h>
21#include <linux/of_pci.h>
22#include <linux/of_platform.h>
23
24#include "../pci.h"
25#include "../pci-bridge-emul.h"
26
27/*
28 * PCIe unit register offsets.
29 */
30#define PCIE_DEV_ID_OFF 0x0000
31#define PCIE_CMD_OFF 0x0004
32#define PCIE_DEV_REV_OFF 0x0008
33#define PCIE_BAR_LO_OFF(n) (0x0010 + ((n) << 3))
34#define PCIE_BAR_HI_OFF(n) (0x0014 + ((n) << 3))
35#define PCIE_CAP_PCIEXP 0x0060
36#define PCIE_HEADER_LOG_4_OFF 0x0128
37#define PCIE_BAR_CTRL_OFF(n) (0x1804 + (((n) - 1) * 4))
38#define PCIE_WIN04_CTRL_OFF(n) (0x1820 + ((n) << 4))
39#define PCIE_WIN04_BASE_OFF(n) (0x1824 + ((n) << 4))
40#define PCIE_WIN04_REMAP_OFF(n) (0x182c + ((n) << 4))
41#define PCIE_WIN5_CTRL_OFF 0x1880
42#define PCIE_WIN5_BASE_OFF 0x1884
43#define PCIE_WIN5_REMAP_OFF 0x188c
44#define PCIE_CONF_ADDR_OFF 0x18f8
45#define PCIE_CONF_ADDR_EN 0x80000000
46#define PCIE_CONF_REG(r) ((((r) & 0xf00) << 16) | ((r) & 0xfc))
47#define PCIE_CONF_BUS(b) (((b) & 0xff) << 16)
48#define PCIE_CONF_DEV(d) (((d) & 0x1f) << 11)
49#define PCIE_CONF_FUNC(f) (((f) & 0x7) << 8)
50#define PCIE_CONF_ADDR(bus, devfn, where) \
51 (PCIE_CONF_BUS(bus) | PCIE_CONF_DEV(PCI_SLOT(devfn)) | \
52 PCIE_CONF_FUNC(PCI_FUNC(devfn)) | PCIE_CONF_REG(where) | \
53 PCIE_CONF_ADDR_EN)
54#define PCIE_CONF_DATA_OFF 0x18fc
55#define PCIE_MASK_OFF 0x1910
56#define PCIE_MASK_ENABLE_INTS 0x0f000000
57#define PCIE_CTRL_OFF 0x1a00
58#define PCIE_CTRL_X1_MODE 0x0001
59#define PCIE_STAT_OFF 0x1a04
60#define PCIE_STAT_BUS 0xff00
61#define PCIE_STAT_DEV 0x1f0000
62#define PCIE_STAT_LINK_DOWN BIT(0)
63#define PCIE_RC_RTSTA 0x1a14
64#define PCIE_DEBUG_CTRL 0x1a60
65#define PCIE_DEBUG_SOFT_RESET BIT(20)
66
67struct mvebu_pcie_port;
68
69/* Structure representing all PCIe interfaces */
70struct mvebu_pcie {
71 struct platform_device *pdev;
72 struct mvebu_pcie_port *ports;
73 struct msi_controller *msi;
74 struct list_head resources;
75 struct resource io;
76 struct resource realio;
77 struct resource mem;
78 struct resource busn;
79 int nports;
80};
81
82struct mvebu_pcie_window {
83 phys_addr_t base;
84 phys_addr_t remap;
85 size_t size;
86};
87
88/* Structure representing one PCIe interface */
89struct mvebu_pcie_port {
90 char *name;
91 void __iomem *base;
92 u32 port;
93 u32 lane;
94 int devfn;
95 unsigned int mem_target;
96 unsigned int mem_attr;
97 unsigned int io_target;
98 unsigned int io_attr;
99 struct clk *clk;
100 struct gpio_desc *reset_gpio;
101 char *reset_name;
102 struct pci_bridge_emul bridge;
103 struct device_node *dn;
104 struct mvebu_pcie *pcie;
105 struct mvebu_pcie_window memwin;
106 struct mvebu_pcie_window iowin;
107 u32 saved_pcie_stat;
108};
109
110static inline void mvebu_writel(struct mvebu_pcie_port *port, u32 val, u32 reg)
111{
112 writel(val, port->base + reg);
113}
114
115static inline u32 mvebu_readl(struct mvebu_pcie_port *port, u32 reg)
116{
117 return readl(port->base + reg);
118}
119
120static inline bool mvebu_has_ioport(struct mvebu_pcie_port *port)
121{
122 return port->io_target != -1 && port->io_attr != -1;
123}
124
125static bool mvebu_pcie_link_up(struct mvebu_pcie_port *port)
126{
127 return !(mvebu_readl(port, PCIE_STAT_OFF) & PCIE_STAT_LINK_DOWN);
128}
129
130static void mvebu_pcie_set_local_bus_nr(struct mvebu_pcie_port *port, int nr)
131{
132 u32 stat;
133
134 stat = mvebu_readl(port, PCIE_STAT_OFF);
135 stat &= ~PCIE_STAT_BUS;
136 stat |= nr << 8;
137 mvebu_writel(port, stat, PCIE_STAT_OFF);
138}
139
140static void mvebu_pcie_set_local_dev_nr(struct mvebu_pcie_port *port, int nr)
141{
142 u32 stat;
143
144 stat = mvebu_readl(port, PCIE_STAT_OFF);
145 stat &= ~PCIE_STAT_DEV;
146 stat |= nr << 16;
147 mvebu_writel(port, stat, PCIE_STAT_OFF);
148}
149
150/*
151 * Setup PCIE BARs and Address Decode Wins:
152 * BAR[0,2] -> disabled, BAR[1] -> covers all DRAM banks
153 * WIN[0-3] -> DRAM bank[0-3]
154 */
155static void mvebu_pcie_setup_wins(struct mvebu_pcie_port *port)
156{
157 const struct mbus_dram_target_info *dram;
158 u32 size;
159 int i;
160
161 dram = mv_mbus_dram_info();
162
163 /* First, disable and clear BARs and windows. */
164 for (i = 1; i < 3; i++) {
165 mvebu_writel(port, 0, PCIE_BAR_CTRL_OFF(i));
166 mvebu_writel(port, 0, PCIE_BAR_LO_OFF(i));
167 mvebu_writel(port, 0, PCIE_BAR_HI_OFF(i));
168 }
169
170 for (i = 0; i < 5; i++) {
171 mvebu_writel(port, 0, PCIE_WIN04_CTRL_OFF(i));
172 mvebu_writel(port, 0, PCIE_WIN04_BASE_OFF(i));
173 mvebu_writel(port, 0, PCIE_WIN04_REMAP_OFF(i));
174 }
175
176 mvebu_writel(port, 0, PCIE_WIN5_CTRL_OFF);
177 mvebu_writel(port, 0, PCIE_WIN5_BASE_OFF);
178 mvebu_writel(port, 0, PCIE_WIN5_REMAP_OFF);
179
180 /* Setup windows for DDR banks. Count total DDR size on the fly. */
181 size = 0;
182 for (i = 0; i < dram->num_cs; i++) {
183 const struct mbus_dram_window *cs = dram->cs + i;
184
185 mvebu_writel(port, cs->base & 0xffff0000,
186 PCIE_WIN04_BASE_OFF(i));
187 mvebu_writel(port, 0, PCIE_WIN04_REMAP_OFF(i));
188 mvebu_writel(port,
189 ((cs->size - 1) & 0xffff0000) |
190 (cs->mbus_attr << 8) |
191 (dram->mbus_dram_target_id << 4) | 1,
192 PCIE_WIN04_CTRL_OFF(i));
193
194 size += cs->size;
195 }
196
197 /* Round up 'size' to the nearest power of two. */
198 if ((size & (size - 1)) != 0)
199 size = 1 << fls(size);
200
201 /* Setup BAR[1] to all DRAM banks. */
202 mvebu_writel(port, dram->cs[0].base, PCIE_BAR_LO_OFF(1));
203 mvebu_writel(port, 0, PCIE_BAR_HI_OFF(1));
204 mvebu_writel(port, ((size - 1) & 0xffff0000) | 1,
205 PCIE_BAR_CTRL_OFF(1));
206}
207
208static void mvebu_pcie_setup_hw(struct mvebu_pcie_port *port)
209{
210 u32 cmd, mask;
211
212 /* Point PCIe unit MBUS decode windows to DRAM space. */
213 mvebu_pcie_setup_wins(port);
214
215 /* Master + slave enable. */
216 cmd = mvebu_readl(port, PCIE_CMD_OFF);
217 cmd |= PCI_COMMAND_IO;
218 cmd |= PCI_COMMAND_MEMORY;
219 cmd |= PCI_COMMAND_MASTER;
220 mvebu_writel(port, cmd, PCIE_CMD_OFF);
221
222 /* Enable interrupt lines A-D. */
223 mask = mvebu_readl(port, PCIE_MASK_OFF);
224 mask |= PCIE_MASK_ENABLE_INTS;
225 mvebu_writel(port, mask, PCIE_MASK_OFF);
226}
227
228static int mvebu_pcie_hw_rd_conf(struct mvebu_pcie_port *port,
229 struct pci_bus *bus,
230 u32 devfn, int where, int size, u32 *val)
231{
232 void __iomem *conf_data = port->base + PCIE_CONF_DATA_OFF;
233
234 mvebu_writel(port, PCIE_CONF_ADDR(bus->number, devfn, where),
235 PCIE_CONF_ADDR_OFF);
236
237 switch (size) {
238 case 1:
239 *val = readb_relaxed(conf_data + (where & 3));
240 break;
241 case 2:
242 *val = readw_relaxed(conf_data + (where & 2));
243 break;
244 case 4:
245 *val = readl_relaxed(conf_data);
246 break;
247 }
248
249 return PCIBIOS_SUCCESSFUL;
250}
251
252static int mvebu_pcie_hw_wr_conf(struct mvebu_pcie_port *port,
253 struct pci_bus *bus,
254 u32 devfn, int where, int size, u32 val)
255{
256 void __iomem *conf_data = port->base + PCIE_CONF_DATA_OFF;
257
258 mvebu_writel(port, PCIE_CONF_ADDR(bus->number, devfn, where),
259 PCIE_CONF_ADDR_OFF);
260
261 switch (size) {
262 case 1:
263 writeb(val, conf_data + (where & 3));
264 break;
265 case 2:
266 writew(val, conf_data + (where & 2));
267 break;
268 case 4:
269 writel(val, conf_data);
270 break;
271 default:
272 return PCIBIOS_BAD_REGISTER_NUMBER;
273 }
274
275 return PCIBIOS_SUCCESSFUL;
276}
277
278/*
279 * Remove windows, starting from the largest ones to the smallest
280 * ones.
281 */
282static void mvebu_pcie_del_windows(struct mvebu_pcie_port *port,
283 phys_addr_t base, size_t size)
284{
285 while (size) {
286 size_t sz = 1 << (fls(size) - 1);
287
288 mvebu_mbus_del_window(base, sz);
289 base += sz;
290 size -= sz;
291 }
292}
293
294/*
295 * MBus windows can only have a power of two size, but PCI BARs do not
296 * have this constraint. Therefore, we have to split the PCI BAR into
297 * areas each having a power of two size. We start from the largest
298 * one (i.e highest order bit set in the size).
299 */
300static void mvebu_pcie_add_windows(struct mvebu_pcie_port *port,
301 unsigned int target, unsigned int attribute,
302 phys_addr_t base, size_t size,
303 phys_addr_t remap)
304{
305 size_t size_mapped = 0;
306
307 while (size) {
308 size_t sz = 1 << (fls(size) - 1);
309 int ret;
310
311 ret = mvebu_mbus_add_window_remap_by_id(target, attribute, base,
312 sz, remap);
313 if (ret) {
314 phys_addr_t end = base + sz - 1;
315
316 dev_err(&port->pcie->pdev->dev,
317 "Could not create MBus window at [mem %pa-%pa]: %d\n",
318 &base, &end, ret);
319 mvebu_pcie_del_windows(port, base - size_mapped,
320 size_mapped);
321 return;
322 }
323
324 size -= sz;
325 size_mapped += sz;
326 base += sz;
327 if (remap != MVEBU_MBUS_NO_REMAP)
328 remap += sz;
329 }
330}
331
332static void mvebu_pcie_set_window(struct mvebu_pcie_port *port,
333 unsigned int target, unsigned int attribute,
334 const struct mvebu_pcie_window *desired,
335 struct mvebu_pcie_window *cur)
336{
337 if (desired->base == cur->base && desired->remap == cur->remap &&
338 desired->size == cur->size)
339 return;
340
341 if (cur->size != 0) {
342 mvebu_pcie_del_windows(port, cur->base, cur->size);
343 cur->size = 0;
344 cur->base = 0;
345
346 /*
347 * If something tries to change the window while it is enabled
348 * the change will not be done atomically. That would be
349 * difficult to do in the general case.
350 */
351 }
352
353 if (desired->size == 0)
354 return;
355
356 mvebu_pcie_add_windows(port, target, attribute, desired->base,
357 desired->size, desired->remap);
358 *cur = *desired;
359}
360
361static void mvebu_pcie_handle_iobase_change(struct mvebu_pcie_port *port)
362{
363 struct mvebu_pcie_window desired = {};
364 struct pci_bridge_emul_conf *conf = &port->bridge.conf;
365
366 /* Are the new iobase/iolimit values invalid? */
367 if (conf->iolimit < conf->iobase ||
368 conf->iolimitupper < conf->iobaseupper ||
369 !(conf->command & PCI_COMMAND_IO)) {
370 mvebu_pcie_set_window(port, port->io_target, port->io_attr,
371 &desired, &port->iowin);
372 return;
373 }
374
375 if (!mvebu_has_ioport(port)) {
376 dev_WARN(&port->pcie->pdev->dev,
377 "Attempt to set IO when IO is disabled\n");
378 return;
379 }
380
381 /*
382 * We read the PCI-to-PCI bridge emulated registers, and
383 * calculate the base address and size of the address decoding
384 * window to setup, according to the PCI-to-PCI bridge
385 * specifications. iobase is the bus address, port->iowin_base
386 * is the CPU address.
387 */
388 desired.remap = ((conf->iobase & 0xF0) << 8) |
389 (conf->iobaseupper << 16);
390 desired.base = port->pcie->io.start + desired.remap;
391 desired.size = ((0xFFF | ((conf->iolimit & 0xF0) << 8) |
392 (conf->iolimitupper << 16)) -
393 desired.remap) +
394 1;
395
396 mvebu_pcie_set_window(port, port->io_target, port->io_attr, &desired,
397 &port->iowin);
398}
399
400static void mvebu_pcie_handle_membase_change(struct mvebu_pcie_port *port)
401{
402 struct mvebu_pcie_window desired = {.remap = MVEBU_MBUS_NO_REMAP};
403 struct pci_bridge_emul_conf *conf = &port->bridge.conf;
404
405 /* Are the new membase/memlimit values invalid? */
406 if (conf->memlimit < conf->membase ||
407 !(conf->command & PCI_COMMAND_MEMORY)) {
408 mvebu_pcie_set_window(port, port->mem_target, port->mem_attr,
409 &desired, &port->memwin);
410 return;
411 }
412
413 /*
414 * We read the PCI-to-PCI bridge emulated registers, and
415 * calculate the base address and size of the address decoding
416 * window to setup, according to the PCI-to-PCI bridge
417 * specifications.
418 */
419 desired.base = ((conf->membase & 0xFFF0) << 16);
420 desired.size = (((conf->memlimit & 0xFFF0) << 16) | 0xFFFFF) -
421 desired.base + 1;
422
423 mvebu_pcie_set_window(port, port->mem_target, port->mem_attr, &desired,
424 &port->memwin);
425}
426
427static pci_bridge_emul_read_status_t
428mvebu_pci_bridge_emul_pcie_conf_read(struct pci_bridge_emul *bridge,
429 int reg, u32 *value)
430{
431 struct mvebu_pcie_port *port = bridge->data;
432
433 switch (reg) {
434 case PCI_EXP_DEVCAP:
435 *value = mvebu_readl(port, PCIE_CAP_PCIEXP + PCI_EXP_DEVCAP);
436 break;
437
438 case PCI_EXP_DEVCTL:
439 *value = mvebu_readl(port, PCIE_CAP_PCIEXP + PCI_EXP_DEVCTL) &
440 ~(PCI_EXP_DEVCTL_URRE | PCI_EXP_DEVCTL_FERE |
441 PCI_EXP_DEVCTL_NFERE | PCI_EXP_DEVCTL_CERE);
442 break;
443
444 case PCI_EXP_LNKCAP:
445 /*
446 * PCIe requires the clock power management capability to be
447 * hard-wired to zero for downstream ports
448 */
449 *value = mvebu_readl(port, PCIE_CAP_PCIEXP + PCI_EXP_LNKCAP) &
450 ~PCI_EXP_LNKCAP_CLKPM;
451 break;
452
453 case PCI_EXP_LNKCTL:
454 *value = mvebu_readl(port, PCIE_CAP_PCIEXP + PCI_EXP_LNKCTL);
455 break;
456
457 case PCI_EXP_SLTCTL:
458 *value = PCI_EXP_SLTSTA_PDS << 16;
459 break;
460
461 case PCI_EXP_RTSTA:
462 *value = mvebu_readl(port, PCIE_RC_RTSTA);
463 break;
464
465 default:
466 return PCI_BRIDGE_EMUL_NOT_HANDLED;
467 }
468
469 return PCI_BRIDGE_EMUL_HANDLED;
470}
471
472static void
473mvebu_pci_bridge_emul_base_conf_write(struct pci_bridge_emul *bridge,
474 int reg, u32 old, u32 new, u32 mask)
475{
476 struct mvebu_pcie_port *port = bridge->data;
477 struct pci_bridge_emul_conf *conf = &bridge->conf;
478
479 switch (reg) {
480 case PCI_COMMAND:
481 {
482 if (!mvebu_has_ioport(port))
483 conf->command &= ~PCI_COMMAND_IO;
484
485 if ((old ^ new) & PCI_COMMAND_IO)
486 mvebu_pcie_handle_iobase_change(port);
487 if ((old ^ new) & PCI_COMMAND_MEMORY)
488 mvebu_pcie_handle_membase_change(port);
489
490 break;
491 }
492
493 case PCI_IO_BASE:
494 /*
495 * We keep bit 1 set, it is a read-only bit that
496 * indicates we support 32 bits addressing for the
497 * I/O
498 */
499 conf->iobase |= PCI_IO_RANGE_TYPE_32;
500 conf->iolimit |= PCI_IO_RANGE_TYPE_32;
501 mvebu_pcie_handle_iobase_change(port);
502 break;
503
504 case PCI_MEMORY_BASE:
505 mvebu_pcie_handle_membase_change(port);
506 break;
507
508 case PCI_IO_BASE_UPPER16:
509 mvebu_pcie_handle_iobase_change(port);
510 break;
511
512 case PCI_PRIMARY_BUS:
513 mvebu_pcie_set_local_bus_nr(port, conf->secondary_bus);
514 break;
515
516 default:
517 break;
518 }
519}
520
521static void
522mvebu_pci_bridge_emul_pcie_conf_write(struct pci_bridge_emul *bridge,
523 int reg, u32 old, u32 new, u32 mask)
524{
525 struct mvebu_pcie_port *port = bridge->data;
526
527 switch (reg) {
528 case PCI_EXP_DEVCTL:
529 /*
530 * Armada370 data says these bits must always
531 * be zero when in root complex mode.
532 */
533 new &= ~(PCI_EXP_DEVCTL_URRE | PCI_EXP_DEVCTL_FERE |
534 PCI_EXP_DEVCTL_NFERE | PCI_EXP_DEVCTL_CERE);
535
536 mvebu_writel(port, new, PCIE_CAP_PCIEXP + PCI_EXP_DEVCTL);
537 break;
538
539 case PCI_EXP_LNKCTL:
540 /*
541 * If we don't support CLKREQ, we must ensure that the
542 * CLKREQ enable bit always reads zero. Since we haven't
543 * had this capability, and it's dependent on board wiring,
544 * disable it for the time being.
545 */
546 new &= ~PCI_EXP_LNKCTL_CLKREQ_EN;
547
548 mvebu_writel(port, new, PCIE_CAP_PCIEXP + PCI_EXP_LNKCTL);
549 break;
550
551 case PCI_EXP_RTSTA:
552 mvebu_writel(port, new, PCIE_RC_RTSTA);
553 break;
554 }
555}
556
557struct pci_bridge_emul_ops mvebu_pci_bridge_emul_ops = {
558 .write_base = mvebu_pci_bridge_emul_base_conf_write,
559 .read_pcie = mvebu_pci_bridge_emul_pcie_conf_read,
560 .write_pcie = mvebu_pci_bridge_emul_pcie_conf_write,
561};
562
563/*
564 * Initialize the configuration space of the PCI-to-PCI bridge
565 * associated with the given PCIe interface.
566 */
567static void mvebu_pci_bridge_emul_init(struct mvebu_pcie_port *port)
568{
569 struct pci_bridge_emul *bridge = &port->bridge;
570
571 bridge->conf.vendor = PCI_VENDOR_ID_MARVELL;
572 bridge->conf.device = mvebu_readl(port, PCIE_DEV_ID_OFF) >> 16;
573 bridge->conf.class_revision =
574 mvebu_readl(port, PCIE_DEV_REV_OFF) & 0xff;
575
576 if (mvebu_has_ioport(port)) {
577 /* We support 32 bits I/O addressing */
578 bridge->conf.iobase = PCI_IO_RANGE_TYPE_32;
579 bridge->conf.iolimit = PCI_IO_RANGE_TYPE_32;
580 }
581
582 bridge->has_pcie = true;
583 bridge->data = port;
584 bridge->ops = &mvebu_pci_bridge_emul_ops;
585
586 pci_bridge_emul_init(bridge, PCI_BRIDGE_EMUL_NO_PREFETCHABLE_BAR);
587}
588
589static inline struct mvebu_pcie *sys_to_pcie(struct pci_sys_data *sys)
590{
591 return sys->private_data;
592}
593
594static struct mvebu_pcie_port *mvebu_pcie_find_port(struct mvebu_pcie *pcie,
595 struct pci_bus *bus,
596 int devfn)
597{
598 int i;
599
600 for (i = 0; i < pcie->nports; i++) {
601 struct mvebu_pcie_port *port = &pcie->ports[i];
602
603 if (bus->number == 0 && port->devfn == devfn)
604 return port;
605 if (bus->number != 0 &&
606 bus->number >= port->bridge.conf.secondary_bus &&
607 bus->number <= port->bridge.conf.subordinate_bus)
608 return port;
609 }
610
611 return NULL;
612}
613
614/* PCI configuration space write function */
615static int mvebu_pcie_wr_conf(struct pci_bus *bus, u32 devfn,
616 int where, int size, u32 val)
617{
618 struct mvebu_pcie *pcie = bus->sysdata;
619 struct mvebu_pcie_port *port;
620 int ret;
621
622 port = mvebu_pcie_find_port(pcie, bus, devfn);
623 if (!port)
624 return PCIBIOS_DEVICE_NOT_FOUND;
625
626 /* Access the emulated PCI-to-PCI bridge */
627 if (bus->number == 0)
628 return pci_bridge_emul_conf_write(&port->bridge, where,
629 size, val);
630
631 if (!mvebu_pcie_link_up(port))
632 return PCIBIOS_DEVICE_NOT_FOUND;
633
634 /* Access the real PCIe interface */
635 ret = mvebu_pcie_hw_wr_conf(port, bus, devfn,
636 where, size, val);
637
638 return ret;
639}
640
641/* PCI configuration space read function */
642static int mvebu_pcie_rd_conf(struct pci_bus *bus, u32 devfn, int where,
643 int size, u32 *val)
644{
645 struct mvebu_pcie *pcie = bus->sysdata;
646 struct mvebu_pcie_port *port;
647 int ret;
648
649 port = mvebu_pcie_find_port(pcie, bus, devfn);
650 if (!port) {
651 *val = 0xffffffff;
652 return PCIBIOS_DEVICE_NOT_FOUND;
653 }
654
655 /* Access the emulated PCI-to-PCI bridge */
656 if (bus->number == 0)
657 return pci_bridge_emul_conf_read(&port->bridge, where,
658 size, val);
659
660 if (!mvebu_pcie_link_up(port)) {
661 *val = 0xffffffff;
662 return PCIBIOS_DEVICE_NOT_FOUND;
663 }
664
665 /* Access the real PCIe interface */
666 ret = mvebu_pcie_hw_rd_conf(port, bus, devfn,
667 where, size, val);
668
669 return ret;
670}
671
672static struct pci_ops mvebu_pcie_ops = {
673 .read = mvebu_pcie_rd_conf,
674 .write = mvebu_pcie_wr_conf,
675};
676
677static resource_size_t mvebu_pcie_align_resource(struct pci_dev *dev,
678 const struct resource *res,
679 resource_size_t start,
680 resource_size_t size,
681 resource_size_t align)
682{
683 if (dev->bus->number != 0)
684 return start;
685
686 /*
687 * On the PCI-to-PCI bridge side, the I/O windows must have at
688 * least a 64 KB size and the memory windows must have at
689 * least a 1 MB size. Moreover, MBus windows need to have a
690 * base address aligned on their size, and their size must be
691 * a power of two. This means that if the BAR doesn't have a
692 * power of two size, several MBus windows will actually be
693 * created. We need to ensure that the biggest MBus window
694 * (which will be the first one) is aligned on its size, which
695 * explains the rounddown_pow_of_two() being done here.
696 */
697 if (res->flags & IORESOURCE_IO)
698 return round_up(start, max_t(resource_size_t, SZ_64K,
699 rounddown_pow_of_two(size)));
700 else if (res->flags & IORESOURCE_MEM)
701 return round_up(start, max_t(resource_size_t, SZ_1M,
702 rounddown_pow_of_two(size)));
703 else
704 return start;
705}
706
707static void __iomem *mvebu_pcie_map_registers(struct platform_device *pdev,
708 struct device_node *np,
709 struct mvebu_pcie_port *port)
710{
711 struct resource regs;
712 int ret = 0;
713
714 ret = of_address_to_resource(np, 0, ®s);
715 if (ret)
716 return ERR_PTR(ret);
717
718 return devm_ioremap_resource(&pdev->dev, ®s);
719}
720
721#define DT_FLAGS_TO_TYPE(flags) (((flags) >> 24) & 0x03)
722#define DT_TYPE_IO 0x1
723#define DT_TYPE_MEM32 0x2
724#define DT_CPUADDR_TO_TARGET(cpuaddr) (((cpuaddr) >> 56) & 0xFF)
725#define DT_CPUADDR_TO_ATTR(cpuaddr) (((cpuaddr) >> 48) & 0xFF)
726
727static int mvebu_get_tgt_attr(struct device_node *np, int devfn,
728 unsigned long type,
729 unsigned int *tgt,
730 unsigned int *attr)
731{
732 const int na = 3, ns = 2;
733 const __be32 *range;
734 int rlen, nranges, rangesz, pna, i;
735
736 *tgt = -1;
737 *attr = -1;
738
739 range = of_get_property(np, "ranges", &rlen);
740 if (!range)
741 return -EINVAL;
742
743 pna = of_n_addr_cells(np);
744 rangesz = pna + na + ns;
745 nranges = rlen / sizeof(__be32) / rangesz;
746
747 for (i = 0; i < nranges; i++, range += rangesz) {
748 u32 flags = of_read_number(range, 1);
749 u32 slot = of_read_number(range + 1, 1);
750 u64 cpuaddr = of_read_number(range + na, pna);
751 unsigned long rtype;
752
753 if (DT_FLAGS_TO_TYPE(flags) == DT_TYPE_IO)
754 rtype = IORESOURCE_IO;
755 else if (DT_FLAGS_TO_TYPE(flags) == DT_TYPE_MEM32)
756 rtype = IORESOURCE_MEM;
757 else
758 continue;
759
760 if (slot == PCI_SLOT(devfn) && type == rtype) {
761 *tgt = DT_CPUADDR_TO_TARGET(cpuaddr);
762 *attr = DT_CPUADDR_TO_ATTR(cpuaddr);
763 return 0;
764 }
765 }
766
767 return -ENOENT;
768}
769
770#ifdef CONFIG_PM_SLEEP
771static int mvebu_pcie_suspend(struct device *dev)
772{
773 struct mvebu_pcie *pcie;
774 int i;
775
776 pcie = dev_get_drvdata(dev);
777 for (i = 0; i < pcie->nports; i++) {
778 struct mvebu_pcie_port *port = pcie->ports + i;
779 port->saved_pcie_stat = mvebu_readl(port, PCIE_STAT_OFF);
780 }
781
782 return 0;
783}
784
785static int mvebu_pcie_resume(struct device *dev)
786{
787 struct mvebu_pcie *pcie;
788 int i;
789
790 pcie = dev_get_drvdata(dev);
791 for (i = 0; i < pcie->nports; i++) {
792 struct mvebu_pcie_port *port = pcie->ports + i;
793 mvebu_writel(port, port->saved_pcie_stat, PCIE_STAT_OFF);
794 mvebu_pcie_setup_hw(port);
795 }
796
797 return 0;
798}
799#endif
800
801static void mvebu_pcie_port_clk_put(void *data)
802{
803 struct mvebu_pcie_port *port = data;
804
805 clk_put(port->clk);
806}
807
808static int mvebu_pcie_parse_port(struct mvebu_pcie *pcie,
809 struct mvebu_pcie_port *port, struct device_node *child)
810{
811 struct device *dev = &pcie->pdev->dev;
812 enum of_gpio_flags flags;
813 int reset_gpio, ret;
814
815 port->pcie = pcie;
816
817 if (of_property_read_u32(child, "marvell,pcie-port", &port->port)) {
818 dev_warn(dev, "ignoring %pOF, missing pcie-port property\n",
819 child);
820 goto skip;
821 }
822
823 if (of_property_read_u32(child, "marvell,pcie-lane", &port->lane))
824 port->lane = 0;
825
826 port->name = devm_kasprintf(dev, GFP_KERNEL, "pcie%d.%d", port->port,
827 port->lane);
828 if (!port->name) {
829 ret = -ENOMEM;
830 goto err;
831 }
832
833 port->devfn = of_pci_get_devfn(child);
834 if (port->devfn < 0)
835 goto skip;
836
837 ret = mvebu_get_tgt_attr(dev->of_node, port->devfn, IORESOURCE_MEM,
838 &port->mem_target, &port->mem_attr);
839 if (ret < 0) {
840 dev_err(dev, "%s: cannot get tgt/attr for mem window\n",
841 port->name);
842 goto skip;
843 }
844
845 if (resource_size(&pcie->io) != 0) {
846 mvebu_get_tgt_attr(dev->of_node, port->devfn, IORESOURCE_IO,
847 &port->io_target, &port->io_attr);
848 } else {
849 port->io_target = -1;
850 port->io_attr = -1;
851 }
852
853 reset_gpio = of_get_named_gpio_flags(child, "reset-gpios", 0, &flags);
854 if (reset_gpio == -EPROBE_DEFER) {
855 ret = reset_gpio;
856 goto err;
857 }
858
859 if (gpio_is_valid(reset_gpio)) {
860 unsigned long gpio_flags;
861
862 port->reset_name = devm_kasprintf(dev, GFP_KERNEL, "%s-reset",
863 port->name);
864 if (!port->reset_name) {
865 ret = -ENOMEM;
866 goto err;
867 }
868
869 if (flags & OF_GPIO_ACTIVE_LOW) {
870 dev_info(dev, "%pOF: reset gpio is active low\n",
871 child);
872 gpio_flags = GPIOF_ACTIVE_LOW |
873 GPIOF_OUT_INIT_LOW;
874 } else {
875 gpio_flags = GPIOF_OUT_INIT_HIGH;
876 }
877
878 ret = devm_gpio_request_one(dev, reset_gpio, gpio_flags,
879 port->reset_name);
880 if (ret) {
881 if (ret == -EPROBE_DEFER)
882 goto err;
883 goto skip;
884 }
885
886 port->reset_gpio = gpio_to_desc(reset_gpio);
887 }
888
889 port->clk = of_clk_get_by_name(child, NULL);
890 if (IS_ERR(port->clk)) {
891 dev_err(dev, "%s: cannot get clock\n", port->name);
892 goto skip;
893 }
894
895 ret = devm_add_action(dev, mvebu_pcie_port_clk_put, port);
896 if (ret < 0) {
897 clk_put(port->clk);
898 goto err;
899 }
900
901 return 1;
902
903skip:
904 ret = 0;
905
906 /* In the case of skipping, we need to free these */
907 devm_kfree(dev, port->reset_name);
908 port->reset_name = NULL;
909 devm_kfree(dev, port->name);
910 port->name = NULL;
911
912err:
913 return ret;
914}
915
916/*
917 * Power up a PCIe port. PCIe requires the refclk to be stable for 100µs
918 * prior to releasing PERST. See table 2-4 in section 2.6.2 AC Specifications
919 * of the PCI Express Card Electromechanical Specification, 1.1.
920 */
921static int mvebu_pcie_powerup(struct mvebu_pcie_port *port)
922{
923 int ret;
924
925 ret = clk_prepare_enable(port->clk);
926 if (ret < 0)
927 return ret;
928
929 if (port->reset_gpio) {
930 u32 reset_udelay = PCI_PM_D3COLD_WAIT * 1000;
931
932 of_property_read_u32(port->dn, "reset-delay-us",
933 &reset_udelay);
934
935 udelay(100);
936
937 gpiod_set_value_cansleep(port->reset_gpio, 0);
938 msleep(reset_udelay / 1000);
939 }
940
941 return 0;
942}
943
944/*
945 * Power down a PCIe port. Strictly, PCIe requires us to place the card
946 * in D3hot state before asserting PERST#.
947 */
948static void mvebu_pcie_powerdown(struct mvebu_pcie_port *port)
949{
950 gpiod_set_value_cansleep(port->reset_gpio, 1);
951
952 clk_disable_unprepare(port->clk);
953}
954
955/*
956 * We can't use devm_of_pci_get_host_bridge_resources() because we
957 * need to parse our special DT properties encoding the MEM and IO
958 * apertures.
959 */
960static int mvebu_pcie_parse_request_resources(struct mvebu_pcie *pcie)
961{
962 struct device *dev = &pcie->pdev->dev;
963 struct device_node *np = dev->of_node;
964 int ret;
965
966 INIT_LIST_HEAD(&pcie->resources);
967
968 /* Get the bus range */
969 ret = of_pci_parse_bus_range(np, &pcie->busn);
970 if (ret) {
971 dev_err(dev, "failed to parse bus-range property: %d\n", ret);
972 return ret;
973 }
974 pci_add_resource(&pcie->resources, &pcie->busn);
975
976 /* Get the PCIe memory aperture */
977 mvebu_mbus_get_pcie_mem_aperture(&pcie->mem);
978 if (resource_size(&pcie->mem) == 0) {
979 dev_err(dev, "invalid memory aperture size\n");
980 return -EINVAL;
981 }
982
983 pcie->mem.name = "PCI MEM";
984 pci_add_resource(&pcie->resources, &pcie->mem);
985
986 /* Get the PCIe IO aperture */
987 mvebu_mbus_get_pcie_io_aperture(&pcie->io);
988
989 if (resource_size(&pcie->io) != 0) {
990 pcie->realio.flags = pcie->io.flags;
991 pcie->realio.start = PCIBIOS_MIN_IO;
992 pcie->realio.end = min_t(resource_size_t,
993 IO_SPACE_LIMIT - SZ_64K,
994 resource_size(&pcie->io) - 1);
995 pcie->realio.name = "PCI I/O";
996
997 pci_add_resource(&pcie->resources, &pcie->realio);
998 }
999
1000 return devm_request_pci_bus_resources(dev, &pcie->resources);
1001}
1002
1003/*
1004 * This is a copy of pci_host_probe(), except that it does the I/O
1005 * remap as the last step, once we are sure we won't fail.
1006 *
1007 * It should be removed once the I/O remap error handling issue has
1008 * been sorted out.
1009 */
1010static int mvebu_pci_host_probe(struct pci_host_bridge *bridge)
1011{
1012 struct mvebu_pcie *pcie;
1013 struct pci_bus *bus, *child;
1014 int ret;
1015
1016 ret = pci_scan_root_bus_bridge(bridge);
1017 if (ret < 0) {
1018 dev_err(bridge->dev.parent, "Scanning root bridge failed");
1019 return ret;
1020 }
1021
1022 pcie = pci_host_bridge_priv(bridge);
1023 if (resource_size(&pcie->io) != 0) {
1024 unsigned int i;
1025
1026 for (i = 0; i < resource_size(&pcie->realio); i += SZ_64K)
1027 pci_ioremap_io(i, pcie->io.start + i);
1028 }
1029
1030 bus = bridge->bus;
1031
1032 /*
1033 * We insert PCI resources into the iomem_resource and
1034 * ioport_resource trees in either pci_bus_claim_resources()
1035 * or pci_bus_assign_resources().
1036 */
1037 if (pci_has_flag(PCI_PROBE_ONLY)) {
1038 pci_bus_claim_resources(bus);
1039 } else {
1040 pci_bus_size_bridges(bus);
1041 pci_bus_assign_resources(bus);
1042
1043 list_for_each_entry(child, &bus->children, node)
1044 pcie_bus_configure_settings(child);
1045 }
1046
1047 pci_bus_add_devices(bus);
1048 return 0;
1049}
1050
1051static int mvebu_pcie_probe(struct platform_device *pdev)
1052{
1053 struct device *dev = &pdev->dev;
1054 struct mvebu_pcie *pcie;
1055 struct pci_host_bridge *bridge;
1056 struct device_node *np = dev->of_node;
1057 struct device_node *child;
1058 int num, i, ret;
1059
1060 bridge = devm_pci_alloc_host_bridge(dev, sizeof(struct mvebu_pcie));
1061 if (!bridge)
1062 return -ENOMEM;
1063
1064 pcie = pci_host_bridge_priv(bridge);
1065 pcie->pdev = pdev;
1066 platform_set_drvdata(pdev, pcie);
1067
1068 ret = mvebu_pcie_parse_request_resources(pcie);
1069 if (ret)
1070 return ret;
1071
1072 num = of_get_available_child_count(np);
1073
1074 pcie->ports = devm_kcalloc(dev, num, sizeof(*pcie->ports), GFP_KERNEL);
1075 if (!pcie->ports)
1076 return -ENOMEM;
1077
1078 i = 0;
1079 for_each_available_child_of_node(np, child) {
1080 struct mvebu_pcie_port *port = &pcie->ports[i];
1081
1082 ret = mvebu_pcie_parse_port(pcie, port, child);
1083 if (ret < 0) {
1084 of_node_put(child);
1085 return ret;
1086 } else if (ret == 0) {
1087 continue;
1088 }
1089
1090 port->dn = child;
1091 i++;
1092 }
1093 pcie->nports = i;
1094
1095 for (i = 0; i < pcie->nports; i++) {
1096 struct mvebu_pcie_port *port = &pcie->ports[i];
1097
1098 child = port->dn;
1099 if (!child)
1100 continue;
1101
1102 ret = mvebu_pcie_powerup(port);
1103 if (ret < 0)
1104 continue;
1105
1106 port->base = mvebu_pcie_map_registers(pdev, child, port);
1107 if (IS_ERR(port->base)) {
1108 dev_err(dev, "%s: cannot map registers\n", port->name);
1109 port->base = NULL;
1110 mvebu_pcie_powerdown(port);
1111 continue;
1112 }
1113
1114 mvebu_pcie_setup_hw(port);
1115 mvebu_pcie_set_local_dev_nr(port, 1);
1116 mvebu_pci_bridge_emul_init(port);
1117 }
1118
1119 pcie->nports = i;
1120
1121 list_splice_init(&pcie->resources, &bridge->windows);
1122 bridge->dev.parent = dev;
1123 bridge->sysdata = pcie;
1124 bridge->busnr = 0;
1125 bridge->ops = &mvebu_pcie_ops;
1126 bridge->map_irq = of_irq_parse_and_map_pci;
1127 bridge->swizzle_irq = pci_common_swizzle;
1128 bridge->align_resource = mvebu_pcie_align_resource;
1129 bridge->msi = pcie->msi;
1130
1131 return mvebu_pci_host_probe(bridge);
1132}
1133
1134static const struct of_device_id mvebu_pcie_of_match_table[] = {
1135 { .compatible = "marvell,armada-xp-pcie", },
1136 { .compatible = "marvell,armada-370-pcie", },
1137 { .compatible = "marvell,dove-pcie", },
1138 { .compatible = "marvell,kirkwood-pcie", },
1139 {},
1140};
1141
1142static const struct dev_pm_ops mvebu_pcie_pm_ops = {
1143 SET_NOIRQ_SYSTEM_SLEEP_PM_OPS(mvebu_pcie_suspend, mvebu_pcie_resume)
1144};
1145
1146static struct platform_driver mvebu_pcie_driver = {
1147 .driver = {
1148 .name = "mvebu-pcie",
1149 .of_match_table = mvebu_pcie_of_match_table,
1150 /* driver unloading/unbinding currently not supported */
1151 .suppress_bind_attrs = true,
1152 .pm = &mvebu_pcie_pm_ops,
1153 },
1154 .probe = mvebu_pcie_probe,
1155};
1156builtin_platform_driver(mvebu_pcie_driver);