1 /*
2 * PCI / PCI-X / PCI-Express support for 4xx parts
3 *
4 * Copyright 2007 Ben. Herrenschmidt <benh@kernel.crashing.org>, IBM Corp.
5 *
6 * Most PCI Express code is coming from Stefan Roese implementation for
7 * arch/ppc in the Denx tree, slightly reworked by me.
8 *
9 * Copyright 2007 DENX Software Engineering, Stefan Roese <sr@denx.de>
10 *
11 * Some of that comes itself from a previous implementation for 440SPE only
12 * by Roland Dreier:
13 *
14 * Copyright (c) 2005 Cisco Systems. All rights reserved.
15 * Roland Dreier <rolandd@cisco.com>
16 *
17 */
18
19 #undef DEBUG
20
21 #include <linux/kernel.h>
22 #include <linux/pci.h>
23 #include <linux/init.h>
24 #include <linux/of.h>
25 #include <linux/delay.h>
26 #include <linux/slab.h>
27
28 #include <asm/io.h>
29 #include <asm/pci-bridge.h>
30 #include <asm/machdep.h>
31 #include <asm/dcr.h>
32 #include <asm/dcr-regs.h>
33 #include <mm/mmu_decl.h>
34
35 #include "pci.h"
36
37 static int dma_offset_set;
38
39 #define U64_TO_U32_LOW(val) ((u32)((val) & 0x00000000ffffffffULL))
40 #define U64_TO_U32_HIGH(val) ((u32)((val) >> 32))
41
42 #define RES_TO_U32_LOW(val) \
43 ((sizeof(resource_size_t) > sizeof(u32)) ? U64_TO_U32_LOW(val) : (val))
44 #define RES_TO_U32_HIGH(val) \
45 ((sizeof(resource_size_t) > sizeof(u32)) ? U64_TO_U32_HIGH(val) : (0))
46
ppc440spe_revA(void)47 static inline int ppc440spe_revA(void)
48 {
49 /* Catch both 440SPe variants, with and without RAID6 support */
50 if ((mfspr(SPRN_PVR) & 0xffefffff) == 0x53421890)
51 return 1;
52 else
53 return 0;
54 }
55
fixup_ppc4xx_pci_bridge(struct pci_dev * dev)56 static void fixup_ppc4xx_pci_bridge(struct pci_dev *dev)
57 {
58 struct pci_controller *hose;
59 int i;
60
61 if (dev->devfn != 0 || dev->bus->self != NULL)
62 return;
63
64 hose = pci_bus_to_host(dev->bus);
65 if (hose == NULL)
66 return;
67
68 if (!of_device_is_compatible(hose->dn, "ibm,plb-pciex") &&
69 !of_device_is_compatible(hose->dn, "ibm,plb-pcix") &&
70 !of_device_is_compatible(hose->dn, "ibm,plb-pci"))
71 return;
72
73 if (of_device_is_compatible(hose->dn, "ibm,plb440epx-pci") ||
74 of_device_is_compatible(hose->dn, "ibm,plb440grx-pci")) {
75 hose->indirect_type |= PPC_INDIRECT_TYPE_BROKEN_MRM;
76 }
77
78 /* Hide the PCI host BARs from the kernel as their content doesn't
79 * fit well in the resource management
80 */
81 for (i = 0; i < DEVICE_COUNT_RESOURCE; i++) {
82 dev->resource[i].start = dev->resource[i].end = 0;
83 dev->resource[i].flags = 0;
84 }
85
86 printk(KERN_INFO "PCI: Hiding 4xx host bridge resources %s\n",
87 pci_name(dev));
88 }
89 DECLARE_PCI_FIXUP_HEADER(PCI_ANY_ID, PCI_ANY_ID, fixup_ppc4xx_pci_bridge);
90
ppc4xx_parse_dma_ranges(struct pci_controller * hose,void __iomem * reg,struct resource * res)91 static int __init ppc4xx_parse_dma_ranges(struct pci_controller *hose,
92 void __iomem *reg,
93 struct resource *res)
94 {
95 u64 size;
96 const u32 *ranges;
97 int rlen;
98 int pna = of_n_addr_cells(hose->dn);
99 int np = pna + 5;
100
101 /* Default */
102 res->start = 0;
103 size = 0x80000000;
104 res->end = size - 1;
105 res->flags = IORESOURCE_MEM | IORESOURCE_PREFETCH;
106
107 /* Get dma-ranges property */
108 ranges = of_get_property(hose->dn, "dma-ranges", &rlen);
109 if (ranges == NULL)
110 goto out;
111
112 /* Walk it */
113 while ((rlen -= np * 4) >= 0) {
114 u32 pci_space = ranges[0];
115 u64 pci_addr = of_read_number(ranges + 1, 2);
116 u64 cpu_addr = of_translate_dma_address(hose->dn, ranges + 3);
117 size = of_read_number(ranges + pna + 3, 2);
118 ranges += np;
119 if (cpu_addr == OF_BAD_ADDR || size == 0)
120 continue;
121
122 /* We only care about memory */
123 if ((pci_space & 0x03000000) != 0x02000000)
124 continue;
125
126 /* We currently only support memory at 0, and pci_addr
127 * within 32 bits space
128 */
129 if (cpu_addr != 0 || pci_addr > 0xffffffff) {
130 printk(KERN_WARNING "%pOF: Ignored unsupported dma range"
131 " 0x%016llx...0x%016llx -> 0x%016llx\n",
132 hose->dn,
133 pci_addr, pci_addr + size - 1, cpu_addr);
134 continue;
135 }
136
137 /* Check if not prefetchable */
138 if (!(pci_space & 0x40000000))
139 res->flags &= ~IORESOURCE_PREFETCH;
140
141
142 /* Use that */
143 res->start = pci_addr;
144 /* Beware of 32 bits resources */
145 if (sizeof(resource_size_t) == sizeof(u32) &&
146 (pci_addr + size) > 0x100000000ull)
147 res->end = 0xffffffff;
148 else
149 res->end = res->start + size - 1;
150 break;
151 }
152
153 /* We only support one global DMA offset */
154 if (dma_offset_set && pci_dram_offset != res->start) {
155 printk(KERN_ERR "%pOF: dma-ranges(s) mismatch\n", hose->dn);
156 return -ENXIO;
157 }
158
159 /* Check that we can fit all of memory as we don't support
160 * DMA bounce buffers
161 */
162 if (size < total_memory) {
163 printk(KERN_ERR "%pOF: dma-ranges too small "
164 "(size=%llx total_memory=%llx)\n",
165 hose->dn, size, (u64)total_memory);
166 return -ENXIO;
167 }
168
169 /* Check we are a power of 2 size and that base is a multiple of size*/
170 if ((size & (size - 1)) != 0 ||
171 (res->start & (size - 1)) != 0) {
172 printk(KERN_ERR "%pOF: dma-ranges unaligned\n", hose->dn);
173 return -ENXIO;
174 }
175
176 /* Check that we are fully contained within 32 bits space if we are not
177 * running on a 460sx or 476fpe which have 64 bit bus addresses.
178 */
179 if (res->end > 0xffffffff &&
180 !(of_device_is_compatible(hose->dn, "ibm,plb-pciex-460sx")
181 || of_device_is_compatible(hose->dn, "ibm,plb-pciex-476fpe"))) {
182 printk(KERN_ERR "%pOF: dma-ranges outside of 32 bits space\n",
183 hose->dn);
184 return -ENXIO;
185 }
186 out:
187 dma_offset_set = 1;
188 pci_dram_offset = res->start;
189 hose->dma_window_base_cur = res->start;
190 hose->dma_window_size = resource_size(res);
191
192 printk(KERN_INFO "4xx PCI DMA offset set to 0x%08lx\n",
193 pci_dram_offset);
194 printk(KERN_INFO "4xx PCI DMA window base to 0x%016llx\n",
195 (unsigned long long)hose->dma_window_base_cur);
196 printk(KERN_INFO "DMA window size 0x%016llx\n",
197 (unsigned long long)hose->dma_window_size);
198 return 0;
199 }
200
201 /*
202 * 4xx PCI 2.x part
203 */
204
ppc4xx_setup_one_pci_PMM(struct pci_controller * hose,void __iomem * reg,u64 plb_addr,u64 pci_addr,u64 size,unsigned int flags,int index)205 static int __init ppc4xx_setup_one_pci_PMM(struct pci_controller *hose,
206 void __iomem *reg,
207 u64 plb_addr,
208 u64 pci_addr,
209 u64 size,
210 unsigned int flags,
211 int index)
212 {
213 u32 ma, pcila, pciha;
214
215 /* Hack warning ! The "old" PCI 2.x cell only let us configure the low
216 * 32-bit of incoming PLB addresses. The top 4 bits of the 36-bit
217 * address are actually hard wired to a value that appears to depend
218 * on the specific SoC. For example, it's 0 on 440EP and 1 on 440EPx.
219 *
220 * The trick here is we just crop those top bits and ignore them when
221 * programming the chip. That means the device-tree has to be right
222 * for the specific part used (we don't print a warning if it's wrong
223 * but on the other hand, you'll crash quickly enough), but at least
224 * this code should work whatever the hard coded value is
225 */
226 plb_addr &= 0xffffffffull;
227
228 /* Note: Due to the above hack, the test below doesn't actually test
229 * if you address is above 4G, but it tests that address and
230 * (address + size) are both contained in the same 4G
231 */
232 if ((plb_addr + size) > 0xffffffffull || !is_power_of_2(size) ||
233 size < 0x1000 || (plb_addr & (size - 1)) != 0) {
234 printk(KERN_WARNING "%pOF: Resource out of range\n", hose->dn);
235 return -1;
236 }
237 ma = (0xffffffffu << ilog2(size)) | 1;
238 if (flags & IORESOURCE_PREFETCH)
239 ma |= 2;
240
241 pciha = RES_TO_U32_HIGH(pci_addr);
242 pcila = RES_TO_U32_LOW(pci_addr);
243
244 writel(plb_addr, reg + PCIL0_PMM0LA + (0x10 * index));
245 writel(pcila, reg + PCIL0_PMM0PCILA + (0x10 * index));
246 writel(pciha, reg + PCIL0_PMM0PCIHA + (0x10 * index));
247 writel(ma, reg + PCIL0_PMM0MA + (0x10 * index));
248
249 return 0;
250 }
251
ppc4xx_configure_pci_PMMs(struct pci_controller * hose,void __iomem * reg)252 static void __init ppc4xx_configure_pci_PMMs(struct pci_controller *hose,
253 void __iomem *reg)
254 {
255 int i, j, found_isa_hole = 0;
256
257 /* Setup outbound memory windows */
258 for (i = j = 0; i < 3; i++) {
259 struct resource *res = &hose->mem_resources[i];
260 resource_size_t offset = hose->mem_offset[i];
261
262 /* we only care about memory windows */
263 if (!(res->flags & IORESOURCE_MEM))
264 continue;
265 if (j > 2) {
266 printk(KERN_WARNING "%pOF: Too many ranges\n", hose->dn);
267 break;
268 }
269
270 /* Configure the resource */
271 if (ppc4xx_setup_one_pci_PMM(hose, reg,
272 res->start,
273 res->start - offset,
274 resource_size(res),
275 res->flags,
276 j) == 0) {
277 j++;
278
279 /* If the resource PCI address is 0 then we have our
280 * ISA memory hole
281 */
282 if (res->start == offset)
283 found_isa_hole = 1;
284 }
285 }
286
287 /* Handle ISA memory hole if not already covered */
288 if (j <= 2 && !found_isa_hole && hose->isa_mem_size)
289 if (ppc4xx_setup_one_pci_PMM(hose, reg, hose->isa_mem_phys, 0,
290 hose->isa_mem_size, 0, j) == 0)
291 printk(KERN_INFO "%pOF: Legacy ISA memory support enabled\n",
292 hose->dn);
293 }
294
ppc4xx_configure_pci_PTMs(struct pci_controller * hose,void __iomem * reg,const struct resource * res)295 static void __init ppc4xx_configure_pci_PTMs(struct pci_controller *hose,
296 void __iomem *reg,
297 const struct resource *res)
298 {
299 resource_size_t size = resource_size(res);
300 u32 sa;
301
302 /* Calculate window size */
303 sa = (0xffffffffu << ilog2(size)) | 1;
304 sa |= 0x1;
305
306 /* RAM is always at 0 local for now */
307 writel(0, reg + PCIL0_PTM1LA);
308 writel(sa, reg + PCIL0_PTM1MS);
309
310 /* Map on PCI side */
311 early_write_config_dword(hose, hose->first_busno, 0,
312 PCI_BASE_ADDRESS_1, res->start);
313 early_write_config_dword(hose, hose->first_busno, 0,
314 PCI_BASE_ADDRESS_2, 0x00000000);
315 early_write_config_word(hose, hose->first_busno, 0,
316 PCI_COMMAND, 0x0006);
317 }
318
ppc4xx_probe_pci_bridge(struct device_node * np)319 static void __init ppc4xx_probe_pci_bridge(struct device_node *np)
320 {
321 /* NYI */
322 struct resource rsrc_cfg;
323 struct resource rsrc_reg;
324 struct resource dma_window;
325 struct pci_controller *hose = NULL;
326 void __iomem *reg = NULL;
327 const int *bus_range;
328 int primary = 0;
329
330 /* Check if device is enabled */
331 if (!of_device_is_available(np)) {
332 printk(KERN_INFO "%pOF: Port disabled via device-tree\n", np);
333 return;
334 }
335
336 /* Fetch config space registers address */
337 if (of_address_to_resource(np, 0, &rsrc_cfg)) {
338 printk(KERN_ERR "%pOF: Can't get PCI config register base !",
339 np);
340 return;
341 }
342 /* Fetch host bridge internal registers address */
343 if (of_address_to_resource(np, 3, &rsrc_reg)) {
344 printk(KERN_ERR "%pOF: Can't get PCI internal register base !",
345 np);
346 return;
347 }
348
349 /* Check if primary bridge */
350 if (of_get_property(np, "primary", NULL))
351 primary = 1;
352
353 /* Get bus range if any */
354 bus_range = of_get_property(np, "bus-range", NULL);
355
356 /* Map registers */
357 reg = ioremap(rsrc_reg.start, resource_size(&rsrc_reg));
358 if (reg == NULL) {
359 printk(KERN_ERR "%pOF: Can't map registers !", np);
360 goto fail;
361 }
362
363 /* Allocate the host controller data structure */
364 hose = pcibios_alloc_controller(np);
365 if (!hose)
366 goto fail;
367
368 hose->first_busno = bus_range ? bus_range[0] : 0x0;
369 hose->last_busno = bus_range ? bus_range[1] : 0xff;
370
371 /* Setup config space */
372 setup_indirect_pci(hose, rsrc_cfg.start, rsrc_cfg.start + 0x4, 0);
373
374 /* Disable all windows */
375 writel(0, reg + PCIL0_PMM0MA);
376 writel(0, reg + PCIL0_PMM1MA);
377 writel(0, reg + PCIL0_PMM2MA);
378 writel(0, reg + PCIL0_PTM1MS);
379 writel(0, reg + PCIL0_PTM2MS);
380
381 /* Parse outbound mapping resources */
382 pci_process_bridge_OF_ranges(hose, np, primary);
383
384 /* Parse inbound mapping resources */
385 if (ppc4xx_parse_dma_ranges(hose, reg, &dma_window) != 0)
386 goto fail;
387
388 /* Configure outbound ranges POMs */
389 ppc4xx_configure_pci_PMMs(hose, reg);
390
391 /* Configure inbound ranges PIMs */
392 ppc4xx_configure_pci_PTMs(hose, reg, &dma_window);
393
394 /* We don't need the registers anymore */
395 iounmap(reg);
396 return;
397
398 fail:
399 if (hose)
400 pcibios_free_controller(hose);
401 if (reg)
402 iounmap(reg);
403 }
404
405 /*
406 * 4xx PCI-X part
407 */
408
ppc4xx_setup_one_pcix_POM(struct pci_controller * hose,void __iomem * reg,u64 plb_addr,u64 pci_addr,u64 size,unsigned int flags,int index)409 static int __init ppc4xx_setup_one_pcix_POM(struct pci_controller *hose,
410 void __iomem *reg,
411 u64 plb_addr,
412 u64 pci_addr,
413 u64 size,
414 unsigned int flags,
415 int index)
416 {
417 u32 lah, lal, pciah, pcial, sa;
418
419 if (!is_power_of_2(size) || size < 0x1000 ||
420 (plb_addr & (size - 1)) != 0) {
421 printk(KERN_WARNING "%pOF: Resource out of range\n",
422 hose->dn);
423 return -1;
424 }
425
426 /* Calculate register values */
427 lah = RES_TO_U32_HIGH(plb_addr);
428 lal = RES_TO_U32_LOW(plb_addr);
429 pciah = RES_TO_U32_HIGH(pci_addr);
430 pcial = RES_TO_U32_LOW(pci_addr);
431 sa = (0xffffffffu << ilog2(size)) | 0x1;
432
433 /* Program register values */
434 if (index == 0) {
435 writel(lah, reg + PCIX0_POM0LAH);
436 writel(lal, reg + PCIX0_POM0LAL);
437 writel(pciah, reg + PCIX0_POM0PCIAH);
438 writel(pcial, reg + PCIX0_POM0PCIAL);
439 writel(sa, reg + PCIX0_POM0SA);
440 } else {
441 writel(lah, reg + PCIX0_POM1LAH);
442 writel(lal, reg + PCIX0_POM1LAL);
443 writel(pciah, reg + PCIX0_POM1PCIAH);
444 writel(pcial, reg + PCIX0_POM1PCIAL);
445 writel(sa, reg + PCIX0_POM1SA);
446 }
447
448 return 0;
449 }
450
ppc4xx_configure_pcix_POMs(struct pci_controller * hose,void __iomem * reg)451 static void __init ppc4xx_configure_pcix_POMs(struct pci_controller *hose,
452 void __iomem *reg)
453 {
454 int i, j, found_isa_hole = 0;
455
456 /* Setup outbound memory windows */
457 for (i = j = 0; i < 3; i++) {
458 struct resource *res = &hose->mem_resources[i];
459 resource_size_t offset = hose->mem_offset[i];
460
461 /* we only care about memory windows */
462 if (!(res->flags & IORESOURCE_MEM))
463 continue;
464 if (j > 1) {
465 printk(KERN_WARNING "%pOF: Too many ranges\n", hose->dn);
466 break;
467 }
468
469 /* Configure the resource */
470 if (ppc4xx_setup_one_pcix_POM(hose, reg,
471 res->start,
472 res->start - offset,
473 resource_size(res),
474 res->flags,
475 j) == 0) {
476 j++;
477
478 /* If the resource PCI address is 0 then we have our
479 * ISA memory hole
480 */
481 if (res->start == offset)
482 found_isa_hole = 1;
483 }
484 }
485
486 /* Handle ISA memory hole if not already covered */
487 if (j <= 1 && !found_isa_hole && hose->isa_mem_size)
488 if (ppc4xx_setup_one_pcix_POM(hose, reg, hose->isa_mem_phys, 0,
489 hose->isa_mem_size, 0, j) == 0)
490 printk(KERN_INFO "%pOF: Legacy ISA memory support enabled\n",
491 hose->dn);
492 }
493
ppc4xx_configure_pcix_PIMs(struct pci_controller * hose,void __iomem * reg,const struct resource * res,int big_pim,int enable_msi_hole)494 static void __init ppc4xx_configure_pcix_PIMs(struct pci_controller *hose,
495 void __iomem *reg,
496 const struct resource *res,
497 int big_pim,
498 int enable_msi_hole)
499 {
500 resource_size_t size = resource_size(res);
501 u32 sa;
502
503 /* RAM is always at 0 */
504 writel(0x00000000, reg + PCIX0_PIM0LAH);
505 writel(0x00000000, reg + PCIX0_PIM0LAL);
506
507 /* Calculate window size */
508 sa = (0xffffffffu << ilog2(size)) | 1;
509 sa |= 0x1;
510 if (res->flags & IORESOURCE_PREFETCH)
511 sa |= 0x2;
512 if (enable_msi_hole)
513 sa |= 0x4;
514 writel(sa, reg + PCIX0_PIM0SA);
515 if (big_pim)
516 writel(0xffffffff, reg + PCIX0_PIM0SAH);
517
518 /* Map on PCI side */
519 writel(0x00000000, reg + PCIX0_BAR0H);
520 writel(res->start, reg + PCIX0_BAR0L);
521 writew(0x0006, reg + PCIX0_COMMAND);
522 }
523
ppc4xx_probe_pcix_bridge(struct device_node * np)524 static void __init ppc4xx_probe_pcix_bridge(struct device_node *np)
525 {
526 struct resource rsrc_cfg;
527 struct resource rsrc_reg;
528 struct resource dma_window;
529 struct pci_controller *hose = NULL;
530 void __iomem *reg = NULL;
531 const int *bus_range;
532 int big_pim = 0, msi = 0, primary = 0;
533
534 /* Fetch config space registers address */
535 if (of_address_to_resource(np, 0, &rsrc_cfg)) {
536 printk(KERN_ERR "%pOF: Can't get PCI-X config register base !",
537 np);
538 return;
539 }
540 /* Fetch host bridge internal registers address */
541 if (of_address_to_resource(np, 3, &rsrc_reg)) {
542 printk(KERN_ERR "%pOF: Can't get PCI-X internal register base !",
543 np);
544 return;
545 }
546
547 /* Check if it supports large PIMs (440GX) */
548 if (of_get_property(np, "large-inbound-windows", NULL))
549 big_pim = 1;
550
551 /* Check if we should enable MSIs inbound hole */
552 if (of_get_property(np, "enable-msi-hole", NULL))
553 msi = 1;
554
555 /* Check if primary bridge */
556 if (of_get_property(np, "primary", NULL))
557 primary = 1;
558
559 /* Get bus range if any */
560 bus_range = of_get_property(np, "bus-range", NULL);
561
562 /* Map registers */
563 reg = ioremap(rsrc_reg.start, resource_size(&rsrc_reg));
564 if (reg == NULL) {
565 printk(KERN_ERR "%pOF: Can't map registers !", np);
566 goto fail;
567 }
568
569 /* Allocate the host controller data structure */
570 hose = pcibios_alloc_controller(np);
571 if (!hose)
572 goto fail;
573
574 hose->first_busno = bus_range ? bus_range[0] : 0x0;
575 hose->last_busno = bus_range ? bus_range[1] : 0xff;
576
577 /* Setup config space */
578 setup_indirect_pci(hose, rsrc_cfg.start, rsrc_cfg.start + 0x4,
579 PPC_INDIRECT_TYPE_SET_CFG_TYPE);
580
581 /* Disable all windows */
582 writel(0, reg + PCIX0_POM0SA);
583 writel(0, reg + PCIX0_POM1SA);
584 writel(0, reg + PCIX0_POM2SA);
585 writel(0, reg + PCIX0_PIM0SA);
586 writel(0, reg + PCIX0_PIM1SA);
587 writel(0, reg + PCIX0_PIM2SA);
588 if (big_pim) {
589 writel(0, reg + PCIX0_PIM0SAH);
590 writel(0, reg + PCIX0_PIM2SAH);
591 }
592
593 /* Parse outbound mapping resources */
594 pci_process_bridge_OF_ranges(hose, np, primary);
595
596 /* Parse inbound mapping resources */
597 if (ppc4xx_parse_dma_ranges(hose, reg, &dma_window) != 0)
598 goto fail;
599
600 /* Configure outbound ranges POMs */
601 ppc4xx_configure_pcix_POMs(hose, reg);
602
603 /* Configure inbound ranges PIMs */
604 ppc4xx_configure_pcix_PIMs(hose, reg, &dma_window, big_pim, msi);
605
606 /* We don't need the registers anymore */
607 iounmap(reg);
608 return;
609
610 fail:
611 if (hose)
612 pcibios_free_controller(hose);
613 if (reg)
614 iounmap(reg);
615 }
616
617 #ifdef CONFIG_PPC4xx_PCI_EXPRESS
618
619 /*
620 * 4xx PCI-Express part
621 *
622 * We support 3 parts currently based on the compatible property:
623 *
624 * ibm,plb-pciex-440spe
625 * ibm,plb-pciex-405ex
626 * ibm,plb-pciex-460ex
627 *
628 * Anything else will be rejected for now as they are all subtly
629 * different unfortunately.
630 *
631 */
632
633 #define MAX_PCIE_BUS_MAPPED 0x40
634
635 struct ppc4xx_pciex_port
636 {
637 struct pci_controller *hose;
638 struct device_node *node;
639 unsigned int index;
640 int endpoint;
641 int link;
642 int has_ibpre;
643 unsigned int sdr_base;
644 dcr_host_t dcrs;
645 struct resource cfg_space;
646 struct resource utl_regs;
647 void __iomem *utl_base;
648 };
649
650 static struct ppc4xx_pciex_port *ppc4xx_pciex_ports;
651 static unsigned int ppc4xx_pciex_port_count;
652
653 struct ppc4xx_pciex_hwops
654 {
655 bool want_sdr;
656 int (*core_init)(struct device_node *np);
657 int (*port_init_hw)(struct ppc4xx_pciex_port *port);
658 int (*setup_utl)(struct ppc4xx_pciex_port *port);
659 void (*check_link)(struct ppc4xx_pciex_port *port);
660 };
661
662 static struct ppc4xx_pciex_hwops *ppc4xx_pciex_hwops;
663
ppc4xx_pciex_wait_on_sdr(struct ppc4xx_pciex_port * port,unsigned int sdr_offset,unsigned int mask,unsigned int value,int timeout_ms)664 static int __init ppc4xx_pciex_wait_on_sdr(struct ppc4xx_pciex_port *port,
665 unsigned int sdr_offset,
666 unsigned int mask,
667 unsigned int value,
668 int timeout_ms)
669 {
670 u32 val;
671
672 while(timeout_ms--) {
673 val = mfdcri(SDR0, port->sdr_base + sdr_offset);
674 if ((val & mask) == value) {
675 pr_debug("PCIE%d: Wait on SDR %x success with tm %d (%08x)\n",
676 port->index, sdr_offset, timeout_ms, val);
677 return 0;
678 }
679 msleep(1);
680 }
681 return -1;
682 }
683
ppc4xx_pciex_port_reset_sdr(struct ppc4xx_pciex_port * port)684 static int __init ppc4xx_pciex_port_reset_sdr(struct ppc4xx_pciex_port *port)
685 {
686 /* Wait for reset to complete */
687 if (ppc4xx_pciex_wait_on_sdr(port, PESDRn_RCSSTS, 1 << 20, 0, 10)) {
688 printk(KERN_WARNING "PCIE%d: PGRST failed\n",
689 port->index);
690 return -1;
691 }
692 return 0;
693 }
694
695
ppc4xx_pciex_check_link_sdr(struct ppc4xx_pciex_port * port)696 static void __init ppc4xx_pciex_check_link_sdr(struct ppc4xx_pciex_port *port)
697 {
698 printk(KERN_INFO "PCIE%d: Checking link...\n", port->index);
699
700 /* Check for card presence detect if supported, if not, just wait for
701 * link unconditionally.
702 *
703 * note that we don't fail if there is no link, we just filter out
704 * config space accesses. That way, it will be easier to implement
705 * hotplug later on.
706 */
707 if (!port->has_ibpre ||
708 !ppc4xx_pciex_wait_on_sdr(port, PESDRn_LOOP,
709 1 << 28, 1 << 28, 100)) {
710 printk(KERN_INFO
711 "PCIE%d: Device detected, waiting for link...\n",
712 port->index);
713 if (ppc4xx_pciex_wait_on_sdr(port, PESDRn_LOOP,
714 0x1000, 0x1000, 2000))
715 printk(KERN_WARNING
716 "PCIE%d: Link up failed\n", port->index);
717 else {
718 printk(KERN_INFO
719 "PCIE%d: link is up !\n", port->index);
720 port->link = 1;
721 }
722 } else
723 printk(KERN_INFO "PCIE%d: No device detected.\n", port->index);
724 }
725
726 #ifdef CONFIG_44x
727
728 /* Check various reset bits of the 440SPe PCIe core */
ppc440spe_pciex_check_reset(struct device_node * np)729 static int __init ppc440spe_pciex_check_reset(struct device_node *np)
730 {
731 u32 valPE0, valPE1, valPE2;
732 int err = 0;
733
734 /* SDR0_PEGPLLLCT1 reset */
735 if (!(mfdcri(SDR0, PESDR0_PLLLCT1) & 0x01000000)) {
736 /*
737 * the PCIe core was probably already initialised
738 * by firmware - let's re-reset RCSSET regs
739 *
740 * -- Shouldn't we also re-reset the whole thing ? -- BenH
741 */
742 pr_debug("PCIE: SDR0_PLLLCT1 already reset.\n");
743 mtdcri(SDR0, PESDR0_440SPE_RCSSET, 0x01010000);
744 mtdcri(SDR0, PESDR1_440SPE_RCSSET, 0x01010000);
745 mtdcri(SDR0, PESDR2_440SPE_RCSSET, 0x01010000);
746 }
747
748 valPE0 = mfdcri(SDR0, PESDR0_440SPE_RCSSET);
749 valPE1 = mfdcri(SDR0, PESDR1_440SPE_RCSSET);
750 valPE2 = mfdcri(SDR0, PESDR2_440SPE_RCSSET);
751
752 /* SDR0_PExRCSSET rstgu */
753 if (!(valPE0 & 0x01000000) ||
754 !(valPE1 & 0x01000000) ||
755 !(valPE2 & 0x01000000)) {
756 printk(KERN_INFO "PCIE: SDR0_PExRCSSET rstgu error\n");
757 err = -1;
758 }
759
760 /* SDR0_PExRCSSET rstdl */
761 if (!(valPE0 & 0x00010000) ||
762 !(valPE1 & 0x00010000) ||
763 !(valPE2 & 0x00010000)) {
764 printk(KERN_INFO "PCIE: SDR0_PExRCSSET rstdl error\n");
765 err = -1;
766 }
767
768 /* SDR0_PExRCSSET rstpyn */
769 if ((valPE0 & 0x00001000) ||
770 (valPE1 & 0x00001000) ||
771 (valPE2 & 0x00001000)) {
772 printk(KERN_INFO "PCIE: SDR0_PExRCSSET rstpyn error\n");
773 err = -1;
774 }
775
776 /* SDR0_PExRCSSET hldplb */
777 if ((valPE0 & 0x10000000) ||
778 (valPE1 & 0x10000000) ||
779 (valPE2 & 0x10000000)) {
780 printk(KERN_INFO "PCIE: SDR0_PExRCSSET hldplb error\n");
781 err = -1;
782 }
783
784 /* SDR0_PExRCSSET rdy */
785 if ((valPE0 & 0x00100000) ||
786 (valPE1 & 0x00100000) ||
787 (valPE2 & 0x00100000)) {
788 printk(KERN_INFO "PCIE: SDR0_PExRCSSET rdy error\n");
789 err = -1;
790 }
791
792 /* SDR0_PExRCSSET shutdown */
793 if ((valPE0 & 0x00000100) ||
794 (valPE1 & 0x00000100) ||
795 (valPE2 & 0x00000100)) {
796 printk(KERN_INFO "PCIE: SDR0_PExRCSSET shutdown error\n");
797 err = -1;
798 }
799
800 return err;
801 }
802
803 /* Global PCIe core initializations for 440SPe core */
ppc440spe_pciex_core_init(struct device_node * np)804 static int __init ppc440spe_pciex_core_init(struct device_node *np)
805 {
806 int time_out = 20;
807
808 /* Set PLL clock receiver to LVPECL */
809 dcri_clrset(SDR0, PESDR0_PLLLCT1, 0, 1 << 28);
810
811 /* Shouldn't we do all the calibration stuff etc... here ? */
812 if (ppc440spe_pciex_check_reset(np))
813 return -ENXIO;
814
815 if (!(mfdcri(SDR0, PESDR0_PLLLCT2) & 0x10000)) {
816 printk(KERN_INFO "PCIE: PESDR_PLLCT2 resistance calibration "
817 "failed (0x%08x)\n",
818 mfdcri(SDR0, PESDR0_PLLLCT2));
819 return -1;
820 }
821
822 /* De-assert reset of PCIe PLL, wait for lock */
823 dcri_clrset(SDR0, PESDR0_PLLLCT1, 1 << 24, 0);
824 udelay(3);
825
826 while (time_out) {
827 if (!(mfdcri(SDR0, PESDR0_PLLLCT3) & 0x10000000)) {
828 time_out--;
829 udelay(1);
830 } else
831 break;
832 }
833 if (!time_out) {
834 printk(KERN_INFO "PCIE: VCO output not locked\n");
835 return -1;
836 }
837
838 pr_debug("PCIE initialization OK\n");
839
840 return 3;
841 }
842
ppc440spe_pciex_init_port_hw(struct ppc4xx_pciex_port * port)843 static int __init ppc440spe_pciex_init_port_hw(struct ppc4xx_pciex_port *port)
844 {
845 u32 val = 1 << 24;
846
847 if (port->endpoint)
848 val = PTYPE_LEGACY_ENDPOINT << 20;
849 else
850 val = PTYPE_ROOT_PORT << 20;
851
852 if (port->index == 0)
853 val |= LNKW_X8 << 12;
854 else
855 val |= LNKW_X4 << 12;
856
857 mtdcri(SDR0, port->sdr_base + PESDRn_DLPSET, val);
858 mtdcri(SDR0, port->sdr_base + PESDRn_UTLSET1, 0x20222222);
859 if (ppc440spe_revA())
860 mtdcri(SDR0, port->sdr_base + PESDRn_UTLSET2, 0x11000000);
861 mtdcri(SDR0, port->sdr_base + PESDRn_440SPE_HSSL0SET1, 0x35000000);
862 mtdcri(SDR0, port->sdr_base + PESDRn_440SPE_HSSL1SET1, 0x35000000);
863 mtdcri(SDR0, port->sdr_base + PESDRn_440SPE_HSSL2SET1, 0x35000000);
864 mtdcri(SDR0, port->sdr_base + PESDRn_440SPE_HSSL3SET1, 0x35000000);
865 if (port->index == 0) {
866 mtdcri(SDR0, port->sdr_base + PESDRn_440SPE_HSSL4SET1,
867 0x35000000);
868 mtdcri(SDR0, port->sdr_base + PESDRn_440SPE_HSSL5SET1,
869 0x35000000);
870 mtdcri(SDR0, port->sdr_base + PESDRn_440SPE_HSSL6SET1,
871 0x35000000);
872 mtdcri(SDR0, port->sdr_base + PESDRn_440SPE_HSSL7SET1,
873 0x35000000);
874 }
875 dcri_clrset(SDR0, port->sdr_base + PESDRn_RCSSET,
876 (1 << 24) | (1 << 16), 1 << 12);
877
878 return ppc4xx_pciex_port_reset_sdr(port);
879 }
880
ppc440speA_pciex_init_port_hw(struct ppc4xx_pciex_port * port)881 static int __init ppc440speA_pciex_init_port_hw(struct ppc4xx_pciex_port *port)
882 {
883 return ppc440spe_pciex_init_port_hw(port);
884 }
885
ppc440speB_pciex_init_port_hw(struct ppc4xx_pciex_port * port)886 static int __init ppc440speB_pciex_init_port_hw(struct ppc4xx_pciex_port *port)
887 {
888 int rc = ppc440spe_pciex_init_port_hw(port);
889
890 port->has_ibpre = 1;
891
892 return rc;
893 }
894
ppc440speA_pciex_init_utl(struct ppc4xx_pciex_port * port)895 static int ppc440speA_pciex_init_utl(struct ppc4xx_pciex_port *port)
896 {
897 /* XXX Check what that value means... I hate magic */
898 dcr_write(port->dcrs, DCRO_PEGPL_SPECIAL, 0x68782800);
899
900 /*
901 * Set buffer allocations and then assert VRB and TXE.
902 */
903 out_be32(port->utl_base + PEUTL_OUTTR, 0x08000000);
904 out_be32(port->utl_base + PEUTL_INTR, 0x02000000);
905 out_be32(port->utl_base + PEUTL_OPDBSZ, 0x10000000);
906 out_be32(port->utl_base + PEUTL_PBBSZ, 0x53000000);
907 out_be32(port->utl_base + PEUTL_IPHBSZ, 0x08000000);
908 out_be32(port->utl_base + PEUTL_IPDBSZ, 0x10000000);
909 out_be32(port->utl_base + PEUTL_RCIRQEN, 0x00f00000);
910 out_be32(port->utl_base + PEUTL_PCTL, 0x80800066);
911
912 return 0;
913 }
914
ppc440speB_pciex_init_utl(struct ppc4xx_pciex_port * port)915 static int ppc440speB_pciex_init_utl(struct ppc4xx_pciex_port *port)
916 {
917 /* Report CRS to the operating system */
918 out_be32(port->utl_base + PEUTL_PBCTL, 0x08000000);
919
920 return 0;
921 }
922
923 static struct ppc4xx_pciex_hwops ppc440speA_pcie_hwops __initdata =
924 {
925 .want_sdr = true,
926 .core_init = ppc440spe_pciex_core_init,
927 .port_init_hw = ppc440speA_pciex_init_port_hw,
928 .setup_utl = ppc440speA_pciex_init_utl,
929 .check_link = ppc4xx_pciex_check_link_sdr,
930 };
931
932 static struct ppc4xx_pciex_hwops ppc440speB_pcie_hwops __initdata =
933 {
934 .want_sdr = true,
935 .core_init = ppc440spe_pciex_core_init,
936 .port_init_hw = ppc440speB_pciex_init_port_hw,
937 .setup_utl = ppc440speB_pciex_init_utl,
938 .check_link = ppc4xx_pciex_check_link_sdr,
939 };
940
ppc460ex_pciex_core_init(struct device_node * np)941 static int __init ppc460ex_pciex_core_init(struct device_node *np)
942 {
943 /* Nothing to do, return 2 ports */
944 return 2;
945 }
946
ppc460ex_pciex_init_port_hw(struct ppc4xx_pciex_port * port)947 static int __init ppc460ex_pciex_init_port_hw(struct ppc4xx_pciex_port *port)
948 {
949 u32 val;
950 u32 utlset1;
951
952 if (port->endpoint)
953 val = PTYPE_LEGACY_ENDPOINT << 20;
954 else
955 val = PTYPE_ROOT_PORT << 20;
956
957 if (port->index == 0) {
958 val |= LNKW_X1 << 12;
959 utlset1 = 0x20000000;
960 } else {
961 val |= LNKW_X4 << 12;
962 utlset1 = 0x20101101;
963 }
964
965 mtdcri(SDR0, port->sdr_base + PESDRn_DLPSET, val);
966 mtdcri(SDR0, port->sdr_base + PESDRn_UTLSET1, utlset1);
967 mtdcri(SDR0, port->sdr_base + PESDRn_UTLSET2, 0x01210000);
968
969 switch (port->index) {
970 case 0:
971 mtdcri(SDR0, PESDR0_460EX_L0CDRCTL, 0x00003230);
972 mtdcri(SDR0, PESDR0_460EX_L0DRV, 0x00000130);
973 mtdcri(SDR0, PESDR0_460EX_L0CLK, 0x00000006);
974
975 mtdcri(SDR0, PESDR0_460EX_PHY_CTL_RST,0x10000000);
976 break;
977
978 case 1:
979 mtdcri(SDR0, PESDR1_460EX_L0CDRCTL, 0x00003230);
980 mtdcri(SDR0, PESDR1_460EX_L1CDRCTL, 0x00003230);
981 mtdcri(SDR0, PESDR1_460EX_L2CDRCTL, 0x00003230);
982 mtdcri(SDR0, PESDR1_460EX_L3CDRCTL, 0x00003230);
983 mtdcri(SDR0, PESDR1_460EX_L0DRV, 0x00000130);
984 mtdcri(SDR0, PESDR1_460EX_L1DRV, 0x00000130);
985 mtdcri(SDR0, PESDR1_460EX_L2DRV, 0x00000130);
986 mtdcri(SDR0, PESDR1_460EX_L3DRV, 0x00000130);
987 mtdcri(SDR0, PESDR1_460EX_L0CLK, 0x00000006);
988 mtdcri(SDR0, PESDR1_460EX_L1CLK, 0x00000006);
989 mtdcri(SDR0, PESDR1_460EX_L2CLK, 0x00000006);
990 mtdcri(SDR0, PESDR1_460EX_L3CLK, 0x00000006);
991
992 mtdcri(SDR0, PESDR1_460EX_PHY_CTL_RST,0x10000000);
993 break;
994 }
995
996 mtdcri(SDR0, port->sdr_base + PESDRn_RCSSET,
997 mfdcri(SDR0, port->sdr_base + PESDRn_RCSSET) |
998 (PESDRx_RCSSET_RSTGU | PESDRx_RCSSET_RSTPYN));
999
1000 /* Poll for PHY reset */
1001 /* XXX FIXME add timeout */
1002 switch (port->index) {
1003 case 0:
1004 while (!(mfdcri(SDR0, PESDR0_460EX_RSTSTA) & 0x1))
1005 udelay(10);
1006 break;
1007 case 1:
1008 while (!(mfdcri(SDR0, PESDR1_460EX_RSTSTA) & 0x1))
1009 udelay(10);
1010 break;
1011 }
1012
1013 mtdcri(SDR0, port->sdr_base + PESDRn_RCSSET,
1014 (mfdcri(SDR0, port->sdr_base + PESDRn_RCSSET) &
1015 ~(PESDRx_RCSSET_RSTGU | PESDRx_RCSSET_RSTDL)) |
1016 PESDRx_RCSSET_RSTPYN);
1017
1018 port->has_ibpre = 1;
1019
1020 return ppc4xx_pciex_port_reset_sdr(port);
1021 }
1022
ppc460ex_pciex_init_utl(struct ppc4xx_pciex_port * port)1023 static int ppc460ex_pciex_init_utl(struct ppc4xx_pciex_port *port)
1024 {
1025 dcr_write(port->dcrs, DCRO_PEGPL_SPECIAL, 0x0);
1026
1027 /*
1028 * Set buffer allocations and then assert VRB and TXE.
1029 */
1030 out_be32(port->utl_base + PEUTL_PBCTL, 0x0800000c);
1031 out_be32(port->utl_base + PEUTL_OUTTR, 0x08000000);
1032 out_be32(port->utl_base + PEUTL_INTR, 0x02000000);
1033 out_be32(port->utl_base + PEUTL_OPDBSZ, 0x04000000);
1034 out_be32(port->utl_base + PEUTL_PBBSZ, 0x00000000);
1035 out_be32(port->utl_base + PEUTL_IPHBSZ, 0x02000000);
1036 out_be32(port->utl_base + PEUTL_IPDBSZ, 0x04000000);
1037 out_be32(port->utl_base + PEUTL_RCIRQEN,0x00f00000);
1038 out_be32(port->utl_base + PEUTL_PCTL, 0x80800066);
1039
1040 return 0;
1041 }
1042
1043 static struct ppc4xx_pciex_hwops ppc460ex_pcie_hwops __initdata =
1044 {
1045 .want_sdr = true,
1046 .core_init = ppc460ex_pciex_core_init,
1047 .port_init_hw = ppc460ex_pciex_init_port_hw,
1048 .setup_utl = ppc460ex_pciex_init_utl,
1049 .check_link = ppc4xx_pciex_check_link_sdr,
1050 };
1051
apm821xx_pciex_core_init(struct device_node * np)1052 static int __init apm821xx_pciex_core_init(struct device_node *np)
1053 {
1054 /* Return the number of pcie port */
1055 return 1;
1056 }
1057
apm821xx_pciex_init_port_hw(struct ppc4xx_pciex_port * port)1058 static int __init apm821xx_pciex_init_port_hw(struct ppc4xx_pciex_port *port)
1059 {
1060 u32 val;
1061
1062 /*
1063 * Do a software reset on PCIe ports.
1064 * This code is to fix the issue that pci drivers doesn't re-assign
1065 * bus number for PCIE devices after Uboot
1066 * scanned and configured all the buses (eg. PCIE NIC IntelPro/1000
1067 * PT quad port, SAS LSI 1064E)
1068 */
1069
1070 mtdcri(SDR0, PESDR0_460EX_PHY_CTL_RST, 0x0);
1071 mdelay(10);
1072
1073 if (port->endpoint)
1074 val = PTYPE_LEGACY_ENDPOINT << 20;
1075 else
1076 val = PTYPE_ROOT_PORT << 20;
1077
1078 val |= LNKW_X1 << 12;
1079
1080 mtdcri(SDR0, port->sdr_base + PESDRn_DLPSET, val);
1081 mtdcri(SDR0, port->sdr_base + PESDRn_UTLSET1, 0x00000000);
1082 mtdcri(SDR0, port->sdr_base + PESDRn_UTLSET2, 0x01010000);
1083
1084 mtdcri(SDR0, PESDR0_460EX_L0CDRCTL, 0x00003230);
1085 mtdcri(SDR0, PESDR0_460EX_L0DRV, 0x00000130);
1086 mtdcri(SDR0, PESDR0_460EX_L0CLK, 0x00000006);
1087
1088 mtdcri(SDR0, PESDR0_460EX_PHY_CTL_RST, 0x10000000);
1089 mdelay(50);
1090 mtdcri(SDR0, PESDR0_460EX_PHY_CTL_RST, 0x30000000);
1091
1092 mtdcri(SDR0, port->sdr_base + PESDRn_RCSSET,
1093 mfdcri(SDR0, port->sdr_base + PESDRn_RCSSET) |
1094 (PESDRx_RCSSET_RSTGU | PESDRx_RCSSET_RSTPYN));
1095
1096 /* Poll for PHY reset */
1097 val = PESDR0_460EX_RSTSTA - port->sdr_base;
1098 if (ppc4xx_pciex_wait_on_sdr(port, val, 0x1, 1, 100)) {
1099 printk(KERN_WARNING "%s: PCIE: Can't reset PHY\n", __func__);
1100 return -EBUSY;
1101 } else {
1102 mtdcri(SDR0, port->sdr_base + PESDRn_RCSSET,
1103 (mfdcri(SDR0, port->sdr_base + PESDRn_RCSSET) &
1104 ~(PESDRx_RCSSET_RSTGU | PESDRx_RCSSET_RSTDL)) |
1105 PESDRx_RCSSET_RSTPYN);
1106
1107 port->has_ibpre = 1;
1108 return 0;
1109 }
1110 }
1111
1112 static struct ppc4xx_pciex_hwops apm821xx_pcie_hwops __initdata = {
1113 .want_sdr = true,
1114 .core_init = apm821xx_pciex_core_init,
1115 .port_init_hw = apm821xx_pciex_init_port_hw,
1116 .setup_utl = ppc460ex_pciex_init_utl,
1117 .check_link = ppc4xx_pciex_check_link_sdr,
1118 };
1119
ppc460sx_pciex_core_init(struct device_node * np)1120 static int __init ppc460sx_pciex_core_init(struct device_node *np)
1121 {
1122 /* HSS drive amplitude */
1123 mtdcri(SDR0, PESDR0_460SX_HSSL0DAMP, 0xB9843211);
1124 mtdcri(SDR0, PESDR0_460SX_HSSL1DAMP, 0xB9843211);
1125 mtdcri(SDR0, PESDR0_460SX_HSSL2DAMP, 0xB9843211);
1126 mtdcri(SDR0, PESDR0_460SX_HSSL3DAMP, 0xB9843211);
1127 mtdcri(SDR0, PESDR0_460SX_HSSL4DAMP, 0xB9843211);
1128 mtdcri(SDR0, PESDR0_460SX_HSSL5DAMP, 0xB9843211);
1129 mtdcri(SDR0, PESDR0_460SX_HSSL6DAMP, 0xB9843211);
1130 mtdcri(SDR0, PESDR0_460SX_HSSL7DAMP, 0xB9843211);
1131
1132 mtdcri(SDR0, PESDR1_460SX_HSSL0DAMP, 0xB9843211);
1133 mtdcri(SDR0, PESDR1_460SX_HSSL1DAMP, 0xB9843211);
1134 mtdcri(SDR0, PESDR1_460SX_HSSL2DAMP, 0xB9843211);
1135 mtdcri(SDR0, PESDR1_460SX_HSSL3DAMP, 0xB9843211);
1136
1137 mtdcri(SDR0, PESDR2_460SX_HSSL0DAMP, 0xB9843211);
1138 mtdcri(SDR0, PESDR2_460SX_HSSL1DAMP, 0xB9843211);
1139 mtdcri(SDR0, PESDR2_460SX_HSSL2DAMP, 0xB9843211);
1140 mtdcri(SDR0, PESDR2_460SX_HSSL3DAMP, 0xB9843211);
1141
1142 /* HSS TX pre-emphasis */
1143 mtdcri(SDR0, PESDR0_460SX_HSSL0COEFA, 0xDCB98987);
1144 mtdcri(SDR0, PESDR0_460SX_HSSL1COEFA, 0xDCB98987);
1145 mtdcri(SDR0, PESDR0_460SX_HSSL2COEFA, 0xDCB98987);
1146 mtdcri(SDR0, PESDR0_460SX_HSSL3COEFA, 0xDCB98987);
1147 mtdcri(SDR0, PESDR0_460SX_HSSL4COEFA, 0xDCB98987);
1148 mtdcri(SDR0, PESDR0_460SX_HSSL5COEFA, 0xDCB98987);
1149 mtdcri(SDR0, PESDR0_460SX_HSSL6COEFA, 0xDCB98987);
1150 mtdcri(SDR0, PESDR0_460SX_HSSL7COEFA, 0xDCB98987);
1151
1152 mtdcri(SDR0, PESDR1_460SX_HSSL0COEFA, 0xDCB98987);
1153 mtdcri(SDR0, PESDR1_460SX_HSSL1COEFA, 0xDCB98987);
1154 mtdcri(SDR0, PESDR1_460SX_HSSL2COEFA, 0xDCB98987);
1155 mtdcri(SDR0, PESDR1_460SX_HSSL3COEFA, 0xDCB98987);
1156
1157 mtdcri(SDR0, PESDR2_460SX_HSSL0COEFA, 0xDCB98987);
1158 mtdcri(SDR0, PESDR2_460SX_HSSL1COEFA, 0xDCB98987);
1159 mtdcri(SDR0, PESDR2_460SX_HSSL2COEFA, 0xDCB98987);
1160 mtdcri(SDR0, PESDR2_460SX_HSSL3COEFA, 0xDCB98987);
1161
1162 /* HSS TX calibration control */
1163 mtdcri(SDR0, PESDR0_460SX_HSSL1CALDRV, 0x22222222);
1164 mtdcri(SDR0, PESDR1_460SX_HSSL1CALDRV, 0x22220000);
1165 mtdcri(SDR0, PESDR2_460SX_HSSL1CALDRV, 0x22220000);
1166
1167 /* HSS TX slew control */
1168 mtdcri(SDR0, PESDR0_460SX_HSSSLEW, 0xFFFFFFFF);
1169 mtdcri(SDR0, PESDR1_460SX_HSSSLEW, 0xFFFF0000);
1170 mtdcri(SDR0, PESDR2_460SX_HSSSLEW, 0xFFFF0000);
1171
1172 /* Set HSS PRBS enabled */
1173 mtdcri(SDR0, PESDR0_460SX_HSSCTLSET, 0x00001130);
1174 mtdcri(SDR0, PESDR2_460SX_HSSCTLSET, 0x00001130);
1175
1176 udelay(100);
1177
1178 /* De-assert PLLRESET */
1179 dcri_clrset(SDR0, PESDR0_PLLLCT2, 0x00000100, 0);
1180
1181 /* Reset DL, UTL, GPL before configuration */
1182 mtdcri(SDR0, PESDR0_460SX_RCSSET,
1183 PESDRx_RCSSET_RSTDL | PESDRx_RCSSET_RSTGU);
1184 mtdcri(SDR0, PESDR1_460SX_RCSSET,
1185 PESDRx_RCSSET_RSTDL | PESDRx_RCSSET_RSTGU);
1186 mtdcri(SDR0, PESDR2_460SX_RCSSET,
1187 PESDRx_RCSSET_RSTDL | PESDRx_RCSSET_RSTGU);
1188
1189 udelay(100);
1190
1191 /*
1192 * If bifurcation is not enabled, u-boot would have disabled the
1193 * third PCIe port
1194 */
1195 if (((mfdcri(SDR0, PESDR1_460SX_HSSCTLSET) & 0x00000001) ==
1196 0x00000001)) {
1197 printk(KERN_INFO "PCI: PCIE bifurcation setup successfully.\n");
1198 printk(KERN_INFO "PCI: Total 3 PCIE ports are present\n");
1199 return 3;
1200 }
1201
1202 printk(KERN_INFO "PCI: Total 2 PCIE ports are present\n");
1203 return 2;
1204 }
1205
ppc460sx_pciex_init_port_hw(struct ppc4xx_pciex_port * port)1206 static int __init ppc460sx_pciex_init_port_hw(struct ppc4xx_pciex_port *port)
1207 {
1208
1209 if (port->endpoint)
1210 dcri_clrset(SDR0, port->sdr_base + PESDRn_UTLSET2,
1211 0x01000000, 0);
1212 else
1213 dcri_clrset(SDR0, port->sdr_base + PESDRn_UTLSET2,
1214 0, 0x01000000);
1215
1216 dcri_clrset(SDR0, port->sdr_base + PESDRn_RCSSET,
1217 (PESDRx_RCSSET_RSTGU | PESDRx_RCSSET_RSTDL),
1218 PESDRx_RCSSET_RSTPYN);
1219
1220 port->has_ibpre = 1;
1221
1222 return ppc4xx_pciex_port_reset_sdr(port);
1223 }
1224
ppc460sx_pciex_init_utl(struct ppc4xx_pciex_port * port)1225 static int ppc460sx_pciex_init_utl(struct ppc4xx_pciex_port *port)
1226 {
1227 /* Max 128 Bytes */
1228 out_be32 (port->utl_base + PEUTL_PBBSZ, 0x00000000);
1229 /* Assert VRB and TXE - per datasheet turn off addr validation */
1230 out_be32(port->utl_base + PEUTL_PCTL, 0x80800000);
1231 return 0;
1232 }
1233
ppc460sx_pciex_check_link(struct ppc4xx_pciex_port * port)1234 static void __init ppc460sx_pciex_check_link(struct ppc4xx_pciex_port *port)
1235 {
1236 void __iomem *mbase;
1237 int attempt = 50;
1238
1239 port->link = 0;
1240
1241 mbase = ioremap(port->cfg_space.start + 0x10000000, 0x1000);
1242 if (mbase == NULL) {
1243 printk(KERN_ERR "%pOF: Can't map internal config space !",
1244 port->node);
1245 goto done;
1246 }
1247
1248 while (attempt && (0 == (in_le32(mbase + PECFG_460SX_DLLSTA)
1249 & PECFG_460SX_DLLSTA_LINKUP))) {
1250 attempt--;
1251 mdelay(10);
1252 }
1253 if (attempt)
1254 port->link = 1;
1255 done:
1256 iounmap(mbase);
1257
1258 }
1259
1260 static struct ppc4xx_pciex_hwops ppc460sx_pcie_hwops __initdata = {
1261 .want_sdr = true,
1262 .core_init = ppc460sx_pciex_core_init,
1263 .port_init_hw = ppc460sx_pciex_init_port_hw,
1264 .setup_utl = ppc460sx_pciex_init_utl,
1265 .check_link = ppc460sx_pciex_check_link,
1266 };
1267
1268 #endif /* CONFIG_44x */
1269
1270 #ifdef CONFIG_40x
1271
ppc405ex_pciex_core_init(struct device_node * np)1272 static int __init ppc405ex_pciex_core_init(struct device_node *np)
1273 {
1274 /* Nothing to do, return 2 ports */
1275 return 2;
1276 }
1277
ppc405ex_pcie_phy_reset(struct ppc4xx_pciex_port * port)1278 static void ppc405ex_pcie_phy_reset(struct ppc4xx_pciex_port *port)
1279 {
1280 /* Assert the PE0_PHY reset */
1281 mtdcri(SDR0, port->sdr_base + PESDRn_RCSSET, 0x01010000);
1282 msleep(1);
1283
1284 /* deassert the PE0_hotreset */
1285 if (port->endpoint)
1286 mtdcri(SDR0, port->sdr_base + PESDRn_RCSSET, 0x01111000);
1287 else
1288 mtdcri(SDR0, port->sdr_base + PESDRn_RCSSET, 0x01101000);
1289
1290 /* poll for phy !reset */
1291 /* XXX FIXME add timeout */
1292 while (!(mfdcri(SDR0, port->sdr_base + PESDRn_405EX_PHYSTA) & 0x00001000))
1293 ;
1294
1295 /* deassert the PE0_gpl_utl_reset */
1296 mtdcri(SDR0, port->sdr_base + PESDRn_RCSSET, 0x00101000);
1297 }
1298
ppc405ex_pciex_init_port_hw(struct ppc4xx_pciex_port * port)1299 static int __init ppc405ex_pciex_init_port_hw(struct ppc4xx_pciex_port *port)
1300 {
1301 u32 val;
1302
1303 if (port->endpoint)
1304 val = PTYPE_LEGACY_ENDPOINT;
1305 else
1306 val = PTYPE_ROOT_PORT;
1307
1308 mtdcri(SDR0, port->sdr_base + PESDRn_DLPSET,
1309 1 << 24 | val << 20 | LNKW_X1 << 12);
1310
1311 mtdcri(SDR0, port->sdr_base + PESDRn_UTLSET1, 0x00000000);
1312 mtdcri(SDR0, port->sdr_base + PESDRn_UTLSET2, 0x01010000);
1313 mtdcri(SDR0, port->sdr_base + PESDRn_405EX_PHYSET1, 0x720F0000);
1314 mtdcri(SDR0, port->sdr_base + PESDRn_405EX_PHYSET2, 0x70600003);
1315
1316 /*
1317 * Only reset the PHY when no link is currently established.
1318 * This is for the Atheros PCIe board which has problems to establish
1319 * the link (again) after this PHY reset. All other currently tested
1320 * PCIe boards don't show this problem.
1321 * This has to be re-tested and fixed in a later release!
1322 */
1323 val = mfdcri(SDR0, port->sdr_base + PESDRn_LOOP);
1324 if (!(val & 0x00001000))
1325 ppc405ex_pcie_phy_reset(port);
1326
1327 dcr_write(port->dcrs, DCRO_PEGPL_CFG, 0x10000000); /* guarded on */
1328
1329 port->has_ibpre = 1;
1330
1331 return ppc4xx_pciex_port_reset_sdr(port);
1332 }
1333
ppc405ex_pciex_init_utl(struct ppc4xx_pciex_port * port)1334 static int ppc405ex_pciex_init_utl(struct ppc4xx_pciex_port *port)
1335 {
1336 dcr_write(port->dcrs, DCRO_PEGPL_SPECIAL, 0x0);
1337
1338 /*
1339 * Set buffer allocations and then assert VRB and TXE.
1340 */
1341 out_be32(port->utl_base + PEUTL_OUTTR, 0x02000000);
1342 out_be32(port->utl_base + PEUTL_INTR, 0x02000000);
1343 out_be32(port->utl_base + PEUTL_OPDBSZ, 0x04000000);
1344 out_be32(port->utl_base + PEUTL_PBBSZ, 0x21000000);
1345 out_be32(port->utl_base + PEUTL_IPHBSZ, 0x02000000);
1346 out_be32(port->utl_base + PEUTL_IPDBSZ, 0x04000000);
1347 out_be32(port->utl_base + PEUTL_RCIRQEN, 0x00f00000);
1348 out_be32(port->utl_base + PEUTL_PCTL, 0x80800066);
1349
1350 out_be32(port->utl_base + PEUTL_PBCTL, 0x08000000);
1351
1352 return 0;
1353 }
1354
1355 static struct ppc4xx_pciex_hwops ppc405ex_pcie_hwops __initdata =
1356 {
1357 .want_sdr = true,
1358 .core_init = ppc405ex_pciex_core_init,
1359 .port_init_hw = ppc405ex_pciex_init_port_hw,
1360 .setup_utl = ppc405ex_pciex_init_utl,
1361 .check_link = ppc4xx_pciex_check_link_sdr,
1362 };
1363
1364 #endif /* CONFIG_40x */
1365
1366 #ifdef CONFIG_476FPE
ppc_476fpe_pciex_core_init(struct device_node * np)1367 static int __init ppc_476fpe_pciex_core_init(struct device_node *np)
1368 {
1369 return 4;
1370 }
1371
ppc_476fpe_pciex_check_link(struct ppc4xx_pciex_port * port)1372 static void __init ppc_476fpe_pciex_check_link(struct ppc4xx_pciex_port *port)
1373 {
1374 u32 timeout_ms = 20;
1375 u32 val = 0, mask = (PECFG_TLDLP_LNKUP|PECFG_TLDLP_PRESENT);
1376 void __iomem *mbase = ioremap(port->cfg_space.start + 0x10000000,
1377 0x1000);
1378
1379 printk(KERN_INFO "PCIE%d: Checking link...\n", port->index);
1380
1381 if (mbase == NULL) {
1382 printk(KERN_WARNING "PCIE%d: failed to get cfg space\n",
1383 port->index);
1384 return;
1385 }
1386
1387 while (timeout_ms--) {
1388 val = in_le32(mbase + PECFG_TLDLP);
1389
1390 if ((val & mask) == mask)
1391 break;
1392 msleep(10);
1393 }
1394
1395 if (val & PECFG_TLDLP_PRESENT) {
1396 printk(KERN_INFO "PCIE%d: link is up !\n", port->index);
1397 port->link = 1;
1398 } else
1399 printk(KERN_WARNING "PCIE%d: Link up failed\n", port->index);
1400
1401 iounmap(mbase);
1402 }
1403
1404 static struct ppc4xx_pciex_hwops ppc_476fpe_pcie_hwops __initdata =
1405 {
1406 .core_init = ppc_476fpe_pciex_core_init,
1407 .check_link = ppc_476fpe_pciex_check_link,
1408 };
1409 #endif /* CONFIG_476FPE */
1410
1411 /* Check that the core has been initied and if not, do it */
ppc4xx_pciex_check_core_init(struct device_node * np)1412 static int __init ppc4xx_pciex_check_core_init(struct device_node *np)
1413 {
1414 static int core_init;
1415 int count = -ENODEV;
1416
1417 if (core_init++)
1418 return 0;
1419
1420 #ifdef CONFIG_44x
1421 if (of_device_is_compatible(np, "ibm,plb-pciex-440spe")) {
1422 if (ppc440spe_revA())
1423 ppc4xx_pciex_hwops = &ppc440speA_pcie_hwops;
1424 else
1425 ppc4xx_pciex_hwops = &ppc440speB_pcie_hwops;
1426 }
1427 if (of_device_is_compatible(np, "ibm,plb-pciex-460ex"))
1428 ppc4xx_pciex_hwops = &ppc460ex_pcie_hwops;
1429 if (of_device_is_compatible(np, "ibm,plb-pciex-460sx"))
1430 ppc4xx_pciex_hwops = &ppc460sx_pcie_hwops;
1431 if (of_device_is_compatible(np, "ibm,plb-pciex-apm821xx"))
1432 ppc4xx_pciex_hwops = &apm821xx_pcie_hwops;
1433 #endif /* CONFIG_44x */
1434 #ifdef CONFIG_40x
1435 if (of_device_is_compatible(np, "ibm,plb-pciex-405ex"))
1436 ppc4xx_pciex_hwops = &ppc405ex_pcie_hwops;
1437 #endif
1438 #ifdef CONFIG_476FPE
1439 if (of_device_is_compatible(np, "ibm,plb-pciex-476fpe")
1440 || of_device_is_compatible(np, "ibm,plb-pciex-476gtr"))
1441 ppc4xx_pciex_hwops = &ppc_476fpe_pcie_hwops;
1442 #endif
1443 if (ppc4xx_pciex_hwops == NULL) {
1444 printk(KERN_WARNING "PCIE: unknown host type %pOF\n", np);
1445 return -ENODEV;
1446 }
1447
1448 count = ppc4xx_pciex_hwops->core_init(np);
1449 if (count > 0) {
1450 ppc4xx_pciex_ports =
1451 kcalloc(count, sizeof(struct ppc4xx_pciex_port),
1452 GFP_KERNEL);
1453 if (ppc4xx_pciex_ports) {
1454 ppc4xx_pciex_port_count = count;
1455 return 0;
1456 }
1457 printk(KERN_WARNING "PCIE: failed to allocate ports array\n");
1458 return -ENOMEM;
1459 }
1460 return -ENODEV;
1461 }
1462
ppc4xx_pciex_port_init_mapping(struct ppc4xx_pciex_port * port)1463 static void __init ppc4xx_pciex_port_init_mapping(struct ppc4xx_pciex_port *port)
1464 {
1465 /* We map PCI Express configuration based on the reg property */
1466 dcr_write(port->dcrs, DCRO_PEGPL_CFGBAH,
1467 RES_TO_U32_HIGH(port->cfg_space.start));
1468 dcr_write(port->dcrs, DCRO_PEGPL_CFGBAL,
1469 RES_TO_U32_LOW(port->cfg_space.start));
1470
1471 /* XXX FIXME: Use size from reg property. For now, map 512M */
1472 dcr_write(port->dcrs, DCRO_PEGPL_CFGMSK, 0xe0000001);
1473
1474 /* We map UTL registers based on the reg property */
1475 dcr_write(port->dcrs, DCRO_PEGPL_REGBAH,
1476 RES_TO_U32_HIGH(port->utl_regs.start));
1477 dcr_write(port->dcrs, DCRO_PEGPL_REGBAL,
1478 RES_TO_U32_LOW(port->utl_regs.start));
1479
1480 /* XXX FIXME: Use size from reg property */
1481 dcr_write(port->dcrs, DCRO_PEGPL_REGMSK, 0x00007001);
1482
1483 /* Disable all other outbound windows */
1484 dcr_write(port->dcrs, DCRO_PEGPL_OMR1MSKL, 0);
1485 dcr_write(port->dcrs, DCRO_PEGPL_OMR2MSKL, 0);
1486 dcr_write(port->dcrs, DCRO_PEGPL_OMR3MSKL, 0);
1487 dcr_write(port->dcrs, DCRO_PEGPL_MSGMSK, 0);
1488 }
1489
ppc4xx_pciex_port_init(struct ppc4xx_pciex_port * port)1490 static int __init ppc4xx_pciex_port_init(struct ppc4xx_pciex_port *port)
1491 {
1492 int rc = 0;
1493
1494 /* Init HW */
1495 if (ppc4xx_pciex_hwops->port_init_hw)
1496 rc = ppc4xx_pciex_hwops->port_init_hw(port);
1497 if (rc != 0)
1498 return rc;
1499
1500 /*
1501 * Initialize mapping: disable all regions and configure
1502 * CFG and REG regions based on resources in the device tree
1503 */
1504 ppc4xx_pciex_port_init_mapping(port);
1505
1506 if (ppc4xx_pciex_hwops->check_link)
1507 ppc4xx_pciex_hwops->check_link(port);
1508
1509 /*
1510 * Map UTL
1511 */
1512 port->utl_base = ioremap(port->utl_regs.start, 0x100);
1513 BUG_ON(port->utl_base == NULL);
1514
1515 /*
1516 * Setup UTL registers --BenH.
1517 */
1518 if (ppc4xx_pciex_hwops->setup_utl)
1519 ppc4xx_pciex_hwops->setup_utl(port);
1520
1521 /*
1522 * Check for VC0 active or PLL Locked and assert RDY.
1523 */
1524 if (port->sdr_base) {
1525 if (of_device_is_compatible(port->node,
1526 "ibm,plb-pciex-460sx")){
1527 if (port->link && ppc4xx_pciex_wait_on_sdr(port,
1528 PESDRn_RCSSTS,
1529 1 << 12, 1 << 12, 5000)) {
1530 printk(KERN_INFO "PCIE%d: PLL not locked\n",
1531 port->index);
1532 port->link = 0;
1533 }
1534 } else if (port->link &&
1535 ppc4xx_pciex_wait_on_sdr(port, PESDRn_RCSSTS,
1536 1 << 16, 1 << 16, 5000)) {
1537 printk(KERN_INFO "PCIE%d: VC0 not active\n",
1538 port->index);
1539 port->link = 0;
1540 }
1541
1542 dcri_clrset(SDR0, port->sdr_base + PESDRn_RCSSET, 0, 1 << 20);
1543 }
1544
1545 msleep(100);
1546
1547 return 0;
1548 }
1549
ppc4xx_pciex_validate_bdf(struct ppc4xx_pciex_port * port,struct pci_bus * bus,unsigned int devfn)1550 static int ppc4xx_pciex_validate_bdf(struct ppc4xx_pciex_port *port,
1551 struct pci_bus *bus,
1552 unsigned int devfn)
1553 {
1554 static int message;
1555
1556 /* Endpoint can not generate upstream(remote) config cycles */
1557 if (port->endpoint && bus->number != port->hose->first_busno)
1558 return PCIBIOS_DEVICE_NOT_FOUND;
1559
1560 /* Check we are within the mapped range */
1561 if (bus->number > port->hose->last_busno) {
1562 if (!message) {
1563 printk(KERN_WARNING "Warning! Probing bus %u"
1564 " out of range !\n", bus->number);
1565 message++;
1566 }
1567 return PCIBIOS_DEVICE_NOT_FOUND;
1568 }
1569
1570 /* The root complex has only one device / function */
1571 if (bus->number == port->hose->first_busno && devfn != 0)
1572 return PCIBIOS_DEVICE_NOT_FOUND;
1573
1574 /* The other side of the RC has only one device as well */
1575 if (bus->number == (port->hose->first_busno + 1) &&
1576 PCI_SLOT(devfn) != 0)
1577 return PCIBIOS_DEVICE_NOT_FOUND;
1578
1579 /* Check if we have a link */
1580 if ((bus->number != port->hose->first_busno) && !port->link)
1581 return PCIBIOS_DEVICE_NOT_FOUND;
1582
1583 return 0;
1584 }
1585
ppc4xx_pciex_get_config_base(struct ppc4xx_pciex_port * port,struct pci_bus * bus,unsigned int devfn)1586 static void __iomem *ppc4xx_pciex_get_config_base(struct ppc4xx_pciex_port *port,
1587 struct pci_bus *bus,
1588 unsigned int devfn)
1589 {
1590 int relbus;
1591
1592 /* Remove the casts when we finally remove the stupid volatile
1593 * in struct pci_controller
1594 */
1595 if (bus->number == port->hose->first_busno)
1596 return (void __iomem *)port->hose->cfg_addr;
1597
1598 relbus = bus->number - (port->hose->first_busno + 1);
1599 return (void __iomem *)port->hose->cfg_data +
1600 ((relbus << 20) | (devfn << 12));
1601 }
1602
ppc4xx_pciex_read_config(struct pci_bus * bus,unsigned int devfn,int offset,int len,u32 * val)1603 static int ppc4xx_pciex_read_config(struct pci_bus *bus, unsigned int devfn,
1604 int offset, int len, u32 *val)
1605 {
1606 struct pci_controller *hose = pci_bus_to_host(bus);
1607 struct ppc4xx_pciex_port *port =
1608 &ppc4xx_pciex_ports[hose->indirect_type];
1609 void __iomem *addr;
1610 u32 gpl_cfg;
1611
1612 BUG_ON(hose != port->hose);
1613
1614 if (ppc4xx_pciex_validate_bdf(port, bus, devfn) != 0)
1615 return PCIBIOS_DEVICE_NOT_FOUND;
1616
1617 addr = ppc4xx_pciex_get_config_base(port, bus, devfn);
1618
1619 /*
1620 * Reading from configuration space of non-existing device can
1621 * generate transaction errors. For the read duration we suppress
1622 * assertion of machine check exceptions to avoid those.
1623 */
1624 gpl_cfg = dcr_read(port->dcrs, DCRO_PEGPL_CFG);
1625 dcr_write(port->dcrs, DCRO_PEGPL_CFG, gpl_cfg | GPL_DMER_MASK_DISA);
1626
1627 /* Make sure no CRS is recorded */
1628 out_be32(port->utl_base + PEUTL_RCSTA, 0x00040000);
1629
1630 switch (len) {
1631 case 1:
1632 *val = in_8((u8 *)(addr + offset));
1633 break;
1634 case 2:
1635 *val = in_le16((u16 *)(addr + offset));
1636 break;
1637 default:
1638 *val = in_le32((u32 *)(addr + offset));
1639 break;
1640 }
1641
1642 pr_debug("pcie-config-read: bus=%3d [%3d..%3d] devfn=0x%04x"
1643 " offset=0x%04x len=%d, addr=0x%p val=0x%08x\n",
1644 bus->number, hose->first_busno, hose->last_busno,
1645 devfn, offset, len, addr + offset, *val);
1646
1647 /* Check for CRS (440SPe rev B does that for us but heh ..) */
1648 if (in_be32(port->utl_base + PEUTL_RCSTA) & 0x00040000) {
1649 pr_debug("Got CRS !\n");
1650 if (len != 4 || offset != 0)
1651 return PCIBIOS_DEVICE_NOT_FOUND;
1652 *val = 0xffff0001;
1653 }
1654
1655 dcr_write(port->dcrs, DCRO_PEGPL_CFG, gpl_cfg);
1656
1657 return PCIBIOS_SUCCESSFUL;
1658 }
1659
ppc4xx_pciex_write_config(struct pci_bus * bus,unsigned int devfn,int offset,int len,u32 val)1660 static int ppc4xx_pciex_write_config(struct pci_bus *bus, unsigned int devfn,
1661 int offset, int len, u32 val)
1662 {
1663 struct pci_controller *hose = pci_bus_to_host(bus);
1664 struct ppc4xx_pciex_port *port =
1665 &ppc4xx_pciex_ports[hose->indirect_type];
1666 void __iomem *addr;
1667 u32 gpl_cfg;
1668
1669 if (ppc4xx_pciex_validate_bdf(port, bus, devfn) != 0)
1670 return PCIBIOS_DEVICE_NOT_FOUND;
1671
1672 addr = ppc4xx_pciex_get_config_base(port, bus, devfn);
1673
1674 /*
1675 * Reading from configuration space of non-existing device can
1676 * generate transaction errors. For the read duration we suppress
1677 * assertion of machine check exceptions to avoid those.
1678 */
1679 gpl_cfg = dcr_read(port->dcrs, DCRO_PEGPL_CFG);
1680 dcr_write(port->dcrs, DCRO_PEGPL_CFG, gpl_cfg | GPL_DMER_MASK_DISA);
1681
1682 pr_debug("pcie-config-write: bus=%3d [%3d..%3d] devfn=0x%04x"
1683 " offset=0x%04x len=%d, addr=0x%p val=0x%08x\n",
1684 bus->number, hose->first_busno, hose->last_busno,
1685 devfn, offset, len, addr + offset, val);
1686
1687 switch (len) {
1688 case 1:
1689 out_8((u8 *)(addr + offset), val);
1690 break;
1691 case 2:
1692 out_le16((u16 *)(addr + offset), val);
1693 break;
1694 default:
1695 out_le32((u32 *)(addr + offset), val);
1696 break;
1697 }
1698
1699 dcr_write(port->dcrs, DCRO_PEGPL_CFG, gpl_cfg);
1700
1701 return PCIBIOS_SUCCESSFUL;
1702 }
1703
1704 static struct pci_ops ppc4xx_pciex_pci_ops =
1705 {
1706 .read = ppc4xx_pciex_read_config,
1707 .write = ppc4xx_pciex_write_config,
1708 };
1709
ppc4xx_setup_one_pciex_POM(struct ppc4xx_pciex_port * port,struct pci_controller * hose,void __iomem * mbase,u64 plb_addr,u64 pci_addr,u64 size,unsigned int flags,int index)1710 static int __init ppc4xx_setup_one_pciex_POM(struct ppc4xx_pciex_port *port,
1711 struct pci_controller *hose,
1712 void __iomem *mbase,
1713 u64 plb_addr,
1714 u64 pci_addr,
1715 u64 size,
1716 unsigned int flags,
1717 int index)
1718 {
1719 u32 lah, lal, pciah, pcial, sa;
1720
1721 if (!is_power_of_2(size) ||
1722 (index < 2 && size < 0x100000) ||
1723 (index == 2 && size < 0x100) ||
1724 (plb_addr & (size - 1)) != 0) {
1725 printk(KERN_WARNING "%pOF: Resource out of range\n", hose->dn);
1726 return -1;
1727 }
1728
1729 /* Calculate register values */
1730 lah = RES_TO_U32_HIGH(plb_addr);
1731 lal = RES_TO_U32_LOW(plb_addr);
1732 pciah = RES_TO_U32_HIGH(pci_addr);
1733 pcial = RES_TO_U32_LOW(pci_addr);
1734 sa = (0xffffffffu << ilog2(size)) | 0x1;
1735
1736 /* Program register values */
1737 switch (index) {
1738 case 0:
1739 out_le32(mbase + PECFG_POM0LAH, pciah);
1740 out_le32(mbase + PECFG_POM0LAL, pcial);
1741 dcr_write(port->dcrs, DCRO_PEGPL_OMR1BAH, lah);
1742 dcr_write(port->dcrs, DCRO_PEGPL_OMR1BAL, lal);
1743 dcr_write(port->dcrs, DCRO_PEGPL_OMR1MSKH, 0x7fffffff);
1744 /*Enabled and single region */
1745 if (of_device_is_compatible(port->node, "ibm,plb-pciex-460sx"))
1746 dcr_write(port->dcrs, DCRO_PEGPL_OMR1MSKL,
1747 sa | DCRO_PEGPL_460SX_OMR1MSKL_UOT
1748 | DCRO_PEGPL_OMRxMSKL_VAL);
1749 else if (of_device_is_compatible(
1750 port->node, "ibm,plb-pciex-476fpe") ||
1751 of_device_is_compatible(
1752 port->node, "ibm,plb-pciex-476gtr"))
1753 dcr_write(port->dcrs, DCRO_PEGPL_OMR1MSKL,
1754 sa | DCRO_PEGPL_476FPE_OMR1MSKL_UOT
1755 | DCRO_PEGPL_OMRxMSKL_VAL);
1756 else
1757 dcr_write(port->dcrs, DCRO_PEGPL_OMR1MSKL,
1758 sa | DCRO_PEGPL_OMR1MSKL_UOT
1759 | DCRO_PEGPL_OMRxMSKL_VAL);
1760 break;
1761 case 1:
1762 out_le32(mbase + PECFG_POM1LAH, pciah);
1763 out_le32(mbase + PECFG_POM1LAL, pcial);
1764 dcr_write(port->dcrs, DCRO_PEGPL_OMR2BAH, lah);
1765 dcr_write(port->dcrs, DCRO_PEGPL_OMR2BAL, lal);
1766 dcr_write(port->dcrs, DCRO_PEGPL_OMR2MSKH, 0x7fffffff);
1767 dcr_write(port->dcrs, DCRO_PEGPL_OMR2MSKL,
1768 sa | DCRO_PEGPL_OMRxMSKL_VAL);
1769 break;
1770 case 2:
1771 out_le32(mbase + PECFG_POM2LAH, pciah);
1772 out_le32(mbase + PECFG_POM2LAL, pcial);
1773 dcr_write(port->dcrs, DCRO_PEGPL_OMR3BAH, lah);
1774 dcr_write(port->dcrs, DCRO_PEGPL_OMR3BAL, lal);
1775 dcr_write(port->dcrs, DCRO_PEGPL_OMR3MSKH, 0x7fffffff);
1776 /* Note that 3 here means enabled | IO space !!! */
1777 dcr_write(port->dcrs, DCRO_PEGPL_OMR3MSKL,
1778 sa | DCRO_PEGPL_OMR3MSKL_IO
1779 | DCRO_PEGPL_OMRxMSKL_VAL);
1780 break;
1781 }
1782
1783 return 0;
1784 }
1785
ppc4xx_configure_pciex_POMs(struct ppc4xx_pciex_port * port,struct pci_controller * hose,void __iomem * mbase)1786 static void __init ppc4xx_configure_pciex_POMs(struct ppc4xx_pciex_port *port,
1787 struct pci_controller *hose,
1788 void __iomem *mbase)
1789 {
1790 int i, j, found_isa_hole = 0;
1791
1792 /* Setup outbound memory windows */
1793 for (i = j = 0; i < 3; i++) {
1794 struct resource *res = &hose->mem_resources[i];
1795 resource_size_t offset = hose->mem_offset[i];
1796
1797 /* we only care about memory windows */
1798 if (!(res->flags & IORESOURCE_MEM))
1799 continue;
1800 if (j > 1) {
1801 printk(KERN_WARNING "%pOF: Too many ranges\n",
1802 port->node);
1803 break;
1804 }
1805
1806 /* Configure the resource */
1807 if (ppc4xx_setup_one_pciex_POM(port, hose, mbase,
1808 res->start,
1809 res->start - offset,
1810 resource_size(res),
1811 res->flags,
1812 j) == 0) {
1813 j++;
1814
1815 /* If the resource PCI address is 0 then we have our
1816 * ISA memory hole
1817 */
1818 if (res->start == offset)
1819 found_isa_hole = 1;
1820 }
1821 }
1822
1823 /* Handle ISA memory hole if not already covered */
1824 if (j <= 1 && !found_isa_hole && hose->isa_mem_size)
1825 if (ppc4xx_setup_one_pciex_POM(port, hose, mbase,
1826 hose->isa_mem_phys, 0,
1827 hose->isa_mem_size, 0, j) == 0)
1828 printk(KERN_INFO "%pOF: Legacy ISA memory support enabled\n",
1829 hose->dn);
1830
1831 /* Configure IO, always 64K starting at 0. We hard wire it to 64K !
1832 * Note also that it -has- to be region index 2 on this HW
1833 */
1834 if (hose->io_resource.flags & IORESOURCE_IO)
1835 ppc4xx_setup_one_pciex_POM(port, hose, mbase,
1836 hose->io_base_phys, 0,
1837 0x10000, IORESOURCE_IO, 2);
1838 }
1839
ppc4xx_configure_pciex_PIMs(struct ppc4xx_pciex_port * port,struct pci_controller * hose,void __iomem * mbase,struct resource * res)1840 static void __init ppc4xx_configure_pciex_PIMs(struct ppc4xx_pciex_port *port,
1841 struct pci_controller *hose,
1842 void __iomem *mbase,
1843 struct resource *res)
1844 {
1845 resource_size_t size = resource_size(res);
1846 u64 sa;
1847
1848 if (port->endpoint) {
1849 resource_size_t ep_addr = 0;
1850 resource_size_t ep_size = 32 << 20;
1851
1852 /* Currently we map a fixed 64MByte window to PLB address
1853 * 0 (SDRAM). This should probably be configurable via a dts
1854 * property.
1855 */
1856
1857 /* Calculate window size */
1858 sa = (0xffffffffffffffffull << ilog2(ep_size));
1859
1860 /* Setup BAR0 */
1861 out_le32(mbase + PECFG_BAR0HMPA, RES_TO_U32_HIGH(sa));
1862 out_le32(mbase + PECFG_BAR0LMPA, RES_TO_U32_LOW(sa) |
1863 PCI_BASE_ADDRESS_MEM_TYPE_64);
1864
1865 /* Disable BAR1 & BAR2 */
1866 out_le32(mbase + PECFG_BAR1MPA, 0);
1867 out_le32(mbase + PECFG_BAR2HMPA, 0);
1868 out_le32(mbase + PECFG_BAR2LMPA, 0);
1869
1870 out_le32(mbase + PECFG_PIM01SAH, RES_TO_U32_HIGH(sa));
1871 out_le32(mbase + PECFG_PIM01SAL, RES_TO_U32_LOW(sa));
1872
1873 out_le32(mbase + PCI_BASE_ADDRESS_0, RES_TO_U32_LOW(ep_addr));
1874 out_le32(mbase + PCI_BASE_ADDRESS_1, RES_TO_U32_HIGH(ep_addr));
1875 } else {
1876 /* Calculate window size */
1877 sa = (0xffffffffffffffffull << ilog2(size));
1878 if (res->flags & IORESOURCE_PREFETCH)
1879 sa |= PCI_BASE_ADDRESS_MEM_PREFETCH;
1880
1881 if (of_device_is_compatible(port->node, "ibm,plb-pciex-460sx") ||
1882 of_device_is_compatible(
1883 port->node, "ibm,plb-pciex-476fpe") ||
1884 of_device_is_compatible(
1885 port->node, "ibm,plb-pciex-476gtr"))
1886 sa |= PCI_BASE_ADDRESS_MEM_TYPE_64;
1887
1888 out_le32(mbase + PECFG_BAR0HMPA, RES_TO_U32_HIGH(sa));
1889 out_le32(mbase + PECFG_BAR0LMPA, RES_TO_U32_LOW(sa));
1890
1891 /* The setup of the split looks weird to me ... let's see
1892 * if it works
1893 */
1894 out_le32(mbase + PECFG_PIM0LAL, 0x00000000);
1895 out_le32(mbase + PECFG_PIM0LAH, 0x00000000);
1896 out_le32(mbase + PECFG_PIM1LAL, 0x00000000);
1897 out_le32(mbase + PECFG_PIM1LAH, 0x00000000);
1898 out_le32(mbase + PECFG_PIM01SAH, 0xffff0000);
1899 out_le32(mbase + PECFG_PIM01SAL, 0x00000000);
1900
1901 out_le32(mbase + PCI_BASE_ADDRESS_0, RES_TO_U32_LOW(res->start));
1902 out_le32(mbase + PCI_BASE_ADDRESS_1, RES_TO_U32_HIGH(res->start));
1903 }
1904
1905 /* Enable inbound mapping */
1906 out_le32(mbase + PECFG_PIMEN, 0x1);
1907
1908 /* Enable I/O, Mem, and Busmaster cycles */
1909 out_le16(mbase + PCI_COMMAND,
1910 in_le16(mbase + PCI_COMMAND) |
1911 PCI_COMMAND_IO | PCI_COMMAND_MEMORY | PCI_COMMAND_MASTER);
1912 }
1913
ppc4xx_pciex_port_setup_hose(struct ppc4xx_pciex_port * port)1914 static void __init ppc4xx_pciex_port_setup_hose(struct ppc4xx_pciex_port *port)
1915 {
1916 struct resource dma_window;
1917 struct pci_controller *hose = NULL;
1918 const int *bus_range;
1919 int primary = 0, busses;
1920 void __iomem *mbase = NULL, *cfg_data = NULL;
1921 const u32 *pval;
1922 u32 val;
1923
1924 /* Check if primary bridge */
1925 if (of_get_property(port->node, "primary", NULL))
1926 primary = 1;
1927
1928 /* Get bus range if any */
1929 bus_range = of_get_property(port->node, "bus-range", NULL);
1930
1931 /* Allocate the host controller data structure */
1932 hose = pcibios_alloc_controller(port->node);
1933 if (!hose)
1934 goto fail;
1935
1936 /* We stick the port number in "indirect_type" so the config space
1937 * ops can retrieve the port data structure easily
1938 */
1939 hose->indirect_type = port->index;
1940
1941 /* Get bus range */
1942 hose->first_busno = bus_range ? bus_range[0] : 0x0;
1943 hose->last_busno = bus_range ? bus_range[1] : 0xff;
1944
1945 /* Because of how big mapping the config space is (1M per bus), we
1946 * limit how many busses we support. In the long run, we could replace
1947 * that with something akin to kmap_atomic instead. We set aside 1 bus
1948 * for the host itself too.
1949 */
1950 busses = hose->last_busno - hose->first_busno; /* This is off by 1 */
1951 if (busses > MAX_PCIE_BUS_MAPPED) {
1952 busses = MAX_PCIE_BUS_MAPPED;
1953 hose->last_busno = hose->first_busno + busses;
1954 }
1955
1956 if (!port->endpoint) {
1957 /* Only map the external config space in cfg_data for
1958 * PCIe root-complexes. External space is 1M per bus
1959 */
1960 cfg_data = ioremap(port->cfg_space.start +
1961 (hose->first_busno + 1) * 0x100000,
1962 busses * 0x100000);
1963 if (cfg_data == NULL) {
1964 printk(KERN_ERR "%pOF: Can't map external config space !",
1965 port->node);
1966 goto fail;
1967 }
1968 hose->cfg_data = cfg_data;
1969 }
1970
1971 /* Always map the host config space in cfg_addr.
1972 * Internal space is 4K
1973 */
1974 mbase = ioremap(port->cfg_space.start + 0x10000000, 0x1000);
1975 if (mbase == NULL) {
1976 printk(KERN_ERR "%pOF: Can't map internal config space !",
1977 port->node);
1978 goto fail;
1979 }
1980 hose->cfg_addr = mbase;
1981
1982 pr_debug("PCIE %pOF, bus %d..%d\n", port->node,
1983 hose->first_busno, hose->last_busno);
1984 pr_debug(" config space mapped at: root @0x%p, other @0x%p\n",
1985 hose->cfg_addr, hose->cfg_data);
1986
1987 /* Setup config space */
1988 hose->ops = &ppc4xx_pciex_pci_ops;
1989 port->hose = hose;
1990 mbase = (void __iomem *)hose->cfg_addr;
1991
1992 if (!port->endpoint) {
1993 /*
1994 * Set bus numbers on our root port
1995 */
1996 out_8(mbase + PCI_PRIMARY_BUS, hose->first_busno);
1997 out_8(mbase + PCI_SECONDARY_BUS, hose->first_busno + 1);
1998 out_8(mbase + PCI_SUBORDINATE_BUS, hose->last_busno);
1999 }
2000
2001 /*
2002 * OMRs are already reset, also disable PIMs
2003 */
2004 out_le32(mbase + PECFG_PIMEN, 0);
2005
2006 /* Parse outbound mapping resources */
2007 pci_process_bridge_OF_ranges(hose, port->node, primary);
2008
2009 /* Parse inbound mapping resources */
2010 if (ppc4xx_parse_dma_ranges(hose, mbase, &dma_window) != 0)
2011 goto fail;
2012
2013 /* Configure outbound ranges POMs */
2014 ppc4xx_configure_pciex_POMs(port, hose, mbase);
2015
2016 /* Configure inbound ranges PIMs */
2017 ppc4xx_configure_pciex_PIMs(port, hose, mbase, &dma_window);
2018
2019 /* The root complex doesn't show up if we don't set some vendor
2020 * and device IDs into it. The defaults below are the same bogus
2021 * one that the initial code in arch/ppc had. This can be
2022 * overwritten by setting the "vendor-id/device-id" properties
2023 * in the pciex node.
2024 */
2025
2026 /* Get the (optional) vendor-/device-id from the device-tree */
2027 pval = of_get_property(port->node, "vendor-id", NULL);
2028 if (pval) {
2029 val = *pval;
2030 } else {
2031 if (!port->endpoint)
2032 val = 0xaaa0 + port->index;
2033 else
2034 val = 0xeee0 + port->index;
2035 }
2036 out_le16(mbase + 0x200, val);
2037
2038 pval = of_get_property(port->node, "device-id", NULL);
2039 if (pval) {
2040 val = *pval;
2041 } else {
2042 if (!port->endpoint)
2043 val = 0xbed0 + port->index;
2044 else
2045 val = 0xfed0 + port->index;
2046 }
2047 out_le16(mbase + 0x202, val);
2048
2049 /* Enable Bus master, memory, and io space */
2050 if (of_device_is_compatible(port->node, "ibm,plb-pciex-460sx"))
2051 out_le16(mbase + 0x204, 0x7);
2052
2053 if (!port->endpoint) {
2054 /* Set Class Code to PCI-PCI bridge and Revision Id to 1 */
2055 out_le32(mbase + 0x208, 0x06040001);
2056
2057 printk(KERN_INFO "PCIE%d: successfully set as root-complex\n",
2058 port->index);
2059 } else {
2060 /* Set Class Code to Processor/PPC */
2061 out_le32(mbase + 0x208, 0x0b200001);
2062
2063 printk(KERN_INFO "PCIE%d: successfully set as endpoint\n",
2064 port->index);
2065 }
2066
2067 return;
2068 fail:
2069 if (hose)
2070 pcibios_free_controller(hose);
2071 if (cfg_data)
2072 iounmap(cfg_data);
2073 if (mbase)
2074 iounmap(mbase);
2075 }
2076
ppc4xx_probe_pciex_bridge(struct device_node * np)2077 static void __init ppc4xx_probe_pciex_bridge(struct device_node *np)
2078 {
2079 struct ppc4xx_pciex_port *port;
2080 const u32 *pval;
2081 int portno;
2082 unsigned int dcrs;
2083
2084 /* First, proceed to core initialization as we assume there's
2085 * only one PCIe core in the system
2086 */
2087 if (ppc4xx_pciex_check_core_init(np))
2088 return;
2089
2090 /* Get the port number from the device-tree */
2091 pval = of_get_property(np, "port", NULL);
2092 if (pval == NULL) {
2093 printk(KERN_ERR "PCIE: Can't find port number for %pOF\n", np);
2094 return;
2095 }
2096 portno = *pval;
2097 if (portno >= ppc4xx_pciex_port_count) {
2098 printk(KERN_ERR "PCIE: port number out of range for %pOF\n",
2099 np);
2100 return;
2101 }
2102 port = &ppc4xx_pciex_ports[portno];
2103 port->index = portno;
2104
2105 /*
2106 * Check if device is enabled
2107 */
2108 if (!of_device_is_available(np)) {
2109 printk(KERN_INFO "PCIE%d: Port disabled via device-tree\n", port->index);
2110 return;
2111 }
2112
2113 port->node = of_node_get(np);
2114 if (ppc4xx_pciex_hwops->want_sdr) {
2115 pval = of_get_property(np, "sdr-base", NULL);
2116 if (pval == NULL) {
2117 printk(KERN_ERR "PCIE: missing sdr-base for %pOF\n",
2118 np);
2119 return;
2120 }
2121 port->sdr_base = *pval;
2122 }
2123
2124 /* Check if device_type property is set to "pci" or "pci-endpoint".
2125 * Resulting from this setup this PCIe port will be configured
2126 * as root-complex or as endpoint.
2127 */
2128 if (of_node_is_type(port->node, "pci-endpoint")) {
2129 port->endpoint = 1;
2130 } else if (of_node_is_type(port->node, "pci")) {
2131 port->endpoint = 0;
2132 } else {
2133 printk(KERN_ERR "PCIE: missing or incorrect device_type for %pOF\n",
2134 np);
2135 return;
2136 }
2137
2138 /* Fetch config space registers address */
2139 if (of_address_to_resource(np, 0, &port->cfg_space)) {
2140 printk(KERN_ERR "%pOF: Can't get PCI-E config space !", np);
2141 return;
2142 }
2143 /* Fetch host bridge internal registers address */
2144 if (of_address_to_resource(np, 1, &port->utl_regs)) {
2145 printk(KERN_ERR "%pOF: Can't get UTL register base !", np);
2146 return;
2147 }
2148
2149 /* Map DCRs */
2150 dcrs = dcr_resource_start(np, 0);
2151 if (dcrs == 0) {
2152 printk(KERN_ERR "%pOF: Can't get DCR register base !", np);
2153 return;
2154 }
2155 port->dcrs = dcr_map(np, dcrs, dcr_resource_len(np, 0));
2156
2157 /* Initialize the port specific registers */
2158 if (ppc4xx_pciex_port_init(port)) {
2159 printk(KERN_WARNING "PCIE%d: Port init failed\n", port->index);
2160 return;
2161 }
2162
2163 /* Setup the linux hose data structure */
2164 ppc4xx_pciex_port_setup_hose(port);
2165 }
2166
2167 #endif /* CONFIG_PPC4xx_PCI_EXPRESS */
2168
ppc4xx_pci_find_bridges(void)2169 static int __init ppc4xx_pci_find_bridges(void)
2170 {
2171 struct device_node *np;
2172
2173 pci_add_flags(PCI_ENABLE_PROC_DOMAINS | PCI_COMPAT_DOMAIN_0);
2174
2175 #ifdef CONFIG_PPC4xx_PCI_EXPRESS
2176 for_each_compatible_node(np, NULL, "ibm,plb-pciex")
2177 ppc4xx_probe_pciex_bridge(np);
2178 #endif
2179 for_each_compatible_node(np, NULL, "ibm,plb-pcix")
2180 ppc4xx_probe_pcix_bridge(np);
2181 for_each_compatible_node(np, NULL, "ibm,plb-pci")
2182 ppc4xx_probe_pci_bridge(np);
2183
2184 return 0;
2185 }
2186 arch_initcall(ppc4xx_pci_find_bridges);
2187
2188