blob: c44c1894cf76b1dd9522b58c1df26f2fc5af47da [file] [log] [blame]
Bruce Griffith27ed80b2014-08-15 11:46:25 -06001/*
2 * This file is part of the coreboot project.
3 *
4 * Copyright (C) 2012 Advanced Micro Devices, Inc.
5 *
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License as published by
8 * the Free Software Foundation; version 2 of the License.
9 *
10 * This program is distributed in the hope that it will be useful,
11 * but WITHOUT ANY WARRANTY; without even the implied warranty of
12 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
13 * GNU General Public License for more details.
14 *
15 * You should have received a copy of the GNU General Public License
16 * along with this program; if not, write to the Free Software
Patrick Georgib890a122015-03-26 15:17:45 +010017 * Foundation, Inc.
Bruce Griffith27ed80b2014-08-15 11:46:25 -060018 */
19
20#include <console/console.h>
21#include <arch/io.h>
22#include <arch/acpi.h>
23#include <stdint.h>
24#include <device/device.h>
25#include <device/pci.h>
26#include <device/pci_ids.h>
27#include <device/hypertransport.h>
28#include <stdlib.h>
29#include <string.h>
30#include <lib.h>
31#include <cpu/cpu.h>
32#include <cbmem.h>
33
34#include <Porting.h>
35#include <AGESA.h>
36#include <FieldAccessors.h>
37#include <Options.h>
38#include <Topology.h>
39#include <cpu/amd/amdfam16.h>
40#include <cpuRegisters.h>
Kyösti Mälkki023ed1f2014-10-22 08:05:36 +030041#include <northbridge/amd/pi/agesawrapper.h>
Kyösti Mälkkie4c17ce2014-10-21 18:22:32 +030042#include <northbridge/amd/pi/agesawrapper_call.h>
Bruce Griffith27ed80b2014-08-15 11:46:25 -060043#include "northbridge.h"
44
45#include <cpu/x86/lapic.h>
46#include <cpu/amd/mtrr.h>
Kyösti Mälkki0b5b5412014-11-26 08:11:07 +020047#include <arch/acpi.h>
48#include <arch/acpigen.h>
Bruce Griffith27ed80b2014-08-15 11:46:25 -060049
50#define MAX_NODE_NUMS (MAX_NODES * MAX_DIES)
51
Bruce Griffith27ed80b2014-08-15 11:46:25 -060052typedef struct dram_base_mask {
53 u32 base; //[47:27] at [28:8]
54 u32 mask; //[47:27] at [28:8] and enable at bit 0
55} dram_base_mask_t;
56
57static unsigned node_nums;
58static unsigned sblink;
59static device_t __f0_dev[MAX_NODE_NUMS];
60static device_t __f1_dev[MAX_NODE_NUMS];
61static device_t __f2_dev[MAX_NODE_NUMS];
62static device_t __f4_dev[MAX_NODE_NUMS];
63static unsigned fx_devs = 0;
64
65static dram_base_mask_t get_dram_base_mask(u32 nodeid)
66{
67 device_t dev;
68 dram_base_mask_t d;
69 dev = __f1_dev[0];
70 u32 temp;
71 temp = pci_read_config32(dev, 0x44 + (nodeid << 3)); //[39:24] at [31:16]
72 d.mask = ((temp & 0xfff80000)>>(8+3)); // mask out DramMask [26:24] too
73 temp = pci_read_config32(dev, 0x144 + (nodeid <<3)) & 0xff; //[47:40] at [7:0]
74 d.mask |= temp<<21;
75 temp = pci_read_config32(dev, 0x40 + (nodeid << 3)); //[39:24] at [31:16]
76 d.mask |= (temp & 1); // enable bit
77 d.base = ((temp & 0xfff80000)>>(8+3)); // mask out DramBase [26:24) too
78 temp = pci_read_config32(dev, 0x140 + (nodeid <<3)) & 0xff; //[47:40] at [7:0]
79 d.base |= temp<<21;
80 return d;
81}
82
83static void set_io_addr_reg(device_t dev, u32 nodeid, u32 linkn, u32 reg,
84 u32 io_min, u32 io_max)
85{
86 u32 i;
87 u32 tempreg;
88 /* io range allocation */
89 tempreg = (nodeid&0xf) | ((nodeid & 0x30)<<(8-4)) | (linkn<<4) | ((io_max&0xf0)<<(12-4)); //limit
90 for (i=0; i<node_nums; i++)
91 pci_write_config32(__f1_dev[i], reg+4, tempreg);
92 tempreg = 3 /*| ( 3<<4)*/ | ((io_min&0xf0)<<(12-4)); //base :ISA and VGA ?
93#if 0
94 // FIXME: can we use VGA reg instead?
95 if (dev->link[link].bridge_ctrl & PCI_BRIDGE_CTL_VGA) {
96 printk(BIOS_SPEW, "%s, enabling legacy VGA IO forwarding for %s link %s\n",
97 __func__, dev_path(dev), link);
98 tempreg |= PCI_IO_BASE_VGA_EN;
99 }
100 if (dev->link[link].bridge_ctrl & PCI_BRIDGE_CTL_NO_ISA) {
101 tempreg |= PCI_IO_BASE_NO_ISA;
102 }
103#endif
104 for (i=0; i<node_nums; i++)
105 pci_write_config32(__f1_dev[i], reg, tempreg);
106}
107
108static void set_mmio_addr_reg(u32 nodeid, u32 linkn, u32 reg, u32 index, u32 mmio_min, u32 mmio_max, u32 nodes)
109{
110 u32 i;
111 u32 tempreg;
112 /* io range allocation */
113 tempreg = (nodeid&0xf) | (linkn<<4) | (mmio_max&0xffffff00); //limit
114 for (i=0; i<nodes; i++)
115 pci_write_config32(__f1_dev[i], reg+4, tempreg);
116 tempreg = 3 | (nodeid & 0x30) | (mmio_min&0xffffff00);
117 for (i=0; i<node_nums; i++)
118 pci_write_config32(__f1_dev[i], reg, tempreg);
119}
120
121static device_t get_node_pci(u32 nodeid, u32 fn)
122{
123#if MAX_NODE_NUMS + CONFIG_CDB >= 32
124 if ((CONFIG_CDB + nodeid) < 32) {
125 return dev_find_slot(CONFIG_CBB, PCI_DEVFN(CONFIG_CDB + nodeid, fn));
126 } else {
127 return dev_find_slot(CONFIG_CBB-1, PCI_DEVFN(CONFIG_CDB + nodeid - 32, fn));
128 }
129#else
130 return dev_find_slot(CONFIG_CBB, PCI_DEVFN(CONFIG_CDB + nodeid, fn));
131#endif
132}
133
134static void get_fx_devs(void)
135{
136 int i;
137 for (i = 0; i < MAX_NODE_NUMS; i++) {
138 __f0_dev[i] = get_node_pci(i, 0);
139 __f1_dev[i] = get_node_pci(i, 1);
140 __f2_dev[i] = get_node_pci(i, 2);
141 __f4_dev[i] = get_node_pci(i, 4);
142 if (__f0_dev[i] != NULL && __f1_dev[i] != NULL)
143 fx_devs = i+1;
144 }
145 if (__f1_dev[0] == NULL || __f0_dev[0] == NULL || fx_devs == 0) {
146 die("Cannot find 0:0x18.[0|1]\n");
147 }
148 printk(BIOS_DEBUG, "fx_devs=0x%x\n", fx_devs);
149}
150
151static u32 f1_read_config32(unsigned reg)
152{
153 if (fx_devs == 0)
154 get_fx_devs();
155 return pci_read_config32(__f1_dev[0], reg);
156}
157
158static void f1_write_config32(unsigned reg, u32 value)
159{
160 int i;
161 if (fx_devs == 0)
162 get_fx_devs();
163 for(i = 0; i < fx_devs; i++) {
164 device_t dev;
165 dev = __f1_dev[i];
166 if (dev && dev->enabled) {
167 pci_write_config32(dev, reg, value);
168 }
169 }
170}
171
172static u32 amdfam16_nodeid(device_t dev)
173{
174#if MAX_NODE_NUMS == 64
175 unsigned busn;
176 busn = dev->bus->secondary;
177 if (busn != CONFIG_CBB) {
178 return (dev->path.pci.devfn >> 3) - CONFIG_CDB + 32;
179 } else {
180 return (dev->path.pci.devfn >> 3) - CONFIG_CDB;
181 }
182
183#else
184 return (dev->path.pci.devfn >> 3) - CONFIG_CDB;
185#endif
186}
187
188static void set_vga_enable_reg(u32 nodeid, u32 linkn)
189{
190 u32 val;
191
192 val = 1 | (nodeid<<4) | (linkn<<12);
193 /* it will routing
194 * (1)mmio 0xa0000:0xbffff
195 * (2)io 0x3b0:0x3bb, 0x3c0:0x3df
196 */
197 f1_write_config32(0xf4, val);
198
199}
200
201/**
202 * @return
203 * @retval 2 resoure does not exist, usable
204 * @retval 0 resource exists, not usable
205 * @retval 1 resource exist, resource has been allocated before
206 */
207static int reg_useable(unsigned reg, device_t goal_dev, unsigned goal_nodeid,
208 unsigned goal_link)
209{
210 struct resource *res;
211 unsigned nodeid, link = 0;
212 int result;
213 res = 0;
214 for (nodeid = 0; !res && (nodeid < fx_devs); nodeid++) {
215 device_t dev;
216 dev = __f0_dev[nodeid];
217 if (!dev)
218 continue;
219 for (link = 0; !res && (link < 8); link++) {
220 res = probe_resource(dev, IOINDEX(0x1000 + reg, link));
221 }
222 }
223 result = 2;
224 if (res) {
225 result = 0;
226 if ((goal_link == (link - 1)) &&
227 (goal_nodeid == (nodeid - 1)) &&
228 (res->flags <= 1)) {
229 result = 1;
230 }
231 }
232 return result;
233}
234
235static struct resource *amdfam16_find_iopair(device_t dev, unsigned nodeid, unsigned link)
236{
237 struct resource *resource;
238 u32 free_reg, reg;
239 resource = 0;
240 free_reg = 0;
241 for (reg = 0xc0; reg <= 0xd8; reg += 0x8) {
242 int result;
243 result = reg_useable(reg, dev, nodeid, link);
244 if (result == 1) {
245 /* I have been allocated this one */
246 break;
247 }
248 else if (result > 1) {
249 /* I have a free register pair */
250 free_reg = reg;
251 }
252 }
253 if (reg > 0xd8) {
254 reg = free_reg; // if no free, the free_reg still be 0
255 }
256
257 resource = new_resource(dev, IOINDEX(0x1000 + reg, link));
258
259 return resource;
260}
261
262static struct resource *amdfam16_find_mempair(device_t dev, u32 nodeid, u32 link)
263{
264 struct resource *resource;
265 u32 free_reg, reg;
266 resource = 0;
267 free_reg = 0;
268 for (reg = 0x80; reg <= 0xb8; reg += 0x8) {
269 int result;
270 result = reg_useable(reg, dev, nodeid, link);
271 if (result == 1) {
272 /* I have been allocated this one */
273 break;
274 }
275 else if (result > 1) {
276 /* I have a free register pair */
277 free_reg = reg;
278 }
279 }
280 if (reg > 0xb8) {
281 reg = free_reg;
282 }
283
284 resource = new_resource(dev, IOINDEX(0x1000 + reg, link));
285 return resource;
286}
287
288static void amdfam16_link_read_bases(device_t dev, u32 nodeid, u32 link)
289{
290 struct resource *resource;
291
292 /* Initialize the io space constraints on the current bus */
293 resource = amdfam16_find_iopair(dev, nodeid, link);
294 if (resource) {
295 u32 align;
296 align = log2(HT_IO_HOST_ALIGN);
297 resource->base = 0;
298 resource->size = 0;
299 resource->align = align;
300 resource->gran = align;
301 resource->limit = 0xffffUL;
302 resource->flags = IORESOURCE_IO | IORESOURCE_BRIDGE;
303 }
304
305 /* Initialize the prefetchable memory constraints on the current bus */
306 resource = amdfam16_find_mempair(dev, nodeid, link);
307 if (resource) {
308 resource->base = 0;
309 resource->size = 0;
310 resource->align = log2(HT_MEM_HOST_ALIGN);
311 resource->gran = log2(HT_MEM_HOST_ALIGN);
312 resource->limit = 0xffffffffffULL;
313 resource->flags = IORESOURCE_MEM | IORESOURCE_PREFETCH;
314 resource->flags |= IORESOURCE_BRIDGE;
315 }
316
317 /* Initialize the memory constraints on the current bus */
318 resource = amdfam16_find_mempair(dev, nodeid, link);
319 if (resource) {
320 resource->base = 0;
321 resource->size = 0;
322 resource->align = log2(HT_MEM_HOST_ALIGN);
323 resource->gran = log2(HT_MEM_HOST_ALIGN);
324 resource->limit = 0xffffffffffULL;
325 resource->flags = IORESOURCE_MEM | IORESOURCE_BRIDGE;
326 }
327
328}
329
330static void read_resources(device_t dev)
331{
332 u32 nodeid;
333 struct bus *link;
334
335 nodeid = amdfam16_nodeid(dev);
336 for (link = dev->link_list; link; link = link->next) {
337 if (link->children) {
338 amdfam16_link_read_bases(dev, nodeid, link->link_num);
339 }
340 }
341}
342
343static void set_resource(device_t dev, struct resource *resource, u32 nodeid)
344{
345 resource_t rbase, rend;
346 unsigned reg, link_num;
347 char buf[50];
348
349 /* Make certain the resource has actually been set */
350 if (!(resource->flags & IORESOURCE_ASSIGNED)) {
351 return;
352 }
353
354 /* If I have already stored this resource don't worry about it */
355 if (resource->flags & IORESOURCE_STORED) {
356 return;
357 }
358
359 /* Only handle PCI memory and IO resources */
360 if (!(resource->flags & (IORESOURCE_MEM | IORESOURCE_IO)))
361 return;
362
363 /* Ensure I am actually looking at a resource of function 1 */
364 if ((resource->index & 0xffff) < 0x1000) {
365 return;
366 }
367 /* Get the base address */
368 rbase = resource->base;
369
370 /* Get the limit (rounded up) */
371 rend = resource_end(resource);
372
373 /* Get the register and link */
374 reg = resource->index & 0xfff; // 4k
375 link_num = IOINDEX_LINK(resource->index);
376
377 if (resource->flags & IORESOURCE_IO) {
378 set_io_addr_reg(dev, nodeid, link_num, reg, rbase>>8, rend>>8);
379 }
380 else if (resource->flags & IORESOURCE_MEM) {
381 set_mmio_addr_reg(nodeid, link_num, reg, (resource->index >>24), rbase>>8, rend>>8, node_nums) ;// [39:8]
382 }
383 resource->flags |= IORESOURCE_STORED;
384 snprintf(buf, sizeof (buf), " <node %x link %x>",
385 nodeid, link_num);
386 report_resource_stored(dev, resource, buf);
387}
388
389/**
390 * I tried to reuse the resource allocation code in set_resource()
391 * but it is too difficult to deal with the resource allocation magic.
392 */
393
394static void create_vga_resource(device_t dev, unsigned nodeid)
395{
396 struct bus *link;
397
398 /* find out which link the VGA card is connected,
399 * we only deal with the 'first' vga card */
400 for (link = dev->link_list; link; link = link->next) {
401 if (link->bridge_ctrl & PCI_BRIDGE_CTL_VGA) {
402#if CONFIG_MULTIPLE_VGA_ADAPTERS
403 extern device_t vga_pri; // the primary vga device, defined in device.c
404 printk(BIOS_DEBUG, "VGA: vga_pri bus num = %d bus range [%d,%d]\n", vga_pri->bus->secondary,
405 link->secondary,link->subordinate);
406 /* We need to make sure the vga_pri is under the link */
407 if((vga_pri->bus->secondary >= link->secondary ) &&
408 (vga_pri->bus->secondary <= link->subordinate )
409 )
410#endif
411 break;
412 }
413 }
414
415 /* no VGA card installed */
416 if (link == NULL)
417 return;
418
419 printk(BIOS_DEBUG, "VGA: %s (aka node %d) link %d has VGA device\n", dev_path(dev), nodeid, sblink);
420 set_vga_enable_reg(nodeid, sblink);
421}
422
423static void set_resources(device_t dev)
424{
425 unsigned nodeid;
426 struct bus *bus;
427 struct resource *res;
428
429 /* Find the nodeid */
430 nodeid = amdfam16_nodeid(dev);
431
432 create_vga_resource(dev, nodeid); //TODO: do we need this?
433
434 /* Set each resource we have found */
435 for (res = dev->resource_list; res; res = res->next) {
436 set_resource(dev, res, nodeid);
437 }
438
439 for (bus = dev->link_list; bus; bus = bus->next) {
440 if (bus->children) {
441 assign_resources(bus);
442 }
443 }
444}
445
446static void northbridge_init(struct device *dev)
447{
448}
Kyösti Mälkki0b5b5412014-11-26 08:11:07 +0200449
Vladimir Serbinenko807127f2014-11-09 13:36:18 +0100450static unsigned long acpi_fill_hest(acpi_hest_t *hest)
Kyösti Mälkki0b5b5412014-11-26 08:11:07 +0200451{
452 void *addr, *current;
453
454 /* Skip the HEST header. */
455 current = (void *)(hest + 1);
456
457 addr = agesawrapper_getlateinitptr(PICK_WHEA_MCE);
458 if (addr != NULL)
459 current += acpi_create_hest_error_source(hest, current, 0, (void *)((u32)addr + 2), *(UINT16 *)addr - 2);
460
461 addr = agesawrapper_getlateinitptr(PICK_WHEA_CMC);
462 if (addr != NULL)
463 current += acpi_create_hest_error_source(hest, current, 1, (void *)((u32)addr + 2), *(UINT16 *)addr - 2);
464
465 return (unsigned long)current;
466}
467
Kyösti Mälkki0b5b5412014-11-26 08:11:07 +0200468static void northbridge_fill_ssdt_generator(void)
469{
470 msr_t msr;
471 char pscope[] = "\\_SB.PCI0";
472
473 acpigen_write_scope(pscope);
474 msr = rdmsr(TOP_MEM);
475 acpigen_write_name_dword("TOM1", msr.lo);
476 msr = rdmsr(TOP_MEM2);
477 /*
478 * Since XP only implements parts of ACPI 2.0, we can't use a qword
479 * here.
480 * See http://www.acpi.info/presentations/S01USMOBS169_OS%2520new.ppt
481 * slide 22ff.
482 * Shift value right by 20 bit to make it fit into 32bit,
483 * giving us 1MB granularity and a limit of almost 4Exabyte of memory.
484 */
485 acpigen_write_name_dword("TOM2", (msr.hi << 12) | msr.lo >> 20);
486 acpigen_pop_len();
487}
488
489static unsigned long agesa_write_acpi_tables(unsigned long current,
490 acpi_rsdp_t *rsdp)
491{
492 acpi_srat_t *srat;
493 acpi_slit_t *slit;
494 acpi_header_t *ssdt;
495 acpi_header_t *alib;
496 acpi_header_t *ivrs;
497 acpi_hest_t *hest;
498
499 /* HEST */
500 current = ALIGN(current, 8);
501 hest = (acpi_hest_t *)current;
Vladimir Serbinenko807127f2014-11-09 13:36:18 +0100502 acpi_write_hest((void *)current, acpi_fill_hest);
Kyösti Mälkki0b5b5412014-11-26 08:11:07 +0200503 acpi_add_table(rsdp, (void *)current);
504 current += ((acpi_header_t *)current)->length;
505
506 current = ALIGN(current, 8);
507 printk(BIOS_DEBUG, "ACPI: * IVRS at %lx\n", current);
508 ivrs = agesawrapper_getlateinitptr(PICK_IVRS);
509 if (ivrs != NULL) {
510 memcpy((void *)current, ivrs, ivrs->length);
511 ivrs = (acpi_header_t *) current;
512 current += ivrs->length;
513 acpi_add_table(rsdp, ivrs);
514 } else {
515 printk(BIOS_DEBUG, " AGESA IVRS table NULL. Skipping.\n");
516 }
517
518 /* SRAT */
519 current = ALIGN(current, 8);
520 printk(BIOS_DEBUG, "ACPI: * SRAT at %lx\n", current);
521 srat = (acpi_srat_t *) agesawrapper_getlateinitptr (PICK_SRAT);
522 if (srat != NULL) {
523 memcpy((void *)current, srat, srat->header.length);
524 srat = (acpi_srat_t *) current;
525 current += srat->header.length;
526 acpi_add_table(rsdp, srat);
527 } else {
528 printk(BIOS_DEBUG, " AGESA SRAT table NULL. Skipping.\n");
529 }
530
531 /* SLIT */
532 current = ALIGN(current, 8);
533 printk(BIOS_DEBUG, "ACPI: * SLIT at %lx\n", current);
534 slit = (acpi_slit_t *) agesawrapper_getlateinitptr (PICK_SLIT);
535 if (slit != NULL) {
536 memcpy((void *)current, slit, slit->header.length);
537 slit = (acpi_slit_t *) current;
538 current += slit->header.length;
539 acpi_add_table(rsdp, slit);
540 } else {
541 printk(BIOS_DEBUG, " AGESA SLIT table NULL. Skipping.\n");
542 }
543
544 /* ALIB */
545 current = ALIGN(current, 16);
546 printk(BIOS_DEBUG, "ACPI: * AGESA ALIB SSDT at %lx\n", current);
547 alib = (acpi_header_t *)agesawrapper_getlateinitptr (PICK_ALIB);
548 if (alib != NULL) {
549 memcpy((void *)current, alib, alib->length);
550 alib = (acpi_header_t *) current;
551 current += alib->length;
552 acpi_add_table(rsdp, (void *)alib);
553 }
554 else {
555 printk(BIOS_DEBUG, " AGESA ALIB SSDT table NULL. Skipping.\n");
556 }
557
558 /* this pstate ssdt may cause Blue Screen: Fixed: Keep this comment for a while. */
559 /* SSDT */
560 current = ALIGN(current, 16);
561 printk(BIOS_DEBUG, "ACPI: * SSDT at %lx\n", current);
562 ssdt = (acpi_header_t *)agesawrapper_getlateinitptr (PICK_PSTATE);
563 if (ssdt != NULL) {
564 memcpy((void *)current, ssdt, ssdt->length);
565 ssdt = (acpi_header_t *) current;
566 current += ssdt->length;
567 }
568 else {
569 printk(BIOS_DEBUG, " AGESA PState table NULL. Skipping.\n");
570 }
571 acpi_add_table(rsdp,ssdt);
572
573 printk(BIOS_DEBUG, "ACPI: * SSDT for PState at %lx\n", current);
574 return current;
575}
576
Bruce Griffith27ed80b2014-08-15 11:46:25 -0600577static struct device_operations northbridge_operations = {
578 .read_resources = read_resources,
579 .set_resources = set_resources,
580 .enable_resources = pci_dev_enable_resources,
581 .init = northbridge_init,
Kyösti Mälkki0b5b5412014-11-26 08:11:07 +0200582 .acpi_fill_ssdt_generator = northbridge_fill_ssdt_generator,
583 .write_acpi_tables = agesa_write_acpi_tables,
Bruce Griffith27ed80b2014-08-15 11:46:25 -0600584 .enable = 0,
585 .ops_pci = 0,
586};
587
588static const struct pci_driver family16_northbridge __pci_driver = {
589 .ops = &northbridge_operations,
590 .vendor = PCI_VENDOR_ID_AMD,
591 .device = PCI_DEVICE_ID_AMD_16H_MODEL_003F_NB_HT,
592};
593
594static const struct pci_driver family10_northbridge __pci_driver = {
595 .ops = &northbridge_operations,
596 .vendor = PCI_VENDOR_ID_AMD,
597 .device = PCI_DEVICE_ID_AMD_10H_NB_HT,
598};
599
Dave Frodin891f71a2015-01-19 15:58:24 -0700600static void fam16_finalize(void *chip_info)
601{
602 device_t dev;
603 u32 value;
604 dev = dev_find_slot(0, PCI_DEVFN(0, 0)); /* clear IoapicSbFeatureEn */
605 pci_write_config32(dev, 0xF8, 0);
606 pci_write_config32(dev, 0xFC, 5); /* TODO: move it to dsdt.asl */
607
608 /* disable No Snoop */
609 dev = dev_find_slot(0, PCI_DEVFN(1, 1));
610 value = pci_read_config32(dev, 0x60);
611 value &= ~(1 << 11);
612 pci_write_config32(dev, 0x60, value);
613}
614
Kyösti Mälkkie4c17ce2014-10-21 18:22:32 +0300615struct chip_operations northbridge_amd_pi_00730F01_ops = {
Bruce Griffith27ed80b2014-08-15 11:46:25 -0600616 CHIP_NAME("AMD FAM16 Northbridge")
617 .enable_dev = 0,
Dave Frodin891f71a2015-01-19 15:58:24 -0700618 .final = fam16_finalize,
Bruce Griffith27ed80b2014-08-15 11:46:25 -0600619};
620
621static void domain_read_resources(device_t dev)
622{
623 unsigned reg;
624
625 /* Find the already assigned resource pairs */
626 get_fx_devs();
627 for (reg = 0x80; reg <= 0xd8; reg+= 0x08) {
628 u32 base, limit;
629 base = f1_read_config32(reg);
630 limit = f1_read_config32(reg + 0x04);
631 /* Is this register allocated? */
632 if ((base & 3) != 0) {
633 unsigned nodeid, reg_link;
634 device_t reg_dev;
635 if (reg<0xc0) { // mmio
636 nodeid = (limit & 0xf) + (base&0x30);
637 } else { // io
638 nodeid = (limit & 0xf) + ((base>>4)&0x30);
639 }
640 reg_link = (limit >> 4) & 7;
641 reg_dev = __f0_dev[nodeid];
642 if (reg_dev) {
643 /* Reserve the resource */
644 struct resource *res;
645 res = new_resource(reg_dev, IOINDEX(0x1000 + reg, reg_link));
646 if (res) {
647 res->flags = 1;
648 }
649 }
650 }
651 }
652 /* FIXME: do we need to check extend conf space?
653 I don't believe that much preset value */
654
655#if !CONFIG_PCI_64BIT_PREF_MEM
656 pci_domain_read_resources(dev);
657
658#else
659 struct bus *link;
660 struct resource *resource;
661 for (link=dev->link_list; link; link = link->next) {
662 /* Initialize the system wide io space constraints */
663 resource = new_resource(dev, 0|(link->link_num<<2));
664 resource->base = 0x400;
665 resource->limit = 0xffffUL;
666 resource->flags = IORESOURCE_IO;
667
668 /* Initialize the system wide prefetchable memory resources constraints */
669 resource = new_resource(dev, 1|(link->link_num<<2));
670 resource->limit = 0xfcffffffffULL;
671 resource->flags = IORESOURCE_MEM | IORESOURCE_PREFETCH;
672
673 /* Initialize the system wide memory resources constraints */
674 resource = new_resource(dev, 2|(link->link_num<<2));
675 resource->limit = 0xfcffffffffULL;
676 resource->flags = IORESOURCE_MEM;
677 }
678#endif
679}
680
681static void domain_enable_resources(device_t dev)
682{
683 if (acpi_is_wakeup_s3())
684 AGESAWRAPPER(fchs3laterestore);
685
686 /* Must be called after PCI enumeration and resource allocation */
687 if (!acpi_is_wakeup_s3())
688 AGESAWRAPPER(amdinitmid);
689
690 printk(BIOS_DEBUG, " ader - leaving domain_enable_resources.\n");
691}
692
693#if CONFIG_HW_MEM_HOLE_SIZEK != 0
694struct hw_mem_hole_info {
695 unsigned hole_startk;
696 int node_id;
697};
698static struct hw_mem_hole_info get_hw_mem_hole_info(void)
699{
700 struct hw_mem_hole_info mem_hole;
701 int i;
702 mem_hole.hole_startk = CONFIG_HW_MEM_HOLE_SIZEK;
703 mem_hole.node_id = -1;
704 for (i = 0; i < node_nums; i++) {
705 dram_base_mask_t d;
706 u32 hole;
707 d = get_dram_base_mask(i);
708 if (!(d.mask & 1)) continue; // no memory on this node
709 hole = pci_read_config32(__f1_dev[i], 0xf0);
710 if (hole & 2) { // we find the hole
711 mem_hole.hole_startk = (hole & (0xff<<24)) >> 10;
712 mem_hole.node_id = i; // record the node No with hole
713 break; // only one hole
714 }
715 }
716
717 /* We need to double check if there is special set on base reg and limit reg
718 * are not continuous instead of hole, it will find out its hole_startk.
719 */
720 if (mem_hole.node_id == -1) {
721 resource_t limitk_pri = 0;
722 for (i=0; i<node_nums; i++) {
723 dram_base_mask_t d;
724 resource_t base_k, limit_k;
725 d = get_dram_base_mask(i);
726 if (!(d.base & 1)) continue;
727 base_k = ((resource_t)(d.base & 0x1fffff00)) <<9;
728 if (base_k > 4 *1024 * 1024) break; // don't need to go to check
729 if (limitk_pri != base_k) { // we find the hole
730 mem_hole.hole_startk = (unsigned)limitk_pri; // must beblow 4G
731 mem_hole.node_id = i;
732 break; //only one hole
733 }
734 limit_k = ((resource_t)(((d.mask & ~1) + 0x000FF) & 0x1fffff00)) << 9;
735 limitk_pri = limit_k;
736 }
737 }
738 return mem_hole;
739}
740#endif
741
742#define ONE_MB_SHIFT 20
743#define ONE_GB_SHIFT 30
744
745static void setup_uma_memory(void)
746{
747#if CONFIG_GFXUMA
748 uint64_t topmem = bsp_topmem();
749 uint64_t topmem2 = bsp_topmem2();
750 uint32_t sysmem_mb, sysmem_gb;
751
752 /* refer to UMA_AUTO size computation in Family16h BKDG. */
753 /* Please reference MemNGetUmaSizeML() */
754 /*
755 * Total system memory UMASize
756 * >= 6G 1024M
757 * >= 4G 512M
758 * >= 2G 256M
759 * < 2G 128M
760 */
761
762 sysmem_mb = (topmem + (16ull << ONE_MB_SHIFT)) >> ONE_MB_SHIFT; // Ignore 16MB allocated for C6 when finding UMA size
763 sysmem_mb += topmem2 ? ((topmem2 >> ONE_MB_SHIFT) - 4096) : 0;
764 sysmem_gb = sysmem_mb >> (ONE_GB_SHIFT - ONE_MB_SHIFT);
765 printk(BIOS_SPEW, "%s: system memory size %luGB, topmem2 size %lluMB, topmem size %lluMB\n", __func__, (unsigned long)sysmem_gb, (topmem2 >> ONE_MB_SHIFT), (topmem >> ONE_MB_SHIFT));
766 if (sysmem_gb >= 6) {
767 uma_memory_size = 1024 << ONE_MB_SHIFT;
768 } else if (sysmem_gb >= 4) {
769 uma_memory_size = 512 << ONE_MB_SHIFT;
770 } else if (sysmem_gb >= 2) {
771 uma_memory_size = 256 << ONE_MB_SHIFT;
772 } else {
773 uma_memory_size = 128 << ONE_MB_SHIFT;
774 }
775 uma_memory_base = topmem - uma_memory_size; /* TOP_MEM1 */
776
777 printk(BIOS_INFO, "%s: uma size 0x%08llx, memory start 0x%08llx\n",
778 __func__, uma_memory_size, uma_memory_base);
779
780 /* TODO: TOP_MEM2 */
781#endif
782}
783
784
785static void domain_set_resources(device_t dev)
786{
787#if CONFIG_PCI_64BIT_PREF_MEM
788 struct resource *io, *mem1, *mem2;
789 struct resource *res;
790#endif
791 unsigned long mmio_basek;
792 u32 pci_tolm;
793 u64 ramtop = 0;
794 int i, idx;
795 struct bus *link;
796#if CONFIG_HW_MEM_HOLE_SIZEK != 0
797 struct hw_mem_hole_info mem_hole;
798 u32 reset_memhole = 1;
799#endif
800
801#if CONFIG_PCI_64BIT_PREF_MEM
802
803 for (link = dev->link_list; link; link = link->next) {
804 /* Now reallocate the pci resources memory with the
805 * highest addresses I can manage.
806 */
807 mem1 = find_resource(dev, 1|(link->link_num<<2));
808 mem2 = find_resource(dev, 2|(link->link_num<<2));
809
810 printk(BIOS_DEBUG, "base1: 0x%08Lx limit1: 0x%08Lx size: 0x%08Lx align: %d\n",
811 mem1->base, mem1->limit, mem1->size, mem1->align);
812 printk(BIOS_DEBUG, "base2: 0x%08Lx limit2: 0x%08Lx size: 0x%08Lx align: %d\n",
813 mem2->base, mem2->limit, mem2->size, mem2->align);
814
815 /* See if both resources have roughly the same limits */
816 if (((mem1->limit <= 0xffffffff) && (mem2->limit <= 0xffffffff)) ||
817 ((mem1->limit > 0xffffffff) && (mem2->limit > 0xffffffff)))
818 {
819 /* If so place the one with the most stringent alignment first */
820 if (mem2->align > mem1->align) {
821 struct resource *tmp;
822 tmp = mem1;
823 mem1 = mem2;
824 mem2 = tmp;
825 }
826 /* Now place the memory as high up as it will go */
827 mem2->base = resource_max(mem2);
828 mem1->limit = mem2->base - 1;
829 mem1->base = resource_max(mem1);
830 }
831 else {
832 /* Place the resources as high up as they will go */
833 mem2->base = resource_max(mem2);
834 mem1->base = resource_max(mem1);
835 }
836
837 printk(BIOS_DEBUG, "base1: 0x%08Lx limit1: 0x%08Lx size: 0x%08Lx align: %d\n",
838 mem1->base, mem1->limit, mem1->size, mem1->align);
839 printk(BIOS_DEBUG, "base2: 0x%08Lx limit2: 0x%08Lx size: 0x%08Lx align: %d\n",
840 mem2->base, mem2->limit, mem2->size, mem2->align);
841 }
842
843 for (res = &dev->resource_list; res; res = res->next)
844 {
845 res->flags |= IORESOURCE_ASSIGNED;
846 res->flags |= IORESOURCE_STORED;
847 report_resource_stored(dev, res, "");
848 }
849#endif
850
851 pci_tolm = 0xffffffffUL;
852 for (link = dev->link_list; link; link = link->next) {
853 pci_tolm = find_pci_tolm(link);
854 }
855
856 // FIXME handle interleaved nodes. If you fix this here, please fix
857 // amdk8, too.
858 mmio_basek = pci_tolm >> 10;
859 /* Round mmio_basek to something the processor can support */
860 mmio_basek &= ~((1 << 6) -1);
861
862 // FIXME improve mtrr.c so we don't use up all of the mtrrs with a 64M
863 // MMIO hole. If you fix this here, please fix amdk8, too.
864 /* Round the mmio hole to 64M */
865 mmio_basek &= ~((64*1024) - 1);
866
867#if CONFIG_HW_MEM_HOLE_SIZEK != 0
868 /* if the hw mem hole is already set in raminit stage, here we will compare
869 * mmio_basek and hole_basek. if mmio_basek is bigger that hole_basek and will
870 * use hole_basek as mmio_basek and we don't need to reset hole.
871 * otherwise We reset the hole to the mmio_basek
872 */
873
874 mem_hole = get_hw_mem_hole_info();
875
876 // Use hole_basek as mmio_basek, and we don't need to reset hole anymore
877 if ((mem_hole.node_id != -1) && (mmio_basek > mem_hole.hole_startk)) {
878 mmio_basek = mem_hole.hole_startk;
879 reset_memhole = 0;
880 }
881#endif
882
883 idx = 0x10;
884 for (i = 0; i < node_nums; i++) {
885 dram_base_mask_t d;
886 resource_t basek, limitk, sizek; // 4 1T
887
888 d = get_dram_base_mask(i);
889
890 if (!(d.mask & 1)) continue;
891 basek = ((resource_t)(d.base & 0x1fffff00)) << 9; // could overflow, we may lost 6 bit here
892 limitk = ((resource_t)(((d.mask & ~1) + 0x000FF) & 0x1fffff00)) << 9 ;
893
894 sizek = limitk - basek;
895
896 /* see if we need a hole from 0xa0000 to 0xbffff */
897 if ((basek < ((8*64)+(8*16))) && (sizek > ((8*64)+(16*16)))) {
898 ram_resource(dev, (idx | i), basek, ((8*64)+(8*16)) - basek);
899 idx += 0x10;
900 basek = (8*64)+(16*16);
901 sizek = limitk - ((8*64)+(16*16));
902
903 }
904
905 //printk(BIOS_DEBUG, "node %d : mmio_basek=%08lx, basek=%08llx, limitk=%08llx\n", i, mmio_basek, basek, limitk);
906
907 /* split the region to accommodate pci memory space */
908 if ((basek < 4*1024*1024 ) && (limitk > mmio_basek)) {
909 if (basek <= mmio_basek) {
910 unsigned pre_sizek;
911 pre_sizek = mmio_basek - basek;
912 if (pre_sizek>0) {
913 ram_resource(dev, (idx | i), basek, pre_sizek);
914 idx += 0x10;
915 sizek -= pre_sizek;
916 if (!ramtop)
917 ramtop = mmio_basek * 1024;
918 }
919 basek = mmio_basek;
920 }
921 if ((basek + sizek) <= 4*1024*1024) {
922 sizek = 0;
923 }
924 else {
925 uint64_t topmem2 = bsp_topmem2();
926 basek = 4*1024*1024;
927 sizek = topmem2/1024 - basek;
928 }
929 }
930
931 ram_resource(dev, (idx | i), basek, sizek);
932 idx += 0x10;
933 printk(BIOS_DEBUG, "node %d: mmio_basek=%08lx, basek=%08llx, limitk=%08llx\n",
934 i, mmio_basek, basek, limitk);
935 if (!ramtop)
936 ramtop = limitk * 1024;
937 }
938
939#if CONFIG_GFXUMA
940 set_top_of_ram(uma_memory_base);
941 uma_resource(dev, 7, uma_memory_base >> 10, uma_memory_size >> 10);
942#else
943 set_top_of_ram(ramtop);
944#endif
945
946 for(link = dev->link_list; link; link = link->next) {
947 if (link->children) {
948 assign_resources(link);
949 }
950 }
951}
952
953static struct device_operations pci_domain_ops = {
954 .read_resources = domain_read_resources,
955 .set_resources = domain_set_resources,
956 .enable_resources = domain_enable_resources,
957 .init = NULL,
958 .scan_bus = pci_domain_scan_bus,
959 .ops_pci_bus = pci_bus_default_ops,
960};
961
962static void sysconf_init(device_t dev) // first node
963{
964 sblink = (pci_read_config32(dev, 0x64)>>8) & 7; // don't forget sublink1
965 node_nums = ((pci_read_config32(dev, 0x60)>>4) & 7) + 1; //NodeCnt[2:0]
966}
967
968static void add_more_links(device_t dev, unsigned total_links)
969{
970 struct bus *link, *last = NULL;
971 int link_num;
972
973 for (link = dev->link_list; link; link = link->next)
974 last = link;
975
976 if (last) {
977 int links = total_links - last->link_num;
978 link_num = last->link_num;
979 if (links > 0) {
980 link = malloc(links*sizeof(*link));
981 if (!link)
982 die("Couldn't allocate more links!\n");
983 memset(link, 0, links*sizeof(*link));
984 last->next = link;
985 }
986 }
987 else {
988 link_num = -1;
989 link = malloc(total_links*sizeof(*link));
990 memset(link, 0, total_links*sizeof(*link));
991 dev->link_list = link;
992 }
993
994 for (link_num = link_num + 1; link_num < total_links; link_num++) {
995 link->link_num = link_num;
996 link->dev = dev;
997 link->next = link + 1;
998 last = link;
999 link = link->next;
1000 }
1001 last->next = NULL;
1002}
1003
1004static u32 cpu_bus_scan(device_t dev, u32 max)
1005{
1006 struct bus *cpu_bus;
1007 device_t dev_mc;
1008#if CONFIG_CBB
1009 device_t pci_domain;
1010#endif
1011 int i,j;
1012 int coreid_bits;
1013 int core_max = 0;
1014 unsigned ApicIdCoreIdSize;
1015 unsigned core_nums;
1016 int siblings = 0;
1017 unsigned int family;
1018 u32 modules = 0;
1019 VOID* modules_ptr = &modules;
1020 BUILD_OPT_CFG* options = NULL;
1021 int ioapic_count = 0;
1022
1023 // TODO Remove the printk's.
1024 printk(BIOS_SPEW, "MullinsPI Debug: Grabbing the AMD Topology Information.\n");
1025 AmdGetValue(AMD_GLOBAL_USER_OPTIONS, (VOID**)&options, sizeof(options));
1026 AmdGetValue(AMD_GLOBAL_NUM_MODULES, &modules_ptr, sizeof(modules));
Alexandru Gagniuc2e0cf142014-12-28 20:38:32 -06001027 modules = *(u32*)modules_ptr;
Bruce Griffith27ed80b2014-08-15 11:46:25 -06001028 ASSERT(modules > 0);
1029 ASSERT(options);
1030 ioapic_count = (int)options->CfgPlatNumIoApics;
1031 ASSERT(ioapic_count > 0);
1032 printk(BIOS_SPEW, "MullinsPI Debug: AMD Topology Number of Modules (@0x%p) is %d\n", modules_ptr, modules);
1033 printk(BIOS_SPEW, "MullinsPI Debug: AMD Topology Number of IOAPICs (@0x%p) is %d\n", options, (int)options->CfgPlatNumIoApics);
1034
1035#if CONFIG_CBB
1036 dev_mc = dev_find_slot(0, PCI_DEVFN(CONFIG_CDB, 0)); //0x00
1037 if (dev_mc && dev_mc->bus) {
1038 printk(BIOS_DEBUG, "%s found", dev_path(dev_mc));
1039 pci_domain = dev_mc->bus->dev;
1040 if (pci_domain && (pci_domain->path.type == DEVICE_PATH_DOMAIN)) {
1041 printk(BIOS_DEBUG, "\n%s move to ",dev_path(dev_mc));
1042 dev_mc->bus->secondary = CONFIG_CBB; // move to 0xff
1043 printk(BIOS_DEBUG, "%s",dev_path(dev_mc));
1044 } else {
1045 printk(BIOS_DEBUG, " but it is not under pci_domain directly ");
1046 }
1047 printk(BIOS_DEBUG, "\n");
1048 }
1049 dev_mc = dev_find_slot(CONFIG_CBB, PCI_DEVFN(CONFIG_CDB, 0));
1050 if (!dev_mc) {
1051 dev_mc = dev_find_slot(0, PCI_DEVFN(0x18, 0));
1052 if (dev_mc && dev_mc->bus) {
1053 printk(BIOS_DEBUG, "%s found\n", dev_path(dev_mc));
1054 pci_domain = dev_mc->bus->dev;
1055 if (pci_domain && (pci_domain->path.type == DEVICE_PATH_DOMAIN)) {
1056 if ((pci_domain->link_list) && (pci_domain->link_list->children == dev_mc)) {
1057 printk(BIOS_DEBUG, "%s move to ",dev_path(dev_mc));
1058 dev_mc->bus->secondary = CONFIG_CBB; // move to 0xff
1059 printk(BIOS_DEBUG, "%s\n",dev_path(dev_mc));
1060 while (dev_mc) {
1061 printk(BIOS_DEBUG, "%s move to ",dev_path(dev_mc));
1062 dev_mc->path.pci.devfn -= PCI_DEVFN(0x18,0);
1063 printk(BIOS_DEBUG, "%s\n",dev_path(dev_mc));
1064 dev_mc = dev_mc->sibling;
1065 }
1066 }
1067 }
1068 }
1069 }
1070#endif
1071 dev_mc = dev_find_slot(CONFIG_CBB, PCI_DEVFN(CONFIG_CDB, 0));
1072 if (!dev_mc) {
1073 printk(BIOS_ERR, "%02x:%02x.0 not found", CONFIG_CBB, CONFIG_CDB);
1074 die("");
1075 }
1076 sysconf_init(dev_mc);
1077#if CONFIG_CBB && (MAX_NODE_NUMS > 32)
1078 if (node_nums>32) { // need to put node 32 to node 63 to bus 0xfe
1079 if (pci_domain->link_list && !pci_domain->link_list->next) {
1080 struct bus *new_link = new_link(pci_domain);
1081 pci_domain->link_list->next = new_link;
1082 new_link->link_num = 1;
1083 new_link->dev = pci_domain;
1084 new_link->children = 0;
1085 printk(BIOS_DEBUG, "%s links now 2\n", dev_path(pci_domain));
1086 }
1087 pci_domain->link_list->next->secondary = CONFIG_CBB - 1;
1088 }
1089#endif
1090
1091 /* Get Max Number of cores(MNC) */
1092 coreid_bits = (cpuid_ecx(AMD_CPUID_ASIZE_PCCOUNT) & 0x0000F000) >> 12;
1093 core_max = 1 << (coreid_bits & 0x000F); //mnc
1094
1095 ApicIdCoreIdSize = ((cpuid_ecx(0x80000008)>>12) & 0xF);
1096 if (ApicIdCoreIdSize) {
1097 core_nums = (1 << ApicIdCoreIdSize) - 1;
1098 } else {
1099 core_nums = 3; //quad core
1100 }
1101
1102 /* Find which cpus are present */
1103 cpu_bus = dev->link_list;
1104 for (i = 0; i < node_nums; i++) {
1105 device_t cdb_dev;
1106 unsigned busn, devn;
1107 struct bus *pbus;
1108
1109 busn = CONFIG_CBB;
1110 devn = CONFIG_CDB + i;
1111 pbus = dev_mc->bus;
1112#if CONFIG_CBB && (MAX_NODE_NUMS > 32)
1113 if (i >= 32) {
1114 busn--;
1115 devn -= 32;
1116 pbus = pci_domain->link_list->next;
1117 }
1118#endif
1119
1120 /* Find the cpu's pci device */
1121 cdb_dev = dev_find_slot(busn, PCI_DEVFN(devn, 0));
1122 if (!cdb_dev) {
1123 /* If I am probing things in a weird order
1124 * ensure all of the cpu's pci devices are found.
1125 */
1126 int fn;
1127 for(fn = 0; fn <= 5; fn++) { //FBDIMM?
1128 cdb_dev = pci_probe_dev(NULL, pbus,
1129 PCI_DEVFN(devn, fn));
1130 }
1131 cdb_dev = dev_find_slot(busn, PCI_DEVFN(devn, 0));
1132 } else {
1133 /* Ok, We need to set the links for that device.
1134 * otherwise the device under it will not be scanned
1135 */
Kyösti Mälkkic5163ed82015-02-04 13:25:37 +02001136
1137 add_more_links(cdb_dev, 4);
Bruce Griffith27ed80b2014-08-15 11:46:25 -06001138 }
1139
1140 family = cpuid_eax(1);
1141 family = (family >> 20) & 0xFF;
1142 if (family == 1) { //f10
1143 u32 dword;
1144 cdb_dev = dev_find_slot(busn, PCI_DEVFN(devn, 3));
1145 dword = pci_read_config32(cdb_dev, 0xe8);
1146 siblings = ((dword & BIT15) >> 13) | ((dword & (BIT13 | BIT12)) >> 12);
1147 } else if (family == 7) {//f16
1148 cdb_dev = dev_find_slot(busn, PCI_DEVFN(devn, 5));
1149 if (cdb_dev && cdb_dev->enabled) {
1150 siblings = pci_read_config32(cdb_dev, 0x84);
1151 siblings &= 0xFF;
1152 }
1153 } else {
1154 siblings = 0; //default one core
1155 }
1156 int enable_node = cdb_dev && cdb_dev->enabled;
1157 printk(BIOS_SPEW, "%s family%xh, core_max=0x%x, core_nums=0x%x, siblings=0x%x\n",
1158 dev_path(cdb_dev), 0x0f + family, core_max, core_nums, siblings);
1159
1160 for (j = 0; j <= siblings; j++ ) {
1161 u32 lapicid_start = 0;
1162
1163 /*
1164 * APIC ID calucation is tightly coupled with AGESA v5 code.
1165 * This calculation MUST match the assignment calculation done
1166 * in LocalApicInitializationAtEarly() function.
1167 * And reference GetLocalApicIdForCore()
1168 *
1169 * Apply apic enumeration rules
1170 * For systems with >= 16 APICs, put the IO-APICs at 0..n and
1171 * put the local-APICs at m..z
1172 *
1173 * This is needed because many IO-APIC devices only have 4 bits
1174 * for their APIC id and therefore must reside at 0..15
1175 */
1176 if ((node_nums * core_max) + ioapic_count >= 0x10) {
1177 lapicid_start = (ioapic_count - 1) / core_max;
1178 lapicid_start = (lapicid_start + 1) * core_max;
1179 printk(BIOS_SPEW, "lpaicid_start=0x%x ", lapicid_start);
1180 }
1181 u32 apic_id = (lapicid_start * (i/modules + 1)) + ((i % modules) ? (j + (siblings + 1)) : j);
1182 printk(BIOS_SPEW, "node 0x%x core 0x%x apicid=0x%x\n",
1183 i, j, apic_id);
1184
1185 device_t cpu = add_cpu_device(cpu_bus, apic_id, enable_node);
1186 if (cpu)
1187 amd_cpu_topology(cpu, i, j);
1188 } //j
1189 }
1190 return max;
1191}
1192
1193static void cpu_bus_init(device_t dev)
1194{
1195 initialize_cpus(dev->link_list);
1196}
1197
Bruce Griffith27ed80b2014-08-15 11:46:25 -06001198static void cpu_bus_read_resources(device_t dev)
1199{
1200#if CONFIG_MMCONF_SUPPORT
1201 struct resource *resource = new_resource(dev, 0xc0010058);
1202 resource->base = CONFIG_MMCONF_BASE_ADDRESS;
1203 resource->size = CONFIG_MMCONF_BUS_NUMBER * 4096*256;
1204 resource->flags = IORESOURCE_MEM | IORESOURCE_RESERVE |
1205 IORESOURCE_FIXED | IORESOURCE_STORED | IORESOURCE_ASSIGNED;
1206#endif
1207}
1208
1209static void cpu_bus_set_resources(device_t dev)
1210{
1211 struct resource *resource = find_resource(dev, 0xc0010058);
1212 if (resource) {
1213 report_resource_stored(dev, resource, " <mmconfig>");
1214 }
1215 pci_dev_set_resources(dev);
1216}
1217
1218static struct device_operations cpu_bus_ops = {
1219 .read_resources = cpu_bus_read_resources,
1220 .set_resources = cpu_bus_set_resources,
Edward O'Callaghan812d2a42014-10-31 08:17:23 +11001221 .enable_resources = DEVICE_NOOP,
Bruce Griffith27ed80b2014-08-15 11:46:25 -06001222 .init = cpu_bus_init,
1223 .scan_bus = cpu_bus_scan,
1224};
1225
1226static void root_complex_enable_dev(struct device *dev)
1227{
1228 static int done = 0;
1229
1230 /* Do not delay UMA setup, as a device on the PCI bus may evaluate
1231 the global uma_memory variables already in its enable function. */
1232 if (!done) {
1233 setup_bsp_ramtop();
1234 setup_uma_memory();
1235 done = 1;
1236 }
1237
1238 /* Set the operations if it is a special bus type */
1239 if (dev->path.type == DEVICE_PATH_DOMAIN) {
1240 dev->ops = &pci_domain_ops;
1241 } else if (dev->path.type == DEVICE_PATH_CPU_CLUSTER) {
1242 dev->ops = &cpu_bus_ops;
1243 }
1244}
1245
Kyösti Mälkkie4c17ce2014-10-21 18:22:32 +03001246struct chip_operations northbridge_amd_pi_00730F01_root_complex_ops = {
Bruce Griffith27ed80b2014-08-15 11:46:25 -06001247 CHIP_NAME("AMD FAM16 Root Complex")
1248 .enable_dev = root_complex_enable_dev,
1249};
1250
1251/*********************************************************************
1252 * Change the vendor / device IDs to match the generic VBIOS header. *
1253 *********************************************************************/
1254u32 map_oprom_vendev(u32 vendev)
1255{
1256 u32 new_vendev;
1257 new_vendev =
1258 ((0x10029850 <= vendev) && (vendev <= 0x1002986F)) ? 0x10029850 : vendev;
1259
1260 if (vendev != new_vendev)
1261 printk(BIOS_NOTICE, "Mapping PCI device %8x to %8x\n", vendev, new_vendev);
1262
1263 return new_vendev;
1264}