blob: 849f40cbd23c4150c7903b53587e971d4399b66c [file] [log] [blame]
Siyuan Wang3e32cc02013-07-09 17:16:20 +08001/*
2 * This file is part of the coreboot project.
3 *
4 * Copyright (C) 2012 Advanced Micro Devices, Inc.
5 *
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License as published by
8 * the Free Software Foundation; version 2 of the License.
9 *
10 * This program is distributed in the hope that it will be useful,
11 * but WITHOUT ANY WARRANTY; without even the implied warranty of
12 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
13 * GNU General Public License for more details.
14 *
15 * You should have received a copy of the GNU General Public License
16 * along with this program; if not, write to the Free Software
17 * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA
18 */
19
20#include <console/console.h>
21#include <arch/io.h>
Kyösti Mälkki8ae16a42014-06-19 20:44:34 +030022#include <arch/acpi.h>
Vladimir Serbinenkodb09b0622014-10-08 22:15:17 +020023#include <arch/acpigen.h>
Siyuan Wang3e32cc02013-07-09 17:16:20 +080024#include <stdint.h>
25#include <device/device.h>
26#include <device/pci.h>
27#include <device/pci_ids.h>
28#include <device/hypertransport.h>
29#include <stdlib.h>
30#include <string.h>
31#include <lib.h>
32#include <cpu/cpu.h>
33#include <cbmem.h>
34
35#include <cpu/x86/lapic.h>
36#include <cpu/amd/mtrr.h>
37
38#include <Porting.h>
39#include <AGESA.h>
40#include <Options.h>
41#include <Topology.h>
42#include <cpu/amd/amdfam16.h>
43#include <cpuRegisters.h>
Kyösti Mälkkif21c2ac2014-10-19 09:35:18 +030044#include <northbridge/amd/agesa/agesawrapper.h>
Kyösti Mälkki7b23ae02014-07-04 16:14:37 +030045#include <northbridge/amd/agesa/agesawrapper_call.h>
Siyuan Wang3e32cc02013-07-09 17:16:20 +080046
47#define MAX_NODE_NUMS (MAX_NODES * MAX_DIES)
48
49#if (defined CONFIG_EXT_CONF_SUPPORT) && CONFIG_EXT_CONF_SUPPORT == 1
50#error CONFIG_EXT_CONF_SUPPORT == 1 not support anymore!
51#endif
52
53typedef struct dram_base_mask {
54 u32 base; //[47:27] at [28:8]
55 u32 mask; //[47:27] at [28:8] and enable at bit 0
56} dram_base_mask_t;
57
58static unsigned node_nums;
59static unsigned sblink;
60static device_t __f0_dev[MAX_NODE_NUMS];
61static device_t __f1_dev[MAX_NODE_NUMS];
62static device_t __f2_dev[MAX_NODE_NUMS];
63static device_t __f4_dev[MAX_NODE_NUMS];
64static unsigned fx_devs = 0;
65
66static dram_base_mask_t get_dram_base_mask(u32 nodeid)
67{
68 device_t dev;
69 dram_base_mask_t d;
70 dev = __f1_dev[0];
71 u32 temp;
72 temp = pci_read_config32(dev, 0x44 + (nodeid << 3)); //[39:24] at [31:16]
73 d.mask = ((temp & 0xfff80000)>>(8+3)); // mask out DramMask [26:24] too
74 temp = pci_read_config32(dev, 0x144 + (nodeid <<3)) & 0xff; //[47:40] at [7:0]
75 d.mask |= temp<<21;
76 temp = pci_read_config32(dev, 0x40 + (nodeid << 3)); //[39:24] at [31:16]
77 d.mask |= (temp & 1); // enable bit
78 d.base = ((temp & 0xfff80000)>>(8+3)); // mask out DramBase [26:24) too
79 temp = pci_read_config32(dev, 0x140 + (nodeid <<3)) & 0xff; //[47:40] at [7:0]
80 d.base |= temp<<21;
81 return d;
82}
83
84static void set_io_addr_reg(device_t dev, u32 nodeid, u32 linkn, u32 reg,
85 u32 io_min, u32 io_max)
86{
87 u32 i;
88 u32 tempreg;
89 /* io range allocation */
90 tempreg = (nodeid&0xf) | ((nodeid & 0x30)<<(8-4)) | (linkn<<4) | ((io_max&0xf0)<<(12-4)); //limit
91 for (i=0; i<node_nums; i++)
92 pci_write_config32(__f1_dev[i], reg+4, tempreg);
93 tempreg = 3 /*| ( 3<<4)*/ | ((io_min&0xf0)<<(12-4)); //base :ISA and VGA ?
94#if 0
95 // FIXME: can we use VGA reg instead?
96 if (dev->link[link].bridge_ctrl & PCI_BRIDGE_CTL_VGA) {
97 printk(BIOS_SPEW, "%s, enabling legacy VGA IO forwarding for %s link %s\n",
98 __func__, dev_path(dev), link);
99 tempreg |= PCI_IO_BASE_VGA_EN;
100 }
101 if (dev->link[link].bridge_ctrl & PCI_BRIDGE_CTL_NO_ISA) {
102 tempreg |= PCI_IO_BASE_NO_ISA;
103 }
104#endif
105 for (i=0; i<node_nums; i++)
106 pci_write_config32(__f1_dev[i], reg, tempreg);
107}
108
109static void set_mmio_addr_reg(u32 nodeid, u32 linkn, u32 reg, u32 index, u32 mmio_min, u32 mmio_max, u32 nodes)
110{
111 u32 i;
112 u32 tempreg;
113 /* io range allocation */
114 tempreg = (nodeid&0xf) | (linkn<<4) | (mmio_max&0xffffff00); //limit
115 for (i=0; i<nodes; i++)
116 pci_write_config32(__f1_dev[i], reg+4, tempreg);
117 tempreg = 3 | (nodeid & 0x30) | (mmio_min&0xffffff00);
118 for (i=0; i<node_nums; i++)
119 pci_write_config32(__f1_dev[i], reg, tempreg);
120}
121
122static device_t get_node_pci(u32 nodeid, u32 fn)
123{
124#if MAX_NODE_NUMS + CONFIG_CDB >= 32
125 if ((CONFIG_CDB + nodeid) < 32) {
126 return dev_find_slot(CONFIG_CBB, PCI_DEVFN(CONFIG_CDB + nodeid, fn));
127 } else {
128 return dev_find_slot(CONFIG_CBB-1, PCI_DEVFN(CONFIG_CDB + nodeid - 32, fn));
129 }
130#else
131 return dev_find_slot(CONFIG_CBB, PCI_DEVFN(CONFIG_CDB + nodeid, fn));
132#endif
133}
134
135static void get_fx_devs(void)
136{
137 int i;
138 for (i = 0; i < MAX_NODE_NUMS; i++) {
139 __f0_dev[i] = get_node_pci(i, 0);
140 __f1_dev[i] = get_node_pci(i, 1);
141 __f2_dev[i] = get_node_pci(i, 2);
142 __f4_dev[i] = get_node_pci(i, 4);
143 if (__f0_dev[i] != NULL && __f1_dev[i] != NULL)
144 fx_devs = i+1;
145 }
146 if (__f1_dev[0] == NULL || __f0_dev[0] == NULL || fx_devs == 0) {
147 die("Cannot find 0:0x18.[0|1]\n");
148 }
149 printk(BIOS_DEBUG, "fx_devs=0x%x\n", fx_devs);
150}
151
152static u32 f1_read_config32(unsigned reg)
153{
154 if (fx_devs == 0)
155 get_fx_devs();
156 return pci_read_config32(__f1_dev[0], reg);
157}
158
159static void f1_write_config32(unsigned reg, u32 value)
160{
161 int i;
162 if (fx_devs == 0)
163 get_fx_devs();
164 for(i = 0; i < fx_devs; i++) {
165 device_t dev;
166 dev = __f1_dev[i];
167 if (dev && dev->enabled) {
168 pci_write_config32(dev, reg, value);
169 }
170 }
171}
172
173static u32 amdfam16_nodeid(device_t dev)
174{
175#if MAX_NODE_NUMS == 64
176 unsigned busn;
177 busn = dev->bus->secondary;
178 if (busn != CONFIG_CBB) {
179 return (dev->path.pci.devfn >> 3) - CONFIG_CDB + 32;
180 } else {
181 return (dev->path.pci.devfn >> 3) - CONFIG_CDB;
182 }
183
184#else
185 return (dev->path.pci.devfn >> 3) - CONFIG_CDB;
186#endif
187}
188
189static void set_vga_enable_reg(u32 nodeid, u32 linkn)
190{
191 u32 val;
192
193 val = 1 | (nodeid<<4) | (linkn<<12);
194 /* it will routing
195 * (1)mmio 0xa0000:0xbffff
196 * (2)io 0x3b0:0x3bb, 0x3c0:0x3df
197 */
198 f1_write_config32(0xf4, val);
199
200}
201
202/**
203 * @return
204 * @retval 2 resoure does not exist, usable
205 * @retval 0 resource exists, not usable
206 * @retval 1 resource exist, resource has been allocated before
207 */
208static int reg_useable(unsigned reg, device_t goal_dev, unsigned goal_nodeid,
209 unsigned goal_link)
210{
211 struct resource *res;
212 unsigned nodeid, link = 0;
213 int result;
214 res = 0;
215 for (nodeid = 0; !res && (nodeid < fx_devs); nodeid++) {
216 device_t dev;
217 dev = __f0_dev[nodeid];
218 if (!dev)
219 continue;
220 for (link = 0; !res && (link < 8); link++) {
221 res = probe_resource(dev, IOINDEX(0x1000 + reg, link));
222 }
223 }
224 result = 2;
225 if (res) {
226 result = 0;
227 if ((goal_link == (link - 1)) &&
228 (goal_nodeid == (nodeid - 1)) &&
229 (res->flags <= 1)) {
230 result = 1;
231 }
232 }
233 return result;
234}
235
236static struct resource *amdfam16_find_iopair(device_t dev, unsigned nodeid, unsigned link)
237{
238 struct resource *resource;
239 u32 free_reg, reg;
240 resource = 0;
241 free_reg = 0;
242 for (reg = 0xc0; reg <= 0xd8; reg += 0x8) {
243 int result;
244 result = reg_useable(reg, dev, nodeid, link);
245 if (result == 1) {
246 /* I have been allocated this one */
247 break;
248 }
249 else if (result > 1) {
250 /* I have a free register pair */
251 free_reg = reg;
252 }
253 }
254 if (reg > 0xd8) {
255 reg = free_reg; // if no free, the free_reg still be 0
256 }
257
258 resource = new_resource(dev, IOINDEX(0x1000 + reg, link));
259
260 return resource;
261}
262
263static struct resource *amdfam16_find_mempair(device_t dev, u32 nodeid, u32 link)
264{
265 struct resource *resource;
266 u32 free_reg, reg;
267 resource = 0;
268 free_reg = 0;
269 for (reg = 0x80; reg <= 0xb8; reg += 0x8) {
270 int result;
271 result = reg_useable(reg, dev, nodeid, link);
272 if (result == 1) {
273 /* I have been allocated this one */
274 break;
275 }
276 else if (result > 1) {
277 /* I have a free register pair */
278 free_reg = reg;
279 }
280 }
281 if (reg > 0xb8) {
282 reg = free_reg;
283 }
284
285 resource = new_resource(dev, IOINDEX(0x1000 + reg, link));
286 return resource;
287}
288
289static void amdfam16_link_read_bases(device_t dev, u32 nodeid, u32 link)
290{
291 struct resource *resource;
292
293 /* Initialize the io space constraints on the current bus */
294 resource = amdfam16_find_iopair(dev, nodeid, link);
295 if (resource) {
296 u32 align;
297 align = log2(HT_IO_HOST_ALIGN);
298 resource->base = 0;
299 resource->size = 0;
300 resource->align = align;
301 resource->gran = align;
302 resource->limit = 0xffffUL;
303 resource->flags = IORESOURCE_IO | IORESOURCE_BRIDGE;
304 }
305
306 /* Initialize the prefetchable memory constraints on the current bus */
307 resource = amdfam16_find_mempair(dev, nodeid, link);
308 if (resource) {
309 resource->base = 0;
310 resource->size = 0;
311 resource->align = log2(HT_MEM_HOST_ALIGN);
312 resource->gran = log2(HT_MEM_HOST_ALIGN);
313 resource->limit = 0xffffffffffULL;
314 resource->flags = IORESOURCE_MEM | IORESOURCE_PREFETCH;
315 resource->flags |= IORESOURCE_BRIDGE;
316 }
317
318 /* Initialize the memory constraints on the current bus */
319 resource = amdfam16_find_mempair(dev, nodeid, link);
320 if (resource) {
321 resource->base = 0;
322 resource->size = 0;
323 resource->align = log2(HT_MEM_HOST_ALIGN);
324 resource->gran = log2(HT_MEM_HOST_ALIGN);
325 resource->limit = 0xffffffffffULL;
326 resource->flags = IORESOURCE_MEM | IORESOURCE_BRIDGE;
327 }
328
329}
330
331static void read_resources(device_t dev)
332{
333 u32 nodeid;
334 struct bus *link;
335
336 nodeid = amdfam16_nodeid(dev);
337 for (link = dev->link_list; link; link = link->next) {
338 if (link->children) {
339 amdfam16_link_read_bases(dev, nodeid, link->link_num);
340 }
341 }
Edward O'Callaghan66c65322014-11-21 01:43:38 +1100342
343 /*
344 * This MMCONF resource must be reserved in the PCI_DOMAIN.
345 * It is not honored by the coreboot resource allocator if it is in
346 * the APIC_CLUSTER.
347 */
348#if CONFIG_MMCONF_SUPPORT
349 struct resource *resource = new_resource(dev, 0xc0010058);
350 resource->base = CONFIG_MMCONF_BASE_ADDRESS;
351 resource->size = CONFIG_MMCONF_BUS_NUMBER * 4096 * 256;
352 resource->flags = IORESOURCE_MEM | IORESOURCE_RESERVE |
353 IORESOURCE_FIXED | IORESOURCE_STORED | IORESOURCE_ASSIGNED;
354#endif
Siyuan Wang3e32cc02013-07-09 17:16:20 +0800355}
356
357static void set_resource(device_t dev, struct resource *resource, u32 nodeid)
358{
359 resource_t rbase, rend;
360 unsigned reg, link_num;
361 char buf[50];
362
363 /* Make certain the resource has actually been set */
364 if (!(resource->flags & IORESOURCE_ASSIGNED)) {
365 return;
366 }
367
368 /* If I have already stored this resource don't worry about it */
369 if (resource->flags & IORESOURCE_STORED) {
370 return;
371 }
372
373 /* Only handle PCI memory and IO resources */
374 if (!(resource->flags & (IORESOURCE_MEM | IORESOURCE_IO)))
375 return;
376
377 /* Ensure I am actually looking at a resource of function 1 */
378 if ((resource->index & 0xffff) < 0x1000) {
379 return;
380 }
381 /* Get the base address */
382 rbase = resource->base;
383
384 /* Get the limit (rounded up) */
385 rend = resource_end(resource);
386
387 /* Get the register and link */
388 reg = resource->index & 0xfff; // 4k
389 link_num = IOINDEX_LINK(resource->index);
390
391 if (resource->flags & IORESOURCE_IO) {
392 set_io_addr_reg(dev, nodeid, link_num, reg, rbase>>8, rend>>8);
393 }
394 else if (resource->flags & IORESOURCE_MEM) {
Edward O'Callaghanae5fd342014-11-20 19:58:09 +1100395 set_mmio_addr_reg(nodeid, link_num, reg, (resource->index >>24), rbase>>8, rend>>8, node_nums);// [39:8]
Siyuan Wang3e32cc02013-07-09 17:16:20 +0800396 }
397 resource->flags |= IORESOURCE_STORED;
Vladimir Serbinenkoa37383d2013-11-26 02:41:26 +0100398 snprintf(buf, sizeof (buf), " <node %x link %x>",
Siyuan Wang3e32cc02013-07-09 17:16:20 +0800399 nodeid, link_num);
400 report_resource_stored(dev, resource, buf);
401}
402
403/**
404 * I tried to reuse the resource allocation code in set_resource()
405 * but it is too difficult to deal with the resource allocation magic.
406 */
407
408static void create_vga_resource(device_t dev, unsigned nodeid)
409{
410 struct bus *link;
411
412 /* find out which link the VGA card is connected,
413 * we only deal with the 'first' vga card */
414 for (link = dev->link_list; link; link = link->next) {
415 if (link->bridge_ctrl & PCI_BRIDGE_CTL_VGA) {
416#if CONFIG_MULTIPLE_VGA_ADAPTERS
417 extern device_t vga_pri; // the primary vga device, defined in device.c
418 printk(BIOS_DEBUG, "VGA: vga_pri bus num = %d bus range [%d,%d]\n", vga_pri->bus->secondary,
419 link->secondary,link->subordinate);
420 /* We need to make sure the vga_pri is under the link */
421 if((vga_pri->bus->secondary >= link->secondary ) &&
422 (vga_pri->bus->secondary <= link->subordinate )
423 )
424#endif
425 break;
426 }
427 }
428
429 /* no VGA card installed */
430 if (link == NULL)
431 return;
432
433 printk(BIOS_DEBUG, "VGA: %s (aka node %d) link %d has VGA device\n", dev_path(dev), nodeid, sblink);
434 set_vga_enable_reg(nodeid, sblink);
435}
436
437static void set_resources(device_t dev)
438{
439 unsigned nodeid;
440 struct bus *bus;
441 struct resource *res;
442
443 /* Find the nodeid */
444 nodeid = amdfam16_nodeid(dev);
445
446 create_vga_resource(dev, nodeid); //TODO: do we need this?
447
448 /* Set each resource we have found */
449 for (res = dev->resource_list; res; res = res->next) {
450 set_resource(dev, res, nodeid);
451 }
452
453 for (bus = dev->link_list; bus; bus = bus->next) {
454 if (bus->children) {
455 assign_resources(bus);
456 }
457 }
Edward O'Callaghan66c65322014-11-21 01:43:38 +1100458
459 /* Print the MMCONF region if it has been reserved. */
460 res = find_resource(dev, 0xc0010058);
461 if (res) {
462 report_resource_stored(dev, res, " <mmconfig>");
463 }
Siyuan Wang3e32cc02013-07-09 17:16:20 +0800464}
465
Siyuan Wang3e32cc02013-07-09 17:16:20 +0800466#if 0 /* TODO: Check if needed. */
467static unsigned scan_chains(device_t dev, unsigned max)
468{
469 unsigned nodeid;
470 struct bus *link;
471 device_t io_hub = NULL;
472 u32 next_unitid = 0x18;
473 nodeid = amdfam16_nodeid(dev);
474 if (nodeid == 0) {
475 for (link = dev->link_list; link; link = link->next) {
476 //if (link->link_num == sblink) { /* devicetree put IO Hub on link_lsit[sblink] */
477 if (link->link_num == 0) { /* devicetree put IO Hub on link_lsit[0] */
478 io_hub = link->children;
479 if (!io_hub || !io_hub->enabled) {
480 die("I can't find the IO Hub, or IO Hub not enabled, please check the device tree.\n");
481 }
482 /* Now that nothing is overlapping it is safe to scan the children. */
483 max = pci_scan_bus(link, 0x00, ((next_unitid - 1) << 3) | 7, 0);
484 }
485 }
486 }
487 return max;
488}
489#endif
Vladimir Serbinenkodb09b0622014-10-08 22:15:17 +0200490
491
492unsigned long acpi_fill_hest(acpi_hest_t *hest)
493{
494 void *addr, *current;
495
496 /* Skip the HEST header. */
497 current = (void *)(hest + 1);
498
499 addr = agesawrapper_getlateinitptr(PICK_WHEA_MCE);
500 if (addr != NULL)
501 current += acpi_create_hest_error_source(hest, current, 0, (void *)((u32)addr + 2), *(UINT16 *)addr - 2);
502
503 addr = agesawrapper_getlateinitptr(PICK_WHEA_CMC);
504 if (addr != NULL)
505 current += acpi_create_hest_error_source(hest, current, 1, (void *)((u32)addr + 2), *(UINT16 *)addr - 2);
506
507 return (unsigned long)current;
508}
509
510/* Implemented with AGESA-specific code. Dummy to keep linker happy. */
511unsigned long acpi_fill_slit(unsigned long current)
512{
513 return current;
514}
515
516/* Implemented with AGESA-specific code. Dummy to keep linker happy. */
517unsigned long acpi_fill_srat(unsigned long current)
518{
519 return current;
520}
521
522static void northbridge_fill_ssdt_generator(void)
523{
524 msr_t msr;
525 char pscope[] = "\\_SB.PCI0";
526
527 acpigen_write_scope(pscope);
528 msr = rdmsr(TOP_MEM);
529 acpigen_write_name_dword("TOM1", msr.lo);
530 msr = rdmsr(TOP_MEM2);
531 /*
532 * Since XP only implements parts of ACPI 2.0, we can't use a qword
533 * here.
534 * See http://www.acpi.info/presentations/S01USMOBS169_OS%2520new.ppt
535 * slide 22ff.
536 * Shift value right by 20 bit to make it fit into 32bit,
537 * giving us 1MB granularity and a limit of almost 4Exabyte of memory.
538 */
539 acpigen_write_name_dword("TOM2", (msr.hi << 12) | msr.lo >> 20);
540 acpigen_pop_len();
541}
542
543static unsigned long agesa_write_acpi_tables(unsigned long current,
544 acpi_rsdp_t *rsdp)
545{
546 acpi_srat_t *srat;
547 acpi_slit_t *slit;
548 acpi_header_t *ssdt;
549 acpi_header_t *alib;
550 acpi_header_t *ivrs;
551 acpi_hest_t *hest;
552
553 /* HEST */
554 current = ALIGN(current, 8);
555 hest = (acpi_hest_t *)current;
556 acpi_write_hest((void *)current);
557 acpi_add_table(rsdp, (void *)current);
558 current += ((acpi_header_t *)current)->length;
559
560 current = ALIGN(current, 8);
561 printk(BIOS_DEBUG, "ACPI: * IVRS at %lx\n", current);
562 ivrs = agesawrapper_getlateinitptr(PICK_IVRS);
563 if (ivrs != NULL) {
564 memcpy((void *)current, ivrs, ivrs->length);
565 ivrs = (acpi_header_t *) current;
566 current += ivrs->length;
567 acpi_add_table(rsdp, ivrs);
568 } else {
569 printk(BIOS_DEBUG, " AGESA IVRS table NULL. Skipping.\n");
570 }
571
572 /* SRAT */
573 current = ALIGN(current, 8);
574 printk(BIOS_DEBUG, "ACPI: * SRAT at %lx\n", current);
575 srat = (acpi_srat_t *) agesawrapper_getlateinitptr (PICK_SRAT);
576 if (srat != NULL) {
577 memcpy((void *)current, srat, srat->header.length);
578 srat = (acpi_srat_t *) current;
579 current += srat->header.length;
580 acpi_add_table(rsdp, srat);
581 } else {
582 printk(BIOS_DEBUG, " AGESA SRAT table NULL. Skipping.\n");
583 }
584
585 /* SLIT */
586 current = ALIGN(current, 8);
587 printk(BIOS_DEBUG, "ACPI: * SLIT at %lx\n", current);
588 slit = (acpi_slit_t *) agesawrapper_getlateinitptr (PICK_SLIT);
589 if (slit != NULL) {
590 memcpy((void *)current, slit, slit->header.length);
591 slit = (acpi_slit_t *) current;
592 current += slit->header.length;
593 acpi_add_table(rsdp, slit);
594 } else {
595 printk(BIOS_DEBUG, " AGESA SLIT table NULL. Skipping.\n");
596 }
597
598 /* ALIB */
599 current = ALIGN(current, 16);
600 printk(BIOS_DEBUG, "ACPI: * AGESA ALIB SSDT at %lx\n", current);
601 alib = (acpi_header_t *)agesawrapper_getlateinitptr (PICK_ALIB);
602 if (alib != NULL) {
603 memcpy((void *)current, alib, alib->length);
604 alib = (acpi_header_t *) current;
605 current += alib->length;
606 acpi_add_table(rsdp, (void *)alib);
607 }
608 else {
609 printk(BIOS_DEBUG, " AGESA ALIB SSDT table NULL. Skipping.\n");
610 }
611
612 /* this pstate ssdt may cause Blue Screen: Fixed: Keep this comment for a while. */
613 /* SSDT */
614 current = ALIGN(current, 16);
615 printk(BIOS_DEBUG, "ACPI: * SSDT at %lx\n", current);
616 ssdt = (acpi_header_t *)agesawrapper_getlateinitptr (PICK_PSTATE);
617 if (ssdt != NULL) {
618 memcpy((void *)current, ssdt, ssdt->length);
619 ssdt = (acpi_header_t *) current;
620 current += ssdt->length;
621 }
622 else {
623 printk(BIOS_DEBUG, " AGESA PState table NULL. Skipping.\n");
624 }
625 acpi_add_table(rsdp,ssdt);
626
627 printk(BIOS_DEBUG, "ACPI: * SSDT for PState at %lx\n", current);
628
629 return current;
630}
631
Siyuan Wang3e32cc02013-07-09 17:16:20 +0800632static struct device_operations northbridge_operations = {
633 .read_resources = read_resources,
634 .set_resources = set_resources,
635 .enable_resources = pci_dev_enable_resources,
Edward O'Callaghand994ef12014-11-21 02:22:33 +1100636 .init = DEVICE_NOOP,
Vladimir Serbinenkodb09b0622014-10-08 22:15:17 +0200637 .acpi_fill_ssdt_generator = northbridge_fill_ssdt_generator,
638 .write_acpi_tables = agesa_write_acpi_tables,
Siyuan Wang3e32cc02013-07-09 17:16:20 +0800639 //.scan_bus = scan_chains, /* TODO: */
640 .enable = 0,
641 .ops_pci = 0,
642};
643
644static const struct pci_driver family16_northbridge __pci_driver = {
645 .ops = &northbridge_operations,
646 .vendor = PCI_VENDOR_ID_AMD,
647 .device = PCI_DEVICE_ID_AMD_16H_MODEL_000F_NB_HT,
648};
649
650static const struct pci_driver family10_northbridge __pci_driver = {
651 .ops = &northbridge_operations,
652 .vendor = PCI_VENDOR_ID_AMD,
653 .device = PCI_DEVICE_ID_AMD_10H_NB_HT,
654};
655
Kyösti Mälkki4ee82c62014-11-25 16:03:12 +0200656static void fam16_finalize(void *chip_info)
657{
658 device_t dev;
659 u32 value;
660 dev = dev_find_slot(0, PCI_DEVFN(0, 0)); /* clear IoapicSbFeatureEn */
661 pci_write_config32(dev, 0xF8, 0);
662 pci_write_config32(dev, 0xFC, 5); /* TODO: move it to dsdt.asl */
663
664 /* disable No Snoop */
665 dev = dev_find_slot(0, PCI_DEVFN(1, 1));
666 value = pci_read_config32(dev, 0x60);
667 value &= ~(1 << 11);
668 pci_write_config32(dev, 0x60, value);
669}
670
Siyuan Wang3e32cc02013-07-09 17:16:20 +0800671struct chip_operations northbridge_amd_agesa_family16kb_ops = {
672 CHIP_NAME("AMD FAM16 Northbridge")
673 .enable_dev = 0,
Kyösti Mälkki4ee82c62014-11-25 16:03:12 +0200674 .final = fam16_finalize,
Siyuan Wang3e32cc02013-07-09 17:16:20 +0800675};
676
677static void domain_read_resources(device_t dev)
678{
679 unsigned reg;
680
681 /* Find the already assigned resource pairs */
682 get_fx_devs();
683 for (reg = 0x80; reg <= 0xd8; reg+= 0x08) {
684 u32 base, limit;
685 base = f1_read_config32(reg);
686 limit = f1_read_config32(reg + 0x04);
687 /* Is this register allocated? */
688 if ((base & 3) != 0) {
689 unsigned nodeid, reg_link;
690 device_t reg_dev;
691 if (reg<0xc0) { // mmio
692 nodeid = (limit & 0xf) + (base&0x30);
693 } else { // io
694 nodeid = (limit & 0xf) + ((base>>4)&0x30);
695 }
696 reg_link = (limit >> 4) & 7;
697 reg_dev = __f0_dev[nodeid];
698 if (reg_dev) {
699 /* Reserve the resource */
700 struct resource *res;
701 res = new_resource(reg_dev, IOINDEX(0x1000 + reg, reg_link));
702 if (res) {
703 res->flags = 1;
704 }
705 }
706 }
707 }
708 /* FIXME: do we need to check extend conf space?
709 I don't believe that much preset value */
710
711#if !CONFIG_PCI_64BIT_PREF_MEM
712 pci_domain_read_resources(dev);
713
714#else
715 struct bus *link;
716 struct resource *resource;
717 for (link=dev->link_list; link; link = link->next) {
718 /* Initialize the system wide io space constraints */
719 resource = new_resource(dev, 0|(link->link_num<<2));
720 resource->base = 0x400;
721 resource->limit = 0xffffUL;
722 resource->flags = IORESOURCE_IO;
723
724 /* Initialize the system wide prefetchable memory resources constraints */
725 resource = new_resource(dev, 1|(link->link_num<<2));
726 resource->limit = 0xfcffffffffULL;
727 resource->flags = IORESOURCE_MEM | IORESOURCE_PREFETCH;
728
729 /* Initialize the system wide memory resources constraints */
730 resource = new_resource(dev, 2|(link->link_num<<2));
731 resource->limit = 0xfcffffffffULL;
732 resource->flags = IORESOURCE_MEM;
733 }
734#endif
735}
736
Siyuan Wang3e32cc02013-07-09 17:16:20 +0800737static void domain_enable_resources(device_t dev)
738{
Kyösti Mälkki8ae16a42014-06-19 20:44:34 +0300739 if (acpi_is_wakeup_s3())
Kyösti Mälkki7b23ae02014-07-04 16:14:37 +0300740 AGESAWRAPPER(fchs3laterestore);
Siyuan Wang3e32cc02013-07-09 17:16:20 +0800741
742 /* Must be called after PCI enumeration and resource allocation */
Kyösti Mälkkib139b5e2014-10-20 07:41:20 +0300743 if (!acpi_is_wakeup_s3()) {
744 /* Enable MMIO on AMD CPU Address Map Controller */
745 agesawrapper_amdinitcpuio();
Siyuan Wang3e32cc02013-07-09 17:16:20 +0800746
Kyösti Mälkkib139b5e2014-10-20 07:41:20 +0300747 AGESAWRAPPER(amdinitmid);
748 }
Siyuan Wang3e32cc02013-07-09 17:16:20 +0800749 printk(BIOS_DEBUG, " ader - leaving domain_enable_resources.\n");
750}
751
752#if CONFIG_HW_MEM_HOLE_SIZEK != 0
753struct hw_mem_hole_info {
754 unsigned hole_startk;
755 int node_id;
756};
757static struct hw_mem_hole_info get_hw_mem_hole_info(void)
758{
759 struct hw_mem_hole_info mem_hole;
760 int i;
761 mem_hole.hole_startk = CONFIG_HW_MEM_HOLE_SIZEK;
762 mem_hole.node_id = -1;
763 for (i = 0; i < node_nums; i++) {
764 dram_base_mask_t d;
765 u32 hole;
766 d = get_dram_base_mask(i);
767 if (!(d.mask & 1)) continue; // no memory on this node
768 hole = pci_read_config32(__f1_dev[i], 0xf0);
769 if (hole & 2) { // we find the hole
770 mem_hole.hole_startk = (hole & (0xff<<24)) >> 10;
771 mem_hole.node_id = i; // record the node No with hole
772 break; // only one hole
773 }
774 }
Kyösti Mälkki2f9b3af2014-06-26 05:30:54 +0300775
776 /* We need to double check if there is special set on base reg and limit reg
777 * are not continuous instead of hole, it will find out its hole_startk.
778 */
Siyuan Wang3e32cc02013-07-09 17:16:20 +0800779 if (mem_hole.node_id == -1) {
780 resource_t limitk_pri = 0;
781 for (i=0; i<node_nums; i++) {
782 dram_base_mask_t d;
783 resource_t base_k, limit_k;
784 d = get_dram_base_mask(i);
785 if (!(d.base & 1)) continue;
786 base_k = ((resource_t)(d.base & 0x1fffff00)) <<9;
787 if (base_k > 4 *1024 * 1024) break; // don't need to go to check
788 if (limitk_pri != base_k) { // we find the hole
789 mem_hole.hole_startk = (unsigned)limitk_pri; // must beblow 4G
790 mem_hole.node_id = i;
791 break; //only one hole
792 }
793 limit_k = ((resource_t)(((d.mask & ~1) + 0x000FF) & 0x1fffff00)) << 9;
794 limitk_pri = limit_k;
795 }
796 }
797 return mem_hole;
798}
799#endif
800
801#define ONE_MB_SHIFT 20
802
803static void setup_uma_memory(void)
804{
805#if CONFIG_GFXUMA
806 uint32_t topmem = (uint32_t) bsp_topmem();
807 uint32_t sys_mem;
808
809 /* refer to UMA Size Consideration in Family16h BKDG. */
810 /* Please reference MemNGetUmaSizeOR () */
811 /*
812 * Total system memory UMASize
813 * >= 2G 512M
814 * >=1G 256M
815 * <1G 64M
816 */
817 sys_mem = topmem + (16 << ONE_MB_SHIFT); // Ignore 16MB allocated for C6 when finding UMA size
818 if ((bsp_topmem2()>>32) || (sys_mem >= 2048 << ONE_MB_SHIFT)) {
819 uma_memory_size = 512 << ONE_MB_SHIFT;
820 } else if (sys_mem >= 1024 << ONE_MB_SHIFT) {
821 uma_memory_size = 256 << ONE_MB_SHIFT;
822 } else {
823 uma_memory_size = 64 << ONE_MB_SHIFT;
824 }
825 uma_memory_base = topmem - uma_memory_size; /* TOP_MEM1 */
826
827 printk(BIOS_INFO, "%s: uma size 0x%08llx, memory start 0x%08llx\n",
828 __func__, uma_memory_size, uma_memory_base);
829
830 /* TODO: TOP_MEM2 */
831#endif
832}
833
834
835static void domain_set_resources(device_t dev)
836{
837#if CONFIG_PCI_64BIT_PREF_MEM
838 struct resource *io, *mem1, *mem2;
839 struct resource *res;
840#endif
841 unsigned long mmio_basek;
842 u32 pci_tolm;
Kyösti Mälkki2b790f62013-09-03 05:25:57 +0300843 u64 ramtop = 0;
Siyuan Wang3e32cc02013-07-09 17:16:20 +0800844 int i, idx;
845 struct bus *link;
846#if CONFIG_HW_MEM_HOLE_SIZEK != 0
847 struct hw_mem_hole_info mem_hole;
848 u32 reset_memhole = 1;
849#endif
850
851#if CONFIG_PCI_64BIT_PREF_MEM
852
853 for (link = dev->link_list; link; link = link->next) {
854 /* Now reallocate the pci resources memory with the
855 * highest addresses I can manage.
856 */
857 mem1 = find_resource(dev, 1|(link->link_num<<2));
858 mem2 = find_resource(dev, 2|(link->link_num<<2));
859
860 printk(BIOS_DEBUG, "base1: 0x%08Lx limit1: 0x%08Lx size: 0x%08Lx align: %d\n",
861 mem1->base, mem1->limit, mem1->size, mem1->align);
862 printk(BIOS_DEBUG, "base2: 0x%08Lx limit2: 0x%08Lx size: 0x%08Lx align: %d\n",
863 mem2->base, mem2->limit, mem2->size, mem2->align);
864
865 /* See if both resources have roughly the same limits */
866 if (((mem1->limit <= 0xffffffff) && (mem2->limit <= 0xffffffff)) ||
867 ((mem1->limit > 0xffffffff) && (mem2->limit > 0xffffffff)))
868 {
869 /* If so place the one with the most stringent alignment first */
870 if (mem2->align > mem1->align) {
871 struct resource *tmp;
872 tmp = mem1;
873 mem1 = mem2;
874 mem2 = tmp;
875 }
876 /* Now place the memory as high up as it will go */
877 mem2->base = resource_max(mem2);
878 mem1->limit = mem2->base - 1;
879 mem1->base = resource_max(mem1);
880 }
881 else {
882 /* Place the resources as high up as they will go */
883 mem2->base = resource_max(mem2);
884 mem1->base = resource_max(mem1);
885 }
886
887 printk(BIOS_DEBUG, "base1: 0x%08Lx limit1: 0x%08Lx size: 0x%08Lx align: %d\n",
888 mem1->base, mem1->limit, mem1->size, mem1->align);
889 printk(BIOS_DEBUG, "base2: 0x%08Lx limit2: 0x%08Lx size: 0x%08Lx align: %d\n",
890 mem2->base, mem2->limit, mem2->size, mem2->align);
891 }
892
893 for (res = &dev->resource_list; res; res = res->next)
894 {
895 res->flags |= IORESOURCE_ASSIGNED;
896 res->flags |= IORESOURCE_STORED;
897 report_resource_stored(dev, res, "");
898 }
899#endif
900
901 pci_tolm = 0xffffffffUL;
902 for (link = dev->link_list; link; link = link->next) {
903 pci_tolm = find_pci_tolm(link);
904 }
905
906 // FIXME handle interleaved nodes. If you fix this here, please fix
907 // amdk8, too.
908 mmio_basek = pci_tolm >> 10;
909 /* Round mmio_basek to something the processor can support */
910 mmio_basek &= ~((1 << 6) -1);
911
912 // FIXME improve mtrr.c so we don't use up all of the mtrrs with a 64M
913 // MMIO hole. If you fix this here, please fix amdk8, too.
914 /* Round the mmio hole to 64M */
915 mmio_basek &= ~((64*1024) - 1);
916
917#if CONFIG_HW_MEM_HOLE_SIZEK != 0
918 /* if the hw mem hole is already set in raminit stage, here we will compare
919 * mmio_basek and hole_basek. if mmio_basek is bigger that hole_basek and will
920 * use hole_basek as mmio_basek and we don't need to reset hole.
921 * otherwise We reset the hole to the mmio_basek
922 */
923
924 mem_hole = get_hw_mem_hole_info();
925
926 // Use hole_basek as mmio_basek, and we don't need to reset hole anymore
927 if ((mem_hole.node_id != -1) && (mmio_basek > mem_hole.hole_startk)) {
928 mmio_basek = mem_hole.hole_startk;
929 reset_memhole = 0;
930 }
931#endif
932
933 idx = 0x10;
934 for (i = 0; i < node_nums; i++) {
935 dram_base_mask_t d;
936 resource_t basek, limitk, sizek; // 4 1T
937
938 d = get_dram_base_mask(i);
939
940 if (!(d.mask & 1)) continue;
941 basek = ((resource_t)(d.base & 0x1fffff00)) << 9; // could overflow, we may lost 6 bit here
Edward O'Callaghanae5fd342014-11-20 19:58:09 +1100942 limitk = ((resource_t)(((d.mask & ~1) + 0x000FF) & 0x1fffff00)) << 9;
Siyuan Wang3e32cc02013-07-09 17:16:20 +0800943
944 sizek = limitk - basek;
945
946 /* see if we need a hole from 0xa0000 to 0xbffff */
947 if ((basek < ((8*64)+(8*16))) && (sizek > ((8*64)+(16*16)))) {
948 ram_resource(dev, (idx | i), basek, ((8*64)+(8*16)) - basek);
949 idx += 0x10;
950 basek = (8*64)+(16*16);
951 sizek = limitk - ((8*64)+(16*16));
952
953 }
954
955 //printk(BIOS_DEBUG, "node %d : mmio_basek=%08lx, basek=%08llx, limitk=%08llx\n", i, mmio_basek, basek, limitk);
956
Kyösti Mälkki26c65432014-06-26 05:30:54 +0300957 /* split the region to accommodate pci memory space */
Siyuan Wang3e32cc02013-07-09 17:16:20 +0800958 if ((basek < 4*1024*1024 ) && (limitk > mmio_basek)) {
959 if (basek <= mmio_basek) {
960 unsigned pre_sizek;
961 pre_sizek = mmio_basek - basek;
962 if (pre_sizek>0) {
963 ram_resource(dev, (idx | i), basek, pre_sizek);
964 idx += 0x10;
965 sizek -= pre_sizek;
Kyösti Mälkki2b790f62013-09-03 05:25:57 +0300966 if (!ramtop)
967 ramtop = mmio_basek * 1024;
Siyuan Wang3e32cc02013-07-09 17:16:20 +0800968 }
969 basek = mmio_basek;
970 }
971 if ((basek + sizek) <= 4*1024*1024) {
972 sizek = 0;
973 }
974 else {
975 uint64_t topmem2 = bsp_topmem2();
976 basek = 4*1024*1024;
977 sizek = topmem2/1024 - basek;
978 }
979 }
980
981 ram_resource(dev, (idx | i), basek, sizek);
982 idx += 0x10;
983 printk(BIOS_DEBUG, "node %d: mmio_basek=%08lx, basek=%08llx, limitk=%08llx\n",
984 i, mmio_basek, basek, limitk);
Kyösti Mälkki2b790f62013-09-03 05:25:57 +0300985 if (!ramtop)
986 ramtop = limitk * 1024;
Siyuan Wang3e32cc02013-07-09 17:16:20 +0800987 }
988
989#if CONFIG_GFXUMA
Kyösti Mälkki2b790f62013-09-03 05:25:57 +0300990 set_top_of_ram(uma_memory_base);
Siyuan Wang3e32cc02013-07-09 17:16:20 +0800991 uma_resource(dev, 7, uma_memory_base >> 10, uma_memory_size >> 10);
Kyösti Mälkki2b790f62013-09-03 05:25:57 +0300992#else
993 set_top_of_ram(ramtop);
Siyuan Wang3e32cc02013-07-09 17:16:20 +0800994#endif
995
996 for(link = dev->link_list; link; link = link->next) {
997 if (link->children) {
998 assign_resources(link);
999 }
1000 }
1001}
1002
1003static struct device_operations pci_domain_ops = {
1004 .read_resources = domain_read_resources,
1005 .set_resources = domain_set_resources,
1006 .enable_resources = domain_enable_resources,
1007 .init = NULL,
1008 .scan_bus = pci_domain_scan_bus,
1009 .ops_pci_bus = pci_bus_default_ops,
1010};
1011
1012static void sysconf_init(device_t dev) // first node
1013{
1014 sblink = (pci_read_config32(dev, 0x64)>>8) & 7; // don't forget sublink1
1015 node_nums = ((pci_read_config32(dev, 0x60)>>4) & 7) + 1; //NodeCnt[2:0]
1016}
1017
1018static void add_more_links(device_t dev, unsigned total_links)
1019{
1020 struct bus *link, *last = NULL;
1021 int link_num;
1022
1023 for (link = dev->link_list; link; link = link->next)
1024 last = link;
1025
1026 if (last) {
1027 int links = total_links - last->link_num;
1028 link_num = last->link_num;
1029 if (links > 0) {
1030 link = malloc(links*sizeof(*link));
1031 if (!link)
1032 die("Couldn't allocate more links!\n");
1033 memset(link, 0, links*sizeof(*link));
1034 last->next = link;
1035 }
1036 }
1037 else {
1038 link_num = -1;
1039 link = malloc(total_links*sizeof(*link));
1040 memset(link, 0, total_links*sizeof(*link));
1041 dev->link_list = link;
1042 }
1043
1044 for (link_num = link_num + 1; link_num < total_links; link_num++) {
1045 link->link_num = link_num;
1046 link->dev = dev;
1047 link->next = link + 1;
1048 last = link;
1049 link = link->next;
1050 }
1051 last->next = NULL;
1052}
1053
1054static u32 cpu_bus_scan(device_t dev, u32 max)
1055{
1056 struct bus *cpu_bus;
1057 device_t dev_mc;
1058#if CONFIG_CBB
1059 device_t pci_domain;
1060#endif
1061 int i,j;
1062 int coreid_bits;
1063 int core_max = 0;
1064 unsigned ApicIdCoreIdSize;
1065 unsigned core_nums;
1066 int siblings = 0;
1067 unsigned int family;
1068
1069#if CONFIG_CBB
1070 dev_mc = dev_find_slot(0, PCI_DEVFN(CONFIG_CDB, 0)); //0x00
1071 if (dev_mc && dev_mc->bus) {
1072 printk(BIOS_DEBUG, "%s found", dev_path(dev_mc));
1073 pci_domain = dev_mc->bus->dev;
1074 if (pci_domain && (pci_domain->path.type == DEVICE_PATH_DOMAIN)) {
1075 printk(BIOS_DEBUG, "\n%s move to ",dev_path(dev_mc));
1076 dev_mc->bus->secondary = CONFIG_CBB; // move to 0xff
1077 printk(BIOS_DEBUG, "%s",dev_path(dev_mc));
1078 } else {
1079 printk(BIOS_DEBUG, " but it is not under pci_domain directly ");
1080 }
1081 printk(BIOS_DEBUG, "\n");
1082 }
1083 dev_mc = dev_find_slot(CONFIG_CBB, PCI_DEVFN(CONFIG_CDB, 0));
1084 if (!dev_mc) {
1085 dev_mc = dev_find_slot(0, PCI_DEVFN(0x18, 0));
1086 if (dev_mc && dev_mc->bus) {
1087 printk(BIOS_DEBUG, "%s found\n", dev_path(dev_mc));
1088 pci_domain = dev_mc->bus->dev;
1089 if (pci_domain && (pci_domain->path.type == DEVICE_PATH_DOMAIN)) {
1090 if ((pci_domain->link_list) && (pci_domain->link_list->children == dev_mc)) {
1091 printk(BIOS_DEBUG, "%s move to ",dev_path(dev_mc));
1092 dev_mc->bus->secondary = CONFIG_CBB; // move to 0xff
1093 printk(BIOS_DEBUG, "%s\n",dev_path(dev_mc));
1094 while (dev_mc) {
1095 printk(BIOS_DEBUG, "%s move to ",dev_path(dev_mc));
1096 dev_mc->path.pci.devfn -= PCI_DEVFN(0x18,0);
1097 printk(BIOS_DEBUG, "%s\n",dev_path(dev_mc));
1098 dev_mc = dev_mc->sibling;
1099 }
1100 }
1101 }
1102 }
1103 }
1104#endif
1105 dev_mc = dev_find_slot(CONFIG_CBB, PCI_DEVFN(CONFIG_CDB, 0));
1106 if (!dev_mc) {
1107 printk(BIOS_ERR, "%02x:%02x.0 not found", CONFIG_CBB, CONFIG_CDB);
1108 die("");
1109 }
1110 sysconf_init(dev_mc);
1111#if CONFIG_CBB && (MAX_NODE_NUMS > 32)
1112 if (node_nums>32) { // need to put node 32 to node 63 to bus 0xfe
1113 if (pci_domain->link_list && !pci_domain->link_list->next) {
1114 struct bus *new_link = new_link(pci_domain);
1115 pci_domain->link_list->next = new_link;
1116 new_link->link_num = 1;
1117 new_link->dev = pci_domain;
1118 new_link->children = 0;
1119 printk(BIOS_DEBUG, "%s links now 2\n", dev_path(pci_domain));
1120 }
1121 pci_domain->link_list->next->secondary = CONFIG_CBB - 1;
1122 }
1123#endif
1124
1125 /* Get Max Number of cores(MNC) */
1126 coreid_bits = (cpuid_ecx(AMD_CPUID_ASIZE_PCCOUNT) & 0x0000F000) >> 12;
1127 core_max = 1 << (coreid_bits & 0x000F); //mnc
1128
1129 ApicIdCoreIdSize = ((cpuid_ecx(0x80000008)>>12) & 0xF);
1130 if (ApicIdCoreIdSize) {
1131 core_nums = (1 << ApicIdCoreIdSize) - 1;
1132 } else {
1133 core_nums = 3; //quad core
1134 }
1135
1136 /* Find which cpus are present */
1137 cpu_bus = dev->link_list;
1138 for (i = 0; i < node_nums; i++) {
1139 device_t cdb_dev;
1140 unsigned busn, devn;
1141 struct bus *pbus;
1142
1143 busn = CONFIG_CBB;
1144 devn = CONFIG_CDB + i;
1145 pbus = dev_mc->bus;
1146#if CONFIG_CBB && (MAX_NODE_NUMS > 32)
1147 if (i >= 32) {
1148 busn--;
1149 devn -= 32;
1150 pbus = pci_domain->link_list->next;
1151 }
1152#endif
1153
1154 /* Find the cpu's pci device */
1155 cdb_dev = dev_find_slot(busn, PCI_DEVFN(devn, 0));
1156 if (!cdb_dev) {
1157 /* If I am probing things in a weird order
1158 * ensure all of the cpu's pci devices are found.
1159 */
1160 int fn;
1161 for(fn = 0; fn <= 5; fn++) { //FBDIMM?
1162 cdb_dev = pci_probe_dev(NULL, pbus,
1163 PCI_DEVFN(devn, fn));
1164 }
1165 cdb_dev = dev_find_slot(busn, PCI_DEVFN(devn, 0));
1166 } else {
1167 /* Ok, We need to set the links for that device.
1168 * otherwise the device under it will not be scanned
1169 */
1170 int linknum;
1171#if CONFIG_HT3_SUPPORT
1172 linknum = 8;
1173#else
1174 linknum = 4;
1175#endif
1176 add_more_links(cdb_dev, linknum);
1177 }
1178
1179 family = cpuid_eax(1);
1180 family = (family >> 20) & 0xFF;
1181 if (family == 1) { //f10
1182 u32 dword;
1183 cdb_dev = dev_find_slot(busn, PCI_DEVFN(devn, 3));
1184 dword = pci_read_config32(cdb_dev, 0xe8);
1185 siblings = ((dword & BIT15) >> 13) | ((dword & (BIT13 | BIT12)) >> 12);
1186 } else if (family == 7) {//f16
1187 cdb_dev = dev_find_slot(busn, PCI_DEVFN(devn, 5));
1188 if (cdb_dev && cdb_dev->enabled) {
1189 siblings = pci_read_config32(cdb_dev, 0x84);
1190 siblings &= 0xFF;
1191 }
1192 } else {
1193 siblings = 0; //default one core
1194 }
1195 int enable_node = cdb_dev && cdb_dev->enabled;
1196 printk(BIOS_SPEW, "%s family%xh, core_max=0x%x, core_nums=0x%x, siblings=0x%x\n",
1197 dev_path(cdb_dev), 0x0f + family, core_max, core_nums, siblings);
1198
1199 for (j = 0; j <= siblings; j++ ) {
1200 extern CONST OPTIONS_CONFIG_TOPOLOGY ROMDATA TopologyConfiguration;
1201 u32 modules = TopologyConfiguration.PlatformNumberOfModules;
1202 u32 lapicid_start = 0;
1203
1204 /*
1205 * APIC ID calucation is tightly coupled with AGESA v5 code.
1206 * This calculation MUST match the assignment calculation done
1207 * in LocalApicInitializationAtEarly() function.
1208 * And reference GetLocalApicIdForCore()
1209 *
1210 * Apply apic enumeration rules
1211 * For systems with >= 16 APICs, put the IO-APICs at 0..n and
1212 * put the local-APICs at m..z
1213 *
1214 * This is needed because many IO-APIC devices only have 4 bits
1215 * for their APIC id and therefore must reside at 0..15
1216 */
1217#ifndef CFG_PLAT_NUM_IO_APICS /* defined in mainboard buildOpts.c */
1218#define CFG_PLAT_NUM_IO_APICS 3
1219#endif
1220 if ((node_nums * core_max) + CFG_PLAT_NUM_IO_APICS >= 0x10) {
1221 lapicid_start = (CFG_PLAT_NUM_IO_APICS - 1) / core_max;
1222 lapicid_start = (lapicid_start + 1) * core_max;
1223 printk(BIOS_SPEW, "lpaicid_start=0x%x ", lapicid_start);
1224 }
1225 u32 apic_id = (lapicid_start * (i/modules + 1)) + ((i % modules) ? (j + (siblings + 1)) : j);
1226 printk(BIOS_SPEW, "node 0x%x core 0x%x apicid=0x%x\n",
1227 i, j, apic_id);
1228
1229 device_t cpu = add_cpu_device(cpu_bus, apic_id, enable_node);
1230 if (cpu)
1231 amd_cpu_topology(cpu, i, j);
1232 } //j
1233 }
1234 return max;
1235}
1236
1237static void cpu_bus_init(device_t dev)
1238{
1239 initialize_cpus(dev->link_list);
1240}
1241
Siyuan Wang3e32cc02013-07-09 17:16:20 +08001242static struct device_operations cpu_bus_ops = {
Edward O'Callaghan66c65322014-11-21 01:43:38 +11001243 .read_resources = DEVICE_NOOP,
1244 .set_resources = DEVICE_NOOP,
Edward O'Callaghan812d2a42014-10-31 08:17:23 +11001245 .enable_resources = DEVICE_NOOP,
Siyuan Wang3e32cc02013-07-09 17:16:20 +08001246 .init = cpu_bus_init,
1247 .scan_bus = cpu_bus_scan,
1248};
1249
1250static void root_complex_enable_dev(struct device *dev)
1251{
1252 static int done = 0;
1253
1254 /* Do not delay UMA setup, as a device on the PCI bus may evaluate
1255 the global uma_memory variables already in its enable function. */
1256 if (!done) {
1257 setup_bsp_ramtop();
1258 setup_uma_memory();
1259 done = 1;
1260 }
1261
1262 /* Set the operations if it is a special bus type */
1263 if (dev->path.type == DEVICE_PATH_DOMAIN) {
1264 dev->ops = &pci_domain_ops;
1265 } else if (dev->path.type == DEVICE_PATH_CPU_CLUSTER) {
1266 dev->ops = &cpu_bus_ops;
1267 }
1268}
1269
1270struct chip_operations northbridge_amd_agesa_family16kb_root_complex_ops = {
1271 CHIP_NAME("AMD FAM16 Root Complex")
1272 .enable_dev = root_complex_enable_dev,
1273};
Bruce Griffith76db07e2013-07-07 02:06:53 -06001274
1275/*********************************************************************
1276 * Change the vendor / device IDs to match the generic VBIOS header. *
1277 *********************************************************************/
1278u32 map_oprom_vendev(u32 vendev)
1279{
1280 u32 new_vendev = vendev;
1281
1282 switch(vendev) {
1283 case 0x10029830:
1284 case 0x10029831:
1285 case 0x10029832:
1286 case 0x10029833:
1287 case 0x10029834:
1288 case 0x10029835:
1289 case 0x10029836:
1290 case 0x10029837:
1291 case 0x10029838:
1292 case 0x10029839:
1293 case 0x1002983A:
1294 case 0x1002983D:
1295 new_vendev = 0x10029830; // This is the default value in AMD-generated VBIOS
1296 break;
1297 default:
1298 break;
1299 }
1300
1301 if (vendev != new_vendev)
1302 printk(BIOS_NOTICE, "Mapping PCI device %8x to %8x\n", vendev, new_vendev);
1303
1304 return new_vendev;
1305}