blob: 430eeef6f620176292166b19272519fd8b07f739 [file] [log] [blame]
Angel Pons4b429832020-04-02 23:48:50 +02001/* SPDX-License-Identifier: GPL-2.0-only */
Bruce Griffith27ed80b2014-08-15 11:46:25 -06002
Michał Żygowski2f399b72020-04-02 19:51:37 +02003#include <commonlib/helpers.h>
Bruce Griffith27ed80b2014-08-15 11:46:25 -06004#include <console/console.h>
Kyösti Mälkkif1b58b72019-03-01 13:43:02 +02005#include <device/pci_ops.h>
Furquan Shaikh76cedd22020-05-02 10:24:23 -07006#include <acpi/acpi.h>
7#include <acpi/acpi_ivrs.h>
Michał Żygowski208318c2020-03-20 15:54:27 +01008#include <arch/ioapic.h>
Felix Held61dd31c2023-06-05 19:38:36 +02009#include <arch/vga.h>
Elyes HAOUAS146d0c22020-07-22 11:47:08 +020010#include <types.h>
Bruce Griffith27ed80b2014-08-15 11:46:25 -060011#include <device/device.h>
12#include <device/pci.h>
13#include <device/pci_ids.h>
Bruce Griffith27ed80b2014-08-15 11:46:25 -060014#include <string.h>
Michał Żygowski2f399b72020-04-02 19:51:37 +020015#include <stdlib.h>
Michał Kopećdc35d2a2021-11-30 17:40:52 +010016#include <cpu/x86/mp.h>
Bruce Griffith27ed80b2014-08-15 11:46:25 -060017#include <Porting.h>
Bruce Griffith27ed80b2014-08-15 11:46:25 -060018#include <Topology.h>
Elyes HAOUAS400ce552018-10-12 10:54:30 +020019#include <cpu/amd/msr.h>
20#include <cpu/amd/mtrr.h>
Furquan Shaikh76cedd22020-05-02 10:24:23 -070021#include <acpi/acpigen.h>
Angel Ponsec5cf152020-11-10 20:42:07 +010022#include <northbridge/amd/nb_common.h>
Kyösti Mälkkied8d2772017-07-15 17:12:44 +030023#include <northbridge/amd/agesa/agesa_helper.h>
Michał Żygowski2f399b72020-04-02 19:51:37 +020024#include <southbridge/amd/pi/hudson/pci_devs.h>
Arthur Heymans44807ac2022-09-13 12:43:37 +020025#include <amdblocks/cpu.h>
Bruce Griffith27ed80b2014-08-15 11:46:25 -060026
Michał Żygowski6ca5b472019-09-10 15:10:22 +020027#define PCIE_CAP_AER BIT(5)
28#define PCIE_CAP_ACS BIT(6)
Bruce Griffith27ed80b2014-08-15 11:46:25 -060029
Felix Held3eaa8502023-12-16 01:37:34 +010030static int get_dram_base_limit(resource_t *basek, resource_t *limitk)
Michał Żygowski88a0ce62021-05-05 09:52:59 +020031{
32 u32 temp;
33
Felix Held7a83ab72023-12-16 23:10:50 +010034 temp = pci_read_config32(DEV_PTR(ht_1), 0x40); //[39:24] at [31:16]
Michał Żygowski88a0ce62021-05-05 09:52:59 +020035 if (!(temp & 1))
36 return 0; // this memory range is not enabled
37 /*
38 * BKDG: {DramBase[39:24], 00_0000h} <= address[39:0] so shift left by 8 bits
39 * for physical address and the convert to KiB by shifting 10 bits left
40 */
41 *basek = ((temp & 0xffff0000)) >> (10 - 8);
42 /*
43 * BKDG address[39:0] <= {DramLimit[39:24], FF_FFFFh} converted as above but
44 * ORed with 0xffff to get real limit before shifting.
45 */
Felix Held7a83ab72023-12-16 23:10:50 +010046 temp = pci_read_config32(DEV_PTR(ht_1), 0x44); //[39:24] at [31:16]
Michał Żygowski88a0ce62021-05-05 09:52:59 +020047 *limitk = ((temp & 0xffff0000) | 0xffff) >> (10 - 8);
48 *limitk += 1; // round up last byte
49
50 return 1;
51}
52
Michał Żygowski58d6f962021-05-05 10:52:08 +020053static void add_fixed_resources(struct device *dev, int index)
54{
55 /* Reserve everything between A segment and 1MB:
56 *
57 * 0xa0000 - 0xbffff: legacy VGA
58 * 0xc0000 - 0xfffff: option ROMs and SeaBIOS (if used)
59 */
Felix Held61dd31c2023-06-05 19:38:36 +020060 mmio_resource_kb(dev, index++, VGA_MMIO_BASE >> 10, VGA_MMIO_SIZE >> 10);
Kyösti Mälkki27d62992022-05-24 20:25:58 +030061 reserved_ram_resource_kb(dev, index++, 0xc0000 >> 10, (0x100000 - 0xc0000) >> 10);
Michał Żygowski58d6f962021-05-05 10:52:08 +020062
Michał Żygowski58d6f962021-05-05 10:52:08 +020063 /* Check if CC6 save area is enabled (bit 18 CC6SaveEn) */
Felix Held7a83ab72023-12-16 23:10:50 +010064 if (pci_read_config32(DEV_PTR(ht_2), 0x118) & (1 << 18)) {
Michał Żygowski58d6f962021-05-05 10:52:08 +020065 /* Add CC6 DRAM UC resource residing at DRAM Limit of size 16MB as per BKDG */
66 resource_t basek, limitk;
Felix Held3eaa8502023-12-16 01:37:34 +010067 if (!get_dram_base_limit(&basek, &limitk))
Michał Żygowski58d6f962021-05-05 10:52:08 +020068 return;
Elyes Haouasf9b535e2022-07-16 09:47:42 +020069 mmio_resource_kb(dev, index++, limitk, 16 * 1024);
Michał Żygowski58d6f962021-05-05 10:52:08 +020070 }
71}
72
Michał Żygowskifb198c62021-05-09 13:54:09 +020073static void nb_read_resources(struct device *dev)
Bruce Griffith27ed80b2014-08-15 11:46:25 -060074{
Kyösti Mälkki5d490382015-05-27 07:58:22 +030075 /*
76 * This MMCONF resource must be reserved in the PCI domain.
77 * It is not honored by the coreboot resource allocator if it is in
78 * the CPU_CLUSTER.
79 */
Elyes HAOUAS400ce552018-10-12 10:54:30 +020080 mmconf_resource(dev, MMIO_CONF_BASE);
Michał Żygowski208318c2020-03-20 15:54:27 +010081
82 /* NB IOAPIC2 resource */
Felix Held8f0075c2023-08-09 19:28:39 +020083 mmio_range(dev, IO_APIC2_ADDR, IO_APIC2_ADDR, 0x1000);
Michał Żygowski58d6f962021-05-05 10:52:08 +020084
85 add_fixed_resources(dev, 0);
Bruce Griffith27ed80b2014-08-15 11:46:25 -060086}
87
Bruce Griffith27ed80b2014-08-15 11:46:25 -060088static void northbridge_init(struct device *dev)
89{
Felix Held0d192892024-02-06 16:55:29 +010090 register_new_ioapic(IO_APIC2_ADDR);
Bruce Griffith27ed80b2014-08-15 11:46:25 -060091}
Kyösti Mälkki0b5b5412014-11-26 08:11:07 +020092
Vladimir Serbinenko807127f2014-11-09 13:36:18 +010093static unsigned long acpi_fill_hest(acpi_hest_t *hest)
Kyösti Mälkki0b5b5412014-11-26 08:11:07 +020094{
95 void *addr, *current;
96
97 /* Skip the HEST header. */
98 current = (void *)(hest + 1);
99
100 addr = agesawrapper_getlateinitptr(PICK_WHEA_MCE);
101 if (addr != NULL)
102 current += acpi_create_hest_error_source(hest, current, 0, (void *)((u32)addr + 2), *(UINT16 *)addr - 2);
103
104 addr = agesawrapper_getlateinitptr(PICK_WHEA_CMC);
105 if (addr != NULL)
106 current += acpi_create_hest_error_source(hest, current, 1, (void *)((u32)addr + 2), *(UINT16 *)addr - 2);
107
108 return (unsigned long)current;
109}
110
Arthur Heymansf9ee87f2023-06-07 15:29:02 +0200111static unsigned long acpi_fill_ivrs_ioapic(acpi_ivrs_t *ivrs, unsigned long current)
Timothy Pearson9ef07d82016-06-13 13:48:58 -0500112{
Michał Żygowski2f399b72020-04-02 19:51:37 +0200113 /* 8-byte IVHD structures must be aligned to the 8-byte boundary. */
114 current = ALIGN_UP(current, 8);
115 ivrs_ivhd_special_t *ivhd_ioapic = (ivrs_ivhd_special_t *)current;
Timothy Pearson9ef07d82016-06-13 13:48:58 -0500116
Michał Żygowski2f399b72020-04-02 19:51:37 +0200117 ivhd_ioapic->type = IVHD_DEV_8_BYTE_EXT_SPECIAL_DEV;
118 ivhd_ioapic->reserved = 0x0000;
119 ivhd_ioapic->dte_setting = IVHD_DTE_LINT_1_PASS | IVHD_DTE_LINT_0_PASS |
120 IVHD_DTE_SYS_MGT_NO_TRANS | IVHD_DTE_NMI_PASS |
121 IVHD_DTE_EXT_INT_PASS | IVHD_DTE_INIT_PASS;
Felix Held0d192892024-02-06 16:55:29 +0100122 ivhd_ioapic->handle = get_ioapic_id(IO_APIC_ADDR);
Michał Żygowski2f399b72020-04-02 19:51:37 +0200123 ivhd_ioapic->source_dev_id = PCI_DEVFN(SMBUS_DEV, SMBUS_FUNC);
124 ivhd_ioapic->variety = IVHD_SPECIAL_DEV_IOAPIC;
125 current += sizeof(ivrs_ivhd_special_t);
126
127 ivhd_ioapic = (ivrs_ivhd_special_t *)current;
Michał Żygowski2f399b72020-04-02 19:51:37 +0200128 ivhd_ioapic->type = IVHD_DEV_8_BYTE_EXT_SPECIAL_DEV;
129 ivhd_ioapic->reserved = 0x0000;
130 ivhd_ioapic->dte_setting = 0x00;
Felix Held0d192892024-02-06 16:55:29 +0100131 ivhd_ioapic->handle = get_ioapic_id(IO_APIC2_ADDR);
Michał Żygowski2f399b72020-04-02 19:51:37 +0200132 ivhd_ioapic->source_dev_id = PCI_DEVFN(0, 1);
133 ivhd_ioapic->variety = IVHD_SPECIAL_DEV_IOAPIC;
134 current += sizeof(ivrs_ivhd_special_t);
135
136 return current;
137}
138
139static unsigned long ivhd_describe_hpet(unsigned long current)
140{
141 /* 8-byte IVHD structures must be aligned to the 8-byte boundary. */
142 current = ALIGN_UP(current, 8);
143 ivrs_ivhd_special_t *ivhd_hpet = (ivrs_ivhd_special_t *)current;
144
145 ivhd_hpet->type = IVHD_DEV_8_BYTE_EXT_SPECIAL_DEV;
146 ivhd_hpet->reserved = 0x0000;
147 ivhd_hpet->dte_setting = 0x00;
148 ivhd_hpet->handle = 0x00;
149 ivhd_hpet->source_dev_id = PCI_DEVFN(SMBUS_DEV, SMBUS_FUNC);
150 ivhd_hpet->variety = IVHD_SPECIAL_DEV_HPET;
151 current += sizeof(ivrs_ivhd_special_t);
152
153 return current;
154}
155
156static unsigned long ivhd_dev_range(unsigned long current, uint16_t start_devid,
157 uint16_t end_devid, uint8_t setting)
158{
159 /* 4-byte IVHD structures must be aligned to the 4-byte boundary. */
160 current = ALIGN_UP(current, 4);
161 ivrs_ivhd_generic_t *ivhd_range = (ivrs_ivhd_generic_t *)current;
162
163 /* Create the start range IVHD entry */
164 ivhd_range->type = IVHD_DEV_4_BYTE_START_RANGE;
165 ivhd_range->dev_id = start_devid;
166 ivhd_range->dte_setting = setting;
167 current += sizeof(ivrs_ivhd_generic_t);
168
169 /* Create the end range IVHD entry */
170 ivhd_range = (ivrs_ivhd_generic_t *)current;
171 ivhd_range->type = IVHD_DEV_4_BYTE_END_RANGE;
172 ivhd_range->dev_id = end_devid;
173 ivhd_range->dte_setting = setting;
174 current += sizeof(ivrs_ivhd_generic_t);
175
176 return current;
177}
178
179static unsigned long add_ivhd_dev_entry(struct device *parent, struct device *dev,
180 unsigned long *current, uint8_t type, uint8_t data)
181{
182 if (type == IVHD_DEV_4_BYTE_SELECT) {
183 /* 4-byte IVHD structures must be aligned to the 4-byte boundary. */
184 *current = ALIGN_UP(*current, 4);
185 ivrs_ivhd_generic_t *ivhd_entry = (ivrs_ivhd_generic_t *)*current;
186
187 ivhd_entry->type = type;
Arthur Heymans7fcd4d52023-08-24 15:12:19 +0200188 ivhd_entry->dev_id = dev->path.pci.devfn | (dev->upstream->secondary << 8);
Michał Żygowski2f399b72020-04-02 19:51:37 +0200189 ivhd_entry->dte_setting = data;
190 *current += sizeof(ivrs_ivhd_generic_t);
191 } else if (type == IVHD_DEV_8_BYTE_ALIAS_SELECT) {
192 /* 8-byte IVHD structures must be aligned to the 8-byte boundary. */
193 *current = ALIGN_UP(*current, 8);
194 ivrs_ivhd_alias_t *ivhd_entry = (ivrs_ivhd_alias_t *)*current;
195
196 ivhd_entry->type = type;
Arthur Heymans7fcd4d52023-08-24 15:12:19 +0200197 ivhd_entry->dev_id = dev->path.pci.devfn | (dev->upstream->secondary << 8);
Michał Żygowski2f399b72020-04-02 19:51:37 +0200198 ivhd_entry->dte_setting = data;
199 ivhd_entry->reserved1 = 0;
200 ivhd_entry->reserved2 = 0;
201 ivhd_entry->source_dev_id = parent->path.pci.devfn |
Arthur Heymans7fcd4d52023-08-24 15:12:19 +0200202 (parent->upstream->secondary << 8);
Michał Żygowski2f399b72020-04-02 19:51:37 +0200203 *current += sizeof(ivrs_ivhd_alias_t);
204 }
205
206 return *current;
207}
208
209static void ivrs_add_device_or_bridge(struct device *parent, struct device *dev,
210 unsigned long *current, uint16_t *ivhd_length)
211{
212 unsigned int header_type, is_pcie;
213 unsigned long current_backup;
214
215 header_type = dev->hdr_type & 0x7f;
216 is_pcie = pci_find_capability(dev, PCI_CAP_ID_PCIE);
217
218 if (((header_type == PCI_HEADER_TYPE_NORMAL) ||
219 (header_type == PCI_HEADER_TYPE_BRIDGE)) && is_pcie) {
220 /* Device or Bridge is PCIe */
221 current_backup = *current;
222 add_ivhd_dev_entry(parent, dev, current, IVHD_DEV_4_BYTE_SELECT, 0x0);
223 *ivhd_length += (*current - current_backup);
224 } else if ((header_type == PCI_HEADER_TYPE_NORMAL) && !is_pcie) {
225 /* Device is legacy PCI or PCI-X */
226 current_backup = *current;
227 add_ivhd_dev_entry(parent, dev, current, IVHD_DEV_8_BYTE_ALIAS_SELECT, 0x0);
228 *ivhd_length += (*current - current_backup);
Timothy Pearson9ef07d82016-06-13 13:48:58 -0500229 }
230}
231
Michał Żygowski2f399b72020-04-02 19:51:37 +0200232static void add_ivhd_device_entries(struct device *parent, struct device *dev,
Timothy Pearson9ef07d82016-06-13 13:48:58 -0500233 unsigned int depth, int linknum, int8_t *root_level,
Michał Żygowski2f399b72020-04-02 19:51:37 +0200234 unsigned long *current, uint16_t *ivhd_length)
Timothy Pearson9ef07d82016-06-13 13:48:58 -0500235{
236 struct device *sibling;
Michał Żygowski2f399b72020-04-02 19:51:37 +0200237
238 if (!root_level) {
239 root_level = malloc(sizeof(int8_t));
240 *root_level = -1;
241 }
Timothy Pearson9ef07d82016-06-13 13:48:58 -0500242
Timothy Pearson9ef07d82016-06-13 13:48:58 -0500243 if (dev->path.type == DEVICE_PATH_PCI) {
244
Arthur Heymans7fcd4d52023-08-24 15:12:19 +0200245 if ((dev->upstream->secondary == 0x0) &&
Timothy Pearson9ef07d82016-06-13 13:48:58 -0500246 (dev->path.pci.devfn == 0x0))
247 *root_level = depth;
248
249 if ((*root_level != -1) && (dev->enabled)) {
Michał Żygowski2f399b72020-04-02 19:51:37 +0200250 if (depth != *root_level)
251 ivrs_add_device_or_bridge(parent, dev, current, ivhd_length);
Timothy Pearson9ef07d82016-06-13 13:48:58 -0500252 }
253 }
254
Arthur Heymans7fcd4d52023-08-24 15:12:19 +0200255 if (dev->downstream) {
256 for (sibling = dev->downstream->children; sibling; sibling = sibling->sibling)
Michał Żygowski2f399b72020-04-02 19:51:37 +0200257 add_ivhd_device_entries(dev, sibling, depth + 1, depth, root_level,
258 current, ivhd_length);
Arthur Heymans80c79a52023-08-24 15:12:19 +0200259 }
Michał Żygowski2f399b72020-04-02 19:51:37 +0200260
261 free(root_level);
Timothy Pearson9ef07d82016-06-13 13:48:58 -0500262}
263
Michał Żygowski2f399b72020-04-02 19:51:37 +0200264#define IOMMU_MMIO32(x) (*((volatile uint32_t *)(x)))
265#define EFR_SUPPORT BIT(27)
266
267static unsigned long acpi_fill_ivrs11(unsigned long current, acpi_ivrs_t *ivrs_agesa)
Timothy Pearson9ef07d82016-06-13 13:48:58 -0500268{
Michał Żygowski2f399b72020-04-02 19:51:37 +0200269 acpi_ivrs_ivhd11_t *ivhd_11;
270 unsigned long current_backup;
Timothy Pearson9ef07d82016-06-13 13:48:58 -0500271
Michał Żygowski2f399b72020-04-02 19:51:37 +0200272 /*
273 * These devices should be already found by previous function.
274 * Do not perform NULL checks.
275 */
276 struct device *nb_dev = pcidev_on_root(0, 0);
277 struct device *iommu_dev = pcidev_on_root(0, 2);
Timothy Pearson9ef07d82016-06-13 13:48:58 -0500278
Michał Żygowski2f399b72020-04-02 19:51:37 +0200279 /*
280 * In order to utilize all features, firmware should expose type 11h
281 * IVHD which supersedes the type 10h.
282 */
283 memset((void *)current, 0, sizeof(acpi_ivrs_ivhd11_t));
284 ivhd_11 = (acpi_ivrs_ivhd11_t *)current;
Timothy Pearson9ef07d82016-06-13 13:48:58 -0500285
Michał Żygowski2f399b72020-04-02 19:51:37 +0200286 /* Enable EFR */
287 ivhd_11->type = IVHD_BLOCK_TYPE_FULL__FIXED;
288 /* For type 11h bits 6 and 7 are reserved */
289 ivhd_11->flags = ivrs_agesa->ivhd.flags & 0x3f;
290 ivhd_11->length = sizeof(struct acpi_ivrs_ivhd_11);
291 /* BDF <bus>:00.2 */
Arthur Heymans7fcd4d52023-08-24 15:12:19 +0200292 ivhd_11->device_id = 0x02 | (nb_dev->upstream->secondary << 8);
Michał Żygowski2f399b72020-04-02 19:51:37 +0200293 /* PCI Capability block 0x40 (type 0xf, "Secure device") */
294 ivhd_11->capability_offset = 0x40;
295 ivhd_11->iommu_base_low = ivrs_agesa->ivhd.iommu_base_low;
296 ivhd_11->iommu_base_high = ivrs_agesa->ivhd.iommu_base_high;
Arthur Heymans7fcd4d52023-08-24 15:12:19 +0200297 ivhd_11->pci_segment_group = nb_dev->upstream->segment_group;
Michał Żygowski2f399b72020-04-02 19:51:37 +0200298 ivhd_11->iommu_info = ivrs_agesa->ivhd.iommu_info;
299 ivhd_11->iommu_attributes.perf_counters =
300 (IOMMU_MMIO32(ivhd_11->iommu_base_low + 0x4000) >> 7) & 0xf;
301 ivhd_11->iommu_attributes.perf_counter_banks =
302 (IOMMU_MMIO32(ivhd_11->iommu_base_low + 0x4000) >> 12) & 0x3f;
303 ivhd_11->iommu_attributes.msi_num_ppr =
304 (pci_read_config32(iommu_dev, ivhd_11->capability_offset + 0x10) >> 27) & 0x1f;
Timothy Pearson9ef07d82016-06-13 13:48:58 -0500305
Michał Żygowski2f399b72020-04-02 19:51:37 +0200306 if (pci_read_config32(iommu_dev, ivhd_11->capability_offset) & EFR_SUPPORT) {
307 ivhd_11->efr_reg_image_low = IOMMU_MMIO32(ivhd_11->iommu_base_low + 0x30);
308 ivhd_11->efr_reg_image_high = IOMMU_MMIO32(ivhd_11->iommu_base_low + 0x34);
309 }
310
311 current += sizeof(acpi_ivrs_ivhd11_t);
312
313 /* Now repeat all the device entries from type 10h */
314 current_backup = current;
315 current = ivhd_dev_range(current, PCI_DEVFN(1, 0), PCI_DEVFN(0x1f, 6), 0);
316 ivhd_11->length += (current - current_backup);
317 add_ivhd_device_entries(NULL, all_devices, 0, -1, NULL, &current, &ivhd_11->length);
318
319 /* Describe HPET */
320 current_backup = current;
321 current = ivhd_describe_hpet(current);
322 ivhd_11->length += (current - current_backup);
323
324 /* Describe IOAPICs */
325 current_backup = current;
326 current = acpi_fill_ivrs_ioapic(ivrs_agesa, current);
327 ivhd_11->length += (current - current_backup);
Timothy Pearson9ef07d82016-06-13 13:48:58 -0500328
329 return current;
330}
331
332static unsigned long acpi_fill_ivrs(acpi_ivrs_t *ivrs, unsigned long current)
333{
Piotr Król063e1562018-07-22 20:52:26 +0200334 acpi_ivrs_t *ivrs_agesa;
Michał Żygowski2f399b72020-04-02 19:51:37 +0200335 unsigned long current_backup;
Timothy Pearson9ef07d82016-06-13 13:48:58 -0500336
Michał Żygowski2f399b72020-04-02 19:51:37 +0200337 struct device *nb_dev = pcidev_on_root(0, 0);
Timothy Pearson9ef07d82016-06-13 13:48:58 -0500338 if (!nb_dev) {
Timothy Pearson9ef07d82016-06-13 13:48:58 -0500339 printk(BIOS_WARNING, "%s: G-series northbridge device not present!\n", __func__);
340 printk(BIOS_WARNING, "%s: IVRS table not generated...\n", __func__);
341
342 return (unsigned long)ivrs;
343 }
344
Michał Żygowski2f399b72020-04-02 19:51:37 +0200345 struct device *iommu_dev = pcidev_on_root(0, 2);
Timothy Pearson9ef07d82016-06-13 13:48:58 -0500346
Michał Żygowski2f399b72020-04-02 19:51:37 +0200347 if (!iommu_dev) {
348 printk(BIOS_WARNING, "%s: IOMMU device not found\n", __func__);
349
350 return (unsigned long)ivrs;
351 }
352
Piotr Król063e1562018-07-22 20:52:26 +0200353 ivrs_agesa = agesawrapper_getlateinitptr(PICK_IVRS);
354 if (ivrs_agesa != NULL) {
Michał Żygowski2f399b72020-04-02 19:51:37 +0200355 ivrs->iv_info = ivrs_agesa->iv_info;
356 ivrs->ivhd.type = IVHD_BLOCK_TYPE_LEGACY__FIXED;
357 ivrs->ivhd.flags = ivrs_agesa->ivhd.flags;
Piotr Król063e1562018-07-22 20:52:26 +0200358 ivrs->ivhd.length = sizeof(struct acpi_ivrs_ivhd);
359 /* BDF <bus>:00.2 */
Arthur Heymans7fcd4d52023-08-24 15:12:19 +0200360 ivrs->ivhd.device_id = 0x02 | (nb_dev->upstream->secondary << 8);
Michał Żygowski2f399b72020-04-02 19:51:37 +0200361 /* PCI Capability block 0x40 (type 0xf, "Secure device") */
Piotr Król063e1562018-07-22 20:52:26 +0200362 ivrs->ivhd.capability_offset = 0x40;
363 ivrs->ivhd.iommu_base_low = ivrs_agesa->ivhd.iommu_base_low;
364 ivrs->ivhd.iommu_base_high = ivrs_agesa->ivhd.iommu_base_high;
Arthur Heymans7fcd4d52023-08-24 15:12:19 +0200365 ivrs->ivhd.pci_segment_group = nb_dev->upstream->segment_group;
Michał Żygowski2f399b72020-04-02 19:51:37 +0200366 ivrs->ivhd.iommu_info = ivrs_agesa->ivhd.iommu_info;
367 ivrs->ivhd.iommu_feature_info = ivrs_agesa->ivhd.iommu_feature_info;
368 /* Enable EFR if supported */
369 if (pci_read_config32(iommu_dev, ivrs->ivhd.capability_offset) & EFR_SUPPORT)
370 ivrs->iv_info |= IVINFO_EFR_SUPPORTED;
Piotr Król063e1562018-07-22 20:52:26 +0200371 } else {
372 printk(BIOS_WARNING, "%s: AGESA returned NULL IVRS\n", __func__);
373
374 return (unsigned long)ivrs;
375 }
Timothy Pearson9ef07d82016-06-13 13:48:58 -0500376
Michał Żygowski2f399b72020-04-02 19:51:37 +0200377 /*
378 * Add all possible PCI devices on bus 0 that can generate transactions
379 * processed by IOMMU. Start with device 00:01.0 since IOMMU does not
380 * translate transactions generated by itself.
381 */
382 current_backup = current;
383 current = ivhd_dev_range(current, PCI_DEVFN(1, 0), PCI_DEVFN(0x1f, 6), 0);
384 ivrs->ivhd.length += (current - current_backup);
385 add_ivhd_device_entries(NULL, all_devices, 0, -1, NULL, &current, &ivrs->ivhd.length);
Timothy Pearson9ef07d82016-06-13 13:48:58 -0500386
Michał Żygowski2f399b72020-04-02 19:51:37 +0200387 /* Describe HPET */
388 current_backup = current;
389 current = ivhd_describe_hpet(current);
390 ivrs->ivhd.length += (current - current_backup);
Timothy Pearson9ef07d82016-06-13 13:48:58 -0500391
392 /* Describe IOAPICs */
Michał Żygowski2f399b72020-04-02 19:51:37 +0200393 current_backup = current;
394 current = acpi_fill_ivrs_ioapic(ivrs_agesa, current);
395 ivrs->ivhd.length += (current - current_backup);
Timothy Pearson9ef07d82016-06-13 13:48:58 -0500396
Michał Żygowski2f399b72020-04-02 19:51:37 +0200397 /* If EFR is not supported, IVHD type 11h is reserved */
398 if (!(ivrs->iv_info & IVINFO_EFR_SUPPORTED))
399 return current;
400
401 return acpi_fill_ivrs11(current, ivrs_agesa);
Timothy Pearson9ef07d82016-06-13 13:48:58 -0500402}
403
Furquan Shaikh7536a392020-04-24 21:59:21 -0700404static void northbridge_fill_ssdt_generator(const struct device *device)
Kyösti Mälkki0b5b5412014-11-26 08:11:07 +0200405{
Kyösti Mälkki0b5b5412014-11-26 08:11:07 +0200406 char pscope[] = "\\_SB.PCI0";
407
408 acpigen_write_scope(pscope);
Felix Helde3453782023-04-20 13:06:08 +0200409 acpigen_write_name_dword("TOM1", get_top_of_mem_below_4gb());
410
Kyösti Mälkki0b5b5412014-11-26 08:11:07 +0200411 /*
412 * Since XP only implements parts of ACPI 2.0, we can't use a qword
413 * here.
414 * See http://www.acpi.info/presentations/S01USMOBS169_OS%2520new.ppt
415 * slide 22ff.
416 * Shift value right by 20 bit to make it fit into 32bit,
417 * giving us 1MB granularity and a limit of almost 4Exabyte of memory.
418 */
Felix Held27af3e62023-04-22 05:59:52 +0200419 acpigen_write_name_dword("TOM2", get_top_of_mem_above_4gb() >> 20);
Kyösti Mälkki0b5b5412014-11-26 08:11:07 +0200420 acpigen_pop_len();
421}
422
Furquan Shaikh0f007d82020-04-24 06:41:18 -0700423static unsigned long agesa_write_acpi_tables(const struct device *device,
Alexander Couzens83fc32f2015-04-12 22:28:37 +0200424 unsigned long current,
Kyösti Mälkki0b5b5412014-11-26 08:11:07 +0200425 acpi_rsdp_t *rsdp)
426{
427 acpi_srat_t *srat;
428 acpi_slit_t *slit;
429 acpi_header_t *ssdt;
430 acpi_header_t *alib;
Timothy Pearson9ef07d82016-06-13 13:48:58 -0500431 acpi_ivrs_t *ivrs;
Kyösti Mälkki0b5b5412014-11-26 08:11:07 +0200432
433 /* HEST */
Elyes Haouasd6b6b222022-10-10 12:34:21 +0200434 current = ALIGN_UP(current, 8);
Vladimir Serbinenko807127f2014-11-09 13:36:18 +0100435 acpi_write_hest((void *)current, acpi_fill_hest);
Kyösti Mälkki0b5b5412014-11-26 08:11:07 +0200436 acpi_add_table(rsdp, (void *)current);
437 current += ((acpi_header_t *)current)->length;
438
Timothy Pearson9ef07d82016-06-13 13:48:58 -0500439 /* IVRS */
Elyes Haouasd6b6b222022-10-10 12:34:21 +0200440 current = ALIGN_UP(current, 8);
Timothy Pearson9ef07d82016-06-13 13:48:58 -0500441 printk(BIOS_DEBUG, "ACPI: * IVRS at %lx\n", current);
Elyes Haouasf9b535e2022-07-16 09:47:42 +0200442 ivrs = (acpi_ivrs_t *)current;
Timothy Pearson9ef07d82016-06-13 13:48:58 -0500443 acpi_create_ivrs(ivrs, acpi_fill_ivrs);
444 current += ivrs->header.length;
445 acpi_add_table(rsdp, ivrs);
Kyösti Mälkki0b5b5412014-11-26 08:11:07 +0200446
447 /* SRAT */
Elyes Haouasd6b6b222022-10-10 12:34:21 +0200448 current = ALIGN_UP(current, 8);
Kyösti Mälkki0b5b5412014-11-26 08:11:07 +0200449 printk(BIOS_DEBUG, "ACPI: * SRAT at %lx\n", current);
Elyes Haouasf9b535e2022-07-16 09:47:42 +0200450 srat = (acpi_srat_t *)agesawrapper_getlateinitptr(PICK_SRAT);
Kyösti Mälkki0b5b5412014-11-26 08:11:07 +0200451 if (srat != NULL) {
452 memcpy((void *)current, srat, srat->header.length);
Elyes Haouasf9b535e2022-07-16 09:47:42 +0200453 srat = (acpi_srat_t *)current;
Kyösti Mälkki0b5b5412014-11-26 08:11:07 +0200454 current += srat->header.length;
455 acpi_add_table(rsdp, srat);
456 } else {
457 printk(BIOS_DEBUG, " AGESA SRAT table NULL. Skipping.\n");
458 }
459
460 /* SLIT */
Elyes Haouasd6b6b222022-10-10 12:34:21 +0200461 current = ALIGN_UP(current, 8);
Kyösti Mälkki0b5b5412014-11-26 08:11:07 +0200462 printk(BIOS_DEBUG, "ACPI: * SLIT at %lx\n", current);
Elyes Haouasf9b535e2022-07-16 09:47:42 +0200463 slit = (acpi_slit_t *)agesawrapper_getlateinitptr(PICK_SLIT);
Kyösti Mälkki0b5b5412014-11-26 08:11:07 +0200464 if (slit != NULL) {
465 memcpy((void *)current, slit, slit->header.length);
Elyes Haouasf9b535e2022-07-16 09:47:42 +0200466 slit = (acpi_slit_t *)current;
Kyösti Mälkki0b5b5412014-11-26 08:11:07 +0200467 current += slit->header.length;
468 acpi_add_table(rsdp, slit);
469 } else {
470 printk(BIOS_DEBUG, " AGESA SLIT table NULL. Skipping.\n");
471 }
472
473 /* ALIB */
Elyes Haouasd6b6b222022-10-10 12:34:21 +0200474 current = ALIGN_UP(current, 16);
Kyösti Mälkki0b5b5412014-11-26 08:11:07 +0200475 printk(BIOS_DEBUG, "ACPI: * AGESA ALIB SSDT at %lx\n", current);
Elyes Haouasf9b535e2022-07-16 09:47:42 +0200476 alib = (acpi_header_t *)agesawrapper_getlateinitptr(PICK_ALIB);
Kyösti Mälkki0b5b5412014-11-26 08:11:07 +0200477 if (alib != NULL) {
478 memcpy((void *)current, alib, alib->length);
Elyes Haouasf9b535e2022-07-16 09:47:42 +0200479 alib = (acpi_header_t *)current;
Kyösti Mälkki0b5b5412014-11-26 08:11:07 +0200480 current += alib->length;
481 acpi_add_table(rsdp, (void *)alib);
482 }
483 else {
484 printk(BIOS_DEBUG, " AGESA ALIB SSDT table NULL. Skipping.\n");
485 }
486
487 /* this pstate ssdt may cause Blue Screen: Fixed: Keep this comment for a while. */
488 /* SSDT */
Elyes Haouasd6b6b222022-10-10 12:34:21 +0200489 current = ALIGN_UP(current, 16);
Kyösti Mälkki0b5b5412014-11-26 08:11:07 +0200490 printk(BIOS_DEBUG, "ACPI: * SSDT at %lx\n", current);
Elyes Haouasf9b535e2022-07-16 09:47:42 +0200491 ssdt = (acpi_header_t *)agesawrapper_getlateinitptr(PICK_PSTATE);
Kyösti Mälkki0b5b5412014-11-26 08:11:07 +0200492 if (ssdt != NULL) {
493 memcpy((void *)current, ssdt, ssdt->length);
Elyes Haouasf9b535e2022-07-16 09:47:42 +0200494 ssdt = (acpi_header_t *)current;
Kyösti Mälkki0b5b5412014-11-26 08:11:07 +0200495 current += ssdt->length;
496 }
497 else {
498 printk(BIOS_DEBUG, " AGESA PState table NULL. Skipping.\n");
499 }
Elyes Haouasf9b535e2022-07-16 09:47:42 +0200500 acpi_add_table(rsdp, ssdt);
Kyösti Mälkki0b5b5412014-11-26 08:11:07 +0200501
502 printk(BIOS_DEBUG, "ACPI: * SSDT for PState at %lx\n", current);
503 return current;
504}
505
Felix Held7b9c6472023-11-16 16:06:49 +0100506struct device_operations amd_pi_northbridge_ops = {
Michał Żygowskifb198c62021-05-09 13:54:09 +0200507 .read_resources = nb_read_resources,
Felix Heldb986e212023-12-16 00:58:09 +0100508 .set_resources = pci_dev_set_resources,
Bruce Griffith27ed80b2014-08-15 11:46:25 -0600509 .enable_resources = pci_dev_enable_resources,
510 .init = northbridge_init,
Michał Żygowskifb198c62021-05-09 13:54:09 +0200511 .ops_pci = &pci_dev_ops_pci,
Nico Huber68680dd2020-03-31 17:34:52 +0200512 .acpi_fill_ssdt = northbridge_fill_ssdt_generator,
Kyösti Mälkki0b5b5412014-11-26 08:11:07 +0200513 .write_acpi_tables = agesa_write_acpi_tables,
Bruce Griffith27ed80b2014-08-15 11:46:25 -0600514};
515
Dave Frodin891f71a2015-01-19 15:58:24 -0700516static void fam16_finalize(void *chip_info)
517{
Kyösti Mälkki90ac7362018-05-20 20:59:52 +0300518 struct device *dev;
Kyösti Mälkki33ff44c2018-05-22 01:15:22 +0300519 dev = pcidev_on_root(0, 0); /* clear IoapicSbFeatureEn */
Elyes Haouasa1f5ad02022-02-17 18:14:08 +0100520
Dave Frodin891f71a2015-01-19 15:58:24 -0700521 pci_write_config32(dev, 0xF8, 0);
522 pci_write_config32(dev, 0xFC, 5); /* TODO: move it to dsdt.asl */
523
Michał Żygowski6ca5b472019-09-10 15:10:22 +0200524 /*
525 * Currently it is impossible to enable ACS with AGESA by setting the
526 * correct bit for AmdInitMid phase. AGESA code path does not call the
527 * right function that enables these functionalities. Disabled ACS
528 * result in multiple PCIe devices to be assigned to the same IOMMU
529 * group. Without IOMMU group separation the devices cannot be passed
530 * through independently.
531 */
532
533 /* Select GPP link core IO Link Strap Control register 0xB0 */
534 pci_write_config32(dev, 0xE0, 0x014000B0);
Michał Żygowski6ca5b472019-09-10 15:10:22 +0200535
536 /* Enable AER (bit 5) and ACS (bit 6 undocumented) */
Elyes Haouasa1f5ad02022-02-17 18:14:08 +0100537 pci_or_config32(dev, 0xE4, PCIE_CAP_AER | PCIE_CAP_ACS);
Michał Żygowski6ca5b472019-09-10 15:10:22 +0200538
539 /* Select GPP link core Wrapper register 0x00 (undocumented) */
540 pci_write_config32(dev, 0xE0, 0x01300000);
Michał Żygowski6ca5b472019-09-10 15:10:22 +0200541
542 /*
543 * Enable ACS capabilities straps including sub-items. From lspci it
544 * looks like these bits enable: Source Validation and Translation
545 * Blocking
546 */
Elyes Haouasa1f5ad02022-02-17 18:14:08 +0100547 pci_or_config32(dev, 0xE4, (BIT(24) | BIT(25) | BIT(26)));
Michał Żygowski6ca5b472019-09-10 15:10:22 +0200548
Dave Frodin891f71a2015-01-19 15:58:24 -0700549 /* disable No Snoop */
Kyösti Mälkki33ff44c2018-05-22 01:15:22 +0300550 dev = pcidev_on_root(1, 1);
Kyösti Mälkki69f6fd42019-01-21 14:19:01 +0200551 if (dev != NULL) {
Elyes Haouasa1f5ad02022-02-17 18:14:08 +0100552 pci_and_config32(dev, 0x60, ~(1 << 11));
Kyösti Mälkki69f6fd42019-01-21 14:19:01 +0200553 }
Dave Frodin891f71a2015-01-19 15:58:24 -0700554}
555
Bruce Griffith27ed80b2014-08-15 11:46:25 -0600556struct hw_mem_hole_info {
Subrata Banikb1434fc2019-03-15 22:20:41 +0530557 unsigned int hole_startk;
Bruce Griffith27ed80b2014-08-15 11:46:25 -0600558 int node_id;
559};
560static struct hw_mem_hole_info get_hw_mem_hole_info(void)
561{
562 struct hw_mem_hole_info mem_hole;
Bruce Griffith27ed80b2014-08-15 11:46:25 -0600563 mem_hole.node_id = -1;
Felix Helda8807202023-11-16 21:29:33 +0100564
565 resource_t basek, limitk;
Felix Held3eaa8502023-12-16 01:37:34 +0100566 if (get_dram_base_limit(&basek, &limitk)) { // memory on this node
Felix Held7a83ab72023-12-16 23:10:50 +0100567 u32 hole = pci_read_config32(DEV_PTR(ht_1), 0xf0);
Bruce Griffith27ed80b2014-08-15 11:46:25 -0600568 if (hole & 2) { // we find the hole
Elyes Haouasf9b535e2022-07-16 09:47:42 +0200569 mem_hole.hole_startk = (hole & (0xff << 24)) >> 10;
Felix Helda8807202023-11-16 21:29:33 +0100570 mem_hole.node_id = 0; // record the node No with hole
Bruce Griffith27ed80b2014-08-15 11:46:25 -0600571 }
572 }
573 return mem_hole;
574}
Bruce Griffith27ed80b2014-08-15 11:46:25 -0600575
Michał Żygowskif5d457d2021-05-09 13:58:04 +0200576static void domain_read_resources(struct device *dev)
Bruce Griffith27ed80b2014-08-15 11:46:25 -0600577{
Bruce Griffith27ed80b2014-08-15 11:46:25 -0600578 unsigned long mmio_basek;
Felix Helddcbb1e82023-12-17 18:20:01 +0100579 unsigned long idx = 0;
Bruce Griffith27ed80b2014-08-15 11:46:25 -0600580 struct hw_mem_hole_info mem_hole;
Felix Held3f234f82023-12-17 18:38:08 +0100581 resource_t basek = 0;
582 resource_t limitk = 0;
583 resource_t sizek;
Bruce Griffith27ed80b2014-08-15 11:46:25 -0600584
Michał Żygowskif5d457d2021-05-09 13:58:04 +0200585 pci_domain_read_resources(dev);
586
Michał Żygowski58d6f962021-05-05 10:52:08 +0200587 /* TOP_MEM MSR is our boundary between DRAM and MMIO under 4G */
Felix Held5e9afe72023-04-20 12:55:55 +0200588 mmio_basek = get_top_of_mem_below_4gb() >> 10;
Bruce Griffith27ed80b2014-08-15 11:46:25 -0600589
Bruce Griffith27ed80b2014-08-15 11:46:25 -0600590 /* if the hw mem hole is already set in raminit stage, here we will compare
591 * mmio_basek and hole_basek. if mmio_basek is bigger that hole_basek and will
592 * use hole_basek as mmio_basek and we don't need to reset hole.
593 * otherwise We reset the hole to the mmio_basek
594 */
595
596 mem_hole = get_hw_mem_hole_info();
597
598 // Use hole_basek as mmio_basek, and we don't need to reset hole anymore
599 if ((mem_hole.node_id != -1) && (mmio_basek > mem_hole.hole_startk)) {
600 mmio_basek = mem_hole.hole_startk;
Bruce Griffith27ed80b2014-08-15 11:46:25 -0600601 }
Bruce Griffith27ed80b2014-08-15 11:46:25 -0600602
Felix Held3f234f82023-12-17 18:38:08 +0100603 get_dram_base_limit(&basek, &limitk);
604 sizek = limitk - basek;
605
606 printk(BIOS_DEBUG, "basek=%08llx, limitk=%08llx, sizek=%08llx,\n",
607 basek, limitk, sizek);
608
609 /* See if we need a hole from 0xa0000 (640K) to 0xfffff (1024K) */
610 if (basek < 640 && sizek > 1024) {
611 ram_resource_kb(dev, idx++, basek, 640 - basek);
612 basek = 1024;
Bruce Griffith27ed80b2014-08-15 11:46:25 -0600613 sizek = limitk - basek;
Bruce Griffith27ed80b2014-08-15 11:46:25 -0600614 }
615
Felix Held3f234f82023-12-17 18:38:08 +0100616 printk(BIOS_DEBUG, "basek=%08llx, limitk=%08llx, sizek=%08llx,\n",
617 basek, limitk, sizek);
618
619 /* split the region to accommodate pci memory space */
620 if ((basek < 4 * 1024 * 1024) && (limitk > mmio_basek)) {
621 if (basek <= mmio_basek) {
622 unsigned int pre_sizek;
623 pre_sizek = mmio_basek - basek;
624 if (pre_sizek > 0) {
625 ram_resource_kb(dev, idx++, basek, pre_sizek);
626 sizek -= pre_sizek;
627 }
628 basek = mmio_basek;
629 }
630 if ((basek + sizek) <= 4 * 1024 * 1024) {
631 sizek = 0;
632 } else {
633 uint64_t topmem2 = get_top_of_mem_above_4gb();
634 basek = 4 * 1024 * 1024;
635 sizek = topmem2 / 1024 - basek;
636 }
637 }
638
639 ram_resource_kb(dev, idx++, basek, sizek);
640 printk(BIOS_DEBUG, "mmio_basek=%08lx, basek=%08llx, limitk=%08llx\n",
641 mmio_basek, basek, limitk);
642
Felix Helddcbb1e82023-12-17 18:20:01 +0100643 add_uma_resource_below_tolm(dev, idx++);
Bruce Griffith27ed80b2014-08-15 11:46:25 -0600644}
645
Aaron Durbinaa090cb2017-09-13 16:01:52 -0600646static const char *domain_acpi_name(const struct device *dev)
Philipp Deppenwiese30670122017-03-01 02:24:33 +0100647{
648 if (dev->path.type == DEVICE_PATH_DOMAIN)
649 return "PCI0";
650
651 return NULL;
652}
653
Felix Held8ccd3142023-11-16 00:58:30 +0100654struct device_operations amd_fam16_mod30_pci_domain_ops = {
Bruce Griffith27ed80b2014-08-15 11:46:25 -0600655 .read_resources = domain_read_resources,
Michał Żygowskif5d457d2021-05-09 13:58:04 +0200656 .set_resources = pci_domain_set_resources,
Arthur Heymans0b0113f2023-08-31 17:09:28 +0200657 .scan_bus = pci_host_bridge_scan_bus,
Philipp Deppenwiese30670122017-03-01 02:24:33 +0100658 .acpi_name = domain_acpi_name,
Bruce Griffith27ed80b2014-08-15 11:46:25 -0600659};
660
Michał Kopećdc35d2a2021-11-30 17:40:52 +0100661void mp_init_cpus(struct bus *cpu_bus)
662{
Arthur Heymans4fcaccf2022-06-02 13:17:37 +0200663 extern const struct mp_ops amd_mp_ops_no_smm;
Michał Kopećdc35d2a2021-11-30 17:40:52 +0100664 /* TODO: Handle mp_init_with_smm failure? */
Arthur Heymans4fcaccf2022-06-02 13:17:37 +0200665 mp_init_with_smm(cpu_bus, &amd_mp_ops_no_smm);
Michał Kopećdc35d2a2021-11-30 17:40:52 +0100666
667 /* The flash is now no longer cacheable. Reset to WP for performance. */
668 mtrr_use_temp_range(OPTIMAL_CACHE_ROM_BASE, OPTIMAL_CACHE_ROM_SIZE,
669 MTRR_TYPE_WRPROT);
Bruce Griffith27ed80b2014-08-15 11:46:25 -0600670}
671
Felix Heldc391bff2023-02-16 19:38:49 +0100672void generate_cpu_entries(const struct device *device)
673{
674 int cpu;
675 const int cores = get_cpu_count();
676
677 printk(BIOS_DEBUG, "ACPI \\_SB report %d core(s)\n", cores);
678
679 /* Generate \_SB.Pxxx */
680 for (cpu = 0; cpu < cores; cpu++) {
681 acpigen_write_processor_device(cpu);
682 acpigen_write_processor_device_end();
683 }
684}
685
Felix Held8ccd3142023-11-16 00:58:30 +0100686struct device_operations amd_fam16_mod30_cpu_bus_ops = {
Felix Heldc391bff2023-02-16 19:38:49 +0100687 .read_resources = noop_read_resources,
688 .set_resources = noop_set_resources,
689 .init = mp_cpu_bus_init,
690 .acpi_fill_ssdt = generate_cpu_entries,
Bruce Griffith27ed80b2014-08-15 11:46:25 -0600691};
692
Felix Held1952d132023-11-16 00:54:30 +0100693struct chip_operations northbridge_amd_pi_00730F01_ops = {
Nicholas Sudsgaardbfb11be2024-01-30 09:53:46 +0900694 .name = "AMD FAM16 Root Complex",
Felix Held1952d132023-11-16 00:54:30 +0100695 .final = fam16_finalize,
Bruce Griffith27ed80b2014-08-15 11:46:25 -0600696};
697
698/*********************************************************************
699 * Change the vendor / device IDs to match the generic VBIOS header. *
700 *********************************************************************/
701u32 map_oprom_vendev(u32 vendev)
702{
703 u32 new_vendev;
704 new_vendev =
705 ((0x10029850 <= vendev) && (vendev <= 0x1002986F)) ? 0x10029850 : vendev;
706
707 if (vendev != new_vendev)
708 printk(BIOS_NOTICE, "Mapping PCI device %8x to %8x\n", vendev, new_vendev);
709
710 return new_vendev;
711}