| /* SPDX-License-Identifier: GPL-2.0-only */ |
| |
| #include <arch/ioapic.h> |
| #include <assert.h> |
| #include <console/console.h> |
| #include <cpu/x86/lapic.h> |
| #include <device/pci.h> |
| #include <fsp/api.h> |
| #include <intelblocks/p2sb.h> |
| #include <post.h> |
| #include <soc/cpu.h> |
| #include <soc/ramstage.h> |
| #include <soc/pm.h> |
| #include <soc/soc_util.h> |
| #include <stdlib.h> |
| |
| /* C620 IOAPIC has 120 redirection entries */ |
| #define C620_IOAPIC_REDIR_ENTRIES 120 |
| |
| struct pci_resource { |
| struct device *dev; |
| struct resource *res; |
| struct pci_resource *next; |
| }; |
| |
| struct stack_dev_resource { |
| uint8_t align; |
| struct pci_resource *children; |
| struct stack_dev_resource *next; |
| }; |
| |
| typedef enum { |
| RES_TYPE_IO = 0, |
| RES_TYPE_NONPREF_MEM, |
| RES_TYPE_PREF_MEM, |
| MAX_RES_TYPES |
| } ResType; |
| |
| static ResType get_res_type(uint64_t flags) |
| { |
| if (flags & IORESOURCE_IO) |
| return RES_TYPE_IO; |
| if (flags & IORESOURCE_MEM) { |
| if (flags & IORESOURCE_PREFETCH) { |
| printk(BIOS_DEBUG, "%s:%d flags: 0x%llx\n", __func__, __LINE__, flags); |
| return RES_TYPE_PREF_MEM; |
| } |
| /* both 64-bit and 32-bit use below 4GB address space */ |
| return RES_TYPE_NONPREF_MEM; |
| } |
| printk(BIOS_ERR, "Invalid resource type 0x%llx\n", flags); |
| die("Invalida resource type"); |
| } |
| |
| static bool need_assignment(uint64_t flags) |
| { |
| if (flags & (IORESOURCE_STORED | IORESOURCE_RESERVE | IORESOURCE_FIXED | |
| IORESOURCE_ASSIGNED)) |
| return false; |
| else |
| return true; |
| } |
| |
| static uint64_t get_resource_base(STACK_RES *stack, ResType res_type) |
| { |
| if (res_type == RES_TYPE_IO) { |
| assert(stack->PciResourceIoBase <= stack->PciResourceIoLimit); |
| return stack->PciResourceIoBase; |
| } |
| if (res_type == RES_TYPE_NONPREF_MEM) { |
| assert(stack->PciResourceMem32Base <= stack->PciResourceMem32Limit); |
| return stack->PciResourceMem32Base; |
| } |
| assert(stack->PciResourceMem64Base <= stack->PciResourceMem64Limit); |
| return stack->PciResourceMem64Base; |
| } |
| |
| static void set_resource_base(STACK_RES *stack, ResType res_type, uint64_t base) |
| { |
| if (res_type == RES_TYPE_IO) { |
| assert(base <= (stack->PciResourceIoLimit + 1)); |
| stack->PciResourceIoBase = base; |
| } else if (res_type == RES_TYPE_NONPREF_MEM) { |
| assert(base <= (stack->PciResourceMem32Limit + 1)); |
| stack->PciResourceMem32Base = base; |
| } else { |
| assert(base <= (stack->PciResourceMem64Limit + 1)); |
| stack->PciResourceMem64Base = base; |
| } |
| } |
| |
| static void assign_stack_resources(struct iiostack_resource *stack_list, |
| struct device *dev, struct resource *bridge); |
| |
| static void xeonsp_cpx_pci_domain_scan_bus(struct device *dev) |
| { |
| DEV_FUNC_ENTER(dev); |
| struct bus *link = dev->link_list; |
| |
| printk(BIOS_SPEW, "%s:%s scanning buses under device %s\n", |
| __FILE__, __func__, dev_path(dev)); |
| while (link != NULL) { |
| if (link->secondary == 0) { // scan only PSTACK buses |
| struct device *d; |
| for (d = link->children; d; d = d->sibling) |
| pci_probe_dev(d, link, d->path.pci.devfn); |
| scan_bridges(link); |
| } else { |
| pci_scan_bus(link, PCI_DEVFN(0, 0), 0xff); |
| } |
| link = link->next; |
| } |
| DEV_FUNC_EXIT(dev); |
| } |
| |
| static void xeonsp_pci_dev_iterator(struct bus *bus, |
| void (*dev_iterator)(struct device *, void *), |
| void (*res_iterator)(struct device *, struct resource *, void *), |
| void *data) |
| { |
| struct device *curdev; |
| struct resource *res; |
| |
| /* Walk through all devices and find which resources they need. */ |
| for (curdev = bus->children; curdev; curdev = curdev->sibling) { |
| struct bus *link; |
| |
| if (!curdev->enabled) |
| continue; |
| |
| if (!curdev->ops || !curdev->ops->read_resources) { |
| if (curdev->path.type != DEVICE_PATH_APIC) |
| printk(BIOS_ERR, "%s missing read_resources\n", |
| dev_path(curdev)); |
| continue; |
| } |
| |
| if (dev_iterator) |
| dev_iterator(curdev, data); |
| |
| if (res_iterator) { |
| for (res = curdev->resource_list; res; res = res->next) |
| res_iterator(curdev, res, data); |
| } |
| |
| /* Read in the resources behind the current device's links. */ |
| for (link = curdev->link_list; link; link = link->next) |
| xeonsp_pci_dev_iterator(link, dev_iterator, res_iterator, data); |
| } |
| } |
| |
| static void xeonsp_pci_dev_read_resources(struct device *dev, void *data) |
| { |
| post_log_path(dev); |
| dev->ops->read_resources(dev); |
| } |
| |
| static void xeonsp_pci_dev_dummy_func(struct device *dev) |
| { |
| } |
| |
| static void xeonsp_reset_pci_op(struct device *dev, void *data) |
| { |
| if (dev->ops) |
| dev->ops->read_resources = xeonsp_pci_dev_dummy_func; |
| } |
| |
| static STACK_RES *find_stack_for_bus(struct iiostack_resource *info, uint8_t bus) |
| { |
| for (int i = 0; i < info->no_of_stacks; ++i) { |
| if (bus >= info->res[i].BusBase && bus <= info->res[i].BusLimit) |
| return &info->res[i]; |
| } |
| return NULL; |
| } |
| |
| static void add_res_to_stack(struct stack_dev_resource **root, |
| struct device *dev, struct resource *res) |
| { |
| struct stack_dev_resource *cur = *root; |
| while (cur) { |
| if (cur->align == res->align || cur->next == NULL) /* equal or last record */ |
| break; |
| else if (cur->align > res->align) { |
| if (cur->next->align < res->align) /* need to insert new record here */ |
| break; |
| cur = cur->next; |
| } else { |
| break; |
| } |
| } |
| |
| struct stack_dev_resource *nr; |
| if (!cur || cur->align != res->align) { /* need to add new record */ |
| nr = malloc(sizeof(struct stack_dev_resource)); |
| if (nr == 0) |
| die("assign_resource_to_stack(): out of memory.\n"); |
| memset(nr, 0, sizeof(struct stack_dev_resource)); |
| nr->align = res->align; |
| if (!cur) { |
| *root = nr; /* head node */ |
| } else if (cur->align > nr->align) { |
| if (cur->next == NULL) { |
| cur->next = nr; |
| } else { |
| nr->next = cur->next; |
| cur->next = nr; |
| } |
| } else { /* insert in the beginning */ |
| nr->next = cur; |
| *root = nr; |
| } |
| } else { |
| nr = cur; |
| } |
| |
| assert(nr != NULL && nr->align == res->align); |
| |
| struct pci_resource *npr = malloc(sizeof(struct pci_resource)); |
| if (npr == NULL) |
| die("%s: out of memory.\n", __func__); |
| npr->res = res; |
| npr->dev = dev; |
| npr->next = NULL; |
| |
| if (nr->children == NULL) { |
| nr->children = npr; |
| } else { |
| struct pci_resource *pr = nr->children; |
| while (pr->next != NULL) |
| pr = pr->next; |
| pr->next = npr; |
| } |
| } |
| |
| static void reserve_dev_resources(STACK_RES *stack, ResType res_type, |
| struct stack_dev_resource *res_root, struct resource *bridge) |
| { |
| uint8_t align; |
| uint64_t orig_base, base; |
| |
| orig_base = get_resource_base(stack, res_type); |
| |
| align = 0; |
| base = orig_base; |
| int first = 1; |
| while (res_root) { /* loop through all devices grouped by alignment requirements */ |
| struct pci_resource *pr = res_root->children; |
| while (pr) { |
| if (first) { |
| if (bridge) { /* takes highest alignment */ |
| if (bridge->align < pr->res->align) |
| bridge->align = pr->res->align; |
| orig_base = ALIGN_UP(orig_base, 1 << bridge->align); |
| } else { |
| orig_base = ALIGN_UP(orig_base, 1 << pr->res->align); |
| } |
| base = orig_base; |
| |
| if (bridge) |
| bridge->base = base; |
| pr->res->base = base; |
| first = 0; |
| } else { |
| pr->res->base = ALIGN_UP(base, 1 << pr->res->align); |
| } |
| pr->res->limit = pr->res->base + pr->res->size - 1; |
| base = pr->res->limit + 1; |
| pr->res->flags |= (IORESOURCE_ASSIGNED); |
| pr = pr->next; |
| } |
| res_root = res_root->next; |
| } |
| |
| if (bridge) { |
| /* this bridge doesn't have any resources, will set it to default window */ |
| if (first) { |
| orig_base = ALIGN_UP(orig_base, 1 << bridge->align); |
| bridge->base = orig_base; |
| base = orig_base + (1ULL << bridge->gran); |
| } |
| |
| bridge->size = ALIGN_UP(base, 1 << bridge->align) - bridge->base; |
| |
| bridge->limit = bridge->base + bridge->size - 1; |
| bridge->flags |= (IORESOURCE_ASSIGNED); |
| base = bridge->limit + 1; |
| } |
| |
| set_resource_base(stack, res_type, base); |
| } |
| |
| static void reclaim_resource_mem(struct stack_dev_resource *res_root) |
| { |
| while (res_root) { /* loop through all devices grouped by alignment requirements */ |
| /* free pci_resource */ |
| struct pci_resource *pr = res_root->children; |
| while (pr) { |
| struct pci_resource *dpr = pr; |
| pr = pr->next; |
| free(dpr); |
| } |
| |
| /* free stack_dev_resource */ |
| struct stack_dev_resource *ddr = res_root; |
| res_root = res_root->next; |
| free(ddr); |
| } |
| } |
| |
| static void assign_bridge_resources(struct iiostack_resource *stack_list, |
| struct device *dev, struct resource *bridge) |
| { |
| struct resource *res; |
| if (!dev->enabled) |
| return; |
| |
| for (res = dev->resource_list; res; res = res->next) { |
| if (!(res->flags & IORESOURCE_BRIDGE) || |
| (bridge && (get_res_type(bridge->flags) != get_res_type(res->flags)))) |
| continue; |
| |
| assign_stack_resources(stack_list, dev, res); |
| |
| if (!bridge) |
| continue; |
| |
| /* for 1st time update, overlading IORESOURCE_ASSIGNED */ |
| if (!(bridge->flags & IORESOURCE_ASSIGNED)) { |
| bridge->base = res->base; |
| bridge->limit = res->limit; |
| bridge->flags |= (IORESOURCE_ASSIGNED); |
| } else { |
| /* update bridge range from child bridge range */ |
| if (res->base < bridge->base) |
| bridge->base = res->base; |
| if (res->limit > bridge->limit) |
| bridge->limit = res->limit; |
| } |
| bridge->size = (bridge->limit - bridge->base + 1); |
| } |
| } |
| |
| static void assign_stack_resources(struct iiostack_resource *stack_list, |
| struct device *dev, struct resource *bridge) |
| { |
| struct bus *bus; |
| |
| /* Read in the resources behind the current device's links. */ |
| for (bus = dev->link_list; bus; bus = bus->next) { |
| struct device *curdev; |
| STACK_RES *stack; |
| |
| /* get IIO stack for this bus */ |
| stack = find_stack_for_bus(stack_list, bus->secondary); |
| assert(stack != NULL); |
| |
| /* Assign resources to bridge */ |
| for (curdev = bus->children; curdev; curdev = curdev->sibling) |
| assign_bridge_resources(stack_list, curdev, bridge); |
| |
| /* Pick non-bridged resources for resource allocation for each resource type */ |
| ResType res_types[MAX_RES_TYPES] = { |
| RES_TYPE_IO, |
| RES_TYPE_NONPREF_MEM, |
| RES_TYPE_PREF_MEM |
| }; |
| |
| uint8_t no_res_types = MAX_RES_TYPES; |
| |
| /* if it is a bridge, only process matching brigge resource type */ |
| if (bridge) { |
| res_types[0] = get_res_type(bridge->flags); |
| no_res_types = 1; |
| } |
| |
| printk(BIOS_DEBUG, "%s:%d no_res_types: %d\n", __func__, __LINE__, |
| no_res_types); |
| |
| /* Process each resource type */ |
| for (int rt = 0; rt < no_res_types; ++rt) { |
| struct stack_dev_resource *res_root = NULL; |
| printk(BIOS_DEBUG, "%s:%d rt: %d\n", __func__, __LINE__, rt); |
| for (curdev = bus->children; curdev; curdev = curdev->sibling) { |
| struct resource *res; |
| printk(BIOS_DEBUG, "%s:%d dev: %s\n", |
| __func__, __LINE__, dev_path(curdev)); |
| if (!curdev->enabled) |
| continue; |
| |
| for (res = curdev->resource_list; res; res = res->next) { |
| printk(BIOS_DEBUG, "%s:%d dev: %s, flags: 0x%lx\n", |
| __func__, __LINE__, |
| dev_path(curdev), res->flags); |
| if (res->size == 0 || |
| get_res_type(res->flags) != res_types[rt] || |
| (res->flags & IORESOURCE_BRIDGE) || |
| !need_assignment(res->flags)) |
| continue; |
| else |
| add_res_to_stack(&res_root, curdev, res); |
| } |
| } |
| |
| /* Allocate resources and update bridge range */ |
| if (res_root || (bridge && !(bridge->flags & IORESOURCE_ASSIGNED))) { |
| reserve_dev_resources(stack, res_types[rt], res_root, bridge); |
| reclaim_resource_mem(res_root); |
| } |
| } |
| } |
| } |
| |
| static void xeonsp_pci_domain_read_resources(struct device *dev) |
| { |
| struct bus *link; |
| |
| DEV_FUNC_ENTER(dev); |
| |
| pci_domain_read_resources(dev); |
| |
| /* |
| * Walk through all devices in this domain and read resources. |
| * Since there is no callback when read resource operation is |
| * complete for all devices, domain read resource function initiates |
| * read resources for all devices and swaps read resource operation |
| * with dummy function to avoid warning. |
| */ |
| for (link = dev->link_list; link; link = link->next) |
| xeonsp_pci_dev_iterator(link, xeonsp_pci_dev_read_resources, NULL, NULL); |
| |
| for (link = dev->link_list; link; link = link->next) |
| xeonsp_pci_dev_iterator(link, xeonsp_reset_pci_op, NULL, NULL); |
| |
| struct iiostack_resource stack_info = {0}; |
| uint8_t pci64bit_alloc_flag = get_iiostack_info(&stack_info); |
| if (!pci64bit_alloc_flag) { |
| /* |
| * Split 32 bit address space between prefetchable and |
| * non-prefetchable windows |
| */ |
| for (int s = 0; s < stack_info.no_of_stacks; ++s) { |
| STACK_RES *res = &stack_info.res[s]; |
| uint64_t length = (res->PciResourceMem32Limit - |
| res->PciResourceMem32Base + 1)/2; |
| res->PciResourceMem64Limit = res->PciResourceMem32Limit; |
| res->PciResourceMem32Limit = (res->PciResourceMem32Base + length - 1); |
| res->PciResourceMem64Base = res->PciResourceMem32Limit + 1; |
| } |
| } |
| |
| |
| /* assign resources */ |
| assign_stack_resources(&stack_info, dev, NULL); |
| |
| DEV_FUNC_EXIT(dev); |
| } |
| |
| static void reset_resource_to_unassigned(struct device *dev, struct resource *res, void *data) |
| { |
| if ((res->flags & (IORESOURCE_IO | IORESOURCE_MEM)) && |
| !(res->flags & (IORESOURCE_FIXED | IORESOURCE_RESERVE))) { |
| res->flags &= ~IORESOURCE_ASSIGNED; |
| } |
| } |
| |
| static void xeonsp_cpx_pci_domain_set_resources(struct device *dev) |
| { |
| DEV_FUNC_ENTER(dev); |
| |
| print_resource_tree(dev, BIOS_SPEW, "Before xeonsp pci domain set resource"); |
| |
| /* reset bus 0 dev resource assignment - need to change them to FSP IIOStack window */ |
| xeonsp_pci_dev_iterator(dev->link_list, NULL, reset_resource_to_unassigned, NULL); |
| |
| /* update dev resources based on IIOStack IO/Mem32/Mem64 windows */ |
| xeonsp_pci_domain_read_resources(dev); |
| |
| struct bus *link = dev->link_list; |
| while (link != NULL) { |
| assign_resources(link); |
| link = link->next; |
| } |
| |
| print_resource_tree(dev, BIOS_SPEW, "After xeonsp pci domain set resource"); |
| |
| DEV_FUNC_EXIT(dev); |
| } |
| |
| void platform_fsp_silicon_init_params_cb(FSPS_UPD *silupd) |
| { |
| /* not implemented yet */ |
| } |
| |
| static struct device_operations pci_domain_ops = { |
| .read_resources = &pci_domain_read_resources, |
| .set_resources = &xeonsp_cpx_pci_domain_set_resources, |
| .scan_bus = &xeonsp_cpx_pci_domain_scan_bus, |
| }; |
| |
| static struct device_operations cpu_bus_ops = { |
| .read_resources = noop_read_resources, |
| .set_resources = noop_set_resources, |
| .init = cpx_init_cpus, |
| }; |
| |
| /* Attach IIO stack bus numbers with dummy device to PCI DOMAIN 0000 device */ |
| static void attach_iio_stacks(struct device *dev) |
| { |
| struct bus *iiostack_bus; |
| struct iiostack_resource stack_info = {0}; |
| |
| DEV_FUNC_ENTER(dev); |
| |
| get_iiostack_info(&stack_info); |
| for (int s = 0; s < stack_info.no_of_stacks; ++s) { |
| /* only non zero bus no. needs to be enumerated */ |
| if (stack_info.res[s].BusBase == 0) |
| continue; |
| |
| iiostack_bus = malloc(sizeof(struct bus)); |
| if (iiostack_bus == NULL) |
| die("%s: out of memory.\n", __func__); |
| memset(iiostack_bus, 0, sizeof(*iiostack_bus)); |
| memcpy(iiostack_bus, dev->bus, sizeof(*iiostack_bus)); |
| iiostack_bus->secondary = stack_info.res[s].BusBase; |
| iiostack_bus->subordinate = stack_info.res[s].BusBase; |
| iiostack_bus->dev = NULL; |
| iiostack_bus->children = NULL; |
| iiostack_bus->next = NULL; |
| iiostack_bus->link_num = 1; |
| |
| if (dev->link_list == NULL) { |
| dev->link_list = iiostack_bus; |
| } else { |
| struct bus *nlink = dev->link_list; |
| while (nlink->next != NULL) |
| nlink = nlink->next; |
| nlink->next = iiostack_bus; |
| } |
| } |
| |
| DEV_FUNC_EXIT(dev); |
| } |
| |
| static void pch_enable_ioapic(const struct device *dev) |
| { |
| uint32_t reg32; |
| |
| set_ioapic_id((void *)IO_APIC_ADDR, 2); |
| |
| /* affirm full set of redirection table entries ("write once") */ |
| reg32 = io_apic_read((void *)IO_APIC_ADDR, 1); |
| |
| reg32 &= ~0x00ff0000; |
| reg32 |= (C620_IOAPIC_REDIR_ENTRIES - 1) << 16; |
| |
| io_apic_write((void *)IO_APIC_ADDR, 1, reg32); |
| |
| /* |
| * Select Boot Configuration register (0x03) and |
| * use Processor System Bus (0x01) to deliver interrupts. |
| */ |
| io_apic_write((void *)IO_APIC_ADDR, 3, 1); |
| } |
| |
| struct pci_operations soc_pci_ops = { |
| .set_subsystem = pci_dev_set_subsystem, |
| }; |
| |
| static void chip_enable_dev(struct device *dev) |
| { |
| /* Set the operations if it is a special bus type */ |
| if (dev->path.type == DEVICE_PATH_DOMAIN) { |
| dev->ops = &pci_domain_ops; |
| attach_iio_stacks(dev); |
| } else if (dev->path.type == DEVICE_PATH_CPU_CLUSTER) { |
| dev->ops = &cpu_bus_ops; |
| } |
| } |
| |
| static void chip_final(void *data) |
| { |
| p2sb_hide(); |
| } |
| |
| static void chip_init(void *data) |
| { |
| printk(BIOS_DEBUG, "coreboot: calling fsp_silicon_init\n"); |
| fsp_silicon_init(false); |
| pch_enable_ioapic(NULL); |
| setup_lapic(); |
| p2sb_unhide(); |
| } |
| |
| struct chip_operations soc_intel_xeon_sp_cpx_ops = { |
| CHIP_NAME("Intel Cooperlake-SP") |
| .enable_dev = chip_enable_dev, |
| .init = chip_init, |
| .final = chip_final, |
| }; |