| /* |
| * This file is part of the coreboot project. |
| * |
| * |
| * This program is free software; you can redistribute it and/or modify |
| * it under the terms of the GNU General Public License as published by |
| * the Free Software Foundation; version 2 of the License. |
| * |
| * This program is distributed in the hope that it will be useful, |
| * but WITHOUT ANY WARRANTY; without even the implied warranty of |
| * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
| * GNU General Public License for more details. |
| */ |
| |
| #include <assert.h> |
| #include <commonlib/helpers.h> |
| #include <console/console.h> |
| #include <device/mmio.h> |
| #include <delay.h> |
| #include <device/pci.h> |
| #include <device/pci_ids.h> |
| #include <device/pci_ops.h> |
| #include <intelblocks/cse.h> |
| #include <soc/iomap.h> |
| #include <soc/pci_devs.h> |
| #include <soc/me.h> |
| #include <string.h> |
| #include <timer.h> |
| |
| #define MAX_HECI_MESSAGE_RETRY_COUNT 5 |
| |
| /* Wait up to 15 sec for HECI to get ready */ |
| #define HECI_DELAY_READY (15 * 1000) |
| /* Wait up to 100 usec between circular buffer polls */ |
| #define HECI_DELAY 100 |
| /* Wait up to 5 sec for CSE to chew something we sent */ |
| #define HECI_SEND_TIMEOUT (5 * 1000) |
| /* Wait up to 5 sec for CSE to blurp a reply */ |
| #define HECI_READ_TIMEOUT (5 * 1000) |
| |
| #define SLOT_SIZE sizeof(uint32_t) |
| |
| #define MMIO_CSE_CB_WW 0x00 |
| #define MMIO_HOST_CSR 0x04 |
| #define MMIO_CSE_CB_RW 0x08 |
| #define MMIO_CSE_CSR 0x0c |
| |
| #define CSR_IE (1 << 0) |
| #define CSR_IS (1 << 1) |
| #define CSR_IG (1 << 2) |
| #define CSR_READY (1 << 3) |
| #define CSR_RESET (1 << 4) |
| #define CSR_RP_START 8 |
| #define CSR_RP (((1 << 8) - 1) << CSR_RP_START) |
| #define CSR_WP_START 16 |
| #define CSR_WP (((1 << 8) - 1) << CSR_WP_START) |
| #define CSR_CBD_START 24 |
| #define CSR_CBD (((1 << 8) - 1) << CSR_CBD_START) |
| |
| #define MEI_HDR_IS_COMPLETE (1 << 31) |
| #define MEI_HDR_LENGTH_START 16 |
| #define MEI_HDR_LENGTH_SIZE 9 |
| #define MEI_HDR_LENGTH (((1 << MEI_HDR_LENGTH_SIZE) - 1) \ |
| << MEI_HDR_LENGTH_START) |
| #define MEI_HDR_HOST_ADDR_START 8 |
| #define MEI_HDR_HOST_ADDR (((1 << 8) - 1) << MEI_HDR_HOST_ADDR_START) |
| #define MEI_HDR_CSE_ADDR_START 0 |
| #define MEI_HDR_CSE_ADDR (((1 << 8) - 1) << MEI_HDR_CSE_ADDR_START) |
| |
| /* Wait up to 5 seconds for CSE to boot from RO(BP1) */ |
| #define CSE_DELAY_BOOT_TO_RO (5 * 1000) |
| |
| static struct cse_device { |
| uintptr_t sec_bar; |
| } cse; |
| |
| /* |
| * Initialize the device with provided temporary BAR. If BAR is 0 use a |
| * default. This is intended for pre-mem usage only where BARs haven't been |
| * assigned yet and devices are not enabled. |
| */ |
| void heci_init(uintptr_t tempbar) |
| { |
| #if defined(__SIMPLE_DEVICE__) |
| pci_devfn_t dev = PCH_DEV_CSE; |
| #else |
| struct device *dev = PCH_DEV_CSE; |
| #endif |
| u8 pcireg; |
| |
| /* Assume it is already initialized, nothing else to do */ |
| if (cse.sec_bar) |
| return; |
| |
| /* Use default pre-ram bar */ |
| if (!tempbar) |
| tempbar = HECI1_BASE_ADDRESS; |
| |
| /* Assign Resources to HECI1 */ |
| /* Clear BIT 1-2 of Command Register */ |
| pcireg = pci_read_config8(dev, PCI_COMMAND); |
| pcireg &= ~(PCI_COMMAND_MASTER | PCI_COMMAND_MEMORY); |
| pci_write_config8(dev, PCI_COMMAND, pcireg); |
| |
| /* Program Temporary BAR for HECI1 */ |
| pci_write_config32(dev, PCI_BASE_ADDRESS_0, tempbar); |
| pci_write_config32(dev, PCI_BASE_ADDRESS_1, 0x0); |
| |
| /* Enable Bus Master and MMIO Space */ |
| pcireg = pci_read_config8(dev, PCI_COMMAND); |
| pcireg |= PCI_COMMAND_MASTER | PCI_COMMAND_MEMORY; |
| pci_write_config8(dev, PCI_COMMAND, pcireg); |
| |
| cse.sec_bar = tempbar; |
| } |
| |
| /* Get HECI BAR 0 from PCI configuration space */ |
| static uint32_t get_cse_bar(void) |
| { |
| uintptr_t bar; |
| |
| bar = pci_read_config32(PCH_DEV_CSE, PCI_BASE_ADDRESS_0); |
| assert(bar != 0); |
| /* |
| * Bits 31-12 are the base address as per EDS for SPI, |
| * Don't care about 0-11 bit |
| */ |
| return bar & ~PCI_BASE_ADDRESS_MEM_ATTR_MASK; |
| } |
| |
| static uint32_t read_bar(uint32_t offset) |
| { |
| /* Load and cache BAR */ |
| if (!cse.sec_bar) |
| cse.sec_bar = get_cse_bar(); |
| return read32((void *)(cse.sec_bar + offset)); |
| } |
| |
| static void write_bar(uint32_t offset, uint32_t val) |
| { |
| /* Load and cache BAR */ |
| if (!cse.sec_bar) |
| cse.sec_bar = get_cse_bar(); |
| return write32((void *)(cse.sec_bar + offset), val); |
| } |
| |
| static uint32_t read_cse_csr(void) |
| { |
| return read_bar(MMIO_CSE_CSR); |
| } |
| |
| static uint32_t read_host_csr(void) |
| { |
| return read_bar(MMIO_HOST_CSR); |
| } |
| |
| static void write_host_csr(uint32_t data) |
| { |
| write_bar(MMIO_HOST_CSR, data); |
| } |
| |
| static size_t filled_slots(uint32_t data) |
| { |
| uint8_t wp, rp; |
| rp = data >> CSR_RP_START; |
| wp = data >> CSR_WP_START; |
| return (uint8_t) (wp - rp); |
| } |
| |
| static size_t cse_filled_slots(void) |
| { |
| return filled_slots(read_cse_csr()); |
| } |
| |
| static size_t host_empty_slots(void) |
| { |
| uint32_t csr; |
| csr = read_host_csr(); |
| |
| return ((csr & CSR_CBD) >> CSR_CBD_START) - filled_slots(csr); |
| } |
| |
| static void clear_int(void) |
| { |
| uint32_t csr; |
| csr = read_host_csr(); |
| csr |= CSR_IS; |
| write_host_csr(csr); |
| } |
| |
| static uint32_t read_slot(void) |
| { |
| return read_bar(MMIO_CSE_CB_RW); |
| } |
| |
| static void write_slot(uint32_t val) |
| { |
| write_bar(MMIO_CSE_CB_WW, val); |
| } |
| |
| static int wait_write_slots(size_t cnt) |
| { |
| struct stopwatch sw; |
| |
| stopwatch_init_msecs_expire(&sw, HECI_SEND_TIMEOUT); |
| while (host_empty_slots() < cnt) { |
| udelay(HECI_DELAY); |
| if (stopwatch_expired(&sw)) { |
| printk(BIOS_ERR, "HECI: timeout, buffer not drained\n"); |
| return 0; |
| } |
| } |
| return 1; |
| } |
| |
| static int wait_read_slots(size_t cnt) |
| { |
| struct stopwatch sw; |
| |
| stopwatch_init_msecs_expire(&sw, HECI_READ_TIMEOUT); |
| while (cse_filled_slots() < cnt) { |
| udelay(HECI_DELAY); |
| if (stopwatch_expired(&sw)) { |
| printk(BIOS_ERR, "HECI: timed out reading answer!\n"); |
| return 0; |
| } |
| } |
| return 1; |
| } |
| |
| /* get number of full 4-byte slots */ |
| static size_t bytes_to_slots(size_t bytes) |
| { |
| return ALIGN_UP(bytes, SLOT_SIZE) / SLOT_SIZE; |
| } |
| |
| static int cse_ready(void) |
| { |
| uint32_t csr; |
| csr = read_cse_csr(); |
| return csr & CSR_READY; |
| } |
| |
| static bool cse_check_hfs1_com(int mode) |
| { |
| union me_hfsts1 hfs1; |
| hfs1.data = me_read_config32(PCI_ME_HFSTS1); |
| return hfs1.fields.operation_mode == mode; |
| } |
| |
| bool cse_is_hfs1_cws_normal(void) |
| { |
| union me_hfsts1 hfs1; |
| hfs1.data = me_read_config32(PCI_ME_HFSTS1); |
| if (hfs1.fields.working_state == ME_HFS1_CWS_NORMAL) |
| return true; |
| return false; |
| } |
| |
| bool cse_is_hfs1_com_normal(void) |
| { |
| return cse_check_hfs1_com(ME_HFS1_COM_NORMAL); |
| } |
| |
| bool cse_is_hfs1_com_secover_mei_msg(void) |
| { |
| return cse_check_hfs1_com(ME_HFS1_COM_SECOVER_MEI_MSG); |
| } |
| |
| bool cse_is_hfs1_com_soft_temp_disable(void) |
| { |
| return cse_check_hfs1_com(ME_HFS1_COM_SOFT_TEMP_DISABLE); |
| } |
| |
| bool cse_is_hfs3_fw_sku_custom(void) |
| { |
| union me_hfsts3 hfs3; |
| hfs3.data = me_read_config32(PCI_ME_HFSTS3); |
| return hfs3.fields.fw_sku == ME_HFS3_FW_SKU_CUSTOM; |
| } |
| |
| /* Makes the host ready to communicate with CSE */ |
| void cse_set_host_ready(void) |
| { |
| uint32_t csr; |
| csr = read_host_csr(); |
| csr &= ~CSR_RESET; |
| csr |= (CSR_IG | CSR_READY); |
| write_host_csr(csr); |
| } |
| |
| /* Polls for ME mode ME_HFS1_COM_SECOVER_MEI_MSG for 15 seconds */ |
| uint8_t cse_wait_sec_override_mode(void) |
| { |
| struct stopwatch sw; |
| stopwatch_init_msecs_expire(&sw, HECI_DELAY_READY); |
| while (!cse_is_hfs1_com_secover_mei_msg()) { |
| udelay(HECI_DELAY); |
| if (stopwatch_expired(&sw)) { |
| printk(BIOS_ERR, "HECI: Timed out waiting for SEC_OVERRIDE mode!\n"); |
| return 0; |
| } |
| } |
| printk(BIOS_DEBUG, "HECI: CSE took %lu ms to enter security override mode\n", |
| stopwatch_duration_msecs(&sw)); |
| return 1; |
| } |
| |
| /* |
| * Polls for CSE's current operation mode 'Soft Temporary Disable'. |
| * The CSE enters the current operation mode when it boots from RO(BP1). |
| */ |
| uint8_t cse_wait_com_soft_temp_disable(void) |
| { |
| struct stopwatch sw; |
| stopwatch_init_msecs_expire(&sw, CSE_DELAY_BOOT_TO_RO); |
| while (!cse_is_hfs1_com_soft_temp_disable()) { |
| udelay(HECI_DELAY); |
| if (stopwatch_expired(&sw)) { |
| printk(BIOS_ERR, "HECI: Timed out waiting for CSE to boot from RO!\n"); |
| return 0; |
| } |
| } |
| printk(BIOS_SPEW, "HECI: CSE took %lu ms to boot from RO\n", |
| stopwatch_duration_msecs(&sw)); |
| return 1; |
| } |
| |
| static int wait_heci_ready(void) |
| { |
| struct stopwatch sw; |
| |
| stopwatch_init_msecs_expire(&sw, HECI_DELAY_READY); |
| while (!cse_ready()) { |
| udelay(HECI_DELAY); |
| if (stopwatch_expired(&sw)) |
| return 0; |
| } |
| |
| return 1; |
| } |
| |
| static void host_gen_interrupt(void) |
| { |
| uint32_t csr; |
| csr = read_host_csr(); |
| csr |= CSR_IG; |
| write_host_csr(csr); |
| } |
| |
| static size_t hdr_get_length(uint32_t hdr) |
| { |
| return (hdr & MEI_HDR_LENGTH) >> MEI_HDR_LENGTH_START; |
| } |
| |
| static int |
| send_one_message(uint32_t hdr, const void *buff) |
| { |
| size_t pend_len, pend_slots, remainder, i; |
| uint32_t tmp; |
| const uint32_t *p = buff; |
| |
| /* Get space for the header */ |
| if (!wait_write_slots(1)) |
| return 0; |
| |
| /* First, write header */ |
| write_slot(hdr); |
| |
| pend_len = hdr_get_length(hdr); |
| pend_slots = bytes_to_slots(pend_len); |
| |
| if (!wait_write_slots(pend_slots)) |
| return 0; |
| |
| /* Write the body in whole slots */ |
| i = 0; |
| while (i < ALIGN_DOWN(pend_len, SLOT_SIZE)) { |
| write_slot(*p++); |
| i += SLOT_SIZE; |
| } |
| |
| remainder = pend_len % SLOT_SIZE; |
| /* Pad to 4 bytes not touching caller's buffer */ |
| if (remainder) { |
| memcpy(&tmp, p, remainder); |
| write_slot(tmp); |
| } |
| |
| host_gen_interrupt(); |
| |
| /* Make sure nothing bad happened during transmission */ |
| if (!cse_ready()) |
| return 0; |
| |
| return pend_len; |
| } |
| |
| int |
| heci_send(const void *msg, size_t len, uint8_t host_addr, uint8_t client_addr) |
| { |
| uint8_t retry; |
| uint32_t csr, hdr; |
| size_t sent, remaining, cb_size, max_length; |
| const uint8_t *p; |
| |
| if (!msg || !len) |
| return 0; |
| |
| clear_int(); |
| |
| for (retry = 0; retry < MAX_HECI_MESSAGE_RETRY_COUNT; retry++) { |
| p = msg; |
| |
| if (!wait_heci_ready()) { |
| printk(BIOS_ERR, "HECI: not ready\n"); |
| continue; |
| } |
| |
| csr = read_host_csr(); |
| cb_size = ((csr & CSR_CBD) >> CSR_CBD_START) * SLOT_SIZE; |
| /* |
| * Reserve one slot for the header. Limit max message |
| * length by 9 bits that are available in the header. |
| */ |
| max_length = MIN(cb_size, (1 << MEI_HDR_LENGTH_SIZE) - 1) |
| - SLOT_SIZE; |
| remaining = len; |
| |
| /* |
| * Fragment the message into smaller messages not exceeding |
| * useful circular buffer length. Mark last message complete. |
| */ |
| do { |
| hdr = MIN(max_length, remaining) |
| << MEI_HDR_LENGTH_START; |
| hdr |= client_addr << MEI_HDR_CSE_ADDR_START; |
| hdr |= host_addr << MEI_HDR_HOST_ADDR_START; |
| hdr |= (MIN(max_length, remaining) == remaining) ? |
| MEI_HDR_IS_COMPLETE : 0; |
| sent = send_one_message(hdr, p); |
| p += sent; |
| remaining -= sent; |
| } while (remaining > 0 && sent != 0); |
| |
| if (!remaining) |
| return 1; |
| } |
| return 0; |
| } |
| |
| static size_t |
| recv_one_message(uint32_t *hdr, void *buff, size_t maxlen) |
| { |
| uint32_t reg, *p = buff; |
| size_t recv_slots, recv_len, remainder, i; |
| |
| /* first get the header */ |
| if (!wait_read_slots(1)) |
| return 0; |
| |
| *hdr = read_slot(); |
| recv_len = hdr_get_length(*hdr); |
| |
| if (!recv_len) |
| printk(BIOS_WARNING, "HECI: message is zero-sized\n"); |
| |
| recv_slots = bytes_to_slots(recv_len); |
| |
| i = 0; |
| if (recv_len > maxlen) { |
| printk(BIOS_ERR, "HECI: response is too big\n"); |
| return 0; |
| } |
| |
| /* wait for the rest of messages to arrive */ |
| wait_read_slots(recv_slots); |
| |
| /* fetch whole slots first */ |
| while (i < ALIGN_DOWN(recv_len, SLOT_SIZE)) { |
| *p++ = read_slot(); |
| i += SLOT_SIZE; |
| } |
| |
| /* |
| * If ME is not ready, something went wrong and |
| * we received junk |
| */ |
| if (!cse_ready()) |
| return 0; |
| |
| remainder = recv_len % SLOT_SIZE; |
| |
| if (remainder) { |
| reg = read_slot(); |
| memcpy(p, ®, remainder); |
| } |
| |
| return recv_len; |
| } |
| |
| int heci_receive(void *buff, size_t *maxlen) |
| { |
| uint8_t retry; |
| size_t left, received; |
| uint32_t hdr = 0; |
| uint8_t *p; |
| |
| if (!buff || !maxlen || !*maxlen) |
| return 0; |
| |
| clear_int(); |
| |
| for (retry = 0; retry < MAX_HECI_MESSAGE_RETRY_COUNT; retry++) { |
| p = buff; |
| left = *maxlen; |
| |
| if (!wait_heci_ready()) { |
| printk(BIOS_ERR, "HECI: not ready\n"); |
| continue; |
| } |
| |
| /* |
| * Receive multiple packets until we meet one marked |
| * complete or we run out of space in caller-provided buffer. |
| */ |
| do { |
| received = recv_one_message(&hdr, p, left); |
| if (!received) { |
| printk(BIOS_ERR, "HECI: Failed to receive!\n"); |
| return 0; |
| } |
| left -= received; |
| p += received; |
| /* If we read out everything ping to send more */ |
| if (!(hdr & MEI_HDR_IS_COMPLETE) && !cse_filled_slots()) |
| host_gen_interrupt(); |
| } while (received && !(hdr & MEI_HDR_IS_COMPLETE) && left > 0); |
| |
| if ((hdr & MEI_HDR_IS_COMPLETE) && received) { |
| *maxlen = p - (uint8_t *) buff; |
| return 1; |
| } |
| } |
| return 0; |
| } |
| |
| int heci_send_receive(const void *snd_msg, size_t snd_sz, void *rcv_msg, size_t *rcv_sz) |
| { |
| if (!heci_send(snd_msg, snd_sz, BIOS_HOST_ADDR, HECI_MKHI_ADDR)) { |
| printk(BIOS_ERR, "HECI: send Failed\n"); |
| return 0; |
| } |
| |
| if (rcv_msg != NULL) { |
| if (!heci_receive(rcv_msg, rcv_sz)) { |
| printk(BIOS_ERR, "HECI: receive Failed\n"); |
| return 0; |
| } |
| } |
| return 1; |
| } |
| |
| /* |
| * Attempt to reset the device. This is useful when host and ME are out |
| * of sync during transmission or ME didn't understand the message. |
| */ |
| int heci_reset(void) |
| { |
| uint32_t csr; |
| |
| /* Send reset request */ |
| csr = read_host_csr(); |
| csr |= (CSR_RESET | CSR_IG); |
| write_host_csr(csr); |
| |
| if (wait_heci_ready()) { |
| /* Device is back on its imaginary feet, clear reset */ |
| cse_set_host_ready(); |
| return 1; |
| } |
| |
| printk(BIOS_CRIT, "HECI: reset failed\n"); |
| |
| return 0; |
| } |
| |
| bool is_cse_enabled(void) |
| { |
| const struct device *cse_dev = pcidev_path_on_root(PCH_DEVFN_CSE); |
| |
| if (!cse_dev || !cse_dev->enabled) { |
| printk(BIOS_WARNING, "HECI: No CSE device\n"); |
| return false; |
| } |
| |
| if (pci_read_config16(PCH_DEV_CSE, PCI_VENDOR_ID) == 0xFFFF) { |
| printk(BIOS_WARNING, "HECI: CSE device is hidden\n"); |
| return false; |
| } |
| |
| return true; |
| } |
| |
| uint32_t me_read_config32(int offset) |
| { |
| return pci_read_config32(PCH_DEV_CSE, offset); |
| } |
| |
| static bool cse_is_global_reset_allowed(void) |
| { |
| /* |
| * Allow sending GLOBAL_RESET command only if: |
| * - CSE's current working state is Normal and current operation mode is Normal. |
| * - (or) CSE's current working state is normal and current operation mode can |
| * be Soft Temp Disable or Security Override Mode if CSE's Firmware SKU is |
| * Custom. |
| */ |
| if (!cse_is_hfs1_cws_normal()) |
| return false; |
| |
| if (cse_is_hfs1_com_normal()) |
| return true; |
| |
| if (cse_is_hfs3_fw_sku_custom()) { |
| if (cse_is_hfs1_com_soft_temp_disable() || cse_is_hfs1_com_secover_mei_msg()) |
| return true; |
| } |
| return false; |
| } |
| |
| /* |
| * Sends GLOBAL_RESET_REQ cmd to CSE.The reset type can be GLOBAL_RESET/CSE_RESET_ONLY. |
| */ |
| int cse_request_global_reset(enum rst_req_type rst_type) |
| { |
| int status; |
| struct mkhi_hdr reply; |
| struct reset_message { |
| struct mkhi_hdr hdr; |
| uint8_t req_origin; |
| uint8_t reset_type; |
| } __packed; |
| struct reset_message msg = { |
| .hdr = { |
| .group_id = MKHI_GROUP_ID_CBM, |
| .command = MKHI_CBM_GLOBAL_RESET_REQ, |
| }, |
| .req_origin = GR_ORIGIN_BIOS_POST, |
| .reset_type = rst_type |
| }; |
| size_t reply_size; |
| |
| printk(BIOS_DEBUG, "HECI: Global Reset(Type:%d) Command\n", rst_type); |
| |
| if (!(rst_type == GLOBAL_RESET || rst_type == CSE_RESET_ONLY)) { |
| printk(BIOS_ERR, "HECI: Unsupported reset type is requested\n"); |
| return 0; |
| } |
| |
| if (!cse_is_global_reset_allowed()) { |
| printk(BIOS_ERR, "HECI: CSE does not meet required prerequisites\n"); |
| return 0; |
| } |
| |
| heci_reset(); |
| |
| reply_size = sizeof(reply); |
| memset(&reply, 0, reply_size); |
| |
| if (rst_type == CSE_RESET_ONLY) |
| status = heci_send(&msg, sizeof(msg), BIOS_HOST_ADDR, HECI_MKHI_ADDR); |
| else |
| status = heci_send_receive(&msg, sizeof(msg), &reply, &reply_size); |
| |
| printk(BIOS_DEBUG, "HECI: Global Reset %s!\n", status ? "success" : "failure"); |
| return status; |
| } |
| |
| static bool cse_is_hmrfpo_enable_allowed(void) |
| { |
| /* |
| * Allow sending HMRFPO ENABLE command only if: |
| * - CSE's current working state is Normal and current operation mode is Normal |
| * - (or) cse's current working state is normal and current operation mode is |
| * Soft Temp Disable if CSE's Firmware SKU is Custom |
| */ |
| if (!cse_is_hfs1_cws_normal()) |
| return false; |
| |
| if (cse_is_hfs1_com_normal()) |
| return true; |
| |
| if (cse_is_hfs3_fw_sku_custom() && cse_is_hfs1_com_soft_temp_disable()) |
| return true; |
| |
| return false; |
| } |
| |
| /* Sends HMRFPO Enable command to CSE */ |
| int cse_hmrfpo_enable(void) |
| { |
| struct hmrfpo_enable_msg { |
| struct mkhi_hdr hdr; |
| uint32_t nonce[2]; |
| } __packed; |
| |
| /* HMRFPO Enable message */ |
| struct hmrfpo_enable_msg msg = { |
| .hdr = { |
| .group_id = MKHI_GROUP_ID_HMRFPO, |
| .command = MKHI_HMRFPO_ENABLE, |
| }, |
| .nonce = {0}, |
| }; |
| |
| /* HMRFPO Enable response */ |
| struct hmrfpo_enable_resp { |
| struct mkhi_hdr hdr; |
| /* Base addr for factory data area, not relevant for client SKUs */ |
| uint32_t fct_base; |
| /* Length of factory data area, not relevant for client SKUs */ |
| uint32_t fct_limit; |
| uint8_t status; |
| uint8_t reserved[3]; |
| } __packed; |
| |
| struct hmrfpo_enable_resp resp; |
| size_t resp_size = sizeof(struct hmrfpo_enable_resp); |
| |
| printk(BIOS_DEBUG, "HECI: Send HMRFPO Enable Command\n"); |
| |
| if (!cse_is_hmrfpo_enable_allowed()) { |
| printk(BIOS_ERR, "HECI: CSE does not meet required prerequisites\n"); |
| return 0; |
| } |
| |
| if (!heci_send_receive(&msg, sizeof(struct hmrfpo_enable_msg), |
| &resp, &resp_size)) |
| return 0; |
| |
| if (resp.hdr.result) { |
| printk(BIOS_ERR, "HECI: Resp Failed:%d\n", resp.hdr.result); |
| return 0; |
| } |
| |
| if (resp.status) { |
| printk(BIOS_ERR, "HECI: HMRFPO_Enable Failed (resp status: %d)\n", resp.status); |
| return 0; |
| } |
| |
| return 1; |
| } |
| |
| /* |
| * Sends HMRFPO Get Status command to CSE to get the HMRFPO status. |
| * The status can be DISABLED/LOCKED/ENABLED |
| */ |
| int cse_hmrfpo_get_status(void) |
| { |
| struct hmrfpo_get_status_msg { |
| struct mkhi_hdr hdr; |
| } __packed; |
| |
| struct hmrfpo_get_status_resp { |
| struct mkhi_hdr hdr; |
| uint8_t status; |
| uint8_t reserved[3]; |
| } __packed; |
| |
| struct hmrfpo_get_status_msg msg = { |
| .hdr = { |
| .group_id = MKHI_GROUP_ID_HMRFPO, |
| .command = MKHI_HMRFPO_GET_STATUS, |
| }, |
| }; |
| struct hmrfpo_get_status_resp resp; |
| size_t resp_size = sizeof(struct hmrfpo_get_status_resp); |
| |
| printk(BIOS_INFO, "HECI: Sending Get HMRFPO Status Command\n"); |
| |
| if (!cse_is_hfs1_cws_normal()) { |
| printk(BIOS_ERR, "HECI: CSE's current working state is not Normal\n"); |
| return -1; |
| } |
| |
| if (!heci_send_receive(&msg, sizeof(struct hmrfpo_get_status_msg), |
| &resp, &resp_size)) { |
| printk(BIOS_ERR, "HECI: HMRFPO send/receive fail\n"); |
| return -1; |
| } |
| |
| if (resp.hdr.result) { |
| printk(BIOS_ERR, "HECI: HMRFPO Resp Failed:%d\n", |
| resp.hdr.result); |
| return -1; |
| } |
| |
| return resp.status; |
| } |
| |
| void print_me_fw_version(void *unused) |
| { |
| struct version { |
| uint16_t minor; |
| uint16_t major; |
| uint16_t build; |
| uint16_t hotfix; |
| } __packed; |
| |
| struct fw_ver_resp { |
| struct mkhi_hdr hdr; |
| struct version code; |
| struct version rec; |
| struct version fitc; |
| } __packed; |
| |
| const struct mkhi_hdr fw_ver_msg = { |
| .group_id = MKHI_GROUP_ID_GEN, |
| .command = MKHI_GEN_GET_FW_VERSION, |
| }; |
| |
| struct fw_ver_resp resp; |
| size_t resp_size = sizeof(resp); |
| |
| /* Ignore if UART debugging is disabled */ |
| if (!CONFIG(CONSOLE_SERIAL)) |
| return; |
| |
| /* Ignore if CSE is disabled */ |
| if (!is_cse_enabled()) |
| return; |
| |
| /* |
| * Ignore if ME Firmware SKU type is custom since |
| * print_boot_partition_info() logs RO(BP1) and RW(BP2) versions. |
| */ |
| if (cse_is_hfs3_fw_sku_custom()) |
| return; |
| |
| /* |
| * Prerequisites: |
| * 1) HFSTS1 Current Working State is Normal |
| * 2) HFSTS1 Current Operation Mode is Normal |
| * 3) It's after DRAM INIT DONE message (taken care of by calling it |
| * during ramstage |
| */ |
| if (!cse_is_hfs1_cws_normal() || !cse_is_hfs1_com_normal()) |
| goto fail; |
| |
| heci_reset(); |
| |
| if (!heci_send_receive(&fw_ver_msg, sizeof(fw_ver_msg), &resp, &resp_size)) |
| goto fail; |
| |
| if (resp.hdr.result) |
| goto fail; |
| |
| printk(BIOS_DEBUG, "ME: Version: %d.%d.%d.%d\n", resp.code.major, |
| resp.code.minor, resp.code.hotfix, resp.code.build); |
| return; |
| |
| fail: |
| printk(BIOS_DEBUG, "ME: Version: Unavailable\n"); |
| } |
| |
| #if ENV_RAMSTAGE |
| |
| static void update_sec_bar(struct device *dev) |
| { |
| cse.sec_bar = find_resource(dev, PCI_BASE_ADDRESS_0)->base; |
| } |
| |
| static void cse_set_resources(struct device *dev) |
| { |
| if (dev->path.pci.devfn == PCH_DEVFN_CSE) |
| update_sec_bar(dev); |
| |
| pci_dev_set_resources(dev); |
| } |
| |
| static struct device_operations cse_ops = { |
| .set_resources = cse_set_resources, |
| .read_resources = pci_dev_read_resources, |
| .enable_resources = pci_dev_enable_resources, |
| .init = pci_dev_init, |
| .ops_pci = &pci_dev_ops_pci, |
| }; |
| |
| static const unsigned short pci_device_ids[] = { |
| PCI_DEVICE_ID_INTEL_APL_CSE0, |
| PCI_DEVICE_ID_INTEL_GLK_CSE0, |
| PCI_DEVICE_ID_INTEL_CNL_CSE0, |
| PCI_DEVICE_ID_INTEL_SKL_CSE0, |
| PCI_DEVICE_ID_INTEL_LWB_CSE0, |
| PCI_DEVICE_ID_INTEL_LWB_CSE0_SUPER, |
| PCI_DEVICE_ID_INTEL_CNP_H_CSE0, |
| PCI_DEVICE_ID_INTEL_ICL_CSE0, |
| PCI_DEVICE_ID_INTEL_CMP_CSE0, |
| PCI_DEVICE_ID_INTEL_CMP_H_CSE0, |
| PCI_DEVICE_ID_INTEL_TGL_CSE0, |
| PCI_DEVICE_ID_INTEL_MCC_CSE0, |
| PCI_DEVICE_ID_INTEL_MCC_CSE1, |
| PCI_DEVICE_ID_INTEL_MCC_CSE2, |
| PCI_DEVICE_ID_INTEL_MCC_CSE3, |
| PCI_DEVICE_ID_INTEL_JSP_CSE0, |
| PCI_DEVICE_ID_INTEL_JSP_CSE1, |
| PCI_DEVICE_ID_INTEL_JSP_CSE2, |
| PCI_DEVICE_ID_INTEL_JSP_CSE3, |
| 0, |
| }; |
| |
| static const struct pci_driver cse_driver __pci_driver = { |
| .ops = &cse_ops, |
| .vendor = PCI_VENDOR_ID_INTEL, |
| /* SoC/chipset needs to provide PCI device ID */ |
| .devices = pci_device_ids |
| }; |
| |
| #endif |