// SPDX-License-Identifier: GPL-2.0+ /* * K3: R5 Common Architecture initialization * * Copyright (C) 2023 Texas Instruments Incorporated - https://www.ti.com/ */ #include #include #include #include #include #include #include #include #include #include #include "../common.h" #if IS_ENABLED(CONFIG_SYS_K3_SPL_ATF) enum { IMAGE_ID_ATF, IMAGE_ID_OPTEE, IMAGE_ID_SPL, IMAGE_ID_DM_FW, IMAGE_ID_TIFSSTUB_HS, IMAGE_ID_TIFSSTUB_FS, IMAGE_ID_T, IMAGE_AMT, }; #if CONFIG_IS_ENABLED(FIT_IMAGE_POST_PROCESS) static const char *image_os_match[IMAGE_AMT] = { "arm-trusted-firmware", "tee", "U-Boot", "DM", "tifsstub-hs", "tifsstub-fs", "tifsstub-gp", }; #endif static struct image_info fit_image_info[IMAGE_AMT]; void init_env(void) { #ifdef CONFIG_SPL_ENV_SUPPORT char *part; env_init(); env_relocate(); switch (spl_boot_device()) { case BOOT_DEVICE_MMC2: part = env_get("bootpart"); env_set("storage_interface", "mmc"); env_set("fw_dev_part", part); break; case BOOT_DEVICE_SPI: env_set("storage_interface", "ubi"); env_set("fw_ubi_mtdpart", "UBI"); env_set("fw_ubi_volume", "UBI0"); break; default: printf("%s from device %u not supported!\n", __func__, spl_boot_device()); return; } #endif } int load_firmware(char *name_fw, char *name_loadaddr, u32 *loadaddr) { struct udevice *fsdev; char *name = NULL; int size = 0; if (!CONFIG_IS_ENABLED(FS_LOADER)) return 0; *loadaddr = 0; #ifdef CONFIG_SPL_ENV_SUPPORT switch (spl_boot_device()) { case BOOT_DEVICE_MMC2: name = env_get(name_fw); *loadaddr = env_get_hex(name_loadaddr, *loadaddr); break; default: printf("Loading rproc fw image from device %u not supported!\n", spl_boot_device()); return 0; } #endif if (!*loadaddr) return 0; if (!get_fs_loader(&fsdev)) { size = request_firmware_into_buf(fsdev, name, (void *)*loadaddr, 0, 0); } return size; } void release_resources_for_core_shutdown(void) { struct ti_sci_handle *ti_sci = get_ti_sci_handle(); struct ti_sci_dev_ops *dev_ops = &ti_sci->ops.dev_ops; struct ti_sci_proc_ops *proc_ops = &ti_sci->ops.proc_ops; int ret; u32 i; /* Iterate through list of devices to put (shutdown) */ for (i = 0; i < ARRAY_SIZE(put_device_ids); i++) { u32 id = put_device_ids[i]; ret = dev_ops->put_device(ti_sci, id); if (ret) panic("Failed to put device %u (%d)\n", id, ret); } /* Iterate through list of cores to put (shutdown) */ for (i = 0; i < ARRAY_SIZE(put_core_ids); i++) { u32 id = put_core_ids[i]; /* * Queue up the core shutdown request. Note that this call * needs to be followed up by an actual invocation of an WFE * or WFI CPU instruction. */ ret = proc_ops->proc_shutdown_no_wait(ti_sci, id); if (ret) panic("Failed sending core %u shutdown message (%d)\n", id, ret); } } void __noreturn jump_to_image_no_args(struct spl_image_info *spl_image) { typedef void __noreturn (*image_entry_noargs_t)(void); struct ti_sci_handle *ti_sci = get_ti_sci_handle(); u32 loadaddr = 0; int ret, size = 0, shut_cpu = 0; /* Release all the exclusive devices held by SPL before starting ATF */ ti_sci->ops.dev_ops.release_exclusive_devices(ti_sci); ret = rproc_init(); if (ret) panic("rproc failed to be initialized (%d)\n", ret); init_env(); if (!fit_image_info[IMAGE_ID_DM_FW].image_start) { size = load_firmware("name_mcur5f0_0fw", "addr_mcur5f0_0load", &loadaddr); } /* * It is assumed that remoteproc device 1 is the corresponding * Cortex-A core which runs ATF. Make sure DT reflects the same. */ if (!fit_image_info[IMAGE_ID_ATF].image_start) fit_image_info[IMAGE_ID_ATF].image_start = spl_image->entry_point; ret = rproc_load(1, fit_image_info[IMAGE_ID_ATF].image_start, 0x200); if (ret) panic("%s: ATF failed to load on rproc (%d)\n", __func__, ret); #if CONFIG_IS_ENABLED(FIT_IMAGE_POST_PROCESS) /* Authenticate ATF */ void *image_addr = (void *)fit_image_info[IMAGE_ID_ATF].image_start; debug("%s: Authenticating image: addr=%lx, size=%ld, os=%s\n", __func__, fit_image_info[IMAGE_ID_ATF].image_start, fit_image_info[IMAGE_ID_ATF].image_len, image_os_match[IMAGE_ID_ATF]); ti_secure_image_post_process(&image_addr, (size_t *)&fit_image_info[IMAGE_ID_ATF].image_len); /* Authenticate OPTEE */ image_addr = (void *)fit_image_info[IMAGE_ID_OPTEE].image_start; debug("%s: Authenticating image: addr=%lx, size=%ld, os=%s\n", __func__, fit_image_info[IMAGE_ID_OPTEE].image_start, fit_image_info[IMAGE_ID_OPTEE].image_len, image_os_match[IMAGE_ID_OPTEE]); ti_secure_image_post_process(&image_addr, (size_t *)&fit_image_info[IMAGE_ID_OPTEE].image_len); #endif if (!fit_image_info[IMAGE_ID_DM_FW].image_len && !(size > 0 && valid_elf_image(loadaddr))) { shut_cpu = 1; goto start_arm64; } if (!fit_image_info[IMAGE_ID_DM_FW].image_start) { loadaddr = load_elf_image_phdr(loadaddr); } else { loadaddr = fit_image_info[IMAGE_ID_DM_FW].image_start; if (valid_elf_image(loadaddr)) loadaddr = load_elf_image_phdr(loadaddr); } debug("%s: jumping to address %x\n", __func__, loadaddr); start_arm64: /* Add an extra newline to differentiate the ATF logs from SPL */ printf("Starting ATF on ARM64 core...\n\n"); ret = rproc_start(1); if (ret) panic("%s: ATF failed to start on rproc (%d)\n", __func__, ret); if (shut_cpu) { debug("Shutting down...\n"); release_resources_for_core_shutdown(); while (1) asm volatile("wfe"); } image_entry_noargs_t image_entry = (image_entry_noargs_t)loadaddr; image_entry(); } #endif void disable_linefill_optimization(void) { u32 actlr; /* * On K3 devices there are 2 conditions where R5F can deadlock: * 1.When software is performing series of store operations to * cacheable write back/write allocate memory region and later * on software execute barrier operation (DSB or DMB). R5F may * hang at the barrier instruction. * 2.When software is performing a mix of load and store operations * within a tight loop and store operations are all writing to * cacheable write back/write allocates memory regions, R5F may * hang at one of the load instruction. * * To avoid the above two conditions disable linefill optimization * inside Cortex R5F. */ asm("mrc p15, 0, %0, c1, c0, 1" : "=r" (actlr)); actlr |= (1 << 13); /* Set DLFO bit */ asm("mcr p15, 0, %0, c1, c0, 1" : : "r" (actlr)); } static void remove_fwl_regions(struct fwl_data fwl_data, size_t num_regions, enum k3_firewall_region_type fwl_type) { struct ti_sci_fwl_ops *fwl_ops; struct ti_sci_handle *ti_sci; struct ti_sci_msg_fwl_region region; size_t j; ti_sci = get_ti_sci_handle(); fwl_ops = &ti_sci->ops.fwl_ops; for (j = 0; j < fwl_data.regions; j++) { region.fwl_id = fwl_data.fwl_id; region.region = j; region.n_permission_regs = 3; fwl_ops->get_fwl_region(ti_sci, ®ion); /* Don't disable the background regions */ if (region.control != 0 && ((region.control >> K3_FIREWALL_BACKGROUND_BIT) & 1) == fwl_type) { pr_debug("Attempting to disable firewall %5d (%25s)\n", region.fwl_id, fwl_data.name); region.control = 0; if (fwl_ops->set_fwl_region(ti_sci, ®ion)) pr_err("Could not disable firewall %5d (%25s)\n", region.fwl_id, fwl_data.name); } } } void remove_fwl_configs(struct fwl_data *fwl_data, size_t fwl_data_size) { size_t i; for (i = 0; i < fwl_data_size; i++) { remove_fwl_regions(fwl_data[i], fwl_data[i].regions, K3_FIREWALL_REGION_FOREGROUND); remove_fwl_regions(fwl_data[i], fwl_data[i].regions, K3_FIREWALL_REGION_BACKGROUND); } } #if CONFIG_IS_ENABLED(FIT_IMAGE_POST_PROCESS) void board_fit_image_post_process(const void *fit, int node, void **p_image, size_t *p_size) { int len; int i; const char *os; u32 addr; os = fdt_getprop(fit, node, "os", &len); addr = fdt_getprop_u32_default_node(fit, node, 0, "entry", -1); debug("%s: processing image: addr=%x, size=%d, os=%s\n", __func__, addr, *p_size, os); for (i = 0; i < IMAGE_AMT; i++) { if (!strcmp(os, image_os_match[i])) { fit_image_info[i].image_start = addr; fit_image_info[i].image_len = *p_size; debug("%s: matched image for ID %d\n", __func__, i); break; } } if (i < IMAGE_AMT && i > IMAGE_ID_DM_FW) { int device_type = get_device_type(); if ((device_type == K3_DEVICE_TYPE_HS_SE && strcmp(os, "tifsstub-hs")) || (device_type == K3_DEVICE_TYPE_HS_FS && strcmp(os, "tifsstub-fs")) || (device_type == K3_DEVICE_TYPE_GP && strcmp(os, "tifsstub-gp"))) { *p_size = 0; } else { debug("tifsstub-type: %s\n", os); } return; } /* * Only DM and the DTBs are being authenticated here, * rest will be authenticated when A72 cluster is up */ if ((i != IMAGE_ID_ATF) && (i != IMAGE_ID_OPTEE)) { ti_secure_image_check_binary(p_image, p_size); ti_secure_image_post_process(p_image, p_size); } else { ti_secure_image_check_binary(p_image, p_size); } } #endif