mirror of
https://github.com/ARM-software/arm-trusted-firmware.git
synced 2025-04-08 05:43:53 +00:00

Errata application is painful for performance. For a start, it's done when the core has just come out of reset, which means branch predictors and caches will be empty so a branch to a workaround function must be fetched from memory and that round trip is very slow. Then it also runs with the I-cache off, which means that the loop to iterate over the workarounds must also be fetched from memory on each iteration. We can remove both branches. First, we can simply apply every erratum directly instead of defining a workaround function and jumping to it. Currently, no errata that need to be applied at both reset and runtime, with the same workaround function, exist. If the need arose in future, this should be achievable with a reset + runtime wrapper combo. Then, we can construct a function that applies each erratum linearly instead of looping over the list. If this function is part of the reset function, then the only "far" branches at reset will be for the checker functions. Importantly, this mitigates the slowdown even when an erratum is disabled. The result is ~50% speedup on N1SDP and ~20% on AArch64 Juno on wakeup from PSCI calls that end in powerdown. This is roughly back to the baseline of v2.9, before the errata framework regressed on performance (or a little better). It is important to note that there are other slowdowns since then that remain unknown. Change-Id: Ie4d5288a331b11fd648e5c4a0b652b74160b07b9 Signed-off-by: Boyan Karatotev <boyan.karatotev@arm.com>
303 lines
8.5 KiB
ArmAsm
303 lines
8.5 KiB
ArmAsm
/*
|
|
* Copyright (c) 2015-2025, Arm Limited and Contributors. All rights reserved.
|
|
*
|
|
* SPDX-License-Identifier: BSD-3-Clause
|
|
*/
|
|
#include <arch.h>
|
|
#include <asm_macros.S>
|
|
#include <assert_macros.S>
|
|
#include <cortex_a72.h>
|
|
#include <cpu_macros.S>
|
|
#include <plat_macros.S>
|
|
#include "wa_cve_2022_23960_bhb_vector.S"
|
|
|
|
#if WORKAROUND_CVE_2022_23960
|
|
wa_cve_2022_23960_bhb_vector_table CORTEX_A72_BHB_LOOP_COUNT, cortex_a72
|
|
#endif /* WORKAROUND_CVE_2022_23960 */
|
|
|
|
cpu_reset_prologue cortex_a72
|
|
|
|
/* ---------------------------------------------
|
|
* Disable L1 data cache and unified L2 cache
|
|
* ---------------------------------------------
|
|
*/
|
|
func cortex_a72_disable_dcache
|
|
mrs x1, sctlr_el3
|
|
bic x1, x1, #SCTLR_C_BIT
|
|
msr sctlr_el3, x1
|
|
isb
|
|
ret
|
|
endfunc cortex_a72_disable_dcache
|
|
|
|
/* ---------------------------------------------
|
|
* Disable all types of L2 prefetches.
|
|
* ---------------------------------------------
|
|
*/
|
|
func cortex_a72_disable_l2_prefetch
|
|
mrs x0, CORTEX_A72_ECTLR_EL1
|
|
orr x0, x0, #CORTEX_A72_ECTLR_DIS_TWD_ACC_PFTCH_BIT
|
|
mov x1, #CORTEX_A72_ECTLR_L2_IPFTCH_DIST_MASK
|
|
orr x1, x1, #CORTEX_A72_ECTLR_L2_DPFTCH_DIST_MASK
|
|
bic x0, x0, x1
|
|
msr CORTEX_A72_ECTLR_EL1, x0
|
|
isb
|
|
ret
|
|
endfunc cortex_a72_disable_l2_prefetch
|
|
|
|
/* ---------------------------------------------
|
|
* Disable the load-store hardware prefetcher.
|
|
* ---------------------------------------------
|
|
*/
|
|
func cortex_a72_disable_hw_prefetcher
|
|
sysreg_bit_set CORTEX_A72_CPUACTLR_EL1, CORTEX_A72_CPUACTLR_EL1_DISABLE_L1_DCACHE_HW_PFTCH
|
|
isb
|
|
dsb ish
|
|
ret
|
|
endfunc cortex_a72_disable_hw_prefetcher
|
|
|
|
/* ---------------------------------------------
|
|
* Disable intra-cluster coherency
|
|
* ---------------------------------------------
|
|
*/
|
|
func cortex_a72_disable_smp
|
|
sysreg_bit_clear CORTEX_A72_ECTLR_EL1, CORTEX_A72_ECTLR_SMP_BIT
|
|
ret
|
|
endfunc cortex_a72_disable_smp
|
|
|
|
/* ---------------------------------------------
|
|
* Disable debug interfaces
|
|
* ---------------------------------------------
|
|
*/
|
|
func cortex_a72_disable_ext_debug
|
|
mov x0, #1
|
|
msr osdlr_el1, x0
|
|
isb
|
|
dsb sy
|
|
ret
|
|
endfunc cortex_a72_disable_ext_debug
|
|
|
|
func check_smccc_arch_workaround_3
|
|
cpu_check_csv2 x0, 1f
|
|
mov x0, #ERRATA_APPLIES
|
|
ret
|
|
1:
|
|
mov x0, #ERRATA_NOT_APPLIES
|
|
ret
|
|
endfunc check_smccc_arch_workaround_3
|
|
|
|
workaround_reset_start cortex_a72, ERRATUM(859971), ERRATA_A72_859971
|
|
sysreg_bit_set CORTEX_A72_CPUACTLR_EL1, CORTEX_A72_CPUACTLR_EL1_DIS_INSTR_PREFETCH
|
|
workaround_reset_end cortex_a72, ERRATUM(859971)
|
|
|
|
check_erratum_ls cortex_a72, ERRATUM(859971), CPU_REV(0, 3)
|
|
|
|
/* Due to the nature of the errata it is applied unconditionally when chosen */
|
|
check_erratum_chosen cortex_a72, ERRATUM(1319367), ERRATA_A72_1319367
|
|
/* erratum workaround is interleaved with generic code */
|
|
add_erratum_entry cortex_a72, ERRATUM(1319367), ERRATA_A72_1319367
|
|
|
|
workaround_reset_start cortex_a72, CVE(2017, 5715), WORKAROUND_CVE_2017_5715
|
|
#if IMAGE_BL31
|
|
override_vector_table wa_cve_2017_5715_mmu_vbar
|
|
#endif
|
|
workaround_reset_end cortex_a72, CVE(2017, 5715)
|
|
|
|
check_erratum_custom_start cortex_a72, CVE(2017, 5715)
|
|
cpu_check_csv2 x0, 1f
|
|
#if WORKAROUND_CVE_2017_5715
|
|
mov x0, #ERRATA_APPLIES
|
|
#else
|
|
mov x0, #ERRATA_MISSING
|
|
#endif
|
|
ret
|
|
1:
|
|
mov x0, #ERRATA_NOT_APPLIES
|
|
ret
|
|
check_erratum_custom_end cortex_a72, CVE(2017, 5715)
|
|
|
|
workaround_reset_start cortex_a72, CVE(2018, 3639), WORKAROUND_CVE_2018_3639
|
|
sysreg_bit_set CORTEX_A72_CPUACTLR_EL1, CORTEX_A72_CPUACTLR_EL1_DIS_LOAD_PASS_STORE
|
|
isb
|
|
dsb sy
|
|
workaround_reset_end cortex_a72, CVE(2018, 3639)
|
|
check_erratum_chosen cortex_a72, CVE(2018, 3639), WORKAROUND_CVE_2018_3639
|
|
|
|
workaround_reset_start cortex_a72, CVE(2022, 23960), WORKAROUND_CVE_2022_23960
|
|
#if IMAGE_BL31
|
|
/* Skip installing vector table again if already done for CVE(2017, 5715) */
|
|
/*
|
|
* The Cortex-A72 generic vectors are overridden to apply the
|
|
* mitigation on exception entry from lower ELs for revisions >= r1p0
|
|
* which has CSV2 implemented.
|
|
*/
|
|
adr x0, wa_cve_vbar_cortex_a72
|
|
mrs x1, vbar_el3
|
|
cmp x0, x1
|
|
b.eq 1f
|
|
msr vbar_el3, x0
|
|
1:
|
|
#endif /* IMAGE_BL31 */
|
|
workaround_reset_end cortex_a72, CVE(2022, 23960)
|
|
|
|
check_erratum_custom_start cortex_a72, CVE(2022, 23960)
|
|
#if WORKAROUND_CVE_2017_5715 || WORKAROUND_CVE_2022_23960
|
|
cpu_check_csv2 x0, 1f
|
|
mov x0, #ERRATA_APPLIES
|
|
ret
|
|
1:
|
|
#if WORKAROUND_CVE_2022_23960
|
|
mov x0, #ERRATA_APPLIES
|
|
#else
|
|
mov x0, #ERRATA_MISSING
|
|
#endif /* WORKAROUND_CVE_2022_23960 */
|
|
ret
|
|
#endif /* WORKAROUND_CVE_2017_5715 || WORKAROUND_CVE_2022_23960 */
|
|
mov x0, #ERRATA_MISSING
|
|
ret
|
|
check_erratum_custom_end cortex_a72, CVE(2022, 23960)
|
|
|
|
cpu_reset_func_start cortex_a72
|
|
|
|
/* ---------------------------------------------
|
|
* Enable the SMP bit.
|
|
* ---------------------------------------------
|
|
*/
|
|
sysreg_bit_set CORTEX_A72_ECTLR_EL1, CORTEX_A72_ECTLR_SMP_BIT
|
|
|
|
cpu_reset_func_end cortex_a72
|
|
|
|
/* ----------------------------------------------------
|
|
* The CPU Ops core power down function for Cortex-A72.
|
|
* ----------------------------------------------------
|
|
*/
|
|
func cortex_a72_core_pwr_dwn
|
|
mov x18, x30
|
|
|
|
/* ---------------------------------------------
|
|
* Turn off caches.
|
|
* ---------------------------------------------
|
|
*/
|
|
bl cortex_a72_disable_dcache
|
|
|
|
/* ---------------------------------------------
|
|
* Disable the L2 prefetches.
|
|
* ---------------------------------------------
|
|
*/
|
|
bl cortex_a72_disable_l2_prefetch
|
|
|
|
/* ---------------------------------------------
|
|
* Disable the load-store hardware prefetcher.
|
|
* ---------------------------------------------
|
|
*/
|
|
bl cortex_a72_disable_hw_prefetcher
|
|
|
|
/* ---------------------------------------------
|
|
* Flush L1 caches.
|
|
* ---------------------------------------------
|
|
*/
|
|
mov x0, #DCCISW
|
|
bl dcsw_op_level1
|
|
|
|
/* ---------------------------------------------
|
|
* Come out of intra cluster coherency
|
|
* ---------------------------------------------
|
|
*/
|
|
bl cortex_a72_disable_smp
|
|
|
|
/* ---------------------------------------------
|
|
* Force the debug interfaces to be quiescent
|
|
* ---------------------------------------------
|
|
*/
|
|
mov x30, x18
|
|
b cortex_a72_disable_ext_debug
|
|
endfunc cortex_a72_core_pwr_dwn
|
|
|
|
/* -------------------------------------------------------
|
|
* The CPU Ops cluster power down function for Cortex-A72.
|
|
* -------------------------------------------------------
|
|
*/
|
|
func cortex_a72_cluster_pwr_dwn
|
|
mov x18, x30
|
|
|
|
/* ---------------------------------------------
|
|
* Turn off caches.
|
|
* ---------------------------------------------
|
|
*/
|
|
bl cortex_a72_disable_dcache
|
|
|
|
/* ---------------------------------------------
|
|
* Disable the L2 prefetches.
|
|
* ---------------------------------------------
|
|
*/
|
|
bl cortex_a72_disable_l2_prefetch
|
|
|
|
/* ---------------------------------------------
|
|
* Disable the load-store hardware prefetcher.
|
|
* ---------------------------------------------
|
|
*/
|
|
bl cortex_a72_disable_hw_prefetcher
|
|
|
|
#if !SKIP_A72_L1_FLUSH_PWR_DWN
|
|
/* ---------------------------------------------
|
|
* Flush L1 caches.
|
|
* ---------------------------------------------
|
|
*/
|
|
mov x0, #DCCISW
|
|
bl dcsw_op_level1
|
|
#endif
|
|
|
|
/* ---------------------------------------------
|
|
* Disable the optional ACP.
|
|
* ---------------------------------------------
|
|
*/
|
|
bl plat_disable_acp
|
|
|
|
/* -------------------------------------------------
|
|
* Flush the L2 caches.
|
|
* -------------------------------------------------
|
|
*/
|
|
mov x0, #DCCISW
|
|
bl dcsw_op_level2
|
|
|
|
/* ---------------------------------------------
|
|
* Come out of intra cluster coherency
|
|
* ---------------------------------------------
|
|
*/
|
|
bl cortex_a72_disable_smp
|
|
|
|
/* ---------------------------------------------
|
|
* Force the debug interfaces to be quiescent
|
|
* ---------------------------------------------
|
|
*/
|
|
mov x30, x18
|
|
b cortex_a72_disable_ext_debug
|
|
endfunc cortex_a72_cluster_pwr_dwn
|
|
|
|
/* ---------------------------------------------
|
|
* This function provides cortex_a72 specific
|
|
* register information for crash reporting.
|
|
* It needs to return with x6 pointing to
|
|
* a list of register names in ascii and
|
|
* x8 - x15 having values of registers to be
|
|
* reported.
|
|
* ---------------------------------------------
|
|
*/
|
|
.section .rodata.cortex_a72_regs, "aS"
|
|
cortex_a72_regs: /* The ascii list of register names to be reported */
|
|
.asciz "cpuectlr_el1", "cpumerrsr_el1", "l2merrsr_el1", ""
|
|
|
|
func cortex_a72_cpu_reg_dump
|
|
adr x6, cortex_a72_regs
|
|
mrs x8, CORTEX_A72_ECTLR_EL1
|
|
mrs x9, CORTEX_A72_MERRSR_EL1
|
|
mrs x10, CORTEX_A72_L2MERRSR_EL1
|
|
ret
|
|
endfunc cortex_a72_cpu_reg_dump
|
|
|
|
declare_cpu_ops_wa cortex_a72, CORTEX_A72_MIDR, \
|
|
cortex_a72_reset_func, \
|
|
check_erratum_cortex_a72_5715, \
|
|
CPU_NO_EXTRA2_FUNC, \
|
|
check_smccc_arch_workaround_3, \
|
|
cortex_a72_core_pwr_dwn, \
|
|
cortex_a72_cluster_pwr_dwn
|