mirror of
https://github.com/ARM-software/arm-trusted-firmware.git
synced 2025-04-15 09:04:17 +00:00

The build option `ENABLE_ASSERTIONS` should be used instead. That way both C and ASM assertions can be enabled or disabled together. All occurrences of `ASM_ASSERTION` in common code and ARM platforms have been replaced by `ENABLE_ASSERTIONS`. ASM_ASSERTION has been removed from the user guide. Change-Id: I51f1991f11b9b7ff83e787c9a3270c274748ec6f Signed-off-by: Antonio Nino Diaz <antonio.ninodiaz@arm.com>
147 lines
4.5 KiB
ArmAsm
147 lines
4.5 KiB
ArmAsm
/*
|
|
* Copyright (c) 2016-2017, ARM Limited and Contributors. All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions are met:
|
|
*
|
|
* Redistributions of source code must retain the above copyright notice, this
|
|
* list of conditions and the following disclaimer.
|
|
*
|
|
* Redistributions in binary form must reproduce the above copyright notice,
|
|
* this list of conditions and the following disclaimer in the documentation
|
|
* and/or other materials provided with the distribution.
|
|
*
|
|
* Neither the name of ARM nor the names of its contributors may be used
|
|
* to endorse or promote products derived from this software without specific
|
|
* prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
|
* AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE
|
|
* LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
|
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
|
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
|
|
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
|
|
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
|
|
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
|
* POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
|
|
#include <arch.h>
|
|
#include <asm_macros.S>
|
|
#include <assert_macros.S>
|
|
#include <cortex_a32.h>
|
|
#include <cpu_macros.S>
|
|
|
|
|
|
/* ---------------------------------------------
|
|
* Disable intra-cluster coherency
|
|
* Clobbers: r0-r1
|
|
* ---------------------------------------------
|
|
*/
|
|
func cortex_a32_disable_smp
|
|
ldcopr16 r0, r1, CORTEX_A32_CPUECTLR_EL1
|
|
bic r0, r0, #CORTEX_A32_CPUECTLR_SMPEN_BIT
|
|
stcopr16 r0, r1, CORTEX_A32_CPUECTLR_EL1
|
|
isb
|
|
dsb sy
|
|
bx lr
|
|
endfunc cortex_a32_disable_smp
|
|
|
|
/* -------------------------------------------------
|
|
* The CPU Ops reset function for Cortex-A32.
|
|
* Clobbers: r0-r1
|
|
* -------------------------------------------------
|
|
*/
|
|
func cortex_a32_reset_func
|
|
/* ---------------------------------------------
|
|
* Enable the SMP bit.
|
|
* ---------------------------------------------
|
|
*/
|
|
ldcopr16 r0, r1, CORTEX_A32_CPUECTLR_EL1
|
|
orr r0, r0, #CORTEX_A32_CPUECTLR_SMPEN_BIT
|
|
stcopr16 r0, r1, CORTEX_A32_CPUECTLR_EL1
|
|
isb
|
|
bx lr
|
|
endfunc cortex_a32_reset_func
|
|
|
|
/* ----------------------------------------------------
|
|
* The CPU Ops core power down function for Cortex-A32.
|
|
* Clobbers: r0-r3
|
|
* ----------------------------------------------------
|
|
*/
|
|
func cortex_a32_core_pwr_dwn
|
|
/* r12 is pushed to meet the 8 byte stack alignment requirement */
|
|
push {r12, lr}
|
|
|
|
/* Assert if cache is enabled */
|
|
#if ENABLE_ASSERTIONS
|
|
ldcopr r0, SCTLR
|
|
tst r0, #SCTLR_C_BIT
|
|
ASM_ASSERT(eq)
|
|
#endif
|
|
|
|
/* ---------------------------------------------
|
|
* Flush L1 caches.
|
|
* ---------------------------------------------
|
|
*/
|
|
mov r0, #DC_OP_CISW
|
|
bl dcsw_op_level1
|
|
|
|
/* ---------------------------------------------
|
|
* Come out of intra cluster coherency
|
|
* ---------------------------------------------
|
|
*/
|
|
pop {r12, lr}
|
|
b cortex_a32_disable_smp
|
|
endfunc cortex_a32_core_pwr_dwn
|
|
|
|
/* -------------------------------------------------------
|
|
* The CPU Ops cluster power down function for Cortex-A32.
|
|
* Clobbers: r0-r3
|
|
* -------------------------------------------------------
|
|
*/
|
|
func cortex_a32_cluster_pwr_dwn
|
|
/* r12 is pushed to meet the 8 byte stack alignment requirement */
|
|
push {r12, lr}
|
|
|
|
/* Assert if cache is enabled */
|
|
#if ENABLE_ASSERTIONS
|
|
ldcopr r0, SCTLR
|
|
tst r0, #SCTLR_C_BIT
|
|
ASM_ASSERT(eq)
|
|
#endif
|
|
|
|
/* ---------------------------------------------
|
|
* Flush L1 cache.
|
|
* ---------------------------------------------
|
|
*/
|
|
mov r0, #DC_OP_CISW
|
|
bl dcsw_op_level1
|
|
|
|
/* ---------------------------------------------
|
|
* Disable the optional ACP.
|
|
* ---------------------------------------------
|
|
*/
|
|
bl plat_disable_acp
|
|
|
|
/* ---------------------------------------------
|
|
* Flush L2 cache.
|
|
* ---------------------------------------------
|
|
*/
|
|
mov r0, #DC_OP_CISW
|
|
bl dcsw_op_level2
|
|
|
|
/* ---------------------------------------------
|
|
* Come out of intra cluster coherency
|
|
* ---------------------------------------------
|
|
*/
|
|
pop {r12, lr}
|
|
b cortex_a32_disable_smp
|
|
endfunc cortex_a32_cluster_pwr_dwn
|
|
|
|
declare_cpu_ops cortex_a32, CORTEX_A32_MIDR, \
|
|
cortex_a32_reset_func, \
|
|
cortex_a32_core_pwr_dwn, \
|
|
cortex_a32_cluster_pwr_dwn
|