mirror of
https://github.com/ARM-software/arm-trusted-firmware.git
synced 2025-04-16 09:34:18 +00:00

Previously exception handlers in BL3-1, X19-X29 were not saved and restored on every SMC/trap into EL3. Instead these registers were 'saved as needed' as a side effect of the A64 ABI used by the C compiler. That approach failed when world switching but was not visible with the TSP/TSPD code because the TSP is 64-bit, did not clobber these registers when running and did not support pre-emption by normal world interrupts. These scenarios showed that the values in these registers can be passed through a world switch, which broke the normal and trusted world assumptions about these registers being preserved. The Ideal solution saves and restores these registers when a world switch occurs - but that type of implementation is more complex. So this patch always saves and restores these registers on entry and exit of EL3. Fixes ARM-software/tf-issues#141 Change-Id: I9a727167bbc594454e81cf78a97ca899dfb11c27
164 lines
4.5 KiB
ArmAsm
164 lines
4.5 KiB
ArmAsm
/*
|
|
* Copyright (c) 2013-2014, ARM Limited and Contributors. All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions are met:
|
|
*
|
|
* Redistributions of source code must retain the above copyright notice, this
|
|
* list of conditions and the following disclaimer.
|
|
*
|
|
* Redistributions in binary form must reproduce the above copyright notice,
|
|
* this list of conditions and the following disclaimer in the documentation
|
|
* and/or other materials provided with the distribution.
|
|
*
|
|
* Neither the name of ARM nor the names of its contributors may be used
|
|
* to endorse or promote products derived from this software without specific
|
|
* prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
|
* AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE
|
|
* LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
|
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
|
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
|
|
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
|
|
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
|
|
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
|
* POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
|
|
#include <arch.h>
|
|
#include <asm_macros.S>
|
|
#include <cm_macros.S>
|
|
#include <psci.h>
|
|
|
|
.globl psci_aff_on_finish_entry
|
|
.globl psci_aff_suspend_finish_entry
|
|
.globl __psci_cpu_off
|
|
.globl __psci_cpu_suspend
|
|
|
|
/* -----------------------------------------------------
|
|
* This cpu has been physically powered up. Depending
|
|
* upon whether it was resumed from suspend or simply
|
|
* turned on, call the common power on finisher with
|
|
* the handlers (chosen depending upon original state).
|
|
* For ease, the finisher is called with coherent
|
|
* stacks. This allows the cluster/cpu finishers to
|
|
* enter coherency and enable the mmu without running
|
|
* into issues. We switch back to normal stacks once
|
|
* all this is done.
|
|
* -----------------------------------------------------
|
|
*/
|
|
func psci_aff_on_finish_entry
|
|
adr x23, psci_afflvl_on_finishers
|
|
b psci_aff_common_finish_entry
|
|
|
|
psci_aff_suspend_finish_entry:
|
|
adr x23, psci_afflvl_suspend_finishers
|
|
|
|
psci_aff_common_finish_entry:
|
|
adr x22, psci_afflvl_power_on_finish
|
|
|
|
/* ---------------------------------------------
|
|
* Exceptions should not occur at this point.
|
|
* Set VBAR in order to handle and report any
|
|
* that do occur
|
|
* ---------------------------------------------
|
|
*/
|
|
adr x0, early_exceptions
|
|
msr vbar_el3, x0
|
|
isb
|
|
|
|
/* ---------------------------------------------
|
|
* Use SP_EL0 for the C runtime stack.
|
|
* ---------------------------------------------
|
|
*/
|
|
msr spsel, #0
|
|
isb
|
|
|
|
bl read_mpidr
|
|
mov x19, x0
|
|
bl platform_set_coherent_stack
|
|
|
|
/* ---------------------------------------------
|
|
* Call the finishers starting from affinity
|
|
* level 0.
|
|
* ---------------------------------------------
|
|
*/
|
|
mov x0, x19
|
|
bl get_power_on_target_afflvl
|
|
cmp x0, xzr
|
|
b.lt _panic
|
|
mov x3, x23
|
|
mov x2, x0
|
|
mov x0, x19
|
|
mov x1, #MPIDR_AFFLVL0
|
|
blr x22
|
|
|
|
/* --------------------------------------------
|
|
* Give ourselves a stack allocated in Normal
|
|
* -IS-WBWA memory
|
|
* --------------------------------------------
|
|
*/
|
|
mov x0, x19
|
|
bl platform_set_stack
|
|
|
|
b el3_exit
|
|
_panic:
|
|
b _panic
|
|
|
|
/* -----------------------------------------------------
|
|
* The following two stubs give the calling cpu a
|
|
* coherent stack to allow flushing of caches without
|
|
* suffering from stack coherency issues
|
|
* -----------------------------------------------------
|
|
*/
|
|
func __psci_cpu_off
|
|
func_prologue
|
|
sub sp, sp, #0x10
|
|
stp x19, x20, [sp, #0]
|
|
mov x19, sp
|
|
bl read_mpidr
|
|
bl platform_set_coherent_stack
|
|
bl psci_cpu_off
|
|
mov x1, #PSCI_E_SUCCESS
|
|
cmp x0, x1
|
|
b.eq final_wfi
|
|
mov sp, x19
|
|
ldp x19, x20, [sp,#0]
|
|
add sp, sp, #0x10
|
|
func_epilogue
|
|
ret
|
|
|
|
func __psci_cpu_suspend
|
|
func_prologue
|
|
sub sp, sp, #0x20
|
|
stp x19, x20, [sp, #0]
|
|
stp x21, x22, [sp, #0x10]
|
|
mov x19, sp
|
|
mov x20, x0
|
|
mov x21, x1
|
|
mov x22, x2
|
|
bl read_mpidr
|
|
bl platform_set_coherent_stack
|
|
mov x0, x20
|
|
mov x1, x21
|
|
mov x2, x22
|
|
bl psci_cpu_suspend
|
|
mov x1, #PSCI_E_SUCCESS
|
|
cmp x0, x1
|
|
b.eq final_wfi
|
|
mov sp, x19
|
|
ldp x21, x22, [sp,#0x10]
|
|
ldp x19, x20, [sp,#0]
|
|
add sp, sp, #0x20
|
|
func_epilogue
|
|
ret
|
|
|
|
func final_wfi
|
|
dsb sy
|
|
wfi
|
|
wfi_spill:
|
|
b wfi_spill
|
|
|