mirror of
https://github.com/ARM-software/arm-trusted-firmware.git
synced 2025-04-18 10:34:19 +00:00

This patch uses the reworked exception handling support to handle runtime service requests through SMCs following the SMC calling convention. This is a giant commit since all the changes are inter-related. It does the following: 1. Replace the old exception handling mechanism with the new one 2. Enforce that SP_EL0 is used C runtime stacks. 3. Ensures that the cold and warm boot paths use the 'cpu_context' structure to program an ERET into the next lower EL. 4. Ensures that SP_EL3 always points to the next 'cpu_context' structure prior to an ERET into the next lower EL 5. Introduces a PSCI SMC handler which completes the use of PSCI as a runtime service Change-Id: I661797f834c0803d2c674d20f504df1b04c2b852 Co-authored-by: Achin Gupta <achin.gupta@arm.com>
216 lines
7.9 KiB
C
216 lines
7.9 KiB
C
/*
|
|
* Copyright (c) 2013-2014, ARM Limited and Contributors. All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions are met:
|
|
*
|
|
* Redistributions of source code must retain the above copyright notice, this
|
|
* list of conditions and the following disclaimer.
|
|
*
|
|
* Redistributions in binary form must reproduce the above copyright notice,
|
|
* this list of conditions and the following disclaimer in the documentation
|
|
* and/or other materials provided with the distribution.
|
|
*
|
|
* Neither the name of ARM nor the names of its contributors may be used
|
|
* to endorse or promote products derived from this software without specific
|
|
* prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
|
* AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE
|
|
* LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
|
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
|
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
|
|
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
|
|
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
|
|
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
|
* POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
|
|
#include <stdio.h>
|
|
#include <errno.h>
|
|
#include <string.h>
|
|
#include <assert.h>
|
|
#include <arch_helpers.h>
|
|
#include <platform.h>
|
|
#include <bl_common.h>
|
|
#include <runtime_svc.h>
|
|
#include <context_mgmt.h>
|
|
|
|
/*******************************************************************************
|
|
* Data structure which holds the pointers to non-secure and secure security
|
|
* state contexts for each cpu. It is aligned to the cache line boundary to
|
|
* allow efficient concurrent manipulation of these pointers on different cpus
|
|
******************************************************************************/
|
|
typedef struct {
|
|
void *ptr[2];
|
|
} __aligned (CACHE_WRITEBACK_GRANULE) context_info;
|
|
|
|
static context_info cm_context_info[PLATFORM_CORE_COUNT];
|
|
|
|
/*******************************************************************************
|
|
* Context management library initialisation routine. This library is used by
|
|
* runtime services to share pointers to 'cpu_context' structures for the secure
|
|
* and non-secure states. Management of the structures and their associated
|
|
* memory is not done by the context management library e.g. the PSCI service
|
|
* manages the cpu context used for entry from and exit to the non-secure state.
|
|
* The Secure payload dispatcher service manages the context(s) corresponding to
|
|
* the secure state. It also uses this library to get access to the non-secure
|
|
* state cpu context pointers.
|
|
* Lastly, this library provides the api to make SP_EL3 point to the cpu context
|
|
* which will used for programming an entry into a lower EL. The same context
|
|
* will used to save state upon exception entry from that EL.
|
|
******************************************************************************/
|
|
void cm_init()
|
|
{
|
|
/*
|
|
* The context management library has only global data to intialize, but
|
|
* that will be done when the BSS is zeroed out
|
|
*/
|
|
}
|
|
|
|
/*******************************************************************************
|
|
* This function returns a pointer to the most recent 'cpu_context' structure
|
|
* that was set as the context for the specified security state. NULL is
|
|
* returned if no such structure has been specified.
|
|
******************************************************************************/
|
|
void *cm_get_context(uint64_t mpidr, uint32_t security_state)
|
|
{
|
|
uint32_t linear_id = platform_get_core_pos(mpidr);
|
|
|
|
assert(security_state <= NON_SECURE);
|
|
|
|
return cm_context_info[linear_id].ptr[security_state];
|
|
}
|
|
|
|
/*******************************************************************************
|
|
* This function sets the pointer to the current 'cpu_context' structure for the
|
|
* specified security state.
|
|
******************************************************************************/
|
|
void cm_set_context(uint64_t mpidr, void *context, uint32_t security_state)
|
|
{
|
|
uint32_t linear_id = platform_get_core_pos(mpidr);
|
|
|
|
assert(security_state <= NON_SECURE);
|
|
|
|
cm_context_info[linear_id].ptr[security_state] = context;
|
|
}
|
|
|
|
/*******************************************************************************
|
|
* The next four functions are used by runtime services to save and restore EL3
|
|
* and EL1 contexts on the 'cpu_context' structure for the specified security
|
|
* state.
|
|
******************************************************************************/
|
|
void cm_el3_sysregs_context_save(uint32_t security_state)
|
|
{
|
|
cpu_context *ctx;
|
|
|
|
ctx = cm_get_context(read_mpidr(), security_state);
|
|
assert(ctx);
|
|
|
|
el3_sysregs_context_save(get_el3state_ctx(ctx));
|
|
}
|
|
|
|
void cm_el3_sysregs_context_restore(uint32_t security_state)
|
|
{
|
|
cpu_context *ctx;
|
|
|
|
ctx = cm_get_context(read_mpidr(), security_state);
|
|
assert(ctx);
|
|
|
|
el3_sysregs_context_restore(get_el3state_ctx(ctx));
|
|
}
|
|
|
|
void cm_el1_sysregs_context_save(uint32_t security_state)
|
|
{
|
|
cpu_context *ctx;
|
|
|
|
ctx = cm_get_context(read_mpidr(), security_state);
|
|
assert(ctx);
|
|
|
|
el1_sysregs_context_save(get_sysregs_ctx(ctx));
|
|
}
|
|
|
|
void cm_el1_sysregs_context_restore(uint32_t security_state)
|
|
{
|
|
cpu_context *ctx;
|
|
|
|
ctx = cm_get_context(read_mpidr(), security_state);
|
|
assert(ctx);
|
|
|
|
el1_sysregs_context_restore(get_sysregs_ctx(ctx));
|
|
}
|
|
|
|
/*******************************************************************************
|
|
* This function function populates 'cpu_context' pertaining to the given
|
|
* security state with the entrypoint, SPSR and SCR values so that an ERET from
|
|
* this securit state correctly restores corresponding values to drop the CPU to
|
|
* the next exception level
|
|
******************************************************************************/
|
|
void cm_set_el3_eret_context(uint32_t security_state, uint64_t entrypoint,
|
|
uint32_t spsr, uint32_t scr)
|
|
{
|
|
cpu_context *ctx;
|
|
el3_state *state;
|
|
|
|
ctx = cm_get_context(read_mpidr(), security_state);
|
|
assert(ctx);
|
|
|
|
/* Populate EL3 state so that we've the right context before doing ERET */
|
|
state = get_el3state_ctx(ctx);
|
|
write_ctx_reg(state, CTX_SPSR_EL3, spsr);
|
|
write_ctx_reg(state, CTX_ELR_EL3, entrypoint);
|
|
write_ctx_reg(state, CTX_SCR_EL3, scr);
|
|
}
|
|
|
|
/*******************************************************************************
|
|
* This function is used to program the context that's used for exception
|
|
* return. This initializes the SP_EL3 to a pointer to a 'cpu_context' set for
|
|
* the required security state
|
|
******************************************************************************/
|
|
void cm_set_next_eret_context(uint32_t security_state)
|
|
{
|
|
cpu_context *ctx;
|
|
#if DEBUG
|
|
uint64_t sp_mode;
|
|
#endif
|
|
|
|
ctx = cm_get_context(read_mpidr(), security_state);
|
|
assert(ctx);
|
|
|
|
#if DEBUG
|
|
/*
|
|
* Check that this function is called with SP_EL0 as the stack
|
|
* pointer
|
|
*/
|
|
__asm__ volatile("mrs %0, SPSel\n"
|
|
: "=r" (sp_mode));
|
|
|
|
assert(sp_mode == MODE_SP_EL0);
|
|
#endif
|
|
|
|
__asm__ volatile("msr spsel, #1\n"
|
|
"mov sp, %0\n"
|
|
"msr spsel, #0\n"
|
|
: : "r" (ctx));
|
|
}
|
|
|
|
/*******************************************************************************
|
|
* This function is used to program exception stack in the 'cpu_context'
|
|
* structure. This is the initial stack used for taking and handling exceptions
|
|
* at EL3. This stack is expected to be initialized once by each security state
|
|
******************************************************************************/
|
|
void cm_init_exception_stack(uint64_t mpidr, uint32_t security_state)
|
|
{
|
|
cpu_context *ctx;
|
|
el3_state *state;
|
|
|
|
ctx = cm_get_context(mpidr, security_state);
|
|
assert(ctx);
|
|
|
|
/* Set exception stack in the context */
|
|
state = get_el3state_ctx(ctx);
|
|
|
|
write_ctx_reg(state, CTX_EXCEPTION_SP, get_exception_stack(mpidr));
|
|
}
|