mirror of
https://github.com/ARM-software/arm-trusted-firmware.git
synced 2025-04-17 10:04:26 +00:00

In order to quantify the overall time spent in the PSCI software implementation, an initial collection of PMF instrumentation points has been added. Instrumentation has been added to the following code paths: - Entry to PSCI SMC handler. The timestamp is captured as early as possible during the runtime exception and stored in memory before entering the PSCI SMC handler. - Exit from PSCI SMC handler. The timestamp is captured after normal return from the PSCI SMC handler or if a low power state was requested it is captured in the bl31 warm boot path before return to normal world. - Entry to low power state. The timestamp is captured before entry to a low power state which implies either standby or power down. As these power states are mutually exclusive, only one timestamp is defined to describe both. It is possible to differentiate between the two power states using the PSCI STAT interface. - Exit from low power state. The timestamp is captured after a standby or power up operation has completed. To calculate the number of cycles spent running code in Trusted Firmware one can perform the following calculation: (exit_psci - enter_psci) - (exit_low_pwr - enter_low_pwr). The resulting number of cycles can be converted to time given the frequency of the counter. Change-Id: Ie3b8f3d16409b6703747093b3a2d5c7429ad0166 Signed-off-by: dp-arm <dimitris.papastamos@arm.com>
488 lines
13 KiB
C
488 lines
13 KiB
C
/*
|
|
* Copyright (c) 2013-2016, ARM Limited and Contributors. All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions are met:
|
|
*
|
|
* Redistributions of source code must retain the above copyright notice, this
|
|
* list of conditions and the following disclaimer.
|
|
*
|
|
* Redistributions in binary form must reproduce the above copyright notice,
|
|
* this list of conditions and the following disclaimer in the documentation
|
|
* and/or other materials provided with the distribution.
|
|
*
|
|
* Neither the name of ARM nor the names of its contributors may be used
|
|
* to endorse or promote products derived from this software without specific
|
|
* prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
|
* AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE
|
|
* LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
|
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
|
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
|
|
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
|
|
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
|
|
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
|
* POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
|
|
#include <arch.h>
|
|
#include <arch_helpers.h>
|
|
#include <assert.h>
|
|
#include <debug.h>
|
|
#include <platform.h>
|
|
#include <pmf.h>
|
|
#include <runtime_instr.h>
|
|
#include <smcc.h>
|
|
#include <string.h>
|
|
#include "psci_private.h"
|
|
|
|
/*******************************************************************************
|
|
* PSCI frontend api for servicing SMCs. Described in the PSCI spec.
|
|
******************************************************************************/
|
|
int psci_cpu_on(u_register_t target_cpu,
|
|
uintptr_t entrypoint,
|
|
u_register_t context_id)
|
|
|
|
{
|
|
int rc;
|
|
entry_point_info_t ep;
|
|
|
|
/* Determine if the cpu exists of not */
|
|
rc = psci_validate_mpidr(target_cpu);
|
|
if (rc != PSCI_E_SUCCESS)
|
|
return PSCI_E_INVALID_PARAMS;
|
|
|
|
/* Validate the entry point and get the entry_point_info */
|
|
rc = psci_validate_entry_point(&ep, entrypoint, context_id);
|
|
if (rc != PSCI_E_SUCCESS)
|
|
return rc;
|
|
|
|
/*
|
|
* To turn this cpu on, specify which power
|
|
* levels need to be turned on
|
|
*/
|
|
return psci_cpu_on_start(target_cpu, &ep);
|
|
}
|
|
|
|
unsigned int psci_version(void)
|
|
{
|
|
return PSCI_MAJOR_VER | PSCI_MINOR_VER;
|
|
}
|
|
|
|
int psci_cpu_suspend(unsigned int power_state,
|
|
uintptr_t entrypoint,
|
|
u_register_t context_id)
|
|
{
|
|
int rc;
|
|
unsigned int target_pwrlvl, is_power_down_state;
|
|
entry_point_info_t ep;
|
|
psci_power_state_t state_info = { {PSCI_LOCAL_STATE_RUN} };
|
|
plat_local_state_t cpu_pd_state;
|
|
|
|
/* Validate the power_state parameter */
|
|
rc = psci_validate_power_state(power_state, &state_info);
|
|
if (rc != PSCI_E_SUCCESS) {
|
|
assert(rc == PSCI_E_INVALID_PARAMS);
|
|
return rc;
|
|
}
|
|
|
|
/*
|
|
* Get the value of the state type bit from the power state parameter.
|
|
*/
|
|
is_power_down_state = psci_get_pstate_type(power_state);
|
|
|
|
/* Sanity check the requested suspend levels */
|
|
assert(psci_validate_suspend_req(&state_info, is_power_down_state)
|
|
== PSCI_E_SUCCESS);
|
|
|
|
target_pwrlvl = psci_find_target_suspend_lvl(&state_info);
|
|
if (target_pwrlvl == PSCI_INVALID_PWR_LVL) {
|
|
ERROR("Invalid target power level for suspend operation\n");
|
|
panic();
|
|
}
|
|
|
|
/* Fast path for CPU standby.*/
|
|
if (is_cpu_standby_req(is_power_down_state, target_pwrlvl)) {
|
|
if (!psci_plat_pm_ops->cpu_standby)
|
|
return PSCI_E_INVALID_PARAMS;
|
|
|
|
/*
|
|
* Set the state of the CPU power domain to the platform
|
|
* specific retention state and enter the standby state.
|
|
*/
|
|
cpu_pd_state = state_info.pwr_domain_state[PSCI_CPU_PWR_LVL];
|
|
psci_set_cpu_local_state(cpu_pd_state);
|
|
|
|
#if ENABLE_PSCI_STAT
|
|
/*
|
|
* Capture time-stamp before CPU standby
|
|
* No cache maintenance is needed as caches
|
|
* are ON through out the CPU standby operation.
|
|
*/
|
|
PMF_CAPTURE_TIMESTAMP(psci_svc, PSCI_STAT_ID_ENTER_LOW_PWR,
|
|
PMF_NO_CACHE_MAINT);
|
|
#endif
|
|
|
|
#if ENABLE_RUNTIME_INSTRUMENTATION
|
|
PMF_CAPTURE_TIMESTAMP(rt_instr_svc,
|
|
RT_INSTR_ENTER_HW_LOW_PWR,
|
|
PMF_NO_CACHE_MAINT);
|
|
#endif
|
|
|
|
psci_plat_pm_ops->cpu_standby(cpu_pd_state);
|
|
|
|
/* Upon exit from standby, set the state back to RUN. */
|
|
psci_set_cpu_local_state(PSCI_LOCAL_STATE_RUN);
|
|
|
|
#if ENABLE_RUNTIME_INSTRUMENTATION
|
|
PMF_CAPTURE_TIMESTAMP(rt_instr_svc,
|
|
RT_INSTR_EXIT_HW_LOW_PWR,
|
|
PMF_NO_CACHE_MAINT);
|
|
#endif
|
|
|
|
#if ENABLE_PSCI_STAT
|
|
/* Capture time-stamp after CPU standby */
|
|
PMF_CAPTURE_TIMESTAMP(psci_svc, PSCI_STAT_ID_EXIT_LOW_PWR,
|
|
PMF_NO_CACHE_MAINT);
|
|
|
|
/* Update PSCI stats */
|
|
psci_stats_update_pwr_up(PSCI_CPU_PWR_LVL, &state_info,
|
|
PMF_NO_CACHE_MAINT);
|
|
#endif
|
|
|
|
return PSCI_E_SUCCESS;
|
|
}
|
|
|
|
/*
|
|
* If a power down state has been requested, we need to verify entry
|
|
* point and program entry information.
|
|
*/
|
|
if (is_power_down_state) {
|
|
rc = psci_validate_entry_point(&ep, entrypoint, context_id);
|
|
if (rc != PSCI_E_SUCCESS)
|
|
return rc;
|
|
}
|
|
|
|
/*
|
|
* Do what is needed to enter the power down state. Upon success,
|
|
* enter the final wfi which will power down this CPU. This function
|
|
* might return if the power down was abandoned for any reason, e.g.
|
|
* arrival of an interrupt
|
|
*/
|
|
psci_cpu_suspend_start(&ep,
|
|
target_pwrlvl,
|
|
&state_info,
|
|
is_power_down_state);
|
|
|
|
return PSCI_E_SUCCESS;
|
|
}
|
|
|
|
|
|
int psci_system_suspend(uintptr_t entrypoint, u_register_t context_id)
|
|
{
|
|
int rc;
|
|
psci_power_state_t state_info;
|
|
entry_point_info_t ep;
|
|
|
|
/* Check if the current CPU is the last ON CPU in the system */
|
|
if (!psci_is_last_on_cpu())
|
|
return PSCI_E_DENIED;
|
|
|
|
/* Validate the entry point and get the entry_point_info */
|
|
rc = psci_validate_entry_point(&ep, entrypoint, context_id);
|
|
if (rc != PSCI_E_SUCCESS)
|
|
return rc;
|
|
|
|
/* Query the psci_power_state for system suspend */
|
|
psci_query_sys_suspend_pwrstate(&state_info);
|
|
|
|
/* Ensure that the psci_power_state makes sense */
|
|
assert(psci_find_target_suspend_lvl(&state_info) == PLAT_MAX_PWR_LVL);
|
|
assert(psci_validate_suspend_req(&state_info, PSTATE_TYPE_POWERDOWN)
|
|
== PSCI_E_SUCCESS);
|
|
assert(is_local_state_off(state_info.pwr_domain_state[PLAT_MAX_PWR_LVL]));
|
|
|
|
/*
|
|
* Do what is needed to enter the system suspend state. This function
|
|
* might return if the power down was abandoned for any reason, e.g.
|
|
* arrival of an interrupt
|
|
*/
|
|
psci_cpu_suspend_start(&ep,
|
|
PLAT_MAX_PWR_LVL,
|
|
&state_info,
|
|
PSTATE_TYPE_POWERDOWN);
|
|
|
|
return PSCI_E_SUCCESS;
|
|
}
|
|
|
|
int psci_cpu_off(void)
|
|
{
|
|
int rc;
|
|
unsigned int target_pwrlvl = PLAT_MAX_PWR_LVL;
|
|
|
|
/*
|
|
* Do what is needed to power off this CPU and possible higher power
|
|
* levels if it able to do so. Upon success, enter the final wfi
|
|
* which will power down this CPU.
|
|
*/
|
|
rc = psci_do_cpu_off(target_pwrlvl);
|
|
|
|
/*
|
|
* The only error cpu_off can return is E_DENIED. So check if that's
|
|
* indeed the case.
|
|
*/
|
|
assert(rc == PSCI_E_DENIED);
|
|
|
|
return rc;
|
|
}
|
|
|
|
int psci_affinity_info(u_register_t target_affinity,
|
|
unsigned int lowest_affinity_level)
|
|
{
|
|
unsigned int target_idx;
|
|
|
|
/* We dont support level higher than PSCI_CPU_PWR_LVL */
|
|
if (lowest_affinity_level > PSCI_CPU_PWR_LVL)
|
|
return PSCI_E_INVALID_PARAMS;
|
|
|
|
/* Calculate the cpu index of the target */
|
|
target_idx = plat_core_pos_by_mpidr(target_affinity);
|
|
if (target_idx == -1)
|
|
return PSCI_E_INVALID_PARAMS;
|
|
|
|
return psci_get_aff_info_state_by_idx(target_idx);
|
|
}
|
|
|
|
int psci_migrate(u_register_t target_cpu)
|
|
{
|
|
int rc;
|
|
u_register_t resident_cpu_mpidr;
|
|
|
|
rc = psci_spd_migrate_info(&resident_cpu_mpidr);
|
|
if (rc != PSCI_TOS_UP_MIG_CAP)
|
|
return (rc == PSCI_TOS_NOT_UP_MIG_CAP) ?
|
|
PSCI_E_DENIED : PSCI_E_NOT_SUPPORTED;
|
|
|
|
/*
|
|
* Migrate should only be invoked on the CPU where
|
|
* the Secure OS is resident.
|
|
*/
|
|
if (resident_cpu_mpidr != read_mpidr_el1())
|
|
return PSCI_E_NOT_PRESENT;
|
|
|
|
/* Check the validity of the specified target cpu */
|
|
rc = psci_validate_mpidr(target_cpu);
|
|
if (rc != PSCI_E_SUCCESS)
|
|
return PSCI_E_INVALID_PARAMS;
|
|
|
|
assert(psci_spd_pm && psci_spd_pm->svc_migrate);
|
|
|
|
rc = psci_spd_pm->svc_migrate(read_mpidr_el1(), target_cpu);
|
|
assert(rc == PSCI_E_SUCCESS || rc == PSCI_E_INTERN_FAIL);
|
|
|
|
return rc;
|
|
}
|
|
|
|
int psci_migrate_info_type(void)
|
|
{
|
|
u_register_t resident_cpu_mpidr;
|
|
|
|
return psci_spd_migrate_info(&resident_cpu_mpidr);
|
|
}
|
|
|
|
long psci_migrate_info_up_cpu(void)
|
|
{
|
|
u_register_t resident_cpu_mpidr;
|
|
int rc;
|
|
|
|
/*
|
|
* Return value of this depends upon what
|
|
* psci_spd_migrate_info() returns.
|
|
*/
|
|
rc = psci_spd_migrate_info(&resident_cpu_mpidr);
|
|
if (rc != PSCI_TOS_NOT_UP_MIG_CAP && rc != PSCI_TOS_UP_MIG_CAP)
|
|
return PSCI_E_INVALID_PARAMS;
|
|
|
|
return resident_cpu_mpidr;
|
|
}
|
|
|
|
int psci_node_hw_state(u_register_t target_cpu,
|
|
unsigned int power_level)
|
|
{
|
|
int rc;
|
|
|
|
/* Validate target_cpu */
|
|
rc = psci_validate_mpidr(target_cpu);
|
|
if (rc != PSCI_E_SUCCESS)
|
|
return PSCI_E_INVALID_PARAMS;
|
|
|
|
/* Validate power_level against PLAT_MAX_PWR_LVL */
|
|
if (power_level > PLAT_MAX_PWR_LVL)
|
|
return PSCI_E_INVALID_PARAMS;
|
|
|
|
/*
|
|
* Dispatch this call to platform to query power controller, and pass on
|
|
* to the caller what it returns
|
|
*/
|
|
assert(psci_plat_pm_ops->get_node_hw_state);
|
|
rc = psci_plat_pm_ops->get_node_hw_state(target_cpu, power_level);
|
|
assert((rc >= HW_ON && rc <= HW_STANDBY) || rc == PSCI_E_NOT_SUPPORTED
|
|
|| rc == PSCI_E_INVALID_PARAMS);
|
|
return rc;
|
|
}
|
|
|
|
int psci_features(unsigned int psci_fid)
|
|
{
|
|
unsigned int local_caps = psci_caps;
|
|
|
|
/* Check if it is a 64 bit function */
|
|
if (((psci_fid >> FUNCID_CC_SHIFT) & FUNCID_CC_MASK) == SMC_64)
|
|
local_caps &= PSCI_CAP_64BIT_MASK;
|
|
|
|
/* Check for invalid fid */
|
|
if (!(is_std_svc_call(psci_fid) && is_valid_fast_smc(psci_fid)
|
|
&& is_psci_fid(psci_fid)))
|
|
return PSCI_E_NOT_SUPPORTED;
|
|
|
|
|
|
/* Check if the psci fid is supported or not */
|
|
if (!(local_caps & define_psci_cap(psci_fid)))
|
|
return PSCI_E_NOT_SUPPORTED;
|
|
|
|
/* Format the feature flags */
|
|
if (psci_fid == PSCI_CPU_SUSPEND_AARCH32 ||
|
|
psci_fid == PSCI_CPU_SUSPEND_AARCH64) {
|
|
/*
|
|
* The trusted firmware does not support OS Initiated Mode.
|
|
*/
|
|
return (FF_PSTATE << FF_PSTATE_SHIFT) |
|
|
((!FF_SUPPORTS_OS_INIT_MODE) << FF_MODE_SUPPORT_SHIFT);
|
|
}
|
|
|
|
/* Return 0 for all other fid's */
|
|
return PSCI_E_SUCCESS;
|
|
}
|
|
|
|
/*******************************************************************************
|
|
* PSCI top level handler for servicing SMCs.
|
|
******************************************************************************/
|
|
u_register_t psci_smc_handler(uint32_t smc_fid,
|
|
u_register_t x1,
|
|
u_register_t x2,
|
|
u_register_t x3,
|
|
u_register_t x4,
|
|
void *cookie,
|
|
void *handle,
|
|
u_register_t flags)
|
|
{
|
|
if (is_caller_secure(flags))
|
|
return SMC_UNK;
|
|
|
|
/* Check the fid against the capabilities */
|
|
if (!(psci_caps & define_psci_cap(smc_fid)))
|
|
return SMC_UNK;
|
|
|
|
if (((smc_fid >> FUNCID_CC_SHIFT) & FUNCID_CC_MASK) == SMC_32) {
|
|
/* 32-bit PSCI function, clear top parameter bits */
|
|
|
|
x1 = (uint32_t)x1;
|
|
x2 = (uint32_t)x2;
|
|
x3 = (uint32_t)x3;
|
|
|
|
switch (smc_fid) {
|
|
case PSCI_VERSION:
|
|
return psci_version();
|
|
|
|
case PSCI_CPU_OFF:
|
|
return psci_cpu_off();
|
|
|
|
case PSCI_CPU_SUSPEND_AARCH32:
|
|
return psci_cpu_suspend(x1, x2, x3);
|
|
|
|
case PSCI_CPU_ON_AARCH32:
|
|
return psci_cpu_on(x1, x2, x3);
|
|
|
|
case PSCI_AFFINITY_INFO_AARCH32:
|
|
return psci_affinity_info(x1, x2);
|
|
|
|
case PSCI_MIG_AARCH32:
|
|
return psci_migrate(x1);
|
|
|
|
case PSCI_MIG_INFO_TYPE:
|
|
return psci_migrate_info_type();
|
|
|
|
case PSCI_MIG_INFO_UP_CPU_AARCH32:
|
|
return psci_migrate_info_up_cpu();
|
|
|
|
case PSCI_NODE_HW_STATE_AARCH32:
|
|
return psci_node_hw_state(x1, x2);
|
|
|
|
case PSCI_SYSTEM_SUSPEND_AARCH32:
|
|
return psci_system_suspend(x1, x2);
|
|
|
|
case PSCI_SYSTEM_OFF:
|
|
psci_system_off();
|
|
/* We should never return from psci_system_off() */
|
|
|
|
case PSCI_SYSTEM_RESET:
|
|
psci_system_reset();
|
|
/* We should never return from psci_system_reset() */
|
|
|
|
case PSCI_FEATURES:
|
|
return psci_features(x1);
|
|
|
|
#if ENABLE_PSCI_STAT
|
|
case PSCI_STAT_RESIDENCY_AARCH32:
|
|
return psci_stat_residency(x1, x2);
|
|
|
|
case PSCI_STAT_COUNT_AARCH32:
|
|
return psci_stat_count(x1, x2);
|
|
#endif
|
|
|
|
default:
|
|
break;
|
|
}
|
|
} else {
|
|
/* 64-bit PSCI function */
|
|
|
|
switch (smc_fid) {
|
|
case PSCI_CPU_SUSPEND_AARCH64:
|
|
return psci_cpu_suspend(x1, x2, x3);
|
|
|
|
case PSCI_CPU_ON_AARCH64:
|
|
return psci_cpu_on(x1, x2, x3);
|
|
|
|
case PSCI_AFFINITY_INFO_AARCH64:
|
|
return psci_affinity_info(x1, x2);
|
|
|
|
case PSCI_MIG_AARCH64:
|
|
return psci_migrate(x1);
|
|
|
|
case PSCI_MIG_INFO_UP_CPU_AARCH64:
|
|
return psci_migrate_info_up_cpu();
|
|
|
|
case PSCI_NODE_HW_STATE_AARCH64:
|
|
return psci_node_hw_state(x1, x2);
|
|
|
|
case PSCI_SYSTEM_SUSPEND_AARCH64:
|
|
return psci_system_suspend(x1, x2);
|
|
|
|
#if ENABLE_PSCI_STAT
|
|
case PSCI_STAT_RESIDENCY_AARCH64:
|
|
return psci_stat_residency(x1, x2);
|
|
|
|
case PSCI_STAT_COUNT_AARCH64:
|
|
return psci_stat_count(x1, x2);
|
|
#endif
|
|
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
|
|
WARN("Unimplemented PSCI Call: 0x%x \n", smc_fid);
|
|
return SMC_UNK;
|
|
}
|