mirror of
https://github.com/ARM-software/arm-trusted-firmware.git
synced 2025-04-15 00:54:22 +00:00

This patch implements the pwr_domain_off_early handler for Tegra platforms. Powering off the boot core on some Tegra platforms is not allowed and the SOC specific helper functions for Tegra194, Tegra210 and Tegra186 implement this restriction. Signed-off-by: Varun Wadekar <vwadekar@nvidia.com> Change-Id: I9d06e0eee12314764adb0422e023a5bec6ed9c1e
515 lines
15 KiB
C
515 lines
15 KiB
C
/*
|
|
* Copyright (c) 2019-2020, NVIDIA CORPORATION. All rights reserved.
|
|
*
|
|
* SPDX-License-Identifier: BSD-3-Clause
|
|
*/
|
|
|
|
#include <arch.h>
|
|
#include <assert.h>
|
|
#include <stdbool.h>
|
|
#include <string.h>
|
|
|
|
#include <arch_helpers.h>
|
|
#include <bpmp_ipc.h>
|
|
#include <common/bl_common.h>
|
|
#include <common/debug.h>
|
|
#include <context.h>
|
|
#include <drivers/delay_timer.h>
|
|
#include <denver.h>
|
|
#include <lib/el3_runtime/context_mgmt.h>
|
|
#include <lib/psci/psci.h>
|
|
#include <mce.h>
|
|
#include <mce_private.h>
|
|
#include <memctrl_v2.h>
|
|
#include <plat/common/platform.h>
|
|
#include <se.h>
|
|
#include <smmu.h>
|
|
#include <t194_nvg.h>
|
|
#include <tegra194_private.h>
|
|
#include <tegra_platform.h>
|
|
#include <tegra_private.h>
|
|
|
|
extern uint32_t __tegra194_cpu_reset_handler_data,
|
|
__tegra194_cpu_reset_handler_end;
|
|
|
|
/* TZDRAM offset for saving SMMU context */
|
|
#define TEGRA194_SMMU_CTX_OFFSET 16U
|
|
|
|
/* state id mask */
|
|
#define TEGRA194_STATE_ID_MASK 0xFU
|
|
/* constants to get power state's wake time */
|
|
#define TEGRA194_WAKE_TIME_MASK 0x0FFFFFF0U
|
|
#define TEGRA194_WAKE_TIME_SHIFT 4U
|
|
/* default core wake mask for CPU_SUSPEND */
|
|
#define TEGRA194_CORE_WAKE_MASK 0x180cU
|
|
|
|
static struct t19x_psci_percpu_data {
|
|
uint32_t wake_time;
|
|
} __aligned(CACHE_WRITEBACK_GRANULE) t19x_percpu_data[PLATFORM_CORE_COUNT];
|
|
|
|
int32_t tegra_soc_validate_power_state(uint32_t power_state,
|
|
psci_power_state_t *req_state)
|
|
{
|
|
uint8_t state_id = (uint8_t)psci_get_pstate_id(power_state) &
|
|
TEGRA194_STATE_ID_MASK;
|
|
uint32_t cpu = plat_my_core_pos();
|
|
int32_t ret = PSCI_E_SUCCESS;
|
|
|
|
/* save the core wake time (in TSC ticks)*/
|
|
t19x_percpu_data[cpu].wake_time = (power_state & TEGRA194_WAKE_TIME_MASK)
|
|
<< TEGRA194_WAKE_TIME_SHIFT;
|
|
|
|
/*
|
|
* Clean t19x_percpu_data[cpu] to DRAM. This needs to be done to ensure
|
|
* that the correct value is read in tegra_soc_pwr_domain_suspend(),
|
|
* which is called with caches disabled. It is possible to read a stale
|
|
* value from DRAM in that function, because the L2 cache is not flushed
|
|
* unless the cluster is entering CC6/CC7.
|
|
*/
|
|
clean_dcache_range((uint64_t)&t19x_percpu_data[cpu],
|
|
sizeof(t19x_percpu_data[cpu]));
|
|
|
|
/* Sanity check the requested state id */
|
|
switch (state_id) {
|
|
case PSTATE_ID_CORE_IDLE:
|
|
|
|
if (psci_get_pstate_type(power_state) != PSTATE_TYPE_STANDBY) {
|
|
ret = PSCI_E_INVALID_PARAMS;
|
|
break;
|
|
}
|
|
|
|
/* Core idle request */
|
|
req_state->pwr_domain_state[MPIDR_AFFLVL0] = PLAT_MAX_RET_STATE;
|
|
req_state->pwr_domain_state[MPIDR_AFFLVL1] = PSCI_LOCAL_STATE_RUN;
|
|
break;
|
|
|
|
default:
|
|
ERROR("%s: unsupported state id (%d)\n", __func__, state_id);
|
|
ret = PSCI_E_INVALID_PARAMS;
|
|
break;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
int32_t tegra_soc_cpu_standby(plat_local_state_t cpu_state)
|
|
{
|
|
uint32_t cpu = plat_my_core_pos();
|
|
mce_cstate_info_t cstate_info = { 0 };
|
|
|
|
/* Program default wake mask */
|
|
cstate_info.wake_mask = TEGRA194_CORE_WAKE_MASK;
|
|
cstate_info.update_wake_mask = 1;
|
|
mce_update_cstate_info(&cstate_info);
|
|
|
|
/* Enter CPU idle */
|
|
(void)mce_command_handler((uint64_t)MCE_CMD_ENTER_CSTATE,
|
|
(uint64_t)TEGRA_NVG_CORE_C6,
|
|
t19x_percpu_data[cpu].wake_time,
|
|
0U);
|
|
|
|
return PSCI_E_SUCCESS;
|
|
}
|
|
|
|
int32_t tegra_soc_pwr_domain_suspend(const psci_power_state_t *target_state)
|
|
{
|
|
const plat_local_state_t *pwr_domain_state;
|
|
uint8_t stateid_afflvl2;
|
|
plat_params_from_bl2_t *params_from_bl2 = bl31_get_plat_params();
|
|
uint64_t mc_ctx_base;
|
|
uint32_t val;
|
|
mce_cstate_info_t sc7_cstate_info = {
|
|
.cluster = (uint32_t)TEGRA_NVG_CLUSTER_CC6,
|
|
.ccplex = (uint32_t)TEGRA_NVG_CG_CG7,
|
|
.system = (uint32_t)TEGRA_NVG_SYSTEM_SC7,
|
|
.system_state_force = 1U,
|
|
.update_wake_mask = 1U,
|
|
};
|
|
int32_t ret = 0;
|
|
|
|
/* get the state ID */
|
|
pwr_domain_state = target_state->pwr_domain_state;
|
|
stateid_afflvl2 = pwr_domain_state[PLAT_MAX_PWR_LVL] &
|
|
TEGRA194_STATE_ID_MASK;
|
|
|
|
if (stateid_afflvl2 == PSTATE_ID_SOC_POWERDN) {
|
|
|
|
/* save 'Secure Boot' Processor Feature Config Register */
|
|
val = mmio_read_32(TEGRA_MISC_BASE + MISCREG_PFCFG);
|
|
mmio_write_32(TEGRA_SCRATCH_BASE + SCRATCH_SECURE_BOOTP_FCFG, val);
|
|
|
|
/* save MC context */
|
|
mc_ctx_base = params_from_bl2->tzdram_base +
|
|
tegra194_get_mc_ctx_offset();
|
|
tegra_mc_save_context((uintptr_t)mc_ctx_base);
|
|
|
|
/*
|
|
* Suspend SE, RNG1 and PKA1 only on silcon and fpga,
|
|
* since VDK does not support atomic se ctx save
|
|
*/
|
|
if (tegra_platform_is_silicon() || tegra_platform_is_fpga()) {
|
|
ret = tegra_se_suspend();
|
|
assert(ret == 0);
|
|
}
|
|
|
|
/* Prepare for system suspend */
|
|
mce_update_cstate_info(&sc7_cstate_info);
|
|
|
|
do {
|
|
val = (uint32_t)mce_command_handler(
|
|
(uint32_t)MCE_CMD_IS_SC7_ALLOWED,
|
|
(uint32_t)TEGRA_NVG_CORE_C7,
|
|
MCE_CORE_SLEEP_TIME_INFINITE,
|
|
0U);
|
|
} while (val == 0U);
|
|
|
|
/* Instruct the MCE to enter system suspend state */
|
|
ret = mce_command_handler(
|
|
(uint64_t)MCE_CMD_ENTER_CSTATE,
|
|
(uint64_t)TEGRA_NVG_CORE_C7,
|
|
MCE_CORE_SLEEP_TIME_INFINITE,
|
|
0U);
|
|
assert(ret == 0);
|
|
|
|
/* set system suspend state for house-keeping */
|
|
tegra194_set_system_suspend_entry();
|
|
}
|
|
|
|
return PSCI_E_SUCCESS;
|
|
}
|
|
|
|
/*******************************************************************************
|
|
* Helper function to check if this is the last ON CPU in the cluster
|
|
******************************************************************************/
|
|
static bool tegra_last_on_cpu_in_cluster(const plat_local_state_t *states,
|
|
uint32_t ncpu)
|
|
{
|
|
plat_local_state_t target;
|
|
bool last_on_cpu = true;
|
|
uint32_t num_cpus = ncpu, pos = 0;
|
|
|
|
do {
|
|
target = states[pos];
|
|
if (target != PLAT_MAX_OFF_STATE) {
|
|
last_on_cpu = false;
|
|
}
|
|
--num_cpus;
|
|
pos++;
|
|
} while (num_cpus != 0U);
|
|
|
|
return last_on_cpu;
|
|
}
|
|
|
|
/*******************************************************************************
|
|
* Helper function to get target power state for the cluster
|
|
******************************************************************************/
|
|
static plat_local_state_t tegra_get_afflvl1_pwr_state(const plat_local_state_t *states,
|
|
uint32_t ncpu)
|
|
{
|
|
uint32_t core_pos = (uint32_t)read_mpidr() & (uint32_t)MPIDR_CPU_MASK;
|
|
plat_local_state_t target = states[core_pos];
|
|
mce_cstate_info_t cstate_info = { 0 };
|
|
|
|
/* CPU off */
|
|
if (target == PLAT_MAX_OFF_STATE) {
|
|
|
|
/* Enable cluster powerdn from last CPU in the cluster */
|
|
if (tegra_last_on_cpu_in_cluster(states, ncpu)) {
|
|
|
|
/* Enable CC6 state and turn off wake mask */
|
|
cstate_info.cluster = (uint32_t)TEGRA_NVG_CLUSTER_CC6;
|
|
cstate_info.ccplex = (uint32_t)TEGRA_NVG_CG_CG7;
|
|
cstate_info.system_state_force = 1;
|
|
cstate_info.update_wake_mask = 1U;
|
|
mce_update_cstate_info(&cstate_info);
|
|
|
|
} else {
|
|
|
|
/* Turn off wake_mask */
|
|
cstate_info.update_wake_mask = 1U;
|
|
mce_update_cstate_info(&cstate_info);
|
|
target = PSCI_LOCAL_STATE_RUN;
|
|
}
|
|
}
|
|
|
|
return target;
|
|
}
|
|
|
|
/*******************************************************************************
|
|
* Platform handler to calculate the proper target power level at the
|
|
* specified affinity level
|
|
******************************************************************************/
|
|
plat_local_state_t tegra_soc_get_target_pwr_state(uint32_t lvl,
|
|
const plat_local_state_t *states,
|
|
uint32_t ncpu)
|
|
{
|
|
plat_local_state_t target = PSCI_LOCAL_STATE_RUN;
|
|
uint32_t cpu = plat_my_core_pos();
|
|
|
|
/* System Suspend */
|
|
if ((lvl == (uint32_t)MPIDR_AFFLVL2) && (states[cpu] == PSTATE_ID_SOC_POWERDN)) {
|
|
target = PSTATE_ID_SOC_POWERDN;
|
|
}
|
|
|
|
/* CPU off, CPU suspend */
|
|
if (lvl == (uint32_t)MPIDR_AFFLVL1) {
|
|
target = tegra_get_afflvl1_pwr_state(states, ncpu);
|
|
}
|
|
|
|
/* target cluster/system state */
|
|
return target;
|
|
}
|
|
|
|
int32_t tegra_soc_pwr_domain_power_down_wfi(const psci_power_state_t *target_state)
|
|
{
|
|
const plat_local_state_t *pwr_domain_state =
|
|
target_state->pwr_domain_state;
|
|
plat_params_from_bl2_t *params_from_bl2 = bl31_get_plat_params();
|
|
uint8_t stateid_afflvl2 = pwr_domain_state[PLAT_MAX_PWR_LVL] &
|
|
TEGRA194_STATE_ID_MASK;
|
|
uint64_t src_len_in_bytes = (uintptr_t)&__BL31_END__ - (uintptr_t)BL31_BASE;
|
|
uint64_t val;
|
|
int32_t ret = PSCI_E_SUCCESS;
|
|
|
|
if (stateid_afflvl2 == PSTATE_ID_SOC_POWERDN) {
|
|
val = params_from_bl2->tzdram_base +
|
|
tegra194_get_cpu_reset_handler_size();
|
|
|
|
/* initialise communication channel with BPMP */
|
|
ret = tegra_bpmp_ipc_init();
|
|
assert(ret == 0);
|
|
|
|
/* Enable SE clock before SE context save */
|
|
ret = tegra_bpmp_ipc_enable_clock(TEGRA194_CLK_SE);
|
|
assert(ret == 0);
|
|
|
|
/*
|
|
* It is very unlikely that the BL31 image would be
|
|
* bigger than 2^32 bytes
|
|
*/
|
|
assert(src_len_in_bytes < UINT32_MAX);
|
|
|
|
if (tegra_se_calculate_save_sha256(BL31_BASE,
|
|
(uint32_t)src_len_in_bytes) != 0) {
|
|
ERROR("Hash calculation failed. Reboot\n");
|
|
(void)tegra_soc_prepare_system_reset();
|
|
}
|
|
|
|
/*
|
|
* The TZRAM loses power when we enter system suspend. To
|
|
* allow graceful exit from system suspend, we need to copy
|
|
* BL3-1 over to TZDRAM.
|
|
*/
|
|
val = params_from_bl2->tzdram_base +
|
|
tegra194_get_cpu_reset_handler_size();
|
|
memcpy((void *)(uintptr_t)val, (void *)(uintptr_t)BL31_BASE,
|
|
src_len_in_bytes);
|
|
|
|
/* Disable SE clock after SE context save */
|
|
ret = tegra_bpmp_ipc_disable_clock(TEGRA194_CLK_SE);
|
|
assert(ret == 0);
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
int32_t tegra_soc_pwr_domain_suspend_pwrdown_early(const psci_power_state_t *target_state)
|
|
{
|
|
return PSCI_E_NOT_SUPPORTED;
|
|
}
|
|
|
|
int32_t tegra_soc_pwr_domain_on(u_register_t mpidr)
|
|
{
|
|
uint64_t target_cpu = mpidr & MPIDR_CPU_MASK;
|
|
uint64_t target_cluster = (mpidr & MPIDR_CLUSTER_MASK) >>
|
|
MPIDR_AFFINITY_BITS;
|
|
int32_t ret = 0;
|
|
|
|
if (target_cluster > ((uint32_t)PLATFORM_CLUSTER_COUNT - 1U)) {
|
|
ERROR("%s: unsupported CPU (0x%lx)\n", __func__ , mpidr);
|
|
return PSCI_E_NOT_PRESENT;
|
|
}
|
|
|
|
/* construct the target CPU # */
|
|
target_cpu += (target_cluster << 1U);
|
|
|
|
ret = mce_command_handler((uint64_t)MCE_CMD_ONLINE_CORE, target_cpu, 0U, 0U);
|
|
if (ret < 0) {
|
|
return PSCI_E_DENIED;
|
|
}
|
|
|
|
return PSCI_E_SUCCESS;
|
|
}
|
|
|
|
int32_t tegra_soc_pwr_domain_on_finish(const psci_power_state_t *target_state)
|
|
{
|
|
const plat_params_from_bl2_t *params_from_bl2 = bl31_get_plat_params();
|
|
uint8_t enable_ccplex_lock_step = params_from_bl2->enable_ccplex_lock_step;
|
|
uint8_t stateid_afflvl2 = target_state->pwr_domain_state[PLAT_MAX_PWR_LVL];
|
|
cpu_context_t *ctx = cm_get_context(NON_SECURE);
|
|
uint64_t actlr_elx;
|
|
|
|
/*
|
|
* Reset power state info for CPUs when onlining, we set
|
|
* deepest power when offlining a core but that may not be
|
|
* requested by non-secure sw which controls idle states. It
|
|
* will re-init this info from non-secure software when the
|
|
* core come online.
|
|
*/
|
|
actlr_elx = read_ctx_reg((get_el1_sysregs_ctx(ctx)), (CTX_ACTLR_EL1));
|
|
actlr_elx &= ~DENVER_CPU_PMSTATE_MASK;
|
|
actlr_elx |= DENVER_CPU_PMSTATE_C1;
|
|
write_ctx_reg((get_el1_sysregs_ctx(ctx)), (CTX_ACTLR_EL1), (actlr_elx));
|
|
|
|
/*
|
|
* Check if we are exiting from deep sleep and restore SE
|
|
* context if we are.
|
|
*/
|
|
if (stateid_afflvl2 == PSTATE_ID_SOC_POWERDN) {
|
|
|
|
#if ENABLE_STRICT_CHECKING_MODE
|
|
/*
|
|
* Enable strict checking after programming the GSC for
|
|
* enabling TZSRAM and TZDRAM
|
|
*/
|
|
mce_enable_strict_checking();
|
|
#endif
|
|
|
|
/* Init SMMU */
|
|
tegra_smmu_init();
|
|
|
|
/* Resume SE, RNG1 and PKA1 */
|
|
tegra_se_resume();
|
|
|
|
/*
|
|
* Program XUSB STREAMIDs
|
|
* ======================
|
|
* T19x XUSB has support for XUSB virtualization. It will
|
|
* have one physical function (PF) and four Virtual functions
|
|
* (VF)
|
|
*
|
|
* There were below two SIDs for XUSB until T186.
|
|
* 1) #define TEGRA_SID_XUSB_HOST 0x1bU
|
|
* 2) #define TEGRA_SID_XUSB_DEV 0x1cU
|
|
*
|
|
* We have below four new SIDs added for VF(s)
|
|
* 3) #define TEGRA_SID_XUSB_VF0 0x5dU
|
|
* 4) #define TEGRA_SID_XUSB_VF1 0x5eU
|
|
* 5) #define TEGRA_SID_XUSB_VF2 0x5fU
|
|
* 6) #define TEGRA_SID_XUSB_VF3 0x60U
|
|
*
|
|
* When virtualization is enabled then we have to disable SID
|
|
* override and program above SIDs in below newly added SID
|
|
* registers in XUSB PADCTL MMIO space. These registers are
|
|
* TZ protected and so need to be done in ATF.
|
|
*
|
|
* a) #define XUSB_PADCTL_HOST_AXI_STREAMID_PF_0 (0x136cU)
|
|
* b) #define XUSB_PADCTL_DEV_AXI_STREAMID_PF_0 (0x139cU)
|
|
* c) #define XUSB_PADCTL_HOST_AXI_STREAMID_VF_0 (0x1370U)
|
|
* d) #define XUSB_PADCTL_HOST_AXI_STREAMID_VF_1 (0x1374U)
|
|
* e) #define XUSB_PADCTL_HOST_AXI_STREAMID_VF_2 (0x1378U)
|
|
* f) #define XUSB_PADCTL_HOST_AXI_STREAMID_VF_3 (0x137cU)
|
|
*
|
|
* This change disables SID override and programs XUSB SIDs
|
|
* in above registers to support both virtualization and
|
|
* non-virtualization platforms
|
|
*/
|
|
if (tegra_platform_is_silicon() || tegra_platform_is_fpga()) {
|
|
|
|
mmio_write_32(TEGRA_XUSB_PADCTL_BASE +
|
|
XUSB_PADCTL_HOST_AXI_STREAMID_PF_0, TEGRA_SID_XUSB_HOST);
|
|
assert(mmio_read_32(TEGRA_XUSB_PADCTL_BASE +
|
|
XUSB_PADCTL_HOST_AXI_STREAMID_PF_0) == TEGRA_SID_XUSB_HOST);
|
|
mmio_write_32(TEGRA_XUSB_PADCTL_BASE +
|
|
XUSB_PADCTL_HOST_AXI_STREAMID_VF_0, TEGRA_SID_XUSB_VF0);
|
|
assert(mmio_read_32(TEGRA_XUSB_PADCTL_BASE +
|
|
XUSB_PADCTL_HOST_AXI_STREAMID_VF_0) == TEGRA_SID_XUSB_VF0);
|
|
mmio_write_32(TEGRA_XUSB_PADCTL_BASE +
|
|
XUSB_PADCTL_HOST_AXI_STREAMID_VF_1, TEGRA_SID_XUSB_VF1);
|
|
assert(mmio_read_32(TEGRA_XUSB_PADCTL_BASE +
|
|
XUSB_PADCTL_HOST_AXI_STREAMID_VF_1) == TEGRA_SID_XUSB_VF1);
|
|
mmio_write_32(TEGRA_XUSB_PADCTL_BASE +
|
|
XUSB_PADCTL_HOST_AXI_STREAMID_VF_2, TEGRA_SID_XUSB_VF2);
|
|
assert(mmio_read_32(TEGRA_XUSB_PADCTL_BASE +
|
|
XUSB_PADCTL_HOST_AXI_STREAMID_VF_2) == TEGRA_SID_XUSB_VF2);
|
|
mmio_write_32(TEGRA_XUSB_PADCTL_BASE +
|
|
XUSB_PADCTL_HOST_AXI_STREAMID_VF_3, TEGRA_SID_XUSB_VF3);
|
|
assert(mmio_read_32(TEGRA_XUSB_PADCTL_BASE +
|
|
XUSB_PADCTL_HOST_AXI_STREAMID_VF_3) == TEGRA_SID_XUSB_VF3);
|
|
mmio_write_32(TEGRA_XUSB_PADCTL_BASE +
|
|
XUSB_PADCTL_DEV_AXI_STREAMID_PF_0, TEGRA_SID_XUSB_DEV);
|
|
assert(mmio_read_32(TEGRA_XUSB_PADCTL_BASE +
|
|
XUSB_PADCTL_DEV_AXI_STREAMID_PF_0) == TEGRA_SID_XUSB_DEV);
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Enable dual execution optimized translations for all ELx.
|
|
*/
|
|
if (enable_ccplex_lock_step != 0U) {
|
|
actlr_elx = read_actlr_el3();
|
|
actlr_elx |= DENVER_CPU_ENABLE_DUAL_EXEC_EL3;
|
|
write_actlr_el3(actlr_elx);
|
|
|
|
actlr_elx = read_actlr_el2();
|
|
actlr_elx |= DENVER_CPU_ENABLE_DUAL_EXEC_EL2;
|
|
write_actlr_el2(actlr_elx);
|
|
|
|
actlr_elx = read_actlr_el1();
|
|
actlr_elx |= DENVER_CPU_ENABLE_DUAL_EXEC_EL1;
|
|
write_actlr_el1(actlr_elx);
|
|
}
|
|
|
|
return PSCI_E_SUCCESS;
|
|
}
|
|
|
|
int32_t tegra_soc_pwr_domain_off_early(const psci_power_state_t *target_state)
|
|
{
|
|
/* Do not power off the boot CPU */
|
|
if (plat_is_my_cpu_primary()) {
|
|
return PSCI_E_DENIED;
|
|
}
|
|
|
|
return PSCI_E_SUCCESS;
|
|
}
|
|
|
|
int32_t tegra_soc_pwr_domain_off(const psci_power_state_t *target_state)
|
|
{
|
|
uint64_t impl = (read_midr() >> MIDR_IMPL_SHIFT) & MIDR_IMPL_MASK;
|
|
int32_t ret = 0;
|
|
|
|
(void)target_state;
|
|
|
|
/* Disable Denver's DCO operations */
|
|
if (impl == DENVER_IMPL) {
|
|
denver_disable_dco();
|
|
}
|
|
|
|
/* Turn off CPU */
|
|
ret = mce_command_handler((uint64_t)MCE_CMD_ENTER_CSTATE,
|
|
(uint64_t)TEGRA_NVG_CORE_C7, MCE_CORE_SLEEP_TIME_INFINITE, 0U);
|
|
assert(ret == 0);
|
|
|
|
return PSCI_E_SUCCESS;
|
|
}
|
|
|
|
__dead2 void tegra_soc_prepare_system_off(void)
|
|
{
|
|
/* System power off */
|
|
mce_system_shutdown();
|
|
|
|
wfi();
|
|
|
|
/* wait for the system to power down */
|
|
for (;;) {
|
|
;
|
|
}
|
|
}
|
|
|
|
int32_t tegra_soc_prepare_system_reset(void)
|
|
{
|
|
/* System reboot */
|
|
mce_system_reboot();
|
|
|
|
return PSCI_E_SUCCESS;
|
|
}
|