mirror of
https://github.com/ARM-software/arm-trusted-firmware.git
synced 2025-04-24 22:05:40 +00:00

Platform may use specific cache line sizes. Since CACHE_WRITEBACK_GRANULE defines the platform specific cache line size, it is used to define the size of the cpu data structure CPU_DATA_SIZE aligned on cache line size. Introduce assembly macro 'mov_imm' for AArch32 to simplify implementation of function '_cpu_data_by_index'. Change-Id: Ic2d49ffe0c3e51649425fd9c8c99559c582ac5a1 Signed-off-by: Etienne Carriere <etienne.carriere@linaro.org>
158 lines
4.8 KiB
C
158 lines
4.8 KiB
C
/*
|
|
* Copyright (c) 2014-2016, ARM Limited and Contributors. All rights reserved.
|
|
*
|
|
* SPDX-License-Identifier: BSD-3-Clause
|
|
*/
|
|
|
|
#ifndef __CPU_DATA_H__
|
|
#define __CPU_DATA_H__
|
|
|
|
#include <platform_def.h> /* CACHE_WRITEBACK_GRANULE required */
|
|
|
|
#ifdef AARCH32
|
|
|
|
#if CRASH_REPORTING
|
|
#error "Crash reporting is not supported in AArch32"
|
|
#endif
|
|
#define CPU_DATA_CPU_OPS_PTR 0x0
|
|
#define CPU_DATA_CRASH_BUF_OFFSET 0x4
|
|
|
|
#else /* AARCH32 */
|
|
|
|
/* Offsets for the cpu_data structure */
|
|
#define CPU_DATA_CRASH_BUF_OFFSET 0x18
|
|
/* need enough space in crash buffer to save 8 registers */
|
|
#define CPU_DATA_CRASH_BUF_SIZE 64
|
|
#define CPU_DATA_CPU_OPS_PTR 0x10
|
|
|
|
#endif /* AARCH32 */
|
|
|
|
#if CRASH_REPORTING
|
|
#define CPU_DATA_CRASH_BUF_END (CPU_DATA_CRASH_BUF_OFFSET + \
|
|
CPU_DATA_CRASH_BUF_SIZE)
|
|
#else
|
|
#define CPU_DATA_CRASH_BUF_END CPU_DATA_CRASH_BUF_OFFSET
|
|
#endif
|
|
|
|
/* cpu_data size is the data size rounded up to the platform cache line size */
|
|
#define CPU_DATA_SIZE (((CPU_DATA_CRASH_BUF_END + \
|
|
CACHE_WRITEBACK_GRANULE - 1) / \
|
|
CACHE_WRITEBACK_GRANULE) * \
|
|
CACHE_WRITEBACK_GRANULE)
|
|
|
|
#if ENABLE_RUNTIME_INSTRUMENTATION
|
|
/* Temporary space to store PMF timestamps from assembly code */
|
|
#define CPU_DATA_PMF_TS_COUNT 1
|
|
#define CPU_DATA_PMF_TS0_OFFSET CPU_DATA_CRASH_BUF_END
|
|
#define CPU_DATA_PMF_TS0_IDX 0
|
|
#endif
|
|
|
|
#ifndef __ASSEMBLY__
|
|
|
|
#include <arch_helpers.h>
|
|
#include <cassert.h>
|
|
#include <platform_def.h>
|
|
#include <psci.h>
|
|
#include <stdint.h>
|
|
|
|
/* Offsets for the cpu_data structure */
|
|
#define CPU_DATA_PSCI_LOCK_OFFSET __builtin_offsetof\
|
|
(cpu_data_t, psci_svc_cpu_data.pcpu_bakery_info)
|
|
|
|
#if PLAT_PCPU_DATA_SIZE
|
|
#define CPU_DATA_PLAT_PCPU_OFFSET __builtin_offsetof\
|
|
(cpu_data_t, platform_cpu_data)
|
|
#endif
|
|
|
|
/*******************************************************************************
|
|
* Function & variable prototypes
|
|
******************************************************************************/
|
|
|
|
/*******************************************************************************
|
|
* Cache of frequently used per-cpu data:
|
|
* Pointers to non-secure and secure security state contexts
|
|
* Address of the crash stack
|
|
* It is aligned to the cache line boundary to allow efficient concurrent
|
|
* manipulation of these pointers on different cpus
|
|
*
|
|
* TODO: Add other commonly used variables to this (tf_issues#90)
|
|
*
|
|
* The data structure and the _cpu_data accessors should not be used directly
|
|
* by components that have per-cpu members. The member access macros should be
|
|
* used for this.
|
|
******************************************************************************/
|
|
typedef struct cpu_data {
|
|
#ifndef AARCH32
|
|
void *cpu_context[2];
|
|
#endif
|
|
uintptr_t cpu_ops_ptr;
|
|
#if CRASH_REPORTING
|
|
u_register_t crash_buf[CPU_DATA_CRASH_BUF_SIZE >> 3];
|
|
#endif
|
|
#if ENABLE_RUNTIME_INSTRUMENTATION
|
|
uint64_t cpu_data_pmf_ts[CPU_DATA_PMF_TS_COUNT];
|
|
#endif
|
|
struct psci_cpu_data psci_svc_cpu_data;
|
|
#if PLAT_PCPU_DATA_SIZE
|
|
uint8_t platform_cpu_data[PLAT_PCPU_DATA_SIZE];
|
|
#endif
|
|
} __aligned(CACHE_WRITEBACK_GRANULE) cpu_data_t;
|
|
|
|
#if CRASH_REPORTING
|
|
/* verify assembler offsets match data structures */
|
|
CASSERT(CPU_DATA_CRASH_BUF_OFFSET == __builtin_offsetof
|
|
(cpu_data_t, crash_buf),
|
|
assert_cpu_data_crash_stack_offset_mismatch);
|
|
#endif
|
|
|
|
CASSERT(CPU_DATA_SIZE == sizeof(cpu_data_t),
|
|
assert_cpu_data_size_mismatch);
|
|
|
|
CASSERT(CPU_DATA_CPU_OPS_PTR == __builtin_offsetof
|
|
(cpu_data_t, cpu_ops_ptr),
|
|
assert_cpu_data_cpu_ops_ptr_offset_mismatch);
|
|
|
|
#if ENABLE_RUNTIME_INSTRUMENTATION
|
|
CASSERT(CPU_DATA_PMF_TS0_OFFSET == __builtin_offsetof
|
|
(cpu_data_t, cpu_data_pmf_ts[0]),
|
|
assert_cpu_data_pmf_ts0_offset_mismatch);
|
|
#endif
|
|
|
|
struct cpu_data *_cpu_data_by_index(uint32_t cpu_index);
|
|
|
|
#ifndef AARCH32
|
|
/* Return the cpu_data structure for the current CPU. */
|
|
static inline struct cpu_data *_cpu_data(void)
|
|
{
|
|
return (cpu_data_t *)read_tpidr_el3();
|
|
}
|
|
#else
|
|
struct cpu_data *_cpu_data(void);
|
|
#endif
|
|
|
|
/**************************************************************************
|
|
* APIs for initialising and accessing per-cpu data
|
|
*************************************************************************/
|
|
|
|
void init_cpu_data_ptr(void);
|
|
void init_cpu_ops(void);
|
|
|
|
#define get_cpu_data(_m) _cpu_data()->_m
|
|
#define set_cpu_data(_m, _v) _cpu_data()->_m = _v
|
|
#define get_cpu_data_by_index(_ix, _m) _cpu_data_by_index(_ix)->_m
|
|
#define set_cpu_data_by_index(_ix, _m, _v) _cpu_data_by_index(_ix)->_m = _v
|
|
|
|
#define flush_cpu_data(_m) flush_dcache_range((uintptr_t) \
|
|
&(_cpu_data()->_m), \
|
|
sizeof(_cpu_data()->_m))
|
|
#define inv_cpu_data(_m) inv_dcache_range((uintptr_t) \
|
|
&(_cpu_data()->_m), \
|
|
sizeof(_cpu_data()->_m))
|
|
#define flush_cpu_data_by_index(_ix, _m) \
|
|
flush_dcache_range((uintptr_t) \
|
|
&(_cpu_data_by_index(_ix)->_m), \
|
|
sizeof(_cpu_data_by_index(_ix)->_m))
|
|
|
|
|
|
#endif /* __ASSEMBLY__ */
|
|
#endif /* __CPU_DATA_H__ */
|