blob: 11ef6e5f47f90213f6c535849a5efa00b6acd6b2 [file] [log] [blame]
Soby Mathew748be1d2016-05-05 14:10:46 +01001/*
Antonio Nino Diaz3c817f42018-03-21 10:49:27 +00002 * Copyright (c) 2016-2018, ARM Limited and Contributors. All rights reserved.
Soby Mathew748be1d2016-05-05 14:10:46 +01003 *
dp-armfa3cf0b2017-05-03 09:38:09 +01004 * SPDX-License-Identifier: BSD-3-Clause
Soby Mathew748be1d2016-05-05 14:10:46 +01005 */
6
Dimitris Papastamosdda48b02017-10-17 14:03:14 +01007#include <amu.h>
Soby Mathew748be1d2016-05-05 14:10:46 +01008#include <arch.h>
9#include <arch_helpers.h>
10#include <assert.h>
11#include <bl_common.h>
12#include <context.h>
13#include <context_mgmt.h>
14#include <platform.h>
15#include <platform_def.h>
Antonio Nino Diaz3c817f42018-03-21 10:49:27 +000016#include <smccc_helpers.h>
Soby Mathew748be1d2016-05-05 14:10:46 +010017#include <string.h>
Douglas Raillarda8954fc2017-01-26 15:54:44 +000018#include <utils.h>
Soby Mathew748be1d2016-05-05 14:10:46 +010019
20/*******************************************************************************
21 * Context management library initialisation routine. This library is used by
22 * runtime services to share pointers to 'cpu_context' structures for the secure
23 * and non-secure states. Management of the structures and their associated
24 * memory is not done by the context management library e.g. the PSCI service
25 * manages the cpu context used for entry from and exit to the non-secure state.
26 * The Secure payload manages the context(s) corresponding to the secure state.
27 * It also uses this library to get access to the non-secure
28 * state cpu context pointers.
29 ******************************************************************************/
30void cm_init(void)
31{
32 /*
33 * The context management library has only global data to initialize, but
34 * that will be done when the BSS is zeroed out
35 */
36}
37
38/*******************************************************************************
39 * The following function initializes the cpu_context 'ctx' for
40 * first use, and sets the initial entrypoint state as specified by the
41 * entry_point_info structure.
42 *
43 * The security state to initialize is determined by the SECURE attribute
Antonio Nino Diaz28dce9e2018-05-22 10:09:10 +010044 * of the entry_point_info.
Soby Mathew748be1d2016-05-05 14:10:46 +010045 *
46 * The EE and ST attributes are used to configure the endianness and secure
47 * timer availability for the new execution context.
48 *
49 * To prepare the register state for entry call cm_prepare_el3_exit() and
50 * el3_exit(). For Secure-EL1 cm_prepare_el3_exit() is equivalent to
51 * cm_e1_sysreg_context_restore().
52 ******************************************************************************/
Antonio Nino Diaz28dce9e2018-05-22 10:09:10 +010053void cm_setup_context(cpu_context_t *ctx, const entry_point_info_t *ep)
Soby Mathew748be1d2016-05-05 14:10:46 +010054{
55 unsigned int security_state;
56 uint32_t scr, sctlr;
57 regs_t *reg_ctx;
58
59 assert(ctx);
60
61 security_state = GET_SECURITY_STATE(ep->h.attr);
62
63 /* Clear any residual register values from the context */
Douglas Raillarda8954fc2017-01-26 15:54:44 +000064 zeromem(ctx, sizeof(*ctx));
Soby Mathew748be1d2016-05-05 14:10:46 +010065
Soby Mathewb4a970a2016-08-31 12:34:33 +010066 reg_ctx = get_regs_ctx(ctx);
67
Soby Mathew748be1d2016-05-05 14:10:46 +010068 /*
69 * Base the context SCR on the current value, adjust for entry point
70 * specific requirements
71 */
72 scr = read_scr();
73 scr &= ~(SCR_NS_BIT | SCR_HCE_BIT);
74
75 if (security_state != SECURE)
76 scr |= SCR_NS_BIT;
77
Soby Mathew748be1d2016-05-05 14:10:46 +010078 if (security_state != SECURE) {
Soby Mathewa993c422016-09-29 14:15:57 +010079 /*
David Cunadofee86532017-04-13 22:38:29 +010080 * Set up SCTLR for the Non-secure context.
81 *
82 * SCTLR.EE: Endianness is taken from the entrypoint attributes.
83 *
84 * SCTLR.M, SCTLR.C and SCTLR.I: These fields must be zero (as
85 * required by PSCI specification)
86 *
87 * Set remaining SCTLR fields to their architecturally defined
88 * values. Some fields reset to an IMPLEMENTATION DEFINED value:
89 *
90 * SCTLR.TE: Set to zero so that exceptions to an Exception
91 * Level executing at PL1 are taken to A32 state.
92 *
93 * SCTLR.V: Set to zero to select the normal exception vectors
94 * with base address held in VBAR.
Soby Mathewa993c422016-09-29 14:15:57 +010095 */
David Cunadofee86532017-04-13 22:38:29 +010096 assert(((ep->spsr >> SPSR_E_SHIFT) & SPSR_E_MASK) ==
97 (EP_GET_EE(ep->h.attr) >> EP_EE_SHIFT));
98
99 sctlr = EP_GET_EE(ep->h.attr) ? SCTLR_EE_BIT : 0;
100 sctlr |= (SCTLR_RESET_VAL & ~(SCTLR_TE_BIT | SCTLR_V_BIT));
Soby Mathew748be1d2016-05-05 14:10:46 +0100101 write_ctx_reg(reg_ctx, CTX_NS_SCTLR, sctlr);
102 }
103
David Cunadofee86532017-04-13 22:38:29 +0100104 /*
105 * The target exception level is based on the spsr mode requested. If
106 * execution is requested to hyp mode, HVC is enabled via SCR.HCE.
107 */
Soby Mathew748be1d2016-05-05 14:10:46 +0100108 if (GET_M32(ep->spsr) == MODE32_hyp)
109 scr |= SCR_HCE_BIT;
110
David Cunadofee86532017-04-13 22:38:29 +0100111 /*
112 * Store the initialised values for SCTLR and SCR in the cpu_context.
113 * The Hyp mode registers are not part of the saved context and are
114 * set-up in cm_prepare_el3_exit().
115 */
Soby Mathew748be1d2016-05-05 14:10:46 +0100116 write_ctx_reg(reg_ctx, CTX_SCR, scr);
117 write_ctx_reg(reg_ctx, CTX_LR, ep->pc);
118 write_ctx_reg(reg_ctx, CTX_SPSR, ep->spsr);
119
120 /*
121 * Store the r0-r3 value from the entrypoint into the context
122 * Use memcpy as we are in control of the layout of the structures
123 */
124 memcpy((void *)reg_ctx, (void *)&ep->args, sizeof(aapcs32_params_t));
125}
126
127/*******************************************************************************
Dimitris Papastamos1e6f93e2017-11-07 09:55:29 +0000128 * Enable architecture extensions on first entry to Non-secure world.
129 * When EL2 is implemented but unused `el2_unused` is non-zero, otherwise
130 * it is zero.
131 ******************************************************************************/
132static void enable_extensions_nonsecure(int el2_unused)
133{
134#if IMAGE_BL32
Dimitris Papastamosdda48b02017-10-17 14:03:14 +0100135#if ENABLE_AMU
136 amu_enable(el2_unused);
137#endif
Dimitris Papastamos1e6f93e2017-11-07 09:55:29 +0000138#endif
139}
140
141/*******************************************************************************
Soby Mathew748be1d2016-05-05 14:10:46 +0100142 * The following function initializes the cpu_context for a CPU specified by
143 * its `cpu_idx` for first use, and sets the initial entrypoint state as
144 * specified by the entry_point_info structure.
145 ******************************************************************************/
146void cm_init_context_by_index(unsigned int cpu_idx,
147 const entry_point_info_t *ep)
148{
149 cpu_context_t *ctx;
150 ctx = cm_get_context_by_index(cpu_idx, GET_SECURITY_STATE(ep->h.attr));
Antonio Nino Diaz28dce9e2018-05-22 10:09:10 +0100151 cm_setup_context(ctx, ep);
Soby Mathew748be1d2016-05-05 14:10:46 +0100152}
153
154/*******************************************************************************
155 * The following function initializes the cpu_context for the current CPU
156 * for first use, and sets the initial entrypoint state as specified by the
157 * entry_point_info structure.
158 ******************************************************************************/
159void cm_init_my_context(const entry_point_info_t *ep)
160{
161 cpu_context_t *ctx;
162 ctx = cm_get_context(GET_SECURITY_STATE(ep->h.attr));
Antonio Nino Diaz28dce9e2018-05-22 10:09:10 +0100163 cm_setup_context(ctx, ep);
Soby Mathew748be1d2016-05-05 14:10:46 +0100164}
165
166/*******************************************************************************
167 * Prepare the CPU system registers for first entry into secure or normal world
168 *
169 * If execution is requested to hyp mode, HSCTLR is initialized
170 * If execution is requested to non-secure PL1, and the CPU supports
171 * HYP mode then HYP mode is disabled by configuring all necessary HYP mode
172 * registers.
173 ******************************************************************************/
174void cm_prepare_el3_exit(uint32_t security_state)
175{
David Cunadofee86532017-04-13 22:38:29 +0100176 uint32_t hsctlr, scr;
Soby Mathew748be1d2016-05-05 14:10:46 +0100177 cpu_context_t *ctx = cm_get_context(security_state);
Dimitris Papastamos1e6f93e2017-11-07 09:55:29 +0000178 int el2_unused = 0;
Soby Mathew748be1d2016-05-05 14:10:46 +0100179
180 assert(ctx);
181
182 if (security_state == NON_SECURE) {
183 scr = read_ctx_reg(get_regs_ctx(ctx), CTX_SCR);
184 if (scr & SCR_HCE_BIT) {
185 /* Use SCTLR value to initialize HSCTLR */
David Cunadofee86532017-04-13 22:38:29 +0100186 hsctlr = read_ctx_reg(get_regs_ctx(ctx),
Soby Mathew748be1d2016-05-05 14:10:46 +0100187 CTX_NS_SCTLR);
David Cunadofee86532017-04-13 22:38:29 +0100188 hsctlr |= HSCTLR_RES1;
Soby Mathew748be1d2016-05-05 14:10:46 +0100189 /* Temporarily set the NS bit to access HSCTLR */
190 write_scr(read_scr() | SCR_NS_BIT);
191 /*
192 * Make sure the write to SCR is complete so that
193 * we can access HSCTLR
194 */
195 isb();
David Cunadofee86532017-04-13 22:38:29 +0100196 write_hsctlr(hsctlr);
Soby Mathew748be1d2016-05-05 14:10:46 +0100197 isb();
198
199 write_scr(read_scr() & ~SCR_NS_BIT);
200 isb();
201 } else if (read_id_pfr1() &
202 (ID_PFR1_VIRTEXT_MASK << ID_PFR1_VIRTEXT_SHIFT)) {
Dimitris Papastamos1e6f93e2017-11-07 09:55:29 +0000203 el2_unused = 1;
204
David Cunado5f55e282016-10-31 17:37:34 +0000205 /*
206 * Set the NS bit to access NS copies of certain banked
207 * registers
208 */
Soby Mathew748be1d2016-05-05 14:10:46 +0100209 write_scr(read_scr() | SCR_NS_BIT);
210 isb();
211
David Cunadofee86532017-04-13 22:38:29 +0100212 /*
213 * Hyp / PL2 present but unused, need to disable safely.
214 * HSCTLR can be ignored in this case.
215 *
216 * Set HCR to its architectural reset value so that
217 * Non-secure operations do not trap to Hyp mode.
218 */
219 write_hcr(HCR_RESET_VAL);
Soby Mathew748be1d2016-05-05 14:10:46 +0100220
David Cunadofee86532017-04-13 22:38:29 +0100221 /*
222 * Set HCPTR to its architectural reset value so that
223 * Non-secure access from EL1 or EL0 to trace and to
224 * Advanced SIMD and floating point functionality does
225 * not trap to Hyp mode.
226 */
227 write_hcptr(HCPTR_RESET_VAL);
Soby Mathew748be1d2016-05-05 14:10:46 +0100228
David Cunadofee86532017-04-13 22:38:29 +0100229 /*
230 * Initialise CNTHCTL. All fields are architecturally
231 * UNKNOWN on reset and are set to zero except for
232 * field(s) listed below.
233 *
234 * CNTHCTL.PL1PCEN: Disable traps to Hyp mode of
235 * Non-secure EL0 and EL1 accessed to the physical
236 * timer registers.
237 *
238 * CNTHCTL.PL1PCTEN: Disable traps to Hyp mode of
239 * Non-secure EL0 and EL1 accessed to the physical
240 * counter registers.
241 */
242 write_cnthctl(CNTHCTL_RESET_VAL |
243 PL1PCEN_BIT | PL1PCTEN_BIT);
Soby Mathew748be1d2016-05-05 14:10:46 +0100244
David Cunadofee86532017-04-13 22:38:29 +0100245 /*
246 * Initialise CNTVOFF to zero as it resets to an
247 * IMPLEMENTATION DEFINED value.
248 */
Soby Mathew748be1d2016-05-05 14:10:46 +0100249 write64_cntvoff(0);
250
David Cunadofee86532017-04-13 22:38:29 +0100251 /*
252 * Set VPIDR and VMPIDR to match MIDR_EL1 and MPIDR
253 * respectively.
254 */
Soby Mathew748be1d2016-05-05 14:10:46 +0100255 write_vpidr(read_midr());
256 write_vmpidr(read_mpidr());
257
258 /*
David Cunadofee86532017-04-13 22:38:29 +0100259 * Initialise VTTBR, setting all fields rather than
260 * relying on the hw. Some fields are architecturally
261 * UNKNOWN at reset.
262 *
263 * VTTBR.VMID: Set to zero which is the architecturally
264 * defined reset value. Even though EL1&0 stage 2
265 * address translation is disabled, cache maintenance
266 * operations depend on the VMID.
267 *
268 * VTTBR.BADDR: Set to zero as EL1&0 stage 2 address
269 * translation is disabled.
Soby Mathew748be1d2016-05-05 14:10:46 +0100270 */
David Cunadofee86532017-04-13 22:38:29 +0100271 write64_vttbr(VTTBR_RESET_VAL &
272 ~((VTTBR_VMID_MASK << VTTBR_VMID_SHIFT)
273 | (VTTBR_BADDR_MASK << VTTBR_BADDR_SHIFT)));
David Cunado5f55e282016-10-31 17:37:34 +0000274
275 /*
David Cunadofee86532017-04-13 22:38:29 +0100276 * Initialise HDCR, setting all the fields rather than
277 * relying on hw.
278 *
279 * HDCR.HPMN: Set to value of PMCR.N which is the
280 * architecturally-defined reset value.
David Cunado5f55e282016-10-31 17:37:34 +0000281 */
David Cunadofee86532017-04-13 22:38:29 +0100282 write_hdcr(HDCR_RESET_VAL |
283 ((read_pmcr() & PMCR_N_BITS) >> PMCR_N_SHIFT));
David Cunadoc14b08e2016-11-25 00:21:59 +0000284
285 /*
David Cunadofee86532017-04-13 22:38:29 +0100286 * Set HSTR to its architectural reset value so that
287 * access to system registers in the cproc=1111
288 * encoding space do not trap to Hyp mode.
289 */
290 write_hstr(HSTR_RESET_VAL);
291 /*
292 * Set CNTHP_CTL to its architectural reset value to
293 * disable the EL2 physical timer and prevent timer
294 * interrupts. Some fields are architecturally UNKNOWN
295 * on reset and are set to zero.
David Cunadoc14b08e2016-11-25 00:21:59 +0000296 */
David Cunadofee86532017-04-13 22:38:29 +0100297 write_cnthp_ctl(CNTHP_CTL_RESET_VAL);
Soby Mathew748be1d2016-05-05 14:10:46 +0100298 isb();
299
300 write_scr(read_scr() & ~SCR_NS_BIT);
301 isb();
302 }
Dimitris Papastamos1e6f93e2017-11-07 09:55:29 +0000303 enable_extensions_nonsecure(el2_unused);
Soby Mathew748be1d2016-05-05 14:10:46 +0100304 }
305}