blob: d0bca64fccad4f6a7f037c4e982c66ae476f1bf0 [file] [log] [blame]
Achin Gupta9ac63c52014-01-16 12:08:03 +00001/*
2 * Copyright (c) 2013-2014, ARM Limited and Contributors. All rights reserved.
3 *
4 * Redistribution and use in source and binary forms, with or without
5 * modification, are permitted provided that the following conditions are met:
6 *
7 * Redistributions of source code must retain the above copyright notice, this
8 * list of conditions and the following disclaimer.
9 *
10 * Redistributions in binary form must reproduce the above copyright notice,
11 * this list of conditions and the following disclaimer in the documentation
12 * and/or other materials provided with the distribution.
13 *
14 * Neither the name of ARM nor the names of its contributors may be used
15 * to endorse or promote products derived from this software without specific
16 * prior written permission.
17 *
18 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
19 * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
20 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
21 * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE
22 * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
23 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
24 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
25 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
26 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
27 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
28 * POSSIBILITY OF SUCH DAMAGE.
29 */
30
Dan Handley2bd4ef22014-04-09 13:14:54 +010031#include <arch.h>
Andrew Thoelke38bde412014-03-18 13:46:55 +000032#include <asm_macros.S>
Dan Handley2bd4ef22014-04-09 13:14:54 +010033#include <context.h>
Achin Gupta9ac63c52014-01-16 12:08:03 +000034
35/* -----------------------------------------------------
36 * The following function strictly follows the AArch64
37 * PCS to use x9-x17 (temporary caller-saved registers)
38 * to save essential EL3 system register context. It
39 * assumes that 'x0' is pointing to a 'el1_sys_regs'
40 * structure where the register context will be saved.
41 * -----------------------------------------------------
42 */
43 .global el3_sysregs_context_save
Andrew Thoelke38bde412014-03-18 13:46:55 +000044func el3_sysregs_context_save
Achin Gupta9ac63c52014-01-16 12:08:03 +000045
46 mrs x9, scr_el3
47 mrs x10, sctlr_el3
48 stp x9, x10, [x0, #CTX_SCR_EL3]
49
50 mrs x11, cptr_el3
51 stp x11, xzr, [x0, #CTX_CPTR_EL3]
52
53 mrs x13, cntfrq_el0
54 mrs x14, mair_el3
55 stp x13, x14, [x0, #CTX_CNTFRQ_EL0]
56
57 mrs x15, tcr_el3
58 mrs x16, ttbr0_el3
59 stp x15, x16, [x0, #CTX_TCR_EL3]
60
61 mrs x17, daif
62 and x17, x17, #(DAIF_ABT_BIT | DAIF_DBG_BIT)
63 stp x17, xzr, [x0, #CTX_DAIF_EL3]
64
65 ret
66
67/* -----------------------------------------------------
68 * The following function strictly follows the AArch64
69 * PCS to use x9-x17 (temporary caller-saved registers)
70 * to restore essential EL3 system register context. It
71 * assumes that 'x0' is pointing to a 'el1_sys_regs'
72 * structure from where the register context will be
73 * restored.
74 *
75 * Note that the sequence differs from that of the save
76 * function as we want the MMU to be enabled last
77 * -----------------------------------------------------
78 */
79 .global el3_sysregs_context_restore
Andrew Thoelke38bde412014-03-18 13:46:55 +000080func el3_sysregs_context_restore
Achin Gupta9ac63c52014-01-16 12:08:03 +000081
82 ldp x11, xzr, [x0, #CTX_CPTR_EL3]
83 msr cptr_el3, x11
84
85 ldp x13, x14, [x0, #CTX_CNTFRQ_EL0]
86 msr cntfrq_el0, x13
87 msr mair_el3, x14
88
89 ldp x15, x16, [x0, #CTX_TCR_EL3]
90 msr tcr_el3, x15
91 msr ttbr0_el3, x16
92
93 ldp x17, xzr, [x0, #CTX_DAIF_EL3]
94 mrs x11, daif
95 orr x17, x17, x11
96 msr daif, x17
97
98 /* Make sure all the above changes are observed */
99 isb
100
101 ldp x9, x10, [x0, #CTX_SCR_EL3]
102 msr scr_el3, x9
103 msr sctlr_el3, x10
104 isb
105
106 ret
107
108/* -----------------------------------------------------
109 * The following function strictly follows the AArch64
110 * PCS to use x9-x17 (temporary caller-saved registers)
111 * to save EL1 system register context. It assumes that
112 * 'x0' is pointing to a 'el1_sys_regs' structure where
113 * the register context will be saved.
114 * -----------------------------------------------------
115 */
116 .global el1_sysregs_context_save
Andrew Thoelke38bde412014-03-18 13:46:55 +0000117func el1_sysregs_context_save
Achin Gupta9ac63c52014-01-16 12:08:03 +0000118
119 mrs x9, spsr_el1
120 mrs x10, elr_el1
121 stp x9, x10, [x0, #CTX_SPSR_EL1]
122
123 mrs x11, spsr_abt
124 mrs x12, spsr_und
125 stp x11, x12, [x0, #CTX_SPSR_ABT]
126
127 mrs x13, spsr_irq
128 mrs x14, spsr_fiq
129 stp x13, x14, [x0, #CTX_SPSR_IRQ]
130
131 mrs x15, sctlr_el1
132 mrs x16, actlr_el1
133 stp x15, x16, [x0, #CTX_SCTLR_EL1]
134
135 mrs x17, cpacr_el1
136 mrs x9, csselr_el1
137 stp x17, x9, [x0, #CTX_CPACR_EL1]
138
139 mrs x10, sp_el1
140 mrs x11, esr_el1
141 stp x10, x11, [x0, #CTX_SP_EL1]
142
143 mrs x12, ttbr0_el1
144 mrs x13, ttbr1_el1
145 stp x12, x13, [x0, #CTX_TTBR0_EL1]
146
147 mrs x14, mair_el1
148 mrs x15, amair_el1
149 stp x14, x15, [x0, #CTX_MAIR_EL1]
150
151 mrs x16, tcr_el1
152 mrs x17, tpidr_el1
153 stp x16, x17, [x0, #CTX_TCR_EL1]
154
155 mrs x9, tpidr_el0
156 mrs x10, tpidrro_el0
157 stp x9, x10, [x0, #CTX_TPIDR_EL0]
158
159 mrs x11, dacr32_el2
160 mrs x12, ifsr32_el2
161 stp x11, x12, [x0, #CTX_DACR32_EL2]
162
163 mrs x13, par_el1
164 mrs x14, far_el1
165 stp x13, x14, [x0, #CTX_PAR_EL1]
166
167 mrs x15, afsr0_el1
168 mrs x16, afsr1_el1
169 stp x15, x16, [x0, #CTX_AFSR0_EL1]
170
171 mrs x17, contextidr_el1
172 mrs x9, vbar_el1
173 stp x17, x9, [x0, #CTX_CONTEXTIDR_EL1]
174
Jeenu Viswambharand1b60152014-05-12 15:28:47 +0100175 /* Save NS timer registers if the build has instructed so */
176#if NS_TIMER_SWITCH
Achin Gupta9ac63c52014-01-16 12:08:03 +0000177 mrs x10, cntp_ctl_el0
178 mrs x11, cntp_cval_el0
179 stp x10, x11, [x0, #CTX_CNTP_CTL_EL0]
180
181 mrs x12, cntv_ctl_el0
182 mrs x13, cntv_cval_el0
183 stp x12, x13, [x0, #CTX_CNTV_CTL_EL0]
184
185 mrs x14, cntkctl_el1
Jeenu Viswambharand1b60152014-05-12 15:28:47 +0100186 str x14, [x0, #CTX_CNTKCTL_EL1]
187#endif
188
Achin Gupta9ac63c52014-01-16 12:08:03 +0000189 mrs x15, fpexc32_el2
Jeenu Viswambharand1b60152014-05-12 15:28:47 +0100190 str x15, [x0, #CTX_FP_FPEXC32_EL2]
Achin Gupta9ac63c52014-01-16 12:08:03 +0000191
192 ret
193
194/* -----------------------------------------------------
195 * The following function strictly follows the AArch64
196 * PCS to use x9-x17 (temporary caller-saved registers)
197 * to restore EL1 system register context. It assumes
198 * that 'x0' is pointing to a 'el1_sys_regs' structure
199 * from where the register context will be restored
200 * -----------------------------------------------------
201 */
202 .global el1_sysregs_context_restore
Andrew Thoelke38bde412014-03-18 13:46:55 +0000203func el1_sysregs_context_restore
Achin Gupta9ac63c52014-01-16 12:08:03 +0000204
205 ldp x9, x10, [x0, #CTX_SPSR_EL1]
206 msr spsr_el1, x9
207 msr elr_el1, x10
208
209 ldp x11, x12, [x0, #CTX_SPSR_ABT]
210 msr spsr_abt, x11
211 msr spsr_und, x12
212
213 ldp x13, x14, [x0, #CTX_SPSR_IRQ]
214 msr spsr_irq, x13
215 msr spsr_fiq, x14
216
217 ldp x15, x16, [x0, #CTX_SCTLR_EL1]
218 msr sctlr_el1, x15
219 msr actlr_el1, x16
220
221 ldp x17, x9, [x0, #CTX_CPACR_EL1]
222 msr cpacr_el1, x17
223 msr csselr_el1, x9
224
225 ldp x10, x11, [x0, #CTX_SP_EL1]
226 msr sp_el1, x10
227 msr esr_el1, x11
228
229 ldp x12, x13, [x0, #CTX_TTBR0_EL1]
230 msr ttbr0_el1, x12
231 msr ttbr1_el1, x13
232
233 ldp x14, x15, [x0, #CTX_MAIR_EL1]
234 msr mair_el1, x14
235 msr amair_el1, x15
236
237 ldp x16, x17, [x0, #CTX_TCR_EL1]
238 msr tcr_el1, x16
239 msr tpidr_el1, x17
240
241 ldp x9, x10, [x0, #CTX_TPIDR_EL0]
242 msr tpidr_el0, x9
243 msr tpidrro_el0, x10
244
245 ldp x11, x12, [x0, #CTX_DACR32_EL2]
246 msr dacr32_el2, x11
247 msr ifsr32_el2, x12
248
249 ldp x13, x14, [x0, #CTX_PAR_EL1]
250 msr par_el1, x13
251 msr far_el1, x14
252
253 ldp x15, x16, [x0, #CTX_AFSR0_EL1]
254 msr afsr0_el1, x15
255 msr afsr1_el1, x16
256
257 ldp x17, x9, [x0, #CTX_CONTEXTIDR_EL1]
258 msr contextidr_el1, x17
259 msr vbar_el1, x9
260
Jeenu Viswambharand1b60152014-05-12 15:28:47 +0100261 /* Restore NS timer registers if the build has instructed so */
262#if NS_TIMER_SWITCH
Achin Gupta9ac63c52014-01-16 12:08:03 +0000263 ldp x10, x11, [x0, #CTX_CNTP_CTL_EL0]
264 msr cntp_ctl_el0, x10
265 msr cntp_cval_el0, x11
266
267 ldp x12, x13, [x0, #CTX_CNTV_CTL_EL0]
268 msr cntv_ctl_el0, x12
269 msr cntv_cval_el0, x13
270
Jeenu Viswambharand1b60152014-05-12 15:28:47 +0100271 ldr x14, [x0, #CTX_CNTKCTL_EL1]
Achin Gupta9ac63c52014-01-16 12:08:03 +0000272 msr cntkctl_el1, x14
Jeenu Viswambharand1b60152014-05-12 15:28:47 +0100273#endif
274
275 ldr x15, [x0, #CTX_FP_FPEXC32_EL2]
Achin Gupta9ac63c52014-01-16 12:08:03 +0000276 msr fpexc32_el2, x15
277
278 /* No explict ISB required here as ERET covers it */
279
280 ret
281
282/* -----------------------------------------------------
283 * The followsing function follows the aapcs_64 strictly
284 * to use x9-x17 (temporary caller-saved registers
285 * according to AArch64 PCS) to save floating point
286 * register context. It assumes that 'x0' is pointing to
287 * a 'fp_regs' structure where the register context will
288 * be saved.
289 *
290 * Access to VFP registers will trap if CPTR_EL3.TFP is
291 * set. However currently we don't use VFP registers
292 * nor set traps in Trusted Firmware, and assume it's
293 * cleared
294 *
295 * TODO: Revisit when VFP is used in secure world
296 * -----------------------------------------------------
297 */
298 .global fpregs_context_save
Andrew Thoelke38bde412014-03-18 13:46:55 +0000299func fpregs_context_save
Achin Gupta9ac63c52014-01-16 12:08:03 +0000300 stp q0, q1, [x0, #CTX_FP_Q0]
301 stp q2, q3, [x0, #CTX_FP_Q2]
302 stp q4, q5, [x0, #CTX_FP_Q4]
303 stp q6, q7, [x0, #CTX_FP_Q6]
304 stp q8, q9, [x0, #CTX_FP_Q8]
305 stp q10, q11, [x0, #CTX_FP_Q10]
306 stp q12, q13, [x0, #CTX_FP_Q12]
307 stp q14, q15, [x0, #CTX_FP_Q14]
308 stp q16, q17, [x0, #CTX_FP_Q16]
309 stp q18, q19, [x0, #CTX_FP_Q18]
310 stp q20, q21, [x0, #CTX_FP_Q20]
311 stp q22, q23, [x0, #CTX_FP_Q22]
312 stp q24, q25, [x0, #CTX_FP_Q24]
313 stp q26, q27, [x0, #CTX_FP_Q26]
314 stp q28, q29, [x0, #CTX_FP_Q28]
315 stp q30, q31, [x0, #CTX_FP_Q30]
316
317 mrs x9, fpsr
318 str x9, [x0, #CTX_FP_FPSR]
319
320 mrs x10, fpcr
321 str x10, [x0, #CTX_FP_FPCR]
322
323 ret
324
325/* -----------------------------------------------------
326 * The following function follows the aapcs_64 strictly
327 * to use x9-x17 (temporary caller-saved registers
328 * according to AArch64 PCS) to restore floating point
329 * register context. It assumes that 'x0' is pointing to
330 * a 'fp_regs' structure from where the register context
331 * will be restored.
332 *
333 * Access to VFP registers will trap if CPTR_EL3.TFP is
334 * set. However currently we don't use VFP registers
335 * nor set traps in Trusted Firmware, and assume it's
336 * cleared
337 *
338 * TODO: Revisit when VFP is used in secure world
339 * -----------------------------------------------------
340 */
341 .global fpregs_context_restore
Andrew Thoelke38bde412014-03-18 13:46:55 +0000342func fpregs_context_restore
Achin Gupta9ac63c52014-01-16 12:08:03 +0000343 ldp q0, q1, [x0, #CTX_FP_Q0]
344 ldp q2, q3, [x0, #CTX_FP_Q2]
345 ldp q4, q5, [x0, #CTX_FP_Q4]
346 ldp q6, q7, [x0, #CTX_FP_Q6]
347 ldp q8, q9, [x0, #CTX_FP_Q8]
348 ldp q10, q11, [x0, #CTX_FP_Q10]
349 ldp q12, q13, [x0, #CTX_FP_Q12]
350 ldp q14, q15, [x0, #CTX_FP_Q14]
351 ldp q16, q17, [x0, #CTX_FP_Q16]
352 ldp q18, q19, [x0, #CTX_FP_Q18]
353 ldp q20, q21, [x0, #CTX_FP_Q20]
354 ldp q22, q23, [x0, #CTX_FP_Q22]
355 ldp q24, q25, [x0, #CTX_FP_Q24]
356 ldp q26, q27, [x0, #CTX_FP_Q26]
357 ldp q28, q29, [x0, #CTX_FP_Q28]
358 ldp q30, q31, [x0, #CTX_FP_Q30]
359
360 ldr x9, [x0, #CTX_FP_FPSR]
361 msr fpsr, x9
362
363 str x10, [x0, #CTX_FP_FPCR]
364 msr fpcr, x10
365
366 /*
367 * No explict ISB required here as ERET to
368 * swtich to secure EL1 or non-secure world
369 * covers it
370 */
371
372 ret