blob: 5fe8bab05780c98b7009399b232defb1c0e53eb6 [file] [log] [blame]
Achin Gupta9ac63c52014-01-16 12:08:03 +00001/*
2 * Copyright (c) 2013-2014, ARM Limited and Contributors. All rights reserved.
3 *
4 * Redistribution and use in source and binary forms, with or without
5 * modification, are permitted provided that the following conditions are met:
6 *
7 * Redistributions of source code must retain the above copyright notice, this
8 * list of conditions and the following disclaimer.
9 *
10 * Redistributions in binary form must reproduce the above copyright notice,
11 * this list of conditions and the following disclaimer in the documentation
12 * and/or other materials provided with the distribution.
13 *
14 * Neither the name of ARM nor the names of its contributors may be used
15 * to endorse or promote products derived from this software without specific
16 * prior written permission.
17 *
18 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
19 * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
20 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
21 * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE
22 * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
23 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
24 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
25 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
26 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
27 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
28 * POSSIBILITY OF SUCH DAMAGE.
29 */
30
31#include <context.h>
32
33/* -----------------------------------------------------
34 * The following function strictly follows the AArch64
35 * PCS to use x9-x17 (temporary caller-saved registers)
36 * to save essential EL3 system register context. It
37 * assumes that 'x0' is pointing to a 'el1_sys_regs'
38 * structure where the register context will be saved.
39 * -----------------------------------------------------
40 */
41 .global el3_sysregs_context_save
42el3_sysregs_context_save:
43
44 mrs x9, scr_el3
45 mrs x10, sctlr_el3
46 stp x9, x10, [x0, #CTX_SCR_EL3]
47
48 mrs x11, cptr_el3
49 stp x11, xzr, [x0, #CTX_CPTR_EL3]
50
51 mrs x13, cntfrq_el0
52 mrs x14, mair_el3
53 stp x13, x14, [x0, #CTX_CNTFRQ_EL0]
54
55 mrs x15, tcr_el3
56 mrs x16, ttbr0_el3
57 stp x15, x16, [x0, #CTX_TCR_EL3]
58
59 mrs x17, daif
60 and x17, x17, #(DAIF_ABT_BIT | DAIF_DBG_BIT)
61 stp x17, xzr, [x0, #CTX_DAIF_EL3]
62
63 ret
64
65/* -----------------------------------------------------
66 * The following function strictly follows the AArch64
67 * PCS to use x9-x17 (temporary caller-saved registers)
68 * to restore essential EL3 system register context. It
69 * assumes that 'x0' is pointing to a 'el1_sys_regs'
70 * structure from where the register context will be
71 * restored.
72 *
73 * Note that the sequence differs from that of the save
74 * function as we want the MMU to be enabled last
75 * -----------------------------------------------------
76 */
77 .global el3_sysregs_context_restore
78el3_sysregs_context_restore:
79
80 ldp x11, xzr, [x0, #CTX_CPTR_EL3]
81 msr cptr_el3, x11
82
83 ldp x13, x14, [x0, #CTX_CNTFRQ_EL0]
84 msr cntfrq_el0, x13
85 msr mair_el3, x14
86
87 ldp x15, x16, [x0, #CTX_TCR_EL3]
88 msr tcr_el3, x15
89 msr ttbr0_el3, x16
90
91 ldp x17, xzr, [x0, #CTX_DAIF_EL3]
92 mrs x11, daif
93 orr x17, x17, x11
94 msr daif, x17
95
96 /* Make sure all the above changes are observed */
97 isb
98
99 ldp x9, x10, [x0, #CTX_SCR_EL3]
100 msr scr_el3, x9
101 msr sctlr_el3, x10
102 isb
103
104 ret
105
106/* -----------------------------------------------------
107 * The following function strictly follows the AArch64
108 * PCS to use x9-x17 (temporary caller-saved registers)
109 * to save EL1 system register context. It assumes that
110 * 'x0' is pointing to a 'el1_sys_regs' structure where
111 * the register context will be saved.
112 * -----------------------------------------------------
113 */
114 .global el1_sysregs_context_save
115el1_sysregs_context_save:
116
117 mrs x9, spsr_el1
118 mrs x10, elr_el1
119 stp x9, x10, [x0, #CTX_SPSR_EL1]
120
121 mrs x11, spsr_abt
122 mrs x12, spsr_und
123 stp x11, x12, [x0, #CTX_SPSR_ABT]
124
125 mrs x13, spsr_irq
126 mrs x14, spsr_fiq
127 stp x13, x14, [x0, #CTX_SPSR_IRQ]
128
129 mrs x15, sctlr_el1
130 mrs x16, actlr_el1
131 stp x15, x16, [x0, #CTX_SCTLR_EL1]
132
133 mrs x17, cpacr_el1
134 mrs x9, csselr_el1
135 stp x17, x9, [x0, #CTX_CPACR_EL1]
136
137 mrs x10, sp_el1
138 mrs x11, esr_el1
139 stp x10, x11, [x0, #CTX_SP_EL1]
140
141 mrs x12, ttbr0_el1
142 mrs x13, ttbr1_el1
143 stp x12, x13, [x0, #CTX_TTBR0_EL1]
144
145 mrs x14, mair_el1
146 mrs x15, amair_el1
147 stp x14, x15, [x0, #CTX_MAIR_EL1]
148
149 mrs x16, tcr_el1
150 mrs x17, tpidr_el1
151 stp x16, x17, [x0, #CTX_TCR_EL1]
152
153 mrs x9, tpidr_el0
154 mrs x10, tpidrro_el0
155 stp x9, x10, [x0, #CTX_TPIDR_EL0]
156
157 mrs x11, dacr32_el2
158 mrs x12, ifsr32_el2
159 stp x11, x12, [x0, #CTX_DACR32_EL2]
160
161 mrs x13, par_el1
162 mrs x14, far_el1
163 stp x13, x14, [x0, #CTX_PAR_EL1]
164
165 mrs x15, afsr0_el1
166 mrs x16, afsr1_el1
167 stp x15, x16, [x0, #CTX_AFSR0_EL1]
168
169 mrs x17, contextidr_el1
170 mrs x9, vbar_el1
171 stp x17, x9, [x0, #CTX_CONTEXTIDR_EL1]
172
173 mrs x10, cntp_ctl_el0
174 mrs x11, cntp_cval_el0
175 stp x10, x11, [x0, #CTX_CNTP_CTL_EL0]
176
177 mrs x12, cntv_ctl_el0
178 mrs x13, cntv_cval_el0
179 stp x12, x13, [x0, #CTX_CNTV_CTL_EL0]
180
181 mrs x14, cntkctl_el1
182 mrs x15, fpexc32_el2
183 stp x14, x15, [x0, #CTX_CNTKCTL_EL1]
184
185 ret
186
187/* -----------------------------------------------------
188 * The following function strictly follows the AArch64
189 * PCS to use x9-x17 (temporary caller-saved registers)
190 * to restore EL1 system register context. It assumes
191 * that 'x0' is pointing to a 'el1_sys_regs' structure
192 * from where the register context will be restored
193 * -----------------------------------------------------
194 */
195 .global el1_sysregs_context_restore
196el1_sysregs_context_restore:
197
198 ldp x9, x10, [x0, #CTX_SPSR_EL1]
199 msr spsr_el1, x9
200 msr elr_el1, x10
201
202 ldp x11, x12, [x0, #CTX_SPSR_ABT]
203 msr spsr_abt, x11
204 msr spsr_und, x12
205
206 ldp x13, x14, [x0, #CTX_SPSR_IRQ]
207 msr spsr_irq, x13
208 msr spsr_fiq, x14
209
210 ldp x15, x16, [x0, #CTX_SCTLR_EL1]
211 msr sctlr_el1, x15
212 msr actlr_el1, x16
213
214 ldp x17, x9, [x0, #CTX_CPACR_EL1]
215 msr cpacr_el1, x17
216 msr csselr_el1, x9
217
218 ldp x10, x11, [x0, #CTX_SP_EL1]
219 msr sp_el1, x10
220 msr esr_el1, x11
221
222 ldp x12, x13, [x0, #CTX_TTBR0_EL1]
223 msr ttbr0_el1, x12
224 msr ttbr1_el1, x13
225
226 ldp x14, x15, [x0, #CTX_MAIR_EL1]
227 msr mair_el1, x14
228 msr amair_el1, x15
229
230 ldp x16, x17, [x0, #CTX_TCR_EL1]
231 msr tcr_el1, x16
232 msr tpidr_el1, x17
233
234 ldp x9, x10, [x0, #CTX_TPIDR_EL0]
235 msr tpidr_el0, x9
236 msr tpidrro_el0, x10
237
238 ldp x11, x12, [x0, #CTX_DACR32_EL2]
239 msr dacr32_el2, x11
240 msr ifsr32_el2, x12
241
242 ldp x13, x14, [x0, #CTX_PAR_EL1]
243 msr par_el1, x13
244 msr far_el1, x14
245
246 ldp x15, x16, [x0, #CTX_AFSR0_EL1]
247 msr afsr0_el1, x15
248 msr afsr1_el1, x16
249
250 ldp x17, x9, [x0, #CTX_CONTEXTIDR_EL1]
251 msr contextidr_el1, x17
252 msr vbar_el1, x9
253
254 ldp x10, x11, [x0, #CTX_CNTP_CTL_EL0]
255 msr cntp_ctl_el0, x10
256 msr cntp_cval_el0, x11
257
258 ldp x12, x13, [x0, #CTX_CNTV_CTL_EL0]
259 msr cntv_ctl_el0, x12
260 msr cntv_cval_el0, x13
261
262 ldp x14, x15, [x0, #CTX_CNTKCTL_EL1]
263 msr cntkctl_el1, x14
264 msr fpexc32_el2, x15
265
266 /* No explict ISB required here as ERET covers it */
267
268 ret
269
270/* -----------------------------------------------------
271 * The followsing function follows the aapcs_64 strictly
272 * to use x9-x17 (temporary caller-saved registers
273 * according to AArch64 PCS) to save floating point
274 * register context. It assumes that 'x0' is pointing to
275 * a 'fp_regs' structure where the register context will
276 * be saved.
277 *
278 * Access to VFP registers will trap if CPTR_EL3.TFP is
279 * set. However currently we don't use VFP registers
280 * nor set traps in Trusted Firmware, and assume it's
281 * cleared
282 *
283 * TODO: Revisit when VFP is used in secure world
284 * -----------------------------------------------------
285 */
286 .global fpregs_context_save
287fpregs_context_save:
288 stp q0, q1, [x0, #CTX_FP_Q0]
289 stp q2, q3, [x0, #CTX_FP_Q2]
290 stp q4, q5, [x0, #CTX_FP_Q4]
291 stp q6, q7, [x0, #CTX_FP_Q6]
292 stp q8, q9, [x0, #CTX_FP_Q8]
293 stp q10, q11, [x0, #CTX_FP_Q10]
294 stp q12, q13, [x0, #CTX_FP_Q12]
295 stp q14, q15, [x0, #CTX_FP_Q14]
296 stp q16, q17, [x0, #CTX_FP_Q16]
297 stp q18, q19, [x0, #CTX_FP_Q18]
298 stp q20, q21, [x0, #CTX_FP_Q20]
299 stp q22, q23, [x0, #CTX_FP_Q22]
300 stp q24, q25, [x0, #CTX_FP_Q24]
301 stp q26, q27, [x0, #CTX_FP_Q26]
302 stp q28, q29, [x0, #CTX_FP_Q28]
303 stp q30, q31, [x0, #CTX_FP_Q30]
304
305 mrs x9, fpsr
306 str x9, [x0, #CTX_FP_FPSR]
307
308 mrs x10, fpcr
309 str x10, [x0, #CTX_FP_FPCR]
310
311 ret
312
313/* -----------------------------------------------------
314 * The following function follows the aapcs_64 strictly
315 * to use x9-x17 (temporary caller-saved registers
316 * according to AArch64 PCS) to restore floating point
317 * register context. It assumes that 'x0' is pointing to
318 * a 'fp_regs' structure from where the register context
319 * will be restored.
320 *
321 * Access to VFP registers will trap if CPTR_EL3.TFP is
322 * set. However currently we don't use VFP registers
323 * nor set traps in Trusted Firmware, and assume it's
324 * cleared
325 *
326 * TODO: Revisit when VFP is used in secure world
327 * -----------------------------------------------------
328 */
329 .global fpregs_context_restore
330fpregs_context_restore:
331 ldp q0, q1, [x0, #CTX_FP_Q0]
332 ldp q2, q3, [x0, #CTX_FP_Q2]
333 ldp q4, q5, [x0, #CTX_FP_Q4]
334 ldp q6, q7, [x0, #CTX_FP_Q6]
335 ldp q8, q9, [x0, #CTX_FP_Q8]
336 ldp q10, q11, [x0, #CTX_FP_Q10]
337 ldp q12, q13, [x0, #CTX_FP_Q12]
338 ldp q14, q15, [x0, #CTX_FP_Q14]
339 ldp q16, q17, [x0, #CTX_FP_Q16]
340 ldp q18, q19, [x0, #CTX_FP_Q18]
341 ldp q20, q21, [x0, #CTX_FP_Q20]
342 ldp q22, q23, [x0, #CTX_FP_Q22]
343 ldp q24, q25, [x0, #CTX_FP_Q24]
344 ldp q26, q27, [x0, #CTX_FP_Q26]
345 ldp q28, q29, [x0, #CTX_FP_Q28]
346 ldp q30, q31, [x0, #CTX_FP_Q30]
347
348 ldr x9, [x0, #CTX_FP_FPSR]
349 msr fpsr, x9
350
351 str x10, [x0, #CTX_FP_FPCR]
352 msr fpcr, x10
353
354 /*
355 * No explict ISB required here as ERET to
356 * swtich to secure EL1 or non-secure world
357 * covers it
358 */
359
360 ret