blob: cd824973249ca6430d5255c5b322338f655a9c11 [file] [log] [blame]
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +00001/*
Dimitris Papastamosb63c6f12018-01-11 15:29:36 +00002 * Copyright (c) 2017-2018, ARM Limited and Contributors. All rights reserved.
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +00003 *
4 * SPDX-License-Identifier: BSD-3-Clause
5 */
6
7#include <arch.h>
Dimitris Papastamos28803632018-01-08 13:57:39 +00008#include <arm_arch_svc.h>
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +00009#include <asm_macros.S>
10#include <context.h>
11
12 .globl workaround_bpiall_vbar0_runtime_exceptions
13
14#define EMIT_BPIALL 0xee070fd5
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +000015#define EMIT_SMC 0xe1600070
Dimitris Papastamos28803632018-01-08 13:57:39 +000016#define ESR_EL3_A64_SMC0 0x5e000000
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +000017
Dimitris Papastamosb63c6f12018-01-11 15:29:36 +000018 .macro enter_workaround _from_vector
19 /*
20 * Save register state to enable a call to AArch32 S-EL1 and return
21 * Identify the original calling vector in w2 (==_from_vector)
22 * Use w3-w6 for additional register state preservation while in S-EL1
23 */
24
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +000025 /* Save GP regs */
26 stp x0, x1, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X0]
27 stp x2, x3, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X2]
28 stp x4, x5, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X4]
29 stp x6, x7, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X6]
30 stp x8, x9, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X8]
31 stp x10, x11, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X10]
32 stp x12, x13, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X12]
33 stp x14, x15, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X14]
34 stp x16, x17, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X16]
35 stp x18, x19, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X18]
36 stp x20, x21, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X20]
37 stp x22, x23, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X22]
38 stp x24, x25, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X24]
39 stp x26, x27, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X26]
40 stp x28, x29, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X28]
41
Dimitris Papastamosb63c6f12018-01-11 15:29:36 +000042 /* Identify the original exception vector */
43 mov w2, \_from_vector
44
45 /* Preserve 32-bit system registers in GP registers through the workaround */
46 mrs x3, esr_el3
47 mrs x4, spsr_el3
48 mrs x5, scr_el3
49 mrs x6, sctlr_el1
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +000050
51 /*
Dimitris Papastamosb63c6f12018-01-11 15:29:36 +000052 * Preserve LR and ELR_EL3 registers in the GP regs context.
53 * Temporarily use the CTX_GPREG_SP_EL0 slot to preserve ELR_EL3
54 * through the workaround. This is OK because at this point the
55 * current state for this context's SP_EL0 is in the live system
56 * register, which is unmodified by the workaround.
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +000057 */
Dimitris Papastamosb63c6f12018-01-11 15:29:36 +000058 mrs x7, elr_el3
59 stp x30, x7, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_LR]
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +000060
61 /*
Dimitris Papastamosb63c6f12018-01-11 15:29:36 +000062 * Load system registers for entry to S-EL1.
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +000063 */
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +000064
Dimitris Papastamosb63c6f12018-01-11 15:29:36 +000065 /* Mask all interrupts and set AArch32 Supervisor mode */
66 movz w8, SPSR_MODE32(MODE32_svc, SPSR_T_ARM, SPSR_E_LITTLE, SPSR_AIF_MASK)
67
68 /* Switch EL3 exception vectors while the workaround is executing. */
69 adr x9, workaround_bpiall_vbar1_runtime_exceptions
70
71 /* Setup SCTLR_EL1 with MMU off and I$ on */
72 ldr x10, stub_sel1_sctlr
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +000073
Dimitris Papastamosb63c6f12018-01-11 15:29:36 +000074 /* Land at the S-EL1 workaround stub */
75 adr x11, aarch32_stub
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +000076
77 /*
78 * Setting SCR_EL3 to all zeroes means that the NS, RW
79 * and SMD bits are configured as expected.
80 */
81 msr scr_el3, xzr
Dimitris Papastamosb63c6f12018-01-11 15:29:36 +000082 msr spsr_el3, x8
83 msr vbar_el3, x9
84 msr sctlr_el1, x10
85 msr elr_el3, x11
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +000086
87 eret
88 .endm
89
90 /* ---------------------------------------------------------------------
91 * This vector table is used at runtime to enter the workaround at
92 * AArch32 S-EL1 for Sync/IRQ/FIQ/SError exceptions. If the workaround
93 * is not enabled, the existing runtime exception vector table is used.
94 * ---------------------------------------------------------------------
95 */
96vector_base workaround_bpiall_vbar0_runtime_exceptions
97
98 /* ---------------------------------------------------------------------
99 * Current EL with SP_EL0 : 0x0 - 0x200
100 * ---------------------------------------------------------------------
101 */
102vector_entry workaround_bpiall_vbar0_sync_exception_sp_el0
103 b sync_exception_sp_el0
Dimitris Papastamosb63c6f12018-01-11 15:29:36 +0000104 nop /* to force 8 byte alignment for the following stub */
105
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +0000106 /*
107 * Since each vector table entry is 128 bytes, we can store the
108 * stub context in the unused space to minimize memory footprint.
109 */
Dimitris Papastamosb63c6f12018-01-11 15:29:36 +0000110stub_sel1_sctlr:
111 .quad SCTLR_AARCH32_EL1_RES1 | SCTLR_I_BIT
112
113aarch32_stub:
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +0000114 .word EMIT_BPIALL
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +0000115 .word EMIT_SMC
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +0000116
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +0000117 check_vector_size workaround_bpiall_vbar0_sync_exception_sp_el0
118
119vector_entry workaround_bpiall_vbar0_irq_sp_el0
120 b irq_sp_el0
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +0000121 check_vector_size workaround_bpiall_vbar0_irq_sp_el0
122
123vector_entry workaround_bpiall_vbar0_fiq_sp_el0
124 b fiq_sp_el0
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +0000125 check_vector_size workaround_bpiall_vbar0_fiq_sp_el0
126
127vector_entry workaround_bpiall_vbar0_serror_sp_el0
128 b serror_sp_el0
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +0000129 check_vector_size workaround_bpiall_vbar0_serror_sp_el0
130
131 /* ---------------------------------------------------------------------
132 * Current EL with SP_ELx: 0x200 - 0x400
133 * ---------------------------------------------------------------------
134 */
135vector_entry workaround_bpiall_vbar0_sync_exception_sp_elx
136 b sync_exception_sp_elx
137 check_vector_size workaround_bpiall_vbar0_sync_exception_sp_elx
138
139vector_entry workaround_bpiall_vbar0_irq_sp_elx
140 b irq_sp_elx
141 check_vector_size workaround_bpiall_vbar0_irq_sp_elx
142
143vector_entry workaround_bpiall_vbar0_fiq_sp_elx
144 b fiq_sp_elx
145 check_vector_size workaround_bpiall_vbar0_fiq_sp_elx
146
147vector_entry workaround_bpiall_vbar0_serror_sp_elx
148 b serror_sp_elx
149 check_vector_size workaround_bpiall_vbar0_serror_sp_elx
150
151 /* ---------------------------------------------------------------------
152 * Lower EL using AArch64 : 0x400 - 0x600
153 * ---------------------------------------------------------------------
154 */
155vector_entry workaround_bpiall_vbar0_sync_exception_aarch64
Dimitris Papastamosb63c6f12018-01-11 15:29:36 +0000156 enter_workaround 1
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +0000157 check_vector_size workaround_bpiall_vbar0_sync_exception_aarch64
158
159vector_entry workaround_bpiall_vbar0_irq_aarch64
Dimitris Papastamosb63c6f12018-01-11 15:29:36 +0000160 enter_workaround 2
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +0000161 check_vector_size workaround_bpiall_vbar0_irq_aarch64
162
163vector_entry workaround_bpiall_vbar0_fiq_aarch64
Dimitris Papastamosb63c6f12018-01-11 15:29:36 +0000164 enter_workaround 4
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +0000165 check_vector_size workaround_bpiall_vbar0_fiq_aarch64
166
167vector_entry workaround_bpiall_vbar0_serror_aarch64
Dimitris Papastamosb63c6f12018-01-11 15:29:36 +0000168 enter_workaround 8
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +0000169 check_vector_size workaround_bpiall_vbar0_serror_aarch64
170
171 /* ---------------------------------------------------------------------
172 * Lower EL using AArch32 : 0x600 - 0x800
173 * ---------------------------------------------------------------------
174 */
175vector_entry workaround_bpiall_vbar0_sync_exception_aarch32
Dimitris Papastamosb63c6f12018-01-11 15:29:36 +0000176 enter_workaround 1
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +0000177 check_vector_size workaround_bpiall_vbar0_sync_exception_aarch32
178
179vector_entry workaround_bpiall_vbar0_irq_aarch32
Dimitris Papastamosb63c6f12018-01-11 15:29:36 +0000180 enter_workaround 2
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +0000181 check_vector_size workaround_bpiall_vbar0_irq_aarch32
182
183vector_entry workaround_bpiall_vbar0_fiq_aarch32
Dimitris Papastamosb63c6f12018-01-11 15:29:36 +0000184 enter_workaround 4
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +0000185 check_vector_size workaround_bpiall_vbar0_fiq_aarch32
186
187vector_entry workaround_bpiall_vbar0_serror_aarch32
Dimitris Papastamosb63c6f12018-01-11 15:29:36 +0000188 enter_workaround 8
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +0000189 check_vector_size workaround_bpiall_vbar0_serror_aarch32
190
191 /* ---------------------------------------------------------------------
192 * This vector table is used while the workaround is executing. It
193 * installs a simple SMC handler to allow the Sync/IRQ/FIQ/SError
194 * workaround stubs to enter EL3 from S-EL1. It restores the previous
195 * EL3 state before proceeding with the normal runtime exception vector.
196 * ---------------------------------------------------------------------
197 */
198vector_base workaround_bpiall_vbar1_runtime_exceptions
199
200 /* ---------------------------------------------------------------------
201 * Current EL with SP_EL0 : 0x0 - 0x200 (UNUSED)
202 * ---------------------------------------------------------------------
203 */
204vector_entry workaround_bpiall_vbar1_sync_exception_sp_el0
205 b report_unhandled_exception
206 check_vector_size workaround_bpiall_vbar1_sync_exception_sp_el0
207
208vector_entry workaround_bpiall_vbar1_irq_sp_el0
209 b report_unhandled_interrupt
210 check_vector_size workaround_bpiall_vbar1_irq_sp_el0
211
212vector_entry workaround_bpiall_vbar1_fiq_sp_el0
213 b report_unhandled_interrupt
214 check_vector_size workaround_bpiall_vbar1_fiq_sp_el0
215
216vector_entry workaround_bpiall_vbar1_serror_sp_el0
217 b report_unhandled_exception
218 check_vector_size workaround_bpiall_vbar1_serror_sp_el0
219
220 /* ---------------------------------------------------------------------
221 * Current EL with SP_ELx: 0x200 - 0x400 (UNUSED)
222 * ---------------------------------------------------------------------
223 */
224vector_entry workaround_bpiall_vbar1_sync_exception_sp_elx
225 b report_unhandled_exception
226 check_vector_size workaround_bpiall_vbar1_sync_exception_sp_elx
227
228vector_entry workaround_bpiall_vbar1_irq_sp_elx
229 b report_unhandled_interrupt
230 check_vector_size workaround_bpiall_vbar1_irq_sp_elx
231
232vector_entry workaround_bpiall_vbar1_fiq_sp_elx
233 b report_unhandled_interrupt
234 check_vector_size workaround_bpiall_vbar1_fiq_sp_elx
235
236vector_entry workaround_bpiall_vbar1_serror_sp_elx
237 b report_unhandled_exception
238 check_vector_size workaround_bpiall_vbar1_serror_sp_elx
239
240 /* ---------------------------------------------------------------------
241 * Lower EL using AArch64 : 0x400 - 0x600 (UNUSED)
242 * ---------------------------------------------------------------------
243 */
244vector_entry workaround_bpiall_vbar1_sync_exception_aarch64
245 b report_unhandled_exception
246 check_vector_size workaround_bpiall_vbar1_sync_exception_aarch64
247
248vector_entry workaround_bpiall_vbar1_irq_aarch64
249 b report_unhandled_interrupt
250 check_vector_size workaround_bpiall_vbar1_irq_aarch64
251
252vector_entry workaround_bpiall_vbar1_fiq_aarch64
253 b report_unhandled_interrupt
254 check_vector_size workaround_bpiall_vbar1_fiq_aarch64
255
256vector_entry workaround_bpiall_vbar1_serror_aarch64
257 b report_unhandled_exception
258 check_vector_size workaround_bpiall_vbar1_serror_aarch64
259
260 /* ---------------------------------------------------------------------
261 * Lower EL using AArch32 : 0x600 - 0x800
262 * ---------------------------------------------------------------------
263 */
264vector_entry workaround_bpiall_vbar1_sync_exception_aarch32
Dimitris Papastamosb63c6f12018-01-11 15:29:36 +0000265 /*
266 * w2 indicates which SEL1 stub was run and thus which original vector was used
267 * w3-w6 contain saved system register state (esr_el3 in w3)
268 * Restore LR and ELR_EL3 register state from the GP regs context
269 */
270 ldp x30, x7, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_LR]
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +0000271
272 /* Apply the restored system register state */
Dimitris Papastamosb63c6f12018-01-11 15:29:36 +0000273 msr esr_el3, x3
274 msr spsr_el3, x4
275 msr scr_el3, x5
276 msr sctlr_el1, x6
277 msr elr_el3, x7
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +0000278
279 /*
280 * Workaround is complete, so swap VBAR_EL3 to point
281 * to workaround entry table in preparation for subsequent
282 * Sync/IRQ/FIQ/SError exceptions.
283 */
Dimitris Papastamosb63c6f12018-01-11 15:29:36 +0000284 adr x0, workaround_bpiall_vbar0_runtime_exceptions
285 msr vbar_el3, x0
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +0000286
287 /*
Dimitris Papastamosb63c6f12018-01-11 15:29:36 +0000288 * Restore all GP regs except x2 and x3 (esr). The value in x2
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +0000289 * indicates the type of the original exception.
290 */
Dimitris Papastamosb63c6f12018-01-11 15:29:36 +0000291 ldp x0, x1, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X0]
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +0000292 ldp x4, x5, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X4]
293 ldp x6, x7, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X6]
294 ldp x8, x9, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X8]
295 ldp x10, x11, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X10]
296 ldp x12, x13, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X12]
297 ldp x14, x15, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X14]
298 ldp x16, x17, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X16]
299 ldp x18, x19, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X18]
300 ldp x20, x21, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X20]
301 ldp x22, x23, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X22]
302 ldp x24, x25, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X24]
303 ldp x26, x27, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X26]
304 ldp x28, x29, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X28]
305
Dimitris Papastamosb63c6f12018-01-11 15:29:36 +0000306 /* Fast path Sync exceptions. Static predictor will fall through. */
307 tbz w2, #0, workaround_not_sync
Dimitris Papastamos28803632018-01-08 13:57:39 +0000308
309 /*
310 * Check if SMC is coming from A64 state on #0
311 * with W0 = SMCCC_ARCH_WORKAROUND_1
312 *
313 * This sequence evaluates as:
314 * (W0==SMCCC_ARCH_WORKAROUND_1) ? (ESR_EL3==SMC#0) : (NE)
315 * allowing use of a single branch operation
316 */
317 orr w2, wzr, #SMCCC_ARCH_WORKAROUND_1
318 cmp w0, w2
319 mov_imm w2, ESR_EL3_A64_SMC0
320 ccmp w3, w2, #0, eq
321 /* Static predictor will predict a fall through */
322 bne 1f
323 eret
3241:
Dimitris Papastamosb63c6f12018-01-11 15:29:36 +0000325 ldp x2, x3, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X2]
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +0000326 b sync_exception_aarch64
327 check_vector_size workaround_bpiall_vbar1_sync_exception_aarch32
328
329vector_entry workaround_bpiall_vbar1_irq_aarch32
330 b report_unhandled_interrupt
Dimitris Papastamosb63c6f12018-01-11 15:29:36 +0000331
332 /*
333 * Post-workaround fan-out for non-sync exceptions
334 */
335workaround_not_sync:
336 tbnz w2, #3, workaround_bpiall_vbar1_serror
337 tbnz w2, #2, workaround_bpiall_vbar1_fiq
338 /* IRQ */
339 ldp x2, x3, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X2]
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +0000340 b irq_aarch64
Dimitris Papastamosb63c6f12018-01-11 15:29:36 +0000341
342workaround_bpiall_vbar1_fiq:
343 ldp x2, x3, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X2]
344 b fiq_aarch64
345
346workaround_bpiall_vbar1_serror:
347 ldp x2, x3, [sp, #CTX_GPREGS_OFFSET + CTX_GPREG_X2]
348 b serror_aarch64
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +0000349 check_vector_size workaround_bpiall_vbar1_irq_aarch32
350
351vector_entry workaround_bpiall_vbar1_fiq_aarch32
352 b report_unhandled_interrupt
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +0000353 check_vector_size workaround_bpiall_vbar1_fiq_aarch32
354
355vector_entry workaround_bpiall_vbar1_serror_aarch32
356 b report_unhandled_exception
Dimitris Papastamosc52ebdc2017-12-18 13:46:21 +0000357 check_vector_size workaround_bpiall_vbar1_serror_aarch32