| /* |
| * Copyright (c) 2013-2024, Arm Limited and Contributors. All rights reserved. |
| * |
| * SPDX-License-Identifier: BSD-3-Clause |
| */ |
| |
| #ifndef CONTEXT_H |
| #define CONTEXT_H |
| |
| #include <lib/el3_runtime/context_el2.h> |
| #include <lib/el3_runtime/cpu_data.h> |
| #include <lib/utils_def.h> |
| |
| /******************************************************************************* |
| * Constants that allow assembler code to access members of and the 'gp_regs' |
| * structure at their correct offsets. |
| ******************************************************************************/ |
| #define CTX_GPREGS_OFFSET U(0x0) |
| #define CTX_GPREG_X0 U(0x0) |
| #define CTX_GPREG_X1 U(0x8) |
| #define CTX_GPREG_X2 U(0x10) |
| #define CTX_GPREG_X3 U(0x18) |
| #define CTX_GPREG_X4 U(0x20) |
| #define CTX_GPREG_X5 U(0x28) |
| #define CTX_GPREG_X6 U(0x30) |
| #define CTX_GPREG_X7 U(0x38) |
| #define CTX_GPREG_X8 U(0x40) |
| #define CTX_GPREG_X9 U(0x48) |
| #define CTX_GPREG_X10 U(0x50) |
| #define CTX_GPREG_X11 U(0x58) |
| #define CTX_GPREG_X12 U(0x60) |
| #define CTX_GPREG_X13 U(0x68) |
| #define CTX_GPREG_X14 U(0x70) |
| #define CTX_GPREG_X15 U(0x78) |
| #define CTX_GPREG_X16 U(0x80) |
| #define CTX_GPREG_X17 U(0x88) |
| #define CTX_GPREG_X18 U(0x90) |
| #define CTX_GPREG_X19 U(0x98) |
| #define CTX_GPREG_X20 U(0xa0) |
| #define CTX_GPREG_X21 U(0xa8) |
| #define CTX_GPREG_X22 U(0xb0) |
| #define CTX_GPREG_X23 U(0xb8) |
| #define CTX_GPREG_X24 U(0xc0) |
| #define CTX_GPREG_X25 U(0xc8) |
| #define CTX_GPREG_X26 U(0xd0) |
| #define CTX_GPREG_X27 U(0xd8) |
| #define CTX_GPREG_X28 U(0xe0) |
| #define CTX_GPREG_X29 U(0xe8) |
| #define CTX_GPREG_LR U(0xf0) |
| #define CTX_GPREG_SP_EL0 U(0xf8) |
| #define CTX_GPREGS_END U(0x100) |
| |
| /******************************************************************************* |
| * Constants that allow assembler code to access members of and the 'el3_state' |
| * structure at their correct offsets. Note that some of the registers are only |
| * 32-bits wide but are stored as 64-bit values for convenience |
| ******************************************************************************/ |
| #define CTX_EL3STATE_OFFSET (CTX_GPREGS_OFFSET + CTX_GPREGS_END) |
| #define CTX_SCR_EL3 U(0x0) |
| #define CTX_ESR_EL3 U(0x8) |
| #define CTX_RUNTIME_SP U(0x10) |
| #define CTX_SPSR_EL3 U(0x18) |
| #define CTX_ELR_EL3 U(0x20) |
| #define CTX_PMCR_EL0 U(0x28) |
| #define CTX_IS_IN_EL3 U(0x30) |
| #define CTX_MDCR_EL3 U(0x38) |
| /* Constants required in supporting nested exception in EL3 */ |
| #define CTX_SAVED_ELR_EL3 U(0x40) |
| /* |
| * General purpose flag, to save various EL3 states |
| * FFH mode : Used to identify if handling nested exception |
| * KFH mode : Used as counter value |
| */ |
| #define CTX_NESTED_EA_FLAG U(0x48) |
| #if FFH_SUPPORT |
| #define CTX_SAVED_ESR_EL3 U(0x50) |
| #define CTX_SAVED_SPSR_EL3 U(0x58) |
| #define CTX_SAVED_GPREG_LR U(0x60) |
| #define CTX_EL3STATE_END U(0x70) /* Align to the next 16 byte boundary */ |
| #else |
| #define CTX_EL3STATE_END U(0x50) /* Align to the next 16 byte boundary */ |
| #endif /* FFH_SUPPORT */ |
| |
| /******************************************************************************* |
| * Constants that allow assembler code to access members of and the |
| * 'el1_sys_regs' structure at their correct offsets. Note that some of the |
| * registers are only 32-bits wide but are stored as 64-bit values for |
| * convenience |
| ******************************************************************************/ |
| #define CTX_EL1_SYSREGS_OFFSET (CTX_EL3STATE_OFFSET + CTX_EL3STATE_END) |
| #define CTX_SPSR_EL1 U(0x0) |
| #define CTX_ELR_EL1 U(0x8) |
| #define CTX_SCTLR_EL1 U(0x10) |
| #define CTX_TCR_EL1 U(0x18) |
| #define CTX_CPACR_EL1 U(0x20) |
| #define CTX_CSSELR_EL1 U(0x28) |
| #define CTX_SP_EL1 U(0x30) |
| #define CTX_ESR_EL1 U(0x38) |
| #define CTX_TTBR0_EL1 U(0x40) |
| #define CTX_TTBR1_EL1 U(0x48) |
| #define CTX_MAIR_EL1 U(0x50) |
| #define CTX_AMAIR_EL1 U(0x58) |
| #define CTX_ACTLR_EL1 U(0x60) |
| #define CTX_TPIDR_EL1 U(0x68) |
| #define CTX_TPIDR_EL0 U(0x70) |
| #define CTX_TPIDRRO_EL0 U(0x78) |
| #define CTX_PAR_EL1 U(0x80) |
| #define CTX_FAR_EL1 U(0x88) |
| #define CTX_AFSR0_EL1 U(0x90) |
| #define CTX_AFSR1_EL1 U(0x98) |
| #define CTX_CONTEXTIDR_EL1 U(0xa0) |
| #define CTX_VBAR_EL1 U(0xa8) |
| #define CTX_MDCCINT_EL1 U(0xb0) |
| #define CTX_MDSCR_EL1 U(0xb8) |
| |
| #define CTX_AARCH64_END U(0xc0) /* Align to the next 16 byte boundary */ |
| |
| /* |
| * If the platform is AArch64-only, there is no need to save and restore these |
| * AArch32 registers. |
| */ |
| #if CTX_INCLUDE_AARCH32_REGS |
| #define CTX_SPSR_ABT (CTX_AARCH64_END + U(0x0)) |
| #define CTX_SPSR_UND (CTX_AARCH64_END + U(0x8)) |
| #define CTX_SPSR_IRQ (CTX_AARCH64_END + U(0x10)) |
| #define CTX_SPSR_FIQ (CTX_AARCH64_END + U(0x18)) |
| #define CTX_DACR32_EL2 (CTX_AARCH64_END + U(0x20)) |
| #define CTX_IFSR32_EL2 (CTX_AARCH64_END + U(0x28)) |
| #define CTX_AARCH32_END (CTX_AARCH64_END + U(0x30)) /* Align to the next 16 byte boundary */ |
| #else |
| #define CTX_AARCH32_END CTX_AARCH64_END |
| #endif /* CTX_INCLUDE_AARCH32_REGS */ |
| |
| /* |
| * If the timer registers aren't saved and restored, we don't have to reserve |
| * space for them in the context |
| */ |
| #if NS_TIMER_SWITCH |
| #define CTX_CNTP_CTL_EL0 (CTX_AARCH32_END + U(0x0)) |
| #define CTX_CNTP_CVAL_EL0 (CTX_AARCH32_END + U(0x8)) |
| #define CTX_CNTV_CTL_EL0 (CTX_AARCH32_END + U(0x10)) |
| #define CTX_CNTV_CVAL_EL0 (CTX_AARCH32_END + U(0x18)) |
| #define CTX_CNTKCTL_EL1 (CTX_AARCH32_END + U(0x20)) |
| #define CTX_TIMER_SYSREGS_END (CTX_AARCH32_END + U(0x30)) /* Align to the next 16 byte boundary */ |
| #else |
| #define CTX_TIMER_SYSREGS_END CTX_AARCH32_END |
| #endif /* NS_TIMER_SWITCH */ |
| |
| #if ENABLE_FEAT_MTE2 |
| #define CTX_TFSRE0_EL1 (CTX_TIMER_SYSREGS_END + U(0x0)) |
| #define CTX_TFSR_EL1 (CTX_TIMER_SYSREGS_END + U(0x8)) |
| #define CTX_RGSR_EL1 (CTX_TIMER_SYSREGS_END + U(0x10)) |
| #define CTX_GCR_EL1 (CTX_TIMER_SYSREGS_END + U(0x18)) |
| #define CTX_MTE_REGS_END (CTX_TIMER_SYSREGS_END + U(0x20)) /* Align to the next 16 byte boundary */ |
| #else |
| #define CTX_MTE_REGS_END CTX_TIMER_SYSREGS_END |
| #endif /* ENABLE_FEAT_MTE2 */ |
| |
| #if ENABLE_FEAT_RAS |
| #define CTX_DISR_EL1 (CTX_MTE_REGS_END + U(0x0)) |
| #define CTX_RAS_REGS_END (CTX_MTE_REGS_END + U(0x10)) /* Align to the next 16 byte boundary */ |
| #else |
| #define CTX_RAS_REGS_END CTX_MTE_REGS_END |
| #endif /* ENABLE_FEAT_RAS */ |
| |
| #if ENABLE_FEAT_S1PIE |
| #define CTX_PIRE0_EL1 (CTX_RAS_REGS_END + U(0x0)) |
| #define CTX_PIR_EL1 (CTX_RAS_REGS_END + U(0x8)) |
| #define CTX_S1PIE_REGS_END (CTX_RAS_REGS_END + U(0x10)) /* Align to the next 16 byte boundary */ |
| #else |
| #define CTX_S1PIE_REGS_END CTX_RAS_REGS_END |
| #endif /* ENABLE_FEAT_S1PIE */ |
| |
| #if ENABLE_FEAT_S1POE |
| #define CTX_POR_EL1 (CTX_S1PIE_REGS_END + U(0x0)) |
| #define CTX_S1POE_REGS_END (CTX_S1PIE_REGS_END + U(0x10)) /* Align to the next 16 byte boundary */ |
| #else |
| #define CTX_S1POE_REGS_END CTX_S1PIE_REGS_END |
| #endif /* ENABLE_FEAT_S1POE */ |
| |
| #if ENABLE_FEAT_S2POE |
| #define CTX_S2POR_EL1 (CTX_S1POE_REGS_END + U(0x0)) |
| #define CTX_S2POE_REGS_END (CTX_S1POE_REGS_END + U(0x10)) /* Align to the next 16 byte boundary */ |
| #else |
| #define CTX_S2POE_REGS_END CTX_S1POE_REGS_END |
| #endif /* ENABLE_FEAT_S2POE */ |
| |
| #if ENABLE_FEAT_TCR2 |
| #define CTX_TCR2_EL1 (CTX_S2POE_REGS_END + U(0x0)) |
| #define CTX_TCR2_REGS_END (CTX_S2POE_REGS_END + U(0x10)) /* Align to the next 16 byte boundary */ |
| #else |
| #define CTX_TCR2_REGS_END CTX_S2POE_REGS_END |
| #endif /* ENABLE_FEAT_TCR2 */ |
| |
| #if ENABLE_TRF_FOR_NS |
| #define CTX_TRFCR_EL1 (CTX_TCR2_REGS_END + U(0x0)) |
| #define CTX_TRF_REGS_END (CTX_TCR2_REGS_END + U(0x10)) /* Align to the next 16 byte boundary */ |
| #else |
| #define CTX_TRF_REGS_END CTX_TCR2_REGS_END |
| #endif /* ENABLE_TRF_FOR_NS */ |
| |
| #if ENABLE_FEAT_CSV2_2 |
| #define CTX_SCXTNUM_EL0 (CTX_TRF_REGS_END + U(0x0)) |
| #define CTX_SCXTNUM_EL1 (CTX_TRF_REGS_END + U(0x8)) |
| #define CTX_CSV2_2_REGS_END (CTX_TRF_REGS_END + U(0x10)) /* Align to the next 16 byte boundary */ |
| #else |
| #define CTX_CSV2_2_REGS_END CTX_TRF_REGS_END |
| #endif /* ENABLE_FEAT_CSV2_2 */ |
| |
| #if ENABLE_FEAT_GCS |
| #define CTX_GCSCR_EL1 (CTX_CSV2_2_REGS_END + U(0x0)) |
| #define CTX_GCSCRE0_EL1 (CTX_CSV2_2_REGS_END + U(0x8)) |
| #define CTX_GCSPR_EL1 (CTX_CSV2_2_REGS_END + U(0x10)) |
| #define CTX_GCSPR_EL0 (CTX_CSV2_2_REGS_END + U(0x18)) |
| #define CTX_GCS_REGS_END (CTX_CSV2_2_REGS_END + U(0x20)) /* Align to the next 16 byte boundary */ |
| #else |
| #define CTX_GCS_REGS_END CTX_CSV2_2_REGS_END |
| #endif /* ENABLE_FEAT_GCS */ |
| |
| /* |
| * End of EL1 system registers. |
| */ |
| #define CTX_EL1_SYSREGS_END CTX_GCS_REGS_END |
| |
| /******************************************************************************* |
| * Constants that allow assembler code to access members of and the 'fp_regs' |
| * structure at their correct offsets. |
| ******************************************************************************/ |
| # define CTX_FPREGS_OFFSET (CTX_EL1_SYSREGS_OFFSET + CTX_EL1_SYSREGS_END) |
| #if CTX_INCLUDE_FPREGS |
| #define CTX_FP_Q0 U(0x0) |
| #define CTX_FP_Q1 U(0x10) |
| #define CTX_FP_Q2 U(0x20) |
| #define CTX_FP_Q3 U(0x30) |
| #define CTX_FP_Q4 U(0x40) |
| #define CTX_FP_Q5 U(0x50) |
| #define CTX_FP_Q6 U(0x60) |
| #define CTX_FP_Q7 U(0x70) |
| #define CTX_FP_Q8 U(0x80) |
| #define CTX_FP_Q9 U(0x90) |
| #define CTX_FP_Q10 U(0xa0) |
| #define CTX_FP_Q11 U(0xb0) |
| #define CTX_FP_Q12 U(0xc0) |
| #define CTX_FP_Q13 U(0xd0) |
| #define CTX_FP_Q14 U(0xe0) |
| #define CTX_FP_Q15 U(0xf0) |
| #define CTX_FP_Q16 U(0x100) |
| #define CTX_FP_Q17 U(0x110) |
| #define CTX_FP_Q18 U(0x120) |
| #define CTX_FP_Q19 U(0x130) |
| #define CTX_FP_Q20 U(0x140) |
| #define CTX_FP_Q21 U(0x150) |
| #define CTX_FP_Q22 U(0x160) |
| #define CTX_FP_Q23 U(0x170) |
| #define CTX_FP_Q24 U(0x180) |
| #define CTX_FP_Q25 U(0x190) |
| #define CTX_FP_Q26 U(0x1a0) |
| #define CTX_FP_Q27 U(0x1b0) |
| #define CTX_FP_Q28 U(0x1c0) |
| #define CTX_FP_Q29 U(0x1d0) |
| #define CTX_FP_Q30 U(0x1e0) |
| #define CTX_FP_Q31 U(0x1f0) |
| #define CTX_FP_FPSR U(0x200) |
| #define CTX_FP_FPCR U(0x208) |
| #if CTX_INCLUDE_AARCH32_REGS |
| #define CTX_FP_FPEXC32_EL2 U(0x210) |
| #define CTX_FPREGS_END U(0x220) /* Align to the next 16 byte boundary */ |
| #else |
| #define CTX_FPREGS_END U(0x210) /* Align to the next 16 byte boundary */ |
| #endif /* CTX_INCLUDE_AARCH32_REGS */ |
| #else |
| #define CTX_FPREGS_END U(0) |
| #endif /* CTX_INCLUDE_FPREGS */ |
| |
| /******************************************************************************* |
| * Registers related to CVE-2018-3639 |
| ******************************************************************************/ |
| #define CTX_CVE_2018_3639_OFFSET (CTX_FPREGS_OFFSET + CTX_FPREGS_END) |
| #define CTX_CVE_2018_3639_DISABLE U(0) |
| #define CTX_CVE_2018_3639_END U(0x10) /* Align to the next 16 byte boundary */ |
| |
| /******************************************************************************* |
| * Registers related to ARMv8.3-PAuth. |
| ******************************************************************************/ |
| #define CTX_PAUTH_REGS_OFFSET (CTX_CVE_2018_3639_OFFSET + CTX_CVE_2018_3639_END) |
| #if CTX_INCLUDE_PAUTH_REGS |
| #define CTX_PACIAKEY_LO U(0x0) |
| #define CTX_PACIAKEY_HI U(0x8) |
| #define CTX_PACIBKEY_LO U(0x10) |
| #define CTX_PACIBKEY_HI U(0x18) |
| #define CTX_PACDAKEY_LO U(0x20) |
| #define CTX_PACDAKEY_HI U(0x28) |
| #define CTX_PACDBKEY_LO U(0x30) |
| #define CTX_PACDBKEY_HI U(0x38) |
| #define CTX_PACGAKEY_LO U(0x40) |
| #define CTX_PACGAKEY_HI U(0x48) |
| #define CTX_PAUTH_REGS_END U(0x50) /* Align to the next 16 byte boundary */ |
| #else |
| #define CTX_PAUTH_REGS_END U(0) |
| #endif /* CTX_INCLUDE_PAUTH_REGS */ |
| |
| /******************************************************************************* |
| * Registers initialised in a per-world context. |
| ******************************************************************************/ |
| #define CTX_CPTR_EL3 U(0x0) |
| #define CTX_ZCR_EL3 U(0x8) |
| #define CTX_MPAM3_EL3 U(0x10) |
| #define CTX_PERWORLD_EL3STATE_END U(0x18) |
| |
| #ifndef __ASSEMBLER__ |
| |
| #include <stdint.h> |
| |
| #include <lib/cassert.h> |
| |
| /* |
| * Common constants to help define the 'cpu_context' structure and its |
| * members below. |
| */ |
| #define DWORD_SHIFT U(3) |
| #define DEFINE_REG_STRUCT(name, num_regs) \ |
| typedef struct name { \ |
| uint64_t ctx_regs[num_regs]; \ |
| } __aligned(16) name##_t |
| |
| /* Constants to determine the size of individual context structures */ |
| #define CTX_GPREG_ALL (CTX_GPREGS_END >> DWORD_SHIFT) |
| #define CTX_EL1_SYSREGS_ALL (CTX_EL1_SYSREGS_END >> DWORD_SHIFT) |
| |
| #if CTX_INCLUDE_FPREGS |
| # define CTX_FPREG_ALL (CTX_FPREGS_END >> DWORD_SHIFT) |
| #endif |
| #define CTX_EL3STATE_ALL (CTX_EL3STATE_END >> DWORD_SHIFT) |
| #define CTX_CVE_2018_3639_ALL (CTX_CVE_2018_3639_END >> DWORD_SHIFT) |
| #if CTX_INCLUDE_PAUTH_REGS |
| # define CTX_PAUTH_REGS_ALL (CTX_PAUTH_REGS_END >> DWORD_SHIFT) |
| #endif |
| |
| /* |
| * AArch64 general purpose register context structure. Usually x0-x18, |
| * lr are saved as the compiler is expected to preserve the remaining |
| * callee saved registers if used by the C runtime and the assembler |
| * does not touch the remaining. But in case of world switch during |
| * exception handling, we need to save the callee registers too. |
| */ |
| DEFINE_REG_STRUCT(gp_regs, CTX_GPREG_ALL); |
| |
| /* |
| * AArch64 EL1 system register context structure for preserving the |
| * architectural state during world switches. |
| */ |
| DEFINE_REG_STRUCT(el1_sysregs, CTX_EL1_SYSREGS_ALL); |
| |
| /* |
| * AArch64 floating point register context structure for preserving |
| * the floating point state during switches from one security state to |
| * another. |
| */ |
| #if CTX_INCLUDE_FPREGS |
| DEFINE_REG_STRUCT(fp_regs, CTX_FPREG_ALL); |
| #endif |
| |
| /* |
| * Miscellaneous registers used by EL3 firmware to maintain its state |
| * across exception entries and exits |
| */ |
| DEFINE_REG_STRUCT(el3_state, CTX_EL3STATE_ALL); |
| |
| /* Function pointer used by CVE-2018-3639 dynamic mitigation */ |
| DEFINE_REG_STRUCT(cve_2018_3639, CTX_CVE_2018_3639_ALL); |
| |
| /* Registers associated to ARMv8.3-PAuth */ |
| #if CTX_INCLUDE_PAUTH_REGS |
| DEFINE_REG_STRUCT(pauth, CTX_PAUTH_REGS_ALL); |
| #endif |
| |
| /* |
| * Macros to access members of any of the above structures using their |
| * offsets |
| */ |
| #define read_ctx_reg(ctx, offset) ((ctx)->ctx_regs[(offset) >> DWORD_SHIFT]) |
| #define write_ctx_reg(ctx, offset, val) (((ctx)->ctx_regs[(offset) >> DWORD_SHIFT]) \ |
| = (uint64_t) (val)) |
| |
| /* |
| * Top-level context structure which is used by EL3 firmware to preserve |
| * the state of a core at the next lower EL in a given security state and |
| * save enough EL3 meta data to be able to return to that EL and security |
| * state. The context management library will be used to ensure that |
| * SP_EL3 always points to an instance of this structure at exception |
| * entry and exit. |
| */ |
| typedef struct cpu_context { |
| gp_regs_t gpregs_ctx; |
| el3_state_t el3state_ctx; |
| el1_sysregs_t el1_sysregs_ctx; |
| |
| #if CTX_INCLUDE_FPREGS |
| fp_regs_t fpregs_ctx; |
| #endif |
| cve_2018_3639_t cve_2018_3639_ctx; |
| |
| #if CTX_INCLUDE_PAUTH_REGS |
| pauth_t pauth_ctx; |
| #endif |
| |
| #if CTX_INCLUDE_EL2_REGS |
| el2_sysregs_t el2_sysregs_ctx; |
| #endif |
| |
| } cpu_context_t; |
| |
| /* |
| * Per-World Context. |
| * It stores registers whose values can be shared across CPUs. |
| */ |
| typedef struct per_world_context { |
| uint64_t ctx_cptr_el3; |
| uint64_t ctx_zcr_el3; |
| uint64_t ctx_mpam3_el3; |
| } per_world_context_t; |
| |
| extern per_world_context_t per_world_context[CPU_DATA_CONTEXT_NUM]; |
| |
| /* Macros to access members of the 'cpu_context_t' structure */ |
| #define get_el3state_ctx(h) (&((cpu_context_t *) h)->el3state_ctx) |
| #if CTX_INCLUDE_FPREGS |
| # define get_fpregs_ctx(h) (&((cpu_context_t *) h)->fpregs_ctx) |
| #endif |
| #define get_el1_sysregs_ctx(h) (&((cpu_context_t *) h)->el1_sysregs_ctx) |
| #if CTX_INCLUDE_EL2_REGS |
| # define get_el2_sysregs_ctx(h) (&((cpu_context_t *) h)->el2_sysregs_ctx) |
| #endif |
| #define get_gpregs_ctx(h) (&((cpu_context_t *) h)->gpregs_ctx) |
| #define get_cve_2018_3639_ctx(h) (&((cpu_context_t *) h)->cve_2018_3639_ctx) |
| #if CTX_INCLUDE_PAUTH_REGS |
| # define get_pauth_ctx(h) (&((cpu_context_t *) h)->pauth_ctx) |
| #endif |
| |
| /* |
| * Compile time assertions related to the 'cpu_context' structure to |
| * ensure that the assembler and the compiler view of the offsets of |
| * the structure members is the same. |
| */ |
| CASSERT(CTX_GPREGS_OFFSET == __builtin_offsetof(cpu_context_t, gpregs_ctx), |
| assert_core_context_gp_offset_mismatch); |
| |
| CASSERT(CTX_EL3STATE_OFFSET == __builtin_offsetof(cpu_context_t, el3state_ctx), |
| assert_core_context_el3state_offset_mismatch); |
| |
| CASSERT(CTX_EL1_SYSREGS_OFFSET == __builtin_offsetof(cpu_context_t, el1_sysregs_ctx), |
| assert_core_context_el1_sys_offset_mismatch); |
| |
| #if CTX_INCLUDE_FPREGS |
| CASSERT(CTX_FPREGS_OFFSET == __builtin_offsetof(cpu_context_t, fpregs_ctx), |
| assert_core_context_fp_offset_mismatch); |
| #endif /* CTX_INCLUDE_FPREGS */ |
| |
| CASSERT(CTX_CVE_2018_3639_OFFSET == __builtin_offsetof(cpu_context_t, cve_2018_3639_ctx), |
| assert_core_context_cve_2018_3639_offset_mismatch); |
| |
| #if CTX_INCLUDE_PAUTH_REGS |
| CASSERT(CTX_PAUTH_REGS_OFFSET == __builtin_offsetof(cpu_context_t, pauth_ctx), |
| assert_core_context_pauth_offset_mismatch); |
| #endif /* CTX_INCLUDE_PAUTH_REGS */ |
| |
| /* |
| * Helper macro to set the general purpose registers that correspond to |
| * parameters in an aapcs_64 call i.e. x0-x7 |
| */ |
| #define set_aapcs_args0(ctx, x0) do { \ |
| write_ctx_reg(get_gpregs_ctx(ctx), CTX_GPREG_X0, x0); \ |
| } while (0) |
| #define set_aapcs_args1(ctx, x0, x1) do { \ |
| write_ctx_reg(get_gpregs_ctx(ctx), CTX_GPREG_X1, x1); \ |
| set_aapcs_args0(ctx, x0); \ |
| } while (0) |
| #define set_aapcs_args2(ctx, x0, x1, x2) do { \ |
| write_ctx_reg(get_gpregs_ctx(ctx), CTX_GPREG_X2, x2); \ |
| set_aapcs_args1(ctx, x0, x1); \ |
| } while (0) |
| #define set_aapcs_args3(ctx, x0, x1, x2, x3) do { \ |
| write_ctx_reg(get_gpregs_ctx(ctx), CTX_GPREG_X3, x3); \ |
| set_aapcs_args2(ctx, x0, x1, x2); \ |
| } while (0) |
| #define set_aapcs_args4(ctx, x0, x1, x2, x3, x4) do { \ |
| write_ctx_reg(get_gpregs_ctx(ctx), CTX_GPREG_X4, x4); \ |
| set_aapcs_args3(ctx, x0, x1, x2, x3); \ |
| } while (0) |
| #define set_aapcs_args5(ctx, x0, x1, x2, x3, x4, x5) do { \ |
| write_ctx_reg(get_gpregs_ctx(ctx), CTX_GPREG_X5, x5); \ |
| set_aapcs_args4(ctx, x0, x1, x2, x3, x4); \ |
| } while (0) |
| #define set_aapcs_args6(ctx, x0, x1, x2, x3, x4, x5, x6) do { \ |
| write_ctx_reg(get_gpregs_ctx(ctx), CTX_GPREG_X6, x6); \ |
| set_aapcs_args5(ctx, x0, x1, x2, x3, x4, x5); \ |
| } while (0) |
| #define set_aapcs_args7(ctx, x0, x1, x2, x3, x4, x5, x6, x7) do { \ |
| write_ctx_reg(get_gpregs_ctx(ctx), CTX_GPREG_X7, x7); \ |
| set_aapcs_args6(ctx, x0, x1, x2, x3, x4, x5, x6); \ |
| } while (0) |
| |
| /******************************************************************************* |
| * Function prototypes |
| ******************************************************************************/ |
| #if CTX_INCLUDE_FPREGS |
| void fpregs_context_save(fp_regs_t *regs); |
| void fpregs_context_restore(fp_regs_t *regs); |
| #endif |
| |
| #endif /* __ASSEMBLER__ */ |
| |
| #endif /* CONTEXT_H */ |