blob: bf3ae1e7de090c15a86e9ec1d1e6b79858726476 [file] [log] [blame]
Antonio Nino Diazf1b84f62018-07-03 11:58:49 +01001/*
2 * Copyright (c) 2017-2018, ARM Limited and Contributors. All rights reserved.
3 *
4 * SPDX-License-Identifier: BSD-3-Clause
5 */
6
Antonio Nino Diaz9d596c42018-07-12 15:43:07 +01007#include <assert.h>
Antonio Nino Diaze0f90632018-12-14 00:18:21 +00008
Antonio Nino Diazf1b84f62018-07-03 11:58:49 +01009#include <platform_def.h>
Antonio Nino Diaze0f90632018-12-14 00:18:21 +000010
11#include <common/debug.h>
12#include <lib/xlat_tables/xlat_tables_defs.h>
13#include <lib/xlat_tables/xlat_tables_v2.h>
Antonio Nino Diazf1b84f62018-07-03 11:58:49 +010014
15#include "xlat_tables_private.h"
16
17/*
Antonio Nino Diaz67f799e2018-07-15 16:42:01 +010018 * MMU configuration register values for the active translation context. Used
19 * from the MMU assembly helpers.
20 */
21uint64_t mmu_cfg_params[MMU_CFG_PARAM_MAX];
22
23/*
Antonio Nino Diazf1b84f62018-07-03 11:58:49 +010024 * Allocate and initialise the default translation context for the BL image
25 * currently executing.
26 */
27REGISTER_XLAT_CONTEXT(tf, MAX_MMAP_REGIONS, MAX_XLAT_TABLES,
28 PLAT_VIRT_ADDR_SPACE_SIZE, PLAT_PHY_ADDR_SPACE_SIZE);
29
30void mmap_add_region(unsigned long long base_pa, uintptr_t base_va, size_t size,
31 unsigned int attr)
32{
33 mmap_region_t mm = MAP_REGION(base_pa, base_va, size, attr);
34
35 mmap_add_region_ctx(&tf_xlat_ctx, &mm);
36}
37
38void mmap_add(const mmap_region_t *mm)
39{
40 mmap_add_ctx(&tf_xlat_ctx, mm);
41}
42
Antonio Nino Diazc0033282018-11-20 16:03:11 +000043void mmap_add_region_alloc_va(unsigned long long base_pa, uintptr_t *base_va,
44 size_t size, unsigned int attr)
45{
46 mmap_region_t mm = MAP_REGION_ALLOC_VA(base_pa, size, attr);
47
48 mmap_add_region_alloc_va_ctx(&tf_xlat_ctx, &mm);
49
50 *base_va = mm.base_va;
51}
52
53void mmap_add_alloc_va(mmap_region_t *mm)
54{
55 while (mm->granularity != 0U) {
56 assert(mm->base_va == 0U);
57 mmap_add_region_alloc_va_ctx(&tf_xlat_ctx, mm);
58 mm++;
59 }
60}
61
Antonio Nino Diazf1b84f62018-07-03 11:58:49 +010062#if PLAT_XLAT_TABLES_DYNAMIC
63
64int mmap_add_dynamic_region(unsigned long long base_pa, uintptr_t base_va,
65 size_t size, unsigned int attr)
66{
67 mmap_region_t mm = MAP_REGION(base_pa, base_va, size, attr);
68
69 return mmap_add_dynamic_region_ctx(&tf_xlat_ctx, &mm);
70}
71
Antonio Nino Diazc0033282018-11-20 16:03:11 +000072int mmap_add_dynamic_region_alloc_va(unsigned long long base_pa,
73 uintptr_t *base_va, size_t size,
74 unsigned int attr)
75{
76 mmap_region_t mm = MAP_REGION_ALLOC_VA(base_pa, size, attr);
77
78 int rc = mmap_add_dynamic_region_alloc_va_ctx(&tf_xlat_ctx, &mm);
79
80 *base_va = mm.base_va;
81
82 return rc;
83}
84
85
Antonio Nino Diazf1b84f62018-07-03 11:58:49 +010086int mmap_remove_dynamic_region(uintptr_t base_va, size_t size)
87{
88 return mmap_remove_dynamic_region_ctx(&tf_xlat_ctx,
89 base_va, size);
90}
91
92#endif /* PLAT_XLAT_TABLES_DYNAMIC */
93
Daniel Boulby5a03a252018-08-30 16:48:56 +010094void __init init_xlat_tables(void)
Antonio Nino Diazf1b84f62018-07-03 11:58:49 +010095{
Antonio Nino Diaz9d596c42018-07-12 15:43:07 +010096 assert(tf_xlat_ctx.xlat_regime == EL_REGIME_INVALID);
97
Antonio Nino Diaz50eb3742018-07-24 10:20:53 +010098 unsigned int current_el = xlat_arch_current_el();
Antonio Nino Diaz9d596c42018-07-12 15:43:07 +010099
Antonio Nino Diaz50eb3742018-07-24 10:20:53 +0100100 if (current_el == 1U) {
Antonio Nino Diaz9d596c42018-07-12 15:43:07 +0100101 tf_xlat_ctx.xlat_regime = EL1_EL0_REGIME;
Antonio Nino Diaz128de8d2018-08-07 19:59:49 +0100102 } else if (current_el == 2U) {
103 tf_xlat_ctx.xlat_regime = EL2_REGIME;
Antonio Nino Diaz9d596c42018-07-12 15:43:07 +0100104 } else {
Antonio Nino Diaz50eb3742018-07-24 10:20:53 +0100105 assert(current_el == 3U);
Antonio Nino Diaz9d596c42018-07-12 15:43:07 +0100106 tf_xlat_ctx.xlat_regime = EL3_REGIME;
107 }
108
Antonio Nino Diazf1b84f62018-07-03 11:58:49 +0100109 init_xlat_tables_ctx(&tf_xlat_ctx);
110}
111
Antonio Nino Diaz6c4c9ee2018-08-05 15:34:10 +0100112int xlat_get_mem_attributes(uintptr_t base_va, uint32_t *attr)
113{
114 return xlat_get_mem_attributes_ctx(&tf_xlat_ctx, base_va, attr);
115}
116
117int xlat_change_mem_attributes(uintptr_t base_va, size_t size, uint32_t attr)
118{
119 return xlat_change_mem_attributes_ctx(&tf_xlat_ctx, base_va, size, attr);
120}
121
Antonio Nino Diazf1b84f62018-07-03 11:58:49 +0100122/*
123 * If dynamic allocation of new regions is disabled then by the time we call the
124 * function enabling the MMU, we'll have registered all the memory regions to
125 * map for the system's lifetime. Therefore, at this point we know the maximum
126 * physical address that will ever be mapped.
127 *
128 * If dynamic allocation is enabled then we can't make any such assumption
129 * because the maximum physical address could get pushed while adding a new
130 * region. Therefore, in this case we have to assume that the whole address
131 * space size might be mapped.
132 */
133#ifdef PLAT_XLAT_TABLES_DYNAMIC
134#define MAX_PHYS_ADDR tf_xlat_ctx.pa_max_address
135#else
136#define MAX_PHYS_ADDR tf_xlat_ctx.max_pa
137#endif
138
139#ifdef AARCH32
140
Antonio Nino Diaz128de8d2018-08-07 19:59:49 +0100141void enable_mmu_svc_mon(unsigned int flags)
142{
Antonio Nino Diaz67f799e2018-07-15 16:42:01 +0100143 setup_mmu_cfg((uint64_t *)&mmu_cfg_params, flags,
144 tf_xlat_ctx.base_table, MAX_PHYS_ADDR,
Antonio Nino Diaz9d596c42018-07-12 15:43:07 +0100145 tf_xlat_ctx.va_max_address, EL1_EL0_REGIME);
Antonio Nino Diaz128de8d2018-08-07 19:59:49 +0100146 enable_mmu_direct_svc_mon(flags);
Antonio Nino Diazf1b84f62018-07-03 11:58:49 +0100147}
148
Antonio Nino Diaz128de8d2018-08-07 19:59:49 +0100149void enable_mmu_hyp(unsigned int flags)
150{
151 setup_mmu_cfg((uint64_t *)&mmu_cfg_params, flags,
152 tf_xlat_ctx.base_table, MAX_PHYS_ADDR,
153 tf_xlat_ctx.va_max_address, EL2_REGIME);
154 enable_mmu_direct_hyp(flags);
155}
156
Antonio Nino Diazf1b84f62018-07-03 11:58:49 +0100157#else
158
159void enable_mmu_el1(unsigned int flags)
160{
Antonio Nino Diaz67f799e2018-07-15 16:42:01 +0100161 setup_mmu_cfg((uint64_t *)&mmu_cfg_params, flags,
162 tf_xlat_ctx.base_table, MAX_PHYS_ADDR,
Antonio Nino Diaz9d596c42018-07-12 15:43:07 +0100163 tf_xlat_ctx.va_max_address, EL1_EL0_REGIME);
Antonio Nino Diazf1b84f62018-07-03 11:58:49 +0100164 enable_mmu_direct_el1(flags);
165}
166
Antonio Nino Diaz128de8d2018-08-07 19:59:49 +0100167void enable_mmu_el2(unsigned int flags)
168{
169 setup_mmu_cfg((uint64_t *)&mmu_cfg_params, flags,
170 tf_xlat_ctx.base_table, MAX_PHYS_ADDR,
171 tf_xlat_ctx.va_max_address, EL2_REGIME);
172 enable_mmu_direct_el2(flags);
173}
174
Antonio Nino Diazf1b84f62018-07-03 11:58:49 +0100175void enable_mmu_el3(unsigned int flags)
176{
Antonio Nino Diaz67f799e2018-07-15 16:42:01 +0100177 setup_mmu_cfg((uint64_t *)&mmu_cfg_params, flags,
178 tf_xlat_ctx.base_table, MAX_PHYS_ADDR,
Antonio Nino Diaz9d596c42018-07-12 15:43:07 +0100179 tf_xlat_ctx.va_max_address, EL3_REGIME);
Antonio Nino Diazf1b84f62018-07-03 11:58:49 +0100180 enable_mmu_direct_el3(flags);
181}
182
183#endif /* AARCH32 */