blob: 7f1d3958a2cfa2266adf48f256c8088282543ff2 [file] [log] [blame]
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +00001/*
Antonio Nino Diaz3f518922018-01-05 11:30:36 +00002 * Copyright (c) 2017-2018, ARM Limited and Contributors. All rights reserved.
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +00003 *
dp-armfa3cf0b2017-05-03 09:38:09 +01004 * SPDX-License-Identifier: BSD-3-Clause
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +00005 */
6
7#include <arch.h>
8#include <arch_helpers.h>
9#include <assert.h>
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +000010#include <common_def.h>
11#include <debug.h>
12#include <errno.h>
13#include <platform_def.h>
14#include <string.h>
15#include <types.h>
16#include <utils.h>
Antonio Nino Diazdcf9d922017-10-04 16:52:15 +010017#include <xlat_tables_arch_private.h>
Sandrine Bailleux090c8492017-05-19 09:59:37 +010018#include <xlat_tables_defs.h>
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +000019#include <xlat_tables_v2.h>
Sandrine Bailleux090c8492017-05-19 09:59:37 +010020
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +000021#include "xlat_tables_private.h"
22
Sandrine Bailleux66342932017-07-18 13:26:36 +010023/*
24 * Each platform can define the size of its physical and virtual address spaces.
25 * If the platform hasn't defined one or both of them, default to
26 * ADDR_SPACE_SIZE. The latter is deprecated, though.
27 */
28#if ERROR_DEPRECATED
29# ifdef ADDR_SPACE_SIZE
30# error "ADDR_SPACE_SIZE is deprecated. Use PLAT_xxx_ADDR_SPACE_SIZE instead."
31# endif
32#elif defined(ADDR_SPACE_SIZE)
33# ifndef PLAT_PHY_ADDR_SPACE_SIZE
34# define PLAT_PHY_ADDR_SPACE_SIZE ADDR_SPACE_SIZE
35# endif
36# ifndef PLAT_VIRT_ADDR_SPACE_SIZE
37# define PLAT_VIRT_ADDR_SPACE_SIZE ADDR_SPACE_SIZE
38# endif
39#endif
40
41/*
42 * Allocate and initialise the default translation context for the BL image
43 * currently executing.
44 */
45REGISTER_XLAT_CONTEXT(tf, MAX_MMAP_REGIONS, MAX_XLAT_TABLES,
46 PLAT_VIRT_ADDR_SPACE_SIZE, PLAT_PHY_ADDR_SPACE_SIZE);
47
Antonio Nino Diazac998032017-02-27 17:23:54 +000048#if PLAT_XLAT_TABLES_DYNAMIC
49
50/*
51 * The following functions assume that they will be called using subtables only.
52 * The base table can't be unmapped, so it is not needed to do any special
53 * handling for it.
54 */
55
56/*
57 * Returns the index of the array corresponding to the specified translation
58 * table.
59 */
60static int xlat_table_get_index(xlat_ctx_t *ctx, const uint64_t *table)
61{
Varun Wadekar66231d12017-06-07 09:57:42 -070062 for (unsigned int i = 0; i < ctx->tables_num; i++)
Antonio Nino Diazac998032017-02-27 17:23:54 +000063 if (ctx->tables[i] == table)
64 return i;
65
66 /*
67 * Maybe we were asked to get the index of the base level table, which
68 * should never happen.
69 */
70 assert(0);
71
72 return -1;
73}
74
75/* Returns a pointer to an empty translation table. */
76static uint64_t *xlat_table_get_empty(xlat_ctx_t *ctx)
77{
Varun Wadekar66231d12017-06-07 09:57:42 -070078 for (unsigned int i = 0; i < ctx->tables_num; i++)
Antonio Nino Diazac998032017-02-27 17:23:54 +000079 if (ctx->tables_mapped_regions[i] == 0)
80 return ctx->tables[i];
81
82 return NULL;
83}
84
85/* Increments region count for a given table. */
86static void xlat_table_inc_regions_count(xlat_ctx_t *ctx, const uint64_t *table)
87{
88 ctx->tables_mapped_regions[xlat_table_get_index(ctx, table)]++;
89}
90
91/* Decrements region count for a given table. */
92static void xlat_table_dec_regions_count(xlat_ctx_t *ctx, const uint64_t *table)
93{
94 ctx->tables_mapped_regions[xlat_table_get_index(ctx, table)]--;
95}
96
97/* Returns 0 if the speficied table isn't empty, otherwise 1. */
98static int xlat_table_is_empty(xlat_ctx_t *ctx, const uint64_t *table)
99{
100 return !ctx->tables_mapped_regions[xlat_table_get_index(ctx, table)];
101}
102
103#else /* PLAT_XLAT_TABLES_DYNAMIC */
104
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000105/* Returns a pointer to the first empty translation table. */
106static uint64_t *xlat_table_get_empty(xlat_ctx_t *ctx)
107{
108 assert(ctx->next_table < ctx->tables_num);
109
110 return ctx->tables[ctx->next_table++];
111}
112
Antonio Nino Diazac998032017-02-27 17:23:54 +0000113#endif /* PLAT_XLAT_TABLES_DYNAMIC */
114
Antonio Nino Diazdcf9d922017-10-04 16:52:15 +0100115/*
116 * Returns a block/page table descriptor for the given level and attributes.
117 */
Antonio Nino Diaze8811472018-04-17 15:10:18 +0100118static uint64_t xlat_desc(const xlat_ctx_t *ctx, uint32_t attr,
119 unsigned long long addr_pa, int level)
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000120{
121 uint64_t desc;
122 int mem_type;
123
124 /* Make sure that the granularity is fine enough to map this address. */
125 assert((addr_pa & XLAT_BLOCK_MASK(level)) == 0);
126
127 desc = addr_pa;
128 /*
129 * There are different translation table descriptors for level 3 and the
130 * rest.
131 */
132 desc |= (level == XLAT_TABLE_LEVEL_MAX) ? PAGE_DESC : BLOCK_DESC;
133 /*
134 * Always set the access flag, as TF doesn't manage access flag faults.
135 * Deduce other fields of the descriptor based on the MT_NS and MT_RW
136 * memory region attributes.
137 */
Antonio Nino Diazdcf9d922017-10-04 16:52:15 +0100138 desc |= LOWER_ATTRS(ACCESS_FLAG);
139
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000140 desc |= (attr & MT_NS) ? LOWER_ATTRS(NS) : 0;
141 desc |= (attr & MT_RW) ? LOWER_ATTRS(AP_RW) : LOWER_ATTRS(AP_RO);
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000142
143 /*
Antonio Nino Diazdcf9d922017-10-04 16:52:15 +0100144 * Do not allow unprivileged access when the mapping is for a privileged
145 * EL. For translation regimes that do not have mappings for access for
146 * lower exception levels, set AP[2] to AP_NO_ACCESS_UNPRIVILEGED.
147 */
148 if (ctx->xlat_regime == EL1_EL0_REGIME) {
149 if (attr & MT_USER) {
150 /* EL0 mapping requested, so we give User access */
151 desc |= LOWER_ATTRS(AP_ACCESS_UNPRIVILEGED);
152 } else {
153 /* EL1 mapping requested, no User access granted */
154 desc |= LOWER_ATTRS(AP_NO_ACCESS_UNPRIVILEGED);
155 }
156 } else {
157 assert(ctx->xlat_regime == EL3_REGIME);
Antonio Nino Diaz49074492018-04-26 12:59:08 +0100158 desc |= LOWER_ATTRS(AP_ONE_VA_RANGE_RES1);
Antonio Nino Diazdcf9d922017-10-04 16:52:15 +0100159 }
160
161 /*
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000162 * Deduce shareability domain and executability of the memory region
163 * from the memory type of the attributes (MT_TYPE).
164 *
165 * Data accesses to device memory and non-cacheable normal memory are
166 * coherent for all observers in the system, and correspondingly are
167 * always treated as being Outer Shareable. Therefore, for these 2 types
168 * of memory, it is not strictly needed to set the shareability field
169 * in the translation tables.
170 */
171 mem_type = MT_TYPE(attr);
172 if (mem_type == MT_DEVICE) {
173 desc |= LOWER_ATTRS(ATTR_DEVICE_INDEX | OSH);
174 /*
175 * Always map device memory as execute-never.
176 * This is to avoid the possibility of a speculative instruction
177 * fetch, which could be an issue if this memory region
178 * corresponds to a read-sensitive peripheral.
179 */
Antonio Nino Diazdcf9d922017-10-04 16:52:15 +0100180 desc |= xlat_arch_regime_get_xn_desc(ctx->xlat_regime);
Antonio Nino Diazefabaa92017-04-27 13:30:22 +0100181
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000182 } else { /* Normal memory */
183 /*
184 * Always map read-write normal memory as execute-never.
185 * (Trusted Firmware doesn't self-modify its code, therefore
186 * R/W memory is reserved for data storage, which must not be
187 * executable.)
188 * Note that setting the XN bit here is for consistency only.
Antonio Nino Diazefabaa92017-04-27 13:30:22 +0100189 * The function that enables the MMU sets the SCTLR_ELx.WXN bit,
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000190 * which makes any writable memory region to be treated as
191 * execute-never, regardless of the value of the XN bit in the
192 * translation table.
193 *
194 * For read-only memory, rely on the MT_EXECUTE/MT_EXECUTE_NEVER
Antonio Nino Diazdcf9d922017-10-04 16:52:15 +0100195 * attribute to figure out the value of the XN bit. The actual
196 * XN bit(s) to set in the descriptor depends on the context's
197 * translation regime and the policy applied in
198 * xlat_arch_regime_get_xn_desc().
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000199 */
Antonio Nino Diazefabaa92017-04-27 13:30:22 +0100200 if ((attr & MT_RW) || (attr & MT_EXECUTE_NEVER)) {
Antonio Nino Diazdcf9d922017-10-04 16:52:15 +0100201 desc |= xlat_arch_regime_get_xn_desc(ctx->xlat_regime);
Antonio Nino Diazefabaa92017-04-27 13:30:22 +0100202 }
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000203
204 if (mem_type == MT_MEMORY) {
205 desc |= LOWER_ATTRS(ATTR_IWBWA_OWBWA_NTR_INDEX | ISH);
206 } else {
207 assert(mem_type == MT_NON_CACHEABLE);
208 desc |= LOWER_ATTRS(ATTR_NON_CACHEABLE_INDEX | OSH);
209 }
210 }
211
212 return desc;
213}
214
215/*
216 * Enumeration of actions that can be made when mapping table entries depending
217 * on the previous value in that entry and information about the region being
218 * mapped.
219 */
220typedef enum {
221
222 /* Do nothing */
223 ACTION_NONE,
224
225 /* Write a block (or page, if in level 3) entry. */
226 ACTION_WRITE_BLOCK_ENTRY,
227
228 /*
229 * Create a new table and write a table entry pointing to it. Recurse
230 * into it for further processing.
231 */
232 ACTION_CREATE_NEW_TABLE,
233
234 /*
235 * There is a table descriptor in this entry, read it and recurse into
236 * that table for further processing.
237 */
238 ACTION_RECURSE_INTO_TABLE,
239
240} action_t;
241
Antonio Nino Diazac998032017-02-27 17:23:54 +0000242#if PLAT_XLAT_TABLES_DYNAMIC
243
244/*
245 * Recursive function that writes to the translation tables and unmaps the
246 * specified region.
247 */
248static void xlat_tables_unmap_region(xlat_ctx_t *ctx, mmap_region_t *mm,
249 const uintptr_t table_base_va,
250 uint64_t *const table_base,
251 const int table_entries,
Varun Wadekar66231d12017-06-07 09:57:42 -0700252 const unsigned int level)
Antonio Nino Diazac998032017-02-27 17:23:54 +0000253{
254 assert(level >= ctx->base_level && level <= XLAT_TABLE_LEVEL_MAX);
255
256 uint64_t *subtable;
257 uint64_t desc;
258
259 uintptr_t table_idx_va;
260 uintptr_t table_idx_end_va; /* End VA of this entry */
261
262 uintptr_t region_end_va = mm->base_va + mm->size - 1;
263
264 int table_idx;
265
266 if (mm->base_va > table_base_va) {
267 /* Find the first index of the table affected by the region. */
268 table_idx_va = mm->base_va & ~XLAT_BLOCK_MASK(level);
269
270 table_idx = (table_idx_va - table_base_va) >>
271 XLAT_ADDR_SHIFT(level);
272
273 assert(table_idx < table_entries);
274 } else {
275 /* Start from the beginning of the table. */
276 table_idx_va = table_base_va;
277 table_idx = 0;
278 }
279
280 while (table_idx < table_entries) {
281
282 table_idx_end_va = table_idx_va + XLAT_BLOCK_SIZE(level) - 1;
283
284 desc = table_base[table_idx];
285 uint64_t desc_type = desc & DESC_MASK;
286
287 action_t action = ACTION_NONE;
288
289 if ((mm->base_va <= table_idx_va) &&
290 (region_end_va >= table_idx_end_va)) {
291
292 /* Region covers all block */
293
294 if (level == 3) {
295 /*
296 * Last level, only page descriptors allowed,
297 * erase it.
298 */
299 assert(desc_type == PAGE_DESC);
300
301 action = ACTION_WRITE_BLOCK_ENTRY;
302 } else {
303 /*
304 * Other levels can have table descriptors. If
305 * so, recurse into it and erase descriptors
306 * inside it as needed. If there is a block
307 * descriptor, just erase it. If an invalid
308 * descriptor is found, this table isn't
309 * actually mapped, which shouldn't happen.
310 */
311 if (desc_type == TABLE_DESC) {
312 action = ACTION_RECURSE_INTO_TABLE;
313 } else {
314 assert(desc_type == BLOCK_DESC);
315 action = ACTION_WRITE_BLOCK_ENTRY;
316 }
317 }
318
319 } else if ((mm->base_va <= table_idx_end_va) ||
320 (region_end_va >= table_idx_va)) {
321
322 /*
323 * Region partially covers block.
324 *
325 * It can't happen in level 3.
326 *
327 * There must be a table descriptor here, if not there
328 * was a problem when mapping the region.
329 */
330
331 assert(level < 3);
332
333 assert(desc_type == TABLE_DESC);
334
335 action = ACTION_RECURSE_INTO_TABLE;
336 }
337
338 if (action == ACTION_WRITE_BLOCK_ENTRY) {
339
340 table_base[table_idx] = INVALID_DESC;
Antonio Nino Diazdcf9d922017-10-04 16:52:15 +0100341 xlat_arch_tlbi_va_regime(table_idx_va, ctx->xlat_regime);
Antonio Nino Diazac998032017-02-27 17:23:54 +0000342
343 } else if (action == ACTION_RECURSE_INTO_TABLE) {
344
345 subtable = (uint64_t *)(uintptr_t)(desc & TABLE_ADDR_MASK);
346
347 /* Recurse to write into subtable */
348 xlat_tables_unmap_region(ctx, mm, table_idx_va,
349 subtable, XLAT_TABLE_ENTRIES,
350 level + 1);
351
352 /*
353 * If the subtable is now empty, remove its reference.
354 */
355 if (xlat_table_is_empty(ctx, subtable)) {
356 table_base[table_idx] = INVALID_DESC;
Antonio Nino Diazdcf9d922017-10-04 16:52:15 +0100357 xlat_arch_tlbi_va_regime(table_idx_va,
358 ctx->xlat_regime);
Antonio Nino Diazac998032017-02-27 17:23:54 +0000359 }
360
361 } else {
362 assert(action == ACTION_NONE);
363 }
364
365 table_idx++;
366 table_idx_va += XLAT_BLOCK_SIZE(level);
367
368 /* If reached the end of the region, exit */
369 if (region_end_va <= table_idx_va)
370 break;
371 }
372
373 if (level > ctx->base_level)
374 xlat_table_dec_regions_count(ctx, table_base);
375}
376
377#endif /* PLAT_XLAT_TABLES_DYNAMIC */
378
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000379/*
380 * From the given arguments, it decides which action to take when mapping the
381 * specified region.
382 */
383static action_t xlat_tables_map_region_action(const mmap_region_t *mm,
384 const int desc_type, const unsigned long long dest_pa,
Sandrine Bailleux12e86442017-07-19 10:11:13 +0100385 const uintptr_t table_entry_base_va, const unsigned int level)
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000386{
387 uintptr_t mm_end_va = mm->base_va + mm->size - 1;
388 uintptr_t table_entry_end_va =
389 table_entry_base_va + XLAT_BLOCK_SIZE(level) - 1;
390
391 /*
392 * The descriptor types allowed depend on the current table level.
393 */
394
395 if ((mm->base_va <= table_entry_base_va) &&
396 (mm_end_va >= table_entry_end_va)) {
397
398 /*
399 * Table entry is covered by region
400 * --------------------------------
401 *
402 * This means that this table entry can describe the whole
403 * translation with this granularity in principle.
404 */
405
406 if (level == 3) {
407 /*
408 * Last level, only page descriptors are allowed.
409 */
410 if (desc_type == PAGE_DESC) {
411 /*
412 * There's another region mapped here, don't
413 * overwrite.
414 */
415 return ACTION_NONE;
416 } else {
417 assert(desc_type == INVALID_DESC);
418 return ACTION_WRITE_BLOCK_ENTRY;
419 }
420
421 } else {
422
423 /*
424 * Other levels. Table descriptors are allowed. Block
425 * descriptors too, but they have some limitations.
426 */
427
428 if (desc_type == TABLE_DESC) {
429 /* There's already a table, recurse into it. */
430 return ACTION_RECURSE_INTO_TABLE;
431
432 } else if (desc_type == INVALID_DESC) {
433 /*
434 * There's nothing mapped here, create a new
435 * entry.
436 *
437 * Check if the destination granularity allows
438 * us to use a block descriptor or we need a
439 * finer table for it.
440 *
441 * Also, check if the current level allows block
442 * descriptors. If not, create a table instead.
443 */
444 if ((dest_pa & XLAT_BLOCK_MASK(level)) ||
Sandrine Bailleux8f23fa82017-09-28 21:58:12 +0100445 (level < MIN_LVL_BLOCK_DESC) ||
446 (mm->granularity < XLAT_BLOCK_SIZE(level)))
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000447 return ACTION_CREATE_NEW_TABLE;
448 else
449 return ACTION_WRITE_BLOCK_ENTRY;
450
451 } else {
452 /*
453 * There's another region mapped here, don't
454 * overwrite.
455 */
456 assert(desc_type == BLOCK_DESC);
457
458 return ACTION_NONE;
459 }
460 }
461
462 } else if ((mm->base_va <= table_entry_end_va) ||
463 (mm_end_va >= table_entry_base_va)) {
464
465 /*
466 * Region partially covers table entry
467 * -----------------------------------
468 *
469 * This means that this table entry can't describe the whole
470 * translation, a finer table is needed.
471
472 * There cannot be partial block overlaps in level 3. If that
473 * happens, some of the preliminary checks when adding the
474 * mmap region failed to detect that PA and VA must at least be
475 * aligned to PAGE_SIZE.
476 */
477 assert(level < 3);
478
479 if (desc_type == INVALID_DESC) {
480 /*
481 * The block is not fully covered by the region. Create
482 * a new table, recurse into it and try to map the
483 * region with finer granularity.
484 */
485 return ACTION_CREATE_NEW_TABLE;
486
487 } else {
488 assert(desc_type == TABLE_DESC);
489 /*
490 * The block is not fully covered by the region, but
491 * there is already a table here. Recurse into it and
492 * try to map with finer granularity.
493 *
494 * PAGE_DESC for level 3 has the same value as
495 * TABLE_DESC, but this code can't run on a level 3
496 * table because there can't be overlaps in level 3.
497 */
498 return ACTION_RECURSE_INTO_TABLE;
499 }
500 }
501
502 /*
503 * This table entry is outside of the region specified in the arguments,
504 * don't write anything to it.
505 */
506 return ACTION_NONE;
507}
508
509/*
510 * Recursive function that writes to the translation tables and maps the
Antonio Nino Diazac998032017-02-27 17:23:54 +0000511 * specified region. On success, it returns the VA of the last byte that was
512 * succesfully mapped. On error, it returns the VA of the next entry that
513 * should have been mapped.
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000514 */
Antonio Nino Diazac998032017-02-27 17:23:54 +0000515static uintptr_t xlat_tables_map_region(xlat_ctx_t *ctx, mmap_region_t *mm,
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000516 const uintptr_t table_base_va,
517 uint64_t *const table_base,
518 const int table_entries,
Varun Wadekar66231d12017-06-07 09:57:42 -0700519 const unsigned int level)
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000520{
521 assert(level >= ctx->base_level && level <= XLAT_TABLE_LEVEL_MAX);
522
523 uintptr_t mm_end_va = mm->base_va + mm->size - 1;
524
525 uintptr_t table_idx_va;
526 unsigned long long table_idx_pa;
527
528 uint64_t *subtable;
529 uint64_t desc;
530
531 int table_idx;
532
533 if (mm->base_va > table_base_va) {
534 /* Find the first index of the table affected by the region. */
535 table_idx_va = mm->base_va & ~XLAT_BLOCK_MASK(level);
536
537 table_idx = (table_idx_va - table_base_va) >>
538 XLAT_ADDR_SHIFT(level);
539
540 assert(table_idx < table_entries);
541 } else {
542 /* Start from the beginning of the table. */
543 table_idx_va = table_base_va;
544 table_idx = 0;
545 }
546
Antonio Nino Diazac998032017-02-27 17:23:54 +0000547#if PLAT_XLAT_TABLES_DYNAMIC
548 if (level > ctx->base_level)
549 xlat_table_inc_regions_count(ctx, table_base);
550#endif
551
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000552 while (table_idx < table_entries) {
553
554 desc = table_base[table_idx];
555
556 table_idx_pa = mm->base_pa + table_idx_va - mm->base_va;
557
558 action_t action = xlat_tables_map_region_action(mm,
559 desc & DESC_MASK, table_idx_pa, table_idx_va, level);
560
561 if (action == ACTION_WRITE_BLOCK_ENTRY) {
562
563 table_base[table_idx] =
Antonio Nino Diaze8811472018-04-17 15:10:18 +0100564 xlat_desc(ctx, (uint32_t)mm->attr, table_idx_pa,
565 level);
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000566
567 } else if (action == ACTION_CREATE_NEW_TABLE) {
568
569 subtable = xlat_table_get_empty(ctx);
Antonio Nino Diazac998032017-02-27 17:23:54 +0000570 if (subtable == NULL) {
571 /* Not enough free tables to map this region */
572 return table_idx_va;
573 }
574
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000575 /* Point to new subtable from this one. */
Antonio Nino Diazac998032017-02-27 17:23:54 +0000576 table_base[table_idx] = TABLE_DESC | (unsigned long)subtable;
577
578 /* Recurse to write into subtable */
579 uintptr_t end_va = xlat_tables_map_region(ctx, mm, table_idx_va,
580 subtable, XLAT_TABLE_ENTRIES,
581 level + 1);
582 if (end_va != table_idx_va + XLAT_BLOCK_SIZE(level) - 1)
583 return end_va;
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000584
585 } else if (action == ACTION_RECURSE_INTO_TABLE) {
586
587 subtable = (uint64_t *)(uintptr_t)(desc & TABLE_ADDR_MASK);
588 /* Recurse to write into subtable */
Antonio Nino Diazac998032017-02-27 17:23:54 +0000589 uintptr_t end_va = xlat_tables_map_region(ctx, mm, table_idx_va,
590 subtable, XLAT_TABLE_ENTRIES,
591 level + 1);
592 if (end_va != table_idx_va + XLAT_BLOCK_SIZE(level) - 1)
593 return end_va;
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000594
595 } else {
596
597 assert(action == ACTION_NONE);
598
599 }
600
601 table_idx++;
602 table_idx_va += XLAT_BLOCK_SIZE(level);
603
604 /* If reached the end of the region, exit */
605 if (mm_end_va <= table_idx_va)
606 break;
607 }
Antonio Nino Diazac998032017-02-27 17:23:54 +0000608
609 return table_idx_va - 1;
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000610}
611
612void print_mmap(mmap_region_t *const mmap)
613{
614#if LOG_LEVEL >= LOG_LEVEL_VERBOSE
615 tf_printf("mmap:\n");
616 mmap_region_t *mm = mmap;
617
618 while (mm->size) {
Sandrine Bailleux8f23fa82017-09-28 21:58:12 +0100619 tf_printf(" VA:%p PA:0x%llx size:0x%zx attr:0x%x",
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000620 (void *)mm->base_va, mm->base_pa,
621 mm->size, mm->attr);
Sandrine Bailleux8f23fa82017-09-28 21:58:12 +0100622 tf_printf(" granularity:0x%zx\n", mm->granularity);
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000623 ++mm;
624 };
625 tf_printf("\n");
626#endif
627}
628
629/*
630 * Function that verifies that a region can be mapped.
631 * Returns:
632 * 0: Success, the mapping is allowed.
633 * EINVAL: Invalid values were used as arguments.
634 * ERANGE: The memory limits were surpassed.
635 * ENOMEM: There is not enough memory in the mmap array.
636 * EPERM: Region overlaps another one in an invalid way.
637 */
Douglas Raillard6a5f8f12017-09-21 08:42:21 +0100638static int mmap_add_region_check(xlat_ctx_t *ctx, const mmap_region_t *mm)
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000639{
Douglas Raillard6a5f8f12017-09-21 08:42:21 +0100640 unsigned long long base_pa = mm->base_pa;
641 uintptr_t base_va = mm->base_va;
642 size_t size = mm->size;
Sandrine Bailleux8f23fa82017-09-28 21:58:12 +0100643 size_t granularity = mm->granularity;
Douglas Raillard6a5f8f12017-09-21 08:42:21 +0100644
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000645 unsigned long long end_pa = base_pa + size - 1;
646 uintptr_t end_va = base_va + size - 1;
647
648 if (!IS_PAGE_ALIGNED(base_pa) || !IS_PAGE_ALIGNED(base_va) ||
649 !IS_PAGE_ALIGNED(size))
650 return -EINVAL;
651
Sandrine Bailleux8f23fa82017-09-28 21:58:12 +0100652 if ((granularity != XLAT_BLOCK_SIZE(1)) &&
653 (granularity != XLAT_BLOCK_SIZE(2)) &&
654 (granularity != XLAT_BLOCK_SIZE(3))) {
655 return -EINVAL;
656 }
657
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000658 /* Check for overflows */
659 if ((base_pa > end_pa) || (base_va > end_va))
660 return -ERANGE;
661
662 if ((base_va + (uintptr_t)size - (uintptr_t)1) > ctx->va_max_address)
663 return -ERANGE;
664
665 if ((base_pa + (unsigned long long)size - 1ULL) > ctx->pa_max_address)
666 return -ERANGE;
667
Douglas Raillard6a5f8f12017-09-21 08:42:21 +0100668 /* Check that there is space in the ctx->mmap array */
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000669 if (ctx->mmap[ctx->mmap_num - 1].size != 0)
670 return -ENOMEM;
671
672 /* Check for PAs and VAs overlaps with all other regions */
Douglas Raillard6a5f8f12017-09-21 08:42:21 +0100673 for (mmap_region_t *mm_cursor = ctx->mmap;
674 mm_cursor->size; ++mm_cursor) {
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000675
Douglas Raillard6a5f8f12017-09-21 08:42:21 +0100676 uintptr_t mm_cursor_end_va = mm_cursor->base_va
677 + mm_cursor->size - 1;
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000678
679 /*
680 * Check if one of the regions is completely inside the other
681 * one.
682 */
683 int fully_overlapped_va =
Douglas Raillard6a5f8f12017-09-21 08:42:21 +0100684 ((base_va >= mm_cursor->base_va) &&
685 (end_va <= mm_cursor_end_va)) ||
686
687 ((mm_cursor->base_va >= base_va) &&
688 (mm_cursor_end_va <= end_va));
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000689
690 /*
691 * Full VA overlaps are only allowed if both regions are
692 * identity mapped (zero offset) or have the same VA to PA
693 * offset. Also, make sure that it's not the exact same area.
Antonio Nino Diazac998032017-02-27 17:23:54 +0000694 * This can only be done with static regions.
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000695 */
696 if (fully_overlapped_va) {
697
Antonio Nino Diazac998032017-02-27 17:23:54 +0000698#if PLAT_XLAT_TABLES_DYNAMIC
Sandrine Bailleux8f23fa82017-09-28 21:58:12 +0100699 if ((mm->attr & MT_DYNAMIC) ||
Douglas Raillard6a5f8f12017-09-21 08:42:21 +0100700 (mm_cursor->attr & MT_DYNAMIC))
Antonio Nino Diazac998032017-02-27 17:23:54 +0000701 return -EPERM;
702#endif /* PLAT_XLAT_TABLES_DYNAMIC */
Douglas Raillard6a5f8f12017-09-21 08:42:21 +0100703 if ((mm_cursor->base_va - mm_cursor->base_pa) !=
704 (base_va - base_pa))
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000705 return -EPERM;
706
Douglas Raillard6a5f8f12017-09-21 08:42:21 +0100707 if ((base_va == mm_cursor->base_va) &&
708 (size == mm_cursor->size))
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000709 return -EPERM;
710
711 } else {
712 /*
713 * If the regions do not have fully overlapping VAs,
714 * then they must have fully separated VAs and PAs.
715 * Partial overlaps are not allowed
716 */
717
Douglas Raillard6a5f8f12017-09-21 08:42:21 +0100718 unsigned long long mm_cursor_end_pa =
719 mm_cursor->base_pa + mm_cursor->size - 1;
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000720
721 int separated_pa =
Douglas Raillard6a5f8f12017-09-21 08:42:21 +0100722 (end_pa < mm_cursor->base_pa) ||
723 (base_pa > mm_cursor_end_pa);
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000724 int separated_va =
Douglas Raillard6a5f8f12017-09-21 08:42:21 +0100725 (end_va < mm_cursor->base_va) ||
726 (base_va > mm_cursor_end_va);
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000727
728 if (!(separated_va && separated_pa))
729 return -EPERM;
730 }
731 }
732
733 return 0;
734}
735
Sandrine Bailleux66342932017-07-18 13:26:36 +0100736void mmap_add_region_ctx(xlat_ctx_t *ctx, const mmap_region_t *mm)
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000737{
John Tsichritzisfdd92482018-05-25 09:12:48 +0100738 mmap_region_t *mm_cursor = ctx->mmap, *mm_destination;
Varun Wadekarccbd2e32018-04-03 10:44:41 -0700739 const mmap_region_t *mm_end = ctx->mmap + ctx->mmap_num;
740 mmap_region_t *mm_last;
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000741 unsigned long long end_pa = mm->base_pa + mm->size - 1;
742 uintptr_t end_va = mm->base_va + mm->size - 1;
743 int ret;
744
745 /* Ignore empty regions */
746 if (!mm->size)
747 return;
748
Antonio Nino Diazac998032017-02-27 17:23:54 +0000749 /* Static regions must be added before initializing the xlat tables. */
750 assert(!ctx->initialized);
751
Douglas Raillard6a5f8f12017-09-21 08:42:21 +0100752 ret = mmap_add_region_check(ctx, mm);
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000753 if (ret != 0) {
754 ERROR("mmap_add_region_check() failed. error %d\n", ret);
755 assert(0);
756 return;
757 }
758
759 /*
760 * Find correct place in mmap to insert new region.
761 *
762 * 1 - Lower region VA end first.
763 * 2 - Smaller region size first.
764 *
765 * VA 0 0xFF
766 *
767 * 1st |------|
768 * 2nd |------------|
769 * 3rd |------|
770 * 4th |---|
771 * 5th |---|
772 * 6th |----------|
773 * 7th |-------------------------------------|
774 *
775 * This is required for overlapping regions only. It simplifies adding
776 * regions with the loop in xlat_tables_init_internal because the outer
777 * ones won't overwrite block or page descriptors of regions added
778 * previously.
Antonio Nino Diazac998032017-02-27 17:23:54 +0000779 *
780 * Overlapping is only allowed for static regions.
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000781 */
782
783 while ((mm_cursor->base_va + mm_cursor->size - 1) < end_va
784 && mm_cursor->size)
785 ++mm_cursor;
786
Yann Gautiereabaea52018-06-14 14:36:20 +0200787 while ((mm_cursor->base_va + mm_cursor->size - 1 == end_va) &&
788 (mm_cursor->size != 0U) && (mm_cursor->size < mm->size))
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000789 ++mm_cursor;
790
Varun Wadekarccbd2e32018-04-03 10:44:41 -0700791 /*
792 * Find the last entry marker in the mmap
793 */
794 mm_last = ctx->mmap;
795 while ((mm_last->size != 0U) && (mm_last < mm_end)) {
796 ++mm_last;
797 }
798
799 /*
800 * Check if we have enough space in the memory mapping table.
801 * This shouldn't happen as we have checked in mmap_add_region_check
802 * that there is free space.
803 */
804 assert(mm_last->size == 0U);
Jeenu Viswambharan58e81482018-04-27 15:06:57 +0100805
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000806 /* Make room for new region by moving other regions up by one place */
John Tsichritzisfdd92482018-05-25 09:12:48 +0100807 mm_destination = mm_cursor + 1;
808 memmove(mm_destination, mm_cursor,
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000809 (uintptr_t)mm_last - (uintptr_t)mm_cursor);
810
811 /*
812 * Check we haven't lost the empty sentinel from the end of the array.
813 * This shouldn't happen as we have checked in mmap_add_region_check
814 * that there is free space.
815 */
Varun Wadekarccbd2e32018-04-03 10:44:41 -0700816 assert(mm_end->size == 0U);
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000817
Douglas Raillardf68d2ed2017-09-12 10:31:49 +0100818 *mm_cursor = *mm;
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +0000819
820 if (end_pa > ctx->max_pa)
821 ctx->max_pa = end_pa;
822 if (end_va > ctx->max_va)
823 ctx->max_va = end_va;
824}
825
Antonio Nino Diaz8643a812018-06-21 14:39:16 +0100826void mmap_add_region(unsigned long long base_pa, uintptr_t base_va, size_t size,
827 unsigned int attr)
Sandrine Bailleux66342932017-07-18 13:26:36 +0100828{
Douglas Raillard35c09f12017-08-31 16:20:25 +0100829 mmap_region_t mm = MAP_REGION(base_pa, base_va, size, attr);
Sandrine Bailleux66342932017-07-18 13:26:36 +0100830 mmap_add_region_ctx(&tf_xlat_ctx, &mm);
831}
832
833
834void mmap_add_ctx(xlat_ctx_t *ctx, const mmap_region_t *mm)
835{
836 while (mm->size) {
837 mmap_add_region_ctx(ctx, mm);
838 mm++;
839 }
840}
841
842void mmap_add(const mmap_region_t *mm)
843{
844 mmap_add_ctx(&tf_xlat_ctx, mm);
845}
846
Antonio Nino Diazac998032017-02-27 17:23:54 +0000847#if PLAT_XLAT_TABLES_DYNAMIC
848
849int mmap_add_dynamic_region_ctx(xlat_ctx_t *ctx, mmap_region_t *mm)
850{
851 mmap_region_t *mm_cursor = ctx->mmap;
852 mmap_region_t *mm_last = mm_cursor + ctx->mmap_num;
853 unsigned long long end_pa = mm->base_pa + mm->size - 1;
854 uintptr_t end_va = mm->base_va + mm->size - 1;
855 int ret;
856
857 /* Nothing to do */
858 if (!mm->size)
859 return 0;
860
Douglas Raillard6a5f8f12017-09-21 08:42:21 +0100861 /* Now this region is a dynamic one */
862 mm->attr |= MT_DYNAMIC;
863
864 ret = mmap_add_region_check(ctx, mm);
Antonio Nino Diazac998032017-02-27 17:23:54 +0000865 if (ret != 0)
866 return ret;
867
868 /*
869 * Find the adequate entry in the mmap array in the same way done for
870 * static regions in mmap_add_region_ctx().
871 */
872
Douglas Raillard6a5f8f12017-09-21 08:42:21 +0100873 while ((mm_cursor->base_va + mm_cursor->size - 1)
874 < end_va && mm_cursor->size)
Antonio Nino Diazac998032017-02-27 17:23:54 +0000875 ++mm_cursor;
876
Douglas Raillard6a5f8f12017-09-21 08:42:21 +0100877 while ((mm_cursor->base_va + mm_cursor->size - 1 == end_va)
878 && (mm_cursor->size < mm->size))
Antonio Nino Diazac998032017-02-27 17:23:54 +0000879 ++mm_cursor;
880
881 /* Make room for new region by moving other regions up by one place */
Douglas Raillard6a5f8f12017-09-21 08:42:21 +0100882 memmove(mm_cursor + 1, mm_cursor,
883 (uintptr_t)mm_last - (uintptr_t)mm_cursor);
Antonio Nino Diazac998032017-02-27 17:23:54 +0000884
885 /*
886 * Check we haven't lost the empty sentinal from the end of the array.
887 * This shouldn't happen as we have checked in mmap_add_region_check
888 * that there is free space.
889 */
890 assert(mm_last->size == 0);
891
Douglas Raillardf68d2ed2017-09-12 10:31:49 +0100892 *mm_cursor = *mm;
Antonio Nino Diazac998032017-02-27 17:23:54 +0000893
894 /*
895 * Update the translation tables if the xlat tables are initialized. If
896 * not, this region will be mapped when they are initialized.
897 */
898 if (ctx->initialized) {
Douglas Raillard6a5f8f12017-09-21 08:42:21 +0100899 uintptr_t end_va = xlat_tables_map_region(ctx, mm_cursor,
900 0, ctx->base_table, ctx->base_table_entries,
901 ctx->base_level);
Antonio Nino Diazac998032017-02-27 17:23:54 +0000902
903 /* Failed to map, remove mmap entry, unmap and return error. */
904 if (end_va != mm_cursor->base_va + mm_cursor->size - 1) {
Douglas Raillard6a5f8f12017-09-21 08:42:21 +0100905 memmove(mm_cursor, mm_cursor + 1,
906 (uintptr_t)mm_last - (uintptr_t)mm_cursor);
Antonio Nino Diazac998032017-02-27 17:23:54 +0000907
908 /*
909 * Check if the mapping function actually managed to map
910 * anything. If not, just return now.
911 */
Antonio Nino Diaz3f518922018-01-05 11:30:36 +0000912 if (mm->base_va >= end_va)
Antonio Nino Diazac998032017-02-27 17:23:54 +0000913 return -ENOMEM;
914
915 /*
Douglas Raillard6a5f8f12017-09-21 08:42:21 +0100916 * Something went wrong after mapping some table
917 * entries, undo every change done up to this point.
Antonio Nino Diazac998032017-02-27 17:23:54 +0000918 */
919 mmap_region_t unmap_mm = {
920 .base_pa = 0,
921 .base_va = mm->base_va,
922 .size = end_va - mm->base_va,
923 .attr = 0
924 };
Antonio Nino Diazdcf9d922017-10-04 16:52:15 +0100925 xlat_tables_unmap_region(ctx, &unmap_mm, 0, ctx->base_table,
926 ctx->base_table_entries, ctx->base_level);
Antonio Nino Diazac998032017-02-27 17:23:54 +0000927
928 return -ENOMEM;
929 }
930
931 /*
932 * Make sure that all entries are written to the memory. There
933 * is no need to invalidate entries when mapping dynamic regions
934 * because new table/block/page descriptors only replace old
935 * invalid descriptors, that aren't TLB cached.
936 */
937 dsbishst();
938 }
939
940 if (end_pa > ctx->max_pa)
941 ctx->max_pa = end_pa;
942 if (end_va > ctx->max_va)
943 ctx->max_va = end_va;
944
945 return 0;
946}
947
Antonio Nino Diaz8643a812018-06-21 14:39:16 +0100948int mmap_add_dynamic_region(unsigned long long base_pa, uintptr_t base_va,
949 size_t size, unsigned int attr)
Sandrine Bailleux66342932017-07-18 13:26:36 +0100950{
Douglas Raillard35c09f12017-08-31 16:20:25 +0100951 mmap_region_t mm = MAP_REGION(base_pa, base_va, size, attr);
Sandrine Bailleux66342932017-07-18 13:26:36 +0100952 return mmap_add_dynamic_region_ctx(&tf_xlat_ctx, &mm);
953}
954
Antonio Nino Diazac998032017-02-27 17:23:54 +0000955/*
956 * Removes the region with given base Virtual Address and size from the given
957 * context.
958 *
959 * Returns:
960 * 0: Success.
961 * EINVAL: Invalid values were used as arguments (region not found).
962 * EPERM: Tried to remove a static region.
963 */
964int mmap_remove_dynamic_region_ctx(xlat_ctx_t *ctx, uintptr_t base_va,
965 size_t size)
966{
967 mmap_region_t *mm = ctx->mmap;
968 mmap_region_t *mm_last = mm + ctx->mmap_num;
969 int update_max_va_needed = 0;
970 int update_max_pa_needed = 0;
971
972 /* Check sanity of mmap array. */
973 assert(mm[ctx->mmap_num].size == 0);
974
975 while (mm->size) {
976 if ((mm->base_va == base_va) && (mm->size == size))
977 break;
978 ++mm;
979 }
980
981 /* Check that the region was found */
982 if (mm->size == 0)
983 return -EINVAL;
984
985 /* If the region is static it can't be removed */
986 if (!(mm->attr & MT_DYNAMIC))
987 return -EPERM;
988
989 /* Check if this region is using the top VAs or PAs. */
990 if ((mm->base_va + mm->size - 1) == ctx->max_va)
991 update_max_va_needed = 1;
992 if ((mm->base_pa + mm->size - 1) == ctx->max_pa)
993 update_max_pa_needed = 1;
994
995 /* Update the translation tables if needed */
996 if (ctx->initialized) {
997 xlat_tables_unmap_region(ctx, mm, 0, ctx->base_table,
998 ctx->base_table_entries,
999 ctx->base_level);
1000 xlat_arch_tlbi_va_sync();
1001 }
1002
1003 /* Remove this region by moving the rest down by one place. */
1004 memmove(mm, mm + 1, (uintptr_t)mm_last - (uintptr_t)mm);
1005
1006 /* Check if we need to update the max VAs and PAs */
1007 if (update_max_va_needed) {
1008 ctx->max_va = 0;
1009 mm = ctx->mmap;
1010 while (mm->size) {
1011 if ((mm->base_va + mm->size - 1) > ctx->max_va)
1012 ctx->max_va = mm->base_va + mm->size - 1;
1013 ++mm;
1014 }
1015 }
1016
1017 if (update_max_pa_needed) {
1018 ctx->max_pa = 0;
1019 mm = ctx->mmap;
1020 while (mm->size) {
1021 if ((mm->base_pa + mm->size - 1) > ctx->max_pa)
1022 ctx->max_pa = mm->base_pa + mm->size - 1;
1023 ++mm;
1024 }
1025 }
1026
1027 return 0;
1028}
1029
Sandrine Bailleux66342932017-07-18 13:26:36 +01001030int mmap_remove_dynamic_region(uintptr_t base_va, size_t size)
1031{
1032 return mmap_remove_dynamic_region_ctx(&tf_xlat_ctx,
1033 base_va, size);
1034}
1035
Antonio Nino Diazac998032017-02-27 17:23:54 +00001036#endif /* PLAT_XLAT_TABLES_DYNAMIC */
1037
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +00001038#if LOG_LEVEL >= LOG_LEVEL_VERBOSE
1039
1040/* Print the attributes of the specified block descriptor. */
Sandrine Bailleuxc3708e22017-10-13 14:17:09 +01001041static void xlat_desc_print(const xlat_ctx_t *ctx, uint64_t desc)
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +00001042{
1043 int mem_type_index = ATTR_INDEX_GET(desc);
Antonio Nino Diazdcf9d922017-10-04 16:52:15 +01001044 xlat_regime_t xlat_regime = ctx->xlat_regime;
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +00001045
1046 if (mem_type_index == ATTR_IWBWA_OWBWA_NTR_INDEX) {
1047 tf_printf("MEM");
1048 } else if (mem_type_index == ATTR_NON_CACHEABLE_INDEX) {
1049 tf_printf("NC");
1050 } else {
1051 assert(mem_type_index == ATTR_DEVICE_INDEX);
1052 tf_printf("DEV");
1053 }
1054
Antonio Nino Diazdcf9d922017-10-04 16:52:15 +01001055 const char *priv_str = "(PRIV)";
1056 const char *user_str = "(USER)";
1057
1058 /*
1059 * Showing Privileged vs Unprivileged only makes sense for EL1&0
1060 * mappings
1061 */
1062 const char *ro_str = "-RO";
1063 const char *rw_str = "-RW";
1064 const char *no_access_str = "-NOACCESS";
1065
1066 if (xlat_regime == EL3_REGIME) {
1067 /* For EL3, the AP[2] bit is all what matters */
Roberto Vargasc8cc3512018-05-09 10:49:24 +01001068 tf_printf("%s", (desc & LOWER_ATTRS(AP_RO)) ? ro_str : rw_str);
Antonio Nino Diazdcf9d922017-10-04 16:52:15 +01001069 } else {
1070 const char *ap_str = (desc & LOWER_ATTRS(AP_RO)) ? ro_str : rw_str;
Roberto Vargasc8cc3512018-05-09 10:49:24 +01001071 tf_printf("%s", ap_str);
1072 tf_printf("%s", priv_str);
Antonio Nino Diazdcf9d922017-10-04 16:52:15 +01001073 /*
1074 * EL0 can only have the same permissions as EL1 or no
1075 * permissions at all.
1076 */
Roberto Vargasc8cc3512018-05-09 10:49:24 +01001077 tf_printf("%s",
1078 (desc & LOWER_ATTRS(AP_ACCESS_UNPRIVILEGED))
Antonio Nino Diazdcf9d922017-10-04 16:52:15 +01001079 ? ap_str : no_access_str);
Roberto Vargasc8cc3512018-05-09 10:49:24 +01001080 tf_printf("%s", user_str);
Antonio Nino Diazdcf9d922017-10-04 16:52:15 +01001081 }
1082
1083 const char *xn_str = "-XN";
1084 const char *exec_str = "-EXEC";
1085
1086 if (xlat_regime == EL3_REGIME) {
1087 /* For EL3, the XN bit is all what matters */
Antonio Nino Diaz8ca46022018-06-21 10:52:44 +01001088 tf_printf("%s", (UPPER_ATTRS(XN) & desc) ? xn_str : exec_str);
Antonio Nino Diazdcf9d922017-10-04 16:52:15 +01001089 } else {
1090 /* For EL0 and EL1, we need to know who has which rights */
Antonio Nino Diaz8ca46022018-06-21 10:52:44 +01001091 tf_printf("%s", (UPPER_ATTRS(PXN) & desc) ? xn_str : exec_str);
Roberto Vargasc8cc3512018-05-09 10:49:24 +01001092 tf_printf("%s", priv_str);
Antonio Nino Diazdcf9d922017-10-04 16:52:15 +01001093
Antonio Nino Diaz8ca46022018-06-21 10:52:44 +01001094 tf_printf("%s", (UPPER_ATTRS(UXN) & desc) ? xn_str : exec_str);
Roberto Vargasc8cc3512018-05-09 10:49:24 +01001095 tf_printf("%s", user_str);
Antonio Nino Diazdcf9d922017-10-04 16:52:15 +01001096 }
1097
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +00001098 tf_printf(LOWER_ATTRS(NS) & desc ? "-NS" : "-S");
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +00001099}
1100
1101static const char * const level_spacers[] = {
Antonio Nino Diaz755e54f2017-02-13 11:35:49 +00001102 "[LV0] ",
1103 " [LV1] ",
1104 " [LV2] ",
1105 " [LV3] "
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +00001106};
1107
Antonio Nino Diaz755e54f2017-02-13 11:35:49 +00001108static const char *invalid_descriptors_ommited =
1109 "%s(%d invalid descriptors omitted)\n";
1110
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +00001111/*
1112 * Recursive function that reads the translation tables passed as an argument
1113 * and prints their status.
1114 */
Antonio Nino Diazdcf9d922017-10-04 16:52:15 +01001115static void xlat_tables_print_internal(xlat_ctx_t *ctx,
1116 const uintptr_t table_base_va,
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +00001117 uint64_t *const table_base, const int table_entries,
Antonio Nino Diazdcf9d922017-10-04 16:52:15 +01001118 const unsigned int level)
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +00001119{
1120 assert(level <= XLAT_TABLE_LEVEL_MAX);
1121
1122 uint64_t desc;
1123 uintptr_t table_idx_va = table_base_va;
1124 int table_idx = 0;
1125
1126 size_t level_size = XLAT_BLOCK_SIZE(level);
1127
Antonio Nino Diaz755e54f2017-02-13 11:35:49 +00001128 /*
1129 * Keep track of how many invalid descriptors are counted in a row.
1130 * Whenever multiple invalid descriptors are found, only the first one
1131 * is printed, and a line is added to inform about how many descriptors
1132 * have been omitted.
1133 */
1134 int invalid_row_count = 0;
1135
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +00001136 while (table_idx < table_entries) {
1137
1138 desc = table_base[table_idx];
1139
1140 if ((desc & DESC_MASK) == INVALID_DESC) {
1141
Antonio Nino Diaz755e54f2017-02-13 11:35:49 +00001142 if (invalid_row_count == 0) {
1143 tf_printf("%sVA:%p size:0x%zx\n",
1144 level_spacers[level],
1145 (void *)table_idx_va, level_size);
1146 }
1147 invalid_row_count++;
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +00001148
1149 } else {
1150
Antonio Nino Diaz755e54f2017-02-13 11:35:49 +00001151 if (invalid_row_count > 1) {
1152 tf_printf(invalid_descriptors_ommited,
1153 level_spacers[level],
1154 invalid_row_count - 1);
1155 }
1156 invalid_row_count = 0;
1157
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +00001158 /*
1159 * Check if this is a table or a block. Tables are only
1160 * allowed in levels other than 3, but DESC_PAGE has the
1161 * same value as DESC_TABLE, so we need to check.
1162 */
1163 if (((desc & DESC_MASK) == TABLE_DESC) &&
1164 (level < XLAT_TABLE_LEVEL_MAX)) {
1165 /*
1166 * Do not print any PA for a table descriptor,
1167 * as it doesn't directly map physical memory
1168 * but instead points to the next translation
1169 * table in the translation table walk.
1170 */
1171 tf_printf("%sVA:%p size:0x%zx\n",
1172 level_spacers[level],
1173 (void *)table_idx_va, level_size);
1174
1175 uintptr_t addr_inner = desc & TABLE_ADDR_MASK;
1176
Antonio Nino Diazdcf9d922017-10-04 16:52:15 +01001177 xlat_tables_print_internal(ctx, table_idx_va,
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +00001178 (uint64_t *)addr_inner,
Antonio Nino Diazdcf9d922017-10-04 16:52:15 +01001179 XLAT_TABLE_ENTRIES, level + 1);
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +00001180 } else {
1181 tf_printf("%sVA:%p PA:0x%llx size:0x%zx ",
1182 level_spacers[level],
1183 (void *)table_idx_va,
1184 (unsigned long long)(desc & TABLE_ADDR_MASK),
1185 level_size);
Antonio Nino Diazdcf9d922017-10-04 16:52:15 +01001186 xlat_desc_print(ctx, desc);
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +00001187 tf_printf("\n");
1188 }
1189 }
1190
1191 table_idx++;
1192 table_idx_va += level_size;
1193 }
Antonio Nino Diaz755e54f2017-02-13 11:35:49 +00001194
1195 if (invalid_row_count > 1) {
1196 tf_printf(invalid_descriptors_ommited,
1197 level_spacers[level], invalid_row_count - 1);
1198 }
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +00001199}
1200
1201#endif /* LOG_LEVEL >= LOG_LEVEL_VERBOSE */
1202
1203void xlat_tables_print(xlat_ctx_t *ctx)
1204{
1205#if LOG_LEVEL >= LOG_LEVEL_VERBOSE
Antonio Nino Diazdcf9d922017-10-04 16:52:15 +01001206 const char *xlat_regime_str;
1207 if (ctx->xlat_regime == EL1_EL0_REGIME) {
1208 xlat_regime_str = "1&0";
1209 } else {
1210 assert(ctx->xlat_regime == EL3_REGIME);
1211 xlat_regime_str = "3";
1212 }
Sandrine Bailleux4e8f1452017-05-26 15:47:08 +01001213 VERBOSE("Translation tables state:\n");
Antonio Nino Diazdcf9d922017-10-04 16:52:15 +01001214 VERBOSE(" Xlat regime: EL%s\n", xlat_regime_str);
Sandrine Bailleux4e8f1452017-05-26 15:47:08 +01001215 VERBOSE(" Max allowed PA: 0x%llx\n", ctx->pa_max_address);
1216 VERBOSE(" Max allowed VA: %p\n", (void *) ctx->va_max_address);
1217 VERBOSE(" Max mapped PA: 0x%llx\n", ctx->max_pa);
1218 VERBOSE(" Max mapped VA: %p\n", (void *) ctx->max_va);
1219
1220 VERBOSE(" Initial lookup level: %i\n", ctx->base_level);
1221 VERBOSE(" Entries @initial lookup level: %i\n",
1222 ctx->base_table_entries);
1223
1224 int used_page_tables;
1225#if PLAT_XLAT_TABLES_DYNAMIC
1226 used_page_tables = 0;
Sandrine Bailleuxde6628d2017-08-01 09:16:38 +01001227 for (unsigned int i = 0; i < ctx->tables_num; ++i) {
Sandrine Bailleux4e8f1452017-05-26 15:47:08 +01001228 if (ctx->tables_mapped_regions[i] != 0)
1229 ++used_page_tables;
1230 }
1231#else
1232 used_page_tables = ctx->next_table;
1233#endif
1234 VERBOSE(" Used %i sub-tables out of %i (spare: %i)\n",
1235 used_page_tables, ctx->tables_num,
1236 ctx->tables_num - used_page_tables);
1237
Antonio Nino Diazdcf9d922017-10-04 16:52:15 +01001238 xlat_tables_print_internal(ctx, 0, ctx->base_table,
1239 ctx->base_table_entries, ctx->base_level);
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +00001240#endif /* LOG_LEVEL >= LOG_LEVEL_VERBOSE */
1241}
1242
Sandrine Bailleux66342932017-07-18 13:26:36 +01001243void init_xlat_tables_ctx(xlat_ctx_t *ctx)
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +00001244{
Antonio Nino Diazdcf9d922017-10-04 16:52:15 +01001245 assert(ctx != NULL);
Sandrine Bailleux66342932017-07-18 13:26:36 +01001246 assert(!ctx->initialized);
Antonio Nino Diazdcf9d922017-10-04 16:52:15 +01001247 assert(ctx->xlat_regime == EL3_REGIME || ctx->xlat_regime == EL1_EL0_REGIME);
1248 assert(!is_mmu_enabled_ctx(ctx));
Sandrine Bailleux66342932017-07-18 13:26:36 +01001249
Antonio Nino Diazdcf9d922017-10-04 16:52:15 +01001250 mmap_region_t *mm = ctx->mmap;
Sandrine Bailleux66342932017-07-18 13:26:36 +01001251
Antonio Nino Diazdcf9d922017-10-04 16:52:15 +01001252 print_mmap(mm);
Sandrine Bailleux66342932017-07-18 13:26:36 +01001253
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +00001254 /* All tables must be zeroed before mapping any region. */
1255
Varun Wadekar66231d12017-06-07 09:57:42 -07001256 for (unsigned int i = 0; i < ctx->base_table_entries; i++)
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +00001257 ctx->base_table[i] = INVALID_DESC;
1258
Varun Wadekar66231d12017-06-07 09:57:42 -07001259 for (unsigned int j = 0; j < ctx->tables_num; j++) {
Antonio Nino Diazac998032017-02-27 17:23:54 +00001260#if PLAT_XLAT_TABLES_DYNAMIC
1261 ctx->tables_mapped_regions[j] = 0;
1262#endif
Varun Wadekar66231d12017-06-07 09:57:42 -07001263 for (unsigned int i = 0; i < XLAT_TABLE_ENTRIES; i++)
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +00001264 ctx->tables[j][i] = INVALID_DESC;
1265 }
1266
Antonio Nino Diazac998032017-02-27 17:23:54 +00001267 while (mm->size) {
1268 uintptr_t end_va = xlat_tables_map_region(ctx, mm, 0, ctx->base_table,
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +00001269 ctx->base_table_entries, ctx->base_level);
1270
Antonio Nino Diazac998032017-02-27 17:23:54 +00001271 if (end_va != mm->base_va + mm->size - 1) {
1272 ERROR("Not enough memory to map region:\n"
1273 " VA:%p PA:0x%llx size:0x%zx attr:0x%x\n",
1274 (void *)mm->base_va, mm->base_pa, mm->size, mm->attr);
1275 panic();
1276 }
1277
1278 mm++;
1279 }
1280
Sandrine Bailleux46c53a22017-07-11 15:11:10 +01001281 assert(ctx->pa_max_address <= xlat_arch_get_max_supported_pa());
Sandrine Bailleux66342932017-07-18 13:26:36 +01001282 assert(ctx->max_va <= ctx->va_max_address);
1283 assert(ctx->max_pa <= ctx->pa_max_address);
1284
Sandrine Bailleuxc5b63772017-05-31 13:31:48 +01001285 ctx->initialized = 1;
1286
1287 xlat_tables_print(ctx);
Sandrine Bailleux66342932017-07-18 13:26:36 +01001288}
1289
1290void init_xlat_tables(void)
1291{
1292 init_xlat_tables_ctx(&tf_xlat_ctx);
Antonio Nino Diaz233c7c12017-03-08 14:40:23 +00001293}
Sandrine Bailleux66342932017-07-18 13:26:36 +01001294
Sandrine Bailleuxc5b63772017-05-31 13:31:48 +01001295/*
1296 * If dynamic allocation of new regions is disabled then by the time we call the
1297 * function enabling the MMU, we'll have registered all the memory regions to
1298 * map for the system's lifetime. Therefore, at this point we know the maximum
1299 * physical address that will ever be mapped.
1300 *
1301 * If dynamic allocation is enabled then we can't make any such assumption
1302 * because the maximum physical address could get pushed while adding a new
1303 * region. Therefore, in this case we have to assume that the whole address
1304 * space size might be mapped.
1305 */
1306#ifdef PLAT_XLAT_TABLES_DYNAMIC
Sandrine Bailleux46c53a22017-07-11 15:11:10 +01001307#define MAX_PHYS_ADDR tf_xlat_ctx.pa_max_address
Sandrine Bailleuxc5b63772017-05-31 13:31:48 +01001308#else
1309#define MAX_PHYS_ADDR tf_xlat_ctx.max_pa
1310#endif
1311
Sandrine Bailleux66342932017-07-18 13:26:36 +01001312#ifdef AARCH32
1313
1314void enable_mmu_secure(unsigned int flags)
1315{
Jeenu Viswambharan58e81482018-04-27 15:06:57 +01001316 setup_mmu_cfg(flags, tf_xlat_ctx.base_table, MAX_PHYS_ADDR,
Sandrine Bailleux46c53a22017-07-11 15:11:10 +01001317 tf_xlat_ctx.va_max_address);
Jeenu Viswambharan58e81482018-04-27 15:06:57 +01001318 enable_mmu_direct(flags);
Sandrine Bailleux66342932017-07-18 13:26:36 +01001319}
1320
1321#else
1322
1323void enable_mmu_el1(unsigned int flags)
1324{
Jeenu Viswambharan58e81482018-04-27 15:06:57 +01001325 setup_mmu_cfg(flags, tf_xlat_ctx.base_table, MAX_PHYS_ADDR,
Sandrine Bailleux46c53a22017-07-11 15:11:10 +01001326 tf_xlat_ctx.va_max_address);
Jeenu Viswambharan58e81482018-04-27 15:06:57 +01001327 enable_mmu_direct_el1(flags);
Sandrine Bailleux66342932017-07-18 13:26:36 +01001328}
1329
1330void enable_mmu_el3(unsigned int flags)
1331{
Jeenu Viswambharan58e81482018-04-27 15:06:57 +01001332 setup_mmu_cfg(flags, tf_xlat_ctx.base_table, MAX_PHYS_ADDR,
Sandrine Bailleux46c53a22017-07-11 15:11:10 +01001333 tf_xlat_ctx.va_max_address);
Jeenu Viswambharan58e81482018-04-27 15:06:57 +01001334 enable_mmu_direct_el3(flags);
Sandrine Bailleux66342932017-07-18 13:26:36 +01001335}
1336
1337#endif /* AARCH32 */
Sandrine Bailleuxc3708e22017-10-13 14:17:09 +01001338
1339/*
1340 * Do a translation table walk to find the block or page descriptor that maps
1341 * virtual_addr.
1342 *
1343 * On success, return the address of the descriptor within the translation
1344 * table. Its lookup level is stored in '*out_level'.
1345 * On error, return NULL.
1346 *
1347 * xlat_table_base
1348 * Base address for the initial lookup level.
1349 * xlat_table_base_entries
1350 * Number of entries in the translation table for the initial lookup level.
1351 * virt_addr_space_size
1352 * Size in bytes of the virtual address space.
1353 */
1354static uint64_t *find_xlat_table_entry(uintptr_t virtual_addr,
1355 void *xlat_table_base,
1356 int xlat_table_base_entries,
1357 unsigned long long virt_addr_space_size,
1358 int *out_level)
1359{
1360 unsigned int start_level;
1361 uint64_t *table;
1362 int entries;
1363
1364 VERBOSE("%s(%p)\n", __func__, (void *)virtual_addr);
1365
1366 start_level = GET_XLAT_TABLE_LEVEL_BASE(virt_addr_space_size);
1367 VERBOSE("Starting translation table walk from level %i\n", start_level);
1368
1369 table = xlat_table_base;
1370 entries = xlat_table_base_entries;
1371
1372 for (unsigned int level = start_level;
1373 level <= XLAT_TABLE_LEVEL_MAX;
1374 ++level) {
1375 int idx;
1376 uint64_t desc;
1377 uint64_t desc_type;
1378
1379 VERBOSE("Table address: %p\n", (void *)table);
1380
1381 idx = XLAT_TABLE_IDX(virtual_addr, level);
1382 VERBOSE("Index into level %i table: %i\n", level, idx);
1383 if (idx >= entries) {
1384 VERBOSE("Invalid address\n");
1385 return NULL;
1386 }
1387
1388 desc = table[idx];
1389 desc_type = desc & DESC_MASK;
1390 VERBOSE("Descriptor at level %i: 0x%llx\n", level,
1391 (unsigned long long)desc);
1392
1393 if (desc_type == INVALID_DESC) {
1394 VERBOSE("Invalid entry (memory not mapped)\n");
1395 return NULL;
1396 }
1397
1398 if (level == XLAT_TABLE_LEVEL_MAX) {
1399 /*
1400 * There can't be table entries at the final lookup
1401 * level.
1402 */
1403 assert(desc_type == PAGE_DESC);
1404 VERBOSE("Descriptor mapping a memory page (size: 0x%llx)\n",
1405 (unsigned long long)XLAT_BLOCK_SIZE(XLAT_TABLE_LEVEL_MAX));
1406 *out_level = level;
1407 return &table[idx];
1408 }
1409
1410 if (desc_type == BLOCK_DESC) {
1411 VERBOSE("Descriptor mapping a memory block (size: 0x%llx)\n",
1412 (unsigned long long)XLAT_BLOCK_SIZE(level));
1413 *out_level = level;
1414 return &table[idx];
1415 }
1416
1417 assert(desc_type == TABLE_DESC);
1418 VERBOSE("Table descriptor, continuing xlat table walk...\n");
1419 table = (uint64_t *)(uintptr_t)(desc & TABLE_ADDR_MASK);
1420 entries = XLAT_TABLE_ENTRIES;
1421 }
1422
1423 /*
1424 * This shouldn't be reached, the translation table walk should end at
1425 * most at level XLAT_TABLE_LEVEL_MAX and return from inside the loop.
1426 */
1427 assert(0);
1428
1429 return NULL;
1430}
1431
1432
1433static int get_mem_attributes_internal(const xlat_ctx_t *ctx, uintptr_t base_va,
Antonio Nino Diaze8811472018-04-17 15:10:18 +01001434 uint32_t *attributes, uint64_t **table_entry,
Sandrine Bailleuxc3708e22017-10-13 14:17:09 +01001435 unsigned long long *addr_pa, int *table_level)
1436{
1437 uint64_t *entry;
1438 uint64_t desc;
1439 int level;
1440 unsigned long long virt_addr_space_size;
1441
1442 /*
1443 * Sanity-check arguments.
1444 */
1445 assert(ctx != NULL);
1446 assert(ctx->initialized);
1447 assert(ctx->xlat_regime == EL1_EL0_REGIME || ctx->xlat_regime == EL3_REGIME);
1448
1449 virt_addr_space_size = (unsigned long long)ctx->va_max_address + 1;
1450 assert(virt_addr_space_size > 0);
1451
1452 entry = find_xlat_table_entry(base_va,
1453 ctx->base_table,
1454 ctx->base_table_entries,
1455 virt_addr_space_size,
1456 &level);
1457 if (entry == NULL) {
1458 WARN("Address %p is not mapped.\n", (void *)base_va);
1459 return -EINVAL;
1460 }
1461
1462 if (addr_pa != NULL) {
1463 *addr_pa = *entry & TABLE_ADDR_MASK;
1464 }
1465
1466 if (table_entry != NULL) {
1467 *table_entry = entry;
1468 }
1469
1470 if (table_level != NULL) {
1471 *table_level = level;
1472 }
1473
1474 desc = *entry;
1475
1476#if LOG_LEVEL >= LOG_LEVEL_VERBOSE
1477 VERBOSE("Attributes: ");
1478 xlat_desc_print(ctx, desc);
1479 tf_printf("\n");
1480#endif /* LOG_LEVEL >= LOG_LEVEL_VERBOSE */
1481
1482 assert(attributes != NULL);
1483 *attributes = 0;
1484
1485 int attr_index = (desc >> ATTR_INDEX_SHIFT) & ATTR_INDEX_MASK;
1486
1487 if (attr_index == ATTR_IWBWA_OWBWA_NTR_INDEX) {
1488 *attributes |= MT_MEMORY;
1489 } else if (attr_index == ATTR_NON_CACHEABLE_INDEX) {
1490 *attributes |= MT_NON_CACHEABLE;
1491 } else {
1492 assert(attr_index == ATTR_DEVICE_INDEX);
1493 *attributes |= MT_DEVICE;
1494 }
1495
1496 int ap2_bit = (desc >> AP2_SHIFT) & 1;
1497
1498 if (ap2_bit == AP2_RW)
1499 *attributes |= MT_RW;
1500
1501 if (ctx->xlat_regime == EL1_EL0_REGIME) {
1502 int ap1_bit = (desc >> AP1_SHIFT) & 1;
1503 if (ap1_bit == AP1_ACCESS_UNPRIVILEGED)
1504 *attributes |= MT_USER;
1505 }
1506
1507 int ns_bit = (desc >> NS_SHIFT) & 1;
1508
1509 if (ns_bit == 1)
1510 *attributes |= MT_NS;
1511
1512 uint64_t xn_mask = xlat_arch_regime_get_xn_desc(ctx->xlat_regime);
1513
1514 if ((desc & xn_mask) == xn_mask) {
1515 *attributes |= MT_EXECUTE_NEVER;
1516 } else {
1517 assert((desc & xn_mask) == 0);
1518 }
1519
1520 return 0;
1521}
1522
1523
1524int get_mem_attributes(const xlat_ctx_t *ctx, uintptr_t base_va,
Antonio Nino Diaze8811472018-04-17 15:10:18 +01001525 uint32_t *attributes)
Sandrine Bailleuxc3708e22017-10-13 14:17:09 +01001526{
1527 return get_mem_attributes_internal(ctx, base_va, attributes,
1528 NULL, NULL, NULL);
1529}
Sandrine Bailleux439c9012017-10-17 12:02:03 +01001530
1531
1532int change_mem_attributes(xlat_ctx_t *ctx,
1533 uintptr_t base_va,
1534 size_t size,
Antonio Nino Diaze8811472018-04-17 15:10:18 +01001535 uint32_t attr)
Sandrine Bailleux439c9012017-10-17 12:02:03 +01001536{
1537 /* Note: This implementation isn't optimized. */
1538
1539 assert(ctx != NULL);
1540 assert(ctx->initialized);
1541
1542 unsigned long long virt_addr_space_size =
1543 (unsigned long long)ctx->va_max_address + 1;
1544 assert(virt_addr_space_size > 0);
1545
1546 if (!IS_PAGE_ALIGNED(base_va)) {
1547 WARN("%s: Address %p is not aligned on a page boundary.\n",
1548 __func__, (void *)base_va);
1549 return -EINVAL;
1550 }
1551
1552 if (size == 0) {
1553 WARN("%s: Size is 0.\n", __func__);
1554 return -EINVAL;
1555 }
1556
1557 if ((size % PAGE_SIZE) != 0) {
1558 WARN("%s: Size 0x%zx is not a multiple of a page size.\n",
1559 __func__, size);
1560 return -EINVAL;
1561 }
1562
1563 if (((attr & MT_EXECUTE_NEVER) == 0) && ((attr & MT_RW) != 0)) {
1564 WARN("%s() doesn't allow to remap memory as read-write and executable.\n",
1565 __func__);
1566 return -EINVAL;
1567 }
1568
1569 int pages_count = size / PAGE_SIZE;
1570
1571 VERBOSE("Changing memory attributes of %i pages starting from address %p...\n",
1572 pages_count, (void *)base_va);
1573
1574 uintptr_t base_va_original = base_va;
1575
1576 /*
1577 * Sanity checks.
1578 */
1579 for (int i = 0; i < pages_count; ++i) {
1580 uint64_t *entry;
1581 uint64_t desc;
1582 int level;
1583
1584 entry = find_xlat_table_entry(base_va,
1585 ctx->base_table,
1586 ctx->base_table_entries,
1587 virt_addr_space_size,
1588 &level);
1589 if (entry == NULL) {
1590 WARN("Address %p is not mapped.\n", (void *)base_va);
1591 return -EINVAL;
1592 }
1593
1594 desc = *entry;
1595
1596 /*
1597 * Check that all the required pages are mapped at page
1598 * granularity.
1599 */
1600 if (((desc & DESC_MASK) != PAGE_DESC) ||
1601 (level != XLAT_TABLE_LEVEL_MAX)) {
1602 WARN("Address %p is not mapped at the right granularity.\n",
1603 (void *)base_va);
1604 WARN("Granularity is 0x%llx, should be 0x%x.\n",
1605 (unsigned long long)XLAT_BLOCK_SIZE(level), PAGE_SIZE);
1606 return -EINVAL;
1607 }
1608
1609 /*
1610 * If the region type is device, it shouldn't be executable.
1611 */
1612 int attr_index = (desc >> ATTR_INDEX_SHIFT) & ATTR_INDEX_MASK;
1613 if (attr_index == ATTR_DEVICE_INDEX) {
1614 if ((attr & MT_EXECUTE_NEVER) == 0) {
1615 WARN("Setting device memory as executable at address %p.",
1616 (void *)base_va);
1617 return -EINVAL;
1618 }
1619 }
1620
1621 base_va += PAGE_SIZE;
1622 }
1623
1624 /* Restore original value. */
1625 base_va = base_va_original;
1626
1627 VERBOSE("%s: All pages are mapped, now changing their attributes...\n",
1628 __func__);
1629
1630 for (int i = 0; i < pages_count; ++i) {
1631
Antonio Nino Diaze8811472018-04-17 15:10:18 +01001632 uint32_t old_attr, new_attr;
Sandrine Bailleux439c9012017-10-17 12:02:03 +01001633 uint64_t *entry;
1634 int level;
1635 unsigned long long addr_pa;
1636
1637 get_mem_attributes_internal(ctx, base_va, &old_attr,
1638 &entry, &addr_pa, &level);
1639
1640 VERBOSE("Old attributes: 0x%x\n", old_attr);
1641
1642 /*
1643 * From attr, only MT_RO/MT_RW, MT_EXECUTE/MT_EXECUTE_NEVER and
1644 * MT_USER/MT_PRIVILEGED are taken into account. Any other
1645 * information is ignored.
1646 */
1647
1648 /* Clean the old attributes so that they can be rebuilt. */
1649 new_attr = old_attr & ~(MT_RW|MT_EXECUTE_NEVER|MT_USER);
1650
1651 /*
1652 * Update attributes, but filter out the ones this function
1653 * isn't allowed to change.
1654 */
1655 new_attr |= attr & (MT_RW|MT_EXECUTE_NEVER|MT_USER);
1656
1657 VERBOSE("New attributes: 0x%x\n", new_attr);
1658
1659 /*
1660 * The break-before-make sequence requires writing an invalid
1661 * descriptor and making sure that the system sees the change
1662 * before writing the new descriptor.
1663 */
1664 *entry = INVALID_DESC;
1665
1666 /* Invalidate any cached copy of this mapping in the TLBs. */
1667 xlat_arch_tlbi_va_regime(base_va, ctx->xlat_regime);
1668
1669 /* Ensure completion of the invalidation. */
1670 xlat_arch_tlbi_va_sync();
1671
1672 /* Write new descriptor */
1673 *entry = xlat_desc(ctx, new_attr, addr_pa, level);
1674
1675 base_va += PAGE_SIZE;
1676 }
1677
1678 /* Ensure that the last descriptor writen is seen by the system. */
1679 dsbish();
1680
1681 return 0;
1682}