blob: 37849c3124afae6aa515e648651158a79f399128 [file] [log] [blame]
Achin Gupta4f6ad662013-10-25 09:08:21 +01001/*
Louis Mayencourt944ade82019-08-08 12:03:26 +01002 * Copyright (c) 2013-2020, ARM Limited and Contributors. All rights reserved.
Achin Gupta4f6ad662013-10-25 09:08:21 +01003 *
dp-armfa3cf0b2017-05-03 09:38:09 +01004 * SPDX-License-Identifier: BSD-3-Clause
Achin Gupta4f6ad662013-10-25 09:08:21 +01005 */
6
Masahiro Yamada0b67e562020-03-09 17:39:48 +09007#include <common/bl_common.ld.h>
Antonio Nino Diaze0f90632018-12-14 00:18:21 +00008#include <lib/xlat_tables/xlat_tables_defs.h>
Achin Gupta4f6ad662013-10-25 09:08:21 +01009
10OUTPUT_FORMAT(PLATFORM_LINKER_FORMAT)
11OUTPUT_ARCH(PLATFORM_LINKER_ARCH)
Jeenu Viswambharan2a30a752014-03-11 11:06:45 +000012ENTRY(bl2_entrypoint)
Achin Gupta4f6ad662013-10-25 09:08:21 +010013
14MEMORY {
Juan Castillofd8c0772014-09-16 10:40:35 +010015 RAM (rwx): ORIGIN = BL2_BASE, LENGTH = BL2_LIMIT - BL2_BASE
Achin Gupta4f6ad662013-10-25 09:08:21 +010016}
17
18
19SECTIONS
20{
21 . = BL2_BASE;
Antonio Nino Diaz2ce2b092017-11-15 11:45:35 +000022 ASSERT(. == ALIGN(PAGE_SIZE),
Sandrine Bailleux8d69a032013-11-27 09:38:52 +000023 "BL2_BASE address is not aligned on a page boundary.")
Achin Gupta4f6ad662013-10-25 09:08:21 +010024
Sandrine Bailleuxf91f1442016-07-08 14:37:40 +010025#if SEPARATE_CODE_AND_RODATA
26 .text . : {
27 __TEXT_START__ = .;
28 *bl2_entrypoint.o(.text*)
Samuel Holland23f5e542019-10-20 16:11:25 -050029 *(SORT_BY_ALIGNMENT(.text*))
Sandrine Bailleuxf91f1442016-07-08 14:37:40 +010030 *(.vectors)
Roberto Vargasd93fde32018-04-11 11:53:31 +010031 . = ALIGN(PAGE_SIZE);
Sandrine Bailleuxf91f1442016-07-08 14:37:40 +010032 __TEXT_END__ = .;
33 } >RAM
34
Roberto Vargas1d04c632018-05-10 11:01:16 +010035 /* .ARM.extab and .ARM.exidx are only added because Clang need them */
36 .ARM.extab . : {
37 *(.ARM.extab* .gnu.linkonce.armextab.*)
38 } >RAM
39
40 .ARM.exidx . : {
41 *(.ARM.exidx* .gnu.linkonce.armexidx.*)
42 } >RAM
43
Sandrine Bailleuxf91f1442016-07-08 14:37:40 +010044 .rodata . : {
45 __RODATA_START__ = .;
Samuel Holland23f5e542019-10-20 16:11:25 -050046 *(SORT_BY_ALIGNMENT(.rodata*))
Sandrine Bailleuxf91f1442016-07-08 14:37:40 +010047
Masahiro Yamada583f8dd2020-03-26 10:57:12 +090048 RODATA_COMMON
Sandrine Bailleuxf91f1442016-07-08 14:37:40 +010049
Roberto Vargasd93fde32018-04-11 11:53:31 +010050 . = ALIGN(PAGE_SIZE);
Sandrine Bailleuxf91f1442016-07-08 14:37:40 +010051 __RODATA_END__ = .;
52 } >RAM
53#else
Sandrine Bailleux8d69a032013-11-27 09:38:52 +000054 ro . : {
55 __RO_START__ = .;
Andrew Thoelkee01ea342014-03-18 07:13:52 +000056 *bl2_entrypoint.o(.text*)
Samuel Holland23f5e542019-10-20 16:11:25 -050057 *(SORT_BY_ALIGNMENT(.text*))
58 *(SORT_BY_ALIGNMENT(.rodata*))
Juan Castillo8e55d932015-04-02 09:48:16 +010059
Masahiro Yamada583f8dd2020-03-26 10:57:12 +090060 RODATA_COMMON
Juan Castillo8e55d932015-04-02 09:48:16 +010061
Achin Guptab739f222014-01-18 16:50:09 +000062 *(.vectors)
Sandrine Bailleux8d69a032013-11-27 09:38:52 +000063 __RO_END_UNALIGNED__ = .;
64 /*
65 * Memory page(s) mapped to this section will be marked as
66 * read-only, executable. No RW data from the next section must
67 * creep in. Ensure the rest of the current memory page is unused.
68 */
Roberto Vargasd93fde32018-04-11 11:53:31 +010069 . = ALIGN(PAGE_SIZE);
Sandrine Bailleux8d69a032013-11-27 09:38:52 +000070 __RO_END__ = .;
Achin Gupta4f6ad662013-10-25 09:08:21 +010071 } >RAM
Sandrine Bailleuxf91f1442016-07-08 14:37:40 +010072#endif
Achin Gupta4f6ad662013-10-25 09:08:21 +010073
Achin Guptae9c4a642015-09-11 16:03:13 +010074 /*
75 * Define a linker symbol to mark start of the RW memory area for this
76 * image.
77 */
78 __RW_START__ = . ;
79
Masahiro Yamadac5864d82020-04-22 10:50:12 +090080 DATA_SECTION >RAM
Masahiro Yamada403990e2020-04-07 13:04:24 +090081 STACK_SECTION >RAM
Masahiro Yamadadd053b62020-03-26 13:16:33 +090082 BSS_SECTION >RAM
Masahiro Yamada0b67e562020-03-09 17:39:48 +090083 XLAT_TABLE_SECTION >RAM
Achin Guptaa0cd9892014-02-09 13:30:38 +000084
Soby Mathew2ae20432015-01-08 18:02:44 +000085#if USE_COHERENT_MEM
Achin Guptaa0cd9892014-02-09 13:30:38 +000086 /*
Sandrine Bailleux8d69a032013-11-27 09:38:52 +000087 * The base address of the coherent memory section must be page-aligned (4K)
88 * to guarantee that the coherent data are stored on their own pages and
89 * are not mixed with normal data. This is required to set up the correct
90 * memory attributes for the coherent data page tables.
91 */
Antonio Nino Diaz2ce2b092017-11-15 11:45:35 +000092 coherent_ram (NOLOAD) : ALIGN(PAGE_SIZE) {
Sandrine Bailleux8d69a032013-11-27 09:38:52 +000093 __COHERENT_RAM_START__ = .;
94 *(tzfw_coherent_mem)
95 __COHERENT_RAM_END_UNALIGNED__ = .;
96 /*
97 * Memory page(s) mapped to this section will be marked
98 * as device memory. No other unexpected data must creep in.
99 * Ensure the rest of the current memory page is unused.
100 */
Roberto Vargasd93fde32018-04-11 11:53:31 +0100101 . = ALIGN(PAGE_SIZE);
Sandrine Bailleux8d69a032013-11-27 09:38:52 +0000102 __COHERENT_RAM_END__ = .;
Achin Gupta4f6ad662013-10-25 09:08:21 +0100103 } >RAM
Soby Mathew2ae20432015-01-08 18:02:44 +0000104#endif
Achin Gupta4f6ad662013-10-25 09:08:21 +0100105
Achin Guptae9c4a642015-09-11 16:03:13 +0100106 /*
107 * Define a linker symbol to mark end of the RW memory area for this
108 * image.
109 */
110 __RW_END__ = .;
Sandrine Bailleux8d69a032013-11-27 09:38:52 +0000111 __BL2_END__ = .;
Achin Gupta4f6ad662013-10-25 09:08:21 +0100112
Sandrine Bailleux8d69a032013-11-27 09:38:52 +0000113 __BSS_SIZE__ = SIZEOF(.bss);
Soby Mathew2ae20432015-01-08 18:02:44 +0000114
115#if USE_COHERENT_MEM
Sandrine Bailleux8d69a032013-11-27 09:38:52 +0000116 __COHERENT_RAM_UNALIGNED_SIZE__ =
117 __COHERENT_RAM_END_UNALIGNED__ - __COHERENT_RAM_START__;
Soby Mathew2ae20432015-01-08 18:02:44 +0000118#endif
Sandrine Bailleux6c8b3592014-05-22 15:28:26 +0100119
120 ASSERT(. <= BL2_LIMIT, "BL2 image has exceeded its limit.")
Achin Gupta4f6ad662013-10-25 09:08:21 +0100121}