blob: 1a589c7e2a0da419d4ceadc5b9442246d583d0ea [file] [log] [blame]
Tom Rini10e47792018-05-06 17:58:06 -04001// SPDX-License-Identifier: GPL-2.0+
wdenkedc48b62002-09-08 17:56:50 +00002/*
3 * (C) Copyright 2002
4 * Wolfgang Denk, DENX Software Engineering, wd@denx.de.
wdenkedc48b62002-09-08 17:56:50 +00005 */
6
7/* for now: just dummy functions to satisfy the linker */
8
wdenkf8062712005-01-09 23:16:25 +00009#include <common.h>
Simon Glass63334482019-11-14 12:57:39 -070010#include <cpu_func.h>
Simon Glass0f2af882020-05-10 11:40:05 -060011#include <log.h>
Thierry Redingc97d9742014-12-09 22:25:22 -070012#include <malloc.h>
Simon Glass274e0b02020-05-10 11:39:56 -060013#include <asm/cache.h>
Simon Glass3ba929a2020-10-30 21:38:53 -060014#include <asm/global_data.h>
wdenkf8062712005-01-09 23:16:25 +000015
Ovidiu Panait68b371a2020-03-29 20:57:39 +030016DECLARE_GLOBAL_DATA_PTR;
17
Wu, Josh22190262015-07-27 11:40:17 +080018/*
19 * Flush range from all levels of d-cache/unified-cache.
20 * Affects the range [start, start + size - 1].
21 */
Jeroen Hofsteed7460772014-06-23 22:07:04 +020022__weak void flush_cache(unsigned long start, unsigned long size)
wdenkedc48b62002-09-08 17:56:50 +000023{
Wu, Josh22190262015-07-27 11:40:17 +080024 flush_dcache_range(start, start + size);
wdenkedc48b62002-09-08 17:56:50 +000025}
Aneesh V3bda3772011-06-16 23:30:50 +000026
27/*
28 * Default implementation:
29 * do a range flush for the entire range
30 */
Jeroen Hofsteed7460772014-06-23 22:07:04 +020031__weak void flush_dcache_all(void)
Aneesh V3bda3772011-06-16 23:30:50 +000032{
33 flush_cache(0, ~0);
34}
Aneesh Vfffbb972011-08-16 04:33:05 +000035
36/*
37 * Default implementation of enable_caches()
38 * Real implementation should be in platform code
39 */
Jeroen Hofsteed7460772014-06-23 22:07:04 +020040__weak void enable_caches(void)
Aneesh Vfffbb972011-08-16 04:33:05 +000041{
42 puts("WARNING: Caches not enabled\n");
43}
Thierry Redingc97d9742014-12-09 22:25:22 -070044
Wu, Joshaaa35452015-07-27 11:40:16 +080045__weak void invalidate_dcache_range(unsigned long start, unsigned long stop)
46{
47 /* An empty stub, real implementation should be in platform code */
48}
49__weak void flush_dcache_range(unsigned long start, unsigned long stop)
50{
51 /* An empty stub, real implementation should be in platform code */
52}
53
Simon Glass85406582016-06-19 19:43:01 -060054int check_cache_range(unsigned long start, unsigned long stop)
55{
56 int ok = 1;
57
58 if (start & (CONFIG_SYS_CACHELINE_SIZE - 1))
59 ok = 0;
60
61 if (stop & (CONFIG_SYS_CACHELINE_SIZE - 1))
62 ok = 0;
63
64 if (!ok) {
Simon Glass143997a2016-06-19 19:43:05 -060065 warn_non_spl("CACHE: Misaligned operation at range [%08lx, %08lx]\n",
66 start, stop);
Simon Glass85406582016-06-19 19:43:01 -060067 }
68
69 return ok;
70}
71
Thierry Redingc97d9742014-12-09 22:25:22 -070072#ifdef CONFIG_SYS_NONCACHED_MEMORY
73/*
74 * Reserve one MMU section worth of address space below the malloc() area that
75 * will be mapped uncached.
76 */
77static unsigned long noncached_start;
78static unsigned long noncached_end;
79static unsigned long noncached_next;
80
Patrice Chotarde2eb7212020-04-28 11:38:03 +020081void noncached_set_region(void)
82{
83#if !CONFIG_IS_ENABLED(SYS_DCACHE_OFF)
84 mmu_set_region_dcache_behaviour(noncached_start,
85 noncached_end - noncached_start,
86 DCACHE_OFF);
87#endif
88}
89
Ovidiu Panait1c45ed92020-11-28 10:43:13 +020090int noncached_init(void)
Thierry Redingc97d9742014-12-09 22:25:22 -070091{
92 phys_addr_t start, end;
93 size_t size;
94
Stephen Warren9b496432019-08-27 11:54:31 -060095 /* If this calculation changes, update board_f.c:reserve_noncached() */
Thierry Redingc97d9742014-12-09 22:25:22 -070096 end = ALIGN(mem_malloc_start, MMU_SECTION_SIZE) - MMU_SECTION_SIZE;
97 size = ALIGN(CONFIG_SYS_NONCACHED_MEMORY, MMU_SECTION_SIZE);
98 start = end - size;
99
100 debug("mapping memory %pa-%pa non-cached\n", &start, &end);
101
102 noncached_start = start;
103 noncached_end = end;
104 noncached_next = start;
105
Patrice Chotarde2eb7212020-04-28 11:38:03 +0200106 noncached_set_region();
Ovidiu Panait1c45ed92020-11-28 10:43:13 +0200107
108 return 0;
Thierry Redingc97d9742014-12-09 22:25:22 -0700109}
110
111phys_addr_t noncached_alloc(size_t size, size_t align)
112{
113 phys_addr_t next = ALIGN(noncached_next, align);
114
115 if (next >= noncached_end || (noncached_end - next) < size)
116 return 0;
117
118 debug("allocated %zu bytes of uncached memory @%pa\n", size, &next);
119 noncached_next = next + size;
120
121 return next;
122}
123#endif /* CONFIG_SYS_NONCACHED_MEMORY */
Albert ARIBAUDa3823222015-10-23 18:06:40 +0200124
Tom Rini1c640a62017-03-18 09:01:44 -0400125#if CONFIG_IS_ENABLED(SYS_THUMB_BUILD)
Albert ARIBAUDa3823222015-10-23 18:06:40 +0200126void invalidate_l2_cache(void)
127{
128 unsigned int val = 0;
129
130 asm volatile("mcr p15, 1, %0, c15, c11, 0 @ invl l2 cache"
131 : : "r" (val) : "cc");
132 isb();
133}
134#endif
Ovidiu Panait68b371a2020-03-29 20:57:39 +0300135
Ovidiu Panait2a2941b2020-03-29 20:57:41 +0300136int arch_reserve_mmu(void)
Ovidiu Panait68b371a2020-03-29 20:57:39 +0300137{
Ovidiu Panait2b618472020-03-29 20:57:40 +0300138 return arm_reserve_mmu();
139}
140
141__weak int arm_reserve_mmu(void)
142{
Ovidiu Panait68b371a2020-03-29 20:57:39 +0300143#if !(CONFIG_IS_ENABLED(SYS_ICACHE_OFF) && CONFIG_IS_ENABLED(SYS_DCACHE_OFF))
144 /* reserve TLB table */
145 gd->arch.tlb_size = PGTABLE_SIZE;
146 gd->relocaddr -= gd->arch.tlb_size;
147
148 /* round down to next 64 kB limit */
149 gd->relocaddr &= ~(0x10000 - 1);
150
151 gd->arch.tlb_addr = gd->relocaddr;
152 debug("TLB table from %08lx to %08lx\n", gd->arch.tlb_addr,
153 gd->arch.tlb_addr + gd->arch.tlb_size);
154
Tom Rini6a5dccc2022-11-16 13:10:41 -0500155#ifdef CFG_SYS_MEM_RESERVE_SECURE
Ovidiu Panait68b371a2020-03-29 20:57:39 +0300156 /*
157 * Record allocated tlb_addr in case gd->tlb_addr to be overwritten
158 * with location within secure ram.
159 */
160 gd->arch.tlb_allocated = gd->arch.tlb_addr;
161#endif
162#endif
163
164 return 0;
165}