| /* |
| * (C) Copyright 2010 |
| * Texas Instruments, <www.ti.com> |
| * Aneesh V <aneesh@ti.com> |
| * |
| * SPDX-License-Identifier: GPL-2.0+ |
| */ |
| #include <linux/types.h> |
| #include <common.h> |
| #include <asm/armv7.h> |
| #include <asm/utils.h> |
| |
| #define ARMV7_DCACHE_INVAL_ALL 1 |
| #define ARMV7_DCACHE_CLEAN_INVAL_ALL 2 |
| #define ARMV7_DCACHE_INVAL_RANGE 3 |
| #define ARMV7_DCACHE_CLEAN_INVAL_RANGE 4 |
| |
| #ifndef CONFIG_SYS_DCACHE_OFF |
| static int check_cache_range(unsigned long start, unsigned long stop) |
| { |
| int ok = 1; |
| |
| if (start & (CONFIG_SYS_CACHELINE_SIZE - 1)) |
| ok = 0; |
| |
| if (stop & (CONFIG_SYS_CACHELINE_SIZE - 1)) |
| ok = 0; |
| |
| if (!ok) |
| debug("CACHE: Misaligned operation at range [%08lx, %08lx]\n", |
| start, stop); |
| |
| return ok; |
| } |
| |
| /* |
| * Write the level and type you want to Cache Size Selection Register(CSSELR) |
| * to get size details from Current Cache Size ID Register(CCSIDR) |
| */ |
| static void set_csselr(u32 level, u32 type) |
| { |
| u32 csselr = level << 1 | type; |
| |
| /* Write to Cache Size Selection Register(CSSELR) */ |
| asm volatile ("mcr p15, 2, %0, c0, c0, 0" : : "r" (csselr)); |
| } |
| |
| static u32 get_ccsidr(void) |
| { |
| u32 ccsidr; |
| |
| /* Read current CP15 Cache Size ID Register */ |
| asm volatile ("mrc p15, 1, %0, c0, c0, 0" : "=r" (ccsidr)); |
| return ccsidr; |
| } |
| |
| static u32 get_clidr(void) |
| { |
| u32 clidr; |
| |
| /* Read current CP15 Cache Level ID Register */ |
| asm volatile ("mrc p15,1,%0,c0,c0,1" : "=r" (clidr)); |
| return clidr; |
| } |
| |
| static void v7_inval_dcache_level_setway(u32 level, u32 num_sets, |
| u32 num_ways, u32 way_shift, |
| u32 log2_line_len) |
| { |
| int way, set; |
| u32 setway; |
| |
| /* |
| * For optimal assembly code: |
| * a. count down |
| * b. have bigger loop inside |
| */ |
| for (way = num_ways - 1; way >= 0 ; way--) { |
| for (set = num_sets - 1; set >= 0; set--) { |
| setway = (level << 1) | (set << log2_line_len) | |
| (way << way_shift); |
| /* Invalidate data/unified cache line by set/way */ |
| asm volatile (" mcr p15, 0, %0, c7, c6, 2" |
| : : "r" (setway)); |
| } |
| } |
| /* DSB to make sure the operation is complete */ |
| DSB; |
| } |
| |
| static void v7_clean_inval_dcache_level_setway(u32 level, u32 num_sets, |
| u32 num_ways, u32 way_shift, |
| u32 log2_line_len) |
| { |
| int way, set; |
| u32 setway; |
| |
| /* |
| * For optimal assembly code: |
| * a. count down |
| * b. have bigger loop inside |
| */ |
| for (way = num_ways - 1; way >= 0 ; way--) { |
| for (set = num_sets - 1; set >= 0; set--) { |
| setway = (level << 1) | (set << log2_line_len) | |
| (way << way_shift); |
| /* |
| * Clean & Invalidate data/unified |
| * cache line by set/way |
| */ |
| asm volatile (" mcr p15, 0, %0, c7, c14, 2" |
| : : "r" (setway)); |
| } |
| } |
| /* DSB to make sure the operation is complete */ |
| DSB; |
| } |
| |
| static void v7_maint_dcache_level_setway(u32 level, u32 operation) |
| { |
| u32 ccsidr; |
| u32 num_sets, num_ways, log2_line_len, log2_num_ways; |
| u32 way_shift; |
| |
| set_csselr(level, ARMV7_CSSELR_IND_DATA_UNIFIED); |
| |
| ccsidr = get_ccsidr(); |
| |
| log2_line_len = ((ccsidr & CCSIDR_LINE_SIZE_MASK) >> |
| CCSIDR_LINE_SIZE_OFFSET) + 2; |
| /* Converting from words to bytes */ |
| log2_line_len += 2; |
| |
| num_ways = ((ccsidr & CCSIDR_ASSOCIATIVITY_MASK) >> |
| CCSIDR_ASSOCIATIVITY_OFFSET) + 1; |
| num_sets = ((ccsidr & CCSIDR_NUM_SETS_MASK) >> |
| CCSIDR_NUM_SETS_OFFSET) + 1; |
| /* |
| * According to ARMv7 ARM number of sets and number of ways need |
| * not be a power of 2 |
| */ |
| log2_num_ways = log_2_n_round_up(num_ways); |
| |
| way_shift = (32 - log2_num_ways); |
| if (operation == ARMV7_DCACHE_INVAL_ALL) { |
| v7_inval_dcache_level_setway(level, num_sets, num_ways, |
| way_shift, log2_line_len); |
| } else if (operation == ARMV7_DCACHE_CLEAN_INVAL_ALL) { |
| v7_clean_inval_dcache_level_setway(level, num_sets, num_ways, |
| way_shift, log2_line_len); |
| } |
| } |
| |
| static void v7_maint_dcache_all(u32 operation) |
| { |
| u32 level, cache_type, level_start_bit = 0; |
| u32 clidr = get_clidr(); |
| |
| for (level = 0; level < 7; level++) { |
| cache_type = (clidr >> level_start_bit) & 0x7; |
| if ((cache_type == ARMV7_CLIDR_CTYPE_DATA_ONLY) || |
| (cache_type == ARMV7_CLIDR_CTYPE_INSTRUCTION_DATA) || |
| (cache_type == ARMV7_CLIDR_CTYPE_UNIFIED)) |
| v7_maint_dcache_level_setway(level, operation); |
| level_start_bit += 3; |
| } |
| } |
| |
| static void v7_dcache_clean_inval_range(u32 start, u32 stop, u32 line_len) |
| { |
| u32 mva; |
| |
| /* Align start to cache line boundary */ |
| start &= ~(line_len - 1); |
| for (mva = start; mva < stop; mva = mva + line_len) { |
| /* DCCIMVAC - Clean & Invalidate data cache by MVA to PoC */ |
| asm volatile ("mcr p15, 0, %0, c7, c14, 1" : : "r" (mva)); |
| } |
| } |
| |
| static void v7_dcache_inval_range(u32 start, u32 stop, u32 line_len) |
| { |
| u32 mva; |
| |
| /* |
| * If start address is not aligned to cache-line do not |
| * invalidate the first cache-line |
| */ |
| if (start & (line_len - 1)) { |
| printf("ERROR: %s - start address is not aligned - 0x%08x\n", |
| __func__, start); |
| /* move to next cache line */ |
| start = (start + line_len - 1) & ~(line_len - 1); |
| } |
| |
| /* |
| * If stop address is not aligned to cache-line do not |
| * invalidate the last cache-line |
| */ |
| if (stop & (line_len - 1)) { |
| printf("ERROR: %s - stop address is not aligned - 0x%08x\n", |
| __func__, stop); |
| /* align to the beginning of this cache line */ |
| stop &= ~(line_len - 1); |
| } |
| |
| for (mva = start; mva < stop; mva = mva + line_len) { |
| /* DCIMVAC - Invalidate data cache by MVA to PoC */ |
| asm volatile ("mcr p15, 0, %0, c7, c6, 1" : : "r" (mva)); |
| } |
| } |
| |
| static void v7_dcache_maint_range(u32 start, u32 stop, u32 range_op) |
| { |
| u32 line_len, ccsidr; |
| |
| ccsidr = get_ccsidr(); |
| line_len = ((ccsidr & CCSIDR_LINE_SIZE_MASK) >> |
| CCSIDR_LINE_SIZE_OFFSET) + 2; |
| /* Converting from words to bytes */ |
| line_len += 2; |
| /* converting from log2(linelen) to linelen */ |
| line_len = 1 << line_len; |
| |
| switch (range_op) { |
| case ARMV7_DCACHE_CLEAN_INVAL_RANGE: |
| v7_dcache_clean_inval_range(start, stop, line_len); |
| break; |
| case ARMV7_DCACHE_INVAL_RANGE: |
| v7_dcache_inval_range(start, stop, line_len); |
| break; |
| } |
| |
| /* DSB to make sure the operation is complete */ |
| DSB; |
| } |
| |
| /* Invalidate TLB */ |
| static void v7_inval_tlb(void) |
| { |
| /* Invalidate entire unified TLB */ |
| asm volatile ("mcr p15, 0, %0, c8, c7, 0" : : "r" (0)); |
| /* Invalidate entire data TLB */ |
| asm volatile ("mcr p15, 0, %0, c8, c6, 0" : : "r" (0)); |
| /* Invalidate entire instruction TLB */ |
| asm volatile ("mcr p15, 0, %0, c8, c5, 0" : : "r" (0)); |
| /* Full system DSB - make sure that the invalidation is complete */ |
| DSB; |
| /* Full system ISB - make sure the instruction stream sees it */ |
| ISB; |
| } |
| |
| void invalidate_dcache_all(void) |
| { |
| v7_maint_dcache_all(ARMV7_DCACHE_INVAL_ALL); |
| |
| v7_outer_cache_inval_all(); |
| } |
| |
| /* |
| * Performs a clean & invalidation of the entire data cache |
| * at all levels |
| */ |
| void flush_dcache_all(void) |
| { |
| v7_maint_dcache_all(ARMV7_DCACHE_CLEAN_INVAL_ALL); |
| |
| v7_outer_cache_flush_all(); |
| } |
| |
| /* |
| * Invalidates range in all levels of D-cache/unified cache used: |
| * Affects the range [start, stop - 1] |
| */ |
| void invalidate_dcache_range(unsigned long start, unsigned long stop) |
| { |
| check_cache_range(start, stop); |
| |
| v7_dcache_maint_range(start, stop, ARMV7_DCACHE_INVAL_RANGE); |
| |
| v7_outer_cache_inval_range(start, stop); |
| } |
| |
| /* |
| * Flush range(clean & invalidate) from all levels of D-cache/unified |
| * cache used: |
| * Affects the range [start, stop - 1] |
| */ |
| void flush_dcache_range(unsigned long start, unsigned long stop) |
| { |
| check_cache_range(start, stop); |
| |
| v7_dcache_maint_range(start, stop, ARMV7_DCACHE_CLEAN_INVAL_RANGE); |
| |
| v7_outer_cache_flush_range(start, stop); |
| } |
| |
| void arm_init_before_mmu(void) |
| { |
| v7_outer_cache_enable(); |
| invalidate_dcache_all(); |
| v7_inval_tlb(); |
| } |
| |
| void mmu_page_table_flush(unsigned long start, unsigned long stop) |
| { |
| flush_dcache_range(start, stop); |
| v7_inval_tlb(); |
| } |
| #else /* #ifndef CONFIG_SYS_DCACHE_OFF */ |
| void invalidate_dcache_all(void) |
| { |
| } |
| |
| void flush_dcache_all(void) |
| { |
| } |
| |
| void arm_init_before_mmu(void) |
| { |
| } |
| |
| void mmu_page_table_flush(unsigned long start, unsigned long stop) |
| { |
| } |
| |
| void arm_init_domains(void) |
| { |
| } |
| #endif /* #ifndef CONFIG_SYS_DCACHE_OFF */ |
| |
| #ifndef CONFIG_SYS_ICACHE_OFF |
| /* Invalidate entire I-cache and branch predictor array */ |
| void invalidate_icache_all(void) |
| { |
| /* |
| * Invalidate all instruction caches to PoU. |
| * Also flushes branch target cache. |
| */ |
| asm volatile ("mcr p15, 0, %0, c7, c5, 0" : : "r" (0)); |
| |
| /* Invalidate entire branch predictor array */ |
| asm volatile ("mcr p15, 0, %0, c7, c5, 6" : : "r" (0)); |
| |
| /* Full system DSB - make sure that the invalidation is complete */ |
| DSB; |
| |
| /* ISB - make sure the instruction stream sees it */ |
| ISB; |
| } |
| #else |
| void invalidate_icache_all(void) |
| { |
| } |
| #endif |
| |
| /* Stub implementations for outer cache operations */ |
| __weak void v7_outer_cache_enable(void) {} |
| __weak void v7_outer_cache_disable(void) {} |
| __weak void v7_outer_cache_flush_all(void) {} |
| __weak void v7_outer_cache_inval_all(void) {} |
| __weak void v7_outer_cache_flush_range(u32 start, u32 end) {} |
| __weak void v7_outer_cache_inval_range(u32 start, u32 end) {} |