Aneesh V | 960f5c0 | 2011-06-16 23:30:47 +0000 | [diff] [blame] | 1 | /* |
| 2 | * (C) Copyright 2010 |
| 3 | * Texas Instruments, <www.ti.com> |
| 4 | * Aneesh V <aneesh@ti.com> |
| 5 | * |
Wolfgang Denk | d79de1d | 2013-07-08 09:37:19 +0200 | [diff] [blame] | 6 | * SPDX-License-Identifier: GPL-2.0+ |
Aneesh V | 960f5c0 | 2011-06-16 23:30:47 +0000 | [diff] [blame] | 7 | */ |
| 8 | #ifndef ARMV7_H |
| 9 | #define ARMV7_H |
Aneesh V | 960f5c0 | 2011-06-16 23:30:47 +0000 | [diff] [blame] | 10 | |
Aneesh V | 162ced3 | 2011-07-21 09:10:04 -0400 | [diff] [blame] | 11 | /* Cortex-A9 revisions */ |
| 12 | #define MIDR_CORTEX_A9_R0P1 0x410FC091 |
| 13 | #define MIDR_CORTEX_A9_R1P2 0x411FC092 |
| 14 | #define MIDR_CORTEX_A9_R1P3 0x411FC093 |
Aneesh V | 0b92f09 | 2011-07-21 09:29:23 -0400 | [diff] [blame] | 15 | #define MIDR_CORTEX_A9_R2P10 0x412FC09A |
Aneesh V | 162ced3 | 2011-07-21 09:10:04 -0400 | [diff] [blame] | 16 | |
Sricharan | 9310ff7 | 2011-11-15 09:49:55 -0500 | [diff] [blame] | 17 | /* Cortex-A15 revisions */ |
| 18 | #define MIDR_CORTEX_A15_R0P0 0x410FC0F0 |
SRICHARAN R | cf85056 | 2013-02-12 01:33:41 +0000 | [diff] [blame] | 19 | #define MIDR_CORTEX_A15_R2P2 0x412FC0F2 |
Sricharan | 9310ff7 | 2011-11-15 09:49:55 -0500 | [diff] [blame] | 20 | |
Andre Przywara | dd5e8da | 2013-09-19 18:06:41 +0200 | [diff] [blame] | 21 | /* Cortex-A7 revisions */ |
| 22 | #define MIDR_CORTEX_A7_R0P0 0x410FC070 |
| 23 | |
| 24 | #define MIDR_PRIMARY_PART_MASK 0xFF0FFFF0 |
| 25 | |
| 26 | /* ID_PFR1 feature fields */ |
| 27 | #define CPUID_ARM_SEC_SHIFT 4 |
| 28 | #define CPUID_ARM_SEC_MASK (0xF << CPUID_ARM_SEC_SHIFT) |
| 29 | #define CPUID_ARM_VIRT_SHIFT 12 |
| 30 | #define CPUID_ARM_VIRT_MASK (0xF << CPUID_ARM_VIRT_SHIFT) |
| 31 | #define CPUID_ARM_GENTIMER_SHIFT 16 |
| 32 | #define CPUID_ARM_GENTIMER_MASK (0xF << CPUID_ARM_GENTIMER_SHIFT) |
| 33 | |
| 34 | /* valid bits in CBAR register / PERIPHBASE value */ |
| 35 | #define CBAR_MASK 0xFFFF8000 |
| 36 | |
Aneesh V | 960f5c0 | 2011-06-16 23:30:47 +0000 | [diff] [blame] | 37 | /* CCSIDR */ |
| 38 | #define CCSIDR_LINE_SIZE_OFFSET 0 |
| 39 | #define CCSIDR_LINE_SIZE_MASK 0x7 |
| 40 | #define CCSIDR_ASSOCIATIVITY_OFFSET 3 |
| 41 | #define CCSIDR_ASSOCIATIVITY_MASK (0x3FF << 3) |
| 42 | #define CCSIDR_NUM_SETS_OFFSET 13 |
| 43 | #define CCSIDR_NUM_SETS_MASK (0x7FFF << 13) |
| 44 | |
| 45 | /* |
| 46 | * Values for InD field in CSSELR |
| 47 | * Selects the type of cache |
| 48 | */ |
| 49 | #define ARMV7_CSSELR_IND_DATA_UNIFIED 0 |
| 50 | #define ARMV7_CSSELR_IND_INSTRUCTION 1 |
| 51 | |
| 52 | /* Values for Ctype fields in CLIDR */ |
| 53 | #define ARMV7_CLIDR_CTYPE_NO_CACHE 0 |
| 54 | #define ARMV7_CLIDR_CTYPE_INSTRUCTION_ONLY 1 |
| 55 | #define ARMV7_CLIDR_CTYPE_DATA_ONLY 2 |
| 56 | #define ARMV7_CLIDR_CTYPE_INSTRUCTION_DATA 3 |
| 57 | #define ARMV7_CLIDR_CTYPE_UNIFIED 4 |
| 58 | |
Andre Przywara | a6bb668 | 2013-09-19 18:06:39 +0200 | [diff] [blame] | 59 | #ifndef __ASSEMBLY__ |
| 60 | #include <linux/types.h> |
Tom Rini | e968973 | 2015-03-02 08:24:45 -0500 | [diff] [blame] | 61 | #include <asm/io.h> |
Andre Przywara | a6bb668 | 2013-09-19 18:06:39 +0200 | [diff] [blame] | 62 | |
Aneesh V | 960f5c0 | 2011-06-16 23:30:47 +0000 | [diff] [blame] | 63 | /* |
| 64 | * CP15 Barrier instructions |
| 65 | * Please note that we have separate barrier instructions in ARMv7 |
| 66 | * However, we use the CP15 based instructtions because we use |
| 67 | * -march=armv5 in U-Boot |
| 68 | */ |
| 69 | #define CP15ISB asm volatile ("mcr p15, 0, %0, c7, c5, 4" : : "r" (0)) |
| 70 | #define CP15DSB asm volatile ("mcr p15, 0, %0, c7, c10, 4" : : "r" (0)) |
| 71 | #define CP15DMB asm volatile ("mcr p15, 0, %0, c7, c10, 5" : : "r" (0)) |
| 72 | |
Akshay Saraswat | e5be413 | 2015-02-20 13:27:13 +0530 | [diff] [blame] | 73 | /* |
| 74 | * Workaround for ARM errata # 798870 |
| 75 | * Set L2ACTLR[7] to reissue any memory transaction in the L2 that has been |
| 76 | * stalled for 1024 cycles to verify that its hazard condition still exists. |
| 77 | */ |
| 78 | static inline void v7_enable_l2_hazard_detect(void) |
| 79 | { |
| 80 | uint32_t val; |
| 81 | |
| 82 | /* L2ACTLR[7]: Enable hazard detect timeout */ |
| 83 | asm volatile ("mrc p15, 1, %0, c15, c0, 0\n\t" : "=r"(val)); |
| 84 | val |= (1 << 7); |
| 85 | asm volatile ("mcr p15, 1, %0, c15, c0, 0\n\t" : : "r"(val)); |
| 86 | } |
| 87 | |
Akshay Saraswat | 57fd639 | 2015-02-20 13:27:14 +0530 | [diff] [blame] | 88 | /* |
| 89 | * Workaround for ARM errata # 799270 |
| 90 | * Ensure that the L2 logic has been used within the previous 256 cycles |
| 91 | * before modifying the ACTLR.SMP bit. This is required during boot before |
| 92 | * MMU has been enabled, or during a specified reset or power down sequence. |
| 93 | */ |
| 94 | static inline void v7_enable_smp(uint32_t address) |
| 95 | { |
| 96 | uint32_t temp, val; |
| 97 | |
| 98 | /* Read auxiliary control register */ |
| 99 | asm volatile ("mrc p15, 0, %0, c1, c0, 1\n\t" : "=r"(val)); |
| 100 | |
| 101 | /* Enable SMP */ |
| 102 | val |= (1 << 6); |
| 103 | |
| 104 | /* Dummy read to assure L2 access */ |
| 105 | temp = readl(address); |
| 106 | temp &= 0; |
| 107 | val |= temp; |
| 108 | |
| 109 | /* Write auxiliary control register */ |
| 110 | asm volatile ("mcr p15, 0, %0, c1, c0, 1\n\t" : : "r"(val)); |
| 111 | |
| 112 | CP15DSB; |
| 113 | CP15ISB; |
| 114 | } |
| 115 | |
Akshay Saraswat | e5be413 | 2015-02-20 13:27:13 +0530 | [diff] [blame] | 116 | void v7_en_l2_hazard_detect(void); |
Aneesh V | 960f5c0 | 2011-06-16 23:30:47 +0000 | [diff] [blame] | 117 | void v7_outer_cache_enable(void); |
| 118 | void v7_outer_cache_disable(void); |
| 119 | void v7_outer_cache_flush_all(void); |
| 120 | void v7_outer_cache_inval_all(void); |
| 121 | void v7_outer_cache_flush_range(u32 start, u32 end); |
| 122 | void v7_outer_cache_inval_range(u32 start, u32 end); |
| 123 | |
Andre Przywara | 8de142c | 2013-09-19 18:06:45 +0200 | [diff] [blame] | 124 | #if defined(CONFIG_ARMV7_NONSEC) || defined(CONFIG_ARMV7_VIRT) |
Andre Przywara | ad5ad74 | 2013-09-19 18:06:42 +0200 | [diff] [blame] | 125 | |
Marc Zyngier | 855ca66 | 2014-07-12 14:24:03 +0100 | [diff] [blame] | 126 | int armv7_init_nonsec(void); |
Ian Campbell | 68bc8f5 | 2014-12-21 09:45:11 +0000 | [diff] [blame] | 127 | bool armv7_boot_nonsec(void); |
Andre Przywara | ad5ad74 | 2013-09-19 18:06:42 +0200 | [diff] [blame] | 128 | |
Andre Przywara | dd5e8da | 2013-09-19 18:06:41 +0200 | [diff] [blame] | 129 | /* defined in assembly file */ |
| 130 | unsigned int _nonsec_init(void); |
Marc Zyngier | 855ca66 | 2014-07-12 14:24:03 +0100 | [diff] [blame] | 131 | void _do_nonsec_entry(void *target_pc, unsigned long r0, |
| 132 | unsigned long r1, unsigned long r2); |
Andre Przywara | dbbe196 | 2013-09-19 18:06:44 +0200 | [diff] [blame] | 133 | void _smp_pen(void); |
Marc Zyngier | 855ca66 | 2014-07-12 14:24:03 +0100 | [diff] [blame] | 134 | |
| 135 | extern char __secure_start[]; |
| 136 | extern char __secure_end[]; |
| 137 | |
Andre Przywara | 8de142c | 2013-09-19 18:06:45 +0200 | [diff] [blame] | 138 | #endif /* CONFIG_ARMV7_NONSEC || CONFIG_ARMV7_VIRT */ |
Andre Przywara | dd5e8da | 2013-09-19 18:06:41 +0200 | [diff] [blame] | 139 | |
Nishanth Menon | aa0294e | 2015-03-09 17:11:59 -0500 | [diff] [blame^] | 140 | void v7_arch_cp15_set_l2aux_ctrl(u32 l2auxctrl, u32 cpu_midr, |
| 141 | u32 cpu_rev_comb, u32 cpu_variant, |
| 142 | u32 cpu_rev); |
Andre Przywara | a6bb668 | 2013-09-19 18:06:39 +0200 | [diff] [blame] | 143 | #endif /* ! __ASSEMBLY__ */ |
| 144 | |
Aneesh V | 960f5c0 | 2011-06-16 23:30:47 +0000 | [diff] [blame] | 145 | #endif |