blob: a082a810749d23ba84716a99c350bd847da58db4 [file] [log] [blame]
Jeenu Viswambharan5c503042017-05-26 14:15:40 +01001/*
Alexei Fedorov6b4a5f02019-04-26 12:07:07 +01002 * Copyright (c) 2017-2019, ARM Limited and Contributors. All rights reserved.
Jeenu Viswambharan5c503042017-05-26 14:15:40 +01003 *
4 * SPDX-License-Identifier: BSD-3-Clause
5 */
6
Alexei Fedorov6b4a5f02019-04-26 12:07:07 +01007#include <common/debug.h>
Daniel Boulby844b4872018-09-18 13:36:39 +01008#include <cdefs.h>
Antonio Nino Diaze0f90632018-12-14 00:18:21 +00009#include <drivers/arm/smmu_v3.h>
Deepika Bhavnanibda60d32019-10-31 14:09:52 -060010#include <drivers/delay_timer.h>
Antonio Nino Diaze0f90632018-12-14 00:18:21 +000011#include <lib/mmio.h>
12
Alexei Fedorov6b4a5f02019-04-26 12:07:07 +010013/* SMMU poll number of retries */
Deepika Bhavnanibda60d32019-10-31 14:09:52 -060014#define SMMU_POLL_TIMEOUT_US U(1000)
Jeenu Viswambharan5c503042017-05-26 14:15:40 +010015
Alexei Fedorov6b4a5f02019-04-26 12:07:07 +010016static int __init smmuv3_poll(uintptr_t smmu_reg, uint32_t mask,
17 uint32_t value)
Jeenu Viswambharan5c503042017-05-26 14:15:40 +010018{
Deepika Bhavnanibda60d32019-10-31 14:09:52 -060019 uint32_t reg_val;
20 uint64_t timeout;
Jeenu Viswambharan5c503042017-05-26 14:15:40 +010021
Deepika Bhavnanibda60d32019-10-31 14:09:52 -060022 /* Set 1ms timeout value */
23 timeout = timeout_init_us(SMMU_POLL_TIMEOUT_US);
Alexei Fedorov6b4a5f02019-04-26 12:07:07 +010024 do {
25 reg_val = mmio_read_32(smmu_reg);
26 if ((reg_val & mask) == value)
27 return 0;
Deepika Bhavnanibda60d32019-10-31 14:09:52 -060028 } while (!timeout_elapsed(timeout));
Jeenu Viswambharan5c503042017-05-26 14:15:40 +010029
Deepika Bhavnanibda60d32019-10-31 14:09:52 -060030 ERROR("Timeout polling SMMUv3 register @%p\n", (void *)smmu_reg);
Alexei Fedorov6b4a5f02019-04-26 12:07:07 +010031 ERROR("Read value 0x%x, expected 0x%x\n", reg_val,
32 value == 0U ? reg_val & ~mask : reg_val | mask);
33 return -1;
Antonio Nino Diazfeacba32018-08-21 16:12:29 +010034}
35
Jeenu Viswambharan5c503042017-05-26 14:15:40 +010036/*
Alexei Fedorov896799a2019-05-09 12:14:40 +010037 * Abort all incoming transactions in order to implement a default
38 * deny policy on reset.
39 */
40int __init smmuv3_security_init(uintptr_t smmu_base)
41{
42 /* Attribute update has completed when SMMU_(S)_GBPA.Update bit is 0 */
43 if (smmuv3_poll(smmu_base + SMMU_GBPA, SMMU_GBPA_UPDATE, 0U) != 0U)
44 return -1;
45
46 /*
47 * SMMU_(S)_CR0 resets to zero with all streams bypassing the SMMU,
48 * so just abort all incoming transactions.
49 */
50 mmio_setbits_32(smmu_base + SMMU_GBPA,
51 SMMU_GBPA_UPDATE | SMMU_GBPA_ABORT);
52
53 if (smmuv3_poll(smmu_base + SMMU_GBPA, SMMU_GBPA_UPDATE, 0U) != 0U)
54 return -1;
55
56 /* Check if the SMMU supports secure state */
57 if ((mmio_read_32(smmu_base + SMMU_S_IDR1) &
58 SMMU_S_IDR1_SECURE_IMPL) == 0U)
59 return 0;
60
61 /* Abort all incoming secure transactions */
62 if (smmuv3_poll(smmu_base + SMMU_S_GBPA, SMMU_S_GBPA_UPDATE, 0U) != 0U)
63 return -1;
64
65 mmio_setbits_32(smmu_base + SMMU_S_GBPA,
66 SMMU_S_GBPA_UPDATE | SMMU_S_GBPA_ABORT);
67
68 return smmuv3_poll(smmu_base + SMMU_S_GBPA, SMMU_S_GBPA_UPDATE, 0U);
69}
70
71/*
Jeenu Viswambharan5c503042017-05-26 14:15:40 +010072 * Initialize the SMMU by invalidating all secure caches and TLBs.
Alexei Fedorov6b4a5f02019-04-26 12:07:07 +010073 * Abort all incoming transactions in order to implement a default
74 * deny policy on reset
Jeenu Viswambharan5c503042017-05-26 14:15:40 +010075 */
Daniel Boulby844b4872018-09-18 13:36:39 +010076int __init smmuv3_init(uintptr_t smmu_base)
Jeenu Viswambharan5c503042017-05-26 14:15:40 +010077{
Alexei Fedorov896799a2019-05-09 12:14:40 +010078 /* Abort all incoming transactions */
79 if (smmuv3_security_init(smmu_base) != 0)
80 return -1;
81
82 /* Check if the SMMU supports secure state */
83 if ((mmio_read_32(smmu_base + SMMU_S_IDR1) &
84 SMMU_S_IDR1_SECURE_IMPL) == 0U)
85 return 0;
Jeenu Viswambharan5c503042017-05-26 14:15:40 +010086 /*
Alexei Fedorov896799a2019-05-09 12:14:40 +010087 * Initiate invalidation of secure caches and TLBs if the SMMU
88 * supports secure state. If not, it's implementation defined
89 * as to how SMMU_S_INIT register is accessed.
Jeenu Viswambharan5c503042017-05-26 14:15:40 +010090 */
Alexei Fedorov896799a2019-05-09 12:14:40 +010091 mmio_write_32(smmu_base + SMMU_S_INIT, SMMU_S_INIT_INV_ALL);
Jeenu Viswambharan5c503042017-05-26 14:15:40 +010092
Alexei Fedorov896799a2019-05-09 12:14:40 +010093 /* Wait for global invalidation operation to finish */
94 return smmuv3_poll(smmu_base + SMMU_S_INIT,
95 SMMU_S_INIT_INV_ALL, 0U);
Jeenu Viswambharan5c503042017-05-26 14:15:40 +010096}