| /* |
| * Copyright (c) 2019, ARM Limited and Contributors. All rights reserved. |
| * |
| * SPDX-License-Identifier: BSD-3-Clause |
| */ |
| |
| #include <assert.h> |
| #include <errno.h> |
| |
| #include <platform_def.h> |
| |
| #include <arch_helpers.h> |
| #include <bl31/bl31.h> |
| #include <common/debug.h> |
| #include <drivers/console.h> |
| #include <drivers/delay_timer.h> |
| #include <lib/bakery_lock.h> |
| #include <lib/mmio.h> |
| #include <plat/common/platform.h> |
| |
| #include <cpus_on_fixed_addr.h> |
| #include <plat_private.h> |
| #include <pmu.h> |
| #include <px30_def.h> |
| #include <secure.h> |
| #include <soc.h> |
| |
| DEFINE_BAKERY_LOCK(rockchip_pd_lock); |
| #define rockchip_pd_lock_init() bakery_lock_init(&rockchip_pd_lock) |
| #define rockchip_pd_lock_get() bakery_lock_get(&rockchip_pd_lock) |
| #define rockchip_pd_lock_rls() bakery_lock_release(&rockchip_pd_lock) |
| |
| static struct psram_data_t *psram_boot_cfg = |
| (struct psram_data_t *)&sys_sleep_flag_sram; |
| |
| /* |
| * There are two ways to powering on or off on core. |
| * 1) Control it power domain into on or off in PMU_PWRDN_CON reg, |
| * it is core_pwr_pd mode |
| * 2) Enable the core power manage in PMU_CORE_PM_CON reg, |
| * then, if the core enter into wfi, it power domain will be |
| * powered off automatically. it is core_pwr_wfi or core_pwr_wfi_int mode |
| * so we need core_pm_cfg_info to distinguish which method be used now. |
| */ |
| |
| static uint32_t cores_pd_cfg_info[PLATFORM_CORE_COUNT] |
| #if USE_COHERENT_MEM |
| __attribute__ ((section(".tzfw_coherent_mem"))) |
| #endif |
| ; |
| |
| struct px30_sleep_ddr_data { |
| uint32_t clk_sel0; |
| uint32_t cru_mode_save; |
| uint32_t cru_pmu_mode_save; |
| uint32_t ddrc_hwlpctl; |
| uint32_t ddrc_pwrctrl; |
| uint32_t ddrgrf_con0; |
| uint32_t ddrgrf_con1; |
| uint32_t ddrstdby_con0; |
| uint32_t gpio0b_iomux; |
| uint32_t gpio0c_iomux; |
| uint32_t pmu_pwrmd_core_l; |
| uint32_t pmu_pwrmd_core_h; |
| uint32_t pmu_pwrmd_cmm_l; |
| uint32_t pmu_pwrmd_cmm_h; |
| uint32_t pmu_wkup_cfg2_l; |
| uint32_t pmu_cru_clksel_con0; |
| uint32_t pmugrf_soc_con0; |
| uint32_t pmusgrf_soc_con0; |
| uint32_t pmic_slp_iomux; |
| uint32_t pgrf_pvtm_con[2]; |
| uint32_t cru_clk_gate[CRU_CLKGATES_CON_CNT]; |
| uint32_t cru_pmu_clk_gate[CRU_PMU_CLKGATE_CON_CNT]; |
| uint32_t cru_plls_con_save[END_PLL_ID][PLL_CON_CNT]; |
| uint32_t cpu_qos[CPU_AXI_QOS_NUM_REGS]; |
| uint32_t gpu_qos[CPU_AXI_QOS_NUM_REGS]; |
| uint32_t isp_128m_qos[CPU_AXI_QOS_NUM_REGS]; |
| uint32_t isp_rd_qos[CPU_AXI_QOS_NUM_REGS]; |
| uint32_t isp_wr_qos[CPU_AXI_QOS_NUM_REGS]; |
| uint32_t isp_m1_qos[CPU_AXI_QOS_NUM_REGS]; |
| uint32_t vip_qos[CPU_AXI_QOS_NUM_REGS]; |
| uint32_t rga_rd_qos[CPU_AXI_QOS_NUM_REGS]; |
| uint32_t rga_wr_qos[CPU_AXI_QOS_NUM_REGS]; |
| uint32_t vop_m0_qos[CPU_AXI_QOS_NUM_REGS]; |
| uint32_t vop_m1_qos[CPU_AXI_QOS_NUM_REGS]; |
| uint32_t vpu_qos[CPU_AXI_QOS_NUM_REGS]; |
| uint32_t vpu_r128_qos[CPU_AXI_QOS_NUM_REGS]; |
| uint32_t dcf_qos[CPU_AXI_QOS_NUM_REGS]; |
| uint32_t dmac_qos[CPU_AXI_QOS_NUM_REGS]; |
| uint32_t crypto_qos[CPU_AXI_QOS_NUM_REGS]; |
| uint32_t gmac_qos[CPU_AXI_QOS_NUM_REGS]; |
| uint32_t emmc_qos[CPU_AXI_QOS_NUM_REGS]; |
| uint32_t nand_qos[CPU_AXI_QOS_NUM_REGS]; |
| uint32_t sdio_qos[CPU_AXI_QOS_NUM_REGS]; |
| uint32_t sfc_qos[CPU_AXI_QOS_NUM_REGS]; |
| uint32_t sdmmc_qos[CPU_AXI_QOS_NUM_REGS]; |
| uint32_t usb_host_qos[CPU_AXI_QOS_NUM_REGS]; |
| uint32_t usb_otg_qos[CPU_AXI_QOS_NUM_REGS]; |
| }; |
| |
| static struct px30_sleep_ddr_data ddr_data |
| #if USE_COHERENT_MEM |
| __attribute__ ((section(".tzfw_coherent_mem"))) |
| #endif |
| ; |
| |
| static inline uint32_t get_cpus_pwr_domain_cfg_info(uint32_t cpu_id) |
| { |
| assert(cpu_id < PLATFORM_CORE_COUNT); |
| return cores_pd_cfg_info[cpu_id]; |
| } |
| |
| static inline void set_cpus_pwr_domain_cfg_info(uint32_t cpu_id, uint32_t value) |
| { |
| assert(cpu_id < PLATFORM_CORE_COUNT); |
| cores_pd_cfg_info[cpu_id] = value; |
| #if !USE_COHERENT_MEM |
| flush_dcache_range((uintptr_t)&cores_pd_cfg_info[cpu_id], |
| sizeof(uint32_t)); |
| #endif |
| } |
| |
| static inline uint32_t pmu_power_domain_st(uint32_t pd) |
| { |
| return mmio_read_32(PMU_BASE + PMU_PWRDN_ST) & BIT(pd) ? |
| pmu_pd_off : |
| pmu_pd_on; |
| } |
| |
| static int pmu_power_domain_ctr(uint32_t pd, uint32_t pd_state) |
| { |
| uint32_t loop = 0; |
| int ret = 0; |
| |
| rockchip_pd_lock_get(); |
| |
| mmio_write_32(PMU_BASE + PMU_PWRDN_CON, |
| BITS_WITH_WMASK(pd_state, 0x1, pd)); |
| dsb(); |
| |
| while ((pmu_power_domain_st(pd) != pd_state) && (loop < PD_CTR_LOOP)) { |
| udelay(1); |
| loop++; |
| } |
| |
| if (pmu_power_domain_st(pd) != pd_state) { |
| WARN("%s: %d, %d, error!\n", __func__, pd, pd_state); |
| ret = -EINVAL; |
| } |
| |
| rockchip_pd_lock_rls(); |
| |
| return ret; |
| } |
| |
| static inline uint32_t pmu_bus_idle_st(uint32_t bus) |
| { |
| return !!((mmio_read_32(PMU_BASE + PMU_BUS_IDLE_ST) & BIT(bus)) && |
| (mmio_read_32(PMU_BASE + PMU_BUS_IDLE_ST) & BIT(bus + 16))); |
| } |
| |
| static void pmu_bus_idle_req(uint32_t bus, uint32_t state) |
| { |
| uint32_t wait_cnt = 0; |
| |
| mmio_write_32(PMU_BASE + PMU_BUS_IDLE_REQ, |
| BITS_WITH_WMASK(state, 0x1, bus)); |
| |
| while (pmu_bus_idle_st(bus) != state && |
| wait_cnt < BUS_IDLE_LOOP) { |
| udelay(1); |
| wait_cnt++; |
| } |
| |
| if (pmu_bus_idle_st(bus) != state) |
| WARN("%s:idle_st=0x%x, bus_id=%d\n", |
| __func__, mmio_read_32(PMU_BASE + PMU_BUS_IDLE_ST), bus); |
| } |
| |
| static void qos_save(void) |
| { |
| /* scu powerdomain will power off, so cpu qos should be saved */ |
| SAVE_QOS(ddr_data.cpu_qos, CPU); |
| |
| if (pmu_power_domain_st(PD_GPU) == pmu_pd_on) |
| SAVE_QOS(ddr_data.gpu_qos, GPU); |
| if (pmu_power_domain_st(PD_VI) == pmu_pd_on) { |
| SAVE_QOS(ddr_data.isp_128m_qos, ISP_128M); |
| SAVE_QOS(ddr_data.isp_rd_qos, ISP_RD); |
| SAVE_QOS(ddr_data.isp_wr_qos, ISP_WR); |
| SAVE_QOS(ddr_data.isp_m1_qos, ISP_M1); |
| SAVE_QOS(ddr_data.vip_qos, VIP); |
| } |
| if (pmu_power_domain_st(PD_VO) == pmu_pd_on) { |
| SAVE_QOS(ddr_data.rga_rd_qos, RGA_RD); |
| SAVE_QOS(ddr_data.rga_wr_qos, RGA_WR); |
| SAVE_QOS(ddr_data.vop_m0_qos, VOP_M0); |
| SAVE_QOS(ddr_data.vop_m1_qos, VOP_M1); |
| } |
| if (pmu_power_domain_st(PD_VPU) == pmu_pd_on) { |
| SAVE_QOS(ddr_data.vpu_qos, VPU); |
| SAVE_QOS(ddr_data.vpu_r128_qos, VPU_R128); |
| } |
| if (pmu_power_domain_st(PD_MMC_NAND) == pmu_pd_on) { |
| SAVE_QOS(ddr_data.emmc_qos, EMMC); |
| SAVE_QOS(ddr_data.nand_qos, NAND); |
| SAVE_QOS(ddr_data.sdio_qos, SDIO); |
| SAVE_QOS(ddr_data.sfc_qos, SFC); |
| } |
| if (pmu_power_domain_st(PD_GMAC) == pmu_pd_on) |
| SAVE_QOS(ddr_data.gmac_qos, GMAC); |
| if (pmu_power_domain_st(PD_CRYPTO) == pmu_pd_on) |
| SAVE_QOS(ddr_data.crypto_qos, CRYPTO); |
| if (pmu_power_domain_st(PD_SDCARD) == pmu_pd_on) |
| SAVE_QOS(ddr_data.sdmmc_qos, SDMMC); |
| if (pmu_power_domain_st(PD_USB) == pmu_pd_on) { |
| SAVE_QOS(ddr_data.usb_host_qos, USB_HOST); |
| SAVE_QOS(ddr_data.usb_otg_qos, USB_OTG); |
| } |
| } |
| |
| static void qos_restore(void) |
| { |
| RESTORE_QOS(ddr_data.cpu_qos, CPU); |
| |
| if (pmu_power_domain_st(PD_GPU) == pmu_pd_on) |
| RESTORE_QOS(ddr_data.gpu_qos, GPU); |
| if (pmu_power_domain_st(PD_VI) == pmu_pd_on) { |
| RESTORE_QOS(ddr_data.isp_128m_qos, ISP_128M); |
| RESTORE_QOS(ddr_data.isp_rd_qos, ISP_RD); |
| RESTORE_QOS(ddr_data.isp_wr_qos, ISP_WR); |
| RESTORE_QOS(ddr_data.isp_m1_qos, ISP_M1); |
| RESTORE_QOS(ddr_data.vip_qos, VIP); |
| } |
| if (pmu_power_domain_st(PD_VO) == pmu_pd_on) { |
| RESTORE_QOS(ddr_data.rga_rd_qos, RGA_RD); |
| RESTORE_QOS(ddr_data.rga_wr_qos, RGA_WR); |
| RESTORE_QOS(ddr_data.vop_m0_qos, VOP_M0); |
| RESTORE_QOS(ddr_data.vop_m1_qos, VOP_M1); |
| } |
| if (pmu_power_domain_st(PD_VPU) == pmu_pd_on) { |
| RESTORE_QOS(ddr_data.vpu_qos, VPU); |
| RESTORE_QOS(ddr_data.vpu_r128_qos, VPU_R128); |
| } |
| if (pmu_power_domain_st(PD_MMC_NAND) == pmu_pd_on) { |
| RESTORE_QOS(ddr_data.emmc_qos, EMMC); |
| RESTORE_QOS(ddr_data.nand_qos, NAND); |
| RESTORE_QOS(ddr_data.sdio_qos, SDIO); |
| RESTORE_QOS(ddr_data.sfc_qos, SFC); |
| } |
| if (pmu_power_domain_st(PD_GMAC) == pmu_pd_on) |
| RESTORE_QOS(ddr_data.gmac_qos, GMAC); |
| if (pmu_power_domain_st(PD_CRYPTO) == pmu_pd_on) |
| RESTORE_QOS(ddr_data.crypto_qos, CRYPTO); |
| if (pmu_power_domain_st(PD_SDCARD) == pmu_pd_on) |
| RESTORE_QOS(ddr_data.sdmmc_qos, SDMMC); |
| if (pmu_power_domain_st(PD_USB) == pmu_pd_on) { |
| RESTORE_QOS(ddr_data.usb_host_qos, USB_HOST); |
| RESTORE_QOS(ddr_data.usb_otg_qos, USB_OTG); |
| } |
| } |
| |
| static int pmu_set_power_domain(uint32_t pd_id, uint32_t pd_state) |
| { |
| uint32_t state; |
| |
| if (pmu_power_domain_st(pd_id) == pd_state) |
| goto out; |
| |
| if (pd_state == pmu_pd_on) |
| pmu_power_domain_ctr(pd_id, pd_state); |
| |
| state = (pd_state == pmu_pd_off) ? bus_idle : bus_active; |
| |
| switch (pd_id) { |
| case PD_GPU: |
| pmu_bus_idle_req(BUS_ID_GPU, state); |
| break; |
| case PD_VI: |
| pmu_bus_idle_req(BUS_ID_VI, state); |
| break; |
| case PD_VO: |
| pmu_bus_idle_req(BUS_ID_VO, state); |
| break; |
| case PD_VPU: |
| pmu_bus_idle_req(BUS_ID_VPU, state); |
| break; |
| case PD_MMC_NAND: |
| pmu_bus_idle_req(BUS_ID_MMC, state); |
| break; |
| case PD_GMAC: |
| pmu_bus_idle_req(BUS_ID_GMAC, state); |
| break; |
| case PD_CRYPTO: |
| pmu_bus_idle_req(BUS_ID_CRYPTO, state); |
| break; |
| case PD_SDCARD: |
| pmu_bus_idle_req(BUS_ID_SDCARD, state); |
| break; |
| case PD_USB: |
| pmu_bus_idle_req(BUS_ID_USB, state); |
| break; |
| default: |
| break; |
| } |
| |
| if (pd_state == pmu_pd_off) |
| pmu_power_domain_ctr(pd_id, pd_state); |
| |
| out: |
| return 0; |
| } |
| |
| static uint32_t pmu_powerdomain_state; |
| |
| static void pmu_power_domains_suspend(void) |
| { |
| uint32_t clkgt_save[CRU_CLKGATES_CON_CNT + CRU_PMU_CLKGATE_CON_CNT]; |
| |
| clk_gate_con_save(clkgt_save); |
| clk_gate_con_disable(); |
| qos_save(); |
| |
| pmu_powerdomain_state = mmio_read_32(PMU_BASE + PMU_PWRDN_ST); |
| pmu_set_power_domain(PD_GPU, pmu_pd_off); |
| pmu_set_power_domain(PD_VI, pmu_pd_off); |
| pmu_set_power_domain(PD_VO, pmu_pd_off); |
| pmu_set_power_domain(PD_VPU, pmu_pd_off); |
| pmu_set_power_domain(PD_MMC_NAND, pmu_pd_off); |
| pmu_set_power_domain(PD_GMAC, pmu_pd_off); |
| pmu_set_power_domain(PD_CRYPTO, pmu_pd_off); |
| pmu_set_power_domain(PD_SDCARD, pmu_pd_off); |
| pmu_set_power_domain(PD_USB, pmu_pd_off); |
| |
| clk_gate_con_restore(clkgt_save); |
| } |
| |
| static void pmu_power_domains_resume(void) |
| { |
| uint32_t clkgt_save[CRU_CLKGATES_CON_CNT + CRU_PMU_CLKGATE_CON_CNT]; |
| |
| clk_gate_con_save(clkgt_save); |
| clk_gate_con_disable(); |
| |
| if (!(pmu_powerdomain_state & BIT(PD_USB))) |
| pmu_set_power_domain(PD_USB, pmu_pd_on); |
| if (!(pmu_powerdomain_state & BIT(PD_SDCARD))) |
| pmu_set_power_domain(PD_SDCARD, pmu_pd_on); |
| if (!(pmu_powerdomain_state & BIT(PD_CRYPTO))) |
| pmu_set_power_domain(PD_CRYPTO, pmu_pd_on); |
| if (!(pmu_powerdomain_state & BIT(PD_GMAC))) |
| pmu_set_power_domain(PD_GMAC, pmu_pd_on); |
| if (!(pmu_powerdomain_state & BIT(PD_MMC_NAND))) |
| pmu_set_power_domain(PD_MMC_NAND, pmu_pd_on); |
| if (!(pmu_powerdomain_state & BIT(PD_VPU))) |
| pmu_set_power_domain(PD_VPU, pmu_pd_on); |
| if (!(pmu_powerdomain_state & BIT(PD_VO))) |
| pmu_set_power_domain(PD_VO, pmu_pd_on); |
| if (!(pmu_powerdomain_state & BIT(PD_VI))) |
| pmu_set_power_domain(PD_VI, pmu_pd_on); |
| if (!(pmu_powerdomain_state & BIT(PD_GPU))) |
| pmu_set_power_domain(PD_GPU, pmu_pd_on); |
| |
| qos_restore(); |
| clk_gate_con_restore(clkgt_save); |
| } |
| |
| static int check_cpu_wfie(uint32_t cpu) |
| { |
| uint32_t loop = 0, wfie_msk = CKECK_WFEI_MSK << cpu; |
| |
| while (!(mmio_read_32(GRF_BASE + GRF_CPU_STATUS1) & wfie_msk) && |
| (loop < WFEI_CHECK_LOOP)) { |
| udelay(1); |
| loop++; |
| } |
| |
| if ((mmio_read_32(GRF_BASE + GRF_CPU_STATUS1) & wfie_msk) == 0) { |
| WARN("%s: %d, %d, error!\n", __func__, cpu, wfie_msk); |
| return -EINVAL; |
| } |
| |
| return 0; |
| } |
| |
| static int cpus_power_domain_on(uint32_t cpu_id) |
| { |
| uint32_t cpu_pd, apm_value, cfg_info, loop = 0; |
| |
| cpu_pd = PD_CPU0 + cpu_id; |
| cfg_info = get_cpus_pwr_domain_cfg_info(cpu_id); |
| |
| if (cfg_info == core_pwr_pd) { |
| /* disable apm cfg */ |
| mmio_write_32(PMU_BASE + PMU_CPUAPM_CON(cpu_id), |
| WITH_16BITS_WMSK(CORES_PM_DISABLE)); |
| if (pmu_power_domain_st(cpu_pd) == pmu_pd_on) { |
| mmio_write_32(PMU_BASE + PMU_CPUAPM_CON(cpu_id), |
| WITH_16BITS_WMSK(CORES_PM_DISABLE)); |
| pmu_power_domain_ctr(cpu_pd, pmu_pd_off); |
| } |
| pmu_power_domain_ctr(cpu_pd, pmu_pd_on); |
| } else { |
| /* wait cpu down */ |
| while (pmu_power_domain_st(cpu_pd) == pmu_pd_on && loop < 100) { |
| udelay(2); |
| loop++; |
| } |
| |
| /* return error if can't wait cpu down */ |
| if (pmu_power_domain_st(cpu_pd) == pmu_pd_on) { |
| WARN("%s:can't wait cpu down\n", __func__); |
| return -EINVAL; |
| } |
| |
| /* power up cpu in power down state */ |
| apm_value = BIT(core_pm_sft_wakeup_en); |
| mmio_write_32(PMU_BASE + PMU_CPUAPM_CON(cpu_id), |
| WITH_16BITS_WMSK(apm_value)); |
| } |
| |
| return 0; |
| } |
| |
| static int cpus_power_domain_off(uint32_t cpu_id, uint32_t pd_cfg) |
| { |
| uint32_t cpu_pd, apm_value; |
| |
| cpu_pd = PD_CPU0 + cpu_id; |
| if (pmu_power_domain_st(cpu_pd) == pmu_pd_off) |
| return 0; |
| |
| if (pd_cfg == core_pwr_pd) { |
| if (check_cpu_wfie(cpu_id)) |
| return -EINVAL; |
| /* disable apm cfg */ |
| mmio_write_32(PMU_BASE + PMU_CPUAPM_CON(cpu_id), |
| WITH_16BITS_WMSK(CORES_PM_DISABLE)); |
| set_cpus_pwr_domain_cfg_info(cpu_id, pd_cfg); |
| pmu_power_domain_ctr(cpu_pd, pmu_pd_off); |
| } else { |
| set_cpus_pwr_domain_cfg_info(cpu_id, pd_cfg); |
| apm_value = BIT(core_pm_en) | BIT(core_pm_dis_int); |
| if (pd_cfg == core_pwr_wfi_int) |
| apm_value |= BIT(core_pm_int_wakeup_en); |
| mmio_write_32(PMU_BASE + PMU_CPUAPM_CON(cpu_id), |
| WITH_16BITS_WMSK(apm_value)); |
| } |
| |
| return 0; |
| } |
| |
| static void nonboot_cpus_off(void) |
| { |
| uint32_t boot_cpu, cpu; |
| |
| boot_cpu = plat_my_core_pos(); |
| |
| for (cpu = 0; cpu < PLATFORM_CORE_COUNT; cpu++) { |
| if (cpu == boot_cpu) |
| continue; |
| cpus_power_domain_off(cpu, core_pwr_pd); |
| } |
| } |
| |
| int rockchip_soc_cores_pwr_dm_on(unsigned long mpidr, |
| uint64_t entrypoint) |
| { |
| uint32_t cpu_id = plat_core_pos_by_mpidr(mpidr); |
| |
| assert(cpu_id < PLATFORM_CORE_COUNT); |
| assert(cpuson_flags[cpu_id] == 0); |
| cpuson_flags[cpu_id] = PMU_CPU_HOTPLUG; |
| cpuson_entry_point[cpu_id] = entrypoint; |
| dsb(); |
| |
| cpus_power_domain_on(cpu_id); |
| |
| return PSCI_E_SUCCESS; |
| } |
| |
| int rockchip_soc_cores_pwr_dm_on_finish(void) |
| { |
| uint32_t cpu_id = plat_my_core_pos(); |
| |
| mmio_write_32(PMU_BASE + PMU_CPUAPM_CON(cpu_id), |
| WITH_16BITS_WMSK(CORES_PM_DISABLE)); |
| return PSCI_E_SUCCESS; |
| } |
| |
| int rockchip_soc_cores_pwr_dm_off(void) |
| { |
| uint32_t cpu_id = plat_my_core_pos(); |
| |
| cpus_power_domain_off(cpu_id, core_pwr_wfi); |
| |
| return PSCI_E_SUCCESS; |
| } |
| |
| int rockchip_soc_cores_pwr_dm_suspend(void) |
| { |
| uint32_t cpu_id = plat_my_core_pos(); |
| |
| assert(cpu_id < PLATFORM_CORE_COUNT); |
| assert(cpuson_flags[cpu_id] == 0); |
| cpuson_flags[cpu_id] = PMU_CPU_AUTO_PWRDN; |
| cpuson_entry_point[cpu_id] = plat_get_sec_entrypoint(); |
| dsb(); |
| |
| cpus_power_domain_off(cpu_id, core_pwr_wfi_int); |
| |
| return PSCI_E_SUCCESS; |
| } |
| |
| int rockchip_soc_cores_pwr_dm_resume(void) |
| { |
| uint32_t cpu_id = plat_my_core_pos(); |
| |
| /* Disable core_pm */ |
| mmio_write_32(PMU_BASE + PMU_CPUAPM_CON(cpu_id), |
| WITH_16BITS_WMSK(CORES_PM_DISABLE)); |
| |
| return PSCI_E_SUCCESS; |
| } |
| |
| #define CLK_MSK_GATING(msk, con) \ |
| mmio_write_32(CRU_BASE + (con), ((msk) << 16) | 0xffff) |
| #define CLK_MSK_UNGATING(msk, con) \ |
| mmio_write_32(CRU_BASE + (con), ((~(msk)) << 16) | 0xffff) |
| |
| static uint32_t clk_ungt_msk[CRU_CLKGATES_CON_CNT] = { |
| 0xe0ff, 0xffff, 0x0000, 0x0000, |
| 0x0000, 0x0380, 0x0000, 0x0000, |
| 0x07c0, 0x0000, 0x0000, 0x000f, |
| 0x0061, 0x1f02, 0x0440, 0x1801, |
| 0x004b, 0x0000 |
| }; |
| |
| static uint32_t clk_pmu_ungt_msk[CRU_PMU_CLKGATE_CON_CNT] = { |
| 0xf1ff, 0x0310 |
| }; |
| |
| void clk_gate_suspend(void) |
| { |
| int i; |
| |
| for (i = 0; i < CRU_CLKGATES_CON_CNT; i++) { |
| ddr_data.cru_clk_gate[i] = |
| mmio_read_32(CRU_BASE + CRU_CLKGATES_CON(i)); |
| mmio_write_32(CRU_BASE + CRU_CLKGATES_CON(i), |
| WITH_16BITS_WMSK(~clk_ungt_msk[i])); |
| } |
| |
| for (i = 0; i < CRU_PMU_CLKGATE_CON_CNT; i++) { |
| ddr_data.cru_pmu_clk_gate[i] = |
| mmio_read_32(PMUCRU_BASE + CRU_PMU_CLKGATES_CON(i)); |
| mmio_write_32(PMUCRU_BASE + CRU_PMU_CLKGATES_CON(i), |
| WITH_16BITS_WMSK(~clk_pmu_ungt_msk[i])); |
| } |
| } |
| |
| void clk_gate_resume(void) |
| { |
| int i; |
| |
| for (i = 0; i < CRU_PMU_CLKGATE_CON_CNT; i++) |
| mmio_write_32(PMUCRU_BASE + CRU_PMU_CLKGATES_CON(i), |
| WITH_16BITS_WMSK(ddr_data.cru_pmu_clk_gate[i])); |
| |
| for (i = 0; i < CRU_CLKGATES_CON_CNT; i++) |
| mmio_write_32(CRU_BASE + CRU_CLKGATES_CON(i), |
| WITH_16BITS_WMSK(ddr_data.cru_clk_gate[i])); |
| } |
| |
| static void pvtm_32k_config(void) |
| { |
| uint32_t pvtm_freq_khz, pvtm_div; |
| |
| ddr_data.pmu_cru_clksel_con0 = |
| mmio_read_32(PMUCRU_BASE + CRU_PMU_CLKSELS_CON(0)); |
| |
| ddr_data.pgrf_pvtm_con[0] = |
| mmio_read_32(PMUGRF_BASE + PMUGRF_PVTM_CON0); |
| ddr_data.pgrf_pvtm_con[1] = |
| mmio_read_32(PMUGRF_BASE + PMUGRF_PVTM_CON1); |
| |
| mmio_write_32(PMUGRF_BASE + PMUGRF_PVTM_CON0, |
| BITS_WITH_WMASK(0, 0x3, pgrf_pvtm_st)); |
| dsb(); |
| mmio_write_32(PMUGRF_BASE + PMUGRF_PVTM_CON0, |
| BITS_WITH_WMASK(1, 0x1, pgrf_pvtm_en)); |
| dsb(); |
| mmio_write_32(PMUGRF_BASE + PMUGRF_PVTM_CON1, PVTM_CALC_CNT); |
| dsb(); |
| |
| mmio_write_32(PMUGRF_BASE + PMUGRF_PVTM_CON0, |
| BITS_WITH_WMASK(1, 0x1, pgrf_pvtm_st)); |
| |
| /* pmugrf_pvtm_st0 will be clear after PVTM start, |
| * which will cost about 6 cycles of pvtm at least. |
| * So we wait 30 cycles of pvtm for security. |
| */ |
| while (mmio_read_32(PMUGRF_BASE + PMUGRF_PVTM_ST1) < 30) |
| ; |
| |
| dsb(); |
| while (!(mmio_read_32(PMUGRF_BASE + PMUGRF_PVTM_ST0) & 0x1)) |
| ; |
| |
| pvtm_freq_khz = |
| (mmio_read_32(PMUGRF_BASE + PMUGRF_PVTM_ST1) * 24000 + |
| PVTM_CALC_CNT / 2) / PVTM_CALC_CNT; |
| pvtm_div = (pvtm_freq_khz + 16) / 32; |
| |
| /* pvtm_div = div_factor << 2 + 1, |
| * so div_factor = (pvtm_div - 1) >> 2. |
| * But the operation ">> 2" will clear the low bit of pvtm_div, |
| * so we don't have to do "- 1" for compasation |
| */ |
| pvtm_div = pvtm_div >> 2; |
| if (pvtm_div > 0x3f) |
| pvtm_div = 0x3f; |
| |
| mmio_write_32(PMUGRF_BASE + PMUGRF_PVTM_CON0, |
| BITS_WITH_WMASK(pvtm_div, 0x3f, pgrf_pvtm_div)); |
| |
| /* select pvtm as 32k source */ |
| mmio_write_32(PMUCRU_BASE + CRU_PMU_CLKSELS_CON(0), |
| BITS_WITH_WMASK(1, 0x3U, 14)); |
| } |
| |
| static void pvtm_32k_config_restore(void) |
| { |
| mmio_write_32(PMUCRU_BASE + CRU_PMU_CLKSELS_CON(0), |
| ddr_data.pmu_cru_clksel_con0 | BITS_WMSK(0x3U, 14)); |
| |
| mmio_write_32(PMUGRF_BASE + PMUGRF_PVTM_CON0, |
| WITH_16BITS_WMSK(ddr_data.pgrf_pvtm_con[0])); |
| mmio_write_32(PMUGRF_BASE + PMUGRF_PVTM_CON1, |
| ddr_data.pgrf_pvtm_con[1]); |
| } |
| |
| static void ddr_sleep_config(void) |
| { |
| /* disable ddr pd, sr */ |
| ddr_data.ddrc_pwrctrl = mmio_read_32(DDR_UPCTL_BASE + 0x30); |
| mmio_write_32(DDR_UPCTL_BASE + 0x30, BITS_WITH_WMASK(0x0, 0x3, 0)); |
| |
| /* disable ddr auto gt */ |
| ddr_data.ddrgrf_con1 = mmio_read_32(DDRGRF_BASE + 0x4); |
| mmio_write_32(DDRGRF_BASE + 0x4, BITS_WITH_WMASK(0x0, 0x1f, 0)); |
| |
| /* disable ddr standby */ |
| ddr_data.ddrstdby_con0 = mmio_read_32(DDR_STDBY_BASE + 0x0); |
| mmio_write_32(DDR_STDBY_BASE + 0x0, BITS_WITH_WMASK(0x0, 0x1, 0)); |
| while ((mmio_read_32(DDR_UPCTL_BASE + 0x4) & 0x7) != 1) |
| ; |
| |
| /* ddr pmu ctrl */ |
| ddr_data.ddrgrf_con0 = mmio_read_32(DDRGRF_BASE + 0x0); |
| mmio_write_32(DDRGRF_BASE + 0x0, BITS_WITH_WMASK(0x0, 0x1, 5)); |
| dsb(); |
| mmio_write_32(DDRGRF_BASE + 0x0, BITS_WITH_WMASK(0x1, 0x1, 4)); |
| |
| /* ddr ret sel */ |
| ddr_data.pmugrf_soc_con0 = |
| mmio_read_32(PMUGRF_BASE + PMUGRF_SOC_CON(0)); |
| mmio_write_32(PMUGRF_BASE + PMUGRF_SOC_CON(0), |
| BITS_WITH_WMASK(0x0, 0x1, 12)); |
| } |
| |
| static void ddr_sleep_config_restore(void) |
| { |
| /* restore ddr ret sel */ |
| mmio_write_32(PMUGRF_BASE + PMUGRF_SOC_CON(0), |
| ddr_data.pmugrf_soc_con0 | BITS_WMSK(0x1, 12)); |
| |
| /* restore ddr pmu ctrl */ |
| mmio_write_32(DDRGRF_BASE + 0x0, |
| ddr_data.ddrgrf_con0 | BITS_WMSK(0x1, 4)); |
| dsb(); |
| mmio_write_32(DDRGRF_BASE + 0x0, |
| ddr_data.ddrgrf_con0 | BITS_WMSK(0x1, 5)); |
| |
| /* restore ddr standby */ |
| mmio_write_32(DDR_STDBY_BASE + 0x0, |
| ddr_data.ddrstdby_con0 | BITS_WMSK(0x1, 0)); |
| |
| /* restore ddr auto gt */ |
| mmio_write_32(DDRGRF_BASE + 0x4, |
| ddr_data.ddrgrf_con1 | BITS_WMSK(0x1f, 0)); |
| |
| /* restore ddr pd, sr */ |
| mmio_write_32(DDR_UPCTL_BASE + 0x30, |
| ddr_data.ddrc_pwrctrl | BITS_WMSK(0x3, 0)); |
| } |
| |
| static void pmu_sleep_config(void) |
| { |
| uint32_t pwrmd_core_lo, pwrmd_core_hi, pwrmd_com_lo, pwrmd_com_hi; |
| uint32_t pmu_wkup_cfg2_lo; |
| uint32_t clk_freq_khz; |
| |
| /* save pmic_sleep iomux gpio0_a4 */ |
| ddr_data.pmic_slp_iomux = mmio_read_32(PMUGRF_BASE + GPIO0A_IOMUX); |
| |
| ddr_data.pmu_pwrmd_core_l = |
| mmio_read_32(PMU_BASE + PMU_PWRMODE_CORE_LO); |
| ddr_data.pmu_pwrmd_core_h = |
| mmio_read_32(PMU_BASE + PMU_PWRMODE_CORE_HI); |
| ddr_data.pmu_pwrmd_cmm_l = |
| mmio_read_32(PMU_BASE + PMU_PWRMODE_COMMON_CON_LO); |
| ddr_data.pmu_pwrmd_cmm_h = |
| mmio_read_32(PMU_BASE + PMU_PWRMODE_COMMON_CON_HI); |
| ddr_data.pmu_wkup_cfg2_l = mmio_read_32(PMU_BASE + PMU_WKUP_CFG2_LO); |
| |
| pwrmd_core_lo = BIT(pmu_global_int_dis) | |
| BIT(pmu_core_src_gt) | |
| BIT(pmu_cpu0_pd) | |
| BIT(pmu_clr_core) | |
| BIT(pmu_scu_pd) | |
| BIT(pmu_l2_idle) | |
| BIT(pmu_l2_flush) | |
| BIT(pmu_clr_bus2main) | |
| BIT(pmu_clr_peri2msch); |
| |
| pwrmd_core_hi = BIT(pmu_dpll_pd_en) | |
| BIT(pmu_apll_pd_en) | |
| BIT(pmu_cpll_pd_en) | |
| BIT(pmu_gpll_pd_en) | |
| BIT(pmu_npll_pd_en); |
| |
| pwrmd_com_lo = BIT(pmu_mode_en) | |
| BIT(pmu_pll_pd) | |
| BIT(pmu_pmu_use_if) | |
| BIT(pmu_alive_use_if) | |
| BIT(pmu_osc_dis) | |
| BIT(pmu_sref_enter) | |
| BIT(pmu_ddrc_gt) | |
| BIT(pmu_clr_pmu) | |
| BIT(pmu_clr_peri_pmu); |
| |
| pwrmd_com_hi = BIT(pmu_clr_bus) | |
| BIT(pmu_clr_msch) | |
| BIT(pmu_wakeup_begin_cfg); |
| |
| pmu_wkup_cfg2_lo = BIT(pmu_cluster_wkup_en) | |
| BIT(pmu_gpio_wkup_en) | |
| BIT(pmu_timer_wkup_en); |
| |
| /* set pmic_sleep iomux gpio0_a4 */ |
| mmio_write_32(PMUGRF_BASE + GPIO0A_IOMUX, |
| BITS_WITH_WMASK(1, 0x3, 8)); |
| |
| clk_freq_khz = 32; |
| |
| mmio_write_32(PMU_BASE + PMU_OSC_CNT_LO, |
| WITH_16BITS_WMSK(clk_freq_khz * 32 & 0xffff)); |
| mmio_write_32(PMU_BASE + PMU_OSC_CNT_HI, |
| WITH_16BITS_WMSK(clk_freq_khz * 32 >> 16)); |
| |
| mmio_write_32(PMU_BASE + PMU_STABLE_CNT_LO, |
| WITH_16BITS_WMSK(clk_freq_khz * 32 & 0xffff)); |
| mmio_write_32(PMU_BASE + PMU_STABLE_CNT_HI, |
| WITH_16BITS_WMSK(clk_freq_khz * 32 >> 16)); |
| |
| mmio_write_32(PMU_BASE + PMU_WAKEUP_RST_CLR_LO, |
| WITH_16BITS_WMSK(clk_freq_khz * 2 & 0xffff)); |
| mmio_write_32(PMU_BASE + PMU_WAKEUP_RST_CLR_HI, |
| WITH_16BITS_WMSK(clk_freq_khz * 2 >> 16)); |
| |
| /* Pmu's clk has switched to 24M back When pmu FSM counts |
| * the follow counters, so we should use 24M to calculate |
| * these counters. |
| */ |
| mmio_write_32(PMU_BASE + PMU_SCU_PWRDN_CNT_LO, |
| WITH_16BITS_WMSK(24000 * 2 & 0xffff)); |
| mmio_write_32(PMU_BASE + PMU_SCU_PWRDN_CNT_HI, |
| WITH_16BITS_WMSK(24000 * 2 >> 16)); |
| |
| mmio_write_32(PMU_BASE + PMU_SCU_PWRUP_CNT_LO, |
| WITH_16BITS_WMSK(24000 * 2 & 0xffff)); |
| mmio_write_32(PMU_BASE + PMU_SCU_PWRUP_CNT_HI, |
| WITH_16BITS_WMSK(24000 * 2 >> 16)); |
| |
| mmio_write_32(PMU_BASE + PMU_PLLLOCK_CNT_LO, |
| WITH_16BITS_WMSK(24000 * 5 & 0xffff)); |
| mmio_write_32(PMU_BASE + PMU_PLLLOCK_CNT_HI, |
| WITH_16BITS_WMSK(24000 * 5 >> 16)); |
| |
| mmio_write_32(PMU_BASE + PMU_PLLRST_CNT_LO, |
| WITH_16BITS_WMSK(24000 * 2 & 0xffff)); |
| mmio_write_32(PMU_BASE + PMU_PLLRST_CNT_HI, |
| WITH_16BITS_WMSK(24000 * 2 >> 16)); |
| |
| /* Config pmu power mode and pmu wakeup source */ |
| mmio_write_32(PMU_BASE + PMU_PWRMODE_CORE_LO, |
| WITH_16BITS_WMSK(pwrmd_core_lo)); |
| mmio_write_32(PMU_BASE + PMU_PWRMODE_CORE_HI, |
| WITH_16BITS_WMSK(pwrmd_core_hi)); |
| |
| mmio_write_32(PMU_BASE + PMU_PWRMODE_COMMON_CON_LO, |
| WITH_16BITS_WMSK(pwrmd_com_lo)); |
| mmio_write_32(PMU_BASE + PMU_PWRMODE_COMMON_CON_HI, |
| WITH_16BITS_WMSK(pwrmd_com_hi)); |
| |
| mmio_write_32(PMU_BASE + PMU_WKUP_CFG2_LO, |
| WITH_16BITS_WMSK(pmu_wkup_cfg2_lo)); |
| } |
| |
| static void pmu_sleep_restore(void) |
| { |
| mmio_write_32(PMU_BASE + PMU_PWRMODE_CORE_LO, |
| WITH_16BITS_WMSK(ddr_data.pmu_pwrmd_core_l)); |
| mmio_write_32(PMU_BASE + PMU_PWRMODE_CORE_HI, |
| WITH_16BITS_WMSK(ddr_data.pmu_pwrmd_core_h)); |
| mmio_write_32(PMU_BASE + PMU_PWRMODE_COMMON_CON_LO, |
| WITH_16BITS_WMSK(ddr_data.pmu_pwrmd_cmm_l)); |
| mmio_write_32(PMU_BASE + PMU_PWRMODE_COMMON_CON_HI, |
| WITH_16BITS_WMSK(ddr_data.pmu_pwrmd_cmm_h)); |
| mmio_write_32(PMU_BASE + PMU_WKUP_CFG2_LO, |
| WITH_16BITS_WMSK(ddr_data.pmu_wkup_cfg2_l)); |
| |
| /* restore pmic_sleep iomux */ |
| mmio_write_32(PMUGRF_BASE + GPIO0A_IOMUX, |
| WITH_16BITS_WMSK(ddr_data.pmic_slp_iomux)); |
| } |
| |
| static void soc_sleep_config(void) |
| { |
| ddr_data.gpio0c_iomux = mmio_read_32(PMUGRF_BASE + GPIO0C_IOMUX); |
| |
| pmu_sleep_config(); |
| |
| ddr_sleep_config(); |
| |
| pvtm_32k_config(); |
| } |
| |
| static void soc_sleep_restore(void) |
| { |
| secure_timer_init(); |
| |
| pvtm_32k_config_restore(); |
| |
| ddr_sleep_config_restore(); |
| |
| pmu_sleep_restore(); |
| |
| mmio_write_32(PMUGRF_BASE + GPIO0C_IOMUX, |
| WITH_16BITS_WMSK(ddr_data.gpio0c_iomux)); |
| } |
| |
| static inline void pm_pll_wait_lock(uint32_t pll_base, uint32_t pll_id) |
| { |
| uint32_t delay = PLL_LOCKED_TIMEOUT; |
| |
| while (delay > 0) { |
| if (mmio_read_32(pll_base + PLL_CON(1)) & |
| PLL_LOCK_MSK) |
| break; |
| delay--; |
| } |
| |
| if (delay == 0) |
| ERROR("Can't wait pll:%d lock\n", pll_id); |
| } |
| |
| static inline void pll_pwr_ctr(uint32_t pll_base, uint32_t pll_id, uint32_t pd) |
| { |
| mmio_write_32(pll_base + PLL_CON(1), |
| BITS_WITH_WMASK(1, 1U, 15)); |
| if (pd) |
| mmio_write_32(pll_base + PLL_CON(1), |
| BITS_WITH_WMASK(1, 1, 14)); |
| else |
| mmio_write_32(pll_base + PLL_CON(1), |
| BITS_WITH_WMASK(0, 1, 14)); |
| } |
| |
| static inline void pll_set_mode(uint32_t pll_id, uint32_t mode) |
| { |
| uint32_t val = BITS_WITH_WMASK(mode, 0x3, PLL_MODE_SHIFT(pll_id)); |
| |
| if (pll_id != GPLL_ID) |
| mmio_write_32(CRU_BASE + CRU_MODE, val); |
| else |
| mmio_write_32(PMUCRU_BASE + CRU_PMU_MODE, |
| BITS_WITH_WMASK(mode, 0x3, 0)); |
| } |
| |
| static inline void pll_suspend(uint32_t pll_id) |
| { |
| int i; |
| uint32_t pll_base; |
| |
| if (pll_id != GPLL_ID) |
| pll_base = CRU_BASE + CRU_PLL_CONS(pll_id, 0); |
| else |
| pll_base = PMUCRU_BASE + CRU_PLL_CONS(0, 0); |
| |
| /* save pll con */ |
| for (i = 0; i < PLL_CON_CNT; i++) |
| ddr_data.cru_plls_con_save[pll_id][i] = |
| mmio_read_32(pll_base + PLL_CON(i)); |
| |
| /* slow mode */ |
| pll_set_mode(pll_id, SLOW_MODE); |
| } |
| |
| static inline void pll_resume(uint32_t pll_id) |
| { |
| uint32_t mode, pll_base; |
| |
| if (pll_id != GPLL_ID) { |
| pll_base = CRU_BASE + CRU_PLL_CONS(pll_id, 0); |
| mode = (ddr_data.cru_mode_save >> PLL_MODE_SHIFT(pll_id)) & 0x3; |
| } else { |
| pll_base = PMUCRU_BASE + CRU_PLL_CONS(0, 0); |
| mode = ddr_data.cru_pmu_mode_save & 0x3; |
| } |
| |
| /* if pll locked before suspend, we should wait atfer resume */ |
| if (ddr_data.cru_plls_con_save[pll_id][1] & PLL_LOCK_MSK) |
| pm_pll_wait_lock(pll_base, pll_id); |
| |
| pll_set_mode(pll_id, mode); |
| } |
| |
| static void pm_plls_suspend(void) |
| { |
| ddr_data.cru_mode_save = mmio_read_32(CRU_BASE + CRU_MODE); |
| ddr_data.cru_pmu_mode_save = mmio_read_32(PMUCRU_BASE + CRU_PMU_MODE); |
| ddr_data.clk_sel0 = mmio_read_32(CRU_BASE + CRU_CLKSELS_CON(0)); |
| |
| pll_suspend(GPLL_ID); |
| pll_suspend(NPLL_ID); |
| pll_suspend(CPLL_ID); |
| pll_suspend(APLL_ID); |
| |
| /* core */ |
| mmio_write_32(CRU_BASE + CRU_CLKSELS_CON(0), |
| BITS_WITH_WMASK(0, 0xf, 0)); |
| |
| /* pclk_dbg */ |
| mmio_write_32(CRU_BASE + CRU_CLKSELS_CON(0), |
| BITS_WITH_WMASK(0, 0xf, 8)); |
| } |
| |
| static void pm_plls_resume(void) |
| { |
| /* pclk_dbg */ |
| mmio_write_32(CRU_BASE + CRU_CLKSELS_CON(0), |
| ddr_data.clk_sel0 | BITS_WMSK(0xf, 8)); |
| |
| /* core */ |
| mmio_write_32(CRU_BASE + CRU_CLKSELS_CON(0), |
| ddr_data.clk_sel0 | BITS_WMSK(0xf, 0)); |
| |
| pll_resume(APLL_ID); |
| pll_resume(CPLL_ID); |
| pll_resume(NPLL_ID); |
| pll_resume(GPLL_ID); |
| } |
| |
| int rockchip_soc_sys_pwr_dm_suspend(void) |
| { |
| pmu_power_domains_suspend(); |
| |
| clk_gate_suspend(); |
| |
| soc_sleep_config(); |
| |
| pm_plls_suspend(); |
| |
| psram_boot_cfg->pm_flag &= ~PM_WARM_BOOT_BIT; |
| |
| return 0; |
| } |
| |
| int rockchip_soc_sys_pwr_dm_resume(void) |
| { |
| psram_boot_cfg->pm_flag |= PM_WARM_BOOT_BIT; |
| |
| pm_plls_resume(); |
| |
| soc_sleep_restore(); |
| |
| clk_gate_resume(); |
| |
| pmu_power_domains_resume(); |
| |
| plat_rockchip_gic_cpuif_enable(); |
| |
| return 0; |
| } |
| |
| void __dead2 rockchip_soc_soft_reset(void) |
| { |
| pll_set_mode(GPLL_ID, SLOW_MODE); |
| pll_set_mode(CPLL_ID, SLOW_MODE); |
| pll_set_mode(NPLL_ID, SLOW_MODE); |
| pll_set_mode(APLL_ID, SLOW_MODE); |
| dsb(); |
| |
| mmio_write_32(CRU_BASE + CRU_GLB_SRST_FST, CRU_GLB_SRST_FST_VALUE); |
| dsb(); |
| |
| /* |
| * Maybe the HW needs some times to reset the system, |
| * so we do not hope the core to execute valid codes. |
| */ |
| psci_power_down_wfi(); |
| } |
| |
| void __dead2 rockchip_soc_system_off(void) |
| { |
| uint32_t val; |
| |
| /* set pmic_sleep pin(gpio0_a4) to gpio mode */ |
| mmio_write_32(PMUGRF_BASE + GPIO0A_IOMUX, BITS_WITH_WMASK(0, 0x3, 8)); |
| |
| /* config output */ |
| val = mmio_read_32(GPIO0_BASE + SWPORTA_DDR); |
| val |= BIT(4); |
| mmio_write_32(GPIO0_BASE + SWPORTA_DDR, val); |
| |
| /* config output high level */ |
| val = mmio_read_32(GPIO0_BASE); |
| val |= BIT(4); |
| mmio_write_32(GPIO0_BASE, val); |
| dsb(); |
| |
| /* |
| * Maybe the HW needs some times to reset the system, |
| * so we do not hope the core to execute valid codes. |
| */ |
| psci_power_down_wfi(); |
| } |
| |
| void rockchip_plat_mmu_el3(void) |
| { |
| /* TODO: support the el3 for px30 SoCs */ |
| } |
| |
| void plat_rockchip_pmu_init(void) |
| { |
| uint32_t cpu; |
| |
| rockchip_pd_lock_init(); |
| |
| for (cpu = 0; cpu < PLATFORM_CORE_COUNT; cpu++) |
| cpuson_flags[cpu] = 0; |
| |
| psram_boot_cfg->ddr_func = (uint64_t)0; |
| psram_boot_cfg->ddr_data = (uint64_t)0; |
| psram_boot_cfg->sp = PSRAM_SP_TOP; |
| psram_boot_cfg->ddr_flag = 0x0; |
| psram_boot_cfg->boot_mpidr = read_mpidr_el1() & 0xffff; |
| psram_boot_cfg->pm_flag = PM_WARM_BOOT_BIT; |
| |
| nonboot_cpus_off(); |
| |
| /* Remap pmu_sram's base address to boot address */ |
| mmio_write_32(PMUSGRF_BASE + PMUSGRF_SOC_CON(0), |
| BITS_WITH_WMASK(1, 0x1, 13)); |
| |
| INFO("%s: pd status %x\n", |
| __func__, mmio_read_32(PMU_BASE + PMU_PWRDN_ST)); |
| } |