/* * Idle processing for ARMv7-based Qualcomm SoCs. * * Copyright (C) 2007 Google, Inc. * Copyright (c) 2007-2009, 2011-2013 The Linux Foundation. All rights reserved. * * This software is licensed under the terms of the GNU General Public * License version 2, as published by the Free Software Foundation, and * may be copied, distributed, and modified under those terms. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * */ #include #include #include #include "idle.h" #include "idle-macros.S" #ifdef CONFIG_MSM_SCM #define SCM_SVC_BOOT 0x1 #define SCM_CMD_TERMINATE_PC 0x2 #endif ENTRY(msm_arch_idle) #ifdef CONFIG_ARCH_MSM_KRAIT mrc p15, 0, r0, c0, c0, 0 bic r1, r0, #0xff movw r2, #0x0400 movt r2, #0x511F movw r3, #0x0600 movt r3, #0x510F cmp r2, r1 cmpne r3, r1 bne go_wfi mrs r0, cpsr cpsid if mrc p15, 7, r1, c15, c0, 5 bic r2, r1, #0x20000 mcr p15, 7, r2, c15, c0, 5 isb go_wfi: wfi bne wfi_done mcr p15, 7, r1, c15, c0, 5 isb msr cpsr_c, r0 wfi_done: bx lr #else wfi #ifdef CONFIG_ARCH_MSM8X60 mrc p14, 1, r1, c1, c5, 4 /* read ETM PDSR to clear sticky bit */ mrc p14, 0, r1, c1, c5, 4 /* read DBG PRSR to clear sticky bit */ isb #endif bx lr #endif ENTRY(msm_pm_collapse) #if defined(CONFIG_MSM_FIQ_SUPPORT) cpsid f #endif ldr r0, =msm_saved_state /* address of msm_saved_state ptr */ ldr r0, [r0] /* load ptr */ #if (NR_CPUS >= 2) mrc p15, 0, r1, c0, c0, 5 /* MPIDR */ ands r1, r1, #15 /* What CPU am I */ mov r2, #CPU_SAVED_STATE_SIZE mul r1, r1, r2 add r0, r0, r1 #endif stmia r0!, {r4-r14} mrc p15, 0, r1, c1, c0, 0 /* MMU control */ mrc p15, 0, r2, c2, c0, 0 /* TTBR0 */ mrc p15, 0, r3, c3, c0, 0 /* dacr */ #ifdef CONFIG_ARCH_MSM_SCORPION /* This instruction is not valid for non scorpion processors */ mrc p15, 3, r4, c15, c0, 3 /* L2CR1 is the L2 cache control reg 1 */ #endif mrc p15, 0, r5, c10, c2, 0 /* PRRR */ mrc p15, 0, r6, c10, c2, 1 /* NMRR */ mrc p15, 0, r7, c1, c0, 1 /* ACTLR */ mrc p15, 0, r8, c2, c0, 1 /* TTBR1 */ mrc p15, 0, r9, c13, c0, 3 /* TPIDRURO */ mrc p15, 0, ip, c13, c0, 1 /* context ID */ stmia r0!, {r1-r9, ip} #if defined(CONFIG_MSM_JTAG) || defined(CONFIG_MSM_JTAG_MM) bl msm_jtag_save_state #endif ldr r0, =msm_pm_flush_l2_flag ldr r0, [r0] mov r1, #0 mcr p15, 2, r1, c0, c0, 0 /*CCSELR*/ isb mrc p15, 1, r1, c0, c0, 0 /*CCSIDR*/ mov r2, #1 and r1, r2, r1, ASR #30 /* Check if the cache is write back */ orr r1, r0, r1 cmp r1, #1 bne skip bl v7_flush_dcache_all ldr r1, =msm_pm_flush_l2_fn ldr r1, [r1] cmp r1, #0 blxne r1 skip: ldr r1, =msm_pm_disable_l2_fn ldr r1, [r1] cmp r1, #0 blxne r1 dmb mrc p15, 0, r0, c0, c0, 5 /* MPIDR */ and r0, r0, #15 /* what CPU am I */ ldr r1, =msm_pc_debug_counters /*load the IMEM debug location */ ldr r1, [r1] cmp r1, #0 beq skip_pc_debug1 add r1, r1, r0, LSL #4 /* debug location for this CPU */ ldr r2, [r1] add r2, #1 str r2, [r1] skip_pc_debug1: #ifdef CONFIG_MSM_SCM ldr r0, =SCM_SVC_BOOT ldr r1, =SCM_CMD_TERMINATE_PC ldr r2, =msm_pm_flush_l2_flag ldr r2, [r2] bl scm_call_atomic1 #else mrc p15, 0, r4, c1, c0, 0 /* read current CR */ bic r0, r4, #(1 << 2) /* clear dcache bit */ bic r0, r0, #(1 << 12) /* clear icache bit */ mcr p15, 0, r0, c1, c0, 0 /* disable d/i cache */ isb SUSPEND_8x25_L2 SET_SMP_COHERENCY OFF wfi DELAY_8x25 300 mcr p15, 0, r4, c1, c0, 0 /* restore d/i cache */ isb ENABLE_8x25_L2 /* enable only l2, no need to restore the reg back */ SET_SMP_COHERENCY ON #endif #if defined(CONFIG_MSM_FIQ_SUPPORT) cpsie f #endif mrc p15, 0, r0, c0, c0, 5 /* MPIDR */ and r0, r0, #15 /* what CPU am I */ ldr r1, =msm_pc_debug_counters /*load the IMEM debug location */ ldr r1, [r1] cmp r1, #0 beq skip_pc_debug2 add r1, r1, r0, LSL #4 /* debug location for this CPU */ add r1, #8 ldr r2, [r1] add r2, #1 str r2, [r1] skip_pc_debug2: ldr r1, =msm_pm_enable_l2_fn ldr r1, [r1] cmp r1, #0 blxne r1 dmb #if defined(CONFIG_MSM_JTAG) || defined(CONFIG_MSM_JTAG_MM) bl msm_jtag_restore_state #endif ldr r0, =msm_saved_state /* address of msm_saved_state ptr */ ldr r0, [r0] /* load ptr */ #if (NR_CPUS >= 2) mrc p15, 0, r1, c0, c0, 5 /* MPIDR */ ands r1, r1, #15 /* What CPU am I */ mov r2, #CPU_SAVED_STATE_SIZE mul r2, r2, r1 add r0, r0, r2 #endif ldmfd r0, {r4-r14} /* restore registers */ mov r0, #0 /* return power collapse failed */ bx lr ENTRY(msm_pm_collapse_exit) #if 0 /* serial debug */ mov r0, #0x80000016 mcr p15, 0, r0, c15, c2, 4 mov r0, #0xA9000000 add r0, r0, #0x00A00000 /* UART1 */ /*add r0, r0, #0x00C00000*/ /* UART3 */ mov r1, #'A' str r1, [r0, #0x00C] #endif ldr r1, =msm_saved_state_phys ldr r2, =msm_pm_collapse_exit adr r3, msm_pm_collapse_exit add r1, r1, r3 sub r1, r1, r2 ldr r1, [r1] add r1, r1, #CPU_SAVED_STATE_SIZE #if (NR_CPUS >= 2) mrc p15, 0, r2, c0, c0, 5 /* MPIDR */ ands r2, r2, #15 /* What CPU am I */ mov r3, #CPU_SAVED_STATE_SIZE mul r2, r2, r3 add r1, r1, r2 #endif ldmdb r1!, {r2-r11} mcr p15, 0, r4, c3, c0, 0 /* dacr */ mcr p15, 0, r3, c2, c0, 0 /* TTBR0 */ #ifdef CONFIG_ARCH_MSM_SCORPION /* This instruction is not valid for non scorpion processors */ mcr p15, 3, r5, c15, c0, 3 /* L2CR1 */ #endif mcr p15, 0, r6, c10, c2, 0 /* PRRR */ mcr p15, 0, r7, c10, c2, 1 /* NMRR */ mcr p15, 0, r8, c1, c0, 1 /* ACTLR */ mcr p15, 0, r9, c2, c0, 1 /* TTBR1 */ mcr p15, 0, r10, c13, c0, 3 /* TPIDRURO */ mcr p15, 0, r11, c13, c0, 1 /* context ID */ isb ldmdb r1!, {r4-r14} ldr r0, =msm_pm_pc_pgd ldr r1, =msm_pm_collapse_exit adr r3, msm_pm_collapse_exit add r0, r0, r3 sub r0, r0, r1 ldr r0, [r0] mrc p15, 0, r1, c2, c0, 0 /* save current TTBR0 */ and r3, r1, #0x7f /* mask to get TTB flags */ orr r0, r0, r3 /* add TTB flags to switch TTBR value */ mcr p15, 0, r0, c2, c0, 0 /* temporary switch TTBR0 */ isb mcr p15, 0, r2, c1, c0, 0 /* MMU control */ isb msm_pm_mapped_pa: /* Switch to virtual */ ldr r0, =msm_pm_pa_to_va mov pc, r0 msm_pm_pa_to_va: mcr p15, 0, r1, c2, c0, 0 /* restore TTBR0 */ isb mcr p15, 0, r3, c8, c7, 0 /* UTLBIALL */ mcr p15, 0, r3, c7, c5, 6 /* BPIALL */ dsb isb #ifdef CONFIG_ARCH_MSM_KRAIT mrc p15, 0, r1, c0, c0, 0 ldr r3, =0xff00fc00 and r3, r1, r3 ldr r1, =0x51000400 cmp r3, r1 mrceq p15, 7, r3, c15, c0, 2 biceq r3, r3, #0x400 mcreq p15, 7, r3, c15, c0, 2 #else RESUME_8x25_L2 SET_SMP_COHERENCY ON #endif ldr r1, =msm_pm_enable_l2_fn ldr r1, [r1] cmp r1, #0 stmfd sp!, {lr} blxne r1 dmb #if defined(CONFIG_MSM_JTAG) || defined(CONFIG_MSM_JTAG_MM) bl msm_jtag_restore_state #endif ldmfd sp!, {lr} mov r0, #1 bx lr nop nop nop nop nop 1: b 1b ENTRY(msm_pm_boot_entry) mrc p15, 0, r0, c0, c0, 5 /* MPIDR */ and r0, r0, #15 /* what CPU am I */ ldr r1, =msm_pc_debug_counters_phys /*phys addr for IMEM reg */ ldr r2, =msm_pm_boot_entry adr r3, msm_pm_boot_entry add r1, r1, r3 /* translate virt to phys addr */ sub r1, r1, r2 ldr r1,[r1] cmp r1, #0 beq skip_pc_debug3 add r1, r1, r0, LSL #4 /* debug location for this CPU */ add r1, #4 /* warmboot entry counter*/ ldr r2, [r1] add r2, #1 str r2, [r1] skip_pc_debug3: ldr r1, =msm_pm_boot_vector ldr r2, =msm_pm_boot_entry adr r3, msm_pm_boot_entry add r1, r1, r3 /* translate virt to phys addr */ sub r1, r1, r2 add r1, r1, r0, LSL #2 /* locate boot vector for our cpu */ ldr pc, [r1] /* jump */ ENTRY(msm_pm_set_l2_flush_flag) ldr r1, =msm_pm_flush_l2_flag str r0, [r1] bx lr ENTRY(msm_pm_get_l2_flush_flag) ldr r1, =msm_pm_flush_l2_flag ldr r0, [r1] bx lr .data .globl msm_pm_pc_pgd msm_pm_pc_pgd: .long 0x0 .globl msm_saved_state msm_saved_state: .long 0x0 .globl msm_saved_state_phys msm_saved_state_phys: .long 0x0 .globl msm_pm_boot_vector msm_pm_boot_vector: .space 4 * NR_CPUS .globl target_type target_type: .long 0x0 .globl apps_power_collapse apps_power_collapse: .long 0x0 .globl l2x0_base_addr l2x0_base_addr: .long 0x0 .globl msm_pc_debug_counters_phys msm_pc_debug_counters_phys: .long 0x0 .globl msm_pc_debug_counters msm_pc_debug_counters: .long 0x0 .globl msm_pm_enable_l2_fn msm_pm_enable_l2_fn: .long 0x0 .globl msm_pm_disable_l2_fn msm_pm_disable_l2_fn: .long 0x0 .globl msm_pm_flush_l2_fn msm_pm_flush_l2_fn: .long 0x0 /* * Default the l2 flush flag to 1 so that caches are flushed during power * collapse unless the L2 driver decides to flush them only during L2 * Power collapse. */ msm_pm_flush_l2_flag: .long 0x1 /* * Save & restore l2x0 registers while system is entering and resuming * from Power Collapse. * 1. aux_ctrl_save (0x0) * 2. data_latency_ctrl (0x4) * 3. prefetch control (0x8) */ l2x0_saved_ctrl_reg_val: .space 4 * 3