| /* |
| * Idle processing for ARMv7-based Qualcomm SoCs. |
| * |
| * Copyright (C) 2007 Google, Inc. |
| * Copyright (c) 2007-2009, Code Aurora Forum. All rights reserved. |
| * |
| * This software is licensed under the terms of the GNU General Public |
| * License version 2, as published by the Free Software Foundation, and |
| * may be copied, distributed, and modified under those terms. |
| * |
| * This program is distributed in the hope that it will be useful, |
| * but WITHOUT ANY WARRANTY; without even the implied warranty of |
| * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
| * GNU General Public License for more details. |
| * |
| */ |
| |
| #include <linux/linkage.h> |
| #include <asm/assembler.h> |
| #include <asm/vfpmacros.h> |
| |
| ENTRY(msm_arch_idle) |
| wfi |
| bx lr |
| |
| ENTRY(msm_pm_collapse) |
| #if defined(CONFIG_MSM_FIQ_SUPPORT) |
| cpsid f |
| #endif |
| |
| ldr r0, =saved_state |
| stmia r0!, {r4-r14} |
| mrc p15, 0, r1, c1, c0, 0 /* MMU control */ |
| mrc p15, 0, r2, c2, c0, 0 /* TTBR0 */ |
| mrc p15, 0, r3, c3, c0, 0 /* dacr */ |
| mrc p15, 3, r4, c15, c0, 3 /* L2CR1 is the L2 cache control reg 1 */ |
| mrc p15, 0, r5, c10, c2, 0 /* PRRR */ |
| mrc p15, 0, r6, c10, c2, 1 /* NMRR */ |
| mrc p15, 0, r7, c1, c0, 1 /* ACTLR */ |
| mrc p15, 0, r8, c2, c0, 1 /* TTBR1 */ |
| mrc p15, 0, r9, c13, c0, 3 /* TPIDRURO */ |
| mrc p15, 0, ip, c13, c0, 1 /* context ID */ |
| stmia r0!, {r1-r9, ip} |
| |
| #ifdef CONFIG_VFP |
| VFPFSTMIA r0, r1 /* Save VFP working registers */ |
| fmrx r1, fpexc |
| fmrx r2, fpscr |
| stmia r0!, {r1, r2} /* Save VFP state registers */ |
| #endif |
| bl v7_flush_dcache_all |
| |
| mrc p15, 0, r1, c1, c0, 0 /* read current CR */ |
| bic r0, r1, #(1 << 2) /* clear dcache bit */ |
| bic r0, r0, #(1 << 12) /* clear icache bit */ |
| mcr p15, 0, r0, c1, c0, 0 /* disable d/i cache */ |
| |
| dsb |
| wfi |
| |
| mcr p15, 0, r1, c1, c0, 0 /* restore d/i cache */ |
| isb |
| |
| #if defined(CONFIG_MSM_FIQ_SUPPORT) |
| cpsie f |
| #endif |
| |
| ldr r0, =saved_state /* restore registers */ |
| ldmfd r0, {r4-r14} |
| mov r0, #0 /* return power collapse failed */ |
| bx lr |
| |
| ENTRY(msm_pm_collapse_exit) |
| #if 0 /* serial debug */ |
| mov r0, #0x80000016 |
| mcr p15, 0, r0, c15, c2, 4 |
| mov r0, #0xA9000000 |
| add r0, r0, #0x00A00000 /* UART1 */ |
| /*add r0, r0, #0x00C00000*/ /* UART3 */ |
| mov r1, #'A' |
| str r1, [r0, #0x00C] |
| #endif |
| |
| #if 0 |
| //; REMOVE FOLLOWING THREE INSTRUCTIONS WHEN POWER COLLAPSE IS ENA |
| //;Make sure the DBGOSLSR[LOCK] bit is cleared to allow access to the debug registers |
| //; Writing anything but the "secret code" to the DBGOSLAR clears the DBGOSLSR[LOCK] bit |
| MCR p14, 0, r0, c1, c0, 4 //; WCP14_DBGOSLAR r0 |
| |
| |
| //; Read the DBGPRSR to clear the DBGPRSR[STICKYPD] |
| //; Any read to DBGPRSR clear the STICKYPD bit |
| //; ISB guarantees the read completes before attempting to |
| //; execute a CP14 instruction. |
| MRC p14, 0, r3, c1, c5, 4 //; RCP14_DBGPRSR r3 |
| ISB |
| #endif |
| |
| #if 0 /* allow jtag reconnect */ |
| 1: |
| b 1b |
| #endif |
| |
| bl __cpu_early_init |
| |
| ldr r1, =saved_state_end |
| ldr r2, =msm_pm_collapse_exit |
| adr r3, msm_pm_collapse_exit |
| add r1, r1, r3 |
| sub r1, r1, r2 |
| #ifdef CONFIG_VFP |
| mrc p15, 0, r2, c1, c0, 2 /* Read CP Access Control Register */ |
| orr r2, r2, #0x00F00000 /* Enable full access for p10,11 */ |
| mcr p15, 0, r2, c1, c0, 2 /* Write CPACR */ |
| isb |
| mov r2, #0x40000000 /* Enable VFP */ |
| fmxr fpexc, r2 |
| isb |
| ldmdb r1!, {r2, r3} /* Read saved VFP state registers */ |
| sub r1, r1, #32*8 /* Jump to start of vfp regs area */ |
| VFPFLDMIA r1, r4 /* Restore VFP working registers, |
| * r1 incremented to end of vfp |
| * regs area */ |
| sub r1, r1, #32*8 /* Jump back to start of vfp regs area */ |
| fmxr fpscr, r3 /* Restore FPSCR */ |
| fmxr fpexc, r2 /* Restore FPEXC last */ |
| #endif |
| ldmdb r1!, {r2-r11} |
| mcr p15, 0, r4, c3, c0, 0 /* dacr */ |
| mcr p15, 0, r3, c2, c0, 0 /* TTBR0 */ |
| mcr p15, 3, r5, c15, c0, 3 /* L2CR1 */ |
| mcr p15, 0, r6, c10, c2, 0 /* PRRR */ |
| mcr p15, 0, r7, c10, c2, 1 /* NMRR */ |
| mcr p15, 0, r8, c1, c0, 1 /* ACTLR */ |
| mcr p15, 0, r9, c2, c0, 1 /* TTBR1 */ |
| mcr p15, 0, r10, c13, c0, 3 /* TPIDRURO */ |
| mcr p15, 0, r11, c13, c0, 1 /* context ID */ |
| isb |
| ldmdb r1!, {r4-r14} |
| /* Add 1:1 map in the PMD to allow smooth switch when turning on MMU */ |
| and r3, r3, #~0x7F /* mask off lower 7 bits of TTB */ |
| adr r0, msm_pm_mapped_pa /* get address of the mapped instr */ |
| lsr r1, r0, #20 /* get the addr range of addr in MB */ |
| lsl r1, r1, #2 /* multiply by 4 to get to the pg index */ |
| add r3, r3, r1 /* pgd + pgd_index(addr) */ |
| ldr r1, [r3] /* save current entry to r1 */ |
| lsr r0, #20 /* align current addr to 1MB boundary */ |
| lsl r0, #20 |
| /* Create new entry for this 1MB page */ |
| orr r0, r0, #0x4 /* PMD_SECT_BUFFERED */ |
| orr r0, r0, #0x400 /* PMD_SECT_AP_WRITE */ |
| orr r0, r0, #0x2 /* PMD_TYPE_SECT|PMD_DOMAIN(DOMAIN_KERNEL) */ |
| str r0, [r3] /* put new entry into the MMU table */ |
| mcr p15, 0, r3, c7, c10, 1 /* flush_pmd */ |
| dsb |
| isb |
| mcr p15, 0, r2, c1, c0, 0 /* MMU control */ |
| isb |
| msm_pm_mapped_pa: |
| /* Switch to virtual */ |
| adr r2, msm_pm_pa_to_va |
| ldr r0, =msm_pm_pa_to_va |
| mov pc, r0 |
| msm_pm_pa_to_va: |
| sub r0, r0, r2 |
| /* Restore r1 in MMU table */ |
| add r3, r3, r0 |
| str r1, [r3] |
| mcr p15, 0, r3, c7, c10, 1 /* flush_pmd */ |
| dsb |
| isb |
| mcr p15, 0, r3, c8, c7, 0 /* UTLBIALL */ |
| mcr p15, 0, r3, c7, c5, 6 /* BPIALL */ |
| dsb |
| isb |
| stmfd sp!, {lr} |
| bl v7_flush_kern_cache_all |
| ldmfd sp!, {lr} |
| mov r0, #1 |
| bx lr |
| nop |
| nop |
| nop |
| nop |
| nop |
| 1: b 1b |
| |
| |
| .data |
| |
| saved_state: |
| .space 4 * 11 /* r4-14 */ |
| .space 4 * 10 /* cp15 */ |
| #ifdef CONFIG_VFP |
| .space 8 * 32 /* VFP working registers */ |
| .space 4 * 2 /* VFP state registers */ |
| #endif |
| saved_state_end: |
| |