123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223 |
- /*
- * Copyright (C) 2013,2014 - ARM Ltd
- * Author: Marc Zyngier <marc.zyngier@arm.com>
- *
- * This program is free software; you can redistribute it and/or modify
- * it under the terms of the GNU General Public License version 2 as
- * published by the Free Software Foundation.
- *
- * This program is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
- * GNU General Public License for more details.
- *
- * You should have received a copy of the GNU General Public License
- * along with this program. If not, see <http://www.gnu.org/licenses/>.
- */
- #include <config.h>
- #include <linux/linkage.h>
- #include <asm/macro.h>
- #include <asm/psci.h>
- .pushsection ._secure.text, "ax"
- .arch_extension sec
- .align 5
- .globl _psci_vectors
- _psci_vectors:
- b default_psci_vector @ reset
- b default_psci_vector @ undef
- b _smc_psci @ smc
- b default_psci_vector @ pabort
- b default_psci_vector @ dabort
- b default_psci_vector @ hyp
- b default_psci_vector @ irq
- b psci_fiq_enter @ fiq
- ENTRY(psci_fiq_enter)
- movs pc, lr
- ENDPROC(psci_fiq_enter)
- .weak psci_fiq_enter
- ENTRY(default_psci_vector)
- movs pc, lr
- ENDPROC(default_psci_vector)
- .weak default_psci_vector
- ENTRY(psci_cpu_suspend)
- ENTRY(psci_cpu_off)
- ENTRY(psci_cpu_on)
- ENTRY(psci_migrate)
- mov r0, #ARM_PSCI_RET_NI @ Return -1 (Not Implemented)
- mov pc, lr
- ENDPROC(psci_migrate)
- ENDPROC(psci_cpu_on)
- ENDPROC(psci_cpu_off)
- ENDPROC(psci_cpu_suspend)
- .weak psci_cpu_suspend
- .weak psci_cpu_off
- .weak psci_cpu_on
- .weak psci_migrate
- _psci_table:
- .word ARM_PSCI_FN_CPU_SUSPEND
- .word psci_cpu_suspend
- .word ARM_PSCI_FN_CPU_OFF
- .word psci_cpu_off
- .word ARM_PSCI_FN_CPU_ON
- .word psci_cpu_on
- .word ARM_PSCI_FN_MIGRATE
- .word psci_migrate
- .word 0
- .word 0
- _smc_psci:
- push {r4-r7,lr}
- @ Switch to secure
- mrc p15, 0, r7, c1, c1, 0
- bic r4, r7, #1
- mcr p15, 0, r4, c1, c1, 0
- isb
- adr r4, _psci_table
- 1: ldr r5, [r4] @ Load PSCI function ID
- ldr r6, [r4, #4] @ Load target PC
- cmp r5, #0 @ If reach the end, bail out
- moveq r0, #ARM_PSCI_RET_INVAL @ Return -2 (Invalid)
- beq 2f
- cmp r0, r5 @ If not matching, try next entry
- addne r4, r4, #8
- bne 1b
- blx r6 @ Execute PSCI function
- @ Switch back to non-secure
- 2: mcr p15, 0, r7, c1, c1, 0
- pop {r4-r7, lr}
- movs pc, lr @ Return to the kernel
- @ Requires dense and single-cluster CPU ID space
- ENTRY(psci_get_cpu_id)
- mrc p15, 0, r0, c0, c0, 5 /* read MPIDR */
- and r0, r0, #0xff /* return CPU ID in cluster */
- bx lr
- ENDPROC(psci_get_cpu_id)
- .weak psci_get_cpu_id
- /* Imported from Linux kernel */
- LENTRY(v7_flush_dcache_all)
- dmb @ ensure ordering with previous memory accesses
- mrc p15, 1, r0, c0, c0, 1 @ read clidr
- ands r3, r0, #0x7000000 @ extract loc from clidr
- mov r3, r3, lsr #23 @ left align loc bit field
- beq finished @ if loc is 0, then no need to clean
- mov r10, #0 @ start clean at cache level 0
- flush_levels:
- add r2, r10, r10, lsr #1 @ work out 3x current cache level
- mov r1, r0, lsr r2 @ extract cache type bits from clidr
- and r1, r1, #7 @ mask of the bits for current cache only
- cmp r1, #2 @ see what cache we have at this level
- blt skip @ skip if no cache, or just i-cache
- mrs r9, cpsr @ make cssr&csidr read atomic
- mcr p15, 2, r10, c0, c0, 0 @ select current cache level in cssr
- isb @ isb to sych the new cssr&csidr
- mrc p15, 1, r1, c0, c0, 0 @ read the new csidr
- msr cpsr_c, r9
- and r2, r1, #7 @ extract the length of the cache lines
- add r2, r2, #4 @ add 4 (line length offset)
- ldr r4, =0x3ff
- ands r4, r4, r1, lsr #3 @ find maximum number on the way size
- clz r5, r4 @ find bit position of way size increment
- ldr r7, =0x7fff
- ands r7, r7, r1, lsr #13 @ extract max number of the index size
- loop1:
- mov r9, r7 @ create working copy of max index
- loop2:
- orr r11, r10, r4, lsl r5 @ factor way and cache number into r11
- orr r11, r11, r9, lsl r2 @ factor index number into r11
- mcr p15, 0, r11, c7, c14, 2 @ clean & invalidate by set/way
- subs r9, r9, #1 @ decrement the index
- bge loop2
- subs r4, r4, #1 @ decrement the way
- bge loop1
- skip:
- add r10, r10, #2 @ increment cache number
- cmp r3, r10
- bgt flush_levels
- finished:
- mov r10, #0 @ swith back to cache level 0
- mcr p15, 2, r10, c0, c0, 0 @ select current cache level in cssr
- dsb st
- isb
- bx lr
- ENDPROC(v7_flush_dcache_all)
- ENTRY(psci_disable_smp)
- mrc p15, 0, r0, c1, c0, 1 @ ACTLR
- bic r0, r0, #(1 << 6) @ Clear SMP bit
- mcr p15, 0, r0, c1, c0, 1 @ ACTLR
- isb
- dsb
- bx lr
- ENDPROC(psci_disable_smp)
- .weak psci_disable_smp
- ENTRY(psci_enable_smp)
- mrc p15, 0, r0, c1, c0, 1 @ ACTLR
- orr r0, r0, #(1 << 6) @ Set SMP bit
- mcr p15, 0, r0, c1, c0, 1 @ ACTLR
- isb
- bx lr
- ENDPROC(psci_enable_smp)
- .weak psci_enable_smp
- ENTRY(psci_cpu_off_common)
- push {lr}
- mrc p15, 0, r0, c1, c0, 0 @ SCTLR
- bic r0, r0, #(1 << 2) @ Clear C bit
- mcr p15, 0, r0, c1, c0, 0 @ SCTLR
- isb
- dsb
- bl v7_flush_dcache_all
- clrex @ Why???
- bl psci_disable_smp
- pop {lr}
- bx lr
- ENDPROC(psci_cpu_off_common)
- @ expects CPU ID in r0 and returns stack top in r0
- ENTRY(psci_get_cpu_stack_top)
- mov r5, #0x400 @ 1kB of stack per CPU
- mul r0, r0, r5
- ldr r5, =psci_text_end @ end of monitor text
- add r5, r5, #0x2000 @ Skip two pages
- lsr r5, r5, #12 @ Align to start of page
- lsl r5, r5, #12
- sub r5, r5, #4 @ reserve 1 word for target PC
- sub r0, r5, r0 @ here's our stack!
- bx lr
- ENDPROC(psci_get_cpu_stack_top)
- ENTRY(psci_cpu_entry)
- bl psci_enable_smp
- bl _nonsec_init
- bl psci_get_cpu_id @ CPU ID => r0
- bl psci_get_cpu_stack_top @ stack top => r0
- ldr r0, [r0] @ target PC at stack top
- b _do_nonsec_entry
- ENDPROC(psci_cpu_entry)
- .popsection
|