adapt omap3 to the new layered cache maintenance framework Signed-off-by: Aneesh V <aneesh@ti.com>master
parent
8b457fa828
commit
45bf05854b
@ -1,263 +0,0 @@ |
||||
/* |
||||
* Copyright (c) 2009 Wind River Systems, Inc. |
||||
* Tom Rix <Tom.Rix@windriver.com>
|
||||
* |
||||
* This file is based on and replaces the existing cache.c file |
||||
* The copyrights for the cache.c file are: |
||||
* |
||||
* (C) Copyright 2008 Texas Insturments |
||||
* |
||||
* (C) Copyright 2002 |
||||
* Sysgo Real-Time Solutions, GmbH <www.elinos.com> |
||||
* Marius Groeger <mgroeger@sysgo.de>
|
||||
* |
||||
* (C) Copyright 2002 |
||||
* Gary Jennejohn, DENX Software Engineering, <gj@denx.de>
|
||||
* |
||||
* See file CREDITS for list of people who contributed to this |
||||
* project. |
||||
* |
||||
* This program is free software; you can redistribute it and/or
|
||||
* modify it under the terms of the GNU General Public License as |
||||
* published by the Free Software Foundation; either version 2 of
|
||||
* the License, or (at your option) any later version. |
||||
* |
||||
* This program is distributed in the hope that it will be useful, |
||||
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
||||
* GNU General Public License for more details. |
||||
* |
||||
* You should have received a copy of the GNU General Public License |
||||
* along with this program; if not, write to the Free Software
|
||||
* Foundation, Inc., 59 Temple Place, Suite 330, Boston, |
||||
* MA 02111-1307 USA |
||||
*/ |
||||
|
||||
#include <asm/arch/omap3.h> |
||||
|
||||
/* |
||||
* omap3 cache code |
||||
*/ |
||||
|
||||
.align 5
|
||||
.global invalidate_dcache
|
||||
.global l2_cache_enable
|
||||
.global l2_cache_disable
|
||||
.global setup_auxcr
|
||||
|
||||
/* |
||||
* invalidate_dcache() |
||||
* |
||||
* Invalidate the whole D-cache. |
||||
* |
||||
* Corrupted registers: r0-r5, r7, r9-r11 |
||||
* |
||||
* - mm - mm_struct describing address space |
||||
*/ |
||||
invalidate_dcache: |
||||
stmfd r13!, {r0 - r5, r7, r9 - r12, r14} |
||||
|
||||
mov r7, r0 @ take a backup of device type
|
||||
cmp r0, #0x3 @ check if the device type is
|
||||
@ GP
|
||||
moveq r12, #0x1 @ set up to invalide L2
|
||||
smi: .word 0x01600070 @ Call SMI monitor (smieq)
|
||||
cmp r7, #0x3 @ compare again in case its
|
||||
@ lost
|
||||
beq finished_inval @ if GP device, inval done
|
||||
@ above
|
||||
|
||||
mrc p15, 1, r0, c0, c0, 1 @ read clidr
|
||||
ands r3, r0, #0x7000000 @ extract loc from clidr
|
||||
mov r3, r3, lsr #23 @ left align loc bit field
|
||||
beq finished_inval @ if loc is 0, then no need to
|
||||
@ clean
|
||||
mov r10, #0 @ start clean at cache level 0
|
||||
inval_loop1: |
||||
add r2, r10, r10, lsr #1 @ work out 3x current cache
|
||||
@ level
|
||||
mov r1, r0, lsr r2 @ extract cache type bits from
|
||||
@ clidr
|
||||
and r1, r1, #7 @ mask of the bits for current
|
||||
@ cache only
|
||||
cmp r1, #2 @ see what cache we have at
|
||||
@ this level
|
||||
blt skip_inval @ skip if no cache, or just
|
||||
@ i-cache
|
||||
mcr p15, 2, r10, c0, c0, 0 @ select current cache level
|
||||
@ in cssr
|
||||
mov r2, #0 @ operand for mcr SBZ
|
||||
mcr p15, 0, r2, c7, c5, 4 @ flush prefetch buffer to
|
||||
@ sych the new cssr&csidr,
|
||||
@ with armv7 this is 'isb',
|
||||
@ but we compile with armv5
|
||||
mrc p15, 1, r1, c0, c0, 0 @ read the new csidr
|
||||
and r2, r1, #7 @ extract the length of the
|
||||
@ cache lines
|
||||
add r2, r2, #4 @ add 4 (line length offset)
|
||||
ldr r4, =0x3ff |
||||
ands r4, r4, r1, lsr #3 @ find maximum number on the
|
||||
@ way size
|
||||
clz r5, r4 @ find bit position of way
|
||||
@ size increment
|
||||
ldr r7, =0x7fff |
||||
ands r7, r7, r1, lsr #13 @ extract max number of the
|
||||
@ index size
|
||||
inval_loop2: |
||||
mov r9, r4 @ create working copy of max
|
||||
@ way size
|
||||
inval_loop3: |
||||
orr r11, r10, r9, lsl r5 @ factor way and cache number
|
||||
@ into r11
|
||||
orr r11, r11, r7, lsl r2 @ factor index number into r11
|
||||
mcr p15, 0, r11, c7, c6, 2 @ invalidate by set/way
|
||||
subs r9, r9, #1 @ decrement the way
|
||||
bge inval_loop3 |
||||
subs r7, r7, #1 @ decrement the index
|
||||
bge inval_loop2 |
||||
skip_inval: |
||||
add r10, r10, #2 @ increment cache number
|
||||
cmp r3, r10 |
||||
bgt inval_loop1 |
||||
finished_inval: |
||||
mov r10, #0 @ swith back to cache level 0
|
||||
mcr p15, 2, r10, c0, c0, 0 @ select current cache level
|
||||
@ in cssr
|
||||
mcr p15, 0, r10, c7, c5, 4 @ flush prefetch buffer,
|
||||
@ with armv7 this is 'isb',
|
||||
@ but we compile with armv5
|
||||
|
||||
ldmfd r13!, {r0 - r5, r7, r9 - r12, pc} |
||||
|
||||
l2_cache_set: |
||||
stmfd r13!, {r4 - r6, lr} |
||||
mov r5, r0 |
||||
bl get_cpu_rev |
||||
mov r4, r0 |
||||
bl get_cpu_family |
||||
@ ES2 onwards we can disable/enable L2 ourselves
|
||||
cmp r0, #CPU_OMAP34XX |
||||
cmpeq r4, #CPU_3XX_ES10 |
||||
mrc 15, 0, r0, cr1, cr0, 1 |
||||
bic r0, r0, #2 |
||||
orr r0, r0, r5, lsl #1 |
||||
mcreq 15, 0, r0, cr1, cr0, 1 |
||||
@ GP Device ROM code API usage here
|
||||
@ r12 = AUXCR Write function and r0 value
|
||||
mov ip, #3 |
||||
@ SMCNE instruction to call ROM Code API
|
||||
.word 0x11600070
|
||||
ldmfd r13!, {r4 - r6, pc} |
||||
|
||||
l2_cache_enable: |
||||
mov r0, #1 |
||||
b l2_cache_set |
||||
|
||||
l2_cache_disable: |
||||
mov r0, #0 |
||||
b l2_cache_set |
||||
|
||||
/****************************************************************************** |
||||
* Routine: setup_auxcr() |
||||
* Description: Write to AuxCR desired value using SMI. |
||||
* general use. |
||||
*****************************************************************************/ |
||||
setup_auxcr: |
||||
mrc p15, 0, r0, c0, c0, 0 @ read main ID register
|
||||
and r2, r0, #0x00f00000 @ variant
|
||||
and r3, r0, #0x0000000f @ revision
|
||||
orr r1, r3, r2, lsr #20-4 @ combine variant and revision
|
||||
mov r12, #0x3 |
||||
mrc p15, 0, r0, c1, c0, 1 |
||||
orr r0, r0, #0x10 @ Enable ASA
|
||||
@ Enable L1NEON on pre-r2p1 (erratum 621766 workaround)
|
||||
cmp r1, #0x21 |
||||
orrlt r0, r0, #1 << 5 |
||||
.word 0xE1600070 @ SMC
|
||||
mov r12, #0x2 |
||||
mrc p15, 1, r0, c9, c0, 2 |
||||
@ Set PLD_FWD bit in L2AUXCR on pre-r2p1 (erratum 725233 workaround)
|
||||
cmp r1, #0x21 |
||||
orrlt r0, r0, #1 << 27 |
||||
.word 0xE1600070 @ SMC
|
||||
bx lr |
||||
|
||||
.align 5
|
||||
.global v7_flush_dcache_all
|
||||
.global v7_flush_cache_all
|
||||
|
||||
/* |
||||
* v7_flush_dcache_all() |
||||
* |
||||
* Flush the whole D-cache. |
||||
* |
||||
* Corrupted registers: r0-r7, r9-r11 (r6 only in Thumb mode) |
||||
* |
||||
* - mm - mm_struct describing address space |
||||
*/ |
||||
v7_flush_dcache_all: |
||||
# dmb @ ensure ordering with previous memory accesses
|
||||
mrc p15, 1, r0, c0, c0, 1 @ read clidr
|
||||
ands r3, r0, #0x7000000 @ extract loc from clidr
|
||||
mov r3, r3, lsr #23 @ left align loc bit field
|
||||
beq finished @ if loc is 0, then no need to clean
|
||||
mov r10, #0 @ start clean at cache level 0
|
||||
loop1: |
||||
add r2, r10, r10, lsr #1 @ work out 3x current cache level
|
||||
mov r1, r0, lsr r2 @ extract cache type bits from clidr
|
||||
and r1, r1, #7 @ mask of the bits for current cache only
|
||||
cmp r1, #2 @ see what cache we have at this level
|
||||
blt skip @ skip if no cache, or just i-cache
|
||||
mcr p15, 2, r10, c0, c0, 0 @ select current cache level in cssr
|
||||
mcr p15, 0, r10, c7, c5, 4 @ flush prefetch buffer,
|
||||
@ with armv7 this is 'isb',
|
||||
@ but we compile with armv5
|
||||
mrc p15, 1, r1, c0, c0, 0 @ read the new csidr
|
||||
and r2, r1, #7 @ extract the length of the cache lines
|
||||
add r2, r2, #4 @ add 4 (line length offset)
|
||||
ldr r4, =0x3ff |
||||
ands r4, r4, r1, lsr #3 @ find maximum number on the way size
|
||||
clz r5, r4 @ find bit position of way size increment
|
||||
ldr r7, =0x7fff |
||||
ands r7, r7, r1, lsr #13 @ extract max number of the index size
|
||||
loop2: |
||||
mov r9, r4 @ create working copy of max way size
|
||||
loop3: |
||||
orr r11, r10, r9, lsl r5 @ factor way and cache number into r11
|
||||
orr r11, r11, r7, lsl r2 @ factor index number into r11
|
||||
mcr p15, 0, r11, c7, c14, 2 @ clean & invalidate by set/way
|
||||
subs r9, r9, #1 @ decrement the way
|
||||
bge loop3 |
||||
subs r7, r7, #1 @ decrement the index
|
||||
bge loop2 |
||||
skip: |
||||
add r10, r10, #2 @ increment cache number
|
||||
cmp r3, r10 |
||||
bgt loop1 |
||||
finished: |
||||
mov r10, #0 @ swith back to cache level 0
|
||||
mcr p15, 2, r10, c0, c0, 0 @ select current cache level in cssr
|
||||
# dsb |
||||
mcr p15, 0, r10, c7, c5, 4 @ flush prefetch buffer,
|
||||
@ with armv7 this is 'isb',
|
||||
@ but we compile with armv5
|
||||
mov pc, lr |
||||
|
||||
/* |
||||
* v7_flush_cache_all() |
||||
* |
||||
* Flush the entire cache system. |
||||
* The data cache flush is now achieved using atomic clean / invalidates |
||||
* working outwards from L1 cache. This is done using Set/Way based cache |
||||
* maintainance instructions. |
||||
* The instruction cache can still be invalidated back to the point of |
||||
* unification in a single instruction. |
||||
* |
||||
*/ |
||||
v7_flush_cache_all: |
||||
stmfd sp!, {r0-r7, r9-r11, lr} |
||||
bl v7_flush_dcache_all |
||||
mov r0, #0 |
||||
mcr p15, 0, r0, c7, c5, 0 @ I+BTB cache invalidate
|
||||
ldmfd sp!, {r0-r7, r9-r11, lr} |
||||
mov pc, lr |
Loading…
Reference in new issue