kernel_samsung_a34x-permissive/arch/arm64/mm/cache-mtk-v8.S

330 lines
7 KiB
ArmAsm
Raw Normal View History

/* SPDX-License-Identifier: GPL-2.0 */
/*
* Copyright (C) 2016 MediaTek Inc.
*
* This program is free software: you can redistribute it and/or modify
* it under the terms of the GNU General Public License version 2 as
* published by the Free Software Foundation.
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*/
#include <asm/assembler.h>
#include <linux/init.h>
#include <linux/linkage.h>
#include <asm/asm-uaccess.h>
.text
.equ SCTLR_C_BIT, 0x00000004
.equ SCTLR_I_BIT, 0x00001000
.equ DCISW, 0x0
.equ DCCISW, 0x1
.equ DCCSW, 0x2
.equ LOC_SHIFT, 24
.equ CLIDR_FIELD_WIDTH, 3
.equ LEVEL_SHIFT, 1
ENTRY(__enable_icache)
mrs x0, SCTLR_EL1
orr x0, x0, #SCTLR_I_BIT
msr SCTLR_EL1, x0
ret
ENDPROC(__enable_icache)
ENTRY(__disable_icache)
mrs x0, SCTLR_EL1
bic x0, x0, #SCTLR_I_BIT
msr SCTLR_EL1, x0
ret
ENDPROC(__disable_icache)
/* might pollute x0 */
.macro __dis_D
mrs x0, SCTLR_EL1
bic x0, x0, #SCTLR_C_BIT
dsb sy
msr SCTLR_EL1, x0
dsb sy
isb sy
.endm
ENTRY(__enable_dcache)
mrs x0, SCTLR_EL1
orr x0, x0, #SCTLR_C_BIT
dsb sy
msr SCTLR_EL1, x0
dsb sy
isb sy
ret
ENDPROC(__enable_dcache)
ENTRY(__disable_dcache)
mrs x0, SCTLR_EL1
bic x0, x0, #SCTLR_C_BIT
dsb sy
msr SCTLR_EL1, x0
dsb sy
isb sy
ret
ENDPROC(__disable_dcache)
ENTRY(__enable_cache)
mrs x0, SCTLR_EL1
orr x0, x0, #SCTLR_I_BIT
orr x0, x0, #SCTLR_C_BIT
dsb sy
msr SCTLR_EL1, x0
dsb sy
isb sy
ret
ENDPROC(__enable_cache)
ENTRY(__disable_cache)
mrs x0, SCTLR_EL1
bic x0, x0, #SCTLR_I_BIT
bic x0, x0, #SCTLR_C_BIT
dsb sy
msr SCTLR_EL1, x0
dsb sy
isb sy
ret
ENDPROC(__disable_cache)
/* ---------------------------------------------------------------
* Data cache operations by set/way to the level specified
*
* The main function, do_dcsw_op requires:
* x0: The operation type (0-2), as defined in arch.h
* x1: The first cache level to operate on
* x3: The last cache level to operate on
* x9: clidr_el1
* and will carry out the operation on each data cache from level 0
* to the level in x3 in sequence
*
* The dcsw_op macro sets up the x3 and x9 parameters based on
* clidr_el1 cache information before invoking the main function
* ---------------------------------------------------------------
*/
ENTRY(do_dcsw_op)
lsl x3, x3, #1
cbz x3, exit
sub x1, x1, #1
lsl x1, x1, #1
mov x10, x1
adr x14, dcsw_loop_table
add x14, x14, x0, lsl #5
mov x0, x9
mov w8, #1
loop:
add x2, x10, x10, lsr #1
lsr x1, x0, x2
and x1, x1, #7
cmp x1, #2
b.lt level_done
msr csselr_el1, x10
isb
mrs x1, ccsidr_el1
and x2, x1, #7
add x2, x2, #4
ubfx x4, x1, #3, #10
clz w5, w4
lsl w9, w4, w5
lsl w16, w8, w5
orr w9, w10, w9
ubfx w6, w1, #13, #15
lsl w17, w8, w2
dsb sy
br x14
.macro dcsw_loop _op
loop2_\_op:
lsl w7, w6, w2
loop3_\_op:
orr w11, w9, w7
dc \_op, x11
subs w7, w7, w17
b.ge loop3_\_op
subs x9, x9, x16
b.ge loop2_\_op
b level_done
.endm
level_done:
add x10, x10, #2
cmp x3, x10
b.gt loop
msr csselr_el1, xzr
dsb sy
isb
exit:
ret
ENDPROC(do_dcsw_op)
dcsw_loop_table:
dcsw_loop isw
dcsw_loop cisw
dcsw_loop csw
.macro __inner_dcache_all mode
mov x0, \mode
mov x1, #1
mrs x9, clidr_el1
ubfx x3, x9, #24, #0x7 /* LOC as last cache level */
b do_dcsw_op
.endm
.macro __inner_dcache_L1 mode
mov x0, \mode
mov x1, #1
mov x3, #1
mrs x9, clidr_el1
b do_dcsw_op
.endm
.macro __inner_dcache_L2 mode
mov x0, \mode
mov x1, #2
mov x3, #2
mrs x9, clidr_el1
b do_dcsw_op
.endm
.macro __inner_dcache_L1_L2 mode
mov x0, \mode
mov x1, #1
mov x3, #2
mrs x9, clidr_el1
b do_dcsw_op
.endm
ENTRY(__inner_flush_dcache_all)
__inner_dcache_all #DCCISW
ENDPROC(__inner_flush_dcache_all)
ENTRY(__inner_flush_dcache_L1)
__inner_dcache_L1 #DCCISW
ENDPROC(__inner_flush_dcache_L1)
ENTRY(__inner_flush_dcache_L2)
__inner_dcache_L2 #DCCISW
ENDPROC(__inner_flush_dcache_L2)
ENTRY(__inner_clean_dcache_all)
__inner_dcache_all #DCCSW
ENDPROC(__inner_clean_dcache_all)
ENTRY(__inner_clean_dcache_L1)
__inner_dcache_L1 #DCCSW
ENDPROC(__inner_clean_dcache_L1)
ENTRY(__inner_clean_dcache_L2)
__inner_dcache_L2 #DCCSW
ENDPROC(__inner_clean_dcache_L2)
ENTRY(__inner_inv_dcache_all)
__inner_dcache_all #DCISW
ENDPROC(__inner_inv_dcache_all)
ENTRY(__inner_inv_dcache_L1)
__inner_dcache_L1 #DCISW
ENDPROC(__inner_clean_dcache_L1)
ENTRY(__inner_inv_dcache_L2)
__inner_dcache_L2 #DCISW
ENDPROC(__inner_clean_dcache_L2)
ENTRY(__disable_dcache__inner_flush_dcache_L1)
__dis_D
__inner_dcache_L1 #DCCISW
ENDPROC(__disable_dcache__inner_flush_dcache_L1)
ENTRY(__disable_dcache__inner_flush_dcache_L1__inner_flush_dcache_L2)
__dis_D
__inner_dcache_L1_L2 #DCCISW
ENDPROC(__disable_dcache__inner_flush_dcache_L1__inner_flush_dcache_L2)
ENTRY(__disable_dcache__inner_clean_dcache_L1__inner_clean_dcache_L2)
__dis_D
__inner_dcache_L1_L2 #DCCSW
ENDPROC(__disable_dcache__inner_clean_dcache_L1__inner_clean_dcache_L2)
ENTRY(dis_D_inner_fL1L2)
__dis_D
/* since we need to do different operations for L1/L2,
and our current implementation would jump from
do_dcsw_op to caller(who invokes the last bl) directly,
we need to construct stack frame by ourself here.
We use two caller-saved registers, x12 & x13, to save lr & sp,
to prevent any memory access during cache operation
NOTICE: any macro or function MUST not corrupt x12 & x13 here
*/
mov x12, x29
mov x13, x30
mov x29, sp
bl __inner_flush_dcache_L1
mov x29, x12
mov x30, x13
__inner_dcache_L2 #DCCSW
ENDPROC(dis_D_inner_fL1L2)
ENTRY(dis_D_inner_flush_all)
__dis_D
__inner_dcache_all #DCCISW
ENDPROC(dis_D_inner_flush_all)
/*
* __xxxx_dcache_user_range(start,size)
*
* Ensure that any D-cache lines for the interval [start, start+size)
* are cleaned or/and invalidated to the PoC.
*
* - start - virtual start address of region (EL0/EL1)
* - size - size in question
*/
ENTRY(__flush_dcache_user_area)
uaccess_ttbr0_enable x2, x3, x4
dcache_by_line_op civac, sy, x0, x1, x2, x3
uaccess_ttbr0_disable x1, x2
ret
ENDPROC(__flush_dcache_user_area)
ENTRY(__clean_dcache_user_area)
uaccess_ttbr0_enable x2, x3, x4
dcache_by_line_op cvac, sy, x0, x1, x2, x3
uaccess_ttbr0_disable x1, x2
ret
ENDPROC(__clean_dcache_user_area)
ENTRY(__inval_dcache_user_area)
add x1, x1, x0
uaccess_ttbr0_enable x2, x3, x4
dcache_line_size x2, x3
sub x3, x2, #1
tst x1, x3 // end cache line aligned?
bic x1, x1, x3
b.eq 1f
dc civac, x1 // clean & invalidate D / U line
1: tst x0, x3 // start cache line aligned?
bic x0, x0, x3
b.eq 2f
dc civac, x0 // clean & invalidate D / U line
b 3f
2: dc civac, x0 // invalidate D / U line
3: add x0, x0, x2
cmp x0, x1
b.lo 2b
dsb sy
uaccess_ttbr0_disable x1, x2
ret
ENDPROC(__inval_dcache_user_area)