/* $NetBSD: cache.S,v 1.2 2018/09/07 17:30:32 jmcneill Exp $ */ /*- * Copyright (c) 2014 Robin Randhawa * Copyright (c) 2015 The FreeBSD Foundation * All rights reserved. * * Portions of this software were developed by Andrew Turner * under sponsorship from the FreeBSD Foundation * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. * * $FreeBSD: head/sys/arm64/arm64/cpufunc_asm.S 313347 2017-02-06 17:50:09Z andrew $ */ #include #define SCTLR_M (1<<0) #define SCTLR_C (1<<2) .text .align 2 /* * Macro to handle the cache. This takes the start address in x0, length * in x1. It will corrupt x0, x1, x2, and x3. */ .macro cache_handle_range dcop = 0, ic = 0, icop = 0 .if \ic == 0 mrs x3, ctr_el0 ubfx x3, x3, #16, #4 /* x3 = D cache shift */ mov x2, #4 /* size of word */ lsl x3, x2, x3 /* x3 = D cache line size */ .else mrs x3, ctr_el0 ubfx x2, x3, #16, #4 /* x2 = D cache shift */ and x3, x3, #15 /* x3 = I cache shift */ cmp x3, x2 bcs 1f mov x3, x2 1: /* x3 = MAX(IcacheShift,DcacheShift) */ mov x2, #4 /* size of word */ lsl x3, x2, x3 /* x3 = cache line size */ .endif sub x4, x3, #1 /* Get the address mask */ and x2, x0, x4 /* Get the low bits of the address */ add x1, x1, x2 /* Add these to the size */ bic x0, x0, x4 /* Clear the low bit of the address */ 1: dc \dcop, x0 dsb ish .if \ic != 0 ic \icop, x0 dsb ish .endif add x0, x0, x3 /* Move to the next line */ subs x1, x1, x3 /* Reduce the size */ b.hi 1b /* Check if we are done */ .if \ic != 0 isb .endif ret .endm /* * void aarch64_dcache_wbinv_range(vaddr_t, vsize_t) */ ENTRY(aarch64_dcache_wbinv_range) cache_handle_range dcop = civac END(aarch64_dcache_wbinv_range) /* * void aarch64_icache_inv_all(void) */ ENTRY(aarch64_icache_inv_all) dsb ish ic ialluis dsb ish isb ret END(aarch64_icache_inv_all) /* * void aarch64_exec_kernel(paddr_t entry, paddr_t dtb) */ ENTRY(aarch64_exec_kernel) mov x20, x0 /* kernel entry point */ mov x21, x1 /* dtb address */ mrs x0, CurrentEL lsr x0, x0, #2 cmp x0, #0x2 b.eq 1f /* Disable MMU and dcache, CurrentEL = EL1 */ mrs x0, sctlr_el1 bic x0, x0, #SCTLR_M bic x0, x0, #SCTLR_C msr sctlr_el1, x0 isb b 2f 1: /* Disable MMU and dcache, CurrentEL = EL2 */ mrs x0, sctlr_el2 bic x0, x0, #SCTLR_M bic x0, x0, #SCTLR_C msr sctlr_el2, x0 isb 2: /* Jump to kernel */ mov x0, x21 mov x1, xzr mov x2, xzr mov x3, xzr br x20 END(aarch64_exec_kernel)