From f4f382d8c0e6d00d433209a837d6a686cc9c99b5 Mon Sep 17 00:00:00 2001 From: Ben Gras Date: Tue, 10 Dec 2013 19:51:07 +0100 Subject: [PATCH] align ARM cpu.h importing & using armreg.h Change-Id: I4793517d936f71b0bb7088fbfe67e73a65fafb11 --- kernel/arch/earm/arch_system.c | 1 + kernel/arch/earm/include/archconst.h | 5 +- kernel/arch/earm/mpx.S | 11 +- kernel/arch/earm/pg_utils.c | 15 +- sys/arch/arm/include/Makefile | 2 +- sys/arch/arm/include/armreg.h | 655 +++++++++++++++++++++++++++ sys/arch/arm/include/cpu.h | 464 ++++++++++++++++--- 7 files changed, 1087 insertions(+), 66 deletions(-) create mode 100644 sys/arch/arm/include/armreg.h diff --git a/kernel/arch/earm/arch_system.c b/kernel/arch/earm/arch_system.c index 57b8ac7ce..142f4d756 100644 --- a/kernel/arch/earm/arch_system.c +++ b/kernel/arch/earm/arch_system.c @@ -9,6 +9,7 @@ #include #include #include +#include #include diff --git a/kernel/arch/earm/include/archconst.h b/kernel/arch/earm/include/archconst.h index 96999e898..8d2edf967 100644 --- a/kernel/arch/earm/include/archconst.h +++ b/kernel/arch/earm/include/archconst.h @@ -5,10 +5,11 @@ #include #include #include +#include /* Program stack words and masks. */ -#define INIT_PSR (MODE_USR | PSR_F) /* initial psr */ -#define INIT_TASK_PSR (MODE_SVC | PSR_F) /* initial psr for tasks */ +#define INIT_PSR (PSR_USR32_MODE | PSR_F) /* initial psr */ +#define INIT_TASK_PSR (PSR_SVC32_MODE | PSR_F) /* initial psr for tasks */ /* Exception vector numbers */ #define RESET_VECTOR 0 diff --git a/kernel/arch/earm/mpx.S b/kernel/arch/earm/mpx.S index d10b4b6f8..c7371e988 100644 --- a/kernel/arch/earm/mpx.S +++ b/kernel/arch/earm/mpx.S @@ -27,6 +27,7 @@ #include #include #include +#include #include "bsp_intr.h" #include "arch_proto.h" /* K_STACK_SIZE */ @@ -46,11 +47,11 @@ IMPORT(svc_stack) */ .macro switch_to_svc lr_offset sub lr, lr, #\lr_offset /* do the adjustment */ - srsdb sp!, #MODE_SVC /* store the saved the return */ + srsdb sp!, #PSR_SVC32_MODE /* store the saved the return */ /* address and program status */ /* register onto the kernel stack */ /* Also modify the stack pointer. */ - cps #MODE_SVC /* do the switch to SVC. */ + cps #PSR_SVC32_MODE /* do the switch to SVC. */ .endm /* @@ -63,8 +64,8 @@ IMPORT(svc_stack) ldr r3, [sp, #8] /* get spsr. */ orr r3, r3, #(PSR_F | PSR_I) /* mask interrupts on return. */ str r3, [sp, #8] /* store spsr. */ - and r3, r3, #PSR_MODE_MASK /* mask the ARM mode. */ - cmp r3, #MODE_USR /* compare it to user mode. */ + and r3, r3, #PSR_MODE /* mask the ARM mode. */ + cmp r3, #PSR_USR32_MODE /* compare it to user mode. */ pop {r3} bne \label /* In-kernel handling. */ .endm @@ -171,7 +172,7 @@ irq_entry_from_kernel: */ ENTRY(svc_entry) /* Store the LR and the SPSR of the current mode onto the SVC stack */ - srsdb sp!, #MODE_SVC + srsdb sp!, #PSR_SVC32_MODE save_process_ctx /* save the pointer to the current process */ diff --git a/kernel/arch/earm/pg_utils.c b/kernel/arch/earm/pg_utils.c index 438f48761..84cef7ed2 100644 --- a/kernel/arch/earm/pg_utils.c +++ b/kernel/arch/earm/pg_utils.c @@ -5,6 +5,7 @@ #include "kernel/kernel.h" #include "arch_proto.h" #include +#include #include #include @@ -214,21 +215,21 @@ void vm_enable_paging(void) sctlr = read_sctlr(); /* Enable MMU */ - sctlr |= SCTLR_M; + sctlr |= CPU_CONTROL_MMU_ENABLE; /* TRE set to zero (default reset value): TEX[2:0] are used, plus C and B bits.*/ - sctlr &= ~SCTLR_TRE; + sctlr &= ~CPU_CONTROL_TR_ENABLE; /* AFE set to zero (default reset value): not using simplified model. */ - sctlr &= ~SCTLR_AFE; + sctlr &= ~CPU_CONTROL_AF_ENABLE; /* Enable instruction ,data cache and branch prediction */ - sctlr |= SCTLR_C; - sctlr |= SCTLR_I; - sctlr |= SCTLR_Z; + sctlr |= CPU_CONTROL_DC_ENABLE; + sctlr |= CPU_CONTROL_IC_ENABLE; + sctlr |= CPU_CONTROL_BPRD_ENABLE; /* Enable barriers */ - sctlr |= SCTLR_CP15BEN; + sctlr |= CPU_CONTROL_32BD_ENABLE; /* Enable L2 cache (cortex-a8) */ #define CORTEX_A8_L2EN (0x02) diff --git a/sys/arch/arm/include/Makefile b/sys/arch/arm/include/Makefile index b33c5406e..dee4cf777 100644 --- a/sys/arch/arm/include/Makefile +++ b/sys/arch/arm/include/Makefile @@ -2,7 +2,7 @@ INCSDIR= /usr/include/arm -INCS= aeabi.h ansi.h asm.h \ +INCS= aeabi.h ansi.h armreg.h asm.h \ bswap.h byte_swap.h \ cdefs.h cpu.h \ disklabel.h \ diff --git a/sys/arch/arm/include/armreg.h b/sys/arch/arm/include/armreg.h new file mode 100644 index 000000000..a7218deb0 --- /dev/null +++ b/sys/arch/arm/include/armreg.h @@ -0,0 +1,655 @@ +/* $NetBSD: armreg.h,v 1.69 2012/09/27 21:48:17 matt Exp $ */ + +/* + * Copyright (c) 1998, 2001 Ben Harris + * Copyright (c) 1994-1996 Mark Brinicombe. + * Copyright (c) 1994 Brini. + * All rights reserved. + * + * This code is derived from software written for Brini by Mark Brinicombe + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions + * are met: + * 1. Redistributions of source code must retain the above copyright + * notice, this list of conditions and the following disclaimer. + * 2. Redistributions in binary form must reproduce the above copyright + * notice, this list of conditions and the following disclaimer in the + * documentation and/or other materials provided with the distribution. + * 3. All advertising materials mentioning features or use of this software + * must display the following acknowledgement: + * This product includes software developed by Brini. + * 4. The name of the company nor the name of the author may be used to + * endorse or promote products derived from this software without specific + * prior written permission. + * + * THIS SOFTWARE IS PROVIDED BY BRINI ``AS IS'' AND ANY EXPRESS OR IMPLIED + * WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF + * MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. + * IN NO EVENT SHALL BRINI OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, + * INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES + * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR + * SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) + * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT + * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY + * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF + * SUCH DAMAGE. + */ + +#ifndef _ARM_ARMREG_H +#define _ARM_ARMREG_H + +/* + * ARM Process Status Register + * + * The picture in the ARM manuals looks like this: + * 3 3 2 2 2 2 + * 1 0 9 8 7 6 8 7 6 5 4 0 + * +-+-+-+-+-+-------------------------------------+-+-+-+---------+ + * |N|Z|C|V|Q| reserved |I|F|T|M M M M M| + * | | | | | | | | | |4 3 2 1 0| + * +-+-+-+-+-+-------------------------------------+-+-+-+---------+ + */ + +#define PSR_FLAGS 0xf0000000 /* flags */ +#define PSR_N_bit (1 << 31) /* negative */ +#define PSR_Z_bit (1 << 30) /* zero */ +#define PSR_C_bit (1 << 29) /* carry */ +#define PSR_V_bit (1 << 28) /* overflow */ + +#define PSR_Q_bit (1 << 27) /* saturation */ + +#define I32_bit (1 << 7) /* IRQ disable */ +#define F32_bit (1 << 6) /* FIQ disable */ +#define IF32_bits (3 << 6) /* IRQ/FIQ disable */ + +#define PSR_T_bit (1 << 5) /* Thumb state */ +#define PSR_J_bit (1 << 24) /* Java mode */ + +#ifdef __minix +/* Minix uses these aliases */ +#define PSR_F F32_bit +#define PSR_I I32_bit +#endif + +#define PSR_MODE 0x0000001f /* mode mask */ +#define PSR_USR26_MODE 0x00000000 +#define PSR_FIQ26_MODE 0x00000001 +#define PSR_IRQ26_MODE 0x00000002 +#define PSR_SVC26_MODE 0x00000003 +#define PSR_USR32_MODE 0x00000010 +#define PSR_FIQ32_MODE 0x00000011 +#define PSR_IRQ32_MODE 0x00000012 +#define PSR_SVC32_MODE 0x00000013 +#define PSR_MON32_MODE 0x00000016 +#define PSR_ABT32_MODE 0x00000017 +#define PSR_HYP32_MODE 0x0000001a +#define PSR_UND32_MODE 0x0000001b +#define PSR_SYS32_MODE 0x0000001f +#define PSR_32_MODE 0x00000010 + +#define PSR_IN_USR_MODE(psr) (!((psr) & 3)) /* XXX */ +#define PSR_IN_32_MODE(psr) ((psr) & PSR_32_MODE) + +/* In 26-bit modes, the PSR is stuffed into R15 along with the PC. */ + +#define R15_MODE 0x00000003 +#define R15_MODE_USR 0x00000000 +#define R15_MODE_FIQ 0x00000001 +#define R15_MODE_IRQ 0x00000002 +#define R15_MODE_SVC 0x00000003 + +#define R15_PC 0x03fffffc + +#define R15_FIQ_DISABLE 0x04000000 +#define R15_IRQ_DISABLE 0x08000000 + +#define R15_FLAGS 0xf0000000 +#define R15_FLAG_N 0x80000000 +#define R15_FLAG_Z 0x40000000 +#define R15_FLAG_C 0x20000000 +#define R15_FLAG_V 0x10000000 + +/* + * Co-processor 15: The system control co-processor. + */ + +#define ARM_CP15_CPU_ID 0 + +/* + * The CPU ID register is theoretically structured, but the definitions of + * the fields keep changing. + */ + +/* The high-order byte is always the implementor */ +#define CPU_ID_IMPLEMENTOR_MASK 0xff000000 +#define CPU_ID_ARM_LTD 0x41000000 /* 'A' */ +#define CPU_ID_DEC 0x44000000 /* 'D' */ +#define CPU_ID_INTEL 0x69000000 /* 'i' */ +#define CPU_ID_TI 0x54000000 /* 'T' */ +#define CPU_ID_MARVELL 0x56000000 /* 'V' */ +#define CPU_ID_FARADAY 0x66000000 /* 'f' */ + +/* How to decide what format the CPUID is in. */ +#define CPU_ID_ISOLD(x) (((x) & 0x0000f000) == 0x00000000) +#define CPU_ID_IS7(x) (((x) & 0x0000f000) == 0x00007000) +#define CPU_ID_ISNEW(x) (!CPU_ID_ISOLD(x) && !CPU_ID_IS7(x)) + +/* On ARM3 and ARM6, this byte holds the foundry ID. */ +#define CPU_ID_FOUNDRY_MASK 0x00ff0000 +#define CPU_ID_FOUNDRY_VLSI 0x00560000 + +/* On ARM7 it holds the architecture and variant (sub-model) */ +#define CPU_ID_7ARCH_MASK 0x00800000 +#define CPU_ID_7ARCH_V3 0x00000000 +#define CPU_ID_7ARCH_V4T 0x00800000 +#define CPU_ID_7VARIANT_MASK 0x007f0000 + +/* On more recent ARMs, it does the same, but in a different format */ +#define CPU_ID_ARCH_MASK 0x000f0000 +#define CPU_ID_ARCH_V3 0x00000000 +#define CPU_ID_ARCH_V4 0x00010000 +#define CPU_ID_ARCH_V4T 0x00020000 +#define CPU_ID_ARCH_V5 0x00030000 +#define CPU_ID_ARCH_V5T 0x00040000 +#define CPU_ID_ARCH_V5TE 0x00050000 +#define CPU_ID_ARCH_V5TEJ 0x00060000 +#define CPU_ID_ARCH_V6 0x00070000 +#define CPU_ID_VARIANT_MASK 0x00f00000 + +/* Next three nybbles are part number */ +#define CPU_ID_PARTNO_MASK 0x0000fff0 + +/* Intel XScale has sub fields in part number */ +#define CPU_ID_XSCALE_COREGEN_MASK 0x0000e000 /* core generation */ +#define CPU_ID_XSCALE_COREREV_MASK 0x00001c00 /* core revision */ +#define CPU_ID_XSCALE_PRODUCT_MASK 0x000003f0 /* product number */ + +/* And finally, the revision number. */ +#define CPU_ID_REVISION_MASK 0x0000000f + +/* Individual CPUs are probably best IDed by everything but the revision. */ +#define CPU_ID_CPU_MASK 0xfffffff0 + +/* Fake CPU IDs for ARMs without CP15 */ +#define CPU_ID_ARM2 0x41560200 +#define CPU_ID_ARM250 0x41560250 + +/* Pre-ARM7 CPUs -- [15:12] == 0 */ +#define CPU_ID_ARM3 0x41560300 +#define CPU_ID_ARM600 0x41560600 +#define CPU_ID_ARM610 0x41560610 +#define CPU_ID_ARM620 0x41560620 + +/* ARM7 CPUs -- [15:12] == 7 */ +#define CPU_ID_ARM700 0x41007000 /* XXX This is a guess. */ +#define CPU_ID_ARM710 0x41007100 +#define CPU_ID_ARM7500 0x41027100 +#define CPU_ID_ARM710A 0x41047100 /* inc ARM7100 */ +#define CPU_ID_ARM7500FE 0x41077100 +#define CPU_ID_ARM710T 0x41807100 +#define CPU_ID_ARM720T 0x41807200 +#define CPU_ID_ARM740T8K 0x41807400 /* XXX no MMU, 8KB cache */ +#define CPU_ID_ARM740T4K 0x41817400 /* XXX no MMU, 4KB cache */ + +/* Post-ARM7 CPUs */ +#define CPU_ID_ARM810 0x41018100 +#define CPU_ID_ARM920T 0x41129200 +#define CPU_ID_ARM922T 0x41029220 +#define CPU_ID_ARM926EJS 0x41069260 +#define CPU_ID_ARM940T 0x41029400 /* XXX no MMU */ +#define CPU_ID_ARM946ES 0x41049460 /* XXX no MMU */ +#define CPU_ID_ARM966ES 0x41049660 /* XXX no MMU */ +#define CPU_ID_ARM966ESR1 0x41059660 /* XXX no MMU */ +#define CPU_ID_ARM1020E 0x4115a200 /* (AKA arm10 rev 1) */ +#define CPU_ID_ARM1022ES 0x4105a220 +#define CPU_ID_ARM1026EJS 0x4106a260 +#define CPU_ID_ARM11MPCORE 0x410fb020 +#define CPU_ID_ARM1136JS 0x4107b360 +#define CPU_ID_ARM1136JSR1 0x4117b360 +#define CPU_ID_ARM1156T2S 0x4107b560 /* MPU only */ +#define CPU_ID_ARM1176JZS 0x410fb760 +#define CPU_ID_ARM11_P(n) ((n & 0xff07f000) == 0x4107b000) +#define CPU_ID_CORTEXA5R0 0x410fc050 +#define CPU_ID_CORTEXA8R1 0x411fc080 +#define CPU_ID_CORTEXA8R2 0x412fc080 +#define CPU_ID_CORTEXA8R3 0x413fc080 +#define CPU_ID_CORTEXA9R2 0x411fc090 +#define CPU_ID_CORTEXA9R3 0x412fc090 +#define CPU_ID_CORTEXA9R4 0x413fc090 +#define CPU_ID_CORTEXA15R2 0x412fc0f0 +#define CPU_ID_CORTEXA15R3 0x413fc0f0 +#define CPU_ID_CORTEX_P(n) ((n & 0xff0ff000) == 0x410fc000) +#define CPU_ID_CORTEX_A8_P(n) ((n & 0xff0ff0f0) == 0x410fc080) +#define CPU_ID_CORTEX_A9_P(n) ((n & 0xff0ff0f0) == 0x410fc090) +#define CPU_ID_SA110 0x4401a100 +#define CPU_ID_SA1100 0x4401a110 +#define CPU_ID_TI925T 0x54029250 +#define CPU_ID_MV88FR571_VD 0x56155710 +#define CPU_ID_MV88SV131 0x56251310 +#define CPU_ID_FA526 0x66015260 +#define CPU_ID_SA1110 0x6901b110 +#define CPU_ID_IXP1200 0x6901c120 +#define CPU_ID_80200 0x69052000 +#define CPU_ID_PXA250 0x69052100 /* sans core revision */ +#define CPU_ID_PXA210 0x69052120 +#define CPU_ID_PXA250A 0x69052100 /* 1st version Core */ +#define CPU_ID_PXA210A 0x69052120 /* 1st version Core */ +#define CPU_ID_PXA250B 0x69052900 /* 3rd version Core */ +#define CPU_ID_PXA210B 0x69052920 /* 3rd version Core */ +#define CPU_ID_PXA250C 0x69052d00 /* 4th version Core */ +#define CPU_ID_PXA210C 0x69052d20 /* 4th version Core */ +#define CPU_ID_PXA27X 0x69054110 +#define CPU_ID_80321_400 0x69052420 +#define CPU_ID_80321_600 0x69052430 +#define CPU_ID_80321_400_B0 0x69052c20 +#define CPU_ID_80321_600_B0 0x69052c30 +#define CPU_ID_80219_400 0x69052e20 +#define CPU_ID_80219_600 0x69052e30 +#define CPU_ID_IXP425_533 0x690541c0 +#define CPU_ID_IXP425_400 0x690541d0 +#define CPU_ID_IXP425_266 0x690541f0 + +/* ARM3-specific coprocessor 15 registers */ +#define ARM3_CP15_FLUSH 1 +#define ARM3_CP15_CONTROL 2 +#define ARM3_CP15_CACHEABLE 3 +#define ARM3_CP15_UPDATEABLE 4 +#define ARM3_CP15_DISRUPTIVE 5 + +/* ARM3 Control register bits */ +#define ARM3_CTL_CACHE_ON 0x00000001 +#define ARM3_CTL_SHARED 0x00000002 +#define ARM3_CTL_MONITOR 0x00000004 + +/* + * Post-ARM3 CP15 registers: + * + * 1 Control register + * + * 2 Translation Table Base + * + * 3 Domain Access Control + * + * 4 Reserved + * + * 5 Fault Status + * + * 6 Fault Address + * + * 7 Cache/write-buffer Control + * + * 8 TLB Control + * + * 9 Cache Lockdown + * + * 10 TLB Lockdown + * + * 11 Reserved + * + * 12 Reserved + * + * 13 Process ID (for FCSE) + * + * 14 Reserved + * + * 15 Implementation Dependent + */ + +/* Some of the definitions below need cleaning up for V3/V4 architectures */ + +/* CPU control register (CP15 register 1) */ +#define CPU_CONTROL_MMU_ENABLE 0x00000001 /* M: MMU/Protection unit enable */ +#define CPU_CONTROL_AFLT_ENABLE 0x00000002 /* A: Alignment fault enable */ +#define CPU_CONTROL_DC_ENABLE 0x00000004 /* C: IDC/DC enable */ +#define CPU_CONTROL_WBUF_ENABLE 0x00000008 /* W: Write buffer enable */ +#define CPU_CONTROL_32BP_ENABLE 0x00000010 /* P: 32-bit exception handlers */ +#define CPU_CONTROL_32BD_ENABLE 0x00000020 /* D: 32-bit addressing */ +#define CPU_CONTROL_LABT_ENABLE 0x00000040 /* L: Late abort enable */ +#define CPU_CONTROL_BEND_ENABLE 0x00000080 /* B: Big-endian mode */ +#define CPU_CONTROL_SYST_ENABLE 0x00000100 /* S: System protection bit */ +#define CPU_CONTROL_ROM_ENABLE 0x00000200 /* R: ROM protection bit */ +#define CPU_CONTROL_CPCLK 0x00000400 /* F: Implementation defined */ +#define CPU_CONTROL_SWP_ENABLE 0x00000400 /* SW: SWP{B} perform normally. */ +#define CPU_CONTROL_BPRD_ENABLE 0x00000800 /* Z: Branch prediction enable */ +#define CPU_CONTROL_IC_ENABLE 0x00001000 /* I: IC enable */ +#define CPU_CONTROL_VECRELOC 0x00002000 /* V: Vector relocation */ +#define CPU_CONTROL_ROUNDROBIN 0x00004000 /* RR: Predictable replacement */ +#define CPU_CONTROL_V4COMPAT 0x00008000 /* L4: ARMv4 compat LDR R15 etc */ +#define CPU_CONTROL_FI_ENABLE 0x00200000 /* FI: Low interrupt latency */ +#define CPU_CONTROL_UNAL_ENABLE 0x00400000 /* U: unaligned data access */ +#define CPU_CONTROL_XP_ENABLE 0x00800000 /* XP: extended page table */ +#define CPU_CONTROL_V_ENABLE 0x01000000 /* VE: Interrupt vectors enable */ +#define CPU_CONTROL_EX_BEND 0x02000000 /* EE: exception endianness */ +#define CPU_CONTROL_NMFI 0x08000000 /* NMFI: Non maskable FIQ */ +#define CPU_CONTROL_TR_ENABLE 0x10000000 /* TRE: */ +#define CPU_CONTROL_AF_ENABLE 0x20000000 /* AFE: Access flag enable */ +#define CPU_CONTROL_TE_ENABLE 0x40000000 /* TE: Thumb Exception enable */ + +#define CPU_CONTROL_IDC_ENABLE CPU_CONTROL_DC_ENABLE + +/* ARMv6/ARMv7 Co-Processor Access Control Register (CP15, 0, c1, c0, 2) */ +#define CPACR_V7_ASEDIS 0x80000000 /* Disable Advanced SIMD Ext. */ +#define CPACR_V7_D32DIS 0x40000000 /* Disable VFP regs 15-31 */ +#define CPACR_CPn(n) (3 << (2*n)) +#define CPACR_NOACCESS 0 /* reset value */ +#define CPACR_PRIVED 1 /* Privileged mode access */ +#define CPACR_RESERVED 2 +#define CPACR_ALL 3 /* Privileged and User mode access */ + +/* ARM11x6 Auxiliary Control Register (CP15 register 1, opcode2 1) */ +#define ARM11X6_AUXCTL_RS 0x00000001 /* return stack */ +#define ARM11X6_AUXCTL_DB 0x00000002 /* dynamic branch prediction */ +#define ARM11X6_AUXCTL_SB 0x00000004 /* static branch prediction */ +#define ARM11X6_AUXCTL_TR 0x00000008 /* MicroTLB replacement strat. */ +#define ARM11X6_AUXCTL_EX 0x00000010 /* exclusive L1/L2 cache */ +#define ARM11X6_AUXCTL_RA 0x00000020 /* clean entire cache disable */ +#define ARM11X6_AUXCTL_RV 0x00000040 /* block transfer cache disable */ +#define ARM11X6_AUXCTL_CZ 0x00000080 /* restrict cache size */ + +/* ARM1136 Auxiliary Control Register (CP15 register 1, opcode2 1) */ +#define ARM1136_AUXCTL_PFI 0x80000000 /* PFI: partial FI mode. */ + /* This is an undocumented flag + * used to work around a cache bug + * in r0 steppings. See errata + * 364296. + */ +/* ARM1176 Auxiliary Control Register (CP15 register 1, opcode2 1) */ +#define ARM1176_AUXCTL_PHD 0x10000000 /* inst. prefetch halting disable */ +#define ARM1176_AUXCTL_BFD 0x20000000 /* branch folding disable */ +#define ARM1176_AUXCTL_FSD 0x40000000 /* force speculative ops disable */ +#define ARM1176_AUXCTL_FIO 0x80000000 /* low intr latency override */ + +/* Cortex-A9 Auxiliary Control Register (CP15 register 1, opcode2 1) */ +#define CORTEXA9_AUXCTL_PARITY 0x00000200 /* Enable parity */ +#define CORTEXA9_AUXCTL_1WAY 0x00000100 /* Alloc in one way only */ +#define CORTEXA9_AUXCTL_EXCL 0x00000080 /* Exclusive cache */ +#define CORTEXA9_AUXCTL_SMP 0x00000040 /* CPU is in SMP mode */ +#define CORTEXA9_AUXCTL_WRZERO 0x00000008 /* Write full line of zeroes */ +#define CORTEXA9_AUXCTL_L1PLD 0x00000004 /* L1 Dside prefetch */ +#define CORTEXA9_AUXCTL_L2PLD 0x00000002 /* L2 Dside prefetch */ +#define CORTEXA9_AUXCTL_FW 0x00000001 /* Forward Cache/TLB ops */ + +/* XScale Auxiliary Control Register (CP15 register 1, opcode2 1) */ +#define XSCALE_AUXCTL_K 0x00000001 /* dis. write buffer coalescing */ +#define XSCALE_AUXCTL_P 0x00000002 /* ECC protect page table access */ +#define XSCALE_AUXCTL_MD_WB_RA 0x00000000 /* mini-D$ wb, read-allocate */ +#define XSCALE_AUXCTL_MD_WB_RWA 0x00000010 /* mini-D$ wb, read/write-allocate */ +#define XSCALE_AUXCTL_MD_WT 0x00000020 /* mini-D$ wt, read-allocate */ +#define XSCALE_AUXCTL_MD_MASK 0x00000030 + +/* ARM11 MPCore Auxiliary Control Register (CP15 register 1, opcode2 1) */ +#define MPCORE_AUXCTL_RS 0x00000001 /* return stack */ +#define MPCORE_AUXCTL_DB 0x00000002 /* dynamic branch prediction */ +#define MPCORE_AUXCTL_SB 0x00000004 /* static branch prediction */ +#define MPCORE_AUXCTL_F 0x00000008 /* instruction folding enable */ +#define MPCORE_AUXCTL_EX 0x00000010 /* exclusive L1/L2 cache */ +#define MPCORE_AUXCTL_SA 0x00000020 /* SMP/AMP */ + +/* Cortex-A9 Auxiliary Control Register (CP15 register 1, opcode 1) */ +#define CORTEXA9_AUXCTL_FW 0x00000001 /* Cache and TLB updates broadcast */ +#define CORTEXA9_AUXCTL_L2_PLD 0x00000002 /* Prefetch hint enable */ +#define CORTEXA9_AUXCTL_L1_PLD 0x00000004 /* Data prefetch hint enable */ +#define CORTEXA9_AUXCTL_WR_ZERO 0x00000008 /* Ena. write full line of 0s mode */ +#define CORTEXA9_AUXCTL_SMP 0x00000040 /* Coherency is active */ +#define CORTEXA9_AUXCTL_EXCL 0x00000080 /* Exclusive cache bit */ +#define CORTEXA9_AUXCTL_ONEWAY 0x00000100 /* Allocate in on cache way only */ +#define CORTEXA9_AUXCTL_PARITY 0x00000200 /* Support parity checking */ + +/* Marvell Feroceon Extra Features Register (CP15 register 1, opcode2 0) */ +#define FC_DCACHE_REPL_LOCK 0x80000000 /* Replace DCache Lock */ +#define FC_DCACHE_STREAM_EN 0x20000000 /* DCache Streaming Switch */ +#define FC_WR_ALLOC_EN 0x10000000 /* Enable Write Allocate */ +#define FC_L2_PREF_DIS 0x01000000 /* L2 Cache Prefetch Disable */ +#define FC_L2_INV_EVICT_LINE 0x00800000 /* L2 Invalidates Uncorrectable Error Line Eviction */ +#define FC_L2CACHE_EN 0x00400000 /* L2 enable */ +#define FC_ICACHE_REPL_LOCK 0x00080000 /* Replace ICache Lock */ +#define FC_GLOB_HIST_REG_EN 0x00040000 /* Branch Global History Register Enable */ +#define FC_BRANCH_TARG_BUF_DIS 0x00020000 /* Branch Target Buffer Disable */ +#define FC_L1_PAR_ERR_EN 0x00010000 /* L1 Parity Error Enable */ + +/* Cache type register definitions 0 */ +#define CPU_CT_FORMAT(x) (((x) >> 29) & 0x7) /* reg format */ +#define CPU_CT_ISIZE(x) ((x) & 0xfff) /* I$ info */ +#define CPU_CT_DSIZE(x) (((x) >> 12) & 0xfff) /* D$ info */ +#define CPU_CT_S (1U << 24) /* split cache */ +#define CPU_CT_CTYPE(x) (((x) >> 25) & 0xf) /* cache type */ + +#define CPU_CT_CTYPE_WT 0 /* write-through */ +#define CPU_CT_CTYPE_WB1 1 /* write-back, clean w/ read */ +#define CPU_CT_CTYPE_WB2 2 /* w/b, clean w/ cp15,7 */ +#define CPU_CT_CTYPE_WB6 6 /* w/b, cp15,7, lockdown fmt A */ +#define CPU_CT_CTYPE_WB7 7 /* w/b, cp15,7, lockdown fmt B */ +#define CPU_CT_CTYPE_WB14 14 /* w/b, cp15,7, lockdown fmt C */ + +#define CPU_CT_xSIZE_LEN(x) ((x) & 0x3) /* line size */ +#define CPU_CT_xSIZE_M (1U << 2) /* multiplier */ +#define CPU_CT_xSIZE_ASSOC(x) (((x) >> 3) & 0x7) /* associativity */ +#define CPU_CT_xSIZE_SIZE(x) (((x) >> 6) & 0x7) /* size */ +#define CPU_CT_xSIZE_P (1U << 11) /* need to page-color */ + +/* format 4 definitions */ +#define CPU_CT4_ILINE(x) ((x) & 0xf) /* I$ line size */ +#define CPU_CT4_DLINE(x) (((x) >> 16) & 0xf) /* D$ line size */ +#define CPU_CT4_L1IPOLICY(x) (((x) >> 14) & 0x3) /* I$ policy */ +#define CPU_CT4_L1_AIVIVT 1 /* ASID tagged VIVT */ +#define CPU_CT4_L1_VIPT 2 /* VIPT */ +#define CPU_CT4_L1_PIPT 3 /* PIPT */ +#define CPU_CT4_ERG(x) (((x) >> 20) & 0xf) /* Cache WriteBack Granule */ +#define CPU_CT4_CWG(x) (((x) >> 24) & 0xf) /* Exclusive Resv. Granule */ + +/* Cache size identifaction register definitions 1, Rd, c0, c0, 0 */ +#define CPU_CSID_CTYPE_WT 0x80000000 /* write-through avail */ +#define CPU_CSID_CTYPE_WB 0x40000000 /* write-back avail */ +#define CPU_CSID_CTYPE_RA 0x20000000 /* read-allocation avail */ +#define CPU_CSID_CTYPE_WA 0x10000000 /* write-allocation avail */ +#define CPU_CSID_NUMSETS(x) (((x) >> 13) & 0x7fff) +#define CPU_CSID_ASSOC(x) (((x) >> 3) & 0x1ff) +#define CPU_CSID_LEN(x) ((x) & 0x07) + +/* Cache size selection register definitions 2, Rd, c0, c0, 0 */ +#define CPU_CSSR_L2 0x00000002 +#define CPU_CSSR_L1 0x00000000 +#define CPU_CSSR_InD 0x00000001 + +/* Fault status register definitions */ + +#define FAULT_TYPE_MASK 0x0f +#define FAULT_USER 0x10 + +#define FAULT_WRTBUF_0 0x00 /* Vector Exception */ +#define FAULT_WRTBUF_1 0x02 /* Terminal Exception */ +#define FAULT_BUSERR_0 0x04 /* External Abort on Linefetch -- Section */ +#define FAULT_BUSERR_1 0x06 /* External Abort on Linefetch -- Page */ +#define FAULT_BUSERR_2 0x08 /* External Abort on Non-linefetch -- Section */ +#define FAULT_BUSERR_3 0x0a /* External Abort on Non-linefetch -- Page */ +#define FAULT_BUSTRNL1 0x0c /* External abort on Translation -- Level 1 */ +#define FAULT_BUSTRNL2 0x0e /* External abort on Translation -- Level 2 */ +#define FAULT_ALIGN_0 0x01 /* Alignment */ +#define FAULT_ALIGN_1 0x03 /* Alignment */ +#define FAULT_TRANS_S 0x05 /* Translation -- Section */ +#define FAULT_TRANS_P 0x07 /* Translation -- Page */ +#define FAULT_DOMAIN_S 0x09 /* Domain -- Section */ +#define FAULT_DOMAIN_P 0x0b /* Domain -- Page */ +#define FAULT_PERM_S 0x0d /* Permission -- Section */ +#define FAULT_PERM_P 0x0f /* Permission -- Page */ + +#define FAULT_IMPRECISE 0x400 /* Imprecise exception (XSCALE) */ + +/* + * Address of the vector page, low and high versions. + */ +#define ARM_VECTORS_LOW 0x00000000U +#define ARM_VECTORS_HIGH 0xffff0000U + +/* + * ARM Instructions + * + * 3 3 2 2 2 + * 1 0 9 8 7 0 + * +-------+-------------------------------------------------------+ + * | cond | instruction dependent | + * |c c c c| | + * +-------+-------------------------------------------------------+ + */ + +#define INSN_SIZE 4 /* Always 4 bytes */ +#define INSN_COND_MASK 0xf0000000 /* Condition mask */ +#define INSN_COND_AL 0xe0000000 /* Always condition */ + +#define THUMB_INSN_SIZE 2 /* Some are 4 bytes. */ + +/* + * Defines and such for arm11 Performance Monitor Counters (p15, c15, c12, 0) + */ +#define ARM11_PMCCTL_E __BIT(0) /* enable all three counters */ +#define ARM11_PMCCTL_P __BIT(1) /* reset both Count Registers to zero */ +#define ARM11_PMCCTL_C __BIT(2) /* reset the Cycle Counter Register to zero */ +#define ARM11_PMCCTL_D __BIT(3) /* cycle count divide by 64 */ +#define ARM11_PMCCTL_EC0 __BIT(4) /* Enable Counter Register 0 interrupt */ +#define ARM11_PMCCTL_EC1 __BIT(5) /* Enable Counter Register 1 interrupt */ +#define ARM11_PMCCTL_ECC __BIT(6) /* Enable Cycle Counter interrupt */ +#define ARM11_PMCCTL_SBZa __BIT(7) /* UNP/SBZ */ +#define ARM11_PMCCTL_CR0 __BIT(8) /* Count Register 0 overflow flag */ +#define ARM11_PMCCTL_CR1 __BIT(9) /* Count Register 1 overflow flag */ +#define ARM11_PMCCTL_CCR __BIT(10) /* Cycle Count Register overflow flag */ +#define ARM11_PMCCTL_X __BIT(11) /* Enable Export of the events to the event bus */ +#define ARM11_PMCCTL_EVT1 __BITS(19,12) /* source of events for Count Register 1 */ +#define ARM11_PMCCTL_EVT0 __BITS(27,20) /* source of events for Count Register 0 */ +#define ARM11_PMCCTL_SBZb __BITS(31,28) /* UNP/SBZ */ +#define ARM11_PMCCTL_SBZ \ + (ARM11_PMCCTL_SBZa | ARM11_PMCCTL_SBZb) + +#define ARM11_PMCEVT_ICACHE_MISS 0 /* Instruction Cache Miss */ +#define ARM11_PMCEVT_ISTREAM_STALL 1 /* Instruction Stream Stall */ +#define ARM11_PMCEVT_IUTLB_MISS 2 /* Instruction uTLB Miss */ +#define ARM11_PMCEVT_DUTLB_MISS 3 /* Data uTLB Miss */ +#define ARM11_PMCEVT_BRANCH 4 /* Branch Inst. Executed */ +#define ARM11_PMCEVT_BRANCH_MISS 6 /* Branch mispredicted */ +#define ARM11_PMCEVT_INST_EXEC 7 /* Instruction Executed */ +#define ARM11_PMCEVT_DCACHE_ACCESS0 9 /* Data Cache Access */ +#define ARM11_PMCEVT_DCACHE_ACCESS1 10 /* Data Cache Access */ +#define ARM11_PMCEVT_DCACHE_MISS 11 /* Data Cache Miss */ +#define ARM11_PMCEVT_DCACHE_WRITEBACK 12 /* Data Cache Writeback */ +#define ARM11_PMCEVT_PC_CHANGE 13 /* Software PC change */ +#define ARM11_PMCEVT_TLB_MISS 15 /* Main TLB Miss */ +#define ARM11_PMCEVT_DATA_ACCESS 16 /* non-cached data access */ +#define ARM11_PMCEVT_LSU_STALL 17 /* Load/Store Unit stall */ +#define ARM11_PMCEVT_WBUF_DRAIN 18 /* Write buffer drained */ +#define ARM11_PMCEVT_ETMEXTOUT0 32 /* ETMEXTOUT[0] asserted */ +#define ARM11_PMCEVT_ETMEXTOUT1 33 /* ETMEXTOUT[1] asserted */ +#define ARM11_PMCEVT_ETMEXTOUT 34 /* ETMEXTOUT[0 & 1] */ +#define ARM11_PMCEVT_CALL_EXEC 35 /* Procedure call executed */ +#define ARM11_PMCEVT_RETURN_EXEC 36 /* Return executed */ +#define ARM11_PMCEVT_RETURN_HIT 37 /* return address predicted */ +#define ARM11_PMCEVT_RETURN_MISS 38 /* return addr. mispredicted */ +#define ARM11_PMCEVT_CYCLE 255 /* Increment each cycle */ + +/* Defines for ARM CORTEX performance counters */ +#define CORTEX_CNTENS_C __BIT(31) /* Enables the cycle counter */ +#define CORTEX_CNTENC_C __BIT(31) /* Disables the cycle counter */ +#define CORTEX_CNTOFL_C __BIT(31) /* Cycle counter overflow flag */ + +#if !defined(__ASSEMBLER__) +#define ARMREG_READ_INLINE(name, __insnstring) \ +static inline uint32_t armreg_##name##_read(void) \ +{ \ + uint32_t __rv; \ + __asm __volatile("mrc " __insnstring : "=r"(__rv)); \ + return __rv; \ +} + +#define ARMREG_WRITE_INLINE(name, __insnstring) \ +static inline void armreg_##name##_write(uint32_t __val) \ +{ \ + __asm __volatile("mcr " __insnstring :: "r"(__val)); \ +} + +/* c0 registers */ +ARMREG_READ_INLINE(midr, "p15,0,%0,c0,c0,0") /* Main ID Register */ +ARMREG_READ_INLINE(ctr, "p15,0,%0,c0,c0,1") /* Cache Type Register */ +ARMREG_READ_INLINE(mpidr, "p15,0,%0,c0,c0,5") /* Multiprocess Affinity Register */ +ARMREG_READ_INLINE(pfr0, "p15,0,%0,c0,c1,0") /* Processor Feature Register 0 */ +ARMREG_READ_INLINE(pfr1, "p15,0,%0,c0,c1,1") /* Processor Feature Register 1 */ +ARMREG_READ_INLINE(mmfr0, "p15,0,%0,c0,c1,4") /* Memory Model Feature Register 0 */ +ARMREG_READ_INLINE(mmfr1, "p15,0,%0,c0,c1,5") /* Memory Model Feature Register 1 */ +ARMREG_READ_INLINE(mmfr2, "p15,0,%0,c0,c1,6") /* Memory Model Feature Register 2 */ +ARMREG_READ_INLINE(mmfr3, "p15,0,%0,c0,c1,7") /* Memory Model Feature Register 3 */ +ARMREG_READ_INLINE(isar0, "p15,0,%0,c0,c2,0") /* Instruction Set Attribute Register 0 */ +ARMREG_READ_INLINE(isar1, "p15,0,%0,c0,c2,1") /* Instruction Set Attribute Register 1 */ +ARMREG_READ_INLINE(isar2, "p15,0,%0,c0,c2,2") /* Instruction Set Attribute Register 2 */ +ARMREG_READ_INLINE(isar3, "p15,0,%0,c0,c2,3") /* Instruction Set Attribute Register 3 */ +ARMREG_READ_INLINE(isar4, "p15,0,%0,c0,c2,4") /* Instruction Set Attribute Register 4 */ +ARMREG_READ_INLINE(isar5, "p15,0,%0,c0,c2,5") /* Instruction Set Attribute Register 5 */ +ARMREG_READ_INLINE(ccsidr, "p15,1,%0,c0,c0,0") /* Cache Size ID Register */ +ARMREG_READ_INLINE(clidr, "p15,1,%0,c0,c0,1") /* Cache Level ID Register */ +ARMREG_READ_INLINE(csselr, "p15,2,%0,c0,c0,0") /* Cache Size Selection Register */ +ARMREG_WRITE_INLINE(csselr, "p15,2,%0,c0,c0,0") /* Cache Size Selection Register */ +/* c1 registers */ +ARMREG_READ_INLINE(auxctl, "p15,0,%0,c1,c0,1") /* Auxiliary Control Register */ +ARMREG_WRITE_INLINE(auxctl, "p15,0,%0,c1,c0,1") /* Auxiliary Control Register */ +ARMREG_READ_INLINE(cpacr, "p15,0,%0,c1,c0,2") /* Co-Processor Access Control Register */ +ARMREG_WRITE_INLINE(cpacr, "p15,0,%0,c1,c0,2") /* Co-Processor Access Control Register */ +/* c2 registers */ +ARMREG_READ_INLINE(ttbr, "p15,0,%0,c2,c0,0") /* Translation Table Base Register 0 */ +ARMREG_WRITE_INLINE(ttbr, "p15,0,%0,c2,c0,0") /* Translation Table Base Register 0 */ +ARMREG_READ_INLINE(ttbr1, "p15,0,%0,c2,c0,1") /* Translation Table Base Register 1 */ +ARMREG_WRITE_INLINE(ttbr1, "p15,0,%0,c2,c0,1") /* Translation Table Base Register 1 */ +ARMREG_READ_INLINE(ttbcr, "p15,0,%0,c2,c0,2") /* Translation Table Base Register */ +ARMREG_WRITE_INLINE(ttbcr, "p15,0,%0,c2,c0,2") /* Translation Table Base Register */ +/* c5 registers */ +ARMREG_READ_INLINE(dfsr, "p15,0,%0,c5,c0,0") /* Data Fault Status Register */ +ARMREG_READ_INLINE(ifsr, "p15,0,%0,c5,c0,1") /* Instruction Fault Status Register */ +/* c6 registers */ +ARMREG_READ_INLINE(dfar, "p15,0,%0,c6,c0,0") /* Data Fault Address Register */ +ARMREG_READ_INLINE(ifar, "p15,0,%0,c6,c0,2") /* Instruction Fault Address Register */ +/* c7 registers */ +ARMREG_WRITE_INLINE(icialluis, "p15,0,%0,c7,c1,0") /* Instruction Inv All (IS) */ +ARMREG_WRITE_INLINE(bpiallis, "p15,0,%0,c7,c1,6") /* Branch Invalidate All (IS) */ +ARMREG_READ_INLINE(par, "p15,0,%0,c7,c4,0") /* Physical Address Register */ +ARMREG_WRITE_INLINE(iciallu, "p15,0,%0,c7,c5,0") /* Instruction Invalidate All */ +ARMREG_WRITE_INLINE(icimvau, "p15,0,%0,c7,c5,1") /* Instruction Invalidate MVA */ +ARMREG_WRITE_INLINE(isb, "p15,0,%0,c7,c5,4") /* Instruction Synchronization Barrier */ +ARMREG_WRITE_INLINE(bpiall, "p15,0,%0,c5,c1,6") /* Breakpoint Invalidate All */ +ARMREG_WRITE_INLINE(dcimvac, "p15,0,%0,c7,c6,1") /* Data Invalidate MVA to PoC */ +ARMREG_WRITE_INLINE(dcisw, "p15,0,%0,c7,c6,2") /* Data Invalidate Set/Way */ +ARMREG_WRITE_INLINE(ats1cpr, "p15,0,%0,c7,c8,0") /* AddrTrans CurState PL1 Read */ +ARMREG_WRITE_INLINE(dccmvac, "p15,0,%0,c7,c10,1") /* Data Clean MVA to PoC */ +ARMREG_WRITE_INLINE(dccsw, "p15,0,%0,c7,c10,2") /* Data Clean Set/Way */ +ARMREG_WRITE_INLINE(dsb, "p15,0,%0,c7,c10,4") /* Data Synchronization Barrier */ +ARMREG_WRITE_INLINE(dmb, "p15,0,%0,c7,c10,5") /* Data Memory Barrier */ +ARMREG_WRITE_INLINE(dccmvau, "p15,0,%0,c7,c14,1") /* Data Clean MVA to PoU */ +ARMREG_WRITE_INLINE(dccimvac, "p15,0,%0,c7,c14,1") /* Data Clean&Inv MVA to PoC */ +ARMREG_WRITE_INLINE(dccisw, "p15,0,%0,c7,c14,2") /* Data Clean&Inv Set/Way */ +/* c9 registers */ +ARMREG_READ_INLINE(pmcr, "p15,0,%0,c9,c12,0") /* PMC Control Register */ +ARMREG_WRITE_INLINE(pmcr, "p15,0,%0,c9,c12,0") /* PMC Control Register */ +ARMREG_READ_INLINE(pmcntenset, "p15,0,%0,c9,c12,1") /* PMC Count Enable Set */ +ARMREG_WRITE_INLINE(pmcntenset, "p15,0,%0,c9,c12,1") /* PMC Count Enable Set */ +ARMREG_READ_INLINE(pmcntenclr, "p15,0,%0,c9,c12,2") /* PMC Count Enable Clear */ +ARMREG_WRITE_INLINE(pmcntenclr, "p15,0,%0,c9,c12,2") /* PMC Count Enable Clear */ +ARMREG_READ_INLINE(pmovsr, "p15,0,%0,c9,c12,3") /* PMC Overflow Flag Status */ +ARMREG_WRITE_INLINE(pmovsr, "p15,0,%0,c9,c12,3") /* PMC Overflow Flag Status */ +ARMREG_READ_INLINE(pmccntr, "p15,0,%0,c9,c13,0") /* PMC Cycle Counter */ +ARMREG_WRITE_INLINE(pmccntr, "p15,0,%0,c9,c13,0") /* PMC Cycle Counter */ +/* c13 registers */ +ARMREG_READ_INLINE(tpidrprw, "p15,0,%0,c13,c0,4") /* PL1 only Thread ID Register */ +ARMREG_WRITE_INLINE(tpidrprw, "p15,0,%0,c13,c0,4") /* PL1 only Thread ID Register */ +ARMREG_READ_INLINE(cbar, "p15,4,%0,c15,c0,0") /* Configuration Base Address Register */ +/* c13 registers */ +ARMREG_READ_INLINE(pmcrv6, "p15,0,%0,c15,c12,0") /* PMC Control Register (armv6) */ +ARMREG_WRITE_INLINE(pmcrv6, "p15,0,%0,c15,c12,0") /* PMC Control Register (armv6) */ +ARMREG_READ_INLINE(pmccntrv6, "p15,0,%0,c15,c12,1") /* PMC Cycle Counter (armv6) */ +ARMREG_WRITE_INLINE(pmccntrv6, "p15,0,%0,c15,c12,1") /* PMC Cycle Counter (armv6) */ + +#endif /* !__ASSEMBLER__ */ + + +#define MPIDR_31 0x80000000 +#define MPIDR_U 0x40000000 // 1 = Uniprocessor +#define MPIDR_MT 0x01000000 // AFF0 for SMT +#define MPIDR_AFF2 0x00ff0000 +#define MPIDR_AFF1 0x0000ff00 +#define MPIDR_AFF0 0x000000ff + +#endif /* _ARM_ARMREG_H */ diff --git a/sys/arch/arm/include/cpu.h b/sys/arch/arm/include/cpu.h index 0ae101984..56f8ec33a 100644 --- a/sys/arch/arm/include/cpu.h +++ b/sys/arch/arm/include/cpu.h @@ -1,55 +1,417 @@ +/* cpu.h,v 1.45.4.7 2008/01/28 18:20:39 matt Exp */ + +/* + * Copyright (c) 1994-1996 Mark Brinicombe. + * Copyright (c) 1994 Brini. + * All rights reserved. + * + * This code is derived from software written for Brini by Mark Brinicombe + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions + * are met: + * 1. Redistributions of source code must retain the above copyright + * notice, this list of conditions and the following disclaimer. + * 2. Redistributions in binary form must reproduce the above copyright + * notice, this list of conditions and the following disclaimer in the + * documentation and/or other materials provided with the distribution. + * 3. All advertising materials mentioning features or use of this software + * must display the following acknowledgement: + * This product includes software developed by Brini. + * 4. The name of the company nor the name of the author may be used to + * endorse or promote products derived from this software without specific + * prior written permission. + * + * THIS SOFTWARE IS PROVIDED BY BRINI ``AS IS'' AND ANY EXPRESS OR IMPLIED + * WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF + * MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. + * IN NO EVENT SHALL BRINI OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, + * INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES + * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR + * SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) + * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT + * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY + * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF + * SUCH DAMAGE. + * + * RiscBSD kernel project + * + * cpu.h + * + * CPU specific symbols + * + * Created : 18/09/94 + * + * Based on kate/katelib/arm6.h + */ + #ifndef _ARM_CPU_H_ #define _ARM_CPU_H_ +/* + * User-visible definitions + */ + +/* CTL_MACHDEP definitions. */ +#define CPU_DEBUG 1 /* int: misc kernel debug control */ +#define CPU_BOOTED_DEVICE 2 /* string: device we booted from */ +#define CPU_BOOTED_KERNEL 3 /* string: kernel we booted */ +#define CPU_CONSDEV 4 /* struct: dev_t of our console */ +#define CPU_POWERSAVE 5 /* int: use CPU powersave mode */ +#define CPU_MAXID 6 /* number of valid machdep ids */ + +#if defined(_KERNEL) || defined(_KMEMUSER) + +/* + * Kernel-only definitions + */ + +#if !defined(_LKM) && defined(_KERNEL_OPT) +#include "opt_multiprocessor.h" +#include "opt_cpuoptions.h" +#include "opt_lockdebug.h" +#include "opt_cputypes.h" +#endif /* !_LKM && _KERNEL_OPT */ + +#include + +#ifndef _LOCORE +#include +#endif /* !_LOCORE */ + +#include + + +#ifndef _LOCORE +/* 1 == use cpu_sleep(), 0 == don't */ +extern int cpu_do_powersave; +#endif + +#ifdef _LOCORE + +#if defined(_ARM_ARCH_6) +#define IRQdisable cpsid i +#define IRQenable cpsie i +#elif defined(__PROG32) +#define IRQdisable \ + stmfd sp!, {r0} ; \ + mrs r0, cpsr ; \ + orr r0, r0, #(I32_bit) ; \ + msr cpsr_c, r0 ; \ + ldmfd sp!, {r0} + +#define IRQenable \ + stmfd sp!, {r0} ; \ + mrs r0, cpsr ; \ + bic r0, r0, #(I32_bit) ; \ + msr cpsr_c, r0 ; \ + ldmfd sp!, {r0} +#else +/* Not yet used in 26-bit code */ +#endif + +#if defined (TPIDRPRW_IS_CURCPU) +#define GET_CURCPU(rX) mrc p15, 0, rX, c13, c0, 4 +#define GET_CURLWP(rX) GET_CURCPU(rX); ldr rX, [rX, #CI_CURLWP] +#elif defined (TPIDRPRW_IS_CURLWP) +#define GET_CURLWP(rX) mrc p15, 0, rX, c13, c0, 4 +#define GET_CURCPU(rX) GET_CURLWP(rX); ldr rX, [rX, #L_CPU] +#elif !defined(MULTIPROCESSOR) +#define GET_CURCPU(rX) ldr rX, =_C_LABEL(cpu_info_store) +#define GET_CURLWP(rX) GET_CURCPU(rX); ldr rX, [rX, #CI_CURLWP] +#endif +#define GET_CURPCB(rX) GET_CURLWP(rX); ldr rX, [rX, #L_PCB] + +#else /* !_LOCORE */ + +#ifdef __PROG32 +#define IRQdisable __set_cpsr_c(I32_bit, I32_bit); +#define IRQenable __set_cpsr_c(I32_bit, 0); +#else +#define IRQdisable set_r15(R15_IRQ_DISABLE, R15_IRQ_DISABLE); +#define IRQenable set_r15(R15_IRQ_DISABLE, 0); +#endif + +#endif /* !_LOCORE */ + +#ifndef _LOCORE + +/* All the CLKF_* macros take a struct clockframe * as an argument. */ + +/* + * CLKF_USERMODE: Return TRUE/FALSE (1/0) depending on whether the + * frame came from USR mode or not. + */ +#ifdef __PROG32 +#define CLKF_USERMODE(frame) ((frame->cf_tf.tf_spsr & PSR_MODE) == PSR_USR32_MODE) +#else +#define CLKF_USERMODE(frame) ((frame->cf_if.if_r15 & R15_MODE) == R15_MODE_USR) +#endif + +/* + * CLKF_INTR: True if we took the interrupt from inside another + * interrupt handler. + */ +#ifdef __PROG32 +/* Hack to treat FPE time as interrupt time so we can measure it */ +#define CLKF_INTR(frame) \ + ((curcpu()->ci_intr_depth > 1) || \ + (frame->cf_tf.tf_spsr & PSR_MODE) == PSR_UND32_MODE) +#else +#define CLKF_INTR(frame) (curcpu()->ci_intr_depth > 1) +#endif + +/* + * CLKF_PC: Extract the program counter from a clockframe + */ +#ifdef __PROG32 +#define CLKF_PC(frame) (frame->cf_tf.tf_pc) +#else +#define CLKF_PC(frame) (frame->cf_if.if_r15 & R15_PC) +#endif + +/* + * LWP_PC: Find out the program counter for the given lwp. + */ +#ifdef __PROG32 +#define LWP_PC(l) (lwp_trapframe(l)->tf_pc) +#else +#define LWP_PC(l) (lwp_trapframe(l)->tf_r15 & R15_PC) +#endif + +/* + * Validate a PC or PSR for a user process. Used by various system calls + * that take a context passed by the user and restore it. + */ + +#ifdef __PROG32 +#define VALID_R15_PSR(r15,psr) \ + (((psr) & PSR_MODE) == PSR_USR32_MODE && \ + ((psr) & (I32_bit | F32_bit)) == 0) +#else +#define VALID_R15_PSR(r15,psr) \ + (((r15) & R15_MODE) == R15_MODE_USR && \ + ((r15) & (R15_IRQ_DISABLE | R15_FIQ_DISABLE)) == 0) +#endif + + + +/* The address of the vector page. */ +extern vaddr_t vector_page; +#ifdef __PROG32 +void arm32_vector_init(vaddr_t, int); + +#define ARM_VEC_RESET (1 << 0) +#define ARM_VEC_UNDEFINED (1 << 1) +#define ARM_VEC_SWI (1 << 2) +#define ARM_VEC_PREFETCH_ABORT (1 << 3) +#define ARM_VEC_DATA_ABORT (1 << 4) +#define ARM_VEC_ADDRESS_EXCEPTION (1 << 5) +#define ARM_VEC_IRQ (1 << 6) +#define ARM_VEC_FIQ (1 << 7) + +#define ARM_NVEC 8 +#define ARM_VEC_ALL 0xffffffff +#endif + +/* + * Per-CPU information. For now we assume one CPU. + */ +static inline int curcpl(void); +static inline void set_curcpl(int); +static inline void cpu_dosoftints(void); + +#include +#include +#include + +struct cpu_info { + struct cpu_data ci_data; /* MI per-cpu data */ + device_t ci_dev; /* Device corresponding to this CPU */ + cpuid_t ci_cpuid; + uint32_t ci_arm_cpuid; /* aggregate CPU id */ + uint32_t ci_arm_cputype; /* CPU type */ + uint32_t ci_arm_cpurev; /* CPU revision */ + uint32_t ci_ctrl; /* The CPU control register */ + int ci_cpl; /* current processor level (spl) */ + int ci_astpending; /* */ + int ci_want_resched; /* resched() was called */ + int ci_intr_depth; /* */ + struct cpu_softc *ci_softc; /* platform softc */ +#ifdef __HAVE_FAST_SOFTINTS + lwp_t *ci_softlwps[SOFTINT_COUNT]; + volatile uint32_t ci_softints; +#endif + lwp_t *ci_curlwp; /* current lwp */ + struct evcnt ci_arm700bugcount; + int32_t ci_mtx_count; + int ci_mtx_oldspl; + register_t ci_undefsave[3]; + uint32_t ci_vfp_id; +#if defined(_ARM_ARCH_7) + uint64_t ci_lastintr; +#endif + struct evcnt ci_abt_evs[FAULT_TYPE_MASK+1]; +#if defined(MP_CPU_INFO_MEMBERS) + MP_CPU_INFO_MEMBERS +#endif +}; + +extern struct cpu_info cpu_info_store; +#if defined(TPIDRPRW_IS_CURLWP) +static inline struct lwp * +_curlwp(void) +{ + return (struct lwp *) armreg_tpidrprw_read(); +} + +static inline void +_curlwp_set(struct lwp *l) +{ + armreg_tpidrprw_write((uintptr_t)l); +} + +#define curlwp (_curlwp()) +static inline struct cpu_info * +curcpu(void) +{ + return curlwp->l_cpu; +} +#elif defined(TPIDRPRW_IS_CURCPU) +static inline struct cpu_info * +curcpu(void) +{ + return (struct cpu_info *) armreg_tpidrprw_read(); +} +#elif !defined(MULTIPROCESSOR) +#define curcpu() (&cpu_info_store) +#else +#error MULTIPROCESSOR requires TPIDRPRW_IS_CURLWP or TPIDRPRW_IS_CURCPU +#endif /* !TPIDRPRW_IS_CURCPU && !TPIDRPRW_IS_CURLWP */ + +#ifndef curlwp +#define curlwp (curcpu()->ci_curlwp) +#endif + +#define CPU_INFO_ITERATOR int +#if defined(MULTIPROCESSOR) +extern struct cpu_info *cpu_info[]; +#define cpu_number() (curcpu()->ci_cpuid) +void cpu_boot_secondary_processors(void); +#define CPU_IS_PRIMARY(ci) ((ci)->ci_cpuid == 0) +#define CPU_INFO_FOREACH(cii, ci) \ + cii = 0, ci = cpu_info[0]; cii < ncpu && (ci = cpu_info[cii]) != NULL; cii++ +#else +#define cpu_number() 0 + +#define CPU_IS_PRIMARY(ci) true +#define CPU_INFO_FOREACH(cii, ci) \ + cii = 0, ci = curcpu(); ci != NULL; ci = NULL +#endif + +#define LWP0_CPU_INFO (&cpu_info_store) + +static inline int +curcpl(void) +{ + return curcpu()->ci_cpl; +} + +static inline void +set_curcpl(int pri) +{ + curcpu()->ci_cpl = pri; +} + +static inline void +cpu_dosoftints(void) +{ +#ifdef __HAVE_FAST_SOFTINTS + void dosoftints(void); +#ifndef __HAVE_PIC_FAST_SOFTINTS + struct cpu_info * const ci = curcpu(); + if (ci->ci_intr_depth == 0 && (ci->ci_softints >> ci->ci_cpl) > 0) + dosoftints(); +#endif +#endif +} + +#ifdef __PROG32 +void cpu_proc_fork(struct proc *, struct proc *); +#else +#define cpu_proc_fork(p1, p2) +#endif + +/* + * Scheduling glue + */ + +#define setsoftast() (curcpu()->ci_astpending = 1) + +/* + * Notify the current process (p) that it has a signal pending, + * process as soon as possible. + */ + +#define cpu_signotify(l) setsoftast() + +/* + * Give a profiling tick to the current process when the user profiling + * buffer pages are invalid. On the i386, request an ast to send us + * through trap(), marking the proc as needing a profiling tick. + */ +#define cpu_need_proftick(l) ((l)->l_pflag |= LP_OWEUPC, setsoftast()) + +/* + * We've already preallocated the stack for the idlelwps for additional CPUs. + * This hook allows to return them. + */ +vaddr_t cpu_uarea_alloc_idlelwp(struct cpu_info *); + +#ifndef acorn26 +/* + * cpu device glue (belongs in cpuvar.h) + */ +void cpu_attach(device_t, cpuid_t); +#endif + +/* + * Random cruft + */ + +struct lwp; + +/* locore.S */ +void atomic_set_bit(u_int *, u_int); +void atomic_clear_bit(u_int *, u_int); + +/* cpuswitch.S */ +struct pcb; +void savectx(struct pcb *); + +/* ast.c */ +void userret(register struct lwp *); + +/* *_machdep.c */ +void bootsync(void); + +/* fault.c */ +int badaddr_read(void *, size_t, void *); + +/* syscall.c */ +void swi_handler(trapframe_t *); + +/* arm_machdep.c */ +void ucas_ras_check(trapframe_t *); + +/* vfp_init.c */ +void vfp_attach(void); +void vfp_discardcontext(void); +void vfp_savecontext(void); +extern const pcu_ops_t arm_vfp_ops; + +#endif /* !_LOCORE */ + +#endif /* _KERNEL */ -/* xPSR - Program Status Registers */ -#define PSR_T (1 << 5) /* Thumb execution state bit */ -#define PSR_F (1 << 6) /* FIQ mask bit */ -#define PSR_I (1 << 7) /* IRQ mask bit */ -#define PSR_A (1 << 8) /* Asynchronous abort mask bit */ -#define PSR_E (1 << 9) /* Endianness execution state bit */ -#define PSR_J (1 << 24) /* Jazelle bit */ -#define PSR_Q (1 << 27) /* Cumulative saturation bit */ -#define PSR_V (1 << 28) /* Overflow condition flag */ -#define PSR_C (1 << 29) /* Carry condition flag */ -#define PSR_Z (1 << 30) /* Zero condition flag */ -#define PSR_N (1 << 31) /* Negative condition flag */ - -#define PSR_MODE_MASK 0x0000001F /* Mode field mask */ - -#define MODE_USR 0x10 /* User mode */ -#define MODE_FIQ 0x11 /* FIQ mode */ -#define MODE_IRQ 0x12 /* IRQ mode */ -#define MODE_SVC 0x13 /* Supervisor mode */ -#define MODE_MON 0x16 /* Monitor mode */ -#define MODE_ABT 0x17 /* Abort mode */ -#define MODE_HYP 0x1A /* Hyp mode */ -#define MODE_UND 0x1B /* Undefined mode */ -#define MODE_SYS 0x1F /* System mode */ - -/* SCTLR - System Control Register */ -#define SCTLR_M (1 << 0) /* MMU enable */ -#define SCTLR_A (1 << 1) /* Alignment check enable */ -#define SCTLR_C (1 << 2) /* Data and Unified Cache enable */ -#define SCTLR_CP15BEN (1 << 5) /* CP15 barrier enable */ -#define SCTLR_SW (1 << 10) /* SWP and SWPB enable */ -#define SCTLR_Z (1 << 11) /* Branch prediction enable */ -#define SCTLR_I (1 << 12) /* Instruction cache enable */ -#define SCTLR_V (1 << 13) /* (High) Vectors bit */ -#define SCTLR_RR (1 << 14) /* Round Robin (cache) select */ -#define SCTLR_HA (1 << 17) /* Hardware Access flag enable */ -#define SCTLR_FI (1 << 21) /* Fast interrupts configuration enable */ -#define SCTLR_VE (1 << 24) /* Interrupt Vectors Enable */ -#define SCTLR_EE (1 << 25) /* Exception Endianness */ -#define SCTLR_NMFI (1 << 27) /* Non-maskable FIQ (NMFI) support */ -#define SCTLR_TRE (1 << 28) /* TEX remap enable */ -#define SCTLR_AFE (1 << 29) /* Access flag enable */ -#define SCTLR_TE (1 << 30) /* Thumb Exception enable */ - -/* ACTLR - Auxiliary Control Register */ -#define A8_ACTLR_L1ALIAS (1 << 0) /* L1 Dcache hw alias check enable */ -#define A8_ACTLR_L2EN (1 << 1) /* L2 cache enable */ -#define A8_ACTLR_L1RSTDIS (1 << 30) /* L1 hw reset disable */ -#define A8_ACTLR_L2RSTDIS (1 << 31) /* L2 hw reset disable */ - -#endif /* _ARM_CPU_H_ */ +#endif /* !_ARM_CPU_H_ */ -- 2.44.0