/* * tegra 2 SoC machine assist * dual arm cortex-a9 processors * * ARM v7 arch. ref. man. §B1.3.3 says that we don't need barriers * around writes to CPSR. * * LDREX/STREX use an exclusive monitor, which is part of the data cache unit * for the L1 cache, so they won't work right if the L1 cache is disabled. */ #include "arm.s" #define LDREX(fp,t) WORD $(0xe<<28|0x01900f9f | (fp)<<16 | (t)<<12) /* `The order of operands is from left to right in dataflow order' - asm man */ #define STREX(f,tp,r) WORD $(0xe<<28|0x01800f90 | (tp)<<16 | (r)<<12 | (f)<<0) #define MAXMB (KiB-1) /* last MB has vectors */ #define TMPSTACK (DRAMSIZE - 64*MiB) /* used only during cpu startup */ /* tas/cas strex debugging limits; started at 10000 */ #define MAXSC 100000 GLOBL testmem(SB), $4 /* * Entered here from Das U-Boot or another Plan 9 kernel with MMU disabled. * Until the MMU is enabled it is OK to call functions provided * they are within ±32MiB relative and do not require any * local variables or more than one argument (i.e. there is * no stack). */ TEXT _start(SB), 1, $-4 CPSMODE(PsrMsvc) CPSID /* interrupts off */ CPSAE SETEND(0) /* little-endian */ BARRIERS CLREX SETZSB MOVW CPSR, R0 ORR $PsrDfiq, R0 MOVW R0, CPSR /* invalidate i-cache and branch-target cache */ MTCP P(CpSC), 0, PC, C(CpCACHE), C(CpCACHEinvi), CpCACHEall BARRIERS /* put cpus other than 0 to sleep until cpu 0 is ready */ CPUID(R1) BEQ cpuinit /* not cpu 0 */ PUTC('Z') PUTC('Z') BARRIERS dowfi: WFI MOVW cpus_proceed(SB), R1 CMP $0, R1 BEQ dowfi BL cpureset(SB) B dowfi cpuinit: DELAY(printloopret, 1) PUTC('\r') DELAY(printloopnl, 1) PUTC('\n') DELAY(printloops, 1) PUTC('P') /* disable the PL310 L2 cache on cpu0 */ MOVW $(PHYSL2BAG+0x100), R1 MOVW $0, R2 MOVW R2, (R1) BARRIERS /* invalidate it */ MOVW $((1<<16)-1), R2 MOVW R2, 0x77c(R1) BARRIERS /* * disable my MMU & caches */ MFCP P(CpSC), 0, R1, C(CpCONTROL), C(0), CpMainctl ORR $CpCsbo, R1 BIC $(CpCsbz|CpCmmu|CpCdcache|CpCicache|CpCpredict), R1 MTCP P(CpSC), 0, R1, C(CpCONTROL), C(0), CpMainctl BARRIERS /* cortex-a9 model-specific initial configuration */ MOVW $0, R1 MTCP P(CpSC), 0, R1, C(CpCONTROL), C(0), CpAuxctl BARRIERS PUTC('l') DELAY(printloop3, 1) MOVW $testmem-KZERO(SB), R0 BL memdiag(SB) PUTC('a') /* clear Mach for cpu 0 */ MOVW $PADDR(MACHADDR), R4 /* address of Mach for cpu 0 */ MOVW $0, R0 _machZ: MOVW R0, (R4) ADD $4, R4 CMP.S $PADDR(L1+L1X(0)), R4 /* end at top-level page table */ BNE _machZ /* * set up the MMU page table for cpu 0 */ PUTC('n') /* clear all PTEs first, to provide a default */ // MOVW $PADDR(L1+L1X(0)), R4 /* address of PTE for 0 */ _ptenv0: ZEROPTE() CMP.S $PADDR(L1+16*KiB), R4 BNE _ptenv0 DELAY(printloop4, 2) PUTC(' ') /* * set up double map of PHYSDRAM, KZERO to PHYSDRAM for first few MBs, * but only if KZERO and PHYSDRAM differ. */ MOVW $PTEDRAM, R2 /* PTE bits */ MOVW $PHYSDRAM, R3 /* pa */ CMP $KZERO, R3 BEQ no2map MOVW $PADDR(L1+L1X(PHYSDRAM)), R4 /* address of PTE for PHYSDRAM */ MOVW $DOUBLEMAPMBS, R5 _ptdbl: FILLPTE() SUB.S $1, R5 BNE _ptdbl no2map: /* * back up and fill in PTEs for memory at KZERO. * trimslice has 1 bank of 1GB at PHYSDRAM. * Map the maximum. */ PUTC('9') MOVW $PTEDRAM, R2 /* PTE bits */ MOVW $PHYSDRAM, R3 MOVW $PADDR(L1+L1X(KZERO)), R4 /* start with PTE for KZERO */ MOVW $MAXMB, R5 /* inner loop count (MBs) */ _ptekrw: /* set PTEs */ FILLPTE() SUB.S $1, R5 /* decrement inner loop count */ BNE _ptekrw /* * back up and fill in PTEs for MMIO */ PUTC(' ') MOVW $PTEIO, R2 /* PTE bits */ MOVW $PHYSIO, R3 MOVW $PADDR(L1+L1X(VIRTIO)), R4 /* start with PTE for VIRTIO */ _ptenv2: FILLPTE() CMP.S $PADDR(L1+L1X(PHYSIOEND)), R4 BNE _ptenv2 /* mmu.c sets up the trap vectors later */ MOVW $(PHYSDRAM | TMPSTACK), SP /* * learn l1 cache characteristics (on cpu 0 only). */ MOVW $(1-1), R0 /* l1 */ SLL $1, R0 /* R0 = (cache - 1) << 1 */ MTCP P(CpSC), CpIDcssel, R0, C(CpID), C(CpIDid), 0 /* select l1 cache */ BARRIERS MFCP P(CpSC), CpIDcsize, R0, C(CpID), C(CpIDid), 0 /* get sets & ways */ MOVW $CACHECONF, R8 /* get log2linelen into l1setsh */ MOVW R0, R1 AND $3, R1 ADD $4, R1 /* l1 & l2 must have same cache line size, thus same set shift */ MOVW R1, 4(R8) /* +4 = l1setsh */ MOVW R1, 12(R8) /* +12 = l2setsh */ /* get nways in R1 */ SRA $3, R0, R1 AND $((1<<10)-1), R1 ADD $1, R1 /* get log2(nways) in R2 (assume nways is 2^n) */ MOVW $(BI2BY*BY2WD - 1), R2 CLZ(1, 1) SUB.S R1, R2 /* R2 = 31 - clz(nways) */ ADD.EQ $1, R2 // MOVW R2, R3 /* print log2(nways): 2 */ MOVW $32, R1 SUB R2, R1 /* R1 = 32 - log2(nways) */ MOVW R1, 0(R8) /* +0 = l1waysh */ BARRIERS MOVW $testmem-KZERO(SB), R0 BL memdiag(SB) /* * the mpcore manual says invalidate d-cache, scu, pl310 in that order, * but says nothing about when to disable them. * * invalidate my caches before enabling */ BL cachedinv(SB) MTCP P(CpSC), 0, PC, C(CpCACHE), C(CpCACHEinvi), CpCACHEall BARRIERS PUTC('f') /* * the mpcore manual says enable scu, d-cache, pl310, smp mode * in that order. we have to reverse the last two; see main(). */ BL scuon(SB) /* * turn my L1 cache on; need it for tas below. */ MFCP P(CpSC), 0, R1, C(CpCONTROL), C(0), CpMainctl ORR $(CpCdcache|CpCicache|CpCalign|CpCpredict), R1 MTCP P(CpSC), 0, R1, C(CpCONTROL), C(0), CpMainctl BARRIERS /* cortex-a9 model-specific configuration */ MOVW $CpACl1pref, R1 MTCP P(CpSC), 0, R1, C(CpCONTROL), C(0), CpAuxctl BARRIERS /* we're supposed to wait until l1 & l2 are on before calling smpon */ PUTC('r') /* set the domain access control */ MOVW $Client, R0 BL dacput(SB) DELAY(printloop5, 2) PUTC('o') BL mmuinvalidate(SB) MOVW $0, R0 BL pidput(SB) /* set the translation table base */ MOVW $PADDR(L1), R0 BL ttbput(SB) PUTC('m') /* * the little dance to turn the MMU on */ BL cacheuwbinv(SB) BL mmuinvalidate(SB) BL mmuenable(SB) PUTC(' ') /* warp the PC into the virtual map */ MOVW $KZERO, R0 BL _r15warp(SB) /* * cpu 0 is now running at KZERO+something! */ BARRIERS MOVW $setR12(SB), R12 /* reload kernel SB */ MOVW $(KZERO | TMPSTACK), SP BL cacheuwbinv(SB) PUTC('B') MOVW $PHYSDRAM, R3 /* pa */ CMP $KZERO, R3 BEQ no2unmap /* undo double map of PHYSDRAM, KZERO & first few MBs */ MOVW $(L1+L1X(PHYSDRAM)), R4 /* addr. of PTE for PHYSDRAM */ MOVW $0, R0 MOVW $DOUBLEMAPMBS, R5 _ptudbl: ZEROPTE() SUB.S $1, R5 BNE _ptudbl no2unmap: BL cachedwb(SB) BL mmuinvalidate(SB) /* * call main in C * pass Mach to main and set up the stack in it */ MOVW $MACHADDR, R0 /* cpu 0 Mach */ MOVW R0, R(MACH) /* m = MACHADDR */ ADD $(MACHSIZE-4), R0, SP /* leave space for link register */ PUTC('e') BL main(SB) /* main(m) */ limbo: BL idlehands(SB) B limbo BL _div(SB) /* hack to load _div, etc. */ /* * called on cpu(s) other than 0, to start them, from _vrst * (reset vector) in lexception.s, with interrupts disabled * and in SVC mode, running in the zero segment (pc is in lower 256MB). * SB is set for the zero segment. */ TEXT cpureset(SB), 1, $-4 CLREX MOVW CPSR, R0 ORR $PsrDfiq, R0 MOVW R0, CPSR MOVW $(PHYSDRAM | TMPSTACK), SP /* stack for cache ops */ /* paranoia: turn my mmu and caches off. */ MFCP P(CpSC), 0, R0, C(CpCONTROL), C(0), CpMainctl ORR $CpCsbo, R0 BIC $(CpCsbz|CpCmmu|CpCdcache|CpCicache|CpCpredict), R0 MTCP P(CpSC), 0, R0, C(CpCONTROL), C(0), CpMainctl BARRIERS /* cortex-a9 model-specific initial configuration */ MOVW $0, R1 MTCP P(CpSC), 0, R1, C(CpCONTROL), C(0), CpAuxctl ISB /* invalidate my caches before enabling */ BL cachedinv(SB) MTCP P(CpSC), 0, PC, C(CpCACHE), C(CpCACHEinvi), CpCACHEall BARRIERS /* * turn my L1 cache on; need it (and mmu) for tas below. * need branch prediction to make delay() timing right. */ MFCP P(CpSC), 0, R0, C(CpCONTROL), C(0), CpMainctl ORR $(CpCdcache|CpCicache|CpCalign|CpCpredict), R0 MTCP P(CpSC), 0, R0, C(CpCONTROL), C(0), CpMainctl BARRIERS /* enable l1 caches coherency, at minimum for ldrex/strex. */ BL smpon(SB) BARRIERS /* * we used to write to PHYSEVP here; now we do it in C, which offers * more assurance that we're up and won't go off the rails. */ /* set the domain access control */ MOVW $Client, R0 BL dacput(SB) BL setmach(SB) /* * redo double map of PHYSDRAM, KZERO in this cpu's ptes. * mmuinit will undo this later. */ MOVW $PHYSDRAM, R3 CMP $KZERO, R3 BEQ noun2map /* launchinit set m->mmul1 to a copy of cpu0's l1 page table */ MOVW 12(R(MACH)), R0 /* m->mmul1 (virtual addr) */ BL k2paddr(SB) /* R0 = PADDR(m->mmul1) */ ADD $L1X(PHYSDRAM), R0, R4 /* R4 = address of PHYSDRAM's PTE */ MOVW $PTEDRAM, R2 /* PTE bits */ MOVW $DOUBLEMAPMBS, R5 _ptrdbl: ORR R3, R2, R1 /* first identity-map 0 to 0, etc. */ MOVW R1, (R4) ADD $4, R4 /* bump PTE address */ ADD $MiB, R3 /* bump pa */ SUB.S $1, R5 BNE _ptrdbl noun2map: MOVW $0, R0 BL pidput(SB) /* set the translation table base to PADDR(m->mmul1) */ MOVW 12(R(MACH)), R0 /* m->mmul1 */ BL k2paddr(SB) /* R0 = PADDR(m->mmul1) */ BL ttbput(SB) /* * the little dance to turn the MMU on */ BL cacheuwbinv(SB) BL mmuinvalidate(SB) BL mmuenable(SB) /* * mmu is now on, with l1 pt at m->mmul1. */ /* warp the PC into the virtual map */ MOVW $KZERO, R0 BL _r15warp(SB) /* * now running at KZERO+something! */ BARRIERS MOVW $setR12(SB), R12 /* reload kernel's SB */ MOVW $(KZERO | TMPSTACK), SP /* stack for cache ops*/ BL setmach(SB) ADD $(MACHSIZE-4), R(MACH), SP /* leave space for link register */ BL cpustart(SB) /* * converts virtual address in R0 to a physical address. */ TEXT k2paddr(SB), 1, $-4 BIC $KSEGM, R0 ADD $PHYSDRAM, R0 RET /* * converts physical address in R0 to a virtual address. */ TEXT p2kaddr(SB), 1, $-4 BIC $KSEGM, R0 ORR $KZERO, R0 RET /* * converts address in R0 to the current segment, as defined by the PC. * clobbers R1. */ TEXT addr2pcseg(SB), 1, $-4 BIC $KSEGM, R0 MOVW PC, R1 AND $KSEGM, R1 /* segment PC is in */ ORR R1, R0 RET /* sets R(MACH), preserves other registers */ TEXT setmach(SB), 1, $-4 MOVM.DB.W [R14], (R13) MOVM.DB.W [R0-R2], (R13) CPUID(R2) SLL $2, R2 /* convert to word index */ MOVW $machaddr(SB), R0 BL addr2pcseg(SB) ADD R2, R0 /* R0 = &machaddr[cpuid] */ MOVW (R0), R0 /* R0 = machaddr[cpuid] */ CMP $0, R0 MOVW.EQ $MACHADDR, R0 /* paranoia: use MACHADDR if 0 */ BL addr2pcseg(SB) MOVW R0, R(MACH) /* m = machaddr[cpuid] */ MOVM.IA.W (R13), [R0-R2] MOVM.IA.W (R13), [R14] RET /* * memory diagnostic * tests word at (R0); modifies R7 and R8 */ TEXT memdiag(SB), 1, $-4 MOVW $0xabcdef89, R7 MOVW R7, (R0) MOVW (R0), R8 CMP R7, R8 BNE mbuggery /* broken memory */ BARRIERS MOVW (R0), R8 CMP R7, R8 BNE mbuggery /* broken memory */ MOVW $0, R7 MOVW R7, (R0) BARRIERS RET /* modifies R0, R3—R6 */ TEXT printhex(SB), 1, $-4 MOVW R0, R3 PUTC('0') PUTC('x') MOVW $(32-4), R5 /* bits to shift right */ nextdig: SRA R5, R3, R4 AND $0xf, R4 ADD $'0', R4 CMP.S $'9', R4 BLE nothex /* if R4 <= 9, jump */ ADD $('a'-('9'+1)), R4 nothex: PUTC(R4) SUB.S $4, R5 BGE nextdig PUTC('\r') PUTC('\n') DELAY(proct, 50) RET mbuggery: PUTC('?') PUTC('m') mtopanic: MOVW $membmsg(SB), R0 MOVW R14, R1 /* get R14's segment ... */ AND $KSEGM, R1 BIC $KSEGM, R0 /* strip segment from address */ ORR R1, R0 /* combine them */ BL panic(SB) mbugloop: WFI B mbugloop DATA membmsg+0(SB)/8,$"memory b" DATA membmsg+8(SB)/6,$"roken\z" GLOBL membmsg(SB), $14 TEXT _r15warp(SB), 1, $-4 BIC $KSEGM, R14 /* link reg, will become PC */ ORR R0, R14 BIC $KSEGM, SP ORR R0, SP RET /* * `single-element' cache operations. * in arm arch v7, they operate on all architected cache levels, so separate * l2 functions are usually unnecessary. */ TEXT cachedwbse(SB), $-4 /* D writeback SE */ MOVW R0, R2 MOVW CPSR, R3 CPSID /* splhi */ BARRIERS /* force outstanding stores to cache */ MOVW R2, R0 MOVW 4(FP), R1 ADD R0, R1 /* R1 is end address */ BIC $(CACHELINESZ-1), R0 /* cache line start */ _dwbse: MTCP P(CpSC), 0, R0, C(CpCACHE), C(CpCACHEwb), CpCACHEse ADD $CACHELINESZ, R0 CMP.S R0, R1 BGT _dwbse B _wait TEXT cachedwbinvse(SB), $-4 /* D writeback+invalidate SE */ MOVW R0, R2 MOVW CPSR, R3 CPSID /* splhi */ BARRIERS /* force outstanding stores to cache */ MOVW R2, R0 MOVW 4(FP), R1 ADD R0, R1 /* R1 is end address */ BIC $(CACHELINESZ-1), R0 /* cache line start */ _dwbinvse: MTCP P(CpSC), 0, R0, C(CpCACHE), C(CpCACHEwbi), CpCACHEse ADD $CACHELINESZ, R0 CMP.S R0, R1 BGT _dwbinvse _wait: /* drain write buffer */ BARRIERS MOVW R3, CPSR /* splx */ RET TEXT cachedinvse(SB), $-4 /* D invalidate SE */ MOVW R0, R2 MOVW CPSR, R3 CPSID /* splhi */ BARRIERS /* force outstanding stores to cache */ MOVW R2, R0 MOVW 4(FP), R1 ADD R0, R1 /* R1 is end address */ /* * if start & end addresses are not on cache-line boundaries, * flush first & last cache lines before invalidating. */ AND.S $(CACHELINESZ-1), R0, R4 BEQ stok BIC $(CACHELINESZ-1), R0, R4 /* cache line start */ MTCP P(CpSC), 0, R4, C(CpCACHE), C(CpCACHEwb), CpCACHEse stok: AND.S $(CACHELINESZ-1), R1, R4 BEQ endok BIC $(CACHELINESZ-1), R1, R4 /* cache line start */ MTCP P(CpSC), 0, R4, C(CpCACHE), C(CpCACHEwb), CpCACHEse endok: BIC $(CACHELINESZ-1), R0 /* cache line start */ _dinvse: MTCP P(CpSC), 0, R0, C(CpCACHE), C(CpCACHEinvd), CpCACHEse ADD $CACHELINESZ, R0 CMP.S R0, R1 BGT _dinvse B _wait /* * enable mmu and high vectors */ TEXT mmuenable(SB), 1, $-4 MFCP P(CpSC), 0, R0, C(CpCONTROL), C(0), CpMainctl ORR $CpCmmu, R0 MTCP P(CpSC), 0, R0, C(CpCONTROL), C(0), CpMainctl BARRIERS RET TEXT mmudisable(SB), 1, $-4 MFCP P(CpSC), 0, R0, C(CpCONTROL), C(0), CpMainctl BIC $CpCmmu, R0 MTCP P(CpSC), 0, R0, C(CpCONTROL), C(0), CpMainctl BARRIERS RET /* * If one of these MCR instructions crashes or hangs the machine, * check your Level 1 page table (at TTB) closely. */ TEXT mmuinvalidate(SB), $-4 /* invalidate all */ MOVW CPSR, R2 CPSID /* interrupts off */ BARRIERS MTCP P(CpSC), 0, PC, C(CpTLB), C(CpTLBinvu), CpTLBinv BARRIERS MOVW R2, CPSR /* interrupts restored */ RET TEXT mmuinvalidateaddr(SB), $-4 /* invalidate single entry */ MTCP P(CpSC), 0, R0, C(CpTLB), C(CpTLBinvu), CpTLBinvse BARRIERS RET TEXT cpidget(SB), 1, $-4 /* main ID */ MFCP P(CpSC), 0, R0, C(CpID), C(CpIDidct), CpIDid RET TEXT cpctget(SB), 1, $-4 /* cache type */ MFCP P(CpSC), 0, R0, C(CpID), C(CpIDidct), CpIDct RET TEXT controlget(SB), 1, $-4 /* system control (sctlr) */ MFCP P(CpSC), 0, R0, C(CpCONTROL), C(0), CpMainctl RET TEXT ttbget(SB), 1, $-4 /* translation table base */ MFCP P(CpSC), 0, R0, C(CpTTB), C(0), CpTTB0 RET TEXT ttbput(SB), 1, $-4 /* translation table base */ MOVW CPSR, R2 CPSID MOVW R0, R1 BARRIERS /* finish prior accesses before changing ttb */ MTCP P(CpSC), 0, R1, C(CpTTB), C(0), CpTTB0 MTCP P(CpSC), 0, R1, C(CpTTB), C(0), CpTTB1 /* non-secure too */ MOVW $0, R0 MTCP P(CpSC), 0, R0, C(CpTTB), C(0), CpTTBctl BARRIERS MOVW R2, CPSR RET TEXT dacget(SB), 1, $-4 /* domain access control */ MFCP P(CpSC), 0, R0, C(CpDAC), C(0) RET TEXT dacput(SB), 1, $-4 /* domain access control */ MOVW R0, R1 BARRIERS MTCP P(CpSC), 0, R1, C(CpDAC), C(0) ISB RET TEXT fsrget(SB), 1, $-4 /* fault status */ MFCP P(CpSC), 0, R0, C(CpFSR), C(0), CpDFSR RET TEXT farget(SB), 1, $-4 /* fault address */ MFCP P(CpSC), 0, R0, C(CpFAR), C(0), CpDFAR RET TEXT getpsr(SB), 1, $-4 MOVW CPSR, R0 RET TEXT getscr(SB), 1, $-4 /* secure configuration */ MFCP P(CpSC), 0, R0, C(CpCONTROL), C(CpCONTROLscr), CpSCRscr RET TEXT pidget(SB), 1, $-4 /* address translation pid */ MFCP P(CpSC), 0, R0, C(CpPID), C(0x0) RET TEXT pidput(SB), 1, $-4 /* address translation pid */ MTCP P(CpSC), 0, R0, C(CpPID), C(0), 0 /* pid, v7a deprecated */ MTCP P(CpSC), 0, R0, C(CpPID), C(0), 1 /* context id, errata 754322 */ ISB RET /* * access to yet more coprocessor registers */ TEXT getauxctl(SB), 1, $-4 /* get cortex-a9 aux. ctl. */ MFCP P(CpSC), 0, R0, C(CpCONTROL), C(0), CpAuxctl RET TEXT putauxctl(SB), 1, $-4 /* put cortex-a9 aux. ctl. */ BARRIERS MTCP P(CpSC), 0, R0, C(CpCONTROL), C(0), CpAuxctl BARRIERS RET TEXT getclvlid(SB), 1, $-4 MFCP P(CpSC), CpIDcsize, R0, C(CpID), C(CpIDidct), CpIDclvlid RET TEXT getcyc(SB), 1, $-4 MFCP P(CpSC), 0, R0, C(CpCLD), C(CpCLDcyc), 0 RET TEXT getdebug(SB), 1, $-4 /* get cortex-a9 debug enable register */ MFCP P(CpSC), 0, R0, C(1), C(1), 1 RET TEXT getpc(SB), 1, $-4 MOVW PC, R0 RET TEXT getsb(SB), 1, $-4 MOVW R12, R0 RET TEXT setsp(SB), 1, $-4 MOVW R0, SP RET TEXT splhi(SB), 1, $-4 MOVW CPSR, R0 /* return old CPSR */ CPSID /* turn off interrupts */ CMP.S $0, R(MACH) MOVW.NE R14, 4(R(MACH)) /* save caller pc in m->splpc */ RET TEXT spllo(SB), 1, $-4 /* start marker for devkprof.c */ MOVW CPSR, R0 /* return old CPSR */ MOVW $0, R1 CMP.S R1, R(MACH) MOVW.NE R1, 4(R(MACH)) /* clear m->splpc */ CPSIE RET TEXT splx(SB), 1, $-4 MOVW CPSR, R3 /* must return old CPSR */ CPSID CMP.S $0, R(MACH) MOVW.NE R14, 4(R(MACH)) /* save caller pc in m->splpc */ MOVW R0, CPSR /* reset interrupt level */ MOVW R3, R0 /* must return old CPSR */ RET TEXT spldone(SB), 1, $0 /* end marker for devkprof.c */ RET TEXT islo(SB), 1, $-4 MOVW CPSR, R0 AND $(PsrDirq), R0 EOR $(PsrDirq), R0 RET TEXT clz(SB), $-4 CLZ(0, 0) /* 0 is R0 */ RET TEXT setlabel(SB), 1, $-4 MOVW SP, 0(R0) MOVW R14, 4(R0) /* pc */ MOVW $0, R0 RET TEXT gotolabel(SB), 1, $-4 MOVW 0(R0), SP MOVW 4(R0), R14 /* pc */ MOVW $1, R0 RET TEXT getcallerpc(SB), 1, $-4 MOVW 0(SP), R0 RET TEXT wfi(SB), $-4 MOVW CPSR, R1 /* * an interrupt should break us out of wfi. masking interrupts * slows interrupt response slightly but prevents recursion. */ // CPSIE CPSID BARRIERS WFI MOVW R1, CPSR RET TEXT coherence(SB), $-4 BARRIERS RET GLOBL cpus_proceed+0(SB), $4 #include "cache.v7.s" TEXT tas(SB), $-4 /* _tas(ulong *) */ /* returns old (R0) after modifying (R0) */ MOVW R0,R5 DMB MOVW $1,R2 /* new value of (R0) */ MOVW $MAXSC, R8 tas1: LDREX(5,7) /* LDREX 0(R5),R7 */ CMP.S $0, R7 /* old value non-zero (lock taken)? */ BNE lockbusy /* we lose */ SUB.S $1, R8 BEQ lockloop2 STREX(2,5,4) /* STREX R2,(R5),R4 */ CMP.S $0, R4 BNE tas1 /* strex failed? try again */ DMB B tas0 lockloop2: PUTC('?') PUTC('l') PUTC('t') BL abort(SB) lockbusy: CLREX tas0: MOVW R7, R0 /* return old value */ RET