Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Ard Biesheuvel | 1253 | 40.50% | 14 | 19.44% |
Catalin Marinas | 635 | 20.52% | 7 | 9.72% |
Russell King | 555 | 17.94% | 16 | 22.22% |
Dave P Martin | 149 | 4.82% | 4 | 5.56% |
Uwe Kleine-König | 112 | 3.62% | 1 | 1.39% |
Will Deacon | 66 | 2.13% | 3 | 4.17% |
Dirk Behme | 60 | 1.94% | 1 | 1.39% |
Nico Pitre | 49 | 1.58% | 2 | 2.78% |
Mark Rutland | 45 | 1.45% | 1 | 1.39% |
Linus Torvalds (pre-git) | 29 | 0.94% | 6 | 8.33% |
Linus Torvalds | 22 | 0.71% | 2 | 2.78% |
Masami Hiramatsu | 21 | 0.68% | 1 | 1.39% |
Marc Zyngier | 17 | 0.55% | 1 | 1.39% |
Ben Dooks | 16 | 0.52% | 1 | 1.39% |
Rabin Vincent | 13 | 0.42% | 1 | 1.39% |
Stefan Agner | 10 | 0.32% | 1 | 1.39% |
Magnus Damm | 9 | 0.29% | 1 | 1.39% |
Viresh Kumar | 7 | 0.23% | 1 | 1.39% |
Vincent Whitchurch | 5 | 0.16% | 1 | 1.39% |
Arnd Bergmann | 4 | 0.13% | 1 | 1.39% |
Andrey Ryabinin | 4 | 0.13% | 1 | 1.39% |
Victor Kamensky | 4 | 0.13% | 1 | 1.39% |
Linus Walleij | 3 | 0.10% | 1 | 1.39% |
Thomas Gleixner | 2 | 0.06% | 1 | 1.39% |
Vladimir Murzin | 2 | 0.06% | 1 | 1.39% |
Lorenzo Pieralisi | 2 | 0.06% | 1 | 1.39% |
Total | 3094 | 72 |
/* SPDX-License-Identifier: GPL-2.0-only */ /* * arch/arm/include/asm/assembler.h * * Copyright (C) 1996-2000 Russell King * * This file contains arm architecture specific defines * for the different processors. * * Do not include any C declarations in this file - it is included by * assembler source. */ #ifndef __ASM_ASSEMBLER_H__ #define __ASM_ASSEMBLER_H__ #ifndef __ASSEMBLY__ #error "Only include this from assembly code" #endif #include <asm/ptrace.h> #include <asm/opcodes-virt.h> #include <asm/asm-offsets.h> #include <asm/page.h> #include <asm/pgtable.h> #include <asm/thread_info.h> #include <asm/uaccess-asm.h> #define IOMEM(x) (x) /* * Endian independent macros for shifting bytes within registers. */ #ifndef __ARMEB__ #define lspull lsr #define lspush lsl #define get_byte_0 lsl #0 #define get_byte_1 lsr #8 #define get_byte_2 lsr #16 #define get_byte_3 lsr #24 #define put_byte_0 lsl #0 #define put_byte_1 lsl #8 #define put_byte_2 lsl #16 #define put_byte_3 lsl #24 #else #define lspull lsl #define lspush lsr #define get_byte_0 lsr #24 #define get_byte_1 lsr #16 #define get_byte_2 lsr #8 #define get_byte_3 lsl #0 #define put_byte_0 lsl #24 #define put_byte_1 lsl #16 #define put_byte_2 lsl #8 #define put_byte_3 lsl #0 #endif /* Select code for any configuration running in BE8 mode */ #ifdef CONFIG_CPU_ENDIAN_BE8 #define ARM_BE8(code...) code #else #define ARM_BE8(code...) #endif /* * Data preload for architectures that support it */ #if __LINUX_ARM_ARCH__ >= 5 #define PLD(code...) code #else #define PLD(code...) #endif /* * This can be used to enable code to cacheline align the destination * pointer when bulk writing to memory. Experiments on StrongARM and * XScale didn't show this a worthwhile thing to do when the cache is not * set to write-allocate (this would need further testing on XScale when WA * is used). * * On Feroceon there is much to gain however, regardless of cache mode. */ #ifdef CONFIG_CPU_FEROCEON #define CALGN(code...) code #else #define CALGN(code...) #endif #define IMM12_MASK 0xfff /* the frame pointer used for stack unwinding */ ARM( fpreg .req r11 ) THUMB( fpreg .req r7 ) /* * Enable and disable interrupts */ #if __LINUX_ARM_ARCH__ >= 6 .macro disable_irq_notrace cpsid i .endm .macro enable_irq_notrace cpsie i .endm #else .macro disable_irq_notrace msr cpsr_c, #PSR_I_BIT | SVC_MODE .endm .macro enable_irq_notrace msr cpsr_c, #SVC_MODE .endm #endif #if __LINUX_ARM_ARCH__ < 7 .macro dsb, args mcr p15, 0, r0, c7, c10, 4 .endm .macro isb, args mcr p15, 0, r0, c7, c5, 4 .endm #endif .macro asm_trace_hardirqs_off, save=1 #if defined(CONFIG_TRACE_IRQFLAGS) .if \save stmdb sp!, {r0-r3, ip, lr} .endif bl trace_hardirqs_off .if \save ldmia sp!, {r0-r3, ip, lr} .endif #endif .endm .macro asm_trace_hardirqs_on, cond=al, save=1 #if defined(CONFIG_TRACE_IRQFLAGS) /* * actually the registers should be pushed and pop'd conditionally, but * after bl the flags are certainly clobbered */ .if \save stmdb sp!, {r0-r3, ip, lr} .endif bl\cond trace_hardirqs_on .if \save ldmia sp!, {r0-r3, ip, lr} .endif #endif .endm .macro disable_irq, save=1 disable_irq_notrace asm_trace_hardirqs_off \save .endm .macro enable_irq asm_trace_hardirqs_on enable_irq_notrace .endm /* * Save the current IRQ state and disable IRQs. Note that this macro * assumes FIQs are enabled, and that the processor is in SVC mode. */ .macro save_and_disable_irqs, oldcpsr #ifdef CONFIG_CPU_V7M mrs \oldcpsr, primask #else mrs \oldcpsr, cpsr #endif disable_irq .endm .macro save_and_disable_irqs_notrace, oldcpsr #ifdef CONFIG_CPU_V7M mrs \oldcpsr, primask #else mrs \oldcpsr, cpsr #endif disable_irq_notrace .endm /* * Restore interrupt state previously stored in a register. We don't * guarantee that this will preserve the flags. */ .macro restore_irqs_notrace, oldcpsr #ifdef CONFIG_CPU_V7M msr primask, \oldcpsr #else msr cpsr_c, \oldcpsr #endif .endm .macro restore_irqs, oldcpsr tst \oldcpsr, #PSR_I_BIT asm_trace_hardirqs_on cond=eq restore_irqs_notrace \oldcpsr .endm /* * Assembly version of "adr rd, BSYM(sym)". This should only be used to * reference local symbols in the same assembly file which are to be * resolved by the assembler. Other usage is undefined. */ .irp c,,eq,ne,cs,cc,mi,pl,vs,vc,hi,ls,ge,lt,gt,le,hs,lo .macro badr\c, rd, sym #ifdef CONFIG_THUMB2_KERNEL adr\c \rd, \sym + 1 #else adr\c \rd, \sym #endif .endm .endr /* * Get current thread_info. */ .macro get_thread_info, rd /* thread_info is the first member of struct task_struct */ get_current \rd .endm /* * Increment/decrement the preempt count. */ #ifdef CONFIG_PREEMPT_COUNT .macro inc_preempt_count, ti, tmp ldr \tmp, [\ti, #TI_PREEMPT] @ get preempt count add \tmp, \tmp, #1 @ increment it str \tmp, [\ti, #TI_PREEMPT] .endm .macro dec_preempt_count, ti, tmp ldr \tmp, [\ti, #TI_PREEMPT] @ get preempt count sub \tmp, \tmp, #1 @ decrement it str \tmp, [\ti, #TI_PREEMPT] .endm #else .macro inc_preempt_count, ti, tmp .endm .macro dec_preempt_count, ti, tmp .endm #endif #define USERL(l, x...) \ 9999: x; \ .pushsection __ex_table,"a"; \ .align 3; \ .long 9999b,l; \ .popsection #define USER(x...) USERL(9001f, x) #ifdef CONFIG_SMP #define ALT_SMP(instr...) \ 9998: instr /* * Note: if you get assembler errors from ALT_UP() when building with * CONFIG_THUMB2_KERNEL, you almost certainly need to use * ALT_SMP( W(instr) ... ) */ #define ALT_UP(instr...) \ .pushsection ".alt.smp.init", "a" ;\ .align 2 ;\ .long 9998b - . ;\ 9997: instr ;\ .if . - 9997b == 2 ;\ nop ;\ .endif ;\ .if . - 9997b != 4 ;\ .error "ALT_UP() content must assemble to exactly 4 bytes";\ .endif ;\ .popsection #define ALT_UP_B(label) \ .pushsection ".alt.smp.init", "a" ;\ .align 2 ;\ .long 9998b - . ;\ W(b) . + (label - 9998b) ;\ .popsection #else #define ALT_SMP(instr...) #define ALT_UP(instr...) instr #define ALT_UP_B(label) b label #endif /* * this_cpu_offset - load the per-CPU offset of this CPU into * register 'rd' */ .macro this_cpu_offset, rd:req #ifdef CONFIG_SMP ALT_SMP(mrc p15, 0, \rd, c13, c0, 4) #ifdef CONFIG_CPU_V6 ALT_UP_B(.L1_\@) .L0_\@: .subsection 1 .L1_\@: ldr_va \rd, __per_cpu_offset b .L0_\@ .previous #endif #else mov \rd, #0 #endif .endm /* * set_current - store the task pointer of this CPU's current task */ .macro set_current, rn:req, tmp:req #if defined(CONFIG_CURRENT_POINTER_IN_TPIDRURO) || defined(CONFIG_SMP) 9998: mcr p15, 0, \rn, c13, c0, 3 @ set TPIDRURO register #ifdef CONFIG_CPU_V6 ALT_UP_B(.L0_\@) .subsection 1 .L0_\@: str_va \rn, __current, \tmp b .L1_\@ .previous .L1_\@: #endif #else str_va \rn, __current, \tmp #endif .endm /* * get_current - load the task pointer of this CPU's current task */ .macro get_current, rd:req #if defined(CONFIG_CURRENT_POINTER_IN_TPIDRURO) || defined(CONFIG_SMP) 9998: mrc p15, 0, \rd, c13, c0, 3 @ get TPIDRURO register #ifdef CONFIG_CPU_V6 ALT_UP_B(.L0_\@) .subsection 1 .L0_\@: ldr_va \rd, __current b .L1_\@ .previous .L1_\@: #endif #else ldr_va \rd, __current #endif .endm /* * reload_current - reload the task pointer of this CPU's current task * into the TLS register */ .macro reload_current, t1:req, t2:req #if defined(CONFIG_CURRENT_POINTER_IN_TPIDRURO) || defined(CONFIG_SMP) #ifdef CONFIG_CPU_V6 ALT_SMP(nop) ALT_UP_B(.L0_\@) #endif ldr_this_cpu \t1, __entry_task, \t1, \t2 mcr p15, 0, \t1, c13, c0, 3 @ store in TPIDRURO .L0_\@: #endif .endm /* * Instruction barrier */ .macro instr_sync #if __LINUX_ARM_ARCH__ >= 7 isb #elif __LINUX_ARM_ARCH__ == 6 mcr p15, 0, r0, c7, c5, 4 #endif .endm /* * SMP data memory barrier */ .macro smp_dmb mode #ifdef CONFIG_SMP #if __LINUX_ARM_ARCH__ >= 7 .ifeqs "\mode","arm" ALT_SMP(dmb ish) .else ALT_SMP(W(dmb) ish) .endif #elif __LINUX_ARM_ARCH__ == 6 ALT_SMP(mcr p15, 0, r0, c7, c10, 5) @ dmb #else #error Incompatible SMP platform #endif .ifeqs "\mode","arm" ALT_UP(nop) .else ALT_UP(W(nop)) .endif #endif .endm /* * Raw SMP data memory barrier */ .macro __smp_dmb mode #if __LINUX_ARM_ARCH__ >= 7 .ifeqs "\mode","arm" dmb ish .else W(dmb) ish .endif #elif __LINUX_ARM_ARCH__ == 6 mcr p15, 0, r0, c7, c10, 5 @ dmb #else .error "Incompatible SMP platform" #endif .endm #if defined(CONFIG_CPU_V7M) /* * setmode is used to assert to be in svc mode during boot. For v7-M * this is done in __v7m_setup, so setmode can be empty here. */ .macro setmode, mode, reg .endm #elif defined(CONFIG_THUMB2_KERNEL) .macro setmode, mode, reg mov \reg, #\mode msr cpsr_c, \reg .endm #else .macro setmode, mode, reg msr cpsr_c, #\mode .endm #endif /* * Helper macro to enter SVC mode cleanly and mask interrupts. reg is * a scratch register for the macro to overwrite. * * This macro is intended for forcing the CPU into SVC mode at boot time. * you cannot return to the original mode. */ .macro safe_svcmode_maskall reg:req #if __LINUX_ARM_ARCH__ >= 6 && !defined(CONFIG_CPU_V7M) mrs \reg , cpsr eor \reg, \reg, #HYP_MODE tst \reg, #MODE_MASK bic \reg , \reg , #MODE_MASK orr \reg , \reg , #PSR_I_BIT | PSR_F_BIT | SVC_MODE THUMB( orr \reg , \reg , #PSR_T_BIT ) bne 1f orr \reg, \reg, #PSR_A_BIT badr lr, 2f msr spsr_cxsf, \reg __MSR_ELR_HYP(14) __ERET 1: msr cpsr_c, \reg 2: #else /* * workaround for possibly broken pre-v6 hardware * (akita, Sharp Zaurus C-1000, PXA270-based) */ setmode PSR_F_BIT | PSR_I_BIT | SVC_MODE, \reg #endif .endm /* * STRT/LDRT access macros with ARM and Thumb-2 variants */ #ifdef CONFIG_THUMB2_KERNEL .macro usraccoff, instr, reg, ptr, inc, off, cond, abort, t=TUSER() 9999: .if \inc == 1 \instr\()b\t\cond\().w \reg, [\ptr, #\off] .elseif \inc == 4 \instr\t\cond\().w \reg, [\ptr, #\off] .else .error "Unsupported inc macro argument" .endif .pushsection __ex_table,"a" .align 3 .long 9999b, \abort .popsection .endm .macro usracc, instr, reg, ptr, inc, cond, rept, abort @ explicit IT instruction needed because of the label @ introduced by the USER macro .ifnc \cond,al .if \rept == 1 itt \cond .elseif \rept == 2 ittt \cond .else .error "Unsupported rept macro argument" .endif .endif @ Slightly optimised to avoid incrementing the pointer twice usraccoff \instr, \reg, \ptr, \inc, 0, \cond, \abort .if \rept == 2 usraccoff \instr, \reg, \ptr, \inc, \inc, \cond, \abort .endif add\cond \ptr, #\rept * \inc .endm #else /* !CONFIG_THUMB2_KERNEL */ .macro usracc, instr, reg, ptr, inc, cond, rept, abort, t=TUSER() .rept \rept 9999: .if \inc == 1 \instr\()b\t\cond \reg, [\ptr], #\inc .elseif \inc == 4 \instr\t\cond \reg, [\ptr], #\inc .else .error "Unsupported inc macro argument" .endif .pushsection __ex_table,"a" .align 3 .long 9999b, \abort .popsection .endr .endm #endif /* CONFIG_THUMB2_KERNEL */ .macro strusr, reg, ptr, inc, cond=al, rept=1, abort=9001f usracc str, \reg, \ptr, \inc, \cond, \rept, \abort .endm .macro ldrusr, reg, ptr, inc, cond=al, rept=1, abort=9001f usracc ldr, \reg, \ptr, \inc, \cond, \rept, \abort .endm /* Utility macro for declaring string literals */ .macro string name:req, string .type \name , #object \name: .asciz "\string" .size \name , . - \name .endm .irp c,,eq,ne,cs,cc,mi,pl,vs,vc,hi,ls,ge,lt,gt,le,hs,lo .macro ret\c, reg #if __LINUX_ARM_ARCH__ < 6 mov\c pc, \reg #else .ifeqs "\reg", "lr" bx\c \reg .else mov\c pc, \reg .endif #endif .endm .endr .macro ret.w, reg ret \reg #ifdef CONFIG_THUMB2_KERNEL nop #endif .endm .macro bug, msg, line #ifdef CONFIG_THUMB2_KERNEL 1: .inst 0xde02 #else 1: .inst 0xe7f001f2 #endif #ifdef CONFIG_DEBUG_BUGVERBOSE .pushsection .rodata.str, "aMS", %progbits, 1 2: .asciz "\msg" .popsection .pushsection __bug_table, "aw" .align 2 .word 1b, 2b .hword \line .popsection #endif .endm #ifdef CONFIG_KPROBES #define _ASM_NOKPROBE(entry) \ .pushsection "_kprobe_blacklist", "aw" ; \ .balign 4 ; \ .long entry; \ .popsection #else #define _ASM_NOKPROBE(entry) #endif .macro __adldst_l, op, reg, sym, tmp, c .if __LINUX_ARM_ARCH__ < 7 ldr\c \tmp, .La\@ .subsection 1 .align 2 .La\@: .long \sym - .Lpc\@ .previous .else .ifnb \c THUMB( ittt \c ) .endif movw\c \tmp, #:lower16:\sym - .Lpc\@ movt\c \tmp, #:upper16:\sym - .Lpc\@ .endif #ifndef CONFIG_THUMB2_KERNEL .set .Lpc\@, . + 8 // PC bias .ifc \op, add add\c \reg, \tmp, pc .else \op\c \reg, [pc, \tmp] .endif #else .Lb\@: add\c \tmp, \tmp, pc /* * In Thumb-2 builds, the PC bias depends on whether we are currently * emitting into a .arm or a .thumb section. The size of the add opcode * above will be 2 bytes when emitting in Thumb mode and 4 bytes when * emitting in ARM mode, so let's use this to account for the bias. */ .set .Lpc\@, . + (. - .Lb\@) .ifnc \op, add \op\c \reg, [\tmp] .endif #endif .endm /* * mov_l - move a constant value or [relocated] address into a register */ .macro mov_l, dst:req, imm:req, cond .if __LINUX_ARM_ARCH__ < 7 ldr\cond \dst, =\imm .else movw\cond \dst, #:lower16:\imm movt\cond \dst, #:upper16:\imm .endif .endm /* * adr_l - adr pseudo-op with unlimited range * * @dst: destination register * @sym: name of the symbol * @cond: conditional opcode suffix */ .macro adr_l, dst:req, sym:req, cond __adldst_l add, \dst, \sym, \dst, \cond .endm /* * ldr_l - ldr <literal> pseudo-op with unlimited range * * @dst: destination register * @sym: name of the symbol * @cond: conditional opcode suffix */ .macro ldr_l, dst:req, sym:req, cond __adldst_l ldr, \dst, \sym, \dst, \cond .endm /* * str_l - str <literal> pseudo-op with unlimited range * * @src: source register * @sym: name of the symbol * @tmp: mandatory scratch register * @cond: conditional opcode suffix */ .macro str_l, src:req, sym:req, tmp:req, cond __adldst_l str, \src, \sym, \tmp, \cond .endm .macro __ldst_va, op, reg, tmp, sym, cond, offset #if __LINUX_ARM_ARCH__ >= 7 || \ !defined(CONFIG_ARM_HAS_GROUP_RELOCS) || \ (defined(MODULE) && defined(CONFIG_ARM_MODULE_PLTS)) mov_l \tmp, \sym, \cond #else /* * Avoid a literal load, by emitting a sequence of ADD/LDR instructions * with the appropriate relocations. The combined sequence has a range * of -/+ 256 MiB, which should be sufficient for the core kernel and * for modules loaded into the module region. */ .globl \sym .reloc .L0_\@, R_ARM_ALU_PC_G0_NC, \sym .reloc .L1_\@, R_ARM_ALU_PC_G1_NC, \sym .reloc .L2_\@, R_ARM_LDR_PC_G2, \sym .L0_\@: sub\cond \tmp, pc, #8 - \offset .L1_\@: sub\cond \tmp, \tmp, #4 - \offset .L2_\@: #endif \op\cond \reg, [\tmp, #\offset] .endm /* * ldr_va - load a 32-bit word from the virtual address of \sym */ .macro ldr_va, rd:req, sym:req, cond, tmp, offset=0 .ifnb \tmp __ldst_va ldr, \rd, \tmp, \sym, \cond, \offset .else __ldst_va ldr, \rd, \rd, \sym, \cond, \offset .endif .endm /* * str_va - store a 32-bit word to the virtual address of \sym */ .macro str_va, rn:req, sym:req, tmp:req, cond __ldst_va str, \rn, \tmp, \sym, \cond, 0 .endm /* * ldr_this_cpu_armv6 - Load a 32-bit word from the per-CPU variable 'sym', * without using a temp register. Supported in ARM mode * only. */ .macro ldr_this_cpu_armv6, rd:req, sym:req this_cpu_offset \rd .globl \sym .reloc .L0_\@, R_ARM_ALU_PC_G0_NC, \sym .reloc .L1_\@, R_ARM_ALU_PC_G1_NC, \sym .reloc .L2_\@, R_ARM_LDR_PC_G2, \sym add \rd, \rd, pc .L0_\@: sub \rd, \rd, #4 .L1_\@: sub \rd, \rd, #0 .L2_\@: ldr \rd, [\rd, #4] .endm /* * ldr_this_cpu - Load a 32-bit word from the per-CPU variable 'sym' * into register 'rd', which may be the stack pointer, * using 't1' and 't2' as general temp registers. These * are permitted to overlap with 'rd' if != sp */ .macro ldr_this_cpu, rd:req, sym:req, t1:req, t2:req #ifndef CONFIG_SMP ldr_va \rd, \sym, tmp=\t1 #elif __LINUX_ARM_ARCH__ >= 7 || \ !defined(CONFIG_ARM_HAS_GROUP_RELOCS) || \ (defined(MODULE) && defined(CONFIG_ARM_MODULE_PLTS)) this_cpu_offset \t1 mov_l \t2, \sym ldr \rd, [\t1, \t2] #else ldr_this_cpu_armv6 \rd, \sym #endif .endm /* * rev_l - byte-swap a 32-bit value * * @val: source/destination register * @tmp: scratch register */ .macro rev_l, val:req, tmp:req .if __LINUX_ARM_ARCH__ < 6 eor \tmp, \val, \val, ror #16 bic \tmp, \tmp, #0x00ff0000 mov \val, \val, ror #8 eor \val, \val, \tmp, lsr #8 .else rev \val, \val .endif .endm .if __LINUX_ARM_ARCH__ < 6 .set .Lrev_l_uses_tmp, 1 .else .set .Lrev_l_uses_tmp, 0 .endif /* * bl_r - branch and link to register * * @dst: target to branch to * @c: conditional opcode suffix */ .macro bl_r, dst:req, c .if __LINUX_ARM_ARCH__ < 6 mov\c lr, pc mov\c pc, \dst .else blx\c \dst .endif .endm #endif /* __ASM_ASSEMBLER_H__ */
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1