Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Palmer Dabbelt | 2871 | 57.55% | 1 | 4.76% |
Atish Patra | 985 | 19.74% | 2 | 9.52% |
Anup Patel | 871 | 17.46% | 3 | 14.29% |
Puranjay Mohan | 104 | 2.08% | 1 | 4.76% |
Alexandre Ghiti | 35 | 0.70% | 2 | 9.52% |
Sami Tolvanen | 31 | 0.62% | 2 | 9.52% |
Sia Jee Heng | 30 | 0.60% | 1 | 4.76% |
Guo Ren | 17 | 0.34% | 2 | 9.52% |
Steven Rostedt | 14 | 0.28% | 1 | 4.76% |
Deepak Gupta | 11 | 0.22% | 1 | 4.76% |
Christoph Hellwig | 8 | 0.16% | 1 | 4.76% |
Vincent Chen | 6 | 0.12% | 1 | 4.76% |
Alan Kao | 3 | 0.06% | 1 | 4.76% |
Thomas Gleixner | 2 | 0.04% | 1 | 4.76% |
Souptick Joarder | 1 | 0.02% | 1 | 4.76% |
Total | 4989 | 21 |
123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514
// SPDX-License-Identifier: GPL-2.0-only /* * Copyright (C) 2012 Regents of the University of California * Copyright (C) 2017 SiFive */ #include <linux/kbuild.h> #include <linux/mm.h> #include <linux/sched.h> #include <linux/ftrace.h> #include <linux/suspend.h> #include <asm/kvm_host.h> #include <asm/thread_info.h> #include <asm/ptrace.h> #include <asm/cpu_ops_sbi.h> #include <asm/stacktrace.h> #include <asm/suspend.h> void asm_offsets(void); void asm_offsets(void) { OFFSET(TASK_THREAD_RA, task_struct, thread.ra); OFFSET(TASK_THREAD_SP, task_struct, thread.sp); OFFSET(TASK_THREAD_S0, task_struct, thread.s[0]); OFFSET(TASK_THREAD_S1, task_struct, thread.s[1]); OFFSET(TASK_THREAD_S2, task_struct, thread.s[2]); OFFSET(TASK_THREAD_S3, task_struct, thread.s[3]); OFFSET(TASK_THREAD_S4, task_struct, thread.s[4]); OFFSET(TASK_THREAD_S5, task_struct, thread.s[5]); OFFSET(TASK_THREAD_S6, task_struct, thread.s[6]); OFFSET(TASK_THREAD_S7, task_struct, thread.s[7]); OFFSET(TASK_THREAD_S8, task_struct, thread.s[8]); OFFSET(TASK_THREAD_S9, task_struct, thread.s[9]); OFFSET(TASK_THREAD_S10, task_struct, thread.s[10]); OFFSET(TASK_THREAD_S11, task_struct, thread.s[11]); OFFSET(TASK_TI_CPU, task_struct, thread_info.cpu); OFFSET(TASK_TI_FLAGS, task_struct, thread_info.flags); OFFSET(TASK_TI_PREEMPT_COUNT, task_struct, thread_info.preempt_count); OFFSET(TASK_TI_KERNEL_SP, task_struct, thread_info.kernel_sp); OFFSET(TASK_TI_USER_SP, task_struct, thread_info.user_sp); #ifdef CONFIG_SHADOW_CALL_STACK OFFSET(TASK_TI_SCS_SP, task_struct, thread_info.scs_sp); #endif #ifdef CONFIG_64BIT OFFSET(TASK_TI_A0, task_struct, thread_info.a0); OFFSET(TASK_TI_A1, task_struct, thread_info.a1); OFFSET(TASK_TI_A2, task_struct, thread_info.a2); #endif OFFSET(TASK_TI_CPU_NUM, task_struct, thread_info.cpu); OFFSET(TASK_THREAD_F0, task_struct, thread.fstate.f[0]); OFFSET(TASK_THREAD_F1, task_struct, thread.fstate.f[1]); OFFSET(TASK_THREAD_F2, task_struct, thread.fstate.f[2]); OFFSET(TASK_THREAD_F3, task_struct, thread.fstate.f[3]); OFFSET(TASK_THREAD_F4, task_struct, thread.fstate.f[4]); OFFSET(TASK_THREAD_F5, task_struct, thread.fstate.f[5]); OFFSET(TASK_THREAD_F6, task_struct, thread.fstate.f[6]); OFFSET(TASK_THREAD_F7, task_struct, thread.fstate.f[7]); OFFSET(TASK_THREAD_F8, task_struct, thread.fstate.f[8]); OFFSET(TASK_THREAD_F9, task_struct, thread.fstate.f[9]); OFFSET(TASK_THREAD_F10, task_struct, thread.fstate.f[10]); OFFSET(TASK_THREAD_F11, task_struct, thread.fstate.f[11]); OFFSET(TASK_THREAD_F12, task_struct, thread.fstate.f[12]); OFFSET(TASK_THREAD_F13, task_struct, thread.fstate.f[13]); OFFSET(TASK_THREAD_F14, task_struct, thread.fstate.f[14]); OFFSET(TASK_THREAD_F15, task_struct, thread.fstate.f[15]); OFFSET(TASK_THREAD_F16, task_struct, thread.fstate.f[16]); OFFSET(TASK_THREAD_F17, task_struct, thread.fstate.f[17]); OFFSET(TASK_THREAD_F18, task_struct, thread.fstate.f[18]); OFFSET(TASK_THREAD_F19, task_struct, thread.fstate.f[19]); OFFSET(TASK_THREAD_F20, task_struct, thread.fstate.f[20]); OFFSET(TASK_THREAD_F21, task_struct, thread.fstate.f[21]); OFFSET(TASK_THREAD_F22, task_struct, thread.fstate.f[22]); OFFSET(TASK_THREAD_F23, task_struct, thread.fstate.f[23]); OFFSET(TASK_THREAD_F24, task_struct, thread.fstate.f[24]); OFFSET(TASK_THREAD_F25, task_struct, thread.fstate.f[25]); OFFSET(TASK_THREAD_F26, task_struct, thread.fstate.f[26]); OFFSET(TASK_THREAD_F27, task_struct, thread.fstate.f[27]); OFFSET(TASK_THREAD_F28, task_struct, thread.fstate.f[28]); OFFSET(TASK_THREAD_F29, task_struct, thread.fstate.f[29]); OFFSET(TASK_THREAD_F30, task_struct, thread.fstate.f[30]); OFFSET(TASK_THREAD_F31, task_struct, thread.fstate.f[31]); OFFSET(TASK_THREAD_FCSR, task_struct, thread.fstate.fcsr); #ifdef CONFIG_STACKPROTECTOR OFFSET(TSK_STACK_CANARY, task_struct, stack_canary); #endif DEFINE(PT_SIZE, sizeof(struct pt_regs)); OFFSET(PT_EPC, pt_regs, epc); OFFSET(PT_RA, pt_regs, ra); OFFSET(PT_FP, pt_regs, s0); OFFSET(PT_S0, pt_regs, s0); OFFSET(PT_S1, pt_regs, s1); OFFSET(PT_S2, pt_regs, s2); OFFSET(PT_S3, pt_regs, s3); OFFSET(PT_S4, pt_regs, s4); OFFSET(PT_S5, pt_regs, s5); OFFSET(PT_S6, pt_regs, s6); OFFSET(PT_S7, pt_regs, s7); OFFSET(PT_S8, pt_regs, s8); OFFSET(PT_S9, pt_regs, s9); OFFSET(PT_S10, pt_regs, s10); OFFSET(PT_S11, pt_regs, s11); OFFSET(PT_SP, pt_regs, sp); OFFSET(PT_TP, pt_regs, tp); OFFSET(PT_A0, pt_regs, a0); OFFSET(PT_A1, pt_regs, a1); OFFSET(PT_A2, pt_regs, a2); OFFSET(PT_A3, pt_regs, a3); OFFSET(PT_A4, pt_regs, a4); OFFSET(PT_A5, pt_regs, a5); OFFSET(PT_A6, pt_regs, a6); OFFSET(PT_A7, pt_regs, a7); OFFSET(PT_T0, pt_regs, t0); OFFSET(PT_T1, pt_regs, t1); OFFSET(PT_T2, pt_regs, t2); OFFSET(PT_T3, pt_regs, t3); OFFSET(PT_T4, pt_regs, t4); OFFSET(PT_T5, pt_regs, t5); OFFSET(PT_T6, pt_regs, t6); OFFSET(PT_GP, pt_regs, gp); OFFSET(PT_ORIG_A0, pt_regs, orig_a0); OFFSET(PT_STATUS, pt_regs, status); OFFSET(PT_BADADDR, pt_regs, badaddr); OFFSET(PT_CAUSE, pt_regs, cause); OFFSET(SUSPEND_CONTEXT_REGS, suspend_context, regs); OFFSET(HIBERN_PBE_ADDR, pbe, address); OFFSET(HIBERN_PBE_ORIG, pbe, orig_address); OFFSET(HIBERN_PBE_NEXT, pbe, next); OFFSET(KVM_ARCH_GUEST_ZERO, kvm_vcpu_arch, guest_context.zero); OFFSET(KVM_ARCH_GUEST_RA, kvm_vcpu_arch, guest_context.ra); OFFSET(KVM_ARCH_GUEST_SP, kvm_vcpu_arch, guest_context.sp); OFFSET(KVM_ARCH_GUEST_GP, kvm_vcpu_arch, guest_context.gp); OFFSET(KVM_ARCH_GUEST_TP, kvm_vcpu_arch, guest_context.tp); OFFSET(KVM_ARCH_GUEST_T0, kvm_vcpu_arch, guest_context.t0); OFFSET(KVM_ARCH_GUEST_T1, kvm_vcpu_arch, guest_context.t1); OFFSET(KVM_ARCH_GUEST_T2, kvm_vcpu_arch, guest_context.t2); OFFSET(KVM_ARCH_GUEST_S0, kvm_vcpu_arch, guest_context.s0); OFFSET(KVM_ARCH_GUEST_S1, kvm_vcpu_arch, guest_context.s1); OFFSET(KVM_ARCH_GUEST_A0, kvm_vcpu_arch, guest_context.a0); OFFSET(KVM_ARCH_GUEST_A1, kvm_vcpu_arch, guest_context.a1); OFFSET(KVM_ARCH_GUEST_A2, kvm_vcpu_arch, guest_context.a2); OFFSET(KVM_ARCH_GUEST_A3, kvm_vcpu_arch, guest_context.a3); OFFSET(KVM_ARCH_GUEST_A4, kvm_vcpu_arch, guest_context.a4); OFFSET(KVM_ARCH_GUEST_A5, kvm_vcpu_arch, guest_context.a5); OFFSET(KVM_ARCH_GUEST_A6, kvm_vcpu_arch, guest_context.a6); OFFSET(KVM_ARCH_GUEST_A7, kvm_vcpu_arch, guest_context.a7); OFFSET(KVM_ARCH_GUEST_S2, kvm_vcpu_arch, guest_context.s2); OFFSET(KVM_ARCH_GUEST_S3, kvm_vcpu_arch, guest_context.s3); OFFSET(KVM_ARCH_GUEST_S4, kvm_vcpu_arch, guest_context.s4); OFFSET(KVM_ARCH_GUEST_S5, kvm_vcpu_arch, guest_context.s5); OFFSET(KVM_ARCH_GUEST_S6, kvm_vcpu_arch, guest_context.s6); OFFSET(KVM_ARCH_GUEST_S7, kvm_vcpu_arch, guest_context.s7); OFFSET(KVM_ARCH_GUEST_S8, kvm_vcpu_arch, guest_context.s8); OFFSET(KVM_ARCH_GUEST_S9, kvm_vcpu_arch, guest_context.s9); OFFSET(KVM_ARCH_GUEST_S10, kvm_vcpu_arch, guest_context.s10); OFFSET(KVM_ARCH_GUEST_S11, kvm_vcpu_arch, guest_context.s11); OFFSET(KVM_ARCH_GUEST_T3, kvm_vcpu_arch, guest_context.t3); OFFSET(KVM_ARCH_GUEST_T4, kvm_vcpu_arch, guest_context.t4); OFFSET(KVM_ARCH_GUEST_T5, kvm_vcpu_arch, guest_context.t5); OFFSET(KVM_ARCH_GUEST_T6, kvm_vcpu_arch, guest_context.t6); OFFSET(KVM_ARCH_GUEST_SEPC, kvm_vcpu_arch, guest_context.sepc); OFFSET(KVM_ARCH_GUEST_SSTATUS, kvm_vcpu_arch, guest_context.sstatus); OFFSET(KVM_ARCH_GUEST_HSTATUS, kvm_vcpu_arch, guest_context.hstatus); OFFSET(KVM_ARCH_GUEST_SCOUNTEREN, kvm_vcpu_arch, guest_csr.scounteren); OFFSET(KVM_ARCH_HOST_ZERO, kvm_vcpu_arch, host_context.zero); OFFSET(KVM_ARCH_HOST_RA, kvm_vcpu_arch, host_context.ra); OFFSET(KVM_ARCH_HOST_SP, kvm_vcpu_arch, host_context.sp); OFFSET(KVM_ARCH_HOST_GP, kvm_vcpu_arch, host_context.gp); OFFSET(KVM_ARCH_HOST_TP, kvm_vcpu_arch, host_context.tp); OFFSET(KVM_ARCH_HOST_T0, kvm_vcpu_arch, host_context.t0); OFFSET(KVM_ARCH_HOST_T1, kvm_vcpu_arch, host_context.t1); OFFSET(KVM_ARCH_HOST_T2, kvm_vcpu_arch, host_context.t2); OFFSET(KVM_ARCH_HOST_S0, kvm_vcpu_arch, host_context.s0); OFFSET(KVM_ARCH_HOST_S1, kvm_vcpu_arch, host_context.s1); OFFSET(KVM_ARCH_HOST_A0, kvm_vcpu_arch, host_context.a0); OFFSET(KVM_ARCH_HOST_A1, kvm_vcpu_arch, host_context.a1); OFFSET(KVM_ARCH_HOST_A2, kvm_vcpu_arch, host_context.a2); OFFSET(KVM_ARCH_HOST_A3, kvm_vcpu_arch, host_context.a3); OFFSET(KVM_ARCH_HOST_A4, kvm_vcpu_arch, host_context.a4); OFFSET(KVM_ARCH_HOST_A5, kvm_vcpu_arch, host_context.a5); OFFSET(KVM_ARCH_HOST_A6, kvm_vcpu_arch, host_context.a6); OFFSET(KVM_ARCH_HOST_A7, kvm_vcpu_arch, host_context.a7); OFFSET(KVM_ARCH_HOST_S2, kvm_vcpu_arch, host_context.s2); OFFSET(KVM_ARCH_HOST_S3, kvm_vcpu_arch, host_context.s3); OFFSET(KVM_ARCH_HOST_S4, kvm_vcpu_arch, host_context.s4); OFFSET(KVM_ARCH_HOST_S5, kvm_vcpu_arch, host_context.s5); OFFSET(KVM_ARCH_HOST_S6, kvm_vcpu_arch, host_context.s6); OFFSET(KVM_ARCH_HOST_S7, kvm_vcpu_arch, host_context.s7); OFFSET(KVM_ARCH_HOST_S8, kvm_vcpu_arch, host_context.s8); OFFSET(KVM_ARCH_HOST_S9, kvm_vcpu_arch, host_context.s9); OFFSET(KVM_ARCH_HOST_S10, kvm_vcpu_arch, host_context.s10); OFFSET(KVM_ARCH_HOST_S11, kvm_vcpu_arch, host_context.s11); OFFSET(KVM_ARCH_HOST_T3, kvm_vcpu_arch, host_context.t3); OFFSET(KVM_ARCH_HOST_T4, kvm_vcpu_arch, host_context.t4); OFFSET(KVM_ARCH_HOST_T5, kvm_vcpu_arch, host_context.t5); OFFSET(KVM_ARCH_HOST_T6, kvm_vcpu_arch, host_context.t6); OFFSET(KVM_ARCH_HOST_SEPC, kvm_vcpu_arch, host_context.sepc); OFFSET(KVM_ARCH_HOST_SSTATUS, kvm_vcpu_arch, host_context.sstatus); OFFSET(KVM_ARCH_HOST_HSTATUS, kvm_vcpu_arch, host_context.hstatus); OFFSET(KVM_ARCH_HOST_SSCRATCH, kvm_vcpu_arch, host_sscratch); OFFSET(KVM_ARCH_HOST_STVEC, kvm_vcpu_arch, host_stvec); OFFSET(KVM_ARCH_HOST_SCOUNTEREN, kvm_vcpu_arch, host_scounteren); OFFSET(KVM_ARCH_TRAP_SEPC, kvm_cpu_trap, sepc); OFFSET(KVM_ARCH_TRAP_SCAUSE, kvm_cpu_trap, scause); OFFSET(KVM_ARCH_TRAP_STVAL, kvm_cpu_trap, stval); OFFSET(KVM_ARCH_TRAP_HTVAL, kvm_cpu_trap, htval); OFFSET(KVM_ARCH_TRAP_HTINST, kvm_cpu_trap, htinst); /* F extension */ OFFSET(KVM_ARCH_FP_F_F0, kvm_cpu_context, fp.f.f[0]); OFFSET(KVM_ARCH_FP_F_F1, kvm_cpu_context, fp.f.f[1]); OFFSET(KVM_ARCH_FP_F_F2, kvm_cpu_context, fp.f.f[2]); OFFSET(KVM_ARCH_FP_F_F3, kvm_cpu_context, fp.f.f[3]); OFFSET(KVM_ARCH_FP_F_F4, kvm_cpu_context, fp.f.f[4]); OFFSET(KVM_ARCH_FP_F_F5, kvm_cpu_context, fp.f.f[5]); OFFSET(KVM_ARCH_FP_F_F6, kvm_cpu_context, fp.f.f[6]); OFFSET(KVM_ARCH_FP_F_F7, kvm_cpu_context, fp.f.f[7]); OFFSET(KVM_ARCH_FP_F_F8, kvm_cpu_context, fp.f.f[8]); OFFSET(KVM_ARCH_FP_F_F9, kvm_cpu_context, fp.f.f[9]); OFFSET(KVM_ARCH_FP_F_F10, kvm_cpu_context, fp.f.f[10]); OFFSET(KVM_ARCH_FP_F_F11, kvm_cpu_context, fp.f.f[11]); OFFSET(KVM_ARCH_FP_F_F12, kvm_cpu_context, fp.f.f[12]); OFFSET(KVM_ARCH_FP_F_F13, kvm_cpu_context, fp.f.f[13]); OFFSET(KVM_ARCH_FP_F_F14, kvm_cpu_context, fp.f.f[14]); OFFSET(KVM_ARCH_FP_F_F15, kvm_cpu_context, fp.f.f[15]); OFFSET(KVM_ARCH_FP_F_F16, kvm_cpu_context, fp.f.f[16]); OFFSET(KVM_ARCH_FP_F_F17, kvm_cpu_context, fp.f.f[17]); OFFSET(KVM_ARCH_FP_F_F18, kvm_cpu_context, fp.f.f[18]); OFFSET(KVM_ARCH_FP_F_F19, kvm_cpu_context, fp.f.f[19]); OFFSET(KVM_ARCH_FP_F_F20, kvm_cpu_context, fp.f.f[20]); OFFSET(KVM_ARCH_FP_F_F21, kvm_cpu_context, fp.f.f[21]); OFFSET(KVM_ARCH_FP_F_F22, kvm_cpu_context, fp.f.f[22]); OFFSET(KVM_ARCH_FP_F_F23, kvm_cpu_context, fp.f.f[23]); OFFSET(KVM_ARCH_FP_F_F24, kvm_cpu_context, fp.f.f[24]); OFFSET(KVM_ARCH_FP_F_F25, kvm_cpu_context, fp.f.f[25]); OFFSET(KVM_ARCH_FP_F_F26, kvm_cpu_context, fp.f.f[26]); OFFSET(KVM_ARCH_FP_F_F27, kvm_cpu_context, fp.f.f[27]); OFFSET(KVM_ARCH_FP_F_F28, kvm_cpu_context, fp.f.f[28]); OFFSET(KVM_ARCH_FP_F_F29, kvm_cpu_context, fp.f.f[29]); OFFSET(KVM_ARCH_FP_F_F30, kvm_cpu_context, fp.f.f[30]); OFFSET(KVM_ARCH_FP_F_F31, kvm_cpu_context, fp.f.f[31]); OFFSET(KVM_ARCH_FP_F_FCSR, kvm_cpu_context, fp.f.fcsr); /* D extension */ OFFSET(KVM_ARCH_FP_D_F0, kvm_cpu_context, fp.d.f[0]); OFFSET(KVM_ARCH_FP_D_F1, kvm_cpu_context, fp.d.f[1]); OFFSET(KVM_ARCH_FP_D_F2, kvm_cpu_context, fp.d.f[2]); OFFSET(KVM_ARCH_FP_D_F3, kvm_cpu_context, fp.d.f[3]); OFFSET(KVM_ARCH_FP_D_F4, kvm_cpu_context, fp.d.f[4]); OFFSET(KVM_ARCH_FP_D_F5, kvm_cpu_context, fp.d.f[5]); OFFSET(KVM_ARCH_FP_D_F6, kvm_cpu_context, fp.d.f[6]); OFFSET(KVM_ARCH_FP_D_F7, kvm_cpu_context, fp.d.f[7]); OFFSET(KVM_ARCH_FP_D_F8, kvm_cpu_context, fp.d.f[8]); OFFSET(KVM_ARCH_FP_D_F9, kvm_cpu_context, fp.d.f[9]); OFFSET(KVM_ARCH_FP_D_F10, kvm_cpu_context, fp.d.f[10]); OFFSET(KVM_ARCH_FP_D_F11, kvm_cpu_context, fp.d.f[11]); OFFSET(KVM_ARCH_FP_D_F12, kvm_cpu_context, fp.d.f[12]); OFFSET(KVM_ARCH_FP_D_F13, kvm_cpu_context, fp.d.f[13]); OFFSET(KVM_ARCH_FP_D_F14, kvm_cpu_context, fp.d.f[14]); OFFSET(KVM_ARCH_FP_D_F15, kvm_cpu_context, fp.d.f[15]); OFFSET(KVM_ARCH_FP_D_F16, kvm_cpu_context, fp.d.f[16]); OFFSET(KVM_ARCH_FP_D_F17, kvm_cpu_context, fp.d.f[17]); OFFSET(KVM_ARCH_FP_D_F18, kvm_cpu_context, fp.d.f[18]); OFFSET(KVM_ARCH_FP_D_F19, kvm_cpu_context, fp.d.f[19]); OFFSET(KVM_ARCH_FP_D_F20, kvm_cpu_context, fp.d.f[20]); OFFSET(KVM_ARCH_FP_D_F21, kvm_cpu_context, fp.d.f[21]); OFFSET(KVM_ARCH_FP_D_F22, kvm_cpu_context, fp.d.f[22]); OFFSET(KVM_ARCH_FP_D_F23, kvm_cpu_context, fp.d.f[23]); OFFSET(KVM_ARCH_FP_D_F24, kvm_cpu_context, fp.d.f[24]); OFFSET(KVM_ARCH_FP_D_F25, kvm_cpu_context, fp.d.f[25]); OFFSET(KVM_ARCH_FP_D_F26, kvm_cpu_context, fp.d.f[26]); OFFSET(KVM_ARCH_FP_D_F27, kvm_cpu_context, fp.d.f[27]); OFFSET(KVM_ARCH_FP_D_F28, kvm_cpu_context, fp.d.f[28]); OFFSET(KVM_ARCH_FP_D_F29, kvm_cpu_context, fp.d.f[29]); OFFSET(KVM_ARCH_FP_D_F30, kvm_cpu_context, fp.d.f[30]); OFFSET(KVM_ARCH_FP_D_F31, kvm_cpu_context, fp.d.f[31]); OFFSET(KVM_ARCH_FP_D_FCSR, kvm_cpu_context, fp.d.fcsr); /* * THREAD_{F,X}* might be larger than a S-type offset can handle, but * these are used in performance-sensitive assembly so we can't resort * to loading the long immediate every time. */ DEFINE(TASK_THREAD_RA_RA, offsetof(struct task_struct, thread.ra) - offsetof(struct task_struct, thread.ra) ); DEFINE(TASK_THREAD_SP_RA, offsetof(struct task_struct, thread.sp) - offsetof(struct task_struct, thread.ra) ); DEFINE(TASK_THREAD_S0_RA, offsetof(struct task_struct, thread.s[0]) - offsetof(struct task_struct, thread.ra) ); DEFINE(TASK_THREAD_S1_RA, offsetof(struct task_struct, thread.s[1]) - offsetof(struct task_struct, thread.ra) ); DEFINE(TASK_THREAD_S2_RA, offsetof(struct task_struct, thread.s[2]) - offsetof(struct task_struct, thread.ra) ); DEFINE(TASK_THREAD_S3_RA, offsetof(struct task_struct, thread.s[3]) - offsetof(struct task_struct, thread.ra) ); DEFINE(TASK_THREAD_S4_RA, offsetof(struct task_struct, thread.s[4]) - offsetof(struct task_struct, thread.ra) ); DEFINE(TASK_THREAD_S5_RA, offsetof(struct task_struct, thread.s[5]) - offsetof(struct task_struct, thread.ra) ); DEFINE(TASK_THREAD_S6_RA, offsetof(struct task_struct, thread.s[6]) - offsetof(struct task_struct, thread.ra) ); DEFINE(TASK_THREAD_S7_RA, offsetof(struct task_struct, thread.s[7]) - offsetof(struct task_struct, thread.ra) ); DEFINE(TASK_THREAD_S8_RA, offsetof(struct task_struct, thread.s[8]) - offsetof(struct task_struct, thread.ra) ); DEFINE(TASK_THREAD_S9_RA, offsetof(struct task_struct, thread.s[9]) - offsetof(struct task_struct, thread.ra) ); DEFINE(TASK_THREAD_S10_RA, offsetof(struct task_struct, thread.s[10]) - offsetof(struct task_struct, thread.ra) ); DEFINE(TASK_THREAD_S11_RA, offsetof(struct task_struct, thread.s[11]) - offsetof(struct task_struct, thread.ra) ); DEFINE(TASK_THREAD_F0_F0, offsetof(struct task_struct, thread.fstate.f[0]) - offsetof(struct task_struct, thread.fstate.f[0]) ); DEFINE(TASK_THREAD_F1_F0, offsetof(struct task_struct, thread.fstate.f[1]) - offsetof(struct task_struct, thread.fstate.f[0]) ); DEFINE(TASK_THREAD_F2_F0, offsetof(struct task_struct, thread.fstate.f[2]) - offsetof(struct task_struct, thread.fstate.f[0]) ); DEFINE(TASK_THREAD_F3_F0, offsetof(struct task_struct, thread.fstate.f[3]) - offsetof(struct task_struct, thread.fstate.f[0]) ); DEFINE(TASK_THREAD_F4_F0, offsetof(struct task_struct, thread.fstate.f[4]) - offsetof(struct task_struct, thread.fstate.f[0]) ); DEFINE(TASK_THREAD_F5_F0, offsetof(struct task_struct, thread.fstate.f[5]) - offsetof(struct task_struct, thread.fstate.f[0]) ); DEFINE(TASK_THREAD_F6_F0, offsetof(struct task_struct, thread.fstate.f[6]) - offsetof(struct task_struct, thread.fstate.f[0]) ); DEFINE(TASK_THREAD_F7_F0, offsetof(struct task_struct, thread.fstate.f[7]) - offsetof(struct task_struct, thread.fstate.f[0]) ); DEFINE(TASK_THREAD_F8_F0, offsetof(struct task_struct, thread.fstate.f[8]) - offsetof(struct task_struct, thread.fstate.f[0]) ); DEFINE(TASK_THREAD_F9_F0, offsetof(struct task_struct, thread.fstate.f[9]) - offsetof(struct task_struct, thread.fstate.f[0]) ); DEFINE(TASK_THREAD_F10_F0, offsetof(struct task_struct, thread.fstate.f[10]) - offsetof(struct task_struct, thread.fstate.f[0]) ); DEFINE(TASK_THREAD_F11_F0, offsetof(struct task_struct, thread.fstate.f[11]) - offsetof(struct task_struct, thread.fstate.f[0]) ); DEFINE(TASK_THREAD_F12_F0, offsetof(struct task_struct, thread.fstate.f[12]) - offsetof(struct task_struct, thread.fstate.f[0]) ); DEFINE(TASK_THREAD_F13_F0, offsetof(struct task_struct, thread.fstate.f[13]) - offsetof(struct task_struct, thread.fstate.f[0]) ); DEFINE(TASK_THREAD_F14_F0, offsetof(struct task_struct, thread.fstate.f[14]) - offsetof(struct task_struct, thread.fstate.f[0]) ); DEFINE(TASK_THREAD_F15_F0, offsetof(struct task_struct, thread.fstate.f[15]) - offsetof(struct task_struct, thread.fstate.f[0]) ); DEFINE(TASK_THREAD_F16_F0, offsetof(struct task_struct, thread.fstate.f[16]) - offsetof(struct task_struct, thread.fstate.f[0]) ); DEFINE(TASK_THREAD_F17_F0, offsetof(struct task_struct, thread.fstate.f[17]) - offsetof(struct task_struct, thread.fstate.f[0]) ); DEFINE(TASK_THREAD_F18_F0, offsetof(struct task_struct, thread.fstate.f[18]) - offsetof(struct task_struct, thread.fstate.f[0]) ); DEFINE(TASK_THREAD_F19_F0, offsetof(struct task_struct, thread.fstate.f[19]) - offsetof(struct task_struct, thread.fstate.f[0]) ); DEFINE(TASK_THREAD_F20_F0, offsetof(struct task_struct, thread.fstate.f[20]) - offsetof(struct task_struct, thread.fstate.f[0]) ); DEFINE(TASK_THREAD_F21_F0, offsetof(struct task_struct, thread.fstate.f[21]) - offsetof(struct task_struct, thread.fstate.f[0]) ); DEFINE(TASK_THREAD_F22_F0, offsetof(struct task_struct, thread.fstate.f[22]) - offsetof(struct task_struct, thread.fstate.f[0]) ); DEFINE(TASK_THREAD_F23_F0, offsetof(struct task_struct, thread.fstate.f[23]) - offsetof(struct task_struct, thread.fstate.f[0]) ); DEFINE(TASK_THREAD_F24_F0, offsetof(struct task_struct, thread.fstate.f[24]) - offsetof(struct task_struct, thread.fstate.f[0]) ); DEFINE(TASK_THREAD_F25_F0, offsetof(struct task_struct, thread.fstate.f[25]) - offsetof(struct task_struct, thread.fstate.f[0]) ); DEFINE(TASK_THREAD_F26_F0, offsetof(struct task_struct, thread.fstate.f[26]) - offsetof(struct task_struct, thread.fstate.f[0]) ); DEFINE(TASK_THREAD_F27_F0, offsetof(struct task_struct, thread.fstate.f[27]) - offsetof(struct task_struct, thread.fstate.f[0]) ); DEFINE(TASK_THREAD_F28_F0, offsetof(struct task_struct, thread.fstate.f[28]) - offsetof(struct task_struct, thread.fstate.f[0]) ); DEFINE(TASK_THREAD_F29_F0, offsetof(struct task_struct, thread.fstate.f[29]) - offsetof(struct task_struct, thread.fstate.f[0]) ); DEFINE(TASK_THREAD_F30_F0, offsetof(struct task_struct, thread.fstate.f[30]) - offsetof(struct task_struct, thread.fstate.f[0]) ); DEFINE(TASK_THREAD_F31_F0, offsetof(struct task_struct, thread.fstate.f[31]) - offsetof(struct task_struct, thread.fstate.f[0]) ); DEFINE(TASK_THREAD_FCSR_F0, offsetof(struct task_struct, thread.fstate.fcsr) - offsetof(struct task_struct, thread.fstate.f[0]) ); /* * We allocate a pt_regs on the stack when entering the kernel. This * ensures the alignment is sane. */ DEFINE(PT_SIZE_ON_STACK, ALIGN(sizeof(struct pt_regs), STACK_ALIGN)); OFFSET(KERNEL_MAP_VIRT_ADDR, kernel_mapping, virt_addr); OFFSET(SBI_HART_BOOT_TASK_PTR_OFFSET, sbi_hart_boot_data, task_ptr); OFFSET(SBI_HART_BOOT_STACK_PTR_OFFSET, sbi_hart_boot_data, stack_ptr); DEFINE(STACKFRAME_SIZE_ON_STACK, ALIGN(sizeof(struct stackframe), STACK_ALIGN)); OFFSET(STACKFRAME_FP, stackframe, fp); OFFSET(STACKFRAME_RA, stackframe, ra); #ifdef CONFIG_DYNAMIC_FTRACE_WITH_ARGS DEFINE(FREGS_SIZE_ON_STACK, ALIGN(sizeof(struct __arch_ftrace_regs), STACK_ALIGN)); DEFINE(FREGS_EPC, offsetof(struct __arch_ftrace_regs, epc)); DEFINE(FREGS_RA, offsetof(struct __arch_ftrace_regs, ra)); DEFINE(FREGS_SP, offsetof(struct __arch_ftrace_regs, sp)); DEFINE(FREGS_S0, offsetof(struct __arch_ftrace_regs, s0)); DEFINE(FREGS_T1, offsetof(struct __arch_ftrace_regs, t1)); DEFINE(FREGS_A0, offsetof(struct __arch_ftrace_regs, a0)); DEFINE(FREGS_A1, offsetof(struct __arch_ftrace_regs, a1)); DEFINE(FREGS_A2, offsetof(struct __arch_ftrace_regs, a2)); DEFINE(FREGS_A3, offsetof(struct __arch_ftrace_regs, a3)); DEFINE(FREGS_A4, offsetof(struct __arch_ftrace_regs, a4)); DEFINE(FREGS_A5, offsetof(struct __arch_ftrace_regs, a5)); DEFINE(FREGS_A6, offsetof(struct __arch_ftrace_regs, a6)); DEFINE(FREGS_A7, offsetof(struct __arch_ftrace_regs, a7)); #endif }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1