Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Mark Brown | 928 | 94.02% | 4 | 28.57% |
Catalin Marinas | 36 | 3.65% | 3 | 21.43% |
Eric W. Biedermann | 7 | 0.71% | 1 | 7.14% |
Jean-Philippe Brucker | 7 | 0.71% | 1 | 7.14% |
Julien Thierry | 3 | 0.30% | 1 | 7.14% |
Joey Gouly | 2 | 0.20% | 1 | 7.14% |
Will Deacon | 2 | 0.20% | 1 | 7.14% |
Thomas Gleixner | 1 | 0.10% | 1 | 7.14% |
Dave P Martin | 1 | 0.10% | 1 | 7.14% |
Total | 987 | 14 |
123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254
// SPDX-License-Identifier: GPL-2.0-only #include <linux/mm.h> #include <linux/mman.h> #include <linux/syscalls.h> #include <linux/types.h> #include <asm/cmpxchg.h> #include <asm/cpufeature.h> #include <asm/gcs.h> #include <asm/page.h> static unsigned long alloc_gcs(unsigned long addr, unsigned long size) { int flags = MAP_ANONYMOUS | MAP_PRIVATE; struct mm_struct *mm = current->mm; unsigned long mapped_addr, unused; if (addr) flags |= MAP_FIXED_NOREPLACE; mmap_write_lock(mm); mapped_addr = do_mmap(NULL, addr, size, PROT_READ, flags, VM_SHADOW_STACK | VM_WRITE, 0, &unused, NULL); mmap_write_unlock(mm); return mapped_addr; } static unsigned long gcs_size(unsigned long size) { if (size) return PAGE_ALIGN(size); /* Allocate RLIMIT_STACK/2 with limits of PAGE_SIZE..2G */ size = PAGE_ALIGN(min_t(unsigned long long, rlimit(RLIMIT_STACK) / 2, SZ_2G)); return max(PAGE_SIZE, size); } unsigned long gcs_alloc_thread_stack(struct task_struct *tsk, const struct kernel_clone_args *args) { unsigned long addr, size; if (!system_supports_gcs()) return 0; if (!task_gcs_el0_enabled(tsk)) return 0; if ((args->flags & (CLONE_VFORK | CLONE_VM)) != CLONE_VM) { tsk->thread.gcspr_el0 = read_sysreg_s(SYS_GCSPR_EL0); return 0; } size = args->stack_size / 2; size = gcs_size(size); addr = alloc_gcs(0, size); if (IS_ERR_VALUE(addr)) return addr; tsk->thread.gcs_base = addr; tsk->thread.gcs_size = size; tsk->thread.gcspr_el0 = addr + size - sizeof(u64); return addr; } SYSCALL_DEFINE3(map_shadow_stack, unsigned long, addr, unsigned long, size, unsigned int, flags) { unsigned long alloc_size; unsigned long __user *cap_ptr; unsigned long cap_val; int ret = 0; int cap_offset; if (!system_supports_gcs()) return -EOPNOTSUPP; if (flags & ~(SHADOW_STACK_SET_TOKEN | SHADOW_STACK_SET_MARKER)) return -EINVAL; if (!PAGE_ALIGNED(addr)) return -EINVAL; if (size == 8 || !IS_ALIGNED(size, 8)) return -EINVAL; /* * An overflow would result in attempting to write the restore token * to the wrong location. Not catastrophic, but just return the right * error code and block it. */ alloc_size = PAGE_ALIGN(size); if (alloc_size < size) return -EOVERFLOW; addr = alloc_gcs(addr, alloc_size); if (IS_ERR_VALUE(addr)) return addr; /* * Put a cap token at the end of the allocated region so it * can be switched to. */ if (flags & SHADOW_STACK_SET_TOKEN) { /* Leave an extra empty frame as a top of stack marker? */ if (flags & SHADOW_STACK_SET_MARKER) cap_offset = 2; else cap_offset = 1; cap_ptr = (unsigned long __user *)(addr + size - (cap_offset * sizeof(unsigned long))); cap_val = GCS_CAP(cap_ptr); put_user_gcs(cap_val, cap_ptr, &ret); if (ret != 0) { vm_munmap(addr, size); return -EFAULT; } /* * Ensure the new cap is ordered before standard * memory accesses to the same location. */ gcsb_dsync(); } return addr; } /* * Apply the GCS mode configured for the specified task to the * hardware. */ void gcs_set_el0_mode(struct task_struct *task) { u64 gcscre0_el1 = GCSCRE0_EL1_nTR; if (task->thread.gcs_el0_mode & PR_SHADOW_STACK_ENABLE) gcscre0_el1 |= GCSCRE0_EL1_RVCHKEN | GCSCRE0_EL1_PCRSEL; if (task->thread.gcs_el0_mode & PR_SHADOW_STACK_WRITE) gcscre0_el1 |= GCSCRE0_EL1_STREn; if (task->thread.gcs_el0_mode & PR_SHADOW_STACK_PUSH) gcscre0_el1 |= GCSCRE0_EL1_PUSHMEn; write_sysreg_s(gcscre0_el1, SYS_GCSCRE0_EL1); } void gcs_free(struct task_struct *task) { if (!system_supports_gcs()) return; /* * When fork() with CLONE_VM fails, the child (tsk) already * has a GCS allocated, and exit_thread() calls this function * to free it. In this case the parent (current) and the * child share the same mm struct. */ if (!task->mm || task->mm != current->mm) return; if (task->thread.gcs_base) vm_munmap(task->thread.gcs_base, task->thread.gcs_size); task->thread.gcspr_el0 = 0; task->thread.gcs_base = 0; task->thread.gcs_size = 0; } int arch_set_shadow_stack_status(struct task_struct *task, unsigned long arg) { unsigned long gcs, size; int ret; if (!system_supports_gcs()) return -EINVAL; if (is_compat_thread(task_thread_info(task))) return -EINVAL; /* Reject unknown flags */ if (arg & ~PR_SHADOW_STACK_SUPPORTED_STATUS_MASK) return -EINVAL; ret = gcs_check_locked(task, arg); if (ret != 0) return ret; /* If we are enabling GCS then make sure we have a stack */ if (arg & PR_SHADOW_STACK_ENABLE && !task_gcs_el0_enabled(task)) { /* Do not allow GCS to be reenabled */ if (task->thread.gcs_base || task->thread.gcspr_el0) return -EINVAL; if (task != current) return -EBUSY; size = gcs_size(0); gcs = alloc_gcs(0, size); if (!gcs) return -ENOMEM; task->thread.gcspr_el0 = gcs + size - sizeof(u64); task->thread.gcs_base = gcs; task->thread.gcs_size = size; if (task == current) write_sysreg_s(task->thread.gcspr_el0, SYS_GCSPR_EL0); } task->thread.gcs_el0_mode = arg; if (task == current) gcs_set_el0_mode(task); return 0; } int arch_get_shadow_stack_status(struct task_struct *task, unsigned long __user *arg) { if (!system_supports_gcs()) return -EINVAL; if (is_compat_thread(task_thread_info(task))) return -EINVAL; return put_user(task->thread.gcs_el0_mode, arg); } int arch_lock_shadow_stack_status(struct task_struct *task, unsigned long arg) { if (!system_supports_gcs()) return -EINVAL; if (is_compat_thread(task_thread_info(task))) return -EINVAL; /* * We support locking unknown bits so applications can prevent * any changes in a future proof manner. */ task->thread.gcs_el0_locked |= arg; return 0; }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1