cregit-Linux how code gets into the kernel

Release 4.14 arch/mips/math-emu/dsemul.c

// SPDX-License-Identifier: GPL-2.0
#include <linux/err.h>
#include <linux/slab.h>
#include <linux/mm_types.h>
#include <linux/sched/task.h>

#include <asm/branch.h>
#include <asm/cacheflush.h>
#include <asm/fpu_emulator.h>
#include <asm/inst.h>
#include <asm/mipsregs.h>
#include <linux/uaccess.h>

/**
 * struct emuframe - The 'emulation' frame structure
 * @emul:       The instruction to 'emulate'.
 * @badinst:    A break instruction to cause a return to the kernel.
 *
 * This structure defines the frames placed within the delay slot emulation
 * page in response to a call to mips_dsemul(). Each thread may be allocated
 * only one frame at any given time. The kernel stores within it the
 * instruction to be 'emulated' followed by a break instruction, then
 * executes the frame in user mode. The break causes a trap to the kernel
 * which leads to do_dsemulret() being called unless the instruction in
 * @emul causes a trap itself, is a branch, or a signal is delivered to
 * the thread. In these cases the allocated frame will either be reused by
 * a subsequent delay slot 'emulation', or be freed during signal delivery or
 * upon thread exit.
 *
 * This approach is used because:
 *
 * - Actually emulating all instructions isn't feasible. We would need to
 *   be able to handle instructions from all revisions of the MIPS ISA,
 *   all ASEs & all vendor instruction set extensions. This would be a
 *   whole lot of work & continual maintenance burden as new instructions
 *   are introduced, and in the case of some vendor extensions may not
 *   even be possible. Thus we need to take the approach of actually
 *   executing the instruction.
 *
 * - We must execute the instruction within user context. If we were to
 *   execute the instruction in kernel mode then it would have access to
 *   kernel resources without very careful checks, leaving us with a
 *   high potential for security or stability issues to arise.
 *
 * - We used to place the frame on the users stack, but this requires
 *   that the stack be executable. This is bad for security so the
 *   per-process page is now used instead.
 *
 * - The instruction in @emul may be something entirely invalid for a
 *   delay slot. The user may (intentionally or otherwise) place a branch
 *   in a delay slot, or a kernel mode instruction, or something else
 *   which generates an exception. Thus we can't rely upon the break in
 *   @badinst always being hit. For this reason we track the index of the
 *   frame allocated to each thread, allowing us to clean it up at later
 *   points such as signal delivery or thread exit.
 *
 * - The user may generate a fake struct emuframe if they wish, invoking
 *   the BRK_MEMU break instruction themselves. We must therefore not
 *   trust that BRK_MEMU means there's actually a valid frame allocated
 *   to the thread, and must not allow the user to do anything they
 *   couldn't already.
 */

struct emuframe {
	
mips_instruction	emul;
	
mips_instruction	badinst;
};


static const int emupage_frame_count = PAGE_SIZE / sizeof(struct emuframe);


static inline __user struct emuframe *dsemul_page(void)
{
	return (__user struct emuframe *)STACK_TOP;
}


static int alloc_emuframe(void) { mm_context_t *mm_ctx = &current->mm->context; int idx; retry: spin_lock(&mm_ctx->bd_emupage_lock); /* Ensure we have an allocation bitmap */ if (!mm_ctx->bd_emupage_allocmap) { mm_ctx->bd_emupage_allocmap = kcalloc(BITS_TO_LONGS(emupage_frame_count), sizeof(unsigned long), GFP_ATOMIC); if (!mm_ctx->bd_emupage_allocmap) { idx = BD_EMUFRAME_NONE; goto out_unlock; } } /* Attempt to allocate a single bit/frame */ idx = bitmap_find_free_region(mm_ctx->bd_emupage_allocmap, emupage_frame_count, 0); if (idx < 0) { /* * Failed to allocate a frame. We'll wait until one becomes * available. We unlock the page so that other threads actually * get the opportunity to free their frames, which means * technically the result of bitmap_full may be incorrect. * However the worst case is that we repeat all this and end up * back here again. */ spin_unlock(&mm_ctx->bd_emupage_lock); if (!wait_event_killable(mm_ctx->bd_emupage_queue, !bitmap_full(mm_ctx->bd_emupage_allocmap, emupage_frame_count))) goto retry; /* Received a fatal signal - just give in */ return BD_EMUFRAME_NONE; } /* Success! */ pr_debug("allocate emuframe %d to %d\n", idx, current->pid); out_unlock: spin_unlock(&mm_ctx->bd_emupage_lock); return idx; }

Contributors

PersonTokensPropCommitsCommitProp
Paul Burton161100.00%1100.00%
Total161100.00%1100.00%


static void free_emuframe(int idx, struct mm_struct *mm) { mm_context_t *mm_ctx = &mm->context; spin_lock(&mm_ctx->bd_emupage_lock); pr_debug("free emuframe %d from %d\n", idx, current->pid); bitmap_clear(mm_ctx->bd_emupage_allocmap, idx, 1); /* If some thread is waiting for a frame, now's its chance */ wake_up(&mm_ctx->bd_emupage_queue); spin_unlock(&mm_ctx->bd_emupage_lock); }

Contributors

PersonTokensPropCommitsCommitProp
Paul Burton6998.57%150.00%
Ralf Bächle11.43%150.00%
Total70100.00%2100.00%


static bool within_emuframe(struct pt_regs *regs) { unsigned long base = (unsigned long)dsemul_page(); if (regs->cp0_epc < base) return false; if (regs->cp0_epc >= (base + PAGE_SIZE)) return false; return true; }

Contributors

PersonTokensPropCommitsCommitProp
Paul Burton4792.16%133.33%
Ralf Bächle47.84%266.67%
Total51100.00%3100.00%


bool dsemul_thread_cleanup(struct task_struct *tsk) { int fr_idx; /* Clear any allocated frame, retrieving its index */ fr_idx = atomic_xchg(&tsk->thread.bd_emu_frame, BD_EMUFRAME_NONE); /* If no frame was allocated, we're done */ if (fr_idx == BD_EMUFRAME_NONE) return false; task_lock(tsk); /* Free the frame that this thread had allocated */ if (tsk->mm) free_emuframe(fr_idx, tsk->mm); task_unlock(tsk); return true; }

Contributors

PersonTokensPropCommitsCommitProp
Paul Burton67100.00%1100.00%
Total67100.00%1100.00%


bool dsemul_thread_rollback(struct pt_regs *regs) { struct emuframe __user *fr; int fr_idx; /* Do nothing if we're not executing from a frame */ if (!within_emuframe(regs)) return false; /* Find the frame being executed */ fr_idx = atomic_read(&current->thread.bd_emu_frame); if (fr_idx == BD_EMUFRAME_NONE) return false; fr = &dsemul_page()[fr_idx]; /* * If the PC is at the emul instruction, roll back to the branch. If * PC is at the badinst (break) instruction, we've already emulated the * instruction so progress to the continue PC. If it's anything else * then something is amiss & the user has branched into some other area * of the emupage - we'll free the allocated frame anyway. */ if (msk_isa16_mode(regs->cp0_epc) == (unsigned long)&fr->emul) regs->cp0_epc = current->thread.bd_emu_branch_pc; else if (msk_isa16_mode(regs->cp0_epc) == (unsigned long)&fr->badinst) regs->cp0_epc = current->thread.bd_emu_cont_pc; atomic_set(&current->thread.bd_emu_frame, BD_EMUFRAME_NONE); free_emuframe(fr_idx, current->mm); return true; }

Contributors

PersonTokensPropCommitsCommitProp
Paul Burton144100.00%1100.00%
Total144100.00%1100.00%


void dsemul_mm_cleanup(struct mm_struct *mm) { mm_context_t *mm_ctx = &mm->context; kfree(mm_ctx->bd_emupage_allocmap); }

Contributors

PersonTokensPropCommitsCommitProp
Paul Burton26100.00%1100.00%
Total26100.00%1100.00%


int mips_dsemul(struct pt_regs *regs, mips_instruction ir, unsigned long branch_pc, unsigned long cont_pc) { int isa16 = get_isa16_mode(regs->cp0_epc); mips_instruction break_math; struct emuframe __user *fr; int err, fr_idx; /* NOP is easy */ if (ir == 0) return -1; /* microMIPS instructions */ if (isa16) { union mips_instruction insn = { .word = ir }; /* NOP16 aka MOVE16 $0, $0 */ if ((ir >> 16) == MM_NOP16) return -1; /* ADDIUPC */ if (insn.mm_a_format.opcode == mm_addiupc_op) { unsigned int rs; s32 v; rs = (((insn.mm_a_format.rs + 0xe) & 0xf) + 2); v = regs->cp0_epc & ~3; v += insn.mm_a_format.simmediate << 2; regs->regs[rs] = (long)v; return -1; } } pr_debug("dsemul 0x%08lx cont at 0x%08lx\n", regs->cp0_epc, cont_pc); /* Allocate a frame if we don't already have one */ fr_idx = atomic_read(&current->thread.bd_emu_frame); if (fr_idx == BD_EMUFRAME_NONE) fr_idx = alloc_emuframe(); if (fr_idx == BD_EMUFRAME_NONE) return SIGBUS; fr = &dsemul_page()[fr_idx]; /* Retrieve the appropriately encoded break instruction */ break_math = BREAK_MATH(isa16); /* Write the instructions to the frame */ if (isa16) { err = __put_user(ir >> 16, (u16 __user *)(&fr->emul)); err |= __put_user(ir & 0xffff, (u16 __user *)((long)(&fr->emul) + 2)); err |= __put_user(break_math >> 16, (u16 __user *)(&fr->badinst)); err |= __put_user(break_math & 0xffff, (u16 __user *)((long)(&fr->badinst) + 2)); } else { err = __put_user(ir, &fr->emul); err |= __put_user(break_math, &fr->badinst); } if (unlikely(err)) { MIPS_FPU_EMU_INC_STATS(errors); free_emuframe(fr_idx, current->mm); return SIGBUS; } /* Record the PC of the branch, PC to continue from & frame index */ current->thread.bd_emu_branch_pc = branch_pc; current->thread.bd_emu_cont_pc = cont_pc; atomic_set(&current->thread.bd_emu_frame, fr_idx); /* Change user register context to execute the frame */ regs->cp0_epc = (unsigned long)&fr->emul | isa16; /* Ensure the icache observes our newly written frame */ flush_cache_sigtramp((unsigned long)&fr->emul); return 0; }

Contributors

PersonTokensPropCommitsCommitProp
Maciej W. Rozycki12828.83%750.00%
Leonid Yegoshin11626.13%17.14%
Ralf Bächle10724.10%321.43%
Paul Burton8920.05%17.14%
David Daney30.68%17.14%
Atsushi Nemoto10.23%17.14%
Total444100.00%14100.00%


bool do_dsemulret(struct pt_regs *xcp) { /* Cleanup the allocated frame, returning if there wasn't one */ if (!dsemul_thread_cleanup(current)) { MIPS_FPU_EMU_INC_STATS(errors); return false; } /* Set EPC to return to post-branch instruction */ xcp->cp0_epc = current->thread.bd_emu_cont_pc; pr_debug("dsemulret to 0x%08lx\n", xcp->cp0_epc); MIPS_FPU_EMU_INC_STATS(ds_emul); return true; }

Contributors

PersonTokensPropCommitsCommitProp
Ralf Bächle2747.37%114.29%
Paul Burton2238.60%228.57%
David Daney610.53%228.57%
Jesper Juhl11.75%114.29%
Leonid Yegoshin11.75%114.29%
Total57100.00%7100.00%


Overall Contributors

PersonTokensPropCommitsCommitProp
Paul Burton67161.17%29.52%
Ralf Bächle16214.77%314.29%
Maciej W. Rozycki12811.67%733.33%
Leonid Yegoshin11710.67%14.76%
David Daney90.82%29.52%
Ingo Molnar60.55%29.52%
Atsushi Nemoto10.09%14.76%
Greg Kroah-Hartman10.09%14.76%
Jesper Juhl10.09%14.76%
Linus Torvalds10.09%14.76%
Total1097100.00%21100.00%
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with cregit.