cregit-Linux how code gets into the kernel

Release 4.11 arch/sh/mm/tlbex_64.c

Directory: arch/sh/mm
/*
 * The SH64 TLB miss.
 *
 * Original code from fault.c
 * Copyright (C) 2000, 2001  Paolo Alberelli
 *
 * Fast PTE->TLB refill path
 * Copyright (C) 2003 Richard.Curnow@superh.com
 *
 * IMPORTANT NOTES :
 * The do_fast_page_fault function is called from a context in entry.S
 * where very few registers have been saved.  In particular, the code in
 * this file must be compiled not to use ANY caller-save registers that
 * are not part of the restricted save set.  Also, it means that code in
 * this file must not make calls to functions elsewhere in the kernel, or
 * else the excepting context will see corruption in its caller-save
 * registers.  Plus, the entry.S save area is non-reentrant, so this code
 * has to run with SR.BL==1, i.e. no interrupts taken inside it and panic
 * on any exception.
 *
 * This file is subject to the terms and conditions of the GNU General Public
 * License.  See the file "COPYING" in the main directory of this archive
 * for more details.
 */
#include <linux/signal.h>
#include <linux/sched.h>
#include <linux/kernel.h>
#include <linux/errno.h>
#include <linux/string.h>
#include <linux/types.h>
#include <linux/ptrace.h>
#include <linux/mman.h>
#include <linux/mm.h>
#include <linux/smp.h>
#include <linux/interrupt.h>
#include <linux/kprobes.h>
#include <asm/tlb.h>
#include <asm/io.h>
#include <linux/uaccess.h>
#include <asm/pgalloc.h>
#include <asm/mmu_context.h>


static int handle_tlbmiss(unsigned long long protection_flags, unsigned long address) { pgd_t *pgd; pud_t *pud; pmd_t *pmd; pte_t *pte; pte_t entry; if (is_vmalloc_addr((void *)address)) { pgd = pgd_offset_k(address); } else { if (unlikely(address >= TASK_SIZE || !current->mm)) return 1; pgd = pgd_offset(current->mm, address); } pud = pud_offset(pgd, address); if (pud_none(*pud) || !pud_present(*pud)) return 1; pmd = pmd_offset(pud, address); if (pmd_none(*pmd) || !pmd_present(*pmd)) return 1; pte = pte_offset_kernel(pmd, address); entry = *pte; if (pte_none(entry) || !pte_present(entry)) return 1; /* * If the page doesn't have sufficient protection bits set to * service the kind of fault being handled, there's not much * point doing the TLB refill. Punt the fault to the general * handler. */ if ((pte_val(entry) & protection_flags) != protection_flags) return 1; update_mmu_cache(NULL, address, pte); return 0; }

Contributors

PersonTokensPropCommitsCommitProp
Paul Mundt198100.00%5100.00%
Total198100.00%5100.00%

/* * Put all this information into one structure so that everything is just * arithmetic relative to a single base address. This reduces the number * of movi/shori pairs needed just to load addresses of static data. */ struct expevt_lookup { unsigned short protection_flags[8]; unsigned char is_text_access[8]; unsigned char is_write_access[8]; }; #define PRU (1<<9) #define PRW (1<<8) #define PRX (1<<7) #define PRR (1<<6) /* Sized as 8 rather than 4 to allow checking the PTE's PRU bit against whether the fault happened in user mode or privileged mode. */ static struct expevt_lookup expevt_lookup_table = { .protection_flags = {PRX, PRX, 0, 0, PRR, PRR, PRW, PRW}, .is_text_access = {1, 1, 0, 0, 0, 0, 0, 0} };
static inline unsigned int expevt_to_fault_code(unsigned long expevt) { if (expevt == 0xa40) return FAULT_CODE_ITLB; else if (expevt == 0x060) return FAULT_CODE_WRITE; return 0; }

Contributors

PersonTokensPropCommitsCommitProp
Paul Mundt34100.00%1100.00%
Total34100.00%1100.00%

/* This routine handles page faults that can be serviced just by refilling a TLB entry from an existing page table entry. (This case represents a very large majority of page faults.) Return 1 if the fault was successfully handled. Return 0 if the fault could not be handled. (This leads into the general fault handling in fault.c which deals with mapping file-backed pages, stack growth, segmentation faults, swapping etc etc) */
asmlinkage int __kprobes do_fast_page_fault(unsigned long long ssr_md, unsigned long long expevt, unsigned long address) { unsigned long long protection_flags; unsigned long long index; unsigned long long expevt4; unsigned int fault_code; /* The next few lines implement a way of hashing EXPEVT into a * small array index which can be used to lookup parameters * specific to the type of TLBMISS being handled. * * Note: * ITLBMISS has EXPEVT==0xa40 * RTLBMISS has EXPEVT==0x040 * WTLBMISS has EXPEVT==0x060 */ expevt4 = (expevt >> 4); /* TODO : xor ssr_md into this expression too. Then we can check * that PRU is set when it needs to be. */ index = expevt4 ^ (expevt4 >> 5); index &= 7; fault_code = expevt_to_fault_code(expevt); protection_flags = expevt_lookup_table.protection_flags[index]; if (expevt_lookup_table.is_text_access[index]) fault_code |= FAULT_CODE_ITLB; if (!ssr_md) fault_code |= FAULT_CODE_USER; set_thread_fault_code(fault_code); return handle_tlbmiss(protection_flags, address); }

Contributors

PersonTokensPropCommitsCommitProp
Paul Mundt115100.00%5100.00%
Total115100.00%5100.00%


Overall Contributors

PersonTokensPropCommitsCommitProp
Paul Mundt49299.80%685.71%
Linus Torvalds10.20%114.29%
Total493100.00%7100.00%
Directory: arch/sh/mm
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with cregit.