Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Ganesh Goudar | 295 | 89.94% | 1 | 50.00% |
Kees Cook | 33 | 10.06% | 1 | 50.00% |
Total | 328 | 2 |
// SPDX-License-Identifier: GPL-2.0 #include "lkdtm.h" #include <linux/slab.h> #include <linux/vmalloc.h> #include <asm/mmu.h> /* Inserts new slb entries */ static void insert_slb_entry(unsigned long p, int ssize, int page_size) { unsigned long flags; flags = SLB_VSID_KERNEL | mmu_psize_defs[page_size].sllp; preempt_disable(); asm volatile("slbmte %0,%1" : : "r" (mk_vsid_data(p, ssize, flags)), "r" (mk_esid_data(p, ssize, SLB_NUM_BOLTED)) : "memory"); asm volatile("slbmte %0,%1" : : "r" (mk_vsid_data(p, ssize, flags)), "r" (mk_esid_data(p, ssize, SLB_NUM_BOLTED + 1)) : "memory"); preempt_enable(); } /* Inject slb multihit on vmalloc-ed address i.e 0xD00... */ static int inject_vmalloc_slb_multihit(void) { char *p; p = vmalloc(PAGE_SIZE); if (!p) return -ENOMEM; insert_slb_entry((unsigned long)p, MMU_SEGSIZE_1T, mmu_vmalloc_psize); /* * This triggers exception, If handled correctly we must recover * from this error. */ p[0] = '!'; vfree(p); return 0; } /* Inject slb multihit on kmalloc-ed address i.e 0xC00... */ static int inject_kmalloc_slb_multihit(void) { char *p; p = kmalloc(2048, GFP_KERNEL); if (!p) return -ENOMEM; insert_slb_entry((unsigned long)p, MMU_SEGSIZE_1T, mmu_linear_psize); /* * This triggers exception, If handled correctly we must recover * from this error. */ p[0] = '!'; kfree(p); return 0; } /* * Few initial SLB entries are bolted. Add a test to inject * multihit in bolted entry 0. */ static void insert_dup_slb_entry_0(void) { unsigned long test_address = PAGE_OFFSET, *test_ptr; unsigned long esid, vsid; unsigned long i = 0; test_ptr = (unsigned long *)test_address; preempt_disable(); asm volatile("slbmfee %0,%1" : "=r" (esid) : "r" (i)); asm volatile("slbmfev %0,%1" : "=r" (vsid) : "r" (i)); /* for i !=0 we would need to mask out the old entry number */ asm volatile("slbmte %0,%1" : : "r" (vsid), "r" (esid | SLB_NUM_BOLTED) : "memory"); asm volatile("slbmfee %0,%1" : "=r" (esid) : "r" (i)); asm volatile("slbmfev %0,%1" : "=r" (vsid) : "r" (i)); /* for i !=0 we would need to mask out the old entry number */ asm volatile("slbmte %0,%1" : : "r" (vsid), "r" (esid | (SLB_NUM_BOLTED + 1)) : "memory"); pr_info("%s accessing test address 0x%lx: 0x%lx\n", __func__, test_address, *test_ptr); preempt_enable(); } static void lkdtm_PPC_SLB_MULTIHIT(void) { if (!radix_enabled()) { pr_info("Injecting SLB multihit errors\n"); /* * These need not be separate tests, And they do pretty * much same thing. In any case we must recover from the * errors introduced by these functions, machine would not * survive these tests in case of failure to handle. */ inject_vmalloc_slb_multihit(); inject_kmalloc_slb_multihit(); insert_dup_slb_entry_0(); pr_info("Recovered from SLB multihit errors\n"); } else { pr_err("XFAIL: This test is for ppc64 and with hash mode MMU only\n"); } } static struct crashtype crashtypes[] = { CRASHTYPE(PPC_SLB_MULTIHIT), }; struct crashtype_category powerpc_crashtypes = { .crashtypes = crashtypes, .len = ARRAY_SIZE(crashtypes), };
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1