Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Paul Mundt | 296 | 94.27% | 4 | 80.00% |
Matt Fleming | 18 | 5.73% | 1 | 20.00% |
Total | 314 | 5 |
/* * arch/sh/mm/tlb-pteaex.c * * TLB operations for SH-X3 CPUs featuring PTE ASID Extensions. * * Copyright (C) 2009 Paul Mundt * * This file is subject to the terms and conditions of the GNU General Public * License. See the file "COPYING" in the main directory of this archive * for more details. */ #include <linux/kernel.h> #include <linux/mm.h> #include <linux/io.h> #include <asm/mmu_context.h> #include <asm/cacheflush.h> void __update_tlb(struct vm_area_struct *vma, unsigned long address, pte_t pte) { unsigned long flags, pteval, vpn; /* * Handle debugger faulting in for debugee. */ if (vma && current->active_mm != vma->vm_mm) return; local_irq_save(flags); /* Set PTEH register */ vpn = address & MMU_VPN_MASK; __raw_writel(vpn, MMU_PTEH); /* Set PTEAEX */ __raw_writel(get_asid(), MMU_PTEAEX); pteval = pte.pte_low; /* Set PTEA register */ #ifdef CONFIG_X2TLB /* * For the extended mode TLB this is trivial, only the ESZ and * EPR bits need to be written out to PTEA, with the remainder of * the protection bits (with the exception of the compat-mode SZ * and PR bits, which are cleared) being written out in PTEL. */ __raw_writel(pte.pte_high, MMU_PTEA); #endif /* Set PTEL register */ pteval &= _PAGE_FLAGS_HARDWARE_MASK; /* drop software flags */ #ifdef CONFIG_CACHE_WRITETHROUGH pteval |= _PAGE_WT; #endif /* conveniently, we want all the software flags to be 0 anyway */ __raw_writel(pteval, MMU_PTEL); /* Load the TLB */ asm volatile("ldtlb": /* no output */ : /* no input */ : "memory"); local_irq_restore(flags); } /* * While SH-X2 extended TLB mode splits out the memory-mapped I/UTLB * data arrays, SH-X3 cores with PTEAEX split out the memory-mapped * address arrays. In compat mode the second array is inaccessible, while * in extended mode, the legacy 8-bit ASID field in address array 1 has * undefined behaviour. */ void local_flush_tlb_one(unsigned long asid, unsigned long page) { jump_to_uncached(); __raw_writel(page, MMU_UTLB_ADDRESS_ARRAY | MMU_PAGE_ASSOC_BIT); __raw_writel(asid, MMU_UTLB_ADDRESS_ARRAY2 | MMU_PAGE_ASSOC_BIT); __raw_writel(page, MMU_ITLB_ADDRESS_ARRAY | MMU_PAGE_ASSOC_BIT); __raw_writel(asid, MMU_ITLB_ADDRESS_ARRAY2 | MMU_PAGE_ASSOC_BIT); back_to_cached(); } void local_flush_tlb_all(void) { unsigned long flags, status; int i; /* * Flush all the TLB. */ local_irq_save(flags); jump_to_uncached(); status = __raw_readl(MMUCR); status = ((status & MMUCR_URB) >> MMUCR_URB_SHIFT); if (status == 0) status = MMUCR_URB_NENTRIES; for (i = 0; i < status; i++) __raw_writel(0x0, MMU_UTLB_ADDRESS_ARRAY | (i << 8)); for (i = 0; i < 4; i++) __raw_writel(0x0, MMU_ITLB_ADDRESS_ARRAY | (i << 8)); back_to_cached(); ctrl_barrier(); local_irq_restore(flags); }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1