cregit-Linux how code gets into the kernel

Release 4.10 arch/xtensa/include/asm/mmu_context.h

/*
 * Switch an MMU context.
 *
 * This file is subject to the terms and conditions of the GNU General Public
 * License.  See the file "COPYING" in the main directory of this archive
 * for more details.
 *
 * Copyright (C) 2001 - 2013 Tensilica Inc.
 */

#ifndef _XTENSA_MMU_CONTEXT_H

#define _XTENSA_MMU_CONTEXT_H

#ifndef CONFIG_MMU
#include <asm/nommu_context.h>
#else

#include <linux/stringify.h>
#include <linux/sched.h>

#include <asm/vectors.h>

#include <asm/pgtable.h>
#include <asm/cacheflush.h>
#include <asm/tlbflush.h>
#include <asm-generic/mm_hooks.h>
#include <asm-generic/percpu.h>

#if (XCHAL_HAVE_TLBS != 1)
# error "Linux must have an MMU!"
#endif

DECLARE_PER_CPU(unsigned long, asid_cache);

#define cpu_asid_cache(cpu) per_cpu(asid_cache, cpu)

/*
 * NO_CONTEXT is the invalid ASID value that we don't ever assign to
 * any user or kernel context.  We use the reserved values in the
 * ASID_INSERT macro below.
 *
 * 0 invalid
 * 1 kernel
 * 2 reserved
 * 3 reserved
 * 4...255 available
 */


#define NO_CONTEXT	0

#define ASID_USER_FIRST	4

#define ASID_MASK	((1 << XCHAL_MMU_ASID_BITS) - 1)

#define ASID_INSERT(x)	(0x03020001 | (((x) & ASID_MASK) << 8))

void init_mmu(void);


static inline void set_rasid_register (unsigned long val) { __asm__ __volatile__ (" wsr %0, rasid\n\t" " isync\n" : : "a" (val)); }

Contributors

PersonTokensPropCommitsCommitProp
chris zankelchris zankel1285.71%133.33%
max filippovmax filippov17.14%133.33%
adrian bunkadrian bunk17.14%133.33%
Total14100.00%3100.00%


static inline unsigned long get_rasid_register (void) { unsigned long tmp; __asm__ __volatile__ (" rsr %0, rasid\n\t" : "=a" (tmp)); return tmp; }

Contributors

PersonTokensPropCommitsCommitProp
chris zankelchris zankel1890.00%133.33%
max filippovmax filippov15.00%133.33%
adrian bunkadrian bunk15.00%133.33%
Total20100.00%3100.00%


static inline void get_new_mmu_context(struct mm_struct *mm, unsigned int cpu) { unsigned long asid = cpu_asid_cache(cpu); if ((++asid & ASID_MASK) == 0) { /* * Start new asid cycle; continue counting with next * incarnation bits; skipping over 0, 1, 2, 3. */ local_flush_tlb_all(); asid += ASID_USER_FIRST; } cpu_asid_cache(cpu) = asid; mm->context.asid[cpu] = asid; mm->context.cpu = cpu; }

Contributors

PersonTokensPropCommitsCommitProp
max filippovmax filippov3852.78%125.00%
chris zankelchris zankel3345.83%250.00%
adrian bunkadrian bunk11.39%125.00%
Total72100.00%4100.00%


static inline void get_mmu_context(struct mm_struct *mm, unsigned int cpu) { /* * Check if our ASID is of an older version and thus invalid. */ if (mm) { unsigned long asid = mm->context.asid[cpu]; if (asid == NO_CONTEXT || ((asid ^ cpu_asid_cache(cpu)) & ~ASID_MASK)) get_new_mmu_context(mm, cpu); } }

Contributors

PersonTokensPropCommitsCommitProp
max filippovmax filippov5384.13%133.33%
chris zankelchris zankel914.29%133.33%
adrian bunkadrian bunk11.59%133.33%
Total63100.00%3100.00%


static inline void activate_context(struct mm_struct *mm, unsigned int cpu) { get_mmu_context(mm, cpu); set_rasid_register(ASID_INSERT(mm->context.asid[cpu])); invalidate_page_directory(); }

Contributors

PersonTokensPropCommitsCommitProp
max filippovmax filippov2765.85%133.33%
chris zankelchris zankel1434.15%266.67%
Total41100.00%3100.00%

/* * Initialize the context related info for a new mm_struct * instance. Valid cpu values are 0..(NR_CPUS-1), so initializing * to -1 says the process has never run on any core. */
static inline int init_new_context(struct task_struct *tsk, struct mm_struct *mm) { int cpu; for_each_possible_cpu(cpu) { mm->context.asid[cpu] = NO_CONTEXT; } mm->context.cpu = -1; return 0; }

Contributors

PersonTokensPropCommitsCommitProp
chris zankelchris zankel2551.02%133.33%
max filippovmax filippov2346.94%133.33%
adrian bunkadrian bunk12.04%133.33%
Total49100.00%3100.00%


static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, struct task_struct *tsk) { unsigned int cpu = smp_processor_id(); int migrated = next->context.cpu != cpu; /* Flush the icache if we migrated to a new core. */ if (migrated) { __invalidate_icache_all(); next->context.cpu = cpu; } if (migrated || prev != next) activate_context(next, cpu); }

Contributors

PersonTokensPropCommitsCommitProp
max filippovmax filippov3649.32%125.00%
chris zankelchris zankel3649.32%250.00%
adrian bunkadrian bunk11.37%125.00%
Total73100.00%4100.00%

#define activate_mm(prev, next) switch_mm((prev), (next), NULL) #define deactivate_mm(tsk, mm) do { } while (0) /* * Destroy context related info for an mm_struct that is about * to be put to rest. */
static inline void destroy_context(struct mm_struct *mm) { invalidate_page_directory(); }

Contributors

PersonTokensPropCommitsCommitProp
chris zankelchris zankel1493.33%150.00%
adrian bunkadrian bunk16.67%150.00%
Total15100.00%2100.00%


static inline void enter_lazy_tlb(struct mm_struct *mm, struct task_struct *tsk) { /* Nothing to do. */ }

Contributors

PersonTokensPropCommitsCommitProp
chris zankelchris zankel18100.00%1100.00%
Total18100.00%1100.00%

#endif /* CONFIG_MMU */ #endif /* _XTENSA_MMU_CONTEXT_H */

Overall Contributors

PersonTokensPropCommitsCommitProp
chris zankelchris zankel24350.41%330.00%
max filippovmax filippov21644.81%330.00%
johannes weinerjohannes weiner132.70%220.00%
adrian bunkadrian bunk71.45%110.00%
jeremy fitzhardingejeremy fitzhardinge30.62%110.00%
Total482100.00%10100.00%
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.