cregit-Linux how code gets into the kernel

Release 4.14 arch/m68k/include/asm/pgtable_mm.h

/* SPDX-License-Identifier: GPL-2.0 */
#ifndef _M68K_PGTABLE_H

#define _M68K_PGTABLE_H

#include <asm-generic/4level-fixup.h>

#include <asm/setup.h>

#ifndef __ASSEMBLY__
#include <asm/processor.h>
#include <linux/sched.h>
#include <linux/threads.h>

/*
 * This file contains the functions and defines necessary to modify and use
 * the m68k page table tree.
 */

#include <asm/virtconvert.h>

/* Certain architectures need to do special things when pte's
 * within a page table are directly modified.  Thus, the following
 * hook is made available.
 */

#define set_pte(pteptr, pteval)					\
	do{                                                     \
                *(pteptr) = (pteval);                           \
        } while(0)

#define set_pte_at(mm,addr,ptep,pteval) set_pte(ptep,pteval)


/* PMD_SHIFT determines the size of the area a second-level page table can map */
#ifdef CONFIG_SUN3

#define PMD_SHIFT       17
#else

#define PMD_SHIFT	22
#endif

#define PMD_SIZE	(1UL << PMD_SHIFT)

#define PMD_MASK	(~(PMD_SIZE-1))

/* PGDIR_SHIFT determines what a third-level page table entry can map */
#ifdef CONFIG_SUN3

#define PGDIR_SHIFT     17
#elif defined(CONFIG_COLDFIRE)

#define PGDIR_SHIFT     22
#else

#define PGDIR_SHIFT	25
#endif

#define PGDIR_SIZE	(1UL << PGDIR_SHIFT)

#define PGDIR_MASK	(~(PGDIR_SIZE-1))

/*
 * entries per page directory level: the m68k is configured as three-level,
 * so we do have PMD level physically.
 */
#ifdef CONFIG_SUN3

#define PTRS_PER_PTE   16

#define __PAGETABLE_PMD_FOLDED

#define PTRS_PER_PMD   1

#define PTRS_PER_PGD   2048
#elif defined(CONFIG_COLDFIRE)

#define PTRS_PER_PTE	512

#define __PAGETABLE_PMD_FOLDED

#define PTRS_PER_PMD	1

#define PTRS_PER_PGD	1024
#else

#define PTRS_PER_PTE	1024

#define PTRS_PER_PMD	8

#define PTRS_PER_PGD	128
#endif

#define USER_PTRS_PER_PGD	(TASK_SIZE/PGDIR_SIZE)

#define FIRST_USER_ADDRESS	0UL

/* Virtual address region for use by kernel_map() */
#ifdef CONFIG_SUN3

#define KMAP_START     0x0DC00000

#define KMAP_END       0x0E000000
#elif defined(CONFIG_COLDFIRE)

#define KMAP_START	0xe0000000

#define KMAP_END	0xf0000000
#else

#define	KMAP_START	0xd0000000

#define	KMAP_END	0xf0000000
#endif

#ifdef CONFIG_SUN3
extern unsigned long m68k_vmalloc_end;

#define VMALLOC_START 0x0f800000

#define VMALLOC_END m68k_vmalloc_end
#elif defined(CONFIG_COLDFIRE)

#define VMALLOC_START	0xd0000000

#define VMALLOC_END	0xe0000000
#else
/* Just any arbitrary offset to the start of the vmalloc VM area: the
 * current 8MB value just means that there will be a 8MB "hole" after the
 * physical memory until the kernel virtual memory starts.  That means that
 * any out-of-bounds memory accesses will hopefully be caught.
 * The vmalloc() routines leaves a hole of 4kB between each vmalloced
 * area for the same reason. ;)
 */

#define VMALLOC_OFFSET	(8*1024*1024)

#define VMALLOC_START (((unsigned long) high_memory + VMALLOC_OFFSET) & ~(VMALLOC_OFFSET-1))

#define VMALLOC_END KMAP_START
#endif

/* zero page used for uninitialized stuff */
extern void *empty_zero_page;

/*
 * ZERO_PAGE is a global shared page that is always zero: used
 * for zero-mapped memory areas etc..
 */

#define ZERO_PAGE(vaddr)	(virt_to_page(empty_zero_page))

/* number of bits that fit into a memory pointer */

#define BITS_PER_PTR			(8*sizeof(unsigned long))

/* to align the pointer to a pointer address */

#define PTR_MASK			(~(sizeof(void*)-1))

/* sizeof(void*)==1<<SIZEOF_PTR_LOG2 */
/* 64-bit machines, beware!  SRB. */

#define SIZEOF_PTR_LOG2			       2

extern void kernel_set_cachemode(void *addr, unsigned long size, int cmode);

/*
 * The m68k doesn't have any external MMU info: the kernel page
 * tables contain all the necessary information.  The Sun3 does, but
 * they are updated on demand.
 */

static inline void update_mmu_cache(struct vm_area_struct *vma, unsigned long address, pte_t *ptep) { }

Contributors

PersonTokensPropCommitsCommitProp
Linus Torvalds (pre-git)1684.21%133.33%
Russell King210.53%133.33%
Andrew Morton15.26%133.33%
Total19100.00%3100.00%

#endif /* !__ASSEMBLY__ */ #define kern_addr_valid(addr) (1) /* MMU-specific headers */ #ifdef CONFIG_SUN3 #include <asm/sun3_pgtable.h> #elif defined(CONFIG_COLDFIRE) #include <asm/mcf_pgtable.h> #else #include <asm/motorola_pgtable.h> #endif #ifndef __ASSEMBLY__ /* * Macro to mark a page protection value as "uncacheable". */ #ifdef CONFIG_COLDFIRE # define pgprot_noncached(prot) (__pgprot(pgprot_val(prot) | CF_PAGE_NOCACHE)) #else #ifdef SUN3_PAGE_NOCACHE # define __SUN3_PAGE_NOCACHE SUN3_PAGE_NOCACHE #else # define __SUN3_PAGE_NOCACHE 0 #endif #define pgprot_noncached(prot) \ (MMU_IS_SUN3 \ ? (__pgprot(pgprot_val(prot) | __SUN3_PAGE_NOCACHE)) \ : ((MMU_IS_851 || MMU_IS_030) \ ? (__pgprot(pgprot_val(prot) | _PAGE_NOCACHE030)) \ : (MMU_IS_040 || MMU_IS_060) \ ? (__pgprot((pgprot_val(prot) & _CACHEMASK040) | _PAGE_NOCACHE_S)) \ : (prot))) #endif /* CONFIG_COLDFIRE */ #include <asm-generic/pgtable.h> #endif /* !__ASSEMBLY__ */ /* * No page table caches to initialise */ #define pgtable_cache_init() do { } while (0) #define check_pgt_cache() do { } while (0) #endif /* _M68K_PGTABLE_H */

Overall Contributors

PersonTokensPropCommitsCommitProp
Linus Torvalds (pre-git)25560.57%1350.00%
Greg Ungerer9422.33%13.85%
Andrew Morton255.94%27.69%
David S. Miller133.09%13.85%
Geert Uytterhoeven112.61%13.85%
Kirill A. Shutemov71.66%27.69%
Linus Torvalds61.43%13.85%
Nicholas Piggin30.71%13.85%
Alexey Dobriyan30.71%13.85%
Russell King20.48%13.85%
Hugh Dickins10.24%13.85%
Greg Kroah-Hartman10.24%13.85%
Total421100.00%26100.00%
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with cregit.