/* SPDX-License-Identifier: GPL-2.0 */ /* $Id: cache.h,v 1.6 2004/03/11 18:08:05 lethal Exp $ * * include/asm-sh/cache.h * * Copyright 1999 (C) Niibe Yutaka * Copyright 2002, 2003 (C) Paul Mundt */ #ifndef __ASM_SH_CACHE_H #define __ASM_SH_CACHE_H #ifdef __KERNEL__ #include <linux/init.h> #include <cpu/cache.h> #define L1_CACHE_BYTES (1 << L1_CACHE_SHIFT) #define __read_mostly __attribute__((__section__(".data..read_mostly"))) #ifndef __ASSEMBLY__ struct cache_info { unsigned int ways; /* Number of cache ways */ unsigned int sets; /* Number of cache sets */ unsigned int linesz; /* Cache line size (bytes) */ unsigned int way_size; /* sets * line size */ /* * way_incr is the address offset for accessing the next way * in memory mapped cache array ops. */ unsigned int way_incr; unsigned int entry_shift; unsigned int entry_mask; /* * Compute a mask which selects the address bits which overlap between * 1. those used to select the cache set during indexing * 2. those in the physical page number. */ unsigned int alias_mask; unsigned int n_aliases; /* Number of aliases */ unsigned long flags; }; #endif /* __ASSEMBLY__ */ #endif /* __KERNEL__ */ #endif /* __ASM_SH_CACHE_H */Overall Contributors
Person | Tokens | Prop | Commits | CommitProp |
Paul Mundt | 54 | 60.67% | 6 | 50.00% |
Richard Curnow | 19 | 21.35% | 1 | 8.33% |
Linus Torvalds (pre-git) | 9 | 10.11% | 1 | 8.33% |
Andrew Morton | 5 | 5.62% | 2 | 16.67% |
Denys Vlasenko | 1 | 1.12% | 1 | 8.33% |
Greg Kroah-Hartman | 1 | 1.12% | 1 | 8.33% |
Total | 89 | 100.00% | 12 | 100.00% |