#ifndef _LINUX_SLAB_DEF_H #define _LINUX_SLAB_DEF_H #include <linux/reciprocal_div.h> /* * Definitions unique to the original Linux SLAB allocator. */ struct kmem_cache { struct array_cache __percpu *cpu_cache; /* 1) Cache tunables. Protected by slab_mutex */ unsigned int batchcount; unsigned int limit; unsigned int shared; unsigned int size; struct reciprocal_value reciprocal_buffer_size; /* 2) touched by every alloc & free from the backend */ unsigned int flags; /* constant flags */ unsigned int num; /* # of objs per slab */ /* 3) cache_grow/shrink */ /* order of pgs per slab (2^n) */ unsigned int gfporder; /* force GFP flags, e.g. GFP_DMA */ gfp_t allocflags; size_t colour; /* cache colouring range */ unsigned int colour_off; /* colour offset */ struct kmem_cache *freelist_cache; unsigned int freelist_size; /* constructor func */ void (*ctor)(void *obj); /* 4) cache creation/removal */ const char *name; struct list_head list; int refcount; int object_size; int align; /* 5) statistics */ #ifdef CONFIG_DEBUG_SLAB unsigned long num_active; unsigned long num_allocations; unsigned long high_mark; unsigned long grown; unsigned long reaped; unsigned long errors; unsigned long max_freeable; unsigned long node_allocs; unsigned long node_frees; unsigned long node_overflow; atomic_t allochit; atomic_t allocmiss; atomic_t freehit; atomic_t freemiss; #ifdef CONFIG_DEBUG_SLAB_LEAK atomic_t store_user_clean; #endif /* * If debugging is enabled, then the allocator can add additional * fields and/or padding to every object. size contains the total * object size including these internal fields, the following two * variables contain the offset to the user object and its size. */ int obj_offset; #endif /* CONFIG_DEBUG_SLAB */ #ifdef CONFIG_MEMCG struct memcg_cache_params memcg_params; #endif #ifdef CONFIG_KASAN struct kasan_cache kasan_info; #endif #ifdef CONFIG_SLAB_FREELIST_RANDOM void *random_seq; #endif struct kmem_cache_node *node[MAX_NUMNODES]; };
static inline void *nearest_obj(struct kmem_cache *cache, struct page *page, void *x) { void *object = x - (x - page->s_mem) % cache->size; void *last_object = page->s_mem + (cache->num - 1) * cache->size; if (unlikely(object > last_object)) return last_object; else return object; }Contributors
Person | Tokens | Prop | Commits | CommitProp | |
alexander potapenko | alexander potapenko | 76 | 100.00% | 1 | 100.00% |
Total | 76 | 100.00% | 1 | 100.00% |
Person | Tokens | Prop | Commits | CommitProp | |
pekka j enberg | pekka j enberg | 142 | 47.81% | 1 | 5.56% |
alexander potapenko | alexander potapenko | 85 | 28.62% | 1 | 5.56% |
christoph lameter | christoph lameter | 25 | 8.42% | 6 | 33.33% |
joonsoo kim | joonsoo kim | 16 | 5.39% | 3 | 16.67% |
thomas garnier | thomas garnier | 9 | 3.03% | 1 | 5.56% |
glauber costa | glauber costa | 9 | 3.03% | 2 | 11.11% |
hannes frederic sowa | hannes frederic sowa | 5 | 1.68% | 1 | 5.56% |
eric dumazet | eric dumazet | 4 | 1.35% | 1 | 5.56% |
fam zheng | fam zheng | 1 | 0.34% | 1 | 5.56% |
johannes weiner | johannes weiner | 1 | 0.34% | 1 | 5.56% |
Total | 297 | 100.00% | 18 | 100.00% |