#ifndef _LINUX_SLAB_DEF_H #define _LINUX_SLAB_DEF_H #include <linux/reciprocal_div.h> /* * Definitions unique to the original Linux SLAB allocator. */ struct kmem_cache { struct array_cache __percpu *cpu_cache; /* 1) Cache tunables. Protected by slab_mutex */ unsigned int batchcount; unsigned int limit; unsigned int shared; unsigned int size; struct reciprocal_value reciprocal_buffer_size; /* 2) touched by every alloc & free from the backend */ unsigned int flags; /* constant flags */ unsigned int num; /* # of objs per slab */ /* 3) cache_grow/shrink */ /* order of pgs per slab (2^n) */ unsigned int gfporder; /* force GFP flags, e.g. GFP_DMA */ gfp_t allocflags; size_t colour; /* cache colouring range */ unsigned int colour_off; /* colour offset */ struct kmem_cache *freelist_cache; unsigned int freelist_size; /* constructor func */ void (*ctor)(void *obj); /* 4) cache creation/removal */ const char *name; struct list_head list; int refcount; int object_size; int align; /* 5) statistics */ #ifdef CONFIG_DEBUG_SLAB unsigned long num_active; unsigned long num_allocations; unsigned long high_mark; unsigned long grown; unsigned long reaped; unsigned long errors; unsigned long max_freeable; unsigned long node_allocs; unsigned long node_frees; unsigned long node_overflow; atomic_t allochit; atomic_t allocmiss; atomic_t freehit; atomic_t freemiss; #ifdef CONFIG_DEBUG_SLAB_LEAK atomic_t store_user_clean; #endif /* * If debugging is enabled, then the allocator can add additional * fields and/or padding to every object. size contains the total * object size including these internal fields, the following two * variables contain the offset to the user object and its size. */ int obj_offset; #endif /* CONFIG_DEBUG_SLAB */ #ifdef CONFIG_MEMCG struct memcg_cache_params memcg_params; #endif #ifdef CONFIG_KASAN struct kasan_cache kasan_info; #endif #ifdef CONFIG_SLAB_FREELIST_RANDOM unsigned int *random_seq; #endif struct kmem_cache_node *node[MAX_NUMNODES]; };
static inline void *nearest_obj(struct kmem_cache *cache, struct page *page, void *x) { void *object = x - (x - page->s_mem) % cache->size; void *last_object = page->s_mem + (cache->num - 1) * cache->size; if (unlikely(object > last_object)) return last_object; else return object; }Contributors
Person | Tokens | Prop | Commits | CommitProp |
Alexander Potapenko | 76 | 100.00% | 1 | 100.00% |
Total | 76 | 100.00% | 1 | 100.00% |
Person | Tokens | Prop | Commits | CommitProp |
Pekka J Enberg | 142 | 47.65% | 1 | 5.26% |
Alexander Potapenko | 85 | 28.52% | 1 | 5.26% |
Christoph Lameter | 25 | 8.39% | 6 | 31.58% |
JoonSoo Kim | 16 | 5.37% | 3 | 15.79% |
Thomas Garnier | 10 | 3.36% | 2 | 10.53% |
Glauber de Oliveira Costa | 9 | 3.02% | 2 | 10.53% |
Hannes Frederic Sowa | 5 | 1.68% | 1 | 5.26% |
Eric Dumazet | 4 | 1.34% | 1 | 5.26% |
Johannes Weiner | 1 | 0.34% | 1 | 5.26% |
Fam Zheng | 1 | 0.34% | 1 | 5.26% |
Total | 298 | 100.00% | 19 | 100.00% |