Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
170 | 98.27% | 7 | 87.50% | |
3 | 1.73% | 1 | 12.50% | |
Total | 173 | 8 |
Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Filipe David Borba Manana | 170 | 98.27% | 7 | 87.50% |
Naohiro Aota | 3 | 1.73% | 1 | 12.50% |
Total | 173 | 8 |
/* SPDX-License-Identifier: GPL-2.0 */ #ifndef BTRFS_LRU_CACHE_H #define BTRFS_LRU_CACHE_H #include <linux/types.h> #include <linux/maple_tree.h> #include <linux/list.h> /* * A cache entry. This is meant to be embedded in a structure of a user of * this module. Similar to how struct list_head and struct rb_node are used. * * Note: it should be embedded as the first element in a struct (offset 0), and * this module assumes it was allocated with kmalloc(), so it calls kfree() when * it needs to free an entry. */ struct btrfs_lru_cache_entry { struct list_head lru_list; u64 key; /* * Optional generation associated to a key. Use 0 if not needed/used. * Entries with the same key and different generations are stored in a * linked list, so use this only for cases where there's a small number * of different generations. */ u64 gen; /* * The maple tree uses unsigned long type for the keys, which is 32 bits * on 32 bits systems, and 64 bits on 64 bits systems. So if we want to * use something like inode numbers as keys, which are always a u64, we * have to deal with this in a special way - we store the key in the * entry itself, as a u64, and the values inserted into the maple tree * are linked lists of entries - so in case we are on a 64 bits system, * that list always has a single entry, while on 32 bits systems it * may have more than one, with each entry having the same value for * their lower 32 bits of the u64 key. */ struct list_head list; }; struct btrfs_lru_cache { struct list_head lru_list; struct maple_tree entries; /* Number of entries stored in the cache. */ unsigned int size; /* Maximum number of entries the cache can have. */ unsigned int max_size; }; #define btrfs_lru_cache_for_each_entry_safe(cache, entry, tmp) \ list_for_each_entry_safe_reverse((entry), (tmp), &(cache)->lru_list, lru_list) static inline struct btrfs_lru_cache_entry *btrfs_lru_cache_lru_entry( struct btrfs_lru_cache *cache) { return list_first_entry_or_null(&cache->lru_list, struct btrfs_lru_cache_entry, lru_list); } void btrfs_lru_cache_init(struct btrfs_lru_cache *cache, unsigned int max_size); struct btrfs_lru_cache_entry *btrfs_lru_cache_lookup(struct btrfs_lru_cache *cache, u64 key, u64 gen); int btrfs_lru_cache_store(struct btrfs_lru_cache *cache, struct btrfs_lru_cache_entry *new_entry, gfp_t gfp); void btrfs_lru_cache_remove(struct btrfs_lru_cache *cache, struct btrfs_lru_cache_entry *entry); void btrfs_lru_cache_clear(struct btrfs_lru_cache *cache); #endif