Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Matthew Auld | 1717 | 99.71% | 5 | 83.33% |
Chris Wilson | 5 | 0.29% | 1 | 16.67% |
Total | 1722 | 6 |
// SPDX-License-Identifier: MIT /* * Copyright © 2019 Intel Corporation */ #include <linux/kmemleak.h> #include <linux/slab.h> #include "i915_buddy.h" #include "i915_gem.h" #include "i915_globals.h" #include "i915_utils.h" static struct i915_global_block { struct i915_global base; struct kmem_cache *slab_blocks; } global; static void i915_global_buddy_shrink(void) { kmem_cache_shrink(global.slab_blocks); } static void i915_global_buddy_exit(void) { kmem_cache_destroy(global.slab_blocks); } static struct i915_global_block global = { { .shrink = i915_global_buddy_shrink, .exit = i915_global_buddy_exit, } }; int __init i915_global_buddy_init(void) { global.slab_blocks = KMEM_CACHE(i915_buddy_block, SLAB_HWCACHE_ALIGN); if (!global.slab_blocks) return -ENOMEM; i915_global_register(&global.base); return 0; } static struct i915_buddy_block *i915_block_alloc(struct i915_buddy_block *parent, unsigned int order, u64 offset) { struct i915_buddy_block *block; block = kmem_cache_zalloc(global.slab_blocks, GFP_KERNEL); if (!block) return NULL; block->header = offset; block->header |= order; block->parent = parent; return block; } static void i915_block_free(struct i915_buddy_block *block) { kmem_cache_free(global.slab_blocks, block); } static void mark_allocated(struct i915_buddy_block *block) { block->header &= ~I915_BUDDY_HEADER_STATE; block->header |= I915_BUDDY_ALLOCATED; list_del(&block->link); } static void mark_free(struct i915_buddy_mm *mm, struct i915_buddy_block *block) { block->header &= ~I915_BUDDY_HEADER_STATE; block->header |= I915_BUDDY_FREE; list_add(&block->link, &mm->free_list[i915_buddy_block_order(block)]); } static void mark_split(struct i915_buddy_block *block) { block->header &= ~I915_BUDDY_HEADER_STATE; block->header |= I915_BUDDY_SPLIT; list_del(&block->link); } int i915_buddy_init(struct i915_buddy_mm *mm, u64 size, u64 chunk_size) { unsigned int i; u64 offset; if (size < chunk_size) return -EINVAL; if (chunk_size < PAGE_SIZE) return -EINVAL; if (!is_power_of_2(chunk_size)) return -EINVAL; size = round_down(size, chunk_size); mm->size = size; mm->chunk_size = chunk_size; mm->max_order = ilog2(size) - ilog2(chunk_size); GEM_BUG_ON(mm->max_order > I915_BUDDY_MAX_ORDER); mm->free_list = kmalloc_array(mm->max_order + 1, sizeof(struct list_head), GFP_KERNEL); if (!mm->free_list) return -ENOMEM; for (i = 0; i <= mm->max_order; ++i) INIT_LIST_HEAD(&mm->free_list[i]); mm->n_roots = hweight64(size); mm->roots = kmalloc_array(mm->n_roots, sizeof(struct i915_buddy_block *), GFP_KERNEL); if (!mm->roots) goto out_free_list; offset = 0; i = 0; /* * Split into power-of-two blocks, in case we are given a size that is * not itself a power-of-two. */ do { struct i915_buddy_block *root; unsigned int order; u64 root_size; root_size = rounddown_pow_of_two(size); order = ilog2(root_size) - ilog2(chunk_size); root = i915_block_alloc(NULL, order, offset); if (!root) goto out_free_roots; mark_free(mm, root); GEM_BUG_ON(i > mm->max_order); GEM_BUG_ON(i915_buddy_block_size(mm, root) < chunk_size); mm->roots[i] = root; offset += root_size; size -= root_size; i++; } while (size); return 0; out_free_roots: while (i--) i915_block_free(mm->roots[i]); kfree(mm->roots); out_free_list: kfree(mm->free_list); return -ENOMEM; } void i915_buddy_fini(struct i915_buddy_mm *mm) { int i; for (i = 0; i < mm->n_roots; ++i) { GEM_WARN_ON(!i915_buddy_block_is_free(mm->roots[i])); i915_block_free(mm->roots[i]); } kfree(mm->roots); kfree(mm->free_list); } static int split_block(struct i915_buddy_mm *mm, struct i915_buddy_block *block) { unsigned int block_order = i915_buddy_block_order(block) - 1; u64 offset = i915_buddy_block_offset(block); GEM_BUG_ON(!i915_buddy_block_is_free(block)); GEM_BUG_ON(!i915_buddy_block_order(block)); block->left = i915_block_alloc(block, block_order, offset); if (!block->left) return -ENOMEM; block->right = i915_block_alloc(block, block_order, offset + (mm->chunk_size << block_order)); if (!block->right) { i915_block_free(block->left); return -ENOMEM; } mark_free(mm, block->left); mark_free(mm, block->right); mark_split(block); return 0; } static struct i915_buddy_block * get_buddy(struct i915_buddy_block *block) { struct i915_buddy_block *parent; parent = block->parent; if (!parent) return NULL; if (parent->left == block) return parent->right; return parent->left; } static void __i915_buddy_free(struct i915_buddy_mm *mm, struct i915_buddy_block *block) { struct i915_buddy_block *parent; while ((parent = block->parent)) { struct i915_buddy_block *buddy; buddy = get_buddy(block); if (!i915_buddy_block_is_free(buddy)) break; list_del(&buddy->link); i915_block_free(block); i915_block_free(buddy); block = parent; } mark_free(mm, block); } void i915_buddy_free(struct i915_buddy_mm *mm, struct i915_buddy_block *block) { GEM_BUG_ON(!i915_buddy_block_is_allocated(block)); __i915_buddy_free(mm, block); } void i915_buddy_free_list(struct i915_buddy_mm *mm, struct list_head *objects) { struct i915_buddy_block *block, *on; list_for_each_entry_safe(block, on, objects, link) { i915_buddy_free(mm, block); cond_resched(); } INIT_LIST_HEAD(objects); } /* * Allocate power-of-two block. The order value here translates to: * * 0 = 2^0 * mm->chunk_size * 1 = 2^1 * mm->chunk_size * 2 = 2^2 * mm->chunk_size * ... */ struct i915_buddy_block * i915_buddy_alloc(struct i915_buddy_mm *mm, unsigned int order) { struct i915_buddy_block *block = NULL; unsigned int i; int err; for (i = order; i <= mm->max_order; ++i) { block = list_first_entry_or_null(&mm->free_list[i], struct i915_buddy_block, link); if (block) break; } if (!block) return ERR_PTR(-ENOSPC); GEM_BUG_ON(!i915_buddy_block_is_free(block)); while (i != order) { err = split_block(mm, block); if (unlikely(err)) goto out_free; /* Go low */ block = block->left; i--; } mark_allocated(block); kmemleak_update_trace(block); return block; out_free: if (i != order) __i915_buddy_free(mm, block); return ERR_PTR(err); } static inline bool overlaps(u64 s1, u64 e1, u64 s2, u64 e2) { return s1 <= e2 && e1 >= s2; } static inline bool contains(u64 s1, u64 e1, u64 s2, u64 e2) { return s1 <= s2 && e1 >= e2; } /* * Allocate range. Note that it's safe to chain together multiple alloc_ranges * with the same blocks list. * * Intended for pre-allocating portions of the address space, for example to * reserve a block for the initial framebuffer or similar, hence the expectation * here is that i915_buddy_alloc() is still the main vehicle for * allocations, so if that's not the case then the drm_mm range allocator is * probably a much better fit, and so you should probably go use that instead. */ int i915_buddy_alloc_range(struct i915_buddy_mm *mm, struct list_head *blocks, u64 start, u64 size) { struct i915_buddy_block *block; struct i915_buddy_block *buddy; LIST_HEAD(allocated); LIST_HEAD(dfs); u64 end; int err; int i; if (size < mm->chunk_size) return -EINVAL; if (!IS_ALIGNED(size | start, mm->chunk_size)) return -EINVAL; if (range_overflows(start, size, mm->size)) return -EINVAL; for (i = 0; i < mm->n_roots; ++i) list_add_tail(&mm->roots[i]->tmp_link, &dfs); end = start + size - 1; do { u64 block_start; u64 block_end; block = list_first_entry_or_null(&dfs, struct i915_buddy_block, tmp_link); if (!block) break; list_del(&block->tmp_link); block_start = i915_buddy_block_offset(block); block_end = block_start + i915_buddy_block_size(mm, block) - 1; if (!overlaps(start, end, block_start, block_end)) continue; if (i915_buddy_block_is_allocated(block)) { err = -ENOSPC; goto err_free; } if (contains(start, end, block_start, block_end)) { if (!i915_buddy_block_is_free(block)) { err = -ENOSPC; goto err_free; } mark_allocated(block); list_add_tail(&block->link, &allocated); continue; } if (!i915_buddy_block_is_split(block)) { err = split_block(mm, block); if (unlikely(err)) goto err_undo; } list_add(&block->right->tmp_link, &dfs); list_add(&block->left->tmp_link, &dfs); } while (1); list_splice_tail(&allocated, blocks); return 0; err_undo: /* * We really don't want to leave around a bunch of split blocks, since * bigger is better, so make sure we merge everything back before we * free the allocated blocks. */ buddy = get_buddy(block); if (buddy && (i915_buddy_block_is_free(block) && i915_buddy_block_is_free(buddy))) __i915_buddy_free(mm, block); err_free: i915_buddy_free_list(mm, &allocated); return err; } #if IS_ENABLED(CONFIG_DRM_I915_SELFTEST) #include "selftests/i915_buddy.c" #endif
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1