Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Matthew Auld | 942 | 70.56% | 5 | 50.00% |
Chris Wilson | 267 | 20.00% | 2 | 20.00% |
Lukasz Fiedorowicz | 74 | 5.54% | 1 | 10.00% |
Ramalingam C | 46 | 3.45% | 1 | 10.00% |
Wambui Karuga | 6 | 0.45% | 1 | 10.00% |
Total | 1335 | 10 |
// SPDX-License-Identifier: MIT /* * Copyright © 2019 Intel Corporation */ #include "intel_memory_region.h" #include "i915_drv.h" /* XXX: Hysterical raisins. BIT(inst) needs to just be (inst) at some point. */ #define REGION_MAP(type, inst) \ BIT((type) + INTEL_MEMORY_TYPE_SHIFT) | BIT(inst) const u32 intel_region_map[] = { [INTEL_REGION_SMEM] = REGION_MAP(INTEL_MEMORY_SYSTEM, 0), [INTEL_REGION_LMEM] = REGION_MAP(INTEL_MEMORY_LOCAL, 0), [INTEL_REGION_STOLEN] = REGION_MAP(INTEL_MEMORY_STOLEN, 0), }; struct intel_memory_region * intel_memory_region_by_type(struct drm_i915_private *i915, enum intel_memory_type mem_type) { struct intel_memory_region *mr; int id; for_each_memory_region(mr, i915, id) if (mr->type == mem_type) return mr; return NULL; } static u64 intel_memory_region_free_pages(struct intel_memory_region *mem, struct list_head *blocks) { struct i915_buddy_block *block, *on; u64 size = 0; list_for_each_entry_safe(block, on, blocks, link) { size += i915_buddy_block_size(&mem->mm, block); i915_buddy_free(&mem->mm, block); } INIT_LIST_HEAD(blocks); return size; } void __intel_memory_region_put_pages_buddy(struct intel_memory_region *mem, struct list_head *blocks) { mutex_lock(&mem->mm_lock); mem->avail += intel_memory_region_free_pages(mem, blocks); mutex_unlock(&mem->mm_lock); } void __intel_memory_region_put_block_buddy(struct i915_buddy_block *block) { struct list_head blocks; INIT_LIST_HEAD(&blocks); list_add(&block->link, &blocks); __intel_memory_region_put_pages_buddy(block->private, &blocks); } int __intel_memory_region_get_pages_buddy(struct intel_memory_region *mem, resource_size_t size, unsigned int flags, struct list_head *blocks) { unsigned int min_order = 0; unsigned long n_pages; GEM_BUG_ON(!IS_ALIGNED(size, mem->mm.chunk_size)); GEM_BUG_ON(!list_empty(blocks)); if (flags & I915_ALLOC_MIN_PAGE_SIZE) { min_order = ilog2(mem->min_page_size) - ilog2(mem->mm.chunk_size); } if (flags & I915_ALLOC_CONTIGUOUS) { size = roundup_pow_of_two(size); min_order = ilog2(size) - ilog2(mem->mm.chunk_size); } if (size > BIT(mem->mm.max_order) * mem->mm.chunk_size) return -E2BIG; n_pages = size >> ilog2(mem->mm.chunk_size); mutex_lock(&mem->mm_lock); do { struct i915_buddy_block *block; unsigned int order; order = fls(n_pages) - 1; GEM_BUG_ON(order > mem->mm.max_order); GEM_BUG_ON(order < min_order); do { block = i915_buddy_alloc(&mem->mm, order); if (!IS_ERR(block)) break; if (order-- == min_order) goto err_free_blocks; } while (1); n_pages -= BIT(order); block->private = mem; list_add(&block->link, blocks); if (!n_pages) break; } while (1); mem->avail -= size; mutex_unlock(&mem->mm_lock); return 0; err_free_blocks: intel_memory_region_free_pages(mem, blocks); mutex_unlock(&mem->mm_lock); return -ENXIO; } struct i915_buddy_block * __intel_memory_region_get_block_buddy(struct intel_memory_region *mem, resource_size_t size, unsigned int flags) { struct i915_buddy_block *block; LIST_HEAD(blocks); int ret; ret = __intel_memory_region_get_pages_buddy(mem, size, flags, &blocks); if (ret) return ERR_PTR(ret); block = list_first_entry(&blocks, typeof(*block), link); list_del_init(&block->link); return block; } int intel_memory_region_init_buddy(struct intel_memory_region *mem) { return i915_buddy_init(&mem->mm, resource_size(&mem->region), PAGE_SIZE); } void intel_memory_region_release_buddy(struct intel_memory_region *mem) { i915_buddy_fini(&mem->mm); } struct intel_memory_region * intel_memory_region_create(struct drm_i915_private *i915, resource_size_t start, resource_size_t size, resource_size_t min_page_size, resource_size_t io_start, const struct intel_memory_region_ops *ops) { struct intel_memory_region *mem; int err; mem = kzalloc(sizeof(*mem), GFP_KERNEL); if (!mem) return ERR_PTR(-ENOMEM); mem->i915 = i915; mem->region = (struct resource)DEFINE_RES_MEM(start, size); mem->io_start = io_start; mem->min_page_size = min_page_size; mem->ops = ops; mem->total = size; mem->avail = mem->total; mutex_init(&mem->objects.lock); INIT_LIST_HEAD(&mem->objects.list); INIT_LIST_HEAD(&mem->objects.purgeable); mutex_init(&mem->mm_lock); if (ops->init) { err = ops->init(mem); if (err) goto err_free; } kref_init(&mem->kref); return mem; err_free: kfree(mem); return ERR_PTR(err); } void intel_memory_region_set_name(struct intel_memory_region *mem, const char *fmt, ...) { va_list ap; va_start(ap, fmt); vsnprintf(mem->name, sizeof(mem->name), fmt, ap); va_end(ap); } static void __intel_memory_region_destroy(struct kref *kref) { struct intel_memory_region *mem = container_of(kref, typeof(*mem), kref); if (mem->ops->release) mem->ops->release(mem); mutex_destroy(&mem->mm_lock); mutex_destroy(&mem->objects.lock); kfree(mem); } struct intel_memory_region * intel_memory_region_get(struct intel_memory_region *mem) { kref_get(&mem->kref); return mem; } void intel_memory_region_put(struct intel_memory_region *mem) { kref_put(&mem->kref, __intel_memory_region_destroy); } /* Global memory region registration -- only slight layer inversions! */ int intel_memory_regions_hw_probe(struct drm_i915_private *i915) { int err, i; for (i = 0; i < ARRAY_SIZE(i915->mm.regions); i++) { struct intel_memory_region *mem = ERR_PTR(-ENODEV); u32 type; if (!HAS_REGION(i915, BIT(i))) continue; type = MEMORY_TYPE_FROM_REGION(intel_region_map[i]); switch (type) { case INTEL_MEMORY_SYSTEM: mem = i915_gem_shmem_setup(i915); break; case INTEL_MEMORY_STOLEN: mem = i915_gem_stolen_setup(i915); break; case INTEL_MEMORY_LOCAL: mem = intel_setup_fake_lmem(i915); break; } if (IS_ERR(mem)) { err = PTR_ERR(mem); drm_err(&i915->drm, "Failed to setup region(%d) type=%d\n", err, type); goto out_cleanup; } mem->id = intel_region_map[i]; mem->type = type; mem->instance = MEMORY_INSTANCE_FROM_REGION(intel_region_map[i]); i915->mm.regions[i] = mem; } return 0; out_cleanup: intel_memory_regions_driver_release(i915); return err; } void intel_memory_regions_driver_release(struct drm_i915_private *i915) { int i; for (i = 0; i < ARRAY_SIZE(i915->mm.regions); i++) { struct intel_memory_region *region = fetch_and_zero(&i915->mm.regions[i]); if (region) intel_memory_region_put(region); } } #if IS_ENABLED(CONFIG_DRM_I915_SELFTEST) #include "selftests/intel_memory_region.c" #include "selftests/mock_region.c" #endif
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1