Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Chris Wilson | 317 | 49.84% | 1 | 9.09% |
Thomas Hellstrom | 194 | 30.50% | 3 | 27.27% |
Tvrtko A. Ursulin | 71 | 11.16% | 2 | 18.18% |
Robert Beckett | 29 | 4.56% | 1 | 9.09% |
Matthew Auld | 25 | 3.93% | 4 | 36.36% |
Total | 636 | 11 |
/* * SPDX-License-Identifier: MIT * * Copyright © 2016 Intel Corporation */ #ifndef I915_SCATTERLIST_H #define I915_SCATTERLIST_H #include <linux/pfn.h> #include <linux/scatterlist.h> #include <linux/dma-mapping.h> #include <xen/xen.h> #include "i915_gem.h" struct drm_mm_node; struct ttm_resource; /* * Optimised SGL iterator for GEM objects */ static __always_inline struct sgt_iter { struct scatterlist *sgp; union { unsigned long pfn; dma_addr_t dma; }; unsigned int curr; unsigned int max; } __sgt_iter(struct scatterlist *sgl, bool dma) { struct sgt_iter s = { .sgp = sgl }; if (dma && s.sgp && sg_dma_len(s.sgp) == 0) { s.sgp = NULL; } else if (s.sgp) { s.max = s.curr = s.sgp->offset; if (dma) { s.dma = sg_dma_address(s.sgp); s.max += sg_dma_len(s.sgp); } else { s.pfn = page_to_pfn(sg_page(s.sgp)); s.max += s.sgp->length; } } return s; } static inline int __sg_page_count(const struct scatterlist *sg) { return sg->length >> PAGE_SHIFT; } static inline int __sg_dma_page_count(const struct scatterlist *sg) { return sg_dma_len(sg) >> PAGE_SHIFT; } static inline struct scatterlist *____sg_next(struct scatterlist *sg) { ++sg; if (unlikely(sg_is_chain(sg))) sg = sg_chain_ptr(sg); return sg; } /** * __sg_next - return the next scatterlist entry in a list * @sg: The current sg entry * * Description: * If the entry is the last, return NULL; otherwise, step to the next * element in the array (@sg@+1). If that's a chain pointer, follow it; * otherwise just return the pointer to the current element. **/ static inline struct scatterlist *__sg_next(struct scatterlist *sg) { return sg_is_last(sg) ? NULL : ____sg_next(sg); } /** * __for_each_sgt_daddr - iterate over the device addresses of the given sg_table * @__dp: Device address (output) * @__iter: 'struct sgt_iter' (iterator state, internal) * @__sgt: sg_table to iterate over (input) * @__step: step size */ #define __for_each_sgt_daddr(__dp, __iter, __sgt, __step) \ for ((__iter) = __sgt_iter((__sgt)->sgl, true); \ ((__dp) = (__iter).dma + (__iter).curr), (__iter).sgp; \ (((__iter).curr += (__step)) >= (__iter).max) ? \ (__iter) = __sgt_iter(__sg_next((__iter).sgp), true), 0 : 0) /** * for_each_sgt_page - iterate over the pages of the given sg_table * @__pp: page pointer (output) * @__iter: 'struct sgt_iter' (iterator state, internal) * @__sgt: sg_table to iterate over (input) */ #define for_each_sgt_page(__pp, __iter, __sgt) \ for ((__iter) = __sgt_iter((__sgt)->sgl, false); \ ((__pp) = (__iter).pfn == 0 ? NULL : \ pfn_to_page((__iter).pfn + ((__iter).curr >> PAGE_SHIFT))); \ (((__iter).curr += PAGE_SIZE) >= (__iter).max) ? \ (__iter) = __sgt_iter(__sg_next((__iter).sgp), false), 0 : 0) /** * i915_sg_dma_sizes - Record the dma segment sizes of a scatterlist * @sg: The scatterlist * * Return: An unsigned int with segment sizes logically or'ed together. * A caller can use this information to determine what hardware page table * entry sizes can be used to map the memory represented by the scatterlist. */ static inline unsigned int i915_sg_dma_sizes(struct scatterlist *sg) { unsigned int page_sizes; page_sizes = 0; while (sg && sg_dma_len(sg)) { GEM_BUG_ON(sg->offset); GEM_BUG_ON(!IS_ALIGNED(sg_dma_len(sg), PAGE_SIZE)); page_sizes |= sg_dma_len(sg); sg = __sg_next(sg); } return page_sizes; } static inline unsigned int i915_sg_segment_size(struct device *dev) { size_t max = min_t(size_t, UINT_MAX, dma_max_mapping_size(dev)); /* * For Xen PV guests pages aren't contiguous in DMA (machine) address * space. The DMA API takes care of that both in dma_alloc_* (by * calling into the hypervisor to make the pages contiguous) and in * dma_map_* (by bounce buffering). But i915 abuses ignores the * coherency aspects of the DMA API and thus can't cope with bounce * buffering actually happening, so add a hack here to force small * allocations and mappings when running in PV mode on Xen. * * Note this will still break if bounce buffering is required for other * reasons, like confidential computing hypervisors or PCIe root ports * with addressing limitations. */ if (xen_pv_domain()) max = PAGE_SIZE; return round_down(max, PAGE_SIZE); } bool i915_sg_trim(struct sg_table *orig_st); /** * struct i915_refct_sgt_ops - Operations structure for struct i915_refct_sgt */ struct i915_refct_sgt_ops { /** * release() - Free the memory of the struct i915_refct_sgt * @ref: struct kref that is embedded in the struct i915_refct_sgt */ void (*release)(struct kref *ref); }; /** * struct i915_refct_sgt - A refcounted scatter-gather table * @kref: struct kref for refcounting * @table: struct sg_table holding the scatter-gather table itself. Note that * @table->sgl = NULL can be used to determine whether a scatter-gather table * is present or not. * @size: The size in bytes of the underlying memory buffer * @ops: The operations structure. */ struct i915_refct_sgt { struct kref kref; struct sg_table table; size_t size; const struct i915_refct_sgt_ops *ops; }; /** * i915_refct_sgt_put - Put a refcounted sg-table * @rsgt the struct i915_refct_sgt to put. */ static inline void i915_refct_sgt_put(struct i915_refct_sgt *rsgt) { if (rsgt) kref_put(&rsgt->kref, rsgt->ops->release); } /** * i915_refct_sgt_get - Get a refcounted sg-table * @rsgt the struct i915_refct_sgt to get. */ static inline struct i915_refct_sgt * i915_refct_sgt_get(struct i915_refct_sgt *rsgt) { kref_get(&rsgt->kref); return rsgt; } /** * __i915_refct_sgt_init - Initialize a refcounted sg-list with a custom * operations structure * @rsgt The struct i915_refct_sgt to initialize. * @size: Size in bytes of the underlying memory buffer. * @ops: A customized operations structure in case the refcounted sg-list * is embedded into another structure. */ static inline void __i915_refct_sgt_init(struct i915_refct_sgt *rsgt, size_t size, const struct i915_refct_sgt_ops *ops) { kref_init(&rsgt->kref); rsgt->table.sgl = NULL; rsgt->size = size; rsgt->ops = ops; } void i915_refct_sgt_init(struct i915_refct_sgt *rsgt, size_t size); struct i915_refct_sgt *i915_rsgt_from_mm_node(const struct drm_mm_node *node, u64 region_start, u32 page_alignment); struct i915_refct_sgt *i915_rsgt_from_buddy_resource(struct ttm_resource *res, u64 region_start, u32 page_alignment); #endif
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1