cregit-Linux how code gets into the kernel

Release 4.7 drivers/gpu/drm/i915/i915_gem_context.c

/*
 * Copyright © 2011-2012 Intel Corporation
 *
 * Permission is hereby granted, free of charge, to any person obtaining a
 * copy of this software and associated documentation files (the "Software"),
 * to deal in the Software without restriction, including without limitation
 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
 * and/or sell copies of the Software, and to permit persons to whom the
 * Software is furnished to do so, subject to the following conditions:
 *
 * The above copyright notice and this permission notice (including the next
 * paragraph) shall be included in all copies or substantial portions of the
 * Software.
 *
 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
 * IN THE SOFTWARE.
 *
 * Authors:
 *    Ben Widawsky <ben@bwidawsk.net>
 *
 */

/*
 * This file implements HW context support. On gen5+ a HW context consists of an
 * opaque GPU object which is referenced at times of context saves and restores.
 * With RC6 enabled, the context is also referenced as the GPU enters and exists
 * from RC6 (GPU has it's own internal power context, except on gen5). Though
 * something like a context does exist for the media ring, the code only
 * supports contexts for the render ring.
 *
 * In software, there is a distinction between contexts created by the user,
 * and the default HW context. The default HW context is used by GPU clients
 * that do not request setup of their own hardware context. The default
 * context's state is never restored to help prevent programming errors. This
 * would happen if a client ran and piggy-backed off another clients GPU state.
 * The default context only exists to give the GPU some offset to load as the
 * current to invoke a save of the context we actually care about. In fact, the
 * code could likely be constructed, albeit in a more complicated fashion, to
 * never use the default context, though that limits the driver's ability to
 * swap out, and/or destroy other contexts.
 *
 * All other contexts are created as a request by the GPU client. These contexts
 * store GPU state, and thus allow GPU clients to not re-emit state (and
 * potentially query certain state) at any time. The kernel driver makes
 * certain that the appropriate commands are inserted.
 *
 * The context life cycle is semi-complicated in that context BOs may live
 * longer than the context itself because of the way the hardware, and object
 * tracking works. Below is a very crude representation of the state machine
 * describing the context life.
 *                                         refcount     pincount     active
 * S0: initial state                          0            0           0
 * S1: context created                        1            0           0
 * S2: context is currently running           2            1           X
 * S3: GPU referenced, but not current        2            0           1
 * S4: context is current, but destroyed      1            1           0
 * S5: like S3, but destroyed                 1            0           1
 *
 * The most common (but not all) transitions:
 * S0->S1: client creates a context
 * S1->S2: client submits execbuf with context
 * S2->S3: other clients submits execbuf with context
 * S3->S1: context object was retired
 * S3->S2: clients submits another execbuf
 * S2->S4: context destroy called with current context
 * S3->S5->S0: destroy path
 * S4->S5->S0: destroy path on current context
 *
 * There are two confusing terms used above:
 *  The "current context" means the context which is currently running on the
 *  GPU. The GPU has loaded its state already and has stored away the gtt
 *  offset of the BO. The GPU is not actively referencing the data at this
 *  offset, but it will on the next context switch. The only way to avoid this
 *  is to do a GPU reset.
 *
 *  An "active context' is one which was previously the "current context" and is
 *  on the active list waiting for the next context switch to occur. Until this
 *  happens, the object must remain at the same gtt offset. It is therefore
 *  possible to destroy a context, but it is still active.
 *
 */

#include <drm/drmP.h>
#include <drm/i915_drm.h>
#include "i915_drv.h"
#include "i915_trace.h"

/* This is a HW constraint. The value below is the largest known requirement
 * I've seen in a spec to date, and that was a workaround for a non-shipping
 * part. It should be safe to decrease this, but it's more future proof as is.
 */

#define GEN6_CONTEXT_ALIGN (64<<10)

#define GEN7_CONTEXT_ALIGN 4096


static size_t get_context_alignment(struct drm_device *dev) { if (IS_GEN6(dev)) return GEN6_CONTEXT_ALIGN; return GEN7_CONTEXT_ALIGN; }

Contributors

PersonTokensPropCommitsCommitProp
ben widawskyben widawsky24100.00%1100.00%
Total24100.00%1100.00%


static int get_context_size(struct drm_device *dev) { struct drm_i915_private *dev_priv = dev->dev_private; int ret; u32 reg; switch (INTEL_INFO(dev)->gen) { case 6: reg = I915_READ(CXT_SIZE); ret = GEN6_CXT_TOTAL_SIZE(reg) * 64; break; case 7: reg = I915_READ(GEN7_CXT_SIZE); if (IS_HASWELL(dev)) ret = HSW_CXT_TOTAL_SIZE; else ret = GEN7_CXT_TOTAL_SIZE(reg) * 64; break; case 8: ret = GEN8_CXT_TOTAL_SIZE; break; default: BUG(); } return ret; }

Contributors

PersonTokensPropCommitsCommitProp
ben widawskyben widawsky104100.00%4100.00%
Total104100.00%4100.00%


static void i915_gem_context_clean(struct intel_context *ctx) { struct i915_hw_ppgtt *ppgtt = ctx->ppgtt; struct i915_vma *vma, *next; if (!ppgtt) return; list_for_each_entry_safe(vma, next, &ppgtt->base.inactive_list, vm_link) { if (WARN_ON(__i915_vma_unbind_no_wait(vma))) break; } }

Contributors

PersonTokensPropCommitsCommitProp
tvrtko ursulintvrtko ursulin5698.25%150.00%
chris wilsonchris wilson11.75%150.00%
Total57100.00%2100.00%


void i915_gem_context_free(struct kref *ctx_ref) { struct intel_context *ctx = container_of(ctx_ref, typeof(*ctx), ref); trace_i915_context_free(ctx); if (i915.enable_execlists) intel_lr_context_free(ctx); /* * This context is going away and we need to remove all VMAs still * around. This is to handle imported shared objects for which * destructor did not run when their handles were closed. */ i915_gem_context_clean(ctx); i915_ppgtt_put(ctx->ppgtt); if (ctx->legacy_hw_ctx.rcs_state) drm_gem_object_unreference(&ctx->legacy_hw_ctx.rcs_state->base); list_del(&ctx->link); kfree(ctx); }

Contributors

PersonTokensPropCommitsCommitProp
ben widawskyben widawsky5156.67%436.36%
mika kuoppalamika kuoppala1516.67%19.09%
oscar mateooscar mateo1011.11%218.18%
tvrtko ursulintvrtko ursulin66.67%19.09%
daniele ceraolo spuriodaniele ceraolo spurio55.56%19.09%
chris wilsonchris wilson22.22%19.09%
daniel vetterdaniel vetter11.11%19.09%
Total90100.00%11100.00%


struct drm_i915_gem_object * i915_gem_alloc_context_obj(struct drm_device *dev, size_t size) { struct drm_i915_gem_object *obj; int ret; obj = i915_gem_alloc_object(dev, size); if (obj == NULL) return ERR_PTR(-ENOMEM); /* * Try to make the context utilize L3 as well as LLC. * * On VLV we don't have L3 controls in the PTEs so we * shouldn't touch the cache level, especially as that * would make the object snooped which might have a * negative performance impact. * * Snooping is required on non-llc platforms in execlist * mode, but since all GGTT accesses use PAT entry 0 we * get snooping anyway regardless of cache_level. * * This is only applicable for Ivy Bridge devices since * later platforms don't have L3 control bits in the PTE. */ if (IS_IVYBRIDGE(dev)) { ret = i915_gem_object_set_cache_level(obj, I915_CACHE_L3_LLC); /* Failure shouldn't ever happen this early */ if (WARN_ON(ret)) { drm_gem_object_unreference(&obj->base); return ERR_PTR(ret); } } return obj; }

Contributors

PersonTokensPropCommitsCommitProp
oscar mateooscar mateo8593.41%125.00%
wayne boyerwayne boyer55.49%250.00%
ben widawskyben widawsky11.10%125.00%
Total91100.00%4100.00%


static struct intel_context * __create_hw_context(struct drm_device *dev, struct drm_i915_file_private *file_priv) { struct drm_i915_private *dev_priv = dev->dev_private; struct intel_context *ctx; int ret; ctx = kzalloc(sizeof(*ctx), GFP_KERNEL); if (ctx == NULL) return ERR_PTR(-ENOMEM); kref_init(&ctx->ref); list_add_tail(&ctx->link, &dev_priv->context_list); ctx->i915 = dev_priv; if (dev_priv->hw_context_size) { struct drm_i915_gem_object *obj = i915_gem_alloc_context_obj(dev, dev_priv->hw_context_size); if (IS_ERR(obj)) { ret = PTR_ERR(obj); goto err_out; } ctx->legacy_hw_ctx.rcs_state = obj; } /* Default context will never have a file_priv */ if (file_priv != NULL) { ret = idr_alloc(&file_priv->context_idr, ctx, DEFAULT_CONTEXT_HANDLE, 0, GFP_KERNEL); if (ret < 0) goto err_out; } else ret = DEFAULT_CONTEXT_HANDLE; ctx->file_priv = file_priv; ctx->user_handle = ret; /* NB: Mark all slices as needing a remap so that when the context first * loads it will restore whatever remap state already exists. If there * is no remap info, it will be a NOP. */ ctx->remap_slice = (1 << NUM_L3_SLICES(dev)) - 1; ctx->hang_stats.ban_period_seconds = DRM_I915_CTX_BAN_PERIOD; return ctx; err_out: i915_gem_context_unreference(ctx); return ERR_PTR(ret); }

Contributors

PersonTokensPropCommitsCommitProp
ben widawskyben widawsky12453.91%531.25%
chris wilsonchris wilson5624.35%425.00%
oscar mateooscar mateo2410.43%425.00%
mika kuoppalamika kuoppala156.52%16.25%
tejun heotejun heo93.91%16.25%
ville syrjalaville syrjala20.87%16.25%
Total230100.00%16100.00%

/** * The default context needs to exist per ring that uses contexts. It stores the * context state of the GPU for applications that don't utilize HW contexts, as * well as an idle case. */
static struct intel_context * i915_gem_create_context(struct drm_device *dev, struct drm_i915_file_private *file_priv) { const bool is_global_default_ctx = file_priv == NULL; struct intel_context *ctx; int ret = 0; BUG_ON(!mutex_is_locked(&dev->struct_mutex)); ctx = __create_hw_context(dev, file_priv); if (IS_ERR(ctx)) return ctx; if (is_global_default_ctx && ctx->legacy_hw_ctx.rcs_state) { /* We may need to do things with the shrinker which * require us to immediately switch back to the default * context. This can cause a problem as pinning the * default context also requires GTT space which may not * be available. To avoid this we always pin the default * context. */ ret = i915_gem_obj_ggtt_pin(ctx->legacy_hw_ctx.rcs_state, get_context_alignment(dev), 0); if (ret) { DRM_DEBUG_DRIVER("Couldn't pin %d\n", ret); goto err_destroy; } } if (USES_FULL_PPGTT(dev)) { struct i915_hw_ppgtt *ppgtt = i915_ppgtt_create(dev, file_priv); if (IS_ERR_OR_NULL(ppgtt)) { DRM_DEBUG_DRIVER("PPGTT setup failed (%ld)\n", PTR_ERR(ppgtt)); ret = PTR_ERR(ppgtt); goto err_unpin; } ctx->ppgtt = ppgtt; } trace_i915_context_create(ctx); return ctx; err_unpin: if (is_global_default_ctx && ctx->legacy_hw_ctx.rcs_state) i915_gem_object_ggtt_unpin(ctx->legacy_hw_ctx.rcs_state); err_destroy: idr_remove(&file_priv->context_idr, ctx->user_handle); i915_gem_context_unreference(ctx); return ERR_PTR(ret); }

Contributors

PersonTokensPropCommitsCommitProp
ben widawskyben widawsky11651.79%942.86%
chris wilsonchris wilson8035.71%419.05%
oscar mateooscar mateo146.25%29.52%
daniel vetterdaniel vetter83.57%419.05%
daniele ceraolo spuriodaniele ceraolo spurio52.23%14.76%
mika kuoppalamika kuoppala10.45%14.76%
Total224100.00%21100.00%


static void i915_gem_context_unpin(struct intel_context *ctx, struct intel_engine_cs *engine) { if (i915.enable_execlists) { intel_lr_context_unpin(ctx, engine); } else { if (engine->id == RCS && ctx->legacy_hw_ctx.rcs_state) i915_gem_object_ggtt_unpin(ctx->legacy_hw_ctx.rcs_state); i915_gem_context_unreference(ctx); } }

Contributors

PersonTokensPropCommitsCommitProp
tvrtko ursulintvrtko ursulin62100.00%2100.00%
Total62100.00%2100.00%


void i915_gem_context_reset(struct drm_device *dev) { struct drm_i915_private *dev_priv = dev->dev_private; int i; if (i915.enable_execlists) { struct intel_context *ctx; list_for_each_entry(ctx, &dev_priv->context_list, link) intel_lr_context_reset(dev_priv, ctx); } for (i = 0; i < I915_NUM_ENGINES; i++) { struct intel_engine_cs *engine = &dev_priv->engine[i]; if (engine->last_context) { i915_gem_context_unpin(engine->last_context, engine); engine->last_context = NULL; } } /* Force the GPU state to be reinitialised on enabling */ dev_priv->kernel_context->legacy_hw_ctx.initialized = false; }

Contributors

PersonTokensPropCommitsCommitProp
ben widawskyben widawsky5448.21%17.14%
thomas danielthomas daniel2724.11%214.29%
tvrtko ursulintvrtko ursulin1311.61%535.71%
chris wilsonchris wilson119.82%214.29%
dave gordondave gordon32.68%17.14%
ville syrjalaville syrjala21.79%17.14%
oscar mateooscar mateo10.89%17.14%
alistair mcaulayalistair mcaulay10.89%17.14%
Total112100.00%14100.00%


int i915_gem_context_init(struct drm_device *dev) { struct drm_i915_private *dev_priv = dev->dev_private; struct intel_context *ctx; /* Init should only be called once per module load. Eventually the * restriction on the context_disabled check can be loosened. */ if (WARN_ON(dev_priv->kernel_context)) return 0; if (intel_vgpu_active(dev) && HAS_LOGICAL_RING_CONTEXTS(dev)) { if (!i915.enable_execlists) { DRM_INFO("Only EXECLIST mode is supported in vgpu.\n"); return -EINVAL; } } if (i915.enable_execlists) { /* NB: intentionally left blank. We will allocate our own * backing objects as we need them, thank you very much */ dev_priv->hw_context_size = 0; } else if (HAS_HW_CONTEXTS(dev)) { dev_priv->hw_context_size = round_up(get_context_size(dev), 4096); if (dev_priv->hw_context_size > (1<<20)) { DRM_DEBUG_DRIVER("Disabling HW Contexts; invalid size %d\n", dev_priv->hw_context_size); dev_priv->hw_context_size = 0; } } ctx = i915_gem_create_context(dev, NULL); if (IS_ERR(ctx)) { DRM_ERROR("Failed to create default global context (error %ld)\n", PTR_ERR(ctx)); return PTR_ERR(ctx); } dev_priv->kernel_context = ctx; DRM_DEBUG_DRIVER("%s context support initialized\n", i915.enable_execlists ? "LR" : dev_priv->hw_context_size ? "HW" : "fake"); return 0; }

Contributors

PersonTokensPropCommitsCommitProp
ben widawskyben widawsky10452.26%1066.67%
chris wilsonchris wilson3819.10%16.67%
zhiyuan lvzhiyuan lv3216.08%16.67%
oscar mateooscar mateo2311.56%213.33%
dave gordondave gordon21.01%16.67%
Total199100.00%15100.00%


void i915_gem_context_fini(struct drm_device *dev) { struct drm_i915_private *dev_priv = dev->dev_private; struct intel_context *dctx = dev_priv->kernel_context; int i; if (dctx->legacy_hw_ctx.rcs_state) { /* The only known way to stop the gpu from accessing the hw context is * to reset it. Do this as the very last operation to avoid confusing * other code, leading to spurious errors. */ intel_gpu_reset(dev, ALL_ENGINES); /* When default context is created and switched to, base object refcount * will be 2 (+1 from object creation and +1 from do_switch()). * i915_gem_context_fini() will be called after gpu_idle() has switched * to default context. So we need to unreference the base object once * to offset the do_switch part, so that i915_gem_context_unreference() * can then free the base object correctly. */ WARN_ON(!dev_priv->engine[RCS].last_context); i915_gem_object_ggtt_unpin(dctx->legacy_hw_ctx.rcs_state); } for (i = I915_NUM_ENGINES; --i >= 0;) { struct intel_engine_cs *engine = &dev_priv->engine[i]; if (engine->last_context) { i915_gem_context_unpin(engine->last_context, engine); engine->last_context = NULL; } } i915_gem_context_unreference(dctx); dev_priv->kernel_context = NULL; }

Contributors

PersonTokensPropCommitsCommitProp
ben widawskyben widawsky7757.89%423.53%
dave gordondave gordon129.02%15.88%
chris wilsonchris wilson129.02%211.76%
tvrtko ursulintvrtko ursulin107.52%423.53%
mika kuoppalamika kuoppala96.77%211.76%
oscar mateooscar mateo86.02%317.65%
daniel vetterdaniel vetter53.76%15.88%
Total133100.00%17100.00%


int i915_gem_context_enable(struct drm_i915_gem_request *req) { struct intel_engine_cs *engine = req->engine; int ret; if (i915.enable_execlists) { if (engine->init_context == NULL) return 0; ret = engine->init_context(req); } else ret = i915_switch_context(req); if (ret) { DRM_ERROR("ring init context: %d\n", ret); return ret; } return 0; }

Contributors

PersonTokensPropCommitsCommitProp
john harrisonjohn harrison2633.77%440.00%
ben widawskyben widawsky2532.47%110.00%
thomas danielthomas daniel2127.27%220.00%
tvrtko ursulintvrtko ursulin45.19%220.00%
chris wilsonchris wilson11.30%110.00%
Total77100.00%10100.00%


static int context_idr_cleanup(int id, void *p, void *data) { struct intel_context *ctx = p; i915_gem_context_unreference(ctx); return 0; }

Contributors

PersonTokensPropCommitsCommitProp
ben widawskyben widawsky2887.50%240.00%
daniel vetterdaniel vetter26.25%120.00%
mika kuoppalamika kuoppala13.12%120.00%
oscar mateooscar mateo13.12%120.00%
Total32100.00%5100.00%


int i915_gem_context_open(struct drm_device *dev, struct drm_file *file) { struct drm_i915_file_private *file_priv = file->driver_priv; struct intel_context *ctx; idr_init(&file_priv->context_idr); mutex_lock(&dev->struct_mutex); ctx = i915_gem_create_context(dev, file_priv); mutex_unlock(&dev->struct_mutex); if (IS_ERR(ctx)) { idr_destroy(&file_priv->context_idr); return PTR_ERR(ctx); } return 0; }

Contributors

PersonTokensPropCommitsCommitProp
ben widawskyben widawsky8090.91%266.67%
oscar mateooscar mateo89.09%133.33%
Total88100.00%3100.00%


void i915_gem_context_close(struct drm_device *dev, struct drm_file *file) { struct drm_i915_file_private *file_priv = file->driver_priv; idr_for_each(&file_priv->context_idr, context_idr_cleanup, NULL); idr_destroy(&file_priv->context_idr); }

Contributors

PersonTokensPropCommitsCommitProp
ben widawskyben widawsky4090.91%360.00%
chris wilsonchris wilson36.82%120.00%
daniel vetterdaniel vetter12.27%120.00%
Total44100.00%5100.00%


struct intel_context * i915_gem_context_get(struct drm_i915_file_private *file_priv, u32 id) { struct intel_context *ctx; ctx = (struct intel_context *)idr_find(&file_priv->context_idr, id); if (!ctx) return ERR_PTR(-ENOENT); return ctx; }

Contributors

PersonTokensPropCommitsCommitProp
ben widawskyben widawsky4994.23%375.00%
oscar mateooscar mateo35.77%125.00%
Total52100.00%4100.00%


static inline int mi_set_context(struct drm_i915_gem_request *req, u32 hw_flags) { struct intel_engine_cs *engine = req->engine; u32 flags = hw_flags | MI_MM_SPACE_GTT; const int num_rings = /* Use an extended w/a on ivb+ if signalling from other rings */ i915_semaphore_is_enabled(engine->dev) ? hweight32(INTEL_INFO(engine->dev)->ring_mask) - 1 : 0; int len, ret; /* w/a: If Flush TLB Invalidation Mode is enabled, driver must do a TLB * invalidation prior to MI_SET_CONTEXT. On GEN6 we don't set the value * explicitly, so we rely on the value at ring init, stored in * itlb_before_ctx_switch. */ if (IS_GEN6(engine->dev)) { ret = engine->flush(req, I915_GEM_GPU_DOMAINS, 0); if (ret) return ret; } /* These flags are for resource streamer on HSW+ */ if (IS_HASWELL(engine->dev) || INTEL_INFO(engine->dev)->gen >= 8) flags |= (HSW_MI_RS_SAVE_STATE_EN | HSW_MI_RS_RESTORE_STATE_EN); else if (INTEL_INFO(engine->dev)->gen < 8) flags |= (MI_SAVE_EXT_STATE_EN | MI_RESTORE_EXT_STATE_EN); len = 4; if (INTEL_INFO(engine->dev)->gen >= 7) len += 2 + (num_rings ? 4*num_rings + 6 : 0); ret = intel_ring_begin(req, len); if (ret) return ret; /* WaProgramMiArbOnOffAroundMiSetContext:ivb,vlv,hsw,bdw,chv */ if (INTEL_INFO(engine->dev)->gen >= 7) { intel_ring_emit(engine, MI_ARB_ON_OFF | MI_ARB_DISABLE); if (num_rings) { struct intel_engine_cs *signaller; intel_ring_emit(engine, MI_LOAD_REGISTER_IMM(num_rings)); for_each_engine(signaller, to_i915(engine->dev)) { if (signaller == engine) continue; intel_ring_emit_reg(engine, RING_PSMI_CTL(signaller->mmio_base)); intel_ring_emit(engine, _MASKED_BIT_ENABLE(GEN6_PSMI_SLEEP_MSG_DISABLE)); } } } intel_ring_emit(engine, MI_NOOP); intel_ring_emit(engine, MI_SET_CONTEXT); intel_ring_emit(engine, i915_gem_obj_ggtt_offset(req->ctx->legacy_hw_ctx.rcs_state) | flags); /* * w/a: MI_SET_CONTEXT must always be followed by MI_NOOP * WaMiSetContext_Hang:snb,ivb,vlv */ intel_ring_emit(engine, MI_NOOP); if (INTEL_INFO(engine->dev)->gen >= 7) { if (num_rings) { struct intel_engine_cs *signaller; i915_reg_t last_reg = {}; /* keep gcc quiet */ intel_ring_emit(engine, MI_LOAD_REGISTER_IMM(num_rings)); for_each_engine(signaller, to_i915(engine->dev)) { if (signaller == engine) continue; last_reg = RING_PSMI_CTL(signaller->mmio_base); intel_ring_emit_reg(engine, last_reg); intel_ring_emit(engine, _MASKED_BIT_DISABLE(GEN6_PSMI_SLEEP_MSG_DISABLE)); } /* Insert a delay before the next switch! */ intel_ring_emit(engine, MI_STORE_REGISTER_MEM | MI_SRM_LRM_GLOBAL_GTT); intel_ring_emit_reg(engine, last_reg); intel_ring_emit(engine, engine->scratch.gtt_offset); intel_ring_emit(engine, MI_NOOP); } intel_ring_emit(engine, MI_ARB_ON_OFF | MI_ARB_ENABLE); } intel_ring_advance(engine); return ret; }

Contributors

PersonTokensPropCommitsCommitProp
chris wilsonchris wilson21146.48%315.00%
ben widawskyben widawsky16135.46%525.00%
tvrtko ursulintvrtko ursulin306.61%315.00%
abdiel janulgueabdiel janulgue224.85%15.00%
john harrisonjohn harrison143.08%315.00%
ville syrjalaville syrjala132.86%420.00%
oscar mateooscar mateo30.66%15.00%
Total454100.00%20100.00%


static inline bool skip_rcs_switch(struct intel_engine_cs *engine, struct intel_context *to) { if (to->remap_slice) return false; if (!to->legacy_hw_ctx.initialized) return false; if (to->ppgtt && !(intel_engine_flag(engine) & to->ppgtt->pd_dirty_rings)) return false; return to == engine->last_context; }

Contributors

PersonTokensPropCommitsCommitProp
ben widawskyben widawsky4363.24%228.57%
chris wilsonchris wilson1927.94%228.57%
daniel vetterdaniel vetter34.41%114.29%
tvrtko ursulintvrtko ursulin34.41%228.57%
Total68100.00%7100.00%


static bool needs_pd_load_pre(struct intel_engine_cs *engine, struct intel_context *to) { if (!to->ppgtt) return false; if (engine->last_context == to && !(intel_engine_flag(engine) & to->ppgtt->pd_dirty_rings)) return false; if (engine->id != RCS) return true; if (INTEL_INFO(engine->dev)->gen < 8) return true; return false; }

Contributors

PersonTokensPropCommitsCommitProp
ben widawskyben widawsky4454.32%133.33%
chris wilsonchris wilson3441.98%133.33%
tvrtko ursulintvrtko ursulin33.70%133.33%
Total81100.00%3100.00%


static bool needs_pd_load_post(struct intel_context *to, u32 hw_flags) { if (!to->ppgtt) return false; if (!IS_GEN8(to->i915)) return false; if (hw_flags & MI_RESTORE_INHIBIT) return true; return false; }

Contributors

PersonTokensPropCommitsCommitProp
ben widawskyben widawsky4795.92%266.67%
chris wilsonchris wilson24.08%133.33%
Total49100.00%3100.00%


static int do_rcs_switch(struct drm_i915_gem_request *req) { struct intel_context *to = req->ctx; struct intel_engine_cs *engine = req->engine; struct intel_context *from; u32 hw_flags; int ret, i; if (skip_rcs_switch(engine, to)) return 0; /* Trying to pin first makes error handling easier. */ ret = i915_gem_obj_ggtt_pin(to->legacy_hw_ctx.rcs_state, get_context_alignment(engine->dev), 0); if (ret) return ret; /* * Pin can switch back to the default context if we end up calling into * evict_everything - as a last ditch gtt defrag effort that also * switches to the default context. Hence we need to reload from here. * * XXX: Doing so is painfully broken! */ from = engine->last_context; /* * Clear this page out of any CPU caches for coherent swap-in/out. Note * that thanks to write = false in this call and us not setting any gpu * write domains when putting a context object onto the active list * (when switching away from it), this won't block. * * XXX: We need a real interface to do this instead of trickery. */ ret = i915_gem_object_set_to_gtt_domain(to->legacy_hw_ctx.rcs_state, false); if (ret) goto unpin_out; if (needs_pd_load_pre(engine, to)) { /* Older GENs and non render rings still want the load first, * "PP_DCLV followed by PP_DIR_BASE register through Load * Register Immediate commands in Ring Buffer before submitting * a context."*/ trace_switch_mm(engine, to); ret = to->ppgtt->switch_mm(to->ppgtt, req); if (ret) goto unpin_out; } if (!to->legacy_hw_ctx.initialized || i915_gem_context_is_default(to)) /* NB: If we inhibit the restore, the context is not allowed to * die because future work may end up depending on valid address * space. This means we must enforce that a page table load * occur when this occurs. */ hw_flags = MI_RESTORE_INHIBIT; else if (to->ppgtt && intel_engine_flag(engine) & to->ppgtt->pd_dirty_rings) hw_flags = MI_FORCE_RESTORE; else hw_flags = 0; /* We should never emit switch_mm more than once */ WARN_ON(needs_pd_load_pre(engine, to) && needs_pd_load_post(to, hw_flags)); if (to != from || (hw_flags & MI_FORCE_RESTORE)) { ret = mi_set_context(req, hw_flags); if (ret) goto unpin_out; } /* The backing object for the context is done after switching to the * *next* context. Therefore we cannot retire the previous context until * the next context has already started running. In fact, the below code * is a bit suboptimal because the retiring can occur simply after the * MI_SET_CONTEXT instead of when the next seqno has completed. */ if (from != NULL) { from->legacy_hw_ctx.rcs_state->base.read_domains = I915_GEM_DOMAIN_INSTRUCTION; i915_vma_move_to_active(i915_gem_obj_to_ggtt(from->legacy_hw_ctx.rcs_state), req); /* As long as MI_SET_CONTEXT is serializing, ie. it flushes the * whole damn pipeline, we don't need to explicitly mark the * object dirty. The only exception is that the context must be * correct in case the object gets swapped out. Ideally we'd be * able to defer doing this until we know the object would be * swapped, but there is no way to do that yet. */ from->legacy_hw_ctx.rcs_state->dirty = 1; /* obj is kept alive until the next request by its active ref */ i915_gem_object_ggtt_unpin(from->legacy_hw_ctx.rcs_state); i915_gem_context_unreference(from); } i915_gem_context_reference(to); engine->last_context = to; /* GEN8 does *not* require an explicit reload if the PDPs have been * setup, and we do not wish to move them. */ if (needs_pd_load_post(to, hw_flags)) { trace_switch_mm(engine, to); ret = to->ppgtt->switch_mm(to->ppgtt, req); /* The hardware context switch is emitted, but we haven't * actually changed the state - so it's probably safe to bail * here. Still, let the user know something dangerous has * happened. */ if (ret) return ret; } if (to->ppgtt) to->ppgtt->pd_dirty_rings &= ~intel_engine_flag(engine); for (i = 0; i < MAX_L3_SLICES; i++) { if (!(to->remap_slice & (1<<i))) continue; ret = i915_gem_l3_remap(req, i); if (ret) return ret; to->remap_slice &= ~(1<<i); } if (!to->legacy_hw_ctx.initialized) { if (engine->init_context) { ret = engine->init_context(req); if (ret) return ret; } to->legacy_hw_ctx.initialized = true; } return 0; unpin_out: i915_gem_object_ggtt_unpin(to->legacy_hw_ctx.rcs_state); return ret; }

Contributors

PersonTokensPropCommitsCommitProp
ben widawskyben widawsky22846.06%926.47%
chris wilsonchris wilson17334.95%720.59%
john harrisonjohn harrison214.24%514.71%
arun siluveryarun siluvery193.84%12.94%
daniel vetterdaniel vetter183.64%514.71%
oscar mateooscar mateo142.83%25.88%
tvrtko ursulintvrtko ursulin132.63%38.82%
daniele ceraolo spuriodaniele ceraolo spurio61.21%12.94%
mika kuoppalamika kuoppala30.61%12.94%
Total495100.00%34100.00%

/** * i915_switch_context() - perform a GPU context switch. * @req: request for which we'll execute the context switch * * The context life cycle is simple. The context refcount is incremented and * decremented by 1 and create and destroy. If the context is in use by the GPU, * it will have a refcount > 1. This allows us to destroy the context abstract * object while letting the normal object tracking destroy the backing BO. * * This function should not be used in execlists mode. Instead the context is * switched by writing to the ELSP and requests keep a reference to their * context. */
int i915_switch_context(struct drm_i915_gem_request *req) { struct intel_engine_cs *engine = req->engine; struct drm_i915_private *dev_priv = req->i915; WARN_ON(i915.enable_execlists); WARN_ON(!mutex_is_locked(&dev_priv->dev->struct_mutex)); if (engine->id != RCS || req->ctx->legacy_hw_ctx.rcs_state == NULL) { struct intel_context *to = req->ctx; if (needs_pd_load_pre(engine, to)) { int ret; trace_switch_mm(engine, to); ret = to->ppgtt->switch_mm(to->ppgtt, req); if (ret) return ret; /* Doing a PD load always reloads the page dirs */ to->ppgtt->pd_dirty_rings &= ~intel_engine_flag(engine); } if (to != engine->last_context) { i915_gem_context_reference(to); if (engine->last_context) i915_gem_context_unreference(engine->last_context); engine->last_context = to; } return 0; } return do_rcs_switch(req); }

Contributors

PersonTokensPropCommitsCommitProp
chris wilsonchris wilson9553.67%321.43%
ben widawskyben widawsky4424.86%428.57%
john harrisonjohn harrison158.47%17.14%
tvrtko ursulintvrtko ursulin84.52%321.43%
thomas danielthomas daniel73.95%17.14%
mika kuoppalamika kuoppala52.82%17.14%
oscar mateooscar mateo31.69%17.14%
Total177100.00%14100.00%


static bool contexts_enabled(struct drm_device *dev) { return i915.enable_execlists || to_i915(dev)->hw_context_size; }

Contributors

PersonTokensPropCommitsCommitProp
chris wilsonchris wilson1878.26%150.00%
oscar mateooscar mateo521.74%150.00%
Total23100.00%2100.00%


int i915_gem_context_create_ioctl(struct drm_device *dev, void *data, struct drm_file *file) { struct drm_i915_gem_context_create *args = data; struct drm_i915_file_private *file_priv = file->driver_priv; struct intel_context *ctx; int ret; if (!contexts_enabled(dev)) return -ENODEV; if (args->pad != 0) return -EINVAL; ret = i915_mutex_lock_interruptible(dev); if (ret) return ret; ctx = i915_gem_create_context(dev, file_priv); mutex_unlock(&dev->struct_mutex); if (IS_ERR(ctx)) return PTR_ERR(ctx); args->ctx_id = ctx->user_handle; DRM_DEBUG_DRIVER("HW context %d created\n", args->ctx_id); return 0; }

Contributors

PersonTokensPropCommitsCommitProp
ben widawskyben widawsky9774.05%440.00%
dan carpenterdan carpenter1410.69%110.00%
chris wilsonchris wilson129.16%110.00%
daniel vetterdaniel vetter53.82%110.00%
oscar mateooscar mateo32.29%330.00%
Total131100.00%10100.00%


int i915_gem_context_destroy_ioctl(struct drm_device *dev, void *data, struct drm_file *file) { struct drm_i915_gem_context_destroy *args = data; struct drm_i915_file_private *file_priv = file->driver_priv; struct intel_context *ctx; int ret; if (args->pad != 0) return -EINVAL; if (args->ctx_id == DEFAULT_CONTEXT_HANDLE) return -ENOENT; ret = i915_mutex_lock_interruptible(dev); if (ret) return ret; ctx = i915_gem_context_get(file_priv, args->ctx_id); if (IS_ERR(ctx)) { mutex_unlock(&dev->struct_mutex); return PTR_ERR(ctx); } idr_remove(&ctx->file_priv->context_idr, ctx->user_handle); i915_gem_context_unreference(ctx); mutex_unlock(&dev->struct_mutex); DRM_DEBUG_DRIVER("HW context %d destroyed\n", args->ctx_id); return 0; }

Contributors

PersonTokensPropCommitsCommitProp
ben widawskyben widawsky12581.17%450.00%
mika kuoppalamika kuoppala149.09%112.50%
chris wilsonchris wilson127.79%112.50%
oscar mateooscar mateo31.95%225.00%
Total154100.00%8100.00%


int i915_gem_context_getparam_ioctl(struct drm_device *dev, void *data, struct drm_file *file) { struct drm_i915_file_private *file_priv = file->driver_priv; struct drm_i915_gem_context_param *args = data; struct intel_context *ctx; int ret; ret = i915_mutex_lock_interruptible(dev); if (ret) return ret; ctx = i915_gem_context_get(file_priv, args->ctx_id); if (IS_ERR(ctx)) { mutex_unlock(&dev->struct_mutex); return PTR_ERR(ctx); } args->size = 0; switch (args->param) { case I915_CONTEXT_PARAM_BAN_PERIOD: args->value = ctx->hang_stats.ban_period_seconds; break; case I915_CONTEXT_PARAM_NO_ZEROMAP: args->value = ctx->flags & CONTEXT_NO_ZEROMAP; break; case I915_CONTEXT_PARAM_GTT_SIZE: if (ctx->ppgtt) args->value = ctx->ppgtt->base.total; else if (to_i915(dev)->mm.aliasing_ppgtt) args->value = to_i915(dev)->mm.aliasing_ppgtt->base.total; else args->value = to_i915(dev)->ggtt.base.total; break; default: ret = -EINVAL; break; } mutex_unlock(&dev->struct_mutex); return ret; }

Contributors

PersonTokensPropCommitsCommitProp
chris wilsonchris wilson20393.12%250.00%
david weinehalldavid weinehall146.42%125.00%
joonas lahtinenjoonas lahtinen10.46%125.00%
Total218100.00%4100.00%


int i915_gem_context_setparam_ioctl(struct drm_device *dev, void *data, struct drm_file *file) { struct drm_i915_file_private *file_priv = file->driver_priv; struct drm_i915_gem_context_param *args = data; struct intel_context *ctx; int ret; ret = i915_mutex_lock_interruptible(dev); if (ret) return ret; ctx = i915_gem_context_get(file_priv, args->ctx_id); if (IS_ERR(ctx)) { mutex_unlock(&dev->struct_mutex); return PTR_ERR(ctx); } switch (args->param) { case I915_CONTEXT_PARAM_BAN_PERIOD: if (args->size) ret = -EINVAL; else if (args->value < ctx->hang_stats.ban_period_seconds && !capable(CAP_SYS_ADMIN)) ret = -EPERM; else ctx->hang_stats.ban_period_seconds = args->value; break; case I915_CONTEXT_PARAM_NO_ZEROMAP: if (args->size) { ret = -EINVAL; } else { ctx->flags &= ~CONTEXT_NO_ZEROMAP; ctx->flags |= args->value ? CONTEXT_NO_ZEROMAP : 0; } break; default: ret = -EINVAL; break; } mutex_unlock(&dev->struct_mutex); return ret; }

Contributors

PersonTokensPropCommitsCommitProp
chris wilsonchris wilson16781.07%150.00%
david weinehalldavid weinehall3918.93%150.00%
Total206100.00%2100.00%


Overall Contributors

PersonTokensPropCommitsCommitProp
ben widawskyben widawsky168445.51%4132.28%
chris wilsonchris wilson115031.08%2217.32%
tvrtko ursulintvrtko ursulin2085.62%97.09%
oscar mateooscar mateo2085.62%86.30%
john harrisonjohn harrison772.08%107.87%
mika kuoppalamika kuoppala631.70%43.15%
thomas danielthomas daniel551.49%32.36%
david weinehalldavid weinehall531.43%10.79%
daniel vetterdaniel vetter431.16%107.87%
zhiyuan lvzhiyuan lv320.86%10.79%
abdiel janulgueabdiel janulgue220.59%10.79%
daniele ceraolo spuriodaniele ceraolo spurio190.51%10.79%
arun siluveryarun siluvery190.51%10.79%
dave gordondave gordon170.46%10.79%
ville syrjalaville syrjala170.46%64.72%
dan carpenterdan carpenter140.38%10.79%
tejun heotejun heo90.24%10.79%
wayne boyerwayne boyer50.14%21.57%
david howellsdavid howells20.05%10.79%
joonas lahtinenjoonas lahtinen10.03%10.79%
alistair mcaulayalistair mcaulay10.03%10.79%
damien lespiaudamien lespiau10.03%10.79%
Total3700100.00%127100.00%
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
{% endraw %}