Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Chris Wilson | 1368 | 87.75% | 21 | 84.00% |
Matthew Brost | 175 | 11.23% | 2 | 8.00% |
John Harrison | 13 | 0.83% | 1 | 4.00% |
Jani Nikula | 3 | 0.19% | 1 | 4.00% |
Total | 1559 | 25 |
// SPDX-License-Identifier: MIT /* * Copyright © 2019 Intel Corporation */ #include "i915_drv.h" #include "i915_request.h" #include "intel_context.h" #include "intel_engine_heartbeat.h" #include "intel_engine_pm.h" #include "intel_engine.h" #include "intel_gt.h" #include "intel_reset.h" /* * While the engine is active, we send a periodic pulse along the engine * to check on its health and to flush any idle-barriers. If that request * is stuck, and we fail to preempt it, we declare the engine hung and * issue a reset -- in the hope that restores progress. */ static bool next_heartbeat(struct intel_engine_cs *engine) { long delay; delay = READ_ONCE(engine->props.heartbeat_interval_ms); if (!delay) return false; delay = msecs_to_jiffies_timeout(delay); if (delay >= HZ) delay = round_jiffies_up_relative(delay); mod_delayed_work(system_highpri_wq, &engine->heartbeat.work, delay + 1); return true; } static struct i915_request * heartbeat_create(struct intel_context *ce, gfp_t gfp) { struct i915_request *rq; intel_context_enter(ce); rq = __i915_request_create(ce, gfp); intel_context_exit(ce); return rq; } static void idle_pulse(struct intel_engine_cs *engine, struct i915_request *rq) { engine->wakeref_serial = READ_ONCE(engine->serial) + 1; i915_request_add_active_barriers(rq); if (!engine->heartbeat.systole && intel_engine_has_heartbeat(engine)) engine->heartbeat.systole = i915_request_get(rq); } static void heartbeat_commit(struct i915_request *rq, const struct i915_sched_attr *attr) { idle_pulse(rq->engine, rq); __i915_request_commit(rq); __i915_request_queue(rq, attr); } static void show_heartbeat(const struct i915_request *rq, struct intel_engine_cs *engine) { struct drm_printer p = drm_debug_printer("heartbeat"); if (!rq) { intel_engine_dump(engine, &p, "%s heartbeat not ticking\n", engine->name); } else { intel_engine_dump(engine, &p, "%s heartbeat {seqno:%llx:%lld, prio:%d} not ticking\n", engine->name, rq->fence.context, rq->fence.seqno, rq->sched.attr.priority); } } static void reset_engine(struct intel_engine_cs *engine, struct i915_request *rq) { if (IS_ENABLED(CONFIG_DRM_I915_DEBUG_GEM)) show_heartbeat(rq, engine); if (intel_engine_uses_guc(engine)) /* * GuC itself is toast or GuC's hang detection * is disabled. Either way, need to find the * hang culprit manually. */ intel_guc_find_hung_context(engine); intel_gt_handle_error(engine->gt, engine->mask, I915_ERROR_CAPTURE, "stopped heartbeat on %s", engine->name); } static void heartbeat(struct work_struct *wrk) { struct i915_sched_attr attr = { .priority = I915_PRIORITY_MIN }; struct intel_engine_cs *engine = container_of(wrk, typeof(*engine), heartbeat.work.work); struct intel_context *ce = engine->kernel_context; struct i915_request *rq; unsigned long serial; /* Just in case everything has gone horribly wrong, give it a kick */ intel_engine_flush_submission(engine); rq = engine->heartbeat.systole; if (rq && i915_request_completed(rq)) { i915_request_put(rq); engine->heartbeat.systole = NULL; } if (!intel_engine_pm_get_if_awake(engine)) return; if (intel_gt_is_wedged(engine->gt)) goto out; if (i915_sched_engine_disabled(engine->sched_engine)) { reset_engine(engine, engine->heartbeat.systole); goto out; } if (engine->heartbeat.systole) { long delay = READ_ONCE(engine->props.heartbeat_interval_ms); /* Safeguard against too-fast worker invocations */ if (!time_after(jiffies, rq->emitted_jiffies + msecs_to_jiffies(delay))) goto out; if (!i915_sw_fence_signaled(&rq->submit)) { /* * Not yet submitted, system is stalled. * * This more often happens for ring submission, * where all contexts are funnelled into a common * ringbuffer. If one context is blocked on an * external fence, not only is it not submitted, * but all other contexts, including the kernel * context are stuck waiting for the signal. */ } else if (engine->sched_engine->schedule && rq->sched.attr.priority < I915_PRIORITY_BARRIER) { /* * Gradually raise the priority of the heartbeat to * give high priority work [which presumably desires * low latency and no jitter] the chance to naturally * complete before being preempted. */ attr.priority = 0; if (rq->sched.attr.priority >= attr.priority) attr.priority = I915_PRIORITY_HEARTBEAT; if (rq->sched.attr.priority >= attr.priority) attr.priority = I915_PRIORITY_BARRIER; local_bh_disable(); engine->sched_engine->schedule(rq, &attr); local_bh_enable(); } else { reset_engine(engine, rq); } rq->emitted_jiffies = jiffies; goto out; } serial = READ_ONCE(engine->serial); if (engine->wakeref_serial == serial) goto out; if (!mutex_trylock(&ce->timeline->mutex)) { /* Unable to lock the kernel timeline, is the engine stuck? */ if (xchg(&engine->heartbeat.blocked, serial) == serial) intel_gt_handle_error(engine->gt, engine->mask, I915_ERROR_CAPTURE, "no heartbeat on %s", engine->name); goto out; } rq = heartbeat_create(ce, GFP_NOWAIT | __GFP_NOWARN); if (IS_ERR(rq)) goto unlock; heartbeat_commit(rq, &attr); unlock: mutex_unlock(&ce->timeline->mutex); out: if (!engine->i915->params.enable_hangcheck || !next_heartbeat(engine)) i915_request_put(fetch_and_zero(&engine->heartbeat.systole)); intel_engine_pm_put(engine); } void intel_engine_unpark_heartbeat(struct intel_engine_cs *engine) { if (!CONFIG_DRM_I915_HEARTBEAT_INTERVAL) return; next_heartbeat(engine); } void intel_engine_park_heartbeat(struct intel_engine_cs *engine) { if (cancel_delayed_work(&engine->heartbeat.work)) i915_request_put(fetch_and_zero(&engine->heartbeat.systole)); } void intel_gt_unpark_heartbeats(struct intel_gt *gt) { struct intel_engine_cs *engine; enum intel_engine_id id; for_each_engine(engine, gt, id) if (intel_engine_pm_is_awake(engine)) intel_engine_unpark_heartbeat(engine); } void intel_gt_park_heartbeats(struct intel_gt *gt) { struct intel_engine_cs *engine; enum intel_engine_id id; for_each_engine(engine, gt, id) intel_engine_park_heartbeat(engine); } void intel_engine_init_heartbeat(struct intel_engine_cs *engine) { INIT_DELAYED_WORK(&engine->heartbeat.work, heartbeat); } static int __intel_engine_pulse(struct intel_engine_cs *engine) { struct i915_sched_attr attr = { .priority = I915_PRIORITY_BARRIER }; struct intel_context *ce = engine->kernel_context; struct i915_request *rq; lockdep_assert_held(&ce->timeline->mutex); GEM_BUG_ON(!intel_engine_has_preemption(engine)); GEM_BUG_ON(!intel_engine_pm_is_awake(engine)); rq = heartbeat_create(ce, GFP_NOWAIT | __GFP_NOWARN); if (IS_ERR(rq)) return PTR_ERR(rq); __set_bit(I915_FENCE_FLAG_SENTINEL, &rq->fence.flags); heartbeat_commit(rq, &attr); GEM_BUG_ON(rq->sched.attr.priority < I915_PRIORITY_BARRIER); return 0; } static unsigned long set_heartbeat(struct intel_engine_cs *engine, unsigned long delay) { unsigned long old; old = xchg(&engine->props.heartbeat_interval_ms, delay); if (delay) intel_engine_unpark_heartbeat(engine); else intel_engine_park_heartbeat(engine); return old; } int intel_engine_set_heartbeat(struct intel_engine_cs *engine, unsigned long delay) { struct intel_context *ce = engine->kernel_context; int err = 0; if (!delay && !intel_engine_has_preempt_reset(engine)) return -ENODEV; intel_engine_pm_get(engine); err = mutex_lock_interruptible(&ce->timeline->mutex); if (err) goto out_rpm; if (delay != engine->props.heartbeat_interval_ms) { unsigned long saved = set_heartbeat(engine, delay); /* recheck current execution */ if (intel_engine_has_preemption(engine)) { err = __intel_engine_pulse(engine); if (err) set_heartbeat(engine, saved); } } mutex_unlock(&ce->timeline->mutex); out_rpm: intel_engine_pm_put(engine); return err; } int intel_engine_pulse(struct intel_engine_cs *engine) { struct intel_context *ce = engine->kernel_context; int err; if (!intel_engine_has_preemption(engine)) return -ENODEV; if (!intel_engine_pm_get_if_awake(engine)) return 0; err = -EINTR; if (!mutex_lock_interruptible(&ce->timeline->mutex)) { err = __intel_engine_pulse(engine); mutex_unlock(&ce->timeline->mutex); } intel_engine_flush_submission(engine); intel_engine_pm_put(engine); return err; } int intel_engine_flush_barriers(struct intel_engine_cs *engine) { struct i915_sched_attr attr = { .priority = I915_PRIORITY_MIN }; struct intel_context *ce = engine->kernel_context; struct i915_request *rq; int err; if (llist_empty(&engine->barrier_tasks)) return 0; if (!intel_engine_pm_get_if_awake(engine)) return 0; if (mutex_lock_interruptible(&ce->timeline->mutex)) { err = -EINTR; goto out_rpm; } rq = heartbeat_create(ce, GFP_KERNEL); if (IS_ERR(rq)) { err = PTR_ERR(rq); goto out_unlock; } heartbeat_commit(rq, &attr); err = 0; out_unlock: mutex_unlock(&ce->timeline->mutex); out_rpm: intel_engine_pm_put(engine); return err; } #if IS_ENABLED(CONFIG_DRM_I915_SELFTEST) #include "selftest_engine_heartbeat.c" #endif
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1