Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Chris Wilson | 1098 | 99.64% | 7 | 87.50% |
Michał Winiarski | 4 | 0.36% | 1 | 12.50% |
Total | 1102 | 8 |
/* * SPDX-License-Identifier: MIT * * Copyright © 2019 Intel Corporation */ #include "intel_context.h" #include "intel_engine_pm.h" #include "intel_gt_requests.h" #include "intel_ring.h" #include "selftest_rc6.h" #include "selftests/i915_random.h" #include "selftests/librapl.h" static u64 rc6_residency(struct intel_rc6 *rc6) { u64 result; /* XXX VLV_GT_MEDIA_RC6? */ result = intel_rc6_residency_ns(rc6, GEN6_GT_GFX_RC6); if (HAS_RC6p(rc6_to_i915(rc6))) result += intel_rc6_residency_ns(rc6, GEN6_GT_GFX_RC6p); if (HAS_RC6pp(rc6_to_i915(rc6))) result += intel_rc6_residency_ns(rc6, GEN6_GT_GFX_RC6pp); return result; } int live_rc6_manual(void *arg) { struct intel_gt *gt = arg; struct intel_rc6 *rc6 = >->rc6; u64 rc0_power, rc6_power; intel_wakeref_t wakeref; ktime_t dt; u64 res[2]; int err = 0; /* * Our claim is that we can "encourage" the GPU to enter rc6 at will. * Let's try it! */ if (!rc6->enabled) return 0; /* bsw/byt use a PCU and decouple RC6 from our manual control */ if (IS_VALLEYVIEW(gt->i915) || IS_CHERRYVIEW(gt->i915)) return 0; wakeref = intel_runtime_pm_get(gt->uncore->rpm); /* Force RC6 off for starters */ __intel_rc6_disable(rc6); msleep(1); /* wakeup is not immediate, takes about 100us on icl */ res[0] = rc6_residency(rc6); dt = ktime_get(); rc0_power = librapl_energy_uJ(); msleep(250); rc0_power = librapl_energy_uJ() - rc0_power; dt = ktime_sub(ktime_get(), dt); res[1] = rc6_residency(rc6); if ((res[1] - res[0]) >> 10) { pr_err("RC6 residency increased by %lldus while disabled for 250ms!\n", (res[1] - res[0]) >> 10); err = -EINVAL; goto out_unlock; } rc0_power = div64_u64(NSEC_PER_SEC * rc0_power, ktime_to_ns(dt)); if (!rc0_power) { pr_err("No power measured while in RC0\n"); err = -EINVAL; goto out_unlock; } /* Manually enter RC6 */ intel_rc6_park(rc6); res[0] = rc6_residency(rc6); intel_uncore_forcewake_flush(rc6_to_uncore(rc6), FORCEWAKE_ALL); dt = ktime_get(); rc6_power = librapl_energy_uJ(); msleep(100); rc6_power = librapl_energy_uJ() - rc6_power; dt = ktime_sub(ktime_get(), dt); res[1] = rc6_residency(rc6); if (res[1] == res[0]) { pr_err("Did not enter RC6! RC6_STATE=%08x, RC6_CONTROL=%08x, residency=%lld\n", intel_uncore_read_fw(gt->uncore, GEN6_RC_STATE), intel_uncore_read_fw(gt->uncore, GEN6_RC_CONTROL), res[0]); err = -EINVAL; } rc6_power = div64_u64(NSEC_PER_SEC * rc6_power, ktime_to_ns(dt)); pr_info("GPU consumed %llduW in RC0 and %llduW in RC6\n", rc0_power, rc6_power); if (2 * rc6_power > rc0_power) { pr_err("GPU leaked energy while in RC6!\n"); err = -EINVAL; goto out_unlock; } /* Restore what should have been the original state! */ intel_rc6_unpark(rc6); out_unlock: intel_runtime_pm_put(gt->uncore->rpm, wakeref); return err; } static const u32 *__live_rc6_ctx(struct intel_context *ce) { struct i915_request *rq; const u32 *result; u32 cmd; u32 *cs; rq = intel_context_create_request(ce); if (IS_ERR(rq)) return ERR_CAST(rq); cs = intel_ring_begin(rq, 4); if (IS_ERR(cs)) { i915_request_add(rq); return cs; } cmd = MI_STORE_REGISTER_MEM | MI_USE_GGTT; if (INTEL_GEN(rq->engine->i915) >= 8) cmd++; *cs++ = cmd; *cs++ = i915_mmio_reg_offset(GEN8_RC6_CTX_INFO); *cs++ = ce->timeline->hwsp_offset + 8; *cs++ = 0; intel_ring_advance(rq, cs); result = rq->hwsp_seqno + 2; i915_request_add(rq); return result; } static struct intel_engine_cs ** randomised_engines(struct intel_gt *gt, struct rnd_state *prng, unsigned int *count) { struct intel_engine_cs *engine, **engines; enum intel_engine_id id; int n; n = 0; for_each_engine(engine, gt, id) n++; if (!n) return NULL; engines = kmalloc_array(n, sizeof(*engines), GFP_KERNEL); if (!engines) return NULL; n = 0; for_each_engine(engine, gt, id) engines[n++] = engine; i915_prandom_shuffle(engines, sizeof(*engines), n, prng); *count = n; return engines; } int live_rc6_ctx_wa(void *arg) { struct intel_gt *gt = arg; struct intel_engine_cs **engines; unsigned int n, count; I915_RND_STATE(prng); int err = 0; /* A read of CTX_INFO upsets rc6. Poke the bear! */ if (INTEL_GEN(gt->i915) < 8) return 0; engines = randomised_engines(gt, &prng, &count); if (!engines) return 0; for (n = 0; n < count; n++) { struct intel_engine_cs *engine = engines[n]; int pass; for (pass = 0; pass < 2; pass++) { struct i915_gpu_error *error = >->i915->gpu_error; struct intel_context *ce; unsigned int resets = i915_reset_engine_count(error, engine); const u32 *res; /* Use a sacrifical context */ ce = intel_context_create(engine); if (IS_ERR(ce)) { err = PTR_ERR(ce); goto out; } intel_engine_pm_get(engine); res = __live_rc6_ctx(ce); intel_engine_pm_put(engine); intel_context_put(ce); if (IS_ERR(res)) { err = PTR_ERR(res); goto out; } if (intel_gt_wait_for_idle(gt, HZ / 5) == -ETIME) { intel_gt_set_wedged(gt); err = -ETIME; goto out; } intel_gt_pm_wait_for_idle(gt); pr_debug("%s: CTX_INFO=%0x\n", engine->name, READ_ONCE(*res)); if (resets != i915_reset_engine_count(error, engine)) { pr_err("%s: GPU reset required\n", engine->name); add_taint_for_CI(gt->i915, TAINT_WARN); err = -EIO; goto out; } } } out: kfree(engines); return err; }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1