Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Matthew Brost | 928 | 91.97% | 1 | 6.25% |
Aravind Iddamsetty | 23 | 2.28% | 2 | 12.50% |
Lucas De Marchi | 22 | 2.18% | 6 | 37.50% |
Daniele Ceraolo Spurio | 13 | 1.29% | 1 | 6.25% |
Rodrigo Vivi | 12 | 1.19% | 4 | 25.00% |
Francois Dugast | 8 | 0.79% | 1 | 6.25% |
Matt Roper | 3 | 0.30% | 1 | 6.25% |
Total | 1009 | 16 |
// SPDX-License-Identifier: MIT /* * Copyright © 2022 Intel Corporation */ #include "xe_force_wake.h" #include <drm/drm_util.h> #include "regs/xe_gt_regs.h" #include "regs/xe_reg_defs.h" #include "xe_gt.h" #include "xe_mmio.h" #define XE_FORCE_WAKE_ACK_TIMEOUT_MS 50 static struct xe_gt * fw_to_gt(struct xe_force_wake *fw) { return fw->gt; } static struct xe_device * fw_to_xe(struct xe_force_wake *fw) { return gt_to_xe(fw_to_gt(fw)); } static void domain_init(struct xe_force_wake_domain *domain, enum xe_force_wake_domain_id id, struct xe_reg reg, struct xe_reg ack, u32 val, u32 mask) { domain->id = id; domain->reg_ctl = reg; domain->reg_ack = ack; domain->val = val; domain->mask = mask; } void xe_force_wake_init_gt(struct xe_gt *gt, struct xe_force_wake *fw) { struct xe_device *xe = gt_to_xe(gt); fw->gt = gt; spin_lock_init(&fw->lock); /* Assuming gen11+ so assert this assumption is correct */ xe_gt_assert(gt, GRAPHICS_VER(gt_to_xe(gt)) >= 11); if (xe->info.graphics_verx100 >= 1270) { domain_init(&fw->domains[XE_FW_DOMAIN_ID_GT], XE_FW_DOMAIN_ID_GT, FORCEWAKE_GT, FORCEWAKE_ACK_GT_MTL, BIT(0), BIT(16)); } else { domain_init(&fw->domains[XE_FW_DOMAIN_ID_GT], XE_FW_DOMAIN_ID_GT, FORCEWAKE_GT, FORCEWAKE_ACK_GT, BIT(0), BIT(16)); } } void xe_force_wake_init_engines(struct xe_gt *gt, struct xe_force_wake *fw) { int i, j; /* Assuming gen11+ so assert this assumption is correct */ xe_gt_assert(gt, GRAPHICS_VER(gt_to_xe(gt)) >= 11); if (!xe_gt_is_media_type(gt)) domain_init(&fw->domains[XE_FW_DOMAIN_ID_RENDER], XE_FW_DOMAIN_ID_RENDER, FORCEWAKE_RENDER, FORCEWAKE_ACK_RENDER, BIT(0), BIT(16)); for (i = XE_HW_ENGINE_VCS0, j = 0; i <= XE_HW_ENGINE_VCS7; ++i, ++j) { if (!(gt->info.engine_mask & BIT(i))) continue; domain_init(&fw->domains[XE_FW_DOMAIN_ID_MEDIA_VDBOX0 + j], XE_FW_DOMAIN_ID_MEDIA_VDBOX0 + j, FORCEWAKE_MEDIA_VDBOX(j), FORCEWAKE_ACK_MEDIA_VDBOX(j), BIT(0), BIT(16)); } for (i = XE_HW_ENGINE_VECS0, j = 0; i <= XE_HW_ENGINE_VECS3; ++i, ++j) { if (!(gt->info.engine_mask & BIT(i))) continue; domain_init(&fw->domains[XE_FW_DOMAIN_ID_MEDIA_VEBOX0 + j], XE_FW_DOMAIN_ID_MEDIA_VEBOX0 + j, FORCEWAKE_MEDIA_VEBOX(j), FORCEWAKE_ACK_MEDIA_VEBOX(j), BIT(0), BIT(16)); } if (gt->info.engine_mask & BIT(XE_HW_ENGINE_GSCCS0)) domain_init(&fw->domains[XE_FW_DOMAIN_ID_GSC], XE_FW_DOMAIN_ID_GSC, FORCEWAKE_GSC, FORCEWAKE_ACK_GSC, BIT(0), BIT(16)); } static void domain_wake(struct xe_gt *gt, struct xe_force_wake_domain *domain) { xe_mmio_write32(gt, domain->reg_ctl, domain->mask | domain->val); } static int domain_wake_wait(struct xe_gt *gt, struct xe_force_wake_domain *domain) { return xe_mmio_wait32(gt, domain->reg_ack, domain->val, domain->val, XE_FORCE_WAKE_ACK_TIMEOUT_MS * USEC_PER_MSEC, NULL, true); } static void domain_sleep(struct xe_gt *gt, struct xe_force_wake_domain *domain) { xe_mmio_write32(gt, domain->reg_ctl, domain->mask); } static int domain_sleep_wait(struct xe_gt *gt, struct xe_force_wake_domain *domain) { return xe_mmio_wait32(gt, domain->reg_ack, domain->val, 0, XE_FORCE_WAKE_ACK_TIMEOUT_MS * USEC_PER_MSEC, NULL, true); } #define for_each_fw_domain_masked(domain__, mask__, fw__, tmp__) \ for (tmp__ = (mask__); tmp__; tmp__ &= ~BIT(ffs(tmp__) - 1)) \ for_each_if((domain__ = ((fw__)->domains + \ (ffs(tmp__) - 1))) && \ domain__->reg_ctl.addr) int xe_force_wake_get(struct xe_force_wake *fw, enum xe_force_wake_domains domains) { struct xe_device *xe = fw_to_xe(fw); struct xe_gt *gt = fw_to_gt(fw); struct xe_force_wake_domain *domain; enum xe_force_wake_domains tmp, woken = 0; unsigned long flags; int ret, ret2 = 0; spin_lock_irqsave(&fw->lock, flags); for_each_fw_domain_masked(domain, domains, fw, tmp) { if (!domain->ref++) { woken |= BIT(domain->id); domain_wake(gt, domain); } } for_each_fw_domain_masked(domain, woken, fw, tmp) { ret = domain_wake_wait(gt, domain); ret2 |= ret; if (ret) drm_notice(&xe->drm, "Force wake domain (%d) failed to ack wake, ret=%d\n", domain->id, ret); } fw->awake_domains |= woken; spin_unlock_irqrestore(&fw->lock, flags); return ret2; } int xe_force_wake_put(struct xe_force_wake *fw, enum xe_force_wake_domains domains) { struct xe_device *xe = fw_to_xe(fw); struct xe_gt *gt = fw_to_gt(fw); struct xe_force_wake_domain *domain; enum xe_force_wake_domains tmp, sleep = 0; unsigned long flags; int ret, ret2 = 0; spin_lock_irqsave(&fw->lock, flags); for_each_fw_domain_masked(domain, domains, fw, tmp) { if (!--domain->ref) { sleep |= BIT(domain->id); domain_sleep(gt, domain); } } for_each_fw_domain_masked(domain, sleep, fw, tmp) { ret = domain_sleep_wait(gt, domain); ret2 |= ret; if (ret) drm_notice(&xe->drm, "Force wake domain (%d) failed to ack sleep, ret=%d\n", domain->id, ret); } fw->awake_domains &= ~sleep; spin_unlock_irqrestore(&fw->lock, flags); return ret2; }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1