Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Rob Clark | 843 | 97.12% | 13 | 76.47% |
Chris Wilson | 12 | 1.38% | 1 | 5.88% |
Jordan Crouse | 10 | 1.15% | 1 | 5.88% |
Thomas Gleixner | 2 | 0.23% | 1 | 5.88% |
Dmitry Eremin-Solenikov | 1 | 0.12% | 1 | 5.88% |
Total | 868 | 17 |
// SPDX-License-Identifier: GPL-2.0-only /* * Copyright (C) 2013-2016 Red Hat * Author: Rob Clark <robdclark@gmail.com> */ #include <linux/dma-fence.h> #include "msm_drv.h" #include "msm_fence.h" #include "msm_gpu.h" static struct msm_gpu *fctx2gpu(struct msm_fence_context *fctx) { struct msm_drm_private *priv = fctx->dev->dev_private; return priv->gpu; } static enum hrtimer_restart deadline_timer(struct hrtimer *t) { struct msm_fence_context *fctx = container_of(t, struct msm_fence_context, deadline_timer); kthread_queue_work(fctx2gpu(fctx)->worker, &fctx->deadline_work); return HRTIMER_NORESTART; } static void deadline_work(struct kthread_work *work) { struct msm_fence_context *fctx = container_of(work, struct msm_fence_context, deadline_work); /* If deadline fence has already passed, nothing to do: */ if (msm_fence_completed(fctx, fctx->next_deadline_fence)) return; msm_devfreq_boost(fctx2gpu(fctx), 2); } struct msm_fence_context * msm_fence_context_alloc(struct drm_device *dev, volatile uint32_t *fenceptr, const char *name) { struct msm_fence_context *fctx; static int index = 0; fctx = kzalloc(sizeof(*fctx), GFP_KERNEL); if (!fctx) return ERR_PTR(-ENOMEM); fctx->dev = dev; strscpy(fctx->name, name, sizeof(fctx->name)); fctx->context = dma_fence_context_alloc(1); fctx->index = index++; fctx->fenceptr = fenceptr; spin_lock_init(&fctx->spinlock); /* * Start out close to the 32b fence rollover point, so we can * catch bugs with fence comparisons. */ fctx->last_fence = 0xffffff00; fctx->completed_fence = fctx->last_fence; *fctx->fenceptr = fctx->last_fence; hrtimer_init(&fctx->deadline_timer, CLOCK_MONOTONIC, HRTIMER_MODE_ABS); fctx->deadline_timer.function = deadline_timer; kthread_init_work(&fctx->deadline_work, deadline_work); fctx->next_deadline = ktime_get(); return fctx; } void msm_fence_context_free(struct msm_fence_context *fctx) { kfree(fctx); } bool msm_fence_completed(struct msm_fence_context *fctx, uint32_t fence) { /* * Note: Check completed_fence first, as fenceptr is in a write-combine * mapping, so it will be more expensive to read. */ return (int32_t)(fctx->completed_fence - fence) >= 0 || (int32_t)(*fctx->fenceptr - fence) >= 0; } /* called from irq handler and workqueue (in recover path) */ void msm_update_fence(struct msm_fence_context *fctx, uint32_t fence) { unsigned long flags; spin_lock_irqsave(&fctx->spinlock, flags); if (fence_after(fence, fctx->completed_fence)) fctx->completed_fence = fence; if (msm_fence_completed(fctx, fctx->next_deadline_fence)) hrtimer_cancel(&fctx->deadline_timer); spin_unlock_irqrestore(&fctx->spinlock, flags); } struct msm_fence { struct dma_fence base; struct msm_fence_context *fctx; }; static inline struct msm_fence *to_msm_fence(struct dma_fence *fence) { return container_of(fence, struct msm_fence, base); } static const char *msm_fence_get_driver_name(struct dma_fence *fence) { return "msm"; } static const char *msm_fence_get_timeline_name(struct dma_fence *fence) { struct msm_fence *f = to_msm_fence(fence); return f->fctx->name; } static bool msm_fence_signaled(struct dma_fence *fence) { struct msm_fence *f = to_msm_fence(fence); return msm_fence_completed(f->fctx, f->base.seqno); } static void msm_fence_set_deadline(struct dma_fence *fence, ktime_t deadline) { struct msm_fence *f = to_msm_fence(fence); struct msm_fence_context *fctx = f->fctx; unsigned long flags; ktime_t now; spin_lock_irqsave(&fctx->spinlock, flags); now = ktime_get(); if (ktime_after(now, fctx->next_deadline) || ktime_before(deadline, fctx->next_deadline)) { fctx->next_deadline = deadline; fctx->next_deadline_fence = max(fctx->next_deadline_fence, (uint32_t)fence->seqno); /* * Set timer to trigger boost 3ms before deadline, or * if we are already less than 3ms before the deadline * schedule boost work immediately. */ deadline = ktime_sub(deadline, ms_to_ktime(3)); if (ktime_after(now, deadline)) { kthread_queue_work(fctx2gpu(fctx)->worker, &fctx->deadline_work); } else { hrtimer_start(&fctx->deadline_timer, deadline, HRTIMER_MODE_ABS); } } spin_unlock_irqrestore(&fctx->spinlock, flags); } static const struct dma_fence_ops msm_fence_ops = { .get_driver_name = msm_fence_get_driver_name, .get_timeline_name = msm_fence_get_timeline_name, .signaled = msm_fence_signaled, .set_deadline = msm_fence_set_deadline, }; struct dma_fence * msm_fence_alloc(void) { struct msm_fence *f; f = kzalloc(sizeof(*f), GFP_KERNEL); if (!f) return ERR_PTR(-ENOMEM); return &f->base; } void msm_fence_init(struct dma_fence *fence, struct msm_fence_context *fctx) { struct msm_fence *f = to_msm_fence(fence); f->fctx = fctx; /* * Until this point, the fence was just some pre-allocated memory, * no-one should have taken a reference to it yet. */ WARN_ON(kref_read(&fence->refcount)); dma_fence_init(&f->base, &msm_fence_ops, &fctx->spinlock, fctx->context, ++fctx->last_fence); }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1