Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Ben Skeggs | 730 | 91.82% | 20 | 76.92% |
Martin Peres | 47 | 5.91% | 4 | 15.38% |
Vince Hsu | 17 | 2.14% | 1 | 3.85% |
Alexandre Courbot | 1 | 0.13% | 1 | 3.85% |
Total | 795 | 26 |
/* * Copyright 2012 Red Hat Inc. * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice shall be included in * all copies or substantial portions of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR * OTHER DEALINGS IN THE SOFTWARE. * * Authors: Ben Skeggs */ #include "priv.h" s64 nvkm_timer_wait_test(struct nvkm_timer_wait *wait) { struct nvkm_subdev *subdev = &wait->tmr->subdev; u64 time = nvkm_timer_read(wait->tmr); if (wait->reads == 0) { wait->time0 = time; wait->time1 = time; } if (wait->time1 == time) { if (wait->reads++ == 16) { nvkm_fatal(subdev, "stalled at %016llx\n", time); return -ETIMEDOUT; } } else { wait->time1 = time; wait->reads = 1; } if (wait->time1 - wait->time0 > wait->limit) return -ETIMEDOUT; return wait->time1 - wait->time0; } void nvkm_timer_wait_init(struct nvkm_device *device, u64 nsec, struct nvkm_timer_wait *wait) { wait->tmr = device->timer; wait->limit = nsec; wait->reads = 0; } u64 nvkm_timer_read(struct nvkm_timer *tmr) { return tmr->func->read(tmr); } void nvkm_timer_alarm_trigger(struct nvkm_timer *tmr) { struct nvkm_alarm *alarm, *atemp; unsigned long flags; LIST_HEAD(exec); /* Process pending alarms. */ spin_lock_irqsave(&tmr->lock, flags); list_for_each_entry_safe(alarm, atemp, &tmr->alarms, head) { /* Have we hit the earliest alarm that hasn't gone off? */ if (alarm->timestamp > nvkm_timer_read(tmr)) { /* Schedule it. If we didn't race, we're done. */ tmr->func->alarm_init(tmr, alarm->timestamp); if (alarm->timestamp > nvkm_timer_read(tmr)) break; } /* Move to completed list. We'll drop the lock before * executing the callback so it can reschedule itself. */ list_del_init(&alarm->head); list_add(&alarm->exec, &exec); } /* Shut down interrupt if no more pending alarms. */ if (list_empty(&tmr->alarms)) tmr->func->alarm_fini(tmr); spin_unlock_irqrestore(&tmr->lock, flags); /* Execute completed callbacks. */ list_for_each_entry_safe(alarm, atemp, &exec, exec) { list_del(&alarm->exec); alarm->func(alarm); } } void nvkm_timer_alarm(struct nvkm_timer *tmr, u32 nsec, struct nvkm_alarm *alarm) { struct nvkm_alarm *list; unsigned long flags; /* Remove alarm from pending list. * * This both protects against the corruption of the list, * and implements alarm rescheduling/cancellation. */ spin_lock_irqsave(&tmr->lock, flags); list_del_init(&alarm->head); if (nsec) { /* Insert into pending list, ordered earliest to latest. */ alarm->timestamp = nvkm_timer_read(tmr) + nsec; list_for_each_entry(list, &tmr->alarms, head) { if (list->timestamp > alarm->timestamp) break; } list_add_tail(&alarm->head, &list->head); /* Update HW if this is now the earliest alarm. */ list = list_first_entry(&tmr->alarms, typeof(*list), head); if (list == alarm) { tmr->func->alarm_init(tmr, alarm->timestamp); /* This shouldn't happen if callers aren't stupid. * * Worst case scenario is that it'll take roughly * 4 seconds for the next alarm to trigger. */ WARN_ON(alarm->timestamp <= nvkm_timer_read(tmr)); } } spin_unlock_irqrestore(&tmr->lock, flags); } static void nvkm_timer_intr(struct nvkm_subdev *subdev) { struct nvkm_timer *tmr = nvkm_timer(subdev); tmr->func->intr(tmr); } static int nvkm_timer_fini(struct nvkm_subdev *subdev, bool suspend) { struct nvkm_timer *tmr = nvkm_timer(subdev); tmr->func->alarm_fini(tmr); return 0; } static int nvkm_timer_init(struct nvkm_subdev *subdev) { struct nvkm_timer *tmr = nvkm_timer(subdev); if (tmr->func->init) tmr->func->init(tmr); tmr->func->time(tmr, ktime_to_ns(ktime_get())); nvkm_timer_alarm_trigger(tmr); return 0; } static void * nvkm_timer_dtor(struct nvkm_subdev *subdev) { return nvkm_timer(subdev); } static const struct nvkm_subdev_func nvkm_timer = { .dtor = nvkm_timer_dtor, .init = nvkm_timer_init, .fini = nvkm_timer_fini, .intr = nvkm_timer_intr, }; int nvkm_timer_new_(const struct nvkm_timer_func *func, struct nvkm_device *device, enum nvkm_subdev_type type, int inst, struct nvkm_timer **ptmr) { struct nvkm_timer *tmr; if (!(tmr = *ptmr = kzalloc(sizeof(*tmr), GFP_KERNEL))) return -ENOMEM; nvkm_subdev_ctor(&nvkm_timer, device, type, inst, &tmr->subdev); tmr->func = func; INIT_LIST_HEAD(&tmr->alarms); spin_lock_init(&tmr->lock); return 0; }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1