Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Chris Wilson | 3070 | 98.33% | 1 | 16.67% |
Lionel Landwerlin | 16 | 0.51% | 1 | 16.67% |
Christian König | 16 | 0.51% | 1 | 16.67% |
Arvind Yadav | 15 | 0.48% | 1 | 16.67% |
Jason A. Donenfeld | 5 | 0.16% | 2 | 33.33% |
Total | 3122 | 6 |
// SPDX-License-Identifier: MIT /* * Copyright © 2019 Intel Corporation */ #include <linux/delay.h> #include <linux/dma-fence.h> #include <linux/dma-fence-chain.h> #include <linux/kernel.h> #include <linux/kthread.h> #include <linux/mm.h> #include <linux/sched/signal.h> #include <linux/slab.h> #include <linux/spinlock.h> #include <linux/random.h> #include "selftest.h" #define CHAIN_SZ (4 << 10) static struct kmem_cache *slab_fences; static inline struct mock_fence { struct dma_fence base; spinlock_t lock; } *to_mock_fence(struct dma_fence *f) { return container_of(f, struct mock_fence, base); } static const char *mock_name(struct dma_fence *f) { return "mock"; } static void mock_fence_release(struct dma_fence *f) { kmem_cache_free(slab_fences, to_mock_fence(f)); } static const struct dma_fence_ops mock_ops = { .get_driver_name = mock_name, .get_timeline_name = mock_name, .release = mock_fence_release, }; static struct dma_fence *mock_fence(void) { struct mock_fence *f; f = kmem_cache_alloc(slab_fences, GFP_KERNEL); if (!f) return NULL; spin_lock_init(&f->lock); dma_fence_init(&f->base, &mock_ops, &f->lock, 0, 0); return &f->base; } static struct dma_fence *mock_chain(struct dma_fence *prev, struct dma_fence *fence, u64 seqno) { struct dma_fence_chain *f; f = dma_fence_chain_alloc(); if (!f) return NULL; dma_fence_chain_init(f, dma_fence_get(prev), dma_fence_get(fence), seqno); return &f->base; } static int sanitycheck(void *arg) { struct dma_fence *f, *chain; int err = 0; f = mock_fence(); if (!f) return -ENOMEM; chain = mock_chain(NULL, f, 1); if (!chain) err = -ENOMEM; dma_fence_enable_sw_signaling(chain); dma_fence_signal(f); dma_fence_put(f); dma_fence_put(chain); return err; } struct fence_chains { unsigned int chain_length; struct dma_fence **fences; struct dma_fence **chains; struct dma_fence *tail; }; static uint64_t seqno_inc(unsigned int i) { return i + 1; } static int fence_chains_init(struct fence_chains *fc, unsigned int count, uint64_t (*seqno_fn)(unsigned int)) { unsigned int i; int err = 0; fc->chains = kvmalloc_array(count, sizeof(*fc->chains), GFP_KERNEL | __GFP_ZERO); if (!fc->chains) return -ENOMEM; fc->fences = kvmalloc_array(count, sizeof(*fc->fences), GFP_KERNEL | __GFP_ZERO); if (!fc->fences) { err = -ENOMEM; goto err_chains; } fc->tail = NULL; for (i = 0; i < count; i++) { fc->fences[i] = mock_fence(); if (!fc->fences[i]) { err = -ENOMEM; goto unwind; } fc->chains[i] = mock_chain(fc->tail, fc->fences[i], seqno_fn(i)); if (!fc->chains[i]) { err = -ENOMEM; goto unwind; } fc->tail = fc->chains[i]; dma_fence_enable_sw_signaling(fc->chains[i]); } fc->chain_length = i; return 0; unwind: for (i = 0; i < count; i++) { dma_fence_put(fc->fences[i]); dma_fence_put(fc->chains[i]); } kvfree(fc->fences); err_chains: kvfree(fc->chains); return err; } static void fence_chains_fini(struct fence_chains *fc) { unsigned int i; for (i = 0; i < fc->chain_length; i++) { dma_fence_signal(fc->fences[i]); dma_fence_put(fc->fences[i]); } kvfree(fc->fences); for (i = 0; i < fc->chain_length; i++) dma_fence_put(fc->chains[i]); kvfree(fc->chains); } static int find_seqno(void *arg) { struct fence_chains fc; struct dma_fence *fence; int err; int i; err = fence_chains_init(&fc, 64, seqno_inc); if (err) return err; fence = dma_fence_get(fc.tail); err = dma_fence_chain_find_seqno(&fence, 0); dma_fence_put(fence); if (err) { pr_err("Reported %d for find_seqno(0)!\n", err); goto err; } for (i = 0; i < fc.chain_length; i++) { fence = dma_fence_get(fc.tail); err = dma_fence_chain_find_seqno(&fence, i + 1); dma_fence_put(fence); if (err) { pr_err("Reported %d for find_seqno(%d:%d)!\n", err, fc.chain_length + 1, i + 1); goto err; } if (fence != fc.chains[i]) { pr_err("Incorrect fence reported by find_seqno(%d:%d)\n", fc.chain_length + 1, i + 1); err = -EINVAL; goto err; } dma_fence_get(fence); err = dma_fence_chain_find_seqno(&fence, i + 1); dma_fence_put(fence); if (err) { pr_err("Error reported for finding self\n"); goto err; } if (fence != fc.chains[i]) { pr_err("Incorrect fence reported by find self\n"); err = -EINVAL; goto err; } dma_fence_get(fence); err = dma_fence_chain_find_seqno(&fence, i + 2); dma_fence_put(fence); if (!err) { pr_err("Error not reported for future fence: find_seqno(%d:%d)!\n", i + 1, i + 2); err = -EINVAL; goto err; } dma_fence_get(fence); err = dma_fence_chain_find_seqno(&fence, i); dma_fence_put(fence); if (err) { pr_err("Error reported for previous fence!\n"); goto err; } if (i > 0 && fence != fc.chains[i - 1]) { pr_err("Incorrect fence reported by find_seqno(%d:%d)\n", i + 1, i); err = -EINVAL; goto err; } } err: fence_chains_fini(&fc); return err; } static int find_signaled(void *arg) { struct fence_chains fc; struct dma_fence *fence; int err; err = fence_chains_init(&fc, 2, seqno_inc); if (err) return err; dma_fence_signal(fc.fences[0]); fence = dma_fence_get(fc.tail); err = dma_fence_chain_find_seqno(&fence, 1); dma_fence_put(fence); if (err) { pr_err("Reported %d for find_seqno()!\n", err); goto err; } if (fence && fence != fc.chains[0]) { pr_err("Incorrect chain-fence.seqno:%lld reported for completed seqno:1\n", fence->seqno); dma_fence_get(fence); err = dma_fence_chain_find_seqno(&fence, 1); dma_fence_put(fence); if (err) pr_err("Reported %d for finding self!\n", err); err = -EINVAL; } err: fence_chains_fini(&fc); return err; } static int find_out_of_order(void *arg) { struct fence_chains fc; struct dma_fence *fence; int err; err = fence_chains_init(&fc, 3, seqno_inc); if (err) return err; dma_fence_signal(fc.fences[1]); fence = dma_fence_get(fc.tail); err = dma_fence_chain_find_seqno(&fence, 2); dma_fence_put(fence); if (err) { pr_err("Reported %d for find_seqno()!\n", err); goto err; } /* * We signaled the middle fence (2) of the 1-2-3 chain. The behavior * of the dma-fence-chain is to make us wait for all the fences up to * the point we want. Since fence 1 is still not signaled, this what * we should get as fence to wait upon (fence 2 being garbage * collected during the traversal of the chain). */ if (fence != fc.chains[0]) { pr_err("Incorrect chain-fence.seqno:%lld reported for completed seqno:2\n", fence ? fence->seqno : 0); err = -EINVAL; } err: fence_chains_fini(&fc); return err; } static uint64_t seqno_inc2(unsigned int i) { return 2 * i + 2; } static int find_gap(void *arg) { struct fence_chains fc; struct dma_fence *fence; int err; int i; err = fence_chains_init(&fc, 64, seqno_inc2); if (err) return err; for (i = 0; i < fc.chain_length; i++) { fence = dma_fence_get(fc.tail); err = dma_fence_chain_find_seqno(&fence, 2 * i + 1); dma_fence_put(fence); if (err) { pr_err("Reported %d for find_seqno(%d:%d)!\n", err, fc.chain_length + 1, 2 * i + 1); goto err; } if (fence != fc.chains[i]) { pr_err("Incorrect fence.seqno:%lld reported by find_seqno(%d:%d)\n", fence->seqno, fc.chain_length + 1, 2 * i + 1); err = -EINVAL; goto err; } dma_fence_get(fence); err = dma_fence_chain_find_seqno(&fence, 2 * i + 2); dma_fence_put(fence); if (err) { pr_err("Error reported for finding self\n"); goto err; } if (fence != fc.chains[i]) { pr_err("Incorrect fence reported by find self\n"); err = -EINVAL; goto err; } } err: fence_chains_fini(&fc); return err; } struct find_race { struct fence_chains fc; atomic_t children; }; static int __find_race(void *arg) { struct find_race *data = arg; int err = 0; while (!kthread_should_stop()) { struct dma_fence *fence = dma_fence_get(data->fc.tail); int seqno; seqno = get_random_u32_inclusive(1, data->fc.chain_length); err = dma_fence_chain_find_seqno(&fence, seqno); if (err) { pr_err("Failed to find fence seqno:%d\n", seqno); dma_fence_put(fence); break; } if (!fence) goto signal; /* * We can only find ourselves if we are on fence we were * looking for. */ if (fence->seqno == seqno) { err = dma_fence_chain_find_seqno(&fence, seqno); if (err) { pr_err("Reported an invalid fence for find-self:%d\n", seqno); dma_fence_put(fence); break; } } dma_fence_put(fence); signal: seqno = get_random_u32_below(data->fc.chain_length - 1); dma_fence_signal(data->fc.fences[seqno]); cond_resched(); } if (atomic_dec_and_test(&data->children)) wake_up_var(&data->children); return err; } static int find_race(void *arg) { struct find_race data; int ncpus = num_online_cpus(); struct task_struct **threads; unsigned long count; int err; int i; err = fence_chains_init(&data.fc, CHAIN_SZ, seqno_inc); if (err) return err; threads = kmalloc_array(ncpus, sizeof(*threads), GFP_KERNEL); if (!threads) { err = -ENOMEM; goto err; } atomic_set(&data.children, 0); for (i = 0; i < ncpus; i++) { threads[i] = kthread_run(__find_race, &data, "dmabuf/%d", i); if (IS_ERR(threads[i])) { ncpus = i; break; } atomic_inc(&data.children); get_task_struct(threads[i]); } wait_var_event_timeout(&data.children, !atomic_read(&data.children), 5 * HZ); for (i = 0; i < ncpus; i++) { int ret; ret = kthread_stop(threads[i]); if (ret && !err) err = ret; put_task_struct(threads[i]); } kfree(threads); count = 0; for (i = 0; i < data.fc.chain_length; i++) if (dma_fence_is_signaled(data.fc.fences[i])) count++; pr_info("Completed %lu cycles\n", count); err: fence_chains_fini(&data.fc); return err; } static int signal_forward(void *arg) { struct fence_chains fc; int err; int i; err = fence_chains_init(&fc, 64, seqno_inc); if (err) return err; for (i = 0; i < fc.chain_length; i++) { dma_fence_signal(fc.fences[i]); if (!dma_fence_is_signaled(fc.chains[i])) { pr_err("chain[%d] not signaled!\n", i); err = -EINVAL; goto err; } if (i + 1 < fc.chain_length && dma_fence_is_signaled(fc.chains[i + 1])) { pr_err("chain[%d] is signaled!\n", i); err = -EINVAL; goto err; } } err: fence_chains_fini(&fc); return err; } static int signal_backward(void *arg) { struct fence_chains fc; int err; int i; err = fence_chains_init(&fc, 64, seqno_inc); if (err) return err; for (i = fc.chain_length; i--; ) { dma_fence_signal(fc.fences[i]); if (i > 0 && dma_fence_is_signaled(fc.chains[i])) { pr_err("chain[%d] is signaled!\n", i); err = -EINVAL; goto err; } } for (i = 0; i < fc.chain_length; i++) { if (!dma_fence_is_signaled(fc.chains[i])) { pr_err("chain[%d] was not signaled!\n", i); err = -EINVAL; goto err; } } err: fence_chains_fini(&fc); return err; } static int __wait_fence_chains(void *arg) { struct fence_chains *fc = arg; if (dma_fence_wait(fc->tail, false)) return -EIO; return 0; } static int wait_forward(void *arg) { struct fence_chains fc; struct task_struct *tsk; int err; int i; err = fence_chains_init(&fc, CHAIN_SZ, seqno_inc); if (err) return err; tsk = kthread_run(__wait_fence_chains, &fc, "dmabuf/wait"); if (IS_ERR(tsk)) { err = PTR_ERR(tsk); goto err; } get_task_struct(tsk); yield_to(tsk, true); for (i = 0; i < fc.chain_length; i++) dma_fence_signal(fc.fences[i]); err = kthread_stop(tsk); put_task_struct(tsk); err: fence_chains_fini(&fc); return err; } static int wait_backward(void *arg) { struct fence_chains fc; struct task_struct *tsk; int err; int i; err = fence_chains_init(&fc, CHAIN_SZ, seqno_inc); if (err) return err; tsk = kthread_run(__wait_fence_chains, &fc, "dmabuf/wait"); if (IS_ERR(tsk)) { err = PTR_ERR(tsk); goto err; } get_task_struct(tsk); yield_to(tsk, true); for (i = fc.chain_length; i--; ) dma_fence_signal(fc.fences[i]); err = kthread_stop(tsk); put_task_struct(tsk); err: fence_chains_fini(&fc); return err; } static void randomise_fences(struct fence_chains *fc) { unsigned int count = fc->chain_length; /* Fisher-Yates shuffle courtesy of Knuth */ while (--count) { unsigned int swp; swp = get_random_u32_below(count + 1); if (swp == count) continue; swap(fc->fences[count], fc->fences[swp]); } } static int wait_random(void *arg) { struct fence_chains fc; struct task_struct *tsk; int err; int i; err = fence_chains_init(&fc, CHAIN_SZ, seqno_inc); if (err) return err; randomise_fences(&fc); tsk = kthread_run(__wait_fence_chains, &fc, "dmabuf/wait"); if (IS_ERR(tsk)) { err = PTR_ERR(tsk); goto err; } get_task_struct(tsk); yield_to(tsk, true); for (i = 0; i < fc.chain_length; i++) dma_fence_signal(fc.fences[i]); err = kthread_stop(tsk); put_task_struct(tsk); err: fence_chains_fini(&fc); return err; } int dma_fence_chain(void) { static const struct subtest tests[] = { SUBTEST(sanitycheck), SUBTEST(find_seqno), SUBTEST(find_signaled), SUBTEST(find_out_of_order), SUBTEST(find_gap), SUBTEST(find_race), SUBTEST(signal_forward), SUBTEST(signal_backward), SUBTEST(wait_forward), SUBTEST(wait_backward), SUBTEST(wait_random), }; int ret; pr_info("sizeof(dma_fence_chain)=%zu\n", sizeof(struct dma_fence_chain)); slab_fences = KMEM_CACHE(mock_fence, SLAB_TYPESAFE_BY_RCU | SLAB_HWCACHE_ALIGN); if (!slab_fences) return -ENOMEM; ret = subtests(tests, NULL); kmem_cache_destroy(slab_fences); return ret; }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1