Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Chris Wilson | 694 | 99.71% | 2 | 66.67% |
Logan Gunthorpe | 2 | 0.29% | 1 | 33.33% |
Total | 696 | 3 |
/* * SPDX-License-Identifier: MIT * * Copyright © 2016 Intel Corporation */ #include "mock_dmabuf.h" static struct sg_table *mock_map_dma_buf(struct dma_buf_attachment *attachment, enum dma_data_direction dir) { struct mock_dmabuf *mock = to_mock(attachment->dmabuf); struct sg_table *st; struct scatterlist *sg; int i, err; st = kmalloc(sizeof(*st), GFP_KERNEL); if (!st) return ERR_PTR(-ENOMEM); err = sg_alloc_table(st, mock->npages, GFP_KERNEL); if (err) goto err_free; sg = st->sgl; for (i = 0; i < mock->npages; i++) { sg_set_page(sg, mock->pages[i], PAGE_SIZE, 0); sg = sg_next(sg); } if (!dma_map_sg(attachment->dev, st->sgl, st->nents, dir)) { err = -ENOMEM; goto err_st; } return st; err_st: sg_free_table(st); err_free: kfree(st); return ERR_PTR(err); } static void mock_unmap_dma_buf(struct dma_buf_attachment *attachment, struct sg_table *st, enum dma_data_direction dir) { dma_unmap_sg(attachment->dev, st->sgl, st->nents, dir); sg_free_table(st); kfree(st); } static void mock_dmabuf_release(struct dma_buf *dma_buf) { struct mock_dmabuf *mock = to_mock(dma_buf); int i; for (i = 0; i < mock->npages; i++) put_page(mock->pages[i]); kfree(mock); } static void *mock_dmabuf_vmap(struct dma_buf *dma_buf) { struct mock_dmabuf *mock = to_mock(dma_buf); return vm_map_ram(mock->pages, mock->npages, 0, PAGE_KERNEL); } static void mock_dmabuf_vunmap(struct dma_buf *dma_buf, void *vaddr) { struct mock_dmabuf *mock = to_mock(dma_buf); vm_unmap_ram(vaddr, mock->npages); } static void *mock_dmabuf_kmap(struct dma_buf *dma_buf, unsigned long page_num) { struct mock_dmabuf *mock = to_mock(dma_buf); return kmap(mock->pages[page_num]); } static void mock_dmabuf_kunmap(struct dma_buf *dma_buf, unsigned long page_num, void *addr) { struct mock_dmabuf *mock = to_mock(dma_buf); return kunmap(mock->pages[page_num]); } static int mock_dmabuf_mmap(struct dma_buf *dma_buf, struct vm_area_struct *vma) { return -ENODEV; } static const struct dma_buf_ops mock_dmabuf_ops = { .map_dma_buf = mock_map_dma_buf, .unmap_dma_buf = mock_unmap_dma_buf, .release = mock_dmabuf_release, .map = mock_dmabuf_kmap, .unmap = mock_dmabuf_kunmap, .mmap = mock_dmabuf_mmap, .vmap = mock_dmabuf_vmap, .vunmap = mock_dmabuf_vunmap, }; static struct dma_buf *mock_dmabuf(int npages) { struct mock_dmabuf *mock; DEFINE_DMA_BUF_EXPORT_INFO(exp_info); struct dma_buf *dmabuf; int i; mock = kmalloc(sizeof(*mock) + npages * sizeof(struct page *), GFP_KERNEL); if (!mock) return ERR_PTR(-ENOMEM); mock->npages = npages; for (i = 0; i < npages; i++) { mock->pages[i] = alloc_page(GFP_KERNEL); if (!mock->pages[i]) goto err; } exp_info.ops = &mock_dmabuf_ops; exp_info.size = npages * PAGE_SIZE; exp_info.flags = O_CLOEXEC; exp_info.priv = mock; dmabuf = dma_buf_export(&exp_info); if (IS_ERR(dmabuf)) goto err; return dmabuf; err: while (i--) put_page(mock->pages[i]); kfree(mock); return ERR_PTR(-ENOMEM); }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1