Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Ben Skeggs | 2657 | 77.85% | 34 | 91.89% |
Danilo Krummrich | 750 | 21.97% | 2 | 5.41% |
Christoph Böhmwalder | 6 | 0.18% | 1 | 2.70% |
Total | 3413 | 37 |
/* * Copyright 2017 Red Hat Inc. * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice shall be included in * all copies or substantial portions of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, DAMAGES OR * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR * OTHER DEALINGS IN THE SOFTWARE. */ #include "uvmm.h" #include "umem.h" #include "ummu.h" #include <core/client.h> #include <core/memory.h> #include <nvif/if000c.h> #include <nvif/unpack.h> static const struct nvkm_object_func nvkm_uvmm; struct nvkm_vmm * nvkm_uvmm_search(struct nvkm_client *client, u64 handle) { struct nvkm_object *object; object = nvkm_object_search(client, handle, &nvkm_uvmm); if (IS_ERR(object)) return (void *)object; return nvkm_vmm_ref(nvkm_uvmm(object)->vmm); } static int nvkm_uvmm_mthd_pfnclr(struct nvkm_uvmm *uvmm, void *argv, u32 argc) { union { struct nvif_vmm_pfnclr_v0 v0; } *args = argv; struct nvkm_vmm *vmm = uvmm->vmm; int ret = -ENOSYS; u64 addr, size; if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, false))) { addr = args->v0.addr; size = args->v0.size; } else return ret; if (nvkm_vmm_in_managed_range(vmm, addr, size) && vmm->managed.raw) return -EINVAL; if (size) { mutex_lock(&vmm->mutex.vmm); ret = nvkm_vmm_pfn_unmap(vmm, addr, size); mutex_unlock(&vmm->mutex.vmm); } return ret; } static int nvkm_uvmm_mthd_pfnmap(struct nvkm_uvmm *uvmm, void *argv, u32 argc) { union { struct nvif_vmm_pfnmap_v0 v0; } *args = argv; struct nvkm_vmm *vmm = uvmm->vmm; int ret = -ENOSYS; u64 addr, size, *phys; u8 page; if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, true))) { page = args->v0.page; addr = args->v0.addr; size = args->v0.size; phys = args->v0.phys; if (argc != (size >> page) * sizeof(args->v0.phys[0])) return -EINVAL; } else return ret; if (nvkm_vmm_in_managed_range(vmm, addr, size) && vmm->managed.raw) return -EINVAL; if (size) { mutex_lock(&vmm->mutex.vmm); ret = nvkm_vmm_pfn_map(vmm, page, addr, size, phys); mutex_unlock(&vmm->mutex.vmm); } return ret; } static int nvkm_uvmm_mthd_unmap(struct nvkm_uvmm *uvmm, void *argv, u32 argc) { union { struct nvif_vmm_unmap_v0 v0; } *args = argv; struct nvkm_vmm *vmm = uvmm->vmm; struct nvkm_vma *vma; int ret = -ENOSYS; u64 addr; if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, false))) { addr = args->v0.addr; } else return ret; if (nvkm_vmm_in_managed_range(vmm, addr, 0) && vmm->managed.raw) return -EINVAL; mutex_lock(&vmm->mutex.vmm); vma = nvkm_vmm_node_search(vmm, addr); if (ret = -ENOENT, !vma || vma->addr != addr) { VMM_DEBUG(vmm, "lookup %016llx: %016llx", addr, vma ? vma->addr : ~0ULL); goto done; } if (ret = -ENOENT, vma->busy) { VMM_DEBUG(vmm, "denied %016llx: %d", addr, vma->busy); goto done; } if (ret = -EINVAL, !vma->memory) { VMM_DEBUG(vmm, "unmapped"); goto done; } nvkm_vmm_unmap_locked(vmm, vma, false); ret = 0; done: mutex_unlock(&vmm->mutex.vmm); return ret; } static int nvkm_uvmm_mthd_map(struct nvkm_uvmm *uvmm, void *argv, u32 argc) { struct nvkm_client *client = uvmm->object.client; union { struct nvif_vmm_map_v0 v0; } *args = argv; u64 addr, size, handle, offset; struct nvkm_vmm *vmm = uvmm->vmm; struct nvkm_vma *vma; struct nvkm_memory *memory; int ret = -ENOSYS; if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, true))) { addr = args->v0.addr; size = args->v0.size; handle = args->v0.memory; offset = args->v0.offset; } else return ret; if (nvkm_vmm_in_managed_range(vmm, addr, size) && vmm->managed.raw) return -EINVAL; memory = nvkm_umem_search(client, handle); if (IS_ERR(memory)) { VMM_DEBUG(vmm, "memory %016llx %ld\n", handle, PTR_ERR(memory)); return PTR_ERR(memory); } mutex_lock(&vmm->mutex.vmm); if (ret = -ENOENT, !(vma = nvkm_vmm_node_search(vmm, addr))) { VMM_DEBUG(vmm, "lookup %016llx", addr); goto fail; } if (ret = -ENOENT, vma->busy) { VMM_DEBUG(vmm, "denied %016llx: %d", addr, vma->busy); goto fail; } if (ret = -EINVAL, vma->mapped && !vma->memory) { VMM_DEBUG(vmm, "pfnmap %016llx", addr); goto fail; } if (ret = -EINVAL, vma->addr != addr || vma->size != size) { if (addr + size > vma->addr + vma->size || vma->memory || (vma->refd == NVKM_VMA_PAGE_NONE && !vma->mapref)) { VMM_DEBUG(vmm, "split %d %d %d " "%016llx %016llx %016llx %016llx", !!vma->memory, vma->refd, vma->mapref, addr, size, vma->addr, (u64)vma->size); goto fail; } vma = nvkm_vmm_node_split(vmm, vma, addr, size); if (!vma) { ret = -ENOMEM; goto fail; } } vma->busy = true; mutex_unlock(&vmm->mutex.vmm); ret = nvkm_memory_map(memory, offset, vmm, vma, argv, argc); if (ret == 0) { /* Successful map will clear vma->busy. */ nvkm_memory_unref(&memory); return 0; } mutex_lock(&vmm->mutex.vmm); vma->busy = false; nvkm_vmm_unmap_region(vmm, vma); fail: mutex_unlock(&vmm->mutex.vmm); nvkm_memory_unref(&memory); return ret; } static int nvkm_uvmm_mthd_put(struct nvkm_uvmm *uvmm, void *argv, u32 argc) { union { struct nvif_vmm_put_v0 v0; } *args = argv; struct nvkm_vmm *vmm = uvmm->vmm; struct nvkm_vma *vma; int ret = -ENOSYS; u64 addr; if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, false))) { addr = args->v0.addr; } else return ret; mutex_lock(&vmm->mutex.vmm); vma = nvkm_vmm_node_search(vmm, args->v0.addr); if (ret = -ENOENT, !vma || vma->addr != addr || vma->part) { VMM_DEBUG(vmm, "lookup %016llx: %016llx %d", addr, vma ? vma->addr : ~0ULL, vma ? vma->part : 0); goto done; } if (ret = -ENOENT, vma->busy) { VMM_DEBUG(vmm, "denied %016llx: %d", addr, vma->busy); goto done; } nvkm_vmm_put_locked(vmm, vma); ret = 0; done: mutex_unlock(&vmm->mutex.vmm); return ret; } static int nvkm_uvmm_mthd_get(struct nvkm_uvmm *uvmm, void *argv, u32 argc) { union { struct nvif_vmm_get_v0 v0; } *args = argv; struct nvkm_vmm *vmm = uvmm->vmm; struct nvkm_vma *vma; int ret = -ENOSYS; bool getref, mapref, sparse; u8 page, align; u64 size; if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, false))) { getref = args->v0.type == NVIF_VMM_GET_V0_PTES; mapref = args->v0.type == NVIF_VMM_GET_V0_ADDR; sparse = args->v0.sparse; page = args->v0.page; align = args->v0.align; size = args->v0.size; } else return ret; mutex_lock(&vmm->mutex.vmm); ret = nvkm_vmm_get_locked(vmm, getref, mapref, sparse, page, align, size, &vma); mutex_unlock(&vmm->mutex.vmm); if (ret) return ret; args->v0.addr = vma->addr; return ret; } static int nvkm_uvmm_mthd_page(struct nvkm_uvmm *uvmm, void *argv, u32 argc) { union { struct nvif_vmm_page_v0 v0; } *args = argv; const struct nvkm_vmm_page *page; int ret = -ENOSYS; u8 type, index, nr; page = uvmm->vmm->func->page; for (nr = 0; page[nr].shift; nr++); if (!(nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, false))) { if ((index = args->v0.index) >= nr) return -EINVAL; type = page[index].type; args->v0.shift = page[index].shift; args->v0.sparse = !!(type & NVKM_VMM_PAGE_SPARSE); args->v0.vram = !!(type & NVKM_VMM_PAGE_VRAM); args->v0.host = !!(type & NVKM_VMM_PAGE_HOST); args->v0.comp = !!(type & NVKM_VMM_PAGE_COMP); } else return -ENOSYS; return 0; } static inline int nvkm_uvmm_page_index(struct nvkm_uvmm *uvmm, u64 size, u8 shift, u8 *refd) { struct nvkm_vmm *vmm = uvmm->vmm; const struct nvkm_vmm_page *page; if (likely(shift)) { for (page = vmm->func->page; page->shift; page++) { if (shift == page->shift) break; } if (!page->shift || !IS_ALIGNED(size, 1ULL << page->shift)) { VMM_DEBUG(vmm, "page %d %016llx", shift, size); return -EINVAL; } } else { return -EINVAL; } *refd = page - vmm->func->page; return 0; } static int nvkm_uvmm_mthd_raw_get(struct nvkm_uvmm *uvmm, struct nvif_vmm_raw_v0 *args) { struct nvkm_vmm *vmm = uvmm->vmm; u8 refd; int ret; if (!nvkm_vmm_in_managed_range(vmm, args->addr, args->size)) return -EINVAL; ret = nvkm_uvmm_page_index(uvmm, args->size, args->shift, &refd); if (ret) return ret; return nvkm_vmm_raw_get(vmm, args->addr, args->size, refd); } static int nvkm_uvmm_mthd_raw_put(struct nvkm_uvmm *uvmm, struct nvif_vmm_raw_v0 *args) { struct nvkm_vmm *vmm = uvmm->vmm; u8 refd; int ret; if (!nvkm_vmm_in_managed_range(vmm, args->addr, args->size)) return -EINVAL; ret = nvkm_uvmm_page_index(uvmm, args->size, args->shift, &refd); if (ret) return ret; nvkm_vmm_raw_put(vmm, args->addr, args->size, refd); return 0; } static int nvkm_uvmm_mthd_raw_map(struct nvkm_uvmm *uvmm, struct nvif_vmm_raw_v0 *args) { struct nvkm_client *client = uvmm->object.client; struct nvkm_vmm *vmm = uvmm->vmm; struct nvkm_vma vma = { .addr = args->addr, .size = args->size, .used = true, .mapref = false, .no_comp = true, }; struct nvkm_memory *memory; void *argv = (void *)(uintptr_t)args->argv; unsigned int argc = args->argc; u64 handle = args->memory; u8 refd; int ret; if (!nvkm_vmm_in_managed_range(vmm, args->addr, args->size)) return -EINVAL; ret = nvkm_uvmm_page_index(uvmm, args->size, args->shift, &refd); if (ret) return ret; vma.page = vma.refd = refd; memory = nvkm_umem_search(client, args->memory); if (IS_ERR(memory)) { VMM_DEBUG(vmm, "memory %016llx %ld\n", handle, PTR_ERR(memory)); return PTR_ERR(memory); } ret = nvkm_memory_map(memory, args->offset, vmm, &vma, argv, argc); nvkm_memory_unref(&vma.memory); nvkm_memory_unref(&memory); return ret; } static int nvkm_uvmm_mthd_raw_unmap(struct nvkm_uvmm *uvmm, struct nvif_vmm_raw_v0 *args) { struct nvkm_vmm *vmm = uvmm->vmm; u8 refd; int ret; if (!nvkm_vmm_in_managed_range(vmm, args->addr, args->size)) return -EINVAL; ret = nvkm_uvmm_page_index(uvmm, args->size, args->shift, &refd); if (ret) return ret; nvkm_vmm_raw_unmap(vmm, args->addr, args->size, args->sparse, refd); return 0; } static int nvkm_uvmm_mthd_raw_sparse(struct nvkm_uvmm *uvmm, struct nvif_vmm_raw_v0 *args) { struct nvkm_vmm *vmm = uvmm->vmm; if (!nvkm_vmm_in_managed_range(vmm, args->addr, args->size)) return -EINVAL; return nvkm_vmm_raw_sparse(vmm, args->addr, args->size, args->ref); } static int nvkm_uvmm_mthd_raw(struct nvkm_uvmm *uvmm, void *argv, u32 argc) { union { struct nvif_vmm_raw_v0 v0; } *args = argv; int ret = -ENOSYS; if (!uvmm->vmm->managed.raw) return -EINVAL; if ((ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, true))) return ret; switch (args->v0.op) { case NVIF_VMM_RAW_V0_GET: return nvkm_uvmm_mthd_raw_get(uvmm, &args->v0); case NVIF_VMM_RAW_V0_PUT: return nvkm_uvmm_mthd_raw_put(uvmm, &args->v0); case NVIF_VMM_RAW_V0_MAP: return nvkm_uvmm_mthd_raw_map(uvmm, &args->v0); case NVIF_VMM_RAW_V0_UNMAP: return nvkm_uvmm_mthd_raw_unmap(uvmm, &args->v0); case NVIF_VMM_RAW_V0_SPARSE: return nvkm_uvmm_mthd_raw_sparse(uvmm, &args->v0); default: return -EINVAL; }; } static int nvkm_uvmm_mthd(struct nvkm_object *object, u32 mthd, void *argv, u32 argc) { struct nvkm_uvmm *uvmm = nvkm_uvmm(object); switch (mthd) { case NVIF_VMM_V0_PAGE : return nvkm_uvmm_mthd_page (uvmm, argv, argc); case NVIF_VMM_V0_GET : return nvkm_uvmm_mthd_get (uvmm, argv, argc); case NVIF_VMM_V0_PUT : return nvkm_uvmm_mthd_put (uvmm, argv, argc); case NVIF_VMM_V0_MAP : return nvkm_uvmm_mthd_map (uvmm, argv, argc); case NVIF_VMM_V0_UNMAP : return nvkm_uvmm_mthd_unmap (uvmm, argv, argc); case NVIF_VMM_V0_PFNMAP: return nvkm_uvmm_mthd_pfnmap(uvmm, argv, argc); case NVIF_VMM_V0_PFNCLR: return nvkm_uvmm_mthd_pfnclr(uvmm, argv, argc); case NVIF_VMM_V0_RAW : return nvkm_uvmm_mthd_raw (uvmm, argv, argc); case NVIF_VMM_V0_MTHD(0x00) ... NVIF_VMM_V0_MTHD(0x7f): if (uvmm->vmm->func->mthd) { return uvmm->vmm->func->mthd(uvmm->vmm, uvmm->object.client, mthd, argv, argc); } break; default: break; } return -EINVAL; } static void * nvkm_uvmm_dtor(struct nvkm_object *object) { struct nvkm_uvmm *uvmm = nvkm_uvmm(object); nvkm_vmm_unref(&uvmm->vmm); return uvmm; } static const struct nvkm_object_func nvkm_uvmm = { .dtor = nvkm_uvmm_dtor, .mthd = nvkm_uvmm_mthd, }; int nvkm_uvmm_new(const struct nvkm_oclass *oclass, void *argv, u32 argc, struct nvkm_object **pobject) { struct nvkm_mmu *mmu = nvkm_ummu(oclass->parent)->mmu; const bool more = oclass->base.maxver >= 0; union { struct nvif_vmm_v0 v0; } *args = argv; const struct nvkm_vmm_page *page; struct nvkm_uvmm *uvmm; int ret = -ENOSYS; u64 addr, size; bool managed, raw; if (!(ret = nvif_unpack(ret, &argv, &argc, args->v0, 0, 0, more))) { managed = args->v0.type == NVIF_VMM_V0_TYPE_MANAGED; raw = args->v0.type == NVIF_VMM_V0_TYPE_RAW; addr = args->v0.addr; size = args->v0.size; } else return ret; if (!(uvmm = kzalloc(sizeof(*uvmm), GFP_KERNEL))) return -ENOMEM; nvkm_object_ctor(&nvkm_uvmm, oclass, &uvmm->object); *pobject = &uvmm->object; if (!mmu->vmm) { ret = mmu->func->vmm.ctor(mmu, managed || raw, addr, size, argv, argc, NULL, "user", &uvmm->vmm); if (ret) return ret; uvmm->vmm->debug = max(uvmm->vmm->debug, oclass->client->debug); } else { if (size) return -EINVAL; uvmm->vmm = nvkm_vmm_ref(mmu->vmm); } uvmm->vmm->managed.raw = raw; if (mmu->func->promote_vmm) { ret = mmu->func->promote_vmm(uvmm->vmm); if (ret) return ret; } page = uvmm->vmm->func->page; args->v0.page_nr = 0; while (page && (page++)->shift) args->v0.page_nr++; args->v0.addr = uvmm->vmm->start; args->v0.size = uvmm->vmm->limit; return 0; }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1