Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Alexandre Courbot | 960 | 55.30% | 6 | 14.29% |
Ben Skeggs | 726 | 41.82% | 34 | 80.95% |
Nicolas Chauvet | 49 | 2.82% | 1 | 2.38% |
Thierry Reding | 1 | 0.06% | 1 | 2.38% |
Total | 1736 | 42 |
/* * Copyright (c) 2014-2015, NVIDIA CORPORATION. All rights reserved. * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice shall be included in * all copies or substantial portions of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER * DEALINGS IN THE SOFTWARE. */ #include "gf100.h" #include "ctxgf100.h" #include <core/firmware.h> #include <subdev/timer.h> #include <nvif/class.h> struct gk20a_fw_av { u32 addr; u32 data; }; int gk20a_gr_av_to_init_(struct nvkm_blob *blob, u8 count, u32 pitch, struct gf100_gr_pack **ppack) { struct gf100_gr_init *init; struct gf100_gr_pack *pack; int nent; int i; nent = (blob->size / sizeof(struct gk20a_fw_av)); pack = vzalloc((sizeof(*pack) * 2) + (sizeof(*init) * (nent + 1))); if (!pack) return -ENOMEM; init = (void *)(pack + 2); pack[0].init = init; for (i = 0; i < nent; i++) { struct gf100_gr_init *ent = &init[i]; struct gk20a_fw_av *av = &((struct gk20a_fw_av *)blob->data)[i]; ent->addr = av->addr; ent->data = av->data; ent->count = ((ent->addr & 0xffff) != 0xe100) ? count : 1; ent->pitch = pitch; } *ppack = pack; return 0; } int gk20a_gr_av_to_init(struct nvkm_blob *blob, struct gf100_gr_pack **ppack) { return gk20a_gr_av_to_init_(blob, 1, 1, ppack); } struct gk20a_fw_aiv { u32 addr; u32 index; u32 data; }; int gk20a_gr_aiv_to_init(struct nvkm_blob *blob, struct gf100_gr_pack **ppack) { struct gf100_gr_init *init; struct gf100_gr_pack *pack; int nent; int i; nent = (blob->size / sizeof(struct gk20a_fw_aiv)); pack = vzalloc((sizeof(*pack) * 2) + (sizeof(*init) * (nent + 1))); if (!pack) return -ENOMEM; init = (void *)(pack + 2); pack[0].init = init; for (i = 0; i < nent; i++) { struct gf100_gr_init *ent = &init[i]; struct gk20a_fw_aiv *av = &((struct gk20a_fw_aiv *)blob->data)[i]; ent->addr = av->addr; ent->data = av->data; ent->count = 1; ent->pitch = 1; } *ppack = pack; return 0; } int gk20a_gr_av_to_method(struct nvkm_blob *blob, struct gf100_gr_pack **ppack) { struct gf100_gr_init *init; struct gf100_gr_pack *pack; /* We don't suppose we will initialize more than 16 classes here... */ static const unsigned int max_classes = 16; u32 classidx = 0, prevclass = 0; int nent; int i; nent = (blob->size / sizeof(struct gk20a_fw_av)); pack = vzalloc((sizeof(*pack) * (max_classes + 1)) + (sizeof(*init) * (nent + max_classes + 1))); if (!pack) return -ENOMEM; init = (void *)(pack + max_classes + 1); for (i = 0; i < nent; i++, init++) { struct gk20a_fw_av *av = &((struct gk20a_fw_av *)blob->data)[i]; u32 class = av->addr & 0xffff; u32 addr = (av->addr & 0xffff0000) >> 14; if (prevclass != class) { if (prevclass) /* Add terminator to the method list. */ init++; pack[classidx].init = init; pack[classidx].type = class; prevclass = class; if (++classidx >= max_classes) { vfree(pack); return -ENOSPC; } } init->addr = addr; init->data = av->data; init->count = 1; init->pitch = 1; } *ppack = pack; return 0; } static int gk20a_gr_wait_mem_scrubbing(struct gf100_gr *gr) { struct nvkm_subdev *subdev = &gr->base.engine.subdev; struct nvkm_device *device = subdev->device; if (nvkm_msec(device, 2000, if (!(nvkm_rd32(device, 0x40910c) & 0x00000006)) break; ) < 0) { nvkm_error(subdev, "FECS mem scrubbing timeout\n"); return -ETIMEDOUT; } if (nvkm_msec(device, 2000, if (!(nvkm_rd32(device, 0x41a10c) & 0x00000006)) break; ) < 0) { nvkm_error(subdev, "GPCCS mem scrubbing timeout\n"); return -ETIMEDOUT; } return 0; } static void gk20a_gr_set_hww_esr_report_mask(struct gf100_gr *gr) { struct nvkm_device *device = gr->base.engine.subdev.device; nvkm_wr32(device, 0x419e44, 0x1ffffe); nvkm_wr32(device, 0x419e4c, 0x7f); } int gk20a_gr_init(struct gf100_gr *gr) { struct nvkm_device *device = gr->base.engine.subdev.device; int ret; /* Clear SCC RAM */ nvkm_wr32(device, 0x40802c, 0x1); gf100_gr_mmio(gr, gr->sw_nonctx); ret = gk20a_gr_wait_mem_scrubbing(gr); if (ret) return ret; ret = gf100_gr_wait_idle(gr); if (ret) return ret; /* MMU debug buffer */ if (gr->func->init_gpc_mmu) gr->func->init_gpc_mmu(gr); /* Set the PE as stream master */ nvkm_mask(device, 0x503018, 0x1, 0x1); /* Zcull init */ gr->func->init_zcull(gr); gr->func->init_rop_active_fbps(gr); /* Enable FIFO access */ nvkm_wr32(device, 0x400500, 0x00010001); /* Enable interrupts */ nvkm_wr32(device, 0x400100, 0xffffffff); nvkm_wr32(device, 0x40013c, 0xffffffff); /* Enable FECS error interrupts */ nvkm_wr32(device, 0x409c24, 0x000f0000); /* Enable hardware warning exceptions */ nvkm_wr32(device, 0x404000, 0xc0000000); nvkm_wr32(device, 0x404600, 0xc0000000); if (gr->func->set_hww_esr_report_mask) gr->func->set_hww_esr_report_mask(gr); /* Enable TPC exceptions per GPC */ nvkm_wr32(device, 0x419d0c, 0x2); nvkm_wr32(device, 0x41ac94, (((1 << gr->tpc_total) - 1) & 0xff) << 16); /* Reset and enable all exceptions */ nvkm_wr32(device, 0x400108, 0xffffffff); nvkm_wr32(device, 0x400138, 0xffffffff); nvkm_wr32(device, 0x400118, 0xffffffff); nvkm_wr32(device, 0x400130, 0xffffffff); nvkm_wr32(device, 0x40011c, 0xffffffff); nvkm_wr32(device, 0x400134, 0xffffffff); gf100_gr_zbc_init(gr); return gf100_gr_init_ctxctl(gr); } static const struct gf100_gr_func gk20a_gr = { .oneinit_tiles = gf100_gr_oneinit_tiles, .oneinit_sm_id = gf100_gr_oneinit_sm_id, .init = gk20a_gr_init, .init_zcull = gf117_gr_init_zcull, .init_rop_active_fbps = gk104_gr_init_rop_active_fbps, .trap_mp = gf100_gr_trap_mp, .set_hww_esr_report_mask = gk20a_gr_set_hww_esr_report_mask, .fecs.reset = gf100_gr_fecs_reset, .rops = gf100_gr_rops, .ppc_nr = 1, .grctx = &gk20a_grctx, .zbc = &gf100_gr_zbc, .sclass = { { -1, -1, FERMI_TWOD_A }, { -1, -1, KEPLER_INLINE_TO_MEMORY_A }, { -1, -1, KEPLER_C, &gf100_fermi }, { -1, -1, KEPLER_COMPUTE_A }, {} } }; int gk20a_gr_load_net(struct gf100_gr *gr, const char *path, const char *name, int ver, int (*load)(struct nvkm_blob *, struct gf100_gr_pack **), struct gf100_gr_pack **ppack) { struct nvkm_blob blob; int ret; ret = nvkm_firmware_load_blob(&gr->base.engine.subdev, path, name, ver, &blob); if (ret) return ret; ret = load(&blob, ppack); nvkm_blob_dtor(&blob); return 0; } int gk20a_gr_load_sw(struct gf100_gr *gr, const char *path, int ver) { if (gk20a_gr_load_net(gr, path, "sw_nonctx", ver, gk20a_gr_av_to_init, &gr->sw_nonctx) || gk20a_gr_load_net(gr, path, "sw_ctx", ver, gk20a_gr_aiv_to_init, &gr->sw_ctx) || gk20a_gr_load_net(gr, path, "sw_bundle_init", ver, gk20a_gr_av_to_init, &gr->bundle) || gk20a_gr_load_net(gr, path, "sw_method_init", ver, gk20a_gr_av_to_method, &gr->method)) return -ENOENT; return 0; } #if IS_ENABLED(CONFIG_ARCH_TEGRA_124_SOC) || IS_ENABLED(CONFIG_ARCH_TEGRA_132_SOC) MODULE_FIRMWARE("nvidia/gk20a/fecs_data.bin"); MODULE_FIRMWARE("nvidia/gk20a/fecs_inst.bin"); MODULE_FIRMWARE("nvidia/gk20a/gpccs_data.bin"); MODULE_FIRMWARE("nvidia/gk20a/gpccs_inst.bin"); MODULE_FIRMWARE("nvidia/gk20a/sw_bundle_init.bin"); MODULE_FIRMWARE("nvidia/gk20a/sw_ctx.bin"); MODULE_FIRMWARE("nvidia/gk20a/sw_method_init.bin"); MODULE_FIRMWARE("nvidia/gk20a/sw_nonctx.bin"); #endif static int gk20a_gr_load(struct gf100_gr *gr, int ver, const struct gf100_gr_fwif *fwif) { struct nvkm_subdev *subdev = &gr->base.engine.subdev; if (nvkm_firmware_load_blob(subdev, "", "fecs_inst", ver, &gr->fecs.inst) || nvkm_firmware_load_blob(subdev, "", "fecs_data", ver, &gr->fecs.data) || nvkm_firmware_load_blob(subdev, "", "gpccs_inst", ver, &gr->gpccs.inst) || nvkm_firmware_load_blob(subdev, "", "gpccs_data", ver, &gr->gpccs.data)) return -ENOENT; gr->firmware = true; return gk20a_gr_load_sw(gr, "", ver); } static const struct gf100_gr_fwif gk20a_gr_fwif[] = { { 0, gk20a_gr_load, &gk20a_gr }, {} }; int gk20a_gr_new(struct nvkm_device *device, enum nvkm_subdev_type type, int inst, struct nvkm_gr **pgr) { return gf100_gr_new_(gk20a_gr_fwif, device, type, inst, pgr); }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1