Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Pratik Patel | 1138 | 52.69% | 1 | 2.70% |
Suzuki K. Poulose | 616 | 28.52% | 17 | 45.95% |
Mathieu J. Poirier | 373 | 17.27% | 13 | 35.14% |
Mike Leach | 16 | 0.74% | 1 | 2.70% |
Robin Murphy | 13 | 0.60% | 1 | 2.70% |
Linus Walleij | 1 | 0.05% | 1 | 2.70% |
Paul Gortmaker | 1 | 0.05% | 1 | 2.70% |
Baoyou Xie | 1 | 0.05% | 1 | 2.70% |
Arvind Yadav | 1 | 0.05% | 1 | 2.70% |
Total | 2160 | 37 |
// SPDX-License-Identifier: GPL-2.0 /* Copyright (c) 2012, The Linux Foundation. All rights reserved. * * Description: CoreSight Trace Memory Controller driver */ #include <linux/kernel.h> #include <linux/init.h> #include <linux/types.h> #include <linux/device.h> #include <linux/idr.h> #include <linux/io.h> #include <linux/err.h> #include <linux/fs.h> #include <linux/miscdevice.h> #include <linux/mutex.h> #include <linux/property.h> #include <linux/uaccess.h> #include <linux/slab.h> #include <linux/dma-mapping.h> #include <linux/spinlock.h> #include <linux/pm_runtime.h> #include <linux/of.h> #include <linux/coresight.h> #include <linux/amba/bus.h> #include "coresight-priv.h" #include "coresight-tmc.h" void tmc_wait_for_tmcready(struct tmc_drvdata *drvdata) { /* Ensure formatter, unformatter and hardware fifo are empty */ if (coresight_timeout(drvdata->base, TMC_STS, TMC_STS_TMCREADY_BIT, 1)) { dev_err(drvdata->dev, "timeout while waiting for TMC to be Ready\n"); } } void tmc_flush_and_stop(struct tmc_drvdata *drvdata) { u32 ffcr; ffcr = readl_relaxed(drvdata->base + TMC_FFCR); ffcr |= TMC_FFCR_STOP_ON_FLUSH; writel_relaxed(ffcr, drvdata->base + TMC_FFCR); ffcr |= BIT(TMC_FFCR_FLUSHMAN_BIT); writel_relaxed(ffcr, drvdata->base + TMC_FFCR); /* Ensure flush completes */ if (coresight_timeout(drvdata->base, TMC_FFCR, TMC_FFCR_FLUSHMAN_BIT, 0)) { dev_err(drvdata->dev, "timeout while waiting for completion of Manual Flush\n"); } tmc_wait_for_tmcready(drvdata); } void tmc_enable_hw(struct tmc_drvdata *drvdata) { writel_relaxed(TMC_CTL_CAPT_EN, drvdata->base + TMC_CTL); } void tmc_disable_hw(struct tmc_drvdata *drvdata) { writel_relaxed(0x0, drvdata->base + TMC_CTL); } static int tmc_read_prepare(struct tmc_drvdata *drvdata) { int ret = 0; switch (drvdata->config_type) { case TMC_CONFIG_TYPE_ETB: case TMC_CONFIG_TYPE_ETF: ret = tmc_read_prepare_etb(drvdata); break; case TMC_CONFIG_TYPE_ETR: ret = tmc_read_prepare_etr(drvdata); break; default: ret = -EINVAL; } if (!ret) dev_dbg(drvdata->dev, "TMC read start\n"); return ret; } static int tmc_read_unprepare(struct tmc_drvdata *drvdata) { int ret = 0; switch (drvdata->config_type) { case TMC_CONFIG_TYPE_ETB: case TMC_CONFIG_TYPE_ETF: ret = tmc_read_unprepare_etb(drvdata); break; case TMC_CONFIG_TYPE_ETR: ret = tmc_read_unprepare_etr(drvdata); break; default: ret = -EINVAL; } if (!ret) dev_dbg(drvdata->dev, "TMC read end\n"); return ret; } static int tmc_open(struct inode *inode, struct file *file) { int ret; struct tmc_drvdata *drvdata = container_of(file->private_data, struct tmc_drvdata, miscdev); ret = tmc_read_prepare(drvdata); if (ret) return ret; nonseekable_open(inode, file); dev_dbg(drvdata->dev, "%s: successfully opened\n", __func__); return 0; } static inline ssize_t tmc_get_sysfs_trace(struct tmc_drvdata *drvdata, loff_t pos, size_t len, char **bufpp) { switch (drvdata->config_type) { case TMC_CONFIG_TYPE_ETB: case TMC_CONFIG_TYPE_ETF: return tmc_etb_get_sysfs_trace(drvdata, pos, len, bufpp); case TMC_CONFIG_TYPE_ETR: return tmc_etr_get_sysfs_trace(drvdata, pos, len, bufpp); } return -EINVAL; } static ssize_t tmc_read(struct file *file, char __user *data, size_t len, loff_t *ppos) { char *bufp; ssize_t actual; struct tmc_drvdata *drvdata = container_of(file->private_data, struct tmc_drvdata, miscdev); actual = tmc_get_sysfs_trace(drvdata, *ppos, len, &bufp); if (actual <= 0) return 0; if (copy_to_user(data, bufp, actual)) { dev_dbg(drvdata->dev, "%s: copy_to_user failed\n", __func__); return -EFAULT; } *ppos += actual; dev_dbg(drvdata->dev, "%zu bytes copied\n", actual); return actual; } static int tmc_release(struct inode *inode, struct file *file) { int ret; struct tmc_drvdata *drvdata = container_of(file->private_data, struct tmc_drvdata, miscdev); ret = tmc_read_unprepare(drvdata); if (ret) return ret; dev_dbg(drvdata->dev, "%s: released\n", __func__); return 0; } static const struct file_operations tmc_fops = { .owner = THIS_MODULE, .open = tmc_open, .read = tmc_read, .release = tmc_release, .llseek = no_llseek, }; static enum tmc_mem_intf_width tmc_get_memwidth(u32 devid) { enum tmc_mem_intf_width memwidth; /* * Excerpt from the TRM: * * DEVID::MEMWIDTH[10:8] * 0x2 Memory interface databus is 32 bits wide. * 0x3 Memory interface databus is 64 bits wide. * 0x4 Memory interface databus is 128 bits wide. * 0x5 Memory interface databus is 256 bits wide. */ switch (BMVAL(devid, 8, 10)) { case 0x2: memwidth = TMC_MEM_INTF_WIDTH_32BITS; break; case 0x3: memwidth = TMC_MEM_INTF_WIDTH_64BITS; break; case 0x4: memwidth = TMC_MEM_INTF_WIDTH_128BITS; break; case 0x5: memwidth = TMC_MEM_INTF_WIDTH_256BITS; break; default: memwidth = 0; } return memwidth; } #define coresight_tmc_reg(name, offset) \ coresight_simple_reg32(struct tmc_drvdata, name, offset) #define coresight_tmc_reg64(name, lo_off, hi_off) \ coresight_simple_reg64(struct tmc_drvdata, name, lo_off, hi_off) coresight_tmc_reg(rsz, TMC_RSZ); coresight_tmc_reg(sts, TMC_STS); coresight_tmc_reg(trg, TMC_TRG); coresight_tmc_reg(ctl, TMC_CTL); coresight_tmc_reg(ffsr, TMC_FFSR); coresight_tmc_reg(ffcr, TMC_FFCR); coresight_tmc_reg(mode, TMC_MODE); coresight_tmc_reg(pscr, TMC_PSCR); coresight_tmc_reg(axictl, TMC_AXICTL); coresight_tmc_reg(devid, CORESIGHT_DEVID); coresight_tmc_reg64(rrp, TMC_RRP, TMC_RRPHI); coresight_tmc_reg64(rwp, TMC_RWP, TMC_RWPHI); coresight_tmc_reg64(dba, TMC_DBALO, TMC_DBAHI); static struct attribute *coresight_tmc_mgmt_attrs[] = { &dev_attr_rsz.attr, &dev_attr_sts.attr, &dev_attr_rrp.attr, &dev_attr_rwp.attr, &dev_attr_trg.attr, &dev_attr_ctl.attr, &dev_attr_ffsr.attr, &dev_attr_ffcr.attr, &dev_attr_mode.attr, &dev_attr_pscr.attr, &dev_attr_devid.attr, &dev_attr_dba.attr, &dev_attr_axictl.attr, NULL, }; static ssize_t trigger_cntr_show(struct device *dev, struct device_attribute *attr, char *buf) { struct tmc_drvdata *drvdata = dev_get_drvdata(dev->parent); unsigned long val = drvdata->trigger_cntr; return sprintf(buf, "%#lx\n", val); } static ssize_t trigger_cntr_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t size) { int ret; unsigned long val; struct tmc_drvdata *drvdata = dev_get_drvdata(dev->parent); ret = kstrtoul(buf, 16, &val); if (ret) return ret; drvdata->trigger_cntr = val; return size; } static DEVICE_ATTR_RW(trigger_cntr); static ssize_t buffer_size_show(struct device *dev, struct device_attribute *attr, char *buf) { struct tmc_drvdata *drvdata = dev_get_drvdata(dev->parent); return sprintf(buf, "%#x\n", drvdata->size); } static ssize_t buffer_size_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t size) { int ret; unsigned long val; struct tmc_drvdata *drvdata = dev_get_drvdata(dev->parent); /* Only permitted for TMC-ETRs */ if (drvdata->config_type != TMC_CONFIG_TYPE_ETR) return -EPERM; ret = kstrtoul(buf, 0, &val); if (ret) return ret; /* The buffer size should be page aligned */ if (val & (PAGE_SIZE - 1)) return -EINVAL; drvdata->size = val; return size; } static DEVICE_ATTR_RW(buffer_size); static struct attribute *coresight_tmc_attrs[] = { &dev_attr_trigger_cntr.attr, &dev_attr_buffer_size.attr, NULL, }; static const struct attribute_group coresight_tmc_group = { .attrs = coresight_tmc_attrs, }; static const struct attribute_group coresight_tmc_mgmt_group = { .attrs = coresight_tmc_mgmt_attrs, .name = "mgmt", }; const struct attribute_group *coresight_tmc_groups[] = { &coresight_tmc_group, &coresight_tmc_mgmt_group, NULL, }; static inline bool tmc_etr_can_use_sg(struct tmc_drvdata *drvdata) { return fwnode_property_present(drvdata->dev->fwnode, "arm,scatter-gather"); } /* Detect and initialise the capabilities of a TMC ETR */ static int tmc_etr_setup_caps(struct tmc_drvdata *drvdata, u32 devid, void *dev_caps) { int rc; u32 dma_mask = 0; /* Set the unadvertised capabilities */ tmc_etr_init_caps(drvdata, (u32)(unsigned long)dev_caps); if (!(devid & TMC_DEVID_NOSCAT) && tmc_etr_can_use_sg(drvdata)) tmc_etr_set_cap(drvdata, TMC_ETR_SG); /* Check if the AXI address width is available */ if (devid & TMC_DEVID_AXIAW_VALID) dma_mask = ((devid >> TMC_DEVID_AXIAW_SHIFT) & TMC_DEVID_AXIAW_MASK); /* * Unless specified in the device configuration, ETR uses a 40-bit * AXI master in place of the embedded SRAM of ETB/ETF. */ switch (dma_mask) { case 32: case 40: case 44: case 48: case 52: dev_info(drvdata->dev, "Detected dma mask %dbits\n", dma_mask); break; default: dma_mask = 40; } rc = dma_set_mask_and_coherent(drvdata->dev, DMA_BIT_MASK(dma_mask)); if (rc) dev_err(drvdata->dev, "Failed to setup DMA mask: %d\n", rc); return rc; } static int tmc_probe(struct amba_device *adev, const struct amba_id *id) { int ret = 0; u32 devid; void __iomem *base; struct device *dev = &adev->dev; struct coresight_platform_data *pdata = NULL; struct tmc_drvdata *drvdata; struct resource *res = &adev->res; struct coresight_desc desc = { 0 }; struct device_node *np = adev->dev.of_node; if (np) { pdata = of_get_coresight_platform_data(dev, np); if (IS_ERR(pdata)) { ret = PTR_ERR(pdata); goto out; } adev->dev.platform_data = pdata; } ret = -ENOMEM; drvdata = devm_kzalloc(dev, sizeof(*drvdata), GFP_KERNEL); if (!drvdata) goto out; drvdata->dev = &adev->dev; dev_set_drvdata(dev, drvdata); /* Validity for the resource is already checked by the AMBA core */ base = devm_ioremap_resource(dev, res); if (IS_ERR(base)) { ret = PTR_ERR(base); goto out; } drvdata->base = base; spin_lock_init(&drvdata->spinlock); devid = readl_relaxed(drvdata->base + CORESIGHT_DEVID); drvdata->config_type = BMVAL(devid, 6, 7); drvdata->memwidth = tmc_get_memwidth(devid); /* This device is not associated with a session */ drvdata->pid = -1; if (drvdata->config_type == TMC_CONFIG_TYPE_ETR) { if (np) ret = of_property_read_u32(np, "arm,buffer-size", &drvdata->size); if (ret) drvdata->size = SZ_1M; } else { drvdata->size = readl_relaxed(drvdata->base + TMC_RSZ) * 4; } desc.pdata = pdata; desc.dev = dev; desc.groups = coresight_tmc_groups; switch (drvdata->config_type) { case TMC_CONFIG_TYPE_ETB: desc.type = CORESIGHT_DEV_TYPE_SINK; desc.subtype.sink_subtype = CORESIGHT_DEV_SUBTYPE_SINK_BUFFER; desc.ops = &tmc_etb_cs_ops; break; case TMC_CONFIG_TYPE_ETR: desc.type = CORESIGHT_DEV_TYPE_SINK; desc.subtype.sink_subtype = CORESIGHT_DEV_SUBTYPE_SINK_BUFFER; desc.ops = &tmc_etr_cs_ops; ret = tmc_etr_setup_caps(drvdata, devid, coresight_get_uci_data(id)); if (ret) goto out; idr_init(&drvdata->idr); mutex_init(&drvdata->idr_mutex); break; case TMC_CONFIG_TYPE_ETF: desc.type = CORESIGHT_DEV_TYPE_LINKSINK; desc.subtype.link_subtype = CORESIGHT_DEV_SUBTYPE_LINK_FIFO; desc.ops = &tmc_etf_cs_ops; break; default: pr_err("%s: Unsupported TMC config\n", pdata->name); ret = -EINVAL; goto out; } drvdata->csdev = coresight_register(&desc); if (IS_ERR(drvdata->csdev)) { ret = PTR_ERR(drvdata->csdev); goto out; } drvdata->miscdev.name = pdata->name; drvdata->miscdev.minor = MISC_DYNAMIC_MINOR; drvdata->miscdev.fops = &tmc_fops; ret = misc_register(&drvdata->miscdev); if (ret) coresight_unregister(drvdata->csdev); else pm_runtime_put(&adev->dev); out: return ret; } static const struct amba_id tmc_ids[] = { CS_AMBA_ID(0x000bb961), /* Coresight SoC 600 TMC-ETR/ETS */ CS_AMBA_ID_DATA(0x000bb9e8, (unsigned long)CORESIGHT_SOC_600_ETR_CAPS), /* Coresight SoC 600 TMC-ETB */ CS_AMBA_ID(0x000bb9e9), /* Coresight SoC 600 TMC-ETF */ CS_AMBA_ID(0x000bb9ea), { 0, 0}, }; static struct amba_driver tmc_driver = { .drv = { .name = "coresight-tmc", .owner = THIS_MODULE, .suppress_bind_attrs = true, }, .probe = tmc_probe, .id_table = tmc_ids, }; builtin_amba_driver(tmc_driver);
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1