Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Claudiu Manoil | 1343 | 91.99% | 4 | 40.00% |
Sebastian Andrzej Siewior | 37 | 2.53% | 1 | 10.00% |
Tudor Laurentiu | 36 | 2.47% | 1 | 10.00% |
Madalin Bucur | 23 | 1.58% | 2 | 20.00% |
Roy Pledge | 17 | 1.16% | 1 | 10.00% |
Rob Herring | 4 | 0.27% | 1 | 10.00% |
Total | 1460 | 10 |
/* Copyright 2008 - 2016 Freescale Semiconductor, Inc. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions are met: * * Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * * Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * * Neither the name of Freescale Semiconductor nor the * names of its contributors may be used to endorse or promote products * derived from this software without specific prior written permission. * * ALTERNATIVELY, this software may be distributed under the terms of the * GNU General Public License ("GPL") as published by the Free Software * Foundation, either version 2 of that License or (at your option) any * later version. * * THIS SOFTWARE IS PROVIDED BY Freescale Semiconductor ``AS IS'' AND ANY * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE * DISCLAIMED. IN NO EVENT SHALL Freescale Semiconductor BE LIABLE FOR ANY * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS * SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. */ #include "qman_priv.h" struct qman_portal *qman_dma_portal; EXPORT_SYMBOL(qman_dma_portal); /* Enable portal interupts (as opposed to polling mode) */ #define CONFIG_FSL_DPA_PIRQ_SLOW 1 #define CONFIG_FSL_DPA_PIRQ_FAST 1 static struct cpumask portal_cpus; /* protect qman global registers and global data shared among portals */ static DEFINE_SPINLOCK(qman_lock); static void portal_set_cpu(struct qm_portal_config *pcfg, int cpu) { #ifdef CONFIG_FSL_PAMU struct device *dev = pcfg->dev; int window_count = 1; struct iommu_domain_geometry geom_attr; struct pamu_stash_attribute stash_attr; int ret; pcfg->iommu_domain = iommu_domain_alloc(&platform_bus_type); if (!pcfg->iommu_domain) { dev_err(dev, "%s(): iommu_domain_alloc() failed", __func__); goto no_iommu; } geom_attr.aperture_start = 0; geom_attr.aperture_end = ((dma_addr_t)1 << min(8 * sizeof(dma_addr_t), (size_t)36)) - 1; geom_attr.force_aperture = true; ret = iommu_domain_set_attr(pcfg->iommu_domain, DOMAIN_ATTR_GEOMETRY, &geom_attr); if (ret < 0) { dev_err(dev, "%s(): iommu_domain_set_attr() = %d", __func__, ret); goto out_domain_free; } ret = iommu_domain_set_attr(pcfg->iommu_domain, DOMAIN_ATTR_WINDOWS, &window_count); if (ret < 0) { dev_err(dev, "%s(): iommu_domain_set_attr() = %d", __func__, ret); goto out_domain_free; } stash_attr.cpu = cpu; stash_attr.cache = PAMU_ATTR_CACHE_L1; ret = iommu_domain_set_attr(pcfg->iommu_domain, DOMAIN_ATTR_FSL_PAMU_STASH, &stash_attr); if (ret < 0) { dev_err(dev, "%s(): iommu_domain_set_attr() = %d", __func__, ret); goto out_domain_free; } ret = iommu_domain_window_enable(pcfg->iommu_domain, 0, 0, 1ULL << 36, IOMMU_READ | IOMMU_WRITE); if (ret < 0) { dev_err(dev, "%s(): iommu_domain_window_enable() = %d", __func__, ret); goto out_domain_free; } ret = iommu_attach_device(pcfg->iommu_domain, dev); if (ret < 0) { dev_err(dev, "%s(): iommu_device_attach() = %d", __func__, ret); goto out_domain_free; } ret = iommu_domain_set_attr(pcfg->iommu_domain, DOMAIN_ATTR_FSL_PAMU_ENABLE, &window_count); if (ret < 0) { dev_err(dev, "%s(): iommu_domain_set_attr() = %d", __func__, ret); goto out_detach_device; } no_iommu: #endif qman_set_sdest(pcfg->channel, cpu); return; #ifdef CONFIG_FSL_PAMU out_detach_device: iommu_detach_device(pcfg->iommu_domain, NULL); out_domain_free: iommu_domain_free(pcfg->iommu_domain); pcfg->iommu_domain = NULL; #endif } static struct qman_portal *init_pcfg(struct qm_portal_config *pcfg) { struct qman_portal *p; u32 irq_sources = 0; /* We need the same LIODN offset for all portals */ qman_liodn_fixup(pcfg->channel); pcfg->iommu_domain = NULL; portal_set_cpu(pcfg, pcfg->cpu); p = qman_create_affine_portal(pcfg, NULL); if (!p) { dev_crit(pcfg->dev, "%s: Portal failure on cpu %d\n", __func__, pcfg->cpu); return NULL; } /* Determine what should be interrupt-vs-poll driven */ #ifdef CONFIG_FSL_DPA_PIRQ_SLOW irq_sources |= QM_PIRQ_EQCI | QM_PIRQ_EQRI | QM_PIRQ_MRI | QM_PIRQ_CSCI; #endif #ifdef CONFIG_FSL_DPA_PIRQ_FAST irq_sources |= QM_PIRQ_DQRI; #endif qman_p_irqsource_add(p, irq_sources); spin_lock(&qman_lock); if (cpumask_equal(&portal_cpus, cpu_possible_mask)) { /* all assigned portals are initialized now */ qman_init_cgr_all(); } if (!qman_dma_portal) qman_dma_portal = p; spin_unlock(&qman_lock); dev_info(pcfg->dev, "Portal initialised, cpu %d\n", pcfg->cpu); return p; } static void qman_portal_update_sdest(const struct qm_portal_config *pcfg, unsigned int cpu) { #ifdef CONFIG_FSL_PAMU /* TODO */ struct pamu_stash_attribute stash_attr; int ret; if (pcfg->iommu_domain) { stash_attr.cpu = cpu; stash_attr.cache = PAMU_ATTR_CACHE_L1; ret = iommu_domain_set_attr(pcfg->iommu_domain, DOMAIN_ATTR_FSL_PAMU_STASH, &stash_attr); if (ret < 0) { dev_err(pcfg->dev, "Failed to update pamu stash setting\n"); return; } } #endif qman_set_sdest(pcfg->channel, cpu); } static int qman_offline_cpu(unsigned int cpu) { struct qman_portal *p; const struct qm_portal_config *pcfg; p = affine_portals[cpu]; if (p) { pcfg = qman_get_qm_portal_config(p); if (pcfg) { /* select any other online CPU */ cpu = cpumask_any_but(cpu_online_mask, cpu); irq_set_affinity(pcfg->irq, cpumask_of(cpu)); qman_portal_update_sdest(pcfg, cpu); } } return 0; } static int qman_online_cpu(unsigned int cpu) { struct qman_portal *p; const struct qm_portal_config *pcfg; p = affine_portals[cpu]; if (p) { pcfg = qman_get_qm_portal_config(p); if (pcfg) { irq_set_affinity(pcfg->irq, cpumask_of(cpu)); qman_portal_update_sdest(pcfg, cpu); } } return 0; } static int qman_portal_probe(struct platform_device *pdev) { struct device *dev = &pdev->dev; struct device_node *node = dev->of_node; struct qm_portal_config *pcfg; struct resource *addr_phys[2]; int irq, cpu, err; u32 val; err = qman_is_probed(); if (!err) return -EPROBE_DEFER; if (err < 0) { dev_err(&pdev->dev, "failing probe due to qman probe error\n"); return -ENODEV; } pcfg = devm_kmalloc(dev, sizeof(*pcfg), GFP_KERNEL); if (!pcfg) return -ENOMEM; pcfg->dev = dev; addr_phys[0] = platform_get_resource(pdev, IORESOURCE_MEM, DPAA_PORTAL_CE); if (!addr_phys[0]) { dev_err(dev, "Can't get %pOF property 'reg::CE'\n", node); return -ENXIO; } addr_phys[1] = platform_get_resource(pdev, IORESOURCE_MEM, DPAA_PORTAL_CI); if (!addr_phys[1]) { dev_err(dev, "Can't get %pOF property 'reg::CI'\n", node); return -ENXIO; } err = of_property_read_u32(node, "cell-index", &val); if (err) { dev_err(dev, "Can't get %pOF property 'cell-index'\n", node); return err; } pcfg->channel = val; pcfg->cpu = -1; irq = platform_get_irq(pdev, 0); if (irq <= 0) { dev_err(dev, "Can't get %pOF IRQ\n", node); return -ENXIO; } pcfg->irq = irq; pcfg->addr_virt_ce = memremap(addr_phys[0]->start, resource_size(addr_phys[0]), QBMAN_MEMREMAP_ATTR); if (!pcfg->addr_virt_ce) { dev_err(dev, "memremap::CE failed\n"); goto err_ioremap1; } pcfg->addr_virt_ci = ioremap(addr_phys[1]->start, resource_size(addr_phys[1])); if (!pcfg->addr_virt_ci) { dev_err(dev, "ioremap::CI failed\n"); goto err_ioremap2; } pcfg->pools = qm_get_pools_sdqcr(); spin_lock(&qman_lock); cpu = cpumask_next_zero(-1, &portal_cpus); if (cpu >= nr_cpu_ids) { /* unassigned portal, skip init */ spin_unlock(&qman_lock); return 0; } cpumask_set_cpu(cpu, &portal_cpus); spin_unlock(&qman_lock); pcfg->cpu = cpu; if (dma_set_mask(dev, DMA_BIT_MASK(40))) { dev_err(dev, "dma_set_mask() failed\n"); goto err_portal_init; } if (!init_pcfg(pcfg)) { dev_err(dev, "portal init failed\n"); goto err_portal_init; } /* clear irq affinity if assigned cpu is offline */ if (!cpu_online(cpu)) qman_offline_cpu(cpu); return 0; err_portal_init: iounmap(pcfg->addr_virt_ci); err_ioremap2: memunmap(pcfg->addr_virt_ce); err_ioremap1: return -ENXIO; } static const struct of_device_id qman_portal_ids[] = { { .compatible = "fsl,qman-portal", }, {} }; MODULE_DEVICE_TABLE(of, qman_portal_ids); static struct platform_driver qman_portal_driver = { .driver = { .name = KBUILD_MODNAME, .of_match_table = qman_portal_ids, }, .probe = qman_portal_probe, }; static int __init qman_portal_driver_register(struct platform_driver *drv) { int ret; ret = platform_driver_register(drv); if (ret < 0) return ret; ret = cpuhp_setup_state_nocalls(CPUHP_AP_ONLINE_DYN, "soc/qman_portal:online", qman_online_cpu, qman_offline_cpu); if (ret < 0) { pr_err("qman: failed to register hotplug callbacks.\n"); platform_driver_unregister(drv); return ret; } return 0; } module_driver(qman_portal_driver, qman_portal_driver_register, platform_driver_unregister);
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1