Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Cyril Chemparathy | 1632 | 50.08% | 1 | 2.63% |
Ravi Gunasekaran | 899 | 27.59% | 1 | 2.63% |
Grygorii Strashko | 435 | 13.35% | 11 | 28.95% |
Mugunthan V N | 183 | 5.62% | 3 | 7.89% |
Sekhar Nori | 26 | 0.80% | 1 | 2.63% |
Maxim Uvarov | 16 | 0.49% | 2 | 5.26% |
Wei Yongjun | 12 | 0.37% | 2 | 5.26% |
Lad Prabhakar | 10 | 0.31% | 2 | 5.26% |
Jingoo Han | 7 | 0.21% | 1 | 2.63% |
Sebastian Andrzej Siewior | 7 | 0.21% | 1 | 2.63% |
Andrew Lunn | 7 | 0.21% | 3 | 7.89% |
Arnd Bergmann | 5 | 0.15% | 1 | 2.63% |
Stephen Boyd | 5 | 0.15% | 1 | 2.63% |
Chi Minghao | 4 | 0.12% | 1 | 2.63% |
Zheng Yongjun | 3 | 0.09% | 1 | 2.63% |
Christian Riesch | 3 | 0.09% | 2 | 5.26% |
Libo Chen | 2 | 0.06% | 1 | 2.63% |
Gustavo A. R. Silva | 1 | 0.03% | 1 | 2.63% |
Bin Liu | 1 | 0.03% | 1 | 2.63% |
George Cherian | 1 | 0.03% | 1 | 2.63% |
Total | 3259 | 38 |
// SPDX-License-Identifier: GPL-2.0+ /* * DaVinci MDIO Module driver * * Copyright (C) 2010 Texas Instruments. * * Shamelessly ripped out of davinci_emac.c, original copyrights follow: * * Copyright (C) 2009 Texas Instruments. * */ #include <linux/module.h> #include <linux/kernel.h> #include <linux/platform_device.h> #include <linux/delay.h> #include <linux/sched.h> #include <linux/slab.h> #include <linux/phy.h> #include <linux/clk.h> #include <linux/err.h> #include <linux/io.h> #include <linux/iopoll.h> #include <linux/pm_runtime.h> #include <linux/davinci_emac.h> #include <linux/of.h> #include <linux/of_device.h> #include <linux/of_mdio.h> #include <linux/pinctrl/consumer.h> #include <linux/mdio-bitbang.h> #include <linux/sys_soc.h> /* * This timeout definition is a worst-case ultra defensive measure against * unexpected controller lock ups. Ideally, we should never ever hit this * scenario in practice. */ #define MDIO_TIMEOUT 100 /* msecs */ #define PHY_REG_MASK 0x1f #define PHY_ID_MASK 0x1f #define DEF_OUT_FREQ 2200000 /* 2.2 MHz */ struct davinci_mdio_of_param { int autosuspend_delay_ms; bool manual_mode; }; struct davinci_mdio_regs { u32 version; u32 control; #define CONTROL_IDLE BIT(31) #define CONTROL_ENABLE BIT(30) #define CONTROL_MAX_DIV (0xffff) #define CONTROL_CLKDIV GENMASK(15, 0) #define MDIO_MAN_MDCLK_O BIT(2) #define MDIO_MAN_OE BIT(1) #define MDIO_MAN_PIN BIT(0) #define MDIO_MANUALMODE BIT(31) #define MDIO_PIN 0 u32 alive; u32 link; u32 linkintraw; u32 linkintmasked; u32 __reserved_0[2]; u32 userintraw; u32 userintmasked; u32 userintmaskset; u32 userintmaskclr; u32 manualif; u32 poll; u32 __reserved_1[18]; struct { u32 access; #define USERACCESS_GO BIT(31) #define USERACCESS_WRITE BIT(30) #define USERACCESS_ACK BIT(29) #define USERACCESS_READ (0) #define USERACCESS_DATA (0xffff) u32 physel; } user[]; }; static const struct mdio_platform_data default_pdata = { .bus_freq = DEF_OUT_FREQ, }; struct davinci_mdio_data { struct mdio_platform_data pdata; struct mdiobb_ctrl bb_ctrl; struct davinci_mdio_regs __iomem *regs; struct clk *clk; struct device *dev; struct mii_bus *bus; bool active_in_suspend; unsigned long access_time; /* jiffies */ /* Indicates that driver shouldn't modify phy_mask in case * if MDIO bus is registered from DT. */ bool skip_scan; u32 clk_div; bool manual_mode; }; static void davinci_mdio_init_clk(struct davinci_mdio_data *data) { u32 mdio_in, div, mdio_out_khz, access_time; mdio_in = clk_get_rate(data->clk); div = (mdio_in / data->pdata.bus_freq) - 1; if (div > CONTROL_MAX_DIV) div = CONTROL_MAX_DIV; data->clk_div = div; /* * One mdio transaction consists of: * 32 bits of preamble * 32 bits of transferred data * 24 bits of bus yield (not needed unless shared?) */ mdio_out_khz = mdio_in / (1000 * (div + 1)); access_time = (88 * 1000) / mdio_out_khz; /* * In the worst case, we could be kicking off a user-access immediately * after the mdio bus scan state-machine triggered its own read. If * so, our request could get deferred by one access cycle. We * defensively allow for 4 access cycles. */ data->access_time = usecs_to_jiffies(access_time * 4); if (!data->access_time) data->access_time = 1; } static void davinci_mdio_enable(struct davinci_mdio_data *data) { /* set enable and clock divider */ writel(data->clk_div | CONTROL_ENABLE, &data->regs->control); } static void davinci_mdio_disable(struct davinci_mdio_data *data) { u32 reg; /* Disable MDIO state machine */ reg = readl(&data->regs->control); reg &= ~CONTROL_CLKDIV; reg |= data->clk_div; reg &= ~CONTROL_ENABLE; writel(reg, &data->regs->control); } static void davinci_mdio_enable_manual_mode(struct davinci_mdio_data *data) { u32 reg; /* set manual mode */ reg = readl(&data->regs->poll); reg |= MDIO_MANUALMODE; writel(reg, &data->regs->poll); } static void davinci_set_mdc(struct mdiobb_ctrl *ctrl, int level) { struct davinci_mdio_data *data; u32 reg; data = container_of(ctrl, struct davinci_mdio_data, bb_ctrl); reg = readl(&data->regs->manualif); if (level) reg |= MDIO_MAN_MDCLK_O; else reg &= ~MDIO_MAN_MDCLK_O; writel(reg, &data->regs->manualif); } static void davinci_set_mdio_dir(struct mdiobb_ctrl *ctrl, int output) { struct davinci_mdio_data *data; u32 reg; data = container_of(ctrl, struct davinci_mdio_data, bb_ctrl); reg = readl(&data->regs->manualif); if (output) reg |= MDIO_MAN_OE; else reg &= ~MDIO_MAN_OE; writel(reg, &data->regs->manualif); } static void davinci_set_mdio_data(struct mdiobb_ctrl *ctrl, int value) { struct davinci_mdio_data *data; u32 reg; data = container_of(ctrl, struct davinci_mdio_data, bb_ctrl); reg = readl(&data->regs->manualif); if (value) reg |= MDIO_MAN_PIN; else reg &= ~MDIO_MAN_PIN; writel(reg, &data->regs->manualif); } static int davinci_get_mdio_data(struct mdiobb_ctrl *ctrl) { struct davinci_mdio_data *data; unsigned long reg; data = container_of(ctrl, struct davinci_mdio_data, bb_ctrl); reg = readl(&data->regs->manualif); return test_bit(MDIO_PIN, ®); } static int davinci_mdiobb_read(struct mii_bus *bus, int phy, int reg) { int ret; ret = pm_runtime_resume_and_get(bus->parent); if (ret < 0) return ret; ret = mdiobb_read(bus, phy, reg); pm_runtime_mark_last_busy(bus->parent); pm_runtime_put_autosuspend(bus->parent); return ret; } static int davinci_mdiobb_write(struct mii_bus *bus, int phy, int reg, u16 val) { int ret; ret = pm_runtime_resume_and_get(bus->parent); if (ret < 0) return ret; ret = mdiobb_write(bus, phy, reg, val); pm_runtime_mark_last_busy(bus->parent); pm_runtime_put_autosuspend(bus->parent); return ret; } static int davinci_mdio_common_reset(struct davinci_mdio_data *data) { u32 phy_mask, ver; int ret; ret = pm_runtime_resume_and_get(data->dev); if (ret < 0) return ret; if (data->manual_mode) { davinci_mdio_disable(data); davinci_mdio_enable_manual_mode(data); } /* wait for scan logic to settle */ msleep(PHY_MAX_ADDR * data->access_time); /* dump hardware version info */ ver = readl(&data->regs->version); dev_info(data->dev, "davinci mdio revision %d.%d, bus freq %ld\n", (ver >> 8) & 0xff, ver & 0xff, data->pdata.bus_freq); if (data->skip_scan) goto done; /* get phy mask from the alive register */ phy_mask = readl(&data->regs->alive); if (phy_mask) { /* restrict mdio bus to live phys only */ dev_info(data->dev, "detected phy mask %x\n", ~phy_mask); phy_mask = ~phy_mask; } else { /* desperately scan all phys */ dev_warn(data->dev, "no live phy, scanning all\n"); phy_mask = 0; } data->bus->phy_mask = phy_mask; done: pm_runtime_mark_last_busy(data->dev); pm_runtime_put_autosuspend(data->dev); return 0; } static int davinci_mdio_reset(struct mii_bus *bus) { struct davinci_mdio_data *data = bus->priv; return davinci_mdio_common_reset(data); } static int davinci_mdiobb_reset(struct mii_bus *bus) { struct mdiobb_ctrl *ctrl = bus->priv; struct davinci_mdio_data *data; data = container_of(ctrl, struct davinci_mdio_data, bb_ctrl); return davinci_mdio_common_reset(data); } /* wait until hardware is ready for another user access */ static inline int wait_for_user_access(struct davinci_mdio_data *data) { struct davinci_mdio_regs __iomem *regs = data->regs; unsigned long timeout = jiffies + msecs_to_jiffies(MDIO_TIMEOUT); u32 reg; while (time_after(timeout, jiffies)) { reg = readl(®s->user[0].access); if ((reg & USERACCESS_GO) == 0) return 0; reg = readl(®s->control); if ((reg & CONTROL_IDLE) == 0) { usleep_range(100, 200); continue; } /* * An emac soft_reset may have clobbered the mdio controller's * state machine. We need to reset and retry the current * operation */ dev_warn(data->dev, "resetting idled controller\n"); davinci_mdio_enable(data); return -EAGAIN; } reg = readl(®s->user[0].access); if ((reg & USERACCESS_GO) == 0) return 0; dev_err(data->dev, "timed out waiting for user access\n"); return -ETIMEDOUT; } /* wait until hardware state machine is idle */ static inline int wait_for_idle(struct davinci_mdio_data *data) { struct davinci_mdio_regs __iomem *regs = data->regs; u32 val, ret; ret = readl_poll_timeout(®s->control, val, val & CONTROL_IDLE, 0, MDIO_TIMEOUT * 1000); if (ret) dev_err(data->dev, "timed out waiting for idle\n"); return ret; } static int davinci_mdio_read(struct mii_bus *bus, int phy_id, int phy_reg) { struct davinci_mdio_data *data = bus->priv; u32 reg; int ret; if (phy_reg & ~PHY_REG_MASK || phy_id & ~PHY_ID_MASK) return -EINVAL; ret = pm_runtime_resume_and_get(data->dev); if (ret < 0) return ret; reg = (USERACCESS_GO | USERACCESS_READ | (phy_reg << 21) | (phy_id << 16)); while (1) { ret = wait_for_user_access(data); if (ret == -EAGAIN) continue; if (ret < 0) break; writel(reg, &data->regs->user[0].access); ret = wait_for_user_access(data); if (ret == -EAGAIN) continue; if (ret < 0) break; reg = readl(&data->regs->user[0].access); ret = (reg & USERACCESS_ACK) ? (reg & USERACCESS_DATA) : -EIO; break; } pm_runtime_mark_last_busy(data->dev); pm_runtime_put_autosuspend(data->dev); return ret; } static int davinci_mdio_write(struct mii_bus *bus, int phy_id, int phy_reg, u16 phy_data) { struct davinci_mdio_data *data = bus->priv; u32 reg; int ret; if (phy_reg & ~PHY_REG_MASK || phy_id & ~PHY_ID_MASK) return -EINVAL; ret = pm_runtime_resume_and_get(data->dev); if (ret < 0) return ret; reg = (USERACCESS_GO | USERACCESS_WRITE | (phy_reg << 21) | (phy_id << 16) | (phy_data & USERACCESS_DATA)); while (1) { ret = wait_for_user_access(data); if (ret == -EAGAIN) continue; if (ret < 0) break; writel(reg, &data->regs->user[0].access); ret = wait_for_user_access(data); if (ret == -EAGAIN) continue; break; } pm_runtime_mark_last_busy(data->dev); pm_runtime_put_autosuspend(data->dev); return ret; } static int davinci_mdio_probe_dt(struct mdio_platform_data *data, struct platform_device *pdev) { struct device_node *node = pdev->dev.of_node; u32 prop; if (!node) return -EINVAL; if (of_property_read_u32(node, "bus_freq", &prop)) { dev_err(&pdev->dev, "Missing bus_freq property in the DT.\n"); return -EINVAL; } data->bus_freq = prop; return 0; } struct k3_mdio_soc_data { bool manual_mode; }; static const struct k3_mdio_soc_data am65_mdio_soc_data = { .manual_mode = true, }; static const struct soc_device_attribute k3_mdio_socinfo[] = { { .family = "AM62X", .revision = "SR1.0", .data = &am65_mdio_soc_data }, { .family = "AM64X", .revision = "SR1.0", .data = &am65_mdio_soc_data }, { .family = "AM64X", .revision = "SR2.0", .data = &am65_mdio_soc_data }, { .family = "AM65X", .revision = "SR1.0", .data = &am65_mdio_soc_data }, { .family = "AM65X", .revision = "SR2.0", .data = &am65_mdio_soc_data }, { .family = "J7200", .revision = "SR1.0", .data = &am65_mdio_soc_data }, { .family = "J7200", .revision = "SR2.0", .data = &am65_mdio_soc_data }, { .family = "J721E", .revision = "SR1.0", .data = &am65_mdio_soc_data }, { .family = "J721E", .revision = "SR2.0", .data = &am65_mdio_soc_data }, { .family = "J721S2", .revision = "SR1.0", .data = &am65_mdio_soc_data}, { /* sentinel */ }, }; #if IS_ENABLED(CONFIG_OF) static const struct davinci_mdio_of_param of_cpsw_mdio_data = { .autosuspend_delay_ms = 100, }; static const struct of_device_id davinci_mdio_of_mtable[] = { { .compatible = "ti,davinci_mdio", }, { .compatible = "ti,cpsw-mdio", .data = &of_cpsw_mdio_data}, { /* sentinel */ }, }; MODULE_DEVICE_TABLE(of, davinci_mdio_of_mtable); #endif static const struct mdiobb_ops davinci_mdiobb_ops = { .owner = THIS_MODULE, .set_mdc = davinci_set_mdc, .set_mdio_dir = davinci_set_mdio_dir, .set_mdio_data = davinci_set_mdio_data, .get_mdio_data = davinci_get_mdio_data, }; static int davinci_mdio_probe(struct platform_device *pdev) { struct mdio_platform_data *pdata = dev_get_platdata(&pdev->dev); struct device *dev = &pdev->dev; struct davinci_mdio_data *data; struct resource *res; struct phy_device *phy; int ret, addr; int autosuspend_delay_ms = -1; data = devm_kzalloc(dev, sizeof(*data), GFP_KERNEL); if (!data) return -ENOMEM; data->manual_mode = false; data->bb_ctrl.ops = &davinci_mdiobb_ops; if (IS_ENABLED(CONFIG_OF) && dev->of_node) { const struct soc_device_attribute *soc_match_data; soc_match_data = soc_device_match(k3_mdio_socinfo); if (soc_match_data && soc_match_data->data) { const struct k3_mdio_soc_data *socdata = soc_match_data->data; data->manual_mode = socdata->manual_mode; } } if (data->manual_mode) data->bus = alloc_mdio_bitbang(&data->bb_ctrl); else data->bus = devm_mdiobus_alloc(dev); if (!data->bus) { dev_err(dev, "failed to alloc mii bus\n"); return -ENOMEM; } if (IS_ENABLED(CONFIG_OF) && dev->of_node) { const struct davinci_mdio_of_param *of_mdio_data; ret = davinci_mdio_probe_dt(&data->pdata, pdev); if (ret) return ret; snprintf(data->bus->id, MII_BUS_ID_SIZE, "%s", pdev->name); of_mdio_data = of_device_get_match_data(&pdev->dev); if (of_mdio_data) { autosuspend_delay_ms = of_mdio_data->autosuspend_delay_ms; } } else { data->pdata = pdata ? (*pdata) : default_pdata; snprintf(data->bus->id, MII_BUS_ID_SIZE, "%s-%x", pdev->name, pdev->id); } data->bus->name = dev_name(dev); if (data->manual_mode) { data->bus->read = davinci_mdiobb_read; data->bus->write = davinci_mdiobb_write; data->bus->reset = davinci_mdiobb_reset; dev_info(dev, "Configuring MDIO in manual mode\n"); } else { data->bus->read = davinci_mdio_read; data->bus->write = davinci_mdio_write; data->bus->reset = davinci_mdio_reset; data->bus->priv = data; } data->bus->parent = dev; data->clk = devm_clk_get(dev, "fck"); if (IS_ERR(data->clk)) { dev_err(dev, "failed to get device clock\n"); return PTR_ERR(data->clk); } dev_set_drvdata(dev, data); data->dev = dev; res = platform_get_resource(pdev, IORESOURCE_MEM, 0); if (!res) return -EINVAL; data->regs = devm_ioremap(dev, res->start, resource_size(res)); if (!data->regs) return -ENOMEM; davinci_mdio_init_clk(data); pm_runtime_set_autosuspend_delay(&pdev->dev, autosuspend_delay_ms); pm_runtime_use_autosuspend(&pdev->dev); pm_runtime_enable(&pdev->dev); /* register the mii bus * Create PHYs from DT only in case if PHY child nodes are explicitly * defined to support backward compatibility with DTs which assume that * Davinci MDIO will always scan the bus for PHYs detection. */ if (dev->of_node && of_get_child_count(dev->of_node)) data->skip_scan = true; ret = of_mdiobus_register(data->bus, dev->of_node); if (ret) goto bail_out; /* scan and dump the bus */ for (addr = 0; addr < PHY_MAX_ADDR; addr++) { phy = mdiobus_get_phy(data->bus, addr); if (phy) { dev_info(dev, "phy[%d]: device %s, driver %s\n", phy->mdio.addr, phydev_name(phy), phy->drv ? phy->drv->name : "unknown"); } } return 0; bail_out: pm_runtime_dont_use_autosuspend(&pdev->dev); pm_runtime_disable(&pdev->dev); return ret; } static int davinci_mdio_remove(struct platform_device *pdev) { struct davinci_mdio_data *data = platform_get_drvdata(pdev); if (data->bus) { mdiobus_unregister(data->bus); if (data->manual_mode) free_mdio_bitbang(data->bus); } pm_runtime_dont_use_autosuspend(&pdev->dev); pm_runtime_disable(&pdev->dev); return 0; } #ifdef CONFIG_PM static int davinci_mdio_runtime_suspend(struct device *dev) { struct davinci_mdio_data *data = dev_get_drvdata(dev); u32 ctrl; /* shutdown the scan state machine */ ctrl = readl(&data->regs->control); ctrl &= ~CONTROL_ENABLE; writel(ctrl, &data->regs->control); if (!data->manual_mode) wait_for_idle(data); return 0; } static int davinci_mdio_runtime_resume(struct device *dev) { struct davinci_mdio_data *data = dev_get_drvdata(dev); if (data->manual_mode) { davinci_mdio_disable(data); davinci_mdio_enable_manual_mode(data); } else { davinci_mdio_enable(data); } return 0; } #endif #ifdef CONFIG_PM_SLEEP static int davinci_mdio_suspend(struct device *dev) { struct davinci_mdio_data *data = dev_get_drvdata(dev); int ret = 0; data->active_in_suspend = !pm_runtime_status_suspended(dev); if (data->active_in_suspend) ret = pm_runtime_force_suspend(dev); if (ret < 0) return ret; /* Select sleep pin state */ pinctrl_pm_select_sleep_state(dev); return 0; } static int davinci_mdio_resume(struct device *dev) { struct davinci_mdio_data *data = dev_get_drvdata(dev); /* Select default pin state */ pinctrl_pm_select_default_state(dev); if (data->active_in_suspend) pm_runtime_force_resume(dev); return 0; } #endif static const struct dev_pm_ops davinci_mdio_pm_ops = { SET_RUNTIME_PM_OPS(davinci_mdio_runtime_suspend, davinci_mdio_runtime_resume, NULL) SET_LATE_SYSTEM_SLEEP_PM_OPS(davinci_mdio_suspend, davinci_mdio_resume) }; static struct platform_driver davinci_mdio_driver = { .driver = { .name = "davinci_mdio", .pm = &davinci_mdio_pm_ops, .of_match_table = of_match_ptr(davinci_mdio_of_mtable), }, .probe = davinci_mdio_probe, .remove = davinci_mdio_remove, }; static int __init davinci_mdio_init(void) { return platform_driver_register(&davinci_mdio_driver); } device_initcall(davinci_mdio_init); static void __exit davinci_mdio_exit(void) { platform_driver_unregister(&davinci_mdio_driver); } module_exit(davinci_mdio_exit); MODULE_LICENSE("GPL"); MODULE_DESCRIPTION("DaVinci MDIO driver");
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1