Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Jose Abreu | 2802 | 93.06% | 34 | 73.91% |
Giuseppe Cavallaro | 156 | 5.18% | 5 | 10.87% |
Ong Boon Leong | 25 | 0.83% | 2 | 4.35% |
Thierry Reding | 10 | 0.33% | 1 | 2.17% |
Niklas Cassel | 8 | 0.27% | 1 | 2.17% |
Herve Codina | 4 | 0.13% | 1 | 2.17% |
Alexandre Torgue | 4 | 0.13% | 1 | 2.17% |
Deepak Sikri | 2 | 0.07% | 1 | 2.17% |
Total | 3011 | 46 |
// SPDX-License-Identifier: (GPL-2.0 OR MIT) /* * Copyright (c) 2018 Synopsys, Inc. and/or its affiliates. * stmmac XGMAC support. */ #include <linux/iopoll.h> #include "stmmac.h" #include "dwxgmac2.h" static int dwxgmac2_dma_reset(void __iomem *ioaddr) { u32 value = readl(ioaddr + XGMAC_DMA_MODE); /* DMA SW reset */ writel(value | XGMAC_SWR, ioaddr + XGMAC_DMA_MODE); return readl_poll_timeout(ioaddr + XGMAC_DMA_MODE, value, !(value & XGMAC_SWR), 0, 100000); } static void dwxgmac2_dma_init(void __iomem *ioaddr, struct stmmac_dma_cfg *dma_cfg, int atds) { u32 value = readl(ioaddr + XGMAC_DMA_SYSBUS_MODE); if (dma_cfg->aal) value |= XGMAC_AAL; if (dma_cfg->eame) value |= XGMAC_EAME; writel(value, ioaddr + XGMAC_DMA_SYSBUS_MODE); } static void dwxgmac2_dma_init_chan(void __iomem *ioaddr, struct stmmac_dma_cfg *dma_cfg, u32 chan) { u32 value = readl(ioaddr + XGMAC_DMA_CH_CONTROL(chan)); if (dma_cfg->pblx8) value |= XGMAC_PBLx8; writel(value, ioaddr + XGMAC_DMA_CH_CONTROL(chan)); writel(XGMAC_DMA_INT_DEFAULT_EN, ioaddr + XGMAC_DMA_CH_INT_EN(chan)); } static void dwxgmac2_dma_init_rx_chan(void __iomem *ioaddr, struct stmmac_dma_cfg *dma_cfg, dma_addr_t phy, u32 chan) { u32 rxpbl = dma_cfg->rxpbl ?: dma_cfg->pbl; u32 value; value = readl(ioaddr + XGMAC_DMA_CH_RX_CONTROL(chan)); value &= ~XGMAC_RxPBL; value |= (rxpbl << XGMAC_RxPBL_SHIFT) & XGMAC_RxPBL; writel(value, ioaddr + XGMAC_DMA_CH_RX_CONTROL(chan)); writel(upper_32_bits(phy), ioaddr + XGMAC_DMA_CH_RxDESC_HADDR(chan)); writel(lower_32_bits(phy), ioaddr + XGMAC_DMA_CH_RxDESC_LADDR(chan)); } static void dwxgmac2_dma_init_tx_chan(void __iomem *ioaddr, struct stmmac_dma_cfg *dma_cfg, dma_addr_t phy, u32 chan) { u32 txpbl = dma_cfg->txpbl ?: dma_cfg->pbl; u32 value; value = readl(ioaddr + XGMAC_DMA_CH_TX_CONTROL(chan)); value &= ~XGMAC_TxPBL; value |= (txpbl << XGMAC_TxPBL_SHIFT) & XGMAC_TxPBL; value |= XGMAC_OSP; writel(value, ioaddr + XGMAC_DMA_CH_TX_CONTROL(chan)); writel(upper_32_bits(phy), ioaddr + XGMAC_DMA_CH_TxDESC_HADDR(chan)); writel(lower_32_bits(phy), ioaddr + XGMAC_DMA_CH_TxDESC_LADDR(chan)); } static void dwxgmac2_dma_axi(void __iomem *ioaddr, struct stmmac_axi *axi) { u32 value = readl(ioaddr + XGMAC_DMA_SYSBUS_MODE); int i; if (axi->axi_lpi_en) value |= XGMAC_EN_LPI; if (axi->axi_xit_frm) value |= XGMAC_LPI_XIT_PKT; value &= ~XGMAC_WR_OSR_LMT; value |= (axi->axi_wr_osr_lmt << XGMAC_WR_OSR_LMT_SHIFT) & XGMAC_WR_OSR_LMT; value &= ~XGMAC_RD_OSR_LMT; value |= (axi->axi_rd_osr_lmt << XGMAC_RD_OSR_LMT_SHIFT) & XGMAC_RD_OSR_LMT; if (!axi->axi_fb) value |= XGMAC_UNDEF; value &= ~XGMAC_BLEN; for (i = 0; i < AXI_BLEN; i++) { switch (axi->axi_blen[i]) { case 256: value |= XGMAC_BLEN256; break; case 128: value |= XGMAC_BLEN128; break; case 64: value |= XGMAC_BLEN64; break; case 32: value |= XGMAC_BLEN32; break; case 16: value |= XGMAC_BLEN16; break; case 8: value |= XGMAC_BLEN8; break; case 4: value |= XGMAC_BLEN4; break; } } writel(value, ioaddr + XGMAC_DMA_SYSBUS_MODE); writel(XGMAC_TDPS, ioaddr + XGMAC_TX_EDMA_CTRL); writel(XGMAC_RDPS, ioaddr + XGMAC_RX_EDMA_CTRL); } static void dwxgmac2_dma_dump_regs(void __iomem *ioaddr, u32 *reg_space) { int i; for (i = (XGMAC_DMA_MODE / 4); i < XGMAC_REGSIZE; i++) reg_space[i] = readl(ioaddr + i * 4); } static void dwxgmac2_dma_rx_mode(void __iomem *ioaddr, int mode, u32 channel, int fifosz, u8 qmode) { u32 value = readl(ioaddr + XGMAC_MTL_RXQ_OPMODE(channel)); unsigned int rqs = fifosz / 256 - 1; if (mode == SF_DMA_MODE) { value |= XGMAC_RSF; } else { value &= ~XGMAC_RSF; value &= ~XGMAC_RTC; if (mode <= 64) value |= 0x0 << XGMAC_RTC_SHIFT; else if (mode <= 96) value |= 0x2 << XGMAC_RTC_SHIFT; else value |= 0x3 << XGMAC_RTC_SHIFT; } value &= ~XGMAC_RQS; value |= (rqs << XGMAC_RQS_SHIFT) & XGMAC_RQS; if ((fifosz >= 4096) && (qmode != MTL_QUEUE_AVB)) { u32 flow = readl(ioaddr + XGMAC_MTL_RXQ_FLOW_CONTROL(channel)); unsigned int rfd, rfa; value |= XGMAC_EHFC; /* Set Threshold for Activating Flow Control to min 2 frames, * i.e. 1500 * 2 = 3000 bytes. * * Set Threshold for Deactivating Flow Control to min 1 frame, * i.e. 1500 bytes. */ switch (fifosz) { case 4096: /* This violates the above formula because of FIFO size * limit therefore overflow may occur in spite of this. */ rfd = 0x03; /* Full-2.5K */ rfa = 0x01; /* Full-1.5K */ break; default: rfd = 0x07; /* Full-4.5K */ rfa = 0x04; /* Full-3K */ break; } flow &= ~XGMAC_RFD; flow |= rfd << XGMAC_RFD_SHIFT; flow &= ~XGMAC_RFA; flow |= rfa << XGMAC_RFA_SHIFT; writel(flow, ioaddr + XGMAC_MTL_RXQ_FLOW_CONTROL(channel)); } writel(value, ioaddr + XGMAC_MTL_RXQ_OPMODE(channel)); /* Enable MTL RX overflow */ value = readl(ioaddr + XGMAC_MTL_QINTEN(channel)); writel(value | XGMAC_RXOIE, ioaddr + XGMAC_MTL_QINTEN(channel)); } static void dwxgmac2_dma_tx_mode(void __iomem *ioaddr, int mode, u32 channel, int fifosz, u8 qmode) { u32 value = readl(ioaddr + XGMAC_MTL_TXQ_OPMODE(channel)); unsigned int tqs = fifosz / 256 - 1; if (mode == SF_DMA_MODE) { value |= XGMAC_TSF; } else { value &= ~XGMAC_TSF; value &= ~XGMAC_TTC; if (mode <= 64) value |= 0x0 << XGMAC_TTC_SHIFT; else if (mode <= 96) value |= 0x2 << XGMAC_TTC_SHIFT; else if (mode <= 128) value |= 0x3 << XGMAC_TTC_SHIFT; else if (mode <= 192) value |= 0x4 << XGMAC_TTC_SHIFT; else if (mode <= 256) value |= 0x5 << XGMAC_TTC_SHIFT; else if (mode <= 384) value |= 0x6 << XGMAC_TTC_SHIFT; else value |= 0x7 << XGMAC_TTC_SHIFT; } /* Use static TC to Queue mapping */ value |= (channel << XGMAC_Q2TCMAP_SHIFT) & XGMAC_Q2TCMAP; value &= ~XGMAC_TXQEN; if (qmode != MTL_QUEUE_AVB) value |= 0x2 << XGMAC_TXQEN_SHIFT; else value |= 0x1 << XGMAC_TXQEN_SHIFT; value &= ~XGMAC_TQS; value |= (tqs << XGMAC_TQS_SHIFT) & XGMAC_TQS; writel(value, ioaddr + XGMAC_MTL_TXQ_OPMODE(channel)); } static void dwxgmac2_enable_dma_irq(void __iomem *ioaddr, u32 chan, bool rx, bool tx) { u32 value = readl(ioaddr + XGMAC_DMA_CH_INT_EN(chan)); if (rx) value |= XGMAC_DMA_INT_DEFAULT_RX; if (tx) value |= XGMAC_DMA_INT_DEFAULT_TX; writel(value, ioaddr + XGMAC_DMA_CH_INT_EN(chan)); } static void dwxgmac2_disable_dma_irq(void __iomem *ioaddr, u32 chan, bool rx, bool tx) { u32 value = readl(ioaddr + XGMAC_DMA_CH_INT_EN(chan)); if (rx) value &= ~XGMAC_DMA_INT_DEFAULT_RX; if (tx) value &= ~XGMAC_DMA_INT_DEFAULT_TX; writel(value, ioaddr + XGMAC_DMA_CH_INT_EN(chan)); } static void dwxgmac2_dma_start_tx(void __iomem *ioaddr, u32 chan) { u32 value; value = readl(ioaddr + XGMAC_DMA_CH_TX_CONTROL(chan)); value |= XGMAC_TXST; writel(value, ioaddr + XGMAC_DMA_CH_TX_CONTROL(chan)); value = readl(ioaddr + XGMAC_TX_CONFIG); value |= XGMAC_CONFIG_TE; writel(value, ioaddr + XGMAC_TX_CONFIG); } static void dwxgmac2_dma_stop_tx(void __iomem *ioaddr, u32 chan) { u32 value; value = readl(ioaddr + XGMAC_DMA_CH_TX_CONTROL(chan)); value &= ~XGMAC_TXST; writel(value, ioaddr + XGMAC_DMA_CH_TX_CONTROL(chan)); value = readl(ioaddr + XGMAC_TX_CONFIG); value &= ~XGMAC_CONFIG_TE; writel(value, ioaddr + XGMAC_TX_CONFIG); } static void dwxgmac2_dma_start_rx(void __iomem *ioaddr, u32 chan) { u32 value; value = readl(ioaddr + XGMAC_DMA_CH_RX_CONTROL(chan)); value |= XGMAC_RXST; writel(value, ioaddr + XGMAC_DMA_CH_RX_CONTROL(chan)); value = readl(ioaddr + XGMAC_RX_CONFIG); value |= XGMAC_CONFIG_RE; writel(value, ioaddr + XGMAC_RX_CONFIG); } static void dwxgmac2_dma_stop_rx(void __iomem *ioaddr, u32 chan) { u32 value; value = readl(ioaddr + XGMAC_DMA_CH_RX_CONTROL(chan)); value &= ~XGMAC_RXST; writel(value, ioaddr + XGMAC_DMA_CH_RX_CONTROL(chan)); } static int dwxgmac2_dma_interrupt(void __iomem *ioaddr, struct stmmac_extra_stats *x, u32 chan, u32 dir) { u32 intr_status = readl(ioaddr + XGMAC_DMA_CH_STATUS(chan)); u32 intr_en = readl(ioaddr + XGMAC_DMA_CH_INT_EN(chan)); int ret = 0; if (dir == DMA_DIR_RX) intr_status &= XGMAC_DMA_STATUS_MSK_RX; else if (dir == DMA_DIR_TX) intr_status &= XGMAC_DMA_STATUS_MSK_TX; /* ABNORMAL interrupts */ if (unlikely(intr_status & XGMAC_AIS)) { if (unlikely(intr_status & XGMAC_RBU)) { x->rx_buf_unav_irq++; ret |= handle_rx; } if (unlikely(intr_status & XGMAC_TPS)) { x->tx_process_stopped_irq++; ret |= tx_hard_error; } if (unlikely(intr_status & XGMAC_FBE)) { x->fatal_bus_error_irq++; ret |= tx_hard_error; } } /* TX/RX NORMAL interrupts */ if (likely(intr_status & XGMAC_NIS)) { x->normal_irq_n++; if (likely(intr_status & XGMAC_RI)) { x->rx_normal_irq_n++; ret |= handle_rx; } if (likely(intr_status & (XGMAC_TI | XGMAC_TBU))) { x->tx_normal_irq_n++; ret |= handle_tx; } } /* Clear interrupts */ writel(intr_en & intr_status, ioaddr + XGMAC_DMA_CH_STATUS(chan)); return ret; } static int dwxgmac2_get_hw_feature(void __iomem *ioaddr, struct dma_features *dma_cap) { u32 hw_cap; /* MAC HW feature 0 */ hw_cap = readl(ioaddr + XGMAC_HW_FEATURE0); dma_cap->vlins = (hw_cap & XGMAC_HWFEAT_SAVLANINS) >> 27; dma_cap->rx_coe = (hw_cap & XGMAC_HWFEAT_RXCOESEL) >> 16; dma_cap->tx_coe = (hw_cap & XGMAC_HWFEAT_TXCOESEL) >> 14; dma_cap->eee = (hw_cap & XGMAC_HWFEAT_EEESEL) >> 13; dma_cap->atime_stamp = (hw_cap & XGMAC_HWFEAT_TSSEL) >> 12; dma_cap->av = (hw_cap & XGMAC_HWFEAT_AVSEL) >> 11; dma_cap->av &= !((hw_cap & XGMAC_HWFEAT_RAVSEL) >> 10); dma_cap->arpoffsel = (hw_cap & XGMAC_HWFEAT_ARPOFFSEL) >> 9; dma_cap->rmon = (hw_cap & XGMAC_HWFEAT_MMCSEL) >> 8; dma_cap->pmt_magic_frame = (hw_cap & XGMAC_HWFEAT_MGKSEL) >> 7; dma_cap->pmt_remote_wake_up = (hw_cap & XGMAC_HWFEAT_RWKSEL) >> 6; dma_cap->vlhash = (hw_cap & XGMAC_HWFEAT_VLHASH) >> 4; dma_cap->mbps_1000 = (hw_cap & XGMAC_HWFEAT_GMIISEL) >> 1; /* MAC HW feature 1 */ hw_cap = readl(ioaddr + XGMAC_HW_FEATURE1); dma_cap->l3l4fnum = (hw_cap & XGMAC_HWFEAT_L3L4FNUM) >> 27; dma_cap->hash_tb_sz = (hw_cap & XGMAC_HWFEAT_HASHTBLSZ) >> 24; dma_cap->rssen = (hw_cap & XGMAC_HWFEAT_RSSEN) >> 20; dma_cap->tsoen = (hw_cap & XGMAC_HWFEAT_TSOEN) >> 18; dma_cap->sphen = (hw_cap & XGMAC_HWFEAT_SPHEN) >> 17; dma_cap->addr64 = (hw_cap & XGMAC_HWFEAT_ADDR64) >> 14; switch (dma_cap->addr64) { case 0: dma_cap->addr64 = 32; break; case 1: dma_cap->addr64 = 40; break; case 2: dma_cap->addr64 = 48; break; default: dma_cap->addr64 = 32; break; } dma_cap->tx_fifo_size = 128 << ((hw_cap & XGMAC_HWFEAT_TXFIFOSIZE) >> 6); dma_cap->rx_fifo_size = 128 << ((hw_cap & XGMAC_HWFEAT_RXFIFOSIZE) >> 0); /* MAC HW feature 2 */ hw_cap = readl(ioaddr + XGMAC_HW_FEATURE2); dma_cap->pps_out_num = (hw_cap & XGMAC_HWFEAT_PPSOUTNUM) >> 24; dma_cap->number_tx_channel = ((hw_cap & XGMAC_HWFEAT_TXCHCNT) >> 18) + 1; dma_cap->number_rx_channel = ((hw_cap & XGMAC_HWFEAT_RXCHCNT) >> 12) + 1; dma_cap->number_tx_queues = ((hw_cap & XGMAC_HWFEAT_TXQCNT) >> 6) + 1; dma_cap->number_rx_queues = ((hw_cap & XGMAC_HWFEAT_RXQCNT) >> 0) + 1; /* MAC HW feature 3 */ hw_cap = readl(ioaddr + XGMAC_HW_FEATURE3); dma_cap->tbssel = (hw_cap & XGMAC_HWFEAT_TBSSEL) >> 27; dma_cap->fpesel = (hw_cap & XGMAC_HWFEAT_FPESEL) >> 26; dma_cap->estwid = (hw_cap & XGMAC_HWFEAT_ESTWID) >> 23; dma_cap->estdep = (hw_cap & XGMAC_HWFEAT_ESTDEP) >> 20; dma_cap->estsel = (hw_cap & XGMAC_HWFEAT_ESTSEL) >> 19; dma_cap->asp = (hw_cap & XGMAC_HWFEAT_ASP) >> 14; dma_cap->dvlan = (hw_cap & XGMAC_HWFEAT_DVLAN) >> 13; dma_cap->frpes = (hw_cap & XGMAC_HWFEAT_FRPES) >> 11; dma_cap->frpbs = (hw_cap & XGMAC_HWFEAT_FRPPB) >> 9; dma_cap->frpsel = (hw_cap & XGMAC_HWFEAT_FRPSEL) >> 3; return 0; } static void dwxgmac2_rx_watchdog(void __iomem *ioaddr, u32 riwt, u32 queue) { writel(riwt & XGMAC_RWT, ioaddr + XGMAC_DMA_CH_Rx_WATCHDOG(queue)); } static void dwxgmac2_set_rx_ring_len(void __iomem *ioaddr, u32 len, u32 chan) { writel(len, ioaddr + XGMAC_DMA_CH_RxDESC_RING_LEN(chan)); } static void dwxgmac2_set_tx_ring_len(void __iomem *ioaddr, u32 len, u32 chan) { writel(len, ioaddr + XGMAC_DMA_CH_TxDESC_RING_LEN(chan)); } static void dwxgmac2_set_rx_tail_ptr(void __iomem *ioaddr, u32 ptr, u32 chan) { writel(ptr, ioaddr + XGMAC_DMA_CH_RxDESC_TAIL_LPTR(chan)); } static void dwxgmac2_set_tx_tail_ptr(void __iomem *ioaddr, u32 ptr, u32 chan) { writel(ptr, ioaddr + XGMAC_DMA_CH_TxDESC_TAIL_LPTR(chan)); } static void dwxgmac2_enable_tso(void __iomem *ioaddr, bool en, u32 chan) { u32 value = readl(ioaddr + XGMAC_DMA_CH_TX_CONTROL(chan)); if (en) value |= XGMAC_TSE; else value &= ~XGMAC_TSE; writel(value, ioaddr + XGMAC_DMA_CH_TX_CONTROL(chan)); } static void dwxgmac2_qmode(void __iomem *ioaddr, u32 channel, u8 qmode) { u32 value = readl(ioaddr + XGMAC_MTL_TXQ_OPMODE(channel)); u32 flow = readl(ioaddr + XGMAC_RX_FLOW_CTRL); value &= ~XGMAC_TXQEN; if (qmode != MTL_QUEUE_AVB) { value |= 0x2 << XGMAC_TXQEN_SHIFT; writel(0, ioaddr + XGMAC_MTL_TCx_ETS_CONTROL(channel)); } else { value |= 0x1 << XGMAC_TXQEN_SHIFT; writel(flow & (~XGMAC_RFE), ioaddr + XGMAC_RX_FLOW_CTRL); } writel(value, ioaddr + XGMAC_MTL_TXQ_OPMODE(channel)); } static void dwxgmac2_set_bfsize(void __iomem *ioaddr, int bfsize, u32 chan) { u32 value; value = readl(ioaddr + XGMAC_DMA_CH_RX_CONTROL(chan)); value &= ~XGMAC_RBSZ; value |= bfsize << XGMAC_RBSZ_SHIFT; writel(value, ioaddr + XGMAC_DMA_CH_RX_CONTROL(chan)); } static void dwxgmac2_enable_sph(void __iomem *ioaddr, bool en, u32 chan) { u32 value = readl(ioaddr + XGMAC_RX_CONFIG); value &= ~XGMAC_CONFIG_HDSMS; value |= XGMAC_CONFIG_HDSMS_256; /* Segment max 256 bytes */ writel(value, ioaddr + XGMAC_RX_CONFIG); value = readl(ioaddr + XGMAC_DMA_CH_CONTROL(chan)); if (en) value |= XGMAC_SPH; else value &= ~XGMAC_SPH; writel(value, ioaddr + XGMAC_DMA_CH_CONTROL(chan)); } static int dwxgmac2_enable_tbs(void __iomem *ioaddr, bool en, u32 chan) { u32 value = readl(ioaddr + XGMAC_DMA_CH_TX_CONTROL(chan)); if (en) value |= XGMAC_EDSE; else value &= ~XGMAC_EDSE; writel(value, ioaddr + XGMAC_DMA_CH_TX_CONTROL(chan)); value = readl(ioaddr + XGMAC_DMA_CH_TX_CONTROL(chan)) & XGMAC_EDSE; if (en && !value) return -EIO; writel(XGMAC_DEF_FTOS, ioaddr + XGMAC_DMA_TBS_CTRL0); writel(XGMAC_DEF_FTOS, ioaddr + XGMAC_DMA_TBS_CTRL1); writel(XGMAC_DEF_FTOS, ioaddr + XGMAC_DMA_TBS_CTRL2); writel(XGMAC_DEF_FTOS, ioaddr + XGMAC_DMA_TBS_CTRL3); return 0; } const struct stmmac_dma_ops dwxgmac210_dma_ops = { .reset = dwxgmac2_dma_reset, .init = dwxgmac2_dma_init, .init_chan = dwxgmac2_dma_init_chan, .init_rx_chan = dwxgmac2_dma_init_rx_chan, .init_tx_chan = dwxgmac2_dma_init_tx_chan, .axi = dwxgmac2_dma_axi, .dump_regs = dwxgmac2_dma_dump_regs, .dma_rx_mode = dwxgmac2_dma_rx_mode, .dma_tx_mode = dwxgmac2_dma_tx_mode, .enable_dma_irq = dwxgmac2_enable_dma_irq, .disable_dma_irq = dwxgmac2_disable_dma_irq, .start_tx = dwxgmac2_dma_start_tx, .stop_tx = dwxgmac2_dma_stop_tx, .start_rx = dwxgmac2_dma_start_rx, .stop_rx = dwxgmac2_dma_stop_rx, .dma_interrupt = dwxgmac2_dma_interrupt, .get_hw_feature = dwxgmac2_get_hw_feature, .rx_watchdog = dwxgmac2_rx_watchdog, .set_rx_ring_len = dwxgmac2_set_rx_ring_len, .set_tx_ring_len = dwxgmac2_set_tx_ring_len, .set_rx_tail_ptr = dwxgmac2_set_rx_tail_ptr, .set_tx_tail_ptr = dwxgmac2_set_tx_tail_ptr, .enable_tso = dwxgmac2_enable_tso, .qmode = dwxgmac2_qmode, .set_bfsize = dwxgmac2_set_bfsize, .enable_sph = dwxgmac2_enable_sph, .enable_tbs = dwxgmac2_enable_tbs, };
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1