Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Jie Deng | 13526 | 99.93% | 3 | 42.86% |
Krzysztof Kozlowski | 4 | 0.03% | 1 | 14.29% |
Arnd Bergmann | 3 | 0.02% | 1 | 14.29% |
Florian Westphal | 2 | 0.01% | 1 | 14.29% |
Jakub Kiciński | 1 | 0.01% | 1 | 14.29% |
Total | 13536 | 7 |
/* Synopsys DesignWare Core Enterprise Ethernet (XLGMAC) Driver * * Copyright (c) 2017 Synopsys, Inc. (www.synopsys.com) * * This program is dual-licensed; you may select either version 2 of * the GNU General Public License ("GPL") or BSD license ("BSD"). * * This Synopsys DWC XLGMAC software driver and associated documentation * (hereinafter the "Software") is an unsupported proprietary work of * Synopsys, Inc. unless otherwise expressly agreed to in writing between * Synopsys and you. The Software IS NOT an item of Licensed Software or a * Licensed Product under any End User Software License Agreement or * Agreement for Licensed Products with Synopsys or any supplement thereto. * Synopsys is a registered trademark of Synopsys, Inc. Other names included * in the SOFTWARE may be the trademarks of their respective owners. */ #include <linux/phy.h> #include <linux/mdio.h> #include <linux/clk.h> #include <linux/bitrev.h> #include <linux/crc32.h> #include <linux/crc32poly.h> #include <linux/dcbnl.h> #include "dwc-xlgmac.h" #include "dwc-xlgmac-reg.h" static int xlgmac_tx_complete(struct xlgmac_dma_desc *dma_desc) { return !XLGMAC_GET_REG_BITS_LE(dma_desc->desc3, TX_NORMAL_DESC3_OWN_POS, TX_NORMAL_DESC3_OWN_LEN); } static int xlgmac_disable_rx_csum(struct xlgmac_pdata *pdata) { u32 regval; regval = readl(pdata->mac_regs + MAC_RCR); regval = XLGMAC_SET_REG_BITS(regval, MAC_RCR_IPC_POS, MAC_RCR_IPC_LEN, 0); writel(regval, pdata->mac_regs + MAC_RCR); return 0; } static int xlgmac_enable_rx_csum(struct xlgmac_pdata *pdata) { u32 regval; regval = readl(pdata->mac_regs + MAC_RCR); regval = XLGMAC_SET_REG_BITS(regval, MAC_RCR_IPC_POS, MAC_RCR_IPC_LEN, 1); writel(regval, pdata->mac_regs + MAC_RCR); return 0; } static int xlgmac_set_mac_address(struct xlgmac_pdata *pdata, const u8 *addr) { unsigned int mac_addr_hi, mac_addr_lo; mac_addr_hi = (addr[5] << 8) | (addr[4] << 0); mac_addr_lo = (addr[3] << 24) | (addr[2] << 16) | (addr[1] << 8) | (addr[0] << 0); writel(mac_addr_hi, pdata->mac_regs + MAC_MACA0HR); writel(mac_addr_lo, pdata->mac_regs + MAC_MACA0LR); return 0; } static void xlgmac_set_mac_reg(struct xlgmac_pdata *pdata, struct netdev_hw_addr *ha, unsigned int *mac_reg) { unsigned int mac_addr_hi, mac_addr_lo; u8 *mac_addr; mac_addr_lo = 0; mac_addr_hi = 0; if (ha) { mac_addr = (u8 *)&mac_addr_lo; mac_addr[0] = ha->addr[0]; mac_addr[1] = ha->addr[1]; mac_addr[2] = ha->addr[2]; mac_addr[3] = ha->addr[3]; mac_addr = (u8 *)&mac_addr_hi; mac_addr[0] = ha->addr[4]; mac_addr[1] = ha->addr[5]; netif_dbg(pdata, drv, pdata->netdev, "adding mac address %pM at %#x\n", ha->addr, *mac_reg); mac_addr_hi = XLGMAC_SET_REG_BITS(mac_addr_hi, MAC_MACA1HR_AE_POS, MAC_MACA1HR_AE_LEN, 1); } writel(mac_addr_hi, pdata->mac_regs + *mac_reg); *mac_reg += MAC_MACA_INC; writel(mac_addr_lo, pdata->mac_regs + *mac_reg); *mac_reg += MAC_MACA_INC; } static int xlgmac_enable_rx_vlan_stripping(struct xlgmac_pdata *pdata) { u32 regval; regval = readl(pdata->mac_regs + MAC_VLANTR); /* Put the VLAN tag in the Rx descriptor */ regval = XLGMAC_SET_REG_BITS(regval, MAC_VLANTR_EVLRXS_POS, MAC_VLANTR_EVLRXS_LEN, 1); /* Don't check the VLAN type */ regval = XLGMAC_SET_REG_BITS(regval, MAC_VLANTR_DOVLTC_POS, MAC_VLANTR_DOVLTC_LEN, 1); /* Check only C-TAG (0x8100) packets */ regval = XLGMAC_SET_REG_BITS(regval, MAC_VLANTR_ERSVLM_POS, MAC_VLANTR_ERSVLM_LEN, 0); /* Don't consider an S-TAG (0x88A8) packet as a VLAN packet */ regval = XLGMAC_SET_REG_BITS(regval, MAC_VLANTR_ESVL_POS, MAC_VLANTR_ESVL_LEN, 0); /* Enable VLAN tag stripping */ regval = XLGMAC_SET_REG_BITS(regval, MAC_VLANTR_EVLS_POS, MAC_VLANTR_EVLS_LEN, 0x3); writel(regval, pdata->mac_regs + MAC_VLANTR); return 0; } static int xlgmac_disable_rx_vlan_stripping(struct xlgmac_pdata *pdata) { u32 regval; regval = readl(pdata->mac_regs + MAC_VLANTR); regval = XLGMAC_SET_REG_BITS(regval, MAC_VLANTR_EVLS_POS, MAC_VLANTR_EVLS_LEN, 0); writel(regval, pdata->mac_regs + MAC_VLANTR); return 0; } static int xlgmac_enable_rx_vlan_filtering(struct xlgmac_pdata *pdata) { u32 regval; regval = readl(pdata->mac_regs + MAC_PFR); /* Enable VLAN filtering */ regval = XLGMAC_SET_REG_BITS(regval, MAC_PFR_VTFE_POS, MAC_PFR_VTFE_LEN, 1); writel(regval, pdata->mac_regs + MAC_PFR); regval = readl(pdata->mac_regs + MAC_VLANTR); /* Enable VLAN Hash Table filtering */ regval = XLGMAC_SET_REG_BITS(regval, MAC_VLANTR_VTHM_POS, MAC_VLANTR_VTHM_LEN, 1); /* Disable VLAN tag inverse matching */ regval = XLGMAC_SET_REG_BITS(regval, MAC_VLANTR_VTIM_POS, MAC_VLANTR_VTIM_LEN, 0); /* Only filter on the lower 12-bits of the VLAN tag */ regval = XLGMAC_SET_REG_BITS(regval, MAC_VLANTR_ETV_POS, MAC_VLANTR_ETV_LEN, 1); /* In order for the VLAN Hash Table filtering to be effective, * the VLAN tag identifier in the VLAN Tag Register must not * be zero. Set the VLAN tag identifier to "1" to enable the * VLAN Hash Table filtering. This implies that a VLAN tag of * 1 will always pass filtering. */ regval = XLGMAC_SET_REG_BITS(regval, MAC_VLANTR_VL_POS, MAC_VLANTR_VL_LEN, 1); writel(regval, pdata->mac_regs + MAC_VLANTR); return 0; } static int xlgmac_disable_rx_vlan_filtering(struct xlgmac_pdata *pdata) { u32 regval; regval = readl(pdata->mac_regs + MAC_PFR); /* Disable VLAN filtering */ regval = XLGMAC_SET_REG_BITS(regval, MAC_PFR_VTFE_POS, MAC_PFR_VTFE_LEN, 0); writel(regval, pdata->mac_regs + MAC_PFR); return 0; } static u32 xlgmac_vid_crc32_le(__le16 vid_le) { unsigned char *data = (unsigned char *)&vid_le; unsigned char data_byte = 0; u32 crc = ~0; u32 temp = 0; int i, bits; bits = get_bitmask_order(VLAN_VID_MASK); for (i = 0; i < bits; i++) { if ((i % 8) == 0) data_byte = data[i / 8]; temp = ((crc & 1) ^ data_byte) & 1; crc >>= 1; data_byte >>= 1; if (temp) crc ^= CRC32_POLY_LE; } return crc; } static int xlgmac_update_vlan_hash_table(struct xlgmac_pdata *pdata) { u16 vlan_hash_table = 0; __le16 vid_le; u32 regval; u32 crc; u16 vid; /* Generate the VLAN Hash Table value */ for_each_set_bit(vid, pdata->active_vlans, VLAN_N_VID) { /* Get the CRC32 value of the VLAN ID */ vid_le = cpu_to_le16(vid); crc = bitrev32(~xlgmac_vid_crc32_le(vid_le)) >> 28; vlan_hash_table |= (1 << crc); } regval = readl(pdata->mac_regs + MAC_VLANHTR); /* Set the VLAN Hash Table filtering register */ regval = XLGMAC_SET_REG_BITS(regval, MAC_VLANHTR_VLHT_POS, MAC_VLANHTR_VLHT_LEN, vlan_hash_table); writel(regval, pdata->mac_regs + MAC_VLANHTR); return 0; } static int xlgmac_set_promiscuous_mode(struct xlgmac_pdata *pdata, unsigned int enable) { unsigned int val = enable ? 1 : 0; u32 regval; regval = XLGMAC_GET_REG_BITS(readl(pdata->mac_regs + MAC_PFR), MAC_PFR_PR_POS, MAC_PFR_PR_LEN); if (regval == val) return 0; netif_dbg(pdata, drv, pdata->netdev, "%s promiscuous mode\n", enable ? "entering" : "leaving"); regval = readl(pdata->mac_regs + MAC_PFR); regval = XLGMAC_SET_REG_BITS(regval, MAC_PFR_PR_POS, MAC_PFR_PR_LEN, val); writel(regval, pdata->mac_regs + MAC_PFR); /* Hardware will still perform VLAN filtering in promiscuous mode */ if (enable) { xlgmac_disable_rx_vlan_filtering(pdata); } else { if (pdata->netdev->features & NETIF_F_HW_VLAN_CTAG_FILTER) xlgmac_enable_rx_vlan_filtering(pdata); } return 0; } static int xlgmac_set_all_multicast_mode(struct xlgmac_pdata *pdata, unsigned int enable) { unsigned int val = enable ? 1 : 0; u32 regval; regval = XLGMAC_GET_REG_BITS(readl(pdata->mac_regs + MAC_PFR), MAC_PFR_PM_POS, MAC_PFR_PM_LEN); if (regval == val) return 0; netif_dbg(pdata, drv, pdata->netdev, "%s allmulti mode\n", enable ? "entering" : "leaving"); regval = readl(pdata->mac_regs + MAC_PFR); regval = XLGMAC_SET_REG_BITS(regval, MAC_PFR_PM_POS, MAC_PFR_PM_LEN, val); writel(regval, pdata->mac_regs + MAC_PFR); return 0; } static void xlgmac_set_mac_addn_addrs(struct xlgmac_pdata *pdata) { struct net_device *netdev = pdata->netdev; struct netdev_hw_addr *ha; unsigned int addn_macs; unsigned int mac_reg; mac_reg = MAC_MACA1HR; addn_macs = pdata->hw_feat.addn_mac; if (netdev_uc_count(netdev) > addn_macs) { xlgmac_set_promiscuous_mode(pdata, 1); } else { netdev_for_each_uc_addr(ha, netdev) { xlgmac_set_mac_reg(pdata, ha, &mac_reg); addn_macs--; } if (netdev_mc_count(netdev) > addn_macs) { xlgmac_set_all_multicast_mode(pdata, 1); } else { netdev_for_each_mc_addr(ha, netdev) { xlgmac_set_mac_reg(pdata, ha, &mac_reg); addn_macs--; } } } /* Clear remaining additional MAC address entries */ while (addn_macs--) xlgmac_set_mac_reg(pdata, NULL, &mac_reg); } static void xlgmac_set_mac_hash_table(struct xlgmac_pdata *pdata) { unsigned int hash_table_shift, hash_table_count; u32 hash_table[XLGMAC_MAC_HASH_TABLE_SIZE]; struct net_device *netdev = pdata->netdev; struct netdev_hw_addr *ha; unsigned int hash_reg; unsigned int i; u32 crc; hash_table_shift = 26 - (pdata->hw_feat.hash_table_size >> 7); hash_table_count = pdata->hw_feat.hash_table_size / 32; memset(hash_table, 0, sizeof(hash_table)); /* Build the MAC Hash Table register values */ netdev_for_each_uc_addr(ha, netdev) { crc = bitrev32(~crc32_le(~0, ha->addr, ETH_ALEN)); crc >>= hash_table_shift; hash_table[crc >> 5] |= (1 << (crc & 0x1f)); } netdev_for_each_mc_addr(ha, netdev) { crc = bitrev32(~crc32_le(~0, ha->addr, ETH_ALEN)); crc >>= hash_table_shift; hash_table[crc >> 5] |= (1 << (crc & 0x1f)); } /* Set the MAC Hash Table registers */ hash_reg = MAC_HTR0; for (i = 0; i < hash_table_count; i++) { writel(hash_table[i], pdata->mac_regs + hash_reg); hash_reg += MAC_HTR_INC; } } static int xlgmac_add_mac_addresses(struct xlgmac_pdata *pdata) { if (pdata->hw_feat.hash_table_size) xlgmac_set_mac_hash_table(pdata); else xlgmac_set_mac_addn_addrs(pdata); return 0; } static void xlgmac_config_mac_address(struct xlgmac_pdata *pdata) { u32 regval; xlgmac_set_mac_address(pdata, pdata->netdev->dev_addr); /* Filtering is done using perfect filtering and hash filtering */ if (pdata->hw_feat.hash_table_size) { regval = readl(pdata->mac_regs + MAC_PFR); regval = XLGMAC_SET_REG_BITS(regval, MAC_PFR_HPF_POS, MAC_PFR_HPF_LEN, 1); regval = XLGMAC_SET_REG_BITS(regval, MAC_PFR_HUC_POS, MAC_PFR_HUC_LEN, 1); regval = XLGMAC_SET_REG_BITS(regval, MAC_PFR_HMC_POS, MAC_PFR_HMC_LEN, 1); writel(regval, pdata->mac_regs + MAC_PFR); } } static void xlgmac_config_jumbo_enable(struct xlgmac_pdata *pdata) { unsigned int val; u32 regval; val = (pdata->netdev->mtu > XLGMAC_STD_PACKET_MTU) ? 1 : 0; regval = readl(pdata->mac_regs + MAC_RCR); regval = XLGMAC_SET_REG_BITS(regval, MAC_RCR_JE_POS, MAC_RCR_JE_LEN, val); writel(regval, pdata->mac_regs + MAC_RCR); } static void xlgmac_config_checksum_offload(struct xlgmac_pdata *pdata) { if (pdata->netdev->features & NETIF_F_RXCSUM) xlgmac_enable_rx_csum(pdata); else xlgmac_disable_rx_csum(pdata); } static void xlgmac_config_vlan_support(struct xlgmac_pdata *pdata) { u32 regval; regval = readl(pdata->mac_regs + MAC_VLANIR); /* Indicate that VLAN Tx CTAGs come from context descriptors */ regval = XLGMAC_SET_REG_BITS(regval, MAC_VLANIR_CSVL_POS, MAC_VLANIR_CSVL_LEN, 0); regval = XLGMAC_SET_REG_BITS(regval, MAC_VLANIR_VLTI_POS, MAC_VLANIR_VLTI_LEN, 1); writel(regval, pdata->mac_regs + MAC_VLANIR); /* Set the current VLAN Hash Table register value */ xlgmac_update_vlan_hash_table(pdata); if (pdata->netdev->features & NETIF_F_HW_VLAN_CTAG_FILTER) xlgmac_enable_rx_vlan_filtering(pdata); else xlgmac_disable_rx_vlan_filtering(pdata); if (pdata->netdev->features & NETIF_F_HW_VLAN_CTAG_RX) xlgmac_enable_rx_vlan_stripping(pdata); else xlgmac_disable_rx_vlan_stripping(pdata); } static int xlgmac_config_rx_mode(struct xlgmac_pdata *pdata) { struct net_device *netdev = pdata->netdev; unsigned int pr_mode, am_mode; pr_mode = ((netdev->flags & IFF_PROMISC) != 0); am_mode = ((netdev->flags & IFF_ALLMULTI) != 0); xlgmac_set_promiscuous_mode(pdata, pr_mode); xlgmac_set_all_multicast_mode(pdata, am_mode); xlgmac_add_mac_addresses(pdata); return 0; } static void xlgmac_prepare_tx_stop(struct xlgmac_pdata *pdata, struct xlgmac_channel *channel) { unsigned int tx_dsr, tx_pos, tx_qidx; unsigned long tx_timeout; unsigned int tx_status; /* Calculate the status register to read and the position within */ if (channel->queue_index < DMA_DSRX_FIRST_QUEUE) { tx_dsr = DMA_DSR0; tx_pos = (channel->queue_index * DMA_DSR_Q_LEN) + DMA_DSR0_TPS_START; } else { tx_qidx = channel->queue_index - DMA_DSRX_FIRST_QUEUE; tx_dsr = DMA_DSR1 + ((tx_qidx / DMA_DSRX_QPR) * DMA_DSRX_INC); tx_pos = ((tx_qidx % DMA_DSRX_QPR) * DMA_DSR_Q_LEN) + DMA_DSRX_TPS_START; } /* The Tx engine cannot be stopped if it is actively processing * descriptors. Wait for the Tx engine to enter the stopped or * suspended state. Don't wait forever though... */ tx_timeout = jiffies + (XLGMAC_DMA_STOP_TIMEOUT * HZ); while (time_before(jiffies, tx_timeout)) { tx_status = readl(pdata->mac_regs + tx_dsr); tx_status = XLGMAC_GET_REG_BITS(tx_status, tx_pos, DMA_DSR_TPS_LEN); if ((tx_status == DMA_TPS_STOPPED) || (tx_status == DMA_TPS_SUSPENDED)) break; usleep_range(500, 1000); } if (!time_before(jiffies, tx_timeout)) netdev_info(pdata->netdev, "timed out waiting for Tx DMA channel %u to stop\n", channel->queue_index); } static void xlgmac_enable_tx(struct xlgmac_pdata *pdata) { struct xlgmac_channel *channel; unsigned int i; u32 regval; /* Enable each Tx DMA channel */ channel = pdata->channel_head; for (i = 0; i < pdata->channel_count; i++, channel++) { if (!channel->tx_ring) break; regval = readl(XLGMAC_DMA_REG(channel, DMA_CH_TCR)); regval = XLGMAC_SET_REG_BITS(regval, DMA_CH_TCR_ST_POS, DMA_CH_TCR_ST_LEN, 1); writel(regval, XLGMAC_DMA_REG(channel, DMA_CH_TCR)); } /* Enable each Tx queue */ for (i = 0; i < pdata->tx_q_count; i++) { regval = readl(XLGMAC_MTL_REG(pdata, i, MTL_Q_TQOMR)); regval = XLGMAC_SET_REG_BITS(regval, MTL_Q_TQOMR_TXQEN_POS, MTL_Q_TQOMR_TXQEN_LEN, MTL_Q_ENABLED); writel(regval, XLGMAC_MTL_REG(pdata, i, MTL_Q_TQOMR)); } /* Enable MAC Tx */ regval = readl(pdata->mac_regs + MAC_TCR); regval = XLGMAC_SET_REG_BITS(regval, MAC_TCR_TE_POS, MAC_TCR_TE_LEN, 1); writel(regval, pdata->mac_regs + MAC_TCR); } static void xlgmac_disable_tx(struct xlgmac_pdata *pdata) { struct xlgmac_channel *channel; unsigned int i; u32 regval; /* Prepare for Tx DMA channel stop */ channel = pdata->channel_head; for (i = 0; i < pdata->channel_count; i++, channel++) { if (!channel->tx_ring) break; xlgmac_prepare_tx_stop(pdata, channel); } /* Disable MAC Tx */ regval = readl(pdata->mac_regs + MAC_TCR); regval = XLGMAC_SET_REG_BITS(regval, MAC_TCR_TE_POS, MAC_TCR_TE_LEN, 0); writel(regval, pdata->mac_regs + MAC_TCR); /* Disable each Tx queue */ for (i = 0; i < pdata->tx_q_count; i++) { regval = readl(XLGMAC_MTL_REG(pdata, i, MTL_Q_TQOMR)); regval = XLGMAC_SET_REG_BITS(regval, MTL_Q_TQOMR_TXQEN_POS, MTL_Q_TQOMR_TXQEN_LEN, 0); writel(regval, XLGMAC_MTL_REG(pdata, i, MTL_Q_TQOMR)); } /* Disable each Tx DMA channel */ channel = pdata->channel_head; for (i = 0; i < pdata->channel_count; i++, channel++) { if (!channel->tx_ring) break; regval = readl(XLGMAC_DMA_REG(channel, DMA_CH_TCR)); regval = XLGMAC_SET_REG_BITS(regval, DMA_CH_TCR_ST_POS, DMA_CH_TCR_ST_LEN, 0); writel(regval, XLGMAC_DMA_REG(channel, DMA_CH_TCR)); } } static void xlgmac_prepare_rx_stop(struct xlgmac_pdata *pdata, unsigned int queue) { unsigned int rx_status, prxq, rxqsts; unsigned long rx_timeout; /* The Rx engine cannot be stopped if it is actively processing * packets. Wait for the Rx queue to empty the Rx fifo. Don't * wait forever though... */ rx_timeout = jiffies + (XLGMAC_DMA_STOP_TIMEOUT * HZ); while (time_before(jiffies, rx_timeout)) { rx_status = readl(XLGMAC_MTL_REG(pdata, queue, MTL_Q_RQDR)); prxq = XLGMAC_GET_REG_BITS(rx_status, MTL_Q_RQDR_PRXQ_POS, MTL_Q_RQDR_PRXQ_LEN); rxqsts = XLGMAC_GET_REG_BITS(rx_status, MTL_Q_RQDR_RXQSTS_POS, MTL_Q_RQDR_RXQSTS_LEN); if ((prxq == 0) && (rxqsts == 0)) break; usleep_range(500, 1000); } if (!time_before(jiffies, rx_timeout)) netdev_info(pdata->netdev, "timed out waiting for Rx queue %u to empty\n", queue); } static void xlgmac_enable_rx(struct xlgmac_pdata *pdata) { struct xlgmac_channel *channel; unsigned int regval, i; /* Enable each Rx DMA channel */ channel = pdata->channel_head; for (i = 0; i < pdata->channel_count; i++, channel++) { if (!channel->rx_ring) break; regval = readl(XLGMAC_DMA_REG(channel, DMA_CH_RCR)); regval = XLGMAC_SET_REG_BITS(regval, DMA_CH_RCR_SR_POS, DMA_CH_RCR_SR_LEN, 1); writel(regval, XLGMAC_DMA_REG(channel, DMA_CH_RCR)); } /* Enable each Rx queue */ regval = 0; for (i = 0; i < pdata->rx_q_count; i++) regval |= (0x02 << (i << 1)); writel(regval, pdata->mac_regs + MAC_RQC0R); /* Enable MAC Rx */ regval = readl(pdata->mac_regs + MAC_RCR); regval = XLGMAC_SET_REG_BITS(regval, MAC_RCR_DCRCC_POS, MAC_RCR_DCRCC_LEN, 1); regval = XLGMAC_SET_REG_BITS(regval, MAC_RCR_CST_POS, MAC_RCR_CST_LEN, 1); regval = XLGMAC_SET_REG_BITS(regval, MAC_RCR_ACS_POS, MAC_RCR_ACS_LEN, 1); regval = XLGMAC_SET_REG_BITS(regval, MAC_RCR_RE_POS, MAC_RCR_RE_LEN, 1); writel(regval, pdata->mac_regs + MAC_RCR); } static void xlgmac_disable_rx(struct xlgmac_pdata *pdata) { struct xlgmac_channel *channel; unsigned int i; u32 regval; /* Disable MAC Rx */ regval = readl(pdata->mac_regs + MAC_RCR); regval = XLGMAC_SET_REG_BITS(regval, MAC_RCR_DCRCC_POS, MAC_RCR_DCRCC_LEN, 0); regval = XLGMAC_SET_REG_BITS(regval, MAC_RCR_CST_POS, MAC_RCR_CST_LEN, 0); regval = XLGMAC_SET_REG_BITS(regval, MAC_RCR_ACS_POS, MAC_RCR_ACS_LEN, 0); regval = XLGMAC_SET_REG_BITS(regval, MAC_RCR_RE_POS, MAC_RCR_RE_LEN, 0); writel(regval, pdata->mac_regs + MAC_RCR); /* Prepare for Rx DMA channel stop */ for (i = 0; i < pdata->rx_q_count; i++) xlgmac_prepare_rx_stop(pdata, i); /* Disable each Rx queue */ writel(0, pdata->mac_regs + MAC_RQC0R); /* Disable each Rx DMA channel */ channel = pdata->channel_head; for (i = 0; i < pdata->channel_count; i++, channel++) { if (!channel->rx_ring) break; regval = readl(XLGMAC_DMA_REG(channel, DMA_CH_RCR)); regval = XLGMAC_SET_REG_BITS(regval, DMA_CH_RCR_SR_POS, DMA_CH_RCR_SR_LEN, 0); writel(regval, XLGMAC_DMA_REG(channel, DMA_CH_RCR)); } } static void xlgmac_tx_start_xmit(struct xlgmac_channel *channel, struct xlgmac_ring *ring) { struct xlgmac_pdata *pdata = channel->pdata; struct xlgmac_desc_data *desc_data; /* Make sure everything is written before the register write */ wmb(); /* Issue a poll command to Tx DMA by writing address * of next immediate free descriptor */ desc_data = XLGMAC_GET_DESC_DATA(ring, ring->cur); writel(lower_32_bits(desc_data->dma_desc_addr), XLGMAC_DMA_REG(channel, DMA_CH_TDTR_LO)); /* Start the Tx timer */ if (pdata->tx_usecs && !channel->tx_timer_active) { channel->tx_timer_active = 1; mod_timer(&channel->tx_timer, jiffies + usecs_to_jiffies(pdata->tx_usecs)); } ring->tx.xmit_more = 0; } static void xlgmac_dev_xmit(struct xlgmac_channel *channel) { struct xlgmac_pdata *pdata = channel->pdata; struct xlgmac_ring *ring = channel->tx_ring; unsigned int tso_context, vlan_context; struct xlgmac_desc_data *desc_data; struct xlgmac_dma_desc *dma_desc; struct xlgmac_pkt_info *pkt_info; unsigned int csum, tso, vlan; int start_index = ring->cur; int cur_index = ring->cur; unsigned int tx_set_ic; int i; pkt_info = &ring->pkt_info; csum = XLGMAC_GET_REG_BITS(pkt_info->attributes, TX_PACKET_ATTRIBUTES_CSUM_ENABLE_POS, TX_PACKET_ATTRIBUTES_CSUM_ENABLE_LEN); tso = XLGMAC_GET_REG_BITS(pkt_info->attributes, TX_PACKET_ATTRIBUTES_TSO_ENABLE_POS, TX_PACKET_ATTRIBUTES_TSO_ENABLE_LEN); vlan = XLGMAC_GET_REG_BITS(pkt_info->attributes, TX_PACKET_ATTRIBUTES_VLAN_CTAG_POS, TX_PACKET_ATTRIBUTES_VLAN_CTAG_LEN); if (tso && (pkt_info->mss != ring->tx.cur_mss)) tso_context = 1; else tso_context = 0; if (vlan && (pkt_info->vlan_ctag != ring->tx.cur_vlan_ctag)) vlan_context = 1; else vlan_context = 0; /* Determine if an interrupt should be generated for this Tx: * Interrupt: * - Tx frame count exceeds the frame count setting * - Addition of Tx frame count to the frame count since the * last interrupt was set exceeds the frame count setting * No interrupt: * - No frame count setting specified (ethtool -C ethX tx-frames 0) * - Addition of Tx frame count to the frame count since the * last interrupt was set does not exceed the frame count setting */ ring->coalesce_count += pkt_info->tx_packets; if (!pdata->tx_frames) tx_set_ic = 0; else if (pkt_info->tx_packets > pdata->tx_frames) tx_set_ic = 1; else if ((ring->coalesce_count % pdata->tx_frames) < pkt_info->tx_packets) tx_set_ic = 1; else tx_set_ic = 0; desc_data = XLGMAC_GET_DESC_DATA(ring, cur_index); dma_desc = desc_data->dma_desc; /* Create a context descriptor if this is a TSO pkt_info */ if (tso_context || vlan_context) { if (tso_context) { netif_dbg(pdata, tx_queued, pdata->netdev, "TSO context descriptor, mss=%u\n", pkt_info->mss); /* Set the MSS size */ dma_desc->desc2 = XLGMAC_SET_REG_BITS_LE( dma_desc->desc2, TX_CONTEXT_DESC2_MSS_POS, TX_CONTEXT_DESC2_MSS_LEN, pkt_info->mss); /* Mark it as a CONTEXT descriptor */ dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( dma_desc->desc3, TX_CONTEXT_DESC3_CTXT_POS, TX_CONTEXT_DESC3_CTXT_LEN, 1); /* Indicate this descriptor contains the MSS */ dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( dma_desc->desc3, TX_CONTEXT_DESC3_TCMSSV_POS, TX_CONTEXT_DESC3_TCMSSV_LEN, 1); ring->tx.cur_mss = pkt_info->mss; } if (vlan_context) { netif_dbg(pdata, tx_queued, pdata->netdev, "VLAN context descriptor, ctag=%u\n", pkt_info->vlan_ctag); /* Mark it as a CONTEXT descriptor */ dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( dma_desc->desc3, TX_CONTEXT_DESC3_CTXT_POS, TX_CONTEXT_DESC3_CTXT_LEN, 1); /* Set the VLAN tag */ dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( dma_desc->desc3, TX_CONTEXT_DESC3_VT_POS, TX_CONTEXT_DESC3_VT_LEN, pkt_info->vlan_ctag); /* Indicate this descriptor contains the VLAN tag */ dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( dma_desc->desc3, TX_CONTEXT_DESC3_VLTV_POS, TX_CONTEXT_DESC3_VLTV_LEN, 1); ring->tx.cur_vlan_ctag = pkt_info->vlan_ctag; } cur_index++; desc_data = XLGMAC_GET_DESC_DATA(ring, cur_index); dma_desc = desc_data->dma_desc; } /* Update buffer address (for TSO this is the header) */ dma_desc->desc0 = cpu_to_le32(lower_32_bits(desc_data->skb_dma)); dma_desc->desc1 = cpu_to_le32(upper_32_bits(desc_data->skb_dma)); /* Update the buffer length */ dma_desc->desc2 = XLGMAC_SET_REG_BITS_LE( dma_desc->desc2, TX_NORMAL_DESC2_HL_B1L_POS, TX_NORMAL_DESC2_HL_B1L_LEN, desc_data->skb_dma_len); /* VLAN tag insertion check */ if (vlan) { dma_desc->desc2 = XLGMAC_SET_REG_BITS_LE( dma_desc->desc2, TX_NORMAL_DESC2_VTIR_POS, TX_NORMAL_DESC2_VTIR_LEN, TX_NORMAL_DESC2_VLAN_INSERT); pdata->stats.tx_vlan_packets++; } /* Timestamp enablement check */ if (XLGMAC_GET_REG_BITS(pkt_info->attributes, TX_PACKET_ATTRIBUTES_PTP_POS, TX_PACKET_ATTRIBUTES_PTP_LEN)) dma_desc->desc2 = XLGMAC_SET_REG_BITS_LE( dma_desc->desc2, TX_NORMAL_DESC2_TTSE_POS, TX_NORMAL_DESC2_TTSE_LEN, 1); /* Mark it as First Descriptor */ dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( dma_desc->desc3, TX_NORMAL_DESC3_FD_POS, TX_NORMAL_DESC3_FD_LEN, 1); /* Mark it as a NORMAL descriptor */ dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( dma_desc->desc3, TX_NORMAL_DESC3_CTXT_POS, TX_NORMAL_DESC3_CTXT_LEN, 0); /* Set OWN bit if not the first descriptor */ if (cur_index != start_index) dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( dma_desc->desc3, TX_NORMAL_DESC3_OWN_POS, TX_NORMAL_DESC3_OWN_LEN, 1); if (tso) { /* Enable TSO */ dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( dma_desc->desc3, TX_NORMAL_DESC3_TSE_POS, TX_NORMAL_DESC3_TSE_LEN, 1); dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( dma_desc->desc3, TX_NORMAL_DESC3_TCPPL_POS, TX_NORMAL_DESC3_TCPPL_LEN, pkt_info->tcp_payload_len); dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( dma_desc->desc3, TX_NORMAL_DESC3_TCPHDRLEN_POS, TX_NORMAL_DESC3_TCPHDRLEN_LEN, pkt_info->tcp_header_len / 4); pdata->stats.tx_tso_packets++; } else { /* Enable CRC and Pad Insertion */ dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( dma_desc->desc3, TX_NORMAL_DESC3_CPC_POS, TX_NORMAL_DESC3_CPC_LEN, 0); /* Enable HW CSUM */ if (csum) dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( dma_desc->desc3, TX_NORMAL_DESC3_CIC_POS, TX_NORMAL_DESC3_CIC_LEN, 0x3); /* Set the total length to be transmitted */ dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( dma_desc->desc3, TX_NORMAL_DESC3_FL_POS, TX_NORMAL_DESC3_FL_LEN, pkt_info->length); } for (i = cur_index - start_index + 1; i < pkt_info->desc_count; i++) { cur_index++; desc_data = XLGMAC_GET_DESC_DATA(ring, cur_index); dma_desc = desc_data->dma_desc; /* Update buffer address */ dma_desc->desc0 = cpu_to_le32(lower_32_bits(desc_data->skb_dma)); dma_desc->desc1 = cpu_to_le32(upper_32_bits(desc_data->skb_dma)); /* Update the buffer length */ dma_desc->desc2 = XLGMAC_SET_REG_BITS_LE( dma_desc->desc2, TX_NORMAL_DESC2_HL_B1L_POS, TX_NORMAL_DESC2_HL_B1L_LEN, desc_data->skb_dma_len); /* Set OWN bit */ dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( dma_desc->desc3, TX_NORMAL_DESC3_OWN_POS, TX_NORMAL_DESC3_OWN_LEN, 1); /* Mark it as NORMAL descriptor */ dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( dma_desc->desc3, TX_NORMAL_DESC3_CTXT_POS, TX_NORMAL_DESC3_CTXT_LEN, 0); /* Enable HW CSUM */ if (csum) dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( dma_desc->desc3, TX_NORMAL_DESC3_CIC_POS, TX_NORMAL_DESC3_CIC_LEN, 0x3); } /* Set LAST bit for the last descriptor */ dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( dma_desc->desc3, TX_NORMAL_DESC3_LD_POS, TX_NORMAL_DESC3_LD_LEN, 1); /* Set IC bit based on Tx coalescing settings */ if (tx_set_ic) dma_desc->desc2 = XLGMAC_SET_REG_BITS_LE( dma_desc->desc2, TX_NORMAL_DESC2_IC_POS, TX_NORMAL_DESC2_IC_LEN, 1); /* Save the Tx info to report back during cleanup */ desc_data->tx.packets = pkt_info->tx_packets; desc_data->tx.bytes = pkt_info->tx_bytes; /* In case the Tx DMA engine is running, make sure everything * is written to the descriptor(s) before setting the OWN bit * for the first descriptor */ dma_wmb(); /* Set OWN bit for the first descriptor */ desc_data = XLGMAC_GET_DESC_DATA(ring, start_index); dma_desc = desc_data->dma_desc; dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( dma_desc->desc3, TX_NORMAL_DESC3_OWN_POS, TX_NORMAL_DESC3_OWN_LEN, 1); if (netif_msg_tx_queued(pdata)) xlgmac_dump_tx_desc(pdata, ring, start_index, pkt_info->desc_count, 1); /* Make sure ownership is written to the descriptor */ smp_wmb(); ring->cur = cur_index + 1; if (!netdev_xmit_more() || netif_xmit_stopped(netdev_get_tx_queue(pdata->netdev, channel->queue_index))) xlgmac_tx_start_xmit(channel, ring); else ring->tx.xmit_more = 1; XLGMAC_PR("%s: descriptors %u to %u written\n", channel->name, start_index & (ring->dma_desc_count - 1), (ring->cur - 1) & (ring->dma_desc_count - 1)); } static void xlgmac_get_rx_tstamp(struct xlgmac_pkt_info *pkt_info, struct xlgmac_dma_desc *dma_desc) { u32 tsa, tsd; u64 nsec; tsa = XLGMAC_GET_REG_BITS_LE(dma_desc->desc3, RX_CONTEXT_DESC3_TSA_POS, RX_CONTEXT_DESC3_TSA_LEN); tsd = XLGMAC_GET_REG_BITS_LE(dma_desc->desc3, RX_CONTEXT_DESC3_TSD_POS, RX_CONTEXT_DESC3_TSD_LEN); if (tsa && !tsd) { nsec = le32_to_cpu(dma_desc->desc1); nsec <<= 32; nsec |= le32_to_cpu(dma_desc->desc0); if (nsec != 0xffffffffffffffffULL) { pkt_info->rx_tstamp = nsec; pkt_info->attributes = XLGMAC_SET_REG_BITS( pkt_info->attributes, RX_PACKET_ATTRIBUTES_RX_TSTAMP_POS, RX_PACKET_ATTRIBUTES_RX_TSTAMP_LEN, 1); } } } static void xlgmac_tx_desc_reset(struct xlgmac_desc_data *desc_data) { struct xlgmac_dma_desc *dma_desc = desc_data->dma_desc; /* Reset the Tx descriptor * Set buffer 1 (lo) address to zero * Set buffer 1 (hi) address to zero * Reset all other control bits (IC, TTSE, B2L & B1L) * Reset all other control bits (OWN, CTXT, FD, LD, CPC, CIC, etc) */ dma_desc->desc0 = 0; dma_desc->desc1 = 0; dma_desc->desc2 = 0; dma_desc->desc3 = 0; /* Make sure ownership is written to the descriptor */ dma_wmb(); } static void xlgmac_tx_desc_init(struct xlgmac_channel *channel) { struct xlgmac_ring *ring = channel->tx_ring; struct xlgmac_desc_data *desc_data; int start_index = ring->cur; int i; /* Initialze all descriptors */ for (i = 0; i < ring->dma_desc_count; i++) { desc_data = XLGMAC_GET_DESC_DATA(ring, i); /* Initialize Tx descriptor */ xlgmac_tx_desc_reset(desc_data); } /* Update the total number of Tx descriptors */ writel(ring->dma_desc_count - 1, XLGMAC_DMA_REG(channel, DMA_CH_TDRLR)); /* Update the starting address of descriptor ring */ desc_data = XLGMAC_GET_DESC_DATA(ring, start_index); writel(upper_32_bits(desc_data->dma_desc_addr), XLGMAC_DMA_REG(channel, DMA_CH_TDLR_HI)); writel(lower_32_bits(desc_data->dma_desc_addr), XLGMAC_DMA_REG(channel, DMA_CH_TDLR_LO)); } static void xlgmac_rx_desc_reset(struct xlgmac_pdata *pdata, struct xlgmac_desc_data *desc_data, unsigned int index) { struct xlgmac_dma_desc *dma_desc = desc_data->dma_desc; unsigned int rx_frames = pdata->rx_frames; unsigned int rx_usecs = pdata->rx_usecs; dma_addr_t hdr_dma, buf_dma; unsigned int inte; if (!rx_usecs && !rx_frames) { /* No coalescing, interrupt for every descriptor */ inte = 1; } else { /* Set interrupt based on Rx frame coalescing setting */ if (rx_frames && !((index + 1) % rx_frames)) inte = 1; else inte = 0; } /* Reset the Rx descriptor * Set buffer 1 (lo) address to header dma address (lo) * Set buffer 1 (hi) address to header dma address (hi) * Set buffer 2 (lo) address to buffer dma address (lo) * Set buffer 2 (hi) address to buffer dma address (hi) and * set control bits OWN and INTE */ hdr_dma = desc_data->rx.hdr.dma_base + desc_data->rx.hdr.dma_off; buf_dma = desc_data->rx.buf.dma_base + desc_data->rx.buf.dma_off; dma_desc->desc0 = cpu_to_le32(lower_32_bits(hdr_dma)); dma_desc->desc1 = cpu_to_le32(upper_32_bits(hdr_dma)); dma_desc->desc2 = cpu_to_le32(lower_32_bits(buf_dma)); dma_desc->desc3 = cpu_to_le32(upper_32_bits(buf_dma)); dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( dma_desc->desc3, RX_NORMAL_DESC3_INTE_POS, RX_NORMAL_DESC3_INTE_LEN, inte); /* Since the Rx DMA engine is likely running, make sure everything * is written to the descriptor(s) before setting the OWN bit * for the descriptor */ dma_wmb(); dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( dma_desc->desc3, RX_NORMAL_DESC3_OWN_POS, RX_NORMAL_DESC3_OWN_LEN, 1); /* Make sure ownership is written to the descriptor */ dma_wmb(); } static void xlgmac_rx_desc_init(struct xlgmac_channel *channel) { struct xlgmac_pdata *pdata = channel->pdata; struct xlgmac_ring *ring = channel->rx_ring; unsigned int start_index = ring->cur; struct xlgmac_desc_data *desc_data; unsigned int i; /* Initialize all descriptors */ for (i = 0; i < ring->dma_desc_count; i++) { desc_data = XLGMAC_GET_DESC_DATA(ring, i); /* Initialize Rx descriptor */ xlgmac_rx_desc_reset(pdata, desc_data, i); } /* Update the total number of Rx descriptors */ writel(ring->dma_desc_count - 1, XLGMAC_DMA_REG(channel, DMA_CH_RDRLR)); /* Update the starting address of descriptor ring */ desc_data = XLGMAC_GET_DESC_DATA(ring, start_index); writel(upper_32_bits(desc_data->dma_desc_addr), XLGMAC_DMA_REG(channel, DMA_CH_RDLR_HI)); writel(lower_32_bits(desc_data->dma_desc_addr), XLGMAC_DMA_REG(channel, DMA_CH_RDLR_LO)); /* Update the Rx Descriptor Tail Pointer */ desc_data = XLGMAC_GET_DESC_DATA(ring, start_index + ring->dma_desc_count - 1); writel(lower_32_bits(desc_data->dma_desc_addr), XLGMAC_DMA_REG(channel, DMA_CH_RDTR_LO)); } static int xlgmac_is_context_desc(struct xlgmac_dma_desc *dma_desc) { /* Rx and Tx share CTXT bit, so check TDES3.CTXT bit */ return XLGMAC_GET_REG_BITS_LE(dma_desc->desc3, TX_NORMAL_DESC3_CTXT_POS, TX_NORMAL_DESC3_CTXT_LEN); } static int xlgmac_is_last_desc(struct xlgmac_dma_desc *dma_desc) { /* Rx and Tx share LD bit, so check TDES3.LD bit */ return XLGMAC_GET_REG_BITS_LE(dma_desc->desc3, TX_NORMAL_DESC3_LD_POS, TX_NORMAL_DESC3_LD_LEN); } static int xlgmac_disable_tx_flow_control(struct xlgmac_pdata *pdata) { unsigned int max_q_count, q_count; unsigned int reg, regval; unsigned int i; /* Clear MTL flow control */ for (i = 0; i < pdata->rx_q_count; i++) { regval = readl(XLGMAC_MTL_REG(pdata, i, MTL_Q_RQOMR)); regval = XLGMAC_SET_REG_BITS(regval, MTL_Q_RQOMR_EHFC_POS, MTL_Q_RQOMR_EHFC_LEN, 0); writel(regval, XLGMAC_MTL_REG(pdata, i, MTL_Q_RQOMR)); } /* Clear MAC flow control */ max_q_count = XLGMAC_MAX_FLOW_CONTROL_QUEUES; q_count = min_t(unsigned int, pdata->tx_q_count, max_q_count); reg = MAC_Q0TFCR; for (i = 0; i < q_count; i++) { regval = readl(pdata->mac_regs + reg); regval = XLGMAC_SET_REG_BITS(regval, MAC_Q0TFCR_TFE_POS, MAC_Q0TFCR_TFE_LEN, 0); writel(regval, pdata->mac_regs + reg); reg += MAC_QTFCR_INC; } return 0; } static int xlgmac_enable_tx_flow_control(struct xlgmac_pdata *pdata) { unsigned int max_q_count, q_count; unsigned int reg, regval; unsigned int i; /* Set MTL flow control */ for (i = 0; i < pdata->rx_q_count; i++) { regval = readl(XLGMAC_MTL_REG(pdata, i, MTL_Q_RQOMR)); regval = XLGMAC_SET_REG_BITS(regval, MTL_Q_RQOMR_EHFC_POS, MTL_Q_RQOMR_EHFC_LEN, 1); writel(regval, XLGMAC_MTL_REG(pdata, i, MTL_Q_RQOMR)); } /* Set MAC flow control */ max_q_count = XLGMAC_MAX_FLOW_CONTROL_QUEUES; q_count = min_t(unsigned int, pdata->tx_q_count, max_q_count); reg = MAC_Q0TFCR; for (i = 0; i < q_count; i++) { regval = readl(pdata->mac_regs + reg); /* Enable transmit flow control */ regval = XLGMAC_SET_REG_BITS(regval, MAC_Q0TFCR_TFE_POS, MAC_Q0TFCR_TFE_LEN, 1); /* Set pause time */ regval = XLGMAC_SET_REG_BITS(regval, MAC_Q0TFCR_PT_POS, MAC_Q0TFCR_PT_LEN, 0xffff); writel(regval, pdata->mac_regs + reg); reg += MAC_QTFCR_INC; } return 0; } static int xlgmac_disable_rx_flow_control(struct xlgmac_pdata *pdata) { u32 regval; regval = readl(pdata->mac_regs + MAC_RFCR); regval = XLGMAC_SET_REG_BITS(regval, MAC_RFCR_RFE_POS, MAC_RFCR_RFE_LEN, 0); writel(regval, pdata->mac_regs + MAC_RFCR); return 0; } static int xlgmac_enable_rx_flow_control(struct xlgmac_pdata *pdata) { u32 regval; regval = readl(pdata->mac_regs + MAC_RFCR); regval = XLGMAC_SET_REG_BITS(regval, MAC_RFCR_RFE_POS, MAC_RFCR_RFE_LEN, 1); writel(regval, pdata->mac_regs + MAC_RFCR); return 0; } static int xlgmac_config_tx_flow_control(struct xlgmac_pdata *pdata) { if (pdata->tx_pause) xlgmac_enable_tx_flow_control(pdata); else xlgmac_disable_tx_flow_control(pdata); return 0; } static int xlgmac_config_rx_flow_control(struct xlgmac_pdata *pdata) { if (pdata->rx_pause) xlgmac_enable_rx_flow_control(pdata); else xlgmac_disable_rx_flow_control(pdata); return 0; } static int xlgmac_config_rx_coalesce(struct xlgmac_pdata *pdata) { struct xlgmac_channel *channel; unsigned int i; u32 regval; channel = pdata->channel_head; for (i = 0; i < pdata->channel_count; i++, channel++) { if (!channel->rx_ring) break; regval = readl(XLGMAC_DMA_REG(channel, DMA_CH_RIWT)); regval = XLGMAC_SET_REG_BITS(regval, DMA_CH_RIWT_RWT_POS, DMA_CH_RIWT_RWT_LEN, pdata->rx_riwt); writel(regval, XLGMAC_DMA_REG(channel, DMA_CH_RIWT)); } return 0; } static void xlgmac_config_flow_control(struct xlgmac_pdata *pdata) { xlgmac_config_tx_flow_control(pdata); xlgmac_config_rx_flow_control(pdata); } static void xlgmac_config_rx_fep_enable(struct xlgmac_pdata *pdata) { unsigned int i; u32 regval; for (i = 0; i < pdata->rx_q_count; i++) { regval = readl(XLGMAC_MTL_REG(pdata, i, MTL_Q_RQOMR)); regval = XLGMAC_SET_REG_BITS(regval, MTL_Q_RQOMR_FEP_POS, MTL_Q_RQOMR_FEP_LEN, 1); writel(regval, XLGMAC_MTL_REG(pdata, i, MTL_Q_RQOMR)); } } static void xlgmac_config_rx_fup_enable(struct xlgmac_pdata *pdata) { unsigned int i; u32 regval; for (i = 0; i < pdata->rx_q_count; i++) { regval = readl(XLGMAC_MTL_REG(pdata, i, MTL_Q_RQOMR)); regval = XLGMAC_SET_REG_BITS(regval, MTL_Q_RQOMR_FUP_POS, MTL_Q_RQOMR_FUP_LEN, 1); writel(regval, XLGMAC_MTL_REG(pdata, i, MTL_Q_RQOMR)); } } static int xlgmac_config_tx_coalesce(struct xlgmac_pdata *pdata) { return 0; } static void xlgmac_config_rx_buffer_size(struct xlgmac_pdata *pdata) { struct xlgmac_channel *channel; unsigned int i; u32 regval; channel = pdata->channel_head; for (i = 0; i < pdata->channel_count; i++, channel++) { if (!channel->rx_ring) break; regval = readl(XLGMAC_DMA_REG(channel, DMA_CH_RCR)); regval = XLGMAC_SET_REG_BITS(regval, DMA_CH_RCR_RBSZ_POS, DMA_CH_RCR_RBSZ_LEN, pdata->rx_buf_size); writel(regval, XLGMAC_DMA_REG(channel, DMA_CH_RCR)); } } static void xlgmac_config_tso_mode(struct xlgmac_pdata *pdata) { struct xlgmac_channel *channel; unsigned int i; u32 regval; channel = pdata->channel_head; for (i = 0; i < pdata->channel_count; i++, channel++) { if (!channel->tx_ring) break; if (pdata->hw_feat.tso) { regval = readl(XLGMAC_DMA_REG(channel, DMA_CH_TCR)); regval = XLGMAC_SET_REG_BITS(regval, DMA_CH_TCR_TSE_POS, DMA_CH_TCR_TSE_LEN, 1); writel(regval, XLGMAC_DMA_REG(channel, DMA_CH_TCR)); } } } static void xlgmac_config_sph_mode(struct xlgmac_pdata *pdata) { struct xlgmac_channel *channel; unsigned int i; u32 regval; channel = pdata->channel_head; for (i = 0; i < pdata->channel_count; i++, channel++) { if (!channel->rx_ring) break; regval = readl(XLGMAC_DMA_REG(channel, DMA_CH_CR)); regval = XLGMAC_SET_REG_BITS(regval, DMA_CH_CR_SPH_POS, DMA_CH_CR_SPH_LEN, 1); writel(regval, XLGMAC_DMA_REG(channel, DMA_CH_CR)); } regval = readl(pdata->mac_regs + MAC_RCR); regval = XLGMAC_SET_REG_BITS(regval, MAC_RCR_HDSMS_POS, MAC_RCR_HDSMS_LEN, XLGMAC_SPH_HDSMS_SIZE); writel(regval, pdata->mac_regs + MAC_RCR); } static unsigned int xlgmac_usec_to_riwt(struct xlgmac_pdata *pdata, unsigned int usec) { unsigned long rate; unsigned int ret; rate = pdata->sysclk_rate; /* Convert the input usec value to the watchdog timer value. Each * watchdog timer value is equivalent to 256 clock cycles. * Calculate the required value as: * ( usec * ( system_clock_mhz / 10^6 ) / 256 */ ret = (usec * (rate / 1000000)) / 256; return ret; } static unsigned int xlgmac_riwt_to_usec(struct xlgmac_pdata *pdata, unsigned int riwt) { unsigned long rate; unsigned int ret; rate = pdata->sysclk_rate; /* Convert the input watchdog timer value to the usec value. Each * watchdog timer value is equivalent to 256 clock cycles. * Calculate the required value as: * ( riwt * 256 ) / ( system_clock_mhz / 10^6 ) */ ret = (riwt * 256) / (rate / 1000000); return ret; } static int xlgmac_config_rx_threshold(struct xlgmac_pdata *pdata, unsigned int val) { unsigned int i; u32 regval; for (i = 0; i < pdata->rx_q_count; i++) { regval = readl(XLGMAC_MTL_REG(pdata, i, MTL_Q_RQOMR)); regval = XLGMAC_SET_REG_BITS(regval, MTL_Q_RQOMR_RTC_POS, MTL_Q_RQOMR_RTC_LEN, val); writel(regval, XLGMAC_MTL_REG(pdata, i, MTL_Q_RQOMR)); } return 0; } static void xlgmac_config_mtl_mode(struct xlgmac_pdata *pdata) { unsigned int i; u32 regval; /* Set Tx to weighted round robin scheduling algorithm */ regval = readl(pdata->mac_regs + MTL_OMR); regval = XLGMAC_SET_REG_BITS(regval, MTL_OMR_ETSALG_POS, MTL_OMR_ETSALG_LEN, MTL_ETSALG_WRR); writel(regval, pdata->mac_regs + MTL_OMR); /* Set Tx traffic classes to use WRR algorithm with equal weights */ for (i = 0; i < pdata->hw_feat.tc_cnt; i++) { regval = readl(XLGMAC_MTL_REG(pdata, i, MTL_TC_ETSCR)); regval = XLGMAC_SET_REG_BITS(regval, MTL_TC_ETSCR_TSA_POS, MTL_TC_ETSCR_TSA_LEN, MTL_TSA_ETS); writel(regval, XLGMAC_MTL_REG(pdata, i, MTL_TC_ETSCR)); regval = readl(XLGMAC_MTL_REG(pdata, i, MTL_TC_QWR)); regval = XLGMAC_SET_REG_BITS(regval, MTL_TC_QWR_QW_POS, MTL_TC_QWR_QW_LEN, 1); writel(regval, XLGMAC_MTL_REG(pdata, i, MTL_TC_QWR)); } /* Set Rx to strict priority algorithm */ regval = readl(pdata->mac_regs + MTL_OMR); regval = XLGMAC_SET_REG_BITS(regval, MTL_OMR_RAA_POS, MTL_OMR_RAA_LEN, MTL_RAA_SP); writel(regval, pdata->mac_regs + MTL_OMR); } static void xlgmac_config_queue_mapping(struct xlgmac_pdata *pdata) { unsigned int ppq, ppq_extra, prio, prio_queues; unsigned int qptc, qptc_extra, queue; unsigned int reg, regval; unsigned int mask; unsigned int i, j; /* Map the MTL Tx Queues to Traffic Classes * Note: Tx Queues >= Traffic Classes */ qptc = pdata->tx_q_count / pdata->hw_feat.tc_cnt; qptc_extra = pdata->tx_q_count % pdata->hw_feat.tc_cnt; for (i = 0, queue = 0; i < pdata->hw_feat.tc_cnt; i++) { for (j = 0; j < qptc; j++) { netif_dbg(pdata, drv, pdata->netdev, "TXq%u mapped to TC%u\n", queue, i); regval = readl(XLGMAC_MTL_REG(pdata, queue, MTL_Q_TQOMR)); regval = XLGMAC_SET_REG_BITS(regval, MTL_Q_TQOMR_Q2TCMAP_POS, MTL_Q_TQOMR_Q2TCMAP_LEN, i); writel(regval, XLGMAC_MTL_REG(pdata, queue, MTL_Q_TQOMR)); queue++; } if (i < qptc_extra) { netif_dbg(pdata, drv, pdata->netdev, "TXq%u mapped to TC%u\n", queue, i); regval = readl(XLGMAC_MTL_REG(pdata, queue, MTL_Q_TQOMR)); regval = XLGMAC_SET_REG_BITS(regval, MTL_Q_TQOMR_Q2TCMAP_POS, MTL_Q_TQOMR_Q2TCMAP_LEN, i); writel(regval, XLGMAC_MTL_REG(pdata, queue, MTL_Q_TQOMR)); queue++; } } /* Map the 8 VLAN priority values to available MTL Rx queues */ prio_queues = min_t(unsigned int, IEEE_8021QAZ_MAX_TCS, pdata->rx_q_count); ppq = IEEE_8021QAZ_MAX_TCS / prio_queues; ppq_extra = IEEE_8021QAZ_MAX_TCS % prio_queues; reg = MAC_RQC2R; regval = 0; for (i = 0, prio = 0; i < prio_queues;) { mask = 0; for (j = 0; j < ppq; j++) { netif_dbg(pdata, drv, pdata->netdev, "PRIO%u mapped to RXq%u\n", prio, i); mask |= (1 << prio); prio++; } if (i < ppq_extra) { netif_dbg(pdata, drv, pdata->netdev, "PRIO%u mapped to RXq%u\n", prio, i); mask |= (1 << prio); prio++; } regval |= (mask << ((i++ % MAC_RQC2_Q_PER_REG) << 3)); if ((i % MAC_RQC2_Q_PER_REG) && (i != prio_queues)) continue; writel(regval, pdata->mac_regs + reg); reg += MAC_RQC2_INC; regval = 0; } /* Configure one to one, MTL Rx queue to DMA Rx channel mapping * ie Q0 <--> CH0, Q1 <--> CH1 ... Q11 <--> CH11 */ reg = MTL_RQDCM0R; regval = readl(pdata->mac_regs + reg); regval |= (MTL_RQDCM0R_Q0MDMACH | MTL_RQDCM0R_Q1MDMACH | MTL_RQDCM0R_Q2MDMACH | MTL_RQDCM0R_Q3MDMACH); writel(regval, pdata->mac_regs + reg); reg += MTL_RQDCM_INC; regval = readl(pdata->mac_regs + reg); regval |= (MTL_RQDCM1R_Q4MDMACH | MTL_RQDCM1R_Q5MDMACH | MTL_RQDCM1R_Q6MDMACH | MTL_RQDCM1R_Q7MDMACH); writel(regval, pdata->mac_regs + reg); reg += MTL_RQDCM_INC; regval = readl(pdata->mac_regs + reg); regval |= (MTL_RQDCM2R_Q8MDMACH | MTL_RQDCM2R_Q9MDMACH | MTL_RQDCM2R_Q10MDMACH | MTL_RQDCM2R_Q11MDMACH); writel(regval, pdata->mac_regs + reg); } static unsigned int xlgmac_calculate_per_queue_fifo( unsigned int fifo_size, unsigned int queue_count) { unsigned int q_fifo_size; unsigned int p_fifo; /* Calculate the configured fifo size */ q_fifo_size = 1 << (fifo_size + 7); /* The configured value may not be the actual amount of fifo RAM */ q_fifo_size = min_t(unsigned int, XLGMAC_MAX_FIFO, q_fifo_size); q_fifo_size = q_fifo_size / queue_count; /* Each increment in the queue fifo size represents 256 bytes of * fifo, with 0 representing 256 bytes. Distribute the fifo equally * between the queues. */ p_fifo = q_fifo_size / 256; if (p_fifo) p_fifo--; return p_fifo; } static void xlgmac_config_tx_fifo_size(struct xlgmac_pdata *pdata) { unsigned int fifo_size; unsigned int i; u32 regval; fifo_size = xlgmac_calculate_per_queue_fifo( pdata->hw_feat.tx_fifo_size, pdata->tx_q_count); for (i = 0; i < pdata->tx_q_count; i++) { regval = readl(XLGMAC_MTL_REG(pdata, i, MTL_Q_TQOMR)); regval = XLGMAC_SET_REG_BITS(regval, MTL_Q_TQOMR_TQS_POS, MTL_Q_TQOMR_TQS_LEN, fifo_size); writel(regval, XLGMAC_MTL_REG(pdata, i, MTL_Q_TQOMR)); } netif_info(pdata, drv, pdata->netdev, "%d Tx hardware queues, %d byte fifo per queue\n", pdata->tx_q_count, ((fifo_size + 1) * 256)); } static void xlgmac_config_rx_fifo_size(struct xlgmac_pdata *pdata) { unsigned int fifo_size; unsigned int i; u32 regval; fifo_size = xlgmac_calculate_per_queue_fifo( pdata->hw_feat.rx_fifo_size, pdata->rx_q_count); for (i = 0; i < pdata->rx_q_count; i++) { regval = readl(XLGMAC_MTL_REG(pdata, i, MTL_Q_RQOMR)); regval = XLGMAC_SET_REG_BITS(regval, MTL_Q_RQOMR_RQS_POS, MTL_Q_RQOMR_RQS_LEN, fifo_size); writel(regval, XLGMAC_MTL_REG(pdata, i, MTL_Q_RQOMR)); } netif_info(pdata, drv, pdata->netdev, "%d Rx hardware queues, %d byte fifo per queue\n", pdata->rx_q_count, ((fifo_size + 1) * 256)); } static void xlgmac_config_flow_control_threshold(struct xlgmac_pdata *pdata) { unsigned int i; u32 regval; for (i = 0; i < pdata->rx_q_count; i++) { regval = readl(XLGMAC_MTL_REG(pdata, i, MTL_Q_RQFCR)); /* Activate flow control when less than 4k left in fifo */ regval = XLGMAC_SET_REG_BITS(regval, MTL_Q_RQFCR_RFA_POS, MTL_Q_RQFCR_RFA_LEN, 2); /* De-activate flow control when more than 6k left in fifo */ regval = XLGMAC_SET_REG_BITS(regval, MTL_Q_RQFCR_RFD_POS, MTL_Q_RQFCR_RFD_LEN, 4); writel(regval, XLGMAC_MTL_REG(pdata, i, MTL_Q_RQFCR)); } } static int xlgmac_config_tx_threshold(struct xlgmac_pdata *pdata, unsigned int val) { unsigned int i; u32 regval; for (i = 0; i < pdata->tx_q_count; i++) { regval = readl(XLGMAC_MTL_REG(pdata, i, MTL_Q_TQOMR)); regval = XLGMAC_SET_REG_BITS(regval, MTL_Q_TQOMR_TTC_POS, MTL_Q_TQOMR_TTC_LEN, val); writel(regval, XLGMAC_MTL_REG(pdata, i, MTL_Q_TQOMR)); } return 0; } static int xlgmac_config_rsf_mode(struct xlgmac_pdata *pdata, unsigned int val) { unsigned int i; u32 regval; for (i = 0; i < pdata->rx_q_count; i++) { regval = readl(XLGMAC_MTL_REG(pdata, i, MTL_Q_RQOMR)); regval = XLGMAC_SET_REG_BITS(regval, MTL_Q_RQOMR_RSF_POS, MTL_Q_RQOMR_RSF_LEN, val); writel(regval, XLGMAC_MTL_REG(pdata, i, MTL_Q_RQOMR)); } return 0; } static int xlgmac_config_tsf_mode(struct xlgmac_pdata *pdata, unsigned int val) { unsigned int i; u32 regval; for (i = 0; i < pdata->tx_q_count; i++) { regval = readl(XLGMAC_MTL_REG(pdata, i, MTL_Q_TQOMR)); regval = XLGMAC_SET_REG_BITS(regval, MTL_Q_TQOMR_TSF_POS, MTL_Q_TQOMR_TSF_LEN, val); writel(regval, XLGMAC_MTL_REG(pdata, i, MTL_Q_TQOMR)); } return 0; } static int xlgmac_config_osp_mode(struct xlgmac_pdata *pdata) { struct xlgmac_channel *channel; unsigned int i; u32 regval; channel = pdata->channel_head; for (i = 0; i < pdata->channel_count; i++, channel++) { if (!channel->tx_ring) break; regval = readl(XLGMAC_DMA_REG(channel, DMA_CH_TCR)); regval = XLGMAC_SET_REG_BITS(regval, DMA_CH_TCR_OSP_POS, DMA_CH_TCR_OSP_LEN, pdata->tx_osp_mode); writel(regval, XLGMAC_DMA_REG(channel, DMA_CH_TCR)); } return 0; } static int xlgmac_config_pblx8(struct xlgmac_pdata *pdata) { struct xlgmac_channel *channel; unsigned int i; u32 regval; channel = pdata->channel_head; for (i = 0; i < pdata->channel_count; i++, channel++) { regval = readl(XLGMAC_DMA_REG(channel, DMA_CH_CR)); regval = XLGMAC_SET_REG_BITS(regval, DMA_CH_CR_PBLX8_POS, DMA_CH_CR_PBLX8_LEN, pdata->pblx8); writel(regval, XLGMAC_DMA_REG(channel, DMA_CH_CR)); } return 0; } static int xlgmac_get_tx_pbl_val(struct xlgmac_pdata *pdata) { u32 regval; regval = readl(XLGMAC_DMA_REG(pdata->channel_head, DMA_CH_TCR)); regval = XLGMAC_GET_REG_BITS(regval, DMA_CH_TCR_PBL_POS, DMA_CH_TCR_PBL_LEN); return regval; } static int xlgmac_config_tx_pbl_val(struct xlgmac_pdata *pdata) { struct xlgmac_channel *channel; unsigned int i; u32 regval; channel = pdata->channel_head; for (i = 0; i < pdata->channel_count; i++, channel++) { if (!channel->tx_ring) break; regval = readl(XLGMAC_DMA_REG(channel, DMA_CH_TCR)); regval = XLGMAC_SET_REG_BITS(regval, DMA_CH_TCR_PBL_POS, DMA_CH_TCR_PBL_LEN, pdata->tx_pbl); writel(regval, XLGMAC_DMA_REG(channel, DMA_CH_TCR)); } return 0; } static int xlgmac_get_rx_pbl_val(struct xlgmac_pdata *pdata) { u32 regval; regval = readl(XLGMAC_DMA_REG(pdata->channel_head, DMA_CH_RCR)); regval = XLGMAC_GET_REG_BITS(regval, DMA_CH_RCR_PBL_POS, DMA_CH_RCR_PBL_LEN); return regval; } static int xlgmac_config_rx_pbl_val(struct xlgmac_pdata *pdata) { struct xlgmac_channel *channel; unsigned int i; u32 regval; channel = pdata->channel_head; for (i = 0; i < pdata->channel_count; i++, channel++) { if (!channel->rx_ring) break; regval = readl(XLGMAC_DMA_REG(channel, DMA_CH_RCR)); regval = XLGMAC_SET_REG_BITS(regval, DMA_CH_RCR_PBL_POS, DMA_CH_RCR_PBL_LEN, pdata->rx_pbl); writel(regval, XLGMAC_DMA_REG(channel, DMA_CH_RCR)); } return 0; } static u64 xlgmac_mmc_read(struct xlgmac_pdata *pdata, unsigned int reg_lo) { bool read_hi; u64 val; switch (reg_lo) { /* These registers are always 64 bit */ case MMC_TXOCTETCOUNT_GB_LO: case MMC_TXOCTETCOUNT_G_LO: case MMC_RXOCTETCOUNT_GB_LO: case MMC_RXOCTETCOUNT_G_LO: read_hi = true; break; default: read_hi = false; } val = (u64)readl(pdata->mac_regs + reg_lo); if (read_hi) val |= ((u64)readl(pdata->mac_regs + reg_lo + 4) << 32); return val; } static void xlgmac_tx_mmc_int(struct xlgmac_pdata *pdata) { unsigned int mmc_isr = readl(pdata->mac_regs + MMC_TISR); struct xlgmac_stats *stats = &pdata->stats; if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_TISR_TXOCTETCOUNT_GB_POS, MMC_TISR_TXOCTETCOUNT_GB_LEN)) stats->txoctetcount_gb += xlgmac_mmc_read(pdata, MMC_TXOCTETCOUNT_GB_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_TISR_TXFRAMECOUNT_GB_POS, MMC_TISR_TXFRAMECOUNT_GB_LEN)) stats->txframecount_gb += xlgmac_mmc_read(pdata, MMC_TXFRAMECOUNT_GB_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_TISR_TXBROADCASTFRAMES_G_POS, MMC_TISR_TXBROADCASTFRAMES_G_LEN)) stats->txbroadcastframes_g += xlgmac_mmc_read(pdata, MMC_TXBROADCASTFRAMES_G_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_TISR_TXMULTICASTFRAMES_G_POS, MMC_TISR_TXMULTICASTFRAMES_G_LEN)) stats->txmulticastframes_g += xlgmac_mmc_read(pdata, MMC_TXMULTICASTFRAMES_G_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_TISR_TX64OCTETS_GB_POS, MMC_TISR_TX64OCTETS_GB_LEN)) stats->tx64octets_gb += xlgmac_mmc_read(pdata, MMC_TX64OCTETS_GB_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_TISR_TX65TO127OCTETS_GB_POS, MMC_TISR_TX65TO127OCTETS_GB_LEN)) stats->tx65to127octets_gb += xlgmac_mmc_read(pdata, MMC_TX65TO127OCTETS_GB_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_TISR_TX128TO255OCTETS_GB_POS, MMC_TISR_TX128TO255OCTETS_GB_LEN)) stats->tx128to255octets_gb += xlgmac_mmc_read(pdata, MMC_TX128TO255OCTETS_GB_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_TISR_TX256TO511OCTETS_GB_POS, MMC_TISR_TX256TO511OCTETS_GB_LEN)) stats->tx256to511octets_gb += xlgmac_mmc_read(pdata, MMC_TX256TO511OCTETS_GB_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_TISR_TX512TO1023OCTETS_GB_POS, MMC_TISR_TX512TO1023OCTETS_GB_LEN)) stats->tx512to1023octets_gb += xlgmac_mmc_read(pdata, MMC_TX512TO1023OCTETS_GB_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_TISR_TX1024TOMAXOCTETS_GB_POS, MMC_TISR_TX1024TOMAXOCTETS_GB_LEN)) stats->tx1024tomaxoctets_gb += xlgmac_mmc_read(pdata, MMC_TX1024TOMAXOCTETS_GB_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_TISR_TXUNICASTFRAMES_GB_POS, MMC_TISR_TXUNICASTFRAMES_GB_LEN)) stats->txunicastframes_gb += xlgmac_mmc_read(pdata, MMC_TXUNICASTFRAMES_GB_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_TISR_TXMULTICASTFRAMES_GB_POS, MMC_TISR_TXMULTICASTFRAMES_GB_LEN)) stats->txmulticastframes_gb += xlgmac_mmc_read(pdata, MMC_TXMULTICASTFRAMES_GB_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_TISR_TXBROADCASTFRAMES_GB_POS, MMC_TISR_TXBROADCASTFRAMES_GB_LEN)) stats->txbroadcastframes_g += xlgmac_mmc_read(pdata, MMC_TXBROADCASTFRAMES_GB_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_TISR_TXUNDERFLOWERROR_POS, MMC_TISR_TXUNDERFLOWERROR_LEN)) stats->txunderflowerror += xlgmac_mmc_read(pdata, MMC_TXUNDERFLOWERROR_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_TISR_TXOCTETCOUNT_G_POS, MMC_TISR_TXOCTETCOUNT_G_LEN)) stats->txoctetcount_g += xlgmac_mmc_read(pdata, MMC_TXOCTETCOUNT_G_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_TISR_TXFRAMECOUNT_G_POS, MMC_TISR_TXFRAMECOUNT_G_LEN)) stats->txframecount_g += xlgmac_mmc_read(pdata, MMC_TXFRAMECOUNT_G_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_TISR_TXPAUSEFRAMES_POS, MMC_TISR_TXPAUSEFRAMES_LEN)) stats->txpauseframes += xlgmac_mmc_read(pdata, MMC_TXPAUSEFRAMES_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_TISR_TXVLANFRAMES_G_POS, MMC_TISR_TXVLANFRAMES_G_LEN)) stats->txvlanframes_g += xlgmac_mmc_read(pdata, MMC_TXVLANFRAMES_G_LO); } static void xlgmac_rx_mmc_int(struct xlgmac_pdata *pdata) { unsigned int mmc_isr = readl(pdata->mac_regs + MMC_RISR); struct xlgmac_stats *stats = &pdata->stats; if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_RISR_RXFRAMECOUNT_GB_POS, MMC_RISR_RXFRAMECOUNT_GB_LEN)) stats->rxframecount_gb += xlgmac_mmc_read(pdata, MMC_RXFRAMECOUNT_GB_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_RISR_RXOCTETCOUNT_GB_POS, MMC_RISR_RXOCTETCOUNT_GB_LEN)) stats->rxoctetcount_gb += xlgmac_mmc_read(pdata, MMC_RXOCTETCOUNT_GB_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_RISR_RXOCTETCOUNT_G_POS, MMC_RISR_RXOCTETCOUNT_G_LEN)) stats->rxoctetcount_g += xlgmac_mmc_read(pdata, MMC_RXOCTETCOUNT_G_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_RISR_RXBROADCASTFRAMES_G_POS, MMC_RISR_RXBROADCASTFRAMES_G_LEN)) stats->rxbroadcastframes_g += xlgmac_mmc_read(pdata, MMC_RXBROADCASTFRAMES_G_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_RISR_RXMULTICASTFRAMES_G_POS, MMC_RISR_RXMULTICASTFRAMES_G_LEN)) stats->rxmulticastframes_g += xlgmac_mmc_read(pdata, MMC_RXMULTICASTFRAMES_G_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_RISR_RXCRCERROR_POS, MMC_RISR_RXCRCERROR_LEN)) stats->rxcrcerror += xlgmac_mmc_read(pdata, MMC_RXCRCERROR_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_RISR_RXRUNTERROR_POS, MMC_RISR_RXRUNTERROR_LEN)) stats->rxrunterror += xlgmac_mmc_read(pdata, MMC_RXRUNTERROR); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_RISR_RXJABBERERROR_POS, MMC_RISR_RXJABBERERROR_LEN)) stats->rxjabbererror += xlgmac_mmc_read(pdata, MMC_RXJABBERERROR); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_RISR_RXUNDERSIZE_G_POS, MMC_RISR_RXUNDERSIZE_G_LEN)) stats->rxundersize_g += xlgmac_mmc_read(pdata, MMC_RXUNDERSIZE_G); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_RISR_RXOVERSIZE_G_POS, MMC_RISR_RXOVERSIZE_G_LEN)) stats->rxoversize_g += xlgmac_mmc_read(pdata, MMC_RXOVERSIZE_G); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_RISR_RX64OCTETS_GB_POS, MMC_RISR_RX64OCTETS_GB_LEN)) stats->rx64octets_gb += xlgmac_mmc_read(pdata, MMC_RX64OCTETS_GB_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_RISR_RX65TO127OCTETS_GB_POS, MMC_RISR_RX65TO127OCTETS_GB_LEN)) stats->rx65to127octets_gb += xlgmac_mmc_read(pdata, MMC_RX65TO127OCTETS_GB_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_RISR_RX128TO255OCTETS_GB_POS, MMC_RISR_RX128TO255OCTETS_GB_LEN)) stats->rx128to255octets_gb += xlgmac_mmc_read(pdata, MMC_RX128TO255OCTETS_GB_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_RISR_RX256TO511OCTETS_GB_POS, MMC_RISR_RX256TO511OCTETS_GB_LEN)) stats->rx256to511octets_gb += xlgmac_mmc_read(pdata, MMC_RX256TO511OCTETS_GB_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_RISR_RX512TO1023OCTETS_GB_POS, MMC_RISR_RX512TO1023OCTETS_GB_LEN)) stats->rx512to1023octets_gb += xlgmac_mmc_read(pdata, MMC_RX512TO1023OCTETS_GB_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_RISR_RX1024TOMAXOCTETS_GB_POS, MMC_RISR_RX1024TOMAXOCTETS_GB_LEN)) stats->rx1024tomaxoctets_gb += xlgmac_mmc_read(pdata, MMC_RX1024TOMAXOCTETS_GB_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_RISR_RXUNICASTFRAMES_G_POS, MMC_RISR_RXUNICASTFRAMES_G_LEN)) stats->rxunicastframes_g += xlgmac_mmc_read(pdata, MMC_RXUNICASTFRAMES_G_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_RISR_RXLENGTHERROR_POS, MMC_RISR_RXLENGTHERROR_LEN)) stats->rxlengtherror += xlgmac_mmc_read(pdata, MMC_RXLENGTHERROR_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_RISR_RXOUTOFRANGETYPE_POS, MMC_RISR_RXOUTOFRANGETYPE_LEN)) stats->rxoutofrangetype += xlgmac_mmc_read(pdata, MMC_RXOUTOFRANGETYPE_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_RISR_RXPAUSEFRAMES_POS, MMC_RISR_RXPAUSEFRAMES_LEN)) stats->rxpauseframes += xlgmac_mmc_read(pdata, MMC_RXPAUSEFRAMES_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_RISR_RXFIFOOVERFLOW_POS, MMC_RISR_RXFIFOOVERFLOW_LEN)) stats->rxfifooverflow += xlgmac_mmc_read(pdata, MMC_RXFIFOOVERFLOW_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_RISR_RXVLANFRAMES_GB_POS, MMC_RISR_RXVLANFRAMES_GB_LEN)) stats->rxvlanframes_gb += xlgmac_mmc_read(pdata, MMC_RXVLANFRAMES_GB_LO); if (XLGMAC_GET_REG_BITS(mmc_isr, MMC_RISR_RXWATCHDOGERROR_POS, MMC_RISR_RXWATCHDOGERROR_LEN)) stats->rxwatchdogerror += xlgmac_mmc_read(pdata, MMC_RXWATCHDOGERROR); } static void xlgmac_read_mmc_stats(struct xlgmac_pdata *pdata) { struct xlgmac_stats *stats = &pdata->stats; u32 regval; /* Freeze counters */ regval = readl(pdata->mac_regs + MMC_CR); regval = XLGMAC_SET_REG_BITS(regval, MMC_CR_MCF_POS, MMC_CR_MCF_LEN, 1); writel(regval, pdata->mac_regs + MMC_CR); stats->txoctetcount_gb += xlgmac_mmc_read(pdata, MMC_TXOCTETCOUNT_GB_LO); stats->txframecount_gb += xlgmac_mmc_read(pdata, MMC_TXFRAMECOUNT_GB_LO); stats->txbroadcastframes_g += xlgmac_mmc_read(pdata, MMC_TXBROADCASTFRAMES_G_LO); stats->txmulticastframes_g += xlgmac_mmc_read(pdata, MMC_TXMULTICASTFRAMES_G_LO); stats->tx64octets_gb += xlgmac_mmc_read(pdata, MMC_TX64OCTETS_GB_LO); stats->tx65to127octets_gb += xlgmac_mmc_read(pdata, MMC_TX65TO127OCTETS_GB_LO); stats->tx128to255octets_gb += xlgmac_mmc_read(pdata, MMC_TX128TO255OCTETS_GB_LO); stats->tx256to511octets_gb += xlgmac_mmc_read(pdata, MMC_TX256TO511OCTETS_GB_LO); stats->tx512to1023octets_gb += xlgmac_mmc_read(pdata, MMC_TX512TO1023OCTETS_GB_LO); stats->tx1024tomaxoctets_gb += xlgmac_mmc_read(pdata, MMC_TX1024TOMAXOCTETS_GB_LO); stats->txunicastframes_gb += xlgmac_mmc_read(pdata, MMC_TXUNICASTFRAMES_GB_LO); stats->txmulticastframes_gb += xlgmac_mmc_read(pdata, MMC_TXMULTICASTFRAMES_GB_LO); stats->txbroadcastframes_g += xlgmac_mmc_read(pdata, MMC_TXBROADCASTFRAMES_GB_LO); stats->txunderflowerror += xlgmac_mmc_read(pdata, MMC_TXUNDERFLOWERROR_LO); stats->txoctetcount_g += xlgmac_mmc_read(pdata, MMC_TXOCTETCOUNT_G_LO); stats->txframecount_g += xlgmac_mmc_read(pdata, MMC_TXFRAMECOUNT_G_LO); stats->txpauseframes += xlgmac_mmc_read(pdata, MMC_TXPAUSEFRAMES_LO); stats->txvlanframes_g += xlgmac_mmc_read(pdata, MMC_TXVLANFRAMES_G_LO); stats->rxframecount_gb += xlgmac_mmc_read(pdata, MMC_RXFRAMECOUNT_GB_LO); stats->rxoctetcount_gb += xlgmac_mmc_read(pdata, MMC_RXOCTETCOUNT_GB_LO); stats->rxoctetcount_g += xlgmac_mmc_read(pdata, MMC_RXOCTETCOUNT_G_LO); stats->rxbroadcastframes_g += xlgmac_mmc_read(pdata, MMC_RXBROADCASTFRAMES_G_LO); stats->rxmulticastframes_g += xlgmac_mmc_read(pdata, MMC_RXMULTICASTFRAMES_G_LO); stats->rxcrcerror += xlgmac_mmc_read(pdata, MMC_RXCRCERROR_LO); stats->rxrunterror += xlgmac_mmc_read(pdata, MMC_RXRUNTERROR); stats->rxjabbererror += xlgmac_mmc_read(pdata, MMC_RXJABBERERROR); stats->rxundersize_g += xlgmac_mmc_read(pdata, MMC_RXUNDERSIZE_G); stats->rxoversize_g += xlgmac_mmc_read(pdata, MMC_RXOVERSIZE_G); stats->rx64octets_gb += xlgmac_mmc_read(pdata, MMC_RX64OCTETS_GB_LO); stats->rx65to127octets_gb += xlgmac_mmc_read(pdata, MMC_RX65TO127OCTETS_GB_LO); stats->rx128to255octets_gb += xlgmac_mmc_read(pdata, MMC_RX128TO255OCTETS_GB_LO); stats->rx256to511octets_gb += xlgmac_mmc_read(pdata, MMC_RX256TO511OCTETS_GB_LO); stats->rx512to1023octets_gb += xlgmac_mmc_read(pdata, MMC_RX512TO1023OCTETS_GB_LO); stats->rx1024tomaxoctets_gb += xlgmac_mmc_read(pdata, MMC_RX1024TOMAXOCTETS_GB_LO); stats->rxunicastframes_g += xlgmac_mmc_read(pdata, MMC_RXUNICASTFRAMES_G_LO); stats->rxlengtherror += xlgmac_mmc_read(pdata, MMC_RXLENGTHERROR_LO); stats->rxoutofrangetype += xlgmac_mmc_read(pdata, MMC_RXOUTOFRANGETYPE_LO); stats->rxpauseframes += xlgmac_mmc_read(pdata, MMC_RXPAUSEFRAMES_LO); stats->rxfifooverflow += xlgmac_mmc_read(pdata, MMC_RXFIFOOVERFLOW_LO); stats->rxvlanframes_gb += xlgmac_mmc_read(pdata, MMC_RXVLANFRAMES_GB_LO); stats->rxwatchdogerror += xlgmac_mmc_read(pdata, MMC_RXWATCHDOGERROR); /* Un-freeze counters */ regval = readl(pdata->mac_regs + MMC_CR); regval = XLGMAC_SET_REG_BITS(regval, MMC_CR_MCF_POS, MMC_CR_MCF_LEN, 0); writel(regval, pdata->mac_regs + MMC_CR); } static void xlgmac_config_mmc(struct xlgmac_pdata *pdata) { u32 regval; regval = readl(pdata->mac_regs + MMC_CR); /* Set counters to reset on read */ regval = XLGMAC_SET_REG_BITS(regval, MMC_CR_ROR_POS, MMC_CR_ROR_LEN, 1); /* Reset the counters */ regval = XLGMAC_SET_REG_BITS(regval, MMC_CR_CR_POS, MMC_CR_CR_LEN, 1); writel(regval, pdata->mac_regs + MMC_CR); } static int xlgmac_write_rss_reg(struct xlgmac_pdata *pdata, unsigned int type, unsigned int index, unsigned int val) { unsigned int wait; int ret = 0; u32 regval; mutex_lock(&pdata->rss_mutex); regval = XLGMAC_GET_REG_BITS(readl(pdata->mac_regs + MAC_RSSAR), MAC_RSSAR_OB_POS, MAC_RSSAR_OB_LEN); if (regval) { ret = -EBUSY; goto unlock; } writel(val, pdata->mac_regs + MAC_RSSDR); regval = readl(pdata->mac_regs + MAC_RSSAR); regval = XLGMAC_SET_REG_BITS(regval, MAC_RSSAR_RSSIA_POS, MAC_RSSAR_RSSIA_LEN, index); regval = XLGMAC_SET_REG_BITS(regval, MAC_RSSAR_ADDRT_POS, MAC_RSSAR_ADDRT_LEN, type); regval = XLGMAC_SET_REG_BITS(regval, MAC_RSSAR_CT_POS, MAC_RSSAR_CT_LEN, 0); regval = XLGMAC_SET_REG_BITS(regval, MAC_RSSAR_OB_POS, MAC_RSSAR_OB_LEN, 1); writel(regval, pdata->mac_regs + MAC_RSSAR); wait = 1000; while (wait--) { regval = XLGMAC_GET_REG_BITS(readl(pdata->mac_regs + MAC_RSSAR), MAC_RSSAR_OB_POS, MAC_RSSAR_OB_LEN); if (!regval) goto unlock; usleep_range(1000, 1500); } ret = -EBUSY; unlock: mutex_unlock(&pdata->rss_mutex); return ret; } static int xlgmac_write_rss_hash_key(struct xlgmac_pdata *pdata) { unsigned int key_regs = sizeof(pdata->rss_key) / sizeof(u32); unsigned int *key = (unsigned int *)&pdata->rss_key; int ret; while (key_regs--) { ret = xlgmac_write_rss_reg(pdata, XLGMAC_RSS_HASH_KEY_TYPE, key_regs, *key++); if (ret) return ret; } return 0; } static int xlgmac_write_rss_lookup_table(struct xlgmac_pdata *pdata) { unsigned int i; int ret; for (i = 0; i < ARRAY_SIZE(pdata->rss_table); i++) { ret = xlgmac_write_rss_reg(pdata, XLGMAC_RSS_LOOKUP_TABLE_TYPE, i, pdata->rss_table[i]); if (ret) return ret; } return 0; } static int xlgmac_set_rss_hash_key(struct xlgmac_pdata *pdata, const u8 *key) { memcpy(pdata->rss_key, key, sizeof(pdata->rss_key)); return xlgmac_write_rss_hash_key(pdata); } static int xlgmac_set_rss_lookup_table(struct xlgmac_pdata *pdata, const u32 *table) { unsigned int i; u32 tval; for (i = 0; i < ARRAY_SIZE(pdata->rss_table); i++) { tval = table[i]; pdata->rss_table[i] = XLGMAC_SET_REG_BITS( pdata->rss_table[i], MAC_RSSDR_DMCH_POS, MAC_RSSDR_DMCH_LEN, tval); } return xlgmac_write_rss_lookup_table(pdata); } static int xlgmac_enable_rss(struct xlgmac_pdata *pdata) { u32 regval; int ret; if (!pdata->hw_feat.rss) return -EOPNOTSUPP; /* Program the hash key */ ret = xlgmac_write_rss_hash_key(pdata); if (ret) return ret; /* Program the lookup table */ ret = xlgmac_write_rss_lookup_table(pdata); if (ret) return ret; /* Set the RSS options */ writel(pdata->rss_options, pdata->mac_regs + MAC_RSSCR); /* Enable RSS */ regval = readl(pdata->mac_regs + MAC_RSSCR); regval = XLGMAC_SET_REG_BITS(regval, MAC_RSSCR_RSSE_POS, MAC_RSSCR_RSSE_LEN, 1); writel(regval, pdata->mac_regs + MAC_RSSCR); return 0; } static int xlgmac_disable_rss(struct xlgmac_pdata *pdata) { u32 regval; if (!pdata->hw_feat.rss) return -EOPNOTSUPP; regval = readl(pdata->mac_regs + MAC_RSSCR); regval = XLGMAC_SET_REG_BITS(regval, MAC_RSSCR_RSSE_POS, MAC_RSSCR_RSSE_LEN, 0); writel(regval, pdata->mac_regs + MAC_RSSCR); return 0; } static void xlgmac_config_rss(struct xlgmac_pdata *pdata) { int ret; if (!pdata->hw_feat.rss) return; if (pdata->netdev->features & NETIF_F_RXHASH) ret = xlgmac_enable_rss(pdata); else ret = xlgmac_disable_rss(pdata); if (ret) netdev_err(pdata->netdev, "error configuring RSS, RSS disabled\n"); } static void xlgmac_enable_dma_interrupts(struct xlgmac_pdata *pdata) { unsigned int dma_ch_isr, dma_ch_ier; struct xlgmac_channel *channel; unsigned int i; channel = pdata->channel_head; for (i = 0; i < pdata->channel_count; i++, channel++) { /* Clear all the interrupts which are set */ dma_ch_isr = readl(XLGMAC_DMA_REG(channel, DMA_CH_SR)); writel(dma_ch_isr, XLGMAC_DMA_REG(channel, DMA_CH_SR)); /* Clear all interrupt enable bits */ dma_ch_ier = 0; /* Enable following interrupts * NIE - Normal Interrupt Summary Enable * AIE - Abnormal Interrupt Summary Enable * FBEE - Fatal Bus Error Enable */ dma_ch_ier = XLGMAC_SET_REG_BITS(dma_ch_ier, DMA_CH_IER_NIE_POS, DMA_CH_IER_NIE_LEN, 1); dma_ch_ier = XLGMAC_SET_REG_BITS(dma_ch_ier, DMA_CH_IER_AIE_POS, DMA_CH_IER_AIE_LEN, 1); dma_ch_ier = XLGMAC_SET_REG_BITS(dma_ch_ier, DMA_CH_IER_FBEE_POS, DMA_CH_IER_FBEE_LEN, 1); if (channel->tx_ring) { /* Enable the following Tx interrupts * TIE - Transmit Interrupt Enable (unless using * per channel interrupts) */ if (!pdata->per_channel_irq) dma_ch_ier = XLGMAC_SET_REG_BITS( dma_ch_ier, DMA_CH_IER_TIE_POS, DMA_CH_IER_TIE_LEN, 1); } if (channel->rx_ring) { /* Enable following Rx interrupts * RBUE - Receive Buffer Unavailable Enable * RIE - Receive Interrupt Enable (unless using * per channel interrupts) */ dma_ch_ier = XLGMAC_SET_REG_BITS( dma_ch_ier, DMA_CH_IER_RBUE_POS, DMA_CH_IER_RBUE_LEN, 1); if (!pdata->per_channel_irq) dma_ch_ier = XLGMAC_SET_REG_BITS( dma_ch_ier, DMA_CH_IER_RIE_POS, DMA_CH_IER_RIE_LEN, 1); } writel(dma_ch_isr, XLGMAC_DMA_REG(channel, DMA_CH_IER)); } } static void xlgmac_enable_mtl_interrupts(struct xlgmac_pdata *pdata) { unsigned int q_count, i; unsigned int mtl_q_isr; q_count = max(pdata->hw_feat.tx_q_cnt, pdata->hw_feat.rx_q_cnt); for (i = 0; i < q_count; i++) { /* Clear all the interrupts which are set */ mtl_q_isr = readl(XLGMAC_MTL_REG(pdata, i, MTL_Q_ISR)); writel(mtl_q_isr, XLGMAC_MTL_REG(pdata, i, MTL_Q_ISR)); /* No MTL interrupts to be enabled */ writel(0, XLGMAC_MTL_REG(pdata, i, MTL_Q_IER)); } } static void xlgmac_enable_mac_interrupts(struct xlgmac_pdata *pdata) { unsigned int mac_ier = 0; u32 regval; /* Enable Timestamp interrupt */ mac_ier = XLGMAC_SET_REG_BITS(mac_ier, MAC_IER_TSIE_POS, MAC_IER_TSIE_LEN, 1); writel(mac_ier, pdata->mac_regs + MAC_IER); /* Enable all counter interrupts */ regval = readl(pdata->mac_regs + MMC_RIER); regval = XLGMAC_SET_REG_BITS(regval, MMC_RIER_ALL_INTERRUPTS_POS, MMC_RIER_ALL_INTERRUPTS_LEN, 0xffffffff); writel(regval, pdata->mac_regs + MMC_RIER); regval = readl(pdata->mac_regs + MMC_TIER); regval = XLGMAC_SET_REG_BITS(regval, MMC_TIER_ALL_INTERRUPTS_POS, MMC_TIER_ALL_INTERRUPTS_LEN, 0xffffffff); writel(regval, pdata->mac_regs + MMC_TIER); } static int xlgmac_set_xlgmii_25000_speed(struct xlgmac_pdata *pdata) { u32 regval; regval = XLGMAC_GET_REG_BITS(readl(pdata->mac_regs + MAC_TCR), MAC_TCR_SS_POS, MAC_TCR_SS_LEN); if (regval == 0x1) return 0; regval = readl(pdata->mac_regs + MAC_TCR); regval = XLGMAC_SET_REG_BITS(regval, MAC_TCR_SS_POS, MAC_TCR_SS_LEN, 0x1); writel(regval, pdata->mac_regs + MAC_TCR); return 0; } static int xlgmac_set_xlgmii_40000_speed(struct xlgmac_pdata *pdata) { u32 regval; regval = XLGMAC_GET_REG_BITS(readl(pdata->mac_regs + MAC_TCR), MAC_TCR_SS_POS, MAC_TCR_SS_LEN); if (regval == 0) return 0; regval = readl(pdata->mac_regs + MAC_TCR); regval = XLGMAC_SET_REG_BITS(regval, MAC_TCR_SS_POS, MAC_TCR_SS_LEN, 0); writel(regval, pdata->mac_regs + MAC_TCR); return 0; } static int xlgmac_set_xlgmii_50000_speed(struct xlgmac_pdata *pdata) { u32 regval; regval = XLGMAC_GET_REG_BITS(readl(pdata->mac_regs + MAC_TCR), MAC_TCR_SS_POS, MAC_TCR_SS_LEN); if (regval == 0x2) return 0; regval = readl(pdata->mac_regs + MAC_TCR); regval = XLGMAC_SET_REG_BITS(regval, MAC_TCR_SS_POS, MAC_TCR_SS_LEN, 0x2); writel(regval, pdata->mac_regs + MAC_TCR); return 0; } static int xlgmac_set_xlgmii_100000_speed(struct xlgmac_pdata *pdata) { u32 regval; regval = XLGMAC_GET_REG_BITS(readl(pdata->mac_regs + MAC_TCR), MAC_TCR_SS_POS, MAC_TCR_SS_LEN); if (regval == 0x3) return 0; regval = readl(pdata->mac_regs + MAC_TCR); regval = XLGMAC_SET_REG_BITS(regval, MAC_TCR_SS_POS, MAC_TCR_SS_LEN, 0x3); writel(regval, pdata->mac_regs + MAC_TCR); return 0; } static void xlgmac_config_mac_speed(struct xlgmac_pdata *pdata) { switch (pdata->phy_speed) { case SPEED_100000: xlgmac_set_xlgmii_100000_speed(pdata); break; case SPEED_50000: xlgmac_set_xlgmii_50000_speed(pdata); break; case SPEED_40000: xlgmac_set_xlgmii_40000_speed(pdata); break; case SPEED_25000: xlgmac_set_xlgmii_25000_speed(pdata); break; } } static int xlgmac_dev_read(struct xlgmac_channel *channel) { struct xlgmac_pdata *pdata = channel->pdata; struct xlgmac_ring *ring = channel->rx_ring; struct net_device *netdev = pdata->netdev; struct xlgmac_desc_data *desc_data; struct xlgmac_dma_desc *dma_desc; struct xlgmac_pkt_info *pkt_info; unsigned int err, etlt, l34t; desc_data = XLGMAC_GET_DESC_DATA(ring, ring->cur); dma_desc = desc_data->dma_desc; pkt_info = &ring->pkt_info; /* Check for data availability */ if (XLGMAC_GET_REG_BITS_LE(dma_desc->desc3, RX_NORMAL_DESC3_OWN_POS, RX_NORMAL_DESC3_OWN_LEN)) return 1; /* Make sure descriptor fields are read after reading the OWN bit */ dma_rmb(); if (netif_msg_rx_status(pdata)) xlgmac_dump_rx_desc(pdata, ring, ring->cur); if (XLGMAC_GET_REG_BITS_LE(dma_desc->desc3, RX_NORMAL_DESC3_CTXT_POS, RX_NORMAL_DESC3_CTXT_LEN)) { /* Timestamp Context Descriptor */ xlgmac_get_rx_tstamp(pkt_info, dma_desc); pkt_info->attributes = XLGMAC_SET_REG_BITS( pkt_info->attributes, RX_PACKET_ATTRIBUTES_CONTEXT_POS, RX_PACKET_ATTRIBUTES_CONTEXT_LEN, 1); pkt_info->attributes = XLGMAC_SET_REG_BITS( pkt_info->attributes, RX_PACKET_ATTRIBUTES_CONTEXT_NEXT_POS, RX_PACKET_ATTRIBUTES_CONTEXT_NEXT_LEN, 0); return 0; } /* Normal Descriptor, be sure Context Descriptor bit is off */ pkt_info->attributes = XLGMAC_SET_REG_BITS( pkt_info->attributes, RX_PACKET_ATTRIBUTES_CONTEXT_POS, RX_PACKET_ATTRIBUTES_CONTEXT_LEN, 0); /* Indicate if a Context Descriptor is next */ if (XLGMAC_GET_REG_BITS_LE(dma_desc->desc3, RX_NORMAL_DESC3_CDA_POS, RX_NORMAL_DESC3_CDA_LEN)) pkt_info->attributes = XLGMAC_SET_REG_BITS( pkt_info->attributes, RX_PACKET_ATTRIBUTES_CONTEXT_NEXT_POS, RX_PACKET_ATTRIBUTES_CONTEXT_NEXT_LEN, 1); /* Get the header length */ if (XLGMAC_GET_REG_BITS_LE(dma_desc->desc3, RX_NORMAL_DESC3_FD_POS, RX_NORMAL_DESC3_FD_LEN)) { desc_data->rx.hdr_len = XLGMAC_GET_REG_BITS_LE(dma_desc->desc2, RX_NORMAL_DESC2_HL_POS, RX_NORMAL_DESC2_HL_LEN); if (desc_data->rx.hdr_len) pdata->stats.rx_split_header_packets++; } /* Get the RSS hash */ if (XLGMAC_GET_REG_BITS_LE(dma_desc->desc3, RX_NORMAL_DESC3_RSV_POS, RX_NORMAL_DESC3_RSV_LEN)) { pkt_info->attributes = XLGMAC_SET_REG_BITS( pkt_info->attributes, RX_PACKET_ATTRIBUTES_RSS_HASH_POS, RX_PACKET_ATTRIBUTES_RSS_HASH_LEN, 1); pkt_info->rss_hash = le32_to_cpu(dma_desc->desc1); l34t = XLGMAC_GET_REG_BITS_LE(dma_desc->desc3, RX_NORMAL_DESC3_L34T_POS, RX_NORMAL_DESC3_L34T_LEN); switch (l34t) { case RX_DESC3_L34T_IPV4_TCP: case RX_DESC3_L34T_IPV4_UDP: case RX_DESC3_L34T_IPV6_TCP: case RX_DESC3_L34T_IPV6_UDP: pkt_info->rss_hash_type = PKT_HASH_TYPE_L4; break; default: pkt_info->rss_hash_type = PKT_HASH_TYPE_L3; } } /* Get the pkt_info length */ desc_data->rx.len = XLGMAC_GET_REG_BITS_LE(dma_desc->desc3, RX_NORMAL_DESC3_PL_POS, RX_NORMAL_DESC3_PL_LEN); if (!XLGMAC_GET_REG_BITS_LE(dma_desc->desc3, RX_NORMAL_DESC3_LD_POS, RX_NORMAL_DESC3_LD_LEN)) { /* Not all the data has been transferred for this pkt_info */ pkt_info->attributes = XLGMAC_SET_REG_BITS( pkt_info->attributes, RX_PACKET_ATTRIBUTES_INCOMPLETE_POS, RX_PACKET_ATTRIBUTES_INCOMPLETE_LEN, 1); return 0; } /* This is the last of the data for this pkt_info */ pkt_info->attributes = XLGMAC_SET_REG_BITS( pkt_info->attributes, RX_PACKET_ATTRIBUTES_INCOMPLETE_POS, RX_PACKET_ATTRIBUTES_INCOMPLETE_LEN, 0); /* Set checksum done indicator as appropriate */ if (netdev->features & NETIF_F_RXCSUM) pkt_info->attributes = XLGMAC_SET_REG_BITS( pkt_info->attributes, RX_PACKET_ATTRIBUTES_CSUM_DONE_POS, RX_PACKET_ATTRIBUTES_CSUM_DONE_LEN, 1); /* Check for errors (only valid in last descriptor) */ err = XLGMAC_GET_REG_BITS_LE(dma_desc->desc3, RX_NORMAL_DESC3_ES_POS, RX_NORMAL_DESC3_ES_LEN); etlt = XLGMAC_GET_REG_BITS_LE(dma_desc->desc3, RX_NORMAL_DESC3_ETLT_POS, RX_NORMAL_DESC3_ETLT_LEN); netif_dbg(pdata, rx_status, netdev, "err=%u, etlt=%#x\n", err, etlt); if (!err || !etlt) { /* No error if err is 0 or etlt is 0 */ if ((etlt == 0x09) && (netdev->features & NETIF_F_HW_VLAN_CTAG_RX)) { pkt_info->attributes = XLGMAC_SET_REG_BITS( pkt_info->attributes, RX_PACKET_ATTRIBUTES_VLAN_CTAG_POS, RX_PACKET_ATTRIBUTES_VLAN_CTAG_LEN, 1); pkt_info->vlan_ctag = XLGMAC_GET_REG_BITS_LE(dma_desc->desc0, RX_NORMAL_DESC0_OVT_POS, RX_NORMAL_DESC0_OVT_LEN); netif_dbg(pdata, rx_status, netdev, "vlan-ctag=%#06x\n", pkt_info->vlan_ctag); } } else { if ((etlt == 0x05) || (etlt == 0x06)) pkt_info->attributes = XLGMAC_SET_REG_BITS( pkt_info->attributes, RX_PACKET_ATTRIBUTES_CSUM_DONE_POS, RX_PACKET_ATTRIBUTES_CSUM_DONE_LEN, 0); else pkt_info->errors = XLGMAC_SET_REG_BITS( pkt_info->errors, RX_PACKET_ERRORS_FRAME_POS, RX_PACKET_ERRORS_FRAME_LEN, 1); } XLGMAC_PR("%s - descriptor=%u (cur=%d)\n", channel->name, ring->cur & (ring->dma_desc_count - 1), ring->cur); return 0; } static int xlgmac_enable_int(struct xlgmac_channel *channel, enum xlgmac_int int_id) { unsigned int dma_ch_ier; dma_ch_ier = readl(XLGMAC_DMA_REG(channel, DMA_CH_IER)); switch (int_id) { case XLGMAC_INT_DMA_CH_SR_TI: dma_ch_ier = XLGMAC_SET_REG_BITS( dma_ch_ier, DMA_CH_IER_TIE_POS, DMA_CH_IER_TIE_LEN, 1); break; case XLGMAC_INT_DMA_CH_SR_TPS: dma_ch_ier = XLGMAC_SET_REG_BITS( dma_ch_ier, DMA_CH_IER_TXSE_POS, DMA_CH_IER_TXSE_LEN, 1); break; case XLGMAC_INT_DMA_CH_SR_TBU: dma_ch_ier = XLGMAC_SET_REG_BITS( dma_ch_ier, DMA_CH_IER_TBUE_POS, DMA_CH_IER_TBUE_LEN, 1); break; case XLGMAC_INT_DMA_CH_SR_RI: dma_ch_ier = XLGMAC_SET_REG_BITS( dma_ch_ier, DMA_CH_IER_RIE_POS, DMA_CH_IER_RIE_LEN, 1); break; case XLGMAC_INT_DMA_CH_SR_RBU: dma_ch_ier = XLGMAC_SET_REG_BITS( dma_ch_ier, DMA_CH_IER_RBUE_POS, DMA_CH_IER_RBUE_LEN, 1); break; case XLGMAC_INT_DMA_CH_SR_RPS: dma_ch_ier = XLGMAC_SET_REG_BITS( dma_ch_ier, DMA_CH_IER_RSE_POS, DMA_CH_IER_RSE_LEN, 1); break; case XLGMAC_INT_DMA_CH_SR_TI_RI: dma_ch_ier = XLGMAC_SET_REG_BITS( dma_ch_ier, DMA_CH_IER_TIE_POS, DMA_CH_IER_TIE_LEN, 1); dma_ch_ier = XLGMAC_SET_REG_BITS( dma_ch_ier, DMA_CH_IER_RIE_POS, DMA_CH_IER_RIE_LEN, 1); break; case XLGMAC_INT_DMA_CH_SR_FBE: dma_ch_ier = XLGMAC_SET_REG_BITS( dma_ch_ier, DMA_CH_IER_FBEE_POS, DMA_CH_IER_FBEE_LEN, 1); break; case XLGMAC_INT_DMA_ALL: dma_ch_ier |= channel->saved_ier; break; default: return -1; } writel(dma_ch_ier, XLGMAC_DMA_REG(channel, DMA_CH_IER)); return 0; } static int xlgmac_disable_int(struct xlgmac_channel *channel, enum xlgmac_int int_id) { unsigned int dma_ch_ier; dma_ch_ier = readl(XLGMAC_DMA_REG(channel, DMA_CH_IER)); switch (int_id) { case XLGMAC_INT_DMA_CH_SR_TI: dma_ch_ier = XLGMAC_SET_REG_BITS( dma_ch_ier, DMA_CH_IER_TIE_POS, DMA_CH_IER_TIE_LEN, 0); break; case XLGMAC_INT_DMA_CH_SR_TPS: dma_ch_ier = XLGMAC_SET_REG_BITS( dma_ch_ier, DMA_CH_IER_TXSE_POS, DMA_CH_IER_TXSE_LEN, 0); break; case XLGMAC_INT_DMA_CH_SR_TBU: dma_ch_ier = XLGMAC_SET_REG_BITS( dma_ch_ier, DMA_CH_IER_TBUE_POS, DMA_CH_IER_TBUE_LEN, 0); break; case XLGMAC_INT_DMA_CH_SR_RI: dma_ch_ier = XLGMAC_SET_REG_BITS( dma_ch_ier, DMA_CH_IER_RIE_POS, DMA_CH_IER_RIE_LEN, 0); break; case XLGMAC_INT_DMA_CH_SR_RBU: dma_ch_ier = XLGMAC_SET_REG_BITS( dma_ch_ier, DMA_CH_IER_RBUE_POS, DMA_CH_IER_RBUE_LEN, 0); break; case XLGMAC_INT_DMA_CH_SR_RPS: dma_ch_ier = XLGMAC_SET_REG_BITS( dma_ch_ier, DMA_CH_IER_RSE_POS, DMA_CH_IER_RSE_LEN, 0); break; case XLGMAC_INT_DMA_CH_SR_TI_RI: dma_ch_ier = XLGMAC_SET_REG_BITS( dma_ch_ier, DMA_CH_IER_TIE_POS, DMA_CH_IER_TIE_LEN, 0); dma_ch_ier = XLGMAC_SET_REG_BITS( dma_ch_ier, DMA_CH_IER_RIE_POS, DMA_CH_IER_RIE_LEN, 0); break; case XLGMAC_INT_DMA_CH_SR_FBE: dma_ch_ier = XLGMAC_SET_REG_BITS( dma_ch_ier, DMA_CH_IER_FBEE_POS, DMA_CH_IER_FBEE_LEN, 0); break; case XLGMAC_INT_DMA_ALL: channel->saved_ier = dma_ch_ier & XLGMAC_DMA_INTERRUPT_MASK; dma_ch_ier &= ~XLGMAC_DMA_INTERRUPT_MASK; break; default: return -1; } writel(dma_ch_ier, XLGMAC_DMA_REG(channel, DMA_CH_IER)); return 0; } static int xlgmac_flush_tx_queues(struct xlgmac_pdata *pdata) { unsigned int i, count; u32 regval; for (i = 0; i < pdata->tx_q_count; i++) { regval = readl(XLGMAC_MTL_REG(pdata, i, MTL_Q_TQOMR)); regval = XLGMAC_SET_REG_BITS(regval, MTL_Q_TQOMR_FTQ_POS, MTL_Q_TQOMR_FTQ_LEN, 1); writel(regval, XLGMAC_MTL_REG(pdata, i, MTL_Q_TQOMR)); } /* Poll Until Poll Condition */ for (i = 0; i < pdata->tx_q_count; i++) { count = 2000; regval = readl(XLGMAC_MTL_REG(pdata, i, MTL_Q_TQOMR)); regval = XLGMAC_GET_REG_BITS(regval, MTL_Q_TQOMR_FTQ_POS, MTL_Q_TQOMR_FTQ_LEN); while (--count && regval) usleep_range(500, 600); if (!count) return -EBUSY; } return 0; } static void xlgmac_config_dma_bus(struct xlgmac_pdata *pdata) { u32 regval; regval = readl(pdata->mac_regs + DMA_SBMR); /* Set enhanced addressing mode */ regval = XLGMAC_SET_REG_BITS(regval, DMA_SBMR_EAME_POS, DMA_SBMR_EAME_LEN, 1); /* Set the System Bus mode */ regval = XLGMAC_SET_REG_BITS(regval, DMA_SBMR_UNDEF_POS, DMA_SBMR_UNDEF_LEN, 1); regval = XLGMAC_SET_REG_BITS(regval, DMA_SBMR_BLEN_256_POS, DMA_SBMR_BLEN_256_LEN, 1); writel(regval, pdata->mac_regs + DMA_SBMR); } static int xlgmac_hw_init(struct xlgmac_pdata *pdata) { struct xlgmac_desc_ops *desc_ops = &pdata->desc_ops; int ret; /* Flush Tx queues */ ret = xlgmac_flush_tx_queues(pdata); if (ret) return ret; /* Initialize DMA related features */ xlgmac_config_dma_bus(pdata); xlgmac_config_osp_mode(pdata); xlgmac_config_pblx8(pdata); xlgmac_config_tx_pbl_val(pdata); xlgmac_config_rx_pbl_val(pdata); xlgmac_config_rx_coalesce(pdata); xlgmac_config_tx_coalesce(pdata); xlgmac_config_rx_buffer_size(pdata); xlgmac_config_tso_mode(pdata); xlgmac_config_sph_mode(pdata); xlgmac_config_rss(pdata); desc_ops->tx_desc_init(pdata); desc_ops->rx_desc_init(pdata); xlgmac_enable_dma_interrupts(pdata); /* Initialize MTL related features */ xlgmac_config_mtl_mode(pdata); xlgmac_config_queue_mapping(pdata); xlgmac_config_tsf_mode(pdata, pdata->tx_sf_mode); xlgmac_config_rsf_mode(pdata, pdata->rx_sf_mode); xlgmac_config_tx_threshold(pdata, pdata->tx_threshold); xlgmac_config_rx_threshold(pdata, pdata->rx_threshold); xlgmac_config_tx_fifo_size(pdata); xlgmac_config_rx_fifo_size(pdata); xlgmac_config_flow_control_threshold(pdata); xlgmac_config_rx_fep_enable(pdata); xlgmac_config_rx_fup_enable(pdata); xlgmac_enable_mtl_interrupts(pdata); /* Initialize MAC related features */ xlgmac_config_mac_address(pdata); xlgmac_config_rx_mode(pdata); xlgmac_config_jumbo_enable(pdata); xlgmac_config_flow_control(pdata); xlgmac_config_mac_speed(pdata); xlgmac_config_checksum_offload(pdata); xlgmac_config_vlan_support(pdata); xlgmac_config_mmc(pdata); xlgmac_enable_mac_interrupts(pdata); return 0; } static int xlgmac_hw_exit(struct xlgmac_pdata *pdata) { unsigned int count = 2000; u32 regval; /* Issue a software reset */ regval = readl(pdata->mac_regs + DMA_MR); regval = XLGMAC_SET_REG_BITS(regval, DMA_MR_SWR_POS, DMA_MR_SWR_LEN, 1); writel(regval, pdata->mac_regs + DMA_MR); usleep_range(10, 15); /* Poll Until Poll Condition */ while (--count && XLGMAC_GET_REG_BITS(readl(pdata->mac_regs + DMA_MR), DMA_MR_SWR_POS, DMA_MR_SWR_LEN)) usleep_range(500, 600); if (!count) return -EBUSY; return 0; } void xlgmac_init_hw_ops(struct xlgmac_hw_ops *hw_ops) { hw_ops->init = xlgmac_hw_init; hw_ops->exit = xlgmac_hw_exit; hw_ops->tx_complete = xlgmac_tx_complete; hw_ops->enable_tx = xlgmac_enable_tx; hw_ops->disable_tx = xlgmac_disable_tx; hw_ops->enable_rx = xlgmac_enable_rx; hw_ops->disable_rx = xlgmac_disable_rx; hw_ops->dev_xmit = xlgmac_dev_xmit; hw_ops->dev_read = xlgmac_dev_read; hw_ops->enable_int = xlgmac_enable_int; hw_ops->disable_int = xlgmac_disable_int; hw_ops->set_mac_address = xlgmac_set_mac_address; hw_ops->config_rx_mode = xlgmac_config_rx_mode; hw_ops->enable_rx_csum = xlgmac_enable_rx_csum; hw_ops->disable_rx_csum = xlgmac_disable_rx_csum; /* For MII speed configuration */ hw_ops->set_xlgmii_25000_speed = xlgmac_set_xlgmii_25000_speed; hw_ops->set_xlgmii_40000_speed = xlgmac_set_xlgmii_40000_speed; hw_ops->set_xlgmii_50000_speed = xlgmac_set_xlgmii_50000_speed; hw_ops->set_xlgmii_100000_speed = xlgmac_set_xlgmii_100000_speed; /* For descriptor related operation */ hw_ops->tx_desc_init = xlgmac_tx_desc_init; hw_ops->rx_desc_init = xlgmac_rx_desc_init; hw_ops->tx_desc_reset = xlgmac_tx_desc_reset; hw_ops->rx_desc_reset = xlgmac_rx_desc_reset; hw_ops->is_last_desc = xlgmac_is_last_desc; hw_ops->is_context_desc = xlgmac_is_context_desc; hw_ops->tx_start_xmit = xlgmac_tx_start_xmit; /* For Flow Control */ hw_ops->config_tx_flow_control = xlgmac_config_tx_flow_control; hw_ops->config_rx_flow_control = xlgmac_config_rx_flow_control; /* For Vlan related config */ hw_ops->enable_rx_vlan_stripping = xlgmac_enable_rx_vlan_stripping; hw_ops->disable_rx_vlan_stripping = xlgmac_disable_rx_vlan_stripping; hw_ops->enable_rx_vlan_filtering = xlgmac_enable_rx_vlan_filtering; hw_ops->disable_rx_vlan_filtering = xlgmac_disable_rx_vlan_filtering; hw_ops->update_vlan_hash_table = xlgmac_update_vlan_hash_table; /* For RX coalescing */ hw_ops->config_rx_coalesce = xlgmac_config_rx_coalesce; hw_ops->config_tx_coalesce = xlgmac_config_tx_coalesce; hw_ops->usec_to_riwt = xlgmac_usec_to_riwt; hw_ops->riwt_to_usec = xlgmac_riwt_to_usec; /* For RX and TX threshold config */ hw_ops->config_rx_threshold = xlgmac_config_rx_threshold; hw_ops->config_tx_threshold = xlgmac_config_tx_threshold; /* For RX and TX Store and Forward Mode config */ hw_ops->config_rsf_mode = xlgmac_config_rsf_mode; hw_ops->config_tsf_mode = xlgmac_config_tsf_mode; /* For TX DMA Operating on Second Frame config */ hw_ops->config_osp_mode = xlgmac_config_osp_mode; /* For RX and TX PBL config */ hw_ops->config_rx_pbl_val = xlgmac_config_rx_pbl_val; hw_ops->get_rx_pbl_val = xlgmac_get_rx_pbl_val; hw_ops->config_tx_pbl_val = xlgmac_config_tx_pbl_val; hw_ops->get_tx_pbl_val = xlgmac_get_tx_pbl_val; hw_ops->config_pblx8 = xlgmac_config_pblx8; /* For MMC statistics support */ hw_ops->tx_mmc_int = xlgmac_tx_mmc_int; hw_ops->rx_mmc_int = xlgmac_rx_mmc_int; hw_ops->read_mmc_stats = xlgmac_read_mmc_stats; /* For Receive Side Scaling */ hw_ops->enable_rss = xlgmac_enable_rss; hw_ops->disable_rss = xlgmac_disable_rss; hw_ops->set_rss_hash_key = xlgmac_set_rss_hash_key; hw_ops->set_rss_lookup_table = xlgmac_set_rss_lookup_table; }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1