Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Iyappan Subramanian | 1006 | 99.80% | 3 | 75.00% |
Thomas Gleixner | 2 | 0.20% | 1 | 25.00% |
Total | 1008 | 4 |
// SPDX-License-Identifier: GPL-2.0-or-later /* Applied Micro X-Gene SoC Ethernet Driver * * Copyright (c) 2015, Applied Micro Circuits Corporation * Author: Iyappan Subramanian <isubramanian@apm.com> */ #include "xgene_enet_main.h" #include "xgene_enet_hw.h" #include "xgene_enet_ring2.h" static void xgene_enet_ring_init(struct xgene_enet_desc_ring *ring) { u32 *ring_cfg = ring->state; u64 addr = ring->dma; if (xgene_enet_ring_owner(ring->id) == RING_OWNER_CPU) { ring_cfg[0] |= SET_VAL(X2_INTLINE, ring->id & RING_BUFNUM_MASK); ring_cfg[3] |= SET_BIT(X2_DEQINTEN); } ring_cfg[0] |= SET_VAL(X2_CFGCRID, 2); addr >>= 8; ring_cfg[2] |= QCOHERENT | SET_VAL(RINGADDRL, addr); addr >>= 27; ring_cfg[3] |= SET_VAL(RINGSIZE, ring->cfgsize) | ACCEPTLERR | SET_VAL(RINGADDRH, addr); ring_cfg[4] |= SET_VAL(X2_SELTHRSH, 1); ring_cfg[5] |= SET_BIT(X2_QBASE_AM) | SET_BIT(X2_MSG_AM); } static void xgene_enet_ring_set_type(struct xgene_enet_desc_ring *ring) { u32 *ring_cfg = ring->state; bool is_bufpool; u32 val; is_bufpool = xgene_enet_is_bufpool(ring->id); val = (is_bufpool) ? RING_BUFPOOL : RING_REGULAR; ring_cfg[4] |= SET_VAL(X2_RINGTYPE, val); if (is_bufpool) ring_cfg[3] |= SET_VAL(RINGMODE, BUFPOOL_MODE); } static void xgene_enet_ring_set_recombbuf(struct xgene_enet_desc_ring *ring) { u32 *ring_cfg = ring->state; ring_cfg[3] |= RECOMBBUF; ring_cfg[4] |= SET_VAL(X2_RECOMTIMEOUT, 0x7); } static void xgene_enet_ring_wr32(struct xgene_enet_desc_ring *ring, u32 offset, u32 data) { struct xgene_enet_pdata *pdata = netdev_priv(ring->ndev); iowrite32(data, pdata->ring_csr_addr + offset); } static void xgene_enet_write_ring_state(struct xgene_enet_desc_ring *ring) { struct xgene_enet_pdata *pdata = netdev_priv(ring->ndev); int i; xgene_enet_ring_wr32(ring, CSR_RING_CONFIG, ring->num); for (i = 0; i < pdata->ring_ops->num_ring_config; i++) { xgene_enet_ring_wr32(ring, CSR_RING_WR_BASE + (i * 4), ring->state[i]); } } static void xgene_enet_clr_ring_state(struct xgene_enet_desc_ring *ring) { memset(ring->state, 0, sizeof(ring->state)); xgene_enet_write_ring_state(ring); } static void xgene_enet_set_ring_state(struct xgene_enet_desc_ring *ring) { enum xgene_ring_owner owner; xgene_enet_ring_set_type(ring); owner = xgene_enet_ring_owner(ring->id); if (owner == RING_OWNER_ETH0 || owner == RING_OWNER_ETH1) xgene_enet_ring_set_recombbuf(ring); xgene_enet_ring_init(ring); xgene_enet_write_ring_state(ring); } static void xgene_enet_set_ring_id(struct xgene_enet_desc_ring *ring) { u32 ring_id_val, ring_id_buf; bool is_bufpool; if (xgene_enet_ring_owner(ring->id) == RING_OWNER_CPU) return; is_bufpool = xgene_enet_is_bufpool(ring->id); ring_id_val = ring->id & GENMASK(9, 0); ring_id_val |= OVERWRITE; ring_id_buf = (ring->num << 9) & GENMASK(18, 9); ring_id_buf |= PREFETCH_BUF_EN; if (is_bufpool) ring_id_buf |= IS_BUFFER_POOL; xgene_enet_ring_wr32(ring, CSR_RING_ID, ring_id_val); xgene_enet_ring_wr32(ring, CSR_RING_ID_BUF, ring_id_buf); } static void xgene_enet_clr_desc_ring_id(struct xgene_enet_desc_ring *ring) { u32 ring_id; ring_id = ring->id | OVERWRITE; xgene_enet_ring_wr32(ring, CSR_RING_ID, ring_id); xgene_enet_ring_wr32(ring, CSR_RING_ID_BUF, 0); } static struct xgene_enet_desc_ring *xgene_enet_setup_ring( struct xgene_enet_desc_ring *ring) { bool is_bufpool; u32 addr, i; xgene_enet_clr_ring_state(ring); xgene_enet_set_ring_state(ring); xgene_enet_set_ring_id(ring); ring->slots = xgene_enet_get_numslots(ring->id, ring->size); is_bufpool = xgene_enet_is_bufpool(ring->id); if (is_bufpool || xgene_enet_ring_owner(ring->id) != RING_OWNER_CPU) return ring; addr = CSR_VMID0_INTR_MBOX + (4 * (ring->id & RING_BUFNUM_MASK)); xgene_enet_ring_wr32(ring, addr, ring->irq_mbox_dma >> 10); for (i = 0; i < ring->slots; i++) xgene_enet_mark_desc_slot_empty(&ring->raw_desc[i]); return ring; } static void xgene_enet_clear_ring(struct xgene_enet_desc_ring *ring) { xgene_enet_clr_desc_ring_id(ring); xgene_enet_clr_ring_state(ring); } static void xgene_enet_wr_cmd(struct xgene_enet_desc_ring *ring, int count) { u32 data = 0; if (xgene_enet_ring_owner(ring->id) == RING_OWNER_CPU) { data = SET_VAL(X2_INTLINE, ring->id & RING_BUFNUM_MASK) | INTR_CLEAR; } data |= (count & GENMASK(16, 0)); iowrite32(data, ring->cmd); } static u32 xgene_enet_ring_len(struct xgene_enet_desc_ring *ring) { u32 __iomem *cmd_base = ring->cmd_base; u32 ring_state, num_msgs; ring_state = ioread32(&cmd_base[1]); num_msgs = GET_VAL(X2_NUMMSGSINQ, ring_state); return num_msgs; } static void xgene_enet_setup_coalescing(struct xgene_enet_desc_ring *ring) { u32 data = 0x77777777; xgene_enet_ring_wr32(ring, CSR_PBM_COAL, 0x8e); xgene_enet_ring_wr32(ring, CSR_PBM_CTICK0, data); xgene_enet_ring_wr32(ring, CSR_PBM_CTICK1, data); xgene_enet_ring_wr32(ring, CSR_PBM_CTICK2, data); xgene_enet_ring_wr32(ring, CSR_PBM_CTICK3, data); xgene_enet_ring_wr32(ring, CSR_THRESHOLD0_SET1, 0x08); xgene_enet_ring_wr32(ring, CSR_THRESHOLD1_SET1, 0x10); } struct xgene_ring_ops xgene_ring2_ops = { .num_ring_config = X2_NUM_RING_CONFIG, .num_ring_id_shift = 13, .setup = xgene_enet_setup_ring, .clear = xgene_enet_clear_ring, .wr_cmd = xgene_enet_wr_cmd, .len = xgene_enet_ring_len, .coalesce = xgene_enet_setup_coalescing, };
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1