Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
VSR Burru | 1513 | 100.00% | 3 | 100.00% |
Total | 1513 | 3 |
// SPDX-License-Identifier: GPL-2.0 /* Marvell Octeon EP (EndPoint) Ethernet Driver * * Copyright (C) 2020 Marvell. * */ #include <linux/pci.h> #include <linux/etherdevice.h> #include <linux/vmalloc.h> #include "octep_config.h" #include "octep_main.h" /* Reset various index of Tx queue data structure. */ static void octep_iq_reset_indices(struct octep_iq *iq) { iq->fill_cnt = 0; iq->host_write_index = 0; iq->octep_read_index = 0; iq->flush_index = 0; iq->pkts_processed = 0; iq->pkt_in_done = 0; atomic_set(&iq->instr_pending, 0); } /** * octep_iq_process_completions() - Process Tx queue completions. * * @iq: Octeon Tx queue data structure. * @budget: max number of completions to be processed in one invocation. */ int octep_iq_process_completions(struct octep_iq *iq, u16 budget) { u32 compl_pkts, compl_bytes, compl_sg; struct octep_device *oct = iq->octep_dev; struct octep_tx_buffer *tx_buffer; struct skb_shared_info *shinfo; u32 fi = iq->flush_index; struct sk_buff *skb; u8 frags, i; compl_pkts = 0; compl_sg = 0; compl_bytes = 0; iq->octep_read_index = oct->hw_ops.update_iq_read_idx(iq); while (likely(budget && (fi != iq->octep_read_index))) { tx_buffer = iq->buff_info + fi; skb = tx_buffer->skb; fi++; if (unlikely(fi == iq->max_count)) fi = 0; compl_bytes += skb->len; compl_pkts++; budget--; if (!tx_buffer->gather) { dma_unmap_single(iq->dev, tx_buffer->dma, tx_buffer->skb->len, DMA_TO_DEVICE); dev_kfree_skb_any(skb); continue; } /* Scatter/Gather */ shinfo = skb_shinfo(skb); frags = shinfo->nr_frags; compl_sg++; dma_unmap_single(iq->dev, tx_buffer->sglist[0].dma_ptr[0], tx_buffer->sglist[0].len[0], DMA_TO_DEVICE); i = 1; /* entry 0 is main skb, unmapped above */ while (frags--) { dma_unmap_page(iq->dev, tx_buffer->sglist[i >> 2].dma_ptr[i & 3], tx_buffer->sglist[i >> 2].len[i & 3], DMA_TO_DEVICE); i++; } dev_kfree_skb_any(skb); } iq->pkts_processed += compl_pkts; atomic_sub(compl_pkts, &iq->instr_pending); iq->stats.instr_completed += compl_pkts; iq->stats.bytes_sent += compl_bytes; iq->stats.sgentry_sent += compl_sg; iq->flush_index = fi; netdev_tx_completed_queue(iq->netdev_q, compl_pkts, compl_bytes); if (unlikely(__netif_subqueue_stopped(iq->netdev, iq->q_no)) && ((iq->max_count - atomic_read(&iq->instr_pending)) > OCTEP_WAKE_QUEUE_THRESHOLD)) netif_wake_subqueue(iq->netdev, iq->q_no); return !budget; } /** * octep_iq_free_pending() - Free Tx buffers for pending completions. * * @iq: Octeon Tx queue data structure. */ static void octep_iq_free_pending(struct octep_iq *iq) { struct octep_tx_buffer *tx_buffer; struct skb_shared_info *shinfo; u32 fi = iq->flush_index; struct sk_buff *skb; u8 frags, i; while (fi != iq->host_write_index) { tx_buffer = iq->buff_info + fi; skb = tx_buffer->skb; fi++; if (unlikely(fi == iq->max_count)) fi = 0; if (!tx_buffer->gather) { dma_unmap_single(iq->dev, tx_buffer->dma, tx_buffer->skb->len, DMA_TO_DEVICE); dev_kfree_skb_any(skb); continue; } /* Scatter/Gather */ shinfo = skb_shinfo(skb); frags = shinfo->nr_frags; dma_unmap_single(iq->dev, tx_buffer->sglist[0].dma_ptr[0], tx_buffer->sglist[0].len[0], DMA_TO_DEVICE); i = 1; /* entry 0 is main skb, unmapped above */ while (frags--) { dma_unmap_page(iq->dev, tx_buffer->sglist[i >> 2].dma_ptr[i & 3], tx_buffer->sglist[i >> 2].len[i & 3], DMA_TO_DEVICE); i++; } dev_kfree_skb_any(skb); } atomic_set(&iq->instr_pending, 0); iq->flush_index = fi; netdev_tx_reset_queue(netdev_get_tx_queue(iq->netdev, iq->q_no)); } /** * octep_clean_iqs() - Clean Tx queues to shutdown the device. * * @oct: Octeon device private data structure. * * Free the buffers in Tx queue descriptors pending completion and * reset queue indices */ void octep_clean_iqs(struct octep_device *oct) { int i; for (i = 0; i < oct->num_iqs; i++) { octep_iq_free_pending(oct->iq[i]); octep_iq_reset_indices(oct->iq[i]); } } /** * octep_setup_iq() - Setup a Tx queue. * * @oct: Octeon device private data structure. * @q_no: Tx queue number to be setup. * * Allocate resources for a Tx queue. */ static int octep_setup_iq(struct octep_device *oct, int q_no) { u32 desc_ring_size, buff_info_size, sglist_size; struct octep_iq *iq; int i; iq = vzalloc(sizeof(*iq)); if (!iq) goto iq_alloc_err; oct->iq[q_no] = iq; iq->octep_dev = oct; iq->netdev = oct->netdev; iq->dev = &oct->pdev->dev; iq->q_no = q_no; iq->max_count = CFG_GET_IQ_NUM_DESC(oct->conf); iq->ring_size_mask = iq->max_count - 1; iq->fill_threshold = CFG_GET_IQ_DB_MIN(oct->conf); iq->netdev_q = netdev_get_tx_queue(iq->netdev, q_no); /* Allocate memory for hardware queue descriptors */ desc_ring_size = OCTEP_IQ_DESC_SIZE * CFG_GET_IQ_NUM_DESC(oct->conf); iq->desc_ring = dma_alloc_coherent(iq->dev, desc_ring_size, &iq->desc_ring_dma, GFP_KERNEL); if (unlikely(!iq->desc_ring)) { dev_err(iq->dev, "Failed to allocate DMA memory for IQ-%d\n", q_no); goto desc_dma_alloc_err; } /* Allocate memory for hardware SGLIST descriptors */ sglist_size = OCTEP_SGLIST_SIZE_PER_PKT * CFG_GET_IQ_NUM_DESC(oct->conf); iq->sglist = dma_alloc_coherent(iq->dev, sglist_size, &iq->sglist_dma, GFP_KERNEL); if (unlikely(!iq->sglist)) { dev_err(iq->dev, "Failed to allocate DMA memory for IQ-%d SGLIST\n", q_no); goto sglist_alloc_err; } /* allocate memory to manage Tx packets pending completion */ buff_info_size = OCTEP_IQ_TXBUFF_INFO_SIZE * iq->max_count; iq->buff_info = vzalloc(buff_info_size); if (!iq->buff_info) { dev_err(iq->dev, "Failed to allocate buff info for IQ-%d\n", q_no); goto buff_info_err; } /* Setup sglist addresses in tx_buffer entries */ for (i = 0; i < CFG_GET_IQ_NUM_DESC(oct->conf); i++) { struct octep_tx_buffer *tx_buffer; tx_buffer = &iq->buff_info[i]; tx_buffer->sglist = &iq->sglist[i * OCTEP_SGLIST_ENTRIES_PER_PKT]; tx_buffer->sglist_dma = iq->sglist_dma + (i * OCTEP_SGLIST_SIZE_PER_PKT); } octep_iq_reset_indices(iq); oct->hw_ops.setup_iq_regs(oct, q_no); oct->num_iqs++; return 0; buff_info_err: dma_free_coherent(iq->dev, sglist_size, iq->sglist, iq->sglist_dma); sglist_alloc_err: dma_free_coherent(iq->dev, desc_ring_size, iq->desc_ring, iq->desc_ring_dma); desc_dma_alloc_err: vfree(iq); oct->iq[q_no] = NULL; iq_alloc_err: return -1; } /** * octep_free_iq() - Free Tx queue resources. * * @iq: Octeon Tx queue data structure. * * Free all the resources allocated for a Tx queue. */ static void octep_free_iq(struct octep_iq *iq) { struct octep_device *oct = iq->octep_dev; u64 desc_ring_size, sglist_size; int q_no = iq->q_no; desc_ring_size = OCTEP_IQ_DESC_SIZE * CFG_GET_IQ_NUM_DESC(oct->conf); vfree(iq->buff_info); if (iq->desc_ring) dma_free_coherent(iq->dev, desc_ring_size, iq->desc_ring, iq->desc_ring_dma); sglist_size = OCTEP_SGLIST_SIZE_PER_PKT * CFG_GET_IQ_NUM_DESC(oct->conf); if (iq->sglist) dma_free_coherent(iq->dev, sglist_size, iq->sglist, iq->sglist_dma); vfree(iq); oct->iq[q_no] = NULL; oct->num_iqs--; } /** * octep_setup_iqs() - setup resources for all Tx queues. * * @oct: Octeon device private data structure. */ int octep_setup_iqs(struct octep_device *oct) { int i; oct->num_iqs = 0; for (i = 0; i < CFG_GET_PORTS_ACTIVE_IO_RINGS(oct->conf); i++) { if (octep_setup_iq(oct, i)) { dev_err(&oct->pdev->dev, "Failed to setup IQ(TxQ)-%d.\n", i); goto iq_setup_err; } dev_dbg(&oct->pdev->dev, "Successfully setup IQ(TxQ)-%d.\n", i); } return 0; iq_setup_err: while (i) { i--; octep_free_iq(oct->iq[i]); } return -1; } /** * octep_free_iqs() - Free resources of all Tx queues. * * @oct: Octeon device private data structure. */ void octep_free_iqs(struct octep_device *oct) { int i; for (i = 0; i < CFG_GET_PORTS_ACTIVE_IO_RINGS(oct->conf); i++) { octep_free_iq(oct->iq[i]); dev_dbg(&oct->pdev->dev, "Successfully destroyed IQ(TxQ)-%d.\n", i); } oct->num_iqs = 0; }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1