Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Haijun Liu | 6857 | 99.55% | 6 | 60.00% |
M Chetan Kumar | 11 | 0.16% | 1 | 10.00% |
Yang Yingliang | 10 | 0.15% | 1 | 10.00% |
Ricardo Martinez | 8 | 0.12% | 1 | 10.00% |
Tejun Heo | 2 | 0.03% | 1 | 10.00% |
Total | 6888 | 10 |
// SPDX-License-Identifier: GPL-2.0-only /* * Copyright (c) 2021, MediaTek Inc. * Copyright (c) 2021-2022, Intel Corporation. * * Authors: * Amir Hanania <amir.hanania@intel.com> * Haijun Liu <haijun.liu@mediatek.com> * Moises Veleta <moises.veleta@intel.com> * Ricardo Martinez <ricardo.martinez@linux.intel.com> * Sreehari Kancharla <sreehari.kancharla@intel.com> * * Contributors: * Andy Shevchenko <andriy.shevchenko@linux.intel.com> * Chiranjeevi Rapolu <chiranjeevi.rapolu@intel.com> * Eliot Lee <eliot.lee@intel.com> */ #include <linux/bits.h> #include <linux/bitops.h> #include <linux/delay.h> #include <linux/device.h> #include <linux/dmapool.h> #include <linux/dma-mapping.h> #include <linux/dma-direction.h> #include <linux/gfp.h> #include <linux/io.h> #include <linux/io-64-nonatomic-lo-hi.h> #include <linux/iopoll.h> #include <linux/irqreturn.h> #include <linux/kernel.h> #include <linux/kthread.h> #include <linux/list.h> #include <linux/netdevice.h> #include <linux/pci.h> #include <linux/pm_runtime.h> #include <linux/sched.h> #include <linux/skbuff.h> #include <linux/slab.h> #include <linux/spinlock.h> #include <linux/types.h> #include <linux/wait.h> #include <linux/workqueue.h> #include "t7xx_cldma.h" #include "t7xx_hif_cldma.h" #include "t7xx_mhccif.h" #include "t7xx_pci.h" #include "t7xx_pcie_mac.h" #include "t7xx_port_proxy.h" #include "t7xx_reg.h" #include "t7xx_state_monitor.h" #define MAX_TX_BUDGET 16 #define MAX_RX_BUDGET 16 #define CHECK_Q_STOP_TIMEOUT_US 1000000 #define CHECK_Q_STOP_STEP_US 10000 #define CLDMA_JUMBO_BUFF_SZ (63 * 1024 + sizeof(struct ccci_header)) static void md_cd_queue_struct_reset(struct cldma_queue *queue, struct cldma_ctrl *md_ctrl, enum mtk_txrx tx_rx, unsigned int index) { queue->dir = tx_rx; queue->index = index; queue->md_ctrl = md_ctrl; queue->tr_ring = NULL; queue->tr_done = NULL; queue->tx_next = NULL; } static void md_cd_queue_struct_init(struct cldma_queue *queue, struct cldma_ctrl *md_ctrl, enum mtk_txrx tx_rx, unsigned int index) { md_cd_queue_struct_reset(queue, md_ctrl, tx_rx, index); init_waitqueue_head(&queue->req_wq); spin_lock_init(&queue->ring_lock); } static void t7xx_cldma_gpd_set_data_ptr(struct cldma_gpd *gpd, dma_addr_t data_ptr) { gpd->data_buff_bd_ptr_h = cpu_to_le32(upper_32_bits(data_ptr)); gpd->data_buff_bd_ptr_l = cpu_to_le32(lower_32_bits(data_ptr)); } static void t7xx_cldma_gpd_set_next_ptr(struct cldma_gpd *gpd, dma_addr_t next_ptr) { gpd->next_gpd_ptr_h = cpu_to_le32(upper_32_bits(next_ptr)); gpd->next_gpd_ptr_l = cpu_to_le32(lower_32_bits(next_ptr)); } static int t7xx_cldma_alloc_and_map_skb(struct cldma_ctrl *md_ctrl, struct cldma_request *req, size_t size, gfp_t gfp_mask) { req->skb = __dev_alloc_skb(size, gfp_mask); if (!req->skb) return -ENOMEM; req->mapped_buff = dma_map_single(md_ctrl->dev, req->skb->data, size, DMA_FROM_DEVICE); if (dma_mapping_error(md_ctrl->dev, req->mapped_buff)) { dev_kfree_skb_any(req->skb); req->skb = NULL; req->mapped_buff = 0; dev_err(md_ctrl->dev, "DMA mapping failed\n"); return -ENOMEM; } return 0; } static int t7xx_cldma_gpd_rx_from_q(struct cldma_queue *queue, int budget, bool *over_budget) { struct cldma_ctrl *md_ctrl = queue->md_ctrl; unsigned int hwo_polling_count = 0; struct t7xx_cldma_hw *hw_info; bool rx_not_done = true; unsigned long flags; int count = 0; hw_info = &md_ctrl->hw_info; do { struct cldma_request *req; struct cldma_gpd *gpd; struct sk_buff *skb; int ret; req = queue->tr_done; if (!req) return -ENODATA; gpd = req->gpd; if ((gpd->flags & GPD_FLAGS_HWO) || !req->skb) { dma_addr_t gpd_addr; if (!pci_device_is_present(to_pci_dev(md_ctrl->dev))) { dev_err(md_ctrl->dev, "PCIe Link disconnected\n"); return -ENODEV; } gpd_addr = ioread64(hw_info->ap_pdn_base + REG_CLDMA_DL_CURRENT_ADDRL_0 + queue->index * sizeof(u64)); if (req->gpd_addr == gpd_addr || hwo_polling_count++ >= 100) return 0; udelay(1); continue; } hwo_polling_count = 0; skb = req->skb; if (req->mapped_buff) { dma_unmap_single(md_ctrl->dev, req->mapped_buff, queue->tr_ring->pkt_size, DMA_FROM_DEVICE); req->mapped_buff = 0; } skb->len = 0; skb_reset_tail_pointer(skb); skb_put(skb, le16_to_cpu(gpd->data_buff_len)); ret = md_ctrl->recv_skb(queue, skb); /* Break processing, will try again later */ if (ret < 0) return ret; req->skb = NULL; t7xx_cldma_gpd_set_data_ptr(gpd, 0); spin_lock_irqsave(&queue->ring_lock, flags); queue->tr_done = list_next_entry_circular(req, &queue->tr_ring->gpd_ring, entry); spin_unlock_irqrestore(&queue->ring_lock, flags); req = queue->rx_refill; ret = t7xx_cldma_alloc_and_map_skb(md_ctrl, req, queue->tr_ring->pkt_size, GFP_KERNEL); if (ret) return ret; gpd = req->gpd; t7xx_cldma_gpd_set_data_ptr(gpd, req->mapped_buff); gpd->data_buff_len = 0; gpd->flags = GPD_FLAGS_IOC | GPD_FLAGS_HWO; spin_lock_irqsave(&queue->ring_lock, flags); queue->rx_refill = list_next_entry_circular(req, &queue->tr_ring->gpd_ring, entry); spin_unlock_irqrestore(&queue->ring_lock, flags); rx_not_done = ++count < budget || !need_resched(); } while (rx_not_done); *over_budget = true; return 0; } static int t7xx_cldma_gpd_rx_collect(struct cldma_queue *queue, int budget) { struct cldma_ctrl *md_ctrl = queue->md_ctrl; struct t7xx_cldma_hw *hw_info; unsigned int pending_rx_int; bool over_budget = false; unsigned long flags; int ret; hw_info = &md_ctrl->hw_info; do { ret = t7xx_cldma_gpd_rx_from_q(queue, budget, &over_budget); if (ret == -ENODATA) return 0; else if (ret) return ret; pending_rx_int = 0; spin_lock_irqsave(&md_ctrl->cldma_lock, flags); if (md_ctrl->rxq_active & BIT(queue->index)) { if (!t7xx_cldma_hw_queue_status(hw_info, queue->index, MTK_RX)) t7xx_cldma_hw_resume_queue(hw_info, queue->index, MTK_RX); pending_rx_int = t7xx_cldma_hw_int_status(hw_info, BIT(queue->index), MTK_RX); if (pending_rx_int) { t7xx_cldma_hw_rx_done(hw_info, pending_rx_int); if (over_budget) { spin_unlock_irqrestore(&md_ctrl->cldma_lock, flags); return -EAGAIN; } } } spin_unlock_irqrestore(&md_ctrl->cldma_lock, flags); } while (pending_rx_int); return 0; } static void t7xx_cldma_rx_done(struct work_struct *work) { struct cldma_queue *queue = container_of(work, struct cldma_queue, cldma_work); struct cldma_ctrl *md_ctrl = queue->md_ctrl; int value; value = t7xx_cldma_gpd_rx_collect(queue, queue->budget); if (value && md_ctrl->rxq_active & BIT(queue->index)) { queue_work(queue->worker, &queue->cldma_work); return; } t7xx_cldma_clear_ip_busy(&md_ctrl->hw_info); t7xx_cldma_hw_irq_en_txrx(&md_ctrl->hw_info, queue->index, MTK_RX); t7xx_cldma_hw_irq_en_eq(&md_ctrl->hw_info, queue->index, MTK_RX); pm_runtime_mark_last_busy(md_ctrl->dev); pm_runtime_put_autosuspend(md_ctrl->dev); } static int t7xx_cldma_gpd_tx_collect(struct cldma_queue *queue) { struct cldma_ctrl *md_ctrl = queue->md_ctrl; unsigned int dma_len, count = 0; struct cldma_request *req; struct cldma_gpd *gpd; unsigned long flags; dma_addr_t dma_free; struct sk_buff *skb; while (!kthread_should_stop()) { spin_lock_irqsave(&queue->ring_lock, flags); req = queue->tr_done; if (!req) { spin_unlock_irqrestore(&queue->ring_lock, flags); break; } gpd = req->gpd; if ((gpd->flags & GPD_FLAGS_HWO) || !req->skb) { spin_unlock_irqrestore(&queue->ring_lock, flags); break; } queue->budget++; dma_free = req->mapped_buff; dma_len = le16_to_cpu(gpd->data_buff_len); skb = req->skb; req->skb = NULL; queue->tr_done = list_next_entry_circular(req, &queue->tr_ring->gpd_ring, entry); spin_unlock_irqrestore(&queue->ring_lock, flags); count++; dma_unmap_single(md_ctrl->dev, dma_free, dma_len, DMA_TO_DEVICE); dev_kfree_skb_any(skb); } if (count) wake_up_nr(&queue->req_wq, count); return count; } static void t7xx_cldma_txq_empty_hndl(struct cldma_queue *queue) { struct cldma_ctrl *md_ctrl = queue->md_ctrl; struct cldma_request *req; dma_addr_t ul_curr_addr; unsigned long flags; bool pending_gpd; if (!(md_ctrl->txq_active & BIT(queue->index))) return; spin_lock_irqsave(&queue->ring_lock, flags); req = list_prev_entry_circular(queue->tx_next, &queue->tr_ring->gpd_ring, entry); spin_unlock_irqrestore(&queue->ring_lock, flags); pending_gpd = (req->gpd->flags & GPD_FLAGS_HWO) && req->skb; spin_lock_irqsave(&md_ctrl->cldma_lock, flags); if (pending_gpd) { struct t7xx_cldma_hw *hw_info = &md_ctrl->hw_info; /* Check current processing TGPD, 64-bit address is in a table by Q index */ ul_curr_addr = ioread64(hw_info->ap_pdn_base + REG_CLDMA_UL_CURRENT_ADDRL_0 + queue->index * sizeof(u64)); if (req->gpd_addr != ul_curr_addr) { spin_unlock_irqrestore(&md_ctrl->cldma_lock, flags); dev_err(md_ctrl->dev, "CLDMA%d queue %d is not empty\n", md_ctrl->hif_id, queue->index); return; } t7xx_cldma_hw_resume_queue(hw_info, queue->index, MTK_TX); } spin_unlock_irqrestore(&md_ctrl->cldma_lock, flags); } static void t7xx_cldma_tx_done(struct work_struct *work) { struct cldma_queue *queue = container_of(work, struct cldma_queue, cldma_work); struct cldma_ctrl *md_ctrl = queue->md_ctrl; struct t7xx_cldma_hw *hw_info; unsigned int l2_tx_int; unsigned long flags; hw_info = &md_ctrl->hw_info; t7xx_cldma_gpd_tx_collect(queue); l2_tx_int = t7xx_cldma_hw_int_status(hw_info, BIT(queue->index) | EQ_STA_BIT(queue->index), MTK_TX); if (l2_tx_int & EQ_STA_BIT(queue->index)) { t7xx_cldma_hw_tx_done(hw_info, EQ_STA_BIT(queue->index)); t7xx_cldma_txq_empty_hndl(queue); } if (l2_tx_int & BIT(queue->index)) { t7xx_cldma_hw_tx_done(hw_info, BIT(queue->index)); queue_work(queue->worker, &queue->cldma_work); return; } spin_lock_irqsave(&md_ctrl->cldma_lock, flags); if (md_ctrl->txq_active & BIT(queue->index)) { t7xx_cldma_clear_ip_busy(hw_info); t7xx_cldma_hw_irq_en_eq(hw_info, queue->index, MTK_TX); t7xx_cldma_hw_irq_en_txrx(hw_info, queue->index, MTK_TX); } spin_unlock_irqrestore(&md_ctrl->cldma_lock, flags); pm_runtime_mark_last_busy(md_ctrl->dev); pm_runtime_put_autosuspend(md_ctrl->dev); } static void t7xx_cldma_ring_free(struct cldma_ctrl *md_ctrl, struct cldma_ring *ring, enum dma_data_direction tx_rx) { struct cldma_request *req_cur, *req_next; list_for_each_entry_safe(req_cur, req_next, &ring->gpd_ring, entry) { if (req_cur->mapped_buff && req_cur->skb) { dma_unmap_single(md_ctrl->dev, req_cur->mapped_buff, ring->pkt_size, tx_rx); req_cur->mapped_buff = 0; } dev_kfree_skb_any(req_cur->skb); if (req_cur->gpd) dma_pool_free(md_ctrl->gpd_dmapool, req_cur->gpd, req_cur->gpd_addr); list_del(&req_cur->entry); kfree(req_cur); } } static struct cldma_request *t7xx_alloc_rx_request(struct cldma_ctrl *md_ctrl, size_t pkt_size) { struct cldma_request *req; int val; req = kzalloc(sizeof(*req), GFP_KERNEL); if (!req) return NULL; req->gpd = dma_pool_zalloc(md_ctrl->gpd_dmapool, GFP_KERNEL, &req->gpd_addr); if (!req->gpd) goto err_free_req; val = t7xx_cldma_alloc_and_map_skb(md_ctrl, req, pkt_size, GFP_KERNEL); if (val) goto err_free_pool; return req; err_free_pool: dma_pool_free(md_ctrl->gpd_dmapool, req->gpd, req->gpd_addr); err_free_req: kfree(req); return NULL; } static int t7xx_cldma_rx_ring_init(struct cldma_ctrl *md_ctrl, struct cldma_ring *ring) { struct cldma_request *req; struct cldma_gpd *gpd; int i; INIT_LIST_HEAD(&ring->gpd_ring); ring->length = MAX_RX_BUDGET; for (i = 0; i < ring->length; i++) { req = t7xx_alloc_rx_request(md_ctrl, ring->pkt_size); if (!req) { t7xx_cldma_ring_free(md_ctrl, ring, DMA_FROM_DEVICE); return -ENOMEM; } gpd = req->gpd; t7xx_cldma_gpd_set_data_ptr(gpd, req->mapped_buff); gpd->rx_data_allow_len = cpu_to_le16(ring->pkt_size); gpd->flags = GPD_FLAGS_IOC | GPD_FLAGS_HWO; INIT_LIST_HEAD(&req->entry); list_add_tail(&req->entry, &ring->gpd_ring); } /* Link previous GPD to next GPD, circular */ list_for_each_entry(req, &ring->gpd_ring, entry) { t7xx_cldma_gpd_set_next_ptr(gpd, req->gpd_addr); gpd = req->gpd; } return 0; } static struct cldma_request *t7xx_alloc_tx_request(struct cldma_ctrl *md_ctrl) { struct cldma_request *req; req = kzalloc(sizeof(*req), GFP_KERNEL); if (!req) return NULL; req->gpd = dma_pool_zalloc(md_ctrl->gpd_dmapool, GFP_KERNEL, &req->gpd_addr); if (!req->gpd) { kfree(req); return NULL; } return req; } static int t7xx_cldma_tx_ring_init(struct cldma_ctrl *md_ctrl, struct cldma_ring *ring) { struct cldma_request *req; struct cldma_gpd *gpd; int i; INIT_LIST_HEAD(&ring->gpd_ring); ring->length = MAX_TX_BUDGET; for (i = 0; i < ring->length; i++) { req = t7xx_alloc_tx_request(md_ctrl); if (!req) { t7xx_cldma_ring_free(md_ctrl, ring, DMA_TO_DEVICE); return -ENOMEM; } gpd = req->gpd; gpd->flags = GPD_FLAGS_IOC; INIT_LIST_HEAD(&req->entry); list_add_tail(&req->entry, &ring->gpd_ring); } /* Link previous GPD to next GPD, circular */ list_for_each_entry(req, &ring->gpd_ring, entry) { t7xx_cldma_gpd_set_next_ptr(gpd, req->gpd_addr); gpd = req->gpd; } return 0; } /** * t7xx_cldma_q_reset() - Reset CLDMA request pointers to their initial values. * @queue: Pointer to the queue structure. * * Called with ring_lock (unless called during initialization phase) */ static void t7xx_cldma_q_reset(struct cldma_queue *queue) { struct cldma_request *req; req = list_first_entry(&queue->tr_ring->gpd_ring, struct cldma_request, entry); queue->tr_done = req; queue->budget = queue->tr_ring->length; if (queue->dir == MTK_TX) queue->tx_next = req; else queue->rx_refill = req; } static void t7xx_cldma_rxq_init(struct cldma_queue *queue) { struct cldma_ctrl *md_ctrl = queue->md_ctrl; queue->dir = MTK_RX; queue->tr_ring = &md_ctrl->rx_ring[queue->index]; t7xx_cldma_q_reset(queue); } static void t7xx_cldma_txq_init(struct cldma_queue *queue) { struct cldma_ctrl *md_ctrl = queue->md_ctrl; queue->dir = MTK_TX; queue->tr_ring = &md_ctrl->tx_ring[queue->index]; t7xx_cldma_q_reset(queue); } static void t7xx_cldma_enable_irq(struct cldma_ctrl *md_ctrl) { t7xx_pcie_mac_set_int(md_ctrl->t7xx_dev, md_ctrl->hw_info.phy_interrupt_id); } static void t7xx_cldma_disable_irq(struct cldma_ctrl *md_ctrl) { t7xx_pcie_mac_clear_int(md_ctrl->t7xx_dev, md_ctrl->hw_info.phy_interrupt_id); } static void t7xx_cldma_irq_work_cb(struct cldma_ctrl *md_ctrl) { unsigned long l2_tx_int_msk, l2_rx_int_msk, l2_tx_int, l2_rx_int, val; struct t7xx_cldma_hw *hw_info = &md_ctrl->hw_info; int i; /* L2 raw interrupt status */ l2_tx_int = ioread32(hw_info->ap_pdn_base + REG_CLDMA_L2TISAR0); l2_rx_int = ioread32(hw_info->ap_pdn_base + REG_CLDMA_L2RISAR0); l2_tx_int_msk = ioread32(hw_info->ap_pdn_base + REG_CLDMA_L2TIMR0); l2_rx_int_msk = ioread32(hw_info->ap_ao_base + REG_CLDMA_L2RIMR0); l2_tx_int &= ~l2_tx_int_msk; l2_rx_int &= ~l2_rx_int_msk; if (l2_tx_int) { if (l2_tx_int & (TQ_ERR_INT_BITMASK | TQ_ACTIVE_START_ERR_INT_BITMASK)) { /* Read and clear L3 TX interrupt status */ val = ioread32(hw_info->ap_pdn_base + REG_CLDMA_L3TISAR0); iowrite32(val, hw_info->ap_pdn_base + REG_CLDMA_L3TISAR0); val = ioread32(hw_info->ap_pdn_base + REG_CLDMA_L3TISAR1); iowrite32(val, hw_info->ap_pdn_base + REG_CLDMA_L3TISAR1); } t7xx_cldma_hw_tx_done(hw_info, l2_tx_int); if (l2_tx_int & (TXRX_STATUS_BITMASK | EMPTY_STATUS_BITMASK)) { for_each_set_bit(i, &l2_tx_int, L2_INT_BIT_COUNT) { if (i < CLDMA_TXQ_NUM) { pm_runtime_get(md_ctrl->dev); t7xx_cldma_hw_irq_dis_eq(hw_info, i, MTK_TX); t7xx_cldma_hw_irq_dis_txrx(hw_info, i, MTK_TX); queue_work(md_ctrl->txq[i].worker, &md_ctrl->txq[i].cldma_work); } else { t7xx_cldma_txq_empty_hndl(&md_ctrl->txq[i - CLDMA_TXQ_NUM]); } } } } if (l2_rx_int) { if (l2_rx_int & (RQ_ERR_INT_BITMASK | RQ_ACTIVE_START_ERR_INT_BITMASK)) { /* Read and clear L3 RX interrupt status */ val = ioread32(hw_info->ap_pdn_base + REG_CLDMA_L3RISAR0); iowrite32(val, hw_info->ap_pdn_base + REG_CLDMA_L3RISAR0); val = ioread32(hw_info->ap_pdn_base + REG_CLDMA_L3RISAR1); iowrite32(val, hw_info->ap_pdn_base + REG_CLDMA_L3RISAR1); } t7xx_cldma_hw_rx_done(hw_info, l2_rx_int); if (l2_rx_int & (TXRX_STATUS_BITMASK | EMPTY_STATUS_BITMASK)) { l2_rx_int |= l2_rx_int >> CLDMA_RXQ_NUM; for_each_set_bit(i, &l2_rx_int, CLDMA_RXQ_NUM) { pm_runtime_get(md_ctrl->dev); t7xx_cldma_hw_irq_dis_eq(hw_info, i, MTK_RX); t7xx_cldma_hw_irq_dis_txrx(hw_info, i, MTK_RX); queue_work(md_ctrl->rxq[i].worker, &md_ctrl->rxq[i].cldma_work); } } } } static bool t7xx_cldma_qs_are_active(struct cldma_ctrl *md_ctrl) { struct t7xx_cldma_hw *hw_info = &md_ctrl->hw_info; unsigned int tx_active; unsigned int rx_active; if (!pci_device_is_present(to_pci_dev(md_ctrl->dev))) return false; tx_active = t7xx_cldma_hw_queue_status(hw_info, CLDMA_ALL_Q, MTK_TX); rx_active = t7xx_cldma_hw_queue_status(hw_info, CLDMA_ALL_Q, MTK_RX); return tx_active || rx_active; } /** * t7xx_cldma_stop() - Stop CLDMA. * @md_ctrl: CLDMA context structure. * * Stop TX and RX queues. Disable L1 and L2 interrupts. * Clear status registers. * * Return: * * 0 - Success. * * -ERROR - Error code from polling cldma_queues_active. */ int t7xx_cldma_stop(struct cldma_ctrl *md_ctrl) { struct t7xx_cldma_hw *hw_info = &md_ctrl->hw_info; bool active; int i, ret; md_ctrl->rxq_active = 0; t7xx_cldma_hw_stop_all_qs(hw_info, MTK_RX); md_ctrl->txq_active = 0; t7xx_cldma_hw_stop_all_qs(hw_info, MTK_TX); md_ctrl->txq_started = 0; t7xx_cldma_disable_irq(md_ctrl); t7xx_cldma_hw_stop(hw_info, MTK_RX); t7xx_cldma_hw_stop(hw_info, MTK_TX); t7xx_cldma_hw_tx_done(hw_info, CLDMA_L2TISAR0_ALL_INT_MASK); t7xx_cldma_hw_rx_done(hw_info, CLDMA_L2RISAR0_ALL_INT_MASK); if (md_ctrl->is_late_init) { for (i = 0; i < CLDMA_TXQ_NUM; i++) flush_work(&md_ctrl->txq[i].cldma_work); for (i = 0; i < CLDMA_RXQ_NUM; i++) flush_work(&md_ctrl->rxq[i].cldma_work); } ret = read_poll_timeout(t7xx_cldma_qs_are_active, active, !active, CHECK_Q_STOP_STEP_US, CHECK_Q_STOP_TIMEOUT_US, true, md_ctrl); if (ret) dev_err(md_ctrl->dev, "Could not stop CLDMA%d queues", md_ctrl->hif_id); return ret; } static void t7xx_cldma_late_release(struct cldma_ctrl *md_ctrl) { int i; if (!md_ctrl->is_late_init) return; for (i = 0; i < CLDMA_TXQ_NUM; i++) t7xx_cldma_ring_free(md_ctrl, &md_ctrl->tx_ring[i], DMA_TO_DEVICE); for (i = 0; i < CLDMA_RXQ_NUM; i++) t7xx_cldma_ring_free(md_ctrl, &md_ctrl->rx_ring[i], DMA_FROM_DEVICE); dma_pool_destroy(md_ctrl->gpd_dmapool); md_ctrl->gpd_dmapool = NULL; md_ctrl->is_late_init = false; } void t7xx_cldma_reset(struct cldma_ctrl *md_ctrl) { unsigned long flags; int i; spin_lock_irqsave(&md_ctrl->cldma_lock, flags); md_ctrl->txq_active = 0; md_ctrl->rxq_active = 0; t7xx_cldma_disable_irq(md_ctrl); spin_unlock_irqrestore(&md_ctrl->cldma_lock, flags); for (i = 0; i < CLDMA_TXQ_NUM; i++) { cancel_work_sync(&md_ctrl->txq[i].cldma_work); spin_lock_irqsave(&md_ctrl->cldma_lock, flags); md_cd_queue_struct_reset(&md_ctrl->txq[i], md_ctrl, MTK_TX, i); spin_unlock_irqrestore(&md_ctrl->cldma_lock, flags); } for (i = 0; i < CLDMA_RXQ_NUM; i++) { cancel_work_sync(&md_ctrl->rxq[i].cldma_work); spin_lock_irqsave(&md_ctrl->cldma_lock, flags); md_cd_queue_struct_reset(&md_ctrl->rxq[i], md_ctrl, MTK_RX, i); spin_unlock_irqrestore(&md_ctrl->cldma_lock, flags); } t7xx_cldma_late_release(md_ctrl); } /** * t7xx_cldma_start() - Start CLDMA. * @md_ctrl: CLDMA context structure. * * Set TX/RX start address. * Start all RX queues and enable L2 interrupt. */ void t7xx_cldma_start(struct cldma_ctrl *md_ctrl) { unsigned long flags; spin_lock_irqsave(&md_ctrl->cldma_lock, flags); if (md_ctrl->is_late_init) { struct t7xx_cldma_hw *hw_info = &md_ctrl->hw_info; int i; t7xx_cldma_enable_irq(md_ctrl); for (i = 0; i < CLDMA_TXQ_NUM; i++) { if (md_ctrl->txq[i].tr_done) t7xx_cldma_hw_set_start_addr(hw_info, i, md_ctrl->txq[i].tr_done->gpd_addr, MTK_TX); } for (i = 0; i < CLDMA_RXQ_NUM; i++) { if (md_ctrl->rxq[i].tr_done) t7xx_cldma_hw_set_start_addr(hw_info, i, md_ctrl->rxq[i].tr_done->gpd_addr, MTK_RX); } /* Enable L2 interrupt */ t7xx_cldma_hw_start_queue(hw_info, CLDMA_ALL_Q, MTK_RX); t7xx_cldma_hw_start(hw_info); md_ctrl->txq_started = 0; md_ctrl->txq_active |= TXRX_STATUS_BITMASK; md_ctrl->rxq_active |= TXRX_STATUS_BITMASK; } spin_unlock_irqrestore(&md_ctrl->cldma_lock, flags); } static void t7xx_cldma_clear_txq(struct cldma_ctrl *md_ctrl, int qnum) { struct cldma_queue *txq = &md_ctrl->txq[qnum]; struct cldma_request *req; struct cldma_gpd *gpd; unsigned long flags; spin_lock_irqsave(&txq->ring_lock, flags); t7xx_cldma_q_reset(txq); list_for_each_entry(req, &txq->tr_ring->gpd_ring, entry) { gpd = req->gpd; gpd->flags &= ~GPD_FLAGS_HWO; t7xx_cldma_gpd_set_data_ptr(gpd, 0); gpd->data_buff_len = 0; dev_kfree_skb_any(req->skb); req->skb = NULL; } spin_unlock_irqrestore(&txq->ring_lock, flags); } static int t7xx_cldma_clear_rxq(struct cldma_ctrl *md_ctrl, int qnum) { struct cldma_queue *rxq = &md_ctrl->rxq[qnum]; struct cldma_request *req; struct cldma_gpd *gpd; unsigned long flags; int ret = 0; spin_lock_irqsave(&rxq->ring_lock, flags); t7xx_cldma_q_reset(rxq); list_for_each_entry(req, &rxq->tr_ring->gpd_ring, entry) { gpd = req->gpd; gpd->flags = GPD_FLAGS_IOC | GPD_FLAGS_HWO; gpd->data_buff_len = 0; if (req->skb) { req->skb->len = 0; skb_reset_tail_pointer(req->skb); } } list_for_each_entry(req, &rxq->tr_ring->gpd_ring, entry) { if (req->skb) continue; ret = t7xx_cldma_alloc_and_map_skb(md_ctrl, req, rxq->tr_ring->pkt_size, GFP_ATOMIC); if (ret) break; t7xx_cldma_gpd_set_data_ptr(req->gpd, req->mapped_buff); } spin_unlock_irqrestore(&rxq->ring_lock, flags); return ret; } void t7xx_cldma_clear_all_qs(struct cldma_ctrl *md_ctrl, enum mtk_txrx tx_rx) { int i; if (tx_rx == MTK_TX) { for (i = 0; i < CLDMA_TXQ_NUM; i++) t7xx_cldma_clear_txq(md_ctrl, i); } else { for (i = 0; i < CLDMA_RXQ_NUM; i++) t7xx_cldma_clear_rxq(md_ctrl, i); } } void t7xx_cldma_stop_all_qs(struct cldma_ctrl *md_ctrl, enum mtk_txrx tx_rx) { struct t7xx_cldma_hw *hw_info = &md_ctrl->hw_info; unsigned long flags; spin_lock_irqsave(&md_ctrl->cldma_lock, flags); t7xx_cldma_hw_irq_dis_eq(hw_info, CLDMA_ALL_Q, tx_rx); t7xx_cldma_hw_irq_dis_txrx(hw_info, CLDMA_ALL_Q, tx_rx); if (tx_rx == MTK_RX) md_ctrl->rxq_active &= ~TXRX_STATUS_BITMASK; else md_ctrl->txq_active &= ~TXRX_STATUS_BITMASK; t7xx_cldma_hw_stop_all_qs(hw_info, tx_rx); spin_unlock_irqrestore(&md_ctrl->cldma_lock, flags); } static int t7xx_cldma_gpd_handle_tx_request(struct cldma_queue *queue, struct cldma_request *tx_req, struct sk_buff *skb) { struct cldma_ctrl *md_ctrl = queue->md_ctrl; struct cldma_gpd *gpd = tx_req->gpd; unsigned long flags; /* Update GPD */ tx_req->mapped_buff = dma_map_single(md_ctrl->dev, skb->data, skb->len, DMA_TO_DEVICE); if (dma_mapping_error(md_ctrl->dev, tx_req->mapped_buff)) { dev_err(md_ctrl->dev, "DMA mapping failed\n"); return -ENOMEM; } t7xx_cldma_gpd_set_data_ptr(gpd, tx_req->mapped_buff); gpd->data_buff_len = cpu_to_le16(skb->len); /* This lock must cover TGPD setting, as even without a resume operation, * CLDMA can send next HWO=1 if last TGPD just finished. */ spin_lock_irqsave(&md_ctrl->cldma_lock, flags); if (md_ctrl->txq_active & BIT(queue->index)) gpd->flags |= GPD_FLAGS_HWO; spin_unlock_irqrestore(&md_ctrl->cldma_lock, flags); tx_req->skb = skb; return 0; } /* Called with cldma_lock */ static void t7xx_cldma_hw_start_send(struct cldma_ctrl *md_ctrl, int qno, struct cldma_request *prev_req) { struct t7xx_cldma_hw *hw_info = &md_ctrl->hw_info; /* Check whether the device was powered off (CLDMA start address is not set) */ if (!t7xx_cldma_tx_addr_is_set(hw_info, qno)) { t7xx_cldma_hw_init(hw_info); t7xx_cldma_hw_set_start_addr(hw_info, qno, prev_req->gpd_addr, MTK_TX); md_ctrl->txq_started &= ~BIT(qno); } if (!t7xx_cldma_hw_queue_status(hw_info, qno, MTK_TX)) { if (md_ctrl->txq_started & BIT(qno)) t7xx_cldma_hw_resume_queue(hw_info, qno, MTK_TX); else t7xx_cldma_hw_start_queue(hw_info, qno, MTK_TX); md_ctrl->txq_started |= BIT(qno); } } /** * t7xx_cldma_set_recv_skb() - Set the callback to handle RX packets. * @md_ctrl: CLDMA context structure. * @recv_skb: Receiving skb callback. */ void t7xx_cldma_set_recv_skb(struct cldma_ctrl *md_ctrl, int (*recv_skb)(struct cldma_queue *queue, struct sk_buff *skb)) { md_ctrl->recv_skb = recv_skb; } /** * t7xx_cldma_send_skb() - Send control data to modem. * @md_ctrl: CLDMA context structure. * @qno: Queue number. * @skb: Socket buffer. * * Return: * * 0 - Success. * * -ENOMEM - Allocation failure. * * -EINVAL - Invalid queue request. * * -EIO - Queue is not active. * * -ETIMEDOUT - Timeout waiting for the device to wake up. */ int t7xx_cldma_send_skb(struct cldma_ctrl *md_ctrl, int qno, struct sk_buff *skb) { struct cldma_request *tx_req; struct cldma_queue *queue; unsigned long flags; int ret; if (qno >= CLDMA_TXQ_NUM) return -EINVAL; ret = pm_runtime_resume_and_get(md_ctrl->dev); if (ret < 0 && ret != -EACCES) return ret; t7xx_pci_disable_sleep(md_ctrl->t7xx_dev); queue = &md_ctrl->txq[qno]; spin_lock_irqsave(&md_ctrl->cldma_lock, flags); if (!(md_ctrl->txq_active & BIT(qno))) { ret = -EIO; spin_unlock_irqrestore(&md_ctrl->cldma_lock, flags); goto allow_sleep; } spin_unlock_irqrestore(&md_ctrl->cldma_lock, flags); do { spin_lock_irqsave(&queue->ring_lock, flags); tx_req = queue->tx_next; if (queue->budget > 0 && !tx_req->skb) { struct list_head *gpd_ring = &queue->tr_ring->gpd_ring; queue->budget--; t7xx_cldma_gpd_handle_tx_request(queue, tx_req, skb); queue->tx_next = list_next_entry_circular(tx_req, gpd_ring, entry); spin_unlock_irqrestore(&queue->ring_lock, flags); if (!t7xx_pci_sleep_disable_complete(md_ctrl->t7xx_dev)) { ret = -ETIMEDOUT; break; } /* Protect the access to the modem for queues operations (resume/start) * which access shared locations by all the queues. * cldma_lock is independent of ring_lock which is per queue. */ spin_lock_irqsave(&md_ctrl->cldma_lock, flags); t7xx_cldma_hw_start_send(md_ctrl, qno, tx_req); spin_unlock_irqrestore(&md_ctrl->cldma_lock, flags); break; } spin_unlock_irqrestore(&queue->ring_lock, flags); if (!t7xx_pci_sleep_disable_complete(md_ctrl->t7xx_dev)) { ret = -ETIMEDOUT; break; } if (!t7xx_cldma_hw_queue_status(&md_ctrl->hw_info, qno, MTK_TX)) { spin_lock_irqsave(&md_ctrl->cldma_lock, flags); t7xx_cldma_hw_resume_queue(&md_ctrl->hw_info, qno, MTK_TX); spin_unlock_irqrestore(&md_ctrl->cldma_lock, flags); } ret = wait_event_interruptible_exclusive(queue->req_wq, queue->budget > 0); } while (!ret); allow_sleep: t7xx_pci_enable_sleep(md_ctrl->t7xx_dev); pm_runtime_mark_last_busy(md_ctrl->dev); pm_runtime_put_autosuspend(md_ctrl->dev); return ret; } static int t7xx_cldma_late_init(struct cldma_ctrl *md_ctrl) { char dma_pool_name[32]; int i, j, ret; if (md_ctrl->is_late_init) { dev_err(md_ctrl->dev, "CLDMA late init was already done\n"); return -EALREADY; } snprintf(dma_pool_name, sizeof(dma_pool_name), "cldma_req_hif%d", md_ctrl->hif_id); md_ctrl->gpd_dmapool = dma_pool_create(dma_pool_name, md_ctrl->dev, sizeof(struct cldma_gpd), GPD_DMAPOOL_ALIGN, 0); if (!md_ctrl->gpd_dmapool) { dev_err(md_ctrl->dev, "DMA pool alloc fail\n"); return -ENOMEM; } for (i = 0; i < CLDMA_TXQ_NUM; i++) { ret = t7xx_cldma_tx_ring_init(md_ctrl, &md_ctrl->tx_ring[i]); if (ret) { dev_err(md_ctrl->dev, "control TX ring init fail\n"); goto err_free_tx_ring; } md_ctrl->tx_ring[i].pkt_size = CLDMA_MTU; } for (j = 0; j < CLDMA_RXQ_NUM; j++) { md_ctrl->rx_ring[j].pkt_size = CLDMA_MTU; if (j == CLDMA_RXQ_NUM - 1) md_ctrl->rx_ring[j].pkt_size = CLDMA_JUMBO_BUFF_SZ; ret = t7xx_cldma_rx_ring_init(md_ctrl, &md_ctrl->rx_ring[j]); if (ret) { dev_err(md_ctrl->dev, "Control RX ring init fail\n"); goto err_free_rx_ring; } } for (i = 0; i < CLDMA_TXQ_NUM; i++) t7xx_cldma_txq_init(&md_ctrl->txq[i]); for (j = 0; j < CLDMA_RXQ_NUM; j++) t7xx_cldma_rxq_init(&md_ctrl->rxq[j]); md_ctrl->is_late_init = true; return 0; err_free_rx_ring: while (j--) t7xx_cldma_ring_free(md_ctrl, &md_ctrl->rx_ring[j], DMA_FROM_DEVICE); err_free_tx_ring: while (i--) t7xx_cldma_ring_free(md_ctrl, &md_ctrl->tx_ring[i], DMA_TO_DEVICE); return ret; } static void __iomem *t7xx_pcie_addr_transfer(void __iomem *addr, u32 addr_trs1, u32 phy_addr) { return addr + phy_addr - addr_trs1; } static void t7xx_hw_info_init(struct cldma_ctrl *md_ctrl) { struct t7xx_addr_base *pbase = &md_ctrl->t7xx_dev->base_addr; struct t7xx_cldma_hw *hw_info = &md_ctrl->hw_info; u32 phy_ao_base, phy_pd_base; if (md_ctrl->hif_id != CLDMA_ID_MD) return; phy_ao_base = CLDMA1_AO_BASE; phy_pd_base = CLDMA1_PD_BASE; hw_info->phy_interrupt_id = CLDMA1_INT; hw_info->hw_mode = MODE_BIT_64; hw_info->ap_ao_base = t7xx_pcie_addr_transfer(pbase->pcie_ext_reg_base, pbase->pcie_dev_reg_trsl_addr, phy_ao_base); hw_info->ap_pdn_base = t7xx_pcie_addr_transfer(pbase->pcie_ext_reg_base, pbase->pcie_dev_reg_trsl_addr, phy_pd_base); } static int t7xx_cldma_default_recv_skb(struct cldma_queue *queue, struct sk_buff *skb) { dev_kfree_skb_any(skb); return 0; } int t7xx_cldma_alloc(enum cldma_id hif_id, struct t7xx_pci_dev *t7xx_dev) { struct device *dev = &t7xx_dev->pdev->dev; struct cldma_ctrl *md_ctrl; md_ctrl = devm_kzalloc(dev, sizeof(*md_ctrl), GFP_KERNEL); if (!md_ctrl) return -ENOMEM; md_ctrl->t7xx_dev = t7xx_dev; md_ctrl->dev = dev; md_ctrl->hif_id = hif_id; md_ctrl->recv_skb = t7xx_cldma_default_recv_skb; t7xx_hw_info_init(md_ctrl); t7xx_dev->md->md_ctrl[hif_id] = md_ctrl; return 0; } static void t7xx_cldma_resume_early(struct t7xx_pci_dev *t7xx_dev, void *entity_param) { struct cldma_ctrl *md_ctrl = entity_param; struct t7xx_cldma_hw *hw_info; unsigned long flags; int qno_t; hw_info = &md_ctrl->hw_info; spin_lock_irqsave(&md_ctrl->cldma_lock, flags); t7xx_cldma_hw_restore(hw_info); for (qno_t = 0; qno_t < CLDMA_TXQ_NUM; qno_t++) { t7xx_cldma_hw_set_start_addr(hw_info, qno_t, md_ctrl->txq[qno_t].tx_next->gpd_addr, MTK_TX); t7xx_cldma_hw_set_start_addr(hw_info, qno_t, md_ctrl->rxq[qno_t].tr_done->gpd_addr, MTK_RX); } t7xx_cldma_enable_irq(md_ctrl); t7xx_cldma_hw_start_queue(hw_info, CLDMA_ALL_Q, MTK_RX); md_ctrl->rxq_active |= TXRX_STATUS_BITMASK; t7xx_cldma_hw_irq_en_eq(hw_info, CLDMA_ALL_Q, MTK_RX); t7xx_cldma_hw_irq_en_txrx(hw_info, CLDMA_ALL_Q, MTK_RX); spin_unlock_irqrestore(&md_ctrl->cldma_lock, flags); } static int t7xx_cldma_resume(struct t7xx_pci_dev *t7xx_dev, void *entity_param) { struct cldma_ctrl *md_ctrl = entity_param; unsigned long flags; spin_lock_irqsave(&md_ctrl->cldma_lock, flags); md_ctrl->txq_active |= TXRX_STATUS_BITMASK; t7xx_cldma_hw_irq_en_txrx(&md_ctrl->hw_info, CLDMA_ALL_Q, MTK_TX); t7xx_cldma_hw_irq_en_eq(&md_ctrl->hw_info, CLDMA_ALL_Q, MTK_TX); spin_unlock_irqrestore(&md_ctrl->cldma_lock, flags); if (md_ctrl->hif_id == CLDMA_ID_MD) t7xx_mhccif_mask_clr(t7xx_dev, D2H_SW_INT_MASK); return 0; } static void t7xx_cldma_suspend_late(struct t7xx_pci_dev *t7xx_dev, void *entity_param) { struct cldma_ctrl *md_ctrl = entity_param; struct t7xx_cldma_hw *hw_info; unsigned long flags; hw_info = &md_ctrl->hw_info; spin_lock_irqsave(&md_ctrl->cldma_lock, flags); t7xx_cldma_hw_irq_dis_eq(hw_info, CLDMA_ALL_Q, MTK_RX); t7xx_cldma_hw_irq_dis_txrx(hw_info, CLDMA_ALL_Q, MTK_RX); md_ctrl->rxq_active &= ~TXRX_STATUS_BITMASK; t7xx_cldma_hw_stop_all_qs(hw_info, MTK_RX); t7xx_cldma_clear_ip_busy(hw_info); t7xx_cldma_disable_irq(md_ctrl); spin_unlock_irqrestore(&md_ctrl->cldma_lock, flags); } static int t7xx_cldma_suspend(struct t7xx_pci_dev *t7xx_dev, void *entity_param) { struct cldma_ctrl *md_ctrl = entity_param; struct t7xx_cldma_hw *hw_info; unsigned long flags; if (md_ctrl->hif_id == CLDMA_ID_MD) t7xx_mhccif_mask_set(t7xx_dev, D2H_SW_INT_MASK); hw_info = &md_ctrl->hw_info; spin_lock_irqsave(&md_ctrl->cldma_lock, flags); t7xx_cldma_hw_irq_dis_eq(hw_info, CLDMA_ALL_Q, MTK_TX); t7xx_cldma_hw_irq_dis_txrx(hw_info, CLDMA_ALL_Q, MTK_TX); md_ctrl->txq_active &= ~TXRX_STATUS_BITMASK; t7xx_cldma_hw_stop_all_qs(hw_info, MTK_TX); md_ctrl->txq_started = 0; spin_unlock_irqrestore(&md_ctrl->cldma_lock, flags); return 0; } static int t7xx_cldma_pm_init(struct cldma_ctrl *md_ctrl) { md_ctrl->pm_entity = kzalloc(sizeof(*md_ctrl->pm_entity), GFP_KERNEL); if (!md_ctrl->pm_entity) return -ENOMEM; md_ctrl->pm_entity->entity_param = md_ctrl; if (md_ctrl->hif_id == CLDMA_ID_MD) md_ctrl->pm_entity->id = PM_ENTITY_ID_CTRL1; else md_ctrl->pm_entity->id = PM_ENTITY_ID_CTRL2; md_ctrl->pm_entity->suspend = t7xx_cldma_suspend; md_ctrl->pm_entity->suspend_late = t7xx_cldma_suspend_late; md_ctrl->pm_entity->resume = t7xx_cldma_resume; md_ctrl->pm_entity->resume_early = t7xx_cldma_resume_early; return t7xx_pci_pm_entity_register(md_ctrl->t7xx_dev, md_ctrl->pm_entity); } static int t7xx_cldma_pm_uninit(struct cldma_ctrl *md_ctrl) { if (!md_ctrl->pm_entity) return -EINVAL; t7xx_pci_pm_entity_unregister(md_ctrl->t7xx_dev, md_ctrl->pm_entity); kfree(md_ctrl->pm_entity); md_ctrl->pm_entity = NULL; return 0; } void t7xx_cldma_hif_hw_init(struct cldma_ctrl *md_ctrl) { struct t7xx_cldma_hw *hw_info = &md_ctrl->hw_info; unsigned long flags; spin_lock_irqsave(&md_ctrl->cldma_lock, flags); t7xx_cldma_hw_stop(hw_info, MTK_TX); t7xx_cldma_hw_stop(hw_info, MTK_RX); t7xx_cldma_hw_rx_done(hw_info, EMPTY_STATUS_BITMASK | TXRX_STATUS_BITMASK); t7xx_cldma_hw_tx_done(hw_info, EMPTY_STATUS_BITMASK | TXRX_STATUS_BITMASK); t7xx_cldma_hw_init(hw_info); spin_unlock_irqrestore(&md_ctrl->cldma_lock, flags); } static irqreturn_t t7xx_cldma_isr_handler(int irq, void *data) { struct cldma_ctrl *md_ctrl = data; u32 interrupt; interrupt = md_ctrl->hw_info.phy_interrupt_id; t7xx_pcie_mac_clear_int(md_ctrl->t7xx_dev, interrupt); t7xx_cldma_irq_work_cb(md_ctrl); t7xx_pcie_mac_clear_int_status(md_ctrl->t7xx_dev, interrupt); t7xx_pcie_mac_set_int(md_ctrl->t7xx_dev, interrupt); return IRQ_HANDLED; } static void t7xx_cldma_destroy_wqs(struct cldma_ctrl *md_ctrl) { int i; for (i = 0; i < CLDMA_TXQ_NUM; i++) { if (md_ctrl->txq[i].worker) { destroy_workqueue(md_ctrl->txq[i].worker); md_ctrl->txq[i].worker = NULL; } } for (i = 0; i < CLDMA_RXQ_NUM; i++) { if (md_ctrl->rxq[i].worker) { destroy_workqueue(md_ctrl->rxq[i].worker); md_ctrl->rxq[i].worker = NULL; } } } /** * t7xx_cldma_init() - Initialize CLDMA. * @md_ctrl: CLDMA context structure. * * Allocate and initialize device power management entity. * Initialize HIF TX/RX queue structure. * Register CLDMA callback ISR with PCIe driver. * * Return: * * 0 - Success. * * -ERROR - Error code from failure sub-initializations. */ int t7xx_cldma_init(struct cldma_ctrl *md_ctrl) { struct t7xx_cldma_hw *hw_info = &md_ctrl->hw_info; int ret, i; md_ctrl->txq_active = 0; md_ctrl->rxq_active = 0; md_ctrl->is_late_init = false; ret = t7xx_cldma_pm_init(md_ctrl); if (ret) return ret; spin_lock_init(&md_ctrl->cldma_lock); for (i = 0; i < CLDMA_TXQ_NUM; i++) { md_cd_queue_struct_init(&md_ctrl->txq[i], md_ctrl, MTK_TX, i); md_ctrl->txq[i].worker = alloc_ordered_workqueue("md_hif%d_tx%d_worker", WQ_MEM_RECLAIM | (i ? 0 : WQ_HIGHPRI), md_ctrl->hif_id, i); if (!md_ctrl->txq[i].worker) goto err_workqueue; INIT_WORK(&md_ctrl->txq[i].cldma_work, t7xx_cldma_tx_done); } for (i = 0; i < CLDMA_RXQ_NUM; i++) { md_cd_queue_struct_init(&md_ctrl->rxq[i], md_ctrl, MTK_RX, i); INIT_WORK(&md_ctrl->rxq[i].cldma_work, t7xx_cldma_rx_done); md_ctrl->rxq[i].worker = alloc_ordered_workqueue("md_hif%d_rx%d_worker", WQ_MEM_RECLAIM, md_ctrl->hif_id, i); if (!md_ctrl->rxq[i].worker) goto err_workqueue; } t7xx_pcie_mac_clear_int(md_ctrl->t7xx_dev, hw_info->phy_interrupt_id); md_ctrl->t7xx_dev->intr_handler[hw_info->phy_interrupt_id] = t7xx_cldma_isr_handler; md_ctrl->t7xx_dev->intr_thread[hw_info->phy_interrupt_id] = NULL; md_ctrl->t7xx_dev->callback_param[hw_info->phy_interrupt_id] = md_ctrl; t7xx_pcie_mac_clear_int_status(md_ctrl->t7xx_dev, hw_info->phy_interrupt_id); return 0; err_workqueue: t7xx_cldma_destroy_wqs(md_ctrl); t7xx_cldma_pm_uninit(md_ctrl); return -ENOMEM; } void t7xx_cldma_switch_cfg(struct cldma_ctrl *md_ctrl) { t7xx_cldma_late_release(md_ctrl); t7xx_cldma_late_init(md_ctrl); } void t7xx_cldma_exit(struct cldma_ctrl *md_ctrl) { t7xx_cldma_stop(md_ctrl); t7xx_cldma_late_release(md_ctrl); t7xx_cldma_destroy_wqs(md_ctrl); t7xx_cldma_pm_uninit(md_ctrl); }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1