Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Alex Maftei (amaftei) | 1390 | 68.88% | 8 | 22.86% |
Edward Cree | 274 | 13.58% | 10 | 28.57% |
Ben Hutchings | 233 | 11.55% | 11 | 31.43% |
Martin Habets | 80 | 3.96% | 1 | 2.86% |
Shradha Shah | 18 | 0.89% | 1 | 2.86% |
Daniel Pieczko | 10 | 0.50% | 1 | 2.86% |
Jon Cooper | 7 | 0.35% | 1 | 2.86% |
Steve Hodgson | 5 | 0.25% | 1 | 2.86% |
Thomas Gleixner | 1 | 0.05% | 1 | 2.86% |
Total | 2018 | 35 |
// SPDX-License-Identifier: GPL-2.0-only /**************************************************************************** * Driver for Solarflare network controllers and boards * Copyright 2019 Solarflare Communications Inc. * * This program is free software; you can redistribute it and/or modify it * under the terms of the GNU General Public License version 2 as published * by the Free Software Foundation, incorporated herein by reference. */ #include "net_driver.h" #include "efx.h" #include "nic.h" #include "mcdi_functions.h" #include "mcdi.h" #include "mcdi_pcol.h" int efx_mcdi_free_vis(struct efx_nic *efx) { MCDI_DECLARE_BUF_ERR(outbuf); size_t outlen; int rc = efx_mcdi_rpc_quiet(efx, MC_CMD_FREE_VIS, NULL, 0, outbuf, sizeof(outbuf), &outlen); /* -EALREADY means nothing to free, so ignore */ if (rc == -EALREADY) rc = 0; if (rc) efx_mcdi_display_error(efx, MC_CMD_FREE_VIS, 0, outbuf, outlen, rc); return rc; } int efx_mcdi_alloc_vis(struct efx_nic *efx, unsigned int min_vis, unsigned int max_vis, unsigned int *vi_base, unsigned int *allocated_vis) { MCDI_DECLARE_BUF(outbuf, MC_CMD_ALLOC_VIS_OUT_LEN); MCDI_DECLARE_BUF(inbuf, MC_CMD_ALLOC_VIS_IN_LEN); size_t outlen; int rc; MCDI_SET_DWORD(inbuf, ALLOC_VIS_IN_MIN_VI_COUNT, min_vis); MCDI_SET_DWORD(inbuf, ALLOC_VIS_IN_MAX_VI_COUNT, max_vis); rc = efx_mcdi_rpc(efx, MC_CMD_ALLOC_VIS, inbuf, sizeof(inbuf), outbuf, sizeof(outbuf), &outlen); if (rc != 0) return rc; if (outlen < MC_CMD_ALLOC_VIS_OUT_LEN) return -EIO; netif_dbg(efx, drv, efx->net_dev, "base VI is A0x%03x\n", MCDI_DWORD(outbuf, ALLOC_VIS_OUT_VI_BASE)); if (vi_base) *vi_base = MCDI_DWORD(outbuf, ALLOC_VIS_OUT_VI_BASE); if (allocated_vis) *allocated_vis = MCDI_DWORD(outbuf, ALLOC_VIS_OUT_VI_COUNT); return 0; } int efx_mcdi_ev_probe(struct efx_channel *channel) { return efx_nic_alloc_buffer(channel->efx, &channel->eventq, (channel->eventq_mask + 1) * sizeof(efx_qword_t), GFP_KERNEL); } int efx_mcdi_ev_init(struct efx_channel *channel, bool v1_cut_thru, bool v2) { MCDI_DECLARE_BUF(inbuf, MC_CMD_INIT_EVQ_V2_IN_LEN(EFX_MAX_EVQ_SIZE * 8 / EFX_BUF_SIZE)); MCDI_DECLARE_BUF(outbuf, MC_CMD_INIT_EVQ_V2_OUT_LEN); size_t entries = channel->eventq.len / EFX_BUF_SIZE; struct efx_nic *efx = channel->efx; size_t inlen, outlen; dma_addr_t dma_addr; int rc, i; /* Fill event queue with all ones (i.e. empty events) */ memset(channel->eventq.addr, 0xff, channel->eventq.len); MCDI_SET_DWORD(inbuf, INIT_EVQ_IN_SIZE, channel->eventq_mask + 1); MCDI_SET_DWORD(inbuf, INIT_EVQ_IN_INSTANCE, channel->channel); /* INIT_EVQ expects index in vector table, not absolute */ MCDI_SET_DWORD(inbuf, INIT_EVQ_IN_IRQ_NUM, channel->channel); MCDI_SET_DWORD(inbuf, INIT_EVQ_IN_TMR_MODE, MC_CMD_INIT_EVQ_IN_TMR_MODE_DIS); MCDI_SET_DWORD(inbuf, INIT_EVQ_IN_TMR_LOAD, 0); MCDI_SET_DWORD(inbuf, INIT_EVQ_IN_TMR_RELOAD, 0); MCDI_SET_DWORD(inbuf, INIT_EVQ_IN_COUNT_MODE, MC_CMD_INIT_EVQ_IN_COUNT_MODE_DIS); MCDI_SET_DWORD(inbuf, INIT_EVQ_IN_COUNT_THRSHLD, 0); if (v2) { /* Use the new generic approach to specifying event queue * configuration, requesting lower latency or higher throughput. * The options that actually get used appear in the output. */ MCDI_POPULATE_DWORD_2(inbuf, INIT_EVQ_V2_IN_FLAGS, INIT_EVQ_V2_IN_FLAG_INTERRUPTING, 1, INIT_EVQ_V2_IN_FLAG_TYPE, MC_CMD_INIT_EVQ_V2_IN_FLAG_TYPE_AUTO); } else { MCDI_POPULATE_DWORD_4(inbuf, INIT_EVQ_IN_FLAGS, INIT_EVQ_IN_FLAG_INTERRUPTING, 1, INIT_EVQ_IN_FLAG_RX_MERGE, 1, INIT_EVQ_IN_FLAG_TX_MERGE, 1, INIT_EVQ_IN_FLAG_CUT_THRU, v1_cut_thru); } dma_addr = channel->eventq.dma_addr; for (i = 0; i < entries; ++i) { MCDI_SET_ARRAY_QWORD(inbuf, INIT_EVQ_IN_DMA_ADDR, i, dma_addr); dma_addr += EFX_BUF_SIZE; } inlen = MC_CMD_INIT_EVQ_IN_LEN(entries); rc = efx_mcdi_rpc(efx, MC_CMD_INIT_EVQ, inbuf, inlen, outbuf, sizeof(outbuf), &outlen); if (outlen >= MC_CMD_INIT_EVQ_V2_OUT_LEN) netif_dbg(efx, drv, efx->net_dev, "Channel %d using event queue flags %08x\n", channel->channel, MCDI_DWORD(outbuf, INIT_EVQ_V2_OUT_FLAGS)); return rc; } void efx_mcdi_ev_remove(struct efx_channel *channel) { efx_nic_free_buffer(channel->efx, &channel->eventq); } void efx_mcdi_ev_fini(struct efx_channel *channel) { MCDI_DECLARE_BUF(inbuf, MC_CMD_FINI_EVQ_IN_LEN); MCDI_DECLARE_BUF_ERR(outbuf); struct efx_nic *efx = channel->efx; size_t outlen; int rc; MCDI_SET_DWORD(inbuf, FINI_EVQ_IN_INSTANCE, channel->channel); rc = efx_mcdi_rpc_quiet(efx, MC_CMD_FINI_EVQ, inbuf, sizeof(inbuf), outbuf, sizeof(outbuf), &outlen); if (rc && rc != -EALREADY) goto fail; return; fail: efx_mcdi_display_error(efx, MC_CMD_FINI_EVQ, MC_CMD_FINI_EVQ_IN_LEN, outbuf, outlen, rc); } int efx_mcdi_tx_init(struct efx_tx_queue *tx_queue) { MCDI_DECLARE_BUF(inbuf, MC_CMD_INIT_TXQ_IN_LEN(EFX_MAX_DMAQ_SIZE * 8 / EFX_BUF_SIZE)); bool csum_offload = tx_queue->type & EFX_TXQ_TYPE_OUTER_CSUM; bool inner_csum = tx_queue->type & EFX_TXQ_TYPE_INNER_CSUM; size_t entries = tx_queue->txd.len / EFX_BUF_SIZE; struct efx_channel *channel = tx_queue->channel; struct efx_nic *efx = tx_queue->efx; dma_addr_t dma_addr; size_t inlen; int rc, i; BUILD_BUG_ON(MC_CMD_INIT_TXQ_OUT_LEN != 0); MCDI_SET_DWORD(inbuf, INIT_TXQ_IN_SIZE, tx_queue->ptr_mask + 1); MCDI_SET_DWORD(inbuf, INIT_TXQ_IN_TARGET_EVQ, channel->channel); MCDI_SET_DWORD(inbuf, INIT_TXQ_IN_LABEL, tx_queue->label); MCDI_SET_DWORD(inbuf, INIT_TXQ_IN_INSTANCE, tx_queue->queue); MCDI_SET_DWORD(inbuf, INIT_TXQ_IN_OWNER_ID, 0); MCDI_SET_DWORD(inbuf, INIT_TXQ_IN_PORT_ID, efx->vport_id); dma_addr = tx_queue->txd.dma_addr; netif_dbg(efx, hw, efx->net_dev, "pushing TXQ %d. %zu entries (%llx)\n", tx_queue->queue, entries, (u64)dma_addr); for (i = 0; i < entries; ++i) { MCDI_SET_ARRAY_QWORD(inbuf, INIT_TXQ_IN_DMA_ADDR, i, dma_addr); dma_addr += EFX_BUF_SIZE; } inlen = MC_CMD_INIT_TXQ_IN_LEN(entries); do { bool tso_v2 = tx_queue->tso_version == 2; /* TSOv2 implies IP header checksum offload for TSO frames, * so we can safely disable IP header checksum offload for * everything else. If we don't have TSOv2, then we have to * enable IP header checksum offload, which is strictly * incorrect but better than breaking TSO. */ MCDI_POPULATE_DWORD_6(inbuf, INIT_TXQ_IN_FLAGS, /* This flag was removed from mcdi_pcol.h for * the non-_EXT version of INIT_TXQ. However, * firmware still honours it. */ INIT_TXQ_EXT_IN_FLAG_TSOV2_EN, tso_v2, INIT_TXQ_IN_FLAG_IP_CSUM_DIS, !(csum_offload && tso_v2), INIT_TXQ_IN_FLAG_TCP_CSUM_DIS, !csum_offload, INIT_TXQ_EXT_IN_FLAG_TIMESTAMP, tx_queue->timestamping, INIT_TXQ_IN_FLAG_INNER_IP_CSUM_EN, inner_csum && !tso_v2, INIT_TXQ_IN_FLAG_INNER_TCP_CSUM_EN, inner_csum); rc = efx_mcdi_rpc_quiet(efx, MC_CMD_INIT_TXQ, inbuf, inlen, NULL, 0, NULL); if (rc == -ENOSPC && tso_v2) { /* Retry without TSOv2 if we're short on contexts. */ tx_queue->tso_version = 0; netif_warn(efx, probe, efx->net_dev, "TSOv2 context not available to segment in " "hardware. TCP performance may be reduced.\n" ); } else if (rc) { efx_mcdi_display_error(efx, MC_CMD_INIT_TXQ, MC_CMD_INIT_TXQ_EXT_IN_LEN, NULL, 0, rc); goto fail; } } while (rc); return 0; fail: return rc; } void efx_mcdi_tx_remove(struct efx_tx_queue *tx_queue) { efx_nic_free_buffer(tx_queue->efx, &tx_queue->txd); } void efx_mcdi_tx_fini(struct efx_tx_queue *tx_queue) { MCDI_DECLARE_BUF(inbuf, MC_CMD_FINI_TXQ_IN_LEN); MCDI_DECLARE_BUF_ERR(outbuf); struct efx_nic *efx = tx_queue->efx; size_t outlen; int rc; MCDI_SET_DWORD(inbuf, FINI_TXQ_IN_INSTANCE, tx_queue->queue); rc = efx_mcdi_rpc_quiet(efx, MC_CMD_FINI_TXQ, inbuf, sizeof(inbuf), outbuf, sizeof(outbuf), &outlen); if (rc && rc != -EALREADY) goto fail; return; fail: efx_mcdi_display_error(efx, MC_CMD_FINI_TXQ, MC_CMD_FINI_TXQ_IN_LEN, outbuf, outlen, rc); } int efx_mcdi_rx_probe(struct efx_rx_queue *rx_queue) { return efx_nic_alloc_buffer(rx_queue->efx, &rx_queue->rxd, (rx_queue->ptr_mask + 1) * sizeof(efx_qword_t), GFP_KERNEL); } void efx_mcdi_rx_init(struct efx_rx_queue *rx_queue) { struct efx_channel *channel = efx_rx_queue_channel(rx_queue); size_t entries = rx_queue->rxd.len / EFX_BUF_SIZE; MCDI_DECLARE_BUF(inbuf, MC_CMD_INIT_RXQ_V4_IN_LEN); struct efx_nic *efx = rx_queue->efx; unsigned int buffer_size; dma_addr_t dma_addr; int rc; int i; BUILD_BUG_ON(MC_CMD_INIT_RXQ_OUT_LEN != 0); rx_queue->scatter_n = 0; rx_queue->scatter_len = 0; if (efx->type->revision == EFX_REV_EF100) buffer_size = efx->rx_page_buf_step; else buffer_size = 0; MCDI_SET_DWORD(inbuf, INIT_RXQ_IN_SIZE, rx_queue->ptr_mask + 1); MCDI_SET_DWORD(inbuf, INIT_RXQ_IN_TARGET_EVQ, channel->channel); MCDI_SET_DWORD(inbuf, INIT_RXQ_IN_LABEL, efx_rx_queue_index(rx_queue)); MCDI_SET_DWORD(inbuf, INIT_RXQ_IN_INSTANCE, efx_rx_queue_index(rx_queue)); MCDI_POPULATE_DWORD_2(inbuf, INIT_RXQ_IN_FLAGS, INIT_RXQ_IN_FLAG_PREFIX, 1, INIT_RXQ_IN_FLAG_TIMESTAMP, 1); MCDI_SET_DWORD(inbuf, INIT_RXQ_IN_OWNER_ID, 0); MCDI_SET_DWORD(inbuf, INIT_RXQ_IN_PORT_ID, efx->vport_id); MCDI_SET_DWORD(inbuf, INIT_RXQ_V4_IN_BUFFER_SIZE_BYTES, buffer_size); dma_addr = rx_queue->rxd.dma_addr; netif_dbg(efx, hw, efx->net_dev, "pushing RXQ %d. %zu entries (%llx)\n", efx_rx_queue_index(rx_queue), entries, (u64)dma_addr); for (i = 0; i < entries; ++i) { MCDI_SET_ARRAY_QWORD(inbuf, INIT_RXQ_IN_DMA_ADDR, i, dma_addr); dma_addr += EFX_BUF_SIZE; } rc = efx_mcdi_rpc(efx, MC_CMD_INIT_RXQ, inbuf, sizeof(inbuf), NULL, 0, NULL); if (rc) netdev_WARN(efx->net_dev, "failed to initialise RXQ %d\n", efx_rx_queue_index(rx_queue)); } void efx_mcdi_rx_remove(struct efx_rx_queue *rx_queue) { efx_nic_free_buffer(rx_queue->efx, &rx_queue->rxd); } void efx_mcdi_rx_fini(struct efx_rx_queue *rx_queue) { MCDI_DECLARE_BUF(inbuf, MC_CMD_FINI_RXQ_IN_LEN); MCDI_DECLARE_BUF_ERR(outbuf); struct efx_nic *efx = rx_queue->efx; size_t outlen; int rc; MCDI_SET_DWORD(inbuf, FINI_RXQ_IN_INSTANCE, efx_rx_queue_index(rx_queue)); rc = efx_mcdi_rpc_quiet(efx, MC_CMD_FINI_RXQ, inbuf, sizeof(inbuf), outbuf, sizeof(outbuf), &outlen); if (rc && rc != -EALREADY) goto fail; return; fail: efx_mcdi_display_error(efx, MC_CMD_FINI_RXQ, MC_CMD_FINI_RXQ_IN_LEN, outbuf, outlen, rc); } int efx_fini_dmaq(struct efx_nic *efx) { struct efx_tx_queue *tx_queue; struct efx_rx_queue *rx_queue; struct efx_channel *channel; int pending; /* If the MC has just rebooted, the TX/RX queues will have already been * torn down, but efx->active_queues needs to be set to zero. */ if (efx->must_realloc_vis) { atomic_set(&efx->active_queues, 0); return 0; } /* Do not attempt to write to the NIC during EEH recovery */ if (efx->state != STATE_RECOVERY) { efx_for_each_channel(channel, efx) { efx_for_each_channel_rx_queue(rx_queue, channel) efx_mcdi_rx_fini(rx_queue); efx_for_each_channel_tx_queue(tx_queue, channel) efx_mcdi_tx_fini(tx_queue); } wait_event_timeout(efx->flush_wq, atomic_read(&efx->active_queues) == 0, msecs_to_jiffies(EFX_MAX_FLUSH_TIME)); pending = atomic_read(&efx->active_queues); if (pending) { netif_err(efx, hw, efx->net_dev, "failed to flush %d queues\n", pending); return -ETIMEDOUT; } } return 0; } int efx_mcdi_window_mode_to_stride(struct efx_nic *efx, u8 vi_window_mode) { switch (vi_window_mode) { case MC_CMD_GET_CAPABILITIES_V3_OUT_VI_WINDOW_MODE_8K: efx->vi_stride = 8192; break; case MC_CMD_GET_CAPABILITIES_V3_OUT_VI_WINDOW_MODE_16K: efx->vi_stride = 16384; break; case MC_CMD_GET_CAPABILITIES_V3_OUT_VI_WINDOW_MODE_64K: efx->vi_stride = 65536; break; default: netif_err(efx, probe, efx->net_dev, "Unrecognised VI window mode %d\n", vi_window_mode); return -EIO; } netif_dbg(efx, probe, efx->net_dev, "vi_stride = %u\n", efx->vi_stride); return 0; } int efx_get_pf_index(struct efx_nic *efx, unsigned int *pf_index) { MCDI_DECLARE_BUF(outbuf, MC_CMD_GET_FUNCTION_INFO_OUT_LEN); size_t outlen; int rc; rc = efx_mcdi_rpc(efx, MC_CMD_GET_FUNCTION_INFO, NULL, 0, outbuf, sizeof(outbuf), &outlen); if (rc) return rc; if (outlen < sizeof(outbuf)) return -EIO; *pf_index = MCDI_DWORD(outbuf, GET_FUNCTION_INFO_OUT_PF); return 0; }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1