Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Michael J. Ruhl | 1273 | 100.00% | 3 | 100.00% |
Total | 1273 | 3 |
// SPDX-License-Identifier: (GPL-2.0 OR BSD-3-Clause) /* * Copyright(c) 2018 Intel Corporation. * * This file is provided under a dual BSD/GPLv2 license. When using or * redistributing this file, you may do so under either license. * * GPL LICENSE SUMMARY * * This program is free software; you can redistribute it and/or modify * it under the terms of version 2 of the GNU General Public License as * published by the Free Software Foundation. * * This program is distributed in the hope that it will be useful, but * WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * General Public License for more details. * * BSD LICENSE * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * * - Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * - Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in * the documentation and/or other materials provided with the * distribution. * - Neither the name of Intel Corporation nor the names of its * contributors may be used to endorse or promote products derived * from this software without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. * */ #include "hfi.h" #include "affinity.h" #include "sdma.h" /** * msix_initialize() - Calculate, request and configure MSIx IRQs * @dd: valid hfi1 devdata * */ int msix_initialize(struct hfi1_devdata *dd) { u32 total; int ret; struct hfi1_msix_entry *entries; /* * MSIx interrupt count: * one for the general, "slow path" interrupt * one per used SDMA engine * one per kernel receive context * one for each VNIC context * ...any new IRQs should be added here. */ total = 1 + dd->num_sdma + dd->n_krcv_queues + dd->num_vnic_contexts; if (total >= CCE_NUM_MSIX_VECTORS) return -EINVAL; ret = pci_alloc_irq_vectors(dd->pcidev, total, total, PCI_IRQ_MSIX); if (ret < 0) { dd_dev_err(dd, "pci_alloc_irq_vectors() failed: %d\n", ret); return ret; } entries = kcalloc(total, sizeof(*dd->msix_info.msix_entries), GFP_KERNEL); if (!entries) { pci_free_irq_vectors(dd->pcidev); return -ENOMEM; } dd->msix_info.msix_entries = entries; spin_lock_init(&dd->msix_info.msix_lock); bitmap_zero(dd->msix_info.in_use_msix, total); dd->msix_info.max_requested = total; dd_dev_info(dd, "%u MSI-X interrupts allocated\n", total); return 0; } /** * msix_request_irq() - Allocate a free MSIx IRQ * @dd: valid devdata * @arg: context information for the IRQ * @handler: IRQ handler * @thread: IRQ thread handler (could be NULL) * @idx: zero base idx if multiple devices are needed * @type: affinty IRQ type * * Allocated an MSIx vector if available, and then create the appropriate * meta data needed to keep track of the pci IRQ request. * * Return: * < 0 Error * >= 0 MSIx vector * */ static int msix_request_irq(struct hfi1_devdata *dd, void *arg, irq_handler_t handler, irq_handler_t thread, u32 idx, enum irq_type type) { unsigned long nr; int irq; int ret; const char *err_info; char name[MAX_NAME_SIZE]; struct hfi1_msix_entry *me; /* Allocate an MSIx vector */ spin_lock(&dd->msix_info.msix_lock); nr = find_first_zero_bit(dd->msix_info.in_use_msix, dd->msix_info.max_requested); if (nr < dd->msix_info.max_requested) __set_bit(nr, dd->msix_info.in_use_msix); spin_unlock(&dd->msix_info.msix_lock); if (nr == dd->msix_info.max_requested) return -ENOSPC; /* Specific verification and determine the name */ switch (type) { case IRQ_GENERAL: /* general interrupt must be MSIx vector 0 */ if (nr) { spin_lock(&dd->msix_info.msix_lock); __clear_bit(nr, dd->msix_info.in_use_msix); spin_unlock(&dd->msix_info.msix_lock); dd_dev_err(dd, "Invalid index %lu for GENERAL IRQ\n", nr); return -EINVAL; } snprintf(name, sizeof(name), DRIVER_NAME "_%d", dd->unit); err_info = "general"; break; case IRQ_SDMA: snprintf(name, sizeof(name), DRIVER_NAME "_%d sdma%d", dd->unit, idx); err_info = "sdma"; break; case IRQ_RCVCTXT: snprintf(name, sizeof(name), DRIVER_NAME "_%d kctxt%d", dd->unit, idx); err_info = "receive context"; break; case IRQ_OTHER: default: return -EINVAL; } name[sizeof(name) - 1] = 0; irq = pci_irq_vector(dd->pcidev, nr); ret = pci_request_irq(dd->pcidev, nr, handler, thread, arg, name); if (ret) { dd_dev_err(dd, "%s: request for IRQ %d failed, MSIx %d, err %d\n", err_info, irq, idx, ret); spin_lock(&dd->msix_info.msix_lock); __clear_bit(nr, dd->msix_info.in_use_msix); spin_unlock(&dd->msix_info.msix_lock); return ret; } /* * assign arg after pci_request_irq call, so it will be * cleaned up */ me = &dd->msix_info.msix_entries[nr]; me->irq = irq; me->arg = arg; me->type = type; /* This is a request, so a failure is not fatal */ ret = hfi1_get_irq_affinity(dd, me); if (ret) dd_dev_err(dd, "unable to pin IRQ %d\n", ret); return nr; } /** * msix_request_rcd_irq() - Helper function for RCVAVAIL IRQs * @rcd: valid rcd context * */ int msix_request_rcd_irq(struct hfi1_ctxtdata *rcd) { int nr; nr = msix_request_irq(rcd->dd, rcd, receive_context_interrupt, receive_context_thread, rcd->ctxt, IRQ_RCVCTXT); if (nr < 0) return nr; /* * Set the interrupt register and mask for this * context's interrupt. */ rcd->ireg = (IS_RCVAVAIL_START + rcd->ctxt) / 64; rcd->imask = ((u64)1) << ((IS_RCVAVAIL_START + rcd->ctxt) % 64); rcd->msix_intr = nr; remap_intr(rcd->dd, IS_RCVAVAIL_START + rcd->ctxt, nr); return 0; } /** * msix_request_smda_ira() - Helper for getting SDMA IRQ resources * @sde: valid sdma engine * */ int msix_request_sdma_irq(struct sdma_engine *sde) { int nr; nr = msix_request_irq(sde->dd, sde, sdma_interrupt, NULL, sde->this_idx, IRQ_SDMA); if (nr < 0) return nr; sde->msix_intr = nr; remap_sdma_interrupts(sde->dd, sde->this_idx, nr); return 0; } /** * enable_sdma_src() - Helper to enable SDMA IRQ srcs * @dd: valid devdata structure * @i: index of SDMA engine */ static void enable_sdma_srcs(struct hfi1_devdata *dd, int i) { set_intr_bits(dd, IS_SDMA_START + i, IS_SDMA_START + i, true); set_intr_bits(dd, IS_SDMA_PROGRESS_START + i, IS_SDMA_PROGRESS_START + i, true); set_intr_bits(dd, IS_SDMA_IDLE_START + i, IS_SDMA_IDLE_START + i, true); set_intr_bits(dd, IS_SDMAENG_ERR_START + i, IS_SDMAENG_ERR_START + i, true); } /** * msix_request_irqs() - Allocate all MSIx IRQs * @dd: valid devdata structure * * Helper function to request the used MSIx IRQs. * */ int msix_request_irqs(struct hfi1_devdata *dd) { int i; int ret; ret = msix_request_irq(dd, dd, general_interrupt, NULL, 0, IRQ_GENERAL); if (ret < 0) return ret; for (i = 0; i < dd->num_sdma; i++) { struct sdma_engine *sde = &dd->per_sdma[i]; ret = msix_request_sdma_irq(sde); if (ret) return ret; enable_sdma_srcs(sde->dd, i); } for (i = 0; i < dd->n_krcv_queues; i++) { struct hfi1_ctxtdata *rcd = hfi1_rcd_get_by_index_safe(dd, i); if (rcd) ret = msix_request_rcd_irq(rcd); hfi1_rcd_put(rcd); if (ret) return ret; } return 0; } /** * msix_free_irq() - Free the specified MSIx resources and IRQ * @dd: valid devdata * @msix_intr: MSIx vector to free. * */ void msix_free_irq(struct hfi1_devdata *dd, u8 msix_intr) { struct hfi1_msix_entry *me; if (msix_intr >= dd->msix_info.max_requested) return; me = &dd->msix_info.msix_entries[msix_intr]; if (!me->arg) /* => no irq, no affinity */ return; hfi1_put_irq_affinity(dd, me); pci_free_irq(dd->pcidev, msix_intr, me->arg); me->arg = NULL; spin_lock(&dd->msix_info.msix_lock); __clear_bit(msix_intr, dd->msix_info.in_use_msix); spin_unlock(&dd->msix_info.msix_lock); } /** * hfi1_clean_up_msix_interrupts() - Free all MSIx IRQ resources * @dd: valid device data data structure * * Free the MSIx and associated PCI resources, if they have been allocated. */ void msix_clean_up_interrupts(struct hfi1_devdata *dd) { int i; struct hfi1_msix_entry *me = dd->msix_info.msix_entries; /* remove irqs - must happen before disabling/turning off */ for (i = 0; i < dd->msix_info.max_requested; i++, me++) msix_free_irq(dd, i); /* clean structures */ kfree(dd->msix_info.msix_entries); dd->msix_info.msix_entries = NULL; dd->msix_info.max_requested = 0; pci_free_irq_vectors(dd->pcidev); } /** * msix_vnic_syncrhonize_irq() - Vnic IRQ synchronize * @dd: valid devdata */ void msix_vnic_synchronize_irq(struct hfi1_devdata *dd) { int i; for (i = 0; i < dd->vnic.num_ctxt; i++) { struct hfi1_ctxtdata *rcd = dd->vnic.ctxt[i]; struct hfi1_msix_entry *me; me = &dd->msix_info.msix_entries[rcd->msix_intr]; synchronize_irq(me->irq); } }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1