Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Srikanth Jampala | 2113 | 99.91% | 5 | 71.43% |
Kees Cook | 1 | 0.05% | 1 | 14.29% |
Greg Kroah-Hartman | 1 | 0.05% | 1 | 14.29% |
Total | 2115 | 7 |
// SPDX-License-Identifier: GPL-2.0 #include <linux/pci.h> #include <linux/printk.h> #include <linux/slab.h> #include "nitrox_dev.h" #include "nitrox_csr.h" #include "nitrox_common.h" #include "nitrox_hal.h" #include "nitrox_mbx.h" /** * One vector for each type of ring * - NPS packet ring, AQMQ ring and ZQMQ ring */ #define NR_RING_VECTORS 3 #define NR_NON_RING_VECTORS 1 /* base entry for packet ring/port */ #define PKT_RING_MSIX_BASE 0 #define NON_RING_MSIX_BASE 192 /** * nps_pkt_slc_isr - IRQ handler for NPS solicit port * @irq: irq number * @data: argument */ static irqreturn_t nps_pkt_slc_isr(int irq, void *data) { struct nitrox_q_vector *qvec = data; union nps_pkt_slc_cnts slc_cnts; struct nitrox_cmdq *cmdq = qvec->cmdq; slc_cnts.value = readq(cmdq->compl_cnt_csr_addr); /* New packet on SLC output port */ if (slc_cnts.s.slc_int) tasklet_hi_schedule(&qvec->resp_tasklet); return IRQ_HANDLED; } static void clear_nps_core_err_intr(struct nitrox_device *ndev) { u64 value; /* Write 1 to clear */ value = nitrox_read_csr(ndev, NPS_CORE_INT); nitrox_write_csr(ndev, NPS_CORE_INT, value); dev_err_ratelimited(DEV(ndev), "NSP_CORE_INT 0x%016llx\n", value); } static void clear_nps_pkt_err_intr(struct nitrox_device *ndev) { union nps_pkt_int pkt_int; unsigned long value, offset; int i; pkt_int.value = nitrox_read_csr(ndev, NPS_PKT_INT); dev_err_ratelimited(DEV(ndev), "NPS_PKT_INT 0x%016llx\n", pkt_int.value); if (pkt_int.s.slc_err) { offset = NPS_PKT_SLC_ERR_TYPE; value = nitrox_read_csr(ndev, offset); nitrox_write_csr(ndev, offset, value); dev_err_ratelimited(DEV(ndev), "NPS_PKT_SLC_ERR_TYPE 0x%016lx\n", value); offset = NPS_PKT_SLC_RERR_LO; value = nitrox_read_csr(ndev, offset); nitrox_write_csr(ndev, offset, value); /* enable the solicit ports */ for_each_set_bit(i, &value, BITS_PER_LONG) enable_pkt_solicit_port(ndev, i); dev_err_ratelimited(DEV(ndev), "NPS_PKT_SLC_RERR_LO 0x%016lx\n", value); offset = NPS_PKT_SLC_RERR_HI; value = nitrox_read_csr(ndev, offset); nitrox_write_csr(ndev, offset, value); dev_err_ratelimited(DEV(ndev), "NPS_PKT_SLC_RERR_HI 0x%016lx\n", value); } if (pkt_int.s.in_err) { offset = NPS_PKT_IN_ERR_TYPE; value = nitrox_read_csr(ndev, offset); nitrox_write_csr(ndev, offset, value); dev_err_ratelimited(DEV(ndev), "NPS_PKT_IN_ERR_TYPE 0x%016lx\n", value); offset = NPS_PKT_IN_RERR_LO; value = nitrox_read_csr(ndev, offset); nitrox_write_csr(ndev, offset, value); /* enable the input ring */ for_each_set_bit(i, &value, BITS_PER_LONG) enable_pkt_input_ring(ndev, i); dev_err_ratelimited(DEV(ndev), "NPS_PKT_IN_RERR_LO 0x%016lx\n", value); offset = NPS_PKT_IN_RERR_HI; value = nitrox_read_csr(ndev, offset); nitrox_write_csr(ndev, offset, value); dev_err_ratelimited(DEV(ndev), "NPS_PKT_IN_RERR_HI 0x%016lx\n", value); } } static void clear_pom_err_intr(struct nitrox_device *ndev) { u64 value; value = nitrox_read_csr(ndev, POM_INT); nitrox_write_csr(ndev, POM_INT, value); dev_err_ratelimited(DEV(ndev), "POM_INT 0x%016llx\n", value); } static void clear_pem_err_intr(struct nitrox_device *ndev) { u64 value; value = nitrox_read_csr(ndev, PEM0_INT); nitrox_write_csr(ndev, PEM0_INT, value); dev_err_ratelimited(DEV(ndev), "PEM(0)_INT 0x%016llx\n", value); } static void clear_lbc_err_intr(struct nitrox_device *ndev) { union lbc_int lbc_int; u64 value, offset; int i; lbc_int.value = nitrox_read_csr(ndev, LBC_INT); dev_err_ratelimited(DEV(ndev), "LBC_INT 0x%016llx\n", lbc_int.value); if (lbc_int.s.dma_rd_err) { for (i = 0; i < NR_CLUSTERS; i++) { offset = EFL_CORE_VF_ERR_INT0X(i); value = nitrox_read_csr(ndev, offset); nitrox_write_csr(ndev, offset, value); offset = EFL_CORE_VF_ERR_INT1X(i); value = nitrox_read_csr(ndev, offset); nitrox_write_csr(ndev, offset, value); } } if (lbc_int.s.cam_soft_err) { dev_err_ratelimited(DEV(ndev), "CAM_SOFT_ERR, invalidating LBC\n"); invalidate_lbc(ndev); } if (lbc_int.s.pref_dat_len_mismatch_err) { offset = LBC_PLM_VF1_64_INT; value = nitrox_read_csr(ndev, offset); nitrox_write_csr(ndev, offset, value); offset = LBC_PLM_VF65_128_INT; value = nitrox_read_csr(ndev, offset); nitrox_write_csr(ndev, offset, value); } if (lbc_int.s.rd_dat_len_mismatch_err) { offset = LBC_ELM_VF1_64_INT; value = nitrox_read_csr(ndev, offset); nitrox_write_csr(ndev, offset, value); offset = LBC_ELM_VF65_128_INT; value = nitrox_read_csr(ndev, offset); nitrox_write_csr(ndev, offset, value); } nitrox_write_csr(ndev, LBC_INT, lbc_int.value); } static void clear_efl_err_intr(struct nitrox_device *ndev) { int i; for (i = 0; i < NR_CLUSTERS; i++) { union efl_core_int core_int; u64 value, offset; offset = EFL_CORE_INTX(i); core_int.value = nitrox_read_csr(ndev, offset); nitrox_write_csr(ndev, offset, core_int.value); dev_err_ratelimited(DEV(ndev), "ELF_CORE(%d)_INT 0x%016llx\n", i, core_int.value); if (core_int.s.se_err) { offset = EFL_CORE_SE_ERR_INTX(i); value = nitrox_read_csr(ndev, offset); nitrox_write_csr(ndev, offset, value); } } } static void clear_bmi_err_intr(struct nitrox_device *ndev) { u64 value; value = nitrox_read_csr(ndev, BMI_INT); nitrox_write_csr(ndev, BMI_INT, value); dev_err_ratelimited(DEV(ndev), "BMI_INT 0x%016llx\n", value); } static void nps_core_int_tasklet(unsigned long data) { struct nitrox_q_vector *qvec = (void *)(uintptr_t)(data); struct nitrox_device *ndev = qvec->ndev; /* if pf mode do queue recovery */ if (ndev->mode == __NDEV_MODE_PF) { } else { /** * if VF(s) enabled communicate the error information * to VF(s) */ } } /** * nps_core_int_isr - interrupt handler for NITROX errors and * mailbox communication */ static irqreturn_t nps_core_int_isr(int irq, void *data) { struct nitrox_q_vector *qvec = data; struct nitrox_device *ndev = qvec->ndev; union nps_core_int_active core_int; core_int.value = nitrox_read_csr(ndev, NPS_CORE_INT_ACTIVE); if (core_int.s.nps_core) clear_nps_core_err_intr(ndev); if (core_int.s.nps_pkt) clear_nps_pkt_err_intr(ndev); if (core_int.s.pom) clear_pom_err_intr(ndev); if (core_int.s.pem) clear_pem_err_intr(ndev); if (core_int.s.lbc) clear_lbc_err_intr(ndev); if (core_int.s.efl) clear_efl_err_intr(ndev); if (core_int.s.bmi) clear_bmi_err_intr(ndev); /* Mailbox interrupt */ if (core_int.s.mbox) nitrox_pf2vf_mbox_handler(ndev); /* If more work callback the ISR, set resend */ core_int.s.resend = 1; nitrox_write_csr(ndev, NPS_CORE_INT_ACTIVE, core_int.value); return IRQ_HANDLED; } void nitrox_unregister_interrupts(struct nitrox_device *ndev) { struct pci_dev *pdev = ndev->pdev; int i; for (i = 0; i < ndev->num_vecs; i++) { struct nitrox_q_vector *qvec; int vec; qvec = ndev->qvec + i; if (!qvec->valid) continue; /* get the vector number */ vec = pci_irq_vector(pdev, i); irq_set_affinity_hint(vec, NULL); free_irq(vec, qvec); tasklet_disable(&qvec->resp_tasklet); tasklet_kill(&qvec->resp_tasklet); qvec->valid = false; } kfree(ndev->qvec); ndev->qvec = NULL; pci_free_irq_vectors(pdev); } int nitrox_register_interrupts(struct nitrox_device *ndev) { struct pci_dev *pdev = ndev->pdev; struct nitrox_q_vector *qvec; int nr_vecs, vec, cpu; int ret, i; /* * PF MSI-X vectors * * Entry 0: NPS PKT ring 0 * Entry 1: AQMQ ring 0 * Entry 2: ZQM ring 0 * Entry 3: NPS PKT ring 1 * Entry 4: AQMQ ring 1 * Entry 5: ZQM ring 1 * .... * Entry 192: NPS_CORE_INT_ACTIVE */ nr_vecs = pci_msix_vec_count(pdev); /* Enable MSI-X */ ret = pci_alloc_irq_vectors(pdev, nr_vecs, nr_vecs, PCI_IRQ_MSIX); if (ret < 0) { dev_err(DEV(ndev), "msix vectors %d alloc failed\n", nr_vecs); return ret; } ndev->num_vecs = nr_vecs; ndev->qvec = kcalloc(nr_vecs, sizeof(*qvec), GFP_KERNEL); if (!ndev->qvec) { pci_free_irq_vectors(pdev); return -ENOMEM; } /* request irqs for packet rings/ports */ for (i = PKT_RING_MSIX_BASE; i < (nr_vecs - 1); i += NR_RING_VECTORS) { qvec = &ndev->qvec[i]; qvec->ring = i / NR_RING_VECTORS; if (qvec->ring >= ndev->nr_queues) break; qvec->cmdq = &ndev->pkt_inq[qvec->ring]; snprintf(qvec->name, IRQ_NAMESZ, "nitrox-pkt%d", qvec->ring); /* get the vector number */ vec = pci_irq_vector(pdev, i); ret = request_irq(vec, nps_pkt_slc_isr, 0, qvec->name, qvec); if (ret) { dev_err(DEV(ndev), "irq failed for pkt ring/port%d\n", qvec->ring); goto irq_fail; } cpu = qvec->ring % num_online_cpus(); irq_set_affinity_hint(vec, get_cpu_mask(cpu)); tasklet_init(&qvec->resp_tasklet, pkt_slc_resp_tasklet, (unsigned long)qvec); qvec->valid = true; } /* request irqs for non ring vectors */ i = NON_RING_MSIX_BASE; qvec = &ndev->qvec[i]; qvec->ndev = ndev; snprintf(qvec->name, IRQ_NAMESZ, "nitrox-core-int%d", i); /* get the vector number */ vec = pci_irq_vector(pdev, i); ret = request_irq(vec, nps_core_int_isr, 0, qvec->name, qvec); if (ret) { dev_err(DEV(ndev), "irq failed for nitrox-core-int%d\n", i); goto irq_fail; } cpu = num_online_cpus(); irq_set_affinity_hint(vec, get_cpu_mask(cpu)); tasklet_init(&qvec->resp_tasklet, nps_core_int_tasklet, (unsigned long)qvec); qvec->valid = true; return 0; irq_fail: nitrox_unregister_interrupts(ndev); return ret; } void nitrox_sriov_unregister_interrupts(struct nitrox_device *ndev) { struct pci_dev *pdev = ndev->pdev; int i; for (i = 0; i < ndev->num_vecs; i++) { struct nitrox_q_vector *qvec; int vec; qvec = ndev->qvec + i; if (!qvec->valid) continue; vec = ndev->iov.msix.vector; irq_set_affinity_hint(vec, NULL); free_irq(vec, qvec); tasklet_disable(&qvec->resp_tasklet); tasklet_kill(&qvec->resp_tasklet); qvec->valid = false; } kfree(ndev->qvec); ndev->qvec = NULL; pci_disable_msix(pdev); } int nitrox_sriov_register_interupts(struct nitrox_device *ndev) { struct pci_dev *pdev = ndev->pdev; struct nitrox_q_vector *qvec; int vec, cpu; int ret; /** * only non ring vectors i.e Entry 192 is available * for PF in SR-IOV mode. */ ndev->iov.msix.entry = NON_RING_MSIX_BASE; ret = pci_enable_msix_exact(pdev, &ndev->iov.msix, NR_NON_RING_VECTORS); if (ret) { dev_err(DEV(ndev), "failed to allocate nps-core-int%d\n", NON_RING_MSIX_BASE); return ret; } qvec = kcalloc(NR_NON_RING_VECTORS, sizeof(*qvec), GFP_KERNEL); if (!qvec) { pci_disable_msix(pdev); return -ENOMEM; } qvec->ndev = ndev; ndev->qvec = qvec; ndev->num_vecs = NR_NON_RING_VECTORS; snprintf(qvec->name, IRQ_NAMESZ, "nitrox-core-int%d", NON_RING_MSIX_BASE); vec = ndev->iov.msix.vector; ret = request_irq(vec, nps_core_int_isr, 0, qvec->name, qvec); if (ret) { dev_err(DEV(ndev), "irq failed for nitrox-core-int%d\n", NON_RING_MSIX_BASE); goto iov_irq_fail; } cpu = num_online_cpus(); irq_set_affinity_hint(vec, get_cpu_mask(cpu)); tasklet_init(&qvec->resp_tasklet, nps_core_int_tasklet, (unsigned long)qvec); qvec->valid = true; return 0; iov_irq_fail: nitrox_sriov_unregister_interrupts(ndev); return ret; }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1