Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Giridhar Malavali | 12283 | 74.74% | 11 | 6.32% |
Armen Baloyan | 913 | 5.56% | 16 | 9.20% |
Andrew Vasquez | 690 | 4.20% | 48 | 27.59% |
James Bottomley | 637 | 3.88% | 5 | 2.87% |
Saurav Kashyap | 552 | 3.36% | 10 | 5.75% |
Anirban Chakraborty | 531 | 3.23% | 6 | 3.45% |
Bart Van Assche | 244 | 1.48% | 22 | 12.64% |
Seokmann Ju | 133 | 0.81% | 4 | 2.30% |
Chad Dupuis | 80 | 0.49% | 9 | 5.17% |
Joe Carnuccio | 44 | 0.27% | 4 | 2.30% |
Sarang Radke | 39 | 0.24% | 1 | 0.57% |
Madhuranath Iyengar | 32 | 0.19% | 2 | 1.15% |
Johannes Thumshirn | 27 | 0.16% | 2 | 1.15% |
FUJITA Tomonori | 26 | 0.16% | 1 | 0.57% |
Harish Zunjarrao | 26 | 0.16% | 2 | 1.15% |
Harihara Kadayam | 25 | 0.15% | 1 | 0.57% |
Lalit Chandivade | 23 | 0.14% | 3 | 1.72% |
Jan Kulich | 20 | 0.12% | 1 | 0.57% |
Michael Hernandez | 18 | 0.11% | 2 | 1.15% |
Daniel Wagner | 17 | 0.10% | 1 | 0.57% |
Christoph Hellwig | 11 | 0.07% | 3 | 1.72% |
Azeem Shaikh | 10 | 0.06% | 1 | 0.57% |
Himanshu Madhani | 6 | 0.04% | 1 | 0.57% |
Ben Hutchings | 6 | 0.04% | 1 | 0.57% |
Tina Ruchandani | 5 | 0.03% | 1 | 0.57% |
gurinder.shergill@hp.com | 5 | 0.03% | 1 | 0.57% |
Yijing Wang | 4 | 0.02% | 1 | 0.57% |
Yani Ioannou | 4 | 0.02% | 1 | 0.57% |
Hannes Reinecke | 4 | 0.02% | 1 | 0.57% |
Quinn Tran | 3 | 0.02% | 1 | 0.57% |
James Smart | 3 | 0.02% | 1 | 0.57% |
Atul Deshmukh | 2 | 0.01% | 1 | 0.57% |
Lee Jones | 2 | 0.01% | 1 | 0.57% |
Oleksandr Khoshaba | 2 | 0.01% | 1 | 0.57% |
Arun Easi | 2 | 0.01% | 2 | 1.15% |
Thomas Gleixner | 2 | 0.01% | 1 | 0.57% |
Masanari Iida | 1 | 0.01% | 1 | 0.57% |
Jason Yan | 1 | 0.01% | 1 | 0.57% |
Joe Lawrence | 1 | 0.01% | 1 | 0.57% |
Tony Jones | 1 | 0.01% | 1 | 0.57% |
Total | 16435 | 174 |
// SPDX-License-Identifier: GPL-2.0-only /* * QLogic Fibre Channel HBA Driver * Copyright (c) 2003-2014 QLogic Corporation */ #include "qla_def.h" #include <linux/delay.h> #include <linux/ktime.h> #include <linux/pci.h> #include <linux/ratelimit.h> #include <linux/vmalloc.h> #include <scsi/scsi_tcq.h> #include <linux/utsname.h> /* QLAFX00 specific Mailbox implementation functions */ /* * qlafx00_mailbox_command * Issue mailbox command and waits for completion. * * Input: * ha = adapter block pointer. * mcp = driver internal mbx struct pointer. * * Output: * mb[MAX_MAILBOX_REGISTER_COUNT] = returned mailbox data. * * Returns: * 0 : QLA_SUCCESS = cmd performed success * 1 : QLA_FUNCTION_FAILED (error encountered) * 6 : QLA_FUNCTION_TIMEOUT (timeout condition encountered) * * Context: * Kernel context. */ static int qlafx00_mailbox_command(scsi_qla_host_t *vha, struct mbx_cmd_32 *mcp) { int rval; unsigned long flags = 0; device_reg_t *reg; uint8_t abort_active; uint8_t io_lock_on; uint16_t command = 0; uint32_t *iptr; __le32 __iomem *optr; uint32_t cnt; uint32_t mboxes; unsigned long wait_time; struct qla_hw_data *ha = vha->hw; scsi_qla_host_t *base_vha = pci_get_drvdata(ha->pdev); if (ha->pdev->error_state == pci_channel_io_perm_failure) { ql_log(ql_log_warn, vha, 0x115c, "PCI channel failed permanently, exiting.\n"); return QLA_FUNCTION_TIMEOUT; } if (vha->device_flags & DFLG_DEV_FAILED) { ql_log(ql_log_warn, vha, 0x115f, "Device in failed state, exiting.\n"); return QLA_FUNCTION_TIMEOUT; } reg = ha->iobase; io_lock_on = base_vha->flags.init_done; rval = QLA_SUCCESS; abort_active = test_bit(ABORT_ISP_ACTIVE, &base_vha->dpc_flags); if (ha->flags.pci_channel_io_perm_failure) { ql_log(ql_log_warn, vha, 0x1175, "Perm failure on EEH timeout MBX, exiting.\n"); return QLA_FUNCTION_TIMEOUT; } if (ha->flags.isp82xx_fw_hung) { /* Setting Link-Down error */ mcp->mb[0] = MBS_LINK_DOWN_ERROR; ql_log(ql_log_warn, vha, 0x1176, "FW hung = %d.\n", ha->flags.isp82xx_fw_hung); rval = QLA_FUNCTION_FAILED; goto premature_exit; } /* * Wait for active mailbox commands to finish by waiting at most tov * seconds. This is to serialize actual issuing of mailbox cmds during * non ISP abort time. */ if (!wait_for_completion_timeout(&ha->mbx_cmd_comp, mcp->tov * HZ)) { /* Timeout occurred. Return error. */ ql_log(ql_log_warn, vha, 0x1177, "Cmd access timeout, cmd=0x%x, Exiting.\n", mcp->mb[0]); return QLA_FUNCTION_TIMEOUT; } ha->flags.mbox_busy = 1; /* Save mailbox command for debug */ ha->mcp32 = mcp; ql_dbg(ql_dbg_mbx, vha, 0x1178, "Prepare to issue mbox cmd=0x%x.\n", mcp->mb[0]); spin_lock_irqsave(&ha->hardware_lock, flags); /* Load mailbox registers. */ optr = ®->ispfx00.mailbox0; iptr = mcp->mb; command = mcp->mb[0]; mboxes = mcp->out_mb; for (cnt = 0; cnt < ha->mbx_count; cnt++) { if (mboxes & BIT_0) wrt_reg_dword(optr, *iptr); mboxes >>= 1; optr++; iptr++; } /* Issue set host interrupt command to send cmd out. */ ha->flags.mbox_int = 0; clear_bit(MBX_INTERRUPT, &ha->mbx_cmd_flags); ql_dump_buffer(ql_dbg_mbx + ql_dbg_buffer, vha, 0x1172, (uint8_t *)mcp->mb, 16); ql_dump_buffer(ql_dbg_mbx + ql_dbg_buffer, vha, 0x1173, ((uint8_t *)mcp->mb + 0x10), 16); ql_dump_buffer(ql_dbg_mbx + ql_dbg_buffer, vha, 0x1174, ((uint8_t *)mcp->mb + 0x20), 8); /* Unlock mbx registers and wait for interrupt */ ql_dbg(ql_dbg_mbx, vha, 0x1179, "Going to unlock irq & waiting for interrupts. " "jiffies=%lx.\n", jiffies); /* Wait for mbx cmd completion until timeout */ if ((!abort_active && io_lock_on) || IS_NOPOLLING_TYPE(ha)) { set_bit(MBX_INTR_WAIT, &ha->mbx_cmd_flags); QLAFX00_SET_HST_INTR(ha, ha->mbx_intr_code); spin_unlock_irqrestore(&ha->hardware_lock, flags); WARN_ON_ONCE(wait_for_completion_timeout(&ha->mbx_intr_comp, mcp->tov * HZ) != 0); } else { ql_dbg(ql_dbg_mbx, vha, 0x112c, "Cmd=%x Polling Mode.\n", command); QLAFX00_SET_HST_INTR(ha, ha->mbx_intr_code); spin_unlock_irqrestore(&ha->hardware_lock, flags); wait_time = jiffies + mcp->tov * HZ; /* wait at most tov secs */ while (!ha->flags.mbox_int) { if (time_after(jiffies, wait_time)) break; /* Check for pending interrupts. */ qla2x00_poll(ha->rsp_q_map[0]); if (!ha->flags.mbox_int && !(IS_QLA2200(ha) && command == MBC_LOAD_RISC_RAM_EXTENDED)) usleep_range(10000, 11000); } /* while */ ql_dbg(ql_dbg_mbx, vha, 0x112d, "Waited %d sec.\n", (uint)((jiffies - (wait_time - (mcp->tov * HZ)))/HZ)); } /* Check whether we timed out */ if (ha->flags.mbox_int) { uint32_t *iptr2; ql_dbg(ql_dbg_mbx, vha, 0x112e, "Cmd=%x completed.\n", command); /* Got interrupt. Clear the flag. */ ha->flags.mbox_int = 0; clear_bit(MBX_INTERRUPT, &ha->mbx_cmd_flags); if (ha->mailbox_out32[0] != MBS_COMMAND_COMPLETE) rval = QLA_FUNCTION_FAILED; /* Load return mailbox registers. */ iptr2 = mcp->mb; iptr = (uint32_t *)&ha->mailbox_out32[0]; mboxes = mcp->in_mb; for (cnt = 0; cnt < ha->mbx_count; cnt++) { if (mboxes & BIT_0) *iptr2 = *iptr; mboxes >>= 1; iptr2++; iptr++; } } else { rval = QLA_FUNCTION_TIMEOUT; } ha->flags.mbox_busy = 0; /* Clean up */ ha->mcp32 = NULL; if ((abort_active || !io_lock_on) && !IS_NOPOLLING_TYPE(ha)) { ql_dbg(ql_dbg_mbx, vha, 0x113a, "checking for additional resp interrupt.\n"); /* polling mode for non isp_abort commands. */ qla2x00_poll(ha->rsp_q_map[0]); } if (rval == QLA_FUNCTION_TIMEOUT && mcp->mb[0] != MBC_GEN_SYSTEM_ERROR) { if (!io_lock_on || (mcp->flags & IOCTL_CMD) || ha->flags.eeh_busy) { /* not in dpc. schedule it for dpc to take over. */ ql_dbg(ql_dbg_mbx, vha, 0x115d, "Timeout, schedule isp_abort_needed.\n"); if (!test_bit(ISP_ABORT_NEEDED, &vha->dpc_flags) && !test_bit(ABORT_ISP_ACTIVE, &vha->dpc_flags) && !test_bit(ISP_ABORT_RETRY, &vha->dpc_flags)) { ql_log(ql_log_info, base_vha, 0x115e, "Mailbox cmd timeout occurred, cmd=0x%x, " "mb[0]=0x%x, eeh_busy=0x%x. Scheduling ISP " "abort.\n", command, mcp->mb[0], ha->flags.eeh_busy); set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags); qla2xxx_wake_dpc(vha); } } else if (!abort_active) { /* call abort directly since we are in the DPC thread */ ql_dbg(ql_dbg_mbx, vha, 0x1160, "Timeout, calling abort_isp.\n"); if (!test_bit(ISP_ABORT_NEEDED, &vha->dpc_flags) && !test_bit(ABORT_ISP_ACTIVE, &vha->dpc_flags) && !test_bit(ISP_ABORT_RETRY, &vha->dpc_flags)) { ql_log(ql_log_info, base_vha, 0x1161, "Mailbox cmd timeout occurred, cmd=0x%x, " "mb[0]=0x%x. Scheduling ISP abort ", command, mcp->mb[0]); set_bit(ABORT_ISP_ACTIVE, &vha->dpc_flags); clear_bit(ISP_ABORT_NEEDED, &vha->dpc_flags); if (ha->isp_ops->abort_isp(vha)) { /* Failed. retry later. */ set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags); } clear_bit(ABORT_ISP_ACTIVE, &vha->dpc_flags); ql_dbg(ql_dbg_mbx, vha, 0x1162, "Finished abort_isp.\n"); } } } premature_exit: /* Allow next mbx cmd to come in. */ complete(&ha->mbx_cmd_comp); if (rval) { ql_log(ql_log_warn, base_vha, 0x1163, "**** Failed=%x mbx[0]=%x, mb[1]=%x, mb[2]=%x, mb[3]=%x, cmd=%x ****.\n", rval, mcp->mb[0], mcp->mb[1], mcp->mb[2], mcp->mb[3], command); } else { ql_dbg(ql_dbg_mbx, base_vha, 0x1164, "Done %s.\n", __func__); } return rval; } /* * qlafx00_driver_shutdown * Indicate a driver shutdown to firmware. * * Input: * ha = adapter block pointer. * * Returns: * local function return status code. * * Context: * Kernel context. */ int qlafx00_driver_shutdown(scsi_qla_host_t *vha, int tmo) { int rval; struct mbx_cmd_32 mc; struct mbx_cmd_32 *mcp = &mc; ql_dbg(ql_dbg_mbx + ql_dbg_verbose, vha, 0x1166, "Entered %s.\n", __func__); mcp->mb[0] = MBC_MR_DRV_SHUTDOWN; mcp->out_mb = MBX_0; mcp->in_mb = MBX_0; if (tmo) mcp->tov = tmo; else mcp->tov = MBX_TOV_SECONDS; mcp->flags = 0; rval = qlafx00_mailbox_command(vha, mcp); if (rval != QLA_SUCCESS) { ql_dbg(ql_dbg_mbx, vha, 0x1167, "Failed=%x.\n", rval); } else { ql_dbg(ql_dbg_mbx + ql_dbg_verbose, vha, 0x1168, "Done %s.\n", __func__); } return rval; } /* * qlafx00_get_firmware_state * Get adapter firmware state. * * Input: * ha = adapter block pointer. * TARGET_QUEUE_LOCK must be released. * ADAPTER_STATE_LOCK must be released. * * Returns: * qla7xxx local function return status code. * * Context: * Kernel context. */ static int qlafx00_get_firmware_state(scsi_qla_host_t *vha, uint32_t *states) { int rval; struct mbx_cmd_32 mc; struct mbx_cmd_32 *mcp = &mc; ql_dbg(ql_dbg_mbx + ql_dbg_verbose, vha, 0x1169, "Entered %s.\n", __func__); mcp->mb[0] = MBC_GET_FIRMWARE_STATE; mcp->out_mb = MBX_0; mcp->in_mb = MBX_1|MBX_0; mcp->tov = MBX_TOV_SECONDS; mcp->flags = 0; rval = qlafx00_mailbox_command(vha, mcp); /* Return firmware states. */ states[0] = mcp->mb[1]; if (rval != QLA_SUCCESS) { ql_dbg(ql_dbg_mbx, vha, 0x116a, "Failed=%x mb[0]=%x.\n", rval, mcp->mb[0]); } else { ql_dbg(ql_dbg_mbx + ql_dbg_verbose, vha, 0x116b, "Done %s.\n", __func__); } return rval; } /* * qlafx00_init_firmware * Initialize adapter firmware. * * Input: * ha = adapter block pointer. * dptr = Initialization control block pointer. * size = size of initialization control block. * TARGET_QUEUE_LOCK must be released. * ADAPTER_STATE_LOCK must be released. * * Returns: * qlafx00 local function return status code. * * Context: * Kernel context. */ int qlafx00_init_firmware(scsi_qla_host_t *vha, uint16_t size) { int rval; struct mbx_cmd_32 mc; struct mbx_cmd_32 *mcp = &mc; struct qla_hw_data *ha = vha->hw; ql_dbg(ql_dbg_mbx + ql_dbg_verbose, vha, 0x116c, "Entered %s.\n", __func__); mcp->mb[0] = MBC_INITIALIZE_FIRMWARE; mcp->mb[1] = 0; mcp->mb[2] = MSD(ha->init_cb_dma); mcp->mb[3] = LSD(ha->init_cb_dma); mcp->out_mb = MBX_3|MBX_2|MBX_1|MBX_0; mcp->in_mb = MBX_0; mcp->buf_size = size; mcp->flags = MBX_DMA_OUT; mcp->tov = MBX_TOV_SECONDS; rval = qlafx00_mailbox_command(vha, mcp); if (rval != QLA_SUCCESS) { ql_dbg(ql_dbg_mbx, vha, 0x116d, "Failed=%x mb[0]=%x.\n", rval, mcp->mb[0]); } else { ql_dbg(ql_dbg_mbx + ql_dbg_verbose, vha, 0x116e, "Done %s.\n", __func__); } return rval; } /* * qlafx00_mbx_reg_test */ static int qlafx00_mbx_reg_test(scsi_qla_host_t *vha) { int rval; struct mbx_cmd_32 mc; struct mbx_cmd_32 *mcp = &mc; ql_dbg(ql_dbg_mbx + ql_dbg_verbose, vha, 0x116f, "Entered %s.\n", __func__); mcp->mb[0] = MBC_MAILBOX_REGISTER_TEST; mcp->mb[1] = 0xAAAA; mcp->mb[2] = 0x5555; mcp->mb[3] = 0xAA55; mcp->mb[4] = 0x55AA; mcp->mb[5] = 0xA5A5; mcp->mb[6] = 0x5A5A; mcp->mb[7] = 0x2525; mcp->mb[8] = 0xBBBB; mcp->mb[9] = 0x6666; mcp->mb[10] = 0xBB66; mcp->mb[11] = 0x66BB; mcp->mb[12] = 0xB6B6; mcp->mb[13] = 0x6B6B; mcp->mb[14] = 0x3636; mcp->mb[15] = 0xCCCC; mcp->out_mb = MBX_15|MBX_14|MBX_13|MBX_12|MBX_11|MBX_10|MBX_9|MBX_8| MBX_7|MBX_6|MBX_5|MBX_4|MBX_3|MBX_2|MBX_1|MBX_0; mcp->in_mb = MBX_15|MBX_14|MBX_13|MBX_12|MBX_11|MBX_10|MBX_9|MBX_8| MBX_7|MBX_6|MBX_5|MBX_4|MBX_3|MBX_2|MBX_1|MBX_0; mcp->buf_size = 0; mcp->flags = MBX_DMA_OUT; mcp->tov = MBX_TOV_SECONDS; rval = qlafx00_mailbox_command(vha, mcp); if (rval == QLA_SUCCESS) { if (mcp->mb[17] != 0xAAAA || mcp->mb[18] != 0x5555 || mcp->mb[19] != 0xAA55 || mcp->mb[20] != 0x55AA) rval = QLA_FUNCTION_FAILED; if (mcp->mb[21] != 0xA5A5 || mcp->mb[22] != 0x5A5A || mcp->mb[23] != 0x2525 || mcp->mb[24] != 0xBBBB) rval = QLA_FUNCTION_FAILED; if (mcp->mb[25] != 0x6666 || mcp->mb[26] != 0xBB66 || mcp->mb[27] != 0x66BB || mcp->mb[28] != 0xB6B6) rval = QLA_FUNCTION_FAILED; if (mcp->mb[29] != 0x6B6B || mcp->mb[30] != 0x3636 || mcp->mb[31] != 0xCCCC) rval = QLA_FUNCTION_FAILED; } if (rval != QLA_SUCCESS) { ql_dbg(ql_dbg_mbx, vha, 0x1170, "Failed=%x mb[0]=%x.\n", rval, mcp->mb[0]); } else { ql_dbg(ql_dbg_mbx + ql_dbg_verbose, vha, 0x1171, "Done %s.\n", __func__); } return rval; } /** * qlafx00_pci_config() - Setup ISPFx00 PCI configuration registers. * @vha: HA context * * Returns 0 on success. */ int qlafx00_pci_config(scsi_qla_host_t *vha) { uint16_t w; struct qla_hw_data *ha = vha->hw; pci_set_master(ha->pdev); pci_try_set_mwi(ha->pdev); pci_read_config_word(ha->pdev, PCI_COMMAND, &w); w |= (PCI_COMMAND_PARITY | PCI_COMMAND_SERR); w &= ~PCI_COMMAND_INTX_DISABLE; pci_write_config_word(ha->pdev, PCI_COMMAND, w); /* PCIe -- adjust Maximum Read Request Size (2048). */ if (pci_is_pcie(ha->pdev)) pcie_set_readrq(ha->pdev, 2048); ha->chip_revision = ha->pdev->revision; return QLA_SUCCESS; } /** * qlafx00_soc_cpu_reset() - Perform warm reset of iSA(CPUs being reset on SOC). * @vha: HA context * */ static inline void qlafx00_soc_cpu_reset(scsi_qla_host_t *vha) { unsigned long flags = 0; struct qla_hw_data *ha = vha->hw; int i, core; uint32_t cnt; uint32_t reg_val; spin_lock_irqsave(&ha->hardware_lock, flags); QLAFX00_SET_HBA_SOC_REG(ha, 0x80004, 0); QLAFX00_SET_HBA_SOC_REG(ha, 0x82004, 0); /* stop the XOR DMA engines */ QLAFX00_SET_HBA_SOC_REG(ha, 0x60920, 0x02); QLAFX00_SET_HBA_SOC_REG(ha, 0x60924, 0x02); QLAFX00_SET_HBA_SOC_REG(ha, 0xf0920, 0x02); QLAFX00_SET_HBA_SOC_REG(ha, 0xf0924, 0x02); /* stop the IDMA engines */ reg_val = QLAFX00_GET_HBA_SOC_REG(ha, 0x60840); reg_val &= ~(1<<12); QLAFX00_SET_HBA_SOC_REG(ha, 0x60840, reg_val); reg_val = QLAFX00_GET_HBA_SOC_REG(ha, 0x60844); reg_val &= ~(1<<12); QLAFX00_SET_HBA_SOC_REG(ha, 0x60844, reg_val); reg_val = QLAFX00_GET_HBA_SOC_REG(ha, 0x60848); reg_val &= ~(1<<12); QLAFX00_SET_HBA_SOC_REG(ha, 0x60848, reg_val); reg_val = QLAFX00_GET_HBA_SOC_REG(ha, 0x6084C); reg_val &= ~(1<<12); QLAFX00_SET_HBA_SOC_REG(ha, 0x6084C, reg_val); for (i = 0; i < 100000; i++) { if ((QLAFX00_GET_HBA_SOC_REG(ha, 0xd0000) & 0x10000000) == 0 && (QLAFX00_GET_HBA_SOC_REG(ha, 0x10600) & 0x1) == 0) break; udelay(100); } /* Set all 4 cores in reset */ for (i = 0; i < 4; i++) { QLAFX00_SET_HBA_SOC_REG(ha, (SOC_SW_RST_CONTROL_REG_CORE0 + 8*i), (0xF01)); QLAFX00_SET_HBA_SOC_REG(ha, (SOC_SW_RST_CONTROL_REG_CORE0 + 4 + 8*i), (0x01010101)); } /* Reset all units in Fabric */ QLAFX00_SET_HBA_SOC_REG(ha, SOC_FABRIC_RST_CONTROL_REG, (0x011f0101)); /* */ QLAFX00_SET_HBA_SOC_REG(ha, 0x10610, 1); QLAFX00_SET_HBA_SOC_REG(ha, 0x10600, 0); /* Set all 4 core Memory Power Down Registers */ for (i = 0; i < 5; i++) { QLAFX00_SET_HBA_SOC_REG(ha, (SOC_PWR_MANAGEMENT_PWR_DOWN_REG + 4*i), (0x0)); } /* Reset all interrupt control registers */ for (i = 0; i < 115; i++) { QLAFX00_SET_HBA_SOC_REG(ha, (SOC_INTERRUPT_SOURCE_I_CONTROL_REG + 4*i), (0x0)); } /* Reset Timers control registers. per core */ for (core = 0; core < 4; core++) for (i = 0; i < 8; i++) QLAFX00_SET_HBA_SOC_REG(ha, (SOC_CORE_TIMER_REG + 0x100*core + 4*i), (0x0)); /* Reset per core IRQ ack register */ for (core = 0; core < 4; core++) QLAFX00_SET_HBA_SOC_REG(ha, (SOC_IRQ_ACK_REG + 0x100*core), (0x3FF)); /* Set Fabric control and config to defaults */ QLAFX00_SET_HBA_SOC_REG(ha, SOC_FABRIC_CONTROL_REG, (0x2)); QLAFX00_SET_HBA_SOC_REG(ha, SOC_FABRIC_CONFIG_REG, (0x3)); /* Kick in Fabric units */ QLAFX00_SET_HBA_SOC_REG(ha, SOC_FABRIC_RST_CONTROL_REG, (0x0)); /* Kick in Core0 to start boot process */ QLAFX00_SET_HBA_SOC_REG(ha, SOC_SW_RST_CONTROL_REG_CORE0, (0xF00)); spin_unlock_irqrestore(&ha->hardware_lock, flags); /* Wait 10secs for soft-reset to complete. */ for (cnt = 10; cnt; cnt--) { msleep(1000); barrier(); } } /** * qlafx00_soft_reset() - Soft Reset ISPFx00. * @vha: HA context * * Returns 0 on success. */ int qlafx00_soft_reset(scsi_qla_host_t *vha) { struct qla_hw_data *ha = vha->hw; int rval = QLA_FUNCTION_FAILED; if (unlikely(pci_channel_offline(ha->pdev) && ha->flags.pci_channel_io_perm_failure)) return rval; ha->isp_ops->disable_intrs(ha); qlafx00_soc_cpu_reset(vha); return QLA_SUCCESS; } /** * qlafx00_chip_diag() - Test ISPFx00 for proper operation. * @vha: HA context * * Returns 0 on success. */ int qlafx00_chip_diag(scsi_qla_host_t *vha) { int rval = 0; struct qla_hw_data *ha = vha->hw; struct req_que *req = ha->req_q_map[0]; ha->fw_transfer_size = REQUEST_ENTRY_SIZE * req->length; rval = qlafx00_mbx_reg_test(vha); if (rval) { ql_log(ql_log_warn, vha, 0x1165, "Failed mailbox send register test\n"); } else { /* Flag a successful rval */ rval = QLA_SUCCESS; } return rval; } void qlafx00_config_rings(struct scsi_qla_host *vha) { struct qla_hw_data *ha = vha->hw; struct device_reg_fx00 __iomem *reg = &ha->iobase->ispfx00; wrt_reg_dword(®->req_q_in, 0); wrt_reg_dword(®->req_q_out, 0); wrt_reg_dword(®->rsp_q_in, 0); wrt_reg_dword(®->rsp_q_out, 0); /* PCI posting */ rd_reg_dword(®->rsp_q_out); } char * qlafx00_pci_info_str(struct scsi_qla_host *vha, char *str, size_t str_len) { struct qla_hw_data *ha = vha->hw; if (pci_is_pcie(ha->pdev)) strscpy(str, "PCIe iSA", str_len); return str; } char * qlafx00_fw_version_str(struct scsi_qla_host *vha, char *str, size_t size) { struct qla_hw_data *ha = vha->hw; snprintf(str, size, "%s", ha->mr.fw_version); return str; } void qlafx00_enable_intrs(struct qla_hw_data *ha) { unsigned long flags = 0; spin_lock_irqsave(&ha->hardware_lock, flags); ha->interrupts_on = 1; QLAFX00_ENABLE_ICNTRL_REG(ha); spin_unlock_irqrestore(&ha->hardware_lock, flags); } void qlafx00_disable_intrs(struct qla_hw_data *ha) { unsigned long flags = 0; spin_lock_irqsave(&ha->hardware_lock, flags); ha->interrupts_on = 0; QLAFX00_DISABLE_ICNTRL_REG(ha); spin_unlock_irqrestore(&ha->hardware_lock, flags); } int qlafx00_abort_target(fc_port_t *fcport, uint64_t l, int tag) { return qla2x00_async_tm_cmd(fcport, TCF_TARGET_RESET, l, tag); } int qlafx00_lun_reset(fc_port_t *fcport, uint64_t l, int tag) { return qla2x00_async_tm_cmd(fcport, TCF_LUN_RESET, l, tag); } int qlafx00_iospace_config(struct qla_hw_data *ha) { if (pci_request_selected_regions(ha->pdev, ha->bars, QLA2XXX_DRIVER_NAME)) { ql_log_pci(ql_log_fatal, ha->pdev, 0x014e, "Failed to reserve PIO/MMIO regions (%s), aborting.\n", pci_name(ha->pdev)); goto iospace_error_exit; } /* Use MMIO operations for all accesses. */ if (!(pci_resource_flags(ha->pdev, 0) & IORESOURCE_MEM)) { ql_log_pci(ql_log_warn, ha->pdev, 0x014f, "Invalid pci I/O region size (%s).\n", pci_name(ha->pdev)); goto iospace_error_exit; } if (pci_resource_len(ha->pdev, 0) < BAR0_LEN_FX00) { ql_log_pci(ql_log_warn, ha->pdev, 0x0127, "Invalid PCI mem BAR0 region size (%s), aborting\n", pci_name(ha->pdev)); goto iospace_error_exit; } ha->cregbase = ioremap(pci_resource_start(ha->pdev, 0), BAR0_LEN_FX00); if (!ha->cregbase) { ql_log_pci(ql_log_fatal, ha->pdev, 0x0128, "cannot remap MMIO (%s), aborting\n", pci_name(ha->pdev)); goto iospace_error_exit; } if (!(pci_resource_flags(ha->pdev, 2) & IORESOURCE_MEM)) { ql_log_pci(ql_log_warn, ha->pdev, 0x0129, "region #2 not an MMIO resource (%s), aborting\n", pci_name(ha->pdev)); goto iospace_error_exit; } if (pci_resource_len(ha->pdev, 2) < BAR2_LEN_FX00) { ql_log_pci(ql_log_warn, ha->pdev, 0x012a, "Invalid PCI mem BAR2 region size (%s), aborting\n", pci_name(ha->pdev)); goto iospace_error_exit; } ha->iobase = ioremap(pci_resource_start(ha->pdev, 2), BAR2_LEN_FX00); if (!ha->iobase) { ql_log_pci(ql_log_fatal, ha->pdev, 0x012b, "cannot remap MMIO (%s), aborting\n", pci_name(ha->pdev)); goto iospace_error_exit; } /* Determine queue resources */ ha->max_req_queues = ha->max_rsp_queues = 1; ql_log_pci(ql_log_info, ha->pdev, 0x012c, "Bars 0x%x, iobase0 0x%p, iobase2 0x%p\n", ha->bars, ha->cregbase, ha->iobase); return 0; iospace_error_exit: return -ENOMEM; } static void qlafx00_save_queue_ptrs(struct scsi_qla_host *vha) { struct qla_hw_data *ha = vha->hw; struct req_que *req = ha->req_q_map[0]; struct rsp_que *rsp = ha->rsp_q_map[0]; req->length_fx00 = req->length; req->ring_fx00 = req->ring; req->dma_fx00 = req->dma; rsp->length_fx00 = rsp->length; rsp->ring_fx00 = rsp->ring; rsp->dma_fx00 = rsp->dma; ql_dbg(ql_dbg_init, vha, 0x012d, "req: %p, ring_fx00: %p, length_fx00: 0x%x," "req->dma_fx00: 0x%llx\n", req, req->ring_fx00, req->length_fx00, (u64)req->dma_fx00); ql_dbg(ql_dbg_init, vha, 0x012e, "rsp: %p, ring_fx00: %p, length_fx00: 0x%x," "rsp->dma_fx00: 0x%llx\n", rsp, rsp->ring_fx00, rsp->length_fx00, (u64)rsp->dma_fx00); } static int qlafx00_config_queues(struct scsi_qla_host *vha) { struct qla_hw_data *ha = vha->hw; struct req_que *req = ha->req_q_map[0]; struct rsp_que *rsp = ha->rsp_q_map[0]; dma_addr_t bar2_hdl = pci_resource_start(ha->pdev, 2); req->length = ha->req_que_len; req->ring = (void __force *)ha->iobase + ha->req_que_off; req->dma = bar2_hdl + ha->req_que_off; if ((!req->ring) || (req->length == 0)) { ql_log_pci(ql_log_info, ha->pdev, 0x012f, "Unable to allocate memory for req_ring\n"); return QLA_FUNCTION_FAILED; } ql_dbg(ql_dbg_init, vha, 0x0130, "req: %p req_ring pointer %p req len 0x%x " "req off 0x%x\n, req->dma: 0x%llx", req, req->ring, req->length, ha->req_que_off, (u64)req->dma); rsp->length = ha->rsp_que_len; rsp->ring = (void __force *)ha->iobase + ha->rsp_que_off; rsp->dma = bar2_hdl + ha->rsp_que_off; if ((!rsp->ring) || (rsp->length == 0)) { ql_log_pci(ql_log_info, ha->pdev, 0x0131, "Unable to allocate memory for rsp_ring\n"); return QLA_FUNCTION_FAILED; } ql_dbg(ql_dbg_init, vha, 0x0132, "rsp: %p rsp_ring pointer %p rsp len 0x%x " "rsp off 0x%x, rsp->dma: 0x%llx\n", rsp, rsp->ring, rsp->length, ha->rsp_que_off, (u64)rsp->dma); return QLA_SUCCESS; } static int qlafx00_init_fw_ready(scsi_qla_host_t *vha) { int rval = 0; unsigned long wtime; uint16_t wait_time; /* Wait time */ struct qla_hw_data *ha = vha->hw; struct device_reg_fx00 __iomem *reg = &ha->iobase->ispfx00; uint32_t aenmbx, aenmbx7 = 0; uint32_t pseudo_aen; uint32_t state[5]; bool done = false; /* 30 seconds wait - Adjust if required */ wait_time = 30; pseudo_aen = rd_reg_dword(®->pseudoaen); if (pseudo_aen == 1) { aenmbx7 = rd_reg_dword(®->initval7); ha->mbx_intr_code = MSW(aenmbx7); ha->rqstq_intr_code = LSW(aenmbx7); rval = qlafx00_driver_shutdown(vha, 10); if (rval != QLA_SUCCESS) qlafx00_soft_reset(vha); } /* wait time before firmware ready */ wtime = jiffies + (wait_time * HZ); do { aenmbx = rd_reg_dword(®->aenmailbox0); barrier(); ql_dbg(ql_dbg_mbx, vha, 0x0133, "aenmbx: 0x%x\n", aenmbx); switch (aenmbx) { case MBA_FW_NOT_STARTED: case MBA_FW_STARTING: break; case MBA_SYSTEM_ERR: case MBA_REQ_TRANSFER_ERR: case MBA_RSP_TRANSFER_ERR: case MBA_FW_INIT_FAILURE: qlafx00_soft_reset(vha); break; case MBA_FW_RESTART_CMPLT: /* Set the mbx and rqstq intr code */ aenmbx7 = rd_reg_dword(®->aenmailbox7); ha->mbx_intr_code = MSW(aenmbx7); ha->rqstq_intr_code = LSW(aenmbx7); ha->req_que_off = rd_reg_dword(®->aenmailbox1); ha->rsp_que_off = rd_reg_dword(®->aenmailbox3); ha->req_que_len = rd_reg_dword(®->aenmailbox5); ha->rsp_que_len = rd_reg_dword(®->aenmailbox6); wrt_reg_dword(®->aenmailbox0, 0); rd_reg_dword_relaxed(®->aenmailbox0); ql_dbg(ql_dbg_init, vha, 0x0134, "f/w returned mbx_intr_code: 0x%x, " "rqstq_intr_code: 0x%x\n", ha->mbx_intr_code, ha->rqstq_intr_code); QLAFX00_CLR_INTR_REG(ha, QLAFX00_HST_INT_STS_BITS); rval = QLA_SUCCESS; done = true; break; default: if ((aenmbx & 0xFF00) == MBA_FW_INIT_INPROGRESS) break; /* If fw is apparently not ready. In order to continue, * we might need to issue Mbox cmd, but the problem is * that the DoorBell vector values that come with the * 8060 AEN are most likely gone by now (and thus no * bell would be rung on the fw side when mbox cmd is * issued). We have to therefore grab the 8060 AEN * shadow regs (filled in by FW when the last 8060 * AEN was being posted). * Do the following to determine what is needed in * order to get the FW ready: * 1. reload the 8060 AEN values from the shadow regs * 2. clear int status to get rid of possible pending * interrupts * 3. issue Get FW State Mbox cmd to determine fw state * Set the mbx and rqstq intr code from Shadow Regs */ aenmbx7 = rd_reg_dword(®->initval7); ha->mbx_intr_code = MSW(aenmbx7); ha->rqstq_intr_code = LSW(aenmbx7); ha->req_que_off = rd_reg_dword(®->initval1); ha->rsp_que_off = rd_reg_dword(®->initval3); ha->req_que_len = rd_reg_dword(®->initval5); ha->rsp_que_len = rd_reg_dword(®->initval6); ql_dbg(ql_dbg_init, vha, 0x0135, "f/w returned mbx_intr_code: 0x%x, " "rqstq_intr_code: 0x%x\n", ha->mbx_intr_code, ha->rqstq_intr_code); QLAFX00_CLR_INTR_REG(ha, QLAFX00_HST_INT_STS_BITS); /* Get the FW state */ rval = qlafx00_get_firmware_state(vha, state); if (rval != QLA_SUCCESS) { /* Retry if timer has not expired */ break; } if (state[0] == FSTATE_FX00_CONFIG_WAIT) { /* Firmware is waiting to be * initialized by driver */ rval = QLA_SUCCESS; done = true; break; } /* Issue driver shutdown and wait until f/w recovers. * Driver should continue to poll until 8060 AEN is * received indicating firmware recovery. */ ql_dbg(ql_dbg_init, vha, 0x0136, "Sending Driver shutdown fw_state 0x%x\n", state[0]); rval = qlafx00_driver_shutdown(vha, 10); if (rval != QLA_SUCCESS) { rval = QLA_FUNCTION_FAILED; break; } msleep(500); wtime = jiffies + (wait_time * HZ); break; } if (!done) { if (time_after_eq(jiffies, wtime)) { ql_dbg(ql_dbg_init, vha, 0x0137, "Init f/w failed: aen[7]: 0x%x\n", rd_reg_dword(®->aenmailbox7)); rval = QLA_FUNCTION_FAILED; done = true; break; } /* Delay for a while */ msleep(500); } } while (!done); if (rval) ql_dbg(ql_dbg_init, vha, 0x0138, "%s **** FAILED ****.\n", __func__); else ql_dbg(ql_dbg_init, vha, 0x0139, "%s **** SUCCESS ****.\n", __func__); return rval; } /* * qlafx00_fw_ready() - Waits for firmware ready. * @ha: HA context * * Returns 0 on success. */ int qlafx00_fw_ready(scsi_qla_host_t *vha) { int rval; unsigned long wtime; uint16_t wait_time; /* Wait time if loop is coming ready */ uint32_t state[5]; rval = QLA_SUCCESS; wait_time = 10; /* wait time before firmware ready */ wtime = jiffies + (wait_time * HZ); /* Wait for ISP to finish init */ if (!vha->flags.init_done) ql_dbg(ql_dbg_init, vha, 0x013a, "Waiting for init to complete...\n"); do { rval = qlafx00_get_firmware_state(vha, state); if (rval == QLA_SUCCESS) { if (state[0] == FSTATE_FX00_INITIALIZED) { ql_dbg(ql_dbg_init, vha, 0x013b, "fw_state=%x\n", state[0]); rval = QLA_SUCCESS; break; } } rval = QLA_FUNCTION_FAILED; if (time_after_eq(jiffies, wtime)) break; /* Delay for a while */ msleep(500); ql_dbg(ql_dbg_init, vha, 0x013c, "fw_state=%x curr time=%lx.\n", state[0], jiffies); } while (1); if (rval) ql_dbg(ql_dbg_init, vha, 0x013d, "Firmware ready **** FAILED ****.\n"); else ql_dbg(ql_dbg_init, vha, 0x013e, "Firmware ready **** SUCCESS ****.\n"); return rval; } static int qlafx00_find_all_targets(scsi_qla_host_t *vha, struct list_head *new_fcports) { int rval; uint16_t tgt_id; fc_port_t *fcport, *new_fcport; int found; struct qla_hw_data *ha = vha->hw; rval = QLA_SUCCESS; if (!test_bit(LOOP_RESYNC_ACTIVE, &vha->dpc_flags)) return QLA_FUNCTION_FAILED; if ((atomic_read(&vha->loop_down_timer) || STATE_TRANSITION(vha))) { atomic_set(&vha->loop_down_timer, 0); set_bit(LOOP_RESYNC_NEEDED, &vha->dpc_flags); return QLA_FUNCTION_FAILED; } ql_dbg(ql_dbg_disc + ql_dbg_init, vha, 0x2088, "Listing Target bit map...\n"); ql_dump_buffer(ql_dbg_disc + ql_dbg_init, vha, 0x2089, ha->gid_list, 32); /* Allocate temporary rmtport for any new rmtports discovered. */ new_fcport = qla2x00_alloc_fcport(vha, GFP_KERNEL); if (new_fcport == NULL) return QLA_MEMORY_ALLOC_FAILED; for_each_set_bit(tgt_id, (void *)ha->gid_list, QLAFX00_TGT_NODE_LIST_SIZE) { /* Send get target node info */ new_fcport->tgt_id = tgt_id; rval = qlafx00_fx_disc(vha, new_fcport, FXDISC_GET_TGT_NODE_INFO); if (rval != QLA_SUCCESS) { ql_log(ql_log_warn, vha, 0x208a, "Target info scan failed -- assuming zero-entry " "result...\n"); continue; } /* Locate matching device in database. */ found = 0; list_for_each_entry(fcport, &vha->vp_fcports, list) { if (memcmp(new_fcport->port_name, fcport->port_name, WWN_SIZE)) continue; found++; /* * If tgt_id is same and state FCS_ONLINE, nothing * changed. */ if (fcport->tgt_id == new_fcport->tgt_id && atomic_read(&fcport->state) == FCS_ONLINE) break; /* * Tgt ID changed or device was marked to be updated. */ ql_dbg(ql_dbg_disc + ql_dbg_init, vha, 0x208b, "TGT-ID Change(%s): Present tgt id: " "0x%x state: 0x%x " "wwnn = %llx wwpn = %llx.\n", __func__, fcport->tgt_id, atomic_read(&fcport->state), (unsigned long long)wwn_to_u64(fcport->node_name), (unsigned long long)wwn_to_u64(fcport->port_name)); ql_log(ql_log_info, vha, 0x208c, "TGT-ID Announce(%s): Discovered tgt " "id 0x%x wwnn = %llx " "wwpn = %llx.\n", __func__, new_fcport->tgt_id, (unsigned long long) wwn_to_u64(new_fcport->node_name), (unsigned long long) wwn_to_u64(new_fcport->port_name)); if (atomic_read(&fcport->state) != FCS_ONLINE) { fcport->old_tgt_id = fcport->tgt_id; fcport->tgt_id = new_fcport->tgt_id; ql_log(ql_log_info, vha, 0x208d, "TGT-ID: New fcport Added: %p\n", fcport); qla2x00_update_fcport(vha, fcport); } else { ql_log(ql_log_info, vha, 0x208e, " Existing TGT-ID %x did not get " " offline event from firmware.\n", fcport->old_tgt_id); qla2x00_mark_device_lost(vha, fcport, 0); set_bit(LOOP_RESYNC_NEEDED, &vha->dpc_flags); qla2x00_free_fcport(new_fcport); return rval; } break; } if (found) continue; /* If device was not in our fcports list, then add it. */ list_add_tail(&new_fcport->list, new_fcports); /* Allocate a new replacement fcport. */ new_fcport = qla2x00_alloc_fcport(vha, GFP_KERNEL); if (new_fcport == NULL) return QLA_MEMORY_ALLOC_FAILED; } qla2x00_free_fcport(new_fcport); return rval; } /* * qlafx00_configure_all_targets * Setup target devices with node ID's. * * Input: * ha = adapter block pointer. * * Returns: * 0 = success. * BIT_0 = error */ static int qlafx00_configure_all_targets(scsi_qla_host_t *vha) { int rval; fc_port_t *fcport, *rmptemp; LIST_HEAD(new_fcports); rval = qlafx00_fx_disc(vha, &vha->hw->mr.fcport, FXDISC_GET_TGT_NODE_LIST); if (rval != QLA_SUCCESS) { set_bit(LOOP_RESYNC_NEEDED, &vha->dpc_flags); return rval; } rval = qlafx00_find_all_targets(vha, &new_fcports); if (rval != QLA_SUCCESS) { set_bit(LOOP_RESYNC_NEEDED, &vha->dpc_flags); return rval; } /* * Delete all previous devices marked lost. */ list_for_each_entry(fcport, &vha->vp_fcports, list) { if (test_bit(LOOP_RESYNC_NEEDED, &vha->dpc_flags)) break; if (atomic_read(&fcport->state) == FCS_DEVICE_LOST) { if (fcport->port_type != FCT_INITIATOR) qla2x00_mark_device_lost(vha, fcport, 0); } } /* * Add the new devices to our devices list. */ list_for_each_entry_safe(fcport, rmptemp, &new_fcports, list) { if (test_bit(LOOP_RESYNC_NEEDED, &vha->dpc_flags)) break; qla2x00_update_fcport(vha, fcport); list_move_tail(&fcport->list, &vha->vp_fcports); ql_log(ql_log_info, vha, 0x208f, "Attach new target id 0x%x wwnn = %llx " "wwpn = %llx.\n", fcport->tgt_id, (unsigned long long)wwn_to_u64(fcport->node_name), (unsigned long long)wwn_to_u64(fcport->port_name)); } /* Free all new device structures not processed. */ list_for_each_entry_safe(fcport, rmptemp, &new_fcports, list) { list_del(&fcport->list); qla2x00_free_fcport(fcport); } return rval; } /* * qlafx00_configure_devices * Updates Fibre Channel Device Database with what is actually on loop. * * Input: * ha = adapter block pointer. * * Returns: * 0 = success. * 1 = error. * 2 = database was full and device was not configured. */ int qlafx00_configure_devices(scsi_qla_host_t *vha) { int rval; unsigned long flags; rval = QLA_SUCCESS; flags = vha->dpc_flags; ql_dbg(ql_dbg_disc, vha, 0x2090, "Configure devices -- dpc flags =0x%lx\n", flags); rval = qlafx00_configure_all_targets(vha); if (rval == QLA_SUCCESS) { if (test_bit(LOOP_RESYNC_NEEDED, &vha->dpc_flags)) { rval = QLA_FUNCTION_FAILED; } else { atomic_set(&vha->loop_state, LOOP_READY); ql_log(ql_log_info, vha, 0x2091, "Device Ready\n"); } } if (rval) { ql_dbg(ql_dbg_disc, vha, 0x2092, "%s *** FAILED ***.\n", __func__); } else { ql_dbg(ql_dbg_disc, vha, 0x2093, "%s: exiting normally.\n", __func__); } return rval; } static void qlafx00_abort_isp_cleanup(scsi_qla_host_t *vha, bool critemp) { struct qla_hw_data *ha = vha->hw; fc_port_t *fcport; vha->flags.online = 0; ha->mr.fw_hbt_en = 0; if (!critemp) { ha->flags.chip_reset_done = 0; clear_bit(ISP_ABORT_NEEDED, &vha->dpc_flags); vha->qla_stats.total_isp_aborts++; ql_log(ql_log_info, vha, 0x013f, "Performing ISP error recovery - ha = %p.\n", ha); ha->isp_ops->reset_chip(vha); } if (atomic_read(&vha->loop_state) != LOOP_DOWN) { atomic_set(&vha->loop_state, LOOP_DOWN); atomic_set(&vha->loop_down_timer, QLAFX00_LOOP_DOWN_TIME); } else { if (!atomic_read(&vha->loop_down_timer)) atomic_set(&vha->loop_down_timer, QLAFX00_LOOP_DOWN_TIME); } /* Clear all async request states across all VPs. */ list_for_each_entry(fcport, &vha->vp_fcports, list) { fcport->flags = 0; if (atomic_read(&fcport->state) == FCS_ONLINE) qla2x00_set_fcport_state(fcport, FCS_DEVICE_LOST); } if (!ha->flags.eeh_busy) { if (critemp) { qla2x00_abort_all_cmds(vha, DID_NO_CONNECT << 16); } else { /* Requeue all commands in outstanding command list. */ qla2x00_abort_all_cmds(vha, DID_RESET << 16); } } qla2x00_free_irqs(vha); if (critemp) set_bit(FX00_CRITEMP_RECOVERY, &vha->dpc_flags); else set_bit(FX00_RESET_RECOVERY, &vha->dpc_flags); /* Clear the Interrupts */ QLAFX00_CLR_INTR_REG(ha, QLAFX00_HST_INT_STS_BITS); ql_log(ql_log_info, vha, 0x0140, "%s Done done - ha=%p.\n", __func__, ha); } /** * qlafx00_init_response_q_entries() - Initializes response queue entries. * @rsp: response queue * * Beginning of request ring has initialization control block already built * by nvram config routine. * * Returns 0 on success. */ void qlafx00_init_response_q_entries(struct rsp_que *rsp) { uint16_t cnt; response_t *pkt; rsp->ring_ptr = rsp->ring; rsp->ring_index = 0; rsp->status_srb = NULL; pkt = rsp->ring_ptr; for (cnt = 0; cnt < rsp->length; cnt++) { pkt->signature = RESPONSE_PROCESSED; wrt_reg_dword((void __force __iomem *)&pkt->signature, RESPONSE_PROCESSED); pkt++; } } int qlafx00_rescan_isp(scsi_qla_host_t *vha) { uint32_t status = QLA_FUNCTION_FAILED; struct qla_hw_data *ha = vha->hw; struct device_reg_fx00 __iomem *reg = &ha->iobase->ispfx00; uint32_t aenmbx7; qla2x00_request_irqs(ha, ha->rsp_q_map[0]); aenmbx7 = rd_reg_dword(®->aenmailbox7); ha->mbx_intr_code = MSW(aenmbx7); ha->rqstq_intr_code = LSW(aenmbx7); ha->req_que_off = rd_reg_dword(®->aenmailbox1); ha->rsp_que_off = rd_reg_dword(®->aenmailbox3); ha->req_que_len = rd_reg_dword(®->aenmailbox5); ha->rsp_que_len = rd_reg_dword(®->aenmailbox6); ql_dbg(ql_dbg_disc, vha, 0x2094, "fw returned mbx_intr_code: 0x%x, rqstq_intr_code: 0x%x " " Req que offset 0x%x Rsp que offset 0x%x\n", ha->mbx_intr_code, ha->rqstq_intr_code, ha->req_que_off, ha->rsp_que_len); /* Clear the Interrupts */ QLAFX00_CLR_INTR_REG(ha, QLAFX00_HST_INT_STS_BITS); status = qla2x00_init_rings(vha); if (!status) { vha->flags.online = 1; /* if no cable then assume it's good */ if ((vha->device_flags & DFLG_NO_CABLE)) status = 0; /* Register system information */ if (qlafx00_fx_disc(vha, &vha->hw->mr.fcport, FXDISC_REG_HOST_INFO)) ql_dbg(ql_dbg_disc, vha, 0x2095, "failed to register host info\n"); } scsi_unblock_requests(vha->host); return status; } void qlafx00_timer_routine(scsi_qla_host_t *vha) { struct qla_hw_data *ha = vha->hw; uint32_t fw_heart_beat; uint32_t aenmbx0; struct device_reg_fx00 __iomem *reg = &ha->iobase->ispfx00; uint32_t tempc; /* Check firmware health */ if (ha->mr.fw_hbt_cnt) ha->mr.fw_hbt_cnt--; else { if ((!ha->flags.mr_reset_hdlr_active) && (!test_bit(UNLOADING, &vha->dpc_flags)) && (!test_bit(ABORT_ISP_ACTIVE, &vha->dpc_flags)) && (ha->mr.fw_hbt_en)) { fw_heart_beat = rd_reg_dword(®->fwheartbeat); if (fw_heart_beat != ha->mr.old_fw_hbt_cnt) { ha->mr.old_fw_hbt_cnt = fw_heart_beat; ha->mr.fw_hbt_miss_cnt = 0; } else { ha->mr.fw_hbt_miss_cnt++; if (ha->mr.fw_hbt_miss_cnt == QLAFX00_HEARTBEAT_MISS_CNT) { set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags); qla2xxx_wake_dpc(vha); ha->mr.fw_hbt_miss_cnt = 0; } } } ha->mr.fw_hbt_cnt = QLAFX00_HEARTBEAT_INTERVAL; } if (test_bit(FX00_RESET_RECOVERY, &vha->dpc_flags)) { /* Reset recovery to be performed in timer routine */ aenmbx0 = rd_reg_dword(®->aenmailbox0); if (ha->mr.fw_reset_timer_exp) { set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags); qla2xxx_wake_dpc(vha); ha->mr.fw_reset_timer_exp = 0; } else if (aenmbx0 == MBA_FW_RESTART_CMPLT) { /* Wake up DPC to rescan the targets */ set_bit(FX00_TARGET_SCAN, &vha->dpc_flags); clear_bit(FX00_RESET_RECOVERY, &vha->dpc_flags); qla2xxx_wake_dpc(vha); ha->mr.fw_reset_timer_tick = QLAFX00_RESET_INTERVAL; } else if ((aenmbx0 == MBA_FW_STARTING) && (!ha->mr.fw_hbt_en)) { ha->mr.fw_hbt_en = 1; } else if (!ha->mr.fw_reset_timer_tick) { if (aenmbx0 == ha->mr.old_aenmbx0_state) ha->mr.fw_reset_timer_exp = 1; ha->mr.fw_reset_timer_tick = QLAFX00_RESET_INTERVAL; } else if (aenmbx0 == 0xFFFFFFFF) { uint32_t data0, data1; data0 = QLAFX00_RD_REG(ha, QLAFX00_BAR1_BASE_ADDR_REG); data1 = QLAFX00_RD_REG(ha, QLAFX00_PEX0_WIN0_BASE_ADDR_REG); data0 &= 0xffff0000; data1 &= 0x0000ffff; QLAFX00_WR_REG(ha, QLAFX00_PEX0_WIN0_BASE_ADDR_REG, (data0 | data1)); } else if ((aenmbx0 & 0xFF00) == MBA_FW_POLL_STATE) { ha->mr.fw_reset_timer_tick = QLAFX00_MAX_RESET_INTERVAL; } else if (aenmbx0 == MBA_FW_RESET_FCT) { ha->mr.fw_reset_timer_tick = QLAFX00_MAX_RESET_INTERVAL; } if (ha->mr.old_aenmbx0_state != aenmbx0) { ha->mr.old_aenmbx0_state = aenmbx0; ha->mr.fw_reset_timer_tick = QLAFX00_RESET_INTERVAL; } ha->mr.fw_reset_timer_tick--; } if (test_bit(FX00_CRITEMP_RECOVERY, &vha->dpc_flags)) { /* * Critical temperature recovery to be * performed in timer routine */ if (ha->mr.fw_critemp_timer_tick == 0) { tempc = QLAFX00_GET_TEMPERATURE(ha); ql_dbg(ql_dbg_timer, vha, 0x6012, "ISPFx00(%s): Critical temp timer, " "current SOC temperature: %d\n", __func__, tempc); if (tempc < ha->mr.critical_temperature) { set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags); clear_bit(FX00_CRITEMP_RECOVERY, &vha->dpc_flags); qla2xxx_wake_dpc(vha); } ha->mr.fw_critemp_timer_tick = QLAFX00_CRITEMP_INTERVAL; } else { ha->mr.fw_critemp_timer_tick--; } } if (ha->mr.host_info_resend) { /* * Incomplete host info might be sent to firmware * durinng system boot - info should be resend */ if (ha->mr.hinfo_resend_timer_tick == 0) { ha->mr.host_info_resend = false; set_bit(FX00_HOST_INFO_RESEND, &vha->dpc_flags); ha->mr.hinfo_resend_timer_tick = QLAFX00_HINFO_RESEND_INTERVAL; qla2xxx_wake_dpc(vha); } else { ha->mr.hinfo_resend_timer_tick--; } } } /* * qlfx00a_reset_initialize * Re-initialize after a iSA device reset. * * Input: * ha = adapter block pointer. * * Returns: * 0 = success */ int qlafx00_reset_initialize(scsi_qla_host_t *vha) { struct qla_hw_data *ha = vha->hw; if (vha->device_flags & DFLG_DEV_FAILED) { ql_dbg(ql_dbg_init, vha, 0x0142, "Device in failed state\n"); return QLA_SUCCESS; } ha->flags.mr_reset_hdlr_active = 1; if (vha->flags.online) { scsi_block_requests(vha->host); qlafx00_abort_isp_cleanup(vha, false); } ql_log(ql_log_info, vha, 0x0143, "(%s): succeeded.\n", __func__); ha->flags.mr_reset_hdlr_active = 0; return QLA_SUCCESS; } /* * qlafx00_abort_isp * Resets ISP and aborts all outstanding commands. * * Input: * ha = adapter block pointer. * * Returns: * 0 = success */ int qlafx00_abort_isp(scsi_qla_host_t *vha) { struct qla_hw_data *ha = vha->hw; if (vha->flags.online) { if (unlikely(pci_channel_offline(ha->pdev) && ha->flags.pci_channel_io_perm_failure)) { clear_bit(ISP_ABORT_RETRY, &vha->dpc_flags); return QLA_SUCCESS; } scsi_block_requests(vha->host); qlafx00_abort_isp_cleanup(vha, false); } else { scsi_block_requests(vha->host); clear_bit(ISP_ABORT_NEEDED, &vha->dpc_flags); vha->qla_stats.total_isp_aborts++; ha->isp_ops->reset_chip(vha); set_bit(FX00_RESET_RECOVERY, &vha->dpc_flags); /* Clear the Interrupts */ QLAFX00_CLR_INTR_REG(ha, QLAFX00_HST_INT_STS_BITS); } ql_log(ql_log_info, vha, 0x0145, "(%s): succeeded.\n", __func__); return QLA_SUCCESS; } static inline fc_port_t* qlafx00_get_fcport(struct scsi_qla_host *vha, int tgt_id) { fc_port_t *fcport; /* Check for matching device in remote port list. */ list_for_each_entry(fcport, &vha->vp_fcports, list) { if (fcport->tgt_id == tgt_id) { ql_dbg(ql_dbg_async, vha, 0x5072, "Matching fcport(%p) found with TGT-ID: 0x%x " "and Remote TGT_ID: 0x%x\n", fcport, fcport->tgt_id, tgt_id); return fcport; } } return NULL; } static void qlafx00_tgt_detach(struct scsi_qla_host *vha, int tgt_id) { fc_port_t *fcport; ql_log(ql_log_info, vha, 0x5073, "Detach TGT-ID: 0x%x\n", tgt_id); fcport = qlafx00_get_fcport(vha, tgt_id); if (!fcport) return; qla2x00_mark_device_lost(vha, fcport, 0); return; } void qlafx00_process_aen(struct scsi_qla_host *vha, struct qla_work_evt *evt) { uint32_t aen_code, aen_data; aen_code = FCH_EVT_VENDOR_UNIQUE; aen_data = evt->u.aenfx.evtcode; switch (evt->u.aenfx.evtcode) { case QLAFX00_MBA_PORT_UPDATE: /* Port database update */ if (evt->u.aenfx.mbx[1] == 0) { if (evt->u.aenfx.mbx[2] == 1) { if (!vha->flags.fw_tgt_reported) vha->flags.fw_tgt_reported = 1; atomic_set(&vha->loop_down_timer, 0); atomic_set(&vha->loop_state, LOOP_UP); set_bit(LOOP_RESYNC_NEEDED, &vha->dpc_flags); qla2xxx_wake_dpc(vha); } else if (evt->u.aenfx.mbx[2] == 2) { qlafx00_tgt_detach(vha, evt->u.aenfx.mbx[3]); } } else if (evt->u.aenfx.mbx[1] == 0xffff) { if (evt->u.aenfx.mbx[2] == 1) { if (!vha->flags.fw_tgt_reported) vha->flags.fw_tgt_reported = 1; set_bit(LOOP_RESYNC_NEEDED, &vha->dpc_flags); } else if (evt->u.aenfx.mbx[2] == 2) { vha->device_flags |= DFLG_NO_CABLE; qla2x00_mark_all_devices_lost(vha); } } break; case QLAFX00_MBA_LINK_UP: aen_code = FCH_EVT_LINKUP; aen_data = 0; break; case QLAFX00_MBA_LINK_DOWN: aen_code = FCH_EVT_LINKDOWN; aen_data = 0; break; case QLAFX00_MBA_TEMP_CRIT: /* Critical temperature event */ ql_log(ql_log_info, vha, 0x5082, "Process critical temperature event " "aenmb[0]: %x\n", evt->u.aenfx.evtcode); scsi_block_requests(vha->host); qlafx00_abort_isp_cleanup(vha, true); scsi_unblock_requests(vha->host); break; } fc_host_post_event(vha->host, fc_get_event_number(), aen_code, aen_data); } static void qlafx00_update_host_attr(scsi_qla_host_t *vha, struct port_info_data *pinfo) { u64 port_name = 0, node_name = 0; port_name = (unsigned long long)wwn_to_u64(pinfo->port_name); node_name = (unsigned long long)wwn_to_u64(pinfo->node_name); fc_host_node_name(vha->host) = node_name; fc_host_port_name(vha->host) = port_name; if (!pinfo->port_type) vha->hw->current_topology = ISP_CFG_F; if (pinfo->link_status == QLAFX00_LINK_STATUS_UP) atomic_set(&vha->loop_state, LOOP_READY); else if (pinfo->link_status == QLAFX00_LINK_STATUS_DOWN) atomic_set(&vha->loop_state, LOOP_DOWN); vha->hw->link_data_rate = (uint16_t)pinfo->link_config; } static void qla2x00_fxdisc_iocb_timeout(void *data) { srb_t *sp = data; struct srb_iocb *lio = &sp->u.iocb_cmd; complete(&lio->u.fxiocb.fxiocb_comp); } static void qla2x00_fxdisc_sp_done(srb_t *sp, int res) { struct srb_iocb *lio = &sp->u.iocb_cmd; complete(&lio->u.fxiocb.fxiocb_comp); } int qlafx00_fx_disc(scsi_qla_host_t *vha, fc_port_t *fcport, uint16_t fx_type) { srb_t *sp; struct srb_iocb *fdisc; int rval = QLA_FUNCTION_FAILED; struct qla_hw_data *ha = vha->hw; struct host_system_info *phost_info; struct register_host_info *preg_hsi; struct new_utsname *p_sysid = NULL; /* ref: INIT */ sp = qla2x00_get_sp(vha, fcport, GFP_KERNEL); if (!sp) goto done; sp->type = SRB_FXIOCB_DCMD; sp->name = "fxdisc"; qla2x00_init_async_sp(sp, FXDISC_TIMEOUT, qla2x00_fxdisc_sp_done); sp->u.iocb_cmd.timeout = qla2x00_fxdisc_iocb_timeout; fdisc = &sp->u.iocb_cmd; switch (fx_type) { case FXDISC_GET_CONFIG_INFO: fdisc->u.fxiocb.flags = SRB_FXDISC_RESP_DMA_VALID; fdisc->u.fxiocb.rsp_len = sizeof(struct config_info_data); break; case FXDISC_GET_PORT_INFO: fdisc->u.fxiocb.flags = SRB_FXDISC_RESP_DMA_VALID | SRB_FXDISC_REQ_DWRD_VALID; fdisc->u.fxiocb.rsp_len = QLAFX00_PORT_DATA_INFO; fdisc->u.fxiocb.req_data = cpu_to_le32(fcport->port_id); break; case FXDISC_GET_TGT_NODE_INFO: fdisc->u.fxiocb.flags = SRB_FXDISC_RESP_DMA_VALID | SRB_FXDISC_REQ_DWRD_VALID; fdisc->u.fxiocb.rsp_len = QLAFX00_TGT_NODE_INFO; fdisc->u.fxiocb.req_data = cpu_to_le32(fcport->tgt_id); break; case FXDISC_GET_TGT_NODE_LIST: fdisc->u.fxiocb.flags = SRB_FXDISC_RESP_DMA_VALID | SRB_FXDISC_REQ_DWRD_VALID; fdisc->u.fxiocb.rsp_len = QLAFX00_TGT_NODE_LIST_SIZE; break; case FXDISC_REG_HOST_INFO: fdisc->u.fxiocb.flags = SRB_FXDISC_REQ_DMA_VALID; fdisc->u.fxiocb.req_len = sizeof(struct register_host_info); p_sysid = utsname(); if (!p_sysid) { ql_log(ql_log_warn, vha, 0x303c, "Not able to get the system information\n"); goto done_free_sp; } break; case FXDISC_ABORT_IOCTL: default: break; } if (fdisc->u.fxiocb.flags & SRB_FXDISC_REQ_DMA_VALID) { fdisc->u.fxiocb.req_addr = dma_alloc_coherent(&ha->pdev->dev, fdisc->u.fxiocb.req_len, &fdisc->u.fxiocb.req_dma_handle, GFP_KERNEL); if (!fdisc->u.fxiocb.req_addr) goto done_free_sp; if (fx_type == FXDISC_REG_HOST_INFO) { preg_hsi = (struct register_host_info *) fdisc->u.fxiocb.req_addr; phost_info = &preg_hsi->hsi; memset(preg_hsi, 0, sizeof(struct register_host_info)); phost_info->os_type = OS_TYPE_LINUX; strscpy(phost_info->sysname, p_sysid->sysname, sizeof(phost_info->sysname)); strscpy(phost_info->nodename, p_sysid->nodename, sizeof(phost_info->nodename)); if (!strcmp(phost_info->nodename, "(none)")) ha->mr.host_info_resend = true; strscpy(phost_info->release, p_sysid->release, sizeof(phost_info->release)); strscpy(phost_info->version, p_sysid->version, sizeof(phost_info->version)); strscpy(phost_info->machine, p_sysid->machine, sizeof(phost_info->machine)); strscpy(phost_info->domainname, p_sysid->domainname, sizeof(phost_info->domainname)); strscpy(phost_info->hostdriver, QLA2XXX_VERSION, sizeof(phost_info->hostdriver)); preg_hsi->utc = (uint64_t)ktime_get_real_seconds(); ql_dbg(ql_dbg_init, vha, 0x0149, "ISP%04X: Host registration with firmware\n", ha->pdev->device); ql_dbg(ql_dbg_init, vha, 0x014a, "os_type = '%d', sysname = '%s', nodname = '%s'\n", phost_info->os_type, phost_info->sysname, phost_info->nodename); ql_dbg(ql_dbg_init, vha, 0x014b, "release = '%s', version = '%s'\n", phost_info->release, phost_info->version); ql_dbg(ql_dbg_init, vha, 0x014c, "machine = '%s' " "domainname = '%s', hostdriver = '%s'\n", phost_info->machine, phost_info->domainname, phost_info->hostdriver); ql_dump_buffer(ql_dbg_init + ql_dbg_disc, vha, 0x014d, phost_info, sizeof(*phost_info)); } } if (fdisc->u.fxiocb.flags & SRB_FXDISC_RESP_DMA_VALID) { fdisc->u.fxiocb.rsp_addr = dma_alloc_coherent(&ha->pdev->dev, fdisc->u.fxiocb.rsp_len, &fdisc->u.fxiocb.rsp_dma_handle, GFP_KERNEL); if (!fdisc->u.fxiocb.rsp_addr) goto done_unmap_req; } fdisc->u.fxiocb.req_func_type = cpu_to_le16(fx_type); rval = qla2x00_start_sp(sp); if (rval != QLA_SUCCESS) goto done_unmap_dma; wait_for_completion(&fdisc->u.fxiocb.fxiocb_comp); if (fx_type == FXDISC_GET_CONFIG_INFO) { struct config_info_data *pinfo = (struct config_info_data *) fdisc->u.fxiocb.rsp_addr; strscpy(vha->hw->model_number, pinfo->model_num, ARRAY_SIZE(vha->hw->model_number)); strscpy(vha->hw->model_desc, pinfo->model_description, ARRAY_SIZE(vha->hw->model_desc)); memcpy(&vha->hw->mr.symbolic_name, pinfo->symbolic_name, sizeof(vha->hw->mr.symbolic_name)); memcpy(&vha->hw->mr.serial_num, pinfo->serial_num, sizeof(vha->hw->mr.serial_num)); memcpy(&vha->hw->mr.hw_version, pinfo->hw_version, sizeof(vha->hw->mr.hw_version)); memcpy(&vha->hw->mr.fw_version, pinfo->fw_version, sizeof(vha->hw->mr.fw_version)); strim(vha->hw->mr.fw_version); memcpy(&vha->hw->mr.uboot_version, pinfo->uboot_version, sizeof(vha->hw->mr.uboot_version)); memcpy(&vha->hw->mr.fru_serial_num, pinfo->fru_serial_num, sizeof(vha->hw->mr.fru_serial_num)); vha->hw->mr.critical_temperature = (pinfo->nominal_temp_value) ? pinfo->nominal_temp_value : QLAFX00_CRITEMP_THRSHLD; ha->mr.extended_io_enabled = (pinfo->enabled_capabilities & QLAFX00_EXTENDED_IO_EN_MASK) != 0; } else if (fx_type == FXDISC_GET_PORT_INFO) { struct port_info_data *pinfo = (struct port_info_data *) fdisc->u.fxiocb.rsp_addr; memcpy(vha->node_name, pinfo->node_name, WWN_SIZE); memcpy(vha->port_name, pinfo->port_name, WWN_SIZE); vha->d_id.b.domain = pinfo->port_id[0]; vha->d_id.b.area = pinfo->port_id[1]; vha->d_id.b.al_pa = pinfo->port_id[2]; qlafx00_update_host_attr(vha, pinfo); ql_dump_buffer(ql_dbg_init + ql_dbg_buffer, vha, 0x0141, pinfo, 16); } else if (fx_type == FXDISC_GET_TGT_NODE_INFO) { struct qlafx00_tgt_node_info *pinfo = (struct qlafx00_tgt_node_info *) fdisc->u.fxiocb.rsp_addr; memcpy(fcport->node_name, pinfo->tgt_node_wwnn, WWN_SIZE); memcpy(fcport->port_name, pinfo->tgt_node_wwpn, WWN_SIZE); fcport->port_type = FCT_TARGET; ql_dump_buffer(ql_dbg_init + ql_dbg_buffer, vha, 0x0144, pinfo, 16); } else if (fx_type == FXDISC_GET_TGT_NODE_LIST) { struct qlafx00_tgt_node_info *pinfo = (struct qlafx00_tgt_node_info *) fdisc->u.fxiocb.rsp_addr; ql_dump_buffer(ql_dbg_init + ql_dbg_buffer, vha, 0x0146, pinfo, 16); memcpy(vha->hw->gid_list, pinfo, QLAFX00_TGT_NODE_LIST_SIZE); } else if (fx_type == FXDISC_ABORT_IOCTL) fdisc->u.fxiocb.result = (fdisc->u.fxiocb.result == cpu_to_le32(QLAFX00_IOCTL_ICOB_ABORT_SUCCESS)) ? cpu_to_le32(QLA_SUCCESS) : cpu_to_le32(QLA_FUNCTION_FAILED); rval = le32_to_cpu(fdisc->u.fxiocb.result); done_unmap_dma: if (fdisc->u.fxiocb.rsp_addr) dma_free_coherent(&ha->pdev->dev, fdisc->u.fxiocb.rsp_len, fdisc->u.fxiocb.rsp_addr, fdisc->u.fxiocb.rsp_dma_handle); done_unmap_req: if (fdisc->u.fxiocb.req_addr) dma_free_coherent(&ha->pdev->dev, fdisc->u.fxiocb.req_len, fdisc->u.fxiocb.req_addr, fdisc->u.fxiocb.req_dma_handle); done_free_sp: /* ref: INIT */ kref_put(&sp->cmd_kref, qla2x00_sp_release); done: return rval; } /* * qlafx00_initialize_adapter * Initialize board. * * Input: * ha = adapter block pointer. * * Returns: * 0 = success */ int qlafx00_initialize_adapter(scsi_qla_host_t *vha) { int rval; struct qla_hw_data *ha = vha->hw; uint32_t tempc; /* Clear adapter flags. */ vha->flags.online = 0; ha->flags.chip_reset_done = 0; vha->flags.reset_active = 0; ha->flags.pci_channel_io_perm_failure = 0; ha->flags.eeh_busy = 0; atomic_set(&vha->loop_down_timer, LOOP_DOWN_TIME); atomic_set(&vha->loop_state, LOOP_DOWN); vha->device_flags = DFLG_NO_CABLE; vha->dpc_flags = 0; vha->flags.management_server_logged_in = 0; ha->isp_abort_cnt = 0; ha->beacon_blink_led = 0; set_bit(0, ha->req_qid_map); set_bit(0, ha->rsp_qid_map); ql_dbg(ql_dbg_init, vha, 0x0147, "Configuring PCI space...\n"); rval = ha->isp_ops->pci_config(vha); if (rval) { ql_log(ql_log_warn, vha, 0x0148, "Unable to configure PCI space.\n"); return rval; } rval = qlafx00_init_fw_ready(vha); if (rval != QLA_SUCCESS) return rval; qlafx00_save_queue_ptrs(vha); rval = qlafx00_config_queues(vha); if (rval != QLA_SUCCESS) return rval; /* * Allocate the array of outstanding commands * now that we know the firmware resources. */ rval = qla2x00_alloc_outstanding_cmds(ha, vha->req); if (rval != QLA_SUCCESS) return rval; rval = qla2x00_init_rings(vha); ha->flags.chip_reset_done = 1; tempc = QLAFX00_GET_TEMPERATURE(ha); ql_dbg(ql_dbg_init, vha, 0x0152, "ISPFx00(%s): Critical temp timer, current SOC temperature: 0x%x\n", __func__, tempc); return rval; } uint32_t qlafx00_fw_state_show(struct device *dev, struct device_attribute *attr, char *buf) { scsi_qla_host_t *vha = shost_priv(class_to_shost(dev)); int rval = QLA_FUNCTION_FAILED; uint32_t state[1]; if (qla2x00_reset_active(vha)) ql_log(ql_log_warn, vha, 0x70ce, "ISP reset active.\n"); else if (!vha->hw->flags.eeh_busy) { rval = qlafx00_get_firmware_state(vha, state); } if (rval != QLA_SUCCESS) memset(state, -1, sizeof(state)); return state[0]; } void qlafx00_get_host_speed(struct Scsi_Host *shost) { struct qla_hw_data *ha = ((struct scsi_qla_host *) (shost_priv(shost)))->hw; u32 speed = FC_PORTSPEED_UNKNOWN; switch (ha->link_data_rate) { case QLAFX00_PORT_SPEED_2G: speed = FC_PORTSPEED_2GBIT; break; case QLAFX00_PORT_SPEED_4G: speed = FC_PORTSPEED_4GBIT; break; case QLAFX00_PORT_SPEED_8G: speed = FC_PORTSPEED_8GBIT; break; case QLAFX00_PORT_SPEED_10G: speed = FC_PORTSPEED_10GBIT; break; } fc_host_speed(shost) = speed; } /** QLAFX00 specific ISR implementation functions */ static inline void qlafx00_handle_sense(srb_t *sp, uint8_t *sense_data, uint32_t par_sense_len, uint32_t sense_len, struct rsp_que *rsp, int res) { struct scsi_qla_host *vha = sp->vha; struct scsi_cmnd *cp = GET_CMD_SP(sp); uint32_t track_sense_len; SET_FW_SENSE_LEN(sp, sense_len); if (sense_len >= SCSI_SENSE_BUFFERSIZE) sense_len = SCSI_SENSE_BUFFERSIZE; SET_CMD_SENSE_LEN(sp, sense_len); SET_CMD_SENSE_PTR(sp, cp->sense_buffer); track_sense_len = sense_len; if (sense_len > par_sense_len) sense_len = par_sense_len; memcpy(cp->sense_buffer, sense_data, sense_len); SET_FW_SENSE_LEN(sp, GET_FW_SENSE_LEN(sp) - sense_len); SET_CMD_SENSE_PTR(sp, cp->sense_buffer + sense_len); track_sense_len -= sense_len; SET_CMD_SENSE_LEN(sp, track_sense_len); ql_dbg(ql_dbg_io, vha, 0x304d, "sense_len=0x%x par_sense_len=0x%x track_sense_len=0x%x.\n", sense_len, par_sense_len, track_sense_len); if (GET_FW_SENSE_LEN(sp) > 0) { rsp->status_srb = sp; cp->result = res; } if (sense_len) { ql_dbg(ql_dbg_io + ql_dbg_buffer, vha, 0x3039, "Check condition Sense data, nexus%ld:%d:%llu cmd=%p.\n", sp->vha->host_no, cp->device->id, cp->device->lun, cp); ql_dump_buffer(ql_dbg_io + ql_dbg_buffer, vha, 0x3049, cp->sense_buffer, sense_len); } } static void qlafx00_tm_iocb_entry(scsi_qla_host_t *vha, struct req_que *req, struct tsk_mgmt_entry_fx00 *pkt, srb_t *sp, __le16 sstatus, __le16 cpstatus) { struct srb_iocb *tmf; tmf = &sp->u.iocb_cmd; if (cpstatus != cpu_to_le16((uint16_t)CS_COMPLETE) || (sstatus & cpu_to_le16((uint16_t)SS_RESPONSE_INFO_LEN_VALID))) cpstatus = cpu_to_le16((uint16_t)CS_INCOMPLETE); tmf->u.tmf.comp_status = cpstatus; sp->done(sp, 0); } static void qlafx00_abort_iocb_entry(scsi_qla_host_t *vha, struct req_que *req, struct abort_iocb_entry_fx00 *pkt) { const char func[] = "ABT_IOCB"; srb_t *sp; struct srb_iocb *abt; sp = qla2x00_get_sp_from_handle(vha, func, req, pkt); if (!sp) return; abt = &sp->u.iocb_cmd; abt->u.abt.comp_status = pkt->tgt_id_sts; sp->done(sp, 0); } static void qlafx00_ioctl_iosb_entry(scsi_qla_host_t *vha, struct req_que *req, struct ioctl_iocb_entry_fx00 *pkt) { const char func[] = "IOSB_IOCB"; srb_t *sp; struct bsg_job *bsg_job; struct fc_bsg_reply *bsg_reply; struct srb_iocb *iocb_job; int res = 0; struct qla_mt_iocb_rsp_fx00 fstatus; uint8_t *fw_sts_ptr; sp = qla2x00_get_sp_from_handle(vha, func, req, pkt); if (!sp) return; if (sp->type == SRB_FXIOCB_DCMD) { iocb_job = &sp->u.iocb_cmd; iocb_job->u.fxiocb.seq_number = pkt->seq_no; iocb_job->u.fxiocb.fw_flags = pkt->fw_iotcl_flags; iocb_job->u.fxiocb.result = pkt->status; if (iocb_job->u.fxiocb.flags & SRB_FXDISC_RSP_DWRD_VALID) iocb_job->u.fxiocb.req_data = pkt->dataword_r; } else { bsg_job = sp->u.bsg_job; bsg_reply = bsg_job->reply; memset(&fstatus, 0, sizeof(struct qla_mt_iocb_rsp_fx00)); fstatus.reserved_1 = pkt->reserved_0; fstatus.func_type = pkt->comp_func_num; fstatus.ioctl_flags = pkt->fw_iotcl_flags; fstatus.ioctl_data = pkt->dataword_r; fstatus.adapid = pkt->adapid; fstatus.reserved_2 = pkt->dataword_r_extra; fstatus.res_count = pkt->residuallen; fstatus.status = pkt->status; fstatus.seq_number = pkt->seq_no; memcpy(fstatus.reserved_3, pkt->reserved_2, 20 * sizeof(uint8_t)); fw_sts_ptr = bsg_job->reply + sizeof(struct fc_bsg_reply); memcpy(fw_sts_ptr, &fstatus, sizeof(fstatus)); bsg_job->reply_len = sizeof(struct fc_bsg_reply) + sizeof(struct qla_mt_iocb_rsp_fx00) + sizeof(uint8_t); ql_dump_buffer(ql_dbg_user + ql_dbg_verbose, sp->vha, 0x5080, pkt, sizeof(*pkt)); ql_dump_buffer(ql_dbg_user + ql_dbg_verbose, sp->vha, 0x5074, fw_sts_ptr, sizeof(fstatus)); res = bsg_reply->result = DID_OK << 16; bsg_reply->reply_payload_rcv_len = bsg_job->reply_payload.payload_len; } sp->done(sp, res); } /** * qlafx00_status_entry() - Process a Status IOCB entry. * @vha: SCSI driver HA context * @rsp: response queue * @pkt: Entry pointer */ static void qlafx00_status_entry(scsi_qla_host_t *vha, struct rsp_que *rsp, void *pkt) { srb_t *sp; fc_port_t *fcport; struct scsi_cmnd *cp; struct sts_entry_fx00 *sts; __le16 comp_status; __le16 scsi_status; __le16 lscsi_status; int32_t resid; uint32_t sense_len, par_sense_len, rsp_info_len, resid_len, fw_resid_len; uint8_t *rsp_info = NULL, *sense_data = NULL; struct qla_hw_data *ha = vha->hw; uint32_t hindex, handle; uint16_t que; struct req_que *req; int logit = 1; int res = 0; sts = (struct sts_entry_fx00 *) pkt; comp_status = sts->comp_status; scsi_status = sts->scsi_status & cpu_to_le16((uint16_t)SS_MASK); hindex = sts->handle; handle = LSW(hindex); que = MSW(hindex); req = ha->req_q_map[que]; /* Validate handle. */ if (handle < req->num_outstanding_cmds) sp = req->outstanding_cmds[handle]; else sp = NULL; if (sp == NULL) { ql_dbg(ql_dbg_io, vha, 0x3034, "Invalid status handle (0x%x).\n", handle); set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags); qla2xxx_wake_dpc(vha); return; } if (sp->type == SRB_TM_CMD) { req->outstanding_cmds[handle] = NULL; qlafx00_tm_iocb_entry(vha, req, pkt, sp, scsi_status, comp_status); return; } /* Fast path completion. */ if (comp_status == CS_COMPLETE && scsi_status == 0) { qla2x00_process_completed_request(vha, req, handle); return; } req->outstanding_cmds[handle] = NULL; cp = GET_CMD_SP(sp); if (cp == NULL) { ql_dbg(ql_dbg_io, vha, 0x3048, "Command already returned (0x%x/%p).\n", handle, sp); return; } lscsi_status = scsi_status & cpu_to_le16((uint16_t)STATUS_MASK); fcport = sp->fcport; sense_len = par_sense_len = rsp_info_len = resid_len = fw_resid_len = 0; if (scsi_status & cpu_to_le16((uint16_t)SS_SENSE_LEN_VALID)) sense_len = sts->sense_len; if (scsi_status & cpu_to_le16(((uint16_t)SS_RESIDUAL_UNDER | (uint16_t)SS_RESIDUAL_OVER))) resid_len = le32_to_cpu(sts->residual_len); if (comp_status == cpu_to_le16((uint16_t)CS_DATA_UNDERRUN)) fw_resid_len = le32_to_cpu(sts->residual_len); rsp_info = sense_data = sts->data; par_sense_len = sizeof(sts->data); /* Check for overrun. */ if (comp_status == CS_COMPLETE && scsi_status & cpu_to_le16((uint16_t)SS_RESIDUAL_OVER)) comp_status = cpu_to_le16((uint16_t)CS_DATA_OVERRUN); /* * Based on Host and scsi status generate status code for Linux */ switch (le16_to_cpu(comp_status)) { case CS_COMPLETE: case CS_QUEUE_FULL: if (scsi_status == 0) { res = DID_OK << 16; break; } if (scsi_status & cpu_to_le16(((uint16_t)SS_RESIDUAL_UNDER | (uint16_t)SS_RESIDUAL_OVER))) { resid = resid_len; scsi_set_resid(cp, resid); if (!lscsi_status && ((unsigned)(scsi_bufflen(cp) - resid) < cp->underflow)) { ql_dbg(ql_dbg_io, fcport->vha, 0x3050, "Mid-layer underflow " "detected (0x%x of 0x%x bytes).\n", resid, scsi_bufflen(cp)); res = DID_ERROR << 16; break; } } res = DID_OK << 16 | le16_to_cpu(lscsi_status); if (lscsi_status == cpu_to_le16((uint16_t)SAM_STAT_TASK_SET_FULL)) { ql_dbg(ql_dbg_io, fcport->vha, 0x3051, "QUEUE FULL detected.\n"); break; } logit = 0; if (lscsi_status != cpu_to_le16((uint16_t)SS_CHECK_CONDITION)) break; memset(cp->sense_buffer, 0, SCSI_SENSE_BUFFERSIZE); if (!(scsi_status & cpu_to_le16((uint16_t)SS_SENSE_LEN_VALID))) break; qlafx00_handle_sense(sp, sense_data, par_sense_len, sense_len, rsp, res); break; case CS_DATA_UNDERRUN: /* Use F/W calculated residual length. */ if (IS_FWI2_CAPABLE(ha) || IS_QLAFX00(ha)) resid = fw_resid_len; else resid = resid_len; scsi_set_resid(cp, resid); if (scsi_status & cpu_to_le16((uint16_t)SS_RESIDUAL_UNDER)) { if ((IS_FWI2_CAPABLE(ha) || IS_QLAFX00(ha)) && fw_resid_len != resid_len) { ql_dbg(ql_dbg_io, fcport->vha, 0x3052, "Dropped frame(s) detected " "(0x%x of 0x%x bytes).\n", resid, scsi_bufflen(cp)); res = DID_ERROR << 16 | le16_to_cpu(lscsi_status); goto check_scsi_status; } if (!lscsi_status && ((unsigned)(scsi_bufflen(cp) - resid) < cp->underflow)) { ql_dbg(ql_dbg_io, fcport->vha, 0x3053, "Mid-layer underflow " "detected (0x%x of 0x%x bytes, " "cp->underflow: 0x%x).\n", resid, scsi_bufflen(cp), cp->underflow); res = DID_ERROR << 16; break; } } else if (lscsi_status != cpu_to_le16((uint16_t)SAM_STAT_TASK_SET_FULL) && lscsi_status != cpu_to_le16((uint16_t)SAM_STAT_BUSY)) { /* * scsi status of task set and busy are considered * to be task not completed. */ ql_dbg(ql_dbg_io, fcport->vha, 0x3054, "Dropped frame(s) detected (0x%x " "of 0x%x bytes).\n", resid, scsi_bufflen(cp)); res = DID_ERROR << 16 | le16_to_cpu(lscsi_status); goto check_scsi_status; } else { ql_dbg(ql_dbg_io, fcport->vha, 0x3055, "scsi_status: 0x%x, lscsi_status: 0x%x\n", scsi_status, lscsi_status); } res = DID_OK << 16 | le16_to_cpu(lscsi_status); logit = 0; check_scsi_status: /* * Check to see if SCSI Status is non zero. If so report SCSI * Status. */ if (lscsi_status != 0) { if (lscsi_status == cpu_to_le16((uint16_t)SAM_STAT_TASK_SET_FULL)) { ql_dbg(ql_dbg_io, fcport->vha, 0x3056, "QUEUE FULL detected.\n"); logit = 1; break; } if (lscsi_status != cpu_to_le16((uint16_t)SS_CHECK_CONDITION)) break; memset(cp->sense_buffer, 0, SCSI_SENSE_BUFFERSIZE); if (!(scsi_status & cpu_to_le16((uint16_t)SS_SENSE_LEN_VALID))) break; qlafx00_handle_sense(sp, sense_data, par_sense_len, sense_len, rsp, res); } break; case CS_PORT_LOGGED_OUT: case CS_PORT_CONFIG_CHG: case CS_PORT_BUSY: case CS_INCOMPLETE: case CS_PORT_UNAVAILABLE: case CS_TIMEOUT: case CS_RESET: /* * We are going to have the fc class block the rport * while we try to recover so instruct the mid layer * to requeue until the class decides how to handle this. */ res = DID_TRANSPORT_DISRUPTED << 16; ql_dbg(ql_dbg_io, fcport->vha, 0x3057, "Port down status: port-state=0x%x.\n", atomic_read(&fcport->state)); if (atomic_read(&fcport->state) == FCS_ONLINE) qla2x00_mark_device_lost(fcport->vha, fcport, 1); break; case CS_ABORTED: res = DID_RESET << 16; break; default: res = DID_ERROR << 16; break; } if (logit) ql_dbg(ql_dbg_io, fcport->vha, 0x3058, "FCP command status: 0x%x-0x%x (0x%x) nexus=%ld:%d:%llu " "tgt_id: 0x%x lscsi_status: 0x%x cdb=%10phN len=0x%x " "rsp_info=%p resid=0x%x fw_resid=0x%x sense_len=0x%x, " "par_sense_len=0x%x, rsp_info_len=0x%x\n", comp_status, scsi_status, res, vha->host_no, cp->device->id, cp->device->lun, fcport->tgt_id, lscsi_status, cp->cmnd, scsi_bufflen(cp), rsp_info, resid_len, fw_resid_len, sense_len, par_sense_len, rsp_info_len); if (rsp->status_srb == NULL) sp->done(sp, res); else WARN_ON_ONCE(true); } /** * qlafx00_status_cont_entry() - Process a Status Continuations entry. * @rsp: response queue * @pkt: Entry pointer * * Extended sense data. */ static void qlafx00_status_cont_entry(struct rsp_que *rsp, sts_cont_entry_t *pkt) { uint8_t sense_sz = 0; struct qla_hw_data *ha = rsp->hw; struct scsi_qla_host *vha = pci_get_drvdata(ha->pdev); srb_t *sp = rsp->status_srb; struct scsi_cmnd *cp; uint32_t sense_len; uint8_t *sense_ptr; if (!sp) { ql_dbg(ql_dbg_io, vha, 0x3037, "no SP, sp = %p\n", sp); return; } if (!GET_FW_SENSE_LEN(sp)) { ql_dbg(ql_dbg_io, vha, 0x304b, "no fw sense data, sp = %p\n", sp); return; } cp = GET_CMD_SP(sp); if (cp == NULL) { ql_log(ql_log_warn, vha, 0x303b, "cmd is NULL: already returned to OS (sp=%p).\n", sp); rsp->status_srb = NULL; return; } if (!GET_CMD_SENSE_LEN(sp)) { ql_dbg(ql_dbg_io, vha, 0x304c, "no sense data, sp = %p\n", sp); } else { sense_len = GET_CMD_SENSE_LEN(sp); sense_ptr = GET_CMD_SENSE_PTR(sp); ql_dbg(ql_dbg_io, vha, 0x304f, "sp=%p sense_len=0x%x sense_ptr=%p.\n", sp, sense_len, sense_ptr); if (sense_len > sizeof(pkt->data)) sense_sz = sizeof(pkt->data); else sense_sz = sense_len; /* Move sense data. */ ql_dump_buffer(ql_dbg_io + ql_dbg_buffer, vha, 0x304e, pkt, sizeof(*pkt)); memcpy(sense_ptr, pkt->data, sense_sz); ql_dump_buffer(ql_dbg_io + ql_dbg_buffer, vha, 0x304a, sense_ptr, sense_sz); sense_len -= sense_sz; sense_ptr += sense_sz; SET_CMD_SENSE_PTR(sp, sense_ptr); SET_CMD_SENSE_LEN(sp, sense_len); } sense_len = GET_FW_SENSE_LEN(sp); sense_len = (sense_len > sizeof(pkt->data)) ? (sense_len - sizeof(pkt->data)) : 0; SET_FW_SENSE_LEN(sp, sense_len); /* Place command on done queue. */ if (sense_len == 0) { rsp->status_srb = NULL; sp->done(sp, cp->result); } else { WARN_ON_ONCE(true); } } /** * qlafx00_multistatus_entry() - Process Multi response queue entries. * @vha: SCSI driver HA context * @rsp: response queue * @pkt: received packet */ static void qlafx00_multistatus_entry(struct scsi_qla_host *vha, struct rsp_que *rsp, void *pkt) { srb_t *sp; struct multi_sts_entry_fx00 *stsmfx; struct qla_hw_data *ha = vha->hw; uint32_t handle, hindex, handle_count, i; uint16_t que; struct req_que *req; __le32 *handle_ptr; stsmfx = (struct multi_sts_entry_fx00 *) pkt; handle_count = stsmfx->handle_count; if (handle_count > MAX_HANDLE_COUNT) { ql_dbg(ql_dbg_io, vha, 0x3035, "Invalid handle count (0x%x).\n", handle_count); set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags); qla2xxx_wake_dpc(vha); return; } handle_ptr = &stsmfx->handles[0]; for (i = 0; i < handle_count; i++) { hindex = le32_to_cpu(*handle_ptr); handle = LSW(hindex); que = MSW(hindex); req = ha->req_q_map[que]; /* Validate handle. */ if (handle < req->num_outstanding_cmds) sp = req->outstanding_cmds[handle]; else sp = NULL; if (sp == NULL) { ql_dbg(ql_dbg_io, vha, 0x3044, "Invalid status handle (0x%x).\n", handle); set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags); qla2xxx_wake_dpc(vha); return; } qla2x00_process_completed_request(vha, req, handle); handle_ptr++; } } /** * qlafx00_error_entry() - Process an error entry. * @vha: SCSI driver HA context * @rsp: response queue * @pkt: Entry pointer */ static void qlafx00_error_entry(scsi_qla_host_t *vha, struct rsp_que *rsp, struct sts_entry_fx00 *pkt) { srb_t *sp; struct qla_hw_data *ha = vha->hw; const char func[] = "ERROR-IOCB"; uint16_t que = 0; struct req_que *req = NULL; int res = DID_ERROR << 16; req = ha->req_q_map[que]; sp = qla2x00_get_sp_from_handle(vha, func, req, pkt); if (sp) { sp->done(sp, res); return; } set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags); qla2xxx_wake_dpc(vha); } /** * qlafx00_process_response_queue() - Process response queue entries. * @vha: SCSI driver HA context * @rsp: response queue */ static void qlafx00_process_response_queue(struct scsi_qla_host *vha, struct rsp_que *rsp) { struct sts_entry_fx00 *pkt; response_t *lptr; uint16_t lreq_q_in = 0; uint16_t lreq_q_out = 0; lreq_q_in = rd_reg_dword(rsp->rsp_q_in); lreq_q_out = rsp->ring_index; while (lreq_q_in != lreq_q_out) { lptr = rsp->ring_ptr; memcpy_fromio(rsp->rsp_pkt, (void __iomem *)lptr, sizeof(rsp->rsp_pkt)); pkt = (struct sts_entry_fx00 *)rsp->rsp_pkt; rsp->ring_index++; lreq_q_out++; if (rsp->ring_index == rsp->length) { lreq_q_out = 0; rsp->ring_index = 0; rsp->ring_ptr = rsp->ring; } else { rsp->ring_ptr++; } if (pkt->entry_status != 0 && pkt->entry_type != IOCTL_IOSB_TYPE_FX00) { ql_dbg(ql_dbg_async, vha, 0x507f, "type of error status in response: 0x%x\n", pkt->entry_status); qlafx00_error_entry(vha, rsp, (struct sts_entry_fx00 *)pkt); continue; } switch (pkt->entry_type) { case STATUS_TYPE_FX00: qlafx00_status_entry(vha, rsp, pkt); break; case STATUS_CONT_TYPE_FX00: qlafx00_status_cont_entry(rsp, (sts_cont_entry_t *)pkt); break; case MULTI_STATUS_TYPE_FX00: qlafx00_multistatus_entry(vha, rsp, pkt); break; case ABORT_IOCB_TYPE_FX00: qlafx00_abort_iocb_entry(vha, rsp->req, (struct abort_iocb_entry_fx00 *)pkt); break; case IOCTL_IOSB_TYPE_FX00: qlafx00_ioctl_iosb_entry(vha, rsp->req, (struct ioctl_iocb_entry_fx00 *)pkt); break; default: /* Type Not Supported. */ ql_dbg(ql_dbg_async, vha, 0x5081, "Received unknown response pkt type %x " "entry status=%x.\n", pkt->entry_type, pkt->entry_status); break; } } /* Adjust ring index */ wrt_reg_dword(rsp->rsp_q_out, rsp->ring_index); } /** * qlafx00_async_event() - Process aynchronous events. * @vha: SCSI driver HA context */ static void qlafx00_async_event(scsi_qla_host_t *vha) { struct qla_hw_data *ha = vha->hw; struct device_reg_fx00 __iomem *reg; int data_size = 1; reg = &ha->iobase->ispfx00; /* Setup to process RIO completion. */ switch (ha->aenmb[0]) { case QLAFX00_MBA_SYSTEM_ERR: /* System Error */ ql_log(ql_log_warn, vha, 0x5079, "ISP System Error - mbx1=%x\n", ha->aenmb[0]); set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags); break; case QLAFX00_MBA_SHUTDOWN_RQSTD: /* Shutdown requested */ ql_dbg(ql_dbg_async, vha, 0x5076, "Asynchronous FW shutdown requested.\n"); set_bit(ISP_ABORT_NEEDED, &vha->dpc_flags); qla2xxx_wake_dpc(vha); break; case QLAFX00_MBA_PORT_UPDATE: /* Port database update */ ha->aenmb[1] = rd_reg_dword(®->aenmailbox1); ha->aenmb[2] = rd_reg_dword(®->aenmailbox2); ha->aenmb[3] = rd_reg_dword(®->aenmailbox3); ql_dbg(ql_dbg_async, vha, 0x5077, "Asynchronous port Update received " "aenmb[0]: %x, aenmb[1]: %x, aenmb[2]: %x, aenmb[3]: %x\n", ha->aenmb[0], ha->aenmb[1], ha->aenmb[2], ha->aenmb[3]); data_size = 4; break; case QLAFX00_MBA_TEMP_OVER: /* Over temperature event */ ql_log(ql_log_info, vha, 0x5085, "Asynchronous over temperature event received " "aenmb[0]: %x\n", ha->aenmb[0]); break; case QLAFX00_MBA_TEMP_NORM: /* Normal temperature event */ ql_log(ql_log_info, vha, 0x5086, "Asynchronous normal temperature event received " "aenmb[0]: %x\n", ha->aenmb[0]); break; case QLAFX00_MBA_TEMP_CRIT: /* Critical temperature event */ ql_log(ql_log_info, vha, 0x5083, "Asynchronous critical temperature event received " "aenmb[0]: %x\n", ha->aenmb[0]); break; default: ha->aenmb[1] = rd_reg_dword(®->aenmailbox1); ha->aenmb[2] = rd_reg_dword(®->aenmailbox2); ha->aenmb[3] = rd_reg_dword(®->aenmailbox3); ha->aenmb[4] = rd_reg_dword(®->aenmailbox4); ha->aenmb[5] = rd_reg_dword(®->aenmailbox5); ha->aenmb[6] = rd_reg_dword(®->aenmailbox6); ha->aenmb[7] = rd_reg_dword(®->aenmailbox7); ql_dbg(ql_dbg_async, vha, 0x5078, "AEN:%04x %04x %04x %04x :%04x %04x %04x %04x\n", ha->aenmb[0], ha->aenmb[1], ha->aenmb[2], ha->aenmb[3], ha->aenmb[4], ha->aenmb[5], ha->aenmb[6], ha->aenmb[7]); break; } qlafx00_post_aenfx_work(vha, ha->aenmb[0], (uint32_t *)ha->aenmb, data_size); } /** * qlafx00_mbx_completion() - Process mailbox command completions. * @vha: SCSI driver HA context * @mb0: value to be written into mailbox register 0 */ static void qlafx00_mbx_completion(scsi_qla_host_t *vha, uint32_t mb0) { uint16_t cnt; __le32 __iomem *wptr; struct qla_hw_data *ha = vha->hw; struct device_reg_fx00 __iomem *reg = &ha->iobase->ispfx00; if (!ha->mcp32) ql_dbg(ql_dbg_async, vha, 0x507e, "MBX pointer ERROR.\n"); /* Load return mailbox registers. */ ha->flags.mbox_int = 1; ha->mailbox_out32[0] = mb0; wptr = ®->mailbox17; for (cnt = 1; cnt < ha->mbx_count; cnt++) { ha->mailbox_out32[cnt] = rd_reg_dword(wptr); wptr++; } } /** * qlafx00_intr_handler() - Process interrupts for the ISPFX00. * @irq: interrupt number * @dev_id: SCSI driver HA context * * Called by system whenever the host adapter generates an interrupt. * * Returns handled flag. */ irqreturn_t qlafx00_intr_handler(int irq, void *dev_id) { scsi_qla_host_t *vha; struct qla_hw_data *ha; struct device_reg_fx00 __iomem *reg; int status; unsigned long iter; uint32_t stat; uint32_t mb[8]; struct rsp_que *rsp; unsigned long flags; uint32_t clr_intr = 0; uint32_t intr_stat = 0; rsp = (struct rsp_que *) dev_id; if (!rsp) { ql_log(ql_log_info, NULL, 0x507d, "%s: NULL response queue pointer.\n", __func__); return IRQ_NONE; } ha = rsp->hw; reg = &ha->iobase->ispfx00; status = 0; if (unlikely(pci_channel_offline(ha->pdev))) return IRQ_HANDLED; spin_lock_irqsave(&ha->hardware_lock, flags); vha = pci_get_drvdata(ha->pdev); for (iter = 50; iter--; clr_intr = 0) { stat = QLAFX00_RD_INTR_REG(ha); if (qla2x00_check_reg32_for_disconnect(vha, stat)) break; intr_stat = stat & QLAFX00_HST_INT_STS_BITS; if (!intr_stat) break; if (stat & QLAFX00_INTR_MB_CMPLT) { mb[0] = rd_reg_dword(®->mailbox16); qlafx00_mbx_completion(vha, mb[0]); status |= MBX_INTERRUPT; clr_intr |= QLAFX00_INTR_MB_CMPLT; } if (intr_stat & QLAFX00_INTR_ASYNC_CMPLT) { ha->aenmb[0] = rd_reg_dword(®->aenmailbox0); qlafx00_async_event(vha); clr_intr |= QLAFX00_INTR_ASYNC_CMPLT; } if (intr_stat & QLAFX00_INTR_RSP_CMPLT) { qlafx00_process_response_queue(vha, rsp); clr_intr |= QLAFX00_INTR_RSP_CMPLT; } QLAFX00_CLR_INTR_REG(ha, clr_intr); QLAFX00_RD_INTR_REG(ha); } qla2x00_handle_mbx_completion(ha, status); spin_unlock_irqrestore(&ha->hardware_lock, flags); return IRQ_HANDLED; } /** QLAFX00 specific IOCB implementation functions */ static inline cont_a64_entry_t * qlafx00_prep_cont_type1_iocb(struct req_que *req, cont_a64_entry_t *lcont_pkt) { cont_a64_entry_t *cont_pkt; /* Adjust ring index. */ req->ring_index++; if (req->ring_index == req->length) { req->ring_index = 0; req->ring_ptr = req->ring; } else { req->ring_ptr++; } cont_pkt = (cont_a64_entry_t *)req->ring_ptr; /* Load packet defaults. */ lcont_pkt->entry_type = CONTINUE_A64_TYPE_FX00; return cont_pkt; } static inline void qlafx00_build_scsi_iocbs(srb_t *sp, struct cmd_type_7_fx00 *cmd_pkt, uint16_t tot_dsds, struct cmd_type_7_fx00 *lcmd_pkt) { uint16_t avail_dsds; struct dsd64 *cur_dsd; scsi_qla_host_t *vha; struct scsi_cmnd *cmd; struct scatterlist *sg; int i, cont; struct req_que *req; cont_a64_entry_t lcont_pkt; cont_a64_entry_t *cont_pkt; vha = sp->vha; req = vha->req; cmd = GET_CMD_SP(sp); cont = 0; cont_pkt = NULL; /* Update entry type to indicate Command Type 3 IOCB */ lcmd_pkt->entry_type = FX00_COMMAND_TYPE_7; /* No data transfer */ if (!scsi_bufflen(cmd) || cmd->sc_data_direction == DMA_NONE) { lcmd_pkt->byte_count = cpu_to_le32(0); return; } /* Set transfer direction */ if (cmd->sc_data_direction == DMA_TO_DEVICE) { lcmd_pkt->cntrl_flags = TMF_WRITE_DATA; vha->qla_stats.output_bytes += scsi_bufflen(cmd); } else if (cmd->sc_data_direction == DMA_FROM_DEVICE) { lcmd_pkt->cntrl_flags = TMF_READ_DATA; vha->qla_stats.input_bytes += scsi_bufflen(cmd); } /* One DSD is available in the Command Type 3 IOCB */ avail_dsds = 1; cur_dsd = &lcmd_pkt->dsd; /* Load data segments */ scsi_for_each_sg(cmd, sg, tot_dsds, i) { /* Allocate additional continuation packets? */ if (avail_dsds == 0) { /* * Five DSDs are available in the Continuation * Type 1 IOCB. */ memset(&lcont_pkt, 0, REQUEST_ENTRY_SIZE); cont_pkt = qlafx00_prep_cont_type1_iocb(req, &lcont_pkt); cur_dsd = lcont_pkt.dsd; avail_dsds = 5; cont = 1; } append_dsd64(&cur_dsd, sg); avail_dsds--; if (avail_dsds == 0 && cont == 1) { cont = 0; memcpy_toio((void __iomem *)cont_pkt, &lcont_pkt, sizeof(lcont_pkt)); } } if (avail_dsds != 0 && cont == 1) { memcpy_toio((void __iomem *)cont_pkt, &lcont_pkt, sizeof(lcont_pkt)); } } /** * qlafx00_start_scsi() - Send a SCSI command to the ISP * @sp: command to send to the ISP * * Returns non-zero if a failure occurred, else zero. */ int qlafx00_start_scsi(srb_t *sp) { int nseg; unsigned long flags; uint32_t handle; uint16_t cnt; uint16_t req_cnt; uint16_t tot_dsds; struct req_que *req = NULL; struct rsp_que *rsp = NULL; struct scsi_cmnd *cmd = GET_CMD_SP(sp); struct scsi_qla_host *vha = sp->vha; struct qla_hw_data *ha = vha->hw; struct cmd_type_7_fx00 *cmd_pkt; struct cmd_type_7_fx00 lcmd_pkt; struct scsi_lun llun; /* Setup device pointers. */ rsp = ha->rsp_q_map[0]; req = vha->req; /* So we know we haven't pci_map'ed anything yet */ tot_dsds = 0; /* Acquire ring specific lock */ spin_lock_irqsave(&ha->hardware_lock, flags); handle = qla2xxx_get_next_handle(req); if (handle == 0) goto queuing_error; /* Map the sg table so we have an accurate count of sg entries needed */ if (scsi_sg_count(cmd)) { nseg = dma_map_sg(&ha->pdev->dev, scsi_sglist(cmd), scsi_sg_count(cmd), cmd->sc_data_direction); if (unlikely(!nseg)) goto queuing_error; } else nseg = 0; tot_dsds = nseg; req_cnt = qla24xx_calc_iocbs(vha, tot_dsds); if (req->cnt < (req_cnt + 2)) { cnt = rd_reg_dword_relaxed(req->req_q_out); if (req->ring_index < cnt) req->cnt = cnt - req->ring_index; else req->cnt = req->length - (req->ring_index - cnt); if (req->cnt < (req_cnt + 2)) goto queuing_error; } /* Build command packet. */ req->current_outstanding_cmd = handle; req->outstanding_cmds[handle] = sp; sp->handle = handle; cmd->host_scribble = (unsigned char *)(unsigned long)handle; req->cnt -= req_cnt; cmd_pkt = (struct cmd_type_7_fx00 *)req->ring_ptr; memset(&lcmd_pkt, 0, REQUEST_ENTRY_SIZE); lcmd_pkt.handle = make_handle(req->id, sp->handle); lcmd_pkt.reserved_0 = 0; lcmd_pkt.port_path_ctrl = 0; lcmd_pkt.reserved_1 = 0; lcmd_pkt.dseg_count = cpu_to_le16(tot_dsds); lcmd_pkt.tgt_idx = cpu_to_le16(sp->fcport->tgt_id); int_to_scsilun(cmd->device->lun, &llun); host_to_adap((uint8_t *)&llun, (uint8_t *)&lcmd_pkt.lun, sizeof(lcmd_pkt.lun)); /* Load SCSI command packet. */ host_to_adap(cmd->cmnd, lcmd_pkt.fcp_cdb, sizeof(lcmd_pkt.fcp_cdb)); lcmd_pkt.byte_count = cpu_to_le32((uint32_t)scsi_bufflen(cmd)); /* Build IOCB segments */ qlafx00_build_scsi_iocbs(sp, cmd_pkt, tot_dsds, &lcmd_pkt); /* Set total data segment count. */ lcmd_pkt.entry_count = (uint8_t)req_cnt; /* Specify response queue number where completion should happen */ lcmd_pkt.entry_status = (uint8_t) rsp->id; ql_dump_buffer(ql_dbg_io + ql_dbg_buffer, vha, 0x302e, cmd->cmnd, cmd->cmd_len); ql_dump_buffer(ql_dbg_io + ql_dbg_buffer, vha, 0x3032, &lcmd_pkt, sizeof(lcmd_pkt)); memcpy_toio((void __iomem *)cmd_pkt, &lcmd_pkt, REQUEST_ENTRY_SIZE); wmb(); /* Adjust ring index. */ req->ring_index++; if (req->ring_index == req->length) { req->ring_index = 0; req->ring_ptr = req->ring; } else req->ring_ptr++; sp->flags |= SRB_DMA_VALID; /* Set chip new ring index. */ wrt_reg_dword(req->req_q_in, req->ring_index); QLAFX00_SET_HST_INTR(ha, ha->rqstq_intr_code); spin_unlock_irqrestore(&ha->hardware_lock, flags); return QLA_SUCCESS; queuing_error: if (tot_dsds) scsi_dma_unmap(cmd); spin_unlock_irqrestore(&ha->hardware_lock, flags); return QLA_FUNCTION_FAILED; } void qlafx00_tm_iocb(srb_t *sp, struct tsk_mgmt_entry_fx00 *ptm_iocb) { struct srb_iocb *fxio = &sp->u.iocb_cmd; scsi_qla_host_t *vha = sp->vha; struct req_que *req = vha->req; struct tsk_mgmt_entry_fx00 tm_iocb; struct scsi_lun llun; memset(&tm_iocb, 0, sizeof(struct tsk_mgmt_entry_fx00)); tm_iocb.entry_type = TSK_MGMT_IOCB_TYPE_FX00; tm_iocb.entry_count = 1; tm_iocb.handle = make_handle(req->id, sp->handle); tm_iocb.reserved_0 = 0; tm_iocb.tgt_id = cpu_to_le16(sp->fcport->tgt_id); tm_iocb.control_flags = cpu_to_le32(fxio->u.tmf.flags); if (tm_iocb.control_flags == cpu_to_le32((uint32_t)TCF_LUN_RESET)) { int_to_scsilun(fxio->u.tmf.lun, &llun); host_to_adap((uint8_t *)&llun, (uint8_t *)&tm_iocb.lun, sizeof(struct scsi_lun)); } memcpy(ptm_iocb, &tm_iocb, sizeof(struct tsk_mgmt_entry_fx00)); wmb(); } void qlafx00_abort_iocb(srb_t *sp, struct abort_iocb_entry_fx00 *pabt_iocb) { struct srb_iocb *fxio = &sp->u.iocb_cmd; scsi_qla_host_t *vha = sp->vha; struct req_que *req = vha->req; struct abort_iocb_entry_fx00 abt_iocb; memset(&abt_iocb, 0, sizeof(struct abort_iocb_entry_fx00)); abt_iocb.entry_type = ABORT_IOCB_TYPE_FX00; abt_iocb.entry_count = 1; abt_iocb.handle = make_handle(req->id, sp->handle); abt_iocb.abort_handle = make_handle(req->id, fxio->u.abt.cmd_hndl); abt_iocb.tgt_id_sts = cpu_to_le16(sp->fcport->tgt_id); abt_iocb.req_que_no = cpu_to_le16(req->id); memcpy(pabt_iocb, &abt_iocb, sizeof(struct abort_iocb_entry_fx00)); wmb(); } void qlafx00_fxdisc_iocb(srb_t *sp, struct fxdisc_entry_fx00 *pfxiocb) { struct srb_iocb *fxio = &sp->u.iocb_cmd; struct qla_mt_iocb_rqst_fx00 *piocb_rqst; struct bsg_job *bsg_job; struct fc_bsg_request *bsg_request; struct fxdisc_entry_fx00 fx_iocb; uint8_t entry_cnt = 1; memset(&fx_iocb, 0, sizeof(struct fxdisc_entry_fx00)); fx_iocb.entry_type = FX00_IOCB_TYPE; fx_iocb.handle = sp->handle; fx_iocb.entry_count = entry_cnt; if (sp->type == SRB_FXIOCB_DCMD) { fx_iocb.func_num = sp->u.iocb_cmd.u.fxiocb.req_func_type; fx_iocb.adapid = fxio->u.fxiocb.adapter_id; fx_iocb.adapid_hi = fxio->u.fxiocb.adapter_id_hi; fx_iocb.reserved_0 = fxio->u.fxiocb.reserved_0; fx_iocb.reserved_1 = fxio->u.fxiocb.reserved_1; fx_iocb.dataword_extra = fxio->u.fxiocb.req_data_extra; if (fxio->u.fxiocb.flags & SRB_FXDISC_REQ_DMA_VALID) { fx_iocb.req_dsdcnt = cpu_to_le16(1); fx_iocb.req_xfrcnt = cpu_to_le16(fxio->u.fxiocb.req_len); put_unaligned_le64(fxio->u.fxiocb.req_dma_handle, &fx_iocb.dseg_rq[0].address); fx_iocb.dseg_rq[0].length = cpu_to_le32(fxio->u.fxiocb.req_len); } if (fxio->u.fxiocb.flags & SRB_FXDISC_RESP_DMA_VALID) { fx_iocb.rsp_dsdcnt = cpu_to_le16(1); fx_iocb.rsp_xfrcnt = cpu_to_le16(fxio->u.fxiocb.rsp_len); put_unaligned_le64(fxio->u.fxiocb.rsp_dma_handle, &fx_iocb.dseg_rsp[0].address); fx_iocb.dseg_rsp[0].length = cpu_to_le32(fxio->u.fxiocb.rsp_len); } if (fxio->u.fxiocb.flags & SRB_FXDISC_REQ_DWRD_VALID) { fx_iocb.dataword = fxio->u.fxiocb.req_data; } fx_iocb.flags = fxio->u.fxiocb.flags; } else { struct scatterlist *sg; bsg_job = sp->u.bsg_job; bsg_request = bsg_job->request; piocb_rqst = (struct qla_mt_iocb_rqst_fx00 *) &bsg_request->rqst_data.h_vendor.vendor_cmd[1]; fx_iocb.func_num = piocb_rqst->func_type; fx_iocb.adapid = piocb_rqst->adapid; fx_iocb.adapid_hi = piocb_rqst->adapid_hi; fx_iocb.reserved_0 = piocb_rqst->reserved_0; fx_iocb.reserved_1 = piocb_rqst->reserved_1; fx_iocb.dataword_extra = piocb_rqst->dataword_extra; fx_iocb.dataword = piocb_rqst->dataword; fx_iocb.req_xfrcnt = piocb_rqst->req_len; fx_iocb.rsp_xfrcnt = piocb_rqst->rsp_len; if (piocb_rqst->flags & SRB_FXDISC_REQ_DMA_VALID) { int avail_dsds, tot_dsds; cont_a64_entry_t lcont_pkt; cont_a64_entry_t *cont_pkt = NULL; struct dsd64 *cur_dsd; int index = 0, cont = 0; fx_iocb.req_dsdcnt = cpu_to_le16(bsg_job->request_payload.sg_cnt); tot_dsds = bsg_job->request_payload.sg_cnt; cur_dsd = &fx_iocb.dseg_rq[0]; avail_dsds = 1; for_each_sg(bsg_job->request_payload.sg_list, sg, tot_dsds, index) { /* Allocate additional continuation packets? */ if (avail_dsds == 0) { /* * Five DSDs are available in the Cont. * Type 1 IOCB. */ memset(&lcont_pkt, 0, REQUEST_ENTRY_SIZE); cont_pkt = qlafx00_prep_cont_type1_iocb( sp->vha->req, &lcont_pkt); cur_dsd = lcont_pkt.dsd; avail_dsds = 5; cont = 1; entry_cnt++; } append_dsd64(&cur_dsd, sg); avail_dsds--; if (avail_dsds == 0 && cont == 1) { cont = 0; memcpy_toio( (void __iomem *)cont_pkt, &lcont_pkt, REQUEST_ENTRY_SIZE); ql_dump_buffer( ql_dbg_user + ql_dbg_verbose, sp->vha, 0x3042, (uint8_t *)&lcont_pkt, REQUEST_ENTRY_SIZE); } } if (avail_dsds != 0 && cont == 1) { memcpy_toio((void __iomem *)cont_pkt, &lcont_pkt, REQUEST_ENTRY_SIZE); ql_dump_buffer(ql_dbg_user + ql_dbg_verbose, sp->vha, 0x3043, (uint8_t *)&lcont_pkt, REQUEST_ENTRY_SIZE); } } if (piocb_rqst->flags & SRB_FXDISC_RESP_DMA_VALID) { int avail_dsds, tot_dsds; cont_a64_entry_t lcont_pkt; cont_a64_entry_t *cont_pkt = NULL; struct dsd64 *cur_dsd; int index = 0, cont = 0; fx_iocb.rsp_dsdcnt = cpu_to_le16(bsg_job->reply_payload.sg_cnt); tot_dsds = bsg_job->reply_payload.sg_cnt; cur_dsd = &fx_iocb.dseg_rsp[0]; avail_dsds = 1; for_each_sg(bsg_job->reply_payload.sg_list, sg, tot_dsds, index) { /* Allocate additional continuation packets? */ if (avail_dsds == 0) { /* * Five DSDs are available in the Cont. * Type 1 IOCB. */ memset(&lcont_pkt, 0, REQUEST_ENTRY_SIZE); cont_pkt = qlafx00_prep_cont_type1_iocb( sp->vha->req, &lcont_pkt); cur_dsd = lcont_pkt.dsd; avail_dsds = 5; cont = 1; entry_cnt++; } append_dsd64(&cur_dsd, sg); avail_dsds--; if (avail_dsds == 0 && cont == 1) { cont = 0; memcpy_toio((void __iomem *)cont_pkt, &lcont_pkt, REQUEST_ENTRY_SIZE); ql_dump_buffer( ql_dbg_user + ql_dbg_verbose, sp->vha, 0x3045, (uint8_t *)&lcont_pkt, REQUEST_ENTRY_SIZE); } } if (avail_dsds != 0 && cont == 1) { memcpy_toio((void __iomem *)cont_pkt, &lcont_pkt, REQUEST_ENTRY_SIZE); ql_dump_buffer(ql_dbg_user + ql_dbg_verbose, sp->vha, 0x3046, (uint8_t *)&lcont_pkt, REQUEST_ENTRY_SIZE); } } if (piocb_rqst->flags & SRB_FXDISC_REQ_DWRD_VALID) fx_iocb.dataword = piocb_rqst->dataword; fx_iocb.flags = piocb_rqst->flags; fx_iocb.entry_count = entry_cnt; } ql_dump_buffer(ql_dbg_user + ql_dbg_verbose, sp->vha, 0x3047, &fx_iocb, sizeof(fx_iocb)); memcpy_toio((void __iomem *)pfxiocb, &fx_iocb, sizeof(fx_iocb)); wmb(); }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1