Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Quinn Tran | 6927 | 30.94% | 53 | 29.94% |
James Bottomley | 4299 | 19.20% | 4 | 2.26% |
Joe Carnuccio | 3309 | 14.78% | 10 | 5.65% |
Himanshu Madhani | 2161 | 9.65% | 5 | 2.82% |
Andrew Vasquez | 2158 | 9.64% | 34 | 19.21% |
Giridhar Malavali | 880 | 3.93% | 8 | 4.52% |
Saurav Kashyap | 708 | 3.16% | 5 | 2.82% |
Anirban Chakraborty | 519 | 2.32% | 4 | 2.26% |
Chad Dupuis | 386 | 1.72% | 5 | 2.82% |
Duane Grigsby | 342 | 1.53% | 2 | 1.13% |
Bart Van Assche | 279 | 1.25% | 10 | 5.65% |
Michael Hernandez | 97 | 0.43% | 2 | 1.13% |
Bill Kuzeja | 51 | 0.23% | 1 | 0.56% |
Martin Wilck | 37 | 0.17% | 1 | 0.56% |
Seokmann Ju | 36 | 0.16% | 2 | 1.13% |
Alexei Potashnik | 36 | 0.16% | 3 | 1.69% |
Daniel Wagner | 27 | 0.12% | 1 | 0.56% |
Armen Baloyan | 22 | 0.10% | 1 | 0.56% |
Madhuranath Iyengar | 19 | 0.08% | 3 | 1.69% |
Nicholas Bellinger | 19 | 0.08% | 1 | 0.56% |
Arun Easi | 19 | 0.08% | 4 | 2.26% |
Roman Bolshakov | 14 | 0.06% | 1 | 0.56% |
Lalit Chandivade | 6 | 0.03% | 1 | 0.56% |
Lee Jones | 5 | 0.02% | 1 | 0.56% |
Sawan Chandak | 5 | 0.02% | 2 | 1.13% |
Oleksandr Khoshaba | 5 | 0.02% | 1 | 0.56% |
David S. Miller | 4 | 0.02% | 1 | 0.56% |
Bikash Hazarika | 3 | 0.01% | 2 | 1.13% |
Luis R. Rodriguez | 2 | 0.01% | 1 | 0.56% |
Thomas Gleixner | 2 | 0.01% | 1 | 0.56% |
Manish Rangankar | 2 | 0.01% | 1 | 0.56% |
Gustavo A. R. Silva | 2 | 0.01% | 1 | 0.56% |
Dan Carpenter | 2 | 0.01% | 1 | 0.56% |
Sarang Radke | 1 | 0.00% | 1 | 0.56% |
Milan P. Gandhi | 1 | 0.00% | 1 | 0.56% |
Colin Ian King | 1 | 0.00% | 1 | 0.56% |
Julia Lawall | 1 | 0.00% | 1 | 0.56% |
Total | 22387 | 177 |
// SPDX-License-Identifier: GPL-2.0-only /* * QLogic Fibre Channel HBA Driver * Copyright (c) 2003-2014 QLogic Corporation */ #include "qla_def.h" #include "qla_target.h" #include <linux/utsname.h> static int qla2x00_sns_ga_nxt(scsi_qla_host_t *, fc_port_t *); static int qla2x00_sns_gid_pt(scsi_qla_host_t *, sw_info_t *); static int qla2x00_sns_gpn_id(scsi_qla_host_t *, sw_info_t *); static int qla2x00_sns_gnn_id(scsi_qla_host_t *, sw_info_t *); static int qla2x00_sns_rft_id(scsi_qla_host_t *); static int qla2x00_sns_rnn_id(scsi_qla_host_t *); static int qla_async_rftid(scsi_qla_host_t *, port_id_t *); static int qla_async_rffid(scsi_qla_host_t *, port_id_t *, u8, u8); static int qla_async_rnnid(scsi_qla_host_t *, port_id_t *, u8*); static int qla_async_rsnn_nn(scsi_qla_host_t *); /** * qla2x00_prep_ms_iocb() - Prepare common MS/CT IOCB fields for SNS CT query. * @vha: HA context * @arg: CT arguments * * Returns a pointer to the @vha's ms_iocb. */ void * qla2x00_prep_ms_iocb(scsi_qla_host_t *vha, struct ct_arg *arg) { struct qla_hw_data *ha = vha->hw; ms_iocb_entry_t *ms_pkt; ms_pkt = (ms_iocb_entry_t *)arg->iocb; memset(ms_pkt, 0, sizeof(ms_iocb_entry_t)); ms_pkt->entry_type = MS_IOCB_TYPE; ms_pkt->entry_count = 1; SET_TARGET_ID(ha, ms_pkt->loop_id, SIMPLE_NAME_SERVER); ms_pkt->control_flags = cpu_to_le16(CF_READ | CF_HEAD_TAG); ms_pkt->timeout = cpu_to_le16(ha->r_a_tov / 10 * 2); ms_pkt->cmd_dsd_count = cpu_to_le16(1); ms_pkt->total_dsd_count = cpu_to_le16(2); ms_pkt->rsp_bytecount = cpu_to_le32(arg->rsp_size); ms_pkt->req_bytecount = cpu_to_le32(arg->req_size); put_unaligned_le64(arg->req_dma, &ms_pkt->req_dsd.address); ms_pkt->req_dsd.length = ms_pkt->req_bytecount; put_unaligned_le64(arg->rsp_dma, &ms_pkt->rsp_dsd.address); ms_pkt->rsp_dsd.length = ms_pkt->rsp_bytecount; vha->qla_stats.control_requests++; return (ms_pkt); } /** * qla24xx_prep_ms_iocb() - Prepare common CT IOCB fields for SNS CT query. * @vha: HA context * @arg: CT arguments * * Returns a pointer to the @ha's ms_iocb. */ void * qla24xx_prep_ms_iocb(scsi_qla_host_t *vha, struct ct_arg *arg) { struct qla_hw_data *ha = vha->hw; struct ct_entry_24xx *ct_pkt; ct_pkt = (struct ct_entry_24xx *)arg->iocb; memset(ct_pkt, 0, sizeof(struct ct_entry_24xx)); ct_pkt->entry_type = CT_IOCB_TYPE; ct_pkt->entry_count = 1; ct_pkt->nport_handle = cpu_to_le16(arg->nport_handle); ct_pkt->timeout = cpu_to_le16(ha->r_a_tov / 10 * 2); ct_pkt->cmd_dsd_count = cpu_to_le16(1); ct_pkt->rsp_dsd_count = cpu_to_le16(1); ct_pkt->rsp_byte_count = cpu_to_le32(arg->rsp_size); ct_pkt->cmd_byte_count = cpu_to_le32(arg->req_size); put_unaligned_le64(arg->req_dma, &ct_pkt->dsd[0].address); ct_pkt->dsd[0].length = ct_pkt->cmd_byte_count; put_unaligned_le64(arg->rsp_dma, &ct_pkt->dsd[1].address); ct_pkt->dsd[1].length = ct_pkt->rsp_byte_count; ct_pkt->vp_index = vha->vp_idx; vha->qla_stats.control_requests++; return (ct_pkt); } /** * qla2x00_prep_ct_req() - Prepare common CT request fields for SNS query. * @p: CT request buffer * @cmd: GS command * @rsp_size: response size in bytes * * Returns a pointer to the intitialized @ct_req. */ static inline struct ct_sns_req * qla2x00_prep_ct_req(struct ct_sns_pkt *p, uint16_t cmd, uint16_t rsp_size) { memset(p, 0, sizeof(struct ct_sns_pkt)); p->p.req.header.revision = 0x01; p->p.req.header.gs_type = 0xFC; p->p.req.header.gs_subtype = 0x02; p->p.req.command = cpu_to_be16(cmd); p->p.req.max_rsp_size = cpu_to_be16((rsp_size - 16) / 4); return &p->p.req; } int qla2x00_chk_ms_status(scsi_qla_host_t *vha, ms_iocb_entry_t *ms_pkt, struct ct_sns_rsp *ct_rsp, const char *routine) { int rval; uint16_t comp_status; struct qla_hw_data *ha = vha->hw; bool lid_is_sns = false; rval = QLA_FUNCTION_FAILED; if (ms_pkt->entry_status != 0) { ql_dbg(ql_dbg_disc, vha, 0x2031, "%s failed, error status (%x) on port_id: %02x%02x%02x.\n", routine, ms_pkt->entry_status, vha->d_id.b.domain, vha->d_id.b.area, vha->d_id.b.al_pa); } else { if (IS_FWI2_CAPABLE(ha)) comp_status = le16_to_cpu( ((struct ct_entry_24xx *)ms_pkt)->comp_status); else comp_status = le16_to_cpu(ms_pkt->status); switch (comp_status) { case CS_COMPLETE: case CS_DATA_UNDERRUN: case CS_DATA_OVERRUN: /* Overrun? */ if (ct_rsp->header.response != cpu_to_be16(CT_ACCEPT_RESPONSE)) { ql_dbg(ql_dbg_disc + ql_dbg_buffer, vha, 0x2077, "%s failed rejected request on port_id: %02x%02x%02x Completion status 0x%x, response 0x%x\n", routine, vha->d_id.b.domain, vha->d_id.b.area, vha->d_id.b.al_pa, comp_status, ct_rsp->header.response); ql_dump_buffer(ql_dbg_disc + ql_dbg_buffer, vha, 0x2078, ct_rsp, offsetof(typeof(*ct_rsp), rsp)); rval = QLA_INVALID_COMMAND; } else rval = QLA_SUCCESS; break; case CS_PORT_LOGGED_OUT: if (IS_FWI2_CAPABLE(ha)) { if (le16_to_cpu(ms_pkt->loop_id.extended) == NPH_SNS) lid_is_sns = true; } else { if (le16_to_cpu(ms_pkt->loop_id.extended) == SIMPLE_NAME_SERVER) lid_is_sns = true; } if (lid_is_sns) { ql_dbg(ql_dbg_async, vha, 0x502b, "%s failed, Name server has logged out", routine); rval = QLA_NOT_LOGGED_IN; set_bit(LOOP_RESYNC_NEEDED, &vha->dpc_flags); set_bit(LOCAL_LOOP_UPDATE, &vha->dpc_flags); } break; case CS_TIMEOUT: rval = QLA_FUNCTION_TIMEOUT; fallthrough; default: ql_dbg(ql_dbg_disc, vha, 0x2033, "%s failed, completion status (%x) on port_id: " "%02x%02x%02x.\n", routine, comp_status, vha->d_id.b.domain, vha->d_id.b.area, vha->d_id.b.al_pa); break; } } return rval; } /** * qla2x00_ga_nxt() - SNS scan for fabric devices via GA_NXT command. * @vha: HA context * @fcport: fcport entry to updated * * Returns 0 on success. */ int qla2x00_ga_nxt(scsi_qla_host_t *vha, fc_port_t *fcport) { int rval; ms_iocb_entry_t *ms_pkt; struct ct_sns_req *ct_req; struct ct_sns_rsp *ct_rsp; struct qla_hw_data *ha = vha->hw; struct ct_arg arg; if (IS_QLA2100(ha) || IS_QLA2200(ha)) return qla2x00_sns_ga_nxt(vha, fcport); arg.iocb = ha->ms_iocb; arg.req_dma = ha->ct_sns_dma; arg.rsp_dma = ha->ct_sns_dma; arg.req_size = GA_NXT_REQ_SIZE; arg.rsp_size = GA_NXT_RSP_SIZE; arg.nport_handle = NPH_SNS; /* Issue GA_NXT */ /* Prepare common MS IOCB */ ms_pkt = ha->isp_ops->prep_ms_iocb(vha, &arg); /* Prepare CT request */ ct_req = qla2x00_prep_ct_req(ha->ct_sns, GA_NXT_CMD, GA_NXT_RSP_SIZE); ct_rsp = &ha->ct_sns->p.rsp; /* Prepare CT arguments -- port_id */ ct_req->req.port_id.port_id = port_id_to_be_id(fcport->d_id); /* Execute MS IOCB */ rval = qla2x00_issue_iocb(vha, ha->ms_iocb, ha->ms_iocb_dma, sizeof(ms_iocb_entry_t)); if (rval != QLA_SUCCESS) { /*EMPTY*/ ql_dbg(ql_dbg_disc, vha, 0x2062, "GA_NXT issue IOCB failed (%d).\n", rval); } else if (qla2x00_chk_ms_status(vha, ms_pkt, ct_rsp, "GA_NXT") != QLA_SUCCESS) { rval = QLA_FUNCTION_FAILED; } else { /* Populate fc_port_t entry. */ fcport->d_id = be_to_port_id(ct_rsp->rsp.ga_nxt.port_id); memcpy(fcport->node_name, ct_rsp->rsp.ga_nxt.node_name, WWN_SIZE); memcpy(fcport->port_name, ct_rsp->rsp.ga_nxt.port_name, WWN_SIZE); fcport->fc4_type = (ct_rsp->rsp.ga_nxt.fc4_types[2] & BIT_0) ? FS_FC4TYPE_FCP : FC4_TYPE_OTHER; if (ct_rsp->rsp.ga_nxt.port_type != NS_N_PORT_TYPE && ct_rsp->rsp.ga_nxt.port_type != NS_NL_PORT_TYPE) fcport->d_id.b.domain = 0xf0; ql_dbg(ql_dbg_disc, vha, 0x2063, "GA_NXT entry - nn %8phN pn %8phN " "port_id=%02x%02x%02x.\n", fcport->node_name, fcport->port_name, fcport->d_id.b.domain, fcport->d_id.b.area, fcport->d_id.b.al_pa); } return (rval); } static inline int qla2x00_gid_pt_rsp_size(scsi_qla_host_t *vha) { return vha->hw->max_fibre_devices * 4 + 16; } /** * qla2x00_gid_pt() - SNS scan for fabric devices via GID_PT command. * @vha: HA context * @list: switch info entries to populate * * NOTE: Non-Nx_Ports are not requested. * * Returns 0 on success. */ int qla2x00_gid_pt(scsi_qla_host_t *vha, sw_info_t *list) { int rval; uint16_t i; ms_iocb_entry_t *ms_pkt; struct ct_sns_req *ct_req; struct ct_sns_rsp *ct_rsp; struct ct_sns_gid_pt_data *gid_data; struct qla_hw_data *ha = vha->hw; uint16_t gid_pt_rsp_size; struct ct_arg arg; if (IS_QLA2100(ha) || IS_QLA2200(ha)) return qla2x00_sns_gid_pt(vha, list); gid_data = NULL; gid_pt_rsp_size = qla2x00_gid_pt_rsp_size(vha); arg.iocb = ha->ms_iocb; arg.req_dma = ha->ct_sns_dma; arg.rsp_dma = ha->ct_sns_dma; arg.req_size = GID_PT_REQ_SIZE; arg.rsp_size = gid_pt_rsp_size; arg.nport_handle = NPH_SNS; /* Issue GID_PT */ /* Prepare common MS IOCB */ ms_pkt = ha->isp_ops->prep_ms_iocb(vha, &arg); /* Prepare CT request */ ct_req = qla2x00_prep_ct_req(ha->ct_sns, GID_PT_CMD, gid_pt_rsp_size); ct_rsp = &ha->ct_sns->p.rsp; /* Prepare CT arguments -- port_type */ ct_req->req.gid_pt.port_type = NS_NX_PORT_TYPE; /* Execute MS IOCB */ rval = qla2x00_issue_iocb(vha, ha->ms_iocb, ha->ms_iocb_dma, sizeof(ms_iocb_entry_t)); if (rval != QLA_SUCCESS) { /*EMPTY*/ ql_dbg(ql_dbg_disc, vha, 0x2055, "GID_PT issue IOCB failed (%d).\n", rval); } else if (qla2x00_chk_ms_status(vha, ms_pkt, ct_rsp, "GID_PT") != QLA_SUCCESS) { rval = QLA_FUNCTION_FAILED; } else { /* Set port IDs in switch info list. */ for (i = 0; i < ha->max_fibre_devices; i++) { gid_data = &ct_rsp->rsp.gid_pt.entries[i]; list[i].d_id = be_to_port_id(gid_data->port_id); memset(list[i].fabric_port_name, 0, WWN_SIZE); list[i].fp_speed = PORT_SPEED_UNKNOWN; /* Last one exit. */ if (gid_data->control_byte & BIT_7) { list[i].d_id.b.rsvd_1 = gid_data->control_byte; break; } } /* * If we've used all available slots, then the switch is * reporting back more devices than we can handle with this * single call. Return a failed status, and let GA_NXT handle * the overload. */ if (i == ha->max_fibre_devices) rval = QLA_FUNCTION_FAILED; } return (rval); } /** * qla2x00_gpn_id() - SNS Get Port Name (GPN_ID) query. * @vha: HA context * @list: switch info entries to populate * * Returns 0 on success. */ int qla2x00_gpn_id(scsi_qla_host_t *vha, sw_info_t *list) { int rval = QLA_SUCCESS; uint16_t i; ms_iocb_entry_t *ms_pkt; struct ct_sns_req *ct_req; struct ct_sns_rsp *ct_rsp; struct qla_hw_data *ha = vha->hw; struct ct_arg arg; if (IS_QLA2100(ha) || IS_QLA2200(ha)) return qla2x00_sns_gpn_id(vha, list); arg.iocb = ha->ms_iocb; arg.req_dma = ha->ct_sns_dma; arg.rsp_dma = ha->ct_sns_dma; arg.req_size = GPN_ID_REQ_SIZE; arg.rsp_size = GPN_ID_RSP_SIZE; arg.nport_handle = NPH_SNS; for (i = 0; i < ha->max_fibre_devices; i++) { /* Issue GPN_ID */ /* Prepare common MS IOCB */ ms_pkt = ha->isp_ops->prep_ms_iocb(vha, &arg); /* Prepare CT request */ ct_req = qla2x00_prep_ct_req(ha->ct_sns, GPN_ID_CMD, GPN_ID_RSP_SIZE); ct_rsp = &ha->ct_sns->p.rsp; /* Prepare CT arguments -- port_id */ ct_req->req.port_id.port_id = port_id_to_be_id(list[i].d_id); /* Execute MS IOCB */ rval = qla2x00_issue_iocb(vha, ha->ms_iocb, ha->ms_iocb_dma, sizeof(ms_iocb_entry_t)); if (rval != QLA_SUCCESS) { /*EMPTY*/ ql_dbg(ql_dbg_disc, vha, 0x2056, "GPN_ID issue IOCB failed (%d).\n", rval); break; } else if (qla2x00_chk_ms_status(vha, ms_pkt, ct_rsp, "GPN_ID") != QLA_SUCCESS) { rval = QLA_FUNCTION_FAILED; break; } else { /* Save portname */ memcpy(list[i].port_name, ct_rsp->rsp.gpn_id.port_name, WWN_SIZE); } /* Last device exit. */ if (list[i].d_id.b.rsvd_1 != 0) break; } return (rval); } /** * qla2x00_gnn_id() - SNS Get Node Name (GNN_ID) query. * @vha: HA context * @list: switch info entries to populate * * Returns 0 on success. */ int qla2x00_gnn_id(scsi_qla_host_t *vha, sw_info_t *list) { int rval = QLA_SUCCESS; uint16_t i; struct qla_hw_data *ha = vha->hw; ms_iocb_entry_t *ms_pkt; struct ct_sns_req *ct_req; struct ct_sns_rsp *ct_rsp; struct ct_arg arg; if (IS_QLA2100(ha) || IS_QLA2200(ha)) return qla2x00_sns_gnn_id(vha, list); arg.iocb = ha->ms_iocb; arg.req_dma = ha->ct_sns_dma; arg.rsp_dma = ha->ct_sns_dma; arg.req_size = GNN_ID_REQ_SIZE; arg.rsp_size = GNN_ID_RSP_SIZE; arg.nport_handle = NPH_SNS; for (i = 0; i < ha->max_fibre_devices; i++) { /* Issue GNN_ID */ /* Prepare common MS IOCB */ ms_pkt = ha->isp_ops->prep_ms_iocb(vha, &arg); /* Prepare CT request */ ct_req = qla2x00_prep_ct_req(ha->ct_sns, GNN_ID_CMD, GNN_ID_RSP_SIZE); ct_rsp = &ha->ct_sns->p.rsp; /* Prepare CT arguments -- port_id */ ct_req->req.port_id.port_id = port_id_to_be_id(list[i].d_id); /* Execute MS IOCB */ rval = qla2x00_issue_iocb(vha, ha->ms_iocb, ha->ms_iocb_dma, sizeof(ms_iocb_entry_t)); if (rval != QLA_SUCCESS) { /*EMPTY*/ ql_dbg(ql_dbg_disc, vha, 0x2057, "GNN_ID issue IOCB failed (%d).\n", rval); break; } else if (qla2x00_chk_ms_status(vha, ms_pkt, ct_rsp, "GNN_ID") != QLA_SUCCESS) { rval = QLA_FUNCTION_FAILED; break; } else { /* Save nodename */ memcpy(list[i].node_name, ct_rsp->rsp.gnn_id.node_name, WWN_SIZE); ql_dbg(ql_dbg_disc, vha, 0x2058, "GID_PT entry - nn %8phN pn %8phN " "portid=%02x%02x%02x.\n", list[i].node_name, list[i].port_name, list[i].d_id.b.domain, list[i].d_id.b.area, list[i].d_id.b.al_pa); } /* Last device exit. */ if (list[i].d_id.b.rsvd_1 != 0) break; } return (rval); } static void qla2x00_async_sns_sp_done(srb_t *sp, int rc) { struct scsi_qla_host *vha = sp->vha; struct ct_sns_pkt *ct_sns; struct qla_work_evt *e; sp->rc = rc; if (rc == QLA_SUCCESS) { ql_dbg(ql_dbg_disc, vha, 0x204f, "Async done-%s exiting normally.\n", sp->name); } else if (rc == QLA_FUNCTION_TIMEOUT) { ql_dbg(ql_dbg_disc, vha, 0x204f, "Async done-%s timeout\n", sp->name); } else { ct_sns = (struct ct_sns_pkt *)sp->u.iocb_cmd.u.ctarg.rsp; memset(ct_sns, 0, sizeof(*ct_sns)); sp->retry_count++; if (sp->retry_count > 3) goto err; ql_dbg(ql_dbg_disc, vha, 0x204f, "Async done-%s fail rc %x. Retry count %d\n", sp->name, rc, sp->retry_count); e = qla2x00_alloc_work(vha, QLA_EVT_SP_RETRY); if (!e) goto err2; e->u.iosb.sp = sp; qla2x00_post_work(vha, e); return; } err: e = qla2x00_alloc_work(vha, QLA_EVT_UNMAP); err2: if (!e) { /* please ignore kernel warning. otherwise, we have mem leak. */ if (sp->u.iocb_cmd.u.ctarg.req) { dma_free_coherent(&vha->hw->pdev->dev, sp->u.iocb_cmd.u.ctarg.req_allocated_size, sp->u.iocb_cmd.u.ctarg.req, sp->u.iocb_cmd.u.ctarg.req_dma); sp->u.iocb_cmd.u.ctarg.req = NULL; } if (sp->u.iocb_cmd.u.ctarg.rsp) { dma_free_coherent(&vha->hw->pdev->dev, sp->u.iocb_cmd.u.ctarg.rsp_allocated_size, sp->u.iocb_cmd.u.ctarg.rsp, sp->u.iocb_cmd.u.ctarg.rsp_dma); sp->u.iocb_cmd.u.ctarg.rsp = NULL; } /* ref: INIT */ kref_put(&sp->cmd_kref, qla2x00_sp_release); return; } e->u.iosb.sp = sp; qla2x00_post_work(vha, e); } /** * qla2x00_rft_id() - SNS Register FC-4 TYPEs (RFT_ID) supported by the HBA. * @vha: HA context * * Returns 0 on success. */ int qla2x00_rft_id(scsi_qla_host_t *vha) { struct qla_hw_data *ha = vha->hw; if (IS_QLA2100(ha) || IS_QLA2200(ha)) return qla2x00_sns_rft_id(vha); return qla_async_rftid(vha, &vha->d_id); } static int qla_async_rftid(scsi_qla_host_t *vha, port_id_t *d_id) { int rval = QLA_MEMORY_ALLOC_FAILED; struct ct_sns_req *ct_req; srb_t *sp; struct ct_sns_pkt *ct_sns; if (!vha->flags.online) goto done; /* ref: INIT */ sp = qla2x00_get_sp(vha, NULL, GFP_KERNEL); if (!sp) goto done; sp->type = SRB_CT_PTHRU_CMD; sp->name = "rft_id"; qla2x00_init_async_sp(sp, qla2x00_get_async_timeout(vha) + 2, qla2x00_async_sns_sp_done); sp->u.iocb_cmd.u.ctarg.req = dma_alloc_coherent(&vha->hw->pdev->dev, sizeof(struct ct_sns_pkt), &sp->u.iocb_cmd.u.ctarg.req_dma, GFP_KERNEL); sp->u.iocb_cmd.u.ctarg.req_allocated_size = sizeof(struct ct_sns_pkt); if (!sp->u.iocb_cmd.u.ctarg.req) { ql_log(ql_log_warn, vha, 0xd041, "%s: Failed to allocate ct_sns request.\n", __func__); goto done_free_sp; } sp->u.iocb_cmd.u.ctarg.rsp = dma_alloc_coherent(&vha->hw->pdev->dev, sizeof(struct ct_sns_pkt), &sp->u.iocb_cmd.u.ctarg.rsp_dma, GFP_KERNEL); sp->u.iocb_cmd.u.ctarg.rsp_allocated_size = sizeof(struct ct_sns_pkt); if (!sp->u.iocb_cmd.u.ctarg.rsp) { ql_log(ql_log_warn, vha, 0xd042, "%s: Failed to allocate ct_sns request.\n", __func__); goto done_free_sp; } ct_sns = (struct ct_sns_pkt *)sp->u.iocb_cmd.u.ctarg.rsp; memset(ct_sns, 0, sizeof(*ct_sns)); ct_sns = (struct ct_sns_pkt *)sp->u.iocb_cmd.u.ctarg.req; /* Prepare CT request */ ct_req = qla2x00_prep_ct_req(ct_sns, RFT_ID_CMD, RFT_ID_RSP_SIZE); /* Prepare CT arguments -- port_id, FC-4 types */ ct_req->req.rft_id.port_id = port_id_to_be_id(vha->d_id); ct_req->req.rft_id.fc4_types[2] = 0x01; /* FCP-3 */ if (vha->flags.nvme_enabled && qla_ini_mode_enabled(vha)) ct_req->req.rft_id.fc4_types[6] = 1; /* NVMe type 28h */ sp->u.iocb_cmd.u.ctarg.req_size = RFT_ID_REQ_SIZE; sp->u.iocb_cmd.u.ctarg.rsp_size = RFT_ID_RSP_SIZE; sp->u.iocb_cmd.u.ctarg.nport_handle = NPH_SNS; ql_dbg(ql_dbg_disc, vha, 0xffff, "Async-%s - hdl=%x portid %06x.\n", sp->name, sp->handle, d_id->b24); rval = qla2x00_start_sp(sp); if (rval != QLA_SUCCESS) { ql_dbg(ql_dbg_disc, vha, 0x2043, "RFT_ID issue IOCB failed (%d).\n", rval); goto done_free_sp; } return rval; done_free_sp: /* ref: INIT */ kref_put(&sp->cmd_kref, qla2x00_sp_release); done: return rval; } /** * qla2x00_rff_id() - SNS Register FC-4 Features (RFF_ID) supported by the HBA. * @vha: HA context * @type: not used * * Returns 0 on success. */ int qla2x00_rff_id(scsi_qla_host_t *vha, u8 type) { struct qla_hw_data *ha = vha->hw; if (IS_QLA2100(ha) || IS_QLA2200(ha)) { ql_dbg(ql_dbg_disc, vha, 0x2046, "RFF_ID call not supported on ISP2100/ISP2200.\n"); return (QLA_SUCCESS); } return qla_async_rffid(vha, &vha->d_id, qlt_rff_id(vha), type); } static int qla_async_rffid(scsi_qla_host_t *vha, port_id_t *d_id, u8 fc4feature, u8 fc4type) { int rval = QLA_MEMORY_ALLOC_FAILED; struct ct_sns_req *ct_req; srb_t *sp; struct ct_sns_pkt *ct_sns; /* ref: INIT */ sp = qla2x00_get_sp(vha, NULL, GFP_KERNEL); if (!sp) goto done; sp->type = SRB_CT_PTHRU_CMD; sp->name = "rff_id"; qla2x00_init_async_sp(sp, qla2x00_get_async_timeout(vha) + 2, qla2x00_async_sns_sp_done); sp->u.iocb_cmd.u.ctarg.req = dma_alloc_coherent(&vha->hw->pdev->dev, sizeof(struct ct_sns_pkt), &sp->u.iocb_cmd.u.ctarg.req_dma, GFP_KERNEL); sp->u.iocb_cmd.u.ctarg.req_allocated_size = sizeof(struct ct_sns_pkt); if (!sp->u.iocb_cmd.u.ctarg.req) { ql_log(ql_log_warn, vha, 0xd041, "%s: Failed to allocate ct_sns request.\n", __func__); goto done_free_sp; } sp->u.iocb_cmd.u.ctarg.rsp = dma_alloc_coherent(&vha->hw->pdev->dev, sizeof(struct ct_sns_pkt), &sp->u.iocb_cmd.u.ctarg.rsp_dma, GFP_KERNEL); sp->u.iocb_cmd.u.ctarg.rsp_allocated_size = sizeof(struct ct_sns_pkt); if (!sp->u.iocb_cmd.u.ctarg.rsp) { ql_log(ql_log_warn, vha, 0xd042, "%s: Failed to allocate ct_sns request.\n", __func__); goto done_free_sp; } ct_sns = (struct ct_sns_pkt *)sp->u.iocb_cmd.u.ctarg.rsp; memset(ct_sns, 0, sizeof(*ct_sns)); ct_sns = (struct ct_sns_pkt *)sp->u.iocb_cmd.u.ctarg.req; /* Prepare CT request */ ct_req = qla2x00_prep_ct_req(ct_sns, RFF_ID_CMD, RFF_ID_RSP_SIZE); /* Prepare CT arguments -- port_id, FC-4 feature, FC-4 type */ ct_req->req.rff_id.port_id = port_id_to_be_id(*d_id); ct_req->req.rff_id.fc4_feature = fc4feature; ct_req->req.rff_id.fc4_type = fc4type; /* SCSI-FCP or FC-NVMe */ sp->u.iocb_cmd.u.ctarg.req_size = RFF_ID_REQ_SIZE; sp->u.iocb_cmd.u.ctarg.rsp_size = RFF_ID_RSP_SIZE; sp->u.iocb_cmd.u.ctarg.nport_handle = NPH_SNS; ql_dbg(ql_dbg_disc, vha, 0xffff, "Async-%s - hdl=%x portid %06x feature %x type %x.\n", sp->name, sp->handle, d_id->b24, fc4feature, fc4type); rval = qla2x00_start_sp(sp); if (rval != QLA_SUCCESS) { ql_dbg(ql_dbg_disc, vha, 0x2047, "RFF_ID issue IOCB failed (%d).\n", rval); goto done_free_sp; } return rval; done_free_sp: /* ref: INIT */ kref_put(&sp->cmd_kref, qla2x00_sp_release); done: return rval; } /** * qla2x00_rnn_id() - SNS Register Node Name (RNN_ID) of the HBA. * @vha: HA context * * Returns 0 on success. */ int qla2x00_rnn_id(scsi_qla_host_t *vha) { struct qla_hw_data *ha = vha->hw; if (IS_QLA2100(ha) || IS_QLA2200(ha)) return qla2x00_sns_rnn_id(vha); return qla_async_rnnid(vha, &vha->d_id, vha->node_name); } static int qla_async_rnnid(scsi_qla_host_t *vha, port_id_t *d_id, u8 *node_name) { int rval = QLA_MEMORY_ALLOC_FAILED; struct ct_sns_req *ct_req; srb_t *sp; struct ct_sns_pkt *ct_sns; /* ref: INIT */ sp = qla2x00_get_sp(vha, NULL, GFP_KERNEL); if (!sp) goto done; sp->type = SRB_CT_PTHRU_CMD; sp->name = "rnid"; qla2x00_init_async_sp(sp, qla2x00_get_async_timeout(vha) + 2, qla2x00_async_sns_sp_done); sp->u.iocb_cmd.u.ctarg.req = dma_alloc_coherent(&vha->hw->pdev->dev, sizeof(struct ct_sns_pkt), &sp->u.iocb_cmd.u.ctarg.req_dma, GFP_KERNEL); sp->u.iocb_cmd.u.ctarg.req_allocated_size = sizeof(struct ct_sns_pkt); if (!sp->u.iocb_cmd.u.ctarg.req) { ql_log(ql_log_warn, vha, 0xd041, "%s: Failed to allocate ct_sns request.\n", __func__); goto done_free_sp; } sp->u.iocb_cmd.u.ctarg.rsp = dma_alloc_coherent(&vha->hw->pdev->dev, sizeof(struct ct_sns_pkt), &sp->u.iocb_cmd.u.ctarg.rsp_dma, GFP_KERNEL); sp->u.iocb_cmd.u.ctarg.rsp_allocated_size = sizeof(struct ct_sns_pkt); if (!sp->u.iocb_cmd.u.ctarg.rsp) { ql_log(ql_log_warn, vha, 0xd042, "%s: Failed to allocate ct_sns request.\n", __func__); goto done_free_sp; } ct_sns = (struct ct_sns_pkt *)sp->u.iocb_cmd.u.ctarg.rsp; memset(ct_sns, 0, sizeof(*ct_sns)); ct_sns = (struct ct_sns_pkt *)sp->u.iocb_cmd.u.ctarg.req; /* Prepare CT request */ ct_req = qla2x00_prep_ct_req(ct_sns, RNN_ID_CMD, RNN_ID_RSP_SIZE); /* Prepare CT arguments -- port_id, node_name */ ct_req->req.rnn_id.port_id = port_id_to_be_id(vha->d_id); memcpy(ct_req->req.rnn_id.node_name, vha->node_name, WWN_SIZE); sp->u.iocb_cmd.u.ctarg.req_size = RNN_ID_REQ_SIZE; sp->u.iocb_cmd.u.ctarg.rsp_size = RNN_ID_RSP_SIZE; sp->u.iocb_cmd.u.ctarg.nport_handle = NPH_SNS; ql_dbg(ql_dbg_disc, vha, 0xffff, "Async-%s - hdl=%x portid %06x\n", sp->name, sp->handle, d_id->b24); rval = qla2x00_start_sp(sp); if (rval != QLA_SUCCESS) { ql_dbg(ql_dbg_disc, vha, 0x204d, "RNN_ID issue IOCB failed (%d).\n", rval); goto done_free_sp; } return rval; done_free_sp: /* ref: INIT */ kref_put(&sp->cmd_kref, qla2x00_sp_release); done: return rval; } size_t qla2x00_get_sym_node_name(scsi_qla_host_t *vha, uint8_t *snn, size_t size) { struct qla_hw_data *ha = vha->hw; if (IS_QLAFX00(ha)) return scnprintf(snn, size, "%s FW:v%s DVR:v%s", ha->model_number, ha->mr.fw_version, qla2x00_version_str); return scnprintf(snn, size, "%s FW:v%d.%02d.%02d DVR:v%s", ha->model_number, ha->fw_major_version, ha->fw_minor_version, ha->fw_subminor_version, qla2x00_version_str); } /** * qla2x00_rsnn_nn() - SNS Register Symbolic Node Name (RSNN_NN) of the HBA. * @vha: HA context * * Returns 0 on success. */ int qla2x00_rsnn_nn(scsi_qla_host_t *vha) { struct qla_hw_data *ha = vha->hw; if (IS_QLA2100(ha) || IS_QLA2200(ha)) { ql_dbg(ql_dbg_disc, vha, 0x2050, "RSNN_ID call unsupported on ISP2100/ISP2200.\n"); return (QLA_SUCCESS); } return qla_async_rsnn_nn(vha); } static int qla_async_rsnn_nn(scsi_qla_host_t *vha) { int rval = QLA_MEMORY_ALLOC_FAILED; struct ct_sns_req *ct_req; srb_t *sp; struct ct_sns_pkt *ct_sns; /* ref: INIT */ sp = qla2x00_get_sp(vha, NULL, GFP_KERNEL); if (!sp) goto done; sp->type = SRB_CT_PTHRU_CMD; sp->name = "rsnn_nn"; qla2x00_init_async_sp(sp, qla2x00_get_async_timeout(vha) + 2, qla2x00_async_sns_sp_done); sp->u.iocb_cmd.u.ctarg.req = dma_alloc_coherent(&vha->hw->pdev->dev, sizeof(struct ct_sns_pkt), &sp->u.iocb_cmd.u.ctarg.req_dma, GFP_KERNEL); sp->u.iocb_cmd.u.ctarg.req_allocated_size = sizeof(struct ct_sns_pkt); if (!sp->u.iocb_cmd.u.ctarg.req) { ql_log(ql_log_warn, vha, 0xd041, "%s: Failed to allocate ct_sns request.\n", __func__); goto done_free_sp; } sp->u.iocb_cmd.u.ctarg.rsp = dma_alloc_coherent(&vha->hw->pdev->dev, sizeof(struct ct_sns_pkt), &sp->u.iocb_cmd.u.ctarg.rsp_dma, GFP_KERNEL); sp->u.iocb_cmd.u.ctarg.rsp_allocated_size = sizeof(struct ct_sns_pkt); if (!sp->u.iocb_cmd.u.ctarg.rsp) { ql_log(ql_log_warn, vha, 0xd042, "%s: Failed to allocate ct_sns request.\n", __func__); goto done_free_sp; } ct_sns = (struct ct_sns_pkt *)sp->u.iocb_cmd.u.ctarg.rsp; memset(ct_sns, 0, sizeof(*ct_sns)); ct_sns = (struct ct_sns_pkt *)sp->u.iocb_cmd.u.ctarg.req; /* Prepare CT request */ ct_req = qla2x00_prep_ct_req(ct_sns, RSNN_NN_CMD, RSNN_NN_RSP_SIZE); /* Prepare CT arguments -- node_name, symbolic node_name, size */ memcpy(ct_req->req.rsnn_nn.node_name, vha->node_name, WWN_SIZE); /* Prepare the Symbolic Node Name */ qla2x00_get_sym_node_name(vha, ct_req->req.rsnn_nn.sym_node_name, sizeof(ct_req->req.rsnn_nn.sym_node_name)); ct_req->req.rsnn_nn.name_len = (uint8_t)strlen(ct_req->req.rsnn_nn.sym_node_name); sp->u.iocb_cmd.u.ctarg.req_size = 24 + 1 + ct_req->req.rsnn_nn.name_len; sp->u.iocb_cmd.u.ctarg.rsp_size = RSNN_NN_RSP_SIZE; sp->u.iocb_cmd.u.ctarg.nport_handle = NPH_SNS; ql_dbg(ql_dbg_disc, vha, 0xffff, "Async-%s - hdl=%x.\n", sp->name, sp->handle); rval = qla2x00_start_sp(sp); if (rval != QLA_SUCCESS) { ql_dbg(ql_dbg_disc, vha, 0x2043, "RFT_ID issue IOCB failed (%d).\n", rval); goto done_free_sp; } return rval; done_free_sp: /* ref: INIT */ kref_put(&sp->cmd_kref, qla2x00_sp_release); done: return rval; } /** * qla2x00_prep_sns_cmd() - Prepare common SNS command request fields for query. * @vha: HA context * @cmd: GS command * @scmd_len: Subcommand length * @data_size: response size in bytes * * Returns a pointer to the @ha's sns_cmd. */ static inline struct sns_cmd_pkt * qla2x00_prep_sns_cmd(scsi_qla_host_t *vha, uint16_t cmd, uint16_t scmd_len, uint16_t data_size) { uint16_t wc; struct sns_cmd_pkt *sns_cmd; struct qla_hw_data *ha = vha->hw; sns_cmd = ha->sns_cmd; memset(sns_cmd, 0, sizeof(struct sns_cmd_pkt)); wc = data_size / 2; /* Size in 16bit words. */ sns_cmd->p.cmd.buffer_length = cpu_to_le16(wc); put_unaligned_le64(ha->sns_cmd_dma, &sns_cmd->p.cmd.buffer_address); sns_cmd->p.cmd.subcommand_length = cpu_to_le16(scmd_len); sns_cmd->p.cmd.subcommand = cpu_to_le16(cmd); wc = (data_size - 16) / 4; /* Size in 32bit words. */ sns_cmd->p.cmd.size = cpu_to_le16(wc); vha->qla_stats.control_requests++; return (sns_cmd); } /** * qla2x00_sns_ga_nxt() - SNS scan for fabric devices via GA_NXT command. * @vha: HA context * @fcport: fcport entry to updated * * This command uses the old Exectute SNS Command mailbox routine. * * Returns 0 on success. */ static int qla2x00_sns_ga_nxt(scsi_qla_host_t *vha, fc_port_t *fcport) { int rval = QLA_SUCCESS; struct qla_hw_data *ha = vha->hw; struct sns_cmd_pkt *sns_cmd; /* Issue GA_NXT. */ /* Prepare SNS command request. */ sns_cmd = qla2x00_prep_sns_cmd(vha, GA_NXT_CMD, GA_NXT_SNS_SCMD_LEN, GA_NXT_SNS_DATA_SIZE); /* Prepare SNS command arguments -- port_id. */ sns_cmd->p.cmd.param[0] = fcport->d_id.b.al_pa; sns_cmd->p.cmd.param[1] = fcport->d_id.b.area; sns_cmd->p.cmd.param[2] = fcport->d_id.b.domain; /* Execute SNS command. */ rval = qla2x00_send_sns(vha, ha->sns_cmd_dma, GA_NXT_SNS_CMD_SIZE / 2, sizeof(struct sns_cmd_pkt)); if (rval != QLA_SUCCESS) { /*EMPTY*/ ql_dbg(ql_dbg_disc, vha, 0x205f, "GA_NXT Send SNS failed (%d).\n", rval); } else if (sns_cmd->p.gan_data[8] != 0x80 || sns_cmd->p.gan_data[9] != 0x02) { ql_dbg(ql_dbg_disc + ql_dbg_buffer, vha, 0x2084, "GA_NXT failed, rejected request ga_nxt_rsp:\n"); ql_dump_buffer(ql_dbg_disc + ql_dbg_buffer, vha, 0x2074, sns_cmd->p.gan_data, 16); rval = QLA_FUNCTION_FAILED; } else { /* Populate fc_port_t entry. */ fcport->d_id.b.domain = sns_cmd->p.gan_data[17]; fcport->d_id.b.area = sns_cmd->p.gan_data[18]; fcport->d_id.b.al_pa = sns_cmd->p.gan_data[19]; memcpy(fcport->node_name, &sns_cmd->p.gan_data[284], WWN_SIZE); memcpy(fcport->port_name, &sns_cmd->p.gan_data[20], WWN_SIZE); if (sns_cmd->p.gan_data[16] != NS_N_PORT_TYPE && sns_cmd->p.gan_data[16] != NS_NL_PORT_TYPE) fcport->d_id.b.domain = 0xf0; ql_dbg(ql_dbg_disc, vha, 0x2061, "GA_NXT entry - nn %8phN pn %8phN " "port_id=%02x%02x%02x.\n", fcport->node_name, fcport->port_name, fcport->d_id.b.domain, fcport->d_id.b.area, fcport->d_id.b.al_pa); } return (rval); } /** * qla2x00_sns_gid_pt() - SNS scan for fabric devices via GID_PT command. * @vha: HA context * @list: switch info entries to populate * * This command uses the old Exectute SNS Command mailbox routine. * * NOTE: Non-Nx_Ports are not requested. * * Returns 0 on success. */ static int qla2x00_sns_gid_pt(scsi_qla_host_t *vha, sw_info_t *list) { int rval; struct qla_hw_data *ha = vha->hw; uint16_t i; uint8_t *entry; struct sns_cmd_pkt *sns_cmd; uint16_t gid_pt_sns_data_size; gid_pt_sns_data_size = qla2x00_gid_pt_rsp_size(vha); /* Issue GID_PT. */ /* Prepare SNS command request. */ sns_cmd = qla2x00_prep_sns_cmd(vha, GID_PT_CMD, GID_PT_SNS_SCMD_LEN, gid_pt_sns_data_size); /* Prepare SNS command arguments -- port_type. */ sns_cmd->p.cmd.param[0] = NS_NX_PORT_TYPE; /* Execute SNS command. */ rval = qla2x00_send_sns(vha, ha->sns_cmd_dma, GID_PT_SNS_CMD_SIZE / 2, sizeof(struct sns_cmd_pkt)); if (rval != QLA_SUCCESS) { /*EMPTY*/ ql_dbg(ql_dbg_disc, vha, 0x206d, "GID_PT Send SNS failed (%d).\n", rval); } else if (sns_cmd->p.gid_data[8] != 0x80 || sns_cmd->p.gid_data[9] != 0x02) { ql_dbg(ql_dbg_disc, vha, 0x202f, "GID_PT failed, rejected request, gid_rsp:\n"); ql_dump_buffer(ql_dbg_disc + ql_dbg_buffer, vha, 0x2081, sns_cmd->p.gid_data, 16); rval = QLA_FUNCTION_FAILED; } else { /* Set port IDs in switch info list. */ for (i = 0; i < ha->max_fibre_devices; i++) { entry = &sns_cmd->p.gid_data[(i * 4) + 16]; list[i].d_id.b.domain = entry[1]; list[i].d_id.b.area = entry[2]; list[i].d_id.b.al_pa = entry[3]; /* Last one exit. */ if (entry[0] & BIT_7) { list[i].d_id.b.rsvd_1 = entry[0]; break; } } /* * If we've used all available slots, then the switch is * reporting back more devices that we can handle with this * single call. Return a failed status, and let GA_NXT handle * the overload. */ if (i == ha->max_fibre_devices) rval = QLA_FUNCTION_FAILED; } return (rval); } /** * qla2x00_sns_gpn_id() - SNS Get Port Name (GPN_ID) query. * @vha: HA context * @list: switch info entries to populate * * This command uses the old Exectute SNS Command mailbox routine. * * Returns 0 on success. */ static int qla2x00_sns_gpn_id(scsi_qla_host_t *vha, sw_info_t *list) { int rval = QLA_SUCCESS; struct qla_hw_data *ha = vha->hw; uint16_t i; struct sns_cmd_pkt *sns_cmd; for (i = 0; i < ha->max_fibre_devices; i++) { /* Issue GPN_ID */ /* Prepare SNS command request. */ sns_cmd = qla2x00_prep_sns_cmd(vha, GPN_ID_CMD, GPN_ID_SNS_SCMD_LEN, GPN_ID_SNS_DATA_SIZE); /* Prepare SNS command arguments -- port_id. */ sns_cmd->p.cmd.param[0] = list[i].d_id.b.al_pa; sns_cmd->p.cmd.param[1] = list[i].d_id.b.area; sns_cmd->p.cmd.param[2] = list[i].d_id.b.domain; /* Execute SNS command. */ rval = qla2x00_send_sns(vha, ha->sns_cmd_dma, GPN_ID_SNS_CMD_SIZE / 2, sizeof(struct sns_cmd_pkt)); if (rval != QLA_SUCCESS) { /*EMPTY*/ ql_dbg(ql_dbg_disc, vha, 0x2032, "GPN_ID Send SNS failed (%d).\n", rval); } else if (sns_cmd->p.gpn_data[8] != 0x80 || sns_cmd->p.gpn_data[9] != 0x02) { ql_dbg(ql_dbg_disc + ql_dbg_buffer, vha, 0x207e, "GPN_ID failed, rejected request, gpn_rsp:\n"); ql_dump_buffer(ql_dbg_disc + ql_dbg_buffer, vha, 0x207f, sns_cmd->p.gpn_data, 16); rval = QLA_FUNCTION_FAILED; } else { /* Save portname */ memcpy(list[i].port_name, &sns_cmd->p.gpn_data[16], WWN_SIZE); } /* Last device exit. */ if (list[i].d_id.b.rsvd_1 != 0) break; } return (rval); } /** * qla2x00_sns_gnn_id() - SNS Get Node Name (GNN_ID) query. * @vha: HA context * @list: switch info entries to populate * * This command uses the old Exectute SNS Command mailbox routine. * * Returns 0 on success. */ static int qla2x00_sns_gnn_id(scsi_qla_host_t *vha, sw_info_t *list) { int rval = QLA_SUCCESS; struct qla_hw_data *ha = vha->hw; uint16_t i; struct sns_cmd_pkt *sns_cmd; for (i = 0; i < ha->max_fibre_devices; i++) { /* Issue GNN_ID */ /* Prepare SNS command request. */ sns_cmd = qla2x00_prep_sns_cmd(vha, GNN_ID_CMD, GNN_ID_SNS_SCMD_LEN, GNN_ID_SNS_DATA_SIZE); /* Prepare SNS command arguments -- port_id. */ sns_cmd->p.cmd.param[0] = list[i].d_id.b.al_pa; sns_cmd->p.cmd.param[1] = list[i].d_id.b.area; sns_cmd->p.cmd.param[2] = list[i].d_id.b.domain; /* Execute SNS command. */ rval = qla2x00_send_sns(vha, ha->sns_cmd_dma, GNN_ID_SNS_CMD_SIZE / 2, sizeof(struct sns_cmd_pkt)); if (rval != QLA_SUCCESS) { /*EMPTY*/ ql_dbg(ql_dbg_disc, vha, 0x203f, "GNN_ID Send SNS failed (%d).\n", rval); } else if (sns_cmd->p.gnn_data[8] != 0x80 || sns_cmd->p.gnn_data[9] != 0x02) { ql_dbg(ql_dbg_disc + ql_dbg_buffer, vha, 0x2082, "GNN_ID failed, rejected request, gnn_rsp:\n"); ql_dump_buffer(ql_dbg_disc + ql_dbg_buffer, vha, 0x207a, sns_cmd->p.gnn_data, 16); rval = QLA_FUNCTION_FAILED; } else { /* Save nodename */ memcpy(list[i].node_name, &sns_cmd->p.gnn_data[16], WWN_SIZE); ql_dbg(ql_dbg_disc, vha, 0x206e, "GID_PT entry - nn %8phN pn %8phN " "port_id=%02x%02x%02x.\n", list[i].node_name, list[i].port_name, list[i].d_id.b.domain, list[i].d_id.b.area, list[i].d_id.b.al_pa); } /* Last device exit. */ if (list[i].d_id.b.rsvd_1 != 0) break; } return (rval); } /** * qla2x00_sns_rft_id() - SNS Register FC-4 TYPEs (RFT_ID) supported by the HBA. * @vha: HA context * * This command uses the old Exectute SNS Command mailbox routine. * * Returns 0 on success. */ static int qla2x00_sns_rft_id(scsi_qla_host_t *vha) { int rval; struct qla_hw_data *ha = vha->hw; struct sns_cmd_pkt *sns_cmd; /* Issue RFT_ID. */ /* Prepare SNS command request. */ sns_cmd = qla2x00_prep_sns_cmd(vha, RFT_ID_CMD, RFT_ID_SNS_SCMD_LEN, RFT_ID_SNS_DATA_SIZE); /* Prepare SNS command arguments -- port_id, FC-4 types */ sns_cmd->p.cmd.param[0] = vha->d_id.b.al_pa; sns_cmd->p.cmd.param[1] = vha->d_id.b.area; sns_cmd->p.cmd.param[2] = vha->d_id.b.domain; sns_cmd->p.cmd.param[5] = 0x01; /* FCP-3 */ /* Execute SNS command. */ rval = qla2x00_send_sns(vha, ha->sns_cmd_dma, RFT_ID_SNS_CMD_SIZE / 2, sizeof(struct sns_cmd_pkt)); if (rval != QLA_SUCCESS) { /*EMPTY*/ ql_dbg(ql_dbg_disc, vha, 0x2060, "RFT_ID Send SNS failed (%d).\n", rval); } else if (sns_cmd->p.rft_data[8] != 0x80 || sns_cmd->p.rft_data[9] != 0x02) { ql_dbg(ql_dbg_disc + ql_dbg_buffer, vha, 0x2083, "RFT_ID failed, rejected request rft_rsp:\n"); ql_dump_buffer(ql_dbg_disc + ql_dbg_buffer, vha, 0x2080, sns_cmd->p.rft_data, 16); rval = QLA_FUNCTION_FAILED; } else { ql_dbg(ql_dbg_disc, vha, 0x2073, "RFT_ID exiting normally.\n"); } return (rval); } /** * qla2x00_sns_rnn_id() - SNS Register Node Name (RNN_ID) of the HBA. * @vha: HA context * * This command uses the old Exectute SNS Command mailbox routine. * * Returns 0 on success. */ static int qla2x00_sns_rnn_id(scsi_qla_host_t *vha) { int rval; struct qla_hw_data *ha = vha->hw; struct sns_cmd_pkt *sns_cmd; /* Issue RNN_ID. */ /* Prepare SNS command request. */ sns_cmd = qla2x00_prep_sns_cmd(vha, RNN_ID_CMD, RNN_ID_SNS_SCMD_LEN, RNN_ID_SNS_DATA_SIZE); /* Prepare SNS command arguments -- port_id, nodename. */ sns_cmd->p.cmd.param[0] = vha->d_id.b.al_pa; sns_cmd->p.cmd.param[1] = vha->d_id.b.area; sns_cmd->p.cmd.param[2] = vha->d_id.b.domain; sns_cmd->p.cmd.param[4] = vha->node_name[7]; sns_cmd->p.cmd.param[5] = vha->node_name[6]; sns_cmd->p.cmd.param[6] = vha->node_name[5]; sns_cmd->p.cmd.param[7] = vha->node_name[4]; sns_cmd->p.cmd.param[8] = vha->node_name[3]; sns_cmd->p.cmd.param[9] = vha->node_name[2]; sns_cmd->p.cmd.param[10] = vha->node_name[1]; sns_cmd->p.cmd.param[11] = vha->node_name[0]; /* Execute SNS command. */ rval = qla2x00_send_sns(vha, ha->sns_cmd_dma, RNN_ID_SNS_CMD_SIZE / 2, sizeof(struct sns_cmd_pkt)); if (rval != QLA_SUCCESS) { /*EMPTY*/ ql_dbg(ql_dbg_disc, vha, 0x204a, "RNN_ID Send SNS failed (%d).\n", rval); } else if (sns_cmd->p.rnn_data[8] != 0x80 || sns_cmd->p.rnn_data[9] != 0x02) { ql_dbg(ql_dbg_disc + ql_dbg_buffer, vha, 0x207b, "RNN_ID failed, rejected request, rnn_rsp:\n"); ql_dump_buffer(ql_dbg_disc + ql_dbg_buffer, vha, 0x207c, sns_cmd->p.rnn_data, 16); rval = QLA_FUNCTION_FAILED; } else { ql_dbg(ql_dbg_disc, vha, 0x204c, "RNN_ID exiting normally.\n"); } return (rval); } /** * qla2x00_mgmt_svr_login() - Login to fabric Management Service. * @vha: HA context * * Returns 0 on success. */ int qla2x00_mgmt_svr_login(scsi_qla_host_t *vha) { int ret, rval; uint16_t mb[MAILBOX_REGISTER_COUNT]; struct qla_hw_data *ha = vha->hw; ret = QLA_SUCCESS; if (vha->flags.management_server_logged_in) return ret; rval = ha->isp_ops->fabric_login(vha, vha->mgmt_svr_loop_id, 0xff, 0xff, 0xfa, mb, BIT_1); if (rval != QLA_SUCCESS || mb[0] != MBS_COMMAND_COMPLETE) { if (rval == QLA_MEMORY_ALLOC_FAILED) ql_dbg(ql_dbg_disc, vha, 0x2085, "Failed management_server login: loopid=%x " "rval=%d\n", vha->mgmt_svr_loop_id, rval); else ql_dbg(ql_dbg_disc, vha, 0x2024, "Failed management_server login: loopid=%x " "mb[0]=%x mb[1]=%x mb[2]=%x mb[6]=%x mb[7]=%x.\n", vha->mgmt_svr_loop_id, mb[0], mb[1], mb[2], mb[6], mb[7]); ret = QLA_FUNCTION_FAILED; } else vha->flags.management_server_logged_in = 1; return ret; } /** * qla2x00_prep_ms_fdmi_iocb() - Prepare common MS IOCB fields for FDMI query. * @vha: HA context * @req_size: request size in bytes * @rsp_size: response size in bytes * * Returns a pointer to the @ha's ms_iocb. */ void * qla2x00_prep_ms_fdmi_iocb(scsi_qla_host_t *vha, uint32_t req_size, uint32_t rsp_size) { ms_iocb_entry_t *ms_pkt; struct qla_hw_data *ha = vha->hw; ms_pkt = ha->ms_iocb; memset(ms_pkt, 0, sizeof(ms_iocb_entry_t)); ms_pkt->entry_type = MS_IOCB_TYPE; ms_pkt->entry_count = 1; SET_TARGET_ID(ha, ms_pkt->loop_id, vha->mgmt_svr_loop_id); ms_pkt->control_flags = cpu_to_le16(CF_READ | CF_HEAD_TAG); ms_pkt->timeout = cpu_to_le16(ha->r_a_tov / 10 * 2); ms_pkt->cmd_dsd_count = cpu_to_le16(1); ms_pkt->total_dsd_count = cpu_to_le16(2); ms_pkt->rsp_bytecount = cpu_to_le32(rsp_size); ms_pkt->req_bytecount = cpu_to_le32(req_size); put_unaligned_le64(ha->ct_sns_dma, &ms_pkt->req_dsd.address); ms_pkt->req_dsd.length = ms_pkt->req_bytecount; put_unaligned_le64(ha->ct_sns_dma, &ms_pkt->rsp_dsd.address); ms_pkt->rsp_dsd.length = ms_pkt->rsp_bytecount; return ms_pkt; } /** * qla24xx_prep_ms_fdmi_iocb() - Prepare common MS IOCB fields for FDMI query. * @vha: HA context * @req_size: request size in bytes * @rsp_size: response size in bytes * * Returns a pointer to the @ha's ms_iocb. */ void * qla24xx_prep_ms_fdmi_iocb(scsi_qla_host_t *vha, uint32_t req_size, uint32_t rsp_size) { struct ct_entry_24xx *ct_pkt; struct qla_hw_data *ha = vha->hw; ct_pkt = (struct ct_entry_24xx *)ha->ms_iocb; memset(ct_pkt, 0, sizeof(struct ct_entry_24xx)); ct_pkt->entry_type = CT_IOCB_TYPE; ct_pkt->entry_count = 1; ct_pkt->nport_handle = cpu_to_le16(vha->mgmt_svr_loop_id); ct_pkt->timeout = cpu_to_le16(ha->r_a_tov / 10 * 2); ct_pkt->cmd_dsd_count = cpu_to_le16(1); ct_pkt->rsp_dsd_count = cpu_to_le16(1); ct_pkt->rsp_byte_count = cpu_to_le32(rsp_size); ct_pkt->cmd_byte_count = cpu_to_le32(req_size); put_unaligned_le64(ha->ct_sns_dma, &ct_pkt->dsd[0].address); ct_pkt->dsd[0].length = ct_pkt->cmd_byte_count; put_unaligned_le64(ha->ct_sns_dma, &ct_pkt->dsd[1].address); ct_pkt->dsd[1].length = ct_pkt->rsp_byte_count; ct_pkt->vp_index = vha->vp_idx; return ct_pkt; } static void qla2x00_update_ms_fdmi_iocb(scsi_qla_host_t *vha, uint32_t req_size) { struct qla_hw_data *ha = vha->hw; ms_iocb_entry_t *ms_pkt = ha->ms_iocb; struct ct_entry_24xx *ct_pkt = (struct ct_entry_24xx *)ha->ms_iocb; if (IS_FWI2_CAPABLE(ha)) { ct_pkt->cmd_byte_count = cpu_to_le32(req_size); ct_pkt->dsd[0].length = ct_pkt->cmd_byte_count; } else { ms_pkt->req_bytecount = cpu_to_le32(req_size); ms_pkt->req_dsd.length = ms_pkt->req_bytecount; } } /** * qla2x00_prep_ct_fdmi_req() - Prepare common CT request fields for SNS query. * @p: CT request buffer * @cmd: GS command * @rsp_size: response size in bytes * * Returns a pointer to the intitialized @ct_req. */ static inline struct ct_sns_req * qla2x00_prep_ct_fdmi_req(struct ct_sns_pkt *p, uint16_t cmd, uint16_t rsp_size) { memset(p, 0, sizeof(struct ct_sns_pkt)); p->p.req.header.revision = 0x01; p->p.req.header.gs_type = 0xFA; p->p.req.header.gs_subtype = 0x10; p->p.req.command = cpu_to_be16(cmd); p->p.req.max_rsp_size = cpu_to_be16((rsp_size - 16) / 4); return &p->p.req; } uint qla25xx_fdmi_port_speed_capability(struct qla_hw_data *ha) { uint speeds = 0; if (IS_CNA_CAPABLE(ha)) return FDMI_PORT_SPEED_10GB; if (IS_QLA28XX(ha) || IS_QLA27XX(ha)) { if (ha->max_supported_speed == 2) { if (ha->min_supported_speed <= 6) speeds |= FDMI_PORT_SPEED_64GB; } if (ha->max_supported_speed == 2 || ha->max_supported_speed == 1) { if (ha->min_supported_speed <= 5) speeds |= FDMI_PORT_SPEED_32GB; } if (ha->max_supported_speed == 2 || ha->max_supported_speed == 1 || ha->max_supported_speed == 0) { if (ha->min_supported_speed <= 4) speeds |= FDMI_PORT_SPEED_16GB; } if (ha->max_supported_speed == 1 || ha->max_supported_speed == 0) { if (ha->min_supported_speed <= 3) speeds |= FDMI_PORT_SPEED_8GB; } if (ha->max_supported_speed == 0) { if (ha->min_supported_speed <= 2) speeds |= FDMI_PORT_SPEED_4GB; } return speeds; } if (IS_QLA2031(ha)) { if ((ha->pdev->subsystem_vendor == 0x103C) && ((ha->pdev->subsystem_device == 0x8002) || (ha->pdev->subsystem_device == 0x8086))) { speeds = FDMI_PORT_SPEED_16GB; } else { speeds = FDMI_PORT_SPEED_16GB|FDMI_PORT_SPEED_8GB| FDMI_PORT_SPEED_4GB; } return speeds; } if (IS_QLA25XX(ha) || IS_QLAFX00(ha)) return FDMI_PORT_SPEED_8GB|FDMI_PORT_SPEED_4GB| FDMI_PORT_SPEED_2GB|FDMI_PORT_SPEED_1GB; if (IS_QLA24XX_TYPE(ha)) return FDMI_PORT_SPEED_4GB|FDMI_PORT_SPEED_2GB| FDMI_PORT_SPEED_1GB; if (IS_QLA23XX(ha)) return FDMI_PORT_SPEED_2GB|FDMI_PORT_SPEED_1GB; return FDMI_PORT_SPEED_1GB; } uint qla25xx_fdmi_port_speed_currently(struct qla_hw_data *ha) { switch (ha->link_data_rate) { case PORT_SPEED_1GB: return FDMI_PORT_SPEED_1GB; case PORT_SPEED_2GB: return FDMI_PORT_SPEED_2GB; case PORT_SPEED_4GB: return FDMI_PORT_SPEED_4GB; case PORT_SPEED_8GB: return FDMI_PORT_SPEED_8GB; case PORT_SPEED_10GB: return FDMI_PORT_SPEED_10GB; case PORT_SPEED_16GB: return FDMI_PORT_SPEED_16GB; case PORT_SPEED_32GB: return FDMI_PORT_SPEED_32GB; case PORT_SPEED_64GB: return FDMI_PORT_SPEED_64GB; default: return FDMI_PORT_SPEED_UNKNOWN; } } /** * qla2x00_hba_attributes() - perform HBA attributes registration * @vha: HA context * @entries: number of entries to use * @callopt: Option to issue extended or standard FDMI * command parameter * * Returns 0 on success. */ static unsigned long qla2x00_hba_attributes(scsi_qla_host_t *vha, void *entries, unsigned int callopt) { struct qla_hw_data *ha = vha->hw; struct new_utsname *p_sysid = utsname(); struct ct_fdmi_hba_attr *eiter; uint16_t alen; unsigned long size = 0; /* Nodename. */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_HBA_NODE_NAME); memcpy(eiter->a.node_name, vha->node_name, sizeof(eiter->a.node_name)); alen = sizeof(eiter->a.node_name); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20a0, "NODENAME = %016llx.\n", wwn_to_u64(eiter->a.node_name)); /* Manufacturer. */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_HBA_MANUFACTURER); alen = scnprintf( eiter->a.manufacturer, sizeof(eiter->a.manufacturer), "%s", QLA2XXX_MANUFACTURER); alen += FDMI_ATTR_ALIGNMENT(alen); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20a1, "MANUFACTURER = %s.\n", eiter->a.manufacturer); /* Serial number. */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_HBA_SERIAL_NUMBER); alen = 0; if (IS_FWI2_CAPABLE(ha)) { alen = qla2xxx_get_vpd_field(vha, "SN", eiter->a.serial_num, sizeof(eiter->a.serial_num)); } if (!alen) { uint32_t sn = ((ha->serial0 & 0x1f) << 16) | (ha->serial2 << 8) | ha->serial1; alen = scnprintf( eiter->a.serial_num, sizeof(eiter->a.serial_num), "%c%05d", 'A' + sn / 100000, sn % 100000); } alen += FDMI_ATTR_ALIGNMENT(alen); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20a2, "SERIAL NUMBER = %s.\n", eiter->a.serial_num); /* Model name. */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_HBA_MODEL); alen = scnprintf( eiter->a.model, sizeof(eiter->a.model), "%s", ha->model_number); alen += FDMI_ATTR_ALIGNMENT(alen); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20a3, "MODEL NAME = %s.\n", eiter->a.model); /* Model description. */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_HBA_MODEL_DESCRIPTION); alen = scnprintf( eiter->a.model_desc, sizeof(eiter->a.model_desc), "%s", ha->model_desc); alen += FDMI_ATTR_ALIGNMENT(alen); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20a4, "MODEL DESCRIPTION = %s.\n", eiter->a.model_desc); /* Hardware version. */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_HBA_HARDWARE_VERSION); alen = 0; if (IS_FWI2_CAPABLE(ha)) { if (!alen) { alen = qla2xxx_get_vpd_field(vha, "MN", eiter->a.hw_version, sizeof(eiter->a.hw_version)); } if (!alen) { alen = qla2xxx_get_vpd_field(vha, "EC", eiter->a.hw_version, sizeof(eiter->a.hw_version)); } } if (!alen) { alen = scnprintf( eiter->a.hw_version, sizeof(eiter->a.hw_version), "HW:%s", ha->adapter_id); } alen += FDMI_ATTR_ALIGNMENT(alen); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20a5, "HARDWARE VERSION = %s.\n", eiter->a.hw_version); /* Driver version. */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_HBA_DRIVER_VERSION); alen = scnprintf( eiter->a.driver_version, sizeof(eiter->a.driver_version), "%s", qla2x00_version_str); alen += FDMI_ATTR_ALIGNMENT(alen); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20a6, "DRIVER VERSION = %s.\n", eiter->a.driver_version); /* Option ROM version. */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_HBA_OPTION_ROM_VERSION); alen = scnprintf( eiter->a.orom_version, sizeof(eiter->a.orom_version), "%d.%02d", ha->bios_revision[1], ha->bios_revision[0]); alen += FDMI_ATTR_ALIGNMENT(alen); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20a7, "OPTROM VERSION = %d.%02d.\n", eiter->a.orom_version[1], eiter->a.orom_version[0]); /* Firmware version */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_HBA_FIRMWARE_VERSION); ha->isp_ops->fw_version_str(vha, eiter->a.fw_version, sizeof(eiter->a.fw_version)); alen += FDMI_ATTR_ALIGNMENT(alen); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20a8, "FIRMWARE VERSION = %s.\n", eiter->a.fw_version); /* OS Name and Version */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_HBA_OS_NAME_AND_VERSION); alen = 0; if (p_sysid) { alen = scnprintf( eiter->a.os_version, sizeof(eiter->a.os_version), "%s %s %s", p_sysid->sysname, p_sysid->release, p_sysid->machine); } if (!alen) { alen = scnprintf( eiter->a.os_version, sizeof(eiter->a.os_version), "%s %s", "Linux", fc_host_system_hostname(vha->host)); } alen += FDMI_ATTR_ALIGNMENT(alen); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20a9, "OS VERSION = %s.\n", eiter->a.os_version); if (callopt == CALLOPT_FDMI1) goto done; /* MAX CT Payload Length */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_HBA_MAXIMUM_CT_PAYLOAD_LENGTH); eiter->a.max_ct_len = cpu_to_be32(ha->frame_payload_size >> 2); alen = sizeof(eiter->a.max_ct_len); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20aa, "CT PAYLOAD LENGTH = 0x%x.\n", be32_to_cpu(eiter->a.max_ct_len)); /* Node Symbolic Name */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_HBA_NODE_SYMBOLIC_NAME); alen = qla2x00_get_sym_node_name(vha, eiter->a.sym_name, sizeof(eiter->a.sym_name)); alen += FDMI_ATTR_ALIGNMENT(alen); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20ab, "SYMBOLIC NAME = %s.\n", eiter->a.sym_name); /* Vendor Specific information */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_HBA_VENDOR_SPECIFIC_INFO); eiter->a.vendor_specific_info = cpu_to_be32(PCI_VENDOR_ID_QLOGIC); alen = sizeof(eiter->a.vendor_specific_info); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20ac, "VENDOR SPECIFIC INFO = 0x%x.\n", be32_to_cpu(eiter->a.vendor_specific_info)); /* Num Ports */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_HBA_NUM_PORTS); eiter->a.num_ports = cpu_to_be32(1); alen = sizeof(eiter->a.num_ports); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20ad, "PORT COUNT = %x.\n", be32_to_cpu(eiter->a.num_ports)); /* Fabric Name */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_HBA_FABRIC_NAME); memcpy(eiter->a.fabric_name, vha->fabric_node_name, sizeof(eiter->a.fabric_name)); alen = sizeof(eiter->a.fabric_name); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20ae, "FABRIC NAME = %016llx.\n", wwn_to_u64(eiter->a.fabric_name)); /* BIOS Version */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_HBA_BOOT_BIOS_NAME); alen = scnprintf( eiter->a.bios_name, sizeof(eiter->a.bios_name), "BIOS %d.%02d", ha->bios_revision[1], ha->bios_revision[0]); alen += FDMI_ATTR_ALIGNMENT(alen); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20af, "BIOS NAME = %s\n", eiter->a.bios_name); /* Vendor Identifier */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_HBA_VENDOR_IDENTIFIER); alen = scnprintf( eiter->a.vendor_identifier, sizeof(eiter->a.vendor_identifier), "%s", "QLGC"); alen += FDMI_ATTR_ALIGNMENT(alen); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20b0, "VENDOR IDENTIFIER = %s.\n", eiter->a.vendor_identifier); done: return size; } /** * qla2x00_port_attributes() - perform Port attributes registration * @vha: HA context * @entries: number of entries to use * @callopt: Option to issue extended or standard FDMI * command parameter * * Returns 0 on success. */ static unsigned long qla2x00_port_attributes(scsi_qla_host_t *vha, void *entries, unsigned int callopt) { struct qla_hw_data *ha = vha->hw; struct new_utsname *p_sysid = utsname(); char *hostname = p_sysid ? p_sysid->nodename : fc_host_system_hostname(vha->host); struct ct_fdmi_port_attr *eiter; uint16_t alen; unsigned long size = 0; /* FC4 types. */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_PORT_FC4_TYPES); eiter->a.fc4_types[0] = 0x00; eiter->a.fc4_types[1] = 0x00; eiter->a.fc4_types[2] = 0x01; eiter->a.fc4_types[3] = 0x00; alen = sizeof(eiter->a.fc4_types); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20c0, "FC4 TYPES = %016llx.\n", *(uint64_t *)eiter->a.fc4_types); if (vha->flags.nvme_enabled) { eiter->a.fc4_types[6] = 1; /* NVMe type 28h */ ql_dbg(ql_dbg_disc, vha, 0x211f, "NVME FC4 Type = %02x 0x0 0x0 0x0 0x0 0x0.\n", eiter->a.fc4_types[6]); } /* Supported speed. */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_PORT_SUPPORT_SPEED); eiter->a.sup_speed = cpu_to_be32( qla25xx_fdmi_port_speed_capability(ha)); alen = sizeof(eiter->a.sup_speed); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20c1, "SUPPORTED SPEED = %x.\n", be32_to_cpu(eiter->a.sup_speed)); /* Current speed. */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_PORT_CURRENT_SPEED); eiter->a.cur_speed = cpu_to_be32( qla25xx_fdmi_port_speed_currently(ha)); alen = sizeof(eiter->a.cur_speed); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20c2, "CURRENT SPEED = %x.\n", be32_to_cpu(eiter->a.cur_speed)); /* Max frame size. */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_PORT_MAX_FRAME_SIZE); eiter->a.max_frame_size = cpu_to_be32(ha->frame_payload_size); alen = sizeof(eiter->a.max_frame_size); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20c3, "MAX FRAME SIZE = %x.\n", be32_to_cpu(eiter->a.max_frame_size)); /* OS device name. */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_PORT_OS_DEVICE_NAME); alen = scnprintf( eiter->a.os_dev_name, sizeof(eiter->a.os_dev_name), "%s:host%lu", QLA2XXX_DRIVER_NAME, vha->host_no); alen += FDMI_ATTR_ALIGNMENT(alen); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20c4, "OS DEVICE NAME = %s.\n", eiter->a.os_dev_name); /* Hostname. */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_PORT_HOST_NAME); if (!*hostname || !strncmp(hostname, "(none)", 6)) hostname = "Linux-default"; alen = scnprintf( eiter->a.host_name, sizeof(eiter->a.host_name), "%s", hostname); alen += FDMI_ATTR_ALIGNMENT(alen); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20c5, "HOSTNAME = %s.\n", eiter->a.host_name); if (callopt == CALLOPT_FDMI1) goto done; /* Node Name */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_PORT_NODE_NAME); memcpy(eiter->a.node_name, vha->node_name, sizeof(eiter->a.node_name)); alen = sizeof(eiter->a.node_name); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20c6, "NODENAME = %016llx.\n", wwn_to_u64(eiter->a.node_name)); /* Port Name */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_PORT_NAME); memcpy(eiter->a.port_name, vha->port_name, sizeof(eiter->a.port_name)); alen = sizeof(eiter->a.port_name); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20c7, "PORTNAME = %016llx.\n", wwn_to_u64(eiter->a.port_name)); /* Port Symbolic Name */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_PORT_SYM_NAME); alen = qla2x00_get_sym_node_name(vha, eiter->a.port_sym_name, sizeof(eiter->a.port_sym_name)); alen += FDMI_ATTR_ALIGNMENT(alen); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20c8, "PORT SYMBOLIC NAME = %s\n", eiter->a.port_sym_name); /* Port Type */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_PORT_TYPE); eiter->a.port_type = cpu_to_be32(NS_NX_PORT_TYPE); alen = sizeof(eiter->a.port_type); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20c9, "PORT TYPE = %x.\n", be32_to_cpu(eiter->a.port_type)); /* Supported Class of Service */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_PORT_SUPP_COS); eiter->a.port_supported_cos = cpu_to_be32(FC_CLASS_3); alen = sizeof(eiter->a.port_supported_cos); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20ca, "SUPPORTED COS = %08x\n", be32_to_cpu(eiter->a.port_supported_cos)); /* Port Fabric Name */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_PORT_FABRIC_NAME); memcpy(eiter->a.fabric_name, vha->fabric_node_name, sizeof(eiter->a.fabric_name)); alen = sizeof(eiter->a.fabric_name); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20cb, "FABRIC NAME = %016llx.\n", wwn_to_u64(eiter->a.fabric_name)); /* FC4_type */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_PORT_FC4_TYPE); eiter->a.port_fc4_type[0] = 0x00; eiter->a.port_fc4_type[1] = 0x00; eiter->a.port_fc4_type[2] = 0x01; eiter->a.port_fc4_type[3] = 0x00; alen = sizeof(eiter->a.port_fc4_type); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20cc, "PORT ACTIVE FC4 TYPE = %016llx.\n", *(uint64_t *)eiter->a.port_fc4_type); /* Port State */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_PORT_STATE); eiter->a.port_state = cpu_to_be32(2); alen = sizeof(eiter->a.port_state); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20cd, "PORT_STATE = %x.\n", be32_to_cpu(eiter->a.port_state)); /* Number of Ports */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_PORT_COUNT); eiter->a.num_ports = cpu_to_be32(1); alen = sizeof(eiter->a.num_ports); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20ce, "PORT COUNT = %x.\n", be32_to_cpu(eiter->a.num_ports)); /* Port Identifier */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_PORT_IDENTIFIER); eiter->a.port_id = cpu_to_be32(vha->d_id.b24); alen = sizeof(eiter->a.port_id); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20cf, "PORT ID = %x.\n", be32_to_cpu(eiter->a.port_id)); if (callopt == CALLOPT_FDMI2 || !ql2xsmartsan) goto done; /* Smart SAN Service Category (Populate Smart SAN Initiator)*/ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_SMARTSAN_SERVICE); alen = scnprintf( eiter->a.smartsan_service, sizeof(eiter->a.smartsan_service), "%s", "Smart SAN Initiator"); alen += FDMI_ATTR_ALIGNMENT(alen); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20d0, "SMARTSAN SERVICE CATEGORY = %s.\n", eiter->a.smartsan_service); /* Smart SAN GUID (NWWN+PWWN) */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_SMARTSAN_GUID); memcpy(eiter->a.smartsan_guid, vha->node_name, WWN_SIZE); memcpy(eiter->a.smartsan_guid + WWN_SIZE, vha->port_name, WWN_SIZE); alen = sizeof(eiter->a.smartsan_guid); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20d1, "Smart SAN GUID = %016llx-%016llx\n", wwn_to_u64(eiter->a.smartsan_guid), wwn_to_u64(eiter->a.smartsan_guid + WWN_SIZE)); /* Smart SAN Version (populate "Smart SAN Version 1.0") */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_SMARTSAN_VERSION); alen = scnprintf( eiter->a.smartsan_version, sizeof(eiter->a.smartsan_version), "%s", "Smart SAN Version 2.0"); alen += FDMI_ATTR_ALIGNMENT(alen); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20d2, "SMARTSAN VERSION = %s\n", eiter->a.smartsan_version); /* Smart SAN Product Name (Specify Adapter Model No) */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_SMARTSAN_PROD_NAME); alen = scnprintf(eiter->a.smartsan_prod_name, sizeof(eiter->a.smartsan_prod_name), "ISP%04x", ha->pdev->device); alen += FDMI_ATTR_ALIGNMENT(alen); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20d3, "SMARTSAN PRODUCT NAME = %s\n", eiter->a.smartsan_prod_name); /* Smart SAN Port Info (specify: 1=Physical, 2=NPIV, 3=SRIOV) */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_SMARTSAN_PORT_INFO); eiter->a.smartsan_port_info = cpu_to_be32(vha->vp_idx ? 2 : 1); alen = sizeof(eiter->a.smartsan_port_info); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20d4, "SMARTSAN PORT INFO = %x\n", eiter->a.smartsan_port_info); /* Smart SAN Security Support */ eiter = entries + size; eiter->type = cpu_to_be16(FDMI_SMARTSAN_SECURITY_SUPPORT); eiter->a.smartsan_security_support = cpu_to_be32(1); alen = sizeof(eiter->a.smartsan_security_support); alen += FDMI_ATTR_TYPELEN(eiter); eiter->len = cpu_to_be16(alen); size += alen; ql_dbg(ql_dbg_disc, vha, 0x20d6, "SMARTSAN SECURITY SUPPORT = %d\n", be32_to_cpu(eiter->a.smartsan_security_support)); done: return size; } /** * qla2x00_fdmi_rhba() - perform RHBA FDMI registration * @vha: HA context * @callopt: Option to issue FDMI registration * * Returns 0 on success. */ static int qla2x00_fdmi_rhba(scsi_qla_host_t *vha, unsigned int callopt) { struct qla_hw_data *ha = vha->hw; unsigned long size = 0; unsigned int rval, count; ms_iocb_entry_t *ms_pkt; struct ct_sns_req *ct_req; struct ct_sns_rsp *ct_rsp; void *entries; count = callopt != CALLOPT_FDMI1 ? FDMI2_HBA_ATTR_COUNT : FDMI1_HBA_ATTR_COUNT; size = RHBA_RSP_SIZE; ql_dbg(ql_dbg_disc, vha, 0x20e0, "RHBA (callopt=%x count=%u size=%lu).\n", callopt, count, size); /* Request size adjusted after CT preparation */ ms_pkt = ha->isp_ops->prep_ms_fdmi_iocb(vha, 0, size); /* Prepare CT request */ ct_req = qla2x00_prep_ct_fdmi_req(ha->ct_sns, RHBA_CMD, size); ct_rsp = &ha->ct_sns->p.rsp; /* Prepare FDMI command entries */ memcpy(ct_req->req.rhba.hba_identifier, vha->port_name, sizeof(ct_req->req.rhba.hba_identifier)); size += sizeof(ct_req->req.rhba.hba_identifier); ct_req->req.rhba.entry_count = cpu_to_be32(1); size += sizeof(ct_req->req.rhba.entry_count); memcpy(ct_req->req.rhba.port_name, vha->port_name, sizeof(ct_req->req.rhba.port_name)); size += sizeof(ct_req->req.rhba.port_name); /* Attribute count */ ct_req->req.rhba.attrs.count = cpu_to_be32(count); size += sizeof(ct_req->req.rhba.attrs.count); /* Attribute block */ entries = &ct_req->req.rhba.attrs.entry; size += qla2x00_hba_attributes(vha, entries, callopt); /* Update MS request size. */ qla2x00_update_ms_fdmi_iocb(vha, size + 16); ql_dbg(ql_dbg_disc, vha, 0x20e1, "RHBA %016llx %016llx.\n", wwn_to_u64(ct_req->req.rhba.hba_identifier), wwn_to_u64(ct_req->req.rhba.port_name)); ql_dump_buffer(ql_dbg_disc + ql_dbg_buffer, vha, 0x20e2, entries, size); /* Execute MS IOCB */ rval = qla2x00_issue_iocb(vha, ha->ms_iocb, ha->ms_iocb_dma, sizeof(*ha->ms_iocb)); if (rval) { ql_dbg(ql_dbg_disc, vha, 0x20e3, "RHBA iocb failed (%d).\n", rval); return rval; } rval = qla2x00_chk_ms_status(vha, ms_pkt, ct_rsp, "RHBA"); if (rval) { if (ct_rsp->header.reason_code == CT_REASON_CANNOT_PERFORM && ct_rsp->header.explanation_code == CT_EXPL_ALREADY_REGISTERED) { ql_dbg(ql_dbg_disc, vha, 0x20e4, "RHBA already registered.\n"); return QLA_ALREADY_REGISTERED; } ql_dbg(ql_dbg_disc, vha, 0x20e5, "RHBA failed, CT Reason %#x, CT Explanation %#x\n", ct_rsp->header.reason_code, ct_rsp->header.explanation_code); return rval; } ql_dbg(ql_dbg_disc, vha, 0x20e6, "RHBA exiting normally.\n"); return rval; } static int qla2x00_fdmi_dhba(scsi_qla_host_t *vha) { int rval; struct qla_hw_data *ha = vha->hw; ms_iocb_entry_t *ms_pkt; struct ct_sns_req *ct_req; struct ct_sns_rsp *ct_rsp; /* Issue RPA */ /* Prepare common MS IOCB */ ms_pkt = ha->isp_ops->prep_ms_fdmi_iocb(vha, DHBA_REQ_SIZE, DHBA_RSP_SIZE); /* Prepare CT request */ ct_req = qla2x00_prep_ct_fdmi_req(ha->ct_sns, DHBA_CMD, DHBA_RSP_SIZE); ct_rsp = &ha->ct_sns->p.rsp; /* Prepare FDMI command arguments -- portname. */ memcpy(ct_req->req.dhba.port_name, vha->port_name, WWN_SIZE); ql_dbg(ql_dbg_disc, vha, 0x2036, "DHBA portname = %8phN.\n", ct_req->req.dhba.port_name); /* Execute MS IOCB */ rval = qla2x00_issue_iocb(vha, ha->ms_iocb, ha->ms_iocb_dma, sizeof(ms_iocb_entry_t)); if (rval != QLA_SUCCESS) { /*EMPTY*/ ql_dbg(ql_dbg_disc, vha, 0x2037, "DHBA issue IOCB failed (%d).\n", rval); } else if (qla2x00_chk_ms_status(vha, ms_pkt, ct_rsp, "DHBA") != QLA_SUCCESS) { rval = QLA_FUNCTION_FAILED; } else { ql_dbg(ql_dbg_disc, vha, 0x2038, "DHBA exiting normally.\n"); } return rval; } /** * qla2x00_fdmi_rprt() - perform RPRT registration * @vha: HA context * @callopt: Option to issue extended or standard FDMI * command parameter * * Returns 0 on success. */ static int qla2x00_fdmi_rprt(scsi_qla_host_t *vha, int callopt) { struct scsi_qla_host *base_vha = pci_get_drvdata(vha->hw->pdev); struct qla_hw_data *ha = vha->hw; ulong size = 0; uint rval, count; ms_iocb_entry_t *ms_pkt; struct ct_sns_req *ct_req; struct ct_sns_rsp *ct_rsp; void *entries; count = callopt == CALLOPT_FDMI2_SMARTSAN && ql2xsmartsan ? FDMI2_SMARTSAN_PORT_ATTR_COUNT : callopt != CALLOPT_FDMI1 ? FDMI2_PORT_ATTR_COUNT : FDMI1_PORT_ATTR_COUNT; size = RPRT_RSP_SIZE; ql_dbg(ql_dbg_disc, vha, 0x20e8, "RPRT (callopt=%x count=%u size=%lu).\n", callopt, count, size); /* Request size adjusted after CT preparation */ ms_pkt = ha->isp_ops->prep_ms_fdmi_iocb(vha, 0, size); /* Prepare CT request */ ct_req = qla2x00_prep_ct_fdmi_req(ha->ct_sns, RPRT_CMD, size); ct_rsp = &ha->ct_sns->p.rsp; /* Prepare FDMI command entries */ memcpy(ct_req->req.rprt.hba_identifier, base_vha->port_name, sizeof(ct_req->req.rprt.hba_identifier)); size += sizeof(ct_req->req.rprt.hba_identifier); memcpy(ct_req->req.rprt.port_name, vha->port_name, sizeof(ct_req->req.rprt.port_name)); size += sizeof(ct_req->req.rprt.port_name); /* Attribute count */ ct_req->req.rprt.attrs.count = cpu_to_be32(count); size += sizeof(ct_req->req.rprt.attrs.count); /* Attribute block */ entries = ct_req->req.rprt.attrs.entry; size += qla2x00_port_attributes(vha, entries, callopt); /* Update MS request size. */ qla2x00_update_ms_fdmi_iocb(vha, size + 16); ql_dbg(ql_dbg_disc, vha, 0x20e9, "RPRT %016llx %016llx.\n", wwn_to_u64(ct_req->req.rprt.port_name), wwn_to_u64(ct_req->req.rprt.port_name)); ql_dump_buffer(ql_dbg_disc + ql_dbg_buffer, vha, 0x20ea, entries, size); /* Execute MS IOCB */ rval = qla2x00_issue_iocb(vha, ha->ms_iocb, ha->ms_iocb_dma, sizeof(*ha->ms_iocb)); if (rval) { ql_dbg(ql_dbg_disc, vha, 0x20eb, "RPRT iocb failed (%d).\n", rval); return rval; } rval = qla2x00_chk_ms_status(vha, ms_pkt, ct_rsp, "RPRT"); if (rval) { if (ct_rsp->header.reason_code == CT_REASON_CANNOT_PERFORM && ct_rsp->header.explanation_code == CT_EXPL_ALREADY_REGISTERED) { ql_dbg(ql_dbg_disc, vha, 0x20ec, "RPRT already registered.\n"); return QLA_ALREADY_REGISTERED; } ql_dbg(ql_dbg_disc, vha, 0x20ed, "RPRT failed, CT Reason code: %#x, CT Explanation %#x\n", ct_rsp->header.reason_code, ct_rsp->header.explanation_code); return rval; } ql_dbg(ql_dbg_disc, vha, 0x20ee, "RPRT exiting normally.\n"); return rval; } /** * qla2x00_fdmi_rpa() - perform RPA registration * @vha: HA context * @callopt: Option to issue FDMI registration * * Returns 0 on success. */ static int qla2x00_fdmi_rpa(scsi_qla_host_t *vha, uint callopt) { struct qla_hw_data *ha = vha->hw; ulong size = 0; uint rval, count; ms_iocb_entry_t *ms_pkt; struct ct_sns_req *ct_req; struct ct_sns_rsp *ct_rsp; void *entries; count = callopt == CALLOPT_FDMI2_SMARTSAN && ql2xsmartsan ? FDMI2_SMARTSAN_PORT_ATTR_COUNT : callopt != CALLOPT_FDMI1 ? FDMI2_PORT_ATTR_COUNT : FDMI1_PORT_ATTR_COUNT; size = callopt != CALLOPT_FDMI1 ? SMARTSAN_RPA_RSP_SIZE : RPA_RSP_SIZE; ql_dbg(ql_dbg_disc, vha, 0x20f0, "RPA (callopt=%x count=%u size=%lu).\n", callopt, count, size); /* Request size adjusted after CT preparation */ ms_pkt = ha->isp_ops->prep_ms_fdmi_iocb(vha, 0, size); /* Prepare CT request */ ct_req = qla2x00_prep_ct_fdmi_req(ha->ct_sns, RPA_CMD, size); ct_rsp = &ha->ct_sns->p.rsp; /* Prepare FDMI command entries. */ memcpy(ct_req->req.rpa.port_name, vha->port_name, sizeof(ct_req->req.rpa.port_name)); size += sizeof(ct_req->req.rpa.port_name); /* Attribute count */ ct_req->req.rpa.attrs.count = cpu_to_be32(count); size += sizeof(ct_req->req.rpa.attrs.count); /* Attribute block */ entries = ct_req->req.rpa.attrs.entry; size += qla2x00_port_attributes(vha, entries, callopt); /* Update MS request size. */ qla2x00_update_ms_fdmi_iocb(vha, size + 16); ql_dbg(ql_dbg_disc, vha, 0x20f1, "RPA %016llx.\n", wwn_to_u64(ct_req->req.rpa.port_name)); ql_dump_buffer(ql_dbg_disc + ql_dbg_buffer, vha, 0x20f2, entries, size); /* Execute MS IOCB */ rval = qla2x00_issue_iocb(vha, ha->ms_iocb, ha->ms_iocb_dma, sizeof(*ha->ms_iocb)); if (rval) { ql_dbg(ql_dbg_disc, vha, 0x20f3, "RPA iocb failed (%d).\n", rval); return rval; } rval = qla2x00_chk_ms_status(vha, ms_pkt, ct_rsp, "RPA"); if (rval) { if (ct_rsp->header.reason_code == CT_REASON_CANNOT_PERFORM && ct_rsp->header.explanation_code == CT_EXPL_ALREADY_REGISTERED) { ql_dbg(ql_dbg_disc, vha, 0x20f4, "RPA already registered.\n"); return QLA_ALREADY_REGISTERED; } ql_dbg(ql_dbg_disc, vha, 0x20f5, "RPA failed, CT Reason code: %#x, CT Explanation %#x\n", ct_rsp->header.reason_code, ct_rsp->header.explanation_code); return rval; } ql_dbg(ql_dbg_disc, vha, 0x20f6, "RPA exiting normally.\n"); return rval; } /** * qla2x00_fdmi_register() - * @vha: HA context * * Returns 0 on success. */ int qla2x00_fdmi_register(scsi_qla_host_t *vha) { int rval = QLA_SUCCESS; struct qla_hw_data *ha = vha->hw; if (IS_QLA2100(ha) || IS_QLA2200(ha) || IS_QLAFX00(ha)) return rval; rval = qla2x00_mgmt_svr_login(vha); if (rval) return rval; /* For npiv/vport send rprt only */ if (vha->vp_idx) { if (ql2xsmartsan) rval = qla2x00_fdmi_rprt(vha, CALLOPT_FDMI2_SMARTSAN); if (rval || !ql2xsmartsan) rval = qla2x00_fdmi_rprt(vha, CALLOPT_FDMI2); if (rval) rval = qla2x00_fdmi_rprt(vha, CALLOPT_FDMI1); return rval; } /* Try fdmi2 first, if fails then try fdmi1 */ rval = qla2x00_fdmi_rhba(vha, CALLOPT_FDMI2); if (rval) { if (rval != QLA_ALREADY_REGISTERED) goto try_fdmi; rval = qla2x00_fdmi_dhba(vha); if (rval) goto try_fdmi; rval = qla2x00_fdmi_rhba(vha, CALLOPT_FDMI2); if (rval) goto try_fdmi; } if (ql2xsmartsan) rval = qla2x00_fdmi_rpa(vha, CALLOPT_FDMI2_SMARTSAN); if (rval || !ql2xsmartsan) rval = qla2x00_fdmi_rpa(vha, CALLOPT_FDMI2); if (rval) goto try_fdmi; return rval; try_fdmi: rval = qla2x00_fdmi_rhba(vha, CALLOPT_FDMI1); if (rval) { if (rval != QLA_ALREADY_REGISTERED) return rval; rval = qla2x00_fdmi_dhba(vha); if (rval) return rval; rval = qla2x00_fdmi_rhba(vha, CALLOPT_FDMI1); if (rval) return rval; } rval = qla2x00_fdmi_rpa(vha, CALLOPT_FDMI1); return rval; } /** * qla2x00_gfpn_id() - SNS Get Fabric Port Name (GFPN_ID) query. * @vha: HA context * @list: switch info entries to populate * * Returns 0 on success. */ int qla2x00_gfpn_id(scsi_qla_host_t *vha, sw_info_t *list) { int rval = QLA_SUCCESS; uint16_t i; struct qla_hw_data *ha = vha->hw; ms_iocb_entry_t *ms_pkt; struct ct_sns_req *ct_req; struct ct_sns_rsp *ct_rsp; struct ct_arg arg; if (!IS_IIDMA_CAPABLE(ha)) return QLA_FUNCTION_FAILED; arg.iocb = ha->ms_iocb; arg.req_dma = ha->ct_sns_dma; arg.rsp_dma = ha->ct_sns_dma; arg.req_size = GFPN_ID_REQ_SIZE; arg.rsp_size = GFPN_ID_RSP_SIZE; arg.nport_handle = NPH_SNS; for (i = 0; i < ha->max_fibre_devices; i++) { /* Issue GFPN_ID */ /* Prepare common MS IOCB */ ms_pkt = ha->isp_ops->prep_ms_iocb(vha, &arg); /* Prepare CT request */ ct_req = qla2x00_prep_ct_req(ha->ct_sns, GFPN_ID_CMD, GFPN_ID_RSP_SIZE); ct_rsp = &ha->ct_sns->p.rsp; /* Prepare CT arguments -- port_id */ ct_req->req.port_id.port_id = port_id_to_be_id(list[i].d_id); /* Execute MS IOCB */ rval = qla2x00_issue_iocb(vha, ha->ms_iocb, ha->ms_iocb_dma, sizeof(ms_iocb_entry_t)); if (rval != QLA_SUCCESS) { /*EMPTY*/ ql_dbg(ql_dbg_disc, vha, 0x2023, "GFPN_ID issue IOCB failed (%d).\n", rval); break; } else if (qla2x00_chk_ms_status(vha, ms_pkt, ct_rsp, "GFPN_ID") != QLA_SUCCESS) { rval = QLA_FUNCTION_FAILED; break; } else { /* Save fabric portname */ memcpy(list[i].fabric_port_name, ct_rsp->rsp.gfpn_id.port_name, WWN_SIZE); } /* Last device exit. */ if (list[i].d_id.b.rsvd_1 != 0) break; } return (rval); } static inline struct ct_sns_req * qla24xx_prep_ct_fm_req(struct ct_sns_pkt *p, uint16_t cmd, uint16_t rsp_size) { memset(p, 0, sizeof(struct ct_sns_pkt)); p->p.req.header.revision = 0x01; p->p.req.header.gs_type = 0xFA; p->p.req.header.gs_subtype = 0x01; p->p.req.command = cpu_to_be16(cmd); p->p.req.max_rsp_size = cpu_to_be16((rsp_size - 16) / 4); return &p->p.req; } static uint16_t qla2x00_port_speed_capability(uint16_t speed) { switch (speed) { case BIT_15: return PORT_SPEED_1GB; case BIT_14: return PORT_SPEED_2GB; case BIT_13: return PORT_SPEED_4GB; case BIT_12: return PORT_SPEED_10GB; case BIT_11: return PORT_SPEED_8GB; case BIT_10: return PORT_SPEED_16GB; case BIT_8: return PORT_SPEED_32GB; case BIT_7: return PORT_SPEED_64GB; default: return PORT_SPEED_UNKNOWN; } } /** * qla2x00_gpsc() - FCS Get Port Speed Capabilities (GPSC) query. * @vha: HA context * @list: switch info entries to populate * * Returns 0 on success. */ int qla2x00_gpsc(scsi_qla_host_t *vha, sw_info_t *list) { int rval; uint16_t i; struct qla_hw_data *ha = vha->hw; ms_iocb_entry_t *ms_pkt; struct ct_sns_req *ct_req; struct ct_sns_rsp *ct_rsp; struct ct_arg arg; if (!IS_IIDMA_CAPABLE(ha)) return QLA_FUNCTION_FAILED; if (!ha->flags.gpsc_supported) return QLA_FUNCTION_FAILED; rval = qla2x00_mgmt_svr_login(vha); if (rval) return rval; arg.iocb = ha->ms_iocb; arg.req_dma = ha->ct_sns_dma; arg.rsp_dma = ha->ct_sns_dma; arg.req_size = GPSC_REQ_SIZE; arg.rsp_size = GPSC_RSP_SIZE; arg.nport_handle = vha->mgmt_svr_loop_id; for (i = 0; i < ha->max_fibre_devices; i++) { /* Issue GFPN_ID */ /* Prepare common MS IOCB */ ms_pkt = qla24xx_prep_ms_iocb(vha, &arg); /* Prepare CT request */ ct_req = qla24xx_prep_ct_fm_req(ha->ct_sns, GPSC_CMD, GPSC_RSP_SIZE); ct_rsp = &ha->ct_sns->p.rsp; /* Prepare CT arguments -- port_name */ memcpy(ct_req->req.gpsc.port_name, list[i].fabric_port_name, WWN_SIZE); /* Execute MS IOCB */ rval = qla2x00_issue_iocb(vha, ha->ms_iocb, ha->ms_iocb_dma, sizeof(ms_iocb_entry_t)); if (rval != QLA_SUCCESS) { /*EMPTY*/ ql_dbg(ql_dbg_disc, vha, 0x2059, "GPSC issue IOCB failed (%d).\n", rval); } else if ((rval = qla2x00_chk_ms_status(vha, ms_pkt, ct_rsp, "GPSC")) != QLA_SUCCESS) { /* FM command unsupported? */ if (rval == QLA_INVALID_COMMAND && (ct_rsp->header.reason_code == CT_REASON_INVALID_COMMAND_CODE || ct_rsp->header.reason_code == CT_REASON_COMMAND_UNSUPPORTED)) { ql_dbg(ql_dbg_disc, vha, 0x205a, "GPSC command unsupported, disabling " "query.\n"); ha->flags.gpsc_supported = 0; rval = QLA_FUNCTION_FAILED; break; } rval = QLA_FUNCTION_FAILED; } else { list->fp_speed = qla2x00_port_speed_capability( be16_to_cpu(ct_rsp->rsp.gpsc.speed)); ql_dbg(ql_dbg_disc, vha, 0x205b, "GPSC ext entry - fpn " "%8phN speeds=%04x speed=%04x.\n", list[i].fabric_port_name, be16_to_cpu(ct_rsp->rsp.gpsc.speeds), be16_to_cpu(ct_rsp->rsp.gpsc.speed)); } /* Last device exit. */ if (list[i].d_id.b.rsvd_1 != 0) break; } return (rval); } /** * qla2x00_gff_id() - SNS Get FC-4 Features (GFF_ID) query. * * @vha: HA context * @list: switch info entries to populate * */ void qla2x00_gff_id(scsi_qla_host_t *vha, sw_info_t *list) { int rval; uint16_t i; ms_iocb_entry_t *ms_pkt; struct ct_sns_req *ct_req; struct ct_sns_rsp *ct_rsp; struct qla_hw_data *ha = vha->hw; uint8_t fcp_scsi_features = 0, nvme_features = 0; struct ct_arg arg; for (i = 0; i < ha->max_fibre_devices; i++) { /* Set default FC4 Type as UNKNOWN so the default is to * Process this port */ list[i].fc4_type = 0; /* Do not attempt GFF_ID if we are not FWI_2 capable */ if (!IS_FWI2_CAPABLE(ha)) continue; arg.iocb = ha->ms_iocb; arg.req_dma = ha->ct_sns_dma; arg.rsp_dma = ha->ct_sns_dma; arg.req_size = GFF_ID_REQ_SIZE; arg.rsp_size = GFF_ID_RSP_SIZE; arg.nport_handle = NPH_SNS; /* Prepare common MS IOCB */ ms_pkt = ha->isp_ops->prep_ms_iocb(vha, &arg); /* Prepare CT request */ ct_req = qla2x00_prep_ct_req(ha->ct_sns, GFF_ID_CMD, GFF_ID_RSP_SIZE); ct_rsp = &ha->ct_sns->p.rsp; /* Prepare CT arguments -- port_id */ ct_req->req.port_id.port_id = port_id_to_be_id(list[i].d_id); /* Execute MS IOCB */ rval = qla2x00_issue_iocb(vha, ha->ms_iocb, ha->ms_iocb_dma, sizeof(ms_iocb_entry_t)); if (rval != QLA_SUCCESS) { ql_dbg(ql_dbg_disc, vha, 0x205c, "GFF_ID issue IOCB failed (%d).\n", rval); } else if (qla2x00_chk_ms_status(vha, ms_pkt, ct_rsp, "GFF_ID") != QLA_SUCCESS) { ql_dbg(ql_dbg_disc, vha, 0x205d, "GFF_ID IOCB status had a failure status code.\n"); } else { fcp_scsi_features = ct_rsp->rsp.gff_id.fc4_features[GFF_FCP_SCSI_OFFSET]; fcp_scsi_features &= 0x0f; if (fcp_scsi_features) { list[i].fc4_type = FS_FC4TYPE_FCP; list[i].fc4_features = fcp_scsi_features; } nvme_features = ct_rsp->rsp.gff_id.fc4_features[GFF_NVME_OFFSET]; nvme_features &= 0xf; if (nvme_features) { list[i].fc4_type |= FS_FC4TYPE_NVME; list[i].fc4_features = nvme_features; } } /* Last device exit. */ if (list[i].d_id.b.rsvd_1 != 0) break; } } int qla24xx_post_gpsc_work(struct scsi_qla_host *vha, fc_port_t *fcport) { struct qla_work_evt *e; e = qla2x00_alloc_work(vha, QLA_EVT_GPSC); if (!e) return QLA_FUNCTION_FAILED; e->u.fcport.fcport = fcport; return qla2x00_post_work(vha, e); } void qla24xx_handle_gpsc_event(scsi_qla_host_t *vha, struct event_arg *ea) { struct fc_port *fcport = ea->fcport; ql_dbg(ql_dbg_disc, vha, 0x20d8, "%s %8phC DS %d LS %d rc %d login %d|%d rscn %d|%d lid %d\n", __func__, fcport->port_name, fcport->disc_state, fcport->fw_login_state, ea->rc, ea->sp->gen2, fcport->login_gen, ea->sp->gen2, fcport->rscn_gen|ea->sp->gen1, fcport->loop_id); if (fcport->disc_state == DSC_DELETE_PEND) return; /* We will figure-out what happen after AUTH completes */ if (fcport->disc_state == DSC_LOGIN_AUTH_PEND) return; if (ea->sp->gen2 != fcport->login_gen) { /* target side must have changed it. */ ql_dbg(ql_dbg_disc, vha, 0x20d3, "%s %8phC generation changed\n", __func__, fcport->port_name); return; } else if (ea->sp->gen1 != fcport->rscn_gen) { return; } qla_post_iidma_work(vha, fcport); } static void qla24xx_async_gpsc_sp_done(srb_t *sp, int res) { struct scsi_qla_host *vha = sp->vha; struct qla_hw_data *ha = vha->hw; fc_port_t *fcport = sp->fcport; struct ct_sns_rsp *ct_rsp; struct event_arg ea; ct_rsp = &fcport->ct_desc.ct_sns->p.rsp; ql_dbg(ql_dbg_disc, vha, 0x2053, "Async done-%s res %x, WWPN %8phC \n", sp->name, res, fcport->port_name); fcport->flags &= ~(FCF_ASYNC_SENT | FCF_ASYNC_ACTIVE); if (res == QLA_FUNCTION_TIMEOUT) goto done; if (res == (DID_ERROR << 16)) { /* entry status error */ goto done; } else if (res) { if ((ct_rsp->header.reason_code == CT_REASON_INVALID_COMMAND_CODE) || (ct_rsp->header.reason_code == CT_REASON_COMMAND_UNSUPPORTED)) { ql_dbg(ql_dbg_disc, vha, 0x2019, "GPSC command unsupported, disabling query.\n"); ha->flags.gpsc_supported = 0; goto done; } } else { fcport->fp_speed = qla2x00_port_speed_capability( be16_to_cpu(ct_rsp->rsp.gpsc.speed)); ql_dbg(ql_dbg_disc, vha, 0x2054, "Async-%s OUT WWPN %8phC speeds=%04x speed=%04x.\n", sp->name, fcport->fabric_port_name, be16_to_cpu(ct_rsp->rsp.gpsc.speeds), be16_to_cpu(ct_rsp->rsp.gpsc.speed)); } memset(&ea, 0, sizeof(ea)); ea.rc = res; ea.fcport = fcport; ea.sp = sp; qla24xx_handle_gpsc_event(vha, &ea); done: /* ref: INIT */ kref_put(&sp->cmd_kref, qla2x00_sp_release); } int qla24xx_async_gpsc(scsi_qla_host_t *vha, fc_port_t *fcport) { int rval = QLA_FUNCTION_FAILED; struct ct_sns_req *ct_req; srb_t *sp; if (!vha->flags.online || (fcport->flags & FCF_ASYNC_SENT)) return rval; /* ref: INIT */ sp = qla2x00_get_sp(vha, fcport, GFP_KERNEL); if (!sp) goto done; sp->type = SRB_CT_PTHRU_CMD; sp->name = "gpsc"; sp->gen1 = fcport->rscn_gen; sp->gen2 = fcport->login_gen; qla2x00_init_async_sp(sp, qla2x00_get_async_timeout(vha) + 2, qla24xx_async_gpsc_sp_done); /* CT_IU preamble */ ct_req = qla24xx_prep_ct_fm_req(fcport->ct_desc.ct_sns, GPSC_CMD, GPSC_RSP_SIZE); /* GPSC req */ memcpy(ct_req->req.gpsc.port_name, fcport->fabric_port_name, WWN_SIZE); sp->u.iocb_cmd.u.ctarg.req = fcport->ct_desc.ct_sns; sp->u.iocb_cmd.u.ctarg.req_dma = fcport->ct_desc.ct_sns_dma; sp->u.iocb_cmd.u.ctarg.rsp = fcport->ct_desc.ct_sns; sp->u.iocb_cmd.u.ctarg.rsp_dma = fcport->ct_desc.ct_sns_dma; sp->u.iocb_cmd.u.ctarg.req_size = GPSC_REQ_SIZE; sp->u.iocb_cmd.u.ctarg.rsp_size = GPSC_RSP_SIZE; sp->u.iocb_cmd.u.ctarg.nport_handle = vha->mgmt_svr_loop_id; ql_dbg(ql_dbg_disc, vha, 0x205e, "Async-%s %8phC hdl=%x loopid=%x portid=%02x%02x%02x.\n", sp->name, fcport->port_name, sp->handle, fcport->loop_id, fcport->d_id.b.domain, fcport->d_id.b.area, fcport->d_id.b.al_pa); rval = qla2x00_start_sp(sp); if (rval != QLA_SUCCESS) goto done_free_sp; return rval; done_free_sp: /* ref: INIT */ kref_put(&sp->cmd_kref, qla2x00_sp_release); done: return rval; } void qla24xx_sp_unmap(scsi_qla_host_t *vha, srb_t *sp) { struct srb_iocb *c = &sp->u.iocb_cmd; switch (sp->type) { case SRB_ELS_DCMD: qla2x00_els_dcmd2_free(vha, &c->u.els_plogi); break; case SRB_CT_PTHRU_CMD: default: if (sp->u.iocb_cmd.u.ctarg.req) { dma_free_coherent(&vha->hw->pdev->dev, sp->u.iocb_cmd.u.ctarg.req_allocated_size, sp->u.iocb_cmd.u.ctarg.req, sp->u.iocb_cmd.u.ctarg.req_dma); sp->u.iocb_cmd.u.ctarg.req = NULL; } if (sp->u.iocb_cmd.u.ctarg.rsp) { dma_free_coherent(&vha->hw->pdev->dev, sp->u.iocb_cmd.u.ctarg.rsp_allocated_size, sp->u.iocb_cmd.u.ctarg.rsp, sp->u.iocb_cmd.u.ctarg.rsp_dma); sp->u.iocb_cmd.u.ctarg.rsp = NULL; } break; } /* ref: INIT */ kref_put(&sp->cmd_kref, qla2x00_sp_release); } void qla24xx_async_gffid_sp_done(srb_t *sp, int res) { struct scsi_qla_host *vha = sp->vha; fc_port_t *fcport = sp->fcport; struct ct_sns_rsp *ct_rsp; uint8_t fc4_scsi_feat; uint8_t fc4_nvme_feat; ql_dbg(ql_dbg_disc, vha, 0x2133, "Async done-%s res %x ID %x. %8phC\n", sp->name, res, fcport->d_id.b24, fcport->port_name); ct_rsp = sp->u.iocb_cmd.u.ctarg.rsp; fc4_scsi_feat = ct_rsp->rsp.gff_id.fc4_features[GFF_FCP_SCSI_OFFSET]; fc4_nvme_feat = ct_rsp->rsp.gff_id.fc4_features[GFF_NVME_OFFSET]; sp->rc = res; /* * FC-GS-7, 5.2.3.12 FC-4 Features - format * The format of the FC-4 Features object, as defined by the FC-4, * Shall be an array of 4-bit values, one for each type code value */ if (!res) { if (fc4_scsi_feat & 0xf) { /* w1 b00:03 */ fcport->fc4_type = FS_FC4TYPE_FCP; fcport->fc4_features = fc4_scsi_feat & 0xf; } if (fc4_nvme_feat & 0xf) { /* w5 [00:03]/28h */ fcport->fc4_type |= FS_FC4TYPE_NVME; fcport->fc4_features = fc4_nvme_feat & 0xf; } } if (sp->flags & SRB_WAKEUP_ON_COMP) { complete(sp->comp); } else { if (sp->u.iocb_cmd.u.ctarg.req) { dma_free_coherent(&vha->hw->pdev->dev, sp->u.iocb_cmd.u.ctarg.req_allocated_size, sp->u.iocb_cmd.u.ctarg.req, sp->u.iocb_cmd.u.ctarg.req_dma); sp->u.iocb_cmd.u.ctarg.req = NULL; } if (sp->u.iocb_cmd.u.ctarg.rsp) { dma_free_coherent(&vha->hw->pdev->dev, sp->u.iocb_cmd.u.ctarg.rsp_allocated_size, sp->u.iocb_cmd.u.ctarg.rsp, sp->u.iocb_cmd.u.ctarg.rsp_dma); sp->u.iocb_cmd.u.ctarg.rsp = NULL; } /* ref: INIT */ kref_put(&sp->cmd_kref, qla2x00_sp_release); /* we should not be here */ dump_stack(); } } /* Get FC4 Feature with Nport ID. */ int qla24xx_async_gffid(scsi_qla_host_t *vha, fc_port_t *fcport, bool wait) { int rval = QLA_FUNCTION_FAILED; struct ct_sns_req *ct_req; srb_t *sp; DECLARE_COMPLETION_ONSTACK(comp); /* this routine does not have handling for no wait */ if (!vha->flags.online || !wait) return rval; /* ref: INIT */ sp = qla2x00_get_sp(vha, fcport, GFP_KERNEL); if (!sp) return rval; sp->type = SRB_CT_PTHRU_CMD; sp->name = "gffid"; sp->gen1 = fcport->rscn_gen; sp->gen2 = fcport->login_gen; qla2x00_init_async_sp(sp, qla2x00_get_async_timeout(vha) + 2, qla24xx_async_gffid_sp_done); sp->comp = ∁ sp->u.iocb_cmd.timeout = qla2x00_els_dcmd2_iocb_timeout; if (wait) sp->flags = SRB_WAKEUP_ON_COMP; sp->u.iocb_cmd.u.ctarg.req_allocated_size = sizeof(struct ct_sns_pkt); sp->u.iocb_cmd.u.ctarg.req = dma_alloc_coherent(&vha->hw->pdev->dev, sp->u.iocb_cmd.u.ctarg.req_allocated_size, &sp->u.iocb_cmd.u.ctarg.req_dma, GFP_KERNEL); if (!sp->u.iocb_cmd.u.ctarg.req) { ql_log(ql_log_warn, vha, 0xd041, "%s: Failed to allocate ct_sns request.\n", __func__); goto done_free_sp; } sp->u.iocb_cmd.u.ctarg.rsp_allocated_size = sizeof(struct ct_sns_pkt); sp->u.iocb_cmd.u.ctarg.rsp = dma_alloc_coherent(&vha->hw->pdev->dev, sp->u.iocb_cmd.u.ctarg.rsp_allocated_size, &sp->u.iocb_cmd.u.ctarg.rsp_dma, GFP_KERNEL); if (!sp->u.iocb_cmd.u.ctarg.rsp) { ql_log(ql_log_warn, vha, 0xd041, "%s: Failed to allocate ct_sns response.\n", __func__); goto done_free_sp; } /* CT_IU preamble */ ct_req = qla2x00_prep_ct_req(sp->u.iocb_cmd.u.ctarg.req, GFF_ID_CMD, GFF_ID_RSP_SIZE); ct_req->req.gff_id.port_id[0] = fcport->d_id.b.domain; ct_req->req.gff_id.port_id[1] = fcport->d_id.b.area; ct_req->req.gff_id.port_id[2] = fcport->d_id.b.al_pa; sp->u.iocb_cmd.u.ctarg.req_size = GFF_ID_REQ_SIZE; sp->u.iocb_cmd.u.ctarg.rsp_size = GFF_ID_RSP_SIZE; sp->u.iocb_cmd.u.ctarg.nport_handle = NPH_SNS; rval = qla2x00_start_sp(sp); if (rval != QLA_SUCCESS) { rval = QLA_FUNCTION_FAILED; goto done_free_sp; } else { ql_dbg(ql_dbg_disc, vha, 0x3074, "Async-%s hdl=%x portid %06x\n", sp->name, sp->handle, fcport->d_id.b24); } wait_for_completion(sp->comp); rval = sp->rc; done_free_sp: if (sp->u.iocb_cmd.u.ctarg.req) { dma_free_coherent(&vha->hw->pdev->dev, sp->u.iocb_cmd.u.ctarg.req_allocated_size, sp->u.iocb_cmd.u.ctarg.req, sp->u.iocb_cmd.u.ctarg.req_dma); sp->u.iocb_cmd.u.ctarg.req = NULL; } if (sp->u.iocb_cmd.u.ctarg.rsp) { dma_free_coherent(&vha->hw->pdev->dev, sp->u.iocb_cmd.u.ctarg.rsp_allocated_size, sp->u.iocb_cmd.u.ctarg.rsp, sp->u.iocb_cmd.u.ctarg.rsp_dma); sp->u.iocb_cmd.u.ctarg.rsp = NULL; } /* ref: INIT */ kref_put(&sp->cmd_kref, qla2x00_sp_release); return rval; } /* GPN_FT + GNN_FT*/ static int qla2x00_is_a_vp(scsi_qla_host_t *vha, u64 wwn) { struct qla_hw_data *ha = vha->hw; scsi_qla_host_t *vp; unsigned long flags; u64 twwn; int rc = 0; if (!ha->num_vhosts) return 0; spin_lock_irqsave(&ha->vport_slock, flags); list_for_each_entry(vp, &ha->vp_list, list) { twwn = wwn_to_u64(vp->port_name); if (wwn == twwn) { rc = 1; break; } } spin_unlock_irqrestore(&ha->vport_slock, flags); return rc; } void qla24xx_async_gnnft_done(scsi_qla_host_t *vha, srb_t *sp) { fc_port_t *fcport; u32 i, rc; bool found; struct fab_scan_rp *rp, *trp; unsigned long flags; u8 recheck = 0; u16 dup = 0, dup_cnt = 0; ql_dbg(ql_dbg_disc + ql_dbg_verbose, vha, 0xffff, "%s enter\n", __func__); if (sp->gen1 != vha->hw->base_qpair->chip_reset) { ql_dbg(ql_dbg_disc, vha, 0xffff, "%s scan stop due to chip reset %x/%x\n", sp->name, sp->gen1, vha->hw->base_qpair->chip_reset); goto out; } rc = sp->rc; if (rc) { vha->scan.scan_retry++; if (vha->scan.scan_retry < MAX_SCAN_RETRIES) { set_bit(LOCAL_LOOP_UPDATE, &vha->dpc_flags); set_bit(LOOP_RESYNC_NEEDED, &vha->dpc_flags); goto out; } else { ql_dbg(ql_dbg_disc, vha, 0xffff, "%s: Fabric scan failed for %d retries.\n", __func__, vha->scan.scan_retry); /* * Unable to scan any rports. logout loop below * will unregister all sessions. */ list_for_each_entry(fcport, &vha->vp_fcports, list) { if ((fcport->flags & FCF_FABRIC_DEVICE) != 0) { fcport->scan_state = QLA_FCPORT_SCAN; if (fcport->loop_id == FC_NO_LOOP_ID) fcport->logout_on_delete = 0; else fcport->logout_on_delete = 1; } } goto login_logout; } } vha->scan.scan_retry = 0; list_for_each_entry(fcport, &vha->vp_fcports, list) fcport->scan_state = QLA_FCPORT_SCAN; for (i = 0; i < vha->hw->max_fibre_devices; i++) { u64 wwn; int k; rp = &vha->scan.l[i]; found = false; wwn = wwn_to_u64(rp->port_name); if (wwn == 0) continue; /* Remove duplicate NPORT ID entries from switch data base */ for (k = i + 1; k < vha->hw->max_fibre_devices; k++) { trp = &vha->scan.l[k]; if (rp->id.b24 == trp->id.b24) { dup = 1; dup_cnt++; ql_dbg(ql_dbg_disc + ql_dbg_verbose, vha, 0xffff, "Detected duplicate NPORT ID from switch data base: ID %06x WWN %8phN WWN %8phN\n", rp->id.b24, rp->port_name, trp->port_name); memset(trp, 0, sizeof(*trp)); } } if (!memcmp(rp->port_name, vha->port_name, WWN_SIZE)) continue; /* Bypass reserved domain fields. */ if ((rp->id.b.domain & 0xf0) == 0xf0) continue; /* Bypass virtual ports of the same host. */ if (qla2x00_is_a_vp(vha, wwn)) continue; list_for_each_entry(fcport, &vha->vp_fcports, list) { if (memcmp(rp->port_name, fcport->port_name, WWN_SIZE)) continue; fcport->scan_state = QLA_FCPORT_FOUND; fcport->last_rscn_gen = fcport->rscn_gen; fcport->fc4_type = rp->fc4type; found = true; if (fcport->scan_needed) { if (NVME_PRIORITY(vha->hw, fcport)) fcport->do_prli_nvme = 1; else fcport->do_prli_nvme = 0; } /* * If device was not a fabric device before. */ if ((fcport->flags & FCF_FABRIC_DEVICE) == 0) { qla2x00_clear_loop_id(fcport); fcport->flags |= FCF_FABRIC_DEVICE; } else if (fcport->d_id.b24 != rp->id.b24 || (fcport->scan_needed && fcport->port_type != FCT_INITIATOR && fcport->port_type != FCT_NVME_INITIATOR)) { qlt_schedule_sess_for_deletion(fcport); } fcport->d_id.b24 = rp->id.b24; fcport->scan_needed = 0; break; } if (!found) { ql_dbg(ql_dbg_disc, vha, 0xffff, "%s %d %8phC post new sess\n", __func__, __LINE__, rp->port_name); qla24xx_post_newsess_work(vha, &rp->id, rp->port_name, rp->node_name, NULL, rp->fc4type); } } if (dup) { ql_log(ql_log_warn, vha, 0xffff, "Detected %d duplicate NPORT ID(s) from switch data base\n", dup_cnt); } login_logout: /* * Logout all previous fabric dev marked lost, except FCP2 devices. */ list_for_each_entry(fcport, &vha->vp_fcports, list) { if ((fcport->flags & FCF_FABRIC_DEVICE) == 0) { fcport->scan_needed = 0; continue; } if (fcport->scan_state != QLA_FCPORT_FOUND) { bool do_delete = false; if (fcport->scan_needed && fcport->disc_state == DSC_LOGIN_PEND) { /* Cable got disconnected after we sent * a login. Do delete to prevent timeout. */ fcport->logout_on_delete = 1; do_delete = true; } fcport->scan_needed = 0; if (((qla_dual_mode_enabled(vha) || qla_ini_mode_enabled(vha)) && atomic_read(&fcport->state) == FCS_ONLINE) || do_delete) { if (fcport->loop_id != FC_NO_LOOP_ID) { if (fcport->flags & FCF_FCP2_DEVICE) continue; ql_log(ql_log_warn, vha, 0x20f0, "%s %d %8phC post del sess\n", __func__, __LINE__, fcport->port_name); fcport->tgt_link_down_time = 0; qlt_schedule_sess_for_deletion(fcport); continue; } } } else { if (fcport->scan_needed || fcport->disc_state != DSC_LOGIN_COMPLETE) { if (fcport->login_retry == 0) { fcport->login_retry = vha->hw->login_retry_count; ql_dbg(ql_dbg_disc, vha, 0x20a3, "Port login retry %8phN, lid 0x%04x retry cnt=%d.\n", fcport->port_name, fcport->loop_id, fcport->login_retry); } fcport->scan_needed = 0; qla24xx_fcport_handle_login(vha, fcport); } } } recheck = 1; out: qla24xx_sp_unmap(vha, sp); spin_lock_irqsave(&vha->work_lock, flags); vha->scan.scan_flags &= ~SF_SCANNING; spin_unlock_irqrestore(&vha->work_lock, flags); if (recheck) { list_for_each_entry(fcport, &vha->vp_fcports, list) { if (fcport->scan_needed) { set_bit(LOCAL_LOOP_UPDATE, &vha->dpc_flags); set_bit(LOOP_RESYNC_NEEDED, &vha->dpc_flags); break; } } } } static int qla2x00_post_gnnft_gpnft_done_work(struct scsi_qla_host *vha, srb_t *sp, int cmd) { struct qla_work_evt *e; if (cmd != QLA_EVT_GPNFT_DONE && cmd != QLA_EVT_GNNFT_DONE) return QLA_PARAMETER_ERROR; e = qla2x00_alloc_work(vha, cmd); if (!e) return QLA_FUNCTION_FAILED; e->u.iosb.sp = sp; return qla2x00_post_work(vha, e); } static int qla2x00_post_nvme_gpnft_work(struct scsi_qla_host *vha, srb_t *sp, int cmd) { struct qla_work_evt *e; if (cmd != QLA_EVT_GPNFT) return QLA_PARAMETER_ERROR; e = qla2x00_alloc_work(vha, cmd); if (!e) return QLA_FUNCTION_FAILED; e->u.gpnft.fc4_type = FC4_TYPE_NVME; e->u.gpnft.sp = sp; return qla2x00_post_work(vha, e); } static void qla2x00_find_free_fcp_nvme_slot(struct scsi_qla_host *vha, struct srb *sp) { struct qla_hw_data *ha = vha->hw; int num_fibre_dev = ha->max_fibre_devices; struct ct_sns_req *ct_req = (struct ct_sns_req *)sp->u.iocb_cmd.u.ctarg.req; struct ct_sns_gpnft_rsp *ct_rsp = (struct ct_sns_gpnft_rsp *)sp->u.iocb_cmd.u.ctarg.rsp; struct ct_sns_gpn_ft_data *d; struct fab_scan_rp *rp; u16 cmd = be16_to_cpu(ct_req->command); u8 fc4_type = sp->gen2; int i, j, k; port_id_t id; u8 found; u64 wwn; j = 0; for (i = 0; i < num_fibre_dev; i++) { d = &ct_rsp->entries[i]; id.b.rsvd_1 = 0; id.b.domain = d->port_id[0]; id.b.area = d->port_id[1]; id.b.al_pa = d->port_id[2]; wwn = wwn_to_u64(d->port_name); if (id.b24 == 0 || wwn == 0) continue; if (fc4_type == FC4_TYPE_FCP_SCSI) { if (cmd == GPN_FT_CMD) { rp = &vha->scan.l[j]; rp->id = id; memcpy(rp->port_name, d->port_name, 8); j++; rp->fc4type = FS_FC4TYPE_FCP; } else { for (k = 0; k < num_fibre_dev; k++) { rp = &vha->scan.l[k]; if (id.b24 == rp->id.b24) { memcpy(rp->node_name, d->port_name, 8); break; } } } } else { /* Search if the fibre device supports FC4_TYPE_NVME */ if (cmd == GPN_FT_CMD) { found = 0; for (k = 0; k < num_fibre_dev; k++) { rp = &vha->scan.l[k]; if (!memcmp(rp->port_name, d->port_name, 8)) { /* * Supports FC-NVMe & FCP */ rp->fc4type |= FS_FC4TYPE_NVME; found = 1; break; } } /* We found new FC-NVMe only port */ if (!found) { for (k = 0; k < num_fibre_dev; k++) { rp = &vha->scan.l[k]; if (wwn_to_u64(rp->port_name)) { continue; } else { rp->id = id; memcpy(rp->port_name, d->port_name, 8); rp->fc4type = FS_FC4TYPE_NVME; break; } } } } else { for (k = 0; k < num_fibre_dev; k++) { rp = &vha->scan.l[k]; if (id.b24 == rp->id.b24) { memcpy(rp->node_name, d->port_name, 8); break; } } } } } } static void qla2x00_async_gpnft_gnnft_sp_done(srb_t *sp, int res) { struct scsi_qla_host *vha = sp->vha; struct ct_sns_req *ct_req = (struct ct_sns_req *)sp->u.iocb_cmd.u.ctarg.req; u16 cmd = be16_to_cpu(ct_req->command); u8 fc4_type = sp->gen2; unsigned long flags; int rc; /* gen2 field is holding the fc4type */ ql_dbg(ql_dbg_disc, vha, 0xffff, "Async done-%s res %x FC4Type %x\n", sp->name, res, sp->gen2); sp->rc = res; if (res) { unsigned long flags; const char *name = sp->name; if (res == QLA_OS_TIMER_EXPIRED) { /* switch is ignoring all commands. * This might be a zone disable behavior. * This means we hit 64s timeout. * 22s GPNFT + 44s Abort = 64s */ ql_dbg(ql_dbg_disc, vha, 0xffff, "%s: Switch Zone check please .\n", name); qla2x00_mark_all_devices_lost(vha); } /* * We are in an Interrupt context, queue up this * sp for GNNFT_DONE work. This will allow all * the resource to get freed up. */ rc = qla2x00_post_gnnft_gpnft_done_work(vha, sp, QLA_EVT_GNNFT_DONE); if (rc) { /* Cleanup here to prevent memory leak */ qla24xx_sp_unmap(vha, sp); spin_lock_irqsave(&vha->work_lock, flags); vha->scan.scan_flags &= ~SF_SCANNING; vha->scan.scan_retry++; spin_unlock_irqrestore(&vha->work_lock, flags); if (vha->scan.scan_retry < MAX_SCAN_RETRIES) { set_bit(LOCAL_LOOP_UPDATE, &vha->dpc_flags); set_bit(LOOP_RESYNC_NEEDED, &vha->dpc_flags); qla2xxx_wake_dpc(vha); } else { ql_dbg(ql_dbg_disc, vha, 0xffff, "Async done-%s rescan failed on all retries.\n", name); } } return; } qla2x00_find_free_fcp_nvme_slot(vha, sp); if ((fc4_type == FC4_TYPE_FCP_SCSI) && vha->flags.nvme_enabled && cmd == GNN_FT_CMD) { spin_lock_irqsave(&vha->work_lock, flags); vha->scan.scan_flags &= ~SF_SCANNING; spin_unlock_irqrestore(&vha->work_lock, flags); sp->rc = res; rc = qla2x00_post_nvme_gpnft_work(vha, sp, QLA_EVT_GPNFT); if (rc) { qla24xx_sp_unmap(vha, sp); set_bit(LOCAL_LOOP_UPDATE, &vha->dpc_flags); set_bit(LOOP_RESYNC_NEEDED, &vha->dpc_flags); } return; } if (cmd == GPN_FT_CMD) { rc = qla2x00_post_gnnft_gpnft_done_work(vha, sp, QLA_EVT_GPNFT_DONE); } else { rc = qla2x00_post_gnnft_gpnft_done_work(vha, sp, QLA_EVT_GNNFT_DONE); } if (rc) { qla24xx_sp_unmap(vha, sp); set_bit(LOCAL_LOOP_UPDATE, &vha->dpc_flags); set_bit(LOOP_RESYNC_NEEDED, &vha->dpc_flags); return; } } /* * Get WWNN list for fc4_type * * It is assumed the same SRB is re-used from GPNFT to avoid * mem free & re-alloc */ static int qla24xx_async_gnnft(scsi_qla_host_t *vha, struct srb *sp, u8 fc4_type) { int rval = QLA_FUNCTION_FAILED; struct ct_sns_req *ct_req; struct ct_sns_pkt *ct_sns; unsigned long flags; if (!vha->flags.online) { spin_lock_irqsave(&vha->work_lock, flags); vha->scan.scan_flags &= ~SF_SCANNING; spin_unlock_irqrestore(&vha->work_lock, flags); goto done_free_sp; } if (!sp->u.iocb_cmd.u.ctarg.req || !sp->u.iocb_cmd.u.ctarg.rsp) { ql_log(ql_log_warn, vha, 0xffff, "%s: req %p rsp %p are not setup\n", __func__, sp->u.iocb_cmd.u.ctarg.req, sp->u.iocb_cmd.u.ctarg.rsp); spin_lock_irqsave(&vha->work_lock, flags); vha->scan.scan_flags &= ~SF_SCANNING; spin_unlock_irqrestore(&vha->work_lock, flags); WARN_ON(1); set_bit(LOCAL_LOOP_UPDATE, &vha->dpc_flags); set_bit(LOOP_RESYNC_NEEDED, &vha->dpc_flags); goto done_free_sp; } ql_dbg(ql_dbg_disc, vha, 0xfffff, "%s: FC4Type %x, CT-PASSTHRU %s command ctarg rsp size %d, ctarg req size %d\n", __func__, fc4_type, sp->name, sp->u.iocb_cmd.u.ctarg.rsp_size, sp->u.iocb_cmd.u.ctarg.req_size); sp->type = SRB_CT_PTHRU_CMD; sp->name = "gnnft"; sp->gen1 = vha->hw->base_qpair->chip_reset; sp->gen2 = fc4_type; qla2x00_init_async_sp(sp, qla2x00_get_async_timeout(vha) + 2, qla2x00_async_gpnft_gnnft_sp_done); memset(sp->u.iocb_cmd.u.ctarg.rsp, 0, sp->u.iocb_cmd.u.ctarg.rsp_size); memset(sp->u.iocb_cmd.u.ctarg.req, 0, sp->u.iocb_cmd.u.ctarg.req_size); ct_sns = (struct ct_sns_pkt *)sp->u.iocb_cmd.u.ctarg.req; /* CT_IU preamble */ ct_req = qla2x00_prep_ct_req(ct_sns, GNN_FT_CMD, sp->u.iocb_cmd.u.ctarg.rsp_size); /* GPN_FT req */ ct_req->req.gpn_ft.port_type = fc4_type; sp->u.iocb_cmd.u.ctarg.req_size = GNN_FT_REQ_SIZE; sp->u.iocb_cmd.u.ctarg.nport_handle = NPH_SNS; ql_dbg(ql_dbg_disc, vha, 0xffff, "Async-%s hdl=%x FC4Type %x.\n", sp->name, sp->handle, ct_req->req.gpn_ft.port_type); rval = qla2x00_start_sp(sp); if (rval != QLA_SUCCESS) { goto done_free_sp; } return rval; done_free_sp: if (sp->u.iocb_cmd.u.ctarg.req) { dma_free_coherent(&vha->hw->pdev->dev, sp->u.iocb_cmd.u.ctarg.req_allocated_size, sp->u.iocb_cmd.u.ctarg.req, sp->u.iocb_cmd.u.ctarg.req_dma); sp->u.iocb_cmd.u.ctarg.req = NULL; } if (sp->u.iocb_cmd.u.ctarg.rsp) { dma_free_coherent(&vha->hw->pdev->dev, sp->u.iocb_cmd.u.ctarg.rsp_allocated_size, sp->u.iocb_cmd.u.ctarg.rsp, sp->u.iocb_cmd.u.ctarg.rsp_dma); sp->u.iocb_cmd.u.ctarg.rsp = NULL; } /* ref: INIT */ kref_put(&sp->cmd_kref, qla2x00_sp_release); spin_lock_irqsave(&vha->work_lock, flags); vha->scan.scan_flags &= ~SF_SCANNING; if (vha->scan.scan_flags == 0) { ql_dbg(ql_dbg_disc, vha, 0xffff, "%s: schedule\n", __func__); vha->scan.scan_flags |= SF_QUEUED; schedule_delayed_work(&vha->scan.scan_work, 5); } spin_unlock_irqrestore(&vha->work_lock, flags); return rval; } /* GNNFT */ void qla24xx_async_gpnft_done(scsi_qla_host_t *vha, srb_t *sp) { ql_dbg(ql_dbg_disc + ql_dbg_verbose, vha, 0xffff, "%s enter\n", __func__); qla24xx_async_gnnft(vha, sp, sp->gen2); } /* Get WWPN list for certain fc4_type */ int qla24xx_async_gpnft(scsi_qla_host_t *vha, u8 fc4_type, srb_t *sp) { int rval = QLA_FUNCTION_FAILED; struct ct_sns_req *ct_req; struct ct_sns_pkt *ct_sns; u32 rspsz; unsigned long flags; ql_dbg(ql_dbg_disc + ql_dbg_verbose, vha, 0xffff, "%s enter\n", __func__); if (!vha->flags.online) return rval; spin_lock_irqsave(&vha->work_lock, flags); if (vha->scan.scan_flags & SF_SCANNING) { spin_unlock_irqrestore(&vha->work_lock, flags); ql_dbg(ql_dbg_disc + ql_dbg_verbose, vha, 0xffff, "%s: scan active\n", __func__); return rval; } vha->scan.scan_flags |= SF_SCANNING; spin_unlock_irqrestore(&vha->work_lock, flags); if (fc4_type == FC4_TYPE_FCP_SCSI) { ql_dbg(ql_dbg_disc + ql_dbg_verbose, vha, 0xffff, "%s: Performing FCP Scan\n", __func__); if (sp) { /* ref: INIT */ kref_put(&sp->cmd_kref, qla2x00_sp_release); } /* ref: INIT */ sp = qla2x00_get_sp(vha, NULL, GFP_KERNEL); if (!sp) { spin_lock_irqsave(&vha->work_lock, flags); vha->scan.scan_flags &= ~SF_SCANNING; spin_unlock_irqrestore(&vha->work_lock, flags); return rval; } sp->u.iocb_cmd.u.ctarg.req = dma_alloc_coherent(&vha->hw->pdev->dev, sizeof(struct ct_sns_pkt), &sp->u.iocb_cmd.u.ctarg.req_dma, GFP_KERNEL); sp->u.iocb_cmd.u.ctarg.req_allocated_size = sizeof(struct ct_sns_pkt); if (!sp->u.iocb_cmd.u.ctarg.req) { ql_log(ql_log_warn, vha, 0xffff, "Failed to allocate ct_sns request.\n"); spin_lock_irqsave(&vha->work_lock, flags); vha->scan.scan_flags &= ~SF_SCANNING; spin_unlock_irqrestore(&vha->work_lock, flags); qla2x00_rel_sp(sp); return rval; } sp->u.iocb_cmd.u.ctarg.req_size = GPN_FT_REQ_SIZE; rspsz = sizeof(struct ct_sns_gpnft_rsp) + vha->hw->max_fibre_devices * sizeof(struct ct_sns_gpn_ft_data); sp->u.iocb_cmd.u.ctarg.rsp = dma_alloc_coherent(&vha->hw->pdev->dev, rspsz, &sp->u.iocb_cmd.u.ctarg.rsp_dma, GFP_KERNEL); sp->u.iocb_cmd.u.ctarg.rsp_allocated_size = rspsz; if (!sp->u.iocb_cmd.u.ctarg.rsp) { ql_log(ql_log_warn, vha, 0xffff, "Failed to allocate ct_sns request.\n"); spin_lock_irqsave(&vha->work_lock, flags); vha->scan.scan_flags &= ~SF_SCANNING; spin_unlock_irqrestore(&vha->work_lock, flags); dma_free_coherent(&vha->hw->pdev->dev, sp->u.iocb_cmd.u.ctarg.req_allocated_size, sp->u.iocb_cmd.u.ctarg.req, sp->u.iocb_cmd.u.ctarg.req_dma); sp->u.iocb_cmd.u.ctarg.req = NULL; /* ref: INIT */ qla2x00_rel_sp(sp); return rval; } sp->u.iocb_cmd.u.ctarg.rsp_size = rspsz; ql_dbg(ql_dbg_disc + ql_dbg_verbose, vha, 0xffff, "%s scan list size %d\n", __func__, vha->scan.size); memset(vha->scan.l, 0, vha->scan.size); } else if (!sp) { ql_dbg(ql_dbg_disc, vha, 0xffff, "NVME scan did not provide SP\n"); return rval; } sp->type = SRB_CT_PTHRU_CMD; sp->name = "gpnft"; sp->gen1 = vha->hw->base_qpair->chip_reset; sp->gen2 = fc4_type; qla2x00_init_async_sp(sp, qla2x00_get_async_timeout(vha) + 2, qla2x00_async_gpnft_gnnft_sp_done); rspsz = sp->u.iocb_cmd.u.ctarg.rsp_size; memset(sp->u.iocb_cmd.u.ctarg.rsp, 0, sp->u.iocb_cmd.u.ctarg.rsp_size); memset(sp->u.iocb_cmd.u.ctarg.req, 0, sp->u.iocb_cmd.u.ctarg.req_size); ct_sns = (struct ct_sns_pkt *)sp->u.iocb_cmd.u.ctarg.req; /* CT_IU preamble */ ct_req = qla2x00_prep_ct_req(ct_sns, GPN_FT_CMD, rspsz); /* GPN_FT req */ ct_req->req.gpn_ft.port_type = fc4_type; sp->u.iocb_cmd.u.ctarg.nport_handle = NPH_SNS; ql_dbg(ql_dbg_disc, vha, 0xffff, "Async-%s hdl=%x FC4Type %x.\n", sp->name, sp->handle, ct_req->req.gpn_ft.port_type); rval = qla2x00_start_sp(sp); if (rval != QLA_SUCCESS) { goto done_free_sp; } return rval; done_free_sp: if (sp->u.iocb_cmd.u.ctarg.req) { dma_free_coherent(&vha->hw->pdev->dev, sp->u.iocb_cmd.u.ctarg.req_allocated_size, sp->u.iocb_cmd.u.ctarg.req, sp->u.iocb_cmd.u.ctarg.req_dma); sp->u.iocb_cmd.u.ctarg.req = NULL; } if (sp->u.iocb_cmd.u.ctarg.rsp) { dma_free_coherent(&vha->hw->pdev->dev, sp->u.iocb_cmd.u.ctarg.rsp_allocated_size, sp->u.iocb_cmd.u.ctarg.rsp, sp->u.iocb_cmd.u.ctarg.rsp_dma); sp->u.iocb_cmd.u.ctarg.rsp = NULL; } /* ref: INIT */ kref_put(&sp->cmd_kref, qla2x00_sp_release); spin_lock_irqsave(&vha->work_lock, flags); vha->scan.scan_flags &= ~SF_SCANNING; if (vha->scan.scan_flags == 0) { ql_dbg(ql_dbg_disc + ql_dbg_verbose, vha, 0xffff, "%s: Scan scheduled.\n", __func__); vha->scan.scan_flags |= SF_QUEUED; schedule_delayed_work(&vha->scan.scan_work, 5); } spin_unlock_irqrestore(&vha->work_lock, flags); return rval; } void qla_scan_work_fn(struct work_struct *work) { struct fab_scan *s = container_of(to_delayed_work(work), struct fab_scan, scan_work); struct scsi_qla_host *vha = container_of(s, struct scsi_qla_host, scan); unsigned long flags; ql_dbg(ql_dbg_disc, vha, 0xffff, "%s: schedule loop resync\n", __func__); set_bit(LOCAL_LOOP_UPDATE, &vha->dpc_flags); set_bit(LOOP_RESYNC_NEEDED, &vha->dpc_flags); qla2xxx_wake_dpc(vha); spin_lock_irqsave(&vha->work_lock, flags); vha->scan.scan_flags &= ~SF_QUEUED; spin_unlock_irqrestore(&vha->work_lock, flags); } /* GPFN_ID */ void qla24xx_handle_gfpnid_event(scsi_qla_host_t *vha, struct event_arg *ea) { fc_port_t *fcport = ea->fcport; ql_dbg(ql_dbg_disc, vha, 0xffff, "%s %8phC DS %d LS %d rc %d login %d|%d rscn %d|%d fcpcnt %d\n", __func__, fcport->port_name, fcport->disc_state, fcport->fw_login_state, ea->rc, fcport->login_gen, ea->sp->gen2, fcport->rscn_gen, ea->sp->gen1, vha->fcport_count); if (fcport->disc_state == DSC_DELETE_PEND) return; if (ea->sp->gen2 != fcport->login_gen) { /* target side must have changed it. */ ql_dbg(ql_dbg_disc, vha, 0x20d3, "%s %8phC generation changed\n", __func__, fcport->port_name); return; } else if (ea->sp->gen1 != fcport->rscn_gen) { return; } qla24xx_post_gpsc_work(vha, fcport); } static void qla2x00_async_gfpnid_sp_done(srb_t *sp, int res) { struct scsi_qla_host *vha = sp->vha; fc_port_t *fcport = sp->fcport; u8 *fpn = fcport->ct_desc.ct_sns->p.rsp.rsp.gfpn_id.port_name; struct event_arg ea; u64 wwn; wwn = wwn_to_u64(fpn); if (wwn) memcpy(fcport->fabric_port_name, fpn, WWN_SIZE); memset(&ea, 0, sizeof(ea)); ea.fcport = fcport; ea.sp = sp; ea.rc = res; ql_dbg(ql_dbg_disc, vha, 0x204f, "Async done-%s res %x, WWPN %8phC %8phC\n", sp->name, res, fcport->port_name, fcport->fabric_port_name); qla24xx_handle_gfpnid_event(vha, &ea); /* ref: INIT */ kref_put(&sp->cmd_kref, qla2x00_sp_release); } int qla24xx_async_gfpnid(scsi_qla_host_t *vha, fc_port_t *fcport) { int rval = QLA_FUNCTION_FAILED; struct ct_sns_req *ct_req; srb_t *sp; if (!vha->flags.online || (fcport->flags & FCF_ASYNC_SENT)) return rval; /* ref: INIT */ sp = qla2x00_get_sp(vha, fcport, GFP_ATOMIC); if (!sp) goto done; sp->type = SRB_CT_PTHRU_CMD; sp->name = "gfpnid"; sp->gen1 = fcport->rscn_gen; sp->gen2 = fcport->login_gen; qla2x00_init_async_sp(sp, qla2x00_get_async_timeout(vha) + 2, qla2x00_async_gfpnid_sp_done); /* CT_IU preamble */ ct_req = qla2x00_prep_ct_req(fcport->ct_desc.ct_sns, GFPN_ID_CMD, GFPN_ID_RSP_SIZE); /* GFPN_ID req */ ct_req->req.port_id.port_id = port_id_to_be_id(fcport->d_id); /* req & rsp use the same buffer */ sp->u.iocb_cmd.u.ctarg.req = fcport->ct_desc.ct_sns; sp->u.iocb_cmd.u.ctarg.req_dma = fcport->ct_desc.ct_sns_dma; sp->u.iocb_cmd.u.ctarg.rsp = fcport->ct_desc.ct_sns; sp->u.iocb_cmd.u.ctarg.rsp_dma = fcport->ct_desc.ct_sns_dma; sp->u.iocb_cmd.u.ctarg.req_size = GFPN_ID_REQ_SIZE; sp->u.iocb_cmd.u.ctarg.rsp_size = GFPN_ID_RSP_SIZE; sp->u.iocb_cmd.u.ctarg.nport_handle = NPH_SNS; ql_dbg(ql_dbg_disc, vha, 0xffff, "Async-%s - %8phC hdl=%x loopid=%x portid %06x.\n", sp->name, fcport->port_name, sp->handle, fcport->loop_id, fcport->d_id.b24); rval = qla2x00_start_sp(sp); if (rval != QLA_SUCCESS) goto done_free_sp; return rval; done_free_sp: /* ref: INIT */ kref_put(&sp->cmd_kref, qla2x00_sp_release); done: return rval; } int qla24xx_post_gfpnid_work(struct scsi_qla_host *vha, fc_port_t *fcport) { struct qla_work_evt *e; int ls; ls = atomic_read(&vha->loop_state); if (((ls != LOOP_READY) && (ls != LOOP_UP)) || test_bit(UNLOADING, &vha->dpc_flags)) return 0; e = qla2x00_alloc_work(vha, QLA_EVT_GFPNID); if (!e) return QLA_FUNCTION_FAILED; e->u.fcport.fcport = fcport; return qla2x00_post_work(vha, e); }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1