Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
James Smart | 5427 | 100.00% | 2 | 100.00% |
Total | 5427 | 2 |
// SPDX-License-Identifier: GPL-2.0 /* * Copyright (C) 2021 Broadcom. All Rights Reserved. The term * “Broadcom” refers to Broadcom Inc. and/or its subsidiaries. */ /* * This file implements remote node state machines for: * - Fabric logins. * - Fabric controller events. * - Name/directory services interaction. * - Point-to-point logins. */ /* * fabric_sm Node State Machine: Fabric States * ns_sm Node State Machine: Name/Directory Services States * p2p_sm Node State Machine: Point-to-Point Node States */ #include "efc.h" static void efc_fabric_initiate_shutdown(struct efc_node *node) { struct efc *efc = node->efc; node->els_io_enabled = false; if (node->attached) { int rc; /* issue hw node free; don't care if succeeds right away * or sometime later, will check node->attached later in * shutdown process */ rc = efc_cmd_node_detach(efc, &node->rnode); if (rc < 0) { node_printf(node, "Failed freeing HW node, rc=%d\n", rc); } } /* * node has either been detached or is in the process of being detached, * call common node's initiate cleanup function */ efc_node_initiate_cleanup(node); } static void __efc_fabric_common(const char *funcname, struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node *node = NULL; node = ctx->app; switch (evt) { case EFC_EVT_DOMAIN_ATTACH_OK: break; case EFC_EVT_SHUTDOWN: node->shutdown_reason = EFC_NODE_SHUTDOWN_DEFAULT; efc_fabric_initiate_shutdown(node); break; default: /* call default event handler common to all nodes */ __efc_node_common(funcname, ctx, evt, arg); } } void __efc_fabric_init(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node *node = ctx->app; struct efc *efc = node->efc; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); switch (evt) { case EFC_EVT_REENTER: efc_log_debug(efc, ">>> reenter !!\n"); fallthrough; case EFC_EVT_ENTER: /* send FLOGI */ efc_send_flogi(node); efc_node_transition(node, __efc_fabric_flogi_wait_rsp, NULL); break; default: __efc_fabric_common(__func__, ctx, evt, arg); } } void efc_fabric_set_topology(struct efc_node *node, enum efc_nport_topology topology) { node->nport->topology = topology; } void efc_fabric_notify_topology(struct efc_node *node) { struct efc_node *tmp_node; unsigned long index; /* * now loop through the nodes in the nport * and send topology notification */ xa_for_each(&node->nport->lookup, index, tmp_node) { if (tmp_node != node) { efc_node_post_event(tmp_node, EFC_EVT_NPORT_TOPOLOGY_NOTIFY, &node->nport->topology); } } } static bool efc_rnode_is_nport(struct fc_els_flogi *rsp) { return !(ntohs(rsp->fl_csp.sp_features) & FC_SP_FT_FPORT); } void __efc_fabric_flogi_wait_rsp(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node_cb *cbdata = arg; struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); switch (evt) { case EFC_EVT_SRRS_ELS_REQ_OK: { if (efc_node_check_els_req(ctx, evt, arg, ELS_FLOGI, __efc_fabric_common, __func__)) { return; } WARN_ON(!node->els_req_cnt); node->els_req_cnt--; memcpy(node->nport->domain->flogi_service_params, cbdata->els_rsp.virt, sizeof(struct fc_els_flogi)); /* Check to see if the fabric is an F_PORT or and N_PORT */ if (!efc_rnode_is_nport(cbdata->els_rsp.virt)) { /* sm: if not nport / efc_domain_attach */ /* ext_status has the fc_id, attach domain */ efc_fabric_set_topology(node, EFC_NPORT_TOPO_FABRIC); efc_fabric_notify_topology(node); WARN_ON(node->nport->domain->attached); efc_domain_attach(node->nport->domain, cbdata->ext_status); efc_node_transition(node, __efc_fabric_wait_domain_attach, NULL); break; } /* sm: if nport and p2p_winner / efc_domain_attach */ efc_fabric_set_topology(node, EFC_NPORT_TOPO_P2P); if (efc_p2p_setup(node->nport)) { node_printf(node, "p2p setup failed, shutting down node\n"); node->shutdown_reason = EFC_NODE_SHUTDOWN_DEFAULT; efc_fabric_initiate_shutdown(node); break; } if (node->nport->p2p_winner) { efc_node_transition(node, __efc_p2p_wait_domain_attach, NULL); if (node->nport->domain->attached && !node->nport->domain->domain_notify_pend) { /* * already attached, * just send ATTACH_OK */ node_printf(node, "p2p winner, domain already attached\n"); efc_node_post_event(node, EFC_EVT_DOMAIN_ATTACH_OK, NULL); } } else { /* * peer is p2p winner; * PLOGI will be received on the * remote SID=1 node; * this node has served its purpose */ node->shutdown_reason = EFC_NODE_SHUTDOWN_DEFAULT; efc_fabric_initiate_shutdown(node); } break; } case EFC_EVT_ELS_REQ_ABORTED: case EFC_EVT_SRRS_ELS_REQ_RJT: case EFC_EVT_SRRS_ELS_REQ_FAIL: { struct efc_nport *nport = node->nport; /* * with these errors, we have no recovery, * so shutdown the nport, leave the link * up and the domain ready */ if (efc_node_check_els_req(ctx, evt, arg, ELS_FLOGI, __efc_fabric_common, __func__)) { return; } node_printf(node, "FLOGI failed evt=%s, shutting down nport [%s]\n", efc_sm_event_name(evt), nport->display_name); WARN_ON(!node->els_req_cnt); node->els_req_cnt--; efc_sm_post_event(&nport->sm, EFC_EVT_SHUTDOWN, NULL); break; } default: __efc_fabric_common(__func__, ctx, evt, arg); } } void __efc_vport_fabric_init(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); switch (evt) { case EFC_EVT_ENTER: /* sm: / send FDISC */ efc_send_fdisc(node); efc_node_transition(node, __efc_fabric_fdisc_wait_rsp, NULL); break; default: __efc_fabric_common(__func__, ctx, evt, arg); } } void __efc_fabric_fdisc_wait_rsp(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node_cb *cbdata = arg; struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); switch (evt) { case EFC_EVT_SRRS_ELS_REQ_OK: { /* fc_id is in ext_status */ if (efc_node_check_els_req(ctx, evt, arg, ELS_FDISC, __efc_fabric_common, __func__)) { return; } WARN_ON(!node->els_req_cnt); node->els_req_cnt--; /* sm: / efc_nport_attach */ efc_nport_attach(node->nport, cbdata->ext_status); efc_node_transition(node, __efc_fabric_wait_domain_attach, NULL); break; } case EFC_EVT_SRRS_ELS_REQ_RJT: case EFC_EVT_SRRS_ELS_REQ_FAIL: { if (efc_node_check_els_req(ctx, evt, arg, ELS_FDISC, __efc_fabric_common, __func__)) { return; } WARN_ON(!node->els_req_cnt); node->els_req_cnt--; efc_log_err(node->efc, "FDISC failed, shutting down nport\n"); /* sm: / shutdown nport */ efc_sm_post_event(&node->nport->sm, EFC_EVT_SHUTDOWN, NULL); break; } default: __efc_fabric_common(__func__, ctx, evt, arg); } } static int efc_start_ns_node(struct efc_nport *nport) { struct efc_node *ns; /* Instantiate a name services node */ ns = efc_node_find(nport, FC_FID_DIR_SERV); if (!ns) { ns = efc_node_alloc(nport, FC_FID_DIR_SERV, false, false); if (!ns) return -EIO; } /* * for found ns, should we be transitioning from here? * breaks transition only * 1. from within state machine or * 2. if after alloc */ if (ns->efc->nodedb_mask & EFC_NODEDB_PAUSE_NAMESERVER) efc_node_pause(ns, __efc_ns_init); else efc_node_transition(ns, __efc_ns_init, NULL); return 0; } static int efc_start_fabctl_node(struct efc_nport *nport) { struct efc_node *fabctl; fabctl = efc_node_find(nport, FC_FID_FCTRL); if (!fabctl) { fabctl = efc_node_alloc(nport, FC_FID_FCTRL, false, false); if (!fabctl) return -EIO; } /* * for found ns, should we be transitioning from here? * breaks transition only * 1. from within state machine or * 2. if after alloc */ efc_node_transition(fabctl, __efc_fabctl_init, NULL); return 0; } void __efc_fabric_wait_domain_attach(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); switch (evt) { case EFC_EVT_ENTER: efc_node_hold_frames(node); break; case EFC_EVT_EXIT: efc_node_accept_frames(node); break; case EFC_EVT_DOMAIN_ATTACH_OK: case EFC_EVT_NPORT_ATTACH_OK: { int rc; rc = efc_start_ns_node(node->nport); if (rc) return; /* sm: if enable_ini / start fabctl node */ /* Instantiate the fabric controller (sends SCR) */ if (node->nport->enable_rscn) { rc = efc_start_fabctl_node(node->nport); if (rc) return; } efc_node_transition(node, __efc_fabric_idle, NULL); break; } default: __efc_fabric_common(__func__, ctx, evt, arg); } } void __efc_fabric_idle(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); switch (evt) { case EFC_EVT_DOMAIN_ATTACH_OK: break; default: __efc_fabric_common(__func__, ctx, evt, arg); } } void __efc_ns_init(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); switch (evt) { case EFC_EVT_ENTER: /* sm: / send PLOGI */ efc_send_plogi(node); efc_node_transition(node, __efc_ns_plogi_wait_rsp, NULL); break; default: __efc_fabric_common(__func__, ctx, evt, arg); } } void __efc_ns_plogi_wait_rsp(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node_cb *cbdata = arg; struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); switch (evt) { case EFC_EVT_SRRS_ELS_REQ_OK: { int rc; /* Save service parameters */ if (efc_node_check_els_req(ctx, evt, arg, ELS_PLOGI, __efc_fabric_common, __func__)) { return; } WARN_ON(!node->els_req_cnt); node->els_req_cnt--; /* sm: / save sparams, efc_node_attach */ efc_node_save_sparms(node, cbdata->els_rsp.virt); rc = efc_node_attach(node); efc_node_transition(node, __efc_ns_wait_node_attach, NULL); if (rc < 0) efc_node_post_event(node, EFC_EVT_NODE_ATTACH_FAIL, NULL); break; } default: __efc_fabric_common(__func__, ctx, evt, arg); } } void __efc_ns_wait_node_attach(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); switch (evt) { case EFC_EVT_ENTER: efc_node_hold_frames(node); break; case EFC_EVT_EXIT: efc_node_accept_frames(node); break; case EFC_EVT_NODE_ATTACH_OK: node->attached = true; /* sm: / send RFTID */ efc_ns_send_rftid(node); efc_node_transition(node, __efc_ns_rftid_wait_rsp, NULL); break; case EFC_EVT_NODE_ATTACH_FAIL: /* node attach failed, shutdown the node */ node->attached = false; node_printf(node, "Node attach failed\n"); node->shutdown_reason = EFC_NODE_SHUTDOWN_DEFAULT; efc_fabric_initiate_shutdown(node); break; case EFC_EVT_SHUTDOWN: node_printf(node, "Shutdown event received\n"); node->shutdown_reason = EFC_NODE_SHUTDOWN_DEFAULT; efc_node_transition(node, __efc_fabric_wait_attach_evt_shutdown, NULL); break; /* * if receive RSCN just ignore, * we haven't sent GID_PT yet (ACC sent by fabctl node) */ case EFC_EVT_RSCN_RCVD: break; default: __efc_fabric_common(__func__, ctx, evt, arg); } } void __efc_fabric_wait_attach_evt_shutdown(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); switch (evt) { case EFC_EVT_ENTER: efc_node_hold_frames(node); break; case EFC_EVT_EXIT: efc_node_accept_frames(node); break; /* wait for any of these attach events and then shutdown */ case EFC_EVT_NODE_ATTACH_OK: node->attached = true; node_printf(node, "Attach evt=%s, proceed to shutdown\n", efc_sm_event_name(evt)); efc_fabric_initiate_shutdown(node); break; case EFC_EVT_NODE_ATTACH_FAIL: node->attached = false; node_printf(node, "Attach evt=%s, proceed to shutdown\n", efc_sm_event_name(evt)); efc_fabric_initiate_shutdown(node); break; /* ignore shutdown event as we're already in shutdown path */ case EFC_EVT_SHUTDOWN: node_printf(node, "Shutdown event received\n"); break; default: __efc_fabric_common(__func__, ctx, evt, arg); } } void __efc_ns_rftid_wait_rsp(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); switch (evt) { case EFC_EVT_SRRS_ELS_REQ_OK: if (efc_node_check_ns_req(ctx, evt, arg, FC_NS_RFT_ID, __efc_fabric_common, __func__)) { return; } WARN_ON(!node->els_req_cnt); node->els_req_cnt--; /* sm: / send RFFID */ efc_ns_send_rffid(node); efc_node_transition(node, __efc_ns_rffid_wait_rsp, NULL); break; /* * if receive RSCN just ignore, * we haven't sent GID_PT yet (ACC sent by fabctl node) */ case EFC_EVT_RSCN_RCVD: break; default: __efc_fabric_common(__func__, ctx, evt, arg); } } void __efc_ns_rffid_wait_rsp(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); /* * Waits for an RFFID response event; * if rscn enabled, a GIDPT name services request is issued. */ switch (evt) { case EFC_EVT_SRRS_ELS_REQ_OK: { if (efc_node_check_ns_req(ctx, evt, arg, FC_NS_RFF_ID, __efc_fabric_common, __func__)) { return; } WARN_ON(!node->els_req_cnt); node->els_req_cnt--; if (node->nport->enable_rscn) { /* sm: if enable_rscn / send GIDPT */ efc_ns_send_gidpt(node); efc_node_transition(node, __efc_ns_gidpt_wait_rsp, NULL); } else { /* if 'T' only, we're done, go to idle */ efc_node_transition(node, __efc_ns_idle, NULL); } break; } /* * if receive RSCN just ignore, * we haven't sent GID_PT yet (ACC sent by fabctl node) */ case EFC_EVT_RSCN_RCVD: break; default: __efc_fabric_common(__func__, ctx, evt, arg); } } static int efc_process_gidpt_payload(struct efc_node *node, void *data, u32 gidpt_len) { u32 i, j; struct efc_node *newnode; struct efc_nport *nport = node->nport; struct efc *efc = node->efc; u32 port_id = 0, port_count, plist_count; struct efc_node *n; struct efc_node **active_nodes; int residual; struct { struct fc_ct_hdr hdr; struct fc_gid_pn_resp pn_rsp; } *rsp; struct fc_gid_pn_resp *gidpt; unsigned long index; rsp = data; gidpt = &rsp->pn_rsp; residual = be16_to_cpu(rsp->hdr.ct_mr_size); if (residual != 0) efc_log_debug(node->efc, "residual is %u words\n", residual); if (be16_to_cpu(rsp->hdr.ct_cmd) == FC_FS_RJT) { node_printf(node, "GIDPT request failed: rsn x%x rsn_expl x%x\n", rsp->hdr.ct_reason, rsp->hdr.ct_explan); return -EIO; } plist_count = (gidpt_len - sizeof(struct fc_ct_hdr)) / sizeof(*gidpt); /* Count the number of nodes */ port_count = 0; xa_for_each(&nport->lookup, index, n) { port_count++; } /* Allocate a buffer for all nodes */ active_nodes = kzalloc(port_count * sizeof(*active_nodes), GFP_ATOMIC); if (!active_nodes) { node_printf(node, "efc_malloc failed\n"); return -EIO; } /* Fill buffer with fc_id of active nodes */ i = 0; xa_for_each(&nport->lookup, index, n) { port_id = n->rnode.fc_id; switch (port_id) { case FC_FID_FLOGI: case FC_FID_FCTRL: case FC_FID_DIR_SERV: break; default: if (port_id != FC_FID_DOM_MGR) active_nodes[i++] = n; break; } } /* update the active nodes buffer */ for (i = 0; i < plist_count; i++) { hton24(gidpt[i].fp_fid, port_id); for (j = 0; j < port_count; j++) { if (active_nodes[j] && port_id == active_nodes[j]->rnode.fc_id) { active_nodes[j] = NULL; } } if (gidpt[i].fp_resvd & FC_NS_FID_LAST) break; } /* Those remaining in the active_nodes[] are now gone ! */ for (i = 0; i < port_count; i++) { /* * if we're an initiator and the remote node * is a target, then post the node missing event. * if we're target and we have enabled * target RSCN, then post the node missing event. */ if (!active_nodes[i]) continue; if ((node->nport->enable_ini && active_nodes[i]->targ) || (node->nport->enable_tgt && enable_target_rscn(efc))) { efc_node_post_event(active_nodes[i], EFC_EVT_NODE_MISSING, NULL); } else { node_printf(node, "GID_PT: skipping non-tgt port_id x%06x\n", active_nodes[i]->rnode.fc_id); } } kfree(active_nodes); for (i = 0; i < plist_count; i++) { hton24(gidpt[i].fp_fid, port_id); /* Don't create node for ourselves */ if (port_id == node->rnode.nport->fc_id) { if (gidpt[i].fp_resvd & FC_NS_FID_LAST) break; continue; } newnode = efc_node_find(nport, port_id); if (!newnode) { if (!node->nport->enable_ini) continue; newnode = efc_node_alloc(nport, port_id, false, false); if (!newnode) { efc_log_err(efc, "efc_node_alloc() failed\n"); return -EIO; } /* * send PLOGI automatically * if initiator */ efc_node_init_device(newnode, true); } if (node->nport->enable_ini && newnode->targ) { efc_node_post_event(newnode, EFC_EVT_NODE_REFOUND, NULL); } if (gidpt[i].fp_resvd & FC_NS_FID_LAST) break; } return 0; } void __efc_ns_gidpt_wait_rsp(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node_cb *cbdata = arg; struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); /* * Wait for a GIDPT response from the name server. Process the FC_IDs * that are reported by creating new remote ports, as needed. */ switch (evt) { case EFC_EVT_SRRS_ELS_REQ_OK: { if (efc_node_check_ns_req(ctx, evt, arg, FC_NS_GID_PT, __efc_fabric_common, __func__)) { return; } WARN_ON(!node->els_req_cnt); node->els_req_cnt--; /* sm: / process GIDPT payload */ efc_process_gidpt_payload(node, cbdata->els_rsp.virt, cbdata->els_rsp.len); efc_node_transition(node, __efc_ns_idle, NULL); break; } case EFC_EVT_SRRS_ELS_REQ_FAIL: { /* not much we can do; will retry with the next RSCN */ node_printf(node, "GID_PT failed to complete\n"); WARN_ON(!node->els_req_cnt); node->els_req_cnt--; efc_node_transition(node, __efc_ns_idle, NULL); break; } /* if receive RSCN here, queue up another discovery processing */ case EFC_EVT_RSCN_RCVD: { node_printf(node, "RSCN received during GID_PT processing\n"); node->rscn_pending = true; break; } default: __efc_fabric_common(__func__, ctx, evt, arg); } } void __efc_ns_idle(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node *node = ctx->app; struct efc *efc = node->efc; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); /* * Wait for RSCN received events (posted from the fabric controller) * and restart the GIDPT name services query and processing. */ switch (evt) { case EFC_EVT_ENTER: if (!node->rscn_pending) break; node_printf(node, "RSCN pending, restart discovery\n"); node->rscn_pending = false; fallthrough; case EFC_EVT_RSCN_RCVD: { /* sm: / send GIDPT */ /* * If target RSCN processing is enabled, * and this is target only (not initiator), * and tgt_rscn_delay is non-zero, * then we delay issuing the GID_PT */ if (efc->tgt_rscn_delay_msec != 0 && !node->nport->enable_ini && node->nport->enable_tgt && enable_target_rscn(efc)) { efc_node_transition(node, __efc_ns_gidpt_delay, NULL); } else { efc_ns_send_gidpt(node); efc_node_transition(node, __efc_ns_gidpt_wait_rsp, NULL); } break; } default: __efc_fabric_common(__func__, ctx, evt, arg); } } static void gidpt_delay_timer_cb(struct timer_list *t) { struct efc_node *node = from_timer(node, t, gidpt_delay_timer); del_timer(&node->gidpt_delay_timer); efc_node_post_event(node, EFC_EVT_GIDPT_DELAY_EXPIRED, NULL); } void __efc_ns_gidpt_delay(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node *node = ctx->app; struct efc *efc = node->efc; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); switch (evt) { case EFC_EVT_ENTER: { u64 delay_msec, tmp; /* * Compute the delay time. * Set to tgt_rscn_delay, if the time since last GIDPT * is less than tgt_rscn_period, then use tgt_rscn_period. */ delay_msec = efc->tgt_rscn_delay_msec; tmp = jiffies_to_msecs(jiffies) - node->time_last_gidpt_msec; if (tmp < efc->tgt_rscn_period_msec) delay_msec = efc->tgt_rscn_period_msec; timer_setup(&node->gidpt_delay_timer, &gidpt_delay_timer_cb, 0); mod_timer(&node->gidpt_delay_timer, jiffies + msecs_to_jiffies(delay_msec)); break; } case EFC_EVT_GIDPT_DELAY_EXPIRED: node->time_last_gidpt_msec = jiffies_to_msecs(jiffies); efc_ns_send_gidpt(node); efc_node_transition(node, __efc_ns_gidpt_wait_rsp, NULL); break; case EFC_EVT_RSCN_RCVD: { efc_log_debug(efc, "RSCN received while in GIDPT delay - no action\n"); break; } default: __efc_fabric_common(__func__, ctx, evt, arg); } } void __efc_fabctl_init(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node *node = ctx->app; node_sm_trace(); switch (evt) { case EFC_EVT_ENTER: /* no need to login to fabric controller, just send SCR */ efc_send_scr(node); efc_node_transition(node, __efc_fabctl_wait_scr_rsp, NULL); break; case EFC_EVT_NODE_ATTACH_OK: node->attached = true; break; default: __efc_fabric_common(__func__, ctx, evt, arg); } } void __efc_fabctl_wait_scr_rsp(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); /* * Fabric controller node state machine: * Wait for an SCR response from the fabric controller. */ switch (evt) { case EFC_EVT_SRRS_ELS_REQ_OK: if (efc_node_check_els_req(ctx, evt, arg, ELS_SCR, __efc_fabric_common, __func__)) { return; } WARN_ON(!node->els_req_cnt); node->els_req_cnt--; efc_node_transition(node, __efc_fabctl_ready, NULL); break; default: __efc_fabric_common(__func__, ctx, evt, arg); } } static void efc_process_rscn(struct efc_node *node, struct efc_node_cb *cbdata) { struct efc *efc = node->efc; struct efc_nport *nport = node->nport; struct efc_node *ns; /* Forward this event to the name-services node */ ns = efc_node_find(nport, FC_FID_DIR_SERV); if (ns) efc_node_post_event(ns, EFC_EVT_RSCN_RCVD, cbdata); else efc_log_warn(efc, "can't find name server node\n"); } void __efc_fabctl_ready(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node_cb *cbdata = arg; struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); /* * Fabric controller node state machine: Ready. * In this state, the fabric controller sends a RSCN, which is received * by this node and is forwarded to the name services node object; and * the RSCN LS_ACC is sent. */ switch (evt) { case EFC_EVT_RSCN_RCVD: { struct fc_frame_header *hdr = cbdata->header->dma.virt; /* * sm: / process RSCN (forward to name services node), * send LS_ACC */ efc_process_rscn(node, cbdata); efc_send_ls_acc(node, be16_to_cpu(hdr->fh_ox_id)); efc_node_transition(node, __efc_fabctl_wait_ls_acc_cmpl, NULL); break; } default: __efc_fabric_common(__func__, ctx, evt, arg); } } void __efc_fabctl_wait_ls_acc_cmpl(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); switch (evt) { case EFC_EVT_ENTER: efc_node_hold_frames(node); break; case EFC_EVT_EXIT: efc_node_accept_frames(node); break; case EFC_EVT_SRRS_ELS_CMPL_OK: WARN_ON(!node->els_cmpl_cnt); node->els_cmpl_cnt--; efc_node_transition(node, __efc_fabctl_ready, NULL); break; default: __efc_fabric_common(__func__, ctx, evt, arg); } } static uint64_t efc_get_wwpn(struct fc_els_flogi *sp) { return be64_to_cpu(sp->fl_wwnn); } static int efc_rnode_is_winner(struct efc_nport *nport) { struct fc_els_flogi *remote_sp; u64 remote_wwpn; u64 local_wwpn = nport->wwpn; u64 wwn_bump = 0; remote_sp = (struct fc_els_flogi *)nport->domain->flogi_service_params; remote_wwpn = efc_get_wwpn(remote_sp); local_wwpn ^= wwn_bump; efc_log_debug(nport->efc, "r: %llx\n", be64_to_cpu(remote_sp->fl_wwpn)); efc_log_debug(nport->efc, "l: %llx\n", local_wwpn); if (remote_wwpn == local_wwpn) { efc_log_warn(nport->efc, "WWPN of remote node [%08x %08x] matches local WWPN\n", (u32)(local_wwpn >> 32ll), (u32)local_wwpn); return -1; } return (remote_wwpn > local_wwpn); } void __efc_p2p_wait_domain_attach(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node *node = ctx->app; struct efc *efc = node->efc; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); switch (evt) { case EFC_EVT_ENTER: efc_node_hold_frames(node); break; case EFC_EVT_EXIT: efc_node_accept_frames(node); break; case EFC_EVT_DOMAIN_ATTACH_OK: { struct efc_nport *nport = node->nport; struct efc_node *rnode; /* * this transient node (SID=0 (recv'd FLOGI) * or DID=fabric (sent FLOGI)) * is the p2p winner, will use a separate node * to send PLOGI to peer */ WARN_ON(!node->nport->p2p_winner); rnode = efc_node_find(nport, node->nport->p2p_remote_port_id); if (rnode) { /* * the "other" transient p2p node has * already kicked off the * new node from which PLOGI is sent */ node_printf(node, "Node with fc_id x%x already exists\n", rnode->rnode.fc_id); } else { /* * create new node (SID=1, DID=2) * from which to send PLOGI */ rnode = efc_node_alloc(nport, nport->p2p_remote_port_id, false, false); if (!rnode) { efc_log_err(efc, "node alloc failed\n"); return; } efc_fabric_notify_topology(node); /* sm: / allocate p2p remote node */ efc_node_transition(rnode, __efc_p2p_rnode_init, NULL); } /* * the transient node (SID=0 or DID=fabric) * has served its purpose */ if (node->rnode.fc_id == 0) { /* * if this is the SID=0 node, * move to the init state in case peer * has restarted FLOGI discovery and FLOGI is pending */ /* don't send PLOGI on efc_d_init entry */ efc_node_init_device(node, false); } else { /* * if this is the DID=fabric node * (we initiated FLOGI), shut it down */ node->shutdown_reason = EFC_NODE_SHUTDOWN_DEFAULT; efc_fabric_initiate_shutdown(node); } break; } default: __efc_fabric_common(__func__, ctx, evt, arg); } } void __efc_p2p_rnode_init(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node_cb *cbdata = arg; struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); switch (evt) { case EFC_EVT_ENTER: /* sm: / send PLOGI */ efc_send_plogi(node); efc_node_transition(node, __efc_p2p_wait_plogi_rsp, NULL); break; case EFC_EVT_ABTS_RCVD: /* sm: send BA_ACC */ efc_send_bls_acc(node, cbdata->header->dma.virt); break; default: __efc_fabric_common(__func__, ctx, evt, arg); } } void __efc_p2p_wait_flogi_acc_cmpl(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node_cb *cbdata = arg; struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); switch (evt) { case EFC_EVT_ENTER: efc_node_hold_frames(node); break; case EFC_EVT_EXIT: efc_node_accept_frames(node); break; case EFC_EVT_SRRS_ELS_CMPL_OK: WARN_ON(!node->els_cmpl_cnt); node->els_cmpl_cnt--; /* sm: if p2p_winner / domain_attach */ if (node->nport->p2p_winner) { efc_node_transition(node, __efc_p2p_wait_domain_attach, NULL); if (!node->nport->domain->attached) { node_printf(node, "Domain not attached\n"); efc_domain_attach(node->nport->domain, node->nport->p2p_port_id); } else { node_printf(node, "Domain already attached\n"); efc_node_post_event(node, EFC_EVT_DOMAIN_ATTACH_OK, NULL); } } else { /* this node has served its purpose; * we'll expect a PLOGI on a separate * node (remote SID=0x1); return this node * to init state in case peer * restarts discovery -- it may already * have (pending frames may exist). */ /* don't send PLOGI on efc_d_init entry */ efc_node_init_device(node, false); } break; case EFC_EVT_SRRS_ELS_CMPL_FAIL: /* * LS_ACC failed, possibly due to link down; * shutdown node and wait * for FLOGI discovery to restart */ node_printf(node, "FLOGI LS_ACC failed, shutting down\n"); WARN_ON(!node->els_cmpl_cnt); node->els_cmpl_cnt--; node->shutdown_reason = EFC_NODE_SHUTDOWN_DEFAULT; efc_fabric_initiate_shutdown(node); break; case EFC_EVT_ABTS_RCVD: { /* sm: / send BA_ACC */ efc_send_bls_acc(node, cbdata->header->dma.virt); break; } default: __efc_fabric_common(__func__, ctx, evt, arg); } } void __efc_p2p_wait_plogi_rsp(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node_cb *cbdata = arg; struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); switch (evt) { case EFC_EVT_SRRS_ELS_REQ_OK: { int rc; if (efc_node_check_els_req(ctx, evt, arg, ELS_PLOGI, __efc_fabric_common, __func__)) { return; } WARN_ON(!node->els_req_cnt); node->els_req_cnt--; /* sm: / save sparams, efc_node_attach */ efc_node_save_sparms(node, cbdata->els_rsp.virt); rc = efc_node_attach(node); efc_node_transition(node, __efc_p2p_wait_node_attach, NULL); if (rc < 0) efc_node_post_event(node, EFC_EVT_NODE_ATTACH_FAIL, NULL); break; } case EFC_EVT_SRRS_ELS_REQ_FAIL: { if (efc_node_check_els_req(ctx, evt, arg, ELS_PLOGI, __efc_fabric_common, __func__)) { return; } node_printf(node, "PLOGI failed, shutting down\n"); WARN_ON(!node->els_req_cnt); node->els_req_cnt--; node->shutdown_reason = EFC_NODE_SHUTDOWN_DEFAULT; efc_fabric_initiate_shutdown(node); break; } case EFC_EVT_PLOGI_RCVD: { struct fc_frame_header *hdr = cbdata->header->dma.virt; /* if we're in external loopback mode, just send LS_ACC */ if (node->efc->external_loopback) { efc_send_plogi_acc(node, be16_to_cpu(hdr->fh_ox_id)); } else { /* * if this isn't external loopback, * pass to default handler */ __efc_fabric_common(__func__, ctx, evt, arg); } break; } case EFC_EVT_PRLI_RCVD: /* I, or I+T */ /* sent PLOGI and before completion was seen, received the * PRLI from the remote node (WCQEs and RCQEs come in on * different queues and order of processing cannot be assumed) * Save OXID so PRLI can be sent after the attach and continue * to wait for PLOGI response */ efc_process_prli_payload(node, cbdata->payload->dma.virt); efc_send_ls_acc_after_attach(node, cbdata->header->dma.virt, EFC_NODE_SEND_LS_ACC_PRLI); efc_node_transition(node, __efc_p2p_wait_plogi_rsp_recvd_prli, NULL); break; default: __efc_fabric_common(__func__, ctx, evt, arg); } } void __efc_p2p_wait_plogi_rsp_recvd_prli(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node_cb *cbdata = arg; struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); switch (evt) { case EFC_EVT_ENTER: /* * Since we've received a PRLI, we have a port login and will * just need to wait for the PLOGI response to do the node * attach and then we can send the LS_ACC for the PRLI. If, * during this time, we receive FCP_CMNDs (which is possible * since we've already sent a PRLI and our peer may have * accepted). * At this time, we are not waiting on any other unsolicited * frames to continue with the login process. Thus, it will not * hurt to hold frames here. */ efc_node_hold_frames(node); break; case EFC_EVT_EXIT: efc_node_accept_frames(node); break; case EFC_EVT_SRRS_ELS_REQ_OK: { /* PLOGI response received */ int rc; /* Completion from PLOGI sent */ if (efc_node_check_els_req(ctx, evt, arg, ELS_PLOGI, __efc_fabric_common, __func__)) { return; } WARN_ON(!node->els_req_cnt); node->els_req_cnt--; /* sm: / save sparams, efc_node_attach */ efc_node_save_sparms(node, cbdata->els_rsp.virt); rc = efc_node_attach(node); efc_node_transition(node, __efc_p2p_wait_node_attach, NULL); if (rc < 0) efc_node_post_event(node, EFC_EVT_NODE_ATTACH_FAIL, NULL); break; } case EFC_EVT_SRRS_ELS_REQ_FAIL: /* PLOGI response received */ case EFC_EVT_SRRS_ELS_REQ_RJT: /* PLOGI failed, shutdown the node */ if (efc_node_check_els_req(ctx, evt, arg, ELS_PLOGI, __efc_fabric_common, __func__)) { return; } WARN_ON(!node->els_req_cnt); node->els_req_cnt--; node->shutdown_reason = EFC_NODE_SHUTDOWN_DEFAULT; efc_fabric_initiate_shutdown(node); break; default: __efc_fabric_common(__func__, ctx, evt, arg); } } void __efc_p2p_wait_node_attach(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg) { struct efc_node_cb *cbdata = arg; struct efc_node *node = ctx->app; efc_node_evt_set(ctx, evt, __func__); node_sm_trace(); switch (evt) { case EFC_EVT_ENTER: efc_node_hold_frames(node); break; case EFC_EVT_EXIT: efc_node_accept_frames(node); break; case EFC_EVT_NODE_ATTACH_OK: node->attached = true; switch (node->send_ls_acc) { case EFC_NODE_SEND_LS_ACC_PRLI: { efc_d_send_prli_rsp(node->ls_acc_io, node->ls_acc_oxid); node->send_ls_acc = EFC_NODE_SEND_LS_ACC_NONE; node->ls_acc_io = NULL; break; } case EFC_NODE_SEND_LS_ACC_PLOGI: /* Can't happen in P2P */ case EFC_NODE_SEND_LS_ACC_NONE: default: /* Normal case for I */ /* sm: send_plogi_acc is not set / send PLOGI acc */ efc_node_transition(node, __efc_d_port_logged_in, NULL); break; } break; case EFC_EVT_NODE_ATTACH_FAIL: /* node attach failed, shutdown the node */ node->attached = false; node_printf(node, "Node attach failed\n"); node->shutdown_reason = EFC_NODE_SHUTDOWN_DEFAULT; efc_fabric_initiate_shutdown(node); break; case EFC_EVT_SHUTDOWN: node_printf(node, "%s received\n", efc_sm_event_name(evt)); node->shutdown_reason = EFC_NODE_SHUTDOWN_DEFAULT; efc_node_transition(node, __efc_fabric_wait_attach_evt_shutdown, NULL); break; case EFC_EVT_PRLI_RCVD: node_printf(node, "%s: PRLI received before node is attached\n", efc_sm_event_name(evt)); efc_process_prli_payload(node, cbdata->payload->dma.virt); efc_send_ls_acc_after_attach(node, cbdata->header->dma.virt, EFC_NODE_SEND_LS_ACC_PRLI); break; default: __efc_fabric_common(__func__, ctx, evt, arg); } } int efc_p2p_setup(struct efc_nport *nport) { struct efc *efc = nport->efc; int rnode_winner; rnode_winner = efc_rnode_is_winner(nport); /* set nport flags to indicate p2p "winner" */ if (rnode_winner == 1) { nport->p2p_remote_port_id = 0; nport->p2p_port_id = 0; nport->p2p_winner = false; } else if (rnode_winner == 0) { nport->p2p_remote_port_id = 2; nport->p2p_port_id = 1; nport->p2p_winner = true; } else { /* no winner; only okay if external loopback enabled */ if (nport->efc->external_loopback) { /* * External loopback mode enabled; * local nport and remote node * will be registered with an NPortID = 1; */ efc_log_debug(efc, "External loopback mode enabled\n"); nport->p2p_remote_port_id = 1; nport->p2p_port_id = 1; nport->p2p_winner = true; } else { efc_log_warn(efc, "failed to determine p2p winner\n"); return rnode_winner; } } return 0; }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1