Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
James Smart | 5064 | 99.80% | 1 | 50.00% |
Dmitry Bogdanov | 10 | 0.20% | 1 | 50.00% |
Total | 5074 | 2 |
// SPDX-License-Identifier: GPL-2.0 /* * Copyright (C) 2021 Broadcom. All Rights Reserved. The term * “Broadcom” refers to Broadcom Inc. and/or its subsidiaries. */ #include "efct_driver.h" #include "efct_hw.h" #define enable_tsend_auto_resp(efct) 1 #define enable_treceive_auto_resp(efct) 0 #define SCSI_IOFMT "[%04x][i:%04x t:%04x h:%04x]" #define scsi_io_printf(io, fmt, ...) \ efc_log_debug(io->efct, "[%s]" SCSI_IOFMT fmt, \ io->node->display_name, io->instance_index,\ io->init_task_tag, io->tgt_task_tag, io->hw_tag, ##__VA_ARGS__) #define EFCT_LOG_ENABLE_SCSI_TRACE(efct) \ (((efct) != NULL) ? (((efct)->logmask & (1U << 2)) != 0) : 0) #define scsi_io_trace(io, fmt, ...) \ do { \ if (EFCT_LOG_ENABLE_SCSI_TRACE(io->efct)) \ scsi_io_printf(io, fmt, ##__VA_ARGS__); \ } while (0) struct efct_io * efct_scsi_io_alloc(struct efct_node *node) { struct efct *efct; struct efct_xport *xport; struct efct_io *io; unsigned long flags; efct = node->efct; xport = efct->xport; io = efct_io_pool_io_alloc(efct->xport->io_pool); if (!io) { efc_log_err(efct, "IO alloc Failed\n"); atomic_add_return(1, &xport->io_alloc_failed_count); return NULL; } /* initialize refcount */ kref_init(&io->ref); io->release = _efct_scsi_io_free; /* set generic fields */ io->efct = efct; io->node = node; kref_get(&node->ref); /* set type and name */ io->io_type = EFCT_IO_TYPE_IO; io->display_name = "scsi_io"; io->cmd_ini = false; io->cmd_tgt = true; /* Add to node's active_ios list */ INIT_LIST_HEAD(&io->list_entry); spin_lock_irqsave(&node->active_ios_lock, flags); list_add(&io->list_entry, &node->active_ios); spin_unlock_irqrestore(&node->active_ios_lock, flags); return io; } void _efct_scsi_io_free(struct kref *arg) { struct efct_io *io = container_of(arg, struct efct_io, ref); struct efct *efct = io->efct; struct efct_node *node = io->node; unsigned long flags = 0; scsi_io_trace(io, "freeing io 0x%p %s\n", io, io->display_name); if (io->io_free) { efc_log_err(efct, "IO already freed.\n"); return; } spin_lock_irqsave(&node->active_ios_lock, flags); list_del_init(&io->list_entry); spin_unlock_irqrestore(&node->active_ios_lock, flags); kref_put(&node->ref, node->release); io->node = NULL; efct_io_pool_io_free(efct->xport->io_pool, io); } void efct_scsi_io_free(struct efct_io *io) { scsi_io_trace(io, "freeing io 0x%p %s\n", io, io->display_name); WARN_ON(!refcount_read(&io->ref.refcount)); kref_put(&io->ref, io->release); } static void efct_target_io_cb(struct efct_hw_io *hio, u32 length, int status, u32 ext_status, void *app) { u32 flags = 0; struct efct_io *io = app; struct efct *efct; enum efct_scsi_io_status scsi_stat = EFCT_SCSI_STATUS_GOOD; efct_scsi_io_cb_t cb; if (!io || !io->efct) { pr_err("%s: IO can not be NULL\n", __func__); return; } scsi_io_trace(io, "status x%x ext_status x%x\n", status, ext_status); efct = io->efct; io->transferred += length; if (!io->scsi_tgt_cb) { efct_scsi_check_pending(efct); return; } /* Call target server completion */ cb = io->scsi_tgt_cb; /* Clear the callback before invoking the callback */ io->scsi_tgt_cb = NULL; /* if status was good, and auto-good-response was set, * then callback target-server with IO_CMPL_RSP_SENT, * otherwise send IO_CMPL */ if (status == 0 && io->auto_resp) flags |= EFCT_SCSI_IO_CMPL_RSP_SENT; else flags |= EFCT_SCSI_IO_CMPL; switch (status) { case SLI4_FC_WCQE_STATUS_SUCCESS: scsi_stat = EFCT_SCSI_STATUS_GOOD; break; case SLI4_FC_WCQE_STATUS_DI_ERROR: if (ext_status & SLI4_FC_DI_ERROR_GE) scsi_stat = EFCT_SCSI_STATUS_DIF_GUARD_ERR; else if (ext_status & SLI4_FC_DI_ERROR_AE) scsi_stat = EFCT_SCSI_STATUS_DIF_APP_TAG_ERROR; else if (ext_status & SLI4_FC_DI_ERROR_RE) scsi_stat = EFCT_SCSI_STATUS_DIF_REF_TAG_ERROR; else scsi_stat = EFCT_SCSI_STATUS_DIF_UNKNOWN_ERROR; break; case SLI4_FC_WCQE_STATUS_LOCAL_REJECT: switch (ext_status) { case SLI4_FC_LOCAL_REJECT_INVALID_RELOFFSET: case SLI4_FC_LOCAL_REJECT_ABORT_REQUESTED: scsi_stat = EFCT_SCSI_STATUS_ABORTED; break; case SLI4_FC_LOCAL_REJECT_INVALID_RPI: scsi_stat = EFCT_SCSI_STATUS_NEXUS_LOST; break; case SLI4_FC_LOCAL_REJECT_NO_XRI: scsi_stat = EFCT_SCSI_STATUS_NO_IO; break; default: /*we have seen 0x0d(TX_DMA_FAILED err)*/ scsi_stat = EFCT_SCSI_STATUS_ERROR; break; } break; case SLI4_FC_WCQE_STATUS_TARGET_WQE_TIMEOUT: /* target IO timed out */ scsi_stat = EFCT_SCSI_STATUS_TIMEDOUT_AND_ABORTED; break; case SLI4_FC_WCQE_STATUS_SHUTDOWN: /* Target IO cancelled by HW */ scsi_stat = EFCT_SCSI_STATUS_SHUTDOWN; break; default: scsi_stat = EFCT_SCSI_STATUS_ERROR; break; } cb(io, scsi_stat, flags, io->scsi_tgt_cb_arg); efct_scsi_check_pending(efct); } static int efct_scsi_build_sgls(struct efct_hw *hw, struct efct_hw_io *hio, struct efct_scsi_sgl *sgl, u32 sgl_count, enum efct_hw_io_type type) { int rc; u32 i; struct efct *efct = hw->os; /* Initialize HW SGL */ rc = efct_hw_io_init_sges(hw, hio, type); if (rc) { efc_log_err(efct, "efct_hw_io_init_sges failed: %d\n", rc); return -EIO; } for (i = 0; i < sgl_count; i++) { /* Add data SGE */ rc = efct_hw_io_add_sge(hw, hio, sgl[i].addr, sgl[i].len); if (rc) { efc_log_err(efct, "add sge failed cnt=%d rc=%d\n", sgl_count, rc); return rc; } } return 0; } static void efc_log_sgl(struct efct_io *io) { struct efct_hw_io *hio = io->hio; struct sli4_sge *data = NULL; u32 *dword = NULL; u32 i; u32 n_sge; scsi_io_trace(io, "def_sgl at 0x%x 0x%08x\n", upper_32_bits(hio->def_sgl.phys), lower_32_bits(hio->def_sgl.phys)); n_sge = (hio->sgl == &hio->def_sgl) ? hio->n_sge : hio->def_sgl_count; for (i = 0, data = hio->def_sgl.virt; i < n_sge; i++, data++) { dword = (u32 *)data; scsi_io_trace(io, "SGL %2d 0x%08x 0x%08x 0x%08x 0x%08x\n", i, dword[0], dword[1], dword[2], dword[3]); if (dword[2] & (1U << 31)) break; } } static void efct_scsi_check_pending_async_cb(struct efct_hw *hw, int status, u8 *mqe, void *arg) { struct efct_io *io = arg; if (io) { efct_hw_done_t cb = io->hw_cb; if (!io->hw_cb) return; io->hw_cb = NULL; (cb)(io->hio, 0, SLI4_FC_WCQE_STATUS_DISPATCH_ERROR, 0, io); } } static int efct_scsi_io_dispatch_hw_io(struct efct_io *io, struct efct_hw_io *hio) { int rc = 0; struct efct *efct = io->efct; /* Got a HW IO; * update ini/tgt_task_tag with HW IO info and dispatch */ io->hio = hio; if (io->cmd_tgt) io->tgt_task_tag = hio->indicator; else if (io->cmd_ini) io->init_task_tag = hio->indicator; io->hw_tag = hio->reqtag; hio->eq = io->hw_priv; /* Copy WQ steering */ switch (io->wq_steering) { case EFCT_SCSI_WQ_STEERING_CLASS >> EFCT_SCSI_WQ_STEERING_SHIFT: hio->wq_steering = EFCT_HW_WQ_STEERING_CLASS; break; case EFCT_SCSI_WQ_STEERING_REQUEST >> EFCT_SCSI_WQ_STEERING_SHIFT: hio->wq_steering = EFCT_HW_WQ_STEERING_REQUEST; break; case EFCT_SCSI_WQ_STEERING_CPU >> EFCT_SCSI_WQ_STEERING_SHIFT: hio->wq_steering = EFCT_HW_WQ_STEERING_CPU; break; } switch (io->io_type) { case EFCT_IO_TYPE_IO: rc = efct_scsi_build_sgls(&efct->hw, io->hio, io->sgl, io->sgl_count, io->hio_type); if (rc) break; if (EFCT_LOG_ENABLE_SCSI_TRACE(efct)) efc_log_sgl(io); if (io->app_id) io->iparam.fcp_tgt.app_id = io->app_id; io->iparam.fcp_tgt.vpi = io->node->vpi; io->iparam.fcp_tgt.rpi = io->node->rpi; io->iparam.fcp_tgt.s_id = io->node->port_fc_id; io->iparam.fcp_tgt.d_id = io->node->node_fc_id; io->iparam.fcp_tgt.xmit_len = io->wire_len; rc = efct_hw_io_send(&io->efct->hw, io->hio_type, io->hio, &io->iparam, io->hw_cb, io); break; default: scsi_io_printf(io, "Unknown IO type=%d\n", io->io_type); rc = -EIO; break; } return rc; } static int efct_scsi_io_dispatch_no_hw_io(struct efct_io *io) { int rc; switch (io->io_type) { case EFCT_IO_TYPE_ABORT: { struct efct_hw_io *hio_to_abort = NULL; hio_to_abort = io->io_to_abort->hio; if (!hio_to_abort) { /* * If "IO to abort" does not have an * associated HW IO, immediately make callback with * success. The command must have been sent to * the backend, but the data phase has not yet * started, so we don't have a HW IO. * * Note: since the backend shims should be * taking a reference on io_to_abort, it should not * be possible to have been completed and freed by * the backend before the abort got here. */ scsi_io_printf(io, "IO: not active\n"); ((efct_hw_done_t)io->hw_cb)(io->hio, 0, SLI4_FC_WCQE_STATUS_SUCCESS, 0, io); rc = 0; break; } /* HW IO is valid, abort it */ scsi_io_printf(io, "aborting\n"); rc = efct_hw_io_abort(&io->efct->hw, hio_to_abort, io->send_abts, io->hw_cb, io); if (rc) { int status = SLI4_FC_WCQE_STATUS_SUCCESS; efct_hw_done_t cb = io->hw_cb; if (rc != -ENOENT && rc != -EINPROGRESS) { status = -1; scsi_io_printf(io, "Failed to abort IO rc=%d\n", rc); } cb(io->hio, 0, status, 0, io); rc = 0; } break; } default: scsi_io_printf(io, "Unknown IO type=%d\n", io->io_type); rc = -EIO; break; } return rc; } static struct efct_io * efct_scsi_dispatch_pending(struct efct *efct) { struct efct_xport *xport = efct->xport; struct efct_io *io = NULL; struct efct_hw_io *hio; unsigned long flags = 0; int status; spin_lock_irqsave(&xport->io_pending_lock, flags); if (!list_empty(&xport->io_pending_list)) { io = list_first_entry(&xport->io_pending_list, struct efct_io, io_pending_link); list_del_init(&io->io_pending_link); } if (!io) { spin_unlock_irqrestore(&xport->io_pending_lock, flags); return NULL; } if (io->io_type == EFCT_IO_TYPE_ABORT) { hio = NULL; } else { hio = efct_hw_io_alloc(&efct->hw); if (!hio) { /* * No HW IO available.Put IO back on * the front of pending list */ list_add(&xport->io_pending_list, &io->io_pending_link); io = NULL; } else { hio->eq = io->hw_priv; } } /* Must drop the lock before dispatching the IO */ spin_unlock_irqrestore(&xport->io_pending_lock, flags); if (!io) return NULL; /* * We pulled an IO off the pending list, * and either got an HW IO or don't need one */ atomic_sub_return(1, &xport->io_pending_count); if (!hio) status = efct_scsi_io_dispatch_no_hw_io(io); else status = efct_scsi_io_dispatch_hw_io(io, hio); if (status) { /* * Invoke the HW callback, but do so in the * separate execution context,provided by the * NOP mailbox completion processing context * by using efct_hw_async_call() */ if (efct_hw_async_call(&efct->hw, efct_scsi_check_pending_async_cb, io)) { efc_log_debug(efct, "call hw async failed\n"); } } return io; } void efct_scsi_check_pending(struct efct *efct) { struct efct_xport *xport = efct->xport; struct efct_io *io = NULL; int count = 0; unsigned long flags = 0; int dispatch = 0; /* Guard against recursion */ if (atomic_add_return(1, &xport->io_pending_recursing)) { /* This function is already running. Decrement and return. */ atomic_sub_return(1, &xport->io_pending_recursing); return; } while (efct_scsi_dispatch_pending(efct)) count++; if (count) { atomic_sub_return(1, &xport->io_pending_recursing); return; } /* * If nothing was removed from the list, * we might be in a case where we need to abort an * active IO and the abort is on the pending list. * Look for an abort we can dispatch. */ spin_lock_irqsave(&xport->io_pending_lock, flags); list_for_each_entry(io, &xport->io_pending_list, io_pending_link) { if (io->io_type == EFCT_IO_TYPE_ABORT && io->io_to_abort->hio) { /* This IO has a HW IO, so it is * active. Dispatch the abort. */ dispatch = 1; list_del_init(&io->io_pending_link); atomic_sub_return(1, &xport->io_pending_count); break; } } spin_unlock_irqrestore(&xport->io_pending_lock, flags); if (dispatch) { if (efct_scsi_io_dispatch_no_hw_io(io)) { if (efct_hw_async_call(&efct->hw, efct_scsi_check_pending_async_cb, io)) { efc_log_debug(efct, "hw async failed\n"); } } } atomic_sub_return(1, &xport->io_pending_recursing); } int efct_scsi_io_dispatch(struct efct_io *io, void *cb) { struct efct_hw_io *hio; struct efct *efct = io->efct; struct efct_xport *xport = efct->xport; unsigned long flags = 0; io->hw_cb = cb; /* * if this IO already has a HW IO, then this is either * not the first phase of the IO. Send it to the HW. */ if (io->hio) return efct_scsi_io_dispatch_hw_io(io, io->hio); /* * We don't already have a HW IO associated with the IO. First check * the pending list. If not empty, add IO to the tail and process the * pending list. */ spin_lock_irqsave(&xport->io_pending_lock, flags); if (!list_empty(&xport->io_pending_list)) { /* * If this is a low latency request, * the put at the front of the IO pending * queue, otherwise put it at the end of the queue. */ if (io->low_latency) { INIT_LIST_HEAD(&io->io_pending_link); list_add(&xport->io_pending_list, &io->io_pending_link); } else { INIT_LIST_HEAD(&io->io_pending_link); list_add_tail(&io->io_pending_link, &xport->io_pending_list); } spin_unlock_irqrestore(&xport->io_pending_lock, flags); atomic_add_return(1, &xport->io_pending_count); atomic_add_return(1, &xport->io_total_pending); /* process pending list */ efct_scsi_check_pending(efct); return 0; } spin_unlock_irqrestore(&xport->io_pending_lock, flags); /* * We don't have a HW IO associated with the IO and there's nothing * on the pending list. Attempt to allocate a HW IO and dispatch it. */ hio = efct_hw_io_alloc(&io->efct->hw); if (!hio) { /* Couldn't get a HW IO. Save this IO on the pending list */ spin_lock_irqsave(&xport->io_pending_lock, flags); INIT_LIST_HEAD(&io->io_pending_link); list_add_tail(&io->io_pending_link, &xport->io_pending_list); spin_unlock_irqrestore(&xport->io_pending_lock, flags); atomic_add_return(1, &xport->io_total_pending); atomic_add_return(1, &xport->io_pending_count); return 0; } /* We successfully allocated a HW IO; dispatch to HW */ return efct_scsi_io_dispatch_hw_io(io, hio); } int efct_scsi_io_dispatch_abort(struct efct_io *io, void *cb) { struct efct *efct = io->efct; struct efct_xport *xport = efct->xport; unsigned long flags = 0; io->hw_cb = cb; /* * For aborts, we don't need a HW IO, but we still want * to pass through the pending list to preserve ordering. * Thus, if the pending list is not empty, add this abort * to the pending list and process the pending list. */ spin_lock_irqsave(&xport->io_pending_lock, flags); if (!list_empty(&xport->io_pending_list)) { INIT_LIST_HEAD(&io->io_pending_link); list_add_tail(&io->io_pending_link, &xport->io_pending_list); spin_unlock_irqrestore(&xport->io_pending_lock, flags); atomic_add_return(1, &xport->io_pending_count); atomic_add_return(1, &xport->io_total_pending); /* process pending list */ efct_scsi_check_pending(efct); return 0; } spin_unlock_irqrestore(&xport->io_pending_lock, flags); /* nothing on pending list, dispatch abort */ return efct_scsi_io_dispatch_no_hw_io(io); } static inline int efct_scsi_xfer_data(struct efct_io *io, u32 flags, struct efct_scsi_sgl *sgl, u32 sgl_count, u64 xwire_len, enum efct_hw_io_type type, int enable_ar, efct_scsi_io_cb_t cb, void *arg) { struct efct *efct; size_t residual = 0; io->sgl_count = sgl_count; efct = io->efct; scsi_io_trace(io, "%s wire_len %llu\n", (type == EFCT_HW_IO_TARGET_READ) ? "send" : "recv", xwire_len); io->hio_type = type; io->scsi_tgt_cb = cb; io->scsi_tgt_cb_arg = arg; residual = io->exp_xfer_len - io->transferred; io->wire_len = (xwire_len < residual) ? xwire_len : residual; residual = (xwire_len - io->wire_len); memset(&io->iparam, 0, sizeof(io->iparam)); io->iparam.fcp_tgt.ox_id = io->init_task_tag; io->iparam.fcp_tgt.offset = io->transferred; io->iparam.fcp_tgt.cs_ctl = io->cs_ctl; io->iparam.fcp_tgt.timeout = io->timeout; /* if this is the last data phase and there is no residual, enable * auto-good-response */ if (enable_ar && (flags & EFCT_SCSI_LAST_DATAPHASE) && residual == 0 && ((io->transferred + io->wire_len) == io->exp_xfer_len) && (!(flags & EFCT_SCSI_NO_AUTO_RESPONSE))) { io->iparam.fcp_tgt.flags |= SLI4_IO_AUTO_GOOD_RESPONSE; io->auto_resp = true; } else { io->auto_resp = false; } /* save this transfer length */ io->xfer_req = io->wire_len; /* Adjust the transferred count to account for overrun * when the residual is calculated in efct_scsi_send_resp */ io->transferred += residual; /* Adjust the SGL size if there is overrun */ if (residual) { struct efct_scsi_sgl *sgl_ptr = &io->sgl[sgl_count - 1]; while (residual) { size_t len = sgl_ptr->len; if (len > residual) { sgl_ptr->len = len - residual; residual = 0; } else { sgl_ptr->len = 0; residual -= len; io->sgl_count--; } sgl_ptr--; } } /* Set latency and WQ steering */ io->low_latency = (flags & EFCT_SCSI_LOW_LATENCY) != 0; io->wq_steering = (flags & EFCT_SCSI_WQ_STEERING_MASK) >> EFCT_SCSI_WQ_STEERING_SHIFT; io->wq_class = (flags & EFCT_SCSI_WQ_CLASS_MASK) >> EFCT_SCSI_WQ_CLASS_SHIFT; if (efct->xport) { struct efct_xport *xport = efct->xport; if (type == EFCT_HW_IO_TARGET_READ) { xport->fcp_stats.input_requests++; xport->fcp_stats.input_bytes += xwire_len; } else if (type == EFCT_HW_IO_TARGET_WRITE) { xport->fcp_stats.output_requests++; xport->fcp_stats.output_bytes += xwire_len; } } return efct_scsi_io_dispatch(io, efct_target_io_cb); } int efct_scsi_send_rd_data(struct efct_io *io, u32 flags, struct efct_scsi_sgl *sgl, u32 sgl_count, u64 len, efct_scsi_io_cb_t cb, void *arg) { return efct_scsi_xfer_data(io, flags, sgl, sgl_count, len, EFCT_HW_IO_TARGET_READ, enable_tsend_auto_resp(io->efct), cb, arg); } int efct_scsi_recv_wr_data(struct efct_io *io, u32 flags, struct efct_scsi_sgl *sgl, u32 sgl_count, u64 len, efct_scsi_io_cb_t cb, void *arg) { return efct_scsi_xfer_data(io, flags, sgl, sgl_count, len, EFCT_HW_IO_TARGET_WRITE, enable_treceive_auto_resp(io->efct), cb, arg); } int efct_scsi_send_resp(struct efct_io *io, u32 flags, struct efct_scsi_cmd_resp *rsp, efct_scsi_io_cb_t cb, void *arg) { struct efct *efct; int residual; /* Always try auto resp */ bool auto_resp = true; u8 scsi_status = 0; u16 scsi_status_qualifier = 0; u8 *sense_data = NULL; u32 sense_data_length = 0; efct = io->efct; if (rsp) { scsi_status = rsp->scsi_status; scsi_status_qualifier = rsp->scsi_status_qualifier; sense_data = rsp->sense_data; sense_data_length = rsp->sense_data_length; residual = rsp->residual; } else { residual = io->exp_xfer_len - io->transferred; } io->wire_len = 0; io->hio_type = EFCT_HW_IO_TARGET_RSP; io->scsi_tgt_cb = cb; io->scsi_tgt_cb_arg = arg; memset(&io->iparam, 0, sizeof(io->iparam)); io->iparam.fcp_tgt.ox_id = io->init_task_tag; io->iparam.fcp_tgt.offset = 0; io->iparam.fcp_tgt.cs_ctl = io->cs_ctl; io->iparam.fcp_tgt.timeout = io->timeout; /* Set low latency queueing request */ io->low_latency = (flags & EFCT_SCSI_LOW_LATENCY) != 0; io->wq_steering = (flags & EFCT_SCSI_WQ_STEERING_MASK) >> EFCT_SCSI_WQ_STEERING_SHIFT; io->wq_class = (flags & EFCT_SCSI_WQ_CLASS_MASK) >> EFCT_SCSI_WQ_CLASS_SHIFT; if (scsi_status != 0 || residual || sense_data_length) { struct fcp_resp_with_ext *fcprsp = io->rspbuf.virt; u8 *sns_data; if (!fcprsp) { efc_log_err(efct, "NULL response buffer\n"); return -EIO; } sns_data = (u8 *)io->rspbuf.virt + sizeof(*fcprsp); auto_resp = false; memset(fcprsp, 0, sizeof(*fcprsp)); io->wire_len += sizeof(*fcprsp); fcprsp->resp.fr_status = scsi_status; fcprsp->resp.fr_retry_delay = cpu_to_be16(scsi_status_qualifier); /* set residual status if necessary */ if (residual != 0) { /* FCP: if data transferred is less than the * amount expected, then this is an underflow. * If data transferred would have been greater * than the amount expected this is an overflow */ if (residual > 0) { fcprsp->resp.fr_flags |= FCP_RESID_UNDER; fcprsp->ext.fr_resid = cpu_to_be32(residual); } else { fcprsp->resp.fr_flags |= FCP_RESID_OVER; fcprsp->ext.fr_resid = cpu_to_be32(-residual); } } if (EFCT_SCSI_SNS_BUF_VALID(sense_data) && sense_data_length) { if (sense_data_length > SCSI_SENSE_BUFFERSIZE) { efc_log_err(efct, "Sense exceeds max size.\n"); return -EIO; } fcprsp->resp.fr_flags |= FCP_SNS_LEN_VAL; memcpy(sns_data, sense_data, sense_data_length); fcprsp->ext.fr_sns_len = cpu_to_be32(sense_data_length); io->wire_len += sense_data_length; } io->sgl[0].addr = io->rspbuf.phys; io->sgl[0].dif_addr = 0; io->sgl[0].len = io->wire_len; io->sgl_count = 1; } if (auto_resp) io->iparam.fcp_tgt.flags |= SLI4_IO_AUTO_GOOD_RESPONSE; return efct_scsi_io_dispatch(io, efct_target_io_cb); } static int efct_target_bls_resp_cb(struct efct_hw_io *hio, u32 length, int status, u32 ext_status, void *app) { struct efct_io *io = app; struct efct *efct; enum efct_scsi_io_status bls_status; efct = io->efct; /* BLS isn't really a "SCSI" concept, but use SCSI status */ if (status) { io_error_log(io, "s=%#x x=%#x\n", status, ext_status); bls_status = EFCT_SCSI_STATUS_ERROR; } else { bls_status = EFCT_SCSI_STATUS_GOOD; } if (io->bls_cb) { efct_scsi_io_cb_t bls_cb = io->bls_cb; void *bls_cb_arg = io->bls_cb_arg; io->bls_cb = NULL; io->bls_cb_arg = NULL; /* invoke callback */ bls_cb(io, bls_status, 0, bls_cb_arg); } efct_scsi_check_pending(efct); return 0; } static int efct_target_send_bls_resp(struct efct_io *io, efct_scsi_io_cb_t cb, void *arg) { struct efct_node *node = io->node; struct sli_bls_params *bls = &io->iparam.bls; struct efct *efct = node->efct; struct fc_ba_acc *acc; int rc; /* fill out IO structure with everything needed to send BA_ACC */ memset(&io->iparam, 0, sizeof(io->iparam)); bls->ox_id = io->init_task_tag; bls->rx_id = io->abort_rx_id; bls->vpi = io->node->vpi; bls->rpi = io->node->rpi; bls->s_id = U32_MAX; bls->d_id = io->node->node_fc_id; bls->rpi_registered = true; acc = (void *)bls->payload; acc->ba_ox_id = cpu_to_be16(bls->ox_id); acc->ba_rx_id = cpu_to_be16(bls->rx_id); acc->ba_high_seq_cnt = cpu_to_be16(U16_MAX); /* generic io fields have already been populated */ /* set type and BLS-specific fields */ io->io_type = EFCT_IO_TYPE_BLS_RESP; io->display_name = "bls_rsp"; io->hio_type = EFCT_HW_BLS_ACC; io->bls_cb = cb; io->bls_cb_arg = arg; /* dispatch IO */ rc = efct_hw_bls_send(efct, FC_RCTL_BA_ACC, bls, efct_target_bls_resp_cb, io); return rc; } static int efct_bls_send_rjt_cb(struct efct_hw_io *hio, u32 length, int status, u32 ext_status, void *app) { struct efct_io *io = app; efct_scsi_io_free(io); return 0; } struct efct_io * efct_bls_send_rjt(struct efct_io *io, struct fc_frame_header *hdr) { struct efct_node *node = io->node; struct sli_bls_params *bls = &io->iparam.bls; struct efct *efct = node->efct; struct fc_ba_rjt *acc; int rc; /* fill out BLS Response-specific fields */ io->io_type = EFCT_IO_TYPE_BLS_RESP; io->display_name = "ba_rjt"; io->hio_type = EFCT_HW_BLS_RJT; io->init_task_tag = be16_to_cpu(hdr->fh_ox_id); /* fill out iparam fields */ memset(&io->iparam, 0, sizeof(io->iparam)); bls->ox_id = be16_to_cpu(hdr->fh_ox_id); bls->rx_id = be16_to_cpu(hdr->fh_rx_id); bls->vpi = io->node->vpi; bls->rpi = io->node->rpi; bls->s_id = U32_MAX; bls->d_id = io->node->node_fc_id; bls->rpi_registered = true; acc = (void *)bls->payload; acc->br_reason = ELS_RJT_UNAB; acc->br_explan = ELS_EXPL_NONE; rc = efct_hw_bls_send(efct, FC_RCTL_BA_RJT, bls, efct_bls_send_rjt_cb, io); if (rc) { efc_log_err(efct, "efct_scsi_io_dispatch() failed: %d\n", rc); efct_scsi_io_free(io); io = NULL; } return io; } int efct_scsi_send_tmf_resp(struct efct_io *io, enum efct_scsi_tmf_resp rspcode, u8 addl_rsp_info[3], efct_scsi_io_cb_t cb, void *arg) { int rc; struct { struct fcp_resp_with_ext rsp_ext; struct fcp_resp_rsp_info info; } *fcprsp; u8 fcp_rspcode; io->wire_len = 0; switch (rspcode) { case EFCT_SCSI_TMF_FUNCTION_COMPLETE: fcp_rspcode = FCP_TMF_CMPL; break; case EFCT_SCSI_TMF_FUNCTION_SUCCEEDED: case EFCT_SCSI_TMF_FUNCTION_IO_NOT_FOUND: fcp_rspcode = FCP_TMF_CMPL; break; case EFCT_SCSI_TMF_FUNCTION_REJECTED: fcp_rspcode = FCP_TMF_REJECTED; break; case EFCT_SCSI_TMF_INCORRECT_LOGICAL_UNIT_NUMBER: fcp_rspcode = FCP_TMF_INVALID_LUN; break; case EFCT_SCSI_TMF_SERVICE_DELIVERY: fcp_rspcode = FCP_TMF_FAILED; break; default: fcp_rspcode = FCP_TMF_REJECTED; break; } io->hio_type = EFCT_HW_IO_TARGET_RSP; io->scsi_tgt_cb = cb; io->scsi_tgt_cb_arg = arg; if (io->tmf_cmd == EFCT_SCSI_TMF_ABORT_TASK) { rc = efct_target_send_bls_resp(io, cb, arg); return rc; } /* populate the FCP TMF response */ fcprsp = io->rspbuf.virt; memset(fcprsp, 0, sizeof(*fcprsp)); fcprsp->rsp_ext.resp.fr_flags |= FCP_SNS_LEN_VAL; if (addl_rsp_info) { memcpy(fcprsp->info._fr_resvd, addl_rsp_info, sizeof(fcprsp->info._fr_resvd)); } fcprsp->info.rsp_code = fcp_rspcode; io->wire_len = sizeof(*fcprsp); fcprsp->rsp_ext.ext.fr_rsp_len = cpu_to_be32(sizeof(struct fcp_resp_rsp_info)); io->sgl[0].addr = io->rspbuf.phys; io->sgl[0].dif_addr = 0; io->sgl[0].len = io->wire_len; io->sgl_count = 1; memset(&io->iparam, 0, sizeof(io->iparam)); io->iparam.fcp_tgt.ox_id = io->init_task_tag; io->iparam.fcp_tgt.offset = 0; io->iparam.fcp_tgt.cs_ctl = io->cs_ctl; io->iparam.fcp_tgt.timeout = io->timeout; rc = efct_scsi_io_dispatch(io, efct_target_io_cb); return rc; } static int efct_target_abort_cb(struct efct_hw_io *hio, u32 length, int status, u32 ext_status, void *app) { struct efct_io *io = app; struct efct *efct; enum efct_scsi_io_status scsi_status; efct_scsi_io_cb_t abort_cb; void *abort_cb_arg; efct = io->efct; if (!io->abort_cb) goto done; abort_cb = io->abort_cb; abort_cb_arg = io->abort_cb_arg; io->abort_cb = NULL; io->abort_cb_arg = NULL; switch (status) { case SLI4_FC_WCQE_STATUS_SUCCESS: scsi_status = EFCT_SCSI_STATUS_GOOD; break; case SLI4_FC_WCQE_STATUS_LOCAL_REJECT: switch (ext_status) { case SLI4_FC_LOCAL_REJECT_NO_XRI: scsi_status = EFCT_SCSI_STATUS_NO_IO; break; case SLI4_FC_LOCAL_REJECT_ABORT_IN_PROGRESS: scsi_status = EFCT_SCSI_STATUS_ABORT_IN_PROGRESS; break; default: /*we have seen 0x15 (abort in progress)*/ scsi_status = EFCT_SCSI_STATUS_ERROR; break; } break; case SLI4_FC_WCQE_STATUS_FCP_RSP_FAILURE: scsi_status = EFCT_SCSI_STATUS_CHECK_RESPONSE; break; default: scsi_status = EFCT_SCSI_STATUS_ERROR; break; } /* invoke callback */ abort_cb(io->io_to_abort, scsi_status, 0, abort_cb_arg); done: /* done with IO to abort,efct_ref_get(): efct_scsi_tgt_abort_io() */ kref_put(&io->io_to_abort->ref, io->io_to_abort->release); efct_io_pool_io_free(efct->xport->io_pool, io); efct_scsi_check_pending(efct); return 0; } int efct_scsi_tgt_abort_io(struct efct_io *io, efct_scsi_io_cb_t cb, void *arg) { struct efct *efct; struct efct_xport *xport; int rc; struct efct_io *abort_io = NULL; efct = io->efct; xport = efct->xport; /* take a reference on IO being aborted */ if (kref_get_unless_zero(&io->ref) == 0) { /* command no longer active */ scsi_io_printf(io, "command no longer active\n"); return -EIO; } /* * allocate a new IO to send the abort request. Use efct_io_alloc() * directly, as we need an IO object that will not fail allocation * due to allocations being disabled (in efct_scsi_io_alloc()) */ abort_io = efct_io_pool_io_alloc(efct->xport->io_pool); if (!abort_io) { atomic_add_return(1, &xport->io_alloc_failed_count); kref_put(&io->ref, io->release); return -EIO; } /* Save the target server callback and argument */ /* set generic fields */ abort_io->cmd_tgt = true; abort_io->node = io->node; /* set type and abort-specific fields */ abort_io->io_type = EFCT_IO_TYPE_ABORT; abort_io->display_name = "tgt_abort"; abort_io->io_to_abort = io; abort_io->send_abts = false; abort_io->abort_cb = cb; abort_io->abort_cb_arg = arg; /* now dispatch IO */ rc = efct_scsi_io_dispatch_abort(abort_io, efct_target_abort_cb); if (rc) kref_put(&io->ref, io->release); return rc; } void efct_scsi_io_complete(struct efct_io *io) { if (io->io_free) { efc_log_debug(io->efct, "completion for non-busy io tag 0x%x\n", io->tag); return; } scsi_io_trace(io, "freeing io 0x%p %s\n", io, io->display_name); kref_put(&io->ref, io->release); }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1