Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Jing Huang | 4799 | 80.30% | 12 | 29.27% |
Krishna Gudipati | 872 | 14.59% | 9 | 21.95% |
Anil Gurumurthy | 76 | 1.27% | 2 | 4.88% |
Maggie Zhang | 56 | 0.94% | 3 | 7.32% |
Hannes Reinecke | 41 | 0.69% | 2 | 4.88% |
Jakob Normark | 22 | 0.37% | 1 | 2.44% |
Johannes Thumshirn | 21 | 0.35% | 1 | 2.44% |
Jeff Garzik | 19 | 0.32% | 1 | 2.44% |
Arnd Bergmann | 18 | 0.30% | 1 | 2.44% |
Tejun Heo | 15 | 0.25% | 1 | 2.44% |
Dan Carpenter | 13 | 0.22% | 2 | 4.88% |
Christoph Hellwig | 12 | 0.20% | 2 | 4.88% |
Alexey Khoroshilov | 6 | 0.10% | 1 | 2.44% |
Paul Gortmaker | 3 | 0.05% | 1 | 2.44% |
Yong Zhang | 2 | 0.03% | 1 | 2.44% |
Joe Perches | 1 | 0.02% | 1 | 2.44% |
Total | 5976 | 41 |
/* * Copyright (c) 2005-2014 Brocade Communications Systems, Inc. * Copyright (c) 2014- QLogic Corporation. * All rights reserved * www.qlogic.com * * Linux driver for QLogic BR-series Fibre Channel Host Bus Adapter. * * This program is free software; you can redistribute it and/or modify it * under the terms of the GNU General Public License (GPL) Version 2 as * published by the Free Software Foundation * * This program is distributed in the hope that it will be useful, but * WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * General Public License for more details. */ /* * bfad_im.c Linux driver IM module. */ #include <linux/export.h> #include "bfad_drv.h" #include "bfad_im.h" #include "bfa_fcs.h" BFA_TRC_FILE(LDRV, IM); DEFINE_IDR(bfad_im_port_index); struct scsi_transport_template *bfad_im_scsi_transport_template; struct scsi_transport_template *bfad_im_scsi_vport_transport_template; static void bfad_im_itnim_work_handler(struct work_struct *work); static int bfad_im_queuecommand(struct Scsi_Host *h, struct scsi_cmnd *cmnd); static int bfad_im_slave_alloc(struct scsi_device *sdev); static void bfad_im_fc_rport_add(struct bfad_im_port_s *im_port, struct bfad_itnim_s *itnim); void bfa_cb_ioim_done(void *drv, struct bfad_ioim_s *dio, enum bfi_ioim_status io_status, u8 scsi_status, int sns_len, u8 *sns_info, s32 residue) { struct scsi_cmnd *cmnd = (struct scsi_cmnd *)dio; struct bfad_s *bfad = drv; struct bfad_itnim_data_s *itnim_data; struct bfad_itnim_s *itnim; u8 host_status = DID_OK; switch (io_status) { case BFI_IOIM_STS_OK: bfa_trc(bfad, scsi_status); scsi_set_resid(cmnd, 0); if (sns_len > 0) { bfa_trc(bfad, sns_len); if (sns_len > SCSI_SENSE_BUFFERSIZE) sns_len = SCSI_SENSE_BUFFERSIZE; memcpy(cmnd->sense_buffer, sns_info, sns_len); } if (residue > 0) { bfa_trc(bfad, residue); scsi_set_resid(cmnd, residue); if (!sns_len && (scsi_status == SAM_STAT_GOOD) && (scsi_bufflen(cmnd) - residue) < cmnd->underflow) { bfa_trc(bfad, 0); host_status = DID_ERROR; } } cmnd->result = host_status << 16 | scsi_status; break; case BFI_IOIM_STS_TIMEDOUT: cmnd->result = DID_TIME_OUT << 16; break; case BFI_IOIM_STS_PATHTOV: cmnd->result = DID_TRANSPORT_DISRUPTED << 16; break; default: cmnd->result = DID_ERROR << 16; } /* Unmap DMA, if host is NULL, it means a scsi passthru cmd */ if (cmnd->device->host != NULL) scsi_dma_unmap(cmnd); cmnd->host_scribble = NULL; bfa_trc(bfad, cmnd->result); itnim_data = cmnd->device->hostdata; if (itnim_data) { itnim = itnim_data->itnim; if (!cmnd->result && itnim && (bfa_lun_queue_depth > cmnd->device->queue_depth)) { /* Queue depth adjustment for good status completion */ bfad_ramp_up_qdepth(itnim, cmnd->device); } else if (cmnd->result == SAM_STAT_TASK_SET_FULL && itnim) { /* qfull handling */ bfad_handle_qfull(itnim, cmnd->device); } } cmnd->scsi_done(cmnd); } void bfa_cb_ioim_good_comp(void *drv, struct bfad_ioim_s *dio) { struct scsi_cmnd *cmnd = (struct scsi_cmnd *)dio; struct bfad_itnim_data_s *itnim_data; struct bfad_itnim_s *itnim; cmnd->result = DID_OK << 16 | SCSI_STATUS_GOOD; /* Unmap DMA, if host is NULL, it means a scsi passthru cmd */ if (cmnd->device->host != NULL) scsi_dma_unmap(cmnd); cmnd->host_scribble = NULL; /* Queue depth adjustment */ if (bfa_lun_queue_depth > cmnd->device->queue_depth) { itnim_data = cmnd->device->hostdata; if (itnim_data) { itnim = itnim_data->itnim; if (itnim) bfad_ramp_up_qdepth(itnim, cmnd->device); } } cmnd->scsi_done(cmnd); } void bfa_cb_ioim_abort(void *drv, struct bfad_ioim_s *dio) { struct scsi_cmnd *cmnd = (struct scsi_cmnd *)dio; struct bfad_s *bfad = drv; cmnd->result = DID_ERROR << 16; /* Unmap DMA, if host is NULL, it means a scsi passthru cmd */ if (cmnd->device->host != NULL) scsi_dma_unmap(cmnd); bfa_trc(bfad, cmnd->result); cmnd->host_scribble = NULL; } void bfa_cb_tskim_done(void *bfad, struct bfad_tskim_s *dtsk, enum bfi_tskim_status tsk_status) { struct scsi_cmnd *cmnd = (struct scsi_cmnd *)dtsk; wait_queue_head_t *wq; cmnd->SCp.Status |= tsk_status << 1; set_bit(IO_DONE_BIT, (unsigned long *)&cmnd->SCp.Status); wq = (wait_queue_head_t *) cmnd->SCp.ptr; cmnd->SCp.ptr = NULL; if (wq) wake_up(wq); } /* * Scsi_Host_template SCSI host template */ /* * Scsi_Host template entry, returns BFAD PCI info. */ static const char * bfad_im_info(struct Scsi_Host *shost) { static char bfa_buf[256]; struct bfad_im_port_s *im_port = (struct bfad_im_port_s *) shost->hostdata[0]; struct bfad_s *bfad = im_port->bfad; memset(bfa_buf, 0, sizeof(bfa_buf)); snprintf(bfa_buf, sizeof(bfa_buf), "QLogic BR-series FC/FCOE Adapter, hwpath: %s driver: %s", bfad->pci_name, BFAD_DRIVER_VERSION); return bfa_buf; } /* * Scsi_Host template entry, aborts the specified SCSI command. * * Returns: SUCCESS or FAILED. */ static int bfad_im_abort_handler(struct scsi_cmnd *cmnd) { struct Scsi_Host *shost = cmnd->device->host; struct bfad_im_port_s *im_port = (struct bfad_im_port_s *) shost->hostdata[0]; struct bfad_s *bfad = im_port->bfad; struct bfa_ioim_s *hal_io; unsigned long flags; u32 timeout; int rc = FAILED; spin_lock_irqsave(&bfad->bfad_lock, flags); hal_io = (struct bfa_ioim_s *) cmnd->host_scribble; if (!hal_io) { /* IO has been completed, return success */ rc = SUCCESS; goto out; } if (hal_io->dio != (struct bfad_ioim_s *) cmnd) { rc = FAILED; goto out; } bfa_trc(bfad, hal_io->iotag); BFA_LOG(KERN_INFO, bfad, bfa_log_level, "scsi%d: abort cmnd %p iotag %x\n", im_port->shost->host_no, cmnd, hal_io->iotag); (void) bfa_ioim_abort(hal_io); spin_unlock_irqrestore(&bfad->bfad_lock, flags); /* Need to wait until the command get aborted */ timeout = 10; while ((struct bfa_ioim_s *) cmnd->host_scribble == hal_io) { set_current_state(TASK_UNINTERRUPTIBLE); schedule_timeout(timeout); if (timeout < 4 * HZ) timeout *= 2; } cmnd->scsi_done(cmnd); bfa_trc(bfad, hal_io->iotag); BFA_LOG(KERN_INFO, bfad, bfa_log_level, "scsi%d: complete abort 0x%p iotag 0x%x\n", im_port->shost->host_no, cmnd, hal_io->iotag); return SUCCESS; out: spin_unlock_irqrestore(&bfad->bfad_lock, flags); return rc; } static bfa_status_t bfad_im_target_reset_send(struct bfad_s *bfad, struct scsi_cmnd *cmnd, struct bfad_itnim_s *itnim) { struct bfa_tskim_s *tskim; struct bfa_itnim_s *bfa_itnim; bfa_status_t rc = BFA_STATUS_OK; struct scsi_lun scsilun; tskim = bfa_tskim_alloc(&bfad->bfa, (struct bfad_tskim_s *) cmnd); if (!tskim) { BFA_LOG(KERN_ERR, bfad, bfa_log_level, "target reset, fail to allocate tskim\n"); rc = BFA_STATUS_FAILED; goto out; } /* * Set host_scribble to NULL to avoid aborting a task command if * happens. */ cmnd->host_scribble = NULL; cmnd->SCp.Status = 0; bfa_itnim = bfa_fcs_itnim_get_halitn(&itnim->fcs_itnim); /* * bfa_itnim can be NULL if the port gets disconnected and the bfa * and fcs layers have cleaned up their nexus with the targets and * the same has not been cleaned up by the shim */ if (bfa_itnim == NULL) { bfa_tskim_free(tskim); BFA_LOG(KERN_ERR, bfad, bfa_log_level, "target reset, bfa_itnim is NULL\n"); rc = BFA_STATUS_FAILED; goto out; } memset(&scsilun, 0, sizeof(scsilun)); bfa_tskim_start(tskim, bfa_itnim, scsilun, FCP_TM_TARGET_RESET, BFAD_TARGET_RESET_TMO); out: return rc; } /* * Scsi_Host template entry, resets a LUN and abort its all commands. * * Returns: SUCCESS or FAILED. * */ static int bfad_im_reset_lun_handler(struct scsi_cmnd *cmnd) { struct Scsi_Host *shost = cmnd->device->host; struct bfad_im_port_s *im_port = (struct bfad_im_port_s *) shost->hostdata[0]; struct bfad_itnim_data_s *itnim_data = cmnd->device->hostdata; struct bfad_s *bfad = im_port->bfad; struct bfa_tskim_s *tskim; struct bfad_itnim_s *itnim; struct bfa_itnim_s *bfa_itnim; DECLARE_WAIT_QUEUE_HEAD_ONSTACK(wq); int rc = SUCCESS; unsigned long flags; enum bfi_tskim_status task_status; struct scsi_lun scsilun; spin_lock_irqsave(&bfad->bfad_lock, flags); itnim = itnim_data->itnim; if (!itnim) { spin_unlock_irqrestore(&bfad->bfad_lock, flags); rc = FAILED; goto out; } tskim = bfa_tskim_alloc(&bfad->bfa, (struct bfad_tskim_s *) cmnd); if (!tskim) { BFA_LOG(KERN_ERR, bfad, bfa_log_level, "LUN reset, fail to allocate tskim"); spin_unlock_irqrestore(&bfad->bfad_lock, flags); rc = FAILED; goto out; } /* * Set host_scribble to NULL to avoid aborting a task command * if happens. */ cmnd->host_scribble = NULL; cmnd->SCp.ptr = (char *)&wq; cmnd->SCp.Status = 0; bfa_itnim = bfa_fcs_itnim_get_halitn(&itnim->fcs_itnim); /* * bfa_itnim can be NULL if the port gets disconnected and the bfa * and fcs layers have cleaned up their nexus with the targets and * the same has not been cleaned up by the shim */ if (bfa_itnim == NULL) { bfa_tskim_free(tskim); BFA_LOG(KERN_ERR, bfad, bfa_log_level, "lun reset, bfa_itnim is NULL\n"); spin_unlock_irqrestore(&bfad->bfad_lock, flags); rc = FAILED; goto out; } int_to_scsilun(cmnd->device->lun, &scsilun); bfa_tskim_start(tskim, bfa_itnim, scsilun, FCP_TM_LUN_RESET, BFAD_LUN_RESET_TMO); spin_unlock_irqrestore(&bfad->bfad_lock, flags); wait_event(wq, test_bit(IO_DONE_BIT, (unsigned long *)&cmnd->SCp.Status)); task_status = cmnd->SCp.Status >> 1; if (task_status != BFI_TSKIM_STS_OK) { BFA_LOG(KERN_ERR, bfad, bfa_log_level, "LUN reset failure, status: %d\n", task_status); rc = FAILED; } out: return rc; } /* * Scsi_Host template entry, resets the target and abort all commands. */ static int bfad_im_reset_target_handler(struct scsi_cmnd *cmnd) { struct Scsi_Host *shost = cmnd->device->host; struct scsi_target *starget = scsi_target(cmnd->device); struct bfad_im_port_s *im_port = (struct bfad_im_port_s *) shost->hostdata[0]; struct bfad_s *bfad = im_port->bfad; struct bfad_itnim_s *itnim; unsigned long flags; u32 rc, rtn = FAILED; DECLARE_WAIT_QUEUE_HEAD_ONSTACK(wq); enum bfi_tskim_status task_status; spin_lock_irqsave(&bfad->bfad_lock, flags); itnim = bfad_get_itnim(im_port, starget->id); if (itnim) { cmnd->SCp.ptr = (char *)&wq; rc = bfad_im_target_reset_send(bfad, cmnd, itnim); if (rc == BFA_STATUS_OK) { /* wait target reset to complete */ spin_unlock_irqrestore(&bfad->bfad_lock, flags); wait_event(wq, test_bit(IO_DONE_BIT, (unsigned long *)&cmnd->SCp.Status)); spin_lock_irqsave(&bfad->bfad_lock, flags); task_status = cmnd->SCp.Status >> 1; if (task_status != BFI_TSKIM_STS_OK) BFA_LOG(KERN_ERR, bfad, bfa_log_level, "target reset failure," " status: %d\n", task_status); else rtn = SUCCESS; } } spin_unlock_irqrestore(&bfad->bfad_lock, flags); return rtn; } /* * Scsi_Host template entry slave_destroy. */ static void bfad_im_slave_destroy(struct scsi_device *sdev) { sdev->hostdata = NULL; return; } /* * BFA FCS itnim callbacks */ /* * BFA FCS itnim alloc callback, after successful PRLI * Context: Interrupt */ int bfa_fcb_itnim_alloc(struct bfad_s *bfad, struct bfa_fcs_itnim_s **itnim, struct bfad_itnim_s **itnim_drv) { *itnim_drv = kzalloc(sizeof(struct bfad_itnim_s), GFP_ATOMIC); if (*itnim_drv == NULL) return -ENOMEM; (*itnim_drv)->im = bfad->im; *itnim = &(*itnim_drv)->fcs_itnim; (*itnim_drv)->state = ITNIM_STATE_NONE; /* * Initiaze the itnim_work */ INIT_WORK(&(*itnim_drv)->itnim_work, bfad_im_itnim_work_handler); bfad->bfad_flags |= BFAD_RPORT_ONLINE; return 0; } /* * BFA FCS itnim free callback. * Context: Interrupt. bfad_lock is held */ void bfa_fcb_itnim_free(struct bfad_s *bfad, struct bfad_itnim_s *itnim_drv) { struct bfad_port_s *port; wwn_t wwpn; u32 fcid; char wwpn_str[32], fcid_str[16]; struct bfad_im_s *im = itnim_drv->im; /* online to free state transtion should not happen */ WARN_ON(itnim_drv->state == ITNIM_STATE_ONLINE); itnim_drv->queue_work = 1; /* offline request is not yet done, use the same request to free */ if (itnim_drv->state == ITNIM_STATE_OFFLINE_PENDING) itnim_drv->queue_work = 0; itnim_drv->state = ITNIM_STATE_FREE; port = bfa_fcs_itnim_get_drvport(&itnim_drv->fcs_itnim); itnim_drv->im_port = port->im_port; wwpn = bfa_fcs_itnim_get_pwwn(&itnim_drv->fcs_itnim); fcid = bfa_fcs_itnim_get_fcid(&itnim_drv->fcs_itnim); wwn2str(wwpn_str, wwpn); fcid2str(fcid_str, fcid); BFA_LOG(KERN_INFO, bfad, bfa_log_level, "ITNIM FREE scsi%d: FCID: %s WWPN: %s\n", port->im_port->shost->host_no, fcid_str, wwpn_str); /* ITNIM processing */ if (itnim_drv->queue_work) queue_work(im->drv_workq, &itnim_drv->itnim_work); } /* * BFA FCS itnim online callback. * Context: Interrupt. bfad_lock is held */ void bfa_fcb_itnim_online(struct bfad_itnim_s *itnim_drv) { struct bfad_port_s *port; struct bfad_im_s *im = itnim_drv->im; itnim_drv->bfa_itnim = bfa_fcs_itnim_get_halitn(&itnim_drv->fcs_itnim); port = bfa_fcs_itnim_get_drvport(&itnim_drv->fcs_itnim); itnim_drv->state = ITNIM_STATE_ONLINE; itnim_drv->queue_work = 1; itnim_drv->im_port = port->im_port; /* ITNIM processing */ if (itnim_drv->queue_work) queue_work(im->drv_workq, &itnim_drv->itnim_work); } /* * BFA FCS itnim offline callback. * Context: Interrupt. bfad_lock is held */ void bfa_fcb_itnim_offline(struct bfad_itnim_s *itnim_drv) { struct bfad_port_s *port; struct bfad_s *bfad; struct bfad_im_s *im = itnim_drv->im; port = bfa_fcs_itnim_get_drvport(&itnim_drv->fcs_itnim); bfad = port->bfad; if ((bfad->pport.flags & BFAD_PORT_DELETE) || (port->flags & BFAD_PORT_DELETE)) { itnim_drv->state = ITNIM_STATE_OFFLINE; return; } itnim_drv->im_port = port->im_port; itnim_drv->state = ITNIM_STATE_OFFLINE_PENDING; itnim_drv->queue_work = 1; /* ITNIM processing */ if (itnim_drv->queue_work) queue_work(im->drv_workq, &itnim_drv->itnim_work); } /* * Allocate a Scsi_Host for a port. */ int bfad_im_scsi_host_alloc(struct bfad_s *bfad, struct bfad_im_port_s *im_port, struct device *dev) { struct bfad_im_port_pointer *im_portp; int error = 1; mutex_lock(&bfad_mutex); error = idr_alloc(&bfad_im_port_index, im_port, 0, 0, GFP_KERNEL); if (error < 0) { mutex_unlock(&bfad_mutex); printk(KERN_WARNING "idr_alloc failure\n"); goto out; } im_port->idr_id = error; mutex_unlock(&bfad_mutex); im_port->shost = bfad_scsi_host_alloc(im_port, bfad); if (!im_port->shost) { error = 1; goto out_free_idr; } im_portp = shost_priv(im_port->shost); im_portp->p = im_port; im_port->shost->unique_id = im_port->idr_id; im_port->shost->this_id = -1; im_port->shost->max_id = MAX_FCP_TARGET; im_port->shost->max_lun = MAX_FCP_LUN; im_port->shost->max_cmd_len = 16; im_port->shost->can_queue = bfad->cfg_data.ioc_queue_depth; if (im_port->port->pvb_type == BFAD_PORT_PHYS_BASE) im_port->shost->transportt = bfad_im_scsi_transport_template; else im_port->shost->transportt = bfad_im_scsi_vport_transport_template; error = scsi_add_host_with_dma(im_port->shost, dev, &bfad->pcidev->dev); if (error) { printk(KERN_WARNING "scsi_add_host failure %d\n", error); goto out_fc_rel; } return 0; out_fc_rel: scsi_host_put(im_port->shost); im_port->shost = NULL; out_free_idr: mutex_lock(&bfad_mutex); idr_remove(&bfad_im_port_index, im_port->idr_id); mutex_unlock(&bfad_mutex); out: return error; } void bfad_im_scsi_host_free(struct bfad_s *bfad, struct bfad_im_port_s *im_port) { bfa_trc(bfad, bfad->inst_no); BFA_LOG(KERN_INFO, bfad, bfa_log_level, "Free scsi%d\n", im_port->shost->host_no); fc_remove_host(im_port->shost); scsi_remove_host(im_port->shost); scsi_host_put(im_port->shost); mutex_lock(&bfad_mutex); idr_remove(&bfad_im_port_index, im_port->idr_id); mutex_unlock(&bfad_mutex); } static void bfad_im_port_delete_handler(struct work_struct *work) { struct bfad_im_port_s *im_port = container_of(work, struct bfad_im_port_s, port_delete_work); if (im_port->port->pvb_type != BFAD_PORT_PHYS_BASE) { im_port->flags |= BFAD_PORT_DELETE; fc_vport_terminate(im_port->fc_vport); } } bfa_status_t bfad_im_port_new(struct bfad_s *bfad, struct bfad_port_s *port) { int rc = BFA_STATUS_OK; struct bfad_im_port_s *im_port; im_port = kzalloc(sizeof(struct bfad_im_port_s), GFP_ATOMIC); if (im_port == NULL) { rc = BFA_STATUS_ENOMEM; goto ext; } port->im_port = im_port; im_port->port = port; im_port->bfad = bfad; INIT_WORK(&im_port->port_delete_work, bfad_im_port_delete_handler); INIT_LIST_HEAD(&im_port->itnim_mapped_list); INIT_LIST_HEAD(&im_port->binding_list); ext: return rc; } void bfad_im_port_delete(struct bfad_s *bfad, struct bfad_port_s *port) { struct bfad_im_port_s *im_port = port->im_port; queue_work(bfad->im->drv_workq, &im_port->port_delete_work); } void bfad_im_port_clean(struct bfad_im_port_s *im_port) { struct bfad_fcp_binding *bp, *bp_new; unsigned long flags; struct bfad_s *bfad = im_port->bfad; spin_lock_irqsave(&bfad->bfad_lock, flags); list_for_each_entry_safe(bp, bp_new, &im_port->binding_list, list_entry) { list_del(&bp->list_entry); kfree(bp); } /* the itnim_mapped_list must be empty at this time */ WARN_ON(!list_empty(&im_port->itnim_mapped_list)); spin_unlock_irqrestore(&bfad->bfad_lock, flags); } static void bfad_aen_im_notify_handler(struct work_struct *work) { struct bfad_im_s *im = container_of(work, struct bfad_im_s, aen_im_notify_work); struct bfa_aen_entry_s *aen_entry; struct bfad_s *bfad = im->bfad; struct Scsi_Host *shost = bfad->pport.im_port->shost; void *event_data; unsigned long flags; while (!list_empty(&bfad->active_aen_q)) { spin_lock_irqsave(&bfad->bfad_aen_spinlock, flags); bfa_q_deq(&bfad->active_aen_q, &aen_entry); spin_unlock_irqrestore(&bfad->bfad_aen_spinlock, flags); event_data = (char *)aen_entry + sizeof(struct list_head); fc_host_post_vendor_event(shost, fc_get_event_number(), sizeof(struct bfa_aen_entry_s) - sizeof(struct list_head), (char *)event_data, BFAD_NL_VENDOR_ID); spin_lock_irqsave(&bfad->bfad_aen_spinlock, flags); list_add_tail(&aen_entry->qe, &bfad->free_aen_q); spin_unlock_irqrestore(&bfad->bfad_aen_spinlock, flags); } } bfa_status_t bfad_im_probe(struct bfad_s *bfad) { struct bfad_im_s *im; im = kzalloc(sizeof(struct bfad_im_s), GFP_KERNEL); if (im == NULL) return BFA_STATUS_ENOMEM; bfad->im = im; im->bfad = bfad; if (bfad_thread_workq(bfad) != BFA_STATUS_OK) { kfree(im); return BFA_STATUS_FAILED; } INIT_WORK(&im->aen_im_notify_work, bfad_aen_im_notify_handler); return BFA_STATUS_OK; } void bfad_im_probe_undo(struct bfad_s *bfad) { if (bfad->im) { bfad_destroy_workq(bfad->im); kfree(bfad->im); bfad->im = NULL; } } struct Scsi_Host * bfad_scsi_host_alloc(struct bfad_im_port_s *im_port, struct bfad_s *bfad) { struct scsi_host_template *sht; if (im_port->port->pvb_type == BFAD_PORT_PHYS_BASE) sht = &bfad_im_scsi_host_template; else sht = &bfad_im_vport_template; if (max_xfer_size != BFAD_MAX_SECTORS >> 1) sht->max_sectors = max_xfer_size << 1; sht->sg_tablesize = bfad->cfg_data.io_max_sge; return scsi_host_alloc(sht, sizeof(struct bfad_im_port_pointer)); } void bfad_scsi_host_free(struct bfad_s *bfad, struct bfad_im_port_s *im_port) { if (!(im_port->flags & BFAD_PORT_DELETE)) flush_workqueue(bfad->im->drv_workq); bfad_im_scsi_host_free(im_port->bfad, im_port); bfad_im_port_clean(im_port); kfree(im_port); } void bfad_destroy_workq(struct bfad_im_s *im) { if (im && im->drv_workq) { flush_workqueue(im->drv_workq); destroy_workqueue(im->drv_workq); im->drv_workq = NULL; } } bfa_status_t bfad_thread_workq(struct bfad_s *bfad) { struct bfad_im_s *im = bfad->im; bfa_trc(bfad, 0); snprintf(im->drv_workq_name, KOBJ_NAME_LEN, "bfad_wq_%d", bfad->inst_no); im->drv_workq = create_singlethread_workqueue(im->drv_workq_name); if (!im->drv_workq) return BFA_STATUS_FAILED; return BFA_STATUS_OK; } /* * Scsi_Host template entry. * * Description: * OS entry point to adjust the queue_depths on a per-device basis. * Called once per device during the bus scan. * Return non-zero if fails. */ static int bfad_im_slave_configure(struct scsi_device *sdev) { scsi_change_queue_depth(sdev, bfa_lun_queue_depth); return 0; } struct scsi_host_template bfad_im_scsi_host_template = { .module = THIS_MODULE, .name = BFAD_DRIVER_NAME, .info = bfad_im_info, .queuecommand = bfad_im_queuecommand, .eh_timed_out = fc_eh_timed_out, .eh_abort_handler = bfad_im_abort_handler, .eh_device_reset_handler = bfad_im_reset_lun_handler, .eh_target_reset_handler = bfad_im_reset_target_handler, .slave_alloc = bfad_im_slave_alloc, .slave_configure = bfad_im_slave_configure, .slave_destroy = bfad_im_slave_destroy, .this_id = -1, .sg_tablesize = BFAD_IO_MAX_SGE, .cmd_per_lun = 3, .shost_attrs = bfad_im_host_attrs, .max_sectors = BFAD_MAX_SECTORS, .vendor_id = BFA_PCI_VENDOR_ID_BROCADE, }; struct scsi_host_template bfad_im_vport_template = { .module = THIS_MODULE, .name = BFAD_DRIVER_NAME, .info = bfad_im_info, .queuecommand = bfad_im_queuecommand, .eh_timed_out = fc_eh_timed_out, .eh_abort_handler = bfad_im_abort_handler, .eh_device_reset_handler = bfad_im_reset_lun_handler, .eh_target_reset_handler = bfad_im_reset_target_handler, .slave_alloc = bfad_im_slave_alloc, .slave_configure = bfad_im_slave_configure, .slave_destroy = bfad_im_slave_destroy, .this_id = -1, .sg_tablesize = BFAD_IO_MAX_SGE, .cmd_per_lun = 3, .shost_attrs = bfad_im_vport_attrs, .max_sectors = BFAD_MAX_SECTORS, }; bfa_status_t bfad_im_module_init(void) { bfad_im_scsi_transport_template = fc_attach_transport(&bfad_im_fc_function_template); if (!bfad_im_scsi_transport_template) return BFA_STATUS_ENOMEM; bfad_im_scsi_vport_transport_template = fc_attach_transport(&bfad_im_vport_fc_function_template); if (!bfad_im_scsi_vport_transport_template) { fc_release_transport(bfad_im_scsi_transport_template); return BFA_STATUS_ENOMEM; } return BFA_STATUS_OK; } void bfad_im_module_exit(void) { if (bfad_im_scsi_transport_template) fc_release_transport(bfad_im_scsi_transport_template); if (bfad_im_scsi_vport_transport_template) fc_release_transport(bfad_im_scsi_vport_transport_template); idr_destroy(&bfad_im_port_index); } void bfad_ramp_up_qdepth(struct bfad_itnim_s *itnim, struct scsi_device *sdev) { struct scsi_device *tmp_sdev; if (((jiffies - itnim->last_ramp_up_time) > BFA_QUEUE_FULL_RAMP_UP_TIME * HZ) && ((jiffies - itnim->last_queue_full_time) > BFA_QUEUE_FULL_RAMP_UP_TIME * HZ)) { shost_for_each_device(tmp_sdev, sdev->host) { if (bfa_lun_queue_depth > tmp_sdev->queue_depth) { if (tmp_sdev->id != sdev->id) continue; scsi_change_queue_depth(tmp_sdev, tmp_sdev->queue_depth + 1); itnim->last_ramp_up_time = jiffies; } } } } void bfad_handle_qfull(struct bfad_itnim_s *itnim, struct scsi_device *sdev) { struct scsi_device *tmp_sdev; itnim->last_queue_full_time = jiffies; shost_for_each_device(tmp_sdev, sdev->host) { if (tmp_sdev->id != sdev->id) continue; scsi_track_queue_full(tmp_sdev, tmp_sdev->queue_depth - 1); } } struct bfad_itnim_s * bfad_get_itnim(struct bfad_im_port_s *im_port, int id) { struct bfad_itnim_s *itnim = NULL; /* Search the mapped list for this target ID */ list_for_each_entry(itnim, &im_port->itnim_mapped_list, list_entry) { if (id == itnim->scsi_tgt_id) return itnim; } return NULL; } /* * Function is invoked from the SCSI Host Template slave_alloc() entry point. * Has the logic to query the LUN Mask database to check if this LUN needs to * be made visible to the SCSI mid-layer or not. * * Returns BFA_STATUS_OK if this LUN needs to be added to the OS stack. * Returns -ENXIO to notify SCSI mid-layer to not add this LUN to the OS stack. */ static int bfad_im_check_if_make_lun_visible(struct scsi_device *sdev, struct fc_rport *rport) { struct bfad_itnim_data_s *itnim_data = (struct bfad_itnim_data_s *) rport->dd_data; struct bfa_s *bfa = itnim_data->itnim->bfa_itnim->bfa; struct bfa_rport_s *bfa_rport = itnim_data->itnim->bfa_itnim->rport; struct bfa_lun_mask_s *lun_list = bfa_get_lun_mask_list(bfa); int i = 0, ret = -ENXIO; for (i = 0; i < MAX_LUN_MASK_CFG; i++) { if (lun_list[i].state == BFA_IOIM_LUN_MASK_ACTIVE && scsilun_to_int(&lun_list[i].lun) == sdev->lun && lun_list[i].rp_tag == bfa_rport->rport_tag && lun_list[i].lp_tag == (u8)bfa_rport->rport_info.lp_tag) { ret = BFA_STATUS_OK; break; } } return ret; } /* * Scsi_Host template entry slave_alloc */ static int bfad_im_slave_alloc(struct scsi_device *sdev) { struct fc_rport *rport = starget_to_rport(scsi_target(sdev)); struct bfad_itnim_data_s *itnim_data; struct bfa_s *bfa; if (!rport || fc_remote_port_chkready(rport)) return -ENXIO; itnim_data = (struct bfad_itnim_data_s *) rport->dd_data; bfa = itnim_data->itnim->bfa_itnim->bfa; if (bfa_get_lun_mask_status(bfa) == BFA_LUNMASK_ENABLED) { /* * We should not mask LUN 0 - since this will translate * to no LUN / TARGET for SCSI ml resulting no scan. */ if (sdev->lun == 0) { sdev->sdev_bflags |= BLIST_NOREPORTLUN | BLIST_SPARSELUN; goto done; } /* * Query LUN Mask configuration - to expose this LUN * to the SCSI mid-layer or to mask it. */ if (bfad_im_check_if_make_lun_visible(sdev, rport) != BFA_STATUS_OK) return -ENXIO; } done: sdev->hostdata = rport->dd_data; return 0; } u32 bfad_im_supported_speeds(struct bfa_s *bfa) { struct bfa_ioc_attr_s *ioc_attr; u32 supported_speed = 0; ioc_attr = kzalloc(sizeof(struct bfa_ioc_attr_s), GFP_KERNEL); if (!ioc_attr) return 0; bfa_ioc_get_attr(&bfa->ioc, ioc_attr); if (ioc_attr->adapter_attr.max_speed == BFA_PORT_SPEED_16GBPS) supported_speed |= FC_PORTSPEED_16GBIT | FC_PORTSPEED_8GBIT | FC_PORTSPEED_4GBIT | FC_PORTSPEED_2GBIT; else if (ioc_attr->adapter_attr.max_speed == BFA_PORT_SPEED_8GBPS) { if (ioc_attr->adapter_attr.is_mezz) { supported_speed |= FC_PORTSPEED_8GBIT | FC_PORTSPEED_4GBIT | FC_PORTSPEED_2GBIT | FC_PORTSPEED_1GBIT; } else { supported_speed |= FC_PORTSPEED_8GBIT | FC_PORTSPEED_4GBIT | FC_PORTSPEED_2GBIT; } } else if (ioc_attr->adapter_attr.max_speed == BFA_PORT_SPEED_4GBPS) { supported_speed |= FC_PORTSPEED_4GBIT | FC_PORTSPEED_2GBIT | FC_PORTSPEED_1GBIT; } else if (ioc_attr->adapter_attr.max_speed == BFA_PORT_SPEED_10GBPS) { supported_speed |= FC_PORTSPEED_10GBIT; } kfree(ioc_attr); return supported_speed; } void bfad_fc_host_init(struct bfad_im_port_s *im_port) { struct Scsi_Host *host = im_port->shost; struct bfad_s *bfad = im_port->bfad; struct bfad_port_s *port = im_port->port; char symname[BFA_SYMNAME_MAXLEN]; struct bfa_fcport_s *fcport = BFA_FCPORT_MOD(&bfad->bfa); fc_host_node_name(host) = cpu_to_be64((bfa_fcs_lport_get_nwwn(port->fcs_port))); fc_host_port_name(host) = cpu_to_be64((bfa_fcs_lport_get_pwwn(port->fcs_port))); fc_host_max_npiv_vports(host) = bfa_lps_get_max_vport(&bfad->bfa); fc_host_supported_classes(host) = FC_COS_CLASS3; memset(fc_host_supported_fc4s(host), 0, sizeof(fc_host_supported_fc4s(host))); if (supported_fc4s & BFA_LPORT_ROLE_FCP_IM) /* For FCP type 0x08 */ fc_host_supported_fc4s(host)[2] = 1; /* For fibre channel services type 0x20 */ fc_host_supported_fc4s(host)[7] = 1; strlcpy(symname, bfad->bfa_fcs.fabric.bport.port_cfg.sym_name.symname, BFA_SYMNAME_MAXLEN); sprintf(fc_host_symbolic_name(host), "%s", symname); fc_host_supported_speeds(host) = bfad_im_supported_speeds(&bfad->bfa); fc_host_maxframe_size(host) = fcport->cfg.maxfrsize; } static void bfad_im_fc_rport_add(struct bfad_im_port_s *im_port, struct bfad_itnim_s *itnim) { struct fc_rport_identifiers rport_ids; struct fc_rport *fc_rport; struct bfad_itnim_data_s *itnim_data; rport_ids.node_name = cpu_to_be64(bfa_fcs_itnim_get_nwwn(&itnim->fcs_itnim)); rport_ids.port_name = cpu_to_be64(bfa_fcs_itnim_get_pwwn(&itnim->fcs_itnim)); rport_ids.port_id = bfa_hton3b(bfa_fcs_itnim_get_fcid(&itnim->fcs_itnim)); rport_ids.roles = FC_RPORT_ROLE_UNKNOWN; itnim->fc_rport = fc_rport = fc_remote_port_add(im_port->shost, 0, &rport_ids); if (!fc_rport) return; fc_rport->maxframe_size = bfa_fcs_itnim_get_maxfrsize(&itnim->fcs_itnim); fc_rport->supported_classes = bfa_fcs_itnim_get_cos(&itnim->fcs_itnim); itnim_data = fc_rport->dd_data; itnim_data->itnim = itnim; rport_ids.roles |= FC_RPORT_ROLE_FCP_TARGET; if (rport_ids.roles != FC_RPORT_ROLE_UNKNOWN) fc_remote_port_rolechg(fc_rport, rport_ids.roles); if ((fc_rport->scsi_target_id != -1) && (fc_rport->scsi_target_id < MAX_FCP_TARGET)) itnim->scsi_tgt_id = fc_rport->scsi_target_id; itnim->channel = fc_rport->channel; return; } /* * Work queue handler using FC transport service * Context: kernel */ static void bfad_im_itnim_work_handler(struct work_struct *work) { struct bfad_itnim_s *itnim = container_of(work, struct bfad_itnim_s, itnim_work); struct bfad_im_s *im = itnim->im; struct bfad_s *bfad = im->bfad; struct bfad_im_port_s *im_port; unsigned long flags; struct fc_rport *fc_rport; wwn_t wwpn; u32 fcid; char wwpn_str[32], fcid_str[16]; spin_lock_irqsave(&bfad->bfad_lock, flags); im_port = itnim->im_port; bfa_trc(bfad, itnim->state); switch (itnim->state) { case ITNIM_STATE_ONLINE: if (!itnim->fc_rport) { spin_unlock_irqrestore(&bfad->bfad_lock, flags); bfad_im_fc_rport_add(im_port, itnim); spin_lock_irqsave(&bfad->bfad_lock, flags); wwpn = bfa_fcs_itnim_get_pwwn(&itnim->fcs_itnim); fcid = bfa_fcs_itnim_get_fcid(&itnim->fcs_itnim); wwn2str(wwpn_str, wwpn); fcid2str(fcid_str, fcid); list_add_tail(&itnim->list_entry, &im_port->itnim_mapped_list); BFA_LOG(KERN_INFO, bfad, bfa_log_level, "ITNIM ONLINE Target: %d:0:%d " "FCID: %s WWPN: %s\n", im_port->shost->host_no, itnim->scsi_tgt_id, fcid_str, wwpn_str); } else { printk(KERN_WARNING "%s: itnim %llx is already in online state\n", __func__, bfa_fcs_itnim_get_pwwn(&itnim->fcs_itnim)); } break; case ITNIM_STATE_OFFLINE_PENDING: itnim->state = ITNIM_STATE_OFFLINE; if (itnim->fc_rport) { fc_rport = itnim->fc_rport; ((struct bfad_itnim_data_s *) fc_rport->dd_data)->itnim = NULL; itnim->fc_rport = NULL; if (!(im_port->port->flags & BFAD_PORT_DELETE)) { spin_unlock_irqrestore(&bfad->bfad_lock, flags); fc_rport->dev_loss_tmo = bfa_fcpim_path_tov_get(&bfad->bfa) + 1; fc_remote_port_delete(fc_rport); spin_lock_irqsave(&bfad->bfad_lock, flags); } wwpn = bfa_fcs_itnim_get_pwwn(&itnim->fcs_itnim); fcid = bfa_fcs_itnim_get_fcid(&itnim->fcs_itnim); wwn2str(wwpn_str, wwpn); fcid2str(fcid_str, fcid); list_del(&itnim->list_entry); BFA_LOG(KERN_INFO, bfad, bfa_log_level, "ITNIM OFFLINE Target: %d:0:%d " "FCID: %s WWPN: %s\n", im_port->shost->host_no, itnim->scsi_tgt_id, fcid_str, wwpn_str); } break; case ITNIM_STATE_FREE: if (itnim->fc_rport) { fc_rport = itnim->fc_rport; ((struct bfad_itnim_data_s *) fc_rport->dd_data)->itnim = NULL; itnim->fc_rport = NULL; if (!(im_port->port->flags & BFAD_PORT_DELETE)) { spin_unlock_irqrestore(&bfad->bfad_lock, flags); fc_rport->dev_loss_tmo = bfa_fcpim_path_tov_get(&bfad->bfa) + 1; fc_remote_port_delete(fc_rport); spin_lock_irqsave(&bfad->bfad_lock, flags); } list_del(&itnim->list_entry); } kfree(itnim); break; default: WARN_ON(1); break; } spin_unlock_irqrestore(&bfad->bfad_lock, flags); } /* * Scsi_Host template entry, queue a SCSI command to the BFAD. */ static int bfad_im_queuecommand_lck(struct scsi_cmnd *cmnd, void (*done) (struct scsi_cmnd *)) { struct bfad_im_port_s *im_port = (struct bfad_im_port_s *) cmnd->device->host->hostdata[0]; struct bfad_s *bfad = im_port->bfad; struct bfad_itnim_data_s *itnim_data = cmnd->device->hostdata; struct bfad_itnim_s *itnim; struct bfa_ioim_s *hal_io; unsigned long flags; int rc; int sg_cnt = 0; struct fc_rport *rport = starget_to_rport(scsi_target(cmnd->device)); rc = fc_remote_port_chkready(rport); if (rc) { cmnd->result = rc; done(cmnd); return 0; } if (bfad->bfad_flags & BFAD_EEH_BUSY) { if (bfad->bfad_flags & BFAD_EEH_PCI_CHANNEL_IO_PERM_FAILURE) cmnd->result = DID_NO_CONNECT << 16; else cmnd->result = DID_REQUEUE << 16; done(cmnd); return 0; } sg_cnt = scsi_dma_map(cmnd); if (sg_cnt < 0) return SCSI_MLQUEUE_HOST_BUSY; cmnd->scsi_done = done; spin_lock_irqsave(&bfad->bfad_lock, flags); if (!(bfad->bfad_flags & BFAD_HAL_START_DONE)) { printk(KERN_WARNING "bfad%d, queuecommand %p %x failed, BFA stopped\n", bfad->inst_no, cmnd, cmnd->cmnd[0]); cmnd->result = DID_NO_CONNECT << 16; goto out_fail_cmd; } itnim = itnim_data->itnim; if (!itnim) { cmnd->result = DID_IMM_RETRY << 16; goto out_fail_cmd; } hal_io = bfa_ioim_alloc(&bfad->bfa, (struct bfad_ioim_s *) cmnd, itnim->bfa_itnim, sg_cnt); if (!hal_io) { printk(KERN_WARNING "hal_io failure\n"); spin_unlock_irqrestore(&bfad->bfad_lock, flags); scsi_dma_unmap(cmnd); return SCSI_MLQUEUE_HOST_BUSY; } cmnd->host_scribble = (char *)hal_io; bfa_ioim_start(hal_io); spin_unlock_irqrestore(&bfad->bfad_lock, flags); return 0; out_fail_cmd: spin_unlock_irqrestore(&bfad->bfad_lock, flags); scsi_dma_unmap(cmnd); if (done) done(cmnd); return 0; } static DEF_SCSI_QCMD(bfad_im_queuecommand) void bfad_rport_online_wait(struct bfad_s *bfad) { int i; int rport_delay = 10; for (i = 0; !(bfad->bfad_flags & BFAD_PORT_ONLINE) && i < bfa_linkup_delay; i++) { set_current_state(TASK_UNINTERRUPTIBLE); schedule_timeout(HZ); } if (bfad->bfad_flags & BFAD_PORT_ONLINE) { rport_delay = rport_delay < bfa_linkup_delay ? rport_delay : bfa_linkup_delay; for (i = 0; !(bfad->bfad_flags & BFAD_RPORT_ONLINE) && i < rport_delay; i++) { set_current_state(TASK_UNINTERRUPTIBLE); schedule_timeout(HZ); } if (rport_delay > 0 && (bfad->bfad_flags & BFAD_RPORT_ONLINE)) { set_current_state(TASK_UNINTERRUPTIBLE); schedule_timeout(rport_delay * HZ); } } } int bfad_get_linkup_delay(struct bfad_s *bfad) { u8 nwwns = 0; wwn_t wwns[BFA_PREBOOT_BOOTLUN_MAX]; int linkup_delay; /* * Querying for the boot target port wwns * -- read from boot information in flash. * If nwwns > 0 => boot over SAN and set linkup_delay = 30 * else => local boot machine set linkup_delay = 0 */ bfa_iocfc_get_bootwwns(&bfad->bfa, &nwwns, wwns); if (nwwns > 0) /* If Boot over SAN set linkup_delay = 30sec */ linkup_delay = 30; else /* If local boot; no linkup_delay */ linkup_delay = 0; return linkup_delay; }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1