Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
James Bottomley | 3214 | 93.98% | 6 | 33.33% |
Darrick J. Wong | 107 | 3.13% | 3 | 16.67% |
Christoph Hellwig | 52 | 1.52% | 2 | 11.11% |
Jens Axboe | 29 | 0.85% | 1 | 5.56% |
Hannes Reinecke | 8 | 0.23% | 1 | 5.56% |
Al Viro | 5 | 0.15% | 1 | 5.56% |
Thomas Gleixner | 2 | 0.06% | 1 | 5.56% |
Jeff Garzik | 1 | 0.03% | 1 | 5.56% |
Dan J Williams | 1 | 0.03% | 1 | 5.56% |
Harvey Harrison | 1 | 0.03% | 1 | 5.56% |
Total | 3420 | 18 |
// SPDX-License-Identifier: GPL-2.0-only /* * Aic94xx SAS/SATA Tasks * * Copyright (C) 2005 Adaptec, Inc. All rights reserved. * Copyright (C) 2005 Luben Tuikov <luben_tuikov@adaptec.com> */ #include <linux/spinlock.h> #include "aic94xx.h" #include "aic94xx_sas.h" #include "aic94xx_hwi.h" static void asd_unbuild_ata_ascb(struct asd_ascb *a); static void asd_unbuild_smp_ascb(struct asd_ascb *a); static void asd_unbuild_ssp_ascb(struct asd_ascb *a); static void asd_can_dequeue(struct asd_ha_struct *asd_ha, int num) { unsigned long flags; spin_lock_irqsave(&asd_ha->seq.pend_q_lock, flags); asd_ha->seq.can_queue += num; spin_unlock_irqrestore(&asd_ha->seq.pend_q_lock, flags); } /* DMA_... to our direction translation. */ static const u8 data_dir_flags[] = { [DMA_BIDIRECTIONAL] = DATA_DIR_BYRECIPIENT, /* UNSPECIFIED */ [DMA_TO_DEVICE] = DATA_DIR_OUT, /* OUTBOUND */ [DMA_FROM_DEVICE] = DATA_DIR_IN, /* INBOUND */ [DMA_NONE] = DATA_DIR_NONE, /* NO TRANSFER */ }; static int asd_map_scatterlist(struct sas_task *task, struct sg_el *sg_arr, gfp_t gfp_flags) { struct asd_ascb *ascb = task->lldd_task; struct asd_ha_struct *asd_ha = ascb->ha; struct scatterlist *sc; int num_sg, res; if (task->data_dir == DMA_NONE) return 0; if (task->num_scatter == 0) { void *p = task->scatter; dma_addr_t dma = dma_map_single(&asd_ha->pcidev->dev, p, task->total_xfer_len, task->data_dir); sg_arr[0].bus_addr = cpu_to_le64((u64)dma); sg_arr[0].size = cpu_to_le32(task->total_xfer_len); sg_arr[0].flags |= ASD_SG_EL_LIST_EOL; return 0; } /* STP tasks come from libata which has already mapped * the SG list */ if (sas_protocol_ata(task->task_proto)) num_sg = task->num_scatter; else num_sg = dma_map_sg(&asd_ha->pcidev->dev, task->scatter, task->num_scatter, task->data_dir); if (num_sg == 0) return -ENOMEM; if (num_sg > 3) { int i; ascb->sg_arr = asd_alloc_coherent(asd_ha, num_sg*sizeof(struct sg_el), gfp_flags); if (!ascb->sg_arr) { res = -ENOMEM; goto err_unmap; } for_each_sg(task->scatter, sc, num_sg, i) { struct sg_el *sg = &((struct sg_el *)ascb->sg_arr->vaddr)[i]; sg->bus_addr = cpu_to_le64((u64)sg_dma_address(sc)); sg->size = cpu_to_le32((u32)sg_dma_len(sc)); if (i == num_sg-1) sg->flags |= ASD_SG_EL_LIST_EOL; } for_each_sg(task->scatter, sc, 2, i) { sg_arr[i].bus_addr = cpu_to_le64((u64)sg_dma_address(sc)); sg_arr[i].size = cpu_to_le32((u32)sg_dma_len(sc)); } sg_arr[1].next_sg_offs = 2 * sizeof(*sg_arr); sg_arr[1].flags |= ASD_SG_EL_LIST_EOS; memset(&sg_arr[2], 0, sizeof(*sg_arr)); sg_arr[2].bus_addr=cpu_to_le64((u64)ascb->sg_arr->dma_handle); } else { int i; for_each_sg(task->scatter, sc, num_sg, i) { sg_arr[i].bus_addr = cpu_to_le64((u64)sg_dma_address(sc)); sg_arr[i].size = cpu_to_le32((u32)sg_dma_len(sc)); } sg_arr[i-1].flags |= ASD_SG_EL_LIST_EOL; } return 0; err_unmap: if (sas_protocol_ata(task->task_proto)) dma_unmap_sg(&asd_ha->pcidev->dev, task->scatter, task->num_scatter, task->data_dir); return res; } static void asd_unmap_scatterlist(struct asd_ascb *ascb) { struct asd_ha_struct *asd_ha = ascb->ha; struct sas_task *task = ascb->uldd_task; if (task->data_dir == DMA_NONE) return; if (task->num_scatter == 0) { dma_addr_t dma = (dma_addr_t) le64_to_cpu(ascb->scb->ssp_task.sg_element[0].bus_addr); dma_unmap_single(&ascb->ha->pcidev->dev, dma, task->total_xfer_len, task->data_dir); return; } asd_free_coherent(asd_ha, ascb->sg_arr); if (task->task_proto != SAS_PROTOCOL_STP) dma_unmap_sg(&asd_ha->pcidev->dev, task->scatter, task->num_scatter, task->data_dir); } /* ---------- Task complete tasklet ---------- */ static void asd_get_response_tasklet(struct asd_ascb *ascb, struct done_list_struct *dl) { struct asd_ha_struct *asd_ha = ascb->ha; struct sas_task *task = ascb->uldd_task; struct task_status_struct *ts = &task->task_status; unsigned long flags; struct tc_resp_sb_struct { __le16 index_escb; u8 len_lsb; u8 flags; } __attribute__ ((packed)) *resp_sb = (void *) dl->status_block; /* int size = ((resp_sb->flags & 7) << 8) | resp_sb->len_lsb; */ int edb_id = ((resp_sb->flags & 0x70) >> 4)-1; struct asd_ascb *escb; struct asd_dma_tok *edb; void *r; spin_lock_irqsave(&asd_ha->seq.tc_index_lock, flags); escb = asd_tc_index_find(&asd_ha->seq, (int)le16_to_cpu(resp_sb->index_escb)); spin_unlock_irqrestore(&asd_ha->seq.tc_index_lock, flags); if (!escb) { ASD_DPRINTK("Uh-oh! No escb for this dl?!\n"); return; } ts->buf_valid_size = 0; edb = asd_ha->seq.edb_arr[edb_id + escb->edb_index]; r = edb->vaddr; if (task->task_proto == SAS_PROTOCOL_SSP) { struct ssp_response_iu *iu = r + 16 + sizeof(struct ssp_frame_hdr); ts->residual = le32_to_cpu(*(__le32 *)r); sas_ssp_task_response(&asd_ha->pcidev->dev, task, iu); } else { struct ata_task_resp *resp = (void *) &ts->buf[0]; ts->residual = le32_to_cpu(*(__le32 *)r); if (SAS_STATUS_BUF_SIZE >= sizeof(*resp)) { resp->frame_len = le16_to_cpu(*(__le16 *)(r+6)); memcpy(&resp->ending_fis[0], r+16, ATA_RESP_FIS_SIZE); ts->buf_valid_size = sizeof(*resp); } } asd_invalidate_edb(escb, edb_id); } static void asd_task_tasklet_complete(struct asd_ascb *ascb, struct done_list_struct *dl) { struct sas_task *task = ascb->uldd_task; struct task_status_struct *ts = &task->task_status; unsigned long flags; u8 opcode = dl->opcode; asd_can_dequeue(ascb->ha, 1); Again: switch (opcode) { case TC_NO_ERROR: ts->resp = SAS_TASK_COMPLETE; ts->stat = SAM_STAT_GOOD; break; case TC_UNDERRUN: ts->resp = SAS_TASK_COMPLETE; ts->stat = SAS_DATA_UNDERRUN; ts->residual = le32_to_cpu(*(__le32 *)dl->status_block); break; case TC_OVERRUN: ts->resp = SAS_TASK_COMPLETE; ts->stat = SAS_DATA_OVERRUN; ts->residual = 0; break; case TC_SSP_RESP: case TC_ATA_RESP: ts->resp = SAS_TASK_COMPLETE; ts->stat = SAS_PROTO_RESPONSE; asd_get_response_tasklet(ascb, dl); break; case TF_OPEN_REJECT: ts->resp = SAS_TASK_UNDELIVERED; ts->stat = SAS_OPEN_REJECT; if (dl->status_block[1] & 2) ts->open_rej_reason = 1 + dl->status_block[2]; else if (dl->status_block[1] & 1) ts->open_rej_reason = (dl->status_block[2] >> 4)+10; else ts->open_rej_reason = SAS_OREJ_UNKNOWN; break; case TF_OPEN_TO: ts->resp = SAS_TASK_UNDELIVERED; ts->stat = SAS_OPEN_TO; break; case TF_PHY_DOWN: case TU_PHY_DOWN: ts->resp = SAS_TASK_UNDELIVERED; ts->stat = SAS_PHY_DOWN; break; case TI_PHY_DOWN: ts->resp = SAS_TASK_COMPLETE; ts->stat = SAS_PHY_DOWN; break; case TI_BREAK: case TI_PROTO_ERR: case TI_NAK: case TI_ACK_NAK_TO: case TF_SMP_XMIT_RCV_ERR: case TC_ATA_R_ERR_RECV: ts->resp = SAS_TASK_COMPLETE; ts->stat = SAS_INTERRUPTED; break; case TF_BREAK: case TU_BREAK: case TU_ACK_NAK_TO: case TF_SMPRSP_TO: ts->resp = SAS_TASK_UNDELIVERED; ts->stat = SAS_DEV_NO_RESPONSE; break; case TF_NAK_RECV: ts->resp = SAS_TASK_COMPLETE; ts->stat = SAS_NAK_R_ERR; break; case TA_I_T_NEXUS_LOSS: opcode = dl->status_block[0]; goto Again; break; case TF_INV_CONN_HANDLE: ts->resp = SAS_TASK_UNDELIVERED; ts->stat = SAS_DEVICE_UNKNOWN; break; case TF_REQUESTED_N_PENDING: ts->resp = SAS_TASK_UNDELIVERED; ts->stat = SAS_PENDING; break; case TC_TASK_CLEARED: case TA_ON_REQ: ts->resp = SAS_TASK_COMPLETE; ts->stat = SAS_ABORTED_TASK; break; case TF_NO_SMP_CONN: case TF_TMF_NO_CTX: case TF_TMF_NO_TAG: case TF_TMF_TAG_FREE: case TF_TMF_TASK_DONE: case TF_TMF_NO_CONN_HANDLE: case TF_IRTT_TO: case TF_IU_SHORT: case TF_DATA_OFFS_ERR: ts->resp = SAS_TASK_UNDELIVERED; ts->stat = SAS_DEV_NO_RESPONSE; break; case TC_LINK_ADM_RESP: case TC_CONTROL_PHY: case TC_RESUME: case TC_PARTIAL_SG_LIST: default: ASD_DPRINTK("%s: dl opcode: 0x%x?\n", __func__, opcode); break; } switch (task->task_proto) { case SAS_PROTOCOL_SATA: case SAS_PROTOCOL_STP: asd_unbuild_ata_ascb(ascb); break; case SAS_PROTOCOL_SMP: asd_unbuild_smp_ascb(ascb); break; case SAS_PROTOCOL_SSP: asd_unbuild_ssp_ascb(ascb); default: break; } spin_lock_irqsave(&task->task_state_lock, flags); task->task_state_flags &= ~SAS_TASK_STATE_PENDING; task->task_state_flags &= ~SAS_TASK_AT_INITIATOR; task->task_state_flags |= SAS_TASK_STATE_DONE; if (unlikely((task->task_state_flags & SAS_TASK_STATE_ABORTED))) { struct completion *completion = ascb->completion; spin_unlock_irqrestore(&task->task_state_lock, flags); ASD_DPRINTK("task 0x%p done with opcode 0x%x resp 0x%x " "stat 0x%x but aborted by upper layer!\n", task, opcode, ts->resp, ts->stat); if (completion) complete(completion); } else { spin_unlock_irqrestore(&task->task_state_lock, flags); task->lldd_task = NULL; asd_ascb_free(ascb); mb(); task->task_done(task); } } /* ---------- ATA ---------- */ static int asd_build_ata_ascb(struct asd_ascb *ascb, struct sas_task *task, gfp_t gfp_flags) { struct domain_device *dev = task->dev; struct scb *scb; u8 flags; int res = 0; scb = ascb->scb; if (unlikely(task->ata_task.device_control_reg_update)) scb->header.opcode = CONTROL_ATA_DEV; else if (dev->sata_dev.class == ATA_DEV_ATAPI) scb->header.opcode = INITIATE_ATAPI_TASK; else scb->header.opcode = INITIATE_ATA_TASK; scb->ata_task.proto_conn_rate = (1 << 5); /* STP */ if (dev->port->oob_mode == SAS_OOB_MODE) scb->ata_task.proto_conn_rate |= dev->linkrate; scb->ata_task.total_xfer_len = cpu_to_le32(task->total_xfer_len); scb->ata_task.fis = task->ata_task.fis; if (likely(!task->ata_task.device_control_reg_update)) scb->ata_task.fis.flags |= 0x80; /* C=1: update ATA cmd reg */ scb->ata_task.fis.flags &= 0xF0; /* PM_PORT field shall be 0 */ if (dev->sata_dev.class == ATA_DEV_ATAPI) memcpy(scb->ata_task.atapi_packet, task->ata_task.atapi_packet, 16); scb->ata_task.sister_scb = cpu_to_le16(0xFFFF); scb->ata_task.conn_handle = cpu_to_le16( (u16)(unsigned long)dev->lldd_dev); if (likely(!task->ata_task.device_control_reg_update)) { flags = 0; if (task->ata_task.dma_xfer) flags |= DATA_XFER_MODE_DMA; if (task->ata_task.use_ncq && dev->sata_dev.class != ATA_DEV_ATAPI) flags |= ATA_Q_TYPE_NCQ; flags |= data_dir_flags[task->data_dir]; scb->ata_task.ata_flags = flags; scb->ata_task.retry_count = task->ata_task.retry_count; flags = 0; if (task->ata_task.set_affil_pol) flags |= SET_AFFIL_POLICY; if (task->ata_task.stp_affil_pol) flags |= STP_AFFIL_POLICY; scb->ata_task.flags = flags; } ascb->tasklet_complete = asd_task_tasklet_complete; if (likely(!task->ata_task.device_control_reg_update)) res = asd_map_scatterlist(task, scb->ata_task.sg_element, gfp_flags); return res; } static void asd_unbuild_ata_ascb(struct asd_ascb *a) { asd_unmap_scatterlist(a); } /* ---------- SMP ---------- */ static int asd_build_smp_ascb(struct asd_ascb *ascb, struct sas_task *task, gfp_t gfp_flags) { struct asd_ha_struct *asd_ha = ascb->ha; struct domain_device *dev = task->dev; struct scb *scb; dma_map_sg(&asd_ha->pcidev->dev, &task->smp_task.smp_req, 1, DMA_TO_DEVICE); dma_map_sg(&asd_ha->pcidev->dev, &task->smp_task.smp_resp, 1, DMA_FROM_DEVICE); scb = ascb->scb; scb->header.opcode = INITIATE_SMP_TASK; scb->smp_task.proto_conn_rate = dev->linkrate; scb->smp_task.smp_req.bus_addr = cpu_to_le64((u64)sg_dma_address(&task->smp_task.smp_req)); scb->smp_task.smp_req.size = cpu_to_le32((u32)sg_dma_len(&task->smp_task.smp_req)-4); scb->smp_task.smp_resp.bus_addr = cpu_to_le64((u64)sg_dma_address(&task->smp_task.smp_resp)); scb->smp_task.smp_resp.size = cpu_to_le32((u32)sg_dma_len(&task->smp_task.smp_resp)-4); scb->smp_task.sister_scb = cpu_to_le16(0xFFFF); scb->smp_task.conn_handle = cpu_to_le16((u16) (unsigned long)dev->lldd_dev); ascb->tasklet_complete = asd_task_tasklet_complete; return 0; } static void asd_unbuild_smp_ascb(struct asd_ascb *a) { struct sas_task *task = a->uldd_task; BUG_ON(!task); dma_unmap_sg(&a->ha->pcidev->dev, &task->smp_task.smp_req, 1, DMA_TO_DEVICE); dma_unmap_sg(&a->ha->pcidev->dev, &task->smp_task.smp_resp, 1, DMA_FROM_DEVICE); } /* ---------- SSP ---------- */ static int asd_build_ssp_ascb(struct asd_ascb *ascb, struct sas_task *task, gfp_t gfp_flags) { struct domain_device *dev = task->dev; struct scb *scb; int res = 0; scb = ascb->scb; scb->header.opcode = INITIATE_SSP_TASK; scb->ssp_task.proto_conn_rate = (1 << 4); /* SSP */ scb->ssp_task.proto_conn_rate |= dev->linkrate; scb->ssp_task.total_xfer_len = cpu_to_le32(task->total_xfer_len); scb->ssp_task.ssp_frame.frame_type = SSP_DATA; memcpy(scb->ssp_task.ssp_frame.hashed_dest_addr, dev->hashed_sas_addr, HASHED_SAS_ADDR_SIZE); memcpy(scb->ssp_task.ssp_frame.hashed_src_addr, dev->port->ha->hashed_sas_addr, HASHED_SAS_ADDR_SIZE); scb->ssp_task.ssp_frame.tptt = cpu_to_be16(0xFFFF); memcpy(scb->ssp_task.ssp_cmd.lun, task->ssp_task.LUN, 8); if (task->ssp_task.enable_first_burst) scb->ssp_task.ssp_cmd.efb_prio_attr |= EFB_MASK; scb->ssp_task.ssp_cmd.efb_prio_attr |= (task->ssp_task.task_prio << 3); scb->ssp_task.ssp_cmd.efb_prio_attr |= (task->ssp_task.task_attr & 7); memcpy(scb->ssp_task.ssp_cmd.cdb, task->ssp_task.cmd->cmnd, task->ssp_task.cmd->cmd_len); scb->ssp_task.sister_scb = cpu_to_le16(0xFFFF); scb->ssp_task.conn_handle = cpu_to_le16( (u16)(unsigned long)dev->lldd_dev); scb->ssp_task.data_dir = data_dir_flags[task->data_dir]; scb->ssp_task.retry_count = scb->ssp_task.retry_count; ascb->tasklet_complete = asd_task_tasklet_complete; res = asd_map_scatterlist(task, scb->ssp_task.sg_element, gfp_flags); return res; } static void asd_unbuild_ssp_ascb(struct asd_ascb *a) { asd_unmap_scatterlist(a); } /* ---------- Execute Task ---------- */ static int asd_can_queue(struct asd_ha_struct *asd_ha, int num) { int res = 0; unsigned long flags; spin_lock_irqsave(&asd_ha->seq.pend_q_lock, flags); if ((asd_ha->seq.can_queue - num) < 0) res = -SAS_QUEUE_FULL; else asd_ha->seq.can_queue -= num; spin_unlock_irqrestore(&asd_ha->seq.pend_q_lock, flags); return res; } int asd_execute_task(struct sas_task *task, gfp_t gfp_flags) { int res = 0; LIST_HEAD(alist); struct sas_task *t = task; struct asd_ascb *ascb = NULL, *a; struct asd_ha_struct *asd_ha = task->dev->port->ha->lldd_ha; unsigned long flags; res = asd_can_queue(asd_ha, 1); if (res) return res; res = 1; ascb = asd_ascb_alloc_list(asd_ha, &res, gfp_flags); if (res) { res = -ENOMEM; goto out_err; } __list_add(&alist, ascb->list.prev, &ascb->list); list_for_each_entry(a, &alist, list) { a->uldd_task = t; t->lldd_task = a; break; } list_for_each_entry(a, &alist, list) { t = a->uldd_task; a->uldd_timer = 1; if (t->task_proto & SAS_PROTOCOL_STP) t->task_proto = SAS_PROTOCOL_STP; switch (t->task_proto) { case SAS_PROTOCOL_SATA: case SAS_PROTOCOL_STP: res = asd_build_ata_ascb(a, t, gfp_flags); break; case SAS_PROTOCOL_SMP: res = asd_build_smp_ascb(a, t, gfp_flags); break; case SAS_PROTOCOL_SSP: res = asd_build_ssp_ascb(a, t, gfp_flags); break; default: asd_printk("unknown sas_task proto: 0x%x\n", t->task_proto); res = -ENOMEM; break; } if (res) goto out_err_unmap; spin_lock_irqsave(&t->task_state_lock, flags); t->task_state_flags |= SAS_TASK_AT_INITIATOR; spin_unlock_irqrestore(&t->task_state_lock, flags); } list_del_init(&alist); res = asd_post_ascb_list(asd_ha, ascb, 1); if (unlikely(res)) { a = NULL; __list_add(&alist, ascb->list.prev, &ascb->list); goto out_err_unmap; } return 0; out_err_unmap: { struct asd_ascb *b = a; list_for_each_entry(a, &alist, list) { if (a == b) break; t = a->uldd_task; spin_lock_irqsave(&t->task_state_lock, flags); t->task_state_flags &= ~SAS_TASK_AT_INITIATOR; spin_unlock_irqrestore(&t->task_state_lock, flags); switch (t->task_proto) { case SAS_PROTOCOL_SATA: case SAS_PROTOCOL_STP: asd_unbuild_ata_ascb(a); break; case SAS_PROTOCOL_SMP: asd_unbuild_smp_ascb(a); break; case SAS_PROTOCOL_SSP: asd_unbuild_ssp_ascb(a); default: break; } t->lldd_task = NULL; } } list_del_init(&alist); out_err: if (ascb) asd_ascb_free_list(ascb); asd_can_dequeue(asd_ha, 1); return res; }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1