Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Cheng Xu | 3341 | 100.00% | 7 | 100.00% |
Total | 3341 | 7 |
// SPDX-License-Identifier: GPL-2.0 OR BSD-3-Clause /* Authors: Cheng Xu <chengyou@linux.alibaba.com> */ /* Kai Shen <kaishen@linux.alibaba.com> */ /* Copyright (c) 2020-2021, Alibaba Group */ /* Authors: Bernard Metzler <bmt@zurich.ibm.com> */ /* Copyright (c) 2008-2019, IBM Corporation */ #include "erdma_cm.h" #include "erdma_verbs.h" void erdma_qp_llp_close(struct erdma_qp *qp) { struct erdma_qp_attrs qp_attrs; down_write(&qp->state_lock); switch (qp->attrs.state) { case ERDMA_QP_STATE_RTS: case ERDMA_QP_STATE_RTR: case ERDMA_QP_STATE_IDLE: case ERDMA_QP_STATE_TERMINATE: qp_attrs.state = ERDMA_QP_STATE_CLOSING; erdma_modify_qp_internal(qp, &qp_attrs, ERDMA_QP_ATTR_STATE); break; case ERDMA_QP_STATE_CLOSING: qp->attrs.state = ERDMA_QP_STATE_IDLE; break; default: break; } if (qp->cep) { erdma_cep_put(qp->cep); qp->cep = NULL; } up_write(&qp->state_lock); } struct ib_qp *erdma_get_ibqp(struct ib_device *ibdev, int id) { struct erdma_qp *qp = find_qp_by_qpn(to_edev(ibdev), id); if (qp) return &qp->ibqp; return NULL; } static int erdma_modify_qp_state_to_rts(struct erdma_qp *qp, struct erdma_qp_attrs *attrs, enum erdma_qp_attr_mask mask) { int ret; struct erdma_dev *dev = qp->dev; struct erdma_cmdq_modify_qp_req req; struct tcp_sock *tp; struct erdma_cep *cep = qp->cep; struct sockaddr_storage local_addr, remote_addr; if (!(mask & ERDMA_QP_ATTR_LLP_HANDLE)) return -EINVAL; if (!(mask & ERDMA_QP_ATTR_MPA)) return -EINVAL; ret = getname_local(cep->sock, &local_addr); if (ret < 0) return ret; ret = getname_peer(cep->sock, &remote_addr); if (ret < 0) return ret; qp->attrs.state = ERDMA_QP_STATE_RTS; tp = tcp_sk(qp->cep->sock->sk); erdma_cmdq_build_reqhdr(&req.hdr, CMDQ_SUBMOD_RDMA, CMDQ_OPCODE_MODIFY_QP); req.cfg = FIELD_PREP(ERDMA_CMD_MODIFY_QP_STATE_MASK, qp->attrs.state) | FIELD_PREP(ERDMA_CMD_MODIFY_QP_CC_MASK, qp->attrs.cc) | FIELD_PREP(ERDMA_CMD_MODIFY_QP_QPN_MASK, QP_ID(qp)); req.cookie = be32_to_cpu(qp->cep->mpa.ext_data.cookie); req.dip = to_sockaddr_in(remote_addr).sin_addr.s_addr; req.sip = to_sockaddr_in(local_addr).sin_addr.s_addr; req.dport = to_sockaddr_in(remote_addr).sin_port; req.sport = to_sockaddr_in(local_addr).sin_port; req.send_nxt = tp->snd_nxt; /* rsvd tcp seq for mpa-rsp in server. */ if (qp->attrs.qp_type == ERDMA_QP_PASSIVE) req.send_nxt += MPA_DEFAULT_HDR_LEN + qp->attrs.pd_len; req.recv_nxt = tp->rcv_nxt; return erdma_post_cmd_wait(&dev->cmdq, &req, sizeof(req), NULL, NULL); } static int erdma_modify_qp_state_to_stop(struct erdma_qp *qp, struct erdma_qp_attrs *attrs, enum erdma_qp_attr_mask mask) { struct erdma_dev *dev = qp->dev; struct erdma_cmdq_modify_qp_req req; qp->attrs.state = attrs->state; erdma_cmdq_build_reqhdr(&req.hdr, CMDQ_SUBMOD_RDMA, CMDQ_OPCODE_MODIFY_QP); req.cfg = FIELD_PREP(ERDMA_CMD_MODIFY_QP_STATE_MASK, attrs->state) | FIELD_PREP(ERDMA_CMD_MODIFY_QP_QPN_MASK, QP_ID(qp)); return erdma_post_cmd_wait(&dev->cmdq, &req, sizeof(req), NULL, NULL); } int erdma_modify_qp_internal(struct erdma_qp *qp, struct erdma_qp_attrs *attrs, enum erdma_qp_attr_mask mask) { bool need_reflush = false; int drop_conn, ret = 0; if (!mask) return 0; if (!(mask & ERDMA_QP_ATTR_STATE)) return 0; switch (qp->attrs.state) { case ERDMA_QP_STATE_IDLE: case ERDMA_QP_STATE_RTR: if (attrs->state == ERDMA_QP_STATE_RTS) { ret = erdma_modify_qp_state_to_rts(qp, attrs, mask); } else if (attrs->state == ERDMA_QP_STATE_ERROR) { qp->attrs.state = ERDMA_QP_STATE_ERROR; need_reflush = true; if (qp->cep) { erdma_cep_put(qp->cep); qp->cep = NULL; } ret = erdma_modify_qp_state_to_stop(qp, attrs, mask); } break; case ERDMA_QP_STATE_RTS: drop_conn = 0; if (attrs->state == ERDMA_QP_STATE_CLOSING || attrs->state == ERDMA_QP_STATE_TERMINATE || attrs->state == ERDMA_QP_STATE_ERROR) { ret = erdma_modify_qp_state_to_stop(qp, attrs, mask); drop_conn = 1; need_reflush = true; } if (drop_conn) erdma_qp_cm_drop(qp); break; case ERDMA_QP_STATE_TERMINATE: if (attrs->state == ERDMA_QP_STATE_ERROR) qp->attrs.state = ERDMA_QP_STATE_ERROR; break; case ERDMA_QP_STATE_CLOSING: if (attrs->state == ERDMA_QP_STATE_IDLE) { qp->attrs.state = ERDMA_QP_STATE_IDLE; } else if (attrs->state == ERDMA_QP_STATE_ERROR) { ret = erdma_modify_qp_state_to_stop(qp, attrs, mask); qp->attrs.state = ERDMA_QP_STATE_ERROR; } else if (attrs->state != ERDMA_QP_STATE_CLOSING) { return -ECONNABORTED; } break; default: break; } if (need_reflush && !ret && rdma_is_kernel_res(&qp->ibqp.res)) { qp->flags |= ERDMA_QP_IN_FLUSHING; mod_delayed_work(qp->dev->reflush_wq, &qp->reflush_dwork, usecs_to_jiffies(100)); } return ret; } static void erdma_qp_safe_free(struct kref *ref) { struct erdma_qp *qp = container_of(ref, struct erdma_qp, ref); complete(&qp->safe_free); } void erdma_qp_put(struct erdma_qp *qp) { WARN_ON(kref_read(&qp->ref) < 1); kref_put(&qp->ref, erdma_qp_safe_free); } void erdma_qp_get(struct erdma_qp *qp) { kref_get(&qp->ref); } static int fill_inline_data(struct erdma_qp *qp, const struct ib_send_wr *send_wr, u16 wqe_idx, u32 sgl_offset, __le32 *length_field) { u32 remain_size, copy_size, data_off, bytes = 0; char *data; int i = 0; wqe_idx += (sgl_offset >> SQEBB_SHIFT); sgl_offset &= (SQEBB_SIZE - 1); data = get_queue_entry(qp->kern_qp.sq_buf, wqe_idx, qp->attrs.sq_size, SQEBB_SHIFT); while (i < send_wr->num_sge) { bytes += send_wr->sg_list[i].length; if (bytes > (int)ERDMA_MAX_INLINE) return -EINVAL; remain_size = send_wr->sg_list[i].length; data_off = 0; while (1) { copy_size = min(remain_size, SQEBB_SIZE - sgl_offset); memcpy(data + sgl_offset, (void *)(uintptr_t)send_wr->sg_list[i].addr + data_off, copy_size); remain_size -= copy_size; data_off += copy_size; sgl_offset += copy_size; wqe_idx += (sgl_offset >> SQEBB_SHIFT); sgl_offset &= (SQEBB_SIZE - 1); data = get_queue_entry(qp->kern_qp.sq_buf, wqe_idx, qp->attrs.sq_size, SQEBB_SHIFT); if (!remain_size) break; } i++; } *length_field = cpu_to_le32(bytes); return bytes; } static int fill_sgl(struct erdma_qp *qp, const struct ib_send_wr *send_wr, u16 wqe_idx, u32 sgl_offset, __le32 *length_field) { int i = 0; u32 bytes = 0; char *sgl; if (send_wr->num_sge > qp->dev->attrs.max_send_sge) return -EINVAL; if (sgl_offset & 0xF) return -EINVAL; while (i < send_wr->num_sge) { wqe_idx += (sgl_offset >> SQEBB_SHIFT); sgl_offset &= (SQEBB_SIZE - 1); sgl = get_queue_entry(qp->kern_qp.sq_buf, wqe_idx, qp->attrs.sq_size, SQEBB_SHIFT); bytes += send_wr->sg_list[i].length; memcpy(sgl + sgl_offset, &send_wr->sg_list[i], sizeof(struct ib_sge)); sgl_offset += sizeof(struct ib_sge); i++; } *length_field = cpu_to_le32(bytes); return 0; } static int erdma_push_one_sqe(struct erdma_qp *qp, u16 *pi, const struct ib_send_wr *send_wr) { u32 wqe_size, wqebb_cnt, hw_op, flags, sgl_offset; u32 idx = *pi & (qp->attrs.sq_size - 1); enum ib_wr_opcode op = send_wr->opcode; struct erdma_atomic_sqe *atomic_sqe; struct erdma_readreq_sqe *read_sqe; struct erdma_reg_mr_sqe *regmr_sge; struct erdma_write_sqe *write_sqe; struct erdma_send_sqe *send_sqe; struct ib_rdma_wr *rdma_wr; struct erdma_sge *sge; __le32 *length_field; struct erdma_mr *mr; u64 wqe_hdr, *entry; u32 attrs; int ret; entry = get_queue_entry(qp->kern_qp.sq_buf, idx, qp->attrs.sq_size, SQEBB_SHIFT); /* Clear the SQE header section. */ *entry = 0; qp->kern_qp.swr_tbl[idx] = send_wr->wr_id; flags = send_wr->send_flags; wqe_hdr = FIELD_PREP( ERDMA_SQE_HDR_CE_MASK, ((flags & IB_SEND_SIGNALED) || qp->kern_qp.sig_all) ? 1 : 0); wqe_hdr |= FIELD_PREP(ERDMA_SQE_HDR_SE_MASK, flags & IB_SEND_SOLICITED ? 1 : 0); wqe_hdr |= FIELD_PREP(ERDMA_SQE_HDR_FENCE_MASK, flags & IB_SEND_FENCE ? 1 : 0); wqe_hdr |= FIELD_PREP(ERDMA_SQE_HDR_INLINE_MASK, flags & IB_SEND_INLINE ? 1 : 0); wqe_hdr |= FIELD_PREP(ERDMA_SQE_HDR_QPN_MASK, QP_ID(qp)); switch (op) { case IB_WR_RDMA_WRITE: case IB_WR_RDMA_WRITE_WITH_IMM: hw_op = ERDMA_OP_WRITE; if (op == IB_WR_RDMA_WRITE_WITH_IMM) hw_op = ERDMA_OP_WRITE_WITH_IMM; wqe_hdr |= FIELD_PREP(ERDMA_SQE_HDR_OPCODE_MASK, hw_op); rdma_wr = container_of(send_wr, struct ib_rdma_wr, wr); write_sqe = (struct erdma_write_sqe *)entry; write_sqe->imm_data = send_wr->ex.imm_data; write_sqe->sink_stag = cpu_to_le32(rdma_wr->rkey); write_sqe->sink_to_h = cpu_to_le32(upper_32_bits(rdma_wr->remote_addr)); write_sqe->sink_to_l = cpu_to_le32(lower_32_bits(rdma_wr->remote_addr)); length_field = &write_sqe->length; wqe_size = sizeof(struct erdma_write_sqe); sgl_offset = wqe_size; break; case IB_WR_RDMA_READ: case IB_WR_RDMA_READ_WITH_INV: read_sqe = (struct erdma_readreq_sqe *)entry; if (unlikely(send_wr->num_sge != 1)) return -EINVAL; hw_op = ERDMA_OP_READ; if (op == IB_WR_RDMA_READ_WITH_INV) { hw_op = ERDMA_OP_READ_WITH_INV; read_sqe->invalid_stag = cpu_to_le32(send_wr->ex.invalidate_rkey); } wqe_hdr |= FIELD_PREP(ERDMA_SQE_HDR_OPCODE_MASK, hw_op); rdma_wr = container_of(send_wr, struct ib_rdma_wr, wr); read_sqe->length = cpu_to_le32(send_wr->sg_list[0].length); read_sqe->sink_stag = cpu_to_le32(send_wr->sg_list[0].lkey); read_sqe->sink_to_l = cpu_to_le32(lower_32_bits(send_wr->sg_list[0].addr)); read_sqe->sink_to_h = cpu_to_le32(upper_32_bits(send_wr->sg_list[0].addr)); sge = get_queue_entry(qp->kern_qp.sq_buf, idx + 1, qp->attrs.sq_size, SQEBB_SHIFT); sge->addr = cpu_to_le64(rdma_wr->remote_addr); sge->key = cpu_to_le32(rdma_wr->rkey); sge->length = cpu_to_le32(send_wr->sg_list[0].length); wqe_size = sizeof(struct erdma_readreq_sqe) + send_wr->num_sge * sizeof(struct ib_sge); goto out; case IB_WR_SEND: case IB_WR_SEND_WITH_IMM: case IB_WR_SEND_WITH_INV: send_sqe = (struct erdma_send_sqe *)entry; hw_op = ERDMA_OP_SEND; if (op == IB_WR_SEND_WITH_IMM) { hw_op = ERDMA_OP_SEND_WITH_IMM; send_sqe->imm_data = send_wr->ex.imm_data; } else if (op == IB_WR_SEND_WITH_INV) { hw_op = ERDMA_OP_SEND_WITH_INV; send_sqe->invalid_stag = cpu_to_le32(send_wr->ex.invalidate_rkey); } wqe_hdr |= FIELD_PREP(ERDMA_SQE_HDR_OPCODE_MASK, hw_op); length_field = &send_sqe->length; wqe_size = sizeof(struct erdma_send_sqe); sgl_offset = wqe_size; break; case IB_WR_REG_MR: wqe_hdr |= FIELD_PREP(ERDMA_SQE_HDR_OPCODE_MASK, ERDMA_OP_REG_MR); regmr_sge = (struct erdma_reg_mr_sqe *)entry; mr = to_emr(reg_wr(send_wr)->mr); mr->access = ERDMA_MR_ACC_LR | to_erdma_access_flags(reg_wr(send_wr)->access); regmr_sge->addr = cpu_to_le64(mr->ibmr.iova); regmr_sge->length = cpu_to_le32(mr->ibmr.length); regmr_sge->stag = cpu_to_le32(reg_wr(send_wr)->key); attrs = FIELD_PREP(ERDMA_SQE_MR_ACCESS_MASK, mr->access) | FIELD_PREP(ERDMA_SQE_MR_MTT_CNT_MASK, mr->mem.mtt_nents); if (mr->mem.mtt_nents <= ERDMA_MAX_INLINE_MTT_ENTRIES) { attrs |= FIELD_PREP(ERDMA_SQE_MR_MTT_TYPE_MASK, 0); /* Copy SGLs to SQE content to accelerate */ memcpy(get_queue_entry(qp->kern_qp.sq_buf, idx + 1, qp->attrs.sq_size, SQEBB_SHIFT), mr->mem.mtt->buf, MTT_SIZE(mr->mem.mtt_nents)); wqe_size = sizeof(struct erdma_reg_mr_sqe) + MTT_SIZE(mr->mem.mtt_nents); } else { attrs |= FIELD_PREP(ERDMA_SQE_MR_MTT_TYPE_MASK, 1); wqe_size = sizeof(struct erdma_reg_mr_sqe); } regmr_sge->attrs = cpu_to_le32(attrs); goto out; case IB_WR_LOCAL_INV: wqe_hdr |= FIELD_PREP(ERDMA_SQE_HDR_OPCODE_MASK, ERDMA_OP_LOCAL_INV); regmr_sge = (struct erdma_reg_mr_sqe *)entry; regmr_sge->stag = cpu_to_le32(send_wr->ex.invalidate_rkey); wqe_size = sizeof(struct erdma_reg_mr_sqe); goto out; case IB_WR_ATOMIC_CMP_AND_SWP: case IB_WR_ATOMIC_FETCH_AND_ADD: atomic_sqe = (struct erdma_atomic_sqe *)entry; if (op == IB_WR_ATOMIC_CMP_AND_SWP) { wqe_hdr |= FIELD_PREP(ERDMA_SQE_HDR_OPCODE_MASK, ERDMA_OP_ATOMIC_CAS); atomic_sqe->fetchadd_swap_data = cpu_to_le64(atomic_wr(send_wr)->swap); atomic_sqe->cmp_data = cpu_to_le64(atomic_wr(send_wr)->compare_add); } else { wqe_hdr |= FIELD_PREP(ERDMA_SQE_HDR_OPCODE_MASK, ERDMA_OP_ATOMIC_FAA); atomic_sqe->fetchadd_swap_data = cpu_to_le64(atomic_wr(send_wr)->compare_add); } sge = get_queue_entry(qp->kern_qp.sq_buf, idx + 1, qp->attrs.sq_size, SQEBB_SHIFT); sge->addr = cpu_to_le64(atomic_wr(send_wr)->remote_addr); sge->key = cpu_to_le32(atomic_wr(send_wr)->rkey); sge++; sge->addr = cpu_to_le64(send_wr->sg_list[0].addr); sge->key = cpu_to_le32(send_wr->sg_list[0].lkey); sge->length = cpu_to_le32(send_wr->sg_list[0].length); wqe_size = sizeof(*atomic_sqe); goto out; default: return -EOPNOTSUPP; } if (flags & IB_SEND_INLINE) { ret = fill_inline_data(qp, send_wr, idx, sgl_offset, length_field); if (ret < 0) return -EINVAL; wqe_size += ret; wqe_hdr |= FIELD_PREP(ERDMA_SQE_HDR_SGL_LEN_MASK, ret); } else { ret = fill_sgl(qp, send_wr, idx, sgl_offset, length_field); if (ret) return -EINVAL; wqe_size += send_wr->num_sge * sizeof(struct ib_sge); wqe_hdr |= FIELD_PREP(ERDMA_SQE_HDR_SGL_LEN_MASK, send_wr->num_sge); } out: wqebb_cnt = SQEBB_COUNT(wqe_size); wqe_hdr |= FIELD_PREP(ERDMA_SQE_HDR_WQEBB_CNT_MASK, wqebb_cnt - 1); *pi += wqebb_cnt; wqe_hdr |= FIELD_PREP(ERDMA_SQE_HDR_WQEBB_INDEX_MASK, *pi); *entry = wqe_hdr; return 0; } static void kick_sq_db(struct erdma_qp *qp, u16 pi) { u64 db_data = FIELD_PREP(ERDMA_SQE_HDR_QPN_MASK, QP_ID(qp)) | FIELD_PREP(ERDMA_SQE_HDR_WQEBB_INDEX_MASK, pi); *(u64 *)qp->kern_qp.sq_db_info = db_data; writeq(db_data, qp->kern_qp.hw_sq_db); } int erdma_post_send(struct ib_qp *ibqp, const struct ib_send_wr *send_wr, const struct ib_send_wr **bad_send_wr) { struct erdma_qp *qp = to_eqp(ibqp); int ret = 0; const struct ib_send_wr *wr = send_wr; unsigned long flags; u16 sq_pi; if (!send_wr) return -EINVAL; spin_lock_irqsave(&qp->lock, flags); sq_pi = qp->kern_qp.sq_pi; while (wr) { if ((u16)(sq_pi - qp->kern_qp.sq_ci) >= qp->attrs.sq_size) { ret = -ENOMEM; *bad_send_wr = send_wr; break; } ret = erdma_push_one_sqe(qp, &sq_pi, wr); if (ret) { *bad_send_wr = wr; break; } qp->kern_qp.sq_pi = sq_pi; kick_sq_db(qp, sq_pi); wr = wr->next; } spin_unlock_irqrestore(&qp->lock, flags); if (unlikely(qp->flags & ERDMA_QP_IN_FLUSHING)) mod_delayed_work(qp->dev->reflush_wq, &qp->reflush_dwork, usecs_to_jiffies(100)); return ret; } static int erdma_post_recv_one(struct erdma_qp *qp, const struct ib_recv_wr *recv_wr) { struct erdma_rqe *rqe = get_queue_entry(qp->kern_qp.rq_buf, qp->kern_qp.rq_pi, qp->attrs.rq_size, RQE_SHIFT); rqe->qe_idx = cpu_to_le16(qp->kern_qp.rq_pi + 1); rqe->qpn = cpu_to_le32(QP_ID(qp)); if (recv_wr->num_sge == 0) { rqe->length = 0; } else if (recv_wr->num_sge == 1) { rqe->stag = cpu_to_le32(recv_wr->sg_list[0].lkey); rqe->to = cpu_to_le64(recv_wr->sg_list[0].addr); rqe->length = cpu_to_le32(recv_wr->sg_list[0].length); } else { return -EINVAL; } *(u64 *)qp->kern_qp.rq_db_info = *(u64 *)rqe; writeq(*(u64 *)rqe, qp->kern_qp.hw_rq_db); qp->kern_qp.rwr_tbl[qp->kern_qp.rq_pi & (qp->attrs.rq_size - 1)] = recv_wr->wr_id; qp->kern_qp.rq_pi++; return 0; } int erdma_post_recv(struct ib_qp *ibqp, const struct ib_recv_wr *recv_wr, const struct ib_recv_wr **bad_recv_wr) { const struct ib_recv_wr *wr = recv_wr; struct erdma_qp *qp = to_eqp(ibqp); unsigned long flags; int ret; spin_lock_irqsave(&qp->lock, flags); while (wr) { ret = erdma_post_recv_one(qp, wr); if (ret) { *bad_recv_wr = wr; break; } wr = wr->next; } spin_unlock_irqrestore(&qp->lock, flags); if (unlikely(qp->flags & ERDMA_QP_IN_FLUSHING)) mod_delayed_work(qp->dev->reflush_wq, &qp->reflush_dwork, usecs_to_jiffies(100)); return ret; }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1