Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Lijun Ou | 434 | 40.67% | 2 | 11.11% |
Wenpeng Liang | 430 | 40.30% | 7 | 38.89% |
wenglianfa | 133 | 12.46% | 2 | 11.11% |
Wei Hu (Xavier) | 38 | 3.56% | 4 | 22.22% |
Xi Wang | 12 | 1.12% | 1 | 5.56% |
Chengchang Tang | 12 | 1.12% | 1 | 5.56% |
Shaobo Xu | 8 | 0.75% | 1 | 5.56% |
Total | 1067 | 18 |
// SPDX-License-Identifier: (GPL-2.0 OR BSD-2-Clause) // Copyright (c) 2019 Hisilicon Limited. #include <rdma/rdma_cm.h> #include <rdma/restrack.h> #include <uapi/rdma/rdma_netlink.h> #include "hnae3.h" #include "hns_roce_common.h" #include "hns_roce_device.h" #include "hns_roce_hw_v2.h" int hns_roce_fill_res_cq_entry(struct sk_buff *msg, struct ib_cq *ib_cq) { struct hns_roce_cq *hr_cq = to_hr_cq(ib_cq); struct nlattr *table_attr; table_attr = nla_nest_start(msg, RDMA_NLDEV_ATTR_DRIVER); if (!table_attr) return -EMSGSIZE; if (rdma_nl_put_driver_u32(msg, "cq_depth", hr_cq->cq_depth)) goto err; if (rdma_nl_put_driver_u32(msg, "cons_index", hr_cq->cons_index)) goto err; if (rdma_nl_put_driver_u32(msg, "cqe_size", hr_cq->cqe_size)) goto err; if (rdma_nl_put_driver_u32(msg, "arm_sn", hr_cq->arm_sn)) goto err; nla_nest_end(msg, table_attr); return 0; err: nla_nest_cancel(msg, table_attr); return -EMSGSIZE; } int hns_roce_fill_res_cq_entry_raw(struct sk_buff *msg, struct ib_cq *ib_cq) { struct hns_roce_dev *hr_dev = to_hr_dev(ib_cq->device); struct hns_roce_cq *hr_cq = to_hr_cq(ib_cq); struct hns_roce_v2_cq_context context; int ret; if (!hr_dev->hw->query_cqc) return -EINVAL; ret = hr_dev->hw->query_cqc(hr_dev, hr_cq->cqn, &context); if (ret) return -EINVAL; ret = nla_put(msg, RDMA_NLDEV_ATTR_RES_RAW, sizeof(context), &context); return ret; } int hns_roce_fill_res_qp_entry(struct sk_buff *msg, struct ib_qp *ib_qp) { struct hns_roce_qp *hr_qp = to_hr_qp(ib_qp); struct nlattr *table_attr; table_attr = nla_nest_start(msg, RDMA_NLDEV_ATTR_DRIVER); if (!table_attr) return -EMSGSIZE; if (rdma_nl_put_driver_u32_hex(msg, "sq_wqe_cnt", hr_qp->sq.wqe_cnt)) goto err; if (rdma_nl_put_driver_u32_hex(msg, "sq_max_gs", hr_qp->sq.max_gs)) goto err; if (rdma_nl_put_driver_u32_hex(msg, "rq_wqe_cnt", hr_qp->rq.wqe_cnt)) goto err; if (rdma_nl_put_driver_u32_hex(msg, "rq_max_gs", hr_qp->rq.max_gs)) goto err; if (rdma_nl_put_driver_u32_hex(msg, "ext_sge_sge_cnt", hr_qp->sge.sge_cnt)) goto err; nla_nest_end(msg, table_attr); return 0; err: nla_nest_cancel(msg, table_attr); return -EMSGSIZE; } int hns_roce_fill_res_qp_entry_raw(struct sk_buff *msg, struct ib_qp *ib_qp) { struct hns_roce_dev *hr_dev = to_hr_dev(ib_qp->device); struct hns_roce_qp *hr_qp = to_hr_qp(ib_qp); struct hns_roce_full_qp_ctx { struct hns_roce_v2_qp_context qpc; struct hns_roce_v2_scc_context sccc; } context = {}; int ret; if (!hr_dev->hw->query_qpc) return -EINVAL; ret = hr_dev->hw->query_qpc(hr_dev, hr_qp->qpn, &context.qpc); if (ret) return ret; /* If SCC is disabled or the query fails, the queried SCCC will * be all 0. */ if (!(hr_dev->caps.flags & HNS_ROCE_CAP_FLAG_QP_FLOW_CTRL) || !hr_dev->hw->query_sccc) goto out; ret = hr_dev->hw->query_sccc(hr_dev, hr_qp->qpn, &context.sccc); if (ret) ibdev_warn_ratelimited(&hr_dev->ib_dev, "failed to query SCCC, ret = %d.\n", ret); out: ret = nla_put(msg, RDMA_NLDEV_ATTR_RES_RAW, sizeof(context), &context); return ret; } int hns_roce_fill_res_mr_entry(struct sk_buff *msg, struct ib_mr *ib_mr) { struct hns_roce_mr *hr_mr = to_hr_mr(ib_mr); struct nlattr *table_attr; table_attr = nla_nest_start(msg, RDMA_NLDEV_ATTR_DRIVER); if (!table_attr) return -EMSGSIZE; if (rdma_nl_put_driver_u32_hex(msg, "pbl_hop_num", hr_mr->pbl_hop_num)) goto err; if (rdma_nl_put_driver_u32_hex(msg, "ba_pg_shift", hr_mr->pbl_mtr.hem_cfg.ba_pg_shift)) goto err; if (rdma_nl_put_driver_u32_hex(msg, "buf_pg_shift", hr_mr->pbl_mtr.hem_cfg.buf_pg_shift)) goto err; nla_nest_end(msg, table_attr); return 0; err: nla_nest_cancel(msg, table_attr); return -EMSGSIZE; } int hns_roce_fill_res_mr_entry_raw(struct sk_buff *msg, struct ib_mr *ib_mr) { struct hns_roce_dev *hr_dev = to_hr_dev(ib_mr->device); struct hns_roce_mr *hr_mr = to_hr_mr(ib_mr); struct hns_roce_v2_mpt_entry context; int ret; if (!hr_dev->hw->query_mpt) return -EINVAL; ret = hr_dev->hw->query_mpt(hr_dev, hr_mr->key, &context); if (ret) return -EINVAL; ret = nla_put(msg, RDMA_NLDEV_ATTR_RES_RAW, sizeof(context), &context); return ret; } int hns_roce_fill_res_srq_entry(struct sk_buff *msg, struct ib_srq *ib_srq) { struct hns_roce_srq *hr_srq = to_hr_srq(ib_srq); struct nlattr *table_attr; table_attr = nla_nest_start(msg, RDMA_NLDEV_ATTR_DRIVER); if (!table_attr) return -EMSGSIZE; if (rdma_nl_put_driver_u32_hex(msg, "srqn", hr_srq->srqn)) goto err; if (rdma_nl_put_driver_u32_hex(msg, "wqe_cnt", hr_srq->wqe_cnt)) goto err; if (rdma_nl_put_driver_u32_hex(msg, "max_gs", hr_srq->max_gs)) goto err; if (rdma_nl_put_driver_u32_hex(msg, "xrcdn", hr_srq->xrcdn)) goto err; nla_nest_end(msg, table_attr); return 0; err: nla_nest_cancel(msg, table_attr); return -EMSGSIZE; } int hns_roce_fill_res_srq_entry_raw(struct sk_buff *msg, struct ib_srq *ib_srq) { struct hns_roce_dev *hr_dev = to_hr_dev(ib_srq->device); struct hns_roce_srq *hr_srq = to_hr_srq(ib_srq); struct hns_roce_srq_context context; int ret; if (!hr_dev->hw->query_srqc) return -EINVAL; ret = hr_dev->hw->query_srqc(hr_dev, hr_srq->srqn, &context); if (ret) return ret; ret = nla_put(msg, RDMA_NLDEV_ATTR_RES_RAW, sizeof(context), &context); return ret; }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1