Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Gal Pressman | 721 | 91.73% | 8 | 47.06% |
Michal Kalderon | 35 | 4.45% | 1 | 5.88% |
Jason Gunthorpe | 12 | 1.53% | 1 | 5.88% |
Leon Romanovsky | 9 | 1.15% | 5 | 29.41% |
Mark Bloch | 7 | 0.89% | 1 | 5.88% |
Maor Gottlieb | 2 | 0.25% | 1 | 5.88% |
Total | 786 | 17 |
/* SPDX-License-Identifier: GPL-2.0 OR BSD-2-Clause */ /* * Copyright 2018-2021 Amazon.com, Inc. or its affiliates. All rights reserved. */ #ifndef _EFA_H_ #define _EFA_H_ #include <linux/bitops.h> #include <linux/interrupt.h> #include <linux/pci.h> #include <rdma/efa-abi.h> #include <rdma/ib_verbs.h> #include "efa_com_cmd.h" #define DRV_MODULE_NAME "efa" #define DEVICE_NAME "Elastic Fabric Adapter (EFA)" #define EFA_IRQNAME_SIZE 40 #define EFA_MGMNT_MSIX_VEC_IDX 0 #define EFA_COMP_EQS_VEC_BASE 1 struct efa_irq { irq_handler_t handler; void *data; u32 irqn; u32 vector; cpumask_t affinity_hint_mask; char name[EFA_IRQNAME_SIZE]; }; /* Don't use anything other than atomic64 */ struct efa_stats { atomic64_t alloc_pd_err; atomic64_t create_qp_err; atomic64_t create_cq_err; atomic64_t reg_mr_err; atomic64_t alloc_ucontext_err; atomic64_t create_ah_err; atomic64_t mmap_err; atomic64_t keep_alive_rcvd; }; struct efa_dev { struct ib_device ibdev; struct efa_com_dev edev; struct pci_dev *pdev; struct efa_com_get_device_attr_result dev_attr; u64 reg_bar_addr; u64 reg_bar_len; u64 mem_bar_addr; u64 mem_bar_len; u64 db_bar_addr; u64 db_bar_len; int admin_msix_vector_idx; struct efa_irq admin_irq; struct efa_stats stats; /* Array of completion EQs */ struct efa_eq *eqs; unsigned int neqs; /* Only stores CQs with interrupts enabled */ struct xarray cqs_xa; }; struct efa_ucontext { struct ib_ucontext ibucontext; u16 uarn; }; struct efa_pd { struct ib_pd ibpd; u16 pdn; }; struct efa_mr { struct ib_mr ibmr; struct ib_umem *umem; }; struct efa_cq { struct ib_cq ibcq; struct efa_ucontext *ucontext; dma_addr_t dma_addr; void *cpu_addr; struct rdma_user_mmap_entry *mmap_entry; struct rdma_user_mmap_entry *db_mmap_entry; size_t size; u16 cq_idx; /* NULL when no interrupts requested */ struct efa_eq *eq; }; struct efa_qp { struct ib_qp ibqp; dma_addr_t rq_dma_addr; void *rq_cpu_addr; size_t rq_size; enum ib_qp_state state; /* Used for saving mmap_xa entries */ struct rdma_user_mmap_entry *sq_db_mmap_entry; struct rdma_user_mmap_entry *llq_desc_mmap_entry; struct rdma_user_mmap_entry *rq_db_mmap_entry; struct rdma_user_mmap_entry *rq_mmap_entry; u32 qp_handle; u32 max_send_wr; u32 max_recv_wr; u32 max_send_sge; u32 max_recv_sge; u32 max_inline_data; }; struct efa_ah { struct ib_ah ibah; u16 ah; /* dest_addr */ u8 id[EFA_GID_SIZE]; }; struct efa_eq { struct efa_com_eq eeq; struct efa_irq irq; }; int efa_query_device(struct ib_device *ibdev, struct ib_device_attr *props, struct ib_udata *udata); int efa_query_port(struct ib_device *ibdev, u32 port, struct ib_port_attr *props); int efa_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr, int qp_attr_mask, struct ib_qp_init_attr *qp_init_attr); int efa_query_gid(struct ib_device *ibdev, u32 port, int index, union ib_gid *gid); int efa_query_pkey(struct ib_device *ibdev, u32 port, u16 index, u16 *pkey); int efa_alloc_pd(struct ib_pd *ibpd, struct ib_udata *udata); int efa_dealloc_pd(struct ib_pd *ibpd, struct ib_udata *udata); int efa_destroy_qp(struct ib_qp *ibqp, struct ib_udata *udata); int efa_create_qp(struct ib_qp *ibqp, struct ib_qp_init_attr *init_attr, struct ib_udata *udata); int efa_destroy_cq(struct ib_cq *ibcq, struct ib_udata *udata); int efa_create_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *attr, struct ib_udata *udata); struct ib_mr *efa_reg_mr(struct ib_pd *ibpd, u64 start, u64 length, u64 virt_addr, int access_flags, struct ib_udata *udata); struct ib_mr *efa_reg_user_mr_dmabuf(struct ib_pd *ibpd, u64 start, u64 length, u64 virt_addr, int fd, int access_flags, struct ib_udata *udata); int efa_dereg_mr(struct ib_mr *ibmr, struct ib_udata *udata); int efa_get_port_immutable(struct ib_device *ibdev, u32 port_num, struct ib_port_immutable *immutable); int efa_alloc_ucontext(struct ib_ucontext *ibucontext, struct ib_udata *udata); void efa_dealloc_ucontext(struct ib_ucontext *ibucontext); int efa_mmap(struct ib_ucontext *ibucontext, struct vm_area_struct *vma); void efa_mmap_free(struct rdma_user_mmap_entry *rdma_entry); int efa_create_ah(struct ib_ah *ibah, struct rdma_ah_init_attr *init_attr, struct ib_udata *udata); int efa_destroy_ah(struct ib_ah *ibah, u32 flags); int efa_modify_qp(struct ib_qp *ibqp, struct ib_qp_attr *qp_attr, int qp_attr_mask, struct ib_udata *udata); enum rdma_link_layer efa_port_link_layer(struct ib_device *ibdev, u32 port_num); struct rdma_hw_stats *efa_alloc_hw_port_stats(struct ib_device *ibdev, u32 port_num); struct rdma_hw_stats *efa_alloc_hw_device_stats(struct ib_device *ibdev); int efa_get_hw_stats(struct ib_device *ibdev, struct rdma_hw_stats *stats, u32 port_num, int index); #endif /* _EFA_H_ */
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1