Contributors: 4
Author |
Tokens |
Token Proportion |
Commits |
Commit Proportion |
Dennis Dalessandro |
149 |
86.13% |
6 |
66.67% |
Jianxin Xiong |
22 |
12.72% |
1 |
11.11% |
Krzysztof Kozlowski |
1 |
0.58% |
1 |
11.11% |
caihuoqing |
1 |
0.58% |
1 |
11.11% |
Total |
173 |
|
9 |
|
/* SPDX-License-Identifier: GPL-2.0 OR BSD-3-Clause */
/*
* Copyright(c) 2016 Intel Corporation.
*/
#ifndef DEF_RVTMR_H
#define DEF_RVTMR_H
#include <rdma/rdma_vt.h>
struct rvt_mr {
struct ib_mr ibmr;
struct ib_umem *umem;
struct rvt_mregion mr; /* must be last */
};
static inline struct rvt_mr *to_imr(struct ib_mr *ibmr)
{
return container_of(ibmr, struct rvt_mr, ibmr);
}
int rvt_driver_mr_init(struct rvt_dev_info *rdi);
void rvt_mr_exit(struct rvt_dev_info *rdi);
/* Mem Regions */
struct ib_mr *rvt_get_dma_mr(struct ib_pd *pd, int acc);
struct ib_mr *rvt_reg_user_mr(struct ib_pd *pd, u64 start, u64 length,
u64 virt_addr, int mr_access_flags,
struct ib_udata *udata);
int rvt_dereg_mr(struct ib_mr *ibmr, struct ib_udata *udata);
struct ib_mr *rvt_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type,
u32 max_num_sg);
int rvt_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg,
int sg_nents, unsigned int *sg_offset);
#endif /* DEF_RVTMR_H */