Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Maor Gottlieb | 2317 | 74.96% | 8 | 23.53% |
Ariel Levkovich | 501 | 16.21% | 4 | 11.76% |
Yevgeny Kliteynik | 124 | 4.01% | 1 | 2.94% |
Eli Cohen | 31 | 1.00% | 2 | 5.88% |
Artemy Kovalyov | 20 | 0.65% | 1 | 2.94% |
Jason Gunthorpe | 16 | 0.52% | 1 | 2.94% |
Matan Barak | 16 | 0.52% | 2 | 5.88% |
Shun Hao | 15 | 0.49% | 1 | 2.94% |
Leon Romanovsky | 15 | 0.49% | 3 | 8.82% |
Daniel Jurgens | 12 | 0.39% | 2 | 5.88% |
Saeed Mahameed | 8 | 0.26% | 2 | 5.88% |
Haggai Abramonvsky | 6 | 0.19% | 1 | 2.94% |
Huy Nguyen | 3 | 0.10% | 1 | 2.94% |
Aviad Yehezkel | 2 | 0.06% | 1 | 2.94% |
Yishai Hadas | 2 | 0.06% | 1 | 2.94% |
Erez Alfasi | 1 | 0.03% | 1 | 2.94% |
Aviv Heller | 1 | 0.03% | 1 | 2.94% |
Achiad Shochat | 1 | 0.03% | 1 | 2.94% |
Total | 3091 | 34 |
// SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB /* * Copyright (c) 2021, Mellanox Technologies inc. All rights reserved. */ #include <rdma/uverbs_std_types.h> #include "dm.h" #define UVERBS_MODULE_NAME mlx5_ib #include <rdma/uverbs_named_ioctl.h> static int mlx5_cmd_alloc_memic(struct mlx5_dm *dm, phys_addr_t *addr, u64 length, u32 alignment) { struct mlx5_core_dev *dev = dm->dev; u64 num_memic_hw_pages = MLX5_CAP_DEV_MEM(dev, memic_bar_size) >> PAGE_SHIFT; u64 hw_start_addr = MLX5_CAP64_DEV_MEM(dev, memic_bar_start_addr); u32 max_alignment = MLX5_CAP_DEV_MEM(dev, log_max_memic_addr_alignment); u32 num_pages = DIV_ROUND_UP(length, PAGE_SIZE); u32 out[MLX5_ST_SZ_DW(alloc_memic_out)] = {}; u32 in[MLX5_ST_SZ_DW(alloc_memic_in)] = {}; u32 mlx5_alignment; u64 page_idx = 0; int ret = 0; if (!length || (length & MLX5_MEMIC_ALLOC_SIZE_MASK)) return -EINVAL; /* mlx5 device sets alignment as 64*2^driver_value * so normalizing is needed. */ mlx5_alignment = (alignment < MLX5_MEMIC_BASE_ALIGN) ? 0 : alignment - MLX5_MEMIC_BASE_ALIGN; if (mlx5_alignment > max_alignment) return -EINVAL; MLX5_SET(alloc_memic_in, in, opcode, MLX5_CMD_OP_ALLOC_MEMIC); MLX5_SET(alloc_memic_in, in, range_size, num_pages * PAGE_SIZE); MLX5_SET(alloc_memic_in, in, memic_size, length); MLX5_SET(alloc_memic_in, in, log_memic_addr_alignment, mlx5_alignment); while (page_idx < num_memic_hw_pages) { spin_lock(&dm->lock); page_idx = bitmap_find_next_zero_area(dm->memic_alloc_pages, num_memic_hw_pages, page_idx, num_pages, 0); if (page_idx < num_memic_hw_pages) bitmap_set(dm->memic_alloc_pages, page_idx, num_pages); spin_unlock(&dm->lock); if (page_idx >= num_memic_hw_pages) break; MLX5_SET64(alloc_memic_in, in, range_start_addr, hw_start_addr + (page_idx * PAGE_SIZE)); ret = mlx5_cmd_exec_inout(dev, alloc_memic, in, out); if (ret) { spin_lock(&dm->lock); bitmap_clear(dm->memic_alloc_pages, page_idx, num_pages); spin_unlock(&dm->lock); if (ret == -EAGAIN) { page_idx++; continue; } return ret; } *addr = dev->bar_addr + MLX5_GET64(alloc_memic_out, out, memic_start_addr); return 0; } return -ENOMEM; } void mlx5_cmd_dealloc_memic(struct mlx5_dm *dm, phys_addr_t addr, u64 length) { struct mlx5_core_dev *dev = dm->dev; u64 hw_start_addr = MLX5_CAP64_DEV_MEM(dev, memic_bar_start_addr); u32 num_pages = DIV_ROUND_UP(length, PAGE_SIZE); u32 in[MLX5_ST_SZ_DW(dealloc_memic_in)] = {}; u64 start_page_idx; int err; addr -= dev->bar_addr; start_page_idx = (addr - hw_start_addr) >> PAGE_SHIFT; MLX5_SET(dealloc_memic_in, in, opcode, MLX5_CMD_OP_DEALLOC_MEMIC); MLX5_SET64(dealloc_memic_in, in, memic_start_addr, addr); MLX5_SET(dealloc_memic_in, in, memic_size, length); err = mlx5_cmd_exec_in(dev, dealloc_memic, in); if (err) return; spin_lock(&dm->lock); bitmap_clear(dm->memic_alloc_pages, start_page_idx, num_pages); spin_unlock(&dm->lock); } void mlx5_cmd_dealloc_memic_op(struct mlx5_dm *dm, phys_addr_t addr, u8 operation) { u32 in[MLX5_ST_SZ_DW(modify_memic_in)] = {}; struct mlx5_core_dev *dev = dm->dev; MLX5_SET(modify_memic_in, in, opcode, MLX5_CMD_OP_MODIFY_MEMIC); MLX5_SET(modify_memic_in, in, op_mod, MLX5_MODIFY_MEMIC_OP_MOD_DEALLOC); MLX5_SET(modify_memic_in, in, memic_operation_type, operation); MLX5_SET64(modify_memic_in, in, memic_start_addr, addr - dev->bar_addr); mlx5_cmd_exec_in(dev, modify_memic, in); } static int mlx5_cmd_alloc_memic_op(struct mlx5_dm *dm, phys_addr_t addr, u8 operation, phys_addr_t *op_addr) { u32 out[MLX5_ST_SZ_DW(modify_memic_out)] = {}; u32 in[MLX5_ST_SZ_DW(modify_memic_in)] = {}; struct mlx5_core_dev *dev = dm->dev; int err; MLX5_SET(modify_memic_in, in, opcode, MLX5_CMD_OP_MODIFY_MEMIC); MLX5_SET(modify_memic_in, in, op_mod, MLX5_MODIFY_MEMIC_OP_MOD_ALLOC); MLX5_SET(modify_memic_in, in, memic_operation_type, operation); MLX5_SET64(modify_memic_in, in, memic_start_addr, addr - dev->bar_addr); err = mlx5_cmd_exec_inout(dev, modify_memic, in, out); if (err) return err; *op_addr = dev->bar_addr + MLX5_GET64(modify_memic_out, out, memic_operation_addr); return 0; } static int add_dm_mmap_entry(struct ib_ucontext *context, struct mlx5_user_mmap_entry *mentry, u8 mmap_flag, size_t size, u64 address) { mentry->mmap_flag = mmap_flag; mentry->address = address; return rdma_user_mmap_entry_insert_range( context, &mentry->rdma_entry, size, MLX5_IB_MMAP_DEVICE_MEM << 16, (MLX5_IB_MMAP_DEVICE_MEM << 16) + (1UL << 16) - 1); } static void mlx5_ib_dm_memic_free(struct kref *kref) { struct mlx5_ib_dm_memic *dm = container_of(kref, struct mlx5_ib_dm_memic, ref); struct mlx5_ib_dev *dev = to_mdev(dm->base.ibdm.device); mlx5_cmd_dealloc_memic(&dev->dm, dm->base.dev_addr, dm->base.size); kfree(dm); } static int copy_op_to_user(struct mlx5_ib_dm_op_entry *op_entry, struct uverbs_attr_bundle *attrs) { u64 start_offset; u16 page_idx; int err; page_idx = op_entry->mentry.rdma_entry.start_pgoff & 0xFFFF; start_offset = op_entry->op_addr & ~PAGE_MASK; err = uverbs_copy_to(attrs, MLX5_IB_ATTR_DM_MAP_OP_ADDR_RESP_PAGE_INDEX, &page_idx, sizeof(page_idx)); if (err) return err; return uverbs_copy_to(attrs, MLX5_IB_ATTR_DM_MAP_OP_ADDR_RESP_START_OFFSET, &start_offset, sizeof(start_offset)); } static int map_existing_op(struct mlx5_ib_dm_memic *dm, u8 op, struct uverbs_attr_bundle *attrs) { struct mlx5_ib_dm_op_entry *op_entry; op_entry = xa_load(&dm->ops, op); if (!op_entry) return -ENOENT; return copy_op_to_user(op_entry, attrs); } static int UVERBS_HANDLER(MLX5_IB_METHOD_DM_MAP_OP_ADDR)( struct uverbs_attr_bundle *attrs) { struct ib_uobject *uobj = uverbs_attr_get_uobject( attrs, MLX5_IB_ATTR_DM_MAP_OP_ADDR_REQ_HANDLE); struct mlx5_ib_dev *dev = to_mdev(uobj->context->device); struct ib_dm *ibdm = uobj->object; struct mlx5_ib_dm_memic *dm = to_memic(ibdm); struct mlx5_ib_dm_op_entry *op_entry; int err; u8 op; err = uverbs_copy_from(&op, attrs, MLX5_IB_ATTR_DM_MAP_OP_ADDR_REQ_OP); if (err) return err; if (op >= BITS_PER_TYPE(u32)) return -EOPNOTSUPP; if (!(MLX5_CAP_DEV_MEM(dev->mdev, memic_operations) & BIT(op))) return -EOPNOTSUPP; mutex_lock(&dm->ops_xa_lock); err = map_existing_op(dm, op, attrs); if (!err || err != -ENOENT) goto err_unlock; op_entry = kzalloc(sizeof(*op_entry), GFP_KERNEL); if (!op_entry) goto err_unlock; err = mlx5_cmd_alloc_memic_op(&dev->dm, dm->base.dev_addr, op, &op_entry->op_addr); if (err) { kfree(op_entry); goto err_unlock; } op_entry->op = op; op_entry->dm = dm; err = add_dm_mmap_entry(uobj->context, &op_entry->mentry, MLX5_IB_MMAP_TYPE_MEMIC_OP, dm->base.size, op_entry->op_addr & PAGE_MASK); if (err) { mlx5_cmd_dealloc_memic_op(&dev->dm, dm->base.dev_addr, op); kfree(op_entry); goto err_unlock; } /* From this point, entry will be freed by mmap_free */ kref_get(&dm->ref); err = copy_op_to_user(op_entry, attrs); if (err) goto err_remove; err = xa_insert(&dm->ops, op, op_entry, GFP_KERNEL); if (err) goto err_remove; mutex_unlock(&dm->ops_xa_lock); return 0; err_remove: rdma_user_mmap_entry_remove(&op_entry->mentry.rdma_entry); err_unlock: mutex_unlock(&dm->ops_xa_lock); return err; } static struct ib_dm *handle_alloc_dm_memic(struct ib_ucontext *ctx, struct ib_dm_alloc_attr *attr, struct uverbs_attr_bundle *attrs) { struct mlx5_dm *dm_db = &to_mdev(ctx->device)->dm; struct mlx5_ib_dm_memic *dm; u64 start_offset; u16 page_idx; int err; u64 address; if (!MLX5_CAP_DEV_MEM(dm_db->dev, memic)) return ERR_PTR(-EOPNOTSUPP); dm = kzalloc(sizeof(*dm), GFP_KERNEL); if (!dm) return ERR_PTR(-ENOMEM); dm->base.type = MLX5_IB_UAPI_DM_TYPE_MEMIC; dm->base.size = roundup(attr->length, MLX5_MEMIC_BASE_SIZE); dm->base.ibdm.device = ctx->device; kref_init(&dm->ref); xa_init(&dm->ops); mutex_init(&dm->ops_xa_lock); dm->req_length = attr->length; err = mlx5_cmd_alloc_memic(dm_db, &dm->base.dev_addr, dm->base.size, attr->alignment); if (err) { kfree(dm); return ERR_PTR(err); } address = dm->base.dev_addr & PAGE_MASK; err = add_dm_mmap_entry(ctx, &dm->mentry, MLX5_IB_MMAP_TYPE_MEMIC, dm->base.size, address); if (err) { mlx5_cmd_dealloc_memic(dm_db, dm->base.dev_addr, dm->base.size); kfree(dm); return ERR_PTR(err); } page_idx = dm->mentry.rdma_entry.start_pgoff & 0xFFFF; err = uverbs_copy_to(attrs, MLX5_IB_ATTR_ALLOC_DM_RESP_PAGE_INDEX, &page_idx, sizeof(page_idx)); if (err) goto err_copy; start_offset = dm->base.dev_addr & ~PAGE_MASK; err = uverbs_copy_to(attrs, MLX5_IB_ATTR_ALLOC_DM_RESP_START_OFFSET, &start_offset, sizeof(start_offset)); if (err) goto err_copy; return &dm->base.ibdm; err_copy: rdma_user_mmap_entry_remove(&dm->mentry.rdma_entry); return ERR_PTR(err); } static enum mlx5_sw_icm_type get_icm_type(int uapi_type) { switch (uapi_type) { case MLX5_IB_UAPI_DM_TYPE_HEADER_MODIFY_SW_ICM: return MLX5_SW_ICM_TYPE_HEADER_MODIFY; case MLX5_IB_UAPI_DM_TYPE_HEADER_MODIFY_PATTERN_SW_ICM: return MLX5_SW_ICM_TYPE_HEADER_MODIFY_PATTERN; case MLX5_IB_UAPI_DM_TYPE_ENCAP_SW_ICM: return MLX5_SW_ICM_TYPE_SW_ENCAP; case MLX5_IB_UAPI_DM_TYPE_STEERING_SW_ICM: default: return MLX5_SW_ICM_TYPE_STEERING; } } static struct ib_dm *handle_alloc_dm_sw_icm(struct ib_ucontext *ctx, struct ib_dm_alloc_attr *attr, struct uverbs_attr_bundle *attrs, int type) { struct mlx5_core_dev *dev = to_mdev(ctx->device)->mdev; enum mlx5_sw_icm_type icm_type; struct mlx5_ib_dm_icm *dm; u64 act_size; int err; if (!capable(CAP_SYS_RAWIO) || !capable(CAP_NET_RAW)) return ERR_PTR(-EPERM); switch (type) { case MLX5_IB_UAPI_DM_TYPE_STEERING_SW_ICM: case MLX5_IB_UAPI_DM_TYPE_HEADER_MODIFY_SW_ICM: case MLX5_IB_UAPI_DM_TYPE_ENCAP_SW_ICM: if (!(MLX5_CAP_FLOWTABLE_NIC_RX(dev, sw_owner) || MLX5_CAP_FLOWTABLE_NIC_TX(dev, sw_owner) || MLX5_CAP_FLOWTABLE_NIC_RX(dev, sw_owner_v2) || MLX5_CAP_FLOWTABLE_NIC_TX(dev, sw_owner_v2))) return ERR_PTR(-EOPNOTSUPP); break; case MLX5_IB_UAPI_DM_TYPE_HEADER_MODIFY_PATTERN_SW_ICM: if (!MLX5_CAP_FLOWTABLE_NIC_RX(dev, sw_owner_v2) || !MLX5_CAP_FLOWTABLE_NIC_TX(dev, sw_owner_v2)) return ERR_PTR(-EOPNOTSUPP); break; default: return ERR_PTR(-EOPNOTSUPP); } dm = kzalloc(sizeof(*dm), GFP_KERNEL); if (!dm) return ERR_PTR(-ENOMEM); dm->base.type = type; dm->base.ibdm.device = ctx->device; /* Allocation size must a multiple of the basic block size * and a power of 2. */ act_size = round_up(attr->length, MLX5_SW_ICM_BLOCK_SIZE(dev)); act_size = roundup_pow_of_two(act_size); dm->base.size = act_size; icm_type = get_icm_type(type); err = mlx5_dm_sw_icm_alloc(dev, icm_type, act_size, attr->alignment, to_mucontext(ctx)->devx_uid, &dm->base.dev_addr, &dm->obj_id); if (err) goto free; err = uverbs_copy_to(attrs, MLX5_IB_ATTR_ALLOC_DM_RESP_START_OFFSET, &dm->base.dev_addr, sizeof(dm->base.dev_addr)); if (err) { mlx5_dm_sw_icm_dealloc(dev, icm_type, dm->base.size, to_mucontext(ctx)->devx_uid, dm->base.dev_addr, dm->obj_id); goto free; } return &dm->base.ibdm; free: kfree(dm); return ERR_PTR(err); } struct ib_dm *mlx5_ib_alloc_dm(struct ib_device *ibdev, struct ib_ucontext *context, struct ib_dm_alloc_attr *attr, struct uverbs_attr_bundle *attrs) { enum mlx5_ib_uapi_dm_type type; int err; err = uverbs_get_const_default(&type, attrs, MLX5_IB_ATTR_ALLOC_DM_REQ_TYPE, MLX5_IB_UAPI_DM_TYPE_MEMIC); if (err) return ERR_PTR(err); mlx5_ib_dbg(to_mdev(ibdev), "alloc_dm req: dm_type=%d user_length=0x%llx log_alignment=%d\n", type, attr->length, attr->alignment); switch (type) { case MLX5_IB_UAPI_DM_TYPE_MEMIC: return handle_alloc_dm_memic(context, attr, attrs); case MLX5_IB_UAPI_DM_TYPE_STEERING_SW_ICM: case MLX5_IB_UAPI_DM_TYPE_HEADER_MODIFY_SW_ICM: case MLX5_IB_UAPI_DM_TYPE_HEADER_MODIFY_PATTERN_SW_ICM: case MLX5_IB_UAPI_DM_TYPE_ENCAP_SW_ICM: return handle_alloc_dm_sw_icm(context, attr, attrs, type); default: return ERR_PTR(-EOPNOTSUPP); } } static void dm_memic_remove_ops(struct mlx5_ib_dm_memic *dm) { struct mlx5_ib_dm_op_entry *entry; unsigned long idx; mutex_lock(&dm->ops_xa_lock); xa_for_each(&dm->ops, idx, entry) { xa_erase(&dm->ops, idx); rdma_user_mmap_entry_remove(&entry->mentry.rdma_entry); } mutex_unlock(&dm->ops_xa_lock); } static void mlx5_dm_memic_dealloc(struct mlx5_ib_dm_memic *dm) { dm_memic_remove_ops(dm); rdma_user_mmap_entry_remove(&dm->mentry.rdma_entry); } static int mlx5_dm_icm_dealloc(struct mlx5_ib_ucontext *ctx, struct mlx5_ib_dm_icm *dm) { enum mlx5_sw_icm_type type = get_icm_type(dm->base.type); struct mlx5_core_dev *dev = to_mdev(dm->base.ibdm.device)->mdev; int err; err = mlx5_dm_sw_icm_dealloc(dev, type, dm->base.size, ctx->devx_uid, dm->base.dev_addr, dm->obj_id); if (!err) kfree(dm); return 0; } static int mlx5_ib_dealloc_dm(struct ib_dm *ibdm, struct uverbs_attr_bundle *attrs) { struct mlx5_ib_ucontext *ctx = rdma_udata_to_drv_context( &attrs->driver_udata, struct mlx5_ib_ucontext, ibucontext); struct mlx5_ib_dm *dm = to_mdm(ibdm); switch (dm->type) { case MLX5_IB_UAPI_DM_TYPE_MEMIC: mlx5_dm_memic_dealloc(to_memic(ibdm)); return 0; case MLX5_IB_UAPI_DM_TYPE_STEERING_SW_ICM: case MLX5_IB_UAPI_DM_TYPE_HEADER_MODIFY_SW_ICM: case MLX5_IB_UAPI_DM_TYPE_HEADER_MODIFY_PATTERN_SW_ICM: case MLX5_IB_UAPI_DM_TYPE_ENCAP_SW_ICM: return mlx5_dm_icm_dealloc(ctx, to_icm(ibdm)); default: return -EOPNOTSUPP; } } static int UVERBS_HANDLER(MLX5_IB_METHOD_DM_QUERY)( struct uverbs_attr_bundle *attrs) { struct ib_dm *ibdm = uverbs_attr_get_obj(attrs, MLX5_IB_ATTR_QUERY_DM_REQ_HANDLE); struct mlx5_ib_dm *dm = to_mdm(ibdm); struct mlx5_ib_dm_memic *memic; u64 start_offset; u16 page_idx; int err; if (dm->type != MLX5_IB_UAPI_DM_TYPE_MEMIC) return -EOPNOTSUPP; memic = to_memic(ibdm); page_idx = memic->mentry.rdma_entry.start_pgoff & 0xFFFF; err = uverbs_copy_to(attrs, MLX5_IB_ATTR_QUERY_DM_RESP_PAGE_INDEX, &page_idx, sizeof(page_idx)); if (err) return err; start_offset = memic->base.dev_addr & ~PAGE_MASK; err = uverbs_copy_to(attrs, MLX5_IB_ATTR_QUERY_DM_RESP_START_OFFSET, &start_offset, sizeof(start_offset)); if (err) return err; return uverbs_copy_to(attrs, MLX5_IB_ATTR_QUERY_DM_RESP_LENGTH, &memic->req_length, sizeof(memic->req_length)); } void mlx5_ib_dm_mmap_free(struct mlx5_ib_dev *dev, struct mlx5_user_mmap_entry *mentry) { struct mlx5_ib_dm_op_entry *op_entry; struct mlx5_ib_dm_memic *mdm; switch (mentry->mmap_flag) { case MLX5_IB_MMAP_TYPE_MEMIC: mdm = container_of(mentry, struct mlx5_ib_dm_memic, mentry); kref_put(&mdm->ref, mlx5_ib_dm_memic_free); break; case MLX5_IB_MMAP_TYPE_MEMIC_OP: op_entry = container_of(mentry, struct mlx5_ib_dm_op_entry, mentry); mdm = op_entry->dm; mlx5_cmd_dealloc_memic_op(&dev->dm, mdm->base.dev_addr, op_entry->op); kfree(op_entry); kref_put(&mdm->ref, mlx5_ib_dm_memic_free); break; default: WARN_ON(true); } } DECLARE_UVERBS_NAMED_METHOD( MLX5_IB_METHOD_DM_QUERY, UVERBS_ATTR_IDR(MLX5_IB_ATTR_QUERY_DM_REQ_HANDLE, UVERBS_OBJECT_DM, UVERBS_ACCESS_READ, UA_MANDATORY), UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_QUERY_DM_RESP_START_OFFSET, UVERBS_ATTR_TYPE(u64), UA_MANDATORY), UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_QUERY_DM_RESP_PAGE_INDEX, UVERBS_ATTR_TYPE(u16), UA_MANDATORY), UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_QUERY_DM_RESP_LENGTH, UVERBS_ATTR_TYPE(u64), UA_MANDATORY)); ADD_UVERBS_ATTRIBUTES_SIMPLE( mlx5_ib_dm, UVERBS_OBJECT_DM, UVERBS_METHOD_DM_ALLOC, UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_ALLOC_DM_RESP_START_OFFSET, UVERBS_ATTR_TYPE(u64), UA_MANDATORY), UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_ALLOC_DM_RESP_PAGE_INDEX, UVERBS_ATTR_TYPE(u16), UA_OPTIONAL), UVERBS_ATTR_CONST_IN(MLX5_IB_ATTR_ALLOC_DM_REQ_TYPE, enum mlx5_ib_uapi_dm_type, UA_OPTIONAL)); DECLARE_UVERBS_NAMED_METHOD( MLX5_IB_METHOD_DM_MAP_OP_ADDR, UVERBS_ATTR_IDR(MLX5_IB_ATTR_DM_MAP_OP_ADDR_REQ_HANDLE, UVERBS_OBJECT_DM, UVERBS_ACCESS_READ, UA_MANDATORY), UVERBS_ATTR_PTR_IN(MLX5_IB_ATTR_DM_MAP_OP_ADDR_REQ_OP, UVERBS_ATTR_TYPE(u8), UA_MANDATORY), UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_DM_MAP_OP_ADDR_RESP_START_OFFSET, UVERBS_ATTR_TYPE(u64), UA_MANDATORY), UVERBS_ATTR_PTR_OUT(MLX5_IB_ATTR_DM_MAP_OP_ADDR_RESP_PAGE_INDEX, UVERBS_ATTR_TYPE(u16), UA_OPTIONAL)); DECLARE_UVERBS_GLOBAL_METHODS(UVERBS_OBJECT_DM, &UVERBS_METHOD(MLX5_IB_METHOD_DM_MAP_OP_ADDR), &UVERBS_METHOD(MLX5_IB_METHOD_DM_QUERY)); const struct uapi_definition mlx5_ib_dm_defs[] = { UAPI_DEF_CHAIN_OBJ_TREE(UVERBS_OBJECT_DM, &mlx5_ib_dm), UAPI_DEF_CHAIN_OBJ_TREE_NAMED(UVERBS_OBJECT_DM), {}, }; const struct ib_device_ops mlx5_ib_dev_dm_ops = { .alloc_dm = mlx5_ib_alloc_dm, .dealloc_dm = mlx5_ib_dealloc_dm, .reg_dm_mr = mlx5_ib_reg_dm_mr, };
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1