Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Adit Ranadive | 1401 | 98.94% | 1 | 50.00% |
Yuval Shaia | 15 | 1.06% | 1 | 50.00% |
Total | 1416 | 2 |
/* * Copyright (c) 2012-2016 VMware, Inc. All rights reserved. * * This program is free software; you can redistribute it and/or * modify it under the terms of EITHER the GNU General Public License * version 2 as published by the Free Software Foundation or the BSD * 2-Clause License. This program is distributed in the hope that it * will be useful, but WITHOUT ANY WARRANTY; WITHOUT EVEN THE IMPLIED * WARRANTY OF MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. * See the GNU General Public License version 2 for more details at * http://www.gnu.org/licenses/old-licenses/gpl-2.0.en.html. * * You should have received a copy of the GNU General Public License * along with this program available in the file COPYING in the main * directory of this source tree. * * The BSD 2-Clause License * * Redistribution and use in source and binary forms, with or * without modification, are permitted provided that the following * conditions are met: * * - Redistributions of source code must retain the above * copyright notice, this list of conditions and the following * disclaimer. * * - Redistributions in binary form must reproduce the above * copyright notice, this list of conditions and the following * disclaimer in the documentation and/or other materials * provided with the distribution. * * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS * FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE * COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, * INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR * SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, * STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED * OF THE POSSIBILITY OF SUCH DAMAGE. */ #include <linux/list.h> #include <linux/slab.h> #include "pvrdma.h" /** * pvrdma_get_dma_mr - get a DMA memory region * @pd: protection domain * @acc: access flags * * @return: ib_mr pointer on success, otherwise returns an errno. */ struct ib_mr *pvrdma_get_dma_mr(struct ib_pd *pd, int acc) { struct pvrdma_dev *dev = to_vdev(pd->device); struct pvrdma_user_mr *mr; union pvrdma_cmd_req req; union pvrdma_cmd_resp rsp; struct pvrdma_cmd_create_mr *cmd = &req.create_mr; struct pvrdma_cmd_create_mr_resp *resp = &rsp.create_mr_resp; int ret; /* Support only LOCAL_WRITE flag for DMA MRs */ if (acc & ~IB_ACCESS_LOCAL_WRITE) { dev_warn(&dev->pdev->dev, "unsupported dma mr access flags %#x\n", acc); return ERR_PTR(-EOPNOTSUPP); } mr = kzalloc(sizeof(*mr), GFP_KERNEL); if (!mr) return ERR_PTR(-ENOMEM); memset(cmd, 0, sizeof(*cmd)); cmd->hdr.cmd = PVRDMA_CMD_CREATE_MR; cmd->pd_handle = to_vpd(pd)->pd_handle; cmd->access_flags = acc; cmd->flags = PVRDMA_MR_FLAG_DMA; ret = pvrdma_cmd_post(dev, &req, &rsp, PVRDMA_CMD_CREATE_MR_RESP); if (ret < 0) { dev_warn(&dev->pdev->dev, "could not get DMA mem region, error: %d\n", ret); kfree(mr); return ERR_PTR(ret); } mr->mmr.mr_handle = resp->mr_handle; mr->ibmr.lkey = resp->lkey; mr->ibmr.rkey = resp->rkey; return &mr->ibmr; } /** * pvrdma_reg_user_mr - register a userspace memory region * @pd: protection domain * @start: starting address * @length: length of region * @virt_addr: I/O virtual address * @access_flags: access flags for memory region * @udata: user data * * @return: ib_mr pointer on success, otherwise returns an errno. */ struct ib_mr *pvrdma_reg_user_mr(struct ib_pd *pd, u64 start, u64 length, u64 virt_addr, int access_flags, struct ib_udata *udata) { struct pvrdma_dev *dev = to_vdev(pd->device); struct pvrdma_user_mr *mr = NULL; struct ib_umem *umem; union pvrdma_cmd_req req; union pvrdma_cmd_resp rsp; struct pvrdma_cmd_create_mr *cmd = &req.create_mr; struct pvrdma_cmd_create_mr_resp *resp = &rsp.create_mr_resp; int ret; if (length == 0 || length > dev->dsr->caps.max_mr_size) { dev_warn(&dev->pdev->dev, "invalid mem region length\n"); return ERR_PTR(-EINVAL); } umem = ib_umem_get(pd->uobject->context, start, length, access_flags, 0); if (IS_ERR(umem)) { dev_warn(&dev->pdev->dev, "could not get umem for mem region\n"); return ERR_CAST(umem); } if (umem->npages < 0 || umem->npages > PVRDMA_PAGE_DIR_MAX_PAGES) { dev_warn(&dev->pdev->dev, "overflow %d pages in mem region\n", umem->npages); ret = -EINVAL; goto err_umem; } mr = kzalloc(sizeof(*mr), GFP_KERNEL); if (!mr) { ret = -ENOMEM; goto err_umem; } mr->mmr.iova = virt_addr; mr->mmr.size = length; mr->umem = umem; ret = pvrdma_page_dir_init(dev, &mr->pdir, umem->npages, false); if (ret) { dev_warn(&dev->pdev->dev, "could not allocate page directory\n"); goto err_umem; } ret = pvrdma_page_dir_insert_umem(&mr->pdir, mr->umem, 0); if (ret) goto err_pdir; memset(cmd, 0, sizeof(*cmd)); cmd->hdr.cmd = PVRDMA_CMD_CREATE_MR; cmd->start = start; cmd->length = length; cmd->pd_handle = to_vpd(pd)->pd_handle; cmd->access_flags = access_flags; cmd->nchunks = umem->npages; cmd->pdir_dma = mr->pdir.dir_dma; ret = pvrdma_cmd_post(dev, &req, &rsp, PVRDMA_CMD_CREATE_MR_RESP); if (ret < 0) { dev_warn(&dev->pdev->dev, "could not register mem region, error: %d\n", ret); goto err_pdir; } mr->mmr.mr_handle = resp->mr_handle; mr->ibmr.lkey = resp->lkey; mr->ibmr.rkey = resp->rkey; return &mr->ibmr; err_pdir: pvrdma_page_dir_cleanup(dev, &mr->pdir); err_umem: ib_umem_release(umem); kfree(mr); return ERR_PTR(ret); } /** * pvrdma_alloc_mr - allocate a memory region * @pd: protection domain * @mr_type: type of memory region * @max_num_sg: maximum number of pages * * @return: ib_mr pointer on success, otherwise returns an errno. */ struct ib_mr *pvrdma_alloc_mr(struct ib_pd *pd, enum ib_mr_type mr_type, u32 max_num_sg) { struct pvrdma_dev *dev = to_vdev(pd->device); struct pvrdma_user_mr *mr; union pvrdma_cmd_req req; union pvrdma_cmd_resp rsp; struct pvrdma_cmd_create_mr *cmd = &req.create_mr; struct pvrdma_cmd_create_mr_resp *resp = &rsp.create_mr_resp; int size = max_num_sg * sizeof(u64); int ret; if (mr_type != IB_MR_TYPE_MEM_REG || max_num_sg > PVRDMA_MAX_FAST_REG_PAGES) return ERR_PTR(-EINVAL); mr = kzalloc(sizeof(*mr), GFP_KERNEL); if (!mr) return ERR_PTR(-ENOMEM); mr->pages = kzalloc(size, GFP_KERNEL); if (!mr->pages) { ret = -ENOMEM; goto freemr; } ret = pvrdma_page_dir_init(dev, &mr->pdir, max_num_sg, false); if (ret) { dev_warn(&dev->pdev->dev, "failed to allocate page dir for mr\n"); ret = -ENOMEM; goto freepages; } memset(cmd, 0, sizeof(*cmd)); cmd->hdr.cmd = PVRDMA_CMD_CREATE_MR; cmd->pd_handle = to_vpd(pd)->pd_handle; cmd->access_flags = 0; cmd->flags = PVRDMA_MR_FLAG_FRMR; cmd->nchunks = max_num_sg; ret = pvrdma_cmd_post(dev, &req, &rsp, PVRDMA_CMD_CREATE_MR_RESP); if (ret < 0) { dev_warn(&dev->pdev->dev, "could not create FR mem region, error: %d\n", ret); goto freepdir; } mr->max_pages = max_num_sg; mr->mmr.mr_handle = resp->mr_handle; mr->ibmr.lkey = resp->lkey; mr->ibmr.rkey = resp->rkey; mr->page_shift = PAGE_SHIFT; mr->umem = NULL; return &mr->ibmr; freepdir: pvrdma_page_dir_cleanup(dev, &mr->pdir); freepages: kfree(mr->pages); freemr: kfree(mr); return ERR_PTR(ret); } /** * pvrdma_dereg_mr - deregister a memory region * @ibmr: memory region * * @return: 0 on success. */ int pvrdma_dereg_mr(struct ib_mr *ibmr) { struct pvrdma_user_mr *mr = to_vmr(ibmr); struct pvrdma_dev *dev = to_vdev(ibmr->device); union pvrdma_cmd_req req; struct pvrdma_cmd_destroy_mr *cmd = &req.destroy_mr; int ret; memset(cmd, 0, sizeof(*cmd)); cmd->hdr.cmd = PVRDMA_CMD_DESTROY_MR; cmd->mr_handle = mr->mmr.mr_handle; ret = pvrdma_cmd_post(dev, &req, NULL, 0); if (ret < 0) dev_warn(&dev->pdev->dev, "could not deregister mem region, error: %d\n", ret); pvrdma_page_dir_cleanup(dev, &mr->pdir); if (mr->umem) ib_umem_release(mr->umem); kfree(mr->pages); kfree(mr); return 0; } static int pvrdma_set_page(struct ib_mr *ibmr, u64 addr) { struct pvrdma_user_mr *mr = to_vmr(ibmr); if (mr->npages == mr->max_pages) return -ENOMEM; mr->pages[mr->npages++] = addr; return 0; } int pvrdma_map_mr_sg(struct ib_mr *ibmr, struct scatterlist *sg, int sg_nents, unsigned int *sg_offset) { struct pvrdma_user_mr *mr = to_vmr(ibmr); struct pvrdma_dev *dev = to_vdev(ibmr->device); int ret; mr->npages = 0; ret = ib_sg_to_pages(ibmr, sg, sg_nents, sg_offset, pvrdma_set_page); if (ret < 0) dev_warn(&dev->pdev->dev, "could not map sg to pages\n"); return ret; }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1