Release 4.15 drivers/infiniband/hw/cxgb3/iwch_provider.h
  
  
  
/*
 * Copyright (c) 2006 Chelsio, Inc. All rights reserved.
 *
 * This software is available to you under a choice of one of two
 * licenses.  You may choose to be licensed under the terms of the GNU
 * General Public License (GPL) Version 2, available from the file
 * COPYING in the main directory of this source tree, or the
 * OpenIB.org BSD license below:
 *
 *     Redistribution and use in source and binary forms, with or
 *     without modification, are permitted provided that the following
 *     conditions are met:
 *
 *      - Redistributions of source code must retain the above
 *        copyright notice, this list of conditions and the following
 *        disclaimer.
 *
 *      - Redistributions in binary form must reproduce the above
 *        copyright notice, this list of conditions and the following
 *        disclaimer in the documentation and/or other materials
 *        provided with the distribution.
 *
 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
 * SOFTWARE.
 */
#ifndef __IWCH_PROVIDER_H__
#define __IWCH_PROVIDER_H__
#include <linux/list.h>
#include <linux/spinlock.h>
#include <rdma/ib_verbs.h>
#include <asm/types.h>
#include "t3cdev.h"
#include "iwch.h"
#include "cxio_wr.h"
#include "cxio_hal.h"
struct iwch_pd {
	
struct ib_pd ibpd;
	
u32 pdid;
	
struct iwch_dev *rhp;
};
static inline struct iwch_pd *to_iwch_pd(struct ib_pd *ibpd)
{
	return container_of(ibpd, struct iwch_pd, ibpd);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Steve Wise | 25 | 100.00% | 1 | 100.00% | 
| Total | 25 | 100.00% | 1 | 100.00% | 
struct tpt_attributes {
	
u32 stag;
	
u32 state:1;
	
u32 type:2;
	
u32 rsvd:1;
	
enum tpt_mem_perm perms;
	
u32 remote_invaliate_disable:1;
	
u32 zbva:1;
	
u32 mw_bind_enable:1;
	
u32 page_size:5;
	
u32 pdid;
	
u32 qpid;
	
u32 pbl_addr;
	
u32 len;
	
u64 va_fbo;
	
u32 pbl_size;
};
struct iwch_mr {
	
struct ib_mr ibmr;
	
struct ib_umem *umem;
	
struct iwch_dev *rhp;
	
u64 kva;
	
struct tpt_attributes attr;
	
u64 *pages;
	
u32 npages;
};
typedef struct iwch_mw iwch_mw_handle;
static inline struct iwch_mr *to_iwch_mr(struct ib_mr *ibmr)
{
	return container_of(ibmr, struct iwch_mr, ibmr);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Steve Wise | 25 | 100.00% | 1 | 100.00% | 
| Total | 25 | 100.00% | 1 | 100.00% | 
struct iwch_mw {
	
struct ib_mw ibmw;
	
struct iwch_dev *rhp;
	
u64 kva;
	
struct tpt_attributes attr;
};
static inline struct iwch_mw *to_iwch_mw(struct ib_mw *ibmw)
{
	return container_of(ibmw, struct iwch_mw, ibmw);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Steve Wise | 25 | 100.00% | 1 | 100.00% | 
| Total | 25 | 100.00% | 1 | 100.00% | 
struct iwch_cq {
	
struct ib_cq ibcq;
	
struct iwch_dev *rhp;
	
struct t3_cq cq;
	
spinlock_t lock;
	
spinlock_t comp_handler_lock;
	
atomic_t refcnt;
	
wait_queue_head_t wait;
	
u32 __user *user_rptr_addr;
};
static inline struct iwch_cq *to_iwch_cq(struct ib_cq *ibcq)
{
	return container_of(ibcq, struct iwch_cq, ibcq);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Steve Wise | 25 | 100.00% | 1 | 100.00% | 
| Total | 25 | 100.00% | 1 | 100.00% | 
enum IWCH_QP_FLAGS {
	
QP_QUIESCED = 0x01
};
struct iwch_mpa_attributes {
	
u8 initiator;
	
u8 recv_marker_enabled;
	
u8 xmit_marker_enabled;	/* iWARP: enable inbound Read Resp. */
	
u8 crc_enabled;
	
u8 version;	/* 0 or 1 */
};
struct iwch_qp_attributes {
	
u32 scq;
	
u32 rcq;
	
u32 sq_num_entries;
	
u32 rq_num_entries;
	
u32 sq_max_sges;
	
u32 sq_max_sges_rdma_write;
	
u32 rq_max_sges;
	
u32 state;
	
u8 enable_rdma_read;
	
u8 enable_rdma_write;	/* enable inbound Read Resp. */
	
u8 enable_bind;
	
u8 enable_mmid0_fastreg;	/* Enable STAG0 + Fast-register */
	/*
         * Next QP state. If specify the current state, only the
         * QP attributes will be modified.
         */
	
u32 max_ord;
	
u32 max_ird;
	
u32 pd;	/* IN */
	
u32 next_state;
	
char terminate_buffer[52];
	
u32 terminate_msg_len;
	
u8 is_terminate_local;
	
struct iwch_mpa_attributes mpa_attr;	/* IN-OUT */
	
struct iwch_ep *llp_stream_handle;
	
char *stream_msg_buf;	/* Last stream msg. before Idle -> RTS */
	
u32 stream_msg_buf_len;	/* Only on Idle -> RTS */
};
struct iwch_qp {
	
struct ib_qp ibqp;
	
struct iwch_dev *rhp;
	
struct iwch_ep *ep;
	
struct iwch_qp_attributes attr;
	
struct t3_wq wq;
	
spinlock_t lock;
	
atomic_t refcnt;
	
wait_queue_head_t wait;
	
enum IWCH_QP_FLAGS flags;
};
static inline int qp_quiesced(struct iwch_qp *qhp)
{
	return qhp->flags & QP_QUIESCED;
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Steve Wise | 19 | 100.00% | 1 | 100.00% | 
| Total | 19 | 100.00% | 1 | 100.00% | 
static inline struct iwch_qp *to_iwch_qp(struct ib_qp *ibqp)
{
	return container_of(ibqp, struct iwch_qp, ibqp);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Steve Wise | 25 | 100.00% | 1 | 100.00% | 
| Total | 25 | 100.00% | 1 | 100.00% | 
void iwch_qp_add_ref(struct ib_qp *qp);
void iwch_qp_rem_ref(struct ib_qp *qp);
struct iwch_ucontext {
	
struct ib_ucontext ibucontext;
	
struct cxio_ucontext uctx;
	
u32 key;
	
spinlock_t mmap_lock;
	
struct list_head mmaps;
};
static inline struct iwch_ucontext *to_iwch_ucontext(struct ib_ucontext *c)
{
	return container_of(c, struct iwch_ucontext, ibucontext);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Steve Wise | 25 | 100.00% | 1 | 100.00% | 
| Total | 25 | 100.00% | 1 | 100.00% | 
struct iwch_mm_entry {
	
struct list_head entry;
	
u64 addr;
	
u32 key;
	
unsigned len;
};
static inline struct iwch_mm_entry *remove_mmap(struct iwch_ucontext *ucontext,
						u32 key, unsigned len)
{
	struct list_head *pos, *nxt;
	struct iwch_mm_entry *mm;
	spin_lock(&ucontext->mmap_lock);
	list_for_each_safe(pos, nxt, &ucontext->mmaps) {
		mm = list_entry(pos, struct iwch_mm_entry, entry);
		if (mm->key == key && mm->len == len) {
			list_del_init(&mm->entry);
			spin_unlock(&ucontext->mmap_lock);
			pr_debug("%s key 0x%x addr 0x%llx len %d\n",
				 __func__, key,
				 (unsigned long long)mm->addr, mm->len);
			return mm;
		}
	}
	spin_unlock(&ucontext->mmap_lock);
	return NULL;
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Steve Wise | 129 | 98.47% | 1 | 33.33% | 
| Joe Perches | 1 | 0.76% | 1 | 33.33% | 
| Harvey Harrison | 1 | 0.76% | 1 | 33.33% | 
| Total | 131 | 100.00% | 3 | 100.00% | 
static inline void insert_mmap(struct iwch_ucontext *ucontext,
			       struct iwch_mm_entry *mm)
{
	spin_lock(&ucontext->mmap_lock);
	pr_debug("%s key 0x%x addr 0x%llx len %d\n",
		 __func__, mm->key, (unsigned long long)mm->addr, mm->len);
	list_add_tail(&mm->entry, &ucontext->mmaps);
	spin_unlock(&ucontext->mmap_lock);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Steve Wise | 68 | 97.14% | 1 | 33.33% | 
| Joe Perches | 1 | 1.43% | 1 | 33.33% | 
| Harvey Harrison | 1 | 1.43% | 1 | 33.33% | 
| Total | 70 | 100.00% | 3 | 100.00% | 
enum iwch_qp_attr_mask {
	
IWCH_QP_ATTR_NEXT_STATE = 1 << 0,
	
IWCH_QP_ATTR_ENABLE_RDMA_READ = 1 << 7,
	
IWCH_QP_ATTR_ENABLE_RDMA_WRITE = 1 << 8,
	
IWCH_QP_ATTR_ENABLE_RDMA_BIND = 1 << 9,
	
IWCH_QP_ATTR_MAX_ORD = 1 << 11,
	
IWCH_QP_ATTR_MAX_IRD = 1 << 12,
	
IWCH_QP_ATTR_LLP_STREAM_HANDLE = 1 << 22,
	
IWCH_QP_ATTR_STREAM_MSG_BUFFER = 1 << 23,
	
IWCH_QP_ATTR_MPA_ATTR = 1 << 24,
	
IWCH_QP_ATTR_QP_CONTEXT_ACTIVATE = 1 << 25,
	
IWCH_QP_ATTR_VALID_MODIFY = (IWCH_QP_ATTR_ENABLE_RDMA_READ |
				     IWCH_QP_ATTR_ENABLE_RDMA_WRITE |
				     IWCH_QP_ATTR_MAX_ORD |
				     IWCH_QP_ATTR_MAX_IRD |
				     IWCH_QP_ATTR_LLP_STREAM_HANDLE |
				     IWCH_QP_ATTR_STREAM_MSG_BUFFER |
				     IWCH_QP_ATTR_MPA_ATTR |
				     IWCH_QP_ATTR_QP_CONTEXT_ACTIVATE)
};
int iwch_modify_qp(struct iwch_dev *rhp,
				struct iwch_qp *qhp,
				enum iwch_qp_attr_mask mask,
				struct iwch_qp_attributes *attrs,
				int internal);
enum iwch_qp_state {
	
IWCH_QP_STATE_IDLE,
	
IWCH_QP_STATE_RTS,
	
IWCH_QP_STATE_ERROR,
	
IWCH_QP_STATE_TERMINATE,
	
IWCH_QP_STATE_CLOSING,
	
IWCH_QP_STATE_TOT
};
static inline int iwch_convert_state(enum ib_qp_state ib_state)
{
	switch (ib_state) {
	case IB_QPS_RESET:
	case IB_QPS_INIT:
		return IWCH_QP_STATE_IDLE;
	case IB_QPS_RTS:
		return IWCH_QP_STATE_RTS;
	case IB_QPS_SQD:
		return IWCH_QP_STATE_CLOSING;
	case IB_QPS_SQE:
		return IWCH_QP_STATE_TERMINATE;
	case IB_QPS_ERR:
		return IWCH_QP_STATE_ERROR;
	default:
		return -1;
	}
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Steve Wise | 55 | 100.00% | 1 | 100.00% | 
| Total | 55 | 100.00% | 1 | 100.00% | 
static inline u32 iwch_ib_to_tpt_access(int acc)
{
	return (acc & IB_ACCESS_REMOTE_WRITE ? TPT_REMOTE_WRITE : 0) |
	       (acc & IB_ACCESS_REMOTE_READ ? TPT_REMOTE_READ : 0) |
	       (acc & IB_ACCESS_LOCAL_WRITE ? TPT_LOCAL_WRITE : 0) |
	       (acc & IB_ACCESS_MW_BIND ? TPT_MW_BIND : 0) |
	       TPT_LOCAL_READ;
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Steve Wise | 53 | 100.00% | 3 | 100.00% | 
| Total | 53 | 100.00% | 3 | 100.00% | 
static inline u32 iwch_ib_to_tpt_bind_access(int acc)
{
	return (acc & IB_ACCESS_REMOTE_WRITE ? TPT_REMOTE_WRITE : 0) |
	       (acc & IB_ACCESS_REMOTE_READ ? TPT_REMOTE_READ : 0);
}
Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Steve Wise | 31 | 100.00% | 1 | 100.00% | 
| Total | 31 | 100.00% | 1 | 100.00% | 
enum iwch_mmid_state {
	
IWCH_STAG_STATE_VALID,
	
IWCH_STAG_STATE_INVALID
};
enum iwch_qp_query_flags {
	
IWCH_QP_QUERY_CONTEXT_NONE = 0x0,	/* No ctx; Only attrs */
	
IWCH_QP_QUERY_CONTEXT_GET = 0x1,	/* Get ctx + attrs */
	
IWCH_QP_QUERY_CONTEXT_SUSPEND = 0x2,	/* Not Supported */
	/*
         * Quiesce QP context; Consumer
         * will NOT replay outstanding WR
         */
	
IWCH_QP_QUERY_CONTEXT_QUIESCE = 0x4,
	
IWCH_QP_QUERY_CONTEXT_REMOVE = 0x8,
	
IWCH_QP_QUERY_TEST_USERWRITE = 0x32	/* Test special */
};
u16 iwch_rqes_posted(struct iwch_qp *qhp);
int iwch_post_send(struct ib_qp *ibqp, struct ib_send_wr *wr,
		      struct ib_send_wr **bad_wr);
int iwch_post_receive(struct ib_qp *ibqp, struct ib_recv_wr *wr,
		      struct ib_recv_wr **bad_wr);
int iwch_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc);
int iwch_post_terminate(struct iwch_qp *qhp, struct respQ_msg_t *rsp_msg);
int iwch_post_zb_read(struct iwch_ep *ep);
int iwch_register_device(struct iwch_dev *dev);
void iwch_unregister_device(struct iwch_dev *dev);
void stop_read_rep_timer(struct iwch_qp *qhp);
int iwch_register_mem(struct iwch_dev *rhp, struct iwch_pd *php,
		      struct iwch_mr *mhp, int shift);
int iwch_alloc_pbl(struct iwch_mr *mhp, int npages);
void iwch_free_pbl(struct iwch_mr *mhp);
int iwch_write_pbl(struct iwch_mr *mhp, __be64 *pages, int npages, int offset);
#define IWCH_NODE_DESC "cxgb3 Chelsio Communications"
#endif
Overall Contributors
| Person | Tokens | Prop | Commits | CommitProp | 
| Steve Wise | 1233 | 96.03% | 6 | 50.00% | 
| Roland Dreier | 37 | 2.88% | 2 | 16.67% | 
| Sagi Grimberg | 7 | 0.55% | 1 | 8.33% | 
| Kumar Sanghvi | 3 | 0.23% | 1 | 8.33% | 
| Joe Perches | 2 | 0.16% | 1 | 8.33% | 
| Harvey Harrison | 2 | 0.16% | 1 | 8.33% | 
| Total | 1284 | 100.00% | 12 | 100.00% | 
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.