cregit-Linux how code gets into the kernel

Release 4.7 drivers/infiniband/hw/hfi1/verbs.h

/*
 * Copyright(c) 2015, 2016 Intel Corporation.
 *
 * This file is provided under a dual BSD/GPLv2 license.  When using or
 * redistributing this file, you may do so under either license.
 *
 * GPL LICENSE SUMMARY
 *
 * This program is free software; you can redistribute it and/or modify
 * it under the terms of version 2 of the GNU General Public License as
 * published by the Free Software Foundation.
 *
 * This program is distributed in the hope that it will be useful, but
 * WITHOUT ANY WARRANTY; without even the implied warranty of
 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
 * General Public License for more details.
 *
 * BSD LICENSE
 *
 * Redistribution and use in source and binary forms, with or without
 * modification, are permitted provided that the following conditions
 * are met:
 *
 *  - Redistributions of source code must retain the above copyright
 *    notice, this list of conditions and the following disclaimer.
 *  - Redistributions in binary form must reproduce the above copyright
 *    notice, this list of conditions and the following disclaimer in
 *    the documentation and/or other materials provided with the
 *    distribution.
 *  - Neither the name of Intel Corporation nor the names of its
 *    contributors may be used to endorse or promote products derived
 *    from this software without specific prior written permission.
 *
 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
 *
 */

#ifndef HFI1_VERBS_H

#define HFI1_VERBS_H

#include <linux/types.h>
#include <linux/seqlock.h>
#include <linux/kernel.h>
#include <linux/interrupt.h>
#include <linux/kref.h>
#include <linux/workqueue.h>
#include <linux/kthread.h>
#include <linux/completion.h>
#include <linux/slab.h>
#include <rdma/ib_pack.h>
#include <rdma/ib_user_verbs.h>
#include <rdma/ib_mad.h>
#include <rdma/rdma_vt.h>
#include <rdma/rdmavt_qp.h>
#include <rdma/rdmavt_cq.h>

struct hfi1_ctxtdata;
struct hfi1_pportdata;
struct hfi1_devdata;
struct hfi1_packet;

#include "iowait.h"


#define HFI1_MAX_RDMA_ATOMIC     16

#define HFI1_GUIDS_PER_PORT	5

/*
 * Increment this value if any changes that break userspace ABI
 * compatibility are made.
 */

#define HFI1_UVERBS_ABI_VERSION       2


#define IB_SEQ_NAK	(3 << 29)

/* AETH NAK opcode values */

#define IB_RNR_NAK                      0x20

#define IB_NAK_PSN_ERROR                0x60

#define IB_NAK_INVALID_REQUEST          0x61

#define IB_NAK_REMOTE_ACCESS_ERROR      0x62

#define IB_NAK_REMOTE_OPERATIONAL_ERROR 0x63

#define IB_NAK_INVALID_RD_REQUEST       0x64

/* IB Performance Manager status values */

#define IB_PMA_SAMPLE_STATUS_DONE       0x00

#define IB_PMA_SAMPLE_STATUS_STARTED    0x01

#define IB_PMA_SAMPLE_STATUS_RUNNING    0x02

/* Mandatory IB performance counter select values. */

#define IB_PMA_PORT_XMIT_DATA   cpu_to_be16(0x0001)

#define IB_PMA_PORT_RCV_DATA    cpu_to_be16(0x0002)

#define IB_PMA_PORT_XMIT_PKTS   cpu_to_be16(0x0003)

#define IB_PMA_PORT_RCV_PKTS    cpu_to_be16(0x0004)

#define IB_PMA_PORT_XMIT_WAIT   cpu_to_be16(0x0005)


#define HFI1_VENDOR_IPG		cpu_to_be16(0xFFA0)


#define IB_BTH_REQ_ACK		BIT(31)

#define IB_BTH_SOLICITED	BIT(23)

#define IB_BTH_MIG_REQ		BIT(22)


#define IB_GRH_VERSION		6

#define IB_GRH_VERSION_MASK	0xF

#define IB_GRH_VERSION_SHIFT	28

#define IB_GRH_TCLASS_MASK	0xFF

#define IB_GRH_TCLASS_SHIFT	20

#define IB_GRH_FLOW_MASK	0xFFFFF

#define IB_GRH_FLOW_SHIFT	0

#define IB_GRH_NEXT_HDR		0x1B


#define IB_DEFAULT_GID_PREFIX	cpu_to_be64(0xfe80000000000000ULL)

/* flags passed by hfi1_ib_rcv() */
enum {
	
HFI1_HAS_GRH = (1 << 0),
};


struct ib_reth {
	
__be64 vaddr;
	
__be32 rkey;
	
__be32 length;
} 
__packed;


struct ib_atomic_eth {
	
__be32 vaddr[2];        /* unaligned so access as 2 32-bit words */
	
__be32 rkey;
	
__be64 swap_data;
	
__be64 compare_data;
} 
__packed;


union ib_ehdrs {
	struct {
		
__be32 deth[2];
		
__be32 imm_data;
	} 
ud;
	struct {
		
struct ib_reth reth;
		
__be32 imm_data;
	} 
rc;
	struct {
		
__be32 aeth;
		
__be32 atomic_ack_eth[2];
	} 
at;
	
__be32 imm_data;
	
__be32 aeth;
	
__be32 ieth;
	
struct ib_atomic_eth atomic_eth;
}  
__packed;


struct hfi1_other_headers {
	
__be32 bth[3];
	
union ib_ehdrs u;
} 
__packed;

/*
 * Note that UD packets with a GRH header are 8+40+12+8 = 68 bytes
 * long (72 w/ imm_data).  Only the first 56 bytes of the IB header
 * will be in the eager header buffer.  The remaining 12 or 16 bytes
 * are in the data buffer.
 */

struct hfi1_ib_header {
	
__be16 lrh[4];
	union {
		struct {
			
struct ib_grh grh;
			
struct hfi1_other_headers oth;
		} 
l;
		
struct hfi1_other_headers oth;
	} 
u;
} 
__packed;


struct ahg_ib_header {
	
struct sdma_engine *sde;
	
u32 ahgdesc[2];
	
u16 tx_flags;
	
u8 ahgcount;
	
u8 ahgidx;
	
struct hfi1_ib_header ibh;
};


struct hfi1_pio_header {
	
__le64 pbc;
	
struct hfi1_ib_header hdr;
} 
__packed;

/*
 * hfi1 specific data structures that will be hidden from rvt after the queue
 * pair is made common
 */

struct hfi1_qp_priv {
	
struct ahg_ib_header *s_hdr;              /* next header to send */
	
struct sdma_engine *s_sde;                /* current sde */
	
struct send_context *s_sendcontext;       /* current sendcontext */
	
u8 s_sc;		                  /* SC[0..4] for next packet */
	
u8 r_adefered;                            /* number of acks defered */
	
struct iowait s_iowait;
	
struct timer_list s_rnr_timer;
	
struct rvt_qp *owner;
};

/*
 * This structure is used to hold commonly lookedup and computed values during
 * the send engine progress.
 */

struct hfi1_pkt_state {
	
struct hfi1_ibdev *dev;
	
struct hfi1_ibport *ibp;
	
struct hfi1_pportdata *ppd;
	
struct verbs_txreq *s_txreq;
	
unsigned long flags;
};


#define HFI1_PSN_CREDIT  16


struct hfi1_opcode_stats {
	
u64 n_packets;          /* number of packets */
	
u64 n_bytes;            /* total number of bytes */
};


struct hfi1_opcode_stats_perctx {
	
struct hfi1_opcode_stats stats[256];
};


static inline void inc_opstats( u32 tlen, struct hfi1_opcode_stats *stats) { #ifdef CONFIG_DEBUG_FS stats->n_bytes += tlen; stats->n_packets++; #endif }

Contributors

PersonTokensPropCommitsCommitProp
mike marciniszynmike marciniszyn31100.00%1100.00%
Total31100.00%1100.00%

struct hfi1_ibport { struct rvt_qp __rcu *qp[2]; struct rvt_ibport rvp; __be64 guids[HFI1_GUIDS_PER_PORT - 1]; /* writable GUIDs */ /* the first 16 entries are sl_to_vl for !OPA */ u8 sl_to_sc[32]; u8 sc_to_sl[32]; }; struct hfi1_ibdev { struct rvt_dev_info rdi; /* Must be first */ /* QP numbers are shared by all IB ports */ /* protect wait lists */ seqlock_t iowait_lock; struct list_head txwait; /* list for wait verbs_txreq */ struct list_head memwait; /* list for wait kernel memory */ struct list_head txreq_free; struct kmem_cache *verbs_txreq_cache; struct timer_list mem_timer; u64 n_piowait; u64 n_piodrain; u64 n_txwait; u64 n_kmem_wait; #ifdef CONFIG_DEBUG_FS /* per HFI debugfs */ struct dentry *hfi1_ibdev_dbg; /* per HFI symlinks to above */ struct dentry *hfi1_ibdev_link; #endif };
static inline struct hfi1_ibdev *to_idev(struct ib_device *ibdev) { struct rvt_dev_info *rdi; rdi = container_of(ibdev, struct rvt_dev_info, ibdev); return container_of(rdi, struct hfi1_ibdev, rdi); }

Contributors

PersonTokensPropCommitsCommitProp
mike marciniszynmike marciniszyn2354.76%150.00%
dennis dalessandrodennis dalessandro1945.24%150.00%
Total42100.00%2100.00%


static inline struct rvt_qp *iowait_to_qp(struct iowait *s_iowait) { struct hfi1_qp_priv *priv; priv = container_of(s_iowait, struct hfi1_qp_priv, s_iowait); return priv->owner; }

Contributors

PersonTokensPropCommitsCommitProp
dennis dalessandrodennis dalessandro36100.00%2100.00%
Total36100.00%2100.00%

/* * Send if not busy or waiting for I/O and either * a RC response is pending or we can process send work requests. */
static inline int hfi1_send_ok(struct rvt_qp *qp) { return !(qp->s_flags & (RVT_S_BUSY | RVT_S_ANY_WAIT_IO)) && (qp->s_hdrwords || (qp->s_flags & RVT_S_RESP_PENDING) || !(qp->s_flags & RVT_S_ANY_WAIT_SEND)); }

Contributors

PersonTokensPropCommitsCommitProp
mike marciniszynmike marciniszyn4489.80%133.33%
dennis dalessandrodennis dalessandro510.20%266.67%
Total49100.00%3100.00%

/* * This must be called with s_lock held. */ void hfi1_bad_pqkey(struct hfi1_ibport *ibp, __be16 trap_num, u32 key, u32 sl, u32 qp1, u32 qp2, u16 lid1, u16 lid2); void hfi1_cap_mask_chg(struct rvt_dev_info *rdi, u8 port_num); void hfi1_sys_guid_chg(struct hfi1_ibport *ibp); void hfi1_node_desc_chg(struct hfi1_ibport *ibp); int hfi1_process_mad(struct ib_device *ibdev, int mad_flags, u8 port, const struct ib_wc *in_wc, const struct ib_grh *in_grh, const struct ib_mad_hdr *in_mad, size_t in_mad_size, struct ib_mad_hdr *out_mad, size_t *out_mad_size, u16 *out_mad_pkey_index); /* * The PSN_MASK and PSN_SHIFT allow for * 1) comparing two PSNs * 2) returning the PSN with any upper bits masked * 3) returning the difference between to PSNs * * The number of significant bits in the PSN must * necessarily be at least one bit less than * the container holding the PSN. */ #ifndef CONFIG_HFI1_VERBS_31BIT_PSN #define PSN_MASK 0xFFFFFF #define PSN_SHIFT 8 #else #define PSN_MASK 0x7FFFFFFF #define PSN_SHIFT 1 #endif #define PSN_MODIFY_MASK 0xFFFFFF /* * Compare the lower 24 bits of the msn values. * Returns an integer <, ==, or > than zero. */
static inline int cmp_msn(u32 a, u32 b) { return (((int)a) - ((int)b)) << 8; }

Contributors

PersonTokensPropCommitsCommitProp
mike marciniszynmike marciniszyn32100.00%1100.00%
Total32100.00%1100.00%

/* * Compare two PSNs * Returns an integer <, ==, or > than zero. */
static inline int cmp_psn(u32 a, u32 b) { return (((int)a) - ((int)b)) << PSN_SHIFT; }

Contributors

PersonTokensPropCommitsCommitProp
mike marciniszynmike marciniszyn32100.00%1100.00%
Total32100.00%1100.00%

/* * Return masked PSN */
static inline u32 mask_psn(u32 a) { return a & PSN_MASK; }

Contributors

PersonTokensPropCommitsCommitProp
mike marciniszynmike marciniszyn15100.00%1100.00%
Total15100.00%1100.00%

/* * Return delta between two PSNs */
static inline u32 delta_psn(u32 a, u32 b) { return (((int)a - (int)b) << PSN_SHIFT) >> PSN_SHIFT; }

Contributors

PersonTokensPropCommitsCommitProp
mike marciniszynmike marciniszyn32100.00%1100.00%
Total32100.00%1100.00%

struct verbs_txreq; void hfi1_put_txreq(struct verbs_txreq *tx); int hfi1_verbs_send(struct rvt_qp *qp, struct hfi1_pkt_state *ps); void hfi1_copy_sge(struct rvt_sge_state *ss, void *data, u32 length, int release, int copy_last); void hfi1_skip_sge(struct rvt_sge_state *ss, u32 length, int release); void hfi1_cnp_rcv(struct hfi1_packet *packet); void hfi1_uc_rcv(struct hfi1_packet *packet); void hfi1_rc_rcv(struct hfi1_packet *packet); void hfi1_rc_hdrerr( struct hfi1_ctxtdata *rcd, struct hfi1_ib_header *hdr, u32 rcv_flags, struct rvt_qp *qp); u8 ah_to_sc(struct ib_device *ibdev, struct ib_ah_attr *ah_attr); struct ib_ah *hfi1_create_qp0_ah(struct hfi1_ibport *ibp, u16 dlid); void hfi1_rc_rnr_retry(unsigned long arg); void hfi1_add_rnr_timer(struct rvt_qp *qp, u32 to); void hfi1_rc_timeout(unsigned long arg); void hfi1_del_timers_sync(struct rvt_qp *qp); void hfi1_stop_rc_timers(struct rvt_qp *qp); void hfi1_rc_send_complete(struct rvt_qp *qp, struct hfi1_ib_header *hdr); void hfi1_rc_error(struct rvt_qp *qp, enum ib_wc_status err); void hfi1_ud_rcv(struct hfi1_packet *packet); int hfi1_lookup_pkey_idx(struct hfi1_ibport *ibp, u16 pkey); int hfi1_rvt_get_rwqe(struct rvt_qp *qp, int wr_id_only); void hfi1_migrate_qp(struct rvt_qp *qp); int hfi1_check_modify_qp(struct rvt_qp *qp, struct ib_qp_attr *attr, int attr_mask, struct ib_udata *udata); void hfi1_modify_qp(struct rvt_qp *qp, struct ib_qp_attr *attr, int attr_mask, struct ib_udata *udata); int hfi1_check_send_wqe(struct rvt_qp *qp, struct rvt_swqe *wqe); extern const u32 rc_only_opcode; extern const u32 uc_only_opcode;
static inline u8 get_opcode(struct hfi1_ib_header *h) { u16 lnh = be16_to_cpu(h->lrh[0]) & 3; if (lnh == IB_LNH_IBA_LOCAL) return be32_to_cpu(h->u.oth.bth[0]) >> 24; else return be32_to_cpu(h->u.l.oth.bth[0]) >> 24; }

Contributors

PersonTokensPropCommitsCommitProp
mike marciniszynmike marciniszyn70100.00%1100.00%
Total70100.00%1100.00%

int hfi1_ruc_check_hdr(struct hfi1_ibport *ibp, struct hfi1_ib_header *hdr, int has_grh, struct rvt_qp *qp, u32 bth0); u32 hfi1_make_grh(struct hfi1_ibport *ibp, struct ib_grh *hdr, struct ib_global_route *grh, u32 hwords, u32 nwords); void hfi1_make_ruc_header(struct rvt_qp *qp, struct hfi1_other_headers *ohdr, u32 bth0, u32 bth2, int middle, struct hfi1_pkt_state *ps); void _hfi1_do_send(struct work_struct *work); void hfi1_do_send(struct rvt_qp *qp); void hfi1_send_complete(struct rvt_qp *qp, struct rvt_swqe *wqe, enum ib_wc_status status); void hfi1_send_rc_ack(struct hfi1_ctxtdata *, struct rvt_qp *qp, int is_fecn); int hfi1_make_rc_req(struct rvt_qp *qp, struct hfi1_pkt_state *ps); int hfi1_make_uc_req(struct rvt_qp *qp, struct hfi1_pkt_state *ps); int hfi1_make_ud_req(struct rvt_qp *qp, struct hfi1_pkt_state *ps); int hfi1_register_ib_device(struct hfi1_devdata *); void hfi1_unregister_ib_device(struct hfi1_devdata *); void hfi1_ib_rcv(struct hfi1_packet *packet); unsigned hfi1_get_npkeys(struct hfi1_devdata *); int hfi1_verbs_send_dma(struct rvt_qp *qp, struct hfi1_pkt_state *ps, u64 pbc); int hfi1_verbs_send_pio(struct rvt_qp *qp, struct hfi1_pkt_state *ps, u64 pbc); int hfi1_wss_init(void); void hfi1_wss_exit(void); /* platform specific: return the lowest level cache (llc) size, in KiB */
static inline int wss_llc_size(void) { /* assume that the boot CPU value is universal for all CPUs */ return boot_cpu_data.x86_cache_size; }

Contributors

PersonTokensPropCommitsCommitProp
dean luickdean luick15100.00%1100.00%
Total15100.00%1100.00%

/* platform specific: cacheless copy */
static inline void cacheless_memcpy(void *dst, void *src, size_t n) { /* * Use the only available X64 cacheless copy. Add a __user cast * to quiet sparse. The src agument is already in the kernel so * there are no security issues. The extra fault recovery machinery * is not invoked. */ __copy_user_nocache(dst, (void __user *)src, n, 0); }

Contributors

PersonTokensPropCommitsCommitProp
dean luickdean luick35100.00%1100.00%
Total35100.00%1100.00%

extern const enum ib_wc_opcode ib_hfi1_wc_opcode[]; extern const u8 hdr_len_by_opcode[]; extern const int ib_rvt_state_ops[]; extern __be64 ib_hfi1_sys_image_guid; /* in network order */ extern unsigned int hfi1_max_cqes; extern unsigned int hfi1_max_cqs; extern unsigned int hfi1_max_qp_wrs; extern unsigned int hfi1_max_qps; extern unsigned int hfi1_max_sges; extern unsigned int hfi1_max_mcast_grps; extern unsigned int hfi1_max_mcast_qp_attached; extern unsigned int hfi1_max_srqs; extern unsigned int hfi1_max_srq_sges; extern unsigned int hfi1_max_srq_wrs; extern unsigned short piothreshold; extern const u32 ib_hfi1_rnr_table[]; #endif /* HFI1_VERBS_H */

Overall Contributors

PersonTokensPropCommitsCommitProp
mike marciniszynmike marciniszyn146681.31%830.77%
dennis dalessandrodennis dalessandro24113.37%1038.46%
dean luickdean luick673.72%27.69%
ira weinyira weiny110.61%13.85%
jubin johnjubin john80.44%27.69%
harish chegondiharish chegondi50.28%13.85%
jianxin xiongjianxin xiong30.17%13.85%
erik e. kahnerik e. kahn20.11%13.85%
Total1803100.00%26100.00%
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
{% endraw %}