Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Rasesh Mody | 2034 | 94.17% | 4 | 50.00% |
Ivan Vecera | 95 | 4.40% | 3 | 37.50% |
Ben Hutchings | 31 | 1.44% | 1 | 12.50% |
Total | 2160 | 8 |
/* * Linux network driver for QLogic BR-series Converged Network Adapter. * * This program is free software; you can redistribute it and/or modify it * under the terms of the GNU General Public License (GPL) Version 2 as * published by the Free Software Foundation * * This program is distributed in the hope that it will be useful, but * WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * General Public License for more details. */ /* * Copyright (c) 2005-2014 Brocade Communications Systems, Inc. * Copyright (c) 2014-2015 QLogic Corporation * All rights reserved * www.qlogic.com */ /* BNA Hardware and Firmware Interface */ /* Skipping statistics collection to avoid clutter. * Command is no longer needed: * MTU * TxQ Stop * RxQ Stop * RxF Enable/Disable * * HDS-off request is dynamic * keep structures as multiple of 32-bit fields for alignment. * All values must be written in big-endian. */ #ifndef __BFI_ENET_H__ #define __BFI_ENET_H__ #include "bfa_defs.h" #include "bfi.h" #define BFI_ENET_CFG_MAX 32 /* Max resources per PF */ #define BFI_ENET_TXQ_PRIO_MAX 8 #define BFI_ENET_RX_QSET_MAX 16 #define BFI_ENET_TXQ_WI_VECT_MAX 4 #define BFI_ENET_VLAN_ID_MAX 4096 #define BFI_ENET_VLAN_BLOCK_SIZE 512 /* in bits */ #define BFI_ENET_VLAN_BLOCKS_MAX \ (BFI_ENET_VLAN_ID_MAX / BFI_ENET_VLAN_BLOCK_SIZE) #define BFI_ENET_VLAN_WORD_SIZE 32 /* in bits */ #define BFI_ENET_VLAN_WORDS_MAX \ (BFI_ENET_VLAN_BLOCK_SIZE / BFI_ENET_VLAN_WORD_SIZE) #define BFI_ENET_RSS_RIT_MAX 64 /* entries */ #define BFI_ENET_RSS_KEY_LEN 10 /* 32-bit words */ union bfi_addr_be_u { struct { u32 addr_hi; /* Most Significant 32-bits */ u32 addr_lo; /* Least Significant 32-Bits */ } __packed a32; } __packed; /* T X Q U E U E D E F I N E S */ /* TxQ Vector (a.k.a. Tx-Buffer Descriptor) */ /* TxQ Entry Opcodes */ #define BFI_ENET_TXQ_WI_SEND (0x402) /* Single Frame Transmission */ #define BFI_ENET_TXQ_WI_SEND_LSO (0x403) /* Multi-Frame Transmission */ #define BFI_ENET_TXQ_WI_EXTENSION (0x104) /* Extension WI */ /* TxQ Entry Control Flags */ #define BFI_ENET_TXQ_WI_CF_FCOE_CRC BIT(8) #define BFI_ENET_TXQ_WI_CF_IPID_MODE BIT(5) #define BFI_ENET_TXQ_WI_CF_INS_PRIO BIT(4) #define BFI_ENET_TXQ_WI_CF_INS_VLAN BIT(3) #define BFI_ENET_TXQ_WI_CF_UDP_CKSUM BIT(2) #define BFI_ENET_TXQ_WI_CF_TCP_CKSUM BIT(1) #define BFI_ENET_TXQ_WI_CF_IP_CKSUM BIT(0) struct bfi_enet_txq_wi_base { u8 reserved; u8 num_vectors; /* number of vectors present */ u16 opcode; /* BFI_ENET_TXQ_WI_SEND or BFI_ENET_TXQ_WI_SEND_LSO */ u16 flags; /* OR of all the flags */ u16 l4_hdr_size_n_offset; u16 vlan_tag; u16 lso_mss; /* Only 14 LSB are valid */ u32 frame_length; /* Only 24 LSB are valid */ } __packed; struct bfi_enet_txq_wi_ext { u16 reserved; u16 opcode; /* BFI_ENET_TXQ_WI_EXTENSION */ u32 reserved2[3]; } __packed; struct bfi_enet_txq_wi_vector { /* Tx Buffer Descriptor */ u16 reserved; u16 length; /* Only 14 LSB are valid */ union bfi_addr_be_u addr; } __packed; /* TxQ Entry Structure */ struct bfi_enet_txq_entry { union { struct bfi_enet_txq_wi_base base; struct bfi_enet_txq_wi_ext ext; } __packed wi; struct bfi_enet_txq_wi_vector vector[BFI_ENET_TXQ_WI_VECT_MAX]; } __packed; #define wi_hdr wi.base #define wi_ext_hdr wi.ext #define BFI_ENET_TXQ_WI_L4_HDR_N_OFFSET(_hdr_size, _offset) \ (((_hdr_size) << 10) | ((_offset) & 0x3FF)) /* R X Q U E U E D E F I N E S */ struct bfi_enet_rxq_entry { union bfi_addr_be_u rx_buffer; } __packed; /* R X C O M P L E T I O N Q U E U E D E F I N E S */ /* CQ Entry Flags */ #define BFI_ENET_CQ_EF_MAC_ERROR BIT(0) #define BFI_ENET_CQ_EF_FCS_ERROR BIT(1) #define BFI_ENET_CQ_EF_TOO_LONG BIT(2) #define BFI_ENET_CQ_EF_FC_CRC_OK BIT(3) #define BFI_ENET_CQ_EF_RSVD1 BIT(4) #define BFI_ENET_CQ_EF_L4_CKSUM_OK BIT(5) #define BFI_ENET_CQ_EF_L3_CKSUM_OK BIT(6) #define BFI_ENET_CQ_EF_HDS_HEADER BIT(7) #define BFI_ENET_CQ_EF_UDP BIT(8) #define BFI_ENET_CQ_EF_TCP BIT(9) #define BFI_ENET_CQ_EF_IP_OPTIONS BIT(10) #define BFI_ENET_CQ_EF_IPV6 BIT(11) #define BFI_ENET_CQ_EF_IPV4 BIT(12) #define BFI_ENET_CQ_EF_VLAN BIT(13) #define BFI_ENET_CQ_EF_RSS BIT(14) #define BFI_ENET_CQ_EF_RSVD2 BIT(15) #define BFI_ENET_CQ_EF_MCAST_MATCH BIT(16) #define BFI_ENET_CQ_EF_MCAST BIT(17) #define BFI_ENET_CQ_EF_BCAST BIT(18) #define BFI_ENET_CQ_EF_REMOTE BIT(19) #define BFI_ENET_CQ_EF_LOCAL BIT(20) /* CQ Entry Structure */ struct bfi_enet_cq_entry { u32 flags; u16 vlan_tag; u16 length; u32 rss_hash; u8 valid; u8 reserved1; u8 reserved2; u8 rxq_id; } __packed; /* E N E T C O N T R O L P A T H C O M M A N D S */ struct bfi_enet_q { union bfi_addr_u pg_tbl; union bfi_addr_u first_entry; u16 pages; /* # of pages */ u16 page_sz; } __packed; struct bfi_enet_txq { struct bfi_enet_q q; u8 priority; u8 rsvd[3]; } __packed; struct bfi_enet_rxq { struct bfi_enet_q q; u16 rx_buffer_size; u16 rsvd; } __packed; struct bfi_enet_cq { struct bfi_enet_q q; } __packed; struct bfi_enet_ib_cfg { u8 int_pkt_dma; u8 int_enabled; u8 int_pkt_enabled; u8 continuous_coalescing; u8 msix; u8 rsvd[3]; u32 coalescing_timeout; u32 inter_pkt_timeout; u8 inter_pkt_count; u8 rsvd1[3]; } __packed; struct bfi_enet_ib { union bfi_addr_u index_addr; union { u16 msix_index; u16 intx_bitmask; } __packed intr; u16 rsvd; } __packed; /* ENET command messages */ enum bfi_enet_h2i_msgs { /* Rx Commands */ BFI_ENET_H2I_RX_CFG_SET_REQ = 1, BFI_ENET_H2I_RX_CFG_CLR_REQ = 2, BFI_ENET_H2I_RIT_CFG_REQ = 3, BFI_ENET_H2I_RSS_CFG_REQ = 4, BFI_ENET_H2I_RSS_ENABLE_REQ = 5, BFI_ENET_H2I_RX_PROMISCUOUS_REQ = 6, BFI_ENET_H2I_RX_DEFAULT_REQ = 7, BFI_ENET_H2I_MAC_UCAST_SET_REQ = 8, BFI_ENET_H2I_MAC_UCAST_CLR_REQ = 9, BFI_ENET_H2I_MAC_UCAST_ADD_REQ = 10, BFI_ENET_H2I_MAC_UCAST_DEL_REQ = 11, BFI_ENET_H2I_MAC_MCAST_ADD_REQ = 12, BFI_ENET_H2I_MAC_MCAST_DEL_REQ = 13, BFI_ENET_H2I_MAC_MCAST_FILTER_REQ = 14, BFI_ENET_H2I_RX_VLAN_SET_REQ = 15, BFI_ENET_H2I_RX_VLAN_STRIP_ENABLE_REQ = 16, /* Tx Commands */ BFI_ENET_H2I_TX_CFG_SET_REQ = 17, BFI_ENET_H2I_TX_CFG_CLR_REQ = 18, /* Port Commands */ BFI_ENET_H2I_PORT_ADMIN_UP_REQ = 19, BFI_ENET_H2I_SET_PAUSE_REQ = 20, BFI_ENET_H2I_DIAG_LOOPBACK_REQ = 21, /* Get Attributes Command */ BFI_ENET_H2I_GET_ATTR_REQ = 22, /* Statistics Commands */ BFI_ENET_H2I_STATS_GET_REQ = 23, BFI_ENET_H2I_STATS_CLR_REQ = 24, BFI_ENET_H2I_WOL_MAGIC_REQ = 25, BFI_ENET_H2I_WOL_FRAME_REQ = 26, BFI_ENET_H2I_MAX = 27, }; enum bfi_enet_i2h_msgs { /* Rx Responses */ BFI_ENET_I2H_RX_CFG_SET_RSP = BFA_I2HM(BFI_ENET_H2I_RX_CFG_SET_REQ), BFI_ENET_I2H_RX_CFG_CLR_RSP = BFA_I2HM(BFI_ENET_H2I_RX_CFG_CLR_REQ), BFI_ENET_I2H_RIT_CFG_RSP = BFA_I2HM(BFI_ENET_H2I_RIT_CFG_REQ), BFI_ENET_I2H_RSS_CFG_RSP = BFA_I2HM(BFI_ENET_H2I_RSS_CFG_REQ), BFI_ENET_I2H_RSS_ENABLE_RSP = BFA_I2HM(BFI_ENET_H2I_RSS_ENABLE_REQ), BFI_ENET_I2H_RX_PROMISCUOUS_RSP = BFA_I2HM(BFI_ENET_H2I_RX_PROMISCUOUS_REQ), BFI_ENET_I2H_RX_DEFAULT_RSP = BFA_I2HM(BFI_ENET_H2I_RX_DEFAULT_REQ), BFI_ENET_I2H_MAC_UCAST_SET_RSP = BFA_I2HM(BFI_ENET_H2I_MAC_UCAST_SET_REQ), BFI_ENET_I2H_MAC_UCAST_CLR_RSP = BFA_I2HM(BFI_ENET_H2I_MAC_UCAST_CLR_REQ), BFI_ENET_I2H_MAC_UCAST_ADD_RSP = BFA_I2HM(BFI_ENET_H2I_MAC_UCAST_ADD_REQ), BFI_ENET_I2H_MAC_UCAST_DEL_RSP = BFA_I2HM(BFI_ENET_H2I_MAC_UCAST_DEL_REQ), BFI_ENET_I2H_MAC_MCAST_ADD_RSP = BFA_I2HM(BFI_ENET_H2I_MAC_MCAST_ADD_REQ), BFI_ENET_I2H_MAC_MCAST_DEL_RSP = BFA_I2HM(BFI_ENET_H2I_MAC_MCAST_DEL_REQ), BFI_ENET_I2H_MAC_MCAST_FILTER_RSP = BFA_I2HM(BFI_ENET_H2I_MAC_MCAST_FILTER_REQ), BFI_ENET_I2H_RX_VLAN_SET_RSP = BFA_I2HM(BFI_ENET_H2I_RX_VLAN_SET_REQ), BFI_ENET_I2H_RX_VLAN_STRIP_ENABLE_RSP = BFA_I2HM(BFI_ENET_H2I_RX_VLAN_STRIP_ENABLE_REQ), /* Tx Responses */ BFI_ENET_I2H_TX_CFG_SET_RSP = BFA_I2HM(BFI_ENET_H2I_TX_CFG_SET_REQ), BFI_ENET_I2H_TX_CFG_CLR_RSP = BFA_I2HM(BFI_ENET_H2I_TX_CFG_CLR_REQ), /* Port Responses */ BFI_ENET_I2H_PORT_ADMIN_RSP = BFA_I2HM(BFI_ENET_H2I_PORT_ADMIN_UP_REQ), BFI_ENET_I2H_SET_PAUSE_RSP = BFA_I2HM(BFI_ENET_H2I_SET_PAUSE_REQ), BFI_ENET_I2H_DIAG_LOOPBACK_RSP = BFA_I2HM(BFI_ENET_H2I_DIAG_LOOPBACK_REQ), /* Attributes Response */ BFI_ENET_I2H_GET_ATTR_RSP = BFA_I2HM(BFI_ENET_H2I_GET_ATTR_REQ), /* Statistics Responses */ BFI_ENET_I2H_STATS_GET_RSP = BFA_I2HM(BFI_ENET_H2I_STATS_GET_REQ), BFI_ENET_I2H_STATS_CLR_RSP = BFA_I2HM(BFI_ENET_H2I_STATS_CLR_REQ), BFI_ENET_I2H_WOL_MAGIC_RSP = BFA_I2HM(BFI_ENET_H2I_WOL_MAGIC_REQ), BFI_ENET_I2H_WOL_FRAME_RSP = BFA_I2HM(BFI_ENET_H2I_WOL_FRAME_REQ), /* AENs */ BFI_ENET_I2H_LINK_DOWN_AEN = BFA_I2HM(BFI_ENET_H2I_MAX), BFI_ENET_I2H_LINK_UP_AEN = BFA_I2HM(BFI_ENET_H2I_MAX + 1), BFI_ENET_I2H_PORT_ENABLE_AEN = BFA_I2HM(BFI_ENET_H2I_MAX + 2), BFI_ENET_I2H_PORT_DISABLE_AEN = BFA_I2HM(BFI_ENET_H2I_MAX + 3), BFI_ENET_I2H_BW_UPDATE_AEN = BFA_I2HM(BFI_ENET_H2I_MAX + 4), }; /* The following error codes can be returned by the enet commands */ enum bfi_enet_err { BFI_ENET_CMD_OK = 0, BFI_ENET_CMD_FAIL = 1, BFI_ENET_CMD_DUP_ENTRY = 2, /* !< Duplicate entry in CAM */ BFI_ENET_CMD_CAM_FULL = 3, /* !< CAM is full */ BFI_ENET_CMD_NOT_OWNER = 4, /* !< Not permitted, b'cos not owner */ BFI_ENET_CMD_NOT_EXEC = 5, /* !< Was not sent to f/w at all */ BFI_ENET_CMD_WAITING = 6, /* !< Waiting for completion */ BFI_ENET_CMD_PORT_DISABLED = 7, /* !< port in disabled state */ }; /* Generic Request * * bfi_enet_req is used by: * BFI_ENET_H2I_RX_CFG_CLR_REQ * BFI_ENET_H2I_TX_CFG_CLR_REQ */ struct bfi_enet_req { struct bfi_msgq_mhdr mh; } __packed; /* Enable/Disable Request * * bfi_enet_enable_req is used by: * BFI_ENET_H2I_RSS_ENABLE_REQ (enet_id must be zero) * BFI_ENET_H2I_RX_PROMISCUOUS_REQ (enet_id must be zero) * BFI_ENET_H2I_RX_DEFAULT_REQ (enet_id must be zero) * BFI_ENET_H2I_RX_MAC_MCAST_FILTER_REQ * BFI_ENET_H2I_PORT_ADMIN_UP_REQ (enet_id must be zero) */ struct bfi_enet_enable_req { struct bfi_msgq_mhdr mh; u8 enable; /* 1 = enable; 0 = disable */ u8 rsvd[3]; } __packed; /* Generic Response */ struct bfi_enet_rsp { struct bfi_msgq_mhdr mh; u8 error; /*!< if error see cmd_offset */ u8 rsvd; u16 cmd_offset; /*!< offset to invalid parameter */ } __packed; /* GLOBAL CONFIGURATION */ /* bfi_enet_attr_req is used by: * BFI_ENET_H2I_GET_ATTR_REQ */ struct bfi_enet_attr_req { struct bfi_msgq_mhdr mh; } __packed; /* bfi_enet_attr_rsp is used by: * BFI_ENET_I2H_GET_ATTR_RSP */ struct bfi_enet_attr_rsp { struct bfi_msgq_mhdr mh; u8 error; /*!< if error see cmd_offset */ u8 rsvd; u16 cmd_offset; /*!< offset to invalid parameter */ u32 max_cfg; u32 max_ucmac; u32 rit_size; } __packed; /* Tx Configuration * * bfi_enet_tx_cfg is used by: * BFI_ENET_H2I_TX_CFG_SET_REQ */ enum bfi_enet_tx_vlan_mode { BFI_ENET_TX_VLAN_NOP = 0, BFI_ENET_TX_VLAN_INS = 1, BFI_ENET_TX_VLAN_WI = 2, }; struct bfi_enet_tx_cfg { u8 vlan_mode; /*!< processing mode */ u8 rsvd; u16 vlan_id; u8 admit_tagged_frame; u8 apply_vlan_filter; u8 add_to_vswitch; u8 rsvd1[1]; } __packed; struct bfi_enet_tx_cfg_req { struct bfi_msgq_mhdr mh; u8 num_queues; /* # of Tx Queues */ u8 rsvd[3]; struct { struct bfi_enet_txq q; struct bfi_enet_ib ib; } __packed q_cfg[BFI_ENET_TXQ_PRIO_MAX]; struct bfi_enet_ib_cfg ib_cfg; struct bfi_enet_tx_cfg tx_cfg; }; struct bfi_enet_tx_cfg_rsp { struct bfi_msgq_mhdr mh; u8 error; u8 hw_id; /* For debugging */ u8 rsvd[2]; struct { u32 q_dbell; /* PCI base address offset */ u32 i_dbell; /* PCI base address offset */ u8 hw_qid; /* For debugging */ u8 rsvd[3]; } __packed q_handles[BFI_ENET_TXQ_PRIO_MAX]; }; /* Rx Configuration * * bfi_enet_rx_cfg is used by: * BFI_ENET_H2I_RX_CFG_SET_REQ */ enum bfi_enet_rxq_type { BFI_ENET_RXQ_SINGLE = 1, BFI_ENET_RXQ_LARGE_SMALL = 2, BFI_ENET_RXQ_HDS = 3, BFI_ENET_RXQ_HDS_OPT_BASED = 4, }; enum bfi_enet_hds_type { BFI_ENET_HDS_FORCED = 0x01, BFI_ENET_HDS_IPV6_UDP = 0x02, BFI_ENET_HDS_IPV6_TCP = 0x04, BFI_ENET_HDS_IPV4_TCP = 0x08, BFI_ENET_HDS_IPV4_UDP = 0x10, }; struct bfi_enet_rx_cfg { u8 rxq_type; u8 rsvd[1]; u16 frame_size; struct { u8 max_header_size; u8 force_offset; u8 type; u8 rsvd1; } __packed hds; u8 multi_buffer; u8 strip_vlan; u8 drop_untagged; u8 rsvd2; } __packed; /* * Multicast frames are received on the ql of q-set index zero. * On the completion queue. RxQ ID = even is for large/data buffer queues * and RxQ ID = odd is for small/header buffer queues. */ struct bfi_enet_rx_cfg_req { struct bfi_msgq_mhdr mh; u8 num_queue_sets; /* # of Rx Queue Sets */ u8 rsvd[3]; struct { struct bfi_enet_rxq ql; /* large/data/single buffers */ struct bfi_enet_rxq qs; /* small/header buffers */ struct bfi_enet_cq cq; struct bfi_enet_ib ib; } __packed q_cfg[BFI_ENET_RX_QSET_MAX]; struct bfi_enet_ib_cfg ib_cfg; struct bfi_enet_rx_cfg rx_cfg; } __packed; struct bfi_enet_rx_cfg_rsp { struct bfi_msgq_mhdr mh; u8 error; u8 hw_id; /* For debugging */ u8 rsvd[2]; struct { u32 ql_dbell; /* PCI base address offset */ u32 qs_dbell; /* PCI base address offset */ u32 i_dbell; /* PCI base address offset */ u8 hw_lqid; /* For debugging */ u8 hw_sqid; /* For debugging */ u8 hw_cqid; /* For debugging */ u8 rsvd; } __packed q_handles[BFI_ENET_RX_QSET_MAX]; } __packed; /* RIT * * bfi_enet_rit_req is used by: * BFI_ENET_H2I_RIT_CFG_REQ */ struct bfi_enet_rit_req { struct bfi_msgq_mhdr mh; u16 size; /* number of table-entries used */ u8 rsvd[2]; u8 table[BFI_ENET_RSS_RIT_MAX]; } __packed; /* RSS * * bfi_enet_rss_cfg_req is used by: * BFI_ENET_H2I_RSS_CFG_REQ */ enum bfi_enet_rss_type { BFI_ENET_RSS_IPV6 = 0x01, BFI_ENET_RSS_IPV6_TCP = 0x02, BFI_ENET_RSS_IPV4 = 0x04, BFI_ENET_RSS_IPV4_TCP = 0x08 }; struct bfi_enet_rss_cfg { u8 type; u8 mask; u8 rsvd[2]; u32 key[BFI_ENET_RSS_KEY_LEN]; } __packed; struct bfi_enet_rss_cfg_req { struct bfi_msgq_mhdr mh; struct bfi_enet_rss_cfg cfg; } __packed; /* MAC Unicast * * bfi_enet_rx_vlan_req is used by: * BFI_ENET_H2I_MAC_UCAST_SET_REQ * BFI_ENET_H2I_MAC_UCAST_CLR_REQ * BFI_ENET_H2I_MAC_UCAST_ADD_REQ * BFI_ENET_H2I_MAC_UCAST_DEL_REQ */ struct bfi_enet_ucast_req { struct bfi_msgq_mhdr mh; u8 mac_addr[ETH_ALEN]; u8 rsvd[2]; } __packed; /* MAC Unicast + VLAN */ struct bfi_enet_mac_n_vlan_req { struct bfi_msgq_mhdr mh; u16 vlan_id; u8 mac_addr[ETH_ALEN]; } __packed; /* MAC Multicast * * bfi_enet_mac_mfilter_add_req is used by: * BFI_ENET_H2I_MAC_MCAST_ADD_REQ */ struct bfi_enet_mcast_add_req { struct bfi_msgq_mhdr mh; u8 mac_addr[ETH_ALEN]; u8 rsvd[2]; } __packed; /* bfi_enet_mac_mfilter_add_rsp is used by: * BFI_ENET_I2H_MAC_MCAST_ADD_RSP */ struct bfi_enet_mcast_add_rsp { struct bfi_msgq_mhdr mh; u8 error; u8 rsvd; u16 cmd_offset; u16 handle; u8 rsvd1[2]; } __packed; /* bfi_enet_mac_mfilter_del_req is used by: * BFI_ENET_H2I_MAC_MCAST_DEL_REQ */ struct bfi_enet_mcast_del_req { struct bfi_msgq_mhdr mh; u16 handle; u8 rsvd[2]; } __packed; /* VLAN * * bfi_enet_rx_vlan_req is used by: * BFI_ENET_H2I_RX_VLAN_SET_REQ */ struct bfi_enet_rx_vlan_req { struct bfi_msgq_mhdr mh; u8 block_idx; u8 rsvd[3]; u32 bit_mask[BFI_ENET_VLAN_WORDS_MAX]; } __packed; /* PAUSE * * bfi_enet_set_pause_req is used by: * BFI_ENET_H2I_SET_PAUSE_REQ */ struct bfi_enet_set_pause_req { struct bfi_msgq_mhdr mh; u8 rsvd[2]; u8 tx_pause; /* 1 = enable; 0 = disable */ u8 rx_pause; /* 1 = enable; 0 = disable */ } __packed; /* DIAGNOSTICS * * bfi_enet_diag_lb_req is used by: * BFI_ENET_H2I_DIAG_LOOPBACK */ struct bfi_enet_diag_lb_req { struct bfi_msgq_mhdr mh; u8 rsvd[2]; u8 mode; /* cable or Serdes */ u8 enable; /* 1 = enable; 0 = disable */ } __packed; /* enum for Loopback opmodes */ enum { BFI_ENET_DIAG_LB_OPMODE_EXT = 0, BFI_ENET_DIAG_LB_OPMODE_CBL = 1, }; /* STATISTICS * * bfi_enet_stats_req is used by: * BFI_ENET_H2I_STATS_GET_REQ * BFI_ENET_I2H_STATS_CLR_REQ */ struct bfi_enet_stats_req { struct bfi_msgq_mhdr mh; u16 stats_mask; u8 rsvd[2]; u32 rx_enet_mask; u32 tx_enet_mask; union bfi_addr_u host_buffer; } __packed; /* defines for "stats_mask" above. */ #define BFI_ENET_STATS_MAC BIT(0) /* !< MAC Statistics */ #define BFI_ENET_STATS_BPC BIT(1) /* !< Pause Stats from BPC */ #define BFI_ENET_STATS_RAD BIT(2) /* !< Rx Admission Statistics */ #define BFI_ENET_STATS_RX_FC BIT(3) /* !< Rx FC Stats from RxA */ #define BFI_ENET_STATS_TX_FC BIT(4) /* !< Tx FC Stats from TxA */ #define BFI_ENET_STATS_ALL 0x1f /* TxF Frame Statistics */ struct bfi_enet_stats_txf { u64 ucast_octets; u64 ucast; u64 ucast_vlan; u64 mcast_octets; u64 mcast; u64 mcast_vlan; u64 bcast_octets; u64 bcast; u64 bcast_vlan; u64 errors; u64 filter_vlan; /* frames filtered due to VLAN */ u64 filter_mac_sa; /* frames filtered due to SA check */ } __packed; /* RxF Frame Statistics */ struct bfi_enet_stats_rxf { u64 ucast_octets; u64 ucast; u64 ucast_vlan; u64 mcast_octets; u64 mcast; u64 mcast_vlan; u64 bcast_octets; u64 bcast; u64 bcast_vlan; u64 frame_drops; } __packed; /* FC Tx Frame Statistics */ struct bfi_enet_stats_fc_tx { u64 txf_ucast_octets; u64 txf_ucast; u64 txf_ucast_vlan; u64 txf_mcast_octets; u64 txf_mcast; u64 txf_mcast_vlan; u64 txf_bcast_octets; u64 txf_bcast; u64 txf_bcast_vlan; u64 txf_parity_errors; u64 txf_timeout; u64 txf_fid_parity_errors; } __packed; /* FC Rx Frame Statistics */ struct bfi_enet_stats_fc_rx { u64 rxf_ucast_octets; u64 rxf_ucast; u64 rxf_ucast_vlan; u64 rxf_mcast_octets; u64 rxf_mcast; u64 rxf_mcast_vlan; u64 rxf_bcast_octets; u64 rxf_bcast; u64 rxf_bcast_vlan; } __packed; /* RAD Frame Statistics */ struct bfi_enet_stats_rad { u64 rx_frames; u64 rx_octets; u64 rx_vlan_frames; u64 rx_ucast; u64 rx_ucast_octets; u64 rx_ucast_vlan; u64 rx_mcast; u64 rx_mcast_octets; u64 rx_mcast_vlan; u64 rx_bcast; u64 rx_bcast_octets; u64 rx_bcast_vlan; u64 rx_drops; } __packed; /* BPC Tx Registers */ struct bfi_enet_stats_bpc { /* transmit stats */ u64 tx_pause[8]; u64 tx_zero_pause[8]; /*!< Pause cancellation */ /*!<Pause initiation rather than retention */ u64 tx_first_pause[8]; /* receive stats */ u64 rx_pause[8]; u64 rx_zero_pause[8]; /*!< Pause cancellation */ /*!<Pause initiation rather than retention */ u64 rx_first_pause[8]; } __packed; /* MAC Rx Statistics */ struct bfi_enet_stats_mac { u64 stats_clr_cnt; /* times this stats cleared */ u64 frame_64; /* both rx and tx counter */ u64 frame_65_127; /* both rx and tx counter */ u64 frame_128_255; /* both rx and tx counter */ u64 frame_256_511; /* both rx and tx counter */ u64 frame_512_1023; /* both rx and tx counter */ u64 frame_1024_1518; /* both rx and tx counter */ u64 frame_1519_1522; /* both rx and tx counter */ /* receive stats */ u64 rx_bytes; u64 rx_packets; u64 rx_fcs_error; u64 rx_multicast; u64 rx_broadcast; u64 rx_control_frames; u64 rx_pause; u64 rx_unknown_opcode; u64 rx_alignment_error; u64 rx_frame_length_error; u64 rx_code_error; u64 rx_carrier_sense_error; u64 rx_undersize; u64 rx_oversize; u64 rx_fragments; u64 rx_jabber; u64 rx_drop; /* transmit stats */ u64 tx_bytes; u64 tx_packets; u64 tx_multicast; u64 tx_broadcast; u64 tx_pause; u64 tx_deferral; u64 tx_excessive_deferral; u64 tx_single_collision; u64 tx_muliple_collision; u64 tx_late_collision; u64 tx_excessive_collision; u64 tx_total_collision; u64 tx_pause_honored; u64 tx_drop; u64 tx_jabber; u64 tx_fcs_error; u64 tx_control_frame; u64 tx_oversize; u64 tx_undersize; u64 tx_fragments; } __packed; /* Complete statistics, DMAed from fw to host followed by * BFI_ENET_I2H_STATS_GET_RSP */ struct bfi_enet_stats { struct bfi_enet_stats_mac mac_stats; struct bfi_enet_stats_bpc bpc_stats; struct bfi_enet_stats_rad rad_stats; struct bfi_enet_stats_rad rlb_stats; struct bfi_enet_stats_fc_rx fc_rx_stats; struct bfi_enet_stats_fc_tx fc_tx_stats; struct bfi_enet_stats_rxf rxf_stats[BFI_ENET_CFG_MAX]; struct bfi_enet_stats_txf txf_stats[BFI_ENET_CFG_MAX]; } __packed; #endif /* __BFI_ENET_H__ */
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1