Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Yuval Mintz | 2536 | 87.18% | 23 | 60.53% |
Manish Chopra | 218 | 7.49% | 5 | 13.16% |
Rahul Verma | 65 | 2.23% | 2 | 5.26% |
Prabhakar Kushwaha | 31 | 1.07% | 1 | 2.63% |
Shahed Shaikh | 30 | 1.03% | 1 | 2.63% |
Arnd Bergmann | 13 | 0.45% | 1 | 2.63% |
Sudarsana Reddy Kalluru | 8 | 0.28% | 1 | 2.63% |
Omkar Kulkarni | 3 | 0.10% | 1 | 2.63% |
Alexander Lobakin | 2 | 0.07% | 1 | 2.63% |
Jakub Kiciński | 2 | 0.07% | 1 | 2.63% |
Julia Lawall | 1 | 0.03% | 1 | 2.63% |
Total | 2909 | 38 |
/* SPDX-License-Identifier: (GPL-2.0-only OR BSD-3-Clause) */ /* QLogic qed NIC Driver * Copyright (c) 2015-2017 QLogic Corporation */ #ifndef _QED_VF_H #define _QED_VF_H #include "qed_l2.h" #include "qed_mcp.h" #define T_ETH_INDIRECTION_TABLE_SIZE 128 #define T_ETH_RSS_KEY_SIZE 10 struct vf_pf_resc_request { u8 num_rxqs; u8 num_txqs; u8 num_sbs; u8 num_mac_filters; u8 num_vlan_filters; u8 num_mc_filters; u8 num_cids; u8 padding; }; struct hw_sb_info { u16 hw_sb_id; u8 sb_qid; u8 padding[5]; }; #define TLV_BUFFER_SIZE 1024 enum { PFVF_STATUS_WAITING, PFVF_STATUS_SUCCESS, PFVF_STATUS_FAILURE, PFVF_STATUS_NOT_SUPPORTED, PFVF_STATUS_NO_RESOURCE, PFVF_STATUS_FORCED, PFVF_STATUS_MALICIOUS, }; /* vf pf channel tlvs */ /* general tlv header (used for both vf->pf request and pf->vf response) */ struct channel_tlv { u16 type; u16 length; }; /* header of first vf->pf tlv carries the offset used to calculate response * buffer address */ struct vfpf_first_tlv { struct channel_tlv tl; u32 padding; u64 reply_address; }; /* header of pf->vf tlvs, carries the status of handling the request */ struct pfvf_tlv { struct channel_tlv tl; u8 status; u8 padding[3]; }; /* response tlv used for most tlvs */ struct pfvf_def_resp_tlv { struct pfvf_tlv hdr; }; /* used to terminate and pad a tlv list */ struct channel_list_end_tlv { struct channel_tlv tl; u8 padding[4]; }; #define VFPF_ACQUIRE_OS_LINUX (0) #define VFPF_ACQUIRE_OS_WINDOWS (1) #define VFPF_ACQUIRE_OS_ESX (2) #define VFPF_ACQUIRE_OS_SOLARIS (3) #define VFPF_ACQUIRE_OS_LINUX_USERSPACE (4) struct vfpf_acquire_tlv { struct vfpf_first_tlv first_tlv; struct vf_pf_vfdev_info { #define VFPF_ACQUIRE_CAP_PRE_FP_HSI BIT(0) /* VF pre-FP hsi version */ #define VFPF_ACQUIRE_CAP_100G BIT(1) /* VF can support 100g */ /* A requirement for supporting multi-Tx queues on a single queue-zone, * VF would pass qids as additional information whenever passing queue * references. */ #define VFPF_ACQUIRE_CAP_QUEUE_QIDS BIT(2) /* The VF is using the physical bar. While this is mostly internal * to the VF, might affect the number of CIDs supported assuming * QUEUE_QIDS is set. */ #define VFPF_ACQUIRE_CAP_PHYSICAL_BAR BIT(3) u64 capabilities; u8 fw_major; u8 fw_minor; u8 fw_revision; u8 fw_engineering; u32 driver_version; u16 opaque_fid; /* ME register value */ u8 os_type; /* VFPF_ACQUIRE_OS_* value */ u8 eth_fp_hsi_major; u8 eth_fp_hsi_minor; u8 padding[3]; } vfdev_info; struct vf_pf_resc_request resc_request; u64 bulletin_addr; u32 bulletin_size; u32 padding; }; /* receive side scaling tlv */ struct vfpf_vport_update_rss_tlv { struct channel_tlv tl; u8 update_rss_flags; #define VFPF_UPDATE_RSS_CONFIG_FLAG BIT(0) #define VFPF_UPDATE_RSS_CAPS_FLAG BIT(1) #define VFPF_UPDATE_RSS_IND_TABLE_FLAG BIT(2) #define VFPF_UPDATE_RSS_KEY_FLAG BIT(3) u8 rss_enable; u8 rss_caps; u8 rss_table_size_log; /* The table size is 2 ^ rss_table_size_log */ u16 rss_ind_table[T_ETH_INDIRECTION_TABLE_SIZE]; u32 rss_key[T_ETH_RSS_KEY_SIZE]; }; struct pfvf_storm_stats { u32 address; u32 len; }; struct pfvf_stats_info { struct pfvf_storm_stats mstats; struct pfvf_storm_stats pstats; struct pfvf_storm_stats tstats; struct pfvf_storm_stats ustats; }; struct pfvf_acquire_resp_tlv { struct pfvf_tlv hdr; struct pf_vf_pfdev_info { u32 chip_num; u32 mfw_ver; u16 fw_major; u16 fw_minor; u16 fw_rev; u16 fw_eng; u64 capabilities; #define PFVF_ACQUIRE_CAP_DEFAULT_UNTAGGED BIT(0) #define PFVF_ACQUIRE_CAP_100G BIT(1) /* If set, 100g PF */ /* There are old PF versions where the PF might mistakenly override the sanity * mechanism [version-based] and allow a VF that can't be supported to pass * the acquisition phase. * To overcome this, PFs now indicate that they're past that point and the new * VFs would fail probe on the older PFs that fail to do so. */ #define PFVF_ACQUIRE_CAP_POST_FW_OVERRIDE BIT(2) /* PF expects queues to be received with additional qids */ #define PFVF_ACQUIRE_CAP_QUEUE_QIDS BIT(3) u16 db_size; u8 indices_per_sb; u8 os_type; /* These should match the PF's qed_dev values */ u16 chip_rev; u8 dev_type; /* Doorbell bar size configured in HW: log(size) or 0 */ u8 bar_size; struct pfvf_stats_info stats_info; u8 port_mac[ETH_ALEN]; /* It's possible PF had to configure an older fastpath HSI * [in case VF is newer than PF]. This is communicated back * to the VF. It can also be used in case of error due to * non-matching versions to shed light in VF about failure. */ u8 major_fp_hsi; u8 minor_fp_hsi; } pfdev_info; struct pf_vf_resc { #define PFVF_MAX_QUEUES_PER_VF 16 #define PFVF_MAX_SBS_PER_VF 16 struct hw_sb_info hw_sbs[PFVF_MAX_SBS_PER_VF]; u8 hw_qid[PFVF_MAX_QUEUES_PER_VF]; u8 cid[PFVF_MAX_QUEUES_PER_VF]; u8 num_rxqs; u8 num_txqs; u8 num_sbs; u8 num_mac_filters; u8 num_vlan_filters; u8 num_mc_filters; u8 num_cids; u8 padding; } resc; u32 bulletin_size; u32 padding; }; struct pfvf_start_queue_resp_tlv { struct pfvf_tlv hdr; u32 offset; /* offset to consumer/producer of queue */ u8 padding[4]; }; /* Extended queue information - additional index for reference inside qzone. * If communicated between VF/PF, each TLV relating to queues should be * extended by one such [or have a future base TLV that already contains info]. */ struct vfpf_qid_tlv { struct channel_tlv tl; u8 qid; u8 padding[3]; }; /* Setup Queue */ struct vfpf_start_rxq_tlv { struct vfpf_first_tlv first_tlv; /* physical addresses */ u64 rxq_addr; u64 deprecated_sge_addr; u64 cqe_pbl_addr; u16 cqe_pbl_size; u16 hw_sb; u16 rx_qid; u16 hc_rate; /* desired interrupts per sec. */ u16 bd_max_bytes; u16 stat_id; u8 sb_index; u8 padding[3]; }; struct vfpf_start_txq_tlv { struct vfpf_first_tlv first_tlv; /* physical addresses */ u64 pbl_addr; u16 pbl_size; u16 stat_id; u16 tx_qid; u16 hw_sb; u32 flags; /* VFPF_QUEUE_FLG_X flags */ u16 hc_rate; /* desired interrupts per sec. */ u8 sb_index; u8 padding[3]; }; /* Stop RX Queue */ struct vfpf_stop_rxqs_tlv { struct vfpf_first_tlv first_tlv; u16 rx_qid; /* this field is deprecated and should *always* be set to '1' */ u8 num_rxqs; u8 cqe_completion; u8 padding[4]; }; /* Stop TX Queues */ struct vfpf_stop_txqs_tlv { struct vfpf_first_tlv first_tlv; u16 tx_qid; /* this field is deprecated and should *always* be set to '1' */ u8 num_txqs; u8 padding[5]; }; struct vfpf_update_rxq_tlv { struct vfpf_first_tlv first_tlv; u64 deprecated_sge_addr[PFVF_MAX_QUEUES_PER_VF]; u16 rx_qid; u8 num_rxqs; u8 flags; #define VFPF_RXQ_UPD_INIT_SGE_DEPRECATE_FLAG BIT(0) #define VFPF_RXQ_UPD_COMPLETE_CQE_FLAG BIT(1) #define VFPF_RXQ_UPD_COMPLETE_EVENT_FLAG BIT(2) u8 padding[4]; }; /* Set Queue Filters */ struct vfpf_q_mac_vlan_filter { u32 flags; #define VFPF_Q_FILTER_DEST_MAC_VALID 0x01 #define VFPF_Q_FILTER_VLAN_TAG_VALID 0x02 #define VFPF_Q_FILTER_SET_MAC 0x100 /* set/clear */ u8 mac[ETH_ALEN]; u16 vlan_tag; u8 padding[4]; }; /* Start a vport */ struct vfpf_vport_start_tlv { struct vfpf_first_tlv first_tlv; u64 sb_addr[PFVF_MAX_SBS_PER_VF]; u32 tpa_mode; u16 dep1; u16 mtu; u8 vport_id; u8 inner_vlan_removal; u8 only_untagged; u8 max_buffers_per_cqe; u8 padding[4]; }; /* Extended tlvs - need to add rss, mcast, accept mode tlvs */ struct vfpf_vport_update_activate_tlv { struct channel_tlv tl; u8 update_rx; u8 update_tx; u8 active_rx; u8 active_tx; }; struct vfpf_vport_update_tx_switch_tlv { struct channel_tlv tl; u8 tx_switching; u8 padding[3]; }; struct vfpf_vport_update_vlan_strip_tlv { struct channel_tlv tl; u8 remove_vlan; u8 padding[3]; }; struct vfpf_vport_update_mcast_bin_tlv { struct channel_tlv tl; u8 padding[4]; /* There are only 256 approx bins, and in HSI they're divided into * 32-bit values. As old VFs used to set-bit to the values on its side, * the upper half of the array is never expected to contain any data. */ u64 bins[4]; u64 obsolete_bins[4]; }; struct vfpf_vport_update_accept_param_tlv { struct channel_tlv tl; u8 update_rx_mode; u8 update_tx_mode; u8 rx_accept_filter; u8 tx_accept_filter; }; struct vfpf_vport_update_accept_any_vlan_tlv { struct channel_tlv tl; u8 update_accept_any_vlan_flg; u8 accept_any_vlan; u8 padding[2]; }; struct vfpf_vport_update_sge_tpa_tlv { struct channel_tlv tl; u16 sge_tpa_flags; #define VFPF_TPA_IPV4_EN_FLAG BIT(0) #define VFPF_TPA_IPV6_EN_FLAG BIT(1) #define VFPF_TPA_PKT_SPLIT_FLAG BIT(2) #define VFPF_TPA_HDR_DATA_SPLIT_FLAG BIT(3) #define VFPF_TPA_GRO_CONSIST_FLAG BIT(4) u8 update_sge_tpa_flags; #define VFPF_UPDATE_SGE_DEPRECATED_FLAG BIT(0) #define VFPF_UPDATE_TPA_EN_FLAG BIT(1) #define VFPF_UPDATE_TPA_PARAM_FLAG BIT(2) u8 max_buffers_per_cqe; u16 deprecated_sge_buff_size; u16 tpa_max_size; u16 tpa_min_size_to_start; u16 tpa_min_size_to_cont; u8 tpa_max_aggs_num; u8 padding[7]; }; /* Primary tlv as a header for various extended tlvs for * various functionalities in vport update ramrod. */ struct vfpf_vport_update_tlv { struct vfpf_first_tlv first_tlv; }; struct vfpf_ucast_filter_tlv { struct vfpf_first_tlv first_tlv; u8 opcode; u8 type; u8 mac[ETH_ALEN]; u16 vlan; u16 padding[3]; }; /* tunnel update param tlv */ struct vfpf_update_tunn_param_tlv { struct vfpf_first_tlv first_tlv; u8 tun_mode_update_mask; u8 tunn_mode; u8 update_tun_cls; u8 vxlan_clss; u8 l2gre_clss; u8 ipgre_clss; u8 l2geneve_clss; u8 ipgeneve_clss; u8 update_geneve_port; u8 update_vxlan_port; u16 geneve_port; u16 vxlan_port; u8 padding[2]; }; struct pfvf_update_tunn_param_tlv { struct pfvf_tlv hdr; u16 tunn_feature_mask; u8 vxlan_mode; u8 l2geneve_mode; u8 ipgeneve_mode; u8 l2gre_mode; u8 ipgre_mode; u8 vxlan_clss; u8 l2gre_clss; u8 ipgre_clss; u8 l2geneve_clss; u8 ipgeneve_clss; u16 vxlan_udp_port; u16 geneve_udp_port; }; struct tlv_buffer_size { u8 tlv_buffer[TLV_BUFFER_SIZE]; }; struct vfpf_update_coalesce { struct vfpf_first_tlv first_tlv; u16 rx_coal; u16 tx_coal; u16 qid; u8 padding[2]; }; struct vfpf_read_coal_req_tlv { struct vfpf_first_tlv first_tlv; u16 qid; u8 is_rx; u8 padding[5]; }; struct pfvf_read_coal_resp_tlv { struct pfvf_tlv hdr; u16 coal; u8 padding[6]; }; struct vfpf_bulletin_update_mac_tlv { struct vfpf_first_tlv first_tlv; u8 mac[ETH_ALEN]; u8 padding[2]; }; union vfpf_tlvs { struct vfpf_first_tlv first_tlv; struct vfpf_acquire_tlv acquire; struct vfpf_start_rxq_tlv start_rxq; struct vfpf_start_txq_tlv start_txq; struct vfpf_stop_rxqs_tlv stop_rxqs; struct vfpf_stop_txqs_tlv stop_txqs; struct vfpf_update_rxq_tlv update_rxq; struct vfpf_vport_start_tlv start_vport; struct vfpf_vport_update_tlv vport_update; struct vfpf_ucast_filter_tlv ucast_filter; struct vfpf_update_tunn_param_tlv tunn_param_update; struct vfpf_update_coalesce update_coalesce; struct vfpf_read_coal_req_tlv read_coal_req; struct vfpf_bulletin_update_mac_tlv bulletin_update_mac; struct tlv_buffer_size tlv_buf_size; }; union pfvf_tlvs { struct pfvf_def_resp_tlv default_resp; struct pfvf_acquire_resp_tlv acquire_resp; struct tlv_buffer_size tlv_buf_size; struct pfvf_start_queue_resp_tlv queue_start; struct pfvf_update_tunn_param_tlv tunn_param_resp; struct pfvf_read_coal_resp_tlv read_coal_resp; }; enum qed_bulletin_bit { /* Alert the VF that a forced MAC was set by the PF */ MAC_ADDR_FORCED = 0, /* Alert the VF that a forced VLAN was set by the PF */ VLAN_ADDR_FORCED = 2, /* Indicate that `default_only_untagged' contains actual data */ VFPF_BULLETIN_UNTAGGED_DEFAULT = 3, VFPF_BULLETIN_UNTAGGED_DEFAULT_FORCED = 4, /* Alert the VF that suggested mac was sent by the PF. * MAC_ADDR will be disabled in case MAC_ADDR_FORCED is set. */ VFPF_BULLETIN_MAC_ADDR = 5 }; struct qed_bulletin_content { /* crc of structure to ensure is not in mid-update */ u32 crc; u32 version; /* bitmap indicating which fields hold valid values */ u64 valid_bitmap; /* used for MAC_ADDR or MAC_ADDR_FORCED */ u8 mac[ETH_ALEN]; /* If valid, 1 => only untagged Rx if no vlan is configured */ u8 default_only_untagged; u8 padding; /* The following is a 'copy' of qed_mcp_link_state, * qed_mcp_link_params and qed_mcp_link_capabilities. Since it's * possible the structs will increase further along the road we cannot * have it here; Instead we need to have all of its fields. */ u8 req_autoneg; u8 req_autoneg_pause; u8 req_forced_rx; u8 req_forced_tx; u8 padding2[4]; u32 req_adv_speed; u32 req_forced_speed; u32 req_loopback; u32 padding3; u8 link_up; u8 full_duplex; u8 autoneg; u8 autoneg_complete; u8 parallel_detection; u8 pfc_enabled; u8 partner_tx_flow_ctrl_en; u8 partner_rx_flow_ctrl_en; u8 partner_adv_pause; u8 sfp_tx_fault; u16 vxlan_udp_port; u16 geneve_udp_port; u8 padding4[2]; u32 speed; u32 partner_adv_speed; u32 capability_speed; /* Forced vlan */ u16 pvid; u16 padding5; }; struct qed_bulletin { dma_addr_t phys; struct qed_bulletin_content *p_virt; u32 size; }; enum { CHANNEL_TLV_NONE, /* ends tlv sequence */ CHANNEL_TLV_ACQUIRE, CHANNEL_TLV_VPORT_START, CHANNEL_TLV_VPORT_UPDATE, CHANNEL_TLV_VPORT_TEARDOWN, CHANNEL_TLV_START_RXQ, CHANNEL_TLV_START_TXQ, CHANNEL_TLV_STOP_RXQS, CHANNEL_TLV_STOP_TXQS, CHANNEL_TLV_UPDATE_RXQ, CHANNEL_TLV_INT_CLEANUP, CHANNEL_TLV_CLOSE, CHANNEL_TLV_RELEASE, CHANNEL_TLV_LIST_END, CHANNEL_TLV_UCAST_FILTER, CHANNEL_TLV_VPORT_UPDATE_ACTIVATE, CHANNEL_TLV_VPORT_UPDATE_TX_SWITCH, CHANNEL_TLV_VPORT_UPDATE_VLAN_STRIP, CHANNEL_TLV_VPORT_UPDATE_MCAST, CHANNEL_TLV_VPORT_UPDATE_ACCEPT_PARAM, CHANNEL_TLV_VPORT_UPDATE_RSS, CHANNEL_TLV_VPORT_UPDATE_ACCEPT_ANY_VLAN, CHANNEL_TLV_VPORT_UPDATE_SGE_TPA, CHANNEL_TLV_UPDATE_TUNN_PARAM, CHANNEL_TLV_COALESCE_UPDATE, CHANNEL_TLV_QID, CHANNEL_TLV_COALESCE_READ, CHANNEL_TLV_BULLETIN_UPDATE_MAC, CHANNEL_TLV_MAX, /* Required for iterating over vport-update tlvs. * Will break in case non-sequential vport-update tlvs. */ CHANNEL_TLV_VPORT_UPDATE_MAX = CHANNEL_TLV_VPORT_UPDATE_SGE_TPA + 1, }; /* Default number of CIDs [total of both Rx and Tx] to be requested * by default, and maximum possible number. */ #define QED_ETH_VF_DEFAULT_NUM_CIDS (32) #define QED_ETH_VF_MAX_NUM_CIDS (250) /* This data is held in the qed_hwfn structure for VFs only. */ struct qed_vf_iov { union vfpf_tlvs *vf2pf_request; dma_addr_t vf2pf_request_phys; union pfvf_tlvs *pf2vf_reply; dma_addr_t pf2vf_reply_phys; /* Should be taken whenever the mailbox buffers are accessed */ struct mutex mutex; u8 *offset; /* Bulletin Board */ struct qed_bulletin bulletin; struct qed_bulletin_content bulletin_shadow; /* we set aside a copy of the acquire response */ struct pfvf_acquire_resp_tlv acquire_resp; /* In case PF originates prior to the fp-hsi version comparison, * this has to be propagated as it affects the fastpath. */ bool b_pre_fp_hsi; /* Current day VFs are passing the SBs physical address on vport * start, and as they lack an IGU mapping they need to store the * addresses of previously registered SBs. * Even if we were to change configuration flow, due to backward * compatibility [with older PFs] we'd still need to store these. */ struct qed_sb_info *sbs_info[PFVF_MAX_SBS_PER_VF]; /* Determines whether VF utilizes doorbells via limited register * bar or via the doorbell bar. */ bool b_doorbell_bar; }; /** * qed_vf_pf_set_coalesce(): VF - Set Rx/Tx coalesce per VF's relative queue. * Coalesce value '0' will omit the * configuration. * * @p_hwfn: HW device data. * @rx_coal: coalesce value in micro second for rx queue. * @tx_coal: coalesce value in micro second for tx queue. * @p_cid: queue cid. * * Return: Int. * **/ int qed_vf_pf_set_coalesce(struct qed_hwfn *p_hwfn, u16 rx_coal, u16 tx_coal, struct qed_queue_cid *p_cid); /** * qed_vf_pf_get_coalesce(): VF - Get coalesce per VF's relative queue. * * @p_hwfn: HW device data. * @p_coal: coalesce value in micro second for VF queues. * @p_cid: queue cid. * * Return: Int. **/ int qed_vf_pf_get_coalesce(struct qed_hwfn *p_hwfn, u16 *p_coal, struct qed_queue_cid *p_cid); #ifdef CONFIG_QED_SRIOV /** * qed_vf_read_bulletin(): Read the VF bulletin and act on it if needed. * * @p_hwfn: HW device data. * @p_change: qed fills 1 iff bulletin board has changed, 0 otherwise. * * Return: enum _qed_status. */ int qed_vf_read_bulletin(struct qed_hwfn *p_hwfn, u8 *p_change); /** * qed_vf_get_link_params(): Get link parameters for VF from qed * * @p_hwfn: HW device data. * @params: the link params structure to be filled for the VF. * * Return: Void. */ void qed_vf_get_link_params(struct qed_hwfn *p_hwfn, struct qed_mcp_link_params *params); /** * qed_vf_get_link_state(): Get link state for VF from qed. * * @p_hwfn: HW device data. * @link: the link state structure to be filled for the VF * * Return: Void. */ void qed_vf_get_link_state(struct qed_hwfn *p_hwfn, struct qed_mcp_link_state *link); /** * qed_vf_get_link_caps(): Get link capabilities for VF from qed. * * @p_hwfn: HW device data. * @p_link_caps: the link capabilities structure to be filled for the VF * * Return: Void. */ void qed_vf_get_link_caps(struct qed_hwfn *p_hwfn, struct qed_mcp_link_capabilities *p_link_caps); /** * qed_vf_get_num_rxqs(): Get number of Rx queues allocated for VF by qed * * @p_hwfn: HW device data. * @num_rxqs: allocated RX queues * * Return: Void. */ void qed_vf_get_num_rxqs(struct qed_hwfn *p_hwfn, u8 *num_rxqs); /** * qed_vf_get_num_txqs(): Get number of Rx queues allocated for VF by qed * * @p_hwfn: HW device data. * @num_txqs: allocated RX queues * * Return: Void. */ void qed_vf_get_num_txqs(struct qed_hwfn *p_hwfn, u8 *num_txqs); /** * qed_vf_get_num_cids(): Get number of available connections * [both Rx and Tx] for VF * * @p_hwfn: HW device data. * @num_cids: allocated number of connections * * Return: Void. */ void qed_vf_get_num_cids(struct qed_hwfn *p_hwfn, u8 *num_cids); /** * qed_vf_get_port_mac(): Get port mac address for VF. * * @p_hwfn: HW device data. * @port_mac: destination location for port mac * * Return: Void. */ void qed_vf_get_port_mac(struct qed_hwfn *p_hwfn, u8 *port_mac); /** * qed_vf_get_num_vlan_filters(): Get number of VLAN filters allocated * for VF by qed. * * @p_hwfn: HW device data. * @num_vlan_filters: allocated VLAN filters * * Return: Void. */ void qed_vf_get_num_vlan_filters(struct qed_hwfn *p_hwfn, u8 *num_vlan_filters); /** * qed_vf_get_num_mac_filters(): Get number of MAC filters allocated * for VF by qed * * @p_hwfn: HW device data. * @num_mac_filters: allocated MAC filters * * Return: Void. */ void qed_vf_get_num_mac_filters(struct qed_hwfn *p_hwfn, u8 *num_mac_filters); /** * qed_vf_check_mac(): Check if VF can set a MAC address * * @p_hwfn: HW device data. * @mac: Mac. * * Return: bool. */ bool qed_vf_check_mac(struct qed_hwfn *p_hwfn, u8 *mac); /** * qed_vf_get_fw_version(): Set firmware version information * in dev_info from VFs acquire response tlv * * @p_hwfn: HW device data. * @fw_major: FW major. * @fw_minor: FW minor. * @fw_rev: FW rev. * @fw_eng: FW eng. * * Return: Void. */ void qed_vf_get_fw_version(struct qed_hwfn *p_hwfn, u16 *fw_major, u16 *fw_minor, u16 *fw_rev, u16 *fw_eng); /** * qed_vf_hw_prepare(): hw preparation for VF sends ACQUIRE message * * @p_hwfn: HW device data. * * Return: Int. */ int qed_vf_hw_prepare(struct qed_hwfn *p_hwfn); /** * qed_vf_pf_rxq_start(): start the RX Queue by sending a message to the PF * * @p_hwfn: HW device data. * @p_cid: Only relative fields are relevant * @bd_max_bytes: maximum number of bytes per bd * @bd_chain_phys_addr: physical address of bd chain * @cqe_pbl_addr: physical address of pbl * @cqe_pbl_size: pbl size * @pp_prod: pointer to the producer to be used in fastpath * * Return: Int. */ int qed_vf_pf_rxq_start(struct qed_hwfn *p_hwfn, struct qed_queue_cid *p_cid, u16 bd_max_bytes, dma_addr_t bd_chain_phys_addr, dma_addr_t cqe_pbl_addr, u16 cqe_pbl_size, void __iomem **pp_prod); /** * qed_vf_pf_txq_start(): VF - start the TX queue by sending a message to the * PF. * * @p_hwfn: HW device data. * @p_cid: CID. * @pbl_addr: PBL address. * @pbl_size: PBL Size. * @pp_doorbell: pointer to address to which to write the doorbell too. * * Return: Int. */ int qed_vf_pf_txq_start(struct qed_hwfn *p_hwfn, struct qed_queue_cid *p_cid, dma_addr_t pbl_addr, u16 pbl_size, void __iomem **pp_doorbell); /** * qed_vf_pf_rxq_stop(): VF - stop the RX queue by sending a message to the PF. * * @p_hwfn: HW device data. * @p_cid: CID. * @cqe_completion: CQE Completion. * * Return: Int. */ int qed_vf_pf_rxq_stop(struct qed_hwfn *p_hwfn, struct qed_queue_cid *p_cid, bool cqe_completion); /** * qed_vf_pf_txq_stop(): VF - stop the TX queue by sending a message to the PF. * * @p_hwfn: HW device data. * @p_cid: CID. * * Return: Int. */ int qed_vf_pf_txq_stop(struct qed_hwfn *p_hwfn, struct qed_queue_cid *p_cid); /** * qed_vf_pf_vport_update(): VF - send a vport update command. * * @p_hwfn: HW device data. * @p_params: Params * * Return: Int. */ int qed_vf_pf_vport_update(struct qed_hwfn *p_hwfn, struct qed_sp_vport_update_params *p_params); /** * qed_vf_pf_reset(): VF - send a close message to PF. * * @p_hwfn: HW device data. * * Return: enum _qed_status */ int qed_vf_pf_reset(struct qed_hwfn *p_hwfn); /** * qed_vf_pf_release(): VF - free vf`s memories. * * @p_hwfn: HW device data. * * Return: enum _qed_status */ int qed_vf_pf_release(struct qed_hwfn *p_hwfn); /** * qed_vf_get_igu_sb_id(): Get the IGU SB ID for a given * sb_id. For VFs igu sbs don't have to be contiguous * * @p_hwfn: HW device data. * @sb_id: SB ID. * * Return: INLINE u16 */ u16 qed_vf_get_igu_sb_id(struct qed_hwfn *p_hwfn, u16 sb_id); /** * qed_vf_set_sb_info(): Stores [or removes] a configured sb_info. * * @p_hwfn: HW device data. * @sb_id: zero-based SB index [for fastpath] * @p_sb: may be NULL [during removal]. * * Return: Void. */ void qed_vf_set_sb_info(struct qed_hwfn *p_hwfn, u16 sb_id, struct qed_sb_info *p_sb); /** * qed_vf_pf_vport_start(): perform vport start for VF. * * @p_hwfn: HW device data. * @vport_id: Vport ID. * @mtu: MTU. * @inner_vlan_removal: Innter VLAN removal. * @tpa_mode: TPA mode * @max_buffers_per_cqe: Max buffer pre CQE. * @only_untagged: default behavior regarding vlan acceptance * * Return: enum _qed_status */ int qed_vf_pf_vport_start(struct qed_hwfn *p_hwfn, u8 vport_id, u16 mtu, u8 inner_vlan_removal, enum qed_tpa_mode tpa_mode, u8 max_buffers_per_cqe, u8 only_untagged); /** * qed_vf_pf_vport_stop(): stop the VF's vport * * @p_hwfn: HW device data. * * Return: enum _qed_status */ int qed_vf_pf_vport_stop(struct qed_hwfn *p_hwfn); int qed_vf_pf_filter_ucast(struct qed_hwfn *p_hwfn, struct qed_filter_ucast *p_param); void qed_vf_pf_filter_mcast(struct qed_hwfn *p_hwfn, struct qed_filter_mcast *p_filter_cmd); /** * qed_vf_pf_int_cleanup(): clean the SB of the VF * * @p_hwfn: HW device data. * * Return: enum _qed_status */ int qed_vf_pf_int_cleanup(struct qed_hwfn *p_hwfn); /** * __qed_vf_get_link_params(): return the link params in a given bulletin board * * @p_hwfn: HW device data. * @p_params: pointer to a struct to fill with link params * @p_bulletin: Bulletin. * * Return: Void. */ void __qed_vf_get_link_params(struct qed_hwfn *p_hwfn, struct qed_mcp_link_params *p_params, struct qed_bulletin_content *p_bulletin); /** * __qed_vf_get_link_state(): return the link state in a given bulletin board * * @p_hwfn: HW device data. * @p_link: pointer to a struct to fill with link state * @p_bulletin: Bulletin. * * Return: Void. */ void __qed_vf_get_link_state(struct qed_hwfn *p_hwfn, struct qed_mcp_link_state *p_link, struct qed_bulletin_content *p_bulletin); /** * __qed_vf_get_link_caps(): return the link capabilities in a given * bulletin board * * @p_hwfn: HW device data. * @p_link_caps: pointer to a struct to fill with link capabilities * @p_bulletin: Bulletin. * * Return: Void. */ void __qed_vf_get_link_caps(struct qed_hwfn *p_hwfn, struct qed_mcp_link_capabilities *p_link_caps, struct qed_bulletin_content *p_bulletin); void qed_iov_vf_task(struct work_struct *work); void qed_vf_set_vf_start_tunn_update_param(struct qed_tunnel_info *p_tun); int qed_vf_pf_tunnel_param_update(struct qed_hwfn *p_hwfn, struct qed_tunnel_info *p_tunn); u32 qed_vf_hw_bar_size(struct qed_hwfn *p_hwfn, enum BAR_ID bar_id); /** * qed_vf_pf_bulletin_update_mac(): Ask PF to update the MAC address in * it's bulletin board * * @p_hwfn: HW device data. * @p_mac: mac address to be updated in bulletin board * * Return: Int. */ int qed_vf_pf_bulletin_update_mac(struct qed_hwfn *p_hwfn, const u8 *p_mac); #else static inline void qed_vf_get_link_params(struct qed_hwfn *p_hwfn, struct qed_mcp_link_params *params) { } static inline void qed_vf_get_link_state(struct qed_hwfn *p_hwfn, struct qed_mcp_link_state *link) { } static inline void qed_vf_get_link_caps(struct qed_hwfn *p_hwfn, struct qed_mcp_link_capabilities *p_link_caps) { } static inline void qed_vf_get_num_rxqs(struct qed_hwfn *p_hwfn, u8 *num_rxqs) { } static inline void qed_vf_get_num_txqs(struct qed_hwfn *p_hwfn, u8 *num_txqs) { } static inline void qed_vf_get_num_cids(struct qed_hwfn *p_hwfn, u8 *num_cids) { } static inline void qed_vf_get_port_mac(struct qed_hwfn *p_hwfn, u8 *port_mac) { } static inline void qed_vf_get_num_vlan_filters(struct qed_hwfn *p_hwfn, u8 *num_vlan_filters) { } static inline void qed_vf_get_num_mac_filters(struct qed_hwfn *p_hwfn, u8 *num_mac_filters) { } static inline bool qed_vf_check_mac(struct qed_hwfn *p_hwfn, u8 *mac) { return false; } static inline void qed_vf_get_fw_version(struct qed_hwfn *p_hwfn, u16 *fw_major, u16 *fw_minor, u16 *fw_rev, u16 *fw_eng) { } static inline int qed_vf_hw_prepare(struct qed_hwfn *p_hwfn) { return -EINVAL; } static inline int qed_vf_pf_rxq_start(struct qed_hwfn *p_hwfn, struct qed_queue_cid *p_cid, u16 bd_max_bytes, dma_addr_t bd_chain_phys_adr, dma_addr_t cqe_pbl_addr, u16 cqe_pbl_size, void __iomem **pp_prod) { return -EINVAL; } static inline int qed_vf_pf_txq_start(struct qed_hwfn *p_hwfn, struct qed_queue_cid *p_cid, dma_addr_t pbl_addr, u16 pbl_size, void __iomem **pp_doorbell) { return -EINVAL; } static inline int qed_vf_pf_rxq_stop(struct qed_hwfn *p_hwfn, struct qed_queue_cid *p_cid, bool cqe_completion) { return -EINVAL; } static inline int qed_vf_pf_txq_stop(struct qed_hwfn *p_hwfn, struct qed_queue_cid *p_cid) { return -EINVAL; } static inline int qed_vf_pf_vport_update(struct qed_hwfn *p_hwfn, struct qed_sp_vport_update_params *p_params) { return -EINVAL; } static inline int qed_vf_pf_reset(struct qed_hwfn *p_hwfn) { return -EINVAL; } static inline int qed_vf_pf_release(struct qed_hwfn *p_hwfn) { return -EINVAL; } static inline u16 qed_vf_get_igu_sb_id(struct qed_hwfn *p_hwfn, u16 sb_id) { return 0; } static inline void qed_vf_set_sb_info(struct qed_hwfn *p_hwfn, u16 sb_id, struct qed_sb_info *p_sb) { } static inline int qed_vf_pf_vport_start(struct qed_hwfn *p_hwfn, u8 vport_id, u16 mtu, u8 inner_vlan_removal, enum qed_tpa_mode tpa_mode, u8 max_buffers_per_cqe, u8 only_untagged) { return -EINVAL; } static inline int qed_vf_pf_vport_stop(struct qed_hwfn *p_hwfn) { return -EINVAL; } static inline int qed_vf_pf_filter_ucast(struct qed_hwfn *p_hwfn, struct qed_filter_ucast *p_param) { return -EINVAL; } static inline void qed_vf_pf_filter_mcast(struct qed_hwfn *p_hwfn, struct qed_filter_mcast *p_filter_cmd) { } static inline int qed_vf_pf_int_cleanup(struct qed_hwfn *p_hwfn) { return -EINVAL; } static inline void __qed_vf_get_link_params(struct qed_hwfn *p_hwfn, struct qed_mcp_link_params *p_params, struct qed_bulletin_content *p_bulletin) { } static inline void __qed_vf_get_link_state(struct qed_hwfn *p_hwfn, struct qed_mcp_link_state *p_link, struct qed_bulletin_content *p_bulletin) { } static inline void __qed_vf_get_link_caps(struct qed_hwfn *p_hwfn, struct qed_mcp_link_capabilities *p_link_caps, struct qed_bulletin_content *p_bulletin) { } static inline void qed_iov_vf_task(struct work_struct *work) { } static inline void qed_vf_set_vf_start_tunn_update_param(struct qed_tunnel_info *p_tun) { } static inline int qed_vf_pf_tunnel_param_update(struct qed_hwfn *p_hwfn, struct qed_tunnel_info *p_tunn) { return -EINVAL; } static inline int qed_vf_pf_bulletin_update_mac(struct qed_hwfn *p_hwfn, const u8 *p_mac) { return -EINVAL; } static inline u32 qed_vf_hw_bar_size(struct qed_hwfn *p_hwfn, enum BAR_ID bar_id) { return 0; } #endif #endif
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1