Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Varun Prakash | 3640 | 99.07% | 5 | 35.71% |
Hariprasad Shenai | 20 | 0.54% | 2 | 14.29% |
Nicholas Bellinger | 6 | 0.16% | 2 | 14.29% |
Bart Van Assche | 3 | 0.08% | 2 | 14.29% |
Thomas Gleixner | 2 | 0.05% | 1 | 7.14% |
Gustavo A. R. Silva | 2 | 0.05% | 1 | 7.14% |
Pankaj Bharadiya | 1 | 0.03% | 1 | 7.14% |
Total | 3674 | 14 |
// SPDX-License-Identifier: GPL-2.0-only /* * Copyright (c) 2016 Chelsio Communications, Inc. */ #define DRV_NAME "cxgbit" #define DRV_VERSION "1.0.0-ko" #define pr_fmt(fmt) DRV_NAME ": " fmt #include "cxgbit.h" #ifdef CONFIG_CHELSIO_T4_DCB #include <net/dcbevent.h> #include "cxgb4_dcb.h" #endif LIST_HEAD(cdev_list_head); /* cdev list lock */ DEFINE_MUTEX(cdev_list_lock); void _cxgbit_free_cdev(struct kref *kref) { struct cxgbit_device *cdev; cdev = container_of(kref, struct cxgbit_device, kref); cxgbi_ppm_release(cdev2ppm(cdev)); kfree(cdev); } static void cxgbit_set_mdsl(struct cxgbit_device *cdev) { struct cxgb4_lld_info *lldi = &cdev->lldi; u32 mdsl; #define ULP2_MAX_PKT_LEN 16224 #define ISCSI_PDU_NONPAYLOAD_LEN 312 mdsl = min_t(u32, lldi->iscsi_iolen - ISCSI_PDU_NONPAYLOAD_LEN, ULP2_MAX_PKT_LEN - ISCSI_PDU_NONPAYLOAD_LEN); mdsl = min_t(u32, mdsl, 8192); mdsl = min_t(u32, mdsl, (MAX_SKB_FRAGS - 1) * PAGE_SIZE); cdev->mdsl = mdsl; } static void *cxgbit_uld_add(const struct cxgb4_lld_info *lldi) { struct cxgbit_device *cdev; if (is_t4(lldi->adapter_type)) return ERR_PTR(-ENODEV); cdev = kzalloc(sizeof(*cdev), GFP_KERNEL); if (!cdev) return ERR_PTR(-ENOMEM); kref_init(&cdev->kref); spin_lock_init(&cdev->np_lock); cdev->lldi = *lldi; cxgbit_set_mdsl(cdev); if (cxgbit_ddp_init(cdev) < 0) { kfree(cdev); return ERR_PTR(-EINVAL); } if (!test_bit(CDEV_DDP_ENABLE, &cdev->flags)) pr_info("cdev %s ddp init failed\n", pci_name(lldi->pdev)); if (lldi->fw_vers >= 0x10d2b00) set_bit(CDEV_ISO_ENABLE, &cdev->flags); spin_lock_init(&cdev->cskq.lock); INIT_LIST_HEAD(&cdev->cskq.list); mutex_lock(&cdev_list_lock); list_add_tail(&cdev->list, &cdev_list_head); mutex_unlock(&cdev_list_lock); pr_info("cdev %s added for iSCSI target transport\n", pci_name(lldi->pdev)); return cdev; } static void cxgbit_close_conn(struct cxgbit_device *cdev) { struct cxgbit_sock *csk; struct sk_buff *skb; bool wakeup_thread = false; spin_lock_bh(&cdev->cskq.lock); list_for_each_entry(csk, &cdev->cskq.list, list) { skb = alloc_skb(0, GFP_ATOMIC); if (!skb) continue; spin_lock_bh(&csk->rxq.lock); __skb_queue_tail(&csk->rxq, skb); if (skb_queue_len(&csk->rxq) == 1) wakeup_thread = true; spin_unlock_bh(&csk->rxq.lock); if (wakeup_thread) { wake_up(&csk->waitq); wakeup_thread = false; } } spin_unlock_bh(&cdev->cskq.lock); } static void cxgbit_detach_cdev(struct cxgbit_device *cdev) { bool free_cdev = false; spin_lock_bh(&cdev->cskq.lock); if (list_empty(&cdev->cskq.list)) free_cdev = true; spin_unlock_bh(&cdev->cskq.lock); if (free_cdev) { mutex_lock(&cdev_list_lock); list_del(&cdev->list); mutex_unlock(&cdev_list_lock); cxgbit_put_cdev(cdev); } else { cxgbit_close_conn(cdev); } } static int cxgbit_uld_state_change(void *handle, enum cxgb4_state state) { struct cxgbit_device *cdev = handle; switch (state) { case CXGB4_STATE_UP: set_bit(CDEV_STATE_UP, &cdev->flags); pr_info("cdev %s state UP.\n", pci_name(cdev->lldi.pdev)); break; case CXGB4_STATE_START_RECOVERY: clear_bit(CDEV_STATE_UP, &cdev->flags); cxgbit_close_conn(cdev); pr_info("cdev %s state RECOVERY.\n", pci_name(cdev->lldi.pdev)); break; case CXGB4_STATE_DOWN: pr_info("cdev %s state DOWN.\n", pci_name(cdev->lldi.pdev)); break; case CXGB4_STATE_DETACH: clear_bit(CDEV_STATE_UP, &cdev->flags); pr_info("cdev %s state DETACH.\n", pci_name(cdev->lldi.pdev)); cxgbit_detach_cdev(cdev); break; default: pr_info("cdev %s unknown state %d.\n", pci_name(cdev->lldi.pdev), state); break; } return 0; } static void cxgbit_process_ddpvld(struct cxgbit_sock *csk, struct cxgbit_lro_pdu_cb *pdu_cb, u32 ddpvld) { if (ddpvld & (1 << CPL_RX_ISCSI_DDP_STATUS_HCRC_SHIFT)) { pr_info("tid 0x%x, status 0x%x, hcrc bad.\n", csk->tid, ddpvld); pdu_cb->flags |= PDUCBF_RX_HCRC_ERR; } if (ddpvld & (1 << CPL_RX_ISCSI_DDP_STATUS_DCRC_SHIFT)) { pr_info("tid 0x%x, status 0x%x, dcrc bad.\n", csk->tid, ddpvld); pdu_cb->flags |= PDUCBF_RX_DCRC_ERR; } if (ddpvld & (1 << CPL_RX_ISCSI_DDP_STATUS_PAD_SHIFT)) pr_info("tid 0x%x, status 0x%x, pad bad.\n", csk->tid, ddpvld); if ((ddpvld & (1 << CPL_RX_ISCSI_DDP_STATUS_DDP_SHIFT)) && (!(pdu_cb->flags & PDUCBF_RX_DATA))) { pdu_cb->flags |= PDUCBF_RX_DATA_DDPD; } } static void cxgbit_lro_add_packet_rsp(struct sk_buff *skb, u8 op, const __be64 *rsp) { struct cxgbit_lro_cb *lro_cb = cxgbit_skb_lro_cb(skb); struct cxgbit_lro_pdu_cb *pdu_cb = cxgbit_skb_lro_pdu_cb(skb, lro_cb->pdu_idx); struct cpl_rx_iscsi_ddp *cpl = (struct cpl_rx_iscsi_ddp *)(rsp + 1); cxgbit_process_ddpvld(lro_cb->csk, pdu_cb, be32_to_cpu(cpl->ddpvld)); pdu_cb->flags |= PDUCBF_RX_STATUS; pdu_cb->ddigest = ntohl(cpl->ulp_crc); pdu_cb->pdulen = ntohs(cpl->len); if (pdu_cb->flags & PDUCBF_RX_HDR) pdu_cb->complete = true; lro_cb->pdu_totallen += pdu_cb->pdulen; lro_cb->complete = true; lro_cb->pdu_idx++; } static void cxgbit_copy_frags(struct sk_buff *skb, const struct pkt_gl *gl, unsigned int offset) { u8 skb_frag_idx = skb_shinfo(skb)->nr_frags; u8 i; /* usually there's just one frag */ __skb_fill_page_desc(skb, skb_frag_idx, gl->frags[0].page, gl->frags[0].offset + offset, gl->frags[0].size - offset); for (i = 1; i < gl->nfrags; i++) __skb_fill_page_desc(skb, skb_frag_idx + i, gl->frags[i].page, gl->frags[i].offset, gl->frags[i].size); skb_shinfo(skb)->nr_frags += gl->nfrags; /* get a reference to the last page, we don't own it */ get_page(gl->frags[gl->nfrags - 1].page); } static void cxgbit_lro_add_packet_gl(struct sk_buff *skb, u8 op, const struct pkt_gl *gl) { struct cxgbit_lro_cb *lro_cb = cxgbit_skb_lro_cb(skb); struct cxgbit_lro_pdu_cb *pdu_cb = cxgbit_skb_lro_pdu_cb(skb, lro_cb->pdu_idx); u32 len, offset; if (op == CPL_ISCSI_HDR) { struct cpl_iscsi_hdr *cpl = (struct cpl_iscsi_hdr *)gl->va; offset = sizeof(struct cpl_iscsi_hdr); pdu_cb->flags |= PDUCBF_RX_HDR; pdu_cb->seq = ntohl(cpl->seq); len = ntohs(cpl->len); pdu_cb->hdr = gl->va + offset; pdu_cb->hlen = len; pdu_cb->hfrag_idx = skb_shinfo(skb)->nr_frags; if (unlikely(gl->nfrags > 1)) cxgbit_skcb_flags(skb) = 0; lro_cb->complete = false; } else if (op == CPL_ISCSI_DATA) { struct cpl_iscsi_data *cpl = (struct cpl_iscsi_data *)gl->va; offset = sizeof(struct cpl_iscsi_data); pdu_cb->flags |= PDUCBF_RX_DATA; len = ntohs(cpl->len); pdu_cb->dlen = len; pdu_cb->doffset = lro_cb->offset; pdu_cb->nr_dfrags = gl->nfrags; pdu_cb->dfrag_idx = skb_shinfo(skb)->nr_frags; lro_cb->complete = false; } else { struct cpl_rx_iscsi_cmp *cpl; cpl = (struct cpl_rx_iscsi_cmp *)gl->va; offset = sizeof(struct cpl_rx_iscsi_cmp); pdu_cb->flags |= (PDUCBF_RX_HDR | PDUCBF_RX_STATUS); len = be16_to_cpu(cpl->len); pdu_cb->hdr = gl->va + offset; pdu_cb->hlen = len; pdu_cb->hfrag_idx = skb_shinfo(skb)->nr_frags; pdu_cb->ddigest = be32_to_cpu(cpl->ulp_crc); pdu_cb->pdulen = ntohs(cpl->len); if (unlikely(gl->nfrags > 1)) cxgbit_skcb_flags(skb) = 0; cxgbit_process_ddpvld(lro_cb->csk, pdu_cb, be32_to_cpu(cpl->ddpvld)); if (pdu_cb->flags & PDUCBF_RX_DATA_DDPD) { pdu_cb->flags |= PDUCBF_RX_DDP_CMP; pdu_cb->complete = true; } else if (pdu_cb->flags & PDUCBF_RX_DATA) { pdu_cb->complete = true; } lro_cb->pdu_totallen += pdu_cb->hlen + pdu_cb->dlen; lro_cb->complete = true; lro_cb->pdu_idx++; } cxgbit_copy_frags(skb, gl, offset); pdu_cb->frags += gl->nfrags; lro_cb->offset += len; skb->len += len; skb->data_len += len; skb->truesize += len; } static struct sk_buff * cxgbit_lro_init_skb(struct cxgbit_sock *csk, u8 op, const struct pkt_gl *gl, const __be64 *rsp, struct napi_struct *napi) { struct sk_buff *skb; struct cxgbit_lro_cb *lro_cb; skb = napi_alloc_skb(napi, LRO_SKB_MAX_HEADROOM); if (unlikely(!skb)) return NULL; memset(skb->data, 0, LRO_SKB_MAX_HEADROOM); cxgbit_skcb_flags(skb) |= SKCBF_RX_LRO; lro_cb = cxgbit_skb_lro_cb(skb); cxgbit_get_csk(csk); lro_cb->csk = csk; return skb; } static void cxgbit_queue_lro_skb(struct cxgbit_sock *csk, struct sk_buff *skb) { bool wakeup_thread = false; spin_lock(&csk->rxq.lock); __skb_queue_tail(&csk->rxq, skb); if (skb_queue_len(&csk->rxq) == 1) wakeup_thread = true; spin_unlock(&csk->rxq.lock); if (wakeup_thread) wake_up(&csk->waitq); } static void cxgbit_lro_flush(struct t4_lro_mgr *lro_mgr, struct sk_buff *skb) { struct cxgbit_lro_cb *lro_cb = cxgbit_skb_lro_cb(skb); struct cxgbit_sock *csk = lro_cb->csk; csk->lro_skb = NULL; __skb_unlink(skb, &lro_mgr->lroq); cxgbit_queue_lro_skb(csk, skb); cxgbit_put_csk(csk); lro_mgr->lro_pkts++; lro_mgr->lro_session_cnt--; } static void cxgbit_uld_lro_flush(struct t4_lro_mgr *lro_mgr) { struct sk_buff *skb; while ((skb = skb_peek(&lro_mgr->lroq))) cxgbit_lro_flush(lro_mgr, skb); } static int cxgbit_lro_receive(struct cxgbit_sock *csk, u8 op, const __be64 *rsp, const struct pkt_gl *gl, struct t4_lro_mgr *lro_mgr, struct napi_struct *napi) { struct sk_buff *skb; struct cxgbit_lro_cb *lro_cb; if (!csk) { pr_err("%s: csk NULL, op 0x%x.\n", __func__, op); goto out; } if (csk->lro_skb) goto add_packet; start_lro: if (lro_mgr->lro_session_cnt >= MAX_LRO_SESSIONS) { cxgbit_uld_lro_flush(lro_mgr); goto start_lro; } skb = cxgbit_lro_init_skb(csk, op, gl, rsp, napi); if (unlikely(!skb)) goto out; csk->lro_skb = skb; __skb_queue_tail(&lro_mgr->lroq, skb); lro_mgr->lro_session_cnt++; add_packet: skb = csk->lro_skb; lro_cb = cxgbit_skb_lro_cb(skb); if ((gl && (((skb_shinfo(skb)->nr_frags + gl->nfrags) > MAX_SKB_FRAGS) || (lro_cb->pdu_totallen >= LRO_FLUSH_LEN_MAX))) || (lro_cb->pdu_idx >= MAX_SKB_FRAGS)) { cxgbit_lro_flush(lro_mgr, skb); goto start_lro; } if (gl) cxgbit_lro_add_packet_gl(skb, op, gl); else cxgbit_lro_add_packet_rsp(skb, op, rsp); lro_mgr->lro_merged++; return 0; out: return -1; } static int cxgbit_uld_lro_rx_handler(void *hndl, const __be64 *rsp, const struct pkt_gl *gl, struct t4_lro_mgr *lro_mgr, struct napi_struct *napi) { struct cxgbit_device *cdev = hndl; struct cxgb4_lld_info *lldi = &cdev->lldi; struct cpl_tx_data *rpl = NULL; struct cxgbit_sock *csk = NULL; unsigned int tid = 0; struct sk_buff *skb; unsigned int op = *(u8 *)rsp; bool lro_flush = true; switch (op) { case CPL_ISCSI_HDR: case CPL_ISCSI_DATA: case CPL_RX_ISCSI_CMP: case CPL_RX_ISCSI_DDP: case CPL_FW4_ACK: lro_flush = false; fallthrough; case CPL_ABORT_RPL_RSS: case CPL_PASS_ESTABLISH: case CPL_PEER_CLOSE: case CPL_CLOSE_CON_RPL: case CPL_ABORT_REQ_RSS: case CPL_SET_TCB_RPL: case CPL_RX_DATA: rpl = gl ? (struct cpl_tx_data *)gl->va : (struct cpl_tx_data *)(rsp + 1); tid = GET_TID(rpl); csk = lookup_tid(lldi->tids, tid); break; default: break; } if (csk && csk->lro_skb && lro_flush) cxgbit_lro_flush(lro_mgr, csk->lro_skb); if (!gl) { unsigned int len; if (op == CPL_RX_ISCSI_DDP) { if (!cxgbit_lro_receive(csk, op, rsp, NULL, lro_mgr, napi)) return 0; } len = 64 - sizeof(struct rsp_ctrl) - 8; skb = napi_alloc_skb(napi, len); if (!skb) goto nomem; __skb_put(skb, len); skb_copy_to_linear_data(skb, &rsp[1], len); } else { if (unlikely(op != *(u8 *)gl->va)) { pr_info("? FL 0x%p,RSS%#llx,FL %#llx,len %u.\n", gl->va, be64_to_cpu(*rsp), get_unaligned_be64(gl->va), gl->tot_len); return 0; } if ((op == CPL_ISCSI_HDR) || (op == CPL_ISCSI_DATA) || (op == CPL_RX_ISCSI_CMP)) { if (!cxgbit_lro_receive(csk, op, rsp, gl, lro_mgr, napi)) return 0; } #define RX_PULL_LEN 128 skb = cxgb4_pktgl_to_skb(gl, RX_PULL_LEN, RX_PULL_LEN); if (unlikely(!skb)) goto nomem; } rpl = (struct cpl_tx_data *)skb->data; op = rpl->ot.opcode; cxgbit_skcb_rx_opcode(skb) = op; pr_debug("cdev %p, opcode 0x%x(0x%x,0x%x), skb %p.\n", cdev, op, rpl->ot.opcode_tid, ntohl(rpl->ot.opcode_tid), skb); if (op < NUM_CPL_CMDS && cxgbit_cplhandlers[op]) { cxgbit_cplhandlers[op](cdev, skb); } else { pr_err("No handler for opcode 0x%x.\n", op); __kfree_skb(skb); } return 0; nomem: pr_err("%s OOM bailing out.\n", __func__); return 1; } #ifdef CONFIG_CHELSIO_T4_DCB struct cxgbit_dcb_work { struct dcb_app_type dcb_app; struct work_struct work; }; static void cxgbit_update_dcb_priority(struct cxgbit_device *cdev, u8 port_id, u8 dcb_priority, u16 port_num) { struct cxgbit_sock *csk; struct sk_buff *skb; u16 local_port; bool wakeup_thread = false; spin_lock_bh(&cdev->cskq.lock); list_for_each_entry(csk, &cdev->cskq.list, list) { if (csk->port_id != port_id) continue; if (csk->com.local_addr.ss_family == AF_INET6) { struct sockaddr_in6 *sock_in6; sock_in6 = (struct sockaddr_in6 *)&csk->com.local_addr; local_port = ntohs(sock_in6->sin6_port); } else { struct sockaddr_in *sock_in; sock_in = (struct sockaddr_in *)&csk->com.local_addr; local_port = ntohs(sock_in->sin_port); } if (local_port != port_num) continue; if (csk->dcb_priority == dcb_priority) continue; skb = alloc_skb(0, GFP_ATOMIC); if (!skb) continue; spin_lock(&csk->rxq.lock); __skb_queue_tail(&csk->rxq, skb); if (skb_queue_len(&csk->rxq) == 1) wakeup_thread = true; spin_unlock(&csk->rxq.lock); if (wakeup_thread) { wake_up(&csk->waitq); wakeup_thread = false; } } spin_unlock_bh(&cdev->cskq.lock); } static void cxgbit_dcb_workfn(struct work_struct *work) { struct cxgbit_dcb_work *dcb_work; struct net_device *ndev; struct cxgbit_device *cdev = NULL; struct dcb_app_type *iscsi_app; u8 priority, port_id = 0xff; dcb_work = container_of(work, struct cxgbit_dcb_work, work); iscsi_app = &dcb_work->dcb_app; if (iscsi_app->dcbx & DCB_CAP_DCBX_VER_IEEE) { if ((iscsi_app->app.selector != IEEE_8021QAZ_APP_SEL_STREAM) && (iscsi_app->app.selector != IEEE_8021QAZ_APP_SEL_ANY)) goto out; priority = iscsi_app->app.priority; } else if (iscsi_app->dcbx & DCB_CAP_DCBX_VER_CEE) { if (iscsi_app->app.selector != DCB_APP_IDTYPE_PORTNUM) goto out; if (!iscsi_app->app.priority) goto out; priority = ffs(iscsi_app->app.priority) - 1; } else { goto out; } pr_debug("priority for ifid %d is %u\n", iscsi_app->ifindex, priority); ndev = dev_get_by_index(&init_net, iscsi_app->ifindex); if (!ndev) goto out; mutex_lock(&cdev_list_lock); cdev = cxgbit_find_device(ndev, &port_id); dev_put(ndev); if (!cdev) { mutex_unlock(&cdev_list_lock); goto out; } cxgbit_update_dcb_priority(cdev, port_id, priority, iscsi_app->app.protocol); mutex_unlock(&cdev_list_lock); out: kfree(dcb_work); } static int cxgbit_dcbevent_notify(struct notifier_block *nb, unsigned long action, void *data) { struct cxgbit_dcb_work *dcb_work; struct dcb_app_type *dcb_app = data; dcb_work = kzalloc(sizeof(*dcb_work), GFP_ATOMIC); if (!dcb_work) return NOTIFY_DONE; dcb_work->dcb_app = *dcb_app; INIT_WORK(&dcb_work->work, cxgbit_dcb_workfn); schedule_work(&dcb_work->work); return NOTIFY_OK; } #endif static enum target_prot_op cxgbit_get_sup_prot_ops(struct iscsi_conn *conn) { return TARGET_PROT_NORMAL; } static struct iscsit_transport cxgbit_transport = { .name = DRV_NAME, .transport_type = ISCSI_CXGBIT, .rdma_shutdown = false, .priv_size = sizeof(struct cxgbit_cmd), .owner = THIS_MODULE, .iscsit_setup_np = cxgbit_setup_np, .iscsit_accept_np = cxgbit_accept_np, .iscsit_free_np = cxgbit_free_np, .iscsit_free_conn = cxgbit_free_conn, .iscsit_get_login_rx = cxgbit_get_login_rx, .iscsit_put_login_tx = cxgbit_put_login_tx, .iscsit_immediate_queue = iscsit_immediate_queue, .iscsit_response_queue = iscsit_response_queue, .iscsit_get_dataout = iscsit_build_r2ts_for_cmd, .iscsit_queue_data_in = iscsit_queue_rsp, .iscsit_queue_status = iscsit_queue_rsp, .iscsit_xmit_pdu = cxgbit_xmit_pdu, .iscsit_get_r2t_ttt = cxgbit_get_r2t_ttt, .iscsit_get_rx_pdu = cxgbit_get_rx_pdu, .iscsit_validate_params = cxgbit_validate_params, .iscsit_unmap_cmd = cxgbit_unmap_cmd, .iscsit_aborted_task = iscsit_aborted_task, .iscsit_get_sup_prot_ops = cxgbit_get_sup_prot_ops, }; static struct cxgb4_uld_info cxgbit_uld_info = { .name = DRV_NAME, .nrxq = MAX_ULD_QSETS, .ntxq = MAX_ULD_QSETS, .rxq_size = 1024, .lro = true, .add = cxgbit_uld_add, .state_change = cxgbit_uld_state_change, .lro_rx_handler = cxgbit_uld_lro_rx_handler, .lro_flush = cxgbit_uld_lro_flush, }; #ifdef CONFIG_CHELSIO_T4_DCB static struct notifier_block cxgbit_dcbevent_nb = { .notifier_call = cxgbit_dcbevent_notify, }; #endif static int __init cxgbit_init(void) { cxgb4_register_uld(CXGB4_ULD_ISCSIT, &cxgbit_uld_info); iscsit_register_transport(&cxgbit_transport); #ifdef CONFIG_CHELSIO_T4_DCB pr_info("%s dcb enabled.\n", DRV_NAME); register_dcbevent_notifier(&cxgbit_dcbevent_nb); #endif BUILD_BUG_ON(sizeof_field(struct sk_buff, cb) < sizeof(union cxgbit_skb_cb)); return 0; } static void __exit cxgbit_exit(void) { struct cxgbit_device *cdev, *tmp; #ifdef CONFIG_CHELSIO_T4_DCB unregister_dcbevent_notifier(&cxgbit_dcbevent_nb); #endif mutex_lock(&cdev_list_lock); list_for_each_entry_safe(cdev, tmp, &cdev_list_head, list) { list_del(&cdev->list); cxgbit_put_cdev(cdev); } mutex_unlock(&cdev_list_lock); iscsit_unregister_transport(&cxgbit_transport); cxgb4_unregister_uld(CXGB4_ULD_ISCSIT); } module_init(cxgbit_init); module_exit(cxgbit_exit); MODULE_DESCRIPTION("Chelsio iSCSI target offload driver"); MODULE_AUTHOR("Chelsio Communications"); MODULE_VERSION(DRV_VERSION); MODULE_LICENSE("GPL");
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1