Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Rahul Lakkireddy | 3470 | 91.12% | 15 | 42.86% |
Dimitris Michailidis | 176 | 4.62% | 3 | 8.57% |
Hariprasad Shenai | 67 | 1.76% | 3 | 8.57% |
Casey Leedom | 28 | 0.74% | 1 | 2.86% |
Arjun V | 22 | 0.58% | 2 | 5.71% |
Anish Bhatt | 15 | 0.39% | 1 | 2.86% |
Vipul Pandya | 10 | 0.26% | 2 | 5.71% |
Vishal Kulkarni | 8 | 0.21% | 3 | 8.57% |
Dan Carpenter | 6 | 0.16% | 1 | 2.86% |
Yue haibing | 2 | 0.05% | 1 | 2.86% |
Ganesh Goudar | 2 | 0.05% | 1 | 2.86% |
Allen Pais | 1 | 0.03% | 1 | 2.86% |
zhengbin | 1 | 0.03% | 1 | 2.86% |
Total | 3808 | 35 |
// SPDX-License-Identifier: GPL-2.0-only /* Copyright (C) 2019 Chelsio Communications. All rights reserved. */ #include "cxgb4.h" #include "cxgb4_tc_mqprio.h" #include "sched.h" static int cxgb4_mqprio_validate(struct net_device *dev, struct tc_mqprio_qopt_offload *mqprio) { u64 min_rate = 0, max_rate = 0, max_link_rate; struct port_info *pi = netdev2pinfo(dev); struct adapter *adap = netdev2adap(dev); u32 speed, qcount = 0, qoffset = 0; u32 start_a, start_b, end_a, end_b; int ret; u8 i, j; if (!mqprio->qopt.num_tc) return 0; if (mqprio->qopt.hw != TC_MQPRIO_HW_OFFLOAD_TCS) { netdev_err(dev, "Only full TC hardware offload is supported\n"); return -EINVAL; } else if (mqprio->mode != TC_MQPRIO_MODE_CHANNEL) { netdev_err(dev, "Only channel mode offload is supported\n"); return -EINVAL; } else if (mqprio->shaper != TC_MQPRIO_SHAPER_BW_RATE) { netdev_err(dev, "Only bandwidth rate shaper supported\n"); return -EINVAL; } else if (mqprio->qopt.num_tc > adap->params.nsched_cls) { netdev_err(dev, "Only %u traffic classes supported by hardware\n", adap->params.nsched_cls); return -ERANGE; } ret = t4_get_link_params(pi, NULL, &speed, NULL); if (ret) { netdev_err(dev, "Failed to get link speed, ret: %d\n", ret); return -EINVAL; } /* Convert from Mbps to bps */ max_link_rate = (u64)speed * 1000 * 1000; for (i = 0; i < mqprio->qopt.num_tc; i++) { qoffset = max_t(u16, mqprio->qopt.offset[i], qoffset); qcount += mqprio->qopt.count[i]; start_a = mqprio->qopt.offset[i]; end_a = start_a + mqprio->qopt.count[i] - 1; for (j = i + 1; j < mqprio->qopt.num_tc; j++) { start_b = mqprio->qopt.offset[j]; end_b = start_b + mqprio->qopt.count[j] - 1; /* If queue count is 0, then the traffic * belonging to this class will not use * ETHOFLD queues. So, no need to validate * further. */ if (!mqprio->qopt.count[i]) break; if (!mqprio->qopt.count[j]) continue; if (max_t(u32, start_a, start_b) <= min_t(u32, end_a, end_b)) { netdev_err(dev, "Queues can't overlap across tc\n"); return -EINVAL; } } /* Convert byte per second to bits per second */ min_rate += (mqprio->min_rate[i] * 8); max_rate += (mqprio->max_rate[i] * 8); } if (qoffset >= adap->tids.neotids || qcount > adap->tids.neotids) return -ENOMEM; if (min_rate > max_link_rate || max_rate > max_link_rate) { netdev_err(dev, "Total Min/Max (%llu/%llu) Rate > supported (%llu)\n", min_rate, max_rate, max_link_rate); return -EINVAL; } return 0; } static int cxgb4_init_eosw_txq(struct net_device *dev, struct sge_eosw_txq *eosw_txq, u32 eotid, u32 hwqid) { struct adapter *adap = netdev2adap(dev); struct tx_sw_desc *ring; memset(eosw_txq, 0, sizeof(*eosw_txq)); ring = kcalloc(CXGB4_EOSW_TXQ_DEFAULT_DESC_NUM, sizeof(*ring), GFP_KERNEL); if (!ring) return -ENOMEM; eosw_txq->desc = ring; eosw_txq->ndesc = CXGB4_EOSW_TXQ_DEFAULT_DESC_NUM; spin_lock_init(&eosw_txq->lock); eosw_txq->state = CXGB4_EO_STATE_CLOSED; eosw_txq->eotid = eotid; eosw_txq->hwtid = adap->tids.eotid_base + eosw_txq->eotid; eosw_txq->cred = adap->params.ofldq_wr_cred; eosw_txq->hwqid = hwqid; eosw_txq->netdev = dev; tasklet_setup(&eosw_txq->qresume_tsk, cxgb4_ethofld_restart); return 0; } static void cxgb4_clean_eosw_txq(struct net_device *dev, struct sge_eosw_txq *eosw_txq) { struct adapter *adap = netdev2adap(dev); cxgb4_eosw_txq_free_desc(adap, eosw_txq, eosw_txq->ndesc); eosw_txq->pidx = 0; eosw_txq->last_pidx = 0; eosw_txq->cidx = 0; eosw_txq->last_cidx = 0; eosw_txq->flowc_idx = 0; eosw_txq->inuse = 0; eosw_txq->cred = adap->params.ofldq_wr_cred; eosw_txq->ncompl = 0; eosw_txq->last_compl = 0; eosw_txq->state = CXGB4_EO_STATE_CLOSED; } static void cxgb4_free_eosw_txq(struct net_device *dev, struct sge_eosw_txq *eosw_txq) { spin_lock_bh(&eosw_txq->lock); cxgb4_clean_eosw_txq(dev, eosw_txq); kfree(eosw_txq->desc); spin_unlock_bh(&eosw_txq->lock); tasklet_kill(&eosw_txq->qresume_tsk); } static int cxgb4_mqprio_alloc_hw_resources(struct net_device *dev) { struct port_info *pi = netdev2pinfo(dev); struct adapter *adap = netdev2adap(dev); struct sge_ofld_rxq *eorxq; struct sge_eohw_txq *eotxq; int ret, msix = 0; u32 i; /* Allocate ETHOFLD hardware queue structures if not done already */ if (!refcount_read(&adap->tc_mqprio->refcnt)) { adap->sge.eohw_rxq = kcalloc(adap->sge.eoqsets, sizeof(struct sge_ofld_rxq), GFP_KERNEL); if (!adap->sge.eohw_rxq) return -ENOMEM; adap->sge.eohw_txq = kcalloc(adap->sge.eoqsets, sizeof(struct sge_eohw_txq), GFP_KERNEL); if (!adap->sge.eohw_txq) { kfree(adap->sge.eohw_rxq); return -ENOMEM; } refcount_set(&adap->tc_mqprio->refcnt, 1); } else { refcount_inc(&adap->tc_mqprio->refcnt); } if (!(adap->flags & CXGB4_USING_MSIX)) msix = -((int)adap->sge.intrq.abs_id + 1); for (i = 0; i < pi->nqsets; i++) { eorxq = &adap->sge.eohw_rxq[pi->first_qset + i]; eotxq = &adap->sge.eohw_txq[pi->first_qset + i]; /* Allocate Rxqs for receiving ETHOFLD Tx completions */ if (msix >= 0) { msix = cxgb4_get_msix_idx_from_bmap(adap); if (msix < 0) { ret = msix; goto out_free_queues; } eorxq->msix = &adap->msix_info[msix]; snprintf(eorxq->msix->desc, sizeof(eorxq->msix->desc), "%s-eorxq%d", dev->name, i); } init_rspq(adap, &eorxq->rspq, CXGB4_EOHW_RXQ_DEFAULT_INTR_USEC, CXGB4_EOHW_RXQ_DEFAULT_PKT_CNT, CXGB4_EOHW_RXQ_DEFAULT_DESC_NUM, CXGB4_EOHW_RXQ_DEFAULT_DESC_SIZE); eorxq->fl.size = CXGB4_EOHW_FLQ_DEFAULT_DESC_NUM; ret = t4_sge_alloc_rxq(adap, &eorxq->rspq, false, dev, msix, &eorxq->fl, cxgb4_ethofld_rx_handler, NULL, 0); if (ret) goto out_free_queues; /* Allocate ETHOFLD hardware Txqs */ eotxq->q.size = CXGB4_EOHW_TXQ_DEFAULT_DESC_NUM; ret = t4_sge_alloc_ethofld_txq(adap, eotxq, dev, eorxq->rspq.cntxt_id); if (ret) goto out_free_queues; /* Allocate IRQs, set IRQ affinity, and start Rx */ if (adap->flags & CXGB4_USING_MSIX) { ret = request_irq(eorxq->msix->vec, t4_sge_intr_msix, 0, eorxq->msix->desc, &eorxq->rspq); if (ret) goto out_free_msix; cxgb4_set_msix_aff(adap, eorxq->msix->vec, &eorxq->msix->aff_mask, i); } if (adap->flags & CXGB4_FULL_INIT_DONE) cxgb4_enable_rx(adap, &eorxq->rspq); } return 0; out_free_msix: while (i-- > 0) { eorxq = &adap->sge.eohw_rxq[pi->first_qset + i]; if (adap->flags & CXGB4_FULL_INIT_DONE) cxgb4_quiesce_rx(&eorxq->rspq); if (adap->flags & CXGB4_USING_MSIX) { cxgb4_clear_msix_aff(eorxq->msix->vec, eorxq->msix->aff_mask); free_irq(eorxq->msix->vec, &eorxq->rspq); } } out_free_queues: for (i = 0; i < pi->nqsets; i++) { eorxq = &adap->sge.eohw_rxq[pi->first_qset + i]; eotxq = &adap->sge.eohw_txq[pi->first_qset + i]; if (eorxq->rspq.desc) free_rspq_fl(adap, &eorxq->rspq, &eorxq->fl); if (eorxq->msix) cxgb4_free_msix_idx_in_bmap(adap, eorxq->msix->idx); t4_sge_free_ethofld_txq(adap, eotxq); } if (refcount_dec_and_test(&adap->tc_mqprio->refcnt)) { kfree(adap->sge.eohw_txq); kfree(adap->sge.eohw_rxq); } return ret; } static void cxgb4_mqprio_free_hw_resources(struct net_device *dev) { struct port_info *pi = netdev2pinfo(dev); struct adapter *adap = netdev2adap(dev); struct sge_ofld_rxq *eorxq; struct sge_eohw_txq *eotxq; u32 i; /* Return if no ETHOFLD structures have been allocated yet */ if (!refcount_read(&adap->tc_mqprio->refcnt)) return; /* Return if no hardware queues have been allocated */ if (!adap->sge.eohw_rxq[pi->first_qset].rspq.desc) return; for (i = 0; i < pi->nqsets; i++) { eorxq = &adap->sge.eohw_rxq[pi->first_qset + i]; eotxq = &adap->sge.eohw_txq[pi->first_qset + i]; /* Device removal path will already disable NAPI * before unregistering netdevice. So, only disable * NAPI if we're not in device removal path */ if (!(adap->flags & CXGB4_SHUTTING_DOWN)) cxgb4_quiesce_rx(&eorxq->rspq); if (adap->flags & CXGB4_USING_MSIX) { cxgb4_clear_msix_aff(eorxq->msix->vec, eorxq->msix->aff_mask); free_irq(eorxq->msix->vec, &eorxq->rspq); cxgb4_free_msix_idx_in_bmap(adap, eorxq->msix->idx); } free_rspq_fl(adap, &eorxq->rspq, &eorxq->fl); t4_sge_free_ethofld_txq(adap, eotxq); } /* Free up ETHOFLD structures if there are no users */ if (refcount_dec_and_test(&adap->tc_mqprio->refcnt)) { kfree(adap->sge.eohw_txq); kfree(adap->sge.eohw_rxq); } } static int cxgb4_mqprio_alloc_tc(struct net_device *dev, struct tc_mqprio_qopt_offload *mqprio) { struct ch_sched_params p = { .type = SCHED_CLASS_TYPE_PACKET, .u.params.level = SCHED_CLASS_LEVEL_CL_RL, .u.params.mode = SCHED_CLASS_MODE_FLOW, .u.params.rateunit = SCHED_CLASS_RATEUNIT_BITS, .u.params.ratemode = SCHED_CLASS_RATEMODE_ABS, .u.params.class = SCHED_CLS_NONE, .u.params.weight = 0, .u.params.pktsize = dev->mtu, }; struct cxgb4_tc_port_mqprio *tc_port_mqprio; struct port_info *pi = netdev2pinfo(dev); struct adapter *adap = netdev2adap(dev); struct sched_class *e; int ret; u8 i; tc_port_mqprio = &adap->tc_mqprio->port_mqprio[pi->port_id]; p.u.params.channel = pi->tx_chan; for (i = 0; i < mqprio->qopt.num_tc; i++) { /* Convert from bytes per second to Kbps */ p.u.params.minrate = div_u64(mqprio->min_rate[i] * 8, 1000); p.u.params.maxrate = div_u64(mqprio->max_rate[i] * 8, 1000); /* Request larger burst buffer for smaller MTU, so * that hardware can work on more data per burst * cycle. */ if (dev->mtu <= ETH_DATA_LEN) p.u.params.burstsize = 8 * dev->mtu; e = cxgb4_sched_class_alloc(dev, &p); if (!e) { ret = -ENOMEM; goto out_err; } tc_port_mqprio->tc_hwtc_map[i] = e->idx; } return 0; out_err: while (i--) cxgb4_sched_class_free(dev, tc_port_mqprio->tc_hwtc_map[i]); return ret; } static void cxgb4_mqprio_free_tc(struct net_device *dev) { struct cxgb4_tc_port_mqprio *tc_port_mqprio; struct port_info *pi = netdev2pinfo(dev); struct adapter *adap = netdev2adap(dev); u8 i; tc_port_mqprio = &adap->tc_mqprio->port_mqprio[pi->port_id]; for (i = 0; i < tc_port_mqprio->mqprio.qopt.num_tc; i++) cxgb4_sched_class_free(dev, tc_port_mqprio->tc_hwtc_map[i]); } static int cxgb4_mqprio_class_bind(struct net_device *dev, struct sge_eosw_txq *eosw_txq, u8 tc) { struct ch_sched_flowc fe; int ret; init_completion(&eosw_txq->completion); fe.tid = eosw_txq->eotid; fe.class = tc; ret = cxgb4_sched_class_bind(dev, &fe, SCHED_FLOWC); if (ret) return ret; ret = wait_for_completion_timeout(&eosw_txq->completion, CXGB4_FLOWC_WAIT_TIMEOUT); if (!ret) return -ETIMEDOUT; return 0; } static void cxgb4_mqprio_class_unbind(struct net_device *dev, struct sge_eosw_txq *eosw_txq, u8 tc) { struct adapter *adap = netdev2adap(dev); struct ch_sched_flowc fe; /* If we're shutting down, interrupts are disabled and no completions * come back. So, skip waiting for completions in this scenario. */ if (!(adap->flags & CXGB4_SHUTTING_DOWN)) init_completion(&eosw_txq->completion); fe.tid = eosw_txq->eotid; fe.class = tc; cxgb4_sched_class_unbind(dev, &fe, SCHED_FLOWC); if (!(adap->flags & CXGB4_SHUTTING_DOWN)) wait_for_completion_timeout(&eosw_txq->completion, CXGB4_FLOWC_WAIT_TIMEOUT); } static int cxgb4_mqprio_enable_offload(struct net_device *dev, struct tc_mqprio_qopt_offload *mqprio) { struct cxgb4_tc_port_mqprio *tc_port_mqprio; u32 qoffset, qcount, tot_qcount, qid, hwqid; struct port_info *pi = netdev2pinfo(dev); struct adapter *adap = netdev2adap(dev); struct sge_eosw_txq *eosw_txq; int eotid, ret; u16 i, j; u8 hwtc; ret = cxgb4_mqprio_alloc_hw_resources(dev); if (ret) return -ENOMEM; tc_port_mqprio = &adap->tc_mqprio->port_mqprio[pi->port_id]; for (i = 0; i < mqprio->qopt.num_tc; i++) { qoffset = mqprio->qopt.offset[i]; qcount = mqprio->qopt.count[i]; for (j = 0; j < qcount; j++) { eotid = cxgb4_get_free_eotid(&adap->tids); if (eotid < 0) { ret = -ENOMEM; goto out_free_eotids; } qid = qoffset + j; hwqid = pi->first_qset + (eotid % pi->nqsets); eosw_txq = &tc_port_mqprio->eosw_txq[qid]; ret = cxgb4_init_eosw_txq(dev, eosw_txq, eotid, hwqid); if (ret) goto out_free_eotids; cxgb4_alloc_eotid(&adap->tids, eotid, eosw_txq); hwtc = tc_port_mqprio->tc_hwtc_map[i]; ret = cxgb4_mqprio_class_bind(dev, eosw_txq, hwtc); if (ret) goto out_free_eotids; } } memcpy(&tc_port_mqprio->mqprio, mqprio, sizeof(struct tc_mqprio_qopt_offload)); /* Inform the stack about the configured tc params. * * Set the correct queue map. If no queue count has been * specified, then send the traffic through default NIC * queues; instead of ETHOFLD queues. */ ret = netdev_set_num_tc(dev, mqprio->qopt.num_tc); if (ret) goto out_free_eotids; tot_qcount = pi->nqsets; for (i = 0; i < mqprio->qopt.num_tc; i++) { qcount = mqprio->qopt.count[i]; if (qcount) { qoffset = mqprio->qopt.offset[i] + pi->nqsets; } else { qcount = pi->nqsets; qoffset = 0; } ret = netdev_set_tc_queue(dev, i, qcount, qoffset); if (ret) goto out_reset_tc; tot_qcount += mqprio->qopt.count[i]; } ret = netif_set_real_num_tx_queues(dev, tot_qcount); if (ret) goto out_reset_tc; tc_port_mqprio->state = CXGB4_MQPRIO_STATE_ACTIVE; return 0; out_reset_tc: netdev_reset_tc(dev); i = mqprio->qopt.num_tc; out_free_eotids: while (i-- > 0) { qoffset = mqprio->qopt.offset[i]; qcount = mqprio->qopt.count[i]; for (j = 0; j < qcount; j++) { eosw_txq = &tc_port_mqprio->eosw_txq[qoffset + j]; hwtc = tc_port_mqprio->tc_hwtc_map[i]; cxgb4_mqprio_class_unbind(dev, eosw_txq, hwtc); cxgb4_free_eotid(&adap->tids, eosw_txq->eotid); cxgb4_free_eosw_txq(dev, eosw_txq); } } cxgb4_mqprio_free_hw_resources(dev); return ret; } static void cxgb4_mqprio_disable_offload(struct net_device *dev) { struct cxgb4_tc_port_mqprio *tc_port_mqprio; struct port_info *pi = netdev2pinfo(dev); struct adapter *adap = netdev2adap(dev); struct sge_eosw_txq *eosw_txq; u32 qoffset, qcount; u16 i, j; u8 hwtc; tc_port_mqprio = &adap->tc_mqprio->port_mqprio[pi->port_id]; if (tc_port_mqprio->state != CXGB4_MQPRIO_STATE_ACTIVE) return; netdev_reset_tc(dev); netif_set_real_num_tx_queues(dev, pi->nqsets); for (i = 0; i < tc_port_mqprio->mqprio.qopt.num_tc; i++) { qoffset = tc_port_mqprio->mqprio.qopt.offset[i]; qcount = tc_port_mqprio->mqprio.qopt.count[i]; for (j = 0; j < qcount; j++) { eosw_txq = &tc_port_mqprio->eosw_txq[qoffset + j]; hwtc = tc_port_mqprio->tc_hwtc_map[i]; cxgb4_mqprio_class_unbind(dev, eosw_txq, hwtc); cxgb4_free_eotid(&adap->tids, eosw_txq->eotid); cxgb4_free_eosw_txq(dev, eosw_txq); } } cxgb4_mqprio_free_hw_resources(dev); /* Free up the traffic classes */ cxgb4_mqprio_free_tc(dev); memset(&tc_port_mqprio->mqprio, 0, sizeof(struct tc_mqprio_qopt_offload)); tc_port_mqprio->state = CXGB4_MQPRIO_STATE_DISABLED; } int cxgb4_setup_tc_mqprio(struct net_device *dev, struct tc_mqprio_qopt_offload *mqprio) { struct adapter *adap = netdev2adap(dev); bool needs_bring_up = false; int ret; ret = cxgb4_mqprio_validate(dev, mqprio); if (ret) return ret; mutex_lock(&adap->tc_mqprio->mqprio_mutex); /* To configure tc params, the current allocated EOTIDs must * be freed up. However, they can't be freed up if there's * traffic running on the interface. So, ensure interface is * down before configuring tc params. */ if (netif_running(dev)) { netif_tx_stop_all_queues(dev); netif_carrier_off(dev); needs_bring_up = true; } cxgb4_mqprio_disable_offload(dev); /* If requested for clear, then just return since resources are * already freed up by now. */ if (!mqprio->qopt.num_tc) goto out; /* Allocate free available traffic classes and configure * their rate parameters. */ ret = cxgb4_mqprio_alloc_tc(dev, mqprio); if (ret) goto out; ret = cxgb4_mqprio_enable_offload(dev, mqprio); if (ret) { cxgb4_mqprio_free_tc(dev); goto out; } out: if (needs_bring_up) { netif_tx_start_all_queues(dev); netif_carrier_on(dev); } mutex_unlock(&adap->tc_mqprio->mqprio_mutex); return ret; } void cxgb4_mqprio_stop_offload(struct adapter *adap) { struct cxgb4_tc_port_mqprio *tc_port_mqprio; struct net_device *dev; u8 i; if (!adap->tc_mqprio || !adap->tc_mqprio->port_mqprio) return; mutex_lock(&adap->tc_mqprio->mqprio_mutex); for_each_port(adap, i) { dev = adap->port[i]; if (!dev) continue; tc_port_mqprio = &adap->tc_mqprio->port_mqprio[i]; if (!tc_port_mqprio->mqprio.qopt.num_tc) continue; cxgb4_mqprio_disable_offload(dev); } mutex_unlock(&adap->tc_mqprio->mqprio_mutex); } int cxgb4_init_tc_mqprio(struct adapter *adap) { struct cxgb4_tc_port_mqprio *tc_port_mqprio, *port_mqprio; struct cxgb4_tc_mqprio *tc_mqprio; struct sge_eosw_txq *eosw_txq; int ret = 0; u8 i; tc_mqprio = kzalloc(sizeof(*tc_mqprio), GFP_KERNEL); if (!tc_mqprio) return -ENOMEM; tc_port_mqprio = kcalloc(adap->params.nports, sizeof(*tc_port_mqprio), GFP_KERNEL); if (!tc_port_mqprio) { ret = -ENOMEM; goto out_free_mqprio; } mutex_init(&tc_mqprio->mqprio_mutex); tc_mqprio->port_mqprio = tc_port_mqprio; for (i = 0; i < adap->params.nports; i++) { port_mqprio = &tc_mqprio->port_mqprio[i]; eosw_txq = kcalloc(adap->tids.neotids, sizeof(*eosw_txq), GFP_KERNEL); if (!eosw_txq) { ret = -ENOMEM; goto out_free_ports; } port_mqprio->eosw_txq = eosw_txq; } adap->tc_mqprio = tc_mqprio; refcount_set(&adap->tc_mqprio->refcnt, 0); return 0; out_free_ports: for (i = 0; i < adap->params.nports; i++) { port_mqprio = &tc_mqprio->port_mqprio[i]; kfree(port_mqprio->eosw_txq); } kfree(tc_port_mqprio); out_free_mqprio: kfree(tc_mqprio); return ret; } void cxgb4_cleanup_tc_mqprio(struct adapter *adap) { struct cxgb4_tc_port_mqprio *port_mqprio; u8 i; if (adap->tc_mqprio) { mutex_lock(&adap->tc_mqprio->mqprio_mutex); if (adap->tc_mqprio->port_mqprio) { for (i = 0; i < adap->params.nports; i++) { struct net_device *dev = adap->port[i]; if (dev) cxgb4_mqprio_disable_offload(dev); port_mqprio = &adap->tc_mqprio->port_mqprio[i]; kfree(port_mqprio->eosw_txq); } kfree(adap->tc_mqprio->port_mqprio); } mutex_unlock(&adap->tc_mqprio->mqprio_mutex); kfree(adap->tc_mqprio); } }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1