Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Jakub Kiciński | 4365 | 99.98% | 17 | 94.44% |
Ahmed S. Darwish | 1 | 0.02% | 1 | 5.56% |
Total | 4366 | 18 |
// SPDX-License-Identifier: (GPL-2.0-only OR BSD-2-Clause) /* Copyright (C) 2018 Netronome Systems, Inc. */ #include <linux/rtnetlink.h> #include <net/pkt_cls.h> #include <net/pkt_sched.h> #include <net/red.h> #include "../nfpcore/nfp_cpp.h" #include "../nfp_app.h" #include "../nfp_main.h" #include "../nfp_net.h" #include "../nfp_port.h" #include "main.h" static bool nfp_abm_qdisc_is_red(struct nfp_qdisc *qdisc) { return qdisc->type == NFP_QDISC_RED || qdisc->type == NFP_QDISC_GRED; } static bool nfp_abm_qdisc_child_valid(struct nfp_qdisc *qdisc, unsigned int id) { return qdisc->children[id] && qdisc->children[id] != NFP_QDISC_UNTRACKED; } static void *nfp_abm_qdisc_tree_deref_slot(void __rcu **slot) { return rtnl_dereference(*slot); } static void nfp_abm_stats_propagate(struct nfp_alink_stats *parent, struct nfp_alink_stats *child) { parent->tx_pkts += child->tx_pkts; parent->tx_bytes += child->tx_bytes; parent->backlog_pkts += child->backlog_pkts; parent->backlog_bytes += child->backlog_bytes; parent->overlimits += child->overlimits; parent->drops += child->drops; } static void nfp_abm_stats_update_red(struct nfp_abm_link *alink, struct nfp_qdisc *qdisc, unsigned int queue) { struct nfp_cpp *cpp = alink->abm->app->cpp; unsigned int i; int err; if (!qdisc->offloaded) return; for (i = 0; i < qdisc->red.num_bands; i++) { err = nfp_abm_ctrl_read_q_stats(alink, i, queue, &qdisc->red.band[i].stats); if (err) nfp_err(cpp, "RED stats (%d, %d) read failed with error %d\n", i, queue, err); err = nfp_abm_ctrl_read_q_xstats(alink, i, queue, &qdisc->red.band[i].xstats); if (err) nfp_err(cpp, "RED xstats (%d, %d) read failed with error %d\n", i, queue, err); } } static void nfp_abm_stats_update_mq(struct nfp_abm_link *alink, struct nfp_qdisc *qdisc) { unsigned int i; if (qdisc->type != NFP_QDISC_MQ) return; for (i = 0; i < alink->total_queues; i++) if (nfp_abm_qdisc_child_valid(qdisc, i)) nfp_abm_stats_update_red(alink, qdisc->children[i], i); } static void __nfp_abm_stats_update(struct nfp_abm_link *alink, u64 time_now) { alink->last_stats_update = time_now; if (alink->root_qdisc) nfp_abm_stats_update_mq(alink, alink->root_qdisc); } static void nfp_abm_stats_update(struct nfp_abm_link *alink) { u64 now; /* Limit the frequency of updates - stats of non-leaf qdiscs are a sum * of all their leafs, so we would read the same stat multiple times * for every dump. */ now = ktime_get(); if (now - alink->last_stats_update < NFP_ABM_STATS_REFRESH_IVAL) return; __nfp_abm_stats_update(alink, now); } static void nfp_abm_qdisc_unlink_children(struct nfp_qdisc *qdisc, unsigned int start, unsigned int end) { unsigned int i; for (i = start; i < end; i++) if (nfp_abm_qdisc_child_valid(qdisc, i)) { qdisc->children[i]->use_cnt--; qdisc->children[i] = NULL; } } static void nfp_abm_qdisc_offload_stop(struct nfp_abm_link *alink, struct nfp_qdisc *qdisc) { unsigned int i; /* Don't complain when qdisc is getting unlinked */ if (qdisc->use_cnt) nfp_warn(alink->abm->app->cpp, "Offload of '%08x' stopped\n", qdisc->handle); if (!nfp_abm_qdisc_is_red(qdisc)) return; for (i = 0; i < qdisc->red.num_bands; i++) { qdisc->red.band[i].stats.backlog_pkts = 0; qdisc->red.band[i].stats.backlog_bytes = 0; } } static int __nfp_abm_stats_init(struct nfp_abm_link *alink, unsigned int band, unsigned int queue, struct nfp_alink_stats *prev_stats, struct nfp_alink_xstats *prev_xstats) { u64 backlog_pkts, backlog_bytes; int err; /* Don't touch the backlog, backlog can only be reset after it has * been reported back to the tc qdisc stats. */ backlog_pkts = prev_stats->backlog_pkts; backlog_bytes = prev_stats->backlog_bytes; err = nfp_abm_ctrl_read_q_stats(alink, band, queue, prev_stats); if (err) { nfp_err(alink->abm->app->cpp, "RED stats init (%d, %d) failed with error %d\n", band, queue, err); return err; } err = nfp_abm_ctrl_read_q_xstats(alink, band, queue, prev_xstats); if (err) { nfp_err(alink->abm->app->cpp, "RED xstats init (%d, %d) failed with error %d\n", band, queue, err); return err; } prev_stats->backlog_pkts = backlog_pkts; prev_stats->backlog_bytes = backlog_bytes; return 0; } static int nfp_abm_stats_init(struct nfp_abm_link *alink, struct nfp_qdisc *qdisc, unsigned int queue) { unsigned int i; int err; for (i = 0; i < qdisc->red.num_bands; i++) { err = __nfp_abm_stats_init(alink, i, queue, &qdisc->red.band[i].prev_stats, &qdisc->red.band[i].prev_xstats); if (err) return err; } return 0; } static void nfp_abm_offload_compile_red(struct nfp_abm_link *alink, struct nfp_qdisc *qdisc, unsigned int queue) { bool good_red, good_gred; unsigned int i; good_red = qdisc->type == NFP_QDISC_RED && qdisc->params_ok && qdisc->use_cnt == 1 && !alink->has_prio && !qdisc->children[0]; good_gred = qdisc->type == NFP_QDISC_GRED && qdisc->params_ok && qdisc->use_cnt == 1; qdisc->offload_mark = good_red || good_gred; /* If we are starting offload init prev_stats */ if (qdisc->offload_mark && !qdisc->offloaded) if (nfp_abm_stats_init(alink, qdisc, queue)) qdisc->offload_mark = false; if (!qdisc->offload_mark) return; for (i = 0; i < alink->abm->num_bands; i++) { enum nfp_abm_q_action act; nfp_abm_ctrl_set_q_lvl(alink, i, queue, qdisc->red.band[i].threshold); act = qdisc->red.band[i].ecn ? NFP_ABM_ACT_MARK_DROP : NFP_ABM_ACT_DROP; nfp_abm_ctrl_set_q_act(alink, i, queue, act); } } static void nfp_abm_offload_compile_mq(struct nfp_abm_link *alink, struct nfp_qdisc *qdisc) { unsigned int i; qdisc->offload_mark = qdisc->type == NFP_QDISC_MQ; if (!qdisc->offload_mark) return; for (i = 0; i < alink->total_queues; i++) { struct nfp_qdisc *child = qdisc->children[i]; if (!nfp_abm_qdisc_child_valid(qdisc, i)) continue; nfp_abm_offload_compile_red(alink, child, i); } } void nfp_abm_qdisc_offload_update(struct nfp_abm_link *alink) { struct nfp_abm *abm = alink->abm; struct radix_tree_iter iter; struct nfp_qdisc *qdisc; void __rcu **slot; size_t i; /* Mark all thresholds as unconfigured */ for (i = 0; i < abm->num_bands; i++) __bitmap_set(abm->threshold_undef, i * NFP_NET_MAX_RX_RINGS + alink->queue_base, alink->total_queues); /* Clear offload marks */ radix_tree_for_each_slot(slot, &alink->qdiscs, &iter, 0) { qdisc = nfp_abm_qdisc_tree_deref_slot(slot); qdisc->offload_mark = false; } if (alink->root_qdisc) nfp_abm_offload_compile_mq(alink, alink->root_qdisc); /* Refresh offload status */ radix_tree_for_each_slot(slot, &alink->qdiscs, &iter, 0) { qdisc = nfp_abm_qdisc_tree_deref_slot(slot); if (!qdisc->offload_mark && qdisc->offloaded) nfp_abm_qdisc_offload_stop(alink, qdisc); qdisc->offloaded = qdisc->offload_mark; } /* Reset the unconfigured thresholds */ for (i = 0; i < abm->num_thresholds; i++) if (test_bit(i, abm->threshold_undef)) __nfp_abm_ctrl_set_q_lvl(abm, i, NFP_ABM_LVL_INFINITY); __nfp_abm_stats_update(alink, ktime_get()); } static void nfp_abm_qdisc_clear_mq(struct net_device *netdev, struct nfp_abm_link *alink, struct nfp_qdisc *qdisc) { struct radix_tree_iter iter; unsigned int mq_refs = 0; void __rcu **slot; if (!qdisc->use_cnt) return; /* MQ doesn't notify well on destruction, we need special handling of * MQ's children. */ if (qdisc->type == NFP_QDISC_MQ && qdisc == alink->root_qdisc && netdev->reg_state == NETREG_UNREGISTERING) return; /* Count refs held by MQ instances and clear pointers */ radix_tree_for_each_slot(slot, &alink->qdiscs, &iter, 0) { struct nfp_qdisc *mq = nfp_abm_qdisc_tree_deref_slot(slot); unsigned int i; if (mq->type != NFP_QDISC_MQ || mq->netdev != netdev) continue; for (i = 0; i < mq->num_children; i++) if (mq->children[i] == qdisc) { mq->children[i] = NULL; mq_refs++; } } WARN(qdisc->use_cnt != mq_refs, "non-zero qdisc use count: %d (- %d)\n", qdisc->use_cnt, mq_refs); } static void nfp_abm_qdisc_free(struct net_device *netdev, struct nfp_abm_link *alink, struct nfp_qdisc *qdisc) { struct nfp_port *port = nfp_port_from_netdev(netdev); if (!qdisc) return; nfp_abm_qdisc_clear_mq(netdev, alink, qdisc); WARN_ON(radix_tree_delete(&alink->qdiscs, TC_H_MAJ(qdisc->handle)) != qdisc); kfree(qdisc->children); kfree(qdisc); port->tc_offload_cnt--; } static struct nfp_qdisc * nfp_abm_qdisc_alloc(struct net_device *netdev, struct nfp_abm_link *alink, enum nfp_qdisc_type type, u32 parent_handle, u32 handle, unsigned int children) { struct nfp_port *port = nfp_port_from_netdev(netdev); struct nfp_qdisc *qdisc; int err; qdisc = kzalloc(sizeof(*qdisc), GFP_KERNEL); if (!qdisc) return NULL; if (children) { qdisc->children = kcalloc(children, sizeof(void *), GFP_KERNEL); if (!qdisc->children) goto err_free_qdisc; } qdisc->netdev = netdev; qdisc->type = type; qdisc->parent_handle = parent_handle; qdisc->handle = handle; qdisc->num_children = children; err = radix_tree_insert(&alink->qdiscs, TC_H_MAJ(qdisc->handle), qdisc); if (err) { nfp_err(alink->abm->app->cpp, "Qdisc insertion into radix tree failed: %d\n", err); goto err_free_child_tbl; } port->tc_offload_cnt++; return qdisc; err_free_child_tbl: kfree(qdisc->children); err_free_qdisc: kfree(qdisc); return NULL; } static struct nfp_qdisc * nfp_abm_qdisc_find(struct nfp_abm_link *alink, u32 handle) { return radix_tree_lookup(&alink->qdiscs, TC_H_MAJ(handle)); } static int nfp_abm_qdisc_replace(struct net_device *netdev, struct nfp_abm_link *alink, enum nfp_qdisc_type type, u32 parent_handle, u32 handle, unsigned int children, struct nfp_qdisc **qdisc) { *qdisc = nfp_abm_qdisc_find(alink, handle); if (*qdisc) { if (WARN_ON((*qdisc)->type != type)) return -EINVAL; return 1; } *qdisc = nfp_abm_qdisc_alloc(netdev, alink, type, parent_handle, handle, children); return *qdisc ? 0 : -ENOMEM; } static void nfp_abm_qdisc_destroy(struct net_device *netdev, struct nfp_abm_link *alink, u32 handle) { struct nfp_qdisc *qdisc; qdisc = nfp_abm_qdisc_find(alink, handle); if (!qdisc) return; /* We don't get TC_SETUP_ROOT_QDISC w/ MQ when netdev is unregistered */ if (alink->root_qdisc == qdisc) qdisc->use_cnt--; nfp_abm_qdisc_unlink_children(qdisc, 0, qdisc->num_children); nfp_abm_qdisc_free(netdev, alink, qdisc); if (alink->root_qdisc == qdisc) { alink->root_qdisc = NULL; /* Only root change matters, other changes are acted upon on * the graft notification. */ nfp_abm_qdisc_offload_update(alink); } } static int nfp_abm_qdisc_graft(struct nfp_abm_link *alink, u32 handle, u32 child_handle, unsigned int id) { struct nfp_qdisc *parent, *child; parent = nfp_abm_qdisc_find(alink, handle); if (!parent) return 0; if (WARN(id >= parent->num_children, "graft child out of bound %d >= %d\n", id, parent->num_children)) return -EINVAL; nfp_abm_qdisc_unlink_children(parent, id, id + 1); child = nfp_abm_qdisc_find(alink, child_handle); if (child) child->use_cnt++; else child = NFP_QDISC_UNTRACKED; parent->children[id] = child; nfp_abm_qdisc_offload_update(alink); return 0; } static void nfp_abm_stats_calculate(struct nfp_alink_stats *new, struct nfp_alink_stats *old, struct gnet_stats_basic_sync *bstats, struct gnet_stats_queue *qstats) { _bstats_update(bstats, new->tx_bytes - old->tx_bytes, new->tx_pkts - old->tx_pkts); qstats->qlen += new->backlog_pkts - old->backlog_pkts; qstats->backlog += new->backlog_bytes - old->backlog_bytes; qstats->overlimits += new->overlimits - old->overlimits; qstats->drops += new->drops - old->drops; } static void nfp_abm_stats_red_calculate(struct nfp_alink_xstats *new, struct nfp_alink_xstats *old, struct red_stats *stats) { stats->forced_mark += new->ecn_marked - old->ecn_marked; stats->pdrop += new->pdrop - old->pdrop; } static int nfp_abm_gred_stats(struct nfp_abm_link *alink, u32 handle, struct tc_gred_qopt_offload_stats *stats) { struct nfp_qdisc *qdisc; unsigned int i; nfp_abm_stats_update(alink); qdisc = nfp_abm_qdisc_find(alink, handle); if (!qdisc) return -EOPNOTSUPP; /* If the qdisc offload has stopped we may need to adjust the backlog * counters back so carry on even if qdisc is not currently offloaded. */ for (i = 0; i < qdisc->red.num_bands; i++) { if (!stats->xstats[i]) continue; nfp_abm_stats_calculate(&qdisc->red.band[i].stats, &qdisc->red.band[i].prev_stats, &stats->bstats[i], &stats->qstats[i]); qdisc->red.band[i].prev_stats = qdisc->red.band[i].stats; nfp_abm_stats_red_calculate(&qdisc->red.band[i].xstats, &qdisc->red.band[i].prev_xstats, stats->xstats[i]); qdisc->red.band[i].prev_xstats = qdisc->red.band[i].xstats; } return qdisc->offloaded ? 0 : -EOPNOTSUPP; } static bool nfp_abm_gred_check_params(struct nfp_abm_link *alink, struct tc_gred_qopt_offload *opt) { struct nfp_cpp *cpp = alink->abm->app->cpp; struct nfp_abm *abm = alink->abm; unsigned int i; if (opt->set.grio_on || opt->set.wred_on) { nfp_warn(cpp, "GRED offload failed - GRIO and WRED not supported (p:%08x h:%08x)\n", opt->parent, opt->handle); return false; } if (opt->set.dp_def != alink->def_band) { nfp_warn(cpp, "GRED offload failed - default band must be %d (p:%08x h:%08x)\n", alink->def_band, opt->parent, opt->handle); return false; } if (opt->set.dp_cnt != abm->num_bands) { nfp_warn(cpp, "GRED offload failed - band count must be %d (p:%08x h:%08x)\n", abm->num_bands, opt->parent, opt->handle); return false; } for (i = 0; i < abm->num_bands; i++) { struct tc_gred_vq_qopt_offload_params *band = &opt->set.tab[i]; if (!band->present) return false; if (!band->is_ecn && !nfp_abm_has_drop(abm)) { nfp_warn(cpp, "GRED offload failed - drop is not supported (ECN option required) (p:%08x h:%08x vq:%d)\n", opt->parent, opt->handle, i); return false; } if (band->is_ecn && !nfp_abm_has_mark(abm)) { nfp_warn(cpp, "GRED offload failed - ECN marking not supported (p:%08x h:%08x vq:%d)\n", opt->parent, opt->handle, i); return false; } if (band->is_harddrop) { nfp_warn(cpp, "GRED offload failed - harddrop is not supported (p:%08x h:%08x vq:%d)\n", opt->parent, opt->handle, i); return false; } if (band->min != band->max) { nfp_warn(cpp, "GRED offload failed - threshold mismatch (p:%08x h:%08x vq:%d)\n", opt->parent, opt->handle, i); return false; } if (band->min > S32_MAX) { nfp_warn(cpp, "GRED offload failed - threshold too large %d > %d (p:%08x h:%08x vq:%d)\n", band->min, S32_MAX, opt->parent, opt->handle, i); return false; } } return true; } static int nfp_abm_gred_replace(struct net_device *netdev, struct nfp_abm_link *alink, struct tc_gred_qopt_offload *opt) { struct nfp_qdisc *qdisc; unsigned int i; int ret; ret = nfp_abm_qdisc_replace(netdev, alink, NFP_QDISC_GRED, opt->parent, opt->handle, 0, &qdisc); if (ret < 0) return ret; qdisc->params_ok = nfp_abm_gred_check_params(alink, opt); if (qdisc->params_ok) { qdisc->red.num_bands = opt->set.dp_cnt; for (i = 0; i < qdisc->red.num_bands; i++) { qdisc->red.band[i].ecn = opt->set.tab[i].is_ecn; qdisc->red.band[i].threshold = opt->set.tab[i].min; } } if (qdisc->use_cnt) nfp_abm_qdisc_offload_update(alink); return 0; } int nfp_abm_setup_tc_gred(struct net_device *netdev, struct nfp_abm_link *alink, struct tc_gred_qopt_offload *opt) { switch (opt->command) { case TC_GRED_REPLACE: return nfp_abm_gred_replace(netdev, alink, opt); case TC_GRED_DESTROY: nfp_abm_qdisc_destroy(netdev, alink, opt->handle); return 0; case TC_GRED_STATS: return nfp_abm_gred_stats(alink, opt->handle, &opt->stats); default: return -EOPNOTSUPP; } } static int nfp_abm_red_xstats(struct nfp_abm_link *alink, struct tc_red_qopt_offload *opt) { struct nfp_qdisc *qdisc; nfp_abm_stats_update(alink); qdisc = nfp_abm_qdisc_find(alink, opt->handle); if (!qdisc || !qdisc->offloaded) return -EOPNOTSUPP; nfp_abm_stats_red_calculate(&qdisc->red.band[0].xstats, &qdisc->red.band[0].prev_xstats, opt->xstats); qdisc->red.band[0].prev_xstats = qdisc->red.band[0].xstats; return 0; } static int nfp_abm_red_stats(struct nfp_abm_link *alink, u32 handle, struct tc_qopt_offload_stats *stats) { struct nfp_qdisc *qdisc; nfp_abm_stats_update(alink); qdisc = nfp_abm_qdisc_find(alink, handle); if (!qdisc) return -EOPNOTSUPP; /* If the qdisc offload has stopped we may need to adjust the backlog * counters back so carry on even if qdisc is not currently offloaded. */ nfp_abm_stats_calculate(&qdisc->red.band[0].stats, &qdisc->red.band[0].prev_stats, stats->bstats, stats->qstats); qdisc->red.band[0].prev_stats = qdisc->red.band[0].stats; return qdisc->offloaded ? 0 : -EOPNOTSUPP; } static bool nfp_abm_red_check_params(struct nfp_abm_link *alink, struct tc_red_qopt_offload *opt) { struct nfp_cpp *cpp = alink->abm->app->cpp; struct nfp_abm *abm = alink->abm; if (!opt->set.is_ecn && !nfp_abm_has_drop(abm)) { nfp_warn(cpp, "RED offload failed - drop is not supported (ECN option required) (p:%08x h:%08x)\n", opt->parent, opt->handle); return false; } if (opt->set.is_ecn && !nfp_abm_has_mark(abm)) { nfp_warn(cpp, "RED offload failed - ECN marking not supported (p:%08x h:%08x)\n", opt->parent, opt->handle); return false; } if (opt->set.is_harddrop) { nfp_warn(cpp, "RED offload failed - harddrop is not supported (p:%08x h:%08x)\n", opt->parent, opt->handle); return false; } if (opt->set.min != opt->set.max) { nfp_warn(cpp, "RED offload failed - unsupported min/max parameters (p:%08x h:%08x)\n", opt->parent, opt->handle); return false; } if (opt->set.min > NFP_ABM_LVL_INFINITY) { nfp_warn(cpp, "RED offload failed - threshold too large %d > %d (p:%08x h:%08x)\n", opt->set.min, NFP_ABM_LVL_INFINITY, opt->parent, opt->handle); return false; } return true; } static int nfp_abm_red_replace(struct net_device *netdev, struct nfp_abm_link *alink, struct tc_red_qopt_offload *opt) { struct nfp_qdisc *qdisc; int ret; ret = nfp_abm_qdisc_replace(netdev, alink, NFP_QDISC_RED, opt->parent, opt->handle, 1, &qdisc); if (ret < 0) return ret; /* If limit != 0 child gets reset */ if (opt->set.limit) { if (nfp_abm_qdisc_child_valid(qdisc, 0)) qdisc->children[0]->use_cnt--; qdisc->children[0] = NULL; } else { /* Qdisc was just allocated without a limit will use noop_qdisc, * i.e. a block hole. */ if (!ret) qdisc->children[0] = NFP_QDISC_UNTRACKED; } qdisc->params_ok = nfp_abm_red_check_params(alink, opt); if (qdisc->params_ok) { qdisc->red.num_bands = 1; qdisc->red.band[0].ecn = opt->set.is_ecn; qdisc->red.band[0].threshold = opt->set.min; } if (qdisc->use_cnt == 1) nfp_abm_qdisc_offload_update(alink); return 0; } int nfp_abm_setup_tc_red(struct net_device *netdev, struct nfp_abm_link *alink, struct tc_red_qopt_offload *opt) { switch (opt->command) { case TC_RED_REPLACE: return nfp_abm_red_replace(netdev, alink, opt); case TC_RED_DESTROY: nfp_abm_qdisc_destroy(netdev, alink, opt->handle); return 0; case TC_RED_STATS: return nfp_abm_red_stats(alink, opt->handle, &opt->stats); case TC_RED_XSTATS: return nfp_abm_red_xstats(alink, opt); case TC_RED_GRAFT: return nfp_abm_qdisc_graft(alink, opt->handle, opt->child_handle, 0); default: return -EOPNOTSUPP; } } static int nfp_abm_mq_create(struct net_device *netdev, struct nfp_abm_link *alink, struct tc_mq_qopt_offload *opt) { struct nfp_qdisc *qdisc; int ret; ret = nfp_abm_qdisc_replace(netdev, alink, NFP_QDISC_MQ, TC_H_ROOT, opt->handle, alink->total_queues, &qdisc); if (ret < 0) return ret; qdisc->params_ok = true; qdisc->offloaded = true; nfp_abm_qdisc_offload_update(alink); return 0; } static int nfp_abm_mq_stats(struct nfp_abm_link *alink, u32 handle, struct tc_qopt_offload_stats *stats) { struct nfp_qdisc *qdisc, *red; unsigned int i, j; qdisc = nfp_abm_qdisc_find(alink, handle); if (!qdisc) return -EOPNOTSUPP; nfp_abm_stats_update(alink); /* MQ stats are summed over the children in the core, so we need * to add up the unreported child values. */ memset(&qdisc->mq.stats, 0, sizeof(qdisc->mq.stats)); memset(&qdisc->mq.prev_stats, 0, sizeof(qdisc->mq.prev_stats)); for (i = 0; i < qdisc->num_children; i++) { if (!nfp_abm_qdisc_child_valid(qdisc, i)) continue; if (!nfp_abm_qdisc_is_red(qdisc->children[i])) continue; red = qdisc->children[i]; for (j = 0; j < red->red.num_bands; j++) { nfp_abm_stats_propagate(&qdisc->mq.stats, &red->red.band[j].stats); nfp_abm_stats_propagate(&qdisc->mq.prev_stats, &red->red.band[j].prev_stats); } } nfp_abm_stats_calculate(&qdisc->mq.stats, &qdisc->mq.prev_stats, stats->bstats, stats->qstats); return qdisc->offloaded ? 0 : -EOPNOTSUPP; } int nfp_abm_setup_tc_mq(struct net_device *netdev, struct nfp_abm_link *alink, struct tc_mq_qopt_offload *opt) { switch (opt->command) { case TC_MQ_CREATE: return nfp_abm_mq_create(netdev, alink, opt); case TC_MQ_DESTROY: nfp_abm_qdisc_destroy(netdev, alink, opt->handle); return 0; case TC_MQ_STATS: return nfp_abm_mq_stats(alink, opt->handle, &opt->stats); case TC_MQ_GRAFT: return nfp_abm_qdisc_graft(alink, opt->handle, opt->graft_params.child_handle, opt->graft_params.queue); default: return -EOPNOTSUPP; } } int nfp_abm_setup_root(struct net_device *netdev, struct nfp_abm_link *alink, struct tc_root_qopt_offload *opt) { if (opt->ingress) return -EOPNOTSUPP; if (alink->root_qdisc) alink->root_qdisc->use_cnt--; alink->root_qdisc = nfp_abm_qdisc_find(alink, opt->handle); if (alink->root_qdisc) alink->root_qdisc->use_cnt++; nfp_abm_qdisc_offload_update(alink); return 0; }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1