Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Petr Machata | 5927 | 69.14% | 24 | 43.64% |
Nogah Frankel | 2562 | 29.88% | 23 | 41.82% |
Jakub Kiciński | 55 | 0.64% | 1 | 1.82% |
Ido Schimmel | 14 | 0.16% | 3 | 5.45% |
Jiri Pirko | 10 | 0.12% | 2 | 3.64% |
Nathan Chancellor | 4 | 0.05% | 1 | 1.82% |
Wei Yongjun | 1 | 0.01% | 1 | 1.82% |
Total | 8573 | 55 |
// SPDX-License-Identifier: BSD-3-Clause OR GPL-2.0 /* Copyright (c) 2017-2018 Mellanox Technologies. All rights reserved */ #include <linux/kernel.h> #include <linux/errno.h> #include <linux/netdevice.h> #include <net/pkt_cls.h> #include <net/red.h> #include "spectrum.h" #include "spectrum_span.h" #include "reg.h" #define MLXSW_SP_PRIO_BAND_TO_TCLASS(band) (IEEE_8021QAZ_MAX_TCS - band - 1) #define MLXSW_SP_PRIO_CHILD_TO_TCLASS(child) \ MLXSW_SP_PRIO_BAND_TO_TCLASS((child - 1)) enum mlxsw_sp_qdisc_type { MLXSW_SP_QDISC_NO_QDISC, MLXSW_SP_QDISC_RED, MLXSW_SP_QDISC_PRIO, MLXSW_SP_QDISC_ETS, MLXSW_SP_QDISC_TBF, MLXSW_SP_QDISC_FIFO, }; struct mlxsw_sp_qdisc; struct mlxsw_sp_qdisc_ops { enum mlxsw_sp_qdisc_type type; int (*check_params)(struct mlxsw_sp_port *mlxsw_sp_port, void *params); int (*replace)(struct mlxsw_sp_port *mlxsw_sp_port, u32 handle, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, void *params); int (*destroy)(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc); int (*get_stats)(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, struct tc_qopt_offload_stats *stats_ptr); int (*get_xstats)(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, void *xstats_ptr); void (*clean_stats)(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc); /* unoffload - to be used for a qdisc that stops being offloaded without * being destroyed. */ void (*unoffload)(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, void *params); struct mlxsw_sp_qdisc *(*find_class)(struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, u32 parent); unsigned int num_classes; }; struct mlxsw_sp_qdisc { u32 handle; int tclass_num; u8 prio_bitmap; union { struct red_stats red; } xstats_base; struct mlxsw_sp_qdisc_stats { u64 tx_bytes; u64 tx_packets; u64 drops; u64 overlimits; u64 backlog; } stats_base; struct mlxsw_sp_qdisc_ops *ops; struct mlxsw_sp_qdisc *parent; struct mlxsw_sp_qdisc *qdiscs; unsigned int num_classes; }; struct mlxsw_sp_qdisc_state { struct mlxsw_sp_qdisc root_qdisc; /* When a PRIO or ETS are added, the invisible FIFOs in their bands are * created first. When notifications for these FIFOs arrive, it is not * known what qdisc their parent handle refers to. It could be a * newly-created PRIO that will replace the currently-offloaded one, or * it could be e.g. a RED that will be attached below it. * * As the notifications start to arrive, use them to note what the * future parent handle is, and keep track of which child FIFOs were * seen. Then when the parent is known, retroactively offload those * FIFOs. */ u32 future_handle; bool future_fifos[IEEE_8021QAZ_MAX_TCS]; struct mutex lock; /* Protects qdisc state. */ }; static bool mlxsw_sp_qdisc_compare(struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, u32 handle) { return mlxsw_sp_qdisc->ops && mlxsw_sp_qdisc->handle == handle; } static struct mlxsw_sp_qdisc * mlxsw_sp_qdisc_walk(struct mlxsw_sp_qdisc *qdisc, struct mlxsw_sp_qdisc *(*pre)(struct mlxsw_sp_qdisc *, void *), void *data) { struct mlxsw_sp_qdisc *tmp; unsigned int i; if (pre) { tmp = pre(qdisc, data); if (tmp) return tmp; } if (qdisc->ops) { for (i = 0; i < qdisc->num_classes; i++) { tmp = &qdisc->qdiscs[i]; if (qdisc->ops) { tmp = mlxsw_sp_qdisc_walk(tmp, pre, data); if (tmp) return tmp; } } } return NULL; } static struct mlxsw_sp_qdisc * mlxsw_sp_qdisc_walk_cb_find(struct mlxsw_sp_qdisc *qdisc, void *data) { u32 parent = *(u32 *)data; if (qdisc->ops && TC_H_MAJ(qdisc->handle) == TC_H_MAJ(parent)) { if (qdisc->ops->find_class) return qdisc->ops->find_class(qdisc, parent); } return NULL; } static struct mlxsw_sp_qdisc * mlxsw_sp_qdisc_find(struct mlxsw_sp_port *mlxsw_sp_port, u32 parent, bool root_only) { struct mlxsw_sp_qdisc_state *qdisc_state = mlxsw_sp_port->qdisc; if (!qdisc_state) return NULL; if (parent == TC_H_ROOT) return &qdisc_state->root_qdisc; if (root_only) return NULL; return mlxsw_sp_qdisc_walk(&qdisc_state->root_qdisc, mlxsw_sp_qdisc_walk_cb_find, &parent); } static struct mlxsw_sp_qdisc * mlxsw_sp_qdisc_walk_cb_find_by_handle(struct mlxsw_sp_qdisc *qdisc, void *data) { u32 handle = *(u32 *)data; if (qdisc->ops && qdisc->handle == handle) return qdisc; return NULL; } static struct mlxsw_sp_qdisc * mlxsw_sp_qdisc_find_by_handle(struct mlxsw_sp_port *mlxsw_sp_port, u32 handle) { struct mlxsw_sp_qdisc_state *qdisc_state = mlxsw_sp_port->qdisc; if (!qdisc_state) return NULL; return mlxsw_sp_qdisc_walk(&qdisc_state->root_qdisc, mlxsw_sp_qdisc_walk_cb_find_by_handle, &handle); } static void mlxsw_sp_qdisc_reduce_parent_backlog(struct mlxsw_sp_qdisc *mlxsw_sp_qdisc) { struct mlxsw_sp_qdisc *tmp; for (tmp = mlxsw_sp_qdisc->parent; tmp; tmp = tmp->parent) tmp->stats_base.backlog -= mlxsw_sp_qdisc->stats_base.backlog; } static int mlxsw_sp_qdisc_destroy(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc) { struct mlxsw_sp_qdisc *root_qdisc = &mlxsw_sp_port->qdisc->root_qdisc; int err_hdroom = 0; int err = 0; if (!mlxsw_sp_qdisc) return 0; if (root_qdisc == mlxsw_sp_qdisc) { struct mlxsw_sp_hdroom hdroom = *mlxsw_sp_port->hdroom; hdroom.mode = MLXSW_SP_HDROOM_MODE_DCB; mlxsw_sp_hdroom_prios_reset_buf_idx(&hdroom); mlxsw_sp_hdroom_bufs_reset_lossiness(&hdroom); mlxsw_sp_hdroom_bufs_reset_sizes(mlxsw_sp_port, &hdroom); err_hdroom = mlxsw_sp_hdroom_configure(mlxsw_sp_port, &hdroom); } if (!mlxsw_sp_qdisc->ops) return 0; mlxsw_sp_qdisc_reduce_parent_backlog(mlxsw_sp_qdisc); if (mlxsw_sp_qdisc->ops->destroy) err = mlxsw_sp_qdisc->ops->destroy(mlxsw_sp_port, mlxsw_sp_qdisc); if (mlxsw_sp_qdisc->ops->clean_stats) mlxsw_sp_qdisc->ops->clean_stats(mlxsw_sp_port, mlxsw_sp_qdisc); mlxsw_sp_qdisc->handle = TC_H_UNSPEC; mlxsw_sp_qdisc->ops = NULL; mlxsw_sp_qdisc->num_classes = 0; kfree(mlxsw_sp_qdisc->qdiscs); mlxsw_sp_qdisc->qdiscs = NULL; return err_hdroom ?: err; } static int mlxsw_sp_qdisc_create(struct mlxsw_sp_port *mlxsw_sp_port, u32 handle, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, struct mlxsw_sp_qdisc_ops *ops, void *params) { struct mlxsw_sp_qdisc *root_qdisc = &mlxsw_sp_port->qdisc->root_qdisc; struct mlxsw_sp_hdroom orig_hdroom; unsigned int i; int err; err = ops->check_params(mlxsw_sp_port, params); if (err) return err; if (ops->num_classes) { mlxsw_sp_qdisc->qdiscs = kcalloc(ops->num_classes, sizeof(*mlxsw_sp_qdisc->qdiscs), GFP_KERNEL); if (!mlxsw_sp_qdisc->qdiscs) return -ENOMEM; for (i = 0; i < ops->num_classes; i++) mlxsw_sp_qdisc->qdiscs[i].parent = mlxsw_sp_qdisc; } orig_hdroom = *mlxsw_sp_port->hdroom; if (root_qdisc == mlxsw_sp_qdisc) { struct mlxsw_sp_hdroom hdroom = orig_hdroom; hdroom.mode = MLXSW_SP_HDROOM_MODE_TC; mlxsw_sp_hdroom_prios_reset_buf_idx(&hdroom); mlxsw_sp_hdroom_bufs_reset_lossiness(&hdroom); mlxsw_sp_hdroom_bufs_reset_sizes(mlxsw_sp_port, &hdroom); err = mlxsw_sp_hdroom_configure(mlxsw_sp_port, &hdroom); if (err) goto err_hdroom_configure; } mlxsw_sp_qdisc->num_classes = ops->num_classes; mlxsw_sp_qdisc->ops = ops; mlxsw_sp_qdisc->handle = handle; err = ops->replace(mlxsw_sp_port, handle, mlxsw_sp_qdisc, params); if (err) goto err_replace; return 0; err_replace: mlxsw_sp_qdisc->handle = TC_H_UNSPEC; mlxsw_sp_qdisc->ops = NULL; mlxsw_sp_qdisc->num_classes = 0; mlxsw_sp_hdroom_configure(mlxsw_sp_port, &orig_hdroom); err_hdroom_configure: kfree(mlxsw_sp_qdisc->qdiscs); mlxsw_sp_qdisc->qdiscs = NULL; return err; } static int mlxsw_sp_qdisc_change(struct mlxsw_sp_port *mlxsw_sp_port, u32 handle, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, void *params) { struct mlxsw_sp_qdisc_ops *ops = mlxsw_sp_qdisc->ops; int err; err = ops->check_params(mlxsw_sp_port, params); if (err) goto unoffload; err = ops->replace(mlxsw_sp_port, handle, mlxsw_sp_qdisc, params); if (err) goto unoffload; /* Check if the Qdisc changed. That includes a situation where an * invisible Qdisc replaces another one, or is being added for the * first time. */ if (mlxsw_sp_qdisc->handle != handle) { if (ops->clean_stats) ops->clean_stats(mlxsw_sp_port, mlxsw_sp_qdisc); } mlxsw_sp_qdisc->handle = handle; return 0; unoffload: if (ops->unoffload) ops->unoffload(mlxsw_sp_port, mlxsw_sp_qdisc, params); mlxsw_sp_qdisc_destroy(mlxsw_sp_port, mlxsw_sp_qdisc); return err; } static int mlxsw_sp_qdisc_replace(struct mlxsw_sp_port *mlxsw_sp_port, u32 handle, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, struct mlxsw_sp_qdisc_ops *ops, void *params) { if (mlxsw_sp_qdisc->ops && mlxsw_sp_qdisc->ops->type != ops->type) /* In case this location contained a different qdisc of the * same type we can override the old qdisc configuration. * Otherwise, we need to remove the old qdisc before setting the * new one. */ mlxsw_sp_qdisc_destroy(mlxsw_sp_port, mlxsw_sp_qdisc); if (!mlxsw_sp_qdisc->ops) return mlxsw_sp_qdisc_create(mlxsw_sp_port, handle, mlxsw_sp_qdisc, ops, params); else return mlxsw_sp_qdisc_change(mlxsw_sp_port, handle, mlxsw_sp_qdisc, params); } static int mlxsw_sp_qdisc_get_stats(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, struct tc_qopt_offload_stats *stats_ptr) { if (mlxsw_sp_qdisc && mlxsw_sp_qdisc->ops && mlxsw_sp_qdisc->ops->get_stats) return mlxsw_sp_qdisc->ops->get_stats(mlxsw_sp_port, mlxsw_sp_qdisc, stats_ptr); return -EOPNOTSUPP; } static int mlxsw_sp_qdisc_get_xstats(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, void *xstats_ptr) { if (mlxsw_sp_qdisc && mlxsw_sp_qdisc->ops && mlxsw_sp_qdisc->ops->get_xstats) return mlxsw_sp_qdisc->ops->get_xstats(mlxsw_sp_port, mlxsw_sp_qdisc, xstats_ptr); return -EOPNOTSUPP; } static u64 mlxsw_sp_xstats_backlog(struct mlxsw_sp_port_xstats *xstats, int tclass_num) { return xstats->backlog[tclass_num] + xstats->backlog[tclass_num + 8]; } static u64 mlxsw_sp_xstats_tail_drop(struct mlxsw_sp_port_xstats *xstats, int tclass_num) { return xstats->tail_drop[tclass_num] + xstats->tail_drop[tclass_num + 8]; } static void mlxsw_sp_qdisc_bstats_per_priority_get(struct mlxsw_sp_port_xstats *xstats, u8 prio_bitmap, u64 *tx_packets, u64 *tx_bytes) { int i; *tx_packets = 0; *tx_bytes = 0; for (i = 0; i < IEEE_8021QAZ_MAX_TCS; i++) { if (prio_bitmap & BIT(i)) { *tx_packets += xstats->tx_packets[i]; *tx_bytes += xstats->tx_bytes[i]; } } } static void mlxsw_sp_qdisc_collect_tc_stats(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, u64 *p_tx_bytes, u64 *p_tx_packets, u64 *p_drops, u64 *p_backlog) { int tclass_num = mlxsw_sp_qdisc->tclass_num; struct mlxsw_sp_port_xstats *xstats; u64 tx_bytes, tx_packets; xstats = &mlxsw_sp_port->periodic_hw_stats.xstats; mlxsw_sp_qdisc_bstats_per_priority_get(xstats, mlxsw_sp_qdisc->prio_bitmap, &tx_packets, &tx_bytes); *p_tx_packets += tx_packets; *p_tx_bytes += tx_bytes; *p_drops += xstats->wred_drop[tclass_num] + mlxsw_sp_xstats_tail_drop(xstats, tclass_num); *p_backlog += mlxsw_sp_xstats_backlog(xstats, tclass_num); } static void mlxsw_sp_qdisc_update_stats(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, u64 tx_bytes, u64 tx_packets, u64 drops, u64 backlog, struct tc_qopt_offload_stats *stats_ptr) { struct mlxsw_sp_qdisc_stats *stats_base = &mlxsw_sp_qdisc->stats_base; tx_bytes -= stats_base->tx_bytes; tx_packets -= stats_base->tx_packets; drops -= stats_base->drops; backlog -= stats_base->backlog; _bstats_update(stats_ptr->bstats, tx_bytes, tx_packets); stats_ptr->qstats->drops += drops; stats_ptr->qstats->backlog += mlxsw_sp_cells_bytes(mlxsw_sp, backlog); stats_base->backlog += backlog; stats_base->drops += drops; stats_base->tx_bytes += tx_bytes; stats_base->tx_packets += tx_packets; } static void mlxsw_sp_qdisc_get_tc_stats(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, struct tc_qopt_offload_stats *stats_ptr) { u64 tx_packets = 0; u64 tx_bytes = 0; u64 backlog = 0; u64 drops = 0; mlxsw_sp_qdisc_collect_tc_stats(mlxsw_sp_port, mlxsw_sp_qdisc, &tx_bytes, &tx_packets, &drops, &backlog); mlxsw_sp_qdisc_update_stats(mlxsw_sp_port->mlxsw_sp, mlxsw_sp_qdisc, tx_bytes, tx_packets, drops, backlog, stats_ptr); } static int mlxsw_sp_tclass_congestion_enable(struct mlxsw_sp_port *mlxsw_sp_port, int tclass_num, u32 min, u32 max, u32 probability, bool is_wred, bool is_ecn) { char cwtpm_cmd[MLXSW_REG_CWTPM_LEN]; char cwtp_cmd[MLXSW_REG_CWTP_LEN]; struct mlxsw_sp *mlxsw_sp = mlxsw_sp_port->mlxsw_sp; int err; mlxsw_reg_cwtp_pack(cwtp_cmd, mlxsw_sp_port->local_port, tclass_num); mlxsw_reg_cwtp_profile_pack(cwtp_cmd, MLXSW_REG_CWTP_DEFAULT_PROFILE, roundup(min, MLXSW_REG_CWTP_MIN_VALUE), roundup(max, MLXSW_REG_CWTP_MIN_VALUE), probability); err = mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(cwtp), cwtp_cmd); if (err) return err; mlxsw_reg_cwtpm_pack(cwtpm_cmd, mlxsw_sp_port->local_port, tclass_num, MLXSW_REG_CWTP_DEFAULT_PROFILE, is_wred, is_ecn); return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(cwtpm), cwtpm_cmd); } static int mlxsw_sp_tclass_congestion_disable(struct mlxsw_sp_port *mlxsw_sp_port, int tclass_num) { struct mlxsw_sp *mlxsw_sp = mlxsw_sp_port->mlxsw_sp; char cwtpm_cmd[MLXSW_REG_CWTPM_LEN]; mlxsw_reg_cwtpm_pack(cwtpm_cmd, mlxsw_sp_port->local_port, tclass_num, MLXSW_REG_CWTPM_RESET_PROFILE, false, false); return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(cwtpm), cwtpm_cmd); } static void mlxsw_sp_setup_tc_qdisc_red_clean_stats(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc) { int tclass_num = mlxsw_sp_qdisc->tclass_num; struct mlxsw_sp_qdisc_stats *stats_base; struct mlxsw_sp_port_xstats *xstats; struct red_stats *red_base; xstats = &mlxsw_sp_port->periodic_hw_stats.xstats; stats_base = &mlxsw_sp_qdisc->stats_base; red_base = &mlxsw_sp_qdisc->xstats_base.red; mlxsw_sp_qdisc_bstats_per_priority_get(xstats, mlxsw_sp_qdisc->prio_bitmap, &stats_base->tx_packets, &stats_base->tx_bytes); red_base->prob_drop = xstats->wred_drop[tclass_num]; red_base->pdrop = mlxsw_sp_xstats_tail_drop(xstats, tclass_num); stats_base->overlimits = red_base->prob_drop + red_base->prob_mark; stats_base->drops = red_base->prob_drop + red_base->pdrop; stats_base->backlog = 0; } static int mlxsw_sp_qdisc_red_destroy(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc) { return mlxsw_sp_tclass_congestion_disable(mlxsw_sp_port, mlxsw_sp_qdisc->tclass_num); } static int mlxsw_sp_qdisc_red_check_params(struct mlxsw_sp_port *mlxsw_sp_port, void *params) { struct mlxsw_sp *mlxsw_sp = mlxsw_sp_port->mlxsw_sp; struct tc_red_qopt_offload_params *p = params; if (p->min > p->max) { dev_err(mlxsw_sp->bus_info->dev, "spectrum: RED: min %u is bigger then max %u\n", p->min, p->max); return -EINVAL; } if (p->max > MLXSW_CORE_RES_GET(mlxsw_sp->core, GUARANTEED_SHARED_BUFFER)) { dev_err(mlxsw_sp->bus_info->dev, "spectrum: RED: max value %u is too big\n", p->max); return -EINVAL; } if (p->min == 0 || p->max == 0) { dev_err(mlxsw_sp->bus_info->dev, "spectrum: RED: 0 value is illegal for min and max\n"); return -EINVAL; } return 0; } static int mlxsw_sp_qdisc_red_replace(struct mlxsw_sp_port *mlxsw_sp_port, u32 handle, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, void *params) { struct mlxsw_sp *mlxsw_sp = mlxsw_sp_port->mlxsw_sp; struct tc_red_qopt_offload_params *p = params; int tclass_num = mlxsw_sp_qdisc->tclass_num; u32 min, max; u64 prob; /* calculate probability in percentage */ prob = p->probability; prob *= 100; prob = DIV_ROUND_UP(prob, 1 << 16); prob = DIV_ROUND_UP(prob, 1 << 16); min = mlxsw_sp_bytes_cells(mlxsw_sp, p->min); max = mlxsw_sp_bytes_cells(mlxsw_sp, p->max); return mlxsw_sp_tclass_congestion_enable(mlxsw_sp_port, tclass_num, min, max, prob, !p->is_nodrop, p->is_ecn); } static void mlxsw_sp_qdisc_leaf_unoffload(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, struct gnet_stats_queue *qstats) { u64 backlog; backlog = mlxsw_sp_cells_bytes(mlxsw_sp_port->mlxsw_sp, mlxsw_sp_qdisc->stats_base.backlog); qstats->backlog -= backlog; mlxsw_sp_qdisc->stats_base.backlog = 0; } static void mlxsw_sp_qdisc_red_unoffload(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, void *params) { struct tc_red_qopt_offload_params *p = params; mlxsw_sp_qdisc_leaf_unoffload(mlxsw_sp_port, mlxsw_sp_qdisc, p->qstats); } static int mlxsw_sp_qdisc_get_red_xstats(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, void *xstats_ptr) { struct red_stats *xstats_base = &mlxsw_sp_qdisc->xstats_base.red; int tclass_num = mlxsw_sp_qdisc->tclass_num; struct mlxsw_sp_port_xstats *xstats; struct red_stats *res = xstats_ptr; int early_drops, pdrops; xstats = &mlxsw_sp_port->periodic_hw_stats.xstats; early_drops = xstats->wred_drop[tclass_num] - xstats_base->prob_drop; pdrops = mlxsw_sp_xstats_tail_drop(xstats, tclass_num) - xstats_base->pdrop; res->pdrop += pdrops; res->prob_drop += early_drops; xstats_base->pdrop += pdrops; xstats_base->prob_drop += early_drops; return 0; } static int mlxsw_sp_qdisc_get_red_stats(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, struct tc_qopt_offload_stats *stats_ptr) { int tclass_num = mlxsw_sp_qdisc->tclass_num; struct mlxsw_sp_qdisc_stats *stats_base; struct mlxsw_sp_port_xstats *xstats; u64 overlimits; xstats = &mlxsw_sp_port->periodic_hw_stats.xstats; stats_base = &mlxsw_sp_qdisc->stats_base; mlxsw_sp_qdisc_get_tc_stats(mlxsw_sp_port, mlxsw_sp_qdisc, stats_ptr); overlimits = xstats->wred_drop[tclass_num] - stats_base->overlimits; stats_ptr->qstats->overlimits += overlimits; stats_base->overlimits += overlimits; return 0; } static struct mlxsw_sp_qdisc * mlxsw_sp_qdisc_leaf_find_class(struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, u32 parent) { return NULL; } #define MLXSW_SP_PORT_DEFAULT_TCLASS 0 static struct mlxsw_sp_qdisc_ops mlxsw_sp_qdisc_ops_red = { .type = MLXSW_SP_QDISC_RED, .check_params = mlxsw_sp_qdisc_red_check_params, .replace = mlxsw_sp_qdisc_red_replace, .unoffload = mlxsw_sp_qdisc_red_unoffload, .destroy = mlxsw_sp_qdisc_red_destroy, .get_stats = mlxsw_sp_qdisc_get_red_stats, .get_xstats = mlxsw_sp_qdisc_get_red_xstats, .clean_stats = mlxsw_sp_setup_tc_qdisc_red_clean_stats, .find_class = mlxsw_sp_qdisc_leaf_find_class, }; static int __mlxsw_sp_setup_tc_red(struct mlxsw_sp_port *mlxsw_sp_port, struct tc_red_qopt_offload *p) { struct mlxsw_sp_qdisc *mlxsw_sp_qdisc; mlxsw_sp_qdisc = mlxsw_sp_qdisc_find(mlxsw_sp_port, p->parent, false); if (!mlxsw_sp_qdisc) return -EOPNOTSUPP; if (p->command == TC_RED_REPLACE) return mlxsw_sp_qdisc_replace(mlxsw_sp_port, p->handle, mlxsw_sp_qdisc, &mlxsw_sp_qdisc_ops_red, &p->set); if (!mlxsw_sp_qdisc_compare(mlxsw_sp_qdisc, p->handle)) return -EOPNOTSUPP; switch (p->command) { case TC_RED_DESTROY: return mlxsw_sp_qdisc_destroy(mlxsw_sp_port, mlxsw_sp_qdisc); case TC_RED_XSTATS: return mlxsw_sp_qdisc_get_xstats(mlxsw_sp_port, mlxsw_sp_qdisc, p->xstats); case TC_RED_STATS: return mlxsw_sp_qdisc_get_stats(mlxsw_sp_port, mlxsw_sp_qdisc, &p->stats); default: return -EOPNOTSUPP; } } int mlxsw_sp_setup_tc_red(struct mlxsw_sp_port *mlxsw_sp_port, struct tc_red_qopt_offload *p) { int err; mutex_lock(&mlxsw_sp_port->qdisc->lock); err = __mlxsw_sp_setup_tc_red(mlxsw_sp_port, p); mutex_unlock(&mlxsw_sp_port->qdisc->lock); return err; } static void mlxsw_sp_setup_tc_qdisc_leaf_clean_stats(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc) { u64 backlog_cells = 0; u64 tx_packets = 0; u64 tx_bytes = 0; u64 drops = 0; mlxsw_sp_qdisc_collect_tc_stats(mlxsw_sp_port, mlxsw_sp_qdisc, &tx_bytes, &tx_packets, &drops, &backlog_cells); mlxsw_sp_qdisc->stats_base.tx_packets = tx_packets; mlxsw_sp_qdisc->stats_base.tx_bytes = tx_bytes; mlxsw_sp_qdisc->stats_base.drops = drops; mlxsw_sp_qdisc->stats_base.backlog = 0; } static int mlxsw_sp_qdisc_tbf_destroy(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc) { return mlxsw_sp_port_ets_maxrate_set(mlxsw_sp_port, MLXSW_REG_QEEC_HR_SUBGROUP, mlxsw_sp_qdisc->tclass_num, 0, MLXSW_REG_QEEC_MAS_DIS, 0); } static int mlxsw_sp_qdisc_tbf_bs(struct mlxsw_sp_port *mlxsw_sp_port, u32 max_size, u8 *p_burst_size) { /* TBF burst size is configured in bytes. The ASIC burst size value is * ((2 ^ bs) * 512 bits. Convert the TBF bytes to 512-bit units. */ u32 bs512 = max_size / 64; u8 bs = fls(bs512); if (!bs) return -EINVAL; --bs; /* Demand a power of two. */ if ((1 << bs) != bs512) return -EINVAL; if (bs < mlxsw_sp_port->mlxsw_sp->lowest_shaper_bs || bs > MLXSW_REG_QEEC_HIGHEST_SHAPER_BS) return -EINVAL; *p_burst_size = bs; return 0; } static u32 mlxsw_sp_qdisc_tbf_max_size(u8 bs) { return (1U << bs) * 64; } static u64 mlxsw_sp_qdisc_tbf_rate_kbps(struct tc_tbf_qopt_offload_replace_params *p) { /* TBF interface is in bytes/s, whereas Spectrum ASIC is configured in * Kbits/s. */ return div_u64(p->rate.rate_bytes_ps, 1000) * 8; } static int mlxsw_sp_qdisc_tbf_check_params(struct mlxsw_sp_port *mlxsw_sp_port, void *params) { struct tc_tbf_qopt_offload_replace_params *p = params; struct mlxsw_sp *mlxsw_sp = mlxsw_sp_port->mlxsw_sp; u64 rate_kbps = mlxsw_sp_qdisc_tbf_rate_kbps(p); u8 burst_size; int err; if (rate_kbps >= MLXSW_REG_QEEC_MAS_DIS) { dev_err(mlxsw_sp_port->mlxsw_sp->bus_info->dev, "spectrum: TBF: rate of %lluKbps must be below %u\n", rate_kbps, MLXSW_REG_QEEC_MAS_DIS); return -EINVAL; } err = mlxsw_sp_qdisc_tbf_bs(mlxsw_sp_port, p->max_size, &burst_size); if (err) { u8 highest_shaper_bs = MLXSW_REG_QEEC_HIGHEST_SHAPER_BS; dev_err(mlxsw_sp->bus_info->dev, "spectrum: TBF: invalid burst size of %u, must be a power of two between %u and %u", p->max_size, mlxsw_sp_qdisc_tbf_max_size(mlxsw_sp->lowest_shaper_bs), mlxsw_sp_qdisc_tbf_max_size(highest_shaper_bs)); return -EINVAL; } return 0; } static int mlxsw_sp_qdisc_tbf_replace(struct mlxsw_sp_port *mlxsw_sp_port, u32 handle, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, void *params) { struct tc_tbf_qopt_offload_replace_params *p = params; u64 rate_kbps = mlxsw_sp_qdisc_tbf_rate_kbps(p); u8 burst_size; int err; err = mlxsw_sp_qdisc_tbf_bs(mlxsw_sp_port, p->max_size, &burst_size); if (WARN_ON_ONCE(err)) /* check_params above was supposed to reject this value. */ return -EINVAL; /* Configure subgroup shaper, so that both UC and MC traffic is subject * to shaping. That is unlike RED, however UC queue lengths are going to * be different than MC ones due to different pool and quota * configurations, so the configuration is not applicable. For shaper on * the other hand, subjecting the overall stream to the configured * shaper makes sense. Also note that that is what we do for * ieee_setmaxrate(). */ return mlxsw_sp_port_ets_maxrate_set(mlxsw_sp_port, MLXSW_REG_QEEC_HR_SUBGROUP, mlxsw_sp_qdisc->tclass_num, 0, rate_kbps, burst_size); } static void mlxsw_sp_qdisc_tbf_unoffload(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, void *params) { struct tc_tbf_qopt_offload_replace_params *p = params; mlxsw_sp_qdisc_leaf_unoffload(mlxsw_sp_port, mlxsw_sp_qdisc, p->qstats); } static int mlxsw_sp_qdisc_get_tbf_stats(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, struct tc_qopt_offload_stats *stats_ptr) { mlxsw_sp_qdisc_get_tc_stats(mlxsw_sp_port, mlxsw_sp_qdisc, stats_ptr); return 0; } static struct mlxsw_sp_qdisc_ops mlxsw_sp_qdisc_ops_tbf = { .type = MLXSW_SP_QDISC_TBF, .check_params = mlxsw_sp_qdisc_tbf_check_params, .replace = mlxsw_sp_qdisc_tbf_replace, .unoffload = mlxsw_sp_qdisc_tbf_unoffload, .destroy = mlxsw_sp_qdisc_tbf_destroy, .get_stats = mlxsw_sp_qdisc_get_tbf_stats, .clean_stats = mlxsw_sp_setup_tc_qdisc_leaf_clean_stats, .find_class = mlxsw_sp_qdisc_leaf_find_class, }; static int __mlxsw_sp_setup_tc_tbf(struct mlxsw_sp_port *mlxsw_sp_port, struct tc_tbf_qopt_offload *p) { struct mlxsw_sp_qdisc *mlxsw_sp_qdisc; mlxsw_sp_qdisc = mlxsw_sp_qdisc_find(mlxsw_sp_port, p->parent, false); if (!mlxsw_sp_qdisc) return -EOPNOTSUPP; if (p->command == TC_TBF_REPLACE) return mlxsw_sp_qdisc_replace(mlxsw_sp_port, p->handle, mlxsw_sp_qdisc, &mlxsw_sp_qdisc_ops_tbf, &p->replace_params); if (!mlxsw_sp_qdisc_compare(mlxsw_sp_qdisc, p->handle)) return -EOPNOTSUPP; switch (p->command) { case TC_TBF_DESTROY: return mlxsw_sp_qdisc_destroy(mlxsw_sp_port, mlxsw_sp_qdisc); case TC_TBF_STATS: return mlxsw_sp_qdisc_get_stats(mlxsw_sp_port, mlxsw_sp_qdisc, &p->stats); default: return -EOPNOTSUPP; } } int mlxsw_sp_setup_tc_tbf(struct mlxsw_sp_port *mlxsw_sp_port, struct tc_tbf_qopt_offload *p) { int err; mutex_lock(&mlxsw_sp_port->qdisc->lock); err = __mlxsw_sp_setup_tc_tbf(mlxsw_sp_port, p); mutex_unlock(&mlxsw_sp_port->qdisc->lock); return err; } static int mlxsw_sp_qdisc_fifo_check_params(struct mlxsw_sp_port *mlxsw_sp_port, void *params) { return 0; } static int mlxsw_sp_qdisc_fifo_replace(struct mlxsw_sp_port *mlxsw_sp_port, u32 handle, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, void *params) { return 0; } static int mlxsw_sp_qdisc_get_fifo_stats(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, struct tc_qopt_offload_stats *stats_ptr) { mlxsw_sp_qdisc_get_tc_stats(mlxsw_sp_port, mlxsw_sp_qdisc, stats_ptr); return 0; } static struct mlxsw_sp_qdisc_ops mlxsw_sp_qdisc_ops_fifo = { .type = MLXSW_SP_QDISC_FIFO, .check_params = mlxsw_sp_qdisc_fifo_check_params, .replace = mlxsw_sp_qdisc_fifo_replace, .get_stats = mlxsw_sp_qdisc_get_fifo_stats, .clean_stats = mlxsw_sp_setup_tc_qdisc_leaf_clean_stats, }; static int __mlxsw_sp_setup_tc_fifo(struct mlxsw_sp_port *mlxsw_sp_port, struct tc_fifo_qopt_offload *p) { struct mlxsw_sp_qdisc_state *qdisc_state = mlxsw_sp_port->qdisc; struct mlxsw_sp_qdisc *mlxsw_sp_qdisc; unsigned int band; u32 parent_handle; mlxsw_sp_qdisc = mlxsw_sp_qdisc_find(mlxsw_sp_port, p->parent, false); if (!mlxsw_sp_qdisc && p->handle == TC_H_UNSPEC) { parent_handle = TC_H_MAJ(p->parent); if (parent_handle != qdisc_state->future_handle) { /* This notifications is for a different Qdisc than * previously. Wipe the future cache. */ memset(qdisc_state->future_fifos, 0, sizeof(qdisc_state->future_fifos)); qdisc_state->future_handle = parent_handle; } band = TC_H_MIN(p->parent) - 1; if (band < IEEE_8021QAZ_MAX_TCS) { if (p->command == TC_FIFO_REPLACE) qdisc_state->future_fifos[band] = true; else if (p->command == TC_FIFO_DESTROY) qdisc_state->future_fifos[band] = false; } } if (!mlxsw_sp_qdisc) return -EOPNOTSUPP; if (p->command == TC_FIFO_REPLACE) { return mlxsw_sp_qdisc_replace(mlxsw_sp_port, p->handle, mlxsw_sp_qdisc, &mlxsw_sp_qdisc_ops_fifo, NULL); } if (!mlxsw_sp_qdisc_compare(mlxsw_sp_qdisc, p->handle)) return -EOPNOTSUPP; switch (p->command) { case TC_FIFO_DESTROY: return mlxsw_sp_qdisc_destroy(mlxsw_sp_port, mlxsw_sp_qdisc); case TC_FIFO_STATS: return mlxsw_sp_qdisc_get_stats(mlxsw_sp_port, mlxsw_sp_qdisc, &p->stats); case TC_FIFO_REPLACE: /* Handled above. */ break; } return -EOPNOTSUPP; } int mlxsw_sp_setup_tc_fifo(struct mlxsw_sp_port *mlxsw_sp_port, struct tc_fifo_qopt_offload *p) { int err; mutex_lock(&mlxsw_sp_port->qdisc->lock); err = __mlxsw_sp_setup_tc_fifo(mlxsw_sp_port, p); mutex_unlock(&mlxsw_sp_port->qdisc->lock); return err; } static int __mlxsw_sp_qdisc_ets_destroy(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc) { int i; for (i = 0; i < mlxsw_sp_qdisc->num_classes; i++) { mlxsw_sp_port_prio_tc_set(mlxsw_sp_port, i, MLXSW_SP_PORT_DEFAULT_TCLASS); mlxsw_sp_port_ets_set(mlxsw_sp_port, MLXSW_REG_QEEC_HR_SUBGROUP, i, 0, false, 0); mlxsw_sp_qdisc_destroy(mlxsw_sp_port, &mlxsw_sp_qdisc->qdiscs[i]); mlxsw_sp_qdisc->qdiscs[i].prio_bitmap = 0; } return 0; } static int mlxsw_sp_qdisc_prio_destroy(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc) { return __mlxsw_sp_qdisc_ets_destroy(mlxsw_sp_port, mlxsw_sp_qdisc); } static int __mlxsw_sp_qdisc_ets_check_params(unsigned int nbands) { if (nbands > IEEE_8021QAZ_MAX_TCS) return -EOPNOTSUPP; return 0; } static int mlxsw_sp_qdisc_prio_check_params(struct mlxsw_sp_port *mlxsw_sp_port, void *params) { struct tc_prio_qopt_offload_params *p = params; return __mlxsw_sp_qdisc_ets_check_params(p->bands); } static int __mlxsw_sp_qdisc_ets_replace(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, u32 handle, unsigned int nbands, const unsigned int *quanta, const unsigned int *weights, const u8 *priomap) { struct mlxsw_sp_qdisc_state *qdisc_state = mlxsw_sp_port->qdisc; struct mlxsw_sp_qdisc *child_qdisc; int tclass, i, band, backlog; u8 old_priomap; int err; for (band = 0; band < nbands; band++) { tclass = MLXSW_SP_PRIO_BAND_TO_TCLASS(band); child_qdisc = &mlxsw_sp_qdisc->qdiscs[band]; old_priomap = child_qdisc->prio_bitmap; child_qdisc->prio_bitmap = 0; err = mlxsw_sp_port_ets_set(mlxsw_sp_port, MLXSW_REG_QEEC_HR_SUBGROUP, tclass, 0, !!quanta[band], weights[band]); if (err) return err; for (i = 0; i < IEEE_8021QAZ_MAX_TCS; i++) { if (priomap[i] == band) { child_qdisc->prio_bitmap |= BIT(i); if (BIT(i) & old_priomap) continue; err = mlxsw_sp_port_prio_tc_set(mlxsw_sp_port, i, tclass); if (err) return err; } } child_qdisc->tclass_num = tclass; if (old_priomap != child_qdisc->prio_bitmap && child_qdisc->ops && child_qdisc->ops->clean_stats) { backlog = child_qdisc->stats_base.backlog; child_qdisc->ops->clean_stats(mlxsw_sp_port, child_qdisc); child_qdisc->stats_base.backlog = backlog; } if (handle == qdisc_state->future_handle && qdisc_state->future_fifos[band]) { err = mlxsw_sp_qdisc_replace(mlxsw_sp_port, TC_H_UNSPEC, child_qdisc, &mlxsw_sp_qdisc_ops_fifo, NULL); if (err) return err; } } for (; band < IEEE_8021QAZ_MAX_TCS; band++) { tclass = MLXSW_SP_PRIO_BAND_TO_TCLASS(band); child_qdisc = &mlxsw_sp_qdisc->qdiscs[band]; child_qdisc->prio_bitmap = 0; mlxsw_sp_qdisc_destroy(mlxsw_sp_port, child_qdisc); mlxsw_sp_port_ets_set(mlxsw_sp_port, MLXSW_REG_QEEC_HR_SUBGROUP, tclass, 0, false, 0); } qdisc_state->future_handle = TC_H_UNSPEC; memset(qdisc_state->future_fifos, 0, sizeof(qdisc_state->future_fifos)); return 0; } static int mlxsw_sp_qdisc_prio_replace(struct mlxsw_sp_port *mlxsw_sp_port, u32 handle, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, void *params) { struct tc_prio_qopt_offload_params *p = params; unsigned int zeroes[TCQ_ETS_MAX_BANDS] = {0}; return __mlxsw_sp_qdisc_ets_replace(mlxsw_sp_port, mlxsw_sp_qdisc, handle, p->bands, zeroes, zeroes, p->priomap); } static void __mlxsw_sp_qdisc_ets_unoffload(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, struct gnet_stats_queue *qstats) { u64 backlog; backlog = mlxsw_sp_cells_bytes(mlxsw_sp_port->mlxsw_sp, mlxsw_sp_qdisc->stats_base.backlog); qstats->backlog -= backlog; } static void mlxsw_sp_qdisc_prio_unoffload(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, void *params) { struct tc_prio_qopt_offload_params *p = params; __mlxsw_sp_qdisc_ets_unoffload(mlxsw_sp_port, mlxsw_sp_qdisc, p->qstats); } static int mlxsw_sp_qdisc_get_prio_stats(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, struct tc_qopt_offload_stats *stats_ptr) { struct mlxsw_sp_qdisc *tc_qdisc; u64 tx_packets = 0; u64 tx_bytes = 0; u64 backlog = 0; u64 drops = 0; int i; for (i = 0; i < mlxsw_sp_qdisc->num_classes; i++) { tc_qdisc = &mlxsw_sp_qdisc->qdiscs[i]; mlxsw_sp_qdisc_collect_tc_stats(mlxsw_sp_port, tc_qdisc, &tx_bytes, &tx_packets, &drops, &backlog); } mlxsw_sp_qdisc_update_stats(mlxsw_sp_port->mlxsw_sp, mlxsw_sp_qdisc, tx_bytes, tx_packets, drops, backlog, stats_ptr); return 0; } static void mlxsw_sp_setup_tc_qdisc_prio_clean_stats(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc) { struct mlxsw_sp_qdisc_stats *stats_base; struct mlxsw_sp_port_xstats *xstats; struct rtnl_link_stats64 *stats; int i; xstats = &mlxsw_sp_port->periodic_hw_stats.xstats; stats = &mlxsw_sp_port->periodic_hw_stats.stats; stats_base = &mlxsw_sp_qdisc->stats_base; stats_base->tx_packets = stats->tx_packets; stats_base->tx_bytes = stats->tx_bytes; stats_base->drops = 0; for (i = 0; i < IEEE_8021QAZ_MAX_TCS; i++) { stats_base->drops += mlxsw_sp_xstats_tail_drop(xstats, i); stats_base->drops += xstats->wred_drop[i]; } mlxsw_sp_qdisc->stats_base.backlog = 0; } static struct mlxsw_sp_qdisc * mlxsw_sp_qdisc_prio_find_class(struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, u32 parent) { int child_index = TC_H_MIN(parent); int band = child_index - 1; if (band < 0 || band >= mlxsw_sp_qdisc->num_classes) return NULL; return &mlxsw_sp_qdisc->qdiscs[band]; } static struct mlxsw_sp_qdisc_ops mlxsw_sp_qdisc_ops_prio = { .type = MLXSW_SP_QDISC_PRIO, .check_params = mlxsw_sp_qdisc_prio_check_params, .replace = mlxsw_sp_qdisc_prio_replace, .unoffload = mlxsw_sp_qdisc_prio_unoffload, .destroy = mlxsw_sp_qdisc_prio_destroy, .get_stats = mlxsw_sp_qdisc_get_prio_stats, .clean_stats = mlxsw_sp_setup_tc_qdisc_prio_clean_stats, .find_class = mlxsw_sp_qdisc_prio_find_class, .num_classes = IEEE_8021QAZ_MAX_TCS, }; static int mlxsw_sp_qdisc_ets_check_params(struct mlxsw_sp_port *mlxsw_sp_port, void *params) { struct tc_ets_qopt_offload_replace_params *p = params; return __mlxsw_sp_qdisc_ets_check_params(p->bands); } static int mlxsw_sp_qdisc_ets_replace(struct mlxsw_sp_port *mlxsw_sp_port, u32 handle, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, void *params) { struct tc_ets_qopt_offload_replace_params *p = params; return __mlxsw_sp_qdisc_ets_replace(mlxsw_sp_port, mlxsw_sp_qdisc, handle, p->bands, p->quanta, p->weights, p->priomap); } static void mlxsw_sp_qdisc_ets_unoffload(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, void *params) { struct tc_ets_qopt_offload_replace_params *p = params; __mlxsw_sp_qdisc_ets_unoffload(mlxsw_sp_port, mlxsw_sp_qdisc, p->qstats); } static int mlxsw_sp_qdisc_ets_destroy(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc) { return __mlxsw_sp_qdisc_ets_destroy(mlxsw_sp_port, mlxsw_sp_qdisc); } static struct mlxsw_sp_qdisc_ops mlxsw_sp_qdisc_ops_ets = { .type = MLXSW_SP_QDISC_ETS, .check_params = mlxsw_sp_qdisc_ets_check_params, .replace = mlxsw_sp_qdisc_ets_replace, .unoffload = mlxsw_sp_qdisc_ets_unoffload, .destroy = mlxsw_sp_qdisc_ets_destroy, .get_stats = mlxsw_sp_qdisc_get_prio_stats, .clean_stats = mlxsw_sp_setup_tc_qdisc_prio_clean_stats, .find_class = mlxsw_sp_qdisc_prio_find_class, .num_classes = IEEE_8021QAZ_MAX_TCS, }; /* Linux allows linking of Qdiscs to arbitrary classes (so long as the resulting * graph is free of cycles). These operations do not change the parent handle * though, which means it can be incomplete (if there is more than one class * where the Qdisc in question is grafted) or outright wrong (if the Qdisc was * linked to a different class and then removed from the original class). * * E.g. consider this sequence of operations: * * # tc qdisc add dev swp1 root handle 1: prio * # tc qdisc add dev swp1 parent 1:3 handle 13: red limit 1000000 avpkt 10000 * RED: set bandwidth to 10Mbit * # tc qdisc link dev swp1 handle 13: parent 1:2 * * At this point, both 1:2 and 1:3 have the same RED Qdisc instance as their * child. But RED will still only claim that 1:3 is its parent. If it's removed * from that band, its only parent will be 1:2, but it will continue to claim * that it is in fact 1:3. * * The notification for child Qdisc replace (e.g. TC_RED_REPLACE) comes before * the notification for parent graft (e.g. TC_PRIO_GRAFT). We take the replace * notification to offload the child Qdisc, based on its parent handle, and use * the graft operation to validate that the class where the child is actually * grafted corresponds to the parent handle. If the two don't match, we * unoffload the child. */ static int __mlxsw_sp_qdisc_ets_graft(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, u8 band, u32 child_handle) { struct mlxsw_sp_qdisc *old_qdisc; u32 parent; if (band < mlxsw_sp_qdisc->num_classes && mlxsw_sp_qdisc->qdiscs[band].handle == child_handle) return 0; if (!child_handle) { /* This is an invisible FIFO replacing the original Qdisc. * Ignore it--the original Qdisc's destroy will follow. */ return 0; } /* See if the grafted qdisc is already offloaded on any tclass. If so, * unoffload it. */ old_qdisc = mlxsw_sp_qdisc_find_by_handle(mlxsw_sp_port, child_handle); if (old_qdisc) mlxsw_sp_qdisc_destroy(mlxsw_sp_port, old_qdisc); parent = TC_H_MAKE(mlxsw_sp_qdisc->handle, band + 1); mlxsw_sp_qdisc = mlxsw_sp_qdisc->ops->find_class(mlxsw_sp_qdisc, parent); if (!WARN_ON(!mlxsw_sp_qdisc)) mlxsw_sp_qdisc_destroy(mlxsw_sp_port, mlxsw_sp_qdisc); return -EOPNOTSUPP; } static int mlxsw_sp_qdisc_prio_graft(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_qdisc *mlxsw_sp_qdisc, struct tc_prio_qopt_offload_graft_params *p) { return __mlxsw_sp_qdisc_ets_graft(mlxsw_sp_port, mlxsw_sp_qdisc, p->band, p->child_handle); } static int __mlxsw_sp_setup_tc_prio(struct mlxsw_sp_port *mlxsw_sp_port, struct tc_prio_qopt_offload *p) { struct mlxsw_sp_qdisc *mlxsw_sp_qdisc; mlxsw_sp_qdisc = mlxsw_sp_qdisc_find(mlxsw_sp_port, p->parent, true); if (!mlxsw_sp_qdisc) return -EOPNOTSUPP; if (p->command == TC_PRIO_REPLACE) return mlxsw_sp_qdisc_replace(mlxsw_sp_port, p->handle, mlxsw_sp_qdisc, &mlxsw_sp_qdisc_ops_prio, &p->replace_params); if (!mlxsw_sp_qdisc_compare(mlxsw_sp_qdisc, p->handle)) return -EOPNOTSUPP; switch (p->command) { case TC_PRIO_DESTROY: return mlxsw_sp_qdisc_destroy(mlxsw_sp_port, mlxsw_sp_qdisc); case TC_PRIO_STATS: return mlxsw_sp_qdisc_get_stats(mlxsw_sp_port, mlxsw_sp_qdisc, &p->stats); case TC_PRIO_GRAFT: return mlxsw_sp_qdisc_prio_graft(mlxsw_sp_port, mlxsw_sp_qdisc, &p->graft_params); default: return -EOPNOTSUPP; } } int mlxsw_sp_setup_tc_prio(struct mlxsw_sp_port *mlxsw_sp_port, struct tc_prio_qopt_offload *p) { int err; mutex_lock(&mlxsw_sp_port->qdisc->lock); err = __mlxsw_sp_setup_tc_prio(mlxsw_sp_port, p); mutex_unlock(&mlxsw_sp_port->qdisc->lock); return err; } static int __mlxsw_sp_setup_tc_ets(struct mlxsw_sp_port *mlxsw_sp_port, struct tc_ets_qopt_offload *p) { struct mlxsw_sp_qdisc *mlxsw_sp_qdisc; mlxsw_sp_qdisc = mlxsw_sp_qdisc_find(mlxsw_sp_port, p->parent, true); if (!mlxsw_sp_qdisc) return -EOPNOTSUPP; if (p->command == TC_ETS_REPLACE) return mlxsw_sp_qdisc_replace(mlxsw_sp_port, p->handle, mlxsw_sp_qdisc, &mlxsw_sp_qdisc_ops_ets, &p->replace_params); if (!mlxsw_sp_qdisc_compare(mlxsw_sp_qdisc, p->handle)) return -EOPNOTSUPP; switch (p->command) { case TC_ETS_DESTROY: return mlxsw_sp_qdisc_destroy(mlxsw_sp_port, mlxsw_sp_qdisc); case TC_ETS_STATS: return mlxsw_sp_qdisc_get_stats(mlxsw_sp_port, mlxsw_sp_qdisc, &p->stats); case TC_ETS_GRAFT: return __mlxsw_sp_qdisc_ets_graft(mlxsw_sp_port, mlxsw_sp_qdisc, p->graft_params.band, p->graft_params.child_handle); default: return -EOPNOTSUPP; } } int mlxsw_sp_setup_tc_ets(struct mlxsw_sp_port *mlxsw_sp_port, struct tc_ets_qopt_offload *p) { int err; mutex_lock(&mlxsw_sp_port->qdisc->lock); err = __mlxsw_sp_setup_tc_ets(mlxsw_sp_port, p); mutex_unlock(&mlxsw_sp_port->qdisc->lock); return err; } struct mlxsw_sp_qevent_block { struct list_head binding_list; struct list_head mall_entry_list; struct mlxsw_sp *mlxsw_sp; }; struct mlxsw_sp_qevent_binding { struct list_head list; struct mlxsw_sp_port *mlxsw_sp_port; u32 handle; int tclass_num; enum mlxsw_sp_span_trigger span_trigger; }; static LIST_HEAD(mlxsw_sp_qevent_block_cb_list); static int mlxsw_sp_qevent_span_configure(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_mall_entry *mall_entry, struct mlxsw_sp_qevent_binding *qevent_binding, const struct mlxsw_sp_span_agent_parms *agent_parms, int *p_span_id) { struct mlxsw_sp_port *mlxsw_sp_port = qevent_binding->mlxsw_sp_port; struct mlxsw_sp_span_trigger_parms trigger_parms = {}; int span_id; int err; err = mlxsw_sp_span_agent_get(mlxsw_sp, &span_id, agent_parms); if (err) return err; err = mlxsw_sp_span_analyzed_port_get(mlxsw_sp_port, true); if (err) goto err_analyzed_port_get; trigger_parms.span_id = span_id; trigger_parms.probability_rate = 1; err = mlxsw_sp_span_agent_bind(mlxsw_sp, qevent_binding->span_trigger, mlxsw_sp_port, &trigger_parms); if (err) goto err_agent_bind; err = mlxsw_sp_span_trigger_enable(mlxsw_sp_port, qevent_binding->span_trigger, qevent_binding->tclass_num); if (err) goto err_trigger_enable; *p_span_id = span_id; return 0; err_trigger_enable: mlxsw_sp_span_agent_unbind(mlxsw_sp, qevent_binding->span_trigger, mlxsw_sp_port, &trigger_parms); err_agent_bind: mlxsw_sp_span_analyzed_port_put(mlxsw_sp_port, true); err_analyzed_port_get: mlxsw_sp_span_agent_put(mlxsw_sp, span_id); return err; } static void mlxsw_sp_qevent_span_deconfigure(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_qevent_binding *qevent_binding, int span_id) { struct mlxsw_sp_port *mlxsw_sp_port = qevent_binding->mlxsw_sp_port; struct mlxsw_sp_span_trigger_parms trigger_parms = { .span_id = span_id, }; mlxsw_sp_span_trigger_disable(mlxsw_sp_port, qevent_binding->span_trigger, qevent_binding->tclass_num); mlxsw_sp_span_agent_unbind(mlxsw_sp, qevent_binding->span_trigger, mlxsw_sp_port, &trigger_parms); mlxsw_sp_span_analyzed_port_put(mlxsw_sp_port, true); mlxsw_sp_span_agent_put(mlxsw_sp, span_id); } static int mlxsw_sp_qevent_mirror_configure(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_mall_entry *mall_entry, struct mlxsw_sp_qevent_binding *qevent_binding) { struct mlxsw_sp_span_agent_parms agent_parms = { .to_dev = mall_entry->mirror.to_dev, }; return mlxsw_sp_qevent_span_configure(mlxsw_sp, mall_entry, qevent_binding, &agent_parms, &mall_entry->mirror.span_id); } static void mlxsw_sp_qevent_mirror_deconfigure(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_mall_entry *mall_entry, struct mlxsw_sp_qevent_binding *qevent_binding) { mlxsw_sp_qevent_span_deconfigure(mlxsw_sp, qevent_binding, mall_entry->mirror.span_id); } static int mlxsw_sp_qevent_trap_configure(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_mall_entry *mall_entry, struct mlxsw_sp_qevent_binding *qevent_binding) { struct mlxsw_sp_span_agent_parms agent_parms = { .session_id = MLXSW_SP_SPAN_SESSION_ID_BUFFER, }; int err; err = mlxsw_sp_trap_group_policer_hw_id_get(mlxsw_sp, DEVLINK_TRAP_GROUP_GENERIC_ID_BUFFER_DROPS, &agent_parms.policer_enable, &agent_parms.policer_id); if (err) return err; return mlxsw_sp_qevent_span_configure(mlxsw_sp, mall_entry, qevent_binding, &agent_parms, &mall_entry->trap.span_id); } static void mlxsw_sp_qevent_trap_deconfigure(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_mall_entry *mall_entry, struct mlxsw_sp_qevent_binding *qevent_binding) { mlxsw_sp_qevent_span_deconfigure(mlxsw_sp, qevent_binding, mall_entry->trap.span_id); } static int mlxsw_sp_qevent_entry_configure(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_mall_entry *mall_entry, struct mlxsw_sp_qevent_binding *qevent_binding) { switch (mall_entry->type) { case MLXSW_SP_MALL_ACTION_TYPE_MIRROR: return mlxsw_sp_qevent_mirror_configure(mlxsw_sp, mall_entry, qevent_binding); case MLXSW_SP_MALL_ACTION_TYPE_TRAP: return mlxsw_sp_qevent_trap_configure(mlxsw_sp, mall_entry, qevent_binding); default: /* This should have been validated away. */ WARN_ON(1); return -EOPNOTSUPP; } } static void mlxsw_sp_qevent_entry_deconfigure(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_mall_entry *mall_entry, struct mlxsw_sp_qevent_binding *qevent_binding) { switch (mall_entry->type) { case MLXSW_SP_MALL_ACTION_TYPE_MIRROR: return mlxsw_sp_qevent_mirror_deconfigure(mlxsw_sp, mall_entry, qevent_binding); case MLXSW_SP_MALL_ACTION_TYPE_TRAP: return mlxsw_sp_qevent_trap_deconfigure(mlxsw_sp, mall_entry, qevent_binding); default: WARN_ON(1); return; } } static int mlxsw_sp_qevent_binding_configure(struct mlxsw_sp_qevent_block *qevent_block, struct mlxsw_sp_qevent_binding *qevent_binding) { struct mlxsw_sp_mall_entry *mall_entry; int err; list_for_each_entry(mall_entry, &qevent_block->mall_entry_list, list) { err = mlxsw_sp_qevent_entry_configure(qevent_block->mlxsw_sp, mall_entry, qevent_binding); if (err) goto err_entry_configure; } return 0; err_entry_configure: list_for_each_entry_continue_reverse(mall_entry, &qevent_block->mall_entry_list, list) mlxsw_sp_qevent_entry_deconfigure(qevent_block->mlxsw_sp, mall_entry, qevent_binding); return err; } static void mlxsw_sp_qevent_binding_deconfigure(struct mlxsw_sp_qevent_block *qevent_block, struct mlxsw_sp_qevent_binding *qevent_binding) { struct mlxsw_sp_mall_entry *mall_entry; list_for_each_entry(mall_entry, &qevent_block->mall_entry_list, list) mlxsw_sp_qevent_entry_deconfigure(qevent_block->mlxsw_sp, mall_entry, qevent_binding); } static int mlxsw_sp_qevent_block_configure(struct mlxsw_sp_qevent_block *qevent_block) { struct mlxsw_sp_qevent_binding *qevent_binding; int err; list_for_each_entry(qevent_binding, &qevent_block->binding_list, list) { err = mlxsw_sp_qevent_binding_configure(qevent_block, qevent_binding); if (err) goto err_binding_configure; } return 0; err_binding_configure: list_for_each_entry_continue_reverse(qevent_binding, &qevent_block->binding_list, list) mlxsw_sp_qevent_binding_deconfigure(qevent_block, qevent_binding); return err; } static void mlxsw_sp_qevent_block_deconfigure(struct mlxsw_sp_qevent_block *qevent_block) { struct mlxsw_sp_qevent_binding *qevent_binding; list_for_each_entry(qevent_binding, &qevent_block->binding_list, list) mlxsw_sp_qevent_binding_deconfigure(qevent_block, qevent_binding); } static struct mlxsw_sp_mall_entry * mlxsw_sp_qevent_mall_entry_find(struct mlxsw_sp_qevent_block *block, unsigned long cookie) { struct mlxsw_sp_mall_entry *mall_entry; list_for_each_entry(mall_entry, &block->mall_entry_list, list) if (mall_entry->cookie == cookie) return mall_entry; return NULL; } static int mlxsw_sp_qevent_mall_replace(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_qevent_block *qevent_block, struct tc_cls_matchall_offload *f) { struct mlxsw_sp_mall_entry *mall_entry; struct flow_action_entry *act; int err; /* It should not currently be possible to replace a matchall rule. So * this must be a new rule. */ if (!list_empty(&qevent_block->mall_entry_list)) { NL_SET_ERR_MSG(f->common.extack, "At most one filter supported"); return -EOPNOTSUPP; } if (f->rule->action.num_entries != 1) { NL_SET_ERR_MSG(f->common.extack, "Only singular actions supported"); return -EOPNOTSUPP; } if (f->common.chain_index) { NL_SET_ERR_MSG(f->common.extack, "Only chain 0 is supported"); return -EOPNOTSUPP; } if (f->common.protocol != htons(ETH_P_ALL)) { NL_SET_ERR_MSG(f->common.extack, "Protocol matching not supported"); return -EOPNOTSUPP; } act = &f->rule->action.entries[0]; if (!(act->hw_stats & FLOW_ACTION_HW_STATS_DISABLED)) { NL_SET_ERR_MSG(f->common.extack, "HW counters not supported on qevents"); return -EOPNOTSUPP; } mall_entry = kzalloc(sizeof(*mall_entry), GFP_KERNEL); if (!mall_entry) return -ENOMEM; mall_entry->cookie = f->cookie; if (act->id == FLOW_ACTION_MIRRED) { mall_entry->type = MLXSW_SP_MALL_ACTION_TYPE_MIRROR; mall_entry->mirror.to_dev = act->dev; } else if (act->id == FLOW_ACTION_TRAP) { mall_entry->type = MLXSW_SP_MALL_ACTION_TYPE_TRAP; } else { NL_SET_ERR_MSG(f->common.extack, "Unsupported action"); err = -EOPNOTSUPP; goto err_unsupported_action; } list_add_tail(&mall_entry->list, &qevent_block->mall_entry_list); err = mlxsw_sp_qevent_block_configure(qevent_block); if (err) goto err_block_configure; return 0; err_block_configure: list_del(&mall_entry->list); err_unsupported_action: kfree(mall_entry); return err; } static void mlxsw_sp_qevent_mall_destroy(struct mlxsw_sp_qevent_block *qevent_block, struct tc_cls_matchall_offload *f) { struct mlxsw_sp_mall_entry *mall_entry; mall_entry = mlxsw_sp_qevent_mall_entry_find(qevent_block, f->cookie); if (!mall_entry) return; mlxsw_sp_qevent_block_deconfigure(qevent_block); list_del(&mall_entry->list); kfree(mall_entry); } static int mlxsw_sp_qevent_block_mall_cb(struct mlxsw_sp_qevent_block *qevent_block, struct tc_cls_matchall_offload *f) { struct mlxsw_sp *mlxsw_sp = qevent_block->mlxsw_sp; switch (f->command) { case TC_CLSMATCHALL_REPLACE: return mlxsw_sp_qevent_mall_replace(mlxsw_sp, qevent_block, f); case TC_CLSMATCHALL_DESTROY: mlxsw_sp_qevent_mall_destroy(qevent_block, f); return 0; default: return -EOPNOTSUPP; } } static int mlxsw_sp_qevent_block_cb(enum tc_setup_type type, void *type_data, void *cb_priv) { struct mlxsw_sp_qevent_block *qevent_block = cb_priv; switch (type) { case TC_SETUP_CLSMATCHALL: return mlxsw_sp_qevent_block_mall_cb(qevent_block, type_data); default: return -EOPNOTSUPP; } } static struct mlxsw_sp_qevent_block *mlxsw_sp_qevent_block_create(struct mlxsw_sp *mlxsw_sp, struct net *net) { struct mlxsw_sp_qevent_block *qevent_block; qevent_block = kzalloc(sizeof(*qevent_block), GFP_KERNEL); if (!qevent_block) return NULL; INIT_LIST_HEAD(&qevent_block->binding_list); INIT_LIST_HEAD(&qevent_block->mall_entry_list); qevent_block->mlxsw_sp = mlxsw_sp; return qevent_block; } static void mlxsw_sp_qevent_block_destroy(struct mlxsw_sp_qevent_block *qevent_block) { WARN_ON(!list_empty(&qevent_block->binding_list)); WARN_ON(!list_empty(&qevent_block->mall_entry_list)); kfree(qevent_block); } static void mlxsw_sp_qevent_block_release(void *cb_priv) { struct mlxsw_sp_qevent_block *qevent_block = cb_priv; mlxsw_sp_qevent_block_destroy(qevent_block); } static struct mlxsw_sp_qevent_binding * mlxsw_sp_qevent_binding_create(struct mlxsw_sp_port *mlxsw_sp_port, u32 handle, int tclass_num, enum mlxsw_sp_span_trigger span_trigger) { struct mlxsw_sp_qevent_binding *binding; binding = kzalloc(sizeof(*binding), GFP_KERNEL); if (!binding) return ERR_PTR(-ENOMEM); binding->mlxsw_sp_port = mlxsw_sp_port; binding->handle = handle; binding->tclass_num = tclass_num; binding->span_trigger = span_trigger; return binding; } static void mlxsw_sp_qevent_binding_destroy(struct mlxsw_sp_qevent_binding *binding) { kfree(binding); } static struct mlxsw_sp_qevent_binding * mlxsw_sp_qevent_binding_lookup(struct mlxsw_sp_qevent_block *block, struct mlxsw_sp_port *mlxsw_sp_port, u32 handle, enum mlxsw_sp_span_trigger span_trigger) { struct mlxsw_sp_qevent_binding *qevent_binding; list_for_each_entry(qevent_binding, &block->binding_list, list) if (qevent_binding->mlxsw_sp_port == mlxsw_sp_port && qevent_binding->handle == handle && qevent_binding->span_trigger == span_trigger) return qevent_binding; return NULL; } static int mlxsw_sp_setup_tc_block_qevent_bind(struct mlxsw_sp_port *mlxsw_sp_port, struct flow_block_offload *f, enum mlxsw_sp_span_trigger span_trigger) { struct mlxsw_sp *mlxsw_sp = mlxsw_sp_port->mlxsw_sp; struct mlxsw_sp_qevent_binding *qevent_binding; struct mlxsw_sp_qevent_block *qevent_block; struct flow_block_cb *block_cb; struct mlxsw_sp_qdisc *qdisc; bool register_block = false; int err; block_cb = flow_block_cb_lookup(f->block, mlxsw_sp_qevent_block_cb, mlxsw_sp); if (!block_cb) { qevent_block = mlxsw_sp_qevent_block_create(mlxsw_sp, f->net); if (!qevent_block) return -ENOMEM; block_cb = flow_block_cb_alloc(mlxsw_sp_qevent_block_cb, mlxsw_sp, qevent_block, mlxsw_sp_qevent_block_release); if (IS_ERR(block_cb)) { mlxsw_sp_qevent_block_destroy(qevent_block); return PTR_ERR(block_cb); } register_block = true; } else { qevent_block = flow_block_cb_priv(block_cb); } flow_block_cb_incref(block_cb); qdisc = mlxsw_sp_qdisc_find_by_handle(mlxsw_sp_port, f->sch->handle); if (!qdisc) { NL_SET_ERR_MSG(f->extack, "Qdisc not offloaded"); err = -ENOENT; goto err_find_qdisc; } if (WARN_ON(mlxsw_sp_qevent_binding_lookup(qevent_block, mlxsw_sp_port, f->sch->handle, span_trigger))) { err = -EEXIST; goto err_binding_exists; } qevent_binding = mlxsw_sp_qevent_binding_create(mlxsw_sp_port, f->sch->handle, qdisc->tclass_num, span_trigger); if (IS_ERR(qevent_binding)) { err = PTR_ERR(qevent_binding); goto err_binding_create; } err = mlxsw_sp_qevent_binding_configure(qevent_block, qevent_binding); if (err) goto err_binding_configure; list_add(&qevent_binding->list, &qevent_block->binding_list); if (register_block) { flow_block_cb_add(block_cb, f); list_add_tail(&block_cb->driver_list, &mlxsw_sp_qevent_block_cb_list); } return 0; err_binding_configure: mlxsw_sp_qevent_binding_destroy(qevent_binding); err_binding_create: err_binding_exists: err_find_qdisc: if (!flow_block_cb_decref(block_cb)) flow_block_cb_free(block_cb); return err; } static void mlxsw_sp_setup_tc_block_qevent_unbind(struct mlxsw_sp_port *mlxsw_sp_port, struct flow_block_offload *f, enum mlxsw_sp_span_trigger span_trigger) { struct mlxsw_sp *mlxsw_sp = mlxsw_sp_port->mlxsw_sp; struct mlxsw_sp_qevent_binding *qevent_binding; struct mlxsw_sp_qevent_block *qevent_block; struct flow_block_cb *block_cb; block_cb = flow_block_cb_lookup(f->block, mlxsw_sp_qevent_block_cb, mlxsw_sp); if (!block_cb) return; qevent_block = flow_block_cb_priv(block_cb); qevent_binding = mlxsw_sp_qevent_binding_lookup(qevent_block, mlxsw_sp_port, f->sch->handle, span_trigger); if (!qevent_binding) return; list_del(&qevent_binding->list); mlxsw_sp_qevent_binding_deconfigure(qevent_block, qevent_binding); mlxsw_sp_qevent_binding_destroy(qevent_binding); if (!flow_block_cb_decref(block_cb)) { flow_block_cb_remove(block_cb, f); list_del(&block_cb->driver_list); } } static int mlxsw_sp_setup_tc_block_qevent(struct mlxsw_sp_port *mlxsw_sp_port, struct flow_block_offload *f, enum mlxsw_sp_span_trigger span_trigger) { f->driver_block_list = &mlxsw_sp_qevent_block_cb_list; switch (f->command) { case FLOW_BLOCK_BIND: return mlxsw_sp_setup_tc_block_qevent_bind(mlxsw_sp_port, f, span_trigger); case FLOW_BLOCK_UNBIND: mlxsw_sp_setup_tc_block_qevent_unbind(mlxsw_sp_port, f, span_trigger); return 0; default: return -EOPNOTSUPP; } } int mlxsw_sp_setup_tc_block_qevent_early_drop(struct mlxsw_sp_port *mlxsw_sp_port, struct flow_block_offload *f) { return mlxsw_sp_setup_tc_block_qevent(mlxsw_sp_port, f, MLXSW_SP_SPAN_TRIGGER_EARLY_DROP); } int mlxsw_sp_tc_qdisc_init(struct mlxsw_sp_port *mlxsw_sp_port) { struct mlxsw_sp_qdisc_state *qdisc_state; qdisc_state = kzalloc(sizeof(*qdisc_state), GFP_KERNEL); if (!qdisc_state) return -ENOMEM; mutex_init(&qdisc_state->lock); qdisc_state->root_qdisc.prio_bitmap = 0xff; qdisc_state->root_qdisc.tclass_num = MLXSW_SP_PORT_DEFAULT_TCLASS; mlxsw_sp_port->qdisc = qdisc_state; return 0; } void mlxsw_sp_tc_qdisc_fini(struct mlxsw_sp_port *mlxsw_sp_port) { mutex_destroy(&mlxsw_sp_port->qdisc->lock); kfree(mlxsw_sp_port->qdisc); }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1