Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Jiri Pirko | 1330 | 61.35% | 19 | 38.78% |
Ido Schimmel | 760 | 35.06% | 22 | 44.90% |
Nir Dotan | 39 | 1.80% | 3 | 6.12% |
Arkadi Sharshevsky | 26 | 1.20% | 1 | 2.04% |
David Ahern | 6 | 0.28% | 1 | 2.04% |
Petr Machata | 3 | 0.14% | 1 | 2.04% |
Shalom Toledo | 2 | 0.09% | 1 | 2.04% |
Elad Raz | 2 | 0.09% | 1 | 2.04% |
Total | 2168 | 49 |
// SPDX-License-Identifier: BSD-3-Clause OR GPL-2.0 /* Copyright (c) 2017-2020 Mellanox Technologies. All rights reserved */ #include <linux/kernel.h> #include <linux/errno.h> #include <linux/netdevice.h> #include <net/flow_offload.h> #include "spectrum.h" #include "spectrum_span.h" #include "reg.h" static struct mlxsw_sp_mall_entry * mlxsw_sp_mall_entry_find(struct mlxsw_sp_flow_block *block, unsigned long cookie) { struct mlxsw_sp_mall_entry *mall_entry; list_for_each_entry(mall_entry, &block->mall.list, list) if (mall_entry->cookie == cookie) return mall_entry; return NULL; } static int mlxsw_sp_mall_port_mirror_add(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_mall_entry *mall_entry, struct netlink_ext_ack *extack) { struct mlxsw_sp *mlxsw_sp = mlxsw_sp_port->mlxsw_sp; struct mlxsw_sp_span_agent_parms agent_parms = {}; struct mlxsw_sp_span_trigger_parms parms; enum mlxsw_sp_span_trigger trigger; int err; if (!mall_entry->mirror.to_dev) { NL_SET_ERR_MSG(extack, "Could not find requested device"); return -EINVAL; } agent_parms.to_dev = mall_entry->mirror.to_dev; err = mlxsw_sp_span_agent_get(mlxsw_sp, &mall_entry->mirror.span_id, &agent_parms); if (err) { NL_SET_ERR_MSG(extack, "Failed to get SPAN agent"); return err; } err = mlxsw_sp_span_analyzed_port_get(mlxsw_sp_port, mall_entry->ingress); if (err) { NL_SET_ERR_MSG(extack, "Failed to get analyzed port"); goto err_analyzed_port_get; } trigger = mall_entry->ingress ? MLXSW_SP_SPAN_TRIGGER_INGRESS : MLXSW_SP_SPAN_TRIGGER_EGRESS; parms.span_id = mall_entry->mirror.span_id; parms.probability_rate = 1; err = mlxsw_sp_span_agent_bind(mlxsw_sp, trigger, mlxsw_sp_port, &parms); if (err) { NL_SET_ERR_MSG(extack, "Failed to bind SPAN agent"); goto err_agent_bind; } return 0; err_agent_bind: mlxsw_sp_span_analyzed_port_put(mlxsw_sp_port, mall_entry->ingress); err_analyzed_port_get: mlxsw_sp_span_agent_put(mlxsw_sp, mall_entry->mirror.span_id); return err; } static void mlxsw_sp_mall_port_mirror_del(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_mall_entry *mall_entry) { struct mlxsw_sp *mlxsw_sp = mlxsw_sp_port->mlxsw_sp; struct mlxsw_sp_span_trigger_parms parms; enum mlxsw_sp_span_trigger trigger; trigger = mall_entry->ingress ? MLXSW_SP_SPAN_TRIGGER_INGRESS : MLXSW_SP_SPAN_TRIGGER_EGRESS; parms.span_id = mall_entry->mirror.span_id; mlxsw_sp_span_agent_unbind(mlxsw_sp, trigger, mlxsw_sp_port, &parms); mlxsw_sp_span_analyzed_port_put(mlxsw_sp_port, mall_entry->ingress); mlxsw_sp_span_agent_put(mlxsw_sp, mall_entry->mirror.span_id); } static int mlxsw_sp_mall_port_sample_set(struct mlxsw_sp_port *mlxsw_sp_port, bool enable, u32 rate) { struct mlxsw_sp *mlxsw_sp = mlxsw_sp_port->mlxsw_sp; char mpsc_pl[MLXSW_REG_MPSC_LEN]; mlxsw_reg_mpsc_pack(mpsc_pl, mlxsw_sp_port->local_port, enable, rate); return mlxsw_reg_write(mlxsw_sp->core, MLXSW_REG(mpsc), mpsc_pl); } static int mlxsw_sp_mall_port_sample_add(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_mall_entry *mall_entry, struct netlink_ext_ack *extack) { struct mlxsw_sp *mlxsw_sp = mlxsw_sp_port->mlxsw_sp; struct mlxsw_sp_sample_trigger trigger; int err; if (mall_entry->ingress) trigger.type = MLXSW_SP_SAMPLE_TRIGGER_TYPE_INGRESS; else trigger.type = MLXSW_SP_SAMPLE_TRIGGER_TYPE_EGRESS; trigger.local_port = mlxsw_sp_port->local_port; err = mlxsw_sp_sample_trigger_params_set(mlxsw_sp, &trigger, &mall_entry->sample.params, extack); if (err) return err; err = mlxsw_sp->mall_ops->sample_add(mlxsw_sp, mlxsw_sp_port, mall_entry, extack); if (err) goto err_port_sample_set; return 0; err_port_sample_set: mlxsw_sp_sample_trigger_params_unset(mlxsw_sp, &trigger); return err; } static void mlxsw_sp_mall_port_sample_del(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_mall_entry *mall_entry) { struct mlxsw_sp *mlxsw_sp = mlxsw_sp_port->mlxsw_sp; struct mlxsw_sp_sample_trigger trigger; if (mall_entry->ingress) trigger.type = MLXSW_SP_SAMPLE_TRIGGER_TYPE_INGRESS; else trigger.type = MLXSW_SP_SAMPLE_TRIGGER_TYPE_EGRESS; trigger.local_port = mlxsw_sp_port->local_port; mlxsw_sp->mall_ops->sample_del(mlxsw_sp, mlxsw_sp_port, mall_entry); mlxsw_sp_sample_trigger_params_unset(mlxsw_sp, &trigger); } static int mlxsw_sp_mall_port_rule_add(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_mall_entry *mall_entry, struct netlink_ext_ack *extack) { switch (mall_entry->type) { case MLXSW_SP_MALL_ACTION_TYPE_MIRROR: return mlxsw_sp_mall_port_mirror_add(mlxsw_sp_port, mall_entry, extack); case MLXSW_SP_MALL_ACTION_TYPE_SAMPLE: return mlxsw_sp_mall_port_sample_add(mlxsw_sp_port, mall_entry, extack); default: WARN_ON(1); return -EINVAL; } } static void mlxsw_sp_mall_port_rule_del(struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_mall_entry *mall_entry) { switch (mall_entry->type) { case MLXSW_SP_MALL_ACTION_TYPE_MIRROR: mlxsw_sp_mall_port_mirror_del(mlxsw_sp_port, mall_entry); break; case MLXSW_SP_MALL_ACTION_TYPE_SAMPLE: mlxsw_sp_mall_port_sample_del(mlxsw_sp_port, mall_entry); break; default: WARN_ON(1); } } static void mlxsw_sp_mall_prio_update(struct mlxsw_sp_flow_block *block) { struct mlxsw_sp_mall_entry *mall_entry; if (list_empty(&block->mall.list)) return; block->mall.min_prio = UINT_MAX; block->mall.max_prio = 0; list_for_each_entry(mall_entry, &block->mall.list, list) { if (mall_entry->priority < block->mall.min_prio) block->mall.min_prio = mall_entry->priority; if (mall_entry->priority > block->mall.max_prio) block->mall.max_prio = mall_entry->priority; } } int mlxsw_sp_mall_replace(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_flow_block *block, struct tc_cls_matchall_offload *f) { struct mlxsw_sp_flow_block_binding *binding; struct mlxsw_sp_mall_entry *mall_entry; __be16 protocol = f->common.protocol; struct flow_action_entry *act; unsigned int flower_min_prio; unsigned int flower_max_prio; bool flower_prio_valid; int err; if (!flow_offload_has_one_action(&f->rule->action)) { NL_SET_ERR_MSG(f->common.extack, "Only singular actions are supported"); return -EOPNOTSUPP; } if (f->common.chain_index) { NL_SET_ERR_MSG(f->common.extack, "Only chain 0 is supported"); return -EOPNOTSUPP; } if (mlxsw_sp_flow_block_is_mixed_bound(block)) { NL_SET_ERR_MSG(f->common.extack, "Only not mixed bound blocks are supported"); return -EOPNOTSUPP; } err = mlxsw_sp_flower_prio_get(mlxsw_sp, block, f->common.chain_index, &flower_min_prio, &flower_max_prio); if (err) { if (err != -ENOENT) { NL_SET_ERR_MSG(f->common.extack, "Failed to get flower priorities"); return err; } flower_prio_valid = false; /* No flower filters are installed in specified chain. */ } else { flower_prio_valid = true; } if (protocol != htons(ETH_P_ALL)) { NL_SET_ERR_MSG(f->common.extack, "matchall rules only supported with 'all' protocol"); return -EOPNOTSUPP; } mall_entry = kzalloc(sizeof(*mall_entry), GFP_KERNEL); if (!mall_entry) return -ENOMEM; mall_entry->cookie = f->cookie; mall_entry->priority = f->common.prio; mall_entry->ingress = mlxsw_sp_flow_block_is_ingress_bound(block); if (flower_prio_valid && mall_entry->ingress && mall_entry->priority >= flower_min_prio) { NL_SET_ERR_MSG(f->common.extack, "Failed to add behind existing flower rules"); err = -EOPNOTSUPP; goto errout; } if (flower_prio_valid && !mall_entry->ingress && mall_entry->priority <= flower_max_prio) { NL_SET_ERR_MSG(f->common.extack, "Failed to add in front of existing flower rules"); err = -EOPNOTSUPP; goto errout; } act = &f->rule->action.entries[0]; switch (act->id) { case FLOW_ACTION_MIRRED: mall_entry->type = MLXSW_SP_MALL_ACTION_TYPE_MIRROR; mall_entry->mirror.to_dev = act->dev; break; case FLOW_ACTION_SAMPLE: mall_entry->type = MLXSW_SP_MALL_ACTION_TYPE_SAMPLE; mall_entry->sample.params.psample_group = act->sample.psample_group; mall_entry->sample.params.truncate = act->sample.truncate; mall_entry->sample.params.trunc_size = act->sample.trunc_size; mall_entry->sample.params.rate = act->sample.rate; break; default: err = -EOPNOTSUPP; goto errout; } list_for_each_entry(binding, &block->binding_list, list) { err = mlxsw_sp_mall_port_rule_add(binding->mlxsw_sp_port, mall_entry, f->common.extack); if (err) goto rollback; } block->rule_count++; if (mall_entry->ingress) block->egress_blocker_rule_count++; else block->ingress_blocker_rule_count++; list_add_tail(&mall_entry->list, &block->mall.list); mlxsw_sp_mall_prio_update(block); return 0; rollback: list_for_each_entry_continue_reverse(binding, &block->binding_list, list) mlxsw_sp_mall_port_rule_del(binding->mlxsw_sp_port, mall_entry); errout: kfree(mall_entry); return err; } void mlxsw_sp_mall_destroy(struct mlxsw_sp_flow_block *block, struct tc_cls_matchall_offload *f) { struct mlxsw_sp_flow_block_binding *binding; struct mlxsw_sp_mall_entry *mall_entry; mall_entry = mlxsw_sp_mall_entry_find(block, f->cookie); if (!mall_entry) { NL_SET_ERR_MSG(f->common.extack, "Entry not found"); return; } list_del(&mall_entry->list); if (mall_entry->ingress) block->egress_blocker_rule_count--; else block->ingress_blocker_rule_count--; block->rule_count--; list_for_each_entry(binding, &block->binding_list, list) mlxsw_sp_mall_port_rule_del(binding->mlxsw_sp_port, mall_entry); kfree_rcu(mall_entry, rcu); /* sample RX packets may be in-flight */ mlxsw_sp_mall_prio_update(block); } int mlxsw_sp_mall_port_bind(struct mlxsw_sp_flow_block *block, struct mlxsw_sp_port *mlxsw_sp_port, struct netlink_ext_ack *extack) { struct mlxsw_sp_mall_entry *mall_entry; int err; list_for_each_entry(mall_entry, &block->mall.list, list) { err = mlxsw_sp_mall_port_rule_add(mlxsw_sp_port, mall_entry, extack); if (err) goto rollback; } return 0; rollback: list_for_each_entry_continue_reverse(mall_entry, &block->mall.list, list) mlxsw_sp_mall_port_rule_del(mlxsw_sp_port, mall_entry); return err; } void mlxsw_sp_mall_port_unbind(struct mlxsw_sp_flow_block *block, struct mlxsw_sp_port *mlxsw_sp_port) { struct mlxsw_sp_mall_entry *mall_entry; list_for_each_entry(mall_entry, &block->mall.list, list) mlxsw_sp_mall_port_rule_del(mlxsw_sp_port, mall_entry); } int mlxsw_sp_mall_prio_get(struct mlxsw_sp_flow_block *block, u32 chain_index, unsigned int *p_min_prio, unsigned int *p_max_prio) { if (chain_index || list_empty(&block->mall.list)) /* In case there are no matchall rules, the caller * receives -ENOENT to indicate there is no need * to check the priorities. */ return -ENOENT; *p_min_prio = block->mall.min_prio; *p_max_prio = block->mall.max_prio; return 0; } static int mlxsw_sp1_mall_sample_add(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_mall_entry *mall_entry, struct netlink_ext_ack *extack) { u32 rate = mall_entry->sample.params.rate; if (!mall_entry->ingress) { NL_SET_ERR_MSG(extack, "Sampling is not supported on egress"); return -EOPNOTSUPP; } if (rate > MLXSW_REG_MPSC_RATE_MAX) { NL_SET_ERR_MSG(extack, "Unsupported sampling rate"); return -EOPNOTSUPP; } return mlxsw_sp_mall_port_sample_set(mlxsw_sp_port, true, rate); } static void mlxsw_sp1_mall_sample_del(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_mall_entry *mall_entry) { mlxsw_sp_mall_port_sample_set(mlxsw_sp_port, false, 1); } const struct mlxsw_sp_mall_ops mlxsw_sp1_mall_ops = { .sample_add = mlxsw_sp1_mall_sample_add, .sample_del = mlxsw_sp1_mall_sample_del, }; static int mlxsw_sp2_mall_sample_add(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_mall_entry *mall_entry, struct netlink_ext_ack *extack) { struct mlxsw_sp_span_trigger_parms trigger_parms = {}; struct mlxsw_sp_span_agent_parms agent_parms = { .to_dev = NULL, /* Mirror to CPU. */ .session_id = MLXSW_SP_SPAN_SESSION_ID_SAMPLING, }; u32 rate = mall_entry->sample.params.rate; enum mlxsw_sp_span_trigger span_trigger; int err; err = mlxsw_sp_span_agent_get(mlxsw_sp, &mall_entry->sample.span_id, &agent_parms); if (err) { NL_SET_ERR_MSG(extack, "Failed to get SPAN agent"); return err; } err = mlxsw_sp_span_analyzed_port_get(mlxsw_sp_port, mall_entry->ingress); if (err) { NL_SET_ERR_MSG(extack, "Failed to get analyzed port"); goto err_analyzed_port_get; } span_trigger = mall_entry->ingress ? MLXSW_SP_SPAN_TRIGGER_INGRESS : MLXSW_SP_SPAN_TRIGGER_EGRESS; trigger_parms.span_id = mall_entry->sample.span_id; trigger_parms.probability_rate = rate; err = mlxsw_sp_span_agent_bind(mlxsw_sp, span_trigger, mlxsw_sp_port, &trigger_parms); if (err) { NL_SET_ERR_MSG(extack, "Failed to bind SPAN agent"); goto err_agent_bind; } return 0; err_agent_bind: mlxsw_sp_span_analyzed_port_put(mlxsw_sp_port, mall_entry->ingress); err_analyzed_port_get: mlxsw_sp_span_agent_put(mlxsw_sp, mall_entry->sample.span_id); return err; } static void mlxsw_sp2_mall_sample_del(struct mlxsw_sp *mlxsw_sp, struct mlxsw_sp_port *mlxsw_sp_port, struct mlxsw_sp_mall_entry *mall_entry) { struct mlxsw_sp_span_trigger_parms trigger_parms = {}; enum mlxsw_sp_span_trigger span_trigger; span_trigger = mall_entry->ingress ? MLXSW_SP_SPAN_TRIGGER_INGRESS : MLXSW_SP_SPAN_TRIGGER_EGRESS; trigger_parms.span_id = mall_entry->sample.span_id; mlxsw_sp_span_agent_unbind(mlxsw_sp, span_trigger, mlxsw_sp_port, &trigger_parms); mlxsw_sp_span_analyzed_port_put(mlxsw_sp_port, mall_entry->ingress); mlxsw_sp_span_agent_put(mlxsw_sp, mall_entry->sample.span_id); } const struct mlxsw_sp_mall_ops mlxsw_sp2_mall_ops = { .sample_add = mlxsw_sp2_mall_sample_add, .sample_del = mlxsw_sp2_mall_sample_del, };
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1