Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Felix Fietkau | 2504 | 85.37% | 11 | 31.43% |
Lorenzo Bianconi | 147 | 5.01% | 7 | 20.00% |
David Bentham | 100 | 3.41% | 1 | 2.86% |
Daniel Golle | 42 | 1.43% | 1 | 2.86% |
Pablo Neira Ayuso | 42 | 1.43% | 2 | 5.71% |
John Crispin | 36 | 1.23% | 1 | 2.86% |
Sujuan Chen | 22 | 0.75% | 2 | 5.71% |
Asbjörn Sloth Tönnesen | 21 | 0.72% | 1 | 2.86% |
Dan Carpenter | 6 | 0.20% | 1 | 2.86% |
Linus Torvalds (pre-git) | 5 | 0.17% | 2 | 5.71% |
Sean Wang | 2 | 0.07% | 1 | 2.86% |
Vladimir Oltean | 2 | 0.07% | 1 | 2.86% |
Joe Perches | 1 | 0.03% | 1 | 2.86% |
Thomas Gleixner | 1 | 0.03% | 1 | 2.86% |
Florian Fainelli | 1 | 0.03% | 1 | 2.86% |
DENG Qingfang | 1 | 0.03% | 1 | 2.86% |
Total | 2933 | 35 |
// SPDX-License-Identifier: GPL-2.0-only /* * Copyright (C) 2020 Felix Fietkau <nbd@nbd.name> */ #include <linux/if_ether.h> #include <linux/rhashtable.h> #include <linux/ip.h> #include <linux/ipv6.h> #include <net/flow_offload.h> #include <net/pkt_cls.h> #include <net/dsa.h> #include "mtk_eth_soc.h" #include "mtk_wed.h" struct mtk_flow_data { struct ethhdr eth; union { struct { __be32 src_addr; __be32 dst_addr; } v4; struct { struct in6_addr src_addr; struct in6_addr dst_addr; } v6; }; __be16 src_port; __be16 dst_port; u16 vlan_in; struct { u16 id; __be16 proto; u8 num; } vlan; struct { u16 sid; u8 num; } pppoe; }; static const struct rhashtable_params mtk_flow_ht_params = { .head_offset = offsetof(struct mtk_flow_entry, node), .key_offset = offsetof(struct mtk_flow_entry, cookie), .key_len = sizeof(unsigned long), .automatic_shrinking = true, }; static int mtk_flow_set_ipv4_addr(struct mtk_eth *eth, struct mtk_foe_entry *foe, struct mtk_flow_data *data, bool egress) { return mtk_foe_entry_set_ipv4_tuple(eth, foe, egress, data->v4.src_addr, data->src_port, data->v4.dst_addr, data->dst_port); } static int mtk_flow_set_ipv6_addr(struct mtk_eth *eth, struct mtk_foe_entry *foe, struct mtk_flow_data *data) { return mtk_foe_entry_set_ipv6_tuple(eth, foe, data->v6.src_addr.s6_addr32, data->src_port, data->v6.dst_addr.s6_addr32, data->dst_port); } static void mtk_flow_offload_mangle_eth(const struct flow_action_entry *act, void *eth) { void *dest = eth + act->mangle.offset; const void *src = &act->mangle.val; if (act->mangle.offset > 8) return; if (act->mangle.mask == 0xffff) { src += 2; dest += 2; } memcpy(dest, src, act->mangle.mask ? 2 : 4); } static int mtk_flow_get_wdma_info(struct net_device *dev, const u8 *addr, struct mtk_wdma_info *info) { struct net_device_path_stack stack; struct net_device_path *path; int err; if (!dev) return -ENODEV; if (!IS_ENABLED(CONFIG_NET_MEDIATEK_SOC_WED)) return -1; err = dev_fill_forward_path(dev, addr, &stack); if (err) return err; path = &stack.path[stack.num_paths - 1]; if (path->type != DEV_PATH_MTK_WDMA) return -1; info->wdma_idx = path->mtk_wdma.wdma_idx; info->queue = path->mtk_wdma.queue; info->bss = path->mtk_wdma.bss; info->wcid = path->mtk_wdma.wcid; info->amsdu = path->mtk_wdma.amsdu; return 0; } static int mtk_flow_mangle_ports(const struct flow_action_entry *act, struct mtk_flow_data *data) { u32 val = ntohl(act->mangle.val); switch (act->mangle.offset) { case 0: if (act->mangle.mask == ~htonl(0xffff)) data->dst_port = cpu_to_be16(val); else data->src_port = cpu_to_be16(val >> 16); break; case 2: data->dst_port = cpu_to_be16(val); break; default: return -EINVAL; } return 0; } static int mtk_flow_mangle_ipv4(const struct flow_action_entry *act, struct mtk_flow_data *data) { __be32 *dest; switch (act->mangle.offset) { case offsetof(struct iphdr, saddr): dest = &data->v4.src_addr; break; case offsetof(struct iphdr, daddr): dest = &data->v4.dst_addr; break; default: return -EINVAL; } memcpy(dest, &act->mangle.val, sizeof(u32)); return 0; } static int mtk_flow_get_dsa_port(struct net_device **dev) { #if IS_ENABLED(CONFIG_NET_DSA) struct dsa_port *dp; dp = dsa_port_from_netdev(*dev); if (IS_ERR(dp)) return -ENODEV; if (dp->cpu_dp->tag_ops->proto != DSA_TAG_PROTO_MTK) return -ENODEV; *dev = dsa_port_to_conduit(dp); return dp->index; #else return -ENODEV; #endif } static int mtk_flow_set_output_device(struct mtk_eth *eth, struct mtk_foe_entry *foe, struct net_device *dev, const u8 *dest_mac, int *wed_index) { struct mtk_wdma_info info = {}; int pse_port, dsa_port, queue; if (mtk_flow_get_wdma_info(dev, dest_mac, &info) == 0) { mtk_foe_entry_set_wdma(eth, foe, info.wdma_idx, info.queue, info.bss, info.wcid, info.amsdu); if (mtk_is_netsys_v2_or_greater(eth)) { switch (info.wdma_idx) { case 0: pse_port = PSE_WDMA0_PORT; break; case 1: pse_port = PSE_WDMA1_PORT; break; case 2: pse_port = PSE_WDMA2_PORT; break; default: return -EINVAL; } } else { pse_port = 3; } *wed_index = info.wdma_idx; goto out; } dsa_port = mtk_flow_get_dsa_port(&dev); if (dev == eth->netdev[0]) pse_port = PSE_GDM1_PORT; else if (dev == eth->netdev[1]) pse_port = PSE_GDM2_PORT; else if (dev == eth->netdev[2]) pse_port = PSE_GDM3_PORT; else return -EOPNOTSUPP; if (dsa_port >= 0) { mtk_foe_entry_set_dsa(eth, foe, dsa_port); queue = 3 + dsa_port; } else { queue = pse_port - 1; } mtk_foe_entry_set_queue(eth, foe, queue); out: mtk_foe_entry_set_pse_port(eth, foe, pse_port); return 0; } static int mtk_flow_offload_replace(struct mtk_eth *eth, struct flow_cls_offload *f, int ppe_index) { struct flow_rule *rule = flow_cls_offload_flow_rule(f); struct flow_action_entry *act; struct mtk_flow_data data = {}; struct mtk_foe_entry foe; struct net_device *odev = NULL; struct mtk_flow_entry *entry; int offload_type = 0; int wed_index = -1; u16 addr_type = 0; u8 l4proto = 0; int err = 0; int i; if (rhashtable_lookup(ð->flow_table, &f->cookie, mtk_flow_ht_params)) return -EEXIST; if (flow_rule_match_key(rule, FLOW_DISSECTOR_KEY_META)) { struct flow_match_meta match; flow_rule_match_meta(rule, &match); } else { return -EOPNOTSUPP; } if (flow_rule_match_key(rule, FLOW_DISSECTOR_KEY_CONTROL)) { struct flow_match_control match; flow_rule_match_control(rule, &match); addr_type = match.key->addr_type; if (flow_rule_has_control_flags(match.mask->flags, f->common.extack)) return -EOPNOTSUPP; } else { return -EOPNOTSUPP; } if (flow_rule_match_key(rule, FLOW_DISSECTOR_KEY_BASIC)) { struct flow_match_basic match; flow_rule_match_basic(rule, &match); l4proto = match.key->ip_proto; } else { return -EOPNOTSUPP; } switch (addr_type) { case 0: offload_type = MTK_PPE_PKT_TYPE_BRIDGE; if (flow_rule_match_key(rule, FLOW_DISSECTOR_KEY_ETH_ADDRS)) { struct flow_match_eth_addrs match; flow_rule_match_eth_addrs(rule, &match); memcpy(data.eth.h_dest, match.key->dst, ETH_ALEN); memcpy(data.eth.h_source, match.key->src, ETH_ALEN); } else { return -EOPNOTSUPP; } if (flow_rule_match_key(rule, FLOW_DISSECTOR_KEY_VLAN)) { struct flow_match_vlan match; flow_rule_match_vlan(rule, &match); if (match.key->vlan_tpid != cpu_to_be16(ETH_P_8021Q)) return -EOPNOTSUPP; data.vlan_in = match.key->vlan_id; } break; case FLOW_DISSECTOR_KEY_IPV4_ADDRS: offload_type = MTK_PPE_PKT_TYPE_IPV4_HNAPT; break; case FLOW_DISSECTOR_KEY_IPV6_ADDRS: offload_type = MTK_PPE_PKT_TYPE_IPV6_ROUTE_5T; break; default: return -EOPNOTSUPP; } flow_action_for_each(i, act, &rule->action) { switch (act->id) { case FLOW_ACTION_MANGLE: if (offload_type == MTK_PPE_PKT_TYPE_BRIDGE) return -EOPNOTSUPP; if (act->mangle.htype == FLOW_ACT_MANGLE_HDR_TYPE_ETH) mtk_flow_offload_mangle_eth(act, &data.eth); break; case FLOW_ACTION_REDIRECT: odev = act->dev; break; case FLOW_ACTION_CSUM: break; case FLOW_ACTION_VLAN_PUSH: if (data.vlan.num == 1 || act->vlan.proto != htons(ETH_P_8021Q)) return -EOPNOTSUPP; data.vlan.id = act->vlan.vid; data.vlan.proto = act->vlan.proto; data.vlan.num++; break; case FLOW_ACTION_VLAN_POP: break; case FLOW_ACTION_PPPOE_PUSH: if (data.pppoe.num == 1) return -EOPNOTSUPP; data.pppoe.sid = act->pppoe.sid; data.pppoe.num++; break; default: return -EOPNOTSUPP; } } if (!is_valid_ether_addr(data.eth.h_source) || !is_valid_ether_addr(data.eth.h_dest)) return -EINVAL; err = mtk_foe_entry_prepare(eth, &foe, offload_type, l4proto, 0, data.eth.h_source, data.eth.h_dest); if (err) return err; if (flow_rule_match_key(rule, FLOW_DISSECTOR_KEY_PORTS)) { struct flow_match_ports ports; if (offload_type == MTK_PPE_PKT_TYPE_BRIDGE) return -EOPNOTSUPP; flow_rule_match_ports(rule, &ports); data.src_port = ports.key->src; data.dst_port = ports.key->dst; } else if (offload_type != MTK_PPE_PKT_TYPE_BRIDGE) { return -EOPNOTSUPP; } if (addr_type == FLOW_DISSECTOR_KEY_IPV4_ADDRS) { struct flow_match_ipv4_addrs addrs; flow_rule_match_ipv4_addrs(rule, &addrs); data.v4.src_addr = addrs.key->src; data.v4.dst_addr = addrs.key->dst; mtk_flow_set_ipv4_addr(eth, &foe, &data, false); } if (addr_type == FLOW_DISSECTOR_KEY_IPV6_ADDRS) { struct flow_match_ipv6_addrs addrs; flow_rule_match_ipv6_addrs(rule, &addrs); data.v6.src_addr = addrs.key->src; data.v6.dst_addr = addrs.key->dst; mtk_flow_set_ipv6_addr(eth, &foe, &data); } flow_action_for_each(i, act, &rule->action) { if (act->id != FLOW_ACTION_MANGLE) continue; if (offload_type == MTK_PPE_PKT_TYPE_BRIDGE) return -EOPNOTSUPP; switch (act->mangle.htype) { case FLOW_ACT_MANGLE_HDR_TYPE_TCP: case FLOW_ACT_MANGLE_HDR_TYPE_UDP: err = mtk_flow_mangle_ports(act, &data); break; case FLOW_ACT_MANGLE_HDR_TYPE_IP4: err = mtk_flow_mangle_ipv4(act, &data); break; case FLOW_ACT_MANGLE_HDR_TYPE_ETH: /* handled earlier */ break; default: return -EOPNOTSUPP; } if (err) return err; } if (addr_type == FLOW_DISSECTOR_KEY_IPV4_ADDRS) { err = mtk_flow_set_ipv4_addr(eth, &foe, &data, true); if (err) return err; } if (offload_type == MTK_PPE_PKT_TYPE_BRIDGE) foe.bridge.vlan = data.vlan_in; if (data.vlan.num == 1) { if (data.vlan.proto != htons(ETH_P_8021Q)) return -EOPNOTSUPP; mtk_foe_entry_set_vlan(eth, &foe, data.vlan.id); } if (data.pppoe.num == 1) mtk_foe_entry_set_pppoe(eth, &foe, data.pppoe.sid); err = mtk_flow_set_output_device(eth, &foe, odev, data.eth.h_dest, &wed_index); if (err) return err; if (wed_index >= 0 && (err = mtk_wed_flow_add(wed_index)) < 0) return err; entry = kzalloc(sizeof(*entry), GFP_KERNEL); if (!entry) return -ENOMEM; entry->cookie = f->cookie; memcpy(&entry->data, &foe, sizeof(entry->data)); entry->wed_index = wed_index; entry->ppe_index = ppe_index; err = mtk_foe_entry_commit(eth->ppe[entry->ppe_index], entry); if (err < 0) goto free; err = rhashtable_insert_fast(ð->flow_table, &entry->node, mtk_flow_ht_params); if (err < 0) goto clear; return 0; clear: mtk_foe_entry_clear(eth->ppe[entry->ppe_index], entry); free: kfree(entry); if (wed_index >= 0) mtk_wed_flow_remove(wed_index); return err; } static int mtk_flow_offload_destroy(struct mtk_eth *eth, struct flow_cls_offload *f) { struct mtk_flow_entry *entry; entry = rhashtable_lookup(ð->flow_table, &f->cookie, mtk_flow_ht_params); if (!entry) return -ENOENT; mtk_foe_entry_clear(eth->ppe[entry->ppe_index], entry); rhashtable_remove_fast(ð->flow_table, &entry->node, mtk_flow_ht_params); if (entry->wed_index >= 0) mtk_wed_flow_remove(entry->wed_index); kfree(entry); return 0; } static int mtk_flow_offload_stats(struct mtk_eth *eth, struct flow_cls_offload *f) { struct mtk_flow_entry *entry; struct mtk_foe_accounting diff; u32 idle; entry = rhashtable_lookup(ð->flow_table, &f->cookie, mtk_flow_ht_params); if (!entry) return -ENOENT; idle = mtk_foe_entry_idle_time(eth->ppe[entry->ppe_index], entry); f->stats.lastused = jiffies - idle * HZ; if (entry->hash != 0xFFFF && mtk_foe_entry_get_mib(eth->ppe[entry->ppe_index], entry->hash, &diff)) { f->stats.pkts += diff.packets; f->stats.bytes += diff.bytes; } return 0; } static DEFINE_MUTEX(mtk_flow_offload_mutex); int mtk_flow_offload_cmd(struct mtk_eth *eth, struct flow_cls_offload *cls, int ppe_index) { int err; mutex_lock(&mtk_flow_offload_mutex); switch (cls->command) { case FLOW_CLS_REPLACE: err = mtk_flow_offload_replace(eth, cls, ppe_index); break; case FLOW_CLS_DESTROY: err = mtk_flow_offload_destroy(eth, cls); break; case FLOW_CLS_STATS: err = mtk_flow_offload_stats(eth, cls); break; default: err = -EOPNOTSUPP; break; } mutex_unlock(&mtk_flow_offload_mutex); return err; } static int mtk_eth_setup_tc_block_cb(enum tc_setup_type type, void *type_data, void *cb_priv) { struct flow_cls_offload *cls = type_data; struct net_device *dev = cb_priv; struct mtk_mac *mac; struct mtk_eth *eth; mac = netdev_priv(dev); eth = mac->hw; if (!tc_can_offload(dev)) return -EOPNOTSUPP; if (type != TC_SETUP_CLSFLOWER) return -EOPNOTSUPP; return mtk_flow_offload_cmd(eth, cls, 0); } static int mtk_eth_setup_tc_block(struct net_device *dev, struct flow_block_offload *f) { struct mtk_mac *mac = netdev_priv(dev); struct mtk_eth *eth = mac->hw; static LIST_HEAD(block_cb_list); struct flow_block_cb *block_cb; flow_setup_cb_t *cb; if (!eth->soc->offload_version) return -EOPNOTSUPP; if (f->binder_type != FLOW_BLOCK_BINDER_TYPE_CLSACT_INGRESS) return -EOPNOTSUPP; cb = mtk_eth_setup_tc_block_cb; f->driver_block_list = &block_cb_list; switch (f->command) { case FLOW_BLOCK_BIND: block_cb = flow_block_cb_lookup(f->block, cb, dev); if (block_cb) { flow_block_cb_incref(block_cb); return 0; } block_cb = flow_block_cb_alloc(cb, dev, dev, NULL); if (IS_ERR(block_cb)) return PTR_ERR(block_cb); flow_block_cb_incref(block_cb); flow_block_cb_add(block_cb, f); list_add_tail(&block_cb->driver_list, &block_cb_list); return 0; case FLOW_BLOCK_UNBIND: block_cb = flow_block_cb_lookup(f->block, cb, dev); if (!block_cb) return -ENOENT; if (!flow_block_cb_decref(block_cb)) { flow_block_cb_remove(block_cb, f); list_del(&block_cb->driver_list); } return 0; default: return -EOPNOTSUPP; } } int mtk_eth_setup_tc(struct net_device *dev, enum tc_setup_type type, void *type_data) { switch (type) { case TC_SETUP_BLOCK: case TC_SETUP_FT: return mtk_eth_setup_tc_block(dev, type_data); default: return -EOPNOTSUPP; } } int mtk_eth_offload_init(struct mtk_eth *eth) { return rhashtable_init(ð->flow_table, &mtk_flow_ht_params); }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1