Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Steen Hegelund | 1807 | 53.15% | 1 | 6.25% |
Casper Andersson | 1492 | 43.88% | 10 | 62.50% |
Vladimir Oltean | 92 | 2.71% | 3 | 18.75% |
Yang Yingliang | 7 | 0.21% | 1 | 6.25% |
Tobias Waldekranz | 2 | 0.06% | 1 | 6.25% |
Total | 3400 | 16 |
// SPDX-License-Identifier: GPL-2.0+ /* Microchip Sparx5 Switch driver * * Copyright (c) 2021 Microchip Technology Inc. and its subsidiaries. */ #include <linux/if_bridge.h> #include <net/switchdev.h> #include "sparx5_main_regs.h" #include "sparx5_main.h" static struct workqueue_struct *sparx5_owq; struct sparx5_switchdev_event_work { struct work_struct work; struct switchdev_notifier_fdb_info fdb_info; struct net_device *dev; struct sparx5 *sparx5; unsigned long event; }; static int sparx5_port_attr_pre_bridge_flags(struct sparx5_port *port, struct switchdev_brport_flags flags) { if (flags.mask & ~(BR_FLOOD | BR_MCAST_FLOOD | BR_BCAST_FLOOD)) return -EINVAL; return 0; } static void sparx5_port_update_mcast_ip_flood(struct sparx5_port *port, bool flood_flag) { bool should_flood = flood_flag || port->is_mrouter; int pgid; for (pgid = PGID_IPV4_MC_DATA; pgid <= PGID_IPV6_MC_CTRL; pgid++) sparx5_pgid_update_mask(port, pgid, should_flood); } static void sparx5_port_attr_bridge_flags(struct sparx5_port *port, struct switchdev_brport_flags flags) { if (flags.mask & BR_MCAST_FLOOD) { sparx5_pgid_update_mask(port, PGID_MC_FLOOD, !!(flags.val & BR_MCAST_FLOOD)); sparx5_port_update_mcast_ip_flood(port, !!(flags.val & BR_MCAST_FLOOD)); } if (flags.mask & BR_FLOOD) sparx5_pgid_update_mask(port, PGID_UC_FLOOD, !!(flags.val & BR_FLOOD)); if (flags.mask & BR_BCAST_FLOOD) sparx5_pgid_update_mask(port, PGID_BCAST, !!(flags.val & BR_BCAST_FLOOD)); } static void sparx5_attr_stp_state_set(struct sparx5_port *port, u8 state) { struct sparx5 *sparx5 = port->sparx5; if (!test_bit(port->portno, sparx5->bridge_mask)) { netdev_err(port->ndev, "Controlling non-bridged port %d?\n", port->portno); return; } switch (state) { case BR_STATE_FORWARDING: set_bit(port->portno, sparx5->bridge_fwd_mask); fallthrough; case BR_STATE_LEARNING: set_bit(port->portno, sparx5->bridge_lrn_mask); break; default: /* All other states treated as blocking */ clear_bit(port->portno, sparx5->bridge_fwd_mask); clear_bit(port->portno, sparx5->bridge_lrn_mask); break; } /* apply the bridge_fwd_mask to all the ports */ sparx5_update_fwd(sparx5); } static void sparx5_port_attr_ageing_set(struct sparx5_port *port, unsigned long ageing_clock_t) { unsigned long ageing_jiffies = clock_t_to_jiffies(ageing_clock_t); u32 ageing_time = jiffies_to_msecs(ageing_jiffies); sparx5_set_ageing(port->sparx5, ageing_time); } static void sparx5_port_attr_mrouter_set(struct sparx5_port *port, struct net_device *orig_dev, bool enable) { struct sparx5 *sparx5 = port->sparx5; struct sparx5_mdb_entry *e; bool flood_flag; if ((enable && port->is_mrouter) || (!enable && !port->is_mrouter)) return; /* Add/del mrouter port on all active mdb entries in HW. * Don't change entry port mask, since that represents * ports that actually joined that group. */ mutex_lock(&sparx5->mdb_lock); list_for_each_entry(e, &sparx5->mdb_entries, list) { if (!test_bit(port->portno, e->port_mask) && ether_addr_is_ip_mcast(e->addr)) sparx5_pgid_update_mask(port, e->pgid_idx, enable); } mutex_unlock(&sparx5->mdb_lock); /* Enable/disable flooding depending on if port is mrouter port * or if mcast flood is enabled. */ port->is_mrouter = enable; flood_flag = br_port_flag_is_set(port->ndev, BR_MCAST_FLOOD); sparx5_port_update_mcast_ip_flood(port, flood_flag); } static int sparx5_port_attr_set(struct net_device *dev, const void *ctx, const struct switchdev_attr *attr, struct netlink_ext_ack *extack) { struct sparx5_port *port = netdev_priv(dev); switch (attr->id) { case SWITCHDEV_ATTR_ID_PORT_PRE_BRIDGE_FLAGS: return sparx5_port_attr_pre_bridge_flags(port, attr->u.brport_flags); case SWITCHDEV_ATTR_ID_PORT_BRIDGE_FLAGS: sparx5_port_attr_bridge_flags(port, attr->u.brport_flags); break; case SWITCHDEV_ATTR_ID_PORT_STP_STATE: sparx5_attr_stp_state_set(port, attr->u.stp_state); break; case SWITCHDEV_ATTR_ID_BRIDGE_AGEING_TIME: sparx5_port_attr_ageing_set(port, attr->u.ageing_time); break; case SWITCHDEV_ATTR_ID_BRIDGE_VLAN_FILTERING: /* Used PVID 1 when default_pvid is 0, to avoid * collision with non-bridged ports. */ if (port->pvid == 0) port->pvid = 1; port->vlan_aware = attr->u.vlan_filtering; sparx5_vlan_port_apply(port->sparx5, port); break; case SWITCHDEV_ATTR_ID_PORT_MROUTER: sparx5_port_attr_mrouter_set(port, attr->orig_dev, attr->u.mrouter); break; default: return -EOPNOTSUPP; } return 0; } static int sparx5_port_bridge_join(struct sparx5_port *port, struct net_device *bridge, struct netlink_ext_ack *extack) { struct sparx5 *sparx5 = port->sparx5; struct net_device *ndev = port->ndev; int err; if (bitmap_empty(sparx5->bridge_mask, SPX5_PORTS)) /* First bridged port */ sparx5->hw_bridge_dev = bridge; else if (sparx5->hw_bridge_dev != bridge) /* This is adding the port to a second bridge, this is * unsupported */ return -ENODEV; set_bit(port->portno, sparx5->bridge_mask); err = switchdev_bridge_port_offload(ndev, ndev, NULL, NULL, NULL, false, extack); if (err) goto err_switchdev_offload; /* Remove standalone port entry */ sparx5_mact_forget(sparx5, ndev->dev_addr, 0); /* Port enters in bridge mode therefor don't need to copy to CPU * frames for multicast in case the bridge is not requesting them */ __dev_mc_unsync(ndev, sparx5_mc_unsync); return 0; err_switchdev_offload: clear_bit(port->portno, sparx5->bridge_mask); return err; } static void sparx5_port_bridge_leave(struct sparx5_port *port, struct net_device *bridge) { struct sparx5 *sparx5 = port->sparx5; switchdev_bridge_port_unoffload(port->ndev, NULL, NULL, NULL); clear_bit(port->portno, sparx5->bridge_mask); if (bitmap_empty(sparx5->bridge_mask, SPX5_PORTS)) sparx5->hw_bridge_dev = NULL; /* Clear bridge vlan settings before updating the port settings */ port->vlan_aware = 0; port->pvid = NULL_VID; port->vid = NULL_VID; /* Forward frames to CPU */ sparx5_mact_learn(sparx5, PGID_CPU, port->ndev->dev_addr, 0); /* Port enters in host more therefore restore mc list */ __dev_mc_sync(port->ndev, sparx5_mc_sync, sparx5_mc_unsync); } static int sparx5_port_changeupper(struct net_device *dev, struct netdev_notifier_changeupper_info *info) { struct sparx5_port *port = netdev_priv(dev); struct netlink_ext_ack *extack; int err = 0; extack = netdev_notifier_info_to_extack(&info->info); if (netif_is_bridge_master(info->upper_dev)) { if (info->linking) err = sparx5_port_bridge_join(port, info->upper_dev, extack); else sparx5_port_bridge_leave(port, info->upper_dev); sparx5_vlan_port_apply(port->sparx5, port); } return err; } static int sparx5_port_add_addr(struct net_device *dev, bool up) { struct sparx5_port *port = netdev_priv(dev); struct sparx5 *sparx5 = port->sparx5; u16 vid = port->pvid; if (up) sparx5_mact_learn(sparx5, PGID_CPU, port->ndev->dev_addr, vid); else sparx5_mact_forget(sparx5, port->ndev->dev_addr, vid); return 0; } static int sparx5_netdevice_port_event(struct net_device *dev, struct notifier_block *nb, unsigned long event, void *ptr) { int err = 0; if (!sparx5_netdevice_check(dev)) return 0; switch (event) { case NETDEV_CHANGEUPPER: err = sparx5_port_changeupper(dev, ptr); break; case NETDEV_PRE_UP: err = sparx5_port_add_addr(dev, true); break; case NETDEV_DOWN: err = sparx5_port_add_addr(dev, false); break; } return err; } static int sparx5_netdevice_event(struct notifier_block *nb, unsigned long event, void *ptr) { struct net_device *dev = netdev_notifier_info_to_dev(ptr); int ret = 0; ret = sparx5_netdevice_port_event(dev, nb, event, ptr); return notifier_from_errno(ret); } static void sparx5_switchdev_bridge_fdb_event_work(struct work_struct *work) { struct sparx5_switchdev_event_work *switchdev_work = container_of(work, struct sparx5_switchdev_event_work, work); struct net_device *dev = switchdev_work->dev; struct switchdev_notifier_fdb_info *fdb_info; struct sparx5_port *port; struct sparx5 *sparx5; bool host_addr; u16 vid; rtnl_lock(); if (!sparx5_netdevice_check(dev)) { host_addr = true; sparx5 = switchdev_work->sparx5; } else { host_addr = false; sparx5 = switchdev_work->sparx5; port = netdev_priv(dev); } fdb_info = &switchdev_work->fdb_info; /* Used PVID 1 when default_pvid is 0, to avoid * collision with non-bridged ports. */ if (fdb_info->vid == 0) vid = 1; else vid = fdb_info->vid; switch (switchdev_work->event) { case SWITCHDEV_FDB_ADD_TO_DEVICE: if (host_addr) sparx5_add_mact_entry(sparx5, dev, PGID_CPU, fdb_info->addr, vid); else sparx5_add_mact_entry(sparx5, port->ndev, port->portno, fdb_info->addr, vid); break; case SWITCHDEV_FDB_DEL_TO_DEVICE: sparx5_del_mact_entry(sparx5, fdb_info->addr, vid); break; } rtnl_unlock(); kfree(switchdev_work->fdb_info.addr); kfree(switchdev_work); dev_put(dev); } static void sparx5_schedule_work(struct work_struct *work) { queue_work(sparx5_owq, work); } static int sparx5_switchdev_event(struct notifier_block *nb, unsigned long event, void *ptr) { struct net_device *dev = switchdev_notifier_info_to_dev(ptr); struct sparx5_switchdev_event_work *switchdev_work; struct switchdev_notifier_fdb_info *fdb_info; struct switchdev_notifier_info *info = ptr; struct sparx5 *spx5; int err; spx5 = container_of(nb, struct sparx5, switchdev_nb); switch (event) { case SWITCHDEV_PORT_ATTR_SET: err = switchdev_handle_port_attr_set(dev, ptr, sparx5_netdevice_check, sparx5_port_attr_set); return notifier_from_errno(err); case SWITCHDEV_FDB_ADD_TO_DEVICE: fallthrough; case SWITCHDEV_FDB_DEL_TO_DEVICE: switchdev_work = kzalloc(sizeof(*switchdev_work), GFP_ATOMIC); if (!switchdev_work) return NOTIFY_BAD; switchdev_work->dev = dev; switchdev_work->event = event; switchdev_work->sparx5 = spx5; fdb_info = container_of(info, struct switchdev_notifier_fdb_info, info); INIT_WORK(&switchdev_work->work, sparx5_switchdev_bridge_fdb_event_work); memcpy(&switchdev_work->fdb_info, ptr, sizeof(switchdev_work->fdb_info)); switchdev_work->fdb_info.addr = kzalloc(ETH_ALEN, GFP_ATOMIC); if (!switchdev_work->fdb_info.addr) goto err_addr_alloc; ether_addr_copy((u8 *)switchdev_work->fdb_info.addr, fdb_info->addr); dev_hold(dev); sparx5_schedule_work(&switchdev_work->work); break; } return NOTIFY_DONE; err_addr_alloc: kfree(switchdev_work); return NOTIFY_BAD; } static int sparx5_handle_port_vlan_add(struct net_device *dev, struct notifier_block *nb, const struct switchdev_obj_port_vlan *v) { struct sparx5_port *port = netdev_priv(dev); if (netif_is_bridge_master(dev)) { struct sparx5 *sparx5 = container_of(nb, struct sparx5, switchdev_blocking_nb); /* Flood broadcast to CPU */ sparx5_mact_learn(sparx5, PGID_BCAST, dev->broadcast, v->vid); return 0; } if (!sparx5_netdevice_check(dev)) return -EOPNOTSUPP; return sparx5_vlan_vid_add(port, v->vid, v->flags & BRIDGE_VLAN_INFO_PVID, v->flags & BRIDGE_VLAN_INFO_UNTAGGED); } static int sparx5_alloc_mdb_entry(struct sparx5 *sparx5, const unsigned char *addr, u16 vid, struct sparx5_mdb_entry **entry_out) { struct sparx5_mdb_entry *entry; u16 pgid_idx; int err; entry = kzalloc(sizeof(*entry), GFP_KERNEL); if (!entry) return -ENOMEM; err = sparx5_pgid_alloc_mcast(sparx5, &pgid_idx); if (err) { kfree(entry); return err; } memcpy(entry->addr, addr, ETH_ALEN); entry->vid = vid; entry->pgid_idx = pgid_idx; mutex_lock(&sparx5->mdb_lock); list_add_tail(&entry->list, &sparx5->mdb_entries); mutex_unlock(&sparx5->mdb_lock); *entry_out = entry; return 0; } static void sparx5_free_mdb_entry(struct sparx5 *sparx5, const unsigned char *addr, u16 vid) { struct sparx5_mdb_entry *entry, *tmp; mutex_lock(&sparx5->mdb_lock); list_for_each_entry_safe(entry, tmp, &sparx5->mdb_entries, list) { if ((vid == 0 || entry->vid == vid) && ether_addr_equal(addr, entry->addr)) { list_del(&entry->list); sparx5_pgid_free(sparx5, entry->pgid_idx); kfree(entry); goto out; } } out: mutex_unlock(&sparx5->mdb_lock); } static struct sparx5_mdb_entry *sparx5_mdb_get_entry(struct sparx5 *sparx5, const unsigned char *addr, u16 vid) { struct sparx5_mdb_entry *e, *found = NULL; mutex_lock(&sparx5->mdb_lock); list_for_each_entry(e, &sparx5->mdb_entries, list) { if (ether_addr_equal(e->addr, addr) && e->vid == vid) { found = e; goto out; } } out: mutex_unlock(&sparx5->mdb_lock); return found; } static void sparx5_cpu_copy_ena(struct sparx5 *spx5, u16 pgid, bool enable) { spx5_rmw(ANA_AC_PGID_MISC_CFG_PGID_CPU_COPY_ENA_SET(enable), ANA_AC_PGID_MISC_CFG_PGID_CPU_COPY_ENA, spx5, ANA_AC_PGID_MISC_CFG(pgid)); } static int sparx5_handle_port_mdb_add(struct net_device *dev, struct notifier_block *nb, const struct switchdev_obj_port_mdb *v) { struct sparx5_port *port = netdev_priv(dev); struct sparx5 *spx5 = port->sparx5; struct sparx5_mdb_entry *entry; bool is_host, is_new; int err, i; u16 vid; if (!sparx5_netdevice_check(dev)) return -EOPNOTSUPP; is_host = netif_is_bridge_master(v->obj.orig_dev); /* When VLAN unaware the vlan value is not parsed and we receive vid 0. * Fall back to bridge vid 1. */ if (!br_vlan_enabled(spx5->hw_bridge_dev)) vid = 1; else vid = v->vid; is_new = false; entry = sparx5_mdb_get_entry(spx5, v->addr, vid); if (!entry) { err = sparx5_alloc_mdb_entry(spx5, v->addr, vid, &entry); is_new = true; if (err) return err; } mutex_lock(&spx5->mdb_lock); /* Add any mrouter ports to the new entry */ if (is_new && ether_addr_is_ip_mcast(v->addr)) for (i = 0; i < SPX5_PORTS; i++) if (spx5->ports[i] && spx5->ports[i]->is_mrouter) sparx5_pgid_update_mask(spx5->ports[i], entry->pgid_idx, true); if (is_host && !entry->cpu_copy) { sparx5_cpu_copy_ena(spx5, entry->pgid_idx, true); entry->cpu_copy = true; } else if (!is_host) { sparx5_pgid_update_mask(port, entry->pgid_idx, true); set_bit(port->portno, entry->port_mask); } mutex_unlock(&spx5->mdb_lock); sparx5_mact_learn(spx5, entry->pgid_idx, entry->addr, entry->vid); return 0; } static int sparx5_handle_port_mdb_del(struct net_device *dev, struct notifier_block *nb, const struct switchdev_obj_port_mdb *v) { struct sparx5_port *port = netdev_priv(dev); struct sparx5 *spx5 = port->sparx5; struct sparx5_mdb_entry *entry; bool is_host; u16 vid; if (!sparx5_netdevice_check(dev)) return -EOPNOTSUPP; is_host = netif_is_bridge_master(v->obj.orig_dev); if (!br_vlan_enabled(spx5->hw_bridge_dev)) vid = 1; else vid = v->vid; entry = sparx5_mdb_get_entry(spx5, v->addr, vid); if (!entry) return 0; mutex_lock(&spx5->mdb_lock); if (is_host && entry->cpu_copy) { sparx5_cpu_copy_ena(spx5, entry->pgid_idx, false); entry->cpu_copy = false; } else if (!is_host) { clear_bit(port->portno, entry->port_mask); /* Port not mrouter port or addr is L2 mcast, remove port from mask. */ if (!port->is_mrouter || !ether_addr_is_ip_mcast(v->addr)) sparx5_pgid_update_mask(port, entry->pgid_idx, false); } mutex_unlock(&spx5->mdb_lock); if (bitmap_empty(entry->port_mask, SPX5_PORTS) && !entry->cpu_copy) { /* Clear pgid in case mrouter ports exists * that are not part of the group. */ sparx5_pgid_clear(spx5, entry->pgid_idx); sparx5_mact_forget(spx5, entry->addr, entry->vid); sparx5_free_mdb_entry(spx5, entry->addr, entry->vid); } return 0; } static int sparx5_handle_port_obj_add(struct net_device *dev, struct notifier_block *nb, struct switchdev_notifier_port_obj_info *info) { const struct switchdev_obj *obj = info->obj; int err; switch (obj->id) { case SWITCHDEV_OBJ_ID_PORT_VLAN: err = sparx5_handle_port_vlan_add(dev, nb, SWITCHDEV_OBJ_PORT_VLAN(obj)); break; case SWITCHDEV_OBJ_ID_PORT_MDB: case SWITCHDEV_OBJ_ID_HOST_MDB: err = sparx5_handle_port_mdb_add(dev, nb, SWITCHDEV_OBJ_PORT_MDB(obj)); break; default: err = -EOPNOTSUPP; break; } info->handled = true; return err; } static int sparx5_handle_port_vlan_del(struct net_device *dev, struct notifier_block *nb, u16 vid) { struct sparx5_port *port = netdev_priv(dev); int ret; /* Master bridge? */ if (netif_is_bridge_master(dev)) { struct sparx5 *sparx5 = container_of(nb, struct sparx5, switchdev_blocking_nb); sparx5_mact_forget(sparx5, dev->broadcast, vid); return 0; } if (!sparx5_netdevice_check(dev)) return -EOPNOTSUPP; ret = sparx5_vlan_vid_del(port, vid); if (ret) return ret; return 0; } static int sparx5_handle_port_obj_del(struct net_device *dev, struct notifier_block *nb, struct switchdev_notifier_port_obj_info *info) { const struct switchdev_obj *obj = info->obj; int err; switch (obj->id) { case SWITCHDEV_OBJ_ID_PORT_VLAN: err = sparx5_handle_port_vlan_del(dev, nb, SWITCHDEV_OBJ_PORT_VLAN(obj)->vid); break; case SWITCHDEV_OBJ_ID_PORT_MDB: case SWITCHDEV_OBJ_ID_HOST_MDB: err = sparx5_handle_port_mdb_del(dev, nb, SWITCHDEV_OBJ_PORT_MDB(obj)); break; default: err = -EOPNOTSUPP; break; } info->handled = true; return err; } static int sparx5_switchdev_blocking_event(struct notifier_block *nb, unsigned long event, void *ptr) { struct net_device *dev = switchdev_notifier_info_to_dev(ptr); int err; switch (event) { case SWITCHDEV_PORT_OBJ_ADD: err = sparx5_handle_port_obj_add(dev, nb, ptr); return notifier_from_errno(err); case SWITCHDEV_PORT_OBJ_DEL: err = sparx5_handle_port_obj_del(dev, nb, ptr); return notifier_from_errno(err); case SWITCHDEV_PORT_ATTR_SET: err = switchdev_handle_port_attr_set(dev, ptr, sparx5_netdevice_check, sparx5_port_attr_set); return notifier_from_errno(err); } return NOTIFY_DONE; } int sparx5_register_notifier_blocks(struct sparx5 *s5) { int err; s5->netdevice_nb.notifier_call = sparx5_netdevice_event; err = register_netdevice_notifier(&s5->netdevice_nb); if (err) return err; s5->switchdev_nb.notifier_call = sparx5_switchdev_event; err = register_switchdev_notifier(&s5->switchdev_nb); if (err) goto err_switchdev_nb; s5->switchdev_blocking_nb.notifier_call = sparx5_switchdev_blocking_event; err = register_switchdev_blocking_notifier(&s5->switchdev_blocking_nb); if (err) goto err_switchdev_blocking_nb; sparx5_owq = alloc_ordered_workqueue("sparx5_order", 0); if (!sparx5_owq) { err = -ENOMEM; goto err_switchdev_blocking_nb; } return 0; err_switchdev_blocking_nb: unregister_switchdev_notifier(&s5->switchdev_nb); err_switchdev_nb: unregister_netdevice_notifier(&s5->netdevice_nb); return err; } void sparx5_unregister_notifier_blocks(struct sparx5 *s5) { destroy_workqueue(sparx5_owq); unregister_switchdev_blocking_notifier(&s5->switchdev_blocking_nb); unregister_switchdev_notifier(&s5->switchdev_nb); unregister_netdevice_notifier(&s5->netdevice_nb); }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1