Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Roi Dayan | 1326 | 82.57% | 3 | 15.79% |
David Ahern | 110 | 6.85% | 3 | 15.79% |
Maor Dickman | 38 | 2.37% | 1 | 5.26% |
Mark Zhang | 38 | 2.37% | 1 | 5.26% |
Erez Alfasi | 33 | 2.05% | 2 | 10.53% |
Dima Chumak | 32 | 1.99% | 2 | 10.53% |
Jianbo Liu | 17 | 1.06% | 2 | 10.53% |
Jiri Pirko | 8 | 0.50% | 2 | 10.53% |
Mark Bloch | 3 | 0.19% | 2 | 10.53% |
Saeed Mahameed | 1 | 0.06% | 1 | 5.26% |
Total | 1606 | 19 |
// SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB /* Copyright (c) 2019 Mellanox Technologies. */ #include <linux/netdevice.h> #include <net/nexthop.h> #include "lag.h" #include "lag_mp.h" #include "mlx5_core.h" #include "eswitch.h" #include "lib/mlx5.h" static bool __mlx5_lag_is_multipath(struct mlx5_lag *ldev) { return !!(ldev->flags & MLX5_LAG_FLAG_MULTIPATH); } static bool mlx5_lag_multipath_check_prereq(struct mlx5_lag *ldev) { if (!mlx5_lag_is_ready(ldev)) return false; if (__mlx5_lag_is_active(ldev) && !__mlx5_lag_is_multipath(ldev)) return false; return mlx5_esw_multipath_prereq(ldev->pf[MLX5_LAG_P1].dev, ldev->pf[MLX5_LAG_P2].dev); } bool mlx5_lag_is_multipath(struct mlx5_core_dev *dev) { struct mlx5_lag *ldev; bool res; ldev = mlx5_lag_dev(dev); res = ldev && __mlx5_lag_is_multipath(ldev); return res; } /** * mlx5_lag_set_port_affinity * * @ldev: lag device * @port: * 0 - set normal affinity. * 1 - set affinity to port 1. * 2 - set affinity to port 2. * **/ static void mlx5_lag_set_port_affinity(struct mlx5_lag *ldev, enum mlx5_lag_port_affinity port) { struct lag_tracker tracker; if (!__mlx5_lag_is_multipath(ldev)) return; switch (port) { case MLX5_LAG_NORMAL_AFFINITY: tracker.netdev_state[MLX5_LAG_P1].tx_enabled = true; tracker.netdev_state[MLX5_LAG_P2].tx_enabled = true; tracker.netdev_state[MLX5_LAG_P1].link_up = true; tracker.netdev_state[MLX5_LAG_P2].link_up = true; break; case MLX5_LAG_P1_AFFINITY: tracker.netdev_state[MLX5_LAG_P1].tx_enabled = true; tracker.netdev_state[MLX5_LAG_P1].link_up = true; tracker.netdev_state[MLX5_LAG_P2].tx_enabled = false; tracker.netdev_state[MLX5_LAG_P2].link_up = false; break; case MLX5_LAG_P2_AFFINITY: tracker.netdev_state[MLX5_LAG_P1].tx_enabled = false; tracker.netdev_state[MLX5_LAG_P1].link_up = false; tracker.netdev_state[MLX5_LAG_P2].tx_enabled = true; tracker.netdev_state[MLX5_LAG_P2].link_up = true; break; default: mlx5_core_warn(ldev->pf[MLX5_LAG_P1].dev, "Invalid affinity port %d", port); return; } if (tracker.netdev_state[MLX5_LAG_P1].tx_enabled) mlx5_notifier_call_chain(ldev->pf[MLX5_LAG_P1].dev->priv.events, MLX5_DEV_EVENT_PORT_AFFINITY, (void *)0); if (tracker.netdev_state[MLX5_LAG_P2].tx_enabled) mlx5_notifier_call_chain(ldev->pf[MLX5_LAG_P2].dev->priv.events, MLX5_DEV_EVENT_PORT_AFFINITY, (void *)0); mlx5_modify_lag(ldev, &tracker); } static void mlx5_lag_fib_event_flush(struct notifier_block *nb) { struct lag_mp *mp = container_of(nb, struct lag_mp, fib_nb); flush_workqueue(mp->wq); } struct mlx5_fib_event_work { struct work_struct work; struct mlx5_lag *ldev; unsigned long event; union { struct fib_entry_notifier_info fen_info; struct fib_nh_notifier_info fnh_info; }; }; static void mlx5_lag_fib_route_event(struct mlx5_lag *ldev, unsigned long event, struct fib_info *fi) { struct lag_mp *mp = &ldev->lag_mp; struct fib_nh *fib_nh0, *fib_nh1; unsigned int nhs; /* Handle delete event */ if (event == FIB_EVENT_ENTRY_DEL) { /* stop track */ if (mp->mfi == fi) mp->mfi = NULL; return; } /* Handle add/replace event */ nhs = fib_info_num_path(fi); if (nhs == 1) { if (__mlx5_lag_is_active(ldev)) { struct fib_nh *nh = fib_info_nh(fi, 0); struct net_device *nh_dev = nh->fib_nh_dev; int i = mlx5_lag_dev_get_netdev_idx(ldev, nh_dev); if (i < 0) i = MLX5_LAG_NORMAL_AFFINITY; else ++i; mlx5_lag_set_port_affinity(ldev, i); } return; } if (nhs != 2) return; /* Verify next hops are ports of the same hca */ fib_nh0 = fib_info_nh(fi, 0); fib_nh1 = fib_info_nh(fi, 1); if (!(fib_nh0->fib_nh_dev == ldev->pf[MLX5_LAG_P1].netdev && fib_nh1->fib_nh_dev == ldev->pf[MLX5_LAG_P2].netdev) && !(fib_nh0->fib_nh_dev == ldev->pf[MLX5_LAG_P2].netdev && fib_nh1->fib_nh_dev == ldev->pf[MLX5_LAG_P1].netdev)) { mlx5_core_warn(ldev->pf[MLX5_LAG_P1].dev, "Multipath offload require two ports of the same HCA\n"); return; } /* First time we see multipath route */ if (!mp->mfi && !__mlx5_lag_is_active(ldev)) { struct lag_tracker tracker; tracker = ldev->tracker; mlx5_activate_lag(ldev, &tracker, MLX5_LAG_FLAG_MULTIPATH, false); } mlx5_lag_set_port_affinity(ldev, MLX5_LAG_NORMAL_AFFINITY); mp->mfi = fi; } static void mlx5_lag_fib_nexthop_event(struct mlx5_lag *ldev, unsigned long event, struct fib_nh *fib_nh, struct fib_info *fi) { struct lag_mp *mp = &ldev->lag_mp; /* Check the nh event is related to the route */ if (!mp->mfi || mp->mfi != fi) return; /* nh added/removed */ if (event == FIB_EVENT_NH_DEL) { int i = mlx5_lag_dev_get_netdev_idx(ldev, fib_nh->fib_nh_dev); if (i >= 0) { i = (i + 1) % 2 + 1; /* peer port */ mlx5_lag_set_port_affinity(ldev, i); } } else if (event == FIB_EVENT_NH_ADD && fib_info_num_path(fi) == 2) { mlx5_lag_set_port_affinity(ldev, MLX5_LAG_NORMAL_AFFINITY); } } static void mlx5_lag_fib_update(struct work_struct *work) { struct mlx5_fib_event_work *fib_work = container_of(work, struct mlx5_fib_event_work, work); struct mlx5_lag *ldev = fib_work->ldev; struct fib_nh *fib_nh; /* Protect internal structures from changes */ rtnl_lock(); switch (fib_work->event) { case FIB_EVENT_ENTRY_REPLACE: case FIB_EVENT_ENTRY_DEL: mlx5_lag_fib_route_event(ldev, fib_work->event, fib_work->fen_info.fi); fib_info_put(fib_work->fen_info.fi); break; case FIB_EVENT_NH_ADD: case FIB_EVENT_NH_DEL: fib_nh = fib_work->fnh_info.fib_nh; mlx5_lag_fib_nexthop_event(ldev, fib_work->event, fib_work->fnh_info.fib_nh, fib_nh->nh_parent); fib_info_put(fib_work->fnh_info.fib_nh->nh_parent); break; } rtnl_unlock(); kfree(fib_work); } static struct mlx5_fib_event_work * mlx5_lag_init_fib_work(struct mlx5_lag *ldev, unsigned long event) { struct mlx5_fib_event_work *fib_work; fib_work = kzalloc(sizeof(*fib_work), GFP_ATOMIC); if (WARN_ON(!fib_work)) return NULL; INIT_WORK(&fib_work->work, mlx5_lag_fib_update); fib_work->ldev = ldev; fib_work->event = event; return fib_work; } static int mlx5_lag_fib_event(struct notifier_block *nb, unsigned long event, void *ptr) { struct lag_mp *mp = container_of(nb, struct lag_mp, fib_nb); struct mlx5_lag *ldev = container_of(mp, struct mlx5_lag, lag_mp); struct fib_notifier_info *info = ptr; struct mlx5_fib_event_work *fib_work; struct fib_entry_notifier_info *fen_info; struct fib_nh_notifier_info *fnh_info; struct net_device *fib_dev; struct fib_info *fi; if (info->family != AF_INET) return NOTIFY_DONE; if (!mlx5_lag_multipath_check_prereq(ldev)) return NOTIFY_DONE; switch (event) { case FIB_EVENT_ENTRY_REPLACE: case FIB_EVENT_ENTRY_DEL: fen_info = container_of(info, struct fib_entry_notifier_info, info); fi = fen_info->fi; if (fi->nh) { NL_SET_ERR_MSG_MOD(info->extack, "IPv4 route with nexthop objects is not supported"); return notifier_from_errno(-EINVAL); } fib_dev = fib_info_nh(fen_info->fi, 0)->fib_nh_dev; if (fib_dev != ldev->pf[MLX5_LAG_P1].netdev && fib_dev != ldev->pf[MLX5_LAG_P2].netdev) { return NOTIFY_DONE; } fib_work = mlx5_lag_init_fib_work(ldev, event); if (!fib_work) return NOTIFY_DONE; fib_work->fen_info = *fen_info; /* Take reference on fib_info to prevent it from being * freed while work is queued. Release it afterwards. */ fib_info_hold(fib_work->fen_info.fi); break; case FIB_EVENT_NH_ADD: case FIB_EVENT_NH_DEL: fnh_info = container_of(info, struct fib_nh_notifier_info, info); fib_work = mlx5_lag_init_fib_work(ldev, event); if (!fib_work) return NOTIFY_DONE; fib_work->fnh_info = *fnh_info; fib_info_hold(fib_work->fnh_info.fib_nh->nh_parent); break; default: return NOTIFY_DONE; } queue_work(mp->wq, &fib_work->work); return NOTIFY_DONE; } void mlx5_lag_mp_reset(struct mlx5_lag *ldev) { /* Clear mfi, as it might become stale when a route delete event * has been missed, see mlx5_lag_fib_route_event(). */ ldev->lag_mp.mfi = NULL; } int mlx5_lag_mp_init(struct mlx5_lag *ldev) { struct lag_mp *mp = &ldev->lag_mp; int err; /* always clear mfi, as it might become stale when a route delete event * has been missed */ mp->mfi = NULL; if (mp->fib_nb.notifier_call) return 0; mp->wq = create_singlethread_workqueue("mlx5_lag_mp"); if (!mp->wq) return -ENOMEM; mp->fib_nb.notifier_call = mlx5_lag_fib_event; err = register_fib_notifier(&init_net, &mp->fib_nb, mlx5_lag_fib_event_flush, NULL); if (err) { destroy_workqueue(mp->wq); mp->fib_nb.notifier_call = NULL; } return err; } void mlx5_lag_mp_cleanup(struct mlx5_lag *ldev) { struct lag_mp *mp = &ldev->lag_mp; if (!mp->fib_nb.notifier_call) return; unregister_fib_notifier(&init_net, &mp->fib_nb); destroy_workqueue(mp->wq); mp->fib_nb.notifier_call = NULL; mp->mfi = NULL; }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1