Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Dmitry Bogdanov | 8284 | 87.16% | 3 | 50.00% |
Mark Starovoytov | 1219 | 12.83% | 2 | 33.33% |
Dmitry Bezrukov | 1 | 0.01% | 1 | 16.67% |
Total | 9504 | 6 |
// SPDX-License-Identifier: GPL-2.0-only /* Atlantic Network Driver * Copyright (C) 2020 Marvell International Ltd. */ #include "aq_macsec.h" #include "aq_nic.h" #include <linux/rtnetlink.h> #include "macsec/macsec_api.h" #define AQ_MACSEC_KEY_LEN_128_BIT 16 #define AQ_MACSEC_KEY_LEN_192_BIT 24 #define AQ_MACSEC_KEY_LEN_256_BIT 32 enum aq_clear_type { /* update HW configuration */ AQ_CLEAR_HW = BIT(0), /* update SW configuration (busy bits, pointers) */ AQ_CLEAR_SW = BIT(1), /* update both HW and SW configuration */ AQ_CLEAR_ALL = AQ_CLEAR_HW | AQ_CLEAR_SW, }; static int aq_clear_txsc(struct aq_nic_s *nic, const int txsc_idx, enum aq_clear_type clear_type); static int aq_clear_txsa(struct aq_nic_s *nic, struct aq_macsec_txsc *aq_txsc, const int sa_num, enum aq_clear_type clear_type); static int aq_clear_rxsc(struct aq_nic_s *nic, const int rxsc_idx, enum aq_clear_type clear_type); static int aq_clear_rxsa(struct aq_nic_s *nic, struct aq_macsec_rxsc *aq_rxsc, const int sa_num, enum aq_clear_type clear_type); static int aq_clear_secy(struct aq_nic_s *nic, const struct macsec_secy *secy, enum aq_clear_type clear_type); static int aq_apply_macsec_cfg(struct aq_nic_s *nic); static int aq_apply_secy_cfg(struct aq_nic_s *nic, const struct macsec_secy *secy); static void aq_ether_addr_to_mac(u32 mac[2], unsigned char *emac) { u32 tmp[2] = { 0 }; memcpy(((u8 *)tmp) + 2, emac, ETH_ALEN); mac[0] = swab32(tmp[1]); mac[1] = swab32(tmp[0]); } /* There's a 1:1 mapping between SecY and TX SC */ static int aq_get_txsc_idx_from_secy(struct aq_macsec_cfg *macsec_cfg, const struct macsec_secy *secy) { int i; if (unlikely(!secy)) return -1; for (i = 0; i < AQ_MACSEC_MAX_SC; i++) { if (macsec_cfg->aq_txsc[i].sw_secy == secy) return i; } return -1; } static int aq_get_rxsc_idx_from_rxsc(struct aq_macsec_cfg *macsec_cfg, const struct macsec_rx_sc *rxsc) { int i; if (unlikely(!rxsc)) return -1; for (i = 0; i < AQ_MACSEC_MAX_SC; i++) { if (macsec_cfg->aq_rxsc[i].sw_rxsc == rxsc) return i; } return -1; } static int aq_get_txsc_idx_from_sc_idx(const enum aq_macsec_sc_sa sc_sa, const int sc_idx) { switch (sc_sa) { case aq_macsec_sa_sc_4sa_8sc: return sc_idx >> 2; case aq_macsec_sa_sc_2sa_16sc: return sc_idx >> 1; case aq_macsec_sa_sc_1sa_32sc: return sc_idx; default: WARN_ONCE(true, "Invalid sc_sa"); } return -1; } /* Rotate keys u32[8] */ static void aq_rotate_keys(u32 (*key)[8], const int key_len) { u32 tmp[8] = { 0 }; memcpy(&tmp, key, sizeof(tmp)); memset(*key, 0, sizeof(*key)); if (key_len == AQ_MACSEC_KEY_LEN_128_BIT) { (*key)[0] = swab32(tmp[3]); (*key)[1] = swab32(tmp[2]); (*key)[2] = swab32(tmp[1]); (*key)[3] = swab32(tmp[0]); } else if (key_len == AQ_MACSEC_KEY_LEN_192_BIT) { (*key)[0] = swab32(tmp[5]); (*key)[1] = swab32(tmp[4]); (*key)[2] = swab32(tmp[3]); (*key)[3] = swab32(tmp[2]); (*key)[4] = swab32(tmp[1]); (*key)[5] = swab32(tmp[0]); } else if (key_len == AQ_MACSEC_KEY_LEN_256_BIT) { (*key)[0] = swab32(tmp[7]); (*key)[1] = swab32(tmp[6]); (*key)[2] = swab32(tmp[5]); (*key)[3] = swab32(tmp[4]); (*key)[4] = swab32(tmp[3]); (*key)[5] = swab32(tmp[2]); (*key)[6] = swab32(tmp[1]); (*key)[7] = swab32(tmp[0]); } else { pr_warn("Rotate_keys: invalid key_len\n"); } } #define STATS_2x32_TO_64(stat_field) \ (((u64)stat_field[1] << 32) | stat_field[0]) static int aq_get_macsec_common_stats(struct aq_hw_s *hw, struct aq_macsec_common_stats *stats) { struct aq_mss_ingress_common_counters ingress_counters; struct aq_mss_egress_common_counters egress_counters; int ret; /* MACSEC counters */ ret = aq_mss_get_ingress_common_counters(hw, &ingress_counters); if (unlikely(ret)) return ret; stats->in.ctl_pkts = STATS_2x32_TO_64(ingress_counters.ctl_pkts); stats->in.tagged_miss_pkts = STATS_2x32_TO_64(ingress_counters.tagged_miss_pkts); stats->in.untagged_miss_pkts = STATS_2x32_TO_64(ingress_counters.untagged_miss_pkts); stats->in.notag_pkts = STATS_2x32_TO_64(ingress_counters.notag_pkts); stats->in.untagged_pkts = STATS_2x32_TO_64(ingress_counters.untagged_pkts); stats->in.bad_tag_pkts = STATS_2x32_TO_64(ingress_counters.bad_tag_pkts); stats->in.no_sci_pkts = STATS_2x32_TO_64(ingress_counters.no_sci_pkts); stats->in.unknown_sci_pkts = STATS_2x32_TO_64(ingress_counters.unknown_sci_pkts); stats->in.ctrl_prt_pass_pkts = STATS_2x32_TO_64(ingress_counters.ctrl_prt_pass_pkts); stats->in.unctrl_prt_pass_pkts = STATS_2x32_TO_64(ingress_counters.unctrl_prt_pass_pkts); stats->in.ctrl_prt_fail_pkts = STATS_2x32_TO_64(ingress_counters.ctrl_prt_fail_pkts); stats->in.unctrl_prt_fail_pkts = STATS_2x32_TO_64(ingress_counters.unctrl_prt_fail_pkts); stats->in.too_long_pkts = STATS_2x32_TO_64(ingress_counters.too_long_pkts); stats->in.igpoc_ctl_pkts = STATS_2x32_TO_64(ingress_counters.igpoc_ctl_pkts); stats->in.ecc_error_pkts = STATS_2x32_TO_64(ingress_counters.ecc_error_pkts); stats->in.unctrl_hit_drop_redir = STATS_2x32_TO_64(ingress_counters.unctrl_hit_drop_redir); ret = aq_mss_get_egress_common_counters(hw, &egress_counters); if (unlikely(ret)) return ret; stats->out.ctl_pkts = STATS_2x32_TO_64(egress_counters.ctl_pkt); stats->out.unknown_sa_pkts = STATS_2x32_TO_64(egress_counters.unknown_sa_pkts); stats->out.untagged_pkts = STATS_2x32_TO_64(egress_counters.untagged_pkts); stats->out.too_long = STATS_2x32_TO_64(egress_counters.too_long); stats->out.ecc_error_pkts = STATS_2x32_TO_64(egress_counters.ecc_error_pkts); stats->out.unctrl_hit_drop_redir = STATS_2x32_TO_64(egress_counters.unctrl_hit_drop_redir); return 0; } static int aq_get_rxsa_stats(struct aq_hw_s *hw, const int sa_idx, struct aq_macsec_rx_sa_stats *stats) { struct aq_mss_ingress_sa_counters i_sa_counters; int ret; ret = aq_mss_get_ingress_sa_counters(hw, &i_sa_counters, sa_idx); if (unlikely(ret)) return ret; stats->untagged_hit_pkts = STATS_2x32_TO_64(i_sa_counters.untagged_hit_pkts); stats->ctrl_hit_drop_redir_pkts = STATS_2x32_TO_64(i_sa_counters.ctrl_hit_drop_redir_pkts); stats->not_using_sa = STATS_2x32_TO_64(i_sa_counters.not_using_sa); stats->unused_sa = STATS_2x32_TO_64(i_sa_counters.unused_sa); stats->not_valid_pkts = STATS_2x32_TO_64(i_sa_counters.not_valid_pkts); stats->invalid_pkts = STATS_2x32_TO_64(i_sa_counters.invalid_pkts); stats->ok_pkts = STATS_2x32_TO_64(i_sa_counters.ok_pkts); stats->late_pkts = STATS_2x32_TO_64(i_sa_counters.late_pkts); stats->delayed_pkts = STATS_2x32_TO_64(i_sa_counters.delayed_pkts); stats->unchecked_pkts = STATS_2x32_TO_64(i_sa_counters.unchecked_pkts); stats->validated_octets = STATS_2x32_TO_64(i_sa_counters.validated_octets); stats->decrypted_octets = STATS_2x32_TO_64(i_sa_counters.decrypted_octets); return 0; } static int aq_get_txsa_stats(struct aq_hw_s *hw, const int sa_idx, struct aq_macsec_tx_sa_stats *stats) { struct aq_mss_egress_sa_counters e_sa_counters; int ret; ret = aq_mss_get_egress_sa_counters(hw, &e_sa_counters, sa_idx); if (unlikely(ret)) return ret; stats->sa_hit_drop_redirect = STATS_2x32_TO_64(e_sa_counters.sa_hit_drop_redirect); stats->sa_protected2_pkts = STATS_2x32_TO_64(e_sa_counters.sa_protected2_pkts); stats->sa_protected_pkts = STATS_2x32_TO_64(e_sa_counters.sa_protected_pkts); stats->sa_encrypted_pkts = STATS_2x32_TO_64(e_sa_counters.sa_encrypted_pkts); return 0; } static int aq_get_txsa_next_pn(struct aq_hw_s *hw, const int sa_idx, u32 *pn) { struct aq_mss_egress_sa_record sa_rec; int ret; ret = aq_mss_get_egress_sa_record(hw, &sa_rec, sa_idx); if (likely(!ret)) *pn = sa_rec.next_pn; return ret; } static int aq_get_rxsa_next_pn(struct aq_hw_s *hw, const int sa_idx, u32 *pn) { struct aq_mss_ingress_sa_record sa_rec; int ret; ret = aq_mss_get_ingress_sa_record(hw, &sa_rec, sa_idx); if (likely(!ret)) *pn = (!sa_rec.sat_nextpn) ? sa_rec.next_pn : 0; return ret; } static int aq_get_txsc_stats(struct aq_hw_s *hw, const int sc_idx, struct aq_macsec_tx_sc_stats *stats) { struct aq_mss_egress_sc_counters e_sc_counters; int ret; ret = aq_mss_get_egress_sc_counters(hw, &e_sc_counters, sc_idx); if (unlikely(ret)) return ret; stats->sc_protected_pkts = STATS_2x32_TO_64(e_sc_counters.sc_protected_pkts); stats->sc_encrypted_pkts = STATS_2x32_TO_64(e_sc_counters.sc_encrypted_pkts); stats->sc_protected_octets = STATS_2x32_TO_64(e_sc_counters.sc_protected_octets); stats->sc_encrypted_octets = STATS_2x32_TO_64(e_sc_counters.sc_encrypted_octets); return 0; } static int aq_mdo_dev_open(struct macsec_context *ctx) { struct aq_nic_s *nic = netdev_priv(ctx->netdev); int ret = 0; if (ctx->prepare) return 0; if (netif_carrier_ok(nic->ndev)) ret = aq_apply_secy_cfg(nic, ctx->secy); return ret; } static int aq_mdo_dev_stop(struct macsec_context *ctx) { struct aq_nic_s *nic = netdev_priv(ctx->netdev); int i; if (ctx->prepare) return 0; for (i = 0; i < AQ_MACSEC_MAX_SC; i++) { if (nic->macsec_cfg->txsc_idx_busy & BIT(i)) aq_clear_secy(nic, nic->macsec_cfg->aq_txsc[i].sw_secy, AQ_CLEAR_HW); } return 0; } static int aq_set_txsc(struct aq_nic_s *nic, const int txsc_idx) { struct aq_macsec_txsc *aq_txsc = &nic->macsec_cfg->aq_txsc[txsc_idx]; struct aq_mss_egress_class_record tx_class_rec = { 0 }; const struct macsec_secy *secy = aq_txsc->sw_secy; struct aq_mss_egress_sc_record sc_rec = { 0 }; unsigned int sc_idx = aq_txsc->hw_sc_idx; struct aq_hw_s *hw = nic->aq_hw; int ret = 0; aq_ether_addr_to_mac(tx_class_rec.mac_sa, secy->netdev->dev_addr); put_unaligned_be64((__force u64)secy->sci, tx_class_rec.sci); tx_class_rec.sci_mask = 0; tx_class_rec.sa_mask = 0x3f; tx_class_rec.action = 0; /* forward to SA/SC table */ tx_class_rec.valid = 1; tx_class_rec.sc_idx = sc_idx; tx_class_rec.sc_sa = nic->macsec_cfg->sc_sa; ret = aq_mss_set_egress_class_record(hw, &tx_class_rec, txsc_idx); if (ret) return ret; sc_rec.protect = secy->protect_frames; if (secy->tx_sc.encrypt) sc_rec.tci |= BIT(1); if (secy->tx_sc.scb) sc_rec.tci |= BIT(2); if (secy->tx_sc.send_sci) sc_rec.tci |= BIT(3); if (secy->tx_sc.end_station) sc_rec.tci |= BIT(4); /* The C bit is clear if and only if the Secure Data is * exactly the same as the User Data and the ICV is 16 octets long. */ if (!(secy->icv_len == 16 && !secy->tx_sc.encrypt)) sc_rec.tci |= BIT(0); sc_rec.an_roll = 0; switch (secy->key_len) { case AQ_MACSEC_KEY_LEN_128_BIT: sc_rec.sak_len = 0; break; case AQ_MACSEC_KEY_LEN_192_BIT: sc_rec.sak_len = 1; break; case AQ_MACSEC_KEY_LEN_256_BIT: sc_rec.sak_len = 2; break; default: WARN_ONCE(true, "Invalid sc_sa"); return -EINVAL; } sc_rec.curr_an = secy->tx_sc.encoding_sa; sc_rec.valid = 1; sc_rec.fresh = 1; return aq_mss_set_egress_sc_record(hw, &sc_rec, sc_idx); } static u32 aq_sc_idx_max(const enum aq_macsec_sc_sa sc_sa) { u32 result = 0; switch (sc_sa) { case aq_macsec_sa_sc_4sa_8sc: result = 8; break; case aq_macsec_sa_sc_2sa_16sc: result = 16; break; case aq_macsec_sa_sc_1sa_32sc: result = 32; break; default: break; } return result; } static u32 aq_to_hw_sc_idx(const u32 sc_idx, const enum aq_macsec_sc_sa sc_sa) { switch (sc_sa) { case aq_macsec_sa_sc_4sa_8sc: return sc_idx << 2; case aq_macsec_sa_sc_2sa_16sc: return sc_idx << 1; case aq_macsec_sa_sc_1sa_32sc: return sc_idx; default: WARN_ONCE(true, "Invalid sc_sa"); } return sc_idx; } static enum aq_macsec_sc_sa sc_sa_from_num_an(const int num_an) { enum aq_macsec_sc_sa sc_sa = aq_macsec_sa_sc_not_used; switch (num_an) { case 4: sc_sa = aq_macsec_sa_sc_4sa_8sc; break; case 2: sc_sa = aq_macsec_sa_sc_2sa_16sc; break; case 1: sc_sa = aq_macsec_sa_sc_1sa_32sc; break; default: break; } return sc_sa; } static int aq_mdo_add_secy(struct macsec_context *ctx) { struct aq_nic_s *nic = netdev_priv(ctx->netdev); struct aq_macsec_cfg *cfg = nic->macsec_cfg; const struct macsec_secy *secy = ctx->secy; enum aq_macsec_sc_sa sc_sa; u32 txsc_idx; int ret = 0; if (secy->xpn) return -EOPNOTSUPP; sc_sa = sc_sa_from_num_an(MACSEC_NUM_AN); if (sc_sa == aq_macsec_sa_sc_not_used) return -EINVAL; if (hweight32(cfg->txsc_idx_busy) >= aq_sc_idx_max(sc_sa)) return -ENOSPC; txsc_idx = ffz(cfg->txsc_idx_busy); if (txsc_idx == AQ_MACSEC_MAX_SC) return -ENOSPC; if (ctx->prepare) return 0; cfg->sc_sa = sc_sa; cfg->aq_txsc[txsc_idx].hw_sc_idx = aq_to_hw_sc_idx(txsc_idx, sc_sa); cfg->aq_txsc[txsc_idx].sw_secy = secy; if (netif_carrier_ok(nic->ndev) && netif_running(secy->netdev)) ret = aq_set_txsc(nic, txsc_idx); set_bit(txsc_idx, &cfg->txsc_idx_busy); return ret; } static int aq_mdo_upd_secy(struct macsec_context *ctx) { struct aq_nic_s *nic = netdev_priv(ctx->netdev); const struct macsec_secy *secy = ctx->secy; int txsc_idx; int ret = 0; txsc_idx = aq_get_txsc_idx_from_secy(nic->macsec_cfg, secy); if (txsc_idx < 0) return -ENOENT; if (ctx->prepare) return 0; if (netif_carrier_ok(nic->ndev) && netif_running(secy->netdev)) ret = aq_set_txsc(nic, txsc_idx); return ret; } static int aq_clear_txsc(struct aq_nic_s *nic, const int txsc_idx, enum aq_clear_type clear_type) { struct aq_macsec_txsc *tx_sc = &nic->macsec_cfg->aq_txsc[txsc_idx]; struct aq_mss_egress_class_record tx_class_rec = { 0 }; struct aq_mss_egress_sc_record sc_rec = { 0 }; struct aq_hw_s *hw = nic->aq_hw; int ret = 0; int sa_num; for_each_set_bit (sa_num, &tx_sc->tx_sa_idx_busy, AQ_MACSEC_MAX_SA) { ret = aq_clear_txsa(nic, tx_sc, sa_num, clear_type); if (ret) return ret; } if (clear_type & AQ_CLEAR_HW) { ret = aq_mss_set_egress_class_record(hw, &tx_class_rec, txsc_idx); if (ret) return ret; sc_rec.fresh = 1; ret = aq_mss_set_egress_sc_record(hw, &sc_rec, tx_sc->hw_sc_idx); if (ret) return ret; } if (clear_type & AQ_CLEAR_SW) { clear_bit(txsc_idx, &nic->macsec_cfg->txsc_idx_busy); nic->macsec_cfg->aq_txsc[txsc_idx].sw_secy = NULL; } return ret; } static int aq_mdo_del_secy(struct macsec_context *ctx) { struct aq_nic_s *nic = netdev_priv(ctx->netdev); int ret = 0; if (ctx->prepare) return 0; if (!nic->macsec_cfg) return 0; ret = aq_clear_secy(nic, ctx->secy, AQ_CLEAR_ALL); return ret; } static int aq_update_txsa(struct aq_nic_s *nic, const unsigned int sc_idx, const struct macsec_secy *secy, const struct macsec_tx_sa *tx_sa, const unsigned char *key, const unsigned char an) { const u32 next_pn = tx_sa->next_pn_halves.lower; struct aq_mss_egress_sakey_record key_rec; const unsigned int sa_idx = sc_idx | an; struct aq_mss_egress_sa_record sa_rec; struct aq_hw_s *hw = nic->aq_hw; int ret = 0; memset(&sa_rec, 0, sizeof(sa_rec)); sa_rec.valid = tx_sa->active; sa_rec.fresh = 1; sa_rec.next_pn = next_pn; ret = aq_mss_set_egress_sa_record(hw, &sa_rec, sa_idx); if (ret) return ret; if (!key) return ret; memset(&key_rec, 0, sizeof(key_rec)); memcpy(&key_rec.key, key, secy->key_len); aq_rotate_keys(&key_rec.key, secy->key_len); ret = aq_mss_set_egress_sakey_record(hw, &key_rec, sa_idx); return ret; } static int aq_mdo_add_txsa(struct macsec_context *ctx) { struct aq_nic_s *nic = netdev_priv(ctx->netdev); struct aq_macsec_cfg *cfg = nic->macsec_cfg; const struct macsec_secy *secy = ctx->secy; struct aq_macsec_txsc *aq_txsc; int txsc_idx; int ret = 0; txsc_idx = aq_get_txsc_idx_from_secy(cfg, secy); if (txsc_idx < 0) return -EINVAL; if (ctx->prepare) return 0; aq_txsc = &cfg->aq_txsc[txsc_idx]; set_bit(ctx->sa.assoc_num, &aq_txsc->tx_sa_idx_busy); memcpy(aq_txsc->tx_sa_key[ctx->sa.assoc_num], ctx->sa.key, secy->key_len); if (netif_carrier_ok(nic->ndev) && netif_running(secy->netdev)) ret = aq_update_txsa(nic, aq_txsc->hw_sc_idx, secy, ctx->sa.tx_sa, ctx->sa.key, ctx->sa.assoc_num); return ret; } static int aq_mdo_upd_txsa(struct macsec_context *ctx) { struct aq_nic_s *nic = netdev_priv(ctx->netdev); struct aq_macsec_cfg *cfg = nic->macsec_cfg; const struct macsec_secy *secy = ctx->secy; struct aq_macsec_txsc *aq_txsc; int txsc_idx; int ret = 0; txsc_idx = aq_get_txsc_idx_from_secy(cfg, secy); if (txsc_idx < 0) return -EINVAL; if (ctx->prepare) return 0; aq_txsc = &cfg->aq_txsc[txsc_idx]; if (netif_carrier_ok(nic->ndev) && netif_running(secy->netdev)) ret = aq_update_txsa(nic, aq_txsc->hw_sc_idx, secy, ctx->sa.tx_sa, NULL, ctx->sa.assoc_num); return ret; } static int aq_clear_txsa(struct aq_nic_s *nic, struct aq_macsec_txsc *aq_txsc, const int sa_num, enum aq_clear_type clear_type) { const int sa_idx = aq_txsc->hw_sc_idx | sa_num; struct aq_hw_s *hw = nic->aq_hw; int ret = 0; if (clear_type & AQ_CLEAR_SW) clear_bit(sa_num, &aq_txsc->tx_sa_idx_busy); if ((clear_type & AQ_CLEAR_HW) && netif_carrier_ok(nic->ndev)) { struct aq_mss_egress_sakey_record key_rec; struct aq_mss_egress_sa_record sa_rec; memset(&sa_rec, 0, sizeof(sa_rec)); sa_rec.fresh = 1; ret = aq_mss_set_egress_sa_record(hw, &sa_rec, sa_idx); if (ret) return ret; memset(&key_rec, 0, sizeof(key_rec)); return aq_mss_set_egress_sakey_record(hw, &key_rec, sa_idx); } return 0; } static int aq_mdo_del_txsa(struct macsec_context *ctx) { struct aq_nic_s *nic = netdev_priv(ctx->netdev); struct aq_macsec_cfg *cfg = nic->macsec_cfg; int txsc_idx; int ret = 0; txsc_idx = aq_get_txsc_idx_from_secy(cfg, ctx->secy); if (txsc_idx < 0) return -EINVAL; if (ctx->prepare) return 0; ret = aq_clear_txsa(nic, &cfg->aq_txsc[txsc_idx], ctx->sa.assoc_num, AQ_CLEAR_ALL); return ret; } static int aq_rxsc_validate_frames(const enum macsec_validation_type validate) { switch (validate) { case MACSEC_VALIDATE_DISABLED: return 2; case MACSEC_VALIDATE_CHECK: return 1; case MACSEC_VALIDATE_STRICT: return 0; default: WARN_ONCE(true, "Invalid validation type"); } return 0; } static int aq_set_rxsc(struct aq_nic_s *nic, const u32 rxsc_idx) { const struct aq_macsec_rxsc *aq_rxsc = &nic->macsec_cfg->aq_rxsc[rxsc_idx]; struct aq_mss_ingress_preclass_record pre_class_record; const struct macsec_rx_sc *rx_sc = aq_rxsc->sw_rxsc; const struct macsec_secy *secy = aq_rxsc->sw_secy; const u32 hw_sc_idx = aq_rxsc->hw_sc_idx; struct aq_mss_ingress_sc_record sc_record; struct aq_hw_s *hw = nic->aq_hw; int ret = 0; memset(&pre_class_record, 0, sizeof(pre_class_record)); put_unaligned_be64((__force u64)rx_sc->sci, pre_class_record.sci); pre_class_record.sci_mask = 0xff; /* match all MACSEC ethertype packets */ pre_class_record.eth_type = ETH_P_MACSEC; pre_class_record.eth_type_mask = 0x3; aq_ether_addr_to_mac(pre_class_record.mac_sa, (char *)&rx_sc->sci); pre_class_record.sa_mask = 0x3f; pre_class_record.an_mask = nic->macsec_cfg->sc_sa; pre_class_record.sc_idx = hw_sc_idx; /* strip SecTAG & forward for decryption */ pre_class_record.action = 0x0; pre_class_record.valid = 1; ret = aq_mss_set_ingress_preclass_record(hw, &pre_class_record, 2 * rxsc_idx + 1); if (ret) return ret; /* If SCI is absent, then match by SA alone */ pre_class_record.sci_mask = 0; pre_class_record.sci_from_table = 1; ret = aq_mss_set_ingress_preclass_record(hw, &pre_class_record, 2 * rxsc_idx); if (ret) return ret; memset(&sc_record, 0, sizeof(sc_record)); sc_record.validate_frames = aq_rxsc_validate_frames(secy->validate_frames); if (secy->replay_protect) { sc_record.replay_protect = 1; sc_record.anti_replay_window = secy->replay_window; } sc_record.valid = 1; sc_record.fresh = 1; ret = aq_mss_set_ingress_sc_record(hw, &sc_record, hw_sc_idx); if (ret) return ret; return ret; } static int aq_mdo_add_rxsc(struct macsec_context *ctx) { struct aq_nic_s *nic = netdev_priv(ctx->netdev); struct aq_macsec_cfg *cfg = nic->macsec_cfg; const u32 rxsc_idx_max = aq_sc_idx_max(cfg->sc_sa); u32 rxsc_idx; int ret = 0; if (hweight32(cfg->rxsc_idx_busy) >= rxsc_idx_max) return -ENOSPC; rxsc_idx = ffz(cfg->rxsc_idx_busy); if (rxsc_idx >= rxsc_idx_max) return -ENOSPC; if (ctx->prepare) return 0; cfg->aq_rxsc[rxsc_idx].hw_sc_idx = aq_to_hw_sc_idx(rxsc_idx, cfg->sc_sa); cfg->aq_rxsc[rxsc_idx].sw_secy = ctx->secy; cfg->aq_rxsc[rxsc_idx].sw_rxsc = ctx->rx_sc; if (netif_carrier_ok(nic->ndev) && netif_running(ctx->secy->netdev)) ret = aq_set_rxsc(nic, rxsc_idx); if (ret < 0) return ret; set_bit(rxsc_idx, &cfg->rxsc_idx_busy); return 0; } static int aq_mdo_upd_rxsc(struct macsec_context *ctx) { struct aq_nic_s *nic = netdev_priv(ctx->netdev); int rxsc_idx; int ret = 0; rxsc_idx = aq_get_rxsc_idx_from_rxsc(nic->macsec_cfg, ctx->rx_sc); if (rxsc_idx < 0) return -ENOENT; if (ctx->prepare) return 0; if (netif_carrier_ok(nic->ndev) && netif_running(ctx->secy->netdev)) ret = aq_set_rxsc(nic, rxsc_idx); return ret; } static int aq_clear_rxsc(struct aq_nic_s *nic, const int rxsc_idx, enum aq_clear_type clear_type) { struct aq_macsec_rxsc *rx_sc = &nic->macsec_cfg->aq_rxsc[rxsc_idx]; struct aq_hw_s *hw = nic->aq_hw; int ret = 0; int sa_num; for_each_set_bit (sa_num, &rx_sc->rx_sa_idx_busy, AQ_MACSEC_MAX_SA) { ret = aq_clear_rxsa(nic, rx_sc, sa_num, clear_type); if (ret) return ret; } if (clear_type & AQ_CLEAR_HW) { struct aq_mss_ingress_preclass_record pre_class_record; struct aq_mss_ingress_sc_record sc_record; memset(&pre_class_record, 0, sizeof(pre_class_record)); memset(&sc_record, 0, sizeof(sc_record)); ret = aq_mss_set_ingress_preclass_record(hw, &pre_class_record, 2 * rxsc_idx); if (ret) return ret; ret = aq_mss_set_ingress_preclass_record(hw, &pre_class_record, 2 * rxsc_idx + 1); if (ret) return ret; sc_record.fresh = 1; ret = aq_mss_set_ingress_sc_record(hw, &sc_record, rx_sc->hw_sc_idx); if (ret) return ret; } if (clear_type & AQ_CLEAR_SW) { clear_bit(rxsc_idx, &nic->macsec_cfg->rxsc_idx_busy); rx_sc->sw_secy = NULL; rx_sc->sw_rxsc = NULL; } return ret; } static int aq_mdo_del_rxsc(struct macsec_context *ctx) { struct aq_nic_s *nic = netdev_priv(ctx->netdev); enum aq_clear_type clear_type = AQ_CLEAR_SW; int rxsc_idx; int ret = 0; rxsc_idx = aq_get_rxsc_idx_from_rxsc(nic->macsec_cfg, ctx->rx_sc); if (rxsc_idx < 0) return -ENOENT; if (ctx->prepare) return 0; if (netif_carrier_ok(nic->ndev)) clear_type = AQ_CLEAR_ALL; ret = aq_clear_rxsc(nic, rxsc_idx, clear_type); return ret; } static int aq_update_rxsa(struct aq_nic_s *nic, const unsigned int sc_idx, const struct macsec_secy *secy, const struct macsec_rx_sa *rx_sa, const unsigned char *key, const unsigned char an) { struct aq_mss_ingress_sakey_record sa_key_record; const u32 next_pn = rx_sa->next_pn_halves.lower; struct aq_mss_ingress_sa_record sa_record; struct aq_hw_s *hw = nic->aq_hw; const int sa_idx = sc_idx | an; int ret = 0; memset(&sa_record, 0, sizeof(sa_record)); sa_record.valid = rx_sa->active; sa_record.fresh = 1; sa_record.next_pn = next_pn; ret = aq_mss_set_ingress_sa_record(hw, &sa_record, sa_idx); if (ret) return ret; if (!key) return ret; memset(&sa_key_record, 0, sizeof(sa_key_record)); memcpy(&sa_key_record.key, key, secy->key_len); switch (secy->key_len) { case AQ_MACSEC_KEY_LEN_128_BIT: sa_key_record.key_len = 0; break; case AQ_MACSEC_KEY_LEN_192_BIT: sa_key_record.key_len = 1; break; case AQ_MACSEC_KEY_LEN_256_BIT: sa_key_record.key_len = 2; break; default: return -1; } aq_rotate_keys(&sa_key_record.key, secy->key_len); ret = aq_mss_set_ingress_sakey_record(hw, &sa_key_record, sa_idx); return ret; } static int aq_mdo_add_rxsa(struct macsec_context *ctx) { const struct macsec_rx_sc *rx_sc = ctx->sa.rx_sa->sc; struct aq_nic_s *nic = netdev_priv(ctx->netdev); const struct macsec_secy *secy = ctx->secy; struct aq_macsec_rxsc *aq_rxsc; int rxsc_idx; int ret = 0; rxsc_idx = aq_get_rxsc_idx_from_rxsc(nic->macsec_cfg, rx_sc); if (rxsc_idx < 0) return -EINVAL; if (ctx->prepare) return 0; aq_rxsc = &nic->macsec_cfg->aq_rxsc[rxsc_idx]; set_bit(ctx->sa.assoc_num, &aq_rxsc->rx_sa_idx_busy); memcpy(aq_rxsc->rx_sa_key[ctx->sa.assoc_num], ctx->sa.key, secy->key_len); if (netif_carrier_ok(nic->ndev) && netif_running(secy->netdev)) ret = aq_update_rxsa(nic, aq_rxsc->hw_sc_idx, secy, ctx->sa.rx_sa, ctx->sa.key, ctx->sa.assoc_num); return ret; } static int aq_mdo_upd_rxsa(struct macsec_context *ctx) { const struct macsec_rx_sc *rx_sc = ctx->sa.rx_sa->sc; struct aq_nic_s *nic = netdev_priv(ctx->netdev); struct aq_macsec_cfg *cfg = nic->macsec_cfg; const struct macsec_secy *secy = ctx->secy; int rxsc_idx; int ret = 0; rxsc_idx = aq_get_rxsc_idx_from_rxsc(cfg, rx_sc); if (rxsc_idx < 0) return -EINVAL; if (ctx->prepare) return 0; if (netif_carrier_ok(nic->ndev) && netif_running(secy->netdev)) ret = aq_update_rxsa(nic, cfg->aq_rxsc[rxsc_idx].hw_sc_idx, secy, ctx->sa.rx_sa, NULL, ctx->sa.assoc_num); return ret; } static int aq_clear_rxsa(struct aq_nic_s *nic, struct aq_macsec_rxsc *aq_rxsc, const int sa_num, enum aq_clear_type clear_type) { int sa_idx = aq_rxsc->hw_sc_idx | sa_num; struct aq_hw_s *hw = nic->aq_hw; int ret = 0; if (clear_type & AQ_CLEAR_SW) clear_bit(sa_num, &aq_rxsc->rx_sa_idx_busy); if ((clear_type & AQ_CLEAR_HW) && netif_carrier_ok(nic->ndev)) { struct aq_mss_ingress_sakey_record sa_key_record; struct aq_mss_ingress_sa_record sa_record; memset(&sa_key_record, 0, sizeof(sa_key_record)); memset(&sa_record, 0, sizeof(sa_record)); sa_record.fresh = 1; ret = aq_mss_set_ingress_sa_record(hw, &sa_record, sa_idx); if (ret) return ret; return aq_mss_set_ingress_sakey_record(hw, &sa_key_record, sa_idx); } return ret; } static int aq_mdo_del_rxsa(struct macsec_context *ctx) { const struct macsec_rx_sc *rx_sc = ctx->sa.rx_sa->sc; struct aq_nic_s *nic = netdev_priv(ctx->netdev); struct aq_macsec_cfg *cfg = nic->macsec_cfg; int rxsc_idx; int ret = 0; rxsc_idx = aq_get_rxsc_idx_from_rxsc(cfg, rx_sc); if (rxsc_idx < 0) return -EINVAL; if (ctx->prepare) return 0; ret = aq_clear_rxsa(nic, &cfg->aq_rxsc[rxsc_idx], ctx->sa.assoc_num, AQ_CLEAR_ALL); return ret; } static int aq_mdo_get_dev_stats(struct macsec_context *ctx) { struct aq_nic_s *nic = netdev_priv(ctx->netdev); struct aq_macsec_common_stats *stats = &nic->macsec_cfg->stats; struct aq_hw_s *hw = nic->aq_hw; if (ctx->prepare) return 0; aq_get_macsec_common_stats(hw, stats); ctx->stats.dev_stats->OutPktsUntagged = stats->out.untagged_pkts; ctx->stats.dev_stats->InPktsUntagged = stats->in.untagged_pkts; ctx->stats.dev_stats->OutPktsTooLong = stats->out.too_long; ctx->stats.dev_stats->InPktsNoTag = stats->in.notag_pkts; ctx->stats.dev_stats->InPktsBadTag = stats->in.bad_tag_pkts; ctx->stats.dev_stats->InPktsUnknownSCI = stats->in.unknown_sci_pkts; ctx->stats.dev_stats->InPktsNoSCI = stats->in.no_sci_pkts; ctx->stats.dev_stats->InPktsOverrun = 0; return 0; } static int aq_mdo_get_tx_sc_stats(struct macsec_context *ctx) { struct aq_nic_s *nic = netdev_priv(ctx->netdev); struct aq_macsec_tx_sc_stats *stats; struct aq_hw_s *hw = nic->aq_hw; struct aq_macsec_txsc *aq_txsc; int txsc_idx; txsc_idx = aq_get_txsc_idx_from_secy(nic->macsec_cfg, ctx->secy); if (txsc_idx < 0) return -ENOENT; if (ctx->prepare) return 0; aq_txsc = &nic->macsec_cfg->aq_txsc[txsc_idx]; stats = &aq_txsc->stats; aq_get_txsc_stats(hw, aq_txsc->hw_sc_idx, stats); ctx->stats.tx_sc_stats->OutPktsProtected = stats->sc_protected_pkts; ctx->stats.tx_sc_stats->OutPktsEncrypted = stats->sc_encrypted_pkts; ctx->stats.tx_sc_stats->OutOctetsProtected = stats->sc_protected_octets; ctx->stats.tx_sc_stats->OutOctetsEncrypted = stats->sc_encrypted_octets; return 0; } static int aq_mdo_get_tx_sa_stats(struct macsec_context *ctx) { struct aq_nic_s *nic = netdev_priv(ctx->netdev); struct aq_macsec_cfg *cfg = nic->macsec_cfg; struct aq_macsec_tx_sa_stats *stats; struct aq_hw_s *hw = nic->aq_hw; const struct macsec_secy *secy; struct aq_macsec_txsc *aq_txsc; struct macsec_tx_sa *tx_sa; unsigned int sa_idx; int txsc_idx; u32 next_pn; int ret; txsc_idx = aq_get_txsc_idx_from_secy(cfg, ctx->secy); if (txsc_idx < 0) return -EINVAL; if (ctx->prepare) return 0; aq_txsc = &cfg->aq_txsc[txsc_idx]; sa_idx = aq_txsc->hw_sc_idx | ctx->sa.assoc_num; stats = &aq_txsc->tx_sa_stats[ctx->sa.assoc_num]; ret = aq_get_txsa_stats(hw, sa_idx, stats); if (ret) return ret; ctx->stats.tx_sa_stats->OutPktsProtected = stats->sa_protected_pkts; ctx->stats.tx_sa_stats->OutPktsEncrypted = stats->sa_encrypted_pkts; secy = aq_txsc->sw_secy; tx_sa = rcu_dereference_bh(secy->tx_sc.sa[ctx->sa.assoc_num]); ret = aq_get_txsa_next_pn(hw, sa_idx, &next_pn); if (ret == 0) { spin_lock_bh(&tx_sa->lock); tx_sa->next_pn = next_pn; spin_unlock_bh(&tx_sa->lock); } return ret; } static int aq_mdo_get_rx_sc_stats(struct macsec_context *ctx) { struct aq_nic_s *nic = netdev_priv(ctx->netdev); struct aq_macsec_cfg *cfg = nic->macsec_cfg; struct aq_macsec_rx_sa_stats *stats; struct aq_hw_s *hw = nic->aq_hw; struct aq_macsec_rxsc *aq_rxsc; unsigned int sa_idx; int rxsc_idx; int ret = 0; int i; rxsc_idx = aq_get_rxsc_idx_from_rxsc(cfg, ctx->rx_sc); if (rxsc_idx < 0) return -ENOENT; if (ctx->prepare) return 0; aq_rxsc = &cfg->aq_rxsc[rxsc_idx]; for (i = 0; i < MACSEC_NUM_AN; i++) { if (!test_bit(i, &aq_rxsc->rx_sa_idx_busy)) continue; stats = &aq_rxsc->rx_sa_stats[i]; sa_idx = aq_rxsc->hw_sc_idx | i; ret = aq_get_rxsa_stats(hw, sa_idx, stats); if (ret) break; ctx->stats.rx_sc_stats->InOctetsValidated += stats->validated_octets; ctx->stats.rx_sc_stats->InOctetsDecrypted += stats->decrypted_octets; ctx->stats.rx_sc_stats->InPktsUnchecked += stats->unchecked_pkts; ctx->stats.rx_sc_stats->InPktsDelayed += stats->delayed_pkts; ctx->stats.rx_sc_stats->InPktsOK += stats->ok_pkts; ctx->stats.rx_sc_stats->InPktsInvalid += stats->invalid_pkts; ctx->stats.rx_sc_stats->InPktsLate += stats->late_pkts; ctx->stats.rx_sc_stats->InPktsNotValid += stats->not_valid_pkts; ctx->stats.rx_sc_stats->InPktsNotUsingSA += stats->not_using_sa; ctx->stats.rx_sc_stats->InPktsUnusedSA += stats->unused_sa; } return ret; } static int aq_mdo_get_rx_sa_stats(struct macsec_context *ctx) { struct aq_nic_s *nic = netdev_priv(ctx->netdev); struct aq_macsec_cfg *cfg = nic->macsec_cfg; struct aq_macsec_rx_sa_stats *stats; struct aq_hw_s *hw = nic->aq_hw; struct aq_macsec_rxsc *aq_rxsc; struct macsec_rx_sa *rx_sa; unsigned int sa_idx; int rxsc_idx; u32 next_pn; int ret; rxsc_idx = aq_get_rxsc_idx_from_rxsc(cfg, ctx->rx_sc); if (rxsc_idx < 0) return -EINVAL; if (ctx->prepare) return 0; aq_rxsc = &cfg->aq_rxsc[rxsc_idx]; stats = &aq_rxsc->rx_sa_stats[ctx->sa.assoc_num]; sa_idx = aq_rxsc->hw_sc_idx | ctx->sa.assoc_num; ret = aq_get_rxsa_stats(hw, sa_idx, stats); if (ret) return ret; ctx->stats.rx_sa_stats->InPktsOK = stats->ok_pkts; ctx->stats.rx_sa_stats->InPktsInvalid = stats->invalid_pkts; ctx->stats.rx_sa_stats->InPktsNotValid = stats->not_valid_pkts; ctx->stats.rx_sa_stats->InPktsNotUsingSA = stats->not_using_sa; ctx->stats.rx_sa_stats->InPktsUnusedSA = stats->unused_sa; rx_sa = rcu_dereference_bh(aq_rxsc->sw_rxsc->sa[ctx->sa.assoc_num]); ret = aq_get_rxsa_next_pn(hw, sa_idx, &next_pn); if (ret == 0) { spin_lock_bh(&rx_sa->lock); rx_sa->next_pn = next_pn; spin_unlock_bh(&rx_sa->lock); } return ret; } static int apply_txsc_cfg(struct aq_nic_s *nic, const int txsc_idx) { struct aq_macsec_txsc *aq_txsc = &nic->macsec_cfg->aq_txsc[txsc_idx]; const struct macsec_secy *secy = aq_txsc->sw_secy; struct macsec_tx_sa *tx_sa; int ret = 0; int i; if (!netif_running(secy->netdev)) return ret; ret = aq_set_txsc(nic, txsc_idx); if (ret) return ret; for (i = 0; i < MACSEC_NUM_AN; i++) { tx_sa = rcu_dereference_bh(secy->tx_sc.sa[i]); if (tx_sa) { ret = aq_update_txsa(nic, aq_txsc->hw_sc_idx, secy, tx_sa, aq_txsc->tx_sa_key[i], i); if (ret) return ret; } } return ret; } static int apply_rxsc_cfg(struct aq_nic_s *nic, const int rxsc_idx) { struct aq_macsec_rxsc *aq_rxsc = &nic->macsec_cfg->aq_rxsc[rxsc_idx]; const struct macsec_secy *secy = aq_rxsc->sw_secy; struct macsec_rx_sa *rx_sa; int ret = 0; int i; if (!netif_running(secy->netdev)) return ret; ret = aq_set_rxsc(nic, rxsc_idx); if (ret) return ret; for (i = 0; i < MACSEC_NUM_AN; i++) { rx_sa = rcu_dereference_bh(aq_rxsc->sw_rxsc->sa[i]); if (rx_sa) { ret = aq_update_rxsa(nic, aq_rxsc->hw_sc_idx, secy, rx_sa, aq_rxsc->rx_sa_key[i], i); if (ret) return ret; } } return ret; } static int aq_clear_secy(struct aq_nic_s *nic, const struct macsec_secy *secy, enum aq_clear_type clear_type) { struct macsec_rx_sc *rx_sc; int txsc_idx; int rxsc_idx; int ret = 0; txsc_idx = aq_get_txsc_idx_from_secy(nic->macsec_cfg, secy); if (txsc_idx >= 0) { ret = aq_clear_txsc(nic, txsc_idx, clear_type); if (ret) return ret; } for (rx_sc = rcu_dereference_bh(secy->rx_sc); rx_sc; rx_sc = rcu_dereference_bh(rx_sc->next)) { rxsc_idx = aq_get_rxsc_idx_from_rxsc(nic->macsec_cfg, rx_sc); if (rxsc_idx < 0) continue; ret = aq_clear_rxsc(nic, rxsc_idx, clear_type); if (ret) return ret; } return ret; } static int aq_apply_secy_cfg(struct aq_nic_s *nic, const struct macsec_secy *secy) { struct macsec_rx_sc *rx_sc; int txsc_idx; int rxsc_idx; int ret = 0; txsc_idx = aq_get_txsc_idx_from_secy(nic->macsec_cfg, secy); if (txsc_idx >= 0) apply_txsc_cfg(nic, txsc_idx); for (rx_sc = rcu_dereference_bh(secy->rx_sc); rx_sc && rx_sc->active; rx_sc = rcu_dereference_bh(rx_sc->next)) { rxsc_idx = aq_get_rxsc_idx_from_rxsc(nic->macsec_cfg, rx_sc); if (unlikely(rxsc_idx < 0)) continue; ret = apply_rxsc_cfg(nic, rxsc_idx); if (ret) return ret; } return ret; } static int aq_apply_macsec_cfg(struct aq_nic_s *nic) { int ret = 0; int i; for (i = 0; i < AQ_MACSEC_MAX_SC; i++) { if (nic->macsec_cfg->txsc_idx_busy & BIT(i)) { ret = apply_txsc_cfg(nic, i); if (ret) return ret; } } for (i = 0; i < AQ_MACSEC_MAX_SC; i++) { if (nic->macsec_cfg->rxsc_idx_busy & BIT(i)) { ret = apply_rxsc_cfg(nic, i); if (ret) return ret; } } return ret; } static int aq_sa_from_sa_idx(const enum aq_macsec_sc_sa sc_sa, const int sa_idx) { switch (sc_sa) { case aq_macsec_sa_sc_4sa_8sc: return sa_idx & 3; case aq_macsec_sa_sc_2sa_16sc: return sa_idx & 1; case aq_macsec_sa_sc_1sa_32sc: return 0; default: WARN_ONCE(true, "Invalid sc_sa"); } return -EINVAL; } static int aq_sc_idx_from_sa_idx(const enum aq_macsec_sc_sa sc_sa, const int sa_idx) { switch (sc_sa) { case aq_macsec_sa_sc_4sa_8sc: return sa_idx & ~3; case aq_macsec_sa_sc_2sa_16sc: return sa_idx & ~1; case aq_macsec_sa_sc_1sa_32sc: return sa_idx; default: WARN_ONCE(true, "Invalid sc_sa"); } return -EINVAL; } static void aq_check_txsa_expiration(struct aq_nic_s *nic) { u32 egress_sa_expired, egress_sa_threshold_expired; struct aq_macsec_cfg *cfg = nic->macsec_cfg; struct aq_hw_s *hw = nic->aq_hw; struct aq_macsec_txsc *aq_txsc; const struct macsec_secy *secy; int sc_idx = 0, txsc_idx = 0; enum aq_macsec_sc_sa sc_sa; struct macsec_tx_sa *tx_sa; unsigned char an = 0; int ret; int i; sc_sa = cfg->sc_sa; ret = aq_mss_get_egress_sa_expired(hw, &egress_sa_expired); if (unlikely(ret)) return; ret = aq_mss_get_egress_sa_threshold_expired(hw, &egress_sa_threshold_expired); for (i = 0; i < AQ_MACSEC_MAX_SA; i++) { if (egress_sa_expired & BIT(i)) { an = aq_sa_from_sa_idx(sc_sa, i); sc_idx = aq_sc_idx_from_sa_idx(sc_sa, i); txsc_idx = aq_get_txsc_idx_from_sc_idx(sc_sa, sc_idx); if (txsc_idx < 0) continue; aq_txsc = &cfg->aq_txsc[txsc_idx]; if (!(cfg->txsc_idx_busy & BIT(txsc_idx))) { netdev_warn(nic->ndev, "PN threshold expired on invalid TX SC"); continue; } secy = aq_txsc->sw_secy; if (!netif_running(secy->netdev)) { netdev_warn(nic->ndev, "PN threshold expired on down TX SC"); continue; } if (unlikely(!(aq_txsc->tx_sa_idx_busy & BIT(an)))) { netdev_warn(nic->ndev, "PN threshold expired on invalid TX SA"); continue; } tx_sa = rcu_dereference_bh(secy->tx_sc.sa[an]); macsec_pn_wrapped((struct macsec_secy *)secy, tx_sa); } } aq_mss_set_egress_sa_expired(hw, egress_sa_expired); if (likely(!ret)) aq_mss_set_egress_sa_threshold_expired(hw, egress_sa_threshold_expired); } const struct macsec_ops aq_macsec_ops = { .mdo_dev_open = aq_mdo_dev_open, .mdo_dev_stop = aq_mdo_dev_stop, .mdo_add_secy = aq_mdo_add_secy, .mdo_upd_secy = aq_mdo_upd_secy, .mdo_del_secy = aq_mdo_del_secy, .mdo_add_rxsc = aq_mdo_add_rxsc, .mdo_upd_rxsc = aq_mdo_upd_rxsc, .mdo_del_rxsc = aq_mdo_del_rxsc, .mdo_add_rxsa = aq_mdo_add_rxsa, .mdo_upd_rxsa = aq_mdo_upd_rxsa, .mdo_del_rxsa = aq_mdo_del_rxsa, .mdo_add_txsa = aq_mdo_add_txsa, .mdo_upd_txsa = aq_mdo_upd_txsa, .mdo_del_txsa = aq_mdo_del_txsa, .mdo_get_dev_stats = aq_mdo_get_dev_stats, .mdo_get_tx_sc_stats = aq_mdo_get_tx_sc_stats, .mdo_get_tx_sa_stats = aq_mdo_get_tx_sa_stats, .mdo_get_rx_sc_stats = aq_mdo_get_rx_sc_stats, .mdo_get_rx_sa_stats = aq_mdo_get_rx_sa_stats, }; int aq_macsec_init(struct aq_nic_s *nic) { struct aq_macsec_cfg *cfg; u32 caps_lo; if (!nic->aq_fw_ops->get_link_capabilities) return 0; caps_lo = nic->aq_fw_ops->get_link_capabilities(nic->aq_hw); if (!(caps_lo & BIT(CAPS_LO_MACSEC))) return 0; nic->macsec_cfg = kzalloc(sizeof(*cfg), GFP_KERNEL); if (!nic->macsec_cfg) return -ENOMEM; nic->ndev->features |= NETIF_F_HW_MACSEC; nic->ndev->macsec_ops = &aq_macsec_ops; return 0; } void aq_macsec_free(struct aq_nic_s *nic) { kfree(nic->macsec_cfg); nic->macsec_cfg = NULL; } int aq_macsec_enable(struct aq_nic_s *nic) { u32 ctl_ether_types[1] = { ETH_P_PAE }; struct macsec_msg_fw_response resp = { 0 }; struct macsec_msg_fw_request msg = { 0 }; struct aq_hw_s *hw = nic->aq_hw; int num_ctl_ether_types = 0; int index = 0, tbl_idx; int ret; if (!nic->macsec_cfg) return 0; rtnl_lock(); if (nic->aq_fw_ops->send_macsec_req) { struct macsec_cfg_request cfg = { 0 }; cfg.enabled = 1; cfg.egress_threshold = 0xffffffff; cfg.ingress_threshold = 0xffffffff; cfg.interrupts_enabled = 1; msg.msg_type = macsec_cfg_msg; msg.cfg = cfg; ret = nic->aq_fw_ops->send_macsec_req(hw, &msg, &resp); if (ret) goto unlock; } /* Init Ethertype bypass filters */ for (index = 0; index < ARRAY_SIZE(ctl_ether_types); index++) { struct aq_mss_ingress_prectlf_record rx_prectlf_rec; struct aq_mss_egress_ctlf_record tx_ctlf_rec; if (ctl_ether_types[index] == 0) continue; memset(&tx_ctlf_rec, 0, sizeof(tx_ctlf_rec)); tx_ctlf_rec.eth_type = ctl_ether_types[index]; tx_ctlf_rec.match_type = 4; /* Match eth_type only */ tx_ctlf_rec.match_mask = 0xf; /* match for eth_type */ tx_ctlf_rec.action = 0; /* Bypass MACSEC modules */ tbl_idx = NUMROWS_EGRESSCTLFRECORD - num_ctl_ether_types - 1; aq_mss_set_egress_ctlf_record(hw, &tx_ctlf_rec, tbl_idx); memset(&rx_prectlf_rec, 0, sizeof(rx_prectlf_rec)); rx_prectlf_rec.eth_type = ctl_ether_types[index]; rx_prectlf_rec.match_type = 4; /* Match eth_type only */ rx_prectlf_rec.match_mask = 0xf; /* match for eth_type */ rx_prectlf_rec.action = 0; /* Bypass MACSEC modules */ tbl_idx = NUMROWS_INGRESSPRECTLFRECORD - num_ctl_ether_types - 1; aq_mss_set_ingress_prectlf_record(hw, &rx_prectlf_rec, tbl_idx); num_ctl_ether_types++; } ret = aq_apply_macsec_cfg(nic); unlock: rtnl_unlock(); return ret; } void aq_macsec_work(struct aq_nic_s *nic) { if (!nic->macsec_cfg) return; if (!netif_carrier_ok(nic->ndev)) return; rtnl_lock(); aq_check_txsa_expiration(nic); rtnl_unlock(); } int aq_macsec_rx_sa_cnt(struct aq_nic_s *nic) { struct aq_macsec_cfg *cfg = nic->macsec_cfg; int i, cnt = 0; if (!cfg) return 0; for (i = 0; i < AQ_MACSEC_MAX_SC; i++) { if (!test_bit(i, &cfg->rxsc_idx_busy)) continue; cnt += hweight_long(cfg->aq_rxsc[i].rx_sa_idx_busy); } return cnt; } int aq_macsec_tx_sc_cnt(struct aq_nic_s *nic) { if (!nic->macsec_cfg) return 0; return hweight_long(nic->macsec_cfg->txsc_idx_busy); } int aq_macsec_tx_sa_cnt(struct aq_nic_s *nic) { struct aq_macsec_cfg *cfg = nic->macsec_cfg; int i, cnt = 0; if (!cfg) return 0; for (i = 0; i < AQ_MACSEC_MAX_SC; i++) { if (!test_bit(i, &cfg->txsc_idx_busy)) continue; cnt += hweight_long(cfg->aq_txsc[i].tx_sa_idx_busy); } return cnt; } static int aq_macsec_update_stats(struct aq_nic_s *nic) { struct aq_macsec_cfg *cfg = nic->macsec_cfg; struct aq_hw_s *hw = nic->aq_hw; struct aq_macsec_txsc *aq_txsc; struct aq_macsec_rxsc *aq_rxsc; int i, sa_idx, assoc_num; int ret = 0; aq_get_macsec_common_stats(hw, &cfg->stats); for (i = 0; i < AQ_MACSEC_MAX_SC; i++) { if (!(cfg->txsc_idx_busy & BIT(i))) continue; aq_txsc = &cfg->aq_txsc[i]; ret = aq_get_txsc_stats(hw, aq_txsc->hw_sc_idx, &aq_txsc->stats); if (ret) return ret; for (assoc_num = 0; assoc_num < MACSEC_NUM_AN; assoc_num++) { if (!test_bit(assoc_num, &aq_txsc->tx_sa_idx_busy)) continue; sa_idx = aq_txsc->hw_sc_idx | assoc_num; ret = aq_get_txsa_stats(hw, sa_idx, &aq_txsc->tx_sa_stats[assoc_num]); if (ret) return ret; } } for (i = 0; i < AQ_MACSEC_MAX_SC; i++) { if (!(test_bit(i, &cfg->rxsc_idx_busy))) continue; aq_rxsc = &cfg->aq_rxsc[i]; for (assoc_num = 0; assoc_num < MACSEC_NUM_AN; assoc_num++) { if (!test_bit(assoc_num, &aq_rxsc->rx_sa_idx_busy)) continue; sa_idx = aq_rxsc->hw_sc_idx | assoc_num; ret = aq_get_rxsa_stats(hw, sa_idx, &aq_rxsc->rx_sa_stats[assoc_num]); if (ret) return ret; } } return ret; } u64 *aq_macsec_get_stats(struct aq_nic_s *nic, u64 *data) { struct aq_macsec_cfg *cfg = nic->macsec_cfg; struct aq_macsec_common_stats *common_stats; struct aq_macsec_tx_sc_stats *txsc_stats; struct aq_macsec_tx_sa_stats *txsa_stats; struct aq_macsec_rx_sa_stats *rxsa_stats; struct aq_macsec_txsc *aq_txsc; struct aq_macsec_rxsc *aq_rxsc; unsigned int assoc_num; unsigned int sc_num; unsigned int i = 0U; if (!cfg) return data; aq_macsec_update_stats(nic); common_stats = &cfg->stats; data[i] = common_stats->in.ctl_pkts; data[++i] = common_stats->in.tagged_miss_pkts; data[++i] = common_stats->in.untagged_miss_pkts; data[++i] = common_stats->in.notag_pkts; data[++i] = common_stats->in.untagged_pkts; data[++i] = common_stats->in.bad_tag_pkts; data[++i] = common_stats->in.no_sci_pkts; data[++i] = common_stats->in.unknown_sci_pkts; data[++i] = common_stats->in.ctrl_prt_pass_pkts; data[++i] = common_stats->in.unctrl_prt_pass_pkts; data[++i] = common_stats->in.ctrl_prt_fail_pkts; data[++i] = common_stats->in.unctrl_prt_fail_pkts; data[++i] = common_stats->in.too_long_pkts; data[++i] = common_stats->in.igpoc_ctl_pkts; data[++i] = common_stats->in.ecc_error_pkts; data[++i] = common_stats->in.unctrl_hit_drop_redir; data[++i] = common_stats->out.ctl_pkts; data[++i] = common_stats->out.unknown_sa_pkts; data[++i] = common_stats->out.untagged_pkts; data[++i] = common_stats->out.too_long; data[++i] = common_stats->out.ecc_error_pkts; data[++i] = common_stats->out.unctrl_hit_drop_redir; for (sc_num = 0; sc_num < AQ_MACSEC_MAX_SC; sc_num++) { if (!(test_bit(sc_num, &cfg->txsc_idx_busy))) continue; aq_txsc = &cfg->aq_txsc[sc_num]; txsc_stats = &aq_txsc->stats; data[++i] = txsc_stats->sc_protected_pkts; data[++i] = txsc_stats->sc_encrypted_pkts; data[++i] = txsc_stats->sc_protected_octets; data[++i] = txsc_stats->sc_encrypted_octets; for (assoc_num = 0; assoc_num < MACSEC_NUM_AN; assoc_num++) { if (!test_bit(assoc_num, &aq_txsc->tx_sa_idx_busy)) continue; txsa_stats = &aq_txsc->tx_sa_stats[assoc_num]; data[++i] = txsa_stats->sa_hit_drop_redirect; data[++i] = txsa_stats->sa_protected2_pkts; data[++i] = txsa_stats->sa_protected_pkts; data[++i] = txsa_stats->sa_encrypted_pkts; } } for (sc_num = 0; sc_num < AQ_MACSEC_MAX_SC; sc_num++) { if (!(test_bit(sc_num, &cfg->rxsc_idx_busy))) continue; aq_rxsc = &cfg->aq_rxsc[sc_num]; for (assoc_num = 0; assoc_num < MACSEC_NUM_AN; assoc_num++) { if (!test_bit(assoc_num, &aq_rxsc->rx_sa_idx_busy)) continue; rxsa_stats = &aq_rxsc->rx_sa_stats[assoc_num]; data[++i] = rxsa_stats->untagged_hit_pkts; data[++i] = rxsa_stats->ctrl_hit_drop_redir_pkts; data[++i] = rxsa_stats->not_using_sa; data[++i] = rxsa_stats->unused_sa; data[++i] = rxsa_stats->not_valid_pkts; data[++i] = rxsa_stats->invalid_pkts; data[++i] = rxsa_stats->ok_pkts; data[++i] = rxsa_stats->late_pkts; data[++i] = rxsa_stats->delayed_pkts; data[++i] = rxsa_stats->unchecked_pkts; data[++i] = rxsa_stats->validated_octets; data[++i] = rxsa_stats->decrypted_octets; } } i++; data += i; return data; }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1