Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Subbaraya Sundeep | 1104 | 85.78% | 3 | 13.64% |
Sunil Goutham | 96 | 7.46% | 12 | 54.55% |
George Cherian | 50 | 3.89% | 2 | 9.09% |
Christina Jacob | 21 | 1.63% | 1 | 4.55% |
Aleksey Makarov | 6 | 0.47% | 1 | 4.55% |
SrujanaChalla | 5 | 0.39% | 1 | 4.55% |
Felix Manlunas | 4 | 0.31% | 1 | 4.55% |
Harman Kalra | 1 | 0.08% | 1 | 4.55% |
Total | 1287 | 22 |
// SPDX-License-Identifier: GPL-2.0 /* Marvell RVU Admin Function driver * * Copyright (C) 2021 Marvell. * */ #include <linux/bitfield.h> #include "rvu.h" static int rvu_switch_install_rx_rule(struct rvu *rvu, u16 pcifunc, u16 chan_mask) { struct npc_install_flow_req req = { 0 }; struct npc_install_flow_rsp rsp = { 0 }; struct rvu_pfvf *pfvf; pfvf = rvu_get_pfvf(rvu, pcifunc); /* If the pcifunc is not initialized then nothing to do. * This same function will be called again via rvu_switch_update_rules * after pcifunc is initialized. */ if (!test_bit(NIXLF_INITIALIZED, &pfvf->flags)) return 0; ether_addr_copy(req.packet.dmac, pfvf->mac_addr); eth_broadcast_addr((u8 *)&req.mask.dmac); req.hdr.pcifunc = 0; /* AF is requester */ req.vf = pcifunc; req.features = BIT_ULL(NPC_DMAC); req.channel = pfvf->rx_chan_base; req.chan_mask = chan_mask; req.intf = pfvf->nix_rx_intf; req.op = NIX_RX_ACTION_DEFAULT; req.default_rule = 1; return rvu_mbox_handler_npc_install_flow(rvu, &req, &rsp); } static int rvu_switch_install_tx_rule(struct rvu *rvu, u16 pcifunc, u16 entry) { struct npc_install_flow_req req = { 0 }; struct npc_install_flow_rsp rsp = { 0 }; struct rvu_pfvf *pfvf; u8 lbkid; pfvf = rvu_get_pfvf(rvu, pcifunc); /* If the pcifunc is not initialized then nothing to do. * This same function will be called again via rvu_switch_update_rules * after pcifunc is initialized. */ if (!test_bit(NIXLF_INITIALIZED, &pfvf->flags)) return 0; lbkid = pfvf->nix_blkaddr == BLKADDR_NIX0 ? 0 : 1; ether_addr_copy(req.packet.dmac, pfvf->mac_addr); eth_broadcast_addr((u8 *)&req.mask.dmac); req.hdr.pcifunc = 0; /* AF is requester */ req.vf = pcifunc; req.entry = entry; req.features = BIT_ULL(NPC_DMAC); req.intf = pfvf->nix_tx_intf; req.op = NIX_TX_ACTIONOP_UCAST_CHAN; req.index = (lbkid << 8) | RVU_SWITCH_LBK_CHAN; req.set_cntr = 1; return rvu_mbox_handler_npc_install_flow(rvu, &req, &rsp); } static int rvu_switch_install_rules(struct rvu *rvu) { struct rvu_switch *rswitch = &rvu->rswitch; u16 start = rswitch->start_entry; struct rvu_hwinfo *hw = rvu->hw; u16 pcifunc, entry = 0; int pf, vf, numvfs; int err; for (pf = 1; pf < hw->total_pfs; pf++) { if (!is_pf_cgxmapped(rvu, pf)) continue; pcifunc = pf << 10; /* rvu_get_nix_blkaddr sets up the corresponding NIX block * address and NIX RX and TX interfaces for a pcifunc. * Generally it is called during attach call of a pcifunc but it * is called here since we are pre-installing rules before * nixlfs are attached */ rvu_get_nix_blkaddr(rvu, pcifunc); /* MCAM RX rule for a PF/VF already exists as default unicast * rules installed by AF. Hence change the channel in those * rules to ignore channel so that packets with the required * DMAC received from LBK(by other PF/VFs in system) or from * external world (from wire) are accepted. */ err = rvu_switch_install_rx_rule(rvu, pcifunc, 0x0); if (err) { dev_err(rvu->dev, "RX rule for PF%d failed(%d)\n", pf, err); return err; } err = rvu_switch_install_tx_rule(rvu, pcifunc, start + entry); if (err) { dev_err(rvu->dev, "TX rule for PF%d failed(%d)\n", pf, err); return err; } rswitch->entry2pcifunc[entry++] = pcifunc; rvu_get_pf_numvfs(rvu, pf, &numvfs, NULL); for (vf = 0; vf < numvfs; vf++) { pcifunc = pf << 10 | ((vf + 1) & 0x3FF); rvu_get_nix_blkaddr(rvu, pcifunc); err = rvu_switch_install_rx_rule(rvu, pcifunc, 0x0); if (err) { dev_err(rvu->dev, "RX rule for PF%dVF%d failed(%d)\n", pf, vf, err); return err; } err = rvu_switch_install_tx_rule(rvu, pcifunc, start + entry); if (err) { dev_err(rvu->dev, "TX rule for PF%dVF%d failed(%d)\n", pf, vf, err); return err; } rswitch->entry2pcifunc[entry++] = pcifunc; } } return 0; } void rvu_switch_enable(struct rvu *rvu) { struct npc_mcam_alloc_entry_req alloc_req = { 0 }; struct npc_mcam_alloc_entry_rsp alloc_rsp = { 0 }; struct npc_delete_flow_req uninstall_req = { 0 }; struct npc_mcam_free_entry_req free_req = { 0 }; struct rvu_switch *rswitch = &rvu->rswitch; struct msg_rsp rsp; int ret; alloc_req.contig = true; alloc_req.count = rvu->cgx_mapped_pfs + rvu->cgx_mapped_vfs; ret = rvu_mbox_handler_npc_mcam_alloc_entry(rvu, &alloc_req, &alloc_rsp); if (ret) { dev_err(rvu->dev, "Unable to allocate MCAM entries\n"); goto exit; } if (alloc_rsp.count != alloc_req.count) { dev_err(rvu->dev, "Unable to allocate %d MCAM entries, got %d\n", alloc_req.count, alloc_rsp.count); goto free_entries; } rswitch->entry2pcifunc = kcalloc(alloc_req.count, sizeof(u16), GFP_KERNEL); if (!rswitch->entry2pcifunc) goto free_entries; rswitch->used_entries = alloc_rsp.count; rswitch->start_entry = alloc_rsp.entry; ret = rvu_switch_install_rules(rvu); if (ret) goto uninstall_rules; return; uninstall_rules: uninstall_req.start = rswitch->start_entry; uninstall_req.end = rswitch->start_entry + rswitch->used_entries - 1; rvu_mbox_handler_npc_delete_flow(rvu, &uninstall_req, &rsp); kfree(rswitch->entry2pcifunc); free_entries: free_req.all = 1; rvu_mbox_handler_npc_mcam_free_entry(rvu, &free_req, &rsp); exit: return; } void rvu_switch_disable(struct rvu *rvu) { struct npc_delete_flow_req uninstall_req = { 0 }; struct npc_mcam_free_entry_req free_req = { 0 }; struct rvu_switch *rswitch = &rvu->rswitch; struct rvu_hwinfo *hw = rvu->hw; int pf, vf, numvfs; struct msg_rsp rsp; u16 pcifunc; int err; if (!rswitch->used_entries) return; for (pf = 1; pf < hw->total_pfs; pf++) { if (!is_pf_cgxmapped(rvu, pf)) continue; pcifunc = pf << 10; err = rvu_switch_install_rx_rule(rvu, pcifunc, 0xFFF); if (err) dev_err(rvu->dev, "Reverting RX rule for PF%d failed(%d)\n", pf, err); rvu_get_pf_numvfs(rvu, pf, &numvfs, NULL); for (vf = 0; vf < numvfs; vf++) { pcifunc = pf << 10 | ((vf + 1) & 0x3FF); err = rvu_switch_install_rx_rule(rvu, pcifunc, 0xFFF); if (err) dev_err(rvu->dev, "Reverting RX rule for PF%dVF%d failed(%d)\n", pf, vf, err); } } uninstall_req.start = rswitch->start_entry; uninstall_req.end = rswitch->start_entry + rswitch->used_entries - 1; free_req.all = 1; rvu_mbox_handler_npc_delete_flow(rvu, &uninstall_req, &rsp); rvu_mbox_handler_npc_mcam_free_entry(rvu, &free_req, &rsp); rswitch->used_entries = 0; kfree(rswitch->entry2pcifunc); } void rvu_switch_update_rules(struct rvu *rvu, u16 pcifunc) { struct rvu_switch *rswitch = &rvu->rswitch; u32 max = rswitch->used_entries; u16 entry; if (!rswitch->used_entries) return; for (entry = 0; entry < max; entry++) { if (rswitch->entry2pcifunc[entry] == pcifunc) break; } if (entry >= max) return; rvu_switch_install_tx_rule(rvu, pcifunc, rswitch->start_entry + entry); rvu_switch_install_rx_rule(rvu, pcifunc, 0x0); }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1