Contributors: 8
Author |
Tokens |
Token Proportion |
Commits |
Commit Proportion |
Raju Rangoju |
1002 |
81.27% |
4 |
20.00% |
Dimitris Michailidis |
133 |
10.79% |
2 |
10.00% |
Arjun V |
27 |
2.19% |
4 |
20.00% |
Hariprasad Shenai |
22 |
1.78% |
2 |
10.00% |
Ganesh Goudar |
18 |
1.46% |
3 |
15.00% |
Rahul Lakkireddy |
15 |
1.22% |
2 |
10.00% |
Kumar Sanghvi |
10 |
0.81% |
2 |
10.00% |
Yue haibing |
6 |
0.49% |
1 |
5.00% |
Total |
1233 |
|
20 |
|
// SPDX-License-Identifier: GPL-2.0
/* Copyright (c) 2019 Chelsio Communications, Inc. All rights reserved. */
#include "cxgb4.h"
static int cxgb4_mps_ref_dec_by_mac(struct adapter *adap,
const u8 *addr, const u8 *mask)
{
u8 bitmask[] = { 0xff, 0xff, 0xff, 0xff, 0xff, 0xff };
struct mps_entries_ref *mps_entry, *tmp;
int ret = -EINVAL;
spin_lock_bh(&adap->mps_ref_lock);
list_for_each_entry_safe(mps_entry, tmp, &adap->mps_ref, list) {
if (ether_addr_equal(mps_entry->addr, addr) &&
ether_addr_equal(mps_entry->mask, mask ? mask : bitmask)) {
if (!refcount_dec_and_test(&mps_entry->refcnt)) {
spin_unlock_bh(&adap->mps_ref_lock);
return -EBUSY;
}
list_del(&mps_entry->list);
kfree(mps_entry);
ret = 0;
break;
}
}
spin_unlock_bh(&adap->mps_ref_lock);
return ret;
}
static int cxgb4_mps_ref_dec(struct adapter *adap, u16 idx)
{
struct mps_entries_ref *mps_entry, *tmp;
int ret = -EINVAL;
spin_lock(&adap->mps_ref_lock);
list_for_each_entry_safe(mps_entry, tmp, &adap->mps_ref, list) {
if (mps_entry->idx == idx) {
if (!refcount_dec_and_test(&mps_entry->refcnt)) {
spin_unlock(&adap->mps_ref_lock);
return -EBUSY;
}
list_del(&mps_entry->list);
kfree(mps_entry);
ret = 0;
break;
}
}
spin_unlock(&adap->mps_ref_lock);
return ret;
}
static int cxgb4_mps_ref_inc(struct adapter *adap, const u8 *mac_addr,
u16 idx, const u8 *mask)
{
u8 bitmask[] = { 0xff, 0xff, 0xff, 0xff, 0xff, 0xff };
struct mps_entries_ref *mps_entry;
int ret = 0;
spin_lock_bh(&adap->mps_ref_lock);
list_for_each_entry(mps_entry, &adap->mps_ref, list) {
if (mps_entry->idx == idx) {
refcount_inc(&mps_entry->refcnt);
goto unlock;
}
}
mps_entry = kzalloc(sizeof(*mps_entry), GFP_ATOMIC);
if (!mps_entry) {
ret = -ENOMEM;
goto unlock;
}
ether_addr_copy(mps_entry->mask, mask ? mask : bitmask);
ether_addr_copy(mps_entry->addr, mac_addr);
mps_entry->idx = idx;
refcount_set(&mps_entry->refcnt, 1);
list_add_tail(&mps_entry->list, &adap->mps_ref);
unlock:
spin_unlock_bh(&adap->mps_ref_lock);
return ret;
}
int cxgb4_free_mac_filt(struct adapter *adap, unsigned int viid,
unsigned int naddr, const u8 **addr, bool sleep_ok)
{
int ret, i;
for (i = 0; i < naddr; i++) {
if (!cxgb4_mps_ref_dec_by_mac(adap, addr[i], NULL)) {
ret = t4_free_mac_filt(adap, adap->mbox, viid,
1, &addr[i], sleep_ok);
if (ret < 0)
return ret;
}
}
/* return number of filters freed */
return naddr;
}
int cxgb4_alloc_mac_filt(struct adapter *adap, unsigned int viid,
bool free, unsigned int naddr, const u8 **addr,
u16 *idx, u64 *hash, bool sleep_ok)
{
int ret, i;
ret = t4_alloc_mac_filt(adap, adap->mbox, viid, free,
naddr, addr, idx, hash, sleep_ok);
if (ret < 0)
return ret;
for (i = 0; i < naddr; i++) {
if (idx[i] != 0xffff) {
if (cxgb4_mps_ref_inc(adap, addr[i], idx[i], NULL)) {
ret = -ENOMEM;
goto error;
}
}
}
goto out;
error:
cxgb4_free_mac_filt(adap, viid, naddr, addr, sleep_ok);
out:
/* Returns a negative error number or the number of filters allocated */
return ret;
}
int cxgb4_update_mac_filt(struct port_info *pi, unsigned int viid,
int *tcam_idx, const u8 *addr,
bool persistent, u8 *smt_idx)
{
int ret;
ret = cxgb4_change_mac(pi, viid, tcam_idx,
addr, persistent, smt_idx);
if (ret < 0)
return ret;
cxgb4_mps_ref_inc(pi->adapter, addr, *tcam_idx, NULL);
return ret;
}
int cxgb4_free_raw_mac_filt(struct adapter *adap,
unsigned int viid,
const u8 *addr,
const u8 *mask,
unsigned int idx,
u8 lookup_type,
u8 port_id,
bool sleep_ok)
{
int ret = 0;
if (!cxgb4_mps_ref_dec(adap, idx))
ret = t4_free_raw_mac_filt(adap, viid, addr,
mask, idx, lookup_type,
port_id, sleep_ok);
return ret;
}
int cxgb4_alloc_raw_mac_filt(struct adapter *adap,
unsigned int viid,
const u8 *addr,
const u8 *mask,
unsigned int idx,
u8 lookup_type,
u8 port_id,
bool sleep_ok)
{
int ret;
ret = t4_alloc_raw_mac_filt(adap, viid, addr,
mask, idx, lookup_type,
port_id, sleep_ok);
if (ret < 0)
return ret;
if (cxgb4_mps_ref_inc(adap, addr, ret, mask)) {
ret = -ENOMEM;
t4_free_raw_mac_filt(adap, viid, addr,
mask, idx, lookup_type,
port_id, sleep_ok);
}
return ret;
}
int cxgb4_free_encap_mac_filt(struct adapter *adap, unsigned int viid,
int idx, bool sleep_ok)
{
int ret = 0;
if (!cxgb4_mps_ref_dec(adap, idx))
ret = t4_free_encap_mac_filt(adap, viid, idx, sleep_ok);
return ret;
}
int cxgb4_alloc_encap_mac_filt(struct adapter *adap, unsigned int viid,
const u8 *addr, const u8 *mask,
unsigned int vni, unsigned int vni_mask,
u8 dip_hit, u8 lookup_type, bool sleep_ok)
{
int ret;
ret = t4_alloc_encap_mac_filt(adap, viid, addr, mask, vni, vni_mask,
dip_hit, lookup_type, sleep_ok);
if (ret < 0)
return ret;
if (cxgb4_mps_ref_inc(adap, addr, ret, mask)) {
ret = -ENOMEM;
t4_free_encap_mac_filt(adap, viid, ret, sleep_ok);
}
return ret;
}
int cxgb4_init_mps_ref_entries(struct adapter *adap)
{
spin_lock_init(&adap->mps_ref_lock);
INIT_LIST_HEAD(&adap->mps_ref);
return 0;
}
void cxgb4_free_mps_ref_entries(struct adapter *adap)
{
struct mps_entries_ref *mps_entry, *tmp;
if (list_empty(&adap->mps_ref))
return;
spin_lock(&adap->mps_ref_lock);
list_for_each_entry_safe(mps_entry, tmp, &adap->mps_ref, list) {
list_del(&mps_entry->list);
kfree(mps_entry);
}
spin_unlock(&adap->mps_ref_lock);
}