Release 4.11 drivers/net/wireless/ath/ath9k/main.c
/*
* Copyright (c) 2008-2011 Atheros Communications Inc.
*
* Permission to use, copy, modify, and/or distribute this software for any
* purpose with or without fee is hereby granted, provided that the above
* copyright notice and this permission notice appear in all copies.
*
* THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
* WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
* MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
* ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
* WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
* ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
* OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
*/
#include <linux/nl80211.h>
#include <linux/delay.h>
#include "ath9k.h"
#include "btcoex.h"
u8 ath9k_parse_mpdudensity(u8 mpdudensity)
{
/*
* 802.11n D2.0 defined values for "Minimum MPDU Start Spacing":
* 0 for no restriction
* 1 for 1/4 us
* 2 for 1/2 us
* 3 for 1 us
* 4 for 2 us
* 5 for 4 us
* 6 for 8 us
* 7 for 16 us
*/
switch (mpdudensity) {
case 0:
return 0;
case 1:
case 2:
case 3:
/* Our lower layer calculations limit our precision to
1 microsecond */
return 1;
case 4:
return 2;
case 5:
return 4;
case 6:
return 8;
case 7:
return 16;
default:
return 0;
}
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Sujith Manoharan | 52 | 83.87% | 1 | 25.00% |
Johannes Berg | 6 | 9.68% | 1 | 25.00% |
Jouni Malinen | 3 | 4.84% | 1 | 25.00% |
Sven Eckelmann | 1 | 1.61% | 1 | 25.00% |
Total | 62 | 100.00% | 4 | 100.00% |
static bool ath9k_has_pending_frames(struct ath_softc *sc, struct ath_txq *txq,
bool sw_pending)
{
bool pending = false;
spin_lock_bh(&txq->axq_lock);
if (txq->axq_depth) {
pending = true;
goto out;
}
if (!sw_pending)
goto out;
if (txq->mac80211_qnum >= 0) {
struct ath_acq *acq;
acq = &sc->cur_chan->acq[txq->mac80211_qnum];
if (!list_empty(&acq->acq_new) || !list_empty(&acq->acq_old))
pending = true;
}
out:
spin_unlock_bh(&txq->axq_lock);
return pending;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Vasanthakumar Thiagarajan | 54 | 44.63% | 1 | 20.00% |
Felix Fietkau | 33 | 27.27% | 1 | 20.00% |
Sujith Manoharan | 18 | 14.88% | 2 | 40.00% |
Toke Höiland-Jörgensen | 16 | 13.22% | 1 | 20.00% |
Total | 121 | 100.00% | 5 | 100.00% |
static bool ath9k_setpower(struct ath_softc *sc, enum ath9k_power_mode mode)
{
unsigned long flags;
bool ret;
spin_lock_irqsave(&sc->sc_pm_lock, flags);
ret = ath9k_hw_setpower(sc->sc_ah, mode);
spin_unlock_irqrestore(&sc->sc_pm_lock, flags);
return ret;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Sujith Manoharan | 48 | 85.71% | 2 | 33.33% |
Jouni Malinen | 5 | 8.93% | 2 | 33.33% |
Luis R. Rodriguez | 2 | 3.57% | 1 | 16.67% |
Mohammed Shafi Shajakhan | 1 | 1.79% | 1 | 16.67% |
Total | 56 | 100.00% | 6 | 100.00% |
void ath_ps_full_sleep(unsigned long data)
{
struct ath_softc *sc = (struct ath_softc *) data;
struct ath_common *common = ath9k_hw_common(sc->sc_ah);
bool reset;
spin_lock(&common->cc_lock);
ath_hw_cycle_counters_update(common);
spin_unlock(&common->cc_lock);
ath9k_hw_setrxabort(sc->sc_ah, 1);
ath9k_hw_stopdmarecv(sc->sc_ah, &reset);
ath9k_hw_setpower(sc->sc_ah, ATH9K_PM_FULL_SLEEP);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Felix Fietkau | 85 | 100.00% | 1 | 100.00% |
Total | 85 | 100.00% | 1 | 100.00% |
void ath9k_ps_wakeup(struct ath_softc *sc)
{
struct ath_common *common = ath9k_hw_common(sc->sc_ah);
unsigned long flags;
enum ath9k_power_mode power_mode;
spin_lock_irqsave(&sc->sc_pm_lock, flags);
if (++sc->ps_usecount != 1)
goto unlock;
del_timer_sync(&sc->sleep_timer);
power_mode = sc->sc_ah->power_mode;
ath9k_hw_setpower(sc->sc_ah, ATH9K_PM_AWAKE);
/*
* While the hardware is asleep, the cycle counters contain no
* useful data. Better clear them now so that they don't mess up
* survey data results.
*/
if (power_mode != ATH9K_PM_AWAKE) {
spin_lock(&common->cc_lock);
ath_hw_cycle_counters_update(common);
memset(&common->cc_survey, 0, sizeof(common->cc_survey));
memset(&common->cc_ani, 0, sizeof(common->cc_ani));
spin_unlock(&common->cc_lock);
}
unlock:
spin_unlock_irqrestore(&sc->sc_pm_lock, flags);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Felix Fietkau | 79 | 51.63% | 3 | 50.00% |
Sujith Manoharan | 39 | 25.49% | 1 | 16.67% |
Jouni Malinen | 18 | 11.76% | 1 | 16.67% |
Rajkumar Manoharan | 17 | 11.11% | 1 | 16.67% |
Total | 153 | 100.00% | 6 | 100.00% |
void ath9k_ps_restore(struct ath_softc *sc)
{
struct ath_common *common = ath9k_hw_common(sc->sc_ah);
enum ath9k_power_mode mode;
unsigned long flags;
spin_lock_irqsave(&sc->sc_pm_lock, flags);
if (--sc->ps_usecount != 0)
goto unlock;
if (sc->ps_idle) {
mod_timer(&sc->sleep_timer, jiffies + HZ / 10);
goto unlock;
}
if (sc->ps_enabled &&
!(sc->ps_flags & (PS_WAIT_FOR_BEACON |
PS_WAIT_FOR_CAB |
PS_WAIT_FOR_PSPOLL_DATA |
PS_WAIT_FOR_TX_ACK |
PS_WAIT_FOR_ANI))) {
mode = ATH9K_PM_NETWORK_SLEEP;
if (ath9k_hw_btcoex_is_enabled(sc->sc_ah))
ath9k_btcoex_stop_gen_timer(sc);
} else {
goto unlock;
}
spin_lock(&common->cc_lock);
ath_hw_cycle_counters_update(common);
spin_unlock(&common->cc_lock);
ath9k_hw_setpower(sc->sc_ah, mode);
unlock:
spin_unlock_irqrestore(&sc->sc_pm_lock, flags);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Sujith Manoharan | 73 | 42.94% | 5 | 35.71% |
Felix Fietkau | 57 | 33.53% | 4 | 28.57% |
Jouni Malinen | 19 | 11.18% | 1 | 7.14% |
Rajkumar Manoharan | 16 | 9.41% | 2 | 14.29% |
Vivek Natarajan | 4 | 2.35% | 1 | 7.14% |
Luis R. Rodriguez | 1 | 0.59% | 1 | 7.14% |
Total | 170 | 100.00% | 14 | 100.00% |
static void __ath_cancel_work(struct ath_softc *sc)
{
cancel_work_sync(&sc->paprd_work);
cancel_delayed_work_sync(&sc->hw_check_work);
cancel_delayed_work_sync(&sc->hw_pll_work);
#ifdef CONFIG_ATH9K_BTCOEX_SUPPORT
if (ath9k_hw_mci_is_enabled(sc->sc_ah))
cancel_work_sync(&sc->mci_work);
#endif
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Sujith Manoharan | 29 | 50.88% | 3 | 50.00% |
Felix Fietkau | 20 | 35.09% | 2 | 33.33% |
Rajkumar Manoharan | 8 | 14.04% | 1 | 16.67% |
Total | 57 | 100.00% | 6 | 100.00% |
void ath_cancel_work(struct ath_softc *sc)
{
__ath_cancel_work(sc);
cancel_work_sync(&sc->hw_reset_work);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Felix Fietkau | 15 | 65.22% | 2 | 66.67% |
Sujith Manoharan | 8 | 34.78% | 1 | 33.33% |
Total | 23 | 100.00% | 3 | 100.00% |
void ath_restart_work(struct ath_softc *sc)
{
ieee80211_queue_delayed_work(sc->hw, &sc->hw_check_work,
ATH_HW_CHECK_POLL_INT);
if (AR_SREV_9340(sc->sc_ah) || AR_SREV_9330(sc->sc_ah))
ieee80211_queue_delayed_work(sc->hw, &sc->hw_pll_work,
msecs_to_jiffies(ATH_PLL_WORK_INTERVAL));
ath_start_ani(sc);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Sujith Manoharan | 53 | 85.48% | 3 | 60.00% |
Gabor Juhos | 7 | 11.29% | 1 | 20.00% |
Felix Fietkau | 2 | 3.23% | 1 | 20.00% |
Total | 62 | 100.00% | 5 | 100.00% |
static bool ath_prepare_reset(struct ath_softc *sc)
{
struct ath_hw *ah = sc->sc_ah;
bool ret = true;
ieee80211_stop_queues(sc->hw);
ath_stop_ani(sc);
ath9k_hw_disable_interrupts(ah);
if (AR_SREV_9300_20_OR_LATER(ah)) {
ret &= ath_stoprecv(sc);
ret &= ath_drain_all_txq(sc);
} else {
ret &= ath_drain_all_txq(sc);
ret &= ath_stoprecv(sc);
}
return ret;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Felix Fietkau | 60 | 70.59% | 5 | 71.43% |
Sujith Manoharan | 25 | 29.41% | 2 | 28.57% |
Total | 85 | 100.00% | 7 | 100.00% |
static bool ath_complete_reset(struct ath_softc *sc, bool start)
{
struct ath_hw *ah = sc->sc_ah;
struct ath_common *common = ath9k_hw_common(ah);
unsigned long flags;
ath9k_calculate_summary_state(sc, sc->cur_chan);
ath_startrecv(sc);
ath9k_cmn_update_txpow(ah, sc->cur_chan->cur_txpower,
sc->cur_chan->txpower,
&sc->cur_chan->cur_txpower);
clear_bit(ATH_OP_HW_RESET, &common->op_flags);
if (!sc->cur_chan->offchannel && start) {
/* restore per chanctx TSF timer */
if (sc->cur_chan->tsf_val) {
u32 offset;
offset = ath9k_hw_get_tsf_offset(&sc->cur_chan->tsf_ts,
NULL);
ath9k_hw_settsf64(ah, sc->cur_chan->tsf_val + offset);
}
if (!test_bit(ATH_OP_BEACONS, &common->op_flags))
goto work;
if (ah->opmode == NL80211_IFTYPE_STATION &&
test_bit(ATH_OP_PRIM_STA_VIF, &common->op_flags)) {
spin_lock_irqsave(&sc->sc_pm_lock, flags);
sc->ps_flags |= PS_BEACON_SYNC | PS_WAIT_FOR_BEACON;
spin_unlock_irqrestore(&sc->sc_pm_lock, flags);
} else {
ath9k_set_beacon(sc);
}
work:
ath_restart_work(sc);
ath_txq_schedule_all(sc);
}
sc->gtt_cnt = 0;
ath9k_hw_set_interrupts(ah);
ath9k_hw_enable_interrupts(ah);
ieee80211_wake_queues(sc->hw);
ath9k_p2p_ps_timer(sc);
return true;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Felix Fietkau | 127 | 50.00% | 12 | 44.44% |
Sujith Manoharan | 100 | 39.37% | 10 | 37.04% |
Rajkumar Manoharan | 12 | 4.72% | 2 | 7.41% |
Oleksij Rempel | 9 | 3.54% | 1 | 3.70% |
Luis R. Rodriguez | 5 | 1.97% | 1 | 3.70% |
Senthil Balasubramanian | 1 | 0.39% | 1 | 3.70% |
Total | 254 | 100.00% | 27 | 100.00% |
static int ath_reset_internal(struct ath_softc *sc, struct ath9k_channel *hchan)
{
struct ath_hw *ah = sc->sc_ah;
struct ath_common *common = ath9k_hw_common(ah);
struct ath9k_hw_cal_data *caldata = NULL;
bool fastcc = true;
int r;
__ath_cancel_work(sc);
disable_irq(sc->irq);
tasklet_disable(&sc->intr_tq);
tasklet_disable(&sc->bcon_tasklet);
spin_lock_bh(&sc->sc_pcu_lock);
if (!sc->cur_chan->offchannel) {
fastcc = false;
caldata = &sc->cur_chan->caldata;
}
if (!hchan) {
fastcc = false;
hchan = ah->curchan;
}
if (!ath_prepare_reset(sc))
fastcc = false;
if (ath9k_is_chanctx_enabled())
fastcc = false;
spin_lock_bh(&sc->chan_lock);
sc->cur_chandef = sc->cur_chan->chandef;
spin_unlock_bh(&sc->chan_lock);
ath_dbg(common, CONFIG, "Reset to %u MHz, HT40: %d fastcc: %d\n",
hchan->channel, IS_CHAN_HT40(hchan), fastcc);
r = ath9k_hw_reset(ah, hchan, caldata, fastcc);
if (r) {
ath_err(common,
"Unable to reset channel, reset status %d\n", r);
ath9k_hw_enable_interrupts(ah);
ath9k_queue_reset(sc, RESET_TYPE_BB_HANG);
goto out;
}
if (ath9k_hw_mci_is_enabled(sc->sc_ah) &&
sc->cur_chan->offchannel)
ath9k_mci_set_txpower(sc, true, false);
if (!ath_complete_reset(sc, true))
r = -EIO;
out:
enable_irq(sc->irq);
spin_unlock_bh(&sc->sc_pcu_lock);
tasklet_enable(&sc->bcon_tasklet);
tasklet_enable(&sc->intr_tq);
return r;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Felix Fietkau | 178 | 57.05% | 9 | 34.62% |
Sujith Manoharan | 82 | 26.28% | 7 | 26.92% |
Rajkumar Manoharan | 27 | 8.65% | 3 | 11.54% |
Robert Shade | 12 | 3.85% | 1 | 3.85% |
Senthil Balasubramanian | 7 | 2.24% | 1 | 3.85% |
Luis R. Rodriguez | 3 | 0.96% | 2 | 7.69% |
Joe Perches | 3 | 0.96% | 3 | 11.54% |
Total | 312 | 100.00% | 26 | 100.00% |
static void ath_node_attach(struct ath_softc *sc, struct ieee80211_sta *sta,
struct ieee80211_vif *vif)
{
struct ath_node *an;
an = (struct ath_node *)sta->drv_priv;
an->sc = sc;
an->sta = sta;
an->vif = vif;
memset(&an->key_idx, 0, sizeof(an->key_idx));
ath_tx_node_init(sc, an);
ath_dynack_node_init(sc->sc_ah, an);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Sujith Manoharan | 44 | 50.00% | 2 | 33.33% |
Felix Fietkau | 18 | 20.45% | 2 | 33.33% |
Rajkumar Manoharan | 17 | 19.32% | 1 | 16.67% |
Lorenzo Bianconi | 9 | 10.23% | 1 | 16.67% |
Total | 88 | 100.00% | 6 | 100.00% |
static void ath_node_detach(struct ath_softc *sc, struct ieee80211_sta *sta)
{
struct ath_node *an = (struct ath_node *)sta->drv_priv;
ath_tx_node_cleanup(sc, an);
ath_dynack_node_deinit(sc->sc_ah, an);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Sujith Manoharan | 21 | 45.65% | 1 | 33.33% |
Felix Fietkau | 16 | 34.78% | 1 | 33.33% |
Lorenzo Bianconi | 9 | 19.57% | 1 | 33.33% |
Total | 46 | 100.00% | 3 | 100.00% |
void ath9k_tasklet(unsigned long data)
{
struct ath_softc *sc = (struct ath_softc *)data;
struct ath_hw *ah = sc->sc_ah;
struct ath_common *common = ath9k_hw_common(ah);
enum ath_reset_type type;
unsigned long flags;
u32 status;
u32 rxmask;
spin_lock_irqsave(&sc->intr_lock, flags);
status = sc->intrstatus;
sc->intrstatus = 0;
spin_unlock_irqrestore(&sc->intr_lock, flags);
ath9k_ps_wakeup(sc);
spin_lock(&sc->sc_pcu_lock);
if (status & ATH9K_INT_FATAL) {
type = RESET_TYPE_FATAL_INT;
ath9k_queue_reset(sc, type);
ath_dbg(common, RESET, "FATAL: Skipping interrupts\n");
goto out;
}
if ((ah->config.hw_hang_checks & HW_BB_WATCHDOG) &&
(status & ATH9K_INT_BB_WATCHDOG)) {
spin_lock(&common->cc_lock);
ath_hw_cycle_counters_update(common);
ar9003_hw_bb_watchdog_dbg_info(ah);
spin_unlock(&common->cc_lock);
if (ar9003_hw_bb_watchdog_check(ah)) {
type = RESET_TYPE_BB_WATCHDOG;
ath9k_queue_reset(sc, type);
ath_dbg(common, RESET,
"BB_WATCHDOG: Skipping interrupts\n");
goto out;
}
}
if (status & ATH9K_INT_GTT) {
sc->gtt_cnt++;
if ((sc->gtt_cnt >= MAX_GTT_CNT) && !ath9k_hw_check_alive(ah)) {
type = RESET_TYPE_TX_GTT;
ath9k_queue_reset(sc, type);
ath_dbg(common, RESET,
"GTT: Skipping interrupts\n");
goto out;
}
}
spin_lock_irqsave(&sc->sc_pm_lock, flags);
if ((status & ATH9K_INT_TSFOOR) && sc->ps_enabled) {
/*
* TSF sync does not look correct; remain awake to sync with
* the next Beacon.
*/
ath_dbg(common, PS, "TSFOOR - Sync with next Beacon\n");
sc->ps_flags |= PS_WAIT_FOR_BEACON | PS_BEACON_SYNC;
}
spin_unlock_irqrestore(&sc->sc_pm_lock, flags);
if (ah->caps.hw_caps & ATH9K_HW_CAP_EDMA)
rxmask = (ATH9K_INT_RXHP | ATH9K_INT_RXLP | ATH9K_INT_RXEOL |
ATH9K_INT_RXORN);
else
rxmask = (ATH9K_INT_RX | ATH9K_INT_RXEOL | ATH9K_INT_RXORN);
if (status & rxmask) {
/* Check for high priority Rx first */
if ((ah->caps.hw_caps & ATH9K_HW_CAP_EDMA) &&
(status & ATH9K_INT_RXHP))
ath_rx_tasklet(sc, 0, true);
ath_rx_tasklet(sc, 0, false);
}
if (status & ATH9K_INT_TX) {
if (ah->caps.hw_caps & ATH9K_HW_CAP_EDMA) {
/*
* For EDMA chips, TX completion is enabled for the
* beacon queue, so if a beacon has been transmitted
* successfully after a GTT interrupt, the GTT counter
* gets reset to zero here.
*/
sc->gtt_cnt = 0;
ath_tx_edma_tasklet(sc);
} else {
ath_tx_tasklet(sc);
}
wake_up(&sc->tx_wait);
}
if (status & ATH9K_INT_GENTIMER)
ath_gen_timer_isr(sc->sc_ah);
ath9k_btcoex_handle_interrupt(sc, status);
/* re-enable hardware interrupt */
ath9k_hw_resume_interrupts(ah);
out:
spin_unlock(&sc->sc_pcu_lock);
ath9k_ps_restore(sc);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Sujith Manoharan | 346 | 71.93% | 12 | 35.29% |
Felix Fietkau | 82 | 17.05% | 11 | 32.35% |
Luis R. Rodriguez | 25 | 5.20% | 6 | 17.65% |
Mohammed Shafi Shajakhan | 19 | 3.95% | 2 | 5.88% |
Rajkumar Manoharan | 7 | 1.46% | 2 | 5.88% |
Gabor Juhos | 2 | 0.42% | 1 | 2.94% |
Total | 481 | 100.00% | 34 | 100.00% |
irqreturn_t ath_isr(int irq, void *dev)
{
#define SCHED_INTR ( \
ATH9K_INT_FATAL | \
ATH9K_INT_BB_WATCHDOG | \
ATH9K_INT_RXORN | \
ATH9K_INT_RXEOL | \
ATH9K_INT_RX | \
ATH9K_INT_RXLP | \
ATH9K_INT_RXHP | \
ATH9K_INT_TX | \
ATH9K_INT_BMISS | \
ATH9K_INT_CST | \
ATH9K_INT_GTT | \
ATH9K_INT_TSFOOR | \
ATH9K_INT_GENTIMER | \
ATH9K_INT_MCI)
struct ath_softc *sc = dev;
struct ath_hw *ah = sc->sc_ah;
struct ath_common *common = ath9k_hw_common(ah);
enum ath9k_int status;
u32 sync_cause = 0;
bool sched = false;
/*
* The hardware is not ready/present, don't
* touch anything. Note this can happen early
* on if the IRQ is shared.
*/
if (!ah || test_bit(ATH_OP_INVALID, &common->op_flags))
return IRQ_NONE;
/* shared irq, not for us */
if (!ath9k_hw_intrpend(ah))
return IRQ_NONE;
/*
* Figure out the reason(s) for the interrupt. Note
* that the hal returns a pseudo-ISR that may include
* bits we haven't explicitly enabled so we mask the
* value to insure we only process bits we requested.
*/
ath9k_hw_getisr(ah, &status, &sync_cause); /* NB: clears ISR too */
ath9k_debug_sync_cause(sc, sync_cause);
status &= ah->imask; /* discard unasked-for bits */
if (test_bit(ATH_OP_HW_RESET, &common->op_flags))
return IRQ_HANDLED;
/*
* If there are no status bits set, then this interrupt was not
* for me (should have been caught above).
*/
if (!status)
return IRQ_NONE;
/* Cache the status */
spin_lock(&sc->intr_lock);
sc->intrstatus |= status;
spin_unlock(&sc->intr_lock);
if (status & SCHED_INTR)
sched = true;
/*
* If a FATAL interrupt is received, we have to reset the chip
* immediately.
*/
if (status & ATH9K_INT_FATAL)
goto chip_reset;
if ((ah->config.hw_hang_checks & HW_BB_WATCHDOG) &&
(status & ATH9K_INT_BB_WATCHDOG))
goto chip_reset;
if (status & ATH9K_INT_SWBA)
tasklet_schedule(&sc->bcon_tasklet);
if (status & ATH9K_INT_TXURN)
ath9k_hw_updatetxtriglevel(ah, true);
if (status & ATH9K_INT_RXEOL) {
ah->imask &= ~(ATH9K_INT_RXEOL | ATH9K_INT_RXORN);
ath9k_hw_set_interrupts(ah);
}
if (!(ah->caps.hw_caps & ATH9K_HW_CAP_AUTOSLEEP))
if (status & ATH9K_INT_TIM_TIMER) {
if (ATH_DBG_WARN_ON_ONCE(sc->ps_idle))
goto chip_reset;
/* Clear RxAbort bit so that we can
* receive frames */
ath9k_setpower(sc, ATH9K_PM_AWAKE);
spin_lock(&sc->sc_pm_lock);
ath9k_hw_setrxabort(sc->sc_ah, 0);
sc->ps_flags |= PS_WAIT_FOR_BEACON;
spin_unlock(&sc->sc_pm_lock);
}
chip_reset:
ath_debug_stat_interrupt(sc, status);
if (sched) {
/* turn off every interrupt */
ath9k_hw_kill_interrupts(ah);
tasklet_schedule(&sc->intr_tq);
}
return IRQ_HANDLED;
#undef SCHED_INTR
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Sujith Manoharan | 227 | 63.06% | 7 | 38.89% |
Felix Fietkau | 92 | 25.56% | 8 | 44.44% |
Rajkumar Manoharan | 25 | 6.94% | 1 | 5.56% |
Oleksij Rempel | 13 | 3.61% | 1 | 5.56% |
Wojciech Dubowik | 3 | 0.83% | 1 | 5.56% |
Total | 360 | 100.00% | 18 | 100.00% |
/*
* This function is called when a HW reset cannot be deferred
* and has to be immediate.
*/
int ath_reset(struct ath_softc *sc, struct ath9k_channel *hchan)
{
struct ath_common *common = ath9k_hw_common(sc->sc_ah);
int r;
ath9k_hw_kill_interrupts(sc->sc_ah);
set_bit(ATH_OP_HW_RESET, &common->op_flags);
ath9k_ps_wakeup(sc);
r = ath_reset_internal(sc, hchan);
ath9k_ps_restore(sc);
return r;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Sujith Manoharan | 48 | 69.57% | 3 | 60.00% |
Felix Fietkau | 21 | 30.43% | 2 | 40.00% |
Total | 69 | 100.00% | 5 | 100.00% |
/*
* When a HW reset can be deferred, it is added to the
* hw_reset_work workqueue, but we set ATH_OP_HW_RESET before
* queueing.
*/
void ath9k_queue_reset(struct ath_softc *sc, enum ath_reset_type type)
{
struct ath_common *common = ath9k_hw_common(sc->sc_ah);
#ifdef CONFIG_ATH9K_DEBUGFS
RESET_STAT_INC(sc, type);
#endif
ath9k_hw_kill_interrupts(sc->sc_ah);
set_bit(ATH_OP_HW_RESET, &common->op_flags);
ieee80211_queue_work(sc->hw, &sc->hw_reset_work);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Rajkumar Manoharan | 45 | 67.16% | 1 | 33.33% |
Oleksij Rempel | 15 | 22.39% | 1 | 33.33% |
Felix Fietkau | 7 | 10.45% | 1 | 33.33% |
Total | 67 | 100.00% | 3 | 100.00% |
void ath_reset_work(struct work_struct *work)
{
struct ath_softc *sc = container_of(work, struct ath_softc, hw_reset_work);
ath9k_ps_wakeup(sc);
ath_reset_internal(sc, NULL);
ath9k_ps_restore(sc);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Felix Fietkau | 27 | 64.29% | 1 | 33.33% |
Sujith Manoharan | 15 | 35.71% | 2 | 66.67% |
Total | 42 | 100.00% | 3 | 100.00% |
/**********************/
/* mac80211 callbacks */
/**********************/
static int ath9k_start(struct ieee80211_hw *hw)
{
struct ath_softc *sc = hw->priv;
struct ath_hw *ah = sc->sc_ah;
struct ath_common *common = ath9k_hw_common(ah);
struct ieee80211_channel *curchan = sc->cur_chan->chandef.chan;
struct ath_chanctx *ctx = sc->cur_chan;
struct ath9k_channel *init_channel;
int r;
ath_dbg(common, CONFIG,
"Starting driver with initial channel: %d MHz\n",
curchan->center_freq);
ath9k_ps_wakeup(sc);
mutex_lock(&sc->mutex);
init_channel = ath9k_cmn_get_channel(hw, ah, &ctx->chandef);
sc->cur_chandef = hw->conf.chandef;
/* Reset SERDES registers */
ath9k_hw_configpcipowersave(ah, false);
/*
* The basic interface to setting the hardware in a good
* state is ``reset''. On return the hardware is known to
* be powered up and with interrupts disabled. This must
* be followed by initialization of the appropriate bits
* and then setup of the interrupt mask.
*/
spin_lock_bh(&sc->sc_pcu_lock);
atomic_set(&ah->intr_ref_cnt, -1);
r = ath9k_hw_reset(ah, init_channel, ah->caldata, false);
if (r) {
ath_err(common,
"Unable to reset hardware; reset status %d (freq %u MHz)\n",
r, curchan->center_freq);
ah->reset_power_on = false;
}
/* Setup our intr mask. */
ah->imask = ATH9K_INT_TX | ATH9K_INT_RXEOL |
ATH9K_INT_RXORN | ATH9K_INT_FATAL |
ATH9K_INT_GLOBAL;
if (ah->caps.hw_caps & ATH9K_HW_CAP_EDMA)
ah->imask |= ATH9K_INT_RXHP |
ATH9K_INT_RXLP;
else
ah->imask |= ATH9K_INT_RX;
if (ah->config.hw_hang_checks & HW_BB_WATCHDOG)
ah->imask |= ATH9K_INT_BB_WATCHDOG;
/*
* Enable GTT interrupts only for AR9003/AR9004 chips
* for now.
*/
if (AR_SREV_9300_20_OR_LATER(ah))
ah->imask |= ATH9K_INT_GTT;
if (ah->caps.hw_caps & ATH9K_HW_CAP_HT)
ah->imask |= ATH9K_INT_CST;
ath_mci_enable(sc);
clear_bit(ATH_OP_INVALID, &common->op_flags);
sc->sc_ah->is_monitoring = false;
if (!ath_complete_reset(sc, false))
ah->reset_power_on = false;
if (ah->led_pin >= 0) {
ath9k_hw_set_gpio(ah, ah->led_pin,
(ah->config.led_active_high) ? 1 : 0);
ath9k_hw_gpio_request_out(ah, ah->led_pin, NULL,
AR_GPIO_OUTPUT_MUX_AS_OUTPUT);
}
/*
* Reset key cache to sane defaults (all entries cleared) instead of
* semi-random values after suspend/resume.
*/
ath9k_cmn_init_crypto(sc->sc_ah);
ath9k_hw_reset_tsf(ah);
spin_unlock_bh(&sc->sc_pcu_lock);
mutex_unlock(&sc->mutex);
ath9k_ps_restore(sc);
ath9k_rng_start(sc);
return 0;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Sujith Manoharan | 139 | 34.92% | 9 | 19.57% |
Felix Fietkau | 120 | 30.15% | 12 | 26.09% |
Luis R. Rodriguez | 61 | 15.33% | 7 | 15.22% |
Vasanthakumar Thiagarajan | 17 | 4.27% | 2 | 4.35% |
Giedrius Statkevičius | 15 | 3.77% | 1 | 2.17% |
Rajkumar Manoharan | 12 | 3.02% | 3 | 6.52% |
Jouni Malinen | 8 | 2.01% | 2 | 4.35% |
Senthil Balasubramanian | 6 | 1.51% | 1 | 2.17% |
Miaoqing Pan | 5 | 1.26% | 1 | 2.17% |
Joe Perches | 4 | 1.01% | 3 | 6.52% |
Pavel Roskin | 3 | 0.75% | 1 | 2.17% |
Karl Beldan | 3 | 0.75% | 1 | 2.17% |
Oleksij Rempel | 3 | 0.75% | 1 | 2.17% |
Stanislaw Gruszka | 1 | 0.25% | 1 | 2.17% |
Mohammed Shafi Shajakhan | 1 | 0.25% | 1 | 2.17% |
Total | 398 | 100.00% | 46 | 100.00% |
static void ath9k_tx(struct ieee80211_hw *hw,
struct ieee80211_tx_control *control,
struct sk_buff *skb)
{
struct ath_softc *sc = hw->priv;
struct ath_common *common = ath9k_hw_common(sc->sc_ah);
struct ath_tx_control txctl;
struct ieee80211_hdr *hdr = (struct ieee80211_hdr *) skb->data;
unsigned long flags;
if (sc->ps_enabled) {
/*
* mac80211 does not set PM field for normal data frames, so we
* need to update that based on the current PS mode.
*/
if (ieee80211_is_data(hdr->frame_control) &&
!ieee80211_is_nullfunc(hdr->frame_control) &&
!ieee80211_has_pm(hdr->frame_control)) {
ath_dbg(common, PS,
"Add PM=1 for a TX frame while in PS mode\n");
hdr->frame_control |= cpu_to_le16(IEEE80211_FCTL_PM);
}
}
if (unlikely(sc->sc_ah->power_mode == ATH9K_PM_NETWORK_SLEEP)) {
/*
* We are using PS-Poll and mac80211 can request TX while in
* power save mode. Need to wake up hardware for the TX to be
* completed and if needed, also for RX of buffered frames.
*/
ath9k_ps_wakeup(sc);
spin_lock_irqsave(&sc->sc_pm_lock, flags);
if (!(sc->sc_ah->caps.hw_caps & ATH9K_HW_CAP_AUTOSLEEP))
ath9k_hw_setrxabort(sc->sc_ah, 0);
if (ieee80211_is_pspoll(hdr->frame_control)) {
ath_dbg(common, PS,
"Sending PS-Poll to pick a buffered frame\n");
sc->ps_flags |= PS_WAIT_FOR_PSPOLL_DATA;
} else {
ath_dbg(common, PS, "Wake up to complete TX\n");
sc->ps_flags |= PS_WAIT_FOR_TX_ACK;
}
/*
* The actual restore operation will happen only after
* the ps_flags bit is cleared. We are just dropping
* the ps_usecount here.
*/
spin_unlock_irqrestore(&sc->sc_pm_lock, flags);
ath9k_ps_restore(sc);
}
/*
* Cannot tx while the hardware is in full sleep, it first needs a full
* chip reset to recover from that
*/
if (unlikely(sc->sc_ah->power_mode == ATH9K_PM_FULL_SLEEP)) {
ath_err(common, "TX while HW is in FULL_SLEEP mode\n");
goto exit;
}
memset(&txctl, 0, sizeof(struct ath_tx_control));
txctl.txq = sc->tx.txq_map[skb_get_queue_mapping(skb)];
txctl.sta = control->sta;
ath_dbg(common, XMIT, "transmitting packet, skb: %p\n", skb);
if (ath_tx_start(hw, skb, &txctl) != 0) {
ath_dbg(common, XMIT, "TX failed\n");
TX_STAT_INC(txctl.txq->axq_qnum, txfailed);
goto exit;
}
return;
exit:
ieee80211_free_txskb(hw, skb);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Jouni Malinen | 114 | 31.93% | 5 | 19.23% |
Sujith Manoharan | 85 | 23.81% | 5 | 19.23% |
Luis R. Rodriguez | 44 | 12.32% | 2 | 7.69% |
Vasanthakumar Thiagarajan | 34 | 9.52% | 2 | 7.69% |
Felix Fietkau | 28 | 7.84% | 5 | 19.23% |
Benoit Papillault | 14 | 3.92% | 1 | 3.85% |
Thomas Huehn | 13 | 3.64% | 1 | 3.85% |
Joe Perches | 11 | 3.08% | 2 | 7.69% |
Ben Greear | 11 | 3.08% | 1 | 3.85% |
Johannes Berg | 2 | 0.56% | 1 | 3.85% |
Gabor Juhos | 1 | 0.28% | 1 | 3.85% |
Total | 357 | 100.00% | 26 | 100.00% |
static void ath9k_stop(struct ieee80211_hw *hw)
{
struct ath_softc *sc = hw->priv;
struct ath_hw *ah = sc->sc_ah;
struct ath_common *common = ath9k_hw_common(ah);
bool prev_idle;
ath9k_deinit_channel_context(sc);
ath9k_rng_stop(sc);
mutex_lock(&sc->mutex);
ath_cancel_work(sc);
if (test_bit(ATH_OP_INVALID, &common->op_flags)) {
ath_dbg(common, ANY, "Device not present\n");
mutex_unlock(&sc->mutex);
return;
}
/* Ensure HW is awake when we try to shut it down. */
ath9k_ps_wakeup(sc);
spin_lock_bh(&sc->sc_pcu_lock);
/* prevent tasklets to enable interrupts once we disable them */
ah->imask &= ~ATH9K_INT_GLOBAL;
/* make sure h/w will not generate any interrupt
* before setting the invalid flag. */
ath9k_hw_disable_interrupts(ah);
spin_unlock_bh(&sc->sc_pcu_lock);
/* we can now sync irq and kill any running tasklets, since we already
* disabled interrupts and not holding a spin lock */
synchronize_irq(sc->irq);
tasklet_kill(&sc->intr_tq);
tasklet_kill(&sc->bcon_tasklet);
prev_idle = sc->ps_idle;
sc->ps_idle = true;
spin_lock_bh(&sc->sc_pcu_lock);
if (ah->led_pin >= 0) {
ath9k_hw_set_gpio(ah, ah->led_pin,
(ah->config.led_active_high) ? 0 : 1);
ath9k_hw_gpio_request_in(ah, ah->led_pin, NULL);
}
ath_prepare_reset(sc);
if (sc->rx.frag) {
dev_kfree_skb_any(sc->rx.frag);
sc->rx.frag = NULL;
}
if (!ah->curchan)
ah->curchan = ath9k_cmn_get_channel(hw, ah,
&sc->cur_chan->chandef);
ath9k_hw_reset(ah, ah->curchan, ah->caldata, false);
set_bit(ATH_OP_INVALID, &common->op_flags);
ath9k_hw_phy_disable(ah);
ath9k_hw_configpcipowersave(ah, true);
spin_unlock_bh(&sc->sc_pcu_lock);
ath9k_ps_restore(sc);
sc->ps_idle = prev_idle;
mutex_unlock(&sc->mutex);
ath_dbg(common, CONFIG, "Driver halt\n");
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Felix Fietkau | 161 | 46.13% | 10 | 32.26% |
Sujith Manoharan | 79 | 22.64% | 9 | 29.03% |
Luis R. Rodriguez | 42 | 12.03% | 4 | 12.90% |
Stanislaw Gruszka | 19 | 5.44% | 1 | 3.23% |
Vasanthakumar Thiagarajan | 17 | 4.87% | 1 | 3.23% |
Giedrius Statkevičius | 13 | 3.72% | 1 | 3.23% |
Jouni Malinen | 6 | 1.72% | 1 | 3.23% |
Miaoqing Pan | 5 | 1.43% | 1 | 3.23% |
Joe Perches | 4 | 1.15% | 2 | 6.45% |
Oleksij Rempel | 3 | 0.86% | 1 | 3.23% |
Total | 349 | 100.00% | 31 | 100.00% |
static bool ath9k_uses_beacons(int type)
{
switch (type) {
case NL80211_IFTYPE_AP:
case NL80211_IFTYPE_ADHOC:
case NL80211_IFTYPE_MESH_POINT:
return true;
default:
return false;
}
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Ben Greear | 29 | 93.55% | 1 | 33.33% |
Felix Fietkau | 1 | 3.23% | 1 | 33.33% |
Vasanthakumar Thiagarajan | 1 | 3.23% | 1 | 33.33% |
Total | 31 | 100.00% | 3 | 100.00% |
static void ath9k_vif_iter_set_beacon(struct ath9k_vif_iter_data *iter_data,
struct ieee80211_vif *vif)
{
/* Use the first (configured) interface, but prefering AP interfaces. */
if (!iter_data->primary_beacon_vif) {
iter_data->primary_beacon_vif = vif;
} else {
if (iter_data->primary_beacon_vif->type != NL80211_IFTYPE_AP &&
vif->type == NL80211_IFTYPE_AP)
iter_data->primary_beacon_vif = vif;
}
iter_data->beacons = true;
iter_data->nbcnvifs += 1;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Benjamin Berg | 69 | 100.00% | 1 | 100.00% |
Total | 69 | 100.00% | 1 | 100.00% |
static void ath9k_vif_iter(struct ath9k_vif_iter_data *iter_data,
u8 *mac, struct ieee80211_vif *vif)
{
struct ath_vif *avp = (struct ath_vif *)vif->drv_priv;
int i;
if (iter_data->has_hw_macaddr) {
for (i = 0; i < ETH_ALEN; i++)
iter_data->mask[i] &=
~(iter_data->hw_macaddr[i] ^ mac[i]);
} else {
memcpy(iter_data->hw_macaddr, mac, ETH_ALEN);
iter_data->has_hw_macaddr = true;
}
if (!vif->bss_conf.use_short_slot)
iter_data->slottime = 20;
switch (vif->type) {
case NL80211_IFTYPE_AP:
iter_data->naps++;
if (vif->bss_conf.enable_beacon)
ath9k_vif_iter_set_beacon(iter_data, vif);
break;
case NL80211_IFTYPE_STATION:
iter_data->nstations++;
if (avp->assoc && !iter_data->primary_sta)
iter_data->primary_sta = vif;
break;
case NL80211_IFTYPE_OCB:
iter_data->nocbs++;
break;
case NL80211_IFTYPE_ADHOC:
iter_data->nadhocs++;
if (vif->bss_conf.enable_beacon)
ath9k_vif_iter_set_beacon(iter_data, vif);
break;
case NL80211_IFTYPE_MESH_POINT:
iter_data->nmeshes++;
if (vif->bss_conf.enable_beacon)
ath9k_vif_iter_set_beacon(iter_data, vif);
break;
case NL80211_IFTYPE_WDS:
iter_data->nwds++;
break;
default:
break;
}
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Ben Greear | 115 | 47.72% | 1 | 12.50% |
Rajkumar Manoharan | 50 | 20.75% | 1 | 12.50% |
Benjamin Berg | 26 | 10.79% | 2 | 25.00% |
Felix Fietkau | 23 | 9.54% | 1 | 12.50% |
Sujith Manoharan | 18 | 7.47% | 2 | 25.00% |
Jan Kaisrlik | 9 | 3.73% | 1 | 12.50% |
Total | 241 | 100.00% | 8 | 100.00% |
static void ath9k_update_bssid_mask(struct ath_softc *sc,
struct ath_chanctx *ctx,
struct ath9k_vif_iter_data *iter_data)
{
struct ath_common *common = ath9k_hw_common(sc->sc_ah);
struct ath_vif *avp;
int i;
if (!ath9k_is_chanctx_enabled())
return;
list_for_each_entry(avp, &ctx->vifs, list) {
if (ctx->nvifs_assigned != 1)
continue;
if (!iter_data->has_hw_macaddr)
continue;
ether_addr_copy(common->curbssid, avp->bssid);
/* perm_addr will be used as the p2p device address. */
for (i = 0; i < ETH_ALEN; i++)
iter_data->mask[i] &=
~(iter_data->hw_macaddr[i] ^
sc->hw->wiphy->perm_addr[i]);
}
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Sujith Manoharan | 128 | 100.00% | 1 | 100.00% |
Total | 128 | 100.00% | 1 | 100.00% |
/* Called with sc->mutex held. */
void ath9k_calculate_iter_data(struct ath_softc *sc,
struct ath_chanctx *ctx,
struct ath9k_vif_iter_data *iter_data)
{
struct ath_vif *avp;
/*
* The hardware will use primary station addr together with the
* BSSID mask when matching addresses.
*/
memset(iter_data, 0, sizeof(*iter_data));
eth_broadcast_addr(iter_data->mask);
iter_data->slottime = 9;
list_for_each_entry(avp, &ctx->vifs, list)
ath9k_vif_iter(iter_data, avp->vif->addr, avp->vif);
ath9k_update_bssid_mask(sc, ctx, iter_data);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Ben Greear | 50 | 59.52% | 2 | 33.33% |
Rajkumar Manoharan | 23 | 27.38% | 1 | 16.67% |
Sujith Manoharan | 9 | 10.71% | 1 | 16.67% |
Joe Perches | 1 | 1.19% | 1 | 16.67% |
Benjamin Berg | 1 | 1.19% | 1 | 16.67% |
Total | 84 | 100.00% | 6 | 100.00% |
static void ath9k_set_assoc_state(struct ath_softc *sc,
struct ieee80211_vif *vif, bool changed)
{
struct ath_common *common = ath9k_hw_common(sc->sc_ah);
struct ath_vif *avp = (struct ath_vif *)vif->drv_priv;
unsigned long flags;
set_bit(ATH_OP_PRIM_STA_VIF, &common->op_flags);
ether_addr_copy(common->curbssid, avp->bssid);
common->curaid = avp->aid;
ath9k_hw_write_associd(sc->sc_ah);
if (changed) {
common->last_rssi = ATH_RSSI_DUMMY_MARKER;
sc->sc_ah->stats.avgbrssi = ATH_RSSI_DUMMY_MARKER;
spin_lock_irqsave(&sc->sc_pm_lock, flags);
sc->ps_flags |= PS_BEACON_SYNC | PS_WAIT_FOR_BEACON;
spin_unlock_irqrestore(&sc->sc_pm_lock, flags);
}
if (ath9k_hw_mci_is_enabled(sc->sc_ah))
ath9k_mci_update_wlan_channels(sc, false);
ath_dbg(common, CONFIG,
"Primary Station interface: %pM, BSSID: %pM\n",
vif->addr, common->curbssid);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Rajkumar Manoharan | 86 | 51.19% | 1 | 10.00% |
Ben Greear | 45 | 26.79% | 1 | 10.00% |
Sujith Manoharan | 27 | 16.07% | 4 | 40.00% |
Javier Cardona | 4 | 2.38% | 1 | 10.00% |
Vivek Natarajan | 4 | 2.38% | 1 | 10.00% |
Pat Erley | 1 | 0.60% | 1 | 10.00% |
Johannes Berg | 1 | 0.60% | 1 | 10.00% |
Total | 168 | 100.00% | 10 | 100.00% |
#ifdef CONFIG_ATH9K_CHANNEL_CONTEXT
static void ath9k_set_offchannel_state(struct ath_softc *sc)
{
struct ath_hw *ah = sc->sc_ah;
struct ath_common *common = ath9k_hw_common(ah);
struct ieee80211_vif *vif = NULL;
ath9k_ps_wakeup(sc);
if (sc->offchannel.state < ATH_OFFCHANNEL_ROC_START)
vif = sc->offchannel.scan_vif;
else
vif = sc->offchannel.roc_vif;
if (WARN_ON(!vif))
goto exit;
eth_zero_addr(common->curbssid);
eth_broadcast_addr(common->bssidmask);
memcpy(common->macaddr, vif->addr, ETH_ALEN);
common->curaid = 0;
ah->opmode = vif->type;
ah->imask &= ~ATH9K_INT_SWBA;
ah->imask &= ~ATH9K_INT_TSFOOR;
ah->slottime = 9;
ath_hw_setbssidmask(common);
ath9k_hw_setopmode(ah);
ath9k_hw_write_associd(sc->sc_ah);
ath9k_hw_set_interrupts(ah);
ath9k_hw_init_global_settings(ah);
exit:
ath9k_ps_restore(sc);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Sujith Manoharan | 174 | 99.43% | 2 | 66.67% |
Benjamin Berg | 1 | 0.57% | 1 | 33.33% |
Total | 175 | 100.00% | 3 | 100.00% |
#endif
/* Called with sc->mutex held. */
void ath9k_calculate_summary_state(struct ath_softc *sc,
struct ath_chanctx *ctx)
{
struct ath_hw *ah = sc->sc_ah;
struct ath_common *common = ath9k_hw_common(ah);
struct ath9k_vif_iter_data iter_data;
ath_chanctx_check_active(sc, ctx);
if (ctx != sc->cur_chan)
return;
#ifdef CONFIG_ATH9K_CHANNEL_CONTEXT
if (ctx == &sc->offchannel.chan)
return ath9k_set_offchannel_state(sc);
#endif
ath9k_ps_wakeup(sc);
ath9k_calculate_iter_data(sc, ctx, &iter_data);
if (iter_data.has_hw_macaddr)
memcpy(common->macaddr, iter_data.hw_macaddr, ETH_ALEN);
memcpy(common->bssidmask, iter_data.mask, ETH_ALEN);
ath_hw_setbssidmask(common);
if (iter_data.naps > 0) {
ath9k_hw_set_tsfadjust(ah, true);
ah->opmode = NL80211_IFTYPE_AP;
} else {
ath9k_hw_set_tsfadjust(ah, false);
if (iter_data.beacons)
ath9k_beacon_ensure_primary_slot(sc);
if (iter_data.nmeshes)
ah->opmode = NL80211_IFTYPE_MESH_POINT;
else if (iter_data.nocbs)
ah->opmode = NL80211_IFTYPE_OCB;
else if (iter_data.nwds)
ah->opmode = NL80211_IFTYPE_AP;
else if (iter_data.nadhocs)
ah->opmode = NL80211_IFTYPE_ADHOC;
else
ah->opmode = NL80211_IFTYPE_STATION;
}
ath9k_hw_setopmode(ah);
ctx->switch_after_beacon = false;
if ((iter_data.nstations + iter_data.nadhocs + iter_data.nmeshes) > 0)
ah->imask |= ATH9K_INT_TSFOOR;
else {
ah->imask &= ~ATH9K_INT_TSFOOR;
if (iter_data.naps == 1 && iter_data.beacons)
ctx->switch_after_beacon = true;
}
if (ah->opmode == NL80211_IFTYPE_STATION) {
bool changed = (iter_data.primary_sta != ctx->primary_sta);
if (iter_data.primary_sta) {
iter_data.primary_beacon_vif = iter_data.primary_sta;
iter_data.beacons = true;
ath9k_set_assoc_state(sc, iter_data.primary_sta,
changed);
ctx->primary_sta = iter_data.primary_sta;
} else {
ctx->primary_sta = NULL;
eth_zero_addr(common->curbssid);
common->curaid = 0;
ath9k_hw_write_associd(sc->sc_ah);
if (ath9k_hw_mci_is_enabled(sc->sc_ah))
ath9k_mci_update_wlan_channels(sc, true);
}
}
sc->nbcnvifs = iter_data.nbcnvifs;
ath9k_beacon_config(sc, iter_data.primary_beacon_vif,
iter_data.beacons);
ath9k_hw_set_interrupts(ah);
if (ah->slottime != iter_data.slottime) {
ah->slottime = iter_data.slottime;
ath9k_hw_init_global_settings(ah);
}
if (iter_data.primary_sta)
set_bit(ATH_OP_PRIM_STA_VIF, &common->op_flags);
else
clear_bit(ATH_OP_PRIM_STA_VIF, &common->op_flags);
ath_dbg(common, CONFIG,
"macaddr: %pM, bssid: %pM, bssidmask: %pM\n",
common->macaddr, common->curbssid, common->bssidmask);
ath9k_ps_restore(sc);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Rajkumar Manoharan | 252 | 49.80% | 2 | 8.33% |
Felix Fietkau | 85 | 16.80% | 8 | 33.33% |
Sujith Manoharan | 72 | 14.23% | 6 | 25.00% |
Ben Greear | 28 | 5.53% | 1 | 4.17% |
Benjamin Berg | 25 | 4.94% | 1 | 4.17% |
Luis R. Rodriguez | 17 | 3.36% | 1 | 4.17% |
Jan Kaisrlik | 13 | 2.57% | 1 | 4.17% |
Mohammed Shafi Shajakhan | 11 | 2.17% | 1 | 4.17% |
Jouni Malinen | 1 | 0.20% | 1 | 4.17% |
Joe Perches | 1 | 0.20% | 1 | 4.17% |
Vasanthakumar Thiagarajan | 1 | 0.20% | 1 | 4.17% |
Total | 506 | 100.00% | 24 | 100.00% |
static void ath9k_tpc_vif_iter(void *data, u8 *mac, struct ieee80211_vif *vif)
{
int *power = (int *)data;
if (*power < vif->bss_conf.txpower)
*power = vif->bss_conf.txpower;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Lorenzo Bianconi | 49 | 100.00% | 1 | 100.00% |
Total | 49 | 100.00% | 1 | 100.00% |
/* Called with sc->mutex held. */
void ath9k_set_txpower(struct ath_softc *sc, struct ieee80211_vif *vif)
{
int power;
struct ath_hw *ah = sc->sc_ah;
struct ath_regulatory *reg = ath9k_hw_regulatory(ah);
ath9k_ps_wakeup(sc);
if (ah->tpc_enabled) {
power = (vif) ? vif->bss_conf.txpower : -1;
ieee80211_iterate_active_interfaces_atomic(
sc->hw, IEEE80211_IFACE_ITER_RESUME_ALL,
ath9k_tpc_vif_iter, &power);
if (power == -1)
power = sc->hw->conf.power_level;
} else {
power = sc->hw->conf.power_level;
}
sc->cur_chan->txpower = 2 * power;
ath9k_hw_set_txpowerlimit(ah, sc->cur_chan->txpower, false);
sc->cur_chan->cur_txpower = reg->max_power_level;
ath9k_ps_restore(sc);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Lorenzo Bianconi | 147 | 100.00% | 1 | 100.00% |
Total | 147 | 100.00% | 1 | 100.00% |
static void ath9k_assign_hw_queues(struct ieee80211_hw *hw,
struct ieee80211_vif *vif)
{
int i;
if (!ath9k_is_chanctx_enabled())
return;
for (i = 0; i < IEEE80211_NUM_ACS; i++)
vif->hw_queue[i] = i;
if (vif->type == NL80211_IFTYPE_AP ||
vif->type == NL80211_IFTYPE_MESH_POINT)
vif->cab_queue = hw->queues - 2;
else
vif->cab_queue = IEEE80211_INVAL_HW_QUEUE;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Sujith Manoharan | 73 | 92.41% | 2 | 66.67% |
Chun-Yeow Yeoh | 6 | 7.59% | 1 | 33.33% |
Total | 79 | 100.00% | 3 | 100.00% |
static int ath9k_add_interface(struct ieee80211_hw *hw,
struct ieee80211_vif *vif)
{
struct ath_softc *sc = hw->priv;
struct ath_hw *ah = sc->sc_ah;
struct ath_common *common = ath9k_hw_common(ah);
struct ath_vif *avp = (void *)vif->drv_priv;
struct ath_node *an = &avp->mcast_node;
mutex_lock(&sc->mutex);
if (IS_ENABLED(CONFIG_ATH9K_TX99)) {
if (sc->cur_chan->nvifs >= 1) {
mutex_unlock(&sc->mutex);
return -EOPNOTSUPP;
}
sc->tx99_vif = vif;
}
ath_dbg(common, CONFIG, "Attach a VIF of type: %d\n", vif->type);
sc->cur_chan->nvifs++;
if (vif->type == NL80211_IFTYPE_STATION && ath9k_is_chanctx_enabled())
vif->driver_flags |= IEEE80211_VIF_GET_NOA_UPDATE;
if (ath9k_uses_beacons(vif->type))
ath9k_beacon_assign_slot(sc, vif);
avp->vif = vif;
if (!ath9k_is_chanctx_enabled()) {
avp->chanctx = sc->cur_chan;
list_add_tail(&avp->list, &avp->chanctx->vifs);
}
ath9k_calculate_summary_state(sc, avp->chanctx);
ath9k_assign_hw_queues(hw, vif);
ath9k_set_txpower(sc, vif);
an->sc = sc;
an->sta = NULL;
an->vif = vif;
an->no_ps_filter = true;
ath_tx_node_init(sc, an);
mutex_unlock(&sc->mutex);
return 0;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Rajkumar Manoharan | 163 | 60.59% | 4 | 19.05% |
Felix Fietkau | 38 | 14.13% | 4 | 19.05% |
Ben Greear | 17 | 6.32% | 2 | 9.52% |
Janusz Dziedzic | 17 | 6.32% | 1 | 4.76% |
Sujith Manoharan | 12 | 4.46% | 4 | 19.05% |
Luis R. Rodriguez | 8 | 2.97% | 3 | 14.29% |
Lorenzo Bianconi | 7 | 2.60% | 1 | 4.76% |
Mohammed Shafi Shajakhan | 6 | 2.23% | 1 | 4.76% |
Masahiro Yamada | 1 | 0.37% | 1 | 4.76% |
Total | 269 | 100.00% | 21 | 100.00% |
static int ath9k_change_interface(struct ieee80211_hw *hw,
struct ieee80211_vif *vif,
enum nl80211_iftype new_type,
bool p2p)
{
struct ath_softc *sc = hw->priv;
struct ath_common *common = ath9k_hw_common(sc->sc_ah);
struct ath_vif *avp = (void *)vif->drv_priv;
mutex_lock(&sc->mutex);
if (IS_ENABLED(CONFIG_ATH9K_TX99)) {
mutex_unlock(&sc->mutex);
return -EOPNOTSUPP;
}
ath_dbg(common, CONFIG, "Change Interface\n");
if (ath9k_uses_beacons(vif->type))
ath9k_beacon_remove_slot(sc, vif);
vif->type = new_type;
vif->p2p = p2p;
if (ath9k_uses_beacons(vif->type))
ath9k_beacon_assign_slot(sc, vif);
ath9k_assign_hw_queues(hw, vif);
ath9k_calculate_summary_state(sc, avp->chanctx);
ath9k_set_txpower(sc, vif);
mutex_unlock(&sc->mutex);
return 0;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Rajkumar Manoharan | 113 | 65.32% | 2 | 18.18% |
Senthil Balasubramanian | 24 | 13.87% | 2 | 18.18% |
Sujith Manoharan | 16 | 9.25% | 2 | 18.18% |
Lorenzo Bianconi | 7 | 4.05% | 1 | 9.09% |
Luis R. Rodriguez | 6 | 3.47% | 1 | 9.09% |
Pavel Roskin | 5 | 2.89% | 1 | 9.09% |
Masahiro Yamada | 1 | 0.58% | 1 | 9.09% |
Vasanthakumar Thiagarajan | 1 | 0.58% | 1 | 9.09% |
Total | 173 | 100.00% | 11 | 100.00% |
static void ath9k_remove_interface(struct ieee80211_hw *hw,
struct ieee80211_vif *vif)
{
struct ath_softc *sc = hw->priv;
struct ath_common *common = ath9k_hw_common(sc->sc_ah);
struct ath_vif *avp = (void *)vif->drv_priv;
ath_dbg(common, CONFIG, "Detach Interface\n");
mutex_lock(&sc->mutex);
ath9k_p2p_remove_vif(sc, vif);
sc->cur_chan->nvifs--;
sc->tx99_vif = NULL;
if (!ath9k_is_chanctx_enabled())
list_del(&avp->list);
if (ath9k_uses_beacons(vif->type))
ath9k_beacon_remove_slot(sc, vif);
ath_tx_node_cleanup(sc, &avp->mcast_node);
ath9k_calculate_summary_state(sc, avp->chanctx);
ath9k_set_txpower(sc, NULL);
mutex_unlock(&sc->mutex);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Rajkumar Manoharan | 79 | 52.32% | 1 | 14.29% |
Simon Wunderlich | 50 | 33.11% | 1 | 14.29% |
Ben Greear | 9 | 5.96% | 1 | 14.29% |
Lorenzo Bianconi | 7 | 4.64% | 1 | 14.29% |
Sujith Manoharan | 6 | 3.97% | 3 | 42.86% |
Total | 151 | 100.00% | 7 | 100.00% |
static void ath9k_enable_ps(struct ath_softc *sc)
{
struct ath_hw *ah = sc->sc_ah;
struct ath_common *common = ath9k_hw_common(ah);
if (IS_ENABLED(CONFIG_ATH9K_TX99))
return;
sc->ps_enabled = true;
if (!(ah->caps.hw_caps & ATH9K_HW_CAP_AUTOSLEEP)) {
if ((ah->imask & ATH9K_INT_TIM_TIMER) == 0) {
ah->imask |= ATH9K_INT_TIM_TIMER;
ath9k_hw_set_interrupts(ah);
}
ath9k_hw_setrxabort(ah, 1);
}
ath_dbg(common, PS, "PowerSave enabled\n");
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Simon Wunderlich | 51 | 51.00% | 2 | 50.00% |
Rajkumar Manoharan | 48 | 48.00% | 1 | 25.00% |
Masahiro Yamada | 1 | 1.00% | 1 | 25.00% |
Total | 100 | 100.00% | 4 | 100.00% |
static void ath9k_disable_ps(struct ath_softc *sc)
{
struct ath_hw *ah = sc->sc_ah;
struct ath_common *common = ath9k_hw_common(ah);
if (IS_ENABLED(CONFIG_ATH9K_TX99))
return;
sc->ps_enabled = false;
ath9k_hw_setpower(ah, ATH9K_PM_AWAKE);
if (!(ah->caps.hw_caps & ATH9K_HW_CAP_AUTOSLEEP)) {
ath9k_hw_setrxabort(ah, 0);
sc->ps_flags &= ~(PS_WAIT_FOR_BEACON |
PS_WAIT_FOR_CAB |
PS_WAIT_FOR_PSPOLL_DATA |
PS_WAIT_FOR_TX_ACK);
if (ah->imask & ATH9K_INT_TIM_TIMER) {
ah->imask &= ~ATH9K_INT_TIM_TIMER;
ath9k_hw_set_interrupts(ah);
}
}
ath_dbg(common, PS, "PowerSave disabled\n");
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Rajkumar Manoharan | 66 | 55.46% | 2 | 12.50% |
Felix Fietkau | 26 | 21.85% | 6 | 37.50% |
Luis R. Rodriguez | 17 | 14.29% | 4 | 25.00% |
Sujith Manoharan | 4 | 3.36% | 1 | 6.25% |
Jouni Malinen | 4 | 3.36% | 1 | 6.25% |
Masahiro Yamada | 1 | 0.84% | 1 | 6.25% |
Vasanthakumar Thiagarajan | 1 | 0.84% | 1 | 6.25% |
Total | 119 | 100.00% | 16 | 100.00% |
static int ath9k_config(struct ieee80211_hw *hw, u32 changed)
{
struct ath_softc *sc = hw->priv;
struct ath_hw *ah = sc->sc_ah;
struct ath_common *common = ath9k_hw_common(ah);
struct ieee80211_conf *conf = &hw->conf;
struct ath_chanctx *ctx = sc->cur_chan;
ath9k_ps_wakeup(sc);
mutex_lock(&sc->mutex);
if (changed & IEEE80211_CONF_CHANGE_IDLE) {
sc->ps_idle = !!(conf->flags & IEEE80211_CONF_IDLE);
if (sc->ps_idle) {
ath_cancel_work(sc);
ath9k_stop_btcoex(sc);
} else {
ath9k_start_btcoex(sc);
/*
* The chip needs a reset to properly wake up from
* full sleep
*/
ath_chanctx_set_channel(sc, ctx, &ctx->chandef);
}
}
/*
* We just prepare to enable PS. We have to wait until our AP has
* ACK'd our null data frame to disable RX otherwise we'll ignore
* those ACKs and end up retransmitting the same null data frames.
* IEEE80211_CONF_CHANGE_PS is only passed by mac80211 for STA mode.
*/
if (changed & IEEE80211_CONF_CHANGE_PS) {
unsigned long flags;
spin_lock_irqsave(&sc->sc_pm_lock, flags);
if (conf->flags & IEEE80211_CONF_PS)
ath9k_enable_ps(sc);
else
ath9k_disable_ps(sc);
spin_unlock_irqrestore(&sc->sc_pm_lock, flags);
}
if (changed & IEEE80211_CONF_CHANGE_MONITOR) {
if (conf->flags & IEEE80211_CONF_MONITOR) {
ath_dbg(common, CONFIG, "Monitor mode is enabled\n");
sc->sc_ah->is_monitoring = true;
} else {
ath_dbg(common, CONFIG, "Monitor mode is disabled\n");
sc->sc_ah->is_monitoring = false;
}
}
if (!ath9k_is_chanctx_enabled() && (changed & IEEE80211_CONF_CHANGE_CHANNEL)) {
ctx->offchannel = !!(conf->flags & IEEE80211_CONF_OFFCHANNEL);
ath_chanctx_set_channel(sc, ctx, &hw->conf.chandef);
}
mutex_unlock(&sc->mutex);
ath9k_ps_restore(sc);
return 0;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Rajkumar Manoharan | 224 | 74.67% | 2 | 12.50% |
Felix Fietkau | 20 | 6.67% | 4 | 25.00% |
Jouni Malinen | 18 | 6.00% | 3 | 18.75% |
Johannes Berg | 11 | 3.67% | 2 | 12.50% |
Luis R. Rodriguez | 11 | 3.67% | 2 | 12.50% |
Vasanthakumar Thiagarajan | 9 | 3.00% | 1 | 6.25% |
Sujith Manoharan | 7 | 2.33% | 2 | 12.50% |
Total | 300 | 100.00% | 16 | 100.00% |
#define SUPPORTED_FILTERS \
(FIF_ALLMULTI | \
FIF_CONTROL | \
FIF_PSPOLL | \
FIF_OTHER_BSS | \
FIF_BCN_PRBRESP_PROMISC | \
FIF_PROBE_REQ | \
FIF_FCSFAIL)
/* FIXME: sc->sc_full_reset ? */
static void ath9k_configure_filter(struct ieee80211_hw *hw,
unsigned int changed_flags,
unsigned int *total_flags,
u64 multicast)
{
struct ath_softc *sc = hw->priv;
struct ath_chanctx *ctx;
u32 rfilt;
changed_flags &= SUPPORTED_FILTERS;
*total_flags &= SUPPORTED_FILTERS;
spin_lock_bh(&sc->chan_lock);
ath_for_each_chanctx(sc, ctx)
ctx->rxfilter = *total_flags;
#ifdef CONFIG_ATH9K_CHANNEL_CONTEXT
sc->offchannel.chan.rxfilter = *total_flags;
#endif
spin_unlock_bh(&sc->chan_lock);
ath9k_ps_wakeup(sc);
rfilt = ath_calcrxfilter(sc);
ath9k_hw_setrxfilter(sc->sc_ah, rfilt);
ath9k_ps_restore(sc);
ath_dbg(ath9k_hw_common(sc->sc_ah), CONFIG, "Set HW RX filter: 0x%x\n",
rfilt);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Rajkumar Manoharan | 86 | 63.24% | 1 | 20.00% |
Janusz Dziedzic | 28 | 20.59% | 2 | 40.00% |
Sujith Manoharan | 15 | 11.03% | 1 | 20.00% |
Felix Fietkau | 7 | 5.15% | 1 | 20.00% |
Total | 136 | 100.00% | 5 | 100.00% |
static int ath9k_sta_add(struct ieee80211_hw *hw,
struct ieee80211_vif *vif,
struct ieee80211_sta *sta)
{
struct ath_softc *sc = hw->priv;
struct ath_common *common = ath9k_hw_common(sc->sc_ah);
struct ath_node *an = (struct ath_node *) sta->drv_priv;
struct ieee80211_key_conf ps_key = { };
int key;
ath_node_attach(sc, sta, vif);
if (vif->type != NL80211_IFTYPE_AP &&
vif->type != NL80211_IFTYPE_AP_VLAN)
return 0;
key = ath_key_config(common, vif, sta, &ps_key);
if (key > 0) {
an->ps_key = key;
an->key_idx[0] = key;
}
return 0;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Rajkumar Manoharan | 89 | 67.94% | 1 | 50.00% |
Felix Fietkau | 42 | 32.06% | 1 | 50.00% |
Total | 131 | 100.00% | 2 | 100.00% |
static void ath9k_del_ps_key(struct ath_softc *sc,
struct ieee80211_vif *vif,
struct ieee80211_sta *sta)
{
struct ath_common *common = ath9k_hw_common(sc->sc_ah);
struct ath_node *an = (struct ath_node *) sta->drv_priv;
struct ieee80211_key_conf ps_key = { .hw_key_idx = an->ps_key };
if (!an->ps_key)
return;
ath_key_delete(common, &ps_key);
an->ps_key = 0;
an->key_idx[0] = 0;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Rajkumar Manoharan | 60 | 65.93% | 2 | 50.00% |
Felix Fietkau | 31 | 34.07% | 2 | 50.00% |
Total | 91 | 100.00% | 4 | 100.00% |
static int ath9k_sta_remove(struct ieee80211_hw *hw,
struct ieee80211_vif *vif,
struct ieee80211_sta *sta)
{
struct ath_softc *sc = hw->priv;
ath9k_del_ps_key(sc, vif, sta);
ath_node_detach(sc, sta);
return 0;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Johannes Berg | 33 | 67.35% | 2 | 28.57% |
Felix Fietkau | 11 | 22.45% | 2 | 28.57% |
Vasanthakumar Thiagarajan | 4 | 8.16% | 2 | 28.57% |
Luis R. Rodriguez | 1 | 2.04% | 1 | 14.29% |
Total | 49 | 100.00% | 7 | 100.00% |
static int ath9k_sta_state(struct ieee80211_hw *hw,
struct ieee80211_vif *vif,
struct ieee80211_sta *sta,
enum ieee80211_sta_state old_state,
enum ieee80211_sta_state new_state)
{
struct ath_softc *sc = hw->priv;
struct ath_common *common = ath9k_hw_common(sc->sc_ah);
int ret = 0;
if (old_state == IEEE80211_STA_NOTEXIST &&
new_state == IEEE80211_STA_NONE) {
ret = ath9k_sta_add(hw, vif, sta);
ath_dbg(common, CONFIG,
"Add station: %pM\n", sta->addr);
} else if (old_state == IEEE80211_STA_NONE &&
new_state == IEEE80211_STA_NOTEXIST) {
ret = ath9k_sta_remove(hw, vif, sta);
ath_dbg(common, CONFIG,
"Remove station: %pM\n", sta->addr);
}
if (ath9k_is_chanctx_enabled()) {
if (vif->type == NL80211_IFTYPE_STATION) {
if (old_state == IEEE80211_STA_ASSOC &&
new_state == IEEE80211_STA_AUTHORIZED)
ath_chanctx_event(sc, vif,
ATH_CHANCTX_EVENT_AUTHORIZED);
}
}
return ret;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Sujith Manoharan | 163 | 97.60% | 3 | 75.00% |
Felix Fietkau | 4 | 2.40% | 1 | 25.00% |
Total | 167 | 100.00% | 4 | 100.00% |
static void ath9k_sta_set_tx_filter(struct ath_hw *ah,
struct ath_node *an,
bool set)
{
int i;
for (i = 0; i < ARRAY_SIZE(an->key_idx); i++) {
if (!an->key_idx[i])
continue;
ath9k_hw_set_tx_filter(ah, an->key_idx[i], set);
}
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Rajkumar Manoharan | 67 | 100.00% | 1 | 100.00% |
Total | 67 | 100.00% | 1 | 100.00% |
static void ath9k_sta_notify(struct ieee80211_hw *hw,
struct ieee80211_vif *vif,
enum sta_notify_cmd cmd,
struct ieee80211_sta *sta)
{
struct ath_softc *sc = hw->priv;
struct ath_node *an = (struct ath_node *) sta->drv_priv;
switch (cmd) {
case STA_NOTIFY_SLEEP:
an->sleeping = true;
ath_tx_aggr_sleep(sta, sc, an);
ath9k_sta_set_tx_filter(sc->sc_ah, an, true);
break;
case STA_NOTIFY_AWAKE:
ath9k_sta_set_tx_filter(sc->sc_ah, an, false);
an->sleeping = false;
ath_tx_aggr_wakeup(sc, an);
break;
}
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Felix Fietkau | 88 | 78.57% | 1 | 25.00% |
Rajkumar Manoharan | 22 | 19.64% | 2 | 50.00% |
Johannes Berg | 2 | 1.79% | 1 | 25.00% |
Total | 112 | 100.00% | 4 | 100.00% |
static int ath9k_conf_tx(struct ieee80211_hw *hw,
struct ieee80211_vif *vif, u16 queue,
const struct ieee80211_tx_queue_params *params)
{
struct ath_softc *sc = hw->priv;
struct ath_common *common = ath9k_hw_common(sc->sc_ah);
struct ath_txq *txq;
struct ath9k_tx_queue_info qi;
int ret = 0;
if (queue >= IEEE80211_NUM_ACS)
return 0;
txq = sc->tx.txq_map[queue];
ath9k_ps_wakeup(sc);
mutex_lock(&sc->mutex);
memset(&qi, 0, sizeof(struct ath9k_tx_queue_info));
qi.tqi_aifs = params->aifs;
qi.tqi_cwmin = params->cw_min;
qi.tqi_cwmax = params->cw_max;
qi.tqi_burstTime = params->txop * 32;
ath_dbg(common, CONFIG,
"Configure tx [queue/halq] [%d/%d], aifs: %d, cw_min: %d, cw_max: %d, txop: %d\n",
queue, txq->axq_qnum, params->aifs, params->cw_min,
params->cw_max, params->txop);
ath_update_max_aggr_framelen(sc, queue, qi.tqi_burstTime);
ret = ath_txq_update(sc, txq->axq_qnum, &qi);
if (ret)
ath_err(common, "TXQ Update failed\n");
mutex_unlock(&sc->mutex);
ath9k_ps_restore(sc);
return ret;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Vasanthakumar Thiagarajan | 108 | 48.21% | 2 | 11.76% |
Felix Fietkau | 47 | 20.98% | 5 | 29.41% |
Sujith Manoharan | 23 | 10.27% | 3 | 17.65% |
Johannes Berg | 16 | 7.14% | 1 | 5.88% |
Luis R. Rodriguez | 14 | 6.25% | 1 | 5.88% |
Jouni Malinen | 7 | 3.12% | 1 | 5.88% |
Eliad Peller | 5 | 2.23% | 1 | 5.88% |
Joe Perches | 4 | 1.79% | 3 | 17.65% |
Total | 224 | 100.00% | 17 | 100.00% |
static int ath9k_set_key(struct ieee80211_hw *hw,
enum set_key_cmd cmd,
struct ieee80211_vif *vif,
struct ieee80211_sta *sta,
struct ieee80211_key_conf *key)
{
struct ath_softc *sc = hw->priv;
struct ath_common *common = ath9k_hw_common(sc->sc_ah);
struct ath_node *an = NULL;
int ret = 0, i;
if (ath9k_modparam_nohwcrypt)
return -ENOSPC;
if ((vif->type == NL80211_IFTYPE_ADHOC ||
vif->type == NL80211_IFTYPE_MESH_POINT) &&
(key->cipher == WLAN_CIPHER_SUITE_TKIP ||
key->cipher == WLAN_CIPHER_SUITE_CCMP) &&
!(key->flags & IEEE80211_KEY_FLAG_PAIRWISE)) {
/*
* For now, disable hw crypto for the RSN IBSS group keys. This
* could be optimized in the future to use a modified key cache
* design to support per-STA RX GTK, but until that gets
* implemented, use of software crypto for group addressed
* frames is a acceptable to allow RSN IBSS to be used.
*/
return -EOPNOTSUPP;
}
mutex_lock(&sc->mutex);
ath9k_ps_wakeup(sc);
ath_dbg(common, CONFIG, "Set HW Key %d\n", cmd);
if (sta)
an = (struct ath_node *)sta->drv_priv;
switch (cmd) {
case SET_KEY:
if (sta)
ath9k_del_ps_key(sc, vif, sta);
key->hw_key_idx = 0;
ret = ath_key_config(common, vif, sta, key);
if (ret >= 0) {
key->hw_key_idx = ret;
/* push IV and Michael MIC generation to stack */
key->flags |= IEEE80211_KEY_FLAG_GENERATE_IV;
if (key->cipher == WLAN_CIPHER_SUITE_TKIP)
key->flags |= IEEE80211_KEY_FLAG_GENERATE_MMIC;
if (sc->sc_ah->sw_mgmt_crypto_tx &&
key->cipher == WLAN_CIPHER_SUITE_CCMP)
key->flags |= IEEE80211_KEY_FLAG_SW_MGMT_TX;
ret = 0;
}
if (an && key->hw_key_idx) {
for (i = 0; i < ARRAY_SIZE(an->key_idx); i++) {
if (an->key_idx[i])
continue;
an->key_idx[i] = key->hw_key_idx;
break;
}
WARN_ON(i == ARRAY_SIZE(an->key_idx));
}
break;
case DISABLE_KEY:
ath_key_delete(common, key);
if (an) {
for (i = 0; i < ARRAY_SIZE(an->key_idx); i++) {
if (an->key_idx[i] != key->hw_key_idx)
continue;
an->key_idx[i] = 0;
break;
}
}
key->hw_key_idx = 0;
break;
default:
ret = -EINVAL;
}
ath9k_ps_restore(sc);
mutex_unlock(&sc->mutex);
return ret;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Rajkumar Manoharan | 186 | 45.15% | 5 | 23.81% |
Johannes Berg | 143 | 34.71% | 3 | 14.29% |
Jouni Malinen | 38 | 9.22% | 1 | 4.76% |
Felix Fietkau | 15 | 3.64% | 2 | 9.52% |
Luis R. Rodriguez | 14 | 3.40% | 2 | 9.52% |
Chun-Yeow Yeoh | 9 | 2.18% | 2 | 9.52% |
Joe Perches | 2 | 0.49% | 2 | 9.52% |
Sujith Manoharan | 2 | 0.49% | 1 | 4.76% |
Oleksij Rempel | 1 | 0.24% | 1 | 4.76% |
John W. Linville | 1 | 0.24% | 1 | 4.76% |
Bruno Randolf | 1 | 0.24% | 1 | 4.76% |
Total | 412 | 100.00% | 21 | 100.00% |
static void ath9k_bss_info_changed(struct ieee80211_hw *hw,
struct ieee80211_vif *vif,
struct ieee80211_bss_conf *bss_conf,
u32 changed)
{
#define CHECK_ANI \
(BSS_CHANGED_ASSOC | \
BSS_CHANGED_IBSS | \
BSS_CHANGED_BEACON_ENABLED)
struct ath_softc *sc = hw->priv;
struct ath_hw *ah = sc->sc_ah;
struct ath_common *common = ath9k_hw_common(ah);
struct ath_vif *avp = (void *)vif->drv_priv;
int slottime;
ath9k_ps_wakeup(sc);
mutex_lock(&sc->mutex);
if (changed & BSS_CHANGED_ASSOC) {
ath_dbg(common, CONFIG, "BSSID %pM Changed ASSOC %d\n",
bss_conf->bssid, bss_conf->assoc);
memcpy(avp->bssid, bss_conf->bssid, ETH_ALEN);
avp->aid = bss_conf->aid;
avp->assoc = bss_conf->assoc;
ath9k_calculate_summary_state(sc, avp->chanctx);
}
if ((changed & BSS_CHANGED_IBSS) ||
(changed & BSS_CHANGED_OCB)) {
memcpy(common->curbssid, bss_conf->bssid, ETH_ALEN);
common->curaid = bss_conf->aid;
ath9k_hw_write_associd(sc->sc_ah);
}
if ((changed & BSS_CHANGED_BEACON_ENABLED) ||
(changed & BSS_CHANGED_BEACON_INT) ||
(changed & BSS_CHANGED_BEACON_INFO)) {
ath9k_calculate_summary_state(sc, avp->chanctx);
}
if ((avp->chanctx == sc->cur_chan) &&
(changed & BSS_CHANGED_ERP_SLOT)) {
if (bss_conf->use_short_slot)
slottime = 9;
else
slottime = 20;
if (vif->type == NL80211_IFTYPE_AP) {
/*
* Defer update, so that connected stations can adjust
* their settings at the same time.
* See beacon.c for more details
*/
sc->beacon.slottime = slottime;
sc->beacon.updateslot = UPDATE;
} else {
ah->slottime = slottime;
ath9k_hw_init_global_settings(ah);
}
}
if (changed & BSS_CHANGED_P2P_PS)
ath9k_p2p_bss_info_changed(sc, vif);
if (changed & CHECK_ANI)
ath_check_ani(sc);
if (changed & BSS_CHANGED_TXPOWER) {
ath_dbg(common, CONFIG, "vif %pM power %d dbm power_type %d\n",
vif->addr, bss_conf->txpower, bss_conf->txpower_type);
ath9k_set_txpower(sc, vif);
}
mutex_unlock(&sc->mutex);
ath9k_ps_restore(sc);
#undef CHECK_ANI
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Felix Fietkau | 98 | 26.06% | 5 | 18.52% |
Sujith Manoharan | 76 | 20.21% | 8 | 29.63% |
Rajkumar Manoharan | 73 | 19.41% | 5 | 18.52% |
Johannes Berg | 52 | 13.83% | 2 | 7.41% |
Lorenzo Bianconi | 36 | 9.57% | 1 | 3.70% |
Vasanthakumar Thiagarajan | 14 | 3.72% | 1 | 3.70% |
Luis R. Rodriguez | 11 | 2.93% | 2 | 7.41% |
Jan Kaisrlik | 8 | 2.13% | 1 | 3.70% |
Jouni Malinen | 8 | 2.13% | 2 | 7.41% |
Total | 376 | 100.00% | 27 | 100.00% |
static u64 ath9k_get_tsf(struct ieee80211_hw *hw, struct ieee80211_vif *vif)
{
struct ath_softc *sc = hw->priv;
struct ath_vif *avp = (void *)vif->drv_priv;
u64 tsf;
mutex_lock(&sc->mutex);
ath9k_ps_wakeup(sc);
/* Get current TSF either from HW or kernel time. */
if (sc->cur_chan == avp->chanctx) {
tsf = ath9k_hw_gettsf64(sc->sc_ah);
} else {
tsf = sc->cur_chan->tsf_val +
ath9k_hw_get_tsf_offset(&sc->cur_chan->tsf_ts, NULL);
}
tsf += le64_to_cpu(avp->tsf_adjust);
ath9k_ps_restore(sc);
mutex_unlock(&sc->mutex);
return tsf;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Benjamin Berg | 58 | 46.77% | 2 | 22.22% |
Sujith Manoharan | 27 | 21.77% | 2 | 22.22% |
Luis R. Rodriguez | 17 | 13.71% | 1 | 11.11% |
Vasanthakumar Thiagarajan | 12 | 9.68% | 1 | 11.11% |
Eliad Peller | 5 | 4.03% | 1 | 11.11% |
Felix Fietkau | 4 | 3.23% | 1 | 11.11% |
Jouni Malinen | 1 | 0.81% | 1 | 11.11% |
Total | 124 | 100.00% | 9 | 100.00% |
static void ath9k_set_tsf(struct ieee80211_hw *hw,
struct ieee80211_vif *vif,
u64 tsf)
{
struct ath_softc *sc = hw->priv;
struct ath_vif *avp = (void *)vif->drv_priv;
mutex_lock(&sc->mutex);
ath9k_ps_wakeup(sc);
tsf -= le64_to_cpu(avp->tsf_adjust);
getrawmonotonic(&avp->chanctx->tsf_ts);
if (sc->cur_chan == avp->chanctx)
ath9k_hw_settsf64(sc->sc_ah, tsf);
avp->chanctx->tsf_val = tsf;
ath9k_ps_restore(sc);
mutex_unlock(&sc->mutex);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Benjamin Berg | 50 | 44.25% | 2 | 25.00% |
Sujith Manoharan | 27 | 23.89% | 2 | 25.00% |
Alina Friedrichsen | 23 | 20.35% | 1 | 12.50% |
Jouni Malinen | 6 | 5.31% | 1 | 12.50% |
Eliad Peller | 5 | 4.42% | 1 | 12.50% |
Felix Fietkau | 2 | 1.77% | 1 | 12.50% |
Total | 113 | 100.00% | 8 | 100.00% |
static void ath9k_reset_tsf(struct ieee80211_hw *hw, struct ieee80211_vif *vif)
{
struct ath_softc *sc = hw->priv;
struct ath_vif *avp = (void *)vif->drv_priv;
mutex_lock(&sc->mutex);
ath9k_ps_wakeup(sc);
getrawmonotonic(&avp->chanctx->tsf_ts);
if (sc->cur_chan == avp->chanctx)
ath9k_hw_reset_tsf(sc->sc_ah);
avp->chanctx->tsf_val = 0;
ath9k_ps_restore(sc);
mutex_unlock(&sc->mutex);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Benjamin Berg | 41 | 41.41% | 1 | 12.50% |
Sujith Manoharan | 17 | 17.17% | 1 | 12.50% |
Luis R. Rodriguez | 15 | 15.15% | 2 | 25.00% |
Vasanthakumar Thiagarajan | 13 | 13.13% | 1 | 12.50% |
Jouni Malinen | 6 | 6.06% | 1 | 12.50% |
Eliad Peller | 5 | 5.05% | 1 | 12.50% |
Felix Fietkau | 2 | 2.02% | 1 | 12.50% |
Total | 99 | 100.00% | 8 | 100.00% |
static int ath9k_ampdu_action(struct ieee80211_hw *hw,
struct ieee80211_vif *vif,
struct ieee80211_ampdu_params *params)
{
struct ath_softc *sc = hw->priv;
struct ath_common *common = ath9k_hw_common(sc->sc_ah);
bool flush = false;
int ret = 0;
struct ieee80211_sta *sta = params->sta;
struct ath_node *an = (struct ath_node *)sta->drv_priv;
enum ieee80211_ampdu_mlme_action action = params->action;
u16 tid = params->tid;
u16 *ssn = ¶ms->ssn;
struct ath_atx_tid *atid;
mutex_lock(&sc->mutex);
switch (action) {
case IEEE80211_AMPDU_RX_START:
break;
case IEEE80211_AMPDU_RX_STOP:
break;
case IEEE80211_AMPDU_TX_START:
if (ath9k_is_chanctx_enabled()) {
if (test_bit(ATH_OP_SCANNING, &common->op_flags)) {
ret = -EBUSY;
break;
}
}
ath9k_ps_wakeup(sc);
ret = ath_tx_aggr_start(sc, sta, tid, ssn);
if (!ret)
ieee80211_start_tx_ba_cb_irqsafe(vif, sta->addr, tid);
ath9k_ps_restore(sc);
break;
case IEEE80211_AMPDU_TX_STOP_FLUSH:
case IEEE80211_AMPDU_TX_STOP_FLUSH_CONT:
flush = true;
case IEEE80211_AMPDU_TX_STOP_CONT:
ath9k_ps_wakeup(sc);
ath_tx_aggr_stop(sc, sta, tid);
if (!flush)
ieee80211_stop_tx_ba_cb_irqsafe(vif, sta->addr, tid);
ath9k_ps_restore(sc);
break;
case IEEE80211_AMPDU_TX_OPERATIONAL:
atid = ath_node_to_tid(an, tid);
atid->baw_size = IEEE80211_MIN_AMPDU_BUF <<
sta->ht_cap.ampdu_factor;
break;
default:
ath_err(ath9k_hw_common(sc->sc_ah), "Unknown AMPDU action\n");
}
mutex_unlock(&sc->mutex);
return ret;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Vasanthakumar Thiagarajan | 66 | 22.45% | 1 | 4.55% |
Sujith Manoharan | 63 | 21.43% | 4 | 18.18% |
Luis R. Rodriguez | 42 | 14.29% | 4 | 18.18% |
Sara Sharon | 35 | 11.90% | 1 | 4.55% |
Toke Höiland-Jörgensen | 34 | 11.56% | 1 | 4.55% |
Felix Fietkau | 26 | 8.84% | 4 | 18.18% |
Johannes Berg | 21 | 7.14% | 5 | 22.73% |
Jouni Malinen | 6 | 2.04% | 1 | 4.55% |
Joe Perches | 1 | 0.34% | 1 | 4.55% |
Total | 294 | 100.00% | 22 | 100.00% |
static int ath9k_get_survey(struct ieee80211_hw *hw, int idx,
struct survey_info *survey)
{
struct ath_softc *sc = hw->priv;
struct ath_common *common = ath9k_hw_common(sc->sc_ah);
struct ieee80211_supported_band *sband;
struct ieee80211_channel *chan;
int pos;
if (IS_ENABLED(CONFIG_ATH9K_TX99))
return -EOPNOTSUPP;
spin_lock_bh(&common->cc_lock);
if (idx == 0)
ath_update_survey_stats(sc);
sband = hw->wiphy->bands[NL80211_BAND_2GHZ];
if (sband && idx >= sband->n_channels) {
idx -= sband->n_channels;
sband = NULL;
}
if (!sband)
sband = hw->wiphy->bands[NL80211_BAND_5GHZ];
if (!sband || idx >= sband->n_channels) {
spin_unlock_bh(&common->cc_lock);
return -ENOENT;
}
chan = &sband->channels[idx];
pos = chan->hw_value;
memcpy(survey, &sc->survey[pos], sizeof(*survey));
survey->channel = chan;
spin_unlock_bh(&common->cc_lock);
return 0;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Felix Fietkau | 136 | 65.07% | 4 | 44.44% |
Benoit Papillault | 57 | 27.27% | 1 | 11.11% |
Luis R. Rodriguez | 10 | 4.78% | 1 | 11.11% |
Sujith Manoharan | 3 | 1.44% | 1 | 11.11% |
Johannes Berg | 2 | 0.96% | 1 | 11.11% |
Masahiro Yamada | 1 | 0.48% | 1 | 11.11% |
Total | 209 | 100.00% | 9 | 100.00% |
static void ath9k_enable_dynack(struct ath_softc *sc)
{
#ifdef CONFIG_ATH9K_DYNACK
u32 rfilt;
struct ath_hw *ah = sc->sc_ah;
ath_dynack_reset(ah);
ah->dynack.enabled = true;
rfilt = ath_calcrxfilter(sc);
ath9k_hw_setrxfilter(ah, rfilt);
#endif
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Lorenzo Bianconi | 55 | 100.00% | 1 | 100.00% |
Total | 55 | 100.00% | 1 | 100.00% |
static void ath9k_set_coverage_class(struct ieee80211_hw *hw,
s16 coverage_class)
{
struct ath_softc *sc = hw->priv;
struct ath_hw *ah = sc->sc_ah;
if (IS_ENABLED(CONFIG_ATH9K_TX99))
return;
mutex_lock(&sc->mutex);
if (coverage_class >= 0) {
ah->coverage_class = coverage_class;
if (ah->dynack.enabled) {
u32 rfilt;
ah->dynack.enabled = false;
rfilt = ath_calcrxfilter(sc);
ath9k_hw_setrxfilter(ah, rfilt);
}
ath9k_ps_wakeup(sc);
ath9k_hw_init_global_settings(ah);
ath9k_ps_restore(sc);
} else if (!ah->dynack.enabled) {
ath9k_enable_dynack(sc);
}
mutex_unlock(&sc->mutex);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Lorenzo Bianconi | 60 | 43.80% | 1 | 12.50% |
Felix Fietkau | 58 | 42.34% | 2 | 25.00% |
Luis R. Rodriguez | 16 | 11.68% | 2 | 25.00% |
John W. Linville | 1 | 0.73% | 1 | 12.50% |
Sujith Manoharan | 1 | 0.73% | 1 | 12.50% |
Masahiro Yamada | 1 | 0.73% | 1 | 12.50% |
Total | 137 | 100.00% | 8 | 100.00% |
static bool ath9k_has_tx_pending(struct ath_softc *sc,
bool sw_pending)
{
int i, npend = 0;
for (i = 0; i < ATH9K_NUM_TX_QUEUES; i++) {
if (!ATH_TXQ_SETUP(sc, i))
continue;
npend = ath9k_has_pending_frames(sc, &sc->tx.txq[i],
sw_pending);
if (npend)
break;
}
return !!npend;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Felix Fietkau | 69 | 90.79% | 1 | 33.33% |
Sujith Manoharan | 5 | 6.58% | 1 | 33.33% |
Geert Uytterhoeven | 2 | 2.63% | 1 | 33.33% |
Total | 76 | 100.00% | 3 | 100.00% |
static void ath9k_flush(struct ieee80211_hw *hw, struct ieee80211_vif *vif,
u32 queues, bool drop)
{
struct ath_softc *sc = hw->priv;
struct ath_common *common = ath9k_hw_common(sc->sc_ah);
if (ath9k_is_chanctx_enabled()) {
if (!test_bit(ATH_OP_MULTI_CHANNEL, &common->op_flags))
goto flush;
/*
* If MCC is active, extend the flush timeout
* and wait for the HW/SW queues to become
* empty. This needs to be done outside the
* sc->mutex lock to allow the channel scheduler
* to switch channel contexts.
*
* The vif queues have been stopped in mac80211,
* so there won't be any incoming frames.
*/
__ath9k_flush(hw, queues, drop, true, true);
return;
}
flush:
mutex_lock(&sc->mutex);
__ath9k_flush(hw, queues, drop, true, false);
mutex_unlock(&sc->mutex);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Sujith Manoharan | 56 | 50.00% | 2 | 33.33% |
Felix Fietkau | 26 | 23.21% | 1 | 16.67% |
Vasanthakumar Thiagarajan | 22 | 19.64% | 1 | 16.67% |
Emmanuel Grumbach | 5 | 4.46% | 1 | 16.67% |
Johannes Berg | 3 | 2.68% | 1 | 16.67% |
Total | 112 | 100.00% | 6 | 100.00% |
void __ath9k_flush(struct ieee80211_hw *hw, u32 queues, bool drop,
bool sw_pending, bool timeout_override)
{
struct ath_softc *sc = hw->priv;
struct ath_hw *ah = sc->sc_ah;
struct ath_common *common = ath9k_hw_common(ah);
int timeout;
bool drain_txq;
cancel_delayed_work_sync(&sc->hw_check_work);
if (ah->ah_flags & AH_UNPLUGGED) {
ath_dbg(common, ANY, "Device has been unplugged!\n");
return;
}
if (test_bit(ATH_OP_INVALID, &common->op_flags)) {
ath_dbg(common, ANY, "Device not present\n");
return;
}
spin_lock_bh(&sc->chan_lock);
if (timeout_override)
timeout = HZ / 5;
else
timeout = sc->cur_chan->flush_timeout;
spin_unlock_bh(&sc->chan_lock);
ath_dbg(common, CHAN_CTX,
"Flush timeout: %d\n", jiffies_to_msecs(timeout));
if (wait_event_timeout(sc->tx_wait, !ath9k_has_tx_pending(sc, sw_pending),
timeout) > 0)
drop = false;
if (drop) {
ath9k_ps_wakeup(sc);
spin_lock_bh(&sc->sc_pcu_lock);
drain_txq = ath_drain_all_txq(sc);
spin_unlock_bh(&sc->sc_pcu_lock);
if (!drain_txq)
ath_reset(sc, NULL);
ath9k_ps_restore(sc);
}
ieee80211_queue_delayed_work(hw, &sc->hw_check_work,
ATH_HW_CHECK_POLL_INT);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Sujith Manoharan | 64 | 26.12% | 5 | 22.73% |
Felix Fietkau | 58 | 23.67% | 7 | 31.82% |
Vasanthakumar Thiagarajan | 39 | 15.92% | 1 | 4.55% |
Mohammed Shafi Shajakhan | 35 | 14.29% | 2 | 9.09% |
Rajkumar Manoharan | 19 | 7.76% | 2 | 9.09% |
John W. Linville | 13 | 5.31% | 1 | 4.55% |
Senthil Balasubramanian | 10 | 4.08% | 1 | 4.55% |
Oleksij Rempel | 3 | 1.22% | 1 | 4.55% |
Benoit Papillault | 2 | 0.82% | 1 | 4.55% |
Joe Perches | 2 | 0.82% | 1 | 4.55% |
Total | 245 | 100.00% | 22 | 100.00% |
static bool ath9k_tx_frames_pending(struct ieee80211_hw *hw)
{
struct ath_softc *sc = hw->priv;
return ath9k_has_tx_pending(sc, true);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Vivek Natarajan | 24 | 85.71% | 1 | 33.33% |
Sujith Manoharan | 4 | 14.29% | 2 | 66.67% |
Total | 28 | 100.00% | 3 | 100.00% |
static int ath9k_tx_last_beacon(struct ieee80211_hw *hw)
{
struct ath_softc *sc = hw->priv;
struct ath_hw *ah = sc->sc_ah;
struct ieee80211_vif *vif;
struct ath_vif *avp;
struct ath_buf *bf;
struct ath_tx_status ts;
bool edma = !!(ah->caps.hw_caps & ATH9K_HW_CAP_EDMA);
int status;
vif = sc->beacon.bslot[0];
if (!vif)
return 0;
if (!vif->bss_conf.enable_beacon)
return 0;
avp = (void *)vif->drv_priv;
if (!sc->beacon.tx_processed && !edma) {
tasklet_disable(&sc->bcon_tasklet);
bf = avp->av_bcbuf;
if (!bf || !bf->bf_mpdu)
goto skip;
status = ath9k_hw_txprocdesc(ah, bf->bf_desc, &ts);
if (status == -EINPROGRESS)
goto skip;
sc->beacon.tx_processed = true;
sc->beacon.tx_last = !(ts.ts_status & ATH9K_TXERR_MASK);
skip:
tasklet_enable(&sc->bcon_tasklet);
}
return sc->beacon.tx_last;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Felix Fietkau | 199 | 93.87% | 2 | 50.00% |
Sujith Manoharan | 12 | 5.66% | 1 | 25.00% |
Mohammed Shafi Shajakhan | 1 | 0.47% | 1 | 25.00% |
Total | 212 | 100.00% | 4 | 100.00% |
static int ath9k_get_stats(struct ieee80211_hw *hw,
struct ieee80211_low_level_stats *stats)
{
struct ath_softc *sc = hw->priv;
struct ath_hw *ah = sc->sc_ah;
struct ath9k_mib_stats *mib_stats = &ah->ah_mibStats;
stats->dot11ACKFailureCount = mib_stats->ackrcv_bad;
stats->dot11RTSFailureCount = mib_stats->rts_bad;
stats->dot11FCSErrorCount = mib_stats->fcs_bad;
stats->dot11RTSSuccessCount = mib_stats->rts_good;
return 0;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Mohammed Shafi Shajakhan | 79 | 100.00% | 1 | 100.00% |
Total | 79 | 100.00% | 1 | 100.00% |
static u32 fill_chainmask(u32 cap, u32 new)
{
u32 filled = 0;
int i;
for (i = 0; cap && new; i++, cap >>= 1) {
if (!(cap & BIT(0)))
continue;
if (new & BIT(0))
filled |= BIT(i);
new >>= 1;
}
return filled;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Felix Fietkau | 75 | 100.00% | 1 | 100.00% |
Total | 75 | 100.00% | 1 | 100.00% |
static bool validate_antenna_mask(struct ath_hw *ah, u32 val)
{
if (AR_SREV_9300_20_OR_LATER(ah))
return true;
switch (val & 0x7) {
case 0x1:
case 0x3:
case 0x7:
return true;
case 0x2:
return (ah->caps.rx_chainmask == 1);
default:
return false;
}
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Felix Fietkau | 62 | 100.00% | 2 | 100.00% |
Total | 62 | 100.00% | 2 | 100.00% |
static int ath9k_set_antenna(struct ieee80211_hw *hw, u32 tx_ant, u32 rx_ant)
{
struct ath_softc *sc = hw->priv;
struct ath_hw *ah = sc->sc_ah;
if (ah->caps.rx_chainmask != 1)
rx_ant |= tx_ant;
if (!validate_antenna_mask(ah, rx_ant) || !tx_ant)
return -EINVAL;
sc->ant_rx = rx_ant;
sc->ant_tx = tx_ant;
if (ah->caps.rx_chainmask == 1)
return 0;
/* AR9100 runs into calibration issues if not all rx chains are enabled */
if (AR_SREV_9100(ah))
ah->rxchainmask = 0x7;
else
ah->rxchainmask = fill_chainmask(ah->caps.rx_chainmask, rx_ant);
ah->txchainmask = fill_chainmask(ah->caps.tx_chainmask, tx_ant);
ath9k_cmn_reload_chainmask(ah);
return 0;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Felix Fietkau | 142 | 98.61% | 2 | 66.67% |
Oleksij Rempel | 2 | 1.39% | 1 | 33.33% |
Total | 144 | 100.00% | 3 | 100.00% |
static int ath9k_get_antenna(struct ieee80211_hw *hw, u32 *tx_ant, u32 *rx_ant)
{
struct ath_softc *sc = hw->priv;
*tx_ant = sc->ant_tx;
*rx_ant = sc->ant_rx;
return 0;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Felix Fietkau | 45 | 100.00% | 1 | 100.00% |
Total | 45 | 100.00% | 1 | 100.00% |
static void ath9k_sw_scan_start(struct ieee80211_hw *hw,
struct ieee80211_vif *vif,
const u8 *mac_addr)
{
struct ath_softc *sc = hw->priv;
struct ath_common *common = ath9k_hw_common(sc->sc_ah);
set_bit(ATH_OP_SCANNING, &common->op_flags);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Simon Wunderlich | 22 | 42.31% | 1 | 25.00% |
Oleksij Rempel | 15 | 28.85% | 1 | 25.00% |
Johannes Berg | 10 | 19.23% | 1 | 25.00% |
Sujith Manoharan | 5 | 9.62% | 1 | 25.00% |
Total | 52 | 100.00% | 4 | 100.00% |
static void ath9k_sw_scan_complete(struct ieee80211_hw *hw,
struct ieee80211_vif *vif)
{
struct ath_softc *sc = hw->priv;
struct ath_common *common = ath9k_hw_common(sc->sc_ah);
clear_bit(ATH_OP_SCANNING, &common->op_flags);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Simon Wunderlich | 22 | 46.81% | 1 | 25.00% |
Oleksij Rempel | 15 | 31.91% | 1 | 25.00% |
Johannes Berg | 5 | 10.64% | 1 | 25.00% |
Sujith Manoharan | 5 | 10.64% | 1 | 25.00% |
Total | 47 | 100.00% | 4 | 100.00% |
#ifdef CONFIG_ATH9K_CHANNEL_CONTEXT
static void ath9k_cancel_pending_offchannel(struct ath_softc *sc)
{
struct ath_common *common = ath9k_hw_common(sc->sc_ah);
if (sc->offchannel.roc_vif) {
ath_dbg(common, CHAN_CTX,
"%s: Aborting RoC\n", __func__);
del_timer_sync(&sc->offchannel.timer);
if (sc->offchannel.state >= ATH_OFFCHANNEL_ROC_START)
ath_roc_complete(sc, ATH_ROC_COMPLETE_ABORT);
}
if (test_bit(ATH_OP_SCANNING, &common->op_flags)) {
ath_dbg(common, CHAN_CTX,
"%s: Aborting HW scan\n", __func__);
del_timer_sync(&sc->offchannel.timer);
ath_scan_complete(sc, true);
}
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Sujith Manoharan | 112 | 99.12% | 1 | 50.00% |
Janusz Dziedzic | 1 | 0.88% | 1 | 50.00% |
Total | 113 | 100.00% | 2 | 100.00% |
static int ath9k_hw_scan(struct ieee80211_hw *hw, struct ieee80211_vif *vif,
struct ieee80211_scan_request *hw_req)
{
struct cfg80211_scan_request *req = &hw_req->req;
struct ath_softc *sc = hw->priv;
struct ath_common *common = ath9k_hw_common(sc->sc_ah);
int ret = 0;
mutex_lock(&sc->mutex);
if (WARN_ON(sc->offchannel.scan_req)) {
ret = -EBUSY;
goto out;
}
ath9k_ps_wakeup(sc);
set_bit(ATH_OP_SCANNING, &common->op_flags);
sc->offchannel.scan_vif = vif;
sc->offchannel.scan_req = req;
sc->offchannel.scan_idx = 0;
ath_dbg(common, CHAN_CTX, "HW scan request received on vif: %pM\n",
vif->addr);
if (sc->offchannel.state == ATH_OFFCHANNEL_IDLE) {
ath_dbg(common, CHAN_CTX, "Starting HW scan\n");
ath_offchannel_next(sc);
}
out:
mutex_unlock(&sc->mutex);
return ret;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Felix Fietkau | 141 | 79.66% | 2 | 50.00% |
Sujith Manoharan | 24 | 13.56% | 1 | 25.00% |
John W. Linville | 12 | 6.78% | 1 | 25.00% |
Total | 177 | 100.00% | 4 | 100.00% |
static void ath9k_cancel_hw_scan(struct ieee80211_hw *hw,
struct ieee80211_vif *vif)
{
struct ath_softc *sc = hw->priv;
struct ath_common *common = ath9k_hw_common(sc->sc_ah);
ath_dbg(common, CHAN_CTX, "Cancel HW scan on vif: %pM\n", vif->addr);
mutex_lock(&sc->mutex);
del_timer_sync(&sc->offchannel.timer);
ath_scan_complete(sc, true);
mutex_unlock(&sc->mutex);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Felix Fietkau | 58 | 69.88% | 1 | 50.00% |
Sujith Manoharan | 25 | 30.12% | 1 | 50.00% |
Total | 83 | 100.00% | 2 | 100.00% |
static int ath9k_remain_on_channel(struct ieee80211_hw *hw,
struct ieee80211_vif *vif,
struct ieee80211_channel *chan, int duration,
enum ieee80211_roc_type type)
{
struct ath_softc *sc = hw->priv;
struct ath_common *common = ath9k_hw_common(sc->sc_ah);
int ret = 0;
mutex_lock(&sc->mutex);
if (WARN_ON(sc->offchannel.roc_vif)) {
ret = -EBUSY;
goto out;
}
ath9k_ps_wakeup(sc);
sc->offchannel.roc_vif = vif;
sc->offchannel.roc_chan = chan;
sc->offchannel.roc_duration = duration;
ath_dbg(common, CHAN_CTX,
"RoC request on vif: %pM, type: %d duration: %d\n",
vif->addr, type, duration);
if (sc->offchannel.state == ATH_OFFCHANNEL_IDLE) {
ath_dbg(common, CHAN_CTX, "Starting RoC period\n");
ath_offchannel_next(sc);
}
out:
mutex_unlock(&sc->mutex);
return ret;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Felix Fietkau | 128 | 76.19% | 1 | 50.00% |
Sujith Manoharan | 40 | 23.81% | 1 | 50.00% |
Total | 168 | 100.00% | 2 | 100.00% |
static int ath9k_cancel_remain_on_channel(struct ieee80211_hw *hw)
{
struct ath_softc *sc = hw->priv;
struct ath_common *common = ath9k_hw_common(sc->sc_ah);
mutex_lock(&sc->mutex);
ath_dbg(common, CHAN_CTX, "Cancel RoC\n");
del_timer_sync(&sc->offchannel.timer);
if (sc->offchannel.roc_vif) {
if (sc->offchannel.state >= ATH_OFFCHANNEL_ROC_START)
ath_roc_complete(sc, ATH_ROC_COMPLETE_CANCEL);
}
mutex_unlock(&sc->mutex);
return 0;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Felix Fietkau | 75 | 77.32% | 1 | 33.33% |
Sujith Manoharan | 21 | 21.65% | 1 | 33.33% |
Janusz Dziedzic | 1 | 1.03% | 1 | 33.33% |
Total | 97 | 100.00% | 3 | 100.00% |
static int ath9k_add_chanctx(struct ieee80211_hw *hw,
struct ieee80211_chanctx_conf *conf)
{
struct ath_softc *sc = hw->priv;
struct ath_common *common = ath9k_hw_common(sc->sc_ah);
struct ath_chanctx *ctx, **ptr;
int pos;
mutex_lock(&sc->mutex);
ath_for_each_chanctx(sc, ctx) {
if (ctx->assigned)
continue;
ptr = (void *) conf->drv_priv;
*ptr = ctx;
ctx->assigned = true;
pos = ctx - &sc->chanctx[0];
ctx->hw_queue_base = pos * IEEE80211_NUM_ACS;
ath_dbg(common, CHAN_CTX,
"Add channel context: %d MHz\n",
conf->def.chan->center_freq);
ath_chanctx_set_channel(sc, ctx, &conf->def);
mutex_unlock(&sc->mutex);
return 0;
}
mutex_unlock(&sc->mutex);
return -ENOSPC;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Felix Fietkau | 92 | 55.76% | 3 | 50.00% |
Rajkumar Manoharan | 44 | 26.67% | 2 | 33.33% |
Sujith Manoharan | 29 | 17.58% | 1 | 16.67% |
Total | 165 | 100.00% | 6 | 100.00% |
static void ath9k_remove_chanctx(struct ieee80211_hw *hw,
struct ieee80211_chanctx_conf *conf)
{
struct ath_softc *sc = hw->priv;
struct ath_common *common = ath9k_hw_common(sc->sc_ah);
struct ath_chanctx *ctx = ath_chanctx_get(conf);
mutex_lock(&sc->mutex);
ath_dbg(common, CHAN_CTX,
"Remove channel context: %d MHz\n",
conf->def.chan->center_freq);
ctx->assigned = false;
ctx->hw_queue_base = 0;
ath_chanctx_event(sc, NULL, ATH_CHANCTX_EVENT_UNASSIGN);
mutex_unlock(&sc->mutex);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Felix Fietkau | 66 | 65.35% | 2 | 40.00% |
Sujith Manoharan | 30 | 29.70% | 2 | 40.00% |
Rajkumar Manoharan | 5 | 4.95% | 1 | 20.00% |
Total | 101 | 100.00% | 5 | 100.00% |
static void ath9k_change_chanctx(struct ieee80211_hw *hw,
struct ieee80211_chanctx_conf *conf,
u32 changed)
{
struct ath_softc *sc = hw->priv;
struct ath_common *common = ath9k_hw_common(sc->sc_ah);
struct ath_chanctx *ctx = ath_chanctx_get(conf);
mutex_lock(&sc->mutex);
ath_dbg(common, CHAN_CTX,
"Change channel context: %d MHz\n",
conf->def.chan->center_freq);
ath_chanctx_set_channel(sc, ctx, &conf->def);
mutex_unlock(&sc->mutex);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Felix Fietkau | 66 | 69.47% | 1 | 50.00% |
Sujith Manoharan | 29 | 30.53% | 1 | 50.00% |
Total | 95 | 100.00% | 2 | 100.00% |
static int ath9k_assign_vif_chanctx(struct ieee80211_hw *hw,
struct ieee80211_vif *vif,
struct ieee80211_chanctx_conf *conf)
{
struct ath_softc *sc = hw->priv;
struct ath_common *common = ath9k_hw_common(sc->sc_ah);
struct ath_vif *avp = (void *)vif->drv_priv;
struct ath_chanctx *ctx = ath_chanctx_get(conf);
int i;
ath9k_cancel_pending_offchannel(sc);
mutex_lock(&sc->mutex);
ath_dbg(common, CHAN_CTX,
"Assign VIF (addr: %pM, type: %d, p2p: %d) to channel context: %d MHz\n",
vif->addr, vif->type, vif->p2p,
conf->def.chan->center_freq);
avp->chanctx = ctx;
ctx->nvifs_assigned++;
list_add_tail(&avp->list, &ctx->vifs);
ath9k_calculate_summary_state(sc, ctx);
for (i = 0; i < IEEE80211_NUM_ACS; i++)
vif->hw_queue[i] = ctx->hw_queue_base + i;
mutex_unlock(&sc->mutex);
return 0;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Felix Fietkau | 97 | 54.49% | 1 | 16.67% |
Sujith Manoharan | 51 | 28.65% | 3 | 50.00% |
Rajkumar Manoharan | 30 | 16.85% | 2 | 33.33% |
Total | 178 | 100.00% | 6 | 100.00% |
static void ath9k_unassign_vif_chanctx(struct ieee80211_hw *hw,
struct ieee80211_vif *vif,
struct ieee80211_chanctx_conf *conf)
{
struct ath_softc *sc = hw->priv;
struct ath_common *common = ath9k_hw_common(sc->sc_ah);
struct ath_vif *avp = (void *)vif->drv_priv;
struct ath_chanctx *ctx = ath_chanctx_get(conf);
int ac;
ath9k_cancel_pending_offchannel(sc);
mutex_lock(&sc->mutex);
ath_dbg(common, CHAN_CTX,
"Remove VIF (addr: %pM, type: %d, p2p: %d) from channel context: %d MHz\n",
vif->addr, vif->type, vif->p2p,
conf->def.chan->center_freq);
avp->chanctx = NULL;
ctx->nvifs_assigned--;
list_del(&avp->list);
ath9k_calculate_summary_state(sc, ctx);
for (ac = 0; ac < IEEE80211_NUM_ACS; ac++)
vif->hw_queue[ac] = IEEE80211_INVAL_HW_QUEUE;
mutex_unlock(&sc->mutex);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Felix Fietkau | 89 | 53.61% | 1 | 16.67% |
Sujith Manoharan | 51 | 30.72% | 3 | 50.00% |
Rajkumar Manoharan | 26 | 15.66% | 2 | 33.33% |
Total | 166 | 100.00% | 6 | 100.00% |
static void ath9k_mgd_prepare_tx(struct ieee80211_hw *hw,
struct ieee80211_vif *vif)
{
struct ath_softc *sc = hw->priv;
struct ath_common *common = ath9k_hw_common(sc->sc_ah);
struct ath_vif *avp = (struct ath_vif *) vif->drv_priv;
struct ath_beacon_config *cur_conf;
struct ath_chanctx *go_ctx;
unsigned long timeout;
bool changed = false;
u32 beacon_int;
if (!test_bit(ATH_OP_MULTI_CHANNEL, &common->op_flags))
return;
if (!avp->chanctx)
return;
mutex_lock(&sc->mutex);
spin_lock_bh(&sc->chan_lock);
if (sc->next_chan || (sc->cur_chan != avp->chanctx))
changed = true;
spin_unlock_bh(&sc->chan_lock);
if (!changed)
goto out;
ath9k_cancel_pending_offchannel(sc);
go_ctx = ath_is_go_chanctx_present(sc);
if (go_ctx) {
/*
* Wait till the GO interface gets a chance
* to send out an NoA.
*/
spin_lock_bh(&sc->chan_lock);
sc->sched.mgd_prepare_tx = true;
cur_conf = &go_ctx->beacon;
beacon_int = TU_TO_USEC(cur_conf->beacon_interval);
spin_unlock_bh(&sc->chan_lock);
timeout = usecs_to_jiffies(beacon_int * 2);
init_completion(&sc->go_beacon);
mutex_unlock(&sc->mutex);
if (wait_for_completion_timeout(&sc->go_beacon,
timeout) == 0) {
ath_dbg(common, CHAN_CTX,
"Failed to send new NoA\n");
spin_lock_bh(&sc->chan_lock);
sc->sched.mgd_prepare_tx = false;
spin_unlock_bh(&sc->chan_lock);
}
mutex_lock(&sc->mutex);
}
ath_dbg(common, CHAN_CTX,
"%s: Set chanctx state to FORCE_ACTIVE for vif: %pM\n",
__func__, vif->addr);
spin_lock_bh(&sc->chan_lock);
sc->next_chan = avp->chanctx;
sc->sched.state = ATH_CHANCTX_STATE_FORCE_ACTIVE;
spin_unlock_bh(&sc->chan_lock);
ath_chanctx_set_next(sc, true);
out:
mutex_unlock(&sc->mutex);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Sujith Manoharan | 352 | 100.00% | 6 | 100.00% |
Total | 352 | 100.00% | 6 | 100.00% |
void ath9k_fill_chanctx_ops(void)
{
if (!ath9k_is_chanctx_enabled())
return;
ath9k_ops.hw_scan = ath9k_hw_scan;
ath9k_ops.cancel_hw_scan = ath9k_cancel_hw_scan;
ath9k_ops.remain_on_channel = ath9k_remain_on_channel;
ath9k_ops.cancel_remain_on_channel = ath9k_cancel_remain_on_channel;
ath9k_ops.add_chanctx = ath9k_add_chanctx;
ath9k_ops.remove_chanctx = ath9k_remove_chanctx;
ath9k_ops.change_chanctx = ath9k_change_chanctx;
ath9k_ops.assign_vif_chanctx = ath9k_assign_vif_chanctx;
ath9k_ops.unassign_vif_chanctx = ath9k_unassign_vif_chanctx;
ath9k_ops.mgd_prepare_tx = ath9k_mgd_prepare_tx;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Felix Fietkau | 71 | 95.95% | 4 | 66.67% |
Sujith Manoharan | 3 | 4.05% | 2 | 33.33% |
Total | 74 | 100.00% | 6 | 100.00% |
#endif
static int ath9k_get_txpower(struct ieee80211_hw *hw, struct ieee80211_vif *vif,
int *dbm)
{
struct ath_softc *sc = hw->priv;
struct ath_vif *avp = (void *)vif->drv_priv;
mutex_lock(&sc->mutex);
if (avp->chanctx)
*dbm = avp->chanctx->cur_txpower;
else
*dbm = sc->cur_chan->cur_txpower;
mutex_unlock(&sc->mutex);
*dbm /= 2;
return 0;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Felix Fietkau | 91 | 100.00% | 1 | 100.00% |
Total | 91 | 100.00% | 1 | 100.00% |
struct ieee80211_ops ath9k_ops = {
.tx = ath9k_tx,
.start = ath9k_start,
.stop = ath9k_stop,
.add_interface = ath9k_add_interface,
.change_interface = ath9k_change_interface,
.remove_interface = ath9k_remove_interface,
.config = ath9k_config,
.configure_filter = ath9k_configure_filter,
.sta_state = ath9k_sta_state,
.sta_notify = ath9k_sta_notify,
.conf_tx = ath9k_conf_tx,
.bss_info_changed = ath9k_bss_info_changed,
.set_key = ath9k_set_key,
.get_tsf = ath9k_get_tsf,
.set_tsf = ath9k_set_tsf,
.reset_tsf = ath9k_reset_tsf,
.ampdu_action = ath9k_ampdu_action,
.get_survey = ath9k_get_survey,
.rfkill_poll = ath9k_rfkill_poll_state,
.set_coverage_class = ath9k_set_coverage_class,
.flush = ath9k_flush,
.tx_frames_pending = ath9k_tx_frames_pending,
.tx_last_beacon = ath9k_tx_last_beacon,
.release_buffered_frames = ath9k_release_buffered_frames,
.get_stats = ath9k_get_stats,
.set_antenna = ath9k_set_antenna,
.get_antenna = ath9k_get_antenna,
#ifdef CONFIG_ATH9K_WOW
.suspend = ath9k_suspend,
.resume = ath9k_resume,
.set_wakeup = ath9k_set_wakeup,
#endif
#ifdef CONFIG_ATH9K_DEBUGFS
.get_et_sset_count = ath9k_get_et_sset_count,
.get_et_stats = ath9k_get_et_stats,
.get_et_strings = ath9k_get_et_strings,
#endif
#if defined(CONFIG_MAC80211_DEBUGFS) && defined(CONFIG_ATH9K_STATION_STATISTICS)
.sta_add_debugfs = ath9k_sta_add_debugfs,
#endif
.sw_scan_start = ath9k_sw_scan_start,
.sw_scan_complete = ath9k_sw_scan_complete,
.get_txpower = ath9k_get_txpower,
.wake_tx_queue = ath9k_wake_tx_queue,
};
Overall Contributors
Person | Tokens | Prop | Commits | CommitProp |
Felix Fietkau | 3672 | 29.20% | 95 | 27.14% |
Sujith Manoharan | 3452 | 27.45% | 94 | 26.86% |
Rajkumar Manoharan | 2003 | 15.93% | 27 | 7.71% |
Vasanthakumar Thiagarajan | 434 | 3.45% | 5 | 1.43% |
Luis R. Rodriguez | 390 | 3.10% | 23 | 6.57% |
Lorenzo Bianconi | 387 | 3.08% | 3 | 0.86% |
Ben Greear | 326 | 2.59% | 4 | 1.14% |
Johannes Berg | 311 | 2.47% | 18 | 5.14% |
Benjamin Berg | 271 | 2.16% | 4 | 1.14% |
Jouni Malinen | 268 | 2.13% | 12 | 3.43% |
Mohammed Shafi Shajakhan | 247 | 1.96% | 10 | 2.86% |
Simon Wunderlich | 155 | 1.23% | 2 | 0.57% |
Oleksij Rempel | 79 | 0.63% | 2 | 0.57% |
Benoit Papillault | 78 | 0.62% | 2 | 0.57% |
Toke Höiland-Jörgensen | 55 | 0.44% | 2 | 0.57% |
Senthil Balasubramanian | 48 | 0.38% | 5 | 1.43% |
Janusz Dziedzic | 47 | 0.37% | 4 | 1.14% |
Vivek Natarajan | 37 | 0.29% | 3 | 0.86% |
Sara Sharon | 35 | 0.28% | 1 | 0.29% |
Joe Perches | 33 | 0.26% | 4 | 1.14% |
Jan Kaisrlik | 30 | 0.24% | 1 | 0.29% |
Giedrius Statkevičius | 28 | 0.22% | 1 | 0.29% |
Alina Friedrichsen | 28 | 0.22% | 1 | 0.29% |
John W. Linville | 27 | 0.21% | 4 | 1.14% |
Eliad Peller | 20 | 0.16% | 2 | 0.57% |
Stanislaw Gruszka | 20 | 0.16% | 2 | 0.57% |
Chun-Yeow Yeoh | 15 | 0.12% | 3 | 0.86% |
Thomas Huehn | 13 | 0.10% | 1 | 0.29% |
Robert Shade | 12 | 0.10% | 1 | 0.29% |
Gabor Juhos | 10 | 0.08% | 3 | 0.86% |
Miaoqing Pan | 10 | 0.08% | 1 | 0.29% |
Pavel Roskin | 8 | 0.06% | 1 | 0.29% |
Masahiro Yamada | 6 | 0.05% | 1 | 0.29% |
Emmanuel Grumbach | 5 | 0.04% | 1 | 0.29% |
Javier Cardona | 4 | 0.03% | 1 | 0.29% |
Wojciech Dubowik | 3 | 0.02% | 1 | 0.29% |
Karl Beldan | 3 | 0.02% | 1 | 0.29% |
Geert Uytterhoeven | 2 | 0.02% | 1 | 0.29% |
Bruno Randolf | 1 | 0.01% | 1 | 0.29% |
Sven Eckelmann | 1 | 0.01% | 1 | 0.29% |
Pat Erley | 1 | 0.01% | 1 | 0.29% |
Total | 12575 | 100.00% | 350 | 100.00% |
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.