Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Jakub Kiciński | 1560 | 100.00% | 3 | 100.00% |
Total | 1560 | 3 |
/* * Copyright (C) 2014 Felix Fietkau <nbd@openwrt.org> * Copyright (C) 2015 Jakub Kicinski <kubakici@wp.pl> * * This program is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License version 2 * as published by the Free Software Foundation * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. */ #include "mt7601u.h" #include "trace.h" enum mt76_txq_id { MT_TXQ_VO = IEEE80211_AC_VO, MT_TXQ_VI = IEEE80211_AC_VI, MT_TXQ_BE = IEEE80211_AC_BE, MT_TXQ_BK = IEEE80211_AC_BK, MT_TXQ_PSD, MT_TXQ_MCU, __MT_TXQ_MAX }; /* Hardware uses mirrored order of queues with Q0 having the highest priority */ static u8 q2hwq(u8 q) { return q ^ 0x3; } /* Take mac80211 Q id from the skb and translate it to hardware Q id */ static u8 skb2q(struct sk_buff *skb) { int qid = skb_get_queue_mapping(skb); if (WARN_ON(qid >= MT_TXQ_PSD)) { qid = MT_TXQ_BE; skb_set_queue_mapping(skb, qid); } return q2hwq(qid); } /* Note: TX retry reporting is a bit broken. * Retries are reported only once per AMPDU and often come a frame early * i.e. they are reported in the last status preceding the AMPDU. Apart * from the fact that it's hard to know the length of the AMPDU (which is * required to know to how many consecutive frames retries should be * applied), if status comes early on full FIFO it gets lost and retries * of the whole AMPDU become invisible. * As a work-around encode the desired rate in PKT_ID of TX descriptor * and based on that guess the retries (every rate is tried once). * Only downside here is that for MCS0 we have to rely solely on * transmission failures as no retries can ever be reported. * Not having to read EXT_FIFO has a nice effect of doubling the number * of reports which can be fetched. * Also the vendor driver never uses the EXT_FIFO register so it may be * undertested. */ static u8 mt7601u_tx_pktid_enc(struct mt7601u_dev *dev, u8 rate, bool is_probe) { u8 encoded = (rate + 1) + is_probe * 8; /* Because PKT_ID 0 disables status reporting only 15 values are * available but 16 are needed (8 MCS * 2 for encoding is_probe) * - we need to cram together two rates. MCS0 and MCS7 with is_probe * share PKT_ID 9. */ if (is_probe && rate == 7) return encoded - 7; return encoded; } static void mt7601u_tx_pktid_dec(struct mt7601u_dev *dev, struct mt76_tx_status *stat) { u8 req_rate = stat->pktid; u8 eff_rate = stat->rate & 0x7; req_rate -= 1; if (req_rate > 7) { stat->is_probe = true; req_rate -= 8; /* Decide between MCS0 and MCS7 which share pktid 9 */ if (!req_rate && eff_rate) req_rate = 7; } stat->retry = req_rate - eff_rate; } static void mt7601u_tx_skb_remove_dma_overhead(struct sk_buff *skb, struct ieee80211_tx_info *info) { int pkt_len = (unsigned long)info->status.status_driver_data[0]; skb_pull(skb, sizeof(struct mt76_txwi) + 4); if (ieee80211_get_hdrlen_from_skb(skb) % 4) mt76_remove_hdr_pad(skb); skb_trim(skb, pkt_len); } void mt7601u_tx_status(struct mt7601u_dev *dev, struct sk_buff *skb) { struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb); mt7601u_tx_skb_remove_dma_overhead(skb, info); ieee80211_tx_info_clear_status(info); info->status.rates[0].idx = -1; info->flags |= IEEE80211_TX_STAT_ACK; spin_lock(&dev->mac_lock); ieee80211_tx_status(dev->hw, skb); spin_unlock(&dev->mac_lock); } static int mt7601u_skb_rooms(struct mt7601u_dev *dev, struct sk_buff *skb) { int hdr_len = ieee80211_get_hdrlen_from_skb(skb); u32 need_head; need_head = sizeof(struct mt76_txwi) + 4; if (hdr_len % 4) need_head += 2; return skb_cow(skb, need_head); } static struct mt76_txwi * mt7601u_push_txwi(struct mt7601u_dev *dev, struct sk_buff *skb, struct ieee80211_sta *sta, struct mt76_wcid *wcid, int pkt_len) { struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb); struct ieee80211_tx_rate *rate = &info->control.rates[0]; struct mt76_txwi *txwi; unsigned long flags; bool is_probe; u32 pkt_id; u16 rate_ctl; u8 nss; txwi = skb_push(skb, sizeof(struct mt76_txwi)); memset(txwi, 0, sizeof(*txwi)); if (!wcid->tx_rate_set) ieee80211_get_tx_rates(info->control.vif, sta, skb, info->control.rates, 1); spin_lock_irqsave(&dev->lock, flags); if (rate->idx < 0 || !rate->count) rate_ctl = wcid->tx_rate; else rate_ctl = mt76_mac_tx_rate_val(dev, rate, &nss); spin_unlock_irqrestore(&dev->lock, flags); txwi->rate_ctl = cpu_to_le16(rate_ctl); if (!(info->flags & IEEE80211_TX_CTL_NO_ACK)) txwi->ack_ctl |= MT_TXWI_ACK_CTL_REQ; if (info->flags & IEEE80211_TX_CTL_ASSIGN_SEQ) txwi->ack_ctl |= MT_TXWI_ACK_CTL_NSEQ; if ((info->flags & IEEE80211_TX_CTL_AMPDU) && sta) { u8 ba_size = IEEE80211_MIN_AMPDU_BUF; ba_size <<= sta->ht_cap.ampdu_factor; ba_size = min_t(int, 63, ba_size); if (info->flags & IEEE80211_TX_CTL_RATE_CTRL_PROBE) ba_size = 0; txwi->ack_ctl |= FIELD_PREP(MT_TXWI_ACK_CTL_BA_WINDOW, ba_size); txwi->flags = cpu_to_le16(MT_TXWI_FLAGS_AMPDU | FIELD_PREP(MT_TXWI_FLAGS_MPDU_DENSITY, sta->ht_cap.ampdu_density)); if (info->flags & IEEE80211_TX_CTL_RATE_CTRL_PROBE) txwi->flags = 0; } txwi->wcid = wcid->idx; is_probe = !!(info->flags & IEEE80211_TX_CTL_RATE_CTRL_PROBE); pkt_id = mt7601u_tx_pktid_enc(dev, rate_ctl & 0x7, is_probe); pkt_len |= FIELD_PREP(MT_TXWI_LEN_PKTID, pkt_id); txwi->len_ctl = cpu_to_le16(pkt_len); return txwi; } void mt7601u_tx(struct ieee80211_hw *hw, struct ieee80211_tx_control *control, struct sk_buff *skb) { struct ieee80211_tx_info *info = IEEE80211_SKB_CB(skb); struct mt7601u_dev *dev = hw->priv; struct ieee80211_vif *vif = info->control.vif; struct ieee80211_sta *sta = control->sta; struct mt76_sta *msta = NULL; struct mt76_wcid *wcid = dev->mon_wcid; struct mt76_txwi *txwi; int pkt_len = skb->len; int hw_q = skb2q(skb); BUILD_BUG_ON(ARRAY_SIZE(info->status.status_driver_data) < 1); info->status.status_driver_data[0] = (void *)(unsigned long)pkt_len; if (mt7601u_skb_rooms(dev, skb) || mt76_insert_hdr_pad(skb)) { ieee80211_free_txskb(dev->hw, skb); return; } if (sta) { msta = (struct mt76_sta *) sta->drv_priv; wcid = &msta->wcid; } else if (vif) { struct mt76_vif *mvif = (struct mt76_vif *)vif->drv_priv; wcid = &mvif->group_wcid; } txwi = mt7601u_push_txwi(dev, skb, sta, wcid, pkt_len); if (mt7601u_dma_enqueue_tx(dev, skb, wcid, hw_q)) return; trace_mt_tx(dev, skb, msta, txwi); } void mt7601u_tx_stat(struct work_struct *work) { struct mt7601u_dev *dev = container_of(work, struct mt7601u_dev, stat_work.work); struct mt76_tx_status stat; unsigned long flags; int cleaned = 0; while (!test_bit(MT7601U_STATE_REMOVED, &dev->state)) { stat = mt7601u_mac_fetch_tx_status(dev); if (!stat.valid) break; mt7601u_tx_pktid_dec(dev, &stat); mt76_send_tx_status(dev, &stat); cleaned++; } trace_mt_tx_status_cleaned(dev, cleaned); spin_lock_irqsave(&dev->tx_lock, flags); if (cleaned) queue_delayed_work(dev->stat_wq, &dev->stat_work, msecs_to_jiffies(10)); else if (test_and_clear_bit(MT7601U_STATE_MORE_STATS, &dev->state)) queue_delayed_work(dev->stat_wq, &dev->stat_work, msecs_to_jiffies(20)); else clear_bit(MT7601U_STATE_READING_STATS, &dev->state); spin_unlock_irqrestore(&dev->tx_lock, flags); } int mt7601u_conf_tx(struct ieee80211_hw *hw, struct ieee80211_vif *vif, u16 queue, const struct ieee80211_tx_queue_params *params) { struct mt7601u_dev *dev = hw->priv; u8 cw_min = 5, cw_max = 10, hw_q = q2hwq(queue); u32 val; /* TODO: should we do funny things with the parameters? * See what mt7601u_set_default_edca() used to do in init.c. */ if (params->cw_min) cw_min = fls(params->cw_min); if (params->cw_max) cw_max = fls(params->cw_max); WARN_ON(params->txop > 0xff); WARN_ON(params->aifs > 0xf); WARN_ON(cw_min > 0xf); WARN_ON(cw_max > 0xf); val = FIELD_PREP(MT_EDCA_CFG_AIFSN, params->aifs) | FIELD_PREP(MT_EDCA_CFG_CWMIN, cw_min) | FIELD_PREP(MT_EDCA_CFG_CWMAX, cw_max); /* TODO: based on user-controlled EnableTxBurst var vendor drv sets * a really long txop on AC0 (see connect.c:2009) but only on * connect? When not connected should be 0. */ if (!hw_q) val |= 0x60; else val |= FIELD_PREP(MT_EDCA_CFG_TXOP, params->txop); mt76_wr(dev, MT_EDCA_CFG_AC(hw_q), val); val = mt76_rr(dev, MT_WMM_TXOP(hw_q)); val &= ~(MT_WMM_TXOP_MASK << MT_WMM_TXOP_SHIFT(hw_q)); val |= params->txop << MT_WMM_TXOP_SHIFT(hw_q); mt76_wr(dev, MT_WMM_TXOP(hw_q), val); val = mt76_rr(dev, MT_WMM_AIFSN); val &= ~(MT_WMM_AIFSN_MASK << MT_WMM_AIFSN_SHIFT(hw_q)); val |= params->aifs << MT_WMM_AIFSN_SHIFT(hw_q); mt76_wr(dev, MT_WMM_AIFSN, val); val = mt76_rr(dev, MT_WMM_CWMIN); val &= ~(MT_WMM_CWMIN_MASK << MT_WMM_CWMIN_SHIFT(hw_q)); val |= cw_min << MT_WMM_CWMIN_SHIFT(hw_q); mt76_wr(dev, MT_WMM_CWMIN, val); val = mt76_rr(dev, MT_WMM_CWMAX); val &= ~(MT_WMM_CWMAX_MASK << MT_WMM_CWMAX_SHIFT(hw_q)); val |= cw_max << MT_WMM_CWMAX_SHIFT(hw_q); mt76_wr(dev, MT_WMM_CWMAX, val); return 0; }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1