Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Prameela Rani Garnepudi | 740 | 97.50% | 1 | 20.00% |
Siva Rebbagondla | 17 | 2.24% | 2 | 40.00% |
Lee Jones | 1 | 0.13% | 1 | 20.00% |
Eric W. Biedermann | 1 | 0.13% | 1 | 20.00% |
Total | 759 | 5 |
/* * Copyright (c) 2018 Redpine Signals Inc. * * Permission to use, copy, modify, and/or distribute this software for any * purpose with or without fee is hereby granted, provided that the above * copyright notice and this permission notice appear in all copies. * * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE. */ #include "rsi_main.h" #include "rsi_coex.h" #include "rsi_mgmt.h" #include "rsi_hal.h" static enum rsi_coex_queues rsi_coex_determine_coex_q (struct rsi_coex_ctrl_block *coex_cb) { enum rsi_coex_queues q_num = RSI_COEX_Q_INVALID; if (skb_queue_len(&coex_cb->coex_tx_qs[RSI_COEX_Q_COMMON]) > 0) q_num = RSI_COEX_Q_COMMON; if (skb_queue_len(&coex_cb->coex_tx_qs[RSI_COEX_Q_BT]) > 0) q_num = RSI_COEX_Q_BT; if (skb_queue_len(&coex_cb->coex_tx_qs[RSI_COEX_Q_WLAN]) > 0) q_num = RSI_COEX_Q_WLAN; return q_num; } static void rsi_coex_sched_tx_pkts(struct rsi_coex_ctrl_block *coex_cb) { enum rsi_coex_queues coex_q = RSI_COEX_Q_INVALID; struct sk_buff *skb; do { coex_q = rsi_coex_determine_coex_q(coex_cb); rsi_dbg(INFO_ZONE, "queue = %d\n", coex_q); if (coex_q == RSI_COEX_Q_BT) { skb = skb_dequeue(&coex_cb->coex_tx_qs[RSI_COEX_Q_BT]); rsi_send_bt_pkt(coex_cb->priv, skb); } } while (coex_q != RSI_COEX_Q_INVALID); } static void rsi_coex_scheduler_thread(struct rsi_common *common) { struct rsi_coex_ctrl_block *coex_cb = (struct rsi_coex_ctrl_block *)common->coex_cb; u32 timeout = EVENT_WAIT_FOREVER; do { rsi_wait_event(&coex_cb->coex_tx_thread.event, timeout); rsi_reset_event(&coex_cb->coex_tx_thread.event); rsi_coex_sched_tx_pkts(coex_cb); } while (atomic_read(&coex_cb->coex_tx_thread.thread_done) == 0); kthread_complete_and_exit(&coex_cb->coex_tx_thread.completion, 0); } int rsi_coex_recv_pkt(struct rsi_common *common, u8 *msg) { u8 msg_type = msg[RSI_RX_DESC_MSG_TYPE_OFFSET]; switch (msg_type) { case COMMON_CARD_READY_IND: rsi_dbg(INFO_ZONE, "common card ready received\n"); common->hibernate_resume = false; rsi_handle_card_ready(common, msg); break; case SLEEP_NOTIFY_IND: rsi_dbg(INFO_ZONE, "sleep notify received\n"); rsi_mgmt_pkt_recv(common, msg); break; } return 0; } static inline int rsi_map_coex_q(u8 hal_queue) { switch (hal_queue) { case RSI_COEX_Q: return RSI_COEX_Q_COMMON; case RSI_WLAN_Q: return RSI_COEX_Q_WLAN; case RSI_BT_Q: return RSI_COEX_Q_BT; } return RSI_COEX_Q_INVALID; } int rsi_coex_send_pkt(void *priv, struct sk_buff *skb, u8 hal_queue) { struct rsi_common *common = (struct rsi_common *)priv; struct rsi_coex_ctrl_block *coex_cb = (struct rsi_coex_ctrl_block *)common->coex_cb; struct skb_info *tx_params = NULL; enum rsi_coex_queues coex_q; int status; coex_q = rsi_map_coex_q(hal_queue); if (coex_q == RSI_COEX_Q_INVALID) { rsi_dbg(ERR_ZONE, "Invalid coex queue\n"); return -EINVAL; } if (coex_q != RSI_COEX_Q_COMMON && coex_q != RSI_COEX_Q_WLAN) { skb_queue_tail(&coex_cb->coex_tx_qs[coex_q], skb); rsi_set_event(&coex_cb->coex_tx_thread.event); return 0; } if (common->iface_down) { tx_params = (struct skb_info *)&IEEE80211_SKB_CB(skb)->driver_data; if (!(tx_params->flags & INTERNAL_MGMT_PKT)) { rsi_indicate_tx_status(common->priv, skb, -EINVAL); return 0; } } /* Send packet to hal */ if (skb->priority == MGMT_SOFT_Q) status = rsi_send_mgmt_pkt(common, skb); else status = rsi_send_data_pkt(common, skb); return status; } int rsi_coex_attach(struct rsi_common *common) { struct rsi_coex_ctrl_block *coex_cb; int cnt; coex_cb = kzalloc(sizeof(*coex_cb), GFP_KERNEL); if (!coex_cb) return -ENOMEM; common->coex_cb = (void *)coex_cb; coex_cb->priv = common; /* Initialize co-ex queues */ for (cnt = 0; cnt < NUM_COEX_TX_QUEUES; cnt++) skb_queue_head_init(&coex_cb->coex_tx_qs[cnt]); rsi_init_event(&coex_cb->coex_tx_thread.event); /* Initialize co-ex thread */ if (rsi_create_kthread(common, &coex_cb->coex_tx_thread, rsi_coex_scheduler_thread, "Coex-Tx-Thread")) { rsi_dbg(ERR_ZONE, "%s: Unable to init tx thrd\n", __func__); return -EINVAL; } return 0; } void rsi_coex_detach(struct rsi_common *common) { struct rsi_coex_ctrl_block *coex_cb = (struct rsi_coex_ctrl_block *)common->coex_cb; int cnt; rsi_kill_thread(&coex_cb->coex_tx_thread); for (cnt = 0; cnt < NUM_COEX_TX_QUEUES; cnt++) skb_queue_purge(&coex_cb->coex_tx_qs[cnt]); kfree(coex_cb); }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1