• Home
  • Raw
  • Download

Lines Matching refs:wcid

64 		struct mt76_wcid *wcid;  in mt76_tx_status_unlock()  local
66 wcid = rcu_dereference(dev->wcid[cb->wcid]); in mt76_tx_status_unlock()
67 if (wcid) in mt76_tx_status_unlock()
68 status.sta = wcid_to_sta(wcid); in mt76_tx_status_unlock()
114 mt76_tx_status_skb_add(struct mt76_dev *dev, struct mt76_wcid *wcid, in mt76_tx_status_skb_add() argument
121 if (!wcid) in mt76_tx_status_skb_add()
134 pid = mt76_get_next_pkt_id(wcid); in mt76_tx_status_skb_add()
135 cb->wcid = wcid->idx; in mt76_tx_status_skb_add()
147 mt76_tx_status_skb_get(struct mt76_dev *dev, struct mt76_wcid *wcid, int pktid, in mt76_tx_status_skb_get() argument
155 if (wcid && cb->wcid != wcid->idx) in mt76_tx_status_skb_get()
174 mt76_tx_status_check(struct mt76_dev *dev, struct mt76_wcid *wcid, bool flush) in mt76_tx_status_check() argument
179 mt76_tx_status_skb_get(dev, wcid, flush ? -1 : 0, &list); in mt76_tx_status_check()
185 mt76_tx_check_non_aql(struct mt76_dev *dev, struct mt76_wcid *wcid, in mt76_tx_check_non_aql() argument
191 if (!wcid || info->tx_time_est) in mt76_tx_check_non_aql()
194 pending = atomic_dec_return(&wcid->non_aql_packets); in mt76_tx_check_non_aql()
196 atomic_cmpxchg(&wcid->non_aql_packets, pending, 0); in mt76_tx_check_non_aql()
206 struct mt76_wcid *wcid = NULL; in __mt76_tx_complete_skb() local
212 if (wcid_idx < ARRAY_SIZE(dev->wcid)) in __mt76_tx_complete_skb()
213 wcid = rcu_dereference(dev->wcid[wcid_idx]); in __mt76_tx_complete_skb()
215 mt76_tx_check_non_aql(dev, wcid, skb); in __mt76_tx_complete_skb()
233 status.sta = wcid_to_sta(wcid); in __mt76_tx_complete_skb()
251 struct mt76_wcid *wcid, struct ieee80211_sta *sta, in __mt76_tx_queue_skb() argument
262 idx = dev->queue_ops->tx_queue_skb(dev, q, qid, skb, wcid, sta); in __mt76_tx_queue_skb()
266 wcid = (struct mt76_wcid *)sta->drv_priv; in __mt76_tx_queue_skb()
267 q->entry[idx].wcid = wcid->idx; in __mt76_tx_queue_skb()
272 pending = atomic_inc_return(&wcid->non_aql_packets); in __mt76_tx_queue_skb()
281 struct mt76_wcid *wcid, struct sk_buff *skb) in mt76_tx() argument
308 if (wcid && !(wcid->tx_info & MT_WCID_TX_INFO_SET)) in mt76_tx()
318 __mt76_tx_queue_skb(phy, qid, skb, wcid, sta, NULL); in mt76_tx()
347 struct mt76_wcid *wcid = (struct mt76_wcid *)sta->drv_priv; in mt76_queue_ps_skb() local
356 __mt76_tx_queue_skb(phy, MT_TXQ_PSD, skb, wcid, sta, NULL); in mt76_queue_ps_skb()
413 struct mt76_txq *mtxq, struct mt76_wcid *wcid) in mt76_txq_send_burst() argument
424 if (test_bit(MT_WCID_FLAG_PS, &wcid->flags)) in mt76_txq_send_burst()
427 if (atomic_read(&wcid->non_aql_packets) >= MT_MAX_NON_AQL_PKT) in mt76_txq_send_burst()
435 if (!(wcid->tx_info & MT_WCID_TX_INFO_SET)) in mt76_txq_send_burst()
439 idx = __mt76_tx_queue_skb(phy, qid, skb, wcid, txq->sta, &stop); in mt76_txq_send_burst()
455 if (!(wcid->tx_info & MT_WCID_TX_INFO_SET)) in mt76_txq_send_burst()
459 idx = __mt76_tx_queue_skb(phy, qid, skb, wcid, txq->sta, &stop); in mt76_txq_send_burst()
478 struct mt76_wcid *wcid; in mt76_txq_schedule_list() local
497 wcid = rcu_dereference(dev->wcid[mtxq->wcid]); in mt76_txq_schedule_list()
498 if (!wcid || test_bit(MT_WCID_FLAG_PS, &wcid->flags)) in mt76_txq_schedule_list()
517 n_frames = mt76_txq_send_burst(phy, q, mtxq, wcid); in mt76_txq_schedule_list()