Lines Matching refs:mt76_dev
40 mt76_tx_status_lock(struct mt76_dev *dev, struct sk_buff_head *list) in mt76_tx_status_lock()
49 mt76_tx_status_unlock(struct mt76_dev *dev, struct sk_buff_head *list) in mt76_tx_status_unlock()
80 __mt76_tx_status_skb_done(struct mt76_dev *dev, struct sk_buff *skb, u8 flags, in __mt76_tx_status_skb_done()
106 mt76_tx_status_skb_done(struct mt76_dev *dev, struct sk_buff *skb, in mt76_tx_status_skb_done()
114 mt76_tx_status_skb_add(struct mt76_dev *dev, struct mt76_wcid *wcid, in mt76_tx_status_skb_add()
147 mt76_tx_status_skb_get(struct mt76_dev *dev, struct mt76_wcid *wcid, int pktid, in mt76_tx_status_skb_get()
174 mt76_tx_status_check(struct mt76_dev *dev, struct mt76_wcid *wcid, bool flush) in mt76_tx_status_check()
185 mt76_tx_check_non_aql(struct mt76_dev *dev, struct mt76_wcid *wcid, in mt76_tx_check_non_aql()
199 void __mt76_tx_complete_skb(struct mt76_dev *dev, u16 wcid_idx, struct sk_buff *skb, in __mt76_tx_complete_skb()
256 struct mt76_dev *dev = phy->dev; in __mt76_tx_queue_skb()
283 struct mt76_dev *dev = phy->dev; in mt76_tx()
366 struct mt76_dev *dev = phy->dev; in mt76_release_buffered_frames()
415 struct mt76_dev *dev = phy->dev; in mt76_txq_send_burst()
475 struct mt76_dev *dev = phy->dev; in mt76_txq_schedule_list()
560 void mt76_tx_worker_run(struct mt76_dev *dev) in mt76_tx_worker_run()
577 struct mt76_dev *dev = container_of(w, struct mt76_dev, tx_worker); in mt76_tx_worker()
608 struct mt76_dev *dev = phy->dev; in mt76_wake_tx_queue()
658 void mt76_queue_tx_complete(struct mt76_dev *dev, struct mt76_queue *q, in mt76_queue_tx_complete()
671 void __mt76_set_tx_blocked(struct mt76_dev *dev, bool blocked) in __mt76_set_tx_blocked()
691 int mt76_token_consume(struct mt76_dev *dev, struct mt76_txwi_cache **ptxwi) in mt76_token_consume()
712 mt76_token_release(struct mt76_dev *dev, int token, bool *wake) in mt76_token_release()