Lines Matching refs:mt76_dev
40 mt76_tx_status_lock(struct mt76_dev *dev, struct sk_buff_head *list) in mt76_tx_status_lock()
49 mt76_tx_status_unlock(struct mt76_dev *dev, struct sk_buff_head *list) in mt76_tx_status_unlock()
82 __mt76_tx_status_skb_done(struct mt76_dev *dev, struct sk_buff *skb, u8 flags, in __mt76_tx_status_skb_done()
106 mt76_tx_status_skb_done(struct mt76_dev *dev, struct sk_buff *skb, in mt76_tx_status_skb_done()
114 mt76_tx_status_skb_add(struct mt76_dev *dev, struct mt76_wcid *wcid, in mt76_tx_status_skb_add()
156 mt76_tx_status_skb_get(struct mt76_dev *dev, struct mt76_wcid *wcid, int pktid, in mt76_tx_status_skb_get()
198 mt76_tx_status_check(struct mt76_dev *dev, bool flush) in mt76_tx_status_check()
211 mt76_tx_check_non_aql(struct mt76_dev *dev, struct mt76_wcid *wcid, in mt76_tx_check_non_aql()
225 void __mt76_tx_complete_skb(struct mt76_dev *dev, u16 wcid_idx, struct sk_buff *skb, in __mt76_tx_complete_skb()
282 struct mt76_dev *dev = phy->dev; in __mt76_tx_queue_skb()
309 struct mt76_dev *dev = phy->dev; in mt76_tx()
392 struct mt76_dev *dev = phy->dev; in mt76_release_buffered_frames()
441 struct mt76_dev *dev = phy->dev; in mt76_txq_send_burst()
502 struct mt76_dev *dev = phy->dev; in mt76_txq_schedule_list()
587 void mt76_tx_worker_run(struct mt76_dev *dev) in mt76_tx_worker_run()
604 struct mt76_dev *dev = container_of(w, struct mt76_dev, tx_worker); in mt76_tx_worker()
635 struct mt76_dev *dev = phy->dev; in mt76_wake_tx_queue()
685 void mt76_queue_tx_complete(struct mt76_dev *dev, struct mt76_queue *q, in mt76_queue_tx_complete()
698 void __mt76_set_tx_blocked(struct mt76_dev *dev, bool blocked) in __mt76_set_tx_blocked()
718 int mt76_token_consume(struct mt76_dev *dev, struct mt76_txwi_cache **ptxwi) in mt76_token_consume()
739 mt76_token_release(struct mt76_dev *dev, int token, bool *wake) in mt76_token_release()