Lines Matching refs:mt76_dev
42 mt76_alloc_txwi(struct mt76_dev *dev) in mt76_alloc_txwi()
63 mt76_alloc_rxwi(struct mt76_dev *dev) in mt76_alloc_rxwi()
76 __mt76_get_txwi(struct mt76_dev *dev) in __mt76_get_txwi()
92 __mt76_get_rxwi(struct mt76_dev *dev) in __mt76_get_rxwi()
108 mt76_get_txwi(struct mt76_dev *dev) in mt76_get_txwi()
119 mt76_get_rxwi(struct mt76_dev *dev) in mt76_get_rxwi()
131 mt76_put_txwi(struct mt76_dev *dev, struct mt76_txwi_cache *t) in mt76_put_txwi()
143 mt76_put_rxwi(struct mt76_dev *dev, struct mt76_txwi_cache *t) in mt76_put_rxwi()
155 mt76_free_pending_txwi(struct mt76_dev *dev) in mt76_free_pending_txwi()
169 mt76_free_pending_rxwi(struct mt76_dev *dev) in mt76_free_pending_rxwi()
184 mt76_dma_sync_idx(struct mt76_dev *dev, struct mt76_queue *q) in mt76_dma_sync_idx()
193 mt76_dma_queue_reset(struct mt76_dev *dev, struct mt76_queue *q) in mt76_dma_queue_reset()
210 mt76_dma_add_rx_buf(struct mt76_dev *dev, struct mt76_queue *q, in mt76_dma_add_rx_buf()
255 mt76_dma_add_buf(struct mt76_dev *dev, struct mt76_queue *q, in mt76_dma_add_buf()
317 mt76_dma_tx_cleanup_idx(struct mt76_dev *dev, struct mt76_queue *q, int idx, in mt76_dma_tx_cleanup_idx()
341 mt76_dma_kick_queue(struct mt76_dev *dev, struct mt76_queue *q) in mt76_dma_kick_queue()
348 mt76_dma_tx_cleanup(struct mt76_dev *dev, struct mt76_queue *q, bool flush) in mt76_dma_tx_cleanup()
388 mt76_dma_get_buf(struct mt76_dev *dev, struct mt76_queue *q, int idx, in mt76_dma_get_buf()
440 mt76_dma_dequeue(struct mt76_dev *dev, struct mt76_queue *q, bool flush, in mt76_dma_dequeue()
461 mt76_dma_tx_queue_skb_raw(struct mt76_dev *dev, struct mt76_queue *q, in mt76_dma_tx_queue_skb_raw()
491 mt76_dma_tx_queue_skb(struct mt76_dev *dev, struct mt76_queue *q, in mt76_dma_tx_queue_skb()
585 mt76_dma_rx_fill(struct mt76_dev *dev, struct mt76_queue *q, in mt76_dma_rx_fill()
629 int mt76_dma_wed_setup(struct mt76_dev *dev, struct mt76_queue *q, bool reset) in mt76_dma_wed_setup()
683 mt76_dma_alloc_queue(struct mt76_dev *dev, struct mt76_queue *q, in mt76_dma_alloc_queue()
722 mt76_dma_rx_cleanup(struct mt76_dev *dev, struct mt76_queue *q) in mt76_dma_rx_cleanup()
749 mt76_dma_rx_reset(struct mt76_dev *dev, enum mt76_rxq_id qid) in mt76_dma_rx_reset()
771 mt76_add_fragment(struct mt76_dev *dev, struct mt76_queue *q, void *data, in mt76_add_fragment()
798 mt76_dma_rx_process(struct mt76_dev *dev, struct mt76_queue *q, int budget) in mt76_dma_rx_process()
882 struct mt76_dev *dev; in mt76_dma_rx_poll()
885 dev = container_of(napi->dev, struct mt76_dev, napi_dev); in mt76_dma_rx_poll()
906 mt76_dma_init(struct mt76_dev *dev, in mt76_dma_init()
940 void mt76_dma_attach(struct mt76_dev *dev) in mt76_dma_attach()
946 void mt76_dma_cleanup(struct mt76_dev *dev) in mt76_dma_cleanup()