Lines Matching refs:tx_info
462 struct sk_buff *skb, u32 tx_info) in mt76_dma_tx_queue_skb_raw() argument
479 mt76_dma_add_buf(dev, q, &buf, 1, tx_info, skb, NULL); in mt76_dma_tx_queue_skb_raw()
498 struct mt76_tx_info tx_info = { in mt76_dma_tx_queue_skb() local
523 tx_info.buf[n].addr = t->dma_addr; in mt76_dma_tx_queue_skb()
524 tx_info.buf[n++].len = dev->drv->txwi_size; in mt76_dma_tx_queue_skb()
525 tx_info.buf[n].addr = addr; in mt76_dma_tx_queue_skb()
526 tx_info.buf[n++].len = len; in mt76_dma_tx_queue_skb()
529 if (n == ARRAY_SIZE(tx_info.buf)) in mt76_dma_tx_queue_skb()
537 tx_info.buf[n].addr = addr; in mt76_dma_tx_queue_skb()
538 tx_info.buf[n++].len = iter->len; in mt76_dma_tx_queue_skb()
540 tx_info.nbuf = n; in mt76_dma_tx_queue_skb()
542 if (q->queued + (tx_info.nbuf + 1) / 2 >= q->ndesc - 1) { in mt76_dma_tx_queue_skb()
549 ret = dev->drv->tx_prepare_skb(dev, txwi, qid, wcid, sta, &tx_info); in mt76_dma_tx_queue_skb()
555 return mt76_dma_add_buf(dev, q, tx_info.buf, tx_info.nbuf, in mt76_dma_tx_queue_skb()
556 tx_info.info, tx_info.skb, t); in mt76_dma_tx_queue_skb()
560 dma_unmap_single(dev->dma_dev, tx_info.buf[n].addr, in mt76_dma_tx_queue_skb()
561 tx_info.buf[n].len, DMA_TO_DEVICE); in mt76_dma_tx_queue_skb()
569 if (tx_info.skb == phy->test.tx_skb) in mt76_dma_tx_queue_skb()
577 status.skb = tx_info.skb; in mt76_dma_tx_queue_skb()
578 hw = mt76_tx_status_get_hw(dev, tx_info.skb); in mt76_dma_tx_queue_skb()