| /linux/drivers/net/ethernet/stmicro/stmmac/ |
| A D | ring_mode.c | 22 unsigned int entry = tx_q->cur_tx; in jumbo_frm() 29 desc = tx_q->dma_tx + entry; in jumbo_frm() 46 tx_q->tx_skbuff_dma[entry].buf = des2; in jumbo_frm() 47 tx_q->tx_skbuff_dma[entry].len = bmax; in jumbo_frm() 53 tx_q->tx_skbuff[entry] = NULL; in jumbo_frm() 59 desc = tx_q->dma_tx + entry; in jumbo_frm() 66 tx_q->tx_skbuff_dma[entry].buf = des2; in jumbo_frm() 67 tx_q->tx_skbuff_dma[entry].len = len; in jumbo_frm() 80 tx_q->tx_skbuff_dma[entry].buf = des2; in jumbo_frm() 89 tx_q->cur_tx = entry; in jumbo_frm() [all …]
|
| A D | chain_mode.c | 22 unsigned int entry = tx_q->cur_tx; in jumbo_frm() 27 desc = tx_q->dma_tx + entry; in jumbo_frm() 41 tx_q->tx_skbuff_dma[entry].buf = des2; in jumbo_frm() 42 tx_q->tx_skbuff_dma[entry].len = bmax; in jumbo_frm() 48 tx_q->tx_skbuff[entry] = NULL; in jumbo_frm() 50 desc = tx_q->dma_tx + entry; in jumbo_frm() 59 tx_q->tx_skbuff_dma[entry].buf = des2; in jumbo_frm() 73 tx_q->tx_skbuff_dma[entry].len = len; in jumbo_frm() 81 tx_q->cur_tx = entry; in jumbo_frm() 146 unsigned int entry = tx_q->dirty_tx; in clean_desc3() [all …]
|
| A D | stmmac_main.c | 367 if (tx_q->dirty_tx > tx_q->cur_tx) in stmmac_tx_avail() 368 avail = tx_q->dirty_tx - tx_q->cur_tx - 1; in stmmac_tx_avail() 418 if (tx_q->dirty_tx != tx_q->cur_tx) in stmmac_enable_eee_mode() 2792 if (tx_q->dirty_tx != tx_q->cur_tx) in stmmac_tx_clean() 3070 tx_q->tx_tail_addr = tx_q->dma_tx_phy; in stmmac_init_dma_engine() 4101 p = &tx_q->dma_tx[tx_q->cur_tx]; in stmmac_vlan_insert() 4142 desc = &tx_q->dma_tx[tx_q->cur_tx]; in stmmac_tso_allocator() 4180 tx_q->tx_tail_addr = tx_q->dma_tx_phy + (tx_q->cur_tx * desc_size); in stmmac_flush_tx_descriptors() 4365 tx_q->tx_skbuff[tx_q->cur_tx] = skb; in stmmac_tso_xmit() 4388 desc = &tx_q->dma_tx[tx_q->cur_tx]; in stmmac_tso_xmit() [all …]
|
| /linux/drivers/net/ethernet/intel/idpf/ |
| A D | idpf_singleq_txrx.c | 202 u16 i = tx_q->next_to_use; in idpf_tx_singleq_map() 295 nq = netdev_get_tx_queue(tx_q->netdev, tx_q->idx); in idpf_tx_singleq_map() 376 idpf_tx_buf_hw_update(tx_q, tx_q->next_to_use, false); in idpf_tx_singleq_frame() 403 first = &tx_q->tx_buf[tx_q->next_to_use]; in idpf_tx_singleq_frame() 436 .dev = tx_q->dev, in idpf_tx_singleq_clean() 447 ntc -= tx_q->desc_count; in idpf_tx_singleq_clean() 485 tx_buf = tx_q->tx_buf; in idpf_tx_singleq_clean() 501 ntc -= tx_q->desc_count; in idpf_tx_singleq_clean() 502 tx_buf = tx_q->tx_buf; in idpf_tx_singleq_clean() 507 ntc += tx_q->desc_count; in idpf_tx_singleq_clean() [all …]
|
| A D | idpf_txrx.c | 252 tx_q->size = tx_q->desc_count * sizeof(*tx_q->base_tx); in idpf_tx_desc_alloc() 255 tx_q->size = ALIGN(tx_q->size, 4096); in idpf_tx_desc_alloc() 256 tx_q->desc_ring = dmam_alloc_coherent(dev, tx_q->size, &tx_q->dma, in idpf_tx_desc_alloc() 260 tx_q->size); in idpf_tx_desc_alloc() 2075 nq = netdev_get_tx_queue(tx_q->netdev, tx_q->idx); in idpf_tx_clean_complq() 2165 netif_stop_subqueue(tx_q->netdev, tx_q->idx); in idpf_tx_maybe_stop_splitq() 2190 nq = netdev_get_tx_queue(tx_q->netdev, tx_q->idx); in idpf_tx_buf_hw_update() 2506 nq = netdev_get_tx_queue(tx_q->netdev, tx_q->idx); in idpf_tx_splitq_map() 2727 idpf_tx_buf_hw_update(tx_q, tx_q->next_to_use, false); in idpf_tx_drop_skb() 2760 idpf_tx_buf_hw_update(tx_q, tx_q->next_to_use, false); in idpf_tx_splitq_frame() [all …]
|
| A D | idpf_txrx.h | 1038 void idpf_tx_buf_hw_update(struct idpf_tx_queue *tx_q, u32 val, 1041 netdev_tx_t idpf_tx_drop_skb(struct idpf_tx_queue *tx_q, struct sk_buff *skb); 1048 struct idpf_tx_queue *tx_q); 1054 static inline bool idpf_tx_maybe_stop_common(struct idpf_tx_queue *tx_q, in idpf_tx_maybe_stop_common() argument 1057 return !netif_subqueue_maybe_stop(tx_q->netdev, tx_q->idx, in idpf_tx_maybe_stop_common() 1058 IDPF_DESC_UNUSED(tx_q), in idpf_tx_maybe_stop_common()
|
| /linux/drivers/net/ethernet/qualcomm/emac/ |
| A D | emac-mac.c | 587 struct emac_tx_queue *tx_q = &adpt->tx_q; in emac_tx_q_descs_free() local 614 memset(tx_q->tpd.v_addr, 0, tx_q->tpd.size); in emac_tx_q_descs_free() 662 struct emac_tx_queue *tx_q = &adpt->tx_q; in emac_tx_q_bufs_free() local 686 tx_q->tpd.size = tx_q->tpd.count * (adpt->tpd_size * 4); in emac_tx_q_desc_alloc() 1030 tx_q->tpd.last_produce_idx = tx_q->tpd.produce_idx; in emac_tx_tpd_create() 1033 if (++tx_q->tpd.produce_idx == tx_q->tpd.count) in emac_tx_tpd_create() 1188 tpbuf = GET_TPD_BUFFER(tx_q, tx_q->tpd.consume_idx); in emac_mac_tx_process() 1203 if (++tx_q->tpd.consume_idx == tx_q->tpd.count) in emac_mac_tx_process() 1347 tpbuf = GET_TPD_BUFFER(tx_q, tx_q->tpd.produce_idx); in emac_tx_fill_tpd() 1367 tpbuf = GET_TPD_BUFFER(tx_q, tx_q->tpd.produce_idx); in emac_tx_fill_tpd() [all …]
|
| A D | emac-mac.h | 231 struct emac_tx_queue *tx_q, 233 void emac_mac_tx_process(struct emac_adapter *adpt, struct emac_tx_queue *tx_q);
|
| A D | emac.h | 341 struct emac_tx_queue tx_q; member
|
| A D | emac.c | 122 return emac_mac_tx_buf_send(adpt, &adpt->tx_q, skb); in emac_start_xmit() 160 emac_mac_tx_process(adpt, &adpt->tx_q); in emac_isr()
|
| /linux/drivers/net/wwan/t7xx/ |
| A D | t7xx_hif_dpmaif.c | 211 struct dpmaif_tx_queue *tx_q; in t7xx_dpmaif_rxtx_sw_allocs() local 236 tx_q = &dpmaif_ctrl->txq[tx_idx]; in t7xx_dpmaif_rxtx_sw_allocs() 237 tx_q->index = tx_idx; in t7xx_dpmaif_rxtx_sw_allocs() 238 tx_q->dpmaif_ctrl = dpmaif_ctrl; in t7xx_dpmaif_rxtx_sw_allocs() 239 ret = t7xx_dpmaif_txq_init(tx_q); in t7xx_dpmaif_rxtx_sw_allocs() 261 tx_q = &dpmaif_ctrl->txq[i]; in t7xx_dpmaif_rxtx_sw_allocs() 262 t7xx_dpmaif_txq_free(tx_q); in t7xx_dpmaif_rxtx_sw_allocs() 282 struct dpmaif_tx_queue *tx_q; in t7xx_dpmaif_sw_release() local 289 tx_q = &dpmaif_ctrl->txq[i]; in t7xx_dpmaif_sw_release() 290 t7xx_dpmaif_txq_free(tx_q); in t7xx_dpmaif_sw_release()
|
| /linux/drivers/net/ethernet/intel/ice/ |
| A D | ice_dcb_lib.h | 44 static inline bool ice_find_q_in_range(u16 low, u16 high, unsigned int tx_q) in ice_find_q_in_range() argument 46 return (tx_q >= low) && (tx_q < high); in ice_find_q_in_range()
|
| /linux/net/nfc/nci/ |
| A D | uart.c | 44 skb = skb_dequeue(&nu->tx_q); in nci_uart_dequeue() 56 return skb_queue_empty(&nu->tx_q); in nci_uart_queue_empty() 123 skb_queue_head_init(&nu->tx_q); in nci_uart_set_driver() 186 skb_queue_purge(&nu->tx_q); in nci_uart_tty_close() 363 skb_queue_tail(&nu->tx_q, skb); in nci_uart_send()
|
| A D | data.c | 146 spin_lock_irqsave(&ndev->tx_q.lock, flags); in nci_queue_tx_data_frags() 149 __skb_queue_tail(&ndev->tx_q, skb_frag); in nci_queue_tx_data_frags() 151 spin_unlock_irqrestore(&ndev->tx_q.lock, flags); in nci_queue_tx_data_frags() 185 skb_queue_tail(&ndev->tx_q, skb); in nci_send_data()
|
| A D | core.c | 543 skb_queue_purge(&ndev->tx_q); in nci_open_device() 576 skb_queue_purge(&ndev->tx_q); in nci_close_device() 1258 skb_queue_head_init(&ndev->tx_q); in nci_register_device() 1495 skb = skb_dequeue(&ndev->tx_q); in nci_tx_work()
|
| A D | ntf.c | 78 if (!skb_queue_empty(&ndev->tx_q)) in nci_core_conn_credits_ntf_packet() 709 skb_queue_purge(&ndev->tx_q); in nci_rf_deactivate_ntf_packet()
|
| /linux/drivers/net/wireless/mediatek/mt7601u/ |
| A D | dma.c | 312 struct mt7601u_tx_queue *q = &dev->tx_q[ep]; in mt7601u_dma_submit_tx() 477 if (!dev->tx_q) in mt7601u_free_tx() 481 mt7601u_free_tx_queue(&dev->tx_q[i]); in mt7601u_free_tx() 505 dev->tx_q = devm_kcalloc(dev->dev, __MT_EP_OUT_MAX, in mt7601u_alloc_tx() 506 sizeof(*dev->tx_q), GFP_KERNEL); in mt7601u_alloc_tx() 507 if (!dev->tx_q) in mt7601u_alloc_tx() 511 if (mt7601u_alloc_tx_queue(dev, &dev->tx_q[i])) in mt7601u_alloc_tx()
|
| /linux/drivers/net/fddi/skfp/h/ |
| A D | hwmtm.h | 242 #define HWM_GET_TX_USED(smc,queue) (int) (smc)->hw.fp.tx_q[queue].tx_used 261 (smc)->hw.fp.tx_q[queue].tx_curr_put
|
| A D | fplustm.h | 191 struct s_smt_tx_queue tx_q[USED_QUEUES] ; member
|
| /linux/drivers/net/wireless/mediatek/mt76/ |
| A D | mt76_connac_mac.c | 89 if (wcid && pm->tx_q[i].wcid != wcid) in mt76_connac_free_pending_tx_skbs() 92 dev_kfree_skb(pm->tx_q[i].skb); in mt76_connac_free_pending_tx_skbs() 93 pm->tx_q[i].skb = NULL; in mt76_connac_free_pending_tx_skbs() 108 if (!pm->tx_q[qid].skb) { in mt76_connac_pm_queue_skb() 110 pm->tx_q[qid].wcid = wcid; in mt76_connac_pm_queue_skb() 111 pm->tx_q[qid].skb = skb; in mt76_connac_pm_queue_skb() 127 struct mt76_wcid *wcid = pm->tx_q[i].wcid; in mt76_connac_pm_dequeue_skbs() 130 if (!pm->tx_q[i].skb) in mt76_connac_pm_dequeue_skbs() 137 mt76_tx(phy, sta, wcid, pm->tx_q[i].skb); in mt76_connac_pm_dequeue_skbs() 138 pm->tx_q[i].skb = NULL; in mt76_connac_pm_dequeue_skbs()
|
| /linux/drivers/media/radio/wl128x/ |
| A D | fmdrv.h | 206 struct sk_buff_head tx_q; /* TX queue */ member
|
| A D | fmdrv_common.c | 328 if (!skb_queue_empty(&fmdev->tx_q)) in recv_bh_work() 352 skb = skb_dequeue(&fmdev->tx_q); in send_bh_work() 442 skb_queue_tail(&fmdev->tx_q, skb); in fm_send_cmd() 1541 skb_queue_head_init(&fmdev->tx_q); in fmc_prepare() 1595 skb_queue_purge(&fmdev->tx_q); in fmc_release()
|
| /linux/net/bluetooth/ |
| A D | l2cap_core.c | 453 skb_queue_head_init(&chan->tx_q); in l2cap_chan_create() 576 skb_queue_head_init(&chan->tx_q); in l2cap_le_flowctl_init() 686 skb_queue_purge(&chan->tx_q); in l2cap_chan_del() 701 skb_queue_purge(&chan->tx_q); in l2cap_chan_del() 1935 skb = skb_dequeue(&chan->tx_q); in l2cap_streaming_send() 2129 skb_queue_walk(&chan->tx_q, skb) { in l2cap_retransmit_all() 2515 skb_queue_len(&chan->tx_q)); in l2cap_le_flowctl_send() 2705 skb_unlink(acked_skb, &chan->tx_q); in l2cap_process_reqseq() 3139 skb_queue_head_init(&chan->tx_q); in l2cap_ertm_init() 6263 if (!skb_queue_empty(&chan->tx_q)) in l2cap_rx_state_wait_p() [all …]
|
| /linux/include/net/nfc/ |
| A D | nci_core.h | 227 struct sk_buff_head tx_q; member 454 struct sk_buff_head tx_q; member
|
| /linux/drivers/net/ethernet/chelsio/cxgb3/ |
| A D | adapter.h | 208 struct netdev_queue *tx_q; /* associated netdev TX queue */ member
|