Home
last modified time | relevance | path

Searched refs:tx_q (Results 1 – 25 of 25) sorted by relevance

/drivers/net/ethernet/stmicro/stmmac/
A Dring_mode.c22 unsigned int entry = tx_q->cur_tx; in jumbo_frm()
29 desc = tx_q->dma_tx + entry; in jumbo_frm()
46 tx_q->tx_skbuff_dma[entry].buf = des2; in jumbo_frm()
47 tx_q->tx_skbuff_dma[entry].len = bmax; in jumbo_frm()
53 tx_q->tx_skbuff[entry] = NULL; in jumbo_frm()
59 desc = tx_q->dma_tx + entry; in jumbo_frm()
66 tx_q->tx_skbuff_dma[entry].buf = des2; in jumbo_frm()
67 tx_q->tx_skbuff_dma[entry].len = len; in jumbo_frm()
80 tx_q->tx_skbuff_dma[entry].buf = des2; in jumbo_frm()
89 tx_q->cur_tx = entry; in jumbo_frm()
[all …]
A Dchain_mode.c22 unsigned int entry = tx_q->cur_tx; in jumbo_frm()
27 desc = tx_q->dma_tx + entry; in jumbo_frm()
41 tx_q->tx_skbuff_dma[entry].buf = des2; in jumbo_frm()
42 tx_q->tx_skbuff_dma[entry].len = bmax; in jumbo_frm()
48 tx_q->tx_skbuff[entry] = NULL; in jumbo_frm()
50 desc = tx_q->dma_tx + entry; in jumbo_frm()
59 tx_q->tx_skbuff_dma[entry].buf = des2; in jumbo_frm()
73 tx_q->tx_skbuff_dma[entry].len = len; in jumbo_frm()
81 tx_q->cur_tx = entry; in jumbo_frm()
146 unsigned int entry = tx_q->dirty_tx; in clean_desc3()
[all …]
A Dstmmac_main.c397 if (tx_q->dirty_tx > tx_q->cur_tx) in stmmac_tx_avail()
398 avail = tx_q->dirty_tx - tx_q->cur_tx - 1; in stmmac_tx_avail()
432 if (tx_q->dirty_tx != tx_q->cur_tx) in stmmac_eee_tx_busy()
2895 if (tx_q->dirty_tx != tx_q->cur_tx) in stmmac_tx_clean()
3173 tx_q->tx_tail_addr = tx_q->dma_tx_phy; in stmmac_init_dma_engine()
4199 p = &tx_q->dma_tx[tx_q->cur_tx]; in stmmac_vlan_insert()
4240 desc = &tx_q->dma_tx[tx_q->cur_tx]; in stmmac_tso_allocator()
4274 tx_q->tx_tail_addr = tx_q->dma_tx_phy + (tx_q->cur_tx * desc_size); in stmmac_flush_tx_descriptors()
4454 tx_q->tx_skbuff[tx_q->cur_tx] = skb; in stmmac_tso_xmit()
4477 desc = &tx_q->dma_tx[tx_q->cur_tx]; in stmmac_tso_xmit()
[all …]
A Dhwif.h537 int (*jumbo_frm)(struct stmmac_tx_queue *tx_q, struct sk_buff *skb,
542 void (*clean_desc3)(struct stmmac_tx_queue *tx_q, struct dma_desc *p);
/drivers/net/ethernet/intel/idpf/
A Didpf_singleq_txrx.c202 u16 i = tx_q->next_to_use; in idpf_tx_singleq_map()
295 nq = netdev_get_tx_queue(tx_q->netdev, tx_q->idx); in idpf_tx_singleq_map()
374 if (!netif_subqueue_maybe_stop(tx_q->netdev, tx_q->idx, in idpf_tx_singleq_frame()
377 idpf_tx_buf_hw_update(tx_q, tx_q->next_to_use, false); in idpf_tx_singleq_frame()
404 first = &tx_q->tx_buf[tx_q->next_to_use]; in idpf_tx_singleq_frame()
437 .dev = tx_q->dev, in idpf_tx_singleq_clean()
448 ntc -= tx_q->desc_count; in idpf_tx_singleq_clean()
486 tx_buf = tx_q->tx_buf; in idpf_tx_singleq_clean()
503 tx_buf = tx_q->tx_buf; in idpf_tx_singleq_clean()
508 ntc += tx_q->desc_count; in idpf_tx_singleq_clean()
[all …]
A Didpf_txrx.c253 tx_q->size = tx_q->desc_count * sizeof(*tx_q->base_tx); in idpf_tx_desc_alloc()
256 tx_q->size = ALIGN(tx_q->size, 4096); in idpf_tx_desc_alloc()
257 tx_q->desc_ring = dmam_alloc_coherent(dev, tx_q->size, &tx_q->dma, in idpf_tx_desc_alloc()
261 tx_q->size); in idpf_tx_desc_alloc()
2127 nq = netdev_get_tx_queue(tx_q->netdev, tx_q->idx); in idpf_tx_clean_complq()
2210 if (netif_subqueue_maybe_stop(tx_q->netdev, tx_q->idx, in idpf_tx_maybe_stop_splitq()
2237 nq = netdev_get_tx_queue(tx_q->netdev, tx_q->idx); in idpf_tx_buf_hw_update()
2548 nq = netdev_get_tx_queue(tx_q->netdev, tx_q->idx); in idpf_tx_splitq_map()
2769 idpf_tx_buf_hw_update(tx_q, tx_q->next_to_use, false); in idpf_tx_drop_skb()
2870 idpf_tx_buf_hw_update(tx_q, tx_q->next_to_use, false); in idpf_tx_splitq_frame()
[all …]
A Didpf_ptp.c640 int idpf_ptp_request_ts(struct idpf_tx_queue *tx_q, struct sk_buff *skb, in idpf_ptp_request_ts() argument
647 spin_lock(&tx_q->cached_tstamp_caps->latches_lock); in idpf_ptp_request_ts()
649 head = &tx_q->cached_tstamp_caps->latches_free; in idpf_ptp_request_ts()
651 spin_unlock(&tx_q->cached_tstamp_caps->latches_lock); in idpf_ptp_request_ts()
664 &tx_q->cached_tstamp_caps->latches_in_use); in idpf_ptp_request_ts()
665 spin_unlock(&tx_q->cached_tstamp_caps->latches_lock); in idpf_ptp_request_ts()
A Didpf_ptp.h286 int idpf_ptp_request_ts(struct idpf_tx_queue *tx_q, struct sk_buff *skb,
371 static inline int idpf_ptp_request_ts(struct idpf_tx_queue *tx_q, in idpf_ptp_request_ts() argument
A Didpf_txrx.h1037 void idpf_tx_buf_hw_update(struct idpf_tx_queue *tx_q, u32 val,
1040 netdev_tx_t idpf_tx_drop_skb(struct idpf_tx_queue *tx_q, struct sk_buff *skb);
1047 struct idpf_tx_queue *tx_q);
/drivers/net/ethernet/qualcomm/emac/
A Demac-mac.c587 struct emac_tx_queue *tx_q = &adpt->tx_q; in emac_tx_q_descs_free() local
614 memset(tx_q->tpd.v_addr, 0, tx_q->tpd.size); in emac_tx_q_descs_free()
662 struct emac_tx_queue *tx_q = &adpt->tx_q; in emac_tx_q_bufs_free() local
686 tx_q->tpd.size = tx_q->tpd.count * (adpt->tpd_size * 4); in emac_tx_q_desc_alloc()
1030 tx_q->tpd.last_produce_idx = tx_q->tpd.produce_idx; in emac_tx_tpd_create()
1033 if (++tx_q->tpd.produce_idx == tx_q->tpd.count) in emac_tx_tpd_create()
1188 tpbuf = GET_TPD_BUFFER(tx_q, tx_q->tpd.consume_idx); in emac_mac_tx_process()
1203 if (++tx_q->tpd.consume_idx == tx_q->tpd.count) in emac_mac_tx_process()
1347 tpbuf = GET_TPD_BUFFER(tx_q, tx_q->tpd.produce_idx); in emac_tx_fill_tpd()
1367 tpbuf = GET_TPD_BUFFER(tx_q, tx_q->tpd.produce_idx); in emac_tx_fill_tpd()
[all …]
A Demac-mac.h231 struct emac_tx_queue *tx_q,
233 void emac_mac_tx_process(struct emac_adapter *adpt, struct emac_tx_queue *tx_q);
A Demac.c122 return emac_mac_tx_buf_send(adpt, &adpt->tx_q, skb); in emac_start_xmit()
160 emac_mac_tx_process(adpt, &adpt->tx_q); in emac_isr()
A Demac.h341 struct emac_tx_queue tx_q; member
/drivers/net/wwan/t7xx/
A Dt7xx_hif_dpmaif.c211 struct dpmaif_tx_queue *tx_q; in t7xx_dpmaif_rxtx_sw_allocs() local
236 tx_q = &dpmaif_ctrl->txq[tx_idx]; in t7xx_dpmaif_rxtx_sw_allocs()
237 tx_q->index = tx_idx; in t7xx_dpmaif_rxtx_sw_allocs()
238 tx_q->dpmaif_ctrl = dpmaif_ctrl; in t7xx_dpmaif_rxtx_sw_allocs()
239 ret = t7xx_dpmaif_txq_init(tx_q); in t7xx_dpmaif_rxtx_sw_allocs()
261 tx_q = &dpmaif_ctrl->txq[i]; in t7xx_dpmaif_rxtx_sw_allocs()
262 t7xx_dpmaif_txq_free(tx_q); in t7xx_dpmaif_rxtx_sw_allocs()
282 struct dpmaif_tx_queue *tx_q; in t7xx_dpmaif_sw_release() local
289 tx_q = &dpmaif_ctrl->txq[i]; in t7xx_dpmaif_sw_release()
290 t7xx_dpmaif_txq_free(tx_q); in t7xx_dpmaif_sw_release()
/drivers/net/ethernet/intel/ice/
A Dice_dcb_lib.h47 static inline bool ice_find_q_in_range(u16 low, u16 high, unsigned int tx_q) in ice_find_q_in_range() argument
49 return (tx_q >= low) && (tx_q < high); in ice_find_q_in_range()
/drivers/net/wireless/mediatek/mt7601u/
A Ddma.c312 struct mt7601u_tx_queue *q = &dev->tx_q[ep]; in mt7601u_dma_submit_tx()
477 if (!dev->tx_q) in mt7601u_free_tx()
481 mt7601u_free_tx_queue(&dev->tx_q[i]); in mt7601u_free_tx()
505 dev->tx_q = devm_kcalloc(dev->dev, __MT_EP_OUT_MAX, in mt7601u_alloc_tx()
506 sizeof(*dev->tx_q), GFP_KERNEL); in mt7601u_alloc_tx()
507 if (!dev->tx_q) in mt7601u_alloc_tx()
511 if (mt7601u_alloc_tx_queue(dev, &dev->tx_q[i])) in mt7601u_alloc_tx()
A Dmt7601u.h201 struct mt7601u_tx_queue *tx_q; member
/drivers/net/wireless/mediatek/mt76/
A Dmt76_connac_mac.c92 if (wcid && pm->tx_q[i].wcid != wcid) in mt76_connac_free_pending_tx_skbs()
95 dev_kfree_skb(pm->tx_q[i].skb); in mt76_connac_free_pending_tx_skbs()
96 pm->tx_q[i].skb = NULL; in mt76_connac_free_pending_tx_skbs()
111 if (!pm->tx_q[qid].skb) { in mt76_connac_pm_queue_skb()
113 pm->tx_q[qid].wcid = wcid; in mt76_connac_pm_queue_skb()
114 pm->tx_q[qid].skb = skb; in mt76_connac_pm_queue_skb()
130 struct mt76_wcid *wcid = pm->tx_q[i].wcid; in mt76_connac_pm_dequeue_skbs()
133 if (!pm->tx_q[i].skb) in mt76_connac_pm_dequeue_skbs()
140 mt76_tx(phy, sta, wcid, pm->tx_q[i].skb); in mt76_connac_pm_dequeue_skbs()
141 pm->tx_q[i].skb = NULL; in mt76_connac_pm_dequeue_skbs()
A Dmt76_connac.h98 } tx_q[IEEE80211_NUM_ACS]; member
/drivers/net/fddi/skfp/h/
A Dhwmtm.h242 #define HWM_GET_TX_USED(smc,queue) (int) (smc)->hw.fp.tx_q[queue].tx_used
261 (smc)->hw.fp.tx_q[queue].tx_curr_put
A Dfplustm.h191 struct s_smt_tx_queue tx_q[USED_QUEUES] ; member
/drivers/net/ethernet/chelsio/cxgb3/
A Dadapter.h208 struct netdev_queue *tx_q; /* associated netdev TX queue */ member
A Dsge.c1992 netif_tx_wake_queue(qs->tx_q); in restart_tx()
2915 if (__netif_tx_trylock(qs->tx_q)) { in sge_timer_tx()
2918 __netif_tx_unlock(qs->tx_q); in sge_timer_tx()
3156 q->tx_q = netdevq; in t3_sge_alloc_qset()
/drivers/net/fddi/skfp/
A Dfplustm.c305 smc->hw.fp.tx[QUEUE_S] = queue = &smc->hw.fp.tx_q[QUEUE_S] ; in init_tx()
316 smc->hw.fp.tx[QUEUE_A0] = queue = &smc->hw.fp.tx_q[QUEUE_A0] ; in init_tx()
A Dhwmtm.c561 if (smc->hw.fp.tx_q[QUEUE_A0].tx_used) { in mac_drv_repair_descr()
566 if (smc->hw.fp.tx_q[QUEUE_S].tx_used) { in mac_drv_repair_descr()

Completed in 879 milliseconds