Home
last modified time | relevance | path

Searched refs:txq (Results 1 – 25 of 350) sorted by relevance

12345678910>>...14

/linux/drivers/net/wwan/t7xx/
A Dt7xx_hif_dpmaif_tx.c54 struct dpmaif_tx_queue *txq = &dpmaif_ctrl->txq[q_num]; in t7xx_dpmaif_update_drb_rd_idx() local
83 struct dpmaif_tx_queue *txq = &dpmaif_ctrl->txq[q_num]; in t7xx_dpmaif_release_tx_buffer() local
177 queue_work(dpmaif_ctrl->txq[txq->index].worker, in t7xx_dpmaif_tx_done()
178 &dpmaif_ctrl->txq[txq->index].dpmaif_tx_work); in t7xx_dpmaif_tx_done()
253 txq = &dpmaif_ctrl->txq[skb_cb->txq_number]; in t7xx_dpmaif_add_skb_to_ring()
273 if (txq->drb_wr_idx >= txq->drb_size_cnt) in t7xx_dpmaif_add_skb_to_ring()
274 txq->drb_wr_idx -= txq->drb_size_cnt; in t7xx_dpmaif_add_skb_to_ring()
351 txq = &dpmaif_ctrl->txq[DPMAIF_TX_DEFAULT_QUEUE]; in t7xx_select_tx_queue()
556 txq->drb_base, txq->drb_bus_addr); in t7xx_dpmaif_tx_drb_buf_init()
592 txq->drb_base, txq->drb_bus_addr); in t7xx_dpmaif_tx_drb_buf_rel()
[all …]
/linux/drivers/infiniband/hw/hfi1/
A Dipoib_tx.c57 netif_stop_subqueue(txq->priv->netdev, txq->q_idx); in hfi1_ipoib_stop_txq()
64 netif_wake_subqueue(txq->priv->netdev, txq->q_idx); in hfi1_ipoib_wake_txq()
82 if (hfi1_ipoib_used(txq) >= hfi1_ipoib_ring_hwat(txq) && in hfi1_ipoib_check_queue_depth()
353 struct hfi1_ipoib_txq *txq = txp->txq; in hfi1_ipoib_send_dma_common() local
378 tx->txq = txq; in hfi1_ipoib_send_dma_common()
450 iowait_starve_clear(txq->pkts_sent, &txq->wait); in hfi1_ipoib_submit_tx()
460 struct hfi1_ipoib_txq *txq = txp->txq; in hfi1_ipoib_send_dma_single() local
511 struct hfi1_ipoib_txq *txq = txp->txq; in hfi1_ipoib_send_dma_list() local
739 txq->tx_ring.avail = hfi1_ipoib_ring_hwat(txq); in hfi1_ipoib_txreq_init()
854 txq->sde ? txq->sde->this_idx : 0); in hfi1_ipoib_tx_timeout()
[all …]
A Dtrace_tx.h914 __entry->txq = txq;
949 TP_ARGS(txq)
955 TP_ARGS(txq)
961 TP_ARGS(txq)
967 TP_ARGS(txq)
973 TP_ARGS(txq)
979 TP_ARGS(txq)
985 TP_ARGS(txq)
991 TP_ARGS(txq)
1015 __entry->txq = tx->txq;
[all …]
/linux/drivers/net/wireless/intel/iwlwifi/pcie/
A Dtx.c400 &txq->entries[txq->read_ptr].meta; in iwl_pcie_txq_unmap()
460 txq->tfds, txq->dma_addr); in iwl_pcie_txq_free()
475 memset(txq, 0, sizeof(*txq)); in iwl_pcie_txq_free()
683 txq->read_ptr, txq->write_ptr); in iwl_txq_log_scd_error()
696 txq->read_ptr, txq->write_ptr, in iwl_txq_log_scd_error()
1094 (!iwl_txq_used(txq, idx, txq->read_ptr, txq->write_ptr))) { in iwl_pcie_cmdq_reclaim()
1099 txq->write_ptr, txq->read_ptr); in iwl_pcie_cmdq_reclaim()
1593 if (txq->read_ptr == txq->write_ptr && txq->wd_timeout) in iwl_pcie_enqueue_hcmd()
2272 if (txq->read_ptr == txq->write_ptr && txq->wd_timeout) { in iwl_trans_pcie_tx()
2479 txq->read_ptr = txq->write_ptr; in iwl_pcie_set_q_ptrs()
[all …]
A Dtx-gen2.c749 if (iwl_txq_space(trans, txq) < txq->high_mark) { in iwl_txq_gen2_tx()
766 idx = iwl_txq_get_cmd_index(txq, txq->write_ptr); in iwl_txq_gen2_tx()
803 if (txq->read_ptr == txq->write_ptr && txq->wd_timeout) in iwl_txq_gen2_tx()
829 while (txq->write_ptr != txq->read_ptr) { in iwl_txq_gen2_unmap()
868 txq->tfds, txq->dma_addr); in iwl_txq_gen2_free_memory()
871 txq->first_tb_bufs, txq->first_tb_dma); in iwl_txq_gen2_free_memory()
877 txq->bc_tbl.addr, txq->bc_tbl.dma); in iwl_txq_gen2_free_memory()
899 txq = trans_pcie->txqs.txq[txq_id]; in iwl_txq_gen2_free()
937 txq = kzalloc(sizeof(*txq), GFP_KERNEL); in iwl_txq_dyn_alloc_dma()
1007 trans_pcie->txqs.txq[qid] = txq; in iwl_pcie_txq_alloc_response()
[all …]
/linux/drivers/net/ethernet/huawei/hinic/
A Dhinic_tx.c533 hinic_sq_prepare_wqe(txq->sq, sq_wqe, txq->sges, nr_sges); in hinic_lb_xmit_frame()
617 hinic_sq_prepare_wqe(txq->sq, sq_wqe, txq->sges, nr_sges); in hinic_xmit_frame()
810 netif_napi_add_weight(txq->netdev, &txq->napi, free_tx_poll, in tx_request_irq()
832 err = request_irq(sq->irq, tx_irq, 0, txq->irq_name, txq); in tx_request_irq()
866 txq->netdev = netdev; in hinic_init_txq()
867 txq->sq = sq; in hinic_init_txq()
869 txq_stats_init(txq); in hinic_init_txq()
873 txq->sges = devm_kcalloc(&netdev->dev, txq->max_sges, in hinic_init_txq()
875 if (!txq->sges) in hinic_init_txq()
878 txq->free_sges = devm_kcalloc(&netdev->dev, txq->max_sges, in hinic_init_txq()
[all …]
/linux/include/net/
A Dnetdev_queues.h157 netif_tx_stop_queue(txq); \
169 netif_tx_start_queue(txq); \
247 netdev_txq_completed_mb(txq, pkts, bytes); \
254 netif_tx_wake_queue(txq); \
268 struct netdev_queue *txq; \
270 txq = netdev_get_tx_queue(dev, idx); \
276 struct netdev_queue *txq; \
278 txq = netdev_get_tx_queue(dev, idx); \
285 struct netdev_queue *txq; \
287 txq = netdev_get_tx_queue(dev, idx); \
[all …]
/linux/drivers/net/wireless/ath/ath9k/
A Dxmit.c135 struct ath_txq *txq = tid->txq; in ath9k_wake_tx_queue() local
240 .txq = tid->txq, in ath_tid_pull()
281 struct ath_txq *txq = tid->txq; in ath_tx_flush_tid() local
1588 txq = txtid->txq; in ath_tx_aggr_start()
1620 struct ath_txq *txq = txtid->txq; in ath_tx_aggr_stop() local
1659 txq = tid->txq; in ath_tx_aggr_wakeup()
1968 txq = &sc->tx.txq[i]; in ath_drain_all_txq()
2067 txq->axq_qnum, txq->axq_depth); in ath_tx_txqaddbuf()
2079 txq->axq_qnum, txq->axq_link, in ath_tx_txqaddbuf()
2762 txq = &sc->tx.txq[ts.qid]; in ath_tx_edma_tasklet()
[all …]
/linux/drivers/net/ethernet/marvell/
A Dmv643xx_eth.c448 return container_of(txq, struct mv643xx_eth_private, txq[txq->index]); in txq_to_mp()
736 if (txq->tx_curr_desc == txq->tx_ring_size) in txq_put_data_tso()
814 if (txq->tx_curr_desc == txq->tx_ring_size) in txq_put_hdr_tso()
955 if (txq->tx_curr_desc == txq->tx_ring_size) in txq_submit_skb()
1001 txq = mp->txq + queue; in mv643xx_eth_xmit()
1262 struct tx_queue *txq = mp->txq + i; in mv643xx_eth_get_stats() local
2036 struct tx_queue *txq = mp->txq + index; in txq_init() local
2127 txq_reclaim(txq, txq->tx_ring_size, 1); in txq_deinit()
2142 txq->tso_hdrs, txq->tso_hdrs_dma); in txq_deinit()
2212 struct tx_queue *txq = mp->txq + i; in handle_link_event() local
[all …]
A Dmvneta.c135 #define MVNETA_CPU_TXQ_ACCESS(txq) BIT(txq + 8) argument
777 if (txq->txq_get_index == txq->size) in mvneta_txq_inc_get()
785 if (txq->txq_put_index == txq->size) in mvneta_txq_inc_put()
996 txq->next_desc_to_proc = txq->last_desc - 1; in mvneta_txq_desc_put()
1504 for (txq = 0; txq < txq_number; txq++) in mvneta_defaults_set()
1928 if (txq->count <= txq->tx_wake_threshold) in mvneta_txq_done()
2103 if (txq->count + num_frames >= txq->size) in mvneta_xdp_submit_frame()
2982 if (txq->count >= txq->tx_stop_threshold) in mvneta_tx()
3546 txq->last_desc = txq->size - 1; in mvneta_txq_sw_init()
3548 txq->buf = kmalloc_array(txq->size, sizeof(*txq->buf), GFP_KERNEL); in mvneta_txq_sw_init()
[all …]
/linux/drivers/net/ethernet/mellanox/mlx5/core/en/
A Dselq.h30 static inline u16 mlx5e_txq_to_ch_ix(u16 txq, u16 num_channels) in mlx5e_txq_to_ch_ix() argument
32 while (unlikely(txq >= num_channels)) in mlx5e_txq_to_ch_ix()
33 txq -= num_channels; in mlx5e_txq_to_ch_ix()
34 return txq; in mlx5e_txq_to_ch_ix()
37 static inline u16 mlx5e_txq_to_ch_ix_htb(u16 txq, u16 num_channels) in mlx5e_txq_to_ch_ix_htb() argument
39 if (unlikely(txq >= num_channels)) { in mlx5e_txq_to_ch_ix_htb()
40 if (unlikely(txq >= num_channels << 3)) in mlx5e_txq_to_ch_ix_htb()
41 txq %= num_channels; in mlx5e_txq_to_ch_ix_htb()
44 txq -= num_channels; in mlx5e_txq_to_ch_ix_htb()
45 while (txq >= num_channels); in mlx5e_txq_to_ch_ix_htb()
[all …]
/linux/drivers/net/ethernet/qlogic/qede/
A Dqede_fp.c90 idx, txq->sw_tx_cons, txq->sw_tx_prod); in qede_free_tx_pkt()
295 writel(txq->tx_db.raw, txq->doorbell_addr); in qede_update_tx_producer()
313 txq->stopped_cnt++; in qede_xdp_xmit()
329 xdp = txq->sw_tx_ring.xdp + txq->sw_tx_prod; in qede_xdp_xmit()
334 txq->sw_tx_prod = (txq->sw_tx_prod + 1) % txq->num_tx_buffers; in qede_xdp_xmit()
427 txq->sw_tx_cons = (txq->sw_tx_cons + 1) % txq->num_tx_buffers; in qede_xdp_tx_int()
428 txq->xmit_pkts++; in qede_xdp_tx_int()
457 txq->sw_tx_cons = (txq->sw_tx_cons + 1) % txq->num_tx_buffers; in qede_tx_int()
458 txq->xmit_pkts++; in qede_tx_int()
1709 txq->sw_tx_prod = (txq->sw_tx_prod + 1) % txq->num_tx_buffers; in qede_start_xmit()
[all …]
A Dqede_main.c551 fp->id, fp->sb_info->igu_sb_id, txq->index, txq->ndev_txq_id, txq->cos, in qede_txq_fp_log_metadata()
557 le16_to_cpu(*txq->hw_cons_ptr), txq->sw_tx_prod, txq->sw_tx_cons, in qede_txq_fp_log_metadata()
579 txq->index, le16_to_cpu(*txq->hw_cons_ptr), in qede_tx_log_print()
608 txq = &fp->txq[cos]; in qede_tx_timeout()
1030 if (!fp->txq) in qede_alloc_fp_array()
1677 size = sizeof(*txq->sw_tx_ring.xdp) * txq->num_tx_buffers; in qede_alloc_mem_txq()
1800 txq->index, qed_chain_get_cons_idx(&txq->tx_pbl), in qede_empty_tx_queue()
1887 struct qede_tx_queue *txq = &fp->txq[cos]; in qede_init_fp() local
2062 while (txq->sw_tx_cons != txq->sw_tx_prod) { in qede_drain_txq()
2075 txq->index, txq->sw_tx_prod, in qede_drain_txq()
[all …]
/linux/drivers/net/ethernet/atheros/alx/
A Dmain.c159 if (txq->write_idx >= txq->read_idx) in alx_tpd_avail()
160 return txq->count + txq->read_idx - txq->write_idx - 1; in alx_tpd_avail()
200 alx_tpd_avail(txq) > txq->count / 4) in alx_clean_tx_irq()
765 np->txq = txq; in alx_alloc_napis()
1470 if (++txq->write_idx == txq->count) in alx_map_tx_skb()
1473 tpd = &txq->tpd[txq->write_idx]; in alx_map_tx_skb()
1494 if (++txq->write_idx == txq->count) in alx_map_tx_skb()
1496 tpd = &txq->tpd[txq->write_idx]; in alx_map_tx_skb()
1514 txq->bufs[txq->write_idx].skb = skb; in alx_map_tx_skb()
1516 if (++txq->write_idx == txq->count) in alx_map_tx_skb()
[all …]
/linux/include/trace/events/
A Dqdisc.h16 TP_PROTO(struct Qdisc *qdisc, const struct netdev_queue *txq,
19 TP_ARGS(qdisc, txq, packets, skb),
23 __field(const struct netdev_queue *, txq )
35 __entry->txq = txq;
38 __entry->ifindex = txq->dev ? txq->dev->ifindex : 0;
41 __entry->txq_state = txq->state;
51 TP_PROTO(struct Qdisc *qdisc, const struct netdev_queue *txq, struct sk_buff *skb),
53 TP_ARGS(qdisc, txq, skb),
57 __field(const struct netdev_queue *, txq)
66 __entry->txq = txq;
[all …]
/linux/drivers/net/ethernet/freescale/
A Dfec_main.c309 (addr < txq->tso_hdrs_dma + txq->bd.ring_size * TSO_HEADER_SIZE))
338 (const char *)txq->bd.cur) >> txq->bd.dsize_log2) - 1; in fec_enet_get_free_txdesc_num()
363 bdp = txq->bd.base; in fec_dump()
588 bdp = txq->bd.cur; in fec_enet_txq_submit_frag_skb()
626 bdp = txq->bd.cur; in fec_enet_txq_submit_skb()
714 txq->bd.cur = bdp; in fec_enet_txq_submit_skb()
903 txq->bd.cur = bdp; in fec_enet_txq_submit_tso()
3334 txq->tso_hdrs, txq->tso_hdrs_dma); in fec_enet_free_queue()
3351 txq = kzalloc(sizeof(*txq), GFP_KERNEL); in fec_enet_alloc_queue()
3352 if (!txq) { in fec_enet_alloc_queue()
[all …]
/linux/drivers/bluetooth/
A Dbtintel_pcie.c109 static void btintel_pcie_prepare_tx(struct txq *txq, u16 tfd_index, in btintel_pcie_prepare_tx() argument
132 struct txq *txq = &data->txq; in btintel_pcie_send_sync() local
341 struct txq *txq; in btintel_pcie_msix_tx_handle() local
350 txq = &data->txq; in btintel_pcie_msix_tx_handle()
783 struct txq *txq) in btintel_pcie_free_txq_bufs() argument
787 txq->buf_v_addr, txq->buf_p_addr); in btintel_pcie_free_txq_bufs()
788 kfree(txq->bufs); in btintel_pcie_free_txq_bufs()
792 struct txq *txq) in btintel_pcie_setup_txq_bufs() argument
798 txq->bufs = kmalloc_array(txq->count, sizeof(*buf), GFP_KERNEL); in btintel_pcie_setup_txq_bufs()
799 if (!txq->bufs) in btintel_pcie_setup_txq_bufs()
[all …]
/linux/drivers/net/ethernet/chelsio/cxgb4vf/
A Dsge.c1135 netif_tx_stop_queue(txq->txq); in txq_stop()
1263 wr = (void *)&txq->q.desc[txq->q.pidx]; in t4vf_eth_xmit()
1319 txq->tso++; in t4vf_eth_xmit()
2133 if (reclaimable(&txq->q) && __netif_tx_trylock(txq->txq)) { in sge_tx_timer_cb()
2141 __netif_tx_unlock(txq->txq); in sge_tx_timer_cb()
2430 txq->q.desc = alloc_ring(adapter->pdev_dev, txq->q.size, in t4vf_sge_alloc_eth_txq()
2433 &txq->q.phys_addr, &txq->q.sdesc, s->stat_len); in t4vf_sge_alloc_eth_txq()
2482 txq->q.desc, txq->q.phys_addr); in t4vf_sge_alloc_eth_txq()
2490 txq->q.stat = (void *)&txq->q.desc[txq->q.size]; in t4vf_sge_alloc_eth_txq()
2498 txq->txq = devq; in t4vf_sge_alloc_eth_txq()
[all …]
/linux/drivers/net/wireless/mediatek/mt76/
A Dtx.c11 if (!txq->sta) in mt76_txq_get_qid()
14 return txq->ac; in mt76_txq_get_qid()
21 struct ieee80211_txq *txq; in mt76_tx_check_agg_ssn() local
30 txq = sta->txq[tid]; in mt76_tx_check_agg_ssn()
418 struct ieee80211_txq *txq = sta->txq[i]; in mt76_release_buffered_frames() local
481 ieee80211_get_tx_rates(txq->vif, txq->sta, skb, in mt76_txq_send_burst()
503 ieee80211_get_tx_rates(txq->vif, txq->sta, skb, in mt76_txq_send_burst()
544 if (!txq) in mt76_txq_schedule_list()
557 u8 tid = txq->tid; in mt76_txq_schedule_list()
729 struct ieee80211_txq *txq = sta->txq[i]; in mt76_stop_tx_queues() local
[all …]
/linux/drivers/net/ethernet/fungible/funeth/
A Dfuneth_trace.h15 TP_PROTO(const struct funeth_txq *txq,
20 TP_ARGS(txq, len, sqe_idx, ngle),
27 __string(devname, txq->netdev->name)
31 __entry->qidx = txq->qidx;
45 TP_PROTO(const struct funeth_txq *txq,
50 TP_ARGS(txq, sqe_idx, num_sqes, hw_head),
57 __string(devname, txq->netdev->name)
61 __entry->qidx = txq->qidx;
/linux/drivers/net/ethernet/hisilicon/
A Dhisi_femac.c155 struct hisi_femac_queue *txq = &priv->txq; in hisi_femac_xmit_reclaim() local
163 skb = txq->skb[txq->tail]; in hisi_femac_xmit_reclaim()
177 txq->skb[txq->tail] = NULL; in hisi_femac_xmit_reclaim()
178 txq->tail = (txq->tail + 1) % txq->num; in hisi_femac_xmit_reclaim()
387 struct hisi_femac_queue *txq = &priv->txq; in hisi_femac_free_skb_rings() local
412 pos = txq->tail; in hisi_femac_free_skb_rings()
503 struct hisi_femac_queue *txq = &priv->txq; in hisi_femac_net_xmit() local
517 if (unlikely(!CIRC_SPACE(txq->head, txq->tail, in hisi_femac_net_xmit()
533 txq->dma_phys[txq->head] = addr; in hisi_femac_net_xmit()
535 txq->skb[txq->head] = skb; in hisi_femac_net_xmit()
[all …]
/linux/drivers/net/ethernet/intel/idpf/
A Didpf_txrx.h116 #define IDPF_DESC_UNUSED(txq) \ argument
117 ((((txq)->next_to_clean > (txq)->next_to_use) ? 0 : (txq)->desc_count) + \
118 (txq)->next_to_clean - (txq)->next_to_use - 1)
120 #define IDPF_TX_BUF_RSV_UNUSED(txq) ((txq)->stash->buf_stack.top) argument
121 #define IDPF_TX_BUF_RSV_LOW(txq) (IDPF_TX_BUF_RSV_UNUSED(txq) < \ argument
122 (txq)->desc_count >> 2)
128 #define IDPF_TX_COMPLQ_PENDING(txq) \ argument
129 (((txq)->num_completions_pending >= (txq)->complq->num_completions ? \
131 (txq)->num_completions_pending - (txq)->complq->num_completions)
136 ((++(txq)->compl_tag_cur_gen) >= (txq)->compl_tag_gen_max ? \
[all …]
/linux/net/sched/
A Dsch_generic.c87 txq = skb_get_tx_queue(txq->dev, skb); in __skb_dequeue_bad_txq()
99 qdisc_maybe_clear_missed(q, txq); in __skb_dequeue_bad_txq()
260 txq = skb_get_tx_queue(txq->dev, skb); in dequeue_skb()
347 HARD_TX_UNLOCK(dev, txq); in sch_direct_xmit()
392 struct netdev_queue *txq; in qdisc_restart() local
460 __netif_tx_lock(txq, cpu); in netif_freeze_queues()
462 __netif_tx_unlock(txq); in netif_freeze_queues()
485 netif_schedule_queue(txq); in netif_unfreeze_queues()
512 struct netdev_queue *txq; in dev_watchdog() local
515 if (!netif_xmit_stopped(txq)) in dev_watchdog()
[all …]
/linux/drivers/net/wireless/ath/ath5k/
A Dbase.c877 txq->txq_len++; in ath5k_txbuf_setup()
1056 if (!txq->setup) { in ath5k_txq_setup()
1057 txq->qnum = qnum; in ath5k_txq_setup()
1058 txq->link = NULL; in ath5k_txq_setup()
1062 txq->txq_len = 0; in ath5k_txq_setup()
1173 txq->txq_len--; in ath5k_drain_tx_buffs()
1190 if (txq->setup) { in ath5k_txq_release()
1676 if (txq->txq_len >= txq->txq_max && in ath5k_tx_queue()
1822 txq->txq_len--; in ath5k_tx_processq()
1827 if (txq->txq_len < ATH5K_TXQ_LEN_LOW && txq->qnum < 4) in ath5k_tx_processq()
[all …]
/linux/drivers/net/ethernet/chelsio/cxgb3/
A Dsge.c691 if (q->txq[i].desc) { in t3_free_qset()
701 q->txq[i].size * in t3_free_qset()
703 q->txq[i].desc, q->txq[i].phys_addr); in t3_free_qset()
1282 q = &qs->txq[TXQ_ETH]; in t3_eth_xmit()
2256 struct sge_txq *txq = &qs->txq[TXQ_ETH]; in check_ring_db() local
2258 if (txq->cleaned + txq->in_use != txq->processed && in check_ring_db()
2267 struct sge_txq *txq = &qs->txq[TXQ_OFLD]; in check_ring_db() local
2269 if (txq->cleaned + txq->in_use != txq->processed && in check_ring_db()
3064 if (!q->txq[i].desc) in t3_sge_alloc_qset()
3067 q->txq[i].gen = 1; in t3_sge_alloc_qset()
[all …]

Completed in 151 milliseconds

12345678910>>...14