Home
last modified time | relevance | path

Searched refs:txq_idx (Results 1 – 19 of 19) sorted by relevance

/drivers/net/ethernet/microsoft/mana/
A Dmana_bpf.c15 u16 txq_idx = skb_get_queue_mapping(skb); in mana_xdp_tx() local
21 ndevtxq = netdev_get_tx_queue(ndev, txq_idx); in mana_xdp_tx()
A Dmana_en.c257 u16 txq_idx = skb_get_queue_mapping(skb); in mana_start_xmit() local
278 txq = &apc->tx_qp[txq_idx].txq; in mana_start_xmit()
280 cq = &apc->tx_qp[txq_idx].tx_cq; in mana_start_xmit()
421 net_txq = netdev_get_tx_queue(ndev, txq_idx); in mana_start_xmit()
/drivers/net/ethernet/meta/fbnic/
A Dfbnic_txrx.c1357 unsigned int txq_count, unsigned int txq_idx, in fbnic_alloc_napi_vector() argument
1416 db = &uc_addr[FBNIC_QUEUE(txq_idx) + FBNIC_QUEUE_TWQ0_TAIL]; in fbnic_alloc_napi_vector()
1422 fbnic_ring_init(&qt->sub0, db, txq_idx, flags); in fbnic_alloc_napi_vector()
1423 fbn->tx[txq_idx] = &qt->sub0; in fbnic_alloc_napi_vector()
1431 db = &uc_addr[FBNIC_QUEUE(txq_idx) + FBNIC_QUEUE_TCQ_HEAD]; in fbnic_alloc_napi_vector()
1436 txq_idx += v_count; in fbnic_alloc_napi_vector()
1489 1, txq_idx, 0, 0); in fbnic_alloc_napi_vectors()
1495 txq_idx++; in fbnic_alloc_napi_vectors()
1503 int tqpv = DIV_ROUND_UP(num_tx, num_napi - txq_idx); in fbnic_alloc_napi_vectors()
1507 tqpv, txq_idx, rqpv, rxq_idx); in fbnic_alloc_napi_vectors()
[all …]
/drivers/target/iscsi/cxgbit/
A Dcxgbit_cm.c631 cxgb_mk_close_con_req(skb, len, csk->tid, csk->txq_idx, in cxgbit_send_halfclose()
672 cxgb_mk_abort_req(skb, len, csk->tid, csk->txq_idx, in cxgbit_send_abort_req()
943 csk->txq_idx = cxgb4_port_idx(ndev) * step; in cxgbit_offload_init()
979 csk->txq_idx = (port_id * step) + in cxgbit_offload_init()
1483 set_wr_txq(skb, CPL_PRIORITY_DATA, csk->txq_idx); in cxgbit_send_tx_flowc_wr()
1771 cxgb_mk_abort_rpl(rpl_skb, len, csk->tid, csk->txq_idx); in cxgbit_abort_req_rss()
A Dcxgbit.h231 u16 txq_idx; member
A Dcxgbit_target.c244 set_wr_txq(skb, CPL_PRIORITY_DATA, csk->txq_idx); in cxgbit_push_tx_frames()
/drivers/net/ethernet/cavium/thunder/
A Dnicvf_main.c858 unsigned int tx_pkts = 0, tx_bytes = 0, txq_idx; in nicvf_cq_intr_handler() local
921 txq_idx = nicvf_netdev_qidx(nic, cq_idx); in nicvf_cq_intr_handler()
924 if (txq_idx < nic->pnicvf->xdp_tx_queues) { in nicvf_cq_intr_handler()
929 txq_idx -= nic->pnicvf->xdp_tx_queues; in nicvf_cq_intr_handler()
936 txq = netdev_get_tx_queue(netdev, txq_idx); in nicvf_cq_intr_handler()
947 "Transmit queue wakeup SQ%d\n", txq_idx); in nicvf_cq_intr_handler()
/drivers/infiniband/hw/cxgb4/
A Dcm.c642 set_wr_txq(skb, CPL_PRIORITY_DATA, ep->txq_idx); in send_flowc()
655 cxgb_mk_close_con_req(skb, wrlen, ep->hwtid, ep->txq_idx, in send_halfclose()
696 cxgb_mk_abort_req(req_skb, wrlen, ep->hwtid, ep->txq_idx, in send_abort_req()
960 set_wr_txq(skb, CPL_PRIORITY_DATA, ep->txq_idx); in send_mpa_req()
1066 set_wr_txq(skb, CPL_PRIORITY_DATA, ep->txq_idx); in send_mpa_reject()
1118 set_wr_txq(skb, CPL_PRIORITY_DATA, ep->txq_idx); in send_mpa_reject()
1146 set_wr_txq(skb, CPL_PRIORITY_DATA, ep->txq_idx); in send_mpa_reply()
2115 ep->txq_idx = cxgb4_port_idx(pdev) * step; in import_ep()
2133 ep->txq_idx = cxgb4_port_idx(pdev) * step; in import_ep()
2236 ep->txq_idx, ep->tx_chan, ep->smac_idx, ep->rss_qid, in c4iw_reconnect()
[all …]
A Diw_cxgb4.h927 u16 txq_idx; member
A Dqp.c1574 set_wr_txq(skb, CPL_PRIORITY_DATA, qhp->ep->txq_idx); in post_terminate()
1713 set_wr_txq(skb, CPL_PRIORITY_DATA, ep->txq_idx); in rdma_fini()
1777 set_wr_txq(skb, CPL_PRIORITY_DATA, qhp->ep->txq_idx); in rdma_init()
/drivers/net/ethernet/chelsio/inline_crypto/chtls/
A Dchtls_cm.c237 skb_set_queue_mapping(skb, (csk->txq_idx << 1) | CPL_PRIORITY_DATA); in chtls_send_abort()
1233 csk->txq_idx = (rxq_idx < cdev->lldi->ntxq) ? rxq_idx : in chtls_recv_sock()
2061 queue = csk->txq_idx; in bl_abort_syn_rcv()
2094 int queue = csk->txq_idx; in abort_syn_rcv()
2112 int queue = csk->txq_idx; in chtls_abort_req_rss()
A Dchtls.h297 u32 txq_idx; member
A Dchtls_io.c101 skb_set_queue_mapping(skb, (csk->txq_idx << 1) | CPL_PRIORITY_DATA); in create_flowc_wr_skb()
128 csk->txq_idx, in send_flowc_wr()
661 skb_set_queue_mapping(skb, (csk->txq_idx << 1) | in chtls_push_frames()
A Dchtls_hw.c81 skb_set_queue_mapping(skb, (csk->txq_idx << 1) | CPL_PRIORITY_DATA); in chtls_set_tcb_field()
/drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/
A Dtrans.c2511 int iwl_trans_pcie_wait_txq_empty(struct iwl_trans *trans, int txq_idx) in iwl_trans_pcie_wait_txq_empty() argument
2523 if (!test_bit(txq_idx, trans_pcie->txqs.queue_used)) in iwl_trans_pcie_wait_txq_empty()
2526 IWL_DEBUG_TX_QUEUES(trans, "Emptying queue %d...\n", txq_idx); in iwl_trans_pcie_wait_txq_empty()
2527 txq = trans_pcie->txqs.txq[txq_idx]; in iwl_trans_pcie_wait_txq_empty()
2563 "fail to flush all tx fifo queues Q %d\n", txq_idx); in iwl_trans_pcie_wait_txq_empty()
2568 IWL_DEBUG_TX_QUEUES(trans, "Queue %d is now empty.\n", txq_idx); in iwl_trans_pcie_wait_txq_empty()
A Dinternal.h791 int iwl_trans_pcie_wait_txq_empty(struct iwl_trans *trans, int txq_idx);
/drivers/scsi/cxgbi/
A Dlibcxgbi.h120 unsigned short txq_idx; member
/drivers/scsi/cxgbi/cxgb4i/
A Dcxgb4i.c1864 csk->txq_idx = cxgb4_port_idx(ndev) * step; in init_act_open()
1893 csk->flags, csk->tx_chan, csk->txq_idx, csk->rss_qid, in init_act_open()
/drivers/net/ethernet/broadcom/
A Dtg3.c7399 int txq_idx = tp->txq_cnt - 1; in tg3_napi_disable() local
7407 netif_queue_set_napi(tp->dev, txq_idx, in tg3_napi_disable()
7409 txq_idx--; in tg3_napi_disable()
7422 int txq_idx = 0, rxq_idx = 0; in tg3_napi_enable() local
7430 netif_queue_set_napi(tp->dev, txq_idx, in tg3_napi_enable()
7433 txq_idx++; in tg3_napi_enable()

Completed in 117 milliseconds