| /drivers/net/ethernet/microsoft/mana/ |
| A D | mana_bpf.c | 15 u16 txq_idx = skb_get_queue_mapping(skb); in mana_xdp_tx() local 21 ndevtxq = netdev_get_tx_queue(ndev, txq_idx); in mana_xdp_tx()
|
| A D | mana_en.c | 257 u16 txq_idx = skb_get_queue_mapping(skb); in mana_start_xmit() local 278 txq = &apc->tx_qp[txq_idx].txq; in mana_start_xmit() 280 cq = &apc->tx_qp[txq_idx].tx_cq; in mana_start_xmit() 421 net_txq = netdev_get_tx_queue(ndev, txq_idx); in mana_start_xmit()
|
| /drivers/net/ethernet/meta/fbnic/ |
| A D | fbnic_txrx.c | 1357 unsigned int txq_count, unsigned int txq_idx, in fbnic_alloc_napi_vector() argument 1416 db = &uc_addr[FBNIC_QUEUE(txq_idx) + FBNIC_QUEUE_TWQ0_TAIL]; in fbnic_alloc_napi_vector() 1422 fbnic_ring_init(&qt->sub0, db, txq_idx, flags); in fbnic_alloc_napi_vector() 1423 fbn->tx[txq_idx] = &qt->sub0; in fbnic_alloc_napi_vector() 1431 db = &uc_addr[FBNIC_QUEUE(txq_idx) + FBNIC_QUEUE_TCQ_HEAD]; in fbnic_alloc_napi_vector() 1436 txq_idx += v_count; in fbnic_alloc_napi_vector() 1489 1, txq_idx, 0, 0); in fbnic_alloc_napi_vectors() 1495 txq_idx++; in fbnic_alloc_napi_vectors() 1503 int tqpv = DIV_ROUND_UP(num_tx, num_napi - txq_idx); in fbnic_alloc_napi_vectors() 1507 tqpv, txq_idx, rqpv, rxq_idx); in fbnic_alloc_napi_vectors() [all …]
|
| /drivers/target/iscsi/cxgbit/ |
| A D | cxgbit_cm.c | 631 cxgb_mk_close_con_req(skb, len, csk->tid, csk->txq_idx, in cxgbit_send_halfclose() 672 cxgb_mk_abort_req(skb, len, csk->tid, csk->txq_idx, in cxgbit_send_abort_req() 943 csk->txq_idx = cxgb4_port_idx(ndev) * step; in cxgbit_offload_init() 979 csk->txq_idx = (port_id * step) + in cxgbit_offload_init() 1483 set_wr_txq(skb, CPL_PRIORITY_DATA, csk->txq_idx); in cxgbit_send_tx_flowc_wr() 1771 cxgb_mk_abort_rpl(rpl_skb, len, csk->tid, csk->txq_idx); in cxgbit_abort_req_rss()
|
| A D | cxgbit.h | 231 u16 txq_idx; member
|
| A D | cxgbit_target.c | 244 set_wr_txq(skb, CPL_PRIORITY_DATA, csk->txq_idx); in cxgbit_push_tx_frames()
|
| /drivers/net/ethernet/cavium/thunder/ |
| A D | nicvf_main.c | 858 unsigned int tx_pkts = 0, tx_bytes = 0, txq_idx; in nicvf_cq_intr_handler() local 921 txq_idx = nicvf_netdev_qidx(nic, cq_idx); in nicvf_cq_intr_handler() 924 if (txq_idx < nic->pnicvf->xdp_tx_queues) { in nicvf_cq_intr_handler() 929 txq_idx -= nic->pnicvf->xdp_tx_queues; in nicvf_cq_intr_handler() 936 txq = netdev_get_tx_queue(netdev, txq_idx); in nicvf_cq_intr_handler() 947 "Transmit queue wakeup SQ%d\n", txq_idx); in nicvf_cq_intr_handler()
|
| /drivers/infiniband/hw/cxgb4/ |
| A D | cm.c | 642 set_wr_txq(skb, CPL_PRIORITY_DATA, ep->txq_idx); in send_flowc() 655 cxgb_mk_close_con_req(skb, wrlen, ep->hwtid, ep->txq_idx, in send_halfclose() 696 cxgb_mk_abort_req(req_skb, wrlen, ep->hwtid, ep->txq_idx, in send_abort_req() 960 set_wr_txq(skb, CPL_PRIORITY_DATA, ep->txq_idx); in send_mpa_req() 1066 set_wr_txq(skb, CPL_PRIORITY_DATA, ep->txq_idx); in send_mpa_reject() 1118 set_wr_txq(skb, CPL_PRIORITY_DATA, ep->txq_idx); in send_mpa_reject() 1146 set_wr_txq(skb, CPL_PRIORITY_DATA, ep->txq_idx); in send_mpa_reply() 2115 ep->txq_idx = cxgb4_port_idx(pdev) * step; in import_ep() 2133 ep->txq_idx = cxgb4_port_idx(pdev) * step; in import_ep() 2236 ep->txq_idx, ep->tx_chan, ep->smac_idx, ep->rss_qid, in c4iw_reconnect() [all …]
|
| A D | iw_cxgb4.h | 927 u16 txq_idx; member
|
| A D | qp.c | 1574 set_wr_txq(skb, CPL_PRIORITY_DATA, qhp->ep->txq_idx); in post_terminate() 1713 set_wr_txq(skb, CPL_PRIORITY_DATA, ep->txq_idx); in rdma_fini() 1777 set_wr_txq(skb, CPL_PRIORITY_DATA, qhp->ep->txq_idx); in rdma_init()
|
| /drivers/net/ethernet/chelsio/inline_crypto/chtls/ |
| A D | chtls_cm.c | 237 skb_set_queue_mapping(skb, (csk->txq_idx << 1) | CPL_PRIORITY_DATA); in chtls_send_abort() 1233 csk->txq_idx = (rxq_idx < cdev->lldi->ntxq) ? rxq_idx : in chtls_recv_sock() 2061 queue = csk->txq_idx; in bl_abort_syn_rcv() 2094 int queue = csk->txq_idx; in abort_syn_rcv() 2112 int queue = csk->txq_idx; in chtls_abort_req_rss()
|
| A D | chtls.h | 297 u32 txq_idx; member
|
| A D | chtls_io.c | 101 skb_set_queue_mapping(skb, (csk->txq_idx << 1) | CPL_PRIORITY_DATA); in create_flowc_wr_skb() 128 csk->txq_idx, in send_flowc_wr() 661 skb_set_queue_mapping(skb, (csk->txq_idx << 1) | in chtls_push_frames()
|
| A D | chtls_hw.c | 81 skb_set_queue_mapping(skb, (csk->txq_idx << 1) | CPL_PRIORITY_DATA); in chtls_set_tcb_field()
|
| /drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/ |
| A D | trans.c | 2511 int iwl_trans_pcie_wait_txq_empty(struct iwl_trans *trans, int txq_idx) in iwl_trans_pcie_wait_txq_empty() argument 2523 if (!test_bit(txq_idx, trans_pcie->txqs.queue_used)) in iwl_trans_pcie_wait_txq_empty() 2526 IWL_DEBUG_TX_QUEUES(trans, "Emptying queue %d...\n", txq_idx); in iwl_trans_pcie_wait_txq_empty() 2527 txq = trans_pcie->txqs.txq[txq_idx]; in iwl_trans_pcie_wait_txq_empty() 2563 "fail to flush all tx fifo queues Q %d\n", txq_idx); in iwl_trans_pcie_wait_txq_empty() 2568 IWL_DEBUG_TX_QUEUES(trans, "Queue %d is now empty.\n", txq_idx); in iwl_trans_pcie_wait_txq_empty()
|
| A D | internal.h | 791 int iwl_trans_pcie_wait_txq_empty(struct iwl_trans *trans, int txq_idx);
|
| /drivers/scsi/cxgbi/ |
| A D | libcxgbi.h | 120 unsigned short txq_idx; member
|
| /drivers/scsi/cxgbi/cxgb4i/ |
| A D | cxgb4i.c | 1864 csk->txq_idx = cxgb4_port_idx(ndev) * step; in init_act_open() 1893 csk->flags, csk->tx_chan, csk->txq_idx, csk->rss_qid, in init_act_open()
|
| /drivers/net/ethernet/broadcom/ |
| A D | tg3.c | 7399 int txq_idx = tp->txq_cnt - 1; in tg3_napi_disable() local 7407 netif_queue_set_napi(tp->dev, txq_idx, in tg3_napi_disable() 7409 txq_idx--; in tg3_napi_disable() 7422 int txq_idx = 0, rxq_idx = 0; in tg3_napi_enable() local 7430 netif_queue_set_napi(tp->dev, txq_idx, in tg3_napi_enable() 7433 txq_idx++; in tg3_napi_enable()
|