| /drivers/net/ethernet/intel/idpf/ |
| A D | idpf_txrx.c | 182 for (j = 0; j < txq_grp->num_txq; j++) in idpf_tx_desc_rel_all() 1314 u16 *num_txq, u16 *num_rxq) in idpf_vport_calc_numq_per_grp() argument 1319 *num_txq = vport->num_txq; in idpf_vport_calc_numq_per_grp() 1374 tx_qgrp->num_txq = num_txq; in idpf_txq_group_alloc() 1584 u16 num_txq, num_rxq; in idpf_vport_queue_grp_alloc_all() local 1654 for (i = 0; i < vport->num_txq; i++) in idpf_tx_handle_sw_marker() 3905 else if (q_vector->num_txq) in idpf_vport_intr_req_irq() 3968 if (qv->num_txq) { in idpf_vport_intr_ena_irq_all() 3984 if (qv->num_txq || qv->num_rxq) in idpf_vport_intr_ena_irq_all() 4261 u16 num_txq; in idpf_vport_intr_map_vector_to_qs() local [all …]
|
| A D | idpf_ethtool.c | 477 u16 num_txq, num_rxq; in idpf_get_channels() local 482 num_txq = vport_config->user_config.num_req_tx_qs; in idpf_get_channels() 485 combined = min(num_txq, num_rxq); in idpf_get_channels() 498 ch->tx_count = num_txq - combined; in idpf_get_channels() 516 u16 num_txq, num_rxq; in idpf_set_channels() local 532 num_txq = vport_config->user_config.num_req_tx_qs; in idpf_set_channels() 565 vport_config->user_config.num_req_tx_qs = num_txq; in idpf_set_channels() 1118 for (j = 0; j < txq_grp->num_txq; j++) { in idpf_collect_queue_stats() 1186 for (j = 0; j < txq_grp->num_txq; j++, total++) { in idpf_get_ethtool_stats() 1334 if (q_num < vport->num_txq) in idpf_get_q_coalesce() [all …]
|
| A D | idpf_virtchnl.c | 730 for (i = 0; i < vport->num_txq; i++) in idpf_wait_for_marker_event() 738 for (i = 0; i < vport->num_txq; i++) in idpf_wait_for_marker_event() 1314 if (num_regs < vport->num_txq) { in idpf_queue_reg_init() 1321 if (num_regs < vport->num_txq) { in idpf_queue_reg_init() 1854 num_txq = vport->num_txq + vport->num_complq; in idpf_send_ena_dis_queues_msg() 1856 num_q = num_txq + num_rxq; in idpf_send_ena_dis_queues_msg() 1871 if (vport->num_txq != k) in idpf_send_ena_dis_queues_msg() 2028 if (vport->num_txq != k) in idpf_send_map_unmap_queue_vector_msg() 2137 for (i = 0; i < vport->num_txq; i++) in idpf_send_disable_queues_msg() 3567 if (num_ids < vport->num_txq) { in idpf_vport_queue_ids_init() [all …]
|
| A D | idpf_singleq_txrx.c | 541 u16 num_txq = q_vec->num_txq; in idpf_tx_singleq_clean_all() local 545 budget_per_q = num_txq ? max(budget / num_txq, 1) : 0; in idpf_tx_singleq_clean_all() 546 for (i = 0; i < num_txq; i++) { in idpf_tx_singleq_clean_all()
|
| A D | idpf_txrx.h | 396 u16 num_txq; member 919 u16 num_txq; member
|
| A D | idpf.h | 327 u16 num_txq; member
|
| A D | idpf_lib.c | 1317 return netif_set_real_num_tx_queues(vport->netdev, vport->num_txq); in idpf_set_real_num_queues() 1974 err = idpf_send_add_queues_msg(vport, new_vport->num_txq, in idpf_initiate_soft_reset() 1999 idpf_send_add_queues_msg(vport, vport->num_txq, vport->num_complq, in idpf_initiate_soft_reset()
|
| /drivers/net/ethernet/atheros/alx/ |
| A D | main.c | 146 if (r_idx >= alx->num_txq) in alx_tx_queue_mapping() 147 r_idx = r_idx % alx->num_txq; in alx_tx_queue_mapping() 537 for (i = 0; i < alx->num_txq; i++) in alx_free_buffers() 659 alx->num_txq + in alx_alloc_rings() 672 for (i = 0; i < alx->num_txq; i++) { in alx_alloc_rings() 695 for (i = 0; i < alx->num_txq; i++) in alx_free_rings() 759 for (i = 0; i < alx->num_txq; i++) { in alx_alloc_napis() 831 int err, num_vec, num_txq, num_rxq; in alx_enable_msix() local 846 alx->num_txq = num_txq; in alx_enable_msix() 910 alx->num_txq = 1; in alx_init_intr() [all …]
|
| A D | alx.h | 121 int num_txq; member
|
| /drivers/net/ethernet/intel/ice/ |
| A D | ice_sriov.c | 373 u16 num_msix_per_vf, num_txq, num_rxq, avail_qs; in ice_set_per_vf_res() local 400 num_txq = min_t(u16, num_msix_per_vf - ICE_NONQ_VECS_VF, in ice_set_per_vf_res() 404 num_txq = 0; in ice_set_per_vf_res() 405 else if (num_txq > avail_qs) in ice_set_per_vf_res() 406 num_txq = rounddown_pow_of_two(avail_qs); in ice_set_per_vf_res() 416 if (num_txq < ICE_MIN_QS_PER_VF || num_rxq < ICE_MIN_QS_PER_VF) { in ice_set_per_vf_res() 423 pf->vfs.num_qps_per = min_t(int, num_txq, num_rxq); in ice_set_per_vf_res()
|
| A D | ice_lib.c | 189 vsi->num_txq = vsi->req_txq; in ice_vsi_set_num_qs() 590 vsi->num_txq = ch->num_txq; in ice_vsi_alloc_def() 1080 vsi->num_txq = tx_count; in ice_vsi_setup_q_map() 1083 if (vsi->type == ICE_VSI_VF && vsi->num_txq != vsi->num_rxq) { in ice_vsi_setup_q_map() 1088 vsi->num_txq = vsi->num_rxq; in ice_vsi_setup_q_map() 1986 if (vsi->num_txq > ICE_LAN_TXQ_MAX_QDIS) in ice_vsi_stop_tx_rings() 2279 if (!vsi->alloc_txq && vsi->num_txq) in ice_vsi_cfg_tc_lan() 2280 max_txqs[i] = vsi->num_txq; in ice_vsi_cfg_tc_lan() 2908 if (i < vsi->num_txq) in ice_vsi_rebuild_get_coalesce() 3318 vsi->num_txq = new_txq; in ice_vsi_setup_q_map_mqprio() [all …]
|
| A D | ice.h | 162 for ((i) = 0; (i) < (vsi)->num_txq; (i)++) 218 u16 num_txq; member 396 u16 num_txq; /* Used Tx queues */ member
|
| A D | ice_base.c | 823 tx_rings_rem = vsi->num_txq; in ice_vsi_map_rings_to_vectors() 837 q_base = vsi->num_txq - tx_rings_rem; in ice_vsi_map_rings_to_vectors() 1002 return ice_vsi_cfg_txqs(vsi, vsi->tx_rings, vsi->num_txq); in ice_vsi_cfg_lan_txqs()
|
| A D | ice_xsk.c | 172 if (q_idx >= vsi->num_rxq || q_idx >= vsi->num_txq) in ice_qp_dis() 379 if (qid >= vsi->num_rxq || qid >= vsi->num_txq) { in ice_xsk_pool_setup() 1104 if (queue_id >= vsi->num_txq || queue_id >= vsi->num_rxq) in ice_xsk_wakeup()
|
| A D | ice_main.c | 2927 max_txqs[i] = vsi->num_txq; in ice_destroy_xdp_rings() 6935 vsi->num_txq); in ice_update_vsi_ring_stats() 7186 if (!vsi->num_txq || !vsi->num_rxq) in ice_get_stats64() 7373 if (!vsi->num_txq) { in ice_vsi_setup_tx_rings() 8497 if (vsi->num_txq < in ice_validate_mqprio_qopt() 8625 for (i = 0; i < ch->num_txq; i++) { in ice_chnl_cfg_res() 8666 if (ch->num_txq || ch->num_rxq) in ice_chnl_cfg_res() 8797 if (!ch->num_txq || !ch->num_rxq) { in ice_create_q_channel() 8804 vsi->cnt_q_avail, ch->num_txq); in ice_create_q_channel() 8826 vsi->cnt_q_avail -= ch->num_txq; in ice_create_q_channel() [all …]
|
| A D | ice_ethtool.c | 3249 tx_rings = kcalloc(vsi->num_txq, sizeof(*tx_rings), GFP_KERNEL); in ice_set_ringparam() 3756 return min(num_online_cpus(), pf->hw.func_caps.common_cap.num_txq); in ice_get_max_txq() 3809 ch->tx_count = vsi->num_txq - ch->combined_count; in ice_get_channels() 4061 if (q_num < vsi->num_rxq && q_num < vsi->num_txq) { in ice_get_q_coalesce() 4072 } else if (q_num < vsi->num_txq) { in ice_get_q_coalesce() 4229 if (q_num < vsi->num_rxq && q_num < vsi->num_txq) { in ice_set_q_coalesce() 4244 } else if (q_num < vsi->num_txq) { in ice_set_q_coalesce() 4313 if (v_idx >= vsi->num_rxq && v_idx >= vsi->num_txq) in __ice_set_coalesce()
|
| A D | ice_type.h | 276 u16 num_txq; /* Number/Total Tx queues */ member
|
| A D | ice_common.c | 2364 caps->num_txq = number; in ice_parse_common_caps() 2367 caps->num_txq); in ice_parse_common_caps() 3159 func_caps->common_cap.num_txq = 1; in ice_set_safe_mode_caps() 3189 dev_caps->common_cap.num_txq = num_funcs; in ice_set_safe_mode_caps()
|
| /drivers/net/ethernet/brocade/bna/ |
| A D | bna_enet.c | 167 ioceth->attr.num_txq = ntohl(rsp->max_cfg); in bna_bfi_attr_get_rsp() 1679 ioceth->attr.num_txq = BFI_ENET_DEF_TXQ; in bna_attr_init() 1906 attr->num_txq * sizeof(struct bna_tx); in bna_mod_res_req() 1914 attr->num_txq * sizeof(struct bna_txq); in bna_mod_res_req() 2026 bna_num_txq_set(struct bna *bna, int num_txq) in bna_num_txq_set() argument 2029 (num_txq <= bna->ioceth.attr.num_txq)) { in bna_num_txq_set() 2030 bna->ioceth.attr.num_txq = num_txq; in bna_num_txq_set()
|
| A D | bna_types.h | 306 int num_txq; member 476 int num_txq; member 505 int num_txq; member
|
| A D | bna_tx_rx.c | 3097 cfg_req->num_queues = tx->num_txq; in bna_bfi_tx_enet_start() 3098 for (i = 0; i < tx->num_txq; i++) { in bna_bfi_tx_enet_start() 3326 mem_info->num = num_txq; in bna_tx_res_req() 3336 mem_info->num = num_txq; in bna_tx_res_req() 3342 mem_info->num = num_txq; in bna_tx_res_req() 3348 mem_info->num = num_txq; in bna_tx_res_req() 3354 mem_info->num = num_txq; in bna_tx_res_req() 3397 for (i = 0; i < tx_cfg->num_txq; i++) { in bna_tx_create() 3421 tx->num_txq = tx_cfg->num_txq; in bna_tx_create() 3489 if (tx_cfg->num_txq == BFI_TX_MAX_PRIO) in bna_tx_create() [all …]
|
| A D | bna.h | 265 int bna_num_txq_set(struct bna *bna, int num_txq); 306 void bna_tx_res_req(int num_txq, int txq_depth,
|
| A D | bnad.c | 1968 tx_config->num_txq = bnad->num_txq_per_tx; in bnad_setup_tx() 3682 bnad_q_num_adjust(bnad, bna_attr(bna)->num_txq - 1, in bnad_pci_probe()
|
| /drivers/bluetooth/ |
| A D | btintel_pcie.h | 299 num_txq:8, member
|
| /drivers/net/ethernet/intel/ixgbe/ |
| A D | ixgbe_type_e610.h | 850 u32 num_txq; /* Number/Total Tx queues */ member
|