| /drivers/net/ethernet/intel/idpf/ |
| A D | idpf_ethtool.c | 32 cmd->data = vport->num_rxq; in idpf_get_rxnfc() 154 u16 num_rxq; in idpf_add_flow_steer() local 174 if (q_index >= num_rxq) in idpf_add_flow_steer() 477 u16 num_txq, num_rxq; in idpf_get_channels() local 485 combined = min(num_txq, num_rxq); in idpf_get_channels() 516 u16 num_txq, num_rxq; in idpf_set_channels() local 1075 u16 num_rxq; in idpf_collect_queue_stats() local 1080 num_rxq = rxq_grp->singleq.num_rxq; in idpf_collect_queue_stats() 1210 u16 num_rxq; in idpf_get_ethtool_stats() local 1217 num_rxq = rxq_grp->singleq.num_rxq; in idpf_get_ethtool_stats() [all …]
|
| A D | idpf_txrx.c | 529 u16 num_rxq; in idpf_rx_desc_rel_all() local 821 int num_rxq = rx_qgrp->singleq.num_rxq; in idpf_rx_bufs_init_all() local 926 u16 num_rxq; in idpf_rx_desc_alloc_all() local 933 num_rxq = rx_qgrp->singleq.num_rxq; in idpf_rx_desc_alloc_all() 1048 u16 num_rxq; in idpf_rxq_group_rel() local 1062 num_rxq = rx_qgrp->singleq.num_rxq; in idpf_rxq_group_rel() 1324 *num_rxq = vport->num_rxq; in idpf_vport_calc_numq_per_grp() 1470 rx_qgrp->singleq.num_rxq = num_rxq; in idpf_rxq_group_alloc() 4122 u16 num_rxq = q_vec->num_rxq; in idpf_rx_splitq_clean_all() local 4131 budget_per_q = num_rxq ? max(budget / num_rxq, 1) : 0; in idpf_rx_splitq_clean_all() [all …]
|
| A D | idpf_virtchnl.c | 1248 u16 num_rxq = rx_qgrp->singleq.num_rxq; in __idpf_queue_reg_init() local 1707 u16 num_rxq; in idpf_send_config_rx_queues_msg() local 1736 num_rxq = rx_qgrp->singleq.num_rxq; in idpf_send_config_rx_queues_msg() 1855 num_rxq = vport->num_rxq + vport->num_bufq; in idpf_send_ena_dis_queues_msg() 1856 num_q = num_txq + num_rxq; in idpf_send_ena_dis_queues_msg() 1894 num_rxq = rx_qgrp->singleq.num_rxq; in idpf_send_ena_dis_queues_msg() 1932 vport->num_rxq)) in idpf_send_ena_dis_queues_msg() 2033 u16 num_rxq; in idpf_send_map_unmap_queue_vector_msg() local 2038 num_rxq = rx_qgrp->singleq.num_rxq; in idpf_send_map_unmap_queue_vector_msg() 3488 u16 num_rxq; in __idpf_vport_queue_ids_init() local [all …]
|
| A D | idpf_ptp.c | 344 for (i = 0; i < grp->singleq.num_rxq; i++) { in idpf_ptp_update_phctime_rxq_grp() 693 u16 j, num_rxq; in idpf_ptp_set_rx_tstamp() local 696 num_rxq = grp->splitq.num_rxq_sets; in idpf_ptp_set_rx_tstamp() 698 num_rxq = grp->singleq.num_rxq; in idpf_ptp_set_rx_tstamp() 700 for (j = 0; j < num_rxq; j++) { in idpf_ptp_set_rx_tstamp()
|
| A D | idpf_singleq_txrx.c | 1076 u16 num_rxq = q_vec->num_rxq; in idpf_rx_singleq_clean_all() local 1083 budget_per_q = num_rxq ? max(budget / num_rxq, 1) : 0; in idpf_rx_singleq_clean_all() 1084 for (i = 0; i < num_rxq; i++) { in idpf_rx_singleq_clean_all()
|
| A D | idpf_txrx.h | 395 u16 num_rxq; member 890 u16 num_rxq; member
|
| A D | idpf_lib.c | 1313 err = netif_set_real_num_rx_queues(vport->netdev, vport->num_rxq); in idpf_set_real_num_queues() 1359 for (j = 0; j < grp->singleq.num_rxq; j++) { in idpf_rx_init_buf_tail() 1976 new_vport->num_rxq, in idpf_initiate_soft_reset() 2000 vport->num_rxq, vport->num_bufq); in idpf_initiate_soft_reset()
|
| A D | idpf.h | 338 u16 num_rxq; member
|
| /drivers/net/ethernet/intel/ice/ |
| A D | ice_lib.c | 202 vsi->num_rxq = vsi->req_rxq; in ice_vsi_set_num_qs() 589 vsi->num_rxq = ch->num_rxq; in ice_vsi_alloc_def() 1081 vsi->num_rxq = rx_count; in ice_vsi_setup_q_map() 1088 vsi->num_txq = vsi->num_rxq; in ice_vsi_setup_q_map() 1098 ctxt->info.q_mapping[1] = cpu_to_le16(vsi->num_rxq); in ice_vsi_setup_q_map() 1191 qcount = vsi->num_rxq; in ice_chnl_vsi_setup_q_map() 1517 vsi->orig_rss_size <= vsi->num_rxq) { in ice_vsi_cfg_rss_lut_key() 2910 if (i < vsi->num_rxq) in ice_vsi_rebuild_get_coalesce() 3319 vsi->num_rxq = new_rxq; in ice_vsi_setup_q_map_mqprio() 3329 if (tc0_qcount && tc0_qcount < vsi->num_rxq) { in ice_vsi_setup_q_map_mqprio() [all …]
|
| A D | ice_sriov.c | 373 u16 num_msix_per_vf, num_txq, num_rxq, avail_qs; in ice_set_per_vf_res() local 408 num_rxq = min_t(u16, num_msix_per_vf - ICE_NONQ_VECS_VF, in ice_set_per_vf_res() 412 num_rxq = 0; in ice_set_per_vf_res() 413 else if (num_rxq > avail_qs) in ice_set_per_vf_res() 414 num_rxq = rounddown_pow_of_two(avail_qs); in ice_set_per_vf_res() 416 if (num_txq < ICE_MIN_QS_PER_VF || num_rxq < ICE_MIN_QS_PER_VF) { in ice_set_per_vf_res() 423 pf->vfs.num_qps_per = min_t(int, num_txq, num_rxq); in ice_set_per_vf_res()
|
| A D | ice.h | 168 for ((i) = 0; (i) < (vsi)->num_rxq; (i)++) 217 u16 num_rxq; member 398 u16 num_rxq; /* Used Rx queues */ member
|
| A D | ice_base.c | 636 if (q_idx >= vsi->num_rxq) in ice_vsi_cfg_single_rxq() 824 rx_rings_rem = vsi->num_rxq; in ice_vsi_map_rings_to_vectors() 854 q_base = vsi->num_rxq - rx_rings_rem; in ice_vsi_map_rings_to_vectors()
|
| A D | ice_xsk.c | 172 if (q_idx >= vsi->num_rxq || q_idx >= vsi->num_txq) in ice_qp_dis() 379 if (qid >= vsi->num_rxq || qid >= vsi->num_txq) { in ice_xsk_pool_setup() 1104 if (queue_id >= vsi->num_txq || queue_id >= vsi->num_rxq) in ice_xsk_wakeup()
|
| A D | ice_ethtool_fdir.c | 1470 if (!vsi->num_rxq || q_index >= vsi->num_rxq) in ice_update_per_q_fltr() 1809 (*ring >= (ch->base_q + ch->num_rxq))) in ice_update_ring_dest_vsi() 1856 if (ring >= vsi->num_rxq) in ice_set_fdir_input_set()
|
| A D | ice_ethtool.c | 3308 rx_rings = kcalloc(vsi->num_rxq, sizeof(*rx_rings), GFP_KERNEL); in ice_set_ringparam() 3765 return min(num_online_cpus(), pf->hw.func_caps.common_cap.num_rxq); in ice_get_max_rxq() 3808 ch->rx_count = vsi->num_rxq - ch->combined_count; in ice_get_channels() 4061 if (q_num < vsi->num_rxq && q_num < vsi->num_txq) { in ice_get_q_coalesce() 4068 } else if (q_num < vsi->num_rxq) { in ice_get_q_coalesce() 4229 if (q_num < vsi->num_rxq && q_num < vsi->num_txq) { in ice_set_q_coalesce() 4239 } else if (q_num < vsi->num_rxq) { in ice_set_q_coalesce() 4313 if (v_idx >= vsi->num_rxq && v_idx >= vsi->num_txq) in __ice_set_coalesce()
|
| A D | ice_main.c | 4047 pf->max_pf_rxqs = func_caps->common_cap.num_rxq; in ice_set_pf_caps() 7186 if (!vsi->num_txq || !vsi->num_rxq) in ice_get_stats64() 7405 if (!vsi->num_rxq) { in ice_vsi_setup_rx_rings() 8494 if (vsi->num_rxq < in ice_validate_mqprio_qopt() 8666 if (ch->num_txq || ch->num_rxq) in ice_chnl_cfg_res() 8720 vsi->next_base_q = vsi->next_base_q + ch->num_rxq; in ice_setup_hw_channel() 8722 ch->num_rxq); in ice_setup_hw_channel() 8797 if (!ch->num_txq || !ch->num_rxq) { in ice_create_q_channel() 8920 for (i = 0; i < ch->num_rxq; i++) { in ice_remove_q_channels() 9096 ch->num_rxq = vsi->mqprio_qopt.qopt.count[i]; in ice_create_q_channels() [all …]
|
| A D | ice_type.h | 274 u16 num_rxq; /* Number/Total Rx queues */ member
|
| A D | ice_tc_lib.c | 1066 return queue < vsi->num_rxq ? vsi->rx_rings[queue] : NULL; in ice_locate_rx_ring_using_queue() 2043 if (queue >= vsi->num_rxq) { in ice_tc_forward_to_queue()
|
| A D | ice_common.c | 2356 caps->num_rxq = number; in ice_parse_common_caps() 2359 caps->num_rxq); in ice_parse_common_caps() 3158 func_caps->common_cap.num_rxq = 1; in ice_set_safe_mode_caps() 3188 dev_caps->common_cap.num_rxq = num_funcs; in ice_set_safe_mode_caps()
|
| /drivers/net/ethernet/stmicro/stmmac/ |
| A D | stmmac_fpe.c | 50 u32 num_rxq = priv->plat->rx_queues_to_use; in stmmac_fpe_configure_tx() local 59 value |= (num_rxq - 1) << __ffs(reg->fprq_mask); in stmmac_fpe_configure_tx()
|
| A D | dwxgmac2_core.c | 584 struct stmmac_rss *cfg, u32 num_rxq) in dwxgmac2_rss_configure() argument 610 for (i = 0; i < num_rxq; i++) in dwxgmac2_rss_configure()
|
| A D | hwif.h | 398 struct stmmac_rss *cfg, u32 num_rxq);
|
| /drivers/net/ethernet/atheros/alx/ |
| A D | alx.h | 122 int num_rxq; member
|
| A D | main.c | 831 int err, num_vec, num_txq, num_rxq; in alx_enable_msix() local 834 num_rxq = 1; in alx_enable_msix() 835 num_vec = max_t(int, num_txq, num_rxq) + 1; in alx_enable_msix() 847 alx->num_rxq = num_rxq; in alx_enable_msix() 911 alx->num_rxq = 1; in alx_init_intr() 1236 netif_set_real_num_rx_queues(alx->dev, alx->num_rxq); in __alx_open()
|
| /drivers/net/ethernet/intel/ixgbe/ |
| A D | ixgbe_type_e610.h | 848 u32 num_rxq; /* Number/Total Rx queues */ member
|