| /drivers/net/ethernet/aquantia/atlantic/ |
| A D | aq_vec.c | 18 unsigned int tx_rings; member 41 for (i = 0U; self->tx_rings > i; ++i) { in aq_vec_poll() 119 self->tx_rings = 0; in aq_vec_alloc() 144 ++self->tx_rings; in aq_vec_ring_alloc() 190 for (i = 0U; self->tx_rings > i; ++i) { in aq_vec_init() 232 for (i = 0U; self->tx_rings > i; ++i) { in aq_vec_start() 256 for (i = 0U; self->tx_rings > i; ++i) { in aq_vec_stop() 276 for (i = 0U; self->tx_rings > i; ++i) { in aq_vec_deinit() 305 for (i = 0U; self->tx_rings > i; ++i) { in aq_vec_ring_free() 314 self->tx_rings = 0; in aq_vec_ring_free() [all …]
|
| A D | aq_hw.h | 66 u8 tx_rings; member
|
| /drivers/net/ethernet/netronome/nfp/ |
| A D | nfp_net_dp.c | 187 dp->tx_rings = kcalloc(dp->num_tx_rings, sizeof(*dp->tx_rings), in nfp_net_tx_rings_prepare() 189 if (!dp->tx_rings) in nfp_net_tx_rings_prepare() 206 nfp_net_tx_ring_init(&dp->tx_rings[r], dp, in nfp_net_tx_rings_prepare() 209 if (nfp_net_tx_ring_alloc(dp, &dp->tx_rings[r])) in nfp_net_tx_rings_prepare() 212 if (nfp_net_tx_ring_bufs_alloc(dp, &dp->tx_rings[r])) in nfp_net_tx_rings_prepare() 220 nfp_net_tx_ring_bufs_free(dp, &dp->tx_rings[r]); in nfp_net_tx_rings_prepare() 222 nfp_net_tx_ring_free(dp, &dp->tx_rings[r]); in nfp_net_tx_rings_prepare() 228 kfree(dp->tx_rings); in nfp_net_tx_rings_prepare() 237 nfp_net_tx_ring_bufs_free(dp, &dp->tx_rings[r]); in nfp_net_tx_rings_free() 238 nfp_net_tx_ring_free(dp, &dp->tx_rings[r]); in nfp_net_tx_rings_free() [all …]
|
| A D | nfp_net_common.c | 797 idx < dp->num_stack_tx_rings ? &dp->tx_rings[idx] : NULL; in nfp_net_vector_assign_rings() 800 &dp->tx_rings[dp->num_stack_tx_rings + idx] : NULL; in nfp_net_vector_assign_rings() 968 nfp_net_tx_ring_reset(&nn->dp, &nn->dp.tx_rings[r]); in nfp_net_clear_config_and_disable() 1001 nfp_net_tx_ring_hw_cfg_write(nn, &nn->dp.tx_rings[r], r); in nfp_net_set_config_and_enable() 1574 new->tx_rings = NULL; in nfp_net_clone_dp()
|
| /drivers/net/wireless/realtek/rtw88/ |
| A D | pci.c | 168 tx_ring = &rtwpci->tx_rings[i]; in rtw_pci_free_trx_ring() 329 tx_ring = &rtwpci->tx_rings[i]; in rtw_pci_init_trx_ring() 351 tx_ring = &rtwpci->tx_rings[i]; in rtw_pci_init_trx_ring() 419 rtwpci->tx_rings[RTW_TX_QUEUE_BK].r.rp = 0; in rtw_pci_reset_buf_desc() 420 rtwpci->tx_rings[RTW_TX_QUEUE_BK].r.wp = 0; in rtw_pci_reset_buf_desc() 544 tx_ring = &rtwpci->tx_rings[queue]; in rtw_pci_dma_release() 623 tx_ring = &rtwpci->tx_rings[queue]; in rtw_pci_deep_ps_enter() 779 ring = &rtwpci->tx_rings[queue]; in rtw_pci_tx_kick_off_queue() 816 ring = &rtwpci->tx_rings[queue]; in rtw_pci_tx_write_data() 929 ring = &rtwpci->tx_rings[queue]; in rtw_pci_tx_write() [all …]
|
| A D | pci.h | 223 struct rtw_pci_tx_ring tx_rings[RTK_MAX_TX_QUEUE_NUM]; member
|
| /drivers/net/ethernet/intel/ice/ |
| A D | ice_ethtool.c | 3249 tx_rings = kcalloc(vsi->num_txq, sizeof(*tx_rings), GFP_KERNEL); in ice_set_ringparam() 3250 if (!tx_rings) { in ice_set_ringparam() 3257 tx_rings[i] = *vsi->tx_rings[i]; in ice_set_ringparam() 3259 tx_rings[i].desc = NULL; in ice_set_ringparam() 3260 tx_rings[i].tx_buf = NULL; in ice_set_ringparam() 3266 kfree(tx_rings); in ice_set_ringparam() 3352 if (tx_rings) { in ice_set_ringparam() 3355 *vsi->tx_rings[i] = tx_rings[i]; in ice_set_ringparam() 3357 kfree(tx_rings); in ice_set_ringparam() 3394 if (tx_rings) { in ice_set_ringparam() [all …]
|
| A D | ice_lib.c | 85 if (!vsi->tx_rings) in ice_vsi_alloc_arrays() 129 devm_kfree(dev, vsi->tx_rings); in ice_vsi_alloc_arrays() 323 devm_kfree(dev, vsi->tx_rings); in ice_vsi_free_arrays() 324 vsi->tx_rings = NULL; in ice_vsi_free_arrays() 393 ring = vsi->tx_rings[i]; in ice_vsi_alloc_ring_stats() 1363 if (vsi->tx_rings) { in ice_vsi_clear_rings() 1365 if (vsi->tx_rings[i]) { in ice_vsi_clear_rings() 1366 kfree_rcu(vsi->tx_rings[i], rcu); in ice_vsi_clear_rings() 1414 WRITE_ONCE(vsi->tx_rings[i], ring); in ice_vsi_alloc_rings() 2666 if (!vsi->tx_rings) in ice_vsi_free_tx_rings() [all …]
|
| A D | ice_base.c | 840 struct ice_tx_ring *tx_ring = vsi->tx_rings[q_id]; in ice_vsi_map_rings_to_vectors() 953 int ice_vsi_cfg_single_txq(struct ice_vsi *vsi, struct ice_tx_ring **tx_rings, in ice_vsi_cfg_single_txq() argument 958 if (q_idx >= vsi->alloc_txq || !tx_rings || !tx_rings[q_idx]) in ice_vsi_cfg_single_txq() 963 return ice_vsi_cfg_txq(vsi, tx_rings[q_idx], qg_buf); in ice_vsi_cfg_single_txq() 1002 return ice_vsi_cfg_txqs(vsi, vsi->tx_rings, vsi->num_txq); in ice_vsi_cfg_lan_txqs()
|
| A D | ice_base.h | 18 int ice_vsi_cfg_single_txq(struct ice_vsi *vsi, struct ice_tx_ring **tx_rings,
|
| A D | ice_dcb_lib.c | 209 return vsi->tx_rings[queue_index]->dcb_tc; in ice_dcb_get_tc() 226 tx_ring = vsi->tx_rings[i]; in ice_vsi_cfg_dcb_rings() 243 vsi->tx_rings[i]->dcb_tc = n; in ice_vsi_cfg_dcb_rings() 265 vsi->tx_rings[i]->dcb_tc = first_droptc; in ice_vsi_cfg_dcb_rings()
|
| A D | ice_xsk.c | 55 ice_clean_tx_ring(vsi->tx_rings[q_idx]); in ice_qp_clean_rings() 175 tx_ring = vsi->tx_rings[q_idx]; in ice_qp_dis() 222 err = ice_vsi_cfg_single_txq(vsi, vsi->tx_rings, q_idx); in ice_qp_ena()
|
| A D | ice_virtchnl.c | 1579 ring = vsi->tx_rings[q_id]; in ice_vf_vsi_dis_single_txq() 1741 vsi->tx_rings[vsi_q_id]->q_vector = q_vector; in ice_cfg_interrupt() 1978 vsi->tx_rings[i]->quanta_prof_id = quanta_prof_id; in ice_vc_cfg_q_quanta() 2056 vsi->tx_rings[q_idx]->dma = qpi->txq.dma_ring_addr; in ice_vc_cfg_qs_msg() 2057 vsi->tx_rings[q_idx]->count = qpi->txq.ring_len; in ice_vc_cfg_qs_msg() 2064 if (ice_vsi_cfg_single_txq(vsi, vsi->tx_rings, q_idx)) { in ice_vc_cfg_qs_msg()
|
| /drivers/net/ethernet/intel/i40e/ |
| A D | i40e_ethtool.c | 2137 if (!tx_rings) { in i40e_set_ringparam() 2146 tx_rings[i] = *vsi->tx_rings[i]; in i40e_set_ringparam() 2161 kfree(tx_rings); in i40e_set_ringparam() 2162 tx_rings = NULL; in i40e_set_ringparam() 2225 if (tx_rings) { in i40e_set_ringparam() 2229 *vsi->tx_rings[i] = tx_rings[i]; in i40e_set_ringparam() 2232 kfree(tx_rings); in i40e_set_ringparam() 2233 tx_rings = NULL; in i40e_set_ringparam() 2262 if (tx_rings) { in i40e_set_ringparam() 2267 kfree(tx_rings); in i40e_set_ringparam() [all …]
|
| A D | i40e_main.c | 361 if (vsi->tx_rings[i] && vsi->tx_rings[i]->desc) { in i40e_tx_timeout() 474 if (!vsi->tx_rings) in i40e_get_netdev_stats_struct() 3317 if (vsi->tx_rings) { in i40e_vsi_free_tx_resources() 3319 if (vsi->tx_rings[i] && vsi->tx_rings[i]->desc) in i40e_vsi_free_tx_resources() 10663 if (vsi->tx_rings) in i40e_clean_xps_state() 10665 if (vsi->tx_rings[i]) in i40e_clean_xps_state() 11453 if (!vsi->tx_rings) in i40e_vsi_alloc_arrays() 11474 kfree(vsi->tx_rings); in i40e_vsi_alloc_arrays() 11584 kfree(vsi->tx_rings); in i40e_vsi_free_arrays() 11585 vsi->tx_rings = NULL; in i40e_vsi_free_arrays() [all …]
|
| A D | i40e_debugfs.c | 274 struct i40e_ring *tx_ring = READ_ONCE(vsi->tx_rings[i]); in i40e_dbg_dump_vsi_seid() 556 if (!vsi->tx_rings || !vsi->tx_rings[0]->desc) { in i40e_dbg_dump_desc() 568 ring = kmemdup(vsi->tx_rings[ring_id], sizeof(*ring), GFP_KERNEL); in i40e_dbg_dump_desc()
|
| /drivers/thunderbolt/ |
| A D | nhi.c | 529 if (!nhi->tx_rings[i]) { in nhi_alloc_hop() 552 if (ring->is_tx && nhi->tx_rings[ring->hop]) { in nhi_alloc_hop() 566 nhi->tx_rings[ring->hop] = ring; in nhi_alloc_hop() 817 ring->nhi->tx_rings[ring->hop] = NULL; in tb_ring_free() 939 ring = nhi->tx_rings[hop]; in nhi_interrupt_work() 1137 if (nhi->tx_rings[i]) in nhi_shutdown() 1358 nhi->tx_rings = devm_kcalloc(&pdev->dev, nhi->hop_count, in nhi_probe() 1359 sizeof(*nhi->tx_rings), GFP_KERNEL); in nhi_probe() 1362 if (!nhi->tx_rings || !nhi->rx_rings) in nhi_probe()
|
| /drivers/net/ethernet/broadcom/ |
| A D | bcmsysport.c | 435 ring = &priv->tx_rings[q]; in bcm_sysport_update_tx_stats() 501 ring = &priv->tx_rings[i]; in bcm_sysport_get_stats() 633 bcm_sysport_set_tx_coalesce(&priv->tx_rings[i], ec); in bcm_sysport_set_coalesce() 989 bcm_sysport_tx_reclaim(priv, &priv->tx_rings[q]); in bcm_sysport_tx_reclaim_all() 1143 txr = &priv->tx_rings[ring]; in bcm_sysport_rx_isr() 1175 txr = &priv->tx_rings[ring]; in bcm_sysport_tx_isr() 1300 ring = &priv->tx_rings[queue]; in bcm_sysport_xmit() 2339 ring = &priv->tx_rings[q]; in bcm_sysport_map_queues() 2371 ring = &priv->tx_rings[q]; in bcm_sysport_unmap_queues() 2491 priv->tx_rings = devm_kcalloc(&pdev->dev, txq, in bcm_sysport_probe() [all …]
|
| /drivers/net/ethernet/intel/iavf/ |
| A D | iavf_main.c | 1474 kfree(adapter->tx_rings); in iavf_free_queues() 1475 adapter->tx_rings = NULL; in iavf_free_queues() 1588 adapter->tx_rings = kcalloc(num_active_queues, in iavf_alloc_queues() 1590 if (!adapter->tx_rings) in iavf_alloc_queues() 1601 tx_ring = &adapter->tx_rings[i]; in iavf_alloc_queues() 3092 if (adapter->tx_rings[i].q_shaper.bw_min || in iavf_reconfig_qs_bw() 3093 adapter->tx_rings[i].q_shaper.bw_max) { in iavf_reconfig_qs_bw() 3436 if (!adapter->tx_rings) in iavf_free_all_tx_resources() 3440 if (adapter->tx_rings[i].desc) in iavf_free_all_tx_resources() 5109 tx_ring = &adapter->tx_rings[handle->id]; in iavf_shaper_set() [all …]
|
| A D | iavf_virtchnl.c | 410 vqpi->txq.ring_len = adapter->tx_rings[i].count; in iavf_configure_queues() 411 vqpi->txq.dma_ring_addr = adapter->tx_rings[i].dma; in iavf_configure_queues() 1774 if (adapter->tx_rings[i].q_shaper_update) in iavf_cfg_queues_bw() 1786 struct iavf_ring *tx_ring = &adapter->tx_rings[i]; in iavf_cfg_queues_bw() 2945 adapter->tx_rings[i].q_shaper_update = false; in iavf_virtchnl_completion()
|
| A D | iavf_ethtool.c | 355 ring = &adapter->tx_rings[i]; in iavf_get_ethtool_stats() 574 tx_ring = &adapter->tx_rings[queue]; in __iavf_get_coalesce() 634 struct iavf_ring *tx_ring = &adapter->tx_rings[queue]; in iavf_set_itr_per_queue()
|
| /drivers/net/ethernet/broadcom/genet/ |
| A D | bcmgenet.c | 1029 tx_rings[num].stats64, packets), \ 1031 tx_rings[num].stats64, bytes), \ 1033 tx_rings[num].stats64, errors), \ 1035 tx_rings[num].stats64, dropped), \ 2141 ring = &priv->tx_rings[index]; in bcmgenet_xmit() 2810 ring = &priv->tx_rings[i]; in bcmgenet_enable_tx_napi() 2822 ring = &priv->tx_rings[i]; in bcmgenet_disable_tx_napi() 2833 ring = &priv->tx_rings[i]; in bcmgenet_fini_tx_napi() 3203 tx_ring = &priv->tx_rings[index]; in bcmgenet_isr1() 3489 bcmgenet_dump_tx_queue(&priv->tx_rings[q]); in bcmgenet_timeout() [all …]
|
| A D | bcmgenet.h | 606 struct bcmgenet_tx_ring tx_rings[GENET_MAX_MQ_CNT + 1]; member
|
| /drivers/net/wireless/realtek/rtw89/ |
| A D | pci.c | 591 tx_ring = &rtwpci->tx_rings[txch]; in rtw89_pci_release_rpp() 1120 struct rtw89_pci_tx_ring *tx_ring = &rtwpci->tx_rings[txch]; in __rtw89_pci_check_and_reclaim_tx_resource_noio() 1137 struct rtw89_pci_tx_ring *tx_ring = &rtwpci->tx_rings[txch]; in __rtw89_pci_check_and_reclaim_tx_resource() 1230 struct rtw89_pci_tx_ring *tx_ring = &rtwpci->tx_rings[txch]; in rtw89_pci_ops_tx_kick_off() 1250 tx_ring = &rtwpci->tx_rings[txch]; in rtw89_pci_tx_kick_off_pending() 1524 tx_ring = &rtwpci->tx_rings[txch]; in rtw89_pci_tx_write() 1630 tx_ring = &rtwpci->tx_rings[i]; in rtw89_pci_reset_trx_rings() 1777 tx_ring = &rtwpci->tx_rings[i]; in rtw89_pci_switch_bd_idx_addr() 3237 tx_ring = &rtwpci->tx_rings[i]; in rtw89_pci_free_tx_rings() 3435 tx_ring = &rtwpci->tx_rings[i]; in rtw89_pci_alloc_tx_rings() [all …]
|
| /drivers/net/ethernet/sun/ |
| A D | niu.c | 3610 index = (rp - np->tx_rings); in niu_tx_work() 4327 if (np->tx_rings) { in niu_free_channels() 4333 kfree(np->tx_rings); in niu_free_channels() 4334 np->tx_rings = NULL; in niu_free_channels() 4475 struct tx_ring_info *tx_rings; in niu_alloc_channels() local 4531 if (!tx_rings) in niu_alloc_channels() 4536 np->tx_rings = tx_rings; in niu_alloc_channels() 6278 tx_rings = READ_ONCE(np->tx_rings); in niu_get_tx_stats() 6279 if (!tx_rings) in niu_get_tx_stats() 6483 if (np->tx_rings) { in niu_reset_buffers() [all …]
|