| /drivers/net/ethernet/broadcom/bnxt/ |
| A D | bnxt_xdp.c | 193 rx_buf = &rxr->rx_buf_ring[cons]; in bnxt_xdp_buff_init() 245 txr = rxr->bnapi->tx_ring[0]; in bnxt_rx_xdp() 269 rx_buf = &rxr->rx_buf_ring[cons]; in bnxt_rx_xdp() 282 bnxt_xdp_buff_frags_free(rxr, xdp); in bnxt_rx_xdp() 283 bnxt_reuse_rx_data(rxr, cons, page); in bnxt_rx_xdp() 292 NEXT_RX(rxr->rx_prod), xdp); in bnxt_rx_xdp() 293 bnxt_reuse_rx_data(rxr, cons, page); in bnxt_rx_xdp() 302 if (bnxt_alloc_rx_data(bp, rxr, rxr->rx_prod, GFP_ATOMIC)) { in bnxt_rx_xdp() 304 bnxt_xdp_buff_frags_free(rxr, xdp); in bnxt_rx_xdp() 324 bnxt_xdp_buff_frags_free(rxr, xdp); in bnxt_rx_xdp() [all …]
|
| A D | bnxt.c | 1611 rxr->rx_prod = NEXT_RX(rxr->rx_prod); in bnxt_tpa_start() 3113 bnxt_db_write(bp, &rxr->rx_db, rxr->rx_prod); in __bnxt_poll_work_done() 3204 bnxt_db_write(bp, &rxr->rx_db, rxr->rx_prod); in bnxt_poll_nitroa0() 3802 rxr->page_pool = rxr->head_pool = NULL; in bnxt_free_rx_rings() 7294 bnxt_db_write(bp, &rxr->rx_db, rxr->rx_prod); in bnxt_hwrm_rx_agg_ring_alloc() 7390 bnxt_db_write(bp, &rxr->rx_db, rxr->rx_prod); in bnxt_hwrm_ring_alloc() 13975 if (!rxr) in bnxt_dump_rx_sw_state() 14197 bnxt_db_write(bp, &rxr->rx_db, rxr->rx_prod); in bnxt_rx_ring_reset() 15834 memcpy(clone, rxr, sizeof(*rxr)); in bnxt_queue_mem_alloc() 16036 bnxt_db_write(bp, &rxr->rx_db, rxr->rx_prod); in bnxt_queue_start() [all …]
|
| A D | bnxt_xdp.h | 20 bool bnxt_rx_xdp(struct bnxt *bp, struct bnxt_rx_ring_info *rxr, u16 cons, 27 bool bnxt_xdp_attached(struct bnxt *bp, struct bnxt_rx_ring_info *rxr); 29 void bnxt_xdp_buff_init(struct bnxt *bp, struct bnxt_rx_ring_info *rxr, 32 void bnxt_xdp_buff_frags_free(struct bnxt_rx_ring_info *rxr,
|
| A D | bnxt_ethtool.c | 4870 struct bnxt_rx_ring_info *rxr; in bnxt_rx_loopback() local 4878 rxr = bnapi->rx_ring; in bnxt_rx_loopback() 4883 rx_buf = &rxr->rx_buf_ring[cons]; in bnxt_rx_loopback() 4938 struct bnxt_rx_ring_info *rxr = &bp->rx_ring[0]; in bnxt_run_loopback() local 4946 cpr = &rxr->bnapi->cp_ring; in bnxt_run_loopback() 4948 cpr = rxr->rx_cpr; in bnxt_run_loopback()
|
| A D | bnxt.h | 2878 int bnxt_alloc_rx_data(struct bnxt *bp, struct bnxt_rx_ring_info *rxr, 2880 void bnxt_reuse_rx_data(struct bnxt_rx_ring_info *rxr, u16 cons, void *data);
|
| /drivers/net/ethernet/broadcom/ |
| A D | bnx2.c | 774 rxr->rx_buf_ring = in bnx2_alloc_rx_mem() 776 if (!rxr->rx_buf_ring) in bnx2_alloc_rx_mem() 791 rxr->rx_pg_ring = in bnx2_alloc_rx_mem() 3295 BNX2_WR16(bp, rxr->rx_pg_bidx_addr, rxr->rx_pg_prod); in bnx2_rx_int() 3299 BNX2_WR(bp, rxr->rx_bseq_addr, rxr->rx_prod_bseq); in bnx2_rx_int() 5124 rxr->rx_prod = 0; in bnx2_clear_ring_states() 5125 rxr->rx_cons = 0; in bnx2_clear_ring_states() 5233 bnx2_init_rxbd_rings(rxr->rx_desc_ring, rxr->rx_desc_mapping, in bnx2_init_rx_ring() 5291 rxr->rx_prod = prod; in bnx2_init_rx_ring() 5297 BNX2_WR16(bp, rxr->rx_pg_bidx_addr, rxr->rx_pg_prod); in bnx2_init_rx_ring() [all …]
|
| /drivers/net/ethernet/meta/fbnic/ |
| A D | fbnic_txrx.c | 1260 struct fbnic_ring *rxr) in fbnic_remove_rx_ring() argument 1268 WARN_ON(fbn->rx[rxr->q_idx] && fbn->rx[rxr->q_idx] != rxr); in fbnic_remove_rx_ring() 1269 fbn->rx[rxr->q_idx] = NULL; in fbnic_remove_rx_ring() 1609 switch (rxr->doorbell - fbnic_ring_csr_base(rxr)) { in fbnic_alloc_rx_ring_desc() 1628 rxr->desc = dma_alloc_coherent(dev, size, &rxr->dma, in fbnic_alloc_rx_ring_desc() 1630 if (!rxr->desc) in fbnic_alloc_rx_ring_desc() 1635 rxr->size = size; in fbnic_alloc_rx_ring_desc() 1642 size_t size = array_size(sizeof(*rxr->rx_buf), rxr->size_mask + 1); in fbnic_alloc_rx_ring_buffer() 1645 size = sizeof(*rxr->rx_buf) * (rxr->size_mask + 1); in fbnic_alloc_rx_ring_buffer() 1647 size = sizeof(*rxr->pkt); in fbnic_alloc_rx_ring_buffer() [all …]
|
| A D | fbnic_netdev.c | 492 struct fbnic_ring *rxr = fbn->rx[i]; in fbnic_get_stats64() local 494 if (!rxr) in fbnic_get_stats64() 497 stats = &rxr->stats; in fbnic_get_stats64() 528 struct fbnic_ring *rxr = fbn->rx[idx]; in fbnic_get_queue_stats_rx() local 535 if (!rxr) in fbnic_get_queue_stats_rx() 538 stats = &rxr->stats; in fbnic_get_queue_stats_rx()
|
| A D | fbnic_txrx.h | 148 struct fbnic_ring *rxr);
|
| /drivers/net/ethernet/sgi/ |
| A D | ioc3-eth.c | 84 unsigned long *rxr; /* pointer to receiver ring */ member 371 unsigned long *rxr; in ioc3_rx() local 375 rxr = ip->rxr; /* Ring base */ in ioc3_rx() 396 d = rxr[rx_entry]; in ioc3_rx() 405 dma_unmap_single(ip->dma_dev, rxr[rx_entry], in ioc3_rx() 420 d = rxr[rx_entry]; in ioc3_rx() 660 be64_to_cpu(ip->rxr[n_entry]), in ioc3_free_rx_bufs() 684 ip->rxr[i] = cpu_to_be64(ioc3_map(d, PCI64_ATTR_BAR)); in ioc3_alloc_rx_bufs() 887 if (!ip->rxr) { in ioc3eth_probe() 954 if (ip->rxr) in ioc3eth_probe() [all …]
|
| /drivers/net/ethernet/amazon/ena/ |
| A D | ena_netdev.c | 187 struct ena_ring *txr, *rxr; in ena_init_io_rings() local 194 rxr = &adapter->rx_ring[i]; in ena_init_io_rings() 212 ena_init_io_rings_common(adapter, rxr, i); in ena_init_io_rings() 215 rxr->ring_size = adapter->requested_rx_ring_size; in ena_init_io_rings() 216 rxr->rx_copybreak = adapter->rx_copybreak; in ena_init_io_rings() 217 rxr->sgl_size = adapter->max_rx_sgl_size; in ena_init_io_rings() 218 rxr->smoothed_interval = in ena_init_io_rings() 220 rxr->empty_rx_queue = 0; in ena_init_io_rings() 221 rxr->rx_headroom = NET_SKB_PAD; in ena_init_io_rings() 223 rxr->xdp_ring = &adapter->tx_ring[i + adapter->num_io_queues]; in ena_init_io_rings()
|
| /drivers/net/ethernet/intel/iavf/ |
| A D | iavf_txrx.h | 312 bool iavf_alloc_rx_buffers(struct iavf_ring *rxr, u16 cleaned_count);
|
| /drivers/net/ethernet/intel/i40e/ |
| A D | i40e_txrx.h | 461 bool i40e_alloc_rx_buffers(struct i40e_ring *rxr, u16 cleaned_count);
|
| A D | i40e_main.c | 13353 struct i40e_ring *rxr = vsi->rx_rings[queue_pair]; in i40e_queue_pair_toggle_napi() local 13354 struct i40e_q_vector *q_vector = rxr->q_vector; in i40e_queue_pair_toggle_napi() 13429 struct i40e_ring *rxr = vsi->rx_rings[queue_pair]; in i40e_queue_pair_enable_irq() local 13435 i40e_irq_dynamic_enable(vsi, rxr->q_vector->v_idx); in i40e_queue_pair_enable_irq() 13449 struct i40e_ring *rxr = vsi->rx_rings[queue_pair]; in i40e_queue_pair_disable_irq() local 13460 u32 intpf = vsi->base_vector + rxr->q_vector->v_idx; in i40e_queue_pair_disable_irq()
|
| /drivers/net/ethernet/intel/ice/ |
| A D | ice_txrx.h | 495 bool ice_alloc_rx_bufs(struct ice_rx_ring *rxr, unsigned int cleaned_count);
|