| /linux/drivers/net/wwan/t7xx/ |
| A D | t7xx_hif_dpmaif_rx.c | 83 struct dpmaif_rx_queue *rxq = &dpmaif_ctrl->rxq[q_num]; in t7xx_dpmaif_update_bat_wr_idx() local 585 ret = t7xx_dpmaif_rx_buf_alloc(rxq->dpmaif_ctrl, rxq->bat_req, rxq->index, bid_cnt, false); in t7xx_dpmaif_bat_release_and_add() 816 struct dpmaif_rx_queue *rxq = &dpmaif_ctrl->rxq[q_num]; in t7xx_dpmaif_napi_rx_data_collect() local 901 rxq = &dpmaif_ctrl->rxq[qno]; in t7xx_dpmaif_irq_rx_done() 1022 rxq->bat_req = &rxq->dpmaif_ctrl->bat_req; in t7xx_dpmaif_rx_alloc() 1025 rxq->bat_frag = &rxq->dpmaif_ctrl->bat_frag; in t7xx_dpmaif_rx_alloc() 1035 t7xx_dpmaif_bat_free(rxq->dpmaif_ctrl, rxq->bat_req); in t7xx_dpmaif_rx_buf_free() 1036 t7xx_dpmaif_bat_free(rxq->dpmaif_ctrl, rxq->bat_frag); in t7xx_dpmaif_rx_buf_free() 1041 rxq->pit_base, rxq->pit_bus_addr); in t7xx_dpmaif_rx_buf_free() 1073 rxq = &dpmaif_ctrl->rxq[DPF_RX_QNO_DFT]; in t7xx_dpmaif_bat_release_work() [all …]
|
| A D | t7xx_hif_dpmaif.c | 267 rx_q = &dpmaif_ctrl->rxq[i]; in t7xx_dpmaif_rxtx_sw_allocs() 303 struct dpmaif_rx_queue *rxq; in t7xx_dpmaif_start() local 314 rxq = &dpmaif_ctrl->rxq[i]; in t7xx_dpmaif_start() 315 rxq->que_started = true; in t7xx_dpmaif_start() 316 rxq->index = i; in t7xx_dpmaif_start() 317 rxq->budget = rxq->bat_req->bat_size_cnt - 1; in t7xx_dpmaif_start() 372 t7xx_dpmaif_bat_free(rxq->dpmaif_ctrl, rxq->bat_frag); in t7xx_dpmaif_start() 375 t7xx_dpmaif_bat_free(rxq->dpmaif_ctrl, rxq->bat_req); in t7xx_dpmaif_start() 432 struct dpmaif_rx_queue *rxq; in t7xx_dpmaif_start_txrx_qs() local 442 rxq = &dpmaif_ctrl->rxq[que_cnt]; in t7xx_dpmaif_start_txrx_qs() [all …]
|
| /linux/drivers/net/ethernet/huawei/hinic/ |
| A D | hinic_rx.c | 138 skb = netdev_alloc_skb_ip_align(rxq->netdev, rxq->rq->buf_sz); in rx_alloc_skb() 447 rx_alloc_pkts(rxq); in rxq_recv() 482 netif_napi_add_weight(rxq->netdev, &rxq->napi, rx_poll, in rx_add_napi() 525 rx_add_napi(rxq); in rx_request_irq() 545 err = request_irq(rq->irq, rx_irq, 0, rxq->irq_name, rxq); in rx_request_irq() 559 rx_del_napi(rxq); in rx_request_irq() 569 rx_del_napi(rxq); in rx_free_irq() 587 rxq->rq = rq; in hinic_init_rxq() 591 rxq_stats_init(rxq); in hinic_init_rxq() 595 if (!rxq->irq_name) in hinic_init_rxq() [all …]
|
| A D | hinic_rx.h | 44 void hinic_rxq_get_stats(struct hinic_rxq *rxq, struct hinic_rxq_stats *stats); 46 int hinic_init_rxq(struct hinic_rxq *rxq, struct hinic_rq *rq, 49 void hinic_clean_rxq(struct hinic_rxq *rxq);
|
| /linux/drivers/net/wireless/intel/iwlwifi/pcie/ |
| A D | rx.c | 130 return (rxq->read - rxq->write - 1) & (rxq->queue_size - 1); in iwl_rxq_space() 209 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; in iwl_pcie_rxq_check_wrptr() local 238 (u32)rxb->vid, rxq->id, rxq->write); in iwl_pcie_restock_bd() 272 rxq->write = (rxq->write + 1) & (rxq->queue_size - 1); in iwl_pcie_rxmq_restock() 311 rxb = rxq->queue[rxq->write]; in iwl_pcie_rxsq_restock() 322 rxq->queue[rxq->write] = rxb; in iwl_pcie_rxsq_restock() 686 rxq->bd, rxq->bd_dma); in iwl_pcie_free_rxq_dma() 697 rxq->used_bd, rxq->used_bd_dma); in iwl_pcie_free_rxq_dma() 1078 if (rxq && rxq->napi.poll) in iwl_pcie_rx_napi_sync() 1511 rxq = &trans_pcie->rxq[queue]; in iwl_pcie_rx_handle() [all …]
|
| /linux/drivers/infiniband/hw/hfi1/ |
| A D | netdev_rx.c | 194 rx->rxq = kcalloc_node(rx->num_rx_q, sizeof(*rx->rxq), in hfi1_netdev_rxq_init() 197 if (!rx->rxq) { in hfi1_netdev_rxq_init() 203 struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; in hfi1_netdev_rxq_init() local 210 rxq->rx = rx; in hfi1_netdev_rxq_init() 211 rxq->rcd->napi = &rxq->napi; in hfi1_netdev_rxq_init() 230 struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; in hfi1_netdev_rxq_init() local 238 kfree(rx->rxq); in hfi1_netdev_rxq_init() 239 rx->rxq = NULL; in hfi1_netdev_rxq_init() 250 struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; in hfi1_netdev_rxq_deinit() local 268 struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; in enable_queues() local [all …]
|
| A D | vnic_main.c | 295 struct hfi1_vnic_vport_info *vinfo = rxq->vinfo; in hfi1_vnic_decap_skb() 303 vinfo->stats[rxq->idx].rx_oversize++; in hfi1_vnic_decap_skb() 305 vinfo->stats[rxq->idx].rx_runt++; in hfi1_vnic_decap_skb() 336 struct hfi1_vnic_rx_queue *rxq; in hfi1_vnic_bypass_rcv() local 370 rxq = &vinfo->rxq[q_idx]; in hfi1_vnic_bypass_rcv() 389 rc = hfi1_vnic_decap_skb(rxq, skb); in hfi1_vnic_bypass_rcv() 401 napi_gro_receive(&rxq->napi, skb); in hfi1_vnic_bypass_rcv() 599 struct hfi1_vnic_rx_queue *rxq = &vinfo->rxq[i]; in hfi1_vnic_alloc_rn() local 601 rxq->idx = i; in hfi1_vnic_alloc_rn() 602 rxq->vinfo = vinfo; in hfi1_vnic_alloc_rn() [all …]
|
| /linux/drivers/net/ethernet/qlogic/qede/ |
| A D | qede_fp.c | 57 sw_rx_data = &rxq->sw_rx_ring[rxq->sw_rx_prod & NUM_RX_BDS_MAX]; in qede_alloc_rx_buffer() 69 rxq->sw_rx_prod++; in qede_alloc_rx_buffer() 518 rxq->sw_rx_cons++; in qede_rx_bd_ring_consume() 531 curr_prod = &rxq->sw_rx_ring[rxq->sw_rx_prod & NUM_RX_BDS_MAX]; in qede_reuse_page() 552 curr_cons = &rxq->sw_rx_ring[rxq->sw_rx_cons & NUM_RX_BDS_MAX]; in qede_recycle_rx_bd_ring() 668 struct sw_rx_data *current_bd = &rxq->sw_rx_ring[rxq->sw_rx_cons & in qede_fill_frag_skb() 976 struct qede_rx_queue *rxq = fp->rxq; in qede_tpa_end() local 1090 xdp_init_buff(&xdp, rxq->rx_buf_seg_size, &rxq->xdp_rxq); in qede_rx_xdp() 1184 u16 cur_size = pkt_len > rxq->rx_buf_size ? rxq->rx_buf_size : in qede_rx_build_jumbo() 1341 struct qede_rx_queue *rxq = fp->rxq; in qede_rx_int() local [all …]
|
| A D | qede_main.c | 964 if (fp->rxq && xdp_rxq_info_is_reg(&fp->rxq->xdp_rxq)) in qede_free_fp_array() 966 kfree(fp->rxq); in qede_free_fp_array() 1035 fp->rxq = kzalloc(sizeof(*fp->rxq), GFP_KERNEL); in qede_alloc_fp_array() 1036 if (!fp->rxq) in qede_alloc_fp_array() 1532 for (i = rxq->sw_rx_cons; i != rxq->sw_rx_prod; i++) { in qede_free_rx_buffers() 1553 kfree(rxq->sw_rx_ring); in qede_free_mem_rxq() 1586 size = rxq->rx_headroom + in qede_alloc_mem_rxq() 1607 if (!rxq->sw_rx_ring) { in qede_alloc_mem_rxq() 1632 rxq->filled_buffers = 0; in qede_alloc_mem_rxq() 1644 qede_set_tpa_param(rxq); in qede_alloc_mem_rxq() [all …]
|
| /linux/drivers/net/ethernet/microsoft/mana/ |
| A D | mana_en.c | 1486 if (rxq->buf_index == rxq->num_rx_buf) in mana_post_pkt_rxq() 1714 rxbuf_oob = &rxq->rx_oobs[rxq->buf_index]; in mana_process_rx_cqe() 1739 rxq->gdma_id, cq->gdma_id, rxq->rxobj); in mana_process_rx_cqe() 1763 struct mana_rxq *rxq = cq->rxq; in mana_poll_rx_cq() local 2027 if (!rxq) in mana_destroy_rxq() 2072 kfree(rxq); in mana_destroy_rxq() 2202 if (!rxq) in mana_create_rxq() 2239 cq->rxq = rxq; in mana_create_rxq() 2273 rxq->gdma_id = rxq->gdma_rq->id; in mana_create_rxq() 2299 return rxq; in mana_create_rxq() [all …]
|
| A D | mana_bpf.c | 80 u32 mana_run_xdp(struct net_device *ndev, struct mana_rxq *rxq, in mana_run_xdp() argument 88 prog = rcu_dereference(rxq->bpf_prog); in mana_run_xdp() 93 xdp_init_buff(xdp, PAGE_SIZE, &rxq->xdp_rxq); in mana_run_xdp() 98 rx_stats = &rxq->stats; in mana_run_xdp() 107 rxq->xdp_rc = xdp_do_redirect(ndev, xdp, prog); in mana_run_xdp() 108 if (!rxq->xdp_rc) { in mana_run_xdp() 109 rxq->xdp_flush = true; in mana_run_xdp()
|
| /linux/drivers/bluetooth/ |
| A D | btintel_pcie.c | 178 static void btintel_pcie_prepare_rx(struct rxq *rxq, u16 frbd_index) in btintel_pcie_prepare_rx() argument 197 struct rxq *rxq = &data->rxq; in btintel_pcie_submit_rx() local 538 struct rxq *rxq; in btintel_pcie_msix_rx_handle() local 555 rxq = &data->rxq; in btintel_pcie_msix_rx_handle() 827 struct rxq *rxq) in btintel_pcie_free_rxq_bufs() argument 831 rxq->buf_v_addr, rxq->buf_p_addr); in btintel_pcie_free_rxq_bufs() 832 kfree(rxq->bufs); in btintel_pcie_free_rxq_bufs() 836 struct rxq *rxq) in btintel_pcie_setup_rxq_bufs() argument 842 rxq->bufs = kmalloc_array(rxq->count, sizeof(*buf), GFP_KERNEL); in btintel_pcie_setup_rxq_bufs() 843 if (!rxq->bufs) in btintel_pcie_setup_rxq_bufs() [all …]
|
| /linux/drivers/net/ethernet/marvell/ |
| A D | mvneta.c | 134 #define MVNETA_CPU_RXQ_ACCESS(rxq) BIT(rxq) argument 927 prefetch(rxq->descs + rxq->next_desc_to_proc); in mvneta_rxq_next_desc_get() 1500 for (rxq = 0; rxq < rxq_number; rxq++) in mvneta_defaults_set() 2054 rxq->id, i, rxq->refill_num); in mvneta_rx_refill_queue() 3402 if (mvneta_rx_refill(pp, rxq->descs + i, rxq, in mvneta_rxq_fill() 3452 rxq->last_desc = rxq->size - 1; in mvneta_rxq_sw_init() 3465 mvneta_rx_pkts_coal_set(pp, rxq, rxq->pkts_coal); in mvneta_rxq_hw_init() 3466 mvneta_rx_time_coal_set(pp, rxq, rxq->time_coal); in mvneta_rxq_hw_init() 3475 mvneta_rxq_fill(pp, rxq, rxq->size); in mvneta_rxq_hw_init() 4344 int rxq; in mvneta_percpu_elect() local [all …]
|
| A D | mv643xx_eth.c | 443 return container_of(rxq, struct mv643xx_eth_private, rxq[rxq->index]); in rxq_to_mp() 519 rx_desc = &rxq->rx_desc_area[rxq->rx_curr_desc]; in rxq_process() 526 skb = rxq->rx_skb[rxq->rx_curr_desc]; in rxq_process() 527 rxq->rx_skb[rxq->rx_curr_desc] = NULL; in rxq_process() 530 if (rxq->rx_curr_desc == rxq->rx_ring_size) in rxq_process() 624 if (rxq->rx_used_desc == rxq->rx_ring_size) in rxq_refill() 1940 struct rx_queue *rxq = mp->rxq + index; in rxq_init() local 1973 rxq->rx_skb = kcalloc(rxq->rx_ring_size, sizeof(*rxq->rx_skb), in rxq_init() 2010 rxq_disable(rxq); in rxq_deinit() 2029 rxq->rx_desc_area, rxq->rx_desc_dma); in rxq_deinit() [all …]
|
| /linux/net/core/ |
| A D | devmem.c | 108 struct netdev_rx_queue *rxq; in net_devmem_unbind_dmabuf() local 118 rxq->mp_params.mp_priv = NULL; in net_devmem_unbind_dmabuf() 120 rxq_idx = get_netdev_rx_queue_index(rxq); in net_devmem_unbind_dmabuf() 134 struct netdev_rx_queue *rxq; in net_devmem_bind_dmabuf_to_queue() local 143 rxq = __netif_get_rx_queue(dev, rxq_idx); in net_devmem_bind_dmabuf_to_queue() 144 if (rxq->mp_params.mp_priv) { in net_devmem_bind_dmabuf_to_queue() 150 if (rxq->pool) { in net_devmem_bind_dmabuf_to_queue() 161 rxq->mp_params.mp_priv = binding; in net_devmem_bind_dmabuf_to_queue() 170 rxq->mp_params.mp_priv = NULL; in net_devmem_bind_dmabuf_to_queue() 308 struct netdev_rx_queue *rxq; in dev_dmabuf_uninstall() local [all …]
|
| /linux/drivers/net/ethernet/atheros/alx/ |
| A D | main.c | 73 struct alx_rx_queue *rxq = alx->qnapi[0]->rxq; in alx_refill_rx_ring() local 228 rrd = &rxq->rrd[rxq->rrd_read_idx]; in alx_clean_rx_irq() 241 rxb = &rxq->bufs[rxq->read_idx]; in alx_clean_rx_irq() 281 if (++rxq->read_idx == rxq->count) in alx_clean_rx_irq() 283 if (++rxq->rrd_read_idx == rxq->count) in alx_clean_rx_irq() 307 if (np->rxq) in alx_poll() 467 if (np->rxq) { in alx_init_ring_ptrs() 511 if (!rxq->bufs) in alx_free_rxring_buf() 778 rxq = kzalloc(sizeof(*rxq), GFP_KERNEL); in alx_alloc_napis() 779 if (!rxq) in alx_alloc_napis() [all …]
|
| /linux/drivers/net/ethernet/intel/idpf/ |
| A D | idpf_txrx.c | 448 .pp = rxq->pp, in idpf_rx_buf_rel_all() 459 rxq->pp = NULL; in idpf_rx_buf_rel_all() 473 if (!rxq) in idpf_rx_desc_rel() 476 if (rxq->skb) { in idpf_rx_desc_rel() 490 dmam_free_coherent(dev, rxq->size, rxq->desc_ring, rxq->dma); in idpf_rx_desc_rel() 707 if (idpf_rx_singleq_buf_hw_alloc_all(rxq, rxq->desc_count - 1)) in idpf_rx_buf_alloc_singleq() 867 rxq->size = rxq->desc_count * sizeof(union virtchnl2_rx_desc); in idpf_rx_desc_alloc() 870 rxq->size = ALIGN(rxq->size, 4096); in idpf_rx_desc_alloc() 871 rxq->desc_ring = dmam_alloc_coherent(dev, rxq->size, in idpf_rx_desc_alloc() 875 rxq->size); in idpf_rx_desc_alloc() [all …]
|
| /linux/drivers/net/ethernet/hisilicon/ |
| A D | hisi_femac.c | 212 struct hisi_femac_queue *rxq = &priv->rxq; in hisi_femac_rx_refill() local 218 pos = rxq->head; in hisi_femac_rx_refill() 220 if (!CIRC_SPACE(pos, rxq->tail, rxq->num)) in hisi_femac_rx_refill() 238 rxq->skb[pos] = skb; in hisi_femac_rx_refill() 242 rxq->head = pos; in hisi_femac_rx_refill() 248 struct hisi_femac_queue *rxq = &priv->rxq; in hisi_femac_rx() local 253 pos = rxq->tail; in hisi_femac_rx() 292 rxq->tail = pos; in hisi_femac_rx() 388 struct hisi_femac_queue *rxq = &priv->rxq; in hisi_femac_free_skb_rings() local 393 pos = rxq->tail; in hisi_femac_free_skb_rings() [all …]
|
| /linux/drivers/net/wireless/intel/iwlegacy/ |
| A D | 3945-mac.c | 929 struct il_rx_queue *rxq = &il->rxq; in il3945_rx_queue_restock() local 942 rxq->bd[rxq->write] = in il3945_rx_queue_restock() 944 rxq->queue[rxq->write] = rxb; in il3945_rx_queue_restock() 945 rxq->write = (rxq->write + 1) & RX_QUEUE_MASK; in il3945_rx_queue_restock() 956 if (rxq->write_actual != (rxq->write & ~0x7) || in il3945_rx_queue_restock() 957 abs(rxq->write - rxq->read) > 7) { in il3945_rx_queue_restock() 976 struct il_rx_queue *rxq = &il->rxq; in il3945_rx_allocate() local 1070 list_add_tail(&rxq->pool[i].list, &rxq->rx_used); in il3945_rx_queue_reset() 1075 rxq->read = rxq->write = 0; in il3945_rx_queue_reset() 1125 rxq->rb_stts, rxq->rb_stts_dma); in il3945_rx_queue_free() [all …]
|
| /linux/drivers/vhost/ |
| A D | net.c | 152 if (rxq->tail != rxq->head) in vhost_net_buf_get_ptr() 153 return rxq->queue[rxq->head]; in vhost_net_buf_get_ptr() 160 return rxq->tail - rxq->head; in vhost_net_buf_get_size() 165 return rxq->tail == rxq->head; in vhost_net_buf_is_empty() 171 ++rxq->head; in vhost_net_buf_consume() 177 struct vhost_net_buf *rxq = &nvq->rxq; in vhost_net_buf_produce() local 187 struct vhost_net_buf *rxq = &nvq->rxq; in vhost_net_buf_unproduce() local 190 ptr_ring_unconsume(nvq->rx_ring, rxq->queue + rxq->head, in vhost_net_buf_unproduce() 193 rxq->head = rxq->tail = 0; in vhost_net_buf_unproduce() 210 struct vhost_net_buf *rxq = &nvq->rxq; in vhost_net_buf_peek() local [all …]
|
| /linux/tools/testing/selftests/bpf/ |
| A D | xdp_hw_metadata.c | 63 int rxq; variable 408 struct pollfd fds[rxq + 1]; in verify_metadata() 415 for (i = 0; i < rxq; i++) { in verify_metadata() 421 fds[rxq].fd = server_fd; in verify_metadata() 422 fds[rxq].events = POLLIN; in verify_metadata() 423 fds[rxq].revents = 0; in verify_metadata() 443 if (fds[rxq].revents) in verify_metadata() 579 for (i = 0; i < rxq; i++) in cleanup() 670 rxq = rxq_num(ifname); in main() 672 printf("rxq: %d\n", rxq); in main() [all …]
|
| /linux/drivers/net/ethernet/mellanox/mlx5/core/ |
| A D | en_arfs.c | 90 int rxq; member 437 arfs_rule->rxq, arfs_rule->flow_id, in arfs_may_expire_flow() 528 priv->channel_stats[arfs_rule->rxq]->rq.arfs_err++; in arfs_add_rule() 601 priv->channel_stats[arfs_rule->rxq]->rq.arfs_err++; in arfs_add_rule() 614 struct mlx5_flow_handle *rule, u16 rxq) in arfs_modify_rule_rq() argument 623 priv->channel_stats[rxq]->rq.arfs_err++; in arfs_modify_rule_rq() 650 arfs_rule->rxq); in arfs_handle_work() 659 u16 rxq, u32 flow_id) in arfs_alloc_rule() argument 667 priv->channel_stats[rxq]->rq.arfs_err++; in arfs_alloc_rule() 672 rule->rxq = rxq; in arfs_alloc_rule() [all …]
|
| /linux/drivers/net/ethernet/chelsio/cxgb4vf/ |
| A D | cxgb4vf_main.c | 383 int rxq, msi, err; in request_msix_queue_irqs() local 409 while (--rxq >= 0) in request_msix_queue_irqs() 421 int rxq, msi; in free_msix_queue_irqs() local 452 int rxq; in enable_rx() local 477 int rxq; in quiesce_rx() local 650 memset(&rxq->stats, 0, sizeof(rxq->stats)); in setup_sge_queues() 668 IQ_MAP(s, rxq->rspq.abs_id) = &rxq->rspq; in setup_sge_queues() 680 rxq->fl.abs_id = rxq->fl.cntxt_id + s->egr_base; in setup_sge_queues() 681 EQ_MAP(s, rxq->fl.abs_id) = &rxq->fl; in setup_sge_queues() 1242 rxq++; in cxgb4vf_poll_controller() [all …]
|
| /linux/drivers/net/ethernet/fungible/funeth/ |
| A D | funeth_trace.h | 75 TP_PROTO(const struct funeth_rxq *rxq, 81 TP_ARGS(rxq, num_rqes, pkt_len, hash, cls_vec), 90 __string(devname, rxq->netdev->name) 94 __entry->qidx = rxq->qidx; 95 __entry->cq_head = rxq->cq_head;
|
| /linux/drivers/net/ethernet/freescale/ |
| A D | fec_main.c | 489 err = xdp_rxq_info_reg(&rxq->xdp_rxq, fep->netdev, rxq->id, 0); in fec_enet_create_page_pool() 494 rxq->page_pool); in fec_enet_create_page_pool() 504 rxq->page_pool = NULL; in fec_enet_create_page_pool() 962 bdp = rxq->bd.base; in fec_enet_bd_init() 978 rxq->bd.cur = rxq->bd.base; in fec_enet_bd_init() 1725 bdp = rxq->bd.cur; in fec_enet_rx_queue() 1886 rxq->bd.cur = bdp; in fec_enet_rx_queue() 3051 rxq->stats[j] = 0; in fec_enet_clear_ethtool_stats() 3402 bdp = rxq->bd.base; in fec_enet_alloc_rxq_buffers() 3404 err = fec_enet_create_page_pool(fep, rxq, rxq->bd.ring_size); in fec_enet_alloc_rxq_buffers() [all …]
|