Home
last modified time | relevance | path

Searched refs:rxq (Results 1 – 25 of 157) sorted by relevance

1234567

/drivers/net/wwan/t7xx/
A Dt7xx_hif_dpmaif_rx.c84 struct dpmaif_rx_queue *rxq = &dpmaif_ctrl->rxq[q_num]; in t7xx_dpmaif_update_bat_wr_idx() local
586 ret = t7xx_dpmaif_rx_buf_alloc(rxq->dpmaif_ctrl, rxq->bat_req, rxq->index, bid_cnt, false); in t7xx_dpmaif_bat_release_and_add()
817 struct dpmaif_rx_queue *rxq = &dpmaif_ctrl->rxq[q_num]; in t7xx_dpmaif_napi_rx_data_collect() local
902 rxq = &dpmaif_ctrl->rxq[qno]; in t7xx_dpmaif_irq_rx_done()
1023 rxq->bat_req = &rxq->dpmaif_ctrl->bat_req; in t7xx_dpmaif_rx_alloc()
1026 rxq->bat_frag = &rxq->dpmaif_ctrl->bat_frag; in t7xx_dpmaif_rx_alloc()
1036 t7xx_dpmaif_bat_free(rxq->dpmaif_ctrl, rxq->bat_req); in t7xx_dpmaif_rx_buf_free()
1037 t7xx_dpmaif_bat_free(rxq->dpmaif_ctrl, rxq->bat_frag); in t7xx_dpmaif_rx_buf_free()
1042 rxq->pit_base, rxq->pit_bus_addr); in t7xx_dpmaif_rx_buf_free()
1074 rxq = &dpmaif_ctrl->rxq[DPF_RX_QNO_DFT]; in t7xx_dpmaif_bat_release_work()
[all …]
A Dt7xx_hif_dpmaif.c267 rx_q = &dpmaif_ctrl->rxq[i]; in t7xx_dpmaif_rxtx_sw_allocs()
303 struct dpmaif_rx_queue *rxq; in t7xx_dpmaif_start() local
314 rxq = &dpmaif_ctrl->rxq[i]; in t7xx_dpmaif_start()
315 rxq->que_started = true; in t7xx_dpmaif_start()
316 rxq->index = i; in t7xx_dpmaif_start()
317 rxq->budget = rxq->bat_req->bat_size_cnt - 1; in t7xx_dpmaif_start()
372 t7xx_dpmaif_bat_free(rxq->dpmaif_ctrl, rxq->bat_frag); in t7xx_dpmaif_start()
375 t7xx_dpmaif_bat_free(rxq->dpmaif_ctrl, rxq->bat_req); in t7xx_dpmaif_start()
432 struct dpmaif_rx_queue *rxq; in t7xx_dpmaif_start_txrx_qs() local
442 rxq = &dpmaif_ctrl->rxq[que_cnt]; in t7xx_dpmaif_start_txrx_qs()
[all …]
/drivers/net/ethernet/huawei/hinic/
A Dhinic_rx.c138 skb = netdev_alloc_skb_ip_align(rxq->netdev, rxq->rq->buf_sz); in rx_alloc_skb()
447 rx_alloc_pkts(rxq); in rxq_recv()
482 netif_napi_add_weight(rxq->netdev, &rxq->napi, rx_poll, in rx_add_napi()
525 rx_add_napi(rxq); in rx_request_irq()
545 err = request_irq(rq->irq, rx_irq, 0, rxq->irq_name, rxq); in rx_request_irq()
559 rx_del_napi(rxq); in rx_request_irq()
569 rx_del_napi(rxq); in rx_free_irq()
587 rxq->rq = rq; in hinic_init_rxq()
591 rxq_stats_init(rxq); in hinic_init_rxq()
595 if (!rxq->irq_name) in hinic_init_rxq()
[all …]
/drivers/net/wireless/intel/iwlwifi/pcie/gen1_2/
A Drx.c131 return (rxq->read - rxq->write - 1) & (rxq->queue_size - 1); in iwl_rxq_space()
210 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; in iwl_pcie_rxq_check_wrptr() local
239 (u32)rxb->vid, rxq->id, rxq->write); in iwl_pcie_restock_bd()
273 rxq->write = (rxq->write + 1) & (rxq->queue_size - 1); in iwl_pcie_rxmq_restock()
312 rxb = rxq->queue[rxq->write]; in iwl_pcie_rxsq_restock()
323 rxq->queue[rxq->write] = rxb; in iwl_pcie_rxsq_restock()
687 rxq->bd, rxq->bd_dma); in iwl_pcie_free_rxq_dma()
698 rxq->used_bd, rxq->used_bd_dma); in iwl_pcie_free_rxq_dma()
1078 if (rxq && rxq->napi.poll) in iwl_pcie_rx_napi_sync()
1511 rxq = &trans_pcie->rxq[queue]; in iwl_pcie_rx_handle()
[all …]
/drivers/net/ethernet/huawei/hinic3/
A Dhinic3_rx.c81 rx_info = &rxq->rx_info[rxq->next_to_update]; in hinic3_rx_fill_buffers()
84 rxq->buf_len); in hinic3_rx_fill_buffers()
90 rq_wqe_buf_set(rxq->rq, rxq->next_to_update, dma_addr, in hinic3_rx_fill_buffers()
92 rxq->next_to_update = (rxq->next_to_update + 1) & rxq->q_mask; in hinic3_rx_fill_buffers()
96 hinic3_write_db(rxq->rq, rxq->q_id & 3, DB_CFLAG_DP_RQ, in hinic3_rx_fill_buffers()
98 rxq->delta -= i; in hinic3_rx_fill_buffers()
99 rxq->next_to_alloc = rxq->next_to_update; in hinic3_rx_fill_buffers()
117 rxq->buf_len); in hinic3_add_rx_frag()
141 sw_ci = rxq->cons_idx & rxq->q_mask; in packaging_skb()
146 size = rxq->buf_len; in packaging_skb()
[all …]
/drivers/net/ethernet/qlogic/qede/
A Dqede_fp.c57 sw_rx_data = &rxq->sw_rx_ring[rxq->sw_rx_prod & NUM_RX_BDS_MAX]; in qede_alloc_rx_buffer()
69 rxq->sw_rx_prod++; in qede_alloc_rx_buffer()
518 rxq->sw_rx_cons++; in qede_rx_bd_ring_consume()
531 curr_prod = &rxq->sw_rx_ring[rxq->sw_rx_prod & NUM_RX_BDS_MAX]; in qede_reuse_page()
552 curr_cons = &rxq->sw_rx_ring[rxq->sw_rx_cons & NUM_RX_BDS_MAX]; in qede_recycle_rx_bd_ring()
668 struct sw_rx_data *current_bd = &rxq->sw_rx_ring[rxq->sw_rx_cons & in qede_fill_frag_skb()
976 struct qede_rx_queue *rxq = fp->rxq; in qede_tpa_end() local
1090 xdp_init_buff(&xdp, rxq->rx_buf_seg_size, &rxq->xdp_rxq); in qede_rx_xdp()
1184 u16 cur_size = pkt_len > rxq->rx_buf_size ? rxq->rx_buf_size : in qede_rx_build_jumbo()
1341 struct qede_rx_queue *rxq = fp->rxq; in qede_rx_int() local
[all …]
A Dqede_main.c964 if (fp->rxq && xdp_rxq_info_is_reg(&fp->rxq->xdp_rxq)) in qede_free_fp_array()
966 kfree(fp->rxq); in qede_free_fp_array()
1035 fp->rxq = kzalloc(sizeof(*fp->rxq), GFP_KERNEL); in qede_alloc_fp_array()
1036 if (!fp->rxq) in qede_alloc_fp_array()
1532 for (i = rxq->sw_rx_cons; i != rxq->sw_rx_prod; i++) { in qede_free_rx_buffers()
1553 kfree(rxq->sw_rx_ring); in qede_free_mem_rxq()
1586 size = rxq->rx_headroom + in qede_alloc_mem_rxq()
1607 if (!rxq->sw_rx_ring) { in qede_alloc_mem_rxq()
1632 rxq->filled_buffers = 0; in qede_alloc_mem_rxq()
1644 qede_set_tpa_param(rxq); in qede_alloc_mem_rxq()
[all …]
/drivers/infiniband/hw/hfi1/
A Dnetdev_rx.c194 rx->rxq = kcalloc_node(rx->num_rx_q, sizeof(*rx->rxq), in hfi1_netdev_rxq_init()
197 if (!rx->rxq) { in hfi1_netdev_rxq_init()
203 struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; in hfi1_netdev_rxq_init() local
210 rxq->rx = rx; in hfi1_netdev_rxq_init()
211 rxq->rcd->napi = &rxq->napi; in hfi1_netdev_rxq_init()
230 struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; in hfi1_netdev_rxq_init() local
238 kfree(rx->rxq); in hfi1_netdev_rxq_init()
239 rx->rxq = NULL; in hfi1_netdev_rxq_init()
250 struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; in hfi1_netdev_rxq_deinit() local
268 struct hfi1_netdev_rxq *rxq = &rx->rxq[i]; in enable_queues() local
[all …]
A Dvnic_main.c295 struct hfi1_vnic_vport_info *vinfo = rxq->vinfo; in hfi1_vnic_decap_skb()
303 vinfo->stats[rxq->idx].rx_oversize++; in hfi1_vnic_decap_skb()
305 vinfo->stats[rxq->idx].rx_runt++; in hfi1_vnic_decap_skb()
336 struct hfi1_vnic_rx_queue *rxq; in hfi1_vnic_bypass_rcv() local
370 rxq = &vinfo->rxq[q_idx]; in hfi1_vnic_bypass_rcv()
389 rc = hfi1_vnic_decap_skb(rxq, skb); in hfi1_vnic_bypass_rcv()
401 napi_gro_receive(&rxq->napi, skb); in hfi1_vnic_bypass_rcv()
599 struct hfi1_vnic_rx_queue *rxq = &vinfo->rxq[i]; in hfi1_vnic_alloc_rn() local
601 rxq->idx = i; in hfi1_vnic_alloc_rn()
602 rxq->vinfo = vinfo; in hfi1_vnic_alloc_rn()
[all …]
/drivers/net/ethernet/microsoft/mana/
A Dmana_en.c1720 if (rxq->buf_index == rxq->num_rx_buf) in mana_post_pkt_rxq()
1937 rxbuf_oob = &rxq->rx_oobs[rxq->buf_index]; in mana_process_rx_cqe()
1962 rxq->gdma_id, cq->gdma_id, rxq->rxobj); in mana_process_rx_cqe()
1986 struct mana_rxq *rxq = cq->rxq; in mana_poll_rx_cq() local
2287 if (!rxq) in mana_destroy_rxq()
2336 kfree(rxq); in mana_destroy_rxq()
2467 if (!rxq) in mana_create_rxq()
2504 cq->rxq = rxq; in mana_create_rxq()
2538 rxq->gdma_id = rxq->gdma_rq->id; in mana_create_rxq()
2568 return rxq; in mana_create_rxq()
[all …]
A Dmana_bpf.c80 u32 mana_run_xdp(struct net_device *ndev, struct mana_rxq *rxq, in mana_run_xdp() argument
88 prog = rcu_dereference(rxq->bpf_prog); in mana_run_xdp()
93 xdp_init_buff(xdp, PAGE_SIZE, &rxq->xdp_rxq); in mana_run_xdp()
98 rx_stats = &rxq->stats; in mana_run_xdp()
107 rxq->xdp_rc = xdp_do_redirect(ndev, xdp, prog); in mana_run_xdp()
108 if (!rxq->xdp_rc) { in mana_run_xdp()
109 rxq->xdp_flush = true; in mana_run_xdp()
/drivers/net/ethernet/marvell/
A Dmvneta.c134 #define MVNETA_CPU_RXQ_ACCESS(rxq) BIT(rxq) argument
927 prefetch(rxq->descs + rxq->next_desc_to_proc); in mvneta_rxq_next_desc_get()
1500 for (rxq = 0; rxq < rxq_number; rxq++) in mvneta_defaults_set()
2054 rxq->id, i, rxq->refill_num); in mvneta_rx_refill_queue()
3405 if (mvneta_rx_refill(pp, rxq->descs + i, rxq, in mvneta_rxq_fill()
3455 rxq->last_desc = rxq->size - 1; in mvneta_rxq_sw_init()
3468 mvneta_rx_pkts_coal_set(pp, rxq, rxq->pkts_coal); in mvneta_rxq_hw_init()
3469 mvneta_rx_time_coal_set(pp, rxq, rxq->time_coal); in mvneta_rxq_hw_init()
3478 mvneta_rxq_fill(pp, rxq, rxq->size); in mvneta_rxq_hw_init()
4386 int rxq; in mvneta_percpu_elect() local
[all …]
A Dmv643xx_eth.c443 return container_of(rxq, struct mv643xx_eth_private, rxq[rxq->index]); in rxq_to_mp()
519 rx_desc = &rxq->rx_desc_area[rxq->rx_curr_desc]; in rxq_process()
526 skb = rxq->rx_skb[rxq->rx_curr_desc]; in rxq_process()
527 rxq->rx_skb[rxq->rx_curr_desc] = NULL; in rxq_process()
530 if (rxq->rx_curr_desc == rxq->rx_ring_size) in rxq_process()
624 if (rxq->rx_used_desc == rxq->rx_ring_size) in rxq_refill()
1937 struct rx_queue *rxq = mp->rxq + index; in rxq_init() local
1970 rxq->rx_skb = kcalloc(rxq->rx_ring_size, sizeof(*rxq->rx_skb), in rxq_init()
2007 rxq_disable(rxq); in rxq_deinit()
2026 rxq->rx_desc_area, rxq->rx_desc_dma); in rxq_deinit()
[all …]
/drivers/net/ethernet/atheros/alx/
A Dmain.c73 struct alx_rx_queue *rxq = alx->qnapi[0]->rxq; in alx_refill_rx_ring() local
228 rrd = &rxq->rrd[rxq->rrd_read_idx]; in alx_clean_rx_irq()
241 rxb = &rxq->bufs[rxq->read_idx]; in alx_clean_rx_irq()
281 if (++rxq->read_idx == rxq->count) in alx_clean_rx_irq()
283 if (++rxq->rrd_read_idx == rxq->count) in alx_clean_rx_irq()
307 if (np->rxq) in alx_poll()
467 if (np->rxq) { in alx_init_ring_ptrs()
511 if (!rxq->bufs) in alx_free_rxring_buf()
778 rxq = kzalloc(sizeof(*rxq), GFP_KERNEL); in alx_alloc_napis()
779 if (!rxq) in alx_alloc_napis()
[all …]
/drivers/net/ethernet/intel/idpf/
A Didpf_txrx.c449 .pp = rxq->pp, in idpf_rx_buf_rel_all()
460 rxq->pp = NULL; in idpf_rx_buf_rel_all()
474 if (!rxq) in idpf_rx_desc_rel()
477 if (rxq->skb) { in idpf_rx_desc_rel()
491 dmam_free_coherent(dev, rxq->size, rxq->desc_ring, rxq->dma); in idpf_rx_desc_rel()
708 if (idpf_rx_singleq_buf_hw_alloc_all(rxq, rxq->desc_count - 1)) in idpf_rx_buf_alloc_singleq()
868 rxq->size = rxq->desc_count * sizeof(union virtchnl2_rx_desc); in idpf_rx_desc_alloc()
871 rxq->size = ALIGN(rxq->size, 4096); in idpf_rx_desc_alloc()
872 rxq->desc_ring = dmam_alloc_coherent(dev, rxq->size, in idpf_rx_desc_alloc()
876 rxq->size); in idpf_rx_desc_alloc()
[all …]
/drivers/net/ethernet/hisilicon/
A Dhisi_femac.c212 struct hisi_femac_queue *rxq = &priv->rxq; in hisi_femac_rx_refill() local
218 pos = rxq->head; in hisi_femac_rx_refill()
220 if (!CIRC_SPACE(pos, rxq->tail, rxq->num)) in hisi_femac_rx_refill()
238 rxq->skb[pos] = skb; in hisi_femac_rx_refill()
242 rxq->head = pos; in hisi_femac_rx_refill()
248 struct hisi_femac_queue *rxq = &priv->rxq; in hisi_femac_rx() local
253 pos = rxq->tail; in hisi_femac_rx()
292 rxq->tail = pos; in hisi_femac_rx()
388 struct hisi_femac_queue *rxq = &priv->rxq; in hisi_femac_free_skb_rings() local
393 pos = rxq->tail; in hisi_femac_free_skb_rings()
[all …]
/drivers/net/wireless/intel/iwlegacy/
A D3945-mac.c929 struct il_rx_queue *rxq = &il->rxq; in il3945_rx_queue_restock() local
942 rxq->bd[rxq->write] = in il3945_rx_queue_restock()
944 rxq->queue[rxq->write] = rxb; in il3945_rx_queue_restock()
945 rxq->write = (rxq->write + 1) & RX_QUEUE_MASK; in il3945_rx_queue_restock()
956 if (rxq->write_actual != (rxq->write & ~0x7) || in il3945_rx_queue_restock()
957 abs(rxq->write - rxq->read) > 7) { in il3945_rx_queue_restock()
976 struct il_rx_queue *rxq = &il->rxq; in il3945_rx_allocate() local
1070 list_add_tail(&rxq->pool[i].list, &rxq->rx_used); in il3945_rx_queue_reset()
1075 rxq->read = rxq->write = 0; in il3945_rx_queue_reset()
1125 rxq->rb_stts, rxq->rb_stts_dma); in il3945_rx_queue_free()
[all …]
/drivers/bluetooth/
A Dbtintel_pcie.c433 static void btintel_pcie_prepare_rx(struct rxq *rxq, u16 frbd_index) in btintel_pcie_prepare_rx() argument
452 struct rxq *rxq = &data->rxq; in btintel_pcie_submit_rx() local
477 struct rxq *rxq = &data->rxq; in btintel_pcie_start_rx() local
1441 struct rxq *rxq; in btintel_pcie_msix_rx_handle() local
1456 rxq = &data->rxq; in btintel_pcie_msix_rx_handle()
1753 struct rxq *rxq) in btintel_pcie_free_rxq_bufs() argument
1757 rxq->buf_v_addr, rxq->buf_p_addr); in btintel_pcie_free_rxq_bufs()
1758 kfree(rxq->bufs); in btintel_pcie_free_rxq_bufs()
1762 struct rxq *rxq) in btintel_pcie_setup_rxq_bufs() argument
1768 rxq->bufs = kmalloc_array(rxq->count, sizeof(*buf), GFP_KERNEL); in btintel_pcie_setup_rxq_bufs()
[all …]
/drivers/vhost/
A Dnet.c155 if (rxq->tail != rxq->head) in vhost_net_buf_get_ptr()
156 return rxq->queue[rxq->head]; in vhost_net_buf_get_ptr()
163 return rxq->tail - rxq->head; in vhost_net_buf_get_size()
168 return rxq->tail == rxq->head; in vhost_net_buf_is_empty()
174 ++rxq->head; in vhost_net_buf_consume()
180 struct vhost_net_buf *rxq = &nvq->rxq; in vhost_net_buf_produce() local
190 struct vhost_net_buf *rxq = &nvq->rxq; in vhost_net_buf_unproduce() local
193 ptr_ring_unconsume(nvq->rx_ring, rxq->queue + rxq->head, in vhost_net_buf_unproduce()
196 rxq->head = rxq->tail = 0; in vhost_net_buf_unproduce()
213 struct vhost_net_buf *rxq = &nvq->rxq; in vhost_net_buf_peek() local
[all …]
/drivers/net/ethernet/chelsio/cxgb4vf/
A Dcxgb4vf_main.c383 int rxq, msi, err; in request_msix_queue_irqs() local
409 while (--rxq >= 0) in request_msix_queue_irqs()
421 int rxq, msi; in free_msix_queue_irqs() local
452 int rxq; in enable_rx() local
477 int rxq; in quiesce_rx() local
650 memset(&rxq->stats, 0, sizeof(rxq->stats)); in setup_sge_queues()
668 IQ_MAP(s, rxq->rspq.abs_id) = &rxq->rspq; in setup_sge_queues()
680 rxq->fl.abs_id = rxq->fl.cntxt_id + s->egr_base; in setup_sge_queues()
681 EQ_MAP(s, rxq->fl.abs_id) = &rxq->fl; in setup_sge_queues()
1242 rxq++; in cxgb4vf_poll_controller()
[all …]
A Dsge.c1591 rxq->stats.vlan_ex++; in do_gro()
1599 rxq->stats.pkts++; in do_gro()
1600 rxq->stats.rx_cso++; in do_gro()
1630 do_gro(rxq, gl, pkt); in t4vf_ethrx_handler()
1647 rxq->stats.pkts++; in t4vf_ethrx_handler()
1653 rxq->stats.rx_cso++; in t4vf_ethrx_handler()
1658 rxq->stats.rx_cso++; in t4vf_ethrx_handler()
1807 sdesc = &rxq->fl.sdesc[rxq->fl.cidx]; in process_responses()
1869 fl_cap(&rxq->fl) - rxq->fl.avail >= 2*FL_PER_EQ_UNIT) in process_responses()
2571 if (rxq->rspq.desc) in t4vf_free_sge_resources()
[all …]
/drivers/net/ethernet/mellanox/mlx5/core/
A Den_arfs.c90 int rxq; member
437 arfs_rule->rxq, arfs_rule->flow_id, in arfs_may_expire_flow()
528 priv->channel_stats[arfs_rule->rxq]->rq.arfs_err++; in arfs_add_rule()
601 priv->channel_stats[arfs_rule->rxq]->rq.arfs_err++; in arfs_add_rule()
614 struct mlx5_flow_handle *rule, u16 rxq) in arfs_modify_rule_rq() argument
623 priv->channel_stats[rxq]->rq.arfs_err++; in arfs_modify_rule_rq()
650 arfs_rule->rxq); in arfs_handle_work()
659 u16 rxq, u32 flow_id) in arfs_alloc_rule() argument
667 priv->channel_stats[rxq]->rq.arfs_err++; in arfs_alloc_rule()
672 rule->rxq = rxq; in arfs_alloc_rule()
[all …]
/drivers/net/ethernet/freescale/
A Dfec_main.c491 err = xdp_rxq_info_reg(&rxq->xdp_rxq, fep->netdev, rxq->id, 0); in fec_enet_create_page_pool()
496 rxq->page_pool); in fec_enet_create_page_pool()
506 rxq->page_pool = NULL; in fec_enet_create_page_pool()
998 bdp = rxq->bd.base; in fec_enet_bd_init()
1014 rxq->bd.cur = rxq->bd.base; in fec_enet_bd_init()
1782 bdp = rxq->bd.cur; in fec_enet_rx_queue()
1932 rxq->bd.cur = bdp; in fec_enet_rx_queue()
3099 rxq->stats[j] = 0; in fec_enet_clear_ethtool_stats()
3443 bdp = rxq->bd.base; in fec_enet_alloc_rxq_buffers()
3445 err = fec_enet_create_page_pool(fep, rxq, rxq->bd.ring_size); in fec_enet_alloc_rxq_buffers()
[all …]
/drivers/net/ethernet/alacritech/
A Dslicoss.c122 return slic_get_free_queue_descs(rxq->put_idx, rxq->done_idx, rxq->len); in slic_get_free_rx_descs()
396 struct slic_rx_queue *rxq = &sdev->rxq; in slic_refill_rx_queue() local
437 buff = &rxq->rxbuffs[rxq->put_idx]; in slic_refill_rx_queue()
446 rxq->put_idx = slic_next_queue_idx(rxq->put_idx, rxq->len); in slic_refill_rx_queue()
549 struct slic_rx_queue *rxq = &sdev->rxq; in slic_handle_receive() local
559 while (todo && (rxq->done_idx != rxq->put_idx)) { in slic_handle_receive()
560 buff = &rxq->rxbuffs[rxq->done_idx]; in slic_handle_receive()
613 rxq->done_idx = slic_next_queue_idx(rxq->done_idx, rxq->len); in slic_handle_receive()
918 struct slic_rx_queue *rxq = &sdev->rxq; in slic_init_rx_queue() local
923 rxq->put_idx = 0; in slic_init_rx_queue()
[all …]
/drivers/net/ethernet/fungible/funeth/
A Dfuneth_trace.h75 TP_PROTO(const struct funeth_rxq *rxq,
81 TP_ARGS(rxq, num_rqes, pkt_len, hash, cls_vec),
90 __string(devname, rxq->netdev->name)
94 __entry->qidx = rxq->qidx;
95 __entry->cq_head = rxq->cq_head;

Completed in 155 milliseconds

1234567