Home
last modified time | relevance | path

Searched refs:rxqs (Results 1 – 25 of 34) sorted by relevance

12

/linux/drivers/net/ethernet/fungible/funeth/
A Dfuneth_main.c445 struct funeth_rxq **rxqs = qset->rxqs; in fun_free_rings() local
451 rxqs = rtnl_dereference(fp->rxqs); in fun_free_rings()
461 if (rxqs == rtnl_dereference(fp->rxqs)) { in fun_free_rings()
475 qset->rxqs = rxqs; in fun_free_rings()
513 qset->rxqs = rxqs; in fun_alloc_rings()
797 rcu_assign_pointer(fp->rxqs, qset->rxqs); in fun_up()
886 rxqs = rcu_dereference(fp->rxqs); in fun_get_stats64()
1089 rxqs = rtnl_dereference(fp->rxqs); in fun_enter_xdp()
1123 rxqs = rtnl_dereference(fp->rxqs); in fun_end_xdp()
1629 .rxqs = rtnl_dereference(fp->rxqs), in fun_change_num_queues()
[all …]
A Dfuneth.h66 struct funeth_rxq **rxqs; member
87 struct funeth_rxq * __rcu *rxqs; member
A Dfuneth_ethtool.c486 struct funeth_rxq **rxqs; in fun_set_coalesce() local
510 rxqs = rtnl_dereference(fp->rxqs); in fun_set_coalesce()
511 if (!rxqs) in fun_set_coalesce()
515 WRITE_ONCE(rxqs[i]->irq_db_val, db_val); in fun_set_coalesce()
754 struct funeth_rxq **rxqs; in fun_get_ethtool_stats() local
761 rxqs = rtnl_dereference(fp->rxqs); in fun_get_ethtool_stats()
762 if (!rxqs) in fun_get_ethtool_stats()
816 FUN_QSTAT_READ(rxqs[i], start, rxs); in fun_get_ethtool_stats()
/linux/drivers/net/ethernet/huawei/hinic/
A Dhinic_main.c230 if (nic_dev->rxqs) in create_rxqs()
235 if (!nic_dev->rxqs) in create_rxqs()
261 hinic_clean_rxq(&nic_dev->rxqs[i]); in create_rxqs()
264 hinic_rq_debug_rem(nic_dev->rxqs[j].rq); in create_rxqs()
265 hinic_clean_rxq(&nic_dev->rxqs[j]); in create_rxqs()
270 devm_kfree(&netdev->dev, nic_dev->rxqs); in create_rxqs()
283 if (!nic_dev->rxqs) in free_rxqs()
288 hinic_clean_rxq(&nic_dev->rxqs[i]); in free_rxqs()
293 devm_kfree(&netdev->dev, nic_dev->rxqs); in free_rxqs()
294 nic_dev->rxqs = NULL; in free_rxqs()
[all …]
A Dhinic_dev.h98 struct hinic_rxq *rxqs; member
A Dhinic_debugfs.c240 rq = dev->rxqs[rq_id].rq; in hinic_rq_debug_add()
A Dhinic_ethtool.c700 msix_idx = set_rx_coal ? nic_dev->rxqs[q_id].rq->msix_entry : in set_queue_coalesce()
1389 if (!nic_dev->rxqs) in get_drv_queue_stats()
1392 hinic_rxq_get_stats(&nic_dev->rxqs[qid], &rxq_stats); in get_drv_queue_stats()
/linux/net/
A Ddevres.c22 unsigned int txqs, unsigned int rxqs) in devm_alloc_etherdev_mqs() argument
30 dr->ndev = alloc_etherdev_mqs(sizeof_priv, txqs, rxqs); in devm_alloc_etherdev_mqs()
/linux/drivers/net/ethernet/microsoft/mana/
A Dmana_bpf.c145 return rtnl_dereference(apc->rxqs[0]->bpf_prog); in mana_chn_xdp_get()
164 rcu_assign_pointer(apc->rxqs[i]->bpf_prog, prog); in mana_chn_setxdp()
A Dmana_en.c450 rx_stats = &apc->rxqs[q]->stats; in mana_get_stats64()
724 kfree(apc->rxqs); in mana_cleanup_port_context()
725 apc->rxqs = NULL; in mana_cleanup_port_context()
737 apc->rxqs = kcalloc(apc->num_queues, sizeof(struct mana_rxq *), in mana_init_port_context()
740 return !apc->rxqs ? -ENOMEM : 0; in mana_init_port_context()
1322 rxq = apc->rxqs[rxq_idx]; in mana_fence_rqs()
2328 apc->rxqs[i] = rxq; in mana_add_rx_queues()
2331 apc->default_rxobj = apc->rxqs[0]->rxobj; in mana_add_rx_queues()
2343 rxq = apc->rxqs[rxq_idx]; in mana_destroy_vport()
2348 apc->rxqs[rxq_idx] = NULL; in mana_destroy_vport()
[all …]
A Dmana_ethtool.c177 rx_stats = &apc->rxqs[q]->stats; in mana_get_ethtool_stats()
/linux/drivers/net/ethernet/netronome/nfp/
A Dnfp_net_repr.h102 nfp_repr_alloc_mqs(struct nfp_app *app, unsigned int txqs, unsigned int rxqs);
A Dnfp_net_repr.c424 nfp_repr_alloc_mqs(struct nfp_app *app, unsigned int txqs, unsigned int rxqs) in nfp_repr_alloc_mqs() argument
429 netdev = alloc_etherdev_mqs(sizeof(*repr), txqs, rxqs); in nfp_repr_alloc_mqs()
/linux/include/linux/
A Detherdevice.h57 unsigned int rxqs);
63 unsigned int rxqs);
/linux/net/ethernet/
A Deth.c380 unsigned int rxqs) in alloc_etherdev_mqs() argument
383 ether_setup, txqs, rxqs); in alloc_etherdev_mqs()
/linux/drivers/net/can/dev/
A Ddev.c250 unsigned int txqs, unsigned int rxqs) in alloc_candev_mqs() argument
277 txqs, rxqs); in alloc_candev_mqs()
/linux/include/linux/can/
A Ddev.h174 unsigned int txqs, unsigned int rxqs);
/linux/drivers/net/ethernet/marvell/
A Dmvneta.c506 struct mvneta_rx_queue *rxqs; member
1269 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; in mvneta_port_up()
3315 &pp->rxqs[rx_queue]); in mvneta_poll()
3318 &pp->rxqs[rx_queue]); in mvneta_poll()
3656 mvneta_rxq_deinit(pp, &pp->rxqs[queue]); in mvneta_cleanup_rxqs()
4684 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; in mvneta_ethtool_set_coalesce()
4923 if (pp->rxqs[i].page_pool) in mvneta_ethtool_pp_stats()
5364 pp->rxqs = devm_kcalloc(dev, rxq_number, sizeof(*pp->rxqs), GFP_KERNEL); in mvneta_init()
5365 if (!pp->rxqs) in mvneta_init()
5370 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; in mvneta_init()
[all …]
/linux/drivers/infiniband/hw/hfi1/
A Dipoib_main.c245 params->rxqs = dd->num_netdev_contexts; in hfi1_ipoib_rn_get_params()
/linux/drivers/net/ethernet/intel/idpf/
A Didpf_ethtool.c803 rxq = rxq_grp->singleq.rxqs[j]; in idpf_collect_queue_stats()
937 rxq = rxq_grp->singleq.rxqs[j]; in idpf_get_ethtool_stats()
966 return vport->rxq_grps->singleq.rxqs[q_num]->q_vector; in idpf_find_rxq_vec()
A Didpf_txrx.c539 idpf_rx_desc_rel(rx_qgrp->singleq.rxqs[j], dev, in idpf_rx_desc_rel_all()
825 q = rx_qgrp->singleq.rxqs[j]; in idpf_rx_bufs_init_all()
940 q = rx_qgrp->singleq.rxqs[j]; in idpf_rx_desc_alloc_all()
1063 kfree(rx_qgrp->singleq.rxqs[j]); in idpf_rxq_group_rel()
1064 rx_qgrp->singleq.rxqs[j] = NULL; in idpf_rxq_group_rel()
1463 rx_qgrp->singleq.rxqs[j] = in idpf_rxq_group_alloc()
1464 kzalloc(sizeof(*rx_qgrp->singleq.rxqs[j]), in idpf_rxq_group_alloc()
1466 if (!rx_qgrp->singleq.rxqs[j]) { in idpf_rxq_group_alloc()
1535 q = rx_qgrp->singleq.rxqs[j]; in idpf_rxq_group_alloc()
4079 q = rx_qgrp->singleq.rxqs[j]; in idpf_vport_intr_map_vector_to_qs()
A Didpf_virtchnl.c1117 q = rx_qgrp->singleq.rxqs[j]; in __idpf_queue_reg_init()
1607 rxq = rx_qgrp->singleq.rxqs[j]; in idpf_send_config_rx_queues_msg()
1768 cpu_to_le32(rx_qgrp->singleq.rxqs[j]->q_id); in idpf_send_ena_dis_queues_msg()
1910 rxq = rx_qgrp->singleq.rxqs[j]; in idpf_send_map_unmap_queue_vector_msg()
3312 q = rx_qgrp->singleq.rxqs[j]; in __idpf_vport_queue_ids_init()
A Didpf_txrx.h892 struct idpf_rx_queue *rxqs[IDPF_LARGE_MAX_Q]; member
/linux/drivers/net/ethernet/marvell/mvpp2/
A Dmvpp2_main.c727 prxq = port->rxqs[lrxq]->id; in mvpp2_rxq_long_pool_set()
748 prxq = port->rxqs[lrxq]->id; in mvpp2_rxq_short_pool_set()
2318 queue = port->rxqs[lrxq]->id; in mvpp2_defaults_set()
2336 queue = port->rxqs[lrxq]->id; in mvpp2_ingress_enable()
2349 queue = port->rxqs[lrxq]->id; in mvpp2_ingress_disable()
2859 return port->rxqs[queue]; in mvpp2_get_rx_queue()
3315 mvpp2_rxq_deinit(port, port->rxqs[queue]); in mvpp2_cleanup_rxqs()
3327 err = mvpp2_rxq_init(port, port->rxqs[queue]); in mvpp2_setup_rxqs()
6025 port->rxqs = devm_kcalloc(dev, port->nrxqs, sizeof(*port->rxqs), in mvpp2_port_init()
6027 if (!port->rxqs) { in mvpp2_port_init()
[all …]
/linux/include/net/mana/
A Dmana.h439 struct mana_rxq **rxqs; member

Completed in 111 milliseconds

12