| /drivers/net/ethernet/fungible/funeth/ |
| A D | funeth_main.c | 445 struct funeth_rxq **rxqs = qset->rxqs; in fun_free_rings() local 451 rxqs = rtnl_dereference(fp->rxqs); in fun_free_rings() 461 if (rxqs == rtnl_dereference(fp->rxqs)) { in fun_free_rings() 475 qset->rxqs = rxqs; in fun_free_rings() 513 qset->rxqs = rxqs; in fun_alloc_rings() 797 rcu_assign_pointer(fp->rxqs, qset->rxqs); in fun_up() 886 rxqs = rcu_dereference(fp->rxqs); in fun_get_stats64() 1089 rxqs = rtnl_dereference(fp->rxqs); in fun_enter_xdp() 1123 rxqs = rtnl_dereference(fp->rxqs); in fun_end_xdp() 1629 .rxqs = rtnl_dereference(fp->rxqs), in fun_change_num_queues() [all …]
|
| A D | funeth_ethtool.c | 486 struct funeth_rxq **rxqs; in fun_set_coalesce() local 510 rxqs = rtnl_dereference(fp->rxqs); in fun_set_coalesce() 511 if (!rxqs) in fun_set_coalesce() 515 WRITE_ONCE(rxqs[i]->irq_db_val, db_val); in fun_set_coalesce() 754 struct funeth_rxq **rxqs; in fun_get_ethtool_stats() local 761 rxqs = rtnl_dereference(fp->rxqs); in fun_get_ethtool_stats() 762 if (!rxqs) in fun_get_ethtool_stats() 816 FUN_QSTAT_READ(rxqs[i], start, rxs); in fun_get_ethtool_stats()
|
| A D | funeth.h | 66 struct funeth_rxq **rxqs; member 87 struct funeth_rxq * __rcu *rxqs; member
|
| /drivers/net/ethernet/huawei/hinic/ |
| A D | hinic_main.c | 231 if (nic_dev->rxqs) in create_rxqs() 236 if (!nic_dev->rxqs) in create_rxqs() 262 hinic_clean_rxq(&nic_dev->rxqs[i]); in create_rxqs() 266 hinic_clean_rxq(&nic_dev->rxqs[j]); in create_rxqs() 271 devm_kfree(&netdev->dev, nic_dev->rxqs); in create_rxqs() 272 nic_dev->rxqs = NULL; in create_rxqs() 285 if (!nic_dev->rxqs) in free_rxqs() 290 hinic_clean_rxq(&nic_dev->rxqs[i]); in free_rxqs() 295 devm_kfree(&netdev->dev, nic_dev->rxqs); in free_rxqs() 296 nic_dev->rxqs = NULL; in free_rxqs() [all …]
|
| A D | hinic_dev.h | 98 struct hinic_rxq *rxqs; member
|
| A D | hinic_debugfs.c | 240 rq = dev->rxqs[rq_id].rq; in hinic_rq_debug_add()
|
| A D | hinic_ethtool.c | 700 msix_idx = set_rx_coal ? nic_dev->rxqs[q_id].rq->msix_entry : in set_queue_coalesce() 1372 if (!nic_dev->rxqs) in get_drv_queue_stats() 1375 hinic_rxq_get_stats(&nic_dev->rxqs[qid], &rxq_stats); in get_drv_queue_stats()
|
| /drivers/net/ethernet/microsoft/mana/ |
| A D | mana_bpf.c | 145 return rtnl_dereference(apc->rxqs[0]->bpf_prog); in mana_chn_xdp_get() 164 rcu_assign_pointer(apc->rxqs[i]->bpf_prog, prog); in mana_chn_setxdp()
|
| A D | mana_en.c | 489 rx_stats = &apc->rxqs[q]->stats; in mana_get_stats64() 829 kfree(apc->rxqs); in mana_cleanup_port_context() 830 apc->rxqs = NULL; in mana_cleanup_port_context() 845 return !apc->rxqs ? -ENOMEM : 0; in mana_init_port_context() 1556 rxq = apc->rxqs[rxq_idx]; in mana_fence_rqs() 2585 rxq = apc->rxqs[idx]; in mana_create_rxq_debugfs() 2620 apc->rxqs[i] = rxq; in mana_add_rx_queues() 2625 apc->default_rxobj = apc->rxqs[0]->rxobj; in mana_add_rx_queues() 2637 rxq = apc->rxqs[rxq_idx]; in mana_destroy_vport() 2642 apc->rxqs[rxq_idx] = NULL; in mana_destroy_vport() [all …]
|
| A D | mana_ethtool.c | 226 rx_stats = &apc->rxqs[q]->stats; in mana_get_ethtool_stats()
|
| /drivers/net/ethernet/huawei/hinic3/ |
| A D | hinic3_nic_dev.h | 68 struct hinic3_rxq *rxqs; member
|
| /drivers/net/ethernet/netronome/nfp/ |
| A D | nfp_net_repr.h | 102 nfp_repr_alloc_mqs(struct nfp_app *app, unsigned int txqs, unsigned int rxqs);
|
| A D | nfp_net_repr.c | 424 nfp_repr_alloc_mqs(struct nfp_app *app, unsigned int txqs, unsigned int rxqs) in nfp_repr_alloc_mqs() argument 429 netdev = alloc_etherdev_mqs(sizeof(*repr), txqs, rxqs); in nfp_repr_alloc_mqs()
|
| /drivers/net/can/dev/ |
| A D | dev.c | 254 unsigned int txqs, unsigned int rxqs) in alloc_candev_mqs() argument 281 txqs, rxqs); in alloc_candev_mqs()
|
| /drivers/net/ethernet/marvell/ |
| A D | mvneta.c | 510 struct mvneta_rx_queue *rxqs; member 1269 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; in mvneta_port_up() 3318 &pp->rxqs[rx_queue]); in mvneta_poll() 3321 &pp->rxqs[rx_queue]); in mvneta_poll() 3659 mvneta_rxq_deinit(pp, &pp->rxqs[queue]); in mvneta_cleanup_rxqs() 4730 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; in mvneta_ethtool_set_coalesce() 4967 if (pp->rxqs[i].page_pool) in mvneta_ethtool_pp_stats() 5387 pp->rxqs = devm_kcalloc(dev, rxq_number, sizeof(*pp->rxqs), GFP_KERNEL); in mvneta_init() 5388 if (!pp->rxqs) in mvneta_init() 5393 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; in mvneta_init() [all …]
|
| /drivers/net/ethernet/intel/idpf/ |
| A D | idpf_txrx.c | 540 idpf_rx_desc_rel(rx_qgrp->singleq.rxqs[j], dev, in idpf_rx_desc_rel_all() 826 q = rx_qgrp->singleq.rxqs[j]; in idpf_rx_bufs_init_all() 941 q = rx_qgrp->singleq.rxqs[j]; in idpf_rx_desc_alloc_all() 1064 kfree(rx_qgrp->singleq.rxqs[j]); in idpf_rxq_group_rel() 1065 rx_qgrp->singleq.rxqs[j] = NULL; in idpf_rxq_group_rel() 1472 rx_qgrp->singleq.rxqs[j] = in idpf_rxq_group_alloc() 1473 kzalloc(sizeof(*rx_qgrp->singleq.rxqs[j]), in idpf_rxq_group_alloc() 1475 if (!rx_qgrp->singleq.rxqs[j]) { in idpf_rxq_group_alloc() 1544 q = rx_qgrp->singleq.rxqs[j]; in idpf_rxq_group_alloc() 4233 q = rx_qgrp->singleq.rxqs[j]; in idpf_vport_intr_map_vector_to_qs()
|
| A D | idpf_ptp.c | 345 rxq = grp->singleq.rxqs[i]; in idpf_ptp_update_phctime_rxq_grp() 704 rx_queue = grp->singleq.rxqs[j]; in idpf_ptp_set_rx_tstamp()
|
| A D | idpf_ethtool.c | 1091 rxq = rxq_grp->singleq.rxqs[j]; in idpf_collect_queue_stats() 1225 rxq = rxq_grp->singleq.rxqs[j]; in idpf_get_ethtool_stats() 1254 return vport->rxq_grps->singleq.rxqs[q_num]->q_vector; in idpf_find_rxq_vec()
|
| A D | idpf_virtchnl.c | 1253 q = rx_qgrp->singleq.rxqs[j]; in __idpf_queue_reg_init() 1743 rxq = rx_qgrp->singleq.rxqs[j]; in idpf_send_config_rx_queues_msg() 1904 cpu_to_le32(rx_qgrp->singleq.rxqs[j]->q_id); in idpf_send_ena_dis_queues_msg() 2046 rxq = rx_qgrp->singleq.rxqs[j]; in idpf_send_map_unmap_queue_vector_msg() 3501 q = rx_qgrp->singleq.rxqs[j]; in __idpf_vport_queue_ids_init()
|
| A D | idpf_txrx.h | 891 struct idpf_rx_queue *rxqs[IDPF_LARGE_MAX_Q]; member
|
| /drivers/infiniband/hw/hfi1/ |
| A D | ipoib_main.c | 245 params->rxqs = dd->num_netdev_contexts; in hfi1_ipoib_rn_get_params()
|
| /drivers/net/ethernet/marvell/mvpp2/ |
| A D | mvpp2_main.c | 727 prxq = port->rxqs[lrxq]->id; in mvpp2_rxq_long_pool_set() 748 prxq = port->rxqs[lrxq]->id; in mvpp2_rxq_short_pool_set() 2305 queue = port->rxqs[lrxq]->id; in mvpp2_defaults_set() 2323 queue = port->rxqs[lrxq]->id; in mvpp2_ingress_enable() 2336 queue = port->rxqs[lrxq]->id; in mvpp2_ingress_disable() 2846 return port->rxqs[queue]; in mvpp2_get_rx_queue() 3302 mvpp2_rxq_deinit(port, port->rxqs[queue]); in mvpp2_cleanup_rxqs() 3314 err = mvpp2_rxq_init(port, port->rxqs[queue]); in mvpp2_setup_rxqs() 6047 port->rxqs = devm_kcalloc(dev, port->nrxqs, sizeof(*port->rxqs), in mvpp2_port_init() 6049 if (!port->rxqs) { in mvpp2_port_init() [all …]
|
| A D | mvpp2.h | 1212 struct mvpp2_rx_queue **rxqs; member
|
| /drivers/net/ethernet/mellanox/mlx5/core/ipoib/ |
| A D | ipoib.c | 847 .rxqs = nch, in mlx5_rdma_rn_get_params()
|
| /drivers/net/ethernet/ibm/ |
| A D | ibmvnic.c | 195 struct ibmvnic_sub_crq_queue **rxqs; in ibmvnic_clean_affinity() local 200 rxqs = adapter->rx_scrq; in ibmvnic_clean_affinity() 210 if (rxqs) { in ibmvnic_clean_affinity() 212 ibmvnic_clean_queue_affinity(adapter, rxqs[i]); in ibmvnic_clean_affinity() 259 struct ibmvnic_sub_crq_queue **rxqs = adapter->rx_scrq; in ibmvnic_set_affinity() local 290 queue = rxqs[i_rxqs++]; in ibmvnic_set_affinity()
|