/linux-6.3-rc2/drivers/net/ethernet/sun/ |
A D | sunvnet.c | 127 port->q_index, port->switch_port ? "s" : "q", in vnet_get_strings() 131 port->q_index); in vnet_get_strings() 134 port->q_index); in vnet_get_strings() 137 port->q_index); in vnet_get_strings() 140 port->q_index); in vnet_get_strings() 143 port->q_index); in vnet_get_strings() 146 port->q_index); in vnet_get_strings() 181 data[i++] = port->q_index; in vnet_get_ethtool_stats() 245 return port->q_index; in vnet_select_queue()
|
A D | sunvnet_common.h | 90 u16 q_index; member
|
A D | sunvnet_common.c | 726 txq = netdev_get_tx_queue(dev, port->q_index); in vnet_ack() 764 port->q_index); in maybe_tx_wakeup() 1259 txq = netdev_get_tx_queue(dev, port->q_index); in vnet_handle_offloads() 1803 port->q_index = smallest; in sunvnet_port_add_txq_common() 1810 port->vp->q_used[port->q_index]--; in sunvnet_port_rm_txq_common() 1811 port->q_index = 0; in sunvnet_port_rm_txq_common()
|
A D | ldmvsw.c | 111 return port->q_index; in vsw_select_queue()
|
/linux-6.3-rc2/drivers/net/ethernet/intel/ice/ |
A D | ice_base.c | 217 return ring->q_index - ring->ch->base_q; in ice_calc_txq_handle() 223 return ring->q_index - vsi->tc_cfg.tc_info[tc].qoffset; in ice_calc_txq_handle() 267 ring->q_index); in ice_cfg_xps_tx_ring() 380 pf_q = vsi->rxq_map[ring->q_index]; in ice_setup_rx_ctx() 504 ring->q_index, in ice_vsi_cfg_rxq() 522 ring->q_index); in ice_vsi_cfg_rxq() 528 ring->q_index, in ice_vsi_cfg_rxq() 545 ring->q_index, err); in ice_vsi_cfg_rxq() 554 num_bufs, ring->q_index); in ice_vsi_cfg_rxq() 562 u16 pf_q = ring->vsi->rxq_map[ring->q_index]; in ice_vsi_cfg_rxq() [all …]
|
A D | ice_arfs.c | 123 e->flow_id, e->fltr_info.q_index); in ice_arfs_del_flow_rules() 164 ep->arfs_entry->fltr_info.q_index); in ice_arfs_add_flow_rules() 184 if (rps_may_expire_flow(vsi->netdev, arfs_entry->fltr_info.q_index, in ice_arfs_is_flow_expired() 314 fltr_info->q_index = rxq_idx; in ice_arfs_build_entry() 452 if (fltr_info->q_index == rxq_idx || in ice_rx_flow_steer() 457 fltr_info->q_index = rxq_idx; in ice_rx_flow_steer()
|
A D | ice_trace.h | 75 __entry->q_vector->rx.rx_ring->q_index, 102 __entry->q_vector->tx.tx_ring->q_index,
|
A D | ice_txrx.h | 311 u16 q_index; /* Queue number of ring */ member 369 u16 q_index; /* Queue number of ring */ member
|
A D | ice_ethtool_fdir.c | 1330 static void ice_update_per_q_fltr(struct ice_vsi *vsi, u32 q_index, bool inc) in ice_update_per_q_fltr() argument 1334 if (!vsi->num_rxq || q_index >= vsi->num_rxq) in ice_update_per_q_fltr() 1337 rx_ring = vsi->rx_rings[q_index]; in ice_update_per_q_fltr() 1697 u16 dest_vsi, q_index = 0; in ice_set_fdir_input_set() local 1728 q_index = ring; in ice_set_fdir_input_set() 1732 input->q_index = q_index; in ice_set_fdir_input_set()
|
A D | ice_fdir.h | 184 u16 q_index; member
|
A D | ice_fdir.c | 689 fdir_fltr_ctx.qindex = input->q_index; in ice_fdir_get_prgm_desc() 1294 rule->q_index != input->q_index) in ice_fdir_is_dup_fltr()
|
A D | ice_txrx.c | 144 return netdev_get_tx_queue(ring->netdev, ring->q_index); in txring_txq() 519 rx_ring->q_index, rx_ring->q_vector->napi.napi_id)) in ice_setup_rx_ring() 1009 skb_record_rx_queue(skb, rx_ring->q_index); in ice_build_skb() 1061 skb_record_rx_queue(skb, rx_ring->q_index); in ice_construct_skb()
|
A D | ice_eswitch.c | 205 tx_ring->q_index = 0; in ice_eswitch_remap_rings_to_vectors()
|
/linux-6.3-rc2/net/core/ |
A D | netpoll.c | 110 unsigned int q_index; in queue_process() local 119 q_index = skb_get_queue_mapping(skb); in queue_process() 120 if (unlikely(q_index >= dev->real_num_tx_queues)) { in queue_process() 121 q_index = q_index % dev->real_num_tx_queues; in queue_process() 122 skb_set_queue_mapping(skb, q_index); in queue_process() 124 txq = netdev_get_tx_queue(dev, q_index); in queue_process()
|
/linux-6.3-rc2/drivers/net/ethernet/cavium/liquidio/ |
A D | octeon_iq.h | 173 int q_index; member 394 int q_index, union oct_txpciq iq_no, u32 num_descs,
|
A D | request_manager.c | 201 int q_index, in octeon_setup_iq() argument 225 oct->instr_queue[iq_no]->q_index = q_index; in octeon_setup_iq()
|
/linux-6.3-rc2/sound/soc/intel/skylake/ |
A D | skl-messages.c | 986 static void skl_free_queue(struct skl_module_pin *mpin, int q_index) in skl_free_queue() argument 988 if (mpin[q_index].is_dynamic) { in skl_free_queue() 989 mpin[q_index].in_use = false; in skl_free_queue() 990 mpin[q_index].id.module_id = 0; in skl_free_queue() 991 mpin[q_index].id.instance_id = 0; in skl_free_queue() 992 mpin[q_index].id.pvt_id = 0; in skl_free_queue() 994 mpin[q_index].pin_state = SKL_PIN_UNBIND; in skl_free_queue() 995 mpin[q_index].tgt_mcfg = NULL; in skl_free_queue()
|
/linux-6.3-rc2/drivers/net/ethernet/intel/iavf/ |
A D | iavf_fdir.h | 108 u32 q_index; member
|
A D | iavf_ethtool.c | 1187 fsp->ring_cookie = rule->q_index; in iavf_get_ethtool_fdir_entry() 1244 u32 flow_type, q_index = 0; in iavf_add_fdir_fltr_info() local 1251 q_index = fsp->ring_cookie; in iavf_add_fdir_fltr_info() 1252 if (q_index >= adapter->num_active_queues) in iavf_add_fdir_fltr_info() 1260 fltr->q_index = q_index; in iavf_add_fdir_fltr_info()
|
A D | iavf_fdir.c | 597 vc_msg->rule_cfg.action_set.actions[0].act_conf.queue.index = fltr->q_index; in iavf_fill_fdir_add_msg()
|
/linux-6.3-rc2/drivers/gpu/drm/amd/amdkfd/ |
A D | kfd_process_queue_manager.c | 696 unsigned int *q_index, in criu_checkpoint_queues_device() argument 756 *q_index = *q_index + 1; in criu_checkpoint_queues_device() 768 int ret = 0, pdd_index, q_index = 0; in kfd_criu_checkpoint_queues() local 777 ret = criu_checkpoint_queues_device(pdd, user_priv_data, &q_index, in kfd_criu_checkpoint_queues()
|
/linux-6.3-rc2/drivers/net/ethernet/freescale/dpaa2/ |
A D | dpni.h | 954 u8 q_index, 963 u8 q_index,
|
/linux-6.3-rc2/drivers/scsi/ibmvscsi/ |
A D | ibmvfc.c | 2433 int wait, i, q_index, q_size; in ibmvfc_wait_for_ops() local 2450 for (q_index = 0; q_index < q_size; q_index++) { in ibmvfc_wait_for_ops() 2451 spin_lock(&queues[q_index].l_lock); in ibmvfc_wait_for_ops() 2452 for (i = 0; i < queues[q_index].evt_pool.size; i++) { in ibmvfc_wait_for_ops() 2453 evt = &queues[q_index].evt_pool.events[i]; in ibmvfc_wait_for_ops() 2461 spin_unlock(&queues[q_index].l_lock); in ibmvfc_wait_for_ops() 2471 for (q_index = 0; q_index < q_size; q_index++) { in ibmvfc_wait_for_ops() 2472 spin_lock(&queues[q_index].l_lock); in ibmvfc_wait_for_ops() 2473 for (i = 0; i < queues[q_index].evt_pool.size; i++) { in ibmvfc_wait_for_ops() 2474 evt = &queues[q_index].evt_pool.events[i]; in ibmvfc_wait_for_ops() [all …]
|
/linux-6.3-rc2/drivers/scsi/pm8001/ |
A D | pm80xx_hwi.c | 4269 u32 q_index; in pm80xx_chip_ssp_io_req() local 4290 q_index = pm80xx_chip_get_q_index(task); in pm80xx_chip_ssp_io_req() 4361 task->ssp_task.cmd->cmnd[0], q_index); in pm80xx_chip_ssp_io_req() 4408 return pm8001_mpi_build_cmd(pm8001_ha, q_index, opc, &ssp_cmd, in pm80xx_chip_ssp_io_req() 4409 sizeof(ssp_cmd), q_index); in pm80xx_chip_ssp_io_req() 4419 u32 tag = ccb->ccb_tag, q_index; in pm80xx_chip_sata_req() local 4429 q_index = pm80xx_chip_get_q_index(task); in pm80xx_chip_sata_req() 4535 sata_cmd.sata_fis.command, q_index); in pm80xx_chip_sata_req() 4609 return pm8001_mpi_build_cmd(pm8001_ha, q_index, opc, &sata_cmd, in pm80xx_chip_sata_req() 4610 sizeof(sata_cmd), q_index); in pm80xx_chip_sata_req()
|
/linux-6.3-rc2/drivers/net/ethernet/intel/i40e/ |
A D | i40e_ethtool.c | 3477 fsp->ring_cookie = rule->q_index; in i40e_get_ethtool_fdir_entry() 4794 u16 dest_vsi = 0, q_index = 0; in i40e_add_fdir_ethtool() local 4859 q_index = ring; in i40e_add_fdir_ethtool() 4868 input->q_index = q_index; in i40e_add_fdir_ethtool() 5055 if (rule->dest_ctl != drop && count <= rule->q_index) { in i40e_set_channels() 5058 rule->fd_id, rule->q_index); in i40e_set_channels()
|