| /drivers/net/ethernet/marvell/octeon_ep/ |
| A D | octep_pfvf_mbox.c | 49 vf_id, oct->vf_info[vf_id].mbox_version); in octep_pfvf_validate_version() 114 vf_mbox_queue = vf_id * max_rings_per_vf; in octep_send_notification() 162 vf_id); in octep_pfvf_set_mac_addr() 171 vf_id); in octep_pfvf_set_mac_addr() 188 oct->vf_info[vf_id].mac_addr); in octep_pfvf_get_mac_addr() 196 vf_id); in octep_pfvf_get_mac_addr() 212 vf_id); in octep_pfvf_dev_remove() 280 oct->mbox[ring]->vf_id = i; in octep_setup_pfvf_mbox() 318 struct octep_mbox *mbox, int vf_id, in octep_pfvf_pf_get_data() argument 414 int vf_id; in octep_pfvf_mbox_work() local [all …]
|
| /drivers/net/ethernet/huawei/hinic/ |
| A D | hinic_sriov.c | 70 HW_VF_ID_TO_OS(vf_id), err, in hinic_notify_vf_link_status() 103 u8 qos, int vf_id) in hinic_set_vf_vlan() argument 228 HW_VF_ID_TO_OS(vf_id)); in hinic_init_vf_config() 235 vf_info->pf_qos, vf_id); in hinic_init_vf_config() 238 HW_VF_ID_TO_OS(vf_id)); in hinic_init_vf_config() 295 if (vf_id > nic_io->max_vfs) in hinic_unregister_vf_msg_handler() 314 vf_id); in hinic_change_vf_mtu_msg_handler() 356 HW_VF_ID_TO_OS(vf_id)); in hinic_set_vf_mac_msg_handler() 688 vf_id); in hinic_kill_vf_vlan() 694 HW_VF_ID_TO_OS(vf_id)); in hinic_kill_vf_vlan() [all …]
|
| A D | hinic_hw_mbox.h | 67 typedef int (*hinic_pf_mbox_cb)(void *handle, u16 vf_id, u8 cmd, void *buf_in, 126 int (*cmd_msg_handler)(void *hwdev, u16 vf_id, 136 u16 vf_id, u8 cmd, void *buf_in, 167 enum hinic_mod_type mod, u16 vf_id, u8 cmd, void *buf_in,
|
| /drivers/net/ethernet/intel/ice/ |
| A D | ice_sriov.h | 31 int __ice_set_vf_mac(struct ice_pf *pf, u16 vf_id, const u8 *mac); 32 int ice_set_vf_mac(struct net_device *netdev, int vf_id, u8 *mac); 44 ice_set_vf_bw(struct net_device *netdev, int vf_id, int min_tx_rate, 56 ice_get_vf_stats(struct net_device *netdev, int vf_id, 87 u16 __always_unused vf_id, const u8 __always_unused *mac) in __ice_set_vf_mac() argument 94 int __always_unused vf_id, u8 __always_unused *mac) in ice_set_vf_mac() argument 101 int __always_unused vf_id, in ice_get_vf_cfg() argument 109 int __always_unused vf_id, bool __always_unused trusted) in ice_set_vf_trust() argument 116 int __always_unused vf_id, u16 __always_unused vid, in ice_set_vf_port_vlan() argument 124 int __always_unused vf_id, bool __always_unused ena) in ice_set_vf_spoofchk() argument [all …]
|
| A D | ice_sriov.c | 483 vf->vf_id, retval); in ice_start_vfs() 490 vf->vf_id, retval); in ice_start_vfs() 697 for (u16 vf_id = 0; vf_id < num_vfs; vf_id++) { in ice_create_vf_entries() local 706 vf->vf_id = vf_id; in ice_create_vf_entries() 1266 ivi->vf = vf_id; in ice_get_vf_cfg() 1338 vf->vf_id); in __ice_set_vf_mac() 1343 mac, vf_id); in __ice_set_vf_mac() 1512 min_tx_rate, vf->vf_id, in ice_min_tx_rate_oversubscribed() 1569 vf->vf_id); in ice_set_vf_bw() 1580 vf->vf_id); in ice_set_vf_bw() [all …]
|
| A D | ice_vf_lib.c | 31 if (vf->vf_id == vf_id) { in ice_get_vf_by_id() 304 vf->vf_id); in ice_vf_rebuild_vsi() 506 vf->vf_id); in ice_vf_rebuild_host_cfg() 510 vf->vf_id); in ice_vf_rebuild_host_cfg() 514 vf->vf_id); in ice_vf_rebuild_host_cfg() 518 vf->vf_id); in ice_vf_rebuild_host_cfg() 875 vf->vf_id); in ice_reset_vf() 900 vf->vf_id); in ice_reset_vf() 953 vf->vf_id); in ice_reset_vf() 1088 vf->vf_id); in ice_check_vf_init() [all …]
|
| A D | ice_virtchnl_fdir.c | 142 vf->vf_id); in ice_vf_start_ctrl_vsi() 149 vf->vf_id); in ice_vf_start_ctrl_vsi() 666 vf->vf_id); in ice_vc_fdir_write_flow_prof() 673 vf->vf_id); in ice_vc_fdir_write_flow_prof() 760 vf->vf_id); in ice_vc_fdir_config_input_set() 792 vf->vf_id); in ice_vc_fdir_config_input_set() 1202 vf->vf_id); in ice_vc_fdir_parse_action() 1609 vf->vf_id); in ice_vf_verify_rx_desc() 1618 vf->vf_id); in ice_vf_verify_rx_desc() 1893 vf->vf_id); in ice_vc_fdir_set_irq_ctx() [all …]
|
| A D | ice_virtchnl.c | 787 vf->vf_id); in ice_vc_handle_rss_cfg() 794 vf->vf_id); in ice_vc_handle_rss_cfg() 808 vf->vf_id); in ice_vc_handle_rss_cfg() 1268 vf->vf_id); in ice_vc_cfg_promiscuous_mode_msg() 2290 vf->vf_id); in ice_vc_can_add_mac() 2321 vf->vf_id); in ice_vc_add_mac_addr() 2416 vf->vf_id); in ice_vc_del_mac_addr() 2476 vf->vf_id); in ice_vc_handle_mac_addr_msg() 2567 vf->vf_id); in ice_vc_request_qs_msg() 4304 vf->vf_id); in ice_vc_repr_cfg_promiscuous_mode() [all …]
|
| A D | ice_vf_lib.h | 104 u16 vf_id; /* VF ID in the PF space */ member 241 struct ice_vf *ice_get_vf_by_id(struct ice_pf *pf, u16 vf_id); 246 int vf_id = pci_iov_vf_id(vf_dev); in ice_get_vf_by_dev() local 248 if (vf_id < 0) in ice_get_vf_by_dev() 275 static inline struct ice_vf *ice_get_vf_by_id(struct ice_pf *pf, u16 vf_id) in ice_get_vf_by_id() argument
|
| /drivers/vfio/pci/pds/ |
| A D | cmds.c | 94 .vf_id = cpu_to_le16(pds_vfio->vf_id), in pds_vfio_suspend_wait_device_cmd() 134 .vf_id = cpu_to_le16(pds_vfio->vf_id), in pds_vfio_suspend_device_cmd() 168 .vf_id = cpu_to_le16(pds_vfio->vf_id), in pds_vfio_resume_device_cmd() 185 .vf_id = cpu_to_le16(pds_vfio->vf_id), in pds_vfio_get_lm_state_size_cmd() 286 .vf_id = cpu_to_le16(pds_vfio->vf_id), in pds_vfio_get_lm_state_cmd() 324 .vf_id = cpu_to_le16(pds_vfio->vf_id), in pds_vfio_set_lm_state_cmd() 364 .vf_id = cpu_to_le16(pds_vfio->vf_id), in pds_vfio_send_host_vf_lm_status_cmd() 393 .vf_id = cpu_to_le16(pds_vfio->vf_id), in pds_vfio_dirty_status_cmd() 434 .vf_id = cpu_to_le16(pds_vfio->vf_id), in pds_vfio_dirty_enable_cmd() 459 .vf_id = cpu_to_le16(pds_vfio->vf_id), in pds_vfio_dirty_disable_cmd() [all …]
|
| A D | vfio_dev.c | 122 int err, vf_id, pci_id; in pds_vfio_init_device() local 124 vf_id = pci_iov_vf_id(pdev); in pds_vfio_init_device() 125 if (vf_id < 0) in pds_vfio_init_device() 126 return vf_id; in pds_vfio_init_device() 132 pds_vfio->vf_id = vf_id; in pds_vfio_init_device() 143 __func__, pci_dev_id(pci_physfn(pdev)), pci_id, vf_id, in pds_vfio_init_device()
|
| /drivers/virtio/ |
| A D | virtio_pci_admin_legacy_io.c | 43 int vf_id; in virtio_pci_admin_legacy_io_write() local 49 vf_id = pci_iov_vf_id(pdev); in virtio_pci_admin_legacy_io_write() 50 if (vf_id < 0) in virtio_pci_admin_legacy_io_write() 51 return vf_id; in virtio_pci_admin_legacy_io_write() 117 int vf_id; in virtio_pci_admin_legacy_io_read() local 123 vf_id = pci_iov_vf_id(pdev); in virtio_pci_admin_legacy_io_read() 124 if (vf_id < 0) in virtio_pci_admin_legacy_io_read() 125 return vf_id; in virtio_pci_admin_legacy_io_read() 203 int vf_id; in virtio_pci_admin_legacy_io_notify_info() local 210 if (vf_id < 0) in virtio_pci_admin_legacy_io_notify_info() [all …]
|
| A D | virtio_pci_modern.c | 922 int vf_id; in virtio_pci_admin_mode_set() local 929 if (vf_id < 0) in virtio_pci_admin_mode_set() 930 return vf_id; in virtio_pci_admin_mode_set() 972 int vf_id; in virtio_pci_admin_obj_create() local 979 if (vf_id < 0) in virtio_pci_admin_obj_create() 980 return vf_id; in virtio_pci_admin_obj_create() 1044 int vf_id; in virtio_pci_admin_obj_destroy() local 1051 if (vf_id < 0) in virtio_pci_admin_obj_destroy() 1099 int vf_id; in virtio_pci_admin_dev_parts_metadata_get() local 1163 int vf_id; in virtio_pci_admin_dev_parts_get() local [all …]
|
| /drivers/net/ethernet/intel/i40e/ |
| A D | i40e_virtchnl_pf.c | 160 u16 vf_id; in i40e_restore_all_vfs_msi_state() local 246 vf->vf_id); in i40e_vc_reset_vf() 250 vf->vf_id); in i40e_vc_reset_vf() 265 return (vsi && (vsi->vf_id == vf->vf_id)); in i40e_vc_isvalid_vsi_id() 1484 vf->vf_id); in i40e_trigger_vf_reset() 1592 vf->vf_id); in i40e_reset_vf() 1796 for (vf_id = 0; vf_id < tmp; vf_id++) { in i40e_free_vfs() 2775 vf->vf_id, in i40e_vc_request_queues_msg() 4304 int vf_id; in i40e_vc_process_vflr_event() local 4320 for (vf_id = 0; vf_id < pf->num_alloc_vfs; vf_id++) { in i40e_vc_process_vflr_event() [all …]
|
| A D | i40e_virtchnl_pf.h | 78 s16 vf_id; member 128 int i40e_vc_process_vf_msg(struct i40e_pf *pf, s16 vf_id, u32 v_opcode, 137 int i40e_ndo_set_vf_mac(struct net_device *netdev, int vf_id, u8 *mac); 138 int i40e_ndo_set_vf_port_vlan(struct net_device *netdev, int vf_id, 140 int i40e_ndo_set_vf_bw(struct net_device *netdev, int vf_id, int min_tx_rate, 142 int i40e_ndo_set_vf_trust(struct net_device *netdev, int vf_id, bool setting); 144 int vf_id, struct ifla_vf_info *ivi); 145 int i40e_ndo_set_vf_link_state(struct net_device *netdev, int vf_id, int link); 146 int i40e_ndo_set_vf_spoofchk(struct net_device *netdev, int vf_id, bool enable); 153 int i40e_get_vf_stats(struct net_device *netdev, int vf_id,
|
| A D | i40e_client.c | 16 u32 vf_id, u8 *msg, u16 len); 28 bool is_vf, u32 vf_id, 82 i40e_notify_client_of_vf_msg(struct i40e_vsi *vsi, u32 vf_id, u8 *msg, u16 len) in i40e_notify_client_of_vf_msg() argument 99 vf_id, msg, len); in i40e_notify_client_of_vf_msg() 193 void i40e_notify_client_of_vf_reset(struct i40e_pf *pf, u32 vf_id) in i40e_notify_client_of_vf_reset() argument 208 cdev->client->ops->vf_reset(&cdev->lan_info, cdev->client, vf_id); in i40e_notify_client_of_vf_reset() 245 int i40e_vf_client_capable(struct i40e_pf *pf, u32 vf_id) in i40e_vf_client_capable() argument 262 vf_id); in i40e_vf_client_capable() 537 u32 vf_id, u8 *msg, u16 len) in i40e_client_virtchnl_send() argument 543 err = i40e_aq_send_msg_to_vf(hw, vf_id, VIRTCHNL_OP_RDMA, in i40e_client_virtchnl_send() [all …]
|
| /drivers/vdpa/pds/ |
| A D | cmds.c | 23 .vdpa_init.vf_id = cpu_to_le16(pdsv->vdpa_aux->vf_id), in pds_vdpa_init_hw() 45 .vdpa.vf_id = cpu_to_le16(pdsv->vdpa_aux->vf_id), in pds_vdpa_cmd_reset() 65 .vdpa_status.vf_id = cpu_to_le16(pdsv->vdpa_aux->vf_id), in pds_vdpa_cmd_set_status() 86 .vdpa_setattr.vf_id = cpu_to_le16(pdsv->vdpa_aux->vf_id), in pds_vdpa_cmd_set_mac() 109 .vdpa_setattr.vf_id = cpu_to_le16(pdsv->vdpa_aux->vf_id), in pds_vdpa_cmd_set_max_vq_pairs() 133 .vdpa_vq_init.vf_id = cpu_to_le16(pdsv->vdpa_aux->vf_id), in pds_vdpa_cmd_init_vq() 167 .vdpa_vq_reset.vf_id = cpu_to_le16(pdsv->vdpa_aux->vf_id), in pds_vdpa_cmd_reset_vq()
|
| /drivers/scsi/bfa/ |
| A D | bfad_bsg.h | 176 u16 vf_id; member 292 u16 vf_id; member 300 u16 vf_id; member 308 u16 vf_id; member 316 u16 vf_id; member 326 u16 vf_id; member 337 u16 vf_id; member 346 u16 vf_id; member 358 u16 vf_id; member 376 u16 vf_id; member [all …]
|
| /drivers/net/ethernet/broadcom/bnxt/ |
| A D | bnxt_sriov.c | 86 vf = &bp->pf.vf[vf_id]; in bnxt_set_vf_spoofchk() 170 vf = &bp->pf.vf[vf_id]; in bnxt_set_vf_trust() 191 ivi->vf = vf_id; in bnxt_get_vf_config() 192 vf = &bp->pf.vf[vf_id]; in bnxt_get_vf_config() 231 vf = &bp->pf.vf[vf_id]; in bnxt_set_vf_mac() 270 vf = &bp->pf.vf[vf_id]; in bnxt_set_vf_vlan() 304 vf = &bp->pf.vf[vf_id]; in bnxt_set_vf_bw() 308 max_tx_rate, vf_id); in bnxt_set_vf_bw() 345 vf = &bp->pf.vf[vf_id]; in bnxt_set_vf_link_state() 500 vf = &bp->pf.vf[vf_id]; in __bnxt_set_vf_params() [all …]
|
| /drivers/crypto/marvell/octeontx2/ |
| A D | otx2_cptpf_mbox.c | 387 mdev = &mbox->dev[vf->vf_id]; in otx2_cptpf_vfpf_mbox_handler() 397 (vf->vf_id + 1)); in otx2_cptpf_vfpf_mbox_handler() 414 otx2_mbox_msg_send(mbox, vf->vf_id); in otx2_cptpf_vfpf_mbox_handler() 515 int vf_id, int size) in forward_to_vf() argument 532 vf_id--; in forward_to_vf() 533 if (vf_id >= cptpf->enabled_vfs) { in forward_to_vf() 536 vf_id, cptpf->enabled_vfs); in forward_to_vf() 545 "Forwarding to VF%d failed.\n", vf_id); in forward_to_vf() 565 int offset, vf_id, i; in otx2_cptpf_afpf_mbox_handler() local 581 if (vf_id > 0) in otx2_cptpf_afpf_mbox_handler() [all …]
|
| /drivers/net/ethernet/cavium/thunder/ |
| A D | nic.h | 285 u8 vf_id; member 427 u8 vf_id; member 463 u8 vf_id; member 470 u8 vf_id; member 477 u8 vf_id; member 485 u8 vf_id; member 492 u8 vf_id; member 502 u8 vf_id; member 520 u8 vf_id; member 526 u8 vf_id; member [all …]
|
| A D | nic_main.c | 165 mbx.nic_cfg.vf_id = vf; in nic_mbx_send_ready() 240 mbx.bgx_stats.vf_id = bgx->vf_id; in nic_get_bgx_stats() 480 vnic = cfg->vf_id; in nic_config_cpi() 504 qset = cfg->vf_id; in nic_config_cpi() 545 nic->cpi_base[cfg->vf_id] = cpi_base; in nic_config_cpi() 581 qset = cfg->vf_id; in nic_config_rss() 749 nic->pqs_vf[sqs_id] = sqs->vf_id; in nic_alloc_sqs() 755 mbx.sqs_alloc.vf_id = sqs->vf_id; in nic_alloc_sqs() 764 if (lbk->vf_id >= nic->num_vf_en) in nic_config_loopback() 1033 lmac = mbx.mac.vf_id; in nic_handle_mbx_intr() [all …]
|
| /drivers/net/ethernet/amd/pds_core/ |
| A D | devlink.c | 46 int vf_id; in pdsc_dl_enable_set() local 56 for (vf_id = 0; vf_id < pdsc->num_vfs; vf_id++) { in pdsc_dl_enable_set() 57 struct pdsc *vf = pdsc->vfs[vf_id].vf; in pdsc_dl_enable_set() 61 &pdsc->vfs[vf_id].padev); in pdsc_dl_enable_set() 63 pdsc_auxbus_dev_del(vf, pdsc, &pdsc->vfs[vf_id].padev); in pdsc_dl_enable_set()
|
| /drivers/net/ethernet/mellanox/mlx5/core/ |
| A D | sriov.c | 337 int vf_id, in mlx5_sriov_blocking_notifier_unregister() argument 344 if (WARN_ON(vf_id < 0 || vf_id >= sriov->num_vfs)) in mlx5_sriov_blocking_notifier_unregister() 347 vfs_ctx = &sriov->vfs_ctx[vf_id]; in mlx5_sriov_blocking_notifier_unregister() 363 int vf_id, in mlx5_sriov_blocking_notifier_register() argument 370 if (vf_id < 0 || vf_id >= sriov->num_vfs) in mlx5_sriov_blocking_notifier_register() 373 vfs_ctx = &sriov->vfs_ctx[vf_id]; in mlx5_sriov_blocking_notifier_register()
|
| /drivers/crypto/intel/qat/qat_common/ |
| A D | qat_mig_dev.c | 11 struct qat_mig_dev *qat_vfmig_create(struct pci_dev *pdev, int vf_id) in qat_vfmig_create() argument 31 mdev->vf_id = vf_id; in qat_vfmig_create()
|