Searched refs:mvdev (Results 1 – 9 of 9) sorted by relevance
| /drivers/vdpa/mlx5/net/ |
| A D | mlx5_vnet.c | 283 struct mlx5_vdpa_dev *mvdev = &ndev->mvdev; in create_tis() local 884 struct mlx5_vdpa_dev *mvdev = &ndev->mvdev; in create_virtqueue() local 1226 struct mlx5_vdpa_dev *mvdev = &ndev->mvdev; in query_virtqueues() local 1311 struct mlx5_vdpa_dev *mvdev = &ndev->mvdev; in fill_modify_virtqueue_cmd() local 2368 mvdev = wqent->mvdev; in mlx5_cvq_kick_handler() 2438 if (!mvdev->wq || !mvdev->cvq.ready) in mlx5_vdpa_kick_vq() 2775 mvdev->max_idx = mvdev->max_vqs; in update_cvq_info() 2823 mvdev = wqent->mvdev; in update_carrier() 2844 wqent->mvdev = &ndev->mvdev; in queue_link_work() 3887 mvdev = &ndev->mvdev; in mlx5_vdpa_dev_add() [all …]
|
| A D | debug.c | 102 node->ucast_counter.mdev = ndev->mvdev.mdev; in mlx5_vdpa_add_rx_counters() 103 node->mcast_counter.mdev = ndev->mvdev.mdev; in mlx5_vdpa_add_rx_counters() 142 mdev = ndev->mvdev.mdev; in mlx5_vdpa_add_debugfs() 143 ndev->debugfs = debugfs_create_dir(dev_name(&ndev->mvdev.vdev.dev), in mlx5_vdpa_add_debugfs()
|
| A D | mlx5_vnet.h | 10 container_of(__mvdev, struct mlx5_vdpa_net, mvdev) 44 struct mlx5_vdpa_dev mvdev; member
|
| /drivers/vdpa/mlx5/core/ |
| A D | resources.c | 233 if (!mvdev->cvq.iotlb) in init_ctrl_vq() 237 vringh_set_iotlb(&mvdev->cvq.vring, mvdev->cvq.iotlb, &mvdev->cvq.iommu_lock); in init_ctrl_vq() 244 vhost_iotlb_free(mvdev->cvq.iotlb); in cleanup_ctrl_vq() 265 err = create_uctx(mvdev, &res->uid); in mlx5_vdpa_alloc_resources() 286 err = init_ctrl_vq(mvdev); in mlx5_vdpa_alloc_resources() 297 dealloc_pd(mvdev, res->pdn, res->uid); in mlx5_vdpa_alloc_resources() 299 destroy_uctx(mvdev, res->uid); in mlx5_vdpa_alloc_resources() 313 cleanup_ctrl_vq(mvdev); in mlx5_vdpa_free_resources() 316 dealloc_pd(mvdev, res->pdn, res->uid); in mlx5_vdpa_free_resources() 317 destroy_uctx(mvdev, res->uid); in mlx5_vdpa_free_resources() [all …]
|
| A D | mr.c | 356 fill_indir(mvdev, mr, in); in create_indirect_key() 437 destroy_direct_mr(mvdev, mr); in unmap_direct_mr() 486 unmap_direct_mr(mvdev, dmr); in add_direct_chain() 560 unmap_direct_mr(mvdev, dmr); in create_user_mr() 637 unmap_direct_mr(mvdev, dmr); in destroy_user_mr() 648 destroy_user_mr(mvdev, mr); in _mlx5_vdpa_destroy_mr() 650 destroy_dma_mr(mvdev, mr); in _mlx5_vdpa_destroy_mr() 669 struct mlx5_vdpa_dev *mvdev; in mlx5_vdpa_mr_gc_handler() local 763 if (!mvdev->res.valid) in mlx5_vdpa_clean_mrs() 807 destroy_user_mr(mvdev, mr); in _mlx5_vdpa_create_mr() [all …]
|
| A D | mlx5_vdpa.h | 72 struct mlx5_vdpa_dev *mvdev; member 144 int mlx5_vdpa_alloc_resources(struct mlx5_vdpa_dev *mvdev); 145 void mlx5_vdpa_free_resources(struct mlx5_vdpa_dev *mvdev); 151 int mlx5_vdpa_init_mr_resources(struct mlx5_vdpa_dev *mvdev); 153 void mlx5_vdpa_clean_mrs(struct mlx5_vdpa_dev *mvdev); 154 void mlx5_vdpa_get_mr(struct mlx5_vdpa_dev *mvdev, 156 void mlx5_vdpa_put_mr(struct mlx5_vdpa_dev *mvdev, 158 void mlx5_vdpa_update_mr(struct mlx5_vdpa_dev *mvdev, 161 int mlx5_vdpa_update_cvq_iotlb(struct mlx5_vdpa_dev *mvdev, 164 int mlx5_vdpa_create_dma_mr(struct mlx5_vdpa_dev *mvdev); [all …]
|
| /drivers/vfio/pci/mlx5/ |
| A D | main.c | 305 struct mlx5vf_pci_core_device *mvdev = migf->mvdev; in mlx5vf_mig_file_save_work() local 389 if (mvdev->chunk_mode) { in mlx5vf_prep_stop_copy() 429 if (mvdev->chunk_mode) { in mlx5vf_prep_stop_copy() 462 struct mlx5vf_pci_core_device *mvdev = migf->mvdev; in mlx5vf_precopy_ioctl() local 623 migf->mvdev = mvdev; in mlx5vf_pci_save_device_data() 1009 migf->mvdev = mvdev; in mlx5vf_pci_resume_device_data() 1053 if (mvdev->saving_migf) { in mlx5vf_disable_fds() 1279 if (!mvdev->migrate_cap) in mlx5vf_pci_aer_reset_done() 1310 if (mvdev->migrate_cap) in mlx5vf_pci_open_device() 1390 if (IS_ERR(mvdev)) in mlx5vf_pci_probe() [all …]
|
| A D | cmd.c | 46 if (mvdev->mdev_detach) in mlx5vf_cmd_suspend_vhca() 78 if (mvdev->mdev_detach) in mlx5vf_cmd_resume_vhca() 214 mlx5_sriov_blocking_notifier_unregister(mvdev->mdev, mvdev->vf_id, in mlx5vf_cmd_remove_migratable() 215 &mvdev->nb); in mlx5vf_cmd_remove_migratable() 230 if (!mvdev->mdev) in mlx5vf_cmd_set_migratable() 244 ret = mlx5vf_is_migratable(mvdev->mdev, mvdev->vf_id + 1); in mlx5vf_cmd_set_migratable() 248 if (mlx5vf_cmd_get_vhca_id(mvdev->mdev, mvdev->vf_id + 1, in mlx5vf_cmd_set_migratable() 253 if (!mvdev->cb_wq) in mlx5vf_cmd_set_migratable() 259 ret = mlx5_sriov_blocking_notifier_register(mvdev->mdev, mvdev->vf_id, in mlx5vf_cmd_set_migratable() 424 struct mlx5vf_pci_core_device *mvdev = buf->migf->mvdev; in mlx5vf_dma_data_buffer() local [all …]
|
| A D | cmd.h | 109 struct mlx5vf_pci_core_device *mvdev; member 197 int mlx5vf_cmd_suspend_vhca(struct mlx5vf_pci_core_device *mvdev, u16 op_mod); 198 int mlx5vf_cmd_resume_vhca(struct mlx5vf_pci_core_device *mvdev, u16 op_mod); 199 int mlx5vf_cmd_query_vhca_migration_state(struct mlx5vf_pci_core_device *mvdev, 202 void mlx5vf_cmd_set_migratable(struct mlx5vf_pci_core_device *mvdev, 205 void mlx5vf_cmd_remove_migratable(struct mlx5vf_pci_core_device *mvdev); 206 void mlx5vf_cmd_close_migratable(struct mlx5vf_pci_core_device *mvdev); 207 int mlx5vf_cmd_save_vhca_state(struct mlx5vf_pci_core_device *mvdev, 211 int mlx5vf_cmd_load_vhca_state(struct mlx5vf_pci_core_device *mvdev, 236 void mlx5vf_state_mutex_unlock(struct mlx5vf_pci_core_device *mvdev); [all …]
|
Completed in 36 milliseconds