Searched refs:hive (Results 1 – 11 of 11) sorted by relevance
422 kfree(hive); in amdgpu_xgmi_hive_release()685 hive = kzalloc(sizeof(*hive), GFP_KERNEL); in amdgpu_get_xgmi_hive()751 if (hive) in amdgpu_get_xgmi_hive()759 if (hive) in amdgpu_put_xgmi_hive()772 if (!hive) in amdgpu_xgmi_set_pstate()811 hive->pstate = hive->hi_req_count ? in amdgpu_xgmi_set_pstate()1125 adev->hive = hive; in amdgpu_xgmi_add_device()1140 struct amdgpu_hive_info *hive = adev->hive; in amdgpu_xgmi_remove_device() local1145 if (!hive) in amdgpu_xgmi_remove_device()1664 reset_context.hive = hive; in amdgpu_xgmi_reset_on_init_work()[all …]
56 struct amdgpu_hive_info *hive = NULL; in amdgpu_job_core_dump() local59 hive = amdgpu_get_xgmi_hive(adev); in amdgpu_job_core_dump()60 if (hive) in amdgpu_job_core_dump()61 mutex_lock(&hive->hive_lock); in amdgpu_job_core_dump()67 if (!amdgpu_sriov_vf(adev) && (adev->gmc.xgmi.num_physical_nodes > 1) && hive) { in amdgpu_job_core_dump()68 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) in amdgpu_job_core_dump()82 if (hive) { in amdgpu_job_core_dump()83 mutex_unlock(&hive->hive_lock); in amdgpu_job_core_dump()84 amdgpu_put_xgmi_hive(hive); in amdgpu_job_core_dump()
5532 if (hive) in amdgpu_device_reset_sriov()6457 if (hive) in amdgpu_device_gpu_recover()6461 reset_context->hive = hive; in amdgpu_device_gpu_recover()6498 if (hive) { in amdgpu_device_gpu_recover()6908 if (hive) in amdgpu_pci_error_detected()6918 if (hive) { in amdgpu_pci_error_detected()7008 if (hive) { in amdgpu_pci_slot_reset()7010 reset_context.hive = hive; in amdgpu_pci_slot_reset()7035 if (hive) { in amdgpu_pci_slot_reset()7067 if (hive) { in amdgpu_pci_resume()[all …]
99 void amdgpu_put_xgmi_hive(struct amdgpu_hive_info *hive);100 int amdgpu_xgmi_update_topology(struct amdgpu_hive_info *hive, struct amdgpu_device *adev);118 struct amdgpu_hive_info *hive,
1206 struct amdgpu_hive_info *hive; in current_memory_partition_store() local1230 hive = amdgpu_get_xgmi_hive(adev); in current_memory_partition_store()1231 if (hive) { in current_memory_partition_store()1232 atomic_set(&hive->requested_nps_mode, mode); in current_memory_partition_store()1233 amdgpu_put_xgmi_hive(hive); in current_memory_partition_store()1409 struct amdgpu_hive_info *hive; in amdgpu_gmc_prepare_nps_mode_change() local1416 hive = amdgpu_get_xgmi_hive(adev); in amdgpu_gmc_prepare_nps_mode_change()1417 if (hive) { in amdgpu_gmc_prepare_nps_mode_change()1418 req_nps_mode = atomic_read(&hive->requested_nps_mode); in amdgpu_gmc_prepare_nps_mode_change()1421 amdgpu_put_xgmi_hive(hive); in amdgpu_gmc_prepare_nps_mode_change()[all …]
162 reset_context->hive == NULL) { in aldebaran_mode2_perform_reset()352 reset_context->hive == NULL) { in aldebaran_mode2_restore_hwcontext()401 if (reset_context->hive && in aldebaran_mode2_restore_hwcontext()403 r = amdgpu_xgmi_update_topology(reset_context->hive, in aldebaran_mode2_restore_hwcontext()
2600 if (hive) { in amdgpu_ras_set_fed_all()2614 if (hive) { in amdgpu_ras_in_recovery()2616 amdgpu_put_xgmi_hive(hive); in amdgpu_ras_in_recovery()2643 if (hive) { in amdgpu_ras_do_recovery()2712 if (hive) { in amdgpu_ras_do_recovery()2714 amdgpu_put_xgmi_hive(hive); in amdgpu_ras_do_recovery()3924 ras->event_mgr = hive ? &hive->event_mgr : &ras->__event_mgr; in amdgpu_ras_event_mgr_init()3932 if (hive) in amdgpu_ras_event_mgr_init()3933 amdgpu_put_xgmi_hive(hive); in amdgpu_ras_event_mgr_init()4750 if (hive) { in amdgpu_ras_reset_gpu()[all …]
52 struct amdgpu_hive_info *hive; member
1536 struct amdgpu_hive_info *hive; in psp_xgmi_reflect_topology_info() local1542 hive = amdgpu_get_xgmi_hive(psp->adev); in psp_xgmi_reflect_topology_info()1543 if (WARN_ON(!hive)) in psp_xgmi_reflect_topology_info()1546 list_for_each_entry(mirror_adev, &hive->device_list, gmc.xgmi.head) { in psp_xgmi_reflect_topology_info()1573 amdgpu_put_xgmi_hive(hive); in psp_xgmi_reflect_topology_info()
946 struct amdgpu_hive_info *hive; member
40 ((adev)->hive ? (void *)(adev)->hive : (void *)(adev))
Completed in 55 milliseconds