| /linux/drivers/gpu/drm/amd/amdgpu/ |
| A D | amdgpu_xgmi.c | 360 if (!adev->gmc.xgmi.hive_id) in amdgpu_get_xgmi_hive() 495 adev->gmc.xgmi.node_id, in amdgpu_xgmi_update_topology() 595 adev->gmc.xgmi.node_id = adev->gmc.xgmi.physical_node_id + 16; in amdgpu_xgmi_add_device() 603 adev->gmc.xgmi.node_id, adev->gmc.xgmi.hive_id); in amdgpu_xgmi_add_device() 625 adev->gmc.xgmi.node_id; in amdgpu_xgmi_add_device() 685 adev->gmc.xgmi.physical_node_id, adev->gmc.xgmi.hive_id); in amdgpu_xgmi_add_device() 689 adev->gmc.xgmi.physical_node_id, adev->gmc.xgmi.hive_id, in amdgpu_xgmi_add_device() 747 if (!adev->gmc.xgmi.ras_if) in amdgpu_xgmi_ras_late_init() 767 adev->gmc.xgmi.ras_if) { in amdgpu_xgmi_ras_fini() 781 struct amdgpu_xgmi *xgmi = &adev->gmc.xgmi; in amdgpu_xgmi_get_relative_phy_addr() local [all …]
|
| A D | gfxhub_v1_1.c | 88 if (max_region || adev->gmc.xgmi.connected_to_cpu) { in gfxhub_v1_1_get_xgmi_info() 89 adev->gmc.xgmi.num_physical_nodes = max_region + 1; in gfxhub_v1_1_get_xgmi_info() 91 if (adev->gmc.xgmi.num_physical_nodes > max_num_physical_nodes) in gfxhub_v1_1_get_xgmi_info() 95 adev->gmc.xgmi.physical_node_id = in gfxhub_v1_1_get_xgmi_info() 99 adev->gmc.xgmi.physical_node_id = in gfxhub_v1_1_get_xgmi_info() 104 if (adev->gmc.xgmi.physical_node_id > max_physical_node_id) in gfxhub_v1_1_get_xgmi_info() 107 adev->gmc.xgmi.node_segment_size = seg_size; in gfxhub_v1_1_get_xgmi_info()
|
| A D | amdgpu_gmc.c | 48 u64 vram_size = adev->gmc.xgmi.node_segment_size * adev->gmc.xgmi.num_physical_nodes; in amdgpu_gmc_pdb0_alloc() 208 if (mc->xgmi.num_physical_nodes == 0) { in amdgpu_gmc_vram_location() 236 u64 hive_vram_end = mc->xgmi.node_segment_size * mc->xgmi.num_physical_nodes - 1; in amdgpu_gmc_sysvm_location() 237 mc->vram_start = mc->xgmi.node_segment_size * mc->xgmi.physical_node_id; in amdgpu_gmc_sysvm_location() 451 if (!adev->gmc.xgmi.connected_to_cpu) in amdgpu_gmc_ras_late_init() 454 if (adev->gmc.xgmi.ras_funcs && in amdgpu_gmc_ras_late_init() 502 if (adev->gmc.xgmi.ras_funcs && in amdgpu_gmc_ras_fini() 503 adev->gmc.xgmi.ras_funcs->ras_fini) in amdgpu_gmc_ras_fini() 504 adev->gmc.xgmi.ras_funcs->ras_fini(adev); in amdgpu_gmc_ras_fini() 740 u64 vram_size = adev->gmc.xgmi.node_segment_size * adev->gmc.xgmi.num_physical_nodes; in amdgpu_gmc_init_pdb0() [all …]
|
| A D | aldebaran.c | 51 if (adev->gmc.xgmi.connected_to_cpu) { in aldebaran_get_reset_handler() 150 gmc.xgmi.head) { in aldebaran_mode2_perform_reset() 159 gmc.xgmi.head) { in aldebaran_mode2_perform_reset() 161 if (tmp_adev->gmc.xgmi.num_physical_nodes > 1) { in aldebaran_mode2_perform_reset() 179 gmc.xgmi.head) { in aldebaran_mode2_perform_reset() 180 if (tmp_adev->gmc.xgmi.num_physical_nodes > 1) { in aldebaran_mode2_perform_reset() 190 gmc.xgmi.head) { in aldebaran_mode2_perform_reset() 331 gmc.xgmi.head) { in aldebaran_mode2_restore_hwcontext() 349 tmp_adev->gmc.xgmi.num_physical_nodes > 1) in aldebaran_mode2_restore_hwcontext()
|
| A D | gmc_v9_0.c | 1179 if (!adev->gmc.xgmi.connected_to_cpu) in gmc_v9_0_set_umc_funcs() 1254 adev->gmc.xgmi.supported = true; in gmc_v9_0_early_init() 1257 adev->gmc.xgmi.supported = true; in gmc_v9_0_early_init() 1258 adev->gmc.xgmi.connected_to_cpu = in gmc_v9_0_early_init() 1325 base += adev->gmc.xgmi.physical_node_id * adev->gmc.xgmi.node_segment_size; in gmc_v9_0_vram_gtt_location() 1338 adev->gmc.xgmi.physical_node_id * adev->gmc.xgmi.node_segment_size; in gmc_v9_0_vram_gtt_location() 1382 (adev->gmc.xgmi.supported && in gmc_v9_0_mc_init() 1386 adev->gmc.xgmi.physical_node_id * in gmc_v9_0_mc_init() 1387 adev->gmc.xgmi.node_segment_size; in gmc_v9_0_mc_init() 1599 if (adev->gmc.xgmi.supported) { in gmc_v9_0_sw_init() [all …]
|
| A D | amdgpu_xgmi.h | 70 adev->gmc.xgmi.hive_id && in amdgpu_xgmi_same_hive() 71 adev->gmc.xgmi.hive_id == bo_adev->gmc.xgmi.hive_id); in amdgpu_xgmi_same_hive()
|
| A D | amdgpu_amdkfd.c | 543 return adev->gmc.xgmi.hive_id; in amdgpu_amdkfd_get_hive_id() 561 adev->gmc.xgmi.physical_node_id, in amdgpu_amdkfd_get_xgmi_hops_count() 562 peer_adev->gmc.xgmi.physical_node_id, ret); in amdgpu_amdkfd_get_xgmi_hops_count() 583 adev->gmc.xgmi.physical_node_id, in amdgpu_amdkfd_get_xgmi_bandwidth_mbytes() 584 peer_adev->gmc.xgmi.physical_node_id, num_links); in amdgpu_amdkfd_get_xgmi_bandwidth_mbytes() 791 if (!adev->gmc.xgmi.connected_to_cpu) in amdgpu_amdkfd_ras_poison_consumption_handler()
|
| A D | gfxhub_v2_1.c | 519 adev->gmc.xgmi.num_physical_nodes = max_region + 1; in gfxhub_v2_1_get_xgmi_info() 520 if (adev->gmc.xgmi.num_physical_nodes > max_num_physical_nodes) in gfxhub_v2_1_get_xgmi_info() 523 adev->gmc.xgmi.physical_node_id = in gfxhub_v2_1_get_xgmi_info() 525 if (adev->gmc.xgmi.physical_node_id > max_physical_node_id) in gfxhub_v2_1_get_xgmi_info() 528 adev->gmc.xgmi.node_segment_size = REG_GET_FIELD( in gfxhub_v2_1_get_xgmi_info()
|
| A D | amdgpu_device.c | 1291 if (adev->gmc.xgmi.pending_reset) in amdgpu_device_need_post() 2394 if (adev->gmc.xgmi.num_physical_nodes > 1) in amdgpu_device_ip_init() 2398 if (!adev->gmc.xgmi.pending_reset) in amdgpu_device_ip_init() 2879 if (adev->gmc.xgmi.pending_reset && in amdgpu_device_ip_suspend_phase2() 3581 if (adev->gmc.xgmi.num_physical_nodes) { in amdgpu_device_init() 3583 adev->gmc.xgmi.pending_reset = true; in amdgpu_device_init() 3735 if (!adev->gmc.xgmi.pending_reset) { in amdgpu_device_init() 3777 if (adev->gmc.xgmi.pending_reset) in amdgpu_device_init() 3807 if (!adev->gmc.xgmi.connected_to_cpu) { in amdgpu_device_unmap_mmio() 5618 if (adev->gmc.xgmi.connected_to_cpu) in amdgpu_device_flush_hdp() [all …]
|
| A D | amdgpu_ras.c | 949 if (adev->gmc.xgmi.ras_funcs && in amdgpu_ras_query_error_status() 1101 if (adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_ras_error_inject() 1838 list_add_tail(&adev->gmc.xgmi.head, &device_list); in amdgpu_ras_do_recovery() 1843 device_list_handle, gmc.xgmi.head) { in amdgpu_ras_do_recovery() 2094 if (adev->gmc.xgmi.pending_reset) in amdgpu_ras_recovery_init() 2115 (adev->gmc.xgmi.connected_to_cpu)) in amdgpu_ras_recovery_init() 2205 if (!adev->gmc.xgmi.connected_to_cpu) { in amdgpu_ras_check_supported() 2316 if (!adev->gmc.xgmi.connected_to_cpu) in amdgpu_ras_init() 2374 if (adev->gmc.xgmi.connected_to_cpu) in amdgpu_persistent_edc_harvesting_supported() 2619 if (adev && adev->gmc.xgmi.connected_to_cpu && in find_adev() [all …]
|
| A D | gmc_v10_0.c | 750 base += adev->gmc.xgmi.physical_node_id * adev->gmc.xgmi.node_segment_size; in gmc_v10_0_vram_gtt_location() 761 adev->gmc.xgmi.physical_node_id * adev->gmc.xgmi.node_segment_size; in gmc_v10_0_vram_gtt_location() 917 if (adev->gmc.xgmi.supported) { in gmc_v10_0_sw_init()
|
| A D | amdgpu_psp.c | 1016 adev->gmc.xgmi.connected_to_cpu)) in psp_xgmi_terminate() 1133 uint64_t src_node_id = psp->adev->gmc.xgmi.node_id; in psp_xgmi_reflect_topology_info() 1139 list_for_each_entry(mirror_adev, &hive->device_list, gmc.xgmi.head) { in psp_xgmi_reflect_topology_info() 1143 if (mirror_adev->gmc.xgmi.node_id != dst_node_id) in psp_xgmi_reflect_topology_info() 1483 if (!adev->gmc.xgmi.connected_to_cpu) in psp_ras_initialize() 2556 if (adev->gmc.xgmi.num_physical_nodes > 1 && in psp_suspend() 2652 if (adev->gmc.xgmi.num_physical_nodes > 1) { in psp_resume() 2949 if (adev->gmc.xgmi.connected_to_cpu || in psp_init_sos_base_fw()
|
| A D | amdgpu_gmc.h | 254 struct amdgpu_xgmi xgmi; member
|
| A D | psp_v11_0.c | 155 le32_to_cpu(ta_hdr->xgmi.fw_version); in psp_v11_0_init_microcode() 157 le32_to_cpu(ta_hdr->xgmi.size_bytes); in psp_v11_0_init_microcode()
|
| A D | amdgpu_ib.c | 379 } else if (adev->gmc.xgmi.hive_id) { in amdgpu_ib_ring_tests()
|
| A D | amdgpu_ucode.h | 139 struct psp_fw_legacy_bin_desc xgmi; member
|
| A D | gfxhub_v1_0.c | 206 if (adev->gmc.xgmi.connected_to_cpu) { in gfxhub_v1_0_init_cache_regs()
|
| A D | soc15.c | 562 if (adev->gmc.xgmi.supported && adev->gmc.xgmi.connected_to_cpu) in soc15_asic_reset_method()
|
| A D | amdgpu_vram_mgr.c | 466 if (adev->gmc.xgmi.connected_to_cpu) in amdgpu_vram_mgr_new()
|
| A D | amdgpu_amdkfd_gpuvm.c | 434 if (adev->gmc.xgmi.connected_to_cpu || in get_pte_flags() 442 if (adev->gmc.xgmi.connected_to_cpu) in get_pte_flags()
|
| A D | amdgpu_object.c | 1033 if (!adev->gmc.xgmi.connected_to_cpu) { in amdgpu_bo_init()
|
| A D | mmhub_v1_7.c | 212 if (adev->gmc.xgmi.connected_to_cpu) { in mmhub_v1_7_init_cache_regs()
|
| A D | amdgpu_discovery.c | 1201 adev->gmc.xgmi.supported = true; in amdgpu_discovery_set_ip_blocks()
|
| A D | amdgpu_drv.c | 2122 adev->gmc.xgmi.pending_reset = false; in amdgpu_drv_delayed_reset_work_handler()
|
| /linux/drivers/gpu/drm/amd/amdkfd/ |
| A D | kfd_topology.c | 1302 if (gpu->hive_id && adev->gmc.xgmi.connected_to_cpu) { in kfd_topology_add_device()
|