Lines Matching refs:gmc

355 	last = min(pos + size, adev->gmc.visible_vram_size);  in amdgpu_device_aper_access()
1184 int rbar_size = pci_rebar_bytes_to_size(adev->gmc.real_vram_size); in amdgpu_device_resize_fb_bar()
1196 if (adev->gmc.real_vram_size && in amdgpu_device_resize_fb_bar()
1197 (pci_resource_len(adev->pdev, 0) >= adev->gmc.real_vram_size)) in amdgpu_device_resize_fb_bar()
1291 if (adev->gmc.xgmi.pending_reset) in amdgpu_device_need_post()
2394 if (adev->gmc.xgmi.num_physical_nodes > 1) in amdgpu_device_ip_init()
2398 if (!adev->gmc.xgmi.pending_reset) in amdgpu_device_ip_init()
2620 adev->gmc.xgmi.num_physical_nodes > 1) in amdgpu_device_ip_late_init()
2623 if (adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_device_ip_late_init()
2639 if (mgpu_info.num_dgpu == adev->gmc.xgmi.num_physical_nodes) { in amdgpu_device_ip_late_init()
2738 if (adev->gmc.xgmi.num_physical_nodes > 1) in amdgpu_device_ip_fini()
2879 if (adev->gmc.xgmi.pending_reset && in amdgpu_device_ip_suspend_phase2()
3414 adev->gmc.gart_size = 512 * 1024 * 1024; in amdgpu_device_init()
3421 adev->gmc.gmc_funcs = NULL; in amdgpu_device_init()
3581 if (adev->gmc.xgmi.num_physical_nodes) { in amdgpu_device_init()
3583 adev->gmc.xgmi.pending_reset = true; in amdgpu_device_init()
3735 if (!adev->gmc.xgmi.pending_reset) { in amdgpu_device_init()
3777 if (adev->gmc.xgmi.pending_reset) in amdgpu_device_init()
3807 if (!adev->gmc.xgmi.connected_to_cpu) { in amdgpu_device_unmap_mmio()
3808 arch_phys_wc_del(adev->gmc.vram_mtrr); in amdgpu_device_unmap_mmio()
3809 arch_io_free_memtype_wc(adev->gmc.aper_base, adev->gmc.aper_size); in amdgpu_device_unmap_mmio()
4565 if (tmp_adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_do_asic_reset()
4566 tmp_adev->gmc.xgmi.pending_reset = false; in amdgpu_do_asic_reset()
4582 if (tmp_adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_do_asic_reset()
4646 tmp_adev->gmc.xgmi.num_physical_nodes > 1) in amdgpu_do_asic_reset()
4675 tmp_adev->gmc.xgmi.num_physical_nodes > 1) in amdgpu_do_asic_reset()
4752 if (adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_device_lock_hive_adev()
4757 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) { in amdgpu_device_lock_hive_adev()
4766 if (!list_is_first(&tmp_adev->gmc.xgmi.head, &hive->device_list)) { in amdgpu_device_lock_hive_adev()
4775 list_for_each_entry_continue_reverse(tmp_adev, &hive->device_list, gmc.xgmi.head) { in amdgpu_device_lock_hive_adev()
5005 if (adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_device_gpu_recover()
5006 list_for_each_entry(tmp_adev, &hive->device_list, gmc.xgmi.head) in amdgpu_device_gpu_recover()
5399 if (adev->gmc.xgmi.num_physical_nodes > 1) { in amdgpu_pci_error_detected()
5618 if (adev->gmc.xgmi.connected_to_cpu) in amdgpu_device_flush_hdp()
5634 if (adev->gmc.xgmi.connected_to_cpu) in amdgpu_device_invalidate_hdp()