| /drivers/gpu/drm/amd/amdgpu/ |
| A D | amdgpu_ttm.c | 1164 pool = &adev->mman.bdev.pool; in amdgpu_ttm_tt_populate() 1209 pool = &adev->mman.bdev.pool; in amdgpu_ttm_tt_unpopulate() 1500 if (!adev->mman.sdma_access_ptr) in amdgpu_ttm_access_memory_sdma() 1819 if (!adev->mman.ttm_pools) in amdgpu_ttm_pools_init() 1840 kfree(adev->mman.ttm_pools); in amdgpu_ttm_pools_fini() 1841 adev->mman.ttm_pools = NULL; in amdgpu_ttm_pools_fini() 1878 adev->mman.initialized = true; in amdgpu_ttm_init() 1925 if (adev->mman.discovery_bin) { in amdgpu_ttm_init() 2052 if (!adev->mman.initialized) in amdgpu_ttm_fini() 2095 adev->mman.initialized = false; in amdgpu_ttm_fini() [all …]
|
| A D | amdgpu_preempt_mgr.c | 42 struct ttm_resource_manager *man = &adev->mman.preempt_mgr; in mem_info_preempt_used_show() 102 struct ttm_resource_manager *man = &adev->mman.preempt_mgr; in amdgpu_preempt_mgr_init() 108 ttm_resource_manager_init(man, &adev->mman.bdev, (1 << 30)); in amdgpu_preempt_mgr_init() 116 ttm_set_driver_manager(&adev->mman.bdev, AMDGPU_PL_PREEMPT, man); in amdgpu_preempt_mgr_init() 131 struct ttm_resource_manager *man = &adev->mman.preempt_mgr; in amdgpu_preempt_mgr_fini() 136 ret = ttm_resource_manager_evict_all(&adev->mman.bdev, man); in amdgpu_preempt_mgr_fini() 144 ttm_set_driver_manager(&adev->mman.bdev, AMDGPU_PL_PREEMPT, NULL); in amdgpu_preempt_mgr_fini()
|
| A D | amdgpu_gtt_mgr.c | 51 man = ttm_manager_type(&adev->mman.bdev, TTM_PL_TT); in amdgpu_mem_info_gtt_total_show() 69 struct ttm_resource_manager *man = &adev->mman.gtt_mgr.manager; in amdgpu_mem_info_gtt_used_show() 196 adev = container_of(mgr, typeof(*adev), mman.gtt_mgr); in amdgpu_gtt_mgr_recover() 277 struct amdgpu_gtt_mgr *mgr = &adev->mman.gtt_mgr; in amdgpu_gtt_mgr_init() 284 ttm_resource_manager_init(man, &adev->mman.bdev, gtt_size); in amdgpu_gtt_mgr_init() 291 ttm_set_driver_manager(&adev->mman.bdev, TTM_PL_TT, &mgr->manager); in amdgpu_gtt_mgr_init() 306 struct amdgpu_gtt_mgr *mgr = &adev->mman.gtt_mgr; in amdgpu_gtt_mgr_fini() 312 ret = ttm_resource_manager_evict_all(&adev->mman.bdev, man); in amdgpu_gtt_mgr_fini() 321 ttm_set_driver_manager(&adev->mman.bdev, TTM_PL_TT, NULL); in amdgpu_gtt_mgr_fini()
|
| A D | amdgpu_discovery.c | 457 adev->mman.discovery_bin = kzalloc(adev->mman.discovery_tmr_size, GFP_KERNEL); in amdgpu_discovery_init() 458 if (!adev->mman.discovery_bin) in amdgpu_discovery_init() 471 adev, adev->mman.discovery_bin); in amdgpu_discovery_init() 610 kfree(adev->mman.discovery_bin); in amdgpu_discovery_init() 611 adev->mman.discovery_bin = NULL; in amdgpu_discovery_init() 623 kfree(adev->mman.discovery_bin); in amdgpu_discovery_fini() 624 adev->mman.discovery_bin = NULL; in amdgpu_discovery_fini() 1223 if (!adev->mman.discovery_bin) in amdgpu_discovery_sysfs_init() 1580 if (!adev->mman.discovery_bin) { in amdgpu_discovery_get_gfx_info() 1690 if (!adev->mman.discovery_bin) { in amdgpu_discovery_get_mall_info() [all …]
|
| A D | amdgpu_virt.c | 367 struct amdgpu_vram_mgr *mgr = &adev->mman.vram_mgr; in amdgpu_virt_ras_reserve_bps() 385 amdgpu_vram_mgr_reserve_range(&adev->mman.vram_mgr, in amdgpu_virt_ras_reserve_bps() 426 if (adev->mman.fw_vram_usage_va) in amdgpu_virt_add_bad_page() 427 vram_usage_va = adev->mman.fw_vram_usage_va; in amdgpu_virt_add_bad_page() 429 vram_usage_va = adev->mman.drv_vram_usage_va; in amdgpu_virt_add_bad_page() 604 amdgpu_vram_mgr_vis_usage(&adev->mman.vram_mgr) >> 20; in amdgpu_virt_write_vf2pf_data() 677 if (adev->mman.fw_vram_usage_va && adev->mman.drv_vram_usage_va) { in amdgpu_virt_init_data_exchange() 679 } else if (adev->mman.fw_vram_usage_va || adev->mman.drv_vram_usage_va) { in amdgpu_virt_init_data_exchange() 702 if (adev->mman.fw_vram_usage_va || adev->mman.drv_vram_usage_va) { in amdgpu_virt_exchange_data() 703 if (adev->mman.fw_vram_usage_va) { in amdgpu_virt_exchange_data() [all …]
|
| A D | amdgpu_gmc.c | 689 mutex_lock(&adev->mman.gtt_window_lock); in amdgpu_gmc_flush_gpu_tlb() 702 mutex_unlock(&adev->mman.gtt_window_lock); in amdgpu_gmc_flush_gpu_tlb() 710 mutex_unlock(&adev->mman.gtt_window_lock); in amdgpu_gmc_flush_gpu_tlb() 980 adev->mman.stolen_reserved_offset = 0; in amdgpu_gmc_get_vbios_allocations() 981 adev->mman.stolen_reserved_size = 0; in amdgpu_gmc_get_vbios_allocations() 993 adev->mman.keep_stolen_vga_memory = true; in amdgpu_gmc_get_vbios_allocations() 1006 adev->mman.keep_stolen_vga_memory = true; in amdgpu_gmc_get_vbios_allocations() 1019 if (adev->mman.keep_stolen_vga_memory) in amdgpu_gmc_get_vbios_allocations() 1029 adev->mman.stolen_extended_size = size - adev->mman.stolen_vga_size; in amdgpu_gmc_get_vbios_allocations() 1031 adev->mman.stolen_vga_size = size; in amdgpu_gmc_get_vbios_allocations() [all …]
|
| A D | amdgpu_vram_mgr.c | 52 return container_of(mgr, struct amdgpu_device, mman.vram_mgr); in to_amdgpu_device() 141 struct ttm_resource_manager *man = &adev->mman.vram_mgr.manager; in amdgpu_mem_info_vram_used_show() 162 amdgpu_vram_mgr_vis_usage(&adev->mman.vram_mgr)); in amdgpu_mem_info_vis_vram_used_show() 793 struct amdgpu_vram_mgr *mgr = &adev->mman.vram_mgr; in amdgpu_vram_mgr_clear_reset_blocks() 923 struct amdgpu_vram_mgr *mgr = &adev->mman.vram_mgr; in amdgpu_vram_mgr_init() 930 ttm_resource_manager_init(man, &adev->mman.bdev, in amdgpu_vram_mgr_init() 949 ttm_set_driver_manager(&adev->mman.bdev, TTM_PL_VRAM, &mgr->manager); in amdgpu_vram_mgr_init() 964 struct amdgpu_vram_mgr *mgr = &adev->mman.vram_mgr; in amdgpu_vram_mgr_fini() 971 ret = ttm_resource_manager_evict_all(&adev->mman.bdev, man); in amdgpu_vram_mgr_fini() 988 ttm_set_driver_manager(&adev->mman.bdev, TTM_PL_VRAM, NULL); in amdgpu_vram_mgr_fini()
|
| A D | amdgpu_sdma.h | 178 #define amdgpu_emit_copy_buffer(adev, ib, s, d, b, t) (adev)->mman.buffer_funcs->emit_copy_buffer((… 179 #define amdgpu_emit_fill_buffer(adev, ib, s, d, b) (adev)->mman.buffer_funcs->emit_fill_buffer((ib)…
|
| A D | psp_v11_0.c | 518 if (adev->gmc.visible_vram_size < sz || !adev->mman.aper_base_kaddr) { in psp_v11_0_memory_training() 521 adev->mman.aper_base_kaddr); in psp_v11_0_memory_training() 532 memcpy_fromio(buf, adev->mman.aper_base_kaddr, sz); in psp_v11_0_memory_training() 541 memcpy_toio(adev->mman.aper_base_kaddr, buf, sz); in psp_v11_0_memory_training()
|
| A D | amdgpu_atomfirmware.c | 123 adev->mman.fw_vram_usage_start_offset = (start_addr & in amdgpu_atomfirmware_allocate_fb_v2_1() 125 adev->mman.fw_vram_usage_size = fw_size << 10; in amdgpu_atomfirmware_allocate_fb_v2_1() 155 adev->mman.fw_vram_usage_start_offset = (fw_start_addr & in amdgpu_atomfirmware_allocate_fb_v2_2() 157 adev->mman.fw_vram_usage_size = fw_size << 10; in amdgpu_atomfirmware_allocate_fb_v2_2() 164 adev->mman.drv_vram_usage_start_offset = (drv_start_addr & in amdgpu_atomfirmware_allocate_fb_v2_2() 166 adev->mman.drv_vram_usage_size = drv_size << 10; in amdgpu_atomfirmware_allocate_fb_v2_2()
|
| A D | amdgpu_benchmark.c | 40 struct amdgpu_ring *ring = adev->mman.buffer_funcs_ring; in amdgpu_benchmark_do_move() 101 if (adev->mman.buffer_funcs) { in amdgpu_benchmark_move()
|
| A D | psp_v14_0.c | 484 if (adev->gmc.visible_vram_size < sz || !adev->mman.aper_base_kaddr) { in psp_v14_0_memory_training() 487 adev->mman.aper_base_kaddr); in psp_v14_0_memory_training() 498 memcpy_fromio(buf, adev->mman.aper_base_kaddr, sz); in psp_v14_0_memory_training() 507 memcpy_toio(adev->mman.aper_base_kaddr, buf, sz); in psp_v14_0_memory_training()
|
| A D | psp_v13_0.c | 619 if (adev->gmc.visible_vram_size < sz || !adev->mman.aper_base_kaddr) { in psp_v13_0_memory_training() 622 adev->mman.aper_base_kaddr); in psp_v13_0_memory_training() 633 memcpy_fromio(buf, adev->mman.aper_base_kaddr, sz); in psp_v13_0_memory_training() 642 memcpy_toio(adev->mman.aper_base_kaddr, buf, sz); in psp_v13_0_memory_training()
|
| A D | amdgpu_device.c | 618 if (!adev->mman.aper_base_kaddr) in amdgpu_device_aper_access() 1851 if (adev->mman.keep_stolen_vga_memory) in amdgpu_device_seamless_boot_supported() 2594 if (adev->mman.discovery_bin) in amdgpu_device_parse_gpu_info_fw() 4362 adev->mman.buffer_funcs = NULL; in amdgpu_device_init() 4363 adev->mman.buffer_funcs_ring = NULL; in amdgpu_device_init() 4857 if (adev->mman.aper_base_kaddr) in amdgpu_device_unmap_mmio() 4858 iounmap(adev->mman.aper_base_kaddr); in amdgpu_device_unmap_mmio() 4859 adev->mman.aper_base_kaddr = NULL; in amdgpu_device_unmap_mmio() 4881 if (adev->mman.initialized) in amdgpu_device_fini_hw() 4924 if (adev->mman.initialized) in amdgpu_device_fini_hw() [all …]
|
| A D | gmc_v12_0.c | 315 mutex_lock(&adev->mman.gtt_window_lock); in gmc_v12_0_flush_gpu_tlb() 317 mutex_unlock(&adev->mman.gtt_window_lock); in gmc_v12_0_flush_gpu_tlb() 870 amdgpu_gtt_mgr_recover(&adev->mman.gtt_mgr); in gmc_v12_0_gart_enable()
|
| A D | amdgpu_kms.c | 761 ui64 = ttm_resource_manager_usage(&adev->mman.vram_mgr.manager); in amdgpu_info_ioctl() 764 ui64 = amdgpu_vram_mgr_vis_usage(&adev->mman.vram_mgr); in amdgpu_info_ioctl() 767 ui64 = ttm_resource_manager_usage(&adev->mman.gtt_mgr.manager); in amdgpu_info_ioctl() 790 vram_gtt.gtt_size = ttm_manager_type(&adev->mman.bdev, TTM_PL_TT)->size; in amdgpu_info_ioctl() 798 &adev->mman.gtt_mgr.manager; in amdgpu_info_ioctl() 800 &adev->mman.vram_mgr.manager; in amdgpu_info_ioctl() 818 amdgpu_vram_mgr_vis_usage(&adev->mman.vram_mgr); in amdgpu_info_ioctl()
|
| A D | amdgpu_object.c | 560 man = ttm_manager_type(&adev->mman.bdev, TTM_PL_TT); in amdgpu_bo_validate_size() 562 man = ttm_manager_type(&adev->mman.bdev, TTM_PL_VRAM); in amdgpu_bo_validate_size() 689 bo->tbo.bdev = &adev->mman.bdev; in amdgpu_bo_create() 703 r = ttm_bo_init_reserved(&adev->mman.bdev, &bo->tbo, bp->type, in amdgpu_bo_create()
|
| A D | si_dma.c | 829 adev->mman.buffer_funcs = &si_dma_buffer_funcs; in si_dma_set_buffer_funcs() 830 adev->mman.buffer_funcs_ring = &adev->sdma.instance[0].ring; in si_dma_set_buffer_funcs()
|
| /drivers/gpu/drm/qxl/ |
| A D | qxl_ttm.c | 42 struct qxl_mman *mman; in qxl_get_qdev() local 45 mman = container_of(bdev, struct qxl_mman, bdev); in qxl_get_qdev() 46 qdev = container_of(mman, struct qxl_device, mman); in qxl_get_qdev() 187 return ttm_range_man_init(&qdev->mman.bdev, type, false, size); in qxl_ttm_init_mem_type() 196 r = ttm_device_init(&qdev->mman.bdev, &qxl_bo_driver, NULL, in qxl_ttm_init() 228 ttm_range_man_fini(&qdev->mman.bdev, TTM_PL_VRAM); in qxl_ttm_fini() 229 ttm_range_man_fini(&qdev->mman.bdev, TTM_PL_PRIV); in qxl_ttm_fini() 230 ttm_device_fini(&qdev->mman.bdev); in qxl_ttm_fini() 237 ttm_resource_manager_create_debugfs(ttm_manager_type(&qdev->mman.bdev, in qxl_ttm_debugfs_init() 240 ttm_resource_manager_create_debugfs(ttm_manager_type(&qdev->mman.bdev, in qxl_ttm_debugfs_init()
|
| A D | qxl_object.c | 139 r = ttm_bo_init_reserved(&qdev->mman.bdev, &bo->tbo, type, in qxl_bo_create() 408 man = ttm_manager_type(&qdev->mman.bdev, TTM_PL_PRIV); in qxl_surf_evict() 409 return ttm_resource_manager_evict_all(&qdev->mman.bdev, man); in qxl_surf_evict() 416 man = ttm_manager_type(&qdev->mman.bdev, TTM_PL_VRAM); in qxl_vram_evict() 417 return ttm_resource_manager_evict_all(&qdev->mman.bdev, man); in qxl_vram_evict()
|
| /drivers/gpu/drm/radeon/ |
| A D | radeon_ttm.c | 62 struct radeon_mman *mman; in radeon_get_rdev() local 65 mman = container_of(bdev, struct radeon_mman, bdev); in radeon_get_rdev() 66 rdev = container_of(mman, struct radeon_device, mman); in radeon_get_rdev() 78 return ttm_range_man_init(&rdev->mman.bdev, TTM_PL_TT, in radeon_ttm_init_gtt() 570 return ttm_pool_free(&rdev->mman.bdev.pool, ttm); in radeon_ttm_tt_unpopulate() 692 rdev->mman.initialized = true; in radeon_ttm_init() 737 if (!rdev->mman.initialized) in radeon_ttm_fini() 749 ttm_range_man_fini(&rdev->mman.bdev, TTM_PL_TT); in radeon_ttm_fini() 750 ttm_device_fini(&rdev->mman.bdev); in radeon_ttm_fini() 752 rdev->mman.initialized = false; in radeon_ttm_fini() [all …]
|
| A D | radeon_object.c | 204 r = ttm_bo_init_validate(&rdev->mman.bdev, &bo->tbo, type, in radeon_bo_create() 349 struct ttm_device *bdev = &rdev->mman.bdev; in radeon_bo_evict_vram() 418 ttm_manager_type(&rdev->mman.bdev, TTM_PL_VRAM); in radeon_bo_get_threshold_for_moves()
|
| A D | radeon_kms.c | 547 man = ttm_manager_type(&rdev->mman.bdev, TTM_PL_VRAM); in radeon_info_ioctl() 553 man = ttm_manager_type(&rdev->mman.bdev, TTM_PL_TT); in radeon_info_ioctl()
|
| /drivers/gpu/drm/amd/amdkfd/ |
| A D | kfd_migrate.c | 64 num_dw = ALIGN(adev->mman.buffer_funcs->copy_num_dw, 8); in svm_migrate_gart_map() 67 r = amdgpu_job_alloc_with_ib(adev, &adev->mman.high_pr, in svm_migrate_gart_map() 130 struct amdgpu_ring *ring = adev->mman.buffer_funcs_ring; in svm_migrate_copy_memory_gart() 136 mutex_lock(&adev->mman.gtt_window_lock); in svm_migrate_copy_memory_gart() 172 mutex_unlock(&adev->mman.gtt_window_lock); in svm_migrate_copy_memory_gart()
|
| /drivers/gpu/drm/i915/selftests/ |
| A D | i915_live_selftests.h | 33 selftest(mman, i915_gem_mman_live_selftests)
|