Home
last modified time | relevance | path

Searched refs:mman (Results 1 – 25 of 55) sorted by relevance

123

/drivers/gpu/drm/amd/amdgpu/
A Damdgpu_ttm.c1164 pool = &adev->mman.bdev.pool; in amdgpu_ttm_tt_populate()
1209 pool = &adev->mman.bdev.pool; in amdgpu_ttm_tt_unpopulate()
1500 if (!adev->mman.sdma_access_ptr) in amdgpu_ttm_access_memory_sdma()
1819 if (!adev->mman.ttm_pools) in amdgpu_ttm_pools_init()
1840 kfree(adev->mman.ttm_pools); in amdgpu_ttm_pools_fini()
1841 adev->mman.ttm_pools = NULL; in amdgpu_ttm_pools_fini()
1878 adev->mman.initialized = true; in amdgpu_ttm_init()
1925 if (adev->mman.discovery_bin) { in amdgpu_ttm_init()
2052 if (!adev->mman.initialized) in amdgpu_ttm_fini()
2095 adev->mman.initialized = false; in amdgpu_ttm_fini()
[all …]
A Damdgpu_preempt_mgr.c42 struct ttm_resource_manager *man = &adev->mman.preempt_mgr; in mem_info_preempt_used_show()
102 struct ttm_resource_manager *man = &adev->mman.preempt_mgr; in amdgpu_preempt_mgr_init()
108 ttm_resource_manager_init(man, &adev->mman.bdev, (1 << 30)); in amdgpu_preempt_mgr_init()
116 ttm_set_driver_manager(&adev->mman.bdev, AMDGPU_PL_PREEMPT, man); in amdgpu_preempt_mgr_init()
131 struct ttm_resource_manager *man = &adev->mman.preempt_mgr; in amdgpu_preempt_mgr_fini()
136 ret = ttm_resource_manager_evict_all(&adev->mman.bdev, man); in amdgpu_preempt_mgr_fini()
144 ttm_set_driver_manager(&adev->mman.bdev, AMDGPU_PL_PREEMPT, NULL); in amdgpu_preempt_mgr_fini()
A Damdgpu_gtt_mgr.c51 man = ttm_manager_type(&adev->mman.bdev, TTM_PL_TT); in amdgpu_mem_info_gtt_total_show()
69 struct ttm_resource_manager *man = &adev->mman.gtt_mgr.manager; in amdgpu_mem_info_gtt_used_show()
196 adev = container_of(mgr, typeof(*adev), mman.gtt_mgr); in amdgpu_gtt_mgr_recover()
277 struct amdgpu_gtt_mgr *mgr = &adev->mman.gtt_mgr; in amdgpu_gtt_mgr_init()
284 ttm_resource_manager_init(man, &adev->mman.bdev, gtt_size); in amdgpu_gtt_mgr_init()
291 ttm_set_driver_manager(&adev->mman.bdev, TTM_PL_TT, &mgr->manager); in amdgpu_gtt_mgr_init()
306 struct amdgpu_gtt_mgr *mgr = &adev->mman.gtt_mgr; in amdgpu_gtt_mgr_fini()
312 ret = ttm_resource_manager_evict_all(&adev->mman.bdev, man); in amdgpu_gtt_mgr_fini()
321 ttm_set_driver_manager(&adev->mman.bdev, TTM_PL_TT, NULL); in amdgpu_gtt_mgr_fini()
A Damdgpu_discovery.c457 adev->mman.discovery_bin = kzalloc(adev->mman.discovery_tmr_size, GFP_KERNEL); in amdgpu_discovery_init()
458 if (!adev->mman.discovery_bin) in amdgpu_discovery_init()
471 adev, adev->mman.discovery_bin); in amdgpu_discovery_init()
610 kfree(adev->mman.discovery_bin); in amdgpu_discovery_init()
611 adev->mman.discovery_bin = NULL; in amdgpu_discovery_init()
623 kfree(adev->mman.discovery_bin); in amdgpu_discovery_fini()
624 adev->mman.discovery_bin = NULL; in amdgpu_discovery_fini()
1223 if (!adev->mman.discovery_bin) in amdgpu_discovery_sysfs_init()
1580 if (!adev->mman.discovery_bin) { in amdgpu_discovery_get_gfx_info()
1690 if (!adev->mman.discovery_bin) { in amdgpu_discovery_get_mall_info()
[all …]
A Damdgpu_virt.c367 struct amdgpu_vram_mgr *mgr = &adev->mman.vram_mgr; in amdgpu_virt_ras_reserve_bps()
385 amdgpu_vram_mgr_reserve_range(&adev->mman.vram_mgr, in amdgpu_virt_ras_reserve_bps()
426 if (adev->mman.fw_vram_usage_va) in amdgpu_virt_add_bad_page()
427 vram_usage_va = adev->mman.fw_vram_usage_va; in amdgpu_virt_add_bad_page()
429 vram_usage_va = adev->mman.drv_vram_usage_va; in amdgpu_virt_add_bad_page()
604 amdgpu_vram_mgr_vis_usage(&adev->mman.vram_mgr) >> 20; in amdgpu_virt_write_vf2pf_data()
677 if (adev->mman.fw_vram_usage_va && adev->mman.drv_vram_usage_va) { in amdgpu_virt_init_data_exchange()
679 } else if (adev->mman.fw_vram_usage_va || adev->mman.drv_vram_usage_va) { in amdgpu_virt_init_data_exchange()
702 if (adev->mman.fw_vram_usage_va || adev->mman.drv_vram_usage_va) { in amdgpu_virt_exchange_data()
703 if (adev->mman.fw_vram_usage_va) { in amdgpu_virt_exchange_data()
[all …]
A Damdgpu_gmc.c689 mutex_lock(&adev->mman.gtt_window_lock); in amdgpu_gmc_flush_gpu_tlb()
702 mutex_unlock(&adev->mman.gtt_window_lock); in amdgpu_gmc_flush_gpu_tlb()
710 mutex_unlock(&adev->mman.gtt_window_lock); in amdgpu_gmc_flush_gpu_tlb()
980 adev->mman.stolen_reserved_offset = 0; in amdgpu_gmc_get_vbios_allocations()
981 adev->mman.stolen_reserved_size = 0; in amdgpu_gmc_get_vbios_allocations()
993 adev->mman.keep_stolen_vga_memory = true; in amdgpu_gmc_get_vbios_allocations()
1006 adev->mman.keep_stolen_vga_memory = true; in amdgpu_gmc_get_vbios_allocations()
1019 if (adev->mman.keep_stolen_vga_memory) in amdgpu_gmc_get_vbios_allocations()
1029 adev->mman.stolen_extended_size = size - adev->mman.stolen_vga_size; in amdgpu_gmc_get_vbios_allocations()
1031 adev->mman.stolen_vga_size = size; in amdgpu_gmc_get_vbios_allocations()
[all …]
A Damdgpu_vram_mgr.c52 return container_of(mgr, struct amdgpu_device, mman.vram_mgr); in to_amdgpu_device()
141 struct ttm_resource_manager *man = &adev->mman.vram_mgr.manager; in amdgpu_mem_info_vram_used_show()
162 amdgpu_vram_mgr_vis_usage(&adev->mman.vram_mgr)); in amdgpu_mem_info_vis_vram_used_show()
793 struct amdgpu_vram_mgr *mgr = &adev->mman.vram_mgr; in amdgpu_vram_mgr_clear_reset_blocks()
923 struct amdgpu_vram_mgr *mgr = &adev->mman.vram_mgr; in amdgpu_vram_mgr_init()
930 ttm_resource_manager_init(man, &adev->mman.bdev, in amdgpu_vram_mgr_init()
949 ttm_set_driver_manager(&adev->mman.bdev, TTM_PL_VRAM, &mgr->manager); in amdgpu_vram_mgr_init()
964 struct amdgpu_vram_mgr *mgr = &adev->mman.vram_mgr; in amdgpu_vram_mgr_fini()
971 ret = ttm_resource_manager_evict_all(&adev->mman.bdev, man); in amdgpu_vram_mgr_fini()
988 ttm_set_driver_manager(&adev->mman.bdev, TTM_PL_VRAM, NULL); in amdgpu_vram_mgr_fini()
A Damdgpu_sdma.h178 #define amdgpu_emit_copy_buffer(adev, ib, s, d, b, t) (adev)->mman.buffer_funcs->emit_copy_buffer((…
179 #define amdgpu_emit_fill_buffer(adev, ib, s, d, b) (adev)->mman.buffer_funcs->emit_fill_buffer((ib)…
A Dpsp_v11_0.c518 if (adev->gmc.visible_vram_size < sz || !adev->mman.aper_base_kaddr) { in psp_v11_0_memory_training()
521 adev->mman.aper_base_kaddr); in psp_v11_0_memory_training()
532 memcpy_fromio(buf, adev->mman.aper_base_kaddr, sz); in psp_v11_0_memory_training()
541 memcpy_toio(adev->mman.aper_base_kaddr, buf, sz); in psp_v11_0_memory_training()
A Damdgpu_atomfirmware.c123 adev->mman.fw_vram_usage_start_offset = (start_addr & in amdgpu_atomfirmware_allocate_fb_v2_1()
125 adev->mman.fw_vram_usage_size = fw_size << 10; in amdgpu_atomfirmware_allocate_fb_v2_1()
155 adev->mman.fw_vram_usage_start_offset = (fw_start_addr & in amdgpu_atomfirmware_allocate_fb_v2_2()
157 adev->mman.fw_vram_usage_size = fw_size << 10; in amdgpu_atomfirmware_allocate_fb_v2_2()
164 adev->mman.drv_vram_usage_start_offset = (drv_start_addr & in amdgpu_atomfirmware_allocate_fb_v2_2()
166 adev->mman.drv_vram_usage_size = drv_size << 10; in amdgpu_atomfirmware_allocate_fb_v2_2()
A Damdgpu_benchmark.c40 struct amdgpu_ring *ring = adev->mman.buffer_funcs_ring; in amdgpu_benchmark_do_move()
101 if (adev->mman.buffer_funcs) { in amdgpu_benchmark_move()
A Dpsp_v14_0.c484 if (adev->gmc.visible_vram_size < sz || !adev->mman.aper_base_kaddr) { in psp_v14_0_memory_training()
487 adev->mman.aper_base_kaddr); in psp_v14_0_memory_training()
498 memcpy_fromio(buf, adev->mman.aper_base_kaddr, sz); in psp_v14_0_memory_training()
507 memcpy_toio(adev->mman.aper_base_kaddr, buf, sz); in psp_v14_0_memory_training()
A Dpsp_v13_0.c619 if (adev->gmc.visible_vram_size < sz || !adev->mman.aper_base_kaddr) { in psp_v13_0_memory_training()
622 adev->mman.aper_base_kaddr); in psp_v13_0_memory_training()
633 memcpy_fromio(buf, adev->mman.aper_base_kaddr, sz); in psp_v13_0_memory_training()
642 memcpy_toio(adev->mman.aper_base_kaddr, buf, sz); in psp_v13_0_memory_training()
A Damdgpu_device.c618 if (!adev->mman.aper_base_kaddr) in amdgpu_device_aper_access()
1851 if (adev->mman.keep_stolen_vga_memory) in amdgpu_device_seamless_boot_supported()
2594 if (adev->mman.discovery_bin) in amdgpu_device_parse_gpu_info_fw()
4362 adev->mman.buffer_funcs = NULL; in amdgpu_device_init()
4363 adev->mman.buffer_funcs_ring = NULL; in amdgpu_device_init()
4857 if (adev->mman.aper_base_kaddr) in amdgpu_device_unmap_mmio()
4858 iounmap(adev->mman.aper_base_kaddr); in amdgpu_device_unmap_mmio()
4859 adev->mman.aper_base_kaddr = NULL; in amdgpu_device_unmap_mmio()
4881 if (adev->mman.initialized) in amdgpu_device_fini_hw()
4924 if (adev->mman.initialized) in amdgpu_device_fini_hw()
[all …]
A Dgmc_v12_0.c315 mutex_lock(&adev->mman.gtt_window_lock); in gmc_v12_0_flush_gpu_tlb()
317 mutex_unlock(&adev->mman.gtt_window_lock); in gmc_v12_0_flush_gpu_tlb()
870 amdgpu_gtt_mgr_recover(&adev->mman.gtt_mgr); in gmc_v12_0_gart_enable()
A Damdgpu_kms.c761 ui64 = ttm_resource_manager_usage(&adev->mman.vram_mgr.manager); in amdgpu_info_ioctl()
764 ui64 = amdgpu_vram_mgr_vis_usage(&adev->mman.vram_mgr); in amdgpu_info_ioctl()
767 ui64 = ttm_resource_manager_usage(&adev->mman.gtt_mgr.manager); in amdgpu_info_ioctl()
790 vram_gtt.gtt_size = ttm_manager_type(&adev->mman.bdev, TTM_PL_TT)->size; in amdgpu_info_ioctl()
798 &adev->mman.gtt_mgr.manager; in amdgpu_info_ioctl()
800 &adev->mman.vram_mgr.manager; in amdgpu_info_ioctl()
818 amdgpu_vram_mgr_vis_usage(&adev->mman.vram_mgr); in amdgpu_info_ioctl()
A Damdgpu_object.c560 man = ttm_manager_type(&adev->mman.bdev, TTM_PL_TT); in amdgpu_bo_validate_size()
562 man = ttm_manager_type(&adev->mman.bdev, TTM_PL_VRAM); in amdgpu_bo_validate_size()
689 bo->tbo.bdev = &adev->mman.bdev; in amdgpu_bo_create()
703 r = ttm_bo_init_reserved(&adev->mman.bdev, &bo->tbo, bp->type, in amdgpu_bo_create()
A Dsi_dma.c829 adev->mman.buffer_funcs = &si_dma_buffer_funcs; in si_dma_set_buffer_funcs()
830 adev->mman.buffer_funcs_ring = &adev->sdma.instance[0].ring; in si_dma_set_buffer_funcs()
/drivers/gpu/drm/qxl/
A Dqxl_ttm.c42 struct qxl_mman *mman; in qxl_get_qdev() local
45 mman = container_of(bdev, struct qxl_mman, bdev); in qxl_get_qdev()
46 qdev = container_of(mman, struct qxl_device, mman); in qxl_get_qdev()
187 return ttm_range_man_init(&qdev->mman.bdev, type, false, size); in qxl_ttm_init_mem_type()
196 r = ttm_device_init(&qdev->mman.bdev, &qxl_bo_driver, NULL, in qxl_ttm_init()
228 ttm_range_man_fini(&qdev->mman.bdev, TTM_PL_VRAM); in qxl_ttm_fini()
229 ttm_range_man_fini(&qdev->mman.bdev, TTM_PL_PRIV); in qxl_ttm_fini()
230 ttm_device_fini(&qdev->mman.bdev); in qxl_ttm_fini()
237 ttm_resource_manager_create_debugfs(ttm_manager_type(&qdev->mman.bdev, in qxl_ttm_debugfs_init()
240 ttm_resource_manager_create_debugfs(ttm_manager_type(&qdev->mman.bdev, in qxl_ttm_debugfs_init()
A Dqxl_object.c139 r = ttm_bo_init_reserved(&qdev->mman.bdev, &bo->tbo, type, in qxl_bo_create()
408 man = ttm_manager_type(&qdev->mman.bdev, TTM_PL_PRIV); in qxl_surf_evict()
409 return ttm_resource_manager_evict_all(&qdev->mman.bdev, man); in qxl_surf_evict()
416 man = ttm_manager_type(&qdev->mman.bdev, TTM_PL_VRAM); in qxl_vram_evict()
417 return ttm_resource_manager_evict_all(&qdev->mman.bdev, man); in qxl_vram_evict()
/drivers/gpu/drm/radeon/
A Dradeon_ttm.c62 struct radeon_mman *mman; in radeon_get_rdev() local
65 mman = container_of(bdev, struct radeon_mman, bdev); in radeon_get_rdev()
66 rdev = container_of(mman, struct radeon_device, mman); in radeon_get_rdev()
78 return ttm_range_man_init(&rdev->mman.bdev, TTM_PL_TT, in radeon_ttm_init_gtt()
570 return ttm_pool_free(&rdev->mman.bdev.pool, ttm); in radeon_ttm_tt_unpopulate()
692 rdev->mman.initialized = true; in radeon_ttm_init()
737 if (!rdev->mman.initialized) in radeon_ttm_fini()
749 ttm_range_man_fini(&rdev->mman.bdev, TTM_PL_TT); in radeon_ttm_fini()
750 ttm_device_fini(&rdev->mman.bdev); in radeon_ttm_fini()
752 rdev->mman.initialized = false; in radeon_ttm_fini()
[all …]
A Dradeon_object.c204 r = ttm_bo_init_validate(&rdev->mman.bdev, &bo->tbo, type, in radeon_bo_create()
349 struct ttm_device *bdev = &rdev->mman.bdev; in radeon_bo_evict_vram()
418 ttm_manager_type(&rdev->mman.bdev, TTM_PL_VRAM); in radeon_bo_get_threshold_for_moves()
A Dradeon_kms.c547 man = ttm_manager_type(&rdev->mman.bdev, TTM_PL_VRAM); in radeon_info_ioctl()
553 man = ttm_manager_type(&rdev->mman.bdev, TTM_PL_TT); in radeon_info_ioctl()
/drivers/gpu/drm/amd/amdkfd/
A Dkfd_migrate.c64 num_dw = ALIGN(adev->mman.buffer_funcs->copy_num_dw, 8); in svm_migrate_gart_map()
67 r = amdgpu_job_alloc_with_ib(adev, &adev->mman.high_pr, in svm_migrate_gart_map()
130 struct amdgpu_ring *ring = adev->mman.buffer_funcs_ring; in svm_migrate_copy_memory_gart()
136 mutex_lock(&adev->mman.gtt_window_lock); in svm_migrate_copy_memory_gart()
172 mutex_unlock(&adev->mman.gtt_window_lock); in svm_migrate_copy_memory_gart()
/drivers/gpu/drm/i915/selftests/
A Di915_live_selftests.h33 selftest(mman, i915_gem_mman_live_selftests)

Completed in 78 milliseconds

123