Lines Matching refs:ttm

407 	if (WARN(amdgpu_ttm_tt_get_usermm(bo->tbo.ttm),  in amdgpu_amdkfd_bo_validate()
552 struct ttm_tt *src_ttm = mem->bo->tbo.ttm; in kfd_mem_dmamap_userptr()
553 struct ttm_tt *ttm = bo->tbo.ttm; in kfd_mem_dmamap_userptr() local
556 if (WARN_ON(ttm->num_pages != src_ttm->num_pages)) in kfd_mem_dmamap_userptr()
559 ttm->sg = kmalloc(sizeof(*ttm->sg), GFP_KERNEL); in kfd_mem_dmamap_userptr()
560 if (unlikely(!ttm->sg)) in kfd_mem_dmamap_userptr()
564 ret = sg_alloc_table_from_pages(ttm->sg, src_ttm->pages, in kfd_mem_dmamap_userptr()
565 ttm->num_pages, 0, in kfd_mem_dmamap_userptr()
566 (u64)ttm->num_pages << PAGE_SHIFT, in kfd_mem_dmamap_userptr()
571 ret = dma_map_sgtable(adev->dev, ttm->sg, direction, 0); in kfd_mem_dmamap_userptr()
583 dma_unmap_sgtable(adev->dev, ttm->sg, direction, 0); in kfd_mem_dmamap_userptr()
586 sg_free_table(ttm->sg); in kfd_mem_dmamap_userptr()
588 kfree(ttm->sg); in kfd_mem_dmamap_userptr()
589 ttm->sg = NULL; in kfd_mem_dmamap_userptr()
636 struct ttm_tt *ttm = bo->tbo.ttm; in kfd_mem_dmamap_sg_bo() local
644 if (unlikely(ttm->sg)) { in kfd_mem_dmamap_sg_bo()
661 ttm->sg = create_sg_table(dma_addr, mem->bo->tbo.sg->sgl->length); in kfd_mem_dmamap_sg_bo()
662 if (unlikely(!ttm->sg)) { in kfd_mem_dmamap_sg_bo()
675 sg_free_table(ttm->sg); in kfd_mem_dmamap_sg_bo()
676 kfree(ttm->sg); in kfd_mem_dmamap_sg_bo()
677 ttm->sg = NULL; in kfd_mem_dmamap_sg_bo()
713 struct ttm_tt *ttm = bo->tbo.ttm; in kfd_mem_dmaunmap_userptr() local
715 if (unlikely(!ttm->sg)) in kfd_mem_dmaunmap_userptr()
721 dma_unmap_sgtable(adev->dev, ttm->sg, direction, 0); in kfd_mem_dmaunmap_userptr()
722 sg_free_table(ttm->sg); in kfd_mem_dmaunmap_userptr()
723 kfree(ttm->sg); in kfd_mem_dmaunmap_userptr()
724 ttm->sg = NULL; in kfd_mem_dmaunmap_userptr()
759 struct ttm_tt *ttm = bo->tbo.ttm; in kfd_mem_dmaunmap_sg_bo() local
762 if (unlikely(!ttm->sg)) { in kfd_mem_dmaunmap_sg_bo()
772 dma_unmap_resource(adev->dev, ttm->sg->sgl->dma_address, in kfd_mem_dmaunmap_sg_bo()
773 ttm->sg->sgl->length, dir, DMA_ATTR_SKIP_CPU_SYNC); in kfd_mem_dmaunmap_sg_bo()
774 sg_free_table(ttm->sg); in kfd_mem_dmaunmap_sg_bo()
775 kfree(ttm->sg); in kfd_mem_dmaunmap_sg_bo()
776 ttm->sg = NULL; in kfd_mem_dmaunmap_sg_bo()
900 (amdgpu_ttm_tt_get_usermm(mem->bo->tbo.ttm) && reuse_dmamap(adev, bo_adev)) || in kfd_mem_attach()
915 } else if (amdgpu_ttm_tt_get_usermm(mem->bo->tbo.ttm)) { in kfd_mem_attach()
1075 ret = amdgpu_ttm_tt_get_user_pages(bo, bo->tbo.ttm->pages, &range); in init_user_pages()
1096 amdgpu_ttm_tt_get_user_pages_done(bo->tbo.ttm, range); in init_user_pages()
1780 bo->tbo.ttm->sg = sg; in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu()
1892 if (amdgpu_ttm_tt_get_usermm(mem->bo->tbo.ttm)) { in amdgpu_amdkfd_gpuvm_free_memory_of_gpu()
1895 amdgpu_ttm_tt_discard_user_pages(mem->bo->tbo.ttm, mem->range); in amdgpu_amdkfd_gpuvm_free_memory_of_gpu()
1995 if (amdgpu_ttm_tt_get_usermm(bo->tbo.ttm)) { in amdgpu_amdkfd_gpuvm_map_memory_to_gpu()
2026 if (amdgpu_ttm_tt_get_usermm(bo->tbo.ttm) && in amdgpu_amdkfd_gpuvm_map_memory_to_gpu()
2089 if (entry->bo_va->base.bo->tbo.ttm && in amdgpu_amdkfd_gpuvm_dmaunmap_mem()
2090 !entry->bo_va->base.bo->tbo.ttm->sg) in amdgpu_amdkfd_gpuvm_dmaunmap_mem()
2241 if (amdgpu_ttm_tt_get_usermm(bo->tbo.ttm)) { in amdgpu_amdkfd_gpuvm_map_gtt_bo_to_kernel()
2522 amdgpu_ttm_tt_discard_user_pages(bo->tbo.ttm, mem->range); in update_invalid_user_pages()
2547 ret = amdgpu_ttm_tt_get_user_pages(bo, bo->tbo.ttm->pages, in update_invalid_user_pages()
2653 if (bo->tbo.ttm->pages[0]) { in validate_invalid_user_pages()
2716 mem->bo->tbo.ttm, mem->range); in confirm_valid_user_pages_locked()