Lines Matching refs:ib

361 				struct radeon_ib *ib,  in radeon_vm_set_pages()  argument
370 radeon_asic_vm_copy_pages(rdev, ib, pe, src, count); in radeon_vm_set_pages()
373 radeon_asic_vm_write_pages(rdev, ib, pe, addr, in radeon_vm_set_pages()
377 radeon_asic_vm_set_pages(rdev, ib, pe, addr, in radeon_vm_set_pages()
392 struct radeon_ib ib; in radeon_vm_clear_bo() local
408 r = radeon_ib_get(rdev, R600_RING_TYPE_DMA_INDEX, &ib, NULL, 256); in radeon_vm_clear_bo()
412 ib.length_dw = 0; in radeon_vm_clear_bo()
414 radeon_vm_set_pages(rdev, &ib, addr, 0, entries, 0, 0); in radeon_vm_clear_bo()
415 radeon_asic_vm_pad_ib(rdev, &ib); in radeon_vm_clear_bo()
416 WARN_ON(ib.length_dw > 64); in radeon_vm_clear_bo()
418 r = radeon_ib_schedule(rdev, &ib, NULL, false); in radeon_vm_clear_bo()
422 ib.fence->is_vm_update = true; in radeon_vm_clear_bo()
423 radeon_bo_fence(bo, ib.fence, false); in radeon_vm_clear_bo()
426 radeon_ib_free(rdev, &ib); in radeon_vm_clear_bo()
648 struct radeon_ib ib; in radeon_vm_update_page_directory() local
661 r = radeon_ib_get(rdev, R600_RING_TYPE_DMA_INDEX, &ib, NULL, ndw * 4); in radeon_vm_update_page_directory()
664 ib.length_dw = 0; in radeon_vm_update_page_directory()
684 radeon_vm_set_pages(rdev, &ib, last_pde, in radeon_vm_update_page_directory()
698 radeon_vm_set_pages(rdev, &ib, last_pde, last_pt, count, in radeon_vm_update_page_directory()
701 if (ib.length_dw != 0) { in radeon_vm_update_page_directory()
702 radeon_asic_vm_pad_ib(rdev, &ib); in radeon_vm_update_page_directory()
704 radeon_sync_resv(rdev, &ib.sync, pd->tbo.base.resv, true); in radeon_vm_update_page_directory()
705 WARN_ON(ib.length_dw > ndw); in radeon_vm_update_page_directory()
706 r = radeon_ib_schedule(rdev, &ib, NULL, false); in radeon_vm_update_page_directory()
708 radeon_ib_free(rdev, &ib); in radeon_vm_update_page_directory()
711 ib.fence->is_vm_update = true; in radeon_vm_update_page_directory()
712 radeon_bo_fence(pd, ib.fence, false); in radeon_vm_update_page_directory()
714 radeon_ib_free(rdev, &ib); in radeon_vm_update_page_directory()
732 struct radeon_ib *ib, in radeon_vm_frag_ptes() argument
772 radeon_vm_set_pages(rdev, ib, pe_start, addr, count, in radeon_vm_frag_ptes()
780 radeon_vm_set_pages(rdev, ib, pe_start, addr, count, in radeon_vm_frag_ptes()
787 radeon_vm_set_pages(rdev, ib, frag_start, addr, count, in radeon_vm_frag_ptes()
794 radeon_vm_set_pages(rdev, ib, frag_end, addr, count, in radeon_vm_frag_ptes()
816 struct radeon_ib *ib, in radeon_vm_update_ptes() argument
833 radeon_sync_resv(rdev, &ib->sync, pt->tbo.base.resv, true); in radeon_vm_update_ptes()
849 radeon_vm_frag_ptes(rdev, ib, last_pte, in radeon_vm_update_ptes()
866 radeon_vm_frag_ptes(rdev, ib, last_pte, in radeon_vm_update_ptes()
916 struct radeon_ib ib; in radeon_vm_bo_update() local
999 r = radeon_ib_get(rdev, R600_RING_TYPE_DMA_INDEX, &ib, NULL, ndw * 4); in radeon_vm_bo_update()
1002 ib.length_dw = 0; in radeon_vm_bo_update()
1008 radeon_sync_fence(&ib.sync, vm->ids[i].last_id_use); in radeon_vm_bo_update()
1011 r = radeon_vm_update_ptes(rdev, vm, &ib, bo_va->it.start, in radeon_vm_bo_update()
1015 radeon_ib_free(rdev, &ib); in radeon_vm_bo_update()
1019 radeon_asic_vm_pad_ib(rdev, &ib); in radeon_vm_bo_update()
1020 WARN_ON(ib.length_dw > ndw); in radeon_vm_bo_update()
1022 r = radeon_ib_schedule(rdev, &ib, NULL, false); in radeon_vm_bo_update()
1024 radeon_ib_free(rdev, &ib); in radeon_vm_bo_update()
1027 ib.fence->is_vm_update = true; in radeon_vm_bo_update()
1028 radeon_vm_fence_pts(vm, bo_va->it.start, bo_va->it.last + 1, ib.fence); in radeon_vm_bo_update()
1030 bo_va->last_pt_update = radeon_fence_ref(ib.fence); in radeon_vm_bo_update()
1031 radeon_ib_free(rdev, &ib); in radeon_vm_bo_update()