Lines Matching refs:ttm
444 container_of(bo->ttm, struct vmw_ttm_tt, dma_ttm); in vmw_bo_sg_table()
451 struct ttm_tt *ttm, struct ttm_resource *bo_mem) in vmw_ttm_bind() argument
454 container_of(ttm, struct vmw_ttm_tt, dma_ttm); in vmw_ttm_bind()
473 ttm->num_pages, vmw_be->gmr_id); in vmw_ttm_bind()
478 vmw_mob_create(ttm->num_pages); in vmw_ttm_bind()
484 &vmw_be->vsgt, ttm->num_pages, in vmw_ttm_bind()
495 struct ttm_tt *ttm) in vmw_ttm_unbind() argument
498 container_of(ttm, struct vmw_ttm_tt, dma_ttm); in vmw_ttm_unbind()
520 static void vmw_ttm_destroy(struct ttm_device *bdev, struct ttm_tt *ttm) in vmw_ttm_destroy() argument
523 container_of(ttm, struct vmw_ttm_tt, dma_ttm); in vmw_ttm_destroy()
526 ttm_tt_fini(ttm); in vmw_ttm_destroy()
535 struct ttm_tt *ttm, struct ttm_operation_ctx *ctx) in vmw_ttm_populate() argument
541 if (ttm_tt_is_populated(ttm)) in vmw_ttm_populate()
544 ret = ttm_pool_alloc(&bdev->pool, ttm, ctx); in vmw_ttm_populate()
548 for (i = 0; i < ttm->num_pages; ++i) { in vmw_ttm_populate()
549 ret = ttm_mem_global_alloc_page(&ttm_mem_glob, ttm->pages[i], in vmw_ttm_populate()
558 ttm_mem_global_free_page(&ttm_mem_glob, ttm->pages[i], in vmw_ttm_populate()
560 ttm_pool_free(&bdev->pool, ttm); in vmw_ttm_populate()
565 struct ttm_tt *ttm) in vmw_ttm_unpopulate() argument
567 struct vmw_ttm_tt *vmw_tt = container_of(ttm, struct vmw_ttm_tt, in vmw_ttm_unpopulate()
571 vmw_ttm_unbind(bdev, ttm); in vmw_ttm_unpopulate()
580 for (i = 0; i < ttm->num_pages; ++i) in vmw_ttm_unpopulate()
581 ttm_mem_global_free_page(&ttm_mem_glob, ttm->pages[i], in vmw_ttm_unpopulate()
584 ttm_pool_free(&bdev->pool, ttm); in vmw_ttm_unpopulate()
684 ret = vmw_ttm_bind(bo->bdev, bo->ttm, new_mem); in vmw_move()
700 vmw_ttm_unbind(bo->bdev, bo->ttm); in vmw_move()
746 ret = vmw_ttm_populate(bo->bdev, bo->ttm, &ctx); in vmw_bo_create_and_populate()
749 container_of(bo->ttm, struct vmw_ttm_tt, dma_ttm); in vmw_bo_create_and_populate()