Lines Matching refs:mob
87 struct vmw_mob *mob);
88 static void vmw_mob_pt_setup(struct vmw_mob *mob,
124 struct vmw_mob *mob; in vmw_setup_otable_base() local
135 mob = vmw_mob_create(otable->size >> PAGE_SHIFT); in vmw_setup_otable_base()
136 if (unlikely(mob == NULL)) { in vmw_setup_otable_base()
142 mob->pt_level = VMW_MOBFMT_PTDEPTH_0; in vmw_setup_otable_base()
143 mob->pt_root_page = vmw_piter_dma_addr(&iter); in vmw_setup_otable_base()
145 ret = vmw_mob_pt_populate(dev_priv, mob); in vmw_setup_otable_base()
149 vmw_mob_pt_setup(mob, iter, otable->size >> PAGE_SHIFT); in vmw_setup_otable_base()
150 mob->pt_level += VMW_MOBFMT_PTDEPTH_1 - SVGA3D_MOBFMT_PT_1; in vmw_setup_otable_base()
163 cmd->body.baseAddress = mob->pt_root_page >> PAGE_SHIFT; in vmw_setup_otable_base()
166 cmd->body.ptDepth = mob->pt_level; in vmw_setup_otable_base()
173 BUG_ON(mob->pt_level == VMW_MOBFMT_PTDEPTH_2); in vmw_setup_otable_base()
176 otable->page_table = mob; in vmw_setup_otable_base()
182 vmw_mob_destroy(mob); in vmw_setup_otable_base()
394 struct vmw_mob *mob = kzalloc(sizeof(*mob), GFP_KERNEL); in vmw_mob_create() local
396 if (unlikely(!mob)) in vmw_mob_create()
399 mob->num_pages = vmw_mob_calculate_pt_pages(data_pages); in vmw_mob_create()
401 return mob; in vmw_mob_create()
415 struct vmw_mob *mob) in vmw_mob_pt_populate() argument
417 BUG_ON(mob->pt_bo != NULL); in vmw_mob_pt_populate()
419 return vmw_bo_create_and_populate(dev_priv, mob->num_pages * PAGE_SIZE, in vmw_mob_pt_populate()
421 &mob->pt_bo); in vmw_mob_pt_populate()
497 static void vmw_mob_pt_setup(struct vmw_mob *mob, in vmw_mob_pt_setup() argument
502 struct ttm_buffer_object *bo = &mob->pt_bo->tbo; in vmw_mob_pt_setup()
516 mob->pt_level = 0; in vmw_mob_pt_setup()
518 ++mob->pt_level; in vmw_mob_pt_setup()
519 BUG_ON(mob->pt_level > 2); in vmw_mob_pt_setup()
527 mob->pt_root_page = vmw_piter_dma_addr(&save_pt_iter); in vmw_mob_pt_setup()
536 void vmw_mob_destroy(struct vmw_mob *mob) in vmw_mob_destroy() argument
538 if (mob->pt_bo) { in vmw_mob_destroy()
539 vmw_bo_unpin_unlocked(&mob->pt_bo->tbo); in vmw_mob_destroy()
540 vmw_bo_unreference(&mob->pt_bo); in vmw_mob_destroy()
542 kfree(mob); in vmw_mob_destroy()
552 struct vmw_mob *mob) in vmw_mob_unbind() argument
559 struct ttm_buffer_object *bo = &mob->pt_bo->tbo; in vmw_mob_unbind()
573 cmd->body.mobid = mob->id; in vmw_mob_unbind()
600 struct vmw_mob *mob, in vmw_mob_bind() argument
613 mob->id = mob_id; in vmw_mob_bind()
619 mob->pt_level = VMW_MOBFMT_PTDEPTH_0; in vmw_mob_bind()
620 mob->pt_root_page = vmw_piter_dma_addr(&data_iter); in vmw_mob_bind()
621 } else if (unlikely(mob->pt_bo == NULL)) { in vmw_mob_bind()
622 ret = vmw_mob_pt_populate(dev_priv, mob); in vmw_mob_bind()
626 vmw_mob_pt_setup(mob, data_iter, num_data_pages); in vmw_mob_bind()
628 mob->pt_level += VMW_MOBFMT_PTDEPTH_1 - SVGA3D_MOBFMT_PT_1; in vmw_mob_bind()
640 cmd->body.ptDepth = mob->pt_level; in vmw_mob_bind()
641 cmd->body.base = mob->pt_root_page >> PAGE_SHIFT; in vmw_mob_bind()
651 vmw_bo_unpin_unlocked(&mob->pt_bo->tbo); in vmw_mob_bind()
652 vmw_bo_unreference(&mob->pt_bo); in vmw_mob_bind()