Lines Matching refs:mob
87 struct vmw_mob *mob);
88 static void vmw_mob_pt_setup(struct vmw_mob *mob,
124 struct vmw_mob *mob; in vmw_setup_otable_base() local
135 mob = vmw_mob_create(otable->size >> PAGE_SHIFT); in vmw_setup_otable_base()
136 if (unlikely(mob == NULL)) { in vmw_setup_otable_base()
142 mob->pt_level = VMW_MOBFMT_PTDEPTH_0; in vmw_setup_otable_base()
143 mob->pt_root_page = vmw_piter_dma_addr(&iter); in vmw_setup_otable_base()
145 ret = vmw_mob_pt_populate(dev_priv, mob); in vmw_setup_otable_base()
149 vmw_mob_pt_setup(mob, iter, otable->size >> PAGE_SHIFT); in vmw_setup_otable_base()
150 mob->pt_level += VMW_MOBFMT_PTDEPTH_1 - SVGA3D_MOBFMT_PT_1; in vmw_setup_otable_base()
163 cmd->body.baseAddress = mob->pt_root_page >> PAGE_SHIFT; in vmw_setup_otable_base()
166 cmd->body.ptDepth = mob->pt_level; in vmw_setup_otable_base()
173 BUG_ON(mob->pt_level == VMW_MOBFMT_PTDEPTH_2); in vmw_setup_otable_base()
176 otable->page_table = mob; in vmw_setup_otable_base()
182 vmw_mob_destroy(mob); in vmw_setup_otable_base()
393 struct vmw_mob *mob = kzalloc(sizeof(*mob), GFP_KERNEL); in vmw_mob_create() local
395 if (unlikely(!mob)) in vmw_mob_create()
398 mob->num_pages = vmw_mob_calculate_pt_pages(data_pages); in vmw_mob_create()
400 return mob; in vmw_mob_create()
414 struct vmw_mob *mob) in vmw_mob_pt_populate() argument
416 BUG_ON(mob->pt_bo != NULL); in vmw_mob_pt_populate()
418 return vmw_bo_create_and_populate(dev_priv, mob->num_pages * PAGE_SIZE, in vmw_mob_pt_populate()
420 &mob->pt_bo); in vmw_mob_pt_populate()
496 static void vmw_mob_pt_setup(struct vmw_mob *mob, in vmw_mob_pt_setup() argument
501 struct ttm_buffer_object *bo = &mob->pt_bo->tbo; in vmw_mob_pt_setup()
515 mob->pt_level = 0; in vmw_mob_pt_setup()
517 ++mob->pt_level; in vmw_mob_pt_setup()
518 BUG_ON(mob->pt_level > 2); in vmw_mob_pt_setup()
526 mob->pt_root_page = vmw_piter_dma_addr(&save_pt_iter); in vmw_mob_pt_setup()
535 void vmw_mob_destroy(struct vmw_mob *mob) in vmw_mob_destroy() argument
537 if (mob->pt_bo) { in vmw_mob_destroy()
538 vmw_bo_unpin_unlocked(&mob->pt_bo->tbo); in vmw_mob_destroy()
539 vmw_bo_unreference(&mob->pt_bo); in vmw_mob_destroy()
541 kfree(mob); in vmw_mob_destroy()
551 struct vmw_mob *mob) in vmw_mob_unbind() argument
558 struct ttm_buffer_object *bo = &mob->pt_bo->tbo; in vmw_mob_unbind()
572 cmd->body.mobid = mob->id; in vmw_mob_unbind()
599 struct vmw_mob *mob, in vmw_mob_bind() argument
612 mob->id = mob_id; in vmw_mob_bind()
618 mob->pt_level = VMW_MOBFMT_PTDEPTH_0; in vmw_mob_bind()
619 mob->pt_root_page = vmw_piter_dma_addr(&data_iter); in vmw_mob_bind()
620 } else if (unlikely(mob->pt_bo == NULL)) { in vmw_mob_bind()
621 ret = vmw_mob_pt_populate(dev_priv, mob); in vmw_mob_bind()
625 vmw_mob_pt_setup(mob, data_iter, num_data_pages); in vmw_mob_bind()
627 mob->pt_level += VMW_MOBFMT_PTDEPTH_1 - SVGA3D_MOBFMT_PT_1; in vmw_mob_bind()
639 cmd->body.ptDepth = mob->pt_level; in vmw_mob_bind()
640 cmd->body.base = mob->pt_root_page >> PAGE_SHIFT; in vmw_mob_bind()
650 vmw_bo_unpin_unlocked(&mob->pt_bo->tbo); in vmw_mob_bind()
651 vmw_bo_unreference(&mob->pt_bo); in vmw_mob_bind()