Home
last modified time | relevance | path

Searched refs:xe_assert (Results 1 – 25 of 38) sorted by relevance

12

/linux/drivers/gpu/drm/xe/
A Dxe_sriov_pf_helpers.h26 xe_assert((xe), (vfid) <= xe_sriov_pf_get_totalvfs(xe))
36 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_get_totalvfs()
42 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_master_mutex()
A Dxe_bo.c147 xe_assert(xe, resource_is_vram(res)); in res_to_mem_region()
562 xe_assert(xe, attach); in xe_bo_move_dmabuf()
563 xe_assert(xe, ttm_bo->ttm); in xe_bo_move_dmabuf()
759 xe_assert(xe, migrate); in xe_bo_move()
798 xe_assert(xe, new_mem->start == in xe_bo_move()
1020 xe_assert(xe, locked); in xe_ttm_bo_lock_in_destructor()
1478 xe_assert(xe, tile); in __xe_bo_create_locked()
1657 xe_assert(xe, IS_DGFX(xe)); in xe_managed_bo_reinit_in_vram()
1700 xe_assert(xe, !bo->vm); in xe_bo_pin_external()
1800 xe_assert(xe, !bo->vm); in xe_bo_unpin_external()
[all …]
A Dxe_pci_sriov.c71 xe_assert(xe, IS_SRIOV_PF(xe)); in pf_enable_vfs()
72 xe_assert(xe, num_vfs > 0); in pf_enable_vfs()
73 xe_assert(xe, num_vfs <= total_vfs); in pf_enable_vfs()
114 xe_assert(xe, IS_SRIOV_PF(xe)); in pf_disable_vfs()
A Dxe_sriov_pf.c56 xe_assert(xe, totalvfs <= U16_MAX); in xe_sriov_pf_readiness()
83 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_init_early()
99 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_print_vfs_summary()
A Dxe_sriov.c77 xe_assert(xe, !xe->sriov.__mode); in xe_sriov_probe_early()
79 xe_assert(xe, xe->sriov.__mode); in xe_sriov_probe_early()
115 xe_assert(xe, !xe->sriov.wq); in xe_sriov_init()
A Dxe_hw_engine_group.c152 xe_assert(xe, group); in xe_hw_engine_group_add_exec_queue()
153 xe_assert(xe, !(q->flags & EXEC_QUEUE_FLAG_VM)); in xe_hw_engine_group_add_exec_queue()
154 xe_assert(xe, q->vm); in xe_hw_engine_group_add_exec_queue()
191 xe_assert(xe, group); in xe_hw_engine_group_del_exec_queue()
192 xe_assert(xe, q->vm); in xe_hw_engine_group_del_exec_queue()
A Dxe_pat.c198 xe_assert(xe, !err); in xelp_dump()
229 xe_assert(xe, !err); in xehp_dump()
258 xe_assert(xe, !err); in xehpc_dump()
292 xe_assert(xe, !err); in xelpg_dump()
374 xe_assert(xe, !err); in xe2_dump()
449 xe_assert(xe, !xe->pat.ops || xe->pat.ops->dump); in xe_pat_init_early()
450 xe_assert(xe, !xe->pat.ops || xe->pat.ops->program_graphics); in xe_pat_init_early()
451 xe_assert(xe, !xe->pat.ops || MEDIA_VER(xe) < 13 || xe->pat.ops->program_media); in xe_pat_init_early()
A Dxe_gt_sriov_vf_debugfs.c58 xe_assert(xe, IS_SRIOV_VF(xe)); in xe_gt_sriov_vf_debugfs_register()
59 xe_assert(xe, root->d_inode->i_private == gt); in xe_gt_sriov_vf_debugfs_register()
A Dxe_gt_ccs_mode.c25 xe_assert(xe, xe_gt_ccs_mode_enabled(gt)); in __xe_gt_apply_ccs_mode()
27 xe_assert(xe, num_engines && num_engines <= num_slices); in __xe_gt_apply_ccs_mode()
28 xe_assert(xe, !(num_slices % num_engines)); in __xe_gt_apply_ccs_mode()
A Dxe_guc_log.c58 xe_assert(xe, log->bo); in xe_guc_log_print()
63 xe_assert(xe, !(size % (DW_PER_READ * sizeof(u32)))); in xe_guc_log_print()
A Dxe_guc.c48 xe_assert(xe, addr >= xe_wopcm_size(guc_to_xe(guc))); in guc_bo_ggtt_addr()
49 xe_assert(xe, addr < GUC_GGTT_TOP); in guc_bo_ggtt_addr()
50 xe_assert(xe, bo->size <= GUC_GGTT_TOP - addr); in guc_bo_ggtt_addr()
937 xe_assert(xe, !xe_guc_ct_enabled(&guc->ct)); in xe_guc_mmio_send_recv()
938 xe_assert(xe, len); in xe_guc_mmio_send_recv()
939 xe_assert(xe, len <= VF_SW_FLAG_COUNT); in xe_guc_mmio_send_recv()
940 xe_assert(xe, len <= MED_VF_SW_FLAG_COUNT); in xe_guc_mmio_send_recv()
941 xe_assert(xe, FIELD_GET(GUC_HXG_MSG_0_ORIGIN, request[0]) == in xe_guc_mmio_send_recv()
943 xe_assert(xe, FIELD_GET(GUC_HXG_MSG_0_TYPE, request[0]) == in xe_guc_mmio_send_recv()
1065 xe_assert(xe, len <= 2); in guc_self_cfg()
[all …]
A Dxe_guc_submit.c419 xe_assert(xe, exec_queue_registered(q)); in init_policies()
458 xe_assert(xe, xe_exec_queue_is_parallel(q)); in __register_mlrc_exec_queue()
481 xe_assert(xe, len <= MAX_MLRC_REG_SIZE); in __register_mlrc_exec_queue()
515 xe_assert(xe, !exec_queue_registered(q)); in register_exec_queue()
645 xe_assert(xe, i == wqi_size / sizeof(u32)); in wq_item_append()
651 xe_assert(xe, q->guc->wqi_tail <= WQ_SIZE); in wq_item_append()
671 xe_assert(xe, exec_queue_registered(q)); in submit_exec_queue()
868 xe_assert(xe, xe_exec_queue_is_lr(q)); in xe_guc_exec_queue_lr_cleanup()
1299 xe_assert(xe, q->guc->suspend_pending); in suspend_fence_signal()
1615 xe_assert(xe, !q->guc->suspend_pending); in guc_exec_queue_resume()
[all …]
A Dxe_uc_fw.c229 xe_assert(xe, uc_fw->type < ARRAY_SIZE(blobs_all)); in uc_fw_auto_select()
292 xe_assert(xe, !(size % 4)); in xe_uc_fw_copy_rsa()
293 xe_assert(xe, xe_uc_fw_is_available(uc_fw)); in xe_uc_fw_copy_rsa()
470 xe_assert(xe, manifest_entry); in parse_cpd_header()
522 xe_assert(xe, xe->info.platform != XE_DG2); in parse_cpd_header()
661 xe_assert(xe, !uc_fw->status); in uc_fw_request()
662 xe_assert(xe, !uc_fw->path); in uc_fw_request()
855 xe_assert(xe, !xe_uc_fw_is_loaded(uc_fw)); in xe_uc_fw_upload()
A Dxe_vm.c74 xe_assert(xe, xe_vma_is_userptr(vma)); in xe_vma_userptr_pin_pages()
170 xe_assert(vm->xe, link != list); in arm_preempt_fences()
818 xe_assert(vm->xe, vma->tile_present); in xe_vm_rebind()
921 xe_assert(vm->xe, start < end); in xe_vma_create()
922 xe_assert(vm->xe, end < vm->size); in xe_vma_create()
1301 xe_assert(xe, !(flags & ~XE_PTE_PS64)); in xelp_pte_encode_addr()
1620 xe_assert(xe, xe->info.has_asid); in xe_vm_close_and_put()
1624 xe_assert(xe, lookup == vm); in xe_vm_close_and_put()
1643 xe_assert(xe, !vm->size); in vm_destroy_work_func()
1938 xe_assert(vm->xe, bo); in vm_bind_ioctl_ops_create()
[all …]
A Dxe_hmm.c95 xe_assert(xe, !is_device_private_page(pages[i])); in xe_build_sg()
133 xe_assert(xe, userptr->sg); in xe_hmm_userptr_free_sg()
A Dxe_sriov.h24 xe_assert(xe, xe->sriov.__mode); in xe_device_sriov_mode()
A Dxe_vm.h172 xe_assert(xe_vma_vm(vma)->xe, xe_vma_is_userptr(vma)); in to_userptr_vma()
219 xe_assert(vm->xe, xe_vm_in_preempt_fence_mode(vm)); in xe_vm_queue_rebind_worker()
A Dxe_drm_client.c158 xe_assert(xe, !kref_read(&bo->ttm.base.refcount)); in xe_drm_client_remove_bo()
246 xe_assert(xef->xe, !list_empty(&bo->client_link)); in show_meminfo()
A Dxe_mocs.c651 xe_assert(xe, info->unused_entries_index != 0); in get_mocs_settings()
653 xe_assert(xe, info->ops && info->ops->dump); in get_mocs_settings()
654 xe_assert(xe, info->table_size <= info->num_mocs_regs); in get_mocs_settings()
794 xe_assert(xe, !ret); in xe_mocs_dump()
A Dxe_gt_topology.c114 xe_assert(xe, find_last_bit(pattern, XE_MAX_L3_BANK_MASK_BITS) < patternbits || in gen_l3_mask_from_pattern()
116 xe_assert(xe, !mask || patternbits * (__fls(mask) + 1) <= XE_MAX_L3_BANK_MASK_BITS); in gen_l3_mask_from_pattern()
A Dxe_assert.h108 #define xe_assert(xe, condition) xe_assert_msg((xe), condition, "") macro
A Dxe_irq.c560 xe_assert(xe, IS_SRIOV_VF(xe)); in vf_irq_reset()
565 xe_assert(xe, xe_device_has_memirq(xe)); in vf_irq_reset()
617 xe_assert(xe, xe_device_has_memirq(xe)); in vf_irq_postinstall()
A Dxe_execlist.c338 xe_assert(xe, !xe_device_uc_enabled(xe)); in execlist_exec_queue_init()
387 xe_assert(xe, !xe_device_uc_enabled(xe)); in execlist_exec_queue_fini_async()
A Dxe_migrate.c153 xe_assert(xe, IS_ALIGNED(xe->mem.vram.usable_size, SZ_2M)); in xe_migrate_program_identity()
181 xe_assert(xe, pos == vram_limit); in xe_migrate_program_identity()
313 xe_assert(xe, xe->mem.vram.actual_physical_size <= in xe_migrate_prepare_vm()
326 xe_assert(xe, xe->mem.vram.actual_physical_size <= (MAX_NUM_PTE - in xe_migrate_prepare_vm()
600 xe_assert(xe, (va & (SZ_64K - 1)) == in emit_pte()
846 xe_assert(xe, IS_ALIGNED(ccs_it.start, PAGE_SIZE)); in xe_migrate_copy()
/linux/Documentation/gpu/xe/
A Dxe_debugging.rst7 .. kernel-doc:: drivers/gpu/drm/xe/xe_assert.h

Completed in 47 milliseconds

12