| /drivers/gpu/drm/xe/ |
| A D | xe_sriov_pf_service.c | 29 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_service_init() 48 xe_assert(xe, IS_SRIOV_PF(xe)); in pf_negotiate_version() 49 xe_assert(xe, base.major); in pf_negotiate_version() 50 xe_assert(xe, base.major <= latest.major); in pf_negotiate_version() 51 xe_assert(xe, (base.major < latest.major) || (base.minor <= latest.minor)); in pf_negotiate_version() 77 xe_assert(xe, base.major == latest.major); in pf_negotiate_version() 90 xe_assert(xe, major || minor); in pf_connect() 200 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_service_print_versions()
|
| A D | xe_sriov_pf_helpers.h | 26 xe_assert((xe), (vfid) <= xe_sriov_pf_get_totalvfs(xe)) 36 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_get_totalvfs() 42 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_master_mutex()
|
| A D | xe_hmm.c | 44 xe_assert(xe, hmm_pfn & HMM_PFN_VALID); in xe_alloc_sg() 111 xe_assert(xe, !is_device_private_page(page)); in xe_build_sg() 118 xe_assert(xe, i >= npages); in xe_build_sg() 124 xe_assert(xe, i < npages); in xe_build_sg() 143 xe_assert(vm->xe, !userptr->mapped); in xe_hmm_userptr_set_mapped() 185 xe_assert(xe_vma_vm(&uvma->vma)->xe, userptr->sg); in xe_hmm_userptr_free_sg()
|
| A D | xe_pci_sriov.c | 76 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_pci_pf_get_vf_dev() 153 xe_assert(xe, IS_SRIOV_PF(xe)); in pf_enable_vfs() 154 xe_assert(xe, num_vfs > 0); in pf_enable_vfs() 155 xe_assert(xe, num_vfs <= total_vfs); in pf_enable_vfs() 211 xe_assert(xe, IS_SRIOV_PF(xe)); in pf_disable_vfs()
|
| A D | xe_guc.c | 61 xe_assert(xe, addr >= xe_wopcm_size(guc_to_xe(guc))); in guc_bo_ggtt_addr() 62 xe_assert(xe, addr < GUC_GGTT_TOP); in guc_bo_ggtt_addr() 302 xe_assert(xe, !(size % SZ_4K)); in guc_action_register_g2g_buffer() 411 xe_assert(xe, xe == gt_to_xe(far_gt)); in guc_g2g_register() 414 xe_assert(xe, g2g_bo); in guc_g2g_register() 417 xe_assert(xe, slot >= 0); in guc_g2g_register() 640 xe_assert(xe, count <= OPT_IN_MAX_DWORDS); in xe_guc_opt_in_features_enable() 1401 xe_assert(xe, len); in xe_guc_mmio_send_recv() 1402 xe_assert(xe, len <= VF_SW_FLAG_COUNT); in xe_guc_mmio_send_recv() 1403 xe_assert(xe, len <= MED_VF_SW_FLAG_COUNT); in xe_guc_mmio_send_recv() [all …]
|
| A D | xe_hw_engine_group.c | 127 xe_assert(xe, group); in xe_hw_engine_group_add_exec_queue() 128 xe_assert(xe, !(q->flags & EXEC_QUEUE_FLAG_VM)); in xe_hw_engine_group_add_exec_queue() 129 xe_assert(xe, q->vm); in xe_hw_engine_group_add_exec_queue() 167 xe_assert(xe, group); in xe_hw_engine_group_del_exec_queue() 168 xe_assert(xe, q->vm); in xe_hw_engine_group_del_exec_queue()
|
| A D | xe_bo.c | 174 xe_assert(xe, resource_is_vram(res)); in res_to_mem_region() 695 xe_assert(xe, attach); in xe_bo_move_dmabuf() 696 xe_assert(xe, ttm_bo->ttm); in xe_bo_move_dmabuf() 914 xe_assert(xe, migrate); in xe_bo_move() 1126 xe_assert(xe, !bo->backup_obj); in xe_bo_notifier_prepare_pinned() 1450 xe_assert(xe, locked); in xe_ttm_bo_lock_in_destructor() 2047 xe_assert(xe, tile); in __xe_bo_create_locked() 2235 xe_assert(xe, IS_DGFX(xe)); in xe_managed_bo_reinit_in_vram() 2284 xe_assert(xe, !bo->vm); in xe_bo_pin_external() 2372 xe_assert(xe, !bo->vm); in xe_bo_unpin_external() [all …]
|
| A D | xe_sriov.c | 80 xe_assert(xe, !xe->sriov.__mode); in xe_sriov_probe_early() 82 xe_assert(xe, xe->sriov.__mode); in xe_sriov_probe_early() 121 xe_assert(xe, !xe->sriov.wq); in xe_sriov_init()
|
| A D | xe_sriov_pf.c | 61 xe_assert(xe, totalvfs <= U16_MAX); in xe_sriov_pf_readiness() 90 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_init_early() 143 xe_assert(xe, IS_SRIOV_PF(xe)); in xe_sriov_pf_print_vfs_summary()
|
| A D | xe_gt_sriov_vf_debugfs.c | 58 xe_assert(xe, IS_SRIOV_VF(xe)); in xe_gt_sriov_vf_debugfs_register() 59 xe_assert(xe, root->d_inode->i_private == gt); in xe_gt_sriov_vf_debugfs_register()
|
| A D | xe_gt_ccs_mode.c | 25 xe_assert(xe, xe_gt_ccs_mode_enabled(gt)); in __xe_gt_apply_ccs_mode() 27 xe_assert(xe, num_engines && num_engines <= num_slices); in __xe_gt_apply_ccs_mode() 28 xe_assert(xe, !(num_slices % num_engines)); in __xe_gt_apply_ccs_mode()
|
| A D | xe_wa.h | 44 xe_assert(xe__, (xe__)->wa_active.oob_initialized); \ 49 xe_assert(xe__, (xe__)->wa_active.oob_initialized); \
|
| A D | xe_vm.c | 76 xe_assert(xe, xe_vma_is_userptr(vma)); in xe_vma_userptr_pin_pages() 172 xe_assert(vm->xe, link != list); in arm_preempt_fences() 903 xe_assert(vm->xe, vma->tile_present); in xe_vm_rebind() 1182 xe_assert(vm->xe, start < end); in xe_vma_create() 1183 xe_assert(vm->xe, end < vm->size); in xe_vma_create() 1567 xe_assert(xe, !(flags & ~XE_PTE_PS64)); in xelp_pte_encode_addr() 1958 xe_assert(xe, xe->info.has_asid); in xe_vm_close_and_put() 1962 xe_assert(xe, lookup == vm); in xe_vm_close_and_put() 1981 xe_assert(xe, !vm->size); in vm_destroy_work_func() 2330 xe_assert(vm->xe, bo); in vm_bind_ioctl_ops_create() [all …]
|
| A D | xe_shrinker.c | 90 xe_assert(xe, !IS_ERR(ttm_bo)); in __xe_shrinker_walk() 270 xe_assert(shrinker->xe, !shrinker->shrinkable_pages); in xe_shrinker_fini() 271 xe_assert(shrinker->xe, !shrinker->purgeable_pages); in xe_shrinker_fini()
|
| A D | xe_pxp.c | 500 xe_assert(pxp->xe, type == DRM_XE_PXP_TYPE_HWDRM); in xe_pxp_exec_queue_set_type() 541 xe_assert(pxp->xe, type == DRM_XE_PXP_TYPE_HWDRM); in pxp_start() 771 xe_assert(pxp->xe, !bo->pxp_key_instance); in xe_pxp_key_assign() 805 xe_assert(pxp->xe, bo->pxp_key_instance); in xe_pxp_bo_key_check() 940 xe_assert(pxp->xe, pxp->status == XE_PXP_SUSPENDED); in xe_pxp_pm_resume()
|
| A D | xe_sriov.h | 24 xe_assert(xe, xe->sriov.__mode); in xe_device_sriov_mode()
|
| A D | xe_pat.c | 447 xe_assert(xe, !xe->pat.ops || xe->pat.ops->dump); in xe_pat_init_early() 448 xe_assert(xe, !xe->pat.ops || xe->pat.ops->program_graphics); in xe_pat_init_early() 449 xe_assert(xe, !xe->pat.ops || MEDIA_VER(xe) < 13 || xe->pat.ops->program_media); in xe_pat_init_early()
|
| A D | xe_drm_client.c | 158 xe_assert(xe, !kref_read(&bo->ttm.base.refcount)); in xe_drm_client_remove_bo() 246 xe_assert(xef->xe, !list_empty(&bo->client_link)); in show_meminfo()
|
| A D | xe_gt_topology.c | 113 xe_assert(xe, find_last_bit(pattern, XE_MAX_L3_BANK_MASK_BITS) < patternbits || in gen_l3_mask_from_pattern() 115 xe_assert(xe, !mask || patternbits * (__fls(mask) + 1) <= XE_MAX_L3_BANK_MASK_BITS); in gen_l3_mask_from_pattern()
|
| A D | xe_svm.c | 596 xe_assert(vm->xe, xe_vm_is_closed(vm)); in xe_svm_close() 608 xe_assert(vm->xe, xe_vm_is_closed(vm)); in xe_svm_fini() 778 xe_assert(vm->xe, IS_DGFX(vm->xe)); in xe_svm_range_needs_migrate_to_vram() 830 xe_assert(vm->xe, xe_vma_is_cpu_addr_mirror(vma)); in xe_svm_handle_pagefault() 1016 xe_assert(tile_to_xe(tile), range->base.flags.migrate_devmem); in xe_svm_alloc_vram()
|
| A D | xe_uc_fw.c | 312 xe_assert(xe, !(size % 4)); in xe_uc_fw_copy_rsa() 313 xe_assert(xe, xe_uc_fw_is_available(uc_fw)); in xe_uc_fw_copy_rsa() 490 xe_assert(xe, manifest_entry); in parse_cpd_header() 542 xe_assert(xe, xe->info.platform != XE_DG2); in parse_cpd_header() 898 xe_assert(xe, !xe_uc_fw_is_loaded(uc_fw)); in xe_uc_fw_upload()
|
| A D | xe_vm.h | 182 xe_assert(xe_vma_vm(vma)->xe, xe_vma_is_userptr(vma)); in to_userptr_vma() 240 xe_assert(vm->xe, xe_vm_in_preempt_fence_mode(vm)); in xe_vm_queue_rebind_worker()
|
| A D | xe_assert.h | 108 #define xe_assert(xe, condition) xe_assert_msg((xe), condition, "") macro
|
| A D | xe_mocs.c | 652 xe_assert(xe, info->unused_entries_index != 0); in get_mocs_settings() 654 xe_assert(xe, info->ops && info->ops->dump); in get_mocs_settings() 655 xe_assert(xe, info->table_size <= info->num_mocs_regs); in get_mocs_settings()
|
| A D | xe_migrate.c | 153 xe_assert(xe, IS_ALIGNED(xe->mem.vram.usable_size, SZ_2M)); in xe_migrate_program_identity() 181 xe_assert(xe, pos == vram_limit); in xe_migrate_program_identity() 313 xe_assert(xe, xe->mem.vram.actual_physical_size <= in xe_migrate_prepare_vm() 326 xe_assert(xe, xe->mem.vram.actual_physical_size <= (MAX_NUM_PTE - in xe_migrate_prepare_vm() 600 xe_assert(xe, (va & (SZ_64K - 1)) == in emit_pte() 844 xe_assert(xe, type_device); in xe_migrate_copy() 850 xe_assert(xe, IS_ALIGNED(ccs_it.start, PAGE_SIZE)); in xe_migrate_copy() 1636 xe_assert(xe, npages * PAGE_SIZE <= MAX_PREEMPTDISABLE_TRANSFER); in xe_migrate_vram()
|