| /drivers/gpu/drm/xe/tests/ |
| A D | xe_migrate.c | 77 bool big = xe_bo_size(bo) >= SZ_2M; in test_copy() 83 xe_bo_size(bo), in test_copy() 108 xe_map_memset(xe, &remote->vmap, 0, 0xd0, xe_bo_size(remote)); in test_copy() 124 xe_map_memset(xe, &bo->vmap, 0, 0xd0, xe_bo_size(bo)); in test_copy() 134 retval = xe_map_rd(xe, &bo->vmap, xe_bo_size(bo) - 8, u64); in test_copy() 142 xe_map_memset(xe, &bo->vmap, 0, 0xc0, xe_bo_size(bo)); in test_copy() 248 xe_res_first(pt->ttm.resource, 0, xe_bo_size(pt), &src_it); in xe_migrate_sanity_test() 250 xe_res_first_sg(xe_bo_sg(pt), 0, xe_bo_size(pt), &src_it); in xe_migrate_sanity_test() 279 xe_map_memset(xe, &tiny->vmap, 0, 0x22, xe_bo_size(tiny)); in xe_migrate_sanity_test() 301 xe_map_memset(xe, &big->vmap, 0, 0x11, xe_bo_size(big)); in xe_migrate_sanity_test() [all …]
|
| A D | xe_guc_buf_kunit.c | 46 xe_bo_size(bo), SZ_4K)); in replacement_xe_managed_bo_create_pin_map()
|
| A D | xe_bo.c | 109 offset = xe_device_ccs_bytes(tile_to_xe(tile), xe_bo_size(bo)); in ccs_test_migrate()
|
| /drivers/gpu/drm/xe/ |
| A D | xe_migrate.c | 217 pt29_ofs = xe_bo_size(bo) - 3 * XE_PAGE_SIZE; in xe_migrate_prepare_vm() 239 for (i = 0; i < xe_bo_size(batch); in xe_migrate_prepare_vm() 250 xe_tile_assert(tile, xe_bo_size(batch) == SZ_1M); in xe_migrate_prepare_vm() 254 xe_tile_assert(tile, xe_bo_size(batch) == SZ_512K); in xe_migrate_prepare_vm() 256 for (i = 0; i < xe_bo_size(batch); in xe_migrate_prepare_vm() 309 u64 pt30_ofs = xe_bo_size(bo) - 2 * XE_PAGE_SIZE; in xe_migrate_prepare_vm() 324 u64 pt31_ofs = xe_bo_size(bo) - XE_PAGE_SIZE; in xe_migrate_prepare_vm() 771 u64 size = xe_bo_size(src_bo); in xe_migrate_copy() 794 if (src_bo != dst_bo && XE_WARN_ON(xe_bo_size(src_bo) != xe_bo_size(dst_bo))) in xe_migrate_copy() 1067 u64 size = xe_bo_size(bo); in xe_migrate_clear() [all …]
|
| A D | xe_ggtt.c | 424 xe_map_memset(xe, &ggtt->scratch->vmap, 0, 0, xe_bo_size(ggtt->scratch)); in xe_ggtt_init() 696 end = start + xe_bo_size(bo); in xe_ggtt_map_bo() 702 for (xe_res_first_sg(xe_bo_sg(bo), 0, xe_bo_size(bo), &cur); in xe_ggtt_map_bo() 710 for (xe_res_first(bo->ttm.resource, 0, xe_bo_size(bo), &cur); in xe_ggtt_map_bo() 746 xe_tile_assert(ggtt->tile, bo->ggtt_node[tile_id]->base.size == xe_bo_size(bo)); in __xe_ggtt_insert_bo_at() 765 xe_bo_size(bo), alignment, 0, start, end, 0); in __xe_ggtt_insert_bo_at() 826 xe_tile_assert(ggtt->tile, bo->ggtt_node[tile_id]->base.size == xe_bo_size(bo)); in xe_ggtt_remove_bo()
|
| A D | xe_bo.h | 246 static inline size_t xe_bo_size(struct xe_bo *bo) in xe_bo_size() function 259 XE_WARN_ON(ggtt_node->base.size > xe_bo_size(bo)); in __xe_bo_ggtt_addr() 313 return PAGE_ALIGN(xe_bo_size(bo)); in xe_bo_ccs_pages_start()
|
| A D | xe_trace_bo.h | 36 __entry->size = xe_bo_size(bo); 76 __entry->size = xe_bo_size(bo);
|
| A D | xe_lmtt.c | 84 xe_map_memset(lmtt_to_xe(lmtt), &bo->vmap, 0, 0, xe_bo_size(bo)); in lmtt_pt_alloc() 454 lmtt_assert(lmtt, IS_ALIGNED(xe_bo_size(bo), page_size)); in lmtt_insert_bo() 458 xe_res_first(bo->ttm.resource, 0, xe_bo_size(bo), &cur); in lmtt_insert_bo()
|
| A D | xe_gt_sriov_pf_config.c | 294 cfg[n++] = lower_32_bits(xe_bo_size(config->lmem_obj)); in encode_config() 295 cfg[n++] = upper_32_bits(xe_bo_size(config->lmem_obj)); in encode_config() 1322 return bo ? xe_bo_size(bo) : 0; in pf_get_vf_config_lmem() 1421 offset += xe_bo_size(bo); in pf_update_vf_lmtt() 1502 err = pf_push_vf_cfg_lmem(gt, vfid, xe_bo_size(bo)); in pf_provision_vf_lmem() 1507 vfid, xe_bo_size(bo), xe_bo_size(bo) / SZ_1M); in pf_provision_vf_lmem() 2627 string_get_size(xe_bo_size(config->lmem_obj), 1, STRING_UNITS_2, in xe_gt_sriov_pf_config_print_lmem() 2630 n, xe_bo_size(config->lmem_obj), buf); in xe_gt_sriov_pf_config_print_lmem()
|
| A D | xe_bo.c | 455 extra_pages = DIV_ROUND_UP(xe_device_ccs_bytes(xe, xe_bo_size(bo)), in xe_ttm_tt_create() 1142 backup = ___xe_bo_create_locked(xe, NULL, NULL, bo->ttm.base.resv, NULL, xe_bo_size(bo), in xe_bo_notifier_prepare_pinned() 1221 NULL, xe_bo_size(bo), in xe_bo_evict_pinned() 1275 xe_bo_size(bo)); in xe_bo_evict_pinned() 1368 xe_bo_size(bo)); in xe_bo_restore_pinned() 1579 xe_bo_size(bo) - (offset & PAGE_MASK), &cursor); in xe_ttm_access_memory() 2056 start + xe_bo_size(bo), U64_MAX); in __xe_bo_create_locked() 2239 xe_bo_size(*src), dst_flags); in xe_managed_bo_reinit_in_vram() 2529 ret = ttm_bo_kmap(&bo->ttm, 0, xe_bo_size(bo) >> PAGE_SHIFT, &bo->kmap); in xe_bo_vmap()
|
| A D | xe_gsc.c | 63 xe_bo_size(gsc->private) - fw_size); in memcpy_fw() 86 bb->cs[bb->len++] = (xe_bo_size(gsc->private) / SZ_4K) | in emit_gsc_upload()
|
| A D | xe_drm_client.c | 170 u64 sz = xe_bo_size(bo); in bo_meminfo()
|
| A D | xe_guc_log.c | 82 snapshot->size = xe_bo_size(log->bo); in xe_guc_log_snapshot_alloc()
|
| A D | xe_huc.c | 174 xe_bo_size(huc->fw.bo)); in huc_auth_via_gsccs()
|
| A D | xe_oa.c | 406 int size_exponent = __ffs(xe_bo_size(stream->oa_buffer.bo)); in xe_oa_init_oa_buffer() 438 memset(stream->oa_buffer.vaddr, 0, xe_bo_size(stream->oa_buffer.bo)); in xe_oa_init_oa_buffer() 1068 xe_bo_size(stream->oa_buffer.bo) > SZ_16M ? in oag_buf_size_select() 1585 struct drm_xe_oa_stream_info info = { .oa_buf_size = xe_bo_size(stream->oa_buffer.bo), }; in xe_oa_info_locked() 1671 if (vma->vm_end - vma->vm_start != xe_bo_size(stream->oa_buffer.bo)) { in xe_oa_mmap()
|
| A D | xe_guc_ads.c | 893 xe_map_memset(ads_to_xe(ads), ads_to_map(ads), 0, 0, xe_bo_size(ads->bo)); in xe_guc_ads_populate_minimal() 917 xe_map_memset(ads_to_xe(ads), ads_to_map(ads), 0, 0, xe_bo_size(ads->bo)); in xe_guc_ads_populate()
|
| A D | xe_lrc.c | 749 u32 offset = xe_bo_size(lrc->bo) - LRC_WA_BB_SIZE - in __xe_lrc_indirect_ring_offset() 760 return xe_bo_size(lrc->bo) - LRC_WA_BB_SIZE - LRC_INDIRECT_CTX_BO_SIZE; in __xe_lrc_indirect_ctx_offset() 765 return xe_bo_size(lrc->bo) - LRC_WA_BB_SIZE; in __xe_lrc_wa_bb_offset()
|
| A D | xe_guc.c | 63 xe_assert(xe, xe_bo_size(bo) <= GUC_GGTT_TOP - addr); in guc_bo_ggtt_addr() 424 xe_assert(xe, (buf - base + G2G_BUFFER_SIZE) <= xe_bo_size(g2g_bo)); in guc_g2g_register()
|
| A D | xe_guc_ct.c | 476 xe_map_memset(xe, &ct->bo->vmap, 0, 0, xe_bo_size(ct->bo)); in xe_guc_ct_enable() 1935 snapshot->ctb_size = xe_bo_size(ct->bo); in guc_ct_snapshot_alloc()
|
| A D | xe_vm.c | 3469 if (XE_IOCTL_DBG(xe, range > xe_bo_size(bo)) || in xe_vm_bind_ioctl_validate_bo() 3471 xe_bo_size(bo) - range)) { in xe_vm_bind_ioctl_validate_bo() 3774 ops = vm_bind_ioctl_ops_create(vm, &vops, bo, 0, addr, xe_bo_size(bo), in xe_vm_bind_kernel_bo()
|
| /drivers/gpu/drm/xe/display/ |
| A D | xe_hdcp_gsc.c | 88 xe_map_memset(xe, &bo->vmap, 0, 0, xe_bo_size(bo)); in intel_hdcp_gsc_initialize_message()
|
| A D | xe_fb_pin.c | 230 ret = xe_ggtt_node_insert_locked(vma->node, xe_bo_size(bo), align, 0); in __xe_pin_fb_vma_ggtt()
|