Home
last modified time | relevance | path

Searched refs:xe_vma (Results 1 – 14 of 14) sorted by relevance

/drivers/gpu/drm/xe/
A Dxe_trace_bo.h88 DECLARE_EVENT_CLASS(xe_vma,
89 TP_PROTO(struct xe_vma *vma),
118 DEFINE_EVENT(xe_vma, xe_vma_flush,
119 TP_PROTO(struct xe_vma *vma),
124 TP_PROTO(struct xe_vma *vma),
128 DEFINE_EVENT(xe_vma, xe_vma_acc,
129 TP_PROTO(struct xe_vma *vma),
133 DEFINE_EVENT(xe_vma, xe_vma_bind,
134 TP_PROTO(struct xe_vma *vma),
139 TP_PROTO(struct xe_vma *vma),
[all …]
A Dxe_vm.h66 struct xe_vma *
98 return container_of(gpuva, struct xe_vma, gpuva); in gpuva_to_vma()
110 static inline u64 xe_vma_start(struct xe_vma *vma) in xe_vma_start()
115 static inline u64 xe_vma_size(struct xe_vma *vma) in xe_vma_size()
120 static inline u64 xe_vma_end(struct xe_vma *vma) in xe_vma_end()
125 static inline u64 xe_vma_bo_offset(struct xe_vma *vma) in xe_vma_bo_offset()
141 static inline bool xe_vma_read_only(struct xe_vma *vma) in xe_vma_read_only()
146 static inline u64 xe_vma_userptr(struct xe_vma *vma) in xe_vma_userptr()
151 static inline bool xe_vma_is_null(struct xe_vma *vma) in xe_vma_is_null()
225 struct xe_vma *vma,
[all …]
A Dxe_svm.h20 struct xe_vma;
63 int xe_svm_handle_pagefault(struct xe_vm *vm, struct xe_vma *vma,
77 struct xe_vma *vma, struct drm_gpusvm_ctx *ctx);
82 bool xe_svm_range_needs_migrate_to_vram(struct xe_svm_range *range, struct xe_vma *vma,
91 u64 xe_svm_find_vma_start(struct xe_vm *vm, u64 addr, u64 end, struct xe_vma *vma);
172 struct xe_vma;
215 int xe_svm_handle_pagefault(struct xe_vm *vm, struct xe_vma *vma, in xe_svm_handle_pagefault()
248 struct xe_vma *vma, struct drm_gpusvm_ctx *ctx) in xe_svm_range_find_or_insert()
281 bool xe_svm_range_needs_migrate_to_vram(struct xe_svm_range *range, struct xe_vma *vma, in xe_svm_range_needs_migrate_to_vram()
301 u64 xe_svm_find_vma_start(struct xe_vm *vm, u64 addr, u64 end, struct xe_vma *vma) in xe_svm_find_vma_start()
A Dxe_vm_types.h80 struct xe_vma { struct
146 struct xe_vma vma; argument
306 struct xe_vma *last_fault_vma;
337 struct xe_vma *vma;
357 struct xe_vma *prev;
359 struct xe_vma *next;
381 struct xe_vma *vma;
A Dxe_pt_types.h15 struct xe_vma;
43 u64 (*pte_encode_vma)(u64 pte, struct xe_vma *vma,
82 struct xe_vma *vma;
A Dxe_vm.c805 struct xe_vma *vma; in xe_vma_svm_prefetch_op_fini()
982 struct xe_vma *vma, in xe_vm_populate_range_rebind()
996 struct xe_vma *vma, in xe_vm_ops_add_range_rebind()
1173 struct xe_vma *vma; in xe_vma_create()
1308 struct xe_vma *vma = in vma_destroy_work_func()
1317 struct xe_vma *vma = container_of(cb, struct xe_vma, destroy_cb); in vma_destroy_cb()
1400 struct xe_vma *
2223 struct xe_vma *vma; in print_op()
2450 struct xe_vma *vma; in new_vma()
2608 struct xe_vma *vma; in vm_bind_ioctl_ops_parse()
[all …]
A Dxe_pt.c303 struct xe_vma *vma;
983 static void xe_pt_cancel_bind(struct xe_vma *vma, in xe_pt_cancel_bind()
1039 static void xe_pt_commit(struct xe_vma *vma, in xe_pt_commit()
1066 static void xe_pt_abort_bind(struct xe_vma *vma, in xe_pt_abort_bind()
1711 struct xe_vma *vma, in xe_pt_stage_unbind()
1758 static void xe_pt_abort_unbind(struct xe_vma *vma, in xe_pt_abort_unbind()
1784 xe_pt_commit_prepare_unbind(struct xe_vma *vma, in xe_pt_commit_prepare_unbind()
1849 struct xe_vma *vma, bool invalidate_on_bind) in bind_op_prepare()
1958 struct xe_vma *vma) in unbind_op_prepare()
2212 struct xe_vma *vma, struct dma_fence *fence, in bind_op_commit()
[all …]
A Dxe_pt.h20 struct xe_vma;
47 bool xe_pt_zap_ptes(struct xe_tile *tile, struct xe_vma *vma);
A Dxe_gt_pagefault.c70 static bool vma_is_valid(struct xe_tile *tile, struct xe_vma *vma) in vma_is_valid()
76 static int xe_pf_begin(struct drm_exec *exec, struct xe_vma *vma, in xe_pf_begin()
107 static int handle_vma_pagefault(struct xe_gt *gt, struct xe_vma *vma, in handle_vma_pagefault()
189 struct xe_vma *vma = NULL; in handle_pagefault()
533 static struct xe_vma *get_acc_vma(struct xe_vm *vm, struct acc *acc) in get_acc_vma()
547 struct xe_vma *vma; in handle_acc()
A Dxe_gt_tlb_invalidation.h16 struct xe_vma;
A Dxe_hmm.c151 struct xe_vma *vma = &uvma->vma; in xe_hmm_userptr_unmap()
221 struct xe_vma *vma = &uvma->vma; in xe_hmm_userptr_populate_range()
A Dxe_migrate.h25 struct xe_vma;
A Dxe_svm.c678 u64 xe_svm_find_vma_start(struct xe_vm *vm, u64 start, u64 end, struct xe_vma *vma) in xe_svm_find_vma_start()
769 bool xe_svm_range_needs_migrate_to_vram(struct xe_svm_range *range, struct xe_vma *vma, in xe_svm_range_needs_migrate_to_vram()
806 int xe_svm_handle_pagefault(struct xe_vm *vm, struct xe_vma *vma, in xe_svm_handle_pagefault()
963 struct xe_vma *vma, struct drm_gpusvm_ctx *ctx) in xe_svm_range_find_or_insert()
A Dxe_bo.c664 struct xe_vma *vma = gpuva_to_vma(gpuva); in xe_bo_trigger_rebind()

Completed in 757 milliseconds