Home
last modified time | relevance | path

Searched refs:mqd_mem_obj (Results 1 – 11 of 11) sorted by relevance

/drivers/gpu/drm/amd/amdkfd/
A Dkfd_mqd_manager.c51 struct kfd_mem_obj *mqd_mem_obj; in allocate_hiq_mqd() local
54 if (!mqd_mem_obj) in allocate_hiq_mqd()
61 return mqd_mem_obj; in allocate_hiq_mqd()
67 struct kfd_mem_obj *mqd_mem_obj; in allocate_sdma_mqd() local
71 if (!mqd_mem_obj) in allocate_sdma_mqd()
88 return mqd_mem_obj; in allocate_sdma_mqd()
92 struct kfd_mem_obj *mqd_mem_obj) in free_mqd_hiq_sdma() argument
94 WARN_ON(!mqd_mem_obj->gtt_mem); in free_mqd_hiq_sdma()
95 kfree(mqd_mem_obj); in free_mqd_hiq_sdma()
227 if (mqd_mem_obj->gtt_mem) { in kfd_free_mqd_cp()
[all …]
A Dkfd_mqd_manager_cik.c79 struct kfd_mem_obj *mqd_mem_obj; in allocate_mqd() local
82 &mqd_mem_obj)) in allocate_mqd()
85 return mqd_mem_obj; in allocate_mqd()
89 struct kfd_mem_obj *mqd_mem_obj, uint64_t *gart_addr, in init_mqd() argument
95 m = (struct cik_mqd *) mqd_mem_obj->cpu_ptr; in init_mqd()
96 addr = mqd_mem_obj->gpu_addr; in init_mqd()
154 *gart_addr = mqd_mem_obj->gpu_addr; in init_mqd_sdma()
269 m = (struct cik_mqd *) mqd_mem_obj->cpu_ptr; in restore_mqd()
270 addr = mqd_mem_obj->gpu_addr; in restore_mqd()
308 addr = mqd_mem_obj->gpu_addr; in restore_mqd_sdma()
[all …]
A Dkfd_mqd_manager_v9.c136 if (!mqd_mem_obj) in allocate_mqd()
142 &(mqd_mem_obj->gtt_mem), in allocate_mqd()
143 &(mqd_mem_obj->gpu_addr), in allocate_mqd()
147 kfree(mqd_mem_obj); in allocate_mqd()
152 &mqd_mem_obj); in allocate_mqd()
157 return mqd_mem_obj; in allocate_mqd()
168 addr = mqd_mem_obj->gpu_addr; in init_mqd()
420 addr = mqd_mem_obj->gpu_addr; in restore_mqd()
546 addr = mqd_mem_obj->gpu_addr; in restore_mqd_sdma()
669 mqd_mem_obj->gtt_mem : NULL; in get_xcc_mqd()
[all …]
A Dkfd_mqd_manager_vi.c82 struct kfd_mem_obj *mqd_mem_obj; in allocate_mqd() local
85 &mqd_mem_obj)) in allocate_mqd()
88 return mqd_mem_obj; in allocate_mqd()
98 m = (struct vi_mqd *) mqd_mem_obj->cpu_ptr; in init_mqd()
99 addr = mqd_mem_obj->gpu_addr; in init_mqd()
301 m = (struct vi_mqd *) mqd_mem_obj->cpu_ptr; in restore_mqd()
302 addr = mqd_mem_obj->gpu_addr; in restore_mqd()
325 init_mqd(mm, mqd, mqd_mem_obj, gart_addr, q); in init_mqd_hiq()
346 m = (struct vi_sdma_mqd *) mqd_mem_obj->cpu_ptr; in init_mqd_sdma()
352 *gart_addr = mqd_mem_obj->gpu_addr; in init_mqd_sdma()
[all …]
A Dkfd_mqd_manager_v10.c79 struct kfd_mem_obj *mqd_mem_obj; in allocate_mqd() local
82 &mqd_mem_obj)) in allocate_mqd()
85 return mqd_mem_obj; in allocate_mqd()
95 m = (struct v10_compute_mqd *) mqd_mem_obj->cpu_ptr; in init_mqd()
96 addr = mqd_mem_obj->gpu_addr; in init_mqd()
293 addr = mqd_mem_obj->gpu_addr; in restore_mqd()
316 init_mqd(mm, mqd, mqd_mem_obj, gart_addr, q); in init_mqd_hiq()
350 m = (struct v10_sdma_mqd *) mqd_mem_obj->cpu_ptr; in init_mqd_sdma()
356 *gart_addr = mqd_mem_obj->gpu_addr; in init_mqd_sdma()
412 m = (struct v10_sdma_mqd *) mqd_mem_obj->cpu_ptr; in restore_mqd_sdma()
[all …]
A Dkfd_mqd_manager_v11.c105 struct kfd_mem_obj *mqd_mem_obj; in allocate_mqd() local
117 if (kfd_gtt_sa_allocate(node, size, &mqd_mem_obj)) in allocate_mqd()
120 return mqd_mem_obj; in allocate_mqd()
124 struct kfd_mem_obj *mqd_mem_obj, uint64_t *gart_addr, in init_mqd() argument
132 m = (struct v11_compute_mqd *) mqd_mem_obj->cpu_ptr; in init_mqd()
133 addr = mqd_mem_obj->gpu_addr; in init_mqd()
344 m = (struct v11_compute_mqd *) mqd_mem_obj->cpu_ptr; in restore_mqd()
345 addr = mqd_mem_obj->gpu_addr; in restore_mqd()
369 init_mqd(mm, mqd, mqd_mem_obj, gart_addr, q); in init_mqd_hiq()
404 m = (struct v11_sdma_mqd *) mqd_mem_obj->cpu_ptr; in init_mqd_sdma()
[all …]
A Dkfd_mqd_manager.h75 struct kfd_mem_obj *mqd_mem_obj, uint64_t *gart_addr,
93 struct kfd_mem_obj *mqd_mem_obj);
113 struct kfd_mem_obj *mqd_mem_obj, uint64_t *gart_addr,
162 struct kfd_mem_obj *mqd_mem_obj);
177 struct kfd_mem_obj *mqd_mem_obj);
196 struct kfd_mem_obj *mqd_mem_obj, uint32_t virtual_xcc_id);
A Dkfd_mqd_manager_v12.c86 struct kfd_mem_obj *mqd_mem_obj; in allocate_mqd() local
92 if (kfd_gtt_sa_allocate(node, PAGE_SIZE, &mqd_mem_obj)) in allocate_mqd()
95 return mqd_mem_obj; in allocate_mqd()
99 struct kfd_mem_obj *mqd_mem_obj, uint64_t *gart_addr, in init_mqd() argument
105 m = (struct v12_compute_mqd *) mqd_mem_obj->cpu_ptr; in init_mqd()
106 addr = mqd_mem_obj->gpu_addr; in init_mqd()
290 struct kfd_mem_obj *mqd_mem_obj, uint64_t *gart_addr, in init_mqd_hiq() argument
295 init_mqd(mm, mqd, mqd_mem_obj, gart_addr, q); in init_mqd_hiq()
304 struct kfd_mem_obj *mqd_mem_obj, uint64_t *gart_addr, in init_mqd_sdma() argument
309 m = (struct v12_sdma_mqd *) mqd_mem_obj->cpu_ptr; in init_mqd_sdma()
[all …]
A Dkfd_kernel_queue.c149 kq->queue->mqd_mem_obj = kq->mqd_mgr->allocate_mqd(kq->mqd_mgr->dev, in kq_initialize()
151 if (!kq->queue->mqd_mem_obj) in kq_initialize()
154 kq->queue->mqd_mem_obj, in kq_initialize()
182 kq->mqd_mgr->free_mqd(kq->mqd_mgr, kq->queue->mqd, kq->queue->mqd_mem_obj); in kq_initialize()
216 kq->queue->mqd_mem_obj); in kq_uninitialize()
A Dkfd_device_queue_manager.c676 q->mqd_mem_obj = mqd_mgr->allocate_mqd(mqd_mgr->dev, &q->properties); in create_queue_nocpsch()
679 if (!q->mqd_mem_obj) { in create_queue_nocpsch()
689 mqd_mgr->init_mqd(mqd_mgr, &q->mqd, q->mqd_mem_obj, in create_queue_nocpsch()
724 mqd_mgr->free_mqd(mqd_mgr, q->mqd, q->mqd_mem_obj); in create_queue_nocpsch()
938 mqd_mgr->free_mqd(mqd_mgr, q->mqd, q->mqd_mem_obj); in destroy_queue_nocpsch()
2001 if (!q->mqd_mem_obj) { in create_queue_cpsch()
2021 mqd_mgr->init_mqd(mqd_mgr, &q->mqd, q->mqd_mem_obj, in create_queue_cpsch()
2056 mqd_mgr->free_mqd(mqd_mgr, q->mqd, q->mqd_mem_obj); in create_queue_cpsch()
2590 mqd_mgr->free_mqd(mqd_mgr, q->mqd, q->mqd_mem_obj); in destroy_queue_cpsch()
2661 mqd_mgr->free_mqd(mqd_mgr, q->mqd, q->mqd_mem_obj); in process_termination_nocpsch()
[all …]
A Dkfd_priv.h598 struct kfd_mem_obj *mqd_mem_obj; member

Completed in 39 milliseconds