| /drivers/gpu/drm/amd/amdkfd/ |
| A D | kfd_device_queue_manager_v9.c | 30 struct qcm_process_device *qpd); 82 qpd->sh_mem_ape1_limit = 0; in set_cache_memory_policy_v9() 83 qpd->sh_mem_ape1_base = 0; in set_cache_memory_policy_v9() 84 qpd->sh_mem_bases = compute_sh_mem_bases_64bit(qpd_to_pdd(qpd)); in set_cache_memory_policy_v9() 87 qpd->sh_mem_config); in set_cache_memory_policy_v9() 96 pdd = qpd_to_pdd(qpd); in update_qpd_v9() 99 if (qpd->sh_mem_config == 0) { in update_qpd_v9() 108 qpd->sh_mem_config |= in update_qpd_v9() 111 qpd->sh_mem_ape1_limit = 0; in update_qpd_v9() 112 qpd->sh_mem_ape1_base = 0; in update_qpd_v9() [all …]
|
| A D | kfd_device_queue_manager.c | 166 qpd = cur->qpd; in kfd_hws_hang() 312 qpd = cur->qpd; in remove_all_kfd_queues_mes() 339 qpd = cur->qpd; in add_all_kfd_queues_mes() 576 qpd->sh_hidden_private_base, qpd->vmid); in allocate_vmid() 1403 n->qpd = qpd; in register_process() 1443 if (qpd == cur->qpd) { in unregister_process() 2190 qpd = cur->qpd; in find_queue_by_address() 2280 qpd = cur->qpd; in set_sdma_queue_as_reset() 2667 if (qpd == cur->qpd) { in process_termination_nocpsch() 2816 if (qpd == cur->qpd) { in process_termination_cpsch() [all …]
|
| A D | kfd_device_queue_manager_cik.c | 39 struct qcm_process_device *qpd, 46 struct qcm_process_device *qpd); 49 struct qcm_process_device *qpd); 88 struct qcm_process_device *qpd, in set_cache_memory_policy_cik() argument 102 qpd->sh_mem_ape1_base = 1; in set_cache_memory_policy_cik() 103 qpd->sh_mem_ape1_limit = 0; in set_cache_memory_policy_cik() 124 qpd->sh_mem_ape1_base = base >> 16; in set_cache_memory_policy_cik() 136 qpd->sh_mem_config = (qpd->sh_mem_config & PTR32) in set_cache_memory_policy_cik() 147 qpd->pqm->process->is_32bit_user_mode, temp, qpd->sh_mem_bases); in set_cache_memory_policy_cik() 154 struct qcm_process_device *qpd) in update_qpd_cik() argument [all …]
|
| A D | kfd_packet_manager_v9.c | 39 container_of(qpd, struct kfd_process_device, qpd); in pm_map_process_v9() 53 packet->bitfields14.num_gws = (qpd->mapped_gws_queue) ? qpd->num_gws : 0; in pm_map_process_v9() 54 packet->bitfields14.num_oac = qpd->num_oac; in pm_map_process_v9() 56 packet->bitfields14.num_queues = (qpd->is_debug) ? 0 : qpd->queue_count; in pm_map_process_v9() 65 packet->sh_mem_bases = qpd->sh_mem_bases; in pm_map_process_v9() 66 if (qpd->tba_addr) { in pm_map_process_v9() 97 container_of(qpd, struct kfd_process_device, qpd); in pm_map_process_aldebaran() 113 packet->bitfields14.num_gws = (qpd->mapped_gws_queue) ? qpd->num_gws : 0; in pm_map_process_aldebaran() 116 packet->bitfields14.num_queues = (qpd->is_debug) ? 0 : qpd->queue_count; in pm_map_process_aldebaran() 129 packet->sh_mem_bases = qpd->sh_mem_bases; in pm_map_process_aldebaran() [all …]
|
| A D | kfd_device_queue_manager_vi.c | 46 struct qcm_process_device *qpd); 49 struct qcm_process_device *qpd); 89 struct qcm_process_device *qpd, in set_cache_memory_policy_vi() argument 103 qpd->sh_mem_ape1_base = 1; in set_cache_memory_policy_vi() 104 qpd->sh_mem_ape1_limit = 0; in set_cache_memory_policy_vi() 125 qpd->sh_mem_ape1_base = base >> 16; in set_cache_memory_policy_vi() 126 qpd->sh_mem_ape1_limit = limit >> 16; in set_cache_memory_policy_vi() 137 qpd->sh_mem_config = in set_cache_memory_policy_vi() 150 temp, qpd->sh_mem_bases); in set_cache_memory_policy_vi() 156 struct qcm_process_device *qpd) in update_qpd_vi() argument [all …]
|
| A D | kfd_doorbell.c | 228 qpd->doorbell_bitmap); in init_doorbell_bitmap() 240 if (!pdd->qpd.proc_doorbells) { in kfd_get_process_doorbells() 247 pdd->qpd.proc_doorbells, in kfd_get_process_doorbells() 256 struct qcm_process_device *qpd = &pdd->qpd; in kfd_alloc_process_doorbells() local 261 if (!qpd->doorbell_bitmap) { in kfd_alloc_process_doorbells() 278 &qpd->proc_doorbells, in kfd_alloc_process_doorbells() 288 bitmap_free(qpd->doorbell_bitmap); in kfd_alloc_process_doorbells() 289 qpd->doorbell_bitmap = NULL; in kfd_alloc_process_doorbells() 295 struct qcm_process_device *qpd = &pdd->qpd; in kfd_free_process_doorbells() local 297 if (qpd->doorbell_bitmap) { in kfd_free_process_doorbells() [all …]
|
| A D | kfd_device_queue_manager_v11.c | 30 struct qcm_process_device *qpd); 32 struct qcm_process_device *qpd); 34 struct qcm_process_device *qpd, 60 struct qcm_process_device *qpd, in set_cache_memory_policy_v11() argument 67 qpd->sh_mem_config = (SH_MEM_ALIGNMENT_MODE_UNALIGNED << in set_cache_memory_policy_v11() 71 qpd->sh_mem_ape1_limit = 0; in set_cache_memory_policy_v11() 72 qpd->sh_mem_ape1_base = 0; in set_cache_memory_policy_v11() 73 qpd->sh_mem_bases = compute_sh_mem_bases_64bit(qpd_to_pdd(qpd)); in set_cache_memory_policy_v11() 75 pr_debug("sh_mem_bases 0x%X\n", qpd->sh_mem_bases); in set_cache_memory_policy_v11() 80 struct qcm_process_device *qpd) in update_qpd_v11() argument [all …]
|
| A D | kfd_device_queue_manager_v12.c | 30 struct qcm_process_device *qpd); 32 struct qcm_process_device *qpd); 34 struct qcm_process_device *qpd, 60 struct qcm_process_device *qpd, in set_cache_memory_policy_v12() argument 67 qpd->sh_mem_config = (SH_MEM_ALIGNMENT_MODE_UNALIGNED << in set_cache_memory_policy_v12() 71 qpd->sh_mem_ape1_limit = 0; in set_cache_memory_policy_v12() 72 qpd->sh_mem_ape1_base = 0; in set_cache_memory_policy_v12() 73 qpd->sh_mem_bases = compute_sh_mem_bases_64bit(qpd_to_pdd(qpd)); in set_cache_memory_policy_v12() 75 pr_debug("sh_mem_bases 0x%X\n", qpd->sh_mem_bases); in set_cache_memory_policy_v12() 80 struct qcm_process_device *qpd) in update_qpd_v12() argument [all …]
|
| A D | kfd_device_queue_manager_v10.c | 31 struct qcm_process_device *qpd); 33 struct qcm_process_device *qpd); 35 struct qcm_process_device *qpd, 61 struct qcm_process_device *qpd, in set_cache_memory_policy_v10() argument 68 qpd->sh_mem_config = (SH_MEM_ALIGNMENT_MODE_UNALIGNED << in set_cache_memory_policy_v10() 71 qpd->sh_mem_ape1_limit = 0; in set_cache_memory_policy_v10() 72 qpd->sh_mem_ape1_base = 0; in set_cache_memory_policy_v10() 73 qpd->sh_mem_bases = compute_sh_mem_bases_64bit(qpd_to_pdd(qpd)); in set_cache_memory_policy_v10() 75 pr_debug("sh_mem_bases 0x%X\n", qpd->sh_mem_bases); in set_cache_memory_policy_v10() 80 struct qcm_process_device *qpd) in update_qpd_v10() argument [all …]
|
| A D | kfd_process.c | 120 qpd = &pdd->qpd; in kfd_sdma_activity_worker() 794 struct qcm_process_device *qpd = &pdd->qpd; in kfd_process_device_reserve_ib_mem() local 803 if (qpd->ib_kaddr || !qpd->ib_base) in kfd_process_device_reserve_ib_mem() 820 struct qcm_process_device *qpd = &pdd->qpd; in kfd_process_device_destroy_ib_mem() local 822 if (!qpd->ib_kaddr || !qpd->ib_base) in kfd_process_device_destroy_ib_mem() 1371 qpd->tba_addr, qpd->tma_addr, qpd->cwsr_kaddr); in kfd_process_init_cwsr_apu() 1382 struct qcm_process_device *qpd = &pdd->qpd; in kfd_process_device_init_cwsr_dgpu() local 1401 qpd->tba_addr = qpd->cwsr_base; in kfd_process_device_init_cwsr_dgpu() 1410 qpd->tba_addr, qpd->tma_addr, qpd->cwsr_kaddr); in kfd_process_device_init_cwsr_dgpu() 1418 struct qcm_process_device *qpd = &pdd->qpd; in kfd_process_device_destroy_cwsr_dgpu() local [all …]
|
| A D | kfd_device_queue_manager.h | 42 struct qcm_process_device *qpd; member 139 struct qcm_process_device *qpd, 145 struct qcm_process_device *qpd, 152 struct qcm_process_device *qpd); 155 struct qcm_process_device *qpd); 165 struct qcm_process_device *qpd); 169 struct qcm_process_device *qpd); 180 struct qcm_process_device *qpd); 217 struct qcm_process_device *qpd); 310 struct qcm_process_device *qpd); [all …]
|
| A D | kfd_packet_manager_vi.c | 43 struct qcm_process_device *qpd) in pm_map_process_vi() argument 45 struct kfd_process_device *pdd = qpd_to_pdd(qpd); in pm_map_process_vi() 54 packet->bitfields2.diq_enable = (qpd->is_debug) ? 1 : 0; in pm_map_process_vi() 58 packet->bitfields10.gds_size = qpd->gds_size; in pm_map_process_vi() 59 packet->bitfields10.num_gws = qpd->num_gws; in pm_map_process_vi() 60 packet->bitfields10.num_oac = qpd->num_oac; in pm_map_process_vi() 61 packet->bitfields10.num_queues = (qpd->is_debug) ? 0 : qpd->queue_count; in pm_map_process_vi() 63 packet->sh_mem_config = qpd->sh_mem_config; in pm_map_process_vi() 64 packet->sh_mem_bases = qpd->sh_mem_bases; in pm_map_process_vi() 65 packet->sh_mem_ape1_base = qpd->sh_mem_ape1_base; in pm_map_process_vi() [all …]
|
| A D | kfd_packet_manager.c | 147 struct qcm_process_device *qpd; in pm_create_runlist_ib() local 158 qpd = cur->qpd; in pm_create_runlist_ib() 161 xnack_enabled = qpd->pqm->process->xnack_enabled; in pm_create_runlist_ib() 184 qpd = cur->qpd; in pm_create_runlist_ib() 186 if (qpd->pqm->process->xnack_enabled != xnack_enabled) in pm_create_runlist_ib() 203 list_for_each_entry(kq, &qpd->priv_queue_list, list) { in pm_create_runlist_ib() 209 kq->queue->queue, qpd->is_debug); in pm_create_runlist_ib() 214 qpd->is_debug); in pm_create_runlist_ib() 223 list_for_each_entry(q, &qpd->queues_list, list) { in pm_create_runlist_ib() 229 q->queue, qpd->is_debug); in pm_create_runlist_ib() [all …]
|
| A D | kfd_process_queue_manager.c | 131 if (gws && pdd->qpd.num_gws) in pqm_set_gws() 134 if (!gws && pdd->qpd.num_gws == 0) in pqm_set_gws() 208 pdd->qpd.num_gws = 0; in pqm_clean_queue_resource() 351 if (pdd->qpd.queue_count >= max_queues) in pqm_create_queue() 364 if (list_empty(&pdd->qpd.queues_list) && in pqm_create_queue() 365 list_empty(&pdd->qpd.priv_queue_list)) in pqm_create_queue() 443 kq, &pdd->qpd); in pqm_create_queue() 471 pdd->qpd.proc_doorbells, in pqm_create_queue() 499 if (list_empty(&pdd->qpd.queues_list) && in pqm_create_queue() 500 list_empty(&pdd->qpd.priv_queue_list)) in pqm_create_queue() [all …]
|
| A D | kfd_flat_memory.c | 336 pdd->qpd.cwsr_base = SVM_CWSR_BASE; in kfd_init_apertures_vi() 337 pdd->qpd.ib_base = SVM_IB_BASE; in kfd_init_apertures_vi() 359 pdd->qpd.cwsr_base = AMDGPU_VA_RESERVED_TRAP_START(pdd->dev->adev); in kfd_init_apertures_v9()
|
| A D | kfd_debug.c | 613 kfd_process_set_trap_debug_flag(&pdd->qpd, false); in kfd_dbg_trap_deactivate() 626 release_debug_trap_vmid(pdd->dev->dqm, &pdd->qpd)) in kfd_dbg_trap_deactivate() 705 r = reserve_debug_trap_vmid(pdd->dev->dqm, &pdd->qpd); in kfd_dbg_trap_activate() 744 kfd_process_set_trap_debug_flag(&pdd->qpd, true); in kfd_dbg_trap_activate() 785 if (pdd->qpd.num_gws && (!kfd_dbg_has_gws_support(pdd->dev) || in kfd_dbg_trap_enable() 951 struct qcm_process_device *qpd = &pdd->qpd; in kfd_dbg_trap_query_exception_info() local 953 list_for_each_entry(queue, &qpd->queues_list, list) { in kfd_dbg_trap_query_exception_info()
|
| A D | kfd_priv.h | 753 struct qcm_process_device qpd; member 860 #define qpd_to_pdd(x) container_of(x, struct kfd_process_device, qpd) 1189 void kfd_process_set_trap_handler(struct qcm_process_device *qpd, 1192 void kfd_process_set_trap_debug_flag(struct qcm_process_device *qpd, 1428 struct qcm_process_device *qpd);
|
| A D | kfd_chardev.c | 354 if (!pdd->qpd.proc_doorbells) { in kfd_ioctl_create_queue() 604 &pdd->qpd, in kfd_ioctl_set_memory_policy() 640 kfd_process_set_trap_handler(&pdd->qpd, args->tba_addr, args->tma_addr); in kfd_ioctl_set_trap_handler() 915 pdd->qpd.sh_hidden_private_base = args->va_addr; in kfd_ioctl_set_scratch_backing_va() 920 pdd->qpd.vmid != 0 && dev->kfd2kgd->set_scratch_backing_va) in kfd_ioctl_set_scratch_backing_va() 922 dev->adev, args->va_addr, pdd->qpd.vmid); in kfd_ioctl_set_scratch_backing_va() 2277 if (!pdd->qpd.proc_doorbells) { in criu_restore_devices() 2773 if (pdd->qpd.queue_count) in runtime_enable()
|
| /drivers/infiniband/hw/cxgb4/ |
| A D | device.c | 253 space = qpd->bufsize - qpd->pos - 1; in dump_qp() 267 cc = snprintf(qpd->buf + qpd->pos, space, in dump_qp() 288 cc = snprintf(qpd->buf + qpd->pos, space, in dump_qp() 304 cc = snprintf(qpd->buf + qpd->pos, space, in dump_qp() 317 if (!qpd) { in qp_release() 322 kfree(qpd); in qp_release() 333 qpd = kmalloc(sizeof(*qpd), GFP_KERNEL); in qp_open() 334 if (!qpd) in qp_open() 348 qpd->buf = vmalloc(qpd->bufsize); in qp_open() 350 kfree(qpd); in qp_open() [all …]
|