Lines Matching refs:pdd
168 struct kfd_process_device *pdd = qpd_to_pdd(qpd); in kfd_hws_hang() local
170 pdd->has_reset_queue = true; in kfd_hws_hang()
204 struct kfd_process_device *pdd = qpd_to_pdd(qpd); in add_queue_mes() local
215 queue_input.process_id = pdd->pasid; in add_queue_mes()
221 queue_input.process_context_addr = pdd->proc_ctx_gpu_addr; in add_queue_mes()
534 struct kfd_process_device *pdd = qpd_to_pdd(qpd); in allocate_vmid() local
553 dqm->vmid_pasid[allocated_vmid] = pdd->pasid; in allocate_vmid()
555 set_pasid_vmid_mapping(dqm, pdd->pasid, allocated_vmid); in allocate_vmid()
791 struct kfd_process_device *pdd; in dbgdev_wave_reset_wavefronts() local
808 pdd = kfd_get_process_device_data(dev, p); in dbgdev_wave_reset_wavefronts()
809 if (!pdd) in dbgdev_wave_reset_wavefronts()
821 if (status && queried_pasid == pdd->pasid) { in dbgdev_wave_reset_wavefronts()
918 struct kfd_process_device *pdd = qpd_to_pdd(qpd); in destroy_queue_nocpsch() local
935 pdd->sdma_past_activity_counter += sdma_val; in destroy_queue_nocpsch()
949 struct kfd_process_device *pdd; in update_queue() local
953 pdd = kfd_get_process_device_data(q->device, q->process); in update_queue()
954 if (!pdd) { in update_queue()
970 retval = remove_queue_mes(dqm, q, &pdd->qpd); in update_queue()
973 if (pdd->has_reset_queue) { in update_queue()
1012 increment_queue_count(dqm, &pdd->qpd, q); in update_queue()
1014 decrement_queue_count(dqm, &pdd->qpd, q); in update_queue()
1018 pdd->qpd.mapped_gws_queue = true; in update_queue()
1024 pdd->qpd.mapped_gws_queue = false; in update_queue()
1033 retval = add_queue_mes(dqm, q, &pdd->qpd); in update_queue()
1061 struct kfd_process_device *pdd, in suspend_single_queue() argument
1070 pdd->process->lead_thread->pid, in suspend_single_queue()
1085 int r = remove_queue_mes(dqm, q, &pdd->qpd); in suspend_single_queue()
1091 decrement_queue_count(dqm, &pdd->qpd, q); in suspend_single_queue()
1110 struct kfd_process_device *pdd; in resume_single_queue() local
1115 pdd = qpd_to_pdd(qpd); in resume_single_queue()
1118 pdd->process->lead_thread->pid, in resume_single_queue()
1125 int r = add_queue_mes(dqm, q, &pdd->qpd); in resume_single_queue()
1143 struct kfd_process_device *pdd; in evict_process_queues_nocpsch() local
1150 pdd = qpd_to_pdd(qpd); in evict_process_queues_nocpsch()
1152 pdd->process->lead_thread->pid); in evict_process_queues_nocpsch()
1154 pdd->last_evict_timestamp = get_jiffies_64(); in evict_process_queues_nocpsch()
1193 struct kfd_process_device *pdd; in evict_process_queues_cpsch() local
1200 pdd = qpd_to_pdd(qpd); in evict_process_queues_cpsch()
1206 if (!pdd->drm_priv) in evict_process_queues_cpsch()
1210 pdd->process->lead_thread->pid); in evict_process_queues_cpsch()
1234 pdd->last_evict_timestamp = get_jiffies_64(); in evict_process_queues_cpsch()
1253 struct kfd_process_device *pdd; in restore_process_queues_nocpsch() local
1258 pdd = qpd_to_pdd(qpd); in restore_process_queues_nocpsch()
1260 pd_base = amdgpu_amdkfd_gpuvm_get_process_page_dir(pdd->drm_priv); in restore_process_queues_nocpsch()
1271 pdd->process->lead_thread->pid); in restore_process_queues_nocpsch()
1282 kfd_flush_tlb(pdd, TLB_FLUSH_LEGACY); in restore_process_queues_nocpsch()
1288 mm = get_task_mm(pdd->process->lead_thread); in restore_process_queues_nocpsch()
1319 eviction_duration = get_jiffies_64() - pdd->last_evict_timestamp; in restore_process_queues_nocpsch()
1320 atomic64_add(eviction_duration, &pdd->evict_duration_counter); in restore_process_queues_nocpsch()
1333 struct kfd_process_device *pdd; in restore_process_queues_cpsch() local
1337 pdd = qpd_to_pdd(qpd); in restore_process_queues_cpsch()
1351 if (!pdd->drm_priv) in restore_process_queues_cpsch()
1355 pdd->process->lead_thread->pid); in restore_process_queues_cpsch()
1358 qpd->page_table_base = amdgpu_amdkfd_gpuvm_get_process_page_dir(pdd->drm_priv); in restore_process_queues_cpsch()
1368 increment_queue_count(dqm, &pdd->qpd, q); in restore_process_queues_cpsch()
1382 eviction_duration = get_jiffies_64() - pdd->last_evict_timestamp; in restore_process_queues_cpsch()
1383 atomic64_add(eviction_duration, &pdd->evict_duration_counter); in restore_process_queues_cpsch()
1395 struct kfd_process_device *pdd; in register_process() local
1405 pdd = qpd_to_pdd(qpd); in register_process()
1407 pd_base = amdgpu_amdkfd_gpuvm_get_process_page_dir(pdd->drm_priv); in register_process()
2128 struct kfd_process_device *pdd = qpd_to_pdd(qpd); in set_queue_as_reset() local
2131 q->properties.queue_id, pdd->process->lead_thread->pid); in set_queue_as_reset()
2133 pdd->has_reset_queue = true; in set_queue_as_reset()
2214 struct kfd_process_device *pdd; in reset_hung_queues() local
2222 pdd = kfd_get_process_device_data(dqm->dev, q->process); in reset_hung_queues()
2223 if (!pdd) { in reset_hung_queues()
2238 set_queue_as_reset(dqm, q, &pdd->qpd); in reset_hung_queues()
2480 struct kfd_process_device *pdd = kfd_get_process_device_data(q->device, in wait_on_destroy_queue() local
2484 if (WARN_ON(!pdd)) in wait_on_destroy_queue()
2487 if (pdd->qpd.is_debug) in wait_on_destroy_queue()
2492 if (pdd->process->debug_trap_enabled && q->properties.is_suspended) { in wait_on_destroy_queue()
2512 struct kfd_process_device *pdd = qpd_to_pdd(qpd); in destroy_queue_cpsch() local
2553 pdd->sdma_past_activity_counter += sdma_val; in destroy_queue_cpsch()
3052 struct kfd_process_device *pdd = NULL; in kfd_dqm_suspend_bad_queue_mes() local
3053 struct kfd_process *p = kfd_lookup_process_by_pasid(pasid, &pdd); in kfd_dqm_suspend_bad_queue_mes()
3060 if (!pdd) in kfd_dqm_suspend_bad_queue_mes()
3065 if (pdd) { in kfd_dqm_suspend_bad_queue_mes()
3066 qpd = &pdd->qpd; in kfd_dqm_suspend_bad_queue_mes()
3139 int kfd_evict_process_device(struct kfd_process_device *pdd) in kfd_evict_process_device() argument
3145 p = pdd->process; in kfd_evict_process_device()
3146 dqm = pdd->dev->dqm; in kfd_evict_process_device()
3151 ret = kfd_dqm_evict_pasid_mes(dqm, &pdd->qpd); in kfd_evict_process_device()
3153 ret = dqm->ops.evict_process_queues(dqm, &pdd->qpd); in kfd_evict_process_device()
3303 struct kfd_process_device *pdd = p->pdds[i]; in copy_context_work_handler() local
3304 struct device_queue_manager *dqm = pdd->dev->dqm; in copy_context_work_handler()
3305 struct qcm_process_device *qpd = &pdd->qpd; in copy_context_work_handler()
3360 struct kfd_process_device *pdd = p->pdds[i]; in resume_queues() local
3361 struct device_queue_manager *dqm = pdd->dev->dqm; in resume_queues()
3363 struct qcm_process_device *qpd = &pdd->qpd; in resume_queues()
3380 int err = resume_single_queue(dqm, &pdd->qpd, q); in resume_queues()
3463 struct kfd_process_device *pdd = p->pdds[i]; in suspend_queues() local
3464 struct device_queue_manager *dqm = pdd->dev->dqm; in suspend_queues()
3466 struct qcm_process_device *qpd = &pdd->qpd; in suspend_queues()
3480 int err = suspend_single_queue(dqm, pdd, q); in suspend_queues()