Home
last modified time | relevance | path

Searched refs:to_gt (Results 1 – 25 of 82) sorted by relevance

1234

/drivers/gpu/drm/i915/selftests/
A Dmock_gem_device.c50 struct intel_gt *gt = to_gt(i915); in mock_device_flush()
69 intel_gt_driver_remove(to_gt(i915)); in mock_device_release()
73 mock_fini_ggtt(to_gt(i915)->ggtt); in mock_device_release()
213 to_gt(i915)->awake = INTEL_WAKEREF_MOCK_GT; in mock_gem_device()
231 ret = intel_gt_assign_ggtt(to_gt(i915)); in mock_gem_device()
235 mock_init_ggtt(to_gt(i915)); in mock_gem_device()
236 to_gt(i915)->vm = i915_vm_get(&to_gt(i915)->ggtt->vm); in mock_gem_device()
238 to_gt(i915)->info.engine_mask = BIT(0); in mock_gem_device()
241 if (!to_gt(i915)->engine[RCS0]) in mock_gem_device()
244 if (mock_engine_init(to_gt(i915)->engine[RCS0])) in mock_gem_device()
[all …]
A Di915_gem.c45 struct i915_ggtt *ggtt = to_gt(i915)->ggtt; in trash_stolen()
107 i915_ggtt_suspend(to_gt(i915)->ggtt); in igt_pm_suspend()
117 i915_ggtt_suspend(to_gt(i915)->ggtt); in igt_pm_hibernate()
133 i915_ggtt_resume(to_gt(i915)->ggtt); in igt_pm_resume()
135 setup_private_pat(to_gt(i915)); in igt_pm_resume()
258 if (intel_gt_is_wedged(to_gt(i915))) in i915_gem_live_selftests()
A Di915_selftest.c163 struct intel_huc *huc = &to_gt(i915)->uc.huc; in __wait_gsc_huc_load_completed()
271 if (IS_GFX_GT_IP_RANGE(to_gt(i915), IP_VER(12, 00), IP_VER(12, 74))) { in i915_live_selftests()
379 if (intel_gt_pm_wait_for_idle(to_gt(i915))) in __i915_live_setup()
382 return intel_gt_terminally_wedged(to_gt(i915)); in __i915_live_setup()
A Dmock_uncore.c45 intel_uncore_init_early(uncore, to_gt(i915)); in mock_uncore_init()
/drivers/gpu/drm/i915/
A Di915_getparam.c21 const struct sseu_dev_info *sseu = &to_gt(i915)->info.sseu; in i915_getparam_ioctl()
39 value = to_gt(i915)->ggtt->num_fences; in i915_getparam_ioctl()
90 intel_has_gpu_reset(to_gt(i915)); in i915_getparam_ioctl()
91 if (value && intel_has_reset_engine(to_gt(i915))) in i915_getparam_ioctl()
108 value = intel_huc_check_status(&to_gt(i915)->uc.huc); in i915_getparam_ioctl()
160 if (intel_uc_uses_guc_submission(&to_gt(i915)->uc)) in i915_getparam_ioctl()
188 value = to_gt(i915)->clock_frequency; in i915_getparam_ioctl()
A Di915_debugfs.c71 intel_gt_info_print(&to_gt(i915)->info, &p); in i915_capabilities()
288 struct intel_gt *gt = to_gt(i915); in i915_frequency_info()
327 swizzle_string(to_gt(dev_priv)->ggtt->bit_6_swizzle_x)); in i915_swizzle_info()
329 swizzle_string(to_gt(dev_priv)->ggtt->bit_6_swizzle_y)); in i915_swizzle_info()
376 struct intel_rps *rps = &to_gt(dev_priv)->rps; in i915_rps_boost_info()
444 str_yes_no(to_gt(i915)->awake), in i915_engine_info()
445 atomic_read(&to_gt(i915)->wakeref.count), in i915_engine_info()
446 ktime_to_ms(intel_gt_get_awake_time(to_gt(i915)))); in i915_engine_info()
448 to_gt(i915)->clock_frequency, in i915_engine_info()
449 to_gt(i915)->clock_period_ns); in i915_engine_info()
[all …]
A Di915_irq.c160 struct intel_gt *gt = to_gt(dev_priv); in ivb_parity_work()
564 struct intel_gt *gt = to_gt(i915); in gen11_irq_handler()
621 struct intel_gt *gt = to_gt(i915); in dg1_irq_handler()
678 gen5_gt_irq_reset(to_gt(dev_priv)); in ilk_irq_reset()
690 gen5_gt_irq_reset(to_gt(dev_priv)); in valleyview_irq_reset()
702 gen8_gt_irq_reset(to_gt(dev_priv)); in gen8_irq_reset()
710 struct intel_gt *gt = to_gt(dev_priv); in gen11_irq_reset()
750 gen8_gt_irq_reset(to_gt(dev_priv)); in cherryview_irq_reset()
761 gen5_gt_irq_postinstall(to_gt(dev_priv)); in ilk_irq_postinstall()
770 gen5_gt_irq_postinstall(to_gt(dev_priv)); in valleyview_irq_postinstall()
[all …]
A Di915_gem.c93 struct i915_ggtt *ggtt = to_gt(i915)->ggtt; in i915_gem_get_aperture_ioctl()
306 struct i915_ggtt *ggtt = to_gt(i915)->ggtt; in i915_gem_gtt_prepare()
367 struct i915_ggtt *ggtt = to_gt(i915)->ggtt; in i915_gem_gtt_cleanup()
383 struct i915_ggtt *ggtt = to_gt(i915)->ggtt; in i915_gem_gtt_pread()
546 struct i915_ggtt *ggtt = to_gt(i915)->ggtt; in i915_gem_gtt_pwrite_fast()
855 &to_gt(i915)->ggtt->userfault_list, userfault_link) in i915_gem_runtime_suspend()
867 for (i = 0; i < to_gt(i915)->ggtt->num_fences; i++) { in i915_gem_runtime_suspend()
868 struct i915_fence_reg *reg = &to_gt(i915)->ggtt->fence_regs[i]; in i915_gem_runtime_suspend()
909 struct i915_ggtt *ggtt = to_gt(i915)->ggtt; in i915_gem_object_ggtt_pin_ww()
1237 i915_ggtt_resume(to_gt(dev_priv)->ggtt); in i915_gem_init()
A Di915_gem_gtt.c59 struct i915_ggtt *ggtt = to_gt(i915)->ggtt; in i915_gem_gtt_finish_pages()
108 GEM_BUG_ON(vm == &to_gt(vm->i915)->ggtt->alias->vm); in i915_gem_gtt_reserve()
208 GEM_BUG_ON(vm == &to_gt(vm->i915)->ggtt->alias->vm); in i915_gem_gtt_insert()
A Di915_query.c93 const struct sseu_dev_info *sseu = &to_gt(dev_priv)->info.sseu; in query_topology_info()
535 struct intel_gt *gt = to_gt(i915); in query_hwconfig_blob()
561 struct intel_guc *guc = &to_gt(i915)->uc.guc; in query_guc_submission_version()
565 if (!intel_uc_uses_guc_submission(&to_gt(i915)->uc)) in query_guc_submission_version()
A Dintel_gvt.c89 *mmio = intel_uncore_read_notrace(to_gt(dev_priv)->uncore, in save_mmio()
168 if (intel_uc_wants_guc_submission(&to_gt(dev_priv)->uc)) { in intel_gvt_init_device()
/drivers/gpu/drm/i915/gem/
A Di915_gem_tiling.c186 struct i915_ggtt *ggtt = to_gt(i915)->ggtt; in i915_gem_object_fence_prepare()
225 return to_gt(i915)->ggtt->bit_6_swizzle_x == I915_BIT_6_SWIZZLE_9_10_17 && in i915_gem_object_needs_bit17_swizzle()
350 if (!to_gt(i915)->ggtt->num_fences) in i915_gem_set_tiling_ioctl()
376 args->swizzle_mode = to_gt(i915)->ggtt->bit_6_swizzle_x; in i915_gem_set_tiling_ioctl()
378 args->swizzle_mode = to_gt(i915)->ggtt->bit_6_swizzle_y; in i915_gem_set_tiling_ioctl()
433 if (!to_gt(i915)->ggtt->num_fences) in i915_gem_get_tiling_ioctl()
449 args->swizzle_mode = to_gt(i915)->ggtt->bit_6_swizzle_x; in i915_gem_get_tiling_ioctl()
452 args->swizzle_mode = to_gt(i915)->ggtt->bit_6_swizzle_y; in i915_gem_get_tiling_ioctl()
A Di915_gem_mman.c346 struct i915_ggtt *ggtt = to_gt(i915)->ggtt; in vm_fault_gtt()
469 mutex_lock(&to_gt(i915)->ggtt->vm.mutex); in vm_fault_gtt()
471 list_add(&obj->userfault_link, &to_gt(i915)->ggtt->userfault_list); in vm_fault_gtt()
472 mutex_unlock(&to_gt(i915)->ggtt->vm.mutex); in vm_fault_gtt()
593 mutex_lock(&to_gt(i915)->ggtt->vm.mutex); in i915_gem_object_release_mmap_gtt()
611 mutex_unlock(&to_gt(i915)->ggtt->vm.mutex); in i915_gem_object_release_mmap_gtt()
746 err = intel_gt_retire_requests_timeout(to_gt(i915), MAX_SCHEDULE_TIMEOUT, in mmap_offset_attach()
840 else if (!i915_ggtt_has_aperture(to_gt(i915)->ggtt)) in i915_gem_dumb_mmap_offset()
888 if (!i915_ggtt_has_aperture(to_gt(i915)->ggtt)) in i915_gem_mmap_offset_ioctl()
1121 struct i915_ggtt *ggtt = to_gt(i915)->ggtt; in i915_gem_fb_mmap()
A Di915_gem_ttm_move.c202 if (!to_gt(i915)->migrate.context || intel_gt_is_wedged(to_gt(i915))) in i915_ttm_accel_move()
215 intel_engine_pm_get(to_gt(i915)->migrate.context->engine); in i915_ttm_accel_move()
216 ret = intel_context_migrate_clear(to_gt(i915)->migrate.context, deps, in i915_ttm_accel_move()
229 intel_engine_pm_get(to_gt(i915)->migrate.context->engine); in i915_ttm_accel_move()
230 ret = intel_context_migrate_copy(to_gt(i915)->migrate.context, in i915_ttm_accel_move()
242 intel_engine_pm_put(to_gt(i915)->migrate.context->engine); in i915_ttm_accel_move()
A Di915_gem_phys.c79 intel_gt_chipset_flush(to_gt(i915)); in i915_gem_object_get_pages_phys()
163 intel_gt_chipset_flush(to_gt(i915)); in i915_gem_object_pwrite_phys()
/drivers/gpu/drm/i915/gem/selftests/
A Di915_gem_mman.c155 intel_gt_flush_ggtt_writes(to_gt(i915)); in check_partial_mapping()
251 intel_gt_flush_ggtt_writes(to_gt(i915)); in check_partial_mappings()
633 intel_gt_pm_get_untracked(to_gt(i915)); in disable_retire_worker()
640 intel_gt_pm_put_untracked(to_gt(i915)); in restore_retire_worker()
668 GEM_BUG_ON(!to_gt(i915)->awake); in igt_mmap_offset_exhaustion()
669 intel_gt_retire_requests(to_gt(i915)); in igt_mmap_offset_exhaustion()
745 if (intel_gt_is_wedged(to_gt(i915))) in igt_mmap_offset_exhaustion()
962 intel_gt_flush_ggtt_writes(to_gt(i915)); in __igt_mmap()
1466 intel_gt_flush_ggtt_writes(to_gt(i915)); in __igt_mmap_access()
1482 intel_gt_flush_ggtt_writes(to_gt(i915)); in __igt_mmap_access()
[all …]
/drivers/gpu/drm/i915/gt/
A Dselftest_gt_pm.c192 if (intel_gt_is_wedged(to_gt(i915))) in intel_gt_pm_live_selftests()
195 return intel_gt_live_subtests(tests, to_gt(i915)); in intel_gt_pm_live_selftests()
209 if (intel_gt_is_wedged(to_gt(i915))) in intel_gt_pm_late_selftests()
212 return intel_gt_live_subtests(tests, to_gt(i915)); in intel_gt_pm_late_selftests()
A Dmock_engine.c348 GEM_BUG_ON(!to_gt(i915)->uncore); in mock_engine()
356 engine->base.gt = to_gt(i915); in mock_engine()
357 engine->base.uncore = to_gt(i915)->uncore; in mock_engine()
380 to_gt(i915)->engine[id] = &engine->base; in mock_engine()
381 to_gt(i915)->engine_class[0][id] = &engine->base; in mock_engine()
A Dselftest_engine.c15 struct intel_gt *gt = to_gt(i915); in intel_engine_live_selftests()
A Dselftest_ring_submission.c294 if (to_gt(i915)->submission_method > INTEL_SUBMISSION_RING) in intel_ring_submission_live_selftests()
297 return intel_gt_live_subtests(tests, to_gt(i915)); in intel_ring_submission_live_selftests()
A Dselftest_engine_heartbeat.c271 if (intel_gt_is_wedged(to_gt(i915))) in intel_heartbeat_live_selftests()
277 err = intel_gt_live_subtests(tests, to_gt(i915)); in intel_heartbeat_live_selftests()
A Dintel_ggtt.c100 ret = ggtt_init_hw(to_gt(i915)->ggtt); in i915_ggtt_init_hw()
1064 ret = init_ggtt(to_gt(i915)->ggtt); in i915_init_ggtt()
1069 ret = init_aliasing_ppgtt(to_gt(i915)->ggtt); in i915_init_ggtt()
1071 cleanup_init_ggtt(to_gt(i915)->ggtt); in i915_init_ggtt()
1124 struct i915_ggtt *ggtt = to_gt(i915)->ggtt; in i915_ggtt_driver_release()
1139 struct i915_ggtt *ggtt = to_gt(i915)->ggtt; in i915_ggtt_driver_late_release()
1573 ret = ggtt_probe_hw(to_gt(i915)->ggtt, to_gt(i915)); in i915_ggtt_probe_hw()
/drivers/gpu/drm/i915/pxp/
A Dintel_pxp.c165 intel_huc_is_loaded_by_gsc(&to_gt(i915)->uc.huc) && intel_uc_uses_huc(&to_gt(i915)->uc)) in find_gt_for_required_teelink()
166 return to_gt(i915); in find_gt_for_required_teelink()
191 if (IS_ENABLED(CONFIG_INTEL_MEI_PXP) && !i915->media_gt && VDBOX_MASK(to_gt(i915))) in find_gt_for_required_protected_content()
192 return to_gt(i915); in find_gt_for_required_protected_content()
202 if (intel_gt_is_wedged(to_gt(i915))) in intel_pxp_init()
/drivers/gpu/drm/i915/display/
A Dintel_display_rps.c85 intel_rps_mark_interactive(&to_gt(i915)->rps, interactive); in intel_display_rps_mark_interactive()
107 gen5_rps_irq_handler(&to_gt(i915)->rps); in ilk_display_rps_irq_handler()
A Dintel_plane_initial.c74 struct i915_ggtt *ggtt = to_gt(i915)->ggtt; in initial_plane_phys()
202 struct i915_ggtt *ggtt = to_gt(i915)->ggtt; in initial_plane_vma()
217 vma = i915_vma_instance(obj, &to_gt(i915)->ggtt->vm, NULL); in initial_plane_vma()

Completed in 59 milliseconds

1234