Lines Matching refs:pc

90 static struct xe_guc *pc_to_guc(struct xe_guc_pc *pc)  in pc_to_guc()  argument
92 return container_of(pc, struct xe_guc, pc); in pc_to_guc()
95 static struct xe_guc_ct *pc_to_ct(struct xe_guc_pc *pc) in pc_to_ct() argument
97 return &pc_to_guc(pc)->ct; in pc_to_ct()
100 static struct xe_gt *pc_to_gt(struct xe_guc_pc *pc) in pc_to_gt() argument
102 return guc_to_gt(pc_to_guc(pc)); in pc_to_gt()
105 static struct xe_device *pc_to_xe(struct xe_guc_pc *pc) in pc_to_xe() argument
107 return guc_to_xe(pc_to_guc(pc)); in pc_to_xe()
110 static struct iosys_map *pc_to_maps(struct xe_guc_pc *pc) in pc_to_maps() argument
112 return &pc->bo->vmap; in pc_to_maps()
127 static int wait_for_pc_state(struct xe_guc_pc *pc, in wait_for_pc_state() argument
134 xe_device_assert_mem_access(pc_to_xe(pc)); in wait_for_pc_state()
137 if (slpc_shared_data_read(pc, header.global_state) == state) in wait_for_pc_state()
150 static int wait_for_flush_complete(struct xe_guc_pc *pc) in wait_for_flush_complete() argument
154 if (!wait_var_event_timeout(&pc->flush_freq_limit, in wait_for_flush_complete()
155 !atomic_read(&pc->flush_freq_limit), in wait_for_flush_complete()
162 static int wait_for_act_freq_limit(struct xe_guc_pc *pc, u32 freq) in wait_for_act_freq_limit() argument
168 if (xe_guc_pc_get_act_freq(pc) <= freq) in wait_for_act_freq_limit()
180 static int pc_action_reset(struct xe_guc_pc *pc) in pc_action_reset() argument
182 struct xe_guc_ct *ct = pc_to_ct(pc); in pc_action_reset()
186 xe_bo_ggtt_addr(pc->bo), in pc_action_reset()
192 if (ret && !(xe_device_wedged(pc_to_xe(pc)) && ret == -ECANCELED)) in pc_action_reset()
193 xe_gt_err(pc_to_gt(pc), "GuC PC reset failed: %pe\n", in pc_action_reset()
199 static int pc_action_query_task_state(struct xe_guc_pc *pc) in pc_action_query_task_state() argument
201 struct xe_guc_ct *ct = pc_to_ct(pc); in pc_action_query_task_state()
205 xe_bo_ggtt_addr(pc->bo), in pc_action_query_task_state()
210 if (wait_for_pc_state(pc, SLPC_GLOBAL_STATE_RUNNING, in pc_action_query_task_state()
216 if (ret && !(xe_device_wedged(pc_to_xe(pc)) && ret == -ECANCELED)) in pc_action_query_task_state()
217 xe_gt_err(pc_to_gt(pc), "GuC PC query task state failed: %pe\n", in pc_action_query_task_state()
223 static int pc_action_set_param(struct xe_guc_pc *pc, u8 id, u32 value) in pc_action_set_param() argument
225 struct xe_guc_ct *ct = pc_to_ct(pc); in pc_action_set_param()
234 if (wait_for_pc_state(pc, SLPC_GLOBAL_STATE_RUNNING, in pc_action_set_param()
239 if (ret && !(xe_device_wedged(pc_to_xe(pc)) && ret == -ECANCELED)) in pc_action_set_param()
240 xe_gt_err(pc_to_gt(pc), "GuC PC set param[%u]=%u failed: %pe\n", in pc_action_set_param()
246 static int pc_action_unset_param(struct xe_guc_pc *pc, u8 id) in pc_action_unset_param() argument
253 struct xe_guc_ct *ct = &pc_to_guc(pc)->ct; in pc_action_unset_param()
256 if (wait_for_pc_state(pc, SLPC_GLOBAL_STATE_RUNNING, in pc_action_unset_param()
261 if (ret && !(xe_device_wedged(pc_to_xe(pc)) && ret == -ECANCELED)) in pc_action_unset_param()
262 xe_gt_err(pc_to_gt(pc), "GuC PC unset param failed: %pe", in pc_action_unset_param()
268 static int pc_action_setup_gucrc(struct xe_guc_pc *pc, u32 mode) in pc_action_setup_gucrc() argument
270 struct xe_guc_ct *ct = pc_to_ct(pc); in pc_action_setup_gucrc()
278 if (ret && !(xe_device_wedged(pc_to_xe(pc)) && ret == -ECANCELED)) in pc_action_setup_gucrc()
279 xe_gt_err(pc_to_gt(pc), "GuC RC enable mode=%u failed: %pe\n", in pc_action_setup_gucrc()
296 static u32 pc_get_min_freq(struct xe_guc_pc *pc) in pc_get_min_freq() argument
301 slpc_shared_data_read(pc, task_state_data.freq)); in pc_get_min_freq()
306 static void pc_set_manual_rp_ctrl(struct xe_guc_pc *pc, bool enable) in pc_set_manual_rp_ctrl() argument
308 struct xe_gt *gt = pc_to_gt(pc); in pc_set_manual_rp_ctrl()
315 static void pc_set_cur_freq(struct xe_guc_pc *pc, u32 freq) in pc_set_cur_freq() argument
317 struct xe_gt *gt = pc_to_gt(pc); in pc_set_cur_freq()
320 pc_set_manual_rp_ctrl(pc, true); in pc_set_cur_freq()
329 pc_set_manual_rp_ctrl(pc, false); in pc_set_cur_freq()
332 static int pc_set_min_freq(struct xe_guc_pc *pc, u32 freq) in pc_set_min_freq() argument
338 if (freq < pc->rpn_freq || freq > pc->rp0_freq) in pc_set_min_freq()
345 pc_action_set_param(pc, SLPC_PARAM_IGNORE_EFFICIENT_FREQUENCY, in pc_set_min_freq()
346 freq < pc->rpe_freq); in pc_set_min_freq()
348 return pc_action_set_param(pc, in pc_set_min_freq()
353 static int pc_get_max_freq(struct xe_guc_pc *pc) in pc_get_max_freq() argument
358 slpc_shared_data_read(pc, task_state_data.freq)); in pc_get_max_freq()
363 static int pc_set_max_freq(struct xe_guc_pc *pc, u32 freq) in pc_set_max_freq() argument
370 if (freq < pc->rpn_freq || freq > pc->rp0_freq) in pc_set_max_freq()
373 return pc_action_set_param(pc, in pc_set_max_freq()
378 static void mtl_update_rpa_value(struct xe_guc_pc *pc) in mtl_update_rpa_value() argument
380 struct xe_gt *gt = pc_to_gt(pc); in mtl_update_rpa_value()
388 pc->rpa_freq = decode_freq(REG_FIELD_GET(MTL_RPA_MASK, reg)); in mtl_update_rpa_value()
391 static void mtl_update_rpe_value(struct xe_guc_pc *pc) in mtl_update_rpe_value() argument
393 struct xe_gt *gt = pc_to_gt(pc); in mtl_update_rpe_value()
401 pc->rpe_freq = decode_freq(REG_FIELD_GET(MTL_RPE_MASK, reg)); in mtl_update_rpe_value()
404 static void tgl_update_rpa_value(struct xe_guc_pc *pc) in tgl_update_rpa_value() argument
406 struct xe_gt *gt = pc_to_gt(pc); in tgl_update_rpa_value()
417 pc->rpa_freq = REG_FIELD_GET(RP0_MASK, reg) * GT_FREQUENCY_MULTIPLIER; in tgl_update_rpa_value()
420 pc->rpa_freq = REG_FIELD_GET(RPA_MASK, reg) * GT_FREQUENCY_MULTIPLIER; in tgl_update_rpa_value()
424 static void tgl_update_rpe_value(struct xe_guc_pc *pc) in tgl_update_rpe_value() argument
426 struct xe_gt *gt = pc_to_gt(pc); in tgl_update_rpe_value()
437 pc->rpe_freq = REG_FIELD_GET(RP1_MASK, reg) * GT_FREQUENCY_MULTIPLIER; in tgl_update_rpe_value()
440 pc->rpe_freq = REG_FIELD_GET(RPE_MASK, reg) * GT_FREQUENCY_MULTIPLIER; in tgl_update_rpe_value()
444 static void pc_update_rp_values(struct xe_guc_pc *pc) in pc_update_rp_values() argument
446 struct xe_gt *gt = pc_to_gt(pc); in pc_update_rp_values()
450 mtl_update_rpa_value(pc); in pc_update_rp_values()
451 mtl_update_rpe_value(pc); in pc_update_rp_values()
453 tgl_update_rpa_value(pc); in pc_update_rp_values()
454 tgl_update_rpe_value(pc); in pc_update_rp_values()
462 pc->rpn_freq = min(pc->rpn_freq, pc->rpe_freq); in pc_update_rp_values()
471 u32 xe_guc_pc_get_act_freq(struct xe_guc_pc *pc) in xe_guc_pc_get_act_freq() argument
473 struct xe_gt *gt = pc_to_gt(pc); in xe_guc_pc_get_act_freq()
506 u32 xe_guc_pc_get_cur_freq_fw(struct xe_guc_pc *pc) in xe_guc_pc_get_cur_freq_fw() argument
508 struct xe_gt *gt = pc_to_gt(pc); in xe_guc_pc_get_cur_freq_fw()
523 int xe_guc_pc_get_cur_freq(struct xe_guc_pc *pc, u32 *freq) in xe_guc_pc_get_cur_freq() argument
525 struct xe_gt *gt = pc_to_gt(pc); in xe_guc_pc_get_cur_freq()
550 u32 xe_guc_pc_get_rp0_freq(struct xe_guc_pc *pc) in xe_guc_pc_get_rp0_freq() argument
552 return pc->rp0_freq; in xe_guc_pc_get_rp0_freq()
561 u32 xe_guc_pc_get_rpa_freq(struct xe_guc_pc *pc) in xe_guc_pc_get_rpa_freq() argument
563 pc_update_rp_values(pc); in xe_guc_pc_get_rpa_freq()
565 return pc->rpa_freq; in xe_guc_pc_get_rpa_freq()
574 u32 xe_guc_pc_get_rpe_freq(struct xe_guc_pc *pc) in xe_guc_pc_get_rpe_freq() argument
576 pc_update_rp_values(pc); in xe_guc_pc_get_rpe_freq()
578 return pc->rpe_freq; in xe_guc_pc_get_rpe_freq()
587 u32 xe_guc_pc_get_rpn_freq(struct xe_guc_pc *pc) in xe_guc_pc_get_rpn_freq() argument
589 return pc->rpn_freq; in xe_guc_pc_get_rpn_freq()
592 static int xe_guc_pc_get_min_freq_locked(struct xe_guc_pc *pc, u32 *freq) in xe_guc_pc_get_min_freq_locked() argument
596 lockdep_assert_held(&pc->freq_lock); in xe_guc_pc_get_min_freq_locked()
599 if (!pc->freq_ready) in xe_guc_pc_get_min_freq_locked()
602 ret = pc_action_query_task_state(pc); in xe_guc_pc_get_min_freq_locked()
606 *freq = pc_get_min_freq(pc); in xe_guc_pc_get_min_freq_locked()
619 int xe_guc_pc_get_min_freq(struct xe_guc_pc *pc, u32 *freq) in xe_guc_pc_get_min_freq() argument
621 guard(mutex)(&pc->freq_lock); in xe_guc_pc_get_min_freq()
623 return xe_guc_pc_get_min_freq_locked(pc, freq); in xe_guc_pc_get_min_freq()
626 static int xe_guc_pc_set_min_freq_locked(struct xe_guc_pc *pc, u32 freq) in xe_guc_pc_set_min_freq_locked() argument
630 lockdep_assert_held(&pc->freq_lock); in xe_guc_pc_set_min_freq_locked()
633 if (!pc->freq_ready) in xe_guc_pc_set_min_freq_locked()
636 ret = pc_set_min_freq(pc, freq); in xe_guc_pc_set_min_freq_locked()
640 pc->user_requested_min = freq; in xe_guc_pc_set_min_freq_locked()
654 int xe_guc_pc_set_min_freq(struct xe_guc_pc *pc, u32 freq) in xe_guc_pc_set_min_freq() argument
656 guard(mutex)(&pc->freq_lock); in xe_guc_pc_set_min_freq()
658 return xe_guc_pc_set_min_freq_locked(pc, freq); in xe_guc_pc_set_min_freq()
661 static int xe_guc_pc_get_max_freq_locked(struct xe_guc_pc *pc, u32 *freq) in xe_guc_pc_get_max_freq_locked() argument
665 lockdep_assert_held(&pc->freq_lock); in xe_guc_pc_get_max_freq_locked()
668 if (!pc->freq_ready) in xe_guc_pc_get_max_freq_locked()
671 ret = pc_action_query_task_state(pc); in xe_guc_pc_get_max_freq_locked()
675 *freq = pc_get_max_freq(pc); in xe_guc_pc_get_max_freq_locked()
688 int xe_guc_pc_get_max_freq(struct xe_guc_pc *pc, u32 *freq) in xe_guc_pc_get_max_freq() argument
690 guard(mutex)(&pc->freq_lock); in xe_guc_pc_get_max_freq()
692 return xe_guc_pc_get_max_freq_locked(pc, freq); in xe_guc_pc_get_max_freq()
695 static int xe_guc_pc_set_max_freq_locked(struct xe_guc_pc *pc, u32 freq) in xe_guc_pc_set_max_freq_locked() argument
699 lockdep_assert_held(&pc->freq_lock); in xe_guc_pc_set_max_freq_locked()
702 if (!pc->freq_ready) in xe_guc_pc_set_max_freq_locked()
705 ret = pc_set_max_freq(pc, freq); in xe_guc_pc_set_max_freq_locked()
709 pc->user_requested_max = freq; in xe_guc_pc_set_max_freq_locked()
723 int xe_guc_pc_set_max_freq(struct xe_guc_pc *pc, u32 freq) in xe_guc_pc_set_max_freq() argument
725 if (XE_WA(pc_to_gt(pc), 22019338487)) { in xe_guc_pc_set_max_freq()
726 if (wait_for_flush_complete(pc) != 0) in xe_guc_pc_set_max_freq()
730 guard(mutex)(&pc->freq_lock); in xe_guc_pc_set_max_freq()
732 return xe_guc_pc_set_max_freq_locked(pc, freq); in xe_guc_pc_set_max_freq()
739 enum xe_gt_idle_state xe_guc_pc_c_status(struct xe_guc_pc *pc) in xe_guc_pc_c_status() argument
741 struct xe_gt *gt = pc_to_gt(pc); in xe_guc_pc_c_status()
766 u64 xe_guc_pc_rc6_residency(struct xe_guc_pc *pc) in xe_guc_pc_rc6_residency() argument
768 struct xe_gt *gt = pc_to_gt(pc); in xe_guc_pc_rc6_residency()
780 u64 xe_guc_pc_mc6_residency(struct xe_guc_pc *pc) in xe_guc_pc_mc6_residency() argument
782 struct xe_gt *gt = pc_to_gt(pc); in xe_guc_pc_mc6_residency()
790 static void mtl_init_fused_rp_values(struct xe_guc_pc *pc) in mtl_init_fused_rp_values() argument
792 struct xe_gt *gt = pc_to_gt(pc); in mtl_init_fused_rp_values()
795 xe_device_assert_mem_access(pc_to_xe(pc)); in mtl_init_fused_rp_values()
802 pc->rp0_freq = decode_freq(REG_FIELD_GET(MTL_RP0_CAP_MASK, reg)); in mtl_init_fused_rp_values()
804 pc->rpn_freq = decode_freq(REG_FIELD_GET(MTL_RPN_CAP_MASK, reg)); in mtl_init_fused_rp_values()
807 static void tgl_init_fused_rp_values(struct xe_guc_pc *pc) in tgl_init_fused_rp_values() argument
809 struct xe_gt *gt = pc_to_gt(pc); in tgl_init_fused_rp_values()
813 xe_device_assert_mem_access(pc_to_xe(pc)); in tgl_init_fused_rp_values()
819 pc->rp0_freq = REG_FIELD_GET(RP0_MASK, reg) * GT_FREQUENCY_MULTIPLIER; in tgl_init_fused_rp_values()
820 pc->rpn_freq = REG_FIELD_GET(RPN_MASK, reg) * GT_FREQUENCY_MULTIPLIER; in tgl_init_fused_rp_values()
823 static void pc_init_fused_rp_values(struct xe_guc_pc *pc) in pc_init_fused_rp_values() argument
825 struct xe_gt *gt = pc_to_gt(pc); in pc_init_fused_rp_values()
829 mtl_init_fused_rp_values(pc); in pc_init_fused_rp_values()
831 tgl_init_fused_rp_values(pc); in pc_init_fused_rp_values()
834 static u32 pc_max_freq_cap(struct xe_guc_pc *pc) in pc_max_freq_cap() argument
836 struct xe_gt *gt = pc_to_gt(pc); in pc_max_freq_cap()
840 return min(LNL_MERT_FREQ_CAP, pc->rp0_freq); in pc_max_freq_cap()
842 return min(BMG_MERT_FREQ_CAP, pc->rp0_freq); in pc_max_freq_cap()
844 return pc->rp0_freq; in pc_max_freq_cap()
853 void xe_guc_pc_raise_unslice(struct xe_guc_pc *pc) in xe_guc_pc_raise_unslice() argument
855 struct xe_gt *gt = pc_to_gt(pc); in xe_guc_pc_raise_unslice()
858 pc_set_cur_freq(pc, pc_max_freq_cap(pc)); in xe_guc_pc_raise_unslice()
865 void xe_guc_pc_init_early(struct xe_guc_pc *pc) in xe_guc_pc_init_early() argument
867 struct xe_gt *gt = pc_to_gt(pc); in xe_guc_pc_init_early()
870 pc_init_fused_rp_values(pc); in xe_guc_pc_init_early()
873 static int pc_adjust_freq_bounds(struct xe_guc_pc *pc) in pc_adjust_freq_bounds() argument
875 struct xe_tile *tile = gt_to_tile(pc_to_gt(pc)); in pc_adjust_freq_bounds()
878 lockdep_assert_held(&pc->freq_lock); in pc_adjust_freq_bounds()
880 ret = pc_action_query_task_state(pc); in pc_adjust_freq_bounds()
889 if (pc_get_max_freq(pc) > pc->rp0_freq) { in pc_adjust_freq_bounds()
890 ret = pc_set_max_freq(pc, pc->rp0_freq); in pc_adjust_freq_bounds()
899 if (pc_get_min_freq(pc) > pc->rp0_freq) in pc_adjust_freq_bounds()
900 ret = pc_set_min_freq(pc, pc->rp0_freq); in pc_adjust_freq_bounds()
903 ret = pc_set_min_freq(pc, max(BMG_MIN_FREQ, pc_get_min_freq(pc))); in pc_adjust_freq_bounds()
909 static int pc_adjust_requested_freq(struct xe_guc_pc *pc) in pc_adjust_requested_freq() argument
913 lockdep_assert_held(&pc->freq_lock); in pc_adjust_requested_freq()
915 if (pc->user_requested_min != 0) { in pc_adjust_requested_freq()
916 ret = pc_set_min_freq(pc, pc->user_requested_min); in pc_adjust_requested_freq()
921 if (pc->user_requested_max != 0) { in pc_adjust_requested_freq()
922 ret = pc_set_max_freq(pc, pc->user_requested_max); in pc_adjust_requested_freq()
930 static bool needs_flush_freq_limit(struct xe_guc_pc *pc) in needs_flush_freq_limit() argument
932 struct xe_gt *gt = pc_to_gt(pc); in needs_flush_freq_limit()
935 pc->rp0_freq > BMG_MERT_FLUSH_FREQ_CAP; in needs_flush_freq_limit()
944 void xe_guc_pc_apply_flush_freq_limit(struct xe_guc_pc *pc) in xe_guc_pc_apply_flush_freq_limit() argument
946 struct xe_gt *gt = pc_to_gt(pc); in xe_guc_pc_apply_flush_freq_limit()
950 if (!needs_flush_freq_limit(pc)) in xe_guc_pc_apply_flush_freq_limit()
953 guard(mutex)(&pc->freq_lock); in xe_guc_pc_apply_flush_freq_limit()
955 ret = xe_guc_pc_get_max_freq_locked(pc, &max_freq); in xe_guc_pc_apply_flush_freq_limit()
957 ret = pc_set_max_freq(pc, BMG_MERT_FLUSH_FREQ_CAP); in xe_guc_pc_apply_flush_freq_limit()
964 atomic_set(&pc->flush_freq_limit, 1); in xe_guc_pc_apply_flush_freq_limit()
971 if (pc->user_requested_max != 0) in xe_guc_pc_apply_flush_freq_limit()
972 pc->stashed_max_freq = pc->user_requested_max; in xe_guc_pc_apply_flush_freq_limit()
974 pc->stashed_max_freq = max_freq; in xe_guc_pc_apply_flush_freq_limit()
981 ret = wait_for_act_freq_limit(pc, BMG_MERT_FLUSH_FREQ_CAP); in xe_guc_pc_apply_flush_freq_limit()
993 void xe_guc_pc_remove_flush_freq_limit(struct xe_guc_pc *pc) in xe_guc_pc_remove_flush_freq_limit() argument
995 struct xe_gt *gt = pc_to_gt(pc); in xe_guc_pc_remove_flush_freq_limit()
998 if (!needs_flush_freq_limit(pc)) in xe_guc_pc_remove_flush_freq_limit()
1001 if (!atomic_read(&pc->flush_freq_limit)) in xe_guc_pc_remove_flush_freq_limit()
1004 mutex_lock(&pc->freq_lock); in xe_guc_pc_remove_flush_freq_limit()
1006 ret = pc_set_max_freq(&gt->uc.guc.pc, pc->stashed_max_freq); in xe_guc_pc_remove_flush_freq_limit()
1009 pc->stashed_max_freq, ret); in xe_guc_pc_remove_flush_freq_limit()
1011 atomic_set(&pc->flush_freq_limit, 0); in xe_guc_pc_remove_flush_freq_limit()
1012 mutex_unlock(&pc->freq_lock); in xe_guc_pc_remove_flush_freq_limit()
1013 wake_up_var(&pc->flush_freq_limit); in xe_guc_pc_remove_flush_freq_limit()
1016 static int pc_set_mert_freq_cap(struct xe_guc_pc *pc) in pc_set_mert_freq_cap() argument
1020 if (!XE_WA(pc_to_gt(pc), 22019338487)) in pc_set_mert_freq_cap()
1023 guard(mutex)(&pc->freq_lock); in pc_set_mert_freq_cap()
1028 ret = xe_guc_pc_get_min_freq_locked(pc, &pc->stashed_min_freq); in pc_set_mert_freq_cap()
1030 ret = xe_guc_pc_get_max_freq_locked(pc, &pc->stashed_max_freq); in pc_set_mert_freq_cap()
1037 ret = pc_set_min_freq(pc, min(pc->rpe_freq, pc_max_freq_cap(pc))); in pc_set_mert_freq_cap()
1039 ret = pc_set_max_freq(pc, min(pc->rp0_freq, pc_max_freq_cap(pc))); in pc_set_mert_freq_cap()
1051 int xe_guc_pc_restore_stashed_freq(struct xe_guc_pc *pc) in xe_guc_pc_restore_stashed_freq() argument
1055 if (IS_SRIOV_VF(pc_to_xe(pc)) || pc_to_xe(pc)->info.skip_guc_pc) in xe_guc_pc_restore_stashed_freq()
1058 mutex_lock(&pc->freq_lock); in xe_guc_pc_restore_stashed_freq()
1059 ret = pc_set_max_freq(pc, pc->stashed_max_freq); in xe_guc_pc_restore_stashed_freq()
1061 ret = pc_set_min_freq(pc, pc->stashed_min_freq); in xe_guc_pc_restore_stashed_freq()
1062 mutex_unlock(&pc->freq_lock); in xe_guc_pc_restore_stashed_freq()
1075 int xe_guc_pc_gucrc_disable(struct xe_guc_pc *pc) in xe_guc_pc_gucrc_disable() argument
1077 struct xe_device *xe = pc_to_xe(pc); in xe_guc_pc_gucrc_disable()
1078 struct xe_gt *gt = pc_to_gt(pc); in xe_guc_pc_gucrc_disable()
1085 ret = pc_action_setup_gucrc(pc, GUCRC_HOST_CONTROL); in xe_guc_pc_gucrc_disable()
1109 int xe_guc_pc_override_gucrc_mode(struct xe_guc_pc *pc, enum slpc_gucrc_mode mode) in xe_guc_pc_override_gucrc_mode() argument
1113 xe_pm_runtime_get(pc_to_xe(pc)); in xe_guc_pc_override_gucrc_mode()
1114 ret = pc_action_set_param(pc, SLPC_PARAM_PWRGATE_RC_MODE, mode); in xe_guc_pc_override_gucrc_mode()
1115 xe_pm_runtime_put(pc_to_xe(pc)); in xe_guc_pc_override_gucrc_mode()
1126 int xe_guc_pc_unset_gucrc_mode(struct xe_guc_pc *pc) in xe_guc_pc_unset_gucrc_mode() argument
1130 xe_pm_runtime_get(pc_to_xe(pc)); in xe_guc_pc_unset_gucrc_mode()
1131 ret = pc_action_unset_param(pc, SLPC_PARAM_PWRGATE_RC_MODE); in xe_guc_pc_unset_gucrc_mode()
1132 xe_pm_runtime_put(pc_to_xe(pc)); in xe_guc_pc_unset_gucrc_mode()
1137 static void pc_init_pcode_freq(struct xe_guc_pc *pc) in pc_init_pcode_freq() argument
1139 u32 min = DIV_ROUND_CLOSEST(pc->rpn_freq, GT_FREQUENCY_MULTIPLIER); in pc_init_pcode_freq()
1140 u32 max = DIV_ROUND_CLOSEST(pc->rp0_freq, GT_FREQUENCY_MULTIPLIER); in pc_init_pcode_freq()
1142 XE_WARN_ON(xe_pcode_init_min_freq_table(gt_to_tile(pc_to_gt(pc)), min, max)); in pc_init_pcode_freq()
1145 static int pc_init_freqs(struct xe_guc_pc *pc) in pc_init_freqs() argument
1149 mutex_lock(&pc->freq_lock); in pc_init_freqs()
1151 ret = pc_adjust_freq_bounds(pc); in pc_init_freqs()
1155 ret = pc_adjust_requested_freq(pc); in pc_init_freqs()
1159 pc_update_rp_values(pc); in pc_init_freqs()
1161 pc_init_pcode_freq(pc); in pc_init_freqs()
1167 pc->freq_ready = true; in pc_init_freqs()
1170 mutex_unlock(&pc->freq_lock); in pc_init_freqs()
1174 static int pc_action_set_strategy(struct xe_guc_pc *pc, u32 val) in pc_action_set_strategy() argument
1178 ret = pc_action_set_param(pc, in pc_action_set_strategy()
1189 int xe_guc_pc_start(struct xe_guc_pc *pc) in xe_guc_pc_start() argument
1191 struct xe_device *xe = pc_to_xe(pc); in xe_guc_pc_start()
1192 struct xe_gt *gt = pc_to_gt(pc); in xe_guc_pc_start()
1211 pc_set_cur_freq(pc, UINT_MAX); in xe_guc_pc_start()
1217 xe_map_memset(xe, &pc->bo->vmap, 0, 0, size); in xe_guc_pc_start()
1218 slpc_shared_data_write(pc, header.size, size); in xe_guc_pc_start()
1221 ret = pc_action_reset(pc); in xe_guc_pc_start()
1225 if (wait_for_pc_state(pc, SLPC_GLOBAL_STATE_RUNNING, in xe_guc_pc_start()
1228 xe_guc_pc_get_act_freq(pc), get_cur_freq(gt), in xe_guc_pc_start()
1231 if (wait_for_pc_state(pc, SLPC_GLOBAL_STATE_RUNNING, in xe_guc_pc_start()
1242 ret = pc_init_freqs(pc); in xe_guc_pc_start()
1246 ret = pc_set_mert_freq_cap(pc); in xe_guc_pc_start()
1251 xe_guc_pc_gucrc_disable(pc); in xe_guc_pc_start()
1256 ret = pc_action_setup_gucrc(pc, GUCRC_FIRMWARE_CONTROL); in xe_guc_pc_start()
1261 ret = pc_action_set_strategy(pc, SLPC_OPTIMIZED_STRATEGY_COMPUTE); in xe_guc_pc_start()
1272 int xe_guc_pc_stop(struct xe_guc_pc *pc) in xe_guc_pc_stop() argument
1274 struct xe_device *xe = pc_to_xe(pc); in xe_guc_pc_stop()
1277 xe_gt_idle_disable_c6(pc_to_gt(pc)); in xe_guc_pc_stop()
1281 mutex_lock(&pc->freq_lock); in xe_guc_pc_stop()
1282 pc->freq_ready = false; in xe_guc_pc_stop()
1283 mutex_unlock(&pc->freq_lock); in xe_guc_pc_stop()
1294 struct xe_guc_pc *pc = arg; in xe_guc_pc_fini_hw() local
1295 struct xe_device *xe = pc_to_xe(pc); in xe_guc_pc_fini_hw()
1301 fw_ref = xe_force_wake_get(gt_to_fw(pc_to_gt(pc)), XE_FORCEWAKE_ALL); in xe_guc_pc_fini_hw()
1302 xe_guc_pc_gucrc_disable(pc); in xe_guc_pc_fini_hw()
1303 XE_WARN_ON(xe_guc_pc_stop(pc)); in xe_guc_pc_fini_hw()
1306 pc_set_cur_freq(pc, min(pc_max_freq_cap(pc), pc->rpe_freq)); in xe_guc_pc_fini_hw()
1308 xe_force_wake_put(gt_to_fw(pc_to_gt(pc)), fw_ref); in xe_guc_pc_fini_hw()
1315 int xe_guc_pc_init(struct xe_guc_pc *pc) in xe_guc_pc_init() argument
1317 struct xe_gt *gt = pc_to_gt(pc); in xe_guc_pc_init()
1327 err = drmm_mutex_init(&xe->drm, &pc->freq_lock); in xe_guc_pc_init()
1339 pc->bo = bo; in xe_guc_pc_init()
1341 return devm_add_action_or_reset(xe->drm.dev, xe_guc_pc_fini_hw, pc); in xe_guc_pc_init()
1344 static const char *pc_get_state_string(struct xe_guc_pc *pc) in pc_get_state_string() argument
1346 switch (slpc_shared_data_read(pc, header.global_state)) { in pc_get_state_string()
1369 void xe_guc_pc_print(struct xe_guc_pc *pc, struct drm_printer *p) in xe_guc_pc_print() argument
1372 drm_printf(p, "\tSize: %x\n", slpc_shared_data_read(pc, header.size)); in xe_guc_pc_print()
1373 drm_printf(p, "\tGlobal State: %s\n", pc_get_state_string(pc)); in xe_guc_pc_print()
1375 if (pc_action_query_task_state(pc)) in xe_guc_pc_print()
1380 str_yes_no(slpc_shared_data_read(pc, task_state_data.status) & in xe_guc_pc_print()
1383 str_yes_no(slpc_shared_data_read(pc, task_state_data.status) & in xe_guc_pc_print()
1386 str_yes_no(slpc_shared_data_read(pc, task_state_data.status) & in xe_guc_pc_print()
1389 str_yes_no(slpc_shared_data_read(pc, task_state_data.status) & in xe_guc_pc_print()
1392 str_yes_no(slpc_shared_data_read(pc, task_state_data.status) & in xe_guc_pc_print()
1395 str_yes_no(slpc_shared_data_read(pc, task_state_data.status) & in xe_guc_pc_print()
1398 str_yes_no(slpc_shared_data_read(pc, task_state_data.status) & in xe_guc_pc_print()