Lines Matching full:pc
59 * DOC: GuC Power Conservation (PC)
61 * GuC Power Conservation (PC) supports multiple features for the most
79 * Render-C states is also a GuC PC feature that is now enabled in Xe for
84 static struct xe_guc *pc_to_guc(struct xe_guc_pc *pc) in pc_to_guc() argument
86 return container_of(pc, struct xe_guc, pc); in pc_to_guc()
89 static struct xe_guc_ct *pc_to_ct(struct xe_guc_pc *pc) in pc_to_ct() argument
91 return &pc_to_guc(pc)->ct; in pc_to_ct()
94 static struct xe_gt *pc_to_gt(struct xe_guc_pc *pc) in pc_to_gt() argument
96 return guc_to_gt(pc_to_guc(pc)); in pc_to_gt()
99 static struct xe_device *pc_to_xe(struct xe_guc_pc *pc) in pc_to_xe() argument
101 return guc_to_xe(pc_to_guc(pc)); in pc_to_xe()
104 static struct iosys_map *pc_to_maps(struct xe_guc_pc *pc) in pc_to_maps() argument
106 return &pc->bo->vmap; in pc_to_maps()
121 static int wait_for_pc_state(struct xe_guc_pc *pc, in wait_for_pc_state() argument
128 xe_device_assert_mem_access(pc_to_xe(pc)); in wait_for_pc_state()
131 if (slpc_shared_data_read(pc, header.global_state) == state) in wait_for_pc_state()
144 static int pc_action_reset(struct xe_guc_pc *pc) in pc_action_reset() argument
146 struct xe_guc_ct *ct = pc_to_ct(pc); in pc_action_reset()
150 xe_bo_ggtt_addr(pc->bo), in pc_action_reset()
157 xe_gt_err(pc_to_gt(pc), "GuC PC reset failed: %pe\n", in pc_action_reset()
163 static int pc_action_query_task_state(struct xe_guc_pc *pc) in pc_action_query_task_state() argument
165 struct xe_guc_ct *ct = pc_to_ct(pc); in pc_action_query_task_state()
169 xe_bo_ggtt_addr(pc->bo), in pc_action_query_task_state()
174 if (wait_for_pc_state(pc, SLPC_GLOBAL_STATE_RUNNING, in pc_action_query_task_state()
181 xe_gt_err(pc_to_gt(pc), "GuC PC query task state failed: %pe\n", in pc_action_query_task_state()
187 static int pc_action_set_param(struct xe_guc_pc *pc, u8 id, u32 value) in pc_action_set_param() argument
189 struct xe_guc_ct *ct = pc_to_ct(pc); in pc_action_set_param()
198 if (wait_for_pc_state(pc, SLPC_GLOBAL_STATE_RUNNING, in pc_action_set_param()
204 xe_gt_err(pc_to_gt(pc), "GuC PC set param[%u]=%u failed: %pe\n", in pc_action_set_param()
210 static int pc_action_unset_param(struct xe_guc_pc *pc, u8 id) in pc_action_unset_param() argument
217 struct xe_guc_ct *ct = &pc_to_guc(pc)->ct; in pc_action_unset_param()
220 if (wait_for_pc_state(pc, SLPC_GLOBAL_STATE_RUNNING, in pc_action_unset_param()
226 xe_gt_err(pc_to_gt(pc), "GuC PC unset param failed: %pe", in pc_action_unset_param()
232 static int pc_action_setup_gucrc(struct xe_guc_pc *pc, u32 mode) in pc_action_setup_gucrc() argument
234 struct xe_guc_ct *ct = pc_to_ct(pc); in pc_action_setup_gucrc()
243 xe_gt_err(pc_to_gt(pc), "GuC RC enable mode=%u failed: %pe\n", in pc_action_setup_gucrc()
260 static u32 pc_get_min_freq(struct xe_guc_pc *pc) in pc_get_min_freq() argument
265 slpc_shared_data_read(pc, task_state_data.freq)); in pc_get_min_freq()
270 static void pc_set_manual_rp_ctrl(struct xe_guc_pc *pc, bool enable) in pc_set_manual_rp_ctrl() argument
272 struct xe_gt *gt = pc_to_gt(pc); in pc_set_manual_rp_ctrl()
279 static void pc_set_cur_freq(struct xe_guc_pc *pc, u32 freq) in pc_set_cur_freq() argument
281 struct xe_gt *gt = pc_to_gt(pc); in pc_set_cur_freq()
284 pc_set_manual_rp_ctrl(pc, true); in pc_set_cur_freq()
293 pc_set_manual_rp_ctrl(pc, false); in pc_set_cur_freq()
296 static int pc_set_min_freq(struct xe_guc_pc *pc, u32 freq) in pc_set_min_freq() argument
302 if (freq < pc->rpn_freq || freq > pc->rp0_freq) in pc_set_min_freq()
309 pc_action_set_param(pc, SLPC_PARAM_IGNORE_EFFICIENT_FREQUENCY, in pc_set_min_freq()
310 freq < pc->rpe_freq); in pc_set_min_freq()
312 return pc_action_set_param(pc, in pc_set_min_freq()
317 static int pc_get_max_freq(struct xe_guc_pc *pc) in pc_get_max_freq() argument
322 slpc_shared_data_read(pc, task_state_data.freq)); in pc_get_max_freq()
327 static int pc_set_max_freq(struct xe_guc_pc *pc, u32 freq) in pc_set_max_freq() argument
334 if (freq < pc->rpn_freq || freq > pc->rp0_freq) in pc_set_max_freq()
337 return pc_action_set_param(pc, in pc_set_max_freq()
342 static void mtl_update_rpa_value(struct xe_guc_pc *pc) in mtl_update_rpa_value() argument
344 struct xe_gt *gt = pc_to_gt(pc); in mtl_update_rpa_value()
352 pc->rpa_freq = decode_freq(REG_FIELD_GET(MTL_RPA_MASK, reg)); in mtl_update_rpa_value()
355 static void mtl_update_rpe_value(struct xe_guc_pc *pc) in mtl_update_rpe_value() argument
357 struct xe_gt *gt = pc_to_gt(pc); in mtl_update_rpe_value()
365 pc->rpe_freq = decode_freq(REG_FIELD_GET(MTL_RPE_MASK, reg)); in mtl_update_rpe_value()
368 static void tgl_update_rpa_value(struct xe_guc_pc *pc) in tgl_update_rpa_value() argument
370 struct xe_gt *gt = pc_to_gt(pc); in tgl_update_rpa_value()
381 pc->rpa_freq = REG_FIELD_GET(RP0_MASK, reg) * GT_FREQUENCY_MULTIPLIER; in tgl_update_rpa_value()
384 pc->rpa_freq = REG_FIELD_GET(RPA_MASK, reg) * GT_FREQUENCY_MULTIPLIER; in tgl_update_rpa_value()
388 static void tgl_update_rpe_value(struct xe_guc_pc *pc) in tgl_update_rpe_value() argument
390 struct xe_gt *gt = pc_to_gt(pc); in tgl_update_rpe_value()
401 pc->rpe_freq = REG_FIELD_GET(RP1_MASK, reg) * GT_FREQUENCY_MULTIPLIER; in tgl_update_rpe_value()
404 pc->rpe_freq = REG_FIELD_GET(RPE_MASK, reg) * GT_FREQUENCY_MULTIPLIER; in tgl_update_rpe_value()
408 static void pc_update_rp_values(struct xe_guc_pc *pc) in pc_update_rp_values() argument
410 struct xe_gt *gt = pc_to_gt(pc); in pc_update_rp_values()
414 mtl_update_rpa_value(pc); in pc_update_rp_values()
415 mtl_update_rpe_value(pc); in pc_update_rp_values()
417 tgl_update_rpa_value(pc); in pc_update_rp_values()
418 tgl_update_rpe_value(pc); in pc_update_rp_values()
426 pc->rpn_freq = min(pc->rpn_freq, pc->rpe_freq); in pc_update_rp_values()
431 * @pc: The GuC PC
435 u32 xe_guc_pc_get_act_freq(struct xe_guc_pc *pc) in xe_guc_pc_get_act_freq() argument
437 struct xe_gt *gt = pc_to_gt(pc); in xe_guc_pc_get_act_freq()
466 * @pc: The GuC PC
470 * -EAGAIN if GuC PC not ready (likely in middle of a reset).
472 int xe_guc_pc_get_cur_freq(struct xe_guc_pc *pc, u32 *freq) in xe_guc_pc_get_cur_freq() argument
474 struct xe_gt *gt = pc_to_gt(pc); in xe_guc_pc_get_cur_freq()
495 * @pc: The GuC PC
499 u32 xe_guc_pc_get_rp0_freq(struct xe_guc_pc *pc) in xe_guc_pc_get_rp0_freq() argument
501 return pc->rp0_freq; in xe_guc_pc_get_rp0_freq()
506 * @pc: The GuC PC
510 u32 xe_guc_pc_get_rpa_freq(struct xe_guc_pc *pc) in xe_guc_pc_get_rpa_freq() argument
512 pc_update_rp_values(pc); in xe_guc_pc_get_rpa_freq()
514 return pc->rpa_freq; in xe_guc_pc_get_rpa_freq()
519 * @pc: The GuC PC
523 u32 xe_guc_pc_get_rpe_freq(struct xe_guc_pc *pc) in xe_guc_pc_get_rpe_freq() argument
525 pc_update_rp_values(pc); in xe_guc_pc_get_rpe_freq()
527 return pc->rpe_freq; in xe_guc_pc_get_rpe_freq()
532 * @pc: The GuC PC
536 u32 xe_guc_pc_get_rpn_freq(struct xe_guc_pc *pc) in xe_guc_pc_get_rpn_freq() argument
538 return pc->rpn_freq; in xe_guc_pc_get_rpn_freq()
543 * @pc: The GuC PC
547 * -EAGAIN if GuC PC not ready (likely in middle of a reset).
549 int xe_guc_pc_get_min_freq(struct xe_guc_pc *pc, u32 *freq) in xe_guc_pc_get_min_freq() argument
553 xe_device_assert_mem_access(pc_to_xe(pc)); in xe_guc_pc_get_min_freq()
555 mutex_lock(&pc->freq_lock); in xe_guc_pc_get_min_freq()
556 if (!pc->freq_ready) { in xe_guc_pc_get_min_freq()
562 ret = pc_action_query_task_state(pc); in xe_guc_pc_get_min_freq()
566 *freq = pc_get_min_freq(pc); in xe_guc_pc_get_min_freq()
569 mutex_unlock(&pc->freq_lock); in xe_guc_pc_get_min_freq()
575 * @pc: The GuC PC
579 * -EAGAIN if GuC PC not ready (likely in middle of a reset),
582 int xe_guc_pc_set_min_freq(struct xe_guc_pc *pc, u32 freq) in xe_guc_pc_set_min_freq() argument
586 mutex_lock(&pc->freq_lock); in xe_guc_pc_set_min_freq()
587 if (!pc->freq_ready) { in xe_guc_pc_set_min_freq()
593 ret = pc_set_min_freq(pc, freq); in xe_guc_pc_set_min_freq()
597 pc->user_requested_min = freq; in xe_guc_pc_set_min_freq()
600 mutex_unlock(&pc->freq_lock); in xe_guc_pc_set_min_freq()
606 * @pc: The GuC PC
610 * -EAGAIN if GuC PC not ready (likely in middle of a reset).
612 int xe_guc_pc_get_max_freq(struct xe_guc_pc *pc, u32 *freq) in xe_guc_pc_get_max_freq() argument
616 mutex_lock(&pc->freq_lock); in xe_guc_pc_get_max_freq()
617 if (!pc->freq_ready) { in xe_guc_pc_get_max_freq()
623 ret = pc_action_query_task_state(pc); in xe_guc_pc_get_max_freq()
627 *freq = pc_get_max_freq(pc); in xe_guc_pc_get_max_freq()
630 mutex_unlock(&pc->freq_lock); in xe_guc_pc_get_max_freq()
636 * @pc: The GuC PC
640 * -EAGAIN if GuC PC not ready (likely in middle of a reset),
643 int xe_guc_pc_set_max_freq(struct xe_guc_pc *pc, u32 freq) in xe_guc_pc_set_max_freq() argument
647 mutex_lock(&pc->freq_lock); in xe_guc_pc_set_max_freq()
648 if (!pc->freq_ready) { in xe_guc_pc_set_max_freq()
654 ret = pc_set_max_freq(pc, freq); in xe_guc_pc_set_max_freq()
658 pc->user_requested_max = freq; in xe_guc_pc_set_max_freq()
661 mutex_unlock(&pc->freq_lock); in xe_guc_pc_set_max_freq()
667 * @pc: XE_GuC_PC instance
669 enum xe_gt_idle_state xe_guc_pc_c_status(struct xe_guc_pc *pc) in xe_guc_pc_c_status() argument
671 struct xe_gt *gt = pc_to_gt(pc); in xe_guc_pc_c_status()
694 * @pc: Xe_GuC_PC instance
696 u64 xe_guc_pc_rc6_residency(struct xe_guc_pc *pc) in xe_guc_pc_rc6_residency() argument
698 struct xe_gt *gt = pc_to_gt(pc); in xe_guc_pc_rc6_residency()
708 * @pc: Xe_GuC_PC instance
710 u64 xe_guc_pc_mc6_residency(struct xe_guc_pc *pc) in xe_guc_pc_mc6_residency() argument
712 struct xe_gt *gt = pc_to_gt(pc); in xe_guc_pc_mc6_residency()
720 static void mtl_init_fused_rp_values(struct xe_guc_pc *pc) in mtl_init_fused_rp_values() argument
722 struct xe_gt *gt = pc_to_gt(pc); in mtl_init_fused_rp_values()
725 xe_device_assert_mem_access(pc_to_xe(pc)); in mtl_init_fused_rp_values()
732 pc->rp0_freq = decode_freq(REG_FIELD_GET(MTL_RP0_CAP_MASK, reg)); in mtl_init_fused_rp_values()
734 pc->rpn_freq = decode_freq(REG_FIELD_GET(MTL_RPN_CAP_MASK, reg)); in mtl_init_fused_rp_values()
737 static void tgl_init_fused_rp_values(struct xe_guc_pc *pc) in tgl_init_fused_rp_values() argument
739 struct xe_gt *gt = pc_to_gt(pc); in tgl_init_fused_rp_values()
743 xe_device_assert_mem_access(pc_to_xe(pc)); in tgl_init_fused_rp_values()
749 pc->rp0_freq = REG_FIELD_GET(RP0_MASK, reg) * GT_FREQUENCY_MULTIPLIER; in tgl_init_fused_rp_values()
750 pc->rpn_freq = REG_FIELD_GET(RPN_MASK, reg) * GT_FREQUENCY_MULTIPLIER; in tgl_init_fused_rp_values()
753 static void pc_init_fused_rp_values(struct xe_guc_pc *pc) in pc_init_fused_rp_values() argument
755 struct xe_gt *gt = pc_to_gt(pc); in pc_init_fused_rp_values()
759 mtl_init_fused_rp_values(pc); in pc_init_fused_rp_values()
761 tgl_init_fused_rp_values(pc); in pc_init_fused_rp_values()
764 static u32 pc_max_freq_cap(struct xe_guc_pc *pc) in pc_max_freq_cap() argument
766 struct xe_gt *gt = pc_to_gt(pc); in pc_max_freq_cap()
770 return min(LNL_MERT_FREQ_CAP, pc->rp0_freq); in pc_max_freq_cap()
772 return min(BMG_MERT_FREQ_CAP, pc->rp0_freq); in pc_max_freq_cap()
774 return pc->rp0_freq; in pc_max_freq_cap()
781 * @pc: Xe_GuC_PC instance
783 void xe_guc_pc_raise_unslice(struct xe_guc_pc *pc) in xe_guc_pc_raise_unslice() argument
785 struct xe_gt *gt = pc_to_gt(pc); in xe_guc_pc_raise_unslice()
788 pc_set_cur_freq(pc, pc_max_freq_cap(pc)); in xe_guc_pc_raise_unslice()
793 * @pc: Xe_GuC_PC instance
795 void xe_guc_pc_init_early(struct xe_guc_pc *pc) in xe_guc_pc_init_early() argument
797 struct xe_gt *gt = pc_to_gt(pc); in xe_guc_pc_init_early()
800 pc_init_fused_rp_values(pc); in xe_guc_pc_init_early()
803 static int pc_adjust_freq_bounds(struct xe_guc_pc *pc) in pc_adjust_freq_bounds() argument
807 lockdep_assert_held(&pc->freq_lock); in pc_adjust_freq_bounds()
809 ret = pc_action_query_task_state(pc); in pc_adjust_freq_bounds()
818 if (pc_get_max_freq(pc) > pc->rp0_freq) { in pc_adjust_freq_bounds()
819 ret = pc_set_max_freq(pc, pc->rp0_freq); in pc_adjust_freq_bounds()
828 if (pc_get_min_freq(pc) > pc->rp0_freq) in pc_adjust_freq_bounds()
829 ret = pc_set_min_freq(pc, pc->rp0_freq); in pc_adjust_freq_bounds()
835 static int pc_adjust_requested_freq(struct xe_guc_pc *pc) in pc_adjust_requested_freq() argument
839 lockdep_assert_held(&pc->freq_lock); in pc_adjust_requested_freq()
841 if (pc->user_requested_min != 0) { in pc_adjust_requested_freq()
842 ret = pc_set_min_freq(pc, pc->user_requested_min); in pc_adjust_requested_freq()
847 if (pc->user_requested_max != 0) { in pc_adjust_requested_freq()
848 ret = pc_set_max_freq(pc, pc->user_requested_max); in pc_adjust_requested_freq()
856 static int pc_set_mert_freq_cap(struct xe_guc_pc *pc) in pc_set_mert_freq_cap() argument
860 if (XE_WA(pc_to_gt(pc), 22019338487)) { in pc_set_mert_freq_cap()
864 ret = xe_guc_pc_get_min_freq(pc, &pc->stashed_min_freq); in pc_set_mert_freq_cap()
866 ret = xe_guc_pc_get_max_freq(pc, &pc->stashed_max_freq); in pc_set_mert_freq_cap()
873 mutex_lock(&pc->freq_lock); in pc_set_mert_freq_cap()
874 ret = pc_set_min_freq(pc, min(pc->rpe_freq, pc_max_freq_cap(pc))); in pc_set_mert_freq_cap()
876 ret = pc_set_max_freq(pc, min(pc->rp0_freq, pc_max_freq_cap(pc))); in pc_set_mert_freq_cap()
877 mutex_unlock(&pc->freq_lock); in pc_set_mert_freq_cap()
885 * @pc: The GuC PC
890 int xe_guc_pc_restore_stashed_freq(struct xe_guc_pc *pc) in xe_guc_pc_restore_stashed_freq() argument
894 if (IS_SRIOV_VF(pc_to_xe(pc)) || pc_to_xe(pc)->info.skip_guc_pc) in xe_guc_pc_restore_stashed_freq()
897 mutex_lock(&pc->freq_lock); in xe_guc_pc_restore_stashed_freq()
898 ret = pc_set_max_freq(pc, pc->stashed_max_freq); in xe_guc_pc_restore_stashed_freq()
900 ret = pc_set_min_freq(pc, pc->stashed_min_freq); in xe_guc_pc_restore_stashed_freq()
901 mutex_unlock(&pc->freq_lock); in xe_guc_pc_restore_stashed_freq()
908 * @pc: Xe_GuC_PC instance
914 int xe_guc_pc_gucrc_disable(struct xe_guc_pc *pc) in xe_guc_pc_gucrc_disable() argument
916 struct xe_device *xe = pc_to_xe(pc); in xe_guc_pc_gucrc_disable()
917 struct xe_gt *gt = pc_to_gt(pc); in xe_guc_pc_gucrc_disable()
924 ret = pc_action_setup_gucrc(pc, GUCRC_HOST_CONTROL); in xe_guc_pc_gucrc_disable()
943 * @pc: Xe_GuC_PC instance
948 int xe_guc_pc_override_gucrc_mode(struct xe_guc_pc *pc, enum slpc_gucrc_mode mode) in xe_guc_pc_override_gucrc_mode() argument
952 xe_pm_runtime_get(pc_to_xe(pc)); in xe_guc_pc_override_gucrc_mode()
953 ret = pc_action_set_param(pc, SLPC_PARAM_PWRGATE_RC_MODE, mode); in xe_guc_pc_override_gucrc_mode()
954 xe_pm_runtime_put(pc_to_xe(pc)); in xe_guc_pc_override_gucrc_mode()
961 * @pc: Xe_GuC_PC instance
965 int xe_guc_pc_unset_gucrc_mode(struct xe_guc_pc *pc) in xe_guc_pc_unset_gucrc_mode() argument
969 xe_pm_runtime_get(pc_to_xe(pc)); in xe_guc_pc_unset_gucrc_mode()
970 ret = pc_action_unset_param(pc, SLPC_PARAM_PWRGATE_RC_MODE); in xe_guc_pc_unset_gucrc_mode()
971 xe_pm_runtime_put(pc_to_xe(pc)); in xe_guc_pc_unset_gucrc_mode()
976 static void pc_init_pcode_freq(struct xe_guc_pc *pc) in pc_init_pcode_freq() argument
978 u32 min = DIV_ROUND_CLOSEST(pc->rpn_freq, GT_FREQUENCY_MULTIPLIER); in pc_init_pcode_freq()
979 u32 max = DIV_ROUND_CLOSEST(pc->rp0_freq, GT_FREQUENCY_MULTIPLIER); in pc_init_pcode_freq()
981 XE_WARN_ON(xe_pcode_init_min_freq_table(gt_to_tile(pc_to_gt(pc)), min, max)); in pc_init_pcode_freq()
984 static int pc_init_freqs(struct xe_guc_pc *pc) in pc_init_freqs() argument
988 mutex_lock(&pc->freq_lock); in pc_init_freqs()
990 ret = pc_adjust_freq_bounds(pc); in pc_init_freqs()
994 ret = pc_adjust_requested_freq(pc); in pc_init_freqs()
998 pc_update_rp_values(pc); in pc_init_freqs()
1000 pc_init_pcode_freq(pc); in pc_init_freqs()
1006 pc->freq_ready = true; in pc_init_freqs()
1009 mutex_unlock(&pc->freq_lock); in pc_init_freqs()
1013 static int pc_action_set_strategy(struct xe_guc_pc *pc, u32 val) in pc_action_set_strategy() argument
1017 ret = pc_action_set_param(pc, in pc_action_set_strategy()
1026 * @pc: Xe_GuC_PC instance
1028 int xe_guc_pc_start(struct xe_guc_pc *pc) in xe_guc_pc_start() argument
1030 struct xe_device *xe = pc_to_xe(pc); in xe_guc_pc_start()
1031 struct xe_gt *gt = pc_to_gt(pc); in xe_guc_pc_start()
1050 pc_set_cur_freq(pc, UINT_MAX); in xe_guc_pc_start()
1056 memset(pc->bo->vmap.vaddr, 0, size); in xe_guc_pc_start()
1057 slpc_shared_data_write(pc, header.size, size); in xe_guc_pc_start()
1060 ret = pc_action_reset(pc); in xe_guc_pc_start()
1064 if (wait_for_pc_state(pc, SLPC_GLOBAL_STATE_RUNNING, in xe_guc_pc_start()
1066 …xe_gt_warn(gt, "GuC PC start taking longer than normal [freq = %dMHz (req = %dMHz), perf_limit_rea… in xe_guc_pc_start()
1067 xe_guc_pc_get_act_freq(pc), get_cur_freq(gt), in xe_guc_pc_start()
1070 if (wait_for_pc_state(pc, SLPC_GLOBAL_STATE_RUNNING, in xe_guc_pc_start()
1072 …xe_gt_err(gt, "GuC PC Start failed: Dynamic GT frequency control and GT sleep states are now disab… in xe_guc_pc_start()
1077 xe_gt_warn(gt, "GuC PC excessive start time: %lldms", in xe_guc_pc_start()
1081 ret = pc_init_freqs(pc); in xe_guc_pc_start()
1085 ret = pc_set_mert_freq_cap(pc); in xe_guc_pc_start()
1090 xe_guc_pc_gucrc_disable(pc); in xe_guc_pc_start()
1095 ret = pc_action_setup_gucrc(pc, GUCRC_FIRMWARE_CONTROL); in xe_guc_pc_start()
1100 ret = pc_action_set_strategy(pc, SLPC_OPTIMIZED_STRATEGY_COMPUTE); in xe_guc_pc_start()
1109 * @pc: Xe_GuC_PC instance
1111 int xe_guc_pc_stop(struct xe_guc_pc *pc) in xe_guc_pc_stop() argument
1113 struct xe_device *xe = pc_to_xe(pc); in xe_guc_pc_stop()
1116 xe_gt_idle_disable_c6(pc_to_gt(pc)); in xe_guc_pc_stop()
1120 mutex_lock(&pc->freq_lock); in xe_guc_pc_stop()
1121 pc->freq_ready = false; in xe_guc_pc_stop()
1122 mutex_unlock(&pc->freq_lock); in xe_guc_pc_stop()
1133 struct xe_guc_pc *pc = arg; in xe_guc_pc_fini_hw() local
1134 struct xe_device *xe = pc_to_xe(pc); in xe_guc_pc_fini_hw()
1140 fw_ref = xe_force_wake_get(gt_to_fw(pc_to_gt(pc)), XE_FORCEWAKE_ALL); in xe_guc_pc_fini_hw()
1141 xe_guc_pc_gucrc_disable(pc); in xe_guc_pc_fini_hw()
1142 XE_WARN_ON(xe_guc_pc_stop(pc)); in xe_guc_pc_fini_hw()
1145 pc_set_cur_freq(pc, min(pc_max_freq_cap(pc), pc->rpe_freq)); in xe_guc_pc_fini_hw()
1147 xe_force_wake_put(gt_to_fw(pc_to_gt(pc)), fw_ref); in xe_guc_pc_fini_hw()
1152 * @pc: Xe_GuC_PC instance
1154 int xe_guc_pc_init(struct xe_guc_pc *pc) in xe_guc_pc_init() argument
1156 struct xe_gt *gt = pc_to_gt(pc); in xe_guc_pc_init()
1166 err = drmm_mutex_init(&xe->drm, &pc->freq_lock); in xe_guc_pc_init()
1177 pc->bo = bo; in xe_guc_pc_init()
1179 return devm_add_action_or_reset(xe->drm.dev, xe_guc_pc_fini_hw, pc); in xe_guc_pc_init()
1182 static const char *pc_get_state_string(struct xe_guc_pc *pc) in pc_get_state_string() argument
1184 switch (slpc_shared_data_read(pc, header.global_state)) { in pc_get_state_string()
1204 * @pc: Xe_GuC_PC instance
1207 void xe_guc_pc_print(struct xe_guc_pc *pc, struct drm_printer *p) in xe_guc_pc_print() argument
1210 drm_printf(p, "\tSize: %x\n", slpc_shared_data_read(pc, header.size)); in xe_guc_pc_print()
1211 drm_printf(p, "\tGlobal State: %s\n", pc_get_state_string(pc)); in xe_guc_pc_print()
1213 if (pc_action_query_task_state(pc)) in xe_guc_pc_print()
1218 str_yes_no(slpc_shared_data_read(pc, task_state_data.status) & in xe_guc_pc_print()
1221 str_yes_no(slpc_shared_data_read(pc, task_state_data.status) & in xe_guc_pc_print()
1224 str_yes_no(slpc_shared_data_read(pc, task_state_data.status) & in xe_guc_pc_print()
1227 str_yes_no(slpc_shared_data_read(pc, task_state_data.status) & in xe_guc_pc_print()
1230 str_yes_no(slpc_shared_data_read(pc, task_state_data.status) & in xe_guc_pc_print()
1233 str_yes_no(slpc_shared_data_read(pc, task_state_data.status) & in xe_guc_pc_print()
1236 str_yes_no(slpc_shared_data_read(pc, task_state_data.status) & in xe_guc_pc_print()