Lines Matching +full:cs +full:- +full:x

1 // SPDX-License-Identifier: MIT
26 #define CPU_LATENCY 0 /* -1 to disable pm_qos, 0 to disable cstates */
37 return -1; in cmp_u64()
49 return -1; in cmp_u32()
68 #define CS_GPR(x) GEN8_RING_CS_GPR(engine->mmio_base, x) in create_spin_counter() argument
72 u32 *base, *cs; in create_spin_counter() local
76 obj = i915_gem_object_create_internal(vm->i915, 64 << 10); in create_spin_counter()
80 end = obj->base.size / sizeof(u32) - 1; in create_spin_counter()
99 cs = base; in create_spin_counter()
101 *cs++ = MI_LOAD_REGISTER_IMM(__NGPR__ * 2); in create_spin_counter()
103 *cs++ = i915_mmio_reg_offset(CS_GPR(i)); in create_spin_counter()
104 *cs++ = 0; in create_spin_counter()
105 *cs++ = i915_mmio_reg_offset(CS_GPR(i)) + 4; in create_spin_counter()
106 *cs++ = 0; in create_spin_counter()
109 *cs++ = MI_LOAD_REGISTER_IMM(1); in create_spin_counter()
110 *cs++ = i915_mmio_reg_offset(CS_GPR(INC)); in create_spin_counter()
111 *cs++ = 1; in create_spin_counter()
113 loop = cs - base; in create_spin_counter()
117 *cs++ = MI_MATH(4); in create_spin_counter()
118 *cs++ = MI_MATH_LOAD(MI_MATH_REG_SRCA, MI_MATH_REG(COUNT)); in create_spin_counter()
119 *cs++ = MI_MATH_LOAD(MI_MATH_REG_SRCB, MI_MATH_REG(INC)); in create_spin_counter()
120 *cs++ = MI_MATH_ADD; in create_spin_counter()
121 *cs++ = MI_MATH_STORE(MI_MATH_REG(COUNT), MI_MATH_REG_ACCU); in create_spin_counter()
124 *cs++ = MI_STORE_REGISTER_MEM_GEN8; in create_spin_counter()
125 *cs++ = i915_mmio_reg_offset(CS_GPR(COUNT)); in create_spin_counter()
126 *cs++ = lower_32_bits(i915_vma_offset(vma) + end * sizeof(*cs)); in create_spin_counter()
127 *cs++ = upper_32_bits(i915_vma_offset(vma) + end * sizeof(*cs)); in create_spin_counter()
131 *cs++ = MI_BATCH_BUFFER_START_GEN8; in create_spin_counter()
132 *cs++ = lower_32_bits(i915_vma_offset(vma) + loop * sizeof(*cs)); in create_spin_counter()
133 *cs++ = upper_32_bits(i915_vma_offset(vma) + loop * sizeof(*cs)); in create_spin_counter()
134 GEM_BUG_ON(cs - base > end); in create_spin_counter()
190 mutex_lock(&rps->lock); in rps_set_check()
193 mutex_unlock(&rps->lock); in rps_set_check()
196 GEM_BUG_ON(rps->last_freq != freq); in rps_set_check()
197 mutex_unlock(&rps->lock); in rps_set_check()
207 pr_info("P_STATE_CAP[%x]: 0x%08x\n", in show_pstate_limits()
212 pr_info("P_STATE_LIMITS[%x]: 0x%08x\n", in show_pstate_limits()
222 struct intel_rps *rps = &gt->rps; in live_rps_clock_interval()
230 if (!intel_rps_is_enabled(rps) || GRAPHICS_VER(gt->i915) < 6) in live_rps_clock_interval()
234 return -ENOMEM; in live_rps_clock_interval()
237 saved_work = rps->work.func; in live_rps_clock_interval()
238 rps->work.func = dummy_rps_work; in live_rps_clock_interval()
241 intel_rps_disable(&gt->rps); in live_rps_clock_interval()
256 engine->kernel_context, in live_rps_clock_interval()
268 engine->name); in live_rps_clock_interval()
271 intel_gt_set_wedged(engine->gt); in live_rps_clock_interval()
272 err = -EIO; in live_rps_clock_interval()
276 intel_uncore_forcewake_get(gt->uncore, FORCEWAKE_ALL); in live_rps_clock_interval()
278 intel_uncore_write_fw(gt->uncore, GEN6_RP_CUR_UP_EI, 0); in live_rps_clock_interval()
281 intel_uncore_write_fw(gt->uncore, in live_rps_clock_interval()
283 intel_uncore_write_fw(gt->uncore, in live_rps_clock_interval()
286 intel_uncore_write_fw(gt->uncore, GEN6_RP_CONTROL, in live_rps_clock_interval()
289 if (wait_for(intel_uncore_read_fw(gt->uncore, in live_rps_clock_interval()
294 engine->name); in live_rps_clock_interval()
295 err = -ENODEV; in live_rps_clock_interval()
304 cycles_[i] = -intel_uncore_read_fw(gt->uncore, GEN6_RP_CUR_UP_EI); in live_rps_clock_interval()
309 cycles_[i] += intel_uncore_read_fw(gt->uncore, GEN6_RP_CUR_UP_EI); in live_rps_clock_interval()
322 intel_uncore_write_fw(gt->uncore, GEN6_RP_CONTROL, 0); in live_rps_clock_interval()
323 intel_uncore_forcewake_put(gt->uncore, FORCEWAKE_ALL); in live_rps_clock_interval()
334 engine->name, cycles, time, dt, expected, in live_rps_clock_interval()
335 gt->clock_frequency / 1000); in live_rps_clock_interval()
340 engine->name); in live_rps_clock_interval()
341 err = -EINVAL; in live_rps_clock_interval()
347 engine->name); in live_rps_clock_interval()
348 err = -EINVAL; in live_rps_clock_interval()
352 if (igt_flush_test(gt->i915)) in live_rps_clock_interval()
353 err = -EIO; in live_rps_clock_interval()
358 intel_rps_enable(&gt->rps); in live_rps_clock_interval()
364 rps->work.func = saved_work; in live_rps_clock_interval()
366 if (err == -ENODEV) /* skipped, don't report a fail */ in live_rps_clock_interval()
375 struct intel_rps *rps = &gt->rps; in live_rps_control()
393 if (IS_CHERRYVIEW(gt->i915)) /* XXX fragile PCU */ in live_rps_control()
397 return -ENOMEM; in live_rps_control()
400 saved_work = rps->work.func; in live_rps_control()
401 rps->work.func = dummy_rps_work; in live_rps_control()
416 engine->kernel_context, in live_rps_control()
427 engine->name); in live_rps_control()
430 intel_gt_set_wedged(engine->gt); in live_rps_control()
431 err = -EIO; in live_rps_control()
435 if (rps_set_check(rps, rps->min_freq) != rps->min_freq) { in live_rps_control()
436 pr_err("%s: could not set minimum frequency [%x], only %x!\n", in live_rps_control()
437 engine->name, rps->min_freq, read_cagf(rps)); in live_rps_control()
441 err = -EINVAL; in live_rps_control()
445 for (f = rps->min_freq + 1; f < rps->max_freq; f++) { in live_rps_control()
452 if (rps_set_check(rps, rps->min_freq) != rps->min_freq) { in live_rps_control()
453 pr_err("%s: could not restore minimum frequency [%x], only %x!\n", in live_rps_control()
454 engine->name, rps->min_freq, read_cagf(rps)); in live_rps_control()
458 err = -EINVAL; in live_rps_control()
467 min = rps_set_check(rps, rps->min_freq); in live_rps_control()
473 pr_info("%s: range:[%x:%uMHz, %x:%uMHz] limit:[%x:%uMHz], %x:%x response %lluns:%lluns\n", in live_rps_control()
474 engine->name, in live_rps_control()
475 rps->min_freq, intel_gpu_freq(rps, rps->min_freq), in live_rps_control()
476 rps->max_freq, intel_gpu_freq(rps, rps->max_freq), in live_rps_control()
480 if (limit != rps->max_freq) { in live_rps_control()
481 u32 throttle = intel_uncore_read(gt->uncore, in live_rps_control()
484 pr_warn("%s: GPU throttled with reasons 0x%08x\n", in live_rps_control()
485 engine->name, throttle & GT0_PERF_LIMIT_REASONS_MASK); in live_rps_control()
489 if (igt_flush_test(gt->i915)) { in live_rps_control()
490 err = -EIO; in live_rps_control()
499 rps->work.func = saved_work; in live_rps_control()
514 min_gpu_freq = rps->min_freq; in show_pcu_config()
515 max_gpu_freq = rps->max_freq; in show_pcu_config()
522 wakeref = intel_runtime_pm_get(rps_to_uncore(rps)->rpm); in show_pcu_config()
528 snb_pcode_read(rps_to_gt(rps)->uncore, GEN6_PCODE_READ_MIN_FREQ_TABLE, in show_pcu_config()
537 intel_runtime_pm_put(rps_to_uncore(rps)->rpm, wakeref); in show_pcu_config()
547 dc = READ_ONCE(*cntr) - dc; in __measure_frequency()
548 dt = ktime_get() - dt; in __measure_frequency()
555 u64 x[5]; in measure_frequency_at() local
560 x[i] = __measure_frequency(cntr, 2); in measure_frequency_at()
564 sort(x, 5, sizeof(*x), cmp_u64, NULL); in measure_frequency_at()
565 return div_u64(x[1] + 2 * x[2] + x[3], 4); in measure_frequency_at()
573 dc = intel_uncore_read_fw(engine->uncore, CS_GPR(0)); in __measure_cs_frequency()
576 dc = intel_uncore_read_fw(engine->uncore, CS_GPR(0)) - dc; in __measure_cs_frequency()
577 dt = ktime_get() - dt; in __measure_cs_frequency()
586 u64 x[5]; in measure_cs_frequency_at() local
591 x[i] = __measure_cs_frequency(engine, 2); in measure_cs_frequency_at()
595 sort(x, 5, sizeof(*x), cmp_u64, NULL); in measure_cs_frequency_at()
596 return div_u64(x[1] + 2 * x[2] + x[3], 4); in measure_cs_frequency_at()
599 static bool scaled_within(u64 x, u64 y, u32 f_n, u32 f_d) in scaled_within() argument
601 return f_d * x > f_n * y && f_n * x < f_d * y; in scaled_within()
608 struct intel_rps *rps = &gt->rps; in live_rps_frequency_cs()
623 if (GRAPHICS_VER(gt->i915) < 8) /* for CS simplicity */ in live_rps_frequency_cs()
630 saved_work = rps->work.func; in live_rps_frequency_cs()
631 rps->work.func = dummy_rps_work; in live_rps_frequency_cs()
645 engine->kernel_context->vm, false, in live_rps_frequency_cs()
661 err = rq->engine->emit_bb_start(rq, in live_rps_frequency_cs()
668 if (wait_for(intel_uncore_read(engine->uncore, CS_GPR(0)), in live_rps_frequency_cs()
671 engine->name); in live_rps_frequency_cs()
675 min.freq = rps->min_freq; in live_rps_frequency_cs()
678 max.freq = rps->max_freq; in live_rps_frequency_cs()
682 engine->name, in live_rps_frequency_cs()
693 pr_err("%s: CS did not scale with frequency! scaled min:%llu, max:%llu\n", in live_rps_frequency_cs()
694 engine->name, in live_rps_frequency_cs()
699 for (f = min.freq + 1; f <= rps->max_freq; f++) { in live_rps_frequency_cs()
707 pr_info("%s: %x:%uMHz: %lluKHz [%d%%]\n", in live_rps_frequency_cs()
708 engine->name, in live_rps_frequency_cs()
716 err = -EINTR; /* ignore error, continue on with test */ in live_rps_frequency_cs()
721 i915_gem_object_flush_map(vma->obj); in live_rps_frequency_cs()
722 i915_gem_object_unpin_map(vma->obj); in live_rps_frequency_cs()
728 if (igt_flush_test(gt->i915)) in live_rps_frequency_cs()
729 err = -EIO; in live_rps_frequency_cs()
735 rps->work.func = saved_work; in live_rps_frequency_cs()
747 struct intel_rps *rps = &gt->rps; in live_rps_frequency_srm()
762 if (GRAPHICS_VER(gt->i915) < 8) /* for CS simplicity */ in live_rps_frequency_srm()
769 saved_work = rps->work.func; in live_rps_frequency_srm()
770 rps->work.func = dummy_rps_work; in live_rps_frequency_srm()
784 engine->kernel_context->vm, true, in live_rps_frequency_srm()
800 err = rq->engine->emit_bb_start(rq, in live_rps_frequency_srm()
809 engine->name); in live_rps_frequency_srm()
813 min.freq = rps->min_freq; in live_rps_frequency_srm()
816 max.freq = rps->max_freq; in live_rps_frequency_srm()
820 engine->name, in live_rps_frequency_srm()
831 pr_err("%s: CS did not scale with frequency! scaled min:%llu, max:%llu\n", in live_rps_frequency_srm()
832 engine->name, in live_rps_frequency_srm()
837 for (f = min.freq + 1; f <= rps->max_freq; f++) { in live_rps_frequency_srm()
845 pr_info("%s: %x:%uMHz: %lluKHz [%d%%]\n", in live_rps_frequency_srm()
846 engine->name, in live_rps_frequency_srm()
854 err = -EINTR; /* ignore error, continue on with test */ in live_rps_frequency_srm()
859 i915_gem_object_flush_map(vma->obj); in live_rps_frequency_srm()
860 i915_gem_object_unpin_map(vma->obj); in live_rps_frequency_srm()
866 if (igt_flush_test(gt->i915)) in live_rps_frequency_srm()
867 err = -EIO; in live_rps_frequency_srm()
873 rps->work.func = saved_work; in live_rps_frequency_srm()
888 GEM_BUG_ON(rps->pm_iir); in sleep_for_ei()
899 struct intel_uncore *uncore = engine->uncore; in __rps_up_interrupt()
906 rps_set_check(rps, rps->min_freq); in __rps_up_interrupt()
908 rq = igt_spinner_create_request(spin, engine->kernel_context, MI_NOOP); in __rps_up_interrupt()
917 engine->name); in __rps_up_interrupt()
919 intel_gt_set_wedged(engine->gt); in __rps_up_interrupt()
920 return -EIO; in __rps_up_interrupt()
925 engine->name); in __rps_up_interrupt()
928 return -EINVAL; in __rps_up_interrupt()
931 if (!(rps->pm_events & GEN6_PM_RP_UP_THRESHOLD)) { in __rps_up_interrupt()
933 engine->name); in __rps_up_interrupt()
935 return -EINVAL; in __rps_up_interrupt()
938 if (rps->last_freq != rps->min_freq) { in __rps_up_interrupt()
940 engine->name); in __rps_up_interrupt()
942 return -EINVAL; in __rps_up_interrupt()
946 timeout = intel_gt_pm_interval_to_ns(engine->gt, timeout); in __rps_up_interrupt()
955 if (rps->cur_freq != rps->min_freq) { in __rps_up_interrupt()
957 engine->name, intel_rps_read_actual_frequency(rps)); in __rps_up_interrupt()
958 return -EINVAL; in __rps_up_interrupt()
961 if (!(rps->pm_iir & GEN6_PM_RP_UP_THRESHOLD)) { in __rps_up_interrupt()
962 …err("%s: UP interrupt not recorded for spinner, pm_iir:%x, prev_up:%x, up_threshold:%x, up_ei:%x\n… in __rps_up_interrupt()
963 engine->name, rps->pm_iir, in __rps_up_interrupt()
967 return -EINVAL; in __rps_up_interrupt()
976 struct intel_uncore *uncore = engine->uncore; in __rps_down_interrupt()
979 rps_set_check(rps, rps->max_freq); in __rps_down_interrupt()
981 if (!(rps->pm_events & GEN6_PM_RP_DOWN_THRESHOLD)) { in __rps_down_interrupt()
983 engine->name); in __rps_down_interrupt()
984 return -EINVAL; in __rps_down_interrupt()
987 if (rps->last_freq != rps->max_freq) { in __rps_down_interrupt()
989 engine->name); in __rps_down_interrupt()
990 return -EINVAL; in __rps_down_interrupt()
994 timeout = intel_gt_pm_interval_to_ns(engine->gt, timeout); in __rps_down_interrupt()
999 if (rps->cur_freq != rps->max_freq) { in __rps_down_interrupt()
1001 engine->name, in __rps_down_interrupt()
1003 return -EINVAL; in __rps_down_interrupt()
1006 if (!(rps->pm_iir & (GEN6_PM_RP_DOWN_THRESHOLD | GEN6_PM_RP_DOWN_TIMEOUT))) { in __rps_down_interrupt()
1007 …pt not recorded for idle, pm_iir:%x, prev_down:%x, down_threshold:%x, down_ei:%x [prev_up:%x, up_t… in __rps_down_interrupt()
1008 engine->name, rps->pm_iir, in __rps_down_interrupt()
1015 return -EINVAL; in __rps_down_interrupt()
1024 struct intel_rps *rps = &gt->rps; in live_rps_interrupt()
1037 if (!intel_rps_has_interrupts(rps) || GRAPHICS_VER(gt->i915) < 6) in live_rps_interrupt()
1042 pm_events = rps->pm_events; in live_rps_interrupt()
1045 return -ENODEV; in live_rps_interrupt()
1049 return -ENOMEM; in live_rps_interrupt()
1052 saved_work = rps->work.func; in live_rps_interrupt()
1053 rps->work.func = dummy_rps_work; in live_rps_interrupt()
1058 intel_gt_pm_wait_for_idle(engine->gt); in live_rps_interrupt()
1069 intel_gt_pm_wait_for_idle(engine->gt); in live_rps_interrupt()
1075 intel_rc6_disable(&gt->rc6); in live_rps_interrupt()
1079 intel_rc6_enable(&gt->rc6); in live_rps_interrupt()
1087 if (igt_flush_test(gt->i915)) in live_rps_interrupt()
1088 err = -EIO; in live_rps_interrupt()
1093 rps->work.func = saved_work; in live_rps_interrupt()
1105 dE = librapl_energy_uJ() - dE; in __measure_power()
1106 dt = ktime_get() - dt; in __measure_power()
1113 u64 x[5]; in measure_power() local
1117 x[i] = __measure_power(5); in measure_power()
1122 sort(x, 5, sizeof(*x), cmp_u64, NULL); in measure_power()
1123 return div_u64(x[1] + 2 * x[2] + x[3], 4); in measure_power()
1136 struct intel_rps *rps = &gt->rps; in live_rps_power()
1149 if (!intel_rps_is_enabled(rps) || GRAPHICS_VER(gt->i915) < 6) in live_rps_power()
1152 if (!librapl_supported(gt->i915)) in live_rps_power()
1156 return -ENOMEM; in live_rps_power()
1159 saved_work = rps->work.func; in live_rps_power()
1160 rps->work.func = dummy_rps_work; in live_rps_power()
1175 engine->kernel_context, in live_rps_power()
1187 engine->name); in live_rps_power()
1190 intel_gt_set_wedged(engine->gt); in live_rps_power()
1191 err = -EIO; in live_rps_power()
1195 max.freq = rps->max_freq; in live_rps_power()
1198 min.freq = rps->min_freq; in live_rps_power()
1205 engine->name, in live_rps_power()
1218 engine->name); in live_rps_power()
1219 err = -EINVAL; in live_rps_power()
1223 if (igt_flush_test(gt->i915)) { in live_rps_power()
1224 err = -EIO; in live_rps_power()
1232 rps->work.func = saved_work; in live_rps_power()
1240 struct intel_rps *rps = &gt->rps; in live_rps_dynamic()
1253 if (!intel_rps_is_enabled(rps) || GRAPHICS_VER(gt->i915) < 6) in live_rps_dynamic()
1257 return -ENOMEM; in live_rps_dynamic()
1276 rps->cur_freq = rps->min_freq; in live_rps_dynamic()
1279 intel_rc6_disable(&gt->rc6); in live_rps_dynamic()
1280 GEM_BUG_ON(rps->last_freq != rps->min_freq); in live_rps_dynamic()
1283 engine->kernel_context, in live_rps_dynamic()
1293 max.freq = wait_for_freq(rps, rps->max_freq, 500); in live_rps_dynamic()
1299 min.freq = wait_for_freq(rps, rps->min_freq, 2000); in live_rps_dynamic()
1303 engine->name, in live_rps_dynamic()
1310 engine->name); in live_rps_dynamic()
1311 err = -EINVAL; in live_rps_dynamic()
1315 intel_rc6_enable(&gt->rc6); in live_rps_dynamic()
1318 if (igt_flush_test(gt->i915)) in live_rps_dynamic()
1319 err = -EIO; in live_rps_dynamic()