1 // SPDX-License-Identifier: MIT
3 * Copyright © 2021 Intel Corporation
7 #define H2G_DELAY 50000
8 #define delay_for_h2g() usleep_range(H2G_DELAY, H2G_DELAY + 10000)
9 #define FREQUENCY_REQ_UNIT DIV_ROUND_CLOSEST(GT_FREQUENCY_MULTIPLIER, \
20 struct kthread_worker *worker;
21 struct kthread_work work;
26 static int slpc_set_min_freq(struct intel_guc_slpc *slpc, u32 freq)
30 ret = intel_guc_slpc_set_min_freq(slpc, freq);
32 pr_err("Could not set min frequency to [%u]\n", freq);
33 else /* Delay to ensure h2g completes */
39 static int slpc_set_max_freq(struct intel_guc_slpc *slpc, u32 freq)
43 ret = intel_guc_slpc_set_max_freq(slpc, freq);
45 pr_err("Could not set maximum frequency [%u]\n",
47 else /* Delay to ensure h2g completes */
53 static int slpc_set_freq(struct intel_gt *gt, u32 freq)
56 struct intel_guc_slpc *slpc = >_to_guc(gt)->slpc;
58 err = slpc_set_max_freq(slpc, freq);
60 pr_err("Unable to update max freq");
64 err = slpc_set_min_freq(slpc, freq);
66 pr_err("Unable to update min freq");
73 static int slpc_restore_freq(struct intel_guc_slpc *slpc, u32 min, u32 max)
77 err = slpc_set_max_freq(slpc, max);
79 pr_err("Unable to restore max freq");
83 err = slpc_set_min_freq(slpc, min);
85 pr_err("Unable to restore min freq");
89 err = intel_guc_slpc_set_ignore_eff_freq(slpc, false);
91 pr_err("Unable to restore efficient freq");
98 static u64 measure_power_at_freq(struct intel_gt *gt, int *freq, u64 *power)
102 err = slpc_set_freq(gt, *freq);
105 *freq = intel_rps_read_actual_frequency(>->rps);
106 *power = measure_power(>->rps, freq);
111 static int vary_max_freq(struct intel_guc_slpc *slpc, struct intel_rps *rps,
114 u32 step, max_freq, req_freq;
118 /* Go from max to min in 5 steps */
119 step = (slpc->rp0_freq - slpc->min_freq) / NUM_STEPS;
120 *max_act_freq = slpc->min_freq;
121 for (max_freq = slpc->rp0_freq; max_freq > slpc->min_freq;
123 err = slpc_set_max_freq(slpc, max_freq);
127 req_freq = intel_rps_read_punit_req_frequency(rps);
129 /* GuC requests freq in multiples of 50/3 MHz */
130 if (req_freq > (max_freq + FREQUENCY_REQ_UNIT)) {
131 pr_err("SWReq is %d, should be at most %d\n", req_freq,
132 max_freq + FREQUENCY_REQ_UNIT);
136 act_freq = intel_rps_read_actual_frequency(rps);
137 if (act_freq > *max_act_freq)
138 *max_act_freq = act_freq;
147 static int vary_min_freq(struct intel_guc_slpc *slpc, struct intel_rps *rps,
150 u32 step, min_freq, req_freq;
154 /* Go from min to max in 5 steps */
155 step = (slpc->rp0_freq - slpc->min_freq) / NUM_STEPS;
156 *max_act_freq = slpc->min_freq;
157 for (min_freq = slpc->min_freq; min_freq < slpc->rp0_freq;
159 err = slpc_set_min_freq(slpc, min_freq);
163 req_freq = intel_rps_read_punit_req_frequency(rps);
165 /* GuC requests freq in multiples of 50/3 MHz */
166 if (req_freq < (min_freq - FREQUENCY_REQ_UNIT)) {
167 pr_err("SWReq is %d, should be at least %d\n", req_freq,
168 min_freq - FREQUENCY_REQ_UNIT);
172 act_freq = intel_rps_read_actual_frequency(rps);
173 if (act_freq > *max_act_freq)
174 *max_act_freq = act_freq;
183 static int slpc_power(struct intel_gt *gt, struct intel_engine_cs *engine)
185 struct intel_guc_slpc *slpc = >_to_guc(gt)->slpc;
193 * Our fundamental assumption is that running at lower frequency
194 * actually saves power. Let's see if our RAPL measurement supports
197 if (!librapl_supported(gt->i915))
200 min.freq = slpc->min_freq;
201 err = measure_power_at_freq(gt, &min.freq, &min.power);
206 max.freq = slpc->rp0_freq;
207 err = measure_power_at_freq(gt, &max.freq, &max.power);
212 pr_info("%s: min:%llumW @ %uMHz, max:%llumW @ %uMHz\n",
215 max.power, max.freq);
217 if (10 * min.freq >= 9 * max.freq) {
218 pr_notice("Could not control frequency, ran at [%uMHz, %uMhz]\n",
222 if (11 * min.power > 10 * max.power) {
223 pr_err("%s: did not conserve power when setting lower frequency!\n",
228 /* Restore min/max frequencies */
229 slpc_set_max_freq(slpc, slpc->rp0_freq);
230 slpc_set_min_freq(slpc, slpc->min_freq);
235 static int max_granted_freq(struct intel_guc_slpc *slpc, struct intel_rps *rps, u32 *max_act_freq)
237 struct intel_gt *gt = rps_to_gt(rps);
238 u32 perf_limit_reasons;
241 err = slpc_set_min_freq(slpc, slpc->rp0_freq);
245 *max_act_freq = intel_rps_read_actual_frequency(rps);
246 if (*max_act_freq != slpc->rp0_freq) {
247 /* Check if there was some throttling by pcode */
248 perf_limit_reasons = intel_uncore_read(gt->uncore,
249 intel_gt_perf_limit_reasons_reg(gt));
251 /* If not, this is an error */
252 if (!(perf_limit_reasons & GT0_PERF_LIMIT_REASONS_MASK)) {
253 pr_err("Pcode did not grant max freq\n");
256 pr_info("Pcode throttled frequency 0x%x\n", perf_limit_reasons);
263 static int run_test(struct intel_gt *gt, int test_type)
265 struct intel_guc_slpc *slpc = >_to_guc(gt)->slpc;
266 struct intel_rps *rps = >->rps;
267 struct intel_engine_cs *engine;
268 enum intel_engine_id id;
269 intel_wakeref_t wakeref;
270 struct igt_spinner spin;
271 u32 slpc_min_freq, slpc_max_freq;
274 if (!intel_uc_uses_guc_slpc(>->uc))
277 if (slpc->min_freq == slpc->rp0_freq) {
278 pr_err("Min/Max are fused to the same value\n");
282 if (igt_spinner_init(&spin, gt))
285 if (intel_guc_slpc_get_max_freq(slpc, &slpc_max_freq)) {
286 pr_err("Could not get SLPC max freq\n");
290 if (intel_guc_slpc_get_min_freq(slpc, &slpc_min_freq)) {
291 pr_err("Could not get SLPC min freq\n");
296 * Set min frequency to RPn so that we can test the whole
299 err = slpc_set_min_freq(slpc, slpc->min_freq);
301 pr_err("Unable to update min freq!");
306 * Turn off efficient frequency so RPn/RP0 ranges are obeyed.
308 err = intel_guc_slpc_set_ignore_eff_freq(slpc, true);
310 pr_err("Unable to turn off efficient freq!");
314 intel_gt_pm_wait_for_idle(gt);
315 wakeref = intel_gt_pm_get(gt);
316 for_each_engine(engine, gt, id) {
317 struct i915_request *rq;
320 if (!intel_engine_can_store_dword(engine))
323 st_engine_heartbeat_disable(engine);
325 rq = igt_spinner_create_request(&spin,
326 engine->kernel_context,
330 st_engine_heartbeat_enable(engine);
334 i915_request_add(rq);
336 if (!igt_wait_for_spinner(&spin, rq)) {
337 pr_err("%s: Spinner did not start\n",
339 igt_spinner_end(&spin);
340 st_engine_heartbeat_enable(engine);
341 intel_gt_set_wedged(engine->gt);
348 err = vary_min_freq(slpc, rps, &max_act_freq);
352 err = vary_max_freq(slpc, rps, &max_act_freq);
356 case TILE_INTERACTION:
357 /* Media engines have a different RP0 */
358 if (gt->type != GT_MEDIA && (engine->class == VIDEO_DECODE_CLASS ||
359 engine->class == VIDEO_ENHANCEMENT_CLASS)) {
360 igt_spinner_end(&spin);
361 st_engine_heartbeat_enable(engine);
366 err = max_granted_freq(slpc, rps, &max_act_freq);
370 err = slpc_power(gt, engine);
374 if (test_type != SLPC_POWER) {
375 pr_info("Max actual frequency for %s was %d\n",
376 engine->name, max_act_freq);
378 /* Actual frequency should rise above min */
379 if (max_act_freq <= slpc->min_freq) {
380 pr_err("Actual freq did not rise above min\n");
381 pr_err("Perf Limit Reasons: 0x%x\n",
382 intel_uncore_read(gt->uncore,
383 intel_gt_perf_limit_reasons_reg(gt)));
388 igt_spinner_end(&spin);
389 st_engine_heartbeat_enable(engine);
395 /* Restore min/max/efficient frequencies */
396 err = slpc_restore_freq(slpc, slpc_min_freq, slpc_max_freq);
398 if (igt_flush_test(gt->i915))
401 intel_gt_pm_put(gt, wakeref);
402 igt_spinner_fini(&spin);
403 intel_gt_pm_wait_for_idle(gt);
408 static int live_slpc_vary_min(void *arg)
410 struct drm_i915_private *i915 = arg;
415 for_each_gt(gt, i915, i) {
416 ret = run_test(gt, VARY_MIN);
424 static int live_slpc_vary_max(void *arg)
426 struct drm_i915_private *i915 = arg;
431 for_each_gt(gt, i915, i) {
432 ret = run_test(gt, VARY_MAX);
440 /* check if pcode can grant RP0 */
441 static int live_slpc_max_granted(void *arg)
443 struct drm_i915_private *i915 = arg;
448 for_each_gt(gt, i915, i) {
449 ret = run_test(gt, MAX_GRANTED);
457 static int live_slpc_power(void *arg)
459 struct drm_i915_private *i915 = arg;
464 for_each_gt(gt, i915, i) {
465 ret = run_test(gt, SLPC_POWER);
473 static void slpc_spinner_thread(struct kthread_work *work)
475 struct slpc_thread *thread = container_of(work, typeof(*thread), work);
477 thread->result = run_test(thread->gt, TILE_INTERACTION);
480 static int live_slpc_tile_interaction(void *arg)
482 struct drm_i915_private *i915 = arg;
484 struct slpc_thread *threads;
487 threads = kcalloc(I915_MAX_GT, sizeof(*threads), GFP_KERNEL);
491 for_each_gt(gt, i915, i) {
492 threads[i].worker = kthread_create_worker(0, "igt/slpc_parallel:%d", gt->info.id);
494 if (IS_ERR(threads[i].worker)) {
495 ret = PTR_ERR(threads[i].worker);
500 kthread_init_work(&threads[i].work, slpc_spinner_thread);
501 kthread_queue_work(threads[i].worker, &threads[i].work);
504 for_each_gt(gt, i915, i) {
507 if (IS_ERR_OR_NULL(threads[i].worker))
510 kthread_flush_work(&threads[i].work);
511 status = READ_ONCE(threads[i].result);
512 if (status && !ret) {
513 pr_err("%s GT %d failed ", __func__, gt->info.id);
516 kthread_destroy_worker(threads[i].worker);
523 int intel_slpc_live_selftests(struct drm_i915_private *i915)
525 static const struct i915_subtest tests[] = {
526 SUBTEST(live_slpc_vary_max),
527 SUBTEST(live_slpc_vary_min),
528 SUBTEST(live_slpc_max_granted),
529 SUBTEST(live_slpc_power),
530 SUBTEST(live_slpc_tile_interaction),
536 for_each_gt(gt, i915, i) {
537 if (intel_gt_is_wedged(gt))
541 return i915_live_subtests(tests, i915);