1 // SPDX-License-Identifier: MIT
2 /*
3 * Copyright © 2021 Intel Corporation
4 */
5
6 #include <drm/drm_cache.h>
7 #include <linux/string_helpers.h>
8
9 #include "i915_drv.h"
10 #include "i915_reg.h"
11 #include "intel_guc_slpc.h"
12 #include "intel_mchbar_regs.h"
13 #include "gt/intel_gt.h"
14 #include "gt/intel_gt_regs.h"
15 #include "gt/intel_rps.h"
16
slpc_to_guc(struct intel_guc_slpc * slpc)17 static inline struct intel_guc *slpc_to_guc(struct intel_guc_slpc *slpc)
18 {
19 return container_of(slpc, struct intel_guc, slpc);
20 }
21
slpc_to_gt(struct intel_guc_slpc * slpc)22 static inline struct intel_gt *slpc_to_gt(struct intel_guc_slpc *slpc)
23 {
24 return guc_to_gt(slpc_to_guc(slpc));
25 }
26
slpc_to_i915(struct intel_guc_slpc * slpc)27 static inline struct drm_i915_private *slpc_to_i915(struct intel_guc_slpc *slpc)
28 {
29 return slpc_to_gt(slpc)->i915;
30 }
31
__detect_slpc_supported(struct intel_guc * guc)32 static bool __detect_slpc_supported(struct intel_guc *guc)
33 {
34 /* GuC SLPC is unavailable for pre-Gen12 */
35 return guc->submission_supported &&
36 GRAPHICS_VER(guc_to_gt(guc)->i915) >= 12;
37 }
38
__guc_slpc_selected(struct intel_guc * guc)39 static bool __guc_slpc_selected(struct intel_guc *guc)
40 {
41 if (!intel_guc_slpc_is_supported(guc))
42 return false;
43
44 return guc->submission_selected;
45 }
46
intel_guc_slpc_init_early(struct intel_guc_slpc * slpc)47 void intel_guc_slpc_init_early(struct intel_guc_slpc *slpc)
48 {
49 struct intel_guc *guc = slpc_to_guc(slpc);
50
51 slpc->supported = __detect_slpc_supported(guc);
52 slpc->selected = __guc_slpc_selected(guc);
53 }
54
slpc_mem_set_param(struct slpc_shared_data * data,u32 id,u32 value)55 static void slpc_mem_set_param(struct slpc_shared_data *data,
56 u32 id, u32 value)
57 {
58 GEM_BUG_ON(id >= SLPC_MAX_OVERRIDE_PARAMETERS);
59 /*
60 * When the flag bit is set, corresponding value will be read
61 * and applied by SLPC.
62 */
63 data->override_params.bits[id >> 5] |= (1 << (id % 32));
64 data->override_params.values[id] = value;
65 }
66
slpc_mem_set_enabled(struct slpc_shared_data * data,u8 enable_id,u8 disable_id)67 static void slpc_mem_set_enabled(struct slpc_shared_data *data,
68 u8 enable_id, u8 disable_id)
69 {
70 /*
71 * Enabling a param involves setting the enable_id
72 * to 1 and disable_id to 0.
73 */
74 slpc_mem_set_param(data, enable_id, 1);
75 slpc_mem_set_param(data, disable_id, 0);
76 }
77
slpc_mem_set_disabled(struct slpc_shared_data * data,u8 enable_id,u8 disable_id)78 static void slpc_mem_set_disabled(struct slpc_shared_data *data,
79 u8 enable_id, u8 disable_id)
80 {
81 /*
82 * Disabling a param involves setting the enable_id
83 * to 0 and disable_id to 1.
84 */
85 slpc_mem_set_param(data, disable_id, 1);
86 slpc_mem_set_param(data, enable_id, 0);
87 }
88
slpc_get_state(struct intel_guc_slpc * slpc)89 static u32 slpc_get_state(struct intel_guc_slpc *slpc)
90 {
91 struct slpc_shared_data *data;
92
93 GEM_BUG_ON(!slpc->vma);
94
95 drm_clflush_virt_range(slpc->vaddr, sizeof(u32));
96 data = slpc->vaddr;
97
98 return data->header.global_state;
99 }
100
guc_action_slpc_set_param_nb(struct intel_guc * guc,u8 id,u32 value)101 static int guc_action_slpc_set_param_nb(struct intel_guc *guc, u8 id, u32 value)
102 {
103 u32 request[] = {
104 GUC_ACTION_HOST2GUC_PC_SLPC_REQUEST,
105 SLPC_EVENT(SLPC_EVENT_PARAMETER_SET, 2),
106 id,
107 value,
108 };
109 int ret;
110
111 ret = intel_guc_send_nb(guc, request, ARRAY_SIZE(request), 0);
112
113 return ret > 0 ? -EPROTO : ret;
114 }
115
slpc_set_param_nb(struct intel_guc_slpc * slpc,u8 id,u32 value)116 static int slpc_set_param_nb(struct intel_guc_slpc *slpc, u8 id, u32 value)
117 {
118 struct intel_guc *guc = slpc_to_guc(slpc);
119
120 GEM_BUG_ON(id >= SLPC_MAX_PARAM);
121
122 return guc_action_slpc_set_param_nb(guc, id, value);
123 }
124
guc_action_slpc_set_param(struct intel_guc * guc,u8 id,u32 value)125 static int guc_action_slpc_set_param(struct intel_guc *guc, u8 id, u32 value)
126 {
127 u32 request[] = {
128 GUC_ACTION_HOST2GUC_PC_SLPC_REQUEST,
129 SLPC_EVENT(SLPC_EVENT_PARAMETER_SET, 2),
130 id,
131 value,
132 };
133 int ret;
134
135 ret = intel_guc_send(guc, request, ARRAY_SIZE(request));
136
137 return ret > 0 ? -EPROTO : ret;
138 }
139
guc_action_slpc_unset_param(struct intel_guc * guc,u8 id)140 static int guc_action_slpc_unset_param(struct intel_guc *guc, u8 id)
141 {
142 u32 request[] = {
143 GUC_ACTION_HOST2GUC_PC_SLPC_REQUEST,
144 SLPC_EVENT(SLPC_EVENT_PARAMETER_UNSET, 1),
145 id,
146 };
147
148 return intel_guc_send(guc, request, ARRAY_SIZE(request));
149 }
150
slpc_is_running(struct intel_guc_slpc * slpc)151 static bool slpc_is_running(struct intel_guc_slpc *slpc)
152 {
153 return slpc_get_state(slpc) == SLPC_GLOBAL_STATE_RUNNING;
154 }
155
guc_action_slpc_query(struct intel_guc * guc,u32 offset)156 static int guc_action_slpc_query(struct intel_guc *guc, u32 offset)
157 {
158 u32 request[] = {
159 GUC_ACTION_HOST2GUC_PC_SLPC_REQUEST,
160 SLPC_EVENT(SLPC_EVENT_QUERY_TASK_STATE, 2),
161 offset,
162 0,
163 };
164 int ret;
165
166 ret = intel_guc_send(guc, request, ARRAY_SIZE(request));
167
168 return ret > 0 ? -EPROTO : ret;
169 }
170
slpc_query_task_state(struct intel_guc_slpc * slpc)171 static int slpc_query_task_state(struct intel_guc_slpc *slpc)
172 {
173 struct intel_guc *guc = slpc_to_guc(slpc);
174 struct drm_i915_private *i915 = slpc_to_i915(slpc);
175 u32 offset = intel_guc_ggtt_offset(guc, slpc->vma);
176 int ret;
177
178 ret = guc_action_slpc_query(guc, offset);
179 if (unlikely(ret))
180 i915_probe_error(i915, "Failed to query task state (%pe)\n",
181 ERR_PTR(ret));
182
183 drm_clflush_virt_range(slpc->vaddr, SLPC_PAGE_SIZE_BYTES);
184
185 return ret;
186 }
187
slpc_set_param(struct intel_guc_slpc * slpc,u8 id,u32 value)188 static int slpc_set_param(struct intel_guc_slpc *slpc, u8 id, u32 value)
189 {
190 struct intel_guc *guc = slpc_to_guc(slpc);
191 struct drm_i915_private *i915 = slpc_to_i915(slpc);
192 int ret;
193
194 GEM_BUG_ON(id >= SLPC_MAX_PARAM);
195
196 ret = guc_action_slpc_set_param(guc, id, value);
197 if (ret)
198 i915_probe_error(i915, "Failed to set param %d to %u (%pe)\n",
199 id, value, ERR_PTR(ret));
200
201 return ret;
202 }
203
slpc_unset_param(struct intel_guc_slpc * slpc,u8 id)204 static int slpc_unset_param(struct intel_guc_slpc *slpc, u8 id)
205 {
206 struct intel_guc *guc = slpc_to_guc(slpc);
207
208 GEM_BUG_ON(id >= SLPC_MAX_PARAM);
209
210 return guc_action_slpc_unset_param(guc, id);
211 }
212
slpc_force_min_freq(struct intel_guc_slpc * slpc,u32 freq)213 static int slpc_force_min_freq(struct intel_guc_slpc *slpc, u32 freq)
214 {
215 struct drm_i915_private *i915 = slpc_to_i915(slpc);
216 struct intel_guc *guc = slpc_to_guc(slpc);
217 intel_wakeref_t wakeref;
218 int ret = 0;
219
220 lockdep_assert_held(&slpc->lock);
221
222 if (!intel_guc_is_ready(guc))
223 return -ENODEV;
224
225 /*
226 * This function is a little different as compared to
227 * intel_guc_slpc_set_min_freq(). Softlimit will not be updated
228 * here since this is used to temporarily change min freq,
229 * for example, during a waitboost. Caller is responsible for
230 * checking bounds.
231 */
232
233 with_intel_runtime_pm(&i915->runtime_pm, wakeref) {
234 /* Non-blocking request will avoid stalls */
235 ret = slpc_set_param_nb(slpc,
236 SLPC_PARAM_GLOBAL_MIN_GT_UNSLICE_FREQ_MHZ,
237 freq);
238 if (ret)
239 drm_notice(&i915->drm,
240 "Failed to send set_param for min freq(%d): (%d)\n",
241 freq, ret);
242 }
243
244 return ret;
245 }
246
slpc_boost_work(struct work_struct * work)247 static void slpc_boost_work(struct work_struct *work)
248 {
249 struct intel_guc_slpc *slpc = container_of(work, typeof(*slpc), boost_work);
250 int err;
251
252 /*
253 * Raise min freq to boost. It's possible that
254 * this is greater than current max. But it will
255 * certainly be limited by RP0. An error setting
256 * the min param is not fatal.
257 */
258 mutex_lock(&slpc->lock);
259 if (atomic_read(&slpc->num_waiters)) {
260 err = slpc_force_min_freq(slpc, slpc->boost_freq);
261 if (!err)
262 slpc->num_boosts++;
263 }
264 mutex_unlock(&slpc->lock);
265 }
266
intel_guc_slpc_init(struct intel_guc_slpc * slpc)267 int intel_guc_slpc_init(struct intel_guc_slpc *slpc)
268 {
269 struct intel_guc *guc = slpc_to_guc(slpc);
270 struct drm_i915_private *i915 = slpc_to_i915(slpc);
271 u32 size = PAGE_ALIGN(sizeof(struct slpc_shared_data));
272 int err;
273
274 GEM_BUG_ON(slpc->vma);
275
276 err = intel_guc_allocate_and_map_vma(guc, size, &slpc->vma, (void **)&slpc->vaddr);
277 if (unlikely(err)) {
278 i915_probe_error(i915,
279 "Failed to allocate SLPC struct (err=%pe)\n",
280 ERR_PTR(err));
281 return err;
282 }
283
284 slpc->max_freq_softlimit = 0;
285 slpc->min_freq_softlimit = 0;
286 slpc->min_is_rpmax = false;
287
288 slpc->boost_freq = 0;
289 atomic_set(&slpc->num_waiters, 0);
290 slpc->num_boosts = 0;
291 slpc->media_ratio_mode = SLPC_MEDIA_RATIO_MODE_DYNAMIC_CONTROL;
292
293 mutex_init(&slpc->lock);
294 INIT_WORK(&slpc->boost_work, slpc_boost_work);
295
296 return err;
297 }
298
slpc_global_state_to_string(enum slpc_global_state state)299 static const char *slpc_global_state_to_string(enum slpc_global_state state)
300 {
301 switch (state) {
302 case SLPC_GLOBAL_STATE_NOT_RUNNING:
303 return "not running";
304 case SLPC_GLOBAL_STATE_INITIALIZING:
305 return "initializing";
306 case SLPC_GLOBAL_STATE_RESETTING:
307 return "resetting";
308 case SLPC_GLOBAL_STATE_RUNNING:
309 return "running";
310 case SLPC_GLOBAL_STATE_SHUTTING_DOWN:
311 return "shutting down";
312 case SLPC_GLOBAL_STATE_ERROR:
313 return "error";
314 default:
315 return "unknown";
316 }
317 }
318
slpc_get_state_string(struct intel_guc_slpc * slpc)319 static const char *slpc_get_state_string(struct intel_guc_slpc *slpc)
320 {
321 return slpc_global_state_to_string(slpc_get_state(slpc));
322 }
323
guc_action_slpc_reset(struct intel_guc * guc,u32 offset)324 static int guc_action_slpc_reset(struct intel_guc *guc, u32 offset)
325 {
326 u32 request[] = {
327 GUC_ACTION_HOST2GUC_PC_SLPC_REQUEST,
328 SLPC_EVENT(SLPC_EVENT_RESET, 2),
329 offset,
330 0,
331 };
332 int ret;
333
334 ret = intel_guc_send(guc, request, ARRAY_SIZE(request));
335
336 return ret > 0 ? -EPROTO : ret;
337 }
338
slpc_reset(struct intel_guc_slpc * slpc)339 static int slpc_reset(struct intel_guc_slpc *slpc)
340 {
341 struct drm_i915_private *i915 = slpc_to_i915(slpc);
342 struct intel_guc *guc = slpc_to_guc(slpc);
343 u32 offset = intel_guc_ggtt_offset(guc, slpc->vma);
344 int ret;
345
346 ret = guc_action_slpc_reset(guc, offset);
347
348 if (unlikely(ret < 0)) {
349 i915_probe_error(i915, "SLPC reset action failed (%pe)\n",
350 ERR_PTR(ret));
351 return ret;
352 }
353
354 if (!ret) {
355 if (wait_for(slpc_is_running(slpc), SLPC_RESET_TIMEOUT_MS)) {
356 i915_probe_error(i915, "SLPC not enabled! State = %s\n",
357 slpc_get_state_string(slpc));
358 return -EIO;
359 }
360 }
361
362 return 0;
363 }
364
slpc_decode_min_freq(struct intel_guc_slpc * slpc)365 static u32 slpc_decode_min_freq(struct intel_guc_slpc *slpc)
366 {
367 struct slpc_shared_data *data = slpc->vaddr;
368
369 GEM_BUG_ON(!slpc->vma);
370
371 return DIV_ROUND_CLOSEST(REG_FIELD_GET(SLPC_MIN_UNSLICE_FREQ_MASK,
372 data->task_state_data.freq) *
373 GT_FREQUENCY_MULTIPLIER, GEN9_FREQ_SCALER);
374 }
375
slpc_decode_max_freq(struct intel_guc_slpc * slpc)376 static u32 slpc_decode_max_freq(struct intel_guc_slpc *slpc)
377 {
378 struct slpc_shared_data *data = slpc->vaddr;
379
380 GEM_BUG_ON(!slpc->vma);
381
382 return DIV_ROUND_CLOSEST(REG_FIELD_GET(SLPC_MAX_UNSLICE_FREQ_MASK,
383 data->task_state_data.freq) *
384 GT_FREQUENCY_MULTIPLIER, GEN9_FREQ_SCALER);
385 }
386
slpc_shared_data_reset(struct slpc_shared_data * data)387 static void slpc_shared_data_reset(struct slpc_shared_data *data)
388 {
389 memset(data, 0, sizeof(struct slpc_shared_data));
390
391 data->header.size = sizeof(struct slpc_shared_data);
392
393 /* Enable only GTPERF task, disable others */
394 slpc_mem_set_enabled(data, SLPC_PARAM_TASK_ENABLE_GTPERF,
395 SLPC_PARAM_TASK_DISABLE_GTPERF);
396
397 slpc_mem_set_disabled(data, SLPC_PARAM_TASK_ENABLE_BALANCER,
398 SLPC_PARAM_TASK_DISABLE_BALANCER);
399
400 slpc_mem_set_disabled(data, SLPC_PARAM_TASK_ENABLE_DCC,
401 SLPC_PARAM_TASK_DISABLE_DCC);
402 }
403
404 /**
405 * intel_guc_slpc_set_max_freq() - Set max frequency limit for SLPC.
406 * @slpc: pointer to intel_guc_slpc.
407 * @val: frequency (MHz)
408 *
409 * This function will invoke GuC SLPC action to update the max frequency
410 * limit for unslice.
411 *
412 * Return: 0 on success, non-zero error code on failure.
413 */
intel_guc_slpc_set_max_freq(struct intel_guc_slpc * slpc,u32 val)414 int intel_guc_slpc_set_max_freq(struct intel_guc_slpc *slpc, u32 val)
415 {
416 struct drm_i915_private *i915 = slpc_to_i915(slpc);
417 intel_wakeref_t wakeref;
418 int ret;
419
420 if (val < slpc->min_freq ||
421 val > slpc->rp0_freq ||
422 val < slpc->min_freq_softlimit)
423 return -EINVAL;
424
425 with_intel_runtime_pm(&i915->runtime_pm, wakeref) {
426 ret = slpc_set_param(slpc,
427 SLPC_PARAM_GLOBAL_MAX_GT_UNSLICE_FREQ_MHZ,
428 val);
429
430 /* Return standardized err code for sysfs calls */
431 if (ret)
432 ret = -EIO;
433 }
434
435 if (!ret)
436 slpc->max_freq_softlimit = val;
437
438 return ret;
439 }
440
441 /**
442 * intel_guc_slpc_get_max_freq() - Get max frequency limit for SLPC.
443 * @slpc: pointer to intel_guc_slpc.
444 * @val: pointer to val which will hold max frequency (MHz)
445 *
446 * This function will invoke GuC SLPC action to read the max frequency
447 * limit for unslice.
448 *
449 * Return: 0 on success, non-zero error code on failure.
450 */
intel_guc_slpc_get_max_freq(struct intel_guc_slpc * slpc,u32 * val)451 int intel_guc_slpc_get_max_freq(struct intel_guc_slpc *slpc, u32 *val)
452 {
453 struct drm_i915_private *i915 = slpc_to_i915(slpc);
454 intel_wakeref_t wakeref;
455 int ret = 0;
456
457 with_intel_runtime_pm(&i915->runtime_pm, wakeref) {
458 /* Force GuC to update task data */
459 ret = slpc_query_task_state(slpc);
460
461 if (!ret)
462 *val = slpc_decode_max_freq(slpc);
463 }
464
465 return ret;
466 }
467
468 /**
469 * intel_guc_slpc_set_min_freq() - Set min frequency limit for SLPC.
470 * @slpc: pointer to intel_guc_slpc.
471 * @val: frequency (MHz)
472 *
473 * This function will invoke GuC SLPC action to update the min unslice
474 * frequency.
475 *
476 * Return: 0 on success, non-zero error code on failure.
477 */
intel_guc_slpc_set_min_freq(struct intel_guc_slpc * slpc,u32 val)478 int intel_guc_slpc_set_min_freq(struct intel_guc_slpc *slpc, u32 val)
479 {
480 struct drm_i915_private *i915 = slpc_to_i915(slpc);
481 intel_wakeref_t wakeref;
482 int ret;
483
484 if (val < slpc->min_freq ||
485 val > slpc->rp0_freq ||
486 val > slpc->max_freq_softlimit)
487 return -EINVAL;
488
489 /* Need a lock now since waitboost can be modifying min as well */
490 mutex_lock(&slpc->lock);
491 wakeref = intel_runtime_pm_get(&i915->runtime_pm);
492
493 /* Ignore efficient freq if lower min freq is requested */
494 ret = slpc_set_param(slpc,
495 SLPC_PARAM_IGNORE_EFFICIENT_FREQUENCY,
496 val < slpc->rp1_freq);
497 if (ret) {
498 i915_probe_error(i915, "Failed to toggle efficient freq (%pe)\n",
499 ERR_PTR(ret));
500 goto out;
501 }
502
503 ret = slpc_set_param(slpc,
504 SLPC_PARAM_GLOBAL_MIN_GT_UNSLICE_FREQ_MHZ,
505 val);
506
507 if (!ret)
508 slpc->min_freq_softlimit = val;
509
510 out:
511 intel_runtime_pm_put(&i915->runtime_pm, wakeref);
512 mutex_unlock(&slpc->lock);
513
514 /* Return standardized err code for sysfs calls */
515 if (ret)
516 ret = -EIO;
517
518 return ret;
519 }
520
521 /**
522 * intel_guc_slpc_get_min_freq() - Get min frequency limit for SLPC.
523 * @slpc: pointer to intel_guc_slpc.
524 * @val: pointer to val which will hold min frequency (MHz)
525 *
526 * This function will invoke GuC SLPC action to read the min frequency
527 * limit for unslice.
528 *
529 * Return: 0 on success, non-zero error code on failure.
530 */
intel_guc_slpc_get_min_freq(struct intel_guc_slpc * slpc,u32 * val)531 int intel_guc_slpc_get_min_freq(struct intel_guc_slpc *slpc, u32 *val)
532 {
533 struct drm_i915_private *i915 = slpc_to_i915(slpc);
534 intel_wakeref_t wakeref;
535 int ret = 0;
536
537 with_intel_runtime_pm(&i915->runtime_pm, wakeref) {
538 /* Force GuC to update task data */
539 ret = slpc_query_task_state(slpc);
540
541 if (!ret)
542 *val = slpc_decode_min_freq(slpc);
543 }
544
545 return ret;
546 }
547
intel_guc_slpc_set_media_ratio_mode(struct intel_guc_slpc * slpc,u32 val)548 int intel_guc_slpc_set_media_ratio_mode(struct intel_guc_slpc *slpc, u32 val)
549 {
550 struct drm_i915_private *i915 = slpc_to_i915(slpc);
551 intel_wakeref_t wakeref;
552 int ret = 0;
553
554 if (!HAS_MEDIA_RATIO_MODE(i915))
555 return -ENODEV;
556
557 with_intel_runtime_pm(&i915->runtime_pm, wakeref)
558 ret = slpc_set_param(slpc,
559 SLPC_PARAM_MEDIA_FF_RATIO_MODE,
560 val);
561 return ret;
562 }
563
intel_guc_pm_intrmsk_enable(struct intel_gt * gt)564 void intel_guc_pm_intrmsk_enable(struct intel_gt *gt)
565 {
566 u32 pm_intrmsk_mbz = 0;
567
568 /*
569 * Allow GuC to receive ARAT timer expiry event.
570 * This interrupt register is setup by RPS code
571 * when host based Turbo is enabled.
572 */
573 pm_intrmsk_mbz |= ARAT_EXPIRED_INTRMSK;
574
575 intel_uncore_rmw(gt->uncore,
576 GEN6_PMINTRMSK, pm_intrmsk_mbz, 0);
577 }
578
slpc_set_softlimits(struct intel_guc_slpc * slpc)579 static int slpc_set_softlimits(struct intel_guc_slpc *slpc)
580 {
581 int ret = 0;
582
583 /*
584 * Softlimits are initially equivalent to platform limits
585 * unless they have deviated from defaults, in which case,
586 * we retain the values and set min/max accordingly.
587 */
588 if (!slpc->max_freq_softlimit) {
589 slpc->max_freq_softlimit = slpc->rp0_freq;
590 slpc_to_gt(slpc)->defaults.max_freq = slpc->max_freq_softlimit;
591 } else if (slpc->max_freq_softlimit != slpc->rp0_freq) {
592 ret = intel_guc_slpc_set_max_freq(slpc,
593 slpc->max_freq_softlimit);
594 }
595
596 if (unlikely(ret))
597 return ret;
598
599 if (!slpc->min_freq_softlimit) {
600 ret = intel_guc_slpc_get_min_freq(slpc, &slpc->min_freq_softlimit);
601 if (unlikely(ret))
602 return ret;
603 slpc_to_gt(slpc)->defaults.min_freq = slpc->min_freq_softlimit;
604 } else if (slpc->min_freq_softlimit != slpc->min_freq) {
605 return intel_guc_slpc_set_min_freq(slpc,
606 slpc->min_freq_softlimit);
607 }
608
609 return 0;
610 }
611
is_slpc_min_freq_rpmax(struct intel_guc_slpc * slpc)612 static bool is_slpc_min_freq_rpmax(struct intel_guc_slpc *slpc)
613 {
614 struct drm_i915_private *i915 = slpc_to_i915(slpc);
615 int slpc_min_freq;
616 int ret;
617
618 ret = intel_guc_slpc_get_min_freq(slpc, &slpc_min_freq);
619 if (ret) {
620 drm_err(&i915->drm,
621 "Failed to get min freq: (%d)\n",
622 ret);
623 return false;
624 }
625
626 if (slpc_min_freq == SLPC_MAX_FREQ_MHZ)
627 return true;
628 else
629 return false;
630 }
631
update_server_min_softlimit(struct intel_guc_slpc * slpc)632 static void update_server_min_softlimit(struct intel_guc_slpc *slpc)
633 {
634 /* For server parts, SLPC min will be at RPMax.
635 * Use min softlimit to clamp it to RP0 instead.
636 */
637 if (!slpc->min_freq_softlimit &&
638 is_slpc_min_freq_rpmax(slpc)) {
639 slpc->min_is_rpmax = true;
640 slpc->min_freq_softlimit = slpc->rp0_freq;
641 (slpc_to_gt(slpc))->defaults.min_freq = slpc->min_freq_softlimit;
642 }
643 }
644
slpc_use_fused_rp0(struct intel_guc_slpc * slpc)645 static int slpc_use_fused_rp0(struct intel_guc_slpc *slpc)
646 {
647 /* Force SLPC to used platform rp0 */
648 return slpc_set_param(slpc,
649 SLPC_PARAM_GLOBAL_MAX_GT_UNSLICE_FREQ_MHZ,
650 slpc->rp0_freq);
651 }
652
slpc_get_rp_values(struct intel_guc_slpc * slpc)653 static void slpc_get_rp_values(struct intel_guc_slpc *slpc)
654 {
655 struct intel_rps *rps = &slpc_to_gt(slpc)->rps;
656 struct intel_rps_freq_caps caps;
657
658 gen6_rps_get_freq_caps(rps, &caps);
659 slpc->rp0_freq = intel_gpu_freq(rps, caps.rp0_freq);
660 slpc->rp1_freq = intel_gpu_freq(rps, caps.rp1_freq);
661 slpc->min_freq = intel_gpu_freq(rps, caps.min_freq);
662
663 if (!slpc->boost_freq)
664 slpc->boost_freq = slpc->rp0_freq;
665 }
666
667 /**
668 * intel_guc_slpc_override_gucrc_mode() - override GUCRC mode
669 * @slpc: pointer to intel_guc_slpc.
670 * @mode: new value of the mode.
671 *
672 * This function will override the GUCRC mode.
673 *
674 * Return: 0 on success, non-zero error code on failure.
675 */
intel_guc_slpc_override_gucrc_mode(struct intel_guc_slpc * slpc,u32 mode)676 int intel_guc_slpc_override_gucrc_mode(struct intel_guc_slpc *slpc, u32 mode)
677 {
678 int ret;
679 struct drm_i915_private *i915 = slpc_to_i915(slpc);
680 intel_wakeref_t wakeref;
681
682 if (mode >= SLPC_GUCRC_MODE_MAX)
683 return -EINVAL;
684
685 with_intel_runtime_pm(&i915->runtime_pm, wakeref) {
686 ret = slpc_set_param(slpc, SLPC_PARAM_PWRGATE_RC_MODE, mode);
687 if (ret)
688 drm_err(&i915->drm,
689 "Override gucrc mode %d failed %d\n",
690 mode, ret);
691 }
692
693 return ret;
694 }
695
intel_guc_slpc_unset_gucrc_mode(struct intel_guc_slpc * slpc)696 int intel_guc_slpc_unset_gucrc_mode(struct intel_guc_slpc *slpc)
697 {
698 struct drm_i915_private *i915 = slpc_to_i915(slpc);
699 intel_wakeref_t wakeref;
700 int ret = 0;
701
702 with_intel_runtime_pm(&i915->runtime_pm, wakeref) {
703 ret = slpc_unset_param(slpc, SLPC_PARAM_PWRGATE_RC_MODE);
704 if (ret)
705 drm_err(&i915->drm,
706 "Unsetting gucrc mode failed %d\n",
707 ret);
708 }
709
710 return ret;
711 }
712
713 /*
714 * intel_guc_slpc_enable() - Start SLPC
715 * @slpc: pointer to intel_guc_slpc.
716 *
717 * SLPC is enabled by setting up the shared data structure and
718 * sending reset event to GuC SLPC. Initial data is setup in
719 * intel_guc_slpc_init. Here we send the reset event. We do
720 * not currently need a slpc_disable since this is taken care
721 * of automatically when a reset/suspend occurs and the GuC
722 * CTB is destroyed.
723 *
724 * Return: 0 on success, non-zero error code on failure.
725 */
intel_guc_slpc_enable(struct intel_guc_slpc * slpc)726 int intel_guc_slpc_enable(struct intel_guc_slpc *slpc)
727 {
728 struct drm_i915_private *i915 = slpc_to_i915(slpc);
729 int ret;
730
731 GEM_BUG_ON(!slpc->vma);
732
733 slpc_shared_data_reset(slpc->vaddr);
734
735 ret = slpc_reset(slpc);
736 if (unlikely(ret < 0)) {
737 i915_probe_error(i915, "SLPC Reset event returned (%pe)\n",
738 ERR_PTR(ret));
739 return ret;
740 }
741
742 ret = slpc_query_task_state(slpc);
743 if (unlikely(ret < 0))
744 return ret;
745
746 intel_guc_pm_intrmsk_enable(to_gt(i915));
747
748 slpc_get_rp_values(slpc);
749
750 /* Handle the case where min=max=RPmax */
751 update_server_min_softlimit(slpc);
752
753 /* Set SLPC max limit to RP0 */
754 ret = slpc_use_fused_rp0(slpc);
755 if (unlikely(ret)) {
756 i915_probe_error(i915, "Failed to set SLPC max to RP0 (%pe)\n",
757 ERR_PTR(ret));
758 return ret;
759 }
760
761 /* Revert SLPC min/max to softlimits if necessary */
762 ret = slpc_set_softlimits(slpc);
763 if (unlikely(ret)) {
764 i915_probe_error(i915, "Failed to set SLPC softlimits (%pe)\n",
765 ERR_PTR(ret));
766 return ret;
767 }
768
769 /* Set cached media freq ratio mode */
770 intel_guc_slpc_set_media_ratio_mode(slpc, slpc->media_ratio_mode);
771
772 return 0;
773 }
774
intel_guc_slpc_set_boost_freq(struct intel_guc_slpc * slpc,u32 val)775 int intel_guc_slpc_set_boost_freq(struct intel_guc_slpc *slpc, u32 val)
776 {
777 int ret = 0;
778
779 if (val < slpc->min_freq || val > slpc->rp0_freq)
780 return -EINVAL;
781
782 mutex_lock(&slpc->lock);
783
784 if (slpc->boost_freq != val) {
785 /* Apply only if there are active waiters */
786 if (atomic_read(&slpc->num_waiters)) {
787 ret = slpc_force_min_freq(slpc, val);
788 if (ret) {
789 ret = -EIO;
790 goto done;
791 }
792 }
793
794 slpc->boost_freq = val;
795 }
796
797 done:
798 mutex_unlock(&slpc->lock);
799 return ret;
800 }
801
intel_guc_slpc_dec_waiters(struct intel_guc_slpc * slpc)802 void intel_guc_slpc_dec_waiters(struct intel_guc_slpc *slpc)
803 {
804 /*
805 * Return min back to the softlimit.
806 * This is called during request retire,
807 * so we don't need to fail that if the
808 * set_param fails.
809 */
810 mutex_lock(&slpc->lock);
811 if (atomic_dec_and_test(&slpc->num_waiters))
812 slpc_force_min_freq(slpc, slpc->min_freq_softlimit);
813 mutex_unlock(&slpc->lock);
814 }
815
intel_guc_slpc_print_info(struct intel_guc_slpc * slpc,struct drm_printer * p)816 int intel_guc_slpc_print_info(struct intel_guc_slpc *slpc, struct drm_printer *p)
817 {
818 struct drm_i915_private *i915 = slpc_to_i915(slpc);
819 struct slpc_shared_data *data = slpc->vaddr;
820 struct slpc_task_state_data *slpc_tasks;
821 intel_wakeref_t wakeref;
822 int ret = 0;
823
824 GEM_BUG_ON(!slpc->vma);
825
826 with_intel_runtime_pm(&i915->runtime_pm, wakeref) {
827 ret = slpc_query_task_state(slpc);
828
829 if (!ret) {
830 slpc_tasks = &data->task_state_data;
831
832 drm_printf(p, "\tSLPC state: %s\n", slpc_get_state_string(slpc));
833 drm_printf(p, "\tGTPERF task active: %s\n",
834 str_yes_no(slpc_tasks->status & SLPC_GTPERF_TASK_ENABLED));
835 drm_printf(p, "\tMax freq: %u MHz\n",
836 slpc_decode_max_freq(slpc));
837 drm_printf(p, "\tMin freq: %u MHz\n",
838 slpc_decode_min_freq(slpc));
839 drm_printf(p, "\twaitboosts: %u\n",
840 slpc->num_boosts);
841 }
842 }
843
844 return ret;
845 }
846
intel_guc_slpc_fini(struct intel_guc_slpc * slpc)847 void intel_guc_slpc_fini(struct intel_guc_slpc *slpc)
848 {
849 if (!slpc->vma)
850 return;
851
852 i915_vma_unpin_and_release(&slpc->vma, I915_VMA_RELEASE_MAP);
853 }
854