xref: /linux/drivers/gpu/drm/i915/gt/uc/intel_guc_slpc.c (revision fe7fad476ec8153a8b8767a08114e3e4a58a837e)
1 // SPDX-License-Identifier: MIT
2 /*
3  * Copyright © 2021 Intel Corporation
4  */
5 
6 #include <drm/drm_cache.h>
7 #include <linux/string_helpers.h>
8 
9 #include "i915_drv.h"
10 #include "i915_reg.h"
11 #include "intel_guc_slpc.h"
12 #include "intel_guc_print.h"
13 #include "intel_mchbar_regs.h"
14 #include "gt/intel_gt.h"
15 #include "gt/intel_gt_regs.h"
16 #include "gt/intel_rps.h"
17 
18 static inline struct intel_guc *slpc_to_guc(struct intel_guc_slpc *slpc)
19 {
20 	return container_of(slpc, struct intel_guc, slpc);
21 }
22 
23 static inline struct intel_gt *slpc_to_gt(struct intel_guc_slpc *slpc)
24 {
25 	return guc_to_gt(slpc_to_guc(slpc));
26 }
27 
28 static inline struct drm_i915_private *slpc_to_i915(struct intel_guc_slpc *slpc)
29 {
30 	return slpc_to_gt(slpc)->i915;
31 }
32 
33 static bool __detect_slpc_supported(struct intel_guc *guc)
34 {
35 	/* GuC SLPC is unavailable for pre-Gen12 */
36 	return guc->submission_supported &&
37 		GRAPHICS_VER(guc_to_i915(guc)) >= 12;
38 }
39 
40 static bool __guc_slpc_selected(struct intel_guc *guc)
41 {
42 	if (!intel_guc_slpc_is_supported(guc))
43 		return false;
44 
45 	return guc->submission_selected;
46 }
47 
48 void intel_guc_slpc_init_early(struct intel_guc_slpc *slpc)
49 {
50 	struct intel_guc *guc = slpc_to_guc(slpc);
51 
52 	slpc->supported = __detect_slpc_supported(guc);
53 	slpc->selected = __guc_slpc_selected(guc);
54 }
55 
56 static void slpc_mem_set_param(struct slpc_shared_data *data,
57 			       u32 id, u32 value)
58 {
59 	GEM_BUG_ON(id >= SLPC_MAX_OVERRIDE_PARAMETERS);
60 	/*
61 	 * When the flag bit is set, corresponding value will be read
62 	 * and applied by SLPC.
63 	 */
64 	data->override_params.bits[id >> 5] |= (1 << (id % 32));
65 	data->override_params.values[id] = value;
66 }
67 
68 static void slpc_mem_set_enabled(struct slpc_shared_data *data,
69 				 u8 enable_id, u8 disable_id)
70 {
71 	/*
72 	 * Enabling a param involves setting the enable_id
73 	 * to 1 and disable_id to 0.
74 	 */
75 	slpc_mem_set_param(data, enable_id, 1);
76 	slpc_mem_set_param(data, disable_id, 0);
77 }
78 
79 static void slpc_mem_set_disabled(struct slpc_shared_data *data,
80 				  u8 enable_id, u8 disable_id)
81 {
82 	/*
83 	 * Disabling a param involves setting the enable_id
84 	 * to 0 and disable_id to 1.
85 	 */
86 	slpc_mem_set_param(data, disable_id, 1);
87 	slpc_mem_set_param(data, enable_id, 0);
88 }
89 
90 static u32 slpc_get_state(struct intel_guc_slpc *slpc)
91 {
92 	struct slpc_shared_data *data;
93 
94 	GEM_BUG_ON(!slpc->vma);
95 
96 	drm_clflush_virt_range(slpc->vaddr, sizeof(u32));
97 	data = slpc->vaddr;
98 
99 	return data->header.global_state;
100 }
101 
102 static int guc_action_slpc_set_param_nb(struct intel_guc *guc, u8 id, u32 value)
103 {
104 	u32 request[] = {
105 		GUC_ACTION_HOST2GUC_PC_SLPC_REQUEST,
106 		SLPC_EVENT(SLPC_EVENT_PARAMETER_SET, 2),
107 		id,
108 		value,
109 	};
110 	int ret;
111 
112 	ret = intel_guc_send_nb(guc, request, ARRAY_SIZE(request), 0);
113 
114 	return ret > 0 ? -EPROTO : ret;
115 }
116 
117 static int slpc_set_param_nb(struct intel_guc_slpc *slpc, u8 id, u32 value)
118 {
119 	struct intel_guc *guc = slpc_to_guc(slpc);
120 
121 	GEM_BUG_ON(id >= SLPC_MAX_PARAM);
122 
123 	return guc_action_slpc_set_param_nb(guc, id, value);
124 }
125 
126 static int guc_action_slpc_set_param(struct intel_guc *guc, u8 id, u32 value)
127 {
128 	u32 request[] = {
129 		GUC_ACTION_HOST2GUC_PC_SLPC_REQUEST,
130 		SLPC_EVENT(SLPC_EVENT_PARAMETER_SET, 2),
131 		id,
132 		value,
133 	};
134 	int ret;
135 
136 	ret = intel_guc_send(guc, request, ARRAY_SIZE(request));
137 
138 	return ret > 0 ? -EPROTO : ret;
139 }
140 
141 static bool slpc_is_running(struct intel_guc_slpc *slpc)
142 {
143 	return slpc_get_state(slpc) == SLPC_GLOBAL_STATE_RUNNING;
144 }
145 
146 static int guc_action_slpc_query(struct intel_guc *guc, u32 offset)
147 {
148 	u32 request[] = {
149 		GUC_ACTION_HOST2GUC_PC_SLPC_REQUEST,
150 		SLPC_EVENT(SLPC_EVENT_QUERY_TASK_STATE, 2),
151 		offset,
152 		0,
153 	};
154 	int ret;
155 
156 	ret = intel_guc_send(guc, request, ARRAY_SIZE(request));
157 
158 	return ret > 0 ? -EPROTO : ret;
159 }
160 
161 static int slpc_query_task_state(struct intel_guc_slpc *slpc)
162 {
163 	struct intel_guc *guc = slpc_to_guc(slpc);
164 	u32 offset = intel_guc_ggtt_offset(guc, slpc->vma);
165 	int ret;
166 
167 	ret = guc_action_slpc_query(guc, offset);
168 	if (unlikely(ret))
169 		guc_probe_error(guc, "Failed to query task state: %pe\n", ERR_PTR(ret));
170 
171 	drm_clflush_virt_range(slpc->vaddr, SLPC_PAGE_SIZE_BYTES);
172 
173 	return ret;
174 }
175 
176 static int slpc_set_param(struct intel_guc_slpc *slpc, u8 id, u32 value)
177 {
178 	struct intel_guc *guc = slpc_to_guc(slpc);
179 	int ret;
180 
181 	GEM_BUG_ON(id >= SLPC_MAX_PARAM);
182 
183 	ret = guc_action_slpc_set_param(guc, id, value);
184 	if (ret)
185 		guc_probe_error(guc, "Failed to set param %d to %u: %pe\n",
186 				id, value, ERR_PTR(ret));
187 
188 	return ret;
189 }
190 
191 static int slpc_force_min_freq(struct intel_guc_slpc *slpc, u32 freq)
192 {
193 	struct intel_guc *guc = slpc_to_guc(slpc);
194 	struct drm_i915_private *i915 = slpc_to_i915(slpc);
195 	intel_wakeref_t wakeref;
196 	int ret = 0;
197 
198 	lockdep_assert_held(&slpc->lock);
199 
200 	if (!intel_guc_is_ready(guc))
201 		return -ENODEV;
202 
203 	/*
204 	 * This function is a little different as compared to
205 	 * intel_guc_slpc_set_min_freq(). Softlimit will not be updated
206 	 * here since this is used to temporarily change min freq,
207 	 * for example, during a waitboost. Caller is responsible for
208 	 * checking bounds.
209 	 */
210 
211 	with_intel_runtime_pm(&i915->runtime_pm, wakeref) {
212 		/* Non-blocking request will avoid stalls */
213 		ret = slpc_set_param_nb(slpc,
214 					SLPC_PARAM_GLOBAL_MIN_GT_UNSLICE_FREQ_MHZ,
215 					freq);
216 		if (ret)
217 			guc_notice(guc, "Failed to send set_param for min freq(%d): %pe\n",
218 				   freq, ERR_PTR(ret));
219 	}
220 
221 	return ret;
222 }
223 
224 static void slpc_boost_work(struct work_struct *work)
225 {
226 	struct intel_guc_slpc *slpc = container_of(work, typeof(*slpc), boost_work);
227 	int err;
228 
229 	/*
230 	 * Raise min freq to boost. It's possible that
231 	 * this is greater than current max. But it will
232 	 * certainly be limited by RP0. An error setting
233 	 * the min param is not fatal.
234 	 */
235 	mutex_lock(&slpc->lock);
236 	if (atomic_read(&slpc->num_waiters)) {
237 		err = slpc_force_min_freq(slpc, slpc->boost_freq);
238 		if (!err)
239 			slpc->num_boosts++;
240 	}
241 	mutex_unlock(&slpc->lock);
242 }
243 
244 int intel_guc_slpc_init(struct intel_guc_slpc *slpc)
245 {
246 	struct intel_guc *guc = slpc_to_guc(slpc);
247 	u32 size = PAGE_ALIGN(sizeof(struct slpc_shared_data));
248 	int err;
249 
250 	GEM_BUG_ON(slpc->vma);
251 
252 	err = intel_guc_allocate_and_map_vma(guc, size, &slpc->vma, (void **)&slpc->vaddr);
253 	if (unlikely(err)) {
254 		guc_probe_error(guc, "Failed to allocate SLPC struct: %pe\n", ERR_PTR(err));
255 		return err;
256 	}
257 
258 	slpc->max_freq_softlimit = 0;
259 	slpc->min_freq_softlimit = 0;
260 	slpc->ignore_eff_freq = false;
261 	slpc->min_is_rpmax = false;
262 
263 	slpc->boost_freq = 0;
264 	atomic_set(&slpc->num_waiters, 0);
265 	slpc->num_boosts = 0;
266 	slpc->media_ratio_mode = SLPC_MEDIA_RATIO_MODE_DYNAMIC_CONTROL;
267 
268 	mutex_init(&slpc->lock);
269 	INIT_WORK(&slpc->boost_work, slpc_boost_work);
270 
271 	return err;
272 }
273 
274 static const char *slpc_global_state_to_string(enum slpc_global_state state)
275 {
276 	switch (state) {
277 	case SLPC_GLOBAL_STATE_NOT_RUNNING:
278 		return "not running";
279 	case SLPC_GLOBAL_STATE_INITIALIZING:
280 		return "initializing";
281 	case SLPC_GLOBAL_STATE_RESETTING:
282 		return "resetting";
283 	case SLPC_GLOBAL_STATE_RUNNING:
284 		return "running";
285 	case SLPC_GLOBAL_STATE_SHUTTING_DOWN:
286 		return "shutting down";
287 	case SLPC_GLOBAL_STATE_ERROR:
288 		return "error";
289 	default:
290 		return "unknown";
291 	}
292 }
293 
294 static const char *slpc_get_state_string(struct intel_guc_slpc *slpc)
295 {
296 	return slpc_global_state_to_string(slpc_get_state(slpc));
297 }
298 
299 static int guc_action_slpc_reset(struct intel_guc *guc, u32 offset)
300 {
301 	u32 request[] = {
302 		GUC_ACTION_HOST2GUC_PC_SLPC_REQUEST,
303 		SLPC_EVENT(SLPC_EVENT_RESET, 2),
304 		offset,
305 		0,
306 	};
307 	int ret;
308 
309 	ret = intel_guc_send(guc, request, ARRAY_SIZE(request));
310 
311 	return ret > 0 ? -EPROTO : ret;
312 }
313 
314 static int slpc_reset(struct intel_guc_slpc *slpc)
315 {
316 	struct intel_guc *guc = slpc_to_guc(slpc);
317 	u32 offset = intel_guc_ggtt_offset(guc, slpc->vma);
318 	int ret;
319 
320 	ret = guc_action_slpc_reset(guc, offset);
321 
322 	if (unlikely(ret < 0)) {
323 		guc_probe_error(guc, "SLPC reset action failed: %pe\n", ERR_PTR(ret));
324 		return ret;
325 	}
326 
327 	if (!ret) {
328 		if (wait_for(slpc_is_running(slpc), SLPC_RESET_TIMEOUT_MS)) {
329 			guc_probe_error(guc, "SLPC not enabled! State = %s\n",
330 					slpc_get_state_string(slpc));
331 			return -EIO;
332 		}
333 	}
334 
335 	return 0;
336 }
337 
338 static u32 slpc_decode_min_freq(struct intel_guc_slpc *slpc)
339 {
340 	struct slpc_shared_data *data = slpc->vaddr;
341 
342 	GEM_BUG_ON(!slpc->vma);
343 
344 	return	DIV_ROUND_CLOSEST(REG_FIELD_GET(SLPC_MIN_UNSLICE_FREQ_MASK,
345 				  data->task_state_data.freq) *
346 				  GT_FREQUENCY_MULTIPLIER, GEN9_FREQ_SCALER);
347 }
348 
349 static u32 slpc_decode_max_freq(struct intel_guc_slpc *slpc)
350 {
351 	struct slpc_shared_data *data = slpc->vaddr;
352 
353 	GEM_BUG_ON(!slpc->vma);
354 
355 	return	DIV_ROUND_CLOSEST(REG_FIELD_GET(SLPC_MAX_UNSLICE_FREQ_MASK,
356 				  data->task_state_data.freq) *
357 				  GT_FREQUENCY_MULTIPLIER, GEN9_FREQ_SCALER);
358 }
359 
360 static void slpc_shared_data_reset(struct intel_guc_slpc *slpc)
361 {
362 	struct drm_i915_private *i915 = slpc_to_i915(slpc);
363 	struct slpc_shared_data *data = slpc->vaddr;
364 
365 	memset(data, 0, sizeof(struct slpc_shared_data));
366 	data->header.size = sizeof(struct slpc_shared_data);
367 
368 	/* Enable only GTPERF task, disable others */
369 	slpc_mem_set_enabled(data, SLPC_PARAM_TASK_ENABLE_GTPERF,
370 			     SLPC_PARAM_TASK_DISABLE_GTPERF);
371 
372 	/*
373 	 * Don't allow balancer related algorithms on platforms before
374 	 * Xe_LPG, where GuC started to restrict it to TDP limited scenarios.
375 	 */
376 	if (GRAPHICS_VER_FULL(i915) < IP_VER(12, 70)) {
377 		slpc_mem_set_disabled(data, SLPC_PARAM_TASK_ENABLE_BALANCER,
378 				      SLPC_PARAM_TASK_DISABLE_BALANCER);
379 
380 		slpc_mem_set_disabled(data, SLPC_PARAM_TASK_ENABLE_DCC,
381 				      SLPC_PARAM_TASK_DISABLE_DCC);
382 	}
383 }
384 
385 /**
386  * intel_guc_slpc_set_max_freq() - Set max frequency limit for SLPC.
387  * @slpc: pointer to intel_guc_slpc.
388  * @val: frequency (MHz)
389  *
390  * This function will invoke GuC SLPC action to update the max frequency
391  * limit for unslice.
392  *
393  * Return: 0 on success, non-zero error code on failure.
394  */
395 int intel_guc_slpc_set_max_freq(struct intel_guc_slpc *slpc, u32 val)
396 {
397 	struct drm_i915_private *i915 = slpc_to_i915(slpc);
398 	intel_wakeref_t wakeref;
399 	int ret;
400 
401 	if (val < slpc->min_freq ||
402 	    val > slpc->rp0_freq ||
403 	    val < slpc->min_freq_softlimit)
404 		return -EINVAL;
405 
406 	with_intel_runtime_pm(&i915->runtime_pm, wakeref) {
407 		ret = slpc_set_param(slpc,
408 				     SLPC_PARAM_GLOBAL_MAX_GT_UNSLICE_FREQ_MHZ,
409 				     val);
410 
411 		/* Return standardized err code for sysfs calls */
412 		if (ret)
413 			ret = -EIO;
414 	}
415 
416 	if (!ret)
417 		slpc->max_freq_softlimit = val;
418 
419 	return ret;
420 }
421 
422 /**
423  * intel_guc_slpc_get_max_freq() - Get max frequency limit for SLPC.
424  * @slpc: pointer to intel_guc_slpc.
425  * @val: pointer to val which will hold max frequency (MHz)
426  *
427  * This function will invoke GuC SLPC action to read the max frequency
428  * limit for unslice.
429  *
430  * Return: 0 on success, non-zero error code on failure.
431  */
432 int intel_guc_slpc_get_max_freq(struct intel_guc_slpc *slpc, u32 *val)
433 {
434 	struct drm_i915_private *i915 = slpc_to_i915(slpc);
435 	intel_wakeref_t wakeref;
436 	int ret = 0;
437 
438 	with_intel_runtime_pm(&i915->runtime_pm, wakeref) {
439 		/* Force GuC to update task data */
440 		ret = slpc_query_task_state(slpc);
441 
442 		if (!ret)
443 			*val = slpc_decode_max_freq(slpc);
444 	}
445 
446 	return ret;
447 }
448 
449 int intel_guc_slpc_set_ignore_eff_freq(struct intel_guc_slpc *slpc, bool val)
450 {
451 	struct drm_i915_private *i915 = slpc_to_i915(slpc);
452 	intel_wakeref_t wakeref;
453 	int ret;
454 
455 	mutex_lock(&slpc->lock);
456 	wakeref = intel_runtime_pm_get(&i915->runtime_pm);
457 
458 	ret = slpc_set_param(slpc,
459 			     SLPC_PARAM_IGNORE_EFFICIENT_FREQUENCY,
460 			     val);
461 	if (ret) {
462 		guc_probe_error(slpc_to_guc(slpc), "Failed to set efficient freq(%d): %pe\n",
463 				val, ERR_PTR(ret));
464 	} else {
465 		slpc->ignore_eff_freq = val;
466 
467 		/* Set min to RPn when we disable efficient freq */
468 		if (val)
469 			ret = slpc_set_param(slpc,
470 					     SLPC_PARAM_GLOBAL_MIN_GT_UNSLICE_FREQ_MHZ,
471 					     slpc->min_freq);
472 	}
473 
474 	intel_runtime_pm_put(&i915->runtime_pm, wakeref);
475 	mutex_unlock(&slpc->lock);
476 	return ret;
477 }
478 
479 /**
480  * intel_guc_slpc_set_min_freq() - Set min frequency limit for SLPC.
481  * @slpc: pointer to intel_guc_slpc.
482  * @val: frequency (MHz)
483  *
484  * This function will invoke GuC SLPC action to update the min unslice
485  * frequency.
486  *
487  * Return: 0 on success, non-zero error code on failure.
488  */
489 int intel_guc_slpc_set_min_freq(struct intel_guc_slpc *slpc, u32 val)
490 {
491 	struct drm_i915_private *i915 = slpc_to_i915(slpc);
492 	intel_wakeref_t wakeref;
493 	int ret;
494 
495 	if (val < slpc->min_freq ||
496 	    val > slpc->rp0_freq ||
497 	    val > slpc->max_freq_softlimit)
498 		return -EINVAL;
499 
500 	/* Need a lock now since waitboost can be modifying min as well */
501 	mutex_lock(&slpc->lock);
502 	wakeref = intel_runtime_pm_get(&i915->runtime_pm);
503 
504 	ret = slpc_set_param(slpc,
505 			     SLPC_PARAM_GLOBAL_MIN_GT_UNSLICE_FREQ_MHZ,
506 			     val);
507 
508 	if (!ret)
509 		slpc->min_freq_softlimit = val;
510 
511 	intel_runtime_pm_put(&i915->runtime_pm, wakeref);
512 	mutex_unlock(&slpc->lock);
513 
514 	/* Return standardized err code for sysfs calls */
515 	if (ret)
516 		ret = -EIO;
517 
518 	return ret;
519 }
520 
521 /**
522  * intel_guc_slpc_get_min_freq() - Get min frequency limit for SLPC.
523  * @slpc: pointer to intel_guc_slpc.
524  * @val: pointer to val which will hold min frequency (MHz)
525  *
526  * This function will invoke GuC SLPC action to read the min frequency
527  * limit for unslice.
528  *
529  * Return: 0 on success, non-zero error code on failure.
530  */
531 int intel_guc_slpc_get_min_freq(struct intel_guc_slpc *slpc, u32 *val)
532 {
533 	struct drm_i915_private *i915 = slpc_to_i915(slpc);
534 	intel_wakeref_t wakeref;
535 	int ret = 0;
536 
537 	with_intel_runtime_pm(&i915->runtime_pm, wakeref) {
538 		/* Force GuC to update task data */
539 		ret = slpc_query_task_state(slpc);
540 
541 		if (!ret)
542 			*val = slpc_decode_min_freq(slpc);
543 	}
544 
545 	return ret;
546 }
547 
548 int intel_guc_slpc_set_strategy(struct intel_guc_slpc *slpc, u32 val)
549 {
550 	struct drm_i915_private *i915 = slpc_to_i915(slpc);
551 	intel_wakeref_t wakeref;
552 	int ret = 0;
553 
554 	with_intel_runtime_pm(&i915->runtime_pm, wakeref)
555 		ret = slpc_set_param(slpc,
556 				     SLPC_PARAM_STRATEGIES,
557 				     val);
558 
559 	return ret;
560 }
561 
562 int intel_guc_slpc_set_media_ratio_mode(struct intel_guc_slpc *slpc, u32 val)
563 {
564 	struct drm_i915_private *i915 = slpc_to_i915(slpc);
565 	intel_wakeref_t wakeref;
566 	int ret = 0;
567 
568 	if (!HAS_MEDIA_RATIO_MODE(i915))
569 		return -ENODEV;
570 
571 	with_intel_runtime_pm(&i915->runtime_pm, wakeref)
572 		ret = slpc_set_param(slpc,
573 				     SLPC_PARAM_MEDIA_FF_RATIO_MODE,
574 				     val);
575 	return ret;
576 }
577 
578 void intel_guc_pm_intrmsk_enable(struct intel_gt *gt)
579 {
580 	u32 pm_intrmsk_mbz = 0;
581 
582 	/*
583 	 * Allow GuC to receive ARAT timer expiry event.
584 	 * This interrupt register is setup by RPS code
585 	 * when host based Turbo is enabled.
586 	 */
587 	pm_intrmsk_mbz |= ARAT_EXPIRED_INTRMSK;
588 
589 	intel_uncore_rmw(gt->uncore,
590 			 GEN6_PMINTRMSK, pm_intrmsk_mbz, 0);
591 }
592 
593 static int slpc_set_softlimits(struct intel_guc_slpc *slpc)
594 {
595 	int ret = 0;
596 
597 	/*
598 	 * Softlimits are initially equivalent to platform limits
599 	 * unless they have deviated from defaults, in which case,
600 	 * we retain the values and set min/max accordingly.
601 	 */
602 	if (!slpc->max_freq_softlimit) {
603 		slpc->max_freq_softlimit = slpc->rp0_freq;
604 		slpc_to_gt(slpc)->defaults.max_freq = slpc->max_freq_softlimit;
605 	} else if (slpc->max_freq_softlimit != slpc->rp0_freq) {
606 		ret = intel_guc_slpc_set_max_freq(slpc,
607 						  slpc->max_freq_softlimit);
608 	}
609 
610 	if (unlikely(ret))
611 		return ret;
612 
613 	if (!slpc->min_freq_softlimit) {
614 		/* Min softlimit is initialized to RPn */
615 		slpc->min_freq_softlimit = slpc->min_freq;
616 		slpc_to_gt(slpc)->defaults.min_freq = slpc->min_freq_softlimit;
617 	} else {
618 		return intel_guc_slpc_set_min_freq(slpc,
619 						   slpc->min_freq_softlimit);
620 	}
621 
622 	return 0;
623 }
624 
625 static bool is_slpc_min_freq_rpmax(struct intel_guc_slpc *slpc)
626 {
627 	int slpc_min_freq;
628 	int ret;
629 
630 	ret = intel_guc_slpc_get_min_freq(slpc, &slpc_min_freq);
631 	if (ret) {
632 		guc_err(slpc_to_guc(slpc), "Failed to get min freq: %pe\n", ERR_PTR(ret));
633 		return false;
634 	}
635 
636 	if (slpc_min_freq == SLPC_MAX_FREQ_MHZ)
637 		return true;
638 	else
639 		return false;
640 }
641 
642 static void update_server_min_softlimit(struct intel_guc_slpc *slpc)
643 {
644 	/* For server parts, SLPC min will be at RPMax.
645 	 * Use min softlimit to clamp it to RP0 instead.
646 	 */
647 	if (!slpc->min_freq_softlimit &&
648 	    is_slpc_min_freq_rpmax(slpc)) {
649 		slpc->min_is_rpmax = true;
650 		slpc->min_freq_softlimit = slpc->rp0_freq;
651 		(slpc_to_gt(slpc))->defaults.min_freq = slpc->min_freq_softlimit;
652 	}
653 }
654 
655 static int slpc_use_fused_rp0(struct intel_guc_slpc *slpc)
656 {
657 	/* Force SLPC to used platform rp0 */
658 	return slpc_set_param(slpc,
659 			      SLPC_PARAM_GLOBAL_MAX_GT_UNSLICE_FREQ_MHZ,
660 			      slpc->rp0_freq);
661 }
662 
663 static void slpc_get_rp_values(struct intel_guc_slpc *slpc)
664 {
665 	struct intel_rps *rps = &slpc_to_gt(slpc)->rps;
666 	struct intel_rps_freq_caps caps;
667 
668 	gen6_rps_get_freq_caps(rps, &caps);
669 	slpc->rp0_freq = intel_gpu_freq(rps, caps.rp0_freq);
670 	slpc->rp1_freq = intel_gpu_freq(rps, caps.rp1_freq);
671 	slpc->min_freq = intel_gpu_freq(rps, caps.min_freq);
672 
673 	if (!slpc->boost_freq)
674 		slpc->boost_freq = slpc->rp0_freq;
675 }
676 
677 /*
678  * intel_guc_slpc_enable() - Start SLPC
679  * @slpc: pointer to intel_guc_slpc.
680  *
681  * SLPC is enabled by setting up the shared data structure and
682  * sending reset event to GuC SLPC. Initial data is setup in
683  * intel_guc_slpc_init. Here we send the reset event. We do
684  * not currently need a slpc_disable since this is taken care
685  * of automatically when a reset/suspend occurs and the GuC
686  * CTB is destroyed.
687  *
688  * Return: 0 on success, non-zero error code on failure.
689  */
690 int intel_guc_slpc_enable(struct intel_guc_slpc *slpc)
691 {
692 	struct intel_guc *guc = slpc_to_guc(slpc);
693 	int ret;
694 
695 	GEM_BUG_ON(!slpc->vma);
696 
697 	slpc_shared_data_reset(slpc);
698 
699 	ret = slpc_reset(slpc);
700 	if (unlikely(ret < 0)) {
701 		guc_probe_error(guc, "SLPC Reset event returned: %pe\n", ERR_PTR(ret));
702 		return ret;
703 	}
704 
705 	ret = slpc_query_task_state(slpc);
706 	if (unlikely(ret < 0))
707 		return ret;
708 
709 	intel_guc_pm_intrmsk_enable(slpc_to_gt(slpc));
710 
711 	slpc_get_rp_values(slpc);
712 
713 	/* Handle the case where min=max=RPmax */
714 	update_server_min_softlimit(slpc);
715 
716 	/* Set SLPC max limit to RP0 */
717 	ret = slpc_use_fused_rp0(slpc);
718 	if (unlikely(ret)) {
719 		guc_probe_error(guc, "Failed to set SLPC max to RP0: %pe\n", ERR_PTR(ret));
720 		return ret;
721 	}
722 
723 	/* Set cached value of ignore efficient freq */
724 	intel_guc_slpc_set_ignore_eff_freq(slpc, slpc->ignore_eff_freq);
725 
726 	/* Revert SLPC min/max to softlimits if necessary */
727 	ret = slpc_set_softlimits(slpc);
728 	if (unlikely(ret)) {
729 		guc_probe_error(guc, "Failed to set SLPC softlimits: %pe\n", ERR_PTR(ret));
730 		return ret;
731 	}
732 
733 	/* Set cached media freq ratio mode */
734 	intel_guc_slpc_set_media_ratio_mode(slpc, slpc->media_ratio_mode);
735 
736 	/* Enable SLPC Optimized Strategy for compute */
737 	intel_guc_slpc_set_strategy(slpc, SLPC_OPTIMIZED_STRATEGY_COMPUTE);
738 
739 	return 0;
740 }
741 
742 int intel_guc_slpc_set_boost_freq(struct intel_guc_slpc *slpc, u32 val)
743 {
744 	int ret = 0;
745 
746 	if (val < slpc->min_freq || val > slpc->rp0_freq)
747 		return -EINVAL;
748 
749 	mutex_lock(&slpc->lock);
750 
751 	if (slpc->boost_freq != val) {
752 		/* Apply only if there are active waiters */
753 		if (atomic_read(&slpc->num_waiters)) {
754 			ret = slpc_force_min_freq(slpc, val);
755 			if (ret) {
756 				ret = -EIO;
757 				goto done;
758 			}
759 		}
760 
761 		slpc->boost_freq = val;
762 	}
763 
764 done:
765 	mutex_unlock(&slpc->lock);
766 	return ret;
767 }
768 
769 void intel_guc_slpc_dec_waiters(struct intel_guc_slpc *slpc)
770 {
771 	/*
772 	 * Return min back to the softlimit.
773 	 * This is called during request retire,
774 	 * so we don't need to fail that if the
775 	 * set_param fails.
776 	 */
777 	mutex_lock(&slpc->lock);
778 	if (atomic_dec_and_test(&slpc->num_waiters))
779 		slpc_force_min_freq(slpc, slpc->min_freq_softlimit);
780 	mutex_unlock(&slpc->lock);
781 }
782 
783 int intel_guc_slpc_print_info(struct intel_guc_slpc *slpc, struct drm_printer *p)
784 {
785 	struct drm_i915_private *i915 = slpc_to_i915(slpc);
786 	struct slpc_shared_data *data = slpc->vaddr;
787 	struct slpc_task_state_data *slpc_tasks;
788 	intel_wakeref_t wakeref;
789 	int ret = 0;
790 
791 	GEM_BUG_ON(!slpc->vma);
792 
793 	with_intel_runtime_pm(&i915->runtime_pm, wakeref) {
794 		ret = slpc_query_task_state(slpc);
795 
796 		if (!ret) {
797 			slpc_tasks = &data->task_state_data;
798 
799 			drm_printf(p, "\tSLPC state: %s\n", slpc_get_state_string(slpc));
800 			drm_printf(p, "\tGTPERF task active: %s\n",
801 				   str_yes_no(slpc_tasks->status & SLPC_GTPERF_TASK_ENABLED));
802 			drm_printf(p, "\tDCC enabled: %s\n",
803 				   str_yes_no(slpc_tasks->status &
804 					      SLPC_DCC_TASK_ENABLED));
805 			drm_printf(p, "\tDCC in: %s\n",
806 				   str_yes_no(slpc_tasks->status & SLPC_IN_DCC));
807 			drm_printf(p, "\tBalancer enabled: %s\n",
808 				   str_yes_no(slpc_tasks->status &
809 					      SLPC_BALANCER_ENABLED));
810 			drm_printf(p, "\tIBC enabled: %s\n",
811 				   str_yes_no(slpc_tasks->status &
812 					      SLPC_IBC_TASK_ENABLED));
813 			drm_printf(p, "\tBalancer IA LMT enabled: %s\n",
814 				   str_yes_no(slpc_tasks->status &
815 					      SLPC_BALANCER_IA_LMT_ENABLED));
816 			drm_printf(p, "\tBalancer IA LMT active: %s\n",
817 				   str_yes_no(slpc_tasks->status &
818 					      SLPC_BALANCER_IA_LMT_ACTIVE));
819 			drm_printf(p, "\tMax freq: %u MHz\n",
820 				   slpc_decode_max_freq(slpc));
821 			drm_printf(p, "\tMin freq: %u MHz\n",
822 				   slpc_decode_min_freq(slpc));
823 			drm_printf(p, "\twaitboosts: %u\n",
824 				   slpc->num_boosts);
825 			drm_printf(p, "\tBoosts outstanding: %u\n",
826 				   atomic_read(&slpc->num_waiters));
827 		}
828 	}
829 
830 	return ret;
831 }
832 
833 void intel_guc_slpc_fini(struct intel_guc_slpc *slpc)
834 {
835 	if (!slpc->vma)
836 		return;
837 
838 	i915_vma_unpin_and_release(&slpc->vma, I915_VMA_RELEASE_MAP);
839 }
840