1 // SPDX-License-Identifier: MIT 2 /* 3 * Copyright © 2021 Intel Corporation 4 */ 5 6 #include <drm/drm_cache.h> 7 #include <linux/string_helpers.h> 8 9 #include "i915_drv.h" 10 #include "i915_reg.h" 11 #include "intel_guc_slpc.h" 12 #include "intel_guc_print.h" 13 #include "intel_mchbar_regs.h" 14 #include "gt/intel_gt.h" 15 #include "gt/intel_gt_regs.h" 16 #include "gt/intel_rps.h" 17 18 static inline struct intel_guc *slpc_to_guc(struct intel_guc_slpc *slpc) 19 { 20 return container_of(slpc, struct intel_guc, slpc); 21 } 22 23 static inline struct intel_gt *slpc_to_gt(struct intel_guc_slpc *slpc) 24 { 25 return guc_to_gt(slpc_to_guc(slpc)); 26 } 27 28 static inline struct drm_i915_private *slpc_to_i915(struct intel_guc_slpc *slpc) 29 { 30 return slpc_to_gt(slpc)->i915; 31 } 32 33 static bool __detect_slpc_supported(struct intel_guc *guc) 34 { 35 /* GuC SLPC is unavailable for pre-Gen12 */ 36 return guc->submission_supported && 37 GRAPHICS_VER(guc_to_i915(guc)) >= 12; 38 } 39 40 static bool __guc_slpc_selected(struct intel_guc *guc) 41 { 42 if (!intel_guc_slpc_is_supported(guc)) 43 return false; 44 45 return guc->submission_selected; 46 } 47 48 void intel_guc_slpc_init_early(struct intel_guc_slpc *slpc) 49 { 50 struct intel_guc *guc = slpc_to_guc(slpc); 51 52 slpc->supported = __detect_slpc_supported(guc); 53 slpc->selected = __guc_slpc_selected(guc); 54 } 55 56 static void slpc_mem_set_param(struct slpc_shared_data *data, 57 u32 id, u32 value) 58 { 59 GEM_BUG_ON(id >= SLPC_MAX_OVERRIDE_PARAMETERS); 60 /* 61 * When the flag bit is set, corresponding value will be read 62 * and applied by SLPC. 63 */ 64 data->override_params.bits[id >> 5] |= (1 << (id % 32)); 65 data->override_params.values[id] = value; 66 } 67 68 static void slpc_mem_set_enabled(struct slpc_shared_data *data, 69 u8 enable_id, u8 disable_id) 70 { 71 /* 72 * Enabling a param involves setting the enable_id 73 * to 1 and disable_id to 0. 74 */ 75 slpc_mem_set_param(data, enable_id, 1); 76 slpc_mem_set_param(data, disable_id, 0); 77 } 78 79 static void slpc_mem_set_disabled(struct slpc_shared_data *data, 80 u8 enable_id, u8 disable_id) 81 { 82 /* 83 * Disabling a param involves setting the enable_id 84 * to 0 and disable_id to 1. 85 */ 86 slpc_mem_set_param(data, disable_id, 1); 87 slpc_mem_set_param(data, enable_id, 0); 88 } 89 90 static u32 slpc_get_state(struct intel_guc_slpc *slpc) 91 { 92 struct slpc_shared_data *data; 93 94 GEM_BUG_ON(!slpc->vma); 95 96 drm_clflush_virt_range(slpc->vaddr, sizeof(u32)); 97 data = slpc->vaddr; 98 99 return data->header.global_state; 100 } 101 102 static int guc_action_slpc_set_param_nb(struct intel_guc *guc, u8 id, u32 value) 103 { 104 u32 request[] = { 105 GUC_ACTION_HOST2GUC_PC_SLPC_REQUEST, 106 SLPC_EVENT(SLPC_EVENT_PARAMETER_SET, 2), 107 id, 108 value, 109 }; 110 int ret; 111 112 ret = intel_guc_send_nb(guc, request, ARRAY_SIZE(request), 0); 113 114 return ret > 0 ? -EPROTO : ret; 115 } 116 117 static int slpc_set_param_nb(struct intel_guc_slpc *slpc, u8 id, u32 value) 118 { 119 struct intel_guc *guc = slpc_to_guc(slpc); 120 121 GEM_BUG_ON(id >= SLPC_MAX_PARAM); 122 123 return guc_action_slpc_set_param_nb(guc, id, value); 124 } 125 126 static int guc_action_slpc_set_param(struct intel_guc *guc, u8 id, u32 value) 127 { 128 u32 request[] = { 129 GUC_ACTION_HOST2GUC_PC_SLPC_REQUEST, 130 SLPC_EVENT(SLPC_EVENT_PARAMETER_SET, 2), 131 id, 132 value, 133 }; 134 int ret; 135 136 ret = intel_guc_send(guc, request, ARRAY_SIZE(request)); 137 138 return ret > 0 ? -EPROTO : ret; 139 } 140 141 static bool slpc_is_running(struct intel_guc_slpc *slpc) 142 { 143 return slpc_get_state(slpc) == SLPC_GLOBAL_STATE_RUNNING; 144 } 145 146 static int guc_action_slpc_query(struct intel_guc *guc, u32 offset) 147 { 148 u32 request[] = { 149 GUC_ACTION_HOST2GUC_PC_SLPC_REQUEST, 150 SLPC_EVENT(SLPC_EVENT_QUERY_TASK_STATE, 2), 151 offset, 152 0, 153 }; 154 int ret; 155 156 ret = intel_guc_send(guc, request, ARRAY_SIZE(request)); 157 158 return ret > 0 ? -EPROTO : ret; 159 } 160 161 static int slpc_query_task_state(struct intel_guc_slpc *slpc) 162 { 163 struct intel_guc *guc = slpc_to_guc(slpc); 164 u32 offset = intel_guc_ggtt_offset(guc, slpc->vma); 165 int ret; 166 167 ret = guc_action_slpc_query(guc, offset); 168 if (unlikely(ret)) 169 guc_probe_error(guc, "Failed to query task state: %pe\n", ERR_PTR(ret)); 170 171 drm_clflush_virt_range(slpc->vaddr, SLPC_PAGE_SIZE_BYTES); 172 173 return ret; 174 } 175 176 static int slpc_set_param(struct intel_guc_slpc *slpc, u8 id, u32 value) 177 { 178 struct intel_guc *guc = slpc_to_guc(slpc); 179 int ret; 180 181 GEM_BUG_ON(id >= SLPC_MAX_PARAM); 182 183 ret = guc_action_slpc_set_param(guc, id, value); 184 if (ret) 185 guc_probe_error(guc, "Failed to set param %d to %u: %pe\n", 186 id, value, ERR_PTR(ret)); 187 188 return ret; 189 } 190 191 static int slpc_force_min_freq(struct intel_guc_slpc *slpc, u32 freq) 192 { 193 struct intel_guc *guc = slpc_to_guc(slpc); 194 struct drm_i915_private *i915 = slpc_to_i915(slpc); 195 intel_wakeref_t wakeref; 196 int ret = 0; 197 198 lockdep_assert_held(&slpc->lock); 199 200 if (!intel_guc_is_ready(guc)) 201 return -ENODEV; 202 203 /* 204 * This function is a little different as compared to 205 * intel_guc_slpc_set_min_freq(). Softlimit will not be updated 206 * here since this is used to temporarily change min freq, 207 * for example, during a waitboost. Caller is responsible for 208 * checking bounds. 209 */ 210 211 with_intel_runtime_pm(&i915->runtime_pm, wakeref) { 212 /* Non-blocking request will avoid stalls */ 213 ret = slpc_set_param_nb(slpc, 214 SLPC_PARAM_GLOBAL_MIN_GT_UNSLICE_FREQ_MHZ, 215 freq); 216 if (ret) 217 guc_notice(guc, "Failed to send set_param for min freq(%d): %pe\n", 218 freq, ERR_PTR(ret)); 219 } 220 221 return ret; 222 } 223 224 static void slpc_boost_work(struct work_struct *work) 225 { 226 struct intel_guc_slpc *slpc = container_of(work, typeof(*slpc), boost_work); 227 int err; 228 229 /* 230 * Raise min freq to boost. It's possible that 231 * this is greater than current max. But it will 232 * certainly be limited by RP0. An error setting 233 * the min param is not fatal. 234 */ 235 mutex_lock(&slpc->lock); 236 if (atomic_read(&slpc->num_waiters)) { 237 err = slpc_force_min_freq(slpc, slpc->boost_freq); 238 if (!err) 239 slpc->num_boosts++; 240 } 241 mutex_unlock(&slpc->lock); 242 } 243 244 int intel_guc_slpc_init(struct intel_guc_slpc *slpc) 245 { 246 struct intel_guc *guc = slpc_to_guc(slpc); 247 u32 size = PAGE_ALIGN(sizeof(struct slpc_shared_data)); 248 int err; 249 250 GEM_BUG_ON(slpc->vma); 251 252 err = intel_guc_allocate_and_map_vma(guc, size, &slpc->vma, (void **)&slpc->vaddr); 253 if (unlikely(err)) { 254 guc_probe_error(guc, "Failed to allocate SLPC struct: %pe\n", ERR_PTR(err)); 255 return err; 256 } 257 258 slpc->max_freq_softlimit = 0; 259 slpc->min_freq_softlimit = 0; 260 slpc->ignore_eff_freq = false; 261 slpc->min_is_rpmax = false; 262 263 slpc->boost_freq = 0; 264 atomic_set(&slpc->num_waiters, 0); 265 slpc->num_boosts = 0; 266 slpc->media_ratio_mode = SLPC_MEDIA_RATIO_MODE_DYNAMIC_CONTROL; 267 268 mutex_init(&slpc->lock); 269 INIT_WORK(&slpc->boost_work, slpc_boost_work); 270 271 return err; 272 } 273 274 static const char *slpc_global_state_to_string(enum slpc_global_state state) 275 { 276 switch (state) { 277 case SLPC_GLOBAL_STATE_NOT_RUNNING: 278 return "not running"; 279 case SLPC_GLOBAL_STATE_INITIALIZING: 280 return "initializing"; 281 case SLPC_GLOBAL_STATE_RESETTING: 282 return "resetting"; 283 case SLPC_GLOBAL_STATE_RUNNING: 284 return "running"; 285 case SLPC_GLOBAL_STATE_SHUTTING_DOWN: 286 return "shutting down"; 287 case SLPC_GLOBAL_STATE_ERROR: 288 return "error"; 289 default: 290 return "unknown"; 291 } 292 } 293 294 static const char *slpc_get_state_string(struct intel_guc_slpc *slpc) 295 { 296 return slpc_global_state_to_string(slpc_get_state(slpc)); 297 } 298 299 static int guc_action_slpc_reset(struct intel_guc *guc, u32 offset) 300 { 301 u32 request[] = { 302 GUC_ACTION_HOST2GUC_PC_SLPC_REQUEST, 303 SLPC_EVENT(SLPC_EVENT_RESET, 2), 304 offset, 305 0, 306 }; 307 int ret; 308 309 ret = intel_guc_send(guc, request, ARRAY_SIZE(request)); 310 311 return ret > 0 ? -EPROTO : ret; 312 } 313 314 static int slpc_reset(struct intel_guc_slpc *slpc) 315 { 316 struct intel_guc *guc = slpc_to_guc(slpc); 317 u32 offset = intel_guc_ggtt_offset(guc, slpc->vma); 318 int ret; 319 320 ret = guc_action_slpc_reset(guc, offset); 321 322 if (unlikely(ret < 0)) { 323 guc_probe_error(guc, "SLPC reset action failed: %pe\n", ERR_PTR(ret)); 324 return ret; 325 } 326 327 if (!ret) { 328 if (wait_for(slpc_is_running(slpc), SLPC_RESET_TIMEOUT_MS)) { 329 guc_probe_error(guc, "SLPC not enabled! State = %s\n", 330 slpc_get_state_string(slpc)); 331 return -EIO; 332 } 333 } 334 335 return 0; 336 } 337 338 static u32 slpc_decode_min_freq(struct intel_guc_slpc *slpc) 339 { 340 struct slpc_shared_data *data = slpc->vaddr; 341 342 GEM_BUG_ON(!slpc->vma); 343 344 return DIV_ROUND_CLOSEST(REG_FIELD_GET(SLPC_MIN_UNSLICE_FREQ_MASK, 345 data->task_state_data.freq) * 346 GT_FREQUENCY_MULTIPLIER, GEN9_FREQ_SCALER); 347 } 348 349 static u32 slpc_decode_max_freq(struct intel_guc_slpc *slpc) 350 { 351 struct slpc_shared_data *data = slpc->vaddr; 352 353 GEM_BUG_ON(!slpc->vma); 354 355 return DIV_ROUND_CLOSEST(REG_FIELD_GET(SLPC_MAX_UNSLICE_FREQ_MASK, 356 data->task_state_data.freq) * 357 GT_FREQUENCY_MULTIPLIER, GEN9_FREQ_SCALER); 358 } 359 360 static void slpc_shared_data_reset(struct intel_guc_slpc *slpc) 361 { 362 struct drm_i915_private *i915 = slpc_to_i915(slpc); 363 struct slpc_shared_data *data = slpc->vaddr; 364 365 memset(data, 0, sizeof(struct slpc_shared_data)); 366 data->header.size = sizeof(struct slpc_shared_data); 367 368 /* Enable only GTPERF task, disable others */ 369 slpc_mem_set_enabled(data, SLPC_PARAM_TASK_ENABLE_GTPERF, 370 SLPC_PARAM_TASK_DISABLE_GTPERF); 371 372 /* 373 * Don't allow balancer related algorithms on platforms before 374 * Xe_LPG, where GuC started to restrict it to TDP limited scenarios. 375 */ 376 if (GRAPHICS_VER_FULL(i915) < IP_VER(12, 70)) { 377 slpc_mem_set_disabled(data, SLPC_PARAM_TASK_ENABLE_BALANCER, 378 SLPC_PARAM_TASK_DISABLE_BALANCER); 379 380 slpc_mem_set_disabled(data, SLPC_PARAM_TASK_ENABLE_DCC, 381 SLPC_PARAM_TASK_DISABLE_DCC); 382 } 383 } 384 385 /** 386 * intel_guc_slpc_set_max_freq() - Set max frequency limit for SLPC. 387 * @slpc: pointer to intel_guc_slpc. 388 * @val: frequency (MHz) 389 * 390 * This function will invoke GuC SLPC action to update the max frequency 391 * limit for unslice. 392 * 393 * Return: 0 on success, non-zero error code on failure. 394 */ 395 int intel_guc_slpc_set_max_freq(struct intel_guc_slpc *slpc, u32 val) 396 { 397 struct drm_i915_private *i915 = slpc_to_i915(slpc); 398 intel_wakeref_t wakeref; 399 int ret; 400 401 if (val < slpc->min_freq || 402 val > slpc->rp0_freq || 403 val < slpc->min_freq_softlimit) 404 return -EINVAL; 405 406 with_intel_runtime_pm(&i915->runtime_pm, wakeref) { 407 ret = slpc_set_param(slpc, 408 SLPC_PARAM_GLOBAL_MAX_GT_UNSLICE_FREQ_MHZ, 409 val); 410 411 /* Return standardized err code for sysfs calls */ 412 if (ret) 413 ret = -EIO; 414 } 415 416 if (!ret) 417 slpc->max_freq_softlimit = val; 418 419 return ret; 420 } 421 422 /** 423 * intel_guc_slpc_get_max_freq() - Get max frequency limit for SLPC. 424 * @slpc: pointer to intel_guc_slpc. 425 * @val: pointer to val which will hold max frequency (MHz) 426 * 427 * This function will invoke GuC SLPC action to read the max frequency 428 * limit for unslice. 429 * 430 * Return: 0 on success, non-zero error code on failure. 431 */ 432 int intel_guc_slpc_get_max_freq(struct intel_guc_slpc *slpc, u32 *val) 433 { 434 struct drm_i915_private *i915 = slpc_to_i915(slpc); 435 intel_wakeref_t wakeref; 436 int ret = 0; 437 438 with_intel_runtime_pm(&i915->runtime_pm, wakeref) { 439 /* Force GuC to update task data */ 440 ret = slpc_query_task_state(slpc); 441 442 if (!ret) 443 *val = slpc_decode_max_freq(slpc); 444 } 445 446 return ret; 447 } 448 449 int intel_guc_slpc_set_ignore_eff_freq(struct intel_guc_slpc *slpc, bool val) 450 { 451 struct drm_i915_private *i915 = slpc_to_i915(slpc); 452 intel_wakeref_t wakeref; 453 int ret; 454 455 mutex_lock(&slpc->lock); 456 wakeref = intel_runtime_pm_get(&i915->runtime_pm); 457 458 ret = slpc_set_param(slpc, 459 SLPC_PARAM_IGNORE_EFFICIENT_FREQUENCY, 460 val); 461 if (ret) { 462 guc_probe_error(slpc_to_guc(slpc), "Failed to set efficient freq(%d): %pe\n", 463 val, ERR_PTR(ret)); 464 } else { 465 slpc->ignore_eff_freq = val; 466 467 /* Set min to RPn when we disable efficient freq */ 468 if (val) 469 ret = slpc_set_param(slpc, 470 SLPC_PARAM_GLOBAL_MIN_GT_UNSLICE_FREQ_MHZ, 471 slpc->min_freq); 472 } 473 474 intel_runtime_pm_put(&i915->runtime_pm, wakeref); 475 mutex_unlock(&slpc->lock); 476 return ret; 477 } 478 479 /** 480 * intel_guc_slpc_set_min_freq() - Set min frequency limit for SLPC. 481 * @slpc: pointer to intel_guc_slpc. 482 * @val: frequency (MHz) 483 * 484 * This function will invoke GuC SLPC action to update the min unslice 485 * frequency. 486 * 487 * Return: 0 on success, non-zero error code on failure. 488 */ 489 int intel_guc_slpc_set_min_freq(struct intel_guc_slpc *slpc, u32 val) 490 { 491 struct drm_i915_private *i915 = slpc_to_i915(slpc); 492 intel_wakeref_t wakeref; 493 int ret; 494 495 if (val < slpc->min_freq || 496 val > slpc->rp0_freq || 497 val > slpc->max_freq_softlimit) 498 return -EINVAL; 499 500 /* Need a lock now since waitboost can be modifying min as well */ 501 mutex_lock(&slpc->lock); 502 wakeref = intel_runtime_pm_get(&i915->runtime_pm); 503 504 ret = slpc_set_param(slpc, 505 SLPC_PARAM_GLOBAL_MIN_GT_UNSLICE_FREQ_MHZ, 506 val); 507 508 if (!ret) 509 slpc->min_freq_softlimit = val; 510 511 intel_runtime_pm_put(&i915->runtime_pm, wakeref); 512 mutex_unlock(&slpc->lock); 513 514 /* Return standardized err code for sysfs calls */ 515 if (ret) 516 ret = -EIO; 517 518 return ret; 519 } 520 521 /** 522 * intel_guc_slpc_get_min_freq() - Get min frequency limit for SLPC. 523 * @slpc: pointer to intel_guc_slpc. 524 * @val: pointer to val which will hold min frequency (MHz) 525 * 526 * This function will invoke GuC SLPC action to read the min frequency 527 * limit for unslice. 528 * 529 * Return: 0 on success, non-zero error code on failure. 530 */ 531 int intel_guc_slpc_get_min_freq(struct intel_guc_slpc *slpc, u32 *val) 532 { 533 struct drm_i915_private *i915 = slpc_to_i915(slpc); 534 intel_wakeref_t wakeref; 535 int ret = 0; 536 537 with_intel_runtime_pm(&i915->runtime_pm, wakeref) { 538 /* Force GuC to update task data */ 539 ret = slpc_query_task_state(slpc); 540 541 if (!ret) 542 *val = slpc_decode_min_freq(slpc); 543 } 544 545 return ret; 546 } 547 548 int intel_guc_slpc_set_strategy(struct intel_guc_slpc *slpc, u32 val) 549 { 550 struct drm_i915_private *i915 = slpc_to_i915(slpc); 551 intel_wakeref_t wakeref; 552 int ret = 0; 553 554 with_intel_runtime_pm(&i915->runtime_pm, wakeref) 555 ret = slpc_set_param(slpc, 556 SLPC_PARAM_STRATEGIES, 557 val); 558 559 return ret; 560 } 561 562 int intel_guc_slpc_set_media_ratio_mode(struct intel_guc_slpc *slpc, u32 val) 563 { 564 struct drm_i915_private *i915 = slpc_to_i915(slpc); 565 intel_wakeref_t wakeref; 566 int ret = 0; 567 568 if (!HAS_MEDIA_RATIO_MODE(i915)) 569 return -ENODEV; 570 571 with_intel_runtime_pm(&i915->runtime_pm, wakeref) 572 ret = slpc_set_param(slpc, 573 SLPC_PARAM_MEDIA_FF_RATIO_MODE, 574 val); 575 return ret; 576 } 577 578 void intel_guc_pm_intrmsk_enable(struct intel_gt *gt) 579 { 580 u32 pm_intrmsk_mbz = 0; 581 582 /* 583 * Allow GuC to receive ARAT timer expiry event. 584 * This interrupt register is setup by RPS code 585 * when host based Turbo is enabled. 586 */ 587 pm_intrmsk_mbz |= ARAT_EXPIRED_INTRMSK; 588 589 intel_uncore_rmw(gt->uncore, 590 GEN6_PMINTRMSK, pm_intrmsk_mbz, 0); 591 } 592 593 static int slpc_set_softlimits(struct intel_guc_slpc *slpc) 594 { 595 int ret = 0; 596 597 /* 598 * Softlimits are initially equivalent to platform limits 599 * unless they have deviated from defaults, in which case, 600 * we retain the values and set min/max accordingly. 601 */ 602 if (!slpc->max_freq_softlimit) { 603 slpc->max_freq_softlimit = slpc->rp0_freq; 604 slpc_to_gt(slpc)->defaults.max_freq = slpc->max_freq_softlimit; 605 } else if (slpc->max_freq_softlimit != slpc->rp0_freq) { 606 ret = intel_guc_slpc_set_max_freq(slpc, 607 slpc->max_freq_softlimit); 608 } 609 610 if (unlikely(ret)) 611 return ret; 612 613 if (!slpc->min_freq_softlimit) { 614 /* Min softlimit is initialized to RPn */ 615 slpc->min_freq_softlimit = slpc->min_freq; 616 slpc_to_gt(slpc)->defaults.min_freq = slpc->min_freq_softlimit; 617 } else { 618 return intel_guc_slpc_set_min_freq(slpc, 619 slpc->min_freq_softlimit); 620 } 621 622 return 0; 623 } 624 625 static bool is_slpc_min_freq_rpmax(struct intel_guc_slpc *slpc) 626 { 627 int slpc_min_freq; 628 int ret; 629 630 ret = intel_guc_slpc_get_min_freq(slpc, &slpc_min_freq); 631 if (ret) { 632 guc_err(slpc_to_guc(slpc), "Failed to get min freq: %pe\n", ERR_PTR(ret)); 633 return false; 634 } 635 636 if (slpc_min_freq == SLPC_MAX_FREQ_MHZ) 637 return true; 638 else 639 return false; 640 } 641 642 static void update_server_min_softlimit(struct intel_guc_slpc *slpc) 643 { 644 /* For server parts, SLPC min will be at RPMax. 645 * Use min softlimit to clamp it to RP0 instead. 646 */ 647 if (!slpc->min_freq_softlimit && 648 is_slpc_min_freq_rpmax(slpc)) { 649 slpc->min_is_rpmax = true; 650 slpc->min_freq_softlimit = slpc->rp0_freq; 651 (slpc_to_gt(slpc))->defaults.min_freq = slpc->min_freq_softlimit; 652 } 653 } 654 655 static int slpc_use_fused_rp0(struct intel_guc_slpc *slpc) 656 { 657 /* Force SLPC to used platform rp0 */ 658 return slpc_set_param(slpc, 659 SLPC_PARAM_GLOBAL_MAX_GT_UNSLICE_FREQ_MHZ, 660 slpc->rp0_freq); 661 } 662 663 static void slpc_get_rp_values(struct intel_guc_slpc *slpc) 664 { 665 struct intel_rps *rps = &slpc_to_gt(slpc)->rps; 666 struct intel_rps_freq_caps caps; 667 668 gen6_rps_get_freq_caps(rps, &caps); 669 slpc->rp0_freq = intel_gpu_freq(rps, caps.rp0_freq); 670 slpc->rp1_freq = intel_gpu_freq(rps, caps.rp1_freq); 671 slpc->min_freq = intel_gpu_freq(rps, caps.min_freq); 672 673 if (!slpc->boost_freq) 674 slpc->boost_freq = slpc->rp0_freq; 675 } 676 677 /* 678 * intel_guc_slpc_enable() - Start SLPC 679 * @slpc: pointer to intel_guc_slpc. 680 * 681 * SLPC is enabled by setting up the shared data structure and 682 * sending reset event to GuC SLPC. Initial data is setup in 683 * intel_guc_slpc_init. Here we send the reset event. We do 684 * not currently need a slpc_disable since this is taken care 685 * of automatically when a reset/suspend occurs and the GuC 686 * CTB is destroyed. 687 * 688 * Return: 0 on success, non-zero error code on failure. 689 */ 690 int intel_guc_slpc_enable(struct intel_guc_slpc *slpc) 691 { 692 struct intel_guc *guc = slpc_to_guc(slpc); 693 int ret; 694 695 GEM_BUG_ON(!slpc->vma); 696 697 slpc_shared_data_reset(slpc); 698 699 ret = slpc_reset(slpc); 700 if (unlikely(ret < 0)) { 701 guc_probe_error(guc, "SLPC Reset event returned: %pe\n", ERR_PTR(ret)); 702 return ret; 703 } 704 705 ret = slpc_query_task_state(slpc); 706 if (unlikely(ret < 0)) 707 return ret; 708 709 intel_guc_pm_intrmsk_enable(slpc_to_gt(slpc)); 710 711 slpc_get_rp_values(slpc); 712 713 /* Handle the case where min=max=RPmax */ 714 update_server_min_softlimit(slpc); 715 716 /* Set SLPC max limit to RP0 */ 717 ret = slpc_use_fused_rp0(slpc); 718 if (unlikely(ret)) { 719 guc_probe_error(guc, "Failed to set SLPC max to RP0: %pe\n", ERR_PTR(ret)); 720 return ret; 721 } 722 723 /* Set cached value of ignore efficient freq */ 724 intel_guc_slpc_set_ignore_eff_freq(slpc, slpc->ignore_eff_freq); 725 726 /* Revert SLPC min/max to softlimits if necessary */ 727 ret = slpc_set_softlimits(slpc); 728 if (unlikely(ret)) { 729 guc_probe_error(guc, "Failed to set SLPC softlimits: %pe\n", ERR_PTR(ret)); 730 return ret; 731 } 732 733 /* Set cached media freq ratio mode */ 734 intel_guc_slpc_set_media_ratio_mode(slpc, slpc->media_ratio_mode); 735 736 /* Enable SLPC Optimized Strategy for compute */ 737 intel_guc_slpc_set_strategy(slpc, SLPC_OPTIMIZED_STRATEGY_COMPUTE); 738 739 return 0; 740 } 741 742 int intel_guc_slpc_set_boost_freq(struct intel_guc_slpc *slpc, u32 val) 743 { 744 int ret = 0; 745 746 if (val < slpc->min_freq || val > slpc->rp0_freq) 747 return -EINVAL; 748 749 mutex_lock(&slpc->lock); 750 751 if (slpc->boost_freq != val) { 752 /* Apply only if there are active waiters */ 753 if (atomic_read(&slpc->num_waiters)) { 754 ret = slpc_force_min_freq(slpc, val); 755 if (ret) { 756 ret = -EIO; 757 goto done; 758 } 759 } 760 761 slpc->boost_freq = val; 762 } 763 764 done: 765 mutex_unlock(&slpc->lock); 766 return ret; 767 } 768 769 void intel_guc_slpc_dec_waiters(struct intel_guc_slpc *slpc) 770 { 771 /* 772 * Return min back to the softlimit. 773 * This is called during request retire, 774 * so we don't need to fail that if the 775 * set_param fails. 776 */ 777 mutex_lock(&slpc->lock); 778 if (atomic_dec_and_test(&slpc->num_waiters)) 779 slpc_force_min_freq(slpc, slpc->min_freq_softlimit); 780 mutex_unlock(&slpc->lock); 781 } 782 783 int intel_guc_slpc_print_info(struct intel_guc_slpc *slpc, struct drm_printer *p) 784 { 785 struct drm_i915_private *i915 = slpc_to_i915(slpc); 786 struct slpc_shared_data *data = slpc->vaddr; 787 struct slpc_task_state_data *slpc_tasks; 788 intel_wakeref_t wakeref; 789 int ret = 0; 790 791 GEM_BUG_ON(!slpc->vma); 792 793 with_intel_runtime_pm(&i915->runtime_pm, wakeref) { 794 ret = slpc_query_task_state(slpc); 795 796 if (!ret) { 797 slpc_tasks = &data->task_state_data; 798 799 drm_printf(p, "\tSLPC state: %s\n", slpc_get_state_string(slpc)); 800 drm_printf(p, "\tGTPERF task active: %s\n", 801 str_yes_no(slpc_tasks->status & SLPC_GTPERF_TASK_ENABLED)); 802 drm_printf(p, "\tDCC enabled: %s\n", 803 str_yes_no(slpc_tasks->status & 804 SLPC_DCC_TASK_ENABLED)); 805 drm_printf(p, "\tDCC in: %s\n", 806 str_yes_no(slpc_tasks->status & SLPC_IN_DCC)); 807 drm_printf(p, "\tBalancer enabled: %s\n", 808 str_yes_no(slpc_tasks->status & 809 SLPC_BALANCER_ENABLED)); 810 drm_printf(p, "\tIBC enabled: %s\n", 811 str_yes_no(slpc_tasks->status & 812 SLPC_IBC_TASK_ENABLED)); 813 drm_printf(p, "\tBalancer IA LMT enabled: %s\n", 814 str_yes_no(slpc_tasks->status & 815 SLPC_BALANCER_IA_LMT_ENABLED)); 816 drm_printf(p, "\tBalancer IA LMT active: %s\n", 817 str_yes_no(slpc_tasks->status & 818 SLPC_BALANCER_IA_LMT_ACTIVE)); 819 drm_printf(p, "\tMax freq: %u MHz\n", 820 slpc_decode_max_freq(slpc)); 821 drm_printf(p, "\tMin freq: %u MHz\n", 822 slpc_decode_min_freq(slpc)); 823 drm_printf(p, "\twaitboosts: %u\n", 824 slpc->num_boosts); 825 drm_printf(p, "\tBoosts outstanding: %u\n", 826 atomic_read(&slpc->num_waiters)); 827 } 828 } 829 830 return ret; 831 } 832 833 void intel_guc_slpc_fini(struct intel_guc_slpc *slpc) 834 { 835 if (!slpc->vma) 836 return; 837 838 i915_vma_unpin_and_release(&slpc->vma, I915_VMA_RELEASE_MAP); 839 } 840