1 // SPDX-License-Identifier: MIT 2 /* 3 * Copyright © 2020 Intel Corporation 4 * 5 */ 6 7 #include <drm/drm_print.h> 8 9 #include "intel_crtc.h" 10 #include "intel_de.h" 11 #include "intel_display_regs.h" 12 #include "intel_display_types.h" 13 #include "intel_dmc.h" 14 #include "intel_dmc_regs.h" 15 #include "intel_dp.h" 16 #include "intel_psr.h" 17 #include "intel_vrr.h" 18 #include "intel_vrr_regs.h" 19 #include "skl_prefill.h" 20 #include "skl_watermark.h" 21 22 #define FIXED_POINT_PRECISION 100 23 #define CMRR_PRECISION_TOLERANCE 10 24 25 /* 26 * Tunable parameters for DC Balance correction. 27 * These are captured based on experimentations. 28 */ 29 #define DCB_CORRECTION_SENSITIVITY 30 30 #define DCB_CORRECTION_AGGRESSIVENESS 1000 /* ms × 100; 10 ms */ 31 #define DCB_BLANK_TARGET 50 32 33 bool intel_vrr_is_capable(struct intel_connector *connector) 34 { 35 struct intel_display *display = to_intel_display(connector); 36 const struct drm_display_info *info = &connector->base.display_info; 37 struct intel_dp *intel_dp; 38 39 if (!HAS_VRR(display)) 40 return false; 41 42 /* 43 * DP Sink is capable of VRR video timings if 44 * Ignore MSA bit is set in DPCD. 45 * EDID monitor range also should be atleast 10 for reasonable 46 * Adaptive Sync or Variable Refresh Rate end user experience. 47 */ 48 switch (connector->base.connector_type) { 49 case DRM_MODE_CONNECTOR_eDP: 50 if (!connector->panel.vbt.vrr) 51 return false; 52 fallthrough; 53 case DRM_MODE_CONNECTOR_DisplayPort: 54 if (connector->mst.dp) 55 return false; 56 intel_dp = intel_attached_dp(connector); 57 58 if (!drm_dp_sink_can_do_video_without_timing_msa(intel_dp->dpcd)) 59 return false; 60 61 break; 62 default: 63 return false; 64 } 65 66 return info->monitor_range.max_vfreq - info->monitor_range.min_vfreq > 10; 67 } 68 69 bool intel_vrr_is_in_range(struct intel_connector *connector, int vrefresh) 70 { 71 const struct drm_display_info *info = &connector->base.display_info; 72 73 return intel_vrr_is_capable(connector) && 74 vrefresh >= info->monitor_range.min_vfreq && 75 vrefresh <= info->monitor_range.max_vfreq; 76 } 77 78 bool intel_vrr_possible(const struct intel_crtc_state *crtc_state) 79 { 80 return crtc_state->vrr.flipline; 81 } 82 83 void 84 intel_vrr_check_modeset(struct intel_atomic_state *state) 85 { 86 int i; 87 struct intel_crtc_state *old_crtc_state, *new_crtc_state; 88 struct intel_crtc *crtc; 89 90 for_each_oldnew_intel_crtc_in_state(state, crtc, old_crtc_state, 91 new_crtc_state, i) { 92 if (new_crtc_state->uapi.vrr_enabled != 93 old_crtc_state->uapi.vrr_enabled) 94 new_crtc_state->uapi.mode_changed = true; 95 } 96 } 97 98 static int intel_vrr_extra_vblank_delay(struct intel_display *display) 99 { 100 /* 101 * On ICL/TGL VRR hardware inserts one extra scanline 102 * just after vactive, which pushes the vmin decision 103 * boundary ahead accordingly, and thus reduces the 104 * max guardband length by one scanline. 105 */ 106 return DISPLAY_VER(display) < 13 ? 1 : 0; 107 } 108 109 static int intel_vrr_vmin_flipline_offset(struct intel_display *display) 110 { 111 /* 112 * ICL/TGL hardware imposes flipline>=vmin+1 113 * 114 * We reduce the vmin value to compensate when programming the 115 * hardware. This approach allows flipline to remain set at the 116 * original value, and thus the frame will have the desired 117 * minimum vtotal. 118 */ 119 return DISPLAY_VER(display) < 13 ? 1 : 0; 120 } 121 122 static int intel_vrr_guardband_to_pipeline_full(const struct intel_crtc_state *crtc_state, 123 int guardband) 124 { 125 /* hardware imposes one extra scanline somewhere */ 126 return guardband - crtc_state->framestart_delay - 1; 127 } 128 129 static int intel_vrr_pipeline_full_to_guardband(const struct intel_crtc_state *crtc_state, 130 int pipeline_full) 131 { 132 /* hardware imposes one extra scanline somewhere */ 133 return pipeline_full + crtc_state->framestart_delay + 1; 134 } 135 136 /* 137 * Without VRR registers get latched at: 138 * vblank_start 139 * 140 * With VRR the earliest registers can get latched is: 141 * intel_vrr_vmin_vblank_start(), which if we want to maintain 142 * the correct min vtotal is >=vblank_start+1 143 * 144 * The latest point registers can get latched is the vmax decision boundary: 145 * intel_vrr_vmax_vblank_start() 146 * 147 * Between those two points the vblank exit starts (and hence registers get 148 * latched) ASAP after a push is sent. 149 * 150 * framestart_delay is programmable 1-4. 151 */ 152 153 int intel_vrr_vmin_vtotal(const struct intel_crtc_state *crtc_state) 154 { 155 /* Min vblank actually determined by flipline */ 156 return crtc_state->vrr.vmin; 157 } 158 159 int intel_vrr_vmax_vtotal(const struct intel_crtc_state *crtc_state) 160 { 161 return crtc_state->vrr.vmax; 162 } 163 164 int intel_vrr_vmin_vblank_start(const struct intel_crtc_state *crtc_state) 165 { 166 return intel_vrr_vmin_vtotal(crtc_state) - crtc_state->vrr.guardband; 167 } 168 169 int intel_vrr_vmax_vblank_start(const struct intel_crtc_state *crtc_state) 170 { 171 return intel_vrr_vmax_vtotal(crtc_state) - crtc_state->vrr.guardband; 172 } 173 174 static bool 175 is_cmrr_frac_required(struct intel_crtc_state *crtc_state) 176 { 177 struct intel_display *display = to_intel_display(crtc_state); 178 int calculated_refresh_k, actual_refresh_k, pixel_clock_per_line; 179 struct drm_display_mode *adjusted_mode = &crtc_state->hw.adjusted_mode; 180 181 /* Avoid CMRR for now till we have VRR with fixed timings working */ 182 if (!HAS_CMRR(display) || true) 183 return false; 184 185 actual_refresh_k = 186 drm_mode_vrefresh(adjusted_mode) * FIXED_POINT_PRECISION; 187 pixel_clock_per_line = 188 adjusted_mode->crtc_clock * 1000 / adjusted_mode->crtc_htotal; 189 calculated_refresh_k = 190 pixel_clock_per_line * FIXED_POINT_PRECISION / adjusted_mode->crtc_vtotal; 191 192 if ((actual_refresh_k - calculated_refresh_k) < CMRR_PRECISION_TOLERANCE) 193 return false; 194 195 return true; 196 } 197 198 static unsigned int 199 cmrr_get_vtotal(struct intel_crtc_state *crtc_state, bool video_mode_required) 200 { 201 int multiplier_m = 1, multiplier_n = 1, vtotal, desired_refresh_rate; 202 u64 adjusted_pixel_rate; 203 struct drm_display_mode *adjusted_mode = &crtc_state->hw.adjusted_mode; 204 205 desired_refresh_rate = drm_mode_vrefresh(adjusted_mode); 206 207 if (video_mode_required) { 208 multiplier_m = 1001; 209 multiplier_n = 1000; 210 } 211 212 crtc_state->cmrr.cmrr_n = mul_u32_u32(desired_refresh_rate * adjusted_mode->crtc_htotal, 213 multiplier_n); 214 vtotal = DIV_ROUND_UP_ULL(mul_u32_u32(adjusted_mode->crtc_clock * 1000, multiplier_n), 215 crtc_state->cmrr.cmrr_n); 216 adjusted_pixel_rate = mul_u32_u32(adjusted_mode->crtc_clock * 1000, multiplier_m); 217 crtc_state->cmrr.cmrr_m = do_div(adjusted_pixel_rate, crtc_state->cmrr.cmrr_n); 218 219 return vtotal; 220 } 221 222 static 223 void intel_vrr_compute_cmrr_timings(struct intel_crtc_state *crtc_state) 224 { 225 /* 226 * TODO: Compute precise target refresh rate to determine 227 * if video_mode_required should be true. Currently set to 228 * false due to uncertainty about the precise target 229 * refresh Rate. 230 */ 231 crtc_state->vrr.vmax = cmrr_get_vtotal(crtc_state, false); 232 crtc_state->vrr.vmin = crtc_state->vrr.vmax; 233 crtc_state->vrr.flipline = crtc_state->vrr.vmin; 234 235 crtc_state->cmrr.enable = true; 236 crtc_state->mode_flags |= I915_MODE_FLAG_VRR; 237 } 238 239 static 240 void intel_vrr_compute_vrr_timings(struct intel_crtc_state *crtc_state, 241 int vmin, int vmax) 242 { 243 crtc_state->vrr.vmax = vmax; 244 crtc_state->vrr.vmin = vmin; 245 crtc_state->vrr.flipline = crtc_state->vrr.vmin; 246 247 crtc_state->vrr.enable = true; 248 crtc_state->mode_flags |= I915_MODE_FLAG_VRR; 249 } 250 251 static 252 void intel_vrr_compute_fixed_rr_timings(struct intel_crtc_state *crtc_state) 253 { 254 /* For fixed rr, vmin = vmax = flipline */ 255 crtc_state->vrr.vmax = crtc_state->hw.adjusted_mode.crtc_vtotal; 256 crtc_state->vrr.vmin = crtc_state->vrr.vmax; 257 crtc_state->vrr.flipline = crtc_state->vrr.vmin; 258 } 259 260 static int intel_vrr_hw_value(const struct intel_crtc_state *crtc_state, 261 int value) 262 { 263 struct intel_display *display = to_intel_display(crtc_state); 264 265 /* 266 * On TGL vmin/vmax/flipline also need to be 267 * adjusted by the SCL to maintain correct vtotals. 268 */ 269 if (DISPLAY_VER(display) >= 13) 270 return value; 271 else 272 return value - crtc_state->set_context_latency; 273 } 274 275 static int intel_vrr_vblank_start(const struct intel_crtc_state *crtc_state, 276 int vmin_vmax) 277 { 278 return intel_vrr_hw_value(crtc_state, vmin_vmax) - crtc_state->vrr.guardband; 279 } 280 281 /* 282 * For fixed refresh rate mode Vmin, Vmax and Flipline all are set to 283 * Vtotal value. 284 */ 285 static 286 int intel_vrr_fixed_rr_hw_vtotal(const struct intel_crtc_state *crtc_state) 287 { 288 return intel_vrr_hw_value(crtc_state, crtc_state->hw.adjusted_mode.crtc_vtotal); 289 } 290 291 static 292 int intel_vrr_fixed_rr_hw_vmax(const struct intel_crtc_state *crtc_state) 293 { 294 return intel_vrr_fixed_rr_hw_vtotal(crtc_state); 295 } 296 297 static 298 int intel_vrr_fixed_rr_hw_vmin(const struct intel_crtc_state *crtc_state) 299 { 300 struct intel_display *display = to_intel_display(crtc_state); 301 302 return intel_vrr_fixed_rr_hw_vtotal(crtc_state) - 303 intel_vrr_vmin_flipline_offset(display); 304 } 305 306 static 307 int intel_vrr_fixed_rr_hw_flipline(const struct intel_crtc_state *crtc_state) 308 { 309 return intel_vrr_fixed_rr_hw_vtotal(crtc_state); 310 } 311 312 void intel_vrr_set_fixed_rr_timings(const struct intel_crtc_state *crtc_state) 313 { 314 struct intel_display *display = to_intel_display(crtc_state); 315 enum transcoder cpu_transcoder = crtc_state->cpu_transcoder; 316 317 if (!intel_vrr_possible(crtc_state)) 318 return; 319 320 intel_de_write(display, TRANS_VRR_VMIN(display, cpu_transcoder), 321 intel_vrr_fixed_rr_hw_vmin(crtc_state) - 1); 322 intel_de_write(display, TRANS_VRR_VMAX(display, cpu_transcoder), 323 intel_vrr_fixed_rr_hw_vmax(crtc_state) - 1); 324 intel_de_write(display, TRANS_VRR_FLIPLINE(display, cpu_transcoder), 325 intel_vrr_fixed_rr_hw_flipline(crtc_state) - 1); 326 } 327 328 static 329 int intel_vrr_compute_vmin(struct intel_crtc_state *crtc_state) 330 { 331 /* 332 * To make fixed rr and vrr work seamless the guardband/pipeline full 333 * should be set such that it satisfies both the fixed and variable 334 * timings. 335 * For this set the vmin as crtc_vtotal. With this we never need to 336 * change anything to do with the guardband. 337 */ 338 return crtc_state->hw.adjusted_mode.crtc_vtotal; 339 } 340 341 static 342 int intel_vrr_compute_vmax(struct intel_connector *connector, 343 const struct drm_display_mode *adjusted_mode) 344 { 345 const struct drm_display_info *info = &connector->base.display_info; 346 int vmax; 347 348 vmax = adjusted_mode->crtc_clock * 1000 / 349 (adjusted_mode->crtc_htotal * info->monitor_range.min_vfreq); 350 vmax = max_t(int, vmax, adjusted_mode->crtc_vtotal); 351 352 return vmax; 353 } 354 355 static bool intel_vrr_dc_balance_possible(const struct intel_crtc_state *crtc_state) 356 { 357 struct intel_display *display = to_intel_display(crtc_state); 358 struct intel_crtc *crtc = to_intel_crtc(crtc_state->uapi.crtc); 359 enum pipe pipe = crtc->pipe; 360 361 /* 362 * FIXME: Currently Firmware supports DC Balancing on PIPE A 363 * and PIPE B. Account those limitation while computing DC 364 * Balance parameters. 365 */ 366 return (HAS_VRR_DC_BALANCE(display) && 367 ((pipe == PIPE_A) || (pipe == PIPE_B))); 368 } 369 370 static void 371 intel_vrr_dc_balance_compute_config(struct intel_crtc_state *crtc_state) 372 { 373 int guardband_usec, adjustment_usec; 374 struct drm_display_mode *adjusted_mode = &crtc_state->hw.adjusted_mode; 375 376 if (!intel_vrr_dc_balance_possible(crtc_state) || !crtc_state->vrr.enable) 377 return; 378 379 crtc_state->vrr.dc_balance.vmax = crtc_state->vrr.vmax; 380 crtc_state->vrr.dc_balance.vmin = crtc_state->vrr.vmin; 381 crtc_state->vrr.dc_balance.max_increase = 382 crtc_state->vrr.vmax - crtc_state->vrr.vmin; 383 crtc_state->vrr.dc_balance.max_decrease = 384 crtc_state->vrr.vmax - crtc_state->vrr.vmin; 385 crtc_state->vrr.dc_balance.guardband = 386 DIV_ROUND_UP(crtc_state->vrr.dc_balance.vmax * 387 DCB_CORRECTION_SENSITIVITY, 100); 388 guardband_usec = 389 intel_scanlines_to_usecs(adjusted_mode, 390 crtc_state->vrr.dc_balance.guardband); 391 /* 392 * The correction_aggressiveness/100 is the number of milliseconds to 393 * adjust by when the balance is at twice the guardband. 394 * guardband_slope = correction_aggressiveness / (guardband * 100) 395 */ 396 adjustment_usec = DCB_CORRECTION_AGGRESSIVENESS * 10; 397 crtc_state->vrr.dc_balance.slope = 398 DIV_ROUND_UP(adjustment_usec, guardband_usec); 399 crtc_state->vrr.dc_balance.vblank_target = 400 DIV_ROUND_UP((crtc_state->vrr.vmax - crtc_state->vrr.vmin) * 401 DCB_BLANK_TARGET, 100); 402 crtc_state->vrr.dc_balance.enable = true; 403 } 404 405 void 406 intel_vrr_compute_config(struct intel_crtc_state *crtc_state, 407 struct drm_connector_state *conn_state) 408 { 409 struct intel_display *display = to_intel_display(crtc_state); 410 struct intel_connector *connector = 411 to_intel_connector(conn_state->connector); 412 struct intel_dp *intel_dp = intel_attached_dp(connector); 413 bool is_edp = intel_dp_is_edp(intel_dp); 414 struct drm_display_mode *adjusted_mode = &crtc_state->hw.adjusted_mode; 415 int vmin, vmax; 416 417 if (!HAS_VRR(display)) 418 return; 419 420 if (adjusted_mode->flags & DRM_MODE_FLAG_INTERLACE) 421 return; 422 423 crtc_state->vrr.in_range = 424 intel_vrr_is_in_range(connector, drm_mode_vrefresh(adjusted_mode)); 425 426 /* 427 * Allow fixed refresh rate with VRR Timing Generator. 428 * For now set the vrr.in_range to 0, to allow fixed_rr but skip actual 429 * VRR and LRR. 430 * #TODO For actual VRR with joiner, we need to figure out how to 431 * correctly sequence transcoder level stuff vs. pipe level stuff 432 * in the commit. 433 */ 434 if (crtc_state->joiner_pipes) 435 crtc_state->vrr.in_range = false; 436 437 vmin = intel_vrr_compute_vmin(crtc_state); 438 439 if (crtc_state->vrr.in_range) { 440 if (HAS_LRR(display)) 441 crtc_state->update_lrr = true; 442 vmax = intel_vrr_compute_vmax(connector, adjusted_mode); 443 } else { 444 vmax = vmin; 445 } 446 447 if (crtc_state->uapi.vrr_enabled && vmin < vmax) 448 intel_vrr_compute_vrr_timings(crtc_state, vmin, vmax); 449 else if (is_cmrr_frac_required(crtc_state) && is_edp) 450 intel_vrr_compute_cmrr_timings(crtc_state); 451 else 452 intel_vrr_compute_fixed_rr_timings(crtc_state); 453 454 if (HAS_AS_SDP(display)) { 455 crtc_state->vrr.vsync_start = 456 (crtc_state->hw.adjusted_mode.crtc_vtotal - 457 crtc_state->hw.adjusted_mode.crtc_vsync_start); 458 crtc_state->vrr.vsync_end = 459 (crtc_state->hw.adjusted_mode.crtc_vtotal - 460 crtc_state->hw.adjusted_mode.crtc_vsync_end); 461 } 462 463 intel_vrr_dc_balance_compute_config(crtc_state); 464 } 465 466 static int 467 intel_vrr_max_hw_guardband(const struct intel_crtc_state *crtc_state) 468 { 469 struct intel_display *display = to_intel_display(crtc_state); 470 int max_pipeline_full = REG_FIELD_MAX(VRR_CTL_PIPELINE_FULL_MASK); 471 472 if (DISPLAY_VER(display) >= 13) 473 return REG_FIELD_MAX(XELPD_VRR_CTL_VRR_GUARDBAND_MASK); 474 else 475 return intel_vrr_pipeline_full_to_guardband(crtc_state, 476 max_pipeline_full); 477 } 478 479 static int 480 intel_vrr_max_vblank_guardband(const struct intel_crtc_state *crtc_state) 481 { 482 struct intel_display *display = to_intel_display(crtc_state); 483 const struct drm_display_mode *adjusted_mode = &crtc_state->hw.adjusted_mode; 484 485 return crtc_state->vrr.vmin - 486 adjusted_mode->crtc_vdisplay - 487 crtc_state->set_context_latency - 488 intel_vrr_extra_vblank_delay(display); 489 } 490 491 static int 492 intel_vrr_max_guardband(struct intel_crtc_state *crtc_state) 493 { 494 return min(intel_vrr_max_hw_guardband(crtc_state), 495 intel_vrr_max_vblank_guardband(crtc_state)); 496 } 497 498 static 499 int intel_vrr_compute_optimized_guardband(struct intel_crtc_state *crtc_state) 500 { 501 struct intel_display *display = to_intel_display(crtc_state); 502 struct skl_prefill_ctx prefill_ctx; 503 int prefill_latency_us; 504 int guardband = 0; 505 506 skl_prefill_init_worst(&prefill_ctx, crtc_state); 507 508 /* 509 * The SoC power controller runs SAGV mutually exclusive with package C states, 510 * so the max of package C and SAGV latencies is used to compute the min prefill guardband. 511 * PM delay = max(sagv_latency, pkgc_max_latency (highest enabled wm level 1 and up)) 512 */ 513 prefill_latency_us = max(display->sagv.block_time_us, 514 skl_watermark_max_latency(display, 1)); 515 516 guardband = skl_prefill_min_guardband(&prefill_ctx, 517 crtc_state, 518 prefill_latency_us); 519 520 if (intel_crtc_has_dp_encoder(crtc_state)) { 521 guardband = max(guardband, intel_psr_min_guardband(crtc_state)); 522 guardband = max(guardband, intel_dp_sdp_min_guardband(crtc_state, true)); 523 } 524 525 return guardband; 526 } 527 528 static bool intel_vrr_use_optimized_guardband(const struct intel_crtc_state *crtc_state) 529 { 530 /* 531 * #TODO: Enable optimized guardband for HDMI 532 * For HDMI lot of infoframes are transmitted a line or two after vsync. 533 * Since with optimized guardband the double bufferring point is at delayed vblank, 534 * we need to ensure that vsync happens after delayed vblank for the HDMI case. 535 */ 536 if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_HDMI)) 537 return false; 538 539 return true; 540 } 541 542 void intel_vrr_compute_guardband(struct intel_crtc_state *crtc_state) 543 { 544 struct intel_display *display = to_intel_display(crtc_state); 545 struct drm_display_mode *adjusted_mode = &crtc_state->hw.adjusted_mode; 546 struct drm_display_mode *pipe_mode = &crtc_state->hw.pipe_mode; 547 int guardband; 548 549 if (!intel_vrr_possible(crtc_state)) 550 return; 551 552 if (intel_vrr_use_optimized_guardband(crtc_state)) 553 guardband = intel_vrr_compute_optimized_guardband(crtc_state); 554 else 555 guardband = crtc_state->vrr.vmin - adjusted_mode->crtc_vdisplay; 556 557 crtc_state->vrr.guardband = min(guardband, intel_vrr_max_guardband(crtc_state)); 558 559 if (intel_vrr_always_use_vrr_tg(display)) { 560 adjusted_mode->crtc_vblank_start = 561 adjusted_mode->crtc_vtotal - crtc_state->vrr.guardband; 562 /* 563 * pipe_mode has already been derived from the 564 * original adjusted_mode, keep the two in sync. 565 */ 566 pipe_mode->crtc_vblank_start = 567 adjusted_mode->crtc_vblank_start; 568 } 569 570 if (DISPLAY_VER(display) < 13) 571 crtc_state->vrr.pipeline_full = 572 intel_vrr_guardband_to_pipeline_full(crtc_state, 573 crtc_state->vrr.guardband); 574 } 575 576 static u32 trans_vrr_ctl(const struct intel_crtc_state *crtc_state) 577 { 578 struct intel_display *display = to_intel_display(crtc_state); 579 580 if (DISPLAY_VER(display) >= 14) 581 return VRR_CTL_FLIP_LINE_EN | 582 XELPD_VRR_CTL_VRR_GUARDBAND(crtc_state->vrr.guardband); 583 else if (DISPLAY_VER(display) >= 13) 584 return VRR_CTL_IGN_MAX_SHIFT | VRR_CTL_FLIP_LINE_EN | 585 XELPD_VRR_CTL_VRR_GUARDBAND(crtc_state->vrr.guardband); 586 else 587 return VRR_CTL_IGN_MAX_SHIFT | VRR_CTL_FLIP_LINE_EN | 588 VRR_CTL_PIPELINE_FULL(crtc_state->vrr.pipeline_full) | 589 VRR_CTL_PIPELINE_FULL_OVERRIDE; 590 } 591 592 void intel_vrr_set_transcoder_timings(const struct intel_crtc_state *crtc_state) 593 { 594 struct intel_display *display = to_intel_display(crtc_state); 595 enum transcoder cpu_transcoder = crtc_state->cpu_transcoder; 596 597 if (!HAS_VRR(display)) 598 return; 599 600 /* 601 * Bspec says: 602 * "(note: VRR needs to be programmed after 603 * TRANS_DDI_FUNC_CTL and before TRANS_CONF)." 604 * 605 * In practice it turns out that ICL can hang if 606 * TRANS_VRR_VMAX/FLIPLINE are written before 607 * enabling TRANS_DDI_FUNC_CTL. 608 */ 609 drm_WARN_ON(display->drm, 610 !(intel_de_read(display, TRANS_DDI_FUNC_CTL(display, cpu_transcoder)) & TRANS_DDI_FUNC_ENABLE)); 611 612 /* 613 * This bit seems to have two meanings depending on the platform: 614 * TGL: generate VRR "safe window" for DSB vblank waits 615 * ADL/DG2: make TRANS_SET_CONTEXT_LATENCY effective with VRR 616 */ 617 if (IS_DISPLAY_VER(display, 12, 13)) 618 intel_de_rmw(display, CHICKEN_TRANS(display, cpu_transcoder), 619 0, PIPE_VBLANK_WITH_DELAY); 620 621 if (!intel_vrr_possible(crtc_state)) { 622 intel_de_write(display, 623 TRANS_VRR_CTL(display, cpu_transcoder), 0); 624 return; 625 } 626 627 if (crtc_state->cmrr.enable) { 628 intel_de_write(display, TRANS_CMRR_M_HI(display, cpu_transcoder), 629 upper_32_bits(crtc_state->cmrr.cmrr_m)); 630 intel_de_write(display, TRANS_CMRR_M_LO(display, cpu_transcoder), 631 lower_32_bits(crtc_state->cmrr.cmrr_m)); 632 intel_de_write(display, TRANS_CMRR_N_HI(display, cpu_transcoder), 633 upper_32_bits(crtc_state->cmrr.cmrr_n)); 634 intel_de_write(display, TRANS_CMRR_N_LO(display, cpu_transcoder), 635 lower_32_bits(crtc_state->cmrr.cmrr_n)); 636 } 637 638 intel_vrr_set_fixed_rr_timings(crtc_state); 639 640 if (!intel_vrr_always_use_vrr_tg(display)) 641 intel_de_write(display, TRANS_VRR_CTL(display, cpu_transcoder), 642 trans_vrr_ctl(crtc_state)); 643 644 if (HAS_AS_SDP(display)) 645 intel_de_write(display, 646 TRANS_VRR_VSYNC(display, cpu_transcoder), 647 VRR_VSYNC_END(crtc_state->vrr.vsync_end) | 648 VRR_VSYNC_START(crtc_state->vrr.vsync_start)); 649 650 /* 651 * For BMG and LNL+ onwards the EMP_AS_SDP_TL is used for programming 652 * double buffering point and transmission line for VRR packets for 653 * HDMI2.1/DP/eDP/DP->HDMI2.1 PCON. 654 * Since currently we support VRR only for DP/eDP, so this is programmed 655 * to for Adaptive Sync SDP to Vsync start. 656 */ 657 if (DISPLAY_VERx100(display) == 1401 || DISPLAY_VER(display) >= 20) 658 intel_de_write(display, 659 EMP_AS_SDP_TL(display, cpu_transcoder), 660 EMP_AS_SDP_DB_TL(crtc_state->vrr.vsync_start)); 661 } 662 663 void 664 intel_vrr_dcb_increment_flip_count(struct intel_crtc_state *crtc_state, 665 struct intel_crtc *crtc) 666 { 667 struct intel_display *display = to_intel_display(crtc_state); 668 enum pipe pipe = crtc->pipe; 669 670 if (!crtc_state->vrr.dc_balance.enable) 671 return; 672 673 intel_de_write(display, PIPEDMC_DCB_FLIP_COUNT(pipe), 674 ++crtc->dc_balance.flip_count); 675 } 676 677 void 678 intel_vrr_dcb_reset(const struct intel_crtc_state *old_crtc_state, 679 struct intel_crtc *crtc) 680 { 681 struct intel_display *display = to_intel_display(old_crtc_state); 682 enum pipe pipe = crtc->pipe; 683 684 if (!old_crtc_state->vrr.dc_balance.enable) 685 return; 686 687 intel_de_write(display, PIPEDMC_DCB_FLIP_COUNT(pipe), 0); 688 intel_de_write(display, PIPEDMC_DCB_BALANCE_RESET(pipe), 0); 689 } 690 691 void intel_vrr_send_push(struct intel_dsb *dsb, 692 const struct intel_crtc_state *crtc_state) 693 { 694 struct intel_display *display = to_intel_display(crtc_state); 695 enum transcoder cpu_transcoder = crtc_state->cpu_transcoder; 696 697 if (!crtc_state->vrr.enable) 698 return; 699 700 if (dsb) 701 intel_dsb_nonpost_start(dsb); 702 703 intel_de_write_dsb(display, dsb, 704 TRANS_PUSH(display, cpu_transcoder), 705 TRANS_PUSH_EN | TRANS_PUSH_SEND); 706 707 if (dsb) 708 intel_dsb_nonpost_end(dsb); 709 } 710 711 void intel_vrr_check_push_sent(struct intel_dsb *dsb, 712 const struct intel_crtc_state *crtc_state) 713 { 714 struct intel_display *display = to_intel_display(crtc_state); 715 struct intel_crtc *crtc = to_intel_crtc(crtc_state->uapi.crtc); 716 enum transcoder cpu_transcoder = crtc_state->cpu_transcoder; 717 718 if (!crtc_state->vrr.enable) 719 return; 720 721 /* 722 * Make sure the push send bit has cleared. This should 723 * already be the case as long as the caller makes sure 724 * this is called after the delayed vblank has occurred. 725 */ 726 if (dsb) { 727 int wait_us, count; 728 729 wait_us = 2; 730 count = 1; 731 732 /* 733 * If the bit hasn't cleared the DSB will 734 * raise the poll error interrupt. 735 */ 736 intel_dsb_poll(dsb, TRANS_PUSH(display, cpu_transcoder), 737 TRANS_PUSH_SEND, 0, wait_us, count); 738 } else { 739 if (intel_vrr_is_push_sent(crtc_state)) 740 drm_err(display->drm, "[CRTC:%d:%s] VRR push send still pending\n", 741 crtc->base.base.id, crtc->base.name); 742 } 743 } 744 745 bool intel_vrr_is_push_sent(const struct intel_crtc_state *crtc_state) 746 { 747 struct intel_display *display = to_intel_display(crtc_state); 748 enum transcoder cpu_transcoder = crtc_state->cpu_transcoder; 749 750 if (!crtc_state->vrr.enable) 751 return false; 752 753 return intel_de_read(display, TRANS_PUSH(display, cpu_transcoder)) & TRANS_PUSH_SEND; 754 } 755 756 bool intel_vrr_always_use_vrr_tg(struct intel_display *display) 757 { 758 if (!HAS_VRR(display)) 759 return false; 760 761 if (DISPLAY_VER(display) >= 30) 762 return true; 763 764 return false; 765 } 766 767 static int intel_vrr_hw_vmin(const struct intel_crtc_state *crtc_state) 768 { 769 struct intel_display *display = to_intel_display(crtc_state); 770 771 return intel_vrr_hw_value(crtc_state, crtc_state->vrr.vmin) - 772 intel_vrr_vmin_flipline_offset(display); 773 } 774 775 static int intel_vrr_hw_vmax(const struct intel_crtc_state *crtc_state) 776 { 777 return intel_vrr_hw_value(crtc_state, crtc_state->vrr.vmax); 778 } 779 780 static int intel_vrr_hw_flipline(const struct intel_crtc_state *crtc_state) 781 { 782 return intel_vrr_hw_value(crtc_state, crtc_state->vrr.flipline); 783 } 784 785 static void intel_vrr_set_vrr_timings(const struct intel_crtc_state *crtc_state) 786 { 787 struct intel_display *display = to_intel_display(crtc_state); 788 enum transcoder cpu_transcoder = crtc_state->cpu_transcoder; 789 790 intel_de_write(display, TRANS_VRR_VMIN(display, cpu_transcoder), 791 intel_vrr_hw_vmin(crtc_state) - 1); 792 intel_de_write(display, TRANS_VRR_VMAX(display, cpu_transcoder), 793 intel_vrr_hw_vmax(crtc_state) - 1); 794 intel_de_write(display, TRANS_VRR_FLIPLINE(display, cpu_transcoder), 795 intel_vrr_hw_flipline(crtc_state) - 1); 796 } 797 798 static void 799 intel_vrr_enable_dc_balancing(const struct intel_crtc_state *crtc_state) 800 { 801 struct intel_display *display = to_intel_display(crtc_state); 802 enum transcoder cpu_transcoder = crtc_state->cpu_transcoder; 803 struct intel_crtc *crtc = to_intel_crtc(crtc_state->uapi.crtc); 804 enum pipe pipe = crtc->pipe; 805 u32 vrr_ctl = intel_de_read(display, TRANS_VRR_CTL(display, cpu_transcoder)); 806 807 if (!crtc_state->vrr.dc_balance.enable) 808 return; 809 810 intel_de_write(display, TRANS_VRR_DCB_ADJ_VMAX_CFG(cpu_transcoder), 811 VRR_DCB_ADJ_VMAX(crtc_state->vrr.vmax - 1)); 812 intel_de_write(display, TRANS_VRR_DCB_ADJ_VMAX_CFG_LIVE(cpu_transcoder), 813 VRR_DCB_ADJ_VMAX(crtc_state->vrr.vmax - 1)); 814 intel_de_write(display, TRANS_VRR_DCB_VMAX(cpu_transcoder), 815 VRR_DCB_VMAX(crtc_state->vrr.vmax - 1)); 816 intel_de_write(display, TRANS_VRR_DCB_VMAX_LIVE(cpu_transcoder), 817 VRR_DCB_VMAX(crtc_state->vrr.vmax - 1)); 818 intel_de_write(display, TRANS_VRR_DCB_FLIPLINE(cpu_transcoder), 819 VRR_DCB_FLIPLINE(crtc_state->vrr.flipline - 1)); 820 intel_de_write(display, TRANS_VRR_DCB_FLIPLINE_LIVE(cpu_transcoder), 821 VRR_DCB_FLIPLINE(crtc_state->vrr.flipline - 1)); 822 intel_de_write(display, TRANS_VRR_DCB_ADJ_FLIPLINE_CFG_LIVE(cpu_transcoder), 823 VRR_DCB_ADJ_FLIPLINE(crtc_state->vrr.flipline - 1)); 824 intel_de_write(display, TRANS_VRR_DCB_ADJ_FLIPLINE_CFG(cpu_transcoder), 825 VRR_DCB_ADJ_FLIPLINE(crtc_state->vrr.flipline - 1)); 826 intel_de_write(display, PIPEDMC_DCB_VMIN(pipe), 827 crtc_state->vrr.dc_balance.vmin - 1); 828 intel_de_write(display, PIPEDMC_DCB_VMAX(pipe), 829 crtc_state->vrr.dc_balance.vmax - 1); 830 intel_de_write(display, PIPEDMC_DCB_MAX_INCREASE(pipe), 831 crtc_state->vrr.dc_balance.max_increase); 832 intel_de_write(display, PIPEDMC_DCB_MAX_DECREASE(pipe), 833 crtc_state->vrr.dc_balance.max_decrease); 834 intel_de_write(display, PIPEDMC_DCB_GUARDBAND(pipe), 835 crtc_state->vrr.dc_balance.guardband); 836 intel_de_write(display, PIPEDMC_DCB_SLOPE(pipe), 837 crtc_state->vrr.dc_balance.slope); 838 intel_de_write(display, PIPEDMC_DCB_VBLANK(pipe), 839 crtc_state->vrr.dc_balance.vblank_target); 840 intel_dmc_configure_dc_balance_event(display, pipe, true); 841 intel_de_write(display, TRANS_ADAPTIVE_SYNC_DCB_CTL(cpu_transcoder), 842 ADAPTIVE_SYNC_COUNTER_EN); 843 intel_pipedmc_dcb_enable(NULL, crtc); 844 845 vrr_ctl |= VRR_CTL_DCB_ADJ_ENABLE; 846 intel_de_write(display, TRANS_VRR_CTL(display, cpu_transcoder), vrr_ctl); 847 } 848 849 static void 850 intel_vrr_disable_dc_balancing(const struct intel_crtc_state *old_crtc_state) 851 { 852 struct intel_display *display = to_intel_display(old_crtc_state); 853 enum transcoder cpu_transcoder = old_crtc_state->cpu_transcoder; 854 struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->uapi.crtc); 855 enum pipe pipe = crtc->pipe; 856 u32 vrr_ctl = intel_de_read(display, TRANS_VRR_CTL(display, cpu_transcoder)); 857 858 if (!old_crtc_state->vrr.dc_balance.enable) 859 return; 860 861 intel_pipedmc_dcb_disable(NULL, crtc); 862 intel_dmc_configure_dc_balance_event(display, pipe, false); 863 intel_de_write(display, TRANS_ADAPTIVE_SYNC_DCB_CTL(cpu_transcoder), 0); 864 intel_de_write(display, PIPEDMC_DCB_VMIN(pipe), 0); 865 intel_de_write(display, PIPEDMC_DCB_VMAX(pipe), 0); 866 intel_de_write(display, PIPEDMC_DCB_MAX_INCREASE(pipe), 0); 867 intel_de_write(display, PIPEDMC_DCB_MAX_DECREASE(pipe), 0); 868 intel_de_write(display, PIPEDMC_DCB_GUARDBAND(pipe), 0); 869 intel_de_write(display, PIPEDMC_DCB_SLOPE(pipe), 0); 870 intel_de_write(display, PIPEDMC_DCB_VBLANK(pipe), 0); 871 intel_de_write(display, TRANS_VRR_DCB_ADJ_VMAX_CFG_LIVE(cpu_transcoder), 0); 872 intel_de_write(display, TRANS_VRR_DCB_ADJ_FLIPLINE_CFG_LIVE(cpu_transcoder), 0); 873 intel_de_write(display, TRANS_VRR_DCB_VMAX_LIVE(cpu_transcoder), 0); 874 intel_de_write(display, TRANS_VRR_DCB_FLIPLINE_LIVE(cpu_transcoder), 0); 875 intel_de_write(display, TRANS_VRR_DCB_ADJ_VMAX_CFG(cpu_transcoder), 0); 876 intel_de_write(display, TRANS_VRR_DCB_ADJ_FLIPLINE_CFG(cpu_transcoder), 0); 877 intel_de_write(display, TRANS_VRR_DCB_VMAX(cpu_transcoder), 0); 878 intel_de_write(display, TRANS_VRR_DCB_FLIPLINE(cpu_transcoder), 0); 879 880 vrr_ctl &= ~VRR_CTL_DCB_ADJ_ENABLE; 881 intel_de_write(display, TRANS_VRR_CTL(display, cpu_transcoder), vrr_ctl); 882 } 883 884 static void intel_vrr_tg_enable(const struct intel_crtc_state *crtc_state, 885 bool cmrr_enable) 886 { 887 struct intel_display *display = to_intel_display(crtc_state); 888 enum transcoder cpu_transcoder = crtc_state->cpu_transcoder; 889 u32 vrr_ctl; 890 891 intel_de_write(display, TRANS_PUSH(display, cpu_transcoder), TRANS_PUSH_EN); 892 893 vrr_ctl = VRR_CTL_VRR_ENABLE | trans_vrr_ctl(crtc_state); 894 895 /* 896 * FIXME this might be broken as bspec seems to imply that 897 * even VRR_CTL_CMRR_ENABLE is armed by TRANS_CMRR_N_HI 898 * when enabling CMRR (but not when disabling CMRR?). 899 */ 900 if (cmrr_enable) 901 vrr_ctl |= VRR_CTL_CMRR_ENABLE; 902 903 intel_de_write(display, TRANS_VRR_CTL(display, cpu_transcoder), vrr_ctl); 904 } 905 906 static void intel_vrr_tg_disable(const struct intel_crtc_state *old_crtc_state) 907 { 908 struct intel_display *display = to_intel_display(old_crtc_state); 909 enum transcoder cpu_transcoder = old_crtc_state->cpu_transcoder; 910 911 intel_de_write(display, TRANS_VRR_CTL(display, cpu_transcoder), 912 trans_vrr_ctl(old_crtc_state)); 913 914 if (intel_de_wait_for_clear_ms(display, 915 TRANS_VRR_STATUS(display, cpu_transcoder), 916 VRR_STATUS_VRR_EN_LIVE, 1000)) 917 drm_err(display->drm, "Timed out waiting for VRR live status to clear\n"); 918 919 intel_de_write(display, TRANS_PUSH(display, cpu_transcoder), 0); 920 } 921 922 void intel_vrr_enable(const struct intel_crtc_state *crtc_state) 923 { 924 struct intel_display *display = to_intel_display(crtc_state); 925 926 if (!crtc_state->vrr.enable) 927 return; 928 929 intel_vrr_set_vrr_timings(crtc_state); 930 intel_vrr_enable_dc_balancing(crtc_state); 931 932 if (!intel_vrr_always_use_vrr_tg(display)) 933 intel_vrr_tg_enable(crtc_state, crtc_state->cmrr.enable); 934 } 935 936 void intel_vrr_disable(const struct intel_crtc_state *old_crtc_state) 937 { 938 struct intel_display *display = to_intel_display(old_crtc_state); 939 940 if (!old_crtc_state->vrr.enable) 941 return; 942 943 if (!intel_vrr_always_use_vrr_tg(display)) 944 intel_vrr_tg_disable(old_crtc_state); 945 946 intel_vrr_disable_dc_balancing(old_crtc_state); 947 intel_vrr_set_fixed_rr_timings(old_crtc_state); 948 } 949 950 void intel_vrr_transcoder_enable(const struct intel_crtc_state *crtc_state) 951 { 952 struct intel_display *display = to_intel_display(crtc_state); 953 954 intel_vrr_set_transcoder_timings(crtc_state); 955 956 if (!intel_vrr_possible(crtc_state)) 957 return; 958 959 if (intel_vrr_always_use_vrr_tg(display)) 960 intel_vrr_tg_enable(crtc_state, false); 961 } 962 963 void intel_vrr_transcoder_disable(const struct intel_crtc_state *old_crtc_state) 964 { 965 struct intel_display *display = to_intel_display(old_crtc_state); 966 967 if (!intel_vrr_possible(old_crtc_state)) 968 return; 969 970 if (intel_vrr_always_use_vrr_tg(display)) 971 intel_vrr_tg_disable(old_crtc_state); 972 } 973 974 bool intel_vrr_is_fixed_rr(const struct intel_crtc_state *crtc_state) 975 { 976 return crtc_state->vrr.flipline && 977 crtc_state->vrr.flipline == crtc_state->vrr.vmax && 978 crtc_state->vrr.flipline == crtc_state->vrr.vmin; 979 } 980 981 static 982 void intel_vrr_get_dc_balance_config(struct intel_crtc_state *crtc_state) 983 { 984 u32 reg_val; 985 struct intel_display *display = to_intel_display(crtc_state); 986 struct intel_crtc *crtc = to_intel_crtc(crtc_state->uapi.crtc); 987 enum pipe pipe = crtc->pipe; 988 989 if (!intel_vrr_dc_balance_possible(crtc_state)) 990 return; 991 992 reg_val = intel_de_read(display, PIPEDMC_DCB_VMIN(pipe)); 993 crtc_state->vrr.dc_balance.vmin = reg_val ? reg_val + 1 : 0; 994 995 reg_val = intel_de_read(display, PIPEDMC_DCB_VMAX(pipe)); 996 crtc_state->vrr.dc_balance.vmax = reg_val ? reg_val + 1 : 0; 997 998 crtc_state->vrr.dc_balance.guardband = 999 intel_de_read(display, PIPEDMC_DCB_GUARDBAND(pipe)); 1000 crtc_state->vrr.dc_balance.max_increase = 1001 intel_de_read(display, PIPEDMC_DCB_MAX_INCREASE(pipe)); 1002 crtc_state->vrr.dc_balance.max_decrease = 1003 intel_de_read(display, PIPEDMC_DCB_MAX_DECREASE(pipe)); 1004 crtc_state->vrr.dc_balance.slope = 1005 intel_de_read(display, PIPEDMC_DCB_SLOPE(pipe)); 1006 crtc_state->vrr.dc_balance.vblank_target = 1007 intel_de_read(display, PIPEDMC_DCB_VBLANK(pipe)); 1008 } 1009 1010 void intel_vrr_get_config(struct intel_crtc_state *crtc_state) 1011 { 1012 struct intel_display *display = to_intel_display(crtc_state); 1013 enum transcoder cpu_transcoder = crtc_state->cpu_transcoder; 1014 u32 trans_vrr_ctl, trans_vrr_vsync; 1015 bool vrr_enable; 1016 1017 trans_vrr_ctl = intel_de_read(display, 1018 TRANS_VRR_CTL(display, cpu_transcoder)); 1019 1020 if (HAS_CMRR(display)) 1021 crtc_state->cmrr.enable = (trans_vrr_ctl & VRR_CTL_CMRR_ENABLE); 1022 1023 if (crtc_state->cmrr.enable) { 1024 crtc_state->cmrr.cmrr_n = 1025 intel_de_read64_2x32(display, TRANS_CMRR_N_LO(display, cpu_transcoder), 1026 TRANS_CMRR_N_HI(display, cpu_transcoder)); 1027 crtc_state->cmrr.cmrr_m = 1028 intel_de_read64_2x32(display, TRANS_CMRR_M_LO(display, cpu_transcoder), 1029 TRANS_CMRR_M_HI(display, cpu_transcoder)); 1030 } 1031 1032 if (DISPLAY_VER(display) >= 13) { 1033 crtc_state->vrr.guardband = 1034 REG_FIELD_GET(XELPD_VRR_CTL_VRR_GUARDBAND_MASK, trans_vrr_ctl); 1035 } else { 1036 if (trans_vrr_ctl & VRR_CTL_PIPELINE_FULL_OVERRIDE) { 1037 crtc_state->vrr.pipeline_full = 1038 REG_FIELD_GET(VRR_CTL_PIPELINE_FULL_MASK, trans_vrr_ctl); 1039 1040 crtc_state->vrr.guardband = 1041 intel_vrr_pipeline_full_to_guardband(crtc_state, 1042 crtc_state->vrr.pipeline_full); 1043 } 1044 } 1045 1046 if (trans_vrr_ctl & VRR_CTL_FLIP_LINE_EN) { 1047 crtc_state->vrr.flipline = intel_de_read(display, 1048 TRANS_VRR_FLIPLINE(display, cpu_transcoder)) + 1; 1049 crtc_state->vrr.vmax = intel_de_read(display, 1050 TRANS_VRR_VMAX(display, cpu_transcoder)) + 1; 1051 crtc_state->vrr.vmin = intel_de_read(display, 1052 TRANS_VRR_VMIN(display, cpu_transcoder)) + 1; 1053 1054 if (DISPLAY_VER(display) < 13) { 1055 /* undo what intel_vrr_hw_value() does when writing the values */ 1056 crtc_state->vrr.flipline += crtc_state->set_context_latency; 1057 crtc_state->vrr.vmax += crtc_state->set_context_latency; 1058 crtc_state->vrr.vmin += crtc_state->set_context_latency; 1059 1060 crtc_state->vrr.vmin += intel_vrr_vmin_flipline_offset(display); 1061 } 1062 1063 /* 1064 * For platforms that always use VRR Timing Generator, the VTOTAL.Vtotal 1065 * bits are not filled. Since for these platforms TRAN_VMIN is always 1066 * filled with crtc_vtotal, use TRAN_VRR_VMIN to get the vtotal for 1067 * adjusted_mode. 1068 */ 1069 if (intel_vrr_always_use_vrr_tg(display)) 1070 crtc_state->hw.adjusted_mode.crtc_vtotal = 1071 intel_vrr_vmin_vtotal(crtc_state); 1072 1073 if (HAS_AS_SDP(display)) { 1074 trans_vrr_vsync = 1075 intel_de_read(display, 1076 TRANS_VRR_VSYNC(display, cpu_transcoder)); 1077 crtc_state->vrr.vsync_start = 1078 REG_FIELD_GET(VRR_VSYNC_START_MASK, trans_vrr_vsync); 1079 crtc_state->vrr.vsync_end = 1080 REG_FIELD_GET(VRR_VSYNC_END_MASK, trans_vrr_vsync); 1081 } 1082 } 1083 1084 vrr_enable = trans_vrr_ctl & VRR_CTL_VRR_ENABLE; 1085 1086 if (intel_vrr_always_use_vrr_tg(display)) 1087 crtc_state->vrr.enable = vrr_enable && !intel_vrr_is_fixed_rr(crtc_state); 1088 else 1089 crtc_state->vrr.enable = vrr_enable; 1090 1091 intel_vrr_get_dc_balance_config(crtc_state); 1092 1093 /* 1094 * #TODO: For Both VRR and CMRR the flag I915_MODE_FLAG_VRR is set for mode_flags. 1095 * Since CMRR is currently disabled, set this flag for VRR for now. 1096 * Need to keep this in mind while re-enabling CMRR. 1097 */ 1098 if (crtc_state->vrr.enable) 1099 crtc_state->mode_flags |= I915_MODE_FLAG_VRR; 1100 1101 /* 1102 * For platforms that always use the VRR timing generator, we overwrite 1103 * crtc_vblank_start with vtotal - guardband to reflect the delayed 1104 * vblank start. This works for both default and optimized guardband values. 1105 * On other platforms, we keep the original value from 1106 * intel_get_transcoder_timings() and apply adjustments only in VRR-specific 1107 * paths as needed. 1108 */ 1109 if (intel_vrr_always_use_vrr_tg(display)) 1110 crtc_state->hw.adjusted_mode.crtc_vblank_start = 1111 crtc_state->hw.adjusted_mode.crtc_vtotal - 1112 crtc_state->vrr.guardband; 1113 } 1114 1115 int intel_vrr_safe_window_start(const struct intel_crtc_state *crtc_state) 1116 { 1117 struct intel_display *display = to_intel_display(crtc_state); 1118 1119 if (DISPLAY_VER(display) >= 30) 1120 return crtc_state->hw.adjusted_mode.crtc_vdisplay - 1121 crtc_state->set_context_latency; 1122 else 1123 return crtc_state->hw.adjusted_mode.crtc_vdisplay; 1124 } 1125 1126 static int 1127 intel_vrr_dcb_vmin_vblank_start(const struct intel_crtc_state *crtc_state) 1128 { 1129 return (intel_vrr_dcb_vmin_vblank_start_next(crtc_state) < 0) ? 1130 intel_vrr_dcb_vmin_vblank_start_final(crtc_state) : 1131 intel_vrr_dcb_vmin_vblank_start_next(crtc_state); 1132 } 1133 1134 int intel_vrr_vmin_safe_window_end(const struct intel_crtc_state *crtc_state) 1135 { 1136 int vmin_vblank_start = crtc_state->vrr.dc_balance.enable ? 1137 intel_vrr_dcb_vmin_vblank_start(crtc_state) : 1138 intel_vrr_vmin_vblank_start(crtc_state); 1139 1140 return vmin_vblank_start - crtc_state->set_context_latency; 1141 } 1142 1143 int intel_vrr_dcb_vmin_vblank_start_next(const struct intel_crtc_state *crtc_state) 1144 { 1145 struct intel_display *display = to_intel_display(crtc_state); 1146 enum transcoder cpu_transcoder = crtc_state->cpu_transcoder; 1147 u32 tmp = 0; 1148 1149 tmp = intel_de_read(display, TRANS_VRR_DCB_ADJ_FLIPLINE_CFG_LIVE(cpu_transcoder)); 1150 1151 if (REG_FIELD_GET(VRR_DCB_ADJ_FLIPLINE_CNT_MASK, tmp) == 0) 1152 return -EINVAL; 1153 1154 return intel_vrr_vblank_start(crtc_state, VRR_DCB_ADJ_FLIPLINE(tmp) + 1); 1155 } 1156 1157 int intel_vrr_dcb_vmax_vblank_start_next(const struct intel_crtc_state *crtc_state) 1158 { 1159 struct intel_display *display = to_intel_display(crtc_state); 1160 enum transcoder cpu_transcoder = crtc_state->cpu_transcoder; 1161 u32 tmp = 0; 1162 1163 tmp = intel_de_read(display, TRANS_VRR_DCB_ADJ_VMAX_CFG_LIVE(cpu_transcoder)); 1164 1165 if (REG_FIELD_GET(VRR_DCB_ADJ_VMAX_CNT_MASK, tmp) == 0) 1166 return -EINVAL; 1167 1168 return intel_vrr_vblank_start(crtc_state, VRR_DCB_ADJ_VMAX(tmp) + 1); 1169 } 1170 1171 int intel_vrr_dcb_vmin_vblank_start_final(const struct intel_crtc_state *crtc_state) 1172 { 1173 struct intel_display *display = to_intel_display(crtc_state); 1174 enum transcoder cpu_transcoder = crtc_state->cpu_transcoder; 1175 u32 tmp = 0; 1176 1177 tmp = intel_de_read(display, TRANS_VRR_DCB_FLIPLINE_LIVE(cpu_transcoder)); 1178 1179 return intel_vrr_vblank_start(crtc_state, VRR_DCB_FLIPLINE(tmp) + 1); 1180 } 1181 1182 int intel_vrr_dcb_vmax_vblank_start_final(const struct intel_crtc_state *crtc_state) 1183 { 1184 struct intel_display *display = to_intel_display(crtc_state); 1185 enum transcoder cpu_transcoder = crtc_state->cpu_transcoder; 1186 u32 tmp = 0; 1187 1188 tmp = intel_de_read(display, TRANS_VRR_DCB_VMAX_LIVE(cpu_transcoder)); 1189 1190 return intel_vrr_vblank_start(crtc_state, VRR_DCB_VMAX(tmp) + 1); 1191 } 1192