1 // SPDX-License-Identifier: GPL-2.0 OR MIT 2 /************************************************************************** 3 * 4 * Copyright (c) 2024 Broadcom. All Rights Reserved. The term 5 * “Broadcom” refers to Broadcom Inc. and/or its subsidiaries. 6 * 7 * Permission is hereby granted, free of charge, to any person obtaining a 8 * copy of this software and associated documentation files (the 9 * "Software"), to deal in the Software without restriction, including 10 * without limitation the rights to use, copy, modify, merge, publish, 11 * distribute, sub license, and/or sell copies of the Software, and to 12 * permit persons to whom the Software is furnished to do so, subject to 13 * the following conditions: 14 * 15 * The above copyright notice and this permission notice (including the 16 * next paragraph) shall be included in all copies or substantial portions 17 * of the Software. 18 * 19 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 20 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 21 * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL 22 * THE COPYRIGHT HOLDERS, AUTHORS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, 23 * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR 24 * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE 25 * USE OR OTHER DEALINGS IN THE SOFTWARE. 26 * 27 **************************************************************************/ 28 29 #include "vmwgfx_vkms.h" 30 31 #include "vmwgfx_bo.h" 32 #include "vmwgfx_drv.h" 33 #include "vmwgfx_kms.h" 34 35 #include "vmw_surface_cache.h" 36 37 #include <drm/drm_crtc.h> 38 #include <drm/drm_debugfs_crc.h> 39 #include <drm/drm_print.h> 40 #include <drm/drm_vblank.h> 41 42 #include <linux/crc32.h> 43 #include <linux/delay.h> 44 45 #define GUESTINFO_VBLANK "guestinfo.vmwgfx.vkms_enable" 46 47 static int 48 vmw_surface_sync(struct vmw_private *vmw, 49 struct vmw_surface *surf) 50 { 51 int ret; 52 struct vmw_fence_obj *fence = NULL; 53 struct vmw_bo *bo = surf->res.guest_memory_bo; 54 55 vmw_resource_clean(&surf->res); 56 57 ret = ttm_bo_reserve(&bo->tbo, false, false, NULL); 58 if (ret != 0) { 59 drm_warn(&vmw->drm, "%s: failed reserve\n", __func__); 60 goto done; 61 } 62 63 ret = vmw_execbuf_fence_commands(NULL, vmw, &fence, NULL); 64 if (ret != 0) { 65 drm_warn(&vmw->drm, "%s: failed execbuf\n", __func__); 66 ttm_bo_unreserve(&bo->tbo); 67 goto done; 68 } 69 70 dma_fence_wait(&fence->base, false); 71 dma_fence_put(&fence->base); 72 73 ttm_bo_unreserve(&bo->tbo); 74 done: 75 return ret; 76 } 77 78 static void 79 compute_crc(struct drm_crtc *crtc, 80 struct vmw_surface *surf, 81 u32 *crc) 82 { 83 u8 *mapped_surface; 84 struct vmw_bo *bo = surf->res.guest_memory_bo; 85 const struct SVGA3dSurfaceDesc *desc = 86 vmw_surface_get_desc(surf->metadata.format); 87 u32 row_pitch_bytes; 88 SVGA3dSize blocks; 89 u32 y; 90 91 *crc = 0; 92 93 vmw_surface_get_size_in_blocks(desc, &surf->metadata.base_size, &blocks); 94 row_pitch_bytes = blocks.width * desc->pitchBytesPerBlock; 95 WARN_ON(!bo); 96 mapped_surface = vmw_bo_map_and_cache(bo); 97 98 for (y = 0; y < blocks.height; y++) { 99 *crc = crc32_le(*crc, mapped_surface, row_pitch_bytes); 100 mapped_surface += row_pitch_bytes; 101 } 102 103 vmw_bo_unmap(bo); 104 } 105 106 static void 107 crc_generate_worker(struct work_struct *work) 108 { 109 struct vmw_display_unit *du = 110 container_of(work, struct vmw_display_unit, vkms.crc_generator_work); 111 struct drm_crtc *crtc = &du->crtc; 112 struct vmw_private *vmw = vmw_priv(crtc->dev); 113 bool crc_pending; 114 u64 frame_start, frame_end; 115 u32 crc32 = 0; 116 struct vmw_surface *surf = 0; 117 118 spin_lock_irq(&du->vkms.crc_state_lock); 119 crc_pending = du->vkms.crc_pending; 120 spin_unlock_irq(&du->vkms.crc_state_lock); 121 122 /* 123 * We raced with the vblank hrtimer and previous work already computed 124 * the crc, nothing to do. 125 */ 126 if (!crc_pending) 127 return; 128 129 spin_lock_irq(&du->vkms.crc_state_lock); 130 surf = vmw_surface_reference(du->vkms.surface); 131 spin_unlock_irq(&du->vkms.crc_state_lock); 132 133 if (surf) { 134 if (vmw_surface_sync(vmw, surf)) { 135 drm_warn( 136 crtc->dev, 137 "CRC worker wasn't able to sync the crc surface!\n"); 138 return; 139 } 140 141 compute_crc(crtc, surf, &crc32); 142 vmw_surface_unreference(&surf); 143 } 144 145 spin_lock_irq(&du->vkms.crc_state_lock); 146 frame_start = du->vkms.frame_start; 147 frame_end = du->vkms.frame_end; 148 du->vkms.frame_start = 0; 149 du->vkms.frame_end = 0; 150 du->vkms.crc_pending = false; 151 spin_unlock_irq(&du->vkms.crc_state_lock); 152 153 /* 154 * The worker can fall behind the vblank hrtimer, make sure we catch up. 155 */ 156 while (frame_start <= frame_end) 157 drm_crtc_add_crc_entry(crtc, true, frame_start++, &crc32); 158 } 159 160 static enum hrtimer_restart 161 vmw_vkms_vblank_simulate(struct hrtimer *timer) 162 { 163 struct vmw_display_unit *du = container_of(timer, struct vmw_display_unit, vkms.timer); 164 struct drm_crtc *crtc = &du->crtc; 165 struct vmw_private *vmw = vmw_priv(crtc->dev); 166 bool has_surface = false; 167 u64 ret_overrun; 168 bool locked, ret; 169 170 ret_overrun = hrtimer_forward_now(&du->vkms.timer, 171 du->vkms.period_ns); 172 if (ret_overrun != 1) 173 drm_dbg_driver(crtc->dev, "vblank timer missed %lld frames.\n", 174 ret_overrun - 1); 175 176 locked = vmw_vkms_vblank_trylock(crtc); 177 ret = drm_crtc_handle_vblank(crtc); 178 WARN_ON(!ret); 179 if (!locked) 180 return HRTIMER_RESTART; 181 has_surface = du->vkms.surface != NULL; 182 vmw_vkms_unlock(crtc); 183 184 if (du->vkms.crc_enabled && has_surface) { 185 u64 frame = drm_crtc_accurate_vblank_count(crtc); 186 187 spin_lock(&du->vkms.crc_state_lock); 188 if (!du->vkms.crc_pending) 189 du->vkms.frame_start = frame; 190 else 191 drm_dbg_driver(crtc->dev, 192 "crc worker falling behind, frame_start: %llu, frame_end: %llu\n", 193 du->vkms.frame_start, frame); 194 du->vkms.frame_end = frame; 195 du->vkms.crc_pending = true; 196 spin_unlock(&du->vkms.crc_state_lock); 197 198 ret = queue_work(vmw->crc_workq, &du->vkms.crc_generator_work); 199 if (!ret) 200 drm_dbg_driver(crtc->dev, "Composer worker already queued\n"); 201 } 202 203 return HRTIMER_RESTART; 204 } 205 206 void 207 vmw_vkms_init(struct vmw_private *vmw) 208 { 209 char buffer[64]; 210 const size_t max_buf_len = sizeof(buffer) - 1; 211 size_t buf_len = max_buf_len; 212 int ret; 213 214 vmw->vkms_enabled = false; 215 216 ret = vmw_host_get_guestinfo(GUESTINFO_VBLANK, buffer, &buf_len); 217 if (ret || buf_len > max_buf_len) 218 return; 219 buffer[buf_len] = '\0'; 220 221 ret = kstrtobool(buffer, &vmw->vkms_enabled); 222 if (!ret && vmw->vkms_enabled) { 223 ret = drm_vblank_init(&vmw->drm, VMWGFX_NUM_DISPLAY_UNITS); 224 vmw->vkms_enabled = (ret == 0); 225 } 226 227 vmw->crc_workq = alloc_ordered_workqueue("vmwgfx_crc_generator", 0); 228 if (!vmw->crc_workq) { 229 drm_warn(&vmw->drm, "crc workqueue allocation failed. Disabling vkms."); 230 vmw->vkms_enabled = false; 231 } 232 if (vmw->vkms_enabled) 233 drm_info(&vmw->drm, "VKMS enabled\n"); 234 } 235 236 void 237 vmw_vkms_cleanup(struct vmw_private *vmw) 238 { 239 destroy_workqueue(vmw->crc_workq); 240 } 241 242 bool 243 vmw_vkms_get_vblank_timestamp(struct drm_crtc *crtc, 244 int *max_error, 245 ktime_t *vblank_time, 246 bool in_vblank_irq) 247 { 248 struct drm_device *dev = crtc->dev; 249 struct vmw_private *vmw = vmw_priv(dev); 250 struct vmw_display_unit *du = vmw_crtc_to_du(crtc); 251 struct drm_vblank_crtc *vblank = drm_crtc_vblank_crtc(crtc); 252 253 if (!vmw->vkms_enabled) 254 return false; 255 256 if (!READ_ONCE(vblank->enabled)) { 257 *vblank_time = ktime_get(); 258 return true; 259 } 260 261 *vblank_time = READ_ONCE(du->vkms.timer.node.expires); 262 263 if (WARN_ON(*vblank_time == vblank->time)) 264 return true; 265 266 /* 267 * To prevent races we roll the hrtimer forward before we do any 268 * interrupt processing - this is how real hw works (the interrupt is 269 * only generated after all the vblank registers are updated) and what 270 * the vblank core expects. Therefore we need to always correct the 271 * timestampe by one frame. 272 */ 273 *vblank_time -= du->vkms.period_ns; 274 275 return true; 276 } 277 278 int 279 vmw_vkms_enable_vblank(struct drm_crtc *crtc) 280 { 281 struct drm_device *dev = crtc->dev; 282 struct vmw_private *vmw = vmw_priv(dev); 283 struct drm_vblank_crtc *vblank = drm_crtc_vblank_crtc(crtc); 284 struct vmw_display_unit *du = vmw_crtc_to_du(crtc); 285 286 if (!vmw->vkms_enabled) 287 return -EINVAL; 288 289 drm_calc_timestamping_constants(crtc, &crtc->mode); 290 291 hrtimer_setup(&du->vkms.timer, &vmw_vkms_vblank_simulate, CLOCK_MONOTONIC, 292 HRTIMER_MODE_REL); 293 du->vkms.period_ns = ktime_set(0, vblank->framedur_ns); 294 hrtimer_start(&du->vkms.timer, du->vkms.period_ns, HRTIMER_MODE_REL); 295 296 return 0; 297 } 298 299 void 300 vmw_vkms_disable_vblank(struct drm_crtc *crtc) 301 { 302 struct vmw_display_unit *du = vmw_crtc_to_du(crtc); 303 struct vmw_private *vmw = vmw_priv(crtc->dev); 304 305 if (!vmw->vkms_enabled) 306 return; 307 308 hrtimer_cancel(&du->vkms.timer); 309 du->vkms.surface = NULL; 310 du->vkms.period_ns = ktime_set(0, 0); 311 } 312 313 enum vmw_vkms_lock_state { 314 VMW_VKMS_LOCK_UNLOCKED = 0, 315 VMW_VKMS_LOCK_MODESET = 1, 316 VMW_VKMS_LOCK_VBLANK = 2 317 }; 318 319 void 320 vmw_vkms_crtc_init(struct drm_crtc *crtc) 321 { 322 struct vmw_display_unit *du = vmw_crtc_to_du(crtc); 323 324 atomic_set(&du->vkms.atomic_lock, VMW_VKMS_LOCK_UNLOCKED); 325 spin_lock_init(&du->vkms.crc_state_lock); 326 327 INIT_WORK(&du->vkms.crc_generator_work, crc_generate_worker); 328 du->vkms.surface = NULL; 329 } 330 331 void 332 vmw_vkms_crtc_cleanup(struct drm_crtc *crtc) 333 { 334 struct vmw_display_unit *du = vmw_crtc_to_du(crtc); 335 336 if (du->vkms.surface) 337 vmw_surface_unreference(&du->vkms.surface); 338 WARN_ON(work_pending(&du->vkms.crc_generator_work)); 339 hrtimer_cancel(&du->vkms.timer); 340 } 341 342 void 343 vmw_vkms_crtc_atomic_begin(struct drm_crtc *crtc, 344 struct drm_atomic_state *state) 345 { 346 struct vmw_private *vmw = vmw_priv(crtc->dev); 347 348 if (vmw->vkms_enabled) 349 vmw_vkms_modeset_lock(crtc); 350 } 351 352 void 353 vmw_vkms_crtc_atomic_flush(struct drm_crtc *crtc, 354 struct drm_atomic_state *state) 355 { 356 unsigned long flags; 357 struct vmw_private *vmw = vmw_priv(crtc->dev); 358 359 if (!vmw->vkms_enabled) 360 return; 361 362 if (crtc->state->event) { 363 spin_lock_irqsave(&crtc->dev->event_lock, flags); 364 365 if (drm_crtc_vblank_get(crtc) != 0) 366 drm_crtc_send_vblank_event(crtc, crtc->state->event); 367 else 368 drm_crtc_arm_vblank_event(crtc, crtc->state->event); 369 370 spin_unlock_irqrestore(&crtc->dev->event_lock, flags); 371 372 crtc->state->event = NULL; 373 } 374 375 vmw_vkms_unlock(crtc); 376 } 377 378 void 379 vmw_vkms_crtc_atomic_enable(struct drm_crtc *crtc, 380 struct drm_atomic_state *state) 381 { 382 struct vmw_private *vmw = vmw_priv(crtc->dev); 383 384 if (vmw->vkms_enabled) 385 drm_crtc_vblank_on(crtc); 386 } 387 388 void 389 vmw_vkms_crtc_atomic_disable(struct drm_crtc *crtc, 390 struct drm_atomic_state *state) 391 { 392 struct vmw_private *vmw = vmw_priv(crtc->dev); 393 394 if (vmw->vkms_enabled) 395 drm_crtc_vblank_off(crtc); 396 } 397 398 static bool 399 is_crc_supported(struct drm_crtc *crtc) 400 { 401 struct vmw_private *vmw = vmw_priv(crtc->dev); 402 403 if (!vmw->vkms_enabled) 404 return false; 405 406 if (vmw->active_display_unit != vmw_du_screen_target) 407 return false; 408 409 return true; 410 } 411 412 static const char * const pipe_crc_sources[] = {"auto"}; 413 414 static int 415 crc_parse_source(const char *src_name, 416 bool *enabled) 417 { 418 int ret = 0; 419 420 if (!src_name) { 421 *enabled = false; 422 } else if (strcmp(src_name, "auto") == 0) { 423 *enabled = true; 424 } else { 425 *enabled = false; 426 ret = -EINVAL; 427 } 428 429 return ret; 430 } 431 432 const char *const * 433 vmw_vkms_get_crc_sources(struct drm_crtc *crtc, 434 size_t *count) 435 { 436 *count = 0; 437 if (!is_crc_supported(crtc)) 438 return NULL; 439 440 *count = ARRAY_SIZE(pipe_crc_sources); 441 return pipe_crc_sources; 442 } 443 444 int 445 vmw_vkms_verify_crc_source(struct drm_crtc *crtc, 446 const char *src_name, 447 size_t *values_cnt) 448 { 449 bool enabled; 450 451 if (!is_crc_supported(crtc)) 452 return -EINVAL; 453 454 if (crc_parse_source(src_name, &enabled) < 0) { 455 drm_dbg_driver(crtc->dev, "unknown source '%s'\n", src_name); 456 return -EINVAL; 457 } 458 459 *values_cnt = 1; 460 461 return 0; 462 } 463 464 int 465 vmw_vkms_set_crc_source(struct drm_crtc *crtc, 466 const char *src_name) 467 { 468 struct vmw_display_unit *du = vmw_crtc_to_du(crtc); 469 bool enabled, prev_enabled, locked; 470 int ret; 471 472 if (!is_crc_supported(crtc)) 473 return -EINVAL; 474 475 ret = crc_parse_source(src_name, &enabled); 476 477 if (enabled) 478 drm_crtc_vblank_get(crtc); 479 480 locked = vmw_vkms_modeset_lock_relaxed(crtc); 481 prev_enabled = du->vkms.crc_enabled; 482 du->vkms.crc_enabled = enabled; 483 if (locked) 484 vmw_vkms_unlock(crtc); 485 486 if (prev_enabled) 487 drm_crtc_vblank_put(crtc); 488 489 return ret; 490 } 491 492 void 493 vmw_vkms_set_crc_surface(struct drm_crtc *crtc, 494 struct vmw_surface *surf) 495 { 496 struct vmw_display_unit *du = vmw_crtc_to_du(crtc); 497 struct vmw_private *vmw = vmw_priv(crtc->dev); 498 499 if (vmw->vkms_enabled && du->vkms.surface != surf) { 500 WARN_ON(atomic_read(&du->vkms.atomic_lock) != VMW_VKMS_LOCK_MODESET); 501 if (du->vkms.surface) 502 vmw_surface_unreference(&du->vkms.surface); 503 if (surf) 504 du->vkms.surface = vmw_surface_reference(surf); 505 } 506 } 507 508 /** 509 * vmw_vkms_lock_max_wait_ns - Return the max wait for the vkms lock 510 * @du: The vmw_display_unit from which to grab the vblank timings 511 * 512 * Returns the maximum wait time used to acquire the vkms lock. By 513 * default uses a time of a single frame and in case where vblank 514 * was not initialized for the display unit 1/60th of a second. 515 */ 516 static inline u64 517 vmw_vkms_lock_max_wait_ns(struct vmw_display_unit *du) 518 { 519 s64 nsecs = ktime_to_ns(du->vkms.period_ns); 520 521 return (nsecs > 0) ? nsecs : 16666666; 522 } 523 524 /** 525 * vmw_vkms_modeset_lock - Protects access to crtc during modeset 526 * @crtc: The crtc to lock for vkms 527 * 528 * This function prevents the VKMS timers/callbacks from being called 529 * while a modeset operation is in process. We don't want the callbacks 530 * e.g. the vblank simulator to be trying to access incomplete state 531 * so we need to make sure they execute only when the modeset has 532 * finished. 533 * 534 * Normally this would have been done with a spinlock but locking the 535 * entire atomic modeset with vmwgfx is impossible because kms prepare 536 * executes non-atomic ops (e.g. vmw_validation_prepare holds a mutex to 537 * guard various bits of state). Which means that we need to synchronize 538 * atomic context (the vblank handler) with the non-atomic entirity 539 * of kms - so use an atomic_t to track which part of vkms has access 540 * to the basic vkms state. 541 */ 542 void 543 vmw_vkms_modeset_lock(struct drm_crtc *crtc) 544 { 545 struct vmw_display_unit *du = vmw_crtc_to_du(crtc); 546 const u64 nsecs_delay = 10; 547 const u64 MAX_NSECS_DELAY = vmw_vkms_lock_max_wait_ns(du); 548 u64 total_delay = 0; 549 int ret; 550 551 do { 552 ret = atomic_cmpxchg(&du->vkms.atomic_lock, 553 VMW_VKMS_LOCK_UNLOCKED, 554 VMW_VKMS_LOCK_MODESET); 555 if (ret == VMW_VKMS_LOCK_UNLOCKED || total_delay >= MAX_NSECS_DELAY) 556 break; 557 ndelay(nsecs_delay); 558 total_delay += nsecs_delay; 559 } while (1); 560 561 if (total_delay >= MAX_NSECS_DELAY) { 562 drm_warn(crtc->dev, "VKMS lock expired! total_delay = %lld, ret = %d, cur = %d\n", 563 total_delay, ret, atomic_read(&du->vkms.atomic_lock)); 564 } 565 } 566 567 /** 568 * vmw_vkms_modeset_lock_relaxed - Protects access to crtc during modeset 569 * @crtc: The crtc to lock for vkms 570 * 571 * Much like vmw_vkms_modeset_lock except that when the crtc is currently 572 * in a modeset it will return immediately. 573 * 574 * Returns true if actually locked vkms to modeset or false otherwise. 575 */ 576 bool 577 vmw_vkms_modeset_lock_relaxed(struct drm_crtc *crtc) 578 { 579 struct vmw_display_unit *du = vmw_crtc_to_du(crtc); 580 const u64 nsecs_delay = 10; 581 const u64 MAX_NSECS_DELAY = vmw_vkms_lock_max_wait_ns(du); 582 u64 total_delay = 0; 583 int ret; 584 585 do { 586 ret = atomic_cmpxchg(&du->vkms.atomic_lock, 587 VMW_VKMS_LOCK_UNLOCKED, 588 VMW_VKMS_LOCK_MODESET); 589 if (ret == VMW_VKMS_LOCK_UNLOCKED || 590 ret == VMW_VKMS_LOCK_MODESET || 591 total_delay >= MAX_NSECS_DELAY) 592 break; 593 ndelay(nsecs_delay); 594 total_delay += nsecs_delay; 595 } while (1); 596 597 if (total_delay >= MAX_NSECS_DELAY) { 598 drm_warn(crtc->dev, "VKMS relaxed lock expired!\n"); 599 return false; 600 } 601 602 return ret == VMW_VKMS_LOCK_UNLOCKED; 603 } 604 605 /** 606 * vmw_vkms_vblank_trylock - Protects access to crtc during vblank 607 * @crtc: The crtc to lock for vkms 608 * 609 * Tries to lock vkms for vblank, returns immediately. 610 * 611 * Returns true if locked vkms to vblank or false otherwise. 612 */ 613 bool 614 vmw_vkms_vblank_trylock(struct drm_crtc *crtc) 615 { 616 struct vmw_display_unit *du = vmw_crtc_to_du(crtc); 617 u32 ret; 618 619 ret = atomic_cmpxchg(&du->vkms.atomic_lock, 620 VMW_VKMS_LOCK_UNLOCKED, 621 VMW_VKMS_LOCK_VBLANK); 622 623 return ret == VMW_VKMS_LOCK_UNLOCKED; 624 } 625 626 void 627 vmw_vkms_unlock(struct drm_crtc *crtc) 628 { 629 struct vmw_display_unit *du = vmw_crtc_to_du(crtc); 630 631 /* Release flag; mark it as unlocked. */ 632 atomic_set(&du->vkms.atomic_lock, VMW_VKMS_LOCK_UNLOCKED); 633 } 634