1 // SPDX-License-Identifier: MIT 2 /* 3 * Copyright © 2016 Intel Corporation 4 */ 5 6 #include <linux/kthread.h> 7 8 #include "gem/i915_gem_context.h" 9 10 #include "intel_gt.h" 11 #include "intel_engine_heartbeat.h" 12 #include "intel_engine_pm.h" 13 #include "selftest_engine_heartbeat.h" 14 15 #include "i915_selftest.h" 16 #include "selftests/i915_random.h" 17 #include "selftests/igt_flush_test.h" 18 #include "selftests/igt_reset.h" 19 #include "selftests/igt_atomic.h" 20 21 #include "selftests/mock_drm.h" 22 23 #include "gem/selftests/mock_context.h" 24 #include "gem/selftests/igt_gem_utils.h" 25 26 #define IGT_IDLE_TIMEOUT 50 /* ms; time to wait after flushing between tests */ 27 28 struct hang { 29 struct intel_gt *gt; 30 struct drm_i915_gem_object *hws; 31 struct drm_i915_gem_object *obj; 32 struct i915_gem_context *ctx; 33 u32 *seqno; 34 u32 *batch; 35 }; 36 37 static int hang_init(struct hang *h, struct intel_gt *gt) 38 { 39 void *vaddr; 40 int err; 41 42 memset(h, 0, sizeof(*h)); 43 h->gt = gt; 44 45 h->ctx = kernel_context(gt->i915); 46 if (IS_ERR(h->ctx)) 47 return PTR_ERR(h->ctx); 48 49 GEM_BUG_ON(i915_gem_context_is_bannable(h->ctx)); 50 51 h->hws = i915_gem_object_create_internal(gt->i915, PAGE_SIZE); 52 if (IS_ERR(h->hws)) { 53 err = PTR_ERR(h->hws); 54 goto err_ctx; 55 } 56 57 h->obj = i915_gem_object_create_internal(gt->i915, PAGE_SIZE); 58 if (IS_ERR(h->obj)) { 59 err = PTR_ERR(h->obj); 60 goto err_hws; 61 } 62 63 i915_gem_object_set_cache_coherency(h->hws, I915_CACHE_LLC); 64 vaddr = i915_gem_object_pin_map_unlocked(h->hws, I915_MAP_WB); 65 if (IS_ERR(vaddr)) { 66 err = PTR_ERR(vaddr); 67 goto err_obj; 68 } 69 h->seqno = memset(vaddr, 0xff, PAGE_SIZE); 70 71 vaddr = i915_gem_object_pin_map_unlocked(h->obj, 72 i915_coherent_map_type(gt->i915, h->obj, false)); 73 if (IS_ERR(vaddr)) { 74 err = PTR_ERR(vaddr); 75 goto err_unpin_hws; 76 } 77 h->batch = vaddr; 78 79 return 0; 80 81 err_unpin_hws: 82 i915_gem_object_unpin_map(h->hws); 83 err_obj: 84 i915_gem_object_put(h->obj); 85 err_hws: 86 i915_gem_object_put(h->hws); 87 err_ctx: 88 kernel_context_close(h->ctx); 89 return err; 90 } 91 92 static u64 hws_address(const struct i915_vma *hws, 93 const struct i915_request *rq) 94 { 95 return hws->node.start + offset_in_page(sizeof(u32)*rq->fence.context); 96 } 97 98 static int move_to_active(struct i915_vma *vma, 99 struct i915_request *rq, 100 unsigned int flags) 101 { 102 int err; 103 104 i915_vma_lock(vma); 105 err = i915_request_await_object(rq, vma->obj, 106 flags & EXEC_OBJECT_WRITE); 107 if (err == 0) 108 err = i915_vma_move_to_active(vma, rq, flags); 109 i915_vma_unlock(vma); 110 111 return err; 112 } 113 114 static struct i915_request * 115 hang_create_request(struct hang *h, struct intel_engine_cs *engine) 116 { 117 struct intel_gt *gt = h->gt; 118 struct i915_address_space *vm = i915_gem_context_get_vm_rcu(h->ctx); 119 struct drm_i915_gem_object *obj; 120 struct i915_request *rq = NULL; 121 struct i915_vma *hws, *vma; 122 unsigned int flags; 123 void *vaddr; 124 u32 *batch; 125 int err; 126 127 obj = i915_gem_object_create_internal(gt->i915, PAGE_SIZE); 128 if (IS_ERR(obj)) { 129 i915_vm_put(vm); 130 return ERR_CAST(obj); 131 } 132 133 vaddr = i915_gem_object_pin_map_unlocked(obj, i915_coherent_map_type(gt->i915, obj, false)); 134 if (IS_ERR(vaddr)) { 135 i915_gem_object_put(obj); 136 i915_vm_put(vm); 137 return ERR_CAST(vaddr); 138 } 139 140 i915_gem_object_unpin_map(h->obj); 141 i915_gem_object_put(h->obj); 142 143 h->obj = obj; 144 h->batch = vaddr; 145 146 vma = i915_vma_instance(h->obj, vm, NULL); 147 if (IS_ERR(vma)) { 148 i915_vm_put(vm); 149 return ERR_CAST(vma); 150 } 151 152 hws = i915_vma_instance(h->hws, vm, NULL); 153 if (IS_ERR(hws)) { 154 i915_vm_put(vm); 155 return ERR_CAST(hws); 156 } 157 158 err = i915_vma_pin(vma, 0, 0, PIN_USER); 159 if (err) { 160 i915_vm_put(vm); 161 return ERR_PTR(err); 162 } 163 164 err = i915_vma_pin(hws, 0, 0, PIN_USER); 165 if (err) 166 goto unpin_vma; 167 168 rq = igt_request_alloc(h->ctx, engine); 169 if (IS_ERR(rq)) { 170 err = PTR_ERR(rq); 171 goto unpin_hws; 172 } 173 174 err = move_to_active(vma, rq, 0); 175 if (err) 176 goto cancel_rq; 177 178 err = move_to_active(hws, rq, 0); 179 if (err) 180 goto cancel_rq; 181 182 batch = h->batch; 183 if (INTEL_GEN(gt->i915) >= 8) { 184 *batch++ = MI_STORE_DWORD_IMM_GEN4; 185 *batch++ = lower_32_bits(hws_address(hws, rq)); 186 *batch++ = upper_32_bits(hws_address(hws, rq)); 187 *batch++ = rq->fence.seqno; 188 *batch++ = MI_NOOP; 189 190 memset(batch, 0, 1024); 191 batch += 1024 / sizeof(*batch); 192 193 *batch++ = MI_NOOP; 194 *batch++ = MI_BATCH_BUFFER_START | 1 << 8 | 1; 195 *batch++ = lower_32_bits(vma->node.start); 196 *batch++ = upper_32_bits(vma->node.start); 197 } else if (INTEL_GEN(gt->i915) >= 6) { 198 *batch++ = MI_STORE_DWORD_IMM_GEN4; 199 *batch++ = 0; 200 *batch++ = lower_32_bits(hws_address(hws, rq)); 201 *batch++ = rq->fence.seqno; 202 *batch++ = MI_NOOP; 203 204 memset(batch, 0, 1024); 205 batch += 1024 / sizeof(*batch); 206 207 *batch++ = MI_NOOP; 208 *batch++ = MI_BATCH_BUFFER_START | 1 << 8; 209 *batch++ = lower_32_bits(vma->node.start); 210 } else if (INTEL_GEN(gt->i915) >= 4) { 211 *batch++ = MI_STORE_DWORD_IMM_GEN4 | MI_USE_GGTT; 212 *batch++ = 0; 213 *batch++ = lower_32_bits(hws_address(hws, rq)); 214 *batch++ = rq->fence.seqno; 215 *batch++ = MI_NOOP; 216 217 memset(batch, 0, 1024); 218 batch += 1024 / sizeof(*batch); 219 220 *batch++ = MI_NOOP; 221 *batch++ = MI_BATCH_BUFFER_START | 2 << 6; 222 *batch++ = lower_32_bits(vma->node.start); 223 } else { 224 *batch++ = MI_STORE_DWORD_IMM | MI_MEM_VIRTUAL; 225 *batch++ = lower_32_bits(hws_address(hws, rq)); 226 *batch++ = rq->fence.seqno; 227 *batch++ = MI_NOOP; 228 229 memset(batch, 0, 1024); 230 batch += 1024 / sizeof(*batch); 231 232 *batch++ = MI_NOOP; 233 *batch++ = MI_BATCH_BUFFER_START | 2 << 6; 234 *batch++ = lower_32_bits(vma->node.start); 235 } 236 *batch++ = MI_BATCH_BUFFER_END; /* not reached */ 237 intel_gt_chipset_flush(engine->gt); 238 239 if (rq->engine->emit_init_breadcrumb) { 240 err = rq->engine->emit_init_breadcrumb(rq); 241 if (err) 242 goto cancel_rq; 243 } 244 245 flags = 0; 246 if (INTEL_GEN(gt->i915) <= 5) 247 flags |= I915_DISPATCH_SECURE; 248 249 err = rq->engine->emit_bb_start(rq, vma->node.start, PAGE_SIZE, flags); 250 251 cancel_rq: 252 if (err) { 253 i915_request_set_error_once(rq, err); 254 i915_request_add(rq); 255 } 256 unpin_hws: 257 i915_vma_unpin(hws); 258 unpin_vma: 259 i915_vma_unpin(vma); 260 i915_vm_put(vm); 261 return err ? ERR_PTR(err) : rq; 262 } 263 264 static u32 hws_seqno(const struct hang *h, const struct i915_request *rq) 265 { 266 return READ_ONCE(h->seqno[rq->fence.context % (PAGE_SIZE/sizeof(u32))]); 267 } 268 269 static void hang_fini(struct hang *h) 270 { 271 *h->batch = MI_BATCH_BUFFER_END; 272 intel_gt_chipset_flush(h->gt); 273 274 i915_gem_object_unpin_map(h->obj); 275 i915_gem_object_put(h->obj); 276 277 i915_gem_object_unpin_map(h->hws); 278 i915_gem_object_put(h->hws); 279 280 kernel_context_close(h->ctx); 281 282 igt_flush_test(h->gt->i915); 283 } 284 285 static bool wait_until_running(struct hang *h, struct i915_request *rq) 286 { 287 return !(wait_for_us(i915_seqno_passed(hws_seqno(h, rq), 288 rq->fence.seqno), 289 10) && 290 wait_for(i915_seqno_passed(hws_seqno(h, rq), 291 rq->fence.seqno), 292 1000)); 293 } 294 295 static int igt_hang_sanitycheck(void *arg) 296 { 297 struct intel_gt *gt = arg; 298 struct i915_request *rq; 299 struct intel_engine_cs *engine; 300 enum intel_engine_id id; 301 struct hang h; 302 int err; 303 304 /* Basic check that we can execute our hanging batch */ 305 306 err = hang_init(&h, gt); 307 if (err) 308 return err; 309 310 for_each_engine(engine, gt, id) { 311 struct intel_wedge_me w; 312 long timeout; 313 314 if (!intel_engine_can_store_dword(engine)) 315 continue; 316 317 rq = hang_create_request(&h, engine); 318 if (IS_ERR(rq)) { 319 err = PTR_ERR(rq); 320 pr_err("Failed to create request for %s, err=%d\n", 321 engine->name, err); 322 goto fini; 323 } 324 325 i915_request_get(rq); 326 327 *h.batch = MI_BATCH_BUFFER_END; 328 intel_gt_chipset_flush(engine->gt); 329 330 i915_request_add(rq); 331 332 timeout = 0; 333 intel_wedge_on_timeout(&w, gt, HZ / 10 /* 100ms */) 334 timeout = i915_request_wait(rq, 0, 335 MAX_SCHEDULE_TIMEOUT); 336 if (intel_gt_is_wedged(gt)) 337 timeout = -EIO; 338 339 i915_request_put(rq); 340 341 if (timeout < 0) { 342 err = timeout; 343 pr_err("Wait for request failed on %s, err=%d\n", 344 engine->name, err); 345 goto fini; 346 } 347 } 348 349 fini: 350 hang_fini(&h); 351 return err; 352 } 353 354 static bool wait_for_idle(struct intel_engine_cs *engine) 355 { 356 return wait_for(intel_engine_is_idle(engine), IGT_IDLE_TIMEOUT) == 0; 357 } 358 359 static int igt_reset_nop(void *arg) 360 { 361 struct intel_gt *gt = arg; 362 struct i915_gpu_error *global = >->i915->gpu_error; 363 struct intel_engine_cs *engine; 364 unsigned int reset_count, count; 365 enum intel_engine_id id; 366 IGT_TIMEOUT(end_time); 367 int err = 0; 368 369 /* Check that we can reset during non-user portions of requests */ 370 371 reset_count = i915_reset_count(global); 372 count = 0; 373 do { 374 for_each_engine(engine, gt, id) { 375 struct intel_context *ce; 376 int i; 377 378 ce = intel_context_create(engine); 379 if (IS_ERR(ce)) { 380 err = PTR_ERR(ce); 381 break; 382 } 383 384 for (i = 0; i < 16; i++) { 385 struct i915_request *rq; 386 387 rq = intel_context_create_request(ce); 388 if (IS_ERR(rq)) { 389 err = PTR_ERR(rq); 390 break; 391 } 392 393 i915_request_add(rq); 394 } 395 396 intel_context_put(ce); 397 } 398 399 igt_global_reset_lock(gt); 400 intel_gt_reset(gt, ALL_ENGINES, NULL); 401 igt_global_reset_unlock(gt); 402 403 if (intel_gt_is_wedged(gt)) { 404 err = -EIO; 405 break; 406 } 407 408 if (i915_reset_count(global) != reset_count + ++count) { 409 pr_err("Full GPU reset not recorded!\n"); 410 err = -EINVAL; 411 break; 412 } 413 414 err = igt_flush_test(gt->i915); 415 if (err) 416 break; 417 } while (time_before(jiffies, end_time)); 418 pr_info("%s: %d resets\n", __func__, count); 419 420 if (igt_flush_test(gt->i915)) 421 err = -EIO; 422 return err; 423 } 424 425 static int igt_reset_nop_engine(void *arg) 426 { 427 struct intel_gt *gt = arg; 428 struct i915_gpu_error *global = >->i915->gpu_error; 429 struct intel_engine_cs *engine; 430 enum intel_engine_id id; 431 432 /* Check that we can engine-reset during non-user portions */ 433 434 if (!intel_has_reset_engine(gt)) 435 return 0; 436 437 for_each_engine(engine, gt, id) { 438 unsigned int reset_count, reset_engine_count, count; 439 struct intel_context *ce; 440 IGT_TIMEOUT(end_time); 441 int err; 442 443 ce = intel_context_create(engine); 444 if (IS_ERR(ce)) 445 return PTR_ERR(ce); 446 447 reset_count = i915_reset_count(global); 448 reset_engine_count = i915_reset_engine_count(global, engine); 449 count = 0; 450 451 st_engine_heartbeat_disable(engine); 452 set_bit(I915_RESET_ENGINE + id, >->reset.flags); 453 do { 454 int i; 455 456 if (!wait_for_idle(engine)) { 457 pr_err("%s failed to idle before reset\n", 458 engine->name); 459 err = -EIO; 460 break; 461 } 462 463 for (i = 0; i < 16; i++) { 464 struct i915_request *rq; 465 466 rq = intel_context_create_request(ce); 467 if (IS_ERR(rq)) { 468 struct drm_printer p = 469 drm_info_printer(gt->i915->drm.dev); 470 intel_engine_dump(engine, &p, 471 "%s(%s): failed to submit request\n", 472 __func__, 473 engine->name); 474 475 GEM_TRACE("%s(%s): failed to submit request\n", 476 __func__, 477 engine->name); 478 GEM_TRACE_DUMP(); 479 480 intel_gt_set_wedged(gt); 481 482 err = PTR_ERR(rq); 483 break; 484 } 485 486 i915_request_add(rq); 487 } 488 err = intel_engine_reset(engine, NULL); 489 if (err) { 490 pr_err("intel_engine_reset(%s) failed, err:%d\n", 491 engine->name, err); 492 break; 493 } 494 495 if (i915_reset_count(global) != reset_count) { 496 pr_err("Full GPU reset recorded! (engine reset expected)\n"); 497 err = -EINVAL; 498 break; 499 } 500 501 if (i915_reset_engine_count(global, engine) != 502 reset_engine_count + ++count) { 503 pr_err("%s engine reset not recorded!\n", 504 engine->name); 505 err = -EINVAL; 506 break; 507 } 508 } while (time_before(jiffies, end_time)); 509 clear_bit(I915_RESET_ENGINE + id, >->reset.flags); 510 st_engine_heartbeat_enable(engine); 511 512 pr_info("%s(%s): %d resets\n", __func__, engine->name, count); 513 514 intel_context_put(ce); 515 if (igt_flush_test(gt->i915)) 516 err = -EIO; 517 if (err) 518 return err; 519 } 520 521 return 0; 522 } 523 524 static void force_reset_timeout(struct intel_engine_cs *engine) 525 { 526 engine->reset_timeout.probability = 999; 527 atomic_set(&engine->reset_timeout.times, -1); 528 } 529 530 static void cancel_reset_timeout(struct intel_engine_cs *engine) 531 { 532 memset(&engine->reset_timeout, 0, sizeof(engine->reset_timeout)); 533 } 534 535 static int igt_reset_fail_engine(void *arg) 536 { 537 struct intel_gt *gt = arg; 538 struct intel_engine_cs *engine; 539 enum intel_engine_id id; 540 541 /* Check that we can recover from engine-reset failues */ 542 543 if (!intel_has_reset_engine(gt)) 544 return 0; 545 546 for_each_engine(engine, gt, id) { 547 unsigned int count; 548 struct intel_context *ce; 549 IGT_TIMEOUT(end_time); 550 int err; 551 552 ce = intel_context_create(engine); 553 if (IS_ERR(ce)) 554 return PTR_ERR(ce); 555 556 st_engine_heartbeat_disable(engine); 557 set_bit(I915_RESET_ENGINE + id, >->reset.flags); 558 559 force_reset_timeout(engine); 560 err = intel_engine_reset(engine, NULL); 561 cancel_reset_timeout(engine); 562 if (err == 0) /* timeouts only generated on gen8+ */ 563 goto skip; 564 565 count = 0; 566 do { 567 struct i915_request *last = NULL; 568 int i; 569 570 if (!wait_for_idle(engine)) { 571 pr_err("%s failed to idle before reset\n", 572 engine->name); 573 err = -EIO; 574 break; 575 } 576 577 for (i = 0; i < count % 15; i++) { 578 struct i915_request *rq; 579 580 rq = intel_context_create_request(ce); 581 if (IS_ERR(rq)) { 582 struct drm_printer p = 583 drm_info_printer(gt->i915->drm.dev); 584 intel_engine_dump(engine, &p, 585 "%s(%s): failed to submit request\n", 586 __func__, 587 engine->name); 588 589 GEM_TRACE("%s(%s): failed to submit request\n", 590 __func__, 591 engine->name); 592 GEM_TRACE_DUMP(); 593 594 intel_gt_set_wedged(gt); 595 if (last) 596 i915_request_put(last); 597 598 err = PTR_ERR(rq); 599 goto out; 600 } 601 602 if (last) 603 i915_request_put(last); 604 last = i915_request_get(rq); 605 i915_request_add(rq); 606 } 607 608 if (count & 1) { 609 err = intel_engine_reset(engine, NULL); 610 if (err) { 611 GEM_TRACE_ERR("intel_engine_reset(%s) failed, err:%d\n", 612 engine->name, err); 613 GEM_TRACE_DUMP(); 614 i915_request_put(last); 615 break; 616 } 617 } else { 618 force_reset_timeout(engine); 619 err = intel_engine_reset(engine, NULL); 620 cancel_reset_timeout(engine); 621 if (err != -ETIMEDOUT) { 622 pr_err("intel_engine_reset(%s) did not fail, err:%d\n", 623 engine->name, err); 624 i915_request_put(last); 625 break; 626 } 627 } 628 629 err = 0; 630 if (last) { 631 if (i915_request_wait(last, 0, HZ / 2) < 0) { 632 struct drm_printer p = 633 drm_info_printer(gt->i915->drm.dev); 634 635 intel_engine_dump(engine, &p, 636 "%s(%s): failed to complete request\n", 637 __func__, 638 engine->name); 639 640 GEM_TRACE("%s(%s): failed to complete request\n", 641 __func__, 642 engine->name); 643 GEM_TRACE_DUMP(); 644 645 err = -EIO; 646 } 647 i915_request_put(last); 648 } 649 count++; 650 } while (err == 0 && time_before(jiffies, end_time)); 651 out: 652 pr_info("%s(%s): %d resets\n", __func__, engine->name, count); 653 skip: 654 clear_bit(I915_RESET_ENGINE + id, >->reset.flags); 655 st_engine_heartbeat_enable(engine); 656 intel_context_put(ce); 657 658 if (igt_flush_test(gt->i915)) 659 err = -EIO; 660 if (err) 661 return err; 662 } 663 664 return 0; 665 } 666 667 static int __igt_reset_engine(struct intel_gt *gt, bool active) 668 { 669 struct i915_gpu_error *global = >->i915->gpu_error; 670 struct intel_engine_cs *engine; 671 enum intel_engine_id id; 672 struct hang h; 673 int err = 0; 674 675 /* Check that we can issue an engine reset on an idle engine (no-op) */ 676 677 if (!intel_has_reset_engine(gt)) 678 return 0; 679 680 if (active) { 681 err = hang_init(&h, gt); 682 if (err) 683 return err; 684 } 685 686 for_each_engine(engine, gt, id) { 687 unsigned int reset_count, reset_engine_count; 688 unsigned long count; 689 IGT_TIMEOUT(end_time); 690 691 if (active && !intel_engine_can_store_dword(engine)) 692 continue; 693 694 if (!wait_for_idle(engine)) { 695 pr_err("%s failed to idle before reset\n", 696 engine->name); 697 err = -EIO; 698 break; 699 } 700 701 reset_count = i915_reset_count(global); 702 reset_engine_count = i915_reset_engine_count(global, engine); 703 704 st_engine_heartbeat_disable(engine); 705 set_bit(I915_RESET_ENGINE + id, >->reset.flags); 706 count = 0; 707 do { 708 if (active) { 709 struct i915_request *rq; 710 711 rq = hang_create_request(&h, engine); 712 if (IS_ERR(rq)) { 713 err = PTR_ERR(rq); 714 break; 715 } 716 717 i915_request_get(rq); 718 i915_request_add(rq); 719 720 if (!wait_until_running(&h, rq)) { 721 struct drm_printer p = drm_info_printer(gt->i915->drm.dev); 722 723 pr_err("%s: Failed to start request %llx, at %x\n", 724 __func__, rq->fence.seqno, hws_seqno(&h, rq)); 725 intel_engine_dump(engine, &p, 726 "%s\n", engine->name); 727 728 i915_request_put(rq); 729 err = -EIO; 730 break; 731 } 732 733 i915_request_put(rq); 734 } 735 736 err = intel_engine_reset(engine, NULL); 737 if (err) { 738 pr_err("intel_engine_reset(%s) failed, err:%d\n", 739 engine->name, err); 740 break; 741 } 742 743 if (i915_reset_count(global) != reset_count) { 744 pr_err("Full GPU reset recorded! (engine reset expected)\n"); 745 err = -EINVAL; 746 break; 747 } 748 749 if (i915_reset_engine_count(global, engine) != 750 ++reset_engine_count) { 751 pr_err("%s engine reset not recorded!\n", 752 engine->name); 753 err = -EINVAL; 754 break; 755 } 756 757 count++; 758 } while (time_before(jiffies, end_time)); 759 clear_bit(I915_RESET_ENGINE + id, >->reset.flags); 760 st_engine_heartbeat_enable(engine); 761 pr_info("%s: Completed %lu %s resets\n", 762 engine->name, count, active ? "active" : "idle"); 763 764 if (err) 765 break; 766 767 err = igt_flush_test(gt->i915); 768 if (err) 769 break; 770 } 771 772 if (intel_gt_is_wedged(gt)) 773 err = -EIO; 774 775 if (active) 776 hang_fini(&h); 777 778 return err; 779 } 780 781 static int igt_reset_idle_engine(void *arg) 782 { 783 return __igt_reset_engine(arg, false); 784 } 785 786 static int igt_reset_active_engine(void *arg) 787 { 788 return __igt_reset_engine(arg, true); 789 } 790 791 struct active_engine { 792 struct task_struct *task; 793 struct intel_engine_cs *engine; 794 unsigned long resets; 795 unsigned int flags; 796 }; 797 798 #define TEST_ACTIVE BIT(0) 799 #define TEST_OTHERS BIT(1) 800 #define TEST_SELF BIT(2) 801 #define TEST_PRIORITY BIT(3) 802 803 static int active_request_put(struct i915_request *rq) 804 { 805 int err = 0; 806 807 if (!rq) 808 return 0; 809 810 if (i915_request_wait(rq, 0, 5 * HZ) < 0) { 811 GEM_TRACE("%s timed out waiting for completion of fence %llx:%lld\n", 812 rq->engine->name, 813 rq->fence.context, 814 rq->fence.seqno); 815 GEM_TRACE_DUMP(); 816 817 intel_gt_set_wedged(rq->engine->gt); 818 err = -EIO; 819 } 820 821 i915_request_put(rq); 822 823 return err; 824 } 825 826 static int active_engine(void *data) 827 { 828 I915_RND_STATE(prng); 829 struct active_engine *arg = data; 830 struct intel_engine_cs *engine = arg->engine; 831 struct i915_request *rq[8] = {}; 832 struct intel_context *ce[ARRAY_SIZE(rq)]; 833 unsigned long count; 834 int err = 0; 835 836 for (count = 0; count < ARRAY_SIZE(ce); count++) { 837 ce[count] = intel_context_create(engine); 838 if (IS_ERR(ce[count])) { 839 err = PTR_ERR(ce[count]); 840 while (--count) 841 intel_context_put(ce[count]); 842 return err; 843 } 844 } 845 846 count = 0; 847 while (!kthread_should_stop()) { 848 unsigned int idx = count++ & (ARRAY_SIZE(rq) - 1); 849 struct i915_request *old = rq[idx]; 850 struct i915_request *new; 851 852 new = intel_context_create_request(ce[idx]); 853 if (IS_ERR(new)) { 854 err = PTR_ERR(new); 855 break; 856 } 857 858 rq[idx] = i915_request_get(new); 859 i915_request_add(new); 860 861 if (engine->schedule && arg->flags & TEST_PRIORITY) { 862 struct i915_sched_attr attr = { 863 .priority = 864 i915_prandom_u32_max_state(512, &prng), 865 }; 866 engine->schedule(rq[idx], &attr); 867 } 868 869 err = active_request_put(old); 870 if (err) 871 break; 872 873 cond_resched(); 874 } 875 876 for (count = 0; count < ARRAY_SIZE(rq); count++) { 877 int err__ = active_request_put(rq[count]); 878 879 /* Keep the first error */ 880 if (!err) 881 err = err__; 882 883 intel_context_put(ce[count]); 884 } 885 886 return err; 887 } 888 889 static int __igt_reset_engines(struct intel_gt *gt, 890 const char *test_name, 891 unsigned int flags) 892 { 893 struct i915_gpu_error *global = >->i915->gpu_error; 894 struct intel_engine_cs *engine, *other; 895 enum intel_engine_id id, tmp; 896 struct hang h; 897 int err = 0; 898 899 /* Check that issuing a reset on one engine does not interfere 900 * with any other engine. 901 */ 902 903 if (!intel_has_reset_engine(gt)) 904 return 0; 905 906 if (flags & TEST_ACTIVE) { 907 err = hang_init(&h, gt); 908 if (err) 909 return err; 910 911 if (flags & TEST_PRIORITY) 912 h.ctx->sched.priority = 1024; 913 } 914 915 for_each_engine(engine, gt, id) { 916 struct active_engine threads[I915_NUM_ENGINES] = {}; 917 unsigned long device = i915_reset_count(global); 918 unsigned long count = 0, reported; 919 IGT_TIMEOUT(end_time); 920 921 if (flags & TEST_ACTIVE && 922 !intel_engine_can_store_dword(engine)) 923 continue; 924 925 if (!wait_for_idle(engine)) { 926 pr_err("i915_reset_engine(%s:%s): failed to idle before reset\n", 927 engine->name, test_name); 928 err = -EIO; 929 break; 930 } 931 932 memset(threads, 0, sizeof(threads)); 933 for_each_engine(other, gt, tmp) { 934 struct task_struct *tsk; 935 936 threads[tmp].resets = 937 i915_reset_engine_count(global, other); 938 939 if (other == engine && !(flags & TEST_SELF)) 940 continue; 941 942 if (other != engine && !(flags & TEST_OTHERS)) 943 continue; 944 945 threads[tmp].engine = other; 946 threads[tmp].flags = flags; 947 948 tsk = kthread_run(active_engine, &threads[tmp], 949 "igt/%s", other->name); 950 if (IS_ERR(tsk)) { 951 err = PTR_ERR(tsk); 952 goto unwind; 953 } 954 955 threads[tmp].task = tsk; 956 get_task_struct(tsk); 957 } 958 959 yield(); /* start all threads before we begin */ 960 961 st_engine_heartbeat_disable(engine); 962 set_bit(I915_RESET_ENGINE + id, >->reset.flags); 963 do { 964 struct i915_request *rq = NULL; 965 966 if (flags & TEST_ACTIVE) { 967 rq = hang_create_request(&h, engine); 968 if (IS_ERR(rq)) { 969 err = PTR_ERR(rq); 970 break; 971 } 972 973 i915_request_get(rq); 974 i915_request_add(rq); 975 976 if (!wait_until_running(&h, rq)) { 977 struct drm_printer p = drm_info_printer(gt->i915->drm.dev); 978 979 pr_err("%s: Failed to start request %llx, at %x\n", 980 __func__, rq->fence.seqno, hws_seqno(&h, rq)); 981 intel_engine_dump(engine, &p, 982 "%s\n", engine->name); 983 984 i915_request_put(rq); 985 err = -EIO; 986 break; 987 } 988 } 989 990 err = intel_engine_reset(engine, NULL); 991 if (err) { 992 pr_err("i915_reset_engine(%s:%s): failed, err=%d\n", 993 engine->name, test_name, err); 994 break; 995 } 996 997 count++; 998 999 if (rq) { 1000 if (rq->fence.error != -EIO) { 1001 pr_err("i915_reset_engine(%s:%s):" 1002 " failed to reset request %llx:%lld\n", 1003 engine->name, test_name, 1004 rq->fence.context, 1005 rq->fence.seqno); 1006 i915_request_put(rq); 1007 1008 GEM_TRACE_DUMP(); 1009 intel_gt_set_wedged(gt); 1010 err = -EIO; 1011 break; 1012 } 1013 1014 if (i915_request_wait(rq, 0, HZ / 5) < 0) { 1015 struct drm_printer p = 1016 drm_info_printer(gt->i915->drm.dev); 1017 1018 pr_err("i915_reset_engine(%s:%s):" 1019 " failed to complete request %llx:%lld after reset\n", 1020 engine->name, test_name, 1021 rq->fence.context, 1022 rq->fence.seqno); 1023 intel_engine_dump(engine, &p, 1024 "%s\n", engine->name); 1025 i915_request_put(rq); 1026 1027 GEM_TRACE_DUMP(); 1028 intel_gt_set_wedged(gt); 1029 err = -EIO; 1030 break; 1031 } 1032 1033 i915_request_put(rq); 1034 } 1035 1036 if (!(flags & TEST_SELF) && !wait_for_idle(engine)) { 1037 struct drm_printer p = 1038 drm_info_printer(gt->i915->drm.dev); 1039 1040 pr_err("i915_reset_engine(%s:%s):" 1041 " failed to idle after reset\n", 1042 engine->name, test_name); 1043 intel_engine_dump(engine, &p, 1044 "%s\n", engine->name); 1045 1046 err = -EIO; 1047 break; 1048 } 1049 } while (time_before(jiffies, end_time)); 1050 clear_bit(I915_RESET_ENGINE + id, >->reset.flags); 1051 st_engine_heartbeat_enable(engine); 1052 1053 pr_info("i915_reset_engine(%s:%s): %lu resets\n", 1054 engine->name, test_name, count); 1055 1056 reported = i915_reset_engine_count(global, engine); 1057 reported -= threads[engine->id].resets; 1058 if (reported != count) { 1059 pr_err("i915_reset_engine(%s:%s): reset %lu times, but reported %lu\n", 1060 engine->name, test_name, count, reported); 1061 if (!err) 1062 err = -EINVAL; 1063 } 1064 1065 unwind: 1066 for_each_engine(other, gt, tmp) { 1067 int ret; 1068 1069 if (!threads[tmp].task) 1070 continue; 1071 1072 ret = kthread_stop(threads[tmp].task); 1073 if (ret) { 1074 pr_err("kthread for other engine %s failed, err=%d\n", 1075 other->name, ret); 1076 if (!err) 1077 err = ret; 1078 } 1079 put_task_struct(threads[tmp].task); 1080 1081 if (other->uabi_class != engine->uabi_class && 1082 threads[tmp].resets != 1083 i915_reset_engine_count(global, other)) { 1084 pr_err("Innocent engine %s was reset (count=%ld)\n", 1085 other->name, 1086 i915_reset_engine_count(global, other) - 1087 threads[tmp].resets); 1088 if (!err) 1089 err = -EINVAL; 1090 } 1091 } 1092 1093 if (device != i915_reset_count(global)) { 1094 pr_err("Global reset (count=%ld)!\n", 1095 i915_reset_count(global) - device); 1096 if (!err) 1097 err = -EINVAL; 1098 } 1099 1100 if (err) 1101 break; 1102 1103 err = igt_flush_test(gt->i915); 1104 if (err) 1105 break; 1106 } 1107 1108 if (intel_gt_is_wedged(gt)) 1109 err = -EIO; 1110 1111 if (flags & TEST_ACTIVE) 1112 hang_fini(&h); 1113 1114 return err; 1115 } 1116 1117 static int igt_reset_engines(void *arg) 1118 { 1119 static const struct { 1120 const char *name; 1121 unsigned int flags; 1122 } phases[] = { 1123 { "idle", 0 }, 1124 { "active", TEST_ACTIVE }, 1125 { "others-idle", TEST_OTHERS }, 1126 { "others-active", TEST_OTHERS | TEST_ACTIVE }, 1127 { 1128 "others-priority", 1129 TEST_OTHERS | TEST_ACTIVE | TEST_PRIORITY 1130 }, 1131 { 1132 "self-priority", 1133 TEST_ACTIVE | TEST_PRIORITY | TEST_SELF, 1134 }, 1135 { } 1136 }; 1137 struct intel_gt *gt = arg; 1138 typeof(*phases) *p; 1139 int err; 1140 1141 for (p = phases; p->name; p++) { 1142 if (p->flags & TEST_PRIORITY) { 1143 if (!(gt->i915->caps.scheduler & I915_SCHEDULER_CAP_PRIORITY)) 1144 continue; 1145 } 1146 1147 err = __igt_reset_engines(arg, p->name, p->flags); 1148 if (err) 1149 return err; 1150 } 1151 1152 return 0; 1153 } 1154 1155 static u32 fake_hangcheck(struct intel_gt *gt, intel_engine_mask_t mask) 1156 { 1157 u32 count = i915_reset_count(>->i915->gpu_error); 1158 1159 intel_gt_reset(gt, mask, NULL); 1160 1161 return count; 1162 } 1163 1164 static int igt_reset_wait(void *arg) 1165 { 1166 struct intel_gt *gt = arg; 1167 struct i915_gpu_error *global = >->i915->gpu_error; 1168 struct intel_engine_cs *engine = gt->engine[RCS0]; 1169 struct i915_request *rq; 1170 unsigned int reset_count; 1171 struct hang h; 1172 long timeout; 1173 int err; 1174 1175 if (!engine || !intel_engine_can_store_dword(engine)) 1176 return 0; 1177 1178 /* Check that we detect a stuck waiter and issue a reset */ 1179 1180 igt_global_reset_lock(gt); 1181 1182 err = hang_init(&h, gt); 1183 if (err) 1184 goto unlock; 1185 1186 rq = hang_create_request(&h, engine); 1187 if (IS_ERR(rq)) { 1188 err = PTR_ERR(rq); 1189 goto fini; 1190 } 1191 1192 i915_request_get(rq); 1193 i915_request_add(rq); 1194 1195 if (!wait_until_running(&h, rq)) { 1196 struct drm_printer p = drm_info_printer(gt->i915->drm.dev); 1197 1198 pr_err("%s: Failed to start request %llx, at %x\n", 1199 __func__, rq->fence.seqno, hws_seqno(&h, rq)); 1200 intel_engine_dump(rq->engine, &p, "%s\n", rq->engine->name); 1201 1202 intel_gt_set_wedged(gt); 1203 1204 err = -EIO; 1205 goto out_rq; 1206 } 1207 1208 reset_count = fake_hangcheck(gt, ALL_ENGINES); 1209 1210 timeout = i915_request_wait(rq, 0, 10); 1211 if (timeout < 0) { 1212 pr_err("i915_request_wait failed on a stuck request: err=%ld\n", 1213 timeout); 1214 err = timeout; 1215 goto out_rq; 1216 } 1217 1218 if (i915_reset_count(global) == reset_count) { 1219 pr_err("No GPU reset recorded!\n"); 1220 err = -EINVAL; 1221 goto out_rq; 1222 } 1223 1224 out_rq: 1225 i915_request_put(rq); 1226 fini: 1227 hang_fini(&h); 1228 unlock: 1229 igt_global_reset_unlock(gt); 1230 1231 if (intel_gt_is_wedged(gt)) 1232 return -EIO; 1233 1234 return err; 1235 } 1236 1237 struct evict_vma { 1238 struct completion completion; 1239 struct i915_vma *vma; 1240 }; 1241 1242 static int evict_vma(void *data) 1243 { 1244 struct evict_vma *arg = data; 1245 struct i915_address_space *vm = arg->vma->vm; 1246 struct drm_mm_node evict = arg->vma->node; 1247 int err; 1248 1249 complete(&arg->completion); 1250 1251 mutex_lock(&vm->mutex); 1252 err = i915_gem_evict_for_node(vm, &evict, 0); 1253 mutex_unlock(&vm->mutex); 1254 1255 return err; 1256 } 1257 1258 static int evict_fence(void *data) 1259 { 1260 struct evict_vma *arg = data; 1261 int err; 1262 1263 complete(&arg->completion); 1264 1265 /* Mark the fence register as dirty to force the mmio update. */ 1266 err = i915_gem_object_set_tiling(arg->vma->obj, I915_TILING_Y, 512); 1267 if (err) { 1268 pr_err("Invalid Y-tiling settings; err:%d\n", err); 1269 return err; 1270 } 1271 1272 err = i915_vma_pin(arg->vma, 0, 0, PIN_GLOBAL | PIN_MAPPABLE); 1273 if (err) { 1274 pr_err("Unable to pin vma for Y-tiled fence; err:%d\n", err); 1275 return err; 1276 } 1277 1278 err = i915_vma_pin_fence(arg->vma); 1279 i915_vma_unpin(arg->vma); 1280 if (err) { 1281 pr_err("Unable to pin Y-tiled fence; err:%d\n", err); 1282 return err; 1283 } 1284 1285 i915_vma_unpin_fence(arg->vma); 1286 1287 return 0; 1288 } 1289 1290 static int __igt_reset_evict_vma(struct intel_gt *gt, 1291 struct i915_address_space *vm, 1292 int (*fn)(void *), 1293 unsigned int flags) 1294 { 1295 struct intel_engine_cs *engine = gt->engine[RCS0]; 1296 struct drm_i915_gem_object *obj; 1297 struct task_struct *tsk = NULL; 1298 struct i915_request *rq; 1299 struct evict_vma arg; 1300 struct hang h; 1301 unsigned int pin_flags; 1302 int err; 1303 1304 if (!gt->ggtt->num_fences && flags & EXEC_OBJECT_NEEDS_FENCE) 1305 return 0; 1306 1307 if (!engine || !intel_engine_can_store_dword(engine)) 1308 return 0; 1309 1310 /* Check that we can recover an unbind stuck on a hanging request */ 1311 1312 err = hang_init(&h, gt); 1313 if (err) 1314 return err; 1315 1316 obj = i915_gem_object_create_internal(gt->i915, SZ_1M); 1317 if (IS_ERR(obj)) { 1318 err = PTR_ERR(obj); 1319 goto fini; 1320 } 1321 1322 if (flags & EXEC_OBJECT_NEEDS_FENCE) { 1323 err = i915_gem_object_set_tiling(obj, I915_TILING_X, 512); 1324 if (err) { 1325 pr_err("Invalid X-tiling settings; err:%d\n", err); 1326 goto out_obj; 1327 } 1328 } 1329 1330 arg.vma = i915_vma_instance(obj, vm, NULL); 1331 if (IS_ERR(arg.vma)) { 1332 err = PTR_ERR(arg.vma); 1333 goto out_obj; 1334 } 1335 1336 rq = hang_create_request(&h, engine); 1337 if (IS_ERR(rq)) { 1338 err = PTR_ERR(rq); 1339 goto out_obj; 1340 } 1341 1342 pin_flags = i915_vma_is_ggtt(arg.vma) ? PIN_GLOBAL : PIN_USER; 1343 1344 if (flags & EXEC_OBJECT_NEEDS_FENCE) 1345 pin_flags |= PIN_MAPPABLE; 1346 1347 err = i915_vma_pin(arg.vma, 0, 0, pin_flags); 1348 if (err) { 1349 i915_request_add(rq); 1350 goto out_obj; 1351 } 1352 1353 if (flags & EXEC_OBJECT_NEEDS_FENCE) { 1354 err = i915_vma_pin_fence(arg.vma); 1355 if (err) { 1356 pr_err("Unable to pin X-tiled fence; err:%d\n", err); 1357 i915_vma_unpin(arg.vma); 1358 i915_request_add(rq); 1359 goto out_obj; 1360 } 1361 } 1362 1363 i915_vma_lock(arg.vma); 1364 err = i915_request_await_object(rq, arg.vma->obj, 1365 flags & EXEC_OBJECT_WRITE); 1366 if (err == 0) 1367 err = i915_vma_move_to_active(arg.vma, rq, flags); 1368 i915_vma_unlock(arg.vma); 1369 1370 if (flags & EXEC_OBJECT_NEEDS_FENCE) 1371 i915_vma_unpin_fence(arg.vma); 1372 i915_vma_unpin(arg.vma); 1373 1374 i915_request_get(rq); 1375 i915_request_add(rq); 1376 if (err) 1377 goto out_rq; 1378 1379 if (!wait_until_running(&h, rq)) { 1380 struct drm_printer p = drm_info_printer(gt->i915->drm.dev); 1381 1382 pr_err("%s: Failed to start request %llx, at %x\n", 1383 __func__, rq->fence.seqno, hws_seqno(&h, rq)); 1384 intel_engine_dump(rq->engine, &p, "%s\n", rq->engine->name); 1385 1386 intel_gt_set_wedged(gt); 1387 goto out_reset; 1388 } 1389 1390 init_completion(&arg.completion); 1391 1392 tsk = kthread_run(fn, &arg, "igt/evict_vma"); 1393 if (IS_ERR(tsk)) { 1394 err = PTR_ERR(tsk); 1395 tsk = NULL; 1396 goto out_reset; 1397 } 1398 get_task_struct(tsk); 1399 1400 wait_for_completion(&arg.completion); 1401 1402 if (wait_for(!list_empty(&rq->fence.cb_list), 10)) { 1403 struct drm_printer p = drm_info_printer(gt->i915->drm.dev); 1404 1405 pr_err("igt/evict_vma kthread did not wait\n"); 1406 intel_engine_dump(rq->engine, &p, "%s\n", rq->engine->name); 1407 1408 intel_gt_set_wedged(gt); 1409 goto out_reset; 1410 } 1411 1412 out_reset: 1413 igt_global_reset_lock(gt); 1414 fake_hangcheck(gt, rq->engine->mask); 1415 igt_global_reset_unlock(gt); 1416 1417 if (tsk) { 1418 struct intel_wedge_me w; 1419 1420 /* The reset, even indirectly, should take less than 10ms. */ 1421 intel_wedge_on_timeout(&w, gt, HZ / 10 /* 100ms */) 1422 err = kthread_stop(tsk); 1423 1424 put_task_struct(tsk); 1425 } 1426 1427 out_rq: 1428 i915_request_put(rq); 1429 out_obj: 1430 i915_gem_object_put(obj); 1431 fini: 1432 hang_fini(&h); 1433 if (intel_gt_is_wedged(gt)) 1434 return -EIO; 1435 1436 return err; 1437 } 1438 1439 static int igt_reset_evict_ggtt(void *arg) 1440 { 1441 struct intel_gt *gt = arg; 1442 1443 return __igt_reset_evict_vma(gt, >->ggtt->vm, 1444 evict_vma, EXEC_OBJECT_WRITE); 1445 } 1446 1447 static int igt_reset_evict_ppgtt(void *arg) 1448 { 1449 struct intel_gt *gt = arg; 1450 struct i915_ppgtt *ppgtt; 1451 int err; 1452 1453 /* aliasing == global gtt locking, covered above */ 1454 if (INTEL_PPGTT(gt->i915) < INTEL_PPGTT_FULL) 1455 return 0; 1456 1457 ppgtt = i915_ppgtt_create(gt); 1458 if (IS_ERR(ppgtt)) 1459 return PTR_ERR(ppgtt); 1460 1461 err = __igt_reset_evict_vma(gt, &ppgtt->vm, 1462 evict_vma, EXEC_OBJECT_WRITE); 1463 i915_vm_put(&ppgtt->vm); 1464 1465 return err; 1466 } 1467 1468 static int igt_reset_evict_fence(void *arg) 1469 { 1470 struct intel_gt *gt = arg; 1471 1472 return __igt_reset_evict_vma(gt, >->ggtt->vm, 1473 evict_fence, EXEC_OBJECT_NEEDS_FENCE); 1474 } 1475 1476 static int wait_for_others(struct intel_gt *gt, 1477 struct intel_engine_cs *exclude) 1478 { 1479 struct intel_engine_cs *engine; 1480 enum intel_engine_id id; 1481 1482 for_each_engine(engine, gt, id) { 1483 if (engine == exclude) 1484 continue; 1485 1486 if (!wait_for_idle(engine)) 1487 return -EIO; 1488 } 1489 1490 return 0; 1491 } 1492 1493 static int igt_reset_queue(void *arg) 1494 { 1495 struct intel_gt *gt = arg; 1496 struct i915_gpu_error *global = >->i915->gpu_error; 1497 struct intel_engine_cs *engine; 1498 enum intel_engine_id id; 1499 struct hang h; 1500 int err; 1501 1502 /* Check that we replay pending requests following a hang */ 1503 1504 igt_global_reset_lock(gt); 1505 1506 err = hang_init(&h, gt); 1507 if (err) 1508 goto unlock; 1509 1510 for_each_engine(engine, gt, id) { 1511 struct i915_request *prev; 1512 IGT_TIMEOUT(end_time); 1513 unsigned int count; 1514 1515 if (!intel_engine_can_store_dword(engine)) 1516 continue; 1517 1518 prev = hang_create_request(&h, engine); 1519 if (IS_ERR(prev)) { 1520 err = PTR_ERR(prev); 1521 goto fini; 1522 } 1523 1524 i915_request_get(prev); 1525 i915_request_add(prev); 1526 1527 count = 0; 1528 do { 1529 struct i915_request *rq; 1530 unsigned int reset_count; 1531 1532 rq = hang_create_request(&h, engine); 1533 if (IS_ERR(rq)) { 1534 err = PTR_ERR(rq); 1535 goto fini; 1536 } 1537 1538 i915_request_get(rq); 1539 i915_request_add(rq); 1540 1541 /* 1542 * XXX We don't handle resetting the kernel context 1543 * very well. If we trigger a device reset twice in 1544 * quick succession while the kernel context is 1545 * executing, we may end up skipping the breadcrumb. 1546 * This is really only a problem for the selftest as 1547 * normally there is a large interlude between resets 1548 * (hangcheck), or we focus on resetting just one 1549 * engine and so avoid repeatedly resetting innocents. 1550 */ 1551 err = wait_for_others(gt, engine); 1552 if (err) { 1553 pr_err("%s(%s): Failed to idle other inactive engines after device reset\n", 1554 __func__, engine->name); 1555 i915_request_put(rq); 1556 i915_request_put(prev); 1557 1558 GEM_TRACE_DUMP(); 1559 intel_gt_set_wedged(gt); 1560 goto fini; 1561 } 1562 1563 if (!wait_until_running(&h, prev)) { 1564 struct drm_printer p = drm_info_printer(gt->i915->drm.dev); 1565 1566 pr_err("%s(%s): Failed to start request %llx, at %x\n", 1567 __func__, engine->name, 1568 prev->fence.seqno, hws_seqno(&h, prev)); 1569 intel_engine_dump(engine, &p, 1570 "%s\n", engine->name); 1571 1572 i915_request_put(rq); 1573 i915_request_put(prev); 1574 1575 intel_gt_set_wedged(gt); 1576 1577 err = -EIO; 1578 goto fini; 1579 } 1580 1581 reset_count = fake_hangcheck(gt, BIT(id)); 1582 1583 if (prev->fence.error != -EIO) { 1584 pr_err("GPU reset not recorded on hanging request [fence.error=%d]!\n", 1585 prev->fence.error); 1586 i915_request_put(rq); 1587 i915_request_put(prev); 1588 err = -EINVAL; 1589 goto fini; 1590 } 1591 1592 if (rq->fence.error) { 1593 pr_err("Fence error status not zero [%d] after unrelated reset\n", 1594 rq->fence.error); 1595 i915_request_put(rq); 1596 i915_request_put(prev); 1597 err = -EINVAL; 1598 goto fini; 1599 } 1600 1601 if (i915_reset_count(global) == reset_count) { 1602 pr_err("No GPU reset recorded!\n"); 1603 i915_request_put(rq); 1604 i915_request_put(prev); 1605 err = -EINVAL; 1606 goto fini; 1607 } 1608 1609 i915_request_put(prev); 1610 prev = rq; 1611 count++; 1612 } while (time_before(jiffies, end_time)); 1613 pr_info("%s: Completed %d queued resets\n", 1614 engine->name, count); 1615 1616 *h.batch = MI_BATCH_BUFFER_END; 1617 intel_gt_chipset_flush(engine->gt); 1618 1619 i915_request_put(prev); 1620 1621 err = igt_flush_test(gt->i915); 1622 if (err) 1623 break; 1624 } 1625 1626 fini: 1627 hang_fini(&h); 1628 unlock: 1629 igt_global_reset_unlock(gt); 1630 1631 if (intel_gt_is_wedged(gt)) 1632 return -EIO; 1633 1634 return err; 1635 } 1636 1637 static int igt_handle_error(void *arg) 1638 { 1639 struct intel_gt *gt = arg; 1640 struct i915_gpu_error *global = >->i915->gpu_error; 1641 struct intel_engine_cs *engine = gt->engine[RCS0]; 1642 struct hang h; 1643 struct i915_request *rq; 1644 struct i915_gpu_coredump *error; 1645 int err; 1646 1647 /* Check that we can issue a global GPU and engine reset */ 1648 1649 if (!intel_has_reset_engine(gt)) 1650 return 0; 1651 1652 if (!engine || !intel_engine_can_store_dword(engine)) 1653 return 0; 1654 1655 err = hang_init(&h, gt); 1656 if (err) 1657 return err; 1658 1659 rq = hang_create_request(&h, engine); 1660 if (IS_ERR(rq)) { 1661 err = PTR_ERR(rq); 1662 goto err_fini; 1663 } 1664 1665 i915_request_get(rq); 1666 i915_request_add(rq); 1667 1668 if (!wait_until_running(&h, rq)) { 1669 struct drm_printer p = drm_info_printer(gt->i915->drm.dev); 1670 1671 pr_err("%s: Failed to start request %llx, at %x\n", 1672 __func__, rq->fence.seqno, hws_seqno(&h, rq)); 1673 intel_engine_dump(rq->engine, &p, "%s\n", rq->engine->name); 1674 1675 intel_gt_set_wedged(gt); 1676 1677 err = -EIO; 1678 goto err_request; 1679 } 1680 1681 /* Temporarily disable error capture */ 1682 error = xchg(&global->first_error, (void *)-1); 1683 1684 intel_gt_handle_error(gt, engine->mask, 0, NULL); 1685 1686 xchg(&global->first_error, error); 1687 1688 if (rq->fence.error != -EIO) { 1689 pr_err("Guilty request not identified!\n"); 1690 err = -EINVAL; 1691 goto err_request; 1692 } 1693 1694 err_request: 1695 i915_request_put(rq); 1696 err_fini: 1697 hang_fini(&h); 1698 return err; 1699 } 1700 1701 static int __igt_atomic_reset_engine(struct intel_engine_cs *engine, 1702 const struct igt_atomic_section *p, 1703 const char *mode) 1704 { 1705 struct tasklet_struct * const t = &engine->execlists.tasklet; 1706 int err; 1707 1708 GEM_TRACE("i915_reset_engine(%s:%s) under %s\n", 1709 engine->name, mode, p->name); 1710 1711 if (t->func) 1712 tasklet_disable(t); 1713 if (strcmp(p->name, "softirq")) 1714 local_bh_disable(); 1715 p->critical_section_begin(); 1716 1717 err = __intel_engine_reset_bh(engine, NULL); 1718 1719 p->critical_section_end(); 1720 if (strcmp(p->name, "softirq")) 1721 local_bh_enable(); 1722 if (t->func) { 1723 tasklet_enable(t); 1724 tasklet_hi_schedule(t); 1725 } 1726 1727 if (err) 1728 pr_err("i915_reset_engine(%s:%s) failed under %s\n", 1729 engine->name, mode, p->name); 1730 1731 return err; 1732 } 1733 1734 static int igt_atomic_reset_engine(struct intel_engine_cs *engine, 1735 const struct igt_atomic_section *p) 1736 { 1737 struct i915_request *rq; 1738 struct hang h; 1739 int err; 1740 1741 err = __igt_atomic_reset_engine(engine, p, "idle"); 1742 if (err) 1743 return err; 1744 1745 err = hang_init(&h, engine->gt); 1746 if (err) 1747 return err; 1748 1749 rq = hang_create_request(&h, engine); 1750 if (IS_ERR(rq)) { 1751 err = PTR_ERR(rq); 1752 goto out; 1753 } 1754 1755 i915_request_get(rq); 1756 i915_request_add(rq); 1757 1758 if (wait_until_running(&h, rq)) { 1759 err = __igt_atomic_reset_engine(engine, p, "active"); 1760 } else { 1761 pr_err("%s(%s): Failed to start request %llx, at %x\n", 1762 __func__, engine->name, 1763 rq->fence.seqno, hws_seqno(&h, rq)); 1764 intel_gt_set_wedged(engine->gt); 1765 err = -EIO; 1766 } 1767 1768 if (err == 0) { 1769 struct intel_wedge_me w; 1770 1771 intel_wedge_on_timeout(&w, engine->gt, HZ / 20 /* 50ms */) 1772 i915_request_wait(rq, 0, MAX_SCHEDULE_TIMEOUT); 1773 if (intel_gt_is_wedged(engine->gt)) 1774 err = -EIO; 1775 } 1776 1777 i915_request_put(rq); 1778 out: 1779 hang_fini(&h); 1780 return err; 1781 } 1782 1783 static int igt_reset_engines_atomic(void *arg) 1784 { 1785 struct intel_gt *gt = arg; 1786 const typeof(*igt_atomic_phases) *p; 1787 int err = 0; 1788 1789 /* Check that the engines resets are usable from atomic context */ 1790 1791 if (!intel_has_reset_engine(gt)) 1792 return 0; 1793 1794 if (intel_uc_uses_guc_submission(>->uc)) 1795 return 0; 1796 1797 igt_global_reset_lock(gt); 1798 1799 /* Flush any requests before we get started and check basics */ 1800 if (!igt_force_reset(gt)) 1801 goto unlock; 1802 1803 for (p = igt_atomic_phases; p->name; p++) { 1804 struct intel_engine_cs *engine; 1805 enum intel_engine_id id; 1806 1807 for_each_engine(engine, gt, id) { 1808 err = igt_atomic_reset_engine(engine, p); 1809 if (err) 1810 goto out; 1811 } 1812 } 1813 1814 out: 1815 /* As we poke around the guts, do a full reset before continuing. */ 1816 igt_force_reset(gt); 1817 unlock: 1818 igt_global_reset_unlock(gt); 1819 1820 return err; 1821 } 1822 1823 int intel_hangcheck_live_selftests(struct drm_i915_private *i915) 1824 { 1825 static const struct i915_subtest tests[] = { 1826 SUBTEST(igt_hang_sanitycheck), 1827 SUBTEST(igt_reset_nop), 1828 SUBTEST(igt_reset_nop_engine), 1829 SUBTEST(igt_reset_idle_engine), 1830 SUBTEST(igt_reset_active_engine), 1831 SUBTEST(igt_reset_fail_engine), 1832 SUBTEST(igt_reset_engines), 1833 SUBTEST(igt_reset_engines_atomic), 1834 SUBTEST(igt_reset_queue), 1835 SUBTEST(igt_reset_wait), 1836 SUBTEST(igt_reset_evict_ggtt), 1837 SUBTEST(igt_reset_evict_ppgtt), 1838 SUBTEST(igt_reset_evict_fence), 1839 SUBTEST(igt_handle_error), 1840 }; 1841 struct intel_gt *gt = &i915->gt; 1842 intel_wakeref_t wakeref; 1843 int err; 1844 1845 if (!intel_has_gpu_reset(gt)) 1846 return 0; 1847 1848 if (intel_gt_is_wedged(gt)) 1849 return -EIO; /* we're long past hope of a successful reset */ 1850 1851 wakeref = intel_runtime_pm_get(gt->uncore->rpm); 1852 1853 err = intel_gt_live_subtests(tests, gt); 1854 1855 intel_runtime_pm_put(gt->uncore->rpm, wakeref); 1856 1857 return err; 1858 } 1859