1 // SPDX-License-Identifier: GPL-2.0 or MIT 2 /* Copyright 2018 Marty E. Plummer <hanetzer@startmail.com> */ 3 /* Copyright 2019 Linaro, Ltd, Rob Herring <robh@kernel.org> */ 4 /* Copyright 2023 Collabora ltd. */ 5 6 #include <linux/clk.h> 7 #include <linux/mm.h> 8 #include <linux/platform_device.h> 9 #include <linux/pm_domain.h> 10 #include <linux/pm_runtime.h> 11 #include <linux/regulator/consumer.h> 12 #include <linux/reset.h> 13 14 #include <drm/drm_drv.h> 15 #include <drm/drm_managed.h> 16 17 #include "panthor_devfreq.h" 18 #include "panthor_device.h" 19 #include "panthor_fw.h" 20 #include "panthor_gpu.h" 21 #include "panthor_mmu.h" 22 #include "panthor_regs.h" 23 #include "panthor_sched.h" 24 25 static int panthor_gpu_coherency_init(struct panthor_device *ptdev) 26 { 27 ptdev->coherent = device_get_dma_attr(ptdev->base.dev) == DEV_DMA_COHERENT; 28 29 if (!ptdev->coherent) 30 return 0; 31 32 /* Check if the ACE-Lite coherency protocol is actually supported by the GPU. 33 * ACE protocol has never been supported for command stream frontend GPUs. 34 */ 35 if ((gpu_read(ptdev, GPU_COHERENCY_FEATURES) & 36 GPU_COHERENCY_PROT_BIT(ACE_LITE))) 37 return 0; 38 39 drm_err(&ptdev->base, "Coherency not supported by the device"); 40 return -ENOTSUPP; 41 } 42 43 static int panthor_clk_init(struct panthor_device *ptdev) 44 { 45 ptdev->clks.core = devm_clk_get(ptdev->base.dev, NULL); 46 if (IS_ERR(ptdev->clks.core)) 47 return dev_err_probe(ptdev->base.dev, 48 PTR_ERR(ptdev->clks.core), 49 "get 'core' clock failed"); 50 51 ptdev->clks.stacks = devm_clk_get_optional(ptdev->base.dev, "stacks"); 52 if (IS_ERR(ptdev->clks.stacks)) 53 return dev_err_probe(ptdev->base.dev, 54 PTR_ERR(ptdev->clks.stacks), 55 "get 'stacks' clock failed"); 56 57 ptdev->clks.coregroup = devm_clk_get_optional(ptdev->base.dev, "coregroup"); 58 if (IS_ERR(ptdev->clks.coregroup)) 59 return dev_err_probe(ptdev->base.dev, 60 PTR_ERR(ptdev->clks.coregroup), 61 "get 'coregroup' clock failed"); 62 63 drm_info(&ptdev->base, "clock rate = %lu\n", clk_get_rate(ptdev->clks.core)); 64 return 0; 65 } 66 67 void panthor_device_unplug(struct panthor_device *ptdev) 68 { 69 /* This function can be called from two different path: the reset work 70 * and the platform device remove callback. drm_dev_unplug() doesn't 71 * deal with concurrent callers, so we have to protect drm_dev_unplug() 72 * calls with our own lock, and bail out if the device is already 73 * unplugged. 74 */ 75 mutex_lock(&ptdev->unplug.lock); 76 if (drm_dev_is_unplugged(&ptdev->base)) { 77 /* Someone beat us, release the lock and wait for the unplug 78 * operation to be reported as done. 79 **/ 80 mutex_unlock(&ptdev->unplug.lock); 81 wait_for_completion(&ptdev->unplug.done); 82 return; 83 } 84 85 /* Call drm_dev_unplug() so any access to HW blocks happening after 86 * that point get rejected. 87 */ 88 drm_dev_unplug(&ptdev->base); 89 90 /* We do the rest of the unplug with the unplug lock released, 91 * future callers will wait on ptdev->unplug.done anyway. 92 */ 93 mutex_unlock(&ptdev->unplug.lock); 94 95 drm_WARN_ON(&ptdev->base, pm_runtime_get_sync(ptdev->base.dev) < 0); 96 97 /* Now, try to cleanly shutdown the GPU before the device resources 98 * get reclaimed. 99 */ 100 panthor_sched_unplug(ptdev); 101 panthor_fw_unplug(ptdev); 102 panthor_mmu_unplug(ptdev); 103 panthor_gpu_unplug(ptdev); 104 105 pm_runtime_dont_use_autosuspend(ptdev->base.dev); 106 pm_runtime_put_sync_suspend(ptdev->base.dev); 107 108 /* If PM is disabled, we need to call the suspend handler manually. */ 109 if (!IS_ENABLED(CONFIG_PM)) 110 panthor_device_suspend(ptdev->base.dev); 111 112 /* Report the unplug operation as done to unblock concurrent 113 * panthor_device_unplug() callers. 114 */ 115 complete_all(&ptdev->unplug.done); 116 } 117 118 static void panthor_device_reset_cleanup(struct drm_device *ddev, void *data) 119 { 120 struct panthor_device *ptdev = container_of(ddev, struct panthor_device, base); 121 122 cancel_work_sync(&ptdev->reset.work); 123 destroy_workqueue(ptdev->reset.wq); 124 } 125 126 static void panthor_device_reset_work(struct work_struct *work) 127 { 128 struct panthor_device *ptdev = container_of(work, struct panthor_device, reset.work); 129 int ret = 0, cookie; 130 131 /* If the device is entering suspend, we don't reset. A slow reset will 132 * be forced at resume time instead. 133 */ 134 if (atomic_read(&ptdev->pm.state) != PANTHOR_DEVICE_PM_STATE_ACTIVE) 135 return; 136 137 if (!drm_dev_enter(&ptdev->base, &cookie)) 138 return; 139 140 panthor_sched_pre_reset(ptdev); 141 panthor_fw_pre_reset(ptdev, true); 142 panthor_mmu_pre_reset(ptdev); 143 panthor_gpu_soft_reset(ptdev); 144 panthor_gpu_l2_power_on(ptdev); 145 panthor_mmu_post_reset(ptdev); 146 ret = panthor_fw_post_reset(ptdev); 147 atomic_set(&ptdev->reset.pending, 0); 148 panthor_sched_post_reset(ptdev, ret != 0); 149 drm_dev_exit(cookie); 150 151 if (ret) { 152 panthor_device_unplug(ptdev); 153 drm_err(&ptdev->base, "Failed to boot MCU after reset, making device unusable."); 154 } 155 } 156 157 static bool panthor_device_is_initialized(struct panthor_device *ptdev) 158 { 159 return !!ptdev->scheduler; 160 } 161 162 static void panthor_device_free_page(struct drm_device *ddev, void *data) 163 { 164 __free_page(data); 165 } 166 167 int panthor_device_init(struct panthor_device *ptdev) 168 { 169 u32 *dummy_page_virt; 170 struct resource *res; 171 struct page *p; 172 int ret; 173 174 init_completion(&ptdev->unplug.done); 175 ret = drmm_mutex_init(&ptdev->base, &ptdev->unplug.lock); 176 if (ret) 177 return ret; 178 179 ret = drmm_mutex_init(&ptdev->base, &ptdev->pm.mmio_lock); 180 if (ret) 181 return ret; 182 183 #ifdef CONFIG_DEBUG_FS 184 drmm_mutex_init(&ptdev->base, &ptdev->gems.lock); 185 INIT_LIST_HEAD(&ptdev->gems.node); 186 #endif 187 188 atomic_set(&ptdev->pm.state, PANTHOR_DEVICE_PM_STATE_SUSPENDED); 189 p = alloc_page(GFP_KERNEL | __GFP_ZERO); 190 if (!p) 191 return -ENOMEM; 192 193 ptdev->pm.dummy_latest_flush = p; 194 dummy_page_virt = page_address(p); 195 ret = drmm_add_action_or_reset(&ptdev->base, panthor_device_free_page, 196 ptdev->pm.dummy_latest_flush); 197 if (ret) 198 return ret; 199 200 /* 201 * Set the dummy page holding the latest flush to 1. This will cause the 202 * flush to avoided as we know it isn't necessary if the submission 203 * happens while the dummy page is mapped. Zero cannot be used because 204 * that means 'always flush'. 205 */ 206 *dummy_page_virt = 1; 207 208 INIT_WORK(&ptdev->reset.work, panthor_device_reset_work); 209 ptdev->reset.wq = alloc_ordered_workqueue("panthor-reset-wq", 0); 210 if (!ptdev->reset.wq) 211 return -ENOMEM; 212 213 ret = drmm_add_action_or_reset(&ptdev->base, panthor_device_reset_cleanup, NULL); 214 if (ret) 215 return ret; 216 217 ret = panthor_clk_init(ptdev); 218 if (ret) 219 return ret; 220 221 ret = panthor_devfreq_init(ptdev); 222 if (ret) 223 return ret; 224 225 ptdev->iomem = devm_platform_get_and_ioremap_resource(to_platform_device(ptdev->base.dev), 226 0, &res); 227 if (IS_ERR(ptdev->iomem)) 228 return PTR_ERR(ptdev->iomem); 229 230 ptdev->phys_addr = res->start; 231 232 ret = devm_pm_runtime_enable(ptdev->base.dev); 233 if (ret) 234 return ret; 235 236 ret = pm_runtime_resume_and_get(ptdev->base.dev); 237 if (ret) 238 return ret; 239 240 /* If PM is disabled, we need to call panthor_device_resume() manually. */ 241 if (!IS_ENABLED(CONFIG_PM)) { 242 ret = panthor_device_resume(ptdev->base.dev); 243 if (ret) 244 return ret; 245 } 246 247 ret = panthor_gpu_init(ptdev); 248 if (ret) 249 goto err_rpm_put; 250 251 ret = panthor_gpu_coherency_init(ptdev); 252 if (ret) 253 goto err_unplug_gpu; 254 255 ret = panthor_mmu_init(ptdev); 256 if (ret) 257 goto err_unplug_gpu; 258 259 ret = panthor_fw_init(ptdev); 260 if (ret) 261 goto err_unplug_mmu; 262 263 ret = panthor_sched_init(ptdev); 264 if (ret) 265 goto err_unplug_fw; 266 267 /* ~3 frames */ 268 pm_runtime_set_autosuspend_delay(ptdev->base.dev, 50); 269 pm_runtime_use_autosuspend(ptdev->base.dev); 270 271 ret = drm_dev_register(&ptdev->base, 0); 272 if (ret) 273 goto err_disable_autosuspend; 274 275 pm_runtime_put_autosuspend(ptdev->base.dev); 276 return 0; 277 278 err_disable_autosuspend: 279 pm_runtime_dont_use_autosuspend(ptdev->base.dev); 280 panthor_sched_unplug(ptdev); 281 282 err_unplug_fw: 283 panthor_fw_unplug(ptdev); 284 285 err_unplug_mmu: 286 panthor_mmu_unplug(ptdev); 287 288 err_unplug_gpu: 289 panthor_gpu_unplug(ptdev); 290 291 err_rpm_put: 292 pm_runtime_put_sync_suspend(ptdev->base.dev); 293 return ret; 294 } 295 296 #define PANTHOR_EXCEPTION(id) \ 297 [DRM_PANTHOR_EXCEPTION_ ## id] = { \ 298 .name = #id, \ 299 } 300 301 struct panthor_exception_info { 302 const char *name; 303 }; 304 305 static const struct panthor_exception_info panthor_exception_infos[] = { 306 PANTHOR_EXCEPTION(OK), 307 PANTHOR_EXCEPTION(TERMINATED), 308 PANTHOR_EXCEPTION(KABOOM), 309 PANTHOR_EXCEPTION(EUREKA), 310 PANTHOR_EXCEPTION(ACTIVE), 311 PANTHOR_EXCEPTION(CS_RES_TERM), 312 PANTHOR_EXCEPTION(CS_CONFIG_FAULT), 313 PANTHOR_EXCEPTION(CS_UNRECOVERABLE), 314 PANTHOR_EXCEPTION(CS_ENDPOINT_FAULT), 315 PANTHOR_EXCEPTION(CS_BUS_FAULT), 316 PANTHOR_EXCEPTION(CS_INSTR_INVALID), 317 PANTHOR_EXCEPTION(CS_CALL_STACK_OVERFLOW), 318 PANTHOR_EXCEPTION(CS_INHERIT_FAULT), 319 PANTHOR_EXCEPTION(INSTR_INVALID_PC), 320 PANTHOR_EXCEPTION(INSTR_INVALID_ENC), 321 PANTHOR_EXCEPTION(INSTR_BARRIER_FAULT), 322 PANTHOR_EXCEPTION(DATA_INVALID_FAULT), 323 PANTHOR_EXCEPTION(TILE_RANGE_FAULT), 324 PANTHOR_EXCEPTION(ADDR_RANGE_FAULT), 325 PANTHOR_EXCEPTION(IMPRECISE_FAULT), 326 PANTHOR_EXCEPTION(OOM), 327 PANTHOR_EXCEPTION(CSF_FW_INTERNAL_ERROR), 328 PANTHOR_EXCEPTION(CSF_RES_EVICTION_TIMEOUT), 329 PANTHOR_EXCEPTION(GPU_BUS_FAULT), 330 PANTHOR_EXCEPTION(GPU_SHAREABILITY_FAULT), 331 PANTHOR_EXCEPTION(SYS_SHAREABILITY_FAULT), 332 PANTHOR_EXCEPTION(GPU_CACHEABILITY_FAULT), 333 PANTHOR_EXCEPTION(TRANSLATION_FAULT_0), 334 PANTHOR_EXCEPTION(TRANSLATION_FAULT_1), 335 PANTHOR_EXCEPTION(TRANSLATION_FAULT_2), 336 PANTHOR_EXCEPTION(TRANSLATION_FAULT_3), 337 PANTHOR_EXCEPTION(TRANSLATION_FAULT_4), 338 PANTHOR_EXCEPTION(PERM_FAULT_0), 339 PANTHOR_EXCEPTION(PERM_FAULT_1), 340 PANTHOR_EXCEPTION(PERM_FAULT_2), 341 PANTHOR_EXCEPTION(PERM_FAULT_3), 342 PANTHOR_EXCEPTION(ACCESS_FLAG_1), 343 PANTHOR_EXCEPTION(ACCESS_FLAG_2), 344 PANTHOR_EXCEPTION(ACCESS_FLAG_3), 345 PANTHOR_EXCEPTION(ADDR_SIZE_FAULT_IN), 346 PANTHOR_EXCEPTION(ADDR_SIZE_FAULT_OUT0), 347 PANTHOR_EXCEPTION(ADDR_SIZE_FAULT_OUT1), 348 PANTHOR_EXCEPTION(ADDR_SIZE_FAULT_OUT2), 349 PANTHOR_EXCEPTION(ADDR_SIZE_FAULT_OUT3), 350 PANTHOR_EXCEPTION(MEM_ATTR_FAULT_0), 351 PANTHOR_EXCEPTION(MEM_ATTR_FAULT_1), 352 PANTHOR_EXCEPTION(MEM_ATTR_FAULT_2), 353 PANTHOR_EXCEPTION(MEM_ATTR_FAULT_3), 354 }; 355 356 const char *panthor_exception_name(struct panthor_device *ptdev, u32 exception_code) 357 { 358 if (exception_code >= ARRAY_SIZE(panthor_exception_infos) || 359 !panthor_exception_infos[exception_code].name) 360 return "Unknown exception type"; 361 362 return panthor_exception_infos[exception_code].name; 363 } 364 365 static vm_fault_t panthor_mmio_vm_fault(struct vm_fault *vmf) 366 { 367 struct vm_area_struct *vma = vmf->vma; 368 struct panthor_device *ptdev = vma->vm_private_data; 369 u64 offset = (u64)vma->vm_pgoff << PAGE_SHIFT; 370 unsigned long pfn; 371 pgprot_t pgprot; 372 vm_fault_t ret; 373 bool active; 374 int cookie; 375 376 if (!drm_dev_enter(&ptdev->base, &cookie)) 377 return VM_FAULT_SIGBUS; 378 379 mutex_lock(&ptdev->pm.mmio_lock); 380 active = atomic_read(&ptdev->pm.state) == PANTHOR_DEVICE_PM_STATE_ACTIVE; 381 382 switch (offset) { 383 case DRM_PANTHOR_USER_FLUSH_ID_MMIO_OFFSET: 384 if (active) 385 pfn = __phys_to_pfn(ptdev->phys_addr + CSF_GPU_LATEST_FLUSH_ID); 386 else 387 pfn = page_to_pfn(ptdev->pm.dummy_latest_flush); 388 break; 389 390 default: 391 ret = VM_FAULT_SIGBUS; 392 goto out_unlock; 393 } 394 395 pgprot = vma->vm_page_prot; 396 if (active) 397 pgprot = pgprot_noncached(pgprot); 398 399 ret = vmf_insert_pfn_prot(vma, vmf->address, pfn, pgprot); 400 401 out_unlock: 402 mutex_unlock(&ptdev->pm.mmio_lock); 403 drm_dev_exit(cookie); 404 return ret; 405 } 406 407 static const struct vm_operations_struct panthor_mmio_vm_ops = { 408 .fault = panthor_mmio_vm_fault, 409 }; 410 411 int panthor_device_mmap_io(struct panthor_device *ptdev, struct vm_area_struct *vma) 412 { 413 u64 offset = (u64)vma->vm_pgoff << PAGE_SHIFT; 414 415 if ((vma->vm_flags & VM_SHARED) == 0) 416 return -EINVAL; 417 418 switch (offset) { 419 case DRM_PANTHOR_USER_FLUSH_ID_MMIO_OFFSET: 420 if (vma->vm_end - vma->vm_start != PAGE_SIZE || 421 (vma->vm_flags & (VM_WRITE | VM_EXEC))) 422 return -EINVAL; 423 vm_flags_clear(vma, VM_MAYWRITE); 424 425 break; 426 427 default: 428 return -EINVAL; 429 } 430 431 /* Defer actual mapping to the fault handler. */ 432 vma->vm_private_data = ptdev; 433 vma->vm_ops = &panthor_mmio_vm_ops; 434 vm_flags_set(vma, 435 VM_IO | VM_DONTCOPY | VM_DONTEXPAND | 436 VM_NORESERVE | VM_DONTDUMP | VM_PFNMAP); 437 return 0; 438 } 439 440 static int panthor_device_resume_hw_components(struct panthor_device *ptdev) 441 { 442 int ret; 443 444 panthor_gpu_resume(ptdev); 445 panthor_mmu_resume(ptdev); 446 447 ret = panthor_fw_resume(ptdev); 448 if (!ret) 449 return 0; 450 451 panthor_mmu_suspend(ptdev); 452 panthor_gpu_suspend(ptdev); 453 return ret; 454 } 455 456 int panthor_device_resume(struct device *dev) 457 { 458 struct panthor_device *ptdev = dev_get_drvdata(dev); 459 int ret, cookie; 460 461 if (atomic_read(&ptdev->pm.state) != PANTHOR_DEVICE_PM_STATE_SUSPENDED) 462 return -EINVAL; 463 464 atomic_set(&ptdev->pm.state, PANTHOR_DEVICE_PM_STATE_RESUMING); 465 466 ret = clk_prepare_enable(ptdev->clks.core); 467 if (ret) 468 goto err_set_suspended; 469 470 ret = clk_prepare_enable(ptdev->clks.stacks); 471 if (ret) 472 goto err_disable_core_clk; 473 474 ret = clk_prepare_enable(ptdev->clks.coregroup); 475 if (ret) 476 goto err_disable_stacks_clk; 477 478 panthor_devfreq_resume(ptdev); 479 480 if (panthor_device_is_initialized(ptdev) && 481 drm_dev_enter(&ptdev->base, &cookie)) { 482 /* If there was a reset pending at the time we suspended the 483 * device, we force a slow reset. 484 */ 485 if (atomic_read(&ptdev->reset.pending)) { 486 ptdev->reset.fast = false; 487 atomic_set(&ptdev->reset.pending, 0); 488 } 489 490 ret = panthor_device_resume_hw_components(ptdev); 491 if (ret && ptdev->reset.fast) { 492 drm_err(&ptdev->base, "Fast reset failed, trying a slow reset"); 493 ptdev->reset.fast = false; 494 ret = panthor_device_resume_hw_components(ptdev); 495 } 496 497 if (!ret) 498 panthor_sched_resume(ptdev); 499 500 drm_dev_exit(cookie); 501 502 if (ret) 503 goto err_suspend_devfreq; 504 } 505 506 /* Clear all IOMEM mappings pointing to this device after we've 507 * resumed. This way the fake mappings pointing to the dummy pages 508 * are removed and the real iomem mapping will be restored on next 509 * access. 510 */ 511 mutex_lock(&ptdev->pm.mmio_lock); 512 unmap_mapping_range(ptdev->base.anon_inode->i_mapping, 513 DRM_PANTHOR_USER_MMIO_OFFSET, 0, 1); 514 atomic_set(&ptdev->pm.state, PANTHOR_DEVICE_PM_STATE_ACTIVE); 515 mutex_unlock(&ptdev->pm.mmio_lock); 516 return 0; 517 518 err_suspend_devfreq: 519 panthor_devfreq_suspend(ptdev); 520 clk_disable_unprepare(ptdev->clks.coregroup); 521 522 err_disable_stacks_clk: 523 clk_disable_unprepare(ptdev->clks.stacks); 524 525 err_disable_core_clk: 526 clk_disable_unprepare(ptdev->clks.core); 527 528 err_set_suspended: 529 atomic_set(&ptdev->pm.state, PANTHOR_DEVICE_PM_STATE_SUSPENDED); 530 atomic_set(&ptdev->pm.recovery_needed, 1); 531 return ret; 532 } 533 534 int panthor_device_suspend(struct device *dev) 535 { 536 struct panthor_device *ptdev = dev_get_drvdata(dev); 537 int cookie; 538 539 if (atomic_read(&ptdev->pm.state) != PANTHOR_DEVICE_PM_STATE_ACTIVE) 540 return -EINVAL; 541 542 /* Clear all IOMEM mappings pointing to this device before we 543 * shutdown the power-domain and clocks. Failing to do that results 544 * in external aborts when the process accesses the iomem region. 545 * We change the state and call unmap_mapping_range() with the 546 * mmio_lock held to make sure the vm_fault handler won't set up 547 * invalid mappings. 548 */ 549 mutex_lock(&ptdev->pm.mmio_lock); 550 atomic_set(&ptdev->pm.state, PANTHOR_DEVICE_PM_STATE_SUSPENDING); 551 unmap_mapping_range(ptdev->base.anon_inode->i_mapping, 552 DRM_PANTHOR_USER_MMIO_OFFSET, 0, 1); 553 mutex_unlock(&ptdev->pm.mmio_lock); 554 555 if (panthor_device_is_initialized(ptdev) && 556 drm_dev_enter(&ptdev->base, &cookie)) { 557 cancel_work_sync(&ptdev->reset.work); 558 559 /* We prepare everything as if we were resetting the GPU. 560 * The end of the reset will happen in the resume path though. 561 */ 562 panthor_sched_suspend(ptdev); 563 panthor_fw_suspend(ptdev); 564 panthor_mmu_suspend(ptdev); 565 panthor_gpu_suspend(ptdev); 566 drm_dev_exit(cookie); 567 } 568 569 panthor_devfreq_suspend(ptdev); 570 571 clk_disable_unprepare(ptdev->clks.coregroup); 572 clk_disable_unprepare(ptdev->clks.stacks); 573 clk_disable_unprepare(ptdev->clks.core); 574 atomic_set(&ptdev->pm.state, PANTHOR_DEVICE_PM_STATE_SUSPENDED); 575 return 0; 576 } 577