1 // SPDX-License-Identifier: GPL-2.0-only 2 /* 3 * Copyright (c) 2015-2022, NVIDIA Corporation. 4 */ 5 6 #include <linux/clk.h> 7 #include <linux/delay.h> 8 #include <linux/dma-mapping.h> 9 #include <linux/host1x.h> 10 #include <linux/iommu.h> 11 #include <linux/iopoll.h> 12 #include <linux/module.h> 13 #include <linux/of.h> 14 #include <linux/of_device.h> 15 #include <linux/of_platform.h> 16 #include <linux/platform_device.h> 17 #include <linux/pm_runtime.h> 18 #include <linux/reset.h> 19 20 #include <soc/tegra/mc.h> 21 22 #include "drm.h" 23 #include "falcon.h" 24 #include "riscv.h" 25 #include "vic.h" 26 27 #define NVDEC_FALCON_DEBUGINFO 0x1094 28 #define NVDEC_TFBIF_TRANSCFG 0x2c44 29 30 struct nvdec_config { 31 const char *firmware; 32 unsigned int version; 33 bool supports_sid; 34 bool has_riscv; 35 bool has_extra_clocks; 36 }; 37 38 struct nvdec { 39 struct falcon falcon; 40 41 void __iomem *regs; 42 struct tegra_drm_client client; 43 struct host1x_channel *channel; 44 struct device *dev; 45 struct clk_bulk_data clks[3]; 46 unsigned int num_clks; 47 struct reset_control *reset; 48 49 /* Platform configuration */ 50 const struct nvdec_config *config; 51 52 /* RISC-V specific data */ 53 struct tegra_drm_riscv riscv; 54 phys_addr_t carveout_base; 55 }; 56 57 static inline struct nvdec *to_nvdec(struct tegra_drm_client *client) 58 { 59 return container_of(client, struct nvdec, client); 60 } 61 62 static inline void nvdec_writel(struct nvdec *nvdec, u32 value, 63 unsigned int offset) 64 { 65 writel(value, nvdec->regs + offset); 66 } 67 68 static int nvdec_boot_falcon(struct nvdec *nvdec) 69 { 70 u32 stream_id; 71 int err; 72 73 if (nvdec->config->supports_sid && tegra_dev_iommu_get_stream_id(nvdec->dev, &stream_id)) { 74 u32 value; 75 76 value = TRANSCFG_ATT(1, TRANSCFG_SID_FALCON) | TRANSCFG_ATT(0, TRANSCFG_SID_HW); 77 nvdec_writel(nvdec, value, NVDEC_TFBIF_TRANSCFG); 78 79 nvdec_writel(nvdec, stream_id, VIC_THI_STREAMID0); 80 nvdec_writel(nvdec, stream_id, VIC_THI_STREAMID1); 81 } 82 83 err = falcon_boot(&nvdec->falcon); 84 if (err < 0) 85 return err; 86 87 err = falcon_wait_idle(&nvdec->falcon); 88 if (err < 0) { 89 dev_err(nvdec->dev, "falcon boot timed out\n"); 90 return err; 91 } 92 93 return 0; 94 } 95 96 static int nvdec_wait_debuginfo(struct nvdec *nvdec, const char *phase) 97 { 98 int err; 99 u32 val; 100 101 err = readl_poll_timeout(nvdec->regs + NVDEC_FALCON_DEBUGINFO, val, val == 0x0, 10, 100000); 102 if (err) { 103 dev_err(nvdec->dev, "failed to boot %s, debuginfo=0x%x\n", phase, val); 104 return err; 105 } 106 107 return 0; 108 } 109 110 static int nvdec_boot_riscv(struct nvdec *nvdec) 111 { 112 int err; 113 114 err = reset_control_acquire(nvdec->reset); 115 if (err) 116 return err; 117 118 nvdec_writel(nvdec, 0xabcd1234, NVDEC_FALCON_DEBUGINFO); 119 120 err = tegra_drm_riscv_boot_bootrom(&nvdec->riscv, nvdec->carveout_base, 1, 121 &nvdec->riscv.bl_desc); 122 if (err) { 123 dev_err(nvdec->dev, "failed to execute bootloader\n"); 124 goto release_reset; 125 } 126 127 err = nvdec_wait_debuginfo(nvdec, "bootloader"); 128 if (err) 129 goto release_reset; 130 131 err = reset_control_reset(nvdec->reset); 132 if (err) 133 goto release_reset; 134 135 nvdec_writel(nvdec, 0xabcd1234, NVDEC_FALCON_DEBUGINFO); 136 137 err = tegra_drm_riscv_boot_bootrom(&nvdec->riscv, nvdec->carveout_base, 1, 138 &nvdec->riscv.os_desc); 139 if (err) { 140 dev_err(nvdec->dev, "failed to execute firmware\n"); 141 goto release_reset; 142 } 143 144 err = nvdec_wait_debuginfo(nvdec, "firmware"); 145 if (err) 146 goto release_reset; 147 148 release_reset: 149 reset_control_release(nvdec->reset); 150 151 return err; 152 } 153 154 static int nvdec_init(struct host1x_client *client) 155 { 156 struct tegra_drm_client *drm = host1x_to_drm_client(client); 157 struct drm_device *dev = dev_get_drvdata(client->host); 158 struct tegra_drm *tegra = dev->dev_private; 159 struct nvdec *nvdec = to_nvdec(drm); 160 int err; 161 162 err = host1x_client_iommu_attach(client); 163 if (err < 0 && err != -ENODEV) { 164 dev_err(nvdec->dev, "failed to attach to domain: %d\n", err); 165 return err; 166 } 167 168 nvdec->channel = host1x_channel_request(client); 169 if (!nvdec->channel) { 170 err = -ENOMEM; 171 goto detach; 172 } 173 174 client->syncpts[0] = host1x_syncpt_request(client, 0); 175 if (!client->syncpts[0]) { 176 err = -ENOMEM; 177 goto free_channel; 178 } 179 180 pm_runtime_enable(client->dev); 181 pm_runtime_use_autosuspend(client->dev); 182 pm_runtime_set_autosuspend_delay(client->dev, 500); 183 184 err = tegra_drm_register_client(tegra, drm); 185 if (err < 0) 186 goto disable_rpm; 187 188 /* 189 * Inherit the DMA parameters (such as maximum segment size) from the 190 * parent host1x device. 191 */ 192 client->dev->dma_parms = client->host->dma_parms; 193 194 return 0; 195 196 disable_rpm: 197 pm_runtime_dont_use_autosuspend(client->dev); 198 pm_runtime_force_suspend(client->dev); 199 200 host1x_syncpt_put(client->syncpts[0]); 201 free_channel: 202 host1x_channel_put(nvdec->channel); 203 detach: 204 host1x_client_iommu_detach(client); 205 206 return err; 207 } 208 209 static int nvdec_exit(struct host1x_client *client) 210 { 211 struct tegra_drm_client *drm = host1x_to_drm_client(client); 212 struct drm_device *dev = dev_get_drvdata(client->host); 213 struct tegra_drm *tegra = dev->dev_private; 214 struct nvdec *nvdec = to_nvdec(drm); 215 int err; 216 217 /* avoid a dangling pointer just in case this disappears */ 218 client->dev->dma_parms = NULL; 219 220 err = tegra_drm_unregister_client(tegra, drm); 221 if (err < 0) 222 return err; 223 224 pm_runtime_dont_use_autosuspend(client->dev); 225 pm_runtime_force_suspend(client->dev); 226 227 host1x_syncpt_put(client->syncpts[0]); 228 host1x_channel_put(nvdec->channel); 229 host1x_client_iommu_detach(client); 230 231 nvdec->channel = NULL; 232 233 if (client->group) { 234 dma_unmap_single(nvdec->dev, nvdec->falcon.firmware.phys, 235 nvdec->falcon.firmware.size, DMA_TO_DEVICE); 236 tegra_drm_free(tegra, nvdec->falcon.firmware.size, 237 nvdec->falcon.firmware.virt, 238 nvdec->falcon.firmware.iova); 239 } else { 240 dma_free_coherent(nvdec->dev, nvdec->falcon.firmware.size, 241 nvdec->falcon.firmware.virt, 242 nvdec->falcon.firmware.iova); 243 } 244 245 return 0; 246 } 247 248 static const struct host1x_client_ops nvdec_client_ops = { 249 .init = nvdec_init, 250 .exit = nvdec_exit, 251 }; 252 253 static int nvdec_load_falcon_firmware(struct nvdec *nvdec) 254 { 255 struct host1x_client *client = &nvdec->client.base; 256 struct tegra_drm *tegra = nvdec->client.drm; 257 dma_addr_t iova; 258 size_t size; 259 void *virt; 260 int err; 261 262 if (nvdec->falcon.firmware.virt) 263 return 0; 264 265 err = falcon_read_firmware(&nvdec->falcon, nvdec->config->firmware); 266 if (err < 0) 267 return err; 268 269 size = nvdec->falcon.firmware.size; 270 271 if (!client->group) { 272 virt = dma_alloc_coherent(nvdec->dev, size, &iova, GFP_KERNEL); 273 274 err = dma_mapping_error(nvdec->dev, iova); 275 if (err < 0) 276 return err; 277 } else { 278 virt = tegra_drm_alloc(tegra, size, &iova); 279 } 280 281 nvdec->falcon.firmware.virt = virt; 282 nvdec->falcon.firmware.iova = iova; 283 284 err = falcon_load_firmware(&nvdec->falcon); 285 if (err < 0) 286 goto cleanup; 287 288 /* 289 * In this case we have received an IOVA from the shared domain, so we 290 * need to make sure to get the physical address so that the DMA API 291 * knows what memory pages to flush the cache for. 292 */ 293 if (client->group) { 294 dma_addr_t phys; 295 296 phys = dma_map_single(nvdec->dev, virt, size, DMA_TO_DEVICE); 297 298 err = dma_mapping_error(nvdec->dev, phys); 299 if (err < 0) 300 goto cleanup; 301 302 nvdec->falcon.firmware.phys = phys; 303 } 304 305 return 0; 306 307 cleanup: 308 if (!client->group) 309 dma_free_coherent(nvdec->dev, size, virt, iova); 310 else 311 tegra_drm_free(tegra, size, virt, iova); 312 313 return err; 314 } 315 316 static __maybe_unused int nvdec_runtime_resume(struct device *dev) 317 { 318 struct nvdec *nvdec = dev_get_drvdata(dev); 319 int err; 320 321 err = clk_bulk_prepare_enable(nvdec->num_clks, nvdec->clks); 322 if (err < 0) 323 return err; 324 325 usleep_range(10, 20); 326 327 if (nvdec->config->has_riscv) { 328 err = nvdec_boot_riscv(nvdec); 329 if (err < 0) 330 goto disable; 331 } else { 332 err = nvdec_load_falcon_firmware(nvdec); 333 if (err < 0) 334 goto disable; 335 336 err = nvdec_boot_falcon(nvdec); 337 if (err < 0) 338 goto disable; 339 } 340 341 return 0; 342 343 disable: 344 clk_bulk_disable_unprepare(nvdec->num_clks, nvdec->clks); 345 return err; 346 } 347 348 static __maybe_unused int nvdec_runtime_suspend(struct device *dev) 349 { 350 struct nvdec *nvdec = dev_get_drvdata(dev); 351 352 host1x_channel_stop(nvdec->channel); 353 354 clk_bulk_disable_unprepare(nvdec->num_clks, nvdec->clks); 355 356 return 0; 357 } 358 359 static int nvdec_open_channel(struct tegra_drm_client *client, 360 struct tegra_drm_context *context) 361 { 362 struct nvdec *nvdec = to_nvdec(client); 363 364 context->channel = host1x_channel_get(nvdec->channel); 365 if (!context->channel) 366 return -ENOMEM; 367 368 return 0; 369 } 370 371 static void nvdec_close_channel(struct tegra_drm_context *context) 372 { 373 host1x_channel_put(context->channel); 374 } 375 376 static int nvdec_can_use_memory_ctx(struct tegra_drm_client *client, bool *supported) 377 { 378 *supported = true; 379 380 return 0; 381 } 382 383 static const struct tegra_drm_client_ops nvdec_ops = { 384 .open_channel = nvdec_open_channel, 385 .close_channel = nvdec_close_channel, 386 .submit = tegra_drm_submit, 387 .get_streamid_offset = tegra_drm_get_streamid_offset_thi, 388 .can_use_memory_ctx = nvdec_can_use_memory_ctx, 389 }; 390 391 #define NVIDIA_TEGRA_210_NVDEC_FIRMWARE "nvidia/tegra210/nvdec.bin" 392 393 static const struct nvdec_config nvdec_t210_config = { 394 .firmware = NVIDIA_TEGRA_210_NVDEC_FIRMWARE, 395 .version = 0x21, 396 .supports_sid = false, 397 }; 398 399 #define NVIDIA_TEGRA_186_NVDEC_FIRMWARE "nvidia/tegra186/nvdec.bin" 400 401 static const struct nvdec_config nvdec_t186_config = { 402 .firmware = NVIDIA_TEGRA_186_NVDEC_FIRMWARE, 403 .version = 0x18, 404 .supports_sid = true, 405 }; 406 407 #define NVIDIA_TEGRA_194_NVDEC_FIRMWARE "nvidia/tegra194/nvdec.bin" 408 409 static const struct nvdec_config nvdec_t194_config = { 410 .firmware = NVIDIA_TEGRA_194_NVDEC_FIRMWARE, 411 .version = 0x19, 412 .supports_sid = true, 413 }; 414 415 static const struct nvdec_config nvdec_t234_config = { 416 .version = 0x23, 417 .supports_sid = true, 418 .has_riscv = true, 419 .has_extra_clocks = true, 420 }; 421 422 static const struct of_device_id tegra_nvdec_of_match[] = { 423 { .compatible = "nvidia,tegra210-nvdec", .data = &nvdec_t210_config }, 424 { .compatible = "nvidia,tegra186-nvdec", .data = &nvdec_t186_config }, 425 { .compatible = "nvidia,tegra194-nvdec", .data = &nvdec_t194_config }, 426 { .compatible = "nvidia,tegra234-nvdec", .data = &nvdec_t234_config }, 427 { }, 428 }; 429 MODULE_DEVICE_TABLE(of, tegra_nvdec_of_match); 430 431 static int nvdec_probe(struct platform_device *pdev) 432 { 433 struct device *dev = &pdev->dev; 434 struct host1x_syncpt **syncpts; 435 struct nvdec *nvdec; 436 u32 host_class; 437 int err; 438 439 /* inherit DMA mask from host1x parent */ 440 err = dma_coerce_mask_and_coherent(dev, *dev->parent->dma_mask); 441 if (err < 0) { 442 dev_err(&pdev->dev, "failed to set DMA mask: %d\n", err); 443 return err; 444 } 445 446 nvdec = devm_kzalloc(dev, sizeof(*nvdec), GFP_KERNEL); 447 if (!nvdec) 448 return -ENOMEM; 449 450 nvdec->config = of_device_get_match_data(dev); 451 452 syncpts = devm_kzalloc(dev, sizeof(*syncpts), GFP_KERNEL); 453 if (!syncpts) 454 return -ENOMEM; 455 456 nvdec->regs = devm_platform_get_and_ioremap_resource(pdev, 0, NULL); 457 if (IS_ERR(nvdec->regs)) 458 return PTR_ERR(nvdec->regs); 459 460 nvdec->clks[0].id = "nvdec"; 461 nvdec->num_clks = 1; 462 463 if (nvdec->config->has_extra_clocks) { 464 nvdec->num_clks = 3; 465 nvdec->clks[1].id = "fuse"; 466 nvdec->clks[2].id = "tsec_pka"; 467 } 468 469 err = devm_clk_bulk_get(dev, nvdec->num_clks, nvdec->clks); 470 if (err) { 471 dev_err(&pdev->dev, "failed to get clock(s)\n"); 472 return err; 473 } 474 475 err = clk_set_rate(nvdec->clks[0].clk, ULONG_MAX); 476 if (err < 0) { 477 dev_err(&pdev->dev, "failed to set clock rate\n"); 478 return err; 479 } 480 481 err = of_property_read_u32(dev->of_node, "nvidia,host1x-class", &host_class); 482 if (err < 0) 483 host_class = HOST1X_CLASS_NVDEC; 484 485 if (nvdec->config->has_riscv) { 486 struct tegra_mc *mc; 487 488 mc = devm_tegra_memory_controller_get(dev); 489 if (IS_ERR(mc)) { 490 dev_err_probe(dev, PTR_ERR(mc), 491 "failed to get memory controller handle\n"); 492 return PTR_ERR(mc); 493 } 494 495 err = tegra_mc_get_carveout_info(mc, 1, &nvdec->carveout_base, NULL); 496 if (err) { 497 dev_err(dev, "failed to get carveout info: %d\n", err); 498 return err; 499 } 500 501 nvdec->reset = devm_reset_control_get_exclusive_released(dev, "nvdec"); 502 if (IS_ERR(nvdec->reset)) { 503 dev_err_probe(dev, PTR_ERR(nvdec->reset), "failed to get reset\n"); 504 return PTR_ERR(nvdec->reset); 505 } 506 507 nvdec->riscv.dev = dev; 508 nvdec->riscv.regs = nvdec->regs; 509 510 err = tegra_drm_riscv_read_descriptors(&nvdec->riscv); 511 if (err < 0) 512 return err; 513 } else { 514 nvdec->falcon.dev = dev; 515 nvdec->falcon.regs = nvdec->regs; 516 517 err = falcon_init(&nvdec->falcon); 518 if (err < 0) 519 return err; 520 } 521 522 platform_set_drvdata(pdev, nvdec); 523 524 INIT_LIST_HEAD(&nvdec->client.base.list); 525 nvdec->client.base.ops = &nvdec_client_ops; 526 nvdec->client.base.dev = dev; 527 nvdec->client.base.class = host_class; 528 nvdec->client.base.syncpts = syncpts; 529 nvdec->client.base.num_syncpts = 1; 530 nvdec->dev = dev; 531 532 INIT_LIST_HEAD(&nvdec->client.list); 533 nvdec->client.version = nvdec->config->version; 534 nvdec->client.ops = &nvdec_ops; 535 536 err = host1x_client_register(&nvdec->client.base); 537 if (err < 0) { 538 dev_err(dev, "failed to register host1x client: %d\n", err); 539 goto exit_falcon; 540 } 541 542 return 0; 543 544 exit_falcon: 545 falcon_exit(&nvdec->falcon); 546 547 return err; 548 } 549 550 static int nvdec_remove(struct platform_device *pdev) 551 { 552 struct nvdec *nvdec = platform_get_drvdata(pdev); 553 int err; 554 555 err = host1x_client_unregister(&nvdec->client.base); 556 if (err < 0) { 557 dev_err(&pdev->dev, "failed to unregister host1x client: %d\n", 558 err); 559 return err; 560 } 561 562 falcon_exit(&nvdec->falcon); 563 564 return 0; 565 } 566 567 static const struct dev_pm_ops nvdec_pm_ops = { 568 SET_RUNTIME_PM_OPS(nvdec_runtime_suspend, nvdec_runtime_resume, NULL) 569 SET_SYSTEM_SLEEP_PM_OPS(pm_runtime_force_suspend, 570 pm_runtime_force_resume) 571 }; 572 573 struct platform_driver tegra_nvdec_driver = { 574 .driver = { 575 .name = "tegra-nvdec", 576 .of_match_table = tegra_nvdec_of_match, 577 .pm = &nvdec_pm_ops 578 }, 579 .probe = nvdec_probe, 580 .remove = nvdec_remove, 581 }; 582 583 #if IS_ENABLED(CONFIG_ARCH_TEGRA_210_SOC) 584 MODULE_FIRMWARE(NVIDIA_TEGRA_210_NVDEC_FIRMWARE); 585 #endif 586 #if IS_ENABLED(CONFIG_ARCH_TEGRA_186_SOC) 587 MODULE_FIRMWARE(NVIDIA_TEGRA_186_NVDEC_FIRMWARE); 588 #endif 589 #if IS_ENABLED(CONFIG_ARCH_TEGRA_194_SOC) 590 MODULE_FIRMWARE(NVIDIA_TEGRA_194_NVDEC_FIRMWARE); 591 #endif 592