157692c94SEric Anholt // SPDX-License-Identifier: GPL-2.0+ 257692c94SEric Anholt /* Copyright (C) 2014-2018 Broadcom */ 357692c94SEric Anholt 457692c94SEric Anholt /** 557692c94SEric Anholt * DOC: Interrupt management for the V3D engine 657692c94SEric Anholt * 71584f16cSEric Anholt * When we take a bin, render, or TFU done interrupt, we need to 81584f16cSEric Anholt * signal the fence for that job so that the scheduler can queue up 957692c94SEric Anholt * the next one and unblock any waiters. 1057692c94SEric Anholt * 1157692c94SEric Anholt * When we take the binner out of memory interrupt, we need to 1257692c94SEric Anholt * allocate some new memory and pass it to the binner so that the 1357692c94SEric Anholt * current job can make progress. 1457692c94SEric Anholt */ 1557692c94SEric Anholt 1657692c94SEric Anholt #include "v3d_drv.h" 1757692c94SEric Anholt #include "v3d_regs.h" 1855a9b748SEric Anholt #include "v3d_trace.h" 1957692c94SEric Anholt 2057692c94SEric Anholt #define V3D_CORE_IRQS ((u32)(V3D_INT_OUTOMEM | \ 2157692c94SEric Anholt V3D_INT_FLDONE | \ 2257692c94SEric Anholt V3D_INT_FRDONE | \ 2357692c94SEric Anholt V3D_INT_GMPV)) 2457692c94SEric Anholt 2557692c94SEric Anholt #define V3D_HUB_IRQS ((u32)(V3D_HUB_INT_MMU_WRV | \ 2657692c94SEric Anholt V3D_HUB_INT_MMU_PTI | \ 271584f16cSEric Anholt V3D_HUB_INT_MMU_CAP | \ 281584f16cSEric Anholt V3D_HUB_INT_TFUC)) 2957692c94SEric Anholt 3057692c94SEric Anholt static void 3157692c94SEric Anholt v3d_overflow_mem_work(struct work_struct *work) 3257692c94SEric Anholt { 3357692c94SEric Anholt struct v3d_dev *v3d = 3457692c94SEric Anholt container_of(work, struct v3d_dev, overflow_mem_work); 3557692c94SEric Anholt struct drm_device *dev = &v3d->drm; 3657692c94SEric Anholt struct v3d_bo *bo = v3d_bo_create(dev, NULL /* XXX: GMP */, 256 * 1024); 3757692c94SEric Anholt unsigned long irqflags; 3857692c94SEric Anholt 3957692c94SEric Anholt if (IS_ERR(bo)) { 4057692c94SEric Anholt DRM_ERROR("Couldn't allocate binner overflow mem\n"); 4157692c94SEric Anholt return; 4257692c94SEric Anholt } 4357692c94SEric Anholt 4457692c94SEric Anholt /* We lost a race, and our work task came in after the bin job 4557692c94SEric Anholt * completed and exited. This can happen because the HW 4657692c94SEric Anholt * signals OOM before it's fully OOM, so the binner might just 4757692c94SEric Anholt * barely complete. 4857692c94SEric Anholt * 4957692c94SEric Anholt * If we lose the race and our work task comes in after a new 5057692c94SEric Anholt * bin job got scheduled, that's fine. We'll just give them 5157692c94SEric Anholt * some binner pool anyway. 5257692c94SEric Anholt */ 5357692c94SEric Anholt spin_lock_irqsave(&v3d->job_lock, irqflags); 5457692c94SEric Anholt if (!v3d->bin_job) { 5557692c94SEric Anholt spin_unlock_irqrestore(&v3d->job_lock, irqflags); 5657692c94SEric Anholt goto out; 5757692c94SEric Anholt } 5857692c94SEric Anholt 5957692c94SEric Anholt drm_gem_object_get(&bo->base); 6057692c94SEric Anholt list_add_tail(&bo->unref_head, &v3d->bin_job->unref_list); 6157692c94SEric Anholt spin_unlock_irqrestore(&v3d->job_lock, irqflags); 6257692c94SEric Anholt 6357692c94SEric Anholt V3D_CORE_WRITE(0, V3D_PTB_BPOA, bo->node.start << PAGE_SHIFT); 6457692c94SEric Anholt V3D_CORE_WRITE(0, V3D_PTB_BPOS, bo->base.size); 6557692c94SEric Anholt 6657692c94SEric Anholt out: 6757692c94SEric Anholt drm_gem_object_put_unlocked(&bo->base); 6857692c94SEric Anholt } 6957692c94SEric Anholt 7057692c94SEric Anholt static irqreturn_t 7157692c94SEric Anholt v3d_irq(int irq, void *arg) 7257692c94SEric Anholt { 7357692c94SEric Anholt struct v3d_dev *v3d = arg; 7457692c94SEric Anholt u32 intsts; 7557692c94SEric Anholt irqreturn_t status = IRQ_NONE; 7657692c94SEric Anholt 7757692c94SEric Anholt intsts = V3D_CORE_READ(0, V3D_CTL_INT_STS); 7857692c94SEric Anholt 7957692c94SEric Anholt /* Acknowledge the interrupts we're handling here. */ 8057692c94SEric Anholt V3D_CORE_WRITE(0, V3D_CTL_INT_CLR, intsts); 8157692c94SEric Anholt 8257692c94SEric Anholt if (intsts & V3D_INT_OUTOMEM) { 8357692c94SEric Anholt /* Note that the OOM status is edge signaled, so the 8457692c94SEric Anholt * interrupt won't happen again until the we actually 8557692c94SEric Anholt * add more memory. 8657692c94SEric Anholt */ 8757692c94SEric Anholt schedule_work(&v3d->overflow_mem_work); 8857692c94SEric Anholt status = IRQ_HANDLED; 8957692c94SEric Anholt } 9057692c94SEric Anholt 9157692c94SEric Anholt if (intsts & V3D_INT_FLDONE) { 9255a9b748SEric Anholt struct v3d_fence *fence = 9355a9b748SEric Anholt to_v3d_fence(v3d->bin_job->bin.done_fence); 9455a9b748SEric Anholt 9555a9b748SEric Anholt trace_v3d_bcl_irq(&v3d->drm, fence->seqno); 9655a9b748SEric Anholt dma_fence_signal(&fence->base); 9757692c94SEric Anholt status = IRQ_HANDLED; 9857692c94SEric Anholt } 9957692c94SEric Anholt 10057692c94SEric Anholt if (intsts & V3D_INT_FRDONE) { 10155a9b748SEric Anholt struct v3d_fence *fence = 10255a9b748SEric Anholt to_v3d_fence(v3d->render_job->render.done_fence); 10355a9b748SEric Anholt 10455a9b748SEric Anholt trace_v3d_rcl_irq(&v3d->drm, fence->seqno); 10555a9b748SEric Anholt dma_fence_signal(&fence->base); 10657692c94SEric Anholt status = IRQ_HANDLED; 10757692c94SEric Anholt } 10857692c94SEric Anholt 10957692c94SEric Anholt /* We shouldn't be triggering these if we have GMP in 11057692c94SEric Anholt * always-allowed mode. 11157692c94SEric Anholt */ 11257692c94SEric Anholt if (intsts & V3D_INT_GMPV) 11357692c94SEric Anholt dev_err(v3d->dev, "GMP violation\n"); 11457692c94SEric Anholt 11557692c94SEric Anholt return status; 11657692c94SEric Anholt } 11757692c94SEric Anholt 11857692c94SEric Anholt static irqreturn_t 11957692c94SEric Anholt v3d_hub_irq(int irq, void *arg) 12057692c94SEric Anholt { 12157692c94SEric Anholt struct v3d_dev *v3d = arg; 12257692c94SEric Anholt u32 intsts; 12357692c94SEric Anholt irqreturn_t status = IRQ_NONE; 12457692c94SEric Anholt 12557692c94SEric Anholt intsts = V3D_READ(V3D_HUB_INT_STS); 12657692c94SEric Anholt 12757692c94SEric Anholt /* Acknowledge the interrupts we're handling here. */ 12857692c94SEric Anholt V3D_WRITE(V3D_HUB_INT_CLR, intsts); 12957692c94SEric Anholt 1301584f16cSEric Anholt if (intsts & V3D_HUB_INT_TFUC) { 13155a9b748SEric Anholt struct v3d_fence *fence = 13255a9b748SEric Anholt to_v3d_fence(v3d->tfu_job->done_fence); 13355a9b748SEric Anholt 13455a9b748SEric Anholt trace_v3d_tfu_irq(&v3d->drm, fence->seqno); 13555a9b748SEric Anholt dma_fence_signal(&fence->base); 1361584f16cSEric Anholt status = IRQ_HANDLED; 1371584f16cSEric Anholt } 1381584f16cSEric Anholt 13957692c94SEric Anholt if (intsts & (V3D_HUB_INT_MMU_WRV | 14057692c94SEric Anholt V3D_HUB_INT_MMU_PTI | 14157692c94SEric Anholt V3D_HUB_INT_MMU_CAP)) { 14257692c94SEric Anholt u32 axi_id = V3D_READ(V3D_MMU_VIO_ID); 14357692c94SEric Anholt u64 vio_addr = (u64)V3D_READ(V3D_MMU_VIO_ADDR) << 8; 14457692c94SEric Anholt 14557692c94SEric Anholt dev_err(v3d->dev, "MMU error from client %d at 0x%08llx%s%s%s\n", 14657692c94SEric Anholt axi_id, (long long)vio_addr, 14757692c94SEric Anholt ((intsts & V3D_HUB_INT_MMU_WRV) ? 14857692c94SEric Anholt ", write violation" : ""), 14957692c94SEric Anholt ((intsts & V3D_HUB_INT_MMU_PTI) ? 15057692c94SEric Anholt ", pte invalid" : ""), 15157692c94SEric Anholt ((intsts & V3D_HUB_INT_MMU_CAP) ? 15257692c94SEric Anholt ", cap exceeded" : "")); 15357692c94SEric Anholt status = IRQ_HANDLED; 15457692c94SEric Anholt } 15557692c94SEric Anholt 15657692c94SEric Anholt return status; 15757692c94SEric Anholt } 15857692c94SEric Anholt 159*fc227715SEric Anholt int 16057692c94SEric Anholt v3d_irq_init(struct v3d_dev *v3d) 16157692c94SEric Anholt { 16257692c94SEric Anholt int ret, core; 16357692c94SEric Anholt 16457692c94SEric Anholt INIT_WORK(&v3d->overflow_mem_work, v3d_overflow_mem_work); 16557692c94SEric Anholt 16657692c94SEric Anholt /* Clear any pending interrupts someone might have left around 16757692c94SEric Anholt * for us. 16857692c94SEric Anholt */ 16957692c94SEric Anholt for (core = 0; core < v3d->cores; core++) 17057692c94SEric Anholt V3D_CORE_WRITE(core, V3D_CTL_INT_CLR, V3D_CORE_IRQS); 17157692c94SEric Anholt V3D_WRITE(V3D_HUB_INT_CLR, V3D_HUB_IRQS); 17257692c94SEric Anholt 17357692c94SEric Anholt ret = devm_request_irq(v3d->dev, platform_get_irq(v3d->pdev, 0), 17457692c94SEric Anholt v3d_hub_irq, IRQF_SHARED, 17557692c94SEric Anholt "v3d_hub", v3d); 176*fc227715SEric Anholt if (ret) 177*fc227715SEric Anholt goto fail; 178*fc227715SEric Anholt 17957692c94SEric Anholt ret = devm_request_irq(v3d->dev, platform_get_irq(v3d->pdev, 1), 18057692c94SEric Anholt v3d_irq, IRQF_SHARED, 18157692c94SEric Anholt "v3d_core0", v3d); 18257692c94SEric Anholt if (ret) 183*fc227715SEric Anholt goto fail; 18457692c94SEric Anholt 18557692c94SEric Anholt v3d_irq_enable(v3d); 186*fc227715SEric Anholt return 0; 187*fc227715SEric Anholt 188*fc227715SEric Anholt fail: 189*fc227715SEric Anholt if (ret != -EPROBE_DEFER) 190*fc227715SEric Anholt dev_err(v3d->dev, "IRQ setup failed: %d\n", ret); 191*fc227715SEric Anholt return ret; 19257692c94SEric Anholt } 19357692c94SEric Anholt 19457692c94SEric Anholt void 19557692c94SEric Anholt v3d_irq_enable(struct v3d_dev *v3d) 19657692c94SEric Anholt { 19757692c94SEric Anholt int core; 19857692c94SEric Anholt 19957692c94SEric Anholt /* Enable our set of interrupts, masking out any others. */ 20057692c94SEric Anholt for (core = 0; core < v3d->cores; core++) { 20157692c94SEric Anholt V3D_CORE_WRITE(core, V3D_CTL_INT_MSK_SET, ~V3D_CORE_IRQS); 20257692c94SEric Anholt V3D_CORE_WRITE(core, V3D_CTL_INT_MSK_CLR, V3D_CORE_IRQS); 20357692c94SEric Anholt } 20457692c94SEric Anholt 20557692c94SEric Anholt V3D_WRITE(V3D_HUB_INT_MSK_SET, ~V3D_HUB_IRQS); 20657692c94SEric Anholt V3D_WRITE(V3D_HUB_INT_MSK_CLR, V3D_HUB_IRQS); 20757692c94SEric Anholt } 20857692c94SEric Anholt 20957692c94SEric Anholt void 21057692c94SEric Anholt v3d_irq_disable(struct v3d_dev *v3d) 21157692c94SEric Anholt { 21257692c94SEric Anholt int core; 21357692c94SEric Anholt 21457692c94SEric Anholt /* Disable all interrupts. */ 21557692c94SEric Anholt for (core = 0; core < v3d->cores; core++) 21657692c94SEric Anholt V3D_CORE_WRITE(core, V3D_CTL_INT_MSK_SET, ~0); 21757692c94SEric Anholt V3D_WRITE(V3D_HUB_INT_MSK_SET, ~0); 21857692c94SEric Anholt 21957692c94SEric Anholt /* Clear any pending interrupts we might have left. */ 22057692c94SEric Anholt for (core = 0; core < v3d->cores; core++) 22157692c94SEric Anholt V3D_CORE_WRITE(core, V3D_CTL_INT_CLR, V3D_CORE_IRQS); 22257692c94SEric Anholt V3D_WRITE(V3D_HUB_INT_CLR, V3D_HUB_IRQS); 22357692c94SEric Anholt 22457692c94SEric Anholt cancel_work_sync(&v3d->overflow_mem_work); 22557692c94SEric Anholt } 22657692c94SEric Anholt 22757692c94SEric Anholt /** Reinitializes interrupt registers when a GPU reset is performed. */ 22857692c94SEric Anholt void v3d_irq_reset(struct v3d_dev *v3d) 22957692c94SEric Anholt { 23057692c94SEric Anholt v3d_irq_enable(v3d); 23157692c94SEric Anholt } 232