157692c94SEric Anholt // SPDX-License-Identifier: GPL-2.0+ 257692c94SEric Anholt /* Copyright (C) 2014-2018 Broadcom */ 357692c94SEric Anholt 457692c94SEric Anholt /** 557692c94SEric Anholt * DOC: Interrupt management for the V3D engine 657692c94SEric Anholt * 71584f16cSEric Anholt * When we take a bin, render, or TFU done interrupt, we need to 81584f16cSEric Anholt * signal the fence for that job so that the scheduler can queue up 957692c94SEric Anholt * the next one and unblock any waiters. 1057692c94SEric Anholt * 1157692c94SEric Anholt * When we take the binner out of memory interrupt, we need to 1257692c94SEric Anholt * allocate some new memory and pass it to the binner so that the 1357692c94SEric Anholt * current job can make progress. 1457692c94SEric Anholt */ 1557692c94SEric Anholt 1657692c94SEric Anholt #include "v3d_drv.h" 1757692c94SEric Anholt #include "v3d_regs.h" 1855a9b748SEric Anholt #include "v3d_trace.h" 1957692c94SEric Anholt 2057692c94SEric Anholt #define V3D_CORE_IRQS ((u32)(V3D_INT_OUTOMEM | \ 2157692c94SEric Anholt V3D_INT_FLDONE | \ 2257692c94SEric Anholt V3D_INT_FRDONE | \ 2357692c94SEric Anholt V3D_INT_GMPV)) 2457692c94SEric Anholt 2557692c94SEric Anholt #define V3D_HUB_IRQS ((u32)(V3D_HUB_INT_MMU_WRV | \ 2657692c94SEric Anholt V3D_HUB_INT_MMU_PTI | \ 271584f16cSEric Anholt V3D_HUB_INT_MMU_CAP | \ 281584f16cSEric Anholt V3D_HUB_INT_TFUC)) 2957692c94SEric Anholt 30*eea9b97bSEric Anholt static irqreturn_t 31*eea9b97bSEric Anholt v3d_hub_irq(int irq, void *arg); 32*eea9b97bSEric Anholt 3357692c94SEric Anholt static void 3457692c94SEric Anholt v3d_overflow_mem_work(struct work_struct *work) 3557692c94SEric Anholt { 3657692c94SEric Anholt struct v3d_dev *v3d = 3757692c94SEric Anholt container_of(work, struct v3d_dev, overflow_mem_work); 3857692c94SEric Anholt struct drm_device *dev = &v3d->drm; 3957692c94SEric Anholt struct v3d_bo *bo = v3d_bo_create(dev, NULL /* XXX: GMP */, 256 * 1024); 4057692c94SEric Anholt unsigned long irqflags; 4157692c94SEric Anholt 4257692c94SEric Anholt if (IS_ERR(bo)) { 4357692c94SEric Anholt DRM_ERROR("Couldn't allocate binner overflow mem\n"); 4457692c94SEric Anholt return; 4557692c94SEric Anholt } 4657692c94SEric Anholt 4757692c94SEric Anholt /* We lost a race, and our work task came in after the bin job 4857692c94SEric Anholt * completed and exited. This can happen because the HW 4957692c94SEric Anholt * signals OOM before it's fully OOM, so the binner might just 5057692c94SEric Anholt * barely complete. 5157692c94SEric Anholt * 5257692c94SEric Anholt * If we lose the race and our work task comes in after a new 5357692c94SEric Anholt * bin job got scheduled, that's fine. We'll just give them 5457692c94SEric Anholt * some binner pool anyway. 5557692c94SEric Anholt */ 5657692c94SEric Anholt spin_lock_irqsave(&v3d->job_lock, irqflags); 5757692c94SEric Anholt if (!v3d->bin_job) { 5857692c94SEric Anholt spin_unlock_irqrestore(&v3d->job_lock, irqflags); 5957692c94SEric Anholt goto out; 6057692c94SEric Anholt } 6157692c94SEric Anholt 6257692c94SEric Anholt drm_gem_object_get(&bo->base); 6357692c94SEric Anholt list_add_tail(&bo->unref_head, &v3d->bin_job->unref_list); 6457692c94SEric Anholt spin_unlock_irqrestore(&v3d->job_lock, irqflags); 6557692c94SEric Anholt 6657692c94SEric Anholt V3D_CORE_WRITE(0, V3D_PTB_BPOA, bo->node.start << PAGE_SHIFT); 6757692c94SEric Anholt V3D_CORE_WRITE(0, V3D_PTB_BPOS, bo->base.size); 6857692c94SEric Anholt 6957692c94SEric Anholt out: 7057692c94SEric Anholt drm_gem_object_put_unlocked(&bo->base); 7157692c94SEric Anholt } 7257692c94SEric Anholt 7357692c94SEric Anholt static irqreturn_t 7457692c94SEric Anholt v3d_irq(int irq, void *arg) 7557692c94SEric Anholt { 7657692c94SEric Anholt struct v3d_dev *v3d = arg; 7757692c94SEric Anholt u32 intsts; 7857692c94SEric Anholt irqreturn_t status = IRQ_NONE; 7957692c94SEric Anholt 8057692c94SEric Anholt intsts = V3D_CORE_READ(0, V3D_CTL_INT_STS); 8157692c94SEric Anholt 8257692c94SEric Anholt /* Acknowledge the interrupts we're handling here. */ 8357692c94SEric Anholt V3D_CORE_WRITE(0, V3D_CTL_INT_CLR, intsts); 8457692c94SEric Anholt 8557692c94SEric Anholt if (intsts & V3D_INT_OUTOMEM) { 8657692c94SEric Anholt /* Note that the OOM status is edge signaled, so the 8757692c94SEric Anholt * interrupt won't happen again until the we actually 8857692c94SEric Anholt * add more memory. 8957692c94SEric Anholt */ 9057692c94SEric Anholt schedule_work(&v3d->overflow_mem_work); 9157692c94SEric Anholt status = IRQ_HANDLED; 9257692c94SEric Anholt } 9357692c94SEric Anholt 9457692c94SEric Anholt if (intsts & V3D_INT_FLDONE) { 9555a9b748SEric Anholt struct v3d_fence *fence = 9655a9b748SEric Anholt to_v3d_fence(v3d->bin_job->bin.done_fence); 9755a9b748SEric Anholt 9855a9b748SEric Anholt trace_v3d_bcl_irq(&v3d->drm, fence->seqno); 9955a9b748SEric Anholt dma_fence_signal(&fence->base); 10057692c94SEric Anholt status = IRQ_HANDLED; 10157692c94SEric Anholt } 10257692c94SEric Anholt 10357692c94SEric Anholt if (intsts & V3D_INT_FRDONE) { 10455a9b748SEric Anholt struct v3d_fence *fence = 10555a9b748SEric Anholt to_v3d_fence(v3d->render_job->render.done_fence); 10655a9b748SEric Anholt 10755a9b748SEric Anholt trace_v3d_rcl_irq(&v3d->drm, fence->seqno); 10855a9b748SEric Anholt dma_fence_signal(&fence->base); 10957692c94SEric Anholt status = IRQ_HANDLED; 11057692c94SEric Anholt } 11157692c94SEric Anholt 11257692c94SEric Anholt /* We shouldn't be triggering these if we have GMP in 11357692c94SEric Anholt * always-allowed mode. 11457692c94SEric Anholt */ 11557692c94SEric Anholt if (intsts & V3D_INT_GMPV) 11657692c94SEric Anholt dev_err(v3d->dev, "GMP violation\n"); 11757692c94SEric Anholt 118*eea9b97bSEric Anholt /* V3D 4.2 wires the hub and core IRQs together, so if we & 119*eea9b97bSEric Anholt * didn't see the common one then check hub for MMU IRQs. 120*eea9b97bSEric Anholt */ 121*eea9b97bSEric Anholt if (v3d->single_irq_line && status == IRQ_NONE) 122*eea9b97bSEric Anholt return v3d_hub_irq(irq, arg); 123*eea9b97bSEric Anholt 12457692c94SEric Anholt return status; 12557692c94SEric Anholt } 12657692c94SEric Anholt 12757692c94SEric Anholt static irqreturn_t 12857692c94SEric Anholt v3d_hub_irq(int irq, void *arg) 12957692c94SEric Anholt { 13057692c94SEric Anholt struct v3d_dev *v3d = arg; 13157692c94SEric Anholt u32 intsts; 13257692c94SEric Anholt irqreturn_t status = IRQ_NONE; 13357692c94SEric Anholt 13457692c94SEric Anholt intsts = V3D_READ(V3D_HUB_INT_STS); 13557692c94SEric Anholt 13657692c94SEric Anholt /* Acknowledge the interrupts we're handling here. */ 13757692c94SEric Anholt V3D_WRITE(V3D_HUB_INT_CLR, intsts); 13857692c94SEric Anholt 1391584f16cSEric Anholt if (intsts & V3D_HUB_INT_TFUC) { 14055a9b748SEric Anholt struct v3d_fence *fence = 14155a9b748SEric Anholt to_v3d_fence(v3d->tfu_job->done_fence); 14255a9b748SEric Anholt 14355a9b748SEric Anholt trace_v3d_tfu_irq(&v3d->drm, fence->seqno); 14455a9b748SEric Anholt dma_fence_signal(&fence->base); 1451584f16cSEric Anholt status = IRQ_HANDLED; 1461584f16cSEric Anholt } 1471584f16cSEric Anholt 14857692c94SEric Anholt if (intsts & (V3D_HUB_INT_MMU_WRV | 14957692c94SEric Anholt V3D_HUB_INT_MMU_PTI | 15057692c94SEric Anholt V3D_HUB_INT_MMU_CAP)) { 15157692c94SEric Anholt u32 axi_id = V3D_READ(V3D_MMU_VIO_ID); 15257692c94SEric Anholt u64 vio_addr = (u64)V3D_READ(V3D_MMU_VIO_ADDR) << 8; 15357692c94SEric Anholt 15457692c94SEric Anholt dev_err(v3d->dev, "MMU error from client %d at 0x%08llx%s%s%s\n", 15557692c94SEric Anholt axi_id, (long long)vio_addr, 15657692c94SEric Anholt ((intsts & V3D_HUB_INT_MMU_WRV) ? 15757692c94SEric Anholt ", write violation" : ""), 15857692c94SEric Anholt ((intsts & V3D_HUB_INT_MMU_PTI) ? 15957692c94SEric Anholt ", pte invalid" : ""), 16057692c94SEric Anholt ((intsts & V3D_HUB_INT_MMU_CAP) ? 16157692c94SEric Anholt ", cap exceeded" : "")); 16257692c94SEric Anholt status = IRQ_HANDLED; 16357692c94SEric Anholt } 16457692c94SEric Anholt 16557692c94SEric Anholt return status; 16657692c94SEric Anholt } 16757692c94SEric Anholt 168fc227715SEric Anholt int 16957692c94SEric Anholt v3d_irq_init(struct v3d_dev *v3d) 17057692c94SEric Anholt { 171*eea9b97bSEric Anholt int irq1, ret, core; 17257692c94SEric Anholt 17357692c94SEric Anholt INIT_WORK(&v3d->overflow_mem_work, v3d_overflow_mem_work); 17457692c94SEric Anholt 17557692c94SEric Anholt /* Clear any pending interrupts someone might have left around 17657692c94SEric Anholt * for us. 17757692c94SEric Anholt */ 17857692c94SEric Anholt for (core = 0; core < v3d->cores; core++) 17957692c94SEric Anholt V3D_CORE_WRITE(core, V3D_CTL_INT_CLR, V3D_CORE_IRQS); 18057692c94SEric Anholt V3D_WRITE(V3D_HUB_INT_CLR, V3D_HUB_IRQS); 18157692c94SEric Anholt 182*eea9b97bSEric Anholt irq1 = platform_get_irq(v3d->pdev, 1); 183*eea9b97bSEric Anholt if (irq1 == -EPROBE_DEFER) 184*eea9b97bSEric Anholt return irq1; 185*eea9b97bSEric Anholt if (irq1 > 0) { 186*eea9b97bSEric Anholt ret = devm_request_irq(v3d->dev, irq1, 187*eea9b97bSEric Anholt v3d_irq, IRQF_SHARED, 188*eea9b97bSEric Anholt "v3d_core0", v3d); 189*eea9b97bSEric Anholt if (ret) 190*eea9b97bSEric Anholt goto fail; 19157692c94SEric Anholt ret = devm_request_irq(v3d->dev, platform_get_irq(v3d->pdev, 0), 19257692c94SEric Anholt v3d_hub_irq, IRQF_SHARED, 19357692c94SEric Anholt "v3d_hub", v3d); 194fc227715SEric Anholt if (ret) 195fc227715SEric Anholt goto fail; 196*eea9b97bSEric Anholt } else { 197*eea9b97bSEric Anholt v3d->single_irq_line = true; 198fc227715SEric Anholt 199*eea9b97bSEric Anholt ret = devm_request_irq(v3d->dev, platform_get_irq(v3d->pdev, 0), 20057692c94SEric Anholt v3d_irq, IRQF_SHARED, 201*eea9b97bSEric Anholt "v3d", v3d); 20257692c94SEric Anholt if (ret) 203fc227715SEric Anholt goto fail; 204*eea9b97bSEric Anholt } 20557692c94SEric Anholt 20657692c94SEric Anholt v3d_irq_enable(v3d); 207fc227715SEric Anholt return 0; 208fc227715SEric Anholt 209fc227715SEric Anholt fail: 210fc227715SEric Anholt if (ret != -EPROBE_DEFER) 211fc227715SEric Anholt dev_err(v3d->dev, "IRQ setup failed: %d\n", ret); 212fc227715SEric Anholt return ret; 21357692c94SEric Anholt } 21457692c94SEric Anholt 21557692c94SEric Anholt void 21657692c94SEric Anholt v3d_irq_enable(struct v3d_dev *v3d) 21757692c94SEric Anholt { 21857692c94SEric Anholt int core; 21957692c94SEric Anholt 22057692c94SEric Anholt /* Enable our set of interrupts, masking out any others. */ 22157692c94SEric Anholt for (core = 0; core < v3d->cores; core++) { 22257692c94SEric Anholt V3D_CORE_WRITE(core, V3D_CTL_INT_MSK_SET, ~V3D_CORE_IRQS); 22357692c94SEric Anholt V3D_CORE_WRITE(core, V3D_CTL_INT_MSK_CLR, V3D_CORE_IRQS); 22457692c94SEric Anholt } 22557692c94SEric Anholt 22657692c94SEric Anholt V3D_WRITE(V3D_HUB_INT_MSK_SET, ~V3D_HUB_IRQS); 22757692c94SEric Anholt V3D_WRITE(V3D_HUB_INT_MSK_CLR, V3D_HUB_IRQS); 22857692c94SEric Anholt } 22957692c94SEric Anholt 23057692c94SEric Anholt void 23157692c94SEric Anholt v3d_irq_disable(struct v3d_dev *v3d) 23257692c94SEric Anholt { 23357692c94SEric Anholt int core; 23457692c94SEric Anholt 23557692c94SEric Anholt /* Disable all interrupts. */ 23657692c94SEric Anholt for (core = 0; core < v3d->cores; core++) 23757692c94SEric Anholt V3D_CORE_WRITE(core, V3D_CTL_INT_MSK_SET, ~0); 23857692c94SEric Anholt V3D_WRITE(V3D_HUB_INT_MSK_SET, ~0); 23957692c94SEric Anholt 24057692c94SEric Anholt /* Clear any pending interrupts we might have left. */ 24157692c94SEric Anholt for (core = 0; core < v3d->cores; core++) 24257692c94SEric Anholt V3D_CORE_WRITE(core, V3D_CTL_INT_CLR, V3D_CORE_IRQS); 24357692c94SEric Anholt V3D_WRITE(V3D_HUB_INT_CLR, V3D_HUB_IRQS); 24457692c94SEric Anholt 24557692c94SEric Anholt cancel_work_sync(&v3d->overflow_mem_work); 24657692c94SEric Anholt } 24757692c94SEric Anholt 24857692c94SEric Anholt /** Reinitializes interrupt registers when a GPU reset is performed. */ 24957692c94SEric Anholt void v3d_irq_reset(struct v3d_dev *v3d) 25057692c94SEric Anholt { 25157692c94SEric Anholt v3d_irq_enable(v3d); 25257692c94SEric Anholt } 253