1 /* SPDX-License-Identifier: GPL-2.0 or MIT */ 2 /* Copyright 2018 Marty E. Plummer <hanetzer@startmail.com> */ 3 /* Copyright 2019 Linaro, Ltd, Rob Herring <robh@kernel.org> */ 4 /* Copyright 2023 Collabora ltd. */ 5 6 #ifndef __PANTHOR_DEVICE_H__ 7 #define __PANTHOR_DEVICE_H__ 8 9 #include <linux/atomic.h> 10 #include <linux/io-pgtable.h> 11 #include <linux/regulator/consumer.h> 12 #include <linux/pm_runtime.h> 13 #include <linux/sched.h> 14 #include <linux/spinlock.h> 15 16 #include <drm/drm_device.h> 17 #include <drm/drm_mm.h> 18 #include <drm/gpu_scheduler.h> 19 #include <drm/panthor_drm.h> 20 21 struct panthor_csf; 22 struct panthor_csf_ctx; 23 struct panthor_device; 24 struct panthor_gpu; 25 struct panthor_group_pool; 26 struct panthor_heap_pool; 27 struct panthor_job; 28 struct panthor_mmu; 29 struct panthor_fw; 30 struct panthor_perfcnt; 31 struct panthor_vm; 32 struct panthor_vm_pool; 33 34 /** 35 * enum panthor_device_pm_state - PM state 36 */ 37 enum panthor_device_pm_state { 38 /** @PANTHOR_DEVICE_PM_STATE_SUSPENDED: Device is suspended. */ 39 PANTHOR_DEVICE_PM_STATE_SUSPENDED = 0, 40 41 /** @PANTHOR_DEVICE_PM_STATE_RESUMING: Device is being resumed. */ 42 PANTHOR_DEVICE_PM_STATE_RESUMING, 43 44 /** @PANTHOR_DEVICE_PM_STATE_ACTIVE: Device is active. */ 45 PANTHOR_DEVICE_PM_STATE_ACTIVE, 46 47 /** @PANTHOR_DEVICE_PM_STATE_SUSPENDING: Device is being suspended. */ 48 PANTHOR_DEVICE_PM_STATE_SUSPENDING, 49 }; 50 51 /** 52 * struct panthor_irq - IRQ data 53 * 54 * Used to automate IRQ handling for the 3 different IRQs we have in this driver. 55 */ 56 struct panthor_irq { 57 /** @ptdev: Panthor device */ 58 struct panthor_device *ptdev; 59 60 /** @irq: IRQ number. */ 61 int irq; 62 63 /** @mask: Current mask being applied to xxx_INT_MASK. */ 64 u32 mask; 65 66 /** @suspended: Set to true when the IRQ is suspended. */ 67 atomic_t suspended; 68 }; 69 70 /** 71 * enum panthor_device_profiling_mode - Profiling state 72 */ 73 enum panthor_device_profiling_flags { 74 /** @PANTHOR_DEVICE_PROFILING_DISABLED: Profiling is disabled. */ 75 PANTHOR_DEVICE_PROFILING_DISABLED = 0, 76 77 /** @PANTHOR_DEVICE_PROFILING_CYCLES: Sampling job cycles. */ 78 PANTHOR_DEVICE_PROFILING_CYCLES = BIT(0), 79 80 /** @PANTHOR_DEVICE_PROFILING_TIMESTAMP: Sampling job timestamp. */ 81 PANTHOR_DEVICE_PROFILING_TIMESTAMP = BIT(1), 82 83 /** @PANTHOR_DEVICE_PROFILING_ALL: Sampling everything. */ 84 PANTHOR_DEVICE_PROFILING_ALL = 85 PANTHOR_DEVICE_PROFILING_CYCLES | 86 PANTHOR_DEVICE_PROFILING_TIMESTAMP, 87 }; 88 89 /** 90 * struct panthor_device - Panthor device 91 */ 92 struct panthor_device { 93 /** @base: Base drm_device. */ 94 struct drm_device base; 95 96 /** @phys_addr: Physical address of the iomem region. */ 97 phys_addr_t phys_addr; 98 99 /** @iomem: CPU mapping of the IOMEM region. */ 100 void __iomem *iomem; 101 102 /** @clks: GPU clocks. */ 103 struct { 104 /** @core: Core clock. */ 105 struct clk *core; 106 107 /** @stacks: Stacks clock. This clock is optional. */ 108 struct clk *stacks; 109 110 /** @coregroup: Core group clock. This clock is optional. */ 111 struct clk *coregroup; 112 } clks; 113 114 /** @coherent: True if the CPU/GPU are memory coherent. */ 115 bool coherent; 116 117 /** @gpu_info: GPU information. */ 118 struct drm_panthor_gpu_info gpu_info; 119 120 /** @csif_info: Command stream interface information. */ 121 struct drm_panthor_csif_info csif_info; 122 123 /** @gpu: GPU management data. */ 124 struct panthor_gpu *gpu; 125 126 /** @fw: FW management data. */ 127 struct panthor_fw *fw; 128 129 /** @mmu: MMU management data. */ 130 struct panthor_mmu *mmu; 131 132 /** @scheduler: Scheduler management data. */ 133 struct panthor_scheduler *scheduler; 134 135 /** @devfreq: Device frequency scaling management data. */ 136 struct panthor_devfreq *devfreq; 137 138 /** @unplug: Device unplug related fields. */ 139 struct { 140 /** @lock: Lock used to serialize unplug operations. */ 141 struct mutex lock; 142 143 /** 144 * @done: Completion object signaled when the unplug 145 * operation is done. 146 */ 147 struct completion done; 148 } unplug; 149 150 /** @reset: Reset related fields. */ 151 struct { 152 /** @wq: Ordered worqueud used to schedule reset operations. */ 153 struct workqueue_struct *wq; 154 155 /** @work: Reset work. */ 156 struct work_struct work; 157 158 /** @pending: Set to true if a reset is pending. */ 159 atomic_t pending; 160 161 /** 162 * @fast: True if the post_reset logic can proceed with a fast reset. 163 * 164 * A fast reset is just a reset where the driver doesn't reload the FW sections. 165 * 166 * Any time the firmware is properly suspended, a fast reset can take place. 167 * On the other hand, if the halt operation failed, the driver will reload 168 * all FW sections to make sure we start from a fresh state. 169 */ 170 bool fast; 171 } reset; 172 173 /** @pm: Power management related data. */ 174 struct { 175 /** @state: Power state. */ 176 atomic_t state; 177 178 /** 179 * @mmio_lock: Lock protecting MMIO userspace CPU mappings. 180 * 181 * This is needed to ensure we map the dummy IO pages when 182 * the device is being suspended, and the real IO pages when 183 * the device is being resumed. We can't just do with the 184 * state atomicity to deal with this race. 185 */ 186 struct mutex mmio_lock; 187 188 /** 189 * @dummy_latest_flush: Dummy LATEST_FLUSH page. 190 * 191 * Used to replace the real LATEST_FLUSH page when the GPU 192 * is suspended. 193 */ 194 struct page *dummy_latest_flush; 195 196 /** @recovery_needed: True when a resume attempt failed. */ 197 atomic_t recovery_needed; 198 } pm; 199 200 /** @profile_mask: User-set profiling flags for job accounting. */ 201 u32 profile_mask; 202 203 /** @current_frequency: Device clock frequency at present. Set by DVFS*/ 204 unsigned long current_frequency; 205 206 /** @fast_rate: Maximum device clock frequency. Set by DVFS */ 207 unsigned long fast_rate; 208 }; 209 210 struct panthor_gpu_usage { 211 u64 time; 212 u64 cycles; 213 }; 214 215 /** 216 * struct panthor_file - Panthor file 217 */ 218 struct panthor_file { 219 /** @ptdev: Device attached to this file. */ 220 struct panthor_device *ptdev; 221 222 /** @vms: VM pool attached to this file. */ 223 struct panthor_vm_pool *vms; 224 225 /** @groups: Scheduling group pool attached to this file. */ 226 struct panthor_group_pool *groups; 227 228 /** @stats: cycle and timestamp measures for job execution. */ 229 struct panthor_gpu_usage stats; 230 }; 231 232 int panthor_device_init(struct panthor_device *ptdev); 233 void panthor_device_unplug(struct panthor_device *ptdev); 234 235 /** 236 * panthor_device_schedule_reset() - Schedules a reset operation 237 */ 238 static inline void panthor_device_schedule_reset(struct panthor_device *ptdev) 239 { 240 if (!atomic_cmpxchg(&ptdev->reset.pending, 0, 1) && 241 atomic_read(&ptdev->pm.state) == PANTHOR_DEVICE_PM_STATE_ACTIVE) 242 queue_work(ptdev->reset.wq, &ptdev->reset.work); 243 } 244 245 /** 246 * panthor_device_reset_is_pending() - Checks if a reset is pending. 247 * 248 * Return: true if a reset is pending, false otherwise. 249 */ 250 static inline bool panthor_device_reset_is_pending(struct panthor_device *ptdev) 251 { 252 return atomic_read(&ptdev->reset.pending) != 0; 253 } 254 255 int panthor_device_mmap_io(struct panthor_device *ptdev, 256 struct vm_area_struct *vma); 257 258 int panthor_device_resume(struct device *dev); 259 int panthor_device_suspend(struct device *dev); 260 261 static inline int panthor_device_resume_and_get(struct panthor_device *ptdev) 262 { 263 int ret = pm_runtime_resume_and_get(ptdev->base.dev); 264 265 /* If the resume failed, we need to clear the runtime_error, which 266 * can done by forcing the RPM state to suspended. If multiple 267 * threads called panthor_device_resume_and_get(), we only want 268 * one of them to update the state, hence the cmpxchg. Note that a 269 * thread might enter panthor_device_resume_and_get() and call 270 * pm_runtime_resume_and_get() after another thread had attempted 271 * to resume and failed. This means we will end up with an error 272 * without even attempting a resume ourselves. The only risk here 273 * is to report an error when the second resume attempt might have 274 * succeeded. Given resume errors are not expected, this is probably 275 * something we can live with. 276 */ 277 if (ret && atomic_cmpxchg(&ptdev->pm.recovery_needed, 1, 0) == 1) 278 pm_runtime_set_suspended(ptdev->base.dev); 279 280 return ret; 281 } 282 283 enum drm_panthor_exception_type { 284 DRM_PANTHOR_EXCEPTION_OK = 0x00, 285 DRM_PANTHOR_EXCEPTION_TERMINATED = 0x04, 286 DRM_PANTHOR_EXCEPTION_KABOOM = 0x05, 287 DRM_PANTHOR_EXCEPTION_EUREKA = 0x06, 288 DRM_PANTHOR_EXCEPTION_ACTIVE = 0x08, 289 DRM_PANTHOR_EXCEPTION_CS_RES_TERM = 0x0f, 290 DRM_PANTHOR_EXCEPTION_MAX_NON_FAULT = 0x3f, 291 DRM_PANTHOR_EXCEPTION_CS_CONFIG_FAULT = 0x40, 292 DRM_PANTHOR_EXCEPTION_CS_UNRECOVERABLE = 0x41, 293 DRM_PANTHOR_EXCEPTION_CS_ENDPOINT_FAULT = 0x44, 294 DRM_PANTHOR_EXCEPTION_CS_BUS_FAULT = 0x48, 295 DRM_PANTHOR_EXCEPTION_CS_INSTR_INVALID = 0x49, 296 DRM_PANTHOR_EXCEPTION_CS_CALL_STACK_OVERFLOW = 0x4a, 297 DRM_PANTHOR_EXCEPTION_CS_INHERIT_FAULT = 0x4b, 298 DRM_PANTHOR_EXCEPTION_INSTR_INVALID_PC = 0x50, 299 DRM_PANTHOR_EXCEPTION_INSTR_INVALID_ENC = 0x51, 300 DRM_PANTHOR_EXCEPTION_INSTR_BARRIER_FAULT = 0x55, 301 DRM_PANTHOR_EXCEPTION_DATA_INVALID_FAULT = 0x58, 302 DRM_PANTHOR_EXCEPTION_TILE_RANGE_FAULT = 0x59, 303 DRM_PANTHOR_EXCEPTION_ADDR_RANGE_FAULT = 0x5a, 304 DRM_PANTHOR_EXCEPTION_IMPRECISE_FAULT = 0x5b, 305 DRM_PANTHOR_EXCEPTION_OOM = 0x60, 306 DRM_PANTHOR_EXCEPTION_CSF_FW_INTERNAL_ERROR = 0x68, 307 DRM_PANTHOR_EXCEPTION_CSF_RES_EVICTION_TIMEOUT = 0x69, 308 DRM_PANTHOR_EXCEPTION_GPU_BUS_FAULT = 0x80, 309 DRM_PANTHOR_EXCEPTION_GPU_SHAREABILITY_FAULT = 0x88, 310 DRM_PANTHOR_EXCEPTION_SYS_SHAREABILITY_FAULT = 0x89, 311 DRM_PANTHOR_EXCEPTION_GPU_CACHEABILITY_FAULT = 0x8a, 312 DRM_PANTHOR_EXCEPTION_TRANSLATION_FAULT_0 = 0xc0, 313 DRM_PANTHOR_EXCEPTION_TRANSLATION_FAULT_1 = 0xc1, 314 DRM_PANTHOR_EXCEPTION_TRANSLATION_FAULT_2 = 0xc2, 315 DRM_PANTHOR_EXCEPTION_TRANSLATION_FAULT_3 = 0xc3, 316 DRM_PANTHOR_EXCEPTION_TRANSLATION_FAULT_4 = 0xc4, 317 DRM_PANTHOR_EXCEPTION_PERM_FAULT_0 = 0xc8, 318 DRM_PANTHOR_EXCEPTION_PERM_FAULT_1 = 0xc9, 319 DRM_PANTHOR_EXCEPTION_PERM_FAULT_2 = 0xca, 320 DRM_PANTHOR_EXCEPTION_PERM_FAULT_3 = 0xcb, 321 DRM_PANTHOR_EXCEPTION_ACCESS_FLAG_1 = 0xd9, 322 DRM_PANTHOR_EXCEPTION_ACCESS_FLAG_2 = 0xda, 323 DRM_PANTHOR_EXCEPTION_ACCESS_FLAG_3 = 0xdb, 324 DRM_PANTHOR_EXCEPTION_ADDR_SIZE_FAULT_IN = 0xe0, 325 DRM_PANTHOR_EXCEPTION_ADDR_SIZE_FAULT_OUT0 = 0xe4, 326 DRM_PANTHOR_EXCEPTION_ADDR_SIZE_FAULT_OUT1 = 0xe5, 327 DRM_PANTHOR_EXCEPTION_ADDR_SIZE_FAULT_OUT2 = 0xe6, 328 DRM_PANTHOR_EXCEPTION_ADDR_SIZE_FAULT_OUT3 = 0xe7, 329 DRM_PANTHOR_EXCEPTION_MEM_ATTR_FAULT_0 = 0xe8, 330 DRM_PANTHOR_EXCEPTION_MEM_ATTR_FAULT_1 = 0xe9, 331 DRM_PANTHOR_EXCEPTION_MEM_ATTR_FAULT_2 = 0xea, 332 DRM_PANTHOR_EXCEPTION_MEM_ATTR_FAULT_3 = 0xeb, 333 }; 334 335 /** 336 * panthor_exception_is_fault() - Checks if an exception is a fault. 337 * 338 * Return: true if the exception is a fault, false otherwise. 339 */ 340 static inline bool 341 panthor_exception_is_fault(u32 exception_code) 342 { 343 return exception_code > DRM_PANTHOR_EXCEPTION_MAX_NON_FAULT; 344 } 345 346 const char *panthor_exception_name(struct panthor_device *ptdev, 347 u32 exception_code); 348 349 /** 350 * PANTHOR_IRQ_HANDLER() - Define interrupt handlers and the interrupt 351 * registration function. 352 * 353 * The boiler-plate to gracefully deal with shared interrupts is 354 * auto-generated. All you have to do is call PANTHOR_IRQ_HANDLER() 355 * just after the actual handler. The handler prototype is: 356 * 357 * void (*handler)(struct panthor_device *, u32 status); 358 */ 359 #define PANTHOR_IRQ_HANDLER(__name, __reg_prefix, __handler) \ 360 static irqreturn_t panthor_ ## __name ## _irq_raw_handler(int irq, void *data) \ 361 { \ 362 struct panthor_irq *pirq = data; \ 363 struct panthor_device *ptdev = pirq->ptdev; \ 364 \ 365 if (atomic_read(&pirq->suspended)) \ 366 return IRQ_NONE; \ 367 if (!gpu_read(ptdev, __reg_prefix ## _INT_STAT)) \ 368 return IRQ_NONE; \ 369 \ 370 gpu_write(ptdev, __reg_prefix ## _INT_MASK, 0); \ 371 return IRQ_WAKE_THREAD; \ 372 } \ 373 \ 374 static irqreturn_t panthor_ ## __name ## _irq_threaded_handler(int irq, void *data) \ 375 { \ 376 struct panthor_irq *pirq = data; \ 377 struct panthor_device *ptdev = pirq->ptdev; \ 378 irqreturn_t ret = IRQ_NONE; \ 379 \ 380 while (true) { \ 381 u32 status = gpu_read(ptdev, __reg_prefix ## _INT_RAWSTAT) & pirq->mask; \ 382 \ 383 if (!status) \ 384 break; \ 385 \ 386 gpu_write(ptdev, __reg_prefix ## _INT_CLEAR, status); \ 387 \ 388 __handler(ptdev, status); \ 389 ret = IRQ_HANDLED; \ 390 } \ 391 \ 392 if (!atomic_read(&pirq->suspended)) \ 393 gpu_write(ptdev, __reg_prefix ## _INT_MASK, pirq->mask); \ 394 \ 395 return ret; \ 396 } \ 397 \ 398 static inline void panthor_ ## __name ## _irq_suspend(struct panthor_irq *pirq) \ 399 { \ 400 pirq->mask = 0; \ 401 gpu_write(pirq->ptdev, __reg_prefix ## _INT_MASK, 0); \ 402 synchronize_irq(pirq->irq); \ 403 atomic_set(&pirq->suspended, true); \ 404 } \ 405 \ 406 static inline void panthor_ ## __name ## _irq_resume(struct panthor_irq *pirq, u32 mask) \ 407 { \ 408 atomic_set(&pirq->suspended, false); \ 409 pirq->mask = mask; \ 410 gpu_write(pirq->ptdev, __reg_prefix ## _INT_CLEAR, mask); \ 411 gpu_write(pirq->ptdev, __reg_prefix ## _INT_MASK, mask); \ 412 } \ 413 \ 414 static int panthor_request_ ## __name ## _irq(struct panthor_device *ptdev, \ 415 struct panthor_irq *pirq, \ 416 int irq, u32 mask) \ 417 { \ 418 pirq->ptdev = ptdev; \ 419 pirq->irq = irq; \ 420 panthor_ ## __name ## _irq_resume(pirq, mask); \ 421 \ 422 return devm_request_threaded_irq(ptdev->base.dev, irq, \ 423 panthor_ ## __name ## _irq_raw_handler, \ 424 panthor_ ## __name ## _irq_threaded_handler, \ 425 IRQF_SHARED, KBUILD_MODNAME "-" # __name, \ 426 pirq); \ 427 } 428 429 extern struct workqueue_struct *panthor_cleanup_wq; 430 431 #endif 432