| /linux/drivers/gpu/drm/panthor/ |
| H A D | panthor_pwr.c | 55 static void panthor_pwr_irq_handler(struct panthor_device *ptdev, u32 status) in panthor_pwr_irq_handler() argument 57 spin_lock(&ptdev->pwr->reqs_lock); in panthor_pwr_irq_handler() 58 gpu_write(ptdev, PWR_INT_CLEAR, status); in panthor_pwr_irq_handler() 61 drm_err(&ptdev->base, "PWR_IRQ: COMMAND_NOT_ALLOWED"); in panthor_pwr_irq_handler() 64 drm_err(&ptdev->base, "PWR_IRQ: COMMAND_INVALID"); in panthor_pwr_irq_handler() 66 if (status & ptdev->pwr->pending_reqs) { in panthor_pwr_irq_handler() 67 ptdev->pwr->pending_reqs &= ~status; in panthor_pwr_irq_handler() 68 wake_up_all(&ptdev->pwr->reqs_acked); in panthor_pwr_irq_handler() 70 spin_unlock(&ptdev->pwr->reqs_lock); in panthor_pwr_irq_handler() 74 static void panthor_pwr_write_command(struct panthor_device *ptdev, u32 command, u64 args) in panthor_pwr_write_command() argument [all …]
|
| H A D | panthor_device.h | 71 struct panthor_device *ptdev; member 250 struct panthor_device *ptdev; member 280 int panthor_device_init(struct panthor_device *ptdev); 281 void panthor_device_unplug(struct panthor_device *ptdev); 286 static inline void panthor_device_schedule_reset(struct panthor_device *ptdev) in panthor_device_schedule_reset() argument 288 if (!atomic_cmpxchg(&ptdev->reset.pending, 0, 1) && in panthor_device_schedule_reset() 289 atomic_read(&ptdev->pm.state) == PANTHOR_DEVICE_PM_STATE_ACTIVE) in panthor_device_schedule_reset() 290 queue_work(ptdev->reset.wq, &ptdev->reset.work); in panthor_device_schedule_reset() 298 static inline bool panthor_device_reset_is_pending(struct panthor_device *ptdev) in panthor_device_reset_is_pending() argument 300 return atomic_read(&ptdev->reset.pending) != 0; in panthor_device_reset_is_pending() [all …]
|
| H A D | panthor_mmu.c | 250 struct panthor_device *ptdev; member 442 drm_WARN_ON(&vm->ptdev->base, vm->op_ctx); in alloc_pt() 443 p = alloc_pages_node(dev_to_node(vm->ptdev->base.dev), in alloc_pt() 453 if (drm_WARN_ON(&vm->ptdev->base, size != SZ_4K)) in alloc_pt() 459 if (drm_WARN_ON(&vm->ptdev->base, !vm->op_ctx) || in alloc_pt() 460 drm_WARN_ON(&vm->ptdev->base, in alloc_pt() 495 if (drm_WARN_ON(&vm->ptdev->base, size != SZ_4K)) in free_pt() 502 static int wait_ready(struct panthor_device *ptdev, u32 as_nr) in wait_ready() argument 510 ret = gpu_read_relaxed_poll_timeout_atomic(ptdev, AS_STATUS(as_nr), val, in wait_ready() 515 panthor_device_schedule_reset(ptdev); in wait_ready() [all …]
|
| H A D | panthor_sched.c | 146 struct panthor_device *ptdev; member 535 struct panthor_device *ptdev; member 734 if (!queue_work((group)->ptdev->scheduler->wq, &(group)->wname ## _work)) \ 748 !panthor_device_reset_is_pending((sched)->ptdev)) \ 764 !panthor_device_reset_is_pending((sched)->ptdev)) \ 868 struct panthor_device *ptdev = group->ptdev; in panthor_queue_get_syncwait_obj() local 882 if (drm_WARN_ON(&ptdev->base, IS_ERR_OR_NULL(bo))) in panthor_queue_get_syncwait_obj() 887 if (drm_WARN_ON(&ptdev->base, ret)) in panthor_queue_get_syncwait_obj() 891 if (drm_WARN_ON(&ptdev->base, !queue->syncwait.kmap)) in panthor_queue_get_syncwait_obj() 967 struct panthor_device *ptdev = group->ptdev; in group_release() local [all …]
|
| H A D | panthor_hw.h | 15 int (*soft_reset)(struct panthor_device *ptdev); 18 void (*l2_power_off)(struct panthor_device *ptdev); 21 int (*l2_power_on)(struct panthor_device *ptdev); 34 int panthor_hw_init(struct panthor_device *ptdev); 36 static inline int panthor_hw_soft_reset(struct panthor_device *ptdev) in panthor_hw_soft_reset() argument 38 return ptdev->hw->ops.soft_reset(ptdev); in panthor_hw_soft_reset() 41 static inline int panthor_hw_l2_power_on(struct panthor_device *ptdev) in panthor_hw_l2_power_on() argument 43 return ptdev->hw->ops.l2_power_on(ptdev); in panthor_hw_l2_power_on() 46 static inline void panthor_hw_l2_power_off(struct panthor_device *ptdev) in panthor_hw_l2_power_off() argument 48 ptdev->hw->ops.l2_power_off(ptdev); in panthor_hw_l2_power_off() [all …]
|
| H A D | panthor_gpu.h | 12 int panthor_gpu_init(struct panthor_device *ptdev); 13 void panthor_gpu_unplug(struct panthor_device *ptdev); 14 void panthor_gpu_suspend(struct panthor_device *ptdev); 15 void panthor_gpu_resume(struct panthor_device *ptdev); 17 int panthor_gpu_block_power_on(struct panthor_device *ptdev, 21 int panthor_gpu_block_power_off(struct panthor_device *ptdev, 31 #define panthor_gpu_power_on(ptdev, type, mask, timeout_us) \ argument 32 panthor_gpu_block_power_on(ptdev, #type, \ 43 #define panthor_gpu_power_off(ptdev, type, mask, timeout_us) \ argument 44 panthor_gpu_block_power_off(ptdev, #type, \ [all …]
|
| H A D | panthor_pwr.h | 9 void panthor_pwr_unplug(struct panthor_device *ptdev); 11 int panthor_pwr_init(struct panthor_device *ptdev); 13 int panthor_pwr_reset_soft(struct panthor_device *ptdev); 15 void panthor_pwr_l2_power_off(struct panthor_device *ptdev); 17 int panthor_pwr_l2_power_on(struct panthor_device *ptdev); 19 void panthor_pwr_suspend(struct panthor_device *ptdev); 21 void panthor_pwr_resume(struct panthor_device *ptdev);
|
| H A D | panthor_fw.h | 479 panthor_fw_get_glb_iface(struct panthor_device *ptdev); 482 panthor_fw_get_csg_iface(struct panthor_device *ptdev, u32 csg_slot); 485 panthor_fw_get_cs_iface(struct panthor_device *ptdev, u32 csg_slot, u32 cs_slot); 487 u64 panthor_fw_csg_endpoint_req_get(struct panthor_device *ptdev, 490 void panthor_fw_csg_endpoint_req_set(struct panthor_device *ptdev, 493 void panthor_fw_csg_endpoint_req_update(struct panthor_device *ptdev, 497 int panthor_fw_csg_wait_acks(struct panthor_device *ptdev, u32 csg_id, u32 req_mask, 500 int panthor_fw_glb_wait_acks(struct panthor_device *ptdev, u32 req_mask, u32 *acked, 503 void panthor_fw_ring_csg_doorbells(struct panthor_device *ptdev, u32 csg_slot); 506 panthor_fw_alloc_queue_iface_mem(struct panthor_device *ptdev, [all …]
|
| H A D | panthor_drv.c | 502 struct panthor_device *ptdev = container_of(ctx->file->minor->dev, in panthor_submit_ctx_update_job_sync_signal_fences() local 518 if (drm_WARN_ON(&ptdev->base, !sig_sync)) in panthor_submit_ctx_update_job_sync_signal_fences() 525 if (drm_WARN_ON(&ptdev->base, !sig_sync->fence)) in panthor_submit_ctx_update_job_sync_signal_fences() 603 struct panthor_device *ptdev = container_of(ctx->file->minor->dev, in panthor_submit_ctx_add_sync_deps_to_job() local 625 if (drm_WARN_ON(&ptdev->base, !sig_sync->fence)) in panthor_submit_ctx_add_sync_deps_to_job() 764 static int panthor_query_timestamp_info(struct panthor_device *ptdev, in panthor_query_timestamp_info() argument 769 ret = panthor_device_resume_and_get(ptdev); in panthor_query_timestamp_info() 778 arg->current_timestamp = gpu_read64_counter(ptdev, GPU_TIMESTAMP); in panthor_query_timestamp_info() 779 arg->timestamp_offset = gpu_read64(ptdev, GPU_TIMESTAMP_OFFSET); in panthor_query_timestamp_info() 781 pm_runtime_put(ptdev->base.dev); in panthor_query_timestamp_info() [all …]
|
| H A D | panthor_mmu.h | 19 int panthor_mmu_init(struct panthor_device *ptdev); 20 void panthor_mmu_unplug(struct panthor_device *ptdev); 21 void panthor_mmu_pre_reset(struct panthor_device *ptdev); 22 void panthor_mmu_post_reset(struct panthor_device *ptdev); 23 void panthor_mmu_suspend(struct panthor_device *ptdev); 24 void panthor_mmu_resume(struct panthor_device *ptdev); 44 struct panthor_vm *panthor_vm_create(struct panthor_device *ptdev, bool for_mcu, 62 int panthor_vm_pool_create_vm(struct panthor_device *ptdev,
|
| H A D | panthor_gem.c | 21 void panthor_gem_init(struct panthor_device *ptdev) in panthor_gem_init() argument 29 err = drm_gem_huge_mnt_create(&ptdev->base, "within_size"); in panthor_gem_init() 30 if (drm_gem_get_huge_mnt(&ptdev->base)) in panthor_gem_init() 31 drm_info(&ptdev->base, "Using Transparent Hugepage\n"); in panthor_gem_init() 33 drm_warn(&ptdev->base, "Can't use Transparent Hugepage (%d)\n", in panthor_gem_init() 45 struct panthor_device *ptdev = container_of(bo->base.base.dev, in panthor_gem_debugfs_bo_add() local 51 mutex_lock(&ptdev->gems.lock); in panthor_gem_debugfs_bo_add() 52 list_add_tail(&bo->debugfs.node, &ptdev->gems.node); in panthor_gem_debugfs_bo_add() 53 mutex_unlock(&ptdev->gems.lock); in panthor_gem_debugfs_bo_add() 58 struct panthor_device *ptdev = container_of(bo->base.base.dev, in panthor_gem_debugfs_bo_rm() local [all …]
|
| H A D | panthor_heap.c | 88 struct panthor_device *ptdev; member 106 static int panthor_heap_ctx_stride(struct panthor_device *ptdev) in panthor_heap_ctx_stride() argument 108 u32 l2_features = ptdev->gpu_info.l2_features; in panthor_heap_ctx_stride() 116 return panthor_heap_ctx_stride(pool->ptdev) * id; in panthor_get_heap_ctx_offset() 152 chunk->bo = panthor_kernel_bo_create(pool->ptdev, pool->vm, heap->chunk_size, in panthor_alloc_heap_chunk() 337 memset(gpu_ctx, 0, panthor_heap_ctx_stride(pool->ptdev)); in panthor_heap_create() 375 u32 heap_id = (u32)offset / panthor_heap_ctx_stride(pool->ptdev); in panthor_heap_return_chunk() 439 u32 heap_id = (u32)offset / panthor_heap_ctx_stride(pool->ptdev); in panthor_heap_grow() 536 panthor_heap_pool_create(struct panthor_device *ptdev, struct panthor_vm *vm) in panthor_heap_pool_create() argument 539 panthor_heap_ctx_stride(ptdev), in panthor_heap_pool_create() [all …]
|