Home
last modified time | relevance | path

Searched refs:ptdev (Results 1 – 12 of 12) sorted by relevance

/linux/drivers/gpu/drm/panthor/
H A Dpanthor_pwr.c55 static void panthor_pwr_irq_handler(struct panthor_device *ptdev, u32 status) in panthor_pwr_irq_handler() argument
57 spin_lock(&ptdev->pwr->reqs_lock); in panthor_pwr_irq_handler()
58 gpu_write(ptdev, PWR_INT_CLEAR, status); in panthor_pwr_irq_handler()
61 drm_err(&ptdev->base, "PWR_IRQ: COMMAND_NOT_ALLOWED"); in panthor_pwr_irq_handler()
64 drm_err(&ptdev->base, "PWR_IRQ: COMMAND_INVALID"); in panthor_pwr_irq_handler()
66 if (status & ptdev->pwr->pending_reqs) { in panthor_pwr_irq_handler()
67 ptdev->pwr->pending_reqs &= ~status; in panthor_pwr_irq_handler()
68 wake_up_all(&ptdev->pwr->reqs_acked); in panthor_pwr_irq_handler()
70 spin_unlock(&ptdev->pwr->reqs_lock); in panthor_pwr_irq_handler()
74 static void panthor_pwr_write_command(struct panthor_device *ptdev, u32 command, u64 args) in panthor_pwr_write_command() argument
[all …]
H A Dpanthor_device.h71 struct panthor_device *ptdev; member
250 struct panthor_device *ptdev; member
280 int panthor_device_init(struct panthor_device *ptdev);
281 void panthor_device_unplug(struct panthor_device *ptdev);
286 static inline void panthor_device_schedule_reset(struct panthor_device *ptdev) in panthor_device_schedule_reset() argument
288 if (!atomic_cmpxchg(&ptdev->reset.pending, 0, 1) && in panthor_device_schedule_reset()
289 atomic_read(&ptdev->pm.state) == PANTHOR_DEVICE_PM_STATE_ACTIVE) in panthor_device_schedule_reset()
290 queue_work(ptdev->reset.wq, &ptdev->reset.work); in panthor_device_schedule_reset()
298 static inline bool panthor_device_reset_is_pending(struct panthor_device *ptdev) in panthor_device_reset_is_pending() argument
300 return atomic_read(&ptdev->reset.pending) != 0; in panthor_device_reset_is_pending()
[all …]
H A Dpanthor_mmu.c250 struct panthor_device *ptdev; member
442 drm_WARN_ON(&vm->ptdev->base, vm->op_ctx); in alloc_pt()
443 p = alloc_pages_node(dev_to_node(vm->ptdev->base.dev), in alloc_pt()
453 if (drm_WARN_ON(&vm->ptdev->base, size != SZ_4K)) in alloc_pt()
459 if (drm_WARN_ON(&vm->ptdev->base, !vm->op_ctx) || in alloc_pt()
460 drm_WARN_ON(&vm->ptdev->base, in alloc_pt()
495 if (drm_WARN_ON(&vm->ptdev->base, size != SZ_4K)) in free_pt()
502 static int wait_ready(struct panthor_device *ptdev, u32 as_nr) in wait_ready() argument
510 ret = gpu_read_relaxed_poll_timeout_atomic(ptdev, AS_STATUS(as_nr), val, in wait_ready()
515 panthor_device_schedule_reset(ptdev); in wait_ready()
[all …]
H A Dpanthor_sched.c146 struct panthor_device *ptdev; member
535 struct panthor_device *ptdev; member
734 if (!queue_work((group)->ptdev->scheduler->wq, &(group)->wname ## _work)) \
748 !panthor_device_reset_is_pending((sched)->ptdev)) \
764 !panthor_device_reset_is_pending((sched)->ptdev)) \
868 struct panthor_device *ptdev = group->ptdev; in panthor_queue_get_syncwait_obj() local
882 if (drm_WARN_ON(&ptdev->base, IS_ERR_OR_NULL(bo))) in panthor_queue_get_syncwait_obj()
887 if (drm_WARN_ON(&ptdev->base, ret)) in panthor_queue_get_syncwait_obj()
891 if (drm_WARN_ON(&ptdev->base, !queue->syncwait.kmap)) in panthor_queue_get_syncwait_obj()
967 struct panthor_device *ptdev = group->ptdev; in group_release() local
[all …]
H A Dpanthor_hw.h15 int (*soft_reset)(struct panthor_device *ptdev);
18 void (*l2_power_off)(struct panthor_device *ptdev);
21 int (*l2_power_on)(struct panthor_device *ptdev);
34 int panthor_hw_init(struct panthor_device *ptdev);
36 static inline int panthor_hw_soft_reset(struct panthor_device *ptdev) in panthor_hw_soft_reset() argument
38 return ptdev->hw->ops.soft_reset(ptdev); in panthor_hw_soft_reset()
41 static inline int panthor_hw_l2_power_on(struct panthor_device *ptdev) in panthor_hw_l2_power_on() argument
43 return ptdev->hw->ops.l2_power_on(ptdev); in panthor_hw_l2_power_on()
46 static inline void panthor_hw_l2_power_off(struct panthor_device *ptdev) in panthor_hw_l2_power_off() argument
48 ptdev->hw->ops.l2_power_off(ptdev); in panthor_hw_l2_power_off()
[all …]
H A Dpanthor_gpu.h12 int panthor_gpu_init(struct panthor_device *ptdev);
13 void panthor_gpu_unplug(struct panthor_device *ptdev);
14 void panthor_gpu_suspend(struct panthor_device *ptdev);
15 void panthor_gpu_resume(struct panthor_device *ptdev);
17 int panthor_gpu_block_power_on(struct panthor_device *ptdev,
21 int panthor_gpu_block_power_off(struct panthor_device *ptdev,
31 #define panthor_gpu_power_on(ptdev, type, mask, timeout_us) \ argument
32 panthor_gpu_block_power_on(ptdev, #type, \
43 #define panthor_gpu_power_off(ptdev, type, mask, timeout_us) \ argument
44 panthor_gpu_block_power_off(ptdev, #type, \
[all …]
H A Dpanthor_pwr.h9 void panthor_pwr_unplug(struct panthor_device *ptdev);
11 int panthor_pwr_init(struct panthor_device *ptdev);
13 int panthor_pwr_reset_soft(struct panthor_device *ptdev);
15 void panthor_pwr_l2_power_off(struct panthor_device *ptdev);
17 int panthor_pwr_l2_power_on(struct panthor_device *ptdev);
19 void panthor_pwr_suspend(struct panthor_device *ptdev);
21 void panthor_pwr_resume(struct panthor_device *ptdev);
H A Dpanthor_fw.h479 panthor_fw_get_glb_iface(struct panthor_device *ptdev);
482 panthor_fw_get_csg_iface(struct panthor_device *ptdev, u32 csg_slot);
485 panthor_fw_get_cs_iface(struct panthor_device *ptdev, u32 csg_slot, u32 cs_slot);
487 u64 panthor_fw_csg_endpoint_req_get(struct panthor_device *ptdev,
490 void panthor_fw_csg_endpoint_req_set(struct panthor_device *ptdev,
493 void panthor_fw_csg_endpoint_req_update(struct panthor_device *ptdev,
497 int panthor_fw_csg_wait_acks(struct panthor_device *ptdev, u32 csg_id, u32 req_mask,
500 int panthor_fw_glb_wait_acks(struct panthor_device *ptdev, u32 req_mask, u32 *acked,
503 void panthor_fw_ring_csg_doorbells(struct panthor_device *ptdev, u32 csg_slot);
506 panthor_fw_alloc_queue_iface_mem(struct panthor_device *ptdev,
[all …]
H A Dpanthor_drv.c502 struct panthor_device *ptdev = container_of(ctx->file->minor->dev, in panthor_submit_ctx_update_job_sync_signal_fences() local
518 if (drm_WARN_ON(&ptdev->base, !sig_sync)) in panthor_submit_ctx_update_job_sync_signal_fences()
525 if (drm_WARN_ON(&ptdev->base, !sig_sync->fence)) in panthor_submit_ctx_update_job_sync_signal_fences()
603 struct panthor_device *ptdev = container_of(ctx->file->minor->dev, in panthor_submit_ctx_add_sync_deps_to_job() local
625 if (drm_WARN_ON(&ptdev->base, !sig_sync->fence)) in panthor_submit_ctx_add_sync_deps_to_job()
764 static int panthor_query_timestamp_info(struct panthor_device *ptdev, in panthor_query_timestamp_info() argument
769 ret = panthor_device_resume_and_get(ptdev); in panthor_query_timestamp_info()
778 arg->current_timestamp = gpu_read64_counter(ptdev, GPU_TIMESTAMP); in panthor_query_timestamp_info()
779 arg->timestamp_offset = gpu_read64(ptdev, GPU_TIMESTAMP_OFFSET); in panthor_query_timestamp_info()
781 pm_runtime_put(ptdev->base.dev); in panthor_query_timestamp_info()
[all …]
H A Dpanthor_mmu.h19 int panthor_mmu_init(struct panthor_device *ptdev);
20 void panthor_mmu_unplug(struct panthor_device *ptdev);
21 void panthor_mmu_pre_reset(struct panthor_device *ptdev);
22 void panthor_mmu_post_reset(struct panthor_device *ptdev);
23 void panthor_mmu_suspend(struct panthor_device *ptdev);
24 void panthor_mmu_resume(struct panthor_device *ptdev);
44 struct panthor_vm *panthor_vm_create(struct panthor_device *ptdev, bool for_mcu,
62 int panthor_vm_pool_create_vm(struct panthor_device *ptdev,
H A Dpanthor_gem.c21 void panthor_gem_init(struct panthor_device *ptdev) in panthor_gem_init() argument
29 err = drm_gem_huge_mnt_create(&ptdev->base, "within_size"); in panthor_gem_init()
30 if (drm_gem_get_huge_mnt(&ptdev->base)) in panthor_gem_init()
31 drm_info(&ptdev->base, "Using Transparent Hugepage\n"); in panthor_gem_init()
33 drm_warn(&ptdev->base, "Can't use Transparent Hugepage (%d)\n", in panthor_gem_init()
45 struct panthor_device *ptdev = container_of(bo->base.base.dev, in panthor_gem_debugfs_bo_add() local
51 mutex_lock(&ptdev->gems.lock); in panthor_gem_debugfs_bo_add()
52 list_add_tail(&bo->debugfs.node, &ptdev->gems.node); in panthor_gem_debugfs_bo_add()
53 mutex_unlock(&ptdev->gems.lock); in panthor_gem_debugfs_bo_add()
58 struct panthor_device *ptdev = container_of(bo->base.base.dev, in panthor_gem_debugfs_bo_rm() local
[all …]
H A Dpanthor_heap.c88 struct panthor_device *ptdev; member
106 static int panthor_heap_ctx_stride(struct panthor_device *ptdev) in panthor_heap_ctx_stride() argument
108 u32 l2_features = ptdev->gpu_info.l2_features; in panthor_heap_ctx_stride()
116 return panthor_heap_ctx_stride(pool->ptdev) * id; in panthor_get_heap_ctx_offset()
152 chunk->bo = panthor_kernel_bo_create(pool->ptdev, pool->vm, heap->chunk_size, in panthor_alloc_heap_chunk()
337 memset(gpu_ctx, 0, panthor_heap_ctx_stride(pool->ptdev)); in panthor_heap_create()
375 u32 heap_id = (u32)offset / panthor_heap_ctx_stride(pool->ptdev); in panthor_heap_return_chunk()
439 u32 heap_id = (u32)offset / panthor_heap_ctx_stride(pool->ptdev); in panthor_heap_grow()
536 panthor_heap_pool_create(struct panthor_device *ptdev, struct panthor_vm *vm) in panthor_heap_pool_create() argument
539 panthor_heap_ctx_stride(ptdev), in panthor_heap_pool_create()
[all …]