Home
last modified time | relevance | path

Searched refs:fence_drv (Results 1 – 25 of 27) sorted by relevance

12

/linux/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_fence.c65 struct amdgpu_fence_driver *drv = &ring->fence_drv; in amdgpu_fence_write()
81 struct amdgpu_fence_driver *drv = &ring->fence_drv; in amdgpu_fence_read()
124 seq = ++ring->fence_drv.sync_seq; in amdgpu_fence_emit()
126 &ring->fence_drv.lock, in amdgpu_fence_emit()
130 amdgpu_ring_emit_fence(ring, ring->fence_drv.gpu_addr, in amdgpu_fence_emit()
135 ptr = &ring->fence_drv.fences[seq & ring->fence_drv.num_fences_mask]; in amdgpu_fence_emit()
181 seq = ++ring->fence_drv.sync_seq; in amdgpu_fence_emit_polling()
183 seq - ring->fence_drv.num_fences_mask, in amdgpu_fence_emit_polling()
188 amdgpu_ring_emit_fence(ring, ring->fence_drv in amdgpu_fence_emit_polling()
[all...]
H A Damdgpu_userq_fence.c65 static u64 amdgpu_userq_fence_read(struct amdgpu_userq_fence_driver *fence_drv) in amdgpu_userq_fence_read() argument
67 return le64_to_cpu(*fence_drv->cpu_addr); in amdgpu_userq_fence_read()
71 amdgpu_userq_fence_write(struct amdgpu_userq_fence_driver *fence_drv, in amdgpu_userq_fence_write() argument
74 if (fence_drv->cpu_addr) in amdgpu_userq_fence_write()
75 *fence_drv->cpu_addr = cpu_to_le64(seq); in amdgpu_userq_fence_write()
81 struct amdgpu_userq_fence_driver *fence_drv; in amdgpu_userq_fence_driver_alloc() local
85 fence_drv = kzalloc(sizeof(*fence_drv), GFP_KERNEL); in amdgpu_userq_fence_driver_alloc()
86 if (!fence_drv) in amdgpu_userq_fence_driver_alloc()
90 r = amdgpu_seq64_alloc(adev, &fence_drv in amdgpu_userq_fence_driver_alloc()
126 struct amdgpu_userq_fence_driver *fence_drv; amdgpu_userq_walk_and_drop_fence_drv() local
152 amdgpu_userq_fence_driver_process(struct amdgpu_userq_fence_driver * fence_drv) amdgpu_userq_fence_driver_process() argument
185 struct amdgpu_userq_fence_driver *fence_drv = container_of(ref, amdgpu_userq_fence_driver_destroy() local
220 amdgpu_userq_fence_driver_get(struct amdgpu_userq_fence_driver * fence_drv) amdgpu_userq_fence_driver_get() argument
225 amdgpu_userq_fence_driver_put(struct amdgpu_userq_fence_driver * fence_drv) amdgpu_userq_fence_driver_put() argument
240 struct amdgpu_userq_fence_driver *fence_drv; amdgpu_userq_fence_create() local
317 struct amdgpu_userq_fence_driver *fence_drv = fence->fence_drv; amdgpu_userq_fence_signaled() local
333 struct amdgpu_userq_fence_driver *fence_drv = userq_fence->fence_drv; amdgpu_userq_fence_free() local
427 struct amdgpu_userq_fence_driver *fence_drv = fence->fence_drv; amdgpu_userq_fence_driver_set_error() local
447 struct amdgpu_userq_fence_driver *fence_drv = fence->fence_drv; amdgpu_userq_fence_driver_force_completion() local
919 struct amdgpu_userq_fence_driver *fence_drv; amdgpu_userq_wait_ioctl() local
[all...]
H A Dumsch_mm_v4_0.c310 set_hw_resources.api_status.api_completion_fence_addr = umsch->ring.fence_drv.gpu_addr; in umsch_mm_v4_0_set_hw_resources()
311 set_hw_resources.api_status.api_completion_fence_value = ++umsch->ring.fence_drv.sync_seq; in umsch_mm_v4_0_set_hw_resources()
360 add_queue.api_status.api_completion_fence_addr = umsch->ring.fence_drv.gpu_addr; in umsch_mm_v4_0_add_queue()
361 add_queue.api_status.api_completion_fence_value = ++umsch->ring.fence_drv.sync_seq; in umsch_mm_v4_0_add_queue()
392 remove_queue.api_status.api_completion_fence_addr = umsch->ring.fence_drv.gpu_addr; in umsch_mm_v4_0_remove_queue()
393 remove_queue.api_status.api_completion_fence_value = ++umsch->ring.fence_drv.sync_seq; in umsch_mm_v4_0_remove_queue()
H A Damdgpu_ring_mux.c99 last_seq = atomic_read(&e->ring->fence_drv.last_seq); in amdgpu_mux_resubmit_chunks()
109 le32_to_cpu(*(e->ring->fence_drv.cpu_addr + 2))) { in amdgpu_mux_resubmit_chunks()
474 last_seq = atomic_read(&ring->fence_drv.last_seq); in scan_and_remove_signaled_chunk()
537 chunk->sync_seq = READ_ONCE(ring->fence_drv.sync_seq); in amdgpu_ring_mux_end_ib()
570 mux->seqno_to_resubmit = ring->fence_drv.sync_seq; in amdgpu_mcbp_handle_trailing_fence_irq()
H A Damdgpu_userq.h70 struct amdgpu_userq_fence_driver *fence_drv; member
H A Damdgpu_job.c123 job->base.sched->name, atomic_read(&ring->fence_drv.last_seq), in amdgpu_job_timedout()
124 ring->fence_drv.sync_seq); in amdgpu_job_timedout()
H A Dsdma_v6_0.c1169 uint32_t seq = ring->fence_drv.sync_seq; in sdma_v6_0_ring_emit_pipeline_sync()
1170 uint64_t addr = ring->fence_drv.gpu_addr; in sdma_v6_0_ring_emit_pipeline_sync()
1668 struct amdgpu_userq_fence_driver *fence_drv = NULL; in sdma_v6_0_process_fence_irq()
1675 fence_drv = xa_load(xa, doorbell_offset); in sdma_v6_0_process_illegal_inst_irq()
1676 if (fence_drv) in sdma_v6_0_process_illegal_inst_irq()
1677 amdgpu_userq_fence_driver_process(fence_drv);
1655 struct amdgpu_userq_fence_driver *fence_drv = NULL; sdma_v6_0_process_fence_irq() local
H A Damdgpu_vpe.c569 uint32_t seq = ring->fence_drv.sync_seq; in vpe_ring_emit_pipeline_sync()
570 uint64_t addr = ring->fence_drv.gpu_addr; in vpe_ring_emit_pipeline_sync()
H A Dmes_userqueue.c307 userq_props->fence_address = queue->fence_drv->gpu_addr; in mes_userq_mqd_create()
H A Dmes_v11_0.c208 seq = ++ring->fence_drv.sync_seq; in mes_v11_0_submit_pkt_and_poll_completion()
210 seq - ring->fence_drv.num_fences_mask, in mes_v11_0_submit_pkt_and_poll_completion()
226 ring->fence_drv.gpu_addr; in mes_v11_0_submit_pkt_and_poll_completion()
H A Dmes_v12_0.c191 seq = ++ring->fence_drv.sync_seq; in mes_v12_0_submit_pkt_and_poll_completion()
193 seq - ring->fence_drv.num_fences_mask, in mes_v12_0_submit_pkt_and_poll_completion()
209 ring->fence_drv.gpu_addr; in mes_v12_0_submit_pkt_and_poll_completion()
H A Damdgpu_ring.h308 struct amdgpu_fence_driver fence_drv;
306 struct amdgpu_fence_driver fence_drv; global() member
H A Dgfx_v12_0.c4485 uint32_t seq = ring->fence_drv.sync_seq; in gfx_v12_0_ring_emit_fence()
4486 uint64_t addr = ring->fence_drv.gpu_addr;
4824 struct amdgpu_userq_fence_driver *fence_drv = NULL; in gfx_v12_0_eop_irq()
4829 fence_drv = xa_load(xa, doorbell_offset); in gfx_v12_0_eop_irq() local
4830 if (fence_drv) in gfx_v12_0_eop_irq()
4831 amdgpu_userq_fence_driver_process(fence_drv); in gfx_v12_0_eop_irq()
H A Dgfx_v11_0.c5969 uint32_t seq = ring->fence_drv.sync_seq; in gfx_v11_0_ring_emit_pipeline_sync()
5970 uint64_t addr = ring->fence_drv.gpu_addr;
6458 struct amdgpu_userq_fence_driver *fence_drv = NULL; in gfx_v11_0_eop_irq()
6463 fence_drv = xa_load(xa, doorbell_offset); in gfx_v11_0_eop_irq()
6464 if (fence_drv) in gfx_v11_0_eop_irq()
6465 amdgpu_userq_fence_driver_process(fence_drv); in gfx_v11_0_eop_irq()
6453 struct amdgpu_userq_fence_driver *fence_drv = NULL; gfx_v11_0_eop_irq() local
H A Dgfx_v6_0.c2289 uint32_t seq = ring->fence_drv.sync_seq; in gfx_v6_0_ring_emit_pipeline_sync()
2290 uint64_t addr = ring->fence_drv.gpu_addr; in gfx_v6_0_ring_emit_pipeline_sync()
H A Dgfx_v7_0.c3101 uint32_t seq = ring->fence_drv.sync_seq;
3102 uint64_t addr = ring->fence_drv.gpu_addr;
/linux/drivers/gpu/drm/radeon/
H A Dradeon_fence.c69 struct radeon_fence_driver *drv = &rdev->fence_drv[ring]; in radeon_fence_write()
90 struct radeon_fence_driver *drv = &rdev->fence_drv[ring]; in radeon_fence_read()
119 &rdev->fence_drv[ring].lockup_work, in radeon_fence_schedule_check()
145 (*fence)->seq = seq = ++rdev->fence_drv[ring].sync_seq[ring]; in radeon_fence_emit()
177 seq = atomic64_read(&fence->rdev->fence_drv[fence->ring].last_seq); in radeon_fence_check_signaled()
224 last_seq = atomic64_read(&rdev->fence_drv[ring].last_seq); in radeon_fence_activity()
226 last_emitted = rdev->fence_drv[ring].sync_seq[ring]; in radeon_fence_activity()
251 } while (atomic64_xchg(&rdev->fence_drv[ring].last_seq, seq) > seq); in radeon_fence_activity()
269 struct radeon_fence_driver *fence_drv; in radeon_fence_check_lockup() local
273 fence_drv = container_of(work, struct radeon_fence_driver, in radeon_fence_check_lockup()
[all …]
H A Duvd_v2_2.c43 uint64_t addr = rdev->fence_drv[fence->ring].gpu_addr; in uvd_v2_2_fence_emit()
H A Devergreen_dma.c44 u64 addr = rdev->fence_drv[fence->ring].gpu_addr; in evergreen_dma_fence_ring_emit()
H A Duvd_v1_0.c85 uint64_t addr = rdev->fence_drv[fence->ring].gpu_addr; in uvd_v1_0_fence_emit()
H A Dr600_dma.c290 u64 addr = rdev->fence_drv[fence->ring].gpu_addr; in r600_dma_fence_ring_emit()
H A Dcik_sdma.c203 u64 addr = rdev->fence_drv[fence->ring].gpu_addr; in cik_sdma_fence_ring_emit()
H A Dni.c1380 u64 addr = rdev->fence_drv[fence->ring].gpu_addr; in cayman_fence_ring_emit()
/linux/drivers/gpu/drm/virtio/
H A Dvirtgpu_ioctl.c172 fence = virtio_gpu_fence_alloc(vgdev, vgdev->fence_drv.context, 0); in virtio_gpu_resource_create_ioctl()
259 fence = virtio_gpu_fence_alloc(vgdev, vgdev->fence_drv.context, 0); in virtio_gpu_transfer_from_host_ioctl()
319 fence = virtio_gpu_fence_alloc(vgdev, vgdev->fence_drv.context, in virtio_gpu_transfer_to_host_ioctl()
H A Dvirtgpu_submit.c478 u64 fence_ctx = vgdev->fence_drv.context; in virtio_gpu_execbuffer_ioctl()

12