Searched refs:last_fence (Results 1 – 10 of 10) sorted by relevance
64 fctx->last_fence = 0xffffff00; in msm_fence_context_alloc()65 fctx->completed_fence = fctx->last_fence; in msm_fence_context_alloc()66 *fctx->fenceptr = fctx->last_fence; in msm_fence_context_alloc()200 fctx->context, ++fctx->last_fence); in msm_fence_init()
33 uint32_t last_fence; member
316 if (fence_after(ring->fctx->last_fence, ring->memptrs->fence)) in msm_gpu_active()541 uint32_t last_fence; member
1661 if (q->last_fence) { in xe_exec_queue_last_fence_put_unlocked()1662 dma_fence_put(q->last_fence); in xe_exec_queue_last_fence_put_unlocked()1663 q->last_fence = NULL; in xe_exec_queue_last_fence_put_unlocked()1683 if (q->last_fence && in xe_exec_queue_last_fence_get()1684 test_bit(DMA_FENCE_FLAG_SIGNALED_BIT, &q->last_fence->flags)) in xe_exec_queue_last_fence_get()1687 fence = q->last_fence ? q->last_fence : dma_fence_get_stub(); in xe_exec_queue_last_fence_get()1710 if (q->last_fence && in xe_exec_queue_last_fence_get_for_resume()1711 test_bit(DMA_FENCE_FLAG_SIGNALED_BIT, &q->last_fence->flags)) in xe_exec_queue_last_fence_get_for_resume()1714 fence = q->last_fence ? q->last_fence : dma_fence_get_stub(); in xe_exec_queue_last_fence_get_for_resume()1735 q->last_fence = dma_fence_get(fence); in xe_exec_queue_last_fence_set()[all …]
195 static void free_oa_config_bo(struct xe_oa_config_bo *oa_bo, struct dma_fence *last_fence) in free_oa_config_bo() argument198 xe_bb_free(oa_bo->bb, last_fence); in free_oa_config_bo()710 free_oa_config_bo(oa_bo, stream->last_fence); in xe_oa_free_configs()711 dma_fence_put(stream->last_fence); in xe_oa_free_configs()948 dma_fence_put(stream->last_fence); in xe_oa_update_last_fence()949 stream->last_fence = dma_fence_get(fence); in xe_oa_update_last_fence()
120 dma_fence_put(userq->last_fence); in amdgpu_userq_fence_driver_free()121 userq->last_fence = NULL; in amdgpu_userq_fence_driver_free()282 dma_fence_put(userq->last_fence); in amdgpu_userq_fence_init()283 userq->last_fence = dma_fence_get(&fence->base); in amdgpu_userq_fence_init()440 struct dma_fence *f = userq->last_fence; in amdgpu_userq_fence_driver_force_completion()
83 struct dma_fence *last_fence; member
415 struct dma_fence *f = queue->last_fence; in amdgpu_userq_wait_for_last_fence()1284 struct dma_fence *f = queue->last_fence; in amdgpu_userq_wait_for_signal()
467 struct dma_fence *last_fence; member936 dma_fence_put(queue->fence_ctx.last_fence); in group_free_queue()3313 job->done_fence = dma_fence_get(queue->fence_ctx.last_fence); in queue_run_job()3384 dma_fence_put(queue->fence_ctx.last_fence); in queue_run_job()3385 queue->fence_ctx.last_fence = dma_fence_get(job->done_fence); in queue_run_job()
1151 ring ? ring->id : -1, ring ? ring->fctx->last_fence : 0, in a8xx_fault_detect_irq()