| /linux/drivers/gpu/drm/i915/ |
| H A D | i915_sw_fence.c | 8 #include <linux/dma-fence.h> 49 static inline void debug_fence_init(struct i915_sw_fence *fence) in debug_fence_init() argument 51 debug_object_init(fence, &i915_sw_fence_debug_descr); in debug_fence_init() 54 static inline __maybe_unused void debug_fence_init_onstack(struct i915_sw_fence *fence) in debug_fence_init_onstack() argument 56 debug_object_init_on_stack(fence, &i915_sw_fence_debug_descr); in debug_fence_init_onstack() 59 static inline void debug_fence_activate(struct i915_sw_fence *fence) in debug_fence_activate() argument 61 debug_object_activate(fence, &i915_sw_fence_debug_descr); in debug_fence_activate() 64 static inline void debug_fence_set_state(struct i915_sw_fence *fence, in debug_fence_set_state() argument 67 debug_object_active_state(fence, &i915_sw_fence_debug_descr, old, new); in debug_fence_set_state() 70 static inline void debug_fence_deactivate(struct i915_sw_fence *fence) in debug_fence_deactivate() argument [all …]
|
| H A D | i915_sw_fence.h | 12 #include <linux/dma-fence.h> 42 void __i915_sw_fence_init(struct i915_sw_fence *fence, 47 #define i915_sw_fence_init(fence, fn) \ argument 52 __i915_sw_fence_init((fence), (fn), #fence, &__key); \ 55 #define i915_sw_fence_init(fence, fn) \ argument 58 __i915_sw_fence_init((fence), (fn), NULL, NULL); \ 62 void i915_sw_fence_reinit(struct i915_sw_fence *fence); 65 void i915_sw_fence_fini(struct i915_sw_fence *fence); 67 static inline void i915_sw_fence_fini(struct i915_sw_fence *fence) {} in i915_sw_fence_fini() argument 70 void i915_sw_fence_commit(struct i915_sw_fence *fence); [all …]
|
| H A D | i915_deps.c | 6 #include <linux/dma-fence.h> 18 * the migration fence with the unbind fences if these are coalesced 33 /* Min number of fence pointers in the array when an allocation occurs. */ 61 * This function drops all fence references taken, conditionally frees and 75 static int i915_deps_grow(struct i915_deps *deps, struct dma_fence *fence, in i915_deps_grow() argument 96 deps->fences[deps->num_deps++] = dma_fence_get(fence); in i915_deps_grow() 100 if (ctx->no_wait_gpu && !dma_fence_is_signaled(fence)) { in i915_deps_grow() 105 ret = dma_fence_wait(fence, ctx->interruptible); in i915_deps_grow() 109 ret = fence->error; in i915_deps_grow() 127 * encounters an error during the wait or a fence error, the wait for [all …]
|
| /linux/drivers/dma-buf/ |
| H A D | dma-fence.c | 3 * Fence mechanism for dma-buf and to allow for asynchronous dma access 16 #include <linux/dma-fence.h> 31 * fence context counter: each execution context should have its own 32 * fence context, this allows checking if fences belong to the same 45 * A fence is initialized using dma_fence_init() and completed using 60 * fence to be updated. 69 * DOC: fence cross-driver contract 95 * This means any code required for fence completion cannot acquire a 100 * callbacks. This means any code required for fence completion cannot 105 * for fence completion cannot allocate memory with GFP_NOFS or GFP_NOIO. [all …]
|
| H A D | dma-resv.c | 37 #include <linux/dma-fence-array.h> 49 * can have any number of fences attaches to it. Each fence carries an usage 50 * parameter determining how the operation represented by the fence is using the 60 /* Mask for the lower fence pointer bits */ 69 /* Extract the fence and usage flags from an RCU protected entry in the list. */ 71 struct dma_resv *resv, struct dma_fence **fence, in dma_resv_list_entry() argument 78 *fence = (struct dma_fence *)(tmp & ~DMA_RESV_LIST_MASK); in dma_resv_list_entry() 83 /* Set the fence and usage flags at the specific index in the list. */ 86 struct dma_fence *fence, in dma_resv_list_set() argument 89 long tmp = ((long)fence) | usage; in dma_resv_list_set() [all …]
|
| H A D | sync_file.c | 8 #include <linux/dma-fence-unwrap.h> 58 * @fence: fence to add to the sync_fence 60 * Creates a sync_file containg @fence. This function acquires and additional 61 * reference of @fence for the newly-created &sync_file, if it succeeds. The 65 struct sync_file *sync_file_create(struct dma_fence *fence) in sync_file_create() argument 73 sync_file->fence = dma_fence_get(fence); in sync_file_create() 97 * sync_file_get_fence - get the fence related to the sync_file fd 98 * @fd: sync_file fd to get the fence from 100 * Ensures @fd references a valid sync_file and returns a fence that 101 * represents all fence in the sync_file. On error NULL is returned. [all …]
|
| H A D | st-dma-fence-chain.c | 8 #include <linux/dma-fence.h> 9 #include <linux/dma-fence-chain.h> 62 struct dma_fence *fence, in mock_chain() argument 71 dma_fence_chain_init(f, dma_fence_get(prev), dma_fence_get(fence), in mock_chain() 182 struct dma_fence *fence; in find_seqno() local 190 fence = dma_fence_get(fc.tail); in find_seqno() 191 err = dma_fence_chain_find_seqno(&fence, 0); in find_seqno() 192 dma_fence_put(fence); in find_seqno() 199 fence = dma_fence_get(fc.tail); in find_seqno() 200 err = dma_fence_chain_find_seqno(&fence, i + 1); in find_seqno() [all …]
|
| H A D | dma-fence-array.c | 3 * dma-fence-array: aggregate fences to be waited together 14 #include <linux/dma-fence-array.h> 18 static const char *dma_fence_array_get_driver_name(struct dma_fence *fence) in dma_fence_array_get_driver_name() argument 23 static const char *dma_fence_array_get_timeline_name(struct dma_fence *fence) in dma_fence_array_get_timeline_name() argument 70 static bool dma_fence_array_enable_signaling(struct dma_fence *fence) in dma_fence_array_enable_signaling() argument 72 struct dma_fence_array *array = to_dma_fence_array(fence); in dma_fence_array_enable_signaling() 79 * As we may report that the fence is signaled before all in dma_fence_array_enable_signaling() 82 * early. The core fence handling will only hold the reference in dma_fence_array_enable_signaling() 103 static bool dma_fence_array_signaled(struct dma_fence *fence) in dma_fence_array_signaled() argument 105 struct dma_fence_array *array = to_dma_fence_array(fence); in dma_fence_array_signaled() [all …]
|
| H A D | sw_sync.c | 42 * from the timeline and signal any fence that has a seqno smaller or equal 46 * @value: the seqno to initialise the fence with 48 * @fence: return the fd of the new sync_file with the created fence 53 __s32 fence; /* fd of new fence */ member 57 * struct sw_sync_get_deadline - get the deadline hint of a sw_sync fence 60 * @fence_fd: the sw_sync fence fd (in) 62 * Return the earliest deadline set on the fence. The timebase for the 64 * set on the fence, this ioctl will return -ENOENT. 86 static inline struct sync_pt *dma_fence_to_sync_pt(struct dma_fence *fence) in dma_fence_to_sync_pt() argument 88 if (fence->ops != &timeline_fence_ops) in dma_fence_to_sync_pt() [all …]
|
| /linux/drivers/gpu/drm/vmwgfx/ |
| H A D | vmwgfx_fence.c | 23 struct vmw_fence_obj fence; member 27 * struct vmw_event_fence_action - fence callback that delivers a DRM event. 33 * current time tv_sec val when the fence signals. 35 * be assigned the current time tv_usec val when the fence signals. 48 fman_from_fence(struct vmw_fence_obj *fence) in fman_from_fence() argument 50 return container_of(fence->base.lock, struct vmw_fence_manager, lock); in fman_from_fence() 55 struct vmw_fence_obj *fence = in vmw_fence_obj_destroy() local 57 struct vmw_fence_manager *fman = fman_from_fence(fence); in vmw_fence_obj_destroy() 59 if (!list_empty(&fence->head)) { in vmw_fence_obj_destroy() 60 /* The fence manager still has an implicit reference to this in vmw_fence_obj_destroy() [all …]
|
| /linux/drivers/gpu/drm/i915/gt/ |
| H A D | intel_ggtt_fencing.c | 20 * DOC: fence register handling 39 * dynamically associated with objects. Furthermore fence state is committed to 43 * cases the fence can be removed forcefully with i915_gem_object_put_fence(). 51 static struct drm_i915_private *fence_to_i915(struct i915_fence_reg *fence) in fence_to_i915() argument 53 return fence->ggtt->vm.i915; in fence_to_i915() 56 static struct intel_uncore *fence_to_uncore(struct i915_fence_reg *fence) in fence_to_uncore() argument 58 return fence->ggtt->vm.gt->uncore; in fence_to_uncore() 61 static void i965_write_fence_reg(struct i915_fence_reg *fence) in i965_write_fence_reg() argument 67 if (GRAPHICS_VER(fence_to_i915(fence)) >= 6) { in i965_write_fence_reg() 68 fence_reg_lo = FENCE_REG_GEN6_LO(fence->id); in i965_write_fence_reg() [all …]
|
| /linux/drivers/gpu/drm/radeon/ |
| H A D | radeon_fence.c | 50 * for GPU/CPU synchronization. When the fence is written, 51 * it is expected that all buffers associated with that fence 59 * radeon_fence_write - write a fence value 63 * @ring: ring index the fence is associated with 65 * Writes a fence value to memory or a scratch register (all asics). 80 * radeon_fence_read - read a fence value 83 * @ring: ring index the fence is associated with 85 * Reads a fence value from memory or a scratch register (all asics). 86 * Returns the value of the fence read from memory or register. 124 * radeon_fence_emit - emit a fence on the requested ring [all …]
|
| /linux/drivers/gpu/drm/nouveau/ |
| H A D | nouveau_fence.c | 42 nouveau_fctx(struct nouveau_fence *fence) in nouveau_fctx() argument 44 return container_of(fence->base.lock, struct nouveau_fence_chan, lock); in nouveau_fctx() 48 nouveau_fence_signal(struct nouveau_fence *fence) in nouveau_fence_signal() argument 52 dma_fence_signal_locked(&fence->base); in nouveau_fence_signal() 53 list_del(&fence->head); in nouveau_fence_signal() 54 rcu_assign_pointer(fence->channel, NULL); in nouveau_fence_signal() 56 if (test_bit(DMA_FENCE_FLAG_USER_BITS, &fence->base.flags)) { in nouveau_fence_signal() 57 struct nouveau_fence_chan *fctx = nouveau_fctx(fence); in nouveau_fence_signal() 63 dma_fence_put(&fence->base); in nouveau_fence_signal() 68 nouveau_local_fence(struct dma_fence *fence, struct nouveau_drm *drm) in nouveau_local_fence() argument [all …]
|
| /linux/drivers/gpu/drm/vgem/ |
| H A D | vgem_fence.c | 38 static const char *vgem_fence_get_driver_name(struct dma_fence *fence) in vgem_fence_get_driver_name() argument 43 static const char *vgem_fence_get_timeline_name(struct dma_fence *fence) in vgem_fence_get_timeline_name() argument 50 struct vgem_fence *fence = container_of(base, typeof(*fence), base); in vgem_fence_release() local 52 timer_delete_sync(&fence->timer); in vgem_fence_release() 53 dma_fence_free(&fence->base); in vgem_fence_release() 64 struct vgem_fence *fence = timer_container_of(fence, t, timer); in vgem_fence_timeout() local 66 dma_fence_signal(&fence->base); in vgem_fence_timeout() 72 struct vgem_fence *fence; in vgem_fence_create() local 74 fence = kzalloc_obj(*fence); in vgem_fence_create() 75 if (!fence) in vgem_fence_create() [all …]
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | amdgpu_amdkfd_fence.c | 23 #include <linux/dma-fence.h> 36 /* Eviction Fence 37 * Fence helper functions to deal with KFD memory eviction. 41 * All the BOs in a process share an eviction fence. When process X wants 59 * user queues and signal fence. The work item will also start another delayed 68 struct amdgpu_amdkfd_fence *fence; in amdgpu_amdkfd_fence_create() local 70 fence = kzalloc_obj(*fence); in amdgpu_amdkfd_fence_create() 71 if (fence == NULL) in amdgpu_amdkfd_fence_create() 76 fence->mm = mm; in amdgpu_amdkfd_fence_create() 77 get_task_comm(fence->timeline_name, current); in amdgpu_amdkfd_fence_create() [all …]
|
| H A D | amdgpu_fence.c | 56 * amdgpu_fence_write - write a fence value 58 * @ring: ring the fence is associated with 61 * Writes a fence value to memory (all asics). 72 * amdgpu_fence_read - read a fence value 74 * @ring: ring the fence is associated with 76 * Reads a fence value from memory (all asics). 77 * Returns the value of the fence read from memory. 103 * amdgpu_fence_emit - emit a fence on the requested ring 105 * @ring: ring the fence is associated with 106 * @af: amdgpu fence input [all …]
|
| H A D | amdgpu_userq_fence.c | 27 #include <linux/dma-fence-unwrap.h> 157 struct dma_fence *fence; in amdgpu_userq_fence_driver_process() local 169 fence = &userq_fence->base; in amdgpu_userq_fence_driver_process() 171 if (rptr < fence->seqno) in amdgpu_userq_fence_driver_process() 174 dma_fence_signal(fence); in amdgpu_userq_fence_driver_process() 180 dma_fence_put(fence); in amdgpu_userq_fence_driver_process() 192 struct amdgpu_userq_fence *fence, *tmp; in amdgpu_userq_fence_driver_destroy() local 198 list_for_each_entry_safe(fence, tmp, &fence_drv->fences, link) { in amdgpu_userq_fence_driver_destroy() 199 f = &fence->base; in amdgpu_userq_fence_driver_destroy() 206 list_del(&fence->link); in amdgpu_userq_fence_driver_destroy() [all …]
|
| /linux/include/linux/ |
| H A D | dma-fence-chain.h | 3 * fence-chain: chain fences together in a timeline 13 #include <linux/dma-fence.h> 18 * struct dma_fence_chain - fence to represent an node of a fence chain 19 * @base: fence base class 20 * @prev: previous fence of the chain 22 * @fence: encapsulated fence 23 * @lock: spinlock for fence handling 29 struct dma_fence *fence; member 35 * complection of the fence chain. Never used at the same time 54 * to_dma_fence_chain - cast a fence to a dma_fence_chain [all …]
|
| H A D | dma-fence-array.h | 3 * fence-array: aggregates fence to be waited together 15 #include <linux/dma-fence.h> 19 * struct dma_fence_array_cb - callback helper for fence array 20 * @cb: fence callback structure for signaling 21 * @array: reference to the parent fence array object 29 * struct dma_fence_array - fence to represent an array of fences 30 * @base: fence base class 31 * @lock: spinlock for fence handling 52 * to_dma_fence_array - cast a fence to a dma_fence_array 53 * @fence: fence to cast to a dma_fence_array [all …]
|
| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_tlb_inval.c | 32 static void xe_tlb_inval_fence_fini(struct xe_tlb_inval_fence *fence) in xe_tlb_inval_fence_fini() argument 34 if (WARN_ON_ONCE(!fence->tlb_inval)) in xe_tlb_inval_fence_fini() 37 xe_pm_runtime_put(fence->tlb_inval->xe); in xe_tlb_inval_fence_fini() 38 fence->tlb_inval = NULL; /* fini() should be called once */ in xe_tlb_inval_fence_fini() 42 xe_tlb_inval_fence_signal(struct xe_tlb_inval_fence *fence) in xe_tlb_inval_fence_signal() argument 44 bool stack = test_bit(FENCE_STACK_BIT, &fence->base.flags); in xe_tlb_inval_fence_signal() 46 lockdep_assert_held(&fence->tlb_inval->pending_lock); in xe_tlb_inval_fence_signal() 48 list_del(&fence->link); in xe_tlb_inval_fence_signal() 49 trace_xe_tlb_inval_fence_signal(fence->tlb_inval->xe, fence); in xe_tlb_inval_fence_signal() 50 xe_tlb_inval_fence_fini(fence); in xe_tlb_inval_fence_signal() [all …]
|
| H A D | xe_range_fence.c | 6 #include <linux/dma-fence.h> 21 xe_range_fence_signal_notify(struct dma_fence *fence, struct dma_fence_cb *cb) in xe_range_fence_signal_notify() argument 36 dma_fence_put(rfence->fence); in __xe_range_fence_tree_cleanup() 44 * xe_range_fence_insert() - range fence insert 45 * @tree: range fence tree to insert intoi 46 * @rfence: range fence 47 * @ops: range fence ops 48 * @start: start address of range fence 49 * @last: last address of range fence 50 * @fence: dma fence which signals range fence can be removed + freed [all …]
|
| H A D | xe_sync.c | 8 #include <linux/dma-fence-array.h> 99 static void kick_ufence(struct xe_user_fence *ufence, struct dma_fence *fence) in kick_ufence() argument 103 dma_fence_put(fence); in kick_ufence() 106 static void user_fence_cb(struct dma_fence *fence, struct dma_fence_cb *cb) in user_fence_cb() argument 110 kick_ufence(ufence, fence); in user_fence_cb() 148 sync->fence = drm_syncobj_fence_get(sync->syncobj); in xe_sync_entry_parse() 149 if (XE_IOCTL_DBG(xe, !sync->fence)) { in xe_sync_entry_parse() 177 sync->fence = drm_syncobj_fence_get(sync->syncobj); in xe_sync_entry_parse() 178 if (XE_IOCTL_DBG(xe, !sync->fence)) { in xe_sync_entry_parse() 183 err = dma_fence_chain_find_seqno(&sync->fence, in xe_sync_entry_parse() [all …]
|
| /linux/drivers/gpu/drm/ |
| H A D | drm_syncobj.c | 3 * Parts ported from amdgpu (fence wait code). 41 * - Import and export a syncobj's underlying fence to/from a sync file 42 * - Reset a syncobj (set its fence to NULL) 43 * - Signal a syncobj (set a trivially signaled fence) 44 * - Wait for a syncobj's fence to appear and be signaled 57 * to an already signaled fence depending on whether the 63 * the syncobj, the syncobj's fence is replaced with a fence which will be 67 * struct &dma_fence_chain pointing to the DRM driver's fence and also 68 * pointing to the previous fence that was in the syncobj. The new struct 69 * &dma_fence_chain fence replace the syncobj's fence and will be signaled by [all …]
|
| /linux/drivers/gpu/drm/i915/selftests/ |
| H A D | lib_sw_fence.c | 27 /* Small library of different fence types useful for writing tests */ 30 nop_fence_notify(struct i915_sw_fence *fence, enum i915_sw_fence_notify state) in nop_fence_notify() argument 35 void __onstack_fence_init(struct i915_sw_fence *fence, in __onstack_fence_init() argument 39 debug_fence_init_onstack(fence); in __onstack_fence_init() 41 __init_waitqueue_head(&fence->wait, name, key); in __onstack_fence_init() 42 atomic_set(&fence->pending, 1); in __onstack_fence_init() 43 fence->error = 0; in __onstack_fence_init() 44 fence->fn = nop_fence_notify; in __onstack_fence_init() 47 void onstack_fence_fini(struct i915_sw_fence *fence) in onstack_fence_fini() argument 49 if (!fence->fn) in onstack_fence_fini() [all …]
|
| H A D | i915_sw_fence.c | 32 fence_notify(struct i915_sw_fence *fence, enum i915_sw_fence_notify state) in fence_notify() argument 39 /* Leave the fence for the caller to free it after testing */ in fence_notify() 48 struct i915_sw_fence *fence; in alloc_fence() local 50 fence = kmalloc_obj(*fence); in alloc_fence() 51 if (!fence) in alloc_fence() 54 i915_sw_fence_init(fence, fence_notify); in alloc_fence() 55 return fence; in alloc_fence() 58 static void free_fence(struct i915_sw_fence *fence) in free_fence() argument 60 i915_sw_fence_fini(fence); in free_fence() 61 kfree(fence); in free_fence() [all …]
|