Home
last modified time | relevance | path

Searched refs:num_fences (Results 1 – 25 of 28) sorted by relevance

12

/linux/drivers/dma-buf/
H A Ddma-fence-array.c76 for (i = 0; i < array->num_fences; ++i) { in dma_fence_array_enable_signaling()
129 for (i = 0; i < array->num_fences; ++i) { in dma_fence_array_signaled()
145 for (i = 0; i < array->num_fences; ++i) in dma_fence_array_release()
158 for (i = 0; i < array->num_fences; ++i) in dma_fence_array_set_deadline()
178 struct dma_fence_array *dma_fence_array_alloc(int num_fences) in dma_fence_array_alloc() argument
182 return kzalloc_flex(*array, callbacks, num_fences); in dma_fence_array_alloc()
199 int num_fences, struct dma_fence **fences, in dma_fence_array_init() argument
203 WARN_ON(!num_fences || !fences); in dma_fence_array_init()
205 array->num_fences = num_fences; in dma_fence_array_init()
212 atomic_set(&array->num_pending, signal_on_any ? 1 : num_fences); in dma_fence_array_init()
[all …]
H A Ddma-resv.c65 u32 num_fences, max_fences; member
125 for (i = 0; i < list->num_fences; ++i) { in dma_resv_list_free()
182 int dma_resv_reserve_fences(struct dma_resv *obj, unsigned int num_fences) in dma_resv_reserve_fences() argument
193 if (WARN_ON(!num_fences)) in dma_resv_reserve_fences()
198 if ((old->num_fences + num_fences) <= old->max_fences) in dma_resv_reserve_fences()
200 max = max(old->num_fences + num_fences, old->max_fences * 2); in dma_resv_reserve_fences()
202 max = max(4ul, roundup_pow_of_two(num_fences)); in dma_resv_reserve_fences()
215 for (i = 0, j = 0, k = max; i < (old ? old->num_fences : 0); ++i) { in dma_resv_reserve_fences()
225 new->num_fences = j; in dma_resv_reserve_fences()
271 fences->max_fences = fences->num_fences; in dma_resv_reset_max_fences()
[all …]
H A Dsync_file.c301 unsigned int num_fences; in sync_file_ioctl_fence_info() local
312 num_fences = 0; in sync_file_ioctl_fence_info()
314 ++num_fences; in sync_file_ioctl_fence_info()
322 if (!info.num_fences) { in sync_file_ioctl_fence_info()
329 if (info.num_fences < num_fences) in sync_file_ioctl_fence_info()
332 size = num_fences * sizeof(*fence_info); in sync_file_ioctl_fence_info()
337 num_fences = 0; in sync_file_ioctl_fence_info()
341 status = sync_fill_fence_info(fence, &fence_info[num_fences++]); in sync_file_ioctl_fence_info()
353 info.num_fences = num_fences; in sync_file_ioctl_fence_info()
H A Ddma-fence-unwrap.c96 int dma_fence_dedup_array(struct dma_fence **fences, int num_fences) in dma_fence_dedup_array() argument
100 sort(fences, num_fences, sizeof(*fences), fence_cmp, NULL); in dma_fence_dedup_array()
106 for (i = 1; i < num_fences; i++) { in dma_fence_dedup_array()
118 struct dma_fence *__dma_fence_unwrap_merge(unsigned int num_fences, in __dma_fence_unwrap_merge() argument
129 for (i = 0; i < num_fences; ++i) { in __dma_fence_unwrap_merge()
163 for (i = 0; i < num_fences; ++i) { in __dma_fence_unwrap_merge()
H A Dst-dma-fence-unwrap.c50 static struct dma_fence *mock_array(unsigned int num_fences, ...) in mock_array() argument
57 fences = kzalloc_objs(*fences, num_fences); in mock_array()
61 va_start(valist, num_fences); in mock_array()
62 for (i = 0; i < num_fences; ++i) in mock_array()
66 array = dma_fence_array_create(num_fences, fences, in mock_array()
77 va_start(valist, num_fences); in mock_array()
78 for (i = 0; i < num_fences; ++i) in mock_array()
H A Ddma-buf.c499 unsigned int num_fences; in dma_buf_import_sync_file() local
518 num_fences = 0; in dma_buf_import_sync_file()
520 ++num_fences; in dma_buf_import_sync_file()
522 if (num_fences > 0) { in dma_buf_import_sync_file()
525 ret = dma_resv_reserve_fences(dmabuf->resv, num_fences); in dma_buf_import_sync_file()
/linux/tools/testing/selftests/sync/
H A Dsync.c88 int err, num_fences; in sync_file_info() local
100 num_fences = info->num_fences; in sync_file_info()
102 if (num_fences) { in sync_file_info()
104 info->num_fences = num_fences; in sync_file_info()
106 fence_info = calloc(num_fences, sizeof(*fence_info)); in sync_file_info()
139 count = info->num_fences; in sync_fence_size()
156 for (i = 0 ; i < info->num_fences ; i++) { in sync_fence_count_with_status()
/linux/drivers/gpu/drm/amd/amdgpu/
H A Damdgpu_userq_fence.c660 u16 num_points, num_fences = 0; in amdgpu_userq_wait_ioctl() local
752 if (!wait_info->num_fences) { in amdgpu_userq_wait_ioctl()
767 num_fences++; in amdgpu_userq_wait_ioctl()
784 num_fences++; in amdgpu_userq_wait_ioctl()
795 num_fences++; in amdgpu_userq_wait_ioctl()
804 num_fences++; in amdgpu_userq_wait_ioctl()
813 wait_info->num_fences = num_fences; in amdgpu_userq_wait_ioctl()
816 fence_info = kmalloc_array(wait_info->num_fences, sizeof(*fence_info), GFP_KERNEL); in amdgpu_userq_wait_ioctl()
823 fences = kmalloc_array(wait_info->num_fences, sizeof(*fences), GFP_KERNEL); in amdgpu_userq_wait_ioctl()
836 if (WARN_ON_ONCE(num_fences >= wait_info->num_fences)) { in amdgpu_userq_wait_ioctl()
[all …]
H A Damdgpu_vm.h507 unsigned int num_fences);
509 unsigned int num_fences);
H A Damdgpu_vm.c460 unsigned int num_fences) in amdgpu_vm_lock_pd() argument
464 2 + num_fences); in amdgpu_vm_lock_pd()
476 unsigned int num_fences) in amdgpu_vm_lock_done_list() argument
/linux/include/linux/
H A Ddma-fence-array.h42 unsigned num_fences; member
48 struct dma_fence_array_cb callbacks[] __counted_by(num_fences);
82 struct dma_fence_array *dma_fence_array_alloc(int num_fences);
84 int num_fences, struct dma_fence **fences,
88 struct dma_fence_array *dma_fence_array_create(int num_fences,
/linux/drivers/gpu/drm/i915/gt/
H A Dintel_ggtt_fencing.c515 for (i = 0; i < ggtt->num_fences; i++) in intel_ggtt_restore_fences()
845 int num_fences; in intel_ggtt_init_fences() local
854 num_fences = 0; in intel_ggtt_init_fences()
857 num_fences = 32; in intel_ggtt_init_fences()
861 num_fences = 16; in intel_ggtt_init_fences()
863 num_fences = 8; in intel_ggtt_init_fences()
866 num_fences = intel_uncore_read(uncore, in intel_ggtt_init_fences()
868 ggtt->fence_regs = kzalloc_objs(*ggtt->fence_regs, num_fences); in intel_ggtt_init_fences()
870 num_fences = 0; in intel_ggtt_init_fences()
873 for (i = 0; i < num_fences; i++) { in intel_ggtt_init_fences()
[all …]
H A Dintel_gt_types.h322 #define intel_gt_support_legacy_fencing(gt) ((gt)->ggtt->num_fences > 0)
/linux/drivers/gpu/drm/
H A Ddrm_gpuvm.c1190 unsigned int num_fences) in exec_prepare_obj() argument
1192 return num_fences ? drm_exec_prepare_obj(exec, obj, num_fences) : in exec_prepare_obj()
1213 unsigned int num_fences) in drm_gpuvm_prepare_vm() argument
1215 return exec_prepare_obj(exec, gpuvm->r_obj, num_fences); in drm_gpuvm_prepare_vm()
1222 unsigned int num_fences) in __drm_gpuvm_prepare_objects() argument
1229 ret = exec_prepare_obj(exec, vm_bo->obj, num_fences); in __drm_gpuvm_prepare_objects()
1243 unsigned int num_fences) in drm_gpuvm_prepare_objects_locked() argument
1253 ret = exec_prepare_obj(exec, vm_bo->obj, num_fences); in drm_gpuvm_prepare_objects_locked()
1290 unsigned int num_fences) in drm_gpuvm_prepare_objects() argument
1294 num_fences); in drm_gpuvm_prepare_objects()
[all …]
H A Ddrm_atomic_uapi.c1377 unsigned int *num_fences) in prepare_signaling() argument
1423 (*num_fences + 1), GFP_KERNEL); in prepare_signaling()
1427 memset(&f[*num_fences], 0, sizeof(*f)); in prepare_signaling()
1429 f[*num_fences].out_fence_ptr = fence_ptr; in prepare_signaling()
1436 ret = setup_out_fence(&f[(*num_fences)++], fence); in prepare_signaling()
1462 (*num_fences + 1), GFP_KERNEL); in prepare_signaling()
1466 memset(&f[*num_fences], 0, sizeof(*f)); in prepare_signaling()
1468 f[*num_fences].out_fence_ptr = fence_ptr; in prepare_signaling()
1476 ret = setup_out_fence(&f[(*num_fences)++], fence); in prepare_signaling()
1500 unsigned int num_fences, in complete_signaling() argument
[all …]
/linux/include/drm/
H A Ddrm_exec.h144 unsigned int num_fences);
148 unsigned int num_fences);
/linux/drivers/gpu/drm/i915/gem/
H A Di915_gem_execbuffer.c314 unsigned long num_fences; member
2803 SIZE_MAX / sizeof(*f)) - eb->num_fences) in add_timeline_fence_array()
2815 (eb->num_fences + nfences) * sizeof(*f), in add_timeline_fence_array()
2821 f += eb->num_fences; in add_timeline_fence_array()
2913 eb->num_fences++; in add_timeline_fence_array()
2923 unsigned long num_fences = args->num_cliprects; in add_fence_array() local
2929 if (!num_fences) in add_fence_array()
2934 if (num_fences > min_t(unsigned long, in add_fence_array()
2936 SIZE_MAX / sizeof(*f) - eb->num_fences)) in add_fence_array()
2940 if (!access_ok(user, num_fences * sizeof(*user))) in add_fence_array()
[all …]
/linux/include/uapi/linux/
H A Dsync_file.h73 __u32 num_fences; member
/linux/drivers/gpu/drm/i915/selftests/
H A Di915_request.c286 atomic_long_t num_waits, num_fences; member
317 unsigned int num_waits = 0, num_fences = 0; in __igt_breadcrumbs_smoketest() local
440 num_fences += count; in __igt_breadcrumbs_smoketest()
446 atomic_long_add(num_fences, &t->num_fences); in __igt_breadcrumbs_smoketest()
528 atomic_long_read(&t.num_fences), in mock_breadcrumbs_smoketest()
1731 unsigned long num_waits, num_fences; in live_breadcrumbs_smoketest() local
1830 num_fences = 0; in live_breadcrumbs_smoketest()
1849 num_fences += atomic_long_read(&smoke[idx].num_fences); in live_breadcrumbs_smoketest()
1853 num_waits, num_fences, idx, ncpus); in live_breadcrumbs_smoketest()
/linux/drivers/gpu/drm/xe/
H A Dxe_vm.h269 unsigned int num_fences);
/linux/drivers/gpu/drm/i915/gem/selftests/
H A Di915_gem_coherency.c264 return gt->ggtt->num_fences; in needs_fence_registers()
/linux/drivers/gpu/drm/nouveau/
H A Dnouveau_sched.c280 .num_fences = 1, in nouveau_job_submit()
H A Dnouveau_uvmm.c1221 unsigned int num_fences) in bind_lock_validate() argument
1239 ret = drm_exec_prepare_obj(exec, obj, num_fences); in bind_lock_validate()
1423 ret = bind_lock_validate(job, exec, vme->num_fences); in nouveau_uvmm_bind_job_submit()
/linux/include/uapi/drm/
H A Damdgpu_drm.h561 __u16 num_fences; member
/linux/drivers/gpu/drm/i915/
H A Di915_gpu_error.c1270 for (i = 0; i < ggtt->num_fences; i++) in gt_record_fences()
1275 for (i = 0; i < ggtt->num_fences; i++) in gt_record_fences()
1280 for (i = 0; i < ggtt->num_fences; i++) in gt_record_fences()

12