| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_exec.c | 116 struct drm_xe_sync __user *syncs_user = u64_to_user_ptr(args->syncs); in xe_exec_ioctl() 119 struct xe_sync_entry *syncs = NULL; in xe_exec_ioctl() local 165 syncs = kzalloc_objs(*syncs, args->num_syncs); in xe_exec_ioctl() 166 if (!syncs) { in xe_exec_ioctl() 175 err = xe_sync_entry_parse(xe, xef, &syncs[num_syncs], in xe_exec_ioctl() 183 if (xe_sync_is_ufence(&syncs[num_syncs])) in xe_exec_ioctl() 186 if (!num_in_sync && xe_sync_needs_wait(&syncs[num_syncs])) in xe_exec_ioctl() 209 syncs, num_in_sync ? in xe_exec_ioctl() 243 fence = xe_sync_in_fence_get(syncs, num_syncs, q, vm); in xe_exec_ioctl() 250 xe_sync_entry_signal(&syncs[i], fence); in xe_exec_ioctl() [all …]
|
| H A D | xe_oa.c | 100 struct xe_sync_entry *syncs; member 649 err = xe_sync_entry_add_deps(&stream->syncs[i], job); in xe_oa_submit_bb() 1024 if (stream->syncs[i].flags & DRM_XE_SYNC_FLAG_SIGNAL) in xe_oa_emit_oa_config() 1026 xe_sync_entry_signal(&stream->syncs[i], &ofence->base); in xe_oa_emit_oa_config() 1050 xe_sync_entry_cleanup(&stream->syncs[i]); in xe_oa_emit_oa_config() 1051 kfree(stream->syncs); in xe_oa_emit_oa_config() 1411 param->syncs = kzalloc_objs(*param->syncs, param->num_syncs); in xe_oa_parse_syncs() 1412 if (!param->syncs) { in xe_oa_parse_syncs() 1419 ret = xe_sync_entry_parse(oa->xe, param->xef, ¶m->syncs[num_syncs], in xe_oa_parse_syncs() 1426 if (xe_sync_is_ufence(¶m->syncs[num_syncs])) in xe_oa_parse_syncs() [all …]
|
| H A D | xe_vm.c | 685 struct xe_sync_entry *syncs, u32 num_syncs); 1906 find_ufence_get(struct xe_sync_entry *syncs, u32 num_syncs) in find_ufence_get() argument 1911 struct xe_sync_entry *e = &syncs[i]; in find_ufence_get() 3281 ufence = find_ufence_get(vops->syncs, vops->num_syncs); in vm_bind_ioctl_ops_fini() 3296 xe_sync_entry_signal(vops->syncs + i, fence); in vm_bind_ioctl_ops_fini() 3495 struct xe_sync_entry *syncs, in vm_bind_ioctl_signal_fences() argument 3502 fence = xe_sync_in_fence_get(syncs, num_syncs, in vm_bind_ioctl_signal_fences() 3508 xe_sync_entry_signal(&syncs[i], fence); in vm_bind_ioctl_signal_fences() 3518 struct xe_sync_entry *syncs, u32 num_syncs) in xe_vma_ops_init() argument 3524 vops->syncs = syncs; in xe_vma_ops_init() [all …]
|
| H A D | xe_vm_types.h | 473 struct xe_sync_entry *syncs; member
|
| H A D | xe_pt.c | 1202 static bool no_in_syncs(struct xe_sync_entry *syncs, u32 num_syncs) in no_in_syncs() argument 1207 struct dma_fence *fence = syncs[i].fence; in no_in_syncs() 1296 if (!job && !no_in_syncs(vops->syncs, vops->num_syncs)) in xe_pt_vm_dependencies() 1342 err = xe_sync_entry_add_deps(&vops->syncs[i], job); in xe_pt_vm_dependencies()
|
| /linux/drivers/gpu/drm/v3d/ |
| H A D | v3d_submit.c | 470 u32 __user *offsets, *syncs; in v3d_get_cpu_timestamp_query_params() local 493 syncs = u64_to_user_ptr(timestamp.syncs); in v3d_get_cpu_timestamp_query_params() 505 if (get_user(sync, syncs++)) { in v3d_get_cpu_timestamp_query_params() 531 u32 __user *syncs; in v3d_get_cpu_reset_timestamp_params() local 550 syncs = u64_to_user_ptr(reset.syncs); in v3d_get_cpu_reset_timestamp_params() 557 if (get_user(sync, syncs++)) { in v3d_get_cpu_reset_timestamp_params() 584 u32 __user *offsets, *syncs; in v3d_get_cpu_copy_query_results_params() local 607 syncs = u64_to_user_ptr(copy.syncs); in v3d_get_cpu_copy_query_results_params() 619 if (get_user(sync, syncs++)) { in v3d_get_cpu_copy_query_results_params() 650 u32 __user *syncs, in v3d_copy_query_info() argument [all …]
|
| /linux/drivers/gpu/drm/panthor/ |
| H A D | panthor_drv.c | 178 PANTHOR_UOBJ_DECL(struct drm_panthor_queue_submit, syncs), \ 180 PANTHOR_UOBJ_DECL(struct drm_panthor_vm_bind_op, syncs), \ 456 const struct drm_panthor_obj_array *syncs) in panthor_submit_ctx_add_job() argument 462 ret = PANTHOR_UOBJ_GET_ARRAY(ctx->jobs[idx].syncops, syncs); in panthor_submit_ctx_add_job() 466 ctx->jobs[idx].syncop_count = syncs->count; in panthor_submit_ctx_add_job() 1016 ret = panthor_submit_ctx_add_job(&ctx, i, job, &qsubmit->syncs); in panthor_ioctl_group_submit() 1243 ret = panthor_submit_ctx_add_job(&ctx, i, job, &op->syncs); in panthor_ioctl_vm_bind_async()
|
| H A D | panthor_mmu.c | 2569 if (!op->syncs.count) in panthor_vm_bind_prepare_op_ctx() 2708 if (op->syncs.count) in panthor_vm_bind_exec_sync_op()
|
| /linux/Documentation/gpu/ |
| H A D | drm-vm-bind-async.rst | 258 synchronous VM_BIND, the num_syncs and syncs fields must be zero. Here 298 /** @num_syncs: amount of syncs to wait for or to signal on completion. */ 304 /** @syncs: pointer to struct drm_xe_sync array */ 305 __u64 syncs;
|
| /linux/include/uapi/drm/ |
| H A D | xe_drm.h | 1246 __u64 syncs; member 1508 __u64 syncs; member
|
| /linux/drivers/hid/ |
| H A D | hid-debug.c | 3151 static const char *syncs[SYN_CNT] = { variable 3605 [EV_SYN] = syncs, [EV_KEY] = keys,
|
| /linux/Documentation/userspace-api/media/v4l/ |
| H A D | vidioc-g-dv-timings.rst | 240 There are no horizontal syncs/porches at all in this format.
|
| /linux/kernel/sched/ |
| H A D | ext.c | 87 unsigned long syncs[]; member 4809 new_ksyncs = kvzalloc_node(struct_size(new_ksyncs, syncs, nr_cpu_ids), in alloc_kick_syncs() 5594 ksyncs = rcu_dereference_bh(ksyncs_pcpu)->syncs; in kick_cpus_irq_workfn()
|