| /linux/drivers/gpu/drm/xe/ |
| H A D | xe_exec.c | 116 struct drm_xe_sync __user *syncs_user = u64_to_user_ptr(args->syncs); in xe_exec_ioctl() 119 struct xe_sync_entry *syncs = NULL; in xe_exec_ioctl() local 165 syncs = kzalloc_objs(*syncs, args->num_syncs); in xe_exec_ioctl() 166 if (!syncs) { in xe_exec_ioctl() 175 err = xe_sync_entry_parse(xe, xef, &syncs[num_syncs], in xe_exec_ioctl() 183 if (xe_sync_is_ufence(&syncs[num_syncs])) in xe_exec_ioctl() 186 if (!num_in_sync && xe_sync_needs_wait(&syncs[num_syncs])) in xe_exec_ioctl() 209 syncs, num_in_sync ? in xe_exec_ioctl() 243 fence = xe_sync_in_fence_get(syncs, num_syncs, q, vm); in xe_exec_ioctl() 250 xe_sync_entry_signal(&syncs[i], fence); in xe_exec_ioctl() [all …]
|
| H A D | xe_oa.c | 100 struct xe_sync_entry *syncs; member 648 err = xe_sync_entry_add_deps(&stream->syncs[i], job); in xe_oa_submit_bb() 1023 if (stream->syncs[i].flags & DRM_XE_SYNC_FLAG_SIGNAL) in xe_oa_emit_oa_config() 1025 xe_sync_entry_signal(&stream->syncs[i], &ofence->base); in xe_oa_emit_oa_config() 1049 xe_sync_entry_cleanup(&stream->syncs[i]); in xe_oa_emit_oa_config() 1050 kfree(stream->syncs); in xe_oa_emit_oa_config() 1410 param->syncs = kzalloc_objs(*param->syncs, param->num_syncs); in xe_oa_parse_syncs() 1411 if (!param->syncs) { in xe_oa_parse_syncs() 1418 ret = xe_sync_entry_parse(oa->xe, param->xef, ¶m->syncs[num_syncs], in xe_oa_parse_syncs() 1425 if (xe_sync_is_ufence(¶m->syncs[num_syncs])) in xe_oa_parse_syncs() [all …]
|
| H A D | xe_vm.c | 685 struct xe_sync_entry *syncs, u32 num_syncs); 1906 find_ufence_get(struct xe_sync_entry *syncs, u32 num_syncs) in find_ufence_get() argument 1911 struct xe_sync_entry *e = &syncs[i]; in find_ufence_get() 3295 ufence = find_ufence_get(vops->syncs, vops->num_syncs); in vm_bind_ioctl_ops_fini() 3310 xe_sync_entry_signal(vops->syncs + i, fence); in vm_bind_ioctl_ops_fini() 3509 struct xe_sync_entry *syncs, in vm_bind_ioctl_signal_fences() argument 3516 fence = xe_sync_in_fence_get(syncs, num_syncs, in vm_bind_ioctl_signal_fences() 3522 xe_sync_entry_signal(&syncs[i], fence); in vm_bind_ioctl_signal_fences() 3532 struct xe_sync_entry *syncs, u32 num_syncs) in xe_vma_ops_init() argument 3538 vops->syncs = syncs; in xe_vma_ops_init() [all …]
|
| H A D | xe_vm_types.h | 477 struct xe_sync_entry *syncs; member
|
| H A D | xe_pt.c | 1202 static bool no_in_syncs(struct xe_sync_entry *syncs, u32 num_syncs) in no_in_syncs() argument 1207 struct dma_fence *fence = syncs[i].fence; in no_in_syncs() 1296 if (!job && !no_in_syncs(vops->syncs, vops->num_syncs)) in xe_pt_vm_dependencies() 1342 err = xe_sync_entry_add_deps(&vops->syncs[i], job); in xe_pt_vm_dependencies()
|
| /linux/drivers/gpu/drm/v3d/ |
| H A D | v3d_submit.c | 470 u32 __user *offsets, *syncs; in v3d_get_cpu_timestamp_query_params() local 493 syncs = u64_to_user_ptr(timestamp.syncs); in v3d_get_cpu_timestamp_query_params() 505 if (get_user(sync, syncs++)) { in v3d_get_cpu_timestamp_query_params() 531 u32 __user *syncs; in v3d_get_cpu_reset_timestamp_params() local 550 syncs = u64_to_user_ptr(reset.syncs); in v3d_get_cpu_reset_timestamp_params() 557 if (get_user(sync, syncs++)) { in v3d_get_cpu_reset_timestamp_params() 584 u32 __user *offsets, *syncs; in v3d_get_cpu_copy_query_results_params() local 607 syncs = u64_to_user_ptr(copy.syncs); in v3d_get_cpu_copy_query_results_params() 619 if (get_user(sync, syncs++)) { in v3d_get_cpu_copy_query_results_params() 650 u32 __user *syncs, in v3d_copy_query_info() argument [all …]
|
| /linux/include/uapi/drm/ |
| H A D | v3d_drm.h | 435 __u64 syncs; member 458 __u64 syncs; member 504 __u64 syncs; member 521 __u64 syncs; member 573 __u64 syncs; member
|
| H A D | panthor_drm.h | 562 * and drm_panthor_vm_bind_op::syncs contains at least one element. 599 * @syncs: Array of struct drm_panthor_sync_op synchronization 607 struct drm_panthor_obj_array syncs; 924 /** @syncs: Array of struct drm_panthor_sync_op sync operations. */ 925 struct drm_panthor_obj_array syncs; 1158 * syncs that must go through the DRM_PANTHOR_BO_SYNC ioctl (userland cache 564 struct drm_panthor_obj_array syncs; global() member 873 struct drm_panthor_obj_array syncs; global() member
|
| H A D | xe_drm.h | 1246 __u64 syncs; member 1508 __u64 syncs; member
|
| H A D | asahi_drm.h | 701 __u64 syncs; member
|
| /linux/drivers/gpu/drm/panthor/ |
| H A D | panthor_drv.c | 178 PANTHOR_UOBJ_DECL(struct drm_panthor_queue_submit, syncs), \ 180 PANTHOR_UOBJ_DECL(struct drm_panthor_vm_bind_op, syncs), \ 456 const struct drm_panthor_obj_array *syncs) in panthor_submit_ctx_add_job() argument 462 ret = PANTHOR_UOBJ_GET_ARRAY(ctx->jobs[idx].syncops, syncs); in panthor_submit_ctx_add_job() 466 ctx->jobs[idx].syncop_count = syncs->count; in panthor_submit_ctx_add_job() 1016 ret = panthor_submit_ctx_add_job(&ctx, i, job, &qsubmit->syncs); in panthor_ioctl_group_submit() 1243 ret = panthor_submit_ctx_add_job(&ctx, i, job, &op->syncs); in panthor_ioctl_vm_bind_async()
|
| H A D | panthor_mmu.c | 2569 if (!op->syncs.count) in panthor_vm_bind_prepare_op_ctx() 2708 if (op->syncs.count) in panthor_vm_bind_exec_sync_op()
|
| /linux/Documentation/gpu/ |
| H A D | drm-vm-bind-async.rst | 258 synchronous VM_BIND, the num_syncs and syncs fields must be zero. Here 298 /** @num_syncs: amount of syncs to wait for or to signal on completion. */ 304 /** @syncs: pointer to struct drm_xe_sync array */ 305 __u64 syncs;
|
| /linux/drivers/hid/ |
| H A D | hid-debug.c | 3151 static const char *syncs[SYN_CNT] = { variable 3605 [EV_SYN] = syncs, [EV_KEY] = keys,
|
| /linux/Documentation/userspace-api/media/v4l/ |
| H A D | vidioc-g-dv-timings.rst | 240 There are no horizontal syncs/porches at all in this format.
|
| /linux/kernel/sched/ |
| H A D | ext.c | 87 unsigned long syncs[]; member 2472 unsigned long *ksyncs = rcu_dereference_sched(ks)->syncs; in kick_sync_wait_bal_cb() 4904 new_ksyncs = kvzalloc_node(struct_size(new_ksyncs, syncs, nr_cpu_ids), in alloc_kick_syncs() 5736 ksyncs = rcu_dereference_bh(ksyncs_pcpu)->syncs; in kick_cpus_irq_workfn()
|