Lines Matching defs:vm

45 static struct drm_gem_object *xe_vm_obj(struct xe_vm *vm)  in xe_vm_obj()
72 struct xe_vm *vm = xe_vma_vm(vma); in xe_vma_userptr_pin_pages() local
81 static bool preempt_fences_waiting(struct xe_vm *vm) in preempt_fences_waiting()
107 static int alloc_preempt_fences(struct xe_vm *vm, struct list_head *list, in alloc_preempt_fences()
128 static int wait_for_existing_preempt_fences(struct xe_vm *vm) in wait_for_existing_preempt_fences()
150 static bool xe_vm_is_idle(struct xe_vm *vm) in xe_vm_is_idle()
163 static void arm_preempt_fences(struct xe_vm *vm, struct list_head *list) in arm_preempt_fences()
182 static int add_preempt_fences(struct xe_vm *vm, struct xe_bo *bo) in add_preempt_fences()
206 static void resume_and_reinstall_preempt_fences(struct xe_vm *vm, in resume_and_reinstall_preempt_fences()
222 int xe_vm_add_compute_exec_queue(struct xe_vm *vm, struct xe_exec_queue *q) in xe_vm_add_compute_exec_queue()
284 void xe_vm_remove_compute_exec_queue(struct xe_vm *vm, struct xe_exec_queue *q) in xe_vm_remove_compute_exec_queue()
313 int __xe_vm_userptr_needs_repin(struct xe_vm *vm) in __xe_vm_userptr_needs_repin()
331 void xe_vm_kill(struct xe_vm *vm, bool unlocked) in xe_vm_kill()
387 struct xe_vm *vm = gpuvm_to_vm(vm_bo->vm); in xe_gpuvm_validate() local
419 int xe_vm_validate_rebind(struct xe_vm *vm, struct drm_exec *exec, in xe_vm_validate_rebind()
445 static int xe_preempt_work_begin(struct drm_exec *exec, struct xe_vm *vm, in xe_preempt_work_begin()
484 struct xe_vm *vm = container_of(w, struct xe_vm, preempt.rebind_work); in preempt_rebind_work_func() local
586 static void __vma_userptr_invalidate(struct xe_vm *vm, struct xe_userptr_vma *uvma) in __vma_userptr_invalidate()
637 struct xe_vm *vm = xe_vma_vm(vma); in vma_userptr_invalidate() local
672 struct xe_vm *vm = xe_vma_vm(&uvma->vma); in xe_vma_userptr_force_invalidate() local
691 int xe_vm_userptr_pin(struct xe_vm *vm) in xe_vm_userptr_pin()
775 int xe_vm_userptr_check_repin(struct xe_vm *vm) in xe_vm_userptr_check_repin()
856 int xe_vm_rebind(struct xe_vm *vm, bool rebind_worker) in xe_vm_rebind()
911 struct dma_fence *xe_vma_rebind(struct xe_vm *vm, struct xe_vma *vma, u8 tile_mask) in xe_vma_rebind()
997 struct dma_fence *xe_vm_range_rebind(struct xe_vm *vm, in xe_vm_range_rebind()
1080 struct dma_fence *xe_vm_range_unbind(struct xe_vm *vm, in xe_vm_range_unbind()
1139 static struct xe_vma *xe_vma_create(struct xe_vm *vm, in xe_vma_create()
1247 struct xe_vm *vm = xe_vma_vm(vma); in xe_vma_destroy_late() local
1297 struct xe_vm *vm = xe_vma_vm(vma); in xe_vma_destroy() local
1341 struct xe_vm *vm = xe_vma_vm(vma); in xe_vm_lock_vma() local
1373 xe_vm_find_overlapping_vma(struct xe_vm *vm, u64 start, u64 range) in xe_vm_find_overlapping_vma()
1389 static int xe_vm_insert_vma(struct xe_vm *vm, struct xe_vma *vma) in xe_vm_insert_vma()
1404 static void xe_vm_remove_vma(struct xe_vm *vm, struct xe_vma *vma) in xe_vm_remove_vma()
1578 struct xe_vm *vm) in xe_vm_create_scratch()
1595 static void xe_vm_free_scratch(struct xe_vm *vm) in xe_vm_free_scratch()
1618 struct xe_vm *vm; in xe_vm_create() local
1790 static void xe_vm_close(struct xe_vm *vm) in xe_vm_close()
1832 void xe_vm_close_and_put(struct xe_vm *vm) in xe_vm_close_and_put()
1941 struct xe_vm *vm = in vm_destroy_work_func() local
1973 struct xe_vm *vm = container_of(gpuvm, struct xe_vm, gpuvm); in xe_vm_free() local
1981 struct xe_vm *vm; in xe_vm_lookup() local
1992 u64 xe_vm_pdp4_descriptor(struct xe_vm *vm, struct xe_tile *tile) in xe_vm_pdp4_descriptor()
1999 to_wait_exec_queue(struct xe_vm *vm, struct xe_exec_queue *q) in to_wait_exec_queue()
2030 struct xe_vm *vm; in xe_vm_create_ioctl() local
2115 struct xe_vm *vm; in xe_vm_destroy_ioctl() local
2144 static void prep_vma_destroy(struct xe_vm *vm, struct xe_vma *vma, in prep_vma_destroy()
2209 vm_bind_ioctl_ops_create(struct xe_vm *vm, struct xe_bo *bo, in vm_bind_ioctl_ops_create()
2287 static struct xe_vma *new_vma(struct xe_vm *vm, struct drm_gpuva_op_map *op, in new_vma()
2371 static int xe_vma_op_commit(struct xe_vm *vm, struct xe_vma_op *op) in xe_vma_op_commit()
2434 static int vm_bind_ioctl_ops_parse(struct xe_vm *vm, struct drm_gpuva_ops *ops, in vm_bind_ioctl_ops_parse()
2610 static void xe_vma_op_unwind(struct xe_vm *vm, struct xe_vma_op *op, in xe_vma_op_unwind()
2665 static void vm_bind_ioctl_ops_unwind(struct xe_vm *vm, in vm_bind_ioctl_ops_unwind()
2693 struct xe_vm *vm = xe_vma_vm(vma); in vma_lock_and_validate() local
2722 static int op_lock_and_prep(struct drm_exec *exec, struct xe_vm *vm, in op_lock_and_prep()
2778 struct xe_vm *vm, in vm_bind_ioctl_ops_lock_and_prep()
2837 static int vm_ops_setup_tile_args(struct xe_vm *vm, struct xe_vma_ops *vops) in vm_ops_setup_tile_args()
2863 static struct dma_fence *ops_execute(struct xe_vm *vm, in ops_execute()
2957 static void op_add_ufence(struct xe_vm *vm, struct xe_vma_op *op, in op_add_ufence()
2980 static void vm_bind_ioctl_ops_fini(struct xe_vm *vm, struct xe_vma_ops *vops, in vm_bind_ioctl_ops_fini()
3008 static struct dma_fence *vm_bind_ioctl_ops_execute(struct xe_vm *vm, in vm_bind_ioctl_ops_execute()
3060 static int vm_bind_ioctl_check_args(struct xe_device *xe, struct xe_vm *vm, in vm_bind_ioctl_check_args()
3185 static int vm_bind_ioctl_signal_fences(struct xe_vm *vm, in vm_bind_ioctl_signal_fences()
3208 static void xe_vma_ops_init(struct xe_vma_ops *vops, struct xe_vm *vm, in xe_vma_ops_init()
3283 struct xe_vm *vm; in xe_vm_bind_ioctl() local
3509 struct dma_fence *xe_vm_bind_kernel_bo(struct xe_vm *vm, struct xe_bo *bo, in xe_vm_bind_kernel_bo()
3579 int xe_vm_lock(struct xe_vm *vm, bool intr) in xe_vm_lock()
3593 void xe_vm_unlock(struct xe_vm *vm) in xe_vm_unlock()
3677 int xe_vm_validate_protected(struct xe_vm *vm) in xe_vm_validate_protected()
3717 struct xe_vm_snapshot *xe_vm_snapshot_capture(struct xe_vm *vm) in xe_vm_snapshot_capture()