/linux/tools/testing/vma/ |
H A D | vma.c | 280 /* Helper function providing a dummy vm_ops->close() method.*/ 465 const struct vm_operations_struct vm_ops = { in test_merge_new() local 511 vma_a->vm_ops = &vm_ops; /* This should have no impact. */ in test_merge_new() 548 vma_d->vm_ops = &vm_ops; /* This should have no impact. */ in test_merge_new() 566 vma_d->vm_ops = NULL; /* This would otherwise degrade the merge. */ in test_merge_new() 721 const struct vm_operations_struct vm_ops = { in test_vma_merge_with_close() local 728 * vm_ops->close() hook. in test_vma_merge_with_close() 742 * has a vm_ops in test_vma_merge_with_close() 932 const struct vm_operations_struct vm_ops = { test_vma_merge_new_with_close() local 986 const struct vm_operations_struct vm_ops = { test_merge_existing() local [all...] |
H A D | vma_internal.h | 288 const struct vm_operations_struct *vm_ops; member 359 const struct vm_operations_struct *vm_ops; member 650 vma->vm_ops = &vma_dummy_vm_ops; in vma_init() 788 return !vma->vm_ops; in vma_is_anonymous() 1179 vma->vm_ops = NULL; in vma_set_anonymous()
|
/linux/arch/um/kernel/ |
H A D | tlb.c | 18 struct vm_ops { struct 53 struct vm_ops *ops) in update_pte_range() 94 struct vm_ops *ops) in update_pmd_range() 117 struct vm_ops *ops) in update_pud_range() 140 struct vm_ops *ops) in update_p4d_range() 164 struct vm_ops ops; in um_tlb_sync() 19 mm_idpvm_ops global() argument 21 mmapvm_ops global() argument 24 unmapvm_ops global() argument
|
/linux/mm/ |
H A D | vma.c | 23 const struct vm_operations_struct *vm_ops; member 509 if (vma->vm_ops && vma->vm_ops->may_split) { in __split_vma() 510 err = vma->vm_ops->may_split(vma, addr); in __split_vma() 542 if (new->vm_ops && new->vm_ops->open) in __split_vma() 543 new->vm_ops->open(new); in __split_vma() 775 return !vma->vm_ops || !vma->vm_ops->close; in can_merge_remove_vma() 1878 if (new_vma->vm_ops in copy_vma() 2001 vm_ops_needs_writenotify(const struct vm_operations_struct * vm_ops) vm_ops_needs_writenotify() argument [all...] |
H A D | vma_init.c | 44 dest->vm_ops = src->vm_ops; in vm_area_init_from()
|
H A D | secretmem.c | 132 desc->vm_ops = &secretmem_vm_ops; in secretmem_mmap_prepare() 139 return vma->vm_ops == &secretmem_vm_ops; in vma_is_secretmem()
|
H A D | memory.c | 545 vma->vm_ops ? vma->vm_ops->fault : NULL, in print_bad_pte() 606 if (vma->vm_ops && vma->vm_ops->find_special_page) in vm_normal_page() 607 return vma->vm_ops->find_special_page(vma, addr); in vm_normal_page() 2117 * Why not allow any VMA that has vm_ops->pfn_mkwrite? GUP could in vm_mixed_zeropage_allowed() 2120 * page due to vma->vm_ops->pfn_mkwrite, because what's mapped would in vm_mixed_zeropage_allowed() 2125 return vma->vm_ops && vma->vm_ops->pfn_mkwrite && in vm_mixed_zeropage_allowed() 2561 * This function should only be called from a vm_ops [all...] |
H A D | mremap.c | 1041 if (vma->vm_ops && vma->vm_ops->may_split) { in prep_move_vma() 1043 err = vma->vm_ops->may_split(vma, old_addr); in prep_move_vma() 1045 err = vma->vm_ops->may_split(vma, old_addr + old_len); in prep_move_vma() 1206 else if (vma->vm_ops && vma->vm_ops->mremap) in copy_vma_and_data() 1207 err = vma->vm_ops->mremap(new_vma); in copy_vma_and_data()
|
H A D | vma.h | 244 desc->vm_ops = NULL; in vma_to_desc() 268 vma->vm_ops = desc->vm_ops; in set_vma_from_desc()
|
H A D | internal.h | 177 vma->vm_ops = &vma_dummy_vm_ops; in mmap_file() 189 if (vma->vm_ops && vma->vm_ops->close) { in vma_close() 190 vma->vm_ops->close(vma); in vma_close() 196 vma->vm_ops = &vma_dummy_vm_ops; in vma_close()
|
/linux/drivers/gpu/drm/xen/ |
H A D | xen_drm_front_gem.c | 66 vma->vm_ops = gem_obj->funcs->vm_ops; in xen_drm_front_gem_object_mmap() 111 .vm_ops = &xen_drm_drv_vm_ops,
|
/linux/drivers/gpu/drm/mediatek/ |
H A D | mtk_gem.c | 20 static const struct vm_operations_struct vm_ops = { variable 31 .vm_ops = &vm_ops,
|
/linux/arch/hexagon/kernel/ |
H A D | Makefile | 14 obj-y += vm_entry.o vm_events.o vm_switch.o vm_ops.o vm_init_segtable.o
|
/linux/drivers/accel/amdxdna/ |
H A D | amdxdna_gem.c | 290 vma->vm_ops->close(vma); in amdxdna_insert_pages() 298 vma->vm_ops = NULL; in amdxdna_insert_pages() 311 vma->vm_ops->close(vma); in amdxdna_insert_pages() 359 vma->vm_ops = &drm_gem_shmem_vm_ops; in amdxdna_gem_dmabuf_mmap() 377 vma->vm_ops->close(vma); in amdxdna_gem_dmabuf_mmap() 457 .vm_ops = &drm_gem_shmem_vm_ops,
|
/linux/include/linux/ |
H A D | agp_backend.h | 56 const struct vm_operations_struct *vm_ops; member
|
/linux/drivers/gpu/drm/msm/ |
H A D | msm_gem_vma.c | 82 /** @node: list head in msm_vm_bind_job::vm_ops */ 119 /** @vm_ops: a list of struct msm_vm_op */ 120 struct list_head vm_ops; member 464 list_add_tail(&op->node, &arg->job->vm_ops); in vm_op_enqueue() 656 while (!list_empty(&job->vm_ops)) { in msm_vma_job_run() 658 list_first_entry(&job->vm_ops, struct msm_vm_op, node); in msm_vma_job_run() 719 while (!list_empty(&job->vm_ops)) { in msm_vma_job_free() 721 list_first_entry(&job->vm_ops, struct msm_vm_op, node); in msm_vma_job_free() 919 INIT_LIST_HEAD(&job->vm_ops); in vm_bind_job_create()
|
/linux/drivers/pci/ |
H A D | mmap.c | 47 vma->vm_ops = &pci_phys_vm_ops; in pci_mmap_resource_range()
|
/linux/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_gem.c | 116 * Reset both vm_ops and vm_private_data, so we don't end up with in vmw_gem_mmap() 117 * vm_ops pointing to our implementation if the dma-buf backend in vmw_gem_mmap() 121 vma->vm_ops = NULL; in vmw_gem_mmap() 154 .vm_ops = &vmw_vm_ops,
|
/linux/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_mman.c | 1027 vma->vm_ops = obj->ops->mmap_ops; in i915_gem_object_mmap() 1038 vma->vm_ops = &vm_ops_cpu; in i915_gem_object_mmap() 1046 vma->vm_ops = &vm_ops_cpu; in i915_gem_object_mmap() 1052 vma->vm_ops = &vm_ops_cpu; in i915_gem_object_mmap() 1058 vma->vm_ops = &vm_ops_gtt; in i915_gem_object_mmap() 1144 * When we install vm_ops for mmap we are too late for in i915_gem_fb_mmap() 1145 * the vm_ops->open() which increases the ref_count of in i915_gem_fb_mmap() 1146 * this obj and then it gets decreased by the vm_ops->close(). in i915_gem_fb_mmap()
|
/linux/arch/x86/kernel/cpu/sgx/ |
H A D | encl.h | 95 if (!result || result->vm_ops != &sgx_vm_ops) in sgx_encl_find()
|
/linux/drivers/gpu/drm/ttm/ |
H A D | ttm_bo_vm.c | 493 * Drivers may want to override the vm_ops field. Otherwise we in ttm_bo_mmap_obj() 496 if (!vma->vm_ops) in ttm_bo_mmap_obj() 497 vma->vm_ops = &ttm_bo_vm_ops; in ttm_bo_mmap_obj()
|
/linux/drivers/media/common/videobuf2/ |
H A D | videobuf2-vmalloc.c | 195 vma->vm_ops = &vb2_common_vm_ops; in vb2_vmalloc_mmap() 197 vma->vm_ops->open(vma); in vb2_vmalloc_mmap()
|
/linux/fs/nilfs2/ |
H A D | file.c | 131 desc->vm_ops = &nilfs_file_vm_ops; in nilfs_file_mmap_prepare()
|
/linux/drivers/infiniband/sw/rxe/ |
H A D | rxe_mmap.c | 105 vma->vm_ops = &rxe_vm_ops; in rxe_mmap()
|
/linux/fs/ocfs2/ |
H A D | mmap.c | 175 desc->vm_ops = &ocfs2_file_vm_ops; in ocfs2_mmap_prepare()
|