Home
last modified time | relevance | path

Searched defs:vma (Results 1 – 25 of 722) sorted by relevance

12345678910>>...29

/linux/drivers/gpu/drm/i915/
H A Di915_vma.h51 static inline bool i915_vma_is_active(const struct i915_vma *vma) in i915_vma_is_active() argument
65 i915_vma_move_to_active(struct i915_vma *vma, struct i915_request *rq, in i915_vma_move_to_active() argument
73 static inline bool i915_vma_is_ggtt(const struct i915_vma *vma) in i915_vma_is_ggtt() argument
78 static inline bool i915_vma_is_dpt(const struct i915_vma *vma) in i915_vma_is_dpt() argument
83 static inline bool i915_vma_has_ggtt_write(const struct i915_vma *vma) in i915_vma_has_ggtt_write() argument
88 i915_vma_set_ggtt_write(struct i915_vma * vma) i915_vma_set_ggtt_write() argument
94 i915_vma_unset_ggtt_write(struct i915_vma * vma) i915_vma_unset_ggtt_write() argument
102 i915_vma_is_map_and_fenceable(const struct i915_vma * vma) i915_vma_is_map_and_fenceable() argument
107 i915_vma_set_userfault(struct i915_vma * vma) i915_vma_set_userfault() argument
113 i915_vma_unset_userfault(struct i915_vma * vma) i915_vma_unset_userfault() argument
118 i915_vma_has_userfault(const struct i915_vma * vma) i915_vma_has_userfault() argument
123 i915_vma_is_closed(const struct i915_vma * vma) i915_vma_is_closed() argument
129 __i915_vma_size(const struct i915_vma * vma) __i915_vma_size() argument
144 i915_vma_size(const struct i915_vma * vma) i915_vma_size() argument
151 __i915_vma_offset(const struct i915_vma * vma) __i915_vma_offset() argument
167 i915_vma_offset(const struct i915_vma * vma) i915_vma_offset() argument
173 i915_ggtt_offset(const struct i915_vma * vma) i915_ggtt_offset() argument
183 i915_ggtt_pin_bias(struct i915_vma * vma) i915_ggtt_pin_bias() argument
188 i915_vma_get(struct i915_vma * vma) i915_vma_get() argument
194 i915_vma_tryget(struct i915_vma * vma) i915_vma_tryget() argument
202 i915_vma_put(struct i915_vma * vma) i915_vma_put() argument
208 i915_vma_compare(struct i915_vma * vma,struct i915_address_space * vm,const struct i915_gtt_view * view) i915_vma_compare() argument
276 assert_vma_held(vma) global() argument
278 i915_vma_lock(struct i915_vma * vma) i915_vma_lock() argument
283 i915_vma_unlock(struct i915_vma * vma) i915_vma_unlock() argument
298 i915_vma_pin_count(const struct i915_vma * vma) i915_vma_pin_count() argument
303 i915_vma_is_pinned(const struct i915_vma * vma) i915_vma_is_pinned() argument
308 __i915_vma_pin(struct i915_vma * vma) __i915_vma_pin() argument
314 __i915_vma_unpin(struct i915_vma * vma) __i915_vma_unpin() argument
320 i915_vma_unpin(struct i915_vma * vma) i915_vma_unpin() argument
326 i915_vma_is_bound(const struct i915_vma * vma,unsigned int where) i915_vma_is_bound() argument
338 i915_vma_get_iomap(struct i915_vma * vma) i915_vma_get_iomap() argument
386 __i915_vma_unpin_fence(struct i915_vma * vma) __i915_vma_unpin_fence() argument
401 i915_vma_unpin_fence(struct i915_vma * vma) i915_vma_unpin_fence() argument
407 i915_vma_fence_id(const struct i915_vma * vma) i915_vma_fence_id() argument
414 i915_vma_is_scanout(const struct i915_vma * vma) i915_vma_is_scanout() argument
419 i915_vma_mark_scanout(struct i915_vma * vma) i915_vma_mark_scanout() argument
424 i915_vma_clear_scanout(struct i915_vma * vma) i915_vma_clear_scanout() argument
452 i915_vma_sync(struct i915_vma * vma) i915_vma_sync() argument
468 i915_vma_get_current_resource(struct i915_vma * vma) i915_vma_get_current_resource() argument
[all...]
H A Di915_vma.c48 static inline void assert_vma_held_evict(const struct i915_vma *vma) in assert_vma_held_evict() argument
66 static void i915_vma_free(struct i915_vma *vma) in i915_vma_free() argument
75 static void vma_print_allocator(struct i915_vma *vma, const char *reason) in vma_print_allocator() argument
94 vma_print_allocator(struct i915_vma * vma,const char * reason) vma_print_allocator() argument
107 struct i915_vma *vma = active_to_vma(ref); __i915_vma_active() local
132 struct i915_vma *vma = active_to_vma(ref); __i915_vma_retire() local
151 struct i915_vma *vma; vma_create() local
293 struct i915_vma *vma = rb_entry(rb, struct i915_vma, obj_node); i915_vma_lookup() local
327 struct i915_vma *vma; i915_vma_instance() local
405 i915_vma_wait_for_bind(struct i915_vma * vma) i915_vma_wait_for_bind() argument
425 i915_vma_verify_bind_complete(struct i915_vma * vma) i915_vma_verify_bind_complete() argument
448 i915_vma_resource_init_from_vma(struct i915_vma_resource * vma_res,struct i915_vma * vma) i915_vma_resource_init_from_vma() argument
472 i915_vma_bind(struct i915_vma * vma,unsigned int pat_index,u32 flags,struct i915_vma_work * work,struct i915_vma_resource * vma_res) i915_vma_bind() argument
581 i915_vma_pin_iomap(struct i915_vma * vma) i915_vma_pin_iomap() argument
649 i915_vma_flush_writes(struct i915_vma * vma) i915_vma_flush_writes() argument
655 i915_vma_unpin_iomap(struct i915_vma * vma) i915_vma_unpin_iomap() argument
669 struct i915_vma *vma; i915_vma_unpin_and_release() local
687 i915_vma_misplaced(const struct i915_vma * vma,u64 size,u64 alignment,u64 flags) i915_vma_misplaced() argument
721 __i915_vma_set_map_and_fenceable(struct i915_vma * vma) __i915_vma_set_map_and_fenceable() argument
740 i915_gem_valid_gtt_space(struct i915_vma * vma,unsigned long color) i915_gem_valid_gtt_space() argument
788 i915_vma_insert(struct i915_vma * vma,struct i915_gem_ww_ctx * ww,u64 size,u64 alignment,u64 flags) i915_vma_insert() argument
927 i915_vma_detach(struct i915_vma * vma) i915_vma_detach() argument
940 try_qad_pin(struct i915_vma * vma,unsigned int flags) try_qad_pin() argument
1300 __i915_vma_get_pages(struct i915_vma * vma) __i915_vma_get_pages() argument
1347 i915_vma_get_pages(struct i915_vma * vma) i915_vma_get_pages() argument
1394 __vma_put_pages(struct i915_vma * vma,unsigned int count) __vma_put_pages() argument
1410 i915_vma_put_pages(struct i915_vma * vma) i915_vma_put_pages() argument
1418 vma_unbind_pages(struct i915_vma * vma) vma_unbind_pages() argument
1432 i915_vma_pin_ww(struct i915_vma * vma,struct i915_gem_ww_ctx * ww,u64 size,u64 alignment,u64 flags) i915_vma_pin_ww() argument
1610 i915_vma_pin(struct i915_vma * vma,u64 size,u64 alignment,u64 flags) i915_vma_pin() argument
1641 __i915_ggtt_pin(struct i915_vma * vma,struct i915_gem_ww_ctx * ww,u32 align,unsigned int flags) __i915_ggtt_pin() argument
1676 i915_ggtt_pin(struct i915_vma * vma,struct i915_gem_ww_ctx * ww,u32 align,unsigned int flags) i915_ggtt_pin() argument
1708 struct i915_vma *vma; i915_ggtt_clear_scanout() local
1718 __vma_close(struct i915_vma * vma,struct intel_gt * gt) __vma_close() argument
1736 i915_vma_close(struct i915_vma * vma) i915_vma_close() argument
1753 __i915_vma_remove_closed(struct i915_vma * vma) __i915_vma_remove_closed() argument
1758 i915_vma_reopen(struct i915_vma * vma) i915_vma_reopen() argument
1768 force_unbind(struct i915_vma * vma) force_unbind() argument
1778 release_references(struct i915_vma * vma,struct intel_gt * gt,bool vm_ddestroy) release_references() argument
1830 i915_vma_destroy_locked(struct i915_vma * vma) i915_vma_destroy_locked() argument
1839 i915_vma_destroy(struct i915_vma * vma) i915_vma_destroy() argument
1858 struct i915_vma *vma, *next; i915_vma_parked() local
1901 __i915_vma_iounmap(struct i915_vma * vma) __i915_vma_iounmap() argument
1915 i915_vma_revoke_mmap(struct i915_vma * vma) i915_vma_revoke_mmap() argument
1939 __i915_request_await_bind(struct i915_request * rq,struct i915_vma * vma) __i915_request_await_bind() argument
1944 __i915_vma_move_to_active(struct i915_vma * vma,struct i915_request * rq) __i915_vma_move_to_active() argument
1956 _i915_vma_move_to_active(struct i915_vma * vma,struct i915_request * rq,struct dma_fence * fence,unsigned int flags) _i915_vma_move_to_active() argument
2031 __i915_vma_evict(struct i915_vma * vma,bool async) __i915_vma_evict() argument
2111 __i915_vma_unbind(struct i915_vma * vma) __i915_vma_unbind() argument
2142 __i915_vma_unbind_async(struct i915_vma * vma) __i915_vma_unbind_async() argument
2177 i915_vma_unbind(struct i915_vma * vma) i915_vma_unbind() argument
2215 i915_vma_unbind_async(struct i915_vma * vma,bool trylock_vm) i915_vma_unbind_async() argument
2277 i915_vma_unbind_unlocked(struct i915_vma * vma) i915_vma_unbind_unlocked() argument
2288 i915_vma_make_unshrinkable(struct i915_vma * vma) i915_vma_make_unshrinkable() argument
2294 i915_vma_make_shrinkable(struct i915_vma * vma) i915_vma_make_shrinkable() argument
2299 i915_vma_make_purgeable(struct i915_vma * vma) i915_vma_make_purgeable() argument
[all...]
H A Di915_gem_evict.c67 static bool grab_vma(struct i915_vma *vma, struct i915_gem_ww_ctx *ww) in grab_vma() argument
86 static void ungrab_vma(struct i915_vma *vma) in ungrab_vma() argument
98 mark_free(struct drm_mm_scan * scan,struct i915_gem_ww_ctx * ww,struct i915_vma * vma,unsigned int flags,struct list_head * unwind) mark_free() argument
112 defer_evict(struct i915_vma * vma) defer_evict() argument
157 struct i915_vma *vma, *next; i915_gem_evict_something() local
337 struct i915_vma *vma, *next; i915_gem_evict_for_node() local
478 struct i915_vma *vma, *vn; i915_gem_evict_vm() local
[all...]
/linux/include/linux/
H A Duserfaultfd_k.h145 static inline bool is_mergeable_vm_userfaultfd_ctx(struct vm_area_struct *vma, in is_mergeable_vm_userfaultfd_ctx() argument
162 static inline bool uffd_disable_huge_pmd_share(struct vm_area_struct *vma) in uffd_disable_huge_pmd_share() argument
174 static inline bool uffd_disable_fault_around(struct vm_area_struct *vma) in uffd_disable_fault_around() argument
179 static inline bool userfaultfd_missing(struct vm_area_struct *vma) in userfaultfd_missing() argument
184 static inline bool userfaultfd_wp(struct vm_area_struct *vma) in userfaultfd_wp() argument
189 userfaultfd_minor(struct vm_area_struct * vma) userfaultfd_minor() argument
194 userfaultfd_pte_wp(struct vm_area_struct * vma,pte_t pte) userfaultfd_pte_wp() argument
200 userfaultfd_huge_pmd_wp(struct vm_area_struct * vma,pmd_t pmd) userfaultfd_huge_pmd_wp() argument
206 userfaultfd_armed(struct vm_area_struct * vma) userfaultfd_armed() argument
211 vma_can_userfault(struct vm_area_struct * vma,vm_flags_t vm_flags,bool wp_async) vma_can_userfault() argument
246 vma_has_uffd_without_event_remap(struct vm_area_struct * vma) vma_has_uffd_without_event_remap() argument
303 uffd_wp_range(struct vm_area_struct * vma,unsigned long start,unsigned long len,bool enable_wp) uffd_wp_range() argument
310 is_mergeable_vm_userfaultfd_ctx(struct vm_area_struct * vma,struct vm_userfaultfd_ctx vm_ctx) is_mergeable_vm_userfaultfd_ctx() argument
316 userfaultfd_missing(struct vm_area_struct * vma) userfaultfd_missing() argument
321 userfaultfd_wp(struct vm_area_struct * vma) userfaultfd_wp() argument
326 userfaultfd_minor(struct vm_area_struct * vma) userfaultfd_minor() argument
331 userfaultfd_pte_wp(struct vm_area_struct * vma,pte_t pte) userfaultfd_pte_wp() argument
337 userfaultfd_huge_pmd_wp(struct vm_area_struct * vma,pmd_t pmd) userfaultfd_huge_pmd_wp() argument
344 userfaultfd_armed(struct vm_area_struct * vma) userfaultfd_armed() argument
349 dup_userfaultfd(struct vm_area_struct * vma,struct list_head * l) dup_userfaultfd() argument
363 mremap_userfaultfd_prep(struct vm_area_struct * vma,struct vm_userfaultfd_ctx * ctx) mremap_userfaultfd_prep() argument
379 userfaultfd_remove(struct vm_area_struct * vma,unsigned long start,unsigned long end) userfaultfd_remove() argument
386 userfaultfd_unmap_prep(struct vm_area_struct * vma,unsigned long start,unsigned long end,struct list_head * uf) userfaultfd_unmap_prep() argument
398 uffd_disable_fault_around(struct vm_area_struct * vma) uffd_disable_fault_around() argument
403 userfaultfd_wp_unpopulated(struct vm_area_struct * vma) userfaultfd_wp_unpopulated() argument
408 userfaultfd_wp_async(struct vm_area_struct * vma) userfaultfd_wp_async() argument
413 vma_has_uffd_without_event_remap(struct vm_area_struct * vma) vma_has_uffd_without_event_remap() argument
420 userfaultfd_wp_use_markers(struct vm_area_struct * vma) userfaultfd_wp_use_markers() argument
[all...]
H A Dhuge_mm.h101 #define thp_vma_allowable_order(vma, vm_flags, tva_flags, order) \ argument
217 thp_vma_suitable_order(struct vm_area_struct * vma,unsigned long addr,int order) thp_vma_suitable_order() argument
242 thp_vma_suitable_orders(struct vm_area_struct * vma,unsigned long addr,unsigned long orders) thp_vma_suitable_orders() argument
286 thp_vma_allowable_orders(struct vm_area_struct * vma,vm_flags_t vm_flags,unsigned long tva_flags,unsigned long orders) thp_vma_allowable_orders() argument
321 vma_thp_disabled(struct vm_area_struct * vma,vm_flags_t vm_flags) vma_thp_disabled() argument
422 change_huge_pud(struct mmu_gather * tlb,struct vm_area_struct * vma,pud_t * pudp,unsigned long addr,pgprot_t newprot,unsigned long cp_flags) change_huge_pud() argument
450 pmd_trans_huge_lock(pmd_t * pmd,struct vm_area_struct * vma) pmd_trans_huge_lock() argument
458 pud_trans_huge_lock(pud_t * pud,struct vm_area_struct * vma) pud_trans_huge_lock() argument
515 thp_vma_suitable_order(struct vm_area_struct * vma,unsigned long addr,int order) thp_vma_suitable_order() argument
521 thp_vma_suitable_orders(struct vm_area_struct * vma,unsigned long addr,unsigned long orders) thp_vma_suitable_orders() argument
527 thp_vma_allowable_orders(struct vm_area_struct * vma,vm_flags_t vm_flags,unsigned long tva_flags,unsigned long orders) thp_vma_allowable_orders() argument
578 __split_huge_pmd(struct vm_area_struct * vma,pmd_t * pmd,unsigned long address,bool freeze) __split_huge_pmd() argument
580 split_huge_pmd_address(struct vm_area_struct * vma,unsigned long address,bool freeze) split_huge_pmd_address() argument
582 split_huge_pmd_locked(struct vm_area_struct * vma,unsigned long address,pmd_t * pmd,bool freeze) split_huge_pmd_locked() argument
586 unmap_huge_pmd_locked(struct vm_area_struct * vma,unsigned long addr,pmd_t * pmdp,struct folio * folio) unmap_huge_pmd_locked() argument
596 hugepage_madvise(struct vm_area_struct * vma,vm_flags_t * vm_flags,int advice) hugepage_madvise() argument
602 madvise_collapse(struct vm_area_struct * vma,unsigned long start,unsigned long end,bool * lock_dropped) madvise_collapse() argument
609 vma_adjust_trans_huge(struct vm_area_struct * vma,unsigned long start,unsigned long end,struct vm_area_struct * next) vma_adjust_trans_huge() argument
620 pmd_trans_huge_lock(pmd_t * pmd,struct vm_area_struct * vma) pmd_trans_huge_lock() argument
625 pud_trans_huge_lock(pud_t * pud,struct vm_area_struct * vma) pud_trans_huge_lock() argument
655 follow_devmap_pmd(struct vm_area_struct * vma,unsigned long addr,pmd_t * pmd,int flags,struct dev_pagemap ** pgmap) follow_devmap_pmd() argument
676 __split_huge_pud(struct vm_area_struct * vma,pud_t * pud,unsigned long address) __split_huge_pud() argument
682 change_huge_pud(struct mmu_gather * tlb,struct vm_area_struct * vma,pud_t * pudp,unsigned long addr,pgprot_t newprot,unsigned long cp_flags) change_huge_pud() argument
[all...]
H A Drmap.h84 struct vm_area_struct *vma; member
160 static inline int anon_vma_prepare(struct vm_area_struct *vma) in anon_vma_prepare() argument
168 static inline void anon_vma_merge(struct vm_area_struct *vma, in anon_vma_merge() argument
230 folio_set_large_mapcount(struct folio * folio,int mapcount,struct vm_area_struct * vma) folio_set_large_mapcount() argument
244 folio_add_return_large_mapcount(struct folio * folio,int diff,struct vm_area_struct * vma) folio_add_return_large_mapcount() argument
294 folio_sub_return_large_mapcount(struct folio * folio,int diff,struct vm_area_struct * vma) folio_sub_return_large_mapcount() argument
345 folio_set_large_mapcount(struct folio * folio,int mapcount,struct vm_area_struct * vma) folio_set_large_mapcount() argument
352 folio_add_large_mapcount(struct folio * folio,int diff,struct vm_area_struct * vma) folio_add_large_mapcount() argument
358 folio_add_return_large_mapcount(struct folio * folio,int diff,struct vm_area_struct * vma) folio_add_return_large_mapcount() argument
364 folio_sub_large_mapcount(struct folio * folio,int diff,struct vm_area_struct * vma) folio_sub_large_mapcount() argument
370 folio_sub_return_large_mapcount(struct folio * folio,int diff,struct vm_area_struct * vma) folio_sub_return_large_mapcount() argument
376 folio_inc_large_mapcount(folio,vma) global() argument
378 folio_inc_return_large_mapcount(folio,vma) global() argument
380 folio_dec_large_mapcount(folio,vma) global() argument
382 folio_dec_return_large_mapcount(folio,vma) global() argument
482 folio_add_anon_rmap_pte(folio,page,vma,address,flags) global() argument
490 folio_add_file_rmap_pte(folio,page,vma) global() argument
498 folio_remove_rmap_pte(folio,page,vma) global() argument
512 hugetlb_try_dup_anon_rmap(struct folio * folio,struct vm_area_struct * vma) hugetlb_try_dup_anon_rmap() argument
935 struct vm_area_struct *vma; global() member
1046 anon_vma_prepare(vma) global() argument
[all...]
H A Dhugetlb.h108 struct vm_area_struct *vma; member
254 hugetlb_zap_begin(struct vm_area_struct * vma,unsigned long * start,unsigned long * end) hugetlb_zap_begin() argument
261 hugetlb_zap_end(struct vm_area_struct * vma,struct zap_details * details) hugetlb_zap_end() argument
286 hugetlb_dup_vma_private(struct vm_area_struct * vma) hugetlb_dup_vma_private() argument
290 clear_vma_resv_huge_pages(struct vm_area_struct * vma) clear_vma_resv_huge_pages() argument
306 huge_pmd_unshare(struct mm_struct * mm,struct vm_area_struct * vma,unsigned long addr,pte_t * ptep) huge_pmd_unshare() argument
313 adjust_range_if_pmd_sharing_possible(struct vm_area_struct * vma,unsigned long * start,unsigned long * end) adjust_range_if_pmd_sharing_possible() argument
319 hugetlb_zap_begin(struct vm_area_struct * vma,unsigned long * start,unsigned long * end) hugetlb_zap_begin() argument
325 hugetlb_zap_end(struct vm_area_struct * vma,struct zap_details * details) hugetlb_zap_end() argument
339 move_hugetlb_page_tables(struct vm_area_struct * vma,struct vm_area_struct * new_vma,unsigned long old_addr,unsigned long new_addr,unsigned long len) move_hugetlb_page_tables() argument
362 hugetlb_vma_lock_read(struct vm_area_struct * vma) hugetlb_vma_lock_read() argument
366 hugetlb_vma_unlock_read(struct vm_area_struct * vma) hugetlb_vma_unlock_read() argument
370 hugetlb_vma_lock_write(struct vm_area_struct * vma) hugetlb_vma_lock_write() argument
374 hugetlb_vma_unlock_write(struct vm_area_struct * vma) hugetlb_vma_unlock_write() argument
378 hugetlb_vma_trylock_write(struct vm_area_struct * vma) hugetlb_vma_trylock_write() argument
383 hugetlb_vma_assert_locked(struct vm_area_struct * vma) hugetlb_vma_assert_locked() argument
438 hugetlb_change_protection(struct vm_area_struct * vma,unsigned long address,unsigned long end,pgprot_t newprot,unsigned long cp_flags) hugetlb_change_protection() argument
446 __unmap_hugepage_range(struct mmu_gather * tlb,struct vm_area_struct * vma,unsigned long start,unsigned long end,struct folio * folio,zap_flags_t zap_flags) __unmap_hugepage_range() argument
454 hugetlb_fault(struct mm_struct * mm,struct vm_area_struct * vma,unsigned long address,unsigned int flags) hugetlb_fault() argument
461 hugetlb_unshare_all_pmds(struct vm_area_struct * vma) hugetlb_unshare_all_pmds() argument
463 fixup_hugetlb_reservations(struct vm_area_struct * vma) fixup_hugetlb_reservations() argument
467 hugetlb_split(struct vm_area_struct * vma,unsigned long addr) hugetlb_split() argument
762 hstate_vma(struct vm_area_struct * vma) hstate_vma() argument
1034 huge_ptep_modify_prot_start(struct vm_area_struct * vma,unsigned long addr,pte_t * ptep) huge_ptep_modify_prot_start() argument
1045 huge_ptep_modify_prot_commit(struct vm_area_struct * vma,unsigned long addr,pte_t * ptep,pte_t old_pte,pte_t pte) huge_ptep_modify_prot_commit() argument
1337 flush_hugetlb_tlb_range(vma,addr,end) global() argument
1340 __vma_shareable_lock(struct vm_area_struct * vma) __vma_shareable_lock() argument
1352 hugetlb_walk(struct vm_area_struct * vma,unsigned long addr,unsigned long sz) hugetlb_walk() argument
[all...]
/linux/drivers/gpu/drm/xe/
H A Dxe_vm.h110 static inline u64 xe_vma_start(struct xe_vma *vma) in xe_vma_start() argument
115 static inline u64 xe_vma_size(struct xe_vma *vma) in xe_vma_size() argument
120 static inline u64 xe_vma_end(struct xe_vma *vma) in xe_vma_end() argument
125 static inline u64 xe_vma_bo_offset(struct xe_vma *vma) in xe_vma_bo_offset() argument
130 xe_vma_bo(struct xe_vma * vma) xe_vma_bo() argument
136 xe_vma_vm(struct xe_vma * vma) xe_vma_vm() argument
141 xe_vma_read_only(struct xe_vma * vma) xe_vma_read_only() argument
146 xe_vma_userptr(struct xe_vma * vma) xe_vma_userptr() argument
151 xe_vma_is_null(struct xe_vma * vma) xe_vma_is_null() argument
156 xe_vma_is_cpu_addr_mirror(struct xe_vma * vma) xe_vma_is_cpu_addr_mirror() argument
161 xe_vma_has_no_bo(struct xe_vma * vma) xe_vma_has_no_bo() argument
166 xe_vma_is_userptr(struct xe_vma * vma) xe_vma_is_userptr() argument
180 to_userptr_vma(struct xe_vma * vma) to_userptr_vma() argument
[all...]
H A Dxe_vm.c71 struct xe_vma *vma = &uvma->vma; in xe_vma_userptr_pin_pages() local
589 struct xe_vma *vma = &uvma->vma; in __vma_userptr_invalidate() local
636 struct xe_vma *vma = &uvma->vma; vma_userptr_invalidate() local
805 struct xe_vma *vma; xe_vma_svm_prefetch_op_fini() local
846 xe_vm_populate_rebind(struct xe_vma_op * op,struct xe_vma * vma,u8 tile_mask) xe_vm_populate_rebind() argument
862 xe_vm_ops_add_rebind(struct xe_vma_ops * vops,struct xe_vma * vma,u8 tile_mask) xe_vm_ops_add_rebind() argument
887 struct xe_vma *vma, *next; xe_vm_rebind() local
939 xe_vma_rebind(struct xe_vm * vm,struct xe_vma * vma,u8 tile_mask) xe_vma_rebind() argument
982 xe_vm_populate_range_rebind(struct xe_vma_op * op,struct xe_vma * vma,struct xe_svm_range * range,u8 tile_mask) xe_vm_populate_range_rebind() argument
996 xe_vm_ops_add_range_rebind(struct xe_vma_ops * vops,struct xe_vma * vma,struct xe_svm_range * range,u8 tile_mask) xe_vm_ops_add_range_rebind() argument
1026 xe_vm_range_rebind(struct xe_vm * vm,struct xe_vma * vma,struct xe_svm_range * range,u8 tile_mask) xe_vm_range_rebind() argument
1154 xe_vma_free(struct xe_vma * vma) xe_vma_free() argument
1173 struct xe_vma *vma; xe_vma_create() local
1273 xe_vma_destroy_late(struct xe_vma * vma) xe_vma_destroy_late() argument
1308 struct xe_vma *vma = vma_destroy_work_func() local
1317 struct xe_vma *vma = container_of(cb, struct xe_vma, destroy_cb); vma_destroy_cb() local
1323 xe_vma_destroy(struct xe_vma * vma,struct dma_fence * fence) xe_vma_destroy() argument
1367 xe_vm_lock_vma(struct drm_exec * exec,struct xe_vma * vma) xe_vm_lock_vma() argument
1382 xe_vma_destroy_unlocked(struct xe_vma * vma) xe_vma_destroy_unlocked() argument
1417 xe_vm_insert_vma(struct xe_vm * vm,struct xe_vma * vma) xe_vm_insert_vma() argument
1432 xe_vm_remove_vma(struct xe_vm * vm,struct xe_vma * vma) xe_vm_remove_vma() argument
1543 xelp_pte_encode_vma(u64 pte,struct xe_vma * vma,u16 pat_index,u32 pt_level) xelp_pte_encode_vma() argument
1870 struct xe_vma *vma, *next_vma; xe_vm_close_and_put() local
2172 vma_matches(struct xe_vma * vma,u64 page_addr) vma_matches() argument
2189 struct xe_vma *vma = NULL; xe_vm_find_vma_by_addr() local
2207 prep_vma_destroy(struct xe_vm * vm,struct xe_vma * vma,bool post_commit) prep_vma_destroy() argument
2223 struct xe_vma *vma; print_op() local
2369 struct xe_vma *vma = gpuva_to_vma(op->base.prefetch.va); vm_bind_ioctl_ops_create() local
2450 struct xe_vma *vma; new_vma() local
2497 xe_vma_max_pte_size(struct xe_vma * vma) xe_vma_max_pte_size() argument
2511 xe_vma_set_pte_size(struct xe_vma * vma,u64 size) xe_vma_set_pte_size() argument
2608 struct xe_vma *vma; vm_bind_ioctl_ops_parse() local
2791 struct xe_vma *vma = gpuva_to_vma(op->base.unmap.va); xe_vma_op_unwind() local
2804 struct xe_vma *vma = gpuva_to_vma(op->base.remap.unmap->va); xe_vma_op_unwind() local
2855 vma_lock_and_validate(struct drm_exec * exec,struct xe_vma * vma,bool validate) vma_lock_and_validate() argument
2873 check_ufence(struct xe_vma * vma) check_ufence() argument
2891 struct xe_vma *vma = gpuva_to_vma(op->base.prefetch.va); prefetch_ranges() local
2975 struct xe_vma *vma = gpuva_to_vma(op->base.prefetch.va); op_lock_and_prep() local
3192 vma_add_ufence(struct xe_vma * vma,struct xe_user_fence * ufence) vma_add_ufence() argument
3917 xe_vm_invalidate_vma(struct xe_vma * vma) xe_vm_invalidate_vma() argument
3981 struct xe_vma *vma = gpuva_to_vma(gpuva); xe_vm_validate_protected() local
4035 struct xe_vma *vma = gpuva_to_vma(gpuva); xe_vm_snapshot_capture() local
[all...]
/linux/mm/
H A Dvma.c75 static bool vma_had_uncowed_parents(struct vm_area_struct *vma) in vma_had_uncowed_parents() argument
86 struct vm_area_struct *vma = merge_next ? vmg->next : vmg->prev; in is_mergeable_vma() local
141 init_multi_vma_prep(struct vma_prepare * vp,struct vm_area_struct * vma,struct vma_merge_struct * vmg) init_multi_vma_prep() argument
225 __vma_link_file(struct vm_area_struct * vma,struct address_space * mapping) __vma_link_file() argument
239 __remove_shared_vm_struct(struct vm_area_struct * vma,struct address_space * mapping) __remove_shared_vm_struct() argument
265 anon_vma_interval_tree_pre_update_vma(struct vm_area_struct * vma) anon_vma_interval_tree_pre_update_vma() argument
274 anon_vma_interval_tree_post_update_vma(struct vm_area_struct * vma) anon_vma_interval_tree_post_update_vma() argument
411 init_vma_prep(struct vma_prepare * vp,struct vm_area_struct * vma) init_vma_prep() argument
461 remove_vma(struct vm_area_struct * vma) remove_vma() argument
476 unmap_region(struct ma_state * mas,struct vm_area_struct * vma,struct vm_area_struct * prev,struct vm_area_struct * next) unmap_region() argument
499 __split_vma(struct vma_iterator * vmi,struct vm_area_struct * vma,unsigned long addr,int new_below) __split_vma() argument
592 split_vma(struct vma_iterator * vmi,struct vm_area_struct * vma,unsigned long addr,int new_below) split_vma() argument
648 struct vm_area_struct *vma; validate_mm() local
732 struct vm_area_struct *vma; commit_merge() local
773 can_merge_remove_vma(struct vm_area_struct * vma) can_merge_remove_vma() argument
1185 vma_shrink(struct vma_iterator * vmi,struct vm_area_struct * vma,unsigned long start,unsigned long end,pgoff_t pgoff) vma_shrink() argument
1242 struct vm_area_struct *vma; vms_clean_up_area() local
1265 struct vm_area_struct *vma; vms_complete_munmap_vmas() local
1311 struct vm_area_struct *vma; reattach_vmas() local
1479 init_vma_munmap(struct vma_munmap_struct * vms,struct vma_iterator * vmi,struct vm_area_struct * vma,unsigned long start,unsigned long end,struct list_head * uf,bool unlock) init_vma_munmap() argument
1515 do_vmi_align_munmap(struct vma_iterator * vmi,struct vm_area_struct * vma,struct mm_struct * mm,unsigned long start,unsigned long end,struct list_head * uf,bool unlock) do_vmi_align_munmap() argument
1567 struct vm_area_struct *vma; do_vmi_munmap() local
1602 struct vm_area_struct *vma = vmg->middle; vma_modify() local
1640 vma_modify_flags(struct vma_iterator * vmi,struct vm_area_struct * prev,struct vm_area_struct * vma,unsigned long start,unsigned long end,vm_flags_t vm_flags) vma_modify_flags() argument
1655 vma_modify_name(struct vma_iterator * vmi,struct vm_area_struct * prev,struct vm_area_struct * vma,unsigned long start,unsigned long end,struct anon_vma_name * new_name) vma_modify_name() argument
1670 vma_modify_policy(struct vma_iterator * vmi,struct vm_area_struct * prev,struct vm_area_struct * vma,unsigned long start,unsigned long end,struct mempolicy * new_pol) vma_modify_policy() argument
1684 vma_modify_flags_uffd(struct vma_iterator * vmi,struct vm_area_struct * prev,struct vm_area_struct * vma,unsigned long start,unsigned long end,vm_flags_t vm_flags,struct vm_userfaultfd_ctx new_ctx,bool give_up_on_oom) vma_modify_flags_uffd() argument
1704 vma_merge_extend(struct vma_iterator * vmi,struct vm_area_struct * vma,unsigned long delta) vma_merge_extend() argument
1738 unlink_file_vma_batch_add(struct unlink_vma_file_batch * vb,struct vm_area_struct * vma) unlink_file_vma_batch_add() argument
1761 unlink_file_vma(struct vm_area_struct * vma) unlink_file_vma() argument
1774 vma_link_file(struct vm_area_struct * vma) vma_link_file() argument
1787 vma_link(struct mm_struct * mm,struct vm_area_struct * vma) vma_link() argument
1811 struct vm_area_struct *vma = *vmap; copy_vma() local
1967 find_mergeable_anon_vma(struct vm_area_struct * vma) find_mergeable_anon_vma() argument
2006 vma_is_shared_writable(struct vm_area_struct * vma) vma_is_shared_writable() argument
2012 vma_fs_can_writeback(struct vm_area_struct * vma) vma_fs_can_writeback() argument
2026 vma_needs_dirty_tracking(struct vm_area_struct * vma) vma_needs_dirty_tracking() argument
2049 vma_wants_writenotify(struct vm_area_struct * vma,pgprot_t vm_page_prot) vma_wants_writenotify() argument
2163 struct vm_area_struct *vma; mm_take_all_locks() local
2259 struct vm_area_struct *vma; mm_drop_all_locks() local
2403 __mmap_new_file_vma(struct mmap_state * map,struct vm_area_struct * vma) __mmap_new_file_vma() argument
2454 struct vm_area_struct *vma; __mmap_new_vma() local
2524 __mmap_complete(struct mmap_state * map,struct vm_area_struct * vma) __mmap_complete() argument
2601 set_vma_user_defined_fields(struct vm_area_struct * vma,struct mmap_state * map) set_vma_user_defined_fields() argument
2643 struct vm_area_struct *vma = NULL; __mmap_region() local
2761 do_brk_flags(struct vma_iterator * vmi,struct vm_area_struct * vma,unsigned long addr,unsigned long len,vm_flags_t vm_flags) do_brk_flags() argument
2945 acct_stack_growth(struct vm_area_struct * vma,unsigned long size,unsigned long grow) acct_stack_growth() argument
2984 expand_upwards(struct vm_area_struct * vma,unsigned long address) expand_upwards() argument
3070 expand_downwards(struct vm_area_struct * vma,unsigned long address) expand_downwards() argument
3167 insert_vm_struct(struct mm_struct * mm,struct vm_area_struct * vma) insert_vm_struct() argument
[all...]
H A Dmmap_lock.c48 static inline bool __vma_enter_locked(struct vm_area_struct *vma, bool detaching) in __vma_enter_locked() argument
72 static inline void __vma_exit_locked(struct vm_area_struct *vma, bool *detached) in __vma_exit_locked() argument
78 __vma_start_write(struct vm_area_struct * vma,unsigned int mm_lock_seq) __vma_start_write() argument
106 vma_mark_detached(struct vm_area_struct * vma) vma_mark_detached() argument
139 struct vm_area_struct *vma; lock_vma_under_rcu() local
185 struct vm_area_struct *vma; lock_next_vma_under_mmap_lock() local
210 struct vm_area_struct *vma; lock_next_vma() local
335 struct vm_area_struct *vma; lock_mm_and_find_vma() local
397 struct vm_area_struct *vma; lock_mm_and_find_vma() local
[all...]
H A Dmemory.c373 struct vm_area_struct *vma, unsigned long floor, in free_pgtables() argument
501 print_bad_pte(struct vm_area_struct * vma,unsigned long addr,pte_t pte,struct page * page) print_bad_pte() argument
598 vm_normal_page(struct vm_area_struct * vma,unsigned long addr,pte_t pte) vm_normal_page() argument
654 vm_normal_folio(struct vm_area_struct * vma,unsigned long addr,pte_t pte) vm_normal_folio() argument
665 vm_normal_page_pmd(struct vm_area_struct * vma,unsigned long addr,pmd_t pmd) vm_normal_page_pmd() argument
702 vm_normal_folio_pmd(struct vm_area_struct * vma,unsigned long addr,pmd_t pmd) vm_normal_folio_pmd() argument
739 restore_exclusive_pte(struct vm_area_struct * vma,struct folio * folio,struct page * page,unsigned long address,pte_t * ptep,pte_t orig_pte) restore_exclusive_pte() argument
773 try_restore_exclusive_pte(struct vm_area_struct * vma,unsigned long addr,pte_t * ptep,pte_t orig_pte) try_restore_exclusive_pte() argument
1056 folio_prealloc(struct mm_struct * src_mm,struct vm_area_struct * vma,unsigned long addr,bool need_zero) folio_prealloc() argument
1462 zap_install_uffd_wp_if_needed(struct vm_area_struct * vma,unsigned long addr,pte_t * pte,int nr,struct zap_details * details,pte_t pteval) zap_install_uffd_wp_if_needed() argument
1490 zap_present_folio_ptes(struct mmu_gather * tlb,struct vm_area_struct * vma,struct folio * folio,struct page * page,pte_t * pte,pte_t ptent,unsigned int nr,unsigned long addr,struct zap_details * details,int * rss,bool * force_flush,bool * force_break,bool * any_skipped) zap_present_folio_ptes() argument
1541 zap_present_ptes(struct mmu_gather * tlb,struct vm_area_struct * vma,pte_t * pte,pte_t ptent,unsigned int max_nr,unsigned long addr,struct zap_details * details,int * rss,bool * force_flush,bool * force_break,bool * any_skipped) zap_present_ptes() argument
1587 zap_nonpresent_ptes(struct mmu_gather * tlb,struct vm_area_struct * vma,pte_t * pte,pte_t ptent,unsigned int max_nr,unsigned long addr,struct zap_details * details,int * rss,bool * any_skipped) zap_nonpresent_ptes() argument
1657 do_zap_pte_range(struct mmu_gather * tlb,struct vm_area_struct * vma,pte_t * pte,unsigned long addr,unsigned long end,struct zap_details * details,int * rss,bool * force_flush,bool * force_break,bool * any_skipped) do_zap_pte_range() argument
1693 zap_pte_range(struct mmu_gather * tlb,struct vm_area_struct * vma,pmd_t * pmd,unsigned long addr,unsigned long end,struct zap_details * details) zap_pte_range() argument
1785 zap_pmd_range(struct mmu_gather * tlb,struct vm_area_struct * vma,pud_t * pud,unsigned long addr,unsigned long end,struct zap_details * details) zap_pmd_range() argument
1827 zap_pud_range(struct mmu_gather * tlb,struct vm_area_struct * vma,p4d_t * p4d,unsigned long addr,unsigned long end,struct zap_details * details) zap_pud_range() argument
1856 zap_p4d_range(struct mmu_gather * tlb,struct vm_area_struct * vma,pgd_t * pgd,unsigned long addr,unsigned long end,struct zap_details * details) zap_p4d_range() argument
1875 unmap_page_range(struct mmu_gather * tlb,struct vm_area_struct * vma,unsigned long addr,unsigned long end,struct zap_details * details) unmap_page_range() argument
1896 unmap_single_vma(struct mmu_gather * tlb,struct vm_area_struct * vma,unsigned long start_addr,unsigned long end_addr,struct zap_details * details,bool mm_wr_locked) unmap_single_vma() argument
1958 unmap_vmas(struct mmu_gather * tlb,struct ma_state * mas,struct vm_area_struct * vma,unsigned long start_addr,unsigned long end_addr,unsigned long tree_end,bool mm_wr_locked) unmap_vmas() argument
1996 zap_page_range_single_batched(struct mmu_gather * tlb,struct vm_area_struct * vma,unsigned long address,unsigned long size,struct zap_details * details) zap_page_range_single_batched() argument
2035 zap_page_range_single(struct vm_area_struct * vma,unsigned long address,unsigned long size,struct zap_details * details) zap_page_range_single() argument
2056 zap_vma_ptes(struct vm_area_struct * vma,unsigned long address,unsigned long size) zap_vma_ptes() argument
2099 vm_mixed_zeropage_allowed(struct vm_area_struct * vma) vm_mixed_zeropage_allowed() argument
2129 validate_page_before_insert(struct vm_area_struct * vma,struct page * page) validate_page_before_insert() argument
2148 insert_page_into_pte_locked(struct vm_area_struct * vma,pte_t * pte,unsigned long addr,struct page * page,pgprot_t prot,bool mkwrite) insert_page_into_pte_locked() argument
2189 insert_page(struct vm_area_struct * vma,unsigned long addr,struct page * page,pgprot_t prot,bool mkwrite) insert_page() argument
2210 insert_page_in_batch_locked(struct vm_area_struct * vma,pte_t * pte,unsigned long addr,struct page * page,pgprot_t prot) insert_page_in_batch_locked() argument
2224 insert_pages(struct vm_area_struct * vma,unsigned long addr,struct page ** pages,unsigned long * num,pgprot_t prot) insert_pages() argument
2297 vm_insert_pages(struct vm_area_struct * vma,unsigned long addr,struct page ** pages,unsigned long * num) vm_insert_pages() argument
2344 vm_insert_page(struct vm_area_struct * vma,unsigned long addr,struct page * page) vm_insert_page() argument
2371 __vm_map_pages(struct vm_area_struct * vma,struct page ** pages,unsigned long num,unsigned long offset) __vm_map_pages() argument
2414 vm_map_pages(struct vm_area_struct * vma,struct page ** pages,unsigned long num) vm_map_pages() argument
2434 vm_map_pages_zero(struct vm_area_struct * vma,struct page ** pages,unsigned long num) vm_map_pages_zero() argument
2441 insert_pfn(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,pgprot_t prot,bool mkwrite) insert_pfn() argument
2525 vmf_insert_pfn_prot(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,pgprot_t pgprot) vmf_insert_pfn_prot() argument
2572 vmf_insert_pfn(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn) vmf_insert_pfn() argument
2579 vm_mixed_ok(struct vm_area_struct * vma,unsigned long pfn,bool mkwrite) vm_mixed_ok() argument
2593 __vm_insert_mixed(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,bool mkwrite) __vm_insert_mixed() argument
2659 vmf_insert_mixed(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn) vmf_insert_mixed() argument
2671 vmf_insert_mixed_mkwrite(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn) vmf_insert_mixed_mkwrite() argument
2775 remap_pfn_range_internal(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,unsigned long size,pgprot_t prot) remap_pfn_range_internal() argument
2832 remap_pfn_range_notrack(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,unsigned long size,pgprot_t prot) remap_pfn_range_notrack() argument
2892 remap_pfn_range(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,unsigned long size,pgprot_t prot) remap_pfn_range() argument
2930 remap_pfn_range(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,unsigned long size,pgprot_t prot) remap_pfn_range() argument
2953 vm_iomap_memory(struct vm_area_struct * vma,phys_addr_t start,unsigned long len) vm_iomap_memory() argument
3238 struct vm_area_struct *vma = vmf->vma; __wp_page_copy_user() local
3330 __get_fault_gfp_mask(struct vm_area_struct * vma) __get_fault_gfp_mask() argument
3385 struct vm_area_struct *vma = vmf->vma; fault_dirty_shared_page() local
3439 struct vm_area_struct *vma = vmf->vma; wp_page_reuse() local
3472 struct vm_area_struct *vma = vmf->vma; vmf_can_call_fault() local
3497 struct vm_area_struct *vma = vmf->vma; __vmf_anon_prepare() local
3533 struct vm_area_struct *vma = vmf->vma; wp_page_copy() local
3724 struct vm_area_struct *vma = vmf->vma; wp_pfn_shared() local
3747 struct vm_area_struct *vma = vmf->vma; wp_page_shared() local
3786 __wp_can_reuse_large_anon_folio(struct folio * folio,struct vm_area_struct * vma) __wp_can_reuse_large_anon_folio() argument
3852 __wp_can_reuse_large_anon_folio(struct folio * folio,struct vm_area_struct * vma) __wp_can_reuse_large_anon_folio() argument
3859 wp_can_reuse_anon_folio(struct folio * folio,struct vm_area_struct * vma) wp_can_reuse_anon_folio() argument
3925 struct vm_area_struct *vma = vmf->vma; do_wp_page() local
4016 unmap_mapping_range_vma(struct vm_area_struct * vma,unsigned long start_addr,unsigned long end_addr,struct zap_details * details) unmap_mapping_range_vma() argument
4028 struct vm_area_struct *vma; unmap_mapping_range_tree() local
4150 struct vm_area_struct *vma = vmf->vma; remove_device_exclusive_entry() local
4191 should_try_to_free_swap(struct folio * folio,struct vm_area_struct * vma,unsigned int fault_flags) should_try_to_free_swap() argument
4282 struct vm_area_struct *vma = vmf->vma; __alloc_swap_folio() local
4360 struct vm_area_struct *vma = vmf->vma; alloc_swap_folio() local
4456 struct vm_area_struct *vma = vmf->vma; do_swap_page() local
4917 struct vm_area_struct *vma = vmf->vma; alloc_anon_folio() local
5007 struct vm_area_struct *vma = vmf->vma; do_anonymous_page() local
5127 struct vm_area_struct *vma = vmf->vma; __do_fault() local
5184 struct vm_area_struct *vma = vmf->vma; deposit_prealloc_pte() local
5197 struct vm_area_struct *vma = vmf->vma; do_set_pmd() local
5286 struct vm_area_struct *vma = vmf->vma; set_pte_range() local
5344 struct vm_area_struct *vma = vmf->vma; finish_fault() local
5587 struct vm_area_struct *vma = vmf->vma; do_cow_fault() local
5629 struct vm_area_struct *vma = vmf->vma; do_shared_fault() local
5679 struct vm_area_struct *vma = vmf->vma; do_fault() local
5725 struct vm_area_struct *vma = vmf->vma; numa_migrate_check() local
5768 numa_rebuild_single_mapping(struct vm_fault * vmf,struct vm_area_struct * vma,unsigned long fault_addr,pte_t * fault_pte,bool writable) numa_rebuild_single_mapping() argument
5783 numa_rebuild_large_mapping(struct vm_fault * vmf,struct vm_area_struct * vma,struct folio * folio,pte_t fault_pte,bool ignore_writable,bool pte_write_upgrade) numa_rebuild_large_mapping() argument
5824 struct vm_area_struct *vma = vmf->vma; do_numa_page() local
5915 struct vm_area_struct *vma = vmf->vma; create_huge_pmd() local
5926 struct vm_area_struct *vma = vmf->vma; wp_huge_pmd() local
5959 struct vm_area_struct *vma = vmf->vma; create_huge_pud() local
5973 struct vm_area_struct *vma = vmf->vma; wp_huge_pud() local
6102 __handle_mm_fault(struct vm_area_struct * vma,unsigned long address,unsigned int flags) __handle_mm_fault() argument
6267 lru_gen_enter_fault(struct vm_area_struct * vma) lru_gen_enter_fault() argument
6278 lru_gen_enter_fault(struct vm_area_struct * vma) lru_gen_enter_fault() argument
6287 sanitize_fault_flags(struct vm_area_struct * vma,unsigned int * flags) sanitize_fault_flags() argument
6329 handle_mm_fault(struct vm_area_struct * vma,unsigned long address,unsigned int flags,struct pt_regs * regs) handle_mm_fault() argument
6482 pfnmap_lockdep_assert(struct vm_area_struct * vma) pfnmap_lockdep_assert() argument
6529 struct vm_area_struct *vma = args->vma; follow_pfnmap_start() local
6632 generic_access_phys(struct vm_area_struct * vma,unsigned long addr,void * buf,int len,int write) generic_access_phys() argument
6707 struct vm_area_struct *vma = NULL; __access_remote_vm() local
6834 struct vm_area_struct *vma = NULL; __copy_remote_vm_str() local
6931 struct vm_area_struct *vma; print_vma_addr() local
7062 copy_user_gigantic_page(struct folio * dst,struct folio * src,unsigned long addr_hint,struct vm_area_struct * vma,unsigned int nr_pages) copy_user_gigantic_page() argument
7085 struct vm_area_struct *vma; global() member
7100 copy_user_large_folio(struct folio * dst,struct folio * src,unsigned long addr_hint,struct vm_area_struct * vma) copy_user_large_folio() argument
7175 vma_pgtable_walk_begin(struct vm_area_struct * vma) vma_pgtable_walk_begin() argument
7181 vma_pgtable_walk_end(struct vm_area_struct * vma) vma_pgtable_walk_end() argument
[all...]
H A Dmprotect.c43 static bool maybe_change_pte_writable(struct vm_area_struct *vma, pte_t pte) in maybe_change_pte_writable() argument
63 static bool can_change_private_pte_writable(struct vm_area_struct *vma, in can_change_private_pte_writable() argument
81 static bool can_change_shared_pte_writable(struct vm_area_struct *vma, in can_change_shared_pte_writable() argument
99 bool can_change_pte_writable(struct vm_area_struct *vma, unsigne argument
121 prot_numa_skip(struct vm_area_struct * vma,unsigned long addr,pte_t oldpte,pte_t * pte,int target_node,struct folio * folio) prot_numa_skip() argument
178 prot_commit_flush_ptes(struct vm_area_struct * vma,unsigned long addr,pte_t * ptep,pte_t oldpte,pte_t ptent,int nr_ptes,int idx,bool set_write,struct mmu_gather * tlb) prot_commit_flush_ptes() argument
230 commit_anon_folio_batch(struct vm_area_struct * vma,struct folio * folio,struct page * first_page,unsigned long addr,pte_t * ptep,pte_t oldpte,pte_t ptent,int nr_ptes,struct mmu_gather * tlb) commit_anon_folio_batch() argument
249 set_write_prot_commit_flush_ptes(struct vm_area_struct * vma,struct folio * folio,struct page * page,unsigned long addr,pte_t * ptep,pte_t oldpte,pte_t ptent,int nr_ptes,struct mmu_gather * tlb) set_write_prot_commit_flush_ptes() argument
273 change_pte_range(struct mmu_gather * tlb,struct vm_area_struct * vma,pmd_t * pmd,unsigned long addr,unsigned long end,pgprot_t newprot,unsigned long cp_flags) change_pte_range() argument
453 pgtable_split_needed(struct vm_area_struct * vma,unsigned long cp_flags) pgtable_split_needed() argument
469 pgtable_populate_needed(struct vm_area_struct * vma,unsigned long cp_flags) pgtable_populate_needed() argument
485 change_pmd_prepare(vma,pmd,cp_flags) global() argument
500 change_prepare(vma,high,low,addr,cp_flags) global() argument
512 change_pmd_range(struct mmu_gather * tlb,struct vm_area_struct * vma,pud_t * pud,unsigned long addr,unsigned long end,pgprot_t newprot,unsigned long cp_flags) change_pmd_range() argument
582 change_pud_range(struct mmu_gather * tlb,struct vm_area_struct * vma,p4d_t * p4d,unsigned long addr,unsigned long end,pgprot_t newprot,unsigned long cp_flags) change_pud_range() argument
641 change_p4d_range(struct mmu_gather * tlb,struct vm_area_struct * vma,pgd_t * pgd,unsigned long addr,unsigned long end,pgprot_t newprot,unsigned long cp_flags) change_p4d_range() argument
664 change_protection_range(struct mmu_gather * tlb,struct vm_area_struct * vma,unsigned long addr,unsigned long end,pgprot_t newprot,unsigned long cp_flags) change_protection_range() argument
694 change_protection(struct mmu_gather * tlb,struct vm_area_struct * vma,unsigned long start,unsigned long end,unsigned long cp_flags) change_protection() argument
756 mprotect_fixup(struct vma_iterator * vmi,struct mmu_gather * tlb,struct vm_area_struct * vma,struct vm_area_struct ** pprev,unsigned long start,unsigned long end,vm_flags_t newflags) mprotect_fixup() argument
865 struct vm_area_struct *vma, *prev; do_mprotect_pkey() local
[all...]
H A Dnommu.c92 struct vm_area_struct *vma; in kobjsize() local
147 struct vm_area_struct *vma; in __vmalloc_user_flags() local
344 int vm_insert_page(struct vm_area_struct *vma, unsigne argument
351 vm_insert_pages(struct vm_area_struct * vma,unsigned long addr,struct page ** pages,unsigned long * num) vm_insert_pages() argument
358 vm_map_pages(struct vm_area_struct * vma,struct page ** pages,unsigned long num) vm_map_pages() argument
365 vm_map_pages_zero(struct vm_area_struct * vma,struct page ** pages,unsigned long num) vm_map_pages_zero() argument
561 setup_vma_to_mm(struct vm_area_struct * vma,struct mm_struct * mm) setup_vma_to_mm() argument
577 cleanup_vma_from_mm(struct vm_area_struct * vma) cleanup_vma_from_mm() argument
596 delete_vma_from_mm(struct vm_area_struct * vma) delete_vma_from_mm() argument
615 delete_vma(struct mm_struct * mm,struct vm_area_struct * vma) delete_vma() argument
651 expand_stack_locked(struct vm_area_struct * vma,unsigned long addr) expand_stack_locked() argument
670 struct vm_area_struct *vma; find_vma_exact() local
892 do_mmap_shared_file(struct vm_area_struct * vma) do_mmap_shared_file() argument
913 do_mmap_private(struct vm_area_struct * vma,struct vm_region * region,unsigned long len,unsigned long capabilities) do_mmap_private() argument
1021 struct vm_area_struct *vma; do_mmap() local
1305 split_vma(struct vma_iterator * vmi,struct vm_area_struct * vma,unsigned long addr,int new_below) split_vma() argument
1384 vmi_shrink_vma(struct vma_iterator * vmi,struct vm_area_struct * vma,unsigned long from,unsigned long to) vmi_shrink_vma() argument
1428 struct vm_area_struct *vma; do_munmap() local
1510 struct vm_area_struct *vma; exit_mmap() local
1545 struct vm_area_struct *vma; do_mremap() local
1589 remap_pfn_range(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,unsigned long size,pgprot_t prot) remap_pfn_range() argument
1600 vm_iomap_memory(struct vm_area_struct * vma,phys_addr_t start,unsigned long len) vm_iomap_memory() argument
1610 remap_vmalloc_range(struct vm_area_struct * vma,void * addr,unsigned long pgoff) remap_vmalloc_range() argument
1643 struct vm_area_struct *vma; __access_remote_vm() local
1722 struct vm_area_struct *vma; __copy_remote_vm_str() local
1806 struct vm_area_struct *vma; nommu_shrink_inode_mappings() local
[all...]
H A Dpgtable-generic.c68 int ptep_set_access_flags(struct vm_area_struct *vma, in ptep_set_access_flags() argument
82 int ptep_clear_flush_young(struct vm_area_struct *vma, in ptep_clear_flush_young() argument
94 pte_t ptep_clear_flush(struct vm_area_struct *vma, unsigned long address, in ptep_clear_flush() argument
109 int pmdp_set_access_flags(struct vm_area_struct *vma, in pmdp_set_access_flags() argument
124 pmdp_clear_flush_young(struct vm_area_struct * vma,unsigned long address,pmd_t * pmdp) pmdp_clear_flush_young() argument
137 pmdp_huge_clear_flush(struct vm_area_struct * vma,unsigned long address,pmd_t * pmdp) pmdp_huge_clear_flush() argument
149 pudp_huge_clear_flush(struct vm_area_struct * vma,unsigned long address,pud_t * pudp) pudp_huge_clear_flush() argument
197 pmdp_invalidate(struct vm_area_struct * vma,unsigned long address,pmd_t * pmdp) pmdp_invalidate() argument
208 pmdp_invalidate_ad(struct vm_area_struct * vma,unsigned long address,pmd_t * pmdp) pmdp_invalidate_ad() argument
217 pmdp_collapse_flush(struct vm_area_struct * vma,unsigned long address,pmd_t * pmdp) pmdp_collapse_flush() argument
[all...]
H A Dmadvise.c80 struct vm_area_struct *vma; member
110 struct anon_vma_name *anon_vma_name(struct vm_area_struct *vma) in anon_vma_name() argument
119 static int replace_anon_vma_name(struct vm_area_struct *vma, in replace_anon_vma_name() argument
139 static int replace_anon_vma_name(struct vm_area_struct *vma, in replace_anon_vma_name() argument
155 struct vm_area_struct *vma = madv_behavior->vma; madvise_update_vma() local
190 struct vm_area_struct *vma = walk->private; swapin_walk_pmd_entry() local
236 shmem_swapin_range(struct vm_area_struct * vma,unsigned long start,unsigned long end,struct address_space * mapping) shmem_swapin_range() argument
285 struct vm_area_struct *vma = madv_behavior->vma; madvise_willneed() local
331 can_do_file_pageout(struct vm_area_struct * vma) can_do_file_pageout() argument
364 struct vm_area_struct *vma = walk->vma; madvise_cold_or_pageout_pte_range() local
580 struct vm_area_struct *vma = madv_behavior->vma; madvise_cold_page_range() local
593 can_madv_lru_vma(struct vm_area_struct * vma) can_madv_lru_vma() argument
600 struct vm_area_struct *vma = madv_behavior->vma; madvise_cold() local
615 madvise_pageout_page_range(struct mmu_gather * tlb,struct vm_area_struct * vma,struct madvise_behavior_range * range) madvise_pageout_page_range() argument
632 struct vm_area_struct *vma = madv_behavior->vma; madvise_pageout() local
662 struct vm_area_struct *vma = walk->vma; madvise_free_pte_range() local
806 struct vm_area_struct *vma = madv_behavior->vma; madvise_free_single_vma() local
878 struct vm_area_struct *vma = madv_behavior->vma; madvise_dontneed_free_valid_vma() local
920 struct vm_area_struct *vma; madvise_dontneed_free() local
1017 struct vm_area_struct *vma = madv_behavior->vma; madvise_remove() local
1057 is_valid_guard_vma(struct vm_area_struct * vma,bool allow_locked) is_valid_guard_vma() argument
1135 struct vm_area_struct *vma = madv_behavior->vma; madvise_guard_install() local
1246 struct vm_area_struct *vma = madv_behavior->vma; madvise_guard_remove() local
1287 struct vm_area_struct *vma = madv_behavior->vma; can_madvise_modify() local
1336 struct vm_area_struct *vma = madv_behavior->vma; madvise_vma_behavior() local
1586 struct vm_area_struct *vma; try_vma_read_lock() local
1627 struct vm_area_struct *prev, *vma; madvise_walk_vmas() local
[all...]
H A Dmmap.c81 void vma_set_page_prot(struct vm_area_struct *vma) in vma_set_page_prot() argument
693 struct vm_area_struct *vma, *pre in generic_get_unmapped_area() local
741 struct vm_area_struct *vma, *prev; generic_get_unmapped_area_topdown() local
930 struct vm_area_struct *vma; find_vma_prev() local
957 expand_stack_locked(struct vm_area_struct * vma,unsigned long address) expand_stack_locked() argument
964 struct vm_area_struct *vma, *prev; find_extend_vma_locked() local
979 expand_stack_locked(struct vm_area_struct * vma,unsigned long address) expand_stack_locked() argument
986 struct vm_area_struct *vma; find_extend_vma_locked() local
1006 vma_expand_up(vma,addr) global() argument
1007 vma_expand_down(vma,addr) global() argument
1011 vma_expand_up(vma,addr) global() argument
1012 vma_expand_down(vma,addr) global() argument
1029 struct vm_area_struct *vma, *prev; expand_stack() local
1092 struct vm_area_struct *vma; SYSCALL_DEFINE5() local
1209 struct vm_area_struct *vma = NULL; vm_brk_flags() local
1257 struct vm_area_struct *vma; exit_mmap() local
1368 special_mapping_close(struct vm_area_struct * vma) special_mapping_close() argument
1376 special_mapping_name(struct vm_area_struct * vma) special_mapping_name() argument
1394 special_mapping_split(struct vm_area_struct * vma,unsigned long addr) special_mapping_split() argument
1417 struct vm_area_struct *vma = vmf->vma; special_mapping_fault() local
1447 struct vm_area_struct *vma; __install_special_mapping() local
1476 vma_is_special_mapping(const struct vm_area_struct * vma,const struct vm_special_mapping * sm) vma_is_special_mapping() argument
[all...]
H A Dhuge_memory.c85 static inline bool file_thp_enabled(struct vm_area_struct *vma) in file_thp_enabled() argument
100 unsigned long __thp_vma_allowable_orders(struct vm_area_struct *vma, in __thp_vma_allowable_orders() argument
1052 maybe_pmd_mkwrite(pmd_t pmd,struct vm_area_struct * vma) maybe_pmd_mkwrite() argument
1157 vma_alloc_anon_folio_pmd(struct vm_area_struct * vma,unsigned long addr) vma_alloc_anon_folio_pmd() argument
1201 map_anon_folio_pmd(struct folio * folio,pmd_t * pmd,struct vm_area_struct * vma,unsigned long haddr) map_anon_folio_pmd() argument
1220 struct vm_area_struct *vma = vmf->vma; __do_huge_pmd_anonymous_page() local
1279 vma_thp_gfp_mask(struct vm_area_struct * vma) vma_thp_gfp_mask() argument
1307 set_huge_zero_folio(pgtable_t pgtable,struct mm_struct * mm,struct vm_area_struct * vma,unsigned long haddr,pmd_t * pmd,struct folio * zero_folio) set_huge_zero_folio() argument
1319 struct vm_area_struct *vma = vmf->vma; do_huge_pmd_anonymous_page() local
1382 insert_pmd(struct vm_area_struct * vma,unsigned long addr,pmd_t * pmd,struct folio_or_pfn fop,pgprot_t prot,bool write,pgtable_t pgtable) insert_pmd() argument
1448 struct vm_area_struct *vma = vmf->vma; vmf_insert_pfn_pmd() local
1492 struct vm_area_struct *vma = vmf->vma; vmf_insert_folio_pmd() local
1527 maybe_pud_mkwrite(pud_t pud,struct vm_area_struct * vma) maybe_pud_mkwrite() argument
1534 insert_pud(struct vm_area_struct * vma,unsigned long addr,pud_t * pud,struct folio_or_pfn fop,pgprot_t prot,bool write) insert_pud() argument
1587 struct vm_area_struct *vma = vmf->vma; vmf_insert_pfn_pud() local
1628 struct vm_area_struct *vma = vmf->vma; vmf_insert_folio_pud() local
1653 touch_pmd(struct vm_area_struct * vma,unsigned long addr,pmd_t * pmd,bool write) touch_pmd() argument
1790 touch_pud(struct vm_area_struct * vma,unsigned long addr,pud_t * pud,bool write) touch_pud() argument
1805 copy_huge_pud(struct mm_struct * dst_mm,struct mm_struct * src_mm,pud_t * dst_pud,pud_t * src_pud,unsigned long addr,struct vm_area_struct * vma) copy_huge_pud() argument
1869 struct vm_area_struct *vma = vmf->vma; do_huge_zero_wp_pmd() local
1901 struct vm_area_struct *vma = vmf->vma; do_huge_pmd_wp_page() local
1992 can_change_pmd_writable(struct vm_area_struct * vma,unsigned long addr,pmd_t pmd) can_change_pmd_writable() argument
2025 struct vm_area_struct *vma = vmf->vma; do_huge_pmd_numa_page() local
2103 madvise_free_huge_pmd(struct mmu_gather * tlb,struct vm_area_struct * vma,pmd_t * pmd,unsigned long addr,unsigned long next) madvise_free_huge_pmd() argument
2182 zap_huge_pmd(struct mmu_gather * tlb,struct vm_area_struct * vma,pmd_t * pmd,unsigned long addr) zap_huge_pmd() argument
2260 pmd_move_must_withdraw(spinlock_t * new_pmd_ptl,spinlock_t * old_pmd_ptl,struct vm_area_struct * vma) pmd_move_must_withdraw() argument
2293 move_huge_pmd(struct vm_area_struct * vma,unsigned long old_addr,unsigned long new_addr,pmd_t * old_pmd,pmd_t * new_pmd) move_huge_pmd() argument
2351 change_huge_pmd(struct mmu_gather * tlb,struct vm_area_struct * vma,pmd_t * pmd,unsigned long addr,pgprot_t newprot,unsigned long cp_flags) change_huge_pmd() argument
2490 change_huge_pud(struct mmu_gather * tlb,struct vm_area_struct * vma,pud_t * pudp,unsigned long addr,pgprot_t newprot,unsigned long cp_flags) change_huge_pud() argument
2675 __pmd_trans_huge_lock(pmd_t * pmd,struct vm_area_struct * vma) __pmd_trans_huge_lock() argument
2691 __pud_trans_huge_lock(pud_t * pud,struct vm_area_struct * vma) __pud_trans_huge_lock() argument
2703 zap_huge_pud(struct mmu_gather * tlb,struct vm_area_struct * vma,pud_t * pud,unsigned long addr) zap_huge_pud() argument
2738 __split_huge_pud_locked(struct vm_area_struct * vma,pud_t * pud,unsigned long haddr) __split_huge_pud_locked() argument
2770 __split_huge_pud(struct vm_area_struct * vma,pud_t * pud,unsigned long address) __split_huge_pud() argument
2790 __split_huge_pud(struct vm_area_struct * vma,pud_t * pud,unsigned long address) __split_huge_pud() argument
2796 __split_huge_zero_page_pmd(struct vm_area_struct * vma,unsigned long haddr,pmd_t * pmd) __split_huge_zero_page_pmd() argument
2837 __split_huge_pmd_locked(struct vm_area_struct * vma,pmd_t * pmd,unsigned long haddr,bool freeze) __split_huge_pmd_locked() argument
3058 split_huge_pmd_locked(struct vm_area_struct * vma,unsigned long address,pmd_t * pmd,bool freeze) split_huge_pmd_locked() argument
3066 __split_huge_pmd(struct vm_area_struct * vma,pmd_t * pmd,unsigned long address,bool freeze) __split_huge_pmd() argument
3082 split_huge_pmd_address(struct vm_area_struct * vma,unsigned long address,bool freeze) split_huge_pmd_address() argument
3093 split_huge_pmd_if_needed(struct vm_area_struct * vma,unsigned long address) split_huge_pmd_if_needed() argument
3105 vma_adjust_trans_huge(struct vm_area_struct * vma,unsigned long start,unsigned long end,struct vm_area_struct * next) vma_adjust_trans_huge() argument
3144 __discard_anon_folio_pmd_locked(struct vm_area_struct * vma,unsigned long addr,pmd_t * pmdp,struct folio * folio) __discard_anon_folio_pmd_locked() argument
3207 unmap_huge_pmd_locked(struct vm_area_struct * vma,unsigned long addr,pmd_t * pmdp,struct folio * folio) unmap_huge_pmd_locked() argument
4296 vma_not_suitable_for_thp_split(struct vm_area_struct * vma) vma_not_suitable_for_thp_split() argument
4339 struct vm_area_struct *vma = vma_lookup(mm, addr); split_huge_pages_pid() local
4594 struct vm_area_struct *vma = pvmw->vma; set_pmd_migration_entry() local
4643 struct vm_area_struct *vma = pvmw->vma; remove_migration_pmd() local
[all...]
/linux/tools/testing/vma/
H A Dvma.c18 #define vma_iter_prealloc(vmi, vma) \ argument
85 static int attach_vma(struct mm_struct *mm, struct vm_area_struct *vma) in attach_vma() argument
95 static void detach_free_vma(struct vm_area_struct *vma) in detach_free_vma() argument
108 struct vm_area_struct *vma = alloc_vma(mm, start, end, pgoff, vm_flags); alloc_and_link_vma() local
131 struct vm_area_struct *vma; merge_new() local
153 struct vm_area_struct *vma; merge_existing() local
251 struct vm_area_struct *vma; cleanup_mm() local
269 vma_write_started(struct vm_area_struct * vma) vma_write_started() argument
285 __vma_set_dummy_anon_vma(struct vm_area_struct * vma,struct anon_vma_chain * avc,struct anon_vma * anon_vma) __vma_set_dummy_anon_vma() argument
295 vma_set_dummy_anon_vma(struct vm_area_struct * vma,struct anon_vma_chain * avc) vma_set_dummy_anon_vma() argument
303 struct vm_area_struct *vma; test_simple_merge() local
337 struct vm_area_struct *vma; test_simple_modify() local
399 struct vm_area_struct *vma = alloc_vma(&mm, 0, 0x1000, 0, vm_flags); test_simple_expand() local
427 struct vm_area_struct *vma = alloc_vma(&mm, 0, 0x3000, 0, vm_flags); test_simple_shrink() local
469 struct vm_area_struct *vma, *vma_a, *vma_b, *vma_c, *vma_d; test_merge_new() local
652 struct vm_area_struct *vma_left, *vma; test_vma_merge_special_flags() local
724 struct vm_area_struct *vma_prev, *vma_next, *vma; test_vma_merge_with_close() local
935 struct vm_area_struct *vma; test_vma_merge_new_with_close() local
981 struct vm_area_struct *vma, *vma_prev, *vma_next; test_merge_existing() local
1211 struct vm_area_struct *vma, *vma_prev, *vma_next; test_anon_vma_non_mergeable() local
1305 struct vm_area_struct *vma_prev, *vma_next, *vma; test_dup_anon_vma() local
1463 struct vm_area_struct *vma_prev, *vma; test_vmi_prealloc_fail() local
1524 struct vm_area_struct *vma; test_merge_extend() local
1555 struct vm_area_struct *vma, *vma_new, *vma_next; test_copy_vma() local
1587 struct vm_area_struct *vma_prev, *vma; test_expand_only_mode() local
1626 struct vm_area_struct *vma; test_mmap_region_basic() local
[all...]
/linux/arch/powerpc/include/asm/book3s/64/
H A Dtlbflush.h47 static inline void flush_pmd_tlb_range(struct vm_area_struct *vma, in flush_pmd_tlb_range() argument
55 static inline void flush_pud_tlb_range(struct vm_area_struct *vma, in flush_pud_tlb_range() argument
63 static inline void flush_hugetlb_tlb_range(struct vm_area_struct *vma, in flush_hugetlb_tlb_range() argument
71 static inline void flush_tlb_range(struct vm_area_struct *vma, in flush_tlb_range() argument
91 static inline void local_flush_tlb_page(struct vm_area_struct *vma, in local_flush_tlb_page() argument
120 flush_tlb_page(struct vm_area_struct * vma,unsigned long vmaddr) flush_tlb_page() argument
128 flush_tlb_page(vma,addr) global() argument
132 flush_tlb_fix_spurious_fault(struct vm_area_struct * vma,unsigned long address,pte_t * ptep) flush_tlb_fix_spurious_fault() argument
[all...]
/linux/arch/parisc/kernel/
H A Dcache.c331 __flush_cache_page(struct vm_area_struct *vma, unsigned long vmaddr, in __flush_cache_page() argument
393 void flush_icache_pages(struct vm_area_struct *vma, struct page *page, in flush_icache_pages() argument
476 struct vm_area_struct *vma; in flush_dcache_folio() local
623 flush_cache_page_if_present(struct vm_area_struct * vma,unsigned long vmaddr) flush_cache_page_if_present() argument
648 copy_user_highpage(struct page * to,struct page * from,unsigned long vaddr,struct vm_area_struct * vma) copy_user_highpage() argument
660 copy_to_user_page(struct vm_area_struct * vma,struct page * page,unsigned long user_vaddr,void * dst,void * src,int len) copy_to_user_page() argument
668 copy_from_user_page(struct vm_area_struct * vma,struct page * page,unsigned long user_vaddr,void * dst,void * src,int len) copy_from_user_page() argument
705 flush_cache_pages(struct vm_area_struct * vma,unsigned long start,unsigned long end) flush_cache_pages() argument
715 struct vm_area_struct *vma; mm_total_size() local
729 struct vm_area_struct *vma; flush_cache_mm() local
754 flush_cache_range(struct vm_area_struct * vma,unsigned long start,unsigned long end) flush_cache_range() argument
771 flush_cache_page(struct vm_area_struct * vma,unsigned long vmaddr,unsigned long pfn) flush_cache_page() argument
776 flush_anon_page(struct vm_area_struct * vma,struct page * page,unsigned long vmaddr) flush_anon_page() argument
784 ptep_clear_flush_young(struct vm_area_struct * vma,unsigned long addr,pte_t * ptep) ptep_clear_flush_young() argument
804 ptep_clear_flush(struct vm_area_struct * vma,unsigned long addr,pte_t * ptep) ptep_clear_flush() argument
[all...]
/linux/drivers/gpu/drm/nouveau/
H A Dnouveau_vmm.c29 nouveau_vma_unmap(struct nouveau_vma *vma) in nouveau_vma_unmap() argument
38 nouveau_vma_map(struct nouveau_vma *vma, struct nouveau_mem *mem) in nouveau_vma_map() argument
51 struct nouveau_vma *vma; in nouveau_vma_find() local
64 struct nouveau_vma *vma = *pvma; nouveau_vma_del() local
81 struct nouveau_vma *vma; nouveau_vma_new() local
[all...]
/linux/arch/x86/mm/
H A Dpgtable.c391 int ptep_set_access_flags(struct vm_area_struct *vma, in ptep_set_access_flags() argument
404 int pmdp_set_access_flags(struct vm_area_struct *vma, in pmdp_set_access_flags() argument
425 int pudp_set_access_flags(struct vm_area_struct *vma, unsigned long address, in pudp_set_access_flags() argument
446 int ptep_test_and_clear_young(struct vm_area_struct *vma, in ptep_test_and_clear_young() argument
459 int pmdp_test_and_clear_young(struct vm_area_struct *vma, in pmdp_test_and_clear_young() argument
473 int pudp_test_and_clear_young(struct vm_area_struct *vma, in pudp_test_and_clear_young() argument
486 int ptep_clear_flush_young(struct vm_area_struct *vma, in ptep_clear_flush_young() argument
506 int pmdp_clear_flush_young(struct vm_area_struct *vma, in pmdp_clear_flush_young() argument
520 pmdp_invalidate_ad(struct vm_area_struct * vma,unsigned long address,pmd_t * pmdp) pmdp_invalidate_ad() argument
535 pudp_invalidate(struct vm_area_struct * vma,unsigned long address,pud_t * pudp) pudp_invalidate() argument
802 pte_mkwrite(pte_t pte,struct vm_area_struct * vma) pte_mkwrite() argument
812 pmd_mkwrite(pmd_t pmd,struct vm_area_struct * vma) pmd_mkwrite() argument
822 arch_check_zapped_pte(struct vm_area_struct * vma,pte_t pte) arch_check_zapped_pte() argument
835 arch_check_zapped_pmd(struct vm_area_struct * vma,pmd_t pmd) arch_check_zapped_pmd() argument
842 arch_check_zapped_pud(struct vm_area_struct * vma,pud_t pud) arch_check_zapped_pud() argument
[all...]
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/
H A Dvmm.c802 struct nvkm_vma *vma = kzalloc(sizeof(*vma), GFP_KERNEL); in nvkm_vma_new() local
813 nvkm_vma_tail(struct nvkm_vma *vma, u64 tail) in nvkm_vma_tail() argument
836 nvkm_vmm_free_remove(struct nvkm_vmm * vmm,struct nvkm_vma * vma) nvkm_vmm_free_remove() argument
842 nvkm_vmm_free_delete(struct nvkm_vmm * vmm,struct nvkm_vma * vma) nvkm_vmm_free_delete() argument
850 nvkm_vmm_free_insert(struct nvkm_vmm * vmm,struct nvkm_vma * vma) nvkm_vmm_free_insert() argument
878 nvkm_vmm_node_remove(struct nvkm_vmm * vmm,struct nvkm_vma * vma) nvkm_vmm_node_remove() argument
884 nvkm_vmm_node_delete(struct nvkm_vmm * vmm,struct nvkm_vma * vma) nvkm_vmm_node_delete() argument
892 nvkm_vmm_node_insert(struct nvkm_vmm * vmm,struct nvkm_vma * vma) nvkm_vmm_node_insert() argument
918 struct nvkm_vma *vma = rb_entry(node, typeof(*vma), tree); nvkm_vmm_node_search() local
935 nvkm_vmm_node_merge(struct nvkm_vmm * vmm,struct nvkm_vma * prev,struct nvkm_vma * vma,struct nvkm_vma * next,u64 size) nvkm_vmm_node_merge() argument
977 nvkm_vmm_node_split(struct nvkm_vmm * vmm,struct nvkm_vma * vma,u64 addr,u64 size) nvkm_vmm_node_split() argument
1003 nvkm_vma_dump(struct nvkm_vma * vma) nvkm_vma_dump() argument
1021 struct nvkm_vma *vma; nvkm_vmm_dump() local
1030 struct nvkm_vma *vma; nvkm_vmm_dtor() local
1040 struct nvkm_vma *vma = rb_entry(node, typeof(*vma), tree); nvkm_vmm_dtor() local
1074 struct nvkm_vma *vma; nvkm_vmm_ctor_managed() local
1094 struct nvkm_vma *vma; nvkm_vmm_ctor() local
1211 nvkm_vmm_pfn_split_merge(struct nvkm_vmm * vmm,struct nvkm_vma * vma,u64 addr,u64 size,u8 page,bool map) nvkm_vmm_pfn_split_merge() argument
1236 struct nvkm_vma *vma = nvkm_vmm_node_search(vmm, addr); nvkm_vmm_pfn_unmap() local
1273 struct nvkm_vma *vma, *tmp; nvkm_vmm_pfn_map() local
1393 nvkm_vmm_unmap_region(struct nvkm_vmm * vmm,struct nvkm_vma * vma) nvkm_vmm_unmap_region() argument
1410 nvkm_vmm_unmap_locked(struct nvkm_vmm * vmm,struct nvkm_vma * vma,bool pfn) nvkm_vmm_unmap_locked() argument
1425 nvkm_vmm_unmap(struct nvkm_vmm * vmm,struct nvkm_vma * vma) nvkm_vmm_unmap() argument
1435 nvkm_vmm_map_valid(struct nvkm_vmm * vmm,struct nvkm_vma * vma,void * argv,u32 argc,struct nvkm_vmm_map * map) nvkm_vmm_map_valid() argument
1471 nvkm_vmm_map_choose(struct nvkm_vmm * vmm,struct nvkm_vma * vma,void * argv,u32 argc,struct nvkm_vmm_map * map) nvkm_vmm_map_choose() argument
1483 nvkm_vmm_map_locked(struct nvkm_vmm * vmm,struct nvkm_vma * vma,void * argv,u32 argc,struct nvkm_vmm_map * map) nvkm_vmm_map_locked() argument
1571 nvkm_vmm_map(struct nvkm_vmm * vmm,struct nvkm_vma * vma,void * argv,u32 argc,struct nvkm_vmm_map * map) nvkm_vmm_map() argument
1588 nvkm_vmm_put_region(struct nvkm_vmm * vmm,struct nvkm_vma * vma) nvkm_vmm_put_region() argument
1607 nvkm_vmm_put_locked(struct nvkm_vmm * vmm,struct nvkm_vma * vma) nvkm_vmm_put_locked() argument
1691 struct nvkm_vma *vma = *pvma; nvkm_vmm_put() local
1706 struct nvkm_vma *vma = NULL, *tmp; nvkm_vmm_get_locked() local
[all...]
/linux/drivers/gpu/drm/i915/display/
H A Dintel_fb_pin.c35 struct i915_vma *vma; in intel_fb_pin_to_dpt() local
124 struct i915_vma *vma; intel_fb_pin_to_ggtt() local
224 intel_fb_unpin_vma(struct i915_vma * vma,unsigned long flags) intel_fb_unpin_vma() argument
266 struct i915_vma *vma; intel_plane_pin_fb() local
321 struct i915_vma *vma; intel_plane_unpin_fb() local
338 intel_fb_get_map(struct i915_vma * vma,struct iosys_map * map) intel_fb_get_map() argument
[all...]

12345678910>>...29