/linux/drivers/gpu/drm/i915/ |
H A D | i915_vma.h | 51 static inline bool i915_vma_is_active(const struct i915_vma *vma) in i915_vma_is_active() argument 65 i915_vma_move_to_active(struct i915_vma *vma, struct i915_request *rq, in i915_vma_move_to_active() argument 73 static inline bool i915_vma_is_ggtt(const struct i915_vma *vma) in i915_vma_is_ggtt() argument 78 static inline bool i915_vma_is_dpt(const struct i915_vma *vma) in i915_vma_is_dpt() argument 83 static inline bool i915_vma_has_ggtt_write(const struct i915_vma *vma) in i915_vma_has_ggtt_write() argument 88 i915_vma_set_ggtt_write(struct i915_vma * vma) i915_vma_set_ggtt_write() argument 94 i915_vma_unset_ggtt_write(struct i915_vma * vma) i915_vma_unset_ggtt_write() argument 102 i915_vma_is_map_and_fenceable(const struct i915_vma * vma) i915_vma_is_map_and_fenceable() argument 107 i915_vma_set_userfault(struct i915_vma * vma) i915_vma_set_userfault() argument 113 i915_vma_unset_userfault(struct i915_vma * vma) i915_vma_unset_userfault() argument 118 i915_vma_has_userfault(const struct i915_vma * vma) i915_vma_has_userfault() argument 123 i915_vma_is_closed(const struct i915_vma * vma) i915_vma_is_closed() argument 129 __i915_vma_size(const struct i915_vma * vma) __i915_vma_size() argument 144 i915_vma_size(const struct i915_vma * vma) i915_vma_size() argument 151 __i915_vma_offset(const struct i915_vma * vma) __i915_vma_offset() argument 167 i915_vma_offset(const struct i915_vma * vma) i915_vma_offset() argument 173 i915_ggtt_offset(const struct i915_vma * vma) i915_ggtt_offset() argument 183 i915_ggtt_pin_bias(struct i915_vma * vma) i915_ggtt_pin_bias() argument 188 i915_vma_get(struct i915_vma * vma) i915_vma_get() argument 194 i915_vma_tryget(struct i915_vma * vma) i915_vma_tryget() argument 202 i915_vma_put(struct i915_vma * vma) i915_vma_put() argument 208 i915_vma_compare(struct i915_vma * vma,struct i915_address_space * vm,const struct i915_gtt_view * view) i915_vma_compare() argument 276 assert_vma_held(vma) global() argument 278 i915_vma_lock(struct i915_vma * vma) i915_vma_lock() argument 283 i915_vma_unlock(struct i915_vma * vma) i915_vma_unlock() argument 298 i915_vma_pin_count(const struct i915_vma * vma) i915_vma_pin_count() argument 303 i915_vma_is_pinned(const struct i915_vma * vma) i915_vma_is_pinned() argument 308 __i915_vma_pin(struct i915_vma * vma) __i915_vma_pin() argument 314 __i915_vma_unpin(struct i915_vma * vma) __i915_vma_unpin() argument 320 i915_vma_unpin(struct i915_vma * vma) i915_vma_unpin() argument 326 i915_vma_is_bound(const struct i915_vma * vma,unsigned int where) i915_vma_is_bound() argument 338 i915_vma_get_iomap(struct i915_vma * vma) i915_vma_get_iomap() argument 386 __i915_vma_unpin_fence(struct i915_vma * vma) __i915_vma_unpin_fence() argument 401 i915_vma_unpin_fence(struct i915_vma * vma) i915_vma_unpin_fence() argument 407 i915_vma_fence_id(const struct i915_vma * vma) i915_vma_fence_id() argument 414 i915_vma_is_scanout(const struct i915_vma * vma) i915_vma_is_scanout() argument 419 i915_vma_mark_scanout(struct i915_vma * vma) i915_vma_mark_scanout() argument 424 i915_vma_clear_scanout(struct i915_vma * vma) i915_vma_clear_scanout() argument 452 i915_vma_sync(struct i915_vma * vma) i915_vma_sync() argument 468 i915_vma_get_current_resource(struct i915_vma * vma) i915_vma_get_current_resource() argument [all...] |
H A D | i915_vma.c | 48 static inline void assert_vma_held_evict(const struct i915_vma *vma) in assert_vma_held_evict() argument 66 static void i915_vma_free(struct i915_vma *vma) in i915_vma_free() argument 75 static void vma_print_allocator(struct i915_vma *vma, const char *reason) in vma_print_allocator() argument 94 vma_print_allocator(struct i915_vma * vma,const char * reason) vma_print_allocator() argument 107 struct i915_vma *vma = active_to_vma(ref); __i915_vma_active() local 132 struct i915_vma *vma = active_to_vma(ref); __i915_vma_retire() local 151 struct i915_vma *vma; vma_create() local 293 struct i915_vma *vma = rb_entry(rb, struct i915_vma, obj_node); i915_vma_lookup() local 327 struct i915_vma *vma; i915_vma_instance() local 405 i915_vma_wait_for_bind(struct i915_vma * vma) i915_vma_wait_for_bind() argument 425 i915_vma_verify_bind_complete(struct i915_vma * vma) i915_vma_verify_bind_complete() argument 448 i915_vma_resource_init_from_vma(struct i915_vma_resource * vma_res,struct i915_vma * vma) i915_vma_resource_init_from_vma() argument 472 i915_vma_bind(struct i915_vma * vma,unsigned int pat_index,u32 flags,struct i915_vma_work * work,struct i915_vma_resource * vma_res) i915_vma_bind() argument 581 i915_vma_pin_iomap(struct i915_vma * vma) i915_vma_pin_iomap() argument 649 i915_vma_flush_writes(struct i915_vma * vma) i915_vma_flush_writes() argument 655 i915_vma_unpin_iomap(struct i915_vma * vma) i915_vma_unpin_iomap() argument 669 struct i915_vma *vma; i915_vma_unpin_and_release() local 687 i915_vma_misplaced(const struct i915_vma * vma,u64 size,u64 alignment,u64 flags) i915_vma_misplaced() argument 721 __i915_vma_set_map_and_fenceable(struct i915_vma * vma) __i915_vma_set_map_and_fenceable() argument 740 i915_gem_valid_gtt_space(struct i915_vma * vma,unsigned long color) i915_gem_valid_gtt_space() argument 788 i915_vma_insert(struct i915_vma * vma,struct i915_gem_ww_ctx * ww,u64 size,u64 alignment,u64 flags) i915_vma_insert() argument 927 i915_vma_detach(struct i915_vma * vma) i915_vma_detach() argument 940 try_qad_pin(struct i915_vma * vma,unsigned int flags) try_qad_pin() argument 1300 __i915_vma_get_pages(struct i915_vma * vma) __i915_vma_get_pages() argument 1347 i915_vma_get_pages(struct i915_vma * vma) i915_vma_get_pages() argument 1394 __vma_put_pages(struct i915_vma * vma,unsigned int count) __vma_put_pages() argument 1410 i915_vma_put_pages(struct i915_vma * vma) i915_vma_put_pages() argument 1418 vma_unbind_pages(struct i915_vma * vma) vma_unbind_pages() argument 1432 i915_vma_pin_ww(struct i915_vma * vma,struct i915_gem_ww_ctx * ww,u64 size,u64 alignment,u64 flags) i915_vma_pin_ww() argument 1610 i915_vma_pin(struct i915_vma * vma,u64 size,u64 alignment,u64 flags) i915_vma_pin() argument 1641 __i915_ggtt_pin(struct i915_vma * vma,struct i915_gem_ww_ctx * ww,u32 align,unsigned int flags) __i915_ggtt_pin() argument 1676 i915_ggtt_pin(struct i915_vma * vma,struct i915_gem_ww_ctx * ww,u32 align,unsigned int flags) i915_ggtt_pin() argument 1708 struct i915_vma *vma; i915_ggtt_clear_scanout() local 1718 __vma_close(struct i915_vma * vma,struct intel_gt * gt) __vma_close() argument 1736 i915_vma_close(struct i915_vma * vma) i915_vma_close() argument 1753 __i915_vma_remove_closed(struct i915_vma * vma) __i915_vma_remove_closed() argument 1758 i915_vma_reopen(struct i915_vma * vma) i915_vma_reopen() argument 1768 force_unbind(struct i915_vma * vma) force_unbind() argument 1778 release_references(struct i915_vma * vma,struct intel_gt * gt,bool vm_ddestroy) release_references() argument 1830 i915_vma_destroy_locked(struct i915_vma * vma) i915_vma_destroy_locked() argument 1839 i915_vma_destroy(struct i915_vma * vma) i915_vma_destroy() argument 1858 struct i915_vma *vma, *next; i915_vma_parked() local 1901 __i915_vma_iounmap(struct i915_vma * vma) __i915_vma_iounmap() argument 1915 i915_vma_revoke_mmap(struct i915_vma * vma) i915_vma_revoke_mmap() argument 1939 __i915_request_await_bind(struct i915_request * rq,struct i915_vma * vma) __i915_request_await_bind() argument 1944 __i915_vma_move_to_active(struct i915_vma * vma,struct i915_request * rq) __i915_vma_move_to_active() argument 1956 _i915_vma_move_to_active(struct i915_vma * vma,struct i915_request * rq,struct dma_fence * fence,unsigned int flags) _i915_vma_move_to_active() argument 2031 __i915_vma_evict(struct i915_vma * vma,bool async) __i915_vma_evict() argument 2111 __i915_vma_unbind(struct i915_vma * vma) __i915_vma_unbind() argument 2142 __i915_vma_unbind_async(struct i915_vma * vma) __i915_vma_unbind_async() argument 2177 i915_vma_unbind(struct i915_vma * vma) i915_vma_unbind() argument 2215 i915_vma_unbind_async(struct i915_vma * vma,bool trylock_vm) i915_vma_unbind_async() argument 2277 i915_vma_unbind_unlocked(struct i915_vma * vma) i915_vma_unbind_unlocked() argument 2288 i915_vma_make_unshrinkable(struct i915_vma * vma) i915_vma_make_unshrinkable() argument 2294 i915_vma_make_shrinkable(struct i915_vma * vma) i915_vma_make_shrinkable() argument 2299 i915_vma_make_purgeable(struct i915_vma * vma) i915_vma_make_purgeable() argument [all...] |
H A D | i915_gem_evict.c | 67 static bool grab_vma(struct i915_vma *vma, struct i915_gem_ww_ctx *ww) in grab_vma() argument 86 static void ungrab_vma(struct i915_vma *vma) in ungrab_vma() argument 98 mark_free(struct drm_mm_scan * scan,struct i915_gem_ww_ctx * ww,struct i915_vma * vma,unsigned int flags,struct list_head * unwind) mark_free() argument 112 defer_evict(struct i915_vma * vma) defer_evict() argument 157 struct i915_vma *vma, *next; i915_gem_evict_something() local 337 struct i915_vma *vma, *next; i915_gem_evict_for_node() local 478 struct i915_vma *vma, *vn; i915_gem_evict_vm() local [all...] |
/linux/include/linux/ |
H A D | userfaultfd_k.h | 145 static inline bool is_mergeable_vm_userfaultfd_ctx(struct vm_area_struct *vma, in is_mergeable_vm_userfaultfd_ctx() argument 162 static inline bool uffd_disable_huge_pmd_share(struct vm_area_struct *vma) in uffd_disable_huge_pmd_share() argument 174 static inline bool uffd_disable_fault_around(struct vm_area_struct *vma) in uffd_disable_fault_around() argument 179 static inline bool userfaultfd_missing(struct vm_area_struct *vma) in userfaultfd_missing() argument 184 static inline bool userfaultfd_wp(struct vm_area_struct *vma) in userfaultfd_wp() argument 189 userfaultfd_minor(struct vm_area_struct * vma) userfaultfd_minor() argument 194 userfaultfd_pte_wp(struct vm_area_struct * vma,pte_t pte) userfaultfd_pte_wp() argument 200 userfaultfd_huge_pmd_wp(struct vm_area_struct * vma,pmd_t pmd) userfaultfd_huge_pmd_wp() argument 206 userfaultfd_armed(struct vm_area_struct * vma) userfaultfd_armed() argument 211 vma_can_userfault(struct vm_area_struct * vma,vm_flags_t vm_flags,bool wp_async) vma_can_userfault() argument 246 vma_has_uffd_without_event_remap(struct vm_area_struct * vma) vma_has_uffd_without_event_remap() argument 303 uffd_wp_range(struct vm_area_struct * vma,unsigned long start,unsigned long len,bool enable_wp) uffd_wp_range() argument 310 is_mergeable_vm_userfaultfd_ctx(struct vm_area_struct * vma,struct vm_userfaultfd_ctx vm_ctx) is_mergeable_vm_userfaultfd_ctx() argument 316 userfaultfd_missing(struct vm_area_struct * vma) userfaultfd_missing() argument 321 userfaultfd_wp(struct vm_area_struct * vma) userfaultfd_wp() argument 326 userfaultfd_minor(struct vm_area_struct * vma) userfaultfd_minor() argument 331 userfaultfd_pte_wp(struct vm_area_struct * vma,pte_t pte) userfaultfd_pte_wp() argument 337 userfaultfd_huge_pmd_wp(struct vm_area_struct * vma,pmd_t pmd) userfaultfd_huge_pmd_wp() argument 344 userfaultfd_armed(struct vm_area_struct * vma) userfaultfd_armed() argument 349 dup_userfaultfd(struct vm_area_struct * vma,struct list_head * l) dup_userfaultfd() argument 363 mremap_userfaultfd_prep(struct vm_area_struct * vma,struct vm_userfaultfd_ctx * ctx) mremap_userfaultfd_prep() argument 379 userfaultfd_remove(struct vm_area_struct * vma,unsigned long start,unsigned long end) userfaultfd_remove() argument 386 userfaultfd_unmap_prep(struct vm_area_struct * vma,unsigned long start,unsigned long end,struct list_head * uf) userfaultfd_unmap_prep() argument 398 uffd_disable_fault_around(struct vm_area_struct * vma) uffd_disable_fault_around() argument 403 userfaultfd_wp_unpopulated(struct vm_area_struct * vma) userfaultfd_wp_unpopulated() argument 408 userfaultfd_wp_async(struct vm_area_struct * vma) userfaultfd_wp_async() argument 413 vma_has_uffd_without_event_remap(struct vm_area_struct * vma) vma_has_uffd_without_event_remap() argument 420 userfaultfd_wp_use_markers(struct vm_area_struct * vma) userfaultfd_wp_use_markers() argument [all...] |
H A D | huge_mm.h | 101 #define thp_vma_allowable_order(vma, vm_flags, tva_flags, order) \ argument 217 thp_vma_suitable_order(struct vm_area_struct * vma,unsigned long addr,int order) thp_vma_suitable_order() argument 242 thp_vma_suitable_orders(struct vm_area_struct * vma,unsigned long addr,unsigned long orders) thp_vma_suitable_orders() argument 286 thp_vma_allowable_orders(struct vm_area_struct * vma,vm_flags_t vm_flags,unsigned long tva_flags,unsigned long orders) thp_vma_allowable_orders() argument 321 vma_thp_disabled(struct vm_area_struct * vma,vm_flags_t vm_flags) vma_thp_disabled() argument 422 change_huge_pud(struct mmu_gather * tlb,struct vm_area_struct * vma,pud_t * pudp,unsigned long addr,pgprot_t newprot,unsigned long cp_flags) change_huge_pud() argument 450 pmd_trans_huge_lock(pmd_t * pmd,struct vm_area_struct * vma) pmd_trans_huge_lock() argument 458 pud_trans_huge_lock(pud_t * pud,struct vm_area_struct * vma) pud_trans_huge_lock() argument 515 thp_vma_suitable_order(struct vm_area_struct * vma,unsigned long addr,int order) thp_vma_suitable_order() argument 521 thp_vma_suitable_orders(struct vm_area_struct * vma,unsigned long addr,unsigned long orders) thp_vma_suitable_orders() argument 527 thp_vma_allowable_orders(struct vm_area_struct * vma,vm_flags_t vm_flags,unsigned long tva_flags,unsigned long orders) thp_vma_allowable_orders() argument 578 __split_huge_pmd(struct vm_area_struct * vma,pmd_t * pmd,unsigned long address,bool freeze) __split_huge_pmd() argument 580 split_huge_pmd_address(struct vm_area_struct * vma,unsigned long address,bool freeze) split_huge_pmd_address() argument 582 split_huge_pmd_locked(struct vm_area_struct * vma,unsigned long address,pmd_t * pmd,bool freeze) split_huge_pmd_locked() argument 586 unmap_huge_pmd_locked(struct vm_area_struct * vma,unsigned long addr,pmd_t * pmdp,struct folio * folio) unmap_huge_pmd_locked() argument 596 hugepage_madvise(struct vm_area_struct * vma,vm_flags_t * vm_flags,int advice) hugepage_madvise() argument 602 madvise_collapse(struct vm_area_struct * vma,unsigned long start,unsigned long end,bool * lock_dropped) madvise_collapse() argument 609 vma_adjust_trans_huge(struct vm_area_struct * vma,unsigned long start,unsigned long end,struct vm_area_struct * next) vma_adjust_trans_huge() argument 620 pmd_trans_huge_lock(pmd_t * pmd,struct vm_area_struct * vma) pmd_trans_huge_lock() argument 625 pud_trans_huge_lock(pud_t * pud,struct vm_area_struct * vma) pud_trans_huge_lock() argument 655 follow_devmap_pmd(struct vm_area_struct * vma,unsigned long addr,pmd_t * pmd,int flags,struct dev_pagemap ** pgmap) follow_devmap_pmd() argument 676 __split_huge_pud(struct vm_area_struct * vma,pud_t * pud,unsigned long address) __split_huge_pud() argument 682 change_huge_pud(struct mmu_gather * tlb,struct vm_area_struct * vma,pud_t * pudp,unsigned long addr,pgprot_t newprot,unsigned long cp_flags) change_huge_pud() argument [all...] |
H A D | rmap.h | 84 struct vm_area_struct *vma; member 160 static inline int anon_vma_prepare(struct vm_area_struct *vma) in anon_vma_prepare() argument 168 static inline void anon_vma_merge(struct vm_area_struct *vma, in anon_vma_merge() argument 230 folio_set_large_mapcount(struct folio * folio,int mapcount,struct vm_area_struct * vma) folio_set_large_mapcount() argument 244 folio_add_return_large_mapcount(struct folio * folio,int diff,struct vm_area_struct * vma) folio_add_return_large_mapcount() argument 294 folio_sub_return_large_mapcount(struct folio * folio,int diff,struct vm_area_struct * vma) folio_sub_return_large_mapcount() argument 345 folio_set_large_mapcount(struct folio * folio,int mapcount,struct vm_area_struct * vma) folio_set_large_mapcount() argument 352 folio_add_large_mapcount(struct folio * folio,int diff,struct vm_area_struct * vma) folio_add_large_mapcount() argument 358 folio_add_return_large_mapcount(struct folio * folio,int diff,struct vm_area_struct * vma) folio_add_return_large_mapcount() argument 364 folio_sub_large_mapcount(struct folio * folio,int diff,struct vm_area_struct * vma) folio_sub_large_mapcount() argument 370 folio_sub_return_large_mapcount(struct folio * folio,int diff,struct vm_area_struct * vma) folio_sub_return_large_mapcount() argument 376 folio_inc_large_mapcount(folio,vma) global() argument 378 folio_inc_return_large_mapcount(folio,vma) global() argument 380 folio_dec_large_mapcount(folio,vma) global() argument 382 folio_dec_return_large_mapcount(folio,vma) global() argument 482 folio_add_anon_rmap_pte(folio,page,vma,address,flags) global() argument 490 folio_add_file_rmap_pte(folio,page,vma) global() argument 498 folio_remove_rmap_pte(folio,page,vma) global() argument 512 hugetlb_try_dup_anon_rmap(struct folio * folio,struct vm_area_struct * vma) hugetlb_try_dup_anon_rmap() argument 935 struct vm_area_struct *vma; global() member 1046 anon_vma_prepare(vma) global() argument [all...] |
H A D | hugetlb.h | 108 struct vm_area_struct *vma; member 254 hugetlb_zap_begin(struct vm_area_struct * vma,unsigned long * start,unsigned long * end) hugetlb_zap_begin() argument 261 hugetlb_zap_end(struct vm_area_struct * vma,struct zap_details * details) hugetlb_zap_end() argument 286 hugetlb_dup_vma_private(struct vm_area_struct * vma) hugetlb_dup_vma_private() argument 290 clear_vma_resv_huge_pages(struct vm_area_struct * vma) clear_vma_resv_huge_pages() argument 306 huge_pmd_unshare(struct mm_struct * mm,struct vm_area_struct * vma,unsigned long addr,pte_t * ptep) huge_pmd_unshare() argument 313 adjust_range_if_pmd_sharing_possible(struct vm_area_struct * vma,unsigned long * start,unsigned long * end) adjust_range_if_pmd_sharing_possible() argument 319 hugetlb_zap_begin(struct vm_area_struct * vma,unsigned long * start,unsigned long * end) hugetlb_zap_begin() argument 325 hugetlb_zap_end(struct vm_area_struct * vma,struct zap_details * details) hugetlb_zap_end() argument 339 move_hugetlb_page_tables(struct vm_area_struct * vma,struct vm_area_struct * new_vma,unsigned long old_addr,unsigned long new_addr,unsigned long len) move_hugetlb_page_tables() argument 362 hugetlb_vma_lock_read(struct vm_area_struct * vma) hugetlb_vma_lock_read() argument 366 hugetlb_vma_unlock_read(struct vm_area_struct * vma) hugetlb_vma_unlock_read() argument 370 hugetlb_vma_lock_write(struct vm_area_struct * vma) hugetlb_vma_lock_write() argument 374 hugetlb_vma_unlock_write(struct vm_area_struct * vma) hugetlb_vma_unlock_write() argument 378 hugetlb_vma_trylock_write(struct vm_area_struct * vma) hugetlb_vma_trylock_write() argument 383 hugetlb_vma_assert_locked(struct vm_area_struct * vma) hugetlb_vma_assert_locked() argument 438 hugetlb_change_protection(struct vm_area_struct * vma,unsigned long address,unsigned long end,pgprot_t newprot,unsigned long cp_flags) hugetlb_change_protection() argument 446 __unmap_hugepage_range(struct mmu_gather * tlb,struct vm_area_struct * vma,unsigned long start,unsigned long end,struct folio * folio,zap_flags_t zap_flags) __unmap_hugepage_range() argument 454 hugetlb_fault(struct mm_struct * mm,struct vm_area_struct * vma,unsigned long address,unsigned int flags) hugetlb_fault() argument 461 hugetlb_unshare_all_pmds(struct vm_area_struct * vma) hugetlb_unshare_all_pmds() argument 463 fixup_hugetlb_reservations(struct vm_area_struct * vma) fixup_hugetlb_reservations() argument 467 hugetlb_split(struct vm_area_struct * vma,unsigned long addr) hugetlb_split() argument 762 hstate_vma(struct vm_area_struct * vma) hstate_vma() argument 1034 huge_ptep_modify_prot_start(struct vm_area_struct * vma,unsigned long addr,pte_t * ptep) huge_ptep_modify_prot_start() argument 1045 huge_ptep_modify_prot_commit(struct vm_area_struct * vma,unsigned long addr,pte_t * ptep,pte_t old_pte,pte_t pte) huge_ptep_modify_prot_commit() argument 1337 flush_hugetlb_tlb_range(vma,addr,end) global() argument 1340 __vma_shareable_lock(struct vm_area_struct * vma) __vma_shareable_lock() argument 1352 hugetlb_walk(struct vm_area_struct * vma,unsigned long addr,unsigned long sz) hugetlb_walk() argument [all...] |
/linux/drivers/gpu/drm/xe/ |
H A D | xe_vm.h | 110 static inline u64 xe_vma_start(struct xe_vma *vma) in xe_vma_start() argument 115 static inline u64 xe_vma_size(struct xe_vma *vma) in xe_vma_size() argument 120 static inline u64 xe_vma_end(struct xe_vma *vma) in xe_vma_end() argument 125 static inline u64 xe_vma_bo_offset(struct xe_vma *vma) in xe_vma_bo_offset() argument 130 xe_vma_bo(struct xe_vma * vma) xe_vma_bo() argument 136 xe_vma_vm(struct xe_vma * vma) xe_vma_vm() argument 141 xe_vma_read_only(struct xe_vma * vma) xe_vma_read_only() argument 146 xe_vma_userptr(struct xe_vma * vma) xe_vma_userptr() argument 151 xe_vma_is_null(struct xe_vma * vma) xe_vma_is_null() argument 156 xe_vma_is_cpu_addr_mirror(struct xe_vma * vma) xe_vma_is_cpu_addr_mirror() argument 161 xe_vma_has_no_bo(struct xe_vma * vma) xe_vma_has_no_bo() argument 166 xe_vma_is_userptr(struct xe_vma * vma) xe_vma_is_userptr() argument 180 to_userptr_vma(struct xe_vma * vma) to_userptr_vma() argument [all...] |
H A D | xe_vm.c | 71 struct xe_vma *vma = &uvma->vma; in xe_vma_userptr_pin_pages() local 589 struct xe_vma *vma = &uvma->vma; in __vma_userptr_invalidate() local 636 struct xe_vma *vma = &uvma->vma; vma_userptr_invalidate() local 805 struct xe_vma *vma; xe_vma_svm_prefetch_op_fini() local 846 xe_vm_populate_rebind(struct xe_vma_op * op,struct xe_vma * vma,u8 tile_mask) xe_vm_populate_rebind() argument 862 xe_vm_ops_add_rebind(struct xe_vma_ops * vops,struct xe_vma * vma,u8 tile_mask) xe_vm_ops_add_rebind() argument 887 struct xe_vma *vma, *next; xe_vm_rebind() local 939 xe_vma_rebind(struct xe_vm * vm,struct xe_vma * vma,u8 tile_mask) xe_vma_rebind() argument 982 xe_vm_populate_range_rebind(struct xe_vma_op * op,struct xe_vma * vma,struct xe_svm_range * range,u8 tile_mask) xe_vm_populate_range_rebind() argument 996 xe_vm_ops_add_range_rebind(struct xe_vma_ops * vops,struct xe_vma * vma,struct xe_svm_range * range,u8 tile_mask) xe_vm_ops_add_range_rebind() argument 1026 xe_vm_range_rebind(struct xe_vm * vm,struct xe_vma * vma,struct xe_svm_range * range,u8 tile_mask) xe_vm_range_rebind() argument 1154 xe_vma_free(struct xe_vma * vma) xe_vma_free() argument 1173 struct xe_vma *vma; xe_vma_create() local 1273 xe_vma_destroy_late(struct xe_vma * vma) xe_vma_destroy_late() argument 1308 struct xe_vma *vma = vma_destroy_work_func() local 1317 struct xe_vma *vma = container_of(cb, struct xe_vma, destroy_cb); vma_destroy_cb() local 1323 xe_vma_destroy(struct xe_vma * vma,struct dma_fence * fence) xe_vma_destroy() argument 1367 xe_vm_lock_vma(struct drm_exec * exec,struct xe_vma * vma) xe_vm_lock_vma() argument 1382 xe_vma_destroy_unlocked(struct xe_vma * vma) xe_vma_destroy_unlocked() argument 1417 xe_vm_insert_vma(struct xe_vm * vm,struct xe_vma * vma) xe_vm_insert_vma() argument 1432 xe_vm_remove_vma(struct xe_vm * vm,struct xe_vma * vma) xe_vm_remove_vma() argument 1543 xelp_pte_encode_vma(u64 pte,struct xe_vma * vma,u16 pat_index,u32 pt_level) xelp_pte_encode_vma() argument 1870 struct xe_vma *vma, *next_vma; xe_vm_close_and_put() local 2172 vma_matches(struct xe_vma * vma,u64 page_addr) vma_matches() argument 2189 struct xe_vma *vma = NULL; xe_vm_find_vma_by_addr() local 2207 prep_vma_destroy(struct xe_vm * vm,struct xe_vma * vma,bool post_commit) prep_vma_destroy() argument 2223 struct xe_vma *vma; print_op() local 2369 struct xe_vma *vma = gpuva_to_vma(op->base.prefetch.va); vm_bind_ioctl_ops_create() local 2450 struct xe_vma *vma; new_vma() local 2497 xe_vma_max_pte_size(struct xe_vma * vma) xe_vma_max_pte_size() argument 2511 xe_vma_set_pte_size(struct xe_vma * vma,u64 size) xe_vma_set_pte_size() argument 2608 struct xe_vma *vma; vm_bind_ioctl_ops_parse() local 2791 struct xe_vma *vma = gpuva_to_vma(op->base.unmap.va); xe_vma_op_unwind() local 2804 struct xe_vma *vma = gpuva_to_vma(op->base.remap.unmap->va); xe_vma_op_unwind() local 2855 vma_lock_and_validate(struct drm_exec * exec,struct xe_vma * vma,bool validate) vma_lock_and_validate() argument 2873 check_ufence(struct xe_vma * vma) check_ufence() argument 2891 struct xe_vma *vma = gpuva_to_vma(op->base.prefetch.va); prefetch_ranges() local 2975 struct xe_vma *vma = gpuva_to_vma(op->base.prefetch.va); op_lock_and_prep() local 3192 vma_add_ufence(struct xe_vma * vma,struct xe_user_fence * ufence) vma_add_ufence() argument 3917 xe_vm_invalidate_vma(struct xe_vma * vma) xe_vm_invalidate_vma() argument 3981 struct xe_vma *vma = gpuva_to_vma(gpuva); xe_vm_validate_protected() local 4035 struct xe_vma *vma = gpuva_to_vma(gpuva); xe_vm_snapshot_capture() local [all...] |
/linux/mm/ |
H A D | vma.c | 75 static bool vma_had_uncowed_parents(struct vm_area_struct *vma) in vma_had_uncowed_parents() argument 86 struct vm_area_struct *vma = merge_next ? vmg->next : vmg->prev; in is_mergeable_vma() local 141 init_multi_vma_prep(struct vma_prepare * vp,struct vm_area_struct * vma,struct vma_merge_struct * vmg) init_multi_vma_prep() argument 225 __vma_link_file(struct vm_area_struct * vma,struct address_space * mapping) __vma_link_file() argument 239 __remove_shared_vm_struct(struct vm_area_struct * vma,struct address_space * mapping) __remove_shared_vm_struct() argument 265 anon_vma_interval_tree_pre_update_vma(struct vm_area_struct * vma) anon_vma_interval_tree_pre_update_vma() argument 274 anon_vma_interval_tree_post_update_vma(struct vm_area_struct * vma) anon_vma_interval_tree_post_update_vma() argument 411 init_vma_prep(struct vma_prepare * vp,struct vm_area_struct * vma) init_vma_prep() argument 461 remove_vma(struct vm_area_struct * vma) remove_vma() argument 476 unmap_region(struct ma_state * mas,struct vm_area_struct * vma,struct vm_area_struct * prev,struct vm_area_struct * next) unmap_region() argument 499 __split_vma(struct vma_iterator * vmi,struct vm_area_struct * vma,unsigned long addr,int new_below) __split_vma() argument 592 split_vma(struct vma_iterator * vmi,struct vm_area_struct * vma,unsigned long addr,int new_below) split_vma() argument 648 struct vm_area_struct *vma; validate_mm() local 732 struct vm_area_struct *vma; commit_merge() local 773 can_merge_remove_vma(struct vm_area_struct * vma) can_merge_remove_vma() argument 1185 vma_shrink(struct vma_iterator * vmi,struct vm_area_struct * vma,unsigned long start,unsigned long end,pgoff_t pgoff) vma_shrink() argument 1242 struct vm_area_struct *vma; vms_clean_up_area() local 1265 struct vm_area_struct *vma; vms_complete_munmap_vmas() local 1311 struct vm_area_struct *vma; reattach_vmas() local 1479 init_vma_munmap(struct vma_munmap_struct * vms,struct vma_iterator * vmi,struct vm_area_struct * vma,unsigned long start,unsigned long end,struct list_head * uf,bool unlock) init_vma_munmap() argument 1515 do_vmi_align_munmap(struct vma_iterator * vmi,struct vm_area_struct * vma,struct mm_struct * mm,unsigned long start,unsigned long end,struct list_head * uf,bool unlock) do_vmi_align_munmap() argument 1567 struct vm_area_struct *vma; do_vmi_munmap() local 1602 struct vm_area_struct *vma = vmg->middle; vma_modify() local 1640 vma_modify_flags(struct vma_iterator * vmi,struct vm_area_struct * prev,struct vm_area_struct * vma,unsigned long start,unsigned long end,vm_flags_t vm_flags) vma_modify_flags() argument 1655 vma_modify_name(struct vma_iterator * vmi,struct vm_area_struct * prev,struct vm_area_struct * vma,unsigned long start,unsigned long end,struct anon_vma_name * new_name) vma_modify_name() argument 1670 vma_modify_policy(struct vma_iterator * vmi,struct vm_area_struct * prev,struct vm_area_struct * vma,unsigned long start,unsigned long end,struct mempolicy * new_pol) vma_modify_policy() argument 1684 vma_modify_flags_uffd(struct vma_iterator * vmi,struct vm_area_struct * prev,struct vm_area_struct * vma,unsigned long start,unsigned long end,vm_flags_t vm_flags,struct vm_userfaultfd_ctx new_ctx,bool give_up_on_oom) vma_modify_flags_uffd() argument 1704 vma_merge_extend(struct vma_iterator * vmi,struct vm_area_struct * vma,unsigned long delta) vma_merge_extend() argument 1738 unlink_file_vma_batch_add(struct unlink_vma_file_batch * vb,struct vm_area_struct * vma) unlink_file_vma_batch_add() argument 1761 unlink_file_vma(struct vm_area_struct * vma) unlink_file_vma() argument 1774 vma_link_file(struct vm_area_struct * vma) vma_link_file() argument 1787 vma_link(struct mm_struct * mm,struct vm_area_struct * vma) vma_link() argument 1811 struct vm_area_struct *vma = *vmap; copy_vma() local 1967 find_mergeable_anon_vma(struct vm_area_struct * vma) find_mergeable_anon_vma() argument 2006 vma_is_shared_writable(struct vm_area_struct * vma) vma_is_shared_writable() argument 2012 vma_fs_can_writeback(struct vm_area_struct * vma) vma_fs_can_writeback() argument 2026 vma_needs_dirty_tracking(struct vm_area_struct * vma) vma_needs_dirty_tracking() argument 2049 vma_wants_writenotify(struct vm_area_struct * vma,pgprot_t vm_page_prot) vma_wants_writenotify() argument 2163 struct vm_area_struct *vma; mm_take_all_locks() local 2259 struct vm_area_struct *vma; mm_drop_all_locks() local 2403 __mmap_new_file_vma(struct mmap_state * map,struct vm_area_struct * vma) __mmap_new_file_vma() argument 2454 struct vm_area_struct *vma; __mmap_new_vma() local 2524 __mmap_complete(struct mmap_state * map,struct vm_area_struct * vma) __mmap_complete() argument 2601 set_vma_user_defined_fields(struct vm_area_struct * vma,struct mmap_state * map) set_vma_user_defined_fields() argument 2643 struct vm_area_struct *vma = NULL; __mmap_region() local 2761 do_brk_flags(struct vma_iterator * vmi,struct vm_area_struct * vma,unsigned long addr,unsigned long len,vm_flags_t vm_flags) do_brk_flags() argument 2945 acct_stack_growth(struct vm_area_struct * vma,unsigned long size,unsigned long grow) acct_stack_growth() argument 2984 expand_upwards(struct vm_area_struct * vma,unsigned long address) expand_upwards() argument 3070 expand_downwards(struct vm_area_struct * vma,unsigned long address) expand_downwards() argument 3167 insert_vm_struct(struct mm_struct * mm,struct vm_area_struct * vma) insert_vm_struct() argument [all...] |
H A D | mmap_lock.c | 48 static inline bool __vma_enter_locked(struct vm_area_struct *vma, bool detaching) in __vma_enter_locked() argument 72 static inline void __vma_exit_locked(struct vm_area_struct *vma, bool *detached) in __vma_exit_locked() argument 78 __vma_start_write(struct vm_area_struct * vma,unsigned int mm_lock_seq) __vma_start_write() argument 106 vma_mark_detached(struct vm_area_struct * vma) vma_mark_detached() argument 139 struct vm_area_struct *vma; lock_vma_under_rcu() local 185 struct vm_area_struct *vma; lock_next_vma_under_mmap_lock() local 210 struct vm_area_struct *vma; lock_next_vma() local 335 struct vm_area_struct *vma; lock_mm_and_find_vma() local 397 struct vm_area_struct *vma; lock_mm_and_find_vma() local [all...] |
H A D | memory.c | 373 struct vm_area_struct *vma, unsigned long floor, in free_pgtables() argument 501 print_bad_pte(struct vm_area_struct * vma,unsigned long addr,pte_t pte,struct page * page) print_bad_pte() argument 598 vm_normal_page(struct vm_area_struct * vma,unsigned long addr,pte_t pte) vm_normal_page() argument 654 vm_normal_folio(struct vm_area_struct * vma,unsigned long addr,pte_t pte) vm_normal_folio() argument 665 vm_normal_page_pmd(struct vm_area_struct * vma,unsigned long addr,pmd_t pmd) vm_normal_page_pmd() argument 702 vm_normal_folio_pmd(struct vm_area_struct * vma,unsigned long addr,pmd_t pmd) vm_normal_folio_pmd() argument 739 restore_exclusive_pte(struct vm_area_struct * vma,struct folio * folio,struct page * page,unsigned long address,pte_t * ptep,pte_t orig_pte) restore_exclusive_pte() argument 773 try_restore_exclusive_pte(struct vm_area_struct * vma,unsigned long addr,pte_t * ptep,pte_t orig_pte) try_restore_exclusive_pte() argument 1056 folio_prealloc(struct mm_struct * src_mm,struct vm_area_struct * vma,unsigned long addr,bool need_zero) folio_prealloc() argument 1462 zap_install_uffd_wp_if_needed(struct vm_area_struct * vma,unsigned long addr,pte_t * pte,int nr,struct zap_details * details,pte_t pteval) zap_install_uffd_wp_if_needed() argument 1490 zap_present_folio_ptes(struct mmu_gather * tlb,struct vm_area_struct * vma,struct folio * folio,struct page * page,pte_t * pte,pte_t ptent,unsigned int nr,unsigned long addr,struct zap_details * details,int * rss,bool * force_flush,bool * force_break,bool * any_skipped) zap_present_folio_ptes() argument 1541 zap_present_ptes(struct mmu_gather * tlb,struct vm_area_struct * vma,pte_t * pte,pte_t ptent,unsigned int max_nr,unsigned long addr,struct zap_details * details,int * rss,bool * force_flush,bool * force_break,bool * any_skipped) zap_present_ptes() argument 1587 zap_nonpresent_ptes(struct mmu_gather * tlb,struct vm_area_struct * vma,pte_t * pte,pte_t ptent,unsigned int max_nr,unsigned long addr,struct zap_details * details,int * rss,bool * any_skipped) zap_nonpresent_ptes() argument 1657 do_zap_pte_range(struct mmu_gather * tlb,struct vm_area_struct * vma,pte_t * pte,unsigned long addr,unsigned long end,struct zap_details * details,int * rss,bool * force_flush,bool * force_break,bool * any_skipped) do_zap_pte_range() argument 1693 zap_pte_range(struct mmu_gather * tlb,struct vm_area_struct * vma,pmd_t * pmd,unsigned long addr,unsigned long end,struct zap_details * details) zap_pte_range() argument 1785 zap_pmd_range(struct mmu_gather * tlb,struct vm_area_struct * vma,pud_t * pud,unsigned long addr,unsigned long end,struct zap_details * details) zap_pmd_range() argument 1827 zap_pud_range(struct mmu_gather * tlb,struct vm_area_struct * vma,p4d_t * p4d,unsigned long addr,unsigned long end,struct zap_details * details) zap_pud_range() argument 1856 zap_p4d_range(struct mmu_gather * tlb,struct vm_area_struct * vma,pgd_t * pgd,unsigned long addr,unsigned long end,struct zap_details * details) zap_p4d_range() argument 1875 unmap_page_range(struct mmu_gather * tlb,struct vm_area_struct * vma,unsigned long addr,unsigned long end,struct zap_details * details) unmap_page_range() argument 1896 unmap_single_vma(struct mmu_gather * tlb,struct vm_area_struct * vma,unsigned long start_addr,unsigned long end_addr,struct zap_details * details,bool mm_wr_locked) unmap_single_vma() argument 1958 unmap_vmas(struct mmu_gather * tlb,struct ma_state * mas,struct vm_area_struct * vma,unsigned long start_addr,unsigned long end_addr,unsigned long tree_end,bool mm_wr_locked) unmap_vmas() argument 1996 zap_page_range_single_batched(struct mmu_gather * tlb,struct vm_area_struct * vma,unsigned long address,unsigned long size,struct zap_details * details) zap_page_range_single_batched() argument 2035 zap_page_range_single(struct vm_area_struct * vma,unsigned long address,unsigned long size,struct zap_details * details) zap_page_range_single() argument 2056 zap_vma_ptes(struct vm_area_struct * vma,unsigned long address,unsigned long size) zap_vma_ptes() argument 2099 vm_mixed_zeropage_allowed(struct vm_area_struct * vma) vm_mixed_zeropage_allowed() argument 2129 validate_page_before_insert(struct vm_area_struct * vma,struct page * page) validate_page_before_insert() argument 2148 insert_page_into_pte_locked(struct vm_area_struct * vma,pte_t * pte,unsigned long addr,struct page * page,pgprot_t prot,bool mkwrite) insert_page_into_pte_locked() argument 2189 insert_page(struct vm_area_struct * vma,unsigned long addr,struct page * page,pgprot_t prot,bool mkwrite) insert_page() argument 2210 insert_page_in_batch_locked(struct vm_area_struct * vma,pte_t * pte,unsigned long addr,struct page * page,pgprot_t prot) insert_page_in_batch_locked() argument 2224 insert_pages(struct vm_area_struct * vma,unsigned long addr,struct page ** pages,unsigned long * num,pgprot_t prot) insert_pages() argument 2297 vm_insert_pages(struct vm_area_struct * vma,unsigned long addr,struct page ** pages,unsigned long * num) vm_insert_pages() argument 2344 vm_insert_page(struct vm_area_struct * vma,unsigned long addr,struct page * page) vm_insert_page() argument 2371 __vm_map_pages(struct vm_area_struct * vma,struct page ** pages,unsigned long num,unsigned long offset) __vm_map_pages() argument 2414 vm_map_pages(struct vm_area_struct * vma,struct page ** pages,unsigned long num) vm_map_pages() argument 2434 vm_map_pages_zero(struct vm_area_struct * vma,struct page ** pages,unsigned long num) vm_map_pages_zero() argument 2441 insert_pfn(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,pgprot_t prot,bool mkwrite) insert_pfn() argument 2525 vmf_insert_pfn_prot(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,pgprot_t pgprot) vmf_insert_pfn_prot() argument 2572 vmf_insert_pfn(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn) vmf_insert_pfn() argument 2579 vm_mixed_ok(struct vm_area_struct * vma,unsigned long pfn,bool mkwrite) vm_mixed_ok() argument 2593 __vm_insert_mixed(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,bool mkwrite) __vm_insert_mixed() argument 2659 vmf_insert_mixed(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn) vmf_insert_mixed() argument 2671 vmf_insert_mixed_mkwrite(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn) vmf_insert_mixed_mkwrite() argument 2775 remap_pfn_range_internal(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,unsigned long size,pgprot_t prot) remap_pfn_range_internal() argument 2832 remap_pfn_range_notrack(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,unsigned long size,pgprot_t prot) remap_pfn_range_notrack() argument 2892 remap_pfn_range(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,unsigned long size,pgprot_t prot) remap_pfn_range() argument 2930 remap_pfn_range(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,unsigned long size,pgprot_t prot) remap_pfn_range() argument 2953 vm_iomap_memory(struct vm_area_struct * vma,phys_addr_t start,unsigned long len) vm_iomap_memory() argument 3238 struct vm_area_struct *vma = vmf->vma; __wp_page_copy_user() local 3330 __get_fault_gfp_mask(struct vm_area_struct * vma) __get_fault_gfp_mask() argument 3385 struct vm_area_struct *vma = vmf->vma; fault_dirty_shared_page() local 3439 struct vm_area_struct *vma = vmf->vma; wp_page_reuse() local 3472 struct vm_area_struct *vma = vmf->vma; vmf_can_call_fault() local 3497 struct vm_area_struct *vma = vmf->vma; __vmf_anon_prepare() local 3533 struct vm_area_struct *vma = vmf->vma; wp_page_copy() local 3724 struct vm_area_struct *vma = vmf->vma; wp_pfn_shared() local 3747 struct vm_area_struct *vma = vmf->vma; wp_page_shared() local 3786 __wp_can_reuse_large_anon_folio(struct folio * folio,struct vm_area_struct * vma) __wp_can_reuse_large_anon_folio() argument 3852 __wp_can_reuse_large_anon_folio(struct folio * folio,struct vm_area_struct * vma) __wp_can_reuse_large_anon_folio() argument 3859 wp_can_reuse_anon_folio(struct folio * folio,struct vm_area_struct * vma) wp_can_reuse_anon_folio() argument 3925 struct vm_area_struct *vma = vmf->vma; do_wp_page() local 4016 unmap_mapping_range_vma(struct vm_area_struct * vma,unsigned long start_addr,unsigned long end_addr,struct zap_details * details) unmap_mapping_range_vma() argument 4028 struct vm_area_struct *vma; unmap_mapping_range_tree() local 4150 struct vm_area_struct *vma = vmf->vma; remove_device_exclusive_entry() local 4191 should_try_to_free_swap(struct folio * folio,struct vm_area_struct * vma,unsigned int fault_flags) should_try_to_free_swap() argument 4282 struct vm_area_struct *vma = vmf->vma; __alloc_swap_folio() local 4360 struct vm_area_struct *vma = vmf->vma; alloc_swap_folio() local 4456 struct vm_area_struct *vma = vmf->vma; do_swap_page() local 4917 struct vm_area_struct *vma = vmf->vma; alloc_anon_folio() local 5007 struct vm_area_struct *vma = vmf->vma; do_anonymous_page() local 5127 struct vm_area_struct *vma = vmf->vma; __do_fault() local 5184 struct vm_area_struct *vma = vmf->vma; deposit_prealloc_pte() local 5197 struct vm_area_struct *vma = vmf->vma; do_set_pmd() local 5286 struct vm_area_struct *vma = vmf->vma; set_pte_range() local 5344 struct vm_area_struct *vma = vmf->vma; finish_fault() local 5587 struct vm_area_struct *vma = vmf->vma; do_cow_fault() local 5629 struct vm_area_struct *vma = vmf->vma; do_shared_fault() local 5679 struct vm_area_struct *vma = vmf->vma; do_fault() local 5725 struct vm_area_struct *vma = vmf->vma; numa_migrate_check() local 5768 numa_rebuild_single_mapping(struct vm_fault * vmf,struct vm_area_struct * vma,unsigned long fault_addr,pte_t * fault_pte,bool writable) numa_rebuild_single_mapping() argument 5783 numa_rebuild_large_mapping(struct vm_fault * vmf,struct vm_area_struct * vma,struct folio * folio,pte_t fault_pte,bool ignore_writable,bool pte_write_upgrade) numa_rebuild_large_mapping() argument 5824 struct vm_area_struct *vma = vmf->vma; do_numa_page() local 5915 struct vm_area_struct *vma = vmf->vma; create_huge_pmd() local 5926 struct vm_area_struct *vma = vmf->vma; wp_huge_pmd() local 5959 struct vm_area_struct *vma = vmf->vma; create_huge_pud() local 5973 struct vm_area_struct *vma = vmf->vma; wp_huge_pud() local 6102 __handle_mm_fault(struct vm_area_struct * vma,unsigned long address,unsigned int flags) __handle_mm_fault() argument 6267 lru_gen_enter_fault(struct vm_area_struct * vma) lru_gen_enter_fault() argument 6278 lru_gen_enter_fault(struct vm_area_struct * vma) lru_gen_enter_fault() argument 6287 sanitize_fault_flags(struct vm_area_struct * vma,unsigned int * flags) sanitize_fault_flags() argument 6329 handle_mm_fault(struct vm_area_struct * vma,unsigned long address,unsigned int flags,struct pt_regs * regs) handle_mm_fault() argument 6482 pfnmap_lockdep_assert(struct vm_area_struct * vma) pfnmap_lockdep_assert() argument 6529 struct vm_area_struct *vma = args->vma; follow_pfnmap_start() local 6632 generic_access_phys(struct vm_area_struct * vma,unsigned long addr,void * buf,int len,int write) generic_access_phys() argument 6707 struct vm_area_struct *vma = NULL; __access_remote_vm() local 6834 struct vm_area_struct *vma = NULL; __copy_remote_vm_str() local 6931 struct vm_area_struct *vma; print_vma_addr() local 7062 copy_user_gigantic_page(struct folio * dst,struct folio * src,unsigned long addr_hint,struct vm_area_struct * vma,unsigned int nr_pages) copy_user_gigantic_page() argument 7085 struct vm_area_struct *vma; global() member 7100 copy_user_large_folio(struct folio * dst,struct folio * src,unsigned long addr_hint,struct vm_area_struct * vma) copy_user_large_folio() argument 7175 vma_pgtable_walk_begin(struct vm_area_struct * vma) vma_pgtable_walk_begin() argument 7181 vma_pgtable_walk_end(struct vm_area_struct * vma) vma_pgtable_walk_end() argument [all...] |
H A D | mprotect.c | 43 static bool maybe_change_pte_writable(struct vm_area_struct *vma, pte_t pte) in maybe_change_pte_writable() argument 63 static bool can_change_private_pte_writable(struct vm_area_struct *vma, in can_change_private_pte_writable() argument 81 static bool can_change_shared_pte_writable(struct vm_area_struct *vma, in can_change_shared_pte_writable() argument 99 bool can_change_pte_writable(struct vm_area_struct *vma, unsigne argument 121 prot_numa_skip(struct vm_area_struct * vma,unsigned long addr,pte_t oldpte,pte_t * pte,int target_node,struct folio * folio) prot_numa_skip() argument 178 prot_commit_flush_ptes(struct vm_area_struct * vma,unsigned long addr,pte_t * ptep,pte_t oldpte,pte_t ptent,int nr_ptes,int idx,bool set_write,struct mmu_gather * tlb) prot_commit_flush_ptes() argument 230 commit_anon_folio_batch(struct vm_area_struct * vma,struct folio * folio,struct page * first_page,unsigned long addr,pte_t * ptep,pte_t oldpte,pte_t ptent,int nr_ptes,struct mmu_gather * tlb) commit_anon_folio_batch() argument 249 set_write_prot_commit_flush_ptes(struct vm_area_struct * vma,struct folio * folio,struct page * page,unsigned long addr,pte_t * ptep,pte_t oldpte,pte_t ptent,int nr_ptes,struct mmu_gather * tlb) set_write_prot_commit_flush_ptes() argument 273 change_pte_range(struct mmu_gather * tlb,struct vm_area_struct * vma,pmd_t * pmd,unsigned long addr,unsigned long end,pgprot_t newprot,unsigned long cp_flags) change_pte_range() argument 453 pgtable_split_needed(struct vm_area_struct * vma,unsigned long cp_flags) pgtable_split_needed() argument 469 pgtable_populate_needed(struct vm_area_struct * vma,unsigned long cp_flags) pgtable_populate_needed() argument 485 change_pmd_prepare(vma,pmd,cp_flags) global() argument 500 change_prepare(vma,high,low,addr,cp_flags) global() argument 512 change_pmd_range(struct mmu_gather * tlb,struct vm_area_struct * vma,pud_t * pud,unsigned long addr,unsigned long end,pgprot_t newprot,unsigned long cp_flags) change_pmd_range() argument 582 change_pud_range(struct mmu_gather * tlb,struct vm_area_struct * vma,p4d_t * p4d,unsigned long addr,unsigned long end,pgprot_t newprot,unsigned long cp_flags) change_pud_range() argument 641 change_p4d_range(struct mmu_gather * tlb,struct vm_area_struct * vma,pgd_t * pgd,unsigned long addr,unsigned long end,pgprot_t newprot,unsigned long cp_flags) change_p4d_range() argument 664 change_protection_range(struct mmu_gather * tlb,struct vm_area_struct * vma,unsigned long addr,unsigned long end,pgprot_t newprot,unsigned long cp_flags) change_protection_range() argument 694 change_protection(struct mmu_gather * tlb,struct vm_area_struct * vma,unsigned long start,unsigned long end,unsigned long cp_flags) change_protection() argument 756 mprotect_fixup(struct vma_iterator * vmi,struct mmu_gather * tlb,struct vm_area_struct * vma,struct vm_area_struct ** pprev,unsigned long start,unsigned long end,vm_flags_t newflags) mprotect_fixup() argument 865 struct vm_area_struct *vma, *prev; do_mprotect_pkey() local [all...] |
H A D | nommu.c | 92 struct vm_area_struct *vma; in kobjsize() local 147 struct vm_area_struct *vma; in __vmalloc_user_flags() local 344 int vm_insert_page(struct vm_area_struct *vma, unsigne argument 351 vm_insert_pages(struct vm_area_struct * vma,unsigned long addr,struct page ** pages,unsigned long * num) vm_insert_pages() argument 358 vm_map_pages(struct vm_area_struct * vma,struct page ** pages,unsigned long num) vm_map_pages() argument 365 vm_map_pages_zero(struct vm_area_struct * vma,struct page ** pages,unsigned long num) vm_map_pages_zero() argument 561 setup_vma_to_mm(struct vm_area_struct * vma,struct mm_struct * mm) setup_vma_to_mm() argument 577 cleanup_vma_from_mm(struct vm_area_struct * vma) cleanup_vma_from_mm() argument 596 delete_vma_from_mm(struct vm_area_struct * vma) delete_vma_from_mm() argument 615 delete_vma(struct mm_struct * mm,struct vm_area_struct * vma) delete_vma() argument 651 expand_stack_locked(struct vm_area_struct * vma,unsigned long addr) expand_stack_locked() argument 670 struct vm_area_struct *vma; find_vma_exact() local 892 do_mmap_shared_file(struct vm_area_struct * vma) do_mmap_shared_file() argument 913 do_mmap_private(struct vm_area_struct * vma,struct vm_region * region,unsigned long len,unsigned long capabilities) do_mmap_private() argument 1021 struct vm_area_struct *vma; do_mmap() local 1305 split_vma(struct vma_iterator * vmi,struct vm_area_struct * vma,unsigned long addr,int new_below) split_vma() argument 1384 vmi_shrink_vma(struct vma_iterator * vmi,struct vm_area_struct * vma,unsigned long from,unsigned long to) vmi_shrink_vma() argument 1428 struct vm_area_struct *vma; do_munmap() local 1510 struct vm_area_struct *vma; exit_mmap() local 1545 struct vm_area_struct *vma; do_mremap() local 1589 remap_pfn_range(struct vm_area_struct * vma,unsigned long addr,unsigned long pfn,unsigned long size,pgprot_t prot) remap_pfn_range() argument 1600 vm_iomap_memory(struct vm_area_struct * vma,phys_addr_t start,unsigned long len) vm_iomap_memory() argument 1610 remap_vmalloc_range(struct vm_area_struct * vma,void * addr,unsigned long pgoff) remap_vmalloc_range() argument 1643 struct vm_area_struct *vma; __access_remote_vm() local 1722 struct vm_area_struct *vma; __copy_remote_vm_str() local 1806 struct vm_area_struct *vma; nommu_shrink_inode_mappings() local [all...] |
H A D | pgtable-generic.c | 68 int ptep_set_access_flags(struct vm_area_struct *vma, in ptep_set_access_flags() argument 82 int ptep_clear_flush_young(struct vm_area_struct *vma, in ptep_clear_flush_young() argument 94 pte_t ptep_clear_flush(struct vm_area_struct *vma, unsigned long address, in ptep_clear_flush() argument 109 int pmdp_set_access_flags(struct vm_area_struct *vma, in pmdp_set_access_flags() argument 124 pmdp_clear_flush_young(struct vm_area_struct * vma,unsigned long address,pmd_t * pmdp) pmdp_clear_flush_young() argument 137 pmdp_huge_clear_flush(struct vm_area_struct * vma,unsigned long address,pmd_t * pmdp) pmdp_huge_clear_flush() argument 149 pudp_huge_clear_flush(struct vm_area_struct * vma,unsigned long address,pud_t * pudp) pudp_huge_clear_flush() argument 197 pmdp_invalidate(struct vm_area_struct * vma,unsigned long address,pmd_t * pmdp) pmdp_invalidate() argument 208 pmdp_invalidate_ad(struct vm_area_struct * vma,unsigned long address,pmd_t * pmdp) pmdp_invalidate_ad() argument 217 pmdp_collapse_flush(struct vm_area_struct * vma,unsigned long address,pmd_t * pmdp) pmdp_collapse_flush() argument [all...] |
H A D | madvise.c | 80 struct vm_area_struct *vma; member 110 struct anon_vma_name *anon_vma_name(struct vm_area_struct *vma) in anon_vma_name() argument 119 static int replace_anon_vma_name(struct vm_area_struct *vma, in replace_anon_vma_name() argument 139 static int replace_anon_vma_name(struct vm_area_struct *vma, in replace_anon_vma_name() argument 155 struct vm_area_struct *vma = madv_behavior->vma; madvise_update_vma() local 190 struct vm_area_struct *vma = walk->private; swapin_walk_pmd_entry() local 236 shmem_swapin_range(struct vm_area_struct * vma,unsigned long start,unsigned long end,struct address_space * mapping) shmem_swapin_range() argument 285 struct vm_area_struct *vma = madv_behavior->vma; madvise_willneed() local 331 can_do_file_pageout(struct vm_area_struct * vma) can_do_file_pageout() argument 364 struct vm_area_struct *vma = walk->vma; madvise_cold_or_pageout_pte_range() local 580 struct vm_area_struct *vma = madv_behavior->vma; madvise_cold_page_range() local 593 can_madv_lru_vma(struct vm_area_struct * vma) can_madv_lru_vma() argument 600 struct vm_area_struct *vma = madv_behavior->vma; madvise_cold() local 615 madvise_pageout_page_range(struct mmu_gather * tlb,struct vm_area_struct * vma,struct madvise_behavior_range * range) madvise_pageout_page_range() argument 632 struct vm_area_struct *vma = madv_behavior->vma; madvise_pageout() local 662 struct vm_area_struct *vma = walk->vma; madvise_free_pte_range() local 806 struct vm_area_struct *vma = madv_behavior->vma; madvise_free_single_vma() local 878 struct vm_area_struct *vma = madv_behavior->vma; madvise_dontneed_free_valid_vma() local 920 struct vm_area_struct *vma; madvise_dontneed_free() local 1017 struct vm_area_struct *vma = madv_behavior->vma; madvise_remove() local 1057 is_valid_guard_vma(struct vm_area_struct * vma,bool allow_locked) is_valid_guard_vma() argument 1135 struct vm_area_struct *vma = madv_behavior->vma; madvise_guard_install() local 1246 struct vm_area_struct *vma = madv_behavior->vma; madvise_guard_remove() local 1287 struct vm_area_struct *vma = madv_behavior->vma; can_madvise_modify() local 1336 struct vm_area_struct *vma = madv_behavior->vma; madvise_vma_behavior() local 1586 struct vm_area_struct *vma; try_vma_read_lock() local 1627 struct vm_area_struct *prev, *vma; madvise_walk_vmas() local [all...] |
H A D | mmap.c | 81 void vma_set_page_prot(struct vm_area_struct *vma) in vma_set_page_prot() argument 693 struct vm_area_struct *vma, *pre in generic_get_unmapped_area() local 741 struct vm_area_struct *vma, *prev; generic_get_unmapped_area_topdown() local 930 struct vm_area_struct *vma; find_vma_prev() local 957 expand_stack_locked(struct vm_area_struct * vma,unsigned long address) expand_stack_locked() argument 964 struct vm_area_struct *vma, *prev; find_extend_vma_locked() local 979 expand_stack_locked(struct vm_area_struct * vma,unsigned long address) expand_stack_locked() argument 986 struct vm_area_struct *vma; find_extend_vma_locked() local 1006 vma_expand_up(vma,addr) global() argument 1007 vma_expand_down(vma,addr) global() argument 1011 vma_expand_up(vma,addr) global() argument 1012 vma_expand_down(vma,addr) global() argument 1029 struct vm_area_struct *vma, *prev; expand_stack() local 1092 struct vm_area_struct *vma; SYSCALL_DEFINE5() local 1209 struct vm_area_struct *vma = NULL; vm_brk_flags() local 1257 struct vm_area_struct *vma; exit_mmap() local 1368 special_mapping_close(struct vm_area_struct * vma) special_mapping_close() argument 1376 special_mapping_name(struct vm_area_struct * vma) special_mapping_name() argument 1394 special_mapping_split(struct vm_area_struct * vma,unsigned long addr) special_mapping_split() argument 1417 struct vm_area_struct *vma = vmf->vma; special_mapping_fault() local 1447 struct vm_area_struct *vma; __install_special_mapping() local 1476 vma_is_special_mapping(const struct vm_area_struct * vma,const struct vm_special_mapping * sm) vma_is_special_mapping() argument [all...] |
H A D | huge_memory.c | 85 static inline bool file_thp_enabled(struct vm_area_struct *vma) in file_thp_enabled() argument 100 unsigned long __thp_vma_allowable_orders(struct vm_area_struct *vma, in __thp_vma_allowable_orders() argument 1052 maybe_pmd_mkwrite(pmd_t pmd,struct vm_area_struct * vma) maybe_pmd_mkwrite() argument 1157 vma_alloc_anon_folio_pmd(struct vm_area_struct * vma,unsigned long addr) vma_alloc_anon_folio_pmd() argument 1201 map_anon_folio_pmd(struct folio * folio,pmd_t * pmd,struct vm_area_struct * vma,unsigned long haddr) map_anon_folio_pmd() argument 1220 struct vm_area_struct *vma = vmf->vma; __do_huge_pmd_anonymous_page() local 1279 vma_thp_gfp_mask(struct vm_area_struct * vma) vma_thp_gfp_mask() argument 1307 set_huge_zero_folio(pgtable_t pgtable,struct mm_struct * mm,struct vm_area_struct * vma,unsigned long haddr,pmd_t * pmd,struct folio * zero_folio) set_huge_zero_folio() argument 1319 struct vm_area_struct *vma = vmf->vma; do_huge_pmd_anonymous_page() local 1382 insert_pmd(struct vm_area_struct * vma,unsigned long addr,pmd_t * pmd,struct folio_or_pfn fop,pgprot_t prot,bool write,pgtable_t pgtable) insert_pmd() argument 1448 struct vm_area_struct *vma = vmf->vma; vmf_insert_pfn_pmd() local 1492 struct vm_area_struct *vma = vmf->vma; vmf_insert_folio_pmd() local 1527 maybe_pud_mkwrite(pud_t pud,struct vm_area_struct * vma) maybe_pud_mkwrite() argument 1534 insert_pud(struct vm_area_struct * vma,unsigned long addr,pud_t * pud,struct folio_or_pfn fop,pgprot_t prot,bool write) insert_pud() argument 1587 struct vm_area_struct *vma = vmf->vma; vmf_insert_pfn_pud() local 1628 struct vm_area_struct *vma = vmf->vma; vmf_insert_folio_pud() local 1653 touch_pmd(struct vm_area_struct * vma,unsigned long addr,pmd_t * pmd,bool write) touch_pmd() argument 1790 touch_pud(struct vm_area_struct * vma,unsigned long addr,pud_t * pud,bool write) touch_pud() argument 1805 copy_huge_pud(struct mm_struct * dst_mm,struct mm_struct * src_mm,pud_t * dst_pud,pud_t * src_pud,unsigned long addr,struct vm_area_struct * vma) copy_huge_pud() argument 1869 struct vm_area_struct *vma = vmf->vma; do_huge_zero_wp_pmd() local 1901 struct vm_area_struct *vma = vmf->vma; do_huge_pmd_wp_page() local 1992 can_change_pmd_writable(struct vm_area_struct * vma,unsigned long addr,pmd_t pmd) can_change_pmd_writable() argument 2025 struct vm_area_struct *vma = vmf->vma; do_huge_pmd_numa_page() local 2103 madvise_free_huge_pmd(struct mmu_gather * tlb,struct vm_area_struct * vma,pmd_t * pmd,unsigned long addr,unsigned long next) madvise_free_huge_pmd() argument 2182 zap_huge_pmd(struct mmu_gather * tlb,struct vm_area_struct * vma,pmd_t * pmd,unsigned long addr) zap_huge_pmd() argument 2260 pmd_move_must_withdraw(spinlock_t * new_pmd_ptl,spinlock_t * old_pmd_ptl,struct vm_area_struct * vma) pmd_move_must_withdraw() argument 2293 move_huge_pmd(struct vm_area_struct * vma,unsigned long old_addr,unsigned long new_addr,pmd_t * old_pmd,pmd_t * new_pmd) move_huge_pmd() argument 2351 change_huge_pmd(struct mmu_gather * tlb,struct vm_area_struct * vma,pmd_t * pmd,unsigned long addr,pgprot_t newprot,unsigned long cp_flags) change_huge_pmd() argument 2490 change_huge_pud(struct mmu_gather * tlb,struct vm_area_struct * vma,pud_t * pudp,unsigned long addr,pgprot_t newprot,unsigned long cp_flags) change_huge_pud() argument 2675 __pmd_trans_huge_lock(pmd_t * pmd,struct vm_area_struct * vma) __pmd_trans_huge_lock() argument 2691 __pud_trans_huge_lock(pud_t * pud,struct vm_area_struct * vma) __pud_trans_huge_lock() argument 2703 zap_huge_pud(struct mmu_gather * tlb,struct vm_area_struct * vma,pud_t * pud,unsigned long addr) zap_huge_pud() argument 2738 __split_huge_pud_locked(struct vm_area_struct * vma,pud_t * pud,unsigned long haddr) __split_huge_pud_locked() argument 2770 __split_huge_pud(struct vm_area_struct * vma,pud_t * pud,unsigned long address) __split_huge_pud() argument 2790 __split_huge_pud(struct vm_area_struct * vma,pud_t * pud,unsigned long address) __split_huge_pud() argument 2796 __split_huge_zero_page_pmd(struct vm_area_struct * vma,unsigned long haddr,pmd_t * pmd) __split_huge_zero_page_pmd() argument 2837 __split_huge_pmd_locked(struct vm_area_struct * vma,pmd_t * pmd,unsigned long haddr,bool freeze) __split_huge_pmd_locked() argument 3058 split_huge_pmd_locked(struct vm_area_struct * vma,unsigned long address,pmd_t * pmd,bool freeze) split_huge_pmd_locked() argument 3066 __split_huge_pmd(struct vm_area_struct * vma,pmd_t * pmd,unsigned long address,bool freeze) __split_huge_pmd() argument 3082 split_huge_pmd_address(struct vm_area_struct * vma,unsigned long address,bool freeze) split_huge_pmd_address() argument 3093 split_huge_pmd_if_needed(struct vm_area_struct * vma,unsigned long address) split_huge_pmd_if_needed() argument 3105 vma_adjust_trans_huge(struct vm_area_struct * vma,unsigned long start,unsigned long end,struct vm_area_struct * next) vma_adjust_trans_huge() argument 3144 __discard_anon_folio_pmd_locked(struct vm_area_struct * vma,unsigned long addr,pmd_t * pmdp,struct folio * folio) __discard_anon_folio_pmd_locked() argument 3207 unmap_huge_pmd_locked(struct vm_area_struct * vma,unsigned long addr,pmd_t * pmdp,struct folio * folio) unmap_huge_pmd_locked() argument 4296 vma_not_suitable_for_thp_split(struct vm_area_struct * vma) vma_not_suitable_for_thp_split() argument 4339 struct vm_area_struct *vma = vma_lookup(mm, addr); split_huge_pages_pid() local 4594 struct vm_area_struct *vma = pvmw->vma; set_pmd_migration_entry() local 4643 struct vm_area_struct *vma = pvmw->vma; remove_migration_pmd() local [all...] |
/linux/tools/testing/vma/ |
H A D | vma.c | 18 #define vma_iter_prealloc(vmi, vma) \ argument 85 static int attach_vma(struct mm_struct *mm, struct vm_area_struct *vma) in attach_vma() argument 95 static void detach_free_vma(struct vm_area_struct *vma) in detach_free_vma() argument 108 struct vm_area_struct *vma = alloc_vma(mm, start, end, pgoff, vm_flags); alloc_and_link_vma() local 131 struct vm_area_struct *vma; merge_new() local 153 struct vm_area_struct *vma; merge_existing() local 251 struct vm_area_struct *vma; cleanup_mm() local 269 vma_write_started(struct vm_area_struct * vma) vma_write_started() argument 285 __vma_set_dummy_anon_vma(struct vm_area_struct * vma,struct anon_vma_chain * avc,struct anon_vma * anon_vma) __vma_set_dummy_anon_vma() argument 295 vma_set_dummy_anon_vma(struct vm_area_struct * vma,struct anon_vma_chain * avc) vma_set_dummy_anon_vma() argument 303 struct vm_area_struct *vma; test_simple_merge() local 337 struct vm_area_struct *vma; test_simple_modify() local 399 struct vm_area_struct *vma = alloc_vma(&mm, 0, 0x1000, 0, vm_flags); test_simple_expand() local 427 struct vm_area_struct *vma = alloc_vma(&mm, 0, 0x3000, 0, vm_flags); test_simple_shrink() local 469 struct vm_area_struct *vma, *vma_a, *vma_b, *vma_c, *vma_d; test_merge_new() local 652 struct vm_area_struct *vma_left, *vma; test_vma_merge_special_flags() local 724 struct vm_area_struct *vma_prev, *vma_next, *vma; test_vma_merge_with_close() local 935 struct vm_area_struct *vma; test_vma_merge_new_with_close() local 981 struct vm_area_struct *vma, *vma_prev, *vma_next; test_merge_existing() local 1211 struct vm_area_struct *vma, *vma_prev, *vma_next; test_anon_vma_non_mergeable() local 1305 struct vm_area_struct *vma_prev, *vma_next, *vma; test_dup_anon_vma() local 1463 struct vm_area_struct *vma_prev, *vma; test_vmi_prealloc_fail() local 1524 struct vm_area_struct *vma; test_merge_extend() local 1555 struct vm_area_struct *vma, *vma_new, *vma_next; test_copy_vma() local 1587 struct vm_area_struct *vma_prev, *vma; test_expand_only_mode() local 1626 struct vm_area_struct *vma; test_mmap_region_basic() local [all...] |
/linux/arch/powerpc/include/asm/book3s/64/ |
H A D | tlbflush.h | 47 static inline void flush_pmd_tlb_range(struct vm_area_struct *vma, in flush_pmd_tlb_range() argument 55 static inline void flush_pud_tlb_range(struct vm_area_struct *vma, in flush_pud_tlb_range() argument 63 static inline void flush_hugetlb_tlb_range(struct vm_area_struct *vma, in flush_hugetlb_tlb_range() argument 71 static inline void flush_tlb_range(struct vm_area_struct *vma, in flush_tlb_range() argument 91 static inline void local_flush_tlb_page(struct vm_area_struct *vma, in local_flush_tlb_page() argument 120 flush_tlb_page(struct vm_area_struct * vma,unsigned long vmaddr) flush_tlb_page() argument 128 flush_tlb_page(vma,addr) global() argument 132 flush_tlb_fix_spurious_fault(struct vm_area_struct * vma,unsigned long address,pte_t * ptep) flush_tlb_fix_spurious_fault() argument [all...] |
/linux/arch/parisc/kernel/ |
H A D | cache.c | 331 __flush_cache_page(struct vm_area_struct *vma, unsigned long vmaddr, in __flush_cache_page() argument 393 void flush_icache_pages(struct vm_area_struct *vma, struct page *page, in flush_icache_pages() argument 476 struct vm_area_struct *vma; in flush_dcache_folio() local 623 flush_cache_page_if_present(struct vm_area_struct * vma,unsigned long vmaddr) flush_cache_page_if_present() argument 648 copy_user_highpage(struct page * to,struct page * from,unsigned long vaddr,struct vm_area_struct * vma) copy_user_highpage() argument 660 copy_to_user_page(struct vm_area_struct * vma,struct page * page,unsigned long user_vaddr,void * dst,void * src,int len) copy_to_user_page() argument 668 copy_from_user_page(struct vm_area_struct * vma,struct page * page,unsigned long user_vaddr,void * dst,void * src,int len) copy_from_user_page() argument 705 flush_cache_pages(struct vm_area_struct * vma,unsigned long start,unsigned long end) flush_cache_pages() argument 715 struct vm_area_struct *vma; mm_total_size() local 729 struct vm_area_struct *vma; flush_cache_mm() local 754 flush_cache_range(struct vm_area_struct * vma,unsigned long start,unsigned long end) flush_cache_range() argument 771 flush_cache_page(struct vm_area_struct * vma,unsigned long vmaddr,unsigned long pfn) flush_cache_page() argument 776 flush_anon_page(struct vm_area_struct * vma,struct page * page,unsigned long vmaddr) flush_anon_page() argument 784 ptep_clear_flush_young(struct vm_area_struct * vma,unsigned long addr,pte_t * ptep) ptep_clear_flush_young() argument 804 ptep_clear_flush(struct vm_area_struct * vma,unsigned long addr,pte_t * ptep) ptep_clear_flush() argument [all...] |
/linux/drivers/gpu/drm/nouveau/ |
H A D | nouveau_vmm.c | 29 nouveau_vma_unmap(struct nouveau_vma *vma) in nouveau_vma_unmap() argument 38 nouveau_vma_map(struct nouveau_vma *vma, struct nouveau_mem *mem) in nouveau_vma_map() argument 51 struct nouveau_vma *vma; in nouveau_vma_find() local 64 struct nouveau_vma *vma = *pvma; nouveau_vma_del() local 81 struct nouveau_vma *vma; nouveau_vma_new() local [all...] |
/linux/arch/x86/mm/ |
H A D | pgtable.c | 391 int ptep_set_access_flags(struct vm_area_struct *vma, in ptep_set_access_flags() argument 404 int pmdp_set_access_flags(struct vm_area_struct *vma, in pmdp_set_access_flags() argument 425 int pudp_set_access_flags(struct vm_area_struct *vma, unsigned long address, in pudp_set_access_flags() argument 446 int ptep_test_and_clear_young(struct vm_area_struct *vma, in ptep_test_and_clear_young() argument 459 int pmdp_test_and_clear_young(struct vm_area_struct *vma, in pmdp_test_and_clear_young() argument 473 int pudp_test_and_clear_young(struct vm_area_struct *vma, in pudp_test_and_clear_young() argument 486 int ptep_clear_flush_young(struct vm_area_struct *vma, in ptep_clear_flush_young() argument 506 int pmdp_clear_flush_young(struct vm_area_struct *vma, in pmdp_clear_flush_young() argument 520 pmdp_invalidate_ad(struct vm_area_struct * vma,unsigned long address,pmd_t * pmdp) pmdp_invalidate_ad() argument 535 pudp_invalidate(struct vm_area_struct * vma,unsigned long address,pud_t * pudp) pudp_invalidate() argument 802 pte_mkwrite(pte_t pte,struct vm_area_struct * vma) pte_mkwrite() argument 812 pmd_mkwrite(pmd_t pmd,struct vm_area_struct * vma) pmd_mkwrite() argument 822 arch_check_zapped_pte(struct vm_area_struct * vma,pte_t pte) arch_check_zapped_pte() argument 835 arch_check_zapped_pmd(struct vm_area_struct * vma,pmd_t pmd) arch_check_zapped_pmd() argument 842 arch_check_zapped_pud(struct vm_area_struct * vma,pud_t pud) arch_check_zapped_pud() argument [all...] |
/linux/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
H A D | vmm.c | 802 struct nvkm_vma *vma = kzalloc(sizeof(*vma), GFP_KERNEL); in nvkm_vma_new() local 813 nvkm_vma_tail(struct nvkm_vma *vma, u64 tail) in nvkm_vma_tail() argument 836 nvkm_vmm_free_remove(struct nvkm_vmm * vmm,struct nvkm_vma * vma) nvkm_vmm_free_remove() argument 842 nvkm_vmm_free_delete(struct nvkm_vmm * vmm,struct nvkm_vma * vma) nvkm_vmm_free_delete() argument 850 nvkm_vmm_free_insert(struct nvkm_vmm * vmm,struct nvkm_vma * vma) nvkm_vmm_free_insert() argument 878 nvkm_vmm_node_remove(struct nvkm_vmm * vmm,struct nvkm_vma * vma) nvkm_vmm_node_remove() argument 884 nvkm_vmm_node_delete(struct nvkm_vmm * vmm,struct nvkm_vma * vma) nvkm_vmm_node_delete() argument 892 nvkm_vmm_node_insert(struct nvkm_vmm * vmm,struct nvkm_vma * vma) nvkm_vmm_node_insert() argument 918 struct nvkm_vma *vma = rb_entry(node, typeof(*vma), tree); nvkm_vmm_node_search() local 935 nvkm_vmm_node_merge(struct nvkm_vmm * vmm,struct nvkm_vma * prev,struct nvkm_vma * vma,struct nvkm_vma * next,u64 size) nvkm_vmm_node_merge() argument 977 nvkm_vmm_node_split(struct nvkm_vmm * vmm,struct nvkm_vma * vma,u64 addr,u64 size) nvkm_vmm_node_split() argument 1003 nvkm_vma_dump(struct nvkm_vma * vma) nvkm_vma_dump() argument 1021 struct nvkm_vma *vma; nvkm_vmm_dump() local 1030 struct nvkm_vma *vma; nvkm_vmm_dtor() local 1040 struct nvkm_vma *vma = rb_entry(node, typeof(*vma), tree); nvkm_vmm_dtor() local 1074 struct nvkm_vma *vma; nvkm_vmm_ctor_managed() local 1094 struct nvkm_vma *vma; nvkm_vmm_ctor() local 1211 nvkm_vmm_pfn_split_merge(struct nvkm_vmm * vmm,struct nvkm_vma * vma,u64 addr,u64 size,u8 page,bool map) nvkm_vmm_pfn_split_merge() argument 1236 struct nvkm_vma *vma = nvkm_vmm_node_search(vmm, addr); nvkm_vmm_pfn_unmap() local 1273 struct nvkm_vma *vma, *tmp; nvkm_vmm_pfn_map() local 1393 nvkm_vmm_unmap_region(struct nvkm_vmm * vmm,struct nvkm_vma * vma) nvkm_vmm_unmap_region() argument 1410 nvkm_vmm_unmap_locked(struct nvkm_vmm * vmm,struct nvkm_vma * vma,bool pfn) nvkm_vmm_unmap_locked() argument 1425 nvkm_vmm_unmap(struct nvkm_vmm * vmm,struct nvkm_vma * vma) nvkm_vmm_unmap() argument 1435 nvkm_vmm_map_valid(struct nvkm_vmm * vmm,struct nvkm_vma * vma,void * argv,u32 argc,struct nvkm_vmm_map * map) nvkm_vmm_map_valid() argument 1471 nvkm_vmm_map_choose(struct nvkm_vmm * vmm,struct nvkm_vma * vma,void * argv,u32 argc,struct nvkm_vmm_map * map) nvkm_vmm_map_choose() argument 1483 nvkm_vmm_map_locked(struct nvkm_vmm * vmm,struct nvkm_vma * vma,void * argv,u32 argc,struct nvkm_vmm_map * map) nvkm_vmm_map_locked() argument 1571 nvkm_vmm_map(struct nvkm_vmm * vmm,struct nvkm_vma * vma,void * argv,u32 argc,struct nvkm_vmm_map * map) nvkm_vmm_map() argument 1588 nvkm_vmm_put_region(struct nvkm_vmm * vmm,struct nvkm_vma * vma) nvkm_vmm_put_region() argument 1607 nvkm_vmm_put_locked(struct nvkm_vmm * vmm,struct nvkm_vma * vma) nvkm_vmm_put_locked() argument 1691 struct nvkm_vma *vma = *pvma; nvkm_vmm_put() local 1706 struct nvkm_vma *vma = NULL, *tmp; nvkm_vmm_get_locked() local [all...] |
/linux/drivers/gpu/drm/i915/display/ |
H A D | intel_fb_pin.c | 35 struct i915_vma *vma; in intel_fb_pin_to_dpt() local 124 struct i915_vma *vma; intel_fb_pin_to_ggtt() local 224 intel_fb_unpin_vma(struct i915_vma * vma,unsigned long flags) intel_fb_unpin_vma() argument 266 struct i915_vma *vma; intel_plane_pin_fb() local 321 struct i915_vma *vma; intel_plane_unpin_fb() local 338 intel_fb_get_map(struct i915_vma * vma,struct iosys_map * map) intel_fb_get_map() argument [all...] |