Lines Matching refs:vm_mm
122 if (!vma->vm_mm) /* vdso */ in __thp_vma_allowable_orders()
1194 if (mem_cgroup_charge(folio, vma->vm_mm, gfp)) { in vma_alloc_anon_folio_pmd()
1230 set_pmd_at(vma->vm_mm, haddr, pmd, entry); in map_anon_folio_pmd()
1232 add_mm_counter(vma->vm_mm, MM_ANONPAGES, HPAGE_PMD_NR); in map_anon_folio_pmd()
1235 count_memcg_event_mm(vma->vm_mm, THP_FAULT_ALLOC); in map_anon_folio_pmd()
1250 pgtable = pte_alloc_one(vma->vm_mm); in __do_huge_pmd_anonymous_page()
1256 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in __do_huge_pmd_anonymous_page()
1260 ret = check_stable_address_space(vma->vm_mm); in __do_huge_pmd_anonymous_page()
1268 pte_free(vma->vm_mm, pgtable); in __do_huge_pmd_anonymous_page()
1273 pgtable_trans_huge_deposit(vma->vm_mm, vmf->pmd, pgtable); in __do_huge_pmd_anonymous_page()
1275 mm_inc_nr_ptes(vma->vm_mm); in __do_huge_pmd_anonymous_page()
1285 pte_free(vma->vm_mm, pgtable); in __do_huge_pmd_anonymous_page()
1353 !mm_forbids_zeropage(vma->vm_mm) && in do_huge_pmd_anonymous_page()
1359 pgtable = pte_alloc_one(vma->vm_mm); in do_huge_pmd_anonymous_page()
1362 zero_folio = mm_get_huge_zero_folio(vma->vm_mm); in do_huge_pmd_anonymous_page()
1364 pte_free(vma->vm_mm, pgtable); in do_huge_pmd_anonymous_page()
1368 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in do_huge_pmd_anonymous_page()
1371 ret = check_stable_address_space(vma->vm_mm); in do_huge_pmd_anonymous_page()
1374 pte_free(vma->vm_mm, pgtable); in do_huge_pmd_anonymous_page()
1377 pte_free(vma->vm_mm, pgtable); in do_huge_pmd_anonymous_page()
1381 set_huge_zero_folio(pgtable, vma->vm_mm, vma, in do_huge_pmd_anonymous_page()
1388 pte_free(vma->vm_mm, pgtable); in do_huge_pmd_anonymous_page()
1408 struct mm_struct *mm = vma->vm_mm; in insert_pmd()
1417 pgtable = pte_alloc_one(vma->vm_mm); in insert_pmd()
1539 struct mm_struct *mm = vma->vm_mm; in insert_pud()
1835 vmf->ptl = pud_lock(vmf->vma->vm_mm, vmf->pud); in huge_pud_set_accessed()
1849 vmf->ptl = pmd_lock(vmf->vma->vm_mm, vmf->pmd); in huge_pmd_set_accessed()
1871 mmu_notifier_range_init(&range, MMU_NOTIFY_CLEAR, 0, vma->vm_mm, haddr, in do_huge_zero_wp_pmd()
1874 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in do_huge_zero_wp_pmd()
1877 ret = check_stable_address_space(vma->vm_mm); in do_huge_zero_wp_pmd()
1900 vmf->ptl = pmd_lockptr(vma->vm_mm, vmf->pmd); in do_huge_pmd_wp_page()
2027 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in do_huge_pmd_numa_page()
2072 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in do_huge_pmd_numa_page()
2083 set_pmd_at(vma->vm_mm, haddr, vmf->pmd, pmd); in do_huge_pmd_numa_page()
2291 struct mm_struct *mm = vma->vm_mm; in move_huge_pmd()
2348 struct mm_struct *mm = vma->vm_mm; in change_huge_pmd()
2487 struct mm_struct *mm = vma->vm_mm; in change_huge_pud()
2671 ptl = pmd_lock(vma->vm_mm, pmd); in __pmd_trans_huge_lock()
2688 ptl = pud_lock(vma->vm_mm, pud); in __pud_trans_huge_lock()
2759 add_mm_counter(vma->vm_mm, mm_counter_file(folio), in __split_huge_pud_locked()
2769 mmu_notifier_range_init(&range, MMU_NOTIFY_CLEAR, 0, vma->vm_mm, in __split_huge_pud()
2773 ptl = pud_lock(vma->vm_mm, pud); in __split_huge_pud()
2792 struct mm_struct *mm = vma->vm_mm; in __split_huge_zero_page_pmd()
2833 struct mm_struct *mm = vma->vm_mm; in __split_huge_pmd_locked()
3065 mmu_notifier_range_init(&range, MMU_NOTIFY_CLEAR, 0, vma->vm_mm, in __split_huge_pmd()
3069 ptl = pmd_lock(vma->vm_mm, pmd); in __split_huge_pmd()
3078 pmd_t *pmd = mm_find_pmd(vma->vm_mm, address); in split_huge_pmd_address()
3141 struct mm_struct *mm = vma->vm_mm; in __discard_anon_folio_pmd_locked()
4596 struct mm_struct *mm = vma->vm_mm; in set_pmd_migration_entry()
4645 struct mm_struct *mm = vma->vm_mm; in remove_migration_pmd()