| /linux/arch/arm64/mm/ |
| H A D | hugetlbpage.c | 434 ncontig = num_contig_ptes(huge_page_size(hstate_vma(vma)), &pgsize); in huge_ptep_set_access_flags() 483 ncontig = num_contig_ptes(huge_page_size(hstate_vma(vma)), &pgsize); in huge_ptep_clear_flush() 519 unsigned long psize = huge_page_size(hstate_vma(vma)); in huge_ptep_modify_prot_start() 536 unsigned long psize = huge_page_size(hstate_vma(vma)); in huge_ptep_modify_prot_commit()
|
| /linux/arch/mips/include/asm/ |
| H A D | hugetlb.h | 32 unsigned long sz = huge_page_size(hstate_vma(vma)); in huge_ptep_clear_flush()
|
| /linux/arch/powerpc/include/asm/ |
| H A D | hugetlb.h | 59 unsigned long sz = huge_page_size(hstate_vma(vma)); in huge_ptep_clear_flush()
|
| /linux/arch/loongarch/include/asm/ |
| H A D | hugetlb.h | 41 unsigned long sz = huge_page_size(hstate_vma(vma)); in huge_ptep_clear_flush()
|
| /linux/arch/powerpc/mm/book3s64/ |
| H A D | radix_hugetlbpage.c | 50 unsigned long psize = huge_page_size(hstate_vma(vma)); in radix__huge_ptep_modify_prot_commit()
|
| H A D | hugetlbpage.c | 162 psize = huge_page_size(hstate_vma(vma)); in huge_ptep_modify_prot_commit()
|
| /linux/mm/ |
| H A D | hugetlb.c | 2912 struct hstate *h = hstate_vma(vma); in alloc_hugetlb_folio() 4735 struct hstate *h = hstate_vma(vma); in hugetlb_vm_op_close() 4766 if (addr & ~(huge_page_mask(hstate_vma(vma)))) in hugetlb_vm_op_split() 4806 return huge_page_size(hstate_vma(vma)); in hugetlb_vm_op_pagesize() 4840 unsigned int shift = huge_page_shift(hstate_vma(vma)); in make_huge_pte() 4881 hugetlb_count_add(pages_per_huge_page(hstate_vma(vma)), vma->vm_mm); in hugetlb_install_folio() 4893 struct hstate *h = hstate_vma(src_vma); in copy_hugetlb_page_range() 5073 struct hstate *h = hstate_vma(vma); in move_huge_pte() 5112 struct hstate *h = hstate_vma(vma); in move_hugetlb_page_tables() 5182 struct hstate *h = hstate_vma(vma); in __unmap_hugepage_range() [all …]
|
| H A D | mincore.c | 34 ptl = huge_pte_lock(hstate_vma(walk->vma), walk->mm, pte); in mincore_hugetlb()
|
| H A D | hmm.c | 557 ptl = huge_pte_lock(hstate_vma(vma), walk->mm, pte); in hmm_vma_walk_hugetlb_entry() 563 hmm_pfn_flags_order(huge_page_order(hstate_vma(vma))); in hmm_vma_walk_hugetlb_entry()
|
| H A D | page_vma_mapped.c | 196 struct hstate *hstate = hstate_vma(vma); in page_vma_mapped_walk()
|
| H A D | mmu_gather.c | 486 tlb_change_page_size(tlb, huge_page_size(hstate_vma(vma))); in tlb_gather_mmu_vma()
|
| H A D | madvise.c | 888 if (range->start & ~huge_page_mask(hstate_vma(vma))) in madvise_dontneed_free_valid_vma() 897 range->end = ALIGN_DOWN(range->end, huge_page_size(hstate_vma(vma))); in madvise_dontneed_free_valid_vma()
|
| H A D | pagewalk.c | 343 struct hstate *h = hstate_vma(vma); in walk_hugetlb_range()
|
| H A D | mempolicy.c | 782 ptl = huge_pte_lock(hstate_vma(walk->vma), walk->mm, pte); in queue_folios_hugetlb() 2003 !hugepage_migration_supported(hstate_vma(vma))) in vma_migratable() 2334 *mpol = get_vma_policy(vma, addr, hstate_vma(vma)->order, &ilx); in huge_node()
|
| H A D | migrate.c | 416 struct hstate *h = hstate_vma(vma); in remove_migration_pte() 518 spinlock_t *ptl = huge_pte_lockptr(hstate_vma(vma), vma->vm_mm, ptep); in migration_entry_wait_huge()
|
| H A D | rmap.c | 2022 hsz = huge_page_size(hstate_vma(vma)); in try_to_unmap_one() 2449 hsz = huge_page_size(hstate_vma(vma)); in try_to_migrate_one()
|
| H A D | mremap.c | 1488 struct hstate *h __maybe_unused = hstate_vma(vrm->vma); in align_hugetlb()
|
| /linux/include/linux/ |
| H A D | hugetlb.h | 770 static inline struct hstate *hstate_vma(struct vm_area_struct *vma) in hstate_vma() function 1050 unsigned long psize = huge_page_size(hstate_vma(vma)); in huge_ptep_modify_prot_start() 1062 unsigned long psize = huge_page_size(hstate_vma(vma)); in huge_ptep_modify_prot_commit() 1155 static inline struct hstate *hstate_vma(struct vm_area_struct *vma)
|
| /linux/arch/arm64/include/asm/ |
| H A D | hugetlb.h | 99 unsigned long stride = huge_page_size(hstate_vma(vma)); in flush_hugetlb_tlb_range()
|
| /linux/mm/damon/ |
| H A D | vaddr.c | 340 unsigned long psize = huge_page_size(hstate_vma(vma)); in damon_hugetlb_mkold() 351 addr + huge_page_size(hstate_vma(vma)))) in damon_hugetlb_mkold() 365 struct hstate *h = hstate_vma(walk->vma); in damon_mkold_hugetlb_entry() 485 struct hstate *h = hstate_vma(walk->vma); in damon_young_hugetlb_entry()
|
| /linux/arch/riscv/mm/ |
| H A D | tlbflush.c | 178 stride_size = huge_page_size(hstate_vma(vma)); in flush_tlb_range()
|
| /linux/arch/riscv/kvm/ |
| H A D | mmu.c | 491 vma_pageshift = huge_page_shift(hstate_vma(vma)); in kvm_riscv_mmu_map() 499 gfn = (gpa & huge_page_mask(hstate_vma(vma))) >> PAGE_SHIFT; in kvm_riscv_mmu_map()
|
| /linux/arch/x86/include/asm/ |
| H A D | tlbflush.h | 312 ? huge_page_shift(hstate_vma(vma)) \
|
| /linux/fs/proc/ |
| H A D | task_mmu.c | 1243 ptl = huge_pte_lock(hstate_vma(vma), walk->mm, pte); in smaps_hugetlb_range() 1259 mss->shared_hugetlb += huge_page_size(hstate_vma(vma)); in smaps_hugetlb_range() 1261 mss->private_hugetlb += huge_page_size(hstate_vma(vma)); in smaps_hugetlb_range() 2137 ptl = huge_pte_lock(hstate_vma(vma), walk->mm, ptep); in pagemap_hugetlb_range() 2513 const unsigned long psize = huge_page_size(hstate_vma(vma)); in make_uffd_wp_huge_pte() 2857 ptl = huge_pte_lock(hstate_vma(vma), vma->vm_mm, ptep); in pagemap_scan_hugetlb_entry() 3266 ptl = huge_pte_lock(hstate_vma(walk->vma), walk->mm, pte); in gather_hugetlb_stats()
|
| /linux/arch/powerpc/mm/ |
| H A D | pgtable.c | 290 struct hstate *h = hstate_vma(vma); in huge_ptep_set_access_flags()
|