Home
last modified time | relevance | path

Searched refs:update_mmu_cache (Results 1 – 25 of 25) sorted by relevance

/linux/Documentation/translations/zh_CN/core-api/
H A Dcachetlb.rst81 5) ``void update_mmu_cache(struct vm_area_struct *vma,
/linux/arch/hexagon/include/asm/
H A Dcacheflush.h68 #define update_mmu_cache(vma, addr, ptep) \ macro
/linux/arch/microblaze/include/asm/
H A Dtlbflush.h37 #define update_mmu_cache(vma, addr, pte) \ macro
/linux/arch/sh/mm/
H A Dtlbex_32.c80 update_mmu_cache(NULL, address, pte); in handle_tlbmiss()
/linux/arch/sh/include/asm/
H A Dpgtable.h113 #define update_mmu_cache(vma, addr, ptep) \ macro
/linux/arch/m68k/include/asm/
H A Dpgtable_mm.h145 #define update_mmu_cache(vma, addr, ptep) \ macro
/linux/arch/arc/include/asm/
H A Dpgtable-bits-arcv2.h107 #define update_mmu_cache(vma, addr, ptep) \ macro
/linux/arch/powerpc/include/asm/
H A Dpgtable.h49 #define update_mmu_cache(vma, addr, ptep) \ macro
/linux/arch/riscv/include/asm/
H A Dpgtable.h518 #define update_mmu_cache(vma, addr, ptep) \ macro
529 update_mmu_cache(vma, address, ptep); in update_mmu_cache_pmd()
955 update_mmu_cache(vma, address, ptep); in update_mmu_cache_pud()
/linux/arch/csky/include/asm/
H A Dpgtable.h263 #define update_mmu_cache(vma, addr, ptep) \ macro
/linux/arch/nios2/include/asm/
H A Dpgtable.h285 #define update_mmu_cache(vma, addr, ptep) \ macro
298 * update_mmu_cache will unconditionally execute, handling both in ptep_set_access_flags()
/linux/arch/alpha/include/asm/
H A Dpgtable.h294 extern inline void update_mmu_cache(struct vm_area_struct * vma, in update_mmu_cache() function
/linux/arch/xtensa/include/asm/
H A Dpgtable.h402 #define update_mmu_cache(vma, address, ptep) \ macro
/linux/arch/sparc/include/asm/
H A Dpgtable_32.h314 #define update_mmu_cache(vma, address, ptep) do { } while (0) macro
H A Dpgtable_64.h983 #define update_mmu_cache(vma, addr, ptep) \ macro
/linux/mm/
H A Duserfaultfd.c229 update_mmu_cache(dst_vma, dst_addr, dst_pte); in mfill_atomic_install_pte()
371 update_mmu_cache(dst_vma, dst_addr, dst_pte); in mfill_atomic_pte_zeropage()
453 update_mmu_cache(dst_vma, dst_addr, dst_pte); in mfill_atomic_pte_poison()
H A Dmigrate_device.c690 update_mmu_cache(vma, addr, ptep); in migrate_vma_insert_page()
H A Dhugetlb.c5500 update_mmu_cache(vma, address, ptep); in set_huge_ptep_writable()
6812 update_mmu_cache(vma, vmf.address, vmf.pte); in hugetlb_fault()
6908 update_mmu_cache(dst_vma, dst_addr, dst_pte); in hugetlb_mfill_atomic_pte()
7065 update_mmu_cache(dst_vma, dst_addr, dst_pte); in hugetlb_mfill_atomic_pte()
H A Dmemory.c766 update_mmu_cache(vma, address, ptep); in restore_exclusive_pte()
2167 update_mmu_cache(vma, addr, pte); in insert_page_into_pte_locked()
2471 update_mmu_cache(vma, addr, pte); in insert_pfn()
2485 update_mmu_cache(vma, addr, pte); /* XXX: why not for insert_page? */ in insert_pfn()
5998 * There is also a hook called "update_mmu_cache()" that architectures
H A Dmadvise.c1231 update_mmu_cache(walk->vma, addr, pte); in guard_remove_pte_entry()
H A Dmigrate.c423 update_mmu_cache(vma, pvmw.address, pvmw.pte); in remove_migration_pte()
H A Dgup.c789 update_mmu_cache(vma, address, pte); in follow_pfn_pte()
/linux/arch/loongarch/include/asm/
H A Dpgtable.h446 #define update_mmu_cache(vma, addr, ptep) \ macro
/linux/arch/mips/include/asm/
H A Dpgtable.h501 * update_mmu_cache will unconditionally execute, handling both
588 #define update_mmu_cache(vma, address, ptep) \ macro
/linux/arch/x86/include/asm/
H A Dpgtable.h1495 static inline void update_mmu_cache(struct vm_area_struct *vma, in update_mmu_cache() function