| /linux/arch/xtensa/mm/ |
| H A D | kasan_init.c | 24 for (i = 0; i < PTRS_PER_PTE; ++i) in kasan_early_init() 38 unsigned long n_pmds = n_pages / PTRS_PER_PTE; in populate() 49 for (k = 0; k < PTRS_PER_PTE; ++k, ++j) { in populate() 62 for (i = 0; i < n_pmds ; ++i, pte += PTRS_PER_PTE) in populate() 87 for (i = 0; i < PTRS_PER_PTE; ++i) in kasan_init()
|
| H A D | mmu.c | 30 n_pages = ALIGN(n_pages, PTRS_PER_PTE); in init_pmd() 43 for (i = 0; i < n_pages; i += PTRS_PER_PTE, ++pmd) { in init_pmd()
|
| /linux/arch/hexagon/include/asm/ |
| H A D | pgtable.h | 80 #define PTRS_PER_PTE 1024 macro 84 #define PTRS_PER_PTE 256 macro 88 #define PTRS_PER_PTE 64 macro 92 #define PTRS_PER_PTE 16 macro 96 #define PTRS_PER_PTE 4 macro
|
| /linux/arch/m68k/include/asm/ |
| H A D | pgtable_mm.h | 58 #define PTRS_PER_PTE 16 macro 63 #define PTRS_PER_PTE 512 macro 68 #define PTRS_PER_PTE 64 macro
|
| /linux/arch/arm/include/asm/ |
| H A D | pgtable-2level.h | 70 #define PTRS_PER_PTE 512 macro 74 #define PTE_HWTABLE_PTRS (PTRS_PER_PTE) 76 #define PTE_HWTABLE_SIZE (PTRS_PER_PTE * sizeof(u32))
|
| /linux/arch/arc/mm/ |
| H A D | highmem.c | 68 BUILD_BUG_ON(LAST_PKMAP > PTRS_PER_PTE); in kmap_init() 69 BUILD_BUG_ON(FIX_KMAP_SLOTS > PTRS_PER_PTE); in kmap_init()
|
| /linux/arch/csky/mm/ |
| H A D | init.c | 34 ((PTRS_PER_PGD - USER_PTRS_PER_PGD) * PTRS_PER_PTE) 37 pte_t invalid_pte_table[PTRS_PER_PTE] __page_aligned_bss; 70 __pa(kernel_pte_tables + (PTRS_PER_PTE * (i - USER_PTRS_PER_PGD))); in mmu_init()
|
| /linux/arch/powerpc/mm/book3s64/ |
| H A D | subpage_prot.c | 122 i = (addr >> PAGE_SHIFT) & (PTRS_PER_PTE - 1); in subpage_prot_clear() 123 nw = PTRS_PER_PTE - i; in subpage_prot_clear() 261 i = (addr >> PAGE_SHIFT) & (PTRS_PER_PTE - 1); in SYSCALL_DEFINE3() 262 nw = PTRS_PER_PTE - i; in SYSCALL_DEFINE3()
|
| H A D | hash_64k.c | 91 rpte = __real_pte(__pte(old_pte), ptep, PTRS_PER_PTE); in __hash_page_4K() 218 new_pte |= pte_set_hidx(ptep, rpte, subpg_index, slot, PTRS_PER_PTE); in __hash_page_4K() 269 rpte = __real_pte(__pte(old_pte), ptep, PTRS_PER_PTE); in __hash_page_64K() 334 new_pte |= pte_set_hidx(ptep, rpte, 0, slot, PTRS_PER_PTE); in __hash_page_64K()
|
| H A D | hash_4k.c | 60 rpte = __real_pte(__pte(old_pte), ptep, PTRS_PER_PTE); in __hash_page_4K() 122 new_pte |= pte_set_hidx(ptep, rpte, 0, slot, PTRS_PER_PTE); in __hash_page_4K()
|
| H A D | hash_tlb.c | 84 offset = PTRS_PER_PTE; in hpte_need_flush() 246 for (pte = start_pte; pte < start_pte + PTRS_PER_PTE; pte++) { in flush_hash_table_pmd_range()
|
| /linux/arch/powerpc/mm/kasan/ |
| H A D | init_book3s_64.c | 65 for (i = 0; i < PTRS_PER_PTE; i++) in kasan_init() 82 for (i = 0; i < PTRS_PER_PTE; i++) in kasan_init()
|
| H A D | init_book3e_64.c | 92 for (i = 0; i < PTRS_PER_PTE; i++) in kasan_early_init() 120 for (i = 0; i < PTRS_PER_PTE; i++) in kasan_init()
|
| /linux/arch/arm64/mm/ |
| H A D | kasan_init.c | 26 static pgd_t tmp_pg_dir[PTRS_PER_PTE] __initdata __aligned(PAGE_SIZE); 217 static pte_t tbl[PTRS_PER_PTE] __page_aligned_bss; in kasan_early_init() 274 return (addr >> (shift + PAGE_SHIFT)) % PTRS_PER_PTE; in next_level_idx() 294 clear_next_level(l++, next_level_idx(start), PTRS_PER_PTE); in clear_shadow() 368 for (i = 0; i < PTRS_PER_PTE; i++) in kasan_init_shadow()
|
| /linux/arch/mips/include/asm/ |
| H A D | pgtable-32.h | 89 # define PTRS_PER_PTE (PAGE_SIZE / sizeof(pte_t) / 2) macro 91 # define PTRS_PER_PTE (PAGE_SIZE / sizeof(pte_t)) macro 119 extern pte_t invalid_pte_table[PTRS_PER_PTE];
|
| H A D | pgtable-64.h | 132 #define PTRS_PER_PTE (PAGE_SIZE / sizeof(pte_t)) macro 144 min(PTRS_PER_PGD * PTRS_PER_PUD * PTRS_PER_PMD * PTRS_PER_PTE * PAGE_SIZE, \ 167 extern pte_t invalid_pte_table[PTRS_PER_PTE];
|
| /linux/arch/m68k/mm/ |
| H A D | sun3mmu.c | 64 next_pgtable += PTRS_PER_PTE * sizeof (pte_t); in paging_init() 70 for (i=0; i<PTRS_PER_PTE; ++i, ++pg_table) { in paging_init()
|
| /linux/arch/arm64/kernel/pi/ |
| H A D | map_range.c | 42 tbl += (start >> (lshift + PAGE_SHIFT)) % PTRS_PER_PTE; in map_range() 62 *pte += PTRS_PER_PTE * sizeof(pte_t); in map_range()
|
| /linux/arch/s390/mm/ |
| H A D | pageattr.c | 69 mask = ~(PTRS_PER_PTE * sizeof(pte_t) - 1); in pgt_set() 130 for (i = 0; i < PTRS_PER_PTE; i++) { in split_pmd_page() 137 update_page_count(PG_DIRECT_MAP_4K, PTRS_PER_PTE); in split_pmd_page() 450 nr = PTRS_PER_PTE - (nr & (PTRS_PER_PTE - 1)); in __kernel_map_pages()
|
| H A D | pgalloc.c | 134 memset64((u64 *)table, _PAGE_INVALID, PTRS_PER_PTE); in page_table_alloc_noprof() 135 memset64((u64 *)table + PTRS_PER_PTE, 0, PTRS_PER_PTE); in page_table_alloc_noprof() 177 memset64((u64 *)table, _PAGE_INVALID, PTRS_PER_PTE); in base_pgt_alloc()
|
| /linux/arch/x86/mm/ |
| H A D | init_32.c | 177 for (i = 0; i < PTRS_PER_PTE; i++) in page_table_kmap_check() 189 && lastpte && lastpte + PTRS_PER_PTE != pte); in page_table_kmap_check() 324 addr2 = (pfn + PTRS_PER_PTE-1) * PAGE_SIZE + in kernel_physical_mapping_init() 337 pfn += PTRS_PER_PTE; in kernel_physical_mapping_init() 344 for (; pte_ofs < PTRS_PER_PTE && pfn < end_pfn; in kernel_physical_mapping_init()
|
| /linux/arch/x86/power/ |
| H A D | hibernate_32.c | 111 pfn += PTRS_PER_PTE; in resume_physical_mapping_init() 119 max_pte = pte + PTRS_PER_PTE; in resume_physical_mapping_init()
|
| /linux/arch/loongarch/kvm/ |
| H A D | mmu.c | 128 end = p + PTRS_PER_PTE; in _kvm_pte_init() 738 for (i = 0; i < PTRS_PER_PTE; i++) { in kvm_split_huge() 748 kvm->stat.pages += PTRS_PER_PTE; in kvm_split_huge() 750 return child + (gfn & (PTRS_PER_PTE - 1)); in kvm_split_huge() 881 gfn = gfn & ~(PTRS_PER_PTE - 1); in kvm_map_page() 882 pfn = pfn & ~(PTRS_PER_PTE - 1); in kvm_map_page()
|
| /linux/arch/riscv/kvm/ |
| H A D | tlb.c | 29 if (PTRS_PER_PTE < (gpsz >> order)) { in kvm_riscv_local_hfence_gvma_vmid_gpa() 57 if (PTRS_PER_PTE < (gpsz >> order)) { in kvm_riscv_local_hfence_gvma_gpa() 88 if (PTRS_PER_PTE < (gvsz >> order)) { in kvm_riscv_local_hfence_vvma_asid_gva() 128 if (PTRS_PER_PTE < (gvsz >> order)) { in kvm_riscv_local_hfence_vvma_gva()
|
| H A D | gstage.c | 31 mask = (PTRS_PER_PTE * (1UL << kvm_riscv_gstage_pgd_xbits)) - 1; in gstage_pte_index() 33 mask = PTRS_PER_PTE - 1; in gstage_pte_index() 238 for (i = 0; i < PTRS_PER_PTE; i++) in kvm_riscv_gstage_op_pte()
|