| /linux/mm/ |
| H A D | page_vma_mapped.c | 19 pte_t ptent; in map_pte() local 42 ptent = ptep_get(pvmw->pte); in map_pte() 45 if (!is_swap_pte(ptent)) in map_pte() 47 } else if (is_swap_pte(ptent)) { in map_pte() 65 entry = pte_to_swp_entry(ptent); in map_pte() 69 } else if (!pte_present(ptent)) { in map_pte() 107 pte_t ptent = ptep_get(pvmw->pte); in check_pte() local 111 if (!is_swap_pte(ptent)) in check_pte() 113 entry = pte_to_swp_entry(ptent); in check_pte() 119 } else if (is_swap_pte(ptent)) { in check_pte() [all …]
|
| H A D | mprotect.c | 179 pte_t *ptep, pte_t oldpte, pte_t ptent, int nr_ptes, in prot_commit_flush_ptes() argument 189 ptent = pte_advance_pfn(ptent, idx); in prot_commit_flush_ptes() 192 ptent = pte_mkwrite(ptent, vma); in prot_commit_flush_ptes() 194 modify_prot_commit_ptes(vma, addr, ptep, oldpte, ptent, nr_ptes); in prot_commit_flush_ptes() 195 if (pte_needs_flush(oldpte, ptent)) in prot_commit_flush_ptes() 232 pte_t oldpte, pte_t ptent, int nr_ptes, struct mmu_gather *tlb) in commit_anon_folio_batch() argument 242 prot_commit_flush_ptes(vma, addr, ptep, oldpte, ptent, len, in commit_anon_folio_batch() 251 pte_t oldpte, pte_t ptent, int nr_ptes, struct mmu_gather *tlb) in set_write_prot_commit_flush_ptes() argument 256 set_write = can_change_shared_pte_writable(vma, ptent); in set_write_prot_commit_flush_ptes() 257 prot_commit_flush_ptes(vma, addr, ptep, oldpte, ptent, nr_ptes, in set_write_prot_commit_flush_ptes() [all …]
|
| H A D | mapping_dirty_helpers.c | 38 pte_t ptent = ptep_get(pte); in wp_pte() local 40 if (pte_write(ptent)) { in wp_pte() 43 ptent = pte_wrprotect(old_pte); in wp_pte() 44 ptep_modify_prot_commit(walk->vma, addr, pte, old_pte, ptent); in wp_pte() 94 pte_t ptent = ptep_get(pte); in clean_record_pte() local 96 if (pte_dirty(ptent)) { in clean_record_pte() 101 ptent = pte_mkclean(old_pte); in clean_record_pte() 102 ptep_modify_prot_commit(walk->vma, addr, pte, old_pte, ptent); in clean_record_pte()
|
| H A D | madvise.c | 365 pte_t *start_pte, *pte, ptent; in madvise_cold_or_pageout_pte_range() local 461 ptent = ptep_get(pte); in madvise_cold_or_pageout_pte_range() 473 if (pte_none(ptent)) in madvise_cold_or_pageout_pte_range() 476 if (!pte_present(ptent)) in madvise_cold_or_pageout_pte_range() 479 folio = vm_normal_folio(vma, addr, ptent); in madvise_cold_or_pageout_pte_range() 491 nr = madvise_folio_pte_batch(addr, end, folio, pte, &ptent); in madvise_cold_or_pageout_pte_range() 533 if (!pageout && pte_young(ptent)) { in madvise_cold_or_pageout_pte_range() 664 pte_t *start_pte, *pte, ptent; in madvise_free_pte_range() local 683 ptent = ptep_get(pte); in madvise_free_pte_range() 685 if (pte_none(ptent)) in madvise_free_pte_range() [all …]
|
| H A D | mlock.c | 311 pte_t ptent = ptep_get(pte); in folio_mlock_step() local 316 return folio_pte_batch(folio, pte, ptent, count); in folio_mlock_step() 357 pte_t ptent; in mlock_pte_range() local 385 ptent = ptep_get(pte); in mlock_pte_range() 386 if (!pte_present(ptent)) in mlock_pte_range() 388 folio = vm_normal_folio(vma, addr, ptent); in mlock_pte_range()
|
| H A D | memory.c | 1216 pte_t ptent; in copy_pte_range() local 1273 ptent = ptep_get(src_pte); in copy_pte_range() 1274 if (pte_none(ptent)) { in copy_pte_range() 1278 if (unlikely(!pte_present(ptent))) { in copy_pte_range() 1292 ptent = ptep_get(src_pte); in copy_pte_range() 1293 VM_WARN_ON_ONCE(!pte_present(ptent)); in copy_pte_range() 1304 ptent, addr, max_nr, rss, &prealloc); in copy_pte_range() 1620 struct page *page, pte_t *pte, pte_t ptent, unsigned int nr, in zap_present_folio_ptes() argument 1628 ptent = get_and_clear_full_ptes(mm, addr, pte, nr, tlb->fullmm); in zap_present_folio_ptes() 1629 if (pte_dirty(ptent)) { in zap_present_folio_ptes() [all …]
|
| H A D | highmem.c | 205 pte_t ptent; in flush_all_zero_pkmaps() local 218 ptent = ptep_get(&pkmap_page_table[i]); in flush_all_zero_pkmaps() 219 BUG_ON(pte_none(ptent)); in flush_all_zero_pkmaps() 228 page = pte_page(ptent); in flush_all_zero_pkmaps()
|
| H A D | khugepaged.c | 1571 pte_t ptent = ptep_get(pte); in collapse_pte_mapped_thp() local 1574 if (pte_none(ptent)) in collapse_pte_mapped_thp() 1578 if (!pte_present(ptent)) { in collapse_pte_mapped_thp() 1583 page = vm_normal_page(vma, addr, ptent); in collapse_pte_mapped_thp() 1628 pte_t ptent = ptep_get(pte); in collapse_pte_mapped_thp() local 1632 if (pte_none(ptent)) in collapse_pte_mapped_thp() 1640 if (!pte_present(ptent)) { in collapse_pte_mapped_thp() 1644 page = vm_normal_page(vma, addr, ptent); in collapse_pte_mapped_thp() 1649 nr_batch_ptes = folio_pte_batch(folio, pte, ptent, max_nr_batch_ptes); in collapse_pte_mapped_thp()
|
| H A D | mempolicy.c | 683 pte_t ptent; in queue_folios_pte_range() local 702 ptent = ptep_get(pte); in queue_folios_pte_range() 703 if (pte_none(ptent)) in queue_folios_pte_range() 705 if (!pte_present(ptent)) { in queue_folios_pte_range() 706 if (is_migration_entry(pte_to_swp_entry(ptent))) in queue_folios_pte_range() 710 folio = vm_normal_folio(vma, addr, ptent); in queue_folios_pte_range() 714 nr = folio_pte_batch(folio, pte, ptent, max_nr); in queue_folios_pte_range()
|
| H A D | vmalloc.c | 358 pte_t ptent; in vunmap_pte_range() local 372 ptent = huge_ptep_get_and_clear(&init_mm, addr, pte, size); in vunmap_pte_range() 377 ptent = ptep_get_and_clear(&init_mm, addr, pte); in vunmap_pte_range() 378 WARN_ON(!pte_none(ptent) && !pte_present(ptent)); in vunmap_pte_range() 3546 pte_t ptent; in vmap_pfn_apply() local 3551 ptent = pte_mkspecial(pfn_pte(pfn, data->prot)); in vmap_pfn_apply() 3552 set_pte_at(&init_mm, addr, pte, ptent); in vmap_pfn_apply()
|
| H A D | memory-failure.c | 338 pte_t ptent; in dev_pagemap_mapping_shift() local 360 ptent = ptep_get(pte); in dev_pagemap_mapping_shift() 361 if (pte_present(ptent)) in dev_pagemap_mapping_shift()
|
| H A D | swapfile.c | 2246 pte_t ptent; in unuse_pte_range() local 2254 ptent = ptep_get_lockless(pte); in unuse_pte_range() 2256 if (!is_swap_pte(ptent)) in unuse_pte_range() 2259 entry = pte_to_swp_entry(ptent); in unuse_pte_range()
|
| H A D | vmscan.c | 3559 pte_t ptent = ptep_get(pte + i); in walk_pte_range() local 3564 pfn = get_pte_pfn(ptent, args->vma, addr, pgdat); in walk_pte_range() 3582 if (pte_dirty(ptent)) in walk_pte_range() 4288 pte_t ptent = ptep_get(pte + i); in lru_gen_look_around() local 4290 pfn = get_pte_pfn(ptent, vma, addr, pgdat); in lru_gen_look_around() 4308 if (pte_dirty(ptent)) in lru_gen_look_around()
|
| /linux/fs/proc/ |
| H A D | task_mmu.c | 1013 pte_t ptent = ptep_get(pte); in smaps_pte_entry() local 1015 if (pte_present(ptent)) { in smaps_pte_entry() 1016 page = vm_normal_page(vma, addr, ptent); in smaps_pte_entry() 1017 young = pte_young(ptent); in smaps_pte_entry() 1018 dirty = pte_dirty(ptent); in smaps_pte_entry() 1020 } else if (is_swap_pte(ptent)) { in smaps_pte_entry() 1021 swp_entry_t swpent = pte_to_swp_entry(ptent); in smaps_pte_entry() 1226 pte_t ptent; in smaps_hugetlb_range() local 1229 ptent = huge_ptep_get(walk->mm, addr, pte); in smaps_hugetlb_range() 1230 if (pte_present(ptent)) { in smaps_hugetlb_range() [all …]
|
| /linux/mm/damon/ |
| H A D | vaddr.c | 443 pte_t ptent; in damon_young_pmd_entry() local 484 ptent = ptep_get(pte); in damon_young_pmd_entry() 485 if (!pte_present(ptent)) in damon_young_pmd_entry() 487 folio = damon_get_folio(pte_pfn(ptent)); in damon_young_pmd_entry() 490 if (pte_young(ptent) || !folio_test_idle(folio) || in damon_young_pmd_entry() 770 pte_t ptent; in damos_va_migrate_pte_entry() local 772 ptent = ptep_get(pte); in damos_va_migrate_pte_entry() 773 if (pte_none(ptent) || !pte_present(ptent)) in damos_va_migrate_pte_entry() 776 folio = damon_get_folio(pte_pfn(ptent)); in damos_va_migrate_pte_entry() 909 pte_t *start_pte, *pte, ptent; in damos_va_stat_pmd_entry() local [all …]
|
| /linux/mm/kasan/ |
| H A D | init.c | 348 pte_t ptent; in kasan_remove_pte_table() local 355 ptent = ptep_get(pte); in kasan_remove_pte_table() 357 if (!pte_present(ptent)) in kasan_remove_pte_table() 360 if (WARN_ON(!kasan_early_shadow_page_entry(ptent))) in kasan_remove_pte_table()
|
| /linux/arch/arm64/mm/ |
| H A D | contpte.c | 62 pte_t ptent = __ptep_get_and_clear(mm, addr, ptep); in contpte_convert() local 64 if (pte_dirty(ptent)) in contpte_convert() 67 if (pte_young(ptent)) in contpte_convert()
|
| /linux/drivers/gpu/drm/i915/gem/selftests/ |
| H A D | i915_gem_mman.c | 1680 pte_t ptent = ptep_get(pte); in check_present_pte() local 1682 if (!pte_present(ptent) || pte_none(ptent)) { in check_present_pte() 1693 pte_t ptent = ptep_get(pte); in check_absent_pte() local 1695 if (pte_present(ptent) && !pte_none(ptent)) { in check_absent_pte()
|
| /linux/fs/ |
| H A D | userfaultfd.c | 286 pte_t ptent; in userfaultfd_must_wait() local 326 ptent = ptep_get(pte); in userfaultfd_must_wait() 327 if (pte_none_mostly(ptent)) in userfaultfd_must_wait() 329 if (!pte_write(ptent) && (reason & VM_UFFD_WP)) in userfaultfd_must_wait()
|