| /linux/mm/ |
| H A D | page_vma_mapped.c | 250 pvmw->ptl = pmd_lock(mm, pvmw->pmd); in page_vma_mapped_walk() 284 spinlock_t *ptl = pmd_lock(mm, pvmw->pmd); in page_vma_mapped_walk()
|
| H A D | pt_reclaim.c | 43 pml = pmd_lock(mm, pmd); in try_to_free_pte()
|
| H A D | huge_memory.c | 1256 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in __do_huge_pmd_anonymous_page() 1368 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in do_huge_pmd_anonymous_page() 1422 ptl = pmd_lock(mm, pmd); in insert_pmd() 1672 dst_ptl = pmd_lock(dst_mm, dst_pmd); in copy_huge_pmd() 1696 dst_ptl = pmd_lock(dst_mm, dst_pmd); in copy_huge_pmd() 1849 vmf->ptl = pmd_lock(vmf->vma->vm_mm, vmf->pmd); in huge_pmd_set_accessed() 1874 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in do_huge_zero_wp_pmd() 2027 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in do_huge_pmd_numa_page() 2072 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in do_huge_pmd_numa_page() 2671 ptl = pmd_lock(vma->vm_mm, pmd); in __pmd_trans_huge_lock() [all …]
|
| H A D | khugepaged.c | 755 pmd_ptl = pmd_lock(vma->vm_mm, pmd); in __collapse_huge_page_copy_failed() 1167 pmd_ptl = pmd_lock(mm, pmd); /* probably unnecessary */ in collapse_huge_page() 1609 pml = pmd_lock(mm, pmd); in collapse_pte_mapped_thp() 1672 pml = pmd_lock(mm, pmd); in collapse_pte_mapped_thp() 1766 pml = pmd_lock(mm, pmd); in retract_page_tables()
|
| H A D | migrate_device.c | 78 ptl = pmd_lock(mm, pmdp); in migrate_vma_collect_pmd()
|
| H A D | pagewalk.c | 944 ptl = pmd_lock(vma->vm_mm, pmdp); in folio_walk_start()
|
| H A D | mremap.c | 389 old_ptl = pmd_lock(mm, old_pmd); in move_normal_pmd()
|
| H A D | debug_vm_pgtable.c | 1324 ptl = pmd_lock(args.mm, args.pmdp); in debug_vm_pgtable()
|
| H A D | memory.c | 427 spinlock_t *ptl = pmd_lock(mm, pmd); in pmd_install() 1935 spinlock_t *ptl = pmd_lock(tlb->mm, pmd); in zap_pmd_range() 5369 vmf->ptl = pmd_lock(vma->vm_mm, vmf->pmd); in do_set_pmd() 6716 lock = pmd_lock(mm, pmdp); in follow_pfnmap_start()
|
| H A D | migrate.c | 544 ptl = pmd_lock(mm, pmd); in pmd_migration_entry_wait()
|
| H A D | gup.c | 920 ptl = pmd_lock(mm, pmd); in follow_pmd_mask()
|
| H A D | ksm.c | 2486 ptl = pmd_lock(mm, pmdp); in ksm_next_page_pmd_entry()
|
| /linux/Documentation/translations/zh_CN/mm/ |
| H A D | split_page_table_lock.rst | 32 - pmd_lock()
|
| /linux/mm/damon/ |
| H A D | vaddr.c | 314 ptl = pmd_lock(walk->mm, pmd); in damon_mkold_pmd_entry() 452 ptl = pmd_lock(walk->mm, pmd); in damon_young_pmd_entry() 733 ptl = pmd_lock(walk->mm, pmd); in damos_va_migrate_pmd_entry()
|
| /linux/arch/s390/mm/ |
| H A D | pgtable.c | 791 ptl = pmd_lock(mm, pmdp); in set_guest_storage_key() 902 ptl = pmd_lock(mm, pmdp); in reset_guest_reference_bit() 966 ptl = pmd_lock(mm, pmdp); in get_guest_storage_key()
|
| H A D | gmap.c | 589 ptl = pmd_lock(mm, pmd); in __gmap_link()
|
| /linux/Documentation/mm/ |
| H A D | split_page_table_lock.rst | 35 - pmd_lock()
|
| H A D | transhuge.rst | 99 page table lock (pmd_lock()) and re-run pmd_trans_huge. Taking the
|
| H A D | process_addrs.rst | 516 set. The PMD spin lock is obtained via :c:func:`!pmd_lock`, however PTEs are 652 :c:func:`!pmd_ptdesc` called from :c:func:`!pmd_lock` and ultimately
|
| /linux/include/linux/ |
| H A D | mm.h | 3203 static inline spinlock_t *pmd_lock(struct mm_struct *mm, pmd_t *pmd) in pmd_lock() function
|