Searched refs:hmm_range_fault (Results 1 – 11 of 11) sorted by relevance
129 int hmm_range_fault(struct hmm_range *range);150 ret = hmm_range_fault(&range);186 并如上所述调用 hmm_range_fault()。这将填充至少具有读取权限的范围内的所有页面。199 hmm_range_fault 完成后,标志位被设置为页表的当前状态,即 HMM_PFN_VALID | 如果页
125 int hmm_range_fault(struct hmm_range *range);
719 err = hmm_range_fault(&hmm_range); in drm_gpusvm_check_pages()795 err = hmm_range_fault(&hmm_range); in drm_gpusvm_scan_mm()1432 err = hmm_range_fault(&hmm_range); in drm_gpusvm_get_pages()1704 err = hmm_range_fault(&hmm_range); in drm_gpusvm_range_evict()
204 r = hmm_range_fault(hmm_range); in amdgpu_hmm_range_get_pages()
659 int hmm_range_fault(struct hmm_range *range) in hmm_range_fault() function686 EXPORT_SYMBOL(hmm_range_fault);
162 int hmm_range_fault(struct hmm_range *range);184 ret = hmm_range_fault(&range);224 and calls hmm_range_fault() as described above. This will fill fault all pages239 After hmm_range_fault completes the flag bits are set to the current state of
373 ret = hmm_range_fault(&range); in ib_umem_odp_map_dma_and_lock()
411 ret = hmm_range_fault(range); in mshv_region_hmm_fault_and_lock()
306 ret = hmm_range_fault(range); in dmirror_range_fault()1297 ret = hmm_range_fault(range); in dmirror_range_snapshot()
688 ret = hmm_range_fault(&range); in nouveau_range_fault()
921 ret = hmm_range_fault(&mapp->range); in aie2_populate_range()