Lines Matching full:fault
42 /* Are we prepared to handle this kernel fault? */ in no_context()
62 static inline void mm_fault_error(struct pt_regs *regs, unsigned long addr, vm_fault_t fault) in mm_fault_error() argument
64 if (fault & VM_FAULT_OOM) { in mm_fault_error()
67 * (which will retry the fault, or kill us if we got oom-killed). in mm_fault_error()
75 } else if (fault & (VM_FAULT_SIGBUS | VM_FAULT_HWPOISON | VM_FAULT_HWPOISON_LARGE)) { in mm_fault_error()
230 vm_fault_t fault; in handle_page_fault() local
242 * Fault-in kernel-space virtual memory on-demand. in handle_page_fault()
262 * in an atomic region, then we must not take the fault. in handle_page_fault()
298 fault = handle_mm_fault(vma, addr, flags | FAULT_FLAG_VMA_LOCK, regs); in handle_page_fault()
299 if (!(fault & (VM_FAULT_RETRY | VM_FAULT_COMPLETED))) in handle_page_fault()
302 if (!(fault & VM_FAULT_RETRY)) { in handle_page_fault()
307 if (fault & VM_FAULT_MAJOR) in handle_page_fault()
310 if (fault_signal_pending(fault, regs)) { in handle_page_fault()
338 * If for any reason at all we could not handle the fault, in handle_page_fault()
340 * the fault. in handle_page_fault()
342 fault = handle_mm_fault(vma, addr, flags, regs); in handle_page_fault()
349 if (fault_signal_pending(fault, regs)) { in handle_page_fault()
355 /* The fault is fully completed (including releasing mmap lock) */ in handle_page_fault()
356 if (fault & VM_FAULT_COMPLETED) in handle_page_fault()
359 if (unlikely(fault & VM_FAULT_RETRY)) { in handle_page_fault()
373 if (unlikely(fault & VM_FAULT_ERROR)) { in handle_page_fault()
375 mm_fault_error(regs, addr, fault); in handle_page_fault()