Lines Matching full:fault
94 /* Are we prepared to handle this kernel fault? */ in no_context()
114 static inline void mm_fault_error(struct pt_regs *regs, unsigned long addr, vm_fault_t fault) in mm_fault_error() argument
121 if (fault & VM_FAULT_OOM) { in mm_fault_error()
124 * (which will retry the fault, or kill us if we got oom-killed). in mm_fault_error()
128 } else if (fault & (VM_FAULT_SIGBUS | VM_FAULT_HWPOISON | VM_FAULT_HWPOISON_LARGE)) { in mm_fault_error()
132 } else if (fault & VM_FAULT_SIGSEGV) { in mm_fault_error()
283 vm_fault_t fault; in handle_page_fault() local
295 * Fault-in kernel-space virtual memory on-demand. in handle_page_fault()
315 * in an atomic region, then we must not take the fault. in handle_page_fault()
354 fault = handle_mm_fault(vma, addr, flags | FAULT_FLAG_VMA_LOCK, regs); in handle_page_fault()
355 if (!(fault & (VM_FAULT_RETRY | VM_FAULT_COMPLETED))) in handle_page_fault()
358 if (!(fault & VM_FAULT_RETRY)) { in handle_page_fault()
363 if (fault & VM_FAULT_MAJOR) in handle_page_fault()
366 if (fault_signal_pending(fault, regs)) { in handle_page_fault()
394 * If for any reason at all we could not handle the fault, in handle_page_fault()
396 * the fault. in handle_page_fault()
398 fault = handle_mm_fault(vma, addr, flags, regs); in handle_page_fault()
405 if (fault_signal_pending(fault, regs)) { in handle_page_fault()
411 /* The fault is fully completed (including releasing mmap lock) */ in handle_page_fault()
412 if (fault & VM_FAULT_COMPLETED) in handle_page_fault()
415 if (unlikely(fault & VM_FAULT_RETRY)) { in handle_page_fault()
429 if (unlikely(fault & VM_FAULT_ERROR)) { in handle_page_fault()
431 mm_fault_error(regs, addr, fault); in handle_page_fault()