/linux/arch/riscv/mm/ |
H A D | fault.c | 302 fault = handle_mm_fault(vma, addr, flags | FAULT_FLAG_VMA_LOCK, regs); in handle_page_fault() 346 fault = handle_mm_fault(vma, addr, flags, regs); in handle_page_fault()
|
/linux/samples/ftrace/ |
H A D | ftrace-direct-too.c | 160 ftrace_set_filter_ip(&direct, (unsigned long) handle_mm_fault, 0, 0); in ftrace_direct_init()
|
/linux/arch/sparc/mm/ |
H A D | fault_32.c | 179 fault = handle_mm_fault(vma, address, flags, regs); in do_sparc_fault() 327 switch (handle_mm_fault(vma, address, flags, NULL)) { in force_user_fault()
|
H A D | fault_64.c | 429 fault = handle_mm_fault(vma, address, flags, regs); in do_sparc64_fault()
|
/linux/arch/hexagon/mm/ |
H A D | vm_fault.c | 84 fault = handle_mm_fault(vma, address, flags, regs); in do_page_fault()
|
/linux/arch/powerpc/mm/ |
H A D | fault.c | 498 fault = handle_mm_fault(vma, address, flags | FAULT_FLAG_VMA_LOCK, regs); in ___do_page_fault() 541 fault = handle_mm_fault(vma, address, flags, regs); in ___do_page_fault()
|
H A D | copro_fault.c | 59 *flt = handle_mm_fault(vma, ea, is_write ? FAULT_FLAG_WRITE : 0, NULL); in copro_handle_mm_fault()
|
/linux/arch/m68k/mm/ |
H A D | fault.c | 141 fault = handle_mm_fault(vma, address, flags, regs); in do_page_fault()
|
/linux/arch/arc/mm/ |
H A D | fault.c | 135 fault = handle_mm_fault(vma, address, flags, regs); in do_page_fault()
|
/linux/arch/alpha/mm/ |
H A D | fault.c | 145 fault = handle_mm_fault(vma, address, flags, regs); in do_page_fault()
|
/linux/arch/arm/mm/ |
H A D | fault.c | 328 fault = handle_mm_fault(vma, addr, flags | FAULT_FLAG_VMA_LOCK, regs); in do_page_fault() 367 fault = handle_mm_fault(vma, addr & PAGE_MASK, flags, regs); in do_page_fault()
|
/linux/arch/nios2/mm/ |
H A D | fault.c | 124 fault = handle_mm_fault(vma, address, flags, regs); in do_page_fault()
|
/linux/arch/microblaze/mm/ |
H A D | fault.c | 221 fault = handle_mm_fault(vma, address, flags, regs); in do_page_fault()
|
/linux/arch/openrisc/mm/ |
H A D | fault.c | 166 fault = handle_mm_fault(vma, address, flags, regs); in do_page_fault()
|
/linux/arch/xtensa/mm/ |
H A D | fault.c | 159 fault = handle_mm_fault(vma, address, flags, regs); in do_page_fault()
|
/linux/arch/csky/mm/ |
H A D | fault.c | 263 fault = handle_mm_fault(vma, addr, flags, regs); in do_page_fault()
|
/linux/arch/mips/mm/ |
H A D | fault.c | 150 fault = handle_mm_fault(vma, address, flags, regs); in __do_page_fault()
|
/linux/Documentation/translations/zh_CN/mm/ |
H A D | hmm.rst | 203 CPU pte 没有设置写权限,那么HMM将调用handle_mm_fault()。
|
/linux/arch/um/kernel/ |
H A D | trap.c | 75 fault = handle_mm_fault(vma, address, flags, NULL); in handle_page_fault()
|
/linux/drivers/iommu/ |
H A D | iommu-sva.c | 251 ret = handle_mm_fault(vma, prm->addr, fault_flags, NULL); in iommu_sva_handle_mm()
|
/linux/arch/arm64/mm/ |
H A D | fault.c | 607 fault = handle_mm_fault(vma, addr, mm_flags | FAULT_FLAG_VMA_LOCK, regs); in do_page_fault() 650 fault = handle_mm_fault(vma, addr, mm_flags, regs); in do_page_fault()
|
/linux/arch/parisc/mm/ |
H A D | fault.c | 318 fault = handle_mm_fault(vma, address, flags, regs); in do_page_fault()
|
/linux/arch/sh/mm/ |
H A D | fault.c | 469 fault = handle_mm_fault(vma, address, flags, regs); in do_page_fault()
|
/linux/mm/ |
H A D | hmm.c | 80 if (handle_mm_fault(vma, addr, fault_flags, NULL) & in hmm_vma_fault()
|
H A D | gup.c | 598 * page tables. Return error instead of NULL to skip handle_mm_fault, in no_page_table() 601 * be zero-filled if handle_mm_fault() actually did handle it. in no_page_table() 1187 ret = handle_mm_fault(vma, address, fault_flags, NULL); in faultin_page() 1598 * @fault_flags:flags to pass down to handle_mm_fault() in vma_permits_fault() 1611 * unconditionally call handle_mm_fault() which will in turn perform all the in vma_permits_fault() 1647 ret = handle_mm_fault(vma, address, fault_flags, NULL); in fixup_user_fault() 2605 * FAULT_FLAG_ALLOW_RETRY to handle_mm_fault.
|