/linux/arch/hexagon/include/asm/ |
H A D | cacheflush.h | 61 static inline void update_mmu_cache_range(struct vm_fault *vmf, in update_mmu_cache_range() function 69 update_mmu_cache_range(NULL, vma, addr, ptep, 1)
|
/linux/arch/microblaze/include/asm/ |
H A D | tlbflush.h | 36 #define update_mmu_cache_range(vmf, vma, addr, ptep, nr) do { } while (0) macro 38 update_mmu_cache_range(NULL, vma, addr, ptep, 1)
|
/linux/arch/sh/include/asm/ |
H A D | pgtable.h | 105 static inline void update_mmu_cache_range(struct vm_fault *vmf, in update_mmu_cache_range() function 114 update_mmu_cache_range(NULL, vma, addr, ptep, 1)
|
/linux/arch/m68k/include/asm/ |
H A D | pgtable_mm.h | 139 static inline void update_mmu_cache_range(struct vm_fault *vmf, in update_mmu_cache_range() function 146 update_mmu_cache_range(NULL, vma, addr, ptep, 1)
|
/linux/arch/arc/include/asm/ |
H A D | pgtable-bits-arcv2.h | 104 void update_mmu_cache_range(struct vm_fault *vmf, struct vm_area_struct *vma, 108 update_mmu_cache_range(NULL, vma, addr, ptep, 1)
|
/linux/arch/powerpc/include/asm/ |
H A D | pgtable.h | 50 update_mmu_cache_range(NULL, vma, addr, ptep, 1) 147 static inline void update_mmu_cache_range(struct vm_fault *vmf, in update_mmu_cache_range() function
|
/linux/arch/nios2/include/asm/ |
H A D | pgtable.h | 288 void update_mmu_cache_range(struct vm_fault *vmf, struct vm_area_struct *vma, 292 update_mmu_cache_range(NULL, vma, addr, ptep, 1)
|
/linux/arch/arm/include/asm/ |
H A D | tlbflush.h | 625 void update_mmu_cache_range(struct vm_fault *vmf, struct vm_area_struct *vma, 628 static inline void update_mmu_cache_range(struct vm_fault *vmf, in update_mmu_cache_range() function 636 update_mmu_cache_range(NULL, vma, addr, ptep, 1)
|
/linux/arch/csky/include/asm/ |
H A D | pgtable.h | 266 void update_mmu_cache_range(struct vm_fault *vmf, struct vm_area_struct *vma, 269 update_mmu_cache_range(NULL, vma, addr, ptep, 1)
|
/linux/arch/csky/abiv1/ |
H A D | cacheflush.c | 44 void update_mmu_cache_range(struct vm_fault *vmf, struct vm_area_struct *vma, in update_mmu_cache_range() function
|
/linux/arch/csky/abiv2/ |
H A D | cacheflush.c | 10 void update_mmu_cache_range(struct vm_fault *vmf, struct vm_area_struct *vma, in update_mmu_cache_range() function
|
/linux/arch/openrisc/include/asm/ |
H A D | pgtable.h | 383 static inline void update_mmu_cache_range(struct vm_fault *vmf, in update_mmu_cache_range() function 392 update_mmu_cache_range(NULL, vma, addr, ptep, 1)
|
/linux/arch/xtensa/include/asm/ |
H A D | pgtable.h | 406 void update_mmu_cache_range(struct vm_fault *vmf, struct vm_area_struct *vma, 409 update_mmu_cache_range(NULL, vma, address, ptep, 1)
|
/linux/arch/mips/include/asm/ |
H A D | pgtable.h | 581 static inline void update_mmu_cache_range(struct vm_fault *vmf, function 595 update_mmu_cache_range(NULL, vma, address, ptep, 1) 598 update_mmu_cache_range(NULL, vma, address, ptep, nr)
|
/linux/arch/riscv/include/asm/ |
H A D | pgtable.h | 496 static inline void update_mmu_cache_range(struct vm_fault *vmf, in update_mmu_cache_range() function 521 update_mmu_cache_range(NULL, vma, addr, ptep, 1) 524 update_mmu_cache_range(NULL, vma, addr, ptep, nr)
|
/linux/arch/arc/mm/ |
H A D | tlb.c | 472 void update_mmu_cache_range(struct vm_fault *vmf, struct vm_area_struct *vma, in update_mmu_cache_range() function 534 update_mmu_cache_range(NULL, vma, addr, &pte, HPAGE_PMD_NR); in update_mmu_cache_pmd()
|
/linux/arch/nios2/mm/ |
H A D | cacheflush.c | 209 void update_mmu_cache_range(struct vm_fault *vmf, struct vm_area_struct *vma, in update_mmu_cache_range() function
|
/linux/Documentation/core-api/ |
H A D | cachetlb.rst | 91 5) ``void update_mmu_cache_range(struct vm_fault *vmf, 316 page. See sparc64's flush_dcache_folio and update_mmu_cache_range 322 flag bit. Later, in update_mmu_cache_range(), a check is made 371 flush_dcache_folio and update_mmu_cache_range. In the future, the hope
|
/linux/arch/um/include/asm/ |
H A D | pgtable.h | 296 #define update_mmu_cache_range(vmf, vma, address, ptep, nr) do {} while (0) macro
|
/linux/arch/alpha/include/asm/ |
H A D | pgtable.h | 306 static inline void update_mmu_cache_range(struct vm_fault *vmf, in update_mmu_cache_range() function
|
/linux/arch/sparc/include/asm/ |
H A D | pgtable_32.h | 320 #define update_mmu_cache_range(vmf, vma, address, ptep, nr) do { } while (0) macro
|
H A D | pgtable_64.h | 985 void update_mmu_cache_range(struct vm_fault *, struct vm_area_struct *, 988 update_mmu_cache_range(NULL, vma, addr, ptep, 1)
|
/linux/mm/ |
H A D | memory.c | 3095 update_mmu_cache_range(vmf, vma, addr, vmf->pte, 1); in __wp_page_copy_user() 3275 update_mmu_cache_range(vmf, vma, vmf->address, vmf->pte, 1); in wp_page_reuse() 3438 update_mmu_cache_range(vmf, vma, vmf->address, vmf->pte, 1); in wp_page_copy() 4623 update_mmu_cache_range(vmf, vma, address, ptep, nr_pages); in do_swap_page() 4863 update_mmu_cache_range(vmf, vma, addr, vmf->pte, nr_pages); in do_anonymous_page() 5070 update_mmu_cache_range(vmf, vma, addr, vmf->pte, nr); in set_pte_range() 5529 update_mmu_cache_range(vmf, vma, fault_addr, fault_pte, 1); in numa_rebuild_single_mapping() 5824 update_mmu_cache_range(vmf, vmf->vma, vmf->address, in handle_pte_fault()
|
/linux/arch/arm64/include/asm/ |
H A D | pgtable.h | 1483 static inline void update_mmu_cache_range(struct vm_fault *vmf, in update_mmu_cache_range() function 1495 update_mmu_cache_range(NULL, vma, addr, ptep, 1)
|
/linux/arch/sparc/mm/ |
H A D | init_64.c | 397 void update_mmu_cache_range(struct vm_fault *vmf, struct vm_area_struct *vma, in update_mmu_cache_range() function
|