Lines Matching defs:tlb
39 #include <asm/tlb.h>
86 static long change_pte_range(struct mmu_gather *tlb,
98 tlb_change_page_size(tlb, PAGE_SIZE);
197 tlb_flush_pte_range(tlb, addr, PAGE_SIZE);
361 static inline long change_pmd_range(struct mmu_gather *tlb,
402 ret = change_huge_pmd(tlb, vma, pmd,
417 ret = change_pte_range(tlb, vma, pmd, addr, next, newprot,
431 static inline long change_pud_range(struct mmu_gather *tlb,
469 ret = change_huge_pud(tlb, vma, pudp,
480 pages += change_pmd_range(tlb, vma, pudp, addr, next, newprot,
490 static inline long change_p4d_range(struct mmu_gather *tlb,
506 pages += change_pud_range(tlb, vma, p4d, addr, next, newprot,
513 static long change_protection_range(struct mmu_gather *tlb,
524 tlb_start_vma(tlb, vma);
534 pages += change_p4d_range(tlb, vma, pgd, addr, next, newprot,
538 tlb_end_vma(tlb, vma);
543 long change_protection(struct mmu_gather *tlb,
568 pages = change_protection_range(tlb, vma, start, end, newprot,
605 mprotect_fixup(struct vma_iterator *vmi, struct mmu_gather *tlb,
684 change_protection(tlb, vma, start, end, mm_cp_flags);
720 struct mmu_gather tlb;
781 tlb_gather_mmu(&tlb, current->mm);
840 error = mprotect_fixup(&vmi, &tlb, vma, &prev, nstart, tmp, newflags);
848 tlb_finish_mmu(&tlb);