Lines Matching defs:vma

68 int ptep_set_access_flags(struct vm_area_struct *vma,
74 set_pte_at(vma->vm_mm, address, ptep, entry);
75 flush_tlb_fix_spurious_fault(vma, address, ptep);
82 int ptep_clear_flush_young(struct vm_area_struct *vma,
86 young = ptep_test_and_clear_young(vma, address, ptep);
88 flush_tlb_page(vma, address);
94 pte_t ptep_clear_flush(struct vm_area_struct *vma, unsigned long address,
97 struct mm_struct *mm = (vma)->vm_mm;
101 flush_tlb_page(vma, address);
109 int pmdp_set_access_flags(struct vm_area_struct *vma,
116 set_pmd_at(vma->vm_mm, address, pmdp, entry);
117 flush_pmd_tlb_range(vma, address, address + HPAGE_PMD_SIZE);
124 int pmdp_clear_flush_young(struct vm_area_struct *vma,
129 young = pmdp_test_and_clear_young(vma, address, pmdp);
131 flush_pmd_tlb_range(vma, address, address + HPAGE_PMD_SIZE);
137 pmd_t pmdp_huge_clear_flush(struct vm_area_struct *vma, unsigned long address,
144 pmd = pmdp_huge_get_and_clear(vma->vm_mm, address, pmdp);
145 flush_pmd_tlb_range(vma, address, address + HPAGE_PMD_SIZE);
150 pud_t pudp_huge_clear_flush(struct vm_area_struct *vma, unsigned long address,
157 pud = pudp_huge_get_and_clear(vma->vm_mm, address, pudp);
158 flush_pud_tlb_range(vma, address, address + HPAGE_PUD_SIZE);
198 pmd_t pmdp_invalidate(struct vm_area_struct *vma, unsigned long address,
202 pmd_t old = pmdp_establish(vma, address, pmdp, pmd_mkinvalid(*pmdp));
203 flush_pmd_tlb_range(vma, address, address + HPAGE_PMD_SIZE);
209 pmd_t pmdp_invalidate_ad(struct vm_area_struct *vma, unsigned long address,
213 return pmdp_invalidate(vma, address, pmdp);
218 pmd_t pmdp_collapse_flush(struct vm_area_struct *vma, unsigned long address,
229 pmd = pmdp_huge_get_and_clear(vma->vm_mm, address, pmdp);
232 flush_tlb_range(vma, address, address + HPAGE_PMD_SIZE);
324 * holding mmap_lock or vma lock for read or for write; or in truncate or rmap
362 * pte_offset_map() and co once the vma is detached from mm or mm_users is zero.