| H A D | userfaultfd.c | 1043 static struct folio *check_ptes_for_batched_move(struct vm_area_struct *src_vma, in check_ptes_for_batched_move() argument 1058 folio = vm_normal_folio(src_vma, src_addr, orig_src_pte); in check_ptes_for_batched_move() 1074 struct vm_area_struct *src_vma, in move_present_ptes() argument 1089 flush_cache_range(src_vma, src_addr, src_end); in move_present_ptes() 1137 src_folio = check_ptes_for_batched_move(src_vma, src_addr, in move_present_ptes() 1145 flush_tlb_range(src_vma, src_start, src_addr); in move_present_ptes() 1216 struct vm_area_struct *src_vma, in move_swap_pte() 1234 ptep_clear_flush(src_vma, src_addr, src_pte); in move_zeropage_pte() 1249 struct vm_area_struct *src_vma, 1337 ret = move_zeropage_pte(mm, dst_vma, src_vma, in move_pages_ptes() 1220 move_zeropage_pte(struct mm_struct * mm,struct vm_area_struct * dst_vma,struct vm_area_struct * src_vma,unsigned long dst_addr,unsigned long src_addr,pte_t * dst_pte,pte_t * src_pte,pte_t orig_dst_pte,pte_t orig_src_pte,pmd_t * dst_pmd,pmd_t dst_pmdval,spinlock_t * dst_ptl,spinlock_t * src_ptl) move_zeropage_pte() argument 1253 move_pages_ptes(struct mm_struct * mm,pmd_t * dst_pmd,pmd_t * src_pmd,struct vm_area_struct * dst_vma,struct vm_area_struct * src_vma,unsigned long dst_addr,unsigned long src_addr,unsigned long len,__u64 mode) move_pages_ptes() argument 1541 validate_move_areas(struct userfaultfd_ctx * ctx,struct vm_area_struct * src_vma,struct vm_area_struct * dst_vma) validate_move_areas() argument 1671 uffd_move_unlock(struct vm_area_struct * dst_vma,struct vm_area_struct * src_vma) uffd_move_unlock() argument 1696 uffd_move_unlock(struct vm_area_struct * dst_vma,struct vm_area_struct * src_vma) uffd_move_unlock() argument 1774 struct vm_area_struct *src_vma, *dst_vma; move_pages() local [all...] |