| H A D | userfaultfd.c | 100 unsigned long dst_start, in uffd_mfill_lock() argument 105 dst_vma = uffd_lock_vma(dst_mm, dst_start); in uffd_mfill_lock() 106 if (IS_ERR(dst_vma) || validate_dst_vma(dst_vma, dst_start + len)) in uffd_mfill_lock() 121 unsigned long dst_start, in uffd_mfill_lock() argument 127 dst_vma = find_vma_and_prepare_anon(dst_mm, dst_start); in uffd_mfill_lock() 131 if (validate_dst_vma(dst_vma, dst_start + len)) in uffd_mfill_lock() 495 unsigned long dst_start, in mfill_atomic_hugetlb() argument 524 dst_addr = dst_start; in mfill_atomic_hugetlb() 533 if (dst_start & (vma_hpagesize - 1) || len & (vma_hpagesize - 1)) in mfill_atomic_hugetlb() 542 dst_vma = uffd_mfill_lock(dst_mm, dst_start, le in mfill_atomic_hugetlb() 705 mfill_atomic(struct userfaultfd_ctx * ctx,unsigned long dst_start,unsigned long src_start,unsigned long len,uffd_flags_t flags) mfill_atomic() argument 868 mfill_atomic_copy(struct userfaultfd_ctx * ctx,unsigned long dst_start,unsigned long src_start,unsigned long len,uffd_flags_t flags) mfill_atomic_copy() argument 1578 find_vmas_mm_locked(struct mm_struct * mm,unsigned long dst_start,unsigned long src_start,struct vm_area_struct ** dst_vmap,struct vm_area_struct ** src_vmap) find_vmas_mm_locked() argument 1605 uffd_move_lock(struct mm_struct * mm,unsigned long dst_start,unsigned long src_start,struct vm_area_struct ** dst_vmap,struct vm_area_struct ** src_vmap) uffd_move_lock() argument 1681 uffd_move_lock(struct mm_struct * mm,unsigned long dst_start,unsigned long src_start,struct vm_area_struct ** dst_vmap,struct vm_area_struct ** src_vmap) uffd_move_lock() argument 1770 move_pages(struct userfaultfd_ctx * ctx,unsigned long dst_start,unsigned long src_start,unsigned long len,__u64 mode) move_pages() argument [all...] |