| H A D | userfaultfd.c | 146 /* Check if dst_addr is outside of file's size. Must be called with ptl held. */ 148 unsigned long dst_addr) in mfill_file_over_size() argument 157 offset = linear_page_index(dst_vma, dst_addr); in mfill_file_over_size() 163 * Install PTEs, to map dst_addr (within dst_vma) to page. 170 unsigned long dst_addr, struct page *page, in mfill_atomic_install_pte() argument 193 dst_pte = pte_offset_map_lock(dst_mm, dst_pmd, dst_addr, &ptl); in mfill_atomic_install_pte() 197 if (mfill_file_over_size(dst_vma, dst_addr)) { in mfill_atomic_install_pte() 220 folio_add_new_anon_rmap(folio, dst_vma, dst_addr, RMAP_EXCLUSIVE); in mfill_atomic_install_pte() 230 set_pte_at(dst_mm, dst_addr, dst_pte, _dst_pte); in mfill_atomic_install_pte() 233 update_mmu_cache(dst_vma, dst_addr, dst_pt in mfill_atomic_install_pte() 243 mfill_atomic_pte_copy(pmd_t * dst_pmd,struct vm_area_struct * dst_vma,unsigned long dst_addr,unsigned long src_addr,uffd_flags_t flags,struct folio ** foliop) mfill_atomic_pte_copy() argument 319 mfill_atomic_pte_zeroed_folio(pmd_t * dst_pmd,struct vm_area_struct * dst_vma,unsigned long dst_addr) mfill_atomic_pte_zeroed_folio() argument 351 mfill_atomic_pte_zeropage(pmd_t * dst_pmd,struct vm_area_struct * dst_vma,unsigned long dst_addr) mfill_atomic_pte_zeropage() argument 386 mfill_atomic_pte_continue(pmd_t * dst_pmd,struct vm_area_struct * dst_vma,unsigned long dst_addr,uffd_flags_t flags) mfill_atomic_pte_continue() argument 430 mfill_atomic_pte_poison(pmd_t * dst_pmd,struct vm_area_struct * dst_vma,unsigned long dst_addr,uffd_flags_t flags) mfill_atomic_pte_poison() argument 503 unsigned long src_addr, dst_addr; mfill_atomic_hugetlb() local 662 mfill_atomic_pte(pmd_t * dst_pmd,struct vm_area_struct * dst_vma,unsigned long dst_addr,unsigned long src_addr,uffd_flags_t flags,struct folio ** foliop) mfill_atomic_pte() argument 714 unsigned long src_addr, dst_addr; mfill_atomic() local 1075 move_present_ptes(struct mm_struct * mm,struct vm_area_struct * dst_vma,struct vm_area_struct * src_vma,unsigned long dst_addr,unsigned long src_addr,pte_t * dst_pte,pte_t * src_pte,pte_t orig_dst_pte,pte_t orig_src_pte,pmd_t * dst_pmd,pmd_t dst_pmdval,spinlock_t * dst_ptl,spinlock_t * src_ptl,struct folio ** first_src_folio,unsigned long len) move_present_ptes() argument 1155 move_swap_pte(struct mm_struct * mm,struct vm_area_struct * dst_vma,unsigned long dst_addr,unsigned long src_addr,pte_t * dst_pte,pte_t * src_pte,pte_t orig_dst_pte,pte_t orig_src_pte,pmd_t * dst_pmd,pmd_t dst_pmdval,spinlock_t * dst_ptl,spinlock_t * src_ptl,struct folio * src_folio,struct swap_info_struct * si,swp_entry_t entry) move_swap_pte() argument 1221 move_zeropage_pte(struct mm_struct * mm,struct vm_area_struct * dst_vma,struct vm_area_struct * src_vma,unsigned long dst_addr,unsigned long src_addr,pte_t * dst_pte,pte_t * src_pte,pte_t orig_dst_pte,pte_t orig_src_pte,pmd_t * dst_pmd,pmd_t dst_pmdval,spinlock_t * dst_ptl,spinlock_t * src_ptl) move_zeropage_pte() argument 1254 move_pages_ptes(struct mm_struct * mm,pmd_t * dst_pmd,pmd_t * src_pmd,struct vm_area_struct * dst_vma,struct vm_area_struct * src_vma,unsigned long dst_addr,unsigned long src_addr,unsigned long len,__u64 mode) move_pages_ptes() argument 1517 move_splits_huge_pmd(unsigned long dst_addr,unsigned long src_addr,unsigned long src_end) move_splits_huge_pmd() argument 1525 move_splits_huge_pmd(unsigned long dst_addr,unsigned long src_addr,unsigned long src_end) move_splits_huge_pmd() argument 1775 unsigned long src_addr, dst_addr, src_end; move_pages() local [all...] |