| /linux/mm/ |
| H A D | userfaultfd.c | 1103 static inline bool is_pte_pages_stable(pte_t *dst_pte, pte_t *src_pte, in move_present_ptes() 1107 return pte_same(ptep_get(src_pte), orig_src_pte) && in move_present_ptes() 1121 pte_t *src_pte, pte_t *dst_pte) in move_present_ptes() 1130 orig_src_pte = ptep_get(src_pte); in move_present_ptes() 1152 pte_t *dst_pte, pte_t *src_pte, in move_present_ptes() 1168 if (!is_pte_pages_stable(dst_pte, src_pte, orig_dst_pte, orig_src_pte, in move_swap_pte() 1185 orig_src_pte = ptep_get_and_clear(mm, src_addr, src_pte); in move_swap_pte() 1188 set_pte_at(mm, src_addr, src_pte, orig_src_pte); in move_swap_pte() 1210 src_pte++; in move_swap_pte() 1214 src_pte, dst_pt in move_swap_pte() 1027 is_pte_pages_stable(pte_t * dst_pte,pte_t * src_pte,pte_t orig_dst_pte,pte_t orig_src_pte,pmd_t * dst_pmd,pmd_t dst_pmdval) is_pte_pages_stable() argument 1045 check_ptes_for_batched_move(struct vm_area_struct * src_vma,unsigned long src_addr,pte_t * src_pte,pte_t * dst_pte) check_ptes_for_batched_move() argument 1076 move_present_ptes(struct mm_struct * mm,struct vm_area_struct * dst_vma,struct vm_area_struct * src_vma,unsigned long dst_addr,unsigned long src_addr,pte_t * dst_pte,pte_t * src_pte,pte_t orig_dst_pte,pte_t orig_src_pte,pmd_t * dst_pmd,pmd_t dst_pmdval,spinlock_t * dst_ptl,spinlock_t * src_ptl,struct folio ** first_src_folio,unsigned long len) move_present_ptes() argument 1156 move_swap_pte(struct mm_struct * mm,struct vm_area_struct * dst_vma,unsigned long dst_addr,unsigned long src_addr,pte_t * dst_pte,pte_t * src_pte,pte_t orig_dst_pte,pte_t orig_src_pte,pmd_t * dst_pmd,pmd_t dst_pmdval,spinlock_t * dst_ptl,spinlock_t * src_ptl,struct folio * src_folio,struct swap_info_struct * si,swp_entry_t entry) move_swap_pte() argument 1222 move_zeropage_pte(struct mm_struct * mm,struct vm_area_struct * dst_vma,struct vm_area_struct * src_vma,unsigned long dst_addr,unsigned long src_addr,pte_t * dst_pte,pte_t * src_pte,pte_t orig_dst_pte,pte_t orig_src_pte,pmd_t * dst_pmd,pmd_t dst_pmdval,spinlock_t * dst_ptl,spinlock_t * src_ptl) move_zeropage_pte() argument 1261 pte_t *src_pte = NULL; move_pages_ptes() local [all...] |
| H A D | memory.c | 926 pte_t *dst_pte, pte_t *src_pte, struct vm_area_struct *dst_vma, in try_restore_exclusive_pte() 930 pte_t orig_pte = ptep_get(src_pte); 951 set_pte_at(src_mm, addr, src_pte, pte); in copy_nonpresent_pte() 973 set_pte_at(src_mm, addr, src_pte, pte); in copy_nonpresent_pte() 1007 set_pte_at(src_mm, addr, src_pte, pte); in copy_nonpresent_pte() 1017 if (try_restore_exclusive_pte(src_vma, addr, src_pte, orig_pte)) in copy_nonpresent_pte() 1048 pte_t *dst_pte, pte_t *src_pte, unsigned long addr, int *rss, 1075 if (userfaultfd_pte_wp(dst_vma, ptep_get(src_pte))) in copy_present_page() 1083 struct vm_area_struct *src_vma, pte_t *dst_pte, pte_t *src_pte, in copy_present_page() 1090 wrprotect_ptes(src_mm, addr, src_pte, n in copy_present_page() 939 copy_nonpresent_pte(struct mm_struct * dst_mm,struct mm_struct * src_mm,pte_t * dst_pte,pte_t * src_pte,struct vm_area_struct * dst_vma,struct vm_area_struct * src_vma,unsigned long addr,int * rss) copy_nonpresent_pte() argument 1061 copy_present_page(struct vm_area_struct * dst_vma,struct vm_area_struct * src_vma,pte_t * dst_pte,pte_t * src_pte,unsigned long addr,int * rss,struct folio ** prealloc,struct page * page) copy_present_page() argument 1096 __copy_present_ptes(struct vm_area_struct * dst_vma,struct vm_area_struct * src_vma,pte_t * dst_pte,pte_t * src_pte,pte_t pte,unsigned long addr,int nr) __copy_present_ptes() argument 1127 copy_present_ptes(struct vm_area_struct * dst_vma,struct vm_area_struct * src_vma,pte_t * dst_pte,pte_t * src_pte,pte_t pte,unsigned long addr,int max_nr,int * rss,struct folio ** prealloc) copy_present_ptes() argument 1228 pte_t *src_pte, *dst_pte; copy_pte_range() local [all...] |
| H A D | hugetlb.c | 4873 pte_t *src_pte, *dst_pte, entry; in hugetlb_install_folio() 4905 src_pte = hugetlb_walk(src_vma, addr, sz); in copy_hugetlb_page_range() 4906 if (!src_pte) { in copy_hugetlb_page_range() 4925 src_ptl = huge_pte_lockptr(h, src, src_pte); in copy_hugetlb_page_range() 4927 entry = huge_ptep_get(src_vma->vm_mm, addr, src_pte); in copy_hugetlb_page_range() 4952 set_huge_pte_at(src, addr, src_pte, entry, sz); in copy_hugetlb_page_range() 4964 entry = huge_ptep_get(src_vma->vm_mm, addr, src_pte); in copy_hugetlb_page_range() 5003 src_ptl = huge_pte_lockptr(h, src, src_pte); in copy_hugetlb_page_range() 5005 entry = huge_ptep_get(src_vma->vm_mm, addr, src_pte); in copy_hugetlb_page_range() 5026 huge_ptep_set_wrprotect(src, addr, src_pte); in copy_hugetlb_page_range() 4889 pte_t *src_pte, *dst_pte, entry; copy_hugetlb_page_range() local 5069 move_huge_pte(struct vm_area_struct * vma,unsigned long old_addr,unsigned long new_addr,pte_t * src_pte,pte_t * dst_pte,unsigned long sz) move_huge_pte() argument 5118 pte_t *src_pte, *dst_pte; move_hugetlb_page_tables() local [all...] |