| /linux/mm/ |
| H A D | userfaultfd.c | 1155 spinlock_t *dst_ptl, spinlock_t *src_ptl, in move_swap_pte() 1166 double_pt_lock(dst_ptl, src_ptl); in move_swap_pte() 1226 double_pt_unlock(dst_ptl, src_ptl); in move_zeropage_pte() 1235 spinlock_t *dst_ptl, spinlock_t *src_ptl, in move_zeropage_pte() 1248 double_pt_lock(dst_ptl, src_ptl); 1252 double_pt_unlock(dst_ptl, src_ptl); in move_pages_ptes() 1276 double_pt_unlock(dst_ptl, src_ptl); in move_pages_ptes() 1285 double_pt_unlock(dst_ptl, src_ptl); in move_pages_ptes() 1297 spinlock_t *dst_ptl, spinlock_t *src_ptl) in move_pages_ptes() 1301 double_pt_lock(dst_ptl, src_ptl); in move_pages_ptes() 1079 move_present_ptes(struct mm_struct * mm,struct vm_area_struct * dst_vma,struct vm_area_struct * src_vma,unsigned long dst_addr,unsigned long src_addr,pte_t * dst_pte,pte_t * src_pte,pte_t orig_dst_pte,pte_t orig_src_pte,pmd_t * dst_pmd,pmd_t dst_pmdval,spinlock_t * dst_ptl,spinlock_t * src_ptl,struct folio ** first_src_folio,unsigned long len) move_present_ptes() argument 1159 move_swap_pte(struct mm_struct * mm,struct vm_area_struct * dst_vma,unsigned long dst_addr,unsigned long src_addr,pte_t * dst_pte,pte_t * src_pte,pte_t orig_dst_pte,pte_t orig_src_pte,pmd_t * dst_pmd,pmd_t dst_pmdval,spinlock_t * dst_ptl,spinlock_t * src_ptl,struct folio * src_folio,struct swap_info_struct * si,swp_entry_t entry) move_swap_pte() argument 1225 move_zeropage_pte(struct mm_struct * mm,struct vm_area_struct * dst_vma,struct vm_area_struct * src_vma,unsigned long dst_addr,unsigned long src_addr,pte_t * dst_pte,pte_t * src_pte,pte_t orig_dst_pte,pte_t orig_src_pte,pmd_t * dst_pmd,pmd_t dst_pmdval,spinlock_t * dst_ptl,spinlock_t * src_ptl) move_zeropage_pte() argument 1260 spinlock_t *src_ptl, *dst_ptl; move_pages_ptes() local [all...] |
| H A D | huge_memory.c | 1930 spinlock_t *dst_ptl, *src_ptl; in copy_huge_pmd() 1941 src_ptl = pmd_lockptr(src_mm, src_pmd); in copy_huge_pmd() 1942 spin_lock_nested(src_ptl, SINGLE_DEPTH_NESTING); in copy_huge_pmd() 1965 src_ptl = pmd_lockptr(src_mm, src_pmd); in touch_pud() 1966 spin_lock_nested(src_ptl, SINGLE_DEPTH_NESTING); in touch_pud() 2007 spin_unlock(src_ptl); in huge_pud_set_accessed() 2026 spin_unlock(src_ptl); in huge_pmd_set_accessed() 2050 spinlock_t *dst_ptl, *src_ptl; in do_huge_zero_wp_pmd() 2055 src_ptl = pud_lockptr(src_mm, src_pud); in do_huge_zero_wp_pmd() 2056 spin_lock_nested(src_ptl, SINGLE_DEPTH_NESTIN in do_huge_zero_wp_pmd() 1853 spinlock_t *dst_ptl, *src_ptl; copy_huge_pmd() local 1973 spinlock_t *dst_ptl, *src_ptl; copy_huge_pud() local 2713 spinlock_t *src_ptl, *dst_ptl; move_pages_huge_pmd() local [all...] |
| H A D | hugetlb.c | 4904 spinlock_t *src_ptl, *dst_ptl; in copy_hugetlb_page_range() 4925 src_ptl = huge_pte_lockptr(h, src, src_pte); in copy_hugetlb_page_range() 4926 spin_lock_nested(src_ptl, SINGLE_DEPTH_NESTING); in copy_hugetlb_page_range() 4984 spin_unlock(src_ptl); in copy_hugetlb_page_range() 5003 src_ptl = huge_pte_lockptr(h, src, src_pte); in copy_hugetlb_page_range() 5004 spin_lock_nested(src_ptl, SINGLE_DEPTH_NESTING); in copy_hugetlb_page_range() 5038 spin_unlock(src_ptl); in copy_hugetlb_page_range() 5059 spinlock_t *src_ptl, *dst_ptl; in copy_hugetlb_page_range() 5063 src_ptl = huge_pte_lockptr(h, mm, src_pte); in copy_hugetlb_page_range() 5069 if (src_ptl ! in move_huge_pte() 4920 spinlock_t *src_ptl, *dst_ptl; copy_hugetlb_page_range() local 5075 spinlock_t *src_ptl, *dst_ptl; move_huge_pte() local [all...] |
| H A D | memory.c | 1218 spinlock_t *src_ptl, *dst_ptl; in folio_prealloc() 1250 &src_ptl); in copy_pte_range() 1256 spin_lock_nested(src_ptl, SINGLE_DEPTH_NESTING); in copy_pte_range() 1271 spin_needbreak(src_ptl) || spin_needbreak(dst_ptl)) in copy_pte_range() 1329 pte_unmap_unlock(orig_src_pte, src_ptl); in copy_pte_range() 1231 spinlock_t *src_ptl, *dst_ptl; copy_pte_range() local
|