Lines Matching refs:fault_page
83 struct page *fault_page) in migrate_vma_split_folio() argument
86 struct folio *fault_folio = fault_page ? page_folio(fault_page) : NULL; in migrate_vma_split_folio()
103 new_fault_folio = fault_page ? page_folio(fault_page) : NULL; in migrate_vma_split_folio()
257 struct folio *fault_folio = migrate->fault_page ? in migrate_vma_collect_pmd()
258 page_folio(migrate->fault_page) : NULL; in migrate_vma_collect_pmd()
319 migrate->fault_page); in migrate_vma_collect_pmd()
362 migrate->fault_page); in migrate_vma_collect_pmd()
537 static bool migrate_vma_check_page(struct page *page, struct page *fault_page) in migrate_vma_check_page() argument
546 int extra = 1 + (page == fault_page); in migrate_vma_check_page()
568 struct page *fault_page) in migrate_device_unmap() argument
570 struct folio *fault_folio = fault_page ? in migrate_device_unmap()
571 page_folio(fault_page) : NULL; in migrate_device_unmap()
618 !migrate_vma_check_page(page, fault_page)) { in migrate_device_unmap()
668 migrate->fault_page); in migrate_vma_unmap()
753 if (args->fault_page && !is_device_private_page(args->fault_page)) in migrate_vma_setup()
755 if (args->fault_page && !PageLocked(args->fault_page)) in migrate_vma_setup()
1221 if (migrate && migrate->fault_page == page) in __migrate_device_pages()
1274 struct page *fault_page) in __migrate_device_finalize() argument
1276 struct folio *fault_folio = fault_page ? in __migrate_device_finalize()
1277 page_folio(fault_page) : NULL; in __migrate_device_finalize()
1354 migrate->fault_page); in migrate_vma_finalize()