Searched refs:folio_mapping (Results 1 – 21 of 21) sorted by relevance
2759 struct address_space *mapping = folio_mapping(folio);2818 struct address_space *mapping = folio_mapping(folio); in folio_mark_dirty() 2853 struct address_space *mapping = folio_mapping(folio);2938 struct address_space *mapping = folio_mapping(folio); in folio_clear_dirty_for_io() 2975 struct address_space *mapping = folio_mapping(folio); in wb_inode_writeback_end() 3044 trace_folio_wait_writeback(folio, folio_mapping(folio)); in __folio_start_writeback() 3066 trace_folio_wait_writeback(folio, folio_mapping(folio)); in __folio_start_writeback() 3090 if (mapping_stable_writes(folio_mapping(folio))) in folio_wait_writeback()
30 mapping = folio_mapping(folio); in hwpoison_filter_dev()
1045 mapping = folio_mapping(folio); in me_pagecache_clean()1081 struct address_space *mapping = folio_mapping(folio); in me_pagecache_dirty()1167 mapping = folio_mapping(folio); in me_huge_page()1540 mapping = folio_mapping(folio); in unmap_poisoned_folio()2730 if (folio_mapping(folio)) { in unpoison_memory()2838 ret = mapping_evict_folio(folio_mapping(folio), folio); in soft_offline_in_use_page()
73 struct address_space *mapping = folio_mapping(folio); in __dump_folio()
692 struct address_space *folio_mapping(const struct folio *folio) in folio_mapping() function709 EXPORT_SYMBOL(folio_mapping);
553 if (folio_mapping(folio)) in migrate_vma_check_page() 1165 mapping = folio_mapping(folio); in __migrate_device_pages()
568 ret = !mapping_unevictable(folio_mapping(folio)) && in folio_evictable()598 struct address_space *mapping = folio_mapping(folio); in folio_needs_release()
497 if (folio_mapping(folio) == mapping) in handle_write_error()714 BUG_ON(mapping != folio_mapping(folio)); in __remove_mapping()981 mapping = folio_mapping(folio); in folio_check_dirty_writeback()1213 mapping = folio_mapping(folio); in shrink_folio_list()1395 mapping = folio_mapping(folio); in shrink_folio_list()1460 mapping = folio_mapping(folio); in shrink_folio_list()
1093 struct address_space *mapping = folio_mapping(src); in move_to_new_folio()1491 if (hugetlb_folio_subpool(src) && !folio_mapping(src)) { in unmap_and_move_huge_page()
1201 mapping = folio_mapping(folio); in folio_mkclean()3010 VM_WARN_ON_FOLIO(folio && mapping != folio_mapping(folio), folio); in __rmap_walk_file()
1083 mapping = folio_mapping(folio); in isolate_migratepages_block()1141 mapping = folio_mapping(folio); in isolate_migratepages_block()
180 bool page_in_cache = folio_mapping(folio); in mfill_atomic_install_pte()
2001 if (folio_mapping(folio) != mapping) { in collapse_file()
1239 if (folio_mapping(folio) != mapping) { in shmem_undo_range()
1833 struct address_space *mapping = folio_mapping(folio); in hugetlb_folio_mapping_lock_write()
330 mapping = folio_mapping(folio); in bh_get_inode_and_lblk_num()
562 struct address_space *folio_mapping(const struct folio *folio);581 return folio_mapping(folio); in folio_flush_mapping()
276 filemap_dirty_folio(folio_mapping(dfolio), dfolio); in nilfs_copy_dirty_pages()
766 filemap_dirty_folio(folio_mapping(folio), folio); in nfs_mark_request_dirty()1749 filemap_dirty_folio(folio_mapping(folio), folio); in nfs_commit_resched_write()
289 struct address_space *mapping = folio_mapping(folio);
72 mapping = folio_mapping(folio); in page_cache_pipe_buf_try_steal()