Searched refs:folio_mapping (Results 1 – 22 of 22) sorted by relevance
2803 struct address_space *mapping = folio_mapping(folio); in folio_mark_dirty()2862 struct address_space *mapping = folio_mapping(folio); in __folio_cancel_dirty()2897 struct address_space *mapping = folio_mapping(folio); in folio_clear_dirty_for_io()2982 struct address_space *mapping = folio_mapping(folio); in __folio_end_writeback()3019 struct address_space *mapping = folio_mapping(folio); in __folio_start_writeback()3088 trace_folio_wait_writeback(folio, folio_mapping(folio)); in folio_wait_writeback()3110 trace_folio_wait_writeback(folio, folio_mapping(folio)); in folio_wait_writeback_killable()3134 if (mapping_stable_writes(folio_mapping(folio))) in folio_wait_stable()
30 mapping = folio_mapping(folio); in hwpoison_filter_dev()
1041 mapping = folio_mapping(folio); in me_pagecache_clean()1077 struct address_space *mapping = folio_mapping(folio); in me_pagecache_dirty()1163 mapping = folio_mapping(folio); in me_huge_page()1536 mapping = folio_mapping(folio); in unmap_poisoned_folio()2711 if (folio_mapping(folio)) { in unpoison_memory()2819 ret = mapping_evict_folio(folio_mapping(folio), folio); in soft_offline_in_use_page()
73 struct address_space *mapping = folio_mapping(folio); in __dump_folio()
692 struct address_space *folio_mapping(const struct folio *folio) in folio_mapping() function709 EXPORT_SYMBOL(folio_mapping);
553 if (folio_mapping(folio)) in migrate_vma_check_page()1165 mapping = folio_mapping(folio); in __migrate_device_pages()
499 ret = !mapping_unevictable(folio_mapping(folio)) && in folio_evictable()521 struct address_space *mapping = folio_mapping(folio); in folio_needs_release()
496 if (folio_mapping(folio) == mapping) in handle_write_error()713 BUG_ON(mapping != folio_mapping(folio)); in __remove_mapping()981 mapping = folio_mapping(folio); in folio_check_dirty_writeback()1206 mapping = folio_mapping(folio); in shrink_folio_list()1388 mapping = folio_mapping(folio); in shrink_folio_list()1453 mapping = folio_mapping(folio); in shrink_folio_list()
1092 struct address_space *mapping = folio_mapping(src); in move_to_new_folio()1489 if (hugetlb_folio_subpool(src) && !folio_mapping(src)) { in unmap_and_move_huge_page()
1093 mapping = folio_mapping(folio); in folio_mkclean()2901 VM_WARN_ON_FOLIO(folio && mapping != folio_mapping(folio), folio); in __rmap_walk_file()
1083 mapping = folio_mapping(folio); in isolate_migratepages_block()1141 mapping = folio_mapping(folio); in isolate_migratepages_block()
180 bool page_in_cache = folio_mapping(folio); in mfill_atomic_install_pte()
2009 if (folio_mapping(folio) != mapping) { in collapse_file()
1861 struct address_space *mapping = folio_mapping(folio); in hugetlb_folio_mapping_lock_write()
1211 if (folio_mapping(folio) != mapping) { in shmem_undo_range() 1687 * e.g. folio_mapping(folio) might give an unexpected answer. in shmem_writeout()
111 :functions: folio_mapping
330 mapping = folio_mapping(folio); in bh_get_inode_and_lblk_num()
551 struct address_space *folio_mapping(const struct folio *folio);570 return folio_mapping(folio); in folio_flush_mapping()
276 filemap_dirty_folio(folio_mapping(dfolio), dfolio); in nilfs_copy_dirty_pages()
766 filemap_dirty_folio(folio_mapping(folio), folio); in nfs_mark_request_dirty()1749 filemap_dirty_folio(folio_mapping(folio), folio); in nfs_commit_resched_write()
290 struct address_space *mapping = folio_mapping(folio);
72 mapping = folio_mapping(folio); in page_cache_pipe_buf_try_steal()