| /linux/fs/nilfs2/ |
| H A D | mdt.c | 83 bh = nilfs_grab_buffer(inode, inode->i_mapping, block, 0); in nilfs_mdt_create_block() 123 bh = nilfs_grab_buffer(inode, inode->i_mapping, blkoff, 0); in nilfs_mdt_submit_block() 362 folio = filemap_lock_folio(inode->i_mapping, index); in nilfs_mdt_forget_block() 380 invalidate_inode_pages2_range(inode->i_mapping, index, index) != 0) in nilfs_mdt_forget_block() 464 mapping_set_gfp_mask(inode->i_mapping, gfp_mask); in nilfs_mdt_init() 468 inode->i_mapping->a_ops = &def_mdt_aops; in nilfs_mdt_init() 554 ret = nilfs_copy_dirty_pages(s_inode->i_mapping, inode->i_mapping); in nilfs_mdt_save_to_shadow_map() 558 ret = nilfs_copy_dirty_pages(NILFS_I(s_inode)->i_assoc_inode->i_mapping, in nilfs_mdt_save_to_shadow_map() 559 ii->i_assoc_inode->i_mapping); in nilfs_mdt_save_to_shadow_map() 575 folio = filemap_grab_folio(shadow->inode->i_mapping, in nilfs_mdt_freeze_buffer() [all …]
|
| H A D | gcinode.c | 62 bh = nilfs_grab_buffer(inode, inode->i_mapping, blkoff, 0); in nilfs_gccache_submit_read_data() 126 ret = nilfs_btnode_submit_block(btnc_inode->i_mapping, vbn ? : pbn, pbn, in nilfs_gccache_submit_read_node() 161 mapping_set_gfp_mask(inode->i_mapping, GFP_NOFS); in nilfs_init_gcinode() 162 inode->i_mapping->a_ops = &nilfs_buffer_cache_aops; in nilfs_init_gcinode() 183 nilfs_btnode_cache_clear(ii->i_assoc_inode->i_mapping); in nilfs_remove_all_gcinodes()
|
| H A D | inode.c | 314 mapping_set_gfp_mask(inode->i_mapping, in nilfs_new_inode() 315 mapping_gfp_constraint(inode->i_mapping, ~__GFP_FS)); in nilfs_new_inode() 468 inode->i_mapping->a_ops = &nilfs_aops; in __nilfs_read_inode() 472 inode->i_mapping->a_ops = &nilfs_aops; in __nilfs_read_inode() 476 inode->i_mapping->a_ops = &nilfs_aops; in __nilfs_read_inode() 494 mapping_set_gfp_mask(inode->i_mapping, in __nilfs_read_inode() 495 mapping_gfp_constraint(inode->i_mapping, ~__GFP_FS)); in __nilfs_read_inode() 694 mapping_set_gfp_mask(s_inode->i_mapping, GFP_NOFS); in nilfs_iget_for_shadow() 695 s_inode->i_mapping->a_ops = &nilfs_buffer_cache_aops; in nilfs_iget_for_shadow() 813 block_truncate_page(inode->i_mapping, inode->i_size, nilfs_get_block); in nilfs_truncate()
|
| /linux/virt/kvm/ |
| H A D | guest_memfd.c | 129 folio = __filemap_get_folio(inode->i_mapping, index, in kvm_gmem_get_folio() 135 folio = __filemap_get_folio_mpol(inode->i_mapping, index, in kvm_gmem_get_folio() 137 mapping_gfp_mask(inode->i_mapping), policy); in kvm_gmem_get_folio() 205 kvm_gmem_for_each_file(f, inode->i_mapping) in kvm_gmem_invalidate_begin() 226 kvm_gmem_for_each_file(f, inode->i_mapping) in kvm_gmem_invalidate_end() 239 filemap_invalidate_lock(inode->i_mapping); in kvm_gmem_punch_hole() 243 truncate_inode_pages_range(inode->i_mapping, offset, offset + len - 1); in kvm_gmem_punch_hole() 247 filemap_invalidate_unlock(inode->i_mapping); in kvm_gmem_punch_hole() 254 struct address_space *mapping = inode->i_mapping; in kvm_gmem_allocate() 345 filemap_invalidate_lock(inode->i_mapping); in kvm_gmem_release() [all …]
|
| /linux/fs/coda/ |
| H A D | file.c | 185 if (coda_inode->i_mapping == &coda_inode->i_data) in coda_file_mmap() 186 coda_inode->i_mapping = host_inode->i_mapping; in coda_file_mmap() 190 else if (coda_inode->i_mapping != host_inode->i_mapping) { in coda_file_mmap() 279 if (coda_inode->i_mapping == &host_inode->i_data) { in coda_release() 282 coda_inode->i_mapping = &coda_inode->i_data; in coda_release() 306 err = filemap_write_and_wait_range(coda_inode->i_mapping, start, end); in coda_fsync()
|
| /linux/fs/freevxfs/ |
| H A D | vxfs_inode.c | 158 vip->vfs_inode.i_mapping->a_ops = &vxfs_aops; in vxfs_blkiget() 190 pp = vxfs_get_page(ilistp->i_mapping, ino * VXFS_ISIZE / PAGE_SIZE); in __vxfs_iget() 198 vip->vfs_inode.i_mapping->a_ops = &vxfs_aops; in __vxfs_iget() 278 ip->i_mapping->a_ops = aops; in vxfs_iget() 282 ip->i_mapping->a_ops = aops; in vxfs_iget() 287 ip->i_mapping->a_ops = &vxfs_aops; in vxfs_iget()
|
| /linux/fs/netfs/ |
| H A D | locking.c | 139 if (inode->i_mapping->nrpages != 0) { in netfs_block_buffered() 140 unmap_mapping_range(inode->i_mapping, 0, 0, 0); in netfs_block_buffered() 141 ret = filemap_fdatawait(inode->i_mapping); in netfs_block_buffered()
|
| /linux/mm/ |
| H A D | secretmem.c | 168 struct address_space *mapping = inode->i_mapping; in secretmem_setattr() 205 mapping_set_gfp_mask(inode->i_mapping, GFP_HIGHUSER); in secretmem_file_create() 206 mapping_set_unevictable(inode->i_mapping); in secretmem_file_create() 209 inode->i_mapping->a_ops = &secretmem_aops; in secretmem_file_create()
|
| /linux/fs/xfs/scrub/ |
| H A D | xfile.c | 78 mapping_set_gfp_mask(inode->i_mapping, GFP_KERNEL); in xfile_create() 142 if (filemap_check_wb_err(inode->i_mapping, 0)) { in xfile_load() 203 if (filemap_check_wb_err(inode->i_mapping, 0)) { in xfile_store() 286 if (filemap_check_wb_err(inode->i_mapping, 0)) { in xfile_get_folio()
|
| /linux/fs/jfs/ |
| H A D | jfs_umount.c | 96 filemap_write_and_wait(sbi->direct_inode->i_mapping); in jfs_umount() 148 filemap_write_and_wait(sbi->direct_inode->i_mapping); in jfs_umount_rw()
|
| /linux/fs/ |
| H A D | drop_caches.c | 32 (mapping_empty(inode->i_mapping) && !need_resched())) { in drop_pagecache_sb() 40 invalidate_mapping_pages(inode->i_mapping, 0, -1); in drop_pagecache_sb()
|
| /linux/fs/ramfs/ |
| H A D | file-nommu.c | 69 gfp_t gfp = mapping_gfp_mask(inode->i_mapping); in ramfs_nommu_expand_for_mapping() 107 ret = add_to_page_cache_lru(page, inode->i_mapping, loop, in ramfs_nommu_expand_for_mapping() 227 nr_folios = filemap_get_folios_contig(inode->i_mapping, &pgoff, in ramfs_nommu_get_unmapped_area()
|
| H A D | inode.c | 65 inode->i_mapping->a_ops = &ram_aops; in ramfs_get_inode() 66 mapping_set_gfp_mask(inode->i_mapping, GFP_HIGHUSER); in ramfs_get_inode() 67 mapping_set_unevictable(inode->i_mapping); in ramfs_get_inode()
|
| /linux/fs/iomap/ |
| H A D | buffered-io.c | 711 return __filemap_get_folio(iter->inode->i_mapping, pos >> PAGE_SHIFT, in iomap_get_folio() 712 fgp, mapping_gfp_mask(iter->inode->i_mapping)); in iomap_get_folio() 854 if (!mapping_large_folio_support(iter->inode->i_mapping)) in __iomap_get_folio() 869 if (unlikely(folio->mapping != iter->inode->i_mapping)) { in __iomap_get_folio() 1036 filemap_dirty_folio(inode->i_mapping, folio); in __iomap_write_end() 1090 struct address_space *mapping = iter->inode->i_mapping; in iomap_write_iter() 1300 folio = filemap_lock_folio(inode->i_mapping, in iomap_write_delalloc_scan() 1390 lockdep_assert_held_write(&inode->i_mapping->invalidate_lock); in iomap_write_delalloc_release() 1395 start_byte = mapping_seek_hole_data(inode->i_mapping, in iomap_write_delalloc_release() 1416 data_end = mapping_seek_hole_data(inode->i_mapping, start_byte, in iomap_write_delalloc_release() [all …]
|
| H A D | seek.c | 16 *hole_pos = mapping_seek_hole_data(iter->inode->i_mapping, in iomap_seek_hole_iter() 64 *hole_pos = mapping_seek_hole_data(iter->inode->i_mapping, in iomap_seek_data_iter()
|
| /linux/fs/hfsplus/ |
| H A D | super.c | 35 inode->i_mapping->a_ops = &hfsplus_btree_aops; in hfsplus_system_read_inode() 39 inode->i_mapping->a_ops = &hfsplus_btree_aops; in hfsplus_system_read_inode() 43 inode->i_mapping->a_ops = &hfsplus_aops; in hfsplus_system_read_inode() 50 inode->i_mapping->a_ops = &hfsplus_btree_aops; in hfsplus_system_read_inode() 261 error = filemap_write_and_wait(sbi->cat_tree->inode->i_mapping); in hfsplus_sync_fs() 262 error2 = filemap_write_and_wait(sbi->ext_tree->inode->i_mapping); in hfsplus_sync_fs() 267 filemap_write_and_wait(sbi->attr_tree->inode->i_mapping); in hfsplus_sync_fs() 271 error2 = filemap_write_and_wait(sbi->alloc_file->i_mapping); in hfsplus_sync_fs()
|
| /linux/fs/fuse/ |
| H A D | dax.c | 663 filemap_invalidate_unlock(inode->i_mapping); in fuse_wait_dax_page() 665 filemap_invalidate_lock(inode->i_mapping); in fuse_wait_dax_page() 777 filemap_invalidate_lock_shared(inode->i_mapping); in __fuse_dax_fault() 782 filemap_invalidate_unlock_shared(inode->i_mapping); in __fuse_dax_fault() 788 filemap_invalidate_unlock_shared(inode->i_mapping); in __fuse_dax_fault() 838 ret = filemap_fdatawrite_range(inode->i_mapping, start_pos, end_pos); in dmap_writeback_invalidate() 845 ret = invalidate_inode_pages2_range(inode->i_mapping, in dmap_writeback_invalidate() 922 filemap_invalidate_lock(inode->i_mapping); in inode_inline_reclaim_one_dmap() 983 filemap_invalidate_unlock(inode->i_mapping); in inode_inline_reclaim_one_dmap() 1084 filemap_invalidate_lock(inode->i_mapping); in lookup_and_reclaim_dmap() [all …]
|
| /linux/fs/xfs/ |
| H A D | xfs_reflink.h | 21 mapping_tagged(inode->i_mapping, PAGECACHE_TAG_DIRTY) || in xfs_can_free_cowblocks() 22 mapping_tagged(inode->i_mapping, PAGECACHE_TAG_WRITEBACK) || in xfs_can_free_cowblocks()
|
| /linux/fs/ext2/ |
| H A D | inode.c | 885 ext2_write_failed(inode->i_mapping, offset + length); in ext2_iomap_end() 1188 WARN_ON(!rwsem_is_locked(&inode->i_mapping->invalidate_lock)); in __ext2_truncate_blocks() 1270 filemap_invalidate_lock(inode->i_mapping); in ext2_truncate_blocks() 1272 filemap_invalidate_unlock(inode->i_mapping); in ext2_truncate_blocks() 1293 error = block_truncate_page(inode->i_mapping, in ext2_setsize() 1298 filemap_invalidate_lock(inode->i_mapping); in ext2_setsize() 1301 filemap_invalidate_unlock(inode->i_mapping); in ext2_setsize() 1305 sync_mapping_buffers(inode->i_mapping); in ext2_setsize() 1382 inode->i_mapping->a_ops = &ext2_dax_aops; in ext2_set_file_ops() 1384 inode->i_mapping->a_ops = &ext2_aops; in ext2_set_file_ops() [all …]
|
| /linux/fs/qnx6/ |
| H A D | inode.c | 185 struct address_space *mapping = root->i_mapping; in qnx6_checkroot() 506 inode->i_mapping->a_ops = &qnx6_aops; in qnx6_private_inode() 538 mapping = sbi->inodes->i_mapping; in qnx6_iget() 566 inode->i_mapping->a_ops = &qnx6_aops; in qnx6_iget() 570 inode->i_mapping->a_ops = &qnx6_aops; in qnx6_iget() 574 inode->i_mapping->a_ops = &qnx6_aops; in qnx6_iget()
|
| /linux/drivers/dax/ |
| H A D | device.c | 363 inode->i_mapping = __dax_inode->i_mapping; in dax_open() 364 inode->i_mapping->host = __dax_inode; in dax_open() 365 inode->i_mapping->a_ops = &dev_dax_aops; in dax_open() 366 filp->f_mapping = inode->i_mapping; in dax_open()
|
| /linux/include/linux/ |
| H A D | backing-dev.h | 242 !lockdep_is_held(&inode->i_mapping->i_pages.xa_lock) && in inode_to_wb() 287 xa_lock_irqsave(&inode->i_mapping->i_pages, cookie->flags); in unlocked_inode_to_wb_begin() 305 xa_unlock_irqrestore(&inode->i_mapping->i_pages, cookie->flags); in unlocked_inode_to_wb_end()
|
| /linux/fs/ocfs2/ |
| H A D | mmap.c | 52 struct address_space *mapping = inode->i_mapping; in __ocfs2_page_mkwrite() 75 if ((folio->mapping != inode->i_mapping) || in __ocfs2_page_mkwrite()
|
| /linux/fs/afs/ |
| H A D | inode.c | 35 mapping_gfp_mask(vnode->netfs.inode.i_mapping)) < 0) in afs_init_new_symlink() 191 inode->i_mapping->a_ops = &afs_file_aops; in afs_inode_init_from_status() 192 mapping_set_large_folios(inode->i_mapping); in afs_inode_init_from_status() 198 inode->i_mapping->a_ops = &afs_dir_aops; in afs_inode_init_from_status() 217 inode->i_mapping->a_ops = &afs_dir_aops; in afs_inode_init_from_status() 219 mapping_set_release_always(inode->i_mapping); in afs_inode_init_from_status() 759 afs_single_writepages(inode->i_mapping, &wbc); in afs_evict_inode() 883 ret = filemap_fdatawait_range(inode->i_mapping, from, to); in afs_setattr()
|
| /linux/fs/btrfs/ |
| H A D | verity.c | 455 truncate_inode_pages(inode->vfs_inode.i_mapping, inode->vfs_inode.i_size); in rollback_verity() 717 folio = __filemap_get_folio(inode->i_mapping, index, FGP_ACCESSED, 0); in btrfs_read_merkle_tree_page() 733 folio = filemap_alloc_folio(mapping_gfp_constraint(inode->i_mapping, ~__GFP_FS), in btrfs_read_merkle_tree_page() 738 ret = filemap_add_folio(inode->i_mapping, folio, index, GFP_NOFS); in btrfs_read_merkle_tree_page()
|