| /linux/fs/orangefs/ |
| H A D | inode.c | 44 off = folio_pos(folio); in orangefs_writepage_locked() 272 off = folio_pos(folio); in orangefs_read_folio() 366 if (pos == folio_pos(folio) && in orangefs_write_end() 390 } else if (folio_pos(folio) + offset <= wr->pos && in orangefs_invalidate_folio() 391 wr->pos + wr->len <= folio_pos(folio) + offset + length) { in orangefs_invalidate_folio() 397 } else if (wr->pos < folio_pos(folio) + offset && in orangefs_invalidate_folio() 398 wr->pos + wr->len <= folio_pos(folio) + offset + length && in orangefs_invalidate_folio() 399 folio_pos(folio) + offset < wr->pos + wr->len) { in orangefs_invalidate_folio() 401 x = wr->pos + wr->len - (folio_pos(folio) + offset); in orangefs_invalidate_folio() 407 } else if (folio_pos(folio) + offset <= wr->pos && in orangefs_invalidate_folio() [all …]
|
| /linux/fs/iomap/ |
| H A D | buffered-io.c | 157 *range_start = folio_pos(folio) + (start_blk << inode->i_blkbits); in ifs_find_dirty_range() 416 folio_pos(folio) + off, len, error, in iomap_finish_folio_read() 586 .pos = folio_pos(folio), in iomap_read_folio() 727 trace_iomap_release_folio(folio->mapping->host, folio_pos(folio), in iomap_release_folio() 745 folio_pos(folio) + offset, len); in iomap_invalidate_folio() 804 if (!(iter->flags & IOMAP_UNSHARE) && pos <= folio_pos(folio) && in __iomap_write_begin() 915 WARN_ON_ONCE(pos < folio_pos(folio)); in iomap_trim_folio_range() 916 WARN_ON_ONCE(pos >= folio_pos(folio) + fsize); in iomap_trim_folio_range() 995 if (folio_pos(folio) > iter->pos) { in iomap_write_begin() 996 len = min_t(u64, folio_pos(folio) - iter->pos, in iomap_write_begin() [all …]
|
| /linux/fs/netfs/ |
| H A D | read_pgpriv2.c | 23 loff_t fpos = folio_pos(folio), i_size; in netfs_pgpriv2_copy_folio() 105 creq = netfs_create_write_req(rreq->mapping, NULL, folio_pos(folio), in netfs_pgpriv2_begin_copy_to_cache() 194 fpos = folio_pos(folio); in netfs_pgpriv2_unlock_copied_folios()
|
| H A D | buffered_read.c | 415 rreq = netfs_alloc_request(mapping, file, folio_pos(folio), flen, NETFS_READ_GAPS); in netfs_read_gaps() 510 folio_pos(folio), folio_size(folio), in netfs_read_folio() 665 folio_pos(folio), folio_size(folio), in netfs_write_begin() 722 unsigned long long start = folio_pos(folio); in netfs_prefetch_for_write()
|
| H A D | write_issue.c | 331 loff_t fpos = folio_pos(folio), i_size; in netfs_write_folio() 573 wreq = netfs_create_write_req(mapping, NULL, folio_pos(folio), NETFS_WRITEBACK); in netfs_writepages() 587 WARN_ON_ONCE(wreq && folio_pos(folio) < atomic64_read(&wreq->issued_to)); in netfs_writepages() 712 loff_t fpos = folio_pos(folio); in netfs_write_folio_single()
|
| H A D | write_collect.c | 71 fend = folio_pos(folio) + finfo->dirty_offset + finfo->dirty_len; in netfs_folio_written_back() 149 fpos = folio_pos(folio); in netfs_writeback_unlock_folios()
|
| H A D | buffered_write.c | 186 fpos = folio_pos(folio); in netfs_perform_write() 537 folio_pos(folio), in netfs_page_mkwrite()
|
| H A D | read_collect.c | 49 if (unlikely(folio_pos(folio) < rreq->abandon_to)) { in netfs_unlock_read_folio() 139 fpos = folio_pos(folio); in netfs_read_unlock_folios()
|
| /linux/fs/ |
| H A D | mpage.c | 181 block_in_file = folio_pos(folio) >> blkbits; in do_mpage_readpage() 532 block_in_file = folio_pos(folio) >> blkbits; in mpage_write_folio() 574 if (folio_pos(folio) >= i_size) in mpage_write_folio() 577 if (folio_pos(folio) + length > i_size) { in mpage_write_folio() 586 length = i_size - folio_pos(folio); in mpage_write_folio()
|
| H A D | buffer.c | 1007 sector_t block = div_u64(folio_pos(folio), size); in folio_init_buffers() 1860 block = div_u64(folio_pos(folio), blocksize); in __block_write_full_folio() 2125 block = div_u64(folio_pos(folio), blocksize); in __block_write_begin_int() 2268 size_t start = pos - folio_pos(folio); in block_write_end() 2403 iblock = div_u64(folio_pos(folio), blocksize); in block_read_full_folio() 2625 (folio_pos(folio) >= size)) { in block_page_mkwrite() 2633 if (folio_pos(folio) + end > size) in block_page_mkwrite() 2634 end = size - folio_pos(folio); in block_page_mkwrite() 2736 if (folio_pos(folio) >= i_size) { in block_write_full_folio()
|
| /linux/fs/btrfs/ |
| H A D | extent_io.c | 335 range_start = max_t(u64, folio_pos(folio), start); in lock_delalloc_folios() 390 orig_end <= folio_pos(locked_folio))); in find_lock_delalloc_range() 493 ASSERT(folio_pos(folio) <= start && in end_folio_read() 528 u64 start = folio_pos(folio) + fi.offset; in end_bbio_data_write() 558 btrfs_folio_set_lock(fs_info, folio, folio_pos(folio), folio_size(folio)); in begin_folio_read() 589 u64 start = folio_pos(folio) + fi.offset; in end_bbio_data_read() 796 loff_t file_offset = folio_pos(folio) + pg_offset; in submit_extent_folio() 1003 u64 start = folio_pos(folio); in btrfs_do_readpage() 1340 const u64 start = folio_pos(folio); in btrfs_read_folio() 1352 if (folio_pos(folio) < i_size_read(vfs_inode)) in btrfs_read_folio() [all …]
|
| H A D | defrag.c | 886 lock_start = folio_pos(folio); in defrag_prepare_one_folio() 1180 start + len <= folio_pos(folio)) in defrag_one_locked_target() 1232 btrfs_lock_extent(&inode->io_tree, folio_pos(folios[0]), cur - 1, &cached_state); in defrag_one_range() 1258 btrfs_unlock_extent(&inode->io_tree, folio_pos(folios[0]), cur - 1, &cached_state); in defrag_one_range()
|
| H A D | zlib.c | 479 folio_pos(dest_folio)); in zlib_decompress() 501 folio_pos(dest_folio), to_copy, destlen); in zlib_decompress()
|
| /linux/fs/ext4/ |
| H A D | move_extent.c | 137 block = folio_pos(folio) >> inode->i_blkbits; in mext_folio_mkuptodate() 227 move_len = umin(folio_pos(folio[0]) + folio_size(folio[0]) - orig_pos, in mext_move_begin() 228 folio_pos(folio[1]) + folio_size(folio[1]) - donor_pos); in mext_move_begin() 276 block = folio_pos(folio) >> inode->i_blkbits; in mext_folio_mkwrite()
|
| /linux/fs/nfs/ |
| H A D | file.c | 302 loff_t fpos = folio_pos(folio); in nfs_truncate_last_folio() 496 trace_nfs_invalidate_folio(inode, folio_pos(folio) + offset, length); in nfs_invalidate_folio() 561 inode->i_ino, folio_pos(folio)); in nfs_launder_folio() 565 trace_nfs_launder_folio_done(inode, folio_pos(folio), in nfs_launder_folio() 652 (long long)folio_pos(folio)); in nfs_vm_page_mkwrite()
|
| /linux/fs/freevxfs/ |
| H A D | vxfs_immed.c | 34 void *src = vip->vii_immed.vi_immed + folio_pos(folio); in vxfs_immed_read_folio()
|
| /linux/fs/ubifs/ |
| H A D | file.c | 242 if (pos == folio_pos(folio) && len >= folio_size(folio)) in write_begin_slow() 434 if (pos == folio_pos(folio) && len >= folio_size(folio)) { in ubifs_write_begin() 995 if (folio_pos(folio) >= i_size) { in ubifs_writepage() 1005 if (folio_pos(folio) + len <= i_size) { in ubifs_writepage() 1006 if (folio_pos(folio) + len > synced_i_size) { in ubifs_writepage() 1029 len = i_size - folio_pos(folio); in ubifs_writepage() 1541 folio_pos(folio) >= i_size_read(inode))) { in ubifs_vm_page_mkwrite()
|
| /linux/fs/erofs/ |
| H A D | fileio.c | 96 loff_t pos = folio_pos(folio), ofs; in erofs_fileio_scan_folio() 109 ofs = folio_pos(folio) + cur - map->m_la; in erofs_fileio_scan_folio()
|
| H A D | fscache.c | 217 folio_pos(folio), folio_size(folio)); in erofs_fscache_meta_read_folio() 229 folio_pos(folio), folio_size(folio)); in erofs_fscache_meta_read_folio() 231 ret = erofs_fscache_read_io_async(ctx->cookie, folio_pos(folio), io); in erofs_fscache_meta_read_folio() 331 folio_pos(folio), folio_size(folio)); in erofs_fscache_read_folio()
|
| /linux/fs/minix/ |
| H A D | dir.c | 262 pos = folio_pos(folio) + offset_in_folio(folio, p); in minix_add_link() 289 loff_t pos = folio_pos(folio) + offset_in_folio(folio, de); in minix_delete_entry() 412 loff_t pos = folio_pos(folio) + offset_in_folio(folio, de); in minix_set_link()
|
| /linux/fs/vboxsf/ |
| H A D | file.c | 231 loff_t off = folio_pos(folio); in vboxsf_read_folio() 279 loff_t off = folio_pos(folio); in vboxsf_writepages()
|
| /linux/mm/ |
| H A D | truncate.c | 219 loff_t pos = folio_pos(folio); in truncate_inode_partial_folio() 876 offset = from - folio_pos(folio); in pagecache_isize_extended() 877 end = min_t(unsigned int, to - folio_pos(folio), in pagecache_isize_extended()
|
| /linux/fs/ocfs2/ |
| H A D | mmap.c | 53 loff_t pos = folio_pos(folio); in __ocfs2_page_mkwrite()
|
| H A D | aops.c | 266 loff_t start = folio_pos(folio); in ocfs2_read_folio() 555 u64 offset = folio_pos(folio) + block_start; in ocfs2_should_read_blk() 936 new = new | ((i_size_read(inode) <= folio_pos(folio)) && in ocfs2_prepare_folio_for_write() 937 (folio_pos(folio) <= user_pos)); in ocfs2_prepare_folio_for_write() 1960 if (new_isize > folio_pos(wc->w_target_folio)) in ocfs2_write_end_nolock() 2004 loff_t start_byte = folio_pos(folio) + from; in ocfs2_write_end_nolock()
|
| /linux/fs/jfs/ |
| H A D | jfs_metapage.c | 352 folio_pos(folio)); in last_read_complete() 444 page_start = folio_pos(folio) >> inode->i_blkbits; in metapage_write_folio() 577 page_start = folio_pos(folio) >> inode->i_blkbits; in metapage_read_folio()
|