| /linux/fs/netfs/ |
| H A D | read_pgpriv2.c | 23 loff_t fpos = folio_pos(folio), i_size; in netfs_pgpriv2_copy_folio() 105 creq = netfs_create_write_req(rreq->mapping, NULL, folio_pos(folio), in netfs_pgpriv2_begin_copy_to_cache() 194 fpos = folio_pos(folio); in netfs_pgpriv2_unlock_copied_folios()
|
| H A D | buffered_read.c | 414 rreq = netfs_alloc_request(mapping, file, folio_pos(folio), flen, NETFS_READ_GAPS); in netfs_read_gaps() 509 folio_pos(folio), folio_size(folio), in netfs_read_folio() 664 folio_pos(folio), folio_size(folio), in netfs_write_begin() 721 unsigned long long start = folio_pos(folio); in netfs_prefetch_for_write()
|
| H A D | write_issue.c | 330 loff_t fpos = folio_pos(folio), i_size; in netfs_write_folio() 572 wreq = netfs_create_write_req(mapping, NULL, folio_pos(folio), NETFS_WRITEBACK); in netfs_writepages() 586 WARN_ON_ONCE(wreq && folio_pos(folio) < atomic64_read(&wreq->issued_to)); in netfs_writepages() 711 loff_t fpos = folio_pos(folio); in netfs_write_folio_single()
|
| H A D | write_collect.c | 71 fend = folio_pos(folio) + finfo->dirty_offset + finfo->dirty_len; in netfs_folio_written_back() 149 fpos = folio_pos(folio); in netfs_writeback_unlock_folios()
|
| H A D | read_collect.c | 49 if (unlikely(folio_pos(folio) < rreq->abandon_to)) { in netfs_unlock_read_folio() 139 fpos = folio_pos(folio); in netfs_read_unlock_folios()
|
| H A D | misc.c | 221 unsigned long long fpos = folio_pos(folio), end; in netfs_invalidate_folio()
|
| /linux/fs/iomap/ |
| H A D | buffered-io.c | 157 *range_start = folio_pos(folio) + (start_blk << inode->i_blkbits); in ifs_find_dirty_range() 416 folio_pos(folio) + off, len, error, in iomap_finish_folio_read() 591 .pos = folio_pos(folio), in iomap_read_folio() 732 trace_iomap_release_folio(folio->mapping->host, folio_pos(folio), in iomap_release_folio() 750 folio_pos(folio) + offset, len); in iomap_invalidate_folio() 809 if (!(iter->flags & IOMAP_UNSHARE) && pos <= folio_pos(folio) && in __iomap_write_begin() 920 WARN_ON_ONCE(pos < folio_pos(folio)); in iomap_trim_folio_range() 921 WARN_ON_ONCE(pos >= folio_pos(folio) + fsize); in iomap_trim_folio_range() 1000 if (folio_pos(folio) > iter->pos) { in iomap_write_begin() 1001 len = min_t(u64, folio_pos(folio) - iter->pos, in iomap_write_begin() [all …]
|
| /linux/fs/ |
| H A D | mpage.c | 180 block_in_file = folio_pos(folio) >> blkbits; in do_mpage_readpage() 531 block_in_file = folio_pos(folio) >> blkbits; in mpage_write_folio() 573 if (folio_pos(folio) >= i_size) in mpage_write_folio() 576 if (folio_pos(folio) + length > i_size) { in mpage_write_folio() 585 length = i_size - folio_pos(folio); in mpage_write_folio()
|
| /linux/fs/btrfs/ |
| H A D | extent_io.c | 336 range_start = max_t(u64, folio_pos(folio), start); in lock_delalloc_folios() 391 orig_end <= folio_pos(locked_folio))); in find_lock_delalloc_range() 494 ASSERT(folio_pos(folio) <= start && in end_folio_read() 529 u64 start = folio_pos(folio) + fi.offset; in end_bbio_data_write() 552 btrfs_folio_set_lock(fs_info, folio, folio_pos(folio), folio_size(folio)); in begin_folio_read() 583 u64 start = folio_pos(folio) + fi.offset; in end_bbio_data_read() 790 loff_t file_offset = folio_pos(folio) + pg_offset; in submit_extent_folio() 997 u64 start = folio_pos(folio); in btrfs_do_readpage() 1334 const u64 start = folio_pos(folio); in btrfs_read_folio() 1346 if (folio_pos(folio) < i_size_read(vfs_inode)) in btrfs_read_folio() [all …]
|
| H A D | defrag.c | 886 lock_start = folio_pos(folio); in defrag_prepare_one_folio() 1180 start + len <= folio_pos(folio)) in defrag_one_locked_target() 1232 btrfs_lock_extent(&inode->io_tree, folio_pos(folios[0]), cur - 1, &cached_state); in defrag_one_range() 1258 btrfs_unlock_extent(&inode->io_tree, folio_pos(folios[0]), cur - 1, &cached_state); in defrag_one_range()
|
| H A D | zlib.c | 475 folio_pos(dest_folio)); in zlib_decompress() 497 folio_pos(dest_folio), to_copy, destlen); in zlib_decompress()
|
| H A D | zstd.c | 694 folio_pos(dest_folio)); in zstd_decompress() 718 btrfs_ino(inode), folio_pos(dest_folio)); in zstd_decompress()
|
| /linux/fs/freevxfs/ |
| H A D | vxfs_immed.c | 34 void *src = vip->vii_immed.vi_immed + folio_pos(folio); in vxfs_immed_read_folio()
|
| /linux/fs/erofs/ |
| H A D | fileio.c | 94 loff_t pos = folio_pos(folio), ofs; in erofs_fileio_scan_folio() 107 ofs = folio_pos(folio) + cur - map->m_la; in erofs_fileio_scan_folio()
|
| H A D | fscache.c | 217 folio_pos(folio), folio_size(folio)); in erofs_fscache_meta_read_folio() 229 folio_pos(folio), folio_size(folio)); in erofs_fscache_meta_read_folio() 231 ret = erofs_fscache_read_io_async(ctx->cookie, folio_pos(folio), io); in erofs_fscache_meta_read_folio() 331 folio_pos(folio), folio_size(folio)); in erofs_fscache_read_folio()
|
| /linux/fs/vboxsf/ |
| H A D | file.c | 231 loff_t off = folio_pos(folio); in vboxsf_read_folio() 279 loff_t off = folio_pos(folio); in vboxsf_writepages()
|
| /linux/fs/ext4/ |
| H A D | page-io.c | 429 fscrypt_set_bio_crypt_ctx(bio, inode, folio_pos(folio) + bh_offset(bh), in io_submit_init_bio() 448 folio_pos(folio) + bh_offset(bh))) in io_submit_need_new_bio()
|
| H A D | inode.c | 1752 if (folio_pos(folio) < mpd->start_pos) in mpage_release_unused_pages() 2083 WARN_ON_ONCE(folio_pos(folio) != mpd->start_pos); in mpage_submit_folio() 2100 if (folio_pos(folio) + len > size && in mpage_submit_folio() 2432 if (WARN_ON_ONCE((folio_pos(folio) == pos) || in mpage_submit_partial_folio() 2583 err = ext4_jbd2_inode_add_write(handle, inode, folio_pos(folio), len); in ext4_journal_folio_buffers() 2602 if (folio_pos(folio) + len > size && in mpage_journal_page_buffers() 2678 mpd->next_pos != folio_pos(folio)) in mpage_prepare_extent_to_map() 2724 mpd->start_pos = folio_pos(folio); in mpage_prepare_extent_to_map() 6704 if (folio->mapping != inode->i_mapping || folio_pos(folio) > size) { in ext4_block_page_mkwrite() 6710 if (folio_pos(folio) + len > size) in ext4_block_page_mkwrite() [all …]
|
| /linux/arch/xtensa/mm/ |
| H A D | cache.c | 143 unsigned long temp = folio_pos(folio); in flush_dcache_folio()
|
| /linux/fs/xfs/scrub/ |
| H A D | xfile.c | 309 trace_xfile_put_folio(xf, folio_pos(folio), folio_size(folio)); in xfile_put_folio()
|
| /linux/fs/ntfs/ |
| H A D | aops.c | 27 loff_t pos = folio_pos(folio); in ntfs_iomap_read_end_io()
|
| /linux/arch/arm/mm/ |
| H A D | flush.c | 235 flush_pfn_alias(folio_pfn(folio), folio_pos(folio)); in __flush_dcache_folio()
|
| /linux/fs/ceph/ |
| H A D | addr.c | 731 loff_t page_off = folio_pos(folio); in write_folio_nounlock() 1178 if (folio_pos(folio) >= ceph_wbc->i_size) { in ceph_check_page_before_write() 1183 folio_pos(folio) >= i_size_read(inode)) && in ceph_check_page_before_write() 1223 ceph_wbc->offset = (u64)folio_pos(folio); in ceph_allocate_page_array() 1369 ceph_wbc->offset = (u64)folio_pos(folio); in ceph_process_folio_batch() 2065 loff_t off = folio_pos(folio); in ceph_page_mkwrite()
|
| /linux/fs/nfs/ |
| H A D | write.c | 200 end = folio_pos(folio) + (loff_t)offset + (loff_t)count; in nfs_grow_file() 1310 (long long)(folio_pos(folio) + offset)); in nfs_update_folio() 2044 loff_t range_start = folio_pos(folio); in nfs_wb_folio_reclaim() 2078 loff_t range_start = folio_pos(folio); in nfs_wb_folio()
|
| /linux/lib/ |
| H A D | buildid.c | 60 r->folio_off = folio_pos(r->folio); in freader_get_folio()
|