Home
last modified time | relevance | path

Searched refs:folio_pos (Results 1 – 25 of 54) sorted by relevance

123

/linux/fs/orangefs/
H A Dinode.c44 off = folio_pos(folio); in orangefs_writepage_locked()
272 off = folio_pos(folio); in orangefs_read_folio()
366 if (pos == folio_pos(folio) && in orangefs_write_end()
390 } else if (folio_pos(folio) + offset <= wr->pos && in orangefs_invalidate_folio()
391 wr->pos + wr->len <= folio_pos(folio) + offset + length) { in orangefs_invalidate_folio()
397 } else if (wr->pos < folio_pos(folio) + offset && in orangefs_invalidate_folio()
398 wr->pos + wr->len <= folio_pos(folio) + offset + length && in orangefs_invalidate_folio()
399 folio_pos(folio) + offset < wr->pos + wr->len) { in orangefs_invalidate_folio()
401 x = wr->pos + wr->len - (folio_pos(folio) + offset); in orangefs_invalidate_folio()
407 } else if (folio_pos(folio) + offset <= wr->pos && in orangefs_invalidate_folio()
[all …]
/linux/fs/iomap/
H A Dbuffered-io.c157 *range_start = folio_pos(folio) + (start_blk << inode->i_blkbits); in ifs_find_dirty_range()
416 folio_pos(folio) + off, len, error, in iomap_finish_folio_read()
586 .pos = folio_pos(folio), in iomap_read_folio()
727 trace_iomap_release_folio(folio->mapping->host, folio_pos(folio), in iomap_release_folio()
745 folio_pos(folio) + offset, len); in iomap_invalidate_folio()
804 if (!(iter->flags & IOMAP_UNSHARE) && pos <= folio_pos(folio) && in __iomap_write_begin()
915 WARN_ON_ONCE(pos < folio_pos(folio)); in iomap_trim_folio_range()
916 WARN_ON_ONCE(pos >= folio_pos(folio) + fsize); in iomap_trim_folio_range()
995 if (folio_pos(folio) > iter->pos) { in iomap_write_begin()
996 len = min_t(u64, folio_pos(folio) - iter->pos, in iomap_write_begin()
[all …]
/linux/fs/netfs/
H A Dread_pgpriv2.c23 loff_t fpos = folio_pos(folio), i_size; in netfs_pgpriv2_copy_folio()
105 creq = netfs_create_write_req(rreq->mapping, NULL, folio_pos(folio), in netfs_pgpriv2_begin_copy_to_cache()
194 fpos = folio_pos(folio); in netfs_pgpriv2_unlock_copied_folios()
H A Dbuffered_read.c415 rreq = netfs_alloc_request(mapping, file, folio_pos(folio), flen, NETFS_READ_GAPS); in netfs_read_gaps()
510 folio_pos(folio), folio_size(folio), in netfs_read_folio()
665 folio_pos(folio), folio_size(folio), in netfs_write_begin()
722 unsigned long long start = folio_pos(folio); in netfs_prefetch_for_write()
H A Dwrite_issue.c331 loff_t fpos = folio_pos(folio), i_size; in netfs_write_folio()
573 wreq = netfs_create_write_req(mapping, NULL, folio_pos(folio), NETFS_WRITEBACK); in netfs_writepages()
587 WARN_ON_ONCE(wreq && folio_pos(folio) < atomic64_read(&wreq->issued_to)); in netfs_writepages()
712 loff_t fpos = folio_pos(folio); in netfs_write_folio_single()
H A Dwrite_collect.c71 fend = folio_pos(folio) + finfo->dirty_offset + finfo->dirty_len; in netfs_folio_written_back()
149 fpos = folio_pos(folio); in netfs_writeback_unlock_folios()
H A Dbuffered_write.c186 fpos = folio_pos(folio); in netfs_perform_write()
537 folio_pos(folio), in netfs_page_mkwrite()
H A Dread_collect.c49 if (unlikely(folio_pos(folio) < rreq->abandon_to)) { in netfs_unlock_read_folio()
139 fpos = folio_pos(folio); in netfs_read_unlock_folios()
/linux/fs/
H A Dmpage.c181 block_in_file = folio_pos(folio) >> blkbits; in do_mpage_readpage()
532 block_in_file = folio_pos(folio) >> blkbits; in mpage_write_folio()
574 if (folio_pos(folio) >= i_size) in mpage_write_folio()
577 if (folio_pos(folio) + length > i_size) { in mpage_write_folio()
586 length = i_size - folio_pos(folio); in mpage_write_folio()
H A Dbuffer.c1007 sector_t block = div_u64(folio_pos(folio), size); in folio_init_buffers()
1860 block = div_u64(folio_pos(folio), blocksize); in __block_write_full_folio()
2125 block = div_u64(folio_pos(folio), blocksize); in __block_write_begin_int()
2268 size_t start = pos - folio_pos(folio); in block_write_end()
2403 iblock = div_u64(folio_pos(folio), blocksize); in block_read_full_folio()
2625 (folio_pos(folio) >= size)) { in block_page_mkwrite()
2633 if (folio_pos(folio) + end > size) in block_page_mkwrite()
2634 end = size - folio_pos(folio); in block_page_mkwrite()
2736 if (folio_pos(folio) >= i_size) { in block_write_full_folio()
/linux/fs/btrfs/
H A Dextent_io.c335 range_start = max_t(u64, folio_pos(folio), start); in lock_delalloc_folios()
390 orig_end <= folio_pos(locked_folio))); in find_lock_delalloc_range()
493 ASSERT(folio_pos(folio) <= start && in end_folio_read()
528 u64 start = folio_pos(folio) + fi.offset; in end_bbio_data_write()
558 btrfs_folio_set_lock(fs_info, folio, folio_pos(folio), folio_size(folio)); in begin_folio_read()
589 u64 start = folio_pos(folio) + fi.offset; in end_bbio_data_read()
796 loff_t file_offset = folio_pos(folio) + pg_offset; in submit_extent_folio()
1003 u64 start = folio_pos(folio); in btrfs_do_readpage()
1340 const u64 start = folio_pos(folio); in btrfs_read_folio()
1352 if (folio_pos(folio) < i_size_read(vfs_inode)) in btrfs_read_folio()
[all …]
H A Ddefrag.c886 lock_start = folio_pos(folio); in defrag_prepare_one_folio()
1180 start + len <= folio_pos(folio)) in defrag_one_locked_target()
1232 btrfs_lock_extent(&inode->io_tree, folio_pos(folios[0]), cur - 1, &cached_state); in defrag_one_range()
1258 btrfs_unlock_extent(&inode->io_tree, folio_pos(folios[0]), cur - 1, &cached_state); in defrag_one_range()
H A Dzlib.c479 folio_pos(dest_folio)); in zlib_decompress()
501 folio_pos(dest_folio), to_copy, destlen); in zlib_decompress()
/linux/fs/ext4/
H A Dmove_extent.c137 block = folio_pos(folio) >> inode->i_blkbits; in mext_folio_mkuptodate()
227 move_len = umin(folio_pos(folio[0]) + folio_size(folio[0]) - orig_pos, in mext_move_begin()
228 folio_pos(folio[1]) + folio_size(folio[1]) - donor_pos); in mext_move_begin()
276 block = folio_pos(folio) >> inode->i_blkbits; in mext_folio_mkwrite()
/linux/fs/nfs/
H A Dfile.c302 loff_t fpos = folio_pos(folio); in nfs_truncate_last_folio()
496 trace_nfs_invalidate_folio(inode, folio_pos(folio) + offset, length); in nfs_invalidate_folio()
561 inode->i_ino, folio_pos(folio)); in nfs_launder_folio()
565 trace_nfs_launder_folio_done(inode, folio_pos(folio), in nfs_launder_folio()
652 (long long)folio_pos(folio)); in nfs_vm_page_mkwrite()
/linux/fs/freevxfs/
H A Dvxfs_immed.c34 void *src = vip->vii_immed.vi_immed + folio_pos(folio); in vxfs_immed_read_folio()
/linux/fs/ubifs/
H A Dfile.c242 if (pos == folio_pos(folio) && len >= folio_size(folio)) in write_begin_slow()
434 if (pos == folio_pos(folio) && len >= folio_size(folio)) { in ubifs_write_begin()
995 if (folio_pos(folio) >= i_size) { in ubifs_writepage()
1005 if (folio_pos(folio) + len <= i_size) { in ubifs_writepage()
1006 if (folio_pos(folio) + len > synced_i_size) { in ubifs_writepage()
1029 len = i_size - folio_pos(folio); in ubifs_writepage()
1541 folio_pos(folio) >= i_size_read(inode))) { in ubifs_vm_page_mkwrite()
/linux/fs/erofs/
H A Dfileio.c96 loff_t pos = folio_pos(folio), ofs; in erofs_fileio_scan_folio()
109 ofs = folio_pos(folio) + cur - map->m_la; in erofs_fileio_scan_folio()
H A Dfscache.c217 folio_pos(folio), folio_size(folio)); in erofs_fscache_meta_read_folio()
229 folio_pos(folio), folio_size(folio)); in erofs_fscache_meta_read_folio()
231 ret = erofs_fscache_read_io_async(ctx->cookie, folio_pos(folio), io); in erofs_fscache_meta_read_folio()
331 folio_pos(folio), folio_size(folio)); in erofs_fscache_read_folio()
/linux/fs/minix/
H A Ddir.c262 pos = folio_pos(folio) + offset_in_folio(folio, p); in minix_add_link()
289 loff_t pos = folio_pos(folio) + offset_in_folio(folio, de); in minix_delete_entry()
412 loff_t pos = folio_pos(folio) + offset_in_folio(folio, de); in minix_set_link()
/linux/fs/vboxsf/
H A Dfile.c231 loff_t off = folio_pos(folio); in vboxsf_read_folio()
279 loff_t off = folio_pos(folio); in vboxsf_writepages()
/linux/mm/
H A Dtruncate.c219 loff_t pos = folio_pos(folio); in truncate_inode_partial_folio()
876 offset = from - folio_pos(folio); in pagecache_isize_extended()
877 end = min_t(unsigned int, to - folio_pos(folio), in pagecache_isize_extended()
/linux/fs/ocfs2/
H A Dmmap.c53 loff_t pos = folio_pos(folio); in __ocfs2_page_mkwrite()
H A Daops.c266 loff_t start = folio_pos(folio); in ocfs2_read_folio()
555 u64 offset = folio_pos(folio) + block_start; in ocfs2_should_read_blk()
936 new = new | ((i_size_read(inode) <= folio_pos(folio)) && in ocfs2_prepare_folio_for_write()
937 (folio_pos(folio) <= user_pos)); in ocfs2_prepare_folio_for_write()
1960 if (new_isize > folio_pos(wc->w_target_folio)) in ocfs2_write_end_nolock()
2004 loff_t start_byte = folio_pos(folio) + from; in ocfs2_write_end_nolock()
/linux/fs/jfs/
H A Djfs_metapage.c352 folio_pos(folio)); in last_read_complete()
444 page_start = folio_pos(folio) >> inode->i_blkbits; in metapage_write_folio()
577 page_start = folio_pos(folio) >> inode->i_blkbits; in metapage_read_folio()

123