Home
last modified time | relevance | path

Searched refs:folio_pos (Results 1 – 25 of 40) sorted by relevance

12

/linux/fs/netfs/
H A Dread_pgpriv2.c23 loff_t fpos = folio_pos(folio), i_size; in netfs_pgpriv2_copy_folio()
105 creq = netfs_create_write_req(rreq->mapping, NULL, folio_pos(folio), in netfs_pgpriv2_begin_copy_to_cache()
194 fpos = folio_pos(folio); in netfs_pgpriv2_unlock_copied_folios()
H A Dbuffered_read.c414 rreq = netfs_alloc_request(mapping, file, folio_pos(folio), flen, NETFS_READ_GAPS); in netfs_read_gaps()
509 folio_pos(folio), folio_size(folio), in netfs_read_folio()
664 folio_pos(folio), folio_size(folio), in netfs_write_begin()
721 unsigned long long start = folio_pos(folio); in netfs_prefetch_for_write()
H A Dwrite_issue.c330 loff_t fpos = folio_pos(folio), i_size; in netfs_write_folio()
572 wreq = netfs_create_write_req(mapping, NULL, folio_pos(folio), NETFS_WRITEBACK); in netfs_writepages()
586 WARN_ON_ONCE(wreq && folio_pos(folio) < atomic64_read(&wreq->issued_to)); in netfs_writepages()
711 loff_t fpos = folio_pos(folio); in netfs_write_folio_single()
H A Dwrite_collect.c71 fend = folio_pos(folio) + finfo->dirty_offset + finfo->dirty_len; in netfs_folio_written_back()
149 fpos = folio_pos(folio); in netfs_writeback_unlock_folios()
H A Dread_collect.c49 if (unlikely(folio_pos(folio) < rreq->abandon_to)) { in netfs_unlock_read_folio()
139 fpos = folio_pos(folio); in netfs_read_unlock_folios()
H A Dmisc.c221 unsigned long long fpos = folio_pos(folio), end; in netfs_invalidate_folio()
/linux/fs/iomap/
H A Dbuffered-io.c157 *range_start = folio_pos(folio) + (start_blk << inode->i_blkbits); in ifs_find_dirty_range()
416 folio_pos(folio) + off, len, error, in iomap_finish_folio_read()
591 .pos = folio_pos(folio), in iomap_read_folio()
732 trace_iomap_release_folio(folio->mapping->host, folio_pos(folio), in iomap_release_folio()
750 folio_pos(folio) + offset, len); in iomap_invalidate_folio()
809 if (!(iter->flags & IOMAP_UNSHARE) && pos <= folio_pos(folio) && in __iomap_write_begin()
920 WARN_ON_ONCE(pos < folio_pos(folio)); in iomap_trim_folio_range()
921 WARN_ON_ONCE(pos >= folio_pos(folio) + fsize); in iomap_trim_folio_range()
1000 if (folio_pos(folio) > iter->pos) { in iomap_write_begin()
1001 len = min_t(u64, folio_pos(folio) - iter->pos, in iomap_write_begin()
[all …]
/linux/fs/
H A Dmpage.c180 block_in_file = folio_pos(folio) >> blkbits; in do_mpage_readpage()
531 block_in_file = folio_pos(folio) >> blkbits; in mpage_write_folio()
573 if (folio_pos(folio) >= i_size) in mpage_write_folio()
576 if (folio_pos(folio) + length > i_size) { in mpage_write_folio()
585 length = i_size - folio_pos(folio); in mpage_write_folio()
/linux/fs/btrfs/
H A Dextent_io.c336 range_start = max_t(u64, folio_pos(folio), start); in lock_delalloc_folios()
391 orig_end <= folio_pos(locked_folio))); in find_lock_delalloc_range()
494 ASSERT(folio_pos(folio) <= start && in end_folio_read()
529 u64 start = folio_pos(folio) + fi.offset; in end_bbio_data_write()
552 btrfs_folio_set_lock(fs_info, folio, folio_pos(folio), folio_size(folio)); in begin_folio_read()
583 u64 start = folio_pos(folio) + fi.offset; in end_bbio_data_read()
790 loff_t file_offset = folio_pos(folio) + pg_offset; in submit_extent_folio()
997 u64 start = folio_pos(folio); in btrfs_do_readpage()
1334 const u64 start = folio_pos(folio); in btrfs_read_folio()
1346 if (folio_pos(folio) < i_size_read(vfs_inode)) in btrfs_read_folio()
[all …]
H A Ddefrag.c886 lock_start = folio_pos(folio); in defrag_prepare_one_folio()
1180 start + len <= folio_pos(folio)) in defrag_one_locked_target()
1232 btrfs_lock_extent(&inode->io_tree, folio_pos(folios[0]), cur - 1, &cached_state); in defrag_one_range()
1258 btrfs_unlock_extent(&inode->io_tree, folio_pos(folios[0]), cur - 1, &cached_state); in defrag_one_range()
H A Dzlib.c475 folio_pos(dest_folio)); in zlib_decompress()
497 folio_pos(dest_folio), to_copy, destlen); in zlib_decompress()
H A Dzstd.c694 folio_pos(dest_folio)); in zstd_decompress()
718 btrfs_ino(inode), folio_pos(dest_folio)); in zstd_decompress()
/linux/fs/freevxfs/
H A Dvxfs_immed.c34 void *src = vip->vii_immed.vi_immed + folio_pos(folio); in vxfs_immed_read_folio()
/linux/fs/erofs/
H A Dfileio.c94 loff_t pos = folio_pos(folio), ofs; in erofs_fileio_scan_folio()
107 ofs = folio_pos(folio) + cur - map->m_la; in erofs_fileio_scan_folio()
H A Dfscache.c217 folio_pos(folio), folio_size(folio)); in erofs_fscache_meta_read_folio()
229 folio_pos(folio), folio_size(folio)); in erofs_fscache_meta_read_folio()
231 ret = erofs_fscache_read_io_async(ctx->cookie, folio_pos(folio), io); in erofs_fscache_meta_read_folio()
331 folio_pos(folio), folio_size(folio)); in erofs_fscache_read_folio()
/linux/fs/vboxsf/
H A Dfile.c231 loff_t off = folio_pos(folio); in vboxsf_read_folio()
279 loff_t off = folio_pos(folio); in vboxsf_writepages()
/linux/fs/ext4/
H A Dpage-io.c429 fscrypt_set_bio_crypt_ctx(bio, inode, folio_pos(folio) + bh_offset(bh), in io_submit_init_bio()
448 folio_pos(folio) + bh_offset(bh))) in io_submit_need_new_bio()
H A Dinode.c1752 if (folio_pos(folio) < mpd->start_pos) in mpage_release_unused_pages()
2083 WARN_ON_ONCE(folio_pos(folio) != mpd->start_pos); in mpage_submit_folio()
2100 if (folio_pos(folio) + len > size && in mpage_submit_folio()
2432 if (WARN_ON_ONCE((folio_pos(folio) == pos) || in mpage_submit_partial_folio()
2583 err = ext4_jbd2_inode_add_write(handle, inode, folio_pos(folio), len); in ext4_journal_folio_buffers()
2602 if (folio_pos(folio) + len > size && in mpage_journal_page_buffers()
2678 mpd->next_pos != folio_pos(folio)) in mpage_prepare_extent_to_map()
2724 mpd->start_pos = folio_pos(folio); in mpage_prepare_extent_to_map()
6704 if (folio->mapping != inode->i_mapping || folio_pos(folio) > size) { in ext4_block_page_mkwrite()
6710 if (folio_pos(folio) + len > size) in ext4_block_page_mkwrite()
[all …]
/linux/arch/xtensa/mm/
H A Dcache.c143 unsigned long temp = folio_pos(folio); in flush_dcache_folio()
/linux/fs/xfs/scrub/
H A Dxfile.c309 trace_xfile_put_folio(xf, folio_pos(folio), folio_size(folio)); in xfile_put_folio()
/linux/fs/ntfs/
H A Daops.c27 loff_t pos = folio_pos(folio); in ntfs_iomap_read_end_io()
/linux/arch/arm/mm/
H A Dflush.c235 flush_pfn_alias(folio_pfn(folio), folio_pos(folio)); in __flush_dcache_folio()
/linux/fs/ceph/
H A Daddr.c731 loff_t page_off = folio_pos(folio); in write_folio_nounlock()
1178 if (folio_pos(folio) >= ceph_wbc->i_size) { in ceph_check_page_before_write()
1183 folio_pos(folio) >= i_size_read(inode)) && in ceph_check_page_before_write()
1223 ceph_wbc->offset = (u64)folio_pos(folio); in ceph_allocate_page_array()
1369 ceph_wbc->offset = (u64)folio_pos(folio); in ceph_process_folio_batch()
2065 loff_t off = folio_pos(folio); in ceph_page_mkwrite()
/linux/fs/nfs/
H A Dwrite.c200 end = folio_pos(folio) + (loff_t)offset + (loff_t)count; in nfs_grow_file()
1310 (long long)(folio_pos(folio) + offset)); in nfs_update_folio()
2044 loff_t range_start = folio_pos(folio); in nfs_wb_folio_reclaim()
2078 loff_t range_start = folio_pos(folio); in nfs_wb_folio()
/linux/lib/
H A Dbuildid.c60 r->folio_off = folio_pos(r->folio); in freader_get_folio()

12