| /linux/mm/damon/ |
| H A D | paddr.c | 79 *folio_sz = folio_size(folio); in damon_pa_young() 187 *sz_filter_passed += folio_size(folio) / addr_unit; in damon_pa_pageout() 198 addr += folio_size(folio); in damon_pa_pageout() 228 *sz_filter_passed += folio_size(folio) / addr_unit; in damon_pa_mark_accessed_or_deactivate() 236 addr += folio_size(folio); in damon_pa_mark_accessed_or_deactivate() 276 *sz_filter_passed += folio_size(folio) / addr_unit; in damon_pa_migrate() 282 addr += folio_size(folio); in damon_pa_migrate() 310 *sz_filter_passed += folio_size(folio) / addr_unit; in damon_pa_stat() 311 addr += folio_size(folio); in damon_pa_stat()
|
| /linux/mm/ |
| H A D | page_io.c | 399 bvec_set_folio(&sio->bvec[sio->pages], folio, folio_size(folio), 0); in swap_writepage_fs() 400 sio->len += folio_size(folio); in swap_writepage_fs() 418 bio_add_folio_nofail(&bio, folio, folio_size(folio), 0); in swap_writepage_bdev_sync() 438 bio_add_folio_nofail(bio, folio, folio_size(folio), 0); in swap_writepage_bdev_async() 533 folio_zero_range(folio, 0, folio_size(folio)); in swap_read_folio_zeromap() 561 bvec_set_folio(&sio->bvec[sio->pages], folio, folio_size(folio), 0); in swap_read_folio_fs() 562 sio->len += folio_size(folio); in swap_read_folio_fs() 580 bio_add_folio_nofail(&bio, folio, folio_size(folio), 0); in swap_read_folio_bdev_sync() 602 bio_add_folio_nofail(bio, folio, folio_size(folio), 0); in swap_read_folio_bdev_async()
|
| H A D | truncate.c | 160 folio_invalidate(folio, 0, folio_size(folio)); in truncate_cleanup_folio() 220 size_t size = folio_size(folio); in truncate_inode_partial_folio() 878 folio_size(folio)); in pagecache_isize_extended()
|
| /linux/fs/xfs/scrub/ |
| H A D | xfile.c | 149 len = min_t(ssize_t, count, folio_size(folio) - offset); in xfile_load() 210 len = min_t(ssize_t, count, folio_size(folio) - offset); in xfile_store() 280 if (len > folio_size(folio) - offset_in_folio(folio, pos)) { in xfile_get_folio() 309 trace_xfile_put_folio(xf, folio_pos(folio), folio_size(folio)); in xfile_put_folio()
|
| /linux/fs/iomap/ |
| H A D | buffered-io.c | 296 size_t plen = min_t(loff_t, folio_size(folio) - poff, length); in iomap_adjust_read_range() 392 iomap_set_range_uptodate(folio, offset, folio_size(folio) - offset); in iomap_read_inline_data() 448 ifs->read_bytes_pending = folio_size(folio); in iomap_read_init() 481 ifs->read_bytes_pending -= folio_size(folio) - bytes_submitted; in iomap_read_end() 516 size_t folio_len = folio_size(folio); in iomap_read_folio_iter() 587 .len = folio_size(folio), in iomap_read_folio() 691 count = min(folio_size(folio) - from, count); in iomap_is_partially_uptodate() 728 folio_size(folio)); in iomap_release_folio() 751 if (offset == 0 && len == folio_size(folio)) { in iomap_invalidate_folio() 762 size_t len = folio_size(folio); in iomap_dirty_folio() [all …]
|
| /linux/fs/ |
| H A D | mpage.c | 156 const unsigned blocks_per_folio = folio_size(folio) >> blkbits; in do_mpage_readpage() 269 folio_zero_segment(folio, first_hole << blkbits, folio_size(folio)); in do_mpage_readpage() 461 const unsigned blocks_per_folio = folio_size(folio) >> blkbits; in mpage_write_folio() 576 length = folio_size(folio); in mpage_write_folio() 587 folio_zero_segment(folio, length, folio_size(folio)); in mpage_write_folio() 611 wbc_account_cgroup_owner(wbc, folio, folio_size(folio)); in mpage_write_folio()
|
| H A D | buffer.c | 930 offset = folio_size(folio); in folio_alloc_buffers() 1579 BUG_ON(offset >= folio_size(folio)); in folio_set_bh() 1639 BUG_ON(stop > folio_size(folio) || stop < length); in block_invalidate_folio() 1670 if (length == folio_size(folio)) in block_invalidate_folio() 2120 BUG_ON(to > folio_size(folio)); in __block_write_begin_int() 2354 to = min(folio_size(folio) - from, count); in block_is_partially_uptodate() 2356 if (from < blocksize && to > folio_size(folio) - blocksize) in block_is_partially_uptodate() 2631 end = folio_size(folio); in block_page_mkwrite() 2749 folio_size(folio)); in block_write_full_folio()
|
| /linux/fs/btrfs/ |
| H A D | extent_io.c | 558 btrfs_folio_set_lock(fs_info, folio, folio_pos(folio), folio_size(folio)); in begin_folio_read() 711 eb->folio_size = PAGE_SIZE; in alloc_eb_folio_array() 798 ASSERT(pg_offset + size <= folio_size(folio)); in submit_extent_folio() 1004 const u64 end = start + folio_size(folio) - 1; in btrfs_do_readpage() 1028 folio_size(folio) - zero_offset); in btrfs_do_readpage() 1341 const u64 end = start + folio_size(folio) - 1; in btrfs_read_folio() 1375 ASSERT(start >= folio_start && start + len <= folio_start + folio_size(folio)); in set_delalloc_bitmap() 1393 ASSERT(start >= folio_start && start < folio_start + folio_size(folio)); in find_next_delalloc_bitmap() 1429 const u64 page_end = page_start + folio_size(folio) - 1; in writepage_delalloc() 1735 const u64 folio_end = folio_start + folio_size(folio); in extent_writepage_io() [all …]
|
| H A D | zlib.c | 136 copy_length = min(folio_size(folio) - offset, in copy_data_into_buffer() 256 if (!bio_add_folio(bio, out_folio, folio_size(out_folio), 0)) { in zlib_compress_bio() 294 if (!bio_add_folio(bio, out_folio, folio_size(out_folio), 0)) { in zlib_compress_bio() 360 ASSERT(folio_size(fi.folio) == min_folio_size); in zlib_decompress_bio() 422 ASSERT(folio_size(fi.folio) == min_folio_size); in zlib_decompress_bio()
|
| H A D | accessors.c | 60 const int part = eb->folio_size - oif; \ 90 const int part = eb->folio_size - oif; \
|
| H A D | extent_io.h | 89 u32 folio_size; member 131 ASSERT(eb->folio_size); in offset_in_eb_folio() 132 return start & (eb->folio_size - 1); in offset_in_eb_folio()
|
| H A D | lzo.c | 137 const u32 fsize = folio_size(*out_folio); in write_and_queue_folio() 419 folio_size(cur_folio) - offset_in_folio(cur_folio, *cur_in)); in copy_compressed_segment() 450 ASSERT(folio_size(fi.folio) == sectorsize); in lzo_decompress_bio()
|
| H A D | compression.c | 414 u64 folio_sz = folio_size(folio); in add_ra_bio_pages() 440 cur += folio_size(folio); in add_ra_bio_pages() 457 page_end = (pg_index << PAGE_SHIFT) + folio_size(folio) - 1; in add_ra_bio_pages() 487 zeros = folio_size(folio) - zero_offset; in add_ra_bio_pages() 1088 ASSERT(dest_pgoff + destlen <= folio_size(dest_folio) && destlen <= sectorsize); in btrfs_decompress()
|
| /linux/fs/netfs/ |
| H A D | buffered_read.c | 408 size_t flen = folio_size(folio); in netfs_read_gaps() 510 folio_pos(folio), folio_size(folio), in netfs_read_folio() 563 size_t plen = folio_size(folio); in netfs_skip_folio_read() 665 folio_pos(folio), folio_size(folio), in netfs_write_begin() 723 size_t flen = folio_size(folio); in netfs_prefetch_for_write()
|
| H A D | read_pgpriv2.c | 22 size_t fsize = folio_size(folio), flen = fsize; in netfs_pgpriv2_copy_folio() 195 fsize = folio_size(folio); in netfs_pgpriv2_unlock_copied_folios()
|
| /linux/lib/ |
| H A D | buildid.c | 45 file_off < r->folio_off + folio_size(r->folio)) in freader_get_folio() 117 folio_sz = folio_size(r->folio); in freader_fetch() 129 folio_sz = folio_size(r->folio); in freader_fetch()
|
| /linux/fs/ubifs/ |
| H A D | file.c | 119 folio_zero_range(folio, 0, folio_size(folio)); in do_readpage() 242 if (pos == folio_pos(folio) && len >= folio_size(folio)) in write_begin_slow() 434 if (pos == folio_pos(folio) && len >= folio_size(folio)) { in ubifs_write_begin() 555 if (len == folio_size(folio)) in ubifs_write_end() 608 folio_zero_range(folio, 0, folio_size(folio)); in populate_page() 988 int err, len = folio_size(folio); in ubifs_writepage() 1030 folio_zero_segment(folio, len, folio_size(folio)); in ubifs_writepage() 1291 if (offset || length < folio_size(folio)) in ubifs_invalidate_folio()
|
| /linux/fs/hfs/ |
| H A D | btree.c | 86 folio_zero_range(folio, 0, folio_size(folio)); in hfs_btree_open() 91 size = folio_size(folio); in hfs_btree_open() 102 len = min_t(size_t, folio_size(folio), sb->s_blocksize); in hfs_btree_open()
|
| /linux/arch/nios2/mm/ |
| H A D | cacheflush.c | 172 __flush_dcache(start, start + folio_size(folio)); in __flush_dcache_folio() 196 flush_icache_range(start, start + folio_size(folio)); in flush_dcache_folio()
|
| /linux/arch/arm64/mm/ |
| H A D | flush.c | 59 folio_size(folio)); in __sync_icache_dcache()
|
| /linux/fs/nfs/ |
| H A D | file.c | 304 size_t end = folio_size(folio); in nfs_truncate_last_folio() 443 size_t fsize = folio_size(folio); in nfs_write_end() 491 if (offset != 0 || length < folio_size(folio)) in nfs_invalidate_folio() 566 folio_size(folio), ret); in nfs_launder_folio()
|
| /linux/arch/sh/mm/ |
| H A D | cache.c | 154 folio_size(folio)); in __update_cache() 174 folio_size(folio)); in __flush_anon_page()
|
| /linux/fs/vboxsf/ |
| H A D | file.c | 280 u32 nwrite = folio_size(folio); in vboxsf_writepages() 331 if (!folio_test_uptodate(folio) && nwritten == folio_size(folio)) in vboxsf_write_end()
|
| /linux/fs/orangefs/ |
| H A D | inode.c | 45 wlen = folio_size(folio); in orangefs_writepage_locked() 95 folio_size(ow->folios[i]) - start, start); in orangefs_writepages_work() 273 bvec_set_folio(&bv, folio, folio_size(folio), 0); in orangefs_read_folio() 274 iov_iter_bvec(&iter, ITER_DEST, &bv, 1, folio_size(folio)); in orangefs_read_folio() 277 folio_size(folio), inode->i_size, NULL, NULL, file); in orangefs_read_folio()
|
| /linux/arch/microblaze/include/asm/ |
| H A D | cacheflush.h | 81 flush_dcache_range(addr, addr + folio_size(folio)); in flush_dcache_folio()
|