Home
last modified time | relevance | path

Searched refs:folio_size (Results 1 – 25 of 98) sorted by relevance

1234

/linux/mm/damon/
H A Dpaddr.c79 *folio_sz = folio_size(folio); in damon_pa_young()
187 *sz_filter_passed += folio_size(folio) / addr_unit; in damon_pa_pageout()
198 addr += folio_size(folio); in damon_pa_pageout()
228 *sz_filter_passed += folio_size(folio) / addr_unit; in damon_pa_mark_accessed_or_deactivate()
236 addr += folio_size(folio); in damon_pa_mark_accessed_or_deactivate()
276 *sz_filter_passed += folio_size(folio) / addr_unit; in damon_pa_migrate()
282 addr += folio_size(folio); in damon_pa_migrate()
310 *sz_filter_passed += folio_size(folio) / addr_unit; in damon_pa_stat()
311 addr += folio_size(folio); in damon_pa_stat()
/linux/mm/
H A Dpage_io.c399 bvec_set_folio(&sio->bvec[sio->pages], folio, folio_size(folio), 0); in swap_writepage_fs()
400 sio->len += folio_size(folio); in swap_writepage_fs()
418 bio_add_folio_nofail(&bio, folio, folio_size(folio), 0); in swap_writepage_bdev_sync()
438 bio_add_folio_nofail(bio, folio, folio_size(folio), 0); in swap_writepage_bdev_async()
533 folio_zero_range(folio, 0, folio_size(folio)); in swap_read_folio_zeromap()
561 bvec_set_folio(&sio->bvec[sio->pages], folio, folio_size(folio), 0); in swap_read_folio_fs()
562 sio->len += folio_size(folio); in swap_read_folio_fs()
580 bio_add_folio_nofail(&bio, folio, folio_size(folio), 0); in swap_read_folio_bdev_sync()
602 bio_add_folio_nofail(bio, folio, folio_size(folio), 0); in swap_read_folio_bdev_async()
H A Dtruncate.c160 folio_invalidate(folio, 0, folio_size(folio)); in truncate_cleanup_folio()
220 size_t size = folio_size(folio); in truncate_inode_partial_folio()
878 folio_size(folio)); in pagecache_isize_extended()
/linux/fs/xfs/scrub/
H A Dxfile.c149 len = min_t(ssize_t, count, folio_size(folio) - offset); in xfile_load()
210 len = min_t(ssize_t, count, folio_size(folio) - offset); in xfile_store()
280 if (len > folio_size(folio) - offset_in_folio(folio, pos)) { in xfile_get_folio()
309 trace_xfile_put_folio(xf, folio_pos(folio), folio_size(folio)); in xfile_put_folio()
/linux/fs/iomap/
H A Dbuffered-io.c296 size_t plen = min_t(loff_t, folio_size(folio) - poff, length); in iomap_adjust_read_range()
392 iomap_set_range_uptodate(folio, offset, folio_size(folio) - offset); in iomap_read_inline_data()
448 ifs->read_bytes_pending = folio_size(folio); in iomap_read_init()
481 ifs->read_bytes_pending -= folio_size(folio) - bytes_submitted; in iomap_read_end()
516 size_t folio_len = folio_size(folio); in iomap_read_folio_iter()
587 .len = folio_size(folio), in iomap_read_folio()
691 count = min(folio_size(folio) - from, count); in iomap_is_partially_uptodate()
728 folio_size(folio)); in iomap_release_folio()
751 if (offset == 0 && len == folio_size(folio)) { in iomap_invalidate_folio()
762 size_t len = folio_size(folio); in iomap_dirty_folio()
[all …]
/linux/fs/
H A Dmpage.c156 const unsigned blocks_per_folio = folio_size(folio) >> blkbits; in do_mpage_readpage()
269 folio_zero_segment(folio, first_hole << blkbits, folio_size(folio)); in do_mpage_readpage()
461 const unsigned blocks_per_folio = folio_size(folio) >> blkbits; in mpage_write_folio()
576 length = folio_size(folio); in mpage_write_folio()
587 folio_zero_segment(folio, length, folio_size(folio)); in mpage_write_folio()
611 wbc_account_cgroup_owner(wbc, folio, folio_size(folio)); in mpage_write_folio()
H A Dbuffer.c930 offset = folio_size(folio); in folio_alloc_buffers()
1579 BUG_ON(offset >= folio_size(folio)); in folio_set_bh()
1639 BUG_ON(stop > folio_size(folio) || stop < length); in block_invalidate_folio()
1670 if (length == folio_size(folio)) in block_invalidate_folio()
2120 BUG_ON(to > folio_size(folio)); in __block_write_begin_int()
2354 to = min(folio_size(folio) - from, count); in block_is_partially_uptodate()
2356 if (from < blocksize && to > folio_size(folio) - blocksize) in block_is_partially_uptodate()
2631 end = folio_size(folio); in block_page_mkwrite()
2749 folio_size(folio)); in block_write_full_folio()
/linux/fs/btrfs/
H A Dextent_io.c558 btrfs_folio_set_lock(fs_info, folio, folio_pos(folio), folio_size(folio)); in begin_folio_read()
711 eb->folio_size = PAGE_SIZE; in alloc_eb_folio_array()
798 ASSERT(pg_offset + size <= folio_size(folio)); in submit_extent_folio()
1004 const u64 end = start + folio_size(folio) - 1; in btrfs_do_readpage()
1028 folio_size(folio) - zero_offset); in btrfs_do_readpage()
1341 const u64 end = start + folio_size(folio) - 1; in btrfs_read_folio()
1375 ASSERT(start >= folio_start && start + len <= folio_start + folio_size(folio)); in set_delalloc_bitmap()
1393 ASSERT(start >= folio_start && start < folio_start + folio_size(folio)); in find_next_delalloc_bitmap()
1429 const u64 page_end = page_start + folio_size(folio) - 1; in writepage_delalloc()
1735 const u64 folio_end = folio_start + folio_size(folio); in extent_writepage_io()
[all …]
H A Dzlib.c136 copy_length = min(folio_size(folio) - offset, in copy_data_into_buffer()
256 if (!bio_add_folio(bio, out_folio, folio_size(out_folio), 0)) { in zlib_compress_bio()
294 if (!bio_add_folio(bio, out_folio, folio_size(out_folio), 0)) { in zlib_compress_bio()
360 ASSERT(folio_size(fi.folio) == min_folio_size); in zlib_decompress_bio()
422 ASSERT(folio_size(fi.folio) == min_folio_size); in zlib_decompress_bio()
H A Daccessors.c60 const int part = eb->folio_size - oif; \
90 const int part = eb->folio_size - oif; \
H A Dextent_io.h89 u32 folio_size; member
131 ASSERT(eb->folio_size); in offset_in_eb_folio()
132 return start & (eb->folio_size - 1); in offset_in_eb_folio()
H A Dlzo.c137 const u32 fsize = folio_size(*out_folio); in write_and_queue_folio()
419 folio_size(cur_folio) - offset_in_folio(cur_folio, *cur_in)); in copy_compressed_segment()
450 ASSERT(folio_size(fi.folio) == sectorsize); in lzo_decompress_bio()
H A Dcompression.c414 u64 folio_sz = folio_size(folio); in add_ra_bio_pages()
440 cur += folio_size(folio); in add_ra_bio_pages()
457 page_end = (pg_index << PAGE_SHIFT) + folio_size(folio) - 1; in add_ra_bio_pages()
487 zeros = folio_size(folio) - zero_offset; in add_ra_bio_pages()
1088 ASSERT(dest_pgoff + destlen <= folio_size(dest_folio) && destlen <= sectorsize); in btrfs_decompress()
/linux/fs/netfs/
H A Dbuffered_read.c408 size_t flen = folio_size(folio); in netfs_read_gaps()
510 folio_pos(folio), folio_size(folio), in netfs_read_folio()
563 size_t plen = folio_size(folio); in netfs_skip_folio_read()
665 folio_pos(folio), folio_size(folio), in netfs_write_begin()
723 size_t flen = folio_size(folio); in netfs_prefetch_for_write()
H A Dread_pgpriv2.c22 size_t fsize = folio_size(folio), flen = fsize; in netfs_pgpriv2_copy_folio()
195 fsize = folio_size(folio); in netfs_pgpriv2_unlock_copied_folios()
/linux/lib/
H A Dbuildid.c45 file_off < r->folio_off + folio_size(r->folio)) in freader_get_folio()
117 folio_sz = folio_size(r->folio); in freader_fetch()
129 folio_sz = folio_size(r->folio); in freader_fetch()
/linux/fs/ubifs/
H A Dfile.c119 folio_zero_range(folio, 0, folio_size(folio)); in do_readpage()
242 if (pos == folio_pos(folio) && len >= folio_size(folio)) in write_begin_slow()
434 if (pos == folio_pos(folio) && len >= folio_size(folio)) { in ubifs_write_begin()
555 if (len == folio_size(folio)) in ubifs_write_end()
608 folio_zero_range(folio, 0, folio_size(folio)); in populate_page()
988 int err, len = folio_size(folio); in ubifs_writepage()
1030 folio_zero_segment(folio, len, folio_size(folio)); in ubifs_writepage()
1291 if (offset || length < folio_size(folio)) in ubifs_invalidate_folio()
/linux/fs/hfs/
H A Dbtree.c86 folio_zero_range(folio, 0, folio_size(folio)); in hfs_btree_open()
91 size = folio_size(folio); in hfs_btree_open()
102 len = min_t(size_t, folio_size(folio), sb->s_blocksize); in hfs_btree_open()
/linux/arch/nios2/mm/
H A Dcacheflush.c172 __flush_dcache(start, start + folio_size(folio)); in __flush_dcache_folio()
196 flush_icache_range(start, start + folio_size(folio)); in flush_dcache_folio()
/linux/arch/arm64/mm/
H A Dflush.c59 folio_size(folio)); in __sync_icache_dcache()
/linux/fs/nfs/
H A Dfile.c304 size_t end = folio_size(folio); in nfs_truncate_last_folio()
443 size_t fsize = folio_size(folio); in nfs_write_end()
491 if (offset != 0 || length < folio_size(folio)) in nfs_invalidate_folio()
566 folio_size(folio), ret); in nfs_launder_folio()
/linux/arch/sh/mm/
H A Dcache.c154 folio_size(folio)); in __update_cache()
174 folio_size(folio)); in __flush_anon_page()
/linux/fs/vboxsf/
H A Dfile.c280 u32 nwrite = folio_size(folio); in vboxsf_writepages()
331 if (!folio_test_uptodate(folio) && nwritten == folio_size(folio)) in vboxsf_write_end()
/linux/fs/orangefs/
H A Dinode.c45 wlen = folio_size(folio); in orangefs_writepage_locked()
95 folio_size(ow->folios[i]) - start, start); in orangefs_writepages_work()
273 bvec_set_folio(&bv, folio, folio_size(folio), 0); in orangefs_read_folio()
274 iov_iter_bvec(&iter, ITER_DEST, &bv, 1, folio_size(folio)); in orangefs_read_folio()
277 folio_size(folio), inode->i_size, NULL, NULL, file); in orangefs_read_folio()
/linux/arch/microblaze/include/asm/
H A Dcacheflush.h81 flush_dcache_range(addr, addr + folio_size(folio)); in flush_dcache_folio()

1234