Lines Matching defs:write_bytes
72 size_t write_bytes, struct extent_state **cached, bool noreserve)
79 const u64 end_pos = pos + write_bytes;
83 if (write_bytes == 0)
855 loff_t pos, size_t write_bytes,
861 fgf_set_order(write_bytes);
876 ret = prepare_uptodate_folio(inode, folio, pos, write_bytes);
899 loff_t pos, size_t write_bytes,
909 last_pos = round_up(pos + write_bytes, fs_info->sectorsize) - 1;
957 * Check if we can do nocow write into the range [@pos, @pos + @write_bytes)
960 * @write_bytes: The length to write, will be updated to the nocow writeable
968 * > 0 If we can nocow, and updates @write_bytes.
978 size_t *write_bytes, bool nowait)
994 lockend = round_up(pos + *write_bytes,
1033 *write_bytes = min_t(size_t, *write_bytes, cur_offset - pos);
1135 * write_bytes may get smaller than requested here.
1205 size_t write_bytes = calc_write_bytes(inode, iter, start);
1221 if (unlikely(fault_in_iov_iter_readable(iter, write_bytes)))
1224 ret = reserve_space(inode, data_reserved, start, &write_bytes, nowait,
1231 ASSERT(start + write_bytes <= reserved_start + reserved_len);
1243 ret = prepare_one_folio(&inode->vfs_inode, &folio, start, write_bytes, false);
1261 write_bytes = last_block - start;
1266 write_bytes, &lockstart,
1280 write_bytes, iter);
1283 if (unlikely(copied < write_bytes)) {
1889 size_t write_bytes = reserved_space;
1891 if (btrfs_check_nocow_lock(inode, page_start, &write_bytes, false) <= 0)
1902 if (write_bytes < reserved_space)