Lines Matching defs:csum_size

133 	return (bytes >> fs_info->sectorsize_bits) * fs_info->csum_size;
136 static size_t csum_size_to_bytes(const struct btrfs_fs_info *fs_info, u32 csum_size)
138 ASSERT(IS_ALIGNED(csum_size, fs_info->csum_size));
140 return (csum_size / fs_info->csum_size) << fs_info->sectorsize_bits;
146 fs_info->csum_size);
212 const u32 csum_size = fs_info->csum_size;
234 csums_in_item /= csum_size;
245 csum_offset * csum_size);
286 const u32 csum_size = fs_info->csum_size;
303 csum_len = (itemsize / csum_size) * sectorsize;
321 csum_len = (itemsize / csum_size) * sectorsize;
328 ret * csum_size);
347 const u32 csum_size = fs_info->csum_size;
375 if (nblocks * csum_size > BTRFS_BIO_INLINE_CSUM_SIZE) {
376 bbio->csum = kvcalloc(nblocks, csum_size, GFP_NOFS);
435 (bio_offset >> fs_info->sectorsize_bits) * csum_size;
458 memset(csum_dst, 0, csum_size);
792 index += fs_info->csum_size;
880 const u32 csum_size = fs_info->csum_size;
886 csum_end = btrfs_item_size(leaf, path->slots[0]) / csum_size;
898 new_size *= csum_size;
909 new_size *= csum_size;
933 const u32 csum_size = fs_info->csum_size;
969 csum_end = btrfs_item_size(leaf, path->slots[0]) / csum_size;
1033 offset *= csum_size;
1035 shift_len = (len >> blocksize_bits) * csum_size;
1116 const u32 csum_size = fs_info->csum_size;
1148 if ((item_size / csum_size) >=
1149 MAX_CSUM_ITEMS(fs_info, csum_size)) {
1172 if (btrfs_leaf_free_space(leaf) >= csum_size) {
1182 csum_size, 1);
1199 csum_offset >= MAX_CSUM_ITEMS(fs_info, csum_size)) {
1205 csum_size) {
1255 diff = (csum_offset + extend_nr) * csum_size;
1257 MAX_CSUM_ITEMS(fs_info, csum_size) * csum_size);
1261 diff /= csum_size;
1262 diff *= csum_size;
1281 tmp = min_t(u64, tmp, MAX_CSUM_ITEMS(fs_info, csum_size));
1282 ins_size = csum_size * tmp;
1284 ins_size = csum_size;
1296 csum_offset * csum_size);
1299 ins_size *= csum_size;
1306 ins_size /= csum_size;