| /linux/fs/btrfs/tests/ |
| H A D | free-space-tests.c | 90 static int test_bitmaps(struct btrfs_block_group *cache, u32 sectorsize) in test_bitmaps() argument 130 next_bitmap_offset = (u64)(BITS_PER_BITMAP * sectorsize); in test_bitmaps() 159 u32 sectorsize) in test_bitmaps_and_extents() argument 161 u64 bitmap_offset = (u64)(BITS_PER_BITMAP * sectorsize); in test_bitmaps_and_extents() 396 u32 sectorsize) in test_steal_space_from_bitmap_to_extent() argument 533 ret = btrfs_add_free_space(cache, SZ_128M + SZ_16M, sectorsize); in test_steal_space_from_bitmap_to_extent() 591 if (cache->free_space_ctl->free_space != (SZ_1M + sectorsize)) { in test_steal_space_from_bitmap_to_extent() 592 test_err("cache free space is not 1Mb + %u", sectorsize); in test_steal_space_from_bitmap_to_extent() 614 if (cache->free_space_ctl->free_space != sectorsize) { in test_steal_space_from_bitmap_to_extent() 615 test_err("cache free space is not %u", sectorsize); in test_steal_space_from_bitmap_to_extent() [all …]
|
| H A D | btrfs-tests.c | 118 struct btrfs_fs_info *btrfs_alloc_dummy_fs_info(u32 nodesize, u32 sectorsize) in btrfs_alloc_dummy_fs_info() argument 141 fs_info->sectorsize = sectorsize; in btrfs_alloc_dummy_fs_info() 142 fs_info->sectorsize_bits = ilog2(sectorsize); in btrfs_alloc_dummy_fs_info() 219 cache->full_stripe_len = fs_info->sectorsize; in btrfs_alloc_dummy_block_group() 261 u32 sectorsize, nodesize; in btrfs_run_sanity_tests() local 269 sectorsize = test_sectorsize[i]; in btrfs_run_sanity_tests() 270 for (nodesize = sectorsize; in btrfs_run_sanity_tests() 273 pr_info("BTRFS: selftest: sectorsize: %u nodesize: %u\n", in btrfs_run_sanity_tests() 274 sectorsize, nodesiz in btrfs_run_sanity_tests() [all...] |
| H A D | extent-buffer-tests.c | 13 static int test_btrfs_split_item(u32 sectorsize, u32 nodesize) in test_btrfs_split_item() argument 31 fs_info = btrfs_alloc_dummy_fs_info(nodesize, sectorsize); in test_btrfs_split_item() 219 int btrfs_test_extent_buffer_operations(u32 sectorsize, u32 nodesize) in btrfs_test_extent_buffer_operations() argument 222 return test_btrfs_split_item(sectorsize, nodesize); in btrfs_test_extent_buffer_operations()
|
| H A D | qgroup-tests.c | 197 u32 sectorsize, u32 nodesize) in test_no_shared_qgroup() argument 309 u32 sectorsize, u32 nodesize) in test_multiple_refs() argument 462 int btrfs_test_qgroups(u32 sectorsize, u32 nodesize) in btrfs_test_qgroups() argument 469 fs_info = btrfs_alloc_dummy_fs_info(nodesize, sectorsize); in btrfs_test_qgroups() 542 ret = test_no_shared_qgroup(root, sectorsize, nodesize); in btrfs_test_qgroups() 545 ret = test_multiple_refs(root, sectorsize, nodesize); in btrfs_test_qgroups()
|
| /linux/arch/um/drivers/ |
| H A D | cow_user.c | 29 __s32 sectorsize; member 48 __s32 sectorsize; member 91 __u32 sectorsize; member 103 __u32 sectorsize; member 125 void cow_sizes(int version, __u64 size, int sectorsize, int align, in cow_sizes() argument 130 *bitmap_len_out = (size + sectorsize - 1) / (8 * sectorsize); in cow_sizes() 133 *data_offset_out = (*data_offset_out + sectorsize - 1) / in cow_sizes() 134 sectorsize; in cow_sizes() 135 *data_offset_out *= sectorsize; in cow_sizes() 138 *bitmap_len_out = DIV_ROUND(size, sectorsize); in cow_sizes() [all …]
|
| H A D | cow.h | 8 int sectorsize, int alignment, int *bitmap_offset_out, 19 int sectorsize, int alignment, 22 extern void cow_sizes(int version, __u64 size, int sectorsize, int align,
|
| H A D | ubd_kern.c | 63 int sectorsize; member 610 int fd, err, sectorsize, asked_switch, mode = 0644; in open_ubd_file() local 641 &size, §orsize, &align, bitmap_offset_out); in open_ubd_file() 659 sectorsize, align, &size); in open_ubd_file() 671 cow_sizes(version, size, sectorsize, align, *bitmap_offset_out, in open_ubd_file() 682 int sectorsize, int alignment, int *bitmap_offset_out, in create_cow_file() argument 696 err = init_cow_file(fd, cow_file, backing_file, sectorsize, alignment, in create_cow_file() 1242 io_req->sectorsize = SECTOR_SIZE; in ubd_alloc_req() 1422 nsectors = desc->length / req->sectorsize; in do_io() 1432 start * req->sectorsize; in do_io() [all …]
|
| /linux/fs/btrfs/ |
| H A D | lzo.c | 47 * Variable size. Size up limit should be lzo1x_worst_compress(sectorsize) 48 * which is 4419 for a 4KiB sectorsize. 50 * Example with 4K sectorsize: 70 return lzo1x_worst_compress(fs_info->sectorsize); in workspace_buf_length() 74 return lzo1x_worst_compress(fs_info->sectorsize); in workspace_cbuf_length() 183 const u32 sectorsize = fs_info->sectorsize; 222 u32 copy_len = min_t(u32, sectorsize - *total_out % sectorsize, in copy_compressed_data_to_bio() 249 sector_bytes_left = round_up(*total_out, sectorsize) in copy_compressed_data_to_bio() 199 const u32 sectorsize = fs_info->sectorsize; copy_compressed_data_to_bio() local 284 const u32 sectorsize = fs_info->sectorsize; lzo_compress_bio() local 433 const u32 sectorsize = fs_info->sectorsize; lzo_decompress_bio() local 553 const u32 sectorsize = fs_info->sectorsize; lzo_decompress() local [all...] |
| H A D | tree-checker.c | 129 end = ALIGN(key->offset + len, leaf->fs_info->sectorsize); in file_extent_end() 213 u32 sectorsize = fs_info->sectorsize; in check_extent_data_item() local 217 if (unlikely(!IS_ALIGNED(key->offset, sectorsize))) { in check_extent_data_item() 220 key->offset, sectorsize); in check_extent_data_item() 306 if (unlikely(CHECK_FE_ALIGNED(leaf, slot, fi, ram_bytes, sectorsize) || in check_extent_data_item() 307 CHECK_FE_ALIGNED(leaf, slot, fi, disk_bytenr, sectorsize) || in check_extent_data_item() 308 CHECK_FE_ALIGNED(leaf, slot, fi, disk_num_bytes, sectorsize) || in check_extent_data_item() 309 CHECK_FE_ALIGNED(leaf, slot, fi, offset, sectorsize) || in check_extent_data_item() 310 CHECK_FE_ALIGNED(leaf, slot, fi, num_bytes, sectorsize))) in check_extent_data_item() 369 u32 sectorsize = fs_info->sectorsize; check_csum_item() local 863 btrfs_check_chunk_valid(const struct btrfs_fs_info * fs_info,const struct extent_buffer * leaf,const struct btrfs_chunk * chunk,u64 logical,u32 sectorsize) btrfs_check_chunk_valid() argument 1888 const u32 sectorsize = leaf->fs_info->sectorsize; check_dev_extent_item() local [all...] |
| H A D | bio.c | 188 const u32 step = min(fs_info->sectorsize, PAGE_SIZE); in btrfs_end_repair_bio() 197 ASSERT(repair_bbio->saved_iter.bi_size == fs_info->sectorsize); in btrfs_end_repair_bio() 224 repair_bbio->file_offset, fs_info->sectorsize, in btrfs_end_repair_bio() 247 const u32 sectorsize = fs_info->sectorsize; in repair_one_sector() 248 const u32 step = min(fs_info->sectorsize, PAGE_SIZE); in repair_one_sector() 249 const u32 nr_steps = sectorsize / step; in repair_one_sector() 255 sectorsize); in repair_one_sector() 264 num_copies = btrfs_num_copies(fs_info, logical, sectorsize); in repair_one_sector() 307 const u32 sectorsize in btrfs_check_read_bio() 246 const u32 sectorsize = fs_info->sectorsize; repair_one_sector() local 306 const u32 sectorsize = fs_info->sectorsize; btrfs_check_read_bio() local [all...] |
| H A D | file-item.c | 80 * The start and len must match the file extent item, so thus must be sectorsize 92 ASSERT(IS_ALIGNED(start + len, inode->root->fs_info->sectorsize)); in btrfs_inode_set_file_extent_range() 109 * The start and len must match the file extent item, so thus must be sectorsize 121 ASSERT(IS_ALIGNED(start + len, inode->root->fs_info->sectorsize) || in btrfs_inode_clear_file_extent_range() 130 ASSERT(IS_ALIGNED(bytes, fs_info->sectorsize)); in bytes_to_csum_size() 273 * Return 0 for the range [disk_bytenr, disk_bytenr + sectorsize) has no csum 284 const u32 sectorsize = fs_info->sectorsize; in search_csum_tree() local 291 ASSERT(IS_ALIGNED(disk_bytenr, sectorsize) && in search_csum_tree() 292 IS_ALIGNED(len, sectorsize)); in search_csum_tree() 352 const u32 sectorsize = fs_info->sectorsize; btrfs_lookup_bio_sums() local [all...] |
| H A D | file.c | 48 u64 block_start = round_down(pos, fs_info->sectorsize); in btrfs_drop_folio() 49 u64 block_len = round_up(pos + copied, fs_info->sectorsize) - block_start; in btrfs_drop_folio() 89 start_pos = round_down(pos, fs_info->sectorsize); in btrfs_dirty_folio() 90 num_bytes = round_up(end_pos - start_pos, fs_info->sectorsize); in btrfs_dirty_folio() 415 fs_info->sectorsize); in btrfs_drop_extents() 803 const u32 blocksize = inode_to_fs_info(inode)->sectorsize; in prepare_uptodate_folio() 908 start_pos = round_down(pos, fs_info->sectorsize); in lock_and_cleanup_extent_if_need() 909 last_pos = round_up(pos + write_bytes, fs_info->sectorsize) - 1; in lock_and_cleanup_extent_if_need() 993 lockstart = round_down(pos, fs_info->sectorsize); in btrfs_check_nocow_lock() 995 fs_info->sectorsize) in btrfs_check_nocow_lock() 2898 const u64 sectorsize = inode->root->fs_info->sectorsize; btrfs_zero_range_check_range_boundary() local 2928 const u64 sectorsize = fs_info->sectorsize; btrfs_zero_range() local [all...] |
| H A D | raid56.c | 233 const u32 step = min(rbio->bioc->fs_info->sectorsize, PAGE_SIZE); in memcpy_from_bio_to_stripe() 329 nr_sectors = round_up(PAGE_SIZE, fs_info->sectorsize) >> fs_info->sectorsize_bits; in page_nr_to_num_sectors() 358 const u32 step = min(rbio->bioc->fs_info->sectorsize, PAGE_SIZE); in index_stripe_sectors() 398 * We have ensured PAGE_SIZE is aligned with sectorsize, thus in is_data_stripe_page() 1050 const unsigned int step = min(fs_info->sectorsize, PAGE_SIZE); in sector_paddr_in_rbio() 1051 const unsigned int sector_nsteps = fs_info->sectorsize / step; in sector_paddr_in_rbio() 1058 ASSERT(IS_ALIGNED(PAGE_SIZE, fs_info->sectorsize) || 1059 IS_ALIGNED(fs_info->sectorsize, PAGE_SIZE)); 1221 const u32 sectorsize = rbio->bioc->fs_info->sectorsize; in bio_add_paddrs() 1241 const u32 sectorsize = rbio->bioc->fs_info->sectorsize; rbio_add_io_paddrs() local 1618 const u32 sectorsize = rbio->bioc->fs_info->sectorsize; set_bio_pages_uptodate() local 1840 const u32 sectorsize = fs_info->sectorsize; rbio_add_bio() local [all...] |
| H A D | defrag.c | 266 cur = max(cur + fs_info->sectorsize, range.start); in btrfs_run_defrag_inode() 740 const u32 sectorsize = BTRFS_I(inode)->root->fs_info->sectorsize; in defrag_lookup_extent() local 747 em = btrfs_lookup_extent_mapping(em_tree, start, sectorsize); in defrag_lookup_extent() 766 u64 end = start + sectorsize - 1; in defrag_lookup_extent() 1200 const u32 sectorsize = inode->root->fs_info->sectorsize; in defrag_one_range() local 1207 ASSERT(IS_ALIGNED(start, sectorsize) && IS_ALIGNED(len, sectorsize)); in defrag_one_range() 1278 const u32 sectorsize = inode->root->fs_info->sectorsize; in defrag_one_cluster() local 1301 (max_sectors - *sectors_defragged) * sectorsize); in defrag_one_cluster() 1414 cur = round_down(range->start, fs_info->sectorsize); in btrfs_defrag_file() 1415 last_byte = round_up(last_byte, fs_info->sectorsize) - 1; in btrfs_defrag_file()
|
| H A D | reflink.c | 63 const u32 block_size = fs_info->sectorsize; in copy_inline_to_page() 179 fs_info->sectorsize); in clone_copy_inline_extent() 566 ASSERT(datal <= fs_info->sectorsize); in btrfs_clone() 569 WARN_ON(datal > fs_info->sectorsize)) { in btrfs_clone() 602 fs_info->sectorsize); in btrfs_clone() 683 const u64 bs = fs_info->sectorsize; in btrfs_extent_same() 754 u64 bs = fs_info->sectorsize; in btrfs_clone_files() 833 u64 bs = inode_out->root->fs_info->sectorsize; in btrfs_remap_file_range_prep()
|
| H A D | extent_io.c | 401 ASSERT(max_bytes >= fs_info->sectorsize); in find_lock_delalloc_range() 440 max_bytes = fs_info->sectorsize; in find_lock_delalloc_range() 1004 const size_t blocksize = fs_info->sectorsize; in btrfs_do_readpage() 1034 ASSERT(IS_ALIGNED(cur, fs_info->sectorsize)); in btrfs_do_readpage() 1172 const u32 blocksize = fs_info->sectorsize; in can_skip_one_ordered_range() 1639 const u32 sectorsize = fs_info->sectorsize; in submit_one_sector() 1641 ASSERT(IS_ALIGNED(filepos, sectorsize)); in submit_one_sector() 1646 em = btrfs_get_extent(inode, NULL, filepos, sectorsize); in submit_one_sector() 1660 btrfs_folio_clear_dirty(fs_info, folio, filepos, sectorsize); in submit_one_sector() 523 const u32 sectorsize = fs_info->sectorsize; end_bbio_data_write() local 1644 const u32 sectorsize = fs_info->sectorsize; submit_one_sector() local 2635 const u32 sectorsize = fs_info->sectorsize; extent_write_locked_range() local [all...] |
| H A D | disk-io.c | 518 if (fs_info->sectorsize == PAGE_SIZE) { in btree_dirty_folio() 542 cur = page_start + cur_bit * fs_info->sectorsize; in btree_dirty_folio() 685 /* We don't use the stripesize in selftest, set it as sectorsize */ in btrfs_alloc_dummy_root() 2292 * to use super block sectorsize, which should have been validated. in validate_sys_chunk_array() 2294 const u32 sectorsize = btrfs_super_sectorsize(sb); in validate_sys_chunk_array() 2347 sectorsize); in validate_sys_chunk_array() 2374 u64 sectorsize = btrfs_super_sectorsize(sb); in btrfs_validate_super() 2411 * Check sectorsize and nodesize first, other check will need it. in btrfs_validate_super() 2412 * Check all possible sectorsize(4K, 8K, 16K, 32K, 64K) here. in btrfs_validate_super() 2414 if (unlikely(!is_power_of_2(sectorsize) || sectorsiz in btrfs_validate_super() 2293 const u32 sectorsize = btrfs_super_sectorsize(sb); validate_sys_chunk_array() local 2364 u64 sectorsize = btrfs_super_sectorsize(sb); btrfs_validate_super() local 3295 u32 sectorsize; open_ctree() local [all...] |
| H A D | compression.c | 321 ASSERT(IS_ALIGNED(ordered->file_offset, fs_info->sectorsize)); in btrfs_submit_compressed_write() 322 ASSERT(IS_ALIGNED(ordered->num_bytes, fs_info->sectorsize)); in btrfs_submit_compressed_write() 401 if (fs_info->sectorsize < PAGE_SIZE) in add_ra_bio_pages() 488 (cur + fs_info->sectorsize > btrfs_extent_map_end(em)) || in add_ra_bio_pages() 522 if (fs_info->sectorsize < PAGE_SIZE) 571 em = btrfs_lookup_extent_mapping(em_tree, file_offset, fs_info->sectorsize); in btrfs_submit_compressed_read() 1025 * The bi_size may not be sectorsize aligned, thus the caller still need in btrfs_compress_bio() 1110 const u32 sectorsize = fs_info->sectorsize; in btrfs_alloc_compress_wsm() 1115 * And the @destlen should not exceed sectorsize, a in btrfs_alloc_compress_wsm() 1085 const u32 sectorsize = fs_info->sectorsize; btrfs_decompress() local [all...] |
| H A D | zstd.c | 385 workspace->buf = kmalloc(fs_info->sectorsize, GFP_KERNEL); in zstd_alloc_workspace() 464 if (tot_in + workspace->in_buf.pos > fs_info->sectorsize * 2 && in zstd_compress_bio() 620 workspace->out_buf.size = fs_info->sectorsize; in zstd_decompress_bio() 705 workspace->out_buf.size = fs_info->sectorsize; in zstd_decompress() 685 const u32 sectorsize = fs_info->sectorsize; zstd_decompress() local
|
| H A D | print-tree.c | 161 if (!IS_ALIGNED(offset, eb->fs_info->sectorsize)) in print_extent_item() 163 "\t\t\t(parent %llu not aligned to sectorsize %u)\n", in print_extent_item() 164 offset, eb->fs_info->sectorsize); in print_extent_item() 176 * aligned to sectorsize. in print_extent_item() 178 if (!IS_ALIGNED(offset, eb->fs_info->sectorsize)) in print_extent_item() 180 "\t\t\t(parent %llu not aligned to sectorsize %u)\n", in print_extent_item() 181 offset, eb->fs_info->sectorsize); in print_extent_item() 339 const u32 csum_bytes = (size / fs_info->csum_size) * fs_info->sectorsize; in print_extent_csum()
|
| H A D | free-space-tree.c | 51 bitmap_range = cache->fs_info->sectorsize * BTRFS_FREE_SPACE_BITMAP_BITS; in btrfs_set_free_space_tree_thresholds() 253 fs_info->sectorsize); in btrfs_convert_free_space_to_bitmaps() 255 fs_info->sectorsize); in btrfs_convert_free_space_to_bitmaps() 300 bitmap_range = fs_info->sectorsize * BTRFS_FREE_SPACE_BITMAP_BITS; in btrfs_convert_free_space_to_bitmaps() 395 fs_info->sectorsize * in btrfs_convert_free_space_to_extents() 442 key.objectid = start + start_bit * fs_info->sectorsize; in btrfs_convert_free_space_to_extents() 444 key.offset = (end_bit - start_bit) * fs_info->sectorsize; in btrfs_convert_free_space_to_extents() 528 block_group->fs_info->sectorsize); in btrfs_free_space_test_bit() 619 u64 prev_block = start - block_group->fs_info->sectorsize; in modify_free_space_bitmap() 1594 offset += fs_info->sectorsize; in load_free_space_bitmaps()
|
| H A D | fiemap.c | 644 const u64 sectorsize = inode->root->fs_info->sectorsize; in extent_fiemap() local 659 range_start = round_down(start, sectorsize); in extent_fiemap() 660 range_end = round_up(start + len, sectorsize); in extent_fiemap()
|
| H A D | inode.c | 193 fs_info->sectorsize, nlink, in data_reloc_print_warning_inode() 468 const u32 sectorsize = trans->fs_info->sectorsize; in insert_inline_extent() 481 ASSERT(size <= sectorsize); in insert_inline_extent() 488 ASSERT(compressed_size <= sectorsize); in insert_inline_extent() 542 * We align size to sectorsize for inline extents just for simplicity in insert_inline_extent() 546 ALIGN(size, root->fs_info->sectorsize)); in insert_inline_extent() 590 /* Inline extents are limited to sectorsize. */ in can_cow_file_range_inline() 591 if (size > fs_info->sectorsize) in can_cow_file_range_inline() 595 if (data_len >= fs_info->sectorsize) in can_cow_file_range_inline() 469 const u32 sectorsize = trans->fs_info->sectorsize; insert_inline_extent() local 3051 const u64 sectorsize = root->fs_info->sectorsize; insert_reserved_file_extent() local [all...] |
| /linux/drivers/md/ |
| H A D | dm-log-writes.c | 86 __le32 sectorsize; member 107 u32 sectorsize; member 240 lc->sectorsize - entrylen - datalen); in write_metadata() 243 ret = bio_add_page(bio, page, lc->sectorsize, 0); in write_metadata() 244 if (ret != lc->sectorsize) { in write_metadata() 282 pg_sectorlen = ALIGN(pg_datalen, lc->sectorsize); in write_inline_data() 408 super.sectorsize = cpu_to_le32(lc->sectorsize); in log_super() 551 lc->sectorsize = bdev_logical_block_size(lc->dev->bdev); in log_writes_ctr() 552 lc->sectorshift = ilog2(lc->sectorsize); in log_writes_ctr() 567 lc->next_sector = lc->sectorsize >> SECTOR_SHIFT; in log_writes_ctr() [all …]
|
| /linux/drivers/mtd/nand/raw/atmel/ |
| H A D | pmecc.h | 47 int sectorsize; member
|