/linux/arch/um/drivers/ |
H A D | cow.h | 8 int sectorsize, int alignment, int *bitmap_offset_out, 19 int sectorsize, int alignment, 22 extern void cow_sizes(int version, __u64 size, int sectorsize, int align,
|
H A D | cow_user.c | 29 __s32 sectorsize; member 48 __s32 sectorsize; member 91 __u32 sectorsize; member 103 __u32 sectorsize; member 125 void cow_sizes(int version, __u64 size, int sectorsize, int align, in cow_sizes() argument 130 *bitmap_len_out = (size + sectorsize - 1) / (8 * sectorsize); in cow_sizes() 133 *data_offset_out = (*data_offset_out + sectorsize - 1) / in cow_sizes() 134 sectorsize; in cow_sizes() 135 *data_offset_out *= sectorsize; in cow_sizes() 138 *bitmap_len_out = DIV_ROUND(size, sectorsize); in cow_sizes() [all …]
|
/linux/crypto/ |
H A D | Kconfig | 760 implementation currently can't handle a sectorsize which is not a
|
/linux/drivers/md/ |
H A D | dm-log-writes.c | 86 __le32 sectorsize; member 107 u32 sectorsize; member 240 lc->sectorsize - entrylen - datalen); in write_metadata() 243 ret = bio_add_page(bio, page, lc->sectorsize, 0); in write_metadata() 244 if (ret != lc->sectorsize) { in write_metadata() 282 pg_sectorlen = ALIGN(pg_datalen, lc->sectorsize); in write_inline_data() 408 super.sectorsize = cpu_to_le32(lc->sectorsize); in log_super() 550 lc->sectorsize = bdev_logical_block_size(lc->dev->bdev); in log_writes_ctr() 551 lc->sectorshift = ilog2(lc->sectorsize); in log_writes_ctr() [all...] |
/linux/drivers/mtd/devices/ |
H A D | spear_smi.c | 101 unsigned long sectorsize; member 111 .sectorsize = ssize, \ 894 flash->mtd.erasesize = flash_devices[flash_index].sectorsize; in spear_smi_setup_banks()
|
/linux/drivers/mtd/nand/raw/atmel/ |
H A D | nand-controller.c | 1150 req.ecc.sectorsize = chip->ecc.size; in atmel_nand_pmecc_init() 1152 req.ecc.sectorsize = requirements->step_size; in atmel_nand_pmecc_init() 1154 req.ecc.sectorsize = ATMEL_PMECC_SECTOR_SIZE_AUTO; in atmel_nand_pmecc_init() 1172 chip->ecc.size = req.ecc.sectorsize; in atmel_nand_pmecc_init()
|
H A D | pmecc.c | 226 if (req->ecc.sectorsize == 512) { in atmel_pmecc_create_gf_tables() 260 if (req->ecc.sectorsize == 512) in atmel_pmecc_get_gf_tables() 289 if (req->ecc.sectorsize == ATMEL_PMECC_SECTOR_SIZE_AUTO) { in atmel_pmecc_prepare_user_req() 294 req->ecc.sectorsize = 1024; in atmel_pmecc_prepare_user_req() 296 req->ecc.sectorsize = 512; in atmel_pmecc_prepare_user_req() 299 if (req->ecc.sectorsize != 512 && req->ecc.sectorsize != 1024) in atmel_pmecc_prepare_user_req() 302 if (req->pagesize % req->ecc.sectorsize) in atmel_pmecc_prepare_user_req() 305 req->ecc.nsectors = req->pagesize / req->ecc.sectorsize; in atmel_pmecc_prepare_user_req() 316 nbytes = DIV_ROUND_UP(strength * fls(8 * req->ecc.sectorsize), in atmel_pmecc_prepare_user_req() 400 if (req->ecc.sectorsize == 1024) in atmel_pmecc_create_user() [all …]
|
H A D | pmecc.h | 47 int sectorsize; member
|
/linux/fs/btrfs/ |
H A D | accessors.h | 134 WARN_ON(!IS_ALIGNED(val, eb->fs_info->sectorsize)); in btrfs_set_device_total_bytes() 880 sectorsize, 32);
|
H A D | bio.c | 221 repair_bbio->file_offset, fs_info->sectorsize, in btrfs_end_repair_bio() 245 const u32 sectorsize = fs_info->sectorsize; in repair_one_sector() local 255 num_copies = btrfs_num_copies(fs_info, logical, sectorsize); in repair_one_sector() 291 u32 sectorsize = fs_info->sectorsize; in btrfs_check_read_bio() local 315 bv.bv_len = min(bv.bv_len, sectorsize); in btrfs_check_read_bio() 319 bio_advance_iter_single(&bbio->bio, iter, sectorsize); in btrfs_check_read_bio() 320 offset += sectorsize; in btrfs_check_read_bio()
|
H A D | block-group.c | 492 fs_info->nodesize : fs_info->sectorsize; in fragment_free_space() 3232 cache_size *= fs_info->sectorsize; in cache_save_setup()
|
H A D | compression.c | 381 ASSERT(IS_ALIGNED(ordered->file_offset, fs_info->sectorsize)); in btrfs_submit_compressed_write() 382 ASSERT(IS_ALIGNED(ordered->num_bytes, fs_info->sectorsize)); in btrfs_submit_compressed_write() 443 if (fs_info->sectorsize < PAGE_SIZE) in add_ra_bio_pages() 509 (cur + fs_info->sectorsize > extent_map_end(em)) || in add_ra_bio_pages() 542 if (fs_info->sectorsize < PAGE_SIZE) in add_ra_bio_pages() 580 em = lookup_extent_mapping(em_tree, file_offset, fs_info->sectorsize); in btrfs_submit_compressed_read() 1064 const u32 sectorsize = fs_info->sectorsize; in btrfs_decompress() local 1072 ASSERT(dest_pgoff + destlen <= PAGE_SIZE && destlen <= sectorsize); in btrfs_decompress()
|
H A D | defrag.c | 286 cur = max(cur + fs_info->sectorsize, range.start); in __btrfs_run_defrag_inode() 759 const u32 sectorsize = BTRFS_I(inode)->root->fs_info->sectorsize; in defrag_lookup_extent() local 766 em = lookup_extent_mapping(em_tree, start, sectorsize); in defrag_lookup_extent() 785 u64 end = start + sectorsize - 1; in defrag_lookup_extent() 1209 const u32 sectorsize = inode->root->fs_info->sectorsize; in defrag_one_range() local 1217 ASSERT(IS_ALIGNED(start, sectorsize) && IS_ALIGNED(len, sectorsize)); in defrag_one_range() 1284 const u32 sectorsize = inode->root->fs_info->sectorsize; in defrag_one_cluster() local 1307 (max_sectors - *sectors_defragged) * sectorsize); in defrag_one_cluster() 1404 cur = round_down(range->start, fs_info->sectorsize); in btrfs_defrag_file() 1405 last_byte = round_up(last_byte, fs_info->sectorsize) - 1; in btrfs_defrag_file()
|
H A D | delalloc-space.c | 121 bytes = ALIGN(bytes, fs_info->sectorsize); in btrfs_alloc_data_chunk_ondemand() 138 len = round_up(start + len, fs_info->sectorsize) - in btrfs_check_data_free_space() 139 round_down(start, fs_info->sectorsize); in btrfs_check_data_free_space() 140 start = round_down(start, fs_info->sectorsize); in btrfs_check_data_free_space() 176 ASSERT(IS_ALIGNED(len, fs_info->sectorsize)); in btrfs_free_reserved_data_space_noquota() 195 len = round_up(start + len, fs_info->sectorsize) - in btrfs_free_reserved_data_space() 196 round_down(start, fs_info->sectorsize); in btrfs_free_reserved_data_space() 197 start = round_down(start, fs_info->sectorsize); in btrfs_free_reserved_data_space() 332 num_bytes = ALIGN(num_bytes, fs_info->sectorsize); in btrfs_delalloc_reserve_metadata() 333 disk_num_bytes = ALIGN(disk_num_bytes, fs_info->sectorsize); in btrfs_delalloc_reserve_metadata() [all …]
|
H A D | delayed-inode.c | 1892 round_up(i_size_read(inode), fs_info->sectorsize)); in btrfs_fill_inode()
|
H A D | dev-replace.c | 219 dev_replace->tgtdev->io_width = fs_info->sectorsize; in btrfs_init_dev_replace() 220 dev_replace->tgtdev->io_align = fs_info->sectorsize; in btrfs_init_dev_replace() 221 dev_replace->tgtdev->sector_size = fs_info->sectorsize; in btrfs_init_dev_replace() 305 device->io_width = fs_info->sectorsize; in btrfs_init_dev_replace_tgtdev() 306 device->io_align = fs_info->sectorsize; in btrfs_init_dev_replace_tgtdev() 307 device->sector_size = fs_info->sectorsize; in btrfs_init_dev_replace_tgtdev()
|
H A D | disk-io.c | 554 if (fs_info->sectorsize == PAGE_SIZE) { in btree_dirty_folio() 578 cur = page_start + cur_bit * fs_info->sectorsize; in btree_dirty_folio() 742 /* We don't use the stripesize in selftest, set it as sectorsize */ in btrfs_alloc_dummy_root() 2357 u64 sectorsize = btrfs_super_sectorsize(sb); in btrfs_validate_super() local 2386 * Check sectorsize and nodesize first, other check will need it. in btrfs_validate_super() 2387 * Check all possible sectorsize(4K, 8K, 16K, 32K, 64K) here. in btrfs_validate_super() 2389 if (!is_power_of_2(sectorsize) || sectorsize < 4096 || in btrfs_validate_super() 2390 sectorsize > BTRFS_MAX_METADATA_BLOCKSIZE) { in btrfs_validate_super() 2391 btrfs_err(fs_info, "invalid sectorsize in btrfs_validate_super() 3207 u32 sectorsize; open_ctree() local [all...] |
H A D | extent-tree.c | 372 if (offset && IS_ALIGNED(offset, fs_info->sectorsize)) in btrfs_get_extent_inline_ref_type() 385 IS_ALIGNED(offset, fs_info->sectorsize)) in btrfs_get_extent_inline_ref_type() 4374 WARN_ON(ffe_ctl->num_bytes < fs_info->sectorsize); in find_free_extent() 4677 WARN_ON(num_bytes < fs_info->sectorsize); in btrfs_reserve_extent() 4696 fs_info->sectorsize); in btrfs_reserve_extent()
|
H A D | extent_io.c | 468 const u32 sectorsize = fs_info->sectorsize; in end_bbio_data_write() local 480 if (!IS_ALIGNED(fi.offset, sectorsize)) in end_bbio_data_write() 484 else if (!IS_ALIGNED(fi.length, sectorsize)) in end_bbio_data_write() 598 const u32 sectorsize = fs_info->sectorsize; in end_bbio_data_read() local 623 if (!IS_ALIGNED(fi.offset, sectorsize)) in end_bbio_data_read() 627 else if (!IS_ALIGNED(fi.offset + fi.length, sectorsize)) in end_bbio_data_read() 1038 size_t blocksize = fs_info->sectorsize; in btrfs_do_readpage() 1063 ASSERT(IS_ALIGNED(cur, fs_info->sectorsize)); in btrfs_do_readpage() 1329 *start = page_offset(page) + range_start_bit * fs_info->sectorsize; in find_next_dirty_byte() 1330 *end = page_offset(page) + range_end_bit * fs_info->sectorsize; in find_next_dirty_byte() [all …]
|
H A D | file-item.c | 90 ASSERT(IS_ALIGNED(start + len, inode->root->fs_info->sectorsize)); in btrfs_inode_set_file_extent_range() 118 ASSERT(IS_ALIGNED(start + len, inode->root->fs_info->sectorsize) || in btrfs_inode_clear_file_extent_range() 129 ASSERT(IS_ALIGNED(bytes, fs_info->sectorsize)); in bytes_to_csum_size() 285 const u32 sectorsize = fs_info->sectorsize; in search_csum_tree() local 292 ASSERT(IS_ALIGNED(disk_bytenr, sectorsize) && in search_csum_tree() 293 IS_ALIGNED(len, sectorsize)); in search_csum_tree() 303 csum_len = (itemsize / csum_size) * sectorsize; in search_csum_tree() 321 csum_len = (itemsize / csum_size) * sectorsize; in search_csum_tree() 346 const u32 sectorsize = fs_info->sectorsize; in btrfs_lookup_bio_sums() local 437 file_offset + sectorsize - 1, in btrfs_lookup_bio_sums() [all …]
|
H A D | file.c | 101 u64 block_start = round_down(pos, fs_info->sectorsize); in btrfs_drop_pages() 102 u64 block_len = round_up(pos + copied, fs_info->sectorsize) - block_start; in btrfs_drop_pages() 146 start_pos = round_down(pos, fs_info->sectorsize); in btrfs_dirty_pages() 148 fs_info->sectorsize); in btrfs_dirty_pages() 466 fs_info->sectorsize); in btrfs_drop_extents() 1002 start_pos = round_down(pos, fs_info->sectorsize); in lock_and_cleanup_extent_if_need() 1003 last_pos = round_up(pos + write_bytes, fs_info->sectorsize) - 1; in lock_and_cleanup_extent_if_need() 1091 lockstart = round_down(pos, fs_info->sectorsize); in btrfs_check_nocow_lock() 1093 fs_info->sectorsize) - 1; in btrfs_check_nocow_lock() 1176 start_pos = round_down(pos, fs_info->sectorsize); in btrfs_write_check() [all …]
|
H A D | free-space-cache.c | 2286 if (info->bytes <= fs_info->sectorsize * 8) { in use_bitmap() 2952 ctl->unit = fs_info->sectorsize; in btrfs_init_free_space_ctl() 3576 min_bytes = fs_info->sectorsize; in btrfs_find_space_cluster() 3579 min_bytes = fs_info->sectorsize; in btrfs_find_space_cluster()
|
H A D | free-space-tree.c | 52 bitmap_range = cache->fs_info->sectorsize * BTRFS_FREE_SPACE_BITMAP_BITS; in set_free_space_tree_thresholds() 261 fs_info->sectorsize); in convert_free_space_to_bitmaps() 263 fs_info->sectorsize); in convert_free_space_to_bitmaps() 304 bitmap_range = fs_info->sectorsize * BTRFS_FREE_SPACE_BITMAP_BITS; in convert_free_space_to_bitmaps() 400 fs_info->sectorsize * in convert_free_space_to_extents() 443 key.objectid = start + start_bit * block_group->fs_info->sectorsize; in convert_free_space_to_extents() 445 key.offset = (end_bit - start_bit) * block_group->fs_info->sectorsize; in convert_free_space_to_extents() 532 block_group->fs_info->sectorsize); in free_space_test_bit() 622 u64 prev_block = start - block_group->fs_info->sectorsize; in modify_free_space_bitmap() 1543 offset += fs_info->sectorsize; in load_free_space_bitmaps()
|
H A D | fs.h | 789 u32 sectorsize; member
|
H A D | inode-item.c | 585 fs_info->sectorsize); in btrfs_truncate_inode_items() 586 clear_start = ALIGN(new_size, fs_info->sectorsize); in btrfs_truncate_inode_items() 633 clear_len = fs_info->sectorsize; in btrfs_truncate_inode_items()
|