| /linux/block/ |
| H A D | badblocks.c | 554 sector_t sectors = bad->len; in front_merge() local 562 merged = min_t(sector_t, sectors, BB_END(p[prev]) - s); in front_merge() 564 merged = min_t(sector_t, sectors, BB_MAX_LEN - BB_LEN(p[prev])); in front_merge() 839 static bool _badblocks_set(struct badblocks *bb, sector_t s, sector_t sectors, in _badblocks_set() argument 852 if (sectors == 0) in _badblocks_set() 858 sector_t next = s + sectors; in _badblocks_set() 862 sectors = next - s; in _badblocks_set() 872 bad.len = sectors; in _badblocks_set() 923 BB_END(p[prev]) - s, sectors); in _badblocks_set() 954 sectors -= len; in _badblocks_set() [all …]
|
| /linux/Documentation/admin-guide/device-mapper/ |
| H A D | dm-ebs.rst | 8 size. Its main purpose is to provide emulation of 512 byte sectors on 18 <dev path> <offset> <emulated sectors> [<underlying sectors>] 27 has to be a multiple of <emulated sectors>. 28 <emulated sectors>: 29 Number of sectors defining the logical block size to be emulated; 30 1, 2, 4, 8 sectors of 512 bytes supported. 34 <underlying sectors>: 35 Number of sectors defining the logical block size of <dev path>. 36 2^N supported, e.g. 8 = emulate 8 sectors of 512 bytes = 4KiB. 43 offset 1024 sectors with underlying devices block size automatically set: [all …]
|
| H A D | dm-integrity.rst | 68 dm-integrity won't read of write these sectors 77 not used and data sectors and integrity tags are written 96 data directly in the underlying device sectors. 111 The number of interleaved sectors. This values is rounded down to 120 The number of sectors in one metadata buffer. The value is rounded 183 512-byte sectors that corresponds to one bitmap bit. 202 copy sectors from one journal section to another journal section 206 key and also to disallow the attacker to move sectors from one 224 256 sectors of metadata per data area. With the default buffer_sectors of 231 2. provided data sectors - that is the number of sectors that the user [all …]
|
| H A D | dm-dust.rst | 4 This target emulates the behavior of bad sectors at arbitrary 11 drive with bad sectors). 23 sectors. 25 Normally, a drive that is encountering bad sectors will most likely 26 encounter more bad sectors, at an unknown time or location. 32 simulating a "failure" event where bad sectors start to appear. 53 First, find the size (in 512-byte sectors) of the device to be used:: 114 kernel: device-mapper: dust: enabling read failures on bad sectors 299 be bad sectors that occur on the device in the future, in a different,
|
| /linux/drivers/target/ |
| H A D | target_core_sbc.c | 216 static inline u32 sbc_get_size(struct se_cmd *cmd, u32 sectors) in sbc_get_size() argument 218 return cmd->se_dev->dev_attrib.block_size * sectors; in sbc_get_size() 279 unsigned int sectors = sbc_get_write_same_sectors(cmd); in sbc_setup_write_same() local 288 if (sectors > cmd->se_dev->dev_attrib.max_write_same_len) { in sbc_setup_write_same() 290 sectors, cmd->se_dev->dev_attrib.max_write_same_len); in sbc_setup_write_same() 296 if (((cmd->t_task_lba + sectors) < cmd->t_task_lba) || in sbc_setup_write_same() 297 ((cmd->t_task_lba + sectors) > end_lba)) { in sbc_setup_write_same() 299 (unsigned long long)end_lba, cmd->t_task_lba, sectors); in sbc_setup_write_same() 333 ret = sbc_check_prot(dev, cmd, flags >> 5, sectors, true); in sbc_setup_write_same() 668 u32 sectors, bool is_write) in sbc_check_prot() argument [all …]
|
| /linux/drivers/md/ |
| H A D | raid0.c | 68 sector_t curr_zone_end, sectors; in create_strip_zones() local 88 sectors = rdev1->sectors; in create_strip_zones() 89 sector_div(sectors, mddev->chunk_sectors); in create_strip_zones() 90 rdev1->sectors = sectors * mddev->chunk_sectors; in create_strip_zones() 101 (unsigned long long)rdev1->sectors, in create_strip_zones() 103 (unsigned long long)rdev2->sectors); in create_strip_zones() 109 if (rdev2->sectors == rdev1->sectors) { in create_strip_zones() 198 if (!smallest || (rdev1->sectors < smallest->sectors)) in create_strip_zones() 208 zone->zone_end = smallest->sectors * cnt; in create_strip_zones() 221 zone->dev_start = smallest->sectors; in create_strip_zones() [all …]
|
| H A D | md-linear.c | 55 static sector_t linear_size(struct mddev *mddev, sector_t sectors, int raid_disks) in linear_size() argument 61 WARN_ONCE(sectors || raid_disks, in linear_size() 118 sector_t sectors; in linear_conf() local 128 sectors = rdev->sectors; in linear_conf() 129 sector_div(sectors, mddev->chunk_sectors); in linear_conf() 130 rdev->sectors = sectors * mddev->chunk_sectors; in linear_conf() 133 conf->array_sectors += rdev->sectors; in linear_conf() 145 conf->disks[0].end_sector = conf->disks[0].rdev->sectors; in linear_conf() 150 conf->disks[i].rdev->sectors; in linear_conf() 246 start_sector = tmp_dev->end_sector - tmp_dev->rdev->sectors; in linear_make_request() [all …]
|
| H A D | raid1.c | 65 sector_t hi = lo + r1_bio->sectors; in check_and_add_serial() 340 r1_bio->sector + (r1_bio->sectors); in update_head_pos() 455 sector_t hi = r1_bio->sector + r1_bio->sectors; in raid1_end_write_request() 512 if (rdev_has_badblock(rdev, r1_bio->sector, r1_bio->sectors) && in raid1_end_write_request() 560 sector_t sectors) in align_to_barrier_unit_end() argument 564 WARN_ON(sectors == 0); in align_to_barrier_unit_end() 572 if (len > sectors) in align_to_barrier_unit_end() 573 len = sectors; in align_to_barrier_unit_end() 593 int len = r1_bio->sectors; in choose_first_rdev() 622 rdev->recovery_offset < r1_bio->sector + r1_bio->sectors; in rdev_in_recovery() [all …]
|
| H A D | raid10.c | 348 r10_bio->devs[slot].addr + (r10_bio->sectors); in update_head_pos() 532 r10_bio->sectors) && in raid10_end_write_request() 729 int sectors = r10_bio->sectors; in read_balance() local 749 if (raid1_should_read_first(conf->mddev, this_sector, sectors)) in read_balance() 764 r10_bio->devs[slot].addr + sectors > in read_balance() 771 r10_bio->devs[slot].addr + sectors > rdev->recovery_offset) in read_balance() 775 if (is_badblock(rdev, dev_sector, sectors, in read_balance() 786 if (!do_balance && sectors > bad_sectors) in read_balance() 787 sectors = bad_sectors; in read_balance() 788 if (best_good_sectors > sectors) in read_balance() [all …]
|
| H A D | dm-stats.c | 26 unsigned long long sectors[2]; member 602 p->sectors[idx] += len; in dm_stat_for_entry() 731 shared->tmp.sectors[READ] = 0; in __dm_stat_init_temporary_percpu_totals() 732 shared->tmp.sectors[WRITE] = 0; in __dm_stat_init_temporary_percpu_totals() 749 shared->tmp.sectors[READ] += READ_ONCE(p->sectors[READ]); in __dm_stat_init_temporary_percpu_totals() 750 shared->tmp.sectors[WRITE] += READ_ONCE(p->sectors[WRITE]); in __dm_stat_init_temporary_percpu_totals() 783 p->sectors[READ] -= shared->tmp.sectors[READ]; in __dm_stat_clear() 784 p->sectors[WRITE] -= shared->tmp.sectors[WRITE]; in __dm_stat_clear() 901 shared->tmp.sectors[READ], in dm_stats_print() 905 shared->tmp.sectors[WRITE], in dm_stats_print()
|
| /linux/Documentation/block/ |
| H A D | stat.rst | 31 read sectors sectors number of sectors read 35 write sectors sectors number of sectors written 42 discard sectors sectors number of sectors discarded 67 read sectors, write sectors, discard_sectors 70 These values count the number of sectors read from, written to, or 71 discarded from this block device. The "sectors" in question are the 72 standard UNIX 512-byte sectors, not any device- or filesystem-specific
|
| /linux/drivers/mtd/ |
| H A D | ssfdc.c | 22 unsigned char sectors; member 317 ssfdc->sectors = 32; in ssfdcr_add_mtd() 318 get_chs(mtd->size, NULL, &ssfdc->heads, &ssfdc->sectors); in ssfdcr_add_mtd() 320 ((long)ssfdc->sectors * (long)ssfdc->heads)); in ssfdcr_add_mtd() 323 ssfdc->cylinders, ssfdc->heads , ssfdc->sectors, in ssfdcr_add_mtd() 325 (long)ssfdc->sectors); in ssfdcr_add_mtd() 328 (long)ssfdc->sectors; in ssfdcr_add_mtd() 411 ssfdc->cylinders, ssfdc->heads, ssfdc->sectors); in ssfdcr_getgeo() 414 geo->sectors = ssfdc->sectors; in ssfdcr_getgeo()
|
| H A D | nftlcore.c | 71 nftl->sectors = nftl->mbd.size / temp; in nftl_add_mtd() 73 nftl->sectors++; in nftl_add_mtd() 74 temp = nftl->cylinders * nftl->sectors; in nftl_add_mtd() 79 temp = nftl->heads * nftl->sectors; in nftl_add_mtd() 84 if (nftl->mbd.size != nftl->heads * nftl->cylinders * nftl->sectors) { in nftl_add_mtd() 93 nftl->cylinders, nftl->heads , nftl->sectors, in nftl_add_mtd() 95 (long)nftl->sectors ); in nftl_add_mtd() 770 geo->sectors = nftl->sectors; in nftl_getgeo()
|
| /linux/drivers/usb/storage/ |
| H A D | datafab.c | 60 unsigned long sectors; /* total sector count */ member 138 u32 sectors) in datafab_read_data() argument 153 if (sectors > 0x0FFFFFFF) in datafab_read_data() 162 totallen = sectors * info->ssize; in datafab_read_data() 221 u32 sectors) in datafab_write_data() argument 237 if (sectors > 0x0FFFFFFF) in datafab_write_data() 246 totallen = sectors * info->ssize; in datafab_write_data() 420 info->sectors = ((u32)(reply[117]) << 24) | in datafab_id_device() 582 info->sectors, info->ssize); in datafab_transport() 586 ((__be32 *) ptr)[0] = cpu_to_be32(info->sectors - 1); in datafab_transport()
|
| H A D | jumpshot.c | 96 unsigned long sectors; /* total sector count */ member 156 u32 sectors) in jumpshot_read_data() argument 174 totallen = sectors * info->ssize; in jumpshot_read_data() 233 u32 sectors) in jumpshot_write_data() argument 251 totallen = sectors * info->ssize; in jumpshot_write_data() 354 info->sectors = ((u32)(reply[117]) << 24) | in jumpshot_id_device() 513 info->sectors, info->ssize); in jumpshot_transport() 517 ((__be32 *) ptr)[0] = cpu_to_be32(info->sectors - 1); in jumpshot_transport()
|
| H A D | sddr55.c | 190 unsigned short sectors) { in sddr55_read_data() argument 209 len = min_t(unsigned int, sectors, info->blocksize >> in sddr55_read_data() 217 while (sectors>0) { in sddr55_read_data() 227 pages = min_t(unsigned int, sectors << info->smallpageshift, in sddr55_read_data() 295 sectors -= pages >> info->smallpageshift; in sddr55_read_data() 309 unsigned short sectors) { in sddr55_write_data() argument 336 len = min_t(unsigned int, sectors, info->blocksize >> in sddr55_write_data() 344 while (sectors > 0) { in sddr55_write_data() 354 pages = min_t(unsigned int, sectors << info->smallpageshift, in sddr55_write_data() 509 sectors -= pages >> info->smallpageshift; in sddr55_write_data()
|
| /linux/drivers/md/bcache/ |
| H A D | alloc.c | 83 void bch_rescale_priorities(struct cache_set *c, int sectors) in bch_rescale_priorities() argument 90 atomic_sub(sectors, &c->rescale); in bch_rescale_priorities() 592 unsigned int sectors, in bch_alloc_sectors() argument 640 sectors = min(sectors, b->sectors_free); in bch_alloc_sectors() 642 SET_KEY_OFFSET(k, KEY_OFFSET(k) + sectors); in bch_alloc_sectors() 643 SET_KEY_SIZE(k, sectors); in bch_alloc_sectors() 654 b->sectors_free -= sectors; in bch_alloc_sectors() 657 SET_PTR_OFFSET(&b->key, i, PTR_OFFSET(&b->key, i) + sectors); in bch_alloc_sectors() 659 atomic_long_add(sectors, in bch_alloc_sectors()
|
| H A D | request.c | 117 unsigned int sectors = min(bio_sectors(bio), in bch_data_invalidate() local 123 bio->bi_iter.bi_sector += sectors; in bch_data_invalidate() 124 bio->bi_iter.bi_size -= sectors << 9; in bch_data_invalidate() 129 sectors)); in bch_data_invalidate() 367 unsigned int sectors, congested; in check_should_bypass() local 453 sectors = max(task->sequential_io, in check_should_bypass() 457 sectors >= dc->sequential_cutoff >> 9) { in check_should_bypass() 462 if (congested && sectors >= congested) { in check_should_bypass() 540 unsigned int sectors = KEY_INODE(k) == s->iop.inode in cache_lookup_fn() local 544 int ret = s->d->cache_miss(b, s, bio, sectors); in cache_lookup_fn() [all …]
|
| /linux/Documentation/ABI/testing/ |
| H A D | procfs-diskstats | 15 6 sectors read 19 10 sectors written 32 17 sectors discarded
|
| /linux/fs/hfsplus/ |
| H A D | btree.c | 76 u64 sectors, int file_id) in hfsplus_calc_btree_clump_size() argument 100 if (sectors < 0x200000) { in hfsplus_calc_btree_clump_size() 101 clump_size = sectors << 2; /* 0.8 % */ in hfsplus_calc_btree_clump_size() 106 for (i = 0, sectors = sectors >> 22; in hfsplus_calc_btree_clump_size() 107 sectors && (i < CLUMP_ENTRIES - 1); in hfsplus_calc_btree_clump_size() 108 ++i, sectors = sectors >> 1) { in hfsplus_calc_btree_clump_size()
|
| /linux/include/linux/ |
| H A D | badblocks.h | 51 int badblocks_check(struct badblocks *bb, sector_t s, sector_t sectors, 53 bool badblocks_set(struct badblocks *bb, sector_t s, sector_t sectors, 55 bool badblocks_clear(struct badblocks *bb, sector_t s, sector_t sectors);
|
| /linux/block/partitions/ |
| H A D | ibm.c | 41 return cyl * geo->heads * geo->sectors + in cchh2blk() 42 head * geo->sectors; in cchh2blk() 59 return cyl * geo->heads * geo->sectors + in cchhb2blk() 60 head * geo->sectors + in cchhb2blk() 212 offset + geo->sectors; in find_vol1_partitions() 257 * geo->sectors * secperblk; in find_lnx1_partitions()
|
| /linux/drivers/mtd/nand/ |
| H A D | ecc-mtk.c | 66 u32 sectors; member 150 if (dec & ecc->sectors) { in mtk_ecc_irq() 156 ecc->sectors = 0; in mtk_ecc_irq() 213 if (config->sectors) in mtk_ecc_config() 214 ecc->sectors = 1 << (config->sectors - 1); in mtk_ecc_config() 221 int sectors) in mtk_ecc_get_stats() argument 229 for (i = 0; i < sectors; i++) { in mtk_ecc_get_stats()
|
| /linux/drivers/mtd/spi-nor/ |
| H A D | Kconfig | 14 bool "Use small 4096 B erase sectors" 17 Many flash memories support erasing small (4096 B) sectors. Depending 21 small sectors. On the other hand erasing should be faster when using 22 64 KiB block instead of 16 × 4 KiB sectors.
|
| /linux/arch/x86/boot/ |
| H A D | mtools.conf.in | 11 file="@OBJ@/fdimage" cylinders=80 heads=2 sectors=18 filter 15 file="@OBJ@/fdimage" cylinders=80 heads=2 sectors=36 filter
|