Home
last modified time | relevance | path

Searched refs:sectors (Results 1 – 25 of 173) sorted by relevance

1234567

/linux/block/
H A Dbadblocks.c554 sector_t sectors = bad->len; in front_merge() local
562 merged = min_t(sector_t, sectors, BB_END(p[prev]) - s); in front_merge()
564 merged = min_t(sector_t, sectors, BB_MAX_LEN - BB_LEN(p[prev])); in front_merge()
839 static bool _badblocks_set(struct badblocks *bb, sector_t s, sector_t sectors, in _badblocks_set() argument
852 if (sectors == 0) in _badblocks_set()
858 sector_t next = s + sectors; in _badblocks_set()
862 sectors = next - s; in _badblocks_set()
872 bad.len = sectors; in _badblocks_set()
923 BB_END(p[prev]) - s, sectors); in _badblocks_set()
954 sectors -= len; in _badblocks_set()
[all …]
/linux/Documentation/admin-guide/device-mapper/
H A Ddm-ebs.rst8 size. Its main purpose is to provide emulation of 512 byte sectors on
18 <dev path> <offset> <emulated sectors> [<underlying sectors>]
27 has to be a multiple of <emulated sectors>.
28 <emulated sectors>:
29 Number of sectors defining the logical block size to be emulated;
30 1, 2, 4, 8 sectors of 512 bytes supported.
34 <underlying sectors>:
35 Number of sectors defining the logical block size of <dev path>.
36 2^N supported, e.g. 8 = emulate 8 sectors of 512 bytes = 4KiB.
43 offset 1024 sectors with underlying devices block size automatically set:
[all …]
H A Ddm-integrity.rst68 dm-integrity won't read of write these sectors
77 not used and data sectors and integrity tags are written
96 data directly in the underlying device sectors.
111 The number of interleaved sectors. This values is rounded down to
120 The number of sectors in one metadata buffer. The value is rounded
183 512-byte sectors that corresponds to one bitmap bit.
202 copy sectors from one journal section to another journal section
206 key and also to disallow the attacker to move sectors from one
224 256 sectors of metadata per data area. With the default buffer_sectors of
231 2. provided data sectors - that is the number of sectors that the user
[all …]
H A Ddm-dust.rst4 This target emulates the behavior of bad sectors at arbitrary
11 drive with bad sectors).
23 sectors.
25 Normally, a drive that is encountering bad sectors will most likely
26 encounter more bad sectors, at an unknown time or location.
32 simulating a "failure" event where bad sectors start to appear.
53 First, find the size (in 512-byte sectors) of the device to be used::
114 kernel: device-mapper: dust: enabling read failures on bad sectors
299 be bad sectors that occur on the device in the future, in a different,
/linux/drivers/target/
H A Dtarget_core_sbc.c216 static inline u32 sbc_get_size(struct se_cmd *cmd, u32 sectors) in sbc_get_size() argument
218 return cmd->se_dev->dev_attrib.block_size * sectors; in sbc_get_size()
279 unsigned int sectors = sbc_get_write_same_sectors(cmd); in sbc_setup_write_same() local
288 if (sectors > cmd->se_dev->dev_attrib.max_write_same_len) { in sbc_setup_write_same()
290 sectors, cmd->se_dev->dev_attrib.max_write_same_len); in sbc_setup_write_same()
296 if (((cmd->t_task_lba + sectors) < cmd->t_task_lba) || in sbc_setup_write_same()
297 ((cmd->t_task_lba + sectors) > end_lba)) { in sbc_setup_write_same()
299 (unsigned long long)end_lba, cmd->t_task_lba, sectors); in sbc_setup_write_same()
333 ret = sbc_check_prot(dev, cmd, flags >> 5, sectors, true); in sbc_setup_write_same()
668 u32 sectors, bool is_write) in sbc_check_prot() argument
[all …]
/linux/drivers/md/
H A Draid0.c68 sector_t curr_zone_end, sectors; in create_strip_zones() local
88 sectors = rdev1->sectors; in create_strip_zones()
89 sector_div(sectors, mddev->chunk_sectors); in create_strip_zones()
90 rdev1->sectors = sectors * mddev->chunk_sectors; in create_strip_zones()
101 (unsigned long long)rdev1->sectors, in create_strip_zones()
103 (unsigned long long)rdev2->sectors); in create_strip_zones()
109 if (rdev2->sectors == rdev1->sectors) { in create_strip_zones()
198 if (!smallest || (rdev1->sectors < smallest->sectors)) in create_strip_zones()
208 zone->zone_end = smallest->sectors * cnt; in create_strip_zones()
221 zone->dev_start = smallest->sectors; in create_strip_zones()
[all …]
H A Dmd-linear.c55 static sector_t linear_size(struct mddev *mddev, sector_t sectors, int raid_disks) in linear_size() argument
61 WARN_ONCE(sectors || raid_disks, in linear_size()
118 sector_t sectors; in linear_conf() local
128 sectors = rdev->sectors; in linear_conf()
129 sector_div(sectors, mddev->chunk_sectors); in linear_conf()
130 rdev->sectors = sectors * mddev->chunk_sectors; in linear_conf()
133 conf->array_sectors += rdev->sectors; in linear_conf()
145 conf->disks[0].end_sector = conf->disks[0].rdev->sectors; in linear_conf()
150 conf->disks[i].rdev->sectors; in linear_conf()
246 start_sector = tmp_dev->end_sector - tmp_dev->rdev->sectors; in linear_make_request()
[all …]
H A Draid1.c65 sector_t hi = lo + r1_bio->sectors; in check_and_add_serial()
340 r1_bio->sector + (r1_bio->sectors); in update_head_pos()
455 sector_t hi = r1_bio->sector + r1_bio->sectors; in raid1_end_write_request()
512 if (rdev_has_badblock(rdev, r1_bio->sector, r1_bio->sectors) && in raid1_end_write_request()
560 sector_t sectors) in align_to_barrier_unit_end() argument
564 WARN_ON(sectors == 0); in align_to_barrier_unit_end()
572 if (len > sectors) in align_to_barrier_unit_end()
573 len = sectors; in align_to_barrier_unit_end()
593 int len = r1_bio->sectors; in choose_first_rdev()
622 rdev->recovery_offset < r1_bio->sector + r1_bio->sectors; in rdev_in_recovery()
[all …]
H A Draid10.c348 r10_bio->devs[slot].addr + (r10_bio->sectors); in update_head_pos()
532 r10_bio->sectors) && in raid10_end_write_request()
729 int sectors = r10_bio->sectors; in read_balance() local
749 if (raid1_should_read_first(conf->mddev, this_sector, sectors)) in read_balance()
764 r10_bio->devs[slot].addr + sectors > in read_balance()
771 r10_bio->devs[slot].addr + sectors > rdev->recovery_offset) in read_balance()
775 if (is_badblock(rdev, dev_sector, sectors, in read_balance()
786 if (!do_balance && sectors > bad_sectors) in read_balance()
787 sectors = bad_sectors; in read_balance()
788 if (best_good_sectors > sectors) in read_balance()
[all …]
H A Ddm-stats.c26 unsigned long long sectors[2]; member
602 p->sectors[idx] += len; in dm_stat_for_entry()
731 shared->tmp.sectors[READ] = 0; in __dm_stat_init_temporary_percpu_totals()
732 shared->tmp.sectors[WRITE] = 0; in __dm_stat_init_temporary_percpu_totals()
749 shared->tmp.sectors[READ] += READ_ONCE(p->sectors[READ]); in __dm_stat_init_temporary_percpu_totals()
750 shared->tmp.sectors[WRITE] += READ_ONCE(p->sectors[WRITE]); in __dm_stat_init_temporary_percpu_totals()
783 p->sectors[READ] -= shared->tmp.sectors[READ]; in __dm_stat_clear()
784 p->sectors[WRITE] -= shared->tmp.sectors[WRITE]; in __dm_stat_clear()
901 shared->tmp.sectors[READ], in dm_stats_print()
905 shared->tmp.sectors[WRITE], in dm_stats_print()
/linux/Documentation/block/
H A Dstat.rst31 read sectors sectors number of sectors read
35 write sectors sectors number of sectors written
42 discard sectors sectors number of sectors discarded
67 read sectors, write sectors, discard_sectors
70 These values count the number of sectors read from, written to, or
71 discarded from this block device. The "sectors" in question are the
72 standard UNIX 512-byte sectors, not any device- or filesystem-specific
/linux/drivers/mtd/
H A Dssfdc.c22 unsigned char sectors; member
317 ssfdc->sectors = 32; in ssfdcr_add_mtd()
318 get_chs(mtd->size, NULL, &ssfdc->heads, &ssfdc->sectors); in ssfdcr_add_mtd()
320 ((long)ssfdc->sectors * (long)ssfdc->heads)); in ssfdcr_add_mtd()
323 ssfdc->cylinders, ssfdc->heads , ssfdc->sectors, in ssfdcr_add_mtd()
325 (long)ssfdc->sectors); in ssfdcr_add_mtd()
328 (long)ssfdc->sectors; in ssfdcr_add_mtd()
411 ssfdc->cylinders, ssfdc->heads, ssfdc->sectors); in ssfdcr_getgeo()
414 geo->sectors = ssfdc->sectors; in ssfdcr_getgeo()
H A Dnftlcore.c71 nftl->sectors = nftl->mbd.size / temp; in nftl_add_mtd()
73 nftl->sectors++; in nftl_add_mtd()
74 temp = nftl->cylinders * nftl->sectors; in nftl_add_mtd()
79 temp = nftl->heads * nftl->sectors; in nftl_add_mtd()
84 if (nftl->mbd.size != nftl->heads * nftl->cylinders * nftl->sectors) { in nftl_add_mtd()
93 nftl->cylinders, nftl->heads , nftl->sectors, in nftl_add_mtd()
95 (long)nftl->sectors ); in nftl_add_mtd()
770 geo->sectors = nftl->sectors; in nftl_getgeo()
/linux/drivers/usb/storage/
H A Ddatafab.c60 unsigned long sectors; /* total sector count */ member
138 u32 sectors) in datafab_read_data() argument
153 if (sectors > 0x0FFFFFFF) in datafab_read_data()
162 totallen = sectors * info->ssize; in datafab_read_data()
221 u32 sectors) in datafab_write_data() argument
237 if (sectors > 0x0FFFFFFF) in datafab_write_data()
246 totallen = sectors * info->ssize; in datafab_write_data()
420 info->sectors = ((u32)(reply[117]) << 24) | in datafab_id_device()
582 info->sectors, info->ssize); in datafab_transport()
586 ((__be32 *) ptr)[0] = cpu_to_be32(info->sectors - 1); in datafab_transport()
H A Djumpshot.c96 unsigned long sectors; /* total sector count */ member
156 u32 sectors) in jumpshot_read_data() argument
174 totallen = sectors * info->ssize; in jumpshot_read_data()
233 u32 sectors) in jumpshot_write_data() argument
251 totallen = sectors * info->ssize; in jumpshot_write_data()
354 info->sectors = ((u32)(reply[117]) << 24) | in jumpshot_id_device()
513 info->sectors, info->ssize); in jumpshot_transport()
517 ((__be32 *) ptr)[0] = cpu_to_be32(info->sectors - 1); in jumpshot_transport()
H A Dsddr55.c190 unsigned short sectors) { in sddr55_read_data() argument
209 len = min_t(unsigned int, sectors, info->blocksize >> in sddr55_read_data()
217 while (sectors>0) { in sddr55_read_data()
227 pages = min_t(unsigned int, sectors << info->smallpageshift, in sddr55_read_data()
295 sectors -= pages >> info->smallpageshift; in sddr55_read_data()
309 unsigned short sectors) { in sddr55_write_data() argument
336 len = min_t(unsigned int, sectors, info->blocksize >> in sddr55_write_data()
344 while (sectors > 0) { in sddr55_write_data()
354 pages = min_t(unsigned int, sectors << info->smallpageshift, in sddr55_write_data()
509 sectors -= pages >> info->smallpageshift; in sddr55_write_data()
/linux/drivers/md/bcache/
H A Dalloc.c83 void bch_rescale_priorities(struct cache_set *c, int sectors) in bch_rescale_priorities() argument
90 atomic_sub(sectors, &c->rescale); in bch_rescale_priorities()
592 unsigned int sectors, in bch_alloc_sectors() argument
640 sectors = min(sectors, b->sectors_free); in bch_alloc_sectors()
642 SET_KEY_OFFSET(k, KEY_OFFSET(k) + sectors); in bch_alloc_sectors()
643 SET_KEY_SIZE(k, sectors); in bch_alloc_sectors()
654 b->sectors_free -= sectors; in bch_alloc_sectors()
657 SET_PTR_OFFSET(&b->key, i, PTR_OFFSET(&b->key, i) + sectors); in bch_alloc_sectors()
659 atomic_long_add(sectors, in bch_alloc_sectors()
H A Drequest.c117 unsigned int sectors = min(bio_sectors(bio), in bch_data_invalidate() local
123 bio->bi_iter.bi_sector += sectors; in bch_data_invalidate()
124 bio->bi_iter.bi_size -= sectors << 9; in bch_data_invalidate()
129 sectors)); in bch_data_invalidate()
367 unsigned int sectors, congested; in check_should_bypass() local
453 sectors = max(task->sequential_io, in check_should_bypass()
457 sectors >= dc->sequential_cutoff >> 9) { in check_should_bypass()
462 if (congested && sectors >= congested) { in check_should_bypass()
540 unsigned int sectors = KEY_INODE(k) == s->iop.inode in cache_lookup_fn() local
544 int ret = s->d->cache_miss(b, s, bio, sectors); in cache_lookup_fn()
[all …]
/linux/Documentation/ABI/testing/
H A Dprocfs-diskstats15 6 sectors read
19 10 sectors written
32 17 sectors discarded
/linux/fs/hfsplus/
H A Dbtree.c76 u64 sectors, int file_id) in hfsplus_calc_btree_clump_size() argument
100 if (sectors < 0x200000) { in hfsplus_calc_btree_clump_size()
101 clump_size = sectors << 2; /* 0.8 % */ in hfsplus_calc_btree_clump_size()
106 for (i = 0, sectors = sectors >> 22; in hfsplus_calc_btree_clump_size()
107 sectors && (i < CLUMP_ENTRIES - 1); in hfsplus_calc_btree_clump_size()
108 ++i, sectors = sectors >> 1) { in hfsplus_calc_btree_clump_size()
/linux/include/linux/
H A Dbadblocks.h51 int badblocks_check(struct badblocks *bb, sector_t s, sector_t sectors,
53 bool badblocks_set(struct badblocks *bb, sector_t s, sector_t sectors,
55 bool badblocks_clear(struct badblocks *bb, sector_t s, sector_t sectors);
/linux/block/partitions/
H A Dibm.c41 return cyl * geo->heads * geo->sectors + in cchh2blk()
42 head * geo->sectors; in cchh2blk()
59 return cyl * geo->heads * geo->sectors + in cchhb2blk()
60 head * geo->sectors + in cchhb2blk()
212 offset + geo->sectors; in find_vol1_partitions()
257 * geo->sectors * secperblk; in find_lnx1_partitions()
/linux/drivers/mtd/nand/
H A Decc-mtk.c66 u32 sectors; member
150 if (dec & ecc->sectors) { in mtk_ecc_irq()
156 ecc->sectors = 0; in mtk_ecc_irq()
213 if (config->sectors) in mtk_ecc_config()
214 ecc->sectors = 1 << (config->sectors - 1); in mtk_ecc_config()
221 int sectors) in mtk_ecc_get_stats() argument
229 for (i = 0; i < sectors; i++) { in mtk_ecc_get_stats()
/linux/drivers/mtd/spi-nor/
H A DKconfig14 bool "Use small 4096 B erase sectors"
17 Many flash memories support erasing small (4096 B) sectors. Depending
21 small sectors. On the other hand erasing should be faster when using
22 64 KiB block instead of 16 × 4 KiB sectors.
/linux/arch/x86/boot/
H A Dmtools.conf.in11 file="@OBJ@/fdimage" cylinders=80 heads=2 sectors=18 filter
15 file="@OBJ@/fdimage" cylinders=80 heads=2 sectors=36 filter

1234567