Home
last modified time | relevance | path

Searched refs:chunk_sectors (Results 1 – 22 of 22) sorted by relevance

/linux/drivers/md/
H A Dmd-linear.c74 lim.max_hw_sectors = mddev->chunk_sectors; in linear_set_limits()
76 lim.max_write_zeroes_sectors = mddev->chunk_sectors; in linear_set_limits()
77 lim.max_hw_wzeroes_unmap_sectors = mddev->chunk_sectors; in linear_set_limits()
78 lim.io_min = mddev->chunk_sectors << 9; in linear_set_limits()
127 if (mddev->chunk_sectors) { in linear_conf()
129 sector_div(sectors, mddev->chunk_sectors); in linear_conf()
130 rdev->sectors = sectors * mddev->chunk_sectors; in linear_conf()
299 seq_printf(seq, " %dk rounding", mddev->chunk_sectors / 2); in linear_status()
H A Draid5.c927 if (!sector_div(tmp_sec, conf->chunk_sectors)) in stripe_add_to_batch_list()
2959 : conf->chunk_sectors; in raid5_compute_sector()
3155 : conf->chunk_sectors; in raid5_compute_blocknr()
3486 if (first + conf->chunk_sectors * (count - 1) != last) in stripe_bio_overlaps()
3573 previous ? conf->prev_chunk_sectors : conf->chunk_sectors; in stripe_set_idx()
5340 unsigned int chunk_sectors; in in_chunk_boundary() local
5343 chunk_sectors = min(conf->chunk_sectors, conf->prev_chunk_sectors); in in_chunk_boundary()
5344 return chunk_sectors >= in in_chunk_boundary()
5345 ((sector & (chunk_sectors - 1)) + bio_sectors); in in_chunk_boundary()
5495 unsigned chunk_sects = mddev->chunk_sectors; in chunk_aligned_read()
[all …]
H A Ddm-zoned-target.c994 unsigned int chunk_sectors = dmz_zone_nr_sectors(dmz->metadata); in dmz_io_hints() local
1004 limits->max_hw_discard_sectors = chunk_sectors; in dmz_io_hints()
1005 limits->max_write_zeroes_sectors = chunk_sectors; in dmz_io_hints()
1008 limits->chunk_sectors = chunk_sectors; in dmz_io_hints()
1009 limits->max_sectors = chunk_sectors; in dmz_io_hints()
H A Draid5-ppl.c329 (data_sector >> ilog2(conf->chunk_sectors) == in ppl_log_stripe()
330 data_sector_last >> ilog2(conf->chunk_sectors)) && in ppl_log_stripe()
812 if ((pp_size >> 9) < conf->chunk_sectors) { in ppl_recover_entry()
821 (data_disks - 1) * conf->chunk_sectors + in ppl_recover_entry()
825 strip_sectors = conf->chunk_sectors; in ppl_recover_entry()
859 (disk * conf->chunk_sectors); in ppl_recover_entry()
H A Ddm-raid.c718 mddev->new_chunk_sectors = mddev->chunk_sectors; in rs_set_cur()
731 mddev->chunk_sectors = mddev->new_chunk_sectors; in rs_set_new()
985 if (region_size < rs->md.chunk_sectors) { in validate_region_size()
1166 rs->md.new_chunk_sectors = rs->md.chunk_sectors = value; in parse_raid_params()
1492 if (rs->md.chunk_sectors) in parse_raid_params()
1493 max_io_len = rs->md.chunk_sectors; in parse_raid_params()
1542 uint32_t min_stripes = max(mddev->chunk_sectors, mddev->new_chunk_sectors) / 2; in rs_set_raid456_stripe_cache()
1908 rs->md.new_chunk_sectors != rs->md.chunk_sectors; in rs_is_layout_change()
2150 sb->stripe_sectors = cpu_to_le32(mddev->chunk_sectors); in super_sync()
2261 mddev->chunk_sectors = le32_to_cpu(sb->stripe_sectors); in super_init_validation()
[all …]
H A Ddm-unstripe.c180 limits->chunk_sectors = uc->chunk_size; in unstripe_io_hints()
H A Draid10.c1797 dev_start = (first_stripe_index + 1) * mddev->chunk_sectors; in raid10_handle_discard()
1799 dev_start = first_stripe_index * mddev->chunk_sectors; in raid10_handle_discard()
1804 dev_end = (last_stripe_index + 1) * mddev->chunk_sectors; in raid10_handle_discard()
1806 dev_end = last_stripe_index * mddev->chunk_sectors; in raid10_handle_discard()
1909 seq_printf(seq, " %dK chunks", mddev->chunk_sectors / 2); in raid10_status()
3105 window_size = (chunks + extra_chunk) * conf->mddev->chunk_sectors; in raid10_set_cluster_sync_high()
3770 chunk = mddev->chunk_sectors; in setup_geo()
3940 lim.io_min = mddev->chunk_sectors << 9; in raid10_set_queue_limits()
3941 lim.chunk_sectors = mddev->chunk_sectors; in raid10_set_queue_limits()
4203 mddev->new_chunk_sectors = mddev->chunk_sectors; in raid10_takeover_raid0()
[all …]
H A Dmd.c1493 mddev->chunk_sectors = sb->chunk_size >> 9; in super_90_validate()
1522 mddev->new_chunk_sectors = mddev->chunk_sectors; in super_90_validate()
1669 sb->chunk_size = mddev->chunk_sectors << 9; in super_90_sync()
2000 mddev->chunk_sectors = le32_to_cpu(sb->chunksize); in super_1_validate()
2061 mddev->new_chunk_sectors = mddev->chunk_sectors; in super_1_validate()
2217 sb->chunksize = cpu_to_le32(mddev->chunk_sectors); in super_1_sync()
2779 (mddev->chunk_sectors != le32_to_cpu(sb->chunksize))) in does_sb_need_changing()
4127 mddev->new_chunk_sectors = mddev->chunk_sectors; in level_store()
4149 mddev->chunk_sectors = mddev->new_chunk_sectors; in level_store()
4457 mddev->chunk_sectors != mddev->new_chunk_sectors) in chunk_size_show()
[all …]
H A Ddm-stripe.c461 limits->chunk_sectors = sc->chunk_size; in stripe_io_hints()
H A Dmd.h438 int chunk_sectors; member
H A Ddm-table.c1803 zone_sectors = ti_limits.chunk_sectors; in dm_calculate_queue_limits()
1856 zone_sectors = limits->chunk_sectors; in dm_calculate_queue_limits()
H A Draid1.c3369 if (mddev->chunk_sectors != mddev->new_chunk_sectors || in raid1_reshape()
3372 mddev->new_chunk_sectors = mddev->chunk_sectors; in raid1_reshape()
/linux/drivers/char/
H A Dps3flash.c26 u64 chunk_sectors; member
38 start_sector, priv->chunk_sectors, in ps3flash_read_write_sectors()
118 sector = *pos / dev->bounce_size * priv->chunk_sectors; in ps3flash_read()
151 sector += priv->chunk_sectors; in ps3flash_read()
187 sector = *pos / dev->bounce_size * priv->chunk_sectors; in ps3flash_write()
226 sector += priv->chunk_sectors; in ps3flash_write()
376 priv->chunk_sectors = dev->bounce_size / dev->blk_size; in ps3flash_probe()
/linux/drivers/nvme/host/
H A Dzns.c115 lim->chunk_sectors = ns->head->zsze = in nvme_update_zone_info()
H A Dcore.c2203 lim->chunk_sectors = iob; in nvme_set_chunk_sectors()
/linux/block/
H A Dblk.h396 if (lim->chunk_sectors) in bio_may_need_split()
H A Dblk-core.c615 if (nr_sectors > q->limits.chunk_sectors) in blk_check_zone_append()
H A Dblk-sysfs.c226 QUEUE_SYSFS_LIMIT_SHOW(chunk_sectors) in QUEUE_SYSFS_LIMIT_SHOW()
/linux/drivers/block/null_blk/
H A Dzoned.c168 lim->chunk_sectors = dev->zone_size_sects; in null_init_zoned_dev()
/linux/drivers/block/
H A Dzloop.c1177 .chunk_sectors = opts->zone_size, in zloop_ctl_add()
1309 set_capacity(zlo->disk, (u64)lim.chunk_sectors * zlo->nr_zones); in zloop_ctl_add()
H A Dublk_drv.c508 return p->dev_sectors >> ilog2(p->chunk_sectors); in ublk_get_nr_zones()
575 unsigned int zone_size_sectors = disk->queue->limits.chunk_sectors; in ublk_report_zones()
923 if (ublk_dev_is_zoned(ub) && !p->chunk_sectors) in ublk_validate_params()
4421 .chunk_sectors = p->chunk_sectors, in ublk_ctrl_start_dev()
H A Dvirtio_blk.c761 lim->chunk_sectors = vblk->zone_sectors; in virtblk_read_zoned_limits()