Home
last modified time | relevance | path

Searched refs:bi_sector (Results 1 – 25 of 115) sorted by relevance

12345

/linux/drivers/md/dm-vdo/
H A Dio-submitter.c149 vio->bios_merged.head->bi_iter.bi_sector); in get_bio_list()
151 vio->bios_merged.tail->bi_iter.bi_sector); in get_bio_list()
195 sector_t merge_sector = bio->bi_iter.bi_sector; in get_mergeable_locked()
218 return (vio_merge->bios_merged.tail->bi_iter.bi_sector == merge_sector ? in get_mergeable_locked()
222 return (vio_merge->bios_merged.head->bi_iter.bi_sector == merge_sector ? in get_mergeable_locked()
231 bio_sector = vio->bios_merged.head->bi_iter.bi_sector; in map_merged_vio()
236 bio_sector = vio->bios_merged.tail->bi_iter.bi_sector; in map_merged_vio()
243 vdo_int_map_remove(bio_map, prev_vio->bios_merged.tail->bi_iter.bi_sector); in merge_to_prev_tail()
256 vdo_int_map_remove(bio_map, next_vio->bios_merged.head->bi_iter.bi_sector); in merge_to_next_head()
293 bio->bi_iter.bi_sector, in try_bio_map_merge()
/linux/drivers/md/
H A Ddm-ebs-target.c50 sector_t end_sector = __block_mod(bio->bi_iter.bi_sector, ec->u_bs) + bio_sectors(bio); in __nr_blocks()
72 unsigned int buf_off = to_bytes(__block_mod(iter->bi_sector, ec->u_bs)); in __ebs_rw_bvec()
73 sector_t block = __sector_to_block(ec, iter->bi_sector); in __ebs_rw_bvec()
145 sector_t block, blocks, sector = bio->bi_iter.bi_sector; in __ebs_discard_bio()
169 sector_t blocks, sector = bio->bi_iter.bi_sector; in __ebs_forget_bio()
195 block1 = __sector_to_block(ec, bio->bi_iter.bi_sector); in __ebs_process_bios()
200 if (__block_mod(bio->bi_iter.bi_sector, ec->u_bs)) in __ebs_process_bios()
366 bio->bi_iter.bi_sector = ec->start + dm_target_offset(ti, bio->bi_iter.bi_sector); in ebs_map()
375 if (likely(__block_mod(bio->bi_iter.bi_sector, ec->u_bs) || in ebs_map()
H A Ddm-linear.c82 static sector_t linear_map_sector(struct dm_target *ti, sector_t bi_sector) in linear_map_sector() argument
86 return lc->start + dm_target_offset(ti, bi_sector); in linear_map_sector()
94 bio->bi_iter.bi_sector = linear_map_sector(ti, bio->bi_iter.bi_sector); in linear_map()
H A Ddm-flakey.c345 static sector_t flakey_map_sector(struct dm_target *ti, sector_t bi_sector) in flakey_map_sector() argument
349 return fc->start + dm_target_offset(ti, bi_sector); in flakey_map_sector()
357 bio->bi_iter.bi_sector = flakey_map_sector(ti, bio->bi_iter.bi_sector); in flakey_map_bio()
380 (unsigned long long)start.bi_sector, in corrupt_bio_common()
446 clone->bi_iter.bi_sector = flakey_map_sector(ti, bio->bi_iter.bi_sector); in clone_bio()
H A Dmd-linear.c239 sector_t bio_sector = bio->bi_iter.bi_sector; in linear_make_request()
270 bio->bi_iter.bi_sector = bio->bi_iter.bi_sector - in linear_make_request()
289 (unsigned long long)bio->bi_iter.bi_sector, in linear_make_request()
H A Ddm-stripe.c255 stripe_map_range_sector(sc, bio->bi_iter.bi_sector, in stripe_map_range()
261 bio->bi_iter.bi_sector = begin + in stripe_map_range()
292 stripe_map_sector(sc, bio->bi_iter.bi_sector, in stripe_map()
293 &stripe, &bio->bi_iter.bi_sector); in stripe_map()
295 bio->bi_iter.bi_sector += sc->stripe[stripe].physical_start; in stripe_map()
H A Ddm-io-rewind.c62 bip->bip_iter.bi_sector -= bio_integrity_intervals(bi, bytes_done >> 9); in dm_bio_integrity_rewind()
113 iter->bi_sector -= bytes >> 9; in dm_bio_rewind_iter()
H A Ddm-stats.c635 sector_t bi_sector, sector_t end_sector, in __dm_stat_bio() argument
642 if (end_sector <= s->start || bi_sector >= s->end) in __dm_stat_bio()
644 if (unlikely(bi_sector < s->start)) { in __dm_stat_bio()
648 rel_sector = bi_sector - s->start; in __dm_stat_bio()
649 todo = end_sector - bi_sector; in __dm_stat_bio()
673 sector_t bi_sector, unsigned int bi_sectors, bool end, in dm_stats_account_io() argument
686 end_sector = bi_sector + bi_sectors; in dm_stats_account_io()
695 (bi_sector == (READ_ONCE(last->last_sector) && in dm_stats_account_io()
714 __dm_stat_bio(s, bi_rw, bi_sector, end_sector, end, duration_jiffies, stats_aux); in dm_stats_account_io()
H A Ddm-dust.c230 bio->bi_iter.bi_sector = dd->start + dm_target_offset(ti, bio->bi_iter.bi_sector); in dust_map()
233 r = dust_map_read(dd, bio->bi_iter.bi_sector, dd->fail_read_on_bb); in dust_map()
235 r = dust_map_write(dd, bio->bi_iter.bi_sector, dd->fail_read_on_bb); in dust_map()
H A Draid10.c1131 bio->bi_iter.bi_sector < conf->reshape_progress && in regular_request_wait()
1132 bio->bi_iter.bi_sector + sectors > conf->reshape_progress) { in regular_request_wait()
1140 conf->reshape_progress <= bio->bi_iter.bi_sector || in regular_request_wait()
1141 conf->reshape_progress >= bio->bi_iter.bi_sector + in regular_request_wait()
1231 read_bio->bi_iter.bi_sector = r10_bio->devs[slot].addr + in raid10_read_request()
1266 mbio->bi_iter.bi_sector = (r10_bio->devs[n_copy].addr + in raid10_write_one_disk()
1354 bio->bi_iter.bi_sector, in raid10_write_request()
1366 bio->bi_iter.bi_sector, bio_end_sector(bio))) in raid10_write_request()
1381 ? (bio->bi_iter.bi_sector < conf->reshape_safe && in raid10_write_request()
1382 bio->bi_iter.bi_sector + sectors > conf->reshape_progress) in raid10_write_request()
[all …]
H A Ddm-unstripe.c120 sector_t sector = dm_target_offset(ti, bio->bi_iter.bi_sector); in map_to_core()
140 bio->bi_iter.bi_sector = map_to_core(ti, bio) + uc->physical_start; in unstripe_map()
H A Ddm-writecache.c1306 writecache_discard(wc, bio->bi_iter.bi_sector, in writecache_flush_thread()
1343 read_original_sector(wc, e) - bio->bi_iter.bi_sector; in writecache_map_remap_origin()
1356 e = writecache_find_entry(wc, bio->bi_iter.bi_sector, WFE_RETURN_FOLLOWING); in writecache_map_read()
1357 if (e && read_original_sector(wc, e) == bio->bi_iter.bi_sector) { in writecache_map_read()
1367 bio->bi_iter.bi_sector = cache_sector(wc, e); in writecache_map_read()
1394 write_original_sector_seq_count(wc, f, bio->bi_iter.bi_sector + in writecache_bio_copy_ssd()
1421 bio->bi_iter.bi_sector = start_cache_sec; in writecache_bio_copy_ssd()
1447 e = writecache_find_entry(wc, bio->bi_iter.bi_sector, 0); in writecache_map_write()
1470 e = writecache_find_entry(wc, bio->bi_iter.bi_sector, WFE_RETURN_FOLLOWING); in writecache_map_write()
1480 write_original_sector_seq_count(wc, e, bio->bi_iter.bi_sector, wc->seq_count); in writecache_map_write()
[all …]
H A Ddm-target.c212 static sector_t io_err_map_sector(struct dm_target *ti, sector_t bi_sector) in io_err_map_sector() argument
216 return ioec->start + dm_target_offset(ti, bi_sector); in io_err_map_sector()
H A Draid0.c458 sector_t start = bio->bi_iter.bi_sector; in raid0_handle_discard()
474 zone->zone_end - bio->bi_iter.bi_sector, in raid0_handle_discard()
560 sector_t bio_sector = bio->bi_iter.bi_sector; in raid0_map_submit_bio()
586 bio->bi_iter.bi_sector = sector + zone->dev_start + in raid0_map_submit_bio()
608 sector = bio->bi_iter.bi_sector; in raid0_make_request()
H A Ddm-stats.h33 sector_t bi_sector, unsigned int bi_sectors, bool end,
H A Ddm-thin.c675 sector_t block_nr = bio->bi_iter.bi_sector; in get_bio_block()
692 sector_t b = bio->bi_iter.bi_sector; in get_bio_block_range()
717 sector_t bi_sector = bio->bi_iter.bi_sector; in remap() local
721 bio->bi_iter.bi_sector = in remap()
723 (bi_sector & (pool->sectors_per_block - 1)); in remap()
725 bio->bi_iter.bi_sector = (block * pool->sectors_per_block) + in remap()
726 sector_div(bi_sector, pool->sectors_per_block); in remap()
1984 else if (bio->bi_iter.bi_sector < tc->origin_size) { in process_cell()
1986 bio->bi_iter.bi_size = (tc->origin_size - bio->bi_iter.bi_sector) << SECTOR_SHIFT; in process_cell()
2122 sector_t bi_sector = bio->bi_iter.bi_sector; in __thin_bio_rb_add() local
[all …]
/linux/drivers/md/bcache/
H A Drequest.c114 bio_sectors(bio), (uint64_t) bio->bi_iter.bi_sector); in bch_data_invalidate()
123 bio->bi_iter.bi_sector += sectors; in bch_data_invalidate()
128 bio->bi_iter.bi_sector, in bch_data_invalidate()
220 SET_KEY_OFFSET(k, bio->bi_iter.bi_sector); in CLOSURE_CALLBACK()
411 if (bio->bi_iter.bi_sector & (c->cache->sb.block_size - 1) || in check_should_bypass()
430 hlist_for_each_entry(i, iohash(dc, bio->bi_iter.bi_sector), hash) in check_should_bypass()
431 if (i->last == bio->bi_iter.bi_sector && in check_should_bypass()
534 if (bkey_cmp(k, &KEY(s->iop.inode, bio->bi_iter.bi_sector, 0)) <= 0) in cache_lookup_fn()
538 KEY_START(k) > bio->bi_iter.bi_sector) { in cache_lookup_fn()
542 KEY_START(k) - bio->bi_iter.bi_sector) in cache_lookup_fn()
[all …]
H A Ddebug.c54 bio->bi_iter.bi_sector = PTR_OFFSET(&b->key, 0); in bch_btree_verify()
119 check->bi_iter.bi_sector = bio->bi_iter.bi_sector; in bch_data_verify()
140 (uint64_t) bio->bi_iter.bi_sector); in bch_data_verify()
/linux/fs/btrfs/
H A Dbio.c188 const u64 logical = repair_bbio->saved_iter.bi_sector << SECTOR_SHIFT; in btrfs_end_repair_bio()
253 const u64 logical = round_down(failed_bbio->saved_iter.bi_sector << SECTOR_SHIFT, in repair_one_sector()
281 repair_bio->bi_iter.bi_sector = logical >> SECTOR_SHIFT; in repair_one_sector()
447 stripe->physical = bio->bi_iter.bi_sector << SECTOR_SHIFT; in orig_write_end_io_work()
471 stripe->physical = bio->bi_iter.bi_sector << SECTOR_SHIFT; in clone_write_end_io_work()
489 u64 physical = bio->bi_iter.bi_sector << SECTOR_SHIFT; in btrfs_submit_dev_bio()
509 bio->bi_iter.bi_sector = zone_start >> SECTOR_SHIFT; in btrfs_submit_dev_bio()
515 __func__, bio_op(bio), bio->bi_opf, bio->bi_iter.bi_sector, in btrfs_submit_dev_bio()
554 bio->bi_iter.bi_sector = bioc->stripes[dev_nr].physical >> SECTOR_SHIFT; in btrfs_submit_mirrored_bio()
566 bio->bi_iter.bi_sector = smap->physical >> SECTOR_SHIFT; in btrfs_submit_bio()
[all …]
/linux/include/trace/events/
H A Dbcache.h28 __entry->sector = bio->bi_iter.bi_sector;
29 __entry->orig_sector = bio->bi_iter.bi_sector - 16;
103 __entry->sector = bio->bi_iter.bi_sector;
138 __entry->sector = bio->bi_iter.bi_sector;
169 __entry->sector = bio->bi_iter.bi_sector;
238 __entry->sector = bio->bi_iter.bi_sector;
H A Dblock.h323 __entry->sector = bio->bi_iter.bi_sector;
351 __entry->sector = bio->bi_iter.bi_sector;
504 __entry->sector = bio->bi_iter.bi_sector;
543 __entry->sector = bio->bi_iter.bi_sector;
626 __entry->sector = bio->bi_iter.bi_sector;
/linux/fs/iomap/
H A Dbio.c63 bio->bi_iter.bi_sector = sector; in iomap_bio_read_folio_range()
85 bio.bi_iter.bi_sector = iomap_sector(srcmap, pos); in iomap_bio_read_folio_range_sync()
/linux/mm/
H A Dpage_io.c46 (unsigned long long)bio->bi_iter.bi_sector); in __end_swap_bio_write()
65 (unsigned long long)bio->bi_iter.bi_sector); in __end_swap_bio_read()
417 bio.bi_iter.bi_sector = swap_folio_sector(folio); in swap_writepage_bdev_sync()
436 bio->bi_iter.bi_sector = swap_folio_sector(folio); in swap_writepage_bdev_async()
579 bio.bi_iter.bi_sector = swap_folio_sector(folio); in swap_read_folio_bdev_sync()
600 bio->bi_iter.bi_sector = swap_folio_sector(folio); in swap_read_folio_bdev_async()
/linux/block/
H A Dblk-zoned.c330 bio->bi_iter.bi_sector = sector; in blkdev_zone_mgmt()
1094 sector_t sector = bio->bi_iter.bi_sector; in blk_zone_reset_bio_endio()
1151 sector_t sector = bio->bi_iter.bi_sector; in blk_zone_finish_bio_endio()
1243 bio->bi_iter.bi_sector, bio_sectors(bio)); in disk_zone_wplug_add_bio()
1273 zwplug = disk_get_zone_wplug(disk, bio->bi_iter.bi_sector); in blk_zone_write_plug_bio_merged()
1321 if (bio->bi_iter.bi_sector != req_back_sector || in blk_zone_write_plug_init_request()
1383 bio->bi_iter.bi_sector += zwplug->wp_offset; in blk_zone_wplug_prepare_bio()
1410 sector_t sector = bio->bi_iter.bi_sector; in blk_zone_wplug_handle_write()
1509 zwplug = disk_get_zone_wplug(disk, bio->bi_iter.bi_sector); in blk_zone_wplug_handle_native_zone_append()
1539 !bdev_zone_is_seq(bio->bi_bdev, bio->bi_iter.bi_sector)) { in blk_zone_wplug_handle_zone_mgmt()
[all …]
/linux/drivers/nvdimm/
H A Dnd_virtio.c117 if (bio && bio->bi_iter.bi_sector != -1) { in async_pmem_flush()
125 child->bi_iter.bi_sector = -1; in async_pmem_flush()

12345