| /linux/drivers/md/ |
| H A D | dm-bio-record.h | 26 struct bvec_iter bi_iter; member 37 bd->bi_iter = bio->bi_iter; in dm_bio_record() 49 bio->bi_iter = bd->bi_iter; in dm_bio_restore()
|
| H A D | dm-flakey.c | 357 bio->bi_iter.bi_sector = flakey_map_sector(ti, bio->bi_iter.bi_sector); in flakey_map_bio() 432 struct bvec_iter iter = bio->bi_iter; in clone_bio() 434 if (unlikely(bio->bi_iter.bi_size > UIO_MAXIOV << PAGE_SHIFT)) in clone_bio() 437 size = bio->bi_iter.bi_size; in clone_bio() 446 clone->bi_iter.bi_sector = flakey_map_sector(ti, bio->bi_iter.bi_sector); in clone_bio() 511 pb->saved_iter = bio->bi_iter; in flakey_map() 557 clone->bi_iter); in flakey_map() 560 clone->bi_iter); in flakey_map()
|
| H A D | dm-ebs-target.c | 50 sector_t end_sector = __block_mod(bio->bi_iter.bi_sector, ec->u_bs) + bio_sectors(bio); in __nr_blocks() 145 sector_t block, blocks, sector = bio->bi_iter.bi_sector; in __ebs_discard_bio() 169 sector_t blocks, sector = bio->bi_iter.bi_sector; in __ebs_forget_bio() 195 block1 = __sector_to_block(ec, bio->bi_iter.bi_sector); in __ebs_process_bios() 200 if (__block_mod(bio->bi_iter.bi_sector, ec->u_bs)) in __ebs_process_bios() 366 bio->bi_iter.bi_sector = ec->start + dm_target_offset(ti, bio->bi_iter.bi_sector); in ebs_map() 375 if (likely(__block_mod(bio->bi_iter.bi_sector, ec->u_bs) || in ebs_map()
|
| H A D | dm-stripe.c | 255 stripe_map_range_sector(sc, bio->bi_iter.bi_sector, in stripe_map_range() 261 bio->bi_iter.bi_sector = begin + in stripe_map_range() 263 bio->bi_iter.bi_size = to_bytes(end - begin); in stripe_map_range() 292 stripe_map_sector(sc, bio->bi_iter.bi_sector, in stripe_map() 293 &stripe, &bio->bi_iter.bi_sector); in stripe_map() 295 bio->bi_iter.bi_sector += sc->stripe[stripe].physical_start; in stripe_map()
|
| H A D | md-linear.c | 239 sector_t bio_sector = bio->bi_iter.bi_sector; in linear_make_request() 270 bio->bi_iter.bi_sector = bio->bi_iter.bi_sector - in linear_make_request() 289 (unsigned long long)bio->bi_iter.bi_sector, in linear_make_request()
|
| H A D | dm-log-writes.c | 222 bio->bi_iter.bi_size = 0; in write_metadata() 223 bio->bi_iter.bi_sector = sector; in write_metadata() 275 bio->bi_iter.bi_size = 0; in write_inline_data() 276 bio->bi_iter.bi_sector = sector; in write_inline_data() 356 bio->bi_iter.bi_size = 0; in log_one_block() 357 bio->bi_iter.bi_sector = sector; in log_one_block() 374 bio->bi_iter.bi_size = 0; in log_one_block() 375 bio->bi_iter.bi_sector = sector; in log_one_block() 704 block->sector = bio_to_dev_sectors(lc, bio->bi_iter.bi_sector); in log_writes_map()
|
| H A D | dm-io-rewind.c | 140 dm_bio_rewind_iter(bio, &bio->bi_iter, bytes); in dm_bio_rewind() 154 orig->bi_iter.bi_size)); in dm_io_rewind()
|
| H A D | raid10.c | 1131 bio->bi_iter.bi_sector < conf->reshape_progress && in regular_request_wait() 1132 bio->bi_iter.bi_sector + sectors > conf->reshape_progress) { in regular_request_wait() 1140 conf->reshape_progress <= bio->bi_iter.bi_sector || in regular_request_wait() 1141 conf->reshape_progress >= bio->bi_iter.bi_sector + in regular_request_wait() 1231 read_bio->bi_iter.bi_sector = r10_bio->devs[slot].addr + in raid10_read_request() 1266 mbio->bi_iter.bi_sector = (r10_bio->devs[n_copy].addr + in raid10_write_one_disk() 1354 bio->bi_iter.bi_sector, in raid10_write_request() 1366 bio->bi_iter.bi_sector, bio_end_sector(bio))) in raid10_write_request() 1381 ? (bio->bi_iter.bi_sector < conf->reshape_safe && in raid10_write_request() 1382 bio->bi_iter.bi_sector + sectors > conf->reshape_progress) in raid10_write_request() [all …]
|
| H A D | dm-dust.c | 230 bio->bi_iter.bi_sector = dd->start + dm_target_offset(ti, bio->bi_iter.bi_sector); in dust_map() 233 r = dust_map_read(dd, bio->bi_iter.bi_sector, dd->fail_read_on_bb); in dust_map() 235 r = dust_map_write(dd, bio->bi_iter.bi_sector, dd->fail_read_on_bb); in dust_map()
|
| H A D | dm-unstripe.c | 120 sector_t sector = dm_target_offset(ti, bio->bi_iter.bi_sector); in map_to_core() 140 bio->bi_iter.bi_sector = map_to_core(ti, bio) + uc->physical_start; in unstripe_map()
|
| /linux/block/ |
| H A D | bio.c | 229 bio->bi_iter.bi_sector = 0; in bio_init() 230 bio->bi_iter.bi_size = 0; in bio_init() 231 bio->bi_iter.bi_idx = 0; in bio_init() 232 bio->bi_iter.bi_bvec_done = 0; in bio_init() 316 bio->bi_iter.bi_size += bio->bi_io_vec[i].bv_len; in bio_reuse() 665 if (new_size >= bio->bi_iter.bi_size) in bio_truncate() 695 bio->bi_iter.bi_size = new_size; in bio_truncate() 722 if (unlikely(bio->bi_iter.bi_sector >= maxsector)) in guard_bio_eod() 725 maxsector -= bio->bi_iter.bi_sector; in guard_bio_eod() 726 if (likely((bio->bi_iter.bi_size >> 9) <= maxsector)) in guard_bio_eod() [all …]
|
| H A D | blk-core.c | 545 if (should_fail_request(bdev_whole(bio->bi_bdev), bio->bi_iter.bi_size)) in should_fail_bio() 563 bio->bi_iter.bi_sector > maxsector - nr_sectors)) { in bio_check_eod() 569 bio->bi_iter.bi_sector, nr_sectors, maxsector); in bio_check_eod() 582 if (unlikely(should_fail_request(p, bio->bi_iter.bi_size))) in blk_partition_remap() 585 bio->bi_iter.bi_sector += p->bd_start_sect; in blk_partition_remap() 587 bio->bi_iter.bi_sector - in blk_partition_remap() 607 if (!bdev_is_zone_start(bio->bi_bdev, bio->bi_iter.bi_sector)) in blk_check_zone_append() 762 if (bio->bi_iter.bi_size > queue_atomic_write_unit_max_bytes(q)) in blk_validate_atomic_write_op_size() 765 if (bio->bi_iter.bi_size % queue_atomic_write_unit_min_bytes(q)) in blk_validate_atomic_write_op_size() 919 task_io_account_read(bio->bi_iter.bi_size); in submit_bio()
|
| H A D | blk-crypto-fallback.c | 189 bio->bi_iter.bi_sector = bio_src->bi_iter.bi_sector; in blk_crypto_alloc_enc_bio() 278 bio_iter_iovec(src_bio, src_bio->bi_iter); in __blk_crypto_fallback_encrypt_bio() 314 bio_advance_iter_single(src_bio, &src_bio->bi_iter, in __blk_crypto_fallback_encrypt_bio() 316 if (!src_bio->bi_iter.bi_size) in __blk_crypto_fallback_encrypt_bio() 518 f_ctx->crypt_iter = bio->bi_iter; in blk_crypto_fallback_bio_prep()
|
| H A D | fops.c | 80 bio.bi_iter.bi_sector = pos >> SECTOR_SHIFT; in __blkdev_direct_IO_simple() 90 ret = bio.bi_iter.bi_size; in __blkdev_direct_IO_simple() 212 bio->bi_iter.bi_sector = pos >> SECTOR_SHIFT; in __blkdev_direct_IO() 250 task_io_account_write(bio->bi_iter.bi_size); in __blkdev_direct_IO() 252 dio->size += bio->bi_iter.bi_size; in __blkdev_direct_IO() 253 pos += bio->bi_iter.bi_size; in __blkdev_direct_IO() 338 bio->bi_iter.bi_sector = pos >> SECTOR_SHIFT; in __blkdev_direct_IO_async() 357 dio->size = bio->bi_iter.bi_size; in __blkdev_direct_IO_async() 365 task_io_account_write(bio->bi_iter.bi_size); in __blkdev_direct_IO_async()
|
| /linux/drivers/md/bcache/ |
| H A D | request.c | 114 bio_sectors(bio), (uint64_t) bio->bi_iter.bi_sector); in bch_data_invalidate() 123 bio->bi_iter.bi_sector += sectors; in bch_data_invalidate() 124 bio->bi_iter.bi_size -= sectors << 9; in bch_data_invalidate() 128 bio->bi_iter.bi_sector, in bch_data_invalidate() 220 SET_KEY_OFFSET(k, bio->bi_iter.bi_sector); in CLOSURE_CALLBACK() 411 if (bio->bi_iter.bi_sector & (c->cache->sb.block_size - 1) || in check_should_bypass() 430 hlist_for_each_entry(i, iohash(dc, bio->bi_iter.bi_sector), hash) in check_should_bypass() 431 if (i->last == bio->bi_iter.bi_sector && in check_should_bypass() 440 if (i->sequential + bio->bi_iter.bi_size > i->sequential) in check_should_bypass() 441 i->sequential += bio->bi_iter.bi_size; in check_should_bypass() [all …]
|
| H A D | debug.c | 54 bio->bi_iter.bi_sector = PTR_OFFSET(&b->key, 0); in bch_btree_verify() 55 bio->bi_iter.bi_size = KEY_SIZE(&v->key) << 9; in bch_btree_verify() 119 check->bi_iter.bi_sector = bio->bi_iter.bi_sector; in bch_data_verify() 120 check->bi_iter.bi_size = bio->bi_iter.bi_size; in bch_data_verify() 140 (uint64_t) bio->bi_iter.bi_sector); in bch_data_verify()
|
| /linux/include/trace/events/ |
| H A D | bcache.h | 28 __entry->sector = bio->bi_iter.bi_sector; 29 __entry->orig_sector = bio->bi_iter.bi_sector - 16; 30 __entry->nr_sector = bio->bi_iter.bi_size >> 9; 103 __entry->sector = bio->bi_iter.bi_sector; 104 __entry->nr_sector = bio->bi_iter.bi_size >> 9; 138 __entry->sector = bio->bi_iter.bi_sector; 139 __entry->nr_sector = bio->bi_iter.bi_size >> 9; 169 __entry->sector = bio->bi_iter.bi_sector; 170 __entry->nr_sector = bio->bi_iter.bi_size >> 9; 238 __entry->sector = bio->bi_iter.bi_sector; [all …]
|
| H A D | block.h | 323 __entry->sector = bio->bi_iter.bi_sector; 351 __entry->sector = bio->bi_iter.bi_sector; 504 __entry->sector = bio->bi_iter.bi_sector; 543 __entry->sector = bio->bi_iter.bi_sector; 626 __entry->sector = bio->bi_iter.bi_sector;
|
| /linux/drivers/block/ |
| H A D | brd.c | 140 struct bio_vec bv = bio_iter_iovec(bio, bio->bi_iter); in brd_rw_bvec() 141 sector_t sector = bio->bi_iter.bi_sector; in brd_rw_bvec() 167 bio_advance_iter_single(bio, &bio->bi_iter, bv.bv_len); in brd_rw_bvec() 207 brd_do_discard(brd, bio->bi_iter.bi_sector, in brd_submit_bio() 208 bio->bi_iter.bi_size); in brd_submit_bio() 216 } while (bio->bi_iter.bi_size); in brd_submit_bio()
|
| /linux/fs/erofs/ |
| H A D | fileio.c | 28 if (ret >= 0 && ret != rq->bio.bi_iter.bi_size) in erofs_fileio_ki_complete() 51 rq->iocb.ki_pos = rq->bio.bi_iter.bi_sector << SECTOR_SHIFT; in erofs_fileio_rq_submit() 58 rq->bio.bi_iter.bi_size); in erofs_fileio_rq_submit() 141 io->rq->bio.bi_iter.bi_sector = in erofs_fileio_scan_folio()
|
| /linux/fs/iomap/ |
| H A D | bio.c | 129 bio->bi_iter.bi_sector = iomap_sector(iomap, iter->pos); in iomap_read_alloc_bio() 145 bio->bi_iter.bi_size > iomap_max_bio_size(&iter->iomap) - plen || in iomap_bio_read_folio_range() 168 bio.bi_iter.bi_sector = sector; in iomap_bio_read_folio_range_sync()
|
| H A D | ioend.c | 29 ioend->io_size = bio->bi_iter.bi_size; in iomap_init_ioend() 30 ioend->io_sector = bio->bi_iter.bi_sector; in iomap_init_ioend() 165 bio->bi_iter.bi_sector = iomap_sector(&wpc->iomap, pos); in iomap_alloc_ioend() 177 if (ioend->io_bio.bi_iter.bi_size > in iomap_can_add_to_ioend() 475 if (bio->bi_iter.bi_size <= max_len) in iomap_split_ioend()
|
| /linux/fs/btrfs/ |
| H A D | compression.c | 332 ASSERT(cb->bbio.bio.bi_iter.bi_size == ordered->disk_num_bytes); in btrfs_submit_compressed_write() 333 cb->bbio.bio.bi_iter.bi_sector = ordered->disk_bytenr >> SECTOR_SHIFT; in btrfs_submit_compressed_write() 377 u64 cur = cb->orig_bbio->file_offset + orig_bio->bi_iter.bi_size; in add_ra_bio_pages() 490 orig_bio->bi_iter.bi_sector) { in add_ra_bio_pages() 588 cb->len = bbio->bio.bi_iter.bi_size; in btrfs_submit_compressed_read() 612 ASSERT(cb->bbio.bio.bi_iter.bi_size == compressed_len); in btrfs_submit_compressed_read() 617 cb->len = bbio->bio.bi_iter.bi_size; in btrfs_submit_compressed_read() 618 cb->bbio.bio.bi_iter.bi_sector = bbio->bio.bi_iter.bi_sector; in btrfs_submit_compressed_read() 1255 bvec = bio_iter_iovec(orig_bio, orig_bio->bi_iter); in btrfs_decompress_buf2page() 1284 if (!orig_bio->bi_iter.bi_size) in btrfs_decompress_buf2page()
|
| H A D | raid56.c | 1207 bio->bi_iter.bi_size -= added; in bio_add_paddrs() 1260 u64 last_end = last->bi_iter.bi_sector << SECTOR_SHIFT; in rbio_add_io_paddrs() 1261 last_end += last->bi_iter.bi_size; in rbio_add_io_paddrs() 1279 bio->bi_iter.bi_sector = disk_start >> SECTOR_SHIFT; in rbio_add_io_paddrs() 1293 struct bvec_iter iter = bio->bi_iter; in index_one_bio() 1295 u32 offset = (bio->bi_iter.bi_sector << SECTOR_SHIFT) - in index_one_bio() 1342 trace_info->offset = (bio->bi_iter.bi_sector << SECTOR_SHIFT) - in bio_get_trace_info() 1548 u32 offset = (bio->bi_iter.bi_sector << SECTOR_SHIFT) - in set_rbio_range_error() 1555 bio->bi_iter.bi_size >> fs_info->sectorsize_bits); in set_rbio_range_error() 1563 if (bio->bi_iter.bi_size == 0) { in set_rbio_range_error() [all …]
|
| /linux/drivers/nvdimm/ |
| H A D | nd_virtio.c | 117 if (bio && bio->bi_iter.bi_sector != -1) { in async_pmem_flush() 125 child->bi_iter.bi_sector = -1; in async_pmem_flush()
|