Home
last modified time | relevance | path

Searched refs:bi_iter (Results 1 – 25 of 101) sorted by relevance

12345

/linux/drivers/md/
H A Ddm-bio-record.h26 struct bvec_iter bi_iter; member
37 bd->bi_iter = bio->bi_iter; in dm_bio_record()
49 bio->bi_iter = bd->bi_iter; in dm_bio_restore()
H A Ddm-flakey.c357 bio->bi_iter.bi_sector = flakey_map_sector(ti, bio->bi_iter.bi_sector); in flakey_map_bio()
432 struct bvec_iter iter = bio->bi_iter; in clone_bio()
434 if (unlikely(bio->bi_iter.bi_size > UIO_MAXIOV << PAGE_SHIFT)) in clone_bio()
437 size = bio->bi_iter.bi_size; in clone_bio()
446 clone->bi_iter.bi_sector = flakey_map_sector(ti, bio->bi_iter.bi_sector); in clone_bio()
511 pb->saved_iter = bio->bi_iter; in flakey_map()
557 clone->bi_iter); in flakey_map()
560 clone->bi_iter); in flakey_map()
H A Ddm-ebs-target.c50 sector_t end_sector = __block_mod(bio->bi_iter.bi_sector, ec->u_bs) + bio_sectors(bio); in __nr_blocks()
145 sector_t block, blocks, sector = bio->bi_iter.bi_sector; in __ebs_discard_bio()
169 sector_t blocks, sector = bio->bi_iter.bi_sector; in __ebs_forget_bio()
195 block1 = __sector_to_block(ec, bio->bi_iter.bi_sector); in __ebs_process_bios()
200 if (__block_mod(bio->bi_iter.bi_sector, ec->u_bs)) in __ebs_process_bios()
366 bio->bi_iter.bi_sector = ec->start + dm_target_offset(ti, bio->bi_iter.bi_sector); in ebs_map()
375 if (likely(__block_mod(bio->bi_iter.bi_sector, ec->u_bs) || in ebs_map()
H A Ddm-stripe.c255 stripe_map_range_sector(sc, bio->bi_iter.bi_sector, in stripe_map_range()
261 bio->bi_iter.bi_sector = begin + in stripe_map_range()
263 bio->bi_iter.bi_size = to_bytes(end - begin); in stripe_map_range()
292 stripe_map_sector(sc, bio->bi_iter.bi_sector, in stripe_map()
293 &stripe, &bio->bi_iter.bi_sector); in stripe_map()
295 bio->bi_iter.bi_sector += sc->stripe[stripe].physical_start; in stripe_map()
H A Dmd-linear.c239 sector_t bio_sector = bio->bi_iter.bi_sector; in linear_make_request()
270 bio->bi_iter.bi_sector = bio->bi_iter.bi_sector - in linear_make_request()
289 (unsigned long long)bio->bi_iter.bi_sector, in linear_make_request()
H A Ddm-log-writes.c222 bio->bi_iter.bi_size = 0; in write_metadata()
223 bio->bi_iter.bi_sector = sector; in write_metadata()
275 bio->bi_iter.bi_size = 0; in write_inline_data()
276 bio->bi_iter.bi_sector = sector; in write_inline_data()
356 bio->bi_iter.bi_size = 0; in log_one_block()
357 bio->bi_iter.bi_sector = sector; in log_one_block()
374 bio->bi_iter.bi_size = 0; in log_one_block()
375 bio->bi_iter.bi_sector = sector; in log_one_block()
704 block->sector = bio_to_dev_sectors(lc, bio->bi_iter.bi_sector); in log_writes_map()
H A Ddm-io-rewind.c140 dm_bio_rewind_iter(bio, &bio->bi_iter, bytes); in dm_bio_rewind()
154 orig->bi_iter.bi_size)); in dm_io_rewind()
H A Draid10.c1131 bio->bi_iter.bi_sector < conf->reshape_progress && in regular_request_wait()
1132 bio->bi_iter.bi_sector + sectors > conf->reshape_progress) { in regular_request_wait()
1140 conf->reshape_progress <= bio->bi_iter.bi_sector || in regular_request_wait()
1141 conf->reshape_progress >= bio->bi_iter.bi_sector + in regular_request_wait()
1231 read_bio->bi_iter.bi_sector = r10_bio->devs[slot].addr + in raid10_read_request()
1266 mbio->bi_iter.bi_sector = (r10_bio->devs[n_copy].addr + in raid10_write_one_disk()
1354 bio->bi_iter.bi_sector, in raid10_write_request()
1366 bio->bi_iter.bi_sector, bio_end_sector(bio))) in raid10_write_request()
1381 ? (bio->bi_iter.bi_sector < conf->reshape_safe && in raid10_write_request()
1382 bio->bi_iter.bi_sector + sectors > conf->reshape_progress) in raid10_write_request()
[all …]
H A Ddm-dust.c230 bio->bi_iter.bi_sector = dd->start + dm_target_offset(ti, bio->bi_iter.bi_sector); in dust_map()
233 r = dust_map_read(dd, bio->bi_iter.bi_sector, dd->fail_read_on_bb); in dust_map()
235 r = dust_map_write(dd, bio->bi_iter.bi_sector, dd->fail_read_on_bb); in dust_map()
H A Ddm-unstripe.c120 sector_t sector = dm_target_offset(ti, bio->bi_iter.bi_sector); in map_to_core()
140 bio->bi_iter.bi_sector = map_to_core(ti, bio) + uc->physical_start; in unstripe_map()
/linux/block/
H A Dbio.c229 bio->bi_iter.bi_sector = 0; in bio_init()
230 bio->bi_iter.bi_size = 0; in bio_init()
231 bio->bi_iter.bi_idx = 0; in bio_init()
232 bio->bi_iter.bi_bvec_done = 0; in bio_init()
316 bio->bi_iter.bi_size += bio->bi_io_vec[i].bv_len; in bio_reuse()
665 if (new_size >= bio->bi_iter.bi_size) in bio_truncate()
695 bio->bi_iter.bi_size = new_size; in bio_truncate()
722 if (unlikely(bio->bi_iter.bi_sector >= maxsector)) in guard_bio_eod()
725 maxsector -= bio->bi_iter.bi_sector; in guard_bio_eod()
726 if (likely((bio->bi_iter.bi_size >> 9) <= maxsector)) in guard_bio_eod()
[all …]
H A Dblk-core.c545 if (should_fail_request(bdev_whole(bio->bi_bdev), bio->bi_iter.bi_size)) in should_fail_bio()
563 bio->bi_iter.bi_sector > maxsector - nr_sectors)) { in bio_check_eod()
569 bio->bi_iter.bi_sector, nr_sectors, maxsector); in bio_check_eod()
582 if (unlikely(should_fail_request(p, bio->bi_iter.bi_size))) in blk_partition_remap()
585 bio->bi_iter.bi_sector += p->bd_start_sect; in blk_partition_remap()
587 bio->bi_iter.bi_sector - in blk_partition_remap()
607 if (!bdev_is_zone_start(bio->bi_bdev, bio->bi_iter.bi_sector)) in blk_check_zone_append()
762 if (bio->bi_iter.bi_size > queue_atomic_write_unit_max_bytes(q)) in blk_validate_atomic_write_op_size()
765 if (bio->bi_iter.bi_size % queue_atomic_write_unit_min_bytes(q)) in blk_validate_atomic_write_op_size()
919 task_io_account_read(bio->bi_iter.bi_size); in submit_bio()
H A Dblk-crypto-fallback.c189 bio->bi_iter.bi_sector = bio_src->bi_iter.bi_sector; in blk_crypto_alloc_enc_bio()
278 bio_iter_iovec(src_bio, src_bio->bi_iter); in __blk_crypto_fallback_encrypt_bio()
314 bio_advance_iter_single(src_bio, &src_bio->bi_iter, in __blk_crypto_fallback_encrypt_bio()
316 if (!src_bio->bi_iter.bi_size) in __blk_crypto_fallback_encrypt_bio()
518 f_ctx->crypt_iter = bio->bi_iter; in blk_crypto_fallback_bio_prep()
H A Dfops.c80 bio.bi_iter.bi_sector = pos >> SECTOR_SHIFT; in __blkdev_direct_IO_simple()
90 ret = bio.bi_iter.bi_size; in __blkdev_direct_IO_simple()
212 bio->bi_iter.bi_sector = pos >> SECTOR_SHIFT; in __blkdev_direct_IO()
250 task_io_account_write(bio->bi_iter.bi_size); in __blkdev_direct_IO()
252 dio->size += bio->bi_iter.bi_size; in __blkdev_direct_IO()
253 pos += bio->bi_iter.bi_size; in __blkdev_direct_IO()
338 bio->bi_iter.bi_sector = pos >> SECTOR_SHIFT; in __blkdev_direct_IO_async()
357 dio->size = bio->bi_iter.bi_size; in __blkdev_direct_IO_async()
365 task_io_account_write(bio->bi_iter.bi_size); in __blkdev_direct_IO_async()
/linux/drivers/md/bcache/
H A Drequest.c114 bio_sectors(bio), (uint64_t) bio->bi_iter.bi_sector); in bch_data_invalidate()
123 bio->bi_iter.bi_sector += sectors; in bch_data_invalidate()
124 bio->bi_iter.bi_size -= sectors << 9; in bch_data_invalidate()
128 bio->bi_iter.bi_sector, in bch_data_invalidate()
220 SET_KEY_OFFSET(k, bio->bi_iter.bi_sector); in CLOSURE_CALLBACK()
411 if (bio->bi_iter.bi_sector & (c->cache->sb.block_size - 1) || in check_should_bypass()
430 hlist_for_each_entry(i, iohash(dc, bio->bi_iter.bi_sector), hash) in check_should_bypass()
431 if (i->last == bio->bi_iter.bi_sector && in check_should_bypass()
440 if (i->sequential + bio->bi_iter.bi_size > i->sequential) in check_should_bypass()
441 i->sequential += bio->bi_iter.bi_size; in check_should_bypass()
[all …]
H A Ddebug.c54 bio->bi_iter.bi_sector = PTR_OFFSET(&b->key, 0); in bch_btree_verify()
55 bio->bi_iter.bi_size = KEY_SIZE(&v->key) << 9; in bch_btree_verify()
119 check->bi_iter.bi_sector = bio->bi_iter.bi_sector; in bch_data_verify()
120 check->bi_iter.bi_size = bio->bi_iter.bi_size; in bch_data_verify()
140 (uint64_t) bio->bi_iter.bi_sector); in bch_data_verify()
/linux/include/trace/events/
H A Dbcache.h28 __entry->sector = bio->bi_iter.bi_sector;
29 __entry->orig_sector = bio->bi_iter.bi_sector - 16;
30 __entry->nr_sector = bio->bi_iter.bi_size >> 9;
103 __entry->sector = bio->bi_iter.bi_sector;
104 __entry->nr_sector = bio->bi_iter.bi_size >> 9;
138 __entry->sector = bio->bi_iter.bi_sector;
139 __entry->nr_sector = bio->bi_iter.bi_size >> 9;
169 __entry->sector = bio->bi_iter.bi_sector;
170 __entry->nr_sector = bio->bi_iter.bi_size >> 9;
238 __entry->sector = bio->bi_iter.bi_sector;
[all …]
H A Dblock.h323 __entry->sector = bio->bi_iter.bi_sector;
351 __entry->sector = bio->bi_iter.bi_sector;
504 __entry->sector = bio->bi_iter.bi_sector;
543 __entry->sector = bio->bi_iter.bi_sector;
626 __entry->sector = bio->bi_iter.bi_sector;
/linux/drivers/block/
H A Dbrd.c140 struct bio_vec bv = bio_iter_iovec(bio, bio->bi_iter); in brd_rw_bvec()
141 sector_t sector = bio->bi_iter.bi_sector; in brd_rw_bvec()
167 bio_advance_iter_single(bio, &bio->bi_iter, bv.bv_len); in brd_rw_bvec()
207 brd_do_discard(brd, bio->bi_iter.bi_sector, in brd_submit_bio()
208 bio->bi_iter.bi_size); in brd_submit_bio()
216 } while (bio->bi_iter.bi_size); in brd_submit_bio()
/linux/fs/erofs/
H A Dfileio.c28 if (ret >= 0 && ret != rq->bio.bi_iter.bi_size) in erofs_fileio_ki_complete()
51 rq->iocb.ki_pos = rq->bio.bi_iter.bi_sector << SECTOR_SHIFT; in erofs_fileio_rq_submit()
58 rq->bio.bi_iter.bi_size); in erofs_fileio_rq_submit()
141 io->rq->bio.bi_iter.bi_sector = in erofs_fileio_scan_folio()
/linux/fs/iomap/
H A Dbio.c129 bio->bi_iter.bi_sector = iomap_sector(iomap, iter->pos); in iomap_read_alloc_bio()
145 bio->bi_iter.bi_size > iomap_max_bio_size(&iter->iomap) - plen || in iomap_bio_read_folio_range()
168 bio.bi_iter.bi_sector = sector; in iomap_bio_read_folio_range_sync()
H A Dioend.c29 ioend->io_size = bio->bi_iter.bi_size; in iomap_init_ioend()
30 ioend->io_sector = bio->bi_iter.bi_sector; in iomap_init_ioend()
165 bio->bi_iter.bi_sector = iomap_sector(&wpc->iomap, pos); in iomap_alloc_ioend()
177 if (ioend->io_bio.bi_iter.bi_size > in iomap_can_add_to_ioend()
475 if (bio->bi_iter.bi_size <= max_len) in iomap_split_ioend()
/linux/fs/btrfs/
H A Dcompression.c332 ASSERT(cb->bbio.bio.bi_iter.bi_size == ordered->disk_num_bytes); in btrfs_submit_compressed_write()
333 cb->bbio.bio.bi_iter.bi_sector = ordered->disk_bytenr >> SECTOR_SHIFT; in btrfs_submit_compressed_write()
377 u64 cur = cb->orig_bbio->file_offset + orig_bio->bi_iter.bi_size; in add_ra_bio_pages()
490 orig_bio->bi_iter.bi_sector) { in add_ra_bio_pages()
588 cb->len = bbio->bio.bi_iter.bi_size; in btrfs_submit_compressed_read()
612 ASSERT(cb->bbio.bio.bi_iter.bi_size == compressed_len); in btrfs_submit_compressed_read()
617 cb->len = bbio->bio.bi_iter.bi_size; in btrfs_submit_compressed_read()
618 cb->bbio.bio.bi_iter.bi_sector = bbio->bio.bi_iter.bi_sector; in btrfs_submit_compressed_read()
1255 bvec = bio_iter_iovec(orig_bio, orig_bio->bi_iter); in btrfs_decompress_buf2page()
1284 if (!orig_bio->bi_iter.bi_size) in btrfs_decompress_buf2page()
H A Draid56.c1207 bio->bi_iter.bi_size -= added; in bio_add_paddrs()
1260 u64 last_end = last->bi_iter.bi_sector << SECTOR_SHIFT; in rbio_add_io_paddrs()
1261 last_end += last->bi_iter.bi_size; in rbio_add_io_paddrs()
1279 bio->bi_iter.bi_sector = disk_start >> SECTOR_SHIFT; in rbio_add_io_paddrs()
1293 struct bvec_iter iter = bio->bi_iter; in index_one_bio()
1295 u32 offset = (bio->bi_iter.bi_sector << SECTOR_SHIFT) - in index_one_bio()
1342 trace_info->offset = (bio->bi_iter.bi_sector << SECTOR_SHIFT) - in bio_get_trace_info()
1548 u32 offset = (bio->bi_iter.bi_sector << SECTOR_SHIFT) - in set_rbio_range_error()
1555 bio->bi_iter.bi_size >> fs_info->sectorsize_bits); in set_rbio_range_error()
1563 if (bio->bi_iter.bi_size == 0) { in set_rbio_range_error()
[all …]
/linux/drivers/nvdimm/
H A Dnd_virtio.c117 if (bio && bio->bi_iter.bi_sector != -1) { in async_pmem_flush()
125 child->bi_iter.bi_sector = -1; in async_pmem_flush()

12345