| /linux/drivers/block/ |
| H A D | n64cart.c | 69 (bv->bv_len & (MIN_ALIGNMENT - 1))); in n64cart_do_bvec() 79 n64cart_write_reg(PI_WRITE_REG, bv->bv_len - 1); in n64cart_do_bvec() 83 dma_unmap_page(dev, dma_addr, bv->bv_len, DMA_FROM_DEVICE); in n64cart_do_bvec() 99 pos += bvec.bv_len; in n64cart_submit_bio()
|
| H A D | brd.c | 147 bv.bv_len = min_t(u32, bv.bv_len, PAGE_SIZE - offset); in brd_rw_bvec() 158 memcpy_to_page(page, offset, kaddr, bv.bv_len); in brd_rw_bvec() 161 memcpy_from_page(kaddr, page, offset, bv.bv_len); in brd_rw_bvec() 163 memset(kaddr, 0, bv.bv_len); in brd_rw_bvec() 167 bio_advance_iter_single(bio, &bio->bi_iter, bv.bv_len); in brd_rw_bvec()
|
| /linux/drivers/md/ |
| H A D | dm-ebs-target.c | 71 unsigned int bv_len = bv->bv_len; in __ebs_rw_bvec() local 76 if (unlikely(!bv->bv_page || !bv_len)) in __ebs_rw_bvec() 82 while (bv_len) { in __ebs_rw_bvec() 83 cur_len = min(dm_bufio_get_block_size(ec->bufio) - buf_off, bv_len); in __ebs_rw_bvec() 86 if (op == REQ_OP_READ || buf_off || bv_len < dm_bufio_get_block_size(ec->bufio)) in __ebs_rw_bvec() 113 bv_len -= cur_len; in __ebs_rw_bvec()
|
| H A D | dm-io-rewind.c | 27 while (idx >= 0 && bytes && bytes > bv[idx].bv_len) { in dm_bvec_iter_rewind() 28 bytes -= bv[idx].bv_len; in dm_bvec_iter_rewind() 41 iter->bi_bvec_done = bv[idx].bv_len - bytes; in dm_bvec_iter_rewind()
|
| H A D | dm-log-writes.c | 367 block->vecs[i].bv_len, 0); in log_one_block() 368 if (ret != block->vecs[i].bv_len) { in log_one_block() 380 block->vecs[i].bv_len, 0); in log_one_block() 381 if (ret != block->vecs[i].bv_len) { in log_one_block() 387 sector += block->vecs[i].bv_len >> SECTOR_SHIFT; in log_one_block() 751 block->vecs[i].bv_len = bv.bv_len; in log_writes_map()
|
| /linux/block/ |
| H A D | t10-pi.c | 148 for (j = 0; j < iv.bv_len; j += tuple_sz) { in t10_pi_type1_prepare() 196 for (j = 0; j < iv.bv_len && intervals; j += tuple_sz) { in t10_pi_type1_complete() 322 for (j = 0; j < iv.bv_len; j += tuple_sz) { in ext_pi_type1_prepare() 359 for (j = 0; j < iv.bv_len && intervals; j += tuple_sz) { in ext_pi_type1_complete() 391 iter.data_size = bv.bv_len; in blk_integrity_generate() 428 iter.data_size = bv.bv_len; in blk_integrity_verify_iter()
|
| H A D | blk-crypto-fallback.c | 281 if (!IS_ALIGNED(src_bv.bv_len | src_bv.bv_offset, in __blk_crypto_fallback_encrypt_bio() 287 __bio_add_page(enc_bio, enc_page, src_bv.bv_len, in __blk_crypto_fallback_encrypt_bio() 303 for (i = 0; i < src_bv.bv_len; i += data_unit_size) { in __blk_crypto_fallback_encrypt_bio() 315 src_bv.bv_len); in __blk_crypto_fallback_encrypt_bio() 398 if (!IS_ALIGNED(bv.bv_len | bv.bv_offset, data_unit_size)) in __blk_crypto_fallback_decrypt_bio() 404 for (i = 0; i < bv.bv_len; i += data_unit_size) { in __blk_crypto_fallback_decrypt_bio()
|
| H A D | blk-map.c | 77 bvec->bv_len, in bio_copy_from_iter() 83 if (ret < bvec->bv_len) in bio_copy_from_iter() 108 bvec->bv_len, in bio_copy_to_iter() 114 if (ret < bvec->bv_len) in bio_copy_to_iter() 299 len += bio->bi_io_vec[i].bv_len; in bio_invalidate_vmalloc_pages() 348 p += bvec->bv_len; in bio_copy_kern_endio_read()
|
| H A D | bio.c | 344 bio->bi_iter.bi_size += bio->bi_io_vec[i].bv_len; in bio_reuse() 699 if (done + bv.bv_len > new_size) { in bio_truncate() 707 bv.bv_len - offset); in bio_truncate() 710 done += bv.bv_len; in bio_truncate() 969 size_t bv_end = bv->bv_offset + bv->bv_len; in bvec_try_merge_page() 985 bv->bv_len += len; in bvec_try_merge_page() 1006 if (len > queue_max_segment_size(q) - bv->bv_len) in bvec_try_merge_hw_page() 1227 if (nbytes < bv->bv_len) { in bio_iov_iter_align_down() 1228 bv->bv_len -= nbytes; in bio_iov_iter_align_down() 1236 nbytes -= bv->bv_len; in bio_iov_iter_align_down() [all …]
|
| H A D | blk-integrity.c | 41 if (seg_size + iv.bv_len > queue_max_segment_size(q)) in blk_rq_count_integrity_sg() 44 seg_size += iv.bv_len; in blk_rq_count_integrity_sg() 48 seg_size = iv.bv_len; in blk_rq_count_integrity_sg()
|
| /linux/drivers/md/bcache/ |
| H A D | util.c | 244 start: bv->bv_len = min_t(size_t, PAGE_SIZE - bv->bv_offset, in bch_bio_map() 251 base += bv->bv_len; in bch_bio_map() 254 size -= bv->bv_len; in bch_bio_map()
|
| H A D | debug.c | 136 cache_set_err_on(memcmp(p1, p2, bv.bv_len), in bch_data_verify() 144 bio_advance_iter(check, &citer, bv.bv_len); in bch_data_verify()
|
| /linux/drivers/s390/block/ |
| H A D | dasd_fba.c | 457 if (bv.bv_len & (blksize - 1)) in dasd_fba_build_cp_regular() 460 count += bv.bv_len >> (block->s2b_shift + 9); in dasd_fba_build_cp_regular() 461 if (idal_is_needed (page_address(bv.bv_page), bv.bv_len)) in dasd_fba_build_cp_regular() 462 cidaw += bv.bv_len / blksize; in dasd_fba_build_cp_regular() 504 memcpy(copy + bv.bv_offset, dst, bv.bv_len); in dasd_fba_build_cp_regular() 508 for (off = 0; off < bv.bv_len; off += blksize) { in dasd_fba_build_cp_regular() 582 for (off = 0; off < bv.bv_len; off += blksize) { in dasd_fba_free_cp() 593 memcpy(dst, cda, bv.bv_len); in dasd_fba_free_cp()
|
| H A D | dasd_diag.c | 537 if (bv.bv_len & (blksize - 1)) in dasd_diag_build_cp() 540 count += bv.bv_len >> (block->s2b_shift + 9); in dasd_diag_build_cp() 557 for (off = 0; off < bv.bv_len; off += blksize) { in dasd_diag_build_cp()
|
| H A D | dcssblk.c | 918 !IS_ALIGNED(bvec.bv_len, PAGE_SIZE))) in dcssblk_submit_bio() 922 memcpy(page_addr, __va(source_addr), bvec.bv_len); in dcssblk_submit_bio() 924 memcpy(__va(source_addr), page_addr, bvec.bv_len); in dcssblk_submit_bio() 925 bytes_done += bvec.bv_len; in dcssblk_submit_bio()
|
| /linux/net/ceph/ |
| H A D | messenger_v2.c | 156 it->bvec->bv_len - it->iov_offset), in do_try_sendpage() 173 iov_iter_bvec(&msg.msg_iter, ITER_SOURCE, &bv, 1, bv.bv_len); in do_try_sendpage() 237 iov_iter_bvec(&con->v2.in_iter, ITER_DEST, &con->v2.in_bvec, 1, bv->bv_len); in set_in_bvec() 282 con->v2.out_bvec.bv_len); in set_out_bvec() 294 con->v2.out_bvec.bv_len); in set_out_bvec_zero() 881 ceph_msg_data_advance(cursor, bv.bv_len); in calc_sg_cnt_cursor() 929 sg_set_page(*sg, bv.bv_page, bv.bv_len, bv.bv_offset); in init_sgs_cursor() 932 ceph_msg_data_advance(cursor, bv.bv_len); in init_sgs_cursor() 1091 len = min_t(int, len, bv.bv_len); in process_v2_sparse_read() 1866 con->v2.in_bvec.bv_len); in prepare_read_data_cont() [all …]
|
| /linux/fs/squashfs/ |
| H A D | block.c | 50 int bytes_to_copy = min_t(int, bvec->bv_len - offset, in copy_bio_to_actor() 69 if (offset >= bvec->bv_len) { in copy_bio_to_actor() 330 if (offset < bvec->bv_len - 1) { in squashfs_read_data()
|
| H A D | lzo_wrapper.c | 77 int avail = min(bytes, ((int)bvec->bv_len) - offset); in lzo_uncompress()
|
| H A D | zlib_wrapper.c | 83 avail = min(length, ((int)bvec->bv_len) - offset); in zlib_uncompress()
|
| /linux/lib/ |
| H A D | iov_iter.c | 513 if (likely(size < bvec->bv_len)) in iov_iter_bvec_advance() 515 size -= bvec->bv_len; in iov_iter_bvec_advance() 642 size_t n = (--bvec)->bv_len; in iov_iter_revert() 679 return min(i->count, i->bvec->bv_len - i->iov_offset); in iov_iter_single_seg_count() 829 size_t len = bvec->bv_len - skip; in iov_iter_alignment_bvec() 1054 len = i->bvec->bv_len - skip; in first_bvec_segment() 1116 if (i->iov_offset == i->bvec->bv_len) { in __iov_iter_get_pages_alloc() 1185 size_t len = min(p->bv_len - skip, size); in bvec_npages() 1620 if (i->iov_offset == i->bvec->bv_len) { in iov_iter_extract_bvec_pages() 1649 size += bv.bv_len; in iov_iter_extract_bvec_pages() [all …]
|
| /linux/drivers/infiniband/core/ |
| H A D | rw.c | 238 sg_set_page(sg, bv.bv_page, bv.bv_len, bv.bv_offset); in rdma_rw_init_mr_wrs_bvec() 239 bvec_iter_advance(bvecs, iter, bv.bv_len); in rdma_rw_init_mr_wrs_bvec() 387 ctx->single.sge.length = bv.bv_len; in rdma_rw_init_single_wr_bvec() 453 sge->length = bv.bv_len; in rdma_rw_init_map_wrs_bvec() 456 total_len += bv.bv_len; in rdma_rw_init_map_wrs_bvec() 459 bvec_iter_advance_single(bvecs, iter, bv.bv_len); in rdma_rw_init_map_wrs_bvec() 514 mapped_len, bv.bv_len, dir, 0); in rdma_rw_init_iova_wrs_bvec() 518 mapped_len += bv.bv_len; in rdma_rw_init_iova_wrs_bvec() 519 bvec_iter_advance(bvec, iter, bv.bv_len); in rdma_rw_init_iova_wrs_bvec()
|
| /linux/drivers/xen/ |
| H A D | biomerge.c | 15 return bfn1 + PFN_DOWN(vec1->bv_offset + vec1->bv_len) == bfn2; in xen_biovec_phys_mergeable()
|
| /linux/fs/netfs/ |
| H A D | iterator.c | 123 len = bvecs[ix].bv_len; in netfs_limit_bvec() 132 len = min3(n, bvecs[ix].bv_len - skip, max_size); in netfs_limit_bvec()
|
| /linux/include/linux/ |
| H A D | bio.h | 162 bio_advance_iter_single((bio), &(iter), (bvl).bv_len)) 171 bio_advance_iter_single((bio), &(iter), (bvl).bv_len)) 185 #define bio_iter_last(bvec, iter) ((iter).bi_size == (bvec).bv_len) 289 fi->_seg_count = bvec->bv_len; in bio_first_folio()
|
| /linux/fs/erofs/ |
| H A D | zdata.c | 1502 bvec->bv_len = PAGE_SIZE; in z_erofs_fill_bio_vec() 1528 bvec->bv_len = round_up(zbv.end, bs) - bvec->bv_offset; in z_erofs_fill_bio_vec() 1734 if (cur + bvec.bv_len > end) in z_erofs_submit_queue() 1735 bvec.bv_len = end - cur; in z_erofs_submit_queue() 1736 DBG_BUGON(bvec.bv_len < sb->s_blocksize); in z_erofs_submit_queue() 1762 if (!bio_add_page(bio, bvec.bv_page, bvec.bv_len, in z_erofs_submit_queue() 1765 last_pa = cur + bvec.bv_len; in z_erofs_submit_queue() 1767 } while ((cur += bvec.bv_len) < end); in z_erofs_submit_queue()
|